{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9866363628355225, "eval_steps": 500, "global_step": 28000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 3.523701295841151e-05, "grad_norm": 0.0, "learning_rate": 0.0, "loss": 3.1715, "step": 1 }, { "epoch": 7.047402591682303e-05, "grad_norm": 0.0, "learning_rate": 0.0, "loss": 7.8633, "step": 2 }, { "epoch": 0.00010571103887523455, "grad_norm": 7.17532205581665, "learning_rate": 2.347417840375587e-08, "loss": 2.2478, "step": 3 }, { "epoch": 0.00014094805183364605, "grad_norm": 7.17532205581665, "learning_rate": 2.347417840375587e-08, "loss": 6.189, "step": 4 }, { "epoch": 0.00017618506479205758, "grad_norm": 7.17532205581665, "learning_rate": 2.347417840375587e-08, "loss": 11.3102, "step": 5 }, { "epoch": 0.0002114220777504691, "grad_norm": 10.80892276763916, "learning_rate": 4.694835680751174e-08, "loss": 7.0578, "step": 6 }, { "epoch": 0.0002466590907088806, "grad_norm": 10.80892276763916, "learning_rate": 4.694835680751174e-08, "loss": 7.3608, "step": 7 }, { "epoch": 0.0002818961036672921, "grad_norm": 11.575017929077148, "learning_rate": 7.042253521126761e-08, "loss": 3.2351, "step": 8 }, { "epoch": 0.00031713311662570366, "grad_norm": 16.47764778137207, "learning_rate": 9.389671361502348e-08, "loss": 3.7347, "step": 9 }, { "epoch": 0.00035237012958411516, "grad_norm": 11.561605453491211, "learning_rate": 1.1737089201877935e-07, "loss": 3.0588, "step": 10 }, { "epoch": 0.00038760714254252666, "grad_norm": 8.687694549560547, "learning_rate": 1.4084507042253522e-07, "loss": 2.9906, "step": 11 }, { "epoch": 0.0004228441555009382, "grad_norm": 10.123542785644531, "learning_rate": 1.643192488262911e-07, "loss": 2.6714, "step": 12 }, { "epoch": 0.0004580811684593497, "grad_norm": 13.044445037841797, "learning_rate": 1.8779342723004696e-07, "loss": 3.2261, "step": 13 }, { "epoch": 0.0004933181814177612, "grad_norm": 13.212477684020996, "learning_rate": 2.1126760563380284e-07, "loss": 7.5194, "step": 14 }, { "epoch": 0.0005285551943761727, "grad_norm": 13.896424293518066, "learning_rate": 2.347417840375587e-07, "loss": 12.4173, "step": 15 }, { "epoch": 0.0005637922073345842, "grad_norm": 21.609113693237305, "learning_rate": 2.582159624413146e-07, "loss": 3.2692, "step": 16 }, { "epoch": 0.0005990292202929958, "grad_norm": 8.717239379882812, "learning_rate": 2.8169014084507043e-07, "loss": 2.8184, "step": 17 }, { "epoch": 0.0006342662332514073, "grad_norm": 9.326610565185547, "learning_rate": 3.051643192488263e-07, "loss": 2.5868, "step": 18 }, { "epoch": 0.0006695032462098188, "grad_norm": 9.77458381652832, "learning_rate": 3.286384976525822e-07, "loss": 6.9621, "step": 19 }, { "epoch": 0.0007047402591682303, "grad_norm": 8.224367141723633, "learning_rate": 3.521126760563381e-07, "loss": 2.6936, "step": 20 }, { "epoch": 0.0007399772721266418, "grad_norm": 19.591176986694336, "learning_rate": 3.755868544600939e-07, "loss": 10.9086, "step": 21 }, { "epoch": 0.0007752142850850533, "grad_norm": 12.316106796264648, "learning_rate": 3.990610328638498e-07, "loss": 3.1742, "step": 22 }, { "epoch": 0.0008104512980434648, "grad_norm": 23.561098098754883, "learning_rate": 4.225352112676057e-07, "loss": 11.2532, "step": 23 }, { "epoch": 0.0008456883110018764, "grad_norm": 11.277469635009766, "learning_rate": 4.460093896713615e-07, "loss": 7.0606, "step": 24 }, { "epoch": 0.0008809253239602879, "grad_norm": 11.298648834228516, "learning_rate": 4.694835680751174e-07, "loss": 7.6837, "step": 25 }, { "epoch": 0.0009161623369186994, "grad_norm": 11.167600631713867, "learning_rate": 4.929577464788733e-07, "loss": 2.6069, "step": 26 }, { "epoch": 0.0009513993498771109, "grad_norm": 12.619842529296875, "learning_rate": 5.164319248826292e-07, "loss": 2.5448, "step": 27 }, { "epoch": 0.0009866363628355224, "grad_norm": 10.350058555603027, "learning_rate": 5.39906103286385e-07, "loss": 5.6966, "step": 28 }, { "epoch": 0.001021873375793934, "grad_norm": 12.684341430664062, "learning_rate": 5.633802816901409e-07, "loss": 7.2286, "step": 29 }, { "epoch": 0.0010571103887523454, "grad_norm": 15.365835189819336, "learning_rate": 5.868544600938968e-07, "loss": 11.8556, "step": 30 }, { "epoch": 0.001092347401710757, "grad_norm": 7.850368499755859, "learning_rate": 6.103286384976526e-07, "loss": 2.1907, "step": 31 }, { "epoch": 0.0011275844146691684, "grad_norm": 14.248165130615234, "learning_rate": 6.338028169014085e-07, "loss": 11.2266, "step": 32 }, { "epoch": 0.00116282142762758, "grad_norm": 10.94234848022461, "learning_rate": 6.572769953051644e-07, "loss": 1.619, "step": 33 }, { "epoch": 0.0011980584405859916, "grad_norm": 12.080102920532227, "learning_rate": 6.807511737089202e-07, "loss": 10.9008, "step": 34 }, { "epoch": 0.0012332954535444031, "grad_norm": 2.894597291946411, "learning_rate": 7.042253521126762e-07, "loss": 1.006, "step": 35 }, { "epoch": 0.0012685324665028146, "grad_norm": 11.45083999633789, "learning_rate": 7.27699530516432e-07, "loss": 5.7835, "step": 36 }, { "epoch": 0.0013037694794612261, "grad_norm": 8.906060218811035, "learning_rate": 7.511737089201878e-07, "loss": 6.4892, "step": 37 }, { "epoch": 0.0013390064924196376, "grad_norm": 7.098948001861572, "learning_rate": 7.746478873239437e-07, "loss": 6.1666, "step": 38 }, { "epoch": 0.0013742435053780491, "grad_norm": 3.077409029006958, "learning_rate": 7.981220657276996e-07, "loss": 1.4718, "step": 39 }, { "epoch": 0.0014094805183364606, "grad_norm": 7.718149662017822, "learning_rate": 8.215962441314555e-07, "loss": 5.2575, "step": 40 }, { "epoch": 0.0014447175312948721, "grad_norm": 8.897134780883789, "learning_rate": 8.450704225352114e-07, "loss": 6.0578, "step": 41 }, { "epoch": 0.0014799545442532836, "grad_norm": 9.529160499572754, "learning_rate": 8.685446009389673e-07, "loss": 9.724, "step": 42 }, { "epoch": 0.0015151915572116951, "grad_norm": 3.628211259841919, "learning_rate": 8.92018779342723e-07, "loss": 1.3782, "step": 43 }, { "epoch": 0.0015504285701701066, "grad_norm": 3.370419979095459, "learning_rate": 9.154929577464789e-07, "loss": 1.2991, "step": 44 }, { "epoch": 0.0015856655831285181, "grad_norm": 11.018994331359863, "learning_rate": 9.389671361502348e-07, "loss": 10.1218, "step": 45 }, { "epoch": 0.0016209025960869296, "grad_norm": 5.179286479949951, "learning_rate": 9.624413145539907e-07, "loss": 5.7039, "step": 46 }, { "epoch": 0.0016561396090453411, "grad_norm": 2.2075798511505127, "learning_rate": 9.859154929577465e-07, "loss": 1.4643, "step": 47 }, { "epoch": 0.0016913766220037528, "grad_norm": 3.777865171432495, "learning_rate": 1.0093896713615024e-06, "loss": 1.5029, "step": 48 }, { "epoch": 0.0017266136349621643, "grad_norm": 9.669559478759766, "learning_rate": 1.0328638497652584e-06, "loss": 9.267, "step": 49 }, { "epoch": 0.0017618506479205758, "grad_norm": 4.27978515625, "learning_rate": 1.0563380281690142e-06, "loss": 1.2694, "step": 50 }, { "epoch": 0.0017970876608789873, "grad_norm": 5.254648208618164, "learning_rate": 1.07981220657277e-06, "loss": 4.848, "step": 51 }, { "epoch": 0.0018323246738373988, "grad_norm": 4.982268333435059, "learning_rate": 1.103286384976526e-06, "loss": 1.554, "step": 52 }, { "epoch": 0.0018675616867958103, "grad_norm": 2.7640011310577393, "learning_rate": 1.1267605633802817e-06, "loss": 1.4388, "step": 53 }, { "epoch": 0.0019027986997542218, "grad_norm": 15.176131248474121, "learning_rate": 1.1502347417840376e-06, "loss": 6.7051, "step": 54 }, { "epoch": 0.0019380357127126333, "grad_norm": 4.372971057891846, "learning_rate": 1.1737089201877936e-06, "loss": 1.6327, "step": 55 }, { "epoch": 0.001973272725671045, "grad_norm": 2.4785804748535156, "learning_rate": 1.1971830985915492e-06, "loss": 1.3768, "step": 56 }, { "epoch": 0.0020085097386294565, "grad_norm": 2.9967947006225586, "learning_rate": 1.2206572769953053e-06, "loss": 1.2832, "step": 57 }, { "epoch": 0.002043746751587868, "grad_norm": 5.7214837074279785, "learning_rate": 1.244131455399061e-06, "loss": 2.2408, "step": 58 }, { "epoch": 0.0020789837645462795, "grad_norm": 4.299997806549072, "learning_rate": 1.267605633802817e-06, "loss": 1.6737, "step": 59 }, { "epoch": 0.002114220777504691, "grad_norm": 5.979604721069336, "learning_rate": 1.291079812206573e-06, "loss": 5.0702, "step": 60 }, { "epoch": 0.0021494577904631025, "grad_norm": 2.105067491531372, "learning_rate": 1.3145539906103288e-06, "loss": 1.3554, "step": 61 }, { "epoch": 0.002184694803421514, "grad_norm": 8.889758110046387, "learning_rate": 1.3380281690140844e-06, "loss": 9.7338, "step": 62 }, { "epoch": 0.0022199318163799255, "grad_norm": 9.695563316345215, "learning_rate": 1.3615023474178405e-06, "loss": 9.702, "step": 63 }, { "epoch": 0.002255168829338337, "grad_norm": 2.7488269805908203, "learning_rate": 1.3849765258215963e-06, "loss": 1.1239, "step": 64 }, { "epoch": 0.0022904058422967485, "grad_norm": 5.22703742980957, "learning_rate": 1.4084507042253523e-06, "loss": 5.8515, "step": 65 }, { "epoch": 0.00232564285525516, "grad_norm": 3.564469575881958, "learning_rate": 1.4319248826291082e-06, "loss": 1.7331, "step": 66 }, { "epoch": 0.0023608798682135715, "grad_norm": 6.483605861663818, "learning_rate": 1.455399061032864e-06, "loss": 6.1122, "step": 67 }, { "epoch": 0.0023961168811719833, "grad_norm": 2.469683885574341, "learning_rate": 1.4788732394366198e-06, "loss": 1.48, "step": 68 }, { "epoch": 0.0024313538941303945, "grad_norm": 5.2667555809021, "learning_rate": 1.5023474178403756e-06, "loss": 5.6002, "step": 69 }, { "epoch": 0.0024665909070888063, "grad_norm": 10.12618350982666, "learning_rate": 1.5258215962441317e-06, "loss": 9.6181, "step": 70 }, { "epoch": 0.0025018279200472175, "grad_norm": 8.659031867980957, "learning_rate": 1.5492957746478873e-06, "loss": 8.9491, "step": 71 }, { "epoch": 0.0025370649330056293, "grad_norm": 3.828542947769165, "learning_rate": 1.5727699530516433e-06, "loss": 5.5023, "step": 72 }, { "epoch": 0.0025723019459640405, "grad_norm": 2.5202364921569824, "learning_rate": 1.5962441314553992e-06, "loss": 1.536, "step": 73 }, { "epoch": 0.0026075389589224523, "grad_norm": 2.6565699577331543, "learning_rate": 1.6197183098591552e-06, "loss": 1.5474, "step": 74 }, { "epoch": 0.0026427759718808635, "grad_norm": 2.546172857284546, "learning_rate": 1.643192488262911e-06, "loss": 1.6448, "step": 75 }, { "epoch": 0.0026780129848392752, "grad_norm": 2.0448689460754395, "learning_rate": 1.6666666666666667e-06, "loss": 1.1938, "step": 76 }, { "epoch": 0.0027132499977976865, "grad_norm": 13.103910446166992, "learning_rate": 1.6901408450704227e-06, "loss": 13.3063, "step": 77 }, { "epoch": 0.0027484870107560982, "grad_norm": 2.3271756172180176, "learning_rate": 1.7136150234741785e-06, "loss": 1.2093, "step": 78 }, { "epoch": 0.0027837240237145095, "grad_norm": 4.619046211242676, "learning_rate": 1.7370892018779346e-06, "loss": 5.3767, "step": 79 }, { "epoch": 0.0028189610366729212, "grad_norm": 7.51853609085083, "learning_rate": 1.7605633802816902e-06, "loss": 8.1744, "step": 80 }, { "epoch": 0.002854198049631333, "grad_norm": 4.495392322540283, "learning_rate": 1.784037558685446e-06, "loss": 4.8634, "step": 81 }, { "epoch": 0.0028894350625897442, "grad_norm": 2.330268621444702, "learning_rate": 1.807511737089202e-06, "loss": 1.4431, "step": 82 }, { "epoch": 0.002924672075548156, "grad_norm": 3.7343523502349854, "learning_rate": 1.8309859154929579e-06, "loss": 4.142, "step": 83 }, { "epoch": 0.0029599090885065672, "grad_norm": 2.346879720687866, "learning_rate": 1.854460093896714e-06, "loss": 1.393, "step": 84 }, { "epoch": 0.002995146101464979, "grad_norm": 8.298161506652832, "learning_rate": 1.8779342723004696e-06, "loss": 8.348, "step": 85 }, { "epoch": 0.0030303831144233902, "grad_norm": 2.0137407779693604, "learning_rate": 1.9014084507042254e-06, "loss": 1.1462, "step": 86 }, { "epoch": 0.003065620127381802, "grad_norm": 4.259990215301514, "learning_rate": 1.9248826291079814e-06, "loss": 4.538, "step": 87 }, { "epoch": 0.0031008571403402132, "grad_norm": 2.4016196727752686, "learning_rate": 1.9483568075117375e-06, "loss": 1.4571, "step": 88 }, { "epoch": 0.003136094153298625, "grad_norm": 2.5934789180755615, "learning_rate": 1.971830985915493e-06, "loss": 1.5469, "step": 89 }, { "epoch": 0.0031713311662570362, "grad_norm": 1.9990360736846924, "learning_rate": 1.995305164319249e-06, "loss": 1.0973, "step": 90 }, { "epoch": 0.003206568179215448, "grad_norm": 5.241186141967773, "learning_rate": 2.0187793427230047e-06, "loss": 4.7793, "step": 91 }, { "epoch": 0.0032418051921738592, "grad_norm": 4.900415897369385, "learning_rate": 2.0422535211267608e-06, "loss": 4.8954, "step": 92 }, { "epoch": 0.003277042205132271, "grad_norm": 7.073700428009033, "learning_rate": 2.065727699530517e-06, "loss": 8.3697, "step": 93 }, { "epoch": 0.0033122792180906822, "grad_norm": 6.839330196380615, "learning_rate": 2.0892018779342724e-06, "loss": 7.4914, "step": 94 }, { "epoch": 0.003347516231049094, "grad_norm": 3.0100083351135254, "learning_rate": 2.1126760563380285e-06, "loss": 1.5351, "step": 95 }, { "epoch": 0.0033827532440075057, "grad_norm": 6.993569850921631, "learning_rate": 2.136150234741784e-06, "loss": 7.9249, "step": 96 }, { "epoch": 0.003417990256965917, "grad_norm": 4.4020185470581055, "learning_rate": 2.15962441314554e-06, "loss": 4.8725, "step": 97 }, { "epoch": 0.0034532272699243287, "grad_norm": 3.2103800773620605, "learning_rate": 2.1830985915492958e-06, "loss": 3.3924, "step": 98 }, { "epoch": 0.00348846428288274, "grad_norm": 6.317807197570801, "learning_rate": 2.206572769953052e-06, "loss": 8.4028, "step": 99 }, { "epoch": 0.0035237012958411517, "grad_norm": 2.0183370113372803, "learning_rate": 2.230046948356808e-06, "loss": 1.2718, "step": 100 }, { "epoch": 0.003558938308799563, "grad_norm": 2.44419527053833, "learning_rate": 2.2535211267605635e-06, "loss": 1.3375, "step": 101 }, { "epoch": 0.0035941753217579747, "grad_norm": 2.0335288047790527, "learning_rate": 2.2769953051643195e-06, "loss": 1.3198, "step": 102 }, { "epoch": 0.003629412334716386, "grad_norm": 2.1068503856658936, "learning_rate": 2.300469483568075e-06, "loss": 0.8988, "step": 103 }, { "epoch": 0.0036646493476747977, "grad_norm": 1.8526614904403687, "learning_rate": 2.323943661971831e-06, "loss": 1.1061, "step": 104 }, { "epoch": 0.003699886360633209, "grad_norm": 4.082808971405029, "learning_rate": 2.347417840375587e-06, "loss": 5.0303, "step": 105 }, { "epoch": 0.0037351233735916207, "grad_norm": 4.153285503387451, "learning_rate": 2.370892018779343e-06, "loss": 4.7339, "step": 106 }, { "epoch": 0.003770360386550032, "grad_norm": 3.64225172996521, "learning_rate": 2.3943661971830984e-06, "loss": 4.4511, "step": 107 }, { "epoch": 0.0038055973995084437, "grad_norm": 4.054756164550781, "learning_rate": 2.4178403755868545e-06, "loss": 4.701, "step": 108 }, { "epoch": 0.0038408344124668554, "grad_norm": 3.940004587173462, "learning_rate": 2.4413145539906105e-06, "loss": 4.9178, "step": 109 }, { "epoch": 0.0038760714254252667, "grad_norm": 5.83101224899292, "learning_rate": 2.4647887323943666e-06, "loss": 7.7804, "step": 110 }, { "epoch": 0.003911308438383678, "grad_norm": 3.751282215118408, "learning_rate": 2.488262910798122e-06, "loss": 4.7061, "step": 111 }, { "epoch": 0.00394654545134209, "grad_norm": 2.256781816482544, "learning_rate": 2.5117370892018782e-06, "loss": 1.0109, "step": 112 }, { "epoch": 0.003981782464300501, "grad_norm": 3.4020116329193115, "learning_rate": 2.535211267605634e-06, "loss": 4.6154, "step": 113 }, { "epoch": 0.004017019477258913, "grad_norm": 5.7673563957214355, "learning_rate": 2.5586854460093895e-06, "loss": 7.7943, "step": 114 }, { "epoch": 0.004052256490217324, "grad_norm": 3.561821460723877, "learning_rate": 2.582159624413146e-06, "loss": 4.7412, "step": 115 }, { "epoch": 0.004087493503175736, "grad_norm": 7.756902694702148, "learning_rate": 2.6056338028169015e-06, "loss": 8.5829, "step": 116 }, { "epoch": 0.004122730516134147, "grad_norm": 2.289473056793213, "learning_rate": 2.6291079812206576e-06, "loss": 1.3947, "step": 117 }, { "epoch": 0.004157967529092559, "grad_norm": 2.356501340866089, "learning_rate": 2.652582159624413e-06, "loss": 1.5497, "step": 118 }, { "epoch": 0.00419320454205097, "grad_norm": 3.806591033935547, "learning_rate": 2.676056338028169e-06, "loss": 5.0482, "step": 119 }, { "epoch": 0.004228441555009382, "grad_norm": 1.7784596681594849, "learning_rate": 2.6995305164319253e-06, "loss": 1.003, "step": 120 }, { "epoch": 0.004263678567967793, "grad_norm": 3.8842809200286865, "learning_rate": 2.723004694835681e-06, "loss": 4.7094, "step": 121 }, { "epoch": 0.004298915580926205, "grad_norm": 1.9248768091201782, "learning_rate": 2.746478873239437e-06, "loss": 1.3955, "step": 122 }, { "epoch": 0.004334152593884617, "grad_norm": 3.663862943649292, "learning_rate": 2.7699530516431926e-06, "loss": 4.5985, "step": 123 }, { "epoch": 0.004369389606843028, "grad_norm": 2.5125420093536377, "learning_rate": 2.793427230046948e-06, "loss": 1.3476, "step": 124 }, { "epoch": 0.004404626619801439, "grad_norm": 3.3819854259490967, "learning_rate": 2.8169014084507046e-06, "loss": 3.0568, "step": 125 }, { "epoch": 0.004439863632759851, "grad_norm": 1.6669812202453613, "learning_rate": 2.8403755868544603e-06, "loss": 0.9753, "step": 126 }, { "epoch": 0.004475100645718263, "grad_norm": 7.080601215362549, "learning_rate": 2.8638497652582163e-06, "loss": 7.6769, "step": 127 }, { "epoch": 0.004510337658676674, "grad_norm": 6.418327331542969, "learning_rate": 2.887323943661972e-06, "loss": 7.9498, "step": 128 }, { "epoch": 0.004545574671635085, "grad_norm": 1.6943234205245972, "learning_rate": 2.910798122065728e-06, "loss": 1.3189, "step": 129 }, { "epoch": 0.004580811684593497, "grad_norm": 2.538247585296631, "learning_rate": 2.934272300469484e-06, "loss": 1.4867, "step": 130 }, { "epoch": 0.004616048697551909, "grad_norm": 2.3992631435394287, "learning_rate": 2.9577464788732396e-06, "loss": 1.0023, "step": 131 }, { "epoch": 0.00465128571051032, "grad_norm": 6.511706352233887, "learning_rate": 2.9812206572769952e-06, "loss": 7.3175, "step": 132 }, { "epoch": 0.004686522723468731, "grad_norm": 3.52760648727417, "learning_rate": 3.0046948356807513e-06, "loss": 4.2682, "step": 133 }, { "epoch": 0.004721759736427143, "grad_norm": 3.0231053829193115, "learning_rate": 3.0281690140845073e-06, "loss": 4.7062, "step": 134 }, { "epoch": 0.004756996749385555, "grad_norm": 4.136566162109375, "learning_rate": 3.0516431924882634e-06, "loss": 4.8314, "step": 135 }, { "epoch": 0.0047922337623439665, "grad_norm": 2.6363637447357178, "learning_rate": 3.075117370892019e-06, "loss": 1.0891, "step": 136 }, { "epoch": 0.004827470775302377, "grad_norm": 3.999605178833008, "learning_rate": 3.0985915492957746e-06, "loss": 5.3201, "step": 137 }, { "epoch": 0.004862707788260789, "grad_norm": 6.744475364685059, "learning_rate": 3.122065727699531e-06, "loss": 7.5976, "step": 138 }, { "epoch": 0.004897944801219201, "grad_norm": 3.074354648590088, "learning_rate": 3.1455399061032867e-06, "loss": 1.6017, "step": 139 }, { "epoch": 0.0049331818141776125, "grad_norm": 7.9911065101623535, "learning_rate": 3.1690140845070427e-06, "loss": 8.7684, "step": 140 }, { "epoch": 0.004968418827136023, "grad_norm": 2.186990261077881, "learning_rate": 3.1924882629107983e-06, "loss": 0.9916, "step": 141 }, { "epoch": 0.005003655840094435, "grad_norm": 1.6973457336425781, "learning_rate": 3.215962441314554e-06, "loss": 1.4424, "step": 142 }, { "epoch": 0.005038892853052847, "grad_norm": 6.728829860687256, "learning_rate": 3.2394366197183104e-06, "loss": 7.9184, "step": 143 }, { "epoch": 0.0050741298660112585, "grad_norm": 4.123422622680664, "learning_rate": 3.262910798122066e-06, "loss": 4.6512, "step": 144 }, { "epoch": 0.005109366878969669, "grad_norm": 2.6945600509643555, "learning_rate": 3.286384976525822e-06, "loss": 1.4047, "step": 145 }, { "epoch": 0.005144603891928081, "grad_norm": 1.9533910751342773, "learning_rate": 3.3098591549295777e-06, "loss": 1.3266, "step": 146 }, { "epoch": 0.005179840904886493, "grad_norm": 3.5614171028137207, "learning_rate": 3.3333333333333333e-06, "loss": 4.1498, "step": 147 }, { "epoch": 0.0052150779178449045, "grad_norm": 1.533332347869873, "learning_rate": 3.35680751173709e-06, "loss": 0.9904, "step": 148 }, { "epoch": 0.005250314930803316, "grad_norm": 2.247795820236206, "learning_rate": 3.3802816901408454e-06, "loss": 1.5296, "step": 149 }, { "epoch": 0.005285551943761727, "grad_norm": 2.673602342605591, "learning_rate": 3.4037558685446014e-06, "loss": 1.5223, "step": 150 }, { "epoch": 0.005320788956720139, "grad_norm": 3.933093547821045, "learning_rate": 3.427230046948357e-06, "loss": 5.072, "step": 151 }, { "epoch": 0.0053560259696785505, "grad_norm": 2.3533804416656494, "learning_rate": 3.4507042253521127e-06, "loss": 0.9694, "step": 152 }, { "epoch": 0.005391262982636962, "grad_norm": 2.8111050128936768, "learning_rate": 3.474178403755869e-06, "loss": 1.0125, "step": 153 }, { "epoch": 0.005426499995595373, "grad_norm": 5.44124174118042, "learning_rate": 3.4976525821596248e-06, "loss": 7.5596, "step": 154 }, { "epoch": 0.005461737008553785, "grad_norm": 6.819809436798096, "learning_rate": 3.5211267605633804e-06, "loss": 7.3704, "step": 155 }, { "epoch": 0.0054969740215121965, "grad_norm": 2.178105354309082, "learning_rate": 3.5446009389671364e-06, "loss": 1.6146, "step": 156 }, { "epoch": 0.005532211034470608, "grad_norm": 2.0858325958251953, "learning_rate": 3.568075117370892e-06, "loss": 1.0645, "step": 157 }, { "epoch": 0.005567448047429019, "grad_norm": 1.6790754795074463, "learning_rate": 3.5915492957746485e-06, "loss": 1.0186, "step": 158 }, { "epoch": 0.005602685060387431, "grad_norm": 3.4346423149108887, "learning_rate": 3.615023474178404e-06, "loss": 4.3826, "step": 159 }, { "epoch": 0.0056379220733458425, "grad_norm": 3.5869829654693604, "learning_rate": 3.6384976525821597e-06, "loss": 4.5978, "step": 160 }, { "epoch": 0.005673159086304254, "grad_norm": 5.792527675628662, "learning_rate": 3.6619718309859158e-06, "loss": 7.1769, "step": 161 }, { "epoch": 0.005708396099262666, "grad_norm": 3.684171438217163, "learning_rate": 3.6854460093896714e-06, "loss": 4.1327, "step": 162 }, { "epoch": 0.005743633112221077, "grad_norm": 2.8656868934631348, "learning_rate": 3.708920187793428e-06, "loss": 1.3827, "step": 163 }, { "epoch": 0.0057788701251794885, "grad_norm": 6.512641429901123, "learning_rate": 3.7323943661971835e-06, "loss": 7.5506, "step": 164 }, { "epoch": 0.0058141071381379, "grad_norm": 1.8054708242416382, "learning_rate": 3.755868544600939e-06, "loss": 0.8607, "step": 165 }, { "epoch": 0.005849344151096312, "grad_norm": 6.418720722198486, "learning_rate": 3.779342723004695e-06, "loss": 7.9397, "step": 166 }, { "epoch": 0.005884581164054723, "grad_norm": 3.6921181678771973, "learning_rate": 3.8028169014084508e-06, "loss": 3.8019, "step": 167 }, { "epoch": 0.0059198181770131345, "grad_norm": 1.9929388761520386, "learning_rate": 3.826291079812207e-06, "loss": 1.2605, "step": 168 }, { "epoch": 0.005955055189971546, "grad_norm": 2.068556070327759, "learning_rate": 3.849765258215963e-06, "loss": 1.1083, "step": 169 }, { "epoch": 0.005990292202929958, "grad_norm": 10.684554100036621, "learning_rate": 3.873239436619718e-06, "loss": 11.8492, "step": 170 }, { "epoch": 0.006025529215888369, "grad_norm": 1.9300930500030518, "learning_rate": 3.896713615023475e-06, "loss": 1.2399, "step": 171 }, { "epoch": 0.0060607662288467805, "grad_norm": 1.9144169092178345, "learning_rate": 3.92018779342723e-06, "loss": 1.1553, "step": 172 }, { "epoch": 0.006096003241805192, "grad_norm": 2.314589023590088, "learning_rate": 3.943661971830986e-06, "loss": 1.2163, "step": 173 }, { "epoch": 0.006131240254763604, "grad_norm": 3.520357370376587, "learning_rate": 3.967136150234742e-06, "loss": 3.6558, "step": 174 }, { "epoch": 0.006166477267722016, "grad_norm": 4.248208522796631, "learning_rate": 3.990610328638498e-06, "loss": 4.5762, "step": 175 }, { "epoch": 0.0062017142806804265, "grad_norm": 2.033477544784546, "learning_rate": 4.014084507042254e-06, "loss": 1.44, "step": 176 }, { "epoch": 0.006236951293638838, "grad_norm": 1.6570279598236084, "learning_rate": 4.0375586854460095e-06, "loss": 1.3474, "step": 177 }, { "epoch": 0.00627218830659725, "grad_norm": 4.331385135650635, "learning_rate": 4.0610328638497655e-06, "loss": 4.3169, "step": 178 }, { "epoch": 0.006307425319555662, "grad_norm": 3.2699520587921143, "learning_rate": 4.0845070422535216e-06, "loss": 3.8254, "step": 179 }, { "epoch": 0.0063426623325140725, "grad_norm": 4.184691429138184, "learning_rate": 4.107981220657278e-06, "loss": 4.1618, "step": 180 }, { "epoch": 0.006377899345472484, "grad_norm": 2.0151195526123047, "learning_rate": 4.131455399061034e-06, "loss": 0.9959, "step": 181 }, { "epoch": 0.006413136358430896, "grad_norm": 2.038391590118408, "learning_rate": 4.154929577464789e-06, "loss": 1.0419, "step": 182 }, { "epoch": 0.006448373371389308, "grad_norm": 2.0114307403564453, "learning_rate": 4.178403755868545e-06, "loss": 1.2458, "step": 183 }, { "epoch": 0.0064836103843477185, "grad_norm": 2.4075348377227783, "learning_rate": 4.201877934272301e-06, "loss": 1.421, "step": 184 }, { "epoch": 0.00651884739730613, "grad_norm": 2.1786258220672607, "learning_rate": 4.225352112676057e-06, "loss": 1.3799, "step": 185 }, { "epoch": 0.006554084410264542, "grad_norm": 2.628131866455078, "learning_rate": 4.248826291079813e-06, "loss": 1.0799, "step": 186 }, { "epoch": 0.006589321423222954, "grad_norm": 3.486762046813965, "learning_rate": 4.272300469483568e-06, "loss": 3.8681, "step": 187 }, { "epoch": 0.0066245584361813645, "grad_norm": 4.2091064453125, "learning_rate": 4.295774647887324e-06, "loss": 3.8023, "step": 188 }, { "epoch": 0.006659795449139776, "grad_norm": 3.730940580368042, "learning_rate": 4.31924882629108e-06, "loss": 4.4968, "step": 189 }, { "epoch": 0.006695032462098188, "grad_norm": 2.106771945953369, "learning_rate": 4.342723004694836e-06, "loss": 1.3513, "step": 190 }, { "epoch": 0.0067302694750566, "grad_norm": 4.101808547973633, "learning_rate": 4.3661971830985915e-06, "loss": 4.2113, "step": 191 }, { "epoch": 0.006765506488015011, "grad_norm": 3.5684471130371094, "learning_rate": 4.3896713615023476e-06, "loss": 4.4586, "step": 192 }, { "epoch": 0.006800743500973422, "grad_norm": 1.9564422369003296, "learning_rate": 4.413145539906104e-06, "loss": 0.943, "step": 193 }, { "epoch": 0.006835980513931834, "grad_norm": 3.3400378227233887, "learning_rate": 4.43661971830986e-06, "loss": 3.9059, "step": 194 }, { "epoch": 0.006871217526890246, "grad_norm": 3.954238176345825, "learning_rate": 4.460093896713616e-06, "loss": 4.47, "step": 195 }, { "epoch": 0.006906454539848657, "grad_norm": 3.5008528232574463, "learning_rate": 4.483568075117371e-06, "loss": 4.3964, "step": 196 }, { "epoch": 0.006941691552807068, "grad_norm": 4.289782524108887, "learning_rate": 4.507042253521127e-06, "loss": 4.9467, "step": 197 }, { "epoch": 0.00697692856576548, "grad_norm": 2.2520487308502197, "learning_rate": 4.530516431924883e-06, "loss": 0.9516, "step": 198 }, { "epoch": 0.007012165578723892, "grad_norm": 2.3173248767852783, "learning_rate": 4.553990610328639e-06, "loss": 1.7113, "step": 199 }, { "epoch": 0.007047402591682303, "grad_norm": 2.7772862911224365, "learning_rate": 4.577464788732395e-06, "loss": 1.5322, "step": 200 }, { "epoch": 0.007082639604640714, "grad_norm": 1.700067400932312, "learning_rate": 4.60093896713615e-06, "loss": 1.1839, "step": 201 }, { "epoch": 0.007117876617599126, "grad_norm": 2.5955793857574463, "learning_rate": 4.624413145539906e-06, "loss": 1.161, "step": 202 }, { "epoch": 0.007153113630557538, "grad_norm": 3.18257737159729, "learning_rate": 4.647887323943662e-06, "loss": 3.9933, "step": 203 }, { "epoch": 0.007188350643515949, "grad_norm": 6.234624862670898, "learning_rate": 4.671361502347418e-06, "loss": 6.9501, "step": 204 }, { "epoch": 0.007223587656474361, "grad_norm": 3.9975168704986572, "learning_rate": 4.694835680751174e-06, "loss": 4.8118, "step": 205 }, { "epoch": 0.007258824669432772, "grad_norm": 3.869300603866577, "learning_rate": 4.71830985915493e-06, "loss": 4.2765, "step": 206 }, { "epoch": 0.007294061682391184, "grad_norm": 1.9890484809875488, "learning_rate": 4.741784037558686e-06, "loss": 1.1689, "step": 207 }, { "epoch": 0.007329298695349595, "grad_norm": 2.0692670345306396, "learning_rate": 4.765258215962442e-06, "loss": 1.2853, "step": 208 }, { "epoch": 0.007364535708308007, "grad_norm": 3.9951870441436768, "learning_rate": 4.788732394366197e-06, "loss": 3.7521, "step": 209 }, { "epoch": 0.007399772721266418, "grad_norm": 4.306024074554443, "learning_rate": 4.812206572769954e-06, "loss": 4.0872, "step": 210 }, { "epoch": 0.00743500973422483, "grad_norm": 7.4345221519470215, "learning_rate": 4.835680751173709e-06, "loss": 6.8564, "step": 211 }, { "epoch": 0.007470246747183241, "grad_norm": 4.686479091644287, "learning_rate": 4.859154929577465e-06, "loss": 4.7746, "step": 212 }, { "epoch": 0.007505483760141653, "grad_norm": 2.1049957275390625, "learning_rate": 4.882629107981221e-06, "loss": 0.8566, "step": 213 }, { "epoch": 0.007540720773100064, "grad_norm": 6.784102916717529, "learning_rate": 4.906103286384977e-06, "loss": 6.4612, "step": 214 }, { "epoch": 0.007575957786058476, "grad_norm": 3.7893264293670654, "learning_rate": 4.929577464788733e-06, "loss": 4.9564, "step": 215 }, { "epoch": 0.007611194799016887, "grad_norm": 5.203770637512207, "learning_rate": 4.953051643192488e-06, "loss": 4.7336, "step": 216 }, { "epoch": 0.007646431811975299, "grad_norm": 8.029932022094727, "learning_rate": 4.976525821596244e-06, "loss": 8.262, "step": 217 }, { "epoch": 0.007681668824933711, "grad_norm": 4.295474529266357, "learning_rate": 5e-06, "loss": 3.616, "step": 218 }, { "epoch": 0.007716905837892122, "grad_norm": 1.7509058713912964, "learning_rate": 5.0234741784037565e-06, "loss": 0.9823, "step": 219 }, { "epoch": 0.007752142850850533, "grad_norm": 4.089632034301758, "learning_rate": 5.046948356807512e-06, "loss": 4.5521, "step": 220 }, { "epoch": 0.007787379863808945, "grad_norm": 6.620283126831055, "learning_rate": 5.070422535211268e-06, "loss": 5.9708, "step": 221 }, { "epoch": 0.007822616876767356, "grad_norm": 1.8007495403289795, "learning_rate": 5.093896713615024e-06, "loss": 1.1954, "step": 222 }, { "epoch": 0.007857853889725768, "grad_norm": 7.413559436798096, "learning_rate": 5.117370892018779e-06, "loss": 7.3025, "step": 223 }, { "epoch": 0.00789309090268418, "grad_norm": 5.871467113494873, "learning_rate": 5.140845070422536e-06, "loss": 5.7121, "step": 224 }, { "epoch": 0.00792832791564259, "grad_norm": 2.4824156761169434, "learning_rate": 5.164319248826292e-06, "loss": 1.241, "step": 225 }, { "epoch": 0.007963564928601003, "grad_norm": 1.9717001914978027, "learning_rate": 5.187793427230048e-06, "loss": 1.035, "step": 226 }, { "epoch": 0.007998801941559414, "grad_norm": 5.622662544250488, "learning_rate": 5.211267605633803e-06, "loss": 6.7215, "step": 227 }, { "epoch": 0.008034038954517826, "grad_norm": 2.027954578399658, "learning_rate": 5.234741784037559e-06, "loss": 1.1405, "step": 228 }, { "epoch": 0.008069275967476237, "grad_norm": 4.001363754272461, "learning_rate": 5.258215962441315e-06, "loss": 3.755, "step": 229 }, { "epoch": 0.008104512980434648, "grad_norm": 2.105654716491699, "learning_rate": 5.28169014084507e-06, "loss": 0.9084, "step": 230 }, { "epoch": 0.00813974999339306, "grad_norm": 1.8896523714065552, "learning_rate": 5.305164319248826e-06, "loss": 1.285, "step": 231 }, { "epoch": 0.008174987006351471, "grad_norm": 2.0080769062042236, "learning_rate": 5.328638497652583e-06, "loss": 1.4006, "step": 232 }, { "epoch": 0.008210224019309884, "grad_norm": 4.704226016998291, "learning_rate": 5.352112676056338e-06, "loss": 3.62, "step": 233 }, { "epoch": 0.008245461032268295, "grad_norm": 4.4286723136901855, "learning_rate": 5.3755868544600945e-06, "loss": 4.3368, "step": 234 }, { "epoch": 0.008280698045226706, "grad_norm": 2.2693121433258057, "learning_rate": 5.3990610328638506e-06, "loss": 1.1347, "step": 235 }, { "epoch": 0.008315935058185118, "grad_norm": 2.2912681102752686, "learning_rate": 5.422535211267607e-06, "loss": 1.1089, "step": 236 }, { "epoch": 0.008351172071143529, "grad_norm": 1.8499090671539307, "learning_rate": 5.446009389671362e-06, "loss": 1.1248, "step": 237 }, { "epoch": 0.00838640908410194, "grad_norm": 8.502235412597656, "learning_rate": 5.469483568075118e-06, "loss": 6.4971, "step": 238 }, { "epoch": 0.008421646097060352, "grad_norm": 2.0031116008758545, "learning_rate": 5.492957746478874e-06, "loss": 1.078, "step": 239 }, { "epoch": 0.008456883110018763, "grad_norm": 6.226426601409912, "learning_rate": 5.516431924882629e-06, "loss": 7.0993, "step": 240 }, { "epoch": 0.008492120122977176, "grad_norm": 6.216907501220703, "learning_rate": 5.539906103286385e-06, "loss": 4.5581, "step": 241 }, { "epoch": 0.008527357135935587, "grad_norm": 7.658923149108887, "learning_rate": 5.563380281690142e-06, "loss": 7.118, "step": 242 }, { "epoch": 0.008562594148893998, "grad_norm": 1.9029783010482788, "learning_rate": 5.586854460093896e-06, "loss": 1.3039, "step": 243 }, { "epoch": 0.00859783116185241, "grad_norm": 1.8084652423858643, "learning_rate": 5.610328638497653e-06, "loss": 1.031, "step": 244 }, { "epoch": 0.008633068174810821, "grad_norm": 6.59367036819458, "learning_rate": 5.633802816901409e-06, "loss": 4.8862, "step": 245 }, { "epoch": 0.008668305187769234, "grad_norm": 1.6067242622375488, "learning_rate": 5.6572769953051645e-06, "loss": 1.059, "step": 246 }, { "epoch": 0.008703542200727644, "grad_norm": 3.4344794750213623, "learning_rate": 5.6807511737089205e-06, "loss": 3.8838, "step": 247 }, { "epoch": 0.008738779213686055, "grad_norm": 2.4836158752441406, "learning_rate": 5.7042253521126766e-06, "loss": 1.0622, "step": 248 }, { "epoch": 0.008774016226644468, "grad_norm": 5.6046624183654785, "learning_rate": 5.727699530516433e-06, "loss": 7.0237, "step": 249 }, { "epoch": 0.008809253239602879, "grad_norm": 3.612281322479248, "learning_rate": 5.751173708920188e-06, "loss": 3.7461, "step": 250 }, { "epoch": 0.00884449025256129, "grad_norm": 2.2406187057495117, "learning_rate": 5.774647887323944e-06, "loss": 1.2497, "step": 251 }, { "epoch": 0.008879727265519702, "grad_norm": 3.094374895095825, "learning_rate": 5.798122065727701e-06, "loss": 3.8996, "step": 252 }, { "epoch": 0.008914964278478113, "grad_norm": 2.846076011657715, "learning_rate": 5.821596244131456e-06, "loss": 1.0527, "step": 253 }, { "epoch": 0.008950201291436526, "grad_norm": 7.51240348815918, "learning_rate": 5.845070422535212e-06, "loss": 9.0617, "step": 254 }, { "epoch": 0.008985438304394936, "grad_norm": 4.009683609008789, "learning_rate": 5.868544600938968e-06, "loss": 4.7316, "step": 255 }, { "epoch": 0.009020675317353347, "grad_norm": 2.719306707382202, "learning_rate": 5.892018779342723e-06, "loss": 1.0146, "step": 256 }, { "epoch": 0.00905591233031176, "grad_norm": 2.1774230003356934, "learning_rate": 5.915492957746479e-06, "loss": 1.0514, "step": 257 }, { "epoch": 0.00909114934327017, "grad_norm": 1.8112483024597168, "learning_rate": 5.938967136150235e-06, "loss": 1.3485, "step": 258 }, { "epoch": 0.009126386356228583, "grad_norm": 1.4583948850631714, "learning_rate": 5.9624413145539905e-06, "loss": 1.2194, "step": 259 }, { "epoch": 0.009161623369186994, "grad_norm": 4.069332122802734, "learning_rate": 5.9859154929577465e-06, "loss": 3.6537, "step": 260 }, { "epoch": 0.009196860382145405, "grad_norm": 4.309452533721924, "learning_rate": 6.0093896713615026e-06, "loss": 3.6259, "step": 261 }, { "epoch": 0.009232097395103818, "grad_norm": 5.6802449226379395, "learning_rate": 6.0328638497652595e-06, "loss": 6.6686, "step": 262 }, { "epoch": 0.009267334408062228, "grad_norm": 2.1544904708862305, "learning_rate": 6.056338028169015e-06, "loss": 1.2638, "step": 263 }, { "epoch": 0.00930257142102064, "grad_norm": 1.7856277227401733, "learning_rate": 6.079812206572771e-06, "loss": 1.3597, "step": 264 }, { "epoch": 0.009337808433979052, "grad_norm": 2.140928268432617, "learning_rate": 6.103286384976527e-06, "loss": 1.1462, "step": 265 }, { "epoch": 0.009373045446937463, "grad_norm": 3.4992804527282715, "learning_rate": 6.126760563380282e-06, "loss": 4.7853, "step": 266 }, { "epoch": 0.009408282459895875, "grad_norm": 2.296804904937744, "learning_rate": 6.150234741784038e-06, "loss": 1.0566, "step": 267 }, { "epoch": 0.009443519472854286, "grad_norm": 3.633625030517578, "learning_rate": 6.173708920187794e-06, "loss": 3.8557, "step": 268 }, { "epoch": 0.009478756485812697, "grad_norm": 4.009584903717041, "learning_rate": 6.197183098591549e-06, "loss": 3.9425, "step": 269 }, { "epoch": 0.00951399349877111, "grad_norm": 1.479809045791626, "learning_rate": 6.220657276995305e-06, "loss": 1.0266, "step": 270 }, { "epoch": 0.00954923051172952, "grad_norm": 3.1209990978240967, "learning_rate": 6.244131455399062e-06, "loss": 4.4134, "step": 271 }, { "epoch": 0.009584467524687933, "grad_norm": 2.3073346614837646, "learning_rate": 6.267605633802818e-06, "loss": 1.031, "step": 272 }, { "epoch": 0.009619704537646344, "grad_norm": 2.6208791732788086, "learning_rate": 6.291079812206573e-06, "loss": 1.1387, "step": 273 }, { "epoch": 0.009654941550604755, "grad_norm": 1.6897259950637817, "learning_rate": 6.314553990610329e-06, "loss": 1.1719, "step": 274 }, { "epoch": 0.009690178563563167, "grad_norm": 2.915290594100952, "learning_rate": 6.3380281690140855e-06, "loss": 3.7798, "step": 275 }, { "epoch": 0.009725415576521578, "grad_norm": 1.8168833255767822, "learning_rate": 6.361502347417841e-06, "loss": 1.3072, "step": 276 }, { "epoch": 0.009760652589479989, "grad_norm": 1.6200299263000488, "learning_rate": 6.384976525821597e-06, "loss": 1.2688, "step": 277 }, { "epoch": 0.009795889602438402, "grad_norm": 1.54192054271698, "learning_rate": 6.408450704225353e-06, "loss": 1.1064, "step": 278 }, { "epoch": 0.009831126615396812, "grad_norm": 1.8028688430786133, "learning_rate": 6.431924882629108e-06, "loss": 1.2243, "step": 279 }, { "epoch": 0.009866363628355225, "grad_norm": 1.946547269821167, "learning_rate": 6.455399061032864e-06, "loss": 0.9767, "step": 280 }, { "epoch": 0.009901600641313636, "grad_norm": 1.409428358078003, "learning_rate": 6.478873239436621e-06, "loss": 1.0461, "step": 281 }, { "epoch": 0.009936837654272047, "grad_norm": 3.8140978813171387, "learning_rate": 6.502347417840375e-06, "loss": 4.4097, "step": 282 }, { "epoch": 0.00997207466723046, "grad_norm": 3.5345914363861084, "learning_rate": 6.525821596244132e-06, "loss": 4.0742, "step": 283 }, { "epoch": 0.01000731168018887, "grad_norm": 3.2557108402252197, "learning_rate": 6.549295774647888e-06, "loss": 4.1197, "step": 284 }, { "epoch": 0.010042548693147283, "grad_norm": 2.3245911598205566, "learning_rate": 6.572769953051644e-06, "loss": 1.1396, "step": 285 }, { "epoch": 0.010077785706105694, "grad_norm": 2.992795467376709, "learning_rate": 6.596244131455399e-06, "loss": 0.9966, "step": 286 }, { "epoch": 0.010113022719064104, "grad_norm": 1.5400029420852661, "learning_rate": 6.619718309859155e-06, "loss": 1.1365, "step": 287 }, { "epoch": 0.010148259732022517, "grad_norm": 3.2701117992401123, "learning_rate": 6.6431924882629115e-06, "loss": 4.7189, "step": 288 }, { "epoch": 0.010183496744980928, "grad_norm": 2.797013521194458, "learning_rate": 6.666666666666667e-06, "loss": 3.7644, "step": 289 }, { "epoch": 0.010218733757939339, "grad_norm": 4.177467346191406, "learning_rate": 6.690140845070423e-06, "loss": 6.8612, "step": 290 }, { "epoch": 0.010253970770897751, "grad_norm": 2.086243152618408, "learning_rate": 6.71361502347418e-06, "loss": 1.0531, "step": 291 }, { "epoch": 0.010289207783856162, "grad_norm": 2.0694589614868164, "learning_rate": 6.737089201877935e-06, "loss": 1.0048, "step": 292 }, { "epoch": 0.010324444796814575, "grad_norm": 1.577317714691162, "learning_rate": 6.760563380281691e-06, "loss": 1.101, "step": 293 }, { "epoch": 0.010359681809772986, "grad_norm": 3.184884548187256, "learning_rate": 6.784037558685447e-06, "loss": 4.6022, "step": 294 }, { "epoch": 0.010394918822731396, "grad_norm": 2.5549964904785156, "learning_rate": 6.807511737089203e-06, "loss": 4.0745, "step": 295 }, { "epoch": 0.010430155835689809, "grad_norm": 1.9549720287322998, "learning_rate": 6.830985915492958e-06, "loss": 1.3581, "step": 296 }, { "epoch": 0.01046539284864822, "grad_norm": 2.676440954208374, "learning_rate": 6.854460093896714e-06, "loss": 4.0728, "step": 297 }, { "epoch": 0.010500629861606632, "grad_norm": 1.6529661417007446, "learning_rate": 6.87793427230047e-06, "loss": 1.0527, "step": 298 }, { "epoch": 0.010535866874565043, "grad_norm": 1.4875417947769165, "learning_rate": 6.901408450704225e-06, "loss": 0.9976, "step": 299 }, { "epoch": 0.010571103887523454, "grad_norm": 3.3112666606903076, "learning_rate": 6.924882629107981e-06, "loss": 6.5357, "step": 300 }, { "epoch": 0.010606340900481867, "grad_norm": 1.6948519945144653, "learning_rate": 6.948356807511738e-06, "loss": 1.0162, "step": 301 }, { "epoch": 0.010641577913440278, "grad_norm": 1.9733668565750122, "learning_rate": 6.9718309859154935e-06, "loss": 0.9471, "step": 302 }, { "epoch": 0.010676814926398688, "grad_norm": 2.944026470184326, "learning_rate": 6.9953051643192495e-06, "loss": 4.6456, "step": 303 }, { "epoch": 0.010712051939357101, "grad_norm": 1.2965764999389648, "learning_rate": 7.0187793427230056e-06, "loss": 0.9551, "step": 304 }, { "epoch": 0.010747288952315512, "grad_norm": 2.0309650897979736, "learning_rate": 7.042253521126761e-06, "loss": 1.265, "step": 305 }, { "epoch": 0.010782525965273924, "grad_norm": 1.5881472826004028, "learning_rate": 7.065727699530517e-06, "loss": 0.8254, "step": 306 }, { "epoch": 0.010817762978232335, "grad_norm": 2.8426291942596436, "learning_rate": 7.089201877934273e-06, "loss": 4.0273, "step": 307 }, { "epoch": 0.010852999991190746, "grad_norm": 2.1527762413024902, "learning_rate": 7.112676056338029e-06, "loss": 1.2315, "step": 308 }, { "epoch": 0.010888237004149159, "grad_norm": 2.5771236419677734, "learning_rate": 7.136150234741784e-06, "loss": 4.1992, "step": 309 }, { "epoch": 0.01092347401710757, "grad_norm": 2.8491079807281494, "learning_rate": 7.159624413145541e-06, "loss": 4.5822, "step": 310 }, { "epoch": 0.010958711030065982, "grad_norm": 4.143470764160156, "learning_rate": 7.183098591549297e-06, "loss": 6.8679, "step": 311 }, { "epoch": 0.010993948043024393, "grad_norm": 3.2853147983551025, "learning_rate": 7.206572769953052e-06, "loss": 4.0438, "step": 312 }, { "epoch": 0.011029185055982804, "grad_norm": 2.610133171081543, "learning_rate": 7.230046948356808e-06, "loss": 1.0476, "step": 313 }, { "epoch": 0.011064422068941216, "grad_norm": 3.9282050132751465, "learning_rate": 7.253521126760564e-06, "loss": 4.4495, "step": 314 }, { "epoch": 0.011099659081899627, "grad_norm": 3.3522932529449463, "learning_rate": 7.2769953051643195e-06, "loss": 7.7731, "step": 315 }, { "epoch": 0.011134896094858038, "grad_norm": 1.7992554903030396, "learning_rate": 7.3004694835680755e-06, "loss": 0.9925, "step": 316 }, { "epoch": 0.01117013310781645, "grad_norm": 2.104796886444092, "learning_rate": 7.3239436619718316e-06, "loss": 4.9739, "step": 317 }, { "epoch": 0.011205370120774862, "grad_norm": 2.0702006816864014, "learning_rate": 7.347417840375587e-06, "loss": 0.9831, "step": 318 }, { "epoch": 0.011240607133733274, "grad_norm": 1.717795729637146, "learning_rate": 7.370892018779343e-06, "loss": 0.8669, "step": 319 }, { "epoch": 0.011275844146691685, "grad_norm": 2.559128522872925, "learning_rate": 7.3943661971831e-06, "loss": 3.8081, "step": 320 }, { "epoch": 0.011311081159650096, "grad_norm": 2.9320385456085205, "learning_rate": 7.417840375586856e-06, "loss": 6.4979, "step": 321 }, { "epoch": 0.011346318172608508, "grad_norm": 2.2109715938568115, "learning_rate": 7.441314553990611e-06, "loss": 1.069, "step": 322 }, { "epoch": 0.01138155518556692, "grad_norm": 1.884302020072937, "learning_rate": 7.464788732394367e-06, "loss": 0.8798, "step": 323 }, { "epoch": 0.011416792198525332, "grad_norm": 2.0465826988220215, "learning_rate": 7.488262910798123e-06, "loss": 4.6015, "step": 324 }, { "epoch": 0.011452029211483743, "grad_norm": 3.3119335174560547, "learning_rate": 7.511737089201878e-06, "loss": 8.2375, "step": 325 }, { "epoch": 0.011487266224442154, "grad_norm": 2.5395524501800537, "learning_rate": 7.535211267605634e-06, "loss": 3.44, "step": 326 }, { "epoch": 0.011522503237400566, "grad_norm": 2.400813341140747, "learning_rate": 7.55868544600939e-06, "loss": 1.0838, "step": 327 }, { "epoch": 0.011557740250358977, "grad_norm": 2.879608154296875, "learning_rate": 7.5821596244131455e-06, "loss": 7.3328, "step": 328 }, { "epoch": 0.011592977263317388, "grad_norm": 2.934816598892212, "learning_rate": 7.6056338028169015e-06, "loss": 7.0567, "step": 329 }, { "epoch": 0.0116282142762758, "grad_norm": 1.7331291437149048, "learning_rate": 7.629107981220658e-06, "loss": 1.0055, "step": 330 }, { "epoch": 0.011663451289234211, "grad_norm": 2.222409248352051, "learning_rate": 7.652582159624414e-06, "loss": 1.0462, "step": 331 }, { "epoch": 0.011698688302192624, "grad_norm": 1.848039984703064, "learning_rate": 7.67605633802817e-06, "loss": 1.223, "step": 332 }, { "epoch": 0.011733925315151035, "grad_norm": 2.6071157455444336, "learning_rate": 7.699530516431926e-06, "loss": 3.9107, "step": 333 }, { "epoch": 0.011769162328109446, "grad_norm": 2.957768201828003, "learning_rate": 7.723004694835682e-06, "loss": 4.2836, "step": 334 }, { "epoch": 0.011804399341067858, "grad_norm": 2.842270851135254, "learning_rate": 7.746478873239436e-06, "loss": 7.5356, "step": 335 }, { "epoch": 0.011839636354026269, "grad_norm": 3.12445068359375, "learning_rate": 7.769953051643194e-06, "loss": 3.3509, "step": 336 }, { "epoch": 0.011874873366984682, "grad_norm": 2.799595594406128, "learning_rate": 7.79342723004695e-06, "loss": 3.5825, "step": 337 }, { "epoch": 0.011910110379943092, "grad_norm": 4.682062149047852, "learning_rate": 7.816901408450704e-06, "loss": 7.318, "step": 338 }, { "epoch": 0.011945347392901503, "grad_norm": 3.4375622272491455, "learning_rate": 7.84037558685446e-06, "loss": 4.41, "step": 339 }, { "epoch": 0.011980584405859916, "grad_norm": 1.986032247543335, "learning_rate": 7.863849765258216e-06, "loss": 1.1163, "step": 340 }, { "epoch": 0.012015821418818327, "grad_norm": 1.9632568359375, "learning_rate": 7.887323943661972e-06, "loss": 1.0573, "step": 341 }, { "epoch": 0.012051058431776738, "grad_norm": 4.608703136444092, "learning_rate": 7.910798122065728e-06, "loss": 7.5448, "step": 342 }, { "epoch": 0.01208629544473515, "grad_norm": 2.842829942703247, "learning_rate": 7.934272300469484e-06, "loss": 7.5795, "step": 343 }, { "epoch": 0.012121532457693561, "grad_norm": 1.6826400756835938, "learning_rate": 7.95774647887324e-06, "loss": 1.0841, "step": 344 }, { "epoch": 0.012156769470651974, "grad_norm": 2.4194138050079346, "learning_rate": 7.981220657276996e-06, "loss": 1.0417, "step": 345 }, { "epoch": 0.012192006483610384, "grad_norm": 2.0106773376464844, "learning_rate": 8.004694835680753e-06, "loss": 1.2537, "step": 346 }, { "epoch": 0.012227243496568795, "grad_norm": 1.9119226932525635, "learning_rate": 8.028169014084509e-06, "loss": 0.9191, "step": 347 }, { "epoch": 0.012262480509527208, "grad_norm": 2.1980514526367188, "learning_rate": 8.051643192488263e-06, "loss": 0.8745, "step": 348 }, { "epoch": 0.012297717522485619, "grad_norm": 2.1777870655059814, "learning_rate": 8.075117370892019e-06, "loss": 3.6755, "step": 349 }, { "epoch": 0.012332954535444031, "grad_norm": 2.6617698669433594, "learning_rate": 8.098591549295775e-06, "loss": 3.7869, "step": 350 }, { "epoch": 0.012368191548402442, "grad_norm": 2.643073320388794, "learning_rate": 8.122065727699531e-06, "loss": 3.8362, "step": 351 }, { "epoch": 0.012403428561360853, "grad_norm": 2.236356019973755, "learning_rate": 8.145539906103287e-06, "loss": 4.2061, "step": 352 }, { "epoch": 0.012438665574319266, "grad_norm": 1.8140854835510254, "learning_rate": 8.169014084507043e-06, "loss": 4.3772, "step": 353 }, { "epoch": 0.012473902587277676, "grad_norm": 2.260471820831299, "learning_rate": 8.192488262910797e-06, "loss": 4.148, "step": 354 }, { "epoch": 0.012509139600236087, "grad_norm": 1.726349115371704, "learning_rate": 8.215962441314555e-06, "loss": 1.1929, "step": 355 }, { "epoch": 0.0125443766131945, "grad_norm": 1.6451479196548462, "learning_rate": 8.239436619718311e-06, "loss": 1.1605, "step": 356 }, { "epoch": 0.01257961362615291, "grad_norm": 2.0361626148223877, "learning_rate": 8.262910798122067e-06, "loss": 4.0527, "step": 357 }, { "epoch": 0.012614850639111323, "grad_norm": 2.347341775894165, "learning_rate": 8.286384976525822e-06, "loss": 1.0677, "step": 358 }, { "epoch": 0.012650087652069734, "grad_norm": 2.04463529586792, "learning_rate": 8.309859154929578e-06, "loss": 3.9051, "step": 359 }, { "epoch": 0.012685324665028145, "grad_norm": 1.9530123472213745, "learning_rate": 8.333333333333334e-06, "loss": 1.0204, "step": 360 }, { "epoch": 0.012720561677986558, "grad_norm": 2.1576826572418213, "learning_rate": 8.35680751173709e-06, "loss": 4.4164, "step": 361 }, { "epoch": 0.012755798690944968, "grad_norm": 2.673433542251587, "learning_rate": 8.380281690140846e-06, "loss": 4.0633, "step": 362 }, { "epoch": 0.012791035703903381, "grad_norm": 2.968332052230835, "learning_rate": 8.403755868544602e-06, "loss": 4.3995, "step": 363 }, { "epoch": 0.012826272716861792, "grad_norm": 2.893555164337158, "learning_rate": 8.427230046948356e-06, "loss": 4.1662, "step": 364 }, { "epoch": 0.012861509729820203, "grad_norm": 2.381232500076294, "learning_rate": 8.450704225352114e-06, "loss": 3.0079, "step": 365 }, { "epoch": 0.012896746742778615, "grad_norm": 2.5255520343780518, "learning_rate": 8.47417840375587e-06, "loss": 4.7671, "step": 366 }, { "epoch": 0.012931983755737026, "grad_norm": 2.429490804672241, "learning_rate": 8.497652582159626e-06, "loss": 3.4408, "step": 367 }, { "epoch": 0.012967220768695437, "grad_norm": 2.1161246299743652, "learning_rate": 8.52112676056338e-06, "loss": 4.1137, "step": 368 }, { "epoch": 0.01300245778165385, "grad_norm": 2.261544704437256, "learning_rate": 8.544600938967136e-06, "loss": 3.5179, "step": 369 }, { "epoch": 0.01303769479461226, "grad_norm": 1.7343007326126099, "learning_rate": 8.568075117370892e-06, "loss": 0.8964, "step": 370 }, { "epoch": 0.013072931807570673, "grad_norm": 1.6480615139007568, "learning_rate": 8.591549295774648e-06, "loss": 1.0978, "step": 371 }, { "epoch": 0.013108168820529084, "grad_norm": 1.7105282545089722, "learning_rate": 8.615023474178405e-06, "loss": 1.0016, "step": 372 }, { "epoch": 0.013143405833487495, "grad_norm": 2.2872941493988037, "learning_rate": 8.63849765258216e-06, "loss": 1.2367, "step": 373 }, { "epoch": 0.013178642846445907, "grad_norm": 2.0584661960601807, "learning_rate": 8.661971830985915e-06, "loss": 3.6177, "step": 374 }, { "epoch": 0.013213879859404318, "grad_norm": 2.0558316707611084, "learning_rate": 8.685446009389673e-06, "loss": 0.9207, "step": 375 }, { "epoch": 0.013249116872362729, "grad_norm": 1.901312232017517, "learning_rate": 8.708920187793429e-06, "loss": 1.2697, "step": 376 }, { "epoch": 0.013284353885321142, "grad_norm": 2.0170602798461914, "learning_rate": 8.732394366197183e-06, "loss": 1.1368, "step": 377 }, { "epoch": 0.013319590898279552, "grad_norm": 2.503995180130005, "learning_rate": 8.755868544600939e-06, "loss": 3.9531, "step": 378 }, { "epoch": 0.013354827911237965, "grad_norm": 1.5834543704986572, "learning_rate": 8.779342723004695e-06, "loss": 1.4025, "step": 379 }, { "epoch": 0.013390064924196376, "grad_norm": 2.3775064945220947, "learning_rate": 8.802816901408451e-06, "loss": 3.9194, "step": 380 }, { "epoch": 0.013425301937154787, "grad_norm": 1.8064970970153809, "learning_rate": 8.826291079812207e-06, "loss": 1.1846, "step": 381 }, { "epoch": 0.0134605389501132, "grad_norm": 3.556758165359497, "learning_rate": 8.849765258215963e-06, "loss": 6.895, "step": 382 }, { "epoch": 0.01349577596307161, "grad_norm": 2.8685925006866455, "learning_rate": 8.87323943661972e-06, "loss": 3.6601, "step": 383 }, { "epoch": 0.013531012976030023, "grad_norm": 3.0653574466705322, "learning_rate": 8.896713615023475e-06, "loss": 1.0731, "step": 384 }, { "epoch": 0.013566249988988434, "grad_norm": 6.05435848236084, "learning_rate": 8.920187793427231e-06, "loss": 4.4383, "step": 385 }, { "epoch": 0.013601487001946844, "grad_norm": 1.7964208126068115, "learning_rate": 8.943661971830987e-06, "loss": 1.281, "step": 386 }, { "epoch": 0.013636724014905257, "grad_norm": 5.110411167144775, "learning_rate": 8.967136150234742e-06, "loss": 6.9458, "step": 387 }, { "epoch": 0.013671961027863668, "grad_norm": 1.5369724035263062, "learning_rate": 8.990610328638498e-06, "loss": 0.9129, "step": 388 }, { "epoch": 0.013707198040822079, "grad_norm": 2.66107439994812, "learning_rate": 9.014084507042254e-06, "loss": 3.7609, "step": 389 }, { "epoch": 0.013742435053780491, "grad_norm": 4.562012195587158, "learning_rate": 9.03755868544601e-06, "loss": 4.4225, "step": 390 }, { "epoch": 0.013777672066738902, "grad_norm": 3.2169108390808105, "learning_rate": 9.061032863849766e-06, "loss": 4.2459, "step": 391 }, { "epoch": 0.013812909079697315, "grad_norm": 1.6387608051300049, "learning_rate": 9.084507042253522e-06, "loss": 1.0604, "step": 392 }, { "epoch": 0.013848146092655726, "grad_norm": 1.5896732807159424, "learning_rate": 9.107981220657278e-06, "loss": 0.9598, "step": 393 }, { "epoch": 0.013883383105614136, "grad_norm": 5.934868812561035, "learning_rate": 9.131455399061034e-06, "loss": 6.7612, "step": 394 }, { "epoch": 0.013918620118572549, "grad_norm": 2.4856579303741455, "learning_rate": 9.15492957746479e-06, "loss": 0.9606, "step": 395 }, { "epoch": 0.01395385713153096, "grad_norm": 3.4068562984466553, "learning_rate": 9.178403755868546e-06, "loss": 6.632, "step": 396 }, { "epoch": 0.013989094144489372, "grad_norm": 2.4071836471557617, "learning_rate": 9.2018779342723e-06, "loss": 4.4494, "step": 397 }, { "epoch": 0.014024331157447783, "grad_norm": 2.542024850845337, "learning_rate": 9.225352112676057e-06, "loss": 3.9821, "step": 398 }, { "epoch": 0.014059568170406194, "grad_norm": 2.197418451309204, "learning_rate": 9.248826291079813e-06, "loss": 4.3643, "step": 399 }, { "epoch": 0.014094805183364607, "grad_norm": 3.508121967315674, "learning_rate": 9.272300469483569e-06, "loss": 6.3408, "step": 400 }, { "epoch": 0.014130042196323018, "grad_norm": 3.881662607192993, "learning_rate": 9.295774647887325e-06, "loss": 6.505, "step": 401 }, { "epoch": 0.014165279209281428, "grad_norm": 2.4805357456207275, "learning_rate": 9.31924882629108e-06, "loss": 3.8867, "step": 402 }, { "epoch": 0.014200516222239841, "grad_norm": 2.6266720294952393, "learning_rate": 9.342723004694837e-06, "loss": 3.5808, "step": 403 }, { "epoch": 0.014235753235198252, "grad_norm": 4.636962890625, "learning_rate": 9.366197183098593e-06, "loss": 7.5262, "step": 404 }, { "epoch": 0.014270990248156664, "grad_norm": 2.824331045150757, "learning_rate": 9.389671361502349e-06, "loss": 1.0877, "step": 405 }, { "epoch": 0.014306227261115075, "grad_norm": 3.399836301803589, "learning_rate": 9.413145539906105e-06, "loss": 7.0796, "step": 406 }, { "epoch": 0.014341464274073486, "grad_norm": 1.916932225227356, "learning_rate": 9.43661971830986e-06, "loss": 3.9393, "step": 407 }, { "epoch": 0.014376701287031899, "grad_norm": 2.7756643295288086, "learning_rate": 9.460093896713615e-06, "loss": 3.3438, "step": 408 }, { "epoch": 0.01441193829999031, "grad_norm": 2.081050157546997, "learning_rate": 9.483568075117371e-06, "loss": 1.4977, "step": 409 }, { "epoch": 0.014447175312948722, "grad_norm": 3.422701597213745, "learning_rate": 9.507042253521127e-06, "loss": 7.7301, "step": 410 }, { "epoch": 0.014482412325907133, "grad_norm": 1.798227310180664, "learning_rate": 9.530516431924883e-06, "loss": 1.0628, "step": 411 }, { "epoch": 0.014517649338865544, "grad_norm": 1.8264895677566528, "learning_rate": 9.55399061032864e-06, "loss": 0.9678, "step": 412 }, { "epoch": 0.014552886351823956, "grad_norm": 2.0592398643493652, "learning_rate": 9.577464788732394e-06, "loss": 1.0831, "step": 413 }, { "epoch": 0.014588123364782367, "grad_norm": 2.6472694873809814, "learning_rate": 9.600938967136152e-06, "loss": 6.6183, "step": 414 }, { "epoch": 0.014623360377740778, "grad_norm": 1.728895902633667, "learning_rate": 9.624413145539908e-06, "loss": 1.0317, "step": 415 }, { "epoch": 0.01465859739069919, "grad_norm": 2.8150815963745117, "learning_rate": 9.647887323943664e-06, "loss": 1.123, "step": 416 }, { "epoch": 0.014693834403657602, "grad_norm": 2.4313268661499023, "learning_rate": 9.671361502347418e-06, "loss": 3.9203, "step": 417 }, { "epoch": 0.014729071416616014, "grad_norm": 2.1881418228149414, "learning_rate": 9.694835680751174e-06, "loss": 0.7818, "step": 418 }, { "epoch": 0.014764308429574425, "grad_norm": 2.3101859092712402, "learning_rate": 9.71830985915493e-06, "loss": 6.392, "step": 419 }, { "epoch": 0.014799545442532836, "grad_norm": 2.7211689949035645, "learning_rate": 9.741784037558686e-06, "loss": 5.3799, "step": 420 }, { "epoch": 0.014834782455491248, "grad_norm": 1.8715535402297974, "learning_rate": 9.765258215962442e-06, "loss": 1.1888, "step": 421 }, { "epoch": 0.01487001946844966, "grad_norm": 3.0835886001586914, "learning_rate": 9.788732394366198e-06, "loss": 10.0046, "step": 422 }, { "epoch": 0.014905256481408072, "grad_norm": 1.6927839517593384, "learning_rate": 9.812206572769954e-06, "loss": 1.4534, "step": 423 }, { "epoch": 0.014940493494366483, "grad_norm": 2.3459179401397705, "learning_rate": 9.83568075117371e-06, "loss": 3.5929, "step": 424 }, { "epoch": 0.014975730507324893, "grad_norm": 2.7162883281707764, "learning_rate": 9.859154929577466e-06, "loss": 3.7307, "step": 425 }, { "epoch": 0.015010967520283306, "grad_norm": 2.801145553588867, "learning_rate": 9.882629107981222e-06, "loss": 5.0893, "step": 426 }, { "epoch": 0.015046204533241717, "grad_norm": 2.431332588195801, "learning_rate": 9.906103286384977e-06, "loss": 0.8929, "step": 427 }, { "epoch": 0.015081441546200128, "grad_norm": 2.3686599731445312, "learning_rate": 9.929577464788733e-06, "loss": 1.003, "step": 428 }, { "epoch": 0.01511667855915854, "grad_norm": 3.3232877254486084, "learning_rate": 9.953051643192489e-06, "loss": 3.8281, "step": 429 }, { "epoch": 0.015151915572116951, "grad_norm": 2.584277868270874, "learning_rate": 9.976525821596245e-06, "loss": 3.7623, "step": 430 }, { "epoch": 0.015187152585075364, "grad_norm": 1.464976191520691, "learning_rate": 1e-05, "loss": 0.9113, "step": 431 }, { "epoch": 0.015222389598033775, "grad_norm": 1.5460474491119385, "learning_rate": 1.0023474178403755e-05, "loss": 1.2378, "step": 432 }, { "epoch": 0.015257626610992185, "grad_norm": 2.2827024459838867, "learning_rate": 1.0046948356807513e-05, "loss": 1.2853, "step": 433 }, { "epoch": 0.015292863623950598, "grad_norm": 2.014397621154785, "learning_rate": 1.0070422535211269e-05, "loss": 3.6359, "step": 434 }, { "epoch": 0.015328100636909009, "grad_norm": 2.8575856685638428, "learning_rate": 1.0093896713615023e-05, "loss": 1.3211, "step": 435 }, { "epoch": 0.015363337649867422, "grad_norm": 3.5683705806732178, "learning_rate": 1.0117370892018781e-05, "loss": 6.7173, "step": 436 }, { "epoch": 0.015398574662825832, "grad_norm": 2.213655710220337, "learning_rate": 1.0140845070422535e-05, "loss": 3.8259, "step": 437 }, { "epoch": 0.015433811675784243, "grad_norm": 2.7231802940368652, "learning_rate": 1.0164319248826291e-05, "loss": 4.4818, "step": 438 }, { "epoch": 0.015469048688742656, "grad_norm": 2.5079281330108643, "learning_rate": 1.0187793427230047e-05, "loss": 5.9616, "step": 439 }, { "epoch": 0.015504285701701067, "grad_norm": 1.9521703720092773, "learning_rate": 1.0211267605633803e-05, "loss": 1.0675, "step": 440 }, { "epoch": 0.015539522714659477, "grad_norm": 1.959806203842163, "learning_rate": 1.0234741784037558e-05, "loss": 0.9927, "step": 441 }, { "epoch": 0.01557475972761789, "grad_norm": 2.0852391719818115, "learning_rate": 1.0258215962441316e-05, "loss": 0.8526, "step": 442 }, { "epoch": 0.015609996740576301, "grad_norm": 5.094117641448975, "learning_rate": 1.0281690140845072e-05, "loss": 6.6034, "step": 443 }, { "epoch": 0.015645233753534712, "grad_norm": 2.225939989089966, "learning_rate": 1.0305164319248828e-05, "loss": 4.5891, "step": 444 }, { "epoch": 0.015680470766493124, "grad_norm": 2.176257610321045, "learning_rate": 1.0328638497652584e-05, "loss": 1.0077, "step": 445 }, { "epoch": 0.015715707779451537, "grad_norm": 2.1797478199005127, "learning_rate": 1.0352112676056338e-05, "loss": 6.8023, "step": 446 }, { "epoch": 0.015750944792409946, "grad_norm": 2.2790627479553223, "learning_rate": 1.0375586854460096e-05, "loss": 3.8638, "step": 447 }, { "epoch": 0.01578618180536836, "grad_norm": 1.9245857000350952, "learning_rate": 1.039906103286385e-05, "loss": 1.1735, "step": 448 }, { "epoch": 0.01582141881832677, "grad_norm": 1.6776940822601318, "learning_rate": 1.0422535211267606e-05, "loss": 1.0314, "step": 449 }, { "epoch": 0.01585665583128518, "grad_norm": 1.8242470026016235, "learning_rate": 1.0446009389671364e-05, "loss": 1.3181, "step": 450 }, { "epoch": 0.015891892844243593, "grad_norm": 2.2489068508148193, "learning_rate": 1.0469483568075118e-05, "loss": 1.0687, "step": 451 }, { "epoch": 0.015927129857202005, "grad_norm": 2.079066753387451, "learning_rate": 1.0492957746478873e-05, "loss": 7.07, "step": 452 }, { "epoch": 0.015962366870160418, "grad_norm": 2.154287576675415, "learning_rate": 1.051643192488263e-05, "loss": 3.9502, "step": 453 }, { "epoch": 0.015997603883118827, "grad_norm": 1.916580080986023, "learning_rate": 1.0539906103286386e-05, "loss": 4.0525, "step": 454 }, { "epoch": 0.01603284089607724, "grad_norm": 1.7289687395095825, "learning_rate": 1.056338028169014e-05, "loss": 4.1883, "step": 455 }, { "epoch": 0.016068077909035652, "grad_norm": 2.4038257598876953, "learning_rate": 1.0586854460093898e-05, "loss": 8.3464, "step": 456 }, { "epoch": 0.01610331492199406, "grad_norm": 2.0832910537719727, "learning_rate": 1.0610328638497653e-05, "loss": 4.8723, "step": 457 }, { "epoch": 0.016138551934952474, "grad_norm": 1.9041929244995117, "learning_rate": 1.0633802816901409e-05, "loss": 4.1394, "step": 458 }, { "epoch": 0.016173788947910887, "grad_norm": 2.3373985290527344, "learning_rate": 1.0657276995305167e-05, "loss": 3.4201, "step": 459 }, { "epoch": 0.016209025960869296, "grad_norm": 2.777296304702759, "learning_rate": 1.0680751173708921e-05, "loss": 3.984, "step": 460 }, { "epoch": 0.01624426297382771, "grad_norm": 1.8931972980499268, "learning_rate": 1.0704225352112675e-05, "loss": 0.9811, "step": 461 }, { "epoch": 0.01627949998678612, "grad_norm": 2.1997339725494385, "learning_rate": 1.0727699530516433e-05, "loss": 3.7561, "step": 462 }, { "epoch": 0.01631473699974453, "grad_norm": 2.409224033355713, "learning_rate": 1.0751173708920189e-05, "loss": 6.2385, "step": 463 }, { "epoch": 0.016349974012702943, "grad_norm": 2.5075204372406006, "learning_rate": 1.0774647887323943e-05, "loss": 3.5541, "step": 464 }, { "epoch": 0.016385211025661355, "grad_norm": 1.6470845937728882, "learning_rate": 1.0798122065727701e-05, "loss": 1.1553, "step": 465 }, { "epoch": 0.016420448038619768, "grad_norm": 1.949081301689148, "learning_rate": 1.0821596244131455e-05, "loss": 0.8671, "step": 466 }, { "epoch": 0.016455685051578177, "grad_norm": 1.5028753280639648, "learning_rate": 1.0845070422535213e-05, "loss": 0.9144, "step": 467 }, { "epoch": 0.01649092206453659, "grad_norm": 1.3537009954452515, "learning_rate": 1.0868544600938968e-05, "loss": 0.8167, "step": 468 }, { "epoch": 0.016526159077495002, "grad_norm": 7.483241081237793, "learning_rate": 1.0892018779342724e-05, "loss": 4.703, "step": 469 }, { "epoch": 0.01656139609045341, "grad_norm": 2.2480075359344482, "learning_rate": 1.0915492957746481e-05, "loss": 1.1017, "step": 470 }, { "epoch": 0.016596633103411824, "grad_norm": 2.05956768989563, "learning_rate": 1.0938967136150236e-05, "loss": 0.8671, "step": 471 }, { "epoch": 0.016631870116370236, "grad_norm": 3.013094663619995, "learning_rate": 1.0962441314553992e-05, "loss": 0.9076, "step": 472 }, { "epoch": 0.016667107129328645, "grad_norm": 2.0225467681884766, "learning_rate": 1.0985915492957748e-05, "loss": 1.0562, "step": 473 }, { "epoch": 0.016702344142287058, "grad_norm": 1.9703342914581299, "learning_rate": 1.1009389671361504e-05, "loss": 1.1242, "step": 474 }, { "epoch": 0.01673758115524547, "grad_norm": 1.7998673915863037, "learning_rate": 1.1032863849765258e-05, "loss": 0.9224, "step": 475 }, { "epoch": 0.01677281816820388, "grad_norm": 3.16202712059021, "learning_rate": 1.1056338028169016e-05, "loss": 6.928, "step": 476 }, { "epoch": 0.016808055181162292, "grad_norm": 1.794331669807434, "learning_rate": 1.107981220657277e-05, "loss": 4.1736, "step": 477 }, { "epoch": 0.016843292194120705, "grad_norm": 2.0121424198150635, "learning_rate": 1.1103286384976526e-05, "loss": 7.3583, "step": 478 }, { "epoch": 0.016878529207079117, "grad_norm": 1.8480234146118164, "learning_rate": 1.1126760563380284e-05, "loss": 7.092, "step": 479 }, { "epoch": 0.016913766220037527, "grad_norm": 1.8990559577941895, "learning_rate": 1.1150234741784038e-05, "loss": 1.1489, "step": 480 }, { "epoch": 0.01694900323299594, "grad_norm": 2.2544138431549072, "learning_rate": 1.1173708920187793e-05, "loss": 4.2643, "step": 481 }, { "epoch": 0.016984240245954352, "grad_norm": 1.7743180990219116, "learning_rate": 1.119718309859155e-05, "loss": 4.1252, "step": 482 }, { "epoch": 0.01701947725891276, "grad_norm": 2.118295669555664, "learning_rate": 1.1220657276995307e-05, "loss": 7.2855, "step": 483 }, { "epoch": 0.017054714271871173, "grad_norm": 1.9713571071624756, "learning_rate": 1.1244131455399061e-05, "loss": 1.3232, "step": 484 }, { "epoch": 0.017089951284829586, "grad_norm": 1.9433363676071167, "learning_rate": 1.1267605633802819e-05, "loss": 1.0971, "step": 485 }, { "epoch": 0.017125188297787995, "grad_norm": 2.201974630355835, "learning_rate": 1.1291079812206573e-05, "loss": 1.0899, "step": 486 }, { "epoch": 0.017160425310746408, "grad_norm": 1.646652102470398, "learning_rate": 1.1314553990610329e-05, "loss": 1.0804, "step": 487 }, { "epoch": 0.01719566232370482, "grad_norm": 1.8079637289047241, "learning_rate": 1.1338028169014087e-05, "loss": 0.822, "step": 488 }, { "epoch": 0.01723089933666323, "grad_norm": 1.6560924053192139, "learning_rate": 1.1361502347417841e-05, "loss": 1.2865, "step": 489 }, { "epoch": 0.017266136349621642, "grad_norm": 2.8661415576934814, "learning_rate": 1.1384976525821595e-05, "loss": 4.5809, "step": 490 }, { "epoch": 0.017301373362580055, "grad_norm": 2.727036237716675, "learning_rate": 1.1408450704225353e-05, "loss": 10.9725, "step": 491 }, { "epoch": 0.017336610375538467, "grad_norm": 2.1375811100006104, "learning_rate": 1.143192488262911e-05, "loss": 4.6613, "step": 492 }, { "epoch": 0.017371847388496876, "grad_norm": 2.9289305210113525, "learning_rate": 1.1455399061032865e-05, "loss": 6.9325, "step": 493 }, { "epoch": 0.01740708440145529, "grad_norm": 1.3923015594482422, "learning_rate": 1.1478873239436621e-05, "loss": 0.8397, "step": 494 }, { "epoch": 0.0174423214144137, "grad_norm": 1.6348791122436523, "learning_rate": 1.1502347417840376e-05, "loss": 0.9462, "step": 495 }, { "epoch": 0.01747755842737211, "grad_norm": 2.304156541824341, "learning_rate": 1.1525821596244133e-05, "loss": 3.9202, "step": 496 }, { "epoch": 0.017512795440330523, "grad_norm": 2.658198833465576, "learning_rate": 1.1549295774647888e-05, "loss": 3.5085, "step": 497 }, { "epoch": 0.017548032453288936, "grad_norm": 2.7420387268066406, "learning_rate": 1.1572769953051644e-05, "loss": 3.1997, "step": 498 }, { "epoch": 0.017583269466247345, "grad_norm": 1.9836087226867676, "learning_rate": 1.1596244131455401e-05, "loss": 1.2469, "step": 499 }, { "epoch": 0.017618506479205757, "grad_norm": 1.904054045677185, "learning_rate": 1.1619718309859156e-05, "loss": 1.1128, "step": 500 }, { "epoch": 0.01765374349216417, "grad_norm": 2.0734140872955322, "learning_rate": 1.1643192488262912e-05, "loss": 0.9969, "step": 501 }, { "epoch": 0.01768898050512258, "grad_norm": 2.220015525817871, "learning_rate": 1.1666666666666668e-05, "loss": 0.9504, "step": 502 }, { "epoch": 0.017724217518080992, "grad_norm": 2.2554752826690674, "learning_rate": 1.1690140845070424e-05, "loss": 4.0228, "step": 503 }, { "epoch": 0.017759454531039404, "grad_norm": 1.825788974761963, "learning_rate": 1.1713615023474178e-05, "loss": 3.4124, "step": 504 }, { "epoch": 0.017794691543997817, "grad_norm": 2.266758918762207, "learning_rate": 1.1737089201877936e-05, "loss": 4.3487, "step": 505 }, { "epoch": 0.017829928556956226, "grad_norm": 2.1770904064178467, "learning_rate": 1.176056338028169e-05, "loss": 3.915, "step": 506 }, { "epoch": 0.01786516556991464, "grad_norm": 2.1360225677490234, "learning_rate": 1.1784037558685446e-05, "loss": 1.2802, "step": 507 }, { "epoch": 0.01790040258287305, "grad_norm": 2.0103824138641357, "learning_rate": 1.1807511737089204e-05, "loss": 1.0616, "step": 508 }, { "epoch": 0.01793563959583146, "grad_norm": 1.973430871963501, "learning_rate": 1.1830985915492958e-05, "loss": 1.006, "step": 509 }, { "epoch": 0.017970876608789873, "grad_norm": 1.820097804069519, "learning_rate": 1.1854460093896713e-05, "loss": 0.852, "step": 510 }, { "epoch": 0.018006113621748285, "grad_norm": 2.2512922286987305, "learning_rate": 1.187793427230047e-05, "loss": 3.2274, "step": 511 }, { "epoch": 0.018041350634706695, "grad_norm": 3.629587411880493, "learning_rate": 1.1901408450704227e-05, "loss": 4.353, "step": 512 }, { "epoch": 0.018076587647665107, "grad_norm": 1.9418200254440308, "learning_rate": 1.1924882629107981e-05, "loss": 1.0424, "step": 513 }, { "epoch": 0.01811182466062352, "grad_norm": 2.440890312194824, "learning_rate": 1.1948356807511739e-05, "loss": 3.5051, "step": 514 }, { "epoch": 0.01814706167358193, "grad_norm": 2.304457902908325, "learning_rate": 1.1971830985915493e-05, "loss": 3.1345, "step": 515 }, { "epoch": 0.01818229868654034, "grad_norm": 2.8411800861358643, "learning_rate": 1.199530516431925e-05, "loss": 9.7516, "step": 516 }, { "epoch": 0.018217535699498754, "grad_norm": 1.6378328800201416, "learning_rate": 1.2018779342723005e-05, "loss": 1.0761, "step": 517 }, { "epoch": 0.018252772712457167, "grad_norm": 2.953984498977661, "learning_rate": 1.2042253521126761e-05, "loss": 7.1643, "step": 518 }, { "epoch": 0.018288009725415576, "grad_norm": 2.642794132232666, "learning_rate": 1.2065727699530519e-05, "loss": 4.2733, "step": 519 }, { "epoch": 0.01832324673837399, "grad_norm": 1.8666173219680786, "learning_rate": 1.2089201877934273e-05, "loss": 1.1705, "step": 520 }, { "epoch": 0.0183584837513324, "grad_norm": 2.4644558429718018, "learning_rate": 1.211267605633803e-05, "loss": 4.062, "step": 521 }, { "epoch": 0.01839372076429081, "grad_norm": 1.881150484085083, "learning_rate": 1.2136150234741785e-05, "loss": 4.7448, "step": 522 }, { "epoch": 0.018428957777249223, "grad_norm": 2.204775094985962, "learning_rate": 1.2159624413145541e-05, "loss": 1.097, "step": 523 }, { "epoch": 0.018464194790207635, "grad_norm": 2.6403822898864746, "learning_rate": 1.2183098591549296e-05, "loss": 4.4778, "step": 524 }, { "epoch": 0.018499431803166044, "grad_norm": 1.0843554735183716, "learning_rate": 1.2206572769953053e-05, "loss": 0.7853, "step": 525 }, { "epoch": 0.018534668816124457, "grad_norm": 2.0703530311584473, "learning_rate": 1.2230046948356808e-05, "loss": 3.5935, "step": 526 }, { "epoch": 0.01856990582908287, "grad_norm": 1.7714200019836426, "learning_rate": 1.2253521126760564e-05, "loss": 1.1817, "step": 527 }, { "epoch": 0.01860514284204128, "grad_norm": 2.4104294776916504, "learning_rate": 1.2276995305164322e-05, "loss": 1.0964, "step": 528 }, { "epoch": 0.01864037985499969, "grad_norm": 2.4058494567871094, "learning_rate": 1.2300469483568076e-05, "loss": 1.0416, "step": 529 }, { "epoch": 0.018675616867958104, "grad_norm": 2.6000137329101562, "learning_rate": 1.232394366197183e-05, "loss": 4.1224, "step": 530 }, { "epoch": 0.018710853880916516, "grad_norm": 1.4072374105453491, "learning_rate": 1.2347417840375588e-05, "loss": 0.7826, "step": 531 }, { "epoch": 0.018746090893874925, "grad_norm": 1.622068166732788, "learning_rate": 1.2370892018779344e-05, "loss": 1.0257, "step": 532 }, { "epoch": 0.018781327906833338, "grad_norm": 2.2236247062683105, "learning_rate": 1.2394366197183098e-05, "loss": 3.5311, "step": 533 }, { "epoch": 0.01881656491979175, "grad_norm": 2.397815227508545, "learning_rate": 1.2417840375586856e-05, "loss": 1.0609, "step": 534 }, { "epoch": 0.01885180193275016, "grad_norm": 2.2290918827056885, "learning_rate": 1.244131455399061e-05, "loss": 7.0563, "step": 535 }, { "epoch": 0.018887038945708572, "grad_norm": 1.546568751335144, "learning_rate": 1.2464788732394367e-05, "loss": 0.953, "step": 536 }, { "epoch": 0.018922275958666985, "grad_norm": 2.783137321472168, "learning_rate": 1.2488262910798124e-05, "loss": 3.7916, "step": 537 }, { "epoch": 0.018957512971625394, "grad_norm": 1.1988813877105713, "learning_rate": 1.2511737089201879e-05, "loss": 1.1462, "step": 538 }, { "epoch": 0.018992749984583807, "grad_norm": 2.1787221431732178, "learning_rate": 1.2535211267605636e-05, "loss": 1.2452, "step": 539 }, { "epoch": 0.01902798699754222, "grad_norm": 1.824163556098938, "learning_rate": 1.255868544600939e-05, "loss": 4.1842, "step": 540 }, { "epoch": 0.01906322401050063, "grad_norm": 2.197211265563965, "learning_rate": 1.2582159624413147e-05, "loss": 3.8044, "step": 541 }, { "epoch": 0.01909846102345904, "grad_norm": 1.8718222379684448, "learning_rate": 1.2605633802816903e-05, "loss": 1.0141, "step": 542 }, { "epoch": 0.019133698036417453, "grad_norm": 1.5270521640777588, "learning_rate": 1.2629107981220659e-05, "loss": 0.9052, "step": 543 }, { "epoch": 0.019168935049375866, "grad_norm": 2.2924695014953613, "learning_rate": 1.2652582159624413e-05, "loss": 4.1063, "step": 544 }, { "epoch": 0.019204172062334275, "grad_norm": 2.0031983852386475, "learning_rate": 1.2676056338028171e-05, "loss": 0.9099, "step": 545 }, { "epoch": 0.019239409075292688, "grad_norm": 1.6996504068374634, "learning_rate": 1.2699530516431925e-05, "loss": 1.2002, "step": 546 }, { "epoch": 0.0192746460882511, "grad_norm": 1.9933313131332397, "learning_rate": 1.2723004694835681e-05, "loss": 3.8606, "step": 547 }, { "epoch": 0.01930988310120951, "grad_norm": 1.9132474660873413, "learning_rate": 1.2746478873239439e-05, "loss": 0.8884, "step": 548 }, { "epoch": 0.019345120114167922, "grad_norm": 2.938202142715454, "learning_rate": 1.2769953051643193e-05, "loss": 6.8812, "step": 549 }, { "epoch": 0.019380357127126335, "grad_norm": 2.6450695991516113, "learning_rate": 1.279342723004695e-05, "loss": 3.5772, "step": 550 }, { "epoch": 0.019415594140084744, "grad_norm": 3.789970636367798, "learning_rate": 1.2816901408450705e-05, "loss": 4.1048, "step": 551 }, { "epoch": 0.019450831153043156, "grad_norm": 2.2701101303100586, "learning_rate": 1.2840375586854462e-05, "loss": 0.9126, "step": 552 }, { "epoch": 0.01948606816600157, "grad_norm": 5.181969165802002, "learning_rate": 1.2863849765258216e-05, "loss": 6.902, "step": 553 }, { "epoch": 0.019521305178959978, "grad_norm": 1.8457310199737549, "learning_rate": 1.2887323943661974e-05, "loss": 1.5359, "step": 554 }, { "epoch": 0.01955654219191839, "grad_norm": 1.7909839153289795, "learning_rate": 1.2910798122065728e-05, "loss": 1.1597, "step": 555 }, { "epoch": 0.019591779204876803, "grad_norm": 2.837132453918457, "learning_rate": 1.2934272300469484e-05, "loss": 7.1933, "step": 556 }, { "epoch": 0.019627016217835216, "grad_norm": 1.8798542022705078, "learning_rate": 1.2957746478873242e-05, "loss": 1.0815, "step": 557 }, { "epoch": 0.019662253230793625, "grad_norm": 2.303579092025757, "learning_rate": 1.2981220657276996e-05, "loss": 3.3896, "step": 558 }, { "epoch": 0.019697490243752037, "grad_norm": 1.9496334791183472, "learning_rate": 1.300469483568075e-05, "loss": 0.8287, "step": 559 }, { "epoch": 0.01973272725671045, "grad_norm": 1.708465337753296, "learning_rate": 1.3028169014084508e-05, "loss": 1.1299, "step": 560 }, { "epoch": 0.01976796426966886, "grad_norm": 2.445086717605591, "learning_rate": 1.3051643192488264e-05, "loss": 1.2154, "step": 561 }, { "epoch": 0.01980320128262727, "grad_norm": 2.337462902069092, "learning_rate": 1.307511737089202e-05, "loss": 3.2183, "step": 562 }, { "epoch": 0.019838438295585684, "grad_norm": 2.3560197353363037, "learning_rate": 1.3098591549295776e-05, "loss": 3.7306, "step": 563 }, { "epoch": 0.019873675308544093, "grad_norm": 3.2165396213531494, "learning_rate": 1.312206572769953e-05, "loss": 7.2378, "step": 564 }, { "epoch": 0.019908912321502506, "grad_norm": 1.739155888557434, "learning_rate": 1.3145539906103288e-05, "loss": 4.1839, "step": 565 }, { "epoch": 0.01994414933446092, "grad_norm": 1.78290593624115, "learning_rate": 1.3169014084507044e-05, "loss": 1.1701, "step": 566 }, { "epoch": 0.019979386347419328, "grad_norm": 2.558353900909424, "learning_rate": 1.3192488262910799e-05, "loss": 3.923, "step": 567 }, { "epoch": 0.02001462336037774, "grad_norm": 3.4681012630462646, "learning_rate": 1.3215962441314556e-05, "loss": 9.7171, "step": 568 }, { "epoch": 0.020049860373336153, "grad_norm": 2.463831663131714, "learning_rate": 1.323943661971831e-05, "loss": 3.9231, "step": 569 }, { "epoch": 0.020085097386294565, "grad_norm": 1.931047797203064, "learning_rate": 1.3262910798122067e-05, "loss": 3.2394, "step": 570 }, { "epoch": 0.020120334399252975, "grad_norm": 2.2278482913970947, "learning_rate": 1.3286384976525823e-05, "loss": 3.8854, "step": 571 }, { "epoch": 0.020155571412211387, "grad_norm": 1.444270133972168, "learning_rate": 1.3309859154929579e-05, "loss": 0.8684, "step": 572 }, { "epoch": 0.0201908084251698, "grad_norm": 2.7176051139831543, "learning_rate": 1.3333333333333333e-05, "loss": 3.781, "step": 573 }, { "epoch": 0.02022604543812821, "grad_norm": 3.3386917114257812, "learning_rate": 1.3356807511737091e-05, "loss": 4.1772, "step": 574 }, { "epoch": 0.02026128245108662, "grad_norm": 3.9813592433929443, "learning_rate": 1.3380281690140845e-05, "loss": 4.1806, "step": 575 }, { "epoch": 0.020296519464045034, "grad_norm": 2.6869614124298096, "learning_rate": 1.3403755868544601e-05, "loss": 3.7785, "step": 576 }, { "epoch": 0.020331756477003443, "grad_norm": 1.7403686046600342, "learning_rate": 1.342723004694836e-05, "loss": 0.9667, "step": 577 }, { "epoch": 0.020366993489961856, "grad_norm": 3.4053683280944824, "learning_rate": 1.3450704225352114e-05, "loss": 3.8423, "step": 578 }, { "epoch": 0.02040223050292027, "grad_norm": 2.067662477493286, "learning_rate": 1.347417840375587e-05, "loss": 3.8567, "step": 579 }, { "epoch": 0.020437467515878677, "grad_norm": 2.010652780532837, "learning_rate": 1.3497652582159626e-05, "loss": 1.1863, "step": 580 }, { "epoch": 0.02047270452883709, "grad_norm": 2.2038936614990234, "learning_rate": 1.3521126760563382e-05, "loss": 3.7429, "step": 581 }, { "epoch": 0.020507941541795503, "grad_norm": 1.6560989618301392, "learning_rate": 1.3544600938967136e-05, "loss": 1.3637, "step": 582 }, { "epoch": 0.020543178554753915, "grad_norm": 1.9818902015686035, "learning_rate": 1.3568075117370894e-05, "loss": 1.076, "step": 583 }, { "epoch": 0.020578415567712324, "grad_norm": 1.7128766775131226, "learning_rate": 1.3591549295774648e-05, "loss": 1.095, "step": 584 }, { "epoch": 0.020613652580670737, "grad_norm": 1.21425199508667, "learning_rate": 1.3615023474178406e-05, "loss": 1.0524, "step": 585 }, { "epoch": 0.02064888959362915, "grad_norm": 1.341127634048462, "learning_rate": 1.3638497652582162e-05, "loss": 1.0995, "step": 586 }, { "epoch": 0.02068412660658756, "grad_norm": 1.3663300275802612, "learning_rate": 1.3661971830985916e-05, "loss": 0.9349, "step": 587 }, { "epoch": 0.02071936361954597, "grad_norm": 1.5194741487503052, "learning_rate": 1.3685446009389674e-05, "loss": 0.8476, "step": 588 }, { "epoch": 0.020754600632504384, "grad_norm": 2.238875150680542, "learning_rate": 1.3708920187793428e-05, "loss": 3.81, "step": 589 }, { "epoch": 0.020789837645462793, "grad_norm": 1.9170901775360107, "learning_rate": 1.3732394366197184e-05, "loss": 4.4648, "step": 590 }, { "epoch": 0.020825074658421205, "grad_norm": 1.653408408164978, "learning_rate": 1.375586854460094e-05, "loss": 1.0573, "step": 591 }, { "epoch": 0.020860311671379618, "grad_norm": 2.187702178955078, "learning_rate": 1.3779342723004696e-05, "loss": 0.9657, "step": 592 }, { "epoch": 0.020895548684338027, "grad_norm": 1.1809583902359009, "learning_rate": 1.380281690140845e-05, "loss": 1.0564, "step": 593 }, { "epoch": 0.02093078569729644, "grad_norm": 1.5607110261917114, "learning_rate": 1.3826291079812208e-05, "loss": 3.9586, "step": 594 }, { "epoch": 0.020966022710254852, "grad_norm": 2.069201707839966, "learning_rate": 1.3849765258215963e-05, "loss": 3.3487, "step": 595 }, { "epoch": 0.021001259723213265, "grad_norm": 2.2882776260375977, "learning_rate": 1.3873239436619719e-05, "loss": 3.8754, "step": 596 }, { "epoch": 0.021036496736171674, "grad_norm": 1.50045645236969, "learning_rate": 1.3896713615023477e-05, "loss": 0.9249, "step": 597 }, { "epoch": 0.021071733749130087, "grad_norm": 1.5206578969955444, "learning_rate": 1.3920187793427231e-05, "loss": 0.9524, "step": 598 }, { "epoch": 0.0211069707620885, "grad_norm": 3.3704073429107666, "learning_rate": 1.3943661971830987e-05, "loss": 3.5343, "step": 599 }, { "epoch": 0.021142207775046908, "grad_norm": 2.256195306777954, "learning_rate": 1.3967136150234743e-05, "loss": 3.5598, "step": 600 }, { "epoch": 0.02117744478800532, "grad_norm": 2.012640953063965, "learning_rate": 1.3990610328638499e-05, "loss": 4.123, "step": 601 }, { "epoch": 0.021212681800963733, "grad_norm": 3.030730724334717, "learning_rate": 1.4014084507042253e-05, "loss": 3.6097, "step": 602 }, { "epoch": 0.021247918813922143, "grad_norm": 1.446882963180542, "learning_rate": 1.4037558685446011e-05, "loss": 1.0058, "step": 603 }, { "epoch": 0.021283155826880555, "grad_norm": 2.521361827850342, "learning_rate": 1.4061032863849765e-05, "loss": 3.8577, "step": 604 }, { "epoch": 0.021318392839838968, "grad_norm": 3.0622687339782715, "learning_rate": 1.4084507042253522e-05, "loss": 3.8734, "step": 605 }, { "epoch": 0.021353629852797377, "grad_norm": 1.6843767166137695, "learning_rate": 1.410798122065728e-05, "loss": 1.0002, "step": 606 }, { "epoch": 0.02138886686575579, "grad_norm": 3.1289873123168945, "learning_rate": 1.4131455399061034e-05, "loss": 7.6327, "step": 607 }, { "epoch": 0.021424103878714202, "grad_norm": 2.1347458362579346, "learning_rate": 1.4154929577464788e-05, "loss": 3.707, "step": 608 }, { "epoch": 0.021459340891672615, "grad_norm": 2.815270185470581, "learning_rate": 1.4178403755868546e-05, "loss": 4.3108, "step": 609 }, { "epoch": 0.021494577904631024, "grad_norm": 2.308272123336792, "learning_rate": 1.4201877934272302e-05, "loss": 3.6828, "step": 610 }, { "epoch": 0.021529814917589436, "grad_norm": 1.4634202718734741, "learning_rate": 1.4225352112676058e-05, "loss": 0.968, "step": 611 }, { "epoch": 0.02156505193054785, "grad_norm": 3.3291776180267334, "learning_rate": 1.4248826291079814e-05, "loss": 0.9475, "step": 612 }, { "epoch": 0.021600288943506258, "grad_norm": 2.6378748416900635, "learning_rate": 1.4272300469483568e-05, "loss": 4.6062, "step": 613 }, { "epoch": 0.02163552595646467, "grad_norm": 1.685525894165039, "learning_rate": 1.4295774647887326e-05, "loss": 4.085, "step": 614 }, { "epoch": 0.021670762969423083, "grad_norm": 1.500023365020752, "learning_rate": 1.4319248826291082e-05, "loss": 0.8548, "step": 615 }, { "epoch": 0.021705999982381492, "grad_norm": 2.315807819366455, "learning_rate": 1.4342723004694836e-05, "loss": 4.3234, "step": 616 }, { "epoch": 0.021741236995339905, "grad_norm": 1.94193696975708, "learning_rate": 1.4366197183098594e-05, "loss": 1.0291, "step": 617 }, { "epoch": 0.021776474008298317, "grad_norm": 1.9619560241699219, "learning_rate": 1.4389671361502348e-05, "loss": 0.8463, "step": 618 }, { "epoch": 0.021811711021256727, "grad_norm": 2.2542669773101807, "learning_rate": 1.4413145539906104e-05, "loss": 3.9162, "step": 619 }, { "epoch": 0.02184694803421514, "grad_norm": 2.2518725395202637, "learning_rate": 1.443661971830986e-05, "loss": 3.6261, "step": 620 }, { "epoch": 0.02188218504717355, "grad_norm": 1.3220428228378296, "learning_rate": 1.4460093896713617e-05, "loss": 0.8828, "step": 621 }, { "epoch": 0.021917422060131964, "grad_norm": 1.6645944118499756, "learning_rate": 1.4483568075117371e-05, "loss": 0.7838, "step": 622 }, { "epoch": 0.021952659073090373, "grad_norm": 3.603827953338623, "learning_rate": 1.4507042253521129e-05, "loss": 4.4941, "step": 623 }, { "epoch": 0.021987896086048786, "grad_norm": 1.3673815727233887, "learning_rate": 1.4530516431924883e-05, "loss": 0.9616, "step": 624 }, { "epoch": 0.0220231330990072, "grad_norm": 2.322751045227051, "learning_rate": 1.4553990610328639e-05, "loss": 3.3782, "step": 625 }, { "epoch": 0.022058370111965608, "grad_norm": 2.14922833442688, "learning_rate": 1.4577464788732397e-05, "loss": 4.1354, "step": 626 }, { "epoch": 0.02209360712492402, "grad_norm": 2.0844218730926514, "learning_rate": 1.4600938967136151e-05, "loss": 1.0457, "step": 627 }, { "epoch": 0.022128844137882433, "grad_norm": 1.7212529182434082, "learning_rate": 1.4624413145539907e-05, "loss": 3.837, "step": 628 }, { "epoch": 0.022164081150840842, "grad_norm": 1.4873906373977661, "learning_rate": 1.4647887323943663e-05, "loss": 0.9677, "step": 629 }, { "epoch": 0.022199318163799255, "grad_norm": 2.014631986618042, "learning_rate": 1.467136150234742e-05, "loss": 4.0342, "step": 630 }, { "epoch": 0.022234555176757667, "grad_norm": 1.8279117345809937, "learning_rate": 1.4694835680751174e-05, "loss": 3.6022, "step": 631 }, { "epoch": 0.022269792189716076, "grad_norm": 1.7127782106399536, "learning_rate": 1.4718309859154931e-05, "loss": 4.2146, "step": 632 }, { "epoch": 0.02230502920267449, "grad_norm": 2.0631966590881348, "learning_rate": 1.4741784037558686e-05, "loss": 4.3153, "step": 633 }, { "epoch": 0.0223402662156329, "grad_norm": 1.5217581987380981, "learning_rate": 1.4765258215962443e-05, "loss": 1.1995, "step": 634 }, { "epoch": 0.022375503228591314, "grad_norm": 1.7803494930267334, "learning_rate": 1.47887323943662e-05, "loss": 1.2649, "step": 635 }, { "epoch": 0.022410740241549723, "grad_norm": 1.8644113540649414, "learning_rate": 1.4812206572769954e-05, "loss": 4.6708, "step": 636 }, { "epoch": 0.022445977254508136, "grad_norm": 1.7283841371536255, "learning_rate": 1.4835680751173711e-05, "loss": 0.9505, "step": 637 }, { "epoch": 0.02248121426746655, "grad_norm": 1.9270812273025513, "learning_rate": 1.4859154929577466e-05, "loss": 6.9312, "step": 638 }, { "epoch": 0.022516451280424957, "grad_norm": 1.5444140434265137, "learning_rate": 1.4882629107981222e-05, "loss": 1.1071, "step": 639 }, { "epoch": 0.02255168829338337, "grad_norm": 1.6869206428527832, "learning_rate": 1.4906103286384978e-05, "loss": 1.0206, "step": 640 }, { "epoch": 0.022586925306341783, "grad_norm": 1.9546178579330444, "learning_rate": 1.4929577464788734e-05, "loss": 7.254, "step": 641 }, { "epoch": 0.02262216231930019, "grad_norm": 1.8448961973190308, "learning_rate": 1.4953051643192488e-05, "loss": 3.9368, "step": 642 }, { "epoch": 0.022657399332258604, "grad_norm": 1.346551537513733, "learning_rate": 1.4976525821596246e-05, "loss": 1.0723, "step": 643 }, { "epoch": 0.022692636345217017, "grad_norm": 2.0582263469696045, "learning_rate": 1.5000000000000002e-05, "loss": 3.088, "step": 644 }, { "epoch": 0.022727873358175426, "grad_norm": 2.2102363109588623, "learning_rate": 1.5023474178403756e-05, "loss": 4.2011, "step": 645 }, { "epoch": 0.02276311037113384, "grad_norm": 1.4814493656158447, "learning_rate": 1.5046948356807514e-05, "loss": 1.1182, "step": 646 }, { "epoch": 0.02279834738409225, "grad_norm": 2.383024215698242, "learning_rate": 1.5070422535211269e-05, "loss": 1.1053, "step": 647 }, { "epoch": 0.022833584397050664, "grad_norm": 2.551076889038086, "learning_rate": 1.5093896713615025e-05, "loss": 7.1403, "step": 648 }, { "epoch": 0.022868821410009073, "grad_norm": 2.865299940109253, "learning_rate": 1.511737089201878e-05, "loss": 3.2062, "step": 649 }, { "epoch": 0.022904058422967485, "grad_norm": 2.8811817169189453, "learning_rate": 1.5140845070422537e-05, "loss": 4.6184, "step": 650 }, { "epoch": 0.022939295435925898, "grad_norm": 2.510314464569092, "learning_rate": 1.5164319248826291e-05, "loss": 3.1954, "step": 651 }, { "epoch": 0.022974532448884307, "grad_norm": 2.159675359725952, "learning_rate": 1.5187793427230049e-05, "loss": 1.1094, "step": 652 }, { "epoch": 0.02300976946184272, "grad_norm": 2.2785027027130127, "learning_rate": 1.5211267605633803e-05, "loss": 3.7168, "step": 653 }, { "epoch": 0.023045006474801132, "grad_norm": 2.344705104827881, "learning_rate": 1.5234741784037559e-05, "loss": 7.3052, "step": 654 }, { "epoch": 0.02308024348775954, "grad_norm": 2.6189637184143066, "learning_rate": 1.5258215962441317e-05, "loss": 0.93, "step": 655 }, { "epoch": 0.023115480500717954, "grad_norm": 2.33552885055542, "learning_rate": 1.528169014084507e-05, "loss": 1.0454, "step": 656 }, { "epoch": 0.023150717513676367, "grad_norm": 2.603403091430664, "learning_rate": 1.5305164319248827e-05, "loss": 0.8948, "step": 657 }, { "epoch": 0.023185954526634776, "grad_norm": 2.2928755283355713, "learning_rate": 1.5328638497652583e-05, "loss": 6.6777, "step": 658 }, { "epoch": 0.023221191539593188, "grad_norm": 1.635783314704895, "learning_rate": 1.535211267605634e-05, "loss": 1.1593, "step": 659 }, { "epoch": 0.0232564285525516, "grad_norm": 2.2760469913482666, "learning_rate": 1.5375586854460095e-05, "loss": 6.9806, "step": 660 }, { "epoch": 0.023291665565510013, "grad_norm": 1.8654148578643799, "learning_rate": 1.539906103286385e-05, "loss": 1.1611, "step": 661 }, { "epoch": 0.023326902578468423, "grad_norm": 2.0459015369415283, "learning_rate": 1.5422535211267607e-05, "loss": 3.7196, "step": 662 }, { "epoch": 0.023362139591426835, "grad_norm": 2.2385685443878174, "learning_rate": 1.5446009389671363e-05, "loss": 7.3623, "step": 663 }, { "epoch": 0.023397376604385248, "grad_norm": 1.5783571004867554, "learning_rate": 1.546948356807512e-05, "loss": 0.9182, "step": 664 }, { "epoch": 0.023432613617343657, "grad_norm": 1.2612740993499756, "learning_rate": 1.5492957746478872e-05, "loss": 1.0347, "step": 665 }, { "epoch": 0.02346785063030207, "grad_norm": 2.3302321434020996, "learning_rate": 1.551643192488263e-05, "loss": 3.491, "step": 666 }, { "epoch": 0.023503087643260482, "grad_norm": 1.512233018875122, "learning_rate": 1.5539906103286388e-05, "loss": 1.1495, "step": 667 }, { "epoch": 0.02353832465621889, "grad_norm": 1.7155627012252808, "learning_rate": 1.556338028169014e-05, "loss": 1.1367, "step": 668 }, { "epoch": 0.023573561669177304, "grad_norm": 3.030837059020996, "learning_rate": 1.55868544600939e-05, "loss": 6.4449, "step": 669 }, { "epoch": 0.023608798682135716, "grad_norm": 3.225309371948242, "learning_rate": 1.5610328638497652e-05, "loss": 6.5257, "step": 670 }, { "epoch": 0.023644035695094125, "grad_norm": 2.0286216735839844, "learning_rate": 1.563380281690141e-05, "loss": 1.1819, "step": 671 }, { "epoch": 0.023679272708052538, "grad_norm": 1.6989688873291016, "learning_rate": 1.5657276995305168e-05, "loss": 1.2297, "step": 672 }, { "epoch": 0.02371450972101095, "grad_norm": 1.5795015096664429, "learning_rate": 1.568075117370892e-05, "loss": 1.0524, "step": 673 }, { "epoch": 0.023749746733969363, "grad_norm": 3.2831804752349854, "learning_rate": 1.5704225352112677e-05, "loss": 4.1251, "step": 674 }, { "epoch": 0.023784983746927772, "grad_norm": 1.5111101865768433, "learning_rate": 1.5727699530516433e-05, "loss": 0.906, "step": 675 }, { "epoch": 0.023820220759886185, "grad_norm": 1.7958292961120605, "learning_rate": 1.575117370892019e-05, "loss": 1.1267, "step": 676 }, { "epoch": 0.023855457772844597, "grad_norm": 2.9234678745269775, "learning_rate": 1.5774647887323945e-05, "loss": 6.8918, "step": 677 }, { "epoch": 0.023890694785803007, "grad_norm": 2.4186301231384277, "learning_rate": 1.57981220657277e-05, "loss": 3.4114, "step": 678 }, { "epoch": 0.02392593179876142, "grad_norm": 2.8827733993530273, "learning_rate": 1.5821596244131457e-05, "loss": 6.1226, "step": 679 }, { "epoch": 0.02396116881171983, "grad_norm": 2.3317391872406006, "learning_rate": 1.5845070422535213e-05, "loss": 2.8618, "step": 680 }, { "epoch": 0.02399640582467824, "grad_norm": 1.7705028057098389, "learning_rate": 1.586854460093897e-05, "loss": 1.2495, "step": 681 }, { "epoch": 0.024031642837636653, "grad_norm": 1.4550143480300903, "learning_rate": 1.5892018779342725e-05, "loss": 1.0079, "step": 682 }, { "epoch": 0.024066879850595066, "grad_norm": 2.3505492210388184, "learning_rate": 1.591549295774648e-05, "loss": 3.6366, "step": 683 }, { "epoch": 0.024102116863553475, "grad_norm": 1.7696387767791748, "learning_rate": 1.5938967136150237e-05, "loss": 0.873, "step": 684 }, { "epoch": 0.024137353876511888, "grad_norm": 2.624899387359619, "learning_rate": 1.5962441314553993e-05, "loss": 6.1213, "step": 685 }, { "epoch": 0.0241725908894703, "grad_norm": 1.975450873374939, "learning_rate": 1.598591549295775e-05, "loss": 3.6793, "step": 686 }, { "epoch": 0.024207827902428713, "grad_norm": 2.0466933250427246, "learning_rate": 1.6009389671361505e-05, "loss": 4.2292, "step": 687 }, { "epoch": 0.024243064915387122, "grad_norm": 2.0466341972351074, "learning_rate": 1.6032863849765258e-05, "loss": 1.0523, "step": 688 }, { "epoch": 0.024278301928345535, "grad_norm": 1.7176352739334106, "learning_rate": 1.6056338028169017e-05, "loss": 3.7309, "step": 689 }, { "epoch": 0.024313538941303947, "grad_norm": 1.3854635953903198, "learning_rate": 1.607981220657277e-05, "loss": 1.0059, "step": 690 }, { "epoch": 0.024348775954262356, "grad_norm": 1.7026824951171875, "learning_rate": 1.6103286384976526e-05, "loss": 0.9943, "step": 691 }, { "epoch": 0.02438401296722077, "grad_norm": 2.8511431217193604, "learning_rate": 1.6126760563380285e-05, "loss": 9.7093, "step": 692 }, { "epoch": 0.02441924998017918, "grad_norm": 1.8882349729537964, "learning_rate": 1.6150234741784038e-05, "loss": 0.9647, "step": 693 }, { "epoch": 0.02445448699313759, "grad_norm": 2.2311203479766846, "learning_rate": 1.6173708920187794e-05, "loss": 7.0718, "step": 694 }, { "epoch": 0.024489724006096003, "grad_norm": 2.3665354251861572, "learning_rate": 1.619718309859155e-05, "loss": 6.4556, "step": 695 }, { "epoch": 0.024524961019054416, "grad_norm": 1.7809942960739136, "learning_rate": 1.6220657276995306e-05, "loss": 1.0032, "step": 696 }, { "epoch": 0.024560198032012825, "grad_norm": 1.610893964767456, "learning_rate": 1.6244131455399062e-05, "loss": 1.1135, "step": 697 }, { "epoch": 0.024595435044971237, "grad_norm": 3.0235540866851807, "learning_rate": 1.6267605633802818e-05, "loss": 6.9685, "step": 698 }, { "epoch": 0.02463067205792965, "grad_norm": 1.9014040231704712, "learning_rate": 1.6291079812206574e-05, "loss": 3.696, "step": 699 }, { "epoch": 0.024665909070888063, "grad_norm": 2.1298091411590576, "learning_rate": 1.631455399061033e-05, "loss": 3.8325, "step": 700 }, { "epoch": 0.02470114608384647, "grad_norm": 2.2413828372955322, "learning_rate": 1.6338028169014086e-05, "loss": 3.6655, "step": 701 }, { "epoch": 0.024736383096804884, "grad_norm": 3.0543296337127686, "learning_rate": 1.6361502347417842e-05, "loss": 5.9045, "step": 702 }, { "epoch": 0.024771620109763297, "grad_norm": 1.316107153892517, "learning_rate": 1.6384976525821595e-05, "loss": 1.1633, "step": 703 }, { "epoch": 0.024806857122721706, "grad_norm": 2.6087646484375, "learning_rate": 1.6408450704225354e-05, "loss": 0.9333, "step": 704 }, { "epoch": 0.02484209413568012, "grad_norm": 3.50797176361084, "learning_rate": 1.643192488262911e-05, "loss": 3.8888, "step": 705 }, { "epoch": 0.02487733114863853, "grad_norm": 2.03405499458313, "learning_rate": 1.6455399061032866e-05, "loss": 4.1471, "step": 706 }, { "epoch": 0.02491256816159694, "grad_norm": 2.010382890701294, "learning_rate": 1.6478873239436623e-05, "loss": 1.0296, "step": 707 }, { "epoch": 0.024947805174555353, "grad_norm": 2.987788200378418, "learning_rate": 1.6502347417840375e-05, "loss": 3.9951, "step": 708 }, { "epoch": 0.024983042187513765, "grad_norm": 1.3794481754302979, "learning_rate": 1.6525821596244135e-05, "loss": 0.9957, "step": 709 }, { "epoch": 0.025018279200472174, "grad_norm": 3.1363461017608643, "learning_rate": 1.6549295774647887e-05, "loss": 4.5911, "step": 710 }, { "epoch": 0.025053516213430587, "grad_norm": 1.5481754541397095, "learning_rate": 1.6572769953051643e-05, "loss": 0.808, "step": 711 }, { "epoch": 0.025088753226389, "grad_norm": 2.5229387283325195, "learning_rate": 1.6596244131455403e-05, "loss": 3.7298, "step": 712 }, { "epoch": 0.025123990239347412, "grad_norm": 3.5095043182373047, "learning_rate": 1.6619718309859155e-05, "loss": 3.7259, "step": 713 }, { "epoch": 0.02515922725230582, "grad_norm": 2.2013728618621826, "learning_rate": 1.664319248826291e-05, "loss": 1.0494, "step": 714 }, { "epoch": 0.025194464265264234, "grad_norm": 2.4440062046051025, "learning_rate": 1.6666666666666667e-05, "loss": 6.4815, "step": 715 }, { "epoch": 0.025229701278222647, "grad_norm": 1.749119758605957, "learning_rate": 1.6690140845070424e-05, "loss": 0.942, "step": 716 }, { "epoch": 0.025264938291181056, "grad_norm": 2.8586819171905518, "learning_rate": 1.671361502347418e-05, "loss": 5.5803, "step": 717 }, { "epoch": 0.025300175304139468, "grad_norm": 3.008275270462036, "learning_rate": 1.6737089201877936e-05, "loss": 7.0398, "step": 718 }, { "epoch": 0.02533541231709788, "grad_norm": 1.8712269067764282, "learning_rate": 1.676056338028169e-05, "loss": 1.0693, "step": 719 }, { "epoch": 0.02537064933005629, "grad_norm": 2.775736093521118, "learning_rate": 1.6784037558685448e-05, "loss": 4.2223, "step": 720 }, { "epoch": 0.025405886343014702, "grad_norm": 3.098756790161133, "learning_rate": 1.6807511737089204e-05, "loss": 0.8921, "step": 721 }, { "epoch": 0.025441123355973115, "grad_norm": 2.21622371673584, "learning_rate": 1.683098591549296e-05, "loss": 1.0185, "step": 722 }, { "epoch": 0.025476360368931524, "grad_norm": 1.1401466131210327, "learning_rate": 1.6854460093896712e-05, "loss": 0.9812, "step": 723 }, { "epoch": 0.025511597381889937, "grad_norm": 1.491200566291809, "learning_rate": 1.6877934272300472e-05, "loss": 0.6784, "step": 724 }, { "epoch": 0.02554683439484835, "grad_norm": 1.6203267574310303, "learning_rate": 1.6901408450704228e-05, "loss": 1.0695, "step": 725 }, { "epoch": 0.025582071407806762, "grad_norm": 2.297013282775879, "learning_rate": 1.692488262910798e-05, "loss": 4.1346, "step": 726 }, { "epoch": 0.02561730842076517, "grad_norm": 5.545126914978027, "learning_rate": 1.694835680751174e-05, "loss": 8.7276, "step": 727 }, { "epoch": 0.025652545433723584, "grad_norm": 1.7673890590667725, "learning_rate": 1.6971830985915493e-05, "loss": 1.0397, "step": 728 }, { "epoch": 0.025687782446681996, "grad_norm": 2.8243765830993652, "learning_rate": 1.6995305164319252e-05, "loss": 1.0309, "step": 729 }, { "epoch": 0.025723019459640405, "grad_norm": 2.9238057136535645, "learning_rate": 1.7018779342723005e-05, "loss": 3.851, "step": 730 }, { "epoch": 0.025758256472598818, "grad_norm": 1.4761543273925781, "learning_rate": 1.704225352112676e-05, "loss": 1.1291, "step": 731 }, { "epoch": 0.02579349348555723, "grad_norm": 2.6789634227752686, "learning_rate": 1.706572769953052e-05, "loss": 3.8523, "step": 732 }, { "epoch": 0.02582873049851564, "grad_norm": 1.793650507926941, "learning_rate": 1.7089201877934273e-05, "loss": 1.3889, "step": 733 }, { "epoch": 0.025863967511474052, "grad_norm": 4.157182693481445, "learning_rate": 1.711267605633803e-05, "loss": 6.594, "step": 734 }, { "epoch": 0.025899204524432465, "grad_norm": 2.6624228954315186, "learning_rate": 1.7136150234741785e-05, "loss": 3.247, "step": 735 }, { "epoch": 0.025934441537390874, "grad_norm": 2.125337600708008, "learning_rate": 1.715962441314554e-05, "loss": 1.0845, "step": 736 }, { "epoch": 0.025969678550349286, "grad_norm": 1.424472689628601, "learning_rate": 1.7183098591549297e-05, "loss": 1.012, "step": 737 }, { "epoch": 0.0260049155633077, "grad_norm": 1.545397162437439, "learning_rate": 1.7206572769953053e-05, "loss": 1.1031, "step": 738 }, { "epoch": 0.02604015257626611, "grad_norm": 1.830191969871521, "learning_rate": 1.723004694835681e-05, "loss": 1.0566, "step": 739 }, { "epoch": 0.02607538958922452, "grad_norm": 7.040078639984131, "learning_rate": 1.7253521126760565e-05, "loss": 6.727, "step": 740 }, { "epoch": 0.026110626602182933, "grad_norm": 1.4132791757583618, "learning_rate": 1.727699530516432e-05, "loss": 1.154, "step": 741 }, { "epoch": 0.026145863615141346, "grad_norm": 3.6807665824890137, "learning_rate": 1.7300469483568077e-05, "loss": 8.9236, "step": 742 }, { "epoch": 0.026181100628099755, "grad_norm": 3.0807037353515625, "learning_rate": 1.732394366197183e-05, "loss": 3.572, "step": 743 }, { "epoch": 0.026216337641058168, "grad_norm": 1.725850224494934, "learning_rate": 1.734741784037559e-05, "loss": 1.2672, "step": 744 }, { "epoch": 0.02625157465401658, "grad_norm": 1.8143126964569092, "learning_rate": 1.7370892018779345e-05, "loss": 1.0701, "step": 745 }, { "epoch": 0.02628681166697499, "grad_norm": 1.7821022272109985, "learning_rate": 1.7394366197183098e-05, "loss": 4.3324, "step": 746 }, { "epoch": 0.026322048679933402, "grad_norm": 2.142144203186035, "learning_rate": 1.7417840375586857e-05, "loss": 1.279, "step": 747 }, { "epoch": 0.026357285692891814, "grad_norm": 1.8633967638015747, "learning_rate": 1.744131455399061e-05, "loss": 4.1563, "step": 748 }, { "epoch": 0.026392522705850224, "grad_norm": 1.7867779731750488, "learning_rate": 1.7464788732394366e-05, "loss": 3.9348, "step": 749 }, { "epoch": 0.026427759718808636, "grad_norm": 1.8388991355895996, "learning_rate": 1.7488262910798126e-05, "loss": 3.988, "step": 750 }, { "epoch": 0.02646299673176705, "grad_norm": 2.114176034927368, "learning_rate": 1.7511737089201878e-05, "loss": 7.2867, "step": 751 }, { "epoch": 0.026498233744725458, "grad_norm": 2.4476451873779297, "learning_rate": 1.7535211267605638e-05, "loss": 10.5516, "step": 752 }, { "epoch": 0.02653347075768387, "grad_norm": 1.6936980485916138, "learning_rate": 1.755868544600939e-05, "loss": 4.5442, "step": 753 }, { "epoch": 0.026568707770642283, "grad_norm": 1.9132366180419922, "learning_rate": 1.7582159624413146e-05, "loss": 1.2805, "step": 754 }, { "epoch": 0.026603944783600696, "grad_norm": 1.3782182931900024, "learning_rate": 1.7605633802816902e-05, "loss": 1.0358, "step": 755 }, { "epoch": 0.026639181796559105, "grad_norm": 2.5630104541778564, "learning_rate": 1.762910798122066e-05, "loss": 7.3711, "step": 756 }, { "epoch": 0.026674418809517517, "grad_norm": 2.292625904083252, "learning_rate": 1.7652582159624414e-05, "loss": 0.9614, "step": 757 }, { "epoch": 0.02670965582247593, "grad_norm": 1.6254991292953491, "learning_rate": 1.767605633802817e-05, "loss": 4.4772, "step": 758 }, { "epoch": 0.02674489283543434, "grad_norm": 2.2888221740722656, "learning_rate": 1.7699530516431927e-05, "loss": 1.0347, "step": 759 }, { "epoch": 0.02678012984839275, "grad_norm": 1.7831162214279175, "learning_rate": 1.7723004694835683e-05, "loss": 4.4913, "step": 760 }, { "epoch": 0.026815366861351164, "grad_norm": 2.813429832458496, "learning_rate": 1.774647887323944e-05, "loss": 6.0729, "step": 761 }, { "epoch": 0.026850603874309573, "grad_norm": 1.3284785747528076, "learning_rate": 1.7769953051643195e-05, "loss": 0.9127, "step": 762 }, { "epoch": 0.026885840887267986, "grad_norm": 1.806528925895691, "learning_rate": 1.779342723004695e-05, "loss": 1.0829, "step": 763 }, { "epoch": 0.0269210779002264, "grad_norm": 2.196772336959839, "learning_rate": 1.7816901408450707e-05, "loss": 3.904, "step": 764 }, { "epoch": 0.026956314913184808, "grad_norm": 2.9176347255706787, "learning_rate": 1.7840375586854463e-05, "loss": 6.5346, "step": 765 }, { "epoch": 0.02699155192614322, "grad_norm": 1.5728182792663574, "learning_rate": 1.7863849765258215e-05, "loss": 1.2134, "step": 766 }, { "epoch": 0.027026788939101633, "grad_norm": 1.402588129043579, "learning_rate": 1.7887323943661975e-05, "loss": 1.1095, "step": 767 }, { "epoch": 0.027062025952060045, "grad_norm": 2.722843885421753, "learning_rate": 1.7910798122065727e-05, "loss": 4.4908, "step": 768 }, { "epoch": 0.027097262965018454, "grad_norm": 2.565554141998291, "learning_rate": 1.7934272300469484e-05, "loss": 3.9872, "step": 769 }, { "epoch": 0.027132499977976867, "grad_norm": 2.9077744483947754, "learning_rate": 1.7957746478873243e-05, "loss": 4.2178, "step": 770 }, { "epoch": 0.02716773699093528, "grad_norm": 2.141336679458618, "learning_rate": 1.7981220657276996e-05, "loss": 3.6678, "step": 771 }, { "epoch": 0.02720297400389369, "grad_norm": 2.3111047744750977, "learning_rate": 1.800469483568075e-05, "loss": 7.1181, "step": 772 }, { "epoch": 0.0272382110168521, "grad_norm": 3.992180824279785, "learning_rate": 1.8028169014084508e-05, "loss": 9.279, "step": 773 }, { "epoch": 0.027273448029810514, "grad_norm": 3.5903162956237793, "learning_rate": 1.8051643192488264e-05, "loss": 5.9938, "step": 774 }, { "epoch": 0.027308685042768923, "grad_norm": 1.4029488563537598, "learning_rate": 1.807511737089202e-05, "loss": 0.8897, "step": 775 }, { "epoch": 0.027343922055727336, "grad_norm": 2.583793878555298, "learning_rate": 1.8098591549295776e-05, "loss": 6.8561, "step": 776 }, { "epoch": 0.027379159068685748, "grad_norm": 2.793978452682495, "learning_rate": 1.8122065727699532e-05, "loss": 3.6993, "step": 777 }, { "epoch": 0.027414396081644157, "grad_norm": 1.8546562194824219, "learning_rate": 1.8145539906103288e-05, "loss": 1.441, "step": 778 }, { "epoch": 0.02744963309460257, "grad_norm": 2.2853095531463623, "learning_rate": 1.8169014084507044e-05, "loss": 3.6779, "step": 779 }, { "epoch": 0.027484870107560982, "grad_norm": 2.0705137252807617, "learning_rate": 1.81924882629108e-05, "loss": 4.0127, "step": 780 }, { "epoch": 0.027520107120519395, "grad_norm": 3.0150935649871826, "learning_rate": 1.8215962441314556e-05, "loss": 6.1404, "step": 781 }, { "epoch": 0.027555344133477804, "grad_norm": 1.274058222770691, "learning_rate": 1.8239436619718312e-05, "loss": 1.0839, "step": 782 }, { "epoch": 0.027590581146436217, "grad_norm": 3.8472776412963867, "learning_rate": 1.8262910798122068e-05, "loss": 4.5152, "step": 783 }, { "epoch": 0.02762581815939463, "grad_norm": 1.8047014474868774, "learning_rate": 1.8286384976525824e-05, "loss": 0.9, "step": 784 }, { "epoch": 0.02766105517235304, "grad_norm": 1.6776694059371948, "learning_rate": 1.830985915492958e-05, "loss": 3.8189, "step": 785 }, { "epoch": 0.02769629218531145, "grad_norm": 2.140362501144409, "learning_rate": 1.8333333333333333e-05, "loss": 3.8012, "step": 786 }, { "epoch": 0.027731529198269864, "grad_norm": 1.676584243774414, "learning_rate": 1.8356807511737092e-05, "loss": 0.9852, "step": 787 }, { "epoch": 0.027766766211228273, "grad_norm": 1.7701975107192993, "learning_rate": 1.8380281690140845e-05, "loss": 4.553, "step": 788 }, { "epoch": 0.027802003224186685, "grad_norm": 2.0638136863708496, "learning_rate": 1.84037558685446e-05, "loss": 0.9743, "step": 789 }, { "epoch": 0.027837240237145098, "grad_norm": 2.671175003051758, "learning_rate": 1.842723004694836e-05, "loss": 6.4844, "step": 790 }, { "epoch": 0.027872477250103507, "grad_norm": 2.1653966903686523, "learning_rate": 1.8450704225352113e-05, "loss": 4.739, "step": 791 }, { "epoch": 0.02790771426306192, "grad_norm": 2.5297186374664307, "learning_rate": 1.847417840375587e-05, "loss": 4.4423, "step": 792 }, { "epoch": 0.027942951276020332, "grad_norm": 2.3838160037994385, "learning_rate": 1.8497652582159625e-05, "loss": 3.6267, "step": 793 }, { "epoch": 0.027978188288978745, "grad_norm": 1.371861457824707, "learning_rate": 1.852112676056338e-05, "loss": 0.8655, "step": 794 }, { "epoch": 0.028013425301937154, "grad_norm": 1.7399791479110718, "learning_rate": 1.8544600938967137e-05, "loss": 3.9639, "step": 795 }, { "epoch": 0.028048662314895566, "grad_norm": 2.830822467803955, "learning_rate": 1.8568075117370893e-05, "loss": 7.55, "step": 796 }, { "epoch": 0.02808389932785398, "grad_norm": 1.116413950920105, "learning_rate": 1.859154929577465e-05, "loss": 0.9931, "step": 797 }, { "epoch": 0.028119136340812388, "grad_norm": 3.1188459396362305, "learning_rate": 1.8615023474178405e-05, "loss": 4.6834, "step": 798 }, { "epoch": 0.0281543733537708, "grad_norm": 1.9460358619689941, "learning_rate": 1.863849765258216e-05, "loss": 1.0725, "step": 799 }, { "epoch": 0.028189610366729213, "grad_norm": 2.180040121078491, "learning_rate": 1.8661971830985917e-05, "loss": 1.1438, "step": 800 }, { "epoch": 0.028224847379687622, "grad_norm": 2.549021005630493, "learning_rate": 1.8685446009389673e-05, "loss": 1.1627, "step": 801 }, { "epoch": 0.028260084392646035, "grad_norm": 2.0321848392486572, "learning_rate": 1.870892018779343e-05, "loss": 4.4713, "step": 802 }, { "epoch": 0.028295321405604448, "grad_norm": 1.6105304956436157, "learning_rate": 1.8732394366197186e-05, "loss": 0.9941, "step": 803 }, { "epoch": 0.028330558418562857, "grad_norm": 2.619731903076172, "learning_rate": 1.875586854460094e-05, "loss": 4.1019, "step": 804 }, { "epoch": 0.02836579543152127, "grad_norm": 1.59697425365448, "learning_rate": 1.8779342723004698e-05, "loss": 0.9391, "step": 805 }, { "epoch": 0.028401032444479682, "grad_norm": 1.9836206436157227, "learning_rate": 1.880281690140845e-05, "loss": 1.2031, "step": 806 }, { "epoch": 0.028436269457438094, "grad_norm": 2.5891191959381104, "learning_rate": 1.882629107981221e-05, "loss": 5.8793, "step": 807 }, { "epoch": 0.028471506470396504, "grad_norm": 4.079983711242676, "learning_rate": 1.8849765258215962e-05, "loss": 10.0953, "step": 808 }, { "epoch": 0.028506743483354916, "grad_norm": 2.4436867237091064, "learning_rate": 1.887323943661972e-05, "loss": 6.5892, "step": 809 }, { "epoch": 0.02854198049631333, "grad_norm": 2.5528790950775146, "learning_rate": 1.8896713615023478e-05, "loss": 6.0943, "step": 810 }, { "epoch": 0.028577217509271738, "grad_norm": 1.8108502626419067, "learning_rate": 1.892018779342723e-05, "loss": 3.8814, "step": 811 }, { "epoch": 0.02861245452223015, "grad_norm": 1.675460696220398, "learning_rate": 1.8943661971830987e-05, "loss": 0.9126, "step": 812 }, { "epoch": 0.028647691535188563, "grad_norm": 1.3665422201156616, "learning_rate": 1.8967136150234743e-05, "loss": 0.9718, "step": 813 }, { "epoch": 0.028682928548146972, "grad_norm": 1.324087381362915, "learning_rate": 1.89906103286385e-05, "loss": 1.0517, "step": 814 }, { "epoch": 0.028718165561105385, "grad_norm": 1.3250064849853516, "learning_rate": 1.9014084507042255e-05, "loss": 0.8543, "step": 815 }, { "epoch": 0.028753402574063797, "grad_norm": 1.583824872970581, "learning_rate": 1.903755868544601e-05, "loss": 1.0555, "step": 816 }, { "epoch": 0.028788639587022206, "grad_norm": 2.081749439239502, "learning_rate": 1.9061032863849767e-05, "loss": 1.0531, "step": 817 }, { "epoch": 0.02882387659998062, "grad_norm": 3.3092851638793945, "learning_rate": 1.9084507042253523e-05, "loss": 6.1984, "step": 818 }, { "epoch": 0.02885911361293903, "grad_norm": 1.921750545501709, "learning_rate": 1.910798122065728e-05, "loss": 0.8667, "step": 819 }, { "epoch": 0.028894350625897444, "grad_norm": 2.9170782566070557, "learning_rate": 1.9131455399061035e-05, "loss": 5.753, "step": 820 }, { "epoch": 0.028929587638855853, "grad_norm": 2.5025126934051514, "learning_rate": 1.9154929577464788e-05, "loss": 2.703, "step": 821 }, { "epoch": 0.028964824651814266, "grad_norm": 1.365907073020935, "learning_rate": 1.9178403755868547e-05, "loss": 1.1319, "step": 822 }, { "epoch": 0.02900006166477268, "grad_norm": 2.8988113403320312, "learning_rate": 1.9201877934272303e-05, "loss": 0.938, "step": 823 }, { "epoch": 0.029035298677731088, "grad_norm": 1.7835075855255127, "learning_rate": 1.922535211267606e-05, "loss": 1.0951, "step": 824 }, { "epoch": 0.0290705356906895, "grad_norm": 1.830082654953003, "learning_rate": 1.9248826291079815e-05, "loss": 1.3694, "step": 825 }, { "epoch": 0.029105772703647913, "grad_norm": 3.565176010131836, "learning_rate": 1.9272300469483568e-05, "loss": 3.0503, "step": 826 }, { "epoch": 0.029141009716606322, "grad_norm": 1.9963998794555664, "learning_rate": 1.9295774647887327e-05, "loss": 0.9194, "step": 827 }, { "epoch": 0.029176246729564734, "grad_norm": 2.33343505859375, "learning_rate": 1.9319248826291083e-05, "loss": 1.2025, "step": 828 }, { "epoch": 0.029211483742523147, "grad_norm": 3.058318614959717, "learning_rate": 1.9342723004694836e-05, "loss": 5.9087, "step": 829 }, { "epoch": 0.029246720755481556, "grad_norm": 2.105163335800171, "learning_rate": 1.9366197183098595e-05, "loss": 0.9903, "step": 830 }, { "epoch": 0.02928195776843997, "grad_norm": 4.026034832000732, "learning_rate": 1.9389671361502348e-05, "loss": 4.2531, "step": 831 }, { "epoch": 0.02931719478139838, "grad_norm": 1.1567457914352417, "learning_rate": 1.9413145539906104e-05, "loss": 1.0274, "step": 832 }, { "epoch": 0.029352431794356794, "grad_norm": 1.756487250328064, "learning_rate": 1.943661971830986e-05, "loss": 1.0106, "step": 833 }, { "epoch": 0.029387668807315203, "grad_norm": 3.537179946899414, "learning_rate": 1.9460093896713616e-05, "loss": 6.8469, "step": 834 }, { "epoch": 0.029422905820273616, "grad_norm": 1.9641987085342407, "learning_rate": 1.9483568075117372e-05, "loss": 3.8354, "step": 835 }, { "epoch": 0.029458142833232028, "grad_norm": 2.1946089267730713, "learning_rate": 1.9507042253521128e-05, "loss": 3.4078, "step": 836 }, { "epoch": 0.029493379846190437, "grad_norm": 2.1253085136413574, "learning_rate": 1.9530516431924884e-05, "loss": 0.8657, "step": 837 }, { "epoch": 0.02952861685914885, "grad_norm": 3.1257777214050293, "learning_rate": 1.955399061032864e-05, "loss": 3.2954, "step": 838 }, { "epoch": 0.029563853872107262, "grad_norm": 1.921136498451233, "learning_rate": 1.9577464788732396e-05, "loss": 0.7859, "step": 839 }, { "epoch": 0.02959909088506567, "grad_norm": 1.6441457271575928, "learning_rate": 1.9600938967136152e-05, "loss": 0.9955, "step": 840 }, { "epoch": 0.029634327898024084, "grad_norm": 3.319396495819092, "learning_rate": 1.962441314553991e-05, "loss": 4.6372, "step": 841 }, { "epoch": 0.029669564910982497, "grad_norm": 2.8388612270355225, "learning_rate": 1.9647887323943664e-05, "loss": 6.6407, "step": 842 }, { "epoch": 0.029704801923940906, "grad_norm": 2.2118606567382812, "learning_rate": 1.967136150234742e-05, "loss": 4.0148, "step": 843 }, { "epoch": 0.02974003893689932, "grad_norm": 2.4686882495880127, "learning_rate": 1.9694835680751173e-05, "loss": 1.0669, "step": 844 }, { "epoch": 0.02977527594985773, "grad_norm": 1.5062663555145264, "learning_rate": 1.9718309859154933e-05, "loss": 1.0201, "step": 845 }, { "epoch": 0.029810512962816144, "grad_norm": 1.3561657667160034, "learning_rate": 1.9741784037558685e-05, "loss": 0.9792, "step": 846 }, { "epoch": 0.029845749975774553, "grad_norm": 1.8293639421463013, "learning_rate": 1.9765258215962445e-05, "loss": 1.0337, "step": 847 }, { "epoch": 0.029880986988732965, "grad_norm": 1.5938866138458252, "learning_rate": 1.97887323943662e-05, "loss": 1.1248, "step": 848 }, { "epoch": 0.029916224001691378, "grad_norm": 2.0640339851379395, "learning_rate": 1.9812206572769953e-05, "loss": 1.1748, "step": 849 }, { "epoch": 0.029951461014649787, "grad_norm": 2.0789265632629395, "learning_rate": 1.9835680751173713e-05, "loss": 0.9779, "step": 850 }, { "epoch": 0.0299866980276082, "grad_norm": 2.958968162536621, "learning_rate": 1.9859154929577465e-05, "loss": 4.1487, "step": 851 }, { "epoch": 0.030021935040566612, "grad_norm": 2.581120014190674, "learning_rate": 1.988262910798122e-05, "loss": 4.2497, "step": 852 }, { "epoch": 0.03005717205352502, "grad_norm": 3.0803987979888916, "learning_rate": 1.9906103286384977e-05, "loss": 4.5226, "step": 853 }, { "epoch": 0.030092409066483434, "grad_norm": 2.4773945808410645, "learning_rate": 1.9929577464788734e-05, "loss": 3.9551, "step": 854 }, { "epoch": 0.030127646079441846, "grad_norm": 1.7466543912887573, "learning_rate": 1.995305164319249e-05, "loss": 0.8404, "step": 855 }, { "epoch": 0.030162883092400256, "grad_norm": 2.4057037830352783, "learning_rate": 1.9976525821596246e-05, "loss": 1.3947, "step": 856 }, { "epoch": 0.030198120105358668, "grad_norm": 1.8548486232757568, "learning_rate": 2e-05, "loss": 1.2291, "step": 857 }, { "epoch": 0.03023335711831708, "grad_norm": 1.7482457160949707, "learning_rate": 1.9999999934874362e-05, "loss": 1.2486, "step": 858 }, { "epoch": 0.030268594131275493, "grad_norm": 3.498670816421509, "learning_rate": 1.9999999739497455e-05, "loss": 7.1239, "step": 859 }, { "epoch": 0.030303831144233902, "grad_norm": 2.2986183166503906, "learning_rate": 1.9999999413869275e-05, "loss": 3.6953, "step": 860 }, { "epoch": 0.030339068157192315, "grad_norm": 1.7882410287857056, "learning_rate": 1.9999998957989828e-05, "loss": 3.6816, "step": 861 }, { "epoch": 0.030374305170150728, "grad_norm": 1.586327314376831, "learning_rate": 1.9999998371859122e-05, "loss": 1.1004, "step": 862 }, { "epoch": 0.030409542183109137, "grad_norm": 3.051652193069458, "learning_rate": 1.9999997655477165e-05, "loss": 5.9848, "step": 863 }, { "epoch": 0.03044477919606755, "grad_norm": 2.4199113845825195, "learning_rate": 1.999999680884396e-05, "loss": 6.652, "step": 864 }, { "epoch": 0.030480016209025962, "grad_norm": 1.7570139169692993, "learning_rate": 1.9999995831959523e-05, "loss": 0.9219, "step": 865 }, { "epoch": 0.03051525322198437, "grad_norm": 1.6550226211547852, "learning_rate": 1.999999472482387e-05, "loss": 1.2134, "step": 866 }, { "epoch": 0.030550490234942784, "grad_norm": 1.4887415170669556, "learning_rate": 1.999999348743701e-05, "loss": 1.1286, "step": 867 }, { "epoch": 0.030585727247901196, "grad_norm": 1.4720362424850464, "learning_rate": 1.9999992119798963e-05, "loss": 1.0835, "step": 868 }, { "epoch": 0.030620964260859605, "grad_norm": 4.015078067779541, "learning_rate": 1.9999990621909744e-05, "loss": 4.3026, "step": 869 }, { "epoch": 0.030656201273818018, "grad_norm": 1.6393297910690308, "learning_rate": 1.9999988993769373e-05, "loss": 1.0713, "step": 870 }, { "epoch": 0.03069143828677643, "grad_norm": 2.0057051181793213, "learning_rate": 1.9999987235377872e-05, "loss": 0.9768, "step": 871 }, { "epoch": 0.030726675299734843, "grad_norm": 3.290712594985962, "learning_rate": 1.9999985346735262e-05, "loss": 7.3554, "step": 872 }, { "epoch": 0.030761912312693252, "grad_norm": 1.9060224294662476, "learning_rate": 1.999998332784157e-05, "loss": 3.1918, "step": 873 }, { "epoch": 0.030797149325651665, "grad_norm": 2.626798152923584, "learning_rate": 1.9999981178696823e-05, "loss": 1.1966, "step": 874 }, { "epoch": 0.030832386338610077, "grad_norm": 2.8491132259368896, "learning_rate": 1.9999978899301045e-05, "loss": 3.7251, "step": 875 }, { "epoch": 0.030867623351568486, "grad_norm": 2.6924796104431152, "learning_rate": 1.999997648965427e-05, "loss": 6.8785, "step": 876 }, { "epoch": 0.0309028603645269, "grad_norm": 2.6299803256988525, "learning_rate": 1.9999973949756523e-05, "loss": 6.7857, "step": 877 }, { "epoch": 0.03093809737748531, "grad_norm": 1.3890724182128906, "learning_rate": 1.9999971279607845e-05, "loss": 0.7973, "step": 878 }, { "epoch": 0.03097333439044372, "grad_norm": 3.2629828453063965, "learning_rate": 1.999996847920827e-05, "loss": 3.8742, "step": 879 }, { "epoch": 0.031008571403402133, "grad_norm": 4.175227165222168, "learning_rate": 1.9999965548557828e-05, "loss": 5.8804, "step": 880 }, { "epoch": 0.031043808416360546, "grad_norm": 2.046205759048462, "learning_rate": 1.999996248765656e-05, "loss": 1.0389, "step": 881 }, { "epoch": 0.031079045429318955, "grad_norm": 1.5291146039962769, "learning_rate": 1.9999959296504506e-05, "loss": 1.0222, "step": 882 }, { "epoch": 0.031114282442277368, "grad_norm": 3.856996536254883, "learning_rate": 1.9999955975101715e-05, "loss": 7.2697, "step": 883 }, { "epoch": 0.03114951945523578, "grad_norm": 2.3248322010040283, "learning_rate": 1.9999952523448218e-05, "loss": 4.3035, "step": 884 }, { "epoch": 0.031184756468194193, "grad_norm": 1.5875558853149414, "learning_rate": 1.9999948941544066e-05, "loss": 1.0033, "step": 885 }, { "epoch": 0.031219993481152602, "grad_norm": 2.5059945583343506, "learning_rate": 1.9999945229389307e-05, "loss": 4.5056, "step": 886 }, { "epoch": 0.03125523049411101, "grad_norm": 2.389430284500122, "learning_rate": 1.9999941386983987e-05, "loss": 3.3531, "step": 887 }, { "epoch": 0.031290467507069424, "grad_norm": 1.7298946380615234, "learning_rate": 1.9999937414328158e-05, "loss": 1.1469, "step": 888 }, { "epoch": 0.031325704520027836, "grad_norm": 1.6106432676315308, "learning_rate": 1.999993331142187e-05, "loss": 1.3937, "step": 889 }, { "epoch": 0.03136094153298625, "grad_norm": 1.6394884586334229, "learning_rate": 1.9999929078265174e-05, "loss": 1.2471, "step": 890 }, { "epoch": 0.03139617854594466, "grad_norm": 1.4542113542556763, "learning_rate": 1.999992471485813e-05, "loss": 0.7681, "step": 891 }, { "epoch": 0.031431415558903074, "grad_norm": 1.818955659866333, "learning_rate": 1.9999920221200794e-05, "loss": 0.9942, "step": 892 }, { "epoch": 0.031466652571861486, "grad_norm": 2.618698835372925, "learning_rate": 1.999991559729322e-05, "loss": 3.2094, "step": 893 }, { "epoch": 0.03150188958481989, "grad_norm": 1.5298415422439575, "learning_rate": 1.9999910843135476e-05, "loss": 0.8725, "step": 894 }, { "epoch": 0.031537126597778305, "grad_norm": 3.8483638763427734, "learning_rate": 1.9999905958727616e-05, "loss": 4.5469, "step": 895 }, { "epoch": 0.03157236361073672, "grad_norm": 1.781726598739624, "learning_rate": 1.9999900944069707e-05, "loss": 0.9229, "step": 896 }, { "epoch": 0.03160760062369513, "grad_norm": 2.1177420616149902, "learning_rate": 1.9999895799161814e-05, "loss": 1.1649, "step": 897 }, { "epoch": 0.03164283763665354, "grad_norm": 1.5263479948043823, "learning_rate": 1.9999890524004008e-05, "loss": 1.1795, "step": 898 }, { "epoch": 0.031678074649611955, "grad_norm": 2.120789051055908, "learning_rate": 1.999988511859635e-05, "loss": 3.2496, "step": 899 }, { "epoch": 0.03171331166257036, "grad_norm": 2.0117170810699463, "learning_rate": 1.9999879582938917e-05, "loss": 3.4837, "step": 900 }, { "epoch": 0.03174854867552877, "grad_norm": 1.4383649826049805, "learning_rate": 1.999987391703178e-05, "loss": 1.1377, "step": 901 }, { "epoch": 0.031783785688487186, "grad_norm": 1.6327314376831055, "learning_rate": 1.9999868120875007e-05, "loss": 1.1011, "step": 902 }, { "epoch": 0.0318190227014456, "grad_norm": 2.0238771438598633, "learning_rate": 1.9999862194468682e-05, "loss": 3.0257, "step": 903 }, { "epoch": 0.03185425971440401, "grad_norm": 2.1140973567962646, "learning_rate": 1.9999856137812875e-05, "loss": 7.2755, "step": 904 }, { "epoch": 0.031889496727362424, "grad_norm": 1.7641692161560059, "learning_rate": 1.9999849950907673e-05, "loss": 3.6365, "step": 905 }, { "epoch": 0.031924733740320836, "grad_norm": 2.101785898208618, "learning_rate": 1.9999843633753147e-05, "loss": 3.8979, "step": 906 }, { "epoch": 0.03195997075327924, "grad_norm": 1.5336636304855347, "learning_rate": 1.9999837186349385e-05, "loss": 0.9173, "step": 907 }, { "epoch": 0.031995207766237654, "grad_norm": 2.135427236557007, "learning_rate": 1.999983060869647e-05, "loss": 5.8514, "step": 908 }, { "epoch": 0.03203044477919607, "grad_norm": 2.5888657569885254, "learning_rate": 1.999982390079449e-05, "loss": 4.6881, "step": 909 }, { "epoch": 0.03206568179215448, "grad_norm": 1.69389009475708, "learning_rate": 1.9999817062643525e-05, "loss": 0.8732, "step": 910 }, { "epoch": 0.03210091880511289, "grad_norm": 2.0528404712677, "learning_rate": 1.9999810094243673e-05, "loss": 0.9745, "step": 911 }, { "epoch": 0.032136155818071305, "grad_norm": 2.8447532653808594, "learning_rate": 1.9999802995595018e-05, "loss": 4.6242, "step": 912 }, { "epoch": 0.03217139283102971, "grad_norm": 1.309259057044983, "learning_rate": 1.9999795766697657e-05, "loss": 0.8327, "step": 913 }, { "epoch": 0.03220662984398812, "grad_norm": 2.441037654876709, "learning_rate": 1.9999788407551683e-05, "loss": 3.3121, "step": 914 }, { "epoch": 0.032241866856946536, "grad_norm": 1.9099104404449463, "learning_rate": 1.999978091815719e-05, "loss": 4.2441, "step": 915 }, { "epoch": 0.03227710386990495, "grad_norm": 2.4034547805786133, "learning_rate": 1.9999773298514276e-05, "loss": 4.0901, "step": 916 }, { "epoch": 0.03231234088286336, "grad_norm": 1.2998943328857422, "learning_rate": 1.9999765548623042e-05, "loss": 1.0601, "step": 917 }, { "epoch": 0.03234757789582177, "grad_norm": 2.4008822441101074, "learning_rate": 1.999975766848359e-05, "loss": 1.1536, "step": 918 }, { "epoch": 0.032382814908780186, "grad_norm": 2.6598052978515625, "learning_rate": 1.9999749658096016e-05, "loss": 4.081, "step": 919 }, { "epoch": 0.03241805192173859, "grad_norm": 2.3504388332366943, "learning_rate": 1.9999741517460433e-05, "loss": 7.4881, "step": 920 }, { "epoch": 0.032453288934697004, "grad_norm": 2.7225093841552734, "learning_rate": 1.999973324657694e-05, "loss": 1.0381, "step": 921 }, { "epoch": 0.03248852594765542, "grad_norm": 1.1838845014572144, "learning_rate": 1.9999724845445648e-05, "loss": 1.2239, "step": 922 }, { "epoch": 0.03252376296061383, "grad_norm": 1.090419054031372, "learning_rate": 1.999971631406667e-05, "loss": 0.9384, "step": 923 }, { "epoch": 0.03255899997357224, "grad_norm": 2.0824506282806396, "learning_rate": 1.999970765244011e-05, "loss": 3.7195, "step": 924 }, { "epoch": 0.032594236986530654, "grad_norm": 2.835817337036133, "learning_rate": 1.9999698860566083e-05, "loss": 8.5089, "step": 925 }, { "epoch": 0.03262947399948906, "grad_norm": 2.0730414390563965, "learning_rate": 1.999968993844471e-05, "loss": 3.3477, "step": 926 }, { "epoch": 0.03266471101244747, "grad_norm": 1.83244788646698, "learning_rate": 1.9999680886076094e-05, "loss": 3.483, "step": 927 }, { "epoch": 0.032699948025405885, "grad_norm": 1.501037836074829, "learning_rate": 1.9999671703460366e-05, "loss": 1.1121, "step": 928 }, { "epoch": 0.0327351850383643, "grad_norm": 1.5762557983398438, "learning_rate": 1.999966239059764e-05, "loss": 1.1832, "step": 929 }, { "epoch": 0.03277042205132271, "grad_norm": 2.06174635887146, "learning_rate": 1.9999652947488036e-05, "loss": 0.8256, "step": 930 }, { "epoch": 0.03280565906428112, "grad_norm": 2.8184921741485596, "learning_rate": 1.999964337413168e-05, "loss": 8.2278, "step": 931 }, { "epoch": 0.032840896077239536, "grad_norm": 2.553494691848755, "learning_rate": 1.9999633670528694e-05, "loss": 3.9299, "step": 932 }, { "epoch": 0.03287613309019794, "grad_norm": 2.7581210136413574, "learning_rate": 1.9999623836679206e-05, "loss": 3.4509, "step": 933 }, { "epoch": 0.032911370103156354, "grad_norm": 1.483190655708313, "learning_rate": 1.9999613872583343e-05, "loss": 1.0802, "step": 934 }, { "epoch": 0.032946607116114766, "grad_norm": 1.7279118299484253, "learning_rate": 1.9999603778241237e-05, "loss": 0.9777, "step": 935 }, { "epoch": 0.03298184412907318, "grad_norm": 2.148808717727661, "learning_rate": 1.9999593553653018e-05, "loss": 3.6164, "step": 936 }, { "epoch": 0.03301708114203159, "grad_norm": 2.4387221336364746, "learning_rate": 1.999958319881882e-05, "loss": 3.7548, "step": 937 }, { "epoch": 0.033052318154990004, "grad_norm": 2.487114667892456, "learning_rate": 1.9999572713738775e-05, "loss": 3.966, "step": 938 }, { "epoch": 0.03308755516794841, "grad_norm": 1.2655823230743408, "learning_rate": 1.999956209841302e-05, "loss": 0.9259, "step": 939 }, { "epoch": 0.03312279218090682, "grad_norm": 1.5664846897125244, "learning_rate": 1.9999551352841698e-05, "loss": 1.1269, "step": 940 }, { "epoch": 0.033158029193865235, "grad_norm": 1.5491636991500854, "learning_rate": 1.9999540477024944e-05, "loss": 0.9628, "step": 941 }, { "epoch": 0.03319326620682365, "grad_norm": 1.9494102001190186, "learning_rate": 1.9999529470962905e-05, "loss": 0.8053, "step": 942 }, { "epoch": 0.03322850321978206, "grad_norm": 1.502126932144165, "learning_rate": 1.9999518334655717e-05, "loss": 1.1076, "step": 943 }, { "epoch": 0.03326374023274047, "grad_norm": 1.5071102380752563, "learning_rate": 1.9999507068103532e-05, "loss": 1.1745, "step": 944 }, { "epoch": 0.033298977245698885, "grad_norm": 2.485849380493164, "learning_rate": 1.9999495671306493e-05, "loss": 1.2235, "step": 945 }, { "epoch": 0.03333421425865729, "grad_norm": 1.8726589679718018, "learning_rate": 1.9999484144264748e-05, "loss": 1.1626, "step": 946 }, { "epoch": 0.033369451271615704, "grad_norm": 2.097385883331299, "learning_rate": 1.999947248697845e-05, "loss": 3.7705, "step": 947 }, { "epoch": 0.033404688284574116, "grad_norm": 1.681307315826416, "learning_rate": 1.9999460699447747e-05, "loss": 0.8172, "step": 948 }, { "epoch": 0.03343992529753253, "grad_norm": 1.6959290504455566, "learning_rate": 1.9999448781672797e-05, "loss": 1.0797, "step": 949 }, { "epoch": 0.03347516231049094, "grad_norm": 2.7510266304016113, "learning_rate": 1.999943673365375e-05, "loss": 3.7792, "step": 950 }, { "epoch": 0.033510399323449354, "grad_norm": 1.100172519683838, "learning_rate": 1.9999424555390768e-05, "loss": 0.9796, "step": 951 }, { "epoch": 0.03354563633640776, "grad_norm": 2.9016306400299072, "learning_rate": 1.9999412246884005e-05, "loss": 4.1511, "step": 952 }, { "epoch": 0.03358087334936617, "grad_norm": 3.335414409637451, "learning_rate": 1.9999399808133628e-05, "loss": 3.4626, "step": 953 }, { "epoch": 0.033616110362324585, "grad_norm": 3.5435757637023926, "learning_rate": 1.999938723913979e-05, "loss": 4.2786, "step": 954 }, { "epoch": 0.033651347375283, "grad_norm": 1.7328895330429077, "learning_rate": 1.9999374539902666e-05, "loss": 1.221, "step": 955 }, { "epoch": 0.03368658438824141, "grad_norm": 2.2232422828674316, "learning_rate": 1.999936171042241e-05, "loss": 1.1515, "step": 956 }, { "epoch": 0.03372182140119982, "grad_norm": 2.916959762573242, "learning_rate": 1.9999348750699194e-05, "loss": 6.5377, "step": 957 }, { "epoch": 0.033757058414158235, "grad_norm": 1.8350626230239868, "learning_rate": 1.9999335660733186e-05, "loss": 0.9838, "step": 958 }, { "epoch": 0.03379229542711664, "grad_norm": 2.1062824726104736, "learning_rate": 1.999932244052456e-05, "loss": 1.0295, "step": 959 }, { "epoch": 0.03382753244007505, "grad_norm": 1.8696986436843872, "learning_rate": 1.9999309090073485e-05, "loss": 1.0571, "step": 960 }, { "epoch": 0.033862769453033466, "grad_norm": 1.3422448635101318, "learning_rate": 1.9999295609380134e-05, "loss": 1.1521, "step": 961 }, { "epoch": 0.03389800646599188, "grad_norm": 1.5237586498260498, "learning_rate": 1.9999281998444682e-05, "loss": 0.9584, "step": 962 }, { "epoch": 0.03393324347895029, "grad_norm": 1.585180640220642, "learning_rate": 1.9999268257267314e-05, "loss": 1.0808, "step": 963 }, { "epoch": 0.033968480491908704, "grad_norm": 3.2225191593170166, "learning_rate": 1.9999254385848198e-05, "loss": 4.1872, "step": 964 }, { "epoch": 0.03400371750486711, "grad_norm": 1.750192403793335, "learning_rate": 1.9999240384187524e-05, "loss": 0.8186, "step": 965 }, { "epoch": 0.03403895451782552, "grad_norm": 3.3869316577911377, "learning_rate": 1.9999226252285468e-05, "loss": 6.4928, "step": 966 }, { "epoch": 0.034074191530783934, "grad_norm": 2.8866257667541504, "learning_rate": 1.9999211990142216e-05, "loss": 5.9885, "step": 967 }, { "epoch": 0.03410942854374235, "grad_norm": 1.988308072090149, "learning_rate": 1.9999197597757956e-05, "loss": 1.2411, "step": 968 }, { "epoch": 0.03414466555670076, "grad_norm": 1.4046778678894043, "learning_rate": 1.9999183075132873e-05, "loss": 1.1154, "step": 969 }, { "epoch": 0.03417990256965917, "grad_norm": 2.3629703521728516, "learning_rate": 1.9999168422267157e-05, "loss": 4.1976, "step": 970 }, { "epoch": 0.034215139582617585, "grad_norm": 1.951958417892456, "learning_rate": 1.9999153639160997e-05, "loss": 3.7961, "step": 971 }, { "epoch": 0.03425037659557599, "grad_norm": 2.540215253829956, "learning_rate": 1.9999138725814588e-05, "loss": 6.3019, "step": 972 }, { "epoch": 0.0342856136085344, "grad_norm": 2.214125394821167, "learning_rate": 1.9999123682228124e-05, "loss": 3.9061, "step": 973 }, { "epoch": 0.034320850621492816, "grad_norm": 2.158970832824707, "learning_rate": 1.99991085084018e-05, "loss": 3.9714, "step": 974 }, { "epoch": 0.03435608763445123, "grad_norm": 1.2468169927597046, "learning_rate": 1.9999093204335813e-05, "loss": 0.8754, "step": 975 }, { "epoch": 0.03439132464740964, "grad_norm": 2.4524219036102295, "learning_rate": 1.9999077770030365e-05, "loss": 7.1227, "step": 976 }, { "epoch": 0.03442656166036805, "grad_norm": 1.2298191785812378, "learning_rate": 1.9999062205485655e-05, "loss": 0.8561, "step": 977 }, { "epoch": 0.03446179867332646, "grad_norm": 1.9567428827285767, "learning_rate": 1.9999046510701885e-05, "loss": 4.0679, "step": 978 }, { "epoch": 0.03449703568628487, "grad_norm": 1.4279149770736694, "learning_rate": 1.999903068567926e-05, "loss": 3.4777, "step": 979 }, { "epoch": 0.034532272699243284, "grad_norm": 2.1894984245300293, "learning_rate": 1.999901473041799e-05, "loss": 3.7413, "step": 980 }, { "epoch": 0.0345675097122017, "grad_norm": 1.7080358266830444, "learning_rate": 1.9998998644918276e-05, "loss": 4.551, "step": 981 }, { "epoch": 0.03460274672516011, "grad_norm": 1.5903117656707764, "learning_rate": 1.9998982429180332e-05, "loss": 1.1508, "step": 982 }, { "epoch": 0.03463798373811852, "grad_norm": 2.052504539489746, "learning_rate": 1.9998966083204368e-05, "loss": 3.8034, "step": 983 }, { "epoch": 0.034673220751076934, "grad_norm": 1.5415409803390503, "learning_rate": 1.99989496069906e-05, "loss": 1.0207, "step": 984 }, { "epoch": 0.03470845776403534, "grad_norm": 1.5795629024505615, "learning_rate": 1.9998933000539234e-05, "loss": 1.1423, "step": 985 }, { "epoch": 0.03474369477699375, "grad_norm": 2.2497706413269043, "learning_rate": 1.9998916263850495e-05, "loss": 4.3012, "step": 986 }, { "epoch": 0.034778931789952165, "grad_norm": 2.2751541137695312, "learning_rate": 1.99988993969246e-05, "loss": 4.558, "step": 987 }, { "epoch": 0.03481416880291058, "grad_norm": 1.365843415260315, "learning_rate": 1.9998882399761767e-05, "loss": 0.9847, "step": 988 }, { "epoch": 0.03484940581586899, "grad_norm": 1.6101640462875366, "learning_rate": 1.9998865272362214e-05, "loss": 0.9719, "step": 989 }, { "epoch": 0.0348846428288274, "grad_norm": 2.7308056354522705, "learning_rate": 1.9998848014726166e-05, "loss": 0.7516, "step": 990 }, { "epoch": 0.03491987984178581, "grad_norm": 1.965455412864685, "learning_rate": 1.9998830626853852e-05, "loss": 3.7589, "step": 991 }, { "epoch": 0.03495511685474422, "grad_norm": 2.2347700595855713, "learning_rate": 1.9998813108745496e-05, "loss": 4.4022, "step": 992 }, { "epoch": 0.034990353867702634, "grad_norm": 2.5548019409179688, "learning_rate": 1.9998795460401325e-05, "loss": 4.1228, "step": 993 }, { "epoch": 0.035025590880661046, "grad_norm": 2.1798768043518066, "learning_rate": 1.999877768182157e-05, "loss": 3.4108, "step": 994 }, { "epoch": 0.03506082789361946, "grad_norm": 2.258432388305664, "learning_rate": 1.999875977300646e-05, "loss": 3.1753, "step": 995 }, { "epoch": 0.03509606490657787, "grad_norm": 1.3630884885787964, "learning_rate": 1.9998741733956232e-05, "loss": 0.8858, "step": 996 }, { "epoch": 0.035131301919536284, "grad_norm": 1.5914462804794312, "learning_rate": 1.9998723564671114e-05, "loss": 1.1298, "step": 997 }, { "epoch": 0.03516653893249469, "grad_norm": 1.9918090105056763, "learning_rate": 1.999870526515135e-05, "loss": 3.935, "step": 998 }, { "epoch": 0.0352017759454531, "grad_norm": 2.878378391265869, "learning_rate": 1.999868683539718e-05, "loss": 5.7368, "step": 999 }, { "epoch": 0.035237012958411515, "grad_norm": 1.750314474105835, "learning_rate": 1.999866827540884e-05, "loss": 1.1646, "step": 1000 }, { "epoch": 0.03527224997136993, "grad_norm": 2.9225471019744873, "learning_rate": 1.9998649585186572e-05, "loss": 4.6457, "step": 1001 }, { "epoch": 0.03530748698432834, "grad_norm": 2.0518691539764404, "learning_rate": 1.999863076473062e-05, "loss": 3.9376, "step": 1002 }, { "epoch": 0.03534272399728675, "grad_norm": 1.3176794052124023, "learning_rate": 1.9998611814041226e-05, "loss": 1.0282, "step": 1003 }, { "epoch": 0.03537796101024516, "grad_norm": 3.070824146270752, "learning_rate": 1.999859273311864e-05, "loss": 3.5551, "step": 1004 }, { "epoch": 0.03541319802320357, "grad_norm": 2.6467068195343018, "learning_rate": 1.9998573521963113e-05, "loss": 3.9606, "step": 1005 }, { "epoch": 0.035448435036161983, "grad_norm": 2.221841335296631, "learning_rate": 1.999855418057489e-05, "loss": 3.6022, "step": 1006 }, { "epoch": 0.035483672049120396, "grad_norm": 3.468632459640503, "learning_rate": 1.9998534708954222e-05, "loss": 3.9181, "step": 1007 }, { "epoch": 0.03551890906207881, "grad_norm": 1.4670137166976929, "learning_rate": 1.999851510710137e-05, "loss": 0.8927, "step": 1008 }, { "epoch": 0.03555414607503722, "grad_norm": 2.1172432899475098, "learning_rate": 1.9998495375016585e-05, "loss": 1.0069, "step": 1009 }, { "epoch": 0.035589383087995634, "grad_norm": 1.6409891843795776, "learning_rate": 1.9998475512700125e-05, "loss": 0.9613, "step": 1010 }, { "epoch": 0.03562462010095404, "grad_norm": 1.308687448501587, "learning_rate": 1.9998455520152245e-05, "loss": 0.8185, "step": 1011 }, { "epoch": 0.03565985711391245, "grad_norm": 1.354082703590393, "learning_rate": 1.999843539737321e-05, "loss": 1.1024, "step": 1012 }, { "epoch": 0.035695094126870865, "grad_norm": 1.581165075302124, "learning_rate": 1.999841514436328e-05, "loss": 1.0153, "step": 1013 }, { "epoch": 0.03573033113982928, "grad_norm": 1.2421311140060425, "learning_rate": 1.9998394761122717e-05, "loss": 0.9913, "step": 1014 }, { "epoch": 0.03576556815278769, "grad_norm": 1.9978996515274048, "learning_rate": 1.999837424765179e-05, "loss": 0.8292, "step": 1015 }, { "epoch": 0.0358008051657461, "grad_norm": 3.748291492462158, "learning_rate": 1.9998353603950768e-05, "loss": 4.6621, "step": 1016 }, { "epoch": 0.03583604217870451, "grad_norm": 1.7242645025253296, "learning_rate": 1.999833283001991e-05, "loss": 1.0836, "step": 1017 }, { "epoch": 0.03587127919166292, "grad_norm": 2.3425583839416504, "learning_rate": 1.9998311925859498e-05, "loss": 3.8686, "step": 1018 }, { "epoch": 0.03590651620462133, "grad_norm": 2.7483675479888916, "learning_rate": 1.99982908914698e-05, "loss": 0.9856, "step": 1019 }, { "epoch": 0.035941753217579746, "grad_norm": 1.9093797206878662, "learning_rate": 1.9998269726851088e-05, "loss": 1.5835, "step": 1020 }, { "epoch": 0.03597699023053816, "grad_norm": 1.8515244722366333, "learning_rate": 1.999824843200364e-05, "loss": 1.2383, "step": 1021 }, { "epoch": 0.03601222724349657, "grad_norm": 3.002869129180908, "learning_rate": 1.999822700692773e-05, "loss": 7.2973, "step": 1022 }, { "epoch": 0.036047464256454984, "grad_norm": 1.9526457786560059, "learning_rate": 1.9998205451623644e-05, "loss": 3.6263, "step": 1023 }, { "epoch": 0.03608270126941339, "grad_norm": 3.639843463897705, "learning_rate": 1.9998183766091656e-05, "loss": 8.069, "step": 1024 }, { "epoch": 0.0361179382823718, "grad_norm": 1.6355819702148438, "learning_rate": 1.9998161950332052e-05, "loss": 0.8918, "step": 1025 }, { "epoch": 0.036153175295330214, "grad_norm": 3.3407888412475586, "learning_rate": 1.9998140004345113e-05, "loss": 3.9108, "step": 1026 }, { "epoch": 0.03618841230828863, "grad_norm": 1.832800030708313, "learning_rate": 1.9998117928131128e-05, "loss": 1.1899, "step": 1027 }, { "epoch": 0.03622364932124704, "grad_norm": 2.658440351486206, "learning_rate": 1.9998095721690384e-05, "loss": 1.2192, "step": 1028 }, { "epoch": 0.03625888633420545, "grad_norm": 3.4782590866088867, "learning_rate": 1.999807338502317e-05, "loss": 4.0573, "step": 1029 }, { "epoch": 0.03629412334716386, "grad_norm": 1.3963239192962646, "learning_rate": 1.9998050918129777e-05, "loss": 1.0535, "step": 1030 }, { "epoch": 0.03632936036012227, "grad_norm": 1.4578403234481812, "learning_rate": 1.9998028321010495e-05, "loss": 1.0784, "step": 1031 }, { "epoch": 0.03636459737308068, "grad_norm": 5.707817554473877, "learning_rate": 1.999800559366562e-05, "loss": 5.9968, "step": 1032 }, { "epoch": 0.036399834386039095, "grad_norm": 2.350647211074829, "learning_rate": 1.9997982736095455e-05, "loss": 1.256, "step": 1033 }, { "epoch": 0.03643507139899751, "grad_norm": 1.5156463384628296, "learning_rate": 1.9997959748300285e-05, "loss": 1.2715, "step": 1034 }, { "epoch": 0.03647030841195592, "grad_norm": 1.2927601337432861, "learning_rate": 1.999793663028042e-05, "loss": 1.1314, "step": 1035 }, { "epoch": 0.03650554542491433, "grad_norm": 2.6499392986297607, "learning_rate": 1.999791338203615e-05, "loss": 3.7136, "step": 1036 }, { "epoch": 0.03654078243787274, "grad_norm": 3.1019272804260254, "learning_rate": 1.999789000356779e-05, "loss": 6.8214, "step": 1037 }, { "epoch": 0.03657601945083115, "grad_norm": 2.926253318786621, "learning_rate": 1.9997866494875636e-05, "loss": 3.9051, "step": 1038 }, { "epoch": 0.036611256463789564, "grad_norm": 1.0282522439956665, "learning_rate": 1.9997842855960002e-05, "loss": 0.7342, "step": 1039 }, { "epoch": 0.03664649347674798, "grad_norm": 5.365920543670654, "learning_rate": 1.9997819086821188e-05, "loss": 9.9492, "step": 1040 }, { "epoch": 0.03668173048970639, "grad_norm": 1.2935872077941895, "learning_rate": 1.999779518745951e-05, "loss": 1.0323, "step": 1041 }, { "epoch": 0.0367169675026648, "grad_norm": 1.614148497581482, "learning_rate": 1.9997771157875274e-05, "loss": 1.0356, "step": 1042 }, { "epoch": 0.03675220451562321, "grad_norm": 2.640254259109497, "learning_rate": 1.999774699806879e-05, "loss": 3.7403, "step": 1043 }, { "epoch": 0.03678744152858162, "grad_norm": 3.054328203201294, "learning_rate": 1.9997722708040387e-05, "loss": 4.008, "step": 1044 }, { "epoch": 0.03682267854154003, "grad_norm": 1.1796488761901855, "learning_rate": 1.999769828779037e-05, "loss": 1.2213, "step": 1045 }, { "epoch": 0.036857915554498445, "grad_norm": 2.7531702518463135, "learning_rate": 1.9997673737319056e-05, "loss": 3.9498, "step": 1046 }, { "epoch": 0.03689315256745686, "grad_norm": 2.9587621688842773, "learning_rate": 1.999764905662677e-05, "loss": 3.7152, "step": 1047 }, { "epoch": 0.03692838958041527, "grad_norm": 3.87255859375, "learning_rate": 1.9997624245713828e-05, "loss": 6.6201, "step": 1048 }, { "epoch": 0.03696362659337368, "grad_norm": 1.468935489654541, "learning_rate": 1.9997599304580563e-05, "loss": 1.0393, "step": 1049 }, { "epoch": 0.03699886360633209, "grad_norm": 5.719889163970947, "learning_rate": 1.999757423322729e-05, "loss": 9.4736, "step": 1050 }, { "epoch": 0.0370341006192905, "grad_norm": 2.7897961139678955, "learning_rate": 1.9997549031654336e-05, "loss": 4.2386, "step": 1051 }, { "epoch": 0.037069337632248914, "grad_norm": 1.720990777015686, "learning_rate": 1.9997523699862036e-05, "loss": 0.9462, "step": 1052 }, { "epoch": 0.037104574645207326, "grad_norm": 1.965895175933838, "learning_rate": 1.9997498237850718e-05, "loss": 1.168, "step": 1053 }, { "epoch": 0.03713981165816574, "grad_norm": 2.0750043392181396, "learning_rate": 1.999747264562071e-05, "loss": 0.9047, "step": 1054 }, { "epoch": 0.03717504867112415, "grad_norm": 3.1066339015960693, "learning_rate": 1.9997446923172344e-05, "loss": 4.0717, "step": 1055 }, { "epoch": 0.03721028568408256, "grad_norm": 3.3799819946289062, "learning_rate": 1.9997421070505966e-05, "loss": 4.2172, "step": 1056 }, { "epoch": 0.03724552269704097, "grad_norm": 1.5522677898406982, "learning_rate": 1.9997395087621898e-05, "loss": 0.7534, "step": 1057 }, { "epoch": 0.03728075970999938, "grad_norm": 2.726872205734253, "learning_rate": 1.9997368974520488e-05, "loss": 0.8925, "step": 1058 }, { "epoch": 0.037315996722957795, "grad_norm": 1.6738145351409912, "learning_rate": 1.9997342731202077e-05, "loss": 1.1445, "step": 1059 }, { "epoch": 0.03735123373591621, "grad_norm": 1.8880263566970825, "learning_rate": 1.9997316357666997e-05, "loss": 1.2299, "step": 1060 }, { "epoch": 0.03738647074887462, "grad_norm": 1.383594036102295, "learning_rate": 1.9997289853915604e-05, "loss": 1.1677, "step": 1061 }, { "epoch": 0.03742170776183303, "grad_norm": 1.9141651391983032, "learning_rate": 1.9997263219948234e-05, "loss": 4.0902, "step": 1062 }, { "epoch": 0.03745694477479144, "grad_norm": 2.5556302070617676, "learning_rate": 1.9997236455765238e-05, "loss": 3.835, "step": 1063 }, { "epoch": 0.03749218178774985, "grad_norm": 3.563786745071411, "learning_rate": 1.9997209561366963e-05, "loss": 6.2796, "step": 1064 }, { "epoch": 0.03752741880070826, "grad_norm": 1.966538429260254, "learning_rate": 1.999718253675376e-05, "loss": 0.9303, "step": 1065 }, { "epoch": 0.037562655813666676, "grad_norm": 2.2248051166534424, "learning_rate": 1.9997155381925983e-05, "loss": 3.258, "step": 1066 }, { "epoch": 0.03759789282662509, "grad_norm": 1.748478651046753, "learning_rate": 1.999712809688398e-05, "loss": 0.8343, "step": 1067 }, { "epoch": 0.0376331298395835, "grad_norm": 5.507597923278809, "learning_rate": 1.9997100681628113e-05, "loss": 7.0244, "step": 1068 }, { "epoch": 0.03766836685254191, "grad_norm": 6.197465419769287, "learning_rate": 1.9997073136158732e-05, "loss": 6.8466, "step": 1069 }, { "epoch": 0.03770360386550032, "grad_norm": 1.5149672031402588, "learning_rate": 1.9997045460476206e-05, "loss": 0.9693, "step": 1070 }, { "epoch": 0.03773884087845873, "grad_norm": 4.085206031799316, "learning_rate": 1.9997017654580887e-05, "loss": 4.0236, "step": 1071 }, { "epoch": 0.037774077891417145, "grad_norm": 8.540483474731445, "learning_rate": 1.9996989718473138e-05, "loss": 7.2446, "step": 1072 }, { "epoch": 0.03780931490437556, "grad_norm": 3.259023904800415, "learning_rate": 1.9996961652153328e-05, "loss": 3.3646, "step": 1073 }, { "epoch": 0.03784455191733397, "grad_norm": 3.1345999240875244, "learning_rate": 1.9996933455621816e-05, "loss": 3.5244, "step": 1074 }, { "epoch": 0.03787978893029238, "grad_norm": 2.0035409927368164, "learning_rate": 1.9996905128878975e-05, "loss": 1.1605, "step": 1075 }, { "epoch": 0.03791502594325079, "grad_norm": 1.3735361099243164, "learning_rate": 1.9996876671925168e-05, "loss": 1.3355, "step": 1076 }, { "epoch": 0.0379502629562092, "grad_norm": 1.5698992013931274, "learning_rate": 1.999684808476077e-05, "loss": 1.1076, "step": 1077 }, { "epoch": 0.03798549996916761, "grad_norm": 3.2166199684143066, "learning_rate": 1.9996819367386155e-05, "loss": 0.981, "step": 1078 }, { "epoch": 0.038020736982126026, "grad_norm": 3.7716803550720215, "learning_rate": 1.999679051980169e-05, "loss": 7.0676, "step": 1079 }, { "epoch": 0.03805597399508444, "grad_norm": 5.066321849822998, "learning_rate": 1.9996761542007758e-05, "loss": 6.2816, "step": 1080 }, { "epoch": 0.03809121100804285, "grad_norm": 7.87760591506958, "learning_rate": 1.9996732434004732e-05, "loss": 6.8197, "step": 1081 }, { "epoch": 0.03812644802100126, "grad_norm": 3.467142105102539, "learning_rate": 1.9996703195792997e-05, "loss": 4.6126, "step": 1082 }, { "epoch": 0.03816168503395967, "grad_norm": 3.747415065765381, "learning_rate": 1.9996673827372927e-05, "loss": 4.329, "step": 1083 }, { "epoch": 0.03819692204691808, "grad_norm": 4.537320137023926, "learning_rate": 1.9996644328744903e-05, "loss": 3.5078, "step": 1084 }, { "epoch": 0.038232159059876494, "grad_norm": 2.2962846755981445, "learning_rate": 1.9996614699909317e-05, "loss": 3.4233, "step": 1085 }, { "epoch": 0.03826739607283491, "grad_norm": 16.867233276367188, "learning_rate": 1.999658494086655e-05, "loss": 9.0178, "step": 1086 }, { "epoch": 0.03830263308579332, "grad_norm": 4.21895694732666, "learning_rate": 1.9996555051616993e-05, "loss": 3.227, "step": 1087 }, { "epoch": 0.03833787009875173, "grad_norm": 1.2177337408065796, "learning_rate": 1.999652503216103e-05, "loss": 0.9122, "step": 1088 }, { "epoch": 0.03837310711171014, "grad_norm": 1.2254027128219604, "learning_rate": 1.9996494882499058e-05, "loss": 0.967, "step": 1089 }, { "epoch": 0.03840834412466855, "grad_norm": 1.091170310974121, "learning_rate": 1.9996464602631466e-05, "loss": 0.8448, "step": 1090 }, { "epoch": 0.03844358113762696, "grad_norm": 1.3579344749450684, "learning_rate": 1.9996434192558646e-05, "loss": 1.0841, "step": 1091 }, { "epoch": 0.038478818150585375, "grad_norm": 2.976818084716797, "learning_rate": 1.9996403652281003e-05, "loss": 3.8526, "step": 1092 }, { "epoch": 0.03851405516354379, "grad_norm": 7.001910209655762, "learning_rate": 1.9996372981798926e-05, "loss": 6.6279, "step": 1093 }, { "epoch": 0.0385492921765022, "grad_norm": 1.4809613227844238, "learning_rate": 1.9996342181112817e-05, "loss": 1.2528, "step": 1094 }, { "epoch": 0.038584529189460606, "grad_norm": 1.7031824588775635, "learning_rate": 1.9996311250223077e-05, "loss": 1.2745, "step": 1095 }, { "epoch": 0.03861976620241902, "grad_norm": 1.6809484958648682, "learning_rate": 1.999628018913011e-05, "loss": 3.3788, "step": 1096 }, { "epoch": 0.03865500321537743, "grad_norm": 1.6288936138153076, "learning_rate": 1.999624899783432e-05, "loss": 0.7249, "step": 1097 }, { "epoch": 0.038690240228335844, "grad_norm": 1.9269378185272217, "learning_rate": 1.999621767633612e-05, "loss": 1.1587, "step": 1098 }, { "epoch": 0.03872547724129426, "grad_norm": 1.2017240524291992, "learning_rate": 1.9996186224635904e-05, "loss": 1.0443, "step": 1099 }, { "epoch": 0.03876071425425267, "grad_norm": 11.123282432556152, "learning_rate": 1.999615464273409e-05, "loss": 4.4112, "step": 1100 }, { "epoch": 0.03879595126721108, "grad_norm": 1.4430683851242065, "learning_rate": 1.999612293063109e-05, "loss": 1.2384, "step": 1101 }, { "epoch": 0.03883118828016949, "grad_norm": 1.4884179830551147, "learning_rate": 1.9996091088327314e-05, "loss": 3.2526, "step": 1102 }, { "epoch": 0.0388664252931279, "grad_norm": 2.5407798290252686, "learning_rate": 1.999605911582318e-05, "loss": 10.7227, "step": 1103 }, { "epoch": 0.03890166230608631, "grad_norm": 1.3060328960418701, "learning_rate": 1.9996027013119104e-05, "loss": 1.181, "step": 1104 }, { "epoch": 0.038936899319044725, "grad_norm": 1.9542266130447388, "learning_rate": 1.99959947802155e-05, "loss": 1.0695, "step": 1105 }, { "epoch": 0.03897213633200314, "grad_norm": 3.2510995864868164, "learning_rate": 1.9995962417112792e-05, "loss": 4.5312, "step": 1106 }, { "epoch": 0.03900737334496155, "grad_norm": 1.4714243412017822, "learning_rate": 1.99959299238114e-05, "loss": 1.1874, "step": 1107 }, { "epoch": 0.039042610357919956, "grad_norm": 3.418652296066284, "learning_rate": 1.9995897300311745e-05, "loss": 1.5221, "step": 1108 }, { "epoch": 0.03907784737087837, "grad_norm": 2.364342212677002, "learning_rate": 1.999586454661426e-05, "loss": 1.1409, "step": 1109 }, { "epoch": 0.03911308438383678, "grad_norm": 2.0524768829345703, "learning_rate": 1.999583166271936e-05, "loss": 0.9447, "step": 1110 }, { "epoch": 0.039148321396795194, "grad_norm": 2.3669960498809814, "learning_rate": 1.9995798648627484e-05, "loss": 1.2334, "step": 1111 }, { "epoch": 0.039183558409753606, "grad_norm": 1.2137889862060547, "learning_rate": 1.9995765504339057e-05, "loss": 0.9996, "step": 1112 }, { "epoch": 0.03921879542271202, "grad_norm": 2.000868320465088, "learning_rate": 1.9995732229854506e-05, "loss": 3.9592, "step": 1113 }, { "epoch": 0.03925403243567043, "grad_norm": 2.928051710128784, "learning_rate": 1.9995698825174273e-05, "loss": 3.7786, "step": 1114 }, { "epoch": 0.03928926944862884, "grad_norm": 1.987274169921875, "learning_rate": 1.999566529029879e-05, "loss": 3.486, "step": 1115 }, { "epoch": 0.03932450646158725, "grad_norm": 2.0675930976867676, "learning_rate": 1.9995631625228493e-05, "loss": 1.0572, "step": 1116 }, { "epoch": 0.03935974347454566, "grad_norm": 2.758378267288208, "learning_rate": 1.999559782996382e-05, "loss": 6.3492, "step": 1117 }, { "epoch": 0.039394980487504075, "grad_norm": 2.2366294860839844, "learning_rate": 1.9995563904505214e-05, "loss": 4.1355, "step": 1118 }, { "epoch": 0.03943021750046249, "grad_norm": 1.6582248210906982, "learning_rate": 1.9995529848853114e-05, "loss": 1.1086, "step": 1119 }, { "epoch": 0.0394654545134209, "grad_norm": 2.257234573364258, "learning_rate": 1.9995495663007963e-05, "loss": 6.6124, "step": 1120 }, { "epoch": 0.039500691526379306, "grad_norm": 2.5555331707000732, "learning_rate": 1.999546134697021e-05, "loss": 4.217, "step": 1121 }, { "epoch": 0.03953592853933772, "grad_norm": 1.3452879190444946, "learning_rate": 1.99954269007403e-05, "loss": 1.0292, "step": 1122 }, { "epoch": 0.03957116555229613, "grad_norm": 1.712498664855957, "learning_rate": 1.999539232431868e-05, "loss": 0.9404, "step": 1123 }, { "epoch": 0.03960640256525454, "grad_norm": 3.342646598815918, "learning_rate": 1.99953576177058e-05, "loss": 3.9132, "step": 1124 }, { "epoch": 0.039641639578212956, "grad_norm": 2.3873026371002197, "learning_rate": 1.999532278090212e-05, "loss": 4.0139, "step": 1125 }, { "epoch": 0.03967687659117137, "grad_norm": 2.3873026371002197, "learning_rate": 1.999532278090212e-05, "loss": 3.8593, "step": 1126 }, { "epoch": 0.03971211360412978, "grad_norm": 1.6442582607269287, "learning_rate": 1.999528781390808e-05, "loss": 0.966, "step": 1127 }, { "epoch": 0.03974735061708819, "grad_norm": 2.9062182903289795, "learning_rate": 1.999525271672415e-05, "loss": 6.6482, "step": 1128 }, { "epoch": 0.0397825876300466, "grad_norm": 2.364696741104126, "learning_rate": 1.9995217489350772e-05, "loss": 3.3792, "step": 1129 }, { "epoch": 0.03981782464300501, "grad_norm": 2.012349843978882, "learning_rate": 1.999518213178842e-05, "loss": 1.1832, "step": 1130 }, { "epoch": 0.039853061655963425, "grad_norm": 1.4853695631027222, "learning_rate": 1.9995146644037544e-05, "loss": 1.0356, "step": 1131 }, { "epoch": 0.03988829866892184, "grad_norm": 1.5144319534301758, "learning_rate": 1.999511102609861e-05, "loss": 1.0351, "step": 1132 }, { "epoch": 0.03992353568188025, "grad_norm": 3.932882785797119, "learning_rate": 1.9995075277972082e-05, "loss": 3.5893, "step": 1133 }, { "epoch": 0.039958772694838655, "grad_norm": 1.2553361654281616, "learning_rate": 1.9995039399658426e-05, "loss": 1.1549, "step": 1134 }, { "epoch": 0.03999400970779707, "grad_norm": 1.8206309080123901, "learning_rate": 1.999500339115811e-05, "loss": 0.9912, "step": 1135 }, { "epoch": 0.04002924672075548, "grad_norm": 2.775944232940674, "learning_rate": 1.99949672524716e-05, "loss": 4.2225, "step": 1136 }, { "epoch": 0.04006448373371389, "grad_norm": 2.4047605991363525, "learning_rate": 1.9994930983599367e-05, "loss": 0.91, "step": 1137 }, { "epoch": 0.040099720746672306, "grad_norm": 1.1273597478866577, "learning_rate": 1.9994894584541886e-05, "loss": 0.8937, "step": 1138 }, { "epoch": 0.04013495775963072, "grad_norm": 2.60440993309021, "learning_rate": 1.999485805529963e-05, "loss": 3.7679, "step": 1139 }, { "epoch": 0.04017019477258913, "grad_norm": 3.1257777214050293, "learning_rate": 1.9994821395873077e-05, "loss": 6.9971, "step": 1140 }, { "epoch": 0.04020543178554754, "grad_norm": 2.4764091968536377, "learning_rate": 1.99947846062627e-05, "loss": 3.792, "step": 1141 }, { "epoch": 0.04024066879850595, "grad_norm": 4.579751014709473, "learning_rate": 1.999474768646898e-05, "loss": 4.0613, "step": 1142 }, { "epoch": 0.04027590581146436, "grad_norm": 3.095475912094116, "learning_rate": 1.99947106364924e-05, "loss": 3.267, "step": 1143 }, { "epoch": 0.040311142824422774, "grad_norm": 1.1317017078399658, "learning_rate": 1.9994673456333437e-05, "loss": 0.9903, "step": 1144 }, { "epoch": 0.04034637983738119, "grad_norm": 4.33085823059082, "learning_rate": 1.9994636145992585e-05, "loss": 3.6852, "step": 1145 }, { "epoch": 0.0403816168503396, "grad_norm": 2.481201648712158, "learning_rate": 1.9994598705470317e-05, "loss": 3.6997, "step": 1146 }, { "epoch": 0.040416853863298005, "grad_norm": 1.3125606775283813, "learning_rate": 1.9994561134767133e-05, "loss": 0.7809, "step": 1147 }, { "epoch": 0.04045209087625642, "grad_norm": 2.0787580013275146, "learning_rate": 1.9994523433883516e-05, "loss": 4.3375, "step": 1148 }, { "epoch": 0.04048732788921483, "grad_norm": 4.6417555809021, "learning_rate": 1.9994485602819957e-05, "loss": 6.7957, "step": 1149 }, { "epoch": 0.04052256490217324, "grad_norm": 4.6417555809021, "learning_rate": 1.9994485602819957e-05, "loss": 3.9337, "step": 1150 }, { "epoch": 0.040557801915131655, "grad_norm": 1.4289450645446777, "learning_rate": 1.9994447641576952e-05, "loss": 1.0982, "step": 1151 }, { "epoch": 0.04059303892809007, "grad_norm": 2.778306722640991, "learning_rate": 1.999440955015499e-05, "loss": 3.9329, "step": 1152 }, { "epoch": 0.04062827594104848, "grad_norm": 1.4717669486999512, "learning_rate": 1.9994371328554573e-05, "loss": 1.2876, "step": 1153 }, { "epoch": 0.040663512954006886, "grad_norm": 2.7399768829345703, "learning_rate": 1.9994332976776196e-05, "loss": 0.9786, "step": 1154 }, { "epoch": 0.0406987499669653, "grad_norm": 2.349843740463257, "learning_rate": 1.9994294494820357e-05, "loss": 4.0527, "step": 1155 }, { "epoch": 0.04073398697992371, "grad_norm": 5.115382671356201, "learning_rate": 1.9994255882687562e-05, "loss": 8.3619, "step": 1156 }, { "epoch": 0.040769223992882124, "grad_norm": 1.326279640197754, "learning_rate": 1.9994217140378312e-05, "loss": 1.0936, "step": 1157 }, { "epoch": 0.04080446100584054, "grad_norm": 2.116241455078125, "learning_rate": 1.9994178267893108e-05, "loss": 3.4177, "step": 1158 }, { "epoch": 0.04083969801879895, "grad_norm": 1.4206382036209106, "learning_rate": 1.999413926523246e-05, "loss": 0.961, "step": 1159 }, { "epoch": 0.040874935031757355, "grad_norm": 2.3035809993743896, "learning_rate": 1.9994100132396873e-05, "loss": 3.3947, "step": 1160 }, { "epoch": 0.04091017204471577, "grad_norm": 1.570839762687683, "learning_rate": 1.999406086938686e-05, "loss": 1.125, "step": 1161 }, { "epoch": 0.04094540905767418, "grad_norm": 1.6429487466812134, "learning_rate": 1.9994021476202935e-05, "loss": 1.4942, "step": 1162 }, { "epoch": 0.04098064607063259, "grad_norm": 3.2984843254089355, "learning_rate": 1.9993981952845604e-05, "loss": 3.2739, "step": 1163 }, { "epoch": 0.041015883083591005, "grad_norm": 2.194204330444336, "learning_rate": 1.9993942299315385e-05, "loss": 0.97, "step": 1164 }, { "epoch": 0.04105112009654942, "grad_norm": 1.40186607837677, "learning_rate": 1.9993902515612794e-05, "loss": 0.9039, "step": 1165 }, { "epoch": 0.04108635710950783, "grad_norm": 4.178545951843262, "learning_rate": 1.999386260173835e-05, "loss": 3.6576, "step": 1166 }, { "epoch": 0.041121594122466236, "grad_norm": 2.5689644813537598, "learning_rate": 1.9993822557692573e-05, "loss": 3.8333, "step": 1167 }, { "epoch": 0.04115683113542465, "grad_norm": 3.479887008666992, "learning_rate": 1.9993782383475985e-05, "loss": 4.0525, "step": 1168 }, { "epoch": 0.04119206814838306, "grad_norm": 1.5584560632705688, "learning_rate": 1.9993742079089106e-05, "loss": 1.006, "step": 1169 }, { "epoch": 0.041227305161341474, "grad_norm": 1.9033303260803223, "learning_rate": 1.9993701644532463e-05, "loss": 1.3183, "step": 1170 }, { "epoch": 0.041262542174299886, "grad_norm": 1.5033763647079468, "learning_rate": 1.9993661079806585e-05, "loss": 0.8417, "step": 1171 }, { "epoch": 0.0412977791872583, "grad_norm": 1.2956141233444214, "learning_rate": 1.9993620384911998e-05, "loss": 0.9902, "step": 1172 }, { "epoch": 0.041333016200216705, "grad_norm": 1.7333823442459106, "learning_rate": 1.9993579559849236e-05, "loss": 4.1831, "step": 1173 }, { "epoch": 0.04136825321317512, "grad_norm": 1.2000420093536377, "learning_rate": 1.9993538604618824e-05, "loss": 1.0763, "step": 1174 }, { "epoch": 0.04140349022613353, "grad_norm": 2.6348588466644287, "learning_rate": 1.9993497519221296e-05, "loss": 3.1278, "step": 1175 }, { "epoch": 0.04143872723909194, "grad_norm": 1.827543020248413, "learning_rate": 1.9993456303657192e-05, "loss": 0.9969, "step": 1176 }, { "epoch": 0.041473964252050355, "grad_norm": 2.7051451206207275, "learning_rate": 1.9993414957927052e-05, "loss": 6.8525, "step": 1177 }, { "epoch": 0.04150920126500877, "grad_norm": 2.529736042022705, "learning_rate": 1.99933734820314e-05, "loss": 5.7557, "step": 1178 }, { "epoch": 0.04154443827796718, "grad_norm": 2.163773536682129, "learning_rate": 1.9993331875970793e-05, "loss": 6.832, "step": 1179 }, { "epoch": 0.041579675290925586, "grad_norm": 2.5172760486602783, "learning_rate": 1.9993290139745762e-05, "loss": 6.9757, "step": 1180 }, { "epoch": 0.041614912303884, "grad_norm": 1.5026705265045166, "learning_rate": 1.9993248273356853e-05, "loss": 1.0535, "step": 1181 }, { "epoch": 0.04165014931684241, "grad_norm": 3.039032220840454, "learning_rate": 1.9993206276804614e-05, "loss": 6.8882, "step": 1182 }, { "epoch": 0.04168538632980082, "grad_norm": 2.6783721446990967, "learning_rate": 1.999316415008959e-05, "loss": 3.7176, "step": 1183 }, { "epoch": 0.041720623342759236, "grad_norm": 3.3356430530548096, "learning_rate": 1.999312189321233e-05, "loss": 10.2526, "step": 1184 }, { "epoch": 0.04175586035571765, "grad_norm": 1.8297348022460938, "learning_rate": 1.9993079506173383e-05, "loss": 3.2879, "step": 1185 }, { "epoch": 0.041791097368676054, "grad_norm": 1.0215281248092651, "learning_rate": 1.9993036988973306e-05, "loss": 0.9099, "step": 1186 }, { "epoch": 0.04182633438163447, "grad_norm": 1.2912932634353638, "learning_rate": 1.9992994341612647e-05, "loss": 0.9999, "step": 1187 }, { "epoch": 0.04186157139459288, "grad_norm": 6.227427959442139, "learning_rate": 1.9992951564091965e-05, "loss": 9.3406, "step": 1188 }, { "epoch": 0.04189680840755129, "grad_norm": 2.779431104660034, "learning_rate": 1.9992908656411813e-05, "loss": 4.4077, "step": 1189 }, { "epoch": 0.041932045420509705, "grad_norm": 2.3633735179901123, "learning_rate": 1.9992865618572756e-05, "loss": 3.1992, "step": 1190 }, { "epoch": 0.04196728243346812, "grad_norm": 3.1043148040771484, "learning_rate": 1.9992822450575347e-05, "loss": 7.6066, "step": 1191 }, { "epoch": 0.04200251944642653, "grad_norm": 1.751436710357666, "learning_rate": 1.999277915242016e-05, "loss": 0.9956, "step": 1192 }, { "epoch": 0.042037756459384935, "grad_norm": 2.7614662647247314, "learning_rate": 1.9992735724107746e-05, "loss": 4.1613, "step": 1193 }, { "epoch": 0.04207299347234335, "grad_norm": 4.2806830406188965, "learning_rate": 1.999269216563868e-05, "loss": 6.65, "step": 1194 }, { "epoch": 0.04210823048530176, "grad_norm": 2.426837205886841, "learning_rate": 1.999264847701352e-05, "loss": 1.096, "step": 1195 }, { "epoch": 0.04214346749826017, "grad_norm": 3.5430052280426025, "learning_rate": 1.9992604658232847e-05, "loss": 3.5009, "step": 1196 }, { "epoch": 0.042178704511218586, "grad_norm": 1.3167824745178223, "learning_rate": 1.999256070929722e-05, "loss": 0.8813, "step": 1197 }, { "epoch": 0.042213941524177, "grad_norm": 1.440765619277954, "learning_rate": 1.9992516630207223e-05, "loss": 1.1808, "step": 1198 }, { "epoch": 0.042249178537135404, "grad_norm": 1.9696208238601685, "learning_rate": 1.999247242096342e-05, "loss": 1.0184, "step": 1199 }, { "epoch": 0.042284415550093817, "grad_norm": 3.1499011516571045, "learning_rate": 1.999242808156639e-05, "loss": 3.2491, "step": 1200 }, { "epoch": 0.04231965256305223, "grad_norm": 5.656716346740723, "learning_rate": 1.9992383612016716e-05, "loss": 6.5997, "step": 1201 }, { "epoch": 0.04235488957601064, "grad_norm": 2.885369300842285, "learning_rate": 1.999233901231497e-05, "loss": 3.055, "step": 1202 }, { "epoch": 0.042390126588969054, "grad_norm": 2.244373321533203, "learning_rate": 1.9992294282461736e-05, "loss": 3.878, "step": 1203 }, { "epoch": 0.04242536360192747, "grad_norm": 2.9732484817504883, "learning_rate": 1.9992249422457593e-05, "loss": 3.4157, "step": 1204 }, { "epoch": 0.04246060061488588, "grad_norm": 2.861978769302368, "learning_rate": 1.999220443230313e-05, "loss": 6.6669, "step": 1205 }, { "epoch": 0.042495837627844285, "grad_norm": 3.573509931564331, "learning_rate": 1.9992159311998934e-05, "loss": 6.4863, "step": 1206 }, { "epoch": 0.0425310746408027, "grad_norm": 1.227839469909668, "learning_rate": 1.999211406154559e-05, "loss": 1.1695, "step": 1207 }, { "epoch": 0.04256631165376111, "grad_norm": 1.722206473350525, "learning_rate": 1.9992068680943685e-05, "loss": 1.2852, "step": 1208 }, { "epoch": 0.04260154866671952, "grad_norm": 5.5658793449401855, "learning_rate": 1.9992023170193812e-05, "loss": 6.8359, "step": 1209 }, { "epoch": 0.042636785679677935, "grad_norm": 2.897955894470215, "learning_rate": 1.9991977529296565e-05, "loss": 4.5381, "step": 1210 }, { "epoch": 0.04267202269263635, "grad_norm": 3.058993339538574, "learning_rate": 1.999193175825254e-05, "loss": 3.3356, "step": 1211 }, { "epoch": 0.042707259705594754, "grad_norm": 2.8894717693328857, "learning_rate": 1.9991885857062332e-05, "loss": 3.8103, "step": 1212 }, { "epoch": 0.042742496718553166, "grad_norm": 1.4453061819076538, "learning_rate": 1.9991839825726535e-05, "loss": 0.9535, "step": 1213 }, { "epoch": 0.04277773373151158, "grad_norm": 3.6529924869537354, "learning_rate": 1.9991793664245753e-05, "loss": 12.1624, "step": 1214 }, { "epoch": 0.04281297074446999, "grad_norm": 1.6002594232559204, "learning_rate": 1.9991747372620584e-05, "loss": 1.0224, "step": 1215 }, { "epoch": 0.042848207757428404, "grad_norm": 1.9182184934616089, "learning_rate": 1.9991700950851635e-05, "loss": 3.5895, "step": 1216 }, { "epoch": 0.04288344477038682, "grad_norm": 1.4570732116699219, "learning_rate": 1.999165439893951e-05, "loss": 1.1795, "step": 1217 }, { "epoch": 0.04291868178334523, "grad_norm": 1.6230175495147705, "learning_rate": 1.999160771688481e-05, "loss": 3.4181, "step": 1218 }, { "epoch": 0.042953918796303635, "grad_norm": 2.4258735179901123, "learning_rate": 1.9991560904688146e-05, "loss": 3.3813, "step": 1219 }, { "epoch": 0.04298915580926205, "grad_norm": 2.608717679977417, "learning_rate": 1.9991513962350132e-05, "loss": 3.7576, "step": 1220 }, { "epoch": 0.04302439282222046, "grad_norm": 3.9194440841674805, "learning_rate": 1.9991466889871375e-05, "loss": 3.9933, "step": 1221 }, { "epoch": 0.04305962983517887, "grad_norm": 5.271313190460205, "learning_rate": 1.9991419687252486e-05, "loss": 6.6961, "step": 1222 }, { "epoch": 0.043094866848137285, "grad_norm": 4.689140796661377, "learning_rate": 1.9991372354494086e-05, "loss": 6.9736, "step": 1223 }, { "epoch": 0.0431301038610957, "grad_norm": 6.03997802734375, "learning_rate": 1.9991324891596786e-05, "loss": 3.0632, "step": 1224 }, { "epoch": 0.0431653408740541, "grad_norm": 3.0198538303375244, "learning_rate": 1.9991277298561208e-05, "loss": 3.8959, "step": 1225 }, { "epoch": 0.043200577887012516, "grad_norm": 2.4425456523895264, "learning_rate": 1.9991229575387972e-05, "loss": 4.1776, "step": 1226 }, { "epoch": 0.04323581489997093, "grad_norm": 3.8539984226226807, "learning_rate": 1.9991181722077695e-05, "loss": 9.5873, "step": 1227 }, { "epoch": 0.04327105191292934, "grad_norm": 2.1912729740142822, "learning_rate": 1.999113373863101e-05, "loss": 0.9954, "step": 1228 }, { "epoch": 0.043306288925887754, "grad_norm": 1.5982335805892944, "learning_rate": 1.9991085625048527e-05, "loss": 1.1179, "step": 1229 }, { "epoch": 0.043341525938846166, "grad_norm": 1.5982335805892944, "learning_rate": 1.9991085625048527e-05, "loss": 3.7793, "step": 1230 }, { "epoch": 0.04337676295180458, "grad_norm": 1.5982335805892944, "learning_rate": 1.9991085625048527e-05, "loss": 7.1188, "step": 1231 }, { "epoch": 0.043411999964762985, "grad_norm": 1.5982335805892944, "learning_rate": 1.9991085625048527e-05, "loss": 6.4198, "step": 1232 }, { "epoch": 0.0434472369777214, "grad_norm": 4.32018518447876, "learning_rate": 1.9991037381330885e-05, "loss": 1.0972, "step": 1233 }, { "epoch": 0.04348247399067981, "grad_norm": 4.309384346008301, "learning_rate": 1.999098900747871e-05, "loss": 6.2492, "step": 1234 }, { "epoch": 0.04351771100363822, "grad_norm": 1.9435805082321167, "learning_rate": 1.9990940503492625e-05, "loss": 1.1124, "step": 1235 }, { "epoch": 0.043552948016596635, "grad_norm": 2.066457509994507, "learning_rate": 1.9990891869373273e-05, "loss": 0.9372, "step": 1236 }, { "epoch": 0.04358818502955505, "grad_norm": 3.076425552368164, "learning_rate": 1.999084310512128e-05, "loss": 3.1, "step": 1237 }, { "epoch": 0.04362342204251345, "grad_norm": 1.8911223411560059, "learning_rate": 1.9990794210737283e-05, "loss": 0.8245, "step": 1238 }, { "epoch": 0.043658659055471866, "grad_norm": 6.3205180168151855, "learning_rate": 1.999074518622192e-05, "loss": 8.0918, "step": 1239 }, { "epoch": 0.04369389606843028, "grad_norm": 4.471296787261963, "learning_rate": 1.9990696031575824e-05, "loss": 4.0615, "step": 1240 }, { "epoch": 0.04372913308138869, "grad_norm": 1.7235233783721924, "learning_rate": 1.9990646746799643e-05, "loss": 0.9977, "step": 1241 }, { "epoch": 0.0437643700943471, "grad_norm": 3.5516927242279053, "learning_rate": 1.9990597331894014e-05, "loss": 3.2599, "step": 1242 }, { "epoch": 0.043799607107305516, "grad_norm": 1.4152330160140991, "learning_rate": 1.9990547786859584e-05, "loss": 1.025, "step": 1243 }, { "epoch": 0.04383484412026393, "grad_norm": 1.7445459365844727, "learning_rate": 1.9990498111696997e-05, "loss": 0.8726, "step": 1244 }, { "epoch": 0.043870081133222334, "grad_norm": 4.567641258239746, "learning_rate": 1.9990448306406897e-05, "loss": 6.5033, "step": 1245 }, { "epoch": 0.04390531814618075, "grad_norm": 1.8600736856460571, "learning_rate": 1.9990398370989937e-05, "loss": 0.9096, "step": 1246 }, { "epoch": 0.04394055515913916, "grad_norm": 2.8278465270996094, "learning_rate": 1.9990348305446764e-05, "loss": 3.4647, "step": 1247 }, { "epoch": 0.04397579217209757, "grad_norm": 2.0560643672943115, "learning_rate": 1.9990298109778033e-05, "loss": 0.9461, "step": 1248 }, { "epoch": 0.044011029185055985, "grad_norm": 1.626145601272583, "learning_rate": 1.9990247783984394e-05, "loss": 0.9072, "step": 1249 }, { "epoch": 0.0440462661980144, "grad_norm": 5.000935077667236, "learning_rate": 1.9990197328066508e-05, "loss": 3.3514, "step": 1250 }, { "epoch": 0.0440815032109728, "grad_norm": 3.390763759613037, "learning_rate": 1.999014674202503e-05, "loss": 4.3553, "step": 1251 }, { "epoch": 0.044116740223931215, "grad_norm": 6.38892126083374, "learning_rate": 1.9990096025860617e-05, "loss": 4.5586, "step": 1252 }, { "epoch": 0.04415197723688963, "grad_norm": 1.2914763689041138, "learning_rate": 1.999004517957393e-05, "loss": 1.0317, "step": 1253 }, { "epoch": 0.04418721424984804, "grad_norm": 1.453742504119873, "learning_rate": 1.9989994203165633e-05, "loss": 0.8773, "step": 1254 }, { "epoch": 0.04422245126280645, "grad_norm": 2.223179578781128, "learning_rate": 1.9989943096636387e-05, "loss": 1.1139, "step": 1255 }, { "epoch": 0.044257688275764866, "grad_norm": 6.888592720031738, "learning_rate": 1.9989891859986865e-05, "loss": 6.2911, "step": 1256 }, { "epoch": 0.04429292528872328, "grad_norm": 1.2930799722671509, "learning_rate": 1.9989840493217723e-05, "loss": 0.9527, "step": 1257 }, { "epoch": 0.044328162301681684, "grad_norm": 1.4179959297180176, "learning_rate": 1.9989788996329636e-05, "loss": 1.0612, "step": 1258 }, { "epoch": 0.044363399314640097, "grad_norm": 1.3612865209579468, "learning_rate": 1.9989737369323276e-05, "loss": 0.9179, "step": 1259 }, { "epoch": 0.04439863632759851, "grad_norm": 1.783510684967041, "learning_rate": 1.9989685612199315e-05, "loss": 3.9474, "step": 1260 }, { "epoch": 0.04443387334055692, "grad_norm": 5.368151664733887, "learning_rate": 1.998963372495843e-05, "loss": 6.1834, "step": 1261 }, { "epoch": 0.044469110353515334, "grad_norm": 3.167523145675659, "learning_rate": 1.998958170760129e-05, "loss": 8.6449, "step": 1262 }, { "epoch": 0.04450434736647375, "grad_norm": 2.666649341583252, "learning_rate": 1.9989529560128575e-05, "loss": 3.7079, "step": 1263 }, { "epoch": 0.04453958437943215, "grad_norm": 1.8077822923660278, "learning_rate": 1.9989477282540965e-05, "loss": 1.2611, "step": 1264 }, { "epoch": 0.044574821392390565, "grad_norm": 1.5149831771850586, "learning_rate": 1.9989424874839143e-05, "loss": 0.9974, "step": 1265 }, { "epoch": 0.04461005840534898, "grad_norm": 5.004316329956055, "learning_rate": 1.9989372337023787e-05, "loss": 6.8336, "step": 1266 }, { "epoch": 0.04464529541830739, "grad_norm": 1.425620675086975, "learning_rate": 1.9989319669095585e-05, "loss": 1.3023, "step": 1267 }, { "epoch": 0.0446805324312658, "grad_norm": 3.0774831771850586, "learning_rate": 1.998926687105522e-05, "loss": 3.9332, "step": 1268 }, { "epoch": 0.044715769444224215, "grad_norm": 1.7858388423919678, "learning_rate": 1.9989213942903386e-05, "loss": 1.0708, "step": 1269 }, { "epoch": 0.04475100645718263, "grad_norm": 8.040963172912598, "learning_rate": 1.9989160884640765e-05, "loss": 5.9152, "step": 1270 }, { "epoch": 0.044786243470141034, "grad_norm": 2.6326591968536377, "learning_rate": 1.998910769626805e-05, "loss": 1.213, "step": 1271 }, { "epoch": 0.044821480483099446, "grad_norm": 1.4066827297210693, "learning_rate": 1.9989054377785936e-05, "loss": 1.2561, "step": 1272 }, { "epoch": 0.04485671749605786, "grad_norm": 1.2843296527862549, "learning_rate": 1.9989000929195117e-05, "loss": 0.979, "step": 1273 }, { "epoch": 0.04489195450901627, "grad_norm": 4.704237937927246, "learning_rate": 1.9988947350496287e-05, "loss": 4.1551, "step": 1274 }, { "epoch": 0.044927191521974684, "grad_norm": 1.6225860118865967, "learning_rate": 1.9988893641690147e-05, "loss": 1.2239, "step": 1275 }, { "epoch": 0.0449624285349331, "grad_norm": 3.009521245956421, "learning_rate": 1.9988839802777398e-05, "loss": 6.5214, "step": 1276 }, { "epoch": 0.0449976655478915, "grad_norm": 1.257659673690796, "learning_rate": 1.9988785833758735e-05, "loss": 0.9418, "step": 1277 }, { "epoch": 0.045032902560849915, "grad_norm": 3.671478033065796, "learning_rate": 1.9988731734634862e-05, "loss": 4.2375, "step": 1278 }, { "epoch": 0.04506813957380833, "grad_norm": 1.4941494464874268, "learning_rate": 1.9988677505406485e-05, "loss": 0.9662, "step": 1279 }, { "epoch": 0.04510337658676674, "grad_norm": 1.7199701070785522, "learning_rate": 1.9988623146074318e-05, "loss": 1.0558, "step": 1280 }, { "epoch": 0.04513861359972515, "grad_norm": 1.4661529064178467, "learning_rate": 1.9988568656639056e-05, "loss": 0.8097, "step": 1281 }, { "epoch": 0.045173850612683565, "grad_norm": 2.9375927448272705, "learning_rate": 1.9988514037101417e-05, "loss": 4.0215, "step": 1282 }, { "epoch": 0.04520908762564198, "grad_norm": 3.941328525543213, "learning_rate": 1.998845928746211e-05, "loss": 3.8977, "step": 1283 }, { "epoch": 0.04524432463860038, "grad_norm": 3.9152896404266357, "learning_rate": 1.998840440772185e-05, "loss": 5.8175, "step": 1284 }, { "epoch": 0.045279561651558796, "grad_norm": 3.209404230117798, "learning_rate": 1.9988349397881347e-05, "loss": 3.9554, "step": 1285 }, { "epoch": 0.04531479866451721, "grad_norm": 2.191246747970581, "learning_rate": 1.9988294257941322e-05, "loss": 3.8994, "step": 1286 }, { "epoch": 0.04535003567747562, "grad_norm": 1.4518388509750366, "learning_rate": 1.9988238987902492e-05, "loss": 1.1323, "step": 1287 }, { "epoch": 0.045385272690434034, "grad_norm": 1.6044137477874756, "learning_rate": 1.998818358776558e-05, "loss": 0.8962, "step": 1288 }, { "epoch": 0.045420509703392446, "grad_norm": 2.2071373462677, "learning_rate": 1.99881280575313e-05, "loss": 3.0953, "step": 1289 }, { "epoch": 0.04545574671635085, "grad_norm": 3.5528554916381836, "learning_rate": 1.9988072397200386e-05, "loss": 6.101, "step": 1290 }, { "epoch": 0.045490983729309264, "grad_norm": 2.6832265853881836, "learning_rate": 1.998801660677355e-05, "loss": 1.2078, "step": 1291 }, { "epoch": 0.04552622074226768, "grad_norm": 1.404910683631897, "learning_rate": 1.9987960686251528e-05, "loss": 1.0983, "step": 1292 }, { "epoch": 0.04556145775522609, "grad_norm": 2.0780880451202393, "learning_rate": 1.9987904635635048e-05, "loss": 0.8895, "step": 1293 }, { "epoch": 0.0455966947681845, "grad_norm": 2.161149024963379, "learning_rate": 1.9987848454924835e-05, "loss": 1.1006, "step": 1294 }, { "epoch": 0.045631931781142915, "grad_norm": 2.288499355316162, "learning_rate": 1.9987792144121623e-05, "loss": 1.0574, "step": 1295 }, { "epoch": 0.04566716879410133, "grad_norm": 4.262301445007324, "learning_rate": 1.9987735703226147e-05, "loss": 4.4178, "step": 1296 }, { "epoch": 0.04570240580705973, "grad_norm": 1.4425554275512695, "learning_rate": 1.998767913223914e-05, "loss": 0.9018, "step": 1297 }, { "epoch": 0.045737642820018146, "grad_norm": 5.684732437133789, "learning_rate": 1.9987622431161344e-05, "loss": 3.6111, "step": 1298 }, { "epoch": 0.04577287983297656, "grad_norm": 6.660050392150879, "learning_rate": 1.998756559999349e-05, "loss": 3.4968, "step": 1299 }, { "epoch": 0.04580811684593497, "grad_norm": 1.838180661201477, "learning_rate": 1.9987508638736324e-05, "loss": 0.8963, "step": 1300 }, { "epoch": 0.04584335385889338, "grad_norm": 3.5010604858398438, "learning_rate": 1.9987451547390587e-05, "loss": 3.588, "step": 1301 }, { "epoch": 0.045878590871851796, "grad_norm": 3.246307611465454, "learning_rate": 1.9987394325957016e-05, "loss": 6.0371, "step": 1302 }, { "epoch": 0.0459138278848102, "grad_norm": 1.421073317527771, "learning_rate": 1.9987336974436368e-05, "loss": 0.8443, "step": 1303 }, { "epoch": 0.045949064897768614, "grad_norm": 3.384585380554199, "learning_rate": 1.998727949282938e-05, "loss": 6.4875, "step": 1304 }, { "epoch": 0.04598430191072703, "grad_norm": 2.0303027629852295, "learning_rate": 1.9987221881136806e-05, "loss": 3.6105, "step": 1305 }, { "epoch": 0.04601953892368544, "grad_norm": 2.0033392906188965, "learning_rate": 1.9987164139359396e-05, "loss": 1.0156, "step": 1306 }, { "epoch": 0.04605477593664385, "grad_norm": 1.296903133392334, "learning_rate": 1.9987106267497894e-05, "loss": 0.876, "step": 1307 }, { "epoch": 0.046090012949602265, "grad_norm": 1.8447284698486328, "learning_rate": 1.9987048265553067e-05, "loss": 1.1539, "step": 1308 }, { "epoch": 0.04612524996256068, "grad_norm": 2.3224878311157227, "learning_rate": 1.998699013352566e-05, "loss": 1.0995, "step": 1309 }, { "epoch": 0.04616048697551908, "grad_norm": 1.9349111318588257, "learning_rate": 1.998693187141644e-05, "loss": 1.2111, "step": 1310 }, { "epoch": 0.046195723988477495, "grad_norm": 2.2071399688720703, "learning_rate": 1.9986873479226155e-05, "loss": 3.2345, "step": 1311 }, { "epoch": 0.04623096100143591, "grad_norm": 2.441451072692871, "learning_rate": 1.9986814956955575e-05, "loss": 3.5591, "step": 1312 }, { "epoch": 0.04626619801439432, "grad_norm": 1.524182677268982, "learning_rate": 1.9986756304605458e-05, "loss": 1.0649, "step": 1313 }, { "epoch": 0.04630143502735273, "grad_norm": 5.202826976776123, "learning_rate": 1.9986697522176565e-05, "loss": 3.6771, "step": 1314 }, { "epoch": 0.046336672040311146, "grad_norm": 4.669036865234375, "learning_rate": 1.9986638609669667e-05, "loss": 6.1228, "step": 1315 }, { "epoch": 0.04637190905326955, "grad_norm": 1.4947469234466553, "learning_rate": 1.9986579567085528e-05, "loss": 0.9352, "step": 1316 }, { "epoch": 0.046407146066227964, "grad_norm": 3.1117265224456787, "learning_rate": 1.9986520394424918e-05, "loss": 4.3583, "step": 1317 }, { "epoch": 0.046442383079186376, "grad_norm": 1.534356713294983, "learning_rate": 1.998646109168861e-05, "loss": 1.2112, "step": 1318 }, { "epoch": 0.04647762009214479, "grad_norm": 3.2283527851104736, "learning_rate": 1.998640165887737e-05, "loss": 4.0856, "step": 1319 }, { "epoch": 0.0465128571051032, "grad_norm": 1.7047152519226074, "learning_rate": 1.9986342095991978e-05, "loss": 0.9845, "step": 1320 }, { "epoch": 0.046548094118061614, "grad_norm": 4.896782398223877, "learning_rate": 1.9986282403033207e-05, "loss": 2.9542, "step": 1321 }, { "epoch": 0.04658333113102003, "grad_norm": 1.2586899995803833, "learning_rate": 1.998622258000184e-05, "loss": 0.9745, "step": 1322 }, { "epoch": 0.04661856814397843, "grad_norm": 1.460044026374817, "learning_rate": 1.998616262689865e-05, "loss": 1.1178, "step": 1323 }, { "epoch": 0.046653805156936845, "grad_norm": 4.309746742248535, "learning_rate": 1.998610254372442e-05, "loss": 3.7926, "step": 1324 }, { "epoch": 0.04668904216989526, "grad_norm": 4.309670448303223, "learning_rate": 1.998604233047993e-05, "loss": 6.12, "step": 1325 }, { "epoch": 0.04672427918285367, "grad_norm": 7.444100856781006, "learning_rate": 1.9985981987165966e-05, "loss": 3.4221, "step": 1326 }, { "epoch": 0.04675951619581208, "grad_norm": 2.7478744983673096, "learning_rate": 1.9985921513783317e-05, "loss": 3.4233, "step": 1327 }, { "epoch": 0.046794753208770495, "grad_norm": 1.69447922706604, "learning_rate": 1.9985860910332765e-05, "loss": 1.1308, "step": 1328 }, { "epoch": 0.0468299902217289, "grad_norm": 9.921844482421875, "learning_rate": 1.9985800176815107e-05, "loss": 7.8183, "step": 1329 }, { "epoch": 0.046865227234687314, "grad_norm": 6.341106414794922, "learning_rate": 1.9985739313231128e-05, "loss": 4.1429, "step": 1330 }, { "epoch": 0.046900464247645726, "grad_norm": 7.829446315765381, "learning_rate": 1.998567831958162e-05, "loss": 5.7368, "step": 1331 }, { "epoch": 0.04693570126060414, "grad_norm": 5.069218635559082, "learning_rate": 1.9985617195867384e-05, "loss": 3.5158, "step": 1332 }, { "epoch": 0.04697093827356255, "grad_norm": 1.403138518333435, "learning_rate": 1.998555594208921e-05, "loss": 1.0259, "step": 1333 }, { "epoch": 0.047006175286520964, "grad_norm": 1.278314232826233, "learning_rate": 1.9985494558247896e-05, "loss": 1.0629, "step": 1334 }, { "epoch": 0.047041412299479377, "grad_norm": 1.7912631034851074, "learning_rate": 1.998543304434425e-05, "loss": 1.0115, "step": 1335 }, { "epoch": 0.04707664931243778, "grad_norm": 5.19271993637085, "learning_rate": 1.998537140037906e-05, "loss": 7.3876, "step": 1336 }, { "epoch": 0.047111886325396195, "grad_norm": 1.5113273859024048, "learning_rate": 1.998530962635314e-05, "loss": 1.0865, "step": 1337 }, { "epoch": 0.04714712333835461, "grad_norm": 4.044855117797852, "learning_rate": 1.9985247722267286e-05, "loss": 3.3769, "step": 1338 }, { "epoch": 0.04718236035131302, "grad_norm": 3.415269613265991, "learning_rate": 1.998518568812231e-05, "loss": 3.827, "step": 1339 }, { "epoch": 0.04721759736427143, "grad_norm": 1.8503497838974, "learning_rate": 1.998512352391902e-05, "loss": 1.0161, "step": 1340 }, { "epoch": 0.047252834377229845, "grad_norm": 2.0566766262054443, "learning_rate": 1.9985061229658224e-05, "loss": 0.8741, "step": 1341 }, { "epoch": 0.04728807139018825, "grad_norm": 2.9817817211151123, "learning_rate": 1.9984998805340732e-05, "loss": 6.0675, "step": 1342 }, { "epoch": 0.04732330840314666, "grad_norm": 2.086414098739624, "learning_rate": 1.9984936250967362e-05, "loss": 3.7459, "step": 1343 }, { "epoch": 0.047358545416105076, "grad_norm": 1.2107747793197632, "learning_rate": 1.9984873566538924e-05, "loss": 1.1183, "step": 1344 }, { "epoch": 0.04739378242906349, "grad_norm": 1.1699007749557495, "learning_rate": 1.9984810752056233e-05, "loss": 0.9152, "step": 1345 }, { "epoch": 0.0474290194420219, "grad_norm": 2.75024151802063, "learning_rate": 1.998474780752011e-05, "loss": 4.2405, "step": 1346 }, { "epoch": 0.047464256454980314, "grad_norm": 1.5297645330429077, "learning_rate": 1.998468473293138e-05, "loss": 0.9964, "step": 1347 }, { "epoch": 0.047499493467938726, "grad_norm": 1.8456405401229858, "learning_rate": 1.9984621528290854e-05, "loss": 3.8934, "step": 1348 }, { "epoch": 0.04753473048089713, "grad_norm": 2.128249406814575, "learning_rate": 1.9984558193599365e-05, "loss": 1.0865, "step": 1349 }, { "epoch": 0.047569967493855544, "grad_norm": 2.1887104511260986, "learning_rate": 1.9984494728857735e-05, "loss": 6.1574, "step": 1350 }, { "epoch": 0.04760520450681396, "grad_norm": 2.6483848094940186, "learning_rate": 1.9984431134066785e-05, "loss": 3.7923, "step": 1351 }, { "epoch": 0.04764044151977237, "grad_norm": 2.3087213039398193, "learning_rate": 1.9984367409227347e-05, "loss": 6.8645, "step": 1352 }, { "epoch": 0.04767567853273078, "grad_norm": 2.4022858142852783, "learning_rate": 1.9984303554340254e-05, "loss": 4.6, "step": 1353 }, { "epoch": 0.047710915545689195, "grad_norm": 2.2097020149230957, "learning_rate": 1.9984239569406333e-05, "loss": 3.8293, "step": 1354 }, { "epoch": 0.0477461525586476, "grad_norm": 2.996485948562622, "learning_rate": 1.9984175454426423e-05, "loss": 3.5965, "step": 1355 }, { "epoch": 0.04778138957160601, "grad_norm": 1.4239505529403687, "learning_rate": 1.9984111209401358e-05, "loss": 0.8431, "step": 1356 }, { "epoch": 0.047816626584564426, "grad_norm": 3.4930334091186523, "learning_rate": 1.9984046834331966e-05, "loss": 3.7227, "step": 1357 }, { "epoch": 0.04785186359752284, "grad_norm": 8.448260307312012, "learning_rate": 1.99839823292191e-05, "loss": 3.4058, "step": 1358 }, { "epoch": 0.04788710061048125, "grad_norm": 5.4283647537231445, "learning_rate": 1.9983917694063584e-05, "loss": 3.7033, "step": 1359 }, { "epoch": 0.04792233762343966, "grad_norm": 6.193516254425049, "learning_rate": 1.9983852928866273e-05, "loss": 4.0181, "step": 1360 }, { "epoch": 0.047957574636398076, "grad_norm": 1.560210108757019, "learning_rate": 1.9983788033628008e-05, "loss": 1.1975, "step": 1361 }, { "epoch": 0.04799281164935648, "grad_norm": 1.20010507106781, "learning_rate": 1.9983723008349627e-05, "loss": 1.1497, "step": 1362 }, { "epoch": 0.048028048662314894, "grad_norm": 1.1518354415893555, "learning_rate": 1.9983657853031986e-05, "loss": 0.8801, "step": 1363 }, { "epoch": 0.04806328567527331, "grad_norm": 12.530506134033203, "learning_rate": 1.998359256767593e-05, "loss": 4.8268, "step": 1364 }, { "epoch": 0.04809852268823172, "grad_norm": 1.730760097503662, "learning_rate": 1.9983527152282307e-05, "loss": 1.0576, "step": 1365 }, { "epoch": 0.04813375970119013, "grad_norm": 2.869335174560547, "learning_rate": 1.998346160685197e-05, "loss": 2.4202, "step": 1366 }, { "epoch": 0.048168996714148545, "grad_norm": 1.4024498462677002, "learning_rate": 1.9983395931385776e-05, "loss": 0.8944, "step": 1367 }, { "epoch": 0.04820423372710695, "grad_norm": 4.759669780731201, "learning_rate": 1.9983330125884577e-05, "loss": 6.3391, "step": 1368 }, { "epoch": 0.04823947074006536, "grad_norm": 1.4236637353897095, "learning_rate": 1.9983264190349236e-05, "loss": 0.9279, "step": 1369 }, { "epoch": 0.048274707753023775, "grad_norm": 2.8103182315826416, "learning_rate": 1.99831981247806e-05, "loss": 3.773, "step": 1370 }, { "epoch": 0.04830994476598219, "grad_norm": 3.6286706924438477, "learning_rate": 1.998313192917954e-05, "loss": 3.6264, "step": 1371 }, { "epoch": 0.0483451817789406, "grad_norm": 2.7004544734954834, "learning_rate": 1.9983065603546916e-05, "loss": 3.6973, "step": 1372 }, { "epoch": 0.04838041879189901, "grad_norm": 1.452043056488037, "learning_rate": 1.998299914788359e-05, "loss": 1.3286, "step": 1373 }, { "epoch": 0.048415655804857426, "grad_norm": 2.375455856323242, "learning_rate": 1.998293256219043e-05, "loss": 3.7493, "step": 1374 }, { "epoch": 0.04845089281781583, "grad_norm": 1.5020076036453247, "learning_rate": 1.99828658464683e-05, "loss": 1.0323, "step": 1375 }, { "epoch": 0.048486129830774244, "grad_norm": 3.5594680309295654, "learning_rate": 1.998279900071807e-05, "loss": 3.2342, "step": 1376 }, { "epoch": 0.048521366843732656, "grad_norm": 3.7249085903167725, "learning_rate": 1.9982732024940613e-05, "loss": 6.4395, "step": 1377 }, { "epoch": 0.04855660385669107, "grad_norm": 1.2773300409317017, "learning_rate": 1.99826649191368e-05, "loss": 1.1837, "step": 1378 }, { "epoch": 0.04859184086964948, "grad_norm": 2.7144463062286377, "learning_rate": 1.9982597683307502e-05, "loss": 4.1059, "step": 1379 }, { "epoch": 0.048627077882607894, "grad_norm": 2.4654593467712402, "learning_rate": 1.99825303174536e-05, "loss": 3.8355, "step": 1380 }, { "epoch": 0.0486623148955663, "grad_norm": 3.560061454772949, "learning_rate": 1.9982462821575968e-05, "loss": 6.4981, "step": 1381 }, { "epoch": 0.04869755190852471, "grad_norm": 2.07016658782959, "learning_rate": 1.9982395195675487e-05, "loss": 1.0254, "step": 1382 }, { "epoch": 0.048732788921483125, "grad_norm": 1.444545865058899, "learning_rate": 1.998232743975304e-05, "loss": 0.8615, "step": 1383 }, { "epoch": 0.04876802593444154, "grad_norm": 2.555506706237793, "learning_rate": 1.9982259553809502e-05, "loss": 4.3452, "step": 1384 }, { "epoch": 0.04880326294739995, "grad_norm": 1.737039566040039, "learning_rate": 1.998219153784576e-05, "loss": 1.0784, "step": 1385 }, { "epoch": 0.04883849996035836, "grad_norm": 2.742509365081787, "learning_rate": 1.9982123391862706e-05, "loss": 4.2119, "step": 1386 }, { "epoch": 0.048873736973316775, "grad_norm": 1.488976001739502, "learning_rate": 1.9982055115861222e-05, "loss": 1.3235, "step": 1387 }, { "epoch": 0.04890897398627518, "grad_norm": 3.71272611618042, "learning_rate": 1.99819867098422e-05, "loss": 7.0016, "step": 1388 }, { "epoch": 0.048944210999233594, "grad_norm": 1.4601271152496338, "learning_rate": 1.9981918173806528e-05, "loss": 0.8982, "step": 1389 }, { "epoch": 0.048979448012192006, "grad_norm": 1.7436898946762085, "learning_rate": 1.9981849507755103e-05, "loss": 1.0483, "step": 1390 }, { "epoch": 0.04901468502515042, "grad_norm": 2.5358383655548096, "learning_rate": 1.9981780711688813e-05, "loss": 3.5226, "step": 1391 }, { "epoch": 0.04904992203810883, "grad_norm": 1.6879379749298096, "learning_rate": 1.9981711785608556e-05, "loss": 1.1565, "step": 1392 }, { "epoch": 0.049085159051067244, "grad_norm": 1.114334225654602, "learning_rate": 1.9981642729515237e-05, "loss": 1.1307, "step": 1393 }, { "epoch": 0.04912039606402565, "grad_norm": 4.210112571716309, "learning_rate": 1.998157354340975e-05, "loss": 3.1706, "step": 1394 }, { "epoch": 0.04915563307698406, "grad_norm": 5.4000563621521, "learning_rate": 1.9981504227292993e-05, "loss": 6.8099, "step": 1395 }, { "epoch": 0.049190870089942475, "grad_norm": 1.2560639381408691, "learning_rate": 1.998143478116587e-05, "loss": 0.9147, "step": 1396 }, { "epoch": 0.04922610710290089, "grad_norm": 3.8686883449554443, "learning_rate": 1.998136520502929e-05, "loss": 4.1062, "step": 1397 }, { "epoch": 0.0492613441158593, "grad_norm": 3.3913533687591553, "learning_rate": 1.998129549888416e-05, "loss": 3.6127, "step": 1398 }, { "epoch": 0.04929658112881771, "grad_norm": 5.416416645050049, "learning_rate": 1.998122566273138e-05, "loss": 4.0775, "step": 1399 }, { "epoch": 0.049331818141776125, "grad_norm": 1.341647744178772, "learning_rate": 1.9981155696571867e-05, "loss": 0.8703, "step": 1400 }, { "epoch": 0.04936705515473453, "grad_norm": 1.3566430807113647, "learning_rate": 1.9981085600406528e-05, "loss": 1.2879, "step": 1401 }, { "epoch": 0.04940229216769294, "grad_norm": 6.1918625831604, "learning_rate": 1.998101537423628e-05, "loss": 4.2331, "step": 1402 }, { "epoch": 0.049437529180651356, "grad_norm": 1.5128138065338135, "learning_rate": 1.9980945018062036e-05, "loss": 1.2355, "step": 1403 }, { "epoch": 0.04947276619360977, "grad_norm": 1.5606361627578735, "learning_rate": 1.998087453188471e-05, "loss": 0.9195, "step": 1404 }, { "epoch": 0.04950800320656818, "grad_norm": 1.203833818435669, "learning_rate": 1.9980803915705217e-05, "loss": 0.9618, "step": 1405 }, { "epoch": 0.049543240219526594, "grad_norm": 6.64241361618042, "learning_rate": 1.9980733169524487e-05, "loss": 3.7402, "step": 1406 }, { "epoch": 0.049578477232485, "grad_norm": 4.541465759277344, "learning_rate": 1.9980662293343437e-05, "loss": 3.228, "step": 1407 }, { "epoch": 0.04961371424544341, "grad_norm": 4.1024017333984375, "learning_rate": 1.998059128716299e-05, "loss": 3.9425, "step": 1408 }, { "epoch": 0.049648951258401824, "grad_norm": 3.9278244972229004, "learning_rate": 1.9980520150984064e-05, "loss": 3.5396, "step": 1409 }, { "epoch": 0.04968418827136024, "grad_norm": 2.686765670776367, "learning_rate": 1.9980448884807595e-05, "loss": 3.4928, "step": 1410 }, { "epoch": 0.04971942528431865, "grad_norm": 4.111396789550781, "learning_rate": 1.998037748863451e-05, "loss": 3.7631, "step": 1411 }, { "epoch": 0.04975466229727706, "grad_norm": 2.302682399749756, "learning_rate": 1.9980305962465736e-05, "loss": 1.0074, "step": 1412 }, { "epoch": 0.049789899310235475, "grad_norm": 1.7352657318115234, "learning_rate": 1.9980234306302203e-05, "loss": 0.9468, "step": 1413 }, { "epoch": 0.04982513632319388, "grad_norm": 4.1503400802612305, "learning_rate": 1.998016252014485e-05, "loss": 4.1052, "step": 1414 }, { "epoch": 0.04986037333615229, "grad_norm": 1.6230461597442627, "learning_rate": 1.998009060399461e-05, "loss": 0.9752, "step": 1415 }, { "epoch": 0.049895610349110706, "grad_norm": 1.5834953784942627, "learning_rate": 1.9980018557852414e-05, "loss": 1.4892, "step": 1416 }, { "epoch": 0.04993084736206912, "grad_norm": 1.429458737373352, "learning_rate": 1.9979946381719207e-05, "loss": 0.9477, "step": 1417 }, { "epoch": 0.04996608437502753, "grad_norm": 3.6505258083343506, "learning_rate": 1.9979874075595926e-05, "loss": 3.5078, "step": 1418 }, { "epoch": 0.05000132138798594, "grad_norm": 1.9577893018722534, "learning_rate": 1.9979801639483516e-05, "loss": 1.0162, "step": 1419 }, { "epoch": 0.05003655840094435, "grad_norm": 3.6203036308288574, "learning_rate": 1.997972907338292e-05, "loss": 3.7069, "step": 1420 }, { "epoch": 0.05007179541390276, "grad_norm": 10.748424530029297, "learning_rate": 1.997965637729508e-05, "loss": 3.6375, "step": 1421 }, { "epoch": 0.050107032426861174, "grad_norm": 5.510067939758301, "learning_rate": 1.9979583551220942e-05, "loss": 6.1105, "step": 1422 }, { "epoch": 0.05014226943981959, "grad_norm": 1.9086250066757202, "learning_rate": 1.997951059516146e-05, "loss": 0.8738, "step": 1423 }, { "epoch": 0.050177506452778, "grad_norm": 3.3648433685302734, "learning_rate": 1.9979437509117578e-05, "loss": 4.371, "step": 1424 }, { "epoch": 0.05021274346573641, "grad_norm": 2.5076963901519775, "learning_rate": 1.9979364293090254e-05, "loss": 3.0117, "step": 1425 }, { "epoch": 0.050247980478694824, "grad_norm": 1.6692907810211182, "learning_rate": 1.997929094708044e-05, "loss": 1.2156, "step": 1426 }, { "epoch": 0.05028321749165323, "grad_norm": 3.893237590789795, "learning_rate": 1.9979217471089086e-05, "loss": 4.153, "step": 1427 }, { "epoch": 0.05031845450461164, "grad_norm": 1.7726088762283325, "learning_rate": 1.997914386511716e-05, "loss": 4.0565, "step": 1428 }, { "epoch": 0.050353691517570055, "grad_norm": 4.790618896484375, "learning_rate": 1.9979070129165607e-05, "loss": 3.9595, "step": 1429 }, { "epoch": 0.05038892853052847, "grad_norm": 2.0498948097229004, "learning_rate": 1.9978996263235396e-05, "loss": 3.3615, "step": 1430 }, { "epoch": 0.05042416554348688, "grad_norm": 3.3601574897766113, "learning_rate": 1.997892226732749e-05, "loss": 3.1965, "step": 1431 }, { "epoch": 0.05045940255644529, "grad_norm": 7.202772617340088, "learning_rate": 1.997884814144285e-05, "loss": 3.7033, "step": 1432 }, { "epoch": 0.0504946395694037, "grad_norm": 4.856992244720459, "learning_rate": 1.9978773885582442e-05, "loss": 6.4012, "step": 1433 }, { "epoch": 0.05052987658236211, "grad_norm": 4.325500011444092, "learning_rate": 1.9978699499747233e-05, "loss": 4.0352, "step": 1434 }, { "epoch": 0.050565113595320524, "grad_norm": 7.103274345397949, "learning_rate": 1.997862498393819e-05, "loss": 6.7667, "step": 1435 }, { "epoch": 0.050600350608278936, "grad_norm": 1.5456324815750122, "learning_rate": 1.9978550338156287e-05, "loss": 1.056, "step": 1436 }, { "epoch": 0.05063558762123735, "grad_norm": 1.5486890077590942, "learning_rate": 1.9978475562402494e-05, "loss": 0.8969, "step": 1437 }, { "epoch": 0.05067082463419576, "grad_norm": 1.456492304801941, "learning_rate": 1.9978400656677785e-05, "loss": 1.1118, "step": 1438 }, { "epoch": 0.050706061647154174, "grad_norm": 3.5328924655914307, "learning_rate": 1.997832562098314e-05, "loss": 3.1367, "step": 1439 }, { "epoch": 0.05074129866011258, "grad_norm": 1.8296223878860474, "learning_rate": 1.9978250455319526e-05, "loss": 1.0772, "step": 1440 }, { "epoch": 0.05077653567307099, "grad_norm": 1.212904691696167, "learning_rate": 1.9978175159687936e-05, "loss": 1.1845, "step": 1441 }, { "epoch": 0.050811772686029405, "grad_norm": 7.761581897735596, "learning_rate": 1.9978099734089342e-05, "loss": 3.2877, "step": 1442 }, { "epoch": 0.05084700969898782, "grad_norm": 2.914517641067505, "learning_rate": 1.9978024178524726e-05, "loss": 3.4444, "step": 1443 }, { "epoch": 0.05088224671194623, "grad_norm": 3.6468544006347656, "learning_rate": 1.9977948492995075e-05, "loss": 3.6116, "step": 1444 }, { "epoch": 0.05091748372490464, "grad_norm": 1.7461916208267212, "learning_rate": 1.9977872677501374e-05, "loss": 1.0649, "step": 1445 }, { "epoch": 0.05095272073786305, "grad_norm": 1.3629226684570312, "learning_rate": 1.997779673204461e-05, "loss": 1.1894, "step": 1446 }, { "epoch": 0.05098795775082146, "grad_norm": 2.7907001972198486, "learning_rate": 1.9977720656625773e-05, "loss": 1.0972, "step": 1447 }, { "epoch": 0.051023194763779874, "grad_norm": 5.160362243652344, "learning_rate": 1.9977644451245853e-05, "loss": 3.9473, "step": 1448 }, { "epoch": 0.051058431776738286, "grad_norm": 7.264512538909912, "learning_rate": 1.997756811590584e-05, "loss": 7.334, "step": 1449 }, { "epoch": 0.0510936687896967, "grad_norm": 1.366958498954773, "learning_rate": 1.9977491650606734e-05, "loss": 0.8576, "step": 1450 }, { "epoch": 0.05112890580265511, "grad_norm": 4.511459827423096, "learning_rate": 1.997741505534953e-05, "loss": 5.9099, "step": 1451 }, { "epoch": 0.051164142815613524, "grad_norm": 2.426603317260742, "learning_rate": 1.9977338330135225e-05, "loss": 4.2515, "step": 1452 }, { "epoch": 0.05119937982857193, "grad_norm": 3.808046579360962, "learning_rate": 1.9977261474964814e-05, "loss": 3.7903, "step": 1453 }, { "epoch": 0.05123461684153034, "grad_norm": 1.617801547050476, "learning_rate": 1.9977184489839304e-05, "loss": 1.1237, "step": 1454 }, { "epoch": 0.051269853854488755, "grad_norm": 1.5052493810653687, "learning_rate": 1.9977107374759693e-05, "loss": 1.0491, "step": 1455 }, { "epoch": 0.05130509086744717, "grad_norm": 1.5959696769714355, "learning_rate": 1.9977030129726986e-05, "loss": 1.0351, "step": 1456 }, { "epoch": 0.05134032788040558, "grad_norm": 6.1355881690979, "learning_rate": 1.9976952754742193e-05, "loss": 6.4455, "step": 1457 }, { "epoch": 0.05137556489336399, "grad_norm": 5.645410060882568, "learning_rate": 1.997687524980632e-05, "loss": 4.1403, "step": 1458 }, { "epoch": 0.0514108019063224, "grad_norm": 1.7434263229370117, "learning_rate": 1.9976797614920377e-05, "loss": 1.1942, "step": 1459 }, { "epoch": 0.05144603891928081, "grad_norm": 3.9929168224334717, "learning_rate": 1.997671985008537e-05, "loss": 4.1363, "step": 1460 }, { "epoch": 0.05148127593223922, "grad_norm": 2.955636739730835, "learning_rate": 1.997664195530232e-05, "loss": 3.6958, "step": 1461 }, { "epoch": 0.051516512945197636, "grad_norm": 2.533729314804077, "learning_rate": 1.997656393057224e-05, "loss": 0.8488, "step": 1462 }, { "epoch": 0.05155174995815605, "grad_norm": 1.538735032081604, "learning_rate": 1.997648577589614e-05, "loss": 1.0427, "step": 1463 }, { "epoch": 0.05158698697111446, "grad_norm": 1.1232918500900269, "learning_rate": 1.997640749127504e-05, "loss": 1.2758, "step": 1464 }, { "epoch": 0.051622223984072874, "grad_norm": 2.0890655517578125, "learning_rate": 1.9976329076709967e-05, "loss": 1.1833, "step": 1465 }, { "epoch": 0.05165746099703128, "grad_norm": 2.122364044189453, "learning_rate": 1.997625053220193e-05, "loss": 1.002, "step": 1466 }, { "epoch": 0.05169269800998969, "grad_norm": 1.1843507289886475, "learning_rate": 1.9976171857751964e-05, "loss": 0.9314, "step": 1467 }, { "epoch": 0.051727935022948104, "grad_norm": 1.5881246328353882, "learning_rate": 1.9976093053361088e-05, "loss": 0.9535, "step": 1468 }, { "epoch": 0.05176317203590652, "grad_norm": 3.609884023666382, "learning_rate": 1.9976014119030333e-05, "loss": 3.7446, "step": 1469 }, { "epoch": 0.05179840904886493, "grad_norm": 3.8787543773651123, "learning_rate": 1.9975935054760716e-05, "loss": 4.7534, "step": 1470 }, { "epoch": 0.05183364606182334, "grad_norm": 1.446648120880127, "learning_rate": 1.997585586055328e-05, "loss": 0.8601, "step": 1471 }, { "epoch": 0.05186888307478175, "grad_norm": 1.7519176006317139, "learning_rate": 1.997577653640905e-05, "loss": 0.988, "step": 1472 }, { "epoch": 0.05190412008774016, "grad_norm": 2.2555394172668457, "learning_rate": 1.9975697082329056e-05, "loss": 0.8283, "step": 1473 }, { "epoch": 0.05193935710069857, "grad_norm": 1.4585373401641846, "learning_rate": 1.9975617498314337e-05, "loss": 0.9126, "step": 1474 }, { "epoch": 0.051974594113656986, "grad_norm": 7.79910945892334, "learning_rate": 1.9975537784365933e-05, "loss": 5.4976, "step": 1475 }, { "epoch": 0.0520098311266154, "grad_norm": 1.388500452041626, "learning_rate": 1.9975457940484877e-05, "loss": 1.0455, "step": 1476 }, { "epoch": 0.05204506813957381, "grad_norm": 3.042241096496582, "learning_rate": 1.9975377966672213e-05, "loss": 3.9813, "step": 1477 }, { "epoch": 0.05208030515253222, "grad_norm": 5.503352642059326, "learning_rate": 1.9975297862928978e-05, "loss": 4.205, "step": 1478 }, { "epoch": 0.05211554216549063, "grad_norm": 2.7677934169769287, "learning_rate": 1.9975217629256217e-05, "loss": 3.1109, "step": 1479 }, { "epoch": 0.05215077917844904, "grad_norm": 4.0336995124816895, "learning_rate": 1.9975137265654976e-05, "loss": 3.5068, "step": 1480 }, { "epoch": 0.052186016191407454, "grad_norm": 3.3677706718444824, "learning_rate": 1.99750567721263e-05, "loss": 4.2133, "step": 1481 }, { "epoch": 0.05222125320436587, "grad_norm": 1.8790607452392578, "learning_rate": 1.9974976148671242e-05, "loss": 1.1195, "step": 1482 }, { "epoch": 0.05225649021732428, "grad_norm": 1.8129693269729614, "learning_rate": 1.997489539529085e-05, "loss": 1.0131, "step": 1483 }, { "epoch": 0.05229172723028269, "grad_norm": 1.715898036956787, "learning_rate": 1.997481451198617e-05, "loss": 0.9175, "step": 1484 }, { "epoch": 0.0523269642432411, "grad_norm": 4.060809135437012, "learning_rate": 1.9974733498758265e-05, "loss": 4.0034, "step": 1485 }, { "epoch": 0.05236220125619951, "grad_norm": 1.8062465190887451, "learning_rate": 1.9974652355608184e-05, "loss": 1.108, "step": 1486 }, { "epoch": 0.05239743826915792, "grad_norm": 31.259220123291016, "learning_rate": 1.9974571082536985e-05, "loss": 4.1909, "step": 1487 }, { "epoch": 0.052432675282116335, "grad_norm": 7.443804740905762, "learning_rate": 1.9974489679545728e-05, "loss": 12.3926, "step": 1488 }, { "epoch": 0.05246791229507475, "grad_norm": 3.402747869491577, "learning_rate": 1.9974408146635472e-05, "loss": 5.8267, "step": 1489 }, { "epoch": 0.05250314930803316, "grad_norm": 4.220897674560547, "learning_rate": 1.9974326483807277e-05, "loss": 3.8793, "step": 1490 }, { "epoch": 0.052538386320991566, "grad_norm": 3.7718324661254883, "learning_rate": 1.997424469106221e-05, "loss": 6.3876, "step": 1491 }, { "epoch": 0.05257362333394998, "grad_norm": 2.8721182346343994, "learning_rate": 1.997416276840134e-05, "loss": 1.5641, "step": 1492 }, { "epoch": 0.05260886034690839, "grad_norm": 5.135922908782959, "learning_rate": 1.9974080715825725e-05, "loss": 3.3635, "step": 1493 }, { "epoch": 0.052644097359866804, "grad_norm": 3.825284957885742, "learning_rate": 1.997399853333644e-05, "loss": 0.9144, "step": 1494 }, { "epoch": 0.052679334372825216, "grad_norm": 2.226349353790283, "learning_rate": 1.9973916220934554e-05, "loss": 3.516, "step": 1495 }, { "epoch": 0.05271457138578363, "grad_norm": 3.842905044555664, "learning_rate": 1.9973833778621137e-05, "loss": 6.6138, "step": 1496 }, { "epoch": 0.05274980839874204, "grad_norm": 1.7816256284713745, "learning_rate": 1.9973751206397265e-05, "loss": 0.9056, "step": 1497 }, { "epoch": 0.05278504541170045, "grad_norm": 1.7981475591659546, "learning_rate": 1.9973668504264013e-05, "loss": 0.8722, "step": 1498 }, { "epoch": 0.05282028242465886, "grad_norm": 6.824413776397705, "learning_rate": 1.997358567222246e-05, "loss": 6.1972, "step": 1499 }, { "epoch": 0.05285551943761727, "grad_norm": 6.166100025177002, "learning_rate": 1.997350271027368e-05, "loss": 3.6792, "step": 1500 }, { "epoch": 0.052890756450575685, "grad_norm": 2.263047933578491, "learning_rate": 1.9973419618418758e-05, "loss": 3.0875, "step": 1501 }, { "epoch": 0.0529259934635341, "grad_norm": 6.415526390075684, "learning_rate": 1.9973336396658776e-05, "loss": 3.2802, "step": 1502 }, { "epoch": 0.05296123047649251, "grad_norm": 2.9645941257476807, "learning_rate": 1.997325304499482e-05, "loss": 3.7857, "step": 1503 }, { "epoch": 0.052996467489450916, "grad_norm": 4.232644557952881, "learning_rate": 1.997316956342797e-05, "loss": 3.9929, "step": 1504 }, { "epoch": 0.05303170450240933, "grad_norm": 3.1728157997131348, "learning_rate": 1.9973085951959312e-05, "loss": 4.1953, "step": 1505 }, { "epoch": 0.05306694151536774, "grad_norm": 3.862518310546875, "learning_rate": 1.9973002210589943e-05, "loss": 6.7917, "step": 1506 }, { "epoch": 0.053102178528326154, "grad_norm": 2.199415683746338, "learning_rate": 1.9972918339320947e-05, "loss": 3.3985, "step": 1507 }, { "epoch": 0.053137415541284566, "grad_norm": 3.05991268157959, "learning_rate": 1.9972834338153424e-05, "loss": 4.3118, "step": 1508 }, { "epoch": 0.05317265255424298, "grad_norm": 2.112133502960205, "learning_rate": 1.997275020708846e-05, "loss": 0.9508, "step": 1509 }, { "epoch": 0.05320788956720139, "grad_norm": 1.2140849828720093, "learning_rate": 1.9972665946127156e-05, "loss": 0.8734, "step": 1510 }, { "epoch": 0.0532431265801598, "grad_norm": 9.414030075073242, "learning_rate": 1.9972581555270606e-05, "loss": 6.6463, "step": 1511 }, { "epoch": 0.05327836359311821, "grad_norm": 1.9824469089508057, "learning_rate": 1.997249703451991e-05, "loss": 0.912, "step": 1512 }, { "epoch": 0.05331360060607662, "grad_norm": 3.2840168476104736, "learning_rate": 1.9972412383876174e-05, "loss": 3.4814, "step": 1513 }, { "epoch": 0.053348837619035035, "grad_norm": 1.1798864603042603, "learning_rate": 1.9972327603340497e-05, "loss": 0.7046, "step": 1514 }, { "epoch": 0.05338407463199345, "grad_norm": 2.092924118041992, "learning_rate": 1.9972242692913978e-05, "loss": 3.3112, "step": 1515 }, { "epoch": 0.05341931164495186, "grad_norm": 1.7899194955825806, "learning_rate": 1.9972157652597728e-05, "loss": 3.9436, "step": 1516 }, { "epoch": 0.053454548657910265, "grad_norm": 1.4179010391235352, "learning_rate": 1.9972072482392858e-05, "loss": 0.9685, "step": 1517 }, { "epoch": 0.05348978567086868, "grad_norm": 1.9985733032226562, "learning_rate": 1.997198718230047e-05, "loss": 0.9491, "step": 1518 }, { "epoch": 0.05352502268382709, "grad_norm": 3.1674861907958984, "learning_rate": 1.9971901752321677e-05, "loss": 3.8341, "step": 1519 }, { "epoch": 0.0535602596967855, "grad_norm": 1.7430635690689087, "learning_rate": 1.99718161924576e-05, "loss": 0.9937, "step": 1520 }, { "epoch": 0.053595496709743916, "grad_norm": 3.5301098823547363, "learning_rate": 1.9971730502709344e-05, "loss": 6.475, "step": 1521 }, { "epoch": 0.05363073372270233, "grad_norm": 1.2735214233398438, "learning_rate": 1.9971644683078025e-05, "loss": 1.0148, "step": 1522 }, { "epoch": 0.05366597073566074, "grad_norm": 2.559929847717285, "learning_rate": 1.9971558733564768e-05, "loss": 6.4883, "step": 1523 }, { "epoch": 0.05370120774861915, "grad_norm": 1.7832788228988647, "learning_rate": 1.9971472654170688e-05, "loss": 1.0123, "step": 1524 }, { "epoch": 0.05373644476157756, "grad_norm": 1.4753726720809937, "learning_rate": 1.99713864448969e-05, "loss": 1.0487, "step": 1525 }, { "epoch": 0.05377168177453597, "grad_norm": 2.794783592224121, "learning_rate": 1.997130010574454e-05, "loss": 3.4188, "step": 1526 }, { "epoch": 0.053806918787494384, "grad_norm": 4.322618007659912, "learning_rate": 1.997121363671473e-05, "loss": 3.4597, "step": 1527 }, { "epoch": 0.0538421558004528, "grad_norm": 4.008847713470459, "learning_rate": 1.9971127037808583e-05, "loss": 6.4259, "step": 1528 }, { "epoch": 0.05387739281341121, "grad_norm": 1.9197694063186646, "learning_rate": 1.9971040309027243e-05, "loss": 3.5674, "step": 1529 }, { "epoch": 0.053912629826369615, "grad_norm": 1.7351651191711426, "learning_rate": 1.997095345037183e-05, "loss": 1.2415, "step": 1530 }, { "epoch": 0.05394786683932803, "grad_norm": 2.889105796813965, "learning_rate": 1.9970866461843477e-05, "loss": 4.0243, "step": 1531 }, { "epoch": 0.05398310385228644, "grad_norm": 2.5108752250671387, "learning_rate": 1.997077934344332e-05, "loss": 2.9814, "step": 1532 }, { "epoch": 0.05401834086524485, "grad_norm": 2.9250528812408447, "learning_rate": 1.997069209517249e-05, "loss": 5.1596, "step": 1533 }, { "epoch": 0.054053577878203266, "grad_norm": 2.7809784412384033, "learning_rate": 1.997060471703213e-05, "loss": 3.3256, "step": 1534 }, { "epoch": 0.05408881489116168, "grad_norm": 1.778684377670288, "learning_rate": 1.997051720902337e-05, "loss": 1.1197, "step": 1535 }, { "epoch": 0.05412405190412009, "grad_norm": 4.629664421081543, "learning_rate": 1.997042957114735e-05, "loss": 4.084, "step": 1536 }, { "epoch": 0.054159288917078496, "grad_norm": 2.5812220573425293, "learning_rate": 1.997034180340522e-05, "loss": 3.7197, "step": 1537 }, { "epoch": 0.05419452593003691, "grad_norm": 1.2049226760864258, "learning_rate": 1.9970253905798117e-05, "loss": 1.2008, "step": 1538 }, { "epoch": 0.05422976294299532, "grad_norm": 1.142502784729004, "learning_rate": 1.9970165878327184e-05, "loss": 0.8565, "step": 1539 }, { "epoch": 0.054264999955953734, "grad_norm": 1.2679470777511597, "learning_rate": 1.9970077720993575e-05, "loss": 1.1821, "step": 1540 }, { "epoch": 0.05430023696891215, "grad_norm": 6.591475486755371, "learning_rate": 1.996998943379843e-05, "loss": 8.1392, "step": 1541 }, { "epoch": 0.05433547398187056, "grad_norm": 2.1172993183135986, "learning_rate": 1.9969901016742903e-05, "loss": 3.6326, "step": 1542 }, { "epoch": 0.054370710994828965, "grad_norm": 1.7503728866577148, "learning_rate": 1.996981246982815e-05, "loss": 1.2529, "step": 1543 }, { "epoch": 0.05440594800778738, "grad_norm": 3.1598381996154785, "learning_rate": 1.9969723793055314e-05, "loss": 3.4831, "step": 1544 }, { "epoch": 0.05444118502074579, "grad_norm": 1.5442453622817993, "learning_rate": 1.996963498642556e-05, "loss": 1.3176, "step": 1545 }, { "epoch": 0.0544764220337042, "grad_norm": 5.563621520996094, "learning_rate": 1.9969546049940038e-05, "loss": 3.4194, "step": 1546 }, { "epoch": 0.054511659046662615, "grad_norm": 5.82470178604126, "learning_rate": 1.9969456983599905e-05, "loss": 4.1487, "step": 1547 }, { "epoch": 0.05454689605962103, "grad_norm": 4.370096206665039, "learning_rate": 1.996936778740633e-05, "loss": 2.8828, "step": 1548 }, { "epoch": 0.05458213307257944, "grad_norm": 1.8796156644821167, "learning_rate": 1.9969278461360468e-05, "loss": 1.1495, "step": 1549 }, { "epoch": 0.054617370085537846, "grad_norm": 2.1453001499176025, "learning_rate": 1.9969189005463486e-05, "loss": 3.6715, "step": 1550 }, { "epoch": 0.05465260709849626, "grad_norm": 3.324613094329834, "learning_rate": 1.9969099419716547e-05, "loss": 3.6273, "step": 1551 }, { "epoch": 0.05468784411145467, "grad_norm": 3.114622116088867, "learning_rate": 1.9969009704120816e-05, "loss": 3.5708, "step": 1552 }, { "epoch": 0.054723081124413084, "grad_norm": 1.5084682703018188, "learning_rate": 1.9968919858677465e-05, "loss": 0.906, "step": 1553 }, { "epoch": 0.054758318137371496, "grad_norm": 3.610696315765381, "learning_rate": 1.9968829883387663e-05, "loss": 3.8597, "step": 1554 }, { "epoch": 0.05479355515032991, "grad_norm": 0.9282570481300354, "learning_rate": 1.9968739778252582e-05, "loss": 0.8798, "step": 1555 }, { "epoch": 0.054828792163288315, "grad_norm": 3.4104654788970947, "learning_rate": 1.9968649543273395e-05, "loss": 5.6128, "step": 1556 }, { "epoch": 0.05486402917624673, "grad_norm": 6.62390661239624, "learning_rate": 1.9968559178451277e-05, "loss": 1.2308, "step": 1557 }, { "epoch": 0.05489926618920514, "grad_norm": 4.306822299957275, "learning_rate": 1.9968468683787407e-05, "loss": 3.3736, "step": 1558 }, { "epoch": 0.05493450320216355, "grad_norm": 2.0997841358184814, "learning_rate": 1.9968378059282962e-05, "loss": 0.9269, "step": 1559 }, { "epoch": 0.054969740215121965, "grad_norm": 5.011875629425049, "learning_rate": 1.9968287304939126e-05, "loss": 3.9323, "step": 1560 }, { "epoch": 0.05500497722808038, "grad_norm": 1.7059931755065918, "learning_rate": 1.9968196420757073e-05, "loss": 1.032, "step": 1561 }, { "epoch": 0.05504021424103879, "grad_norm": 1.340867280960083, "learning_rate": 1.9968105406737998e-05, "loss": 1.0944, "step": 1562 }, { "epoch": 0.055075451253997196, "grad_norm": 1.8169196844100952, "learning_rate": 1.9968014262883076e-05, "loss": 1.119, "step": 1563 }, { "epoch": 0.05511068826695561, "grad_norm": 1.9168155193328857, "learning_rate": 1.99679229891935e-05, "loss": 3.2562, "step": 1564 }, { "epoch": 0.05514592527991402, "grad_norm": 1.112612247467041, "learning_rate": 1.9967831585670456e-05, "loss": 0.7485, "step": 1565 }, { "epoch": 0.055181162292872434, "grad_norm": 1.6732800006866455, "learning_rate": 1.9967740052315138e-05, "loss": 0.8926, "step": 1566 }, { "epoch": 0.055216399305830846, "grad_norm": 1.8788007497787476, "learning_rate": 1.9967648389128735e-05, "loss": 1.0988, "step": 1567 }, { "epoch": 0.05525163631878926, "grad_norm": 1.2686384916305542, "learning_rate": 1.996755659611244e-05, "loss": 1.0106, "step": 1568 }, { "epoch": 0.055286873331747664, "grad_norm": 7.670873165130615, "learning_rate": 1.9967464673267455e-05, "loss": 7.0966, "step": 1569 }, { "epoch": 0.05532211034470608, "grad_norm": 3.75314998626709, "learning_rate": 1.996737262059497e-05, "loss": 3.1075, "step": 1570 }, { "epoch": 0.05535734735766449, "grad_norm": 2.143155813217163, "learning_rate": 1.9967280438096187e-05, "loss": 0.9984, "step": 1571 }, { "epoch": 0.0553925843706229, "grad_norm": 1.8306090831756592, "learning_rate": 1.9967188125772305e-05, "loss": 1.2098, "step": 1572 }, { "epoch": 0.055427821383581315, "grad_norm": 1.8356021642684937, "learning_rate": 1.9967095683624534e-05, "loss": 1.0973, "step": 1573 }, { "epoch": 0.05546305839653973, "grad_norm": 2.71346116065979, "learning_rate": 1.9967003111654066e-05, "loss": 3.4646, "step": 1574 }, { "epoch": 0.05549829540949814, "grad_norm": 1.7132538557052612, "learning_rate": 1.9966910409862117e-05, "loss": 0.8733, "step": 1575 }, { "epoch": 0.055533532422456545, "grad_norm": 2.9594602584838867, "learning_rate": 1.9966817578249887e-05, "loss": 3.3451, "step": 1576 }, { "epoch": 0.05556876943541496, "grad_norm": 1.6026034355163574, "learning_rate": 1.996672461681859e-05, "loss": 1.153, "step": 1577 }, { "epoch": 0.05560400644837337, "grad_norm": 1.5128189325332642, "learning_rate": 1.9966631525569437e-05, "loss": 0.8757, "step": 1578 }, { "epoch": 0.05563924346133178, "grad_norm": 4.2715373039245605, "learning_rate": 1.9966538304503635e-05, "loss": 8.646, "step": 1579 }, { "epoch": 0.055674480474290196, "grad_norm": 1.984825849533081, "learning_rate": 1.9966444953622407e-05, "loss": 4.3154, "step": 1580 }, { "epoch": 0.05570971748724861, "grad_norm": 4.407662868499756, "learning_rate": 1.996635147292696e-05, "loss": 3.9356, "step": 1581 }, { "epoch": 0.055744954500207014, "grad_norm": 3.4380745887756348, "learning_rate": 1.9966257862418516e-05, "loss": 5.94, "step": 1582 }, { "epoch": 0.05578019151316543, "grad_norm": 2.5462841987609863, "learning_rate": 1.9966164122098295e-05, "loss": 3.1397, "step": 1583 }, { "epoch": 0.05581542852612384, "grad_norm": 1.616684913635254, "learning_rate": 1.9966070251967514e-05, "loss": 0.8774, "step": 1584 }, { "epoch": 0.05585066553908225, "grad_norm": 1.3443328142166138, "learning_rate": 1.9965976252027402e-05, "loss": 1.1209, "step": 1585 }, { "epoch": 0.055885902552040664, "grad_norm": 2.3908333778381348, "learning_rate": 1.996588212227918e-05, "loss": 3.3989, "step": 1586 }, { "epoch": 0.05592113956499908, "grad_norm": 1.2093946933746338, "learning_rate": 1.9965787862724072e-05, "loss": 1.2488, "step": 1587 }, { "epoch": 0.05595637657795749, "grad_norm": 2.0968029499053955, "learning_rate": 1.996569347336331e-05, "loss": 1.1029, "step": 1588 }, { "epoch": 0.055991613590915895, "grad_norm": 1.9977695941925049, "learning_rate": 1.996559895419812e-05, "loss": 4.0169, "step": 1589 }, { "epoch": 0.05602685060387431, "grad_norm": 1.7641305923461914, "learning_rate": 1.9965504305229733e-05, "loss": 0.7936, "step": 1590 }, { "epoch": 0.05606208761683272, "grad_norm": 1.8750438690185547, "learning_rate": 1.9965409526459384e-05, "loss": 3.0596, "step": 1591 }, { "epoch": 0.05609732462979113, "grad_norm": 1.7088088989257812, "learning_rate": 1.9965314617888304e-05, "loss": 1.066, "step": 1592 }, { "epoch": 0.056132561642749546, "grad_norm": 3.250204563140869, "learning_rate": 1.9965219579517735e-05, "loss": 3.4108, "step": 1593 }, { "epoch": 0.05616779865570796, "grad_norm": 4.006408214569092, "learning_rate": 1.9965124411348912e-05, "loss": 4.3924, "step": 1594 }, { "epoch": 0.056203035668666364, "grad_norm": 3.045475959777832, "learning_rate": 1.9965029113383074e-05, "loss": 4.3492, "step": 1595 }, { "epoch": 0.056238272681624776, "grad_norm": 2.5164854526519775, "learning_rate": 1.996493368562146e-05, "loss": 3.411, "step": 1596 }, { "epoch": 0.05627350969458319, "grad_norm": 2.0784990787506104, "learning_rate": 1.9964838128065316e-05, "loss": 1.0661, "step": 1597 }, { "epoch": 0.0563087467075416, "grad_norm": 1.9709672927856445, "learning_rate": 1.9964742440715888e-05, "loss": 3.2139, "step": 1598 }, { "epoch": 0.056343983720500014, "grad_norm": 3.4262712001800537, "learning_rate": 1.996464662357442e-05, "loss": 3.7174, "step": 1599 }, { "epoch": 0.05637922073345843, "grad_norm": 1.5384619235992432, "learning_rate": 1.9964550676642158e-05, "loss": 2.6799, "step": 1600 }, { "epoch": 0.05641445774641684, "grad_norm": 3.985823631286621, "learning_rate": 1.9964454599920354e-05, "loss": 6.51, "step": 1601 }, { "epoch": 0.056449694759375245, "grad_norm": 1.811880350112915, "learning_rate": 1.9964358393410263e-05, "loss": 0.9515, "step": 1602 }, { "epoch": 0.05648493177233366, "grad_norm": 5.8178606033325195, "learning_rate": 1.9964262057113132e-05, "loss": 9.0482, "step": 1603 }, { "epoch": 0.05652016878529207, "grad_norm": 0.9765056371688843, "learning_rate": 1.996416559103022e-05, "loss": 1.0377, "step": 1604 }, { "epoch": 0.05655540579825048, "grad_norm": 1.1371949911117554, "learning_rate": 1.996406899516278e-05, "loss": 1.0963, "step": 1605 }, { "epoch": 0.056590642811208895, "grad_norm": 4.901075839996338, "learning_rate": 1.9963972269512074e-05, "loss": 4.3934, "step": 1606 }, { "epoch": 0.05662587982416731, "grad_norm": 4.286128520965576, "learning_rate": 1.996387541407936e-05, "loss": 6.096, "step": 1607 }, { "epoch": 0.05666111683712571, "grad_norm": 3.2071096897125244, "learning_rate": 1.99637784288659e-05, "loss": 3.7834, "step": 1608 }, { "epoch": 0.056696353850084126, "grad_norm": 1.185287594795227, "learning_rate": 1.9963681313872952e-05, "loss": 0.9255, "step": 1609 }, { "epoch": 0.05673159086304254, "grad_norm": 2.111907720565796, "learning_rate": 1.9963584069101794e-05, "loss": 2.9877, "step": 1610 }, { "epoch": 0.05676682787600095, "grad_norm": 4.458398818969727, "learning_rate": 1.996348669455368e-05, "loss": 3.8837, "step": 1611 }, { "epoch": 0.056802064888959364, "grad_norm": 1.1178818941116333, "learning_rate": 1.996338919022988e-05, "loss": 1.0211, "step": 1612 }, { "epoch": 0.056837301901917776, "grad_norm": 1.8358889818191528, "learning_rate": 1.996329155613167e-05, "loss": 1.2821, "step": 1613 }, { "epoch": 0.05687253891487619, "grad_norm": 1.247307538986206, "learning_rate": 1.9963193792260317e-05, "loss": 0.7699, "step": 1614 }, { "epoch": 0.056907775927834595, "grad_norm": 1.3421200513839722, "learning_rate": 1.9963095898617096e-05, "loss": 0.9117, "step": 1615 }, { "epoch": 0.05694301294079301, "grad_norm": 4.222758769989014, "learning_rate": 1.996299787520328e-05, "loss": 3.8675, "step": 1616 }, { "epoch": 0.05697824995375142, "grad_norm": 1.7745722532272339, "learning_rate": 1.996289972202015e-05, "loss": 1.1792, "step": 1617 }, { "epoch": 0.05701348696670983, "grad_norm": 1.0757542848587036, "learning_rate": 1.9962801439068985e-05, "loss": 0.9107, "step": 1618 }, { "epoch": 0.057048723979668245, "grad_norm": 2.766761302947998, "learning_rate": 1.996270302635106e-05, "loss": 3.391, "step": 1619 }, { "epoch": 0.05708396099262666, "grad_norm": 3.5104422569274902, "learning_rate": 1.9962604483867654e-05, "loss": 3.8678, "step": 1620 }, { "epoch": 0.05711919800558506, "grad_norm": 2.214451551437378, "learning_rate": 1.996250581162006e-05, "loss": 1.0529, "step": 1621 }, { "epoch": 0.057154435018543476, "grad_norm": 1.3344452381134033, "learning_rate": 1.9962407009609558e-05, "loss": 1.0844, "step": 1622 }, { "epoch": 0.05718967203150189, "grad_norm": 1.5934122800827026, "learning_rate": 1.9962308077837438e-05, "loss": 1.2666, "step": 1623 }, { "epoch": 0.0572249090444603, "grad_norm": 3.4578030109405518, "learning_rate": 1.9962209016304982e-05, "loss": 5.6312, "step": 1624 }, { "epoch": 0.057260146057418713, "grad_norm": 1.3289248943328857, "learning_rate": 1.996210982501349e-05, "loss": 1.0413, "step": 1625 }, { "epoch": 0.057295383070377126, "grad_norm": 3.2181854248046875, "learning_rate": 1.996201050396424e-05, "loss": 3.2702, "step": 1626 }, { "epoch": 0.05733062008333554, "grad_norm": 3.8680527210235596, "learning_rate": 1.9961911053158544e-05, "loss": 3.9355, "step": 1627 }, { "epoch": 0.057365857096293944, "grad_norm": 2.8700060844421387, "learning_rate": 1.9961811472597682e-05, "loss": 2.8833, "step": 1628 }, { "epoch": 0.05740109410925236, "grad_norm": 1.8099019527435303, "learning_rate": 1.996171176228296e-05, "loss": 0.9945, "step": 1629 }, { "epoch": 0.05743633112221077, "grad_norm": 3.6363656520843506, "learning_rate": 1.996161192221567e-05, "loss": 3.7201, "step": 1630 }, { "epoch": 0.05747156813516918, "grad_norm": 4.055882930755615, "learning_rate": 1.9961511952397116e-05, "loss": 6.2449, "step": 1631 }, { "epoch": 0.057506805148127595, "grad_norm": 5.120258331298828, "learning_rate": 1.99614118528286e-05, "loss": 5.0807, "step": 1632 }, { "epoch": 0.05754204216108601, "grad_norm": 1.9591822624206543, "learning_rate": 1.9961311623511427e-05, "loss": 1.0393, "step": 1633 }, { "epoch": 0.05757727917404441, "grad_norm": 3.3376009464263916, "learning_rate": 1.9961211264446903e-05, "loss": 3.6492, "step": 1634 }, { "epoch": 0.057612516187002825, "grad_norm": 3.700725793838501, "learning_rate": 1.996111077563633e-05, "loss": 3.7943, "step": 1635 }, { "epoch": 0.05764775319996124, "grad_norm": 1.4295363426208496, "learning_rate": 1.996101015708102e-05, "loss": 0.9745, "step": 1636 }, { "epoch": 0.05768299021291965, "grad_norm": 2.1946213245391846, "learning_rate": 1.9960909408782287e-05, "loss": 1.1038, "step": 1637 }, { "epoch": 0.05771822722587806, "grad_norm": 3.195096492767334, "learning_rate": 1.996080853074144e-05, "loss": 4.0466, "step": 1638 }, { "epoch": 0.057753464238836476, "grad_norm": 6.000408172607422, "learning_rate": 1.996070752295979e-05, "loss": 7.0369, "step": 1639 }, { "epoch": 0.05778870125179489, "grad_norm": 1.3890401124954224, "learning_rate": 1.996060638543866e-05, "loss": 1.0327, "step": 1640 }, { "epoch": 0.057823938264753294, "grad_norm": 3.6694939136505127, "learning_rate": 1.9960505118179358e-05, "loss": 3.8377, "step": 1641 }, { "epoch": 0.05785917527771171, "grad_norm": 2.2690274715423584, "learning_rate": 1.996040372118321e-05, "loss": 4.2488, "step": 1642 }, { "epoch": 0.05789441229067012, "grad_norm": 1.9106935262680054, "learning_rate": 1.9960302194451536e-05, "loss": 1.0076, "step": 1643 }, { "epoch": 0.05792964930362853, "grad_norm": 3.504237651824951, "learning_rate": 1.996020053798566e-05, "loss": 3.4594, "step": 1644 }, { "epoch": 0.057964886316586944, "grad_norm": 1.5078588724136353, "learning_rate": 1.9960098751786894e-05, "loss": 0.967, "step": 1645 }, { "epoch": 0.05800012332954536, "grad_norm": 4.146419525146484, "learning_rate": 1.9959996835856582e-05, "loss": 3.5311, "step": 1646 }, { "epoch": 0.05803536034250376, "grad_norm": 1.235123634338379, "learning_rate": 1.995989479019604e-05, "loss": 0.8357, "step": 1647 }, { "epoch": 0.058070597355462175, "grad_norm": 3.961588144302368, "learning_rate": 1.99597926148066e-05, "loss": 3.8661, "step": 1648 }, { "epoch": 0.05810583436842059, "grad_norm": 1.8084602355957031, "learning_rate": 1.995969030968959e-05, "loss": 1.1543, "step": 1649 }, { "epoch": 0.058141071381379, "grad_norm": 5.612703800201416, "learning_rate": 1.9959587874846347e-05, "loss": 4.0701, "step": 1650 }, { "epoch": 0.05817630839433741, "grad_norm": 1.3841007947921753, "learning_rate": 1.9959485310278204e-05, "loss": 0.9689, "step": 1651 }, { "epoch": 0.058211545407295825, "grad_norm": 1.239898920059204, "learning_rate": 1.9959382615986494e-05, "loss": 1.0919, "step": 1652 }, { "epoch": 0.05824678242025424, "grad_norm": 2.5789191722869873, "learning_rate": 1.995927979197256e-05, "loss": 4.13, "step": 1653 }, { "epoch": 0.058282019433212644, "grad_norm": 3.965181589126587, "learning_rate": 1.995917683823773e-05, "loss": 3.0282, "step": 1654 }, { "epoch": 0.058317256446171056, "grad_norm": 4.169526100158691, "learning_rate": 1.995907375478336e-05, "loss": 6.6243, "step": 1655 }, { "epoch": 0.05835249345912947, "grad_norm": 2.891258955001831, "learning_rate": 1.9958970541610787e-05, "loss": 3.2932, "step": 1656 }, { "epoch": 0.05838773047208788, "grad_norm": 8.49532413482666, "learning_rate": 1.9958867198721352e-05, "loss": 4.0522, "step": 1657 }, { "epoch": 0.058422967485046294, "grad_norm": 3.5612144470214844, "learning_rate": 1.9958763726116404e-05, "loss": 4.1007, "step": 1658 }, { "epoch": 0.05845820449800471, "grad_norm": 1.2468515634536743, "learning_rate": 1.995866012379729e-05, "loss": 0.794, "step": 1659 }, { "epoch": 0.05849344151096311, "grad_norm": 1.1694594621658325, "learning_rate": 1.9958556391765354e-05, "loss": 1.0035, "step": 1660 }, { "epoch": 0.058528678523921525, "grad_norm": 3.4189577102661133, "learning_rate": 1.995845253002196e-05, "loss": 3.4352, "step": 1661 }, { "epoch": 0.05856391553687994, "grad_norm": 1.4270025491714478, "learning_rate": 1.995834853856845e-05, "loss": 0.8771, "step": 1662 }, { "epoch": 0.05859915254983835, "grad_norm": 1.4330379962921143, "learning_rate": 1.9958244417406183e-05, "loss": 0.7965, "step": 1663 }, { "epoch": 0.05863438956279676, "grad_norm": 3.7286579608917236, "learning_rate": 1.9958140166536514e-05, "loss": 3.9952, "step": 1664 }, { "epoch": 0.058669626575755175, "grad_norm": 1.040845274925232, "learning_rate": 1.9958035785960798e-05, "loss": 0.9489, "step": 1665 }, { "epoch": 0.05870486358871359, "grad_norm": 1.413232445716858, "learning_rate": 1.9957931275680397e-05, "loss": 0.9741, "step": 1666 }, { "epoch": 0.05874010060167199, "grad_norm": 7.554260730743408, "learning_rate": 1.9957826635696677e-05, "loss": 4.0595, "step": 1667 }, { "epoch": 0.058775337614630406, "grad_norm": 11.57861042022705, "learning_rate": 1.9957721866010996e-05, "loss": 7.1299, "step": 1668 }, { "epoch": 0.05881057462758882, "grad_norm": 4.749962329864502, "learning_rate": 1.9957616966624717e-05, "loss": 3.586, "step": 1669 }, { "epoch": 0.05884581164054723, "grad_norm": 1.317338466644287, "learning_rate": 1.995751193753921e-05, "loss": 1.0086, "step": 1670 }, { "epoch": 0.058881048653505644, "grad_norm": 1.813899278640747, "learning_rate": 1.995740677875584e-05, "loss": 2.9734, "step": 1671 }, { "epoch": 0.058916285666464056, "grad_norm": 2.4770090579986572, "learning_rate": 1.995730149027598e-05, "loss": 1.1748, "step": 1672 }, { "epoch": 0.05895152267942246, "grad_norm": 1.2600462436676025, "learning_rate": 1.9957196072100998e-05, "loss": 1.1155, "step": 1673 }, { "epoch": 0.058986759692380875, "grad_norm": 2.705021619796753, "learning_rate": 1.995709052423227e-05, "loss": 4.0955, "step": 1674 }, { "epoch": 0.05902199670533929, "grad_norm": 1.886486530303955, "learning_rate": 1.995698484667117e-05, "loss": 0.8658, "step": 1675 }, { "epoch": 0.0590572337182977, "grad_norm": 2.4092490673065186, "learning_rate": 1.9956879039419075e-05, "loss": 3.9414, "step": 1676 }, { "epoch": 0.05909247073125611, "grad_norm": 1.782264232635498, "learning_rate": 1.995677310247736e-05, "loss": 1.1386, "step": 1677 }, { "epoch": 0.059127707744214525, "grad_norm": 2.3330838680267334, "learning_rate": 1.9956667035847408e-05, "loss": 3.4068, "step": 1678 }, { "epoch": 0.05916294475717294, "grad_norm": 1.219997763633728, "learning_rate": 1.99565608395306e-05, "loss": 0.9835, "step": 1679 }, { "epoch": 0.05919818177013134, "grad_norm": 1.5464398860931396, "learning_rate": 1.995645451352832e-05, "loss": 0.9411, "step": 1680 }, { "epoch": 0.059233418783089756, "grad_norm": 2.483795166015625, "learning_rate": 1.9956348057841947e-05, "loss": 6.5435, "step": 1681 }, { "epoch": 0.05926865579604817, "grad_norm": 2.2411601543426514, "learning_rate": 1.9956241472472877e-05, "loss": 4.0508, "step": 1682 }, { "epoch": 0.05930389280900658, "grad_norm": 1.198318600654602, "learning_rate": 1.995613475742249e-05, "loss": 0.8276, "step": 1683 }, { "epoch": 0.059339129821964993, "grad_norm": 1.7114075422286987, "learning_rate": 1.995602791269218e-05, "loss": 0.6929, "step": 1684 }, { "epoch": 0.059374366834923406, "grad_norm": 4.021466255187988, "learning_rate": 1.995592093828334e-05, "loss": 3.7592, "step": 1685 }, { "epoch": 0.05940960384788181, "grad_norm": 3.3252570629119873, "learning_rate": 1.995581383419736e-05, "loss": 4.1319, "step": 1686 }, { "epoch": 0.059444840860840224, "grad_norm": 2.0480921268463135, "learning_rate": 1.995570660043564e-05, "loss": 0.7507, "step": 1687 }, { "epoch": 0.05948007787379864, "grad_norm": 4.443076133728027, "learning_rate": 1.995559923699957e-05, "loss": 6.3679, "step": 1688 }, { "epoch": 0.05951531488675705, "grad_norm": 3.2655627727508545, "learning_rate": 1.9955491743890556e-05, "loss": 6.5443, "step": 1689 }, { "epoch": 0.05955055189971546, "grad_norm": 1.4553574323654175, "learning_rate": 1.9955384121109992e-05, "loss": 1.2015, "step": 1690 }, { "epoch": 0.059585788912673875, "grad_norm": 1.8772276639938354, "learning_rate": 1.995527636865928e-05, "loss": 1.0964, "step": 1691 }, { "epoch": 0.05962102592563229, "grad_norm": 5.2949910163879395, "learning_rate": 1.9955168486539825e-05, "loss": 6.1081, "step": 1692 }, { "epoch": 0.05965626293859069, "grad_norm": 2.2555272579193115, "learning_rate": 1.9955060474753034e-05, "loss": 1.2131, "step": 1693 }, { "epoch": 0.059691499951549105, "grad_norm": 2.3746497631073, "learning_rate": 1.995495233330031e-05, "loss": 1.0915, "step": 1694 }, { "epoch": 0.05972673696450752, "grad_norm": 2.485771417617798, "learning_rate": 1.9954844062183065e-05, "loss": 3.2378, "step": 1695 }, { "epoch": 0.05976197397746593, "grad_norm": 1.2781822681427002, "learning_rate": 1.995473566140271e-05, "loss": 1.188, "step": 1696 }, { "epoch": 0.05979721099042434, "grad_norm": 4.885021209716797, "learning_rate": 1.9954627130960653e-05, "loss": 3.7681, "step": 1697 }, { "epoch": 0.059832448003382756, "grad_norm": 1.502820611000061, "learning_rate": 1.995451847085831e-05, "loss": 1.1129, "step": 1698 }, { "epoch": 0.05986768501634116, "grad_norm": 0.9473538994789124, "learning_rate": 1.9954409681097097e-05, "loss": 1.1624, "step": 1699 }, { "epoch": 0.059902922029299574, "grad_norm": 1.512351393699646, "learning_rate": 1.9954300761678426e-05, "loss": 0.9746, "step": 1700 }, { "epoch": 0.05993815904225799, "grad_norm": 5.536828517913818, "learning_rate": 1.9954191712603724e-05, "loss": 6.9779, "step": 1701 }, { "epoch": 0.0599733960552164, "grad_norm": 2.8936617374420166, "learning_rate": 1.9954082533874408e-05, "loss": 3.6915, "step": 1702 }, { "epoch": 0.06000863306817481, "grad_norm": 3.560666561126709, "learning_rate": 1.9953973225491895e-05, "loss": 3.7759, "step": 1703 }, { "epoch": 0.060043870081133224, "grad_norm": 5.677957534790039, "learning_rate": 1.9953863787457614e-05, "loss": 9.5421, "step": 1704 }, { "epoch": 0.06007910709409164, "grad_norm": 1.7700036764144897, "learning_rate": 1.995375421977299e-05, "loss": 0.9946, "step": 1705 }, { "epoch": 0.06011434410705004, "grad_norm": 3.444171667098999, "learning_rate": 1.9953644522439448e-05, "loss": 3.8724, "step": 1706 }, { "epoch": 0.060149581120008455, "grad_norm": 4.276729106903076, "learning_rate": 1.995353469545842e-05, "loss": 6.8094, "step": 1707 }, { "epoch": 0.06018481813296687, "grad_norm": 1.1522032022476196, "learning_rate": 1.9953424738831332e-05, "loss": 1.2791, "step": 1708 }, { "epoch": 0.06022005514592528, "grad_norm": 1.2316079139709473, "learning_rate": 1.995331465255962e-05, "loss": 1.1426, "step": 1709 }, { "epoch": 0.06025529215888369, "grad_norm": 3.147228479385376, "learning_rate": 1.995320443664472e-05, "loss": 4.2156, "step": 1710 }, { "epoch": 0.060290529171842105, "grad_norm": 2.682410717010498, "learning_rate": 1.995309409108806e-05, "loss": 3.79, "step": 1711 }, { "epoch": 0.06032576618480051, "grad_norm": 5.621735572814941, "learning_rate": 1.9952983615891087e-05, "loss": 6.8255, "step": 1712 }, { "epoch": 0.060361003197758924, "grad_norm": 1.5477914810180664, "learning_rate": 1.995287301105523e-05, "loss": 1.0718, "step": 1713 }, { "epoch": 0.060396240210717336, "grad_norm": 3.7994632720947266, "learning_rate": 1.9952762276581935e-05, "loss": 3.9188, "step": 1714 }, { "epoch": 0.06043147722367575, "grad_norm": 4.717501640319824, "learning_rate": 1.9952651412472643e-05, "loss": 3.3489, "step": 1715 }, { "epoch": 0.06046671423663416, "grad_norm": 2.8753738403320312, "learning_rate": 1.99525404187288e-05, "loss": 3.8293, "step": 1716 }, { "epoch": 0.060501951249592574, "grad_norm": 3.0391485691070557, "learning_rate": 1.995242929535185e-05, "loss": 3.6411, "step": 1717 }, { "epoch": 0.06053718826255099, "grad_norm": 2.2638843059539795, "learning_rate": 1.9952318042343243e-05, "loss": 3.1484, "step": 1718 }, { "epoch": 0.06057242527550939, "grad_norm": 2.810683250427246, "learning_rate": 1.995220665970442e-05, "loss": 3.7856, "step": 1719 }, { "epoch": 0.060607662288467805, "grad_norm": 4.626611709594727, "learning_rate": 1.995209514743684e-05, "loss": 3.7126, "step": 1720 }, { "epoch": 0.06064289930142622, "grad_norm": 2.211946725845337, "learning_rate": 1.9951983505541953e-05, "loss": 4.0942, "step": 1721 }, { "epoch": 0.06067813631438463, "grad_norm": 3.3002431392669678, "learning_rate": 1.9951871734021217e-05, "loss": 3.6499, "step": 1722 }, { "epoch": 0.06071337332734304, "grad_norm": 2.8487629890441895, "learning_rate": 1.995175983287608e-05, "loss": 3.2028, "step": 1723 }, { "epoch": 0.060748610340301455, "grad_norm": 4.305534362792969, "learning_rate": 1.9951647802108006e-05, "loss": 3.9401, "step": 1724 }, { "epoch": 0.06078384735325986, "grad_norm": 2.6152307987213135, "learning_rate": 1.995153564171845e-05, "loss": 3.6659, "step": 1725 }, { "epoch": 0.06081908436621827, "grad_norm": 3.7630257606506348, "learning_rate": 1.9951423351708872e-05, "loss": 9.5199, "step": 1726 }, { "epoch": 0.060854321379176686, "grad_norm": 2.2772371768951416, "learning_rate": 1.995131093208074e-05, "loss": 1.1712, "step": 1727 }, { "epoch": 0.0608895583921351, "grad_norm": 2.0032172203063965, "learning_rate": 1.995119838283552e-05, "loss": 1.0077, "step": 1728 }, { "epoch": 0.06092479540509351, "grad_norm": 1.5462287664413452, "learning_rate": 1.9951085703974668e-05, "loss": 0.8588, "step": 1729 }, { "epoch": 0.060960032418051924, "grad_norm": 3.1405649185180664, "learning_rate": 1.9950972895499657e-05, "loss": 3.4086, "step": 1730 }, { "epoch": 0.060995269431010336, "grad_norm": 1.1798193454742432, "learning_rate": 1.995085995741196e-05, "loss": 0.8095, "step": 1731 }, { "epoch": 0.06103050644396874, "grad_norm": 1.2383872270584106, "learning_rate": 1.995074688971304e-05, "loss": 0.7599, "step": 1732 }, { "epoch": 0.061065743456927155, "grad_norm": 1.632125735282898, "learning_rate": 1.9950633692404374e-05, "loss": 0.9268, "step": 1733 }, { "epoch": 0.06110098046988557, "grad_norm": 3.6486406326293945, "learning_rate": 1.995052036548744e-05, "loss": 3.3056, "step": 1734 }, { "epoch": 0.06113621748284398, "grad_norm": 3.840303897857666, "learning_rate": 1.9950406908963715e-05, "loss": 4.1035, "step": 1735 }, { "epoch": 0.06117145449580239, "grad_norm": 1.4719020128250122, "learning_rate": 1.9950293322834663e-05, "loss": 0.9391, "step": 1736 }, { "epoch": 0.061206691508760805, "grad_norm": 5.679201602935791, "learning_rate": 1.9950179607101778e-05, "loss": 9.4743, "step": 1737 }, { "epoch": 0.06124192852171921, "grad_norm": 2.991230010986328, "learning_rate": 1.9950065761766536e-05, "loss": 3.2614, "step": 1738 }, { "epoch": 0.06127716553467762, "grad_norm": 1.5486507415771484, "learning_rate": 1.994995178683042e-05, "loss": 1.0852, "step": 1739 }, { "epoch": 0.061312402547636036, "grad_norm": 1.8474047183990479, "learning_rate": 1.9949837682294913e-05, "loss": 0.8517, "step": 1740 }, { "epoch": 0.06134763956059445, "grad_norm": 1.0524146556854248, "learning_rate": 1.9949723448161506e-05, "loss": 0.9242, "step": 1741 }, { "epoch": 0.06138287657355286, "grad_norm": 5.070295810699463, "learning_rate": 1.994960908443168e-05, "loss": 3.5548, "step": 1742 }, { "epoch": 0.06141811358651127, "grad_norm": 5.112912178039551, "learning_rate": 1.994949459110693e-05, "loss": 6.1623, "step": 1743 }, { "epoch": 0.061453350599469686, "grad_norm": 1.337015986442566, "learning_rate": 1.9949379968188744e-05, "loss": 0.8491, "step": 1744 }, { "epoch": 0.06148858761242809, "grad_norm": 8.861716270446777, "learning_rate": 1.994926521567862e-05, "loss": 3.6811, "step": 1745 }, { "epoch": 0.061523824625386504, "grad_norm": 1.164421796798706, "learning_rate": 1.994915033357805e-05, "loss": 1.0258, "step": 1746 }, { "epoch": 0.06155906163834492, "grad_norm": 2.251683235168457, "learning_rate": 1.9949035321888526e-05, "loss": 3.6891, "step": 1747 }, { "epoch": 0.06159429865130333, "grad_norm": 5.4947590827941895, "learning_rate": 1.994892018061155e-05, "loss": 4.1909, "step": 1748 }, { "epoch": 0.06162953566426174, "grad_norm": 1.3893821239471436, "learning_rate": 1.994880490974862e-05, "loss": 1.0118, "step": 1749 }, { "epoch": 0.061664772677220155, "grad_norm": 6.002038478851318, "learning_rate": 1.994868950930124e-05, "loss": 1.3625, "step": 1750 }, { "epoch": 0.06170000969017856, "grad_norm": 2.3460168838500977, "learning_rate": 1.9948573979270915e-05, "loss": 0.861, "step": 1751 }, { "epoch": 0.06173524670313697, "grad_norm": 1.621445655822754, "learning_rate": 1.9948458319659144e-05, "loss": 1.1038, "step": 1752 }, { "epoch": 0.061770483716095385, "grad_norm": 1.818789005279541, "learning_rate": 1.9948342530467436e-05, "loss": 1.051, "step": 1753 }, { "epoch": 0.0618057207290538, "grad_norm": 2.214249849319458, "learning_rate": 1.99482266116973e-05, "loss": 3.4456, "step": 1754 }, { "epoch": 0.06184095774201221, "grad_norm": 1.534448266029358, "learning_rate": 1.9948110563350246e-05, "loss": 1.1631, "step": 1755 }, { "epoch": 0.06187619475497062, "grad_norm": 4.814653396606445, "learning_rate": 1.9947994385427784e-05, "loss": 10.0754, "step": 1756 }, { "epoch": 0.061911431767929036, "grad_norm": 1.1793488264083862, "learning_rate": 1.9947878077931425e-05, "loss": 0.9576, "step": 1757 }, { "epoch": 0.06194666878088744, "grad_norm": 1.8981046676635742, "learning_rate": 1.994776164086269e-05, "loss": 0.8537, "step": 1758 }, { "epoch": 0.061981905793845854, "grad_norm": 2.749681234359741, "learning_rate": 1.9947645074223093e-05, "loss": 4.1718, "step": 1759 }, { "epoch": 0.06201714280680427, "grad_norm": 1.4410388469696045, "learning_rate": 1.994752837801415e-05, "loss": 0.971, "step": 1760 }, { "epoch": 0.06205237981976268, "grad_norm": 1.6827421188354492, "learning_rate": 1.9947411552237382e-05, "loss": 1.0649, "step": 1761 }, { "epoch": 0.06208761683272109, "grad_norm": 3.5471575260162354, "learning_rate": 1.9947294596894314e-05, "loss": 3.842, "step": 1762 }, { "epoch": 0.062122853845679504, "grad_norm": 1.1291972398757935, "learning_rate": 1.9947177511986468e-05, "loss": 0.8394, "step": 1763 }, { "epoch": 0.06215809085863791, "grad_norm": 2.6032776832580566, "learning_rate": 1.9947060297515364e-05, "loss": 3.1459, "step": 1764 }, { "epoch": 0.06219332787159632, "grad_norm": 1.657667875289917, "learning_rate": 1.9946942953482534e-05, "loss": 1.0292, "step": 1765 }, { "epoch": 0.062228564884554735, "grad_norm": 2.795748233795166, "learning_rate": 1.9946825479889503e-05, "loss": 2.7266, "step": 1766 }, { "epoch": 0.06226380189751315, "grad_norm": 2.3294215202331543, "learning_rate": 1.994670787673781e-05, "loss": 3.6873, "step": 1767 }, { "epoch": 0.06229903891047156, "grad_norm": 1.422306776046753, "learning_rate": 1.9946590144028973e-05, "loss": 0.8018, "step": 1768 }, { "epoch": 0.06233427592342997, "grad_norm": 1.140743613243103, "learning_rate": 1.9946472281764534e-05, "loss": 0.898, "step": 1769 }, { "epoch": 0.062369512936388385, "grad_norm": 1.5902818441390991, "learning_rate": 1.9946354289946026e-05, "loss": 0.9305, "step": 1770 }, { "epoch": 0.06240474994934679, "grad_norm": 5.232922077178955, "learning_rate": 1.9946236168574987e-05, "loss": 6.2699, "step": 1771 }, { "epoch": 0.062439986962305204, "grad_norm": 3.524214029312134, "learning_rate": 1.9946117917652956e-05, "loss": 3.4898, "step": 1772 }, { "epoch": 0.062475223975263616, "grad_norm": 2.848109245300293, "learning_rate": 1.9945999537181474e-05, "loss": 3.5337, "step": 1773 }, { "epoch": 0.06251046098822202, "grad_norm": 1.839686632156372, "learning_rate": 1.9945881027162076e-05, "loss": 1.231, "step": 1774 }, { "epoch": 0.06254569800118044, "grad_norm": 1.2569465637207031, "learning_rate": 1.994576238759631e-05, "loss": 1.0845, "step": 1775 }, { "epoch": 0.06258093501413885, "grad_norm": 3.3146817684173584, "learning_rate": 1.994564361848573e-05, "loss": 3.4728, "step": 1776 }, { "epoch": 0.06261617202709727, "grad_norm": 1.6779892444610596, "learning_rate": 1.994552471983187e-05, "loss": 1.3033, "step": 1777 }, { "epoch": 0.06265140904005567, "grad_norm": 1.509232759475708, "learning_rate": 1.9945405691636284e-05, "loss": 0.9768, "step": 1778 }, { "epoch": 0.06268664605301409, "grad_norm": 1.479543924331665, "learning_rate": 1.9945286533900522e-05, "loss": 1.0397, "step": 1779 }, { "epoch": 0.0627218830659725, "grad_norm": 2.444681406021118, "learning_rate": 1.9945167246626134e-05, "loss": 3.4237, "step": 1780 }, { "epoch": 0.0627571200789309, "grad_norm": 1.5234814882278442, "learning_rate": 1.994504782981468e-05, "loss": 0.9567, "step": 1781 }, { "epoch": 0.06279235709188932, "grad_norm": 2.4268600940704346, "learning_rate": 1.994492828346771e-05, "loss": 0.9684, "step": 1782 }, { "epoch": 0.06282759410484773, "grad_norm": 1.387901782989502, "learning_rate": 1.994480860758678e-05, "loss": 0.9631, "step": 1783 }, { "epoch": 0.06286283111780615, "grad_norm": 2.2749643325805664, "learning_rate": 1.994468880217345e-05, "loss": 3.3544, "step": 1784 }, { "epoch": 0.06289806813076455, "grad_norm": 4.180627822875977, "learning_rate": 1.9944568867229287e-05, "loss": 4.0246, "step": 1785 }, { "epoch": 0.06293330514372297, "grad_norm": 2.132216691970825, "learning_rate": 1.9944448802755848e-05, "loss": 1.2044, "step": 1786 }, { "epoch": 0.06296854215668138, "grad_norm": 2.6187784671783447, "learning_rate": 1.994432860875469e-05, "loss": 3.7719, "step": 1787 }, { "epoch": 0.06300377916963978, "grad_norm": 1.1784778833389282, "learning_rate": 1.994420828522739e-05, "loss": 1.3616, "step": 1788 }, { "epoch": 0.0630390161825982, "grad_norm": 1.2398521900177002, "learning_rate": 1.994408783217551e-05, "loss": 0.8049, "step": 1789 }, { "epoch": 0.06307425319555661, "grad_norm": 2.2420737743377686, "learning_rate": 1.994396724960062e-05, "loss": 3.6694, "step": 1790 }, { "epoch": 0.06310949020851503, "grad_norm": 1.272700309753418, "learning_rate": 1.994384653750429e-05, "loss": 1.0495, "step": 1791 }, { "epoch": 0.06314472722147343, "grad_norm": 2.896432399749756, "learning_rate": 1.994372569588809e-05, "loss": 3.4629, "step": 1792 }, { "epoch": 0.06317996423443185, "grad_norm": 1.1768913269042969, "learning_rate": 1.9943604724753595e-05, "loss": 1.0169, "step": 1793 }, { "epoch": 0.06321520124739026, "grad_norm": 1.9027881622314453, "learning_rate": 1.9943483624102383e-05, "loss": 0.9898, "step": 1794 }, { "epoch": 0.06325043826034867, "grad_norm": 1.0436546802520752, "learning_rate": 1.994336239393603e-05, "loss": 1.1711, "step": 1795 }, { "epoch": 0.06328567527330708, "grad_norm": 3.369183301925659, "learning_rate": 1.9943241034256114e-05, "loss": 6.9058, "step": 1796 }, { "epoch": 0.06332091228626549, "grad_norm": 3.0043327808380127, "learning_rate": 1.9943119545064222e-05, "loss": 5.6792, "step": 1797 }, { "epoch": 0.06335614929922391, "grad_norm": 1.0556942224502563, "learning_rate": 1.9942997926361927e-05, "loss": 0.9744, "step": 1798 }, { "epoch": 0.06339138631218232, "grad_norm": 1.154418706893921, "learning_rate": 1.9942876178150822e-05, "loss": 0.9226, "step": 1799 }, { "epoch": 0.06342662332514072, "grad_norm": 1.006284236907959, "learning_rate": 1.9942754300432482e-05, "loss": 0.9313, "step": 1800 }, { "epoch": 0.06346186033809914, "grad_norm": 1.4706772565841675, "learning_rate": 1.9942632293208507e-05, "loss": 0.8544, "step": 1801 }, { "epoch": 0.06349709735105755, "grad_norm": 4.210412502288818, "learning_rate": 1.9942510156480475e-05, "loss": 3.6093, "step": 1802 }, { "epoch": 0.06353233436401597, "grad_norm": 2.2150955200195312, "learning_rate": 1.9942387890249983e-05, "loss": 2.9608, "step": 1803 }, { "epoch": 0.06356757137697437, "grad_norm": 4.828554153442383, "learning_rate": 1.9942265494518622e-05, "loss": 7.0915, "step": 1804 }, { "epoch": 0.06360280838993279, "grad_norm": 4.975152492523193, "learning_rate": 1.994214296928799e-05, "loss": 2.9776, "step": 1805 }, { "epoch": 0.0636380454028912, "grad_norm": 2.5306389331817627, "learning_rate": 1.9942020314559676e-05, "loss": 4.1567, "step": 1806 }, { "epoch": 0.0636732824158496, "grad_norm": 3.7446422576904297, "learning_rate": 1.994189753033528e-05, "loss": 6.4349, "step": 1807 }, { "epoch": 0.06370851942880802, "grad_norm": 3.409446954727173, "learning_rate": 1.9941774616616406e-05, "loss": 3.4713, "step": 1808 }, { "epoch": 0.06374375644176643, "grad_norm": 2.630770683288574, "learning_rate": 1.9941651573404648e-05, "loss": 3.0353, "step": 1809 }, { "epoch": 0.06377899345472485, "grad_norm": 2.4716031551361084, "learning_rate": 1.9941528400701613e-05, "loss": 3.2572, "step": 1810 }, { "epoch": 0.06381423046768325, "grad_norm": 1.3022695779800415, "learning_rate": 1.9941405098508904e-05, "loss": 1.0363, "step": 1811 }, { "epoch": 0.06384946748064167, "grad_norm": 1.8940473794937134, "learning_rate": 1.994128166682813e-05, "loss": 3.5768, "step": 1812 }, { "epoch": 0.06388470449360008, "grad_norm": 3.2464599609375, "learning_rate": 1.9941158105660892e-05, "loss": 3.7142, "step": 1813 }, { "epoch": 0.06391994150655848, "grad_norm": 1.4082045555114746, "learning_rate": 1.9941034415008804e-05, "loss": 1.1269, "step": 1814 }, { "epoch": 0.0639551785195169, "grad_norm": 2.2719335556030273, "learning_rate": 1.9940910594873477e-05, "loss": 0.8198, "step": 1815 }, { "epoch": 0.06399041553247531, "grad_norm": 1.6784977912902832, "learning_rate": 1.9940786645256525e-05, "loss": 0.9193, "step": 1816 }, { "epoch": 0.06402565254543373, "grad_norm": 1.1853193044662476, "learning_rate": 1.994066256615956e-05, "loss": 1.3147, "step": 1817 }, { "epoch": 0.06406088955839213, "grad_norm": 5.178390979766846, "learning_rate": 1.9940538357584195e-05, "loss": 4.0342, "step": 1818 }, { "epoch": 0.06409612657135055, "grad_norm": 1.3717015981674194, "learning_rate": 1.9940414019532056e-05, "loss": 1.0127, "step": 1819 }, { "epoch": 0.06413136358430896, "grad_norm": 1.356590747833252, "learning_rate": 1.9940289552004757e-05, "loss": 1.0402, "step": 1820 }, { "epoch": 0.06416660059726736, "grad_norm": 5.195531368255615, "learning_rate": 1.9940164955003923e-05, "loss": 3.3214, "step": 1821 }, { "epoch": 0.06420183761022578, "grad_norm": 6.992108345031738, "learning_rate": 1.994004022853117e-05, "loss": 7.1956, "step": 1822 }, { "epoch": 0.06423707462318419, "grad_norm": 1.2909224033355713, "learning_rate": 1.993991537258813e-05, "loss": 1.0604, "step": 1823 }, { "epoch": 0.06427231163614261, "grad_norm": 2.933187246322632, "learning_rate": 1.9939790387176423e-05, "loss": 4.1025, "step": 1824 }, { "epoch": 0.06430754864910102, "grad_norm": 2.5838944911956787, "learning_rate": 1.993966527229768e-05, "loss": 3.5561, "step": 1825 }, { "epoch": 0.06434278566205942, "grad_norm": 1.5175273418426514, "learning_rate": 1.9939540027953535e-05, "loss": 1.0825, "step": 1826 }, { "epoch": 0.06437802267501784, "grad_norm": 5.854187965393066, "learning_rate": 1.993941465414561e-05, "loss": 6.7996, "step": 1827 }, { "epoch": 0.06441325968797625, "grad_norm": 3.2538130283355713, "learning_rate": 1.9939289150875545e-05, "loss": 3.6755, "step": 1828 }, { "epoch": 0.06444849670093467, "grad_norm": 1.1401885747909546, "learning_rate": 1.9939163518144978e-05, "loss": 1.0343, "step": 1829 }, { "epoch": 0.06448373371389307, "grad_norm": 3.4108240604400635, "learning_rate": 1.993903775595553e-05, "loss": 2.7527, "step": 1830 }, { "epoch": 0.06451897072685149, "grad_norm": 2.9963266849517822, "learning_rate": 1.9938911864308855e-05, "loss": 4.5852, "step": 1831 }, { "epoch": 0.0645542077398099, "grad_norm": 3.1355011463165283, "learning_rate": 1.993878584320659e-05, "loss": 3.682, "step": 1832 }, { "epoch": 0.0645894447527683, "grad_norm": 2.604444980621338, "learning_rate": 1.9938659692650368e-05, "loss": 0.688, "step": 1833 }, { "epoch": 0.06462468176572672, "grad_norm": 1.1706929206848145, "learning_rate": 1.993853341264184e-05, "loss": 1.2714, "step": 1834 }, { "epoch": 0.06465991877868513, "grad_norm": 3.9648454189300537, "learning_rate": 1.993840700318265e-05, "loss": 7.1679, "step": 1835 }, { "epoch": 0.06469515579164355, "grad_norm": 1.5512213706970215, "learning_rate": 1.993828046427444e-05, "loss": 0.7511, "step": 1836 }, { "epoch": 0.06473039280460195, "grad_norm": 1.239443302154541, "learning_rate": 1.9938153795918862e-05, "loss": 0.9708, "step": 1837 }, { "epoch": 0.06476562981756037, "grad_norm": 2.1914803981781006, "learning_rate": 1.9938026998117563e-05, "loss": 3.4147, "step": 1838 }, { "epoch": 0.06480086683051878, "grad_norm": 1.4007185697555542, "learning_rate": 1.99379000708722e-05, "loss": 0.9056, "step": 1839 }, { "epoch": 0.06483610384347718, "grad_norm": 4.029728412628174, "learning_rate": 1.993777301418442e-05, "loss": 3.4234, "step": 1840 }, { "epoch": 0.0648713408564356, "grad_norm": 1.782063364982605, "learning_rate": 1.993764582805588e-05, "loss": 0.9308, "step": 1841 }, { "epoch": 0.06490657786939401, "grad_norm": 3.400782823562622, "learning_rate": 1.9937518512488238e-05, "loss": 3.5278, "step": 1842 }, { "epoch": 0.06494181488235243, "grad_norm": 4.956619739532471, "learning_rate": 1.9937391067483154e-05, "loss": 3.5087, "step": 1843 }, { "epoch": 0.06497705189531083, "grad_norm": 1.2771759033203125, "learning_rate": 1.993726349304228e-05, "loss": 0.9524, "step": 1844 }, { "epoch": 0.06501228890826925, "grad_norm": 3.113342761993408, "learning_rate": 1.993713578916729e-05, "loss": 3.2968, "step": 1845 }, { "epoch": 0.06504752592122766, "grad_norm": 7.922216415405273, "learning_rate": 1.9937007955859837e-05, "loss": 3.5309, "step": 1846 }, { "epoch": 0.06508276293418606, "grad_norm": 5.2585906982421875, "learning_rate": 1.993687999312159e-05, "loss": 3.7079, "step": 1847 }, { "epoch": 0.06511799994714448, "grad_norm": 3.2300312519073486, "learning_rate": 1.9936751900954216e-05, "loss": 3.9963, "step": 1848 }, { "epoch": 0.06515323696010289, "grad_norm": 7.560552597045898, "learning_rate": 1.993662367935938e-05, "loss": 7.239, "step": 1849 }, { "epoch": 0.06518847397306131, "grad_norm": 6.9471917152404785, "learning_rate": 1.9936495328338757e-05, "loss": 3.8484, "step": 1850 }, { "epoch": 0.06522371098601971, "grad_norm": 2.3000993728637695, "learning_rate": 1.993636684789402e-05, "loss": 0.8415, "step": 1851 }, { "epoch": 0.06525894799897812, "grad_norm": 3.479459524154663, "learning_rate": 1.9936238238026834e-05, "loss": 3.7238, "step": 1852 }, { "epoch": 0.06529418501193654, "grad_norm": 9.251896858215332, "learning_rate": 1.993610949873888e-05, "loss": 3.7467, "step": 1853 }, { "epoch": 0.06532942202489495, "grad_norm": 1.4152004718780518, "learning_rate": 1.9935980630031836e-05, "loss": 0.9949, "step": 1854 }, { "epoch": 0.06536465903785336, "grad_norm": 2.953611373901367, "learning_rate": 1.9935851631907378e-05, "loss": 2.8218, "step": 1855 }, { "epoch": 0.06539989605081177, "grad_norm": 4.25880765914917, "learning_rate": 1.9935722504367184e-05, "loss": 3.4381, "step": 1856 }, { "epoch": 0.06543513306377019, "grad_norm": 1.5196808576583862, "learning_rate": 1.9935593247412944e-05, "loss": 0.9022, "step": 1857 }, { "epoch": 0.0654703700767286, "grad_norm": 1.356437087059021, "learning_rate": 1.9935463861046333e-05, "loss": 0.7914, "step": 1858 }, { "epoch": 0.065505607089687, "grad_norm": 5.335116863250732, "learning_rate": 1.993533434526904e-05, "loss": 3.2535, "step": 1859 }, { "epoch": 0.06554084410264542, "grad_norm": 5.995750427246094, "learning_rate": 1.9935204700082752e-05, "loss": 3.3224, "step": 1860 }, { "epoch": 0.06557608111560383, "grad_norm": 1.2565773725509644, "learning_rate": 1.9935074925489158e-05, "loss": 0.8442, "step": 1861 }, { "epoch": 0.06561131812856225, "grad_norm": 2.2710483074188232, "learning_rate": 1.9934945021489945e-05, "loss": 3.5031, "step": 1862 }, { "epoch": 0.06564655514152065, "grad_norm": 1.2490977048873901, "learning_rate": 1.993481498808681e-05, "loss": 1.2093, "step": 1863 }, { "epoch": 0.06568179215447907, "grad_norm": 13.23220443725586, "learning_rate": 1.9934684825281443e-05, "loss": 6.992, "step": 1864 }, { "epoch": 0.06571702916743748, "grad_norm": 1.3152415752410889, "learning_rate": 1.993455453307554e-05, "loss": 1.1661, "step": 1865 }, { "epoch": 0.06575226618039588, "grad_norm": 1.2782888412475586, "learning_rate": 1.99344241114708e-05, "loss": 0.9218, "step": 1866 }, { "epoch": 0.0657875031933543, "grad_norm": 1.307194471359253, "learning_rate": 1.9934293560468922e-05, "loss": 1.1428, "step": 1867 }, { "epoch": 0.06582274020631271, "grad_norm": 1.1211819648742676, "learning_rate": 1.9934162880071602e-05, "loss": 1.1528, "step": 1868 }, { "epoch": 0.06585797721927113, "grad_norm": 1.7574876546859741, "learning_rate": 1.993403207028055e-05, "loss": 1.0819, "step": 1869 }, { "epoch": 0.06589321423222953, "grad_norm": 3.716069221496582, "learning_rate": 1.993390113109746e-05, "loss": 5.6582, "step": 1870 }, { "epoch": 0.06592845124518795, "grad_norm": 1.78080153465271, "learning_rate": 1.9933770062524042e-05, "loss": 1.2986, "step": 1871 }, { "epoch": 0.06596368825814636, "grad_norm": 1.3606786727905273, "learning_rate": 1.9933638864562005e-05, "loss": 1.361, "step": 1872 }, { "epoch": 0.06599892527110476, "grad_norm": 3.4173426628112793, "learning_rate": 1.993350753721306e-05, "loss": 3.354, "step": 1873 }, { "epoch": 0.06603416228406318, "grad_norm": 0.9426512122154236, "learning_rate": 1.993337608047891e-05, "loss": 1.1868, "step": 1874 }, { "epoch": 0.06606939929702159, "grad_norm": 3.781050682067871, "learning_rate": 1.9933244494361274e-05, "loss": 2.9107, "step": 1875 }, { "epoch": 0.06610463630998001, "grad_norm": 1.024351716041565, "learning_rate": 1.9933112778861862e-05, "loss": 0.8572, "step": 1876 }, { "epoch": 0.06613987332293841, "grad_norm": 4.999166965484619, "learning_rate": 1.993298093398239e-05, "loss": 3.0453, "step": 1877 }, { "epoch": 0.06617511033589682, "grad_norm": 5.261509418487549, "learning_rate": 1.9932848959724573e-05, "loss": 3.6601, "step": 1878 }, { "epoch": 0.06621034734885524, "grad_norm": 1.7705305814743042, "learning_rate": 1.993271685609014e-05, "loss": 1.0568, "step": 1879 }, { "epoch": 0.06624558436181364, "grad_norm": 1.6368725299835205, "learning_rate": 1.9932584623080796e-05, "loss": 0.9776, "step": 1880 }, { "epoch": 0.06628082137477206, "grad_norm": 6.81098747253418, "learning_rate": 1.9932452260698277e-05, "loss": 3.6511, "step": 1881 }, { "epoch": 0.06631605838773047, "grad_norm": 2.8186187744140625, "learning_rate": 1.9932319768944303e-05, "loss": 0.9194, "step": 1882 }, { "epoch": 0.06635129540068889, "grad_norm": 2.086728096008301, "learning_rate": 1.99321871478206e-05, "loss": 1.0084, "step": 1883 }, { "epoch": 0.0663865324136473, "grad_norm": 5.311291694641113, "learning_rate": 1.993205439732889e-05, "loss": 3.111, "step": 1884 }, { "epoch": 0.0664217694266057, "grad_norm": 1.5084294080734253, "learning_rate": 1.993192151747091e-05, "loss": 1.0, "step": 1885 }, { "epoch": 0.06645700643956412, "grad_norm": 5.3273468017578125, "learning_rate": 1.9931788508248386e-05, "loss": 3.5568, "step": 1886 }, { "epoch": 0.06649224345252253, "grad_norm": 2.4086413383483887, "learning_rate": 1.993165536966305e-05, "loss": 0.9075, "step": 1887 }, { "epoch": 0.06652748046548095, "grad_norm": 6.113614559173584, "learning_rate": 1.993152210171664e-05, "loss": 3.4762, "step": 1888 }, { "epoch": 0.06656271747843935, "grad_norm": 5.883537292480469, "learning_rate": 1.993138870441089e-05, "loss": 3.4912, "step": 1889 }, { "epoch": 0.06659795449139777, "grad_norm": 1.0841742753982544, "learning_rate": 1.9931255177747533e-05, "loss": 1.0404, "step": 1890 }, { "epoch": 0.06663319150435618, "grad_norm": 6.551823139190674, "learning_rate": 1.9931121521728315e-05, "loss": 6.745, "step": 1891 }, { "epoch": 0.06666842851731458, "grad_norm": 1.2647004127502441, "learning_rate": 1.9930987736354974e-05, "loss": 0.8328, "step": 1892 }, { "epoch": 0.066703665530273, "grad_norm": 1.5874207019805908, "learning_rate": 1.993085382162925e-05, "loss": 1.2218, "step": 1893 }, { "epoch": 0.06673890254323141, "grad_norm": 1.3358510732650757, "learning_rate": 1.9930719777552895e-05, "loss": 1.198, "step": 1894 }, { "epoch": 0.06677413955618983, "grad_norm": 1.839536190032959, "learning_rate": 1.993058560412765e-05, "loss": 1.2779, "step": 1895 }, { "epoch": 0.06680937656914823, "grad_norm": 1.2315841913223267, "learning_rate": 1.993045130135526e-05, "loss": 0.7833, "step": 1896 }, { "epoch": 0.06684461358210665, "grad_norm": 3.168353796005249, "learning_rate": 1.9930316869237477e-05, "loss": 3.291, "step": 1897 }, { "epoch": 0.06687985059506506, "grad_norm": 2.654975175857544, "learning_rate": 1.993018230777605e-05, "loss": 3.3403, "step": 1898 }, { "epoch": 0.06691508760802346, "grad_norm": 1.4894615411758423, "learning_rate": 1.9930047616972733e-05, "loss": 0.9817, "step": 1899 }, { "epoch": 0.06695032462098188, "grad_norm": 3.956062078475952, "learning_rate": 1.9929912796829284e-05, "loss": 3.6235, "step": 1900 }, { "epoch": 0.06698556163394029, "grad_norm": 2.756962299346924, "learning_rate": 1.9929777847347455e-05, "loss": 3.8908, "step": 1901 }, { "epoch": 0.06702079864689871, "grad_norm": 1.7782062292099, "learning_rate": 1.9929642768529007e-05, "loss": 0.9989, "step": 1902 }, { "epoch": 0.06705603565985711, "grad_norm": 1.1776320934295654, "learning_rate": 1.9929507560375695e-05, "loss": 1.0713, "step": 1903 }, { "epoch": 0.06709127267281552, "grad_norm": 4.7201738357543945, "learning_rate": 1.9929372222889282e-05, "loss": 6.5091, "step": 1904 }, { "epoch": 0.06712650968577394, "grad_norm": 1.8066340684890747, "learning_rate": 1.992923675607153e-05, "loss": 1.2335, "step": 1905 }, { "epoch": 0.06716174669873234, "grad_norm": 2.6978089809417725, "learning_rate": 1.9929101159924206e-05, "loss": 3.7668, "step": 1906 }, { "epoch": 0.06719698371169076, "grad_norm": 3.6927242279052734, "learning_rate": 1.9928965434449073e-05, "loss": 3.3489, "step": 1907 }, { "epoch": 0.06723222072464917, "grad_norm": 1.9786418676376343, "learning_rate": 1.9928829579647903e-05, "loss": 4.0605, "step": 1908 }, { "epoch": 0.06726745773760759, "grad_norm": 1.3298118114471436, "learning_rate": 1.992869359552246e-05, "loss": 1.0653, "step": 1909 }, { "epoch": 0.067302694750566, "grad_norm": 2.3061859607696533, "learning_rate": 1.992855748207452e-05, "loss": 5.5725, "step": 1910 }, { "epoch": 0.0673379317635244, "grad_norm": 3.779684543609619, "learning_rate": 1.9928421239305854e-05, "loss": 3.6248, "step": 1911 }, { "epoch": 0.06737316877648282, "grad_norm": 2.6390180587768555, "learning_rate": 1.992828486721824e-05, "loss": 3.9873, "step": 1912 }, { "epoch": 0.06740840578944123, "grad_norm": 4.070684432983398, "learning_rate": 1.9928148365813446e-05, "loss": 3.5785, "step": 1913 }, { "epoch": 0.06744364280239964, "grad_norm": 1.9124493598937988, "learning_rate": 1.9928011735093257e-05, "loss": 1.0596, "step": 1914 }, { "epoch": 0.06747887981535805, "grad_norm": 2.5135385990142822, "learning_rate": 1.9927874975059448e-05, "loss": 3.3587, "step": 1915 }, { "epoch": 0.06751411682831647, "grad_norm": 1.3316420316696167, "learning_rate": 1.9927738085713808e-05, "loss": 1.1247, "step": 1916 }, { "epoch": 0.06754935384127488, "grad_norm": 1.0971728563308716, "learning_rate": 1.9927601067058113e-05, "loss": 0.9632, "step": 1917 }, { "epoch": 0.06758459085423328, "grad_norm": 1.1572411060333252, "learning_rate": 1.9927463919094147e-05, "loss": 1.2098, "step": 1918 }, { "epoch": 0.0676198278671917, "grad_norm": 2.6015853881835938, "learning_rate": 1.9927326641823697e-05, "loss": 3.4729, "step": 1919 }, { "epoch": 0.0676550648801501, "grad_norm": 1.3308830261230469, "learning_rate": 1.992718923524856e-05, "loss": 0.8362, "step": 1920 }, { "epoch": 0.06769030189310853, "grad_norm": 2.453418016433716, "learning_rate": 1.9927051699370514e-05, "loss": 3.6951, "step": 1921 }, { "epoch": 0.06772553890606693, "grad_norm": 2.9836463928222656, "learning_rate": 1.9926914034191355e-05, "loss": 3.4129, "step": 1922 }, { "epoch": 0.06776077591902535, "grad_norm": 2.256080150604248, "learning_rate": 1.9926776239712877e-05, "loss": 3.8942, "step": 1923 }, { "epoch": 0.06779601293198376, "grad_norm": 1.9367518424987793, "learning_rate": 1.9926638315936876e-05, "loss": 1.1386, "step": 1924 }, { "epoch": 0.06783124994494216, "grad_norm": 2.3915112018585205, "learning_rate": 1.992650026286514e-05, "loss": 3.4427, "step": 1925 }, { "epoch": 0.06786648695790058, "grad_norm": 2.5291407108306885, "learning_rate": 1.9926362080499483e-05, "loss": 3.556, "step": 1926 }, { "epoch": 0.06790172397085899, "grad_norm": 2.4217381477355957, "learning_rate": 1.992622376884169e-05, "loss": 4.1696, "step": 1927 }, { "epoch": 0.06793696098381741, "grad_norm": 1.4172003269195557, "learning_rate": 1.992608532789357e-05, "loss": 1.0839, "step": 1928 }, { "epoch": 0.06797219799677581, "grad_norm": 1.6651214361190796, "learning_rate": 1.992594675765692e-05, "loss": 0.9006, "step": 1929 }, { "epoch": 0.06800743500973422, "grad_norm": 1.3173060417175293, "learning_rate": 1.9925808058133555e-05, "loss": 1.1035, "step": 1930 }, { "epoch": 0.06804267202269264, "grad_norm": 5.487671852111816, "learning_rate": 1.992566922932527e-05, "loss": 4.0627, "step": 1931 }, { "epoch": 0.06807790903565104, "grad_norm": 1.920120358467102, "learning_rate": 1.9925530271233885e-05, "loss": 1.2669, "step": 1932 }, { "epoch": 0.06811314604860946, "grad_norm": 2.917015790939331, "learning_rate": 1.99253911838612e-05, "loss": 4.696, "step": 1933 }, { "epoch": 0.06814838306156787, "grad_norm": 2.1512668132781982, "learning_rate": 1.992525196720903e-05, "loss": 0.9166, "step": 1934 }, { "epoch": 0.06818362007452629, "grad_norm": 1.7290891408920288, "learning_rate": 1.992511262127919e-05, "loss": 0.9665, "step": 1935 }, { "epoch": 0.0682188570874847, "grad_norm": 1.1766407489776611, "learning_rate": 1.9924973146073494e-05, "loss": 0.8532, "step": 1936 }, { "epoch": 0.0682540941004431, "grad_norm": 1.3741146326065063, "learning_rate": 1.992483354159376e-05, "loss": 1.345, "step": 1937 }, { "epoch": 0.06828933111340152, "grad_norm": 1.1526297330856323, "learning_rate": 1.99246938078418e-05, "loss": 1.2978, "step": 1938 }, { "epoch": 0.06832456812635992, "grad_norm": 3.505357503890991, "learning_rate": 1.9924553944819446e-05, "loss": 3.7839, "step": 1939 }, { "epoch": 0.06835980513931834, "grad_norm": 1.6546964645385742, "learning_rate": 1.992441395252851e-05, "loss": 0.9805, "step": 1940 }, { "epoch": 0.06839504215227675, "grad_norm": 5.6365485191345215, "learning_rate": 1.9924273830970815e-05, "loss": 3.5631, "step": 1941 }, { "epoch": 0.06843027916523517, "grad_norm": 3.376462936401367, "learning_rate": 1.992413358014819e-05, "loss": 3.8042, "step": 1942 }, { "epoch": 0.06846551617819358, "grad_norm": 1.4921923875808716, "learning_rate": 1.9923993200062462e-05, "loss": 0.9971, "step": 1943 }, { "epoch": 0.06850075319115198, "grad_norm": 4.435703754425049, "learning_rate": 1.9923852690715463e-05, "loss": 3.6736, "step": 1944 }, { "epoch": 0.0685359902041104, "grad_norm": 2.3252663612365723, "learning_rate": 1.9923712052109013e-05, "loss": 3.5664, "step": 1945 }, { "epoch": 0.0685712272170688, "grad_norm": 1.3767213821411133, "learning_rate": 1.992357128424495e-05, "loss": 1.1153, "step": 1946 }, { "epoch": 0.06860646423002723, "grad_norm": 4.036064147949219, "learning_rate": 1.9923430387125108e-05, "loss": 4.2015, "step": 1947 }, { "epoch": 0.06864170124298563, "grad_norm": 3.9527056217193604, "learning_rate": 1.992328936075132e-05, "loss": 6.3449, "step": 1948 }, { "epoch": 0.06867693825594405, "grad_norm": 2.7260468006134033, "learning_rate": 1.9923148205125424e-05, "loss": 3.7411, "step": 1949 }, { "epoch": 0.06871217526890246, "grad_norm": 4.697734832763672, "learning_rate": 1.992300692024926e-05, "loss": 3.7241, "step": 1950 }, { "epoch": 0.06874741228186086, "grad_norm": 3.372260093688965, "learning_rate": 1.992286550612467e-05, "loss": 3.5375, "step": 1951 }, { "epoch": 0.06878264929481928, "grad_norm": 1.2843133211135864, "learning_rate": 1.992272396275349e-05, "loss": 1.0721, "step": 1952 }, { "epoch": 0.06881788630777769, "grad_norm": 2.134565830230713, "learning_rate": 1.9922582290137562e-05, "loss": 0.8332, "step": 1953 }, { "epoch": 0.0688531233207361, "grad_norm": 3.971930503845215, "learning_rate": 1.992244048827874e-05, "loss": 9.086, "step": 1954 }, { "epoch": 0.06888836033369451, "grad_norm": 4.836906433105469, "learning_rate": 1.9922298557178864e-05, "loss": 3.8216, "step": 1955 }, { "epoch": 0.06892359734665292, "grad_norm": 4.489300727844238, "learning_rate": 1.992215649683979e-05, "loss": 4.2351, "step": 1956 }, { "epoch": 0.06895883435961134, "grad_norm": 7.858764171600342, "learning_rate": 1.992201430726336e-05, "loss": 3.7834, "step": 1957 }, { "epoch": 0.06899407137256974, "grad_norm": 5.05075216293335, "learning_rate": 1.992187198845143e-05, "loss": 3.9881, "step": 1958 }, { "epoch": 0.06902930838552816, "grad_norm": 1.2513514757156372, "learning_rate": 1.992172954040586e-05, "loss": 1.0323, "step": 1959 }, { "epoch": 0.06906454539848657, "grad_norm": 7.9975104331970215, "learning_rate": 1.9921586963128493e-05, "loss": 2.8278, "step": 1960 }, { "epoch": 0.06909978241144499, "grad_norm": 5.626822471618652, "learning_rate": 1.992144425662119e-05, "loss": 4.4195, "step": 1961 }, { "epoch": 0.0691350194244034, "grad_norm": 3.6821463108062744, "learning_rate": 1.9921301420885814e-05, "loss": 3.8817, "step": 1962 }, { "epoch": 0.0691702564373618, "grad_norm": 2.060288667678833, "learning_rate": 1.9921158455924225e-05, "loss": 3.4879, "step": 1963 }, { "epoch": 0.06920549345032022, "grad_norm": 0.9273693561553955, "learning_rate": 1.992101536173828e-05, "loss": 0.782, "step": 1964 }, { "epoch": 0.06924073046327862, "grad_norm": 3.2837138175964355, "learning_rate": 1.992087213832985e-05, "loss": 3.3261, "step": 1965 }, { "epoch": 0.06927596747623704, "grad_norm": 3.936305284500122, "learning_rate": 1.9920728785700793e-05, "loss": 5.9739, "step": 1966 }, { "epoch": 0.06931120448919545, "grad_norm": 1.8743882179260254, "learning_rate": 1.992058530385298e-05, "loss": 0.838, "step": 1967 }, { "epoch": 0.06934644150215387, "grad_norm": 2.2421748638153076, "learning_rate": 1.9920441692788284e-05, "loss": 0.8236, "step": 1968 }, { "epoch": 0.06938167851511227, "grad_norm": 1.4659762382507324, "learning_rate": 1.9920297952508568e-05, "loss": 1.0707, "step": 1969 }, { "epoch": 0.06941691552807068, "grad_norm": 1.0145353078842163, "learning_rate": 1.9920154083015707e-05, "loss": 1.1662, "step": 1970 }, { "epoch": 0.0694521525410291, "grad_norm": 1.1009387969970703, "learning_rate": 1.992001008431158e-05, "loss": 0.7979, "step": 1971 }, { "epoch": 0.0694873895539875, "grad_norm": 1.1346688270568848, "learning_rate": 1.9919865956398053e-05, "loss": 1.0676, "step": 1972 }, { "epoch": 0.06952262656694592, "grad_norm": 3.405643939971924, "learning_rate": 1.991972169927701e-05, "loss": 1.1031, "step": 1973 }, { "epoch": 0.06955786357990433, "grad_norm": 3.449398994445801, "learning_rate": 1.9919577312950327e-05, "loss": 6.37, "step": 1974 }, { "epoch": 0.06959310059286274, "grad_norm": 3.0323853492736816, "learning_rate": 1.9919432797419888e-05, "loss": 5.9387, "step": 1975 }, { "epoch": 0.06962833760582116, "grad_norm": 2.123396158218384, "learning_rate": 1.9919288152687573e-05, "loss": 3.5128, "step": 1976 }, { "epoch": 0.06966357461877956, "grad_norm": 1.445634126663208, "learning_rate": 1.9919143378755263e-05, "loss": 2.786, "step": 1977 }, { "epoch": 0.06969881163173798, "grad_norm": 1.2576732635498047, "learning_rate": 1.9918998475624854e-05, "loss": 0.7703, "step": 1978 }, { "epoch": 0.06973404864469639, "grad_norm": 1.0633530616760254, "learning_rate": 1.991885344329822e-05, "loss": 1.1208, "step": 1979 }, { "epoch": 0.0697692856576548, "grad_norm": 4.09548807144165, "learning_rate": 1.9918708281777258e-05, "loss": 4.0726, "step": 1980 }, { "epoch": 0.06980452267061321, "grad_norm": 2.170320510864258, "learning_rate": 1.9918562991063858e-05, "loss": 1.1188, "step": 1981 }, { "epoch": 0.06983975968357162, "grad_norm": 2.3375935554504395, "learning_rate": 1.9918417571159912e-05, "loss": 3.9206, "step": 1982 }, { "epoch": 0.06987499669653004, "grad_norm": 3.287882089614868, "learning_rate": 1.9918272022067314e-05, "loss": 3.6774, "step": 1983 }, { "epoch": 0.06991023370948844, "grad_norm": 1.4823212623596191, "learning_rate": 1.991812634378796e-05, "loss": 0.9794, "step": 1984 }, { "epoch": 0.06994547072244686, "grad_norm": 1.7326412200927734, "learning_rate": 1.9917980536323746e-05, "loss": 0.9009, "step": 1985 }, { "epoch": 0.06998070773540527, "grad_norm": 1.5904275178909302, "learning_rate": 1.991783459967657e-05, "loss": 0.7149, "step": 1986 }, { "epoch": 0.07001594474836369, "grad_norm": 2.7310330867767334, "learning_rate": 1.9917688533848337e-05, "loss": 3.6455, "step": 1987 }, { "epoch": 0.07005118176132209, "grad_norm": 2.715998888015747, "learning_rate": 1.9917542338840947e-05, "loss": 3.105, "step": 1988 }, { "epoch": 0.0700864187742805, "grad_norm": 1.4816638231277466, "learning_rate": 1.9917396014656302e-05, "loss": 0.8736, "step": 1989 }, { "epoch": 0.07012165578723892, "grad_norm": 3.120732307434082, "learning_rate": 1.9917249561296317e-05, "loss": 3.0169, "step": 1990 }, { "epoch": 0.07015689280019732, "grad_norm": 1.7484415769577026, "learning_rate": 1.991710297876289e-05, "loss": 1.0611, "step": 1991 }, { "epoch": 0.07019212981315574, "grad_norm": 1.9766581058502197, "learning_rate": 1.9916956267057934e-05, "loss": 1.0976, "step": 1992 }, { "epoch": 0.07022736682611415, "grad_norm": 1.2971687316894531, "learning_rate": 1.9916809426183357e-05, "loss": 1.2405, "step": 1993 }, { "epoch": 0.07026260383907257, "grad_norm": 1.7312698364257812, "learning_rate": 1.9916662456141077e-05, "loss": 1.1899, "step": 1994 }, { "epoch": 0.07029784085203097, "grad_norm": 3.57342267036438, "learning_rate": 1.9916515356933002e-05, "loss": 5.2717, "step": 1995 }, { "epoch": 0.07033307786498938, "grad_norm": 7.370770454406738, "learning_rate": 1.9916368128561055e-05, "loss": 3.9619, "step": 1996 }, { "epoch": 0.0703683148779478, "grad_norm": 1.7140642404556274, "learning_rate": 1.991622077102715e-05, "loss": 0.9107, "step": 1997 }, { "epoch": 0.0704035518909062, "grad_norm": 9.024073600769043, "learning_rate": 1.9916073284333202e-05, "loss": 3.7836, "step": 1998 }, { "epoch": 0.07043878890386462, "grad_norm": 6.236386299133301, "learning_rate": 1.9915925668481138e-05, "loss": 4.0055, "step": 1999 }, { "epoch": 0.07047402591682303, "grad_norm": 1.6480985879898071, "learning_rate": 1.9915777923472884e-05, "loss": 1.0339, "step": 2000 }, { "epoch": 0.07050926292978144, "grad_norm": 1.7422772645950317, "learning_rate": 1.9915630049310353e-05, "loss": 1.1269, "step": 2001 }, { "epoch": 0.07054449994273986, "grad_norm": 1.7518936395645142, "learning_rate": 1.991548204599548e-05, "loss": 0.9747, "step": 2002 }, { "epoch": 0.07057973695569826, "grad_norm": 5.98264741897583, "learning_rate": 1.991533391353019e-05, "loss": 3.6357, "step": 2003 }, { "epoch": 0.07061497396865668, "grad_norm": 4.412766456604004, "learning_rate": 1.9915185651916415e-05, "loss": 3.5864, "step": 2004 }, { "epoch": 0.07065021098161509, "grad_norm": 2.8363101482391357, "learning_rate": 1.9915037261156083e-05, "loss": 6.5646, "step": 2005 }, { "epoch": 0.0706854479945735, "grad_norm": 1.1076875925064087, "learning_rate": 1.9914888741251128e-05, "loss": 0.9836, "step": 2006 }, { "epoch": 0.07072068500753191, "grad_norm": 2.4045467376708984, "learning_rate": 1.991474009220348e-05, "loss": 3.4086, "step": 2007 }, { "epoch": 0.07075592202049032, "grad_norm": 2.6246702671051025, "learning_rate": 1.9914591314015083e-05, "loss": 4.433, "step": 2008 }, { "epoch": 0.07079115903344874, "grad_norm": 1.6833581924438477, "learning_rate": 1.9914442406687872e-05, "loss": 1.1352, "step": 2009 }, { "epoch": 0.07082639604640714, "grad_norm": 1.1496182680130005, "learning_rate": 1.9914293370223784e-05, "loss": 0.8536, "step": 2010 }, { "epoch": 0.07086163305936556, "grad_norm": 2.1075427532196045, "learning_rate": 1.9914144204624762e-05, "loss": 3.7289, "step": 2011 }, { "epoch": 0.07089687007232397, "grad_norm": 1.7667484283447266, "learning_rate": 1.991399490989275e-05, "loss": 3.7191, "step": 2012 }, { "epoch": 0.07093210708528239, "grad_norm": 1.6607238054275513, "learning_rate": 1.991384548602969e-05, "loss": 3.0163, "step": 2013 }, { "epoch": 0.07096734409824079, "grad_norm": 2.0607473850250244, "learning_rate": 1.991369593303753e-05, "loss": 1.0568, "step": 2014 }, { "epoch": 0.0710025811111992, "grad_norm": 1.190890908241272, "learning_rate": 1.9913546250918217e-05, "loss": 0.8804, "step": 2015 }, { "epoch": 0.07103781812415762, "grad_norm": 1.589064359664917, "learning_rate": 1.9913396439673706e-05, "loss": 0.9587, "step": 2016 }, { "epoch": 0.07107305513711602, "grad_norm": 1.8128231763839722, "learning_rate": 1.991324649930594e-05, "loss": 0.888, "step": 2017 }, { "epoch": 0.07110829215007444, "grad_norm": 1.1240171194076538, "learning_rate": 1.9913096429816874e-05, "loss": 0.9158, "step": 2018 }, { "epoch": 0.07114352916303285, "grad_norm": 4.169102191925049, "learning_rate": 1.9912946231208464e-05, "loss": 8.782, "step": 2019 }, { "epoch": 0.07117876617599127, "grad_norm": 2.204944610595703, "learning_rate": 1.9912795903482672e-05, "loss": 3.4082, "step": 2020 }, { "epoch": 0.07121400318894967, "grad_norm": 4.410640239715576, "learning_rate": 1.991264544664145e-05, "loss": 3.8334, "step": 2021 }, { "epoch": 0.07124924020190808, "grad_norm": 3.446950674057007, "learning_rate": 1.9912494860686752e-05, "loss": 3.4411, "step": 2022 }, { "epoch": 0.0712844772148665, "grad_norm": 1.2684147357940674, "learning_rate": 1.9912344145620552e-05, "loss": 1.0665, "step": 2023 }, { "epoch": 0.0713197142278249, "grad_norm": 6.9780731201171875, "learning_rate": 1.99121933014448e-05, "loss": 3.2001, "step": 2024 }, { "epoch": 0.07135495124078332, "grad_norm": 3.0276429653167725, "learning_rate": 1.9912042328161472e-05, "loss": 3.9262, "step": 2025 }, { "epoch": 0.07139018825374173, "grad_norm": 3.710993766784668, "learning_rate": 1.9911891225772533e-05, "loss": 3.3414, "step": 2026 }, { "epoch": 0.07142542526670013, "grad_norm": 1.717105507850647, "learning_rate": 1.9911739994279943e-05, "loss": 1.0597, "step": 2027 }, { "epoch": 0.07146066227965855, "grad_norm": 8.350849151611328, "learning_rate": 1.9911588633685678e-05, "loss": 5.7955, "step": 2028 }, { "epoch": 0.07149589929261696, "grad_norm": 0.9940709471702576, "learning_rate": 1.9911437143991713e-05, "loss": 0.9899, "step": 2029 }, { "epoch": 0.07153113630557538, "grad_norm": 3.312065601348877, "learning_rate": 1.9911285525200013e-05, "loss": 3.441, "step": 2030 }, { "epoch": 0.07156637331853379, "grad_norm": 1.7567380666732788, "learning_rate": 1.9911133777312558e-05, "loss": 0.9226, "step": 2031 }, { "epoch": 0.0716016103314922, "grad_norm": 1.0692275762557983, "learning_rate": 1.991098190033132e-05, "loss": 1.0237, "step": 2032 }, { "epoch": 0.07163684734445061, "grad_norm": 5.484818935394287, "learning_rate": 1.9910829894258282e-05, "loss": 3.512, "step": 2033 }, { "epoch": 0.07167208435740902, "grad_norm": 1.6329176425933838, "learning_rate": 1.9910677759095423e-05, "loss": 0.9561, "step": 2034 }, { "epoch": 0.07170732137036744, "grad_norm": 2.6540935039520264, "learning_rate": 1.9910525494844722e-05, "loss": 3.0503, "step": 2035 }, { "epoch": 0.07174255838332584, "grad_norm": 1.8318865299224854, "learning_rate": 1.9910373101508165e-05, "loss": 1.1425, "step": 2036 }, { "epoch": 0.07177779539628426, "grad_norm": 3.1627907752990723, "learning_rate": 1.9910220579087738e-05, "loss": 3.2577, "step": 2037 }, { "epoch": 0.07181303240924267, "grad_norm": 1.3241493701934814, "learning_rate": 1.9910067927585422e-05, "loss": 1.0696, "step": 2038 }, { "epoch": 0.07184826942220109, "grad_norm": 1.3241493701934814, "learning_rate": 1.9910067927585422e-05, "loss": 6.9585, "step": 2039 }, { "epoch": 0.07188350643515949, "grad_norm": 1.1974210739135742, "learning_rate": 1.9909915147003213e-05, "loss": 0.8108, "step": 2040 }, { "epoch": 0.0719187434481179, "grad_norm": 4.27716588973999, "learning_rate": 1.9909762237343094e-05, "loss": 6.5397, "step": 2041 }, { "epoch": 0.07195398046107632, "grad_norm": 6.246834754943848, "learning_rate": 1.9909609198607062e-05, "loss": 3.8729, "step": 2042 }, { "epoch": 0.07198921747403472, "grad_norm": 5.460481643676758, "learning_rate": 1.9909456030797104e-05, "loss": 4.2235, "step": 2043 }, { "epoch": 0.07202445448699314, "grad_norm": 4.8285369873046875, "learning_rate": 1.9909302733915222e-05, "loss": 4.7868, "step": 2044 }, { "epoch": 0.07205969149995155, "grad_norm": 2.379755973815918, "learning_rate": 1.990914930796341e-05, "loss": 3.8111, "step": 2045 }, { "epoch": 0.07209492851290997, "grad_norm": 1.6899808645248413, "learning_rate": 1.9908995752943662e-05, "loss": 0.9474, "step": 2046 }, { "epoch": 0.07213016552586837, "grad_norm": 6.243078231811523, "learning_rate": 1.9908842068857985e-05, "loss": 3.5236, "step": 2047 }, { "epoch": 0.07216540253882678, "grad_norm": 3.0289697647094727, "learning_rate": 1.990868825570838e-05, "loss": 3.9526, "step": 2048 }, { "epoch": 0.0722006395517852, "grad_norm": 4.220940589904785, "learning_rate": 1.9908534313496845e-05, "loss": 3.535, "step": 2049 }, { "epoch": 0.0722358765647436, "grad_norm": 1.368815302848816, "learning_rate": 1.9908380242225388e-05, "loss": 0.7404, "step": 2050 }, { "epoch": 0.07227111357770202, "grad_norm": 4.991447448730469, "learning_rate": 1.9908226041896016e-05, "loss": 6.7358, "step": 2051 }, { "epoch": 0.07230635059066043, "grad_norm": 2.7960338592529297, "learning_rate": 1.9908071712510743e-05, "loss": 3.4287, "step": 2052 }, { "epoch": 0.07234158760361883, "grad_norm": 1.1539868116378784, "learning_rate": 1.990791725407157e-05, "loss": 1.0562, "step": 2053 }, { "epoch": 0.07237682461657725, "grad_norm": 4.668356418609619, "learning_rate": 1.990776266658051e-05, "loss": 4.0071, "step": 2054 }, { "epoch": 0.07241206162953566, "grad_norm": 5.5357232093811035, "learning_rate": 1.9907607950039585e-05, "loss": 8.8007, "step": 2055 }, { "epoch": 0.07244729864249408, "grad_norm": 1.0903888940811157, "learning_rate": 1.99074531044508e-05, "loss": 1.0137, "step": 2056 }, { "epoch": 0.07248253565545248, "grad_norm": 1.3728199005126953, "learning_rate": 1.9907298129816178e-05, "loss": 0.9287, "step": 2057 }, { "epoch": 0.0725177726684109, "grad_norm": 1.0576874017715454, "learning_rate": 1.9907143026137735e-05, "loss": 0.9175, "step": 2058 }, { "epoch": 0.07255300968136931, "grad_norm": 1.3604929447174072, "learning_rate": 1.9906987793417495e-05, "loss": 0.9872, "step": 2059 }, { "epoch": 0.07258824669432772, "grad_norm": 3.340719699859619, "learning_rate": 1.9906832431657475e-05, "loss": 3.4072, "step": 2060 }, { "epoch": 0.07262348370728614, "grad_norm": 4.801457405090332, "learning_rate": 1.99066769408597e-05, "loss": 5.8459, "step": 2061 }, { "epoch": 0.07265872072024454, "grad_norm": 5.639570713043213, "learning_rate": 1.9906521321026196e-05, "loss": 3.7554, "step": 2062 }, { "epoch": 0.07269395773320296, "grad_norm": 1.447880744934082, "learning_rate": 1.9906365572158993e-05, "loss": 0.8204, "step": 2063 }, { "epoch": 0.07272919474616137, "grad_norm": 3.860522985458374, "learning_rate": 1.990620969426011e-05, "loss": 3.75, "step": 2064 }, { "epoch": 0.07276443175911979, "grad_norm": 3.2368226051330566, "learning_rate": 1.9906053687331587e-05, "loss": 4.412, "step": 2065 }, { "epoch": 0.07279966877207819, "grad_norm": 6.360616207122803, "learning_rate": 1.9905897551375454e-05, "loss": 3.8787, "step": 2066 }, { "epoch": 0.0728349057850366, "grad_norm": 4.437685489654541, "learning_rate": 1.9905741286393746e-05, "loss": 4.4805, "step": 2067 }, { "epoch": 0.07287014279799502, "grad_norm": 2.702805519104004, "learning_rate": 1.9905584892388493e-05, "loss": 2.8268, "step": 2068 }, { "epoch": 0.07290537981095342, "grad_norm": 2.101353645324707, "learning_rate": 1.9905428369361735e-05, "loss": 0.9274, "step": 2069 }, { "epoch": 0.07294061682391184, "grad_norm": 2.8727898597717285, "learning_rate": 1.990527171731551e-05, "loss": 3.3838, "step": 2070 }, { "epoch": 0.07297585383687025, "grad_norm": 4.873418807983398, "learning_rate": 1.990511493625186e-05, "loss": 4.4245, "step": 2071 }, { "epoch": 0.07301109084982867, "grad_norm": 2.501396656036377, "learning_rate": 1.990495802617283e-05, "loss": 4.0511, "step": 2072 }, { "epoch": 0.07304632786278707, "grad_norm": 3.6263813972473145, "learning_rate": 1.9904800987080453e-05, "loss": 5.7644, "step": 2073 }, { "epoch": 0.07308156487574548, "grad_norm": 2.9934451580047607, "learning_rate": 1.9904643818976787e-05, "loss": 5.8214, "step": 2074 }, { "epoch": 0.0731168018887039, "grad_norm": 1.6250206232070923, "learning_rate": 1.990448652186387e-05, "loss": 0.7243, "step": 2075 }, { "epoch": 0.0731520389016623, "grad_norm": 1.6309001445770264, "learning_rate": 1.9904329095743757e-05, "loss": 4.1912, "step": 2076 }, { "epoch": 0.07318727591462072, "grad_norm": 2.7599172592163086, "learning_rate": 1.9904171540618496e-05, "loss": 4.3839, "step": 2077 }, { "epoch": 0.07322251292757913, "grad_norm": 1.8068878650665283, "learning_rate": 1.9904013856490138e-05, "loss": 3.5526, "step": 2078 }, { "epoch": 0.07325774994053753, "grad_norm": 3.716113328933716, "learning_rate": 1.9903856043360735e-05, "loss": 5.9648, "step": 2079 }, { "epoch": 0.07329298695349595, "grad_norm": 1.5005320310592651, "learning_rate": 1.9903698101232348e-05, "loss": 1.1359, "step": 2080 }, { "epoch": 0.07332822396645436, "grad_norm": 1.9660619497299194, "learning_rate": 1.9903540030107033e-05, "loss": 3.8916, "step": 2081 }, { "epoch": 0.07336346097941278, "grad_norm": 1.3578081130981445, "learning_rate": 1.9903381829986848e-05, "loss": 1.0738, "step": 2082 }, { "epoch": 0.07339869799237118, "grad_norm": 1.4028428792953491, "learning_rate": 1.990322350087385e-05, "loss": 0.9987, "step": 2083 }, { "epoch": 0.0734339350053296, "grad_norm": 1.8947674036026, "learning_rate": 1.990306504277011e-05, "loss": 1.2432, "step": 2084 }, { "epoch": 0.07346917201828801, "grad_norm": 1.7734878063201904, "learning_rate": 1.990290645567768e-05, "loss": 1.1895, "step": 2085 }, { "epoch": 0.07350440903124641, "grad_norm": 0.931086003780365, "learning_rate": 1.9902747739598633e-05, "loss": 1.1029, "step": 2086 }, { "epoch": 0.07353964604420483, "grad_norm": 2.909633159637451, "learning_rate": 1.9902588894535038e-05, "loss": 6.1978, "step": 2087 }, { "epoch": 0.07357488305716324, "grad_norm": 1.0876342058181763, "learning_rate": 1.990242992048896e-05, "loss": 0.8299, "step": 2088 }, { "epoch": 0.07361012007012166, "grad_norm": 1.6656793355941772, "learning_rate": 1.9902270817462474e-05, "loss": 0.9236, "step": 2089 }, { "epoch": 0.07364535708308007, "grad_norm": 1.616416096687317, "learning_rate": 1.9902111585457644e-05, "loss": 0.8385, "step": 2090 }, { "epoch": 0.07368059409603848, "grad_norm": 1.0388216972351074, "learning_rate": 1.990195222447655e-05, "loss": 0.9465, "step": 2091 }, { "epoch": 0.07371583110899689, "grad_norm": 1.4301042556762695, "learning_rate": 1.990179273452127e-05, "loss": 1.1937, "step": 2092 }, { "epoch": 0.0737510681219553, "grad_norm": 3.926318407058716, "learning_rate": 1.9901633115593878e-05, "loss": 4.0961, "step": 2093 }, { "epoch": 0.07378630513491372, "grad_norm": 7.692649841308594, "learning_rate": 1.9901473367696452e-05, "loss": 6.6369, "step": 2094 }, { "epoch": 0.07382154214787212, "grad_norm": 8.663677215576172, "learning_rate": 1.9901313490831074e-05, "loss": 6.91, "step": 2095 }, { "epoch": 0.07385677916083054, "grad_norm": 1.9703240394592285, "learning_rate": 1.9901153484999827e-05, "loss": 1.1506, "step": 2096 }, { "epoch": 0.07389201617378895, "grad_norm": 3.2678186893463135, "learning_rate": 1.9900993350204795e-05, "loss": 3.7904, "step": 2097 }, { "epoch": 0.07392725318674737, "grad_norm": 1.5074586868286133, "learning_rate": 1.9900833086448062e-05, "loss": 1.1059, "step": 2098 }, { "epoch": 0.07396249019970577, "grad_norm": 1.1730417013168335, "learning_rate": 1.9900672693731718e-05, "loss": 1.1963, "step": 2099 }, { "epoch": 0.07399772721266418, "grad_norm": 1.719745397567749, "learning_rate": 1.9900512172057853e-05, "loss": 3.4027, "step": 2100 }, { "epoch": 0.0740329642256226, "grad_norm": 1.6985406875610352, "learning_rate": 1.9900351521428552e-05, "loss": 0.8299, "step": 2101 }, { "epoch": 0.074068201238581, "grad_norm": 3.355876922607422, "learning_rate": 1.990019074184591e-05, "loss": 10.0401, "step": 2102 }, { "epoch": 0.07410343825153942, "grad_norm": 1.1314924955368042, "learning_rate": 1.990002983331203e-05, "loss": 0.9855, "step": 2103 }, { "epoch": 0.07413867526449783, "grad_norm": 1.3975259065628052, "learning_rate": 1.9899868795828997e-05, "loss": 1.1833, "step": 2104 }, { "epoch": 0.07417391227745623, "grad_norm": 1.5312674045562744, "learning_rate": 1.989970762939891e-05, "loss": 1.0814, "step": 2105 }, { "epoch": 0.07420914929041465, "grad_norm": 1.677883267402649, "learning_rate": 1.9899546334023872e-05, "loss": 4.2219, "step": 2106 }, { "epoch": 0.07424438630337306, "grad_norm": 1.0641990900039673, "learning_rate": 1.989938490970598e-05, "loss": 1.2182, "step": 2107 }, { "epoch": 0.07427962331633148, "grad_norm": 1.3601813316345215, "learning_rate": 1.9899223356447338e-05, "loss": 1.1085, "step": 2108 }, { "epoch": 0.07431486032928988, "grad_norm": 1.6464027166366577, "learning_rate": 1.9899061674250055e-05, "loss": 0.8591, "step": 2109 }, { "epoch": 0.0743500973422483, "grad_norm": 0.9245957732200623, "learning_rate": 1.989889986311623e-05, "loss": 1.1185, "step": 2110 }, { "epoch": 0.07438533435520671, "grad_norm": 1.9501608610153198, "learning_rate": 1.9898737923047972e-05, "loss": 0.8911, "step": 2111 }, { "epoch": 0.07442057136816511, "grad_norm": 1.6711955070495605, "learning_rate": 1.9898575854047394e-05, "loss": 3.2157, "step": 2112 }, { "epoch": 0.07445580838112353, "grad_norm": 2.639782190322876, "learning_rate": 1.9898413656116605e-05, "loss": 6.4219, "step": 2113 }, { "epoch": 0.07449104539408194, "grad_norm": 1.6698213815689087, "learning_rate": 1.9898251329257714e-05, "loss": 0.8908, "step": 2114 }, { "epoch": 0.07452628240704036, "grad_norm": 2.053257942199707, "learning_rate": 1.989808887347284e-05, "loss": 4.4605, "step": 2115 }, { "epoch": 0.07456151941999876, "grad_norm": 1.2217601537704468, "learning_rate": 1.98979262887641e-05, "loss": 0.7346, "step": 2116 }, { "epoch": 0.07459675643295718, "grad_norm": 1.7078356742858887, "learning_rate": 1.989776357513361e-05, "loss": 3.4109, "step": 2117 }, { "epoch": 0.07463199344591559, "grad_norm": 2.943335771560669, "learning_rate": 1.9897600732583482e-05, "loss": 6.8433, "step": 2118 }, { "epoch": 0.074667230458874, "grad_norm": 2.7152068614959717, "learning_rate": 1.989743776111585e-05, "loss": 6.1721, "step": 2119 }, { "epoch": 0.07470246747183241, "grad_norm": 1.0726382732391357, "learning_rate": 1.9897274660732827e-05, "loss": 0.866, "step": 2120 }, { "epoch": 0.07473770448479082, "grad_norm": 1.2921055555343628, "learning_rate": 1.9897111431436543e-05, "loss": 1.0919, "step": 2121 }, { "epoch": 0.07477294149774924, "grad_norm": 1.4897255897521973, "learning_rate": 1.989694807322912e-05, "loss": 0.8482, "step": 2122 }, { "epoch": 0.07480817851070765, "grad_norm": 3.5601487159729004, "learning_rate": 1.989678458611269e-05, "loss": 5.8392, "step": 2123 }, { "epoch": 0.07484341552366607, "grad_norm": 1.4711476564407349, "learning_rate": 1.989662097008938e-05, "loss": 1.1148, "step": 2124 }, { "epoch": 0.07487865253662447, "grad_norm": 1.3359040021896362, "learning_rate": 1.9896457225161318e-05, "loss": 1.1144, "step": 2125 }, { "epoch": 0.07491388954958288, "grad_norm": 1.7045620679855347, "learning_rate": 1.9896293351330644e-05, "loss": 1.1101, "step": 2126 }, { "epoch": 0.0749491265625413, "grad_norm": 6.534839630126953, "learning_rate": 1.9896129348599486e-05, "loss": 6.9501, "step": 2127 }, { "epoch": 0.0749843635754997, "grad_norm": 2.3247268199920654, "learning_rate": 1.9895965216969985e-05, "loss": 3.1014, "step": 2128 }, { "epoch": 0.07501960058845812, "grad_norm": 1.2171075344085693, "learning_rate": 1.9895800956444275e-05, "loss": 0.9613, "step": 2129 }, { "epoch": 0.07505483760141653, "grad_norm": 1.1793009042739868, "learning_rate": 1.9895636567024494e-05, "loss": 0.8685, "step": 2130 }, { "epoch": 0.07509007461437493, "grad_norm": 2.7087011337280273, "learning_rate": 1.989547204871279e-05, "loss": 3.24, "step": 2131 }, { "epoch": 0.07512531162733335, "grad_norm": 3.3285014629364014, "learning_rate": 1.98953074015113e-05, "loss": 6.245, "step": 2132 }, { "epoch": 0.07516054864029176, "grad_norm": 1.5867640972137451, "learning_rate": 1.9895142625422172e-05, "loss": 1.1207, "step": 2133 }, { "epoch": 0.07519578565325018, "grad_norm": 2.642261505126953, "learning_rate": 1.989497772044755e-05, "loss": 3.1905, "step": 2134 }, { "epoch": 0.07523102266620858, "grad_norm": 3.428009033203125, "learning_rate": 1.9894812686589583e-05, "loss": 3.9256, "step": 2135 }, { "epoch": 0.075266259679167, "grad_norm": 1.06634521484375, "learning_rate": 1.9894647523850423e-05, "loss": 1.3629, "step": 2136 }, { "epoch": 0.07530149669212541, "grad_norm": 1.6727371215820312, "learning_rate": 1.9894482232232217e-05, "loss": 0.994, "step": 2137 }, { "epoch": 0.07533673370508381, "grad_norm": 2.4463348388671875, "learning_rate": 1.989431681173712e-05, "loss": 4.1754, "step": 2138 }, { "epoch": 0.07537197071804223, "grad_norm": 3.2008323669433594, "learning_rate": 1.9894151262367286e-05, "loss": 3.4062, "step": 2139 }, { "epoch": 0.07540720773100064, "grad_norm": 1.3107413053512573, "learning_rate": 1.989398558412487e-05, "loss": 0.8835, "step": 2140 }, { "epoch": 0.07544244474395906, "grad_norm": 1.8328626155853271, "learning_rate": 1.989381977701203e-05, "loss": 0.8886, "step": 2141 }, { "epoch": 0.07547768175691746, "grad_norm": 1.4400426149368286, "learning_rate": 1.989365384103093e-05, "loss": 1.1872, "step": 2142 }, { "epoch": 0.07551291876987588, "grad_norm": 1.7474271059036255, "learning_rate": 1.9893487776183726e-05, "loss": 0.8121, "step": 2143 }, { "epoch": 0.07554815578283429, "grad_norm": 1.70265793800354, "learning_rate": 1.9893321582472586e-05, "loss": 0.9263, "step": 2144 }, { "epoch": 0.0755833927957927, "grad_norm": 1.4380112886428833, "learning_rate": 1.989315525989967e-05, "loss": 1.0636, "step": 2145 }, { "epoch": 0.07561862980875111, "grad_norm": 4.543999195098877, "learning_rate": 1.9892988808467152e-05, "loss": 6.8202, "step": 2146 }, { "epoch": 0.07565386682170952, "grad_norm": 0.939825713634491, "learning_rate": 1.989282222817719e-05, "loss": 0.8276, "step": 2147 }, { "epoch": 0.07568910383466794, "grad_norm": 1.380436658859253, "learning_rate": 1.9892655519031958e-05, "loss": 0.732, "step": 2148 }, { "epoch": 0.07572434084762635, "grad_norm": 1.7608847618103027, "learning_rate": 1.989248868103363e-05, "loss": 0.8309, "step": 2149 }, { "epoch": 0.07575957786058476, "grad_norm": 4.3189849853515625, "learning_rate": 1.9892321714184374e-05, "loss": 3.5866, "step": 2150 }, { "epoch": 0.07579481487354317, "grad_norm": 6.360549449920654, "learning_rate": 1.989215461848637e-05, "loss": 6.9385, "step": 2151 }, { "epoch": 0.07583005188650158, "grad_norm": 1.4196778535842896, "learning_rate": 1.989198739394179e-05, "loss": 1.154, "step": 2152 }, { "epoch": 0.07586528889946, "grad_norm": 7.994720458984375, "learning_rate": 1.9891820040552817e-05, "loss": 8.7831, "step": 2153 }, { "epoch": 0.0759005259124184, "grad_norm": 3.046140670776367, "learning_rate": 1.9891652558321625e-05, "loss": 3.6608, "step": 2154 }, { "epoch": 0.07593576292537682, "grad_norm": 1.5903024673461914, "learning_rate": 1.98914849472504e-05, "loss": 1.1329, "step": 2155 }, { "epoch": 0.07597099993833523, "grad_norm": 1.2002060413360596, "learning_rate": 1.989131720734132e-05, "loss": 1.1422, "step": 2156 }, { "epoch": 0.07600623695129363, "grad_norm": 1.3603957891464233, "learning_rate": 1.989114933859658e-05, "loss": 0.7285, "step": 2157 }, { "epoch": 0.07604147396425205, "grad_norm": 1.3701744079589844, "learning_rate": 1.9890981341018356e-05, "loss": 0.9723, "step": 2158 }, { "epoch": 0.07607671097721046, "grad_norm": 3.2562918663024902, "learning_rate": 1.989081321460884e-05, "loss": 4.0032, "step": 2159 }, { "epoch": 0.07611194799016888, "grad_norm": 1.44846773147583, "learning_rate": 1.9890644959370224e-05, "loss": 1.0868, "step": 2160 }, { "epoch": 0.07614718500312728, "grad_norm": 1.6109539270401, "learning_rate": 1.9890476575304696e-05, "loss": 0.7731, "step": 2161 }, { "epoch": 0.0761824220160857, "grad_norm": 2.782531499862671, "learning_rate": 1.989030806241445e-05, "loss": 4.3764, "step": 2162 }, { "epoch": 0.07621765902904411, "grad_norm": 2.569429636001587, "learning_rate": 1.9890139420701683e-05, "loss": 4.5894, "step": 2163 }, { "epoch": 0.07625289604200251, "grad_norm": 0.9916761517524719, "learning_rate": 1.988997065016859e-05, "loss": 0.9792, "step": 2164 }, { "epoch": 0.07628813305496093, "grad_norm": 4.940769672393799, "learning_rate": 1.988980175081737e-05, "loss": 6.7683, "step": 2165 }, { "epoch": 0.07632337006791934, "grad_norm": 1.6894112825393677, "learning_rate": 1.9889632722650222e-05, "loss": 0.9498, "step": 2166 }, { "epoch": 0.07635860708087776, "grad_norm": 1.0477573871612549, "learning_rate": 1.9889463565669346e-05, "loss": 1.1, "step": 2167 }, { "epoch": 0.07639384409383616, "grad_norm": 5.76547384262085, "learning_rate": 1.9889294279876948e-05, "loss": 6.9989, "step": 2168 }, { "epoch": 0.07642908110679458, "grad_norm": 5.527298450469971, "learning_rate": 1.9889124865275236e-05, "loss": 9.479, "step": 2169 }, { "epoch": 0.07646431811975299, "grad_norm": 1.2222999334335327, "learning_rate": 1.988895532186641e-05, "loss": 1.0234, "step": 2170 }, { "epoch": 0.0764995551327114, "grad_norm": 2.3654043674468994, "learning_rate": 1.9888785649652682e-05, "loss": 4.2189, "step": 2171 }, { "epoch": 0.07653479214566981, "grad_norm": 4.4349870681762695, "learning_rate": 1.988861584863626e-05, "loss": 8.0824, "step": 2172 }, { "epoch": 0.07657002915862822, "grad_norm": 2.8157386779785156, "learning_rate": 1.988844591881936e-05, "loss": 4.1039, "step": 2173 }, { "epoch": 0.07660526617158664, "grad_norm": 3.3692541122436523, "learning_rate": 1.988827586020419e-05, "loss": 2.9522, "step": 2174 }, { "epoch": 0.07664050318454504, "grad_norm": 5.154280185699463, "learning_rate": 1.9888105672792968e-05, "loss": 10.8787, "step": 2175 }, { "epoch": 0.07667574019750346, "grad_norm": 3.3753294944763184, "learning_rate": 1.988793535658791e-05, "loss": 3.7992, "step": 2176 }, { "epoch": 0.07671097721046187, "grad_norm": 1.0774283409118652, "learning_rate": 1.9887764911591235e-05, "loss": 1.0901, "step": 2177 }, { "epoch": 0.07674621422342028, "grad_norm": 2.336894989013672, "learning_rate": 1.988759433780516e-05, "loss": 3.1119, "step": 2178 }, { "epoch": 0.0767814512363787, "grad_norm": 1.7646751403808594, "learning_rate": 1.988742363523191e-05, "loss": 0.9632, "step": 2179 }, { "epoch": 0.0768166882493371, "grad_norm": 1.4762903451919556, "learning_rate": 1.988725280387371e-05, "loss": 1.1266, "step": 2180 }, { "epoch": 0.07685192526229552, "grad_norm": 2.3714499473571777, "learning_rate": 1.988708184373278e-05, "loss": 3.5438, "step": 2181 }, { "epoch": 0.07688716227525393, "grad_norm": 1.3781309127807617, "learning_rate": 1.9886910754811352e-05, "loss": 1.0293, "step": 2182 }, { "epoch": 0.07692239928821233, "grad_norm": 1.4386322498321533, "learning_rate": 1.9886739537111653e-05, "loss": 0.8789, "step": 2183 }, { "epoch": 0.07695763630117075, "grad_norm": 1.8289412260055542, "learning_rate": 1.9886568190635908e-05, "loss": 0.8373, "step": 2184 }, { "epoch": 0.07699287331412916, "grad_norm": 1.1953688859939575, "learning_rate": 1.9886396715386357e-05, "loss": 1.0183, "step": 2185 }, { "epoch": 0.07702811032708758, "grad_norm": 3.1383299827575684, "learning_rate": 1.9886225111365227e-05, "loss": 3.4272, "step": 2186 }, { "epoch": 0.07706334734004598, "grad_norm": 8.3735933303833, "learning_rate": 1.9886053378574756e-05, "loss": 3.9221, "step": 2187 }, { "epoch": 0.0770985843530044, "grad_norm": 5.39243221282959, "learning_rate": 1.988588151701718e-05, "loss": 3.2995, "step": 2188 }, { "epoch": 0.0771338213659628, "grad_norm": 1.1425621509552002, "learning_rate": 1.988570952669474e-05, "loss": 0.9712, "step": 2189 }, { "epoch": 0.07716905837892121, "grad_norm": 1.606228232383728, "learning_rate": 1.9885537407609675e-05, "loss": 0.8927, "step": 2190 }, { "epoch": 0.07720429539187963, "grad_norm": 3.8627781867980957, "learning_rate": 1.988536515976422e-05, "loss": 5.7696, "step": 2191 }, { "epoch": 0.07723953240483804, "grad_norm": 1.163545846939087, "learning_rate": 1.988519278316063e-05, "loss": 0.837, "step": 2192 }, { "epoch": 0.07727476941779646, "grad_norm": 2.642773151397705, "learning_rate": 1.9885020277801143e-05, "loss": 3.2283, "step": 2193 }, { "epoch": 0.07731000643075486, "grad_norm": 4.416532039642334, "learning_rate": 1.988484764368801e-05, "loss": 4.2841, "step": 2194 }, { "epoch": 0.07734524344371328, "grad_norm": 3.083031415939331, "learning_rate": 1.9884674880823475e-05, "loss": 2.8179, "step": 2195 }, { "epoch": 0.07738048045667169, "grad_norm": 4.388926982879639, "learning_rate": 1.988450198920979e-05, "loss": 4.144, "step": 2196 }, { "epoch": 0.0774157174696301, "grad_norm": 0.9290681481361389, "learning_rate": 1.988432896884921e-05, "loss": 1.1187, "step": 2197 }, { "epoch": 0.07745095448258851, "grad_norm": 2.4666731357574463, "learning_rate": 1.9884155819743984e-05, "loss": 3.6096, "step": 2198 }, { "epoch": 0.07748619149554692, "grad_norm": 4.337451934814453, "learning_rate": 1.988398254189637e-05, "loss": 3.6589, "step": 2199 }, { "epoch": 0.07752142850850534, "grad_norm": 4.015838623046875, "learning_rate": 1.9883809135308627e-05, "loss": 3.7882, "step": 2200 }, { "epoch": 0.07755666552146374, "grad_norm": 1.8049355745315552, "learning_rate": 1.9883635599983004e-05, "loss": 1.0017, "step": 2201 }, { "epoch": 0.07759190253442216, "grad_norm": 1.3741289377212524, "learning_rate": 1.9883461935921774e-05, "loss": 1.1894, "step": 2202 }, { "epoch": 0.07762713954738057, "grad_norm": 4.389278411865234, "learning_rate": 1.9883288143127193e-05, "loss": 5.5795, "step": 2203 }, { "epoch": 0.07766237656033897, "grad_norm": 3.256654739379883, "learning_rate": 1.9883114221601523e-05, "loss": 3.6105, "step": 2204 }, { "epoch": 0.0776976135732974, "grad_norm": 4.081478595733643, "learning_rate": 1.9882940171347035e-05, "loss": 4.4092, "step": 2205 }, { "epoch": 0.0777328505862558, "grad_norm": 2.180673122406006, "learning_rate": 1.9882765992365986e-05, "loss": 4.0949, "step": 2206 }, { "epoch": 0.07776808759921422, "grad_norm": 1.1895549297332764, "learning_rate": 1.988259168466066e-05, "loss": 1.0658, "step": 2207 }, { "epoch": 0.07780332461217263, "grad_norm": 4.22991418838501, "learning_rate": 1.988241724823331e-05, "loss": 6.8555, "step": 2208 }, { "epoch": 0.07783856162513103, "grad_norm": 3.419987440109253, "learning_rate": 1.988224268308622e-05, "loss": 3.8044, "step": 2209 }, { "epoch": 0.07787379863808945, "grad_norm": 1.063701868057251, "learning_rate": 1.988206798922166e-05, "loss": 0.9256, "step": 2210 }, { "epoch": 0.07790903565104786, "grad_norm": 1.407406210899353, "learning_rate": 1.988189316664191e-05, "loss": 0.848, "step": 2211 }, { "epoch": 0.07794427266400628, "grad_norm": 1.4755879640579224, "learning_rate": 1.9881718215349235e-05, "loss": 1.0513, "step": 2212 }, { "epoch": 0.07797950967696468, "grad_norm": 3.9169583320617676, "learning_rate": 1.9881543135345924e-05, "loss": 3.8646, "step": 2213 }, { "epoch": 0.0780147466899231, "grad_norm": 2.3308544158935547, "learning_rate": 1.9881367926634256e-05, "loss": 2.9201, "step": 2214 }, { "epoch": 0.0780499837028815, "grad_norm": 1.2014387845993042, "learning_rate": 1.9881192589216515e-05, "loss": 0.9868, "step": 2215 }, { "epoch": 0.07808522071583991, "grad_norm": 1.0791605710983276, "learning_rate": 1.9881017123094976e-05, "loss": 0.9867, "step": 2216 }, { "epoch": 0.07812045772879833, "grad_norm": 1.4662691354751587, "learning_rate": 1.9880841528271935e-05, "loss": 0.8658, "step": 2217 }, { "epoch": 0.07815569474175674, "grad_norm": 1.0792105197906494, "learning_rate": 1.988066580474967e-05, "loss": 1.1369, "step": 2218 }, { "epoch": 0.07819093175471516, "grad_norm": 1.410764217376709, "learning_rate": 1.988048995253048e-05, "loss": 0.9873, "step": 2219 }, { "epoch": 0.07822616876767356, "grad_norm": 4.143433094024658, "learning_rate": 1.9880313971616646e-05, "loss": 6.5957, "step": 2220 }, { "epoch": 0.07826140578063198, "grad_norm": 1.9856455326080322, "learning_rate": 1.9880137862010466e-05, "loss": 2.989, "step": 2221 }, { "epoch": 0.07829664279359039, "grad_norm": 1.0782417058944702, "learning_rate": 1.987996162371423e-05, "loss": 0.8274, "step": 2222 }, { "epoch": 0.0783318798065488, "grad_norm": 1.2617286443710327, "learning_rate": 1.9879785256730236e-05, "loss": 1.3025, "step": 2223 }, { "epoch": 0.07836711681950721, "grad_norm": 1.158158302307129, "learning_rate": 1.987960876106078e-05, "loss": 1.044, "step": 2224 }, { "epoch": 0.07840235383246562, "grad_norm": 1.2885125875473022, "learning_rate": 1.9879432136708162e-05, "loss": 1.1054, "step": 2225 }, { "epoch": 0.07843759084542404, "grad_norm": 3.3205041885375977, "learning_rate": 1.987925538367468e-05, "loss": 6.3126, "step": 2226 }, { "epoch": 0.07847282785838244, "grad_norm": 2.8392534255981445, "learning_rate": 1.9879078501962642e-05, "loss": 4.2334, "step": 2227 }, { "epoch": 0.07850806487134086, "grad_norm": 5.9511919021606445, "learning_rate": 1.9878901491574348e-05, "loss": 3.8511, "step": 2228 }, { "epoch": 0.07854330188429927, "grad_norm": 2.167461633682251, "learning_rate": 1.9878724352512103e-05, "loss": 3.7566, "step": 2229 }, { "epoch": 0.07857853889725767, "grad_norm": 4.581006050109863, "learning_rate": 1.987854708477821e-05, "loss": 3.9879, "step": 2230 }, { "epoch": 0.0786137759102161, "grad_norm": 3.7747552394866943, "learning_rate": 1.987836968837499e-05, "loss": 3.6023, "step": 2231 }, { "epoch": 0.0786490129231745, "grad_norm": 1.0485188961029053, "learning_rate": 1.987819216330474e-05, "loss": 0.8392, "step": 2232 }, { "epoch": 0.07868424993613292, "grad_norm": 1.5564601421356201, "learning_rate": 1.9878014509569786e-05, "loss": 0.9795, "step": 2233 }, { "epoch": 0.07871948694909132, "grad_norm": 1.9448113441467285, "learning_rate": 1.987783672717243e-05, "loss": 0.9352, "step": 2234 }, { "epoch": 0.07875472396204973, "grad_norm": 4.964756965637207, "learning_rate": 1.9877658816114994e-05, "loss": 6.6944, "step": 2235 }, { "epoch": 0.07878996097500815, "grad_norm": 5.37228536605835, "learning_rate": 1.9877480776399794e-05, "loss": 9.4406, "step": 2236 }, { "epoch": 0.07882519798796656, "grad_norm": 5.752933025360107, "learning_rate": 1.9877302608029146e-05, "loss": 6.4946, "step": 2237 }, { "epoch": 0.07886043500092497, "grad_norm": 1.6946570873260498, "learning_rate": 1.9877124311005374e-05, "loss": 1.1642, "step": 2238 }, { "epoch": 0.07889567201388338, "grad_norm": 4.501028537750244, "learning_rate": 1.98769458853308e-05, "loss": 3.6678, "step": 2239 }, { "epoch": 0.0789309090268418, "grad_norm": 1.7052031755447388, "learning_rate": 1.987676733100775e-05, "loss": 0.9359, "step": 2240 }, { "epoch": 0.0789661460398002, "grad_norm": 2.753352642059326, "learning_rate": 1.9876588648038548e-05, "loss": 3.5095, "step": 2241 }, { "epoch": 0.07900138305275861, "grad_norm": 4.395729064941406, "learning_rate": 1.9876409836425516e-05, "loss": 6.6741, "step": 2242 }, { "epoch": 0.07903662006571703, "grad_norm": 1.7339757680892944, "learning_rate": 1.9876230896170992e-05, "loss": 1.112, "step": 2243 }, { "epoch": 0.07907185707867544, "grad_norm": 5.548560619354248, "learning_rate": 1.98760518272773e-05, "loss": 5.6477, "step": 2244 }, { "epoch": 0.07910709409163386, "grad_norm": 2.981191396713257, "learning_rate": 1.9875872629746778e-05, "loss": 4.1703, "step": 2245 }, { "epoch": 0.07914233110459226, "grad_norm": 1.4491807222366333, "learning_rate": 1.9875693303581752e-05, "loss": 0.916, "step": 2246 }, { "epoch": 0.07917756811755068, "grad_norm": 5.191670894622803, "learning_rate": 1.9875513848784566e-05, "loss": 3.1099, "step": 2247 }, { "epoch": 0.07921280513050909, "grad_norm": 1.0264536142349243, "learning_rate": 1.9875334265357556e-05, "loss": 1.2206, "step": 2248 }, { "epoch": 0.07924804214346749, "grad_norm": 1.1460891962051392, "learning_rate": 1.9875154553303056e-05, "loss": 0.9182, "step": 2249 }, { "epoch": 0.07928327915642591, "grad_norm": 0.9605761766433716, "learning_rate": 1.9874974712623412e-05, "loss": 0.9767, "step": 2250 }, { "epoch": 0.07931851616938432, "grad_norm": 1.2468457221984863, "learning_rate": 1.9874794743320964e-05, "loss": 1.1492, "step": 2251 }, { "epoch": 0.07935375318234274, "grad_norm": 1.5386258363723755, "learning_rate": 1.9874614645398055e-05, "loss": 0.9882, "step": 2252 }, { "epoch": 0.07938899019530114, "grad_norm": 1.344178318977356, "learning_rate": 1.9874434418857037e-05, "loss": 1.3404, "step": 2253 }, { "epoch": 0.07942422720825956, "grad_norm": 5.3028082847595215, "learning_rate": 1.987425406370025e-05, "loss": 3.9661, "step": 2254 }, { "epoch": 0.07945946422121797, "grad_norm": 1.85908842086792, "learning_rate": 1.9874073579930045e-05, "loss": 0.9062, "step": 2255 }, { "epoch": 0.07949470123417637, "grad_norm": 8.406448364257812, "learning_rate": 1.9873892967548772e-05, "loss": 7.0848, "step": 2256 }, { "epoch": 0.0795299382471348, "grad_norm": 1.1784459352493286, "learning_rate": 1.9873712226558788e-05, "loss": 1.0854, "step": 2257 }, { "epoch": 0.0795651752600932, "grad_norm": 5.872072696685791, "learning_rate": 1.9873531356962446e-05, "loss": 5.6517, "step": 2258 }, { "epoch": 0.07960041227305162, "grad_norm": 4.811725616455078, "learning_rate": 1.9873350358762097e-05, "loss": 3.4878, "step": 2259 }, { "epoch": 0.07963564928601002, "grad_norm": 1.1833192110061646, "learning_rate": 1.9873169231960105e-05, "loss": 1.0458, "step": 2260 }, { "epoch": 0.07967088629896843, "grad_norm": 1.765871286392212, "learning_rate": 1.9872987976558825e-05, "loss": 1.0483, "step": 2261 }, { "epoch": 0.07970612331192685, "grad_norm": 1.4197064638137817, "learning_rate": 1.9872806592560618e-05, "loss": 1.1117, "step": 2262 }, { "epoch": 0.07974136032488525, "grad_norm": 4.717961311340332, "learning_rate": 1.987262507996785e-05, "loss": 6.0232, "step": 2263 }, { "epoch": 0.07977659733784367, "grad_norm": 1.4934931993484497, "learning_rate": 1.987244343878288e-05, "loss": 1.253, "step": 2264 }, { "epoch": 0.07981183435080208, "grad_norm": 3.324324369430542, "learning_rate": 1.9872261669008078e-05, "loss": 3.3193, "step": 2265 }, { "epoch": 0.0798470713637605, "grad_norm": 1.4189918041229248, "learning_rate": 1.987207977064581e-05, "loss": 1.1067, "step": 2266 }, { "epoch": 0.0798823083767189, "grad_norm": 4.004894733428955, "learning_rate": 1.987189774369845e-05, "loss": 2.8108, "step": 2267 }, { "epoch": 0.07991754538967731, "grad_norm": 1.7892265319824219, "learning_rate": 1.9871715588168358e-05, "loss": 1.0457, "step": 2268 }, { "epoch": 0.07995278240263573, "grad_norm": 4.820724010467529, "learning_rate": 1.9871533304057914e-05, "loss": 3.4475, "step": 2269 }, { "epoch": 0.07998801941559414, "grad_norm": 1.0972044467926025, "learning_rate": 1.9871350891369492e-05, "loss": 1.0116, "step": 2270 }, { "epoch": 0.08002325642855256, "grad_norm": 1.5525625944137573, "learning_rate": 1.9871168350105466e-05, "loss": 0.9208, "step": 2271 }, { "epoch": 0.08005849344151096, "grad_norm": 3.831200122833252, "learning_rate": 1.987098568026822e-05, "loss": 3.3699, "step": 2272 }, { "epoch": 0.08009373045446938, "grad_norm": 9.373784065246582, "learning_rate": 1.987080288186012e-05, "loss": 9.3828, "step": 2273 }, { "epoch": 0.08012896746742779, "grad_norm": 2.652459144592285, "learning_rate": 1.987061995488356e-05, "loss": 3.3257, "step": 2274 }, { "epoch": 0.08016420448038619, "grad_norm": 3.5827741622924805, "learning_rate": 1.987043689934092e-05, "loss": 4.0185, "step": 2275 }, { "epoch": 0.08019944149334461, "grad_norm": 1.5432902574539185, "learning_rate": 1.987025371523458e-05, "loss": 1.0262, "step": 2276 }, { "epoch": 0.08023467850630302, "grad_norm": 8.918190956115723, "learning_rate": 1.9870070402566926e-05, "loss": 3.2939, "step": 2277 }, { "epoch": 0.08026991551926144, "grad_norm": 1.398113489151001, "learning_rate": 1.986988696134035e-05, "loss": 0.9215, "step": 2278 }, { "epoch": 0.08030515253221984, "grad_norm": 7.028762340545654, "learning_rate": 1.9869703391557238e-05, "loss": 5.4392, "step": 2279 }, { "epoch": 0.08034038954517826, "grad_norm": 9.487737655639648, "learning_rate": 1.9869519693219982e-05, "loss": 6.142, "step": 2280 }, { "epoch": 0.08037562655813667, "grad_norm": 0.8766900300979614, "learning_rate": 1.9869335866330976e-05, "loss": 1.0726, "step": 2281 }, { "epoch": 0.08041086357109507, "grad_norm": 1.6610965728759766, "learning_rate": 1.9869151910892613e-05, "loss": 1.1673, "step": 2282 }, { "epoch": 0.08044610058405349, "grad_norm": 1.0921986103057861, "learning_rate": 1.9868967826907287e-05, "loss": 1.1234, "step": 2283 }, { "epoch": 0.0804813375970119, "grad_norm": 1.0355315208435059, "learning_rate": 1.9868783614377403e-05, "loss": 1.0918, "step": 2284 }, { "epoch": 0.08051657460997032, "grad_norm": 1.1967153549194336, "learning_rate": 1.986859927330535e-05, "loss": 1.2589, "step": 2285 }, { "epoch": 0.08055181162292872, "grad_norm": 2.1804990768432617, "learning_rate": 1.986841480369354e-05, "loss": 4.342, "step": 2286 }, { "epoch": 0.08058704863588713, "grad_norm": 1.8270803689956665, "learning_rate": 1.9868230205544362e-05, "loss": 1.0055, "step": 2287 }, { "epoch": 0.08062228564884555, "grad_norm": 3.5757226943969727, "learning_rate": 1.9868045478860238e-05, "loss": 2.5409, "step": 2288 }, { "epoch": 0.08065752266180395, "grad_norm": 3.4696662425994873, "learning_rate": 1.9867860623643556e-05, "loss": 3.7574, "step": 2289 }, { "epoch": 0.08069275967476237, "grad_norm": 5.0590972900390625, "learning_rate": 1.9867675639896737e-05, "loss": 5.1733, "step": 2290 }, { "epoch": 0.08072799668772078, "grad_norm": 1.84752357006073, "learning_rate": 1.9867490527622185e-05, "loss": 0.8536, "step": 2291 }, { "epoch": 0.0807632337006792, "grad_norm": 1.1306045055389404, "learning_rate": 1.9867305286822313e-05, "loss": 0.902, "step": 2292 }, { "epoch": 0.0807984707136376, "grad_norm": 3.0862808227539062, "learning_rate": 1.9867119917499533e-05, "loss": 1.1095, "step": 2293 }, { "epoch": 0.08083370772659601, "grad_norm": 5.038862705230713, "learning_rate": 1.9866934419656257e-05, "loss": 6.1878, "step": 2294 }, { "epoch": 0.08086894473955443, "grad_norm": 4.747429370880127, "learning_rate": 1.9866748793294906e-05, "loss": 3.7175, "step": 2295 }, { "epoch": 0.08090418175251284, "grad_norm": 3.525758743286133, "learning_rate": 1.9866563038417893e-05, "loss": 3.9969, "step": 2296 }, { "epoch": 0.08093941876547125, "grad_norm": 1.2976187467575073, "learning_rate": 1.986637715502764e-05, "loss": 1.1421, "step": 2297 }, { "epoch": 0.08097465577842966, "grad_norm": 3.8462228775024414, "learning_rate": 1.9866191143126564e-05, "loss": 4.9506, "step": 2298 }, { "epoch": 0.08100989279138808, "grad_norm": 2.566999912261963, "learning_rate": 1.9866005002717098e-05, "loss": 0.88, "step": 2299 }, { "epoch": 0.08104512980434649, "grad_norm": 3.5262014865875244, "learning_rate": 1.986581873380166e-05, "loss": 1.0178, "step": 2300 }, { "epoch": 0.08108036681730489, "grad_norm": 1.5901387929916382, "learning_rate": 1.986563233638267e-05, "loss": 1.1992, "step": 2301 }, { "epoch": 0.08111560383026331, "grad_norm": 1.0772173404693604, "learning_rate": 1.986544581046257e-05, "loss": 0.9656, "step": 2302 }, { "epoch": 0.08115084084322172, "grad_norm": 4.555544376373291, "learning_rate": 1.9865259156043776e-05, "loss": 6.9322, "step": 2303 }, { "epoch": 0.08118607785618014, "grad_norm": 2.1381077766418457, "learning_rate": 1.9865072373128727e-05, "loss": 3.6417, "step": 2304 }, { "epoch": 0.08122131486913854, "grad_norm": 4.334820747375488, "learning_rate": 1.986488546171985e-05, "loss": 5.5776, "step": 2305 }, { "epoch": 0.08125655188209696, "grad_norm": 2.3698601722717285, "learning_rate": 1.986469842181959e-05, "loss": 4.3274, "step": 2306 }, { "epoch": 0.08129178889505537, "grad_norm": 2.6724021434783936, "learning_rate": 1.9864511253430373e-05, "loss": 0.9361, "step": 2307 }, { "epoch": 0.08132702590801377, "grad_norm": 1.4020928144454956, "learning_rate": 1.9864323956554638e-05, "loss": 0.9431, "step": 2308 }, { "epoch": 0.08136226292097219, "grad_norm": 3.6553568840026855, "learning_rate": 1.986413653119483e-05, "loss": 3.3046, "step": 2309 }, { "epoch": 0.0813974999339306, "grad_norm": 7.736029148101807, "learning_rate": 1.986394897735339e-05, "loss": 5.4539, "step": 2310 }, { "epoch": 0.08143273694688902, "grad_norm": 1.1769877672195435, "learning_rate": 1.9863761295032754e-05, "loss": 0.8774, "step": 2311 }, { "epoch": 0.08146797395984742, "grad_norm": 1.4713977575302124, "learning_rate": 1.986357348423537e-05, "loss": 1.0349, "step": 2312 }, { "epoch": 0.08150321097280583, "grad_norm": 0.9472448229789734, "learning_rate": 1.9863385544963688e-05, "loss": 0.8565, "step": 2313 }, { "epoch": 0.08153844798576425, "grad_norm": 1.1065552234649658, "learning_rate": 1.986319747722015e-05, "loss": 1.122, "step": 2314 }, { "epoch": 0.08157368499872265, "grad_norm": 1.1725174188613892, "learning_rate": 1.986300928100721e-05, "loss": 0.8498, "step": 2315 }, { "epoch": 0.08160892201168107, "grad_norm": 1.7629157304763794, "learning_rate": 1.9862820956327316e-05, "loss": 0.9125, "step": 2316 }, { "epoch": 0.08164415902463948, "grad_norm": 3.984584331512451, "learning_rate": 1.9862632503182927e-05, "loss": 3.6077, "step": 2317 }, { "epoch": 0.0816793960375979, "grad_norm": 1.3728591203689575, "learning_rate": 1.9862443921576487e-05, "loss": 0.8695, "step": 2318 }, { "epoch": 0.0817146330505563, "grad_norm": 1.5633009672164917, "learning_rate": 1.9862255211510464e-05, "loss": 1.0097, "step": 2319 }, { "epoch": 0.08174987006351471, "grad_norm": 1.579769253730774, "learning_rate": 1.986206637298731e-05, "loss": 1.1392, "step": 2320 }, { "epoch": 0.08178510707647313, "grad_norm": 6.289689540863037, "learning_rate": 1.9861877406009486e-05, "loss": 6.9248, "step": 2321 }, { "epoch": 0.08182034408943153, "grad_norm": 3.915605306625366, "learning_rate": 1.9861688310579447e-05, "loss": 3.6498, "step": 2322 }, { "epoch": 0.08185558110238995, "grad_norm": 1.4055002927780151, "learning_rate": 1.9861499086699667e-05, "loss": 1.0541, "step": 2323 }, { "epoch": 0.08189081811534836, "grad_norm": 1.826080560684204, "learning_rate": 1.9861309734372602e-05, "loss": 0.9841, "step": 2324 }, { "epoch": 0.08192605512830678, "grad_norm": 4.826569557189941, "learning_rate": 1.9861120253600723e-05, "loss": 6.835, "step": 2325 }, { "epoch": 0.08196129214126519, "grad_norm": 7.810333728790283, "learning_rate": 1.9860930644386495e-05, "loss": 6.0827, "step": 2326 }, { "epoch": 0.08199652915422359, "grad_norm": 2.311459541320801, "learning_rate": 1.9860740906732392e-05, "loss": 3.5106, "step": 2327 }, { "epoch": 0.08203176616718201, "grad_norm": 1.658400058746338, "learning_rate": 1.986055104064088e-05, "loss": 0.8366, "step": 2328 }, { "epoch": 0.08206700318014042, "grad_norm": 2.1801412105560303, "learning_rate": 1.986036104611444e-05, "loss": 3.6493, "step": 2329 }, { "epoch": 0.08210224019309884, "grad_norm": 4.49276065826416, "learning_rate": 1.986017092315554e-05, "loss": 4.1071, "step": 2330 }, { "epoch": 0.08213747720605724, "grad_norm": 4.440062999725342, "learning_rate": 1.9859980671766653e-05, "loss": 4.401, "step": 2331 }, { "epoch": 0.08217271421901566, "grad_norm": 4.651472091674805, "learning_rate": 1.9859790291950266e-05, "loss": 6.1978, "step": 2332 }, { "epoch": 0.08220795123197407, "grad_norm": 5.124188423156738, "learning_rate": 1.985959978370885e-05, "loss": 3.6314, "step": 2333 }, { "epoch": 0.08224318824493247, "grad_norm": 1.9447323083877563, "learning_rate": 1.9859409147044896e-05, "loss": 0.9041, "step": 2334 }, { "epoch": 0.08227842525789089, "grad_norm": 1.4646011590957642, "learning_rate": 1.9859218381960875e-05, "loss": 1.1055, "step": 2335 }, { "epoch": 0.0823136622708493, "grad_norm": 1.9572275876998901, "learning_rate": 1.9859027488459284e-05, "loss": 1.149, "step": 2336 }, { "epoch": 0.08234889928380772, "grad_norm": 0.9659004807472229, "learning_rate": 1.9858836466542605e-05, "loss": 0.9324, "step": 2337 }, { "epoch": 0.08238413629676612, "grad_norm": 2.3385133743286133, "learning_rate": 1.9858645316213323e-05, "loss": 3.659, "step": 2338 }, { "epoch": 0.08241937330972453, "grad_norm": 1.0584633350372314, "learning_rate": 1.985845403747393e-05, "loss": 1.0623, "step": 2339 }, { "epoch": 0.08245461032268295, "grad_norm": 3.6125648021698, "learning_rate": 1.985826263032692e-05, "loss": 3.3988, "step": 2340 }, { "epoch": 0.08248984733564135, "grad_norm": 1.816268801689148, "learning_rate": 1.985807109477478e-05, "loss": 0.8448, "step": 2341 }, { "epoch": 0.08252508434859977, "grad_norm": 1.7067527770996094, "learning_rate": 1.985787943082001e-05, "loss": 0.9009, "step": 2342 }, { "epoch": 0.08256032136155818, "grad_norm": 1.1280661821365356, "learning_rate": 1.9857687638465103e-05, "loss": 1.0771, "step": 2343 }, { "epoch": 0.0825955583745166, "grad_norm": 1.298140525817871, "learning_rate": 1.9857495717712562e-05, "loss": 0.6636, "step": 2344 }, { "epoch": 0.082630795387475, "grad_norm": 2.7692790031433105, "learning_rate": 1.9857303668564883e-05, "loss": 3.4204, "step": 2345 }, { "epoch": 0.08266603240043341, "grad_norm": 4.213258743286133, "learning_rate": 1.9857111491024566e-05, "loss": 3.4955, "step": 2346 }, { "epoch": 0.08270126941339183, "grad_norm": 4.643462181091309, "learning_rate": 1.9856919185094116e-05, "loss": 3.4275, "step": 2347 }, { "epoch": 0.08273650642635023, "grad_norm": 1.312177062034607, "learning_rate": 1.985672675077604e-05, "loss": 0.8435, "step": 2348 }, { "epoch": 0.08277174343930865, "grad_norm": 7.9447221755981445, "learning_rate": 1.9856534188072843e-05, "loss": 7.1948, "step": 2349 }, { "epoch": 0.08280698045226706, "grad_norm": 4.006362438201904, "learning_rate": 1.985634149698703e-05, "loss": 5.9109, "step": 2350 }, { "epoch": 0.08284221746522548, "grad_norm": 4.4961419105529785, "learning_rate": 1.9856148677521118e-05, "loss": 3.4344, "step": 2351 }, { "epoch": 0.08287745447818388, "grad_norm": 10.318888664245605, "learning_rate": 1.985595572967761e-05, "loss": 6.9172, "step": 2352 }, { "epoch": 0.08291269149114229, "grad_norm": 4.2240071296691895, "learning_rate": 1.9855762653459024e-05, "loss": 3.7198, "step": 2353 }, { "epoch": 0.08294792850410071, "grad_norm": 1.048754334449768, "learning_rate": 1.9855569448867874e-05, "loss": 0.7925, "step": 2354 }, { "epoch": 0.08298316551705912, "grad_norm": 1.0408211946487427, "learning_rate": 1.9855376115906675e-05, "loss": 1.0103, "step": 2355 }, { "epoch": 0.08301840253001753, "grad_norm": 1.1906869411468506, "learning_rate": 1.9855182654577948e-05, "loss": 0.8177, "step": 2356 }, { "epoch": 0.08305363954297594, "grad_norm": 3.152179479598999, "learning_rate": 1.985498906488421e-05, "loss": 3.2003, "step": 2357 }, { "epoch": 0.08308887655593436, "grad_norm": 3.980302572250366, "learning_rate": 1.985479534682799e-05, "loss": 3.7857, "step": 2358 }, { "epoch": 0.08312411356889277, "grad_norm": 0.8814953565597534, "learning_rate": 1.98546015004118e-05, "loss": 0.8161, "step": 2359 }, { "epoch": 0.08315935058185117, "grad_norm": 1.5505801439285278, "learning_rate": 1.9854407525638174e-05, "loss": 0.9094, "step": 2360 }, { "epoch": 0.08319458759480959, "grad_norm": 3.1939008235931396, "learning_rate": 1.985421342250963e-05, "loss": 4.0768, "step": 2361 }, { "epoch": 0.083229824607768, "grad_norm": 3.0880720615386963, "learning_rate": 1.9854019191028703e-05, "loss": 6.0484, "step": 2362 }, { "epoch": 0.08326506162072642, "grad_norm": 3.1660172939300537, "learning_rate": 1.9853824831197922e-05, "loss": 3.6521, "step": 2363 }, { "epoch": 0.08330029863368482, "grad_norm": 2.990793228149414, "learning_rate": 1.9853630343019817e-05, "loss": 3.726, "step": 2364 }, { "epoch": 0.08333553564664323, "grad_norm": 1.3334628343582153, "learning_rate": 1.985343572649692e-05, "loss": 0.8375, "step": 2365 }, { "epoch": 0.08337077265960165, "grad_norm": 3.2340848445892334, "learning_rate": 1.9853240981631766e-05, "loss": 4.8062, "step": 2366 }, { "epoch": 0.08340600967256005, "grad_norm": 2.5488648414611816, "learning_rate": 1.98530461084269e-05, "loss": 3.4349, "step": 2367 }, { "epoch": 0.08344124668551847, "grad_norm": 5.265584945678711, "learning_rate": 1.9852851106884848e-05, "loss": 4.0604, "step": 2368 }, { "epoch": 0.08347648369847688, "grad_norm": 1.799750804901123, "learning_rate": 1.9852655977008158e-05, "loss": 1.074, "step": 2369 }, { "epoch": 0.0835117207114353, "grad_norm": 4.106281280517578, "learning_rate": 1.9852460718799363e-05, "loss": 5.674, "step": 2370 }, { "epoch": 0.0835469577243937, "grad_norm": 5.2855682373046875, "learning_rate": 1.9852265332261017e-05, "loss": 6.5278, "step": 2371 }, { "epoch": 0.08358219473735211, "grad_norm": 3.8243086338043213, "learning_rate": 1.9852069817395662e-05, "loss": 3.8364, "step": 2372 }, { "epoch": 0.08361743175031053, "grad_norm": 3.492004871368408, "learning_rate": 1.985187417420584e-05, "loss": 3.4437, "step": 2373 }, { "epoch": 0.08365266876326893, "grad_norm": 1.8884862661361694, "learning_rate": 1.98516784026941e-05, "loss": 1.0103, "step": 2374 }, { "epoch": 0.08368790577622735, "grad_norm": 1.7455390691757202, "learning_rate": 1.9851482502863e-05, "loss": 0.8725, "step": 2375 }, { "epoch": 0.08372314278918576, "grad_norm": 4.426313877105713, "learning_rate": 1.9851286474715078e-05, "loss": 3.7431, "step": 2376 }, { "epoch": 0.08375837980214418, "grad_norm": 1.3902174234390259, "learning_rate": 1.98510903182529e-05, "loss": 0.8561, "step": 2377 }, { "epoch": 0.08379361681510258, "grad_norm": 1.6211951971054077, "learning_rate": 1.9850894033479015e-05, "loss": 1.1686, "step": 2378 }, { "epoch": 0.08382885382806099, "grad_norm": 5.031213283538818, "learning_rate": 1.985069762039598e-05, "loss": 3.628, "step": 2379 }, { "epoch": 0.08386409084101941, "grad_norm": 4.843041896820068, "learning_rate": 1.985050107900635e-05, "loss": 3.3769, "step": 2380 }, { "epoch": 0.08389932785397781, "grad_norm": 6.167156219482422, "learning_rate": 1.9850304409312693e-05, "loss": 6.3731, "step": 2381 }, { "epoch": 0.08393456486693623, "grad_norm": 1.1342586278915405, "learning_rate": 1.9850107611317563e-05, "loss": 1.011, "step": 2382 }, { "epoch": 0.08396980187989464, "grad_norm": 2.5653233528137207, "learning_rate": 1.9849910685023532e-05, "loss": 3.5855, "step": 2383 }, { "epoch": 0.08400503889285306, "grad_norm": 3.3338873386383057, "learning_rate": 1.9849713630433154e-05, "loss": 3.5065, "step": 2384 }, { "epoch": 0.08404027590581147, "grad_norm": 1.3441885709762573, "learning_rate": 1.9849516447549005e-05, "loss": 0.7624, "step": 2385 }, { "epoch": 0.08407551291876987, "grad_norm": 8.355670928955078, "learning_rate": 1.984931913637365e-05, "loss": 5.7081, "step": 2386 }, { "epoch": 0.08411074993172829, "grad_norm": 1.3467274904251099, "learning_rate": 1.9849121696909656e-05, "loss": 1.2355, "step": 2387 }, { "epoch": 0.0841459869446867, "grad_norm": 1.1249327659606934, "learning_rate": 1.98489241291596e-05, "loss": 0.9153, "step": 2388 }, { "epoch": 0.08418122395764512, "grad_norm": 1.6549633741378784, "learning_rate": 1.984872643312605e-05, "loss": 0.8419, "step": 2389 }, { "epoch": 0.08421646097060352, "grad_norm": 1.7579305171966553, "learning_rate": 1.9848528608811587e-05, "loss": 1.0716, "step": 2390 }, { "epoch": 0.08425169798356193, "grad_norm": 1.2187153100967407, "learning_rate": 1.984833065621878e-05, "loss": 0.9898, "step": 2391 }, { "epoch": 0.08428693499652035, "grad_norm": 1.8362818956375122, "learning_rate": 1.9848132575350214e-05, "loss": 0.9376, "step": 2392 }, { "epoch": 0.08432217200947875, "grad_norm": 1.9588921070098877, "learning_rate": 1.9847934366208467e-05, "loss": 1.0104, "step": 2393 }, { "epoch": 0.08435740902243717, "grad_norm": 23.773746490478516, "learning_rate": 1.984773602879612e-05, "loss": 3.6199, "step": 2394 }, { "epoch": 0.08439264603539558, "grad_norm": 10.485150337219238, "learning_rate": 1.9847537563115757e-05, "loss": 5.7285, "step": 2395 }, { "epoch": 0.084427883048354, "grad_norm": 1.1518522500991821, "learning_rate": 1.984733896916996e-05, "loss": 0.7813, "step": 2396 }, { "epoch": 0.0844631200613124, "grad_norm": 104.99382019042969, "learning_rate": 1.9847140246961324e-05, "loss": 4.5325, "step": 2397 }, { "epoch": 0.08449835707427081, "grad_norm": 3.2963576316833496, "learning_rate": 1.9846941396492426e-05, "loss": 3.9237, "step": 2398 }, { "epoch": 0.08453359408722923, "grad_norm": 3.262380838394165, "learning_rate": 1.984674241776587e-05, "loss": 3.0379, "step": 2399 }, { "epoch": 0.08456883110018763, "grad_norm": 1.9280643463134766, "learning_rate": 1.9846543310784233e-05, "loss": 1.1361, "step": 2400 }, { "epoch": 0.08460406811314605, "grad_norm": 8.063138008117676, "learning_rate": 1.9846344075550118e-05, "loss": 3.8281, "step": 2401 }, { "epoch": 0.08463930512610446, "grad_norm": 2.128427505493164, "learning_rate": 1.9846144712066118e-05, "loss": 1.0759, "step": 2402 }, { "epoch": 0.08467454213906288, "grad_norm": 2.132016658782959, "learning_rate": 1.9845945220334827e-05, "loss": 1.1264, "step": 2403 }, { "epoch": 0.08470977915202128, "grad_norm": 1.738036870956421, "learning_rate": 1.9845745600358842e-05, "loss": 1.0502, "step": 2404 }, { "epoch": 0.08474501616497969, "grad_norm": 15.22509479522705, "learning_rate": 1.984554585214077e-05, "loss": 1.5768, "step": 2405 }, { "epoch": 0.08478025317793811, "grad_norm": 3.0853939056396484, "learning_rate": 1.9845345975683213e-05, "loss": 0.9339, "step": 2406 }, { "epoch": 0.08481549019089651, "grad_norm": 7.814184188842773, "learning_rate": 1.9845145970988766e-05, "loss": 3.4545, "step": 2407 }, { "epoch": 0.08485072720385493, "grad_norm": 22.61159896850586, "learning_rate": 1.984494583806004e-05, "loss": 6.5812, "step": 2408 }, { "epoch": 0.08488596421681334, "grad_norm": 79.87761688232422, "learning_rate": 1.984474557689964e-05, "loss": 3.8929, "step": 2409 }, { "epoch": 0.08492120122977176, "grad_norm": 79.87761688232422, "learning_rate": 1.984474557689964e-05, "loss": 8.6964, "step": 2410 }, { "epoch": 0.08495643824273016, "grad_norm": 420.1209716796875, "learning_rate": 1.9844545187510173e-05, "loss": 7.8821, "step": 2411 }, { "epoch": 0.08499167525568857, "grad_norm": 46.33052444458008, "learning_rate": 1.9844344669894254e-05, "loss": 11.4611, "step": 2412 }, { "epoch": 0.08502691226864699, "grad_norm": 29.86028480529785, "learning_rate": 1.9844144024054488e-05, "loss": 8.4606, "step": 2413 }, { "epoch": 0.0850621492816054, "grad_norm": 22.375423431396484, "learning_rate": 1.9843943249993496e-05, "loss": 10.9351, "step": 2414 }, { "epoch": 0.08509738629456381, "grad_norm": 21.929119110107422, "learning_rate": 1.984374234771389e-05, "loss": 8.2201, "step": 2415 }, { "epoch": 0.08513262330752222, "grad_norm": 15.855874061584473, "learning_rate": 1.984354131721828e-05, "loss": 7.3288, "step": 2416 }, { "epoch": 0.08516786032048063, "grad_norm": 25.713510513305664, "learning_rate": 1.9843340158509297e-05, "loss": 13.1768, "step": 2417 }, { "epoch": 0.08520309733343905, "grad_norm": 8.36083698272705, "learning_rate": 1.984313887158955e-05, "loss": 7.434, "step": 2418 }, { "epoch": 0.08523833434639745, "grad_norm": 7.77210807800293, "learning_rate": 1.9842937456461668e-05, "loss": 13.5791, "step": 2419 }, { "epoch": 0.08527357135935587, "grad_norm": 6.166529655456543, "learning_rate": 1.984273591312827e-05, "loss": 9.3873, "step": 2420 }, { "epoch": 0.08530880837231428, "grad_norm": 9.288372993469238, "learning_rate": 1.9842534241591986e-05, "loss": 11.646, "step": 2421 }, { "epoch": 0.0853440453852727, "grad_norm": 12.547172546386719, "learning_rate": 1.9842332441855437e-05, "loss": 9.6814, "step": 2422 }, { "epoch": 0.0853792823982311, "grad_norm": 6.339139938354492, "learning_rate": 1.9842130513921255e-05, "loss": 6.1316, "step": 2423 }, { "epoch": 0.08541451941118951, "grad_norm": 5.376562118530273, "learning_rate": 1.984192845779207e-05, "loss": 9.1699, "step": 2424 }, { "epoch": 0.08544975642414793, "grad_norm": 4.470653057098389, "learning_rate": 1.984172627347051e-05, "loss": 12.6108, "step": 2425 }, { "epoch": 0.08548499343710633, "grad_norm": 5.286890983581543, "learning_rate": 1.9841523960959213e-05, "loss": 9.4846, "step": 2426 }, { "epoch": 0.08552023045006475, "grad_norm": 3.219811201095581, "learning_rate": 1.9841321520260816e-05, "loss": 6.4129, "step": 2427 }, { "epoch": 0.08555546746302316, "grad_norm": 17.936302185058594, "learning_rate": 1.984111895137795e-05, "loss": 6.8207, "step": 2428 }, { "epoch": 0.08559070447598158, "grad_norm": 6.453421592712402, "learning_rate": 1.984091625431326e-05, "loss": 12.8099, "step": 2429 }, { "epoch": 0.08562594148893998, "grad_norm": 4.037910461425781, "learning_rate": 1.9840713429069378e-05, "loss": 9.9578, "step": 2430 }, { "epoch": 0.08566117850189839, "grad_norm": 10.136518478393555, "learning_rate": 1.9840510475648948e-05, "loss": 7.2168, "step": 2431 }, { "epoch": 0.08569641551485681, "grad_norm": 5.426945686340332, "learning_rate": 1.984030739405462e-05, "loss": 13.041, "step": 2432 }, { "epoch": 0.08573165252781521, "grad_norm": 6.153208255767822, "learning_rate": 1.9840104184289034e-05, "loss": 6.2282, "step": 2433 }, { "epoch": 0.08576688954077363, "grad_norm": 6.7685651779174805, "learning_rate": 1.9839900846354836e-05, "loss": 9.0685, "step": 2434 }, { "epoch": 0.08580212655373204, "grad_norm": 15.416261672973633, "learning_rate": 1.9839697380254677e-05, "loss": 6.5683, "step": 2435 }, { "epoch": 0.08583736356669046, "grad_norm": 3.2968621253967285, "learning_rate": 1.9839493785991206e-05, "loss": 6.4567, "step": 2436 }, { "epoch": 0.08587260057964886, "grad_norm": 5.76990270614624, "learning_rate": 1.9839290063567076e-05, "loss": 10.0445, "step": 2437 }, { "epoch": 0.08590783759260727, "grad_norm": 5.261027812957764, "learning_rate": 1.983908621298494e-05, "loss": 9.6416, "step": 2438 }, { "epoch": 0.08594307460556569, "grad_norm": 6.2540693283081055, "learning_rate": 1.983888223424745e-05, "loss": 6.3039, "step": 2439 }, { "epoch": 0.0859783116185241, "grad_norm": 11.299718856811523, "learning_rate": 1.983867812735727e-05, "loss": 6.1009, "step": 2440 }, { "epoch": 0.08601354863148251, "grad_norm": 3.755725622177124, "learning_rate": 1.9838473892317053e-05, "loss": 10.1496, "step": 2441 }, { "epoch": 0.08604878564444092, "grad_norm": 4.944527626037598, "learning_rate": 1.9838269529129456e-05, "loss": 5.5649, "step": 2442 }, { "epoch": 0.08608402265739933, "grad_norm": 12.8356351852417, "learning_rate": 1.983806503779715e-05, "loss": 10.3653, "step": 2443 }, { "epoch": 0.08611925967035775, "grad_norm": 3.590506076812744, "learning_rate": 1.983786041832279e-05, "loss": 5.8368, "step": 2444 }, { "epoch": 0.08615449668331615, "grad_norm": 2.1475284099578857, "learning_rate": 1.9837655670709045e-05, "loss": 8.4917, "step": 2445 }, { "epoch": 0.08618973369627457, "grad_norm": 3.4861903190612793, "learning_rate": 1.9837450794958582e-05, "loss": 6.1964, "step": 2446 }, { "epoch": 0.08622497070923298, "grad_norm": 3.5153560638427734, "learning_rate": 1.9837245791074075e-05, "loss": 9.7264, "step": 2447 }, { "epoch": 0.0862602077221914, "grad_norm": 4.168884754180908, "learning_rate": 1.9837040659058182e-05, "loss": 9.845, "step": 2448 }, { "epoch": 0.0862954447351498, "grad_norm": 5.064790725708008, "learning_rate": 1.9836835398913584e-05, "loss": 6.95, "step": 2449 }, { "epoch": 0.0863306817481082, "grad_norm": 3.872648239135742, "learning_rate": 1.9836630010642953e-05, "loss": 6.7652, "step": 2450 }, { "epoch": 0.08636591876106663, "grad_norm": 5.905561923980713, "learning_rate": 1.983642449424896e-05, "loss": 12.3474, "step": 2451 }, { "epoch": 0.08640115577402503, "grad_norm": 5.031739234924316, "learning_rate": 1.983621884973429e-05, "loss": 11.928, "step": 2452 }, { "epoch": 0.08643639278698345, "grad_norm": 3.4191582202911377, "learning_rate": 1.983601307710161e-05, "loss": 6.3599, "step": 2453 }, { "epoch": 0.08647162979994186, "grad_norm": 5.241065502166748, "learning_rate": 1.983580717635361e-05, "loss": 11.8291, "step": 2454 }, { "epoch": 0.08650686681290028, "grad_norm": 2.8494889736175537, "learning_rate": 1.983560114749297e-05, "loss": 5.9924, "step": 2455 }, { "epoch": 0.08654210382585868, "grad_norm": 4.189448833465576, "learning_rate": 1.9835394990522368e-05, "loss": 8.6508, "step": 2456 }, { "epoch": 0.08657734083881709, "grad_norm": 6.284221172332764, "learning_rate": 1.98351887054445e-05, "loss": 6.4322, "step": 2457 }, { "epoch": 0.08661257785177551, "grad_norm": 4.161053657531738, "learning_rate": 1.983498229226204e-05, "loss": 8.7491, "step": 2458 }, { "epoch": 0.08664781486473391, "grad_norm": 7.797457218170166, "learning_rate": 1.9834775750977687e-05, "loss": 15.2977, "step": 2459 }, { "epoch": 0.08668305187769233, "grad_norm": 11.220661163330078, "learning_rate": 1.9834569081594125e-05, "loss": 9.1833, "step": 2460 }, { "epoch": 0.08671828889065074, "grad_norm": 3.0312516689300537, "learning_rate": 1.9834362284114048e-05, "loss": 6.3301, "step": 2461 }, { "epoch": 0.08675352590360916, "grad_norm": 3.818690538406372, "learning_rate": 1.983415535854015e-05, "loss": 6.3578, "step": 2462 }, { "epoch": 0.08678876291656756, "grad_norm": 4.431600570678711, "learning_rate": 1.9833948304875125e-05, "loss": 9.4546, "step": 2463 }, { "epoch": 0.08682399992952597, "grad_norm": 6.0265374183654785, "learning_rate": 1.9833741123121672e-05, "loss": 12.4814, "step": 2464 }, { "epoch": 0.08685923694248439, "grad_norm": 3.526989221572876, "learning_rate": 1.983353381328249e-05, "loss": 5.9008, "step": 2465 }, { "epoch": 0.0868944739554428, "grad_norm": 2.8351621627807617, "learning_rate": 1.9833326375360272e-05, "loss": 5.9765, "step": 2466 }, { "epoch": 0.08692971096840121, "grad_norm": 4.757086277008057, "learning_rate": 1.983311880935773e-05, "loss": 5.2589, "step": 2467 }, { "epoch": 0.08696494798135962, "grad_norm": 2.8085575103759766, "learning_rate": 1.983291111527756e-05, "loss": 6.449, "step": 2468 }, { "epoch": 0.08700018499431802, "grad_norm": 2.7987074851989746, "learning_rate": 1.9832703293122472e-05, "loss": 5.9863, "step": 2469 }, { "epoch": 0.08703542200727644, "grad_norm": 3.1042354106903076, "learning_rate": 1.983249534289517e-05, "loss": 6.2911, "step": 2470 }, { "epoch": 0.08707065902023485, "grad_norm": 4.209139347076416, "learning_rate": 1.9832287264598365e-05, "loss": 9.5304, "step": 2471 }, { "epoch": 0.08710589603319327, "grad_norm": 3.3336565494537354, "learning_rate": 1.9832079058234764e-05, "loss": 8.8408, "step": 2472 }, { "epoch": 0.08714113304615168, "grad_norm": 2.826763153076172, "learning_rate": 1.9831870723807086e-05, "loss": 9.1603, "step": 2473 }, { "epoch": 0.0871763700591101, "grad_norm": 2.795421838760376, "learning_rate": 1.9831662261318034e-05, "loss": 6.5627, "step": 2474 }, { "epoch": 0.0872116070720685, "grad_norm": 2.1556169986724854, "learning_rate": 1.9831453670770334e-05, "loss": 6.0885, "step": 2475 }, { "epoch": 0.0872468440850269, "grad_norm": 6.46539831161499, "learning_rate": 1.983124495216669e-05, "loss": 13.3742, "step": 2476 }, { "epoch": 0.08728208109798533, "grad_norm": 2.575937271118164, "learning_rate": 1.9831036105509835e-05, "loss": 7.0612, "step": 2477 }, { "epoch": 0.08731731811094373, "grad_norm": 5.887486457824707, "learning_rate": 1.983082713080248e-05, "loss": 6.0371, "step": 2478 }, { "epoch": 0.08735255512390215, "grad_norm": 5.178842067718506, "learning_rate": 1.9830618028047347e-05, "loss": 8.7077, "step": 2479 }, { "epoch": 0.08738779213686056, "grad_norm": 4.309187889099121, "learning_rate": 1.9830408797247167e-05, "loss": 8.4385, "step": 2480 }, { "epoch": 0.08742302914981898, "grad_norm": 5.8315653800964355, "learning_rate": 1.9830199438404656e-05, "loss": 12.6778, "step": 2481 }, { "epoch": 0.08745826616277738, "grad_norm": 4.913346290588379, "learning_rate": 1.9829989951522547e-05, "loss": 8.7965, "step": 2482 }, { "epoch": 0.08749350317573579, "grad_norm": 3.757371425628662, "learning_rate": 1.9829780336603564e-05, "loss": 5.8105, "step": 2483 }, { "epoch": 0.0875287401886942, "grad_norm": 6.3450798988342285, "learning_rate": 1.9829570593650446e-05, "loss": 6.7287, "step": 2484 }, { "epoch": 0.08756397720165261, "grad_norm": 5.083843231201172, "learning_rate": 1.9829360722665913e-05, "loss": 11.3953, "step": 2485 }, { "epoch": 0.08759921421461103, "grad_norm": 5.34736967086792, "learning_rate": 1.9829150723652705e-05, "loss": 9.0367, "step": 2486 }, { "epoch": 0.08763445122756944, "grad_norm": 4.305384159088135, "learning_rate": 1.982894059661356e-05, "loss": 10.1805, "step": 2487 }, { "epoch": 0.08766968824052786, "grad_norm": 5.16099214553833, "learning_rate": 1.982873034155121e-05, "loss": 6.3236, "step": 2488 }, { "epoch": 0.08770492525348626, "grad_norm": 6.994993686676025, "learning_rate": 1.9828519958468395e-05, "loss": 9.097, "step": 2489 }, { "epoch": 0.08774016226644467, "grad_norm": 4.481929302215576, "learning_rate": 1.9828309447367857e-05, "loss": 6.2833, "step": 2490 }, { "epoch": 0.08777539927940309, "grad_norm": 7.906358242034912, "learning_rate": 1.9828098808252335e-05, "loss": 9.791, "step": 2491 }, { "epoch": 0.0878106362923615, "grad_norm": 4.850246906280518, "learning_rate": 1.9827888041124575e-05, "loss": 9.0805, "step": 2492 }, { "epoch": 0.08784587330531991, "grad_norm": 2.748190402984619, "learning_rate": 1.982767714598732e-05, "loss": 9.0127, "step": 2493 }, { "epoch": 0.08788111031827832, "grad_norm": 3.4471547603607178, "learning_rate": 1.982746612284332e-05, "loss": 9.0553, "step": 2494 }, { "epoch": 0.08791634733123672, "grad_norm": 4.359870433807373, "learning_rate": 1.9827254971695323e-05, "loss": 5.8506, "step": 2495 }, { "epoch": 0.08795158434419514, "grad_norm": 3.9710988998413086, "learning_rate": 1.9827043692546077e-05, "loss": 9.0897, "step": 2496 }, { "epoch": 0.08798682135715355, "grad_norm": 3.933814764022827, "learning_rate": 1.9826832285398335e-05, "loss": 5.7392, "step": 2497 }, { "epoch": 0.08802205837011197, "grad_norm": 3.562778949737549, "learning_rate": 1.982662075025485e-05, "loss": 7.8423, "step": 2498 }, { "epoch": 0.08805729538307037, "grad_norm": 4.900052070617676, "learning_rate": 1.982640908711838e-05, "loss": 9.3639, "step": 2499 }, { "epoch": 0.0880925323960288, "grad_norm": 4.42734956741333, "learning_rate": 1.9826197295991677e-05, "loss": 5.4749, "step": 2500 }, { "epoch": 0.0881277694089872, "grad_norm": 3.943390369415283, "learning_rate": 1.9825985376877503e-05, "loss": 8.1482, "step": 2501 }, { "epoch": 0.0881630064219456, "grad_norm": 4.5685505867004395, "learning_rate": 1.982577332977862e-05, "loss": 6.0936, "step": 2502 }, { "epoch": 0.08819824343490403, "grad_norm": 4.060177803039551, "learning_rate": 1.9825561154697786e-05, "loss": 8.8269, "step": 2503 }, { "epoch": 0.08823348044786243, "grad_norm": 5.02520227432251, "learning_rate": 1.9825348851637767e-05, "loss": 8.7988, "step": 2504 }, { "epoch": 0.08826871746082085, "grad_norm": 4.018466949462891, "learning_rate": 1.9825136420601328e-05, "loss": 8.1561, "step": 2505 }, { "epoch": 0.08830395447377926, "grad_norm": 2.9581961631774902, "learning_rate": 1.9824923861591238e-05, "loss": 5.6414, "step": 2506 }, { "epoch": 0.08833919148673768, "grad_norm": 3.9611399173736572, "learning_rate": 1.982471117461026e-05, "loss": 8.5169, "step": 2507 }, { "epoch": 0.08837442849969608, "grad_norm": 5.026424407958984, "learning_rate": 1.9824498359661164e-05, "loss": 11.7249, "step": 2508 }, { "epoch": 0.08840966551265449, "grad_norm": 5.322770118713379, "learning_rate": 1.982428541674673e-05, "loss": 5.8622, "step": 2509 }, { "epoch": 0.0884449025256129, "grad_norm": 5.075022220611572, "learning_rate": 1.9824072345869728e-05, "loss": 8.5856, "step": 2510 }, { "epoch": 0.08848013953857131, "grad_norm": 4.887312889099121, "learning_rate": 1.9823859147032932e-05, "loss": 10.65, "step": 2511 }, { "epoch": 0.08851537655152973, "grad_norm": 5.785980224609375, "learning_rate": 1.9823645820239114e-05, "loss": 6.9182, "step": 2512 }, { "epoch": 0.08855061356448814, "grad_norm": 8.548970222473145, "learning_rate": 1.9823432365491066e-05, "loss": 9.0312, "step": 2513 }, { "epoch": 0.08858585057744656, "grad_norm": 4.086949825286865, "learning_rate": 1.9823218782791552e-05, "loss": 6.3382, "step": 2514 }, { "epoch": 0.08862108759040496, "grad_norm": 3.401010036468506, "learning_rate": 1.982300507214337e-05, "loss": 6.519, "step": 2515 }, { "epoch": 0.08865632460336337, "grad_norm": 7.164829730987549, "learning_rate": 1.982279123354929e-05, "loss": 8.0201, "step": 2516 }, { "epoch": 0.08869156161632179, "grad_norm": 3.4590024948120117, "learning_rate": 1.9822577267012103e-05, "loss": 7.1324, "step": 2517 }, { "epoch": 0.08872679862928019, "grad_norm": 2.6103439331054688, "learning_rate": 1.98223631725346e-05, "loss": 6.0105, "step": 2518 }, { "epoch": 0.08876203564223861, "grad_norm": 2.729761838912964, "learning_rate": 1.9822148950119563e-05, "loss": 6.113, "step": 2519 }, { "epoch": 0.08879727265519702, "grad_norm": 3.2323601245880127, "learning_rate": 1.9821934599769787e-05, "loss": 8.1693, "step": 2520 }, { "epoch": 0.08883250966815542, "grad_norm": 4.168993949890137, "learning_rate": 1.9821720121488054e-05, "loss": 8.668, "step": 2521 }, { "epoch": 0.08886774668111384, "grad_norm": 4.164851188659668, "learning_rate": 1.9821505515277173e-05, "loss": 6.1728, "step": 2522 }, { "epoch": 0.08890298369407225, "grad_norm": 4.220306873321533, "learning_rate": 1.982129078113993e-05, "loss": 5.7863, "step": 2523 }, { "epoch": 0.08893822070703067, "grad_norm": 5.0011491775512695, "learning_rate": 1.982107591907912e-05, "loss": 9.5154, "step": 2524 }, { "epoch": 0.08897345771998907, "grad_norm": 5.31564474105835, "learning_rate": 1.982086092909755e-05, "loss": 5.6809, "step": 2525 }, { "epoch": 0.0890086947329475, "grad_norm": 6.9058427810668945, "learning_rate": 1.9820645811198017e-05, "loss": 5.2347, "step": 2526 }, { "epoch": 0.0890439317459059, "grad_norm": 2.0804905891418457, "learning_rate": 1.9820430565383314e-05, "loss": 5.5727, "step": 2527 }, { "epoch": 0.0890791687588643, "grad_norm": 5.521117687225342, "learning_rate": 1.982021519165626e-05, "loss": 6.9458, "step": 2528 }, { "epoch": 0.08911440577182272, "grad_norm": 3.770317792892456, "learning_rate": 1.9819999690019646e-05, "loss": 5.9089, "step": 2529 }, { "epoch": 0.08914964278478113, "grad_norm": 4.17829704284668, "learning_rate": 1.9819784060476288e-05, "loss": 6.1124, "step": 2530 }, { "epoch": 0.08918487979773955, "grad_norm": 4.528212547302246, "learning_rate": 1.981956830302899e-05, "loss": 7.9876, "step": 2531 }, { "epoch": 0.08922011681069796, "grad_norm": 4.826838493347168, "learning_rate": 1.9819352417680566e-05, "loss": 5.7922, "step": 2532 }, { "epoch": 0.08925535382365637, "grad_norm": 4.1182026863098145, "learning_rate": 1.9819136404433824e-05, "loss": 8.8496, "step": 2533 }, { "epoch": 0.08929059083661478, "grad_norm": 3.1821327209472656, "learning_rate": 1.9818920263291585e-05, "loss": 6.6235, "step": 2534 }, { "epoch": 0.08932582784957319, "grad_norm": 5.700937271118164, "learning_rate": 1.981870399425665e-05, "loss": 5.3974, "step": 2535 }, { "epoch": 0.0893610648625316, "grad_norm": 4.860317230224609, "learning_rate": 1.9818487597331854e-05, "loss": 6.7876, "step": 2536 }, { "epoch": 0.08939630187549001, "grad_norm": 4.315637588500977, "learning_rate": 1.981827107252e-05, "loss": 8.0724, "step": 2537 }, { "epoch": 0.08943153888844843, "grad_norm": 3.5613789558410645, "learning_rate": 1.9818054419823915e-05, "loss": 8.718, "step": 2538 }, { "epoch": 0.08946677590140684, "grad_norm": 2.621281623840332, "learning_rate": 1.981783763924642e-05, "loss": 5.5082, "step": 2539 }, { "epoch": 0.08950201291436526, "grad_norm": 2.2676587104797363, "learning_rate": 1.9817620730790343e-05, "loss": 5.4491, "step": 2540 }, { "epoch": 0.08953724992732366, "grad_norm": 4.4519524574279785, "learning_rate": 1.9817403694458504e-05, "loss": 8.6058, "step": 2541 }, { "epoch": 0.08957248694028207, "grad_norm": 5.238862037658691, "learning_rate": 1.981718653025373e-05, "loss": 6.1558, "step": 2542 }, { "epoch": 0.08960772395324049, "grad_norm": 4.49267578125, "learning_rate": 1.981696923817885e-05, "loss": 8.8715, "step": 2543 }, { "epoch": 0.08964296096619889, "grad_norm": 2.532457113265991, "learning_rate": 1.98167518182367e-05, "loss": 5.4873, "step": 2544 }, { "epoch": 0.08967819797915731, "grad_norm": 7.455324649810791, "learning_rate": 1.9816534270430104e-05, "loss": 14.1214, "step": 2545 }, { "epoch": 0.08971343499211572, "grad_norm": 3.3474984169006348, "learning_rate": 1.9816316594761896e-05, "loss": 9.6782, "step": 2546 }, { "epoch": 0.08974867200507412, "grad_norm": 3.9851958751678467, "learning_rate": 1.9816098791234916e-05, "loss": 9.533, "step": 2547 }, { "epoch": 0.08978390901803254, "grad_norm": 5.248547554016113, "learning_rate": 1.9815880859852005e-05, "loss": 5.6743, "step": 2548 }, { "epoch": 0.08981914603099095, "grad_norm": 4.990542888641357, "learning_rate": 1.9815662800615986e-05, "loss": 5.9989, "step": 2549 }, { "epoch": 0.08985438304394937, "grad_norm": 3.714938163757324, "learning_rate": 1.9815444613529715e-05, "loss": 7.9138, "step": 2550 }, { "epoch": 0.08988962005690777, "grad_norm": 6.104117393493652, "learning_rate": 1.9815226298596025e-05, "loss": 6.9127, "step": 2551 }, { "epoch": 0.0899248570698662, "grad_norm": 2.2115442752838135, "learning_rate": 1.981500785581776e-05, "loss": 9.2955, "step": 2552 }, { "epoch": 0.0899600940828246, "grad_norm": 6.751791954040527, "learning_rate": 1.9814789285197774e-05, "loss": 14.015, "step": 2553 }, { "epoch": 0.089995331095783, "grad_norm": 2.1982595920562744, "learning_rate": 1.9814570586738902e-05, "loss": 5.9923, "step": 2554 }, { "epoch": 0.09003056810874142, "grad_norm": 3.0341951847076416, "learning_rate": 1.9814351760444e-05, "loss": 8.3146, "step": 2555 }, { "epoch": 0.09006580512169983, "grad_norm": 3.7031948566436768, "learning_rate": 1.9814132806315912e-05, "loss": 9.3229, "step": 2556 }, { "epoch": 0.09010104213465825, "grad_norm": 2.4924750328063965, "learning_rate": 1.98139137243575e-05, "loss": 8.6126, "step": 2557 }, { "epoch": 0.09013627914761665, "grad_norm": 4.1807403564453125, "learning_rate": 1.9813694514571607e-05, "loss": 11.8093, "step": 2558 }, { "epoch": 0.09017151616057507, "grad_norm": 5.071283340454102, "learning_rate": 1.9813475176961098e-05, "loss": 5.7133, "step": 2559 }, { "epoch": 0.09020675317353348, "grad_norm": 4.113607406616211, "learning_rate": 1.981325571152882e-05, "loss": 5.5718, "step": 2560 }, { "epoch": 0.09024199018649189, "grad_norm": 2.2795331478118896, "learning_rate": 1.9813036118277638e-05, "loss": 8.4728, "step": 2561 }, { "epoch": 0.0902772271994503, "grad_norm": 4.403189659118652, "learning_rate": 1.9812816397210414e-05, "loss": 8.4269, "step": 2562 }, { "epoch": 0.09031246421240871, "grad_norm": 7.99702787399292, "learning_rate": 1.981259654833e-05, "loss": 6.1982, "step": 2563 }, { "epoch": 0.09034770122536713, "grad_norm": 2.8951737880706787, "learning_rate": 1.981237657163927e-05, "loss": 8.0102, "step": 2564 }, { "epoch": 0.09038293823832554, "grad_norm": 5.661139965057373, "learning_rate": 1.9812156467141082e-05, "loss": 13.245, "step": 2565 }, { "epoch": 0.09041817525128396, "grad_norm": 2.4572460651397705, "learning_rate": 1.981193623483831e-05, "loss": 7.3182, "step": 2566 }, { "epoch": 0.09045341226424236, "grad_norm": 4.450530529022217, "learning_rate": 1.9811715874733818e-05, "loss": 5.6663, "step": 2567 }, { "epoch": 0.09048864927720077, "grad_norm": 7.97666597366333, "learning_rate": 1.981149538683048e-05, "loss": 9.3484, "step": 2568 }, { "epoch": 0.09052388629015919, "grad_norm": 4.197483062744141, "learning_rate": 1.981127477113116e-05, "loss": 8.2191, "step": 2569 }, { "epoch": 0.09055912330311759, "grad_norm": 4.6316914558410645, "learning_rate": 1.9811054027638734e-05, "loss": 8.091, "step": 2570 }, { "epoch": 0.09059436031607601, "grad_norm": 2.5614750385284424, "learning_rate": 1.9810833156356086e-05, "loss": 8.5803, "step": 2571 }, { "epoch": 0.09062959732903442, "grad_norm": 3.5787365436553955, "learning_rate": 1.9810612157286082e-05, "loss": 5.9338, "step": 2572 }, { "epoch": 0.09066483434199282, "grad_norm": 3.487353563308716, "learning_rate": 1.9810391030431605e-05, "loss": 8.5071, "step": 2573 }, { "epoch": 0.09070007135495124, "grad_norm": 3.429420232772827, "learning_rate": 1.981016977579554e-05, "loss": 11.2396, "step": 2574 }, { "epoch": 0.09073530836790965, "grad_norm": 2.7065274715423584, "learning_rate": 1.980994839338076e-05, "loss": 8.1082, "step": 2575 }, { "epoch": 0.09077054538086807, "grad_norm": 2.8690459728240967, "learning_rate": 1.9809726883190153e-05, "loss": 10.6617, "step": 2576 }, { "epoch": 0.09080578239382647, "grad_norm": 6.48087215423584, "learning_rate": 1.9809505245226603e-05, "loss": 5.4084, "step": 2577 }, { "epoch": 0.09084101940678489, "grad_norm": 2.555077075958252, "learning_rate": 1.9809283479493e-05, "loss": 5.8627, "step": 2578 }, { "epoch": 0.0908762564197433, "grad_norm": 3.5418434143066406, "learning_rate": 1.980906158599223e-05, "loss": 5.1836, "step": 2579 }, { "epoch": 0.0909114934327017, "grad_norm": 3.8272247314453125, "learning_rate": 1.9808839564727182e-05, "loss": 6.1439, "step": 2580 }, { "epoch": 0.09094673044566012, "grad_norm": 3.907687187194824, "learning_rate": 1.9808617415700747e-05, "loss": 10.8201, "step": 2581 }, { "epoch": 0.09098196745861853, "grad_norm": 5.446732521057129, "learning_rate": 1.9808395138915824e-05, "loss": 5.3658, "step": 2582 }, { "epoch": 0.09101720447157695, "grad_norm": 4.834763050079346, "learning_rate": 1.9808172734375303e-05, "loss": 6.2484, "step": 2583 }, { "epoch": 0.09105244148453535, "grad_norm": 5.866298198699951, "learning_rate": 1.9807950202082085e-05, "loss": 11.6948, "step": 2584 }, { "epoch": 0.09108767849749377, "grad_norm": 3.4108831882476807, "learning_rate": 1.9807727542039067e-05, "loss": 9.2746, "step": 2585 }, { "epoch": 0.09112291551045218, "grad_norm": 3.2534494400024414, "learning_rate": 1.9807504754249142e-05, "loss": 7.9734, "step": 2586 }, { "epoch": 0.09115815252341058, "grad_norm": 3.1984808444976807, "learning_rate": 1.9807281838715227e-05, "loss": 5.0234, "step": 2587 }, { "epoch": 0.091193389536369, "grad_norm": 8.630814552307129, "learning_rate": 1.980705879544021e-05, "loss": 6.4881, "step": 2588 }, { "epoch": 0.09122862654932741, "grad_norm": 3.1285250186920166, "learning_rate": 1.9806835624427004e-05, "loss": 5.3745, "step": 2589 }, { "epoch": 0.09126386356228583, "grad_norm": 6.272480010986328, "learning_rate": 1.9806612325678517e-05, "loss": 5.969, "step": 2590 }, { "epoch": 0.09129910057524424, "grad_norm": 2.781111240386963, "learning_rate": 1.980638889919766e-05, "loss": 5.8526, "step": 2591 }, { "epoch": 0.09133433758820265, "grad_norm": 2.7416019439697266, "learning_rate": 1.980616534498733e-05, "loss": 5.7503, "step": 2592 }, { "epoch": 0.09136957460116106, "grad_norm": 4.527207374572754, "learning_rate": 1.980594166305045e-05, "loss": 5.6129, "step": 2593 }, { "epoch": 0.09140481161411947, "grad_norm": 7.770206451416016, "learning_rate": 1.9805717853389932e-05, "loss": 4.8097, "step": 2594 }, { "epoch": 0.09144004862707789, "grad_norm": 2.576413869857788, "learning_rate": 1.980549391600869e-05, "loss": 6.7265, "step": 2595 }, { "epoch": 0.09147528564003629, "grad_norm": 6.809453010559082, "learning_rate": 1.9805269850909638e-05, "loss": 12.3805, "step": 2596 }, { "epoch": 0.09151052265299471, "grad_norm": 3.751664400100708, "learning_rate": 1.98050456580957e-05, "loss": 5.1169, "step": 2597 }, { "epoch": 0.09154575966595312, "grad_norm": 3.995809555053711, "learning_rate": 1.9804821337569793e-05, "loss": 8.2453, "step": 2598 }, { "epoch": 0.09158099667891152, "grad_norm": 7.491372108459473, "learning_rate": 1.980459688933484e-05, "loss": 6.8645, "step": 2599 }, { "epoch": 0.09161623369186994, "grad_norm": 3.5211551189422607, "learning_rate": 1.980437231339376e-05, "loss": 5.6515, "step": 2600 }, { "epoch": 0.09165147070482835, "grad_norm": 3.397296905517578, "learning_rate": 1.9804147609749485e-05, "loss": 6.6598, "step": 2601 }, { "epoch": 0.09168670771778677, "grad_norm": 4.846439838409424, "learning_rate": 1.9803922778404937e-05, "loss": 5.6315, "step": 2602 }, { "epoch": 0.09172194473074517, "grad_norm": 5.350329399108887, "learning_rate": 1.9803697819363047e-05, "loss": 8.585, "step": 2603 }, { "epoch": 0.09175718174370359, "grad_norm": 2.6230664253234863, "learning_rate": 1.9803472732626743e-05, "loss": 6.2043, "step": 2604 }, { "epoch": 0.091792418756662, "grad_norm": 5.198380947113037, "learning_rate": 1.9803247518198957e-05, "loss": 5.8635, "step": 2605 }, { "epoch": 0.0918276557696204, "grad_norm": 3.4468445777893066, "learning_rate": 1.9803022176082625e-05, "loss": 8.0227, "step": 2606 }, { "epoch": 0.09186289278257882, "grad_norm": 4.152608871459961, "learning_rate": 1.980279670628068e-05, "loss": 7.849, "step": 2607 }, { "epoch": 0.09189812979553723, "grad_norm": 5.104734420776367, "learning_rate": 1.980257110879606e-05, "loss": 9.2787, "step": 2608 }, { "epoch": 0.09193336680849565, "grad_norm": 3.688460111618042, "learning_rate": 1.9802345383631706e-05, "loss": 8.6148, "step": 2609 }, { "epoch": 0.09196860382145405, "grad_norm": 7.608479976654053, "learning_rate": 1.980211953079055e-05, "loss": 6.2772, "step": 2610 }, { "epoch": 0.09200384083441247, "grad_norm": 6.77382755279541, "learning_rate": 1.980189355027554e-05, "loss": 6.8327, "step": 2611 }, { "epoch": 0.09203907784737088, "grad_norm": 2.4633750915527344, "learning_rate": 1.9801667442089613e-05, "loss": 8.3538, "step": 2612 }, { "epoch": 0.09207431486032928, "grad_norm": 3.249006509780884, "learning_rate": 1.9801441206235726e-05, "loss": 5.6699, "step": 2613 }, { "epoch": 0.0921095518732877, "grad_norm": 6.733852386474609, "learning_rate": 1.9801214842716817e-05, "loss": 8.4312, "step": 2614 }, { "epoch": 0.09214478888624611, "grad_norm": 3.618844509124756, "learning_rate": 1.9800988351535833e-05, "loss": 6.0182, "step": 2615 }, { "epoch": 0.09218002589920453, "grad_norm": 2.4387569427490234, "learning_rate": 1.980076173269573e-05, "loss": 5.5601, "step": 2616 }, { "epoch": 0.09221526291216293, "grad_norm": 7.809319972991943, "learning_rate": 1.980053498619946e-05, "loss": 10.5199, "step": 2617 }, { "epoch": 0.09225049992512135, "grad_norm": 4.920675754547119, "learning_rate": 1.980030811204997e-05, "loss": 9.9482, "step": 2618 }, { "epoch": 0.09228573693807976, "grad_norm": 5.624327182769775, "learning_rate": 1.9800081110250217e-05, "loss": 8.012, "step": 2619 }, { "epoch": 0.09232097395103817, "grad_norm": 3.8333353996276855, "learning_rate": 1.979985398080316e-05, "loss": 8.7902, "step": 2620 }, { "epoch": 0.09235621096399659, "grad_norm": 2.6008107662200928, "learning_rate": 1.9799626723711757e-05, "loss": 9.0332, "step": 2621 }, { "epoch": 0.09239144797695499, "grad_norm": 4.8351664543151855, "learning_rate": 1.979939933897897e-05, "loss": 10.9561, "step": 2622 }, { "epoch": 0.09242668498991341, "grad_norm": 2.8908886909484863, "learning_rate": 1.9799171826607756e-05, "loss": 11.1244, "step": 2623 }, { "epoch": 0.09246192200287182, "grad_norm": 3.3006200790405273, "learning_rate": 1.979894418660108e-05, "loss": 5.8444, "step": 2624 }, { "epoch": 0.09249715901583022, "grad_norm": 2.283996343612671, "learning_rate": 1.9798716418961907e-05, "loss": 6.1553, "step": 2625 }, { "epoch": 0.09253239602878864, "grad_norm": 4.414186477661133, "learning_rate": 1.9798488523693208e-05, "loss": 5.4944, "step": 2626 }, { "epoch": 0.09256763304174705, "grad_norm": 2.4764156341552734, "learning_rate": 1.9798260500797947e-05, "loss": 5.9028, "step": 2627 }, { "epoch": 0.09260287005470547, "grad_norm": 11.78494930267334, "learning_rate": 1.979803235027909e-05, "loss": 8.821, "step": 2628 }, { "epoch": 0.09263810706766387, "grad_norm": 7.074458599090576, "learning_rate": 1.979780407213962e-05, "loss": 14.3501, "step": 2629 }, { "epoch": 0.09267334408062229, "grad_norm": 5.749920845031738, "learning_rate": 1.9797575666382504e-05, "loss": 11.2034, "step": 2630 }, { "epoch": 0.0927085810935807, "grad_norm": 3.283522129058838, "learning_rate": 1.9797347133010715e-05, "loss": 5.2979, "step": 2631 }, { "epoch": 0.0927438181065391, "grad_norm": 4.509775161743164, "learning_rate": 1.979711847202723e-05, "loss": 11.4172, "step": 2632 }, { "epoch": 0.09277905511949752, "grad_norm": 3.0575144290924072, "learning_rate": 1.979688968343503e-05, "loss": 6.1587, "step": 2633 }, { "epoch": 0.09281429213245593, "grad_norm": 3.7961034774780273, "learning_rate": 1.9796660767237094e-05, "loss": 5.7607, "step": 2634 }, { "epoch": 0.09284952914541435, "grad_norm": 6.878536224365234, "learning_rate": 1.9796431723436404e-05, "loss": 11.9908, "step": 2635 }, { "epoch": 0.09288476615837275, "grad_norm": 5.228306770324707, "learning_rate": 1.979620255203594e-05, "loss": 5.8715, "step": 2636 }, { "epoch": 0.09292000317133117, "grad_norm": 3.180375576019287, "learning_rate": 1.9795973253038693e-05, "loss": 9.0552, "step": 2637 }, { "epoch": 0.09295524018428958, "grad_norm": 4.975851535797119, "learning_rate": 1.979574382644765e-05, "loss": 5.4159, "step": 2638 }, { "epoch": 0.09299047719724798, "grad_norm": 3.281801223754883, "learning_rate": 1.979551427226579e-05, "loss": 7.7414, "step": 2639 }, { "epoch": 0.0930257142102064, "grad_norm": 4.139869689941406, "learning_rate": 1.979528459049611e-05, "loss": 8.5344, "step": 2640 }, { "epoch": 0.09306095122316481, "grad_norm": 4.028489112854004, "learning_rate": 1.9795054781141604e-05, "loss": 11.0278, "step": 2641 }, { "epoch": 0.09309618823612323, "grad_norm": 4.666816234588623, "learning_rate": 1.9794824844205254e-05, "loss": 11.2533, "step": 2642 }, { "epoch": 0.09313142524908163, "grad_norm": 3.9041125774383545, "learning_rate": 1.979459477969007e-05, "loss": 10.0556, "step": 2643 }, { "epoch": 0.09316666226204005, "grad_norm": 4.965904235839844, "learning_rate": 1.979436458759904e-05, "loss": 8.95, "step": 2644 }, { "epoch": 0.09320189927499846, "grad_norm": 2.3052773475646973, "learning_rate": 1.9794134267935163e-05, "loss": 5.5814, "step": 2645 }, { "epoch": 0.09323713628795686, "grad_norm": 5.834864616394043, "learning_rate": 1.9793903820701443e-05, "loss": 10.4647, "step": 2646 }, { "epoch": 0.09327237330091528, "grad_norm": 3.210604190826416, "learning_rate": 1.9793673245900874e-05, "loss": 5.6228, "step": 2647 }, { "epoch": 0.09330761031387369, "grad_norm": 4.173891544342041, "learning_rate": 1.9793442543536462e-05, "loss": 8.1339, "step": 2648 }, { "epoch": 0.09334284732683211, "grad_norm": 4.158992767333984, "learning_rate": 1.9793211713611217e-05, "loss": 9.0026, "step": 2649 }, { "epoch": 0.09337808433979052, "grad_norm": 2.338571071624756, "learning_rate": 1.979298075612814e-05, "loss": 6.1211, "step": 2650 }, { "epoch": 0.09341332135274892, "grad_norm": 2.4406790733337402, "learning_rate": 1.979274967109024e-05, "loss": 6.3996, "step": 2651 }, { "epoch": 0.09344855836570734, "grad_norm": 3.268599510192871, "learning_rate": 1.979251845850053e-05, "loss": 8.5137, "step": 2652 }, { "epoch": 0.09348379537866575, "grad_norm": 4.475321292877197, "learning_rate": 1.979228711836202e-05, "loss": 6.4014, "step": 2653 }, { "epoch": 0.09351903239162417, "grad_norm": 3.6764681339263916, "learning_rate": 1.9792055650677723e-05, "loss": 5.5793, "step": 2654 }, { "epoch": 0.09355426940458257, "grad_norm": 2.6741323471069336, "learning_rate": 1.979182405545065e-05, "loss": 8.5902, "step": 2655 }, { "epoch": 0.09358950641754099, "grad_norm": 2.87005877494812, "learning_rate": 1.9791592332683825e-05, "loss": 8.8719, "step": 2656 }, { "epoch": 0.0936247434304994, "grad_norm": 5.1238017082214355, "learning_rate": 1.9791360482380265e-05, "loss": 6.4472, "step": 2657 }, { "epoch": 0.0936599804434578, "grad_norm": 2.6431002616882324, "learning_rate": 1.9791128504542986e-05, "loss": 8.9744, "step": 2658 }, { "epoch": 0.09369521745641622, "grad_norm": 3.571063280105591, "learning_rate": 1.979089639917501e-05, "loss": 8.7267, "step": 2659 }, { "epoch": 0.09373045446937463, "grad_norm": 2.7948997020721436, "learning_rate": 1.979066416627936e-05, "loss": 8.2511, "step": 2660 }, { "epoch": 0.09376569148233305, "grad_norm": 2.9887990951538086, "learning_rate": 1.9790431805859063e-05, "loss": 5.1269, "step": 2661 }, { "epoch": 0.09380092849529145, "grad_norm": 3.4637908935546875, "learning_rate": 1.9790199317917143e-05, "loss": 8.1638, "step": 2662 }, { "epoch": 0.09383616550824987, "grad_norm": 3.8864810466766357, "learning_rate": 1.9789966702456633e-05, "loss": 10.6379, "step": 2663 }, { "epoch": 0.09387140252120828, "grad_norm": 4.141059398651123, "learning_rate": 1.9789733959480557e-05, "loss": 6.0467, "step": 2664 }, { "epoch": 0.09390663953416668, "grad_norm": 3.033611297607422, "learning_rate": 1.978950108899195e-05, "loss": 8.5047, "step": 2665 }, { "epoch": 0.0939418765471251, "grad_norm": 3.582942485809326, "learning_rate": 1.9789268090993845e-05, "loss": 5.8877, "step": 2666 }, { "epoch": 0.09397711356008351, "grad_norm": 3.063575506210327, "learning_rate": 1.9789034965489278e-05, "loss": 6.0331, "step": 2667 }, { "epoch": 0.09401235057304193, "grad_norm": 4.40071439743042, "learning_rate": 1.978880171248128e-05, "loss": 8.28, "step": 2668 }, { "epoch": 0.09404758758600033, "grad_norm": 3.1036548614501953, "learning_rate": 1.9788568331972895e-05, "loss": 5.2785, "step": 2669 }, { "epoch": 0.09408282459895875, "grad_norm": 2.419748544692993, "learning_rate": 1.978833482396716e-05, "loss": 5.7699, "step": 2670 }, { "epoch": 0.09411806161191716, "grad_norm": 5.463094711303711, "learning_rate": 1.978810118846712e-05, "loss": 6.0686, "step": 2671 }, { "epoch": 0.09415329862487556, "grad_norm": 4.685303211212158, "learning_rate": 1.9787867425475807e-05, "loss": 5.6205, "step": 2672 }, { "epoch": 0.09418853563783398, "grad_norm": 3.8954763412475586, "learning_rate": 1.9787633534996283e-05, "loss": 7.8038, "step": 2673 }, { "epoch": 0.09422377265079239, "grad_norm": 2.8933777809143066, "learning_rate": 1.978739951703158e-05, "loss": 7.6919, "step": 2674 }, { "epoch": 0.09425900966375081, "grad_norm": 3.0685575008392334, "learning_rate": 1.9787165371584755e-05, "loss": 4.8616, "step": 2675 }, { "epoch": 0.09429424667670921, "grad_norm": 4.19233512878418, "learning_rate": 1.9786931098658854e-05, "loss": 6.9432, "step": 2676 }, { "epoch": 0.09432948368966762, "grad_norm": 2.384563684463501, "learning_rate": 1.9786696698256923e-05, "loss": 5.3201, "step": 2677 }, { "epoch": 0.09436472070262604, "grad_norm": 4.148212432861328, "learning_rate": 1.978646217038203e-05, "loss": 7.3069, "step": 2678 }, { "epoch": 0.09439995771558445, "grad_norm": 8.645859718322754, "learning_rate": 1.9786227515037212e-05, "loss": 9.603, "step": 2679 }, { "epoch": 0.09443519472854287, "grad_norm": 4.234386444091797, "learning_rate": 1.978599273222554e-05, "loss": 8.4435, "step": 2680 }, { "epoch": 0.09447043174150127, "grad_norm": 5.22114372253418, "learning_rate": 1.9785757821950064e-05, "loss": 6.1493, "step": 2681 }, { "epoch": 0.09450566875445969, "grad_norm": 6.737751007080078, "learning_rate": 1.9785522784213846e-05, "loss": 10.366, "step": 2682 }, { "epoch": 0.0945409057674181, "grad_norm": 4.089515686035156, "learning_rate": 1.9785287619019948e-05, "loss": 7.9146, "step": 2683 }, { "epoch": 0.0945761427803765, "grad_norm": 4.196671009063721, "learning_rate": 1.9785052326371434e-05, "loss": 9.8459, "step": 2684 }, { "epoch": 0.09461137979333492, "grad_norm": 3.3528103828430176, "learning_rate": 1.978481690627136e-05, "loss": 11.0944, "step": 2685 }, { "epoch": 0.09464661680629333, "grad_norm": 3.2965900897979736, "learning_rate": 1.9784581358722805e-05, "loss": 5.439, "step": 2686 }, { "epoch": 0.09468185381925175, "grad_norm": 3.8661420345306396, "learning_rate": 1.9784345683728833e-05, "loss": 8.2187, "step": 2687 }, { "epoch": 0.09471709083221015, "grad_norm": 3.0676932334899902, "learning_rate": 1.9784109881292507e-05, "loss": 5.0984, "step": 2688 }, { "epoch": 0.09475232784516857, "grad_norm": 3.6145427227020264, "learning_rate": 1.978387395141691e-05, "loss": 5.1772, "step": 2689 }, { "epoch": 0.09478756485812698, "grad_norm": 5.786282539367676, "learning_rate": 1.9783637894105102e-05, "loss": 8.3049, "step": 2690 }, { "epoch": 0.09482280187108538, "grad_norm": 3.5653769969940186, "learning_rate": 1.978340170936017e-05, "loss": 7.7239, "step": 2691 }, { "epoch": 0.0948580388840438, "grad_norm": 4.471859455108643, "learning_rate": 1.978316539718518e-05, "loss": 5.1981, "step": 2692 }, { "epoch": 0.09489327589700221, "grad_norm": 3.1511051654815674, "learning_rate": 1.9782928957583215e-05, "loss": 5.3308, "step": 2693 }, { "epoch": 0.09492851290996063, "grad_norm": 3.252598285675049, "learning_rate": 1.9782692390557355e-05, "loss": 7.4797, "step": 2694 }, { "epoch": 0.09496374992291903, "grad_norm": 5.179481506347656, "learning_rate": 1.978245569611068e-05, "loss": 8.5519, "step": 2695 }, { "epoch": 0.09499898693587745, "grad_norm": 5.09617805480957, "learning_rate": 1.9782218874246274e-05, "loss": 8.655, "step": 2696 }, { "epoch": 0.09503422394883586, "grad_norm": 4.045094966888428, "learning_rate": 1.978198192496722e-05, "loss": 5.7275, "step": 2697 }, { "epoch": 0.09506946096179426, "grad_norm": 5.6187238693237305, "learning_rate": 1.9781744848276606e-05, "loss": 5.3485, "step": 2698 }, { "epoch": 0.09510469797475268, "grad_norm": 4.836103916168213, "learning_rate": 1.978150764417752e-05, "loss": 5.6583, "step": 2699 }, { "epoch": 0.09513993498771109, "grad_norm": 3.973391056060791, "learning_rate": 1.9781270312673052e-05, "loss": 7.7094, "step": 2700 }, { "epoch": 0.09517517200066951, "grad_norm": 2.9820449352264404, "learning_rate": 1.978103285376629e-05, "loss": 5.3697, "step": 2701 }, { "epoch": 0.09521040901362791, "grad_norm": 6.958560943603516, "learning_rate": 1.9780795267460335e-05, "loss": 10.9859, "step": 2702 }, { "epoch": 0.09524564602658632, "grad_norm": 5.4651288986206055, "learning_rate": 1.978055755375827e-05, "loss": 7.4391, "step": 2703 }, { "epoch": 0.09528088303954474, "grad_norm": 5.445653438568115, "learning_rate": 1.9780319712663198e-05, "loss": 12.1628, "step": 2704 }, { "epoch": 0.09531612005250314, "grad_norm": 4.466730117797852, "learning_rate": 1.9780081744178216e-05, "loss": 6.531, "step": 2705 }, { "epoch": 0.09535135706546156, "grad_norm": 3.110821485519409, "learning_rate": 1.977984364830642e-05, "loss": 5.406, "step": 2706 }, { "epoch": 0.09538659407841997, "grad_norm": 7.082142353057861, "learning_rate": 1.977960542505092e-05, "loss": 10.6462, "step": 2707 }, { "epoch": 0.09542183109137839, "grad_norm": 3.337900400161743, "learning_rate": 1.977936707441481e-05, "loss": 7.9067, "step": 2708 }, { "epoch": 0.0954570681043368, "grad_norm": 2.892515182495117, "learning_rate": 1.97791285964012e-05, "loss": 6.5591, "step": 2709 }, { "epoch": 0.0954923051172952, "grad_norm": 3.3560214042663574, "learning_rate": 1.9778889991013198e-05, "loss": 8.1149, "step": 2710 }, { "epoch": 0.09552754213025362, "grad_norm": 3.578510046005249, "learning_rate": 1.97786512582539e-05, "loss": 8.119, "step": 2711 }, { "epoch": 0.09556277914321203, "grad_norm": 2.750878095626831, "learning_rate": 1.977841239812643e-05, "loss": 5.304, "step": 2712 }, { "epoch": 0.09559801615617045, "grad_norm": 5.131900787353516, "learning_rate": 1.977817341063389e-05, "loss": 9.5881, "step": 2713 }, { "epoch": 0.09563325316912885, "grad_norm": 5.220986843109131, "learning_rate": 1.97779342957794e-05, "loss": 10.7994, "step": 2714 }, { "epoch": 0.09566849018208727, "grad_norm": 4.75163459777832, "learning_rate": 1.9777695053566067e-05, "loss": 7.3743, "step": 2715 }, { "epoch": 0.09570372719504568, "grad_norm": 7.883821487426758, "learning_rate": 1.977745568399701e-05, "loss": 6.5327, "step": 2716 }, { "epoch": 0.09573896420800408, "grad_norm": 3.2249491214752197, "learning_rate": 1.977721618707535e-05, "loss": 5.1475, "step": 2717 }, { "epoch": 0.0957742012209625, "grad_norm": 3.330780267715454, "learning_rate": 1.97769765628042e-05, "loss": 4.9215, "step": 2718 }, { "epoch": 0.09580943823392091, "grad_norm": 4.145940780639648, "learning_rate": 1.9776736811186687e-05, "loss": 5.3084, "step": 2719 }, { "epoch": 0.09584467524687933, "grad_norm": 6.391027450561523, "learning_rate": 1.9776496932225934e-05, "loss": 8.0917, "step": 2720 }, { "epoch": 0.09587991225983773, "grad_norm": 2.5144381523132324, "learning_rate": 1.977625692592506e-05, "loss": 6.3841, "step": 2721 }, { "epoch": 0.09591514927279615, "grad_norm": 4.037167549133301, "learning_rate": 1.9776016792287194e-05, "loss": 4.9626, "step": 2722 }, { "epoch": 0.09595038628575456, "grad_norm": 2.6559560298919678, "learning_rate": 1.9775776531315466e-05, "loss": 7.7181, "step": 2723 }, { "epoch": 0.09598562329871296, "grad_norm": 3.2675065994262695, "learning_rate": 1.9775536143013e-05, "loss": 8.1912, "step": 2724 }, { "epoch": 0.09602086031167138, "grad_norm": 2.755295753479004, "learning_rate": 1.9775295627382935e-05, "loss": 4.6697, "step": 2725 }, { "epoch": 0.09605609732462979, "grad_norm": 3.326200008392334, "learning_rate": 1.9775054984428394e-05, "loss": 7.0817, "step": 2726 }, { "epoch": 0.09609133433758821, "grad_norm": 4.4361138343811035, "learning_rate": 1.9774814214152522e-05, "loss": 5.5069, "step": 2727 }, { "epoch": 0.09612657135054661, "grad_norm": 5.306636333465576, "learning_rate": 1.9774573316558447e-05, "loss": 6.2299, "step": 2728 }, { "epoch": 0.09616180836350502, "grad_norm": 2.044090986251831, "learning_rate": 1.977433229164931e-05, "loss": 5.2686, "step": 2729 }, { "epoch": 0.09619704537646344, "grad_norm": 3.8001785278320312, "learning_rate": 1.977409113942825e-05, "loss": 8.2394, "step": 2730 }, { "epoch": 0.09623228238942184, "grad_norm": 4.681027412414551, "learning_rate": 1.9773849859898406e-05, "loss": 8.1912, "step": 2731 }, { "epoch": 0.09626751940238026, "grad_norm": 2.4084384441375732, "learning_rate": 1.9773608453062926e-05, "loss": 5.7435, "step": 2732 }, { "epoch": 0.09630275641533867, "grad_norm": 1.928824543952942, "learning_rate": 1.9773366918924946e-05, "loss": 5.9712, "step": 2733 }, { "epoch": 0.09633799342829709, "grad_norm": 5.692034721374512, "learning_rate": 1.9773125257487622e-05, "loss": 10.1314, "step": 2734 }, { "epoch": 0.0963732304412555, "grad_norm": 3.7429561614990234, "learning_rate": 1.9772883468754096e-05, "loss": 8.6895, "step": 2735 }, { "epoch": 0.0964084674542139, "grad_norm": 4.27713680267334, "learning_rate": 1.9772641552727518e-05, "loss": 8.002, "step": 2736 }, { "epoch": 0.09644370446717232, "grad_norm": 7.144478797912598, "learning_rate": 1.9772399509411037e-05, "loss": 4.5237, "step": 2737 }, { "epoch": 0.09647894148013073, "grad_norm": 4.076171875, "learning_rate": 1.9772157338807808e-05, "loss": 8.5195, "step": 2738 }, { "epoch": 0.09651417849308914, "grad_norm": 3.0627384185791016, "learning_rate": 1.9771915040920982e-05, "loss": 4.7665, "step": 2739 }, { "epoch": 0.09654941550604755, "grad_norm": 3.096759080886841, "learning_rate": 1.9771672615753724e-05, "loss": 8.0384, "step": 2740 }, { "epoch": 0.09658465251900597, "grad_norm": 3.758125066757202, "learning_rate": 1.977143006330918e-05, "loss": 8.6809, "step": 2741 }, { "epoch": 0.09661988953196438, "grad_norm": 6.111944675445557, "learning_rate": 1.9771187383590518e-05, "loss": 6.1362, "step": 2742 }, { "epoch": 0.09665512654492278, "grad_norm": 5.707476615905762, "learning_rate": 1.9770944576600895e-05, "loss": 9.3383, "step": 2743 }, { "epoch": 0.0966903635578812, "grad_norm": 3.0729568004608154, "learning_rate": 1.9770701642343475e-05, "loss": 5.6848, "step": 2744 }, { "epoch": 0.0967256005708396, "grad_norm": 3.0386247634887695, "learning_rate": 1.977045858082142e-05, "loss": 5.8676, "step": 2745 }, { "epoch": 0.09676083758379803, "grad_norm": 4.363292694091797, "learning_rate": 1.97702153920379e-05, "loss": 10.5439, "step": 2746 }, { "epoch": 0.09679607459675643, "grad_norm": 3.413996458053589, "learning_rate": 1.976997207599608e-05, "loss": 7.5883, "step": 2747 }, { "epoch": 0.09683131160971485, "grad_norm": 5.239765644073486, "learning_rate": 1.9769728632699125e-05, "loss": 13.6133, "step": 2748 }, { "epoch": 0.09686654862267326, "grad_norm": 4.589384078979492, "learning_rate": 1.9769485062150212e-05, "loss": 7.489, "step": 2749 }, { "epoch": 0.09690178563563166, "grad_norm": 3.0520131587982178, "learning_rate": 1.9769241364352512e-05, "loss": 7.7124, "step": 2750 }, { "epoch": 0.09693702264859008, "grad_norm": 2.8655030727386475, "learning_rate": 1.97689975393092e-05, "loss": 7.1635, "step": 2751 }, { "epoch": 0.09697225966154849, "grad_norm": 3.8307931423187256, "learning_rate": 1.976875358702345e-05, "loss": 8.3506, "step": 2752 }, { "epoch": 0.09700749667450691, "grad_norm": 3.833514928817749, "learning_rate": 1.9768509507498435e-05, "loss": 7.7909, "step": 2753 }, { "epoch": 0.09704273368746531, "grad_norm": 3.34987211227417, "learning_rate": 1.976826530073735e-05, "loss": 7.1161, "step": 2754 }, { "epoch": 0.09707797070042372, "grad_norm": 4.148176193237305, "learning_rate": 1.9768020966743354e-05, "loss": 7.6666, "step": 2755 }, { "epoch": 0.09711320771338214, "grad_norm": 2.62873911857605, "learning_rate": 1.9767776505519646e-05, "loss": 5.7976, "step": 2756 }, { "epoch": 0.09714844472634054, "grad_norm": 5.896088600158691, "learning_rate": 1.9767531917069407e-05, "loss": 7.3447, "step": 2757 }, { "epoch": 0.09718368173929896, "grad_norm": 2.4789445400238037, "learning_rate": 1.976728720139582e-05, "loss": 5.1654, "step": 2758 }, { "epoch": 0.09721891875225737, "grad_norm": 2.897049903869629, "learning_rate": 1.9767042358502067e-05, "loss": 4.7364, "step": 2759 }, { "epoch": 0.09725415576521579, "grad_norm": 3.1191322803497314, "learning_rate": 1.976679738839135e-05, "loss": 8.2448, "step": 2760 }, { "epoch": 0.0972893927781742, "grad_norm": 2.3534622192382812, "learning_rate": 1.976655229106685e-05, "loss": 4.5532, "step": 2761 }, { "epoch": 0.0973246297911326, "grad_norm": 2.914210796356201, "learning_rate": 1.976630706653176e-05, "loss": 7.248, "step": 2762 }, { "epoch": 0.09735986680409102, "grad_norm": 3.0960981845855713, "learning_rate": 1.9766061714789285e-05, "loss": 6.992, "step": 2763 }, { "epoch": 0.09739510381704942, "grad_norm": 6.033299446105957, "learning_rate": 1.9765816235842602e-05, "loss": 4.9165, "step": 2764 }, { "epoch": 0.09743034083000784, "grad_norm": 2.716491222381592, "learning_rate": 1.9765570629694926e-05, "loss": 7.5218, "step": 2765 }, { "epoch": 0.09746557784296625, "grad_norm": 6.943063735961914, "learning_rate": 1.9765324896349446e-05, "loss": 9.3945, "step": 2766 }, { "epoch": 0.09750081485592467, "grad_norm": 3.479814052581787, "learning_rate": 1.9765079035809362e-05, "loss": 9.947, "step": 2767 }, { "epoch": 0.09753605186888308, "grad_norm": 2.4434831142425537, "learning_rate": 1.9764833048077886e-05, "loss": 5.4121, "step": 2768 }, { "epoch": 0.09757128888184148, "grad_norm": 4.46685266494751, "learning_rate": 1.976458693315821e-05, "loss": 7.0002, "step": 2769 }, { "epoch": 0.0976065258947999, "grad_norm": 3.373274087905884, "learning_rate": 1.9764340691053547e-05, "loss": 4.9187, "step": 2770 }, { "epoch": 0.0976417629077583, "grad_norm": 2.0155606269836426, "learning_rate": 1.9764094321767105e-05, "loss": 5.6892, "step": 2771 }, { "epoch": 0.09767699992071673, "grad_norm": 3.464306116104126, "learning_rate": 1.9763847825302087e-05, "loss": 4.975, "step": 2772 }, { "epoch": 0.09771223693367513, "grad_norm": 3.653812885284424, "learning_rate": 1.976360120166171e-05, "loss": 6.9676, "step": 2773 }, { "epoch": 0.09774747394663355, "grad_norm": 2.6098384857177734, "learning_rate": 1.9763354450849185e-05, "loss": 5.1655, "step": 2774 }, { "epoch": 0.09778271095959196, "grad_norm": 2.3922412395477295, "learning_rate": 1.9763107572867725e-05, "loss": 5.1701, "step": 2775 }, { "epoch": 0.09781794797255036, "grad_norm": 5.839688301086426, "learning_rate": 1.976286056772054e-05, "loss": 7.7689, "step": 2776 }, { "epoch": 0.09785318498550878, "grad_norm": 3.792790651321411, "learning_rate": 1.9762613435410854e-05, "loss": 5.6914, "step": 2777 }, { "epoch": 0.09788842199846719, "grad_norm": 2.988589286804199, "learning_rate": 1.976236617594189e-05, "loss": 5.8014, "step": 2778 }, { "epoch": 0.0979236590114256, "grad_norm": 2.912149667739868, "learning_rate": 1.9762118789316855e-05, "loss": 5.9737, "step": 2779 }, { "epoch": 0.09795889602438401, "grad_norm": 8.042208671569824, "learning_rate": 1.9761871275538986e-05, "loss": 8.1587, "step": 2780 }, { "epoch": 0.09799413303734242, "grad_norm": 3.9235551357269287, "learning_rate": 1.9761623634611497e-05, "loss": 4.819, "step": 2781 }, { "epoch": 0.09802937005030084, "grad_norm": 4.775265216827393, "learning_rate": 1.9761375866537617e-05, "loss": 8.6347, "step": 2782 }, { "epoch": 0.09806460706325924, "grad_norm": 4.771561145782471, "learning_rate": 1.9761127971320572e-05, "loss": 5.4462, "step": 2783 }, { "epoch": 0.09809984407621766, "grad_norm": 4.549435615539551, "learning_rate": 1.9760879948963593e-05, "loss": 7.7863, "step": 2784 }, { "epoch": 0.09813508108917607, "grad_norm": 3.4721527099609375, "learning_rate": 1.976063179946991e-05, "loss": 8.2324, "step": 2785 }, { "epoch": 0.09817031810213449, "grad_norm": 3.7607147693634033, "learning_rate": 1.9760383522842754e-05, "loss": 8.4073, "step": 2786 }, { "epoch": 0.0982055551150929, "grad_norm": 4.297215938568115, "learning_rate": 1.9760135119085358e-05, "loss": 6.7422, "step": 2787 }, { "epoch": 0.0982407921280513, "grad_norm": 7.723025321960449, "learning_rate": 1.975988658820096e-05, "loss": 7.6878, "step": 2788 }, { "epoch": 0.09827602914100972, "grad_norm": 6.664500713348389, "learning_rate": 1.9759637930192796e-05, "loss": 12.0531, "step": 2789 }, { "epoch": 0.09831126615396812, "grad_norm": 3.7787535190582275, "learning_rate": 1.9759389145064105e-05, "loss": 6.2584, "step": 2790 }, { "epoch": 0.09834650316692654, "grad_norm": 3.470071315765381, "learning_rate": 1.9759140232818127e-05, "loss": 4.7576, "step": 2791 }, { "epoch": 0.09838174017988495, "grad_norm": 6.806899547576904, "learning_rate": 1.9758891193458104e-05, "loss": 8.3216, "step": 2792 }, { "epoch": 0.09841697719284337, "grad_norm": 3.6012067794799805, "learning_rate": 1.9758642026987282e-05, "loss": 5.3644, "step": 2793 }, { "epoch": 0.09845221420580177, "grad_norm": 10.679421424865723, "learning_rate": 1.9758392733408904e-05, "loss": 10.0731, "step": 2794 }, { "epoch": 0.09848745121876018, "grad_norm": 3.7657523155212402, "learning_rate": 1.975814331272621e-05, "loss": 5.6881, "step": 2795 }, { "epoch": 0.0985226882317186, "grad_norm": 2.5121817588806152, "learning_rate": 1.9757893764942465e-05, "loss": 5.214, "step": 2796 }, { "epoch": 0.098557925244677, "grad_norm": 4.141789436340332, "learning_rate": 1.975764409006091e-05, "loss": 7.2643, "step": 2797 }, { "epoch": 0.09859316225763542, "grad_norm": 3.9681060314178467, "learning_rate": 1.9757394288084798e-05, "loss": 4.6989, "step": 2798 }, { "epoch": 0.09862839927059383, "grad_norm": 3.882317304611206, "learning_rate": 1.975714435901738e-05, "loss": 7.3788, "step": 2799 }, { "epoch": 0.09866363628355225, "grad_norm": 3.2203192710876465, "learning_rate": 1.9756894302861913e-05, "loss": 5.7655, "step": 2800 }, { "epoch": 0.09869887329651066, "grad_norm": 2.3615238666534424, "learning_rate": 1.975664411962166e-05, "loss": 5.6658, "step": 2801 }, { "epoch": 0.09873411030946906, "grad_norm": 4.402431488037109, "learning_rate": 1.975639380929987e-05, "loss": 5.6929, "step": 2802 }, { "epoch": 0.09876934732242748, "grad_norm": 4.464417934417725, "learning_rate": 1.9756143371899807e-05, "loss": 10.8116, "step": 2803 }, { "epoch": 0.09880458433538589, "grad_norm": 4.615457057952881, "learning_rate": 1.9755892807424738e-05, "loss": 4.8028, "step": 2804 }, { "epoch": 0.0988398213483443, "grad_norm": 4.610134601593018, "learning_rate": 1.975564211587792e-05, "loss": 4.7492, "step": 2805 }, { "epoch": 0.09887505836130271, "grad_norm": 3.3414881229400635, "learning_rate": 1.9755391297262624e-05, "loss": 7.5232, "step": 2806 }, { "epoch": 0.09891029537426112, "grad_norm": 4.251720905303955, "learning_rate": 1.975514035158211e-05, "loss": 5.024, "step": 2807 }, { "epoch": 0.09894553238721954, "grad_norm": 2.995629072189331, "learning_rate": 1.9754889278839647e-05, "loss": 4.6107, "step": 2808 }, { "epoch": 0.09898076940017794, "grad_norm": 3.8718934059143066, "learning_rate": 1.9754638079038514e-05, "loss": 6.3668, "step": 2809 }, { "epoch": 0.09901600641313636, "grad_norm": 6.255846977233887, "learning_rate": 1.9754386752181978e-05, "loss": 9.0888, "step": 2810 }, { "epoch": 0.09905124342609477, "grad_norm": 3.2228004932403564, "learning_rate": 1.975413529827331e-05, "loss": 7.43, "step": 2811 }, { "epoch": 0.09908648043905319, "grad_norm": 3.2921035289764404, "learning_rate": 1.9753883717315788e-05, "loss": 7.4609, "step": 2812 }, { "epoch": 0.09912171745201159, "grad_norm": 5.63803768157959, "learning_rate": 1.9753632009312688e-05, "loss": 9.3062, "step": 2813 }, { "epoch": 0.09915695446497, "grad_norm": 5.238389015197754, "learning_rate": 1.975338017426729e-05, "loss": 8.7124, "step": 2814 }, { "epoch": 0.09919219147792842, "grad_norm": 3.960723876953125, "learning_rate": 1.975312821218287e-05, "loss": 7.8248, "step": 2815 }, { "epoch": 0.09922742849088682, "grad_norm": 3.209660053253174, "learning_rate": 1.9752876123062713e-05, "loss": 5.163, "step": 2816 }, { "epoch": 0.09926266550384524, "grad_norm": 2.390164375305176, "learning_rate": 1.9752623906910104e-05, "loss": 4.4966, "step": 2817 }, { "epoch": 0.09929790251680365, "grad_norm": 2.710799217224121, "learning_rate": 1.9752371563728326e-05, "loss": 4.7112, "step": 2818 }, { "epoch": 0.09933313952976207, "grad_norm": 5.882761478424072, "learning_rate": 1.9752119093520666e-05, "loss": 12.0977, "step": 2819 }, { "epoch": 0.09936837654272047, "grad_norm": 4.707705497741699, "learning_rate": 1.975186649629041e-05, "loss": 5.1208, "step": 2820 }, { "epoch": 0.09940361355567888, "grad_norm": 4.489301681518555, "learning_rate": 1.9751613772040853e-05, "loss": 7.4108, "step": 2821 }, { "epoch": 0.0994388505686373, "grad_norm": 6.033566951751709, "learning_rate": 1.9751360920775286e-05, "loss": 5.4257, "step": 2822 }, { "epoch": 0.0994740875815957, "grad_norm": 4.68234395980835, "learning_rate": 1.9751107942497e-05, "loss": 7.8626, "step": 2823 }, { "epoch": 0.09950932459455412, "grad_norm": 4.588456153869629, "learning_rate": 1.9750854837209288e-05, "loss": 7.6589, "step": 2824 }, { "epoch": 0.09954456160751253, "grad_norm": 1.9907900094985962, "learning_rate": 1.975060160491545e-05, "loss": 5.4685, "step": 2825 }, { "epoch": 0.09957979862047095, "grad_norm": 2.20306134223938, "learning_rate": 1.9750348245618786e-05, "loss": 4.993, "step": 2826 }, { "epoch": 0.09961503563342936, "grad_norm": 3.1544601917266846, "learning_rate": 1.9750094759322597e-05, "loss": 8.9771, "step": 2827 }, { "epoch": 0.09965027264638776, "grad_norm": 3.3908164501190186, "learning_rate": 1.9749841146030178e-05, "loss": 4.2402, "step": 2828 }, { "epoch": 0.09968550965934618, "grad_norm": 3.8442821502685547, "learning_rate": 1.9749587405744837e-05, "loss": 9.5956, "step": 2829 }, { "epoch": 0.09972074667230459, "grad_norm": 4.073941230773926, "learning_rate": 1.9749333538469877e-05, "loss": 6.962, "step": 2830 }, { "epoch": 0.099755983685263, "grad_norm": 3.4642152786254883, "learning_rate": 1.9749079544208605e-05, "loss": 6.7704, "step": 2831 }, { "epoch": 0.09979122069822141, "grad_norm": 2.1147656440734863, "learning_rate": 1.974882542296433e-05, "loss": 5.5009, "step": 2832 }, { "epoch": 0.09982645771117982, "grad_norm": 2.91139817237854, "learning_rate": 1.9748571174740366e-05, "loss": 7.7849, "step": 2833 }, { "epoch": 0.09986169472413824, "grad_norm": 6.551177024841309, "learning_rate": 1.9748316799540022e-05, "loss": 11.8147, "step": 2834 }, { "epoch": 0.09989693173709664, "grad_norm": 3.2296218872070312, "learning_rate": 1.9748062297366608e-05, "loss": 5.2592, "step": 2835 }, { "epoch": 0.09993216875005506, "grad_norm": 7.867426872253418, "learning_rate": 1.9747807668223435e-05, "loss": 10.7143, "step": 2836 }, { "epoch": 0.09996740576301347, "grad_norm": 6.494894504547119, "learning_rate": 1.9747552912113832e-05, "loss": 7.9474, "step": 2837 }, { "epoch": 0.10000264277597189, "grad_norm": 4.971475124359131, "learning_rate": 1.974729802904111e-05, "loss": 4.5861, "step": 2838 }, { "epoch": 0.10003787978893029, "grad_norm": 4.769617557525635, "learning_rate": 1.974704301900859e-05, "loss": 7.9815, "step": 2839 }, { "epoch": 0.1000731168018887, "grad_norm": 2.7510108947753906, "learning_rate": 1.974678788201959e-05, "loss": 4.3505, "step": 2840 }, { "epoch": 0.10010835381484712, "grad_norm": 4.8737640380859375, "learning_rate": 1.9746532618077437e-05, "loss": 5.7771, "step": 2841 }, { "epoch": 0.10014359082780552, "grad_norm": 4.821971416473389, "learning_rate": 1.9746277227185456e-05, "loss": 6.7497, "step": 2842 }, { "epoch": 0.10017882784076394, "grad_norm": 10.41567325592041, "learning_rate": 1.974602170934697e-05, "loss": 10.7629, "step": 2843 }, { "epoch": 0.10021406485372235, "grad_norm": 5.837238311767578, "learning_rate": 1.9745766064565314e-05, "loss": 7.5306, "step": 2844 }, { "epoch": 0.10024930186668077, "grad_norm": 5.808064937591553, "learning_rate": 1.974551029284381e-05, "loss": 8.107, "step": 2845 }, { "epoch": 0.10028453887963917, "grad_norm": 5.71960973739624, "learning_rate": 1.9745254394185796e-05, "loss": 6.7868, "step": 2846 }, { "epoch": 0.10031977589259758, "grad_norm": 4.554245471954346, "learning_rate": 1.97449983685946e-05, "loss": 6.6892, "step": 2847 }, { "epoch": 0.100355012905556, "grad_norm": 4.331148147583008, "learning_rate": 1.974474221607356e-05, "loss": 9.7605, "step": 2848 }, { "epoch": 0.1003902499185144, "grad_norm": 6.459632873535156, "learning_rate": 1.9744485936626013e-05, "loss": 8.8803, "step": 2849 }, { "epoch": 0.10042548693147282, "grad_norm": 3.8774573802948, "learning_rate": 1.9744229530255292e-05, "loss": 7.1188, "step": 2850 }, { "epoch": 0.10046072394443123, "grad_norm": 3.5203564167022705, "learning_rate": 1.974397299696474e-05, "loss": 7.8939, "step": 2851 }, { "epoch": 0.10049596095738965, "grad_norm": 3.918883800506592, "learning_rate": 1.9743716336757704e-05, "loss": 4.7595, "step": 2852 }, { "epoch": 0.10053119797034805, "grad_norm": 7.986271858215332, "learning_rate": 1.9743459549637517e-05, "loss": 10.0541, "step": 2853 }, { "epoch": 0.10056643498330646, "grad_norm": 2.271664619445801, "learning_rate": 1.974320263560753e-05, "loss": 5.4391, "step": 2854 }, { "epoch": 0.10060167199626488, "grad_norm": 2.7177464962005615, "learning_rate": 1.9742945594671085e-05, "loss": 7.7675, "step": 2855 }, { "epoch": 0.10063690900922329, "grad_norm": 3.8635671138763428, "learning_rate": 1.9742688426831534e-05, "loss": 7.1045, "step": 2856 }, { "epoch": 0.1006721460221817, "grad_norm": 2.7785098552703857, "learning_rate": 1.9742431132092227e-05, "loss": 7.2928, "step": 2857 }, { "epoch": 0.10070738303514011, "grad_norm": 2.9423587322235107, "learning_rate": 1.9742173710456513e-05, "loss": 7.8119, "step": 2858 }, { "epoch": 0.10074262004809852, "grad_norm": 1.715941071510315, "learning_rate": 1.9741916161927744e-05, "loss": 4.9902, "step": 2859 }, { "epoch": 0.10077785706105694, "grad_norm": 2.8756356239318848, "learning_rate": 1.9741658486509277e-05, "loss": 8.1548, "step": 2860 }, { "epoch": 0.10081309407401534, "grad_norm": 4.3790459632873535, "learning_rate": 1.9741400684204467e-05, "loss": 10.2379, "step": 2861 }, { "epoch": 0.10084833108697376, "grad_norm": 2.3062851428985596, "learning_rate": 1.974114275501667e-05, "loss": 5.239, "step": 2862 }, { "epoch": 0.10088356809993217, "grad_norm": 2.0262348651885986, "learning_rate": 1.974088469894925e-05, "loss": 4.5948, "step": 2863 }, { "epoch": 0.10091880511289059, "grad_norm": 2.3814992904663086, "learning_rate": 1.9740626516005566e-05, "loss": 8.6026, "step": 2864 }, { "epoch": 0.10095404212584899, "grad_norm": 2.127427577972412, "learning_rate": 1.9740368206188982e-05, "loss": 5.5836, "step": 2865 }, { "epoch": 0.1009892791388074, "grad_norm": 3.607811689376831, "learning_rate": 1.974010976950286e-05, "loss": 10.4531, "step": 2866 }, { "epoch": 0.10102451615176582, "grad_norm": 3.101579189300537, "learning_rate": 1.9739851205950565e-05, "loss": 5.007, "step": 2867 }, { "epoch": 0.10105975316472422, "grad_norm": 3.17021107673645, "learning_rate": 1.973959251553547e-05, "loss": 9.3184, "step": 2868 }, { "epoch": 0.10109499017768264, "grad_norm": 2.6445703506469727, "learning_rate": 1.9739333698260944e-05, "loss": 7.4983, "step": 2869 }, { "epoch": 0.10113022719064105, "grad_norm": 2.376462459564209, "learning_rate": 1.9739074754130354e-05, "loss": 7.1422, "step": 2870 }, { "epoch": 0.10116546420359947, "grad_norm": 5.27662467956543, "learning_rate": 1.9738815683147074e-05, "loss": 6.919, "step": 2871 }, { "epoch": 0.10120070121655787, "grad_norm": 3.271975040435791, "learning_rate": 1.973855648531448e-05, "loss": 6.71, "step": 2872 }, { "epoch": 0.10123593822951628, "grad_norm": 2.8655800819396973, "learning_rate": 1.9738297160635952e-05, "loss": 4.5502, "step": 2873 }, { "epoch": 0.1012711752424747, "grad_norm": 2.650603771209717, "learning_rate": 1.9738037709114856e-05, "loss": 4.6809, "step": 2874 }, { "epoch": 0.1013064122554331, "grad_norm": 3.203644275665283, "learning_rate": 1.9737778130754578e-05, "loss": 4.5867, "step": 2875 }, { "epoch": 0.10134164926839152, "grad_norm": 9.33868408203125, "learning_rate": 1.9737518425558505e-05, "loss": 10.5339, "step": 2876 }, { "epoch": 0.10137688628134993, "grad_norm": 3.6468985080718994, "learning_rate": 1.973725859353001e-05, "loss": 7.9465, "step": 2877 }, { "epoch": 0.10141212329430835, "grad_norm": 2.864104986190796, "learning_rate": 1.9736998634672482e-05, "loss": 5.0018, "step": 2878 }, { "epoch": 0.10144736030726675, "grad_norm": 1.703410267829895, "learning_rate": 1.9736738548989308e-05, "loss": 5.0238, "step": 2879 }, { "epoch": 0.10148259732022516, "grad_norm": 2.7258799076080322, "learning_rate": 1.9736478336483875e-05, "loss": 4.3438, "step": 2880 }, { "epoch": 0.10151783433318358, "grad_norm": 4.661616325378418, "learning_rate": 1.9736217997159568e-05, "loss": 7.1215, "step": 2881 }, { "epoch": 0.10155307134614198, "grad_norm": 2.4013540744781494, "learning_rate": 1.9735957531019786e-05, "loss": 4.4455, "step": 2882 }, { "epoch": 0.1015883083591004, "grad_norm": 2.7099783420562744, "learning_rate": 1.9735696938067912e-05, "loss": 4.9765, "step": 2883 }, { "epoch": 0.10162354537205881, "grad_norm": 3.8869516849517822, "learning_rate": 1.973543621830735e-05, "loss": 7.696, "step": 2884 }, { "epoch": 0.10165878238501722, "grad_norm": 2.301161289215088, "learning_rate": 1.9735175371741485e-05, "loss": 4.5891, "step": 2885 }, { "epoch": 0.10169401939797564, "grad_norm": 2.575265645980835, "learning_rate": 1.9734914398373725e-05, "loss": 5.1542, "step": 2886 }, { "epoch": 0.10172925641093404, "grad_norm": 3.449394702911377, "learning_rate": 1.9734653298207464e-05, "loss": 7.167, "step": 2887 }, { "epoch": 0.10176449342389246, "grad_norm": 3.904064893722534, "learning_rate": 1.9734392071246103e-05, "loss": 7.8083, "step": 2888 }, { "epoch": 0.10179973043685087, "grad_norm": 4.951888084411621, "learning_rate": 1.9734130717493048e-05, "loss": 4.6214, "step": 2889 }, { "epoch": 0.10183496744980929, "grad_norm": 8.33664608001709, "learning_rate": 1.97338692369517e-05, "loss": 7.0598, "step": 2890 }, { "epoch": 0.10187020446276769, "grad_norm": 6.727916240692139, "learning_rate": 1.9733607629625464e-05, "loss": 7.0576, "step": 2891 }, { "epoch": 0.1019054414757261, "grad_norm": 2.5519087314605713, "learning_rate": 1.973334589551775e-05, "loss": 5.2474, "step": 2892 }, { "epoch": 0.10194067848868452, "grad_norm": 2.214207172393799, "learning_rate": 1.9733084034631964e-05, "loss": 5.2222, "step": 2893 }, { "epoch": 0.10197591550164292, "grad_norm": 4.22908878326416, "learning_rate": 1.973282204697152e-05, "loss": 6.7932, "step": 2894 }, { "epoch": 0.10201115251460134, "grad_norm": 9.747177124023438, "learning_rate": 1.9732559932539828e-05, "loss": 9.5897, "step": 2895 }, { "epoch": 0.10204638952755975, "grad_norm": 1.7226827144622803, "learning_rate": 1.97322976913403e-05, "loss": 5.6991, "step": 2896 }, { "epoch": 0.10208162654051817, "grad_norm": 3.7612385749816895, "learning_rate": 1.9732035323376362e-05, "loss": 7.4046, "step": 2897 }, { "epoch": 0.10211686355347657, "grad_norm": 2.6778740882873535, "learning_rate": 1.973177282865142e-05, "loss": 5.0989, "step": 2898 }, { "epoch": 0.10215210056643498, "grad_norm": 4.073547840118408, "learning_rate": 1.9731510207168897e-05, "loss": 4.1566, "step": 2899 }, { "epoch": 0.1021873375793934, "grad_norm": 1.9773709774017334, "learning_rate": 1.9731247458932216e-05, "loss": 6.1876, "step": 2900 }, { "epoch": 0.1022225745923518, "grad_norm": 3.684119462966919, "learning_rate": 1.9730984583944796e-05, "loss": 4.5593, "step": 2901 }, { "epoch": 0.10225781160531022, "grad_norm": 6.485085487365723, "learning_rate": 1.9730721582210063e-05, "loss": 7.2728, "step": 2902 }, { "epoch": 0.10229304861826863, "grad_norm": 2.014636993408203, "learning_rate": 1.9730458453731442e-05, "loss": 5.9158, "step": 2903 }, { "epoch": 0.10232828563122705, "grad_norm": 2.4854063987731934, "learning_rate": 1.973019519851236e-05, "loss": 4.936, "step": 2904 }, { "epoch": 0.10236352264418545, "grad_norm": 5.88689661026001, "learning_rate": 1.9729931816556248e-05, "loss": 7.2159, "step": 2905 }, { "epoch": 0.10239875965714386, "grad_norm": 2.9841997623443604, "learning_rate": 1.972966830786653e-05, "loss": 4.8395, "step": 2906 }, { "epoch": 0.10243399667010228, "grad_norm": 3.7435684204101562, "learning_rate": 1.9729404672446645e-05, "loss": 6.286, "step": 2907 }, { "epoch": 0.10246923368306068, "grad_norm": 1.916637659072876, "learning_rate": 1.9729140910300027e-05, "loss": 4.4535, "step": 2908 }, { "epoch": 0.1025044706960191, "grad_norm": 2.507827043533325, "learning_rate": 1.9728877021430107e-05, "loss": 3.8601, "step": 2909 }, { "epoch": 0.10253970770897751, "grad_norm": 3.5436606407165527, "learning_rate": 1.9728613005840324e-05, "loss": 7.2479, "step": 2910 }, { "epoch": 0.10257494472193592, "grad_norm": 2.3494598865509033, "learning_rate": 1.9728348863534118e-05, "loss": 4.5893, "step": 2911 }, { "epoch": 0.10261018173489433, "grad_norm": 4.491735458374023, "learning_rate": 1.972808459451493e-05, "loss": 7.0137, "step": 2912 }, { "epoch": 0.10264541874785274, "grad_norm": 5.308455467224121, "learning_rate": 1.97278201987862e-05, "loss": 7.9818, "step": 2913 }, { "epoch": 0.10268065576081116, "grad_norm": 5.725704669952393, "learning_rate": 1.9727555676351374e-05, "loss": 6.1734, "step": 2914 }, { "epoch": 0.10271589277376957, "grad_norm": 2.1368157863616943, "learning_rate": 1.9727291027213897e-05, "loss": 5.294, "step": 2915 }, { "epoch": 0.10275112978672798, "grad_norm": 4.227774143218994, "learning_rate": 1.9727026251377214e-05, "loss": 4.0706, "step": 2916 }, { "epoch": 0.10278636679968639, "grad_norm": 2.6525931358337402, "learning_rate": 1.9726761348844775e-05, "loss": 6.9482, "step": 2917 }, { "epoch": 0.1028216038126448, "grad_norm": 4.5975775718688965, "learning_rate": 1.9726496319620026e-05, "loss": 6.9101, "step": 2918 }, { "epoch": 0.10285684082560322, "grad_norm": 3.64111328125, "learning_rate": 1.972623116370643e-05, "loss": 7.2687, "step": 2919 }, { "epoch": 0.10289207783856162, "grad_norm": 2.294781446456909, "learning_rate": 1.972596588110743e-05, "loss": 6.0269, "step": 2920 }, { "epoch": 0.10292731485152004, "grad_norm": 10.209403991699219, "learning_rate": 1.9725700471826486e-05, "loss": 9.5791, "step": 2921 }, { "epoch": 0.10296255186447845, "grad_norm": 2.1054930686950684, "learning_rate": 1.9725434935867058e-05, "loss": 4.496, "step": 2922 }, { "epoch": 0.10299778887743687, "grad_norm": 5.029268264770508, "learning_rate": 1.97251692732326e-05, "loss": 9.896, "step": 2923 }, { "epoch": 0.10303302589039527, "grad_norm": 5.668169021606445, "learning_rate": 1.9724903483926572e-05, "loss": 8.9099, "step": 2924 }, { "epoch": 0.10306826290335368, "grad_norm": 4.035568714141846, "learning_rate": 1.972463756795244e-05, "loss": 7.4935, "step": 2925 }, { "epoch": 0.1031034999163121, "grad_norm": 4.917416095733643, "learning_rate": 1.972437152531366e-05, "loss": 7.4512, "step": 2926 }, { "epoch": 0.1031387369292705, "grad_norm": 5.645336151123047, "learning_rate": 1.9724105356013704e-05, "loss": 6.8634, "step": 2927 }, { "epoch": 0.10317397394222892, "grad_norm": 4.802798748016357, "learning_rate": 1.9723839060056038e-05, "loss": 7.4691, "step": 2928 }, { "epoch": 0.10320921095518733, "grad_norm": 8.080162048339844, "learning_rate": 1.972357263744413e-05, "loss": 10.975, "step": 2929 }, { "epoch": 0.10324444796814575, "grad_norm": 3.4461700916290283, "learning_rate": 1.972330608818145e-05, "loss": 4.8736, "step": 2930 }, { "epoch": 0.10327968498110415, "grad_norm": 7.999200344085693, "learning_rate": 1.9723039412271468e-05, "loss": 7.5572, "step": 2931 }, { "epoch": 0.10331492199406256, "grad_norm": 4.496985912322998, "learning_rate": 1.9722772609717663e-05, "loss": 7.7152, "step": 2932 }, { "epoch": 0.10335015900702098, "grad_norm": 1.8140041828155518, "learning_rate": 1.9722505680523502e-05, "loss": 5.2831, "step": 2933 }, { "epoch": 0.10338539601997938, "grad_norm": 7.653252601623535, "learning_rate": 1.9722238624692467e-05, "loss": 8.3706, "step": 2934 }, { "epoch": 0.1034206330329378, "grad_norm": 3.2816967964172363, "learning_rate": 1.9721971442228037e-05, "loss": 7.5099, "step": 2935 }, { "epoch": 0.10345587004589621, "grad_norm": 6.333223342895508, "learning_rate": 1.9721704133133686e-05, "loss": 9.9915, "step": 2936 }, { "epoch": 0.10349110705885461, "grad_norm": 3.577775716781616, "learning_rate": 1.9721436697412904e-05, "loss": 6.6528, "step": 2937 }, { "epoch": 0.10352634407181303, "grad_norm": 2.695918560028076, "learning_rate": 1.972116913506917e-05, "loss": 6.8846, "step": 2938 }, { "epoch": 0.10356158108477144, "grad_norm": 2.325880765914917, "learning_rate": 1.9720901446105974e-05, "loss": 5.7459, "step": 2939 }, { "epoch": 0.10359681809772986, "grad_norm": 4.504791259765625, "learning_rate": 1.9720633630526795e-05, "loss": 7.9042, "step": 2940 }, { "epoch": 0.10363205511068826, "grad_norm": 6.011647701263428, "learning_rate": 1.9720365688335124e-05, "loss": 7.3953, "step": 2941 }, { "epoch": 0.10366729212364668, "grad_norm": 4.599272727966309, "learning_rate": 1.9720097619534453e-05, "loss": 6.7714, "step": 2942 }, { "epoch": 0.10370252913660509, "grad_norm": 6.229043483734131, "learning_rate": 1.9719829424128272e-05, "loss": 7.7027, "step": 2943 }, { "epoch": 0.1037377661495635, "grad_norm": 2.2302122116088867, "learning_rate": 1.9719561102120074e-05, "loss": 5.2878, "step": 2944 }, { "epoch": 0.10377300316252192, "grad_norm": 2.222252130508423, "learning_rate": 1.9719292653513353e-05, "loss": 4.8184, "step": 2945 }, { "epoch": 0.10380824017548032, "grad_norm": 2.306218147277832, "learning_rate": 1.971902407831161e-05, "loss": 4.3498, "step": 2946 }, { "epoch": 0.10384347718843874, "grad_norm": 15.188657760620117, "learning_rate": 1.9718755376518337e-05, "loss": 7.2702, "step": 2947 }, { "epoch": 0.10387871420139715, "grad_norm": 2.9245717525482178, "learning_rate": 1.9718486548137042e-05, "loss": 4.4146, "step": 2948 }, { "epoch": 0.10391395121435557, "grad_norm": 5.15259313583374, "learning_rate": 1.9718217593171217e-05, "loss": 7.0475, "step": 2949 }, { "epoch": 0.10394918822731397, "grad_norm": 1.6932860612869263, "learning_rate": 1.9717948511624374e-05, "loss": 5.3777, "step": 2950 }, { "epoch": 0.10398442524027238, "grad_norm": 2.365269899368286, "learning_rate": 1.971767930350001e-05, "loss": 5.1591, "step": 2951 }, { "epoch": 0.1040196622532308, "grad_norm": 6.398541450500488, "learning_rate": 1.9717409968801638e-05, "loss": 9.0985, "step": 2952 }, { "epoch": 0.1040548992661892, "grad_norm": 5.633503437042236, "learning_rate": 1.9717140507532764e-05, "loss": 6.7114, "step": 2953 }, { "epoch": 0.10409013627914762, "grad_norm": 2.178568124771118, "learning_rate": 1.9716870919696894e-05, "loss": 3.9219, "step": 2954 }, { "epoch": 0.10412537329210603, "grad_norm": 2.2525203227996826, "learning_rate": 1.9716601205297547e-05, "loss": 5.2616, "step": 2955 }, { "epoch": 0.10416061030506445, "grad_norm": 4.488893508911133, "learning_rate": 1.9716331364338226e-05, "loss": 8.2881, "step": 2956 }, { "epoch": 0.10419584731802285, "grad_norm": 3.1806812286376953, "learning_rate": 1.9716061396822456e-05, "loss": 4.5417, "step": 2957 }, { "epoch": 0.10423108433098126, "grad_norm": 2.698505163192749, "learning_rate": 1.9715791302753747e-05, "loss": 4.4857, "step": 2958 }, { "epoch": 0.10426632134393968, "grad_norm": 5.592583179473877, "learning_rate": 1.9715521082135618e-05, "loss": 7.1929, "step": 2959 }, { "epoch": 0.10430155835689808, "grad_norm": 9.026041984558105, "learning_rate": 1.971525073497159e-05, "loss": 7.2476, "step": 2960 }, { "epoch": 0.1043367953698565, "grad_norm": 9.766456604003906, "learning_rate": 1.9714980261265184e-05, "loss": 6.3355, "step": 2961 }, { "epoch": 0.10437203238281491, "grad_norm": 3.7918286323547363, "learning_rate": 1.9714709661019925e-05, "loss": 7.5316, "step": 2962 }, { "epoch": 0.10440726939577331, "grad_norm": 4.752466201782227, "learning_rate": 1.9714438934239332e-05, "loss": 6.9128, "step": 2963 }, { "epoch": 0.10444250640873173, "grad_norm": 5.2421698570251465, "learning_rate": 1.9714168080926935e-05, "loss": 7.2308, "step": 2964 }, { "epoch": 0.10447774342169014, "grad_norm": 1.8994311094284058, "learning_rate": 1.971389710108626e-05, "loss": 4.2154, "step": 2965 }, { "epoch": 0.10451298043464856, "grad_norm": 1.9373668432235718, "learning_rate": 1.9713625994720838e-05, "loss": 4.2106, "step": 2966 }, { "epoch": 0.10454821744760696, "grad_norm": 3.4838976860046387, "learning_rate": 1.9713354761834203e-05, "loss": 7.2088, "step": 2967 }, { "epoch": 0.10458345446056538, "grad_norm": 6.273225784301758, "learning_rate": 1.9713083402429883e-05, "loss": 10.0465, "step": 2968 }, { "epoch": 0.10461869147352379, "grad_norm": 2.7703115940093994, "learning_rate": 1.9712811916511413e-05, "loss": 3.9944, "step": 2969 }, { "epoch": 0.1046539284864822, "grad_norm": 3.691429376602173, "learning_rate": 1.9712540304082332e-05, "loss": 6.9169, "step": 2970 }, { "epoch": 0.10468916549944061, "grad_norm": 11.45900821685791, "learning_rate": 1.9712268565146176e-05, "loss": 7.1601, "step": 2971 }, { "epoch": 0.10472440251239902, "grad_norm": 2.2581169605255127, "learning_rate": 1.9711996699706487e-05, "loss": 3.8255, "step": 2972 }, { "epoch": 0.10475963952535744, "grad_norm": 3.7030951976776123, "learning_rate": 1.97117247077668e-05, "loss": 7.3409, "step": 2973 }, { "epoch": 0.10479487653831585, "grad_norm": 2.683749198913574, "learning_rate": 1.9711452589330666e-05, "loss": 6.0054, "step": 2974 }, { "epoch": 0.10483011355127426, "grad_norm": 2.10978102684021, "learning_rate": 1.971118034440162e-05, "loss": 5.0911, "step": 2975 }, { "epoch": 0.10486535056423267, "grad_norm": 4.59158992767334, "learning_rate": 1.9710907972983217e-05, "loss": 10.3732, "step": 2976 }, { "epoch": 0.10490058757719108, "grad_norm": 3.048516035079956, "learning_rate": 1.9710635475079e-05, "loss": 4.6583, "step": 2977 }, { "epoch": 0.1049358245901495, "grad_norm": 3.4079957008361816, "learning_rate": 1.971036285069252e-05, "loss": 6.5177, "step": 2978 }, { "epoch": 0.1049710616031079, "grad_norm": 4.154538631439209, "learning_rate": 1.9710090099827324e-05, "loss": 7.1202, "step": 2979 }, { "epoch": 0.10500629861606632, "grad_norm": 2.3062868118286133, "learning_rate": 1.970981722248697e-05, "loss": 5.2684, "step": 2980 }, { "epoch": 0.10504153562902473, "grad_norm": 4.807531356811523, "learning_rate": 1.9709544218675008e-05, "loss": 7.3747, "step": 2981 }, { "epoch": 0.10507677264198313, "grad_norm": 2.4364514350891113, "learning_rate": 1.9709271088394998e-05, "loss": 4.6858, "step": 2982 }, { "epoch": 0.10511200965494155, "grad_norm": 2.526350975036621, "learning_rate": 1.9708997831650492e-05, "loss": 4.4705, "step": 2983 }, { "epoch": 0.10514724666789996, "grad_norm": 3.5159757137298584, "learning_rate": 1.9708724448445055e-05, "loss": 8.068, "step": 2984 }, { "epoch": 0.10518248368085838, "grad_norm": 2.3025248050689697, "learning_rate": 1.9708450938782248e-05, "loss": 4.391, "step": 2985 }, { "epoch": 0.10521772069381678, "grad_norm": 5.477391719818115, "learning_rate": 1.9708177302665627e-05, "loss": 7.8407, "step": 2986 }, { "epoch": 0.1052529577067752, "grad_norm": 4.03801965713501, "learning_rate": 1.9707903540098758e-05, "loss": 7.2155, "step": 2987 }, { "epoch": 0.10528819471973361, "grad_norm": 18.08416748046875, "learning_rate": 1.970762965108521e-05, "loss": 6.418, "step": 2988 }, { "epoch": 0.10532343173269201, "grad_norm": 10.399401664733887, "learning_rate": 1.9707355635628554e-05, "loss": 6.4858, "step": 2989 }, { "epoch": 0.10535866874565043, "grad_norm": 4.508461952209473, "learning_rate": 1.970708149373235e-05, "loss": 7.768, "step": 2990 }, { "epoch": 0.10539390575860884, "grad_norm": 5.6271467208862305, "learning_rate": 1.9706807225400175e-05, "loss": 4.96, "step": 2991 }, { "epoch": 0.10542914277156726, "grad_norm": 4.474921703338623, "learning_rate": 1.9706532830635598e-05, "loss": 7.1299, "step": 2992 }, { "epoch": 0.10546437978452566, "grad_norm": 4.6542253494262695, "learning_rate": 1.9706258309442196e-05, "loss": 4.588, "step": 2993 }, { "epoch": 0.10549961679748408, "grad_norm": 8.300271987915039, "learning_rate": 1.9705983661823543e-05, "loss": 10.1263, "step": 2994 }, { "epoch": 0.10553485381044249, "grad_norm": 4.857484817504883, "learning_rate": 1.9705708887783213e-05, "loss": 8.0118, "step": 2995 }, { "epoch": 0.1055700908234009, "grad_norm": 7.074674129486084, "learning_rate": 1.970543398732479e-05, "loss": 12.7645, "step": 2996 }, { "epoch": 0.10560532783635931, "grad_norm": 3.2544877529144287, "learning_rate": 1.9705158960451856e-05, "loss": 4.2737, "step": 2997 }, { "epoch": 0.10564056484931772, "grad_norm": 2.9979233741760254, "learning_rate": 1.9704883807167987e-05, "loss": 6.7571, "step": 2998 }, { "epoch": 0.10567580186227614, "grad_norm": 2.7269890308380127, "learning_rate": 1.970460852747677e-05, "loss": 3.7686, "step": 2999 }, { "epoch": 0.10571103887523454, "grad_norm": 5.572577953338623, "learning_rate": 1.9704333121381794e-05, "loss": 9.204, "step": 3000 }, { "epoch": 0.10574627588819296, "grad_norm": 3.117575168609619, "learning_rate": 1.9704057588886642e-05, "loss": 4.9717, "step": 3001 }, { "epoch": 0.10578151290115137, "grad_norm": 2.538618803024292, "learning_rate": 1.9703781929994902e-05, "loss": 5.553, "step": 3002 }, { "epoch": 0.10581674991410978, "grad_norm": 2.0783417224884033, "learning_rate": 1.9703506144710167e-05, "loss": 4.7323, "step": 3003 }, { "epoch": 0.1058519869270682, "grad_norm": 5.0619659423828125, "learning_rate": 1.970323023303603e-05, "loss": 7.4842, "step": 3004 }, { "epoch": 0.1058872239400266, "grad_norm": 7.7088236808776855, "learning_rate": 1.9702954194976083e-05, "loss": 8.0876, "step": 3005 }, { "epoch": 0.10592246095298502, "grad_norm": 2.222858190536499, "learning_rate": 1.970267803053392e-05, "loss": 5.2907, "step": 3006 }, { "epoch": 0.10595769796594343, "grad_norm": 2.7259249687194824, "learning_rate": 1.970240173971314e-05, "loss": 4.1112, "step": 3007 }, { "epoch": 0.10599293497890183, "grad_norm": 3.5658180713653564, "learning_rate": 1.9702125322517343e-05, "loss": 4.7076, "step": 3008 }, { "epoch": 0.10602817199186025, "grad_norm": 8.341872215270996, "learning_rate": 1.9701848778950126e-05, "loss": 6.6023, "step": 3009 }, { "epoch": 0.10606340900481866, "grad_norm": 2.256890058517456, "learning_rate": 1.9701572109015093e-05, "loss": 5.1528, "step": 3010 }, { "epoch": 0.10609864601777708, "grad_norm": 3.2169337272644043, "learning_rate": 1.970129531271585e-05, "loss": 4.9127, "step": 3011 }, { "epoch": 0.10613388303073548, "grad_norm": 2.327047824859619, "learning_rate": 1.9701018390055998e-05, "loss": 4.593, "step": 3012 }, { "epoch": 0.1061691200436939, "grad_norm": 4.203036785125732, "learning_rate": 1.9700741341039146e-05, "loss": 4.026, "step": 3013 }, { "epoch": 0.10620435705665231, "grad_norm": 2.963148832321167, "learning_rate": 1.97004641656689e-05, "loss": 4.5279, "step": 3014 }, { "epoch": 0.10623959406961071, "grad_norm": 2.078618049621582, "learning_rate": 1.9700186863948873e-05, "loss": 4.5564, "step": 3015 }, { "epoch": 0.10627483108256913, "grad_norm": 7.84876012802124, "learning_rate": 1.969990943588268e-05, "loss": 8.9361, "step": 3016 }, { "epoch": 0.10631006809552754, "grad_norm": 3.9409265518188477, "learning_rate": 1.969963188147393e-05, "loss": 7.0314, "step": 3017 }, { "epoch": 0.10634530510848596, "grad_norm": 5.756450176239014, "learning_rate": 1.9699354200726236e-05, "loss": 6.9867, "step": 3018 }, { "epoch": 0.10638054212144436, "grad_norm": 2.9917492866516113, "learning_rate": 1.969907639364322e-05, "loss": 6.7047, "step": 3019 }, { "epoch": 0.10641577913440278, "grad_norm": 2.1474812030792236, "learning_rate": 1.96987984602285e-05, "loss": 4.5355, "step": 3020 }, { "epoch": 0.10645101614736119, "grad_norm": 3.3164923191070557, "learning_rate": 1.9698520400485694e-05, "loss": 7.5618, "step": 3021 }, { "epoch": 0.1064862531603196, "grad_norm": 4.370471477508545, "learning_rate": 1.9698242214418424e-05, "loss": 9.181, "step": 3022 }, { "epoch": 0.10652149017327801, "grad_norm": 3.5381438732147217, "learning_rate": 1.9697963902030314e-05, "loss": 7.6139, "step": 3023 }, { "epoch": 0.10655672718623642, "grad_norm": 4.758995532989502, "learning_rate": 1.9697685463324988e-05, "loss": 9.6691, "step": 3024 }, { "epoch": 0.10659196419919484, "grad_norm": 3.4613146781921387, "learning_rate": 1.9697406898306073e-05, "loss": 7.0697, "step": 3025 }, { "epoch": 0.10662720121215324, "grad_norm": 2.422518730163574, "learning_rate": 1.96971282069772e-05, "loss": 5.729, "step": 3026 }, { "epoch": 0.10666243822511166, "grad_norm": 2.506016969680786, "learning_rate": 1.9696849389341996e-05, "loss": 5.2931, "step": 3027 }, { "epoch": 0.10669767523807007, "grad_norm": 2.492926597595215, "learning_rate": 1.969657044540409e-05, "loss": 5.4142, "step": 3028 }, { "epoch": 0.10673291225102848, "grad_norm": 5.124637126922607, "learning_rate": 1.9696291375167127e-05, "loss": 7.6345, "step": 3029 }, { "epoch": 0.1067681492639869, "grad_norm": 2.7657458782196045, "learning_rate": 1.9696012178634727e-05, "loss": 4.5459, "step": 3030 }, { "epoch": 0.1068033862769453, "grad_norm": 1.9417216777801514, "learning_rate": 1.9695732855810536e-05, "loss": 5.9891, "step": 3031 }, { "epoch": 0.10683862328990372, "grad_norm": 2.2615413665771484, "learning_rate": 1.9695453406698187e-05, "loss": 4.0305, "step": 3032 }, { "epoch": 0.10687386030286213, "grad_norm": 5.792564868927002, "learning_rate": 1.9695173831301323e-05, "loss": 9.657, "step": 3033 }, { "epoch": 0.10690909731582053, "grad_norm": 4.17678165435791, "learning_rate": 1.969489412962359e-05, "loss": 9.6937, "step": 3034 }, { "epoch": 0.10694433432877895, "grad_norm": 3.89315128326416, "learning_rate": 1.969461430166862e-05, "loss": 4.3411, "step": 3035 }, { "epoch": 0.10697957134173736, "grad_norm": 2.9392788410186768, "learning_rate": 1.9694334347440066e-05, "loss": 4.7507, "step": 3036 }, { "epoch": 0.10701480835469578, "grad_norm": 3.889247417449951, "learning_rate": 1.9694054266941572e-05, "loss": 6.5744, "step": 3037 }, { "epoch": 0.10705004536765418, "grad_norm": 6.001384735107422, "learning_rate": 1.9693774060176788e-05, "loss": 12.7572, "step": 3038 }, { "epoch": 0.1070852823806126, "grad_norm": 4.308362007141113, "learning_rate": 1.9693493727149357e-05, "loss": 7.4022, "step": 3039 }, { "epoch": 0.107120519393571, "grad_norm": 5.230185031890869, "learning_rate": 1.969321326786294e-05, "loss": 6.3205, "step": 3040 }, { "epoch": 0.10715575640652941, "grad_norm": 2.522167444229126, "learning_rate": 1.9692932682321184e-05, "loss": 5.1327, "step": 3041 }, { "epoch": 0.10719099341948783, "grad_norm": 4.188108444213867, "learning_rate": 1.9692651970527747e-05, "loss": 6.3309, "step": 3042 }, { "epoch": 0.10722623043244624, "grad_norm": 6.366846084594727, "learning_rate": 1.9692371132486282e-05, "loss": 9.3188, "step": 3043 }, { "epoch": 0.10726146744540466, "grad_norm": 2.1156132221221924, "learning_rate": 1.969209016820045e-05, "loss": 5.2821, "step": 3044 }, { "epoch": 0.10729670445836306, "grad_norm": 4.500992774963379, "learning_rate": 1.9691809077673905e-05, "loss": 6.4037, "step": 3045 }, { "epoch": 0.10733194147132148, "grad_norm": 2.6096436977386475, "learning_rate": 1.9691527860910315e-05, "loss": 5.0706, "step": 3046 }, { "epoch": 0.10736717848427989, "grad_norm": 1.7122247219085693, "learning_rate": 1.9691246517913337e-05, "loss": 6.2065, "step": 3047 }, { "epoch": 0.1074024154972383, "grad_norm": 2.8080806732177734, "learning_rate": 1.9690965048686643e-05, "loss": 4.5566, "step": 3048 }, { "epoch": 0.10743765251019671, "grad_norm": 2.4725985527038574, "learning_rate": 1.9690683453233894e-05, "loss": 5.2604, "step": 3049 }, { "epoch": 0.10747288952315512, "grad_norm": 1.9702099561691284, "learning_rate": 1.9690401731558758e-05, "loss": 4.7137, "step": 3050 }, { "epoch": 0.10750812653611354, "grad_norm": 9.346911430358887, "learning_rate": 1.9690119883664905e-05, "loss": 9.0406, "step": 3051 }, { "epoch": 0.10754336354907194, "grad_norm": 7.133174419403076, "learning_rate": 1.9689837909556004e-05, "loss": 8.506, "step": 3052 }, { "epoch": 0.10757860056203036, "grad_norm": 3.603970527648926, "learning_rate": 1.9689555809235734e-05, "loss": 5.0173, "step": 3053 }, { "epoch": 0.10761383757498877, "grad_norm": 9.916817665100098, "learning_rate": 1.9689273582707762e-05, "loss": 10.2054, "step": 3054 }, { "epoch": 0.10764907458794717, "grad_norm": 3.2747912406921387, "learning_rate": 1.968899122997577e-05, "loss": 4.0239, "step": 3055 }, { "epoch": 0.1076843116009056, "grad_norm": 7.792024612426758, "learning_rate": 1.9688708751043432e-05, "loss": 5.872, "step": 3056 }, { "epoch": 0.107719548613864, "grad_norm": 6.385551452636719, "learning_rate": 1.968842614591443e-05, "loss": 10.108, "step": 3057 }, { "epoch": 0.10775478562682242, "grad_norm": 2.3739302158355713, "learning_rate": 1.968814341459244e-05, "loss": 5.2068, "step": 3058 }, { "epoch": 0.10779002263978082, "grad_norm": 3.224120616912842, "learning_rate": 1.9687860557081147e-05, "loss": 4.3448, "step": 3059 }, { "epoch": 0.10782525965273923, "grad_norm": 3.764979362487793, "learning_rate": 1.968757757338424e-05, "loss": 7.5315, "step": 3060 }, { "epoch": 0.10786049666569765, "grad_norm": 2.635068655014038, "learning_rate": 1.96872944635054e-05, "loss": 4.7621, "step": 3061 }, { "epoch": 0.10789573367865606, "grad_norm": 2.9404187202453613, "learning_rate": 1.9687011227448316e-05, "loss": 8.2917, "step": 3062 }, { "epoch": 0.10793097069161448, "grad_norm": 2.5867040157318115, "learning_rate": 1.9686727865216674e-05, "loss": 4.7181, "step": 3063 }, { "epoch": 0.10796620770457288, "grad_norm": 7.890815734863281, "learning_rate": 1.9686444376814172e-05, "loss": 9.4074, "step": 3064 }, { "epoch": 0.1080014447175313, "grad_norm": 3.1994354724884033, "learning_rate": 1.9686160762244494e-05, "loss": 6.7464, "step": 3065 }, { "epoch": 0.1080366817304897, "grad_norm": 3.1378629207611084, "learning_rate": 1.968587702151134e-05, "loss": 7.5152, "step": 3066 }, { "epoch": 0.10807191874344811, "grad_norm": 3.356630563735962, "learning_rate": 1.9685593154618405e-05, "loss": 4.0168, "step": 3067 }, { "epoch": 0.10810715575640653, "grad_norm": 2.822754383087158, "learning_rate": 1.9685309161569384e-05, "loss": 7.0505, "step": 3068 }, { "epoch": 0.10814239276936494, "grad_norm": 3.2966067790985107, "learning_rate": 1.968502504236798e-05, "loss": 4.0399, "step": 3069 }, { "epoch": 0.10817762978232336, "grad_norm": 4.300788879394531, "learning_rate": 1.968474079701789e-05, "loss": 5.2372, "step": 3070 }, { "epoch": 0.10821286679528176, "grad_norm": 2.2557473182678223, "learning_rate": 1.9684456425522814e-05, "loss": 7.658, "step": 3071 }, { "epoch": 0.10824810380824018, "grad_norm": 2.016089677810669, "learning_rate": 1.9684171927886465e-05, "loss": 5.9949, "step": 3072 }, { "epoch": 0.10828334082119859, "grad_norm": 8.933248519897461, "learning_rate": 1.968388730411254e-05, "loss": 6.4688, "step": 3073 }, { "epoch": 0.10831857783415699, "grad_norm": 2.7820024490356445, "learning_rate": 1.9683602554204748e-05, "loss": 6.5975, "step": 3074 }, { "epoch": 0.10835381484711541, "grad_norm": 3.3344924449920654, "learning_rate": 1.9683317678166804e-05, "loss": 7.1776, "step": 3075 }, { "epoch": 0.10838905186007382, "grad_norm": 2.981410026550293, "learning_rate": 1.9683032676002408e-05, "loss": 4.2045, "step": 3076 }, { "epoch": 0.10842428887303224, "grad_norm": 4.39988374710083, "learning_rate": 1.968274754771528e-05, "loss": 10.8467, "step": 3077 }, { "epoch": 0.10845952588599064, "grad_norm": 3.5718345642089844, "learning_rate": 1.9682462293309132e-05, "loss": 7.6682, "step": 3078 }, { "epoch": 0.10849476289894906, "grad_norm": 2.2159218788146973, "learning_rate": 1.968217691278768e-05, "loss": 4.8847, "step": 3079 }, { "epoch": 0.10852999991190747, "grad_norm": 5.021773338317871, "learning_rate": 1.968189140615464e-05, "loss": 6.2359, "step": 3080 }, { "epoch": 0.10856523692486587, "grad_norm": 5.602629661560059, "learning_rate": 1.968160577341373e-05, "loss": 6.8818, "step": 3081 }, { "epoch": 0.1086004739378243, "grad_norm": 4.429756164550781, "learning_rate": 1.9681320014568674e-05, "loss": 7.5156, "step": 3082 }, { "epoch": 0.1086357109507827, "grad_norm": 6.303082466125488, "learning_rate": 1.968103412962319e-05, "loss": 7.692, "step": 3083 }, { "epoch": 0.10867094796374112, "grad_norm": 1.4440443515777588, "learning_rate": 1.9680748118581005e-05, "loss": 6.4124, "step": 3084 }, { "epoch": 0.10870618497669952, "grad_norm": 2.992003917694092, "learning_rate": 1.9680461981445843e-05, "loss": 5.3017, "step": 3085 }, { "epoch": 0.10874142198965793, "grad_norm": 3.981393575668335, "learning_rate": 1.9680175718221428e-05, "loss": 6.554, "step": 3086 }, { "epoch": 0.10877665900261635, "grad_norm": 2.6517207622528076, "learning_rate": 1.967988932891149e-05, "loss": 4.1635, "step": 3087 }, { "epoch": 0.10881189601557475, "grad_norm": 2.4380152225494385, "learning_rate": 1.9679602813519763e-05, "loss": 3.434, "step": 3088 }, { "epoch": 0.10884713302853317, "grad_norm": 4.893213748931885, "learning_rate": 1.9679316172049976e-05, "loss": 4.1312, "step": 3089 }, { "epoch": 0.10888237004149158, "grad_norm": 4.418741226196289, "learning_rate": 1.9679029404505864e-05, "loss": 4.4926, "step": 3090 }, { "epoch": 0.10891760705445, "grad_norm": 8.43238353729248, "learning_rate": 1.967874251089116e-05, "loss": 8.6736, "step": 3091 }, { "epoch": 0.1089528440674084, "grad_norm": 5.2104411125183105, "learning_rate": 1.96784554912096e-05, "loss": 6.5264, "step": 3092 }, { "epoch": 0.10898808108036681, "grad_norm": 6.999913215637207, "learning_rate": 1.967816834546493e-05, "loss": 9.5999, "step": 3093 }, { "epoch": 0.10902331809332523, "grad_norm": 4.982417106628418, "learning_rate": 1.9677881073660877e-05, "loss": 7.5267, "step": 3094 }, { "epoch": 0.10905855510628364, "grad_norm": 3.0223124027252197, "learning_rate": 1.9677593675801195e-05, "loss": 4.5018, "step": 3095 }, { "epoch": 0.10909379211924206, "grad_norm": 2.519319772720337, "learning_rate": 1.967730615188962e-05, "loss": 4.9598, "step": 3096 }, { "epoch": 0.10912902913220046, "grad_norm": 3.4823157787323, "learning_rate": 1.9677018501929902e-05, "loss": 7.2405, "step": 3097 }, { "epoch": 0.10916426614515888, "grad_norm": 2.2899913787841797, "learning_rate": 1.9676730725925785e-05, "loss": 5.5795, "step": 3098 }, { "epoch": 0.10919950315811729, "grad_norm": 2.847787618637085, "learning_rate": 1.9676442823881018e-05, "loss": 6.585, "step": 3099 }, { "epoch": 0.10923474017107569, "grad_norm": 3.3801300525665283, "learning_rate": 1.967615479579935e-05, "loss": 5.8652, "step": 3100 }, { "epoch": 0.10926997718403411, "grad_norm": 2.812626361846924, "learning_rate": 1.9675866641684533e-05, "loss": 6.9752, "step": 3101 }, { "epoch": 0.10930521419699252, "grad_norm": 2.7672712802886963, "learning_rate": 1.967557836154032e-05, "loss": 3.8212, "step": 3102 }, { "epoch": 0.10934045120995094, "grad_norm": 3.604336977005005, "learning_rate": 1.9675289955370468e-05, "loss": 9.4985, "step": 3103 }, { "epoch": 0.10937568822290934, "grad_norm": 3.1150951385498047, "learning_rate": 1.967500142317873e-05, "loss": 8.2585, "step": 3104 }, { "epoch": 0.10941092523586776, "grad_norm": 2.3679609298706055, "learning_rate": 1.9674712764968868e-05, "loss": 4.2798, "step": 3105 }, { "epoch": 0.10944616224882617, "grad_norm": 1.838724970817566, "learning_rate": 1.967442398074464e-05, "loss": 4.4078, "step": 3106 }, { "epoch": 0.10948139926178457, "grad_norm": 4.601446628570557, "learning_rate": 1.9674135070509805e-05, "loss": 7.1196, "step": 3107 }, { "epoch": 0.10951663627474299, "grad_norm": 5.160449981689453, "learning_rate": 1.967384603426813e-05, "loss": 6.3739, "step": 3108 }, { "epoch": 0.1095518732877014, "grad_norm": 3.429349422454834, "learning_rate": 1.9673556872023378e-05, "loss": 4.793, "step": 3109 }, { "epoch": 0.10958711030065982, "grad_norm": 5.930237293243408, "learning_rate": 1.9673267583779315e-05, "loss": 11.0465, "step": 3110 }, { "epoch": 0.10962234731361822, "grad_norm": 4.190768241882324, "learning_rate": 1.9672978169539712e-05, "loss": 8.3564, "step": 3111 }, { "epoch": 0.10965758432657663, "grad_norm": 3.1369807720184326, "learning_rate": 1.9672688629308335e-05, "loss": 5.4218, "step": 3112 }, { "epoch": 0.10969282133953505, "grad_norm": 2.5729048252105713, "learning_rate": 1.9672398963088956e-05, "loss": 5.528, "step": 3113 }, { "epoch": 0.10972805835249345, "grad_norm": 2.4518072605133057, "learning_rate": 1.9672109170885345e-05, "loss": 3.5139, "step": 3114 }, { "epoch": 0.10976329536545187, "grad_norm": 9.986335754394531, "learning_rate": 1.9671819252701284e-05, "loss": 7.3934, "step": 3115 }, { "epoch": 0.10979853237841028, "grad_norm": 11.428364753723145, "learning_rate": 1.9671529208540544e-05, "loss": 12.1721, "step": 3116 }, { "epoch": 0.1098337693913687, "grad_norm": 6.217845916748047, "learning_rate": 1.9671239038406905e-05, "loss": 8.6145, "step": 3117 }, { "epoch": 0.1098690064043271, "grad_norm": 4.228009223937988, "learning_rate": 1.9670948742304143e-05, "loss": 7.2667, "step": 3118 }, { "epoch": 0.10990424341728551, "grad_norm": 5.591989517211914, "learning_rate": 1.9670658320236045e-05, "loss": 9.5859, "step": 3119 }, { "epoch": 0.10993948043024393, "grad_norm": 2.538154125213623, "learning_rate": 1.967036777220639e-05, "loss": 3.8558, "step": 3120 }, { "epoch": 0.10997471744320234, "grad_norm": 5.201537132263184, "learning_rate": 1.9670077098218963e-05, "loss": 7.6722, "step": 3121 }, { "epoch": 0.11000995445616076, "grad_norm": 4.94689416885376, "learning_rate": 1.9669786298277547e-05, "loss": 9.834, "step": 3122 }, { "epoch": 0.11004519146911916, "grad_norm": 3.047593355178833, "learning_rate": 1.9669495372385935e-05, "loss": 4.7205, "step": 3123 }, { "epoch": 0.11008042848207758, "grad_norm": 2.014547824859619, "learning_rate": 1.9669204320547915e-05, "loss": 4.2508, "step": 3124 }, { "epoch": 0.11011566549503599, "grad_norm": 1.944778323173523, "learning_rate": 1.9668913142767277e-05, "loss": 4.9045, "step": 3125 }, { "epoch": 0.11015090250799439, "grad_norm": 4.6392130851745605, "learning_rate": 1.9668621839047813e-05, "loss": 8.2028, "step": 3126 }, { "epoch": 0.11018613952095281, "grad_norm": 2.794748544692993, "learning_rate": 1.9668330409393315e-05, "loss": 4.7613, "step": 3127 }, { "epoch": 0.11022137653391122, "grad_norm": 6.987411022186279, "learning_rate": 1.9668038853807587e-05, "loss": 6.5008, "step": 3128 }, { "epoch": 0.11025661354686964, "grad_norm": 1.4371347427368164, "learning_rate": 1.966774717229442e-05, "loss": 5.4095, "step": 3129 }, { "epoch": 0.11029185055982804, "grad_norm": 7.078964710235596, "learning_rate": 1.9667455364857614e-05, "loss": 7.0285, "step": 3130 }, { "epoch": 0.11032708757278646, "grad_norm": 5.363550186157227, "learning_rate": 1.9667163431500968e-05, "loss": 10.5323, "step": 3131 }, { "epoch": 0.11036232458574487, "grad_norm": 2.6964945793151855, "learning_rate": 1.966687137222829e-05, "loss": 4.1546, "step": 3132 }, { "epoch": 0.11039756159870327, "grad_norm": 3.0843989849090576, "learning_rate": 1.966657918704338e-05, "loss": 6.9397, "step": 3133 }, { "epoch": 0.11043279861166169, "grad_norm": 3.674984931945801, "learning_rate": 1.9666286875950046e-05, "loss": 9.9797, "step": 3134 }, { "epoch": 0.1104680356246201, "grad_norm": 4.097760200500488, "learning_rate": 1.9665994438952092e-05, "loss": 8.8645, "step": 3135 }, { "epoch": 0.11050327263757852, "grad_norm": 3.872375249862671, "learning_rate": 1.966570187605333e-05, "loss": 5.4803, "step": 3136 }, { "epoch": 0.11053850965053692, "grad_norm": 2.7972493171691895, "learning_rate": 1.966540918725757e-05, "loss": 3.6686, "step": 3137 }, { "epoch": 0.11057374666349533, "grad_norm": 2.960157871246338, "learning_rate": 1.9665116372568624e-05, "loss": 4.3326, "step": 3138 }, { "epoch": 0.11060898367645375, "grad_norm": 3.665177583694458, "learning_rate": 1.9664823431990304e-05, "loss": 7.356, "step": 3139 }, { "epoch": 0.11064422068941215, "grad_norm": 4.096410751342773, "learning_rate": 1.9664530365526433e-05, "loss": 6.5126, "step": 3140 }, { "epoch": 0.11067945770237057, "grad_norm": 2.4244675636291504, "learning_rate": 1.9664237173180818e-05, "loss": 4.8045, "step": 3141 }, { "epoch": 0.11071469471532898, "grad_norm": 2.6254079341888428, "learning_rate": 1.9663943854957287e-05, "loss": 4.9695, "step": 3142 }, { "epoch": 0.1107499317282874, "grad_norm": 4.56483268737793, "learning_rate": 1.9663650410859653e-05, "loss": 8.8436, "step": 3143 }, { "epoch": 0.1107851687412458, "grad_norm": 2.629065752029419, "learning_rate": 1.9663356840891743e-05, "loss": 4.4099, "step": 3144 }, { "epoch": 0.11082040575420421, "grad_norm": 4.574554443359375, "learning_rate": 1.9663063145057377e-05, "loss": 6.6277, "step": 3145 }, { "epoch": 0.11085564276716263, "grad_norm": 1.7659008502960205, "learning_rate": 1.9662769323360384e-05, "loss": 4.5004, "step": 3146 }, { "epoch": 0.11089087978012103, "grad_norm": 3.29337215423584, "learning_rate": 1.966247537580459e-05, "loss": 6.4429, "step": 3147 }, { "epoch": 0.11092611679307945, "grad_norm": 2.496642827987671, "learning_rate": 1.9662181302393823e-05, "loss": 5.7558, "step": 3148 }, { "epoch": 0.11096135380603786, "grad_norm": 4.316132068634033, "learning_rate": 1.9661887103131913e-05, "loss": 9.1174, "step": 3149 }, { "epoch": 0.11099659081899628, "grad_norm": 4.360889434814453, "learning_rate": 1.9661592778022696e-05, "loss": 7.2018, "step": 3150 }, { "epoch": 0.11103182783195469, "grad_norm": 5.585057735443115, "learning_rate": 1.966129832707e-05, "loss": 9.3131, "step": 3151 }, { "epoch": 0.11106706484491309, "grad_norm": 2.4746289253234863, "learning_rate": 1.9661003750277663e-05, "loss": 4.267, "step": 3152 }, { "epoch": 0.11110230185787151, "grad_norm": 6.579873561859131, "learning_rate": 1.966070904764952e-05, "loss": 9.984, "step": 3153 }, { "epoch": 0.11113753887082992, "grad_norm": 3.888532876968384, "learning_rate": 1.9660414219189413e-05, "loss": 7.6993, "step": 3154 }, { "epoch": 0.11117277588378834, "grad_norm": 2.908677339553833, "learning_rate": 1.966011926490118e-05, "loss": 5.293, "step": 3155 }, { "epoch": 0.11120801289674674, "grad_norm": 2.5630621910095215, "learning_rate": 1.9659824184788664e-05, "loss": 4.4245, "step": 3156 }, { "epoch": 0.11124324990970516, "grad_norm": 1.9488091468811035, "learning_rate": 1.9659528978855706e-05, "loss": 4.9559, "step": 3157 }, { "epoch": 0.11127848692266357, "grad_norm": 3.1749322414398193, "learning_rate": 1.965923364710615e-05, "loss": 6.5139, "step": 3158 }, { "epoch": 0.11131372393562197, "grad_norm": 2.273484468460083, "learning_rate": 1.9658938189543852e-05, "loss": 4.3387, "step": 3159 }, { "epoch": 0.11134896094858039, "grad_norm": 2.8247554302215576, "learning_rate": 1.965864260617265e-05, "loss": 4.1417, "step": 3160 }, { "epoch": 0.1113841979615388, "grad_norm": 2.482616901397705, "learning_rate": 1.96583468969964e-05, "loss": 4.9496, "step": 3161 }, { "epoch": 0.11141943497449722, "grad_norm": 2.1720662117004395, "learning_rate": 1.965805106201895e-05, "loss": 4.73, "step": 3162 }, { "epoch": 0.11145467198745562, "grad_norm": 3.01501202583313, "learning_rate": 1.965775510124415e-05, "loss": 5.9995, "step": 3163 }, { "epoch": 0.11148990900041403, "grad_norm": 3.6236865520477295, "learning_rate": 1.9657459014675866e-05, "loss": 4.8039, "step": 3164 }, { "epoch": 0.11152514601337245, "grad_norm": 4.32383918762207, "learning_rate": 1.9657162802317944e-05, "loss": 6.4206, "step": 3165 }, { "epoch": 0.11156038302633085, "grad_norm": 3.2551522254943848, "learning_rate": 1.965686646417425e-05, "loss": 6.9981, "step": 3166 }, { "epoch": 0.11159562003928927, "grad_norm": 5.817585468292236, "learning_rate": 1.9656570000248637e-05, "loss": 9.3052, "step": 3167 }, { "epoch": 0.11163085705224768, "grad_norm": 7.021423816680908, "learning_rate": 1.9656273410544973e-05, "loss": 9.7431, "step": 3168 }, { "epoch": 0.1116660940652061, "grad_norm": 2.752859115600586, "learning_rate": 1.9655976695067112e-05, "loss": 4.0567, "step": 3169 }, { "epoch": 0.1117013310781645, "grad_norm": 3.1274380683898926, "learning_rate": 1.9655679853818933e-05, "loss": 4.6341, "step": 3170 }, { "epoch": 0.11173656809112291, "grad_norm": 3.88958740234375, "learning_rate": 1.9655382886804287e-05, "loss": 4.8364, "step": 3171 }, { "epoch": 0.11177180510408133, "grad_norm": 3.720064640045166, "learning_rate": 1.965508579402705e-05, "loss": 6.2874, "step": 3172 }, { "epoch": 0.11180704211703973, "grad_norm": 3.530891180038452, "learning_rate": 1.9654788575491092e-05, "loss": 5.8346, "step": 3173 }, { "epoch": 0.11184227912999815, "grad_norm": 4.238663196563721, "learning_rate": 1.9654491231200284e-05, "loss": 9.2544, "step": 3174 }, { "epoch": 0.11187751614295656, "grad_norm": 3.9857890605926514, "learning_rate": 1.9654193761158497e-05, "loss": 6.4582, "step": 3175 }, { "epoch": 0.11191275315591498, "grad_norm": 2.887631893157959, "learning_rate": 1.9653896165369607e-05, "loss": 4.4644, "step": 3176 }, { "epoch": 0.11194799016887338, "grad_norm": 5.58311653137207, "learning_rate": 1.965359844383749e-05, "loss": 5.8323, "step": 3177 }, { "epoch": 0.11198322718183179, "grad_norm": 4.495744705200195, "learning_rate": 1.9653300596566023e-05, "loss": 6.7012, "step": 3178 }, { "epoch": 0.11201846419479021, "grad_norm": 4.344954013824463, "learning_rate": 1.9653002623559085e-05, "loss": 6.2519, "step": 3179 }, { "epoch": 0.11205370120774862, "grad_norm": 8.09172534942627, "learning_rate": 1.9652704524820557e-05, "loss": 9.8853, "step": 3180 }, { "epoch": 0.11208893822070704, "grad_norm": 4.273255825042725, "learning_rate": 1.9652406300354327e-05, "loss": 6.5363, "step": 3181 }, { "epoch": 0.11212417523366544, "grad_norm": 4.970042705535889, "learning_rate": 1.9652107950164274e-05, "loss": 4.0125, "step": 3182 }, { "epoch": 0.11215941224662386, "grad_norm": 2.12147855758667, "learning_rate": 1.9651809474254288e-05, "loss": 4.0106, "step": 3183 }, { "epoch": 0.11219464925958227, "grad_norm": 1.9322528839111328, "learning_rate": 1.9651510872628247e-05, "loss": 4.684, "step": 3184 }, { "epoch": 0.11222988627254067, "grad_norm": 4.242413520812988, "learning_rate": 1.9651212145290052e-05, "loss": 7.0317, "step": 3185 }, { "epoch": 0.11226512328549909, "grad_norm": 2.692967176437378, "learning_rate": 1.965091329224359e-05, "loss": 3.7908, "step": 3186 }, { "epoch": 0.1123003602984575, "grad_norm": 1.9490489959716797, "learning_rate": 1.965061431349275e-05, "loss": 4.7009, "step": 3187 }, { "epoch": 0.11233559731141592, "grad_norm": 2.5580620765686035, "learning_rate": 1.9650315209041434e-05, "loss": 4.414, "step": 3188 }, { "epoch": 0.11237083432437432, "grad_norm": 5.2327189445495605, "learning_rate": 1.965001597889353e-05, "loss": 6.5619, "step": 3189 }, { "epoch": 0.11240607133733273, "grad_norm": 2.2213470935821533, "learning_rate": 1.964971662305294e-05, "loss": 4.8316, "step": 3190 }, { "epoch": 0.11244130835029115, "grad_norm": 3.987637519836426, "learning_rate": 1.964941714152356e-05, "loss": 4.353, "step": 3191 }, { "epoch": 0.11247654536324955, "grad_norm": 2.3870058059692383, "learning_rate": 1.9649117534309292e-05, "loss": 4.0346, "step": 3192 }, { "epoch": 0.11251178237620797, "grad_norm": 2.9648029804229736, "learning_rate": 1.964881780141404e-05, "loss": 4.2201, "step": 3193 }, { "epoch": 0.11254701938916638, "grad_norm": 2.0411438941955566, "learning_rate": 1.964851794284171e-05, "loss": 4.0308, "step": 3194 }, { "epoch": 0.1125822564021248, "grad_norm": 4.443978309631348, "learning_rate": 1.96482179585962e-05, "loss": 6.97, "step": 3195 }, { "epoch": 0.1126174934150832, "grad_norm": 3.642685651779175, "learning_rate": 1.9647917848681424e-05, "loss": 7.2001, "step": 3196 }, { "epoch": 0.11265273042804161, "grad_norm": 4.920929431915283, "learning_rate": 1.9647617613101293e-05, "loss": 8.8235, "step": 3197 }, { "epoch": 0.11268796744100003, "grad_norm": 3.887687921524048, "learning_rate": 1.964731725185971e-05, "loss": 6.3383, "step": 3198 }, { "epoch": 0.11272320445395843, "grad_norm": 2.194399118423462, "learning_rate": 1.964701676496059e-05, "loss": 4.6148, "step": 3199 }, { "epoch": 0.11275844146691685, "grad_norm": 2.8992271423339844, "learning_rate": 1.964671615240785e-05, "loss": 3.7866, "step": 3200 }, { "epoch": 0.11279367847987526, "grad_norm": 3.2377259731292725, "learning_rate": 1.9646415414205403e-05, "loss": 3.6123, "step": 3201 }, { "epoch": 0.11282891549283368, "grad_norm": 5.313608169555664, "learning_rate": 1.9646114550357164e-05, "loss": 7.4896, "step": 3202 }, { "epoch": 0.11286415250579208, "grad_norm": 3.959839105606079, "learning_rate": 1.9645813560867057e-05, "loss": 3.738, "step": 3203 }, { "epoch": 0.11289938951875049, "grad_norm": 3.5000436305999756, "learning_rate": 1.9645512445739e-05, "loss": 4.1332, "step": 3204 }, { "epoch": 0.11293462653170891, "grad_norm": 3.9897897243499756, "learning_rate": 1.9645211204976914e-05, "loss": 7.2014, "step": 3205 }, { "epoch": 0.11296986354466731, "grad_norm": 2.8191943168640137, "learning_rate": 1.9644909838584722e-05, "loss": 4.4192, "step": 3206 }, { "epoch": 0.11300510055762573, "grad_norm": 4.565830707550049, "learning_rate": 1.9644608346566353e-05, "loss": 8.6732, "step": 3207 }, { "epoch": 0.11304033757058414, "grad_norm": 2.1982533931732178, "learning_rate": 1.964430672892573e-05, "loss": 4.1252, "step": 3208 }, { "epoch": 0.11307557458354256, "grad_norm": 2.5751121044158936, "learning_rate": 1.9644004985666784e-05, "loss": 4.5081, "step": 3209 }, { "epoch": 0.11311081159650097, "grad_norm": 3.7113537788391113, "learning_rate": 1.9643703116793444e-05, "loss": 7.3778, "step": 3210 }, { "epoch": 0.11314604860945937, "grad_norm": 2.563535690307617, "learning_rate": 1.964340112230964e-05, "loss": 4.0871, "step": 3211 }, { "epoch": 0.11318128562241779, "grad_norm": 3.0026023387908936, "learning_rate": 1.9643099002219313e-05, "loss": 6.7433, "step": 3212 }, { "epoch": 0.1132165226353762, "grad_norm": 4.276597023010254, "learning_rate": 1.9642796756526393e-05, "loss": 6.4135, "step": 3213 }, { "epoch": 0.11325175964833462, "grad_norm": 3.749917507171631, "learning_rate": 1.9642494385234813e-05, "loss": 5.3884, "step": 3214 }, { "epoch": 0.11328699666129302, "grad_norm": 4.235332012176514, "learning_rate": 1.964219188834852e-05, "loss": 7.2136, "step": 3215 }, { "epoch": 0.11332223367425143, "grad_norm": 6.563230037689209, "learning_rate": 1.9641889265871444e-05, "loss": 6.7906, "step": 3216 }, { "epoch": 0.11335747068720985, "grad_norm": 2.7467422485351562, "learning_rate": 1.9641586517807534e-05, "loss": 3.9127, "step": 3217 }, { "epoch": 0.11339270770016825, "grad_norm": 3.734947681427002, "learning_rate": 1.9641283644160735e-05, "loss": 4.8486, "step": 3218 }, { "epoch": 0.11342794471312667, "grad_norm": 7.05629825592041, "learning_rate": 1.9640980644934985e-05, "loss": 11.0842, "step": 3219 }, { "epoch": 0.11346318172608508, "grad_norm": 2.5254805088043213, "learning_rate": 1.9640677520134236e-05, "loss": 4.1412, "step": 3220 }, { "epoch": 0.1134984187390435, "grad_norm": 3.466524362564087, "learning_rate": 1.964037426976243e-05, "loss": 6.7619, "step": 3221 }, { "epoch": 0.1135336557520019, "grad_norm": 2.893559694290161, "learning_rate": 1.9640070893823528e-05, "loss": 8.6036, "step": 3222 }, { "epoch": 0.11356889276496031, "grad_norm": 5.387435436248779, "learning_rate": 1.963976739232147e-05, "loss": 3.514, "step": 3223 }, { "epoch": 0.11360412977791873, "grad_norm": 4.089903831481934, "learning_rate": 1.9639463765260216e-05, "loss": 5.8688, "step": 3224 }, { "epoch": 0.11363936679087713, "grad_norm": 4.969803810119629, "learning_rate": 1.9639160012643716e-05, "loss": 7.3505, "step": 3225 }, { "epoch": 0.11367460380383555, "grad_norm": 2.52140212059021, "learning_rate": 1.9638856134475932e-05, "loss": 4.3912, "step": 3226 }, { "epoch": 0.11370984081679396, "grad_norm": 3.503106117248535, "learning_rate": 1.9638552130760817e-05, "loss": 7.2058, "step": 3227 }, { "epoch": 0.11374507782975238, "grad_norm": 2.0154664516448975, "learning_rate": 1.9638248001502334e-05, "loss": 4.1507, "step": 3228 }, { "epoch": 0.11378031484271078, "grad_norm": 3.951103687286377, "learning_rate": 1.963794374670444e-05, "loss": 6.6383, "step": 3229 }, { "epoch": 0.11381555185566919, "grad_norm": 2.110032558441162, "learning_rate": 1.9637639366371104e-05, "loss": 4.2783, "step": 3230 }, { "epoch": 0.11385078886862761, "grad_norm": 2.3538262844085693, "learning_rate": 1.9637334860506286e-05, "loss": 3.938, "step": 3231 }, { "epoch": 0.11388602588158601, "grad_norm": 7.766003608703613, "learning_rate": 1.963703022911396e-05, "loss": 8.8587, "step": 3232 }, { "epoch": 0.11392126289454443, "grad_norm": 3.5183980464935303, "learning_rate": 1.963672547219808e-05, "loss": 3.9237, "step": 3233 }, { "epoch": 0.11395649990750284, "grad_norm": 4.235241413116455, "learning_rate": 1.9636420589762622e-05, "loss": 7.0534, "step": 3234 }, { "epoch": 0.11399173692046126, "grad_norm": 8.607869148254395, "learning_rate": 1.9636115581811564e-05, "loss": 6.1864, "step": 3235 }, { "epoch": 0.11402697393341966, "grad_norm": 4.623208999633789, "learning_rate": 1.963581044834887e-05, "loss": 6.7522, "step": 3236 }, { "epoch": 0.11406221094637807, "grad_norm": 2.7167108058929443, "learning_rate": 1.963550518937852e-05, "loss": 4.1644, "step": 3237 }, { "epoch": 0.11409744795933649, "grad_norm": 3.4258337020874023, "learning_rate": 1.9635199804904488e-05, "loss": 5.5651, "step": 3238 }, { "epoch": 0.1141326849722949, "grad_norm": 8.468326568603516, "learning_rate": 1.9634894294930747e-05, "loss": 5.4886, "step": 3239 }, { "epoch": 0.11416792198525332, "grad_norm": 4.07818078994751, "learning_rate": 1.9634588659461283e-05, "loss": 6.3505, "step": 3240 }, { "epoch": 0.11420315899821172, "grad_norm": 5.553788661956787, "learning_rate": 1.9634282898500075e-05, "loss": 6.4203, "step": 3241 }, { "epoch": 0.11423839601117013, "grad_norm": 12.469511985778809, "learning_rate": 1.9633977012051106e-05, "loss": 9.2285, "step": 3242 }, { "epoch": 0.11427363302412855, "grad_norm": 4.398253440856934, "learning_rate": 1.9633671000118356e-05, "loss": 6.8149, "step": 3243 }, { "epoch": 0.11430887003708695, "grad_norm": 2.7621445655822754, "learning_rate": 1.9633364862705817e-05, "loss": 3.7771, "step": 3244 }, { "epoch": 0.11434410705004537, "grad_norm": 5.213431358337402, "learning_rate": 1.9633058599817473e-05, "loss": 6.1863, "step": 3245 }, { "epoch": 0.11437934406300378, "grad_norm": 2.3031094074249268, "learning_rate": 1.963275221145731e-05, "loss": 4.7924, "step": 3246 }, { "epoch": 0.1144145810759622, "grad_norm": 3.491670608520508, "learning_rate": 1.9632445697629324e-05, "loss": 6.5749, "step": 3247 }, { "epoch": 0.1144498180889206, "grad_norm": 3.308929204940796, "learning_rate": 1.963213905833751e-05, "loss": 3.6851, "step": 3248 }, { "epoch": 0.11448505510187901, "grad_norm": 2.228703022003174, "learning_rate": 1.9631832293585856e-05, "loss": 4.0674, "step": 3249 }, { "epoch": 0.11452029211483743, "grad_norm": 3.066523551940918, "learning_rate": 1.963152540337836e-05, "loss": 3.6199, "step": 3250 }, { "epoch": 0.11455552912779583, "grad_norm": 2.6249277591705322, "learning_rate": 1.9631218387719017e-05, "loss": 4.4293, "step": 3251 }, { "epoch": 0.11459076614075425, "grad_norm": 3.5533151626586914, "learning_rate": 1.9630911246611825e-05, "loss": 6.4195, "step": 3252 }, { "epoch": 0.11462600315371266, "grad_norm": 3.0218756198883057, "learning_rate": 1.9630603980060793e-05, "loss": 6.2182, "step": 3253 }, { "epoch": 0.11466124016667108, "grad_norm": 2.896475076675415, "learning_rate": 1.9630296588069913e-05, "loss": 5.403, "step": 3254 }, { "epoch": 0.11469647717962948, "grad_norm": 2.276446580886841, "learning_rate": 1.9629989070643195e-05, "loss": 5.7856, "step": 3255 }, { "epoch": 0.11473171419258789, "grad_norm": 7.203536510467529, "learning_rate": 1.9629681427784643e-05, "loss": 6.9148, "step": 3256 }, { "epoch": 0.11476695120554631, "grad_norm": 5.067115783691406, "learning_rate": 1.9629373659498262e-05, "loss": 9.8748, "step": 3257 }, { "epoch": 0.11480218821850471, "grad_norm": 3.5966320037841797, "learning_rate": 1.9629065765788063e-05, "loss": 3.8669, "step": 3258 }, { "epoch": 0.11483742523146313, "grad_norm": 2.6373913288116455, "learning_rate": 1.962875774665806e-05, "loss": 4.0942, "step": 3259 }, { "epoch": 0.11487266224442154, "grad_norm": 2.8813531398773193, "learning_rate": 1.9628449602112257e-05, "loss": 4.4979, "step": 3260 }, { "epoch": 0.11490789925737996, "grad_norm": 7.440052032470703, "learning_rate": 1.962814133215467e-05, "loss": 6.1638, "step": 3261 }, { "epoch": 0.11494313627033836, "grad_norm": 5.145183563232422, "learning_rate": 1.9627832936789318e-05, "loss": 6.0919, "step": 3262 }, { "epoch": 0.11497837328329677, "grad_norm": 12.759654998779297, "learning_rate": 1.9627524416020215e-05, "loss": 10.1685, "step": 3263 }, { "epoch": 0.11501361029625519, "grad_norm": 7.460165977478027, "learning_rate": 1.962721576985138e-05, "loss": 9.1502, "step": 3264 }, { "epoch": 0.1150488473092136, "grad_norm": 3.1970436573028564, "learning_rate": 1.9626906998286834e-05, "loss": 3.8979, "step": 3265 }, { "epoch": 0.11508408432217201, "grad_norm": 2.5296740531921387, "learning_rate": 1.96265981013306e-05, "loss": 4.1497, "step": 3266 }, { "epoch": 0.11511932133513042, "grad_norm": 4.58578634262085, "learning_rate": 1.9626289078986695e-05, "loss": 6.6127, "step": 3267 }, { "epoch": 0.11515455834808883, "grad_norm": 1.726863145828247, "learning_rate": 1.9625979931259153e-05, "loss": 4.0634, "step": 3268 }, { "epoch": 0.11518979536104725, "grad_norm": 8.457132339477539, "learning_rate": 1.962567065815199e-05, "loss": 8.586, "step": 3269 }, { "epoch": 0.11522503237400565, "grad_norm": 5.478322982788086, "learning_rate": 1.9625361259669246e-05, "loss": 5.9142, "step": 3270 }, { "epoch": 0.11526026938696407, "grad_norm": 5.405611038208008, "learning_rate": 1.9625051735814945e-05, "loss": 6.6175, "step": 3271 }, { "epoch": 0.11529550639992248, "grad_norm": 4.7051005363464355, "learning_rate": 1.962474208659312e-05, "loss": 6.6301, "step": 3272 }, { "epoch": 0.1153307434128809, "grad_norm": 4.524657726287842, "learning_rate": 1.96244323120078e-05, "loss": 6.8724, "step": 3273 }, { "epoch": 0.1153659804258393, "grad_norm": 2.58618426322937, "learning_rate": 1.9624122412063024e-05, "loss": 3.5014, "step": 3274 }, { "epoch": 0.1154012174387977, "grad_norm": 4.487634658813477, "learning_rate": 1.962381238676283e-05, "loss": 7.1738, "step": 3275 }, { "epoch": 0.11543645445175613, "grad_norm": 3.413637638092041, "learning_rate": 1.9623502236111252e-05, "loss": 4.5765, "step": 3276 }, { "epoch": 0.11547169146471453, "grad_norm": 2.03999924659729, "learning_rate": 1.962319196011233e-05, "loss": 4.9156, "step": 3277 }, { "epoch": 0.11550692847767295, "grad_norm": 5.310688495635986, "learning_rate": 1.962288155877011e-05, "loss": 8.7867, "step": 3278 }, { "epoch": 0.11554216549063136, "grad_norm": 2.353757619857788, "learning_rate": 1.9622571032088632e-05, "loss": 4.3897, "step": 3279 }, { "epoch": 0.11557740250358978, "grad_norm": 6.5077948570251465, "learning_rate": 1.9622260380071938e-05, "loss": 11.791, "step": 3280 }, { "epoch": 0.11561263951654818, "grad_norm": 2.1625826358795166, "learning_rate": 1.9621949602724078e-05, "loss": 5.1397, "step": 3281 }, { "epoch": 0.11564787652950659, "grad_norm": 3.7232940196990967, "learning_rate": 1.9621638700049098e-05, "loss": 6.8947, "step": 3282 }, { "epoch": 0.11568311354246501, "grad_norm": 4.160206317901611, "learning_rate": 1.962132767205105e-05, "loss": 6.8153, "step": 3283 }, { "epoch": 0.11571835055542341, "grad_norm": 2.344501495361328, "learning_rate": 1.9621016518733983e-05, "loss": 4.4696, "step": 3284 }, { "epoch": 0.11575358756838183, "grad_norm": 5.468149185180664, "learning_rate": 1.9620705240101952e-05, "loss": 9.313, "step": 3285 }, { "epoch": 0.11578882458134024, "grad_norm": 2.8187642097473145, "learning_rate": 1.9620393836159003e-05, "loss": 3.8234, "step": 3286 }, { "epoch": 0.11582406159429866, "grad_norm": 4.241171360015869, "learning_rate": 1.9620082306909208e-05, "loss": 6.2039, "step": 3287 }, { "epoch": 0.11585929860725706, "grad_norm": 4.311210632324219, "learning_rate": 1.961977065235661e-05, "loss": 7.2551, "step": 3288 }, { "epoch": 0.11589453562021547, "grad_norm": 4.85788106918335, "learning_rate": 1.9619458872505274e-05, "loss": 4.8997, "step": 3289 }, { "epoch": 0.11592977263317389, "grad_norm": 3.736252546310425, "learning_rate": 1.9619146967359262e-05, "loss": 7.9228, "step": 3290 }, { "epoch": 0.1159650096461323, "grad_norm": 2.562602996826172, "learning_rate": 1.9618834936922634e-05, "loss": 4.824, "step": 3291 }, { "epoch": 0.11600024665909071, "grad_norm": 8.840729713439941, "learning_rate": 1.9618522781199458e-05, "loss": 11.1292, "step": 3292 }, { "epoch": 0.11603548367204912, "grad_norm": 10.113720893859863, "learning_rate": 1.9618210500193798e-05, "loss": 8.7112, "step": 3293 }, { "epoch": 0.11607072068500753, "grad_norm": 6.289178371429443, "learning_rate": 1.9617898093909718e-05, "loss": 6.3551, "step": 3294 }, { "epoch": 0.11610595769796594, "grad_norm": 7.00360107421875, "learning_rate": 1.9617585562351295e-05, "loss": 6.5316, "step": 3295 }, { "epoch": 0.11614119471092435, "grad_norm": 5.944705009460449, "learning_rate": 1.961727290552259e-05, "loss": 7.0725, "step": 3296 }, { "epoch": 0.11617643172388277, "grad_norm": 6.883766174316406, "learning_rate": 1.9616960123427682e-05, "loss": 6.3156, "step": 3297 }, { "epoch": 0.11621166873684118, "grad_norm": 5.069067001342773, "learning_rate": 1.9616647216070644e-05, "loss": 6.8968, "step": 3298 }, { "epoch": 0.1162469057497996, "grad_norm": 2.357891798019409, "learning_rate": 1.9616334183455553e-05, "loss": 4.2639, "step": 3299 }, { "epoch": 0.116282142762758, "grad_norm": 8.959622383117676, "learning_rate": 1.961602102558648e-05, "loss": 6.8772, "step": 3300 }, { "epoch": 0.1163173797757164, "grad_norm": 3.5858750343322754, "learning_rate": 1.9615707742467508e-05, "loss": 3.9895, "step": 3301 }, { "epoch": 0.11635261678867483, "grad_norm": 7.125631332397461, "learning_rate": 1.9615394334102722e-05, "loss": 9.3308, "step": 3302 }, { "epoch": 0.11638785380163323, "grad_norm": 1.9960013628005981, "learning_rate": 1.9615080800496198e-05, "loss": 4.9378, "step": 3303 }, { "epoch": 0.11642309081459165, "grad_norm": 3.195366859436035, "learning_rate": 1.9614767141652024e-05, "loss": 4.1528, "step": 3304 }, { "epoch": 0.11645832782755006, "grad_norm": 2.7495503425598145, "learning_rate": 1.961445335757428e-05, "loss": 3.9642, "step": 3305 }, { "epoch": 0.11649356484050848, "grad_norm": 8.485946655273438, "learning_rate": 1.9614139448267057e-05, "loss": 9.3489, "step": 3306 }, { "epoch": 0.11652880185346688, "grad_norm": 2.232221841812134, "learning_rate": 1.9613825413734443e-05, "loss": 4.9215, "step": 3307 }, { "epoch": 0.11656403886642529, "grad_norm": 4.717319488525391, "learning_rate": 1.961351125398053e-05, "loss": 6.8124, "step": 3308 }, { "epoch": 0.11659927587938371, "grad_norm": 4.460480213165283, "learning_rate": 1.9613196969009408e-05, "loss": 6.1048, "step": 3309 }, { "epoch": 0.11663451289234211, "grad_norm": 4.152261734008789, "learning_rate": 1.961288255882517e-05, "loss": 6.2936, "step": 3310 }, { "epoch": 0.11666974990530053, "grad_norm": 2.1548397541046143, "learning_rate": 1.961256802343191e-05, "loss": 3.9449, "step": 3311 }, { "epoch": 0.11670498691825894, "grad_norm": 3.4089767932891846, "learning_rate": 1.9612253362833732e-05, "loss": 4.3827, "step": 3312 }, { "epoch": 0.11674022393121736, "grad_norm": 4.884377479553223, "learning_rate": 1.9611938577034726e-05, "loss": 9.3257, "step": 3313 }, { "epoch": 0.11677546094417576, "grad_norm": 4.384761333465576, "learning_rate": 1.9611623666038998e-05, "loss": 9.5358, "step": 3314 }, { "epoch": 0.11681069795713417, "grad_norm": 3.8168013095855713, "learning_rate": 1.9611308629850642e-05, "loss": 3.7911, "step": 3315 }, { "epoch": 0.11684593497009259, "grad_norm": 2.529902219772339, "learning_rate": 1.961099346847377e-05, "loss": 4.3316, "step": 3316 }, { "epoch": 0.116881171983051, "grad_norm": 3.8136372566223145, "learning_rate": 1.9610678181912487e-05, "loss": 4.1803, "step": 3317 }, { "epoch": 0.11691640899600941, "grad_norm": 5.838756084442139, "learning_rate": 1.9610362770170895e-05, "loss": 6.0903, "step": 3318 }, { "epoch": 0.11695164600896782, "grad_norm": 7.001745223999023, "learning_rate": 1.9610047233253103e-05, "loss": 5.6263, "step": 3319 }, { "epoch": 0.11698688302192622, "grad_norm": 2.283879041671753, "learning_rate": 1.9609731571163218e-05, "loss": 4.7099, "step": 3320 }, { "epoch": 0.11702212003488464, "grad_norm": 10.855831146240234, "learning_rate": 1.960941578390536e-05, "loss": 9.301, "step": 3321 }, { "epoch": 0.11705735704784305, "grad_norm": 2.960991621017456, "learning_rate": 1.9609099871483636e-05, "loss": 4.4845, "step": 3322 }, { "epoch": 0.11709259406080147, "grad_norm": 2.4752755165100098, "learning_rate": 1.9608783833902162e-05, "loss": 4.3127, "step": 3323 }, { "epoch": 0.11712783107375987, "grad_norm": 5.45631217956543, "learning_rate": 1.960846767116506e-05, "loss": 6.4761, "step": 3324 }, { "epoch": 0.1171630680867183, "grad_norm": 5.630247116088867, "learning_rate": 1.9608151383276434e-05, "loss": 9.4737, "step": 3325 }, { "epoch": 0.1171983050996767, "grad_norm": 3.53605055809021, "learning_rate": 1.9607834970240417e-05, "loss": 7.2392, "step": 3326 }, { "epoch": 0.1172335421126351, "grad_norm": 2.906560182571411, "learning_rate": 1.9607518432061123e-05, "loss": 3.5315, "step": 3327 }, { "epoch": 0.11726877912559353, "grad_norm": 6.357539653778076, "learning_rate": 1.960720176874268e-05, "loss": 6.1263, "step": 3328 }, { "epoch": 0.11730401613855193, "grad_norm": 5.013306617736816, "learning_rate": 1.960688498028921e-05, "loss": 5.6501, "step": 3329 }, { "epoch": 0.11733925315151035, "grad_norm": 3.415881633758545, "learning_rate": 1.960656806670484e-05, "loss": 4.3987, "step": 3330 }, { "epoch": 0.11737449016446876, "grad_norm": 5.075587272644043, "learning_rate": 1.9606251027993695e-05, "loss": 6.5938, "step": 3331 }, { "epoch": 0.11740972717742718, "grad_norm": 3.0728650093078613, "learning_rate": 1.9605933864159908e-05, "loss": 4.8663, "step": 3332 }, { "epoch": 0.11744496419038558, "grad_norm": 2.341499090194702, "learning_rate": 1.9605616575207607e-05, "loss": 3.812, "step": 3333 }, { "epoch": 0.11748020120334399, "grad_norm": 10.15494155883789, "learning_rate": 1.960529916114093e-05, "loss": 12.6402, "step": 3334 }, { "epoch": 0.1175154382163024, "grad_norm": 7.149826526641846, "learning_rate": 1.9604981621964005e-05, "loss": 8.6882, "step": 3335 }, { "epoch": 0.11755067522926081, "grad_norm": 6.995563507080078, "learning_rate": 1.9604663957680974e-05, "loss": 5.9452, "step": 3336 }, { "epoch": 0.11758591224221923, "grad_norm": 6.8017258644104, "learning_rate": 1.9604346168295967e-05, "loss": 5.7864, "step": 3337 }, { "epoch": 0.11762114925517764, "grad_norm": 8.772948265075684, "learning_rate": 1.960402825381313e-05, "loss": 6.9437, "step": 3338 }, { "epoch": 0.11765638626813606, "grad_norm": 8.01378059387207, "learning_rate": 1.9603710214236602e-05, "loss": 6.1354, "step": 3339 }, { "epoch": 0.11769162328109446, "grad_norm": 2.931676149368286, "learning_rate": 1.9603392049570527e-05, "loss": 4.1075, "step": 3340 }, { "epoch": 0.11772686029405287, "grad_norm": 2.107243299484253, "learning_rate": 1.9603073759819045e-05, "loss": 4.1502, "step": 3341 }, { "epoch": 0.11776209730701129, "grad_norm": 3.0977718830108643, "learning_rate": 1.9602755344986302e-05, "loss": 3.8462, "step": 3342 }, { "epoch": 0.1177973343199697, "grad_norm": 3.7802345752716064, "learning_rate": 1.9602436805076453e-05, "loss": 7.3752, "step": 3343 }, { "epoch": 0.11783257133292811, "grad_norm": 2.3605539798736572, "learning_rate": 1.9602118140093638e-05, "loss": 4.3291, "step": 3344 }, { "epoch": 0.11786780834588652, "grad_norm": 2.5193395614624023, "learning_rate": 1.960179935004201e-05, "loss": 4.8484, "step": 3345 }, { "epoch": 0.11790304535884492, "grad_norm": 5.2797651290893555, "learning_rate": 1.9601480434925723e-05, "loss": 6.5288, "step": 3346 }, { "epoch": 0.11793828237180334, "grad_norm": 2.3926842212677, "learning_rate": 1.9601161394748932e-05, "loss": 4.2697, "step": 3347 }, { "epoch": 0.11797351938476175, "grad_norm": 5.7005839347839355, "learning_rate": 1.960084222951579e-05, "loss": 8.852, "step": 3348 }, { "epoch": 0.11800875639772017, "grad_norm": 3.8173229694366455, "learning_rate": 1.9600522939230454e-05, "loss": 6.6409, "step": 3349 }, { "epoch": 0.11804399341067857, "grad_norm": 3.5090110301971436, "learning_rate": 1.9600203523897085e-05, "loss": 8.2242, "step": 3350 }, { "epoch": 0.118079230423637, "grad_norm": 7.565164566040039, "learning_rate": 1.9599883983519844e-05, "loss": 9.3745, "step": 3351 }, { "epoch": 0.1181144674365954, "grad_norm": 5.5799665451049805, "learning_rate": 1.9599564318102887e-05, "loss": 6.907, "step": 3352 }, { "epoch": 0.1181497044495538, "grad_norm": 6.342134475708008, "learning_rate": 1.9599244527650386e-05, "loss": 6.2075, "step": 3353 }, { "epoch": 0.11818494146251222, "grad_norm": 5.91518497467041, "learning_rate": 1.95989246121665e-05, "loss": 7.8285, "step": 3354 }, { "epoch": 0.11822017847547063, "grad_norm": 1.790367841720581, "learning_rate": 1.9598604571655397e-05, "loss": 3.6941, "step": 3355 }, { "epoch": 0.11825541548842905, "grad_norm": 2.1722185611724854, "learning_rate": 1.959828440612125e-05, "loss": 3.9436, "step": 3356 }, { "epoch": 0.11829065250138746, "grad_norm": 2.194903612136841, "learning_rate": 1.9597964115568226e-05, "loss": 3.79, "step": 3357 }, { "epoch": 0.11832588951434587, "grad_norm": 3.1048667430877686, "learning_rate": 1.9597643700000494e-05, "loss": 6.5741, "step": 3358 }, { "epoch": 0.11836112652730428, "grad_norm": 3.2235608100891113, "learning_rate": 1.9597323159422232e-05, "loss": 7.3934, "step": 3359 }, { "epoch": 0.11839636354026269, "grad_norm": 2.9970014095306396, "learning_rate": 1.959700249383761e-05, "loss": 6.1223, "step": 3360 }, { "epoch": 0.1184316005532211, "grad_norm": 3.8768951892852783, "learning_rate": 1.959668170325081e-05, "loss": 10.8499, "step": 3361 }, { "epoch": 0.11846683756617951, "grad_norm": 2.1973016262054443, "learning_rate": 1.9596360787666012e-05, "loss": 4.0706, "step": 3362 }, { "epoch": 0.11850207457913793, "grad_norm": 4.502163887023926, "learning_rate": 1.959603974708739e-05, "loss": 8.9199, "step": 3363 }, { "epoch": 0.11853731159209634, "grad_norm": 2.800128221511841, "learning_rate": 1.959571858151913e-05, "loss": 4.1913, "step": 3364 }, { "epoch": 0.11857254860505476, "grad_norm": 4.057764530181885, "learning_rate": 1.9595397290965406e-05, "loss": 8.3808, "step": 3365 }, { "epoch": 0.11860778561801316, "grad_norm": 3.7530179023742676, "learning_rate": 1.9595075875430418e-05, "loss": 7.1391, "step": 3366 }, { "epoch": 0.11864302263097157, "grad_norm": 3.6415369510650635, "learning_rate": 1.959475433491834e-05, "loss": 6.8042, "step": 3367 }, { "epoch": 0.11867825964392999, "grad_norm": 2.54419207572937, "learning_rate": 1.9594432669433368e-05, "loss": 3.8098, "step": 3368 }, { "epoch": 0.11871349665688839, "grad_norm": 6.550652027130127, "learning_rate": 1.9594110878979683e-05, "loss": 8.9764, "step": 3369 }, { "epoch": 0.11874873366984681, "grad_norm": 4.959092617034912, "learning_rate": 1.9593788963561485e-05, "loss": 12.3618, "step": 3370 }, { "epoch": 0.11878397068280522, "grad_norm": 1.8319395780563354, "learning_rate": 1.9593466923182966e-05, "loss": 5.2139, "step": 3371 }, { "epoch": 0.11881920769576362, "grad_norm": 3.965956211090088, "learning_rate": 1.9593144757848314e-05, "loss": 5.6732, "step": 3372 }, { "epoch": 0.11885444470872204, "grad_norm": 1.881116509437561, "learning_rate": 1.9592822467561733e-05, "loss": 4.292, "step": 3373 }, { "epoch": 0.11888968172168045, "grad_norm": 1.879286766052246, "learning_rate": 1.9592500052327414e-05, "loss": 5.1205, "step": 3374 }, { "epoch": 0.11892491873463887, "grad_norm": 2.35577130317688, "learning_rate": 1.9592177512149563e-05, "loss": 6.5412, "step": 3375 }, { "epoch": 0.11896015574759727, "grad_norm": 1.983557105064392, "learning_rate": 1.959185484703238e-05, "loss": 3.777, "step": 3376 }, { "epoch": 0.1189953927605557, "grad_norm": 4.173167705535889, "learning_rate": 1.959153205698006e-05, "loss": 7.0109, "step": 3377 }, { "epoch": 0.1190306297735141, "grad_norm": 3.951353073120117, "learning_rate": 1.9591209141996818e-05, "loss": 5.642, "step": 3378 }, { "epoch": 0.1190658667864725, "grad_norm": 5.114135265350342, "learning_rate": 1.959088610208685e-05, "loss": 6.8076, "step": 3379 }, { "epoch": 0.11910110379943092, "grad_norm": 2.281482219696045, "learning_rate": 1.9590562937254375e-05, "loss": 3.6033, "step": 3380 }, { "epoch": 0.11913634081238933, "grad_norm": 3.2322309017181396, "learning_rate": 1.959023964750359e-05, "loss": 7.0619, "step": 3381 }, { "epoch": 0.11917157782534775, "grad_norm": 2.2873120307922363, "learning_rate": 1.9589916232838716e-05, "loss": 4.0399, "step": 3382 }, { "epoch": 0.11920681483830615, "grad_norm": 3.3212761878967285, "learning_rate": 1.9589592693263963e-05, "loss": 6.1893, "step": 3383 }, { "epoch": 0.11924205185126457, "grad_norm": 3.478672742843628, "learning_rate": 1.958926902878354e-05, "loss": 9.1139, "step": 3384 }, { "epoch": 0.11927728886422298, "grad_norm": 4.123788833618164, "learning_rate": 1.9588945239401666e-05, "loss": 9.7048, "step": 3385 }, { "epoch": 0.11931252587718139, "grad_norm": 2.341123580932617, "learning_rate": 1.958862132512256e-05, "loss": 3.9628, "step": 3386 }, { "epoch": 0.1193477628901398, "grad_norm": 6.851607799530029, "learning_rate": 1.958829728595044e-05, "loss": 6.1119, "step": 3387 }, { "epoch": 0.11938299990309821, "grad_norm": 3.661654472351074, "learning_rate": 1.9587973121889527e-05, "loss": 5.2965, "step": 3388 }, { "epoch": 0.11941823691605663, "grad_norm": 2.004652976989746, "learning_rate": 1.9587648832944045e-05, "loss": 3.4905, "step": 3389 }, { "epoch": 0.11945347392901504, "grad_norm": 9.510855674743652, "learning_rate": 1.958732441911821e-05, "loss": 6.1564, "step": 3390 }, { "epoch": 0.11948871094197346, "grad_norm": 2.9772560596466064, "learning_rate": 1.958699988041626e-05, "loss": 3.9015, "step": 3391 }, { "epoch": 0.11952394795493186, "grad_norm": 5.784638404846191, "learning_rate": 1.958667521684241e-05, "loss": 9.0445, "step": 3392 }, { "epoch": 0.11955918496789027, "grad_norm": 4.983401298522949, "learning_rate": 1.9586350428400895e-05, "loss": 6.312, "step": 3393 }, { "epoch": 0.11959442198084869, "grad_norm": 5.01987361907959, "learning_rate": 1.9586025515095948e-05, "loss": 6.5248, "step": 3394 }, { "epoch": 0.11962965899380709, "grad_norm": 4.310140609741211, "learning_rate": 1.9585700476931796e-05, "loss": 7.7114, "step": 3395 }, { "epoch": 0.11966489600676551, "grad_norm": 2.016716480255127, "learning_rate": 1.9585375313912672e-05, "loss": 4.1836, "step": 3396 }, { "epoch": 0.11970013301972392, "grad_norm": 8.22018814086914, "learning_rate": 1.9585050026042817e-05, "loss": 7.9231, "step": 3397 }, { "epoch": 0.11973537003268232, "grad_norm": 5.341711044311523, "learning_rate": 1.958472461332646e-05, "loss": 6.8878, "step": 3398 }, { "epoch": 0.11977060704564074, "grad_norm": 4.08292293548584, "learning_rate": 1.9584399075767848e-05, "loss": 9.7039, "step": 3399 }, { "epoch": 0.11980584405859915, "grad_norm": 2.4650890827178955, "learning_rate": 1.9584073413371215e-05, "loss": 3.9981, "step": 3400 }, { "epoch": 0.11984108107155757, "grad_norm": 2.053311347961426, "learning_rate": 1.9583747626140804e-05, "loss": 4.5376, "step": 3401 }, { "epoch": 0.11987631808451597, "grad_norm": 6.119431018829346, "learning_rate": 1.958342171408086e-05, "loss": 7.0141, "step": 3402 }, { "epoch": 0.11991155509747439, "grad_norm": 2.168121814727783, "learning_rate": 1.958309567719563e-05, "loss": 3.7381, "step": 3403 }, { "epoch": 0.1199467921104328, "grad_norm": 3.154888391494751, "learning_rate": 1.9582769515489358e-05, "loss": 3.7428, "step": 3404 }, { "epoch": 0.1199820291233912, "grad_norm": 2.0495243072509766, "learning_rate": 1.958244322896629e-05, "loss": 4.203, "step": 3405 }, { "epoch": 0.12001726613634962, "grad_norm": 2.1188602447509766, "learning_rate": 1.958211681763068e-05, "loss": 5.3538, "step": 3406 }, { "epoch": 0.12005250314930803, "grad_norm": 2.2436881065368652, "learning_rate": 1.9581790281486777e-05, "loss": 4.3082, "step": 3407 }, { "epoch": 0.12008774016226645, "grad_norm": 4.596464157104492, "learning_rate": 1.9581463620538835e-05, "loss": 6.6451, "step": 3408 }, { "epoch": 0.12012297717522485, "grad_norm": 3.6248347759246826, "learning_rate": 1.958113683479111e-05, "loss": 5.9836, "step": 3409 }, { "epoch": 0.12015821418818327, "grad_norm": 2.6349520683288574, "learning_rate": 1.9580809924247856e-05, "loss": 3.8332, "step": 3410 }, { "epoch": 0.12019345120114168, "grad_norm": 2.635502576828003, "learning_rate": 1.9580482888913333e-05, "loss": 4.3365, "step": 3411 }, { "epoch": 0.12022868821410009, "grad_norm": 5.98813533782959, "learning_rate": 1.95801557287918e-05, "loss": 9.8211, "step": 3412 }, { "epoch": 0.1202639252270585, "grad_norm": 2.085982084274292, "learning_rate": 1.957982844388752e-05, "loss": 4.5283, "step": 3413 }, { "epoch": 0.12029916224001691, "grad_norm": 4.999022006988525, "learning_rate": 1.957950103420475e-05, "loss": 9.749, "step": 3414 }, { "epoch": 0.12033439925297533, "grad_norm": 4.7037034034729, "learning_rate": 1.9579173499747762e-05, "loss": 6.7264, "step": 3415 }, { "epoch": 0.12036963626593374, "grad_norm": 2.8260419368743896, "learning_rate": 1.957884584052082e-05, "loss": 6.2937, "step": 3416 }, { "epoch": 0.12040487327889215, "grad_norm": 6.683393478393555, "learning_rate": 1.9578518056528185e-05, "loss": 8.807, "step": 3417 }, { "epoch": 0.12044011029185056, "grad_norm": 5.214310169219971, "learning_rate": 1.9578190147774138e-05, "loss": 6.6149, "step": 3418 }, { "epoch": 0.12047534730480897, "grad_norm": 2.6013357639312744, "learning_rate": 1.957786211426294e-05, "loss": 3.4119, "step": 3419 }, { "epoch": 0.12051058431776739, "grad_norm": 3.166193723678589, "learning_rate": 1.957753395599887e-05, "loss": 6.6392, "step": 3420 }, { "epoch": 0.12054582133072579, "grad_norm": 9.045149803161621, "learning_rate": 1.9577205672986203e-05, "loss": 9.1175, "step": 3421 }, { "epoch": 0.12058105834368421, "grad_norm": 6.05070161819458, "learning_rate": 1.9576877265229207e-05, "loss": 9.0031, "step": 3422 }, { "epoch": 0.12061629535664262, "grad_norm": 5.687436580657959, "learning_rate": 1.957654873273217e-05, "loss": 6.3618, "step": 3423 }, { "epoch": 0.12065153236960102, "grad_norm": 4.751945972442627, "learning_rate": 1.9576220075499362e-05, "loss": 5.9468, "step": 3424 }, { "epoch": 0.12068676938255944, "grad_norm": 6.005201816558838, "learning_rate": 1.957589129353507e-05, "loss": 6.7017, "step": 3425 }, { "epoch": 0.12072200639551785, "grad_norm": 2.0681071281433105, "learning_rate": 1.957556238684357e-05, "loss": 4.4445, "step": 3426 }, { "epoch": 0.12075724340847627, "grad_norm": 2.182626962661743, "learning_rate": 1.9575233355429155e-05, "loss": 4.317, "step": 3427 }, { "epoch": 0.12079248042143467, "grad_norm": 6.305344581604004, "learning_rate": 1.9574904199296104e-05, "loss": 6.0453, "step": 3428 }, { "epoch": 0.12082771743439309, "grad_norm": 3.9272706508636475, "learning_rate": 1.9574574918448707e-05, "loss": 3.5215, "step": 3429 }, { "epoch": 0.1208629544473515, "grad_norm": 6.082813739776611, "learning_rate": 1.9574245512891253e-05, "loss": 8.2106, "step": 3430 }, { "epoch": 0.1208981914603099, "grad_norm": 6.87156867980957, "learning_rate": 1.9573915982628028e-05, "loss": 7.4524, "step": 3431 }, { "epoch": 0.12093342847326832, "grad_norm": 11.05324935913086, "learning_rate": 1.957358632766333e-05, "loss": 8.8359, "step": 3432 }, { "epoch": 0.12096866548622673, "grad_norm": 9.736062049865723, "learning_rate": 1.957325654800145e-05, "loss": 7.7516, "step": 3433 }, { "epoch": 0.12100390249918515, "grad_norm": 5.673093795776367, "learning_rate": 1.9572926643646687e-05, "loss": 7.194, "step": 3434 }, { "epoch": 0.12103913951214355, "grad_norm": 2.905646324157715, "learning_rate": 1.957259661460333e-05, "loss": 3.8137, "step": 3435 }, { "epoch": 0.12107437652510197, "grad_norm": 1.8929686546325684, "learning_rate": 1.9572266460875688e-05, "loss": 3.6096, "step": 3436 }, { "epoch": 0.12110961353806038, "grad_norm": 1.9760023355484009, "learning_rate": 1.957193618246805e-05, "loss": 4.3334, "step": 3437 }, { "epoch": 0.12114485055101878, "grad_norm": 4.761190414428711, "learning_rate": 1.9571605779384726e-05, "loss": 7.2529, "step": 3438 }, { "epoch": 0.1211800875639772, "grad_norm": 3.445845365524292, "learning_rate": 1.957127525163002e-05, "loss": 6.5999, "step": 3439 }, { "epoch": 0.12121532457693561, "grad_norm": 2.6490025520324707, "learning_rate": 1.9570944599208234e-05, "loss": 4.1699, "step": 3440 }, { "epoch": 0.12125056158989403, "grad_norm": 3.91477370262146, "learning_rate": 1.9570613822123678e-05, "loss": 6.7544, "step": 3441 }, { "epoch": 0.12128579860285243, "grad_norm": 2.8496298789978027, "learning_rate": 1.9570282920380653e-05, "loss": 4.1226, "step": 3442 }, { "epoch": 0.12132103561581085, "grad_norm": 4.396988868713379, "learning_rate": 1.9569951893983475e-05, "loss": 6.6657, "step": 3443 }, { "epoch": 0.12135627262876926, "grad_norm": 5.191524028778076, "learning_rate": 1.9569620742936455e-05, "loss": 8.5453, "step": 3444 }, { "epoch": 0.12139150964172767, "grad_norm": 2.847827911376953, "learning_rate": 1.9569289467243905e-05, "loss": 3.2274, "step": 3445 }, { "epoch": 0.12142674665468609, "grad_norm": 2.9256808757781982, "learning_rate": 1.9568958066910145e-05, "loss": 3.779, "step": 3446 }, { "epoch": 0.12146198366764449, "grad_norm": 6.444625377655029, "learning_rate": 1.956862654193948e-05, "loss": 7.4983, "step": 3447 }, { "epoch": 0.12149722068060291, "grad_norm": 3.1992950439453125, "learning_rate": 1.956829489233624e-05, "loss": 6.0594, "step": 3448 }, { "epoch": 0.12153245769356132, "grad_norm": 3.7957346439361572, "learning_rate": 1.9567963118104743e-05, "loss": 6.0121, "step": 3449 }, { "epoch": 0.12156769470651972, "grad_norm": 4.172027587890625, "learning_rate": 1.9567631219249305e-05, "loss": 6.0852, "step": 3450 }, { "epoch": 0.12160293171947814, "grad_norm": 3.494025707244873, "learning_rate": 1.9567299195774246e-05, "loss": 6.7152, "step": 3451 }, { "epoch": 0.12163816873243655, "grad_norm": 2.791334390640259, "learning_rate": 1.9566967047683903e-05, "loss": 3.0845, "step": 3452 }, { "epoch": 0.12167340574539497, "grad_norm": 1.9400272369384766, "learning_rate": 1.9566634774982596e-05, "loss": 4.5315, "step": 3453 }, { "epoch": 0.12170864275835337, "grad_norm": 3.159524917602539, "learning_rate": 1.956630237767465e-05, "loss": 6.2074, "step": 3454 }, { "epoch": 0.12174387977131179, "grad_norm": 2.104814052581787, "learning_rate": 1.9565969855764394e-05, "loss": 4.4769, "step": 3455 }, { "epoch": 0.1217791167842702, "grad_norm": 6.532188415527344, "learning_rate": 1.9565637209256168e-05, "loss": 5.5286, "step": 3456 }, { "epoch": 0.1218143537972286, "grad_norm": 5.778915882110596, "learning_rate": 1.9565304438154295e-05, "loss": 9.9656, "step": 3457 }, { "epoch": 0.12184959081018702, "grad_norm": 1.9560130834579468, "learning_rate": 1.9564971542463115e-05, "loss": 3.8771, "step": 3458 }, { "epoch": 0.12188482782314543, "grad_norm": 5.2521562576293945, "learning_rate": 1.9564638522186962e-05, "loss": 6.1692, "step": 3459 }, { "epoch": 0.12192006483610385, "grad_norm": 2.136955976486206, "learning_rate": 1.9564305377330175e-05, "loss": 4.6138, "step": 3460 }, { "epoch": 0.12195530184906225, "grad_norm": 3.7292752265930176, "learning_rate": 1.9563972107897087e-05, "loss": 5.5028, "step": 3461 }, { "epoch": 0.12199053886202067, "grad_norm": 2.6856472492218018, "learning_rate": 1.956363871389205e-05, "loss": 3.8749, "step": 3462 }, { "epoch": 0.12202577587497908, "grad_norm": 7.275889873504639, "learning_rate": 1.95633051953194e-05, "loss": 9.3958, "step": 3463 }, { "epoch": 0.12206101288793748, "grad_norm": 9.985550880432129, "learning_rate": 1.9562971552183478e-05, "loss": 11.9456, "step": 3464 }, { "epoch": 0.1220962499008959, "grad_norm": 5.255985260009766, "learning_rate": 1.9562637784488637e-05, "loss": 5.6589, "step": 3465 }, { "epoch": 0.12213148691385431, "grad_norm": 5.685122013092041, "learning_rate": 1.956230389223922e-05, "loss": 8.7042, "step": 3466 }, { "epoch": 0.12216672392681273, "grad_norm": 2.0519299507141113, "learning_rate": 1.9561969875439578e-05, "loss": 4.9556, "step": 3467 }, { "epoch": 0.12220196093977113, "grad_norm": 2.343440532684326, "learning_rate": 1.9561635734094055e-05, "loss": 4.0465, "step": 3468 }, { "epoch": 0.12223719795272955, "grad_norm": 2.221970796585083, "learning_rate": 1.9561301468207014e-05, "loss": 3.4985, "step": 3469 }, { "epoch": 0.12227243496568796, "grad_norm": 3.735102653503418, "learning_rate": 1.9560967077782803e-05, "loss": 4.4489, "step": 3470 }, { "epoch": 0.12230767197864637, "grad_norm": 5.632269859313965, "learning_rate": 1.9560632562825775e-05, "loss": 6.9392, "step": 3471 }, { "epoch": 0.12234290899160478, "grad_norm": 2.060637950897217, "learning_rate": 1.9560297923340294e-05, "loss": 3.8406, "step": 3472 }, { "epoch": 0.12237814600456319, "grad_norm": 10.555825233459473, "learning_rate": 1.9559963159330707e-05, "loss": 5.1261, "step": 3473 }, { "epoch": 0.12241338301752161, "grad_norm": 4.291122913360596, "learning_rate": 1.9559628270801388e-05, "loss": 6.9258, "step": 3474 }, { "epoch": 0.12244862003048002, "grad_norm": 2.5133516788482666, "learning_rate": 1.9559293257756695e-05, "loss": 5.8935, "step": 3475 }, { "epoch": 0.12248385704343842, "grad_norm": 4.158650875091553, "learning_rate": 1.9558958120200987e-05, "loss": 6.0457, "step": 3476 }, { "epoch": 0.12251909405639684, "grad_norm": 3.376743793487549, "learning_rate": 1.9558622858138633e-05, "loss": 5.6756, "step": 3477 }, { "epoch": 0.12255433106935525, "grad_norm": 3.0180370807647705, "learning_rate": 1.9558287471573996e-05, "loss": 6.1946, "step": 3478 }, { "epoch": 0.12258956808231367, "grad_norm": 3.4973676204681396, "learning_rate": 1.955795196051145e-05, "loss": 5.3831, "step": 3479 }, { "epoch": 0.12262480509527207, "grad_norm": 2.9731814861297607, "learning_rate": 1.955761632495536e-05, "loss": 3.8165, "step": 3480 }, { "epoch": 0.12266004210823049, "grad_norm": 3.5204906463623047, "learning_rate": 1.95572805649101e-05, "loss": 3.2883, "step": 3481 }, { "epoch": 0.1226952791211889, "grad_norm": 3.2040042877197266, "learning_rate": 1.9556944680380044e-05, "loss": 3.7414, "step": 3482 }, { "epoch": 0.1227305161341473, "grad_norm": 2.0590577125549316, "learning_rate": 1.9556608671369566e-05, "loss": 3.8741, "step": 3483 }, { "epoch": 0.12276575314710572, "grad_norm": 2.8103082180023193, "learning_rate": 1.9556272537883046e-05, "loss": 3.225, "step": 3484 }, { "epoch": 0.12280099016006413, "grad_norm": 2.7076504230499268, "learning_rate": 1.9555936279924855e-05, "loss": 6.1959, "step": 3485 }, { "epoch": 0.12283622717302255, "grad_norm": 3.3136579990386963, "learning_rate": 1.9555599897499377e-05, "loss": 6.9387, "step": 3486 }, { "epoch": 0.12287146418598095, "grad_norm": 4.961054801940918, "learning_rate": 1.9555263390610993e-05, "loss": 11.9041, "step": 3487 }, { "epoch": 0.12290670119893937, "grad_norm": 2.4442286491394043, "learning_rate": 1.9554926759264087e-05, "loss": 6.534, "step": 3488 }, { "epoch": 0.12294193821189778, "grad_norm": 2.9253833293914795, "learning_rate": 1.9554590003463043e-05, "loss": 7.1286, "step": 3489 }, { "epoch": 0.12297717522485618, "grad_norm": 2.7218618392944336, "learning_rate": 1.9554253123212247e-05, "loss": 6.3157, "step": 3490 }, { "epoch": 0.1230124122378146, "grad_norm": 3.5785272121429443, "learning_rate": 1.9553916118516087e-05, "loss": 7.2166, "step": 3491 }, { "epoch": 0.12304764925077301, "grad_norm": 3.3187901973724365, "learning_rate": 1.9553578989378953e-05, "loss": 5.6333, "step": 3492 }, { "epoch": 0.12308288626373143, "grad_norm": 2.2147912979125977, "learning_rate": 1.9553241735805236e-05, "loss": 4.0949, "step": 3493 }, { "epoch": 0.12311812327668983, "grad_norm": 4.906425952911377, "learning_rate": 1.955290435779933e-05, "loss": 6.4373, "step": 3494 }, { "epoch": 0.12315336028964825, "grad_norm": 4.405986309051514, "learning_rate": 1.9552566855365624e-05, "loss": 6.3645, "step": 3495 }, { "epoch": 0.12318859730260666, "grad_norm": 4.041942596435547, "learning_rate": 1.9552229228508518e-05, "loss": 6.6316, "step": 3496 }, { "epoch": 0.12322383431556506, "grad_norm": 4.4193501472473145, "learning_rate": 1.9551891477232413e-05, "loss": 8.4616, "step": 3497 }, { "epoch": 0.12325907132852348, "grad_norm": 2.5892066955566406, "learning_rate": 1.95515536015417e-05, "loss": 3.8197, "step": 3498 }, { "epoch": 0.12329430834148189, "grad_norm": 2.2325737476348877, "learning_rate": 1.955121560144079e-05, "loss": 5.0585, "step": 3499 }, { "epoch": 0.12332954535444031, "grad_norm": 6.606112003326416, "learning_rate": 1.955087747693408e-05, "loss": 7.6512, "step": 3500 }, { "epoch": 0.12336478236739871, "grad_norm": 2.285146951675415, "learning_rate": 1.955053922802597e-05, "loss": 4.1297, "step": 3501 }, { "epoch": 0.12340001938035712, "grad_norm": 2.297922134399414, "learning_rate": 1.9550200854720874e-05, "loss": 4.1075, "step": 3502 }, { "epoch": 0.12343525639331554, "grad_norm": 2.706199884414673, "learning_rate": 1.954986235702319e-05, "loss": 3.5707, "step": 3503 }, { "epoch": 0.12347049340627395, "grad_norm": 7.4477219581604, "learning_rate": 1.9549523734937338e-05, "loss": 8.0285, "step": 3504 }, { "epoch": 0.12350573041923237, "grad_norm": 2.7460436820983887, "learning_rate": 1.9549184988467722e-05, "loss": 5.0138, "step": 3505 }, { "epoch": 0.12354096743219077, "grad_norm": 2.811051845550537, "learning_rate": 1.9548846117618755e-05, "loss": 4.1486, "step": 3506 }, { "epoch": 0.12357620444514919, "grad_norm": 3.6232035160064697, "learning_rate": 1.9548507122394848e-05, "loss": 6.0765, "step": 3507 }, { "epoch": 0.1236114414581076, "grad_norm": 4.39542293548584, "learning_rate": 1.954816800280042e-05, "loss": 6.0612, "step": 3508 }, { "epoch": 0.123646678471066, "grad_norm": 8.056259155273438, "learning_rate": 1.954782875883989e-05, "loss": 12.6002, "step": 3509 }, { "epoch": 0.12368191548402442, "grad_norm": 6.001546859741211, "learning_rate": 1.9547489390517675e-05, "loss": 5.3369, "step": 3510 }, { "epoch": 0.12371715249698283, "grad_norm": 2.2506237030029297, "learning_rate": 1.9547149897838192e-05, "loss": 5.0811, "step": 3511 }, { "epoch": 0.12375238950994125, "grad_norm": 1.5559076070785522, "learning_rate": 1.9546810280805867e-05, "loss": 5.453, "step": 3512 }, { "epoch": 0.12378762652289965, "grad_norm": 4.064017295837402, "learning_rate": 1.954647053942512e-05, "loss": 5.2394, "step": 3513 }, { "epoch": 0.12382286353585807, "grad_norm": 2.7218801975250244, "learning_rate": 1.9546130673700378e-05, "loss": 6.3729, "step": 3514 }, { "epoch": 0.12385810054881648, "grad_norm": 3.9662206172943115, "learning_rate": 1.954579068363607e-05, "loss": 8.4162, "step": 3515 }, { "epoch": 0.12389333756177488, "grad_norm": 2.3774585723876953, "learning_rate": 1.9545450569236625e-05, "loss": 3.646, "step": 3516 }, { "epoch": 0.1239285745747333, "grad_norm": 2.358626127243042, "learning_rate": 1.9545110330506463e-05, "loss": 3.6657, "step": 3517 }, { "epoch": 0.12396381158769171, "grad_norm": 4.0506672859191895, "learning_rate": 1.954476996745003e-05, "loss": 5.5395, "step": 3518 }, { "epoch": 0.12399904860065013, "grad_norm": 3.1218245029449463, "learning_rate": 1.954442948007175e-05, "loss": 5.5703, "step": 3519 }, { "epoch": 0.12403428561360853, "grad_norm": 3.053248167037964, "learning_rate": 1.954408886837606e-05, "loss": 3.4699, "step": 3520 }, { "epoch": 0.12406952262656695, "grad_norm": 4.670015811920166, "learning_rate": 1.9543748132367396e-05, "loss": 5.1685, "step": 3521 }, { "epoch": 0.12410475963952536, "grad_norm": 5.80225944519043, "learning_rate": 1.9543407272050197e-05, "loss": 8.2748, "step": 3522 }, { "epoch": 0.12413999665248376, "grad_norm": 3.1009442806243896, "learning_rate": 1.9543066287428904e-05, "loss": 3.884, "step": 3523 }, { "epoch": 0.12417523366544218, "grad_norm": 4.555724143981934, "learning_rate": 1.9542725178507955e-05, "loss": 5.9321, "step": 3524 }, { "epoch": 0.12421047067840059, "grad_norm": 2.4642317295074463, "learning_rate": 1.95423839452918e-05, "loss": 4.3367, "step": 3525 }, { "epoch": 0.12424570769135901, "grad_norm": 3.6460983753204346, "learning_rate": 1.954204258778487e-05, "loss": 5.7067, "step": 3526 }, { "epoch": 0.12428094470431741, "grad_norm": 5.996573448181152, "learning_rate": 1.9541701105991627e-05, "loss": 6.2718, "step": 3527 }, { "epoch": 0.12431618171727582, "grad_norm": 2.7852022647857666, "learning_rate": 1.9541359499916508e-05, "loss": 2.9154, "step": 3528 }, { "epoch": 0.12435141873023424, "grad_norm": 6.201767444610596, "learning_rate": 1.954101776956397e-05, "loss": 7.6977, "step": 3529 }, { "epoch": 0.12438665574319265, "grad_norm": 2.401268720626831, "learning_rate": 1.9540675914938456e-05, "loss": 3.0316, "step": 3530 }, { "epoch": 0.12442189275615106, "grad_norm": 4.032315254211426, "learning_rate": 1.9540333936044425e-05, "loss": 4.5729, "step": 3531 }, { "epoch": 0.12445712976910947, "grad_norm": 2.147700071334839, "learning_rate": 1.953999183288633e-05, "loss": 3.2791, "step": 3532 }, { "epoch": 0.12449236678206789, "grad_norm": 2.164027452468872, "learning_rate": 1.9539649605468623e-05, "loss": 3.9246, "step": 3533 }, { "epoch": 0.1245276037950263, "grad_norm": 2.1730682849884033, "learning_rate": 1.9539307253795767e-05, "loss": 3.9524, "step": 3534 }, { "epoch": 0.1245628408079847, "grad_norm": 2.776207208633423, "learning_rate": 1.9538964777872218e-05, "loss": 3.3827, "step": 3535 }, { "epoch": 0.12459807782094312, "grad_norm": 2.1433773040771484, "learning_rate": 1.953862217770244e-05, "loss": 3.6858, "step": 3536 }, { "epoch": 0.12463331483390153, "grad_norm": 3.0105957984924316, "learning_rate": 1.9538279453290893e-05, "loss": 3.1945, "step": 3537 }, { "epoch": 0.12466855184685995, "grad_norm": 5.0745320320129395, "learning_rate": 1.9537936604642038e-05, "loss": 6.7857, "step": 3538 }, { "epoch": 0.12470378885981835, "grad_norm": 5.279195308685303, "learning_rate": 1.9537593631760345e-05, "loss": 5.9211, "step": 3539 }, { "epoch": 0.12473902587277677, "grad_norm": 4.4077067375183105, "learning_rate": 1.953725053465028e-05, "loss": 6.1997, "step": 3540 }, { "epoch": 0.12477426288573518, "grad_norm": 2.7670257091522217, "learning_rate": 1.9536907313316313e-05, "loss": 3.9138, "step": 3541 }, { "epoch": 0.12480949989869358, "grad_norm": 8.205036163330078, "learning_rate": 1.9536563967762912e-05, "loss": 6.3818, "step": 3542 }, { "epoch": 0.124844736911652, "grad_norm": 3.16278338432312, "learning_rate": 1.9536220497994552e-05, "loss": 6.479, "step": 3543 }, { "epoch": 0.12487997392461041, "grad_norm": 2.288731336593628, "learning_rate": 1.9535876904015705e-05, "loss": 3.9625, "step": 3544 }, { "epoch": 0.12491521093756883, "grad_norm": 2.187107563018799, "learning_rate": 1.9535533185830846e-05, "loss": 3.5523, "step": 3545 }, { "epoch": 0.12495044795052723, "grad_norm": 4.8099284172058105, "learning_rate": 1.9535189343444452e-05, "loss": 7.93, "step": 3546 }, { "epoch": 0.12498568496348565, "grad_norm": 5.74242639541626, "learning_rate": 1.9534845376861006e-05, "loss": 10.4191, "step": 3547 }, { "epoch": 0.12502092197644404, "grad_norm": 3.003575086593628, "learning_rate": 1.9534501286084978e-05, "loss": 6.4347, "step": 3548 }, { "epoch": 0.12505615898940248, "grad_norm": 5.262754917144775, "learning_rate": 1.953415707112086e-05, "loss": 6.7662, "step": 3549 }, { "epoch": 0.12509139600236088, "grad_norm": 3.1346065998077393, "learning_rate": 1.9533812731973132e-05, "loss": 3.9197, "step": 3550 }, { "epoch": 0.1251266330153193, "grad_norm": 8.418415069580078, "learning_rate": 1.953346826864628e-05, "loss": 6.6501, "step": 3551 }, { "epoch": 0.1251618700282777, "grad_norm": 10.464726448059082, "learning_rate": 1.9533123681144788e-05, "loss": 8.0658, "step": 3552 }, { "epoch": 0.12519710704123613, "grad_norm": 8.362726211547852, "learning_rate": 1.9532778969473146e-05, "loss": 6.2813, "step": 3553 }, { "epoch": 0.12523234405419453, "grad_norm": 2.4361345767974854, "learning_rate": 1.9532434133635842e-05, "loss": 4.4983, "step": 3554 }, { "epoch": 0.12526758106715294, "grad_norm": 3.4729185104370117, "learning_rate": 1.953208917363737e-05, "loss": 4.2924, "step": 3555 }, { "epoch": 0.12530281808011134, "grad_norm": 3.958463430404663, "learning_rate": 1.9531744089482222e-05, "loss": 5.6722, "step": 3556 }, { "epoch": 0.12533805509306975, "grad_norm": 9.090879440307617, "learning_rate": 1.9531398881174896e-05, "loss": 5.611, "step": 3557 }, { "epoch": 0.12537329210602818, "grad_norm": 3.4480888843536377, "learning_rate": 1.953105354871988e-05, "loss": 6.5476, "step": 3558 }, { "epoch": 0.1254085291189866, "grad_norm": 3.1658432483673096, "learning_rate": 1.953070809212168e-05, "loss": 2.9814, "step": 3559 }, { "epoch": 0.125443766131945, "grad_norm": 2.231330394744873, "learning_rate": 1.9530362511384796e-05, "loss": 3.3773, "step": 3560 }, { "epoch": 0.1254790031449034, "grad_norm": 2.406888961791992, "learning_rate": 1.953001680651372e-05, "loss": 3.1492, "step": 3561 }, { "epoch": 0.1255142401578618, "grad_norm": 8.436776161193848, "learning_rate": 1.952967097751297e-05, "loss": 11.5179, "step": 3562 }, { "epoch": 0.12554947717082024, "grad_norm": 4.316087245941162, "learning_rate": 1.9529325024387035e-05, "loss": 5.5872, "step": 3563 }, { "epoch": 0.12558471418377865, "grad_norm": 2.6168134212493896, "learning_rate": 1.9528978947140433e-05, "loss": 3.019, "step": 3564 }, { "epoch": 0.12561995119673705, "grad_norm": 4.544806480407715, "learning_rate": 1.9528632745777662e-05, "loss": 6.2163, "step": 3565 }, { "epoch": 0.12565518820969546, "grad_norm": 6.211276531219482, "learning_rate": 1.9528286420303236e-05, "loss": 9.0517, "step": 3566 }, { "epoch": 0.12569042522265386, "grad_norm": 3.518850088119507, "learning_rate": 1.9527939970721665e-05, "loss": 4.0421, "step": 3567 }, { "epoch": 0.1257256622356123, "grad_norm": 2.4685018062591553, "learning_rate": 1.9527593397037465e-05, "loss": 2.9132, "step": 3568 }, { "epoch": 0.1257608992485707, "grad_norm": 3.2526869773864746, "learning_rate": 1.9527246699255145e-05, "loss": 6.2109, "step": 3569 }, { "epoch": 0.1257961362615291, "grad_norm": 5.890438556671143, "learning_rate": 1.9526899877379226e-05, "loss": 7.6002, "step": 3570 }, { "epoch": 0.1258313732744875, "grad_norm": 4.215062141418457, "learning_rate": 1.9526552931414216e-05, "loss": 6.7802, "step": 3571 }, { "epoch": 0.12586661028744595, "grad_norm": 2.760226011276245, "learning_rate": 1.952620586136465e-05, "loss": 2.9589, "step": 3572 }, { "epoch": 0.12590184730040435, "grad_norm": 3.837008237838745, "learning_rate": 1.9525858667235033e-05, "loss": 6.3241, "step": 3573 }, { "epoch": 0.12593708431336276, "grad_norm": 2.0625407695770264, "learning_rate": 1.9525511349029895e-05, "loss": 4.8784, "step": 3574 }, { "epoch": 0.12597232132632116, "grad_norm": 10.605695724487305, "learning_rate": 1.9525163906753757e-05, "loss": 8.3446, "step": 3575 }, { "epoch": 0.12600755833927957, "grad_norm": 3.3510749340057373, "learning_rate": 1.9524816340411144e-05, "loss": 4.91, "step": 3576 }, { "epoch": 0.126042795352238, "grad_norm": 6.91146993637085, "learning_rate": 1.952446865000659e-05, "loss": 6.2009, "step": 3577 }, { "epoch": 0.1260780323651964, "grad_norm": 2.8388872146606445, "learning_rate": 1.952412083554462e-05, "loss": 4.4201, "step": 3578 }, { "epoch": 0.1261132693781548, "grad_norm": 2.6559526920318604, "learning_rate": 1.952377289702976e-05, "loss": 3.7673, "step": 3579 }, { "epoch": 0.12614850639111322, "grad_norm": 5.82876443862915, "learning_rate": 1.9523424834466542e-05, "loss": 7.5261, "step": 3580 }, { "epoch": 0.12618374340407162, "grad_norm": 2.420469284057617, "learning_rate": 1.9523076647859503e-05, "loss": 4.3618, "step": 3581 }, { "epoch": 0.12621898041703006, "grad_norm": 3.399993896484375, "learning_rate": 1.9522728337213183e-05, "loss": 3.8664, "step": 3582 }, { "epoch": 0.12625421742998846, "grad_norm": 5.392434597015381, "learning_rate": 1.952237990253211e-05, "loss": 8.8572, "step": 3583 }, { "epoch": 0.12628945444294687, "grad_norm": 3.3007190227508545, "learning_rate": 1.9522031343820823e-05, "loss": 3.1783, "step": 3584 }, { "epoch": 0.12632469145590527, "grad_norm": 2.9807450771331787, "learning_rate": 1.9521682661083866e-05, "loss": 3.9804, "step": 3585 }, { "epoch": 0.1263599284688637, "grad_norm": 3.5739846229553223, "learning_rate": 1.9521333854325785e-05, "loss": 6.1725, "step": 3586 }, { "epoch": 0.1263951654818221, "grad_norm": 2.373549461364746, "learning_rate": 1.952098492355111e-05, "loss": 3.1759, "step": 3587 }, { "epoch": 0.12643040249478052, "grad_norm": 2.245457649230957, "learning_rate": 1.95206358687644e-05, "loss": 4.5745, "step": 3588 }, { "epoch": 0.12646563950773893, "grad_norm": 7.391998767852783, "learning_rate": 1.952028668997019e-05, "loss": 6.6952, "step": 3589 }, { "epoch": 0.12650087652069733, "grad_norm": 6.338364124298096, "learning_rate": 1.9519937387173035e-05, "loss": 6.2487, "step": 3590 }, { "epoch": 0.12653611353365576, "grad_norm": 2.5475807189941406, "learning_rate": 1.9519587960377485e-05, "loss": 4.1627, "step": 3591 }, { "epoch": 0.12657135054661417, "grad_norm": 2.0787856578826904, "learning_rate": 1.9519238409588086e-05, "loss": 4.2688, "step": 3592 }, { "epoch": 0.12660658755957258, "grad_norm": 2.271451711654663, "learning_rate": 1.9518888734809397e-05, "loss": 3.2045, "step": 3593 }, { "epoch": 0.12664182457253098, "grad_norm": 3.3401846885681152, "learning_rate": 1.9518538936045966e-05, "loss": 5.4069, "step": 3594 }, { "epoch": 0.1266770615854894, "grad_norm": 3.7811994552612305, "learning_rate": 1.9518189013302356e-05, "loss": 5.8019, "step": 3595 }, { "epoch": 0.12671229859844782, "grad_norm": 2.530721664428711, "learning_rate": 1.9517838966583122e-05, "loss": 2.5234, "step": 3596 }, { "epoch": 0.12674753561140623, "grad_norm": 2.9841806888580322, "learning_rate": 1.9517488795892823e-05, "loss": 3.9476, "step": 3597 }, { "epoch": 0.12678277262436463, "grad_norm": 5.475713729858398, "learning_rate": 1.9517138501236016e-05, "loss": 8.0568, "step": 3598 }, { "epoch": 0.12681800963732304, "grad_norm": 2.929536819458008, "learning_rate": 1.9516788082617273e-05, "loss": 4.2506, "step": 3599 }, { "epoch": 0.12685324665028144, "grad_norm": 3.7699804306030273, "learning_rate": 1.9516437540041153e-05, "loss": 5.4795, "step": 3600 }, { "epoch": 0.12688848366323988, "grad_norm": 5.163819313049316, "learning_rate": 1.9516086873512216e-05, "loss": 6.6582, "step": 3601 }, { "epoch": 0.12692372067619828, "grad_norm": 6.188910007476807, "learning_rate": 1.9515736083035038e-05, "loss": 6.2267, "step": 3602 }, { "epoch": 0.1269589576891567, "grad_norm": 2.507066011428833, "learning_rate": 1.951538516861419e-05, "loss": 6.1987, "step": 3603 }, { "epoch": 0.1269941947021151, "grad_norm": 6.853210926055908, "learning_rate": 1.9515034130254235e-05, "loss": 5.1383, "step": 3604 }, { "epoch": 0.12702943171507353, "grad_norm": 1.790837049484253, "learning_rate": 1.9514682967959747e-05, "loss": 4.451, "step": 3605 }, { "epoch": 0.12706466872803193, "grad_norm": 8.363879203796387, "learning_rate": 1.95143316817353e-05, "loss": 9.1649, "step": 3606 }, { "epoch": 0.12709990574099034, "grad_norm": 5.271599769592285, "learning_rate": 1.9513980271585477e-05, "loss": 5.7123, "step": 3607 }, { "epoch": 0.12713514275394874, "grad_norm": 6.392078876495361, "learning_rate": 1.9513628737514842e-05, "loss": 5.9214, "step": 3608 }, { "epoch": 0.12717037976690715, "grad_norm": 2.503976583480835, "learning_rate": 1.9513277079527987e-05, "loss": 3.0924, "step": 3609 }, { "epoch": 0.12720561677986558, "grad_norm": 2.5718002319335938, "learning_rate": 1.9512925297629484e-05, "loss": 7.2398, "step": 3610 }, { "epoch": 0.127240853792824, "grad_norm": 2.637030601501465, "learning_rate": 1.9512573391823916e-05, "loss": 3.8514, "step": 3611 }, { "epoch": 0.1272760908057824, "grad_norm": 3.3332571983337402, "learning_rate": 1.9512221362115865e-05, "loss": 3.3118, "step": 3612 }, { "epoch": 0.1273113278187408, "grad_norm": 3.485271453857422, "learning_rate": 1.9511869208509924e-05, "loss": 3.0644, "step": 3613 }, { "epoch": 0.1273465648316992, "grad_norm": 2.3642311096191406, "learning_rate": 1.9511516931010675e-05, "loss": 4.0701, "step": 3614 }, { "epoch": 0.12738180184465764, "grad_norm": 4.288758754730225, "learning_rate": 1.9511164529622704e-05, "loss": 6.419, "step": 3615 }, { "epoch": 0.12741703885761604, "grad_norm": 3.466604471206665, "learning_rate": 1.9510812004350604e-05, "loss": 3.2607, "step": 3616 }, { "epoch": 0.12745227587057445, "grad_norm": 4.015239238739014, "learning_rate": 1.9510459355198967e-05, "loss": 6.1179, "step": 3617 }, { "epoch": 0.12748751288353286, "grad_norm": 5.465211868286133, "learning_rate": 1.9510106582172387e-05, "loss": 9.0606, "step": 3618 }, { "epoch": 0.12752274989649126, "grad_norm": 8.241998672485352, "learning_rate": 1.9509753685275454e-05, "loss": 8.7267, "step": 3619 }, { "epoch": 0.1275579869094497, "grad_norm": 4.034603118896484, "learning_rate": 1.950940066451277e-05, "loss": 5.6157, "step": 3620 }, { "epoch": 0.1275932239224081, "grad_norm": 6.461251258850098, "learning_rate": 1.9509047519888932e-05, "loss": 5.5539, "step": 3621 }, { "epoch": 0.1276284609353665, "grad_norm": 9.207945823669434, "learning_rate": 1.9508694251408536e-05, "loss": 6.899, "step": 3622 }, { "epoch": 0.1276636979483249, "grad_norm": 3.200439453125, "learning_rate": 1.9508340859076187e-05, "loss": 2.8328, "step": 3623 }, { "epoch": 0.12769893496128334, "grad_norm": 5.4600510597229, "learning_rate": 1.9507987342896486e-05, "loss": 6.9407, "step": 3624 }, { "epoch": 0.12773417197424175, "grad_norm": 2.4354326725006104, "learning_rate": 1.950763370287404e-05, "loss": 4.915, "step": 3625 }, { "epoch": 0.12776940898720016, "grad_norm": 4.305500030517578, "learning_rate": 1.9507279939013456e-05, "loss": 5.2616, "step": 3626 }, { "epoch": 0.12780464600015856, "grad_norm": 7.35310697555542, "learning_rate": 1.9506926051319337e-05, "loss": 5.8301, "step": 3627 }, { "epoch": 0.12783988301311697, "grad_norm": 1.9211643934249878, "learning_rate": 1.9506572039796295e-05, "loss": 2.775, "step": 3628 }, { "epoch": 0.1278751200260754, "grad_norm": 4.9657416343688965, "learning_rate": 1.9506217904448946e-05, "loss": 6.2724, "step": 3629 }, { "epoch": 0.1279103570390338, "grad_norm": 2.5197417736053467, "learning_rate": 1.950586364528189e-05, "loss": 2.9402, "step": 3630 }, { "epoch": 0.1279455940519922, "grad_norm": 2.934877872467041, "learning_rate": 1.9505509262299757e-05, "loss": 6.0381, "step": 3631 }, { "epoch": 0.12798083106495062, "grad_norm": 6.043972015380859, "learning_rate": 1.950515475550715e-05, "loss": 6.6216, "step": 3632 }, { "epoch": 0.12801606807790902, "grad_norm": 2.3288309574127197, "learning_rate": 1.9504800124908693e-05, "loss": 3.322, "step": 3633 }, { "epoch": 0.12805130509086746, "grad_norm": 5.347126007080078, "learning_rate": 1.9504445370509005e-05, "loss": 6.2721, "step": 3634 }, { "epoch": 0.12808654210382586, "grad_norm": 3.0178213119506836, "learning_rate": 1.9504090492312704e-05, "loss": 3.6927, "step": 3635 }, { "epoch": 0.12812177911678427, "grad_norm": 5.655160427093506, "learning_rate": 1.9503735490324415e-05, "loss": 6.4727, "step": 3636 }, { "epoch": 0.12815701612974267, "grad_norm": 2.1191179752349854, "learning_rate": 1.9503380364548756e-05, "loss": 3.1796, "step": 3637 }, { "epoch": 0.1281922531427011, "grad_norm": 3.9390649795532227, "learning_rate": 1.950302511499036e-05, "loss": 5.989, "step": 3638 }, { "epoch": 0.1282274901556595, "grad_norm": 2.7870914936065674, "learning_rate": 1.9502669741653853e-05, "loss": 4.6019, "step": 3639 }, { "epoch": 0.12826272716861792, "grad_norm": 4.329942226409912, "learning_rate": 1.950231424454386e-05, "loss": 7.326, "step": 3640 }, { "epoch": 0.12829796418157632, "grad_norm": 3.632594108581543, "learning_rate": 1.9501958623665015e-05, "loss": 5.2503, "step": 3641 }, { "epoch": 0.12833320119453473, "grad_norm": 2.6087944507598877, "learning_rate": 1.950160287902195e-05, "loss": 3.5738, "step": 3642 }, { "epoch": 0.12836843820749316, "grad_norm": 4.017384052276611, "learning_rate": 1.9501247010619297e-05, "loss": 5.4477, "step": 3643 }, { "epoch": 0.12840367522045157, "grad_norm": 2.9106411933898926, "learning_rate": 1.9500891018461686e-05, "loss": 3.1407, "step": 3644 }, { "epoch": 0.12843891223340997, "grad_norm": 5.115055084228516, "learning_rate": 1.9500534902553765e-05, "loss": 7.6943, "step": 3645 }, { "epoch": 0.12847414924636838, "grad_norm": 2.1377341747283936, "learning_rate": 1.9500178662900168e-05, "loss": 3.422, "step": 3646 }, { "epoch": 0.12850938625932679, "grad_norm": 2.1410305500030518, "learning_rate": 1.949982229950553e-05, "loss": 3.9444, "step": 3647 }, { "epoch": 0.12854462327228522, "grad_norm": 3.3372557163238525, "learning_rate": 1.9499465812374496e-05, "loss": 5.4049, "step": 3648 }, { "epoch": 0.12857986028524362, "grad_norm": 2.5151357650756836, "learning_rate": 1.949910920151171e-05, "loss": 3.2472, "step": 3649 }, { "epoch": 0.12861509729820203, "grad_norm": 2.0098865032196045, "learning_rate": 1.949875246692182e-05, "loss": 5.4084, "step": 3650 }, { "epoch": 0.12865033431116044, "grad_norm": 2.359485149383545, "learning_rate": 1.9498395608609465e-05, "loss": 3.6438, "step": 3651 }, { "epoch": 0.12868557132411884, "grad_norm": 3.697753429412842, "learning_rate": 1.94980386265793e-05, "loss": 6.1643, "step": 3652 }, { "epoch": 0.12872080833707727, "grad_norm": 2.5529541969299316, "learning_rate": 1.9497681520835968e-05, "loss": 4.4315, "step": 3653 }, { "epoch": 0.12875604535003568, "grad_norm": 5.235386848449707, "learning_rate": 1.949732429138413e-05, "loss": 4.8328, "step": 3654 }, { "epoch": 0.1287912823629941, "grad_norm": 1.9737770557403564, "learning_rate": 1.9496966938228428e-05, "loss": 5.7522, "step": 3655 }, { "epoch": 0.1288265193759525, "grad_norm": 3.7330474853515625, "learning_rate": 1.9496609461373527e-05, "loss": 2.6279, "step": 3656 }, { "epoch": 0.12886175638891093, "grad_norm": 2.090609550476074, "learning_rate": 1.9496251860824074e-05, "loss": 3.3517, "step": 3657 }, { "epoch": 0.12889699340186933, "grad_norm": 2.821841239929199, "learning_rate": 1.9495894136584735e-05, "loss": 5.7481, "step": 3658 }, { "epoch": 0.12893223041482774, "grad_norm": 3.7495744228363037, "learning_rate": 1.9495536288660163e-05, "loss": 4.6796, "step": 3659 }, { "epoch": 0.12896746742778614, "grad_norm": 3.733358144760132, "learning_rate": 1.949517831705502e-05, "loss": 6.0713, "step": 3660 }, { "epoch": 0.12900270444074455, "grad_norm": 3.721564769744873, "learning_rate": 1.949482022177397e-05, "loss": 3.2291, "step": 3661 }, { "epoch": 0.12903794145370298, "grad_norm": 2.3580284118652344, "learning_rate": 1.949446200282168e-05, "loss": 2.5716, "step": 3662 }, { "epoch": 0.1290731784666614, "grad_norm": 2.7424488067626953, "learning_rate": 1.949410366020281e-05, "loss": 4.0566, "step": 3663 }, { "epoch": 0.1291084154796198, "grad_norm": 3.0372531414031982, "learning_rate": 1.9493745193922033e-05, "loss": 4.4655, "step": 3664 }, { "epoch": 0.1291436524925782, "grad_norm": 6.743203639984131, "learning_rate": 1.9493386603984016e-05, "loss": 10.402, "step": 3665 }, { "epoch": 0.1291788895055366, "grad_norm": 2.3248274326324463, "learning_rate": 1.9493027890393426e-05, "loss": 3.2482, "step": 3666 }, { "epoch": 0.12921412651849504, "grad_norm": 4.733700275421143, "learning_rate": 1.949266905315494e-05, "loss": 5.4318, "step": 3667 }, { "epoch": 0.12924936353145344, "grad_norm": 5.851317405700684, "learning_rate": 1.949231009227323e-05, "loss": 7.6053, "step": 3668 }, { "epoch": 0.12928460054441185, "grad_norm": 3.2342231273651123, "learning_rate": 1.9491951007752972e-05, "loss": 6.9324, "step": 3669 }, { "epoch": 0.12931983755737025, "grad_norm": 11.649394989013672, "learning_rate": 1.9491591799598845e-05, "loss": 7.6653, "step": 3670 }, { "epoch": 0.12935507457032866, "grad_norm": 3.887221336364746, "learning_rate": 1.9491232467815522e-05, "loss": 3.4683, "step": 3671 }, { "epoch": 0.1293903115832871, "grad_norm": 12.079549789428711, "learning_rate": 1.949087301240769e-05, "loss": 5.2324, "step": 3672 }, { "epoch": 0.1294255485962455, "grad_norm": 6.877966403961182, "learning_rate": 1.9490513433380028e-05, "loss": 5.6235, "step": 3673 }, { "epoch": 0.1294607856092039, "grad_norm": 3.2767231464385986, "learning_rate": 1.9490153730737218e-05, "loss": 5.3966, "step": 3674 }, { "epoch": 0.1294960226221623, "grad_norm": 2.1302361488342285, "learning_rate": 1.9489793904483948e-05, "loss": 4.33, "step": 3675 }, { "epoch": 0.12953125963512074, "grad_norm": 11.791611671447754, "learning_rate": 1.9489433954624902e-05, "loss": 8.6846, "step": 3676 }, { "epoch": 0.12956649664807915, "grad_norm": 2.9169254302978516, "learning_rate": 1.9489073881164772e-05, "loss": 4.7813, "step": 3677 }, { "epoch": 0.12960173366103755, "grad_norm": 5.52318000793457, "learning_rate": 1.9488713684108248e-05, "loss": 8.7087, "step": 3678 }, { "epoch": 0.12963697067399596, "grad_norm": 3.5338783264160156, "learning_rate": 1.9488353363460014e-05, "loss": 6.2374, "step": 3679 }, { "epoch": 0.12967220768695437, "grad_norm": 3.334146738052368, "learning_rate": 1.948799291922477e-05, "loss": 4.686, "step": 3680 }, { "epoch": 0.1297074446999128, "grad_norm": 4.772420883178711, "learning_rate": 1.9487632351407214e-05, "loss": 7.9242, "step": 3681 }, { "epoch": 0.1297426817128712, "grad_norm": 2.1333391666412354, "learning_rate": 1.9487271660012036e-05, "loss": 2.7428, "step": 3682 }, { "epoch": 0.1297779187258296, "grad_norm": 2.4463467597961426, "learning_rate": 1.9486910845043936e-05, "loss": 2.9873, "step": 3683 }, { "epoch": 0.12981315573878802, "grad_norm": 6.332281112670898, "learning_rate": 1.9486549906507614e-05, "loss": 6.989, "step": 3684 }, { "epoch": 0.12984839275174642, "grad_norm": 2.999624252319336, "learning_rate": 1.948618884440777e-05, "loss": 5.8264, "step": 3685 }, { "epoch": 0.12988362976470486, "grad_norm": 2.3605782985687256, "learning_rate": 1.9485827658749107e-05, "loss": 3.3295, "step": 3686 }, { "epoch": 0.12991886677766326, "grad_norm": 2.4880223274230957, "learning_rate": 1.9485466349536333e-05, "loss": 5.6057, "step": 3687 }, { "epoch": 0.12995410379062167, "grad_norm": 2.6867613792419434, "learning_rate": 1.9485104916774152e-05, "loss": 2.7393, "step": 3688 }, { "epoch": 0.12998934080358007, "grad_norm": 5.442580223083496, "learning_rate": 1.948474336046727e-05, "loss": 6.2665, "step": 3689 }, { "epoch": 0.1300245778165385, "grad_norm": 4.432894229888916, "learning_rate": 1.9484381680620393e-05, "loss": 6.3975, "step": 3690 }, { "epoch": 0.1300598148294969, "grad_norm": 5.281391143798828, "learning_rate": 1.9484019877238244e-05, "loss": 5.6882, "step": 3691 }, { "epoch": 0.13009505184245532, "grad_norm": 7.299768924713135, "learning_rate": 1.9483657950325522e-05, "loss": 4.6556, "step": 3692 }, { "epoch": 0.13013028885541372, "grad_norm": 5.237464904785156, "learning_rate": 1.9483295899886953e-05, "loss": 7.4194, "step": 3693 }, { "epoch": 0.13016552586837213, "grad_norm": 6.208938121795654, "learning_rate": 1.9482933725927246e-05, "loss": 5.2082, "step": 3694 }, { "epoch": 0.13020076288133056, "grad_norm": 2.2636806964874268, "learning_rate": 1.9482571428451114e-05, "loss": 3.2594, "step": 3695 }, { "epoch": 0.13023599989428897, "grad_norm": 2.432258129119873, "learning_rate": 1.9482209007463284e-05, "loss": 3.6926, "step": 3696 }, { "epoch": 0.13027123690724737, "grad_norm": 3.1510798931121826, "learning_rate": 1.9481846462968476e-05, "loss": 5.6013, "step": 3697 }, { "epoch": 0.13030647392020578, "grad_norm": 2.70611834526062, "learning_rate": 1.948148379497141e-05, "loss": 2.9435, "step": 3698 }, { "epoch": 0.13034171093316418, "grad_norm": 3.976790189743042, "learning_rate": 1.9481121003476806e-05, "loss": 5.3283, "step": 3699 }, { "epoch": 0.13037694794612262, "grad_norm": 6.3628411293029785, "learning_rate": 1.9480758088489396e-05, "loss": 5.4077, "step": 3700 }, { "epoch": 0.13041218495908102, "grad_norm": 6.018941402435303, "learning_rate": 1.9480395050013904e-05, "loss": 6.0899, "step": 3701 }, { "epoch": 0.13044742197203943, "grad_norm": 3.782416343688965, "learning_rate": 1.948003188805506e-05, "loss": 6.0332, "step": 3702 }, { "epoch": 0.13048265898499783, "grad_norm": 2.4901814460754395, "learning_rate": 1.947966860261759e-05, "loss": 3.8458, "step": 3703 }, { "epoch": 0.13051789599795624, "grad_norm": 2.539902925491333, "learning_rate": 1.9479305193706232e-05, "loss": 6.1207, "step": 3704 }, { "epoch": 0.13055313301091467, "grad_norm": 3.948230028152466, "learning_rate": 1.9478941661325713e-05, "loss": 5.366, "step": 3705 }, { "epoch": 0.13058837002387308, "grad_norm": 2.147451162338257, "learning_rate": 1.9478578005480774e-05, "loss": 3.9, "step": 3706 }, { "epoch": 0.13062360703683148, "grad_norm": 3.455777406692505, "learning_rate": 1.9478214226176148e-05, "loss": 3.3962, "step": 3707 }, { "epoch": 0.1306588440497899, "grad_norm": 4.086887836456299, "learning_rate": 1.9477850323416575e-05, "loss": 8.0249, "step": 3708 }, { "epoch": 0.13069408106274832, "grad_norm": 4.8435869216918945, "learning_rate": 1.9477486297206795e-05, "loss": 7.7188, "step": 3709 }, { "epoch": 0.13072931807570673, "grad_norm": 2.175868511199951, "learning_rate": 1.9477122147551547e-05, "loss": 2.8777, "step": 3710 }, { "epoch": 0.13076455508866514, "grad_norm": 5.036766529083252, "learning_rate": 1.9476757874455577e-05, "loss": 9.37, "step": 3711 }, { "epoch": 0.13079979210162354, "grad_norm": 2.2908072471618652, "learning_rate": 1.947639347792363e-05, "loss": 3.3206, "step": 3712 }, { "epoch": 0.13083502911458195, "grad_norm": 3.481433868408203, "learning_rate": 1.9476028957960444e-05, "loss": 5.8739, "step": 3713 }, { "epoch": 0.13087026612754038, "grad_norm": 2.6655704975128174, "learning_rate": 1.947566431457078e-05, "loss": 2.769, "step": 3714 }, { "epoch": 0.13090550314049879, "grad_norm": 3.0862414836883545, "learning_rate": 1.9475299547759382e-05, "loss": 5.4161, "step": 3715 }, { "epoch": 0.1309407401534572, "grad_norm": 2.6650779247283936, "learning_rate": 1.9474934657530998e-05, "loss": 3.0114, "step": 3716 }, { "epoch": 0.1309759771664156, "grad_norm": 2.420989751815796, "learning_rate": 1.9474569643890384e-05, "loss": 4.1541, "step": 3717 }, { "epoch": 0.131011214179374, "grad_norm": 4.305671691894531, "learning_rate": 1.9474204506842295e-05, "loss": 7.2217, "step": 3718 }, { "epoch": 0.13104645119233244, "grad_norm": 8.33467960357666, "learning_rate": 1.947383924639148e-05, "loss": 5.4716, "step": 3719 }, { "epoch": 0.13108168820529084, "grad_norm": 2.5060343742370605, "learning_rate": 1.9473473862542705e-05, "loss": 3.6804, "step": 3720 }, { "epoch": 0.13111692521824925, "grad_norm": 5.249693870544434, "learning_rate": 1.9473108355300726e-05, "loss": 5.4575, "step": 3721 }, { "epoch": 0.13115216223120765, "grad_norm": 2.774934768676758, "learning_rate": 1.9472742724670306e-05, "loss": 2.8243, "step": 3722 }, { "epoch": 0.13118739924416606, "grad_norm": 2.4500181674957275, "learning_rate": 1.9472376970656208e-05, "loss": 3.4263, "step": 3723 }, { "epoch": 0.1312226362571245, "grad_norm": 2.4062817096710205, "learning_rate": 1.947201109326319e-05, "loss": 3.4846, "step": 3724 }, { "epoch": 0.1312578732700829, "grad_norm": 3.1023120880126953, "learning_rate": 1.9471645092496016e-05, "loss": 3.373, "step": 3725 }, { "epoch": 0.1312931102830413, "grad_norm": 5.554392337799072, "learning_rate": 1.9471278968359467e-05, "loss": 6.119, "step": 3726 }, { "epoch": 0.1313283472959997, "grad_norm": 2.8507728576660156, "learning_rate": 1.9470912720858296e-05, "loss": 3.7105, "step": 3727 }, { "epoch": 0.13136358430895814, "grad_norm": 2.488538980484009, "learning_rate": 1.9470546349997285e-05, "loss": 3.6764, "step": 3728 }, { "epoch": 0.13139882132191655, "grad_norm": 2.6113409996032715, "learning_rate": 1.94701798557812e-05, "loss": 2.8326, "step": 3729 }, { "epoch": 0.13143405833487495, "grad_norm": 3.3471972942352295, "learning_rate": 1.9469813238214813e-05, "loss": 2.9406, "step": 3730 }, { "epoch": 0.13146929534783336, "grad_norm": 4.702914237976074, "learning_rate": 1.9469446497302906e-05, "loss": 7.7298, "step": 3731 }, { "epoch": 0.13150453236079176, "grad_norm": 4.8440423011779785, "learning_rate": 1.9469079633050252e-05, "loss": 7.9926, "step": 3732 }, { "epoch": 0.1315397693737502, "grad_norm": 1.9543383121490479, "learning_rate": 1.946871264546163e-05, "loss": 4.6044, "step": 3733 }, { "epoch": 0.1315750063867086, "grad_norm": 2.144355535507202, "learning_rate": 1.9468345534541816e-05, "loss": 3.8119, "step": 3734 }, { "epoch": 0.131610243399667, "grad_norm": 4.510964393615723, "learning_rate": 1.94679783002956e-05, "loss": 6.0811, "step": 3735 }, { "epoch": 0.13164548041262542, "grad_norm": 5.954268455505371, "learning_rate": 1.946761094272776e-05, "loss": 8.6553, "step": 3736 }, { "epoch": 0.13168071742558382, "grad_norm": 4.175889492034912, "learning_rate": 1.9467243461843078e-05, "loss": 5.3209, "step": 3737 }, { "epoch": 0.13171595443854225, "grad_norm": 4.038393020629883, "learning_rate": 1.9466875857646344e-05, "loss": 5.9371, "step": 3738 }, { "epoch": 0.13175119145150066, "grad_norm": 2.8880093097686768, "learning_rate": 1.9466508130142346e-05, "loss": 3.0048, "step": 3739 }, { "epoch": 0.13178642846445907, "grad_norm": 2.149552583694458, "learning_rate": 1.946614027933588e-05, "loss": 3.3532, "step": 3740 }, { "epoch": 0.13182166547741747, "grad_norm": 3.3770484924316406, "learning_rate": 1.9465772305231723e-05, "loss": 5.9957, "step": 3741 }, { "epoch": 0.1318569024903759, "grad_norm": 4.19654655456543, "learning_rate": 1.946540420783468e-05, "loss": 6.0368, "step": 3742 }, { "epoch": 0.1318921395033343, "grad_norm": 3.4353370666503906, "learning_rate": 1.9465035987149544e-05, "loss": 5.7314, "step": 3743 }, { "epoch": 0.13192737651629272, "grad_norm": 4.8730854988098145, "learning_rate": 1.9464667643181106e-05, "loss": 8.5569, "step": 3744 }, { "epoch": 0.13196261352925112, "grad_norm": 2.237788200378418, "learning_rate": 1.9464299175934164e-05, "loss": 3.6521, "step": 3745 }, { "epoch": 0.13199785054220953, "grad_norm": 6.775052070617676, "learning_rate": 1.9463930585413524e-05, "loss": 5.1561, "step": 3746 }, { "epoch": 0.13203308755516796, "grad_norm": 5.594513416290283, "learning_rate": 1.946356187162398e-05, "loss": 7.2288, "step": 3747 }, { "epoch": 0.13206832456812637, "grad_norm": 4.088629722595215, "learning_rate": 1.9463193034570336e-05, "loss": 5.1231, "step": 3748 }, { "epoch": 0.13210356158108477, "grad_norm": 2.8129663467407227, "learning_rate": 1.94628240742574e-05, "loss": 3.5572, "step": 3749 }, { "epoch": 0.13213879859404318, "grad_norm": 3.4309940338134766, "learning_rate": 1.9462454990689975e-05, "loss": 2.4234, "step": 3750 }, { "epoch": 0.13217403560700158, "grad_norm": 6.852811336517334, "learning_rate": 1.9462085783872865e-05, "loss": 7.132, "step": 3751 }, { "epoch": 0.13220927261996002, "grad_norm": 6.3814778327941895, "learning_rate": 1.9461716453810886e-05, "loss": 8.4429, "step": 3752 }, { "epoch": 0.13224450963291842, "grad_norm": 5.386456489562988, "learning_rate": 1.9461347000508845e-05, "loss": 6.3824, "step": 3753 }, { "epoch": 0.13227974664587683, "grad_norm": 3.907841682434082, "learning_rate": 1.9460977423971553e-05, "loss": 5.4269, "step": 3754 }, { "epoch": 0.13231498365883523, "grad_norm": 2.4514734745025635, "learning_rate": 1.9460607724203823e-05, "loss": 2.8495, "step": 3755 }, { "epoch": 0.13235022067179364, "grad_norm": 6.510928630828857, "learning_rate": 1.9460237901210474e-05, "loss": 4.9564, "step": 3756 }, { "epoch": 0.13238545768475207, "grad_norm": 2.367307186126709, "learning_rate": 1.9459867954996322e-05, "loss": 5.0434, "step": 3757 }, { "epoch": 0.13242069469771048, "grad_norm": 8.49144172668457, "learning_rate": 1.9459497885566184e-05, "loss": 8.27, "step": 3758 }, { "epoch": 0.13245593171066888, "grad_norm": 3.290327310562134, "learning_rate": 1.9459127692924883e-05, "loss": 3.7687, "step": 3759 }, { "epoch": 0.1324911687236273, "grad_norm": 3.5419728755950928, "learning_rate": 1.9458757377077235e-05, "loss": 6.0645, "step": 3760 }, { "epoch": 0.13252640573658572, "grad_norm": 2.507413148880005, "learning_rate": 1.9458386938028072e-05, "loss": 4.159, "step": 3761 }, { "epoch": 0.13256164274954413, "grad_norm": 2.365159511566162, "learning_rate": 1.9458016375782213e-05, "loss": 3.9625, "step": 3762 }, { "epoch": 0.13259687976250253, "grad_norm": 4.660172462463379, "learning_rate": 1.945764569034448e-05, "loss": 4.3475, "step": 3763 }, { "epoch": 0.13263211677546094, "grad_norm": 3.7131874561309814, "learning_rate": 1.9457274881719715e-05, "loss": 4.0239, "step": 3764 }, { "epoch": 0.13266735378841935, "grad_norm": 4.992259502410889, "learning_rate": 1.9456903949912737e-05, "loss": 3.3688, "step": 3765 }, { "epoch": 0.13270259080137778, "grad_norm": 3.1495864391326904, "learning_rate": 1.9456532894928378e-05, "loss": 3.3964, "step": 3766 }, { "epoch": 0.13273782781433618, "grad_norm": 7.305615425109863, "learning_rate": 1.945616171677148e-05, "loss": 7.3035, "step": 3767 }, { "epoch": 0.1327730648272946, "grad_norm": 4.274827480316162, "learning_rate": 1.9455790415446866e-05, "loss": 5.3673, "step": 3768 }, { "epoch": 0.132808301840253, "grad_norm": 5.573026657104492, "learning_rate": 1.945541899095938e-05, "loss": 5.6295, "step": 3769 }, { "epoch": 0.1328435388532114, "grad_norm": 5.798035621643066, "learning_rate": 1.9455047443313854e-05, "loss": 6.1178, "step": 3770 }, { "epoch": 0.13287877586616983, "grad_norm": 4.320161819458008, "learning_rate": 1.9454675772515132e-05, "loss": 6.2013, "step": 3771 }, { "epoch": 0.13291401287912824, "grad_norm": 4.946331977844238, "learning_rate": 1.9454303978568056e-05, "loss": 6.1356, "step": 3772 }, { "epoch": 0.13294924989208665, "grad_norm": 5.410982608795166, "learning_rate": 1.9453932061477463e-05, "loss": 5.3602, "step": 3773 }, { "epoch": 0.13298448690504505, "grad_norm": 5.684078693389893, "learning_rate": 1.9453560021248204e-05, "loss": 7.1181, "step": 3774 }, { "epoch": 0.13301972391800346, "grad_norm": 3.812028408050537, "learning_rate": 1.9453187857885116e-05, "loss": 4.8324, "step": 3775 }, { "epoch": 0.1330549609309619, "grad_norm": 3.415329933166504, "learning_rate": 1.9452815571393056e-05, "loss": 2.9312, "step": 3776 }, { "epoch": 0.1330901979439203, "grad_norm": 5.711522102355957, "learning_rate": 1.945244316177687e-05, "loss": 9.4814, "step": 3777 }, { "epoch": 0.1331254349568787, "grad_norm": 4.174032211303711, "learning_rate": 1.9452070629041406e-05, "loss": 7.742, "step": 3778 }, { "epoch": 0.1331606719698371, "grad_norm": 4.066784381866455, "learning_rate": 1.945169797319152e-05, "loss": 5.3949, "step": 3779 }, { "epoch": 0.13319590898279554, "grad_norm": 2.705700397491455, "learning_rate": 1.945132519423206e-05, "loss": 2.5456, "step": 3780 }, { "epoch": 0.13323114599575395, "grad_norm": 4.572198867797852, "learning_rate": 1.945095229216789e-05, "loss": 5.3182, "step": 3781 }, { "epoch": 0.13326638300871235, "grad_norm": 4.8209052085876465, "learning_rate": 1.945057926700386e-05, "loss": 7.5525, "step": 3782 }, { "epoch": 0.13330162002167076, "grad_norm": 4.874972820281982, "learning_rate": 1.9450206118744833e-05, "loss": 5.8921, "step": 3783 }, { "epoch": 0.13333685703462916, "grad_norm": 2.095615863800049, "learning_rate": 1.9449832847395666e-05, "loss": 2.8338, "step": 3784 }, { "epoch": 0.1333720940475876, "grad_norm": 2.4802162647247314, "learning_rate": 1.944945945296122e-05, "loss": 3.7567, "step": 3785 }, { "epoch": 0.133407331060546, "grad_norm": 8.180517196655273, "learning_rate": 1.9449085935446367e-05, "loss": 9.0442, "step": 3786 }, { "epoch": 0.1334425680735044, "grad_norm": 5.313084602355957, "learning_rate": 1.944871229485596e-05, "loss": 6.1413, "step": 3787 }, { "epoch": 0.13347780508646281, "grad_norm": 7.478957653045654, "learning_rate": 1.944833853119488e-05, "loss": 5.7476, "step": 3788 }, { "epoch": 0.13351304209942122, "grad_norm": 6.2573699951171875, "learning_rate": 1.9447964644467978e-05, "loss": 10.2906, "step": 3789 }, { "epoch": 0.13354827911237965, "grad_norm": 5.813272953033447, "learning_rate": 1.9447590634680137e-05, "loss": 5.6729, "step": 3790 }, { "epoch": 0.13358351612533806, "grad_norm": 2.5358588695526123, "learning_rate": 1.9447216501836223e-05, "loss": 2.1091, "step": 3791 }, { "epoch": 0.13361875313829646, "grad_norm": 2.688232421875, "learning_rate": 1.9446842245941112e-05, "loss": 3.4332, "step": 3792 }, { "epoch": 0.13365399015125487, "grad_norm": 2.397495985031128, "learning_rate": 1.9446467866999676e-05, "loss": 3.632, "step": 3793 }, { "epoch": 0.1336892271642133, "grad_norm": 5.890760898590088, "learning_rate": 1.9446093365016795e-05, "loss": 5.4818, "step": 3794 }, { "epoch": 0.1337244641771717, "grad_norm": 5.390255928039551, "learning_rate": 1.9445718739997343e-05, "loss": 5.1601, "step": 3795 }, { "epoch": 0.13375970119013011, "grad_norm": 4.089723587036133, "learning_rate": 1.9445343991946203e-05, "loss": 5.2491, "step": 3796 }, { "epoch": 0.13379493820308852, "grad_norm": 2.3697755336761475, "learning_rate": 1.9444969120868254e-05, "loss": 3.5113, "step": 3797 }, { "epoch": 0.13383017521604693, "grad_norm": 4.099724769592285, "learning_rate": 1.9444594126768378e-05, "loss": 5.8805, "step": 3798 }, { "epoch": 0.13386541222900536, "grad_norm": 3.2132625579833984, "learning_rate": 1.944421900965146e-05, "loss": 2.8687, "step": 3799 }, { "epoch": 0.13390064924196377, "grad_norm": 3.1275572776794434, "learning_rate": 1.9443843769522388e-05, "loss": 5.4564, "step": 3800 }, { "epoch": 0.13393588625492217, "grad_norm": 4.970715045928955, "learning_rate": 1.9443468406386046e-05, "loss": 5.2107, "step": 3801 }, { "epoch": 0.13397112326788058, "grad_norm": 3.411074161529541, "learning_rate": 1.9443092920247326e-05, "loss": 5.4894, "step": 3802 }, { "epoch": 0.13400636028083898, "grad_norm": 3.7589919567108154, "learning_rate": 1.9442717311111118e-05, "loss": 4.166, "step": 3803 }, { "epoch": 0.13404159729379742, "grad_norm": 5.780364990234375, "learning_rate": 1.9442341578982314e-05, "loss": 7.5083, "step": 3804 }, { "epoch": 0.13407683430675582, "grad_norm": 2.7324512004852295, "learning_rate": 1.944196572386581e-05, "loss": 2.6679, "step": 3805 }, { "epoch": 0.13411207131971423, "grad_norm": 3.2292914390563965, "learning_rate": 1.94415897457665e-05, "loss": 3.7252, "step": 3806 }, { "epoch": 0.13414730833267263, "grad_norm": 8.474483489990234, "learning_rate": 1.9441213644689277e-05, "loss": 7.7147, "step": 3807 }, { "epoch": 0.13418254534563104, "grad_norm": 8.108183860778809, "learning_rate": 1.944083742063905e-05, "loss": 6.8436, "step": 3808 }, { "epoch": 0.13421778235858947, "grad_norm": 8.396573066711426, "learning_rate": 1.9440461073620707e-05, "loss": 6.9276, "step": 3809 }, { "epoch": 0.13425301937154788, "grad_norm": 9.595176696777344, "learning_rate": 1.9440084603639158e-05, "loss": 7.4996, "step": 3810 }, { "epoch": 0.13428825638450628, "grad_norm": 3.987612247467041, "learning_rate": 1.94397080106993e-05, "loss": 2.8876, "step": 3811 }, { "epoch": 0.1343234933974647, "grad_norm": 3.3195242881774902, "learning_rate": 1.943933129480605e-05, "loss": 2.8755, "step": 3812 }, { "epoch": 0.13435873041042312, "grad_norm": 2.511129856109619, "learning_rate": 1.9438954455964306e-05, "loss": 2.6853, "step": 3813 }, { "epoch": 0.13439396742338153, "grad_norm": 2.709531545639038, "learning_rate": 1.9438577494178975e-05, "loss": 3.9873, "step": 3814 }, { "epoch": 0.13442920443633993, "grad_norm": 4.688258171081543, "learning_rate": 1.943820040945497e-05, "loss": 5.3221, "step": 3815 }, { "epoch": 0.13446444144929834, "grad_norm": 5.709949970245361, "learning_rate": 1.9437823201797202e-05, "loss": 5.8888, "step": 3816 }, { "epoch": 0.13449967846225674, "grad_norm": 8.426027297973633, "learning_rate": 1.943744587121059e-05, "loss": 7.3603, "step": 3817 }, { "epoch": 0.13453491547521518, "grad_norm": 4.226010799407959, "learning_rate": 1.9437068417700037e-05, "loss": 5.3977, "step": 3818 }, { "epoch": 0.13457015248817358, "grad_norm": 3.7461791038513184, "learning_rate": 1.9436690841270473e-05, "loss": 3.3236, "step": 3819 }, { "epoch": 0.134605389501132, "grad_norm": 4.146112442016602, "learning_rate": 1.9436313141926802e-05, "loss": 5.637, "step": 3820 }, { "epoch": 0.1346406265140904, "grad_norm": 5.200303077697754, "learning_rate": 1.943593531967395e-05, "loss": 3.2252, "step": 3821 }, { "epoch": 0.1346758635270488, "grad_norm": 3.3895792961120605, "learning_rate": 1.9435557374516844e-05, "loss": 4.1467, "step": 3822 }, { "epoch": 0.13471110054000723, "grad_norm": 3.4044790267944336, "learning_rate": 1.94351793064604e-05, "loss": 5.3328, "step": 3823 }, { "epoch": 0.13474633755296564, "grad_norm": 6.09254789352417, "learning_rate": 1.943480111550954e-05, "loss": 5.3312, "step": 3824 }, { "epoch": 0.13478157456592404, "grad_norm": 2.831540107727051, "learning_rate": 1.9434422801669198e-05, "loss": 3.0567, "step": 3825 }, { "epoch": 0.13481681157888245, "grad_norm": 4.176555156707764, "learning_rate": 1.9434044364944292e-05, "loss": 4.6442, "step": 3826 }, { "epoch": 0.13485204859184086, "grad_norm": 5.531644344329834, "learning_rate": 1.943366580533976e-05, "loss": 4.3427, "step": 3827 }, { "epoch": 0.1348872856047993, "grad_norm": 8.142505645751953, "learning_rate": 1.943328712286053e-05, "loss": 5.6575, "step": 3828 }, { "epoch": 0.1349225226177577, "grad_norm": 4.172738552093506, "learning_rate": 1.943290831751153e-05, "loss": 5.6677, "step": 3829 }, { "epoch": 0.1349577596307161, "grad_norm": 3.57548451423645, "learning_rate": 1.9432529389297705e-05, "loss": 3.3101, "step": 3830 }, { "epoch": 0.1349929966436745, "grad_norm": 5.142658233642578, "learning_rate": 1.9432150338223973e-05, "loss": 5.6869, "step": 3831 }, { "epoch": 0.13502823365663294, "grad_norm": 9.63972282409668, "learning_rate": 1.9431771164295288e-05, "loss": 4.8763, "step": 3832 }, { "epoch": 0.13506347066959135, "grad_norm": 4.008281707763672, "learning_rate": 1.9431391867516583e-05, "loss": 3.1187, "step": 3833 }, { "epoch": 0.13509870768254975, "grad_norm": 8.088046073913574, "learning_rate": 1.943101244789279e-05, "loss": 5.4452, "step": 3834 }, { "epoch": 0.13513394469550816, "grad_norm": 2.561676502227783, "learning_rate": 1.9430632905428867e-05, "loss": 2.4132, "step": 3835 }, { "epoch": 0.13516918170846656, "grad_norm": 2.75110125541687, "learning_rate": 1.9430253240129745e-05, "loss": 2.7866, "step": 3836 }, { "epoch": 0.135204418721425, "grad_norm": 6.497035980224609, "learning_rate": 1.9429873452000377e-05, "loss": 5.5128, "step": 3837 }, { "epoch": 0.1352396557343834, "grad_norm": 2.2071683406829834, "learning_rate": 1.9429493541045702e-05, "loss": 3.9475, "step": 3838 }, { "epoch": 0.1352748927473418, "grad_norm": 5.211167812347412, "learning_rate": 1.9429113507270674e-05, "loss": 5.8427, "step": 3839 }, { "epoch": 0.1353101297603002, "grad_norm": 2.190990924835205, "learning_rate": 1.9428733350680242e-05, "loss": 2.9222, "step": 3840 }, { "epoch": 0.13534536677325862, "grad_norm": 2.4947781562805176, "learning_rate": 1.9428353071279354e-05, "loss": 3.2145, "step": 3841 }, { "epoch": 0.13538060378621705, "grad_norm": 2.4652256965637207, "learning_rate": 1.942797266907297e-05, "loss": 3.2283, "step": 3842 }, { "epoch": 0.13541584079917546, "grad_norm": 4.956733226776123, "learning_rate": 1.942759214406604e-05, "loss": 5.6457, "step": 3843 }, { "epoch": 0.13545107781213386, "grad_norm": 4.396452903747559, "learning_rate": 1.942721149626352e-05, "loss": 5.0862, "step": 3844 }, { "epoch": 0.13548631482509227, "grad_norm": 3.587498188018799, "learning_rate": 1.9426830725670372e-05, "loss": 2.0161, "step": 3845 }, { "epoch": 0.1355215518380507, "grad_norm": 2.335071563720703, "learning_rate": 1.9426449832291553e-05, "loss": 4.92, "step": 3846 }, { "epoch": 0.1355567888510091, "grad_norm": 2.5688745975494385, "learning_rate": 1.9426068816132023e-05, "loss": 4.4209, "step": 3847 }, { "epoch": 0.1355920258639675, "grad_norm": 3.075705051422119, "learning_rate": 1.9425687677196743e-05, "loss": 2.47, "step": 3848 }, { "epoch": 0.13562726287692592, "grad_norm": 4.122830390930176, "learning_rate": 1.9425306415490682e-05, "loss": 4.5005, "step": 3849 }, { "epoch": 0.13566249988988432, "grad_norm": 3.8124656677246094, "learning_rate": 1.942492503101881e-05, "loss": 5.1563, "step": 3850 }, { "epoch": 0.13569773690284276, "grad_norm": 5.917694091796875, "learning_rate": 1.9424543523786084e-05, "loss": 5.1191, "step": 3851 }, { "epoch": 0.13573297391580116, "grad_norm": 2.6702682971954346, "learning_rate": 1.9424161893797476e-05, "loss": 3.4666, "step": 3852 }, { "epoch": 0.13576821092875957, "grad_norm": 7.139008522033691, "learning_rate": 1.9423780141057963e-05, "loss": 8.7618, "step": 3853 }, { "epoch": 0.13580344794171798, "grad_norm": 2.5930287837982178, "learning_rate": 1.9423398265572512e-05, "loss": 2.933, "step": 3854 }, { "epoch": 0.13583868495467638, "grad_norm": 2.060844898223877, "learning_rate": 1.9423016267346096e-05, "loss": 2.4034, "step": 3855 }, { "epoch": 0.13587392196763481, "grad_norm": 3.18286395072937, "learning_rate": 1.9422634146383698e-05, "loss": 5.8584, "step": 3856 }, { "epoch": 0.13590915898059322, "grad_norm": 5.693269729614258, "learning_rate": 1.9422251902690284e-05, "loss": 5.843, "step": 3857 }, { "epoch": 0.13594439599355163, "grad_norm": 6.241941452026367, "learning_rate": 1.942186953627084e-05, "loss": 5.8505, "step": 3858 }, { "epoch": 0.13597963300651003, "grad_norm": 4.316615104675293, "learning_rate": 1.9421487047130348e-05, "loss": 5.6408, "step": 3859 }, { "epoch": 0.13601487001946844, "grad_norm": 5.035491466522217, "learning_rate": 1.9421104435273786e-05, "loss": 8.4269, "step": 3860 }, { "epoch": 0.13605010703242687, "grad_norm": 2.4964523315429688, "learning_rate": 1.9420721700706138e-05, "loss": 2.7565, "step": 3861 }, { "epoch": 0.13608534404538528, "grad_norm": 3.398298501968384, "learning_rate": 1.942033884343239e-05, "loss": 4.8954, "step": 3862 }, { "epoch": 0.13612058105834368, "grad_norm": 2.8748085498809814, "learning_rate": 1.941995586345753e-05, "loss": 4.6678, "step": 3863 }, { "epoch": 0.1361558180713021, "grad_norm": 2.2450573444366455, "learning_rate": 1.941957276078654e-05, "loss": 1.7946, "step": 3864 }, { "epoch": 0.13619105508426052, "grad_norm": 2.4419896602630615, "learning_rate": 1.9419189535424422e-05, "loss": 2.6164, "step": 3865 }, { "epoch": 0.13622629209721893, "grad_norm": 3.120743751525879, "learning_rate": 1.9418806187376157e-05, "loss": 2.662, "step": 3866 }, { "epoch": 0.13626152911017733, "grad_norm": 3.981746196746826, "learning_rate": 1.9418422716646744e-05, "loss": 5.4289, "step": 3867 }, { "epoch": 0.13629676612313574, "grad_norm": 2.792375087738037, "learning_rate": 1.9418039123241172e-05, "loss": 2.5751, "step": 3868 }, { "epoch": 0.13633200313609414, "grad_norm": 2.78898286819458, "learning_rate": 1.941765540716444e-05, "loss": 3.1091, "step": 3869 }, { "epoch": 0.13636724014905258, "grad_norm": 4.796361446380615, "learning_rate": 1.9417271568421553e-05, "loss": 9.8181, "step": 3870 }, { "epoch": 0.13640247716201098, "grad_norm": 3.315189838409424, "learning_rate": 1.94168876070175e-05, "loss": 2.998, "step": 3871 }, { "epoch": 0.1364377141749694, "grad_norm": 7.172422409057617, "learning_rate": 1.9416503522957286e-05, "loss": 8.4516, "step": 3872 }, { "epoch": 0.1364729511879278, "grad_norm": 7.312485694885254, "learning_rate": 1.9416119316245915e-05, "loss": 7.2922, "step": 3873 }, { "epoch": 0.1365081882008862, "grad_norm": 2.57352614402771, "learning_rate": 1.9415734986888388e-05, "loss": 5.6009, "step": 3874 }, { "epoch": 0.13654342521384463, "grad_norm": 6.009307384490967, "learning_rate": 1.9415350534889716e-05, "loss": 7.5632, "step": 3875 }, { "epoch": 0.13657866222680304, "grad_norm": 2.3019211292266846, "learning_rate": 1.9414965960254903e-05, "loss": 3.6553, "step": 3876 }, { "epoch": 0.13661389923976144, "grad_norm": 4.398622035980225, "learning_rate": 1.9414581262988958e-05, "loss": 3.2056, "step": 3877 }, { "epoch": 0.13664913625271985, "grad_norm": 7.176769256591797, "learning_rate": 1.9414196443096894e-05, "loss": 9.2752, "step": 3878 }, { "epoch": 0.13668437326567826, "grad_norm": 4.5831298828125, "learning_rate": 1.9413811500583723e-05, "loss": 5.1978, "step": 3879 }, { "epoch": 0.1367196102786367, "grad_norm": 5.745423316955566, "learning_rate": 1.941342643545446e-05, "loss": 5.5361, "step": 3880 }, { "epoch": 0.1367548472915951, "grad_norm": 4.136204242706299, "learning_rate": 1.9413041247714115e-05, "loss": 3.6596, "step": 3881 }, { "epoch": 0.1367900843045535, "grad_norm": 3.102619171142578, "learning_rate": 1.941265593736771e-05, "loss": 2.9128, "step": 3882 }, { "epoch": 0.1368253213175119, "grad_norm": 2.287362575531006, "learning_rate": 1.941227050442026e-05, "loss": 2.7677, "step": 3883 }, { "epoch": 0.13686055833047034, "grad_norm": 2.1540443897247314, "learning_rate": 1.941188494887679e-05, "loss": 2.5763, "step": 3884 }, { "epoch": 0.13689579534342874, "grad_norm": 2.7355048656463623, "learning_rate": 1.941149927074232e-05, "loss": 2.8498, "step": 3885 }, { "epoch": 0.13693103235638715, "grad_norm": 2.263251543045044, "learning_rate": 1.9411113470021872e-05, "loss": 2.8119, "step": 3886 }, { "epoch": 0.13696626936934556, "grad_norm": 8.58885669708252, "learning_rate": 1.9410727546720476e-05, "loss": 5.7192, "step": 3887 }, { "epoch": 0.13700150638230396, "grad_norm": 2.391806125640869, "learning_rate": 1.941034150084315e-05, "loss": 2.9547, "step": 3888 }, { "epoch": 0.1370367433952624, "grad_norm": 8.200161933898926, "learning_rate": 1.940995533239493e-05, "loss": 5.4989, "step": 3889 }, { "epoch": 0.1370719804082208, "grad_norm": 3.045487880706787, "learning_rate": 1.9409569041380846e-05, "loss": 2.9795, "step": 3890 }, { "epoch": 0.1371072174211792, "grad_norm": 8.623766899108887, "learning_rate": 1.940918262780592e-05, "loss": 7.4232, "step": 3891 }, { "epoch": 0.1371424544341376, "grad_norm": 3.1248693466186523, "learning_rate": 1.9408796091675194e-05, "loss": 2.747, "step": 3892 }, { "epoch": 0.13717769144709602, "grad_norm": 5.747650623321533, "learning_rate": 1.9408409432993705e-05, "loss": 5.8215, "step": 3893 }, { "epoch": 0.13721292846005445, "grad_norm": 2.6950876712799072, "learning_rate": 1.940802265176648e-05, "loss": 4.1776, "step": 3894 }, { "epoch": 0.13724816547301286, "grad_norm": 5.101363182067871, "learning_rate": 1.9407635747998563e-05, "loss": 5.6847, "step": 3895 }, { "epoch": 0.13728340248597126, "grad_norm": 3.0485928058624268, "learning_rate": 1.940724872169499e-05, "loss": 4.8326, "step": 3896 }, { "epoch": 0.13731863949892967, "grad_norm": 2.022469997406006, "learning_rate": 1.9406861572860805e-05, "loss": 2.3358, "step": 3897 }, { "epoch": 0.1373538765118881, "grad_norm": 1.9459272623062134, "learning_rate": 1.9406474301501052e-05, "loss": 3.3823, "step": 3898 }, { "epoch": 0.1373891135248465, "grad_norm": 3.0260281562805176, "learning_rate": 1.9406086907620772e-05, "loss": 2.4671, "step": 3899 }, { "epoch": 0.1374243505378049, "grad_norm": 2.989227533340454, "learning_rate": 1.940569939122501e-05, "loss": 4.0131, "step": 3900 }, { "epoch": 0.13745958755076332, "grad_norm": 3.5591940879821777, "learning_rate": 1.9405311752318816e-05, "loss": 5.1505, "step": 3901 }, { "epoch": 0.13749482456372172, "grad_norm": 2.5343425273895264, "learning_rate": 1.940492399090724e-05, "loss": 2.7925, "step": 3902 }, { "epoch": 0.13753006157668016, "grad_norm": 2.7156717777252197, "learning_rate": 1.940453610699533e-05, "loss": 2.9662, "step": 3903 }, { "epoch": 0.13756529858963856, "grad_norm": 3.4013586044311523, "learning_rate": 1.940414810058814e-05, "loss": 6.8403, "step": 3904 }, { "epoch": 0.13760053560259697, "grad_norm": 2.4115164279937744, "learning_rate": 1.9403759971690722e-05, "loss": 2.8697, "step": 3905 }, { "epoch": 0.13763577261555537, "grad_norm": 3.497352361679077, "learning_rate": 1.9403371720308133e-05, "loss": 6.5789, "step": 3906 }, { "epoch": 0.13767100962851378, "grad_norm": 3.9758358001708984, "learning_rate": 1.940298334644543e-05, "loss": 5.5621, "step": 3907 }, { "epoch": 0.1377062466414722, "grad_norm": 4.0008344650268555, "learning_rate": 1.940259485010767e-05, "loss": 6.9351, "step": 3908 }, { "epoch": 0.13774148365443062, "grad_norm": 2.100215435028076, "learning_rate": 1.9402206231299915e-05, "loss": 2.3248, "step": 3909 }, { "epoch": 0.13777672066738902, "grad_norm": 4.2730326652526855, "learning_rate": 1.9401817490027226e-05, "loss": 6.0022, "step": 3910 }, { "epoch": 0.13781195768034743, "grad_norm": 5.725156307220459, "learning_rate": 1.940142862629467e-05, "loss": 5.5759, "step": 3911 }, { "epoch": 0.13784719469330584, "grad_norm": 2.9959442615509033, "learning_rate": 1.9401039640107304e-05, "loss": 2.9644, "step": 3912 }, { "epoch": 0.13788243170626427, "grad_norm": 3.8525376319885254, "learning_rate": 1.94006505314702e-05, "loss": 5.0009, "step": 3913 }, { "epoch": 0.13791766871922267, "grad_norm": 3.5279130935668945, "learning_rate": 1.9400261300388425e-05, "loss": 3.026, "step": 3914 }, { "epoch": 0.13795290573218108, "grad_norm": 2.242257833480835, "learning_rate": 1.939987194686705e-05, "loss": 3.8732, "step": 3915 }, { "epoch": 0.13798814274513949, "grad_norm": 2.3983423709869385, "learning_rate": 1.9399482470911145e-05, "loss": 4.6257, "step": 3916 }, { "epoch": 0.13802337975809792, "grad_norm": 7.538886547088623, "learning_rate": 1.9399092872525786e-05, "loss": 9.3021, "step": 3917 }, { "epoch": 0.13805861677105633, "grad_norm": 3.2943837642669678, "learning_rate": 1.9398703151716045e-05, "loss": 3.2716, "step": 3918 }, { "epoch": 0.13809385378401473, "grad_norm": 3.3959336280822754, "learning_rate": 1.9398313308486996e-05, "loss": 4.9082, "step": 3919 }, { "epoch": 0.13812909079697314, "grad_norm": 3.8407702445983887, "learning_rate": 1.939792334284372e-05, "loss": 4.8676, "step": 3920 }, { "epoch": 0.13816432780993154, "grad_norm": 4.966983318328857, "learning_rate": 1.9397533254791294e-05, "loss": 4.6154, "step": 3921 }, { "epoch": 0.13819956482288998, "grad_norm": 8.469594955444336, "learning_rate": 1.9397143044334804e-05, "loss": 9.514, "step": 3922 }, { "epoch": 0.13823480183584838, "grad_norm": 6.015953540802002, "learning_rate": 1.9396752711479326e-05, "loss": 4.61, "step": 3923 }, { "epoch": 0.1382700388488068, "grad_norm": 2.7901434898376465, "learning_rate": 1.9396362256229942e-05, "loss": 3.1579, "step": 3924 }, { "epoch": 0.1383052758617652, "grad_norm": 24.236648559570312, "learning_rate": 1.939597167859175e-05, "loss": 6.4999, "step": 3925 }, { "epoch": 0.1383405128747236, "grad_norm": 12.683844566345215, "learning_rate": 1.939558097856983e-05, "loss": 10.4684, "step": 3926 }, { "epoch": 0.13837574988768203, "grad_norm": 5.965884685516357, "learning_rate": 1.9395190156169267e-05, "loss": 6.759, "step": 3927 }, { "epoch": 0.13841098690064044, "grad_norm": 6.053569316864014, "learning_rate": 1.9394799211395158e-05, "loss": 8.1725, "step": 3928 }, { "epoch": 0.13844622391359884, "grad_norm": 5.546915054321289, "learning_rate": 1.939440814425259e-05, "loss": 5.3506, "step": 3929 }, { "epoch": 0.13848146092655725, "grad_norm": 2.8527069091796875, "learning_rate": 1.9394016954746663e-05, "loss": 3.3048, "step": 3930 }, { "epoch": 0.13851669793951565, "grad_norm": 2.672649621963501, "learning_rate": 1.9393625642882464e-05, "loss": 2.7481, "step": 3931 }, { "epoch": 0.1385519349524741, "grad_norm": 3.564448356628418, "learning_rate": 1.93932342086651e-05, "loss": 3.1766, "step": 3932 }, { "epoch": 0.1385871719654325, "grad_norm": 2.6633269786834717, "learning_rate": 1.9392842652099658e-05, "loss": 2.5578, "step": 3933 }, { "epoch": 0.1386224089783909, "grad_norm": 5.870407581329346, "learning_rate": 1.9392450973191252e-05, "loss": 7.5044, "step": 3934 }, { "epoch": 0.1386576459913493, "grad_norm": 2.3276991844177246, "learning_rate": 1.939205917194497e-05, "loss": 2.6094, "step": 3935 }, { "epoch": 0.13869288300430774, "grad_norm": 4.2179365158081055, "learning_rate": 1.939166724836592e-05, "loss": 6.0191, "step": 3936 }, { "epoch": 0.13872812001726614, "grad_norm": 3.4150352478027344, "learning_rate": 1.9391275202459208e-05, "loss": 5.4387, "step": 3937 }, { "epoch": 0.13876335703022455, "grad_norm": 3.521592140197754, "learning_rate": 1.939088303422994e-05, "loss": 5.3918, "step": 3938 }, { "epoch": 0.13879859404318295, "grad_norm": 4.980461120605469, "learning_rate": 1.9390490743683228e-05, "loss": 7.7808, "step": 3939 }, { "epoch": 0.13883383105614136, "grad_norm": 4.709226608276367, "learning_rate": 1.9390098330824175e-05, "loss": 6.3003, "step": 3940 }, { "epoch": 0.1388690680690998, "grad_norm": 2.73799991607666, "learning_rate": 1.9389705795657897e-05, "loss": 2.1035, "step": 3941 }, { "epoch": 0.1389043050820582, "grad_norm": 3.0964081287384033, "learning_rate": 1.93893131381895e-05, "loss": 2.8489, "step": 3942 }, { "epoch": 0.1389395420950166, "grad_norm": 3.6388134956359863, "learning_rate": 1.938892035842411e-05, "loss": 6.7139, "step": 3943 }, { "epoch": 0.138974779107975, "grad_norm": 3.0959742069244385, "learning_rate": 1.9388527456366833e-05, "loss": 2.506, "step": 3944 }, { "epoch": 0.13901001612093342, "grad_norm": 3.3913817405700684, "learning_rate": 1.9388134432022793e-05, "loss": 1.851, "step": 3945 }, { "epoch": 0.13904525313389185, "grad_norm": 7.839053153991699, "learning_rate": 1.9387741285397102e-05, "loss": 5.3904, "step": 3946 }, { "epoch": 0.13908049014685026, "grad_norm": 3.0954694747924805, "learning_rate": 1.9387348016494887e-05, "loss": 2.7554, "step": 3947 }, { "epoch": 0.13911572715980866, "grad_norm": 3.1305315494537354, "learning_rate": 1.9386954625321267e-05, "loss": 3.3274, "step": 3948 }, { "epoch": 0.13915096417276707, "grad_norm": 5.924757480621338, "learning_rate": 1.9386561111881372e-05, "loss": 6.554, "step": 3949 }, { "epoch": 0.13918620118572547, "grad_norm": 6.997674942016602, "learning_rate": 1.9386167476180317e-05, "loss": 5.7864, "step": 3950 }, { "epoch": 0.1392214381986839, "grad_norm": 4.592194557189941, "learning_rate": 1.9385773718223237e-05, "loss": 2.3715, "step": 3951 }, { "epoch": 0.1392566752116423, "grad_norm": 7.034521579742432, "learning_rate": 1.938537983801526e-05, "loss": 5.4587, "step": 3952 }, { "epoch": 0.13929191222460072, "grad_norm": 5.6547112464904785, "learning_rate": 1.9384985835561513e-05, "loss": 5.3001, "step": 3953 }, { "epoch": 0.13932714923755912, "grad_norm": 3.10793137550354, "learning_rate": 1.9384591710867132e-05, "loss": 2.6568, "step": 3954 }, { "epoch": 0.13936238625051756, "grad_norm": 6.989922523498535, "learning_rate": 1.9384197463937247e-05, "loss": 6.177, "step": 3955 }, { "epoch": 0.13939762326347596, "grad_norm": 4.409595489501953, "learning_rate": 1.9383803094776997e-05, "loss": 2.2847, "step": 3956 }, { "epoch": 0.13943286027643437, "grad_norm": 19.034385681152344, "learning_rate": 1.9383408603391513e-05, "loss": 5.3297, "step": 3957 }, { "epoch": 0.13946809728939277, "grad_norm": 6.2746968269348145, "learning_rate": 1.938301398978594e-05, "loss": 3.6196, "step": 3958 }, { "epoch": 0.13950333430235118, "grad_norm": 4.582457542419434, "learning_rate": 1.9382619253965413e-05, "loss": 4.2479, "step": 3959 }, { "epoch": 0.1395385713153096, "grad_norm": 4.627603054046631, "learning_rate": 1.9382224395935073e-05, "loss": 3.1851, "step": 3960 }, { "epoch": 0.13957380832826802, "grad_norm": 5.0419230461120605, "learning_rate": 1.9381829415700068e-05, "loss": 7.4423, "step": 3961 }, { "epoch": 0.13960904534122642, "grad_norm": 6.775285243988037, "learning_rate": 1.938143431326554e-05, "loss": 6.44, "step": 3962 }, { "epoch": 0.13964428235418483, "grad_norm": 4.71196985244751, "learning_rate": 1.938103908863663e-05, "loss": 3.0091, "step": 3963 }, { "epoch": 0.13967951936714323, "grad_norm": 6.596994876861572, "learning_rate": 1.9380643741818495e-05, "loss": 5.9187, "step": 3964 }, { "epoch": 0.13971475638010167, "grad_norm": 4.537566661834717, "learning_rate": 1.938024827281628e-05, "loss": 2.6292, "step": 3965 }, { "epoch": 0.13974999339306007, "grad_norm": 3.2697436809539795, "learning_rate": 1.9379852681635137e-05, "loss": 3.0098, "step": 3966 }, { "epoch": 0.13978523040601848, "grad_norm": 2.840169906616211, "learning_rate": 1.937945696828022e-05, "loss": 2.7143, "step": 3967 }, { "epoch": 0.13982046741897688, "grad_norm": 3.0188376903533936, "learning_rate": 1.937906113275668e-05, "loss": 3.323, "step": 3968 }, { "epoch": 0.13985570443193532, "grad_norm": 3.833657741546631, "learning_rate": 1.9378665175069674e-05, "loss": 5.8617, "step": 3969 }, { "epoch": 0.13989094144489372, "grad_norm": 5.307819366455078, "learning_rate": 1.9378269095224354e-05, "loss": 5.4701, "step": 3970 }, { "epoch": 0.13992617845785213, "grad_norm": 6.11728572845459, "learning_rate": 1.9377872893225892e-05, "loss": 11.6692, "step": 3971 }, { "epoch": 0.13996141547081054, "grad_norm": 2.8821139335632324, "learning_rate": 1.9377476569079438e-05, "loss": 2.2905, "step": 3972 }, { "epoch": 0.13999665248376894, "grad_norm": 2.7707841396331787, "learning_rate": 1.9377080122790155e-05, "loss": 5.4226, "step": 3973 }, { "epoch": 0.14003188949672737, "grad_norm": 3.8148670196533203, "learning_rate": 1.937668355436321e-05, "loss": 5.4326, "step": 3974 }, { "epoch": 0.14006712650968578, "grad_norm": 2.447397232055664, "learning_rate": 1.937628686380377e-05, "loss": 2.9259, "step": 3975 }, { "epoch": 0.14010236352264419, "grad_norm": 5.402132034301758, "learning_rate": 1.9375890051116996e-05, "loss": 7.3138, "step": 3976 }, { "epoch": 0.1401376005356026, "grad_norm": 4.392209053039551, "learning_rate": 1.9375493116308064e-05, "loss": 3.3019, "step": 3977 }, { "epoch": 0.140172837548561, "grad_norm": 3.145176887512207, "learning_rate": 1.937509605938214e-05, "loss": 2.9693, "step": 3978 }, { "epoch": 0.14020807456151943, "grad_norm": 4.513721942901611, "learning_rate": 1.937469888034439e-05, "loss": 3.1151, "step": 3979 }, { "epoch": 0.14024331157447784, "grad_norm": 2.3984620571136475, "learning_rate": 1.9374301579199996e-05, "loss": 2.1292, "step": 3980 }, { "epoch": 0.14027854858743624, "grad_norm": 6.366188049316406, "learning_rate": 1.937390415595413e-05, "loss": 8.8982, "step": 3981 }, { "epoch": 0.14031378560039465, "grad_norm": 3.9867711067199707, "learning_rate": 1.9373506610611967e-05, "loss": 3.1916, "step": 3982 }, { "epoch": 0.14034902261335305, "grad_norm": 3.2967615127563477, "learning_rate": 1.937310894317869e-05, "loss": 2.4987, "step": 3983 }, { "epoch": 0.1403842596263115, "grad_norm": 11.345477104187012, "learning_rate": 1.9372711153659473e-05, "loss": 4.5673, "step": 3984 }, { "epoch": 0.1404194966392699, "grad_norm": 16.4132022857666, "learning_rate": 1.9372313242059495e-05, "loss": 5.4859, "step": 3985 }, { "epoch": 0.1404547336522283, "grad_norm": 8.490099906921387, "learning_rate": 1.9371915208383947e-05, "loss": 3.5018, "step": 3986 }, { "epoch": 0.1404899706651867, "grad_norm": 10.849737167358398, "learning_rate": 1.9371517052638008e-05, "loss": 3.2405, "step": 3987 }, { "epoch": 0.14052520767814514, "grad_norm": 7.879405975341797, "learning_rate": 1.937111877482687e-05, "loss": 2.5176, "step": 3988 }, { "epoch": 0.14056044469110354, "grad_norm": 10.867566108703613, "learning_rate": 1.9370720374955714e-05, "loss": 12.7306, "step": 3989 }, { "epoch": 0.14059568170406195, "grad_norm": 16.011566162109375, "learning_rate": 1.937032185302973e-05, "loss": 2.9954, "step": 3990 }, { "epoch": 0.14063091871702035, "grad_norm": 7.804217338562012, "learning_rate": 1.936992320905411e-05, "loss": 2.8493, "step": 3991 }, { "epoch": 0.14066615572997876, "grad_norm": 5.687119483947754, "learning_rate": 1.9369524443034048e-05, "loss": 2.4467, "step": 3992 }, { "epoch": 0.1407013927429372, "grad_norm": 7.974573612213135, "learning_rate": 1.936912555497474e-05, "loss": 5.1333, "step": 3993 }, { "epoch": 0.1407366297558956, "grad_norm": 10.225950241088867, "learning_rate": 1.9368726544881375e-05, "loss": 2.5168, "step": 3994 }, { "epoch": 0.140771866768854, "grad_norm": 4.959244728088379, "learning_rate": 1.9368327412759148e-05, "loss": 2.4996, "step": 3995 }, { "epoch": 0.1408071037818124, "grad_norm": 7.257038593292236, "learning_rate": 1.936792815861327e-05, "loss": 8.2929, "step": 3996 }, { "epoch": 0.14084234079477082, "grad_norm": 14.344478607177734, "learning_rate": 1.9367528782448932e-05, "loss": 2.5591, "step": 3997 }, { "epoch": 0.14087757780772925, "grad_norm": 6.25661039352417, "learning_rate": 1.936712928427134e-05, "loss": 5.4419, "step": 3998 }, { "epoch": 0.14091281482068765, "grad_norm": 6.773632526397705, "learning_rate": 1.9366729664085694e-05, "loss": 5.7702, "step": 3999 }, { "epoch": 0.14094805183364606, "grad_norm": 6.9900054931640625, "learning_rate": 1.9366329921897206e-05, "loss": 5.0055, "step": 4000 }, { "epoch": 0.14098328884660447, "grad_norm": 8.704756736755371, "learning_rate": 1.936593005771107e-05, "loss": 2.9892, "step": 4001 }, { "epoch": 0.14101852585956287, "grad_norm": 8.612591743469238, "learning_rate": 1.9365530071532507e-05, "loss": 4.8657, "step": 4002 }, { "epoch": 0.1410537628725213, "grad_norm": 15.487382888793945, "learning_rate": 1.936512996336672e-05, "loss": 2.3262, "step": 4003 }, { "epoch": 0.1410889998854797, "grad_norm": 4.178868293762207, "learning_rate": 1.9364729733218923e-05, "loss": 1.9946, "step": 4004 }, { "epoch": 0.14112423689843812, "grad_norm": 16.907379150390625, "learning_rate": 1.936432938109433e-05, "loss": 5.1936, "step": 4005 }, { "epoch": 0.14115947391139652, "grad_norm": 7.4861578941345215, "learning_rate": 1.936392890699815e-05, "loss": 3.604, "step": 4006 }, { "epoch": 0.14119471092435495, "grad_norm": 7.9051666259765625, "learning_rate": 1.9363528310935605e-05, "loss": 4.0369, "step": 4007 }, { "epoch": 0.14122994793731336, "grad_norm": 3.8941187858581543, "learning_rate": 1.936312759291191e-05, "loss": 1.5737, "step": 4008 }, { "epoch": 0.14126518495027177, "grad_norm": 7.090855598449707, "learning_rate": 1.9362726752932287e-05, "loss": 1.2744, "step": 4009 }, { "epoch": 0.14130042196323017, "grad_norm": 6.761809825897217, "learning_rate": 1.936232579100196e-05, "loss": 1.8137, "step": 4010 }, { "epoch": 0.14133565897618858, "grad_norm": 4.126063823699951, "learning_rate": 1.936192470712614e-05, "loss": 1.7076, "step": 4011 }, { "epoch": 0.141370895989147, "grad_norm": 14.346500396728516, "learning_rate": 1.936152350131006e-05, "loss": 7.9921, "step": 4012 }, { "epoch": 0.14140613300210542, "grad_norm": 8.352307319641113, "learning_rate": 1.9361122173558945e-05, "loss": 4.917, "step": 4013 }, { "epoch": 0.14144137001506382, "grad_norm": 3.7602274417877197, "learning_rate": 1.936072072387802e-05, "loss": 1.5556, "step": 4014 }, { "epoch": 0.14147660702802223, "grad_norm": 10.548344612121582, "learning_rate": 1.9360319152272516e-05, "loss": 4.9163, "step": 4015 }, { "epoch": 0.14151184404098063, "grad_norm": 6.367082595825195, "learning_rate": 1.935991745874766e-05, "loss": 1.6106, "step": 4016 }, { "epoch": 0.14154708105393907, "grad_norm": 8.569482803344727, "learning_rate": 1.935951564330869e-05, "loss": 4.4125, "step": 4017 }, { "epoch": 0.14158231806689747, "grad_norm": 4.581825256347656, "learning_rate": 1.9359113705960838e-05, "loss": 4.3096, "step": 4018 }, { "epoch": 0.14161755507985588, "grad_norm": 8.685966491699219, "learning_rate": 1.9358711646709335e-05, "loss": 6.6355, "step": 4019 }, { "epoch": 0.14165279209281428, "grad_norm": 3.597579002380371, "learning_rate": 1.935830946555942e-05, "loss": 1.4997, "step": 4020 }, { "epoch": 0.14168802910577272, "grad_norm": 8.174110412597656, "learning_rate": 1.9357907162516338e-05, "loss": 4.5746, "step": 4021 }, { "epoch": 0.14172326611873112, "grad_norm": 5.821982383728027, "learning_rate": 1.9357504737585316e-05, "loss": 4.3996, "step": 4022 }, { "epoch": 0.14175850313168953, "grad_norm": 10.664436340332031, "learning_rate": 1.9357102190771603e-05, "loss": 1.3528, "step": 4023 }, { "epoch": 0.14179374014464793, "grad_norm": 12.2892427444458, "learning_rate": 1.9356699522080447e-05, "loss": 7.3259, "step": 4024 }, { "epoch": 0.14182897715760634, "grad_norm": 3.9920177459716797, "learning_rate": 1.9356296731517086e-05, "loss": 1.3892, "step": 4025 }, { "epoch": 0.14186421417056477, "grad_norm": 22.03141975402832, "learning_rate": 1.9355893819086765e-05, "loss": 4.4281, "step": 4026 }, { "epoch": 0.14189945118352318, "grad_norm": 9.429708480834961, "learning_rate": 1.9355490784794742e-05, "loss": 6.2349, "step": 4027 }, { "epoch": 0.14193468819648158, "grad_norm": 2.620211124420166, "learning_rate": 1.9355087628646256e-05, "loss": 1.331, "step": 4028 }, { "epoch": 0.14196992520944, "grad_norm": 3.2224628925323486, "learning_rate": 1.935468435064656e-05, "loss": 1.2469, "step": 4029 }, { "epoch": 0.1420051622223984, "grad_norm": 7.527635097503662, "learning_rate": 1.935428095080091e-05, "loss": 3.565, "step": 4030 }, { "epoch": 0.14204039923535683, "grad_norm": 5.824441909790039, "learning_rate": 1.935387742911456e-05, "loss": 1.407, "step": 4031 }, { "epoch": 0.14207563624831523, "grad_norm": 4.113899230957031, "learning_rate": 1.9353473785592765e-05, "loss": 1.2841, "step": 4032 }, { "epoch": 0.14211087326127364, "grad_norm": 4.8499345779418945, "learning_rate": 1.935307002024078e-05, "loss": 3.7348, "step": 4033 }, { "epoch": 0.14214611027423205, "grad_norm": 2.010728597640991, "learning_rate": 1.935266613306387e-05, "loss": 1.1097, "step": 4034 }, { "epoch": 0.14218134728719045, "grad_norm": 5.126120567321777, "learning_rate": 1.935226212406729e-05, "loss": 3.3307, "step": 4035 }, { "epoch": 0.14221658430014888, "grad_norm": 4.790380477905273, "learning_rate": 1.935185799325631e-05, "loss": 4.1726, "step": 4036 }, { "epoch": 0.1422518213131073, "grad_norm": 12.698725700378418, "learning_rate": 1.935145374063618e-05, "loss": 2.9101, "step": 4037 }, { "epoch": 0.1422870583260657, "grad_norm": 10.0615873336792, "learning_rate": 1.935104936621218e-05, "loss": 4.1965, "step": 4038 }, { "epoch": 0.1423222953390241, "grad_norm": 14.785320281982422, "learning_rate": 1.9350644869989567e-05, "loss": 9.4677, "step": 4039 }, { "epoch": 0.14235753235198254, "grad_norm": 7.758788108825684, "learning_rate": 1.9350240251973615e-05, "loss": 1.0494, "step": 4040 }, { "epoch": 0.14239276936494094, "grad_norm": 7.077477931976318, "learning_rate": 1.9349835512169593e-05, "loss": 3.5132, "step": 4041 }, { "epoch": 0.14242800637789935, "grad_norm": 4.830840587615967, "learning_rate": 1.934943065058277e-05, "loss": 3.9475, "step": 4042 }, { "epoch": 0.14246324339085775, "grad_norm": 10.759770393371582, "learning_rate": 1.9349025667218426e-05, "loss": 7.4463, "step": 4043 }, { "epoch": 0.14249848040381616, "grad_norm": 7.684426784515381, "learning_rate": 1.934862056208183e-05, "loss": 6.5996, "step": 4044 }, { "epoch": 0.1425337174167746, "grad_norm": 11.586591720581055, "learning_rate": 1.9348215335178262e-05, "loss": 6.9467, "step": 4045 }, { "epoch": 0.142568954429733, "grad_norm": 4.131878852844238, "learning_rate": 1.9347809986512995e-05, "loss": 4.1979, "step": 4046 }, { "epoch": 0.1426041914426914, "grad_norm": 4.096446990966797, "learning_rate": 1.9347404516091318e-05, "loss": 1.1361, "step": 4047 }, { "epoch": 0.1426394284556498, "grad_norm": 3.246220827102661, "learning_rate": 1.93469989239185e-05, "loss": 1.2576, "step": 4048 }, { "epoch": 0.1426746654686082, "grad_norm": 6.6559672355651855, "learning_rate": 1.9346593209999833e-05, "loss": 5.4171, "step": 4049 }, { "epoch": 0.14270990248156665, "grad_norm": 4.099889278411865, "learning_rate": 1.93461873743406e-05, "loss": 1.1036, "step": 4050 }, { "epoch": 0.14274513949452505, "grad_norm": 4.886231899261475, "learning_rate": 1.9345781416946085e-05, "loss": 3.8186, "step": 4051 }, { "epoch": 0.14278037650748346, "grad_norm": 1.6393119096755981, "learning_rate": 1.9345375337821575e-05, "loss": 1.0497, "step": 4052 }, { "epoch": 0.14281561352044186, "grad_norm": 2.3199920654296875, "learning_rate": 1.9344969136972363e-05, "loss": 1.1286, "step": 4053 }, { "epoch": 0.14285085053340027, "grad_norm": 2.891963005065918, "learning_rate": 1.9344562814403734e-05, "loss": 1.3982, "step": 4054 }, { "epoch": 0.1428860875463587, "grad_norm": 2.0277419090270996, "learning_rate": 1.9344156370120987e-05, "loss": 1.2363, "step": 4055 }, { "epoch": 0.1429213245593171, "grad_norm": 2.454958200454712, "learning_rate": 1.9343749804129414e-05, "loss": 1.1322, "step": 4056 }, { "epoch": 0.14295656157227551, "grad_norm": 4.317836761474609, "learning_rate": 1.934334311643431e-05, "loss": 3.1178, "step": 4057 }, { "epoch": 0.14299179858523392, "grad_norm": 1.6697163581848145, "learning_rate": 1.934293630704097e-05, "loss": 1.2633, "step": 4058 }, { "epoch": 0.14302703559819235, "grad_norm": 4.2012619972229, "learning_rate": 1.9342529375954692e-05, "loss": 4.4254, "step": 4059 }, { "epoch": 0.14306227261115076, "grad_norm": 1.3835304975509644, "learning_rate": 1.9342122323180782e-05, "loss": 1.2131, "step": 4060 }, { "epoch": 0.14309750962410916, "grad_norm": 2.719238519668579, "learning_rate": 1.9341715148724538e-05, "loss": 4.0388, "step": 4061 }, { "epoch": 0.14313274663706757, "grad_norm": 1.8342206478118896, "learning_rate": 1.9341307852591265e-05, "loss": 1.1478, "step": 4062 }, { "epoch": 0.14316798365002598, "grad_norm": 1.1797378063201904, "learning_rate": 1.934090043478627e-05, "loss": 0.8886, "step": 4063 }, { "epoch": 0.1432032206629844, "grad_norm": 2.612637996673584, "learning_rate": 1.9340492895314852e-05, "loss": 1.2524, "step": 4064 }, { "epoch": 0.14323845767594282, "grad_norm": 1.6968551874160767, "learning_rate": 1.9340085234182327e-05, "loss": 1.1566, "step": 4065 }, { "epoch": 0.14327369468890122, "grad_norm": 3.644427537918091, "learning_rate": 1.9339677451394003e-05, "loss": 3.4069, "step": 4066 }, { "epoch": 0.14330893170185963, "grad_norm": 6.914254665374756, "learning_rate": 1.9339269546955186e-05, "loss": 6.7128, "step": 4067 }, { "epoch": 0.14334416871481803, "grad_norm": 7.351336479187012, "learning_rate": 1.9338861520871197e-05, "loss": 6.5254, "step": 4068 }, { "epoch": 0.14337940572777647, "grad_norm": 7.873252868652344, "learning_rate": 1.9338453373147347e-05, "loss": 3.7793, "step": 4069 }, { "epoch": 0.14341464274073487, "grad_norm": 1.291774868965149, "learning_rate": 1.933804510378895e-05, "loss": 0.9703, "step": 4070 }, { "epoch": 0.14344987975369328, "grad_norm": 3.9767651557922363, "learning_rate": 1.933763671280133e-05, "loss": 3.6582, "step": 4071 }, { "epoch": 0.14348511676665168, "grad_norm": 6.625185012817383, "learning_rate": 1.9337228200189798e-05, "loss": 3.9818, "step": 4072 }, { "epoch": 0.14352035377961012, "grad_norm": 18.284265518188477, "learning_rate": 1.933681956595968e-05, "loss": 1.0624, "step": 4073 }, { "epoch": 0.14355559079256852, "grad_norm": 4.52516508102417, "learning_rate": 1.93364108101163e-05, "loss": 2.7904, "step": 4074 }, { "epoch": 0.14359082780552693, "grad_norm": 4.743220329284668, "learning_rate": 1.9336001932664978e-05, "loss": 1.1431, "step": 4075 }, { "epoch": 0.14362606481848533, "grad_norm": 6.218297958374023, "learning_rate": 1.9335592933611044e-05, "loss": 2.9675, "step": 4076 }, { "epoch": 0.14366130183144374, "grad_norm": 1.9492603540420532, "learning_rate": 1.933518381295982e-05, "loss": 0.9095, "step": 4077 }, { "epoch": 0.14369653884440217, "grad_norm": 1.8270745277404785, "learning_rate": 1.933477457071664e-05, "loss": 1.0283, "step": 4078 }, { "epoch": 0.14373177585736058, "grad_norm": 1.9475892782211304, "learning_rate": 1.9334365206886835e-05, "loss": 1.2201, "step": 4079 }, { "epoch": 0.14376701287031898, "grad_norm": 3.1510095596313477, "learning_rate": 1.9333955721475727e-05, "loss": 3.8458, "step": 4080 }, { "epoch": 0.1438022498832774, "grad_norm": 4.172985076904297, "learning_rate": 1.9333546114488662e-05, "loss": 6.1539, "step": 4081 }, { "epoch": 0.1438374868962358, "grad_norm": 2.2207860946655273, "learning_rate": 1.9333136385930967e-05, "loss": 1.0445, "step": 4082 }, { "epoch": 0.14387272390919423, "grad_norm": 1.882673978805542, "learning_rate": 1.9332726535807984e-05, "loss": 1.1299, "step": 4083 }, { "epoch": 0.14390796092215263, "grad_norm": 4.174638271331787, "learning_rate": 1.9332316564125046e-05, "loss": 6.307, "step": 4084 }, { "epoch": 0.14394319793511104, "grad_norm": 7.352062225341797, "learning_rate": 1.93319064708875e-05, "loss": 3.7578, "step": 4085 }, { "epoch": 0.14397843494806944, "grad_norm": 1.8375787734985352, "learning_rate": 1.933149625610068e-05, "loss": 1.071, "step": 4086 }, { "epoch": 0.14401367196102785, "grad_norm": 3.1062357425689697, "learning_rate": 1.9331085919769937e-05, "loss": 3.3515, "step": 4087 }, { "epoch": 0.14404890897398628, "grad_norm": 4.256103515625, "learning_rate": 1.9330675461900608e-05, "loss": 3.827, "step": 4088 }, { "epoch": 0.1440841459869447, "grad_norm": 0.9580897688865662, "learning_rate": 1.9330264882498042e-05, "loss": 1.0651, "step": 4089 }, { "epoch": 0.1441193829999031, "grad_norm": 1.356645107269287, "learning_rate": 1.9329854181567587e-05, "loss": 1.1244, "step": 4090 }, { "epoch": 0.1441546200128615, "grad_norm": 10.04287338256836, "learning_rate": 1.9329443359114596e-05, "loss": 4.1252, "step": 4091 }, { "epoch": 0.14418985702581993, "grad_norm": 6.094362735748291, "learning_rate": 1.9329032415144415e-05, "loss": 4.1103, "step": 4092 }, { "epoch": 0.14422509403877834, "grad_norm": 6.493232250213623, "learning_rate": 1.93286213496624e-05, "loss": 6.1681, "step": 4093 }, { "epoch": 0.14426033105173675, "grad_norm": 5.110296249389648, "learning_rate": 1.9328210162673904e-05, "loss": 6.6462, "step": 4094 }, { "epoch": 0.14429556806469515, "grad_norm": 4.502829074859619, "learning_rate": 1.932779885418428e-05, "loss": 4.5807, "step": 4095 }, { "epoch": 0.14433080507765356, "grad_norm": 3.4058570861816406, "learning_rate": 1.932738742419889e-05, "loss": 2.943, "step": 4096 }, { "epoch": 0.144366042090612, "grad_norm": 3.0171799659729004, "learning_rate": 1.932697587272309e-05, "loss": 3.6602, "step": 4097 }, { "epoch": 0.1444012791035704, "grad_norm": 1.6033084392547607, "learning_rate": 1.9326564199762238e-05, "loss": 1.0715, "step": 4098 }, { "epoch": 0.1444365161165288, "grad_norm": 5.097052097320557, "learning_rate": 1.93261524053217e-05, "loss": 6.4335, "step": 4099 }, { "epoch": 0.1444717531294872, "grad_norm": 2.3705337047576904, "learning_rate": 1.9325740489406844e-05, "loss": 3.2881, "step": 4100 }, { "epoch": 0.1445069901424456, "grad_norm": 4.966952323913574, "learning_rate": 1.9325328452023025e-05, "loss": 3.8484, "step": 4101 }, { "epoch": 0.14454222715540405, "grad_norm": 2.4167861938476562, "learning_rate": 1.932491629317562e-05, "loss": 3.5504, "step": 4102 }, { "epoch": 0.14457746416836245, "grad_norm": 5.054408550262451, "learning_rate": 1.9324504012869987e-05, "loss": 4.6379, "step": 4103 }, { "epoch": 0.14461270118132086, "grad_norm": 5.2430243492126465, "learning_rate": 1.9324091611111502e-05, "loss": 5.2053, "step": 4104 }, { "epoch": 0.14464793819427926, "grad_norm": 1.2139253616333008, "learning_rate": 1.932367908790554e-05, "loss": 1.1133, "step": 4105 }, { "epoch": 0.14468317520723767, "grad_norm": 2.5112128257751465, "learning_rate": 1.9323266443257466e-05, "loss": 3.4172, "step": 4106 }, { "epoch": 0.1447184122201961, "grad_norm": 1.6673588752746582, "learning_rate": 1.9322853677172662e-05, "loss": 1.2172, "step": 4107 }, { "epoch": 0.1447536492331545, "grad_norm": 1.5668022632598877, "learning_rate": 1.93224407896565e-05, "loss": 1.2191, "step": 4108 }, { "epoch": 0.1447888862461129, "grad_norm": 3.9365732669830322, "learning_rate": 1.932202778071436e-05, "loss": 9.0101, "step": 4109 }, { "epoch": 0.14482412325907132, "grad_norm": 2.0121028423309326, "learning_rate": 1.9321614650351622e-05, "loss": 1.2165, "step": 4110 }, { "epoch": 0.14485936027202975, "grad_norm": 3.9433515071868896, "learning_rate": 1.9321201398573664e-05, "loss": 3.6763, "step": 4111 }, { "epoch": 0.14489459728498816, "grad_norm": 1.269294023513794, "learning_rate": 1.932078802538587e-05, "loss": 1.1889, "step": 4112 }, { "epoch": 0.14492983429794656, "grad_norm": 5.360963344573975, "learning_rate": 1.9320374530793624e-05, "loss": 5.7754, "step": 4113 }, { "epoch": 0.14496507131090497, "grad_norm": 6.42921781539917, "learning_rate": 1.9319960914802317e-05, "loss": 7.1546, "step": 4114 }, { "epoch": 0.14500030832386337, "grad_norm": 1.9163023233413696, "learning_rate": 1.9319547177417324e-05, "loss": 1.2632, "step": 4115 }, { "epoch": 0.1450355453368218, "grad_norm": 2.491023063659668, "learning_rate": 1.931913331864405e-05, "loss": 3.6743, "step": 4116 }, { "epoch": 0.14507078234978021, "grad_norm": 1.51908540725708, "learning_rate": 1.9318719338487875e-05, "loss": 0.9831, "step": 4117 }, { "epoch": 0.14510601936273862, "grad_norm": 3.3546907901763916, "learning_rate": 1.9318305236954192e-05, "loss": 4.6464, "step": 4118 }, { "epoch": 0.14514125637569703, "grad_norm": 1.595555067062378, "learning_rate": 1.9317891014048396e-05, "loss": 1.0831, "step": 4119 }, { "epoch": 0.14517649338865543, "grad_norm": 3.6467061042785645, "learning_rate": 1.9317476669775884e-05, "loss": 3.6616, "step": 4120 }, { "epoch": 0.14521173040161386, "grad_norm": 1.4089614152908325, "learning_rate": 1.9317062204142053e-05, "loss": 0.9273, "step": 4121 }, { "epoch": 0.14524696741457227, "grad_norm": 1.6296602487564087, "learning_rate": 1.9316647617152297e-05, "loss": 0.8425, "step": 4122 }, { "epoch": 0.14528220442753068, "grad_norm": 1.9183989763259888, "learning_rate": 1.931623290881202e-05, "loss": 1.2146, "step": 4123 }, { "epoch": 0.14531744144048908, "grad_norm": 3.086883544921875, "learning_rate": 1.931581807912662e-05, "loss": 3.3078, "step": 4124 }, { "epoch": 0.14535267845344751, "grad_norm": 3.0686707496643066, "learning_rate": 1.931540312810151e-05, "loss": 3.3944, "step": 4125 }, { "epoch": 0.14538791546640592, "grad_norm": 3.733945608139038, "learning_rate": 1.9314988055742084e-05, "loss": 4.038, "step": 4126 }, { "epoch": 0.14542315247936433, "grad_norm": 5.397066116333008, "learning_rate": 1.931457286205375e-05, "loss": 4.4883, "step": 4127 }, { "epoch": 0.14545838949232273, "grad_norm": 2.9444401264190674, "learning_rate": 1.9314157547041922e-05, "loss": 3.6996, "step": 4128 }, { "epoch": 0.14549362650528114, "grad_norm": 2.0754141807556152, "learning_rate": 1.9313742110712003e-05, "loss": 1.0658, "step": 4129 }, { "epoch": 0.14552886351823957, "grad_norm": 4.057022571563721, "learning_rate": 1.9313326553069402e-05, "loss": 4.1571, "step": 4130 }, { "epoch": 0.14556410053119798, "grad_norm": 4.485841751098633, "learning_rate": 1.9312910874119543e-05, "loss": 4.0793, "step": 4131 }, { "epoch": 0.14559933754415638, "grad_norm": 4.433752059936523, "learning_rate": 1.931249507386783e-05, "loss": 3.8035, "step": 4132 }, { "epoch": 0.1456345745571148, "grad_norm": 1.2452988624572754, "learning_rate": 1.931207915231969e-05, "loss": 1.2894, "step": 4133 }, { "epoch": 0.1456698115700732, "grad_norm": 2.178683042526245, "learning_rate": 1.9311663109480522e-05, "loss": 2.7076, "step": 4134 }, { "epoch": 0.14570504858303163, "grad_norm": 8.337366104125977, "learning_rate": 1.931124694535576e-05, "loss": 5.9511, "step": 4135 }, { "epoch": 0.14574028559599003, "grad_norm": 1.3489969968795776, "learning_rate": 1.931083065995082e-05, "loss": 0.9858, "step": 4136 }, { "epoch": 0.14577552260894844, "grad_norm": 4.818933010101318, "learning_rate": 1.9310414253271127e-05, "loss": 3.1107, "step": 4137 }, { "epoch": 0.14581075962190684, "grad_norm": 5.561933994293213, "learning_rate": 1.9309997725322104e-05, "loss": 4.0133, "step": 4138 }, { "epoch": 0.14584599663486525, "grad_norm": 1.9835983514785767, "learning_rate": 1.9309581076109173e-05, "loss": 1.2275, "step": 4139 }, { "epoch": 0.14588123364782368, "grad_norm": 2.5075111389160156, "learning_rate": 1.9309164305637758e-05, "loss": 2.5438, "step": 4140 }, { "epoch": 0.1459164706607821, "grad_norm": 3.046015977859497, "learning_rate": 1.9308747413913297e-05, "loss": 3.6736, "step": 4141 }, { "epoch": 0.1459517076737405, "grad_norm": 1.2007125616073608, "learning_rate": 1.9308330400941214e-05, "loss": 0.82, "step": 4142 }, { "epoch": 0.1459869446866989, "grad_norm": 3.163977861404419, "learning_rate": 1.9307913266726942e-05, "loss": 3.5268, "step": 4143 }, { "epoch": 0.14602218169965733, "grad_norm": 3.5917704105377197, "learning_rate": 1.9307496011275916e-05, "loss": 3.9482, "step": 4144 }, { "epoch": 0.14605741871261574, "grad_norm": 3.7692031860351562, "learning_rate": 1.9307078634593564e-05, "loss": 3.616, "step": 4145 }, { "epoch": 0.14609265572557414, "grad_norm": 5.940003395080566, "learning_rate": 1.930666113668533e-05, "loss": 9.7192, "step": 4146 }, { "epoch": 0.14612789273853255, "grad_norm": 3.7696211338043213, "learning_rate": 1.930624351755665e-05, "loss": 3.566, "step": 4147 }, { "epoch": 0.14616312975149096, "grad_norm": 3.3331427574157715, "learning_rate": 1.9305825777212962e-05, "loss": 2.9103, "step": 4148 }, { "epoch": 0.1461983667644494, "grad_norm": 1.404005527496338, "learning_rate": 1.9305407915659705e-05, "loss": 1.1922, "step": 4149 }, { "epoch": 0.1462336037774078, "grad_norm": 1.4078254699707031, "learning_rate": 1.930498993290233e-05, "loss": 0.8895, "step": 4150 }, { "epoch": 0.1462688407903662, "grad_norm": 6.515452861785889, "learning_rate": 1.9304571828946272e-05, "loss": 4.0138, "step": 4151 }, { "epoch": 0.1463040778033246, "grad_norm": 8.378107070922852, "learning_rate": 1.9304153603796983e-05, "loss": 9.1322, "step": 4152 }, { "epoch": 0.146339314816283, "grad_norm": 1.1271774768829346, "learning_rate": 1.9303735257459905e-05, "loss": 1.1746, "step": 4153 }, { "epoch": 0.14637455182924144, "grad_norm": 4.125293731689453, "learning_rate": 1.9303316789940493e-05, "loss": 3.2796, "step": 4154 }, { "epoch": 0.14640978884219985, "grad_norm": 0.9821047782897949, "learning_rate": 1.9302898201244192e-05, "loss": 1.144, "step": 4155 }, { "epoch": 0.14644502585515826, "grad_norm": 8.83704662322998, "learning_rate": 1.930247949137646e-05, "loss": 3.5263, "step": 4156 }, { "epoch": 0.14648026286811666, "grad_norm": 5.9453511238098145, "learning_rate": 1.930206066034275e-05, "loss": 3.2442, "step": 4157 }, { "epoch": 0.14651549988107507, "grad_norm": 1.2640084028244019, "learning_rate": 1.930164170814851e-05, "loss": 1.0433, "step": 4158 }, { "epoch": 0.1465507368940335, "grad_norm": 4.038901329040527, "learning_rate": 1.93012226347992e-05, "loss": 3.5554, "step": 4159 }, { "epoch": 0.1465859739069919, "grad_norm": 1.4585164785385132, "learning_rate": 1.9300803440300287e-05, "loss": 1.0249, "step": 4160 }, { "epoch": 0.1466212109199503, "grad_norm": 8.007558822631836, "learning_rate": 1.930038412465722e-05, "loss": 6.4454, "step": 4161 }, { "epoch": 0.14665644793290872, "grad_norm": 1.0447643995285034, "learning_rate": 1.9299964687875466e-05, "loss": 1.0596, "step": 4162 }, { "epoch": 0.14669168494586715, "grad_norm": 10.15427017211914, "learning_rate": 1.929954512996049e-05, "loss": 4.2022, "step": 4163 }, { "epoch": 0.14672692195882556, "grad_norm": 5.122784614562988, "learning_rate": 1.929912545091775e-05, "loss": 3.0199, "step": 4164 }, { "epoch": 0.14676215897178396, "grad_norm": 3.9668185710906982, "learning_rate": 1.929870565075272e-05, "loss": 3.3083, "step": 4165 }, { "epoch": 0.14679739598474237, "grad_norm": 1.2363795042037964, "learning_rate": 1.929828572947086e-05, "loss": 1.2656, "step": 4166 }, { "epoch": 0.14683263299770077, "grad_norm": 4.614861488342285, "learning_rate": 1.929786568707765e-05, "loss": 3.678, "step": 4167 }, { "epoch": 0.1468678700106592, "grad_norm": 1.4063780307769775, "learning_rate": 1.929744552357855e-05, "loss": 0.9261, "step": 4168 }, { "epoch": 0.1469031070236176, "grad_norm": 4.0487799644470215, "learning_rate": 1.929702523897904e-05, "loss": 3.6616, "step": 4169 }, { "epoch": 0.14693834403657602, "grad_norm": 2.2636377811431885, "learning_rate": 1.929660483328459e-05, "loss": 3.0084, "step": 4170 }, { "epoch": 0.14697358104953442, "grad_norm": 5.946460247039795, "learning_rate": 1.9296184306500684e-05, "loss": 3.2403, "step": 4171 }, { "epoch": 0.14700881806249283, "grad_norm": 1.7847079038619995, "learning_rate": 1.9295763658632784e-05, "loss": 0.9863, "step": 4172 }, { "epoch": 0.14704405507545126, "grad_norm": 0.9867799282073975, "learning_rate": 1.9295342889686385e-05, "loss": 1.2445, "step": 4173 }, { "epoch": 0.14707929208840967, "grad_norm": 1.3239730596542358, "learning_rate": 1.929492199966696e-05, "loss": 1.052, "step": 4174 }, { "epoch": 0.14711452910136807, "grad_norm": 1.3413820266723633, "learning_rate": 1.929450098857999e-05, "loss": 1.0696, "step": 4175 }, { "epoch": 0.14714976611432648, "grad_norm": 5.692925453186035, "learning_rate": 1.9294079856430962e-05, "loss": 3.3767, "step": 4176 }, { "epoch": 0.1471850031272849, "grad_norm": 3.2980053424835205, "learning_rate": 1.929365860322536e-05, "loss": 3.7179, "step": 4177 }, { "epoch": 0.14722024014024332, "grad_norm": 6.123184680938721, "learning_rate": 1.9293237228968667e-05, "loss": 5.7973, "step": 4178 }, { "epoch": 0.14725547715320172, "grad_norm": 1.1999667882919312, "learning_rate": 1.929281573366638e-05, "loss": 0.9644, "step": 4179 }, { "epoch": 0.14729071416616013, "grad_norm": 1.3879034519195557, "learning_rate": 1.9292394117323982e-05, "loss": 1.0903, "step": 4180 }, { "epoch": 0.14732595117911854, "grad_norm": 9.030693054199219, "learning_rate": 1.9291972379946967e-05, "loss": 8.718, "step": 4181 }, { "epoch": 0.14736118819207697, "grad_norm": 1.299884557723999, "learning_rate": 1.9291550521540832e-05, "loss": 1.3224, "step": 4182 }, { "epoch": 0.14739642520503538, "grad_norm": 3.9505786895751953, "learning_rate": 1.9291128542111063e-05, "loss": 4.3461, "step": 4183 }, { "epoch": 0.14743166221799378, "grad_norm": 5.061427116394043, "learning_rate": 1.9290706441663163e-05, "loss": 3.2195, "step": 4184 }, { "epoch": 0.1474668992309522, "grad_norm": 1.325574517250061, "learning_rate": 1.929028422020263e-05, "loss": 0.9369, "step": 4185 }, { "epoch": 0.1475021362439106, "grad_norm": 5.326825141906738, "learning_rate": 1.928986187773496e-05, "loss": 3.2598, "step": 4186 }, { "epoch": 0.14753737325686903, "grad_norm": 4.870950698852539, "learning_rate": 1.9289439414265656e-05, "loss": 3.3679, "step": 4187 }, { "epoch": 0.14757261026982743, "grad_norm": 1.4108093976974487, "learning_rate": 1.928901682980022e-05, "loss": 1.1704, "step": 4188 }, { "epoch": 0.14760784728278584, "grad_norm": 7.603911876678467, "learning_rate": 1.9288594124344156e-05, "loss": 6.3293, "step": 4189 }, { "epoch": 0.14764308429574424, "grad_norm": 4.462197780609131, "learning_rate": 1.928817129790297e-05, "loss": 3.146, "step": 4190 }, { "epoch": 0.14767832130870265, "grad_norm": 1.7676726579666138, "learning_rate": 1.9287748350482172e-05, "loss": 1.0145, "step": 4191 }, { "epoch": 0.14771355832166108, "grad_norm": 1.4465265274047852, "learning_rate": 1.9287325282087266e-05, "loss": 1.2503, "step": 4192 }, { "epoch": 0.1477487953346195, "grad_norm": 5.578821659088135, "learning_rate": 1.9286902092723767e-05, "loss": 4.5717, "step": 4193 }, { "epoch": 0.1477840323475779, "grad_norm": 1.1939911842346191, "learning_rate": 1.9286478782397187e-05, "loss": 1.2325, "step": 4194 }, { "epoch": 0.1478192693605363, "grad_norm": 5.548928737640381, "learning_rate": 1.9286055351113034e-05, "loss": 3.463, "step": 4195 }, { "epoch": 0.14785450637349473, "grad_norm": 1.294321894645691, "learning_rate": 1.9285631798876826e-05, "loss": 1.0942, "step": 4196 }, { "epoch": 0.14788974338645314, "grad_norm": 2.584657669067383, "learning_rate": 1.9285208125694087e-05, "loss": 4.2631, "step": 4197 }, { "epoch": 0.14792498039941154, "grad_norm": 1.362089991569519, "learning_rate": 1.9284784331570327e-05, "loss": 0.9254, "step": 4198 }, { "epoch": 0.14796021741236995, "grad_norm": 1.048959732055664, "learning_rate": 1.928436041651107e-05, "loss": 1.1133, "step": 4199 }, { "epoch": 0.14799545442532835, "grad_norm": 1.7144578695297241, "learning_rate": 1.9283936380521832e-05, "loss": 1.0926, "step": 4200 }, { "epoch": 0.1480306914382868, "grad_norm": 4.960916519165039, "learning_rate": 1.9283512223608142e-05, "loss": 6.1587, "step": 4201 }, { "epoch": 0.1480659284512452, "grad_norm": 1.0062354803085327, "learning_rate": 1.928308794577552e-05, "loss": 0.9901, "step": 4202 }, { "epoch": 0.1481011654642036, "grad_norm": 4.250844955444336, "learning_rate": 1.92826635470295e-05, "loss": 6.9666, "step": 4203 }, { "epoch": 0.148136402477162, "grad_norm": 7.232949733734131, "learning_rate": 1.92822390273756e-05, "loss": 9.8886, "step": 4204 }, { "epoch": 0.1481716394901204, "grad_norm": 2.7009823322296143, "learning_rate": 1.9281814386819358e-05, "loss": 3.7391, "step": 4205 }, { "epoch": 0.14820687650307884, "grad_norm": 6.713744163513184, "learning_rate": 1.9281389625366297e-05, "loss": 3.706, "step": 4206 }, { "epoch": 0.14824211351603725, "grad_norm": 1.2141296863555908, "learning_rate": 1.928096474302196e-05, "loss": 0.9129, "step": 4207 }, { "epoch": 0.14827735052899566, "grad_norm": 1.6312123537063599, "learning_rate": 1.928053973979187e-05, "loss": 1.2556, "step": 4208 }, { "epoch": 0.14831258754195406, "grad_norm": 4.463647842407227, "learning_rate": 1.928011461568157e-05, "loss": 3.4514, "step": 4209 }, { "epoch": 0.14834782455491247, "grad_norm": 2.1931066513061523, "learning_rate": 1.9279689370696598e-05, "loss": 0.927, "step": 4210 }, { "epoch": 0.1483830615678709, "grad_norm": 1.5370726585388184, "learning_rate": 1.9279264004842482e-05, "loss": 1.1882, "step": 4211 }, { "epoch": 0.1484182985808293, "grad_norm": 2.0230307579040527, "learning_rate": 1.9278838518124775e-05, "loss": 0.9925, "step": 4212 }, { "epoch": 0.1484535355937877, "grad_norm": 1.5972950458526611, "learning_rate": 1.9278412910549015e-05, "loss": 0.8928, "step": 4213 }, { "epoch": 0.14848877260674612, "grad_norm": 8.010602951049805, "learning_rate": 1.9277987182120746e-05, "loss": 5.8226, "step": 4214 }, { "epoch": 0.14852400961970455, "grad_norm": 6.356420516967773, "learning_rate": 1.927756133284551e-05, "loss": 2.8222, "step": 4215 }, { "epoch": 0.14855924663266296, "grad_norm": 6.116845607757568, "learning_rate": 1.9277135362728857e-05, "loss": 3.2516, "step": 4216 }, { "epoch": 0.14859448364562136, "grad_norm": 7.794079780578613, "learning_rate": 1.927670927177633e-05, "loss": 6.5508, "step": 4217 }, { "epoch": 0.14862972065857977, "grad_norm": 4.682091236114502, "learning_rate": 1.927628305999349e-05, "loss": 3.328, "step": 4218 }, { "epoch": 0.14866495767153817, "grad_norm": 13.209299087524414, "learning_rate": 1.9275856727385877e-05, "loss": 4.3054, "step": 4219 }, { "epoch": 0.1487001946844966, "grad_norm": 5.108639240264893, "learning_rate": 1.9275430273959048e-05, "loss": 3.7876, "step": 4220 }, { "epoch": 0.148735431697455, "grad_norm": 1.5993660688400269, "learning_rate": 1.9275003699718562e-05, "loss": 1.2077, "step": 4221 }, { "epoch": 0.14877066871041342, "grad_norm": 4.536537170410156, "learning_rate": 1.927457700466997e-05, "loss": 3.642, "step": 4222 }, { "epoch": 0.14880590572337182, "grad_norm": 1.2175191640853882, "learning_rate": 1.927415018881883e-05, "loss": 0.9888, "step": 4223 }, { "epoch": 0.14884114273633023, "grad_norm": 1.5396636724472046, "learning_rate": 1.92737232521707e-05, "loss": 0.9896, "step": 4224 }, { "epoch": 0.14887637974928866, "grad_norm": 1.32395339012146, "learning_rate": 1.927329619473115e-05, "loss": 1.0397, "step": 4225 }, { "epoch": 0.14891161676224707, "grad_norm": 1.2000492811203003, "learning_rate": 1.927286901650573e-05, "loss": 1.0518, "step": 4226 }, { "epoch": 0.14894685377520547, "grad_norm": 1.3850131034851074, "learning_rate": 1.927244171750001e-05, "loss": 0.9793, "step": 4227 }, { "epoch": 0.14898209078816388, "grad_norm": 1.619225263595581, "learning_rate": 1.9272014297719558e-05, "loss": 1.3617, "step": 4228 }, { "epoch": 0.1490173278011223, "grad_norm": 1.495393991470337, "learning_rate": 1.927158675716994e-05, "loss": 0.9657, "step": 4229 }, { "epoch": 0.14905256481408072, "grad_norm": 7.299126625061035, "learning_rate": 1.927115909585672e-05, "loss": 7.2717, "step": 4230 }, { "epoch": 0.14908780182703912, "grad_norm": 5.8382978439331055, "learning_rate": 1.9270731313785473e-05, "loss": 3.3065, "step": 4231 }, { "epoch": 0.14912303883999753, "grad_norm": 3.2627811431884766, "learning_rate": 1.927030341096177e-05, "loss": 3.6664, "step": 4232 }, { "epoch": 0.14915827585295593, "grad_norm": 1.1362310647964478, "learning_rate": 1.9269875387391184e-05, "loss": 1.1787, "step": 4233 }, { "epoch": 0.14919351286591437, "grad_norm": 2.7419679164886475, "learning_rate": 1.926944724307929e-05, "loss": 3.5929, "step": 4234 }, { "epoch": 0.14922874987887277, "grad_norm": 6.424813270568848, "learning_rate": 1.9269018978031665e-05, "loss": 3.2328, "step": 4235 }, { "epoch": 0.14926398689183118, "grad_norm": 3.7136573791503906, "learning_rate": 1.926859059225389e-05, "loss": 5.8103, "step": 4236 }, { "epoch": 0.14929922390478959, "grad_norm": 1.9547044038772583, "learning_rate": 1.9268162085751538e-05, "loss": 1.1469, "step": 4237 }, { "epoch": 0.149334460917748, "grad_norm": 4.5405354499816895, "learning_rate": 1.9267733458530195e-05, "loss": 6.3709, "step": 4238 }, { "epoch": 0.14936969793070642, "grad_norm": 2.213195562362671, "learning_rate": 1.9267304710595444e-05, "loss": 3.7291, "step": 4239 }, { "epoch": 0.14940493494366483, "grad_norm": 7.265406608581543, "learning_rate": 1.926687584195287e-05, "loss": 5.8979, "step": 4240 }, { "epoch": 0.14944017195662324, "grad_norm": 4.035183429718018, "learning_rate": 1.9266446852608055e-05, "loss": 3.3568, "step": 4241 }, { "epoch": 0.14947540896958164, "grad_norm": 1.5733064413070679, "learning_rate": 1.9266017742566593e-05, "loss": 0.9614, "step": 4242 }, { "epoch": 0.14951064598254005, "grad_norm": 2.6589508056640625, "learning_rate": 1.9265588511834066e-05, "loss": 3.4032, "step": 4243 }, { "epoch": 0.14954588299549848, "grad_norm": 1.5327279567718506, "learning_rate": 1.926515916041607e-05, "loss": 1.017, "step": 4244 }, { "epoch": 0.14958112000845689, "grad_norm": 5.22657585144043, "learning_rate": 1.9264729688318195e-05, "loss": 8.6764, "step": 4245 }, { "epoch": 0.1496163570214153, "grad_norm": 1.2230833768844604, "learning_rate": 1.926430009554604e-05, "loss": 1.2887, "step": 4246 }, { "epoch": 0.1496515940343737, "grad_norm": 2.6274054050445557, "learning_rate": 1.926387038210519e-05, "loss": 3.1599, "step": 4247 }, { "epoch": 0.14968683104733213, "grad_norm": 3.7149932384490967, "learning_rate": 1.926344054800125e-05, "loss": 3.1829, "step": 4248 }, { "epoch": 0.14972206806029054, "grad_norm": 5.260887145996094, "learning_rate": 1.926301059323982e-05, "loss": 3.577, "step": 4249 }, { "epoch": 0.14975730507324894, "grad_norm": 2.6062557697296143, "learning_rate": 1.9262580517826494e-05, "loss": 3.7065, "step": 4250 }, { "epoch": 0.14979254208620735, "grad_norm": 4.319173812866211, "learning_rate": 1.9262150321766875e-05, "loss": 3.5866, "step": 4251 }, { "epoch": 0.14982777909916575, "grad_norm": 1.187296748161316, "learning_rate": 1.9261720005066572e-05, "loss": 0.9124, "step": 4252 }, { "epoch": 0.1498630161121242, "grad_norm": 1.567838430404663, "learning_rate": 1.9261289567731186e-05, "loss": 1.2246, "step": 4253 }, { "epoch": 0.1498982531250826, "grad_norm": 2.2642533779144287, "learning_rate": 1.9260859009766322e-05, "loss": 2.9546, "step": 4254 }, { "epoch": 0.149933490138041, "grad_norm": 5.2177534103393555, "learning_rate": 1.9260428331177586e-05, "loss": 3.5935, "step": 4255 }, { "epoch": 0.1499687271509994, "grad_norm": 5.051828861236572, "learning_rate": 1.92599975319706e-05, "loss": 3.5665, "step": 4256 }, { "epoch": 0.1500039641639578, "grad_norm": 5.198858737945557, "learning_rate": 1.925956661215096e-05, "loss": 3.9761, "step": 4257 }, { "epoch": 0.15003920117691624, "grad_norm": 1.2898907661437988, "learning_rate": 1.9259135571724287e-05, "loss": 0.9796, "step": 4258 }, { "epoch": 0.15007443818987465, "grad_norm": 7.999283313751221, "learning_rate": 1.9258704410696196e-05, "loss": 3.7969, "step": 4259 }, { "epoch": 0.15010967520283305, "grad_norm": 8.435197830200195, "learning_rate": 1.92582731290723e-05, "loss": 9.471, "step": 4260 }, { "epoch": 0.15014491221579146, "grad_norm": 1.8421056270599365, "learning_rate": 1.9257841726858216e-05, "loss": 1.0195, "step": 4261 }, { "epoch": 0.15018014922874987, "grad_norm": 2.0977447032928467, "learning_rate": 1.9257410204059565e-05, "loss": 1.1225, "step": 4262 }, { "epoch": 0.1502153862417083, "grad_norm": 1.2575510740280151, "learning_rate": 1.9256978560681965e-05, "loss": 1.2773, "step": 4263 }, { "epoch": 0.1502506232546667, "grad_norm": 4.855735778808594, "learning_rate": 1.9256546796731042e-05, "loss": 4.1337, "step": 4264 }, { "epoch": 0.1502858602676251, "grad_norm": 0.9135820269584656, "learning_rate": 1.925611491221242e-05, "loss": 1.0596, "step": 4265 }, { "epoch": 0.15032109728058352, "grad_norm": 0.9323664903640747, "learning_rate": 1.9255682907131717e-05, "loss": 0.8267, "step": 4266 }, { "epoch": 0.15035633429354195, "grad_norm": 5.269653797149658, "learning_rate": 1.9255250781494567e-05, "loss": 6.6359, "step": 4267 }, { "epoch": 0.15039157130650035, "grad_norm": 1.4690672159194946, "learning_rate": 1.9254818535306598e-05, "loss": 1.0042, "step": 4268 }, { "epoch": 0.15042680831945876, "grad_norm": 1.609902262687683, "learning_rate": 1.925438616857344e-05, "loss": 1.1188, "step": 4269 }, { "epoch": 0.15046204533241717, "grad_norm": 1.5088826417922974, "learning_rate": 1.925395368130072e-05, "loss": 0.914, "step": 4270 }, { "epoch": 0.15049728234537557, "grad_norm": 2.08062744140625, "learning_rate": 1.9253521073494076e-05, "loss": 1.0464, "step": 4271 }, { "epoch": 0.150532519358334, "grad_norm": 4.096313953399658, "learning_rate": 1.925308834515914e-05, "loss": 3.829, "step": 4272 }, { "epoch": 0.1505677563712924, "grad_norm": 1.611903429031372, "learning_rate": 1.9252655496301554e-05, "loss": 0.9891, "step": 4273 }, { "epoch": 0.15060299338425082, "grad_norm": 1.5370802879333496, "learning_rate": 1.9252222526926947e-05, "loss": 0.9845, "step": 4274 }, { "epoch": 0.15063823039720922, "grad_norm": 1.4361145496368408, "learning_rate": 1.9251789437040965e-05, "loss": 1.124, "step": 4275 }, { "epoch": 0.15067346741016763, "grad_norm": 3.8406219482421875, "learning_rate": 1.925135622664925e-05, "loss": 2.4394, "step": 4276 }, { "epoch": 0.15070870442312606, "grad_norm": 1.572098970413208, "learning_rate": 1.9250922895757432e-05, "loss": 1.0438, "step": 4277 }, { "epoch": 0.15074394143608447, "grad_norm": 1.7641922235488892, "learning_rate": 1.9250489444371175e-05, "loss": 4.0043, "step": 4278 }, { "epoch": 0.15077917844904287, "grad_norm": 1.826680302619934, "learning_rate": 1.925005587249611e-05, "loss": 1.2398, "step": 4279 }, { "epoch": 0.15081441546200128, "grad_norm": 2.3350114822387695, "learning_rate": 1.9249622180137887e-05, "loss": 3.5321, "step": 4280 }, { "epoch": 0.1508496524749597, "grad_norm": 1.068885326385498, "learning_rate": 1.9249188367302162e-05, "loss": 1.1967, "step": 4281 }, { "epoch": 0.15088488948791812, "grad_norm": 5.5263285636901855, "learning_rate": 1.9248754433994578e-05, "loss": 3.5202, "step": 4282 }, { "epoch": 0.15092012650087652, "grad_norm": 6.69893741607666, "learning_rate": 1.924832038022079e-05, "loss": 5.9951, "step": 4283 }, { "epoch": 0.15095536351383493, "grad_norm": 1.4143275022506714, "learning_rate": 1.924788620598645e-05, "loss": 1.1095, "step": 4284 }, { "epoch": 0.15099060052679333, "grad_norm": 1.5110762119293213, "learning_rate": 1.9247451911297216e-05, "loss": 0.9705, "step": 4285 }, { "epoch": 0.15102583753975177, "grad_norm": 5.8120341300964355, "learning_rate": 1.924701749615874e-05, "loss": 2.9008, "step": 4286 }, { "epoch": 0.15106107455271017, "grad_norm": 4.8488030433654785, "learning_rate": 1.9246582960576688e-05, "loss": 6.7173, "step": 4287 }, { "epoch": 0.15109631156566858, "grad_norm": 1.325510025024414, "learning_rate": 1.924614830455671e-05, "loss": 1.1579, "step": 4288 }, { "epoch": 0.15113154857862698, "grad_norm": 1.2602133750915527, "learning_rate": 1.9245713528104473e-05, "loss": 1.002, "step": 4289 }, { "epoch": 0.1511667855915854, "grad_norm": 1.865443468093872, "learning_rate": 1.9245278631225642e-05, "loss": 0.8282, "step": 4290 }, { "epoch": 0.15120202260454382, "grad_norm": 1.036097764968872, "learning_rate": 1.924484361392588e-05, "loss": 1.1438, "step": 4291 }, { "epoch": 0.15123725961750223, "grad_norm": 9.957623481750488, "learning_rate": 1.924440847621085e-05, "loss": 4.0088, "step": 4292 }, { "epoch": 0.15127249663046063, "grad_norm": 5.110799789428711, "learning_rate": 1.9243973218086224e-05, "loss": 4.3912, "step": 4293 }, { "epoch": 0.15130773364341904, "grad_norm": 1.1105167865753174, "learning_rate": 1.9243537839557666e-05, "loss": 0.8343, "step": 4294 }, { "epoch": 0.15134297065637745, "grad_norm": 1.3186657428741455, "learning_rate": 1.924310234063085e-05, "loss": 0.9843, "step": 4295 }, { "epoch": 0.15137820766933588, "grad_norm": 2.1641693115234375, "learning_rate": 1.9242666721311453e-05, "loss": 2.9409, "step": 4296 }, { "epoch": 0.15141344468229428, "grad_norm": 5.132687091827393, "learning_rate": 1.924223098160514e-05, "loss": 3.3659, "step": 4297 }, { "epoch": 0.1514486816952527, "grad_norm": 5.848074913024902, "learning_rate": 1.9241795121517595e-05, "loss": 3.2403, "step": 4298 }, { "epoch": 0.1514839187082111, "grad_norm": 5.547036647796631, "learning_rate": 1.924135914105449e-05, "loss": 6.6304, "step": 4299 }, { "epoch": 0.15151915572116953, "grad_norm": 1.2188438177108765, "learning_rate": 1.9240923040221505e-05, "loss": 1.0187, "step": 4300 }, { "epoch": 0.15155439273412794, "grad_norm": 4.063498020172119, "learning_rate": 1.9240486819024324e-05, "loss": 3.4661, "step": 4301 }, { "epoch": 0.15158962974708634, "grad_norm": 5.946369647979736, "learning_rate": 1.924005047746862e-05, "loss": 3.7438, "step": 4302 }, { "epoch": 0.15162486676004475, "grad_norm": 1.3899524211883545, "learning_rate": 1.9239614015560083e-05, "loss": 1.2314, "step": 4303 }, { "epoch": 0.15166010377300315, "grad_norm": 1.1711790561676025, "learning_rate": 1.9239177433304394e-05, "loss": 1.1693, "step": 4304 }, { "epoch": 0.15169534078596159, "grad_norm": 3.2267775535583496, "learning_rate": 1.9238740730707245e-05, "loss": 3.375, "step": 4305 }, { "epoch": 0.15173057779892, "grad_norm": 1.1017593145370483, "learning_rate": 1.923830390777432e-05, "loss": 1.3081, "step": 4306 }, { "epoch": 0.1517658148118784, "grad_norm": 7.463367938995361, "learning_rate": 1.9237866964511312e-05, "loss": 6.7393, "step": 4307 }, { "epoch": 0.1518010518248368, "grad_norm": 10.725565910339355, "learning_rate": 1.9237429900923907e-05, "loss": 8.394, "step": 4308 }, { "epoch": 0.1518362888377952, "grad_norm": 1.3119516372680664, "learning_rate": 1.9236992717017805e-05, "loss": 0.9795, "step": 4309 }, { "epoch": 0.15187152585075364, "grad_norm": 4.8549089431762695, "learning_rate": 1.9236555412798693e-05, "loss": 3.9426, "step": 4310 }, { "epoch": 0.15190676286371205, "grad_norm": 5.641083717346191, "learning_rate": 1.9236117988272273e-05, "loss": 3.6987, "step": 4311 }, { "epoch": 0.15194199987667045, "grad_norm": 4.260662078857422, "learning_rate": 1.9235680443444236e-05, "loss": 5.8337, "step": 4312 }, { "epoch": 0.15197723688962886, "grad_norm": 1.430526614189148, "learning_rate": 1.9235242778320286e-05, "loss": 0.9208, "step": 4313 }, { "epoch": 0.15201247390258726, "grad_norm": 4.474459648132324, "learning_rate": 1.923480499290612e-05, "loss": 5.4908, "step": 4314 }, { "epoch": 0.1520477109155457, "grad_norm": 1.2229245901107788, "learning_rate": 1.923436708720745e-05, "loss": 1.023, "step": 4315 }, { "epoch": 0.1520829479285041, "grad_norm": 9.306953430175781, "learning_rate": 1.9233929061229967e-05, "loss": 4.0228, "step": 4316 }, { "epoch": 0.1521181849414625, "grad_norm": 7.033900260925293, "learning_rate": 1.9233490914979383e-05, "loss": 6.5586, "step": 4317 }, { "epoch": 0.15215342195442091, "grad_norm": 3.6274216175079346, "learning_rate": 1.9233052648461403e-05, "loss": 6.382, "step": 4318 }, { "epoch": 0.15218865896737935, "grad_norm": 2.201057195663452, "learning_rate": 1.9232614261681733e-05, "loss": 4.0143, "step": 4319 }, { "epoch": 0.15222389598033775, "grad_norm": 8.389933586120605, "learning_rate": 1.923217575464609e-05, "loss": 11.4992, "step": 4320 }, { "epoch": 0.15225913299329616, "grad_norm": 1.0766228437423706, "learning_rate": 1.9231737127360184e-05, "loss": 1.2246, "step": 4321 }, { "epoch": 0.15229437000625456, "grad_norm": 8.047061920166016, "learning_rate": 1.9231298379829724e-05, "loss": 4.5556, "step": 4322 }, { "epoch": 0.15232960701921297, "grad_norm": 7.461942672729492, "learning_rate": 1.9230859512060426e-05, "loss": 3.6884, "step": 4323 }, { "epoch": 0.1523648440321714, "grad_norm": 1.4407579898834229, "learning_rate": 1.9230420524058007e-05, "loss": 0.8348, "step": 4324 }, { "epoch": 0.1524000810451298, "grad_norm": 1.3392711877822876, "learning_rate": 1.9229981415828184e-05, "loss": 1.0224, "step": 4325 }, { "epoch": 0.15243531805808821, "grad_norm": 1.1660261154174805, "learning_rate": 1.922954218737668e-05, "loss": 1.1148, "step": 4326 }, { "epoch": 0.15247055507104662, "grad_norm": 4.884274959564209, "learning_rate": 1.922910283870921e-05, "loss": 6.7484, "step": 4327 }, { "epoch": 0.15250579208400503, "grad_norm": 1.271108865737915, "learning_rate": 1.9228663369831505e-05, "loss": 0.9307, "step": 4328 }, { "epoch": 0.15254102909696346, "grad_norm": 1.4893285036087036, "learning_rate": 1.922822378074928e-05, "loss": 1.074, "step": 4329 }, { "epoch": 0.15257626610992187, "grad_norm": 3.257936954498291, "learning_rate": 1.9227784071468265e-05, "loss": 3.5721, "step": 4330 }, { "epoch": 0.15261150312288027, "grad_norm": 2.7076618671417236, "learning_rate": 1.9227344241994184e-05, "loss": 1.042, "step": 4331 }, { "epoch": 0.15264674013583868, "grad_norm": 1.4144138097763062, "learning_rate": 1.9226904292332776e-05, "loss": 0.9782, "step": 4332 }, { "epoch": 0.1526819771487971, "grad_norm": 1.9072329998016357, "learning_rate": 1.922646422248976e-05, "loss": 4.017, "step": 4333 }, { "epoch": 0.15271721416175552, "grad_norm": 2.61309814453125, "learning_rate": 1.9226024032470873e-05, "loss": 3.2015, "step": 4334 }, { "epoch": 0.15275245117471392, "grad_norm": 3.095940113067627, "learning_rate": 1.9225583722281848e-05, "loss": 3.6442, "step": 4335 }, { "epoch": 0.15278768818767233, "grad_norm": 2.2306294441223145, "learning_rate": 1.9225143291928422e-05, "loss": 3.6714, "step": 4336 }, { "epoch": 0.15282292520063073, "grad_norm": 2.24174165725708, "learning_rate": 1.9224702741416322e-05, "loss": 1.1428, "step": 4337 }, { "epoch": 0.15285816221358917, "grad_norm": 2.5891664028167725, "learning_rate": 1.9224262070751304e-05, "loss": 3.67, "step": 4338 }, { "epoch": 0.15289339922654757, "grad_norm": 4.964269161224365, "learning_rate": 1.922382127993909e-05, "loss": 3.9202, "step": 4339 }, { "epoch": 0.15292863623950598, "grad_norm": 4.8138813972473145, "learning_rate": 1.9223380368985433e-05, "loss": 6.3348, "step": 4340 }, { "epoch": 0.15296387325246438, "grad_norm": 2.6155009269714355, "learning_rate": 1.9222939337896068e-05, "loss": 3.2983, "step": 4341 }, { "epoch": 0.1529991102654228, "grad_norm": 4.43307638168335, "learning_rate": 1.9222498186676746e-05, "loss": 7.0702, "step": 4342 }, { "epoch": 0.15303434727838122, "grad_norm": 1.6387804746627808, "learning_rate": 1.9222056915333208e-05, "loss": 0.9785, "step": 4343 }, { "epoch": 0.15306958429133963, "grad_norm": 4.650184154510498, "learning_rate": 1.9221615523871206e-05, "loss": 3.0971, "step": 4344 }, { "epoch": 0.15310482130429803, "grad_norm": 1.2628145217895508, "learning_rate": 1.922117401229649e-05, "loss": 1.2833, "step": 4345 }, { "epoch": 0.15314005831725644, "grad_norm": 1.2222825288772583, "learning_rate": 1.9220732380614805e-05, "loss": 0.9658, "step": 4346 }, { "epoch": 0.15317529533021484, "grad_norm": 1.4185428619384766, "learning_rate": 1.9220290628831906e-05, "loss": 1.2094, "step": 4347 }, { "epoch": 0.15321053234317328, "grad_norm": 1.9935758113861084, "learning_rate": 1.9219848756953548e-05, "loss": 0.9671, "step": 4348 }, { "epoch": 0.15324576935613168, "grad_norm": 2.676435708999634, "learning_rate": 1.921940676498548e-05, "loss": 3.4763, "step": 4349 }, { "epoch": 0.1532810063690901, "grad_norm": 1.1626461744308472, "learning_rate": 1.9218964652933472e-05, "loss": 1.0277, "step": 4350 }, { "epoch": 0.1533162433820485, "grad_norm": 3.4221079349517822, "learning_rate": 1.9218522420803274e-05, "loss": 3.361, "step": 4351 }, { "epoch": 0.15335148039500693, "grad_norm": 1.268635869026184, "learning_rate": 1.9218080068600643e-05, "loss": 0.7804, "step": 4352 }, { "epoch": 0.15338671740796533, "grad_norm": 1.2874195575714111, "learning_rate": 1.921763759633135e-05, "loss": 1.1473, "step": 4353 }, { "epoch": 0.15342195442092374, "grad_norm": 1.3718430995941162, "learning_rate": 1.921719500400115e-05, "loss": 1.203, "step": 4354 }, { "epoch": 0.15345719143388215, "grad_norm": 1.8819094896316528, "learning_rate": 1.921675229161581e-05, "loss": 1.0018, "step": 4355 }, { "epoch": 0.15349242844684055, "grad_norm": 3.96433687210083, "learning_rate": 1.92163094591811e-05, "loss": 4.122, "step": 4356 }, { "epoch": 0.15352766545979898, "grad_norm": 4.351819038391113, "learning_rate": 1.9215866506702784e-05, "loss": 3.6902, "step": 4357 }, { "epoch": 0.1535629024727574, "grad_norm": 4.374841213226318, "learning_rate": 1.9215423434186634e-05, "loss": 5.931, "step": 4358 }, { "epoch": 0.1535981394857158, "grad_norm": 1.3386284112930298, "learning_rate": 1.921498024163842e-05, "loss": 0.8712, "step": 4359 }, { "epoch": 0.1536333764986742, "grad_norm": 8.801876068115234, "learning_rate": 1.9214536929063913e-05, "loss": 3.8827, "step": 4360 }, { "epoch": 0.1536686135116326, "grad_norm": 4.993210792541504, "learning_rate": 1.9214093496468887e-05, "loss": 3.8775, "step": 4361 }, { "epoch": 0.15370385052459104, "grad_norm": 3.0734169483184814, "learning_rate": 1.9213649943859122e-05, "loss": 3.4401, "step": 4362 }, { "epoch": 0.15373908753754945, "grad_norm": 1.6743816137313843, "learning_rate": 1.921320627124039e-05, "loss": 1.0634, "step": 4363 }, { "epoch": 0.15377432455050785, "grad_norm": 1.6945666074752808, "learning_rate": 1.9212762478618477e-05, "loss": 1.0186, "step": 4364 }, { "epoch": 0.15380956156346626, "grad_norm": 3.630150079727173, "learning_rate": 1.921231856599916e-05, "loss": 3.6934, "step": 4365 }, { "epoch": 0.15384479857642466, "grad_norm": 4.172453880310059, "learning_rate": 1.9211874533388218e-05, "loss": 4.0781, "step": 4366 }, { "epoch": 0.1538800355893831, "grad_norm": 1.3502650260925293, "learning_rate": 1.9211430380791436e-05, "loss": 0.9516, "step": 4367 }, { "epoch": 0.1539152726023415, "grad_norm": 3.7834811210632324, "learning_rate": 1.92109861082146e-05, "loss": 3.9653, "step": 4368 }, { "epoch": 0.1539505096152999, "grad_norm": 1.733062982559204, "learning_rate": 1.9210541715663498e-05, "loss": 0.9655, "step": 4369 }, { "epoch": 0.1539857466282583, "grad_norm": 2.842897653579712, "learning_rate": 1.9210097203143915e-05, "loss": 3.8935, "step": 4370 }, { "epoch": 0.15402098364121675, "grad_norm": 4.228041172027588, "learning_rate": 1.9209652570661643e-05, "loss": 6.0354, "step": 4371 }, { "epoch": 0.15405622065417515, "grad_norm": 5.306982517242432, "learning_rate": 1.9209207818222478e-05, "loss": 6.9593, "step": 4372 }, { "epoch": 0.15409145766713356, "grad_norm": 1.3988497257232666, "learning_rate": 1.9208762945832203e-05, "loss": 1.0282, "step": 4373 }, { "epoch": 0.15412669468009196, "grad_norm": 4.96733283996582, "learning_rate": 1.9208317953496617e-05, "loss": 3.8799, "step": 4374 }, { "epoch": 0.15416193169305037, "grad_norm": 5.439053535461426, "learning_rate": 1.9207872841221518e-05, "loss": 3.943, "step": 4375 }, { "epoch": 0.1541971687060088, "grad_norm": 1.4004848003387451, "learning_rate": 1.9207427609012704e-05, "loss": 0.8228, "step": 4376 }, { "epoch": 0.1542324057189672, "grad_norm": 3.7312989234924316, "learning_rate": 1.9206982256875974e-05, "loss": 3.4108, "step": 4377 }, { "epoch": 0.1542676427319256, "grad_norm": 1.5080183744430542, "learning_rate": 1.9206536784817122e-05, "loss": 0.8885, "step": 4378 }, { "epoch": 0.15430287974488402, "grad_norm": 4.139060020446777, "learning_rate": 1.920609119284196e-05, "loss": 6.7073, "step": 4379 }, { "epoch": 0.15433811675784243, "grad_norm": 1.3347569704055786, "learning_rate": 1.9205645480956283e-05, "loss": 1.0728, "step": 4380 }, { "epoch": 0.15437335377080086, "grad_norm": 1.3329432010650635, "learning_rate": 1.9205199649165903e-05, "loss": 0.7943, "step": 4381 }, { "epoch": 0.15440859078375926, "grad_norm": 1.669750452041626, "learning_rate": 1.9204753697476627e-05, "loss": 0.9134, "step": 4382 }, { "epoch": 0.15444382779671767, "grad_norm": 6.75323486328125, "learning_rate": 1.9204307625894262e-05, "loss": 3.9126, "step": 4383 }, { "epoch": 0.15447906480967608, "grad_norm": 1.3482972383499146, "learning_rate": 1.920386143442462e-05, "loss": 1.0228, "step": 4384 }, { "epoch": 0.1545143018226345, "grad_norm": 3.0188047885894775, "learning_rate": 1.9203415123073502e-05, "loss": 3.2942, "step": 4385 }, { "epoch": 0.15454953883559291, "grad_norm": 1.5520793199539185, "learning_rate": 1.9202968691846735e-05, "loss": 0.9826, "step": 4386 }, { "epoch": 0.15458477584855132, "grad_norm": 3.0473482608795166, "learning_rate": 1.9202522140750128e-05, "loss": 3.2825, "step": 4387 }, { "epoch": 0.15462001286150973, "grad_norm": 3.9041876792907715, "learning_rate": 1.92020754697895e-05, "loss": 2.9188, "step": 4388 }, { "epoch": 0.15465524987446813, "grad_norm": 1.0535176992416382, "learning_rate": 1.9201628678970663e-05, "loss": 1.1806, "step": 4389 }, { "epoch": 0.15469048688742656, "grad_norm": 5.723671913146973, "learning_rate": 1.9201181768299442e-05, "loss": 6.8275, "step": 4390 }, { "epoch": 0.15472572390038497, "grad_norm": 1.0576320886611938, "learning_rate": 1.9200734737781655e-05, "loss": 1.0398, "step": 4391 }, { "epoch": 0.15476096091334338, "grad_norm": 7.800765514373779, "learning_rate": 1.920028758742313e-05, "loss": 3.6426, "step": 4392 }, { "epoch": 0.15479619792630178, "grad_norm": 1.3620020151138306, "learning_rate": 1.9199840317229683e-05, "loss": 1.0664, "step": 4393 }, { "epoch": 0.1548314349392602, "grad_norm": 7.195034027099609, "learning_rate": 1.9199392927207146e-05, "loss": 2.8069, "step": 4394 }, { "epoch": 0.15486667195221862, "grad_norm": 2.881357431411743, "learning_rate": 1.9198945417361346e-05, "loss": 3.4227, "step": 4395 }, { "epoch": 0.15490190896517703, "grad_norm": 2.3536572456359863, "learning_rate": 1.9198497787698107e-05, "loss": 4.1582, "step": 4396 }, { "epoch": 0.15493714597813543, "grad_norm": 4.304935932159424, "learning_rate": 1.9198050038223265e-05, "loss": 4.1078, "step": 4397 }, { "epoch": 0.15497238299109384, "grad_norm": 1.3605365753173828, "learning_rate": 1.9197602168942645e-05, "loss": 0.9542, "step": 4398 }, { "epoch": 0.15500762000405224, "grad_norm": 1.7394917011260986, "learning_rate": 1.919715417986209e-05, "loss": 1.0863, "step": 4399 }, { "epoch": 0.15504285701701068, "grad_norm": 4.775567531585693, "learning_rate": 1.919670607098743e-05, "loss": 5.2439, "step": 4400 }, { "epoch": 0.15507809402996908, "grad_norm": 9.74548625946045, "learning_rate": 1.9196257842324498e-05, "loss": 5.392, "step": 4401 }, { "epoch": 0.1551133310429275, "grad_norm": 1.9302839040756226, "learning_rate": 1.919580949387914e-05, "loss": 1.0757, "step": 4402 }, { "epoch": 0.1551485680558859, "grad_norm": 3.8278236389160156, "learning_rate": 1.919536102565719e-05, "loss": 2.8367, "step": 4403 }, { "epoch": 0.15518380506884433, "grad_norm": 1.8700964450836182, "learning_rate": 1.9194912437664493e-05, "loss": 1.0505, "step": 4404 }, { "epoch": 0.15521904208180273, "grad_norm": 5.942344665527344, "learning_rate": 1.919446372990689e-05, "loss": 3.8707, "step": 4405 }, { "epoch": 0.15525427909476114, "grad_norm": 1.4997315406799316, "learning_rate": 1.9194014902390224e-05, "loss": 1.2285, "step": 4406 }, { "epoch": 0.15528951610771954, "grad_norm": 1.466163992881775, "learning_rate": 1.9193565955120345e-05, "loss": 0.913, "step": 4407 }, { "epoch": 0.15532475312067795, "grad_norm": 1.0559203624725342, "learning_rate": 1.91931168881031e-05, "loss": 0.9732, "step": 4408 }, { "epoch": 0.15535999013363638, "grad_norm": 1.1320654153823853, "learning_rate": 1.919266770134433e-05, "loss": 1.1778, "step": 4409 }, { "epoch": 0.1553952271465948, "grad_norm": 1.040732502937317, "learning_rate": 1.91922183948499e-05, "loss": 0.9549, "step": 4410 }, { "epoch": 0.1554304641595532, "grad_norm": 4.91375207901001, "learning_rate": 1.919176896862565e-05, "loss": 3.9866, "step": 4411 }, { "epoch": 0.1554657011725116, "grad_norm": 1.2286421060562134, "learning_rate": 1.9191319422677436e-05, "loss": 1.2028, "step": 4412 }, { "epoch": 0.15550093818547, "grad_norm": 1.1082165241241455, "learning_rate": 1.919086975701112e-05, "loss": 1.0666, "step": 4413 }, { "epoch": 0.15553617519842844, "grad_norm": 1.3309210538864136, "learning_rate": 1.9190419971632555e-05, "loss": 1.0091, "step": 4414 }, { "epoch": 0.15557141221138684, "grad_norm": 6.743384838104248, "learning_rate": 1.9189970066547596e-05, "loss": 6.1461, "step": 4415 }, { "epoch": 0.15560664922434525, "grad_norm": 1.237170934677124, "learning_rate": 1.918952004176211e-05, "loss": 1.1885, "step": 4416 }, { "epoch": 0.15564188623730366, "grad_norm": 5.43143892288208, "learning_rate": 1.918906989728195e-05, "loss": 3.32, "step": 4417 }, { "epoch": 0.15567712325026206, "grad_norm": 1.854253888130188, "learning_rate": 1.9188619633112987e-05, "loss": 1.0485, "step": 4418 }, { "epoch": 0.1557123602632205, "grad_norm": 2.168031930923462, "learning_rate": 1.9188169249261084e-05, "loss": 1.0329, "step": 4419 }, { "epoch": 0.1557475972761789, "grad_norm": 1.5131680965423584, "learning_rate": 1.9187718745732105e-05, "loss": 1.087, "step": 4420 }, { "epoch": 0.1557828342891373, "grad_norm": 3.8708536624908447, "learning_rate": 1.918726812253192e-05, "loss": 2.8467, "step": 4421 }, { "epoch": 0.1558180713020957, "grad_norm": 7.408769130706787, "learning_rate": 1.9186817379666395e-05, "loss": 6.8524, "step": 4422 }, { "epoch": 0.15585330831505415, "grad_norm": 1.437188982963562, "learning_rate": 1.9186366517141404e-05, "loss": 1.258, "step": 4423 }, { "epoch": 0.15588854532801255, "grad_norm": 5.913534641265869, "learning_rate": 1.918591553496282e-05, "loss": 4.5026, "step": 4424 }, { "epoch": 0.15592378234097096, "grad_norm": 1.3757740259170532, "learning_rate": 1.9185464433136514e-05, "loss": 0.8768, "step": 4425 }, { "epoch": 0.15595901935392936, "grad_norm": 5.39621114730835, "learning_rate": 1.918501321166837e-05, "loss": 3.5245, "step": 4426 }, { "epoch": 0.15599425636688777, "grad_norm": 2.8932948112487793, "learning_rate": 1.918456187056425e-05, "loss": 2.8586, "step": 4427 }, { "epoch": 0.1560294933798462, "grad_norm": 3.2649848461151123, "learning_rate": 1.9184110409830046e-05, "loss": 3.3952, "step": 4428 }, { "epoch": 0.1560647303928046, "grad_norm": 6.25156831741333, "learning_rate": 1.9183658829471636e-05, "loss": 5.6259, "step": 4429 }, { "epoch": 0.156099967405763, "grad_norm": 1.3479769229888916, "learning_rate": 1.9183207129494897e-05, "loss": 1.0526, "step": 4430 }, { "epoch": 0.15613520441872142, "grad_norm": 1.299125075340271, "learning_rate": 1.9182755309905715e-05, "loss": 0.8943, "step": 4431 }, { "epoch": 0.15617044143167982, "grad_norm": 2.2548155784606934, "learning_rate": 1.918230337070998e-05, "loss": 3.4905, "step": 4432 }, { "epoch": 0.15620567844463826, "grad_norm": 2.4663186073303223, "learning_rate": 1.918185131191357e-05, "loss": 2.8265, "step": 4433 }, { "epoch": 0.15624091545759666, "grad_norm": 4.844275951385498, "learning_rate": 1.9181399133522377e-05, "loss": 7.1545, "step": 4434 }, { "epoch": 0.15627615247055507, "grad_norm": 1.0407652854919434, "learning_rate": 1.9180946835542292e-05, "loss": 1.2362, "step": 4435 }, { "epoch": 0.15631138948351347, "grad_norm": 1.7369176149368286, "learning_rate": 1.9180494417979206e-05, "loss": 1.2002, "step": 4436 }, { "epoch": 0.1563466264964719, "grad_norm": 2.4117517471313477, "learning_rate": 1.918004188083901e-05, "loss": 3.1118, "step": 4437 }, { "epoch": 0.1563818635094303, "grad_norm": 4.013784408569336, "learning_rate": 1.9179589224127596e-05, "loss": 6.5834, "step": 4438 }, { "epoch": 0.15641710052238872, "grad_norm": 1.1516488790512085, "learning_rate": 1.917913644785087e-05, "loss": 0.9503, "step": 4439 }, { "epoch": 0.15645233753534712, "grad_norm": 1.605370044708252, "learning_rate": 1.9178683552014712e-05, "loss": 0.7777, "step": 4440 }, { "epoch": 0.15648757454830553, "grad_norm": 4.851962089538574, "learning_rate": 1.9178230536625043e-05, "loss": 3.415, "step": 4441 }, { "epoch": 0.15652281156126396, "grad_norm": 2.331603765487671, "learning_rate": 1.9177777401687745e-05, "loss": 3.1219, "step": 4442 }, { "epoch": 0.15655804857422237, "grad_norm": 1.1823269128799438, "learning_rate": 1.9177324147208728e-05, "loss": 1.0729, "step": 4443 }, { "epoch": 0.15659328558718077, "grad_norm": 1.8633195161819458, "learning_rate": 1.9176870773193898e-05, "loss": 1.0692, "step": 4444 }, { "epoch": 0.15662852260013918, "grad_norm": 5.456241130828857, "learning_rate": 1.9176417279649152e-05, "loss": 5.4949, "step": 4445 }, { "epoch": 0.1566637596130976, "grad_norm": 4.814098834991455, "learning_rate": 1.9175963666580407e-05, "loss": 6.3062, "step": 4446 }, { "epoch": 0.15669899662605602, "grad_norm": 1.629866600036621, "learning_rate": 1.9175509933993565e-05, "loss": 1.1477, "step": 4447 }, { "epoch": 0.15673423363901443, "grad_norm": 1.2427129745483398, "learning_rate": 1.917505608189454e-05, "loss": 1.0591, "step": 4448 }, { "epoch": 0.15676947065197283, "grad_norm": 0.9566769599914551, "learning_rate": 1.9174602110289236e-05, "loss": 0.9045, "step": 4449 }, { "epoch": 0.15680470766493124, "grad_norm": 1.0160080194473267, "learning_rate": 1.9174148019183573e-05, "loss": 0.8515, "step": 4450 }, { "epoch": 0.15683994467788964, "grad_norm": 1.1013648509979248, "learning_rate": 1.9173693808583462e-05, "loss": 1.3011, "step": 4451 }, { "epoch": 0.15687518169084808, "grad_norm": 10.313040733337402, "learning_rate": 1.9173239478494822e-05, "loss": 6.6635, "step": 4452 }, { "epoch": 0.15691041870380648, "grad_norm": 3.3328778743743896, "learning_rate": 1.917278502892357e-05, "loss": 3.735, "step": 4453 }, { "epoch": 0.1569456557167649, "grad_norm": 3.9264414310455322, "learning_rate": 1.9172330459875622e-05, "loss": 6.8892, "step": 4454 }, { "epoch": 0.1569808927297233, "grad_norm": 1.8136636018753052, "learning_rate": 1.9171875771356906e-05, "loss": 1.1452, "step": 4455 }, { "epoch": 0.15701612974268173, "grad_norm": 4.268340587615967, "learning_rate": 1.9171420963373335e-05, "loss": 4.2081, "step": 4456 }, { "epoch": 0.15705136675564013, "grad_norm": 2.2157034873962402, "learning_rate": 1.917096603593084e-05, "loss": 3.0175, "step": 4457 }, { "epoch": 0.15708660376859854, "grad_norm": 1.6140072345733643, "learning_rate": 1.9170510989035344e-05, "loss": 1.083, "step": 4458 }, { "epoch": 0.15712184078155694, "grad_norm": 1.050006628036499, "learning_rate": 1.9170055822692776e-05, "loss": 0.8541, "step": 4459 }, { "epoch": 0.15715707779451535, "grad_norm": 2.0593435764312744, "learning_rate": 1.916960053690906e-05, "loss": 1.0903, "step": 4460 }, { "epoch": 0.15719231480747378, "grad_norm": 1.3610464334487915, "learning_rate": 1.916914513169013e-05, "loss": 1.0526, "step": 4461 }, { "epoch": 0.1572275518204322, "grad_norm": 1.3367799520492554, "learning_rate": 1.9168689607041915e-05, "loss": 1.2153, "step": 4462 }, { "epoch": 0.1572627888333906, "grad_norm": 2.017934560775757, "learning_rate": 1.9168233962970352e-05, "loss": 1.106, "step": 4463 }, { "epoch": 0.157298025846349, "grad_norm": 1.3664995431900024, "learning_rate": 1.916777819948137e-05, "loss": 0.9435, "step": 4464 }, { "epoch": 0.1573332628593074, "grad_norm": 1.28180730342865, "learning_rate": 1.9167322316580918e-05, "loss": 1.1397, "step": 4465 }, { "epoch": 0.15736849987226584, "grad_norm": 2.8659844398498535, "learning_rate": 1.9166866314274917e-05, "loss": 3.5909, "step": 4466 }, { "epoch": 0.15740373688522424, "grad_norm": 3.205780029296875, "learning_rate": 1.9166410192569316e-05, "loss": 3.7936, "step": 4467 }, { "epoch": 0.15743897389818265, "grad_norm": 4.505309104919434, "learning_rate": 1.9165953951470057e-05, "loss": 3.7316, "step": 4468 }, { "epoch": 0.15747421091114105, "grad_norm": 3.727703332901001, "learning_rate": 1.9165497590983076e-05, "loss": 5.9044, "step": 4469 }, { "epoch": 0.15750944792409946, "grad_norm": 5.1362385749816895, "learning_rate": 1.916504111111433e-05, "loss": 6.536, "step": 4470 }, { "epoch": 0.1575446849370579, "grad_norm": 1.6602778434753418, "learning_rate": 1.9164584511869748e-05, "loss": 0.8785, "step": 4471 }, { "epoch": 0.1575799219500163, "grad_norm": 1.2355942726135254, "learning_rate": 1.916412779325529e-05, "loss": 1.2772, "step": 4472 }, { "epoch": 0.1576151589629747, "grad_norm": 1.284651279449463, "learning_rate": 1.9163670955276896e-05, "loss": 0.9481, "step": 4473 }, { "epoch": 0.1576503959759331, "grad_norm": 2.6173462867736816, "learning_rate": 1.9163213997940522e-05, "loss": 0.9858, "step": 4474 }, { "epoch": 0.15768563298889154, "grad_norm": 1.3235485553741455, "learning_rate": 1.916275692125212e-05, "loss": 1.0726, "step": 4475 }, { "epoch": 0.15772087000184995, "grad_norm": 2.636775016784668, "learning_rate": 1.916229972521764e-05, "loss": 3.9643, "step": 4476 }, { "epoch": 0.15775610701480836, "grad_norm": 6.697117805480957, "learning_rate": 1.9161842409843043e-05, "loss": 3.2766, "step": 4477 }, { "epoch": 0.15779134402776676, "grad_norm": 4.675210952758789, "learning_rate": 1.9161384975134275e-05, "loss": 3.6823, "step": 4478 }, { "epoch": 0.15782658104072517, "grad_norm": 1.3691493272781372, "learning_rate": 1.9160927421097306e-05, "loss": 1.0768, "step": 4479 }, { "epoch": 0.1578618180536836, "grad_norm": 4.786154747009277, "learning_rate": 1.9160469747738086e-05, "loss": 3.6856, "step": 4480 }, { "epoch": 0.157897055066642, "grad_norm": 4.164218425750732, "learning_rate": 1.9160011955062584e-05, "loss": 2.784, "step": 4481 }, { "epoch": 0.1579322920796004, "grad_norm": 1.4379631280899048, "learning_rate": 1.9159554043076758e-05, "loss": 0.9015, "step": 4482 }, { "epoch": 0.15796752909255882, "grad_norm": 1.2958436012268066, "learning_rate": 1.9159096011786575e-05, "loss": 1.0178, "step": 4483 }, { "epoch": 0.15800276610551722, "grad_norm": 4.239436626434326, "learning_rate": 1.9158637861198e-05, "loss": 3.0395, "step": 4484 }, { "epoch": 0.15803800311847566, "grad_norm": 1.4491862058639526, "learning_rate": 1.9158179591317002e-05, "loss": 1.0531, "step": 4485 }, { "epoch": 0.15807324013143406, "grad_norm": 7.374385833740234, "learning_rate": 1.9157721202149545e-05, "loss": 6.9771, "step": 4486 }, { "epoch": 0.15810847714439247, "grad_norm": 3.411649465560913, "learning_rate": 1.9157262693701603e-05, "loss": 3.838, "step": 4487 }, { "epoch": 0.15814371415735087, "grad_norm": 0.9675024151802063, "learning_rate": 1.9156804065979147e-05, "loss": 0.9168, "step": 4488 }, { "epoch": 0.1581789511703093, "grad_norm": 5.788609027862549, "learning_rate": 1.9156345318988157e-05, "loss": 6.0699, "step": 4489 }, { "epoch": 0.1582141881832677, "grad_norm": 3.8599231243133545, "learning_rate": 1.9155886452734596e-05, "loss": 3.0753, "step": 4490 }, { "epoch": 0.15824942519622612, "grad_norm": 1.4250082969665527, "learning_rate": 1.915542746722445e-05, "loss": 1.1482, "step": 4491 }, { "epoch": 0.15828466220918452, "grad_norm": 2.937401056289673, "learning_rate": 1.91549683624637e-05, "loss": 3.4973, "step": 4492 }, { "epoch": 0.15831989922214293, "grad_norm": 1.5407435894012451, "learning_rate": 1.9154509138458315e-05, "loss": 0.8932, "step": 4493 }, { "epoch": 0.15835513623510136, "grad_norm": 1.1984946727752686, "learning_rate": 1.9154049795214284e-05, "loss": 0.9427, "step": 4494 }, { "epoch": 0.15839037324805977, "grad_norm": 1.102726697921753, "learning_rate": 1.915359033273759e-05, "loss": 1.1638, "step": 4495 }, { "epoch": 0.15842561026101817, "grad_norm": 3.652987003326416, "learning_rate": 1.915313075103421e-05, "loss": 3.8623, "step": 4496 }, { "epoch": 0.15846084727397658, "grad_norm": 5.11927604675293, "learning_rate": 1.915267105011014e-05, "loss": 5.5961, "step": 4497 }, { "epoch": 0.15849608428693499, "grad_norm": 1.347027063369751, "learning_rate": 1.915221122997137e-05, "loss": 0.9983, "step": 4498 }, { "epoch": 0.15853132129989342, "grad_norm": 2.6907811164855957, "learning_rate": 1.9151751290623876e-05, "loss": 3.1812, "step": 4499 }, { "epoch": 0.15856655831285182, "grad_norm": 4.003100872039795, "learning_rate": 1.9151291232073657e-05, "loss": 4.2478, "step": 4500 }, { "epoch": 0.15860179532581023, "grad_norm": 1.3174288272857666, "learning_rate": 1.9150831054326706e-05, "loss": 1.1571, "step": 4501 }, { "epoch": 0.15863703233876864, "grad_norm": 3.319258689880371, "learning_rate": 1.915037075738901e-05, "loss": 3.9604, "step": 4502 }, { "epoch": 0.15867226935172704, "grad_norm": 1.3786848783493042, "learning_rate": 1.9149910341266574e-05, "loss": 0.9928, "step": 4503 }, { "epoch": 0.15870750636468547, "grad_norm": 1.0229301452636719, "learning_rate": 1.9149449805965386e-05, "loss": 1.2492, "step": 4504 }, { "epoch": 0.15874274337764388, "grad_norm": 4.683528423309326, "learning_rate": 1.9148989151491452e-05, "loss": 3.4618, "step": 4505 }, { "epoch": 0.15877798039060229, "grad_norm": 1.4881502389907837, "learning_rate": 1.9148528377850767e-05, "loss": 1.0091, "step": 4506 }, { "epoch": 0.1588132174035607, "grad_norm": 2.496525287628174, "learning_rate": 1.9148067485049338e-05, "loss": 3.748, "step": 4507 }, { "epoch": 0.15884845441651912, "grad_norm": 7.282220840454102, "learning_rate": 1.9147606473093162e-05, "loss": 4.044, "step": 4508 }, { "epoch": 0.15888369142947753, "grad_norm": 1.200628399848938, "learning_rate": 1.9147145341988246e-05, "loss": 0.8512, "step": 4509 }, { "epoch": 0.15891892844243594, "grad_norm": 5.2737603187561035, "learning_rate": 1.91466840917406e-05, "loss": 3.5767, "step": 4510 }, { "epoch": 0.15895416545539434, "grad_norm": 1.1256252527236938, "learning_rate": 1.9146222722356226e-05, "loss": 1.1343, "step": 4511 }, { "epoch": 0.15898940246835275, "grad_norm": 1.3514889478683472, "learning_rate": 1.9145761233841137e-05, "loss": 1.0944, "step": 4512 }, { "epoch": 0.15902463948131118, "grad_norm": 2.2792978286743164, "learning_rate": 1.914529962620134e-05, "loss": 3.9049, "step": 4513 }, { "epoch": 0.1590598764942696, "grad_norm": 1.6143590211868286, "learning_rate": 1.9144837899442855e-05, "loss": 0.941, "step": 4514 }, { "epoch": 0.159095113507228, "grad_norm": 2.8461532592773438, "learning_rate": 1.9144376053571687e-05, "loss": 3.0422, "step": 4515 }, { "epoch": 0.1591303505201864, "grad_norm": 1.2993537187576294, "learning_rate": 1.914391408859386e-05, "loss": 1.2276, "step": 4516 }, { "epoch": 0.1591655875331448, "grad_norm": 6.124049663543701, "learning_rate": 1.9143452004515383e-05, "loss": 5.9413, "step": 4517 }, { "epoch": 0.15920082454610324, "grad_norm": 3.6114296913146973, "learning_rate": 1.9142989801342282e-05, "loss": 4.0745, "step": 4518 }, { "epoch": 0.15923606155906164, "grad_norm": 1.376220464706421, "learning_rate": 1.9142527479080574e-05, "loss": 1.2003, "step": 4519 }, { "epoch": 0.15927129857202005, "grad_norm": 1.3731485605239868, "learning_rate": 1.914206503773628e-05, "loss": 1.239, "step": 4520 }, { "epoch": 0.15930653558497845, "grad_norm": 2.1960084438323975, "learning_rate": 1.9141602477315424e-05, "loss": 3.9298, "step": 4521 }, { "epoch": 0.15934177259793686, "grad_norm": 4.267776012420654, "learning_rate": 1.914113979782403e-05, "loss": 4.1368, "step": 4522 }, { "epoch": 0.1593770096108953, "grad_norm": 1.2357110977172852, "learning_rate": 1.914067699926813e-05, "loss": 0.9548, "step": 4523 }, { "epoch": 0.1594122466238537, "grad_norm": 1.6774201393127441, "learning_rate": 1.9140214081653748e-05, "loss": 0.9483, "step": 4524 }, { "epoch": 0.1594474836368121, "grad_norm": 1.4313312768936157, "learning_rate": 1.913975104498691e-05, "loss": 0.8379, "step": 4525 }, { "epoch": 0.1594827206497705, "grad_norm": 5.691699504852295, "learning_rate": 1.9139287889273653e-05, "loss": 3.606, "step": 4526 }, { "epoch": 0.15951795766272894, "grad_norm": 1.5815479755401611, "learning_rate": 1.9138824614520004e-05, "loss": 0.9869, "step": 4527 }, { "epoch": 0.15955319467568735, "grad_norm": 4.625561237335205, "learning_rate": 1.9138361220732e-05, "loss": 3.4881, "step": 4528 }, { "epoch": 0.15958843168864575, "grad_norm": 1.1320173740386963, "learning_rate": 1.913789770791568e-05, "loss": 0.8408, "step": 4529 }, { "epoch": 0.15962366870160416, "grad_norm": 7.475245952606201, "learning_rate": 1.9137434076077078e-05, "loss": 7.1242, "step": 4530 }, { "epoch": 0.15965890571456257, "grad_norm": 4.067424297332764, "learning_rate": 1.913697032522223e-05, "loss": 6.2027, "step": 4531 }, { "epoch": 0.159694142727521, "grad_norm": 1.2866601943969727, "learning_rate": 1.9136506455357186e-05, "loss": 1.0745, "step": 4532 }, { "epoch": 0.1597293797404794, "grad_norm": 7.6889872550964355, "learning_rate": 1.913604246648798e-05, "loss": 6.933, "step": 4533 }, { "epoch": 0.1597646167534378, "grad_norm": 2.257415771484375, "learning_rate": 1.913557835862065e-05, "loss": 1.0198, "step": 4534 }, { "epoch": 0.15979985376639622, "grad_norm": 1.3716257810592651, "learning_rate": 1.9135114131761258e-05, "loss": 1.0256, "step": 4535 }, { "epoch": 0.15983509077935462, "grad_norm": 3.789625644683838, "learning_rate": 1.9134649785915833e-05, "loss": 3.4382, "step": 4536 }, { "epoch": 0.15987032779231306, "grad_norm": 4.332027912139893, "learning_rate": 1.9134185321090434e-05, "loss": 3.6377, "step": 4537 }, { "epoch": 0.15990556480527146, "grad_norm": 0.9516092538833618, "learning_rate": 1.913372073729111e-05, "loss": 1.0992, "step": 4538 }, { "epoch": 0.15994080181822987, "grad_norm": 9.469868659973145, "learning_rate": 1.9133256034523906e-05, "loss": 6.4445, "step": 4539 }, { "epoch": 0.15997603883118827, "grad_norm": 1.2021855115890503, "learning_rate": 1.9132791212794883e-05, "loss": 0.8483, "step": 4540 }, { "epoch": 0.1600112758441467, "grad_norm": 3.460423469543457, "learning_rate": 1.9132326272110084e-05, "loss": 3.366, "step": 4541 }, { "epoch": 0.1600465128571051, "grad_norm": 5.291296482086182, "learning_rate": 1.9131861212475578e-05, "loss": 6.2029, "step": 4542 }, { "epoch": 0.16008174987006352, "grad_norm": 6.513247966766357, "learning_rate": 1.9131396033897415e-05, "loss": 4.1567, "step": 4543 }, { "epoch": 0.16011698688302192, "grad_norm": 1.1149572134017944, "learning_rate": 1.9130930736381654e-05, "loss": 1.0801, "step": 4544 }, { "epoch": 0.16015222389598033, "grad_norm": 6.3259782791137695, "learning_rate": 1.9130465319934363e-05, "loss": 9.0376, "step": 4545 }, { "epoch": 0.16018746090893876, "grad_norm": 4.9318108558654785, "learning_rate": 1.912999978456159e-05, "loss": 5.2962, "step": 4546 }, { "epoch": 0.16022269792189717, "grad_norm": 2.7848892211914062, "learning_rate": 1.9129534130269412e-05, "loss": 0.8653, "step": 4547 }, { "epoch": 0.16025793493485557, "grad_norm": 7.86956787109375, "learning_rate": 1.9129068357063888e-05, "loss": 4.3592, "step": 4548 }, { "epoch": 0.16029317194781398, "grad_norm": 4.623654365539551, "learning_rate": 1.9128602464951083e-05, "loss": 6.9693, "step": 4549 }, { "epoch": 0.16032840896077238, "grad_norm": 2.4236528873443604, "learning_rate": 1.912813645393707e-05, "loss": 0.8632, "step": 4550 }, { "epoch": 0.16036364597373082, "grad_norm": 6.199878215789795, "learning_rate": 1.912767032402792e-05, "loss": 3.6289, "step": 4551 }, { "epoch": 0.16039888298668922, "grad_norm": 5.409835338592529, "learning_rate": 1.91272040752297e-05, "loss": 3.8697, "step": 4552 }, { "epoch": 0.16043411999964763, "grad_norm": 6.5459723472595215, "learning_rate": 1.912673770754848e-05, "loss": 3.2872, "step": 4553 }, { "epoch": 0.16046935701260603, "grad_norm": 1.3978022336959839, "learning_rate": 1.912627122099034e-05, "loss": 1.2412, "step": 4554 }, { "epoch": 0.16050459402556444, "grad_norm": 1.248574137687683, "learning_rate": 1.9125804615561357e-05, "loss": 0.9995, "step": 4555 }, { "epoch": 0.16053983103852287, "grad_norm": 0.9961839318275452, "learning_rate": 1.912533789126761e-05, "loss": 0.7944, "step": 4556 }, { "epoch": 0.16057506805148128, "grad_norm": 6.7629170417785645, "learning_rate": 1.912487104811517e-05, "loss": 6.3018, "step": 4557 }, { "epoch": 0.16061030506443968, "grad_norm": 1.7762024402618408, "learning_rate": 1.9124404086110122e-05, "loss": 1.0927, "step": 4558 }, { "epoch": 0.1606455420773981, "grad_norm": 5.054959774017334, "learning_rate": 1.912393700525855e-05, "loss": 4.9576, "step": 4559 }, { "epoch": 0.16068077909035652, "grad_norm": 9.743706703186035, "learning_rate": 1.9123469805566533e-05, "loss": 5.8651, "step": 4560 }, { "epoch": 0.16071601610331493, "grad_norm": 6.659671306610107, "learning_rate": 1.9123002487040162e-05, "loss": 4.1994, "step": 4561 }, { "epoch": 0.16075125311627333, "grad_norm": 1.2388122081756592, "learning_rate": 1.9122535049685524e-05, "loss": 0.9901, "step": 4562 }, { "epoch": 0.16078649012923174, "grad_norm": 7.511290550231934, "learning_rate": 1.91220674935087e-05, "loss": 7.1957, "step": 4563 }, { "epoch": 0.16082172714219015, "grad_norm": 1.2494362592697144, "learning_rate": 1.912159981851579e-05, "loss": 1.1926, "step": 4564 }, { "epoch": 0.16085696415514858, "grad_norm": 7.156467914581299, "learning_rate": 1.912113202471288e-05, "loss": 3.2691, "step": 4565 }, { "epoch": 0.16089220116810699, "grad_norm": 1.1238884925842285, "learning_rate": 1.912066411210606e-05, "loss": 0.8105, "step": 4566 }, { "epoch": 0.1609274381810654, "grad_norm": 5.259537220001221, "learning_rate": 1.9120196080701433e-05, "loss": 3.8969, "step": 4567 }, { "epoch": 0.1609626751940238, "grad_norm": 1.2773399353027344, "learning_rate": 1.9119727930505086e-05, "loss": 1.2152, "step": 4568 }, { "epoch": 0.1609979122069822, "grad_norm": 1.5409581661224365, "learning_rate": 1.9119259661523123e-05, "loss": 1.2051, "step": 4569 }, { "epoch": 0.16103314921994064, "grad_norm": 1.2831881046295166, "learning_rate": 1.9118791273761637e-05, "loss": 0.9817, "step": 4570 }, { "epoch": 0.16106838623289904, "grad_norm": 7.248650550842285, "learning_rate": 1.911832276722674e-05, "loss": 4.0033, "step": 4571 }, { "epoch": 0.16110362324585745, "grad_norm": 1.2708193063735962, "learning_rate": 1.9117854141924524e-05, "loss": 1.1368, "step": 4572 }, { "epoch": 0.16113886025881585, "grad_norm": 5.150350570678711, "learning_rate": 1.9117385397861096e-05, "loss": 3.5516, "step": 4573 }, { "epoch": 0.16117409727177426, "grad_norm": 4.575136184692383, "learning_rate": 1.911691653504256e-05, "loss": 3.8099, "step": 4574 }, { "epoch": 0.1612093342847327, "grad_norm": 4.688542366027832, "learning_rate": 1.911644755347503e-05, "loss": 3.0484, "step": 4575 }, { "epoch": 0.1612445712976911, "grad_norm": 4.794185161590576, "learning_rate": 1.9115978453164607e-05, "loss": 3.8535, "step": 4576 }, { "epoch": 0.1612798083106495, "grad_norm": 1.8259588479995728, "learning_rate": 1.91155092341174e-05, "loss": 1.0332, "step": 4577 }, { "epoch": 0.1613150453236079, "grad_norm": 7.502833366394043, "learning_rate": 1.911503989633953e-05, "loss": 3.8278, "step": 4578 }, { "epoch": 0.16135028233656634, "grad_norm": 1.3956528902053833, "learning_rate": 1.91145704398371e-05, "loss": 0.9453, "step": 4579 }, { "epoch": 0.16138551934952475, "grad_norm": 1.0639299154281616, "learning_rate": 1.911410086461623e-05, "loss": 0.9664, "step": 4580 }, { "epoch": 0.16142075636248315, "grad_norm": 1.3549543619155884, "learning_rate": 1.911363117068304e-05, "loss": 0.9984, "step": 4581 }, { "epoch": 0.16145599337544156, "grad_norm": 1.4583675861358643, "learning_rate": 1.9113161358043637e-05, "loss": 1.103, "step": 4582 }, { "epoch": 0.16149123038839996, "grad_norm": 1.4037530422210693, "learning_rate": 1.9112691426704146e-05, "loss": 0.8684, "step": 4583 }, { "epoch": 0.1615264674013584, "grad_norm": 1.2066148519515991, "learning_rate": 1.9112221376670694e-05, "loss": 1.138, "step": 4584 }, { "epoch": 0.1615617044143168, "grad_norm": 5.634387016296387, "learning_rate": 1.9111751207949392e-05, "loss": 3.2036, "step": 4585 }, { "epoch": 0.1615969414272752, "grad_norm": 6.015910625457764, "learning_rate": 1.9111280920546374e-05, "loss": 7.6835, "step": 4586 }, { "epoch": 0.16163217844023361, "grad_norm": 3.0850536823272705, "learning_rate": 1.911081051446776e-05, "loss": 2.4269, "step": 4587 }, { "epoch": 0.16166741545319202, "grad_norm": 3.0868470668792725, "learning_rate": 1.9110339989719678e-05, "loss": 6.5522, "step": 4588 }, { "epoch": 0.16170265246615045, "grad_norm": 6.039217948913574, "learning_rate": 1.9109869346308255e-05, "loss": 3.3473, "step": 4589 }, { "epoch": 0.16173788947910886, "grad_norm": 5.733422756195068, "learning_rate": 1.910939858423963e-05, "loss": 3.4111, "step": 4590 }, { "epoch": 0.16177312649206727, "grad_norm": 3.1966185569763184, "learning_rate": 1.910892770351992e-05, "loss": 3.0244, "step": 4591 }, { "epoch": 0.16180836350502567, "grad_norm": 3.3905954360961914, "learning_rate": 1.9108456704155275e-05, "loss": 3.63, "step": 4592 }, { "epoch": 0.1618436005179841, "grad_norm": 1.420019268989563, "learning_rate": 1.9107985586151814e-05, "loss": 0.9551, "step": 4593 }, { "epoch": 0.1618788375309425, "grad_norm": 1.2003833055496216, "learning_rate": 1.9107514349515682e-05, "loss": 1.1011, "step": 4594 }, { "epoch": 0.16191407454390092, "grad_norm": 1.12966787815094, "learning_rate": 1.9107042994253016e-05, "loss": 1.0126, "step": 4595 }, { "epoch": 0.16194931155685932, "grad_norm": 3.526947021484375, "learning_rate": 1.9106571520369956e-05, "loss": 3.1722, "step": 4596 }, { "epoch": 0.16198454856981773, "grad_norm": 2.394247055053711, "learning_rate": 1.910609992787264e-05, "loss": 3.9173, "step": 4597 }, { "epoch": 0.16201978558277616, "grad_norm": 3.7578883171081543, "learning_rate": 1.9105628216767216e-05, "loss": 3.274, "step": 4598 }, { "epoch": 0.16205502259573457, "grad_norm": 1.859333872795105, "learning_rate": 1.910515638705982e-05, "loss": 0.8587, "step": 4599 }, { "epoch": 0.16209025960869297, "grad_norm": 2.973053455352783, "learning_rate": 1.910468443875661e-05, "loss": 6.0285, "step": 4600 }, { "epoch": 0.16212549662165138, "grad_norm": 3.192551851272583, "learning_rate": 1.910421237186372e-05, "loss": 3.9524, "step": 4601 }, { "epoch": 0.16216073363460978, "grad_norm": 1.3292746543884277, "learning_rate": 1.9103740186387304e-05, "loss": 0.9566, "step": 4602 }, { "epoch": 0.16219597064756822, "grad_norm": 2.2592549324035645, "learning_rate": 1.9103267882333513e-05, "loss": 3.1798, "step": 4603 }, { "epoch": 0.16223120766052662, "grad_norm": 1.5207031965255737, "learning_rate": 1.9102795459708496e-05, "loss": 1.2236, "step": 4604 }, { "epoch": 0.16226644467348503, "grad_norm": 1.3255168199539185, "learning_rate": 1.9102322918518413e-05, "loss": 0.9971, "step": 4605 }, { "epoch": 0.16230168168644343, "grad_norm": 5.845489025115967, "learning_rate": 1.910185025876941e-05, "loss": 3.6248, "step": 4606 }, { "epoch": 0.16233691869940184, "grad_norm": 2.800504684448242, "learning_rate": 1.9101377480467652e-05, "loss": 4.4353, "step": 4607 }, { "epoch": 0.16237215571236027, "grad_norm": 2.2030816078186035, "learning_rate": 1.910090458361929e-05, "loss": 3.2811, "step": 4608 }, { "epoch": 0.16240739272531868, "grad_norm": 4.170436859130859, "learning_rate": 1.910043156823049e-05, "loss": 3.2011, "step": 4609 }, { "epoch": 0.16244262973827708, "grad_norm": 5.4467997550964355, "learning_rate": 1.9099958434307408e-05, "loss": 4.2224, "step": 4610 }, { "epoch": 0.1624778667512355, "grad_norm": 0.9001838564872742, "learning_rate": 1.9099485181856207e-05, "loss": 0.9721, "step": 4611 }, { "epoch": 0.16251310376419392, "grad_norm": 4.205148696899414, "learning_rate": 1.909901181088305e-05, "loss": 3.1406, "step": 4612 }, { "epoch": 0.16254834077715233, "grad_norm": 2.556375026702881, "learning_rate": 1.909853832139411e-05, "loss": 3.8215, "step": 4613 }, { "epoch": 0.16258357779011073, "grad_norm": 1.8211121559143066, "learning_rate": 1.9098064713395547e-05, "loss": 0.9552, "step": 4614 }, { "epoch": 0.16261881480306914, "grad_norm": 1.4033596515655518, "learning_rate": 1.9097590986893534e-05, "loss": 0.9077, "step": 4615 }, { "epoch": 0.16265405181602755, "grad_norm": 1.1848644018173218, "learning_rate": 1.9097117141894238e-05, "loss": 0.9734, "step": 4616 }, { "epoch": 0.16268928882898598, "grad_norm": 1.6592358350753784, "learning_rate": 1.909664317840383e-05, "loss": 0.9181, "step": 4617 }, { "epoch": 0.16272452584194438, "grad_norm": 1.425878882408142, "learning_rate": 1.909616909642849e-05, "loss": 0.9878, "step": 4618 }, { "epoch": 0.1627597628549028, "grad_norm": 2.2903025150299072, "learning_rate": 1.909569489597439e-05, "loss": 4.1405, "step": 4619 }, { "epoch": 0.1627949998678612, "grad_norm": 1.6156866550445557, "learning_rate": 1.90952205770477e-05, "loss": 1.0464, "step": 4620 }, { "epoch": 0.1628302368808196, "grad_norm": 5.00687313079834, "learning_rate": 1.9094746139654606e-05, "loss": 3.7821, "step": 4621 }, { "epoch": 0.16286547389377803, "grad_norm": 3.0439958572387695, "learning_rate": 1.9094271583801286e-05, "loss": 4.2839, "step": 4622 }, { "epoch": 0.16290071090673644, "grad_norm": 2.977290391921997, "learning_rate": 1.909379690949392e-05, "loss": 4.0884, "step": 4623 }, { "epoch": 0.16293594791969485, "grad_norm": 3.4319159984588623, "learning_rate": 1.9093322116738693e-05, "loss": 6.0221, "step": 4624 }, { "epoch": 0.16297118493265325, "grad_norm": 1.4699604511260986, "learning_rate": 1.9092847205541784e-05, "loss": 0.9621, "step": 4625 }, { "epoch": 0.16300642194561166, "grad_norm": 2.326932907104492, "learning_rate": 1.909237217590938e-05, "loss": 4.0079, "step": 4626 }, { "epoch": 0.1630416589585701, "grad_norm": 1.2095491886138916, "learning_rate": 1.9091897027847673e-05, "loss": 1.0837, "step": 4627 }, { "epoch": 0.1630768959715285, "grad_norm": 2.825608968734741, "learning_rate": 1.9091421761362848e-05, "loss": 3.7863, "step": 4628 }, { "epoch": 0.1631121329844869, "grad_norm": 3.0974462032318115, "learning_rate": 1.90909463764611e-05, "loss": 3.8204, "step": 4629 }, { "epoch": 0.1631473699974453, "grad_norm": 2.084296941757202, "learning_rate": 1.9090470873148614e-05, "loss": 1.1134, "step": 4630 }, { "epoch": 0.16318260701040374, "grad_norm": 1.6057862043380737, "learning_rate": 1.9089995251431588e-05, "loss": 1.1648, "step": 4631 }, { "epoch": 0.16321784402336215, "grad_norm": 1.4867031574249268, "learning_rate": 1.9089519511316215e-05, "loss": 0.9608, "step": 4632 }, { "epoch": 0.16325308103632055, "grad_norm": 1.8761008977890015, "learning_rate": 1.9089043652808693e-05, "loss": 0.9601, "step": 4633 }, { "epoch": 0.16328831804927896, "grad_norm": 1.3985950946807861, "learning_rate": 1.908856767591522e-05, "loss": 1.2318, "step": 4634 }, { "epoch": 0.16332355506223736, "grad_norm": 2.1639368534088135, "learning_rate": 1.9088091580641993e-05, "loss": 3.3304, "step": 4635 }, { "epoch": 0.1633587920751958, "grad_norm": 6.502616882324219, "learning_rate": 1.908761536699522e-05, "loss": 4.0828, "step": 4636 }, { "epoch": 0.1633940290881542, "grad_norm": 1.5661131143569946, "learning_rate": 1.9087139034981098e-05, "loss": 1.4199, "step": 4637 }, { "epoch": 0.1634292661011126, "grad_norm": 1.6415727138519287, "learning_rate": 1.908666258460583e-05, "loss": 0.8851, "step": 4638 }, { "epoch": 0.163464503114071, "grad_norm": 1.155043125152588, "learning_rate": 1.908618601587563e-05, "loss": 1.1232, "step": 4639 }, { "epoch": 0.16349974012702942, "grad_norm": 8.050857543945312, "learning_rate": 1.9085709328796693e-05, "loss": 3.5867, "step": 4640 }, { "epoch": 0.16353497713998785, "grad_norm": 1.7466362714767456, "learning_rate": 1.9085232523375237e-05, "loss": 0.968, "step": 4641 }, { "epoch": 0.16357021415294626, "grad_norm": 5.85666036605835, "learning_rate": 1.908475559961747e-05, "loss": 6.0521, "step": 4642 }, { "epoch": 0.16360545116590466, "grad_norm": 1.395949125289917, "learning_rate": 1.9084278557529604e-05, "loss": 0.8579, "step": 4643 }, { "epoch": 0.16364068817886307, "grad_norm": 5.631197929382324, "learning_rate": 1.9083801397117853e-05, "loss": 3.1945, "step": 4644 }, { "epoch": 0.1636759251918215, "grad_norm": 3.8489339351654053, "learning_rate": 1.9083324118388433e-05, "loss": 3.4505, "step": 4645 }, { "epoch": 0.1637111622047799, "grad_norm": 1.2704458236694336, "learning_rate": 1.9082846721347558e-05, "loss": 1.1266, "step": 4646 }, { "epoch": 0.16374639921773831, "grad_norm": 1.6328755617141724, "learning_rate": 1.908236920600145e-05, "loss": 0.9806, "step": 4647 }, { "epoch": 0.16378163623069672, "grad_norm": 7.080931186676025, "learning_rate": 1.908189157235632e-05, "loss": 6.5639, "step": 4648 }, { "epoch": 0.16381687324365513, "grad_norm": 8.196985244750977, "learning_rate": 1.9081413820418402e-05, "loss": 5.8629, "step": 4649 }, { "epoch": 0.16385211025661356, "grad_norm": 7.179581165313721, "learning_rate": 1.9080935950193908e-05, "loss": 3.2982, "step": 4650 }, { "epoch": 0.16388734726957196, "grad_norm": 4.707386016845703, "learning_rate": 1.908045796168907e-05, "loss": 3.6512, "step": 4651 }, { "epoch": 0.16392258428253037, "grad_norm": 1.59372079372406, "learning_rate": 1.9079979854910106e-05, "loss": 1.2658, "step": 4652 }, { "epoch": 0.16395782129548878, "grad_norm": 0.9908966422080994, "learning_rate": 1.907950162986325e-05, "loss": 1.0626, "step": 4653 }, { "epoch": 0.16399305830844718, "grad_norm": 1.4605780839920044, "learning_rate": 1.907902328655473e-05, "loss": 0.8332, "step": 4654 }, { "epoch": 0.16402829532140561, "grad_norm": 5.739521026611328, "learning_rate": 1.9078544824990772e-05, "loss": 6.6817, "step": 4655 }, { "epoch": 0.16406353233436402, "grad_norm": 7.371153831481934, "learning_rate": 1.9078066245177617e-05, "loss": 8.6023, "step": 4656 }, { "epoch": 0.16409876934732243, "grad_norm": 1.717725396156311, "learning_rate": 1.9077587547121487e-05, "loss": 1.0952, "step": 4657 }, { "epoch": 0.16413400636028083, "grad_norm": 1.4316126108169556, "learning_rate": 1.9077108730828625e-05, "loss": 1.267, "step": 4658 }, { "epoch": 0.16416924337323924, "grad_norm": 3.101724624633789, "learning_rate": 1.9076629796305264e-05, "loss": 3.0697, "step": 4659 }, { "epoch": 0.16420448038619767, "grad_norm": 2.456315517425537, "learning_rate": 1.9076150743557646e-05, "loss": 3.5388, "step": 4660 }, { "epoch": 0.16423971739915608, "grad_norm": 1.3967421054840088, "learning_rate": 1.907567157259201e-05, "loss": 1.2335, "step": 4661 }, { "epoch": 0.16427495441211448, "grad_norm": 4.773200035095215, "learning_rate": 1.9075192283414593e-05, "loss": 3.3587, "step": 4662 }, { "epoch": 0.1643101914250729, "grad_norm": 5.134305477142334, "learning_rate": 1.907471287603164e-05, "loss": 6.6453, "step": 4663 }, { "epoch": 0.16434542843803132, "grad_norm": 9.112104415893555, "learning_rate": 1.9074233350449397e-05, "loss": 3.9293, "step": 4664 }, { "epoch": 0.16438066545098973, "grad_norm": 1.1271319389343262, "learning_rate": 1.9073753706674108e-05, "loss": 1.0984, "step": 4665 }, { "epoch": 0.16441590246394813, "grad_norm": 4.589338302612305, "learning_rate": 1.9073273944712023e-05, "loss": 5.7071, "step": 4666 }, { "epoch": 0.16445113947690654, "grad_norm": 4.773998260498047, "learning_rate": 1.9072794064569388e-05, "loss": 5.3354, "step": 4667 }, { "epoch": 0.16448637648986494, "grad_norm": 8.781486511230469, "learning_rate": 1.9072314066252456e-05, "loss": 3.4565, "step": 4668 }, { "epoch": 0.16452161350282338, "grad_norm": 1.9690508842468262, "learning_rate": 1.9071833949767472e-05, "loss": 1.1476, "step": 4669 }, { "epoch": 0.16455685051578178, "grad_norm": 4.381481170654297, "learning_rate": 1.90713537151207e-05, "loss": 3.1107, "step": 4670 }, { "epoch": 0.1645920875287402, "grad_norm": 4.473895072937012, "learning_rate": 1.907087336231839e-05, "loss": 3.5704, "step": 4671 }, { "epoch": 0.1646273245416986, "grad_norm": 6.068910121917725, "learning_rate": 1.9070392891366802e-05, "loss": 6.4662, "step": 4672 }, { "epoch": 0.164662561554657, "grad_norm": 5.737960338592529, "learning_rate": 1.9069912302272186e-05, "loss": 2.805, "step": 4673 }, { "epoch": 0.16469779856761543, "grad_norm": 2.36474347114563, "learning_rate": 1.9069431595040807e-05, "loss": 2.8729, "step": 4674 }, { "epoch": 0.16473303558057384, "grad_norm": 7.05924129486084, "learning_rate": 1.906895076967893e-05, "loss": 4.3846, "step": 4675 }, { "epoch": 0.16476827259353224, "grad_norm": 1.2518290281295776, "learning_rate": 1.9068469826192814e-05, "loss": 1.0829, "step": 4676 }, { "epoch": 0.16480350960649065, "grad_norm": 4.711747169494629, "learning_rate": 1.906798876458872e-05, "loss": 3.9315, "step": 4677 }, { "epoch": 0.16483874661944906, "grad_norm": 0.9463072419166565, "learning_rate": 1.906750758487292e-05, "loss": 0.9295, "step": 4678 }, { "epoch": 0.1648739836324075, "grad_norm": 7.486643314361572, "learning_rate": 1.9067026287051677e-05, "loss": 3.825, "step": 4679 }, { "epoch": 0.1649092206453659, "grad_norm": 4.179172515869141, "learning_rate": 1.906654487113126e-05, "loss": 3.3407, "step": 4680 }, { "epoch": 0.1649444576583243, "grad_norm": 7.426365375518799, "learning_rate": 1.9066063337117943e-05, "loss": 8.1973, "step": 4681 }, { "epoch": 0.1649796946712827, "grad_norm": 3.668902635574341, "learning_rate": 1.9065581685017998e-05, "loss": 3.4602, "step": 4682 }, { "epoch": 0.16501493168424114, "grad_norm": 6.596862316131592, "learning_rate": 1.9065099914837696e-05, "loss": 3.7357, "step": 4683 }, { "epoch": 0.16505016869719955, "grad_norm": 3.513542413711548, "learning_rate": 1.906461802658331e-05, "loss": 3.5149, "step": 4684 }, { "epoch": 0.16508540571015795, "grad_norm": 3.4306461811065674, "learning_rate": 1.9064136020261124e-05, "loss": 3.8683, "step": 4685 }, { "epoch": 0.16512064272311636, "grad_norm": 5.555293560028076, "learning_rate": 1.9063653895877407e-05, "loss": 6.4296, "step": 4686 }, { "epoch": 0.16515587973607476, "grad_norm": 5.268887996673584, "learning_rate": 1.9063171653438446e-05, "loss": 5.9348, "step": 4687 }, { "epoch": 0.1651911167490332, "grad_norm": 0.9533567428588867, "learning_rate": 1.9062689292950522e-05, "loss": 1.3353, "step": 4688 }, { "epoch": 0.1652263537619916, "grad_norm": 1.2966512441635132, "learning_rate": 1.9062206814419914e-05, "loss": 1.172, "step": 4689 }, { "epoch": 0.16526159077495, "grad_norm": 4.543644905090332, "learning_rate": 1.9061724217852906e-05, "loss": 5.7759, "step": 4690 }, { "epoch": 0.1652968277879084, "grad_norm": 1.0910708904266357, "learning_rate": 1.9061241503255784e-05, "loss": 1.0687, "step": 4691 }, { "epoch": 0.16533206480086682, "grad_norm": 1.2547274827957153, "learning_rate": 1.9060758670634843e-05, "loss": 0.9679, "step": 4692 }, { "epoch": 0.16536730181382525, "grad_norm": 3.596393346786499, "learning_rate": 1.906027571999636e-05, "loss": 3.8199, "step": 4693 }, { "epoch": 0.16540253882678366, "grad_norm": 1.2559804916381836, "learning_rate": 1.9059792651346637e-05, "loss": 0.9476, "step": 4694 }, { "epoch": 0.16543777583974206, "grad_norm": 2.047024965286255, "learning_rate": 1.905930946469196e-05, "loss": 1.0932, "step": 4695 }, { "epoch": 0.16547301285270047, "grad_norm": 5.130107879638672, "learning_rate": 1.905882616003862e-05, "loss": 4.1961, "step": 4696 }, { "epoch": 0.1655082498656589, "grad_norm": 1.2263247966766357, "learning_rate": 1.905834273739292e-05, "loss": 1.1052, "step": 4697 }, { "epoch": 0.1655434868786173, "grad_norm": 4.717258930206299, "learning_rate": 1.905785919676115e-05, "loss": 3.1122, "step": 4698 }, { "epoch": 0.1655787238915757, "grad_norm": 3.951782464981079, "learning_rate": 1.9057375538149606e-05, "loss": 3.6529, "step": 4699 }, { "epoch": 0.16561396090453412, "grad_norm": 4.315159797668457, "learning_rate": 1.9056891761564595e-05, "loss": 8.502, "step": 4700 }, { "epoch": 0.16564919791749252, "grad_norm": 4.5444655418396, "learning_rate": 1.9056407867012417e-05, "loss": 1.0366, "step": 4701 }, { "epoch": 0.16568443493045096, "grad_norm": 2.088557481765747, "learning_rate": 1.905592385449937e-05, "loss": 1.0674, "step": 4702 }, { "epoch": 0.16571967194340936, "grad_norm": 1.1546746492385864, "learning_rate": 1.9055439724031763e-05, "loss": 0.9295, "step": 4703 }, { "epoch": 0.16575490895636777, "grad_norm": 4.0346760749816895, "learning_rate": 1.9054955475615898e-05, "loss": 4.0976, "step": 4704 }, { "epoch": 0.16579014596932617, "grad_norm": 1.9507369995117188, "learning_rate": 1.9054471109258088e-05, "loss": 1.093, "step": 4705 }, { "epoch": 0.16582538298228458, "grad_norm": 1.2956821918487549, "learning_rate": 1.9053986624964635e-05, "loss": 0.9814, "step": 4706 }, { "epoch": 0.165860619995243, "grad_norm": 1.5138550996780396, "learning_rate": 1.9053502022741857e-05, "loss": 1.0006, "step": 4707 }, { "epoch": 0.16589585700820142, "grad_norm": 3.518613815307617, "learning_rate": 1.905301730259606e-05, "loss": 3.4432, "step": 4708 }, { "epoch": 0.16593109402115983, "grad_norm": 1.4717689752578735, "learning_rate": 1.9052532464533558e-05, "loss": 1.1661, "step": 4709 }, { "epoch": 0.16596633103411823, "grad_norm": 2.8010849952697754, "learning_rate": 1.9052047508560667e-05, "loss": 3.2341, "step": 4710 }, { "epoch": 0.16600156804707664, "grad_norm": 1.2383482456207275, "learning_rate": 1.9051562434683707e-05, "loss": 1.0979, "step": 4711 }, { "epoch": 0.16603680506003507, "grad_norm": 1.2938810586929321, "learning_rate": 1.905107724290899e-05, "loss": 0.9535, "step": 4712 }, { "epoch": 0.16607204207299348, "grad_norm": 2.627495527267456, "learning_rate": 1.9050591933242845e-05, "loss": 3.1945, "step": 4713 }, { "epoch": 0.16610727908595188, "grad_norm": 1.7912360429763794, "learning_rate": 1.905010650569158e-05, "loss": 1.0077, "step": 4714 }, { "epoch": 0.1661425160989103, "grad_norm": 1.6505835056304932, "learning_rate": 1.904962096026153e-05, "loss": 0.9237, "step": 4715 }, { "epoch": 0.16617775311186872, "grad_norm": 4.75079870223999, "learning_rate": 1.904913529695901e-05, "loss": 3.3292, "step": 4716 }, { "epoch": 0.16621299012482713, "grad_norm": 1.272021770477295, "learning_rate": 1.9048649515790356e-05, "loss": 1.0715, "step": 4717 }, { "epoch": 0.16624822713778553, "grad_norm": 3.317061185836792, "learning_rate": 1.9048163616761884e-05, "loss": 3.2643, "step": 4718 }, { "epoch": 0.16628346415074394, "grad_norm": 1.3348560333251953, "learning_rate": 1.904767759987993e-05, "loss": 1.1098, "step": 4719 }, { "epoch": 0.16631870116370234, "grad_norm": 2.6628870964050293, "learning_rate": 1.904719146515082e-05, "loss": 3.7857, "step": 4720 }, { "epoch": 0.16635393817666078, "grad_norm": 3.799412965774536, "learning_rate": 1.9046705212580894e-05, "loss": 3.822, "step": 4721 }, { "epoch": 0.16638917518961918, "grad_norm": 4.847794055938721, "learning_rate": 1.9046218842176477e-05, "loss": 3.6934, "step": 4722 }, { "epoch": 0.1664244122025776, "grad_norm": 3.8415653705596924, "learning_rate": 1.904573235394391e-05, "loss": 3.5009, "step": 4723 }, { "epoch": 0.166459649215536, "grad_norm": 6.639586448669434, "learning_rate": 1.9045245747889525e-05, "loss": 5.9725, "step": 4724 }, { "epoch": 0.1664948862284944, "grad_norm": 1.6219712495803833, "learning_rate": 1.9044759024019664e-05, "loss": 0.9248, "step": 4725 }, { "epoch": 0.16653012324145283, "grad_norm": 3.309070587158203, "learning_rate": 1.904427218234066e-05, "loss": 1.0784, "step": 4726 }, { "epoch": 0.16656536025441124, "grad_norm": 1.2891470193862915, "learning_rate": 1.9043785222858862e-05, "loss": 0.8676, "step": 4727 }, { "epoch": 0.16660059726736964, "grad_norm": 6.602024555206299, "learning_rate": 1.904329814558061e-05, "loss": 3.083, "step": 4728 }, { "epoch": 0.16663583428032805, "grad_norm": 4.270992279052734, "learning_rate": 1.9042810950512246e-05, "loss": 2.7689, "step": 4729 }, { "epoch": 0.16667107129328645, "grad_norm": 1.0206817388534546, "learning_rate": 1.904232363766012e-05, "loss": 0.8358, "step": 4730 }, { "epoch": 0.1667063083062449, "grad_norm": 4.738133907318115, "learning_rate": 1.9041836207030573e-05, "loss": 4.151, "step": 4731 }, { "epoch": 0.1667415453192033, "grad_norm": 1.275466799736023, "learning_rate": 1.904134865862996e-05, "loss": 1.1389, "step": 4732 }, { "epoch": 0.1667767823321617, "grad_norm": 1.4681591987609863, "learning_rate": 1.9040860992464628e-05, "loss": 1.0364, "step": 4733 }, { "epoch": 0.1668120193451201, "grad_norm": 3.645394802093506, "learning_rate": 1.904037320854093e-05, "loss": 3.483, "step": 4734 }, { "epoch": 0.16684725635807854, "grad_norm": 1.1831676959991455, "learning_rate": 1.903988530686522e-05, "loss": 1.1145, "step": 4735 }, { "epoch": 0.16688249337103694, "grad_norm": 1.4128998517990112, "learning_rate": 1.9039397287443854e-05, "loss": 1.1129, "step": 4736 }, { "epoch": 0.16691773038399535, "grad_norm": 5.483983516693115, "learning_rate": 1.9038909150283184e-05, "loss": 5.3351, "step": 4737 }, { "epoch": 0.16695296739695376, "grad_norm": 1.965609073638916, "learning_rate": 1.9038420895389572e-05, "loss": 1.102, "step": 4738 }, { "epoch": 0.16698820440991216, "grad_norm": 4.908634185791016, "learning_rate": 1.9037932522769376e-05, "loss": 6.6411, "step": 4739 }, { "epoch": 0.1670234414228706, "grad_norm": 1.5427464246749878, "learning_rate": 1.903744403242896e-05, "loss": 0.9187, "step": 4740 }, { "epoch": 0.167058678435829, "grad_norm": 3.3138058185577393, "learning_rate": 1.9036955424374685e-05, "loss": 3.1066, "step": 4741 }, { "epoch": 0.1670939154487874, "grad_norm": 6.595921039581299, "learning_rate": 1.903646669861291e-05, "loss": 6.1931, "step": 4742 }, { "epoch": 0.1671291524617458, "grad_norm": 1.2327163219451904, "learning_rate": 1.903597785515001e-05, "loss": 1.1042, "step": 4743 }, { "epoch": 0.16716438947470422, "grad_norm": 1.3143327236175537, "learning_rate": 1.9035488893992344e-05, "loss": 0.985, "step": 4744 }, { "epoch": 0.16719962648766265, "grad_norm": 5.253249645233154, "learning_rate": 1.903499981514628e-05, "loss": 6.0447, "step": 4745 }, { "epoch": 0.16723486350062106, "grad_norm": 1.0745198726654053, "learning_rate": 1.90345106186182e-05, "loss": 0.9589, "step": 4746 }, { "epoch": 0.16727010051357946, "grad_norm": 5.114069938659668, "learning_rate": 1.9034021304414465e-05, "loss": 3.3611, "step": 4747 }, { "epoch": 0.16730533752653787, "grad_norm": 3.237917423248291, "learning_rate": 1.9033531872541454e-05, "loss": 3.8894, "step": 4748 }, { "epoch": 0.1673405745394963, "grad_norm": 1.1540688276290894, "learning_rate": 1.903304232300554e-05, "loss": 1.0149, "step": 4749 }, { "epoch": 0.1673758115524547, "grad_norm": 1.259799599647522, "learning_rate": 1.9032552655813098e-05, "loss": 1.0408, "step": 4750 }, { "epoch": 0.1674110485654131, "grad_norm": 1.8293925523757935, "learning_rate": 1.90320628709705e-05, "loss": 1.023, "step": 4751 }, { "epoch": 0.16744628557837152, "grad_norm": 3.0683043003082275, "learning_rate": 1.9031572968484142e-05, "loss": 3.5656, "step": 4752 }, { "epoch": 0.16748152259132992, "grad_norm": 1.1948840618133545, "learning_rate": 1.9031082948360394e-05, "loss": 1.0782, "step": 4753 }, { "epoch": 0.16751675960428836, "grad_norm": 5.539646625518799, "learning_rate": 1.903059281060564e-05, "loss": 3.6418, "step": 4754 }, { "epoch": 0.16755199661724676, "grad_norm": 1.3324459791183472, "learning_rate": 1.9030102555226263e-05, "loss": 0.8784, "step": 4755 }, { "epoch": 0.16758723363020517, "grad_norm": 5.799771785736084, "learning_rate": 1.902961218222865e-05, "loss": 4.2323, "step": 4756 }, { "epoch": 0.16762247064316357, "grad_norm": 5.016611576080322, "learning_rate": 1.9029121691619188e-05, "loss": 3.7016, "step": 4757 }, { "epoch": 0.16765770765612198, "grad_norm": 3.4953434467315674, "learning_rate": 1.9028631083404266e-05, "loss": 3.8907, "step": 4758 }, { "epoch": 0.1676929446690804, "grad_norm": 1.4511040449142456, "learning_rate": 1.9028140357590274e-05, "loss": 1.0183, "step": 4759 }, { "epoch": 0.16772818168203882, "grad_norm": 1.291340947151184, "learning_rate": 1.9027649514183607e-05, "loss": 1.114, "step": 4760 }, { "epoch": 0.16776341869499722, "grad_norm": 1.2428672313690186, "learning_rate": 1.902715855319065e-05, "loss": 0.8906, "step": 4761 }, { "epoch": 0.16779865570795563, "grad_norm": 2.5943901538848877, "learning_rate": 1.9026667474617807e-05, "loss": 3.4283, "step": 4762 }, { "epoch": 0.16783389272091404, "grad_norm": 1.7173523902893066, "learning_rate": 1.9026176278471465e-05, "loss": 1.1896, "step": 4763 }, { "epoch": 0.16786912973387247, "grad_norm": 1.4410243034362793, "learning_rate": 1.9025684964758034e-05, "loss": 0.8774, "step": 4764 }, { "epoch": 0.16790436674683087, "grad_norm": 1.829854965209961, "learning_rate": 1.90251935334839e-05, "loss": 0.8772, "step": 4765 }, { "epoch": 0.16793960375978928, "grad_norm": 1.1940265893936157, "learning_rate": 1.9024701984655476e-05, "loss": 1.0497, "step": 4766 }, { "epoch": 0.16797484077274769, "grad_norm": 4.042057514190674, "learning_rate": 1.9024210318279156e-05, "loss": 5.7012, "step": 4767 }, { "epoch": 0.16801007778570612, "grad_norm": 4.036195755004883, "learning_rate": 1.9023718534361348e-05, "loss": 4.1687, "step": 4768 }, { "epoch": 0.16804531479866452, "grad_norm": 1.0316931009292603, "learning_rate": 1.902322663290846e-05, "loss": 0.9902, "step": 4769 }, { "epoch": 0.16808055181162293, "grad_norm": 3.7941315174102783, "learning_rate": 1.902273461392689e-05, "loss": 3.7948, "step": 4770 }, { "epoch": 0.16811578882458134, "grad_norm": 3.937063217163086, "learning_rate": 1.9022242477423057e-05, "loss": 3.0033, "step": 4771 }, { "epoch": 0.16815102583753974, "grad_norm": 8.698638916015625, "learning_rate": 1.9021750223403362e-05, "loss": 9.8849, "step": 4772 }, { "epoch": 0.16818626285049817, "grad_norm": 5.021594524383545, "learning_rate": 1.9021257851874228e-05, "loss": 6.3383, "step": 4773 }, { "epoch": 0.16822149986345658, "grad_norm": 2.0113015174865723, "learning_rate": 1.9020765362842054e-05, "loss": 1.1224, "step": 4774 }, { "epoch": 0.168256736876415, "grad_norm": 7.111907482147217, "learning_rate": 1.9020272756313264e-05, "loss": 6.6325, "step": 4775 }, { "epoch": 0.1682919738893734, "grad_norm": 3.305945873260498, "learning_rate": 1.9019780032294275e-05, "loss": 3.8799, "step": 4776 }, { "epoch": 0.1683272109023318, "grad_norm": 1.799731731414795, "learning_rate": 1.90192871907915e-05, "loss": 1.2411, "step": 4777 }, { "epoch": 0.16836244791529023, "grad_norm": 0.9832956194877625, "learning_rate": 1.901879423181136e-05, "loss": 1.2275, "step": 4778 }, { "epoch": 0.16839768492824864, "grad_norm": 1.2664119005203247, "learning_rate": 1.901830115536028e-05, "loss": 0.9789, "step": 4779 }, { "epoch": 0.16843292194120704, "grad_norm": 0.9664234519004822, "learning_rate": 1.9017807961444674e-05, "loss": 1.5248, "step": 4780 }, { "epoch": 0.16846815895416545, "grad_norm": 6.780928611755371, "learning_rate": 1.9017314650070973e-05, "loss": 3.3424, "step": 4781 }, { "epoch": 0.16850339596712385, "grad_norm": 1.1203725337982178, "learning_rate": 1.90168212212456e-05, "loss": 1.4161, "step": 4782 }, { "epoch": 0.1685386329800823, "grad_norm": 1.4984195232391357, "learning_rate": 1.9016327674974982e-05, "loss": 1.1929, "step": 4783 }, { "epoch": 0.1685738699930407, "grad_norm": 1.1226704120635986, "learning_rate": 1.901583401126555e-05, "loss": 1.1136, "step": 4784 }, { "epoch": 0.1686091070059991, "grad_norm": 8.077974319458008, "learning_rate": 1.9015340230123727e-05, "loss": 8.3255, "step": 4785 }, { "epoch": 0.1686443440189575, "grad_norm": 3.171696186065674, "learning_rate": 1.9014846331555952e-05, "loss": 3.5783, "step": 4786 }, { "epoch": 0.16867958103191594, "grad_norm": 1.4786157608032227, "learning_rate": 1.901435231556865e-05, "loss": 1.0857, "step": 4787 }, { "epoch": 0.16871481804487434, "grad_norm": 1.762430191040039, "learning_rate": 1.901385818216827e-05, "loss": 0.9068, "step": 4788 }, { "epoch": 0.16875005505783275, "grad_norm": 1.537636399269104, "learning_rate": 1.9013363931361232e-05, "loss": 1.0641, "step": 4789 }, { "epoch": 0.16878529207079115, "grad_norm": 1.3989429473876953, "learning_rate": 1.9012869563153982e-05, "loss": 1.0071, "step": 4790 }, { "epoch": 0.16882052908374956, "grad_norm": 1.3235931396484375, "learning_rate": 1.9012375077552957e-05, "loss": 0.8767, "step": 4791 }, { "epoch": 0.168855766096708, "grad_norm": 5.496764183044434, "learning_rate": 1.90118804745646e-05, "loss": 6.5529, "step": 4792 }, { "epoch": 0.1688910031096664, "grad_norm": 5.436790943145752, "learning_rate": 1.9011385754195356e-05, "loss": 4.0167, "step": 4793 }, { "epoch": 0.1689262401226248, "grad_norm": 1.2550691366195679, "learning_rate": 1.9010890916451662e-05, "loss": 1.0367, "step": 4794 }, { "epoch": 0.1689614771355832, "grad_norm": 3.4113874435424805, "learning_rate": 1.9010395961339962e-05, "loss": 4.1507, "step": 4795 }, { "epoch": 0.16899671414854162, "grad_norm": 7.956758975982666, "learning_rate": 1.900990088886671e-05, "loss": 5.9031, "step": 4796 }, { "epoch": 0.16903195116150005, "grad_norm": 6.2596940994262695, "learning_rate": 1.900940569903835e-05, "loss": 3.8411, "step": 4797 }, { "epoch": 0.16906718817445845, "grad_norm": 2.3103034496307373, "learning_rate": 1.9008910391861335e-05, "loss": 2.7599, "step": 4798 }, { "epoch": 0.16910242518741686, "grad_norm": 4.855148792266846, "learning_rate": 1.9008414967342116e-05, "loss": 5.8245, "step": 4799 }, { "epoch": 0.16913766220037527, "grad_norm": 1.6276901960372925, "learning_rate": 1.900791942548714e-05, "loss": 1.2503, "step": 4800 }, { "epoch": 0.1691728992133337, "grad_norm": 1.2694061994552612, "learning_rate": 1.9007423766302867e-05, "loss": 1.0717, "step": 4801 }, { "epoch": 0.1692081362262921, "grad_norm": 1.2114660739898682, "learning_rate": 1.9006927989795753e-05, "loss": 1.3081, "step": 4802 }, { "epoch": 0.1692433732392505, "grad_norm": 2.0409209728240967, "learning_rate": 1.9006432095972257e-05, "loss": 2.9846, "step": 4803 }, { "epoch": 0.16927861025220892, "grad_norm": 1.5361369848251343, "learning_rate": 1.9005936084838832e-05, "loss": 1.2435, "step": 4804 }, { "epoch": 0.16931384726516732, "grad_norm": 5.339228630065918, "learning_rate": 1.9005439956401946e-05, "loss": 5.3839, "step": 4805 }, { "epoch": 0.16934908427812576, "grad_norm": 1.5929316282272339, "learning_rate": 1.9004943710668055e-05, "loss": 0.8385, "step": 4806 }, { "epoch": 0.16938432129108416, "grad_norm": 4.590425491333008, "learning_rate": 1.9004447347643625e-05, "loss": 3.7282, "step": 4807 }, { "epoch": 0.16941955830404257, "grad_norm": 4.896275997161865, "learning_rate": 1.900395086733512e-05, "loss": 5.2059, "step": 4808 }, { "epoch": 0.16945479531700097, "grad_norm": 1.287474513053894, "learning_rate": 1.9003454269749012e-05, "loss": 0.9961, "step": 4809 }, { "epoch": 0.16949003232995938, "grad_norm": 3.212991952896118, "learning_rate": 1.9002957554891765e-05, "loss": 2.7743, "step": 4810 }, { "epoch": 0.1695252693429178, "grad_norm": 1.273993730545044, "learning_rate": 1.9002460722769846e-05, "loss": 0.9744, "step": 4811 }, { "epoch": 0.16956050635587622, "grad_norm": 6.974312782287598, "learning_rate": 1.900196377338973e-05, "loss": 5.6387, "step": 4812 }, { "epoch": 0.16959574336883462, "grad_norm": 3.1049728393554688, "learning_rate": 1.9001466706757892e-05, "loss": 2.6562, "step": 4813 }, { "epoch": 0.16963098038179303, "grad_norm": 3.382227897644043, "learning_rate": 1.9000969522880802e-05, "loss": 3.7884, "step": 4814 }, { "epoch": 0.16966621739475143, "grad_norm": 1.0851114988327026, "learning_rate": 1.9000472221764934e-05, "loss": 1.241, "step": 4815 }, { "epoch": 0.16970145440770987, "grad_norm": 6.440345287322998, "learning_rate": 1.899997480341677e-05, "loss": 3.874, "step": 4816 }, { "epoch": 0.16973669142066827, "grad_norm": 6.841558456420898, "learning_rate": 1.8999477267842792e-05, "loss": 7.3042, "step": 4817 }, { "epoch": 0.16977192843362668, "grad_norm": 1.313233494758606, "learning_rate": 1.8998979615049473e-05, "loss": 0.9223, "step": 4818 }, { "epoch": 0.16980716544658508, "grad_norm": 5.929816722869873, "learning_rate": 1.89984818450433e-05, "loss": 3.1572, "step": 4819 }, { "epoch": 0.16984240245954352, "grad_norm": 3.61864972114563, "learning_rate": 1.8997983957830752e-05, "loss": 3.3016, "step": 4820 }, { "epoch": 0.16987763947250192, "grad_norm": 6.923081398010254, "learning_rate": 1.8997485953418318e-05, "loss": 5.6997, "step": 4821 }, { "epoch": 0.16991287648546033, "grad_norm": 3.931710958480835, "learning_rate": 1.8996987831812482e-05, "loss": 3.9138, "step": 4822 }, { "epoch": 0.16994811349841873, "grad_norm": 5.5581231117248535, "learning_rate": 1.8996489593019737e-05, "loss": 3.9632, "step": 4823 }, { "epoch": 0.16998335051137714, "grad_norm": 1.2360516786575317, "learning_rate": 1.8995991237046566e-05, "loss": 0.9929, "step": 4824 }, { "epoch": 0.17001858752433557, "grad_norm": 4.341247081756592, "learning_rate": 1.8995492763899465e-05, "loss": 8.5528, "step": 4825 }, { "epoch": 0.17005382453729398, "grad_norm": 3.584247589111328, "learning_rate": 1.8994994173584926e-05, "loss": 3.6446, "step": 4826 }, { "epoch": 0.17008906155025239, "grad_norm": 4.251900672912598, "learning_rate": 1.899449546610944e-05, "loss": 6.543, "step": 4827 }, { "epoch": 0.1701242985632108, "grad_norm": 2.4637906551361084, "learning_rate": 1.8993996641479505e-05, "loss": 3.6811, "step": 4828 }, { "epoch": 0.1701595355761692, "grad_norm": 3.9289517402648926, "learning_rate": 1.899349769970162e-05, "loss": 3.5655, "step": 4829 }, { "epoch": 0.17019477258912763, "grad_norm": 4.928002834320068, "learning_rate": 1.899299864078228e-05, "loss": 5.7985, "step": 4830 }, { "epoch": 0.17023000960208604, "grad_norm": 1.827206015586853, "learning_rate": 1.8992499464727986e-05, "loss": 0.9186, "step": 4831 }, { "epoch": 0.17026524661504444, "grad_norm": 1.6048552989959717, "learning_rate": 1.8992000171545243e-05, "loss": 0.9151, "step": 4832 }, { "epoch": 0.17030048362800285, "grad_norm": 7.5342936515808105, "learning_rate": 1.8991500761240554e-05, "loss": 3.7408, "step": 4833 }, { "epoch": 0.17033572064096125, "grad_norm": 1.2505546808242798, "learning_rate": 1.8991001233820417e-05, "loss": 0.9463, "step": 4834 }, { "epoch": 0.17037095765391969, "grad_norm": 7.9763360023498535, "learning_rate": 1.8990501589291348e-05, "loss": 5.5118, "step": 4835 }, { "epoch": 0.1704061946668781, "grad_norm": 1.514540433883667, "learning_rate": 1.899000182765985e-05, "loss": 0.9638, "step": 4836 }, { "epoch": 0.1704414316798365, "grad_norm": 1.099738597869873, "learning_rate": 1.898950194893243e-05, "loss": 0.9794, "step": 4837 }, { "epoch": 0.1704766686927949, "grad_norm": 1.1358250379562378, "learning_rate": 1.8989001953115608e-05, "loss": 1.2692, "step": 4838 }, { "epoch": 0.17051190570575334, "grad_norm": 1.8293046951293945, "learning_rate": 1.8988501840215886e-05, "loss": 3.2494, "step": 4839 }, { "epoch": 0.17054714271871174, "grad_norm": 1.0821659564971924, "learning_rate": 1.8988001610239782e-05, "loss": 1.1043, "step": 4840 }, { "epoch": 0.17058237973167015, "grad_norm": 1.6814320087432861, "learning_rate": 1.8987501263193815e-05, "loss": 0.9083, "step": 4841 }, { "epoch": 0.17061761674462855, "grad_norm": 3.4441914558410645, "learning_rate": 1.89870007990845e-05, "loss": 3.288, "step": 4842 }, { "epoch": 0.17065285375758696, "grad_norm": 6.482299327850342, "learning_rate": 1.898650021791835e-05, "loss": 3.1428, "step": 4843 }, { "epoch": 0.1706880907705454, "grad_norm": 1.7407819032669067, "learning_rate": 1.8985999519701894e-05, "loss": 1.0961, "step": 4844 }, { "epoch": 0.1707233277835038, "grad_norm": 1.3476327657699585, "learning_rate": 1.898549870444165e-05, "loss": 1.132, "step": 4845 }, { "epoch": 0.1707585647964622, "grad_norm": 4.973916053771973, "learning_rate": 1.898499777214414e-05, "loss": 3.5875, "step": 4846 }, { "epoch": 0.1707938018094206, "grad_norm": 6.089395046234131, "learning_rate": 1.8984496722815887e-05, "loss": 3.4566, "step": 4847 }, { "epoch": 0.17082903882237901, "grad_norm": 1.0654934644699097, "learning_rate": 1.898399555646342e-05, "loss": 1.1224, "step": 4848 }, { "epoch": 0.17086427583533745, "grad_norm": 5.279818058013916, "learning_rate": 1.8983494273093267e-05, "loss": 3.8279, "step": 4849 }, { "epoch": 0.17089951284829585, "grad_norm": 0.9351080656051636, "learning_rate": 1.898299287271196e-05, "loss": 0.9768, "step": 4850 }, { "epoch": 0.17093474986125426, "grad_norm": 7.810608386993408, "learning_rate": 1.8982491355326024e-05, "loss": 6.5734, "step": 4851 }, { "epoch": 0.17096998687421266, "grad_norm": 9.913524627685547, "learning_rate": 1.8981989720941992e-05, "loss": 9.4826, "step": 4852 }, { "epoch": 0.1710052238871711, "grad_norm": 6.158127784729004, "learning_rate": 1.8981487969566403e-05, "loss": 3.8258, "step": 4853 }, { "epoch": 0.1710404609001295, "grad_norm": 4.968443393707275, "learning_rate": 1.898098610120579e-05, "loss": 3.7488, "step": 4854 }, { "epoch": 0.1710756979130879, "grad_norm": 5.531453609466553, "learning_rate": 1.8980484115866686e-05, "loss": 4.5846, "step": 4855 }, { "epoch": 0.17111093492604632, "grad_norm": 1.4303967952728271, "learning_rate": 1.8979982013555633e-05, "loss": 1.0904, "step": 4856 }, { "epoch": 0.17114617193900472, "grad_norm": 1.2456928491592407, "learning_rate": 1.8979479794279172e-05, "loss": 0.9833, "step": 4857 }, { "epoch": 0.17118140895196315, "grad_norm": 2.2181718349456787, "learning_rate": 1.897897745804384e-05, "loss": 3.4088, "step": 4858 }, { "epoch": 0.17121664596492156, "grad_norm": 4.620181560516357, "learning_rate": 1.897847500485619e-05, "loss": 6.5118, "step": 4859 }, { "epoch": 0.17125188297787997, "grad_norm": 1.908649206161499, "learning_rate": 1.897797243472275e-05, "loss": 1.1363, "step": 4860 }, { "epoch": 0.17128711999083837, "grad_norm": 2.450985908508301, "learning_rate": 1.897746974765008e-05, "loss": 3.9851, "step": 4861 }, { "epoch": 0.17132235700379678, "grad_norm": 2.4353742599487305, "learning_rate": 1.8976966943644724e-05, "loss": 3.8342, "step": 4862 }, { "epoch": 0.1713575940167552, "grad_norm": 1.2102354764938354, "learning_rate": 1.8976464022713228e-05, "loss": 1.2122, "step": 4863 }, { "epoch": 0.17139283102971362, "grad_norm": 3.1460328102111816, "learning_rate": 1.8975960984862145e-05, "loss": 7.6775, "step": 4864 }, { "epoch": 0.17142806804267202, "grad_norm": 2.8308591842651367, "learning_rate": 1.897545783009803e-05, "loss": 3.9653, "step": 4865 }, { "epoch": 0.17146330505563043, "grad_norm": 1.6491172313690186, "learning_rate": 1.8974954558427427e-05, "loss": 1.1799, "step": 4866 }, { "epoch": 0.17149854206858883, "grad_norm": 2.267057418823242, "learning_rate": 1.89744511698569e-05, "loss": 3.9584, "step": 4867 }, { "epoch": 0.17153377908154727, "grad_norm": 3.569537401199341, "learning_rate": 1.8973947664393006e-05, "loss": 3.3561, "step": 4868 }, { "epoch": 0.17156901609450567, "grad_norm": 1.6360504627227783, "learning_rate": 1.89734440420423e-05, "loss": 1.2378, "step": 4869 }, { "epoch": 0.17160425310746408, "grad_norm": 1.379067301750183, "learning_rate": 1.8972940302811343e-05, "loss": 0.9011, "step": 4870 }, { "epoch": 0.17163949012042248, "grad_norm": 1.6044561862945557, "learning_rate": 1.897243644670669e-05, "loss": 0.9809, "step": 4871 }, { "epoch": 0.17167472713338092, "grad_norm": 3.172119617462158, "learning_rate": 1.8971932473734917e-05, "loss": 6.1004, "step": 4872 }, { "epoch": 0.17170996414633932, "grad_norm": 3.3690757751464844, "learning_rate": 1.8971428383902575e-05, "loss": 6.9482, "step": 4873 }, { "epoch": 0.17174520115929773, "grad_norm": 2.0763232707977295, "learning_rate": 1.897092417721624e-05, "loss": 3.4957, "step": 4874 }, { "epoch": 0.17178043817225613, "grad_norm": 1.954852819442749, "learning_rate": 1.897041985368247e-05, "loss": 3.3996, "step": 4875 }, { "epoch": 0.17181567518521454, "grad_norm": 5.956501007080078, "learning_rate": 1.8969915413307843e-05, "loss": 9.02, "step": 4876 }, { "epoch": 0.17185091219817297, "grad_norm": 1.4080231189727783, "learning_rate": 1.8969410856098923e-05, "loss": 1.0701, "step": 4877 }, { "epoch": 0.17188614921113138, "grad_norm": 3.216977119445801, "learning_rate": 1.8968906182062285e-05, "loss": 4.2647, "step": 4878 }, { "epoch": 0.17192138622408978, "grad_norm": 1.245534896850586, "learning_rate": 1.89684013912045e-05, "loss": 1.0088, "step": 4879 }, { "epoch": 0.1719566232370482, "grad_norm": 1.2666627168655396, "learning_rate": 1.8967896483532147e-05, "loss": 0.8857, "step": 4880 }, { "epoch": 0.1719918602500066, "grad_norm": 0.9374297857284546, "learning_rate": 1.89673914590518e-05, "loss": 0.8895, "step": 4881 }, { "epoch": 0.17202709726296503, "grad_norm": 1.4053415060043335, "learning_rate": 1.8966886317770036e-05, "loss": 0.8533, "step": 4882 }, { "epoch": 0.17206233427592343, "grad_norm": 2.141831398010254, "learning_rate": 1.8966381059693434e-05, "loss": 1.1305, "step": 4883 }, { "epoch": 0.17209757128888184, "grad_norm": 5.160594463348389, "learning_rate": 1.8965875684828576e-05, "loss": 6.045, "step": 4884 }, { "epoch": 0.17213280830184025, "grad_norm": 2.094111680984497, "learning_rate": 1.8965370193182046e-05, "loss": 3.6698, "step": 4885 }, { "epoch": 0.17216804531479865, "grad_norm": 5.298876762390137, "learning_rate": 1.896486458476043e-05, "loss": 6.3184, "step": 4886 }, { "epoch": 0.17220328232775708, "grad_norm": 2.6743452548980713, "learning_rate": 1.8964358859570308e-05, "loss": 4.0766, "step": 4887 }, { "epoch": 0.1722385193407155, "grad_norm": 1.1715967655181885, "learning_rate": 1.896385301761827e-05, "loss": 1.2209, "step": 4888 }, { "epoch": 0.1722737563536739, "grad_norm": 4.303424835205078, "learning_rate": 1.8963347058910905e-05, "loss": 4.9076, "step": 4889 }, { "epoch": 0.1723089933666323, "grad_norm": 4.2038726806640625, "learning_rate": 1.8962840983454803e-05, "loss": 6.0697, "step": 4890 }, { "epoch": 0.17234423037959073, "grad_norm": 6.2230119705200195, "learning_rate": 1.896233479125656e-05, "loss": 7.054, "step": 4891 }, { "epoch": 0.17237946739254914, "grad_norm": 1.2330905199050903, "learning_rate": 1.8961828482322758e-05, "loss": 0.9564, "step": 4892 }, { "epoch": 0.17241470440550755, "grad_norm": 4.125949382781982, "learning_rate": 1.896132205666e-05, "loss": 3.4498, "step": 4893 }, { "epoch": 0.17244994141846595, "grad_norm": 1.3117074966430664, "learning_rate": 1.8960815514274882e-05, "loss": 0.8555, "step": 4894 }, { "epoch": 0.17248517843142436, "grad_norm": 1.1291557550430298, "learning_rate": 1.8960308855174002e-05, "loss": 0.921, "step": 4895 }, { "epoch": 0.1725204154443828, "grad_norm": 4.029538154602051, "learning_rate": 1.8959802079363955e-05, "loss": 6.9176, "step": 4896 }, { "epoch": 0.1725556524573412, "grad_norm": 1.1458380222320557, "learning_rate": 1.8959295186851344e-05, "loss": 0.9374, "step": 4897 }, { "epoch": 0.1725908894702996, "grad_norm": 2.876336097717285, "learning_rate": 1.895878817764277e-05, "loss": 3.649, "step": 4898 }, { "epoch": 0.172626126483258, "grad_norm": 6.403061866760254, "learning_rate": 1.8958281051744845e-05, "loss": 5.518, "step": 4899 }, { "epoch": 0.1726613634962164, "grad_norm": 1.4887458086013794, "learning_rate": 1.8957773809164167e-05, "loss": 0.9328, "step": 4900 }, { "epoch": 0.17269660050917485, "grad_norm": 1.3303121328353882, "learning_rate": 1.895726644990734e-05, "loss": 1.028, "step": 4901 }, { "epoch": 0.17273183752213325, "grad_norm": 3.5864267349243164, "learning_rate": 1.895675897398098e-05, "loss": 3.1921, "step": 4902 }, { "epoch": 0.17276707453509166, "grad_norm": 2.5181291103363037, "learning_rate": 1.8956251381391695e-05, "loss": 3.3685, "step": 4903 }, { "epoch": 0.17280231154805006, "grad_norm": 3.531933546066284, "learning_rate": 1.8955743672146088e-05, "loss": 3.3989, "step": 4904 }, { "epoch": 0.1728375485610085, "grad_norm": 3.6474881172180176, "learning_rate": 1.8955235846250787e-05, "loss": 4.1681, "step": 4905 }, { "epoch": 0.1728727855739669, "grad_norm": 1.3439518213272095, "learning_rate": 1.8954727903712395e-05, "loss": 0.9158, "step": 4906 }, { "epoch": 0.1729080225869253, "grad_norm": 4.719456672668457, "learning_rate": 1.895421984453753e-05, "loss": 4.1219, "step": 4907 }, { "epoch": 0.17294325959988371, "grad_norm": 2.421895742416382, "learning_rate": 1.895371166873281e-05, "loss": 3.8797, "step": 4908 }, { "epoch": 0.17297849661284212, "grad_norm": 5.619711399078369, "learning_rate": 1.8953203376304858e-05, "loss": 3.5795, "step": 4909 }, { "epoch": 0.17301373362580055, "grad_norm": 6.2442097663879395, "learning_rate": 1.895269496726029e-05, "loss": 3.4912, "step": 4910 }, { "epoch": 0.17304897063875896, "grad_norm": 3.146287679672241, "learning_rate": 1.8952186441605733e-05, "loss": 3.49, "step": 4911 }, { "epoch": 0.17308420765171736, "grad_norm": 2.885300874710083, "learning_rate": 1.89516777993478e-05, "loss": 3.8498, "step": 4912 }, { "epoch": 0.17311944466467577, "grad_norm": 1.3514810800552368, "learning_rate": 1.8951169040493133e-05, "loss": 1.217, "step": 4913 }, { "epoch": 0.17315468167763418, "grad_norm": 4.161513328552246, "learning_rate": 1.8950660165048346e-05, "loss": 6.2998, "step": 4914 }, { "epoch": 0.1731899186905926, "grad_norm": 3.809248447418213, "learning_rate": 1.8950151173020065e-05, "loss": 3.8808, "step": 4915 }, { "epoch": 0.17322515570355101, "grad_norm": 2.4735028743743896, "learning_rate": 1.894964206441493e-05, "loss": 3.2736, "step": 4916 }, { "epoch": 0.17326039271650942, "grad_norm": 1.4321526288986206, "learning_rate": 1.8949132839239562e-05, "loss": 0.8735, "step": 4917 }, { "epoch": 0.17329562972946783, "grad_norm": 1.111638069152832, "learning_rate": 1.8948623497500603e-05, "loss": 1.0935, "step": 4918 }, { "epoch": 0.17333086674242623, "grad_norm": 3.8254363536834717, "learning_rate": 1.8948114039204684e-05, "loss": 6.6478, "step": 4919 }, { "epoch": 0.17336610375538467, "grad_norm": 4.79954195022583, "learning_rate": 1.8947604464358436e-05, "loss": 6.7799, "step": 4920 }, { "epoch": 0.17340134076834307, "grad_norm": 1.4217555522918701, "learning_rate": 1.8947094772968505e-05, "loss": 0.9023, "step": 4921 }, { "epoch": 0.17343657778130148, "grad_norm": 1.0989995002746582, "learning_rate": 1.8946584965041522e-05, "loss": 1.1115, "step": 4922 }, { "epoch": 0.17347181479425988, "grad_norm": 3.029971122741699, "learning_rate": 1.894607504058413e-05, "loss": 3.8641, "step": 4923 }, { "epoch": 0.17350705180721832, "grad_norm": 3.259075164794922, "learning_rate": 1.894556499960297e-05, "loss": 3.5911, "step": 4924 }, { "epoch": 0.17354228882017672, "grad_norm": 1.1066789627075195, "learning_rate": 1.894505484210469e-05, "loss": 1.0238, "step": 4925 }, { "epoch": 0.17357752583313513, "grad_norm": 3.571577787399292, "learning_rate": 1.8944544568095926e-05, "loss": 3.9157, "step": 4926 }, { "epoch": 0.17361276284609353, "grad_norm": 4.212810039520264, "learning_rate": 1.8944034177583334e-05, "loss": 2.9339, "step": 4927 }, { "epoch": 0.17364799985905194, "grad_norm": 5.645987033843994, "learning_rate": 1.8943523670573557e-05, "loss": 6.8846, "step": 4928 }, { "epoch": 0.17368323687201037, "grad_norm": 3.5601565837860107, "learning_rate": 1.8943013047073243e-05, "loss": 4.143, "step": 4929 }, { "epoch": 0.17371847388496878, "grad_norm": 4.40812873840332, "learning_rate": 1.8942502307089046e-05, "loss": 3.744, "step": 4930 }, { "epoch": 0.17375371089792718, "grad_norm": 2.902348041534424, "learning_rate": 1.8941991450627618e-05, "loss": 3.7542, "step": 4931 }, { "epoch": 0.1737889479108856, "grad_norm": 3.676740884780884, "learning_rate": 1.894148047769561e-05, "loss": 3.9357, "step": 4932 }, { "epoch": 0.173824184923844, "grad_norm": 4.568501949310303, "learning_rate": 1.8940969388299685e-05, "loss": 2.9856, "step": 4933 }, { "epoch": 0.17385942193680243, "grad_norm": 1.3226630687713623, "learning_rate": 1.894045818244649e-05, "loss": 0.9757, "step": 4934 }, { "epoch": 0.17389465894976083, "grad_norm": 4.254665374755859, "learning_rate": 1.8939946860142687e-05, "loss": 3.7206, "step": 4935 }, { "epoch": 0.17392989596271924, "grad_norm": 5.41311502456665, "learning_rate": 1.893943542139494e-05, "loss": 2.6155, "step": 4936 }, { "epoch": 0.17396513297567764, "grad_norm": 2.9821841716766357, "learning_rate": 1.893892386620991e-05, "loss": 3.3364, "step": 4937 }, { "epoch": 0.17400036998863605, "grad_norm": 2.6740870475769043, "learning_rate": 1.8938412194594257e-05, "loss": 3.2381, "step": 4938 }, { "epoch": 0.17403560700159448, "grad_norm": 4.317339897155762, "learning_rate": 1.8937900406554646e-05, "loss": 3.5588, "step": 4939 }, { "epoch": 0.1740708440145529, "grad_norm": 1.3426941633224487, "learning_rate": 1.8937388502097743e-05, "loss": 1.1616, "step": 4940 }, { "epoch": 0.1741060810275113, "grad_norm": 2.4281258583068848, "learning_rate": 1.8936876481230215e-05, "loss": 3.7083, "step": 4941 }, { "epoch": 0.1741413180404697, "grad_norm": 4.459590435028076, "learning_rate": 1.8936364343958737e-05, "loss": 3.3101, "step": 4942 }, { "epoch": 0.17417655505342813, "grad_norm": 4.172694683074951, "learning_rate": 1.893585209028997e-05, "loss": 5.61, "step": 4943 }, { "epoch": 0.17421179206638654, "grad_norm": 1.452248454093933, "learning_rate": 1.8935339720230592e-05, "loss": 1.1891, "step": 4944 }, { "epoch": 0.17424702907934495, "grad_norm": 1.0888160467147827, "learning_rate": 1.8934827233787278e-05, "loss": 1.0877, "step": 4945 }, { "epoch": 0.17428226609230335, "grad_norm": 2.3508169651031494, "learning_rate": 1.8934314630966702e-05, "loss": 4.2622, "step": 4946 }, { "epoch": 0.17431750310526176, "grad_norm": 1.422244668006897, "learning_rate": 1.893380191177554e-05, "loss": 0.8217, "step": 4947 }, { "epoch": 0.1743527401182202, "grad_norm": 1.5617367029190063, "learning_rate": 1.8933289076220463e-05, "loss": 0.8928, "step": 4948 }, { "epoch": 0.1743879771311786, "grad_norm": 1.3467025756835938, "learning_rate": 1.8932776124308167e-05, "loss": 1.1506, "step": 4949 }, { "epoch": 0.174423214144137, "grad_norm": 3.4573395252227783, "learning_rate": 1.8932263056045318e-05, "loss": 3.9226, "step": 4950 }, { "epoch": 0.1744584511570954, "grad_norm": 5.421790599822998, "learning_rate": 1.893174987143861e-05, "loss": 5.7059, "step": 4951 }, { "epoch": 0.1744936881700538, "grad_norm": 1.1479971408843994, "learning_rate": 1.893123657049472e-05, "loss": 1.0665, "step": 4952 }, { "epoch": 0.17452892518301225, "grad_norm": 2.4035756587982178, "learning_rate": 1.8930723153220336e-05, "loss": 3.0536, "step": 4953 }, { "epoch": 0.17456416219597065, "grad_norm": 1.6805369853973389, "learning_rate": 1.8930209619622143e-05, "loss": 0.7841, "step": 4954 }, { "epoch": 0.17459939920892906, "grad_norm": 3.595717668533325, "learning_rate": 1.8929695969706837e-05, "loss": 3.1367, "step": 4955 }, { "epoch": 0.17463463622188746, "grad_norm": 10.9033203125, "learning_rate": 1.89291822034811e-05, "loss": 6.9137, "step": 4956 }, { "epoch": 0.17466987323484587, "grad_norm": 6.593461036682129, "learning_rate": 1.892866832095163e-05, "loss": 5.3541, "step": 4957 }, { "epoch": 0.1747051102478043, "grad_norm": 7.066381454467773, "learning_rate": 1.8928154322125116e-05, "loss": 3.7347, "step": 4958 }, { "epoch": 0.1747403472607627, "grad_norm": 1.2516001462936401, "learning_rate": 1.892764020700825e-05, "loss": 0.9995, "step": 4959 }, { "epoch": 0.1747755842737211, "grad_norm": 1.2434682846069336, "learning_rate": 1.892712597560774e-05, "loss": 1.0598, "step": 4960 }, { "epoch": 0.17481082128667952, "grad_norm": 4.796553134918213, "learning_rate": 1.8926611627930277e-05, "loss": 5.7302, "step": 4961 }, { "epoch": 0.17484605829963795, "grad_norm": 1.3676403760910034, "learning_rate": 1.8926097163982556e-05, "loss": 0.9636, "step": 4962 }, { "epoch": 0.17488129531259636, "grad_norm": 1.418010950088501, "learning_rate": 1.8925582583771283e-05, "loss": 1.0032, "step": 4963 }, { "epoch": 0.17491653232555476, "grad_norm": 6.482248783111572, "learning_rate": 1.8925067887303163e-05, "loss": 5.592, "step": 4964 }, { "epoch": 0.17495176933851317, "grad_norm": 1.8519707918167114, "learning_rate": 1.8924553074584896e-05, "loss": 1.0648, "step": 4965 }, { "epoch": 0.17498700635147157, "grad_norm": 1.77716064453125, "learning_rate": 1.8924038145623186e-05, "loss": 3.1923, "step": 4966 }, { "epoch": 0.17502224336443, "grad_norm": 1.4083393812179565, "learning_rate": 1.8923523100424744e-05, "loss": 0.9005, "step": 4967 }, { "epoch": 0.1750574803773884, "grad_norm": 2.5518405437469482, "learning_rate": 1.8923007938996277e-05, "loss": 4.5269, "step": 4968 }, { "epoch": 0.17509271739034682, "grad_norm": 4.804460525512695, "learning_rate": 1.89224926613445e-05, "loss": 3.5301, "step": 4969 }, { "epoch": 0.17512795440330522, "grad_norm": 1.4852280616760254, "learning_rate": 1.892197726747611e-05, "loss": 1.1455, "step": 4970 }, { "epoch": 0.17516319141626363, "grad_norm": 1.2713286876678467, "learning_rate": 1.8921461757397835e-05, "loss": 0.9151, "step": 4971 }, { "epoch": 0.17519842842922206, "grad_norm": 5.7001800537109375, "learning_rate": 1.8920946131116383e-05, "loss": 7.967, "step": 4972 }, { "epoch": 0.17523366544218047, "grad_norm": 1.9352266788482666, "learning_rate": 1.8920430388638474e-05, "loss": 0.9518, "step": 4973 }, { "epoch": 0.17526890245513888, "grad_norm": 5.290163516998291, "learning_rate": 1.891991452997082e-05, "loss": 3.7128, "step": 4974 }, { "epoch": 0.17530413946809728, "grad_norm": 1.1798996925354004, "learning_rate": 1.8919398555120144e-05, "loss": 1.2401, "step": 4975 }, { "epoch": 0.17533937648105571, "grad_norm": 1.8664430379867554, "learning_rate": 1.8918882464093164e-05, "loss": 1.006, "step": 4976 }, { "epoch": 0.17537461349401412, "grad_norm": 6.101049900054932, "learning_rate": 1.8918366256896608e-05, "loss": 3.7219, "step": 4977 }, { "epoch": 0.17540985050697253, "grad_norm": 1.4229240417480469, "learning_rate": 1.891784993353719e-05, "loss": 1.1531, "step": 4978 }, { "epoch": 0.17544508751993093, "grad_norm": 1.499650239944458, "learning_rate": 1.8917333494021646e-05, "loss": 0.8265, "step": 4979 }, { "epoch": 0.17548032453288934, "grad_norm": 3.2877767086029053, "learning_rate": 1.891681693835669e-05, "loss": 4.1067, "step": 4980 }, { "epoch": 0.17551556154584777, "grad_norm": 3.6497957706451416, "learning_rate": 1.8916300266549065e-05, "loss": 5.2474, "step": 4981 }, { "epoch": 0.17555079855880618, "grad_norm": 3.4016458988189697, "learning_rate": 1.891578347860549e-05, "loss": 3.5035, "step": 4982 }, { "epoch": 0.17558603557176458, "grad_norm": 4.800065040588379, "learning_rate": 1.89152665745327e-05, "loss": 3.6837, "step": 4983 }, { "epoch": 0.175621272584723, "grad_norm": 1.3517920970916748, "learning_rate": 1.891474955433743e-05, "loss": 1.0026, "step": 4984 }, { "epoch": 0.1756565095976814, "grad_norm": 6.389777183532715, "learning_rate": 1.891423241802641e-05, "loss": 4.2152, "step": 4985 }, { "epoch": 0.17569174661063983, "grad_norm": 1.5467795133590698, "learning_rate": 1.8913715165606375e-05, "loss": 0.7771, "step": 4986 }, { "epoch": 0.17572698362359823, "grad_norm": 2.355329751968384, "learning_rate": 1.8913197797084064e-05, "loss": 3.9495, "step": 4987 }, { "epoch": 0.17576222063655664, "grad_norm": 4.857553005218506, "learning_rate": 1.8912680312466217e-05, "loss": 5.9647, "step": 4988 }, { "epoch": 0.17579745764951504, "grad_norm": 3.3253469467163086, "learning_rate": 1.8912162711759578e-05, "loss": 3.6626, "step": 4989 }, { "epoch": 0.17583269466247345, "grad_norm": 5.8791022300720215, "learning_rate": 1.8911644994970877e-05, "loss": 2.4964, "step": 4990 }, { "epoch": 0.17586793167543188, "grad_norm": 1.3053544759750366, "learning_rate": 1.891112716210687e-05, "loss": 1.0754, "step": 4991 }, { "epoch": 0.1759031686883903, "grad_norm": 2.209578037261963, "learning_rate": 1.8910609213174296e-05, "loss": 2.4002, "step": 4992 }, { "epoch": 0.1759384057013487, "grad_norm": 3.1385550498962402, "learning_rate": 1.8910091148179902e-05, "loss": 3.7845, "step": 4993 }, { "epoch": 0.1759736427143071, "grad_norm": 4.898540019989014, "learning_rate": 1.8909572967130435e-05, "loss": 6.3327, "step": 4994 }, { "epoch": 0.17600887972726553, "grad_norm": 1.096070408821106, "learning_rate": 1.8909054670032646e-05, "loss": 0.9687, "step": 4995 }, { "epoch": 0.17604411674022394, "grad_norm": 1.5030391216278076, "learning_rate": 1.8908536256893284e-05, "loss": 0.9449, "step": 4996 }, { "epoch": 0.17607935375318234, "grad_norm": 8.08953857421875, "learning_rate": 1.8908017727719103e-05, "loss": 6.9802, "step": 4997 }, { "epoch": 0.17611459076614075, "grad_norm": 5.078701496124268, "learning_rate": 1.8907499082516857e-05, "loss": 3.8766, "step": 4998 }, { "epoch": 0.17614982777909916, "grad_norm": 5.49672269821167, "learning_rate": 1.89069803212933e-05, "loss": 3.6983, "step": 4999 }, { "epoch": 0.1761850647920576, "grad_norm": 4.415379047393799, "learning_rate": 1.8906461444055192e-05, "loss": 3.1168, "step": 5000 }, { "epoch": 0.176220301805016, "grad_norm": 1.9224382638931274, "learning_rate": 1.8905942450809284e-05, "loss": 0.8271, "step": 5001 }, { "epoch": 0.1762555388179744, "grad_norm": 1.0350277423858643, "learning_rate": 1.8905423341562346e-05, "loss": 1.0789, "step": 5002 }, { "epoch": 0.1762907758309328, "grad_norm": 2.482027769088745, "learning_rate": 1.890490411632113e-05, "loss": 0.9157, "step": 5003 }, { "epoch": 0.1763260128438912, "grad_norm": 1.145590901374817, "learning_rate": 1.8904384775092407e-05, "loss": 1.0459, "step": 5004 }, { "epoch": 0.17636124985684964, "grad_norm": 1.3574877977371216, "learning_rate": 1.890386531788294e-05, "loss": 0.8773, "step": 5005 }, { "epoch": 0.17639648686980805, "grad_norm": 8.789825439453125, "learning_rate": 1.890334574469949e-05, "loss": 2.9575, "step": 5006 }, { "epoch": 0.17643172388276646, "grad_norm": 6.063044548034668, "learning_rate": 1.890282605554883e-05, "loss": 3.7729, "step": 5007 }, { "epoch": 0.17646696089572486, "grad_norm": 3.6180076599121094, "learning_rate": 1.8902306250437722e-05, "loss": 3.6923, "step": 5008 }, { "epoch": 0.17650219790868327, "grad_norm": 1.2690564393997192, "learning_rate": 1.8901786329372943e-05, "loss": 0.831, "step": 5009 }, { "epoch": 0.1765374349216417, "grad_norm": 6.905923366546631, "learning_rate": 1.8901266292361267e-05, "loss": 6.1507, "step": 5010 }, { "epoch": 0.1765726719346001, "grad_norm": 1.3614082336425781, "learning_rate": 1.890074613940946e-05, "loss": 1.0974, "step": 5011 }, { "epoch": 0.1766079089475585, "grad_norm": 0.9772441983222961, "learning_rate": 1.89002258705243e-05, "loss": 1.1065, "step": 5012 }, { "epoch": 0.17664314596051692, "grad_norm": 7.021068572998047, "learning_rate": 1.8899705485712566e-05, "loss": 5.8106, "step": 5013 }, { "epoch": 0.17667838297347535, "grad_norm": 9.16516399383545, "learning_rate": 1.8899184984981034e-05, "loss": 12.6625, "step": 5014 }, { "epoch": 0.17671361998643376, "grad_norm": 8.191919326782227, "learning_rate": 1.8898664368336486e-05, "loss": 5.8198, "step": 5015 }, { "epoch": 0.17674885699939216, "grad_norm": 1.222132682800293, "learning_rate": 1.88981436357857e-05, "loss": 0.8168, "step": 5016 }, { "epoch": 0.17678409401235057, "grad_norm": 4.037487030029297, "learning_rate": 1.8897622787335458e-05, "loss": 3.5538, "step": 5017 }, { "epoch": 0.17681933102530897, "grad_norm": 1.2457809448242188, "learning_rate": 1.8897101822992548e-05, "loss": 1.0879, "step": 5018 }, { "epoch": 0.1768545680382674, "grad_norm": 5.6883134841918945, "learning_rate": 1.8896580742763754e-05, "loss": 3.8343, "step": 5019 }, { "epoch": 0.1768898050512258, "grad_norm": 3.631070852279663, "learning_rate": 1.8896059546655858e-05, "loss": 3.3321, "step": 5020 }, { "epoch": 0.17692504206418422, "grad_norm": 5.027080535888672, "learning_rate": 1.8895538234675658e-05, "loss": 6.3532, "step": 5021 }, { "epoch": 0.17696027907714262, "grad_norm": 6.568057537078857, "learning_rate": 1.8895016806829936e-05, "loss": 6.1808, "step": 5022 }, { "epoch": 0.17699551609010103, "grad_norm": 1.30336332321167, "learning_rate": 1.8894495263125486e-05, "loss": 1.121, "step": 5023 }, { "epoch": 0.17703075310305946, "grad_norm": 1.3557451963424683, "learning_rate": 1.8893973603569107e-05, "loss": 0.9742, "step": 5024 }, { "epoch": 0.17706599011601787, "grad_norm": 1.4531680345535278, "learning_rate": 1.889345182816759e-05, "loss": 1.0518, "step": 5025 }, { "epoch": 0.17710122712897627, "grad_norm": 2.3715620040893555, "learning_rate": 1.8892929936927723e-05, "loss": 4.0824, "step": 5026 }, { "epoch": 0.17713646414193468, "grad_norm": 2.010664224624634, "learning_rate": 1.8892407929856315e-05, "loss": 1.0345, "step": 5027 }, { "epoch": 0.1771717011548931, "grad_norm": 2.3318614959716797, "learning_rate": 1.889188580696016e-05, "loss": 3.7317, "step": 5028 }, { "epoch": 0.17720693816785152, "grad_norm": 3.5838632583618164, "learning_rate": 1.889136356824606e-05, "loss": 3.9535, "step": 5029 }, { "epoch": 0.17724217518080992, "grad_norm": 1.106041669845581, "learning_rate": 1.889084121372082e-05, "loss": 1.0503, "step": 5030 }, { "epoch": 0.17727741219376833, "grad_norm": 2.0396132469177246, "learning_rate": 1.8890318743391236e-05, "loss": 4.1593, "step": 5031 }, { "epoch": 0.17731264920672674, "grad_norm": 1.2201493978500366, "learning_rate": 1.888979615726412e-05, "loss": 1.0635, "step": 5032 }, { "epoch": 0.17734788621968517, "grad_norm": 1.7103711366653442, "learning_rate": 1.8889273455346276e-05, "loss": 1.018, "step": 5033 }, { "epoch": 0.17738312323264357, "grad_norm": 1.064066767692566, "learning_rate": 1.8888750637644513e-05, "loss": 1.2106, "step": 5034 }, { "epoch": 0.17741836024560198, "grad_norm": 2.715823173522949, "learning_rate": 1.8888227704165642e-05, "loss": 3.8071, "step": 5035 }, { "epoch": 0.17745359725856039, "grad_norm": 1.3554494380950928, "learning_rate": 1.8887704654916472e-05, "loss": 0.6618, "step": 5036 }, { "epoch": 0.1774888342715188, "grad_norm": 5.9480180740356445, "learning_rate": 1.888718148990382e-05, "loss": 3.7635, "step": 5037 }, { "epoch": 0.17752407128447723, "grad_norm": 4.930188179016113, "learning_rate": 1.888665820913449e-05, "loss": 5.7804, "step": 5038 }, { "epoch": 0.17755930829743563, "grad_norm": 1.224090337753296, "learning_rate": 1.888613481261531e-05, "loss": 1.1357, "step": 5039 }, { "epoch": 0.17759454531039404, "grad_norm": 1.4671449661254883, "learning_rate": 1.888561130035309e-05, "loss": 1.0248, "step": 5040 }, { "epoch": 0.17762978232335244, "grad_norm": 2.7404940128326416, "learning_rate": 1.8885087672354653e-05, "loss": 1.0625, "step": 5041 }, { "epoch": 0.17766501933631085, "grad_norm": 1.104658603668213, "learning_rate": 1.8884563928626815e-05, "loss": 0.859, "step": 5042 }, { "epoch": 0.17770025634926928, "grad_norm": 1.4931312799453735, "learning_rate": 1.8884040069176402e-05, "loss": 1.0669, "step": 5043 }, { "epoch": 0.1777354933622277, "grad_norm": 4.733387470245361, "learning_rate": 1.8883516094010235e-05, "loss": 5.1909, "step": 5044 }, { "epoch": 0.1777707303751861, "grad_norm": 8.538698196411133, "learning_rate": 1.888299200313514e-05, "loss": 6.5806, "step": 5045 }, { "epoch": 0.1778059673881445, "grad_norm": 1.2602696418762207, "learning_rate": 1.888246779655794e-05, "loss": 1.3442, "step": 5046 }, { "epoch": 0.17784120440110293, "grad_norm": 4.395340442657471, "learning_rate": 1.8881943474285467e-05, "loss": 2.9955, "step": 5047 }, { "epoch": 0.17787644141406134, "grad_norm": 1.6460626125335693, "learning_rate": 1.8881419036324548e-05, "loss": 0.799, "step": 5048 }, { "epoch": 0.17791167842701974, "grad_norm": 1.6808311939239502, "learning_rate": 1.8880894482682017e-05, "loss": 1.0808, "step": 5049 }, { "epoch": 0.17794691543997815, "grad_norm": 1.4309207201004028, "learning_rate": 1.8880369813364704e-05, "loss": 1.0993, "step": 5050 }, { "epoch": 0.17798215245293655, "grad_norm": 4.58637809753418, "learning_rate": 1.887984502837944e-05, "loss": 3.8172, "step": 5051 }, { "epoch": 0.178017389465895, "grad_norm": 11.713080406188965, "learning_rate": 1.887932012773307e-05, "loss": 4.4721, "step": 5052 }, { "epoch": 0.1780526264788534, "grad_norm": 1.4026535749435425, "learning_rate": 1.8878795111432415e-05, "loss": 0.8066, "step": 5053 }, { "epoch": 0.1780878634918118, "grad_norm": 3.7418625354766846, "learning_rate": 1.887826997948433e-05, "loss": 2.9505, "step": 5054 }, { "epoch": 0.1781231005047702, "grad_norm": 1.1902966499328613, "learning_rate": 1.8877744731895644e-05, "loss": 1.0531, "step": 5055 }, { "epoch": 0.1781583375177286, "grad_norm": 6.646115779876709, "learning_rate": 1.8877219368673203e-05, "loss": 4.888, "step": 5056 }, { "epoch": 0.17819357453068704, "grad_norm": 3.681942939758301, "learning_rate": 1.8876693889823848e-05, "loss": 3.4895, "step": 5057 }, { "epoch": 0.17822881154364545, "grad_norm": 1.1897892951965332, "learning_rate": 1.8876168295354425e-05, "loss": 0.8793, "step": 5058 }, { "epoch": 0.17826404855660385, "grad_norm": 3.212475538253784, "learning_rate": 1.8875642585271778e-05, "loss": 5.3654, "step": 5059 }, { "epoch": 0.17829928556956226, "grad_norm": 2.8417134284973145, "learning_rate": 1.8875116759582757e-05, "loss": 3.6086, "step": 5060 }, { "epoch": 0.17833452258252067, "grad_norm": 3.5396533012390137, "learning_rate": 1.887459081829421e-05, "loss": 3.9378, "step": 5061 }, { "epoch": 0.1783697595954791, "grad_norm": 0.9519243836402893, "learning_rate": 1.8874064761412988e-05, "loss": 0.9833, "step": 5062 }, { "epoch": 0.1784049966084375, "grad_norm": 1.1572434902191162, "learning_rate": 1.8873538588945944e-05, "loss": 1.008, "step": 5063 }, { "epoch": 0.1784402336213959, "grad_norm": 1.4343563318252563, "learning_rate": 1.8873012300899923e-05, "loss": 1.069, "step": 5064 }, { "epoch": 0.17847547063435432, "grad_norm": 2.4972832202911377, "learning_rate": 1.8872485897281793e-05, "loss": 3.7753, "step": 5065 }, { "epoch": 0.17851070764731275, "grad_norm": 6.052775859832764, "learning_rate": 1.88719593780984e-05, "loss": 7.4231, "step": 5066 }, { "epoch": 0.17854594466027116, "grad_norm": 3.650639533996582, "learning_rate": 1.8871432743356607e-05, "loss": 3.7119, "step": 5067 }, { "epoch": 0.17858118167322956, "grad_norm": 2.4879310131073, "learning_rate": 1.8870905993063277e-05, "loss": 3.5055, "step": 5068 }, { "epoch": 0.17861641868618797, "grad_norm": 3.901780366897583, "learning_rate": 1.887037912722526e-05, "loss": 6.0976, "step": 5069 }, { "epoch": 0.17865165569914637, "grad_norm": 1.127278208732605, "learning_rate": 1.8869852145849427e-05, "loss": 0.9231, "step": 5070 }, { "epoch": 0.1786868927121048, "grad_norm": 1.6099787950515747, "learning_rate": 1.886932504894264e-05, "loss": 1.1239, "step": 5071 }, { "epoch": 0.1787221297250632, "grad_norm": 2.6898601055145264, "learning_rate": 1.8868797836511764e-05, "loss": 3.4424, "step": 5072 }, { "epoch": 0.17875736673802162, "grad_norm": 1.9061006307601929, "learning_rate": 1.8868270508563665e-05, "loss": 1.1907, "step": 5073 }, { "epoch": 0.17879260375098002, "grad_norm": 1.2961902618408203, "learning_rate": 1.886774306510522e-05, "loss": 1.0255, "step": 5074 }, { "epoch": 0.17882784076393843, "grad_norm": 5.634471893310547, "learning_rate": 1.8867215506143286e-05, "loss": 7.3907, "step": 5075 }, { "epoch": 0.17886307777689686, "grad_norm": 3.8372349739074707, "learning_rate": 1.886668783168474e-05, "loss": 3.8979, "step": 5076 }, { "epoch": 0.17889831478985527, "grad_norm": 1.1494801044464111, "learning_rate": 1.8866160041736457e-05, "loss": 1.2363, "step": 5077 }, { "epoch": 0.17893355180281367, "grad_norm": 2.0338292121887207, "learning_rate": 1.886563213630531e-05, "loss": 1.3039, "step": 5078 }, { "epoch": 0.17896878881577208, "grad_norm": 1.3790544271469116, "learning_rate": 1.8865104115398177e-05, "loss": 0.8591, "step": 5079 }, { "epoch": 0.1790040258287305, "grad_norm": 4.350614547729492, "learning_rate": 1.886457597902193e-05, "loss": 3.7666, "step": 5080 }, { "epoch": 0.17903926284168892, "grad_norm": 6.40575647354126, "learning_rate": 1.8864047727183456e-05, "loss": 3.3755, "step": 5081 }, { "epoch": 0.17907449985464732, "grad_norm": 4.473512172698975, "learning_rate": 1.8863519359889627e-05, "loss": 3.9766, "step": 5082 }, { "epoch": 0.17910973686760573, "grad_norm": 5.445770263671875, "learning_rate": 1.8862990877147335e-05, "loss": 3.0321, "step": 5083 }, { "epoch": 0.17914497388056413, "grad_norm": 3.794475793838501, "learning_rate": 1.8862462278963454e-05, "loss": 3.2754, "step": 5084 }, { "epoch": 0.17918021089352257, "grad_norm": 1.0367621183395386, "learning_rate": 1.8861933565344873e-05, "loss": 1.1734, "step": 5085 }, { "epoch": 0.17921544790648097, "grad_norm": 1.1612918376922607, "learning_rate": 1.8861404736298476e-05, "loss": 1.0418, "step": 5086 }, { "epoch": 0.17925068491943938, "grad_norm": 7.906508445739746, "learning_rate": 1.8860875791831155e-05, "loss": 3.0952, "step": 5087 }, { "epoch": 0.17928592193239778, "grad_norm": 3.7849528789520264, "learning_rate": 1.8860346731949805e-05, "loss": 3.3108, "step": 5088 }, { "epoch": 0.1793211589453562, "grad_norm": 3.8351364135742188, "learning_rate": 1.8859817556661303e-05, "loss": 4.1072, "step": 5089 }, { "epoch": 0.17935639595831462, "grad_norm": 1.4603338241577148, "learning_rate": 1.8859288265972548e-05, "loss": 1.2027, "step": 5090 }, { "epoch": 0.17939163297127303, "grad_norm": 1.2447234392166138, "learning_rate": 1.8858758859890438e-05, "loss": 0.9106, "step": 5091 }, { "epoch": 0.17942686998423144, "grad_norm": 1.28029465675354, "learning_rate": 1.8858229338421864e-05, "loss": 0.935, "step": 5092 }, { "epoch": 0.17946210699718984, "grad_norm": 6.434625148773193, "learning_rate": 1.8857699701573725e-05, "loss": 6.2027, "step": 5093 }, { "epoch": 0.17949734401014825, "grad_norm": 5.687079429626465, "learning_rate": 1.885716994935292e-05, "loss": 4.2053, "step": 5094 }, { "epoch": 0.17953258102310668, "grad_norm": 1.2568292617797852, "learning_rate": 1.8856640081766348e-05, "loss": 0.8277, "step": 5095 }, { "epoch": 0.17956781803606509, "grad_norm": 4.020138740539551, "learning_rate": 1.885611009882091e-05, "loss": 4.1564, "step": 5096 }, { "epoch": 0.1796030550490235, "grad_norm": 1.1653940677642822, "learning_rate": 1.885558000052351e-05, "loss": 1.1613, "step": 5097 }, { "epoch": 0.1796382920619819, "grad_norm": 5.314829349517822, "learning_rate": 1.885504978688105e-05, "loss": 3.1305, "step": 5098 }, { "epoch": 0.17967352907494033, "grad_norm": 2.087636709213257, "learning_rate": 1.8854519457900444e-05, "loss": 3.4449, "step": 5099 }, { "epoch": 0.17970876608789874, "grad_norm": 1.9291399717330933, "learning_rate": 1.885398901358859e-05, "loss": 0.9939, "step": 5100 }, { "epoch": 0.17974400310085714, "grad_norm": 4.591325759887695, "learning_rate": 1.8853458453952404e-05, "loss": 3.6093, "step": 5101 }, { "epoch": 0.17977924011381555, "grad_norm": 1.5038223266601562, "learning_rate": 1.885292777899879e-05, "loss": 0.9132, "step": 5102 }, { "epoch": 0.17981447712677395, "grad_norm": 3.250108480453491, "learning_rate": 1.8852396988734666e-05, "loss": 4.0615, "step": 5103 }, { "epoch": 0.1798497141397324, "grad_norm": 5.220009803771973, "learning_rate": 1.8851866083166945e-05, "loss": 5.7391, "step": 5104 }, { "epoch": 0.1798849511526908, "grad_norm": 4.530545711517334, "learning_rate": 1.885133506230254e-05, "loss": 3.9089, "step": 5105 }, { "epoch": 0.1799201881656492, "grad_norm": 5.869213581085205, "learning_rate": 1.8850803926148366e-05, "loss": 2.7832, "step": 5106 }, { "epoch": 0.1799554251786076, "grad_norm": 2.5156030654907227, "learning_rate": 1.8850272674711346e-05, "loss": 3.0151, "step": 5107 }, { "epoch": 0.179990662191566, "grad_norm": 1.6036518812179565, "learning_rate": 1.8849741307998397e-05, "loss": 0.8819, "step": 5108 }, { "epoch": 0.18002589920452444, "grad_norm": 5.782435417175293, "learning_rate": 1.884920982601644e-05, "loss": 3.6693, "step": 5109 }, { "epoch": 0.18006113621748285, "grad_norm": 1.0529011487960815, "learning_rate": 1.8848678228772396e-05, "loss": 1.1206, "step": 5110 }, { "epoch": 0.18009637323044125, "grad_norm": 1.2978812456130981, "learning_rate": 1.8848146516273194e-05, "loss": 0.9653, "step": 5111 }, { "epoch": 0.18013161024339966, "grad_norm": 6.244574069976807, "learning_rate": 1.8847614688525755e-05, "loss": 6.4633, "step": 5112 }, { "epoch": 0.18016684725635806, "grad_norm": 2.253385543823242, "learning_rate": 1.8847082745537006e-05, "loss": 3.0508, "step": 5113 }, { "epoch": 0.1802020842693165, "grad_norm": 1.6643418073654175, "learning_rate": 1.884655068731388e-05, "loss": 1.0297, "step": 5114 }, { "epoch": 0.1802373212822749, "grad_norm": 1.2364935874938965, "learning_rate": 1.8846018513863303e-05, "loss": 1.0399, "step": 5115 }, { "epoch": 0.1802725582952333, "grad_norm": 1.0152239799499512, "learning_rate": 1.884548622519221e-05, "loss": 1.419, "step": 5116 }, { "epoch": 0.18030779530819172, "grad_norm": 1.926336407661438, "learning_rate": 1.884495382130753e-05, "loss": 0.9062, "step": 5117 }, { "epoch": 0.18034303232115015, "grad_norm": 1.485105037689209, "learning_rate": 1.8844421302216202e-05, "loss": 0.8189, "step": 5118 }, { "epoch": 0.18037826933410855, "grad_norm": 3.5519771575927734, "learning_rate": 1.8843888667925155e-05, "loss": 3.4535, "step": 5119 }, { "epoch": 0.18041350634706696, "grad_norm": 1.2896907329559326, "learning_rate": 1.8843355918441335e-05, "loss": 0.8032, "step": 5120 }, { "epoch": 0.18044874336002537, "grad_norm": 1.0932610034942627, "learning_rate": 1.884282305377168e-05, "loss": 1.0058, "step": 5121 }, { "epoch": 0.18048398037298377, "grad_norm": 6.641245365142822, "learning_rate": 1.8842290073923127e-05, "loss": 3.8716, "step": 5122 }, { "epoch": 0.1805192173859422, "grad_norm": 1.152528166770935, "learning_rate": 1.8841756978902617e-05, "loss": 1.021, "step": 5123 }, { "epoch": 0.1805544543989006, "grad_norm": 1.3890889883041382, "learning_rate": 1.8841223768717095e-05, "loss": 0.9554, "step": 5124 }, { "epoch": 0.18058969141185902, "grad_norm": 4.163451194763184, "learning_rate": 1.884069044337351e-05, "loss": 6.0675, "step": 5125 }, { "epoch": 0.18062492842481742, "grad_norm": 6.635937690734863, "learning_rate": 1.8840157002878806e-05, "loss": 6.2706, "step": 5126 }, { "epoch": 0.18066016543777583, "grad_norm": 1.4536030292510986, "learning_rate": 1.8839623447239933e-05, "loss": 1.0633, "step": 5127 }, { "epoch": 0.18069540245073426, "grad_norm": 3.0517916679382324, "learning_rate": 1.8839089776463834e-05, "loss": 2.8472, "step": 5128 }, { "epoch": 0.18073063946369267, "grad_norm": 2.172081470489502, "learning_rate": 1.883855599055747e-05, "loss": 3.8203, "step": 5129 }, { "epoch": 0.18076587647665107, "grad_norm": 1.6897221803665161, "learning_rate": 1.8838022089527783e-05, "loss": 1.0175, "step": 5130 }, { "epoch": 0.18080111348960948, "grad_norm": 1.6246455907821655, "learning_rate": 1.8837488073381735e-05, "loss": 0.9677, "step": 5131 }, { "epoch": 0.1808363505025679, "grad_norm": 3.412142753601074, "learning_rate": 1.8836953942126278e-05, "loss": 0.9397, "step": 5132 }, { "epoch": 0.18087158751552632, "grad_norm": 4.49629545211792, "learning_rate": 1.8836419695768374e-05, "loss": 3.1915, "step": 5133 }, { "epoch": 0.18090682452848472, "grad_norm": 1.2912770509719849, "learning_rate": 1.8835885334314972e-05, "loss": 0.9203, "step": 5134 }, { "epoch": 0.18094206154144313, "grad_norm": 2.103055477142334, "learning_rate": 1.8835350857773043e-05, "loss": 4.0542, "step": 5135 }, { "epoch": 0.18097729855440153, "grad_norm": 1.4435001611709595, "learning_rate": 1.883481626614954e-05, "loss": 0.8962, "step": 5136 }, { "epoch": 0.18101253556735997, "grad_norm": 3.610834836959839, "learning_rate": 1.8834281559451433e-05, "loss": 2.8247, "step": 5137 }, { "epoch": 0.18104777258031837, "grad_norm": 3.1597437858581543, "learning_rate": 1.883374673768568e-05, "loss": 3.9716, "step": 5138 }, { "epoch": 0.18108300959327678, "grad_norm": 3.018505096435547, "learning_rate": 1.8833211800859255e-05, "loss": 3.8396, "step": 5139 }, { "epoch": 0.18111824660623518, "grad_norm": 5.909828186035156, "learning_rate": 1.8832676748979117e-05, "loss": 8.4018, "step": 5140 }, { "epoch": 0.1811534836191936, "grad_norm": 1.3756901025772095, "learning_rate": 1.883214158205224e-05, "loss": 0.9788, "step": 5141 }, { "epoch": 0.18118872063215202, "grad_norm": 2.3535828590393066, "learning_rate": 1.8831606300085593e-05, "loss": 3.3277, "step": 5142 }, { "epoch": 0.18122395764511043, "grad_norm": 3.6338911056518555, "learning_rate": 1.8831070903086153e-05, "loss": 3.1501, "step": 5143 }, { "epoch": 0.18125919465806883, "grad_norm": 1.0501269102096558, "learning_rate": 1.8830535391060886e-05, "loss": 1.0359, "step": 5144 }, { "epoch": 0.18129443167102724, "grad_norm": 0.8850386142730713, "learning_rate": 1.8829999764016774e-05, "loss": 0.883, "step": 5145 }, { "epoch": 0.18132966868398565, "grad_norm": 9.239803314208984, "learning_rate": 1.8829464021960787e-05, "loss": 6.5486, "step": 5146 }, { "epoch": 0.18136490569694408, "grad_norm": 4.680690288543701, "learning_rate": 1.882892816489991e-05, "loss": 3.6426, "step": 5147 }, { "epoch": 0.18140014270990248, "grad_norm": 4.717430114746094, "learning_rate": 1.8828392192841114e-05, "loss": 3.5282, "step": 5148 }, { "epoch": 0.1814353797228609, "grad_norm": 1.0940239429473877, "learning_rate": 1.882785610579139e-05, "loss": 0.7645, "step": 5149 }, { "epoch": 0.1814706167358193, "grad_norm": 1.147639513015747, "learning_rate": 1.882731990375771e-05, "loss": 0.9454, "step": 5150 }, { "epoch": 0.18150585374877773, "grad_norm": 1.3917255401611328, "learning_rate": 1.882678358674707e-05, "loss": 1.0944, "step": 5151 }, { "epoch": 0.18154109076173613, "grad_norm": 4.461108207702637, "learning_rate": 1.8826247154766446e-05, "loss": 5.9159, "step": 5152 }, { "epoch": 0.18157632777469454, "grad_norm": 2.0318760871887207, "learning_rate": 1.882571060782283e-05, "loss": 3.5142, "step": 5153 }, { "epoch": 0.18161156478765295, "grad_norm": 5.358443737030029, "learning_rate": 1.8825173945923207e-05, "loss": 6.6897, "step": 5154 }, { "epoch": 0.18164680180061135, "grad_norm": 1.4011107683181763, "learning_rate": 1.882463716907457e-05, "loss": 1.1397, "step": 5155 }, { "epoch": 0.18168203881356979, "grad_norm": 3.935753107070923, "learning_rate": 1.882410027728391e-05, "loss": 6.3539, "step": 5156 }, { "epoch": 0.1817172758265282, "grad_norm": 3.6028475761413574, "learning_rate": 1.8823563270558223e-05, "loss": 4.084, "step": 5157 }, { "epoch": 0.1817525128394866, "grad_norm": 1.9125233888626099, "learning_rate": 1.8823026148904496e-05, "loss": 3.697, "step": 5158 }, { "epoch": 0.181787749852445, "grad_norm": 1.6542704105377197, "learning_rate": 1.882248891232973e-05, "loss": 1.1371, "step": 5159 }, { "epoch": 0.1818229868654034, "grad_norm": 2.168278932571411, "learning_rate": 1.8821951560840924e-05, "loss": 3.8963, "step": 5160 }, { "epoch": 0.18185822387836184, "grad_norm": 1.4918054342269897, "learning_rate": 1.8821414094445076e-05, "loss": 0.9598, "step": 5161 }, { "epoch": 0.18189346089132025, "grad_norm": 2.929062843322754, "learning_rate": 1.8820876513149184e-05, "loss": 3.5493, "step": 5162 }, { "epoch": 0.18192869790427865, "grad_norm": 2.608887195587158, "learning_rate": 1.882033881696025e-05, "loss": 4.147, "step": 5163 }, { "epoch": 0.18196393491723706, "grad_norm": 1.4766619205474854, "learning_rate": 1.881980100588528e-05, "loss": 0.9222, "step": 5164 }, { "epoch": 0.18199917193019546, "grad_norm": 1.6880534887313843, "learning_rate": 1.8819263079931284e-05, "loss": 1.2683, "step": 5165 }, { "epoch": 0.1820344089431539, "grad_norm": 1.5270462036132812, "learning_rate": 1.881872503910526e-05, "loss": 1.1463, "step": 5166 }, { "epoch": 0.1820696459561123, "grad_norm": 2.8990650177001953, "learning_rate": 1.8818186883414216e-05, "loss": 2.9978, "step": 5167 }, { "epoch": 0.1821048829690707, "grad_norm": 2.4788548946380615, "learning_rate": 1.8817648612865165e-05, "loss": 3.3614, "step": 5168 }, { "epoch": 0.18214011998202911, "grad_norm": 2.782573699951172, "learning_rate": 1.881711022746512e-05, "loss": 3.5677, "step": 5169 }, { "epoch": 0.18217535699498755, "grad_norm": 3.6224205493927, "learning_rate": 1.881657172722109e-05, "loss": 3.7962, "step": 5170 }, { "epoch": 0.18221059400794595, "grad_norm": 4.900236129760742, "learning_rate": 1.8816033112140092e-05, "loss": 6.5089, "step": 5171 }, { "epoch": 0.18224583102090436, "grad_norm": 4.738795280456543, "learning_rate": 1.8815494382229137e-05, "loss": 6.6541, "step": 5172 }, { "epoch": 0.18228106803386276, "grad_norm": 6.040731430053711, "learning_rate": 1.881495553749525e-05, "loss": 5.6928, "step": 5173 }, { "epoch": 0.18231630504682117, "grad_norm": 5.7719950675964355, "learning_rate": 1.8814416577945435e-05, "loss": 6.5593, "step": 5174 }, { "epoch": 0.1823515420597796, "grad_norm": 1.1117522716522217, "learning_rate": 1.881387750358673e-05, "loss": 0.8941, "step": 5175 }, { "epoch": 0.182386779072738, "grad_norm": 5.024579048156738, "learning_rate": 1.881333831442614e-05, "loss": 4.7782, "step": 5176 }, { "epoch": 0.18242201608569641, "grad_norm": 1.4345555305480957, "learning_rate": 1.88127990104707e-05, "loss": 0.8176, "step": 5177 }, { "epoch": 0.18245725309865482, "grad_norm": 1.727594256401062, "learning_rate": 1.8812259591727433e-05, "loss": 0.902, "step": 5178 }, { "epoch": 0.18249249011161323, "grad_norm": 7.475857734680176, "learning_rate": 1.881172005820336e-05, "loss": 7.3719, "step": 5179 }, { "epoch": 0.18252772712457166, "grad_norm": 1.2707674503326416, "learning_rate": 1.881118040990551e-05, "loss": 1.1467, "step": 5180 }, { "epoch": 0.18256296413753006, "grad_norm": 1.1528400182724, "learning_rate": 1.8810640646840913e-05, "loss": 1.2224, "step": 5181 }, { "epoch": 0.18259820115048847, "grad_norm": 3.796569347381592, "learning_rate": 1.8810100769016603e-05, "loss": 3.2015, "step": 5182 }, { "epoch": 0.18263343816344688, "grad_norm": 1.599098563194275, "learning_rate": 1.8809560776439603e-05, "loss": 1.1035, "step": 5183 }, { "epoch": 0.1826686751764053, "grad_norm": 1.1702157258987427, "learning_rate": 1.8809020669116953e-05, "loss": 0.9189, "step": 5184 }, { "epoch": 0.18270391218936372, "grad_norm": 5.316476345062256, "learning_rate": 1.880848044705569e-05, "loss": 3.631, "step": 5185 }, { "epoch": 0.18273914920232212, "grad_norm": 6.01800537109375, "learning_rate": 1.8807940110262842e-05, "loss": 5.9235, "step": 5186 }, { "epoch": 0.18277438621528053, "grad_norm": 1.333339810371399, "learning_rate": 1.8807399658745454e-05, "loss": 0.9192, "step": 5187 }, { "epoch": 0.18280962322823893, "grad_norm": 2.457611322402954, "learning_rate": 1.8806859092510566e-05, "loss": 3.1202, "step": 5188 }, { "epoch": 0.18284486024119737, "grad_norm": 1.591680645942688, "learning_rate": 1.880631841156521e-05, "loss": 0.7692, "step": 5189 }, { "epoch": 0.18288009725415577, "grad_norm": 1.0692524909973145, "learning_rate": 1.8805777615916442e-05, "loss": 0.9151, "step": 5190 }, { "epoch": 0.18291533426711418, "grad_norm": 1.4166748523712158, "learning_rate": 1.8805236705571295e-05, "loss": 0.9215, "step": 5191 }, { "epoch": 0.18295057128007258, "grad_norm": 4.634408950805664, "learning_rate": 1.8804695680536823e-05, "loss": 2.9407, "step": 5192 }, { "epoch": 0.182985808293031, "grad_norm": 1.3755412101745605, "learning_rate": 1.8804154540820067e-05, "loss": 0.7681, "step": 5193 }, { "epoch": 0.18302104530598942, "grad_norm": 9.450085639953613, "learning_rate": 1.880361328642807e-05, "loss": 3.8311, "step": 5194 }, { "epoch": 0.18305628231894783, "grad_norm": 1.2833610773086548, "learning_rate": 1.8803071917367894e-05, "loss": 1.0315, "step": 5195 }, { "epoch": 0.18309151933190623, "grad_norm": 9.223179817199707, "learning_rate": 1.8802530433646585e-05, "loss": 6.4084, "step": 5196 }, { "epoch": 0.18312675634486464, "grad_norm": 11.602048873901367, "learning_rate": 1.8801988835271198e-05, "loss": 6.5113, "step": 5197 }, { "epoch": 0.18316199335782304, "grad_norm": 1.0082387924194336, "learning_rate": 1.880144712224878e-05, "loss": 0.872, "step": 5198 }, { "epoch": 0.18319723037078148, "grad_norm": 1.4308547973632812, "learning_rate": 1.8800905294586392e-05, "loss": 0.8923, "step": 5199 }, { "epoch": 0.18323246738373988, "grad_norm": 7.864358901977539, "learning_rate": 1.8800363352291095e-05, "loss": 3.8357, "step": 5200 }, { "epoch": 0.1832677043966983, "grad_norm": 1.411439061164856, "learning_rate": 1.8799821295369943e-05, "loss": 1.2537, "step": 5201 }, { "epoch": 0.1833029414096567, "grad_norm": 1.540300726890564, "learning_rate": 1.879927912383e-05, "loss": 0.9022, "step": 5202 }, { "epoch": 0.18333817842261513, "grad_norm": 1.416761040687561, "learning_rate": 1.879873683767832e-05, "loss": 1.1685, "step": 5203 }, { "epoch": 0.18337341543557353, "grad_norm": 4.878976345062256, "learning_rate": 1.8798194436921975e-05, "loss": 3.6198, "step": 5204 }, { "epoch": 0.18340865244853194, "grad_norm": 3.533961534500122, "learning_rate": 1.879765192156803e-05, "loss": 2.7308, "step": 5205 }, { "epoch": 0.18344388946149034, "grad_norm": 1.149982213973999, "learning_rate": 1.8797109291623545e-05, "loss": 0.988, "step": 5206 }, { "epoch": 0.18347912647444875, "grad_norm": 7.394337177276611, "learning_rate": 1.879656654709559e-05, "loss": 3.7035, "step": 5207 }, { "epoch": 0.18351436348740718, "grad_norm": 1.6616857051849365, "learning_rate": 1.8796023687991236e-05, "loss": 0.9816, "step": 5208 }, { "epoch": 0.1835496005003656, "grad_norm": 1.5618425607681274, "learning_rate": 1.8795480714317557e-05, "loss": 1.1731, "step": 5209 }, { "epoch": 0.183584837513324, "grad_norm": 3.4754703044891357, "learning_rate": 1.8794937626081618e-05, "loss": 3.0567, "step": 5210 }, { "epoch": 0.1836200745262824, "grad_norm": 7.027205467224121, "learning_rate": 1.8794394423290493e-05, "loss": 3.9955, "step": 5211 }, { "epoch": 0.1836553115392408, "grad_norm": 0.9448646306991577, "learning_rate": 1.8793851105951264e-05, "loss": 0.8586, "step": 5212 }, { "epoch": 0.18369054855219924, "grad_norm": 6.645773887634277, "learning_rate": 1.8793307674071002e-05, "loss": 3.297, "step": 5213 }, { "epoch": 0.18372578556515765, "grad_norm": 3.4629881381988525, "learning_rate": 1.8792764127656792e-05, "loss": 3.745, "step": 5214 }, { "epoch": 0.18376102257811605, "grad_norm": 1.6055185794830322, "learning_rate": 1.8792220466715704e-05, "loss": 1.1396, "step": 5215 }, { "epoch": 0.18379625959107446, "grad_norm": 3.6669511795043945, "learning_rate": 1.8791676691254826e-05, "loss": 3.5414, "step": 5216 }, { "epoch": 0.18383149660403286, "grad_norm": 1.3555755615234375, "learning_rate": 1.879113280128124e-05, "loss": 0.9289, "step": 5217 }, { "epoch": 0.1838667336169913, "grad_norm": 4.957563877105713, "learning_rate": 1.879058879680203e-05, "loss": 6.0794, "step": 5218 }, { "epoch": 0.1839019706299497, "grad_norm": 1.0573093891143799, "learning_rate": 1.879004467782428e-05, "loss": 0.9692, "step": 5219 }, { "epoch": 0.1839372076429081, "grad_norm": 7.441522598266602, "learning_rate": 1.878950044435508e-05, "loss": 6.1948, "step": 5220 }, { "epoch": 0.1839724446558665, "grad_norm": 3.9010629653930664, "learning_rate": 1.8788956096401513e-05, "loss": 5.9603, "step": 5221 }, { "epoch": 0.18400768166882495, "grad_norm": 4.691671848297119, "learning_rate": 1.8788411633970677e-05, "loss": 5.8633, "step": 5222 }, { "epoch": 0.18404291868178335, "grad_norm": 7.362890720367432, "learning_rate": 1.878786705706966e-05, "loss": 3.5962, "step": 5223 }, { "epoch": 0.18407815569474176, "grad_norm": 5.9322028160095215, "learning_rate": 1.8787322365705552e-05, "loss": 3.6592, "step": 5224 }, { "epoch": 0.18411339270770016, "grad_norm": 6.484999179840088, "learning_rate": 1.8786777559885453e-05, "loss": 6.0974, "step": 5225 }, { "epoch": 0.18414862972065857, "grad_norm": 1.4537721872329712, "learning_rate": 1.878623263961646e-05, "loss": 0.8886, "step": 5226 }, { "epoch": 0.184183866733617, "grad_norm": 1.534278154373169, "learning_rate": 1.8785687604905665e-05, "loss": 0.9732, "step": 5227 }, { "epoch": 0.1842191037465754, "grad_norm": 4.295928955078125, "learning_rate": 1.878514245576017e-05, "loss": 4.8746, "step": 5228 }, { "epoch": 0.1842543407595338, "grad_norm": 2.1456215381622314, "learning_rate": 1.8784597192187075e-05, "loss": 3.7632, "step": 5229 }, { "epoch": 0.18428957777249222, "grad_norm": 6.828533172607422, "learning_rate": 1.878405181419348e-05, "loss": 6.6715, "step": 5230 }, { "epoch": 0.18432481478545062, "grad_norm": 4.85668420791626, "learning_rate": 1.87835063217865e-05, "loss": 8.3022, "step": 5231 }, { "epoch": 0.18436005179840906, "grad_norm": 3.488983154296875, "learning_rate": 1.8782960714973223e-05, "loss": 3.4605, "step": 5232 }, { "epoch": 0.18439528881136746, "grad_norm": 1.0920506715774536, "learning_rate": 1.8782414993760765e-05, "loss": 0.944, "step": 5233 }, { "epoch": 0.18443052582432587, "grad_norm": 2.6919267177581787, "learning_rate": 1.8781869158156233e-05, "loss": 3.248, "step": 5234 }, { "epoch": 0.18446576283728428, "grad_norm": 1.2921550273895264, "learning_rate": 1.8781323208166738e-05, "loss": 0.9276, "step": 5235 }, { "epoch": 0.1845009998502427, "grad_norm": 1.5907883644104004, "learning_rate": 1.8780777143799393e-05, "loss": 0.9699, "step": 5236 }, { "epoch": 0.18453623686320111, "grad_norm": 5.484589099884033, "learning_rate": 1.8780230965061302e-05, "loss": 5.2155, "step": 5237 }, { "epoch": 0.18457147387615952, "grad_norm": 1.3259183168411255, "learning_rate": 1.8779684671959587e-05, "loss": 1.0525, "step": 5238 }, { "epoch": 0.18460671088911793, "grad_norm": 0.9202318787574768, "learning_rate": 1.8779138264501357e-05, "loss": 1.1432, "step": 5239 }, { "epoch": 0.18464194790207633, "grad_norm": 1.6193538904190063, "learning_rate": 1.8778591742693737e-05, "loss": 0.8054, "step": 5240 }, { "epoch": 0.18467718491503476, "grad_norm": 1.49186110496521, "learning_rate": 1.877804510654384e-05, "loss": 0.9425, "step": 5241 }, { "epoch": 0.18471242192799317, "grad_norm": 5.006694316864014, "learning_rate": 1.8777498356058786e-05, "loss": 3.4804, "step": 5242 }, { "epoch": 0.18474765894095158, "grad_norm": 8.880630493164062, "learning_rate": 1.8776951491245698e-05, "loss": 6.0756, "step": 5243 }, { "epoch": 0.18478289595390998, "grad_norm": 0.9308595061302185, "learning_rate": 1.87764045121117e-05, "loss": 0.9887, "step": 5244 }, { "epoch": 0.1848181329668684, "grad_norm": 6.857015609741211, "learning_rate": 1.877585741866392e-05, "loss": 6.4735, "step": 5245 }, { "epoch": 0.18485336997982682, "grad_norm": 2.8365776538848877, "learning_rate": 1.8775310210909473e-05, "loss": 3.6226, "step": 5246 }, { "epoch": 0.18488860699278523, "grad_norm": 1.7582767009735107, "learning_rate": 1.877476288885549e-05, "loss": 0.7878, "step": 5247 }, { "epoch": 0.18492384400574363, "grad_norm": 10.621428489685059, "learning_rate": 1.877421545250911e-05, "loss": 3.4111, "step": 5248 }, { "epoch": 0.18495908101870204, "grad_norm": 5.7469282150268555, "learning_rate": 1.8773667901877452e-05, "loss": 3.031, "step": 5249 }, { "epoch": 0.18499431803166044, "grad_norm": 1.0641106367111206, "learning_rate": 1.8773120236967655e-05, "loss": 1.2384, "step": 5250 }, { "epoch": 0.18502955504461888, "grad_norm": 1.4755713939666748, "learning_rate": 1.8772572457786844e-05, "loss": 0.9705, "step": 5251 }, { "epoch": 0.18506479205757728, "grad_norm": 1.0079786777496338, "learning_rate": 1.8772024564342164e-05, "loss": 0.8359, "step": 5252 }, { "epoch": 0.1851000290705357, "grad_norm": 2.668323516845703, "learning_rate": 1.8771476556640748e-05, "loss": 3.4466, "step": 5253 }, { "epoch": 0.1851352660834941, "grad_norm": 1.471691608428955, "learning_rate": 1.877092843468973e-05, "loss": 0.9648, "step": 5254 }, { "epoch": 0.18517050309645253, "grad_norm": 6.398085594177246, "learning_rate": 1.877038019849625e-05, "loss": 3.2534, "step": 5255 }, { "epoch": 0.18520574010941093, "grad_norm": 2.455756187438965, "learning_rate": 1.876983184806745e-05, "loss": 3.1721, "step": 5256 }, { "epoch": 0.18524097712236934, "grad_norm": 1.322587251663208, "learning_rate": 1.8769283383410477e-05, "loss": 0.9339, "step": 5257 }, { "epoch": 0.18527621413532774, "grad_norm": 3.635920763015747, "learning_rate": 1.8768734804532468e-05, "loss": 3.2665, "step": 5258 }, { "epoch": 0.18531145114828615, "grad_norm": 6.642467498779297, "learning_rate": 1.8768186111440573e-05, "loss": 3.9977, "step": 5259 }, { "epoch": 0.18534668816124458, "grad_norm": 3.3203330039978027, "learning_rate": 1.8767637304141936e-05, "loss": 3.0231, "step": 5260 }, { "epoch": 0.185381925174203, "grad_norm": 4.455447673797607, "learning_rate": 1.8767088382643708e-05, "loss": 3.7585, "step": 5261 }, { "epoch": 0.1854171621871614, "grad_norm": 5.824244499206543, "learning_rate": 1.8766539346953037e-05, "loss": 3.7707, "step": 5262 }, { "epoch": 0.1854523992001198, "grad_norm": 1.6717382669448853, "learning_rate": 1.8765990197077074e-05, "loss": 1.1178, "step": 5263 }, { "epoch": 0.1854876362130782, "grad_norm": 1.849105954170227, "learning_rate": 1.876544093302297e-05, "loss": 1.1351, "step": 5264 }, { "epoch": 0.18552287322603664, "grad_norm": 1.46061110496521, "learning_rate": 1.8764891554797882e-05, "loss": 0.899, "step": 5265 }, { "epoch": 0.18555811023899504, "grad_norm": 2.722766160964966, "learning_rate": 1.8764342062408968e-05, "loss": 3.6206, "step": 5266 }, { "epoch": 0.18559334725195345, "grad_norm": 10.7405424118042, "learning_rate": 1.8763792455863377e-05, "loss": 5.8638, "step": 5267 }, { "epoch": 0.18562858426491186, "grad_norm": 7.366532802581787, "learning_rate": 1.876324273516828e-05, "loss": 3.1302, "step": 5268 }, { "epoch": 0.18566382127787026, "grad_norm": 5.752516269683838, "learning_rate": 1.8762692900330822e-05, "loss": 5.5949, "step": 5269 }, { "epoch": 0.1856990582908287, "grad_norm": 1.5118082761764526, "learning_rate": 1.8762142951358176e-05, "loss": 0.9648, "step": 5270 }, { "epoch": 0.1857342953037871, "grad_norm": 3.7056636810302734, "learning_rate": 1.8761592888257504e-05, "loss": 3.7434, "step": 5271 }, { "epoch": 0.1857695323167455, "grad_norm": 3.532907724380493, "learning_rate": 1.8761042711035967e-05, "loss": 3.3136, "step": 5272 }, { "epoch": 0.1858047693297039, "grad_norm": 1.5106624364852905, "learning_rate": 1.8760492419700735e-05, "loss": 1.0892, "step": 5273 }, { "epoch": 0.18584000634266234, "grad_norm": 7.044299602508545, "learning_rate": 1.875994201425897e-05, "loss": 3.1615, "step": 5274 }, { "epoch": 0.18587524335562075, "grad_norm": 1.6687216758728027, "learning_rate": 1.8759391494717847e-05, "loss": 0.9286, "step": 5275 }, { "epoch": 0.18591048036857916, "grad_norm": 2.0595250129699707, "learning_rate": 1.8758840861084533e-05, "loss": 0.9095, "step": 5276 }, { "epoch": 0.18594571738153756, "grad_norm": 2.8373615741729736, "learning_rate": 1.87582901133662e-05, "loss": 2.7176, "step": 5277 }, { "epoch": 0.18598095439449597, "grad_norm": 4.013903617858887, "learning_rate": 1.8757739251570024e-05, "loss": 3.6581, "step": 5278 }, { "epoch": 0.1860161914074544, "grad_norm": 3.0144429206848145, "learning_rate": 1.8757188275703178e-05, "loss": 3.4027, "step": 5279 }, { "epoch": 0.1860514284204128, "grad_norm": 1.3063238859176636, "learning_rate": 1.875663718577284e-05, "loss": 0.8686, "step": 5280 }, { "epoch": 0.1860866654333712, "grad_norm": 5.398768901824951, "learning_rate": 1.8756085981786186e-05, "loss": 3.3727, "step": 5281 }, { "epoch": 0.18612190244632962, "grad_norm": 7.589937210083008, "learning_rate": 1.8755534663750396e-05, "loss": 5.9946, "step": 5282 }, { "epoch": 0.18615713945928802, "grad_norm": 4.412134647369385, "learning_rate": 1.8754983231672655e-05, "loss": 3.8221, "step": 5283 }, { "epoch": 0.18619237647224646, "grad_norm": 4.799152374267578, "learning_rate": 1.875443168556014e-05, "loss": 4.7849, "step": 5284 }, { "epoch": 0.18622761348520486, "grad_norm": 4.452366828918457, "learning_rate": 1.8753880025420037e-05, "loss": 4.0215, "step": 5285 }, { "epoch": 0.18626285049816327, "grad_norm": 2.0763704776763916, "learning_rate": 1.8753328251259532e-05, "loss": 0.9702, "step": 5286 }, { "epoch": 0.18629808751112167, "grad_norm": 2.005938768386841, "learning_rate": 1.875277636308581e-05, "loss": 1.0997, "step": 5287 }, { "epoch": 0.1863333245240801, "grad_norm": 5.773597240447998, "learning_rate": 1.8752224360906064e-05, "loss": 5.7685, "step": 5288 }, { "epoch": 0.1863685615370385, "grad_norm": 4.110316276550293, "learning_rate": 1.875167224472748e-05, "loss": 3.2228, "step": 5289 }, { "epoch": 0.18640379854999692, "grad_norm": 4.9605793952941895, "learning_rate": 1.8751120014557253e-05, "loss": 2.8759, "step": 5290 }, { "epoch": 0.18643903556295532, "grad_norm": 3.6913766860961914, "learning_rate": 1.8750567670402565e-05, "loss": 5.3365, "step": 5291 }, { "epoch": 0.18647427257591373, "grad_norm": 5.768638610839844, "learning_rate": 1.8750015212270623e-05, "loss": 3.139, "step": 5292 }, { "epoch": 0.18650950958887216, "grad_norm": 1.3481967449188232, "learning_rate": 1.874946264016862e-05, "loss": 1.1159, "step": 5293 }, { "epoch": 0.18654474660183057, "grad_norm": 1.5141961574554443, "learning_rate": 1.874890995410375e-05, "loss": 1.0683, "step": 5294 }, { "epoch": 0.18657998361478897, "grad_norm": 5.142585754394531, "learning_rate": 1.8748357154083215e-05, "loss": 3.54, "step": 5295 }, { "epoch": 0.18661522062774738, "grad_norm": 1.3617860078811646, "learning_rate": 1.874780424011421e-05, "loss": 0.948, "step": 5296 }, { "epoch": 0.18665045764070579, "grad_norm": 6.234852313995361, "learning_rate": 1.8747251212203944e-05, "loss": 4.7131, "step": 5297 }, { "epoch": 0.18668569465366422, "grad_norm": 1.1059634685516357, "learning_rate": 1.8746698070359618e-05, "loss": 0.9189, "step": 5298 }, { "epoch": 0.18672093166662262, "grad_norm": 4.421064853668213, "learning_rate": 1.8746144814588428e-05, "loss": 3.3494, "step": 5299 }, { "epoch": 0.18675616867958103, "grad_norm": 4.553877830505371, "learning_rate": 1.8745591444897595e-05, "loss": 6.615, "step": 5300 }, { "epoch": 0.18679140569253944, "grad_norm": 2.8178322315216064, "learning_rate": 1.8745037961294316e-05, "loss": 3.367, "step": 5301 }, { "epoch": 0.18682664270549784, "grad_norm": 8.303633689880371, "learning_rate": 1.8744484363785807e-05, "loss": 3.4165, "step": 5302 }, { "epoch": 0.18686187971845628, "grad_norm": 1.5834641456604004, "learning_rate": 1.874393065237927e-05, "loss": 0.8079, "step": 5303 }, { "epoch": 0.18689711673141468, "grad_norm": 5.161906719207764, "learning_rate": 1.8743376827081928e-05, "loss": 4.2931, "step": 5304 }, { "epoch": 0.1869323537443731, "grad_norm": 3.772308826446533, "learning_rate": 1.8742822887900986e-05, "loss": 3.3819, "step": 5305 }, { "epoch": 0.1869675907573315, "grad_norm": 3.4741711616516113, "learning_rate": 1.8742268834843663e-05, "loss": 3.6169, "step": 5306 }, { "epoch": 0.18700282777028993, "grad_norm": 4.267179012298584, "learning_rate": 1.8741714667917174e-05, "loss": 3.8315, "step": 5307 }, { "epoch": 0.18703806478324833, "grad_norm": 1.2071657180786133, "learning_rate": 1.8741160387128738e-05, "loss": 1.0447, "step": 5308 }, { "epoch": 0.18707330179620674, "grad_norm": 1.261852741241455, "learning_rate": 1.8740605992485576e-05, "loss": 1.0955, "step": 5309 }, { "epoch": 0.18710853880916514, "grad_norm": 2.0214884281158447, "learning_rate": 1.8740051483994907e-05, "loss": 0.9726, "step": 5310 }, { "epoch": 0.18714377582212355, "grad_norm": 1.6375149488449097, "learning_rate": 1.8739496861663952e-05, "loss": 0.8292, "step": 5311 }, { "epoch": 0.18717901283508198, "grad_norm": 7.060153007507324, "learning_rate": 1.8738942125499938e-05, "loss": 4.5237, "step": 5312 }, { "epoch": 0.1872142498480404, "grad_norm": 15.346633911132812, "learning_rate": 1.873838727551009e-05, "loss": 3.743, "step": 5313 }, { "epoch": 0.1872494868609988, "grad_norm": 1.21311616897583, "learning_rate": 1.8737832311701636e-05, "loss": 1.2205, "step": 5314 }, { "epoch": 0.1872847238739572, "grad_norm": 4.702340602874756, "learning_rate": 1.8737277234081802e-05, "loss": 3.6611, "step": 5315 }, { "epoch": 0.1873199608869156, "grad_norm": 4.3131327629089355, "learning_rate": 1.8736722042657823e-05, "loss": 4.368, "step": 5316 }, { "epoch": 0.18735519789987404, "grad_norm": 2.9753458499908447, "learning_rate": 1.8736166737436922e-05, "loss": 3.2868, "step": 5317 }, { "epoch": 0.18739043491283244, "grad_norm": 1.0142860412597656, "learning_rate": 1.8735611318426337e-05, "loss": 0.9204, "step": 5318 }, { "epoch": 0.18742567192579085, "grad_norm": 1.4103766679763794, "learning_rate": 1.87350557856333e-05, "loss": 0.9234, "step": 5319 }, { "epoch": 0.18746090893874925, "grad_norm": 1.412716269493103, "learning_rate": 1.873450013906505e-05, "loss": 0.737, "step": 5320 }, { "epoch": 0.18749614595170766, "grad_norm": 2.534515380859375, "learning_rate": 1.8733944378728825e-05, "loss": 4.2079, "step": 5321 }, { "epoch": 0.1875313829646661, "grad_norm": 1.2593553066253662, "learning_rate": 1.8733388504631862e-05, "loss": 0.8042, "step": 5322 }, { "epoch": 0.1875666199776245, "grad_norm": 4.0511555671691895, "learning_rate": 1.87328325167814e-05, "loss": 4.0422, "step": 5323 }, { "epoch": 0.1876018569905829, "grad_norm": 8.246416091918945, "learning_rate": 1.8732276415184683e-05, "loss": 8.7211, "step": 5324 }, { "epoch": 0.1876370940035413, "grad_norm": 1.3830459117889404, "learning_rate": 1.873172019984895e-05, "loss": 0.8478, "step": 5325 }, { "epoch": 0.18767233101649974, "grad_norm": 1.246234655380249, "learning_rate": 1.8731163870781454e-05, "loss": 0.7761, "step": 5326 }, { "epoch": 0.18770756802945815, "grad_norm": 6.377997398376465, "learning_rate": 1.8730607427989433e-05, "loss": 6.245, "step": 5327 }, { "epoch": 0.18774280504241656, "grad_norm": 1.2534546852111816, "learning_rate": 1.873005087148014e-05, "loss": 1.008, "step": 5328 }, { "epoch": 0.18777804205537496, "grad_norm": 1.077067255973816, "learning_rate": 1.8729494201260823e-05, "loss": 1.0853, "step": 5329 }, { "epoch": 0.18781327906833337, "grad_norm": 6.086742401123047, "learning_rate": 1.872893741733873e-05, "loss": 3.5655, "step": 5330 }, { "epoch": 0.1878485160812918, "grad_norm": 9.843290328979492, "learning_rate": 1.8728380519721114e-05, "loss": 6.4339, "step": 5331 }, { "epoch": 0.1878837530942502, "grad_norm": 5.4190287590026855, "learning_rate": 1.8727823508415233e-05, "loss": 3.0857, "step": 5332 }, { "epoch": 0.1879189901072086, "grad_norm": 4.5547308921813965, "learning_rate": 1.872726638342834e-05, "loss": 6.1467, "step": 5333 }, { "epoch": 0.18795422712016702, "grad_norm": 2.9571073055267334, "learning_rate": 1.8726709144767685e-05, "loss": 3.9869, "step": 5334 }, { "epoch": 0.18798946413312542, "grad_norm": 5.636338233947754, "learning_rate": 1.8726151792440535e-05, "loss": 3.4009, "step": 5335 }, { "epoch": 0.18802470114608386, "grad_norm": 3.838719129562378, "learning_rate": 1.8725594326454148e-05, "loss": 5.5662, "step": 5336 }, { "epoch": 0.18805993815904226, "grad_norm": 5.1544036865234375, "learning_rate": 1.8725036746815783e-05, "loss": 4.1192, "step": 5337 }, { "epoch": 0.18809517517200067, "grad_norm": 1.5396003723144531, "learning_rate": 1.8724479053532702e-05, "loss": 0.9083, "step": 5338 }, { "epoch": 0.18813041218495907, "grad_norm": 1.720369815826416, "learning_rate": 1.872392124661217e-05, "loss": 1.226, "step": 5339 }, { "epoch": 0.1881656491979175, "grad_norm": 1.8253517150878906, "learning_rate": 1.8723363326061453e-05, "loss": 1.3933, "step": 5340 }, { "epoch": 0.1882008862108759, "grad_norm": 1.014765977859497, "learning_rate": 1.8722805291887818e-05, "loss": 0.7969, "step": 5341 }, { "epoch": 0.18823612322383432, "grad_norm": 2.042163372039795, "learning_rate": 1.8722247144098533e-05, "loss": 0.9474, "step": 5342 }, { "epoch": 0.18827136023679272, "grad_norm": 4.020740985870361, "learning_rate": 1.8721688882700866e-05, "loss": 3.5273, "step": 5343 }, { "epoch": 0.18830659724975113, "grad_norm": 4.819957256317139, "learning_rate": 1.8721130507702094e-05, "loss": 3.3371, "step": 5344 }, { "epoch": 0.18834183426270956, "grad_norm": 1.1681972742080688, "learning_rate": 1.872057201910948e-05, "loss": 0.8685, "step": 5345 }, { "epoch": 0.18837707127566797, "grad_norm": 1.1494234800338745, "learning_rate": 1.872001341693031e-05, "loss": 1.0683, "step": 5346 }, { "epoch": 0.18841230828862637, "grad_norm": 0.9645292162895203, "learning_rate": 1.871945470117185e-05, "loss": 1.0716, "step": 5347 }, { "epoch": 0.18844754530158478, "grad_norm": 1.4410206079483032, "learning_rate": 1.8718895871841388e-05, "loss": 1.0326, "step": 5348 }, { "epoch": 0.18848278231454318, "grad_norm": 6.1896162033081055, "learning_rate": 1.8718336928946192e-05, "loss": 3.9409, "step": 5349 }, { "epoch": 0.18851801932750162, "grad_norm": 4.561097621917725, "learning_rate": 1.8717777872493547e-05, "loss": 3.2994, "step": 5350 }, { "epoch": 0.18855325634046002, "grad_norm": 1.4382344484329224, "learning_rate": 1.871721870249074e-05, "loss": 1.203, "step": 5351 }, { "epoch": 0.18858849335341843, "grad_norm": 9.634200096130371, "learning_rate": 1.871665941894504e-05, "loss": 8.4052, "step": 5352 }, { "epoch": 0.18862373036637683, "grad_norm": 1.0323569774627686, "learning_rate": 1.871610002186375e-05, "loss": 1.3672, "step": 5353 }, { "epoch": 0.18865896737933524, "grad_norm": 1.2878309488296509, "learning_rate": 1.871554051125414e-05, "loss": 1.3293, "step": 5354 }, { "epoch": 0.18869420439229367, "grad_norm": 1.632026195526123, "learning_rate": 1.871498088712351e-05, "loss": 0.9675, "step": 5355 }, { "epoch": 0.18872944140525208, "grad_norm": 4.505435466766357, "learning_rate": 1.871442114947914e-05, "loss": 3.2546, "step": 5356 }, { "epoch": 0.18876467841821049, "grad_norm": 4.7887163162231445, "learning_rate": 1.8713861298328325e-05, "loss": 2.8145, "step": 5357 }, { "epoch": 0.1887999154311689, "grad_norm": 1.6394188404083252, "learning_rate": 1.871330133367836e-05, "loss": 0.9879, "step": 5358 }, { "epoch": 0.18883515244412732, "grad_norm": 1.5648130178451538, "learning_rate": 1.871274125553653e-05, "loss": 1.2742, "step": 5359 }, { "epoch": 0.18887038945708573, "grad_norm": 1.362646222114563, "learning_rate": 1.871218106391014e-05, "loss": 0.7881, "step": 5360 }, { "epoch": 0.18890562647004414, "grad_norm": 5.45576286315918, "learning_rate": 1.8711620758806477e-05, "loss": 3.5579, "step": 5361 }, { "epoch": 0.18894086348300254, "grad_norm": 1.8703322410583496, "learning_rate": 1.871106034023285e-05, "loss": 0.987, "step": 5362 }, { "epoch": 0.18897610049596095, "grad_norm": 1.537475347518921, "learning_rate": 1.8710499808196548e-05, "loss": 1.2409, "step": 5363 }, { "epoch": 0.18901133750891938, "grad_norm": 1.8502728939056396, "learning_rate": 1.8709939162704877e-05, "loss": 0.9822, "step": 5364 }, { "epoch": 0.18904657452187779, "grad_norm": 5.063783645629883, "learning_rate": 1.870937840376514e-05, "loss": 2.8583, "step": 5365 }, { "epoch": 0.1890818115348362, "grad_norm": 1.2215089797973633, "learning_rate": 1.8708817531384638e-05, "loss": 1.0563, "step": 5366 }, { "epoch": 0.1891170485477946, "grad_norm": 1.2329457998275757, "learning_rate": 1.8708256545570677e-05, "loss": 0.7841, "step": 5367 }, { "epoch": 0.189152285560753, "grad_norm": 7.122326374053955, "learning_rate": 1.870769544633057e-05, "loss": 8.1626, "step": 5368 }, { "epoch": 0.18918752257371144, "grad_norm": 8.444014549255371, "learning_rate": 1.870713423367162e-05, "loss": 5.1679, "step": 5369 }, { "epoch": 0.18922275958666984, "grad_norm": 9.2506685256958, "learning_rate": 1.8706572907601134e-05, "loss": 3.2836, "step": 5370 }, { "epoch": 0.18925799659962825, "grad_norm": 6.55551290512085, "learning_rate": 1.870601146812643e-05, "loss": 3.1265, "step": 5371 }, { "epoch": 0.18929323361258665, "grad_norm": 1.3462419509887695, "learning_rate": 1.8705449915254817e-05, "loss": 0.8011, "step": 5372 }, { "epoch": 0.18932847062554506, "grad_norm": 1.2748847007751465, "learning_rate": 1.870488824899361e-05, "loss": 1.1209, "step": 5373 }, { "epoch": 0.1893637076385035, "grad_norm": 6.184300899505615, "learning_rate": 1.8704326469350125e-05, "loss": 3.1416, "step": 5374 }, { "epoch": 0.1893989446514619, "grad_norm": 1.6599202156066895, "learning_rate": 1.8703764576331677e-05, "loss": 1.2061, "step": 5375 }, { "epoch": 0.1894341816644203, "grad_norm": 1.4606019258499146, "learning_rate": 1.8703202569945586e-05, "loss": 0.9794, "step": 5376 }, { "epoch": 0.1894694186773787, "grad_norm": 1.5564509630203247, "learning_rate": 1.8702640450199176e-05, "loss": 0.9553, "step": 5377 }, { "epoch": 0.18950465569033714, "grad_norm": 6.650216579437256, "learning_rate": 1.8702078217099764e-05, "loss": 5.8009, "step": 5378 }, { "epoch": 0.18953989270329555, "grad_norm": 1.103208065032959, "learning_rate": 1.870151587065468e-05, "loss": 1.0349, "step": 5379 }, { "epoch": 0.18957512971625395, "grad_norm": 4.469745635986328, "learning_rate": 1.8700953410871238e-05, "loss": 3.9395, "step": 5380 }, { "epoch": 0.18961036672921236, "grad_norm": 0.7963800430297852, "learning_rate": 1.8700390837756768e-05, "loss": 1.1359, "step": 5381 }, { "epoch": 0.18964560374217077, "grad_norm": 6.43704891204834, "learning_rate": 1.86998281513186e-05, "loss": 6.2829, "step": 5382 }, { "epoch": 0.1896808407551292, "grad_norm": 4.907464981079102, "learning_rate": 1.8699265351564064e-05, "loss": 3.802, "step": 5383 }, { "epoch": 0.1897160777680876, "grad_norm": 1.1641473770141602, "learning_rate": 1.869870243850049e-05, "loss": 1.1007, "step": 5384 }, { "epoch": 0.189751314781046, "grad_norm": 2.261897087097168, "learning_rate": 1.8698139412135203e-05, "loss": 3.0223, "step": 5385 }, { "epoch": 0.18978655179400442, "grad_norm": 3.1227290630340576, "learning_rate": 1.8697576272475547e-05, "loss": 3.82, "step": 5386 }, { "epoch": 0.18982178880696282, "grad_norm": 1.3853284120559692, "learning_rate": 1.8697013019528852e-05, "loss": 0.9789, "step": 5387 }, { "epoch": 0.18985702581992125, "grad_norm": 4.659107208251953, "learning_rate": 1.869644965330245e-05, "loss": 6.5043, "step": 5388 }, { "epoch": 0.18989226283287966, "grad_norm": 5.474796772003174, "learning_rate": 1.869588617380369e-05, "loss": 6.8501, "step": 5389 }, { "epoch": 0.18992749984583807, "grad_norm": 1.4624463319778442, "learning_rate": 1.8695322581039903e-05, "loss": 1.037, "step": 5390 }, { "epoch": 0.18996273685879647, "grad_norm": 1.9230073690414429, "learning_rate": 1.869475887501843e-05, "loss": 0.9196, "step": 5391 }, { "epoch": 0.1899979738717549, "grad_norm": 4.108871936798096, "learning_rate": 1.8694195055746617e-05, "loss": 3.9983, "step": 5392 }, { "epoch": 0.1900332108847133, "grad_norm": 1.2303589582443237, "learning_rate": 1.8693631123231806e-05, "loss": 1.028, "step": 5393 }, { "epoch": 0.19006844789767172, "grad_norm": 3.8054239749908447, "learning_rate": 1.8693067077481338e-05, "loss": 6.5536, "step": 5394 }, { "epoch": 0.19010368491063012, "grad_norm": 1.3176342248916626, "learning_rate": 1.8692502918502572e-05, "loss": 0.9431, "step": 5395 }, { "epoch": 0.19013892192358853, "grad_norm": 4.699667930603027, "learning_rate": 1.869193864630284e-05, "loss": 3.746, "step": 5396 }, { "epoch": 0.19017415893654696, "grad_norm": 1.0629146099090576, "learning_rate": 1.8691374260889503e-05, "loss": 0.9509, "step": 5397 }, { "epoch": 0.19020939594950537, "grad_norm": 1.3964449167251587, "learning_rate": 1.869080976226991e-05, "loss": 1.0278, "step": 5398 }, { "epoch": 0.19024463296246377, "grad_norm": 9.252100944519043, "learning_rate": 1.8690245150451414e-05, "loss": 6.6834, "step": 5399 }, { "epoch": 0.19027986997542218, "grad_norm": 4.647685527801514, "learning_rate": 1.8689680425441366e-05, "loss": 3.4962, "step": 5400 }, { "epoch": 0.19031510698838058, "grad_norm": 1.2970716953277588, "learning_rate": 1.868911558724712e-05, "loss": 0.8355, "step": 5401 }, { "epoch": 0.19035034400133902, "grad_norm": 5.164874076843262, "learning_rate": 1.8688550635876045e-05, "loss": 3.9117, "step": 5402 }, { "epoch": 0.19038558101429742, "grad_norm": 1.2188522815704346, "learning_rate": 1.8687985571335487e-05, "loss": 1.1704, "step": 5403 }, { "epoch": 0.19042081802725583, "grad_norm": 4.689910888671875, "learning_rate": 1.868742039363281e-05, "loss": 6.2078, "step": 5404 }, { "epoch": 0.19045605504021423, "grad_norm": 4.70474100112915, "learning_rate": 1.8686855102775375e-05, "loss": 2.8063, "step": 5405 }, { "epoch": 0.19049129205317264, "grad_norm": 3.2962868213653564, "learning_rate": 1.8686289698770544e-05, "loss": 3.3863, "step": 5406 }, { "epoch": 0.19052652906613107, "grad_norm": 4.599512100219727, "learning_rate": 1.8685724181625685e-05, "loss": 4.0296, "step": 5407 }, { "epoch": 0.19056176607908948, "grad_norm": 3.2818896770477295, "learning_rate": 1.8685158551348162e-05, "loss": 3.7322, "step": 5408 }, { "epoch": 0.19059700309204788, "grad_norm": 3.6696934700012207, "learning_rate": 1.8684592807945347e-05, "loss": 5.9965, "step": 5409 }, { "epoch": 0.1906322401050063, "grad_norm": 4.420136451721191, "learning_rate": 1.8684026951424602e-05, "loss": 3.4573, "step": 5410 }, { "epoch": 0.19066747711796472, "grad_norm": 1.130384087562561, "learning_rate": 1.8683460981793296e-05, "loss": 0.842, "step": 5411 }, { "epoch": 0.19070271413092313, "grad_norm": 3.4518682956695557, "learning_rate": 1.868289489905881e-05, "loss": 2.9855, "step": 5412 }, { "epoch": 0.19073795114388153, "grad_norm": 3.925584077835083, "learning_rate": 1.8682328703228507e-05, "loss": 3.905, "step": 5413 }, { "epoch": 0.19077318815683994, "grad_norm": 1.8207440376281738, "learning_rate": 1.8681762394309772e-05, "loss": 2.8837, "step": 5414 }, { "epoch": 0.19080842516979835, "grad_norm": 5.541433334350586, "learning_rate": 1.868119597230997e-05, "loss": 3.5839, "step": 5415 }, { "epoch": 0.19084366218275678, "grad_norm": 1.0633171796798706, "learning_rate": 1.8680629437236488e-05, "loss": 0.8819, "step": 5416 }, { "epoch": 0.19087889919571518, "grad_norm": 5.516900539398193, "learning_rate": 1.86800627890967e-05, "loss": 5.5663, "step": 5417 }, { "epoch": 0.1909141362086736, "grad_norm": 4.631567001342773, "learning_rate": 1.867949602789799e-05, "loss": 5.9977, "step": 5418 }, { "epoch": 0.190949373221632, "grad_norm": 4.2416672706604, "learning_rate": 1.867892915364774e-05, "loss": 3.2491, "step": 5419 }, { "epoch": 0.1909846102345904, "grad_norm": 3.4936392307281494, "learning_rate": 1.867836216635333e-05, "loss": 3.6487, "step": 5420 }, { "epoch": 0.19101984724754884, "grad_norm": 1.9342941045761108, "learning_rate": 1.8677795066022145e-05, "loss": 1.1081, "step": 5421 }, { "epoch": 0.19105508426050724, "grad_norm": 1.5353240966796875, "learning_rate": 1.867722785266158e-05, "loss": 0.9377, "step": 5422 }, { "epoch": 0.19109032127346565, "grad_norm": 2.9798660278320312, "learning_rate": 1.8676660526279013e-05, "loss": 3.8898, "step": 5423 }, { "epoch": 0.19112555828642405, "grad_norm": 1.4124946594238281, "learning_rate": 1.867609308688184e-05, "loss": 1.0539, "step": 5424 }, { "epoch": 0.19116079529938246, "grad_norm": 1.244264841079712, "learning_rate": 1.8675525534477447e-05, "loss": 1.122, "step": 5425 }, { "epoch": 0.1911960323123409, "grad_norm": 1.2295001745224, "learning_rate": 1.8674957869073228e-05, "loss": 1.3359, "step": 5426 }, { "epoch": 0.1912312693252993, "grad_norm": 2.278655767440796, "learning_rate": 1.867439009067658e-05, "loss": 3.1977, "step": 5427 }, { "epoch": 0.1912665063382577, "grad_norm": 4.2454657554626465, "learning_rate": 1.8673822199294897e-05, "loss": 3.837, "step": 5428 }, { "epoch": 0.1913017433512161, "grad_norm": 1.814981460571289, "learning_rate": 1.8673254194935573e-05, "loss": 0.8256, "step": 5429 }, { "epoch": 0.19133698036417454, "grad_norm": 1.386620044708252, "learning_rate": 1.867268607760601e-05, "loss": 1.1859, "step": 5430 }, { "epoch": 0.19137221737713295, "grad_norm": 6.222021579742432, "learning_rate": 1.8672117847313603e-05, "loss": 5.8862, "step": 5431 }, { "epoch": 0.19140745439009135, "grad_norm": 2.4629123210906982, "learning_rate": 1.867154950406576e-05, "loss": 3.7632, "step": 5432 }, { "epoch": 0.19144269140304976, "grad_norm": 1.070385456085205, "learning_rate": 1.8670981047869877e-05, "loss": 1.0931, "step": 5433 }, { "epoch": 0.19147792841600816, "grad_norm": 2.0800046920776367, "learning_rate": 1.8670412478733362e-05, "loss": 1.1206, "step": 5434 }, { "epoch": 0.1915131654289666, "grad_norm": 1.2645031213760376, "learning_rate": 1.866984379666362e-05, "loss": 1.3871, "step": 5435 }, { "epoch": 0.191548402441925, "grad_norm": 5.922311305999756, "learning_rate": 1.8669275001668058e-05, "loss": 7.1089, "step": 5436 }, { "epoch": 0.1915836394548834, "grad_norm": 3.357668161392212, "learning_rate": 1.866870609375409e-05, "loss": 2.5918, "step": 5437 }, { "epoch": 0.19161887646784181, "grad_norm": 4.608434677124023, "learning_rate": 1.8668137072929113e-05, "loss": 4.0289, "step": 5438 }, { "epoch": 0.19165411348080022, "grad_norm": 1.338428020477295, "learning_rate": 1.866756793920055e-05, "loss": 0.902, "step": 5439 }, { "epoch": 0.19168935049375865, "grad_norm": 5.6799092292785645, "learning_rate": 1.8666998692575812e-05, "loss": 6.3065, "step": 5440 }, { "epoch": 0.19172458750671706, "grad_norm": 1.1454373598098755, "learning_rate": 1.866642933306231e-05, "loss": 0.7331, "step": 5441 }, { "epoch": 0.19175982451967546, "grad_norm": 4.086359977722168, "learning_rate": 1.8665859860667462e-05, "loss": 5.3152, "step": 5442 }, { "epoch": 0.19179506153263387, "grad_norm": 2.5748291015625, "learning_rate": 1.8665290275398687e-05, "loss": 3.9474, "step": 5443 }, { "epoch": 0.1918302985455923, "grad_norm": 3.743903160095215, "learning_rate": 1.86647205772634e-05, "loss": 4.1606, "step": 5444 }, { "epoch": 0.1918655355585507, "grad_norm": 4.525179862976074, "learning_rate": 1.8664150766269026e-05, "loss": 5.7125, "step": 5445 }, { "epoch": 0.19190077257150912, "grad_norm": 5.53812837600708, "learning_rate": 1.8663580842422985e-05, "loss": 2.8751, "step": 5446 }, { "epoch": 0.19193600958446752, "grad_norm": 4.019644260406494, "learning_rate": 1.8663010805732696e-05, "loss": 6.807, "step": 5447 }, { "epoch": 0.19197124659742593, "grad_norm": 0.9767042398452759, "learning_rate": 1.8662440656205595e-05, "loss": 1.0018, "step": 5448 }, { "epoch": 0.19200648361038436, "grad_norm": 1.8628137111663818, "learning_rate": 1.8661870393849093e-05, "loss": 1.0089, "step": 5449 }, { "epoch": 0.19204172062334277, "grad_norm": 1.2871193885803223, "learning_rate": 1.8661300018670633e-05, "loss": 1.0295, "step": 5450 }, { "epoch": 0.19207695763630117, "grad_norm": 2.214681625366211, "learning_rate": 1.866072953067763e-05, "loss": 1.2442, "step": 5451 }, { "epoch": 0.19211219464925958, "grad_norm": 9.395074844360352, "learning_rate": 1.8660158929877526e-05, "loss": 8.8833, "step": 5452 }, { "epoch": 0.19214743166221798, "grad_norm": 3.857293128967285, "learning_rate": 1.8659588216277748e-05, "loss": 3.965, "step": 5453 }, { "epoch": 0.19218266867517642, "grad_norm": 7.059340000152588, "learning_rate": 1.865901738988573e-05, "loss": 5.9093, "step": 5454 }, { "epoch": 0.19221790568813482, "grad_norm": 1.166043758392334, "learning_rate": 1.8658446450708908e-05, "loss": 0.7856, "step": 5455 }, { "epoch": 0.19225314270109323, "grad_norm": 4.685854911804199, "learning_rate": 1.8657875398754717e-05, "loss": 6.7536, "step": 5456 }, { "epoch": 0.19228837971405163, "grad_norm": 5.251813888549805, "learning_rate": 1.8657304234030598e-05, "loss": 5.7679, "step": 5457 }, { "epoch": 0.19232361672701004, "grad_norm": 1.2227863073349, "learning_rate": 1.8656732956543985e-05, "loss": 0.8294, "step": 5458 }, { "epoch": 0.19235885373996847, "grad_norm": 6.5802154541015625, "learning_rate": 1.8656161566302328e-05, "loss": 5.8915, "step": 5459 }, { "epoch": 0.19239409075292688, "grad_norm": 8.837882995605469, "learning_rate": 1.865559006331306e-05, "loss": 3.1267, "step": 5460 }, { "epoch": 0.19242932776588528, "grad_norm": 4.5172038078308105, "learning_rate": 1.865501844758363e-05, "loss": 3.0861, "step": 5461 }, { "epoch": 0.1924645647788437, "grad_norm": 1.592944622039795, "learning_rate": 1.8654446719121483e-05, "loss": 0.8139, "step": 5462 }, { "epoch": 0.19249980179180212, "grad_norm": 4.082601070404053, "learning_rate": 1.8653874877934064e-05, "loss": 3.636, "step": 5463 }, { "epoch": 0.19253503880476053, "grad_norm": 1.1330615282058716, "learning_rate": 1.8653302924028823e-05, "loss": 1.142, "step": 5464 }, { "epoch": 0.19257027581771893, "grad_norm": 1.172275424003601, "learning_rate": 1.8652730857413208e-05, "loss": 1.2777, "step": 5465 }, { "epoch": 0.19260551283067734, "grad_norm": 1.1498229503631592, "learning_rate": 1.865215867809467e-05, "loss": 1.0655, "step": 5466 }, { "epoch": 0.19264074984363574, "grad_norm": 1.1545768976211548, "learning_rate": 1.8651586386080668e-05, "loss": 1.0344, "step": 5467 }, { "epoch": 0.19267598685659418, "grad_norm": 4.195642948150635, "learning_rate": 1.8651013981378648e-05, "loss": 3.1363, "step": 5468 }, { "epoch": 0.19271122386955258, "grad_norm": 1.0818990468978882, "learning_rate": 1.865044146399607e-05, "loss": 1.1692, "step": 5469 }, { "epoch": 0.192746460882511, "grad_norm": 3.941831111907959, "learning_rate": 1.864986883394039e-05, "loss": 3.2473, "step": 5470 }, { "epoch": 0.1927816978954694, "grad_norm": 10.787920951843262, "learning_rate": 1.8649296091219065e-05, "loss": 4.1309, "step": 5471 }, { "epoch": 0.1928169349084278, "grad_norm": 3.7453160285949707, "learning_rate": 1.864872323583956e-05, "loss": 3.7114, "step": 5472 }, { "epoch": 0.19285217192138623, "grad_norm": 3.04767107963562, "learning_rate": 1.864815026780933e-05, "loss": 5.0472, "step": 5473 }, { "epoch": 0.19288740893434464, "grad_norm": 4.104214668273926, "learning_rate": 1.864757718713584e-05, "loss": 3.6534, "step": 5474 }, { "epoch": 0.19292264594730305, "grad_norm": 6.998157501220703, "learning_rate": 1.864700399382656e-05, "loss": 9.1172, "step": 5475 }, { "epoch": 0.19295788296026145, "grad_norm": 5.8110527992248535, "learning_rate": 1.8646430687888948e-05, "loss": 3.5627, "step": 5476 }, { "epoch": 0.19299311997321986, "grad_norm": 1.1708158254623413, "learning_rate": 1.8645857269330476e-05, "loss": 1.2537, "step": 5477 }, { "epoch": 0.1930283569861783, "grad_norm": 4.826091289520264, "learning_rate": 1.8645283738158615e-05, "loss": 5.6246, "step": 5478 }, { "epoch": 0.1930635939991367, "grad_norm": 2.0672576427459717, "learning_rate": 1.8644710094380828e-05, "loss": 0.9343, "step": 5479 }, { "epoch": 0.1930988310120951, "grad_norm": 1.3797544240951538, "learning_rate": 1.8644136338004593e-05, "loss": 1.2187, "step": 5480 }, { "epoch": 0.1931340680250535, "grad_norm": 9.540209770202637, "learning_rate": 1.864356246903738e-05, "loss": 6.5758, "step": 5481 }, { "epoch": 0.19316930503801194, "grad_norm": 1.1791889667510986, "learning_rate": 1.8642988487486663e-05, "loss": 0.8171, "step": 5482 }, { "epoch": 0.19320454205097035, "grad_norm": 1.4974781274795532, "learning_rate": 1.8642414393359923e-05, "loss": 0.8162, "step": 5483 }, { "epoch": 0.19323977906392875, "grad_norm": 2.74273419380188, "learning_rate": 1.8641840186664636e-05, "loss": 3.7659, "step": 5484 }, { "epoch": 0.19327501607688716, "grad_norm": 4.3753342628479, "learning_rate": 1.8641265867408272e-05, "loss": 3.844, "step": 5485 }, { "epoch": 0.19331025308984556, "grad_norm": 1.4142627716064453, "learning_rate": 1.864069143559833e-05, "loss": 1.142, "step": 5486 }, { "epoch": 0.193345490102804, "grad_norm": 1.4084064960479736, "learning_rate": 1.8640116891242272e-05, "loss": 1.0839, "step": 5487 }, { "epoch": 0.1933807271157624, "grad_norm": 2.565779209136963, "learning_rate": 1.8639542234347597e-05, "loss": 3.7036, "step": 5488 }, { "epoch": 0.1934159641287208, "grad_norm": 5.059309005737305, "learning_rate": 1.863896746492178e-05, "loss": 4.0562, "step": 5489 }, { "epoch": 0.1934512011416792, "grad_norm": 1.452860713005066, "learning_rate": 1.8638392582972312e-05, "loss": 0.8935, "step": 5490 }, { "epoch": 0.19348643815463762, "grad_norm": 5.274214267730713, "learning_rate": 1.8637817588506684e-05, "loss": 5.6395, "step": 5491 }, { "epoch": 0.19352167516759605, "grad_norm": 3.625114679336548, "learning_rate": 1.863724248153238e-05, "loss": 3.8458, "step": 5492 }, { "epoch": 0.19355691218055446, "grad_norm": 1.3969732522964478, "learning_rate": 1.863666726205689e-05, "loss": 1.0089, "step": 5493 }, { "epoch": 0.19359214919351286, "grad_norm": 1.2893081903457642, "learning_rate": 1.863609193008771e-05, "loss": 0.9725, "step": 5494 }, { "epoch": 0.19362738620647127, "grad_norm": 7.5894365310668945, "learning_rate": 1.8635516485632336e-05, "loss": 6.6995, "step": 5495 }, { "epoch": 0.1936626232194297, "grad_norm": 1.4827826023101807, "learning_rate": 1.8634940928698256e-05, "loss": 0.901, "step": 5496 }, { "epoch": 0.1936978602323881, "grad_norm": 9.417181968688965, "learning_rate": 1.863436525929297e-05, "loss": 5.775, "step": 5497 }, { "epoch": 0.19373309724534651, "grad_norm": 1.2260072231292725, "learning_rate": 1.863378947742398e-05, "loss": 0.7749, "step": 5498 }, { "epoch": 0.19376833425830492, "grad_norm": 1.2157622575759888, "learning_rate": 1.863321358309878e-05, "loss": 0.8742, "step": 5499 }, { "epoch": 0.19380357127126333, "grad_norm": 12.833342552185059, "learning_rate": 1.863263757632487e-05, "loss": 6.0613, "step": 5500 }, { "epoch": 0.19383880828422176, "grad_norm": 6.068732261657715, "learning_rate": 1.8632061457109764e-05, "loss": 6.0101, "step": 5501 }, { "epoch": 0.19387404529718016, "grad_norm": 4.806881904602051, "learning_rate": 1.863148522546095e-05, "loss": 4.0671, "step": 5502 }, { "epoch": 0.19390928231013857, "grad_norm": 6.538262844085693, "learning_rate": 1.8630908881385944e-05, "loss": 2.9664, "step": 5503 }, { "epoch": 0.19394451932309698, "grad_norm": 16.86343002319336, "learning_rate": 1.8630332424892253e-05, "loss": 6.1951, "step": 5504 }, { "epoch": 0.19397975633605538, "grad_norm": 6.480633735656738, "learning_rate": 1.8629755855987382e-05, "loss": 6.1828, "step": 5505 }, { "epoch": 0.19401499334901381, "grad_norm": 8.01082706451416, "learning_rate": 1.862917917467884e-05, "loss": 3.7757, "step": 5506 }, { "epoch": 0.19405023036197222, "grad_norm": 4.382617950439453, "learning_rate": 1.8628602380974137e-05, "loss": 3.7963, "step": 5507 }, { "epoch": 0.19408546737493063, "grad_norm": 4.888894557952881, "learning_rate": 1.86280254748808e-05, "loss": 2.7516, "step": 5508 }, { "epoch": 0.19412070438788903, "grad_norm": 5.4880781173706055, "learning_rate": 1.862744845640632e-05, "loss": 4.1705, "step": 5509 }, { "epoch": 0.19415594140084744, "grad_norm": 3.313551425933838, "learning_rate": 1.862687132555823e-05, "loss": 2.5878, "step": 5510 }, { "epoch": 0.19419117841380587, "grad_norm": 4.261222839355469, "learning_rate": 1.8626294082344042e-05, "loss": 3.4561, "step": 5511 }, { "epoch": 0.19422641542676428, "grad_norm": 1.1272028684616089, "learning_rate": 1.8625716726771274e-05, "loss": 0.9653, "step": 5512 }, { "epoch": 0.19426165243972268, "grad_norm": 4.1094136238098145, "learning_rate": 1.862513925884745e-05, "loss": 3.4918, "step": 5513 }, { "epoch": 0.1942968894526811, "grad_norm": 1.1823912858963013, "learning_rate": 1.8624561678580084e-05, "loss": 0.9269, "step": 5514 }, { "epoch": 0.19433212646563952, "grad_norm": 4.302534103393555, "learning_rate": 1.862398398597671e-05, "loss": 3.6039, "step": 5515 }, { "epoch": 0.19436736347859793, "grad_norm": 3.850132703781128, "learning_rate": 1.8623406181044837e-05, "loss": 5.8844, "step": 5516 }, { "epoch": 0.19440260049155633, "grad_norm": 3.0520482063293457, "learning_rate": 1.8622828263792007e-05, "loss": 4.0667, "step": 5517 }, { "epoch": 0.19443783750451474, "grad_norm": 6.0970916748046875, "learning_rate": 1.8622250234225738e-05, "loss": 6.2849, "step": 5518 }, { "epoch": 0.19447307451747314, "grad_norm": 4.174197673797607, "learning_rate": 1.8621672092353562e-05, "loss": 3.7731, "step": 5519 }, { "epoch": 0.19450831153043158, "grad_norm": 1.3190584182739258, "learning_rate": 1.8621093838183006e-05, "loss": 1.0461, "step": 5520 }, { "epoch": 0.19454354854338998, "grad_norm": 1.233933448791504, "learning_rate": 1.8620515471721607e-05, "loss": 0.8986, "step": 5521 }, { "epoch": 0.1945787855563484, "grad_norm": 1.3817014694213867, "learning_rate": 1.8619936992976896e-05, "loss": 0.9995, "step": 5522 }, { "epoch": 0.1946140225693068, "grad_norm": 3.762903928756714, "learning_rate": 1.861935840195641e-05, "loss": 3.9321, "step": 5523 }, { "epoch": 0.1946492595822652, "grad_norm": 11.399393081665039, "learning_rate": 1.861877969866768e-05, "loss": 5.5021, "step": 5524 }, { "epoch": 0.19468449659522363, "grad_norm": 1.5903165340423584, "learning_rate": 1.8618200883118246e-05, "loss": 0.9154, "step": 5525 }, { "epoch": 0.19471973360818204, "grad_norm": 1.6822750568389893, "learning_rate": 1.861762195531565e-05, "loss": 0.9779, "step": 5526 }, { "epoch": 0.19475497062114044, "grad_norm": 2.5369012355804443, "learning_rate": 1.861704291526743e-05, "loss": 3.673, "step": 5527 }, { "epoch": 0.19479020763409885, "grad_norm": 1.0635448694229126, "learning_rate": 1.861646376298113e-05, "loss": 1.0073, "step": 5528 }, { "epoch": 0.19482544464705726, "grad_norm": 4.251285076141357, "learning_rate": 1.8615884498464285e-05, "loss": 5.0786, "step": 5529 }, { "epoch": 0.1948606816600157, "grad_norm": 2.2871627807617188, "learning_rate": 1.8615305121724453e-05, "loss": 0.8287, "step": 5530 }, { "epoch": 0.1948959186729741, "grad_norm": 1.2081596851348877, "learning_rate": 1.861472563276917e-05, "loss": 1.1847, "step": 5531 }, { "epoch": 0.1949311556859325, "grad_norm": 3.702577829360962, "learning_rate": 1.8614146031605992e-05, "loss": 4.088, "step": 5532 }, { "epoch": 0.1949663926988909, "grad_norm": 1.515548825263977, "learning_rate": 1.8613566318242463e-05, "loss": 1.1303, "step": 5533 }, { "epoch": 0.19500162971184934, "grad_norm": 1.3704769611358643, "learning_rate": 1.8612986492686137e-05, "loss": 1.0017, "step": 5534 }, { "epoch": 0.19503686672480774, "grad_norm": 3.552311897277832, "learning_rate": 1.8612406554944563e-05, "loss": 3.8701, "step": 5535 }, { "epoch": 0.19507210373776615, "grad_norm": 2.0858540534973145, "learning_rate": 1.86118265050253e-05, "loss": 2.9678, "step": 5536 }, { "epoch": 0.19510734075072456, "grad_norm": 1.9314868450164795, "learning_rate": 1.8611246342935896e-05, "loss": 0.9306, "step": 5537 }, { "epoch": 0.19514257776368296, "grad_norm": 2.702329397201538, "learning_rate": 1.8610666068683914e-05, "loss": 3.741, "step": 5538 }, { "epoch": 0.1951778147766414, "grad_norm": 6.559401512145996, "learning_rate": 1.861008568227691e-05, "loss": 6.0781, "step": 5539 }, { "epoch": 0.1952130517895998, "grad_norm": 4.450520038604736, "learning_rate": 1.8609505183722438e-05, "loss": 4.1206, "step": 5540 }, { "epoch": 0.1952482888025582, "grad_norm": 2.434725046157837, "learning_rate": 1.860892457302807e-05, "loss": 3.8482, "step": 5541 }, { "epoch": 0.1952835258155166, "grad_norm": 4.133828163146973, "learning_rate": 1.860834385020136e-05, "loss": 6.2271, "step": 5542 }, { "epoch": 0.19531876282847502, "grad_norm": 9.44981861114502, "learning_rate": 1.8607763015249877e-05, "loss": 6.3306, "step": 5543 }, { "epoch": 0.19535399984143345, "grad_norm": 5.223462104797363, "learning_rate": 1.8607182068181183e-05, "loss": 4.2919, "step": 5544 }, { "epoch": 0.19538923685439186, "grad_norm": 8.292447090148926, "learning_rate": 1.8606601009002844e-05, "loss": 5.7251, "step": 5545 }, { "epoch": 0.19542447386735026, "grad_norm": 3.930065393447876, "learning_rate": 1.8606019837722432e-05, "loss": 3.6996, "step": 5546 }, { "epoch": 0.19545971088030867, "grad_norm": 4.461363315582275, "learning_rate": 1.860543855434752e-05, "loss": 3.5662, "step": 5547 }, { "epoch": 0.1954949478932671, "grad_norm": 6.462808132171631, "learning_rate": 1.860485715888567e-05, "loss": 4.1557, "step": 5548 }, { "epoch": 0.1955301849062255, "grad_norm": 1.3719861507415771, "learning_rate": 1.860427565134446e-05, "loss": 1.0373, "step": 5549 }, { "epoch": 0.1955654219191839, "grad_norm": 3.2750394344329834, "learning_rate": 1.860369403173146e-05, "loss": 3.1076, "step": 5550 }, { "epoch": 0.19560065893214232, "grad_norm": 1.6347286701202393, "learning_rate": 1.8603112300054257e-05, "loss": 1.128, "step": 5551 }, { "epoch": 0.19563589594510072, "grad_norm": 1.3978699445724487, "learning_rate": 1.8602530456320415e-05, "loss": 1.0541, "step": 5552 }, { "epoch": 0.19567113295805916, "grad_norm": 4.128976821899414, "learning_rate": 1.860194850053752e-05, "loss": 3.3989, "step": 5553 }, { "epoch": 0.19570636997101756, "grad_norm": 2.509737968444824, "learning_rate": 1.860136643271315e-05, "loss": 3.6683, "step": 5554 }, { "epoch": 0.19574160698397597, "grad_norm": 1.3320467472076416, "learning_rate": 1.8600784252854886e-05, "loss": 1.0367, "step": 5555 }, { "epoch": 0.19577684399693437, "grad_norm": 1.3530168533325195, "learning_rate": 1.860020196097031e-05, "loss": 1.0838, "step": 5556 }, { "epoch": 0.19581208100989278, "grad_norm": 4.83633279800415, "learning_rate": 1.8599619557067012e-05, "loss": 3.4049, "step": 5557 }, { "epoch": 0.1958473180228512, "grad_norm": 5.335212707519531, "learning_rate": 1.859903704115257e-05, "loss": 5.9602, "step": 5558 }, { "epoch": 0.19588255503580962, "grad_norm": 1.4113999605178833, "learning_rate": 1.8598454413234577e-05, "loss": 1.1063, "step": 5559 }, { "epoch": 0.19591779204876802, "grad_norm": 1.3951770067214966, "learning_rate": 1.859787167332062e-05, "loss": 0.8716, "step": 5560 }, { "epoch": 0.19595302906172643, "grad_norm": 1.1203453540802002, "learning_rate": 1.8597288821418288e-05, "loss": 1.1953, "step": 5561 }, { "epoch": 0.19598826607468484, "grad_norm": 1.958795189857483, "learning_rate": 1.8596705857535175e-05, "loss": 0.8748, "step": 5562 }, { "epoch": 0.19602350308764327, "grad_norm": 3.328721523284912, "learning_rate": 1.8596122781678873e-05, "loss": 3.7399, "step": 5563 }, { "epoch": 0.19605874010060168, "grad_norm": 3.300785779953003, "learning_rate": 1.8595539593856976e-05, "loss": 3.2739, "step": 5564 }, { "epoch": 0.19609397711356008, "grad_norm": 1.476280927658081, "learning_rate": 1.8594956294077083e-05, "loss": 0.8088, "step": 5565 }, { "epoch": 0.1961292141265185, "grad_norm": 1.9545446634292603, "learning_rate": 1.8594372882346787e-05, "loss": 3.6163, "step": 5566 }, { "epoch": 0.19616445113947692, "grad_norm": 4.0933661460876465, "learning_rate": 1.8593789358673694e-05, "loss": 3.1556, "step": 5567 }, { "epoch": 0.19619968815243533, "grad_norm": 1.1718076467514038, "learning_rate": 1.8593205723065397e-05, "loss": 1.1191, "step": 5568 }, { "epoch": 0.19623492516539373, "grad_norm": 5.3459906578063965, "learning_rate": 1.85926219755295e-05, "loss": 6.5646, "step": 5569 }, { "epoch": 0.19627016217835214, "grad_norm": 4.533999919891357, "learning_rate": 1.8592038116073605e-05, "loss": 6.3145, "step": 5570 }, { "epoch": 0.19630539919131054, "grad_norm": 1.1630548238754272, "learning_rate": 1.8591454144705324e-05, "loss": 1.3683, "step": 5571 }, { "epoch": 0.19634063620426898, "grad_norm": 1.0629143714904785, "learning_rate": 1.8590870061432257e-05, "loss": 1.0615, "step": 5572 }, { "epoch": 0.19637587321722738, "grad_norm": 1.2844640016555786, "learning_rate": 1.859028586626201e-05, "loss": 1.0622, "step": 5573 }, { "epoch": 0.1964111102301858, "grad_norm": 3.9442224502563477, "learning_rate": 1.85897015592022e-05, "loss": 2.8568, "step": 5574 }, { "epoch": 0.1964463472431442, "grad_norm": 1.0691413879394531, "learning_rate": 1.8589117140260427e-05, "loss": 0.834, "step": 5575 }, { "epoch": 0.1964815842561026, "grad_norm": 1.4879167079925537, "learning_rate": 1.8588532609444313e-05, "loss": 1.172, "step": 5576 }, { "epoch": 0.19651682126906103, "grad_norm": 5.491315841674805, "learning_rate": 1.8587947966761472e-05, "loss": 3.4368, "step": 5577 }, { "epoch": 0.19655205828201944, "grad_norm": 4.30325984954834, "learning_rate": 1.858736321221951e-05, "loss": 2.3476, "step": 5578 }, { "epoch": 0.19658729529497784, "grad_norm": 1.400832176208496, "learning_rate": 1.8586778345826045e-05, "loss": 0.9894, "step": 5579 }, { "epoch": 0.19662253230793625, "grad_norm": 1.3621182441711426, "learning_rate": 1.8586193367588703e-05, "loss": 0.8484, "step": 5580 }, { "epoch": 0.19665776932089465, "grad_norm": 1.2565621137619019, "learning_rate": 1.85856082775151e-05, "loss": 0.7762, "step": 5581 }, { "epoch": 0.1966930063338531, "grad_norm": 1.3259145021438599, "learning_rate": 1.858502307561285e-05, "loss": 1.0587, "step": 5582 }, { "epoch": 0.1967282433468115, "grad_norm": 8.264845848083496, "learning_rate": 1.8584437761889587e-05, "loss": 6.0713, "step": 5583 }, { "epoch": 0.1967634803597699, "grad_norm": 3.0545637607574463, "learning_rate": 1.8583852336352926e-05, "loss": 3.9585, "step": 5584 }, { "epoch": 0.1967987173727283, "grad_norm": 1.073101282119751, "learning_rate": 1.8583266799010495e-05, "loss": 1.1771, "step": 5585 }, { "epoch": 0.19683395438568674, "grad_norm": 5.210782527923584, "learning_rate": 1.858268114986992e-05, "loss": 5.6493, "step": 5586 }, { "epoch": 0.19686919139864514, "grad_norm": 1.4471107721328735, "learning_rate": 1.858209538893883e-05, "loss": 1.3554, "step": 5587 }, { "epoch": 0.19690442841160355, "grad_norm": 1.414160966873169, "learning_rate": 1.8581509516224853e-05, "loss": 0.9815, "step": 5588 }, { "epoch": 0.19693966542456195, "grad_norm": 6.638930797576904, "learning_rate": 1.8580923531735625e-05, "loss": 5.3463, "step": 5589 }, { "epoch": 0.19697490243752036, "grad_norm": 3.079010009765625, "learning_rate": 1.858033743547877e-05, "loss": 2.4553, "step": 5590 }, { "epoch": 0.1970101394504788, "grad_norm": 3.382376194000244, "learning_rate": 1.857975122746193e-05, "loss": 3.2075, "step": 5591 }, { "epoch": 0.1970453764634372, "grad_norm": 1.2353181838989258, "learning_rate": 1.857916490769274e-05, "loss": 1.0115, "step": 5592 }, { "epoch": 0.1970806134763956, "grad_norm": 6.552923202514648, "learning_rate": 1.857857847617883e-05, "loss": 3.2153, "step": 5593 }, { "epoch": 0.197115850489354, "grad_norm": 1.587033748626709, "learning_rate": 1.8577991932927847e-05, "loss": 1.0824, "step": 5594 }, { "epoch": 0.19715108750231242, "grad_norm": 11.497488021850586, "learning_rate": 1.857740527794742e-05, "loss": 6.6601, "step": 5595 }, { "epoch": 0.19718632451527085, "grad_norm": 7.2978644371032715, "learning_rate": 1.85768185112452e-05, "loss": 6.7041, "step": 5596 }, { "epoch": 0.19722156152822926, "grad_norm": 7.279857158660889, "learning_rate": 1.8576231632828826e-05, "loss": 2.8292, "step": 5597 }, { "epoch": 0.19725679854118766, "grad_norm": 1.7173590660095215, "learning_rate": 1.8575644642705947e-05, "loss": 0.9111, "step": 5598 }, { "epoch": 0.19729203555414607, "grad_norm": 1.2330998182296753, "learning_rate": 1.8575057540884198e-05, "loss": 0.9109, "step": 5599 }, { "epoch": 0.1973272725671045, "grad_norm": 1.3942841291427612, "learning_rate": 1.8574470327371235e-05, "loss": 1.3285, "step": 5600 }, { "epoch": 0.1973625095800629, "grad_norm": 3.4038829803466797, "learning_rate": 1.8573883002174705e-05, "loss": 3.2386, "step": 5601 }, { "epoch": 0.1973977465930213, "grad_norm": 5.3780694007873535, "learning_rate": 1.8573295565302254e-05, "loss": 5.0406, "step": 5602 }, { "epoch": 0.19743298360597972, "grad_norm": 1.6115460395812988, "learning_rate": 1.857270801676154e-05, "loss": 1.1999, "step": 5603 }, { "epoch": 0.19746822061893812, "grad_norm": 1.141356348991394, "learning_rate": 1.8572120356560206e-05, "loss": 0.872, "step": 5604 }, { "epoch": 0.19750345763189656, "grad_norm": 1.490473747253418, "learning_rate": 1.8571532584705916e-05, "loss": 0.9419, "step": 5605 }, { "epoch": 0.19753869464485496, "grad_norm": 3.6430561542510986, "learning_rate": 1.857094470120632e-05, "loss": 3.5575, "step": 5606 }, { "epoch": 0.19757393165781337, "grad_norm": 1.2286784648895264, "learning_rate": 1.857035670606908e-05, "loss": 0.8026, "step": 5607 }, { "epoch": 0.19760916867077177, "grad_norm": 2.025681257247925, "learning_rate": 1.8569768599301854e-05, "loss": 1.0963, "step": 5608 }, { "epoch": 0.19764440568373018, "grad_norm": 3.9720191955566406, "learning_rate": 1.8569180380912296e-05, "loss": 5.9954, "step": 5609 }, { "epoch": 0.1976796426966886, "grad_norm": 3.2282097339630127, "learning_rate": 1.8568592050908078e-05, "loss": 4.1752, "step": 5610 }, { "epoch": 0.19771487970964702, "grad_norm": 4.813391208648682, "learning_rate": 1.856800360929685e-05, "loss": 3.8012, "step": 5611 }, { "epoch": 0.19775011672260542, "grad_norm": 2.905585765838623, "learning_rate": 1.856741505608629e-05, "loss": 4.0593, "step": 5612 }, { "epoch": 0.19778535373556383, "grad_norm": 1.143973469734192, "learning_rate": 1.8566826391284052e-05, "loss": 0.9503, "step": 5613 }, { "epoch": 0.19782059074852223, "grad_norm": 3.649411201477051, "learning_rate": 1.856623761489781e-05, "loss": 4.0247, "step": 5614 }, { "epoch": 0.19785582776148067, "grad_norm": 1.1082783937454224, "learning_rate": 1.8565648726935234e-05, "loss": 0.8173, "step": 5615 }, { "epoch": 0.19789106477443907, "grad_norm": 1.5968042612075806, "learning_rate": 1.8565059727403995e-05, "loss": 0.9159, "step": 5616 }, { "epoch": 0.19792630178739748, "grad_norm": 3.855914354324341, "learning_rate": 1.8564470616311756e-05, "loss": 4.5272, "step": 5617 }, { "epoch": 0.19796153880035589, "grad_norm": 5.297291278839111, "learning_rate": 1.8563881393666197e-05, "loss": 5.9324, "step": 5618 }, { "epoch": 0.19799677581331432, "grad_norm": 1.6796786785125732, "learning_rate": 1.8563292059474995e-05, "loss": 1.0118, "step": 5619 }, { "epoch": 0.19803201282627272, "grad_norm": 1.1880826950073242, "learning_rate": 1.8562702613745823e-05, "loss": 0.9478, "step": 5620 }, { "epoch": 0.19806724983923113, "grad_norm": 4.627697944641113, "learning_rate": 1.8562113056486356e-05, "loss": 3.3701, "step": 5621 }, { "epoch": 0.19810248685218954, "grad_norm": 3.5837841033935547, "learning_rate": 1.8561523387704278e-05, "loss": 2.466, "step": 5622 }, { "epoch": 0.19813772386514794, "grad_norm": 4.496248245239258, "learning_rate": 1.8560933607407266e-05, "loss": 3.0539, "step": 5623 }, { "epoch": 0.19817296087810637, "grad_norm": 12.406963348388672, "learning_rate": 1.8560343715603006e-05, "loss": 4.0218, "step": 5624 }, { "epoch": 0.19820819789106478, "grad_norm": 3.1978938579559326, "learning_rate": 1.8559753712299177e-05, "loss": 3.6115, "step": 5625 }, { "epoch": 0.19824343490402319, "grad_norm": 15.011578559875488, "learning_rate": 1.8559163597503464e-05, "loss": 6.8023, "step": 5626 }, { "epoch": 0.1982786719169816, "grad_norm": 5.387842655181885, "learning_rate": 1.8558573371223558e-05, "loss": 5.7765, "step": 5627 }, { "epoch": 0.19831390892994, "grad_norm": 4.374691963195801, "learning_rate": 1.855798303346714e-05, "loss": 4.0987, "step": 5628 }, { "epoch": 0.19834914594289843, "grad_norm": 2.1041312217712402, "learning_rate": 1.8557392584241905e-05, "loss": 3.0673, "step": 5629 }, { "epoch": 0.19838438295585684, "grad_norm": 1.5039148330688477, "learning_rate": 1.8556802023555543e-05, "loss": 0.817, "step": 5630 }, { "epoch": 0.19841961996881524, "grad_norm": 1.246497392654419, "learning_rate": 1.8556211351415742e-05, "loss": 0.8454, "step": 5631 }, { "epoch": 0.19845485698177365, "grad_norm": 1.1165592670440674, "learning_rate": 1.8555620567830197e-05, "loss": 0.9547, "step": 5632 }, { "epoch": 0.19849009399473205, "grad_norm": 1.726104497909546, "learning_rate": 1.855502967280661e-05, "loss": 1.1025, "step": 5633 }, { "epoch": 0.1985253310076905, "grad_norm": 1.1552815437316895, "learning_rate": 1.8554438666352665e-05, "loss": 1.0157, "step": 5634 }, { "epoch": 0.1985605680206489, "grad_norm": 1.328172206878662, "learning_rate": 1.855384754847607e-05, "loss": 1.287, "step": 5635 }, { "epoch": 0.1985958050336073, "grad_norm": 1.4796620607376099, "learning_rate": 1.8553256319184522e-05, "loss": 0.9669, "step": 5636 }, { "epoch": 0.1986310420465657, "grad_norm": 3.1017627716064453, "learning_rate": 1.8552664978485723e-05, "loss": 3.3847, "step": 5637 }, { "epoch": 0.19866627905952414, "grad_norm": 2.7262914180755615, "learning_rate": 1.855207352638737e-05, "loss": 3.3782, "step": 5638 }, { "epoch": 0.19870151607248254, "grad_norm": 1.2732348442077637, "learning_rate": 1.855148196289717e-05, "loss": 0.9542, "step": 5639 }, { "epoch": 0.19873675308544095, "grad_norm": 1.2710893154144287, "learning_rate": 1.855089028802283e-05, "loss": 0.799, "step": 5640 }, { "epoch": 0.19877199009839935, "grad_norm": 1.3619334697723389, "learning_rate": 1.855029850177205e-05, "loss": 0.8957, "step": 5641 }, { "epoch": 0.19880722711135776, "grad_norm": 3.10259747505188, "learning_rate": 1.8549706604152548e-05, "loss": 2.9835, "step": 5642 }, { "epoch": 0.1988424641243162, "grad_norm": 1.7478876113891602, "learning_rate": 1.8549114595172024e-05, "loss": 0.8436, "step": 5643 }, { "epoch": 0.1988777011372746, "grad_norm": 3.0924453735351562, "learning_rate": 1.8548522474838194e-05, "loss": 3.1425, "step": 5644 }, { "epoch": 0.198912938150233, "grad_norm": 1.3340632915496826, "learning_rate": 1.8547930243158773e-05, "loss": 0.7052, "step": 5645 }, { "epoch": 0.1989481751631914, "grad_norm": 7.556122303009033, "learning_rate": 1.854733790014147e-05, "loss": 4.0647, "step": 5646 }, { "epoch": 0.19898341217614982, "grad_norm": 5.952414512634277, "learning_rate": 1.8546745445794006e-05, "loss": 6.0398, "step": 5647 }, { "epoch": 0.19901864918910825, "grad_norm": 3.3960189819335938, "learning_rate": 1.854615288012409e-05, "loss": 3.7179, "step": 5648 }, { "epoch": 0.19905388620206665, "grad_norm": 6.455657482147217, "learning_rate": 1.8545560203139444e-05, "loss": 9.2918, "step": 5649 }, { "epoch": 0.19908912321502506, "grad_norm": 1.5779787302017212, "learning_rate": 1.854496741484779e-05, "loss": 1.2791, "step": 5650 }, { "epoch": 0.19912436022798347, "grad_norm": 2.7706925868988037, "learning_rate": 1.854437451525684e-05, "loss": 3.9703, "step": 5651 }, { "epoch": 0.1991595972409419, "grad_norm": 4.271864414215088, "learning_rate": 1.8543781504374332e-05, "loss": 3.4335, "step": 5652 }, { "epoch": 0.1991948342539003, "grad_norm": 4.4521942138671875, "learning_rate": 1.854318838220798e-05, "loss": 3.1255, "step": 5653 }, { "epoch": 0.1992300712668587, "grad_norm": 3.036910057067871, "learning_rate": 1.854259514876551e-05, "loss": 3.625, "step": 5654 }, { "epoch": 0.19926530827981712, "grad_norm": 3.7770938873291016, "learning_rate": 1.8542001804054647e-05, "loss": 3.2278, "step": 5655 }, { "epoch": 0.19930054529277552, "grad_norm": 1.5870577096939087, "learning_rate": 1.8541408348083125e-05, "loss": 0.8896, "step": 5656 }, { "epoch": 0.19933578230573396, "grad_norm": 1.5815297365188599, "learning_rate": 1.8540814780858673e-05, "loss": 1.0947, "step": 5657 }, { "epoch": 0.19937101931869236, "grad_norm": 0.8922855854034424, "learning_rate": 1.8540221102389017e-05, "loss": 1.1342, "step": 5658 }, { "epoch": 0.19940625633165077, "grad_norm": 5.371492862701416, "learning_rate": 1.8539627312681896e-05, "loss": 6.7767, "step": 5659 }, { "epoch": 0.19944149334460917, "grad_norm": 5.074540138244629, "learning_rate": 1.8539033411745043e-05, "loss": 3.834, "step": 5660 }, { "epoch": 0.19947673035756758, "grad_norm": 1.7242767810821533, "learning_rate": 1.853843939958619e-05, "loss": 1.1597, "step": 5661 }, { "epoch": 0.199511967370526, "grad_norm": 1.0146247148513794, "learning_rate": 1.8537845276213074e-05, "loss": 1.0322, "step": 5662 }, { "epoch": 0.19954720438348442, "grad_norm": 1.5175166130065918, "learning_rate": 1.8537251041633437e-05, "loss": 0.86, "step": 5663 }, { "epoch": 0.19958244139644282, "grad_norm": 4.65997838973999, "learning_rate": 1.8536656695855016e-05, "loss": 2.9475, "step": 5664 }, { "epoch": 0.19961767840940123, "grad_norm": 7.088008880615234, "learning_rate": 1.853606223888556e-05, "loss": 6.872, "step": 5665 }, { "epoch": 0.19965291542235963, "grad_norm": 1.586513876914978, "learning_rate": 1.8535467670732803e-05, "loss": 1.2016, "step": 5666 }, { "epoch": 0.19968815243531807, "grad_norm": 4.110030174255371, "learning_rate": 1.853487299140449e-05, "loss": 2.8879, "step": 5667 }, { "epoch": 0.19972338944827647, "grad_norm": 7.763827323913574, "learning_rate": 1.853427820090837e-05, "loss": 6.207, "step": 5668 }, { "epoch": 0.19975862646123488, "grad_norm": 5.531436443328857, "learning_rate": 1.8533683299252192e-05, "loss": 4.2817, "step": 5669 }, { "epoch": 0.19979386347419328, "grad_norm": 2.005047559738159, "learning_rate": 1.8533088286443698e-05, "loss": 0.8564, "step": 5670 }, { "epoch": 0.19982910048715172, "grad_norm": 1.7680720090866089, "learning_rate": 1.8532493162490647e-05, "loss": 0.984, "step": 5671 }, { "epoch": 0.19986433750011012, "grad_norm": 5.171701431274414, "learning_rate": 1.8531897927400785e-05, "loss": 3.0422, "step": 5672 }, { "epoch": 0.19989957451306853, "grad_norm": 3.821389675140381, "learning_rate": 1.8531302581181864e-05, "loss": 3.5343, "step": 5673 }, { "epoch": 0.19993481152602693, "grad_norm": 6.510008811950684, "learning_rate": 1.853070712384164e-05, "loss": 6.0438, "step": 5674 }, { "epoch": 0.19997004853898534, "grad_norm": 7.98109245300293, "learning_rate": 1.853011155538787e-05, "loss": 3.7272, "step": 5675 }, { "epoch": 0.20000528555194377, "grad_norm": 5.233918190002441, "learning_rate": 1.8529515875828307e-05, "loss": 3.0048, "step": 5676 }, { "epoch": 0.20004052256490218, "grad_norm": 3.428715467453003, "learning_rate": 1.8528920085170717e-05, "loss": 2.8303, "step": 5677 }, { "epoch": 0.20007575957786058, "grad_norm": 1.018355369567871, "learning_rate": 1.8528324183422856e-05, "loss": 1.1111, "step": 5678 }, { "epoch": 0.200110996590819, "grad_norm": 0.8933417797088623, "learning_rate": 1.8527728170592482e-05, "loss": 1.1782, "step": 5679 }, { "epoch": 0.2001462336037774, "grad_norm": 1.040324091911316, "learning_rate": 1.8527132046687367e-05, "loss": 0.9025, "step": 5680 }, { "epoch": 0.20018147061673583, "grad_norm": 1.38370943069458, "learning_rate": 1.852653581171527e-05, "loss": 0.9125, "step": 5681 }, { "epoch": 0.20021670762969423, "grad_norm": 3.0802204608917236, "learning_rate": 1.8525939465683957e-05, "loss": 6.2925, "step": 5682 }, { "epoch": 0.20025194464265264, "grad_norm": 2.8018577098846436, "learning_rate": 1.85253430086012e-05, "loss": 3.7759, "step": 5683 }, { "epoch": 0.20028718165561105, "grad_norm": 2.295659303665161, "learning_rate": 1.852474644047476e-05, "loss": 2.6083, "step": 5684 }, { "epoch": 0.20032241866856945, "grad_norm": 3.7728683948516846, "learning_rate": 1.8524149761312412e-05, "loss": 3.2797, "step": 5685 }, { "epoch": 0.20035765568152789, "grad_norm": 2.2753005027770996, "learning_rate": 1.852355297112193e-05, "loss": 3.5686, "step": 5686 }, { "epoch": 0.2003928926944863, "grad_norm": 2.3080058097839355, "learning_rate": 1.8522956069911082e-05, "loss": 2.762, "step": 5687 }, { "epoch": 0.2004281297074447, "grad_norm": 2.8967573642730713, "learning_rate": 1.852235905768765e-05, "loss": 2.9715, "step": 5688 }, { "epoch": 0.2004633667204031, "grad_norm": 1.334884762763977, "learning_rate": 1.8521761934459398e-05, "loss": 0.869, "step": 5689 }, { "epoch": 0.20049860373336154, "grad_norm": 1.1104803085327148, "learning_rate": 1.852116470023412e-05, "loss": 1.038, "step": 5690 }, { "epoch": 0.20053384074631994, "grad_norm": 5.306014060974121, "learning_rate": 1.8520567355019582e-05, "loss": 5.9358, "step": 5691 }, { "epoch": 0.20056907775927835, "grad_norm": 4.271637439727783, "learning_rate": 1.8519969898823568e-05, "loss": 3.5863, "step": 5692 }, { "epoch": 0.20060431477223675, "grad_norm": 1.376979947090149, "learning_rate": 1.851937233165386e-05, "loss": 0.9422, "step": 5693 }, { "epoch": 0.20063955178519516, "grad_norm": 4.799999237060547, "learning_rate": 1.8518774653518247e-05, "loss": 3.6418, "step": 5694 }, { "epoch": 0.2006747887981536, "grad_norm": 14.162007331848145, "learning_rate": 1.8518176864424505e-05, "loss": 3.8587, "step": 5695 }, { "epoch": 0.200710025811112, "grad_norm": 1.4399234056472778, "learning_rate": 1.8517578964380428e-05, "loss": 0.9774, "step": 5696 }, { "epoch": 0.2007452628240704, "grad_norm": 1.1535409688949585, "learning_rate": 1.8516980953393795e-05, "loss": 1.0177, "step": 5697 }, { "epoch": 0.2007804998370288, "grad_norm": 10.500577926635742, "learning_rate": 1.8516382831472405e-05, "loss": 4.1293, "step": 5698 }, { "epoch": 0.20081573684998721, "grad_norm": 1.3448278903961182, "learning_rate": 1.851578459862404e-05, "loss": 1.0136, "step": 5699 }, { "epoch": 0.20085097386294565, "grad_norm": 1.1931318044662476, "learning_rate": 1.85151862548565e-05, "loss": 0.751, "step": 5700 }, { "epoch": 0.20088621087590405, "grad_norm": 1.5865925550460815, "learning_rate": 1.851458780017757e-05, "loss": 3.6845, "step": 5701 }, { "epoch": 0.20092144788886246, "grad_norm": 8.507208824157715, "learning_rate": 1.851398923459505e-05, "loss": 4.278, "step": 5702 }, { "epoch": 0.20095668490182086, "grad_norm": 0.9486337304115295, "learning_rate": 1.851339055811674e-05, "loss": 0.9028, "step": 5703 }, { "epoch": 0.2009919219147793, "grad_norm": 1.1454646587371826, "learning_rate": 1.851279177075043e-05, "loss": 0.7703, "step": 5704 }, { "epoch": 0.2010271589277377, "grad_norm": 8.996068954467773, "learning_rate": 1.851219287250392e-05, "loss": 6.0045, "step": 5705 }, { "epoch": 0.2010623959406961, "grad_norm": 5.761268615722656, "learning_rate": 1.8511593863385018e-05, "loss": 6.2845, "step": 5706 }, { "epoch": 0.20109763295365451, "grad_norm": 8.004684448242188, "learning_rate": 1.8510994743401522e-05, "loss": 3.274, "step": 5707 }, { "epoch": 0.20113286996661292, "grad_norm": 8.55846881866455, "learning_rate": 1.8510395512561234e-05, "loss": 6.1982, "step": 5708 }, { "epoch": 0.20116810697957135, "grad_norm": 0.7483789920806885, "learning_rate": 1.850979617087196e-05, "loss": 1.1177, "step": 5709 }, { "epoch": 0.20120334399252976, "grad_norm": 3.925821542739868, "learning_rate": 1.850919671834151e-05, "loss": 3.1917, "step": 5710 }, { "epoch": 0.20123858100548817, "grad_norm": 3.0661022663116455, "learning_rate": 1.850859715497768e-05, "loss": 3.58, "step": 5711 }, { "epoch": 0.20127381801844657, "grad_norm": 1.3426672220230103, "learning_rate": 1.8507997480788296e-05, "loss": 0.8679, "step": 5712 }, { "epoch": 0.20130905503140498, "grad_norm": 4.076005935668945, "learning_rate": 1.850739769578116e-05, "loss": 4.0987, "step": 5713 }, { "epoch": 0.2013442920443634, "grad_norm": 1.6894500255584717, "learning_rate": 1.8506797799964084e-05, "loss": 1.1174, "step": 5714 }, { "epoch": 0.20137952905732182, "grad_norm": 1.6554768085479736, "learning_rate": 1.8506197793344883e-05, "loss": 0.7917, "step": 5715 }, { "epoch": 0.20141476607028022, "grad_norm": 4.669020175933838, "learning_rate": 1.8505597675931373e-05, "loss": 6.0939, "step": 5716 }, { "epoch": 0.20145000308323863, "grad_norm": 6.288637161254883, "learning_rate": 1.8504997447731372e-05, "loss": 5.3694, "step": 5717 }, { "epoch": 0.20148524009619703, "grad_norm": 2.0704426765441895, "learning_rate": 1.850439710875269e-05, "loss": 0.9417, "step": 5718 }, { "epoch": 0.20152047710915547, "grad_norm": 2.6334736347198486, "learning_rate": 1.8503796659003156e-05, "loss": 3.3235, "step": 5719 }, { "epoch": 0.20155571412211387, "grad_norm": 8.245495796203613, "learning_rate": 1.850319609849059e-05, "loss": 6.4564, "step": 5720 }, { "epoch": 0.20159095113507228, "grad_norm": 6.304879188537598, "learning_rate": 1.850259542722281e-05, "loss": 3.3961, "step": 5721 }, { "epoch": 0.20162618814803068, "grad_norm": 7.0429205894470215, "learning_rate": 1.850199464520764e-05, "loss": 6.3462, "step": 5722 }, { "epoch": 0.20166142516098912, "grad_norm": 3.3357491493225098, "learning_rate": 1.8501393752452906e-05, "loss": 3.2352, "step": 5723 }, { "epoch": 0.20169666217394752, "grad_norm": 3.230720043182373, "learning_rate": 1.850079274896644e-05, "loss": 3.1839, "step": 5724 }, { "epoch": 0.20173189918690593, "grad_norm": 10.829914093017578, "learning_rate": 1.850019163475606e-05, "loss": 5.6879, "step": 5725 }, { "epoch": 0.20176713619986433, "grad_norm": 1.3907581567764282, "learning_rate": 1.8499590409829605e-05, "loss": 1.0619, "step": 5726 }, { "epoch": 0.20180237321282274, "grad_norm": 1.296996831893921, "learning_rate": 1.84989890741949e-05, "loss": 1.0102, "step": 5727 }, { "epoch": 0.20183761022578117, "grad_norm": 2.029762029647827, "learning_rate": 1.849838762785978e-05, "loss": 1.0624, "step": 5728 }, { "epoch": 0.20187284723873958, "grad_norm": 6.023399353027344, "learning_rate": 1.849778607083208e-05, "loss": 5.7439, "step": 5729 }, { "epoch": 0.20190808425169798, "grad_norm": 4.463565349578857, "learning_rate": 1.849718440311963e-05, "loss": 6.3138, "step": 5730 }, { "epoch": 0.2019433212646564, "grad_norm": 4.321865081787109, "learning_rate": 1.8496582624730278e-05, "loss": 3.959, "step": 5731 }, { "epoch": 0.2019785582776148, "grad_norm": 4.470033645629883, "learning_rate": 1.8495980735671848e-05, "loss": 3.4467, "step": 5732 }, { "epoch": 0.20201379529057323, "grad_norm": 5.410175323486328, "learning_rate": 1.849537873595219e-05, "loss": 3.3601, "step": 5733 }, { "epoch": 0.20204903230353163, "grad_norm": 2.1170687675476074, "learning_rate": 1.8494776625579143e-05, "loss": 3.9228, "step": 5734 }, { "epoch": 0.20208426931649004, "grad_norm": 4.928195953369141, "learning_rate": 1.8494174404560546e-05, "loss": 3.754, "step": 5735 }, { "epoch": 0.20211950632944845, "grad_norm": 6.077911376953125, "learning_rate": 1.8493572072904243e-05, "loss": 4.2358, "step": 5736 }, { "epoch": 0.20215474334240685, "grad_norm": 7.147084712982178, "learning_rate": 1.849296963061809e-05, "loss": 3.1943, "step": 5737 }, { "epoch": 0.20218998035536528, "grad_norm": 5.1938157081604, "learning_rate": 1.849236707770992e-05, "loss": 3.0107, "step": 5738 }, { "epoch": 0.2022252173683237, "grad_norm": 3.0992114543914795, "learning_rate": 1.8491764414187584e-05, "loss": 3.699, "step": 5739 }, { "epoch": 0.2022604543812821, "grad_norm": 4.627120018005371, "learning_rate": 1.8491161640058942e-05, "loss": 3.4195, "step": 5740 }, { "epoch": 0.2022956913942405, "grad_norm": 2.339402914047241, "learning_rate": 1.8490558755331833e-05, "loss": 4.0986, "step": 5741 }, { "epoch": 0.20233092840719893, "grad_norm": 2.2144880294799805, "learning_rate": 1.8489955760014116e-05, "loss": 3.6917, "step": 5742 }, { "epoch": 0.20236616542015734, "grad_norm": 4.694209575653076, "learning_rate": 1.8489352654113646e-05, "loss": 3.5336, "step": 5743 }, { "epoch": 0.20240140243311575, "grad_norm": 7.767014980316162, "learning_rate": 1.8488749437638277e-05, "loss": 7.3406, "step": 5744 }, { "epoch": 0.20243663944607415, "grad_norm": 1.024317979812622, "learning_rate": 1.848814611059586e-05, "loss": 1.1372, "step": 5745 }, { "epoch": 0.20247187645903256, "grad_norm": 1.208748698234558, "learning_rate": 1.8487542672994264e-05, "loss": 1.029, "step": 5746 }, { "epoch": 0.202507113471991, "grad_norm": 1.3960105180740356, "learning_rate": 1.8486939124841338e-05, "loss": 0.8809, "step": 5747 }, { "epoch": 0.2025423504849494, "grad_norm": 4.117757320404053, "learning_rate": 1.848633546614495e-05, "loss": 6.7973, "step": 5748 }, { "epoch": 0.2025775874979078, "grad_norm": 2.558195114135742, "learning_rate": 1.8485731696912966e-05, "loss": 2.7464, "step": 5749 }, { "epoch": 0.2026128245108662, "grad_norm": 1.7936856746673584, "learning_rate": 1.848512781715324e-05, "loss": 3.3112, "step": 5750 }, { "epoch": 0.2026480615238246, "grad_norm": 0.9873171448707581, "learning_rate": 1.848452382687365e-05, "loss": 1.1664, "step": 5751 }, { "epoch": 0.20268329853678305, "grad_norm": 4.565374851226807, "learning_rate": 1.8483919726082053e-05, "loss": 3.6631, "step": 5752 }, { "epoch": 0.20271853554974145, "grad_norm": 2.5225627422332764, "learning_rate": 1.848331551478632e-05, "loss": 3.1051, "step": 5753 }, { "epoch": 0.20275377256269986, "grad_norm": 2.7897889614105225, "learning_rate": 1.848271119299432e-05, "loss": 3.2196, "step": 5754 }, { "epoch": 0.20278900957565826, "grad_norm": 0.9503687024116516, "learning_rate": 1.848210676071393e-05, "loss": 1.3, "step": 5755 }, { "epoch": 0.2028242465886167, "grad_norm": 2.344949960708618, "learning_rate": 1.8481502217953017e-05, "loss": 3.0119, "step": 5756 }, { "epoch": 0.2028594836015751, "grad_norm": 4.335964202880859, "learning_rate": 1.848089756471946e-05, "loss": 6.9554, "step": 5757 }, { "epoch": 0.2028947206145335, "grad_norm": 5.015671253204346, "learning_rate": 1.8480292801021133e-05, "loss": 3.5903, "step": 5758 }, { "epoch": 0.2029299576274919, "grad_norm": 4.151744365692139, "learning_rate": 1.8479687926865907e-05, "loss": 3.744, "step": 5759 }, { "epoch": 0.20296519464045032, "grad_norm": 5.159723281860352, "learning_rate": 1.8479082942261665e-05, "loss": 3.1134, "step": 5760 }, { "epoch": 0.20300043165340875, "grad_norm": 1.9025778770446777, "learning_rate": 1.8478477847216292e-05, "loss": 3.5143, "step": 5761 }, { "epoch": 0.20303566866636716, "grad_norm": 4.17675256729126, "learning_rate": 1.8477872641737665e-05, "loss": 3.2595, "step": 5762 }, { "epoch": 0.20307090567932556, "grad_norm": 1.2695949077606201, "learning_rate": 1.8477267325833663e-05, "loss": 1.0656, "step": 5763 }, { "epoch": 0.20310614269228397, "grad_norm": 4.751701831817627, "learning_rate": 1.8476661899512177e-05, "loss": 3.114, "step": 5764 }, { "epoch": 0.20314137970524238, "grad_norm": 1.4748339653015137, "learning_rate": 1.8476056362781088e-05, "loss": 1.1036, "step": 5765 }, { "epoch": 0.2031766167182008, "grad_norm": 1.6881259679794312, "learning_rate": 1.8475450715648287e-05, "loss": 1.1181, "step": 5766 }, { "epoch": 0.20321185373115921, "grad_norm": 1.4872528314590454, "learning_rate": 1.8474844958121662e-05, "loss": 0.896, "step": 5767 }, { "epoch": 0.20324709074411762, "grad_norm": 5.082772731781006, "learning_rate": 1.8474239090209097e-05, "loss": 3.4491, "step": 5768 }, { "epoch": 0.20328232775707603, "grad_norm": 1.215807557106018, "learning_rate": 1.8473633111918493e-05, "loss": 0.9952, "step": 5769 }, { "epoch": 0.20331756477003443, "grad_norm": 7.182880401611328, "learning_rate": 1.8473027023257737e-05, "loss": 3.5195, "step": 5770 }, { "epoch": 0.20335280178299286, "grad_norm": 6.91159200668335, "learning_rate": 1.8472420824234728e-05, "loss": 5.7758, "step": 5771 }, { "epoch": 0.20338803879595127, "grad_norm": 3.9215750694274902, "learning_rate": 1.8471814514857352e-05, "loss": 3.5639, "step": 5772 }, { "epoch": 0.20342327580890968, "grad_norm": 4.471780300140381, "learning_rate": 1.847120809513352e-05, "loss": 3.2687, "step": 5773 }, { "epoch": 0.20345851282186808, "grad_norm": 3.645725965499878, "learning_rate": 1.8470601565071117e-05, "loss": 3.1737, "step": 5774 }, { "epoch": 0.20349374983482652, "grad_norm": 2.784933090209961, "learning_rate": 1.8469994924678053e-05, "loss": 3.207, "step": 5775 }, { "epoch": 0.20352898684778492, "grad_norm": 1.534568428993225, "learning_rate": 1.8469388173962228e-05, "loss": 1.0885, "step": 5776 }, { "epoch": 0.20356422386074333, "grad_norm": 5.36988639831543, "learning_rate": 1.8468781312931538e-05, "loss": 2.9715, "step": 5777 }, { "epoch": 0.20359946087370173, "grad_norm": 6.961754322052002, "learning_rate": 1.8468174341593896e-05, "loss": 7.9678, "step": 5778 }, { "epoch": 0.20363469788666014, "grad_norm": 1.292141318321228, "learning_rate": 1.8467567259957206e-05, "loss": 0.9075, "step": 5779 }, { "epoch": 0.20366993489961857, "grad_norm": 1.8929884433746338, "learning_rate": 1.846696006802937e-05, "loss": 1.0129, "step": 5780 }, { "epoch": 0.20370517191257698, "grad_norm": 1.1245476007461548, "learning_rate": 1.8466352765818302e-05, "loss": 1.1077, "step": 5781 }, { "epoch": 0.20374040892553538, "grad_norm": 1.5384470224380493, "learning_rate": 1.846574535333191e-05, "loss": 1.0349, "step": 5782 }, { "epoch": 0.2037756459384938, "grad_norm": 1.7081447839736938, "learning_rate": 1.846513783057811e-05, "loss": 3.1273, "step": 5783 }, { "epoch": 0.2038108829514522, "grad_norm": 4.988692283630371, "learning_rate": 1.846453019756481e-05, "loss": 3.3262, "step": 5784 }, { "epoch": 0.20384611996441063, "grad_norm": 4.123498439788818, "learning_rate": 1.8463922454299922e-05, "loss": 3.4843, "step": 5785 }, { "epoch": 0.20388135697736903, "grad_norm": 2.6536545753479004, "learning_rate": 1.846331460079137e-05, "loss": 3.506, "step": 5786 }, { "epoch": 0.20391659399032744, "grad_norm": 4.276508808135986, "learning_rate": 1.8462706637047064e-05, "loss": 8.5449, "step": 5787 }, { "epoch": 0.20395183100328584, "grad_norm": 6.09000825881958, "learning_rate": 1.8462098563074932e-05, "loss": 6.1106, "step": 5788 }, { "epoch": 0.20398706801624425, "grad_norm": 7.784259796142578, "learning_rate": 1.846149037888288e-05, "loss": 3.9329, "step": 5789 }, { "epoch": 0.20402230502920268, "grad_norm": 1.2082180976867676, "learning_rate": 1.8460882084478843e-05, "loss": 1.2053, "step": 5790 }, { "epoch": 0.2040575420421611, "grad_norm": 1.0952649116516113, "learning_rate": 1.846027367987074e-05, "loss": 0.771, "step": 5791 }, { "epoch": 0.2040927790551195, "grad_norm": 1.0055105686187744, "learning_rate": 1.8459665165066494e-05, "loss": 1.0486, "step": 5792 }, { "epoch": 0.2041280160680779, "grad_norm": 4.380404949188232, "learning_rate": 1.8459056540074032e-05, "loss": 6.4783, "step": 5793 }, { "epoch": 0.20416325308103633, "grad_norm": 2.720712423324585, "learning_rate": 1.845844780490128e-05, "loss": 3.4772, "step": 5794 }, { "epoch": 0.20419849009399474, "grad_norm": 1.1370075941085815, "learning_rate": 1.8457838959556167e-05, "loss": 1.1455, "step": 5795 }, { "epoch": 0.20423372710695314, "grad_norm": 3.9880290031433105, "learning_rate": 1.8457230004046628e-05, "loss": 5.47, "step": 5796 }, { "epoch": 0.20426896411991155, "grad_norm": 6.223191261291504, "learning_rate": 1.8456620938380585e-05, "loss": 6.5785, "step": 5797 }, { "epoch": 0.20430420113286996, "grad_norm": 0.9682824611663818, "learning_rate": 1.8456011762565982e-05, "loss": 1.0286, "step": 5798 }, { "epoch": 0.2043394381458284, "grad_norm": 1.0353893041610718, "learning_rate": 1.8455402476610747e-05, "loss": 0.8492, "step": 5799 }, { "epoch": 0.2043746751587868, "grad_norm": 1.197681188583374, "learning_rate": 1.8454793080522817e-05, "loss": 1.1282, "step": 5800 }, { "epoch": 0.2044099121717452, "grad_norm": 4.449649333953857, "learning_rate": 1.8454183574310128e-05, "loss": 3.4152, "step": 5801 }, { "epoch": 0.2044451491847036, "grad_norm": 1.2078882455825806, "learning_rate": 1.8453573957980625e-05, "loss": 1.0866, "step": 5802 }, { "epoch": 0.204480386197662, "grad_norm": 5.329006195068359, "learning_rate": 1.8452964231542242e-05, "loss": 3.9364, "step": 5803 }, { "epoch": 0.20451562321062045, "grad_norm": 5.2239885330200195, "learning_rate": 1.8452354395002926e-05, "loss": 2.9555, "step": 5804 }, { "epoch": 0.20455086022357885, "grad_norm": 1.3322639465332031, "learning_rate": 1.8451744448370614e-05, "loss": 3.0456, "step": 5805 }, { "epoch": 0.20458609723653726, "grad_norm": 1.5389422178268433, "learning_rate": 1.8451134391653256e-05, "loss": 0.7441, "step": 5806 }, { "epoch": 0.20462133424949566, "grad_norm": 3.9465417861938477, "learning_rate": 1.8450524224858795e-05, "loss": 5.7342, "step": 5807 }, { "epoch": 0.2046565712624541, "grad_norm": 3.490678071975708, "learning_rate": 1.844991394799518e-05, "loss": 3.0525, "step": 5808 }, { "epoch": 0.2046918082754125, "grad_norm": 2.0802063941955566, "learning_rate": 1.8449303561070363e-05, "loss": 0.991, "step": 5809 }, { "epoch": 0.2047270452883709, "grad_norm": 10.832542419433594, "learning_rate": 1.8448693064092285e-05, "loss": 5.6706, "step": 5810 }, { "epoch": 0.2047622823013293, "grad_norm": 1.2876311540603638, "learning_rate": 1.8448082457068907e-05, "loss": 1.0701, "step": 5811 }, { "epoch": 0.20479751931428772, "grad_norm": 1.4101659059524536, "learning_rate": 1.844747174000818e-05, "loss": 0.8498, "step": 5812 }, { "epoch": 0.20483275632724615, "grad_norm": 2.6376960277557373, "learning_rate": 1.844686091291805e-05, "loss": 3.8506, "step": 5813 }, { "epoch": 0.20486799334020456, "grad_norm": 3.3555779457092285, "learning_rate": 1.844624997580649e-05, "loss": 3.449, "step": 5814 }, { "epoch": 0.20490323035316296, "grad_norm": 2.7955970764160156, "learning_rate": 1.8445638928681444e-05, "loss": 3.0233, "step": 5815 }, { "epoch": 0.20493846736612137, "grad_norm": 4.230649948120117, "learning_rate": 1.8445027771550874e-05, "loss": 3.2191, "step": 5816 }, { "epoch": 0.20497370437907977, "grad_norm": 1.6153203248977661, "learning_rate": 1.8444416504422743e-05, "loss": 0.9668, "step": 5817 }, { "epoch": 0.2050089413920382, "grad_norm": 1.2779587507247925, "learning_rate": 1.844380512730501e-05, "loss": 1.0531, "step": 5818 }, { "epoch": 0.2050441784049966, "grad_norm": 1.3760275840759277, "learning_rate": 1.8443193640205642e-05, "loss": 0.8927, "step": 5819 }, { "epoch": 0.20507941541795502, "grad_norm": 1.6219223737716675, "learning_rate": 1.84425820431326e-05, "loss": 0.9517, "step": 5820 }, { "epoch": 0.20511465243091342, "grad_norm": 8.271270751953125, "learning_rate": 1.8441970336093856e-05, "loss": 8.5612, "step": 5821 }, { "epoch": 0.20514988944387183, "grad_norm": 1.3551819324493408, "learning_rate": 1.8441358519097367e-05, "loss": 1.1029, "step": 5822 }, { "epoch": 0.20518512645683026, "grad_norm": 5.457330226898193, "learning_rate": 1.844074659215111e-05, "loss": 3.2758, "step": 5823 }, { "epoch": 0.20522036346978867, "grad_norm": 5.976620674133301, "learning_rate": 1.8440134555263053e-05, "loss": 7.0191, "step": 5824 }, { "epoch": 0.20525560048274707, "grad_norm": 6.0242533683776855, "learning_rate": 1.843952240844117e-05, "loss": 6.0115, "step": 5825 }, { "epoch": 0.20529083749570548, "grad_norm": 1.4458891153335571, "learning_rate": 1.843891015169343e-05, "loss": 1.252, "step": 5826 }, { "epoch": 0.2053260745086639, "grad_norm": 1.1428865194320679, "learning_rate": 1.8438297785027812e-05, "loss": 0.871, "step": 5827 }, { "epoch": 0.20536131152162232, "grad_norm": 1.4253426790237427, "learning_rate": 1.843768530845229e-05, "loss": 1.1898, "step": 5828 }, { "epoch": 0.20539654853458073, "grad_norm": 1.1573905944824219, "learning_rate": 1.843707272197484e-05, "loss": 0.9658, "step": 5829 }, { "epoch": 0.20543178554753913, "grad_norm": 1.2070069313049316, "learning_rate": 1.8436460025603447e-05, "loss": 0.9107, "step": 5830 }, { "epoch": 0.20546702256049754, "grad_norm": 6.619481086730957, "learning_rate": 1.8435847219346086e-05, "loss": 8.152, "step": 5831 }, { "epoch": 0.20550225957345597, "grad_norm": 1.0778061151504517, "learning_rate": 1.843523430321074e-05, "loss": 1.0024, "step": 5832 }, { "epoch": 0.20553749658641438, "grad_norm": 4.261948585510254, "learning_rate": 1.8434621277205395e-05, "loss": 3.6164, "step": 5833 }, { "epoch": 0.20557273359937278, "grad_norm": 9.024306297302246, "learning_rate": 1.843400814133803e-05, "loss": 3.4758, "step": 5834 }, { "epoch": 0.2056079706123312, "grad_norm": 7.120002269744873, "learning_rate": 1.8433394895616635e-05, "loss": 6.0486, "step": 5835 }, { "epoch": 0.2056432076252896, "grad_norm": 1.8123706579208374, "learning_rate": 1.84327815400492e-05, "loss": 4.0958, "step": 5836 }, { "epoch": 0.20567844463824803, "grad_norm": 1.1586114168167114, "learning_rate": 1.843216807464371e-05, "loss": 1.0016, "step": 5837 }, { "epoch": 0.20571368165120643, "grad_norm": 9.498849868774414, "learning_rate": 1.8431554499408158e-05, "loss": 7.0514, "step": 5838 }, { "epoch": 0.20574891866416484, "grad_norm": 5.365651607513428, "learning_rate": 1.843094081435053e-05, "loss": 3.3893, "step": 5839 }, { "epoch": 0.20578415567712324, "grad_norm": 2.8501272201538086, "learning_rate": 1.8430327019478826e-05, "loss": 3.2553, "step": 5840 }, { "epoch": 0.20581939269008165, "grad_norm": 1.0596637725830078, "learning_rate": 1.842971311480104e-05, "loss": 1.0319, "step": 5841 }, { "epoch": 0.20585462970304008, "grad_norm": 1.7422513961791992, "learning_rate": 1.8429099100325167e-05, "loss": 0.7816, "step": 5842 }, { "epoch": 0.2058898667159985, "grad_norm": 1.5648523569107056, "learning_rate": 1.8428484976059207e-05, "loss": 1.0384, "step": 5843 }, { "epoch": 0.2059251037289569, "grad_norm": 8.248551368713379, "learning_rate": 1.842787074201115e-05, "loss": 5.7005, "step": 5844 }, { "epoch": 0.2059603407419153, "grad_norm": 4.127204418182373, "learning_rate": 1.8427256398189008e-05, "loss": 3.8323, "step": 5845 }, { "epoch": 0.20599557775487373, "grad_norm": 1.0604716539382935, "learning_rate": 1.8426641944600775e-05, "loss": 1.0056, "step": 5846 }, { "epoch": 0.20603081476783214, "grad_norm": 1.4320359230041504, "learning_rate": 1.8426027381254463e-05, "loss": 0.8412, "step": 5847 }, { "epoch": 0.20606605178079054, "grad_norm": 1.8029017448425293, "learning_rate": 1.8425412708158065e-05, "loss": 0.8541, "step": 5848 }, { "epoch": 0.20610128879374895, "grad_norm": 5.327215194702148, "learning_rate": 1.8424797925319597e-05, "loss": 3.0632, "step": 5849 }, { "epoch": 0.20613652580670735, "grad_norm": 8.024277687072754, "learning_rate": 1.842418303274706e-05, "loss": 3.5881, "step": 5850 }, { "epoch": 0.2061717628196658, "grad_norm": 3.474656105041504, "learning_rate": 1.8423568030448466e-05, "loss": 3.021, "step": 5851 }, { "epoch": 0.2062069998326242, "grad_norm": 3.77864933013916, "learning_rate": 1.8422952918431826e-05, "loss": 3.6175, "step": 5852 }, { "epoch": 0.2062422368455826, "grad_norm": 1.487597107887268, "learning_rate": 1.8422337696705154e-05, "loss": 1.3179, "step": 5853 }, { "epoch": 0.206277473858541, "grad_norm": 1.280488133430481, "learning_rate": 1.8421722365276457e-05, "loss": 0.9949, "step": 5854 }, { "epoch": 0.2063127108714994, "grad_norm": 1.059118628501892, "learning_rate": 1.8421106924153754e-05, "loss": 0.9274, "step": 5855 }, { "epoch": 0.20634794788445784, "grad_norm": 1.283667802810669, "learning_rate": 1.8420491373345065e-05, "loss": 0.9062, "step": 5856 }, { "epoch": 0.20638318489741625, "grad_norm": 1.1218675374984741, "learning_rate": 1.8419875712858396e-05, "loss": 0.9309, "step": 5857 }, { "epoch": 0.20641842191037466, "grad_norm": 2.9258666038513184, "learning_rate": 1.8419259942701778e-05, "loss": 5.7302, "step": 5858 }, { "epoch": 0.20645365892333306, "grad_norm": 1.7138428688049316, "learning_rate": 1.8418644062883225e-05, "loss": 0.913, "step": 5859 }, { "epoch": 0.2064888959362915, "grad_norm": 4.334902286529541, "learning_rate": 1.8418028073410762e-05, "loss": 3.3596, "step": 5860 }, { "epoch": 0.2065241329492499, "grad_norm": 5.933111667633057, "learning_rate": 1.841741197429241e-05, "loss": 5.3097, "step": 5861 }, { "epoch": 0.2065593699622083, "grad_norm": 2.0651023387908936, "learning_rate": 1.8416795765536197e-05, "loss": 2.8814, "step": 5862 }, { "epoch": 0.2065946069751667, "grad_norm": 1.4139788150787354, "learning_rate": 1.8416179447150147e-05, "loss": 1.0706, "step": 5863 }, { "epoch": 0.20662984398812512, "grad_norm": 1.1134636402130127, "learning_rate": 1.8415563019142286e-05, "loss": 0.9329, "step": 5864 }, { "epoch": 0.20666508100108355, "grad_norm": 1.3401007652282715, "learning_rate": 1.8414946481520647e-05, "loss": 1.1808, "step": 5865 }, { "epoch": 0.20670031801404196, "grad_norm": 1.5666128396987915, "learning_rate": 1.8414329834293253e-05, "loss": 0.7653, "step": 5866 }, { "epoch": 0.20673555502700036, "grad_norm": 1.2571903467178345, "learning_rate": 1.8413713077468144e-05, "loss": 0.9316, "step": 5867 }, { "epoch": 0.20677079203995877, "grad_norm": 1.8831524848937988, "learning_rate": 1.841309621105335e-05, "loss": 0.9504, "step": 5868 }, { "epoch": 0.20680602905291717, "grad_norm": 1.2850781679153442, "learning_rate": 1.8412479235056908e-05, "loss": 0.9898, "step": 5869 }, { "epoch": 0.2068412660658756, "grad_norm": 0.9908248782157898, "learning_rate": 1.8411862149486853e-05, "loss": 0.9836, "step": 5870 }, { "epoch": 0.206876503078834, "grad_norm": 1.1369050741195679, "learning_rate": 1.8411244954351217e-05, "loss": 1.0534, "step": 5871 }, { "epoch": 0.20691174009179242, "grad_norm": 20.11525535583496, "learning_rate": 1.8410627649658048e-05, "loss": 6.3008, "step": 5872 }, { "epoch": 0.20694697710475082, "grad_norm": 6.951309680938721, "learning_rate": 1.841001023541538e-05, "loss": 4.1344, "step": 5873 }, { "epoch": 0.20698221411770923, "grad_norm": 1.4490243196487427, "learning_rate": 1.840939271163126e-05, "loss": 1.154, "step": 5874 }, { "epoch": 0.20701745113066766, "grad_norm": 1.1622530221939087, "learning_rate": 1.840877507831373e-05, "loss": 1.2801, "step": 5875 }, { "epoch": 0.20705268814362607, "grad_norm": 1.0723280906677246, "learning_rate": 1.840815733547083e-05, "loss": 1.0758, "step": 5876 }, { "epoch": 0.20708792515658447, "grad_norm": 1.3691750764846802, "learning_rate": 1.840753948311061e-05, "loss": 0.9746, "step": 5877 }, { "epoch": 0.20712316216954288, "grad_norm": 10.719688415527344, "learning_rate": 1.840692152124112e-05, "loss": 10.8222, "step": 5878 }, { "epoch": 0.2071583991825013, "grad_norm": 5.0103302001953125, "learning_rate": 1.8406303449870405e-05, "loss": 3.8699, "step": 5879 }, { "epoch": 0.20719363619545972, "grad_norm": 1.256680965423584, "learning_rate": 1.840568526900652e-05, "loss": 0.9829, "step": 5880 }, { "epoch": 0.20722887320841812, "grad_norm": 2.0641732215881348, "learning_rate": 1.8405066978657508e-05, "loss": 1.0265, "step": 5881 }, { "epoch": 0.20726411022137653, "grad_norm": 1.052147388458252, "learning_rate": 1.8404448578831427e-05, "loss": 1.0737, "step": 5882 }, { "epoch": 0.20729934723433494, "grad_norm": 2.3500618934631348, "learning_rate": 1.840383006953634e-05, "loss": 0.8593, "step": 5883 }, { "epoch": 0.20733458424729337, "grad_norm": 1.9452338218688965, "learning_rate": 1.840321145078029e-05, "loss": 3.454, "step": 5884 }, { "epoch": 0.20736982126025177, "grad_norm": 1.2989122867584229, "learning_rate": 1.840259272257134e-05, "loss": 0.9523, "step": 5885 }, { "epoch": 0.20740505827321018, "grad_norm": 3.1820616722106934, "learning_rate": 1.8401973884917553e-05, "loss": 6.2348, "step": 5886 }, { "epoch": 0.20744029528616859, "grad_norm": 2.518317461013794, "learning_rate": 1.8401354937826985e-05, "loss": 4.0412, "step": 5887 }, { "epoch": 0.207475532299127, "grad_norm": 2.404994249343872, "learning_rate": 1.84007358813077e-05, "loss": 3.5104, "step": 5888 }, { "epoch": 0.20751076931208542, "grad_norm": 4.257167816162109, "learning_rate": 1.8400116715367758e-05, "loss": 8.5121, "step": 5889 }, { "epoch": 0.20754600632504383, "grad_norm": 4.4688215255737305, "learning_rate": 1.8399497440015223e-05, "loss": 5.5914, "step": 5890 }, { "epoch": 0.20758124333800224, "grad_norm": 1.7990561723709106, "learning_rate": 1.839887805525817e-05, "loss": 1.0057, "step": 5891 }, { "epoch": 0.20761648035096064, "grad_norm": 1.0995855331420898, "learning_rate": 1.8398258561104655e-05, "loss": 0.8977, "step": 5892 }, { "epoch": 0.20765171736391905, "grad_norm": 1.7048544883728027, "learning_rate": 1.8397638957562756e-05, "loss": 1.1343, "step": 5893 }, { "epoch": 0.20768695437687748, "grad_norm": 1.9667726755142212, "learning_rate": 1.8397019244640536e-05, "loss": 3.5652, "step": 5894 }, { "epoch": 0.2077221913898359, "grad_norm": 4.55929708480835, "learning_rate": 1.8396399422346077e-05, "loss": 5.8308, "step": 5895 }, { "epoch": 0.2077574284027943, "grad_norm": 1.297804355621338, "learning_rate": 1.8395779490687444e-05, "loss": 0.9674, "step": 5896 }, { "epoch": 0.2077926654157527, "grad_norm": 4.99895715713501, "learning_rate": 1.839515944967271e-05, "loss": 3.0379, "step": 5897 }, { "epoch": 0.20782790242871113, "grad_norm": 1.985856294631958, "learning_rate": 1.8394539299309956e-05, "loss": 1.0666, "step": 5898 }, { "epoch": 0.20786313944166954, "grad_norm": 2.392824649810791, "learning_rate": 1.839391903960726e-05, "loss": 2.5791, "step": 5899 }, { "epoch": 0.20789837645462794, "grad_norm": 1.1541099548339844, "learning_rate": 1.8393298670572698e-05, "loss": 1.2033, "step": 5900 }, { "epoch": 0.20793361346758635, "grad_norm": 3.012186050415039, "learning_rate": 1.839267819221435e-05, "loss": 2.609, "step": 5901 }, { "epoch": 0.20796885048054475, "grad_norm": 4.516303062438965, "learning_rate": 1.8392057604540305e-05, "loss": 6.6868, "step": 5902 }, { "epoch": 0.2080040874935032, "grad_norm": 6.4953083992004395, "learning_rate": 1.839143690755864e-05, "loss": 9.9155, "step": 5903 }, { "epoch": 0.2080393245064616, "grad_norm": 7.237643718719482, "learning_rate": 1.8390816101277436e-05, "loss": 6.8893, "step": 5904 }, { "epoch": 0.20807456151942, "grad_norm": 5.9741973876953125, "learning_rate": 1.8390195185704786e-05, "loss": 3.5261, "step": 5905 }, { "epoch": 0.2081097985323784, "grad_norm": 3.036741256713867, "learning_rate": 1.8389574160848775e-05, "loss": 3.9034, "step": 5906 }, { "epoch": 0.2081450355453368, "grad_norm": 1.5451443195343018, "learning_rate": 1.838895302671749e-05, "loss": 0.8807, "step": 5907 }, { "epoch": 0.20818027255829524, "grad_norm": 5.14731502532959, "learning_rate": 1.8388331783319023e-05, "loss": 3.2271, "step": 5908 }, { "epoch": 0.20821550957125365, "grad_norm": 2.9343440532684326, "learning_rate": 1.838771043066147e-05, "loss": 2.8556, "step": 5909 }, { "epoch": 0.20825074658421205, "grad_norm": 5.413039207458496, "learning_rate": 1.8387088968752916e-05, "loss": 3.8326, "step": 5910 }, { "epoch": 0.20828598359717046, "grad_norm": 1.3555734157562256, "learning_rate": 1.838646739760146e-05, "loss": 1.0268, "step": 5911 }, { "epoch": 0.2083212206101289, "grad_norm": 1.3684496879577637, "learning_rate": 1.83858457172152e-05, "loss": 1.0617, "step": 5912 }, { "epoch": 0.2083564576230873, "grad_norm": 1.834359884262085, "learning_rate": 1.8385223927602234e-05, "loss": 1.1329, "step": 5913 }, { "epoch": 0.2083916946360457, "grad_norm": 8.915048599243164, "learning_rate": 1.8384602028770655e-05, "loss": 5.6551, "step": 5914 }, { "epoch": 0.2084269316490041, "grad_norm": 13.74530029296875, "learning_rate": 1.8383980020728564e-05, "loss": 5.4223, "step": 5915 }, { "epoch": 0.20846216866196252, "grad_norm": 1.1259390115737915, "learning_rate": 1.8383357903484068e-05, "loss": 1.0304, "step": 5916 }, { "epoch": 0.20849740567492095, "grad_norm": 1.1998203992843628, "learning_rate": 1.8382735677045268e-05, "loss": 1.1432, "step": 5917 }, { "epoch": 0.20853264268787935, "grad_norm": 1.3299894332885742, "learning_rate": 1.838211334142027e-05, "loss": 0.9683, "step": 5918 }, { "epoch": 0.20856787970083776, "grad_norm": 10.620336532592773, "learning_rate": 1.8381490896617175e-05, "loss": 5.4015, "step": 5919 }, { "epoch": 0.20860311671379617, "grad_norm": 5.863000392913818, "learning_rate": 1.8380868342644093e-05, "loss": 2.8615, "step": 5920 }, { "epoch": 0.20863835372675457, "grad_norm": 6.574859619140625, "learning_rate": 1.8380245679509133e-05, "loss": 3.4948, "step": 5921 }, { "epoch": 0.208673590739713, "grad_norm": 4.0785603523254395, "learning_rate": 1.8379622907220408e-05, "loss": 3.2989, "step": 5922 }, { "epoch": 0.2087088277526714, "grad_norm": 6.7042622566223145, "learning_rate": 1.8379000025786026e-05, "loss": 3.9684, "step": 5923 }, { "epoch": 0.20874406476562982, "grad_norm": 1.290036916732788, "learning_rate": 1.83783770352141e-05, "loss": 1.4411, "step": 5924 }, { "epoch": 0.20877930177858822, "grad_norm": 6.769360065460205, "learning_rate": 1.837775393551275e-05, "loss": 6.2748, "step": 5925 }, { "epoch": 0.20881453879154663, "grad_norm": 3.316922426223755, "learning_rate": 1.8377130726690083e-05, "loss": 3.9442, "step": 5926 }, { "epoch": 0.20884977580450506, "grad_norm": 1.7622814178466797, "learning_rate": 1.8376507408754223e-05, "loss": 1.3009, "step": 5927 }, { "epoch": 0.20888501281746347, "grad_norm": 1.66750168800354, "learning_rate": 1.8375883981713288e-05, "loss": 0.8379, "step": 5928 }, { "epoch": 0.20892024983042187, "grad_norm": 9.356221199035645, "learning_rate": 1.8375260445575394e-05, "loss": 3.7253, "step": 5929 }, { "epoch": 0.20895548684338028, "grad_norm": 0.9717733263969421, "learning_rate": 1.837463680034867e-05, "loss": 0.8141, "step": 5930 }, { "epoch": 0.2089907238563387, "grad_norm": 10.174981117248535, "learning_rate": 1.837401304604123e-05, "loss": 7.8071, "step": 5931 }, { "epoch": 0.20902596086929712, "grad_norm": 3.203948974609375, "learning_rate": 1.837338918266121e-05, "loss": 3.6593, "step": 5932 }, { "epoch": 0.20906119788225552, "grad_norm": 1.9330363273620605, "learning_rate": 1.8372765210216724e-05, "loss": 3.3042, "step": 5933 }, { "epoch": 0.20909643489521393, "grad_norm": 1.3886854648590088, "learning_rate": 1.8372141128715908e-05, "loss": 1.0781, "step": 5934 }, { "epoch": 0.20913167190817233, "grad_norm": 2.5590155124664307, "learning_rate": 1.8371516938166885e-05, "loss": 3.6646, "step": 5935 }, { "epoch": 0.20916690892113077, "grad_norm": 1.0795528888702393, "learning_rate": 1.837089263857779e-05, "loss": 1.0368, "step": 5936 }, { "epoch": 0.20920214593408917, "grad_norm": 6.552496910095215, "learning_rate": 1.8370268229956748e-05, "loss": 3.4377, "step": 5937 }, { "epoch": 0.20923738294704758, "grad_norm": 1.6036500930786133, "learning_rate": 1.8369643712311898e-05, "loss": 0.9561, "step": 5938 }, { "epoch": 0.20927261996000598, "grad_norm": 2.350050926208496, "learning_rate": 1.8369019085651374e-05, "loss": 3.7575, "step": 5939 }, { "epoch": 0.2093078569729644, "grad_norm": 3.916442394256592, "learning_rate": 1.836839434998331e-05, "loss": 3.5409, "step": 5940 }, { "epoch": 0.20934309398592282, "grad_norm": 1.4181183576583862, "learning_rate": 1.8367769505315844e-05, "loss": 1.2524, "step": 5941 }, { "epoch": 0.20937833099888123, "grad_norm": 1.5832475423812866, "learning_rate": 1.836714455165711e-05, "loss": 0.8632, "step": 5942 }, { "epoch": 0.20941356801183963, "grad_norm": 4.514467716217041, "learning_rate": 1.8366519489015258e-05, "loss": 5.4644, "step": 5943 }, { "epoch": 0.20944880502479804, "grad_norm": 1.1478261947631836, "learning_rate": 1.8365894317398424e-05, "loss": 0.8543, "step": 5944 }, { "epoch": 0.20948404203775645, "grad_norm": 1.6901088953018188, "learning_rate": 1.836526903681475e-05, "loss": 1.0716, "step": 5945 }, { "epoch": 0.20951927905071488, "grad_norm": 6.056687355041504, "learning_rate": 1.8364643647272378e-05, "loss": 3.8188, "step": 5946 }, { "epoch": 0.20955451606367329, "grad_norm": 1.4049259424209595, "learning_rate": 1.8364018148779455e-05, "loss": 0.9708, "step": 5947 }, { "epoch": 0.2095897530766317, "grad_norm": 1.4642863273620605, "learning_rate": 1.836339254134414e-05, "loss": 0.9416, "step": 5948 }, { "epoch": 0.2096249900895901, "grad_norm": 7.681103706359863, "learning_rate": 1.8362766824974563e-05, "loss": 3.9977, "step": 5949 }, { "epoch": 0.20966022710254853, "grad_norm": 3.662146806716919, "learning_rate": 1.8362140999678883e-05, "loss": 3.1588, "step": 5950 }, { "epoch": 0.20969546411550694, "grad_norm": 1.2390706539154053, "learning_rate": 1.8361515065465257e-05, "loss": 0.9883, "step": 5951 }, { "epoch": 0.20973070112846534, "grad_norm": 1.1078070402145386, "learning_rate": 1.8360889022341828e-05, "loss": 1.2952, "step": 5952 }, { "epoch": 0.20976593814142375, "grad_norm": 4.60974645614624, "learning_rate": 1.8360262870316755e-05, "loss": 3.9157, "step": 5953 }, { "epoch": 0.20980117515438215, "grad_norm": 3.7260844707489014, "learning_rate": 1.8359636609398194e-05, "loss": 5.0074, "step": 5954 }, { "epoch": 0.20983641216734059, "grad_norm": 3.704559326171875, "learning_rate": 1.8359010239594303e-05, "loss": 3.6066, "step": 5955 }, { "epoch": 0.209871649180299, "grad_norm": 1.165209412574768, "learning_rate": 1.8358383760913238e-05, "loss": 1.0626, "step": 5956 }, { "epoch": 0.2099068861932574, "grad_norm": 4.457885265350342, "learning_rate": 1.8357757173363156e-05, "loss": 4.6081, "step": 5957 }, { "epoch": 0.2099421232062158, "grad_norm": 6.197649002075195, "learning_rate": 1.8357130476952224e-05, "loss": 3.5479, "step": 5958 }, { "epoch": 0.2099773602191742, "grad_norm": 1.2633994817733765, "learning_rate": 1.8356503671688607e-05, "loss": 0.7228, "step": 5959 }, { "epoch": 0.21001259723213264, "grad_norm": 1.5298902988433838, "learning_rate": 1.8355876757580463e-05, "loss": 1.1955, "step": 5960 }, { "epoch": 0.21004783424509105, "grad_norm": 8.598087310791016, "learning_rate": 1.8355249734635955e-05, "loss": 3.6824, "step": 5961 }, { "epoch": 0.21008307125804945, "grad_norm": 1.2232449054718018, "learning_rate": 1.835462260286326e-05, "loss": 0.9062, "step": 5962 }, { "epoch": 0.21011830827100786, "grad_norm": 1.314107060432434, "learning_rate": 1.8353995362270544e-05, "loss": 1.1518, "step": 5963 }, { "epoch": 0.21015354528396626, "grad_norm": 3.553473472595215, "learning_rate": 1.8353368012865968e-05, "loss": 2.7169, "step": 5964 }, { "epoch": 0.2101887822969247, "grad_norm": 4.69014835357666, "learning_rate": 1.8352740554657713e-05, "loss": 3.5755, "step": 5965 }, { "epoch": 0.2102240193098831, "grad_norm": 1.6462844610214233, "learning_rate": 1.8352112987653948e-05, "loss": 1.2458, "step": 5966 }, { "epoch": 0.2102592563228415, "grad_norm": 1.1273677349090576, "learning_rate": 1.8351485311862848e-05, "loss": 1.072, "step": 5967 }, { "epoch": 0.21029449333579991, "grad_norm": 1.4699243307113647, "learning_rate": 1.8350857527292585e-05, "loss": 0.859, "step": 5968 }, { "epoch": 0.21032973034875835, "grad_norm": 2.758685350418091, "learning_rate": 1.8350229633951342e-05, "loss": 3.441, "step": 5969 }, { "epoch": 0.21036496736171675, "grad_norm": 11.340042114257812, "learning_rate": 1.8349601631847294e-05, "loss": 3.9661, "step": 5970 }, { "epoch": 0.21040020437467516, "grad_norm": 3.3336989879608154, "learning_rate": 1.834897352098862e-05, "loss": 3.9424, "step": 5971 }, { "epoch": 0.21043544138763356, "grad_norm": 6.263433456420898, "learning_rate": 1.8348345301383503e-05, "loss": 3.6617, "step": 5972 }, { "epoch": 0.21047067840059197, "grad_norm": 6.59311580657959, "learning_rate": 1.8347716973040124e-05, "loss": 6.8958, "step": 5973 }, { "epoch": 0.2105059154135504, "grad_norm": 1.3538219928741455, "learning_rate": 1.834708853596667e-05, "loss": 1.1131, "step": 5974 }, { "epoch": 0.2105411524265088, "grad_norm": 3.558725357055664, "learning_rate": 1.8346459990171323e-05, "loss": 3.2705, "step": 5975 }, { "epoch": 0.21057638943946722, "grad_norm": 1.5170422792434692, "learning_rate": 1.8345831335662275e-05, "loss": 0.9745, "step": 5976 }, { "epoch": 0.21061162645242562, "grad_norm": 5.73563289642334, "learning_rate": 1.8345202572447706e-05, "loss": 2.6912, "step": 5977 }, { "epoch": 0.21064686346538403, "grad_norm": 1.3385484218597412, "learning_rate": 1.8344573700535814e-05, "loss": 0.9033, "step": 5978 }, { "epoch": 0.21068210047834246, "grad_norm": 8.336281776428223, "learning_rate": 1.8343944719934782e-05, "loss": 5.954, "step": 5979 }, { "epoch": 0.21071733749130087, "grad_norm": 1.1662768125534058, "learning_rate": 1.8343315630652813e-05, "loss": 0.875, "step": 5980 }, { "epoch": 0.21075257450425927, "grad_norm": 2.3560876846313477, "learning_rate": 1.8342686432698092e-05, "loss": 3.4485, "step": 5981 }, { "epoch": 0.21078781151721768, "grad_norm": 1.2086673974990845, "learning_rate": 1.8342057126078815e-05, "loss": 0.8148, "step": 5982 }, { "epoch": 0.2108230485301761, "grad_norm": 4.372016429901123, "learning_rate": 1.834142771080319e-05, "loss": 3.473, "step": 5983 }, { "epoch": 0.21085828554313452, "grad_norm": 3.5223124027252197, "learning_rate": 1.83407981868794e-05, "loss": 5.6848, "step": 5984 }, { "epoch": 0.21089352255609292, "grad_norm": 2.2571609020233154, "learning_rate": 1.834016855431565e-05, "loss": 3.36, "step": 5985 }, { "epoch": 0.21092875956905133, "grad_norm": 0.9751814603805542, "learning_rate": 1.8339538813120144e-05, "loss": 1.2628, "step": 5986 }, { "epoch": 0.21096399658200973, "grad_norm": 2.0507829189300537, "learning_rate": 1.8338908963301084e-05, "loss": 1.289, "step": 5987 }, { "epoch": 0.21099923359496817, "grad_norm": 2.7245466709136963, "learning_rate": 1.833827900486667e-05, "loss": 4.1455, "step": 5988 }, { "epoch": 0.21103447060792657, "grad_norm": 3.813202381134033, "learning_rate": 1.833764893782511e-05, "loss": 3.3532, "step": 5989 }, { "epoch": 0.21106970762088498, "grad_norm": 3.616959571838379, "learning_rate": 1.8337018762184613e-05, "loss": 5.2262, "step": 5990 }, { "epoch": 0.21110494463384338, "grad_norm": 1.0726792812347412, "learning_rate": 1.833638847795338e-05, "loss": 0.8929, "step": 5991 }, { "epoch": 0.2111401816468018, "grad_norm": 3.4229393005371094, "learning_rate": 1.8335758085139632e-05, "loss": 2.6053, "step": 5992 }, { "epoch": 0.21117541865976022, "grad_norm": 1.3618124723434448, "learning_rate": 1.8335127583751567e-05, "loss": 1.0907, "step": 5993 }, { "epoch": 0.21121065567271863, "grad_norm": 2.570114850997925, "learning_rate": 1.833449697379741e-05, "loss": 3.1127, "step": 5994 }, { "epoch": 0.21124589268567703, "grad_norm": 12.762019157409668, "learning_rate": 1.833386625528536e-05, "loss": 4.0626, "step": 5995 }, { "epoch": 0.21128112969863544, "grad_norm": 5.518186569213867, "learning_rate": 1.8333235428223647e-05, "loss": 6.0357, "step": 5996 }, { "epoch": 0.21131636671159384, "grad_norm": 2.8759806156158447, "learning_rate": 1.833260449262048e-05, "loss": 2.98, "step": 5997 }, { "epoch": 0.21135160372455228, "grad_norm": 2.78206729888916, "learning_rate": 1.8331973448484077e-05, "loss": 4.0257, "step": 5998 }, { "epoch": 0.21138684073751068, "grad_norm": 4.915917873382568, "learning_rate": 1.833134229582266e-05, "loss": 5.7579, "step": 5999 }, { "epoch": 0.2114220777504691, "grad_norm": 1.1730468273162842, "learning_rate": 1.8330711034644448e-05, "loss": 0.8941, "step": 6000 }, { "epoch": 0.2114573147634275, "grad_norm": 1.5879547595977783, "learning_rate": 1.833007966495766e-05, "loss": 1.3577, "step": 6001 }, { "epoch": 0.21149255177638593, "grad_norm": 5.109942436218262, "learning_rate": 1.8329448186770523e-05, "loss": 3.2945, "step": 6002 }, { "epoch": 0.21152778878934433, "grad_norm": 1.2623392343521118, "learning_rate": 1.832881660009127e-05, "loss": 1.0396, "step": 6003 }, { "epoch": 0.21156302580230274, "grad_norm": 3.9746437072753906, "learning_rate": 1.832818490492811e-05, "loss": 2.7685, "step": 6004 }, { "epoch": 0.21159826281526115, "grad_norm": 5.964762210845947, "learning_rate": 1.832755310128929e-05, "loss": 6.2163, "step": 6005 }, { "epoch": 0.21163349982821955, "grad_norm": 6.386673927307129, "learning_rate": 1.8326921189183023e-05, "loss": 3.572, "step": 6006 }, { "epoch": 0.21166873684117798, "grad_norm": 1.161244511604309, "learning_rate": 1.8326289168617552e-05, "loss": 0.8824, "step": 6007 }, { "epoch": 0.2117039738541364, "grad_norm": 3.5691912174224854, "learning_rate": 1.83256570396011e-05, "loss": 4.4375, "step": 6008 }, { "epoch": 0.2117392108670948, "grad_norm": 2.5080296993255615, "learning_rate": 1.8325024802141906e-05, "loss": 3.5569, "step": 6009 }, { "epoch": 0.2117744478800532, "grad_norm": 0.9478074908256531, "learning_rate": 1.8324392456248207e-05, "loss": 0.8555, "step": 6010 }, { "epoch": 0.2118096848930116, "grad_norm": 2.008570432662964, "learning_rate": 1.8323760001928233e-05, "loss": 3.3849, "step": 6011 }, { "epoch": 0.21184492190597004, "grad_norm": 3.19002366065979, "learning_rate": 1.8323127439190223e-05, "loss": 3.3225, "step": 6012 }, { "epoch": 0.21188015891892845, "grad_norm": 0.9632800221443176, "learning_rate": 1.832249476804242e-05, "loss": 0.8158, "step": 6013 }, { "epoch": 0.21191539593188685, "grad_norm": 1.2687335014343262, "learning_rate": 1.8321861988493062e-05, "loss": 1.1171, "step": 6014 }, { "epoch": 0.21195063294484526, "grad_norm": 4.3944878578186035, "learning_rate": 1.832122910055039e-05, "loss": 2.9924, "step": 6015 }, { "epoch": 0.21198586995780366, "grad_norm": 4.3423261642456055, "learning_rate": 1.832059610422265e-05, "loss": 6.9751, "step": 6016 }, { "epoch": 0.2120211069707621, "grad_norm": 3.9291329383850098, "learning_rate": 1.8319962999518085e-05, "loss": 3.4558, "step": 6017 }, { "epoch": 0.2120563439837205, "grad_norm": 1.3580821752548218, "learning_rate": 1.8319329786444947e-05, "loss": 0.9825, "step": 6018 }, { "epoch": 0.2120915809966789, "grad_norm": 3.010218620300293, "learning_rate": 1.8318696465011473e-05, "loss": 3.7513, "step": 6019 }, { "epoch": 0.2121268180096373, "grad_norm": 2.9419760704040527, "learning_rate": 1.831806303522592e-05, "loss": 6.1211, "step": 6020 }, { "epoch": 0.21216205502259575, "grad_norm": 3.154932737350464, "learning_rate": 1.8317429497096536e-05, "loss": 3.6819, "step": 6021 }, { "epoch": 0.21219729203555415, "grad_norm": 2.5107357501983643, "learning_rate": 1.8316795850631574e-05, "loss": 2.9195, "step": 6022 }, { "epoch": 0.21223252904851256, "grad_norm": 2.269455909729004, "learning_rate": 1.8316162095839286e-05, "loss": 3.1619, "step": 6023 }, { "epoch": 0.21226776606147096, "grad_norm": 2.268298625946045, "learning_rate": 1.8315528232727928e-05, "loss": 3.5447, "step": 6024 }, { "epoch": 0.21230300307442937, "grad_norm": 2.518907308578491, "learning_rate": 1.8314894261305758e-05, "loss": 3.0216, "step": 6025 }, { "epoch": 0.2123382400873878, "grad_norm": 1.1395878791809082, "learning_rate": 1.8314260181581027e-05, "loss": 1.0165, "step": 6026 }, { "epoch": 0.2123734771003462, "grad_norm": 0.8480083346366882, "learning_rate": 1.8313625993562e-05, "loss": 0.8595, "step": 6027 }, { "epoch": 0.21240871411330461, "grad_norm": 2.518402099609375, "learning_rate": 1.8312991697256934e-05, "loss": 6.3721, "step": 6028 }, { "epoch": 0.21244395112626302, "grad_norm": 1.4340626001358032, "learning_rate": 1.8312357292674095e-05, "loss": 1.1299, "step": 6029 }, { "epoch": 0.21247918813922143, "grad_norm": 1.2617132663726807, "learning_rate": 1.831172277982174e-05, "loss": 1.235, "step": 6030 }, { "epoch": 0.21251442515217986, "grad_norm": 0.8113403916358948, "learning_rate": 1.831108815870814e-05, "loss": 1.0677, "step": 6031 }, { "epoch": 0.21254966216513826, "grad_norm": 1.249159812927246, "learning_rate": 1.8310453429341555e-05, "loss": 1.1334, "step": 6032 }, { "epoch": 0.21258489917809667, "grad_norm": 3.7445731163024902, "learning_rate": 1.8309818591730256e-05, "loss": 9.5013, "step": 6033 }, { "epoch": 0.21262013619105508, "grad_norm": 1.2860904932022095, "learning_rate": 1.8309183645882515e-05, "loss": 0.9624, "step": 6034 }, { "epoch": 0.2126553732040135, "grad_norm": 4.939641952514648, "learning_rate": 1.8308548591806596e-05, "loss": 3.5165, "step": 6035 }, { "epoch": 0.21269061021697191, "grad_norm": 2.794074773788452, "learning_rate": 1.8307913429510774e-05, "loss": 3.3152, "step": 6036 }, { "epoch": 0.21272584722993032, "grad_norm": 3.6080658435821533, "learning_rate": 1.830727815900332e-05, "loss": 3.8708, "step": 6037 }, { "epoch": 0.21276108424288873, "grad_norm": 1.2579177618026733, "learning_rate": 1.830664278029251e-05, "loss": 1.0409, "step": 6038 }, { "epoch": 0.21279632125584713, "grad_norm": 2.1572418212890625, "learning_rate": 1.8306007293386622e-05, "loss": 3.3002, "step": 6039 }, { "epoch": 0.21283155826880557, "grad_norm": 2.071143865585327, "learning_rate": 1.830537169829393e-05, "loss": 3.4571, "step": 6040 }, { "epoch": 0.21286679528176397, "grad_norm": 1.243861198425293, "learning_rate": 1.8304735995022712e-05, "loss": 0.8278, "step": 6041 }, { "epoch": 0.21290203229472238, "grad_norm": 4.687307357788086, "learning_rate": 1.8304100183581252e-05, "loss": 2.9145, "step": 6042 }, { "epoch": 0.21293726930768078, "grad_norm": 1.449914813041687, "learning_rate": 1.8303464263977825e-05, "loss": 0.784, "step": 6043 }, { "epoch": 0.2129725063206392, "grad_norm": 3.7006893157958984, "learning_rate": 1.8302828236220725e-05, "loss": 3.4603, "step": 6044 }, { "epoch": 0.21300774333359762, "grad_norm": 3.550065040588379, "learning_rate": 1.8302192100318222e-05, "loss": 3.0942, "step": 6045 }, { "epoch": 0.21304298034655603, "grad_norm": 1.2136616706848145, "learning_rate": 1.8301555856278618e-05, "loss": 0.9184, "step": 6046 }, { "epoch": 0.21307821735951443, "grad_norm": 1.2476298809051514, "learning_rate": 1.8300919504110183e-05, "loss": 0.8734, "step": 6047 }, { "epoch": 0.21311345437247284, "grad_norm": 1.551035761833191, "learning_rate": 1.830028304382122e-05, "loss": 1.0654, "step": 6048 }, { "epoch": 0.21314869138543124, "grad_norm": 3.2593588829040527, "learning_rate": 1.8299646475420013e-05, "loss": 3.2943, "step": 6049 }, { "epoch": 0.21318392839838968, "grad_norm": 1.7156167030334473, "learning_rate": 1.829900979891485e-05, "loss": 0.889, "step": 6050 }, { "epoch": 0.21321916541134808, "grad_norm": 1.7362871170043945, "learning_rate": 1.8298373014314027e-05, "loss": 1.0697, "step": 6051 }, { "epoch": 0.2132544024243065, "grad_norm": 5.411688327789307, "learning_rate": 1.829773612162584e-05, "loss": 3.7301, "step": 6052 }, { "epoch": 0.2132896394372649, "grad_norm": 2.1507046222686768, "learning_rate": 1.8297099120858583e-05, "loss": 2.8869, "step": 6053 }, { "epoch": 0.21332487645022333, "grad_norm": 1.1465390920639038, "learning_rate": 1.8296462012020556e-05, "loss": 0.9991, "step": 6054 }, { "epoch": 0.21336011346318173, "grad_norm": 7.895127296447754, "learning_rate": 1.829582479512005e-05, "loss": 6.1141, "step": 6055 }, { "epoch": 0.21339535047614014, "grad_norm": 2.63723087310791, "learning_rate": 1.829518747016537e-05, "loss": 2.8919, "step": 6056 }, { "epoch": 0.21343058748909854, "grad_norm": 5.329637050628662, "learning_rate": 1.829455003716482e-05, "loss": 3.6557, "step": 6057 }, { "epoch": 0.21346582450205695, "grad_norm": 1.5842680931091309, "learning_rate": 1.8293912496126697e-05, "loss": 0.7981, "step": 6058 }, { "epoch": 0.21350106151501538, "grad_norm": 4.048157691955566, "learning_rate": 1.829327484705931e-05, "loss": 3.9812, "step": 6059 }, { "epoch": 0.2135362985279738, "grad_norm": 1.629859209060669, "learning_rate": 1.8292637089970955e-05, "loss": 0.9437, "step": 6060 }, { "epoch": 0.2135715355409322, "grad_norm": 3.31563663482666, "learning_rate": 1.829199922486995e-05, "loss": 3.244, "step": 6061 }, { "epoch": 0.2136067725538906, "grad_norm": 1.499737024307251, "learning_rate": 1.8291361251764598e-05, "loss": 0.8632, "step": 6062 }, { "epoch": 0.213642009566849, "grad_norm": 1.3894706964492798, "learning_rate": 1.829072317066321e-05, "loss": 0.8935, "step": 6063 }, { "epoch": 0.21367724657980744, "grad_norm": 1.2724175453186035, "learning_rate": 1.8290084981574098e-05, "loss": 0.8997, "step": 6064 }, { "epoch": 0.21371248359276585, "grad_norm": 4.578188419342041, "learning_rate": 1.828944668450557e-05, "loss": 5.7973, "step": 6065 }, { "epoch": 0.21374772060572425, "grad_norm": 1.31739342212677, "learning_rate": 1.8288808279465946e-05, "loss": 0.9638, "step": 6066 }, { "epoch": 0.21378295761868266, "grad_norm": 4.1363067626953125, "learning_rate": 1.8288169766463537e-05, "loss": 4.2676, "step": 6067 }, { "epoch": 0.21381819463164106, "grad_norm": 1.9163650274276733, "learning_rate": 1.8287531145506663e-05, "loss": 0.9177, "step": 6068 }, { "epoch": 0.2138534316445995, "grad_norm": 1.1248005628585815, "learning_rate": 1.8286892416603635e-05, "loss": 1.0431, "step": 6069 }, { "epoch": 0.2138886686575579, "grad_norm": 1.320936679840088, "learning_rate": 1.8286253579762783e-05, "loss": 1.2215, "step": 6070 }, { "epoch": 0.2139239056705163, "grad_norm": 5.7412896156311035, "learning_rate": 1.828561463499242e-05, "loss": 6.7806, "step": 6071 }, { "epoch": 0.2139591426834747, "grad_norm": 1.5130348205566406, "learning_rate": 1.8284975582300874e-05, "loss": 1.1278, "step": 6072 }, { "epoch": 0.21399437969643315, "grad_norm": 1.1137460470199585, "learning_rate": 1.828433642169646e-05, "loss": 1.1618, "step": 6073 }, { "epoch": 0.21402961670939155, "grad_norm": 4.125887393951416, "learning_rate": 1.8283697153187512e-05, "loss": 1.4056, "step": 6074 }, { "epoch": 0.21406485372234996, "grad_norm": 1.5215505361557007, "learning_rate": 1.8283057776782352e-05, "loss": 1.0111, "step": 6075 }, { "epoch": 0.21410009073530836, "grad_norm": 3.4829821586608887, "learning_rate": 1.8282418292489312e-05, "loss": 3.814, "step": 6076 }, { "epoch": 0.21413532774826677, "grad_norm": 1.2946586608886719, "learning_rate": 1.828177870031672e-05, "loss": 1.1423, "step": 6077 }, { "epoch": 0.2141705647612252, "grad_norm": 34.217254638671875, "learning_rate": 1.8281139000272904e-05, "loss": 1.1962, "step": 6078 }, { "epoch": 0.2142058017741836, "grad_norm": 2.9774436950683594, "learning_rate": 1.8280499192366193e-05, "loss": 3.7816, "step": 6079 }, { "epoch": 0.214241038787142, "grad_norm": 11.97471809387207, "learning_rate": 1.827985927660493e-05, "loss": 1.1178, "step": 6080 }, { "epoch": 0.21427627580010042, "grad_norm": 1.0029851198196411, "learning_rate": 1.8279219252997443e-05, "loss": 1.1162, "step": 6081 }, { "epoch": 0.21431151281305882, "grad_norm": 5.630227565765381, "learning_rate": 1.8278579121552075e-05, "loss": 3.6898, "step": 6082 }, { "epoch": 0.21434674982601726, "grad_norm": 4.127596378326416, "learning_rate": 1.8277938882277155e-05, "loss": 6.067, "step": 6083 }, { "epoch": 0.21438198683897566, "grad_norm": 3.367907762527466, "learning_rate": 1.8277298535181027e-05, "loss": 3.6496, "step": 6084 }, { "epoch": 0.21441722385193407, "grad_norm": 1.3284715414047241, "learning_rate": 1.827665808027203e-05, "loss": 1.0036, "step": 6085 }, { "epoch": 0.21445246086489247, "grad_norm": 1.7487242221832275, "learning_rate": 1.8276017517558512e-05, "loss": 0.7705, "step": 6086 }, { "epoch": 0.2144876978778509, "grad_norm": 1.6254934072494507, "learning_rate": 1.827537684704881e-05, "loss": 0.9219, "step": 6087 }, { "epoch": 0.2145229348908093, "grad_norm": 2.646771192550659, "learning_rate": 1.8274736068751267e-05, "loss": 3.3089, "step": 6088 }, { "epoch": 0.21455817190376772, "grad_norm": 1.819110631942749, "learning_rate": 1.8274095182674237e-05, "loss": 1.1781, "step": 6089 }, { "epoch": 0.21459340891672612, "grad_norm": 6.323482036590576, "learning_rate": 1.8273454188826065e-05, "loss": 6.1138, "step": 6090 }, { "epoch": 0.21462864592968453, "grad_norm": 3.0189244747161865, "learning_rate": 1.827281308721509e-05, "loss": 3.8415, "step": 6091 }, { "epoch": 0.21466388294264296, "grad_norm": 3.3320255279541016, "learning_rate": 1.8272171877849675e-05, "loss": 5.2345, "step": 6092 }, { "epoch": 0.21469911995560137, "grad_norm": 1.2646982669830322, "learning_rate": 1.8271530560738167e-05, "loss": 0.8923, "step": 6093 }, { "epoch": 0.21473435696855978, "grad_norm": 5.404839515686035, "learning_rate": 1.827088913588892e-05, "loss": 2.9573, "step": 6094 }, { "epoch": 0.21476959398151818, "grad_norm": 2.617110252380371, "learning_rate": 1.827024760331029e-05, "loss": 3.7569, "step": 6095 }, { "epoch": 0.2148048309944766, "grad_norm": 4.774488925933838, "learning_rate": 1.8269605963010632e-05, "loss": 5.888, "step": 6096 }, { "epoch": 0.21484006800743502, "grad_norm": 1.3053035736083984, "learning_rate": 1.8268964214998303e-05, "loss": 0.9522, "step": 6097 }, { "epoch": 0.21487530502039343, "grad_norm": 0.9902704358100891, "learning_rate": 1.8268322359281662e-05, "loss": 1.0149, "step": 6098 }, { "epoch": 0.21491054203335183, "grad_norm": 5.567933559417725, "learning_rate": 1.8267680395869067e-05, "loss": 2.6524, "step": 6099 }, { "epoch": 0.21494577904631024, "grad_norm": 2.999387741088867, "learning_rate": 1.8267038324768882e-05, "loss": 3.6751, "step": 6100 }, { "epoch": 0.21498101605926864, "grad_norm": 8.882722854614258, "learning_rate": 1.826639614598947e-05, "loss": 5.978, "step": 6101 }, { "epoch": 0.21501625307222708, "grad_norm": 3.9071998596191406, "learning_rate": 1.8265753859539198e-05, "loss": 3.5519, "step": 6102 }, { "epoch": 0.21505149008518548, "grad_norm": 1.241716742515564, "learning_rate": 1.8265111465426423e-05, "loss": 1.2022, "step": 6103 }, { "epoch": 0.2150867270981439, "grad_norm": 4.20201301574707, "learning_rate": 1.826446896365952e-05, "loss": 3.0706, "step": 6104 }, { "epoch": 0.2151219641111023, "grad_norm": 1.8000518083572388, "learning_rate": 1.8263826354246862e-05, "loss": 1.3852, "step": 6105 }, { "epoch": 0.21515720112406073, "grad_norm": 7.246269226074219, "learning_rate": 1.826318363719681e-05, "loss": 3.4964, "step": 6106 }, { "epoch": 0.21519243813701913, "grad_norm": 5.384750843048096, "learning_rate": 1.8262540812517736e-05, "loss": 3.8925, "step": 6107 }, { "epoch": 0.21522767514997754, "grad_norm": 2.2671024799346924, "learning_rate": 1.8261897880218015e-05, "loss": 0.7691, "step": 6108 }, { "epoch": 0.21526291216293594, "grad_norm": 1.8340173959732056, "learning_rate": 1.8261254840306023e-05, "loss": 0.932, "step": 6109 }, { "epoch": 0.21529814917589435, "grad_norm": 1.0629771947860718, "learning_rate": 1.8260611692790138e-05, "loss": 0.9711, "step": 6110 }, { "epoch": 0.21533338618885278, "grad_norm": 4.178859233856201, "learning_rate": 1.825996843767873e-05, "loss": 3.3537, "step": 6111 }, { "epoch": 0.2153686232018112, "grad_norm": 4.598895072937012, "learning_rate": 1.8259325074980178e-05, "loss": 3.3817, "step": 6112 }, { "epoch": 0.2154038602147696, "grad_norm": 1.2315192222595215, "learning_rate": 1.825868160470287e-05, "loss": 0.886, "step": 6113 }, { "epoch": 0.215439097227728, "grad_norm": 4.203127861022949, "learning_rate": 1.825803802685518e-05, "loss": 4.0675, "step": 6114 }, { "epoch": 0.2154743342406864, "grad_norm": 2.6483795642852783, "learning_rate": 1.8257394341445492e-05, "loss": 3.2568, "step": 6115 }, { "epoch": 0.21550957125364484, "grad_norm": 2.485480546951294, "learning_rate": 1.8256750548482193e-05, "loss": 1.0312, "step": 6116 }, { "epoch": 0.21554480826660324, "grad_norm": 5.157074451446533, "learning_rate": 1.8256106647973665e-05, "loss": 5.9966, "step": 6117 }, { "epoch": 0.21558004527956165, "grad_norm": 3.56392765045166, "learning_rate": 1.8255462639928297e-05, "loss": 3.4971, "step": 6118 }, { "epoch": 0.21561528229252006, "grad_norm": 4.71872091293335, "learning_rate": 1.8254818524354475e-05, "loss": 6.1641, "step": 6119 }, { "epoch": 0.21565051930547846, "grad_norm": 2.7933542728424072, "learning_rate": 1.825417430126059e-05, "loss": 3.2475, "step": 6120 }, { "epoch": 0.2156857563184369, "grad_norm": 0.9919718503952026, "learning_rate": 1.8253529970655038e-05, "loss": 1.0248, "step": 6121 }, { "epoch": 0.2157209933313953, "grad_norm": 3.5099985599517822, "learning_rate": 1.8252885532546204e-05, "loss": 4.0366, "step": 6122 }, { "epoch": 0.2157562303443537, "grad_norm": 2.495413064956665, "learning_rate": 1.8252240986942486e-05, "loss": 2.897, "step": 6123 }, { "epoch": 0.2157914673573121, "grad_norm": 4.389000415802002, "learning_rate": 1.8251596333852272e-05, "loss": 1.1669, "step": 6124 }, { "epoch": 0.21582670437027054, "grad_norm": 3.5870256423950195, "learning_rate": 1.8250951573283972e-05, "loss": 1.0315, "step": 6125 }, { "epoch": 0.21586194138322895, "grad_norm": 3.7415616512298584, "learning_rate": 1.825030670524597e-05, "loss": 3.1454, "step": 6126 }, { "epoch": 0.21589717839618736, "grad_norm": 5.441995620727539, "learning_rate": 1.824966172974668e-05, "loss": 3.1722, "step": 6127 }, { "epoch": 0.21593241540914576, "grad_norm": 2.907256841659546, "learning_rate": 1.8249016646794492e-05, "loss": 2.6043, "step": 6128 }, { "epoch": 0.21596765242210417, "grad_norm": 1.373005986213684, "learning_rate": 1.824837145639781e-05, "loss": 1.0562, "step": 6129 }, { "epoch": 0.2160028894350626, "grad_norm": 1.3919472694396973, "learning_rate": 1.8247726158565038e-05, "loss": 0.8991, "step": 6130 }, { "epoch": 0.216038126448021, "grad_norm": 2.6094744205474854, "learning_rate": 1.8247080753304583e-05, "loss": 4.2594, "step": 6131 }, { "epoch": 0.2160733634609794, "grad_norm": 4.370606422424316, "learning_rate": 1.824643524062485e-05, "loss": 3.8354, "step": 6132 }, { "epoch": 0.21610860047393782, "grad_norm": 1.3486051559448242, "learning_rate": 1.8245789620534247e-05, "loss": 1.0987, "step": 6133 }, { "epoch": 0.21614383748689622, "grad_norm": 3.14418888092041, "learning_rate": 1.8245143893041192e-05, "loss": 4.2525, "step": 6134 }, { "epoch": 0.21617907449985466, "grad_norm": 1.0672937631607056, "learning_rate": 1.824449805815408e-05, "loss": 0.9405, "step": 6135 }, { "epoch": 0.21621431151281306, "grad_norm": 1.3330155611038208, "learning_rate": 1.824385211588133e-05, "loss": 0.9856, "step": 6136 }, { "epoch": 0.21624954852577147, "grad_norm": 2.4956798553466797, "learning_rate": 1.8243206066231358e-05, "loss": 3.251, "step": 6137 }, { "epoch": 0.21628478553872987, "grad_norm": 3.4441752433776855, "learning_rate": 1.8242559909212577e-05, "loss": 3.7841, "step": 6138 }, { "epoch": 0.2163200225516883, "grad_norm": 1.7709778547286987, "learning_rate": 1.8241913644833404e-05, "loss": 1.0485, "step": 6139 }, { "epoch": 0.2163552595646467, "grad_norm": 1.0297178030014038, "learning_rate": 1.8241267273102258e-05, "loss": 0.9969, "step": 6140 }, { "epoch": 0.21639049657760512, "grad_norm": 1.559669017791748, "learning_rate": 1.824062079402755e-05, "loss": 1.0637, "step": 6141 }, { "epoch": 0.21642573359056352, "grad_norm": 1.279518723487854, "learning_rate": 1.823997420761771e-05, "loss": 0.9423, "step": 6142 }, { "epoch": 0.21646097060352193, "grad_norm": 1.3647840023040771, "learning_rate": 1.8239327513881154e-05, "loss": 1.234, "step": 6143 }, { "epoch": 0.21649620761648036, "grad_norm": 3.7578468322753906, "learning_rate": 1.823868071282631e-05, "loss": 3.4796, "step": 6144 }, { "epoch": 0.21653144462943877, "grad_norm": 6.345157146453857, "learning_rate": 1.82380338044616e-05, "loss": 6.3599, "step": 6145 }, { "epoch": 0.21656668164239717, "grad_norm": 1.7727465629577637, "learning_rate": 1.8237386788795452e-05, "loss": 2.7651, "step": 6146 }, { "epoch": 0.21660191865535558, "grad_norm": 3.1771504878997803, "learning_rate": 1.823673966583629e-05, "loss": 5.3796, "step": 6147 }, { "epoch": 0.21663715566831399, "grad_norm": 3.6412155628204346, "learning_rate": 1.8236092435592546e-05, "loss": 6.8183, "step": 6148 }, { "epoch": 0.21667239268127242, "grad_norm": 4.76554012298584, "learning_rate": 1.8235445098072647e-05, "loss": 3.9942, "step": 6149 }, { "epoch": 0.21670762969423082, "grad_norm": 7.417636394500732, "learning_rate": 1.8234797653285025e-05, "loss": 6.4391, "step": 6150 }, { "epoch": 0.21674286670718923, "grad_norm": 1.3665049076080322, "learning_rate": 1.823415010123812e-05, "loss": 1.0205, "step": 6151 }, { "epoch": 0.21677810372014764, "grad_norm": 7.1942925453186035, "learning_rate": 1.8233502441940355e-05, "loss": 5.7664, "step": 6152 }, { "epoch": 0.21681334073310604, "grad_norm": 1.4877878427505493, "learning_rate": 1.8232854675400173e-05, "loss": 1.2685, "step": 6153 }, { "epoch": 0.21684857774606447, "grad_norm": 1.6197247505187988, "learning_rate": 1.8232206801626013e-05, "loss": 0.9145, "step": 6154 }, { "epoch": 0.21688381475902288, "grad_norm": 1.5563671588897705, "learning_rate": 1.823155882062631e-05, "loss": 1.0027, "step": 6155 }, { "epoch": 0.21691905177198129, "grad_norm": 1.7840038537979126, "learning_rate": 1.8230910732409507e-05, "loss": 0.9474, "step": 6156 }, { "epoch": 0.2169542887849397, "grad_norm": 8.815959930419922, "learning_rate": 1.8230262536984038e-05, "loss": 2.9307, "step": 6157 }, { "epoch": 0.21698952579789813, "grad_norm": 1.46760892868042, "learning_rate": 1.8229614234358354e-05, "loss": 1.2819, "step": 6158 }, { "epoch": 0.21702476281085653, "grad_norm": 1.4566936492919922, "learning_rate": 1.8228965824540897e-05, "loss": 1.157, "step": 6159 }, { "epoch": 0.21705999982381494, "grad_norm": 5.0308709144592285, "learning_rate": 1.8228317307540112e-05, "loss": 4.2791, "step": 6160 }, { "epoch": 0.21709523683677334, "grad_norm": 1.1839003562927246, "learning_rate": 1.8227668683364444e-05, "loss": 1.1603, "step": 6161 }, { "epoch": 0.21713047384973175, "grad_norm": 1.2934536933898926, "learning_rate": 1.8227019952022344e-05, "loss": 0.8622, "step": 6162 }, { "epoch": 0.21716571086269018, "grad_norm": 5.013031959533691, "learning_rate": 1.8226371113522262e-05, "loss": 1.9761, "step": 6163 }, { "epoch": 0.2172009478756486, "grad_norm": 4.8384294509887695, "learning_rate": 1.8225722167872648e-05, "loss": 3.5364, "step": 6164 }, { "epoch": 0.217236184888607, "grad_norm": 4.463468551635742, "learning_rate": 1.8225073115081954e-05, "loss": 4.7603, "step": 6165 }, { "epoch": 0.2172714219015654, "grad_norm": 1.3081531524658203, "learning_rate": 1.8224423955158635e-05, "loss": 0.9006, "step": 6166 }, { "epoch": 0.2173066589145238, "grad_norm": 14.975038528442383, "learning_rate": 1.8223774688111147e-05, "loss": 8.6661, "step": 6167 }, { "epoch": 0.21734189592748224, "grad_norm": 7.883781909942627, "learning_rate": 1.8223125313947947e-05, "loss": 2.8793, "step": 6168 }, { "epoch": 0.21737713294044064, "grad_norm": 3.3738958835601807, "learning_rate": 1.8222475832677493e-05, "loss": 4.1113, "step": 6169 }, { "epoch": 0.21741236995339905, "grad_norm": 3.214113712310791, "learning_rate": 1.8221826244308243e-05, "loss": 3.3185, "step": 6170 }, { "epoch": 0.21744760696635745, "grad_norm": 4.885164737701416, "learning_rate": 1.8221176548848658e-05, "loss": 3.4734, "step": 6171 }, { "epoch": 0.21748284397931586, "grad_norm": 4.149317264556885, "learning_rate": 1.82205267463072e-05, "loss": 3.1292, "step": 6172 }, { "epoch": 0.2175180809922743, "grad_norm": 1.2365543842315674, "learning_rate": 1.8219876836692338e-05, "loss": 1.131, "step": 6173 }, { "epoch": 0.2175533180052327, "grad_norm": 1.27383291721344, "learning_rate": 1.8219226820012533e-05, "loss": 1.0966, "step": 6174 }, { "epoch": 0.2175885550181911, "grad_norm": 1.168022632598877, "learning_rate": 1.8218576696276248e-05, "loss": 0.9523, "step": 6175 }, { "epoch": 0.2176237920311495, "grad_norm": 0.9663659334182739, "learning_rate": 1.8217926465491955e-05, "loss": 0.8814, "step": 6176 }, { "epoch": 0.21765902904410794, "grad_norm": 1.163620114326477, "learning_rate": 1.8217276127668123e-05, "loss": 0.924, "step": 6177 }, { "epoch": 0.21769426605706635, "grad_norm": 6.548211097717285, "learning_rate": 1.8216625682813224e-05, "loss": 9.3124, "step": 6178 }, { "epoch": 0.21772950307002475, "grad_norm": 3.257990598678589, "learning_rate": 1.821597513093573e-05, "loss": 3.5279, "step": 6179 }, { "epoch": 0.21776474008298316, "grad_norm": 1.4209718704223633, "learning_rate": 1.8215324472044113e-05, "loss": 1.0442, "step": 6180 }, { "epoch": 0.21779997709594157, "grad_norm": 1.2808204889297485, "learning_rate": 1.8214673706146845e-05, "loss": 0.7784, "step": 6181 }, { "epoch": 0.2178352141089, "grad_norm": 1.682841420173645, "learning_rate": 1.8214022833252408e-05, "loss": 0.969, "step": 6182 }, { "epoch": 0.2178704511218584, "grad_norm": 1.6469597816467285, "learning_rate": 1.821337185336928e-05, "loss": 0.9978, "step": 6183 }, { "epoch": 0.2179056881348168, "grad_norm": 7.829648017883301, "learning_rate": 1.8212720766505936e-05, "loss": 5.6596, "step": 6184 }, { "epoch": 0.21794092514777522, "grad_norm": 1.2167402505874634, "learning_rate": 1.8212069572670854e-05, "loss": 1.0533, "step": 6185 }, { "epoch": 0.21797616216073362, "grad_norm": 3.557687520980835, "learning_rate": 1.8211418271872523e-05, "loss": 3.6384, "step": 6186 }, { "epoch": 0.21801139917369206, "grad_norm": 1.8032784461975098, "learning_rate": 1.8210766864119423e-05, "loss": 3.5328, "step": 6187 }, { "epoch": 0.21804663618665046, "grad_norm": 1.524290919303894, "learning_rate": 1.8210115349420035e-05, "loss": 0.9864, "step": 6188 }, { "epoch": 0.21808187319960887, "grad_norm": 1.4363669157028198, "learning_rate": 1.8209463727782856e-05, "loss": 0.9479, "step": 6189 }, { "epoch": 0.21811711021256727, "grad_norm": 6.146406650543213, "learning_rate": 1.8208811999216363e-05, "loss": 3.4462, "step": 6190 }, { "epoch": 0.2181523472255257, "grad_norm": 4.796647071838379, "learning_rate": 1.8208160163729046e-05, "loss": 3.0099, "step": 6191 }, { "epoch": 0.2181875842384841, "grad_norm": 3.029738664627075, "learning_rate": 1.8207508221329397e-05, "loss": 3.0734, "step": 6192 }, { "epoch": 0.21822282125144252, "grad_norm": 1.0822272300720215, "learning_rate": 1.820685617202591e-05, "loss": 0.7855, "step": 6193 }, { "epoch": 0.21825805826440092, "grad_norm": 6.28423547744751, "learning_rate": 1.8206204015827074e-05, "loss": 6.5548, "step": 6194 }, { "epoch": 0.21829329527735933, "grad_norm": 1.0462640523910522, "learning_rate": 1.820555175274139e-05, "loss": 0.7758, "step": 6195 }, { "epoch": 0.21832853229031776, "grad_norm": 1.6078763008117676, "learning_rate": 1.8204899382777346e-05, "loss": 1.1096, "step": 6196 }, { "epoch": 0.21836376930327617, "grad_norm": 1.7664328813552856, "learning_rate": 1.8204246905943442e-05, "loss": 0.9554, "step": 6197 }, { "epoch": 0.21839900631623457, "grad_norm": 3.342784881591797, "learning_rate": 1.8203594322248176e-05, "loss": 3.471, "step": 6198 }, { "epoch": 0.21843424332919298, "grad_norm": 4.37334680557251, "learning_rate": 1.8202941631700052e-05, "loss": 6.8285, "step": 6199 }, { "epoch": 0.21846948034215138, "grad_norm": 5.221663475036621, "learning_rate": 1.8202288834307565e-05, "loss": 3.6726, "step": 6200 }, { "epoch": 0.21850471735510982, "grad_norm": 1.8026328086853027, "learning_rate": 1.8201635930079223e-05, "loss": 1.1233, "step": 6201 }, { "epoch": 0.21853995436806822, "grad_norm": 2.3651182651519775, "learning_rate": 1.820098291902353e-05, "loss": 0.901, "step": 6202 }, { "epoch": 0.21857519138102663, "grad_norm": 11.823358535766602, "learning_rate": 1.8200329801148985e-05, "loss": 10.6403, "step": 6203 }, { "epoch": 0.21861042839398503, "grad_norm": 1.7785215377807617, "learning_rate": 1.8199676576464104e-05, "loss": 1.0037, "step": 6204 }, { "epoch": 0.21864566540694344, "grad_norm": 1.6151087284088135, "learning_rate": 1.819902324497739e-05, "loss": 1.0447, "step": 6205 }, { "epoch": 0.21868090241990187, "grad_norm": 3.662559986114502, "learning_rate": 1.8198369806697354e-05, "loss": 3.7969, "step": 6206 }, { "epoch": 0.21871613943286028, "grad_norm": 1.1349502801895142, "learning_rate": 1.8197716261632506e-05, "loss": 1.3601, "step": 6207 }, { "epoch": 0.21875137644581868, "grad_norm": 1.821969985961914, "learning_rate": 1.8197062609791363e-05, "loss": 1.0994, "step": 6208 }, { "epoch": 0.2187866134587771, "grad_norm": 2.655385971069336, "learning_rate": 1.819640885118243e-05, "loss": 3.5794, "step": 6209 }, { "epoch": 0.21882185047173552, "grad_norm": 1.045040488243103, "learning_rate": 1.819575498581423e-05, "loss": 1.0, "step": 6210 }, { "epoch": 0.21885708748469393, "grad_norm": 1.222458839416504, "learning_rate": 1.8195101013695277e-05, "loss": 0.8827, "step": 6211 }, { "epoch": 0.21889232449765234, "grad_norm": 1.3864532709121704, "learning_rate": 1.8194446934834095e-05, "loss": 0.9708, "step": 6212 }, { "epoch": 0.21892756151061074, "grad_norm": 1.131751537322998, "learning_rate": 1.819379274923919e-05, "loss": 1.1383, "step": 6213 }, { "epoch": 0.21896279852356915, "grad_norm": 4.030974388122559, "learning_rate": 1.8193138456919094e-05, "loss": 3.8179, "step": 6214 }, { "epoch": 0.21899803553652758, "grad_norm": 4.574079513549805, "learning_rate": 1.8192484057882324e-05, "loss": 3.7478, "step": 6215 }, { "epoch": 0.21903327254948599, "grad_norm": 3.0654611587524414, "learning_rate": 1.8191829552137407e-05, "loss": 4.4116, "step": 6216 }, { "epoch": 0.2190685095624444, "grad_norm": 4.364604949951172, "learning_rate": 1.8191174939692868e-05, "loss": 3.3062, "step": 6217 }, { "epoch": 0.2191037465754028, "grad_norm": 1.2845194339752197, "learning_rate": 1.819052022055723e-05, "loss": 0.9896, "step": 6218 }, { "epoch": 0.2191389835883612, "grad_norm": 1.22438383102417, "learning_rate": 1.8189865394739025e-05, "loss": 1.1876, "step": 6219 }, { "epoch": 0.21917422060131964, "grad_norm": 3.927044153213501, "learning_rate": 1.8189210462246777e-05, "loss": 6.0361, "step": 6220 }, { "epoch": 0.21920945761427804, "grad_norm": 4.382330894470215, "learning_rate": 1.818855542308902e-05, "loss": 2.8134, "step": 6221 }, { "epoch": 0.21924469462723645, "grad_norm": 2.3928678035736084, "learning_rate": 1.818790027727429e-05, "loss": 3.1887, "step": 6222 }, { "epoch": 0.21927993164019485, "grad_norm": 3.261282444000244, "learning_rate": 1.818724502481111e-05, "loss": 3.647, "step": 6223 }, { "epoch": 0.21931516865315326, "grad_norm": 1.499502182006836, "learning_rate": 1.8186589665708024e-05, "loss": 1.0435, "step": 6224 }, { "epoch": 0.2193504056661117, "grad_norm": 1.9746983051300049, "learning_rate": 1.8185934199973563e-05, "loss": 0.6912, "step": 6225 }, { "epoch": 0.2193856426790701, "grad_norm": 3.638798713684082, "learning_rate": 1.818527862761627e-05, "loss": 2.6246, "step": 6226 }, { "epoch": 0.2194208796920285, "grad_norm": 2.825760841369629, "learning_rate": 1.8184622948644675e-05, "loss": 3.5605, "step": 6227 }, { "epoch": 0.2194561167049869, "grad_norm": 1.8512376546859741, "learning_rate": 1.8183967163067328e-05, "loss": 0.872, "step": 6228 }, { "epoch": 0.21949135371794534, "grad_norm": 2.3952817916870117, "learning_rate": 1.8183311270892764e-05, "loss": 0.8426, "step": 6229 }, { "epoch": 0.21952659073090375, "grad_norm": 0.8399893641471863, "learning_rate": 1.818265527212953e-05, "loss": 0.843, "step": 6230 }, { "epoch": 0.21956182774386215, "grad_norm": 3.438652992248535, "learning_rate": 1.8181999166786165e-05, "loss": 3.3522, "step": 6231 }, { "epoch": 0.21959706475682056, "grad_norm": 1.567299723625183, "learning_rate": 1.8181342954871222e-05, "loss": 0.8586, "step": 6232 }, { "epoch": 0.21963230176977896, "grad_norm": 1.4087408781051636, "learning_rate": 1.8180686636393247e-05, "loss": 1.1289, "step": 6233 }, { "epoch": 0.2196675387827374, "grad_norm": 5.226989269256592, "learning_rate": 1.8180030211360783e-05, "loss": 3.0582, "step": 6234 }, { "epoch": 0.2197027757956958, "grad_norm": 3.830549955368042, "learning_rate": 1.8179373679782382e-05, "loss": 3.5685, "step": 6235 }, { "epoch": 0.2197380128086542, "grad_norm": 4.290233135223389, "learning_rate": 1.81787170416666e-05, "loss": 3.9457, "step": 6236 }, { "epoch": 0.21977324982161262, "grad_norm": 3.985887289047241, "learning_rate": 1.8178060297021988e-05, "loss": 3.6297, "step": 6237 }, { "epoch": 0.21980848683457102, "grad_norm": 1.3044365644454956, "learning_rate": 1.8177403445857094e-05, "loss": 0.9782, "step": 6238 }, { "epoch": 0.21984372384752945, "grad_norm": 7.7844109535217285, "learning_rate": 1.817674648818048e-05, "loss": 5.8632, "step": 6239 }, { "epoch": 0.21987896086048786, "grad_norm": 1.1987518072128296, "learning_rate": 1.8176089424000706e-05, "loss": 1.2709, "step": 6240 }, { "epoch": 0.21991419787344627, "grad_norm": 1.653790831565857, "learning_rate": 1.817543225332632e-05, "loss": 0.9782, "step": 6241 }, { "epoch": 0.21994943488640467, "grad_norm": 3.8698697090148926, "learning_rate": 1.817477497616589e-05, "loss": 3.8035, "step": 6242 }, { "epoch": 0.2199846718993631, "grad_norm": 4.728116035461426, "learning_rate": 1.8174117592527976e-05, "loss": 3.2023, "step": 6243 }, { "epoch": 0.2200199089123215, "grad_norm": 4.619442939758301, "learning_rate": 1.8173460102421136e-05, "loss": 6.7648, "step": 6244 }, { "epoch": 0.22005514592527992, "grad_norm": 7.2048821449279785, "learning_rate": 1.817280250585394e-05, "loss": 7.6121, "step": 6245 }, { "epoch": 0.22009038293823832, "grad_norm": 5.90346097946167, "learning_rate": 1.8172144802834946e-05, "loss": 3.5857, "step": 6246 }, { "epoch": 0.22012561995119673, "grad_norm": 1.7563557624816895, "learning_rate": 1.817148699337273e-05, "loss": 0.9985, "step": 6247 }, { "epoch": 0.22016085696415516, "grad_norm": 1.436448574066162, "learning_rate": 1.8170829077475852e-05, "loss": 0.9501, "step": 6248 }, { "epoch": 0.22019609397711357, "grad_norm": 1.163367509841919, "learning_rate": 1.8170171055152885e-05, "loss": 1.0117, "step": 6249 }, { "epoch": 0.22023133099007197, "grad_norm": 3.1564369201660156, "learning_rate": 1.81695129264124e-05, "loss": 3.397, "step": 6250 }, { "epoch": 0.22026656800303038, "grad_norm": 2.8043487071990967, "learning_rate": 1.816885469126297e-05, "loss": 3.2252, "step": 6251 }, { "epoch": 0.22030180501598878, "grad_norm": 1.0609667301177979, "learning_rate": 1.8168196349713164e-05, "loss": 1.001, "step": 6252 }, { "epoch": 0.22033704202894722, "grad_norm": 4.5882439613342285, "learning_rate": 1.816753790177156e-05, "loss": 3.6314, "step": 6253 }, { "epoch": 0.22037227904190562, "grad_norm": 3.677778959274292, "learning_rate": 1.8166879347446738e-05, "loss": 3.0073, "step": 6254 }, { "epoch": 0.22040751605486403, "grad_norm": 1.4895422458648682, "learning_rate": 1.816622068674727e-05, "loss": 0.791, "step": 6255 }, { "epoch": 0.22044275306782243, "grad_norm": 4.059277534484863, "learning_rate": 1.8165561919681738e-05, "loss": 4.0643, "step": 6256 }, { "epoch": 0.22047799008078084, "grad_norm": 1.4608125686645508, "learning_rate": 1.816490304625872e-05, "loss": 0.8932, "step": 6257 }, { "epoch": 0.22051322709373927, "grad_norm": 1.4916071891784668, "learning_rate": 1.8164244066486803e-05, "loss": 1.2012, "step": 6258 }, { "epoch": 0.22054846410669768, "grad_norm": 1.5401206016540527, "learning_rate": 1.8163584980374564e-05, "loss": 1.0282, "step": 6259 }, { "epoch": 0.22058370111965608, "grad_norm": 1.987892985343933, "learning_rate": 1.8162925787930593e-05, "loss": 0.9535, "step": 6260 }, { "epoch": 0.2206189381326145, "grad_norm": 1.3538936376571655, "learning_rate": 1.8162266489163473e-05, "loss": 0.8162, "step": 6261 }, { "epoch": 0.22065417514557292, "grad_norm": 4.548225402832031, "learning_rate": 1.816160708408179e-05, "loss": 6.0925, "step": 6262 }, { "epoch": 0.22068941215853133, "grad_norm": 1.3279589414596558, "learning_rate": 1.8160947572694137e-05, "loss": 1.3708, "step": 6263 }, { "epoch": 0.22072464917148973, "grad_norm": 1.4399385452270508, "learning_rate": 1.8160287955009103e-05, "loss": 0.9182, "step": 6264 }, { "epoch": 0.22075988618444814, "grad_norm": 1.579825520515442, "learning_rate": 1.815962823103528e-05, "loss": 1.0225, "step": 6265 }, { "epoch": 0.22079512319740655, "grad_norm": 1.7289780378341675, "learning_rate": 1.8158968400781258e-05, "loss": 0.8825, "step": 6266 }, { "epoch": 0.22083036021036498, "grad_norm": 1.1995248794555664, "learning_rate": 1.8158308464255635e-05, "loss": 0.9267, "step": 6267 }, { "epoch": 0.22086559722332338, "grad_norm": 5.163799285888672, "learning_rate": 1.8157648421467e-05, "loss": 6.0171, "step": 6268 }, { "epoch": 0.2209008342362818, "grad_norm": 1.229925513267517, "learning_rate": 1.8156988272423963e-05, "loss": 1.443, "step": 6269 }, { "epoch": 0.2209360712492402, "grad_norm": 3.2288589477539062, "learning_rate": 1.8156328017135108e-05, "loss": 3.9081, "step": 6270 }, { "epoch": 0.2209713082621986, "grad_norm": 1.2862799167633057, "learning_rate": 1.8155667655609047e-05, "loss": 1.1363, "step": 6271 }, { "epoch": 0.22100654527515703, "grad_norm": 12.660760879516602, "learning_rate": 1.815500718785437e-05, "loss": 3.0093, "step": 6272 }, { "epoch": 0.22104178228811544, "grad_norm": 1.693751573562622, "learning_rate": 1.815434661387969e-05, "loss": 1.0543, "step": 6273 }, { "epoch": 0.22107701930107385, "grad_norm": 5.559362888336182, "learning_rate": 1.8153685933693604e-05, "loss": 3.7419, "step": 6274 }, { "epoch": 0.22111225631403225, "grad_norm": 3.926947832107544, "learning_rate": 1.815302514730472e-05, "loss": 2.7227, "step": 6275 }, { "epoch": 0.22114749332699066, "grad_norm": 1.1491284370422363, "learning_rate": 1.815236425472165e-05, "loss": 0.9649, "step": 6276 }, { "epoch": 0.2211827303399491, "grad_norm": 11.521368026733398, "learning_rate": 1.815170325595299e-05, "loss": 6.6794, "step": 6277 }, { "epoch": 0.2212179673529075, "grad_norm": 3.324352264404297, "learning_rate": 1.8151042151007365e-05, "loss": 3.6487, "step": 6278 }, { "epoch": 0.2212532043658659, "grad_norm": 1.7877960205078125, "learning_rate": 1.815038093989337e-05, "loss": 1.0395, "step": 6279 }, { "epoch": 0.2212884413788243, "grad_norm": 2.3784255981445312, "learning_rate": 1.8149719622619627e-05, "loss": 3.5038, "step": 6280 }, { "epoch": 0.22132367839178274, "grad_norm": 2.0307838916778564, "learning_rate": 1.814905819919475e-05, "loss": 1.0464, "step": 6281 }, { "epoch": 0.22135891540474115, "grad_norm": 1.8808774948120117, "learning_rate": 1.814839666962735e-05, "loss": 0.9243, "step": 6282 }, { "epoch": 0.22139415241769955, "grad_norm": 1.4186078310012817, "learning_rate": 1.814773503392605e-05, "loss": 0.812, "step": 6283 }, { "epoch": 0.22142938943065796, "grad_norm": 1.130757212638855, "learning_rate": 1.8147073292099455e-05, "loss": 0.9253, "step": 6284 }, { "epoch": 0.22146462644361636, "grad_norm": 2.0281269550323486, "learning_rate": 1.81464114441562e-05, "loss": 0.8655, "step": 6285 }, { "epoch": 0.2214998634565748, "grad_norm": 1.3207881450653076, "learning_rate": 1.8145749490104894e-05, "loss": 0.8164, "step": 6286 }, { "epoch": 0.2215351004695332, "grad_norm": 5.818493366241455, "learning_rate": 1.8145087429954166e-05, "loss": 5.978, "step": 6287 }, { "epoch": 0.2215703374824916, "grad_norm": 1.554302453994751, "learning_rate": 1.8144425263712633e-05, "loss": 1.1413, "step": 6288 }, { "epoch": 0.22160557449545001, "grad_norm": 1.2912095785140991, "learning_rate": 1.8143762991388927e-05, "loss": 1.057, "step": 6289 }, { "epoch": 0.22164081150840842, "grad_norm": 6.24142599105835, "learning_rate": 1.814310061299167e-05, "loss": 3.833, "step": 6290 }, { "epoch": 0.22167604852136685, "grad_norm": 3.592409133911133, "learning_rate": 1.814243812852949e-05, "loss": 3.1149, "step": 6291 }, { "epoch": 0.22171128553432526, "grad_norm": 3.296726942062378, "learning_rate": 1.8141775538011016e-05, "loss": 3.1859, "step": 6292 }, { "epoch": 0.22174652254728366, "grad_norm": 3.3451650142669678, "learning_rate": 1.8141112841444875e-05, "loss": 4.0147, "step": 6293 }, { "epoch": 0.22178175956024207, "grad_norm": 5.675346851348877, "learning_rate": 1.8140450038839707e-05, "loss": 4.0287, "step": 6294 }, { "epoch": 0.2218169965732005, "grad_norm": 3.363410234451294, "learning_rate": 1.813978713020414e-05, "loss": 3.7114, "step": 6295 }, { "epoch": 0.2218522335861589, "grad_norm": 1.3796175718307495, "learning_rate": 1.813912411554681e-05, "loss": 1.2497, "step": 6296 }, { "epoch": 0.22188747059911731, "grad_norm": 4.940084934234619, "learning_rate": 1.8138460994876348e-05, "loss": 5.9491, "step": 6297 }, { "epoch": 0.22192270761207572, "grad_norm": 1.3252023458480835, "learning_rate": 1.8137797768201396e-05, "loss": 0.7807, "step": 6298 }, { "epoch": 0.22195794462503413, "grad_norm": 1.6872519254684448, "learning_rate": 1.813713443553059e-05, "loss": 0.9625, "step": 6299 }, { "epoch": 0.22199318163799256, "grad_norm": 1.399013876914978, "learning_rate": 1.8136470996872574e-05, "loss": 0.8878, "step": 6300 }, { "epoch": 0.22202841865095096, "grad_norm": 1.549848198890686, "learning_rate": 1.813580745223599e-05, "loss": 1.0026, "step": 6301 }, { "epoch": 0.22206365566390937, "grad_norm": 3.060114860534668, "learning_rate": 1.813514380162947e-05, "loss": 3.2838, "step": 6302 }, { "epoch": 0.22209889267686778, "grad_norm": 4.72307014465332, "learning_rate": 1.813448004506167e-05, "loss": 3.7961, "step": 6303 }, { "epoch": 0.22213412968982618, "grad_norm": 4.006984710693359, "learning_rate": 1.8133816182541234e-05, "loss": 2.8293, "step": 6304 }, { "epoch": 0.22216936670278462, "grad_norm": 5.171941757202148, "learning_rate": 1.81331522140768e-05, "loss": 5.578, "step": 6305 }, { "epoch": 0.22220460371574302, "grad_norm": 4.371060848236084, "learning_rate": 1.8132488139677024e-05, "loss": 3.5929, "step": 6306 }, { "epoch": 0.22223984072870143, "grad_norm": 1.0434449911117554, "learning_rate": 1.8131823959350557e-05, "loss": 1.1403, "step": 6307 }, { "epoch": 0.22227507774165983, "grad_norm": 7.369613170623779, "learning_rate": 1.8131159673106046e-05, "loss": 3.7445, "step": 6308 }, { "epoch": 0.22231031475461824, "grad_norm": 5.477825164794922, "learning_rate": 1.8130495280952143e-05, "loss": 3.9944, "step": 6309 }, { "epoch": 0.22234555176757667, "grad_norm": 1.5423628091812134, "learning_rate": 1.8129830782897506e-05, "loss": 0.8839, "step": 6310 }, { "epoch": 0.22238078878053508, "grad_norm": 0.9670150876045227, "learning_rate": 1.8129166178950783e-05, "loss": 0.8834, "step": 6311 }, { "epoch": 0.22241602579349348, "grad_norm": 0.9100595116615295, "learning_rate": 1.812850146912064e-05, "loss": 1.0213, "step": 6312 }, { "epoch": 0.2224512628064519, "grad_norm": 7.251976490020752, "learning_rate": 1.8127836653415728e-05, "loss": 4.1146, "step": 6313 }, { "epoch": 0.22248649981941032, "grad_norm": 3.7030575275421143, "learning_rate": 1.8127171731844705e-05, "loss": 4.042, "step": 6314 }, { "epoch": 0.22252173683236873, "grad_norm": 4.6456708908081055, "learning_rate": 1.812650670441624e-05, "loss": 6.0075, "step": 6315 }, { "epoch": 0.22255697384532713, "grad_norm": 1.1686701774597168, "learning_rate": 1.8125841571138986e-05, "loss": 1.2002, "step": 6316 }, { "epoch": 0.22259221085828554, "grad_norm": 6.93436336517334, "learning_rate": 1.8125176332021616e-05, "loss": 6.3746, "step": 6317 }, { "epoch": 0.22262744787124394, "grad_norm": 2.90354061126709, "learning_rate": 1.8124510987072782e-05, "loss": 2.94, "step": 6318 }, { "epoch": 0.22266268488420238, "grad_norm": 7.979600429534912, "learning_rate": 1.8123845536301164e-05, "loss": 6.1795, "step": 6319 }, { "epoch": 0.22269792189716078, "grad_norm": 1.0596165657043457, "learning_rate": 1.812317997971542e-05, "loss": 1.0835, "step": 6320 }, { "epoch": 0.2227331589101192, "grad_norm": 3.2420711517333984, "learning_rate": 1.8122514317324224e-05, "loss": 4.0486, "step": 6321 }, { "epoch": 0.2227683959230776, "grad_norm": 1.1777364015579224, "learning_rate": 1.812184854913624e-05, "loss": 1.063, "step": 6322 }, { "epoch": 0.222803632936036, "grad_norm": 1.8026126623153687, "learning_rate": 1.8121182675160146e-05, "loss": 1.0157, "step": 6323 }, { "epoch": 0.22283886994899443, "grad_norm": 1.1255525350570679, "learning_rate": 1.8120516695404615e-05, "loss": 1.0242, "step": 6324 }, { "epoch": 0.22287410696195284, "grad_norm": 1.5591646432876587, "learning_rate": 1.811985060987832e-05, "loss": 0.9426, "step": 6325 }, { "epoch": 0.22290934397491124, "grad_norm": 2.6583058834075928, "learning_rate": 1.8119184418589934e-05, "loss": 3.8306, "step": 6326 }, { "epoch": 0.22294458098786965, "grad_norm": 1.1840531826019287, "learning_rate": 1.811851812154814e-05, "loss": 0.9588, "step": 6327 }, { "epoch": 0.22297981800082806, "grad_norm": 4.403388023376465, "learning_rate": 1.811785171876161e-05, "loss": 5.7295, "step": 6328 }, { "epoch": 0.2230150550137865, "grad_norm": 4.554251194000244, "learning_rate": 1.8117185210239028e-05, "loss": 4.0397, "step": 6329 }, { "epoch": 0.2230502920267449, "grad_norm": 2.961430788040161, "learning_rate": 1.8116518595989076e-05, "loss": 3.896, "step": 6330 }, { "epoch": 0.2230855290397033, "grad_norm": 1.0187692642211914, "learning_rate": 1.8115851876020434e-05, "loss": 0.8101, "step": 6331 }, { "epoch": 0.2231207660526617, "grad_norm": 1.4776277542114258, "learning_rate": 1.811518505034179e-05, "loss": 1.0793, "step": 6332 }, { "epoch": 0.22315600306562014, "grad_norm": 1.456076979637146, "learning_rate": 1.8114518118961824e-05, "loss": 0.903, "step": 6333 }, { "epoch": 0.22319124007857855, "grad_norm": 1.3484240770339966, "learning_rate": 1.8113851081889226e-05, "loss": 1.116, "step": 6334 }, { "epoch": 0.22322647709153695, "grad_norm": 1.7373266220092773, "learning_rate": 1.8113183939132684e-05, "loss": 1.0668, "step": 6335 }, { "epoch": 0.22326171410449536, "grad_norm": 4.474918365478516, "learning_rate": 1.811251669070089e-05, "loss": 3.0381, "step": 6336 }, { "epoch": 0.22329695111745376, "grad_norm": 4.4346394538879395, "learning_rate": 1.811184933660253e-05, "loss": 6.2026, "step": 6337 }, { "epoch": 0.2233321881304122, "grad_norm": 3.440526247024536, "learning_rate": 1.8111181876846303e-05, "loss": 3.4994, "step": 6338 }, { "epoch": 0.2233674251433706, "grad_norm": 7.168257713317871, "learning_rate": 1.8110514311440893e-05, "loss": 6.4228, "step": 6339 }, { "epoch": 0.223402662156329, "grad_norm": 1.0843561887741089, "learning_rate": 1.8109846640395007e-05, "loss": 1.0227, "step": 6340 }, { "epoch": 0.2234378991692874, "grad_norm": 1.8525521755218506, "learning_rate": 1.8109178863717334e-05, "loss": 0.9265, "step": 6341 }, { "epoch": 0.22347313618224582, "grad_norm": 1.9471049308776855, "learning_rate": 1.8108510981416573e-05, "loss": 1.2266, "step": 6342 }, { "epoch": 0.22350837319520425, "grad_norm": 1.865190029144287, "learning_rate": 1.810784299350142e-05, "loss": 0.9286, "step": 6343 }, { "epoch": 0.22354361020816266, "grad_norm": 0.9946943521499634, "learning_rate": 1.8107174899980584e-05, "loss": 1.085, "step": 6344 }, { "epoch": 0.22357884722112106, "grad_norm": 1.7758228778839111, "learning_rate": 1.8106506700862762e-05, "loss": 1.1722, "step": 6345 }, { "epoch": 0.22361408423407947, "grad_norm": 5.549161434173584, "learning_rate": 1.8105838396156655e-05, "loss": 3.6515, "step": 6346 }, { "epoch": 0.2236493212470379, "grad_norm": 1.8289411067962646, "learning_rate": 1.810516998587097e-05, "loss": 0.9222, "step": 6347 }, { "epoch": 0.2236845582599963, "grad_norm": 3.0121028423309326, "learning_rate": 1.8104501470014417e-05, "loss": 2.911, "step": 6348 }, { "epoch": 0.2237197952729547, "grad_norm": 1.848095178604126, "learning_rate": 1.8103832848595698e-05, "loss": 0.9037, "step": 6349 }, { "epoch": 0.22375503228591312, "grad_norm": 3.762413740158081, "learning_rate": 1.810316412162352e-05, "loss": 6.3503, "step": 6350 }, { "epoch": 0.22379026929887152, "grad_norm": 1.166260838508606, "learning_rate": 1.8102495289106607e-05, "loss": 1.1024, "step": 6351 }, { "epoch": 0.22382550631182996, "grad_norm": 1.856249213218689, "learning_rate": 1.810182635105365e-05, "loss": 0.7293, "step": 6352 }, { "epoch": 0.22386074332478836, "grad_norm": 1.2912789583206177, "learning_rate": 1.810115730747338e-05, "loss": 1.062, "step": 6353 }, { "epoch": 0.22389598033774677, "grad_norm": 1.9931938648223877, "learning_rate": 1.81004881583745e-05, "loss": 0.821, "step": 6354 }, { "epoch": 0.22393121735070518, "grad_norm": 1.8310829401016235, "learning_rate": 1.8099818903765732e-05, "loss": 0.7067, "step": 6355 }, { "epoch": 0.22396645436366358, "grad_norm": 3.6403446197509766, "learning_rate": 1.8099149543655793e-05, "loss": 3.3715, "step": 6356 }, { "epoch": 0.22400169137662201, "grad_norm": 4.571660041809082, "learning_rate": 1.8098480078053393e-05, "loss": 4.0621, "step": 6357 }, { "epoch": 0.22403692838958042, "grad_norm": 5.252298355102539, "learning_rate": 1.8097810506967263e-05, "loss": 4.2961, "step": 6358 }, { "epoch": 0.22407216540253883, "grad_norm": 1.7311453819274902, "learning_rate": 1.809714083040612e-05, "loss": 1.1521, "step": 6359 }, { "epoch": 0.22410740241549723, "grad_norm": 4.104881286621094, "learning_rate": 1.8096471048378686e-05, "loss": 3.6948, "step": 6360 }, { "epoch": 0.22414263942845564, "grad_norm": 3.322993040084839, "learning_rate": 1.8095801160893685e-05, "loss": 3.4043, "step": 6361 }, { "epoch": 0.22417787644141407, "grad_norm": 1.535740613937378, "learning_rate": 1.809513116795984e-05, "loss": 0.8866, "step": 6362 }, { "epoch": 0.22421311345437248, "grad_norm": 3.785259246826172, "learning_rate": 1.8094461069585885e-05, "loss": 3.292, "step": 6363 }, { "epoch": 0.22424835046733088, "grad_norm": 1.241895079612732, "learning_rate": 1.809379086578054e-05, "loss": 1.2228, "step": 6364 }, { "epoch": 0.2242835874802893, "grad_norm": 3.3260788917541504, "learning_rate": 1.809312055655254e-05, "loss": 0.8823, "step": 6365 }, { "epoch": 0.22431882449324772, "grad_norm": 3.4177420139312744, "learning_rate": 1.8092450141910612e-05, "loss": 3.3382, "step": 6366 }, { "epoch": 0.22435406150620613, "grad_norm": 2.7243072986602783, "learning_rate": 1.8091779621863493e-05, "loss": 3.4496, "step": 6367 }, { "epoch": 0.22438929851916453, "grad_norm": 1.0443888902664185, "learning_rate": 1.8091108996419912e-05, "loss": 1.0735, "step": 6368 }, { "epoch": 0.22442453553212294, "grad_norm": 4.11816930770874, "learning_rate": 1.8090438265588606e-05, "loss": 3.7808, "step": 6369 }, { "epoch": 0.22445977254508134, "grad_norm": 1.5608710050582886, "learning_rate": 1.8089767429378308e-05, "loss": 0.8974, "step": 6370 }, { "epoch": 0.22449500955803978, "grad_norm": 4.113483905792236, "learning_rate": 1.8089096487797762e-05, "loss": 3.1578, "step": 6371 }, { "epoch": 0.22453024657099818, "grad_norm": 1.4386281967163086, "learning_rate": 1.8088425440855705e-05, "loss": 1.0001, "step": 6372 }, { "epoch": 0.2245654835839566, "grad_norm": 2.067012310028076, "learning_rate": 1.8087754288560876e-05, "loss": 0.983, "step": 6373 }, { "epoch": 0.224600720596915, "grad_norm": 1.8094489574432373, "learning_rate": 1.8087083030922014e-05, "loss": 0.7312, "step": 6374 }, { "epoch": 0.2246359576098734, "grad_norm": 1.0358651876449585, "learning_rate": 1.8086411667947868e-05, "loss": 0.9519, "step": 6375 }, { "epoch": 0.22467119462283183, "grad_norm": 3.7327067852020264, "learning_rate": 1.8085740199647178e-05, "loss": 3.3548, "step": 6376 }, { "epoch": 0.22470643163579024, "grad_norm": 8.412622451782227, "learning_rate": 1.8085068626028692e-05, "loss": 6.2594, "step": 6377 }, { "epoch": 0.22474166864874864, "grad_norm": 0.9684485197067261, "learning_rate": 1.8084396947101157e-05, "loss": 0.7337, "step": 6378 }, { "epoch": 0.22477690566170705, "grad_norm": 1.5009583234786987, "learning_rate": 1.8083725162873325e-05, "loss": 0.9683, "step": 6379 }, { "epoch": 0.22481214267466545, "grad_norm": 1.2116914987564087, "learning_rate": 1.8083053273353942e-05, "loss": 0.9152, "step": 6380 }, { "epoch": 0.2248473796876239, "grad_norm": 3.9490976333618164, "learning_rate": 1.8082381278551758e-05, "loss": 3.3684, "step": 6381 }, { "epoch": 0.2248826167005823, "grad_norm": 1.259232521057129, "learning_rate": 1.8081709178475532e-05, "loss": 0.9099, "step": 6382 }, { "epoch": 0.2249178537135407, "grad_norm": 1.5297483205795288, "learning_rate": 1.8081036973134012e-05, "loss": 1.0042, "step": 6383 }, { "epoch": 0.2249530907264991, "grad_norm": 7.313451290130615, "learning_rate": 1.8080364662535956e-05, "loss": 4.2731, "step": 6384 }, { "epoch": 0.22498832773945754, "grad_norm": 1.6945998668670654, "learning_rate": 1.807969224669012e-05, "loss": 0.8217, "step": 6385 }, { "epoch": 0.22502356475241594, "grad_norm": 5.084982395172119, "learning_rate": 1.8079019725605268e-05, "loss": 3.348, "step": 6386 }, { "epoch": 0.22505880176537435, "grad_norm": 10.113840103149414, "learning_rate": 1.807834709929015e-05, "loss": 5.3038, "step": 6387 }, { "epoch": 0.22509403877833276, "grad_norm": 3.0553932189941406, "learning_rate": 1.8077674367753534e-05, "loss": 3.7638, "step": 6388 }, { "epoch": 0.22512927579129116, "grad_norm": 1.8755987882614136, "learning_rate": 1.8077001531004178e-05, "loss": 0.8293, "step": 6389 }, { "epoch": 0.2251645128042496, "grad_norm": 5.237509250640869, "learning_rate": 1.8076328589050854e-05, "loss": 3.0557, "step": 6390 }, { "epoch": 0.225199749817208, "grad_norm": 1.3906856775283813, "learning_rate": 1.807565554190232e-05, "loss": 1.0943, "step": 6391 }, { "epoch": 0.2252349868301664, "grad_norm": 5.169200897216797, "learning_rate": 1.807498238956734e-05, "loss": 4.3536, "step": 6392 }, { "epoch": 0.2252702238431248, "grad_norm": 4.133227348327637, "learning_rate": 1.8074309132054688e-05, "loss": 3.7462, "step": 6393 }, { "epoch": 0.22530546085608322, "grad_norm": 1.9760172367095947, "learning_rate": 1.807363576937313e-05, "loss": 0.9193, "step": 6394 }, { "epoch": 0.22534069786904165, "grad_norm": 1.3409932851791382, "learning_rate": 1.8072962301531438e-05, "loss": 1.1098, "step": 6395 }, { "epoch": 0.22537593488200006, "grad_norm": 7.732637405395508, "learning_rate": 1.8072288728538385e-05, "loss": 3.7968, "step": 6396 }, { "epoch": 0.22541117189495846, "grad_norm": 1.3603297472000122, "learning_rate": 1.807161505040274e-05, "loss": 0.7087, "step": 6397 }, { "epoch": 0.22544640890791687, "grad_norm": 1.2501163482666016, "learning_rate": 1.8070941267133286e-05, "loss": 1.1481, "step": 6398 }, { "epoch": 0.2254816459208753, "grad_norm": 17.56235694885254, "learning_rate": 1.8070267378738787e-05, "loss": 5.8178, "step": 6399 }, { "epoch": 0.2255168829338337, "grad_norm": 6.871896266937256, "learning_rate": 1.806959338522803e-05, "loss": 6.3048, "step": 6400 }, { "epoch": 0.2255521199467921, "grad_norm": 1.6750842332839966, "learning_rate": 1.8068919286609797e-05, "loss": 0.8784, "step": 6401 }, { "epoch": 0.22558735695975052, "grad_norm": 3.997645139694214, "learning_rate": 1.8068245082892857e-05, "loss": 3.9339, "step": 6402 }, { "epoch": 0.22562259397270892, "grad_norm": 3.9707202911376953, "learning_rate": 1.8067570774086e-05, "loss": 3.4659, "step": 6403 }, { "epoch": 0.22565783098566736, "grad_norm": 2.3662617206573486, "learning_rate": 1.8066896360198e-05, "loss": 2.9552, "step": 6404 }, { "epoch": 0.22569306799862576, "grad_norm": 5.8332719802856445, "learning_rate": 1.8066221841237654e-05, "loss": 3.6594, "step": 6405 }, { "epoch": 0.22572830501158417, "grad_norm": 1.3287326097488403, "learning_rate": 1.8065547217213736e-05, "loss": 0.7818, "step": 6406 }, { "epoch": 0.22576354202454257, "grad_norm": 3.2080917358398438, "learning_rate": 1.8064872488135042e-05, "loss": 3.1068, "step": 6407 }, { "epoch": 0.22579877903750098, "grad_norm": 3.192955493927002, "learning_rate": 1.8064197654010355e-05, "loss": 6.522, "step": 6408 }, { "epoch": 0.2258340160504594, "grad_norm": 3.523834466934204, "learning_rate": 1.8063522714848466e-05, "loss": 3.8739, "step": 6409 }, { "epoch": 0.22586925306341782, "grad_norm": 1.0395914316177368, "learning_rate": 1.8062847670658168e-05, "loss": 0.9296, "step": 6410 }, { "epoch": 0.22590449007637622, "grad_norm": 1.3015483617782593, "learning_rate": 1.8062172521448253e-05, "loss": 1.0152, "step": 6411 }, { "epoch": 0.22593972708933463, "grad_norm": 5.164167404174805, "learning_rate": 1.806149726722751e-05, "loss": 5.8388, "step": 6412 }, { "epoch": 0.22597496410229304, "grad_norm": 1.3804450035095215, "learning_rate": 1.8060821908004743e-05, "loss": 0.8485, "step": 6413 }, { "epoch": 0.22601020111525147, "grad_norm": 3.5146737098693848, "learning_rate": 1.8060146443788736e-05, "loss": 3.595, "step": 6414 }, { "epoch": 0.22604543812820987, "grad_norm": 1.6251407861709595, "learning_rate": 1.80594708745883e-05, "loss": 0.9047, "step": 6415 }, { "epoch": 0.22608067514116828, "grad_norm": 1.8141517639160156, "learning_rate": 1.805879520041223e-05, "loss": 0.8599, "step": 6416 }, { "epoch": 0.22611591215412669, "grad_norm": 1.3962998390197754, "learning_rate": 1.8058119421269327e-05, "loss": 0.7913, "step": 6417 }, { "epoch": 0.22615114916708512, "grad_norm": 3.4666976928710938, "learning_rate": 1.805744353716839e-05, "loss": 3.6614, "step": 6418 }, { "epoch": 0.22618638618004352, "grad_norm": 1.3318166732788086, "learning_rate": 1.8056767548118225e-05, "loss": 0.8394, "step": 6419 }, { "epoch": 0.22622162319300193, "grad_norm": 1.2102612257003784, "learning_rate": 1.8056091454127635e-05, "loss": 0.8708, "step": 6420 }, { "epoch": 0.22625686020596034, "grad_norm": 4.3830342292785645, "learning_rate": 1.805541525520543e-05, "loss": 6.0633, "step": 6421 }, { "epoch": 0.22629209721891874, "grad_norm": 3.126770257949829, "learning_rate": 1.8054738951360412e-05, "loss": 2.8565, "step": 6422 }, { "epoch": 0.22632733423187718, "grad_norm": 5.428778648376465, "learning_rate": 1.8054062542601398e-05, "loss": 6.8007, "step": 6423 }, { "epoch": 0.22636257124483558, "grad_norm": 3.7234857082366943, "learning_rate": 1.805338602893719e-05, "loss": 3.5162, "step": 6424 }, { "epoch": 0.226397808257794, "grad_norm": 4.289783000946045, "learning_rate": 1.8052709410376603e-05, "loss": 3.6967, "step": 6425 }, { "epoch": 0.2264330452707524, "grad_norm": 3.7777438163757324, "learning_rate": 1.805203268692845e-05, "loss": 4.0546, "step": 6426 }, { "epoch": 0.2264682822837108, "grad_norm": 2.9434926509857178, "learning_rate": 1.8051355858601545e-05, "loss": 3.3892, "step": 6427 }, { "epoch": 0.22650351929666923, "grad_norm": 1.1357554197311401, "learning_rate": 1.8050678925404708e-05, "loss": 0.9135, "step": 6428 }, { "epoch": 0.22653875630962764, "grad_norm": 2.6056325435638428, "learning_rate": 1.8050001887346753e-05, "loss": 3.0717, "step": 6429 }, { "epoch": 0.22657399332258604, "grad_norm": 3.8349971771240234, "learning_rate": 1.8049324744436494e-05, "loss": 6.6161, "step": 6430 }, { "epoch": 0.22660923033554445, "grad_norm": 1.2749645709991455, "learning_rate": 1.8048647496682757e-05, "loss": 0.7342, "step": 6431 }, { "epoch": 0.22664446734850285, "grad_norm": 1.2740761041641235, "learning_rate": 1.804797014409436e-05, "loss": 1.0254, "step": 6432 }, { "epoch": 0.2266797043614613, "grad_norm": 4.068939685821533, "learning_rate": 1.8047292686680128e-05, "loss": 6.0459, "step": 6433 }, { "epoch": 0.2267149413744197, "grad_norm": 4.876964569091797, "learning_rate": 1.8046615124448885e-05, "loss": 5.3838, "step": 6434 }, { "epoch": 0.2267501783873781, "grad_norm": 1.3497871160507202, "learning_rate": 1.804593745740945e-05, "loss": 0.7909, "step": 6435 }, { "epoch": 0.2267854154003365, "grad_norm": 1.207394003868103, "learning_rate": 1.804525968557066e-05, "loss": 0.9665, "step": 6436 }, { "epoch": 0.22682065241329494, "grad_norm": 3.426405429840088, "learning_rate": 1.804458180894133e-05, "loss": 5.2117, "step": 6437 }, { "epoch": 0.22685588942625334, "grad_norm": 6.105180263519287, "learning_rate": 1.8043903827530305e-05, "loss": 8.5426, "step": 6438 }, { "epoch": 0.22689112643921175, "grad_norm": 1.4134796857833862, "learning_rate": 1.8043225741346406e-05, "loss": 0.8789, "step": 6439 }, { "epoch": 0.22692636345217015, "grad_norm": 5.786111831665039, "learning_rate": 1.8042547550398466e-05, "loss": 6.3454, "step": 6440 }, { "epoch": 0.22696160046512856, "grad_norm": 4.047916412353516, "learning_rate": 1.8041869254695322e-05, "loss": 3.6351, "step": 6441 }, { "epoch": 0.226996837478087, "grad_norm": 3.1101529598236084, "learning_rate": 1.8041190854245804e-05, "loss": 3.0844, "step": 6442 }, { "epoch": 0.2270320744910454, "grad_norm": 1.494735598564148, "learning_rate": 1.804051234905875e-05, "loss": 0.8633, "step": 6443 }, { "epoch": 0.2270673115040038, "grad_norm": 4.0807647705078125, "learning_rate": 1.8039833739143003e-05, "loss": 3.6227, "step": 6444 }, { "epoch": 0.2271025485169622, "grad_norm": 6.311440944671631, "learning_rate": 1.8039155024507396e-05, "loss": 8.3719, "step": 6445 }, { "epoch": 0.22713778552992062, "grad_norm": 3.1550371646881104, "learning_rate": 1.803847620516077e-05, "loss": 3.815, "step": 6446 }, { "epoch": 0.22717302254287905, "grad_norm": 4.240302085876465, "learning_rate": 1.8037797281111967e-05, "loss": 3.1874, "step": 6447 }, { "epoch": 0.22720825955583746, "grad_norm": 5.467045307159424, "learning_rate": 1.803711825236983e-05, "loss": 3.735, "step": 6448 }, { "epoch": 0.22724349656879586, "grad_norm": 1.6070022583007812, "learning_rate": 1.8036439118943205e-05, "loss": 1.3575, "step": 6449 }, { "epoch": 0.22727873358175427, "grad_norm": 1.9511178731918335, "learning_rate": 1.8035759880840938e-05, "loss": 1.2975, "step": 6450 }, { "epoch": 0.2273139705947127, "grad_norm": 2.976323127746582, "learning_rate": 1.803508053807187e-05, "loss": 3.7792, "step": 6451 }, { "epoch": 0.2273492076076711, "grad_norm": 1.2558046579360962, "learning_rate": 1.803440109064486e-05, "loss": 1.0723, "step": 6452 }, { "epoch": 0.2273844446206295, "grad_norm": 1.2100293636322021, "learning_rate": 1.803372153856875e-05, "loss": 0.8333, "step": 6453 }, { "epoch": 0.22741968163358792, "grad_norm": 1.045578122138977, "learning_rate": 1.8033041881852392e-05, "loss": 1.2207, "step": 6454 }, { "epoch": 0.22745491864654632, "grad_norm": 6.74677848815918, "learning_rate": 1.8032362120504646e-05, "loss": 5.5938, "step": 6455 }, { "epoch": 0.22749015565950476, "grad_norm": 3.213184118270874, "learning_rate": 1.8031682254534356e-05, "loss": 4.6549, "step": 6456 }, { "epoch": 0.22752539267246316, "grad_norm": 1.2080121040344238, "learning_rate": 1.803100228395038e-05, "loss": 1.029, "step": 6457 }, { "epoch": 0.22756062968542157, "grad_norm": 4.4846510887146, "learning_rate": 1.803032220876158e-05, "loss": 5.783, "step": 6458 }, { "epoch": 0.22759586669837997, "grad_norm": 1.1194769144058228, "learning_rate": 1.8029642028976807e-05, "loss": 1.2917, "step": 6459 }, { "epoch": 0.22763110371133838, "grad_norm": 4.941207408905029, "learning_rate": 1.8028961744604925e-05, "loss": 3.4585, "step": 6460 }, { "epoch": 0.2276663407242968, "grad_norm": 1.800539255142212, "learning_rate": 1.8028281355654792e-05, "loss": 1.0688, "step": 6461 }, { "epoch": 0.22770157773725522, "grad_norm": 6.809935569763184, "learning_rate": 1.8027600862135274e-05, "loss": 6.492, "step": 6462 }, { "epoch": 0.22773681475021362, "grad_norm": 1.1301872730255127, "learning_rate": 1.8026920264055233e-05, "loss": 0.9309, "step": 6463 }, { "epoch": 0.22777205176317203, "grad_norm": 2.456780433654785, "learning_rate": 1.802623956142353e-05, "loss": 3.5202, "step": 6464 }, { "epoch": 0.22780728877613043, "grad_norm": 4.181118488311768, "learning_rate": 1.8025558754249037e-05, "loss": 3.3133, "step": 6465 }, { "epoch": 0.22784252578908887, "grad_norm": 3.1100001335144043, "learning_rate": 1.802487784254062e-05, "loss": 3.7717, "step": 6466 }, { "epoch": 0.22787776280204727, "grad_norm": 4.026744842529297, "learning_rate": 1.8024196826307144e-05, "loss": 7.1692, "step": 6467 }, { "epoch": 0.22791299981500568, "grad_norm": 2.0348856449127197, "learning_rate": 1.8023515705557485e-05, "loss": 1.1807, "step": 6468 }, { "epoch": 0.22794823682796408, "grad_norm": 1.3619725704193115, "learning_rate": 1.802283448030051e-05, "loss": 0.878, "step": 6469 }, { "epoch": 0.22798347384092252, "grad_norm": 3.685441255569458, "learning_rate": 1.8022153150545093e-05, "loss": 3.5907, "step": 6470 }, { "epoch": 0.22801871085388092, "grad_norm": 1.4681824445724487, "learning_rate": 1.802147171630011e-05, "loss": 0.9302, "step": 6471 }, { "epoch": 0.22805394786683933, "grad_norm": 5.32115364074707, "learning_rate": 1.8020790177574438e-05, "loss": 8.7521, "step": 6472 }, { "epoch": 0.22808918487979774, "grad_norm": 3.1367313861846924, "learning_rate": 1.802010853437695e-05, "loss": 3.8016, "step": 6473 }, { "epoch": 0.22812442189275614, "grad_norm": 2.3210103511810303, "learning_rate": 1.8019426786716527e-05, "loss": 3.3542, "step": 6474 }, { "epoch": 0.22815965890571457, "grad_norm": 2.0722169876098633, "learning_rate": 1.801874493460205e-05, "loss": 0.9716, "step": 6475 }, { "epoch": 0.22819489591867298, "grad_norm": 2.727196455001831, "learning_rate": 1.8018062978042397e-05, "loss": 3.5879, "step": 6476 }, { "epoch": 0.22823013293163139, "grad_norm": 3.699082851409912, "learning_rate": 1.8017380917046454e-05, "loss": 3.6073, "step": 6477 }, { "epoch": 0.2282653699445898, "grad_norm": 2.3753886222839355, "learning_rate": 1.8016698751623103e-05, "loss": 2.6805, "step": 6478 }, { "epoch": 0.2283006069575482, "grad_norm": 1.7767913341522217, "learning_rate": 1.801601648178123e-05, "loss": 1.0254, "step": 6479 }, { "epoch": 0.22833584397050663, "grad_norm": 3.0291216373443604, "learning_rate": 1.8015334107529718e-05, "loss": 3.8836, "step": 6480 }, { "epoch": 0.22837108098346504, "grad_norm": 1.5172251462936401, "learning_rate": 1.8014651628877462e-05, "loss": 0.912, "step": 6481 }, { "epoch": 0.22840631799642344, "grad_norm": 6.911700248718262, "learning_rate": 1.8013969045833345e-05, "loss": 5.6697, "step": 6482 }, { "epoch": 0.22844155500938185, "grad_norm": 7.287688732147217, "learning_rate": 1.8013286358406265e-05, "loss": 8.5326, "step": 6483 }, { "epoch": 0.22847679202234025, "grad_norm": 5.815036773681641, "learning_rate": 1.8012603566605105e-05, "loss": 3.8637, "step": 6484 }, { "epoch": 0.22851202903529869, "grad_norm": 1.230963945388794, "learning_rate": 1.8011920670438762e-05, "loss": 0.9261, "step": 6485 }, { "epoch": 0.2285472660482571, "grad_norm": 2.219282627105713, "learning_rate": 1.8011237669916135e-05, "loss": 1.0945, "step": 6486 }, { "epoch": 0.2285825030612155, "grad_norm": 3.117405414581299, "learning_rate": 1.8010554565046115e-05, "loss": 1.1344, "step": 6487 }, { "epoch": 0.2286177400741739, "grad_norm": 1.3775886297225952, "learning_rate": 1.8009871355837605e-05, "loss": 1.0396, "step": 6488 }, { "epoch": 0.22865297708713234, "grad_norm": 1.4501311779022217, "learning_rate": 1.8009188042299498e-05, "loss": 0.7563, "step": 6489 }, { "epoch": 0.22868821410009074, "grad_norm": 4.0658040046691895, "learning_rate": 1.8008504624440695e-05, "loss": 3.5775, "step": 6490 }, { "epoch": 0.22872345111304915, "grad_norm": 3.6869852542877197, "learning_rate": 1.80078211022701e-05, "loss": 3.8326, "step": 6491 }, { "epoch": 0.22875868812600755, "grad_norm": 1.3876663446426392, "learning_rate": 1.8007137475796615e-05, "loss": 1.0245, "step": 6492 }, { "epoch": 0.22879392513896596, "grad_norm": 5.403814315795898, "learning_rate": 1.8006453745029143e-05, "loss": 5.7881, "step": 6493 }, { "epoch": 0.2288291621519244, "grad_norm": 3.4332730770111084, "learning_rate": 1.8005769909976594e-05, "loss": 3.1051, "step": 6494 }, { "epoch": 0.2288643991648828, "grad_norm": 2.4988768100738525, "learning_rate": 1.8005085970647872e-05, "loss": 1.0355, "step": 6495 }, { "epoch": 0.2288996361778412, "grad_norm": 0.9098329544067383, "learning_rate": 1.8004401927051885e-05, "loss": 0.949, "step": 6496 }, { "epoch": 0.2289348731907996, "grad_norm": 2.8605306148529053, "learning_rate": 1.800371777919754e-05, "loss": 2.9567, "step": 6497 }, { "epoch": 0.22897011020375801, "grad_norm": 3.8867926597595215, "learning_rate": 1.8003033527093753e-05, "loss": 3.3137, "step": 6498 }, { "epoch": 0.22900534721671645, "grad_norm": 1.6387156248092651, "learning_rate": 1.8002349170749438e-05, "loss": 0.7852, "step": 6499 }, { "epoch": 0.22904058422967485, "grad_norm": 1.385168194770813, "learning_rate": 1.8001664710173503e-05, "loss": 0.7964, "step": 6500 }, { "epoch": 0.22907582124263326, "grad_norm": 5.616709232330322, "learning_rate": 1.8000980145374867e-05, "loss": 3.4022, "step": 6501 }, { "epoch": 0.22911105825559167, "grad_norm": 1.3951592445373535, "learning_rate": 1.8000295476362443e-05, "loss": 0.8368, "step": 6502 }, { "epoch": 0.2291462952685501, "grad_norm": 3.2768373489379883, "learning_rate": 1.7999610703145153e-05, "loss": 3.3054, "step": 6503 }, { "epoch": 0.2291815322815085, "grad_norm": 0.8854705691337585, "learning_rate": 1.799892582573192e-05, "loss": 0.9454, "step": 6504 }, { "epoch": 0.2292167692944669, "grad_norm": 1.6775808334350586, "learning_rate": 1.7998240844131652e-05, "loss": 1.0947, "step": 6505 }, { "epoch": 0.22925200630742532, "grad_norm": 1.1345489025115967, "learning_rate": 1.799755575835328e-05, "loss": 0.9533, "step": 6506 }, { "epoch": 0.22928724332038372, "grad_norm": 3.644300937652588, "learning_rate": 1.7996870568405727e-05, "loss": 3.1662, "step": 6507 }, { "epoch": 0.22932248033334215, "grad_norm": 7.7287774085998535, "learning_rate": 1.7996185274297913e-05, "loss": 3.6125, "step": 6508 }, { "epoch": 0.22935771734630056, "grad_norm": 1.328270435333252, "learning_rate": 1.7995499876038772e-05, "loss": 0.7931, "step": 6509 }, { "epoch": 0.22939295435925897, "grad_norm": 1.495725154876709, "learning_rate": 1.7994814373637225e-05, "loss": 0.9253, "step": 6510 }, { "epoch": 0.22942819137221737, "grad_norm": 6.885375499725342, "learning_rate": 1.79941287671022e-05, "loss": 3.9418, "step": 6511 }, { "epoch": 0.22946342838517578, "grad_norm": 6.4763288497924805, "learning_rate": 1.799344305644263e-05, "loss": 3.1842, "step": 6512 }, { "epoch": 0.2294986653981342, "grad_norm": 8.91511058807373, "learning_rate": 1.7992757241667445e-05, "loss": 3.6805, "step": 6513 }, { "epoch": 0.22953390241109262, "grad_norm": 1.3693088293075562, "learning_rate": 1.7992071322785585e-05, "loss": 0.9065, "step": 6514 }, { "epoch": 0.22956913942405102, "grad_norm": 6.627005577087402, "learning_rate": 1.799138529980597e-05, "loss": 3.6974, "step": 6515 }, { "epoch": 0.22960437643700943, "grad_norm": 2.103052854537964, "learning_rate": 1.7990699172737548e-05, "loss": 1.0497, "step": 6516 }, { "epoch": 0.22963961344996783, "grad_norm": 9.627921104431152, "learning_rate": 1.7990012941589253e-05, "loss": 6.0991, "step": 6517 }, { "epoch": 0.22967485046292627, "grad_norm": 5.2162041664123535, "learning_rate": 1.7989326606370018e-05, "loss": 3.3462, "step": 6518 }, { "epoch": 0.22971008747588467, "grad_norm": 0.9915599226951599, "learning_rate": 1.7988640167088787e-05, "loss": 0.9954, "step": 6519 }, { "epoch": 0.22974532448884308, "grad_norm": 4.768499374389648, "learning_rate": 1.7987953623754503e-05, "loss": 5.5013, "step": 6520 }, { "epoch": 0.22978056150180148, "grad_norm": 1.0831881761550903, "learning_rate": 1.7987266976376105e-05, "loss": 0.9089, "step": 6521 }, { "epoch": 0.22981579851475992, "grad_norm": 3.8907134532928467, "learning_rate": 1.7986580224962532e-05, "loss": 3.0861, "step": 6522 }, { "epoch": 0.22985103552771832, "grad_norm": 2.8400251865386963, "learning_rate": 1.798589336952274e-05, "loss": 0.9963, "step": 6523 }, { "epoch": 0.22988627254067673, "grad_norm": 1.3254399299621582, "learning_rate": 1.7985206410065667e-05, "loss": 1.0189, "step": 6524 }, { "epoch": 0.22992150955363513, "grad_norm": 3.1998674869537354, "learning_rate": 1.7984519346600263e-05, "loss": 3.0182, "step": 6525 }, { "epoch": 0.22995674656659354, "grad_norm": 6.553544521331787, "learning_rate": 1.7983832179135477e-05, "loss": 4.9356, "step": 6526 }, { "epoch": 0.22999198357955197, "grad_norm": 1.0827655792236328, "learning_rate": 1.7983144907680263e-05, "loss": 1.1321, "step": 6527 }, { "epoch": 0.23002722059251038, "grad_norm": 7.077274322509766, "learning_rate": 1.798245753224357e-05, "loss": 6.4619, "step": 6528 }, { "epoch": 0.23006245760546878, "grad_norm": 2.077566385269165, "learning_rate": 1.7981770052834347e-05, "loss": 0.935, "step": 6529 }, { "epoch": 0.2300976946184272, "grad_norm": 4.59989070892334, "learning_rate": 1.7981082469461554e-05, "loss": 3.4659, "step": 6530 }, { "epoch": 0.2301329316313856, "grad_norm": 4.808114528656006, "learning_rate": 1.7980394782134145e-05, "loss": 3.9337, "step": 6531 }, { "epoch": 0.23016816864434403, "grad_norm": 3.941987991333008, "learning_rate": 1.7979706990861076e-05, "loss": 3.1155, "step": 6532 }, { "epoch": 0.23020340565730243, "grad_norm": 1.4071866273880005, "learning_rate": 1.797901909565131e-05, "loss": 0.89, "step": 6533 }, { "epoch": 0.23023864267026084, "grad_norm": 3.6628799438476562, "learning_rate": 1.7978331096513802e-05, "loss": 3.7929, "step": 6534 }, { "epoch": 0.23027387968321925, "grad_norm": 1.2481756210327148, "learning_rate": 1.7977642993457517e-05, "loss": 0.8792, "step": 6535 }, { "epoch": 0.23030911669617765, "grad_norm": 1.3918042182922363, "learning_rate": 1.7976954786491415e-05, "loss": 0.9597, "step": 6536 }, { "epoch": 0.23034435370913608, "grad_norm": 1.023167610168457, "learning_rate": 1.797626647562446e-05, "loss": 0.8909, "step": 6537 }, { "epoch": 0.2303795907220945, "grad_norm": 1.181125283241272, "learning_rate": 1.797557806086562e-05, "loss": 0.8732, "step": 6538 }, { "epoch": 0.2304148277350529, "grad_norm": 2.416442632675171, "learning_rate": 1.7974889542223858e-05, "loss": 3.4145, "step": 6539 }, { "epoch": 0.2304500647480113, "grad_norm": 1.3276913166046143, "learning_rate": 1.7974200919708145e-05, "loss": 0.9772, "step": 6540 }, { "epoch": 0.23048530176096974, "grad_norm": 5.460262775421143, "learning_rate": 1.7973512193327446e-05, "loss": 3.4718, "step": 6541 }, { "epoch": 0.23052053877392814, "grad_norm": 6.901228427886963, "learning_rate": 1.797282336309074e-05, "loss": 3.9304, "step": 6542 }, { "epoch": 0.23055577578688655, "grad_norm": 1.3748207092285156, "learning_rate": 1.7972134429006992e-05, "loss": 0.9404, "step": 6543 }, { "epoch": 0.23059101279984495, "grad_norm": 6.825654029846191, "learning_rate": 1.7971445391085176e-05, "loss": 3.6619, "step": 6544 }, { "epoch": 0.23062624981280336, "grad_norm": 1.0287935733795166, "learning_rate": 1.797075624933427e-05, "loss": 0.8376, "step": 6545 }, { "epoch": 0.2306614868257618, "grad_norm": 4.450038433074951, "learning_rate": 1.7970067003763248e-05, "loss": 3.9348, "step": 6546 }, { "epoch": 0.2306967238387202, "grad_norm": 2.480008840560913, "learning_rate": 1.796937765438109e-05, "loss": 3.4407, "step": 6547 }, { "epoch": 0.2307319608516786, "grad_norm": 1.4963511228561401, "learning_rate": 1.796868820119677e-05, "loss": 0.9293, "step": 6548 }, { "epoch": 0.230767197864637, "grad_norm": 1.5547001361846924, "learning_rate": 1.7967998644219273e-05, "loss": 0.7965, "step": 6549 }, { "epoch": 0.2308024348775954, "grad_norm": 1.2551270723342896, "learning_rate": 1.7967308983457577e-05, "loss": 0.9041, "step": 6550 }, { "epoch": 0.23083767189055385, "grad_norm": 5.340678691864014, "learning_rate": 1.7966619218920666e-05, "loss": 5.4484, "step": 6551 }, { "epoch": 0.23087290890351225, "grad_norm": 2.360746145248413, "learning_rate": 1.7965929350617528e-05, "loss": 3.3603, "step": 6552 }, { "epoch": 0.23090814591647066, "grad_norm": 1.441655158996582, "learning_rate": 1.7965239378557143e-05, "loss": 1.095, "step": 6553 }, { "epoch": 0.23094338292942906, "grad_norm": 1.236246943473816, "learning_rate": 1.7964549302748503e-05, "loss": 0.832, "step": 6554 }, { "epoch": 0.2309786199423875, "grad_norm": 2.197939872741699, "learning_rate": 1.796385912320059e-05, "loss": 2.6839, "step": 6555 }, { "epoch": 0.2310138569553459, "grad_norm": 1.884284257888794, "learning_rate": 1.79631688399224e-05, "loss": 0.8353, "step": 6556 }, { "epoch": 0.2310490939683043, "grad_norm": 5.584051609039307, "learning_rate": 1.7962478452922923e-05, "loss": 5.8706, "step": 6557 }, { "epoch": 0.23108433098126271, "grad_norm": 1.191778540611267, "learning_rate": 1.7961787962211148e-05, "loss": 0.9978, "step": 6558 }, { "epoch": 0.23111956799422112, "grad_norm": 7.7381086349487305, "learning_rate": 1.7961097367796073e-05, "loss": 6.0469, "step": 6559 }, { "epoch": 0.23115480500717955, "grad_norm": 1.995512843132019, "learning_rate": 1.7960406669686686e-05, "loss": 0.9645, "step": 6560 }, { "epoch": 0.23119004202013796, "grad_norm": 2.0893354415893555, "learning_rate": 1.795971586789199e-05, "loss": 1.1767, "step": 6561 }, { "epoch": 0.23122527903309636, "grad_norm": 1.6405956745147705, "learning_rate": 1.795902496242098e-05, "loss": 1.1137, "step": 6562 }, { "epoch": 0.23126051604605477, "grad_norm": 1.3272883892059326, "learning_rate": 1.7958333953282657e-05, "loss": 1.1003, "step": 6563 }, { "epoch": 0.23129575305901318, "grad_norm": 3.1628401279449463, "learning_rate": 1.7957642840486023e-05, "loss": 3.8123, "step": 6564 }, { "epoch": 0.2313309900719716, "grad_norm": 0.8053258061408997, "learning_rate": 1.7956951624040076e-05, "loss": 0.8654, "step": 6565 }, { "epoch": 0.23136622708493002, "grad_norm": 4.335508346557617, "learning_rate": 1.795626030395382e-05, "loss": 4.1304, "step": 6566 }, { "epoch": 0.23140146409788842, "grad_norm": 1.4847694635391235, "learning_rate": 1.795556888023626e-05, "loss": 0.9933, "step": 6567 }, { "epoch": 0.23143670111084683, "grad_norm": 0.8556928634643555, "learning_rate": 1.7954877352896403e-05, "loss": 0.9538, "step": 6568 }, { "epoch": 0.23147193812380523, "grad_norm": 4.944585800170898, "learning_rate": 1.7954185721943254e-05, "loss": 3.3, "step": 6569 }, { "epoch": 0.23150717513676367, "grad_norm": 3.5578219890594482, "learning_rate": 1.7953493987385823e-05, "loss": 3.586, "step": 6570 }, { "epoch": 0.23154241214972207, "grad_norm": 1.6779416799545288, "learning_rate": 1.795280214923312e-05, "loss": 0.9569, "step": 6571 }, { "epoch": 0.23157764916268048, "grad_norm": 2.6985204219818115, "learning_rate": 1.7952110207494156e-05, "loss": 3.2453, "step": 6572 }, { "epoch": 0.23161288617563888, "grad_norm": 4.4735283851623535, "learning_rate": 1.7951418162177944e-05, "loss": 5.4467, "step": 6573 }, { "epoch": 0.23164812318859732, "grad_norm": 1.0837280750274658, "learning_rate": 1.7950726013293496e-05, "loss": 0.8668, "step": 6574 }, { "epoch": 0.23168336020155572, "grad_norm": 2.8714828491210938, "learning_rate": 1.795003376084983e-05, "loss": 3.5561, "step": 6575 }, { "epoch": 0.23171859721451413, "grad_norm": 5.177307605743408, "learning_rate": 1.794934140485596e-05, "loss": 8.3172, "step": 6576 }, { "epoch": 0.23175383422747253, "grad_norm": 4.585881233215332, "learning_rate": 1.7948648945320906e-05, "loss": 5.6177, "step": 6577 }, { "epoch": 0.23178907124043094, "grad_norm": 2.2361645698547363, "learning_rate": 1.7947956382253692e-05, "loss": 3.5975, "step": 6578 }, { "epoch": 0.23182430825338937, "grad_norm": 4.813451766967773, "learning_rate": 1.7947263715663327e-05, "loss": 3.6322, "step": 6579 }, { "epoch": 0.23185954526634778, "grad_norm": 3.5328710079193115, "learning_rate": 1.7946570945558844e-05, "loss": 3.5447, "step": 6580 }, { "epoch": 0.23189478227930618, "grad_norm": 3.375053882598877, "learning_rate": 1.7945878071949262e-05, "loss": 4.1778, "step": 6581 }, { "epoch": 0.2319300192922646, "grad_norm": 1.1556910276412964, "learning_rate": 1.7945185094843602e-05, "loss": 0.8458, "step": 6582 }, { "epoch": 0.231965256305223, "grad_norm": 1.2486590147018433, "learning_rate": 1.7944492014250898e-05, "loss": 0.985, "step": 6583 }, { "epoch": 0.23200049331818143, "grad_norm": 3.1041951179504395, "learning_rate": 1.7943798830180173e-05, "loss": 3.3459, "step": 6584 }, { "epoch": 0.23203573033113983, "grad_norm": 4.263928413391113, "learning_rate": 1.7943105542640455e-05, "loss": 3.555, "step": 6585 }, { "epoch": 0.23207096734409824, "grad_norm": 1.7044165134429932, "learning_rate": 1.7942412151640777e-05, "loss": 1.0563, "step": 6586 }, { "epoch": 0.23210620435705664, "grad_norm": 1.2190524339675903, "learning_rate": 1.7941718657190167e-05, "loss": 0.7898, "step": 6587 }, { "epoch": 0.23214144137001505, "grad_norm": 1.9867513179779053, "learning_rate": 1.794102505929766e-05, "loss": 0.8888, "step": 6588 }, { "epoch": 0.23217667838297348, "grad_norm": 1.54514479637146, "learning_rate": 1.7940331357972295e-05, "loss": 0.7939, "step": 6589 }, { "epoch": 0.2322119153959319, "grad_norm": 3.221223831176758, "learning_rate": 1.7939637553223098e-05, "loss": 4.0394, "step": 6590 }, { "epoch": 0.2322471524088903, "grad_norm": 3.09162974357605, "learning_rate": 1.793894364505911e-05, "loss": 3.6453, "step": 6591 }, { "epoch": 0.2322823894218487, "grad_norm": 4.339426040649414, "learning_rate": 1.793824963348937e-05, "loss": 6.7486, "step": 6592 }, { "epoch": 0.23231762643480713, "grad_norm": 1.6717183589935303, "learning_rate": 1.793755551852292e-05, "loss": 1.0546, "step": 6593 }, { "epoch": 0.23235286344776554, "grad_norm": 3.876668930053711, "learning_rate": 1.7936861300168796e-05, "loss": 5.8574, "step": 6594 }, { "epoch": 0.23238810046072395, "grad_norm": 3.331343650817871, "learning_rate": 1.7936166978436045e-05, "loss": 0.8962, "step": 6595 }, { "epoch": 0.23242333747368235, "grad_norm": 4.133443355560303, "learning_rate": 1.7935472553333703e-05, "loss": 6.4286, "step": 6596 }, { "epoch": 0.23245857448664076, "grad_norm": 2.6453135013580322, "learning_rate": 1.7934778024870825e-05, "loss": 3.3663, "step": 6597 }, { "epoch": 0.2324938114995992, "grad_norm": 2.658050060272217, "learning_rate": 1.7934083393056452e-05, "loss": 2.7721, "step": 6598 }, { "epoch": 0.2325290485125576, "grad_norm": 6.643539905548096, "learning_rate": 1.7933388657899632e-05, "loss": 6.1757, "step": 6599 }, { "epoch": 0.232564285525516, "grad_norm": 3.007305860519409, "learning_rate": 1.7932693819409413e-05, "loss": 3.7446, "step": 6600 }, { "epoch": 0.2325995225384744, "grad_norm": 3.5161571502685547, "learning_rate": 1.7931998877594846e-05, "loss": 3.4665, "step": 6601 }, { "epoch": 0.2326347595514328, "grad_norm": 1.4759161472320557, "learning_rate": 1.7931303832464984e-05, "loss": 1.029, "step": 6602 }, { "epoch": 0.23266999656439125, "grad_norm": 1.553680658340454, "learning_rate": 1.793060868402888e-05, "loss": 1.1853, "step": 6603 }, { "epoch": 0.23270523357734965, "grad_norm": 1.374979019165039, "learning_rate": 1.7929913432295585e-05, "loss": 1.1806, "step": 6604 }, { "epoch": 0.23274047059030806, "grad_norm": 1.8901994228363037, "learning_rate": 1.7929218077274162e-05, "loss": 1.0655, "step": 6605 }, { "epoch": 0.23277570760326646, "grad_norm": 0.949558675289154, "learning_rate": 1.792852261897366e-05, "loss": 1.1171, "step": 6606 }, { "epoch": 0.2328109446162249, "grad_norm": 1.5587265491485596, "learning_rate": 1.7927827057403144e-05, "loss": 1.0246, "step": 6607 }, { "epoch": 0.2328461816291833, "grad_norm": 1.541548252105713, "learning_rate": 1.7927131392571668e-05, "loss": 0.853, "step": 6608 }, { "epoch": 0.2328814186421417, "grad_norm": 4.176589012145996, "learning_rate": 1.79264356244883e-05, "loss": 5.5246, "step": 6609 }, { "epoch": 0.2329166556551001, "grad_norm": 6.942305564880371, "learning_rate": 1.7925739753162093e-05, "loss": 6.9299, "step": 6610 }, { "epoch": 0.23295189266805852, "grad_norm": 2.570438861846924, "learning_rate": 1.7925043778602117e-05, "loss": 3.1268, "step": 6611 }, { "epoch": 0.23298712968101695, "grad_norm": 1.3330711126327515, "learning_rate": 1.7924347700817438e-05, "loss": 0.817, "step": 6612 }, { "epoch": 0.23302236669397536, "grad_norm": 3.5497653484344482, "learning_rate": 1.792365151981712e-05, "loss": 3.3145, "step": 6613 }, { "epoch": 0.23305760370693376, "grad_norm": 7.446408748626709, "learning_rate": 1.7922955235610232e-05, "loss": 6.266, "step": 6614 }, { "epoch": 0.23309284071989217, "grad_norm": 5.53207540512085, "learning_rate": 1.7922258848205842e-05, "loss": 3.5122, "step": 6615 }, { "epoch": 0.23312807773285057, "grad_norm": 4.733396530151367, "learning_rate": 1.7921562357613025e-05, "loss": 3.3286, "step": 6616 }, { "epoch": 0.233163314745809, "grad_norm": 5.236940383911133, "learning_rate": 1.7920865763840845e-05, "loss": 5.8297, "step": 6617 }, { "epoch": 0.23319855175876741, "grad_norm": 4.005953788757324, "learning_rate": 1.792016906689838e-05, "loss": 3.3749, "step": 6618 }, { "epoch": 0.23323378877172582, "grad_norm": 5.263134002685547, "learning_rate": 1.7919472266794708e-05, "loss": 4.3916, "step": 6619 }, { "epoch": 0.23326902578468423, "grad_norm": 3.4152114391326904, "learning_rate": 1.7918775363538894e-05, "loss": 2.8975, "step": 6620 }, { "epoch": 0.23330426279764263, "grad_norm": 2.4959654808044434, "learning_rate": 1.791807835714003e-05, "loss": 3.2445, "step": 6621 }, { "epoch": 0.23333949981060106, "grad_norm": 1.1911070346832275, "learning_rate": 1.791738124760718e-05, "loss": 0.7282, "step": 6622 }, { "epoch": 0.23337473682355947, "grad_norm": 2.3927316665649414, "learning_rate": 1.7916684034949434e-05, "loss": 0.9044, "step": 6623 }, { "epoch": 0.23340997383651788, "grad_norm": 5.391152381896973, "learning_rate": 1.791598671917587e-05, "loss": 2.8463, "step": 6624 }, { "epoch": 0.23344521084947628, "grad_norm": 1.2292100191116333, "learning_rate": 1.791528930029557e-05, "loss": 1.0164, "step": 6625 }, { "epoch": 0.23348044786243471, "grad_norm": 6.751328468322754, "learning_rate": 1.791459177831762e-05, "loss": 3.362, "step": 6626 }, { "epoch": 0.23351568487539312, "grad_norm": 4.242156505584717, "learning_rate": 1.79138941532511e-05, "loss": 3.79, "step": 6627 }, { "epoch": 0.23355092188835153, "grad_norm": 3.807621479034424, "learning_rate": 1.7913196425105103e-05, "loss": 3.4211, "step": 6628 }, { "epoch": 0.23358615890130993, "grad_norm": 5.207540035247803, "learning_rate": 1.7912498593888717e-05, "loss": 3.7728, "step": 6629 }, { "epoch": 0.23362139591426834, "grad_norm": 0.9204218983650208, "learning_rate": 1.7911800659611028e-05, "loss": 0.8253, "step": 6630 }, { "epoch": 0.23365663292722677, "grad_norm": 5.520493507385254, "learning_rate": 1.7911102622281124e-05, "loss": 2.4788, "step": 6631 }, { "epoch": 0.23369186994018518, "grad_norm": 1.2978851795196533, "learning_rate": 1.7910404481908104e-05, "loss": 1.036, "step": 6632 }, { "epoch": 0.23372710695314358, "grad_norm": 1.3608918190002441, "learning_rate": 1.7909706238501058e-05, "loss": 0.9827, "step": 6633 }, { "epoch": 0.233762343966102, "grad_norm": 1.4060062170028687, "learning_rate": 1.7909007892069083e-05, "loss": 1.1451, "step": 6634 }, { "epoch": 0.2337975809790604, "grad_norm": 0.9481850266456604, "learning_rate": 1.7908309442621265e-05, "loss": 1.0168, "step": 6635 }, { "epoch": 0.23383281799201883, "grad_norm": 4.809579372406006, "learning_rate": 1.7907610890166716e-05, "loss": 3.6271, "step": 6636 }, { "epoch": 0.23386805500497723, "grad_norm": 11.095614433288574, "learning_rate": 1.7906912234714523e-05, "loss": 3.7049, "step": 6637 }, { "epoch": 0.23390329201793564, "grad_norm": 6.596689701080322, "learning_rate": 1.7906213476273796e-05, "loss": 3.9979, "step": 6638 }, { "epoch": 0.23393852903089404, "grad_norm": 1.117795467376709, "learning_rate": 1.790551461485363e-05, "loss": 1.07, "step": 6639 }, { "epoch": 0.23397376604385245, "grad_norm": 1.6343045234680176, "learning_rate": 1.790481565046313e-05, "loss": 1.0401, "step": 6640 }, { "epoch": 0.23400900305681088, "grad_norm": 2.753363609313965, "learning_rate": 1.7904116583111393e-05, "loss": 3.2946, "step": 6641 }, { "epoch": 0.2340442400697693, "grad_norm": 9.933328628540039, "learning_rate": 1.7903417412807536e-05, "loss": 8.5354, "step": 6642 }, { "epoch": 0.2340794770827277, "grad_norm": 1.4908665418624878, "learning_rate": 1.7902718139560658e-05, "loss": 0.8701, "step": 6643 }, { "epoch": 0.2341147140956861, "grad_norm": 1.879471778869629, "learning_rate": 1.7902018763379873e-05, "loss": 0.7514, "step": 6644 }, { "epoch": 0.23414995110864453, "grad_norm": 2.4038784503936768, "learning_rate": 1.7901319284274285e-05, "loss": 3.4831, "step": 6645 }, { "epoch": 0.23418518812160294, "grad_norm": 6.2241740226745605, "learning_rate": 1.790061970225301e-05, "loss": 3.5879, "step": 6646 }, { "epoch": 0.23422042513456134, "grad_norm": 11.583026885986328, "learning_rate": 1.789992001732515e-05, "loss": 5.353, "step": 6647 }, { "epoch": 0.23425566214751975, "grad_norm": 4.321437835693359, "learning_rate": 1.789922022949983e-05, "loss": 3.5452, "step": 6648 }, { "epoch": 0.23429089916047816, "grad_norm": 6.096137046813965, "learning_rate": 1.789852033878616e-05, "loss": 3.7369, "step": 6649 }, { "epoch": 0.2343261361734366, "grad_norm": 5.271198749542236, "learning_rate": 1.7897820345193256e-05, "loss": 10.1553, "step": 6650 }, { "epoch": 0.234361373186395, "grad_norm": 5.509321212768555, "learning_rate": 1.7897120248730235e-05, "loss": 4.3736, "step": 6651 }, { "epoch": 0.2343966101993534, "grad_norm": 2.0788822174072266, "learning_rate": 1.789642004940622e-05, "loss": 3.4728, "step": 6652 }, { "epoch": 0.2344318472123118, "grad_norm": 5.226025581359863, "learning_rate": 1.7895719747230325e-05, "loss": 3.6458, "step": 6653 }, { "epoch": 0.2344670842252702, "grad_norm": 4.487889289855957, "learning_rate": 1.7895019342211678e-05, "loss": 3.0458, "step": 6654 }, { "epoch": 0.23450232123822864, "grad_norm": 1.3401880264282227, "learning_rate": 1.7894318834359395e-05, "loss": 0.8717, "step": 6655 }, { "epoch": 0.23453755825118705, "grad_norm": 4.229952335357666, "learning_rate": 1.7893618223682604e-05, "loss": 3.6422, "step": 6656 }, { "epoch": 0.23457279526414546, "grad_norm": 1.662569284439087, "learning_rate": 1.789291751019043e-05, "loss": 0.8826, "step": 6657 }, { "epoch": 0.23460803227710386, "grad_norm": 3.100841760635376, "learning_rate": 1.7892216693892e-05, "loss": 3.2978, "step": 6658 }, { "epoch": 0.2346432692900623, "grad_norm": 7.548559188842773, "learning_rate": 1.7891515774796447e-05, "loss": 1.2499, "step": 6659 }, { "epoch": 0.2346785063030207, "grad_norm": 3.8576083183288574, "learning_rate": 1.789081475291289e-05, "loss": 2.6554, "step": 6660 }, { "epoch": 0.2347137433159791, "grad_norm": 1.737402319908142, "learning_rate": 1.789011362825047e-05, "loss": 0.8636, "step": 6661 }, { "epoch": 0.2347489803289375, "grad_norm": 1.5509483814239502, "learning_rate": 1.788941240081831e-05, "loss": 1.1641, "step": 6662 }, { "epoch": 0.23478421734189592, "grad_norm": 1.3160820007324219, "learning_rate": 1.7888711070625553e-05, "loss": 0.9389, "step": 6663 }, { "epoch": 0.23481945435485435, "grad_norm": 0.9071500897407532, "learning_rate": 1.7888009637681328e-05, "loss": 1.0435, "step": 6664 }, { "epoch": 0.23485469136781276, "grad_norm": 2.36924409866333, "learning_rate": 1.788730810199477e-05, "loss": 1.3772, "step": 6665 }, { "epoch": 0.23488992838077116, "grad_norm": 4.585844039916992, "learning_rate": 1.788660646357502e-05, "loss": 3.5766, "step": 6666 }, { "epoch": 0.23492516539372957, "grad_norm": 2.162031412124634, "learning_rate": 1.788590472243122e-05, "loss": 1.1829, "step": 6667 }, { "epoch": 0.23496040240668797, "grad_norm": 5.278262615203857, "learning_rate": 1.7885202878572502e-05, "loss": 5.2951, "step": 6668 }, { "epoch": 0.2349956394196464, "grad_norm": 6.102953910827637, "learning_rate": 1.7884500932008016e-05, "loss": 4.5178, "step": 6669 }, { "epoch": 0.2350308764326048, "grad_norm": 1.6550002098083496, "learning_rate": 1.78837988827469e-05, "loss": 0.7495, "step": 6670 }, { "epoch": 0.23506611344556322, "grad_norm": 1.1570900678634644, "learning_rate": 1.7883096730798296e-05, "loss": 1.0662, "step": 6671 }, { "epoch": 0.23510135045852162, "grad_norm": 1.5432039499282837, "learning_rate": 1.788239447617136e-05, "loss": 0.9156, "step": 6672 }, { "epoch": 0.23513658747148003, "grad_norm": 4.8556623458862305, "learning_rate": 1.7881692118875226e-05, "loss": 6.0732, "step": 6673 }, { "epoch": 0.23517182448443846, "grad_norm": 4.929468631744385, "learning_rate": 1.788098965891905e-05, "loss": 3.0314, "step": 6674 }, { "epoch": 0.23520706149739687, "grad_norm": 3.9774010181427, "learning_rate": 1.788028709631198e-05, "loss": 5.9014, "step": 6675 }, { "epoch": 0.23524229851035527, "grad_norm": 1.969011664390564, "learning_rate": 1.7879584431063165e-05, "loss": 0.8033, "step": 6676 }, { "epoch": 0.23527753552331368, "grad_norm": 3.5774669647216797, "learning_rate": 1.787888166318176e-05, "loss": 3.4941, "step": 6677 }, { "epoch": 0.2353127725362721, "grad_norm": 8.237168312072754, "learning_rate": 1.7878178792676916e-05, "loss": 6.0601, "step": 6678 }, { "epoch": 0.23534800954923052, "grad_norm": 2.4375925064086914, "learning_rate": 1.787747581955779e-05, "loss": 3.3591, "step": 6679 }, { "epoch": 0.23538324656218892, "grad_norm": 3.323220729827881, "learning_rate": 1.7876772743833544e-05, "loss": 2.7733, "step": 6680 }, { "epoch": 0.23541848357514733, "grad_norm": 4.666815280914307, "learning_rate": 1.7876069565513323e-05, "loss": 3.4364, "step": 6681 }, { "epoch": 0.23545372058810574, "grad_norm": 2.1187641620635986, "learning_rate": 1.7875366284606298e-05, "loss": 0.978, "step": 6682 }, { "epoch": 0.23548895760106417, "grad_norm": 3.7104156017303467, "learning_rate": 1.787466290112162e-05, "loss": 3.0333, "step": 6683 }, { "epoch": 0.23552419461402258, "grad_norm": 4.587975978851318, "learning_rate": 1.7873959415068455e-05, "loss": 4.0048, "step": 6684 }, { "epoch": 0.23555943162698098, "grad_norm": 4.903585910797119, "learning_rate": 1.787325582645597e-05, "loss": 4.3056, "step": 6685 }, { "epoch": 0.2355946686399394, "grad_norm": 1.6666007041931152, "learning_rate": 1.787255213529332e-05, "loss": 1.2745, "step": 6686 }, { "epoch": 0.2356299056528978, "grad_norm": 1.2896689176559448, "learning_rate": 1.7871848341589677e-05, "loss": 0.866, "step": 6687 }, { "epoch": 0.23566514266585623, "grad_norm": 0.89671790599823, "learning_rate": 1.7871144445354206e-05, "loss": 0.8088, "step": 6688 }, { "epoch": 0.23570037967881463, "grad_norm": 1.1379482746124268, "learning_rate": 1.7870440446596075e-05, "loss": 1.1729, "step": 6689 }, { "epoch": 0.23573561669177304, "grad_norm": 2.472324848175049, "learning_rate": 1.7869736345324458e-05, "loss": 0.8386, "step": 6690 }, { "epoch": 0.23577085370473144, "grad_norm": 4.928866863250732, "learning_rate": 1.7869032141548518e-05, "loss": 6.1215, "step": 6691 }, { "epoch": 0.23580609071768985, "grad_norm": 1.7622941732406616, "learning_rate": 1.7868327835277438e-05, "loss": 0.9863, "step": 6692 }, { "epoch": 0.23584132773064828, "grad_norm": 3.220186471939087, "learning_rate": 1.7867623426520382e-05, "loss": 1.1086, "step": 6693 }, { "epoch": 0.2358765647436067, "grad_norm": 3.4804487228393555, "learning_rate": 1.786691891528653e-05, "loss": 3.6733, "step": 6694 }, { "epoch": 0.2359118017565651, "grad_norm": 3.509697675704956, "learning_rate": 1.7866214301585058e-05, "loss": 3.3658, "step": 6695 }, { "epoch": 0.2359470387695235, "grad_norm": 1.5007063150405884, "learning_rate": 1.786550958542514e-05, "loss": 0.9161, "step": 6696 }, { "epoch": 0.23598227578248193, "grad_norm": 2.8012309074401855, "learning_rate": 1.786480476681596e-05, "loss": 2.6376, "step": 6697 }, { "epoch": 0.23601751279544034, "grad_norm": 1.5081382989883423, "learning_rate": 1.7864099845766692e-05, "loss": 0.943, "step": 6698 }, { "epoch": 0.23605274980839874, "grad_norm": 1.5628854036331177, "learning_rate": 1.7863394822286527e-05, "loss": 1.1961, "step": 6699 }, { "epoch": 0.23608798682135715, "grad_norm": 1.9804800748825073, "learning_rate": 1.7862689696384635e-05, "loss": 0.957, "step": 6700 }, { "epoch": 0.23612322383431555, "grad_norm": 1.1989476680755615, "learning_rate": 1.7861984468070216e-05, "loss": 0.8341, "step": 6701 }, { "epoch": 0.236158460847274, "grad_norm": 1.8880360126495361, "learning_rate": 1.7861279137352444e-05, "loss": 0.969, "step": 6702 }, { "epoch": 0.2361936978602324, "grad_norm": 6.043784141540527, "learning_rate": 1.7860573704240515e-05, "loss": 5.6866, "step": 6703 }, { "epoch": 0.2362289348731908, "grad_norm": 1.9199752807617188, "learning_rate": 1.7859868168743607e-05, "loss": 2.9427, "step": 6704 }, { "epoch": 0.2362641718861492, "grad_norm": 3.2930920124053955, "learning_rate": 1.7859162530870916e-05, "loss": 3.211, "step": 6705 }, { "epoch": 0.2362994088991076, "grad_norm": 2.1153101921081543, "learning_rate": 1.7858456790631634e-05, "loss": 3.0607, "step": 6706 }, { "epoch": 0.23633464591206604, "grad_norm": 1.7977573871612549, "learning_rate": 1.785775094803495e-05, "loss": 0.8886, "step": 6707 }, { "epoch": 0.23636988292502445, "grad_norm": 1.4326729774475098, "learning_rate": 1.785704500309006e-05, "loss": 0.759, "step": 6708 }, { "epoch": 0.23640511993798285, "grad_norm": 1.517636775970459, "learning_rate": 1.7856338955806155e-05, "loss": 0.9701, "step": 6709 }, { "epoch": 0.23644035695094126, "grad_norm": 1.8248621225357056, "learning_rate": 1.7855632806192437e-05, "loss": 0.8153, "step": 6710 }, { "epoch": 0.2364755939638997, "grad_norm": 5.953883171081543, "learning_rate": 1.7854926554258098e-05, "loss": 5.9769, "step": 6711 }, { "epoch": 0.2365108309768581, "grad_norm": 1.0768190622329712, "learning_rate": 1.7854220200012343e-05, "loss": 1.0003, "step": 6712 }, { "epoch": 0.2365460679898165, "grad_norm": 1.1520487070083618, "learning_rate": 1.785351374346437e-05, "loss": 1.0053, "step": 6713 }, { "epoch": 0.2365813050027749, "grad_norm": 6.628714561462402, "learning_rate": 1.7852807184623376e-05, "loss": 5.9453, "step": 6714 }, { "epoch": 0.23661654201573332, "grad_norm": 6.173202991485596, "learning_rate": 1.7852100523498574e-05, "loss": 3.2495, "step": 6715 }, { "epoch": 0.23665177902869175, "grad_norm": 3.0403873920440674, "learning_rate": 1.7851393760099158e-05, "loss": 2.9304, "step": 6716 }, { "epoch": 0.23668701604165016, "grad_norm": 3.0683858394622803, "learning_rate": 1.785068689443434e-05, "loss": 4.0831, "step": 6717 }, { "epoch": 0.23672225305460856, "grad_norm": 1.1147115230560303, "learning_rate": 1.7849979926513324e-05, "loss": 1.0285, "step": 6718 }, { "epoch": 0.23675749006756697, "grad_norm": 10.392866134643555, "learning_rate": 1.7849272856345318e-05, "loss": 6.3645, "step": 6719 }, { "epoch": 0.23679272708052537, "grad_norm": 1.1284620761871338, "learning_rate": 1.784856568393954e-05, "loss": 1.0229, "step": 6720 }, { "epoch": 0.2368279640934838, "grad_norm": 5.060233116149902, "learning_rate": 1.7847858409305186e-05, "loss": 3.297, "step": 6721 }, { "epoch": 0.2368632011064422, "grad_norm": 1.932072401046753, "learning_rate": 1.7847151032451482e-05, "loss": 0.8878, "step": 6722 }, { "epoch": 0.23689843811940062, "grad_norm": 1.3656929731369019, "learning_rate": 1.7846443553387634e-05, "loss": 0.8615, "step": 6723 }, { "epoch": 0.23693367513235902, "grad_norm": 1.8694831132888794, "learning_rate": 1.784573597212286e-05, "loss": 1.1233, "step": 6724 }, { "epoch": 0.23696891214531743, "grad_norm": 2.5393788814544678, "learning_rate": 1.7845028288666375e-05, "loss": 3.3863, "step": 6725 }, { "epoch": 0.23700414915827586, "grad_norm": 1.3444322347640991, "learning_rate": 1.7844320503027398e-05, "loss": 0.9248, "step": 6726 }, { "epoch": 0.23703938617123427, "grad_norm": 4.794533729553223, "learning_rate": 1.7843612615215143e-05, "loss": 3.3565, "step": 6727 }, { "epoch": 0.23707462318419267, "grad_norm": 2.5293173789978027, "learning_rate": 1.7842904625238843e-05, "loss": 3.1298, "step": 6728 }, { "epoch": 0.23710986019715108, "grad_norm": 5.119068145751953, "learning_rate": 1.7842196533107706e-05, "loss": 4.9807, "step": 6729 }, { "epoch": 0.2371450972101095, "grad_norm": 1.3505898714065552, "learning_rate": 1.7841488338830962e-05, "loss": 1.1012, "step": 6730 }, { "epoch": 0.23718033422306792, "grad_norm": 1.1955387592315674, "learning_rate": 1.784078004241783e-05, "loss": 0.8941, "step": 6731 }, { "epoch": 0.23721557123602632, "grad_norm": 1.1466429233551025, "learning_rate": 1.7840071643877542e-05, "loss": 1.1535, "step": 6732 }, { "epoch": 0.23725080824898473, "grad_norm": 1.5240241289138794, "learning_rate": 1.7839363143219322e-05, "loss": 1.0451, "step": 6733 }, { "epoch": 0.23728604526194313, "grad_norm": 0.9965428709983826, "learning_rate": 1.7838654540452398e-05, "loss": 0.8433, "step": 6734 }, { "epoch": 0.23732128227490157, "grad_norm": 3.7371764183044434, "learning_rate": 1.7837945835586002e-05, "loss": 2.9058, "step": 6735 }, { "epoch": 0.23735651928785997, "grad_norm": 1.8749041557312012, "learning_rate": 1.783723702862936e-05, "loss": 0.8601, "step": 6736 }, { "epoch": 0.23739175630081838, "grad_norm": 2.056238889694214, "learning_rate": 1.783652811959171e-05, "loss": 0.952, "step": 6737 }, { "epoch": 0.23742699331377679, "grad_norm": 5.374650478363037, "learning_rate": 1.783581910848228e-05, "loss": 3.4304, "step": 6738 }, { "epoch": 0.2374622303267352, "grad_norm": 10.427715301513672, "learning_rate": 1.7835109995310313e-05, "loss": 3.4324, "step": 6739 }, { "epoch": 0.23749746733969362, "grad_norm": 1.0012078285217285, "learning_rate": 1.7834400780085035e-05, "loss": 1.0992, "step": 6740 }, { "epoch": 0.23753270435265203, "grad_norm": 9.411230087280273, "learning_rate": 1.783369146281569e-05, "loss": 3.684, "step": 6741 }, { "epoch": 0.23756794136561044, "grad_norm": 1.555137038230896, "learning_rate": 1.7832982043511514e-05, "loss": 1.0833, "step": 6742 }, { "epoch": 0.23760317837856884, "grad_norm": 1.1871466636657715, "learning_rate": 1.7832272522181753e-05, "loss": 0.7629, "step": 6743 }, { "epoch": 0.23763841539152725, "grad_norm": 1.2248671054840088, "learning_rate": 1.7831562898835642e-05, "loss": 0.9217, "step": 6744 }, { "epoch": 0.23767365240448568, "grad_norm": 5.849634647369385, "learning_rate": 1.783085317348243e-05, "loss": 3.6628, "step": 6745 }, { "epoch": 0.23770888941744409, "grad_norm": 5.735135555267334, "learning_rate": 1.7830143346131355e-05, "loss": 3.9061, "step": 6746 }, { "epoch": 0.2377441264304025, "grad_norm": 0.9695709943771362, "learning_rate": 1.782943341679167e-05, "loss": 1.0296, "step": 6747 }, { "epoch": 0.2377793634433609, "grad_norm": 0.9350019693374634, "learning_rate": 1.7828723385472613e-05, "loss": 1.2438, "step": 6748 }, { "epoch": 0.23781460045631933, "grad_norm": 1.1025458574295044, "learning_rate": 1.7828013252183438e-05, "loss": 0.8431, "step": 6749 }, { "epoch": 0.23784983746927774, "grad_norm": 1.465380311012268, "learning_rate": 1.7827303016933394e-05, "loss": 0.7198, "step": 6750 }, { "epoch": 0.23788507448223614, "grad_norm": 1.9653161764144897, "learning_rate": 1.782659267973173e-05, "loss": 0.987, "step": 6751 }, { "epoch": 0.23792031149519455, "grad_norm": 3.2775685787200928, "learning_rate": 1.78258822405877e-05, "loss": 3.1958, "step": 6752 }, { "epoch": 0.23795554850815295, "grad_norm": 7.102406978607178, "learning_rate": 1.7825171699510556e-05, "loss": 5.6655, "step": 6753 }, { "epoch": 0.2379907855211114, "grad_norm": 2.6441638469696045, "learning_rate": 1.7824461056509554e-05, "loss": 2.5093, "step": 6754 }, { "epoch": 0.2380260225340698, "grad_norm": 1.1437132358551025, "learning_rate": 1.782375031159395e-05, "loss": 1.368, "step": 6755 }, { "epoch": 0.2380612595470282, "grad_norm": 1.2100155353546143, "learning_rate": 1.7823039464773004e-05, "loss": 1.1046, "step": 6756 }, { "epoch": 0.2380964965599866, "grad_norm": 1.761968731880188, "learning_rate": 1.782232851605597e-05, "loss": 1.0296, "step": 6757 }, { "epoch": 0.238131733572945, "grad_norm": 1.1453086137771606, "learning_rate": 1.782161746545211e-05, "loss": 1.0973, "step": 6758 }, { "epoch": 0.23816697058590344, "grad_norm": 5.219735622406006, "learning_rate": 1.782090631297069e-05, "loss": 6.0596, "step": 6759 }, { "epoch": 0.23820220759886185, "grad_norm": 2.256537675857544, "learning_rate": 1.7820195058620967e-05, "loss": 4.0085, "step": 6760 }, { "epoch": 0.23823744461182025, "grad_norm": 1.634698510169983, "learning_rate": 1.7819483702412203e-05, "loss": 0.9107, "step": 6761 }, { "epoch": 0.23827268162477866, "grad_norm": 6.038050651550293, "learning_rate": 1.7818772244353674e-05, "loss": 3.734, "step": 6762 }, { "epoch": 0.2383079186377371, "grad_norm": 1.1797187328338623, "learning_rate": 1.7818060684454638e-05, "loss": 0.894, "step": 6763 }, { "epoch": 0.2383431556506955, "grad_norm": 7.105073928833008, "learning_rate": 1.7817349022724366e-05, "loss": 5.6784, "step": 6764 }, { "epoch": 0.2383783926636539, "grad_norm": 0.9098158478736877, "learning_rate": 1.781663725917213e-05, "loss": 1.2229, "step": 6765 }, { "epoch": 0.2384136296766123, "grad_norm": 2.6417455673217773, "learning_rate": 1.781592539380719e-05, "loss": 3.631, "step": 6766 }, { "epoch": 0.23844886668957072, "grad_norm": 1.3110032081604004, "learning_rate": 1.7815213426638834e-05, "loss": 0.8465, "step": 6767 }, { "epoch": 0.23848410370252915, "grad_norm": 1.2564113140106201, "learning_rate": 1.7814501357676324e-05, "loss": 0.8951, "step": 6768 }, { "epoch": 0.23851934071548755, "grad_norm": 1.2378755807876587, "learning_rate": 1.7813789186928942e-05, "loss": 0.956, "step": 6769 }, { "epoch": 0.23855457772844596, "grad_norm": 9.221689224243164, "learning_rate": 1.7813076914405956e-05, "loss": 3.8949, "step": 6770 }, { "epoch": 0.23858981474140437, "grad_norm": 4.390349388122559, "learning_rate": 1.7812364540116648e-05, "loss": 4.0065, "step": 6771 }, { "epoch": 0.23862505175436277, "grad_norm": 1.1520475149154663, "learning_rate": 1.78116520640703e-05, "loss": 0.8623, "step": 6772 }, { "epoch": 0.2386602887673212, "grad_norm": 1.1929353475570679, "learning_rate": 1.7810939486276186e-05, "loss": 0.8765, "step": 6773 }, { "epoch": 0.2386955257802796, "grad_norm": 1.494659423828125, "learning_rate": 1.781022680674359e-05, "loss": 0.8086, "step": 6774 }, { "epoch": 0.23873076279323802, "grad_norm": 1.2837060689926147, "learning_rate": 1.7809514025481798e-05, "loss": 1.0504, "step": 6775 }, { "epoch": 0.23876599980619642, "grad_norm": 1.399651050567627, "learning_rate": 1.780880114250009e-05, "loss": 1.0797, "step": 6776 }, { "epoch": 0.23880123681915483, "grad_norm": 4.054288387298584, "learning_rate": 1.780808815780775e-05, "loss": 5.8825, "step": 6777 }, { "epoch": 0.23883647383211326, "grad_norm": 1.1216490268707275, "learning_rate": 1.780737507141407e-05, "loss": 1.1284, "step": 6778 }, { "epoch": 0.23887171084507167, "grad_norm": 4.936560153961182, "learning_rate": 1.7806661883328336e-05, "loss": 6.1267, "step": 6779 }, { "epoch": 0.23890694785803007, "grad_norm": 1.6280863285064697, "learning_rate": 1.7805948593559835e-05, "loss": 1.0855, "step": 6780 }, { "epoch": 0.23894218487098848, "grad_norm": 2.8061835765838623, "learning_rate": 1.780523520211786e-05, "loss": 6.1205, "step": 6781 }, { "epoch": 0.2389774218839469, "grad_norm": 4.014029026031494, "learning_rate": 1.78045217090117e-05, "loss": 4.2995, "step": 6782 }, { "epoch": 0.23901265889690532, "grad_norm": 1.6057733297348022, "learning_rate": 1.7803808114250652e-05, "loss": 0.8923, "step": 6783 }, { "epoch": 0.23904789590986372, "grad_norm": 2.6621453762054443, "learning_rate": 1.780309441784401e-05, "loss": 3.0285, "step": 6784 }, { "epoch": 0.23908313292282213, "grad_norm": 1.3880399465560913, "learning_rate": 1.7802380619801068e-05, "loss": 0.9014, "step": 6785 }, { "epoch": 0.23911836993578053, "grad_norm": 4.562219619750977, "learning_rate": 1.7801666720131125e-05, "loss": 6.684, "step": 6786 }, { "epoch": 0.23915360694873897, "grad_norm": 8.464895248413086, "learning_rate": 1.780095271884348e-05, "loss": 6.929, "step": 6787 }, { "epoch": 0.23918884396169737, "grad_norm": 2.869190216064453, "learning_rate": 1.7800238615947428e-05, "loss": 3.3779, "step": 6788 }, { "epoch": 0.23922408097465578, "grad_norm": 1.3252032995224, "learning_rate": 1.779952441145228e-05, "loss": 0.9555, "step": 6789 }, { "epoch": 0.23925931798761418, "grad_norm": 8.581703186035156, "learning_rate": 1.779881010536733e-05, "loss": 6.331, "step": 6790 }, { "epoch": 0.2392945550005726, "grad_norm": 1.5600576400756836, "learning_rate": 1.7798095697701887e-05, "loss": 0.8542, "step": 6791 }, { "epoch": 0.23932979201353102, "grad_norm": 2.1716606616973877, "learning_rate": 1.7797381188465252e-05, "loss": 3.3218, "step": 6792 }, { "epoch": 0.23936502902648943, "grad_norm": 1.0934653282165527, "learning_rate": 1.7796666577666735e-05, "loss": 0.8254, "step": 6793 }, { "epoch": 0.23940026603944783, "grad_norm": 2.314560890197754, "learning_rate": 1.7795951865315645e-05, "loss": 3.0041, "step": 6794 }, { "epoch": 0.23943550305240624, "grad_norm": 3.2393336296081543, "learning_rate": 1.7795237051421286e-05, "loss": 3.277, "step": 6795 }, { "epoch": 0.23947074006536465, "grad_norm": 1.1731570959091187, "learning_rate": 1.779452213599297e-05, "loss": 1.0173, "step": 6796 }, { "epoch": 0.23950597707832308, "grad_norm": 7.669901371002197, "learning_rate": 1.7793807119040013e-05, "loss": 4.4433, "step": 6797 }, { "epoch": 0.23954121409128148, "grad_norm": 1.7614409923553467, "learning_rate": 1.7793092000571722e-05, "loss": 0.9349, "step": 6798 }, { "epoch": 0.2395764511042399, "grad_norm": 2.9912965297698975, "learning_rate": 1.779237678059742e-05, "loss": 3.2193, "step": 6799 }, { "epoch": 0.2396116881171983, "grad_norm": 4.79044246673584, "learning_rate": 1.7791661459126417e-05, "loss": 6.5338, "step": 6800 }, { "epoch": 0.23964692513015673, "grad_norm": 1.0366371870040894, "learning_rate": 1.7790946036168028e-05, "loss": 1.2769, "step": 6801 }, { "epoch": 0.23968216214311514, "grad_norm": 4.636332988739014, "learning_rate": 1.7790230511731577e-05, "loss": 5.8156, "step": 6802 }, { "epoch": 0.23971739915607354, "grad_norm": 5.81246280670166, "learning_rate": 1.7789514885826377e-05, "loss": 5.5201, "step": 6803 }, { "epoch": 0.23975263616903195, "grad_norm": 3.225705862045288, "learning_rate": 1.7788799158461757e-05, "loss": 3.0796, "step": 6804 }, { "epoch": 0.23978787318199035, "grad_norm": 5.417290210723877, "learning_rate": 1.7788083329647036e-05, "loss": 6.2269, "step": 6805 }, { "epoch": 0.23982311019494879, "grad_norm": 0.9850209951400757, "learning_rate": 1.778736739939154e-05, "loss": 1.1322, "step": 6806 }, { "epoch": 0.2398583472079072, "grad_norm": 2.44629168510437, "learning_rate": 1.7786651367704588e-05, "loss": 3.287, "step": 6807 }, { "epoch": 0.2398935842208656, "grad_norm": 1.1786941289901733, "learning_rate": 1.778593523459551e-05, "loss": 0.8718, "step": 6808 }, { "epoch": 0.239928821233824, "grad_norm": 3.1277101039886475, "learning_rate": 1.7785219000073633e-05, "loss": 4.0018, "step": 6809 }, { "epoch": 0.2399640582467824, "grad_norm": 3.1830544471740723, "learning_rate": 1.778450266414829e-05, "loss": 3.1671, "step": 6810 }, { "epoch": 0.23999929525974084, "grad_norm": 5.2632670402526855, "learning_rate": 1.778378622682881e-05, "loss": 3.7019, "step": 6811 }, { "epoch": 0.24003453227269925, "grad_norm": 2.9651591777801514, "learning_rate": 1.7783069688124516e-05, "loss": 3.1618, "step": 6812 }, { "epoch": 0.24006976928565765, "grad_norm": 1.2782129049301147, "learning_rate": 1.778235304804475e-05, "loss": 1.127, "step": 6813 }, { "epoch": 0.24010500629861606, "grad_norm": 1.5019707679748535, "learning_rate": 1.778163630659885e-05, "loss": 1.1901, "step": 6814 }, { "epoch": 0.2401402433115745, "grad_norm": 2.86500883102417, "learning_rate": 1.778091946379614e-05, "loss": 2.8418, "step": 6815 }, { "epoch": 0.2401754803245329, "grad_norm": 3.8034021854400635, "learning_rate": 1.7780202519645963e-05, "loss": 3.1383, "step": 6816 }, { "epoch": 0.2402107173374913, "grad_norm": 3.8513877391815186, "learning_rate": 1.777948547415766e-05, "loss": 3.4266, "step": 6817 }, { "epoch": 0.2402459543504497, "grad_norm": 2.8830134868621826, "learning_rate": 1.7778768327340564e-05, "loss": 3.3078, "step": 6818 }, { "epoch": 0.24028119136340811, "grad_norm": 10.317183494567871, "learning_rate": 1.7778051079204022e-05, "loss": 9.9825, "step": 6819 }, { "epoch": 0.24031642837636655, "grad_norm": 4.057264804840088, "learning_rate": 1.7777333729757377e-05, "loss": 3.5921, "step": 6820 }, { "epoch": 0.24035166538932495, "grad_norm": 2.421138048171997, "learning_rate": 1.7776616279009965e-05, "loss": 0.9404, "step": 6821 }, { "epoch": 0.24038690240228336, "grad_norm": 4.55916690826416, "learning_rate": 1.7775898726971138e-05, "loss": 3.1119, "step": 6822 }, { "epoch": 0.24042213941524176, "grad_norm": 1.4070360660552979, "learning_rate": 1.777518107365024e-05, "loss": 0.9895, "step": 6823 }, { "epoch": 0.24045737642820017, "grad_norm": 6.0498881340026855, "learning_rate": 1.7774463319056613e-05, "loss": 3.7402, "step": 6824 }, { "epoch": 0.2404926134411586, "grad_norm": 4.927470684051514, "learning_rate": 1.7773745463199617e-05, "loss": 3.5217, "step": 6825 }, { "epoch": 0.240527850454117, "grad_norm": 1.0424574613571167, "learning_rate": 1.7773027506088594e-05, "loss": 1.0877, "step": 6826 }, { "epoch": 0.24056308746707541, "grad_norm": 1.2584969997406006, "learning_rate": 1.7772309447732895e-05, "loss": 0.9681, "step": 6827 }, { "epoch": 0.24059832448003382, "grad_norm": 7.176041603088379, "learning_rate": 1.7771591288141882e-05, "loss": 6.1338, "step": 6828 }, { "epoch": 0.24063356149299223, "grad_norm": 4.245421409606934, "learning_rate": 1.7770873027324896e-05, "loss": 3.73, "step": 6829 }, { "epoch": 0.24066879850595066, "grad_norm": 1.1531956195831299, "learning_rate": 1.77701546652913e-05, "loss": 0.914, "step": 6830 }, { "epoch": 0.24070403551890907, "grad_norm": 1.5615607500076294, "learning_rate": 1.776943620205045e-05, "loss": 1.1984, "step": 6831 }, { "epoch": 0.24073927253186747, "grad_norm": 2.789339303970337, "learning_rate": 1.7768717637611706e-05, "loss": 3.2966, "step": 6832 }, { "epoch": 0.24077450954482588, "grad_norm": 4.931205749511719, "learning_rate": 1.7767998971984423e-05, "loss": 6.2058, "step": 6833 }, { "epoch": 0.2408097465577843, "grad_norm": 5.078200340270996, "learning_rate": 1.7767280205177967e-05, "loss": 3.7032, "step": 6834 }, { "epoch": 0.24084498357074272, "grad_norm": 1.2607486248016357, "learning_rate": 1.7766561337201692e-05, "loss": 0.9007, "step": 6835 }, { "epoch": 0.24088022058370112, "grad_norm": 1.3887343406677246, "learning_rate": 1.776584236806497e-05, "loss": 0.9823, "step": 6836 }, { "epoch": 0.24091545759665953, "grad_norm": 1.1635009050369263, "learning_rate": 1.7765123297777163e-05, "loss": 0.9479, "step": 6837 }, { "epoch": 0.24095069460961793, "grad_norm": 10.910747528076172, "learning_rate": 1.7764404126347632e-05, "loss": 3.8418, "step": 6838 }, { "epoch": 0.24098593162257637, "grad_norm": 1.7191309928894043, "learning_rate": 1.7763684853785748e-05, "loss": 0.9309, "step": 6839 }, { "epoch": 0.24102116863553477, "grad_norm": 1.9854172468185425, "learning_rate": 1.7762965480100883e-05, "loss": 0.886, "step": 6840 }, { "epoch": 0.24105640564849318, "grad_norm": 1.1686546802520752, "learning_rate": 1.7762246005302402e-05, "loss": 1.1027, "step": 6841 }, { "epoch": 0.24109164266145158, "grad_norm": 6.725341320037842, "learning_rate": 1.776152642939968e-05, "loss": 4.5438, "step": 6842 }, { "epoch": 0.24112687967441, "grad_norm": 6.424999713897705, "learning_rate": 1.7760806752402085e-05, "loss": 3.677, "step": 6843 }, { "epoch": 0.24116211668736842, "grad_norm": 9.207942008972168, "learning_rate": 1.7760086974318995e-05, "loss": 7.1099, "step": 6844 }, { "epoch": 0.24119735370032683, "grad_norm": 6.126710891723633, "learning_rate": 1.7759367095159785e-05, "loss": 2.9243, "step": 6845 }, { "epoch": 0.24123259071328523, "grad_norm": 1.1047945022583008, "learning_rate": 1.7758647114933828e-05, "loss": 1.0964, "step": 6846 }, { "epoch": 0.24126782772624364, "grad_norm": 3.653041362762451, "learning_rate": 1.7757927033650503e-05, "loss": 3.337, "step": 6847 }, { "epoch": 0.24130306473920204, "grad_norm": 1.660101294517517, "learning_rate": 1.7757206851319195e-05, "loss": 1.118, "step": 6848 }, { "epoch": 0.24133830175216048, "grad_norm": 7.277949810028076, "learning_rate": 1.7756486567949276e-05, "loss": 6.5517, "step": 6849 }, { "epoch": 0.24137353876511888, "grad_norm": 3.9007983207702637, "learning_rate": 1.7755766183550133e-05, "loss": 3.464, "step": 6850 }, { "epoch": 0.2414087757780773, "grad_norm": 1.7565011978149414, "learning_rate": 1.7755045698131143e-05, "loss": 0.9627, "step": 6851 }, { "epoch": 0.2414440127910357, "grad_norm": 1.5736231803894043, "learning_rate": 1.7754325111701703e-05, "loss": 0.9162, "step": 6852 }, { "epoch": 0.24147924980399413, "grad_norm": 1.6067014932632446, "learning_rate": 1.7753604424271183e-05, "loss": 0.7811, "step": 6853 }, { "epoch": 0.24151448681695253, "grad_norm": 4.781179428100586, "learning_rate": 1.775288363584898e-05, "loss": 5.1294, "step": 6854 }, { "epoch": 0.24154972382991094, "grad_norm": 4.530481815338135, "learning_rate": 1.7752162746444483e-05, "loss": 4.0474, "step": 6855 }, { "epoch": 0.24158496084286935, "grad_norm": 1.779180884361267, "learning_rate": 1.7751441756067077e-05, "loss": 0.9137, "step": 6856 }, { "epoch": 0.24162019785582775, "grad_norm": 9.924589157104492, "learning_rate": 1.7750720664726152e-05, "loss": 5.7293, "step": 6857 }, { "epoch": 0.24165543486878618, "grad_norm": 1.3402107954025269, "learning_rate": 1.7749999472431107e-05, "loss": 1.0227, "step": 6858 }, { "epoch": 0.2416906718817446, "grad_norm": 1.6043519973754883, "learning_rate": 1.774927817919133e-05, "loss": 0.7311, "step": 6859 }, { "epoch": 0.241725908894703, "grad_norm": 1.008535623550415, "learning_rate": 1.7748556785016218e-05, "loss": 1.1916, "step": 6860 }, { "epoch": 0.2417611459076614, "grad_norm": 3.8358705043792725, "learning_rate": 1.774783528991517e-05, "loss": 3.3031, "step": 6861 }, { "epoch": 0.2417963829206198, "grad_norm": 1.634294867515564, "learning_rate": 1.7747113693897577e-05, "loss": 0.8558, "step": 6862 }, { "epoch": 0.24183161993357824, "grad_norm": 4.9047322273254395, "learning_rate": 1.7746391996972838e-05, "loss": 6.0845, "step": 6863 }, { "epoch": 0.24186685694653665, "grad_norm": 4.437271595001221, "learning_rate": 1.774567019915036e-05, "loss": 2.6306, "step": 6864 }, { "epoch": 0.24190209395949505, "grad_norm": 4.133745193481445, "learning_rate": 1.7744948300439542e-05, "loss": 2.6877, "step": 6865 }, { "epoch": 0.24193733097245346, "grad_norm": 1.0535370111465454, "learning_rate": 1.7744226300849785e-05, "loss": 0.8191, "step": 6866 }, { "epoch": 0.2419725679854119, "grad_norm": 1.1274627447128296, "learning_rate": 1.7743504200390494e-05, "loss": 0.834, "step": 6867 }, { "epoch": 0.2420078049983703, "grad_norm": 3.2939651012420654, "learning_rate": 1.7742781999071072e-05, "loss": 3.9549, "step": 6868 }, { "epoch": 0.2420430420113287, "grad_norm": 1.8468235731124878, "learning_rate": 1.774205969690093e-05, "loss": 0.9242, "step": 6869 }, { "epoch": 0.2420782790242871, "grad_norm": 1.6109861135482788, "learning_rate": 1.7741337293889476e-05, "loss": 1.0784, "step": 6870 }, { "epoch": 0.2421135160372455, "grad_norm": 4.485990524291992, "learning_rate": 1.7740614790046118e-05, "loss": 3.2502, "step": 6871 }, { "epoch": 0.24214875305020395, "grad_norm": 1.34273362159729, "learning_rate": 1.7739892185380263e-05, "loss": 1.0111, "step": 6872 }, { "epoch": 0.24218399006316235, "grad_norm": 1.2136547565460205, "learning_rate": 1.773916947990133e-05, "loss": 1.0365, "step": 6873 }, { "epoch": 0.24221922707612076, "grad_norm": 4.914548873901367, "learning_rate": 1.7738446673618725e-05, "loss": 6.1218, "step": 6874 }, { "epoch": 0.24225446408907916, "grad_norm": 4.400508403778076, "learning_rate": 1.7737723766541867e-05, "loss": 3.4153, "step": 6875 }, { "epoch": 0.24228970110203757, "grad_norm": 1.5588010549545288, "learning_rate": 1.7737000758680174e-05, "loss": 1.0271, "step": 6876 }, { "epoch": 0.242324938114996, "grad_norm": 1.7378389835357666, "learning_rate": 1.773627765004306e-05, "loss": 1.014, "step": 6877 }, { "epoch": 0.2423601751279544, "grad_norm": 4.893517017364502, "learning_rate": 1.773555444063994e-05, "loss": 4.0402, "step": 6878 }, { "epoch": 0.2423954121409128, "grad_norm": 1.3843880891799927, "learning_rate": 1.7734831130480245e-05, "loss": 1.0033, "step": 6879 }, { "epoch": 0.24243064915387122, "grad_norm": 4.010061740875244, "learning_rate": 1.7734107719573384e-05, "loss": 3.1023, "step": 6880 }, { "epoch": 0.24246588616682963, "grad_norm": 3.8618264198303223, "learning_rate": 1.773338420792879e-05, "loss": 3.8566, "step": 6881 }, { "epoch": 0.24250112317978806, "grad_norm": 2.2758045196533203, "learning_rate": 1.7732660595555873e-05, "loss": 3.293, "step": 6882 }, { "epoch": 0.24253636019274646, "grad_norm": 4.197578430175781, "learning_rate": 1.7731936882464072e-05, "loss": 2.83, "step": 6883 }, { "epoch": 0.24257159720570487, "grad_norm": 4.904784202575684, "learning_rate": 1.7731213068662808e-05, "loss": 3.0516, "step": 6884 }, { "epoch": 0.24260683421866328, "grad_norm": 2.0717880725860596, "learning_rate": 1.7730489154161506e-05, "loss": 1.0988, "step": 6885 }, { "epoch": 0.2426420712316217, "grad_norm": 4.736367225646973, "learning_rate": 1.7729765138969604e-05, "loss": 6.4988, "step": 6886 }, { "epoch": 0.24267730824458011, "grad_norm": 3.8238110542297363, "learning_rate": 1.7729041023096526e-05, "loss": 3.5745, "step": 6887 }, { "epoch": 0.24271254525753852, "grad_norm": 5.777342319488525, "learning_rate": 1.77283168065517e-05, "loss": 6.0483, "step": 6888 }, { "epoch": 0.24274778227049693, "grad_norm": 1.99532949924469, "learning_rate": 1.7727592489344565e-05, "loss": 0.8193, "step": 6889 }, { "epoch": 0.24278301928345533, "grad_norm": 1.4463423490524292, "learning_rate": 1.772686807148455e-05, "loss": 0.7756, "step": 6890 }, { "epoch": 0.24281825629641376, "grad_norm": 1.3420287370681763, "learning_rate": 1.77261435529811e-05, "loss": 0.7736, "step": 6891 }, { "epoch": 0.24285349330937217, "grad_norm": 6.726180076599121, "learning_rate": 1.7725418933843644e-05, "loss": 3.5961, "step": 6892 }, { "epoch": 0.24288873032233058, "grad_norm": 3.6275155544281006, "learning_rate": 1.7724694214081618e-05, "loss": 3.4339, "step": 6893 }, { "epoch": 0.24292396733528898, "grad_norm": 3.6103720664978027, "learning_rate": 1.772396939370447e-05, "loss": 4.0734, "step": 6894 }, { "epoch": 0.2429592043482474, "grad_norm": 1.6281416416168213, "learning_rate": 1.7723244472721638e-05, "loss": 1.0372, "step": 6895 }, { "epoch": 0.24299444136120582, "grad_norm": 1.1087980270385742, "learning_rate": 1.7722519451142562e-05, "loss": 1.0803, "step": 6896 }, { "epoch": 0.24302967837416423, "grad_norm": 5.074366092681885, "learning_rate": 1.7721794328976684e-05, "loss": 3.4812, "step": 6897 }, { "epoch": 0.24306491538712263, "grad_norm": 1.9031668901443481, "learning_rate": 1.7721069106233457e-05, "loss": 1.2388, "step": 6898 }, { "epoch": 0.24310015240008104, "grad_norm": 2.2720396518707275, "learning_rate": 1.7720343782922316e-05, "loss": 0.9857, "step": 6899 }, { "epoch": 0.24313538941303944, "grad_norm": 1.6594058275222778, "learning_rate": 1.7719618359052714e-05, "loss": 1.0612, "step": 6900 }, { "epoch": 0.24317062642599788, "grad_norm": 3.0739047527313232, "learning_rate": 1.77188928346341e-05, "loss": 4.1604, "step": 6901 }, { "epoch": 0.24320586343895628, "grad_norm": 2.2162625789642334, "learning_rate": 1.7718167209675924e-05, "loss": 0.8106, "step": 6902 }, { "epoch": 0.2432411004519147, "grad_norm": 1.412376046180725, "learning_rate": 1.7717441484187637e-05, "loss": 0.9978, "step": 6903 }, { "epoch": 0.2432763374648731, "grad_norm": 6.211417198181152, "learning_rate": 1.7716715658178695e-05, "loss": 3.9742, "step": 6904 }, { "epoch": 0.24331157447783153, "grad_norm": 7.514272689819336, "learning_rate": 1.7715989731658543e-05, "loss": 5.2316, "step": 6905 }, { "epoch": 0.24334681149078993, "grad_norm": 9.511852264404297, "learning_rate": 1.7715263704636644e-05, "loss": 8.7064, "step": 6906 }, { "epoch": 0.24338204850374834, "grad_norm": 6.5855607986450195, "learning_rate": 1.7714537577122455e-05, "loss": 6.2824, "step": 6907 }, { "epoch": 0.24341728551670674, "grad_norm": 6.666676044464111, "learning_rate": 1.771381134912543e-05, "loss": 5.686, "step": 6908 }, { "epoch": 0.24345252252966515, "grad_norm": 6.621350288391113, "learning_rate": 1.7713085020655028e-05, "loss": 3.3937, "step": 6909 }, { "epoch": 0.24348775954262358, "grad_norm": 4.937229156494141, "learning_rate": 1.7712358591720712e-05, "loss": 5.3121, "step": 6910 }, { "epoch": 0.243522996555582, "grad_norm": 1.69504714012146, "learning_rate": 1.7711632062331944e-05, "loss": 1.0592, "step": 6911 }, { "epoch": 0.2435582335685404, "grad_norm": 1.3381328582763672, "learning_rate": 1.771090543249818e-05, "loss": 0.9443, "step": 6912 }, { "epoch": 0.2435934705814988, "grad_norm": 1.1971861124038696, "learning_rate": 1.77101787022289e-05, "loss": 0.8493, "step": 6913 }, { "epoch": 0.2436287075944572, "grad_norm": 4.286947250366211, "learning_rate": 1.7709451871533555e-05, "loss": 2.5373, "step": 6914 }, { "epoch": 0.24366394460741564, "grad_norm": 1.2163180112838745, "learning_rate": 1.7708724940421617e-05, "loss": 0.7858, "step": 6915 }, { "epoch": 0.24369918162037404, "grad_norm": 3.7225632667541504, "learning_rate": 1.7707997908902557e-05, "loss": 3.3471, "step": 6916 }, { "epoch": 0.24373441863333245, "grad_norm": 1.8096299171447754, "learning_rate": 1.770727077698584e-05, "loss": 0.9566, "step": 6917 }, { "epoch": 0.24376965564629086, "grad_norm": 4.83967924118042, "learning_rate": 1.7706543544680944e-05, "loss": 3.3402, "step": 6918 }, { "epoch": 0.24380489265924926, "grad_norm": 1.470017910003662, "learning_rate": 1.770581621199733e-05, "loss": 0.9293, "step": 6919 }, { "epoch": 0.2438401296722077, "grad_norm": 1.6306416988372803, "learning_rate": 1.7705088778944483e-05, "loss": 0.7212, "step": 6920 }, { "epoch": 0.2438753666851661, "grad_norm": 1.3790712356567383, "learning_rate": 1.7704361245531873e-05, "loss": 0.9074, "step": 6921 }, { "epoch": 0.2439106036981245, "grad_norm": 10.943537712097168, "learning_rate": 1.7703633611768975e-05, "loss": 3.8603, "step": 6922 }, { "epoch": 0.2439458407110829, "grad_norm": 4.654026508331299, "learning_rate": 1.7702905877665267e-05, "loss": 6.0298, "step": 6923 }, { "epoch": 0.24398107772404135, "grad_norm": 4.32839822769165, "learning_rate": 1.770217804323023e-05, "loss": 3.5603, "step": 6924 }, { "epoch": 0.24401631473699975, "grad_norm": 8.611961364746094, "learning_rate": 1.7701450108473343e-05, "loss": 3.5034, "step": 6925 }, { "epoch": 0.24405155174995816, "grad_norm": 4.110537052154541, "learning_rate": 1.7700722073404087e-05, "loss": 3.6479, "step": 6926 }, { "epoch": 0.24408678876291656, "grad_norm": 1.5232549905776978, "learning_rate": 1.7699993938031944e-05, "loss": 0.8218, "step": 6927 }, { "epoch": 0.24412202577587497, "grad_norm": 1.7231038808822632, "learning_rate": 1.76992657023664e-05, "loss": 1.0612, "step": 6928 }, { "epoch": 0.2441572627888334, "grad_norm": 7.352329730987549, "learning_rate": 1.769853736641694e-05, "loss": 3.6514, "step": 6929 }, { "epoch": 0.2441924998017918, "grad_norm": 4.168332099914551, "learning_rate": 1.7697808930193053e-05, "loss": 3.0947, "step": 6930 }, { "epoch": 0.2442277368147502, "grad_norm": 2.112154483795166, "learning_rate": 1.7697080393704216e-05, "loss": 1.0363, "step": 6931 }, { "epoch": 0.24426297382770862, "grad_norm": 1.9067054986953735, "learning_rate": 1.7696351756959936e-05, "loss": 3.5893, "step": 6932 }, { "epoch": 0.24429821084066702, "grad_norm": 3.6830153465270996, "learning_rate": 1.769562301996969e-05, "loss": 3.6498, "step": 6933 }, { "epoch": 0.24433344785362546, "grad_norm": 1.5738576650619507, "learning_rate": 1.7694894182742973e-05, "loss": 1.1355, "step": 6934 }, { "epoch": 0.24436868486658386, "grad_norm": 3.2428653240203857, "learning_rate": 1.7694165245289278e-05, "loss": 3.0259, "step": 6935 }, { "epoch": 0.24440392187954227, "grad_norm": 4.479694843292236, "learning_rate": 1.7693436207618103e-05, "loss": 3.1274, "step": 6936 }, { "epoch": 0.24443915889250067, "grad_norm": 4.806639671325684, "learning_rate": 1.769270706973894e-05, "loss": 3.7201, "step": 6937 }, { "epoch": 0.2444743959054591, "grad_norm": 1.2960033416748047, "learning_rate": 1.7691977831661288e-05, "loss": 0.6487, "step": 6938 }, { "epoch": 0.2445096329184175, "grad_norm": 1.9331516027450562, "learning_rate": 1.7691248493394644e-05, "loss": 0.7534, "step": 6939 }, { "epoch": 0.24454486993137592, "grad_norm": 1.8932043313980103, "learning_rate": 1.7690519054948513e-05, "loss": 0.919, "step": 6940 }, { "epoch": 0.24458010694433432, "grad_norm": 7.783337593078613, "learning_rate": 1.7689789516332387e-05, "loss": 3.6735, "step": 6941 }, { "epoch": 0.24461534395729273, "grad_norm": 7.544166088104248, "learning_rate": 1.7689059877555774e-05, "loss": 3.2499, "step": 6942 }, { "epoch": 0.24465058097025116, "grad_norm": 1.7657105922698975, "learning_rate": 1.7688330138628177e-05, "loss": 0.9181, "step": 6943 }, { "epoch": 0.24468581798320957, "grad_norm": 5.168910026550293, "learning_rate": 1.76876002995591e-05, "loss": 5.7278, "step": 6944 }, { "epoch": 0.24472105499616797, "grad_norm": 1.8272615671157837, "learning_rate": 1.768687036035805e-05, "loss": 0.9273, "step": 6945 }, { "epoch": 0.24475629200912638, "grad_norm": 1.3763149976730347, "learning_rate": 1.7686140321034536e-05, "loss": 1.122, "step": 6946 }, { "epoch": 0.2447915290220848, "grad_norm": 1.3011316061019897, "learning_rate": 1.7685410181598063e-05, "loss": 1.232, "step": 6947 }, { "epoch": 0.24482676603504322, "grad_norm": 8.628157615661621, "learning_rate": 1.7684679942058144e-05, "loss": 8.0463, "step": 6948 }, { "epoch": 0.24486200304800163, "grad_norm": 16.531400680541992, "learning_rate": 1.7683949602424292e-05, "loss": 6.8449, "step": 6949 }, { "epoch": 0.24489724006096003, "grad_norm": 7.16672945022583, "learning_rate": 1.7683219162706016e-05, "loss": 5.5543, "step": 6950 }, { "epoch": 0.24493247707391844, "grad_norm": 5.51406192779541, "learning_rate": 1.7682488622912833e-05, "loss": 4.3534, "step": 6951 }, { "epoch": 0.24496771408687684, "grad_norm": 4.6006669998168945, "learning_rate": 1.7681757983054255e-05, "loss": 3.1315, "step": 6952 }, { "epoch": 0.24500295109983528, "grad_norm": 7.700827121734619, "learning_rate": 1.7681027243139804e-05, "loss": 6.4722, "step": 6953 }, { "epoch": 0.24503818811279368, "grad_norm": 1.0554280281066895, "learning_rate": 1.7680296403178992e-05, "loss": 0.9498, "step": 6954 }, { "epoch": 0.2450734251257521, "grad_norm": 0.9056572318077087, "learning_rate": 1.767956546318134e-05, "loss": 0.9893, "step": 6955 }, { "epoch": 0.2451086621387105, "grad_norm": 3.70382022857666, "learning_rate": 1.767883442315637e-05, "loss": 5.0253, "step": 6956 }, { "epoch": 0.24514389915166893, "grad_norm": 3.7431204319000244, "learning_rate": 1.7678103283113604e-05, "loss": 5.9655, "step": 6957 }, { "epoch": 0.24517913616462733, "grad_norm": 1.6630860567092896, "learning_rate": 1.7677372043062567e-05, "loss": 0.7817, "step": 6958 }, { "epoch": 0.24521437317758574, "grad_norm": 3.5126876831054688, "learning_rate": 1.767664070301278e-05, "loss": 5.9704, "step": 6959 }, { "epoch": 0.24524961019054414, "grad_norm": 2.2465944290161133, "learning_rate": 1.7675909262973768e-05, "loss": 1.1867, "step": 6960 }, { "epoch": 0.24528484720350255, "grad_norm": 2.3694064617156982, "learning_rate": 1.7675177722955062e-05, "loss": 3.7748, "step": 6961 }, { "epoch": 0.24532008421646098, "grad_norm": 1.2433335781097412, "learning_rate": 1.7674446082966187e-05, "loss": 0.9003, "step": 6962 }, { "epoch": 0.2453553212294194, "grad_norm": 1.1830612421035767, "learning_rate": 1.7673714343016676e-05, "loss": 0.9752, "step": 6963 }, { "epoch": 0.2453905582423778, "grad_norm": 1.126796007156372, "learning_rate": 1.767298250311606e-05, "loss": 1.0018, "step": 6964 }, { "epoch": 0.2454257952553362, "grad_norm": 3.043699264526367, "learning_rate": 1.7672250563273864e-05, "loss": 3.2529, "step": 6965 }, { "epoch": 0.2454610322682946, "grad_norm": 1.6242797374725342, "learning_rate": 1.767151852349963e-05, "loss": 1.0373, "step": 6966 }, { "epoch": 0.24549626928125304, "grad_norm": 4.345557689666748, "learning_rate": 1.7670786383802892e-05, "loss": 3.0719, "step": 6967 }, { "epoch": 0.24553150629421144, "grad_norm": 1.897499442100525, "learning_rate": 1.7670054144193183e-05, "loss": 1.0072, "step": 6968 }, { "epoch": 0.24556674330716985, "grad_norm": 1.0558159351348877, "learning_rate": 1.766932180468004e-05, "loss": 0.9838, "step": 6969 }, { "epoch": 0.24560198032012825, "grad_norm": 1.3889859914779663, "learning_rate": 1.7668589365273003e-05, "loss": 1.285, "step": 6970 }, { "epoch": 0.24563721733308666, "grad_norm": 3.726627826690674, "learning_rate": 1.7667856825981617e-05, "loss": 3.5571, "step": 6971 }, { "epoch": 0.2456724543460451, "grad_norm": 1.1336783170700073, "learning_rate": 1.7667124186815415e-05, "loss": 1.1489, "step": 6972 }, { "epoch": 0.2457076913590035, "grad_norm": 2.7888364791870117, "learning_rate": 1.7666391447783947e-05, "loss": 3.8325, "step": 6973 }, { "epoch": 0.2457429283719619, "grad_norm": 3.510122060775757, "learning_rate": 1.766565860889675e-05, "loss": 3.2568, "step": 6974 }, { "epoch": 0.2457781653849203, "grad_norm": 2.355766534805298, "learning_rate": 1.7664925670163377e-05, "loss": 3.0009, "step": 6975 }, { "epoch": 0.24581340239787874, "grad_norm": 6.849942207336426, "learning_rate": 1.766419263159337e-05, "loss": 3.8423, "step": 6976 }, { "epoch": 0.24584863941083715, "grad_norm": 1.0199034214019775, "learning_rate": 1.766345949319628e-05, "loss": 0.9121, "step": 6977 }, { "epoch": 0.24588387642379556, "grad_norm": 1.4508544206619263, "learning_rate": 1.7662726254981648e-05, "loss": 1.1233, "step": 6978 }, { "epoch": 0.24591911343675396, "grad_norm": 1.1172610521316528, "learning_rate": 1.7661992916959033e-05, "loss": 0.795, "step": 6979 }, { "epoch": 0.24595435044971237, "grad_norm": 2.860344409942627, "learning_rate": 1.7661259479137984e-05, "loss": 2.8451, "step": 6980 }, { "epoch": 0.2459895874626708, "grad_norm": 3.2233500480651855, "learning_rate": 1.7660525941528054e-05, "loss": 5.1275, "step": 6981 }, { "epoch": 0.2460248244756292, "grad_norm": 6.013513088226318, "learning_rate": 1.76597923041388e-05, "loss": 9.301, "step": 6982 }, { "epoch": 0.2460600614885876, "grad_norm": 0.8032238483428955, "learning_rate": 1.7659058566979776e-05, "loss": 0.9667, "step": 6983 }, { "epoch": 0.24609529850154602, "grad_norm": 1.2269667387008667, "learning_rate": 1.7658324730060535e-05, "loss": 0.8601, "step": 6984 }, { "epoch": 0.24613053551450442, "grad_norm": 4.64187479019165, "learning_rate": 1.765759079339064e-05, "loss": 5.6408, "step": 6985 }, { "epoch": 0.24616577252746286, "grad_norm": 1.355959415435791, "learning_rate": 1.7656856756979653e-05, "loss": 0.9244, "step": 6986 }, { "epoch": 0.24620100954042126, "grad_norm": 1.7274333238601685, "learning_rate": 1.7656122620837127e-05, "loss": 1.1519, "step": 6987 }, { "epoch": 0.24623624655337967, "grad_norm": 1.695468783378601, "learning_rate": 1.765538838497263e-05, "loss": 1.0063, "step": 6988 }, { "epoch": 0.24627148356633807, "grad_norm": 2.1756975650787354, "learning_rate": 1.7654654049395726e-05, "loss": 0.8162, "step": 6989 }, { "epoch": 0.2463067205792965, "grad_norm": 6.885067462921143, "learning_rate": 1.7653919614115976e-05, "loss": 8.9683, "step": 6990 }, { "epoch": 0.2463419575922549, "grad_norm": 1.2974390983581543, "learning_rate": 1.7653185079142948e-05, "loss": 1.0349, "step": 6991 }, { "epoch": 0.24637719460521332, "grad_norm": 2.4663538932800293, "learning_rate": 1.765245044448621e-05, "loss": 1.094, "step": 6992 }, { "epoch": 0.24641243161817172, "grad_norm": 1.4720686674118042, "learning_rate": 1.7651715710155327e-05, "loss": 0.9132, "step": 6993 }, { "epoch": 0.24644766863113013, "grad_norm": 3.376011848449707, "learning_rate": 1.765098087615988e-05, "loss": 3.3038, "step": 6994 }, { "epoch": 0.24648290564408856, "grad_norm": 2.8055336475372314, "learning_rate": 1.7650245942509426e-05, "loss": 4.2371, "step": 6995 }, { "epoch": 0.24651814265704697, "grad_norm": 6.493950843811035, "learning_rate": 1.7649510909213546e-05, "loss": 6.5498, "step": 6996 }, { "epoch": 0.24655337967000537, "grad_norm": 2.526865005493164, "learning_rate": 1.764877577628181e-05, "loss": 1.0234, "step": 6997 }, { "epoch": 0.24658861668296378, "grad_norm": 1.1454938650131226, "learning_rate": 1.76480405437238e-05, "loss": 1.1632, "step": 6998 }, { "epoch": 0.24662385369592218, "grad_norm": 3.6232657432556152, "learning_rate": 1.7647305211549084e-05, "loss": 3.7191, "step": 6999 }, { "epoch": 0.24665909070888062, "grad_norm": 0.9691055417060852, "learning_rate": 1.7646569779767245e-05, "loss": 0.7862, "step": 7000 }, { "epoch": 0.24669432772183902, "grad_norm": 3.7812860012054443, "learning_rate": 1.764583424838786e-05, "loss": 3.0476, "step": 7001 }, { "epoch": 0.24672956473479743, "grad_norm": 3.587878704071045, "learning_rate": 1.7645098617420512e-05, "loss": 3.6485, "step": 7002 }, { "epoch": 0.24676480174775584, "grad_norm": 5.309279441833496, "learning_rate": 1.7644362886874774e-05, "loss": 3.9059, "step": 7003 }, { "epoch": 0.24680003876071424, "grad_norm": 4.205740451812744, "learning_rate": 1.7643627056760242e-05, "loss": 3.1117, "step": 7004 }, { "epoch": 0.24683527577367267, "grad_norm": 3.7810497283935547, "learning_rate": 1.7642891127086495e-05, "loss": 4.3542, "step": 7005 }, { "epoch": 0.24687051278663108, "grad_norm": 2.876216411590576, "learning_rate": 1.7642155097863112e-05, "loss": 3.6633, "step": 7006 }, { "epoch": 0.24690574979958949, "grad_norm": 0.9519078731536865, "learning_rate": 1.764141896909969e-05, "loss": 1.1691, "step": 7007 }, { "epoch": 0.2469409868125479, "grad_norm": 2.6826412677764893, "learning_rate": 1.764068274080581e-05, "loss": 3.4872, "step": 7008 }, { "epoch": 0.24697622382550632, "grad_norm": 1.4912736415863037, "learning_rate": 1.7639946412991065e-05, "loss": 1.2035, "step": 7009 }, { "epoch": 0.24701146083846473, "grad_norm": 1.2781838178634644, "learning_rate": 1.7639209985665044e-05, "loss": 1.0834, "step": 7010 }, { "epoch": 0.24704669785142314, "grad_norm": 1.9417182207107544, "learning_rate": 1.763847345883734e-05, "loss": 0.993, "step": 7011 }, { "epoch": 0.24708193486438154, "grad_norm": 2.448275089263916, "learning_rate": 1.7637736832517548e-05, "loss": 0.913, "step": 7012 }, { "epoch": 0.24711717187733995, "grad_norm": 2.0390985012054443, "learning_rate": 1.763700010671526e-05, "loss": 3.5406, "step": 7013 }, { "epoch": 0.24715240889029838, "grad_norm": 1.178244948387146, "learning_rate": 1.763626328144007e-05, "loss": 1.0147, "step": 7014 }, { "epoch": 0.2471876459032568, "grad_norm": 4.10121488571167, "learning_rate": 1.763552635670158e-05, "loss": 6.0512, "step": 7015 }, { "epoch": 0.2472228829162152, "grad_norm": 2.009018898010254, "learning_rate": 1.763478933250939e-05, "loss": 3.8063, "step": 7016 }, { "epoch": 0.2472581199291736, "grad_norm": 1.3948804140090942, "learning_rate": 1.7634052208873093e-05, "loss": 0.904, "step": 7017 }, { "epoch": 0.247293356942132, "grad_norm": 5.031387805938721, "learning_rate": 1.7633314985802295e-05, "loss": 4.4706, "step": 7018 }, { "epoch": 0.24732859395509044, "grad_norm": 1.3276517391204834, "learning_rate": 1.7632577663306593e-05, "loss": 0.8127, "step": 7019 }, { "epoch": 0.24736383096804884, "grad_norm": 3.2398481369018555, "learning_rate": 1.76318402413956e-05, "loss": 3.6723, "step": 7020 }, { "epoch": 0.24739906798100725, "grad_norm": 3.0600080490112305, "learning_rate": 1.763110272007891e-05, "loss": 3.5185, "step": 7021 }, { "epoch": 0.24743430499396565, "grad_norm": 4.128720283508301, "learning_rate": 1.763036509936614e-05, "loss": 3.6985, "step": 7022 }, { "epoch": 0.24746954200692406, "grad_norm": 1.0498758554458618, "learning_rate": 1.762962737926689e-05, "loss": 0.9967, "step": 7023 }, { "epoch": 0.2475047790198825, "grad_norm": 1.3027215003967285, "learning_rate": 1.7628889559790776e-05, "loss": 0.8441, "step": 7024 }, { "epoch": 0.2475400160328409, "grad_norm": 1.2723709344863892, "learning_rate": 1.76281516409474e-05, "loss": 1.1773, "step": 7025 }, { "epoch": 0.2475752530457993, "grad_norm": 2.9628970623016357, "learning_rate": 1.7627413622746377e-05, "loss": 3.8781, "step": 7026 }, { "epoch": 0.2476104900587577, "grad_norm": 1.5034689903259277, "learning_rate": 1.762667550519732e-05, "loss": 0.9525, "step": 7027 }, { "epoch": 0.24764572707171614, "grad_norm": 1.3081682920455933, "learning_rate": 1.7625937288309844e-05, "loss": 0.7748, "step": 7028 }, { "epoch": 0.24768096408467455, "grad_norm": 3.7882492542266846, "learning_rate": 1.7625198972093564e-05, "loss": 5.9805, "step": 7029 }, { "epoch": 0.24771620109763295, "grad_norm": 1.2459787130355835, "learning_rate": 1.7624460556558093e-05, "loss": 0.9485, "step": 7030 }, { "epoch": 0.24775143811059136, "grad_norm": 4.3871636390686035, "learning_rate": 1.7623722041713055e-05, "loss": 5.4108, "step": 7031 }, { "epoch": 0.24778667512354977, "grad_norm": 1.0561238527297974, "learning_rate": 1.7622983427568067e-05, "loss": 1.0371, "step": 7032 }, { "epoch": 0.2478219121365082, "grad_norm": 3.3762571811676025, "learning_rate": 1.7622244714132745e-05, "loss": 3.7006, "step": 7033 }, { "epoch": 0.2478571491494666, "grad_norm": 3.581312894821167, "learning_rate": 1.762150590141672e-05, "loss": 2.7756, "step": 7034 }, { "epoch": 0.247892386162425, "grad_norm": 2.7395248413085938, "learning_rate": 1.7620766989429604e-05, "loss": 2.9482, "step": 7035 }, { "epoch": 0.24792762317538342, "grad_norm": 1.710313081741333, "learning_rate": 1.762002797818103e-05, "loss": 0.9726, "step": 7036 }, { "epoch": 0.24796286018834182, "grad_norm": 5.528052806854248, "learning_rate": 1.7619288867680622e-05, "loss": 3.4434, "step": 7037 }, { "epoch": 0.24799809720130025, "grad_norm": 1.2091385126113892, "learning_rate": 1.7618549657938003e-05, "loss": 0.888, "step": 7038 }, { "epoch": 0.24803333421425866, "grad_norm": 1.4596805572509766, "learning_rate": 1.761781034896281e-05, "loss": 0.7422, "step": 7039 }, { "epoch": 0.24806857122721707, "grad_norm": 3.380855083465576, "learning_rate": 1.7617070940764664e-05, "loss": 3.727, "step": 7040 }, { "epoch": 0.24810380824017547, "grad_norm": 1.1988581418991089, "learning_rate": 1.7616331433353197e-05, "loss": 0.9103, "step": 7041 }, { "epoch": 0.2481390452531339, "grad_norm": 2.1381492614746094, "learning_rate": 1.761559182673804e-05, "loss": 0.963, "step": 7042 }, { "epoch": 0.2481742822660923, "grad_norm": 2.4814517498016357, "learning_rate": 1.7614852120928832e-05, "loss": 1.0285, "step": 7043 }, { "epoch": 0.24820951927905072, "grad_norm": 6.559317588806152, "learning_rate": 1.761411231593521e-05, "loss": 5.0987, "step": 7044 }, { "epoch": 0.24824475629200912, "grad_norm": 1.6447120904922485, "learning_rate": 1.76133724117668e-05, "loss": 0.7973, "step": 7045 }, { "epoch": 0.24827999330496753, "grad_norm": 4.519289493560791, "learning_rate": 1.7612632408433245e-05, "loss": 5.7203, "step": 7046 }, { "epoch": 0.24831523031792596, "grad_norm": 6.176080703735352, "learning_rate": 1.7611892305944182e-05, "loss": 3.7716, "step": 7047 }, { "epoch": 0.24835046733088437, "grad_norm": 11.51388931274414, "learning_rate": 1.761115210430926e-05, "loss": 8.6108, "step": 7048 }, { "epoch": 0.24838570434384277, "grad_norm": 4.417227268218994, "learning_rate": 1.7610411803538104e-05, "loss": 6.3157, "step": 7049 }, { "epoch": 0.24842094135680118, "grad_norm": 1.540253758430481, "learning_rate": 1.7609671403640364e-05, "loss": 1.0768, "step": 7050 }, { "epoch": 0.24845617836975958, "grad_norm": 5.550596714019775, "learning_rate": 1.760893090462569e-05, "loss": 3.2612, "step": 7051 }, { "epoch": 0.24849141538271802, "grad_norm": 1.7058502435684204, "learning_rate": 1.760819030650372e-05, "loss": 1.0833, "step": 7052 }, { "epoch": 0.24852665239567642, "grad_norm": 1.3211696147918701, "learning_rate": 1.76074496092841e-05, "loss": 0.9385, "step": 7053 }, { "epoch": 0.24856188940863483, "grad_norm": 6.029996871948242, "learning_rate": 1.7606708812976484e-05, "loss": 3.7285, "step": 7054 }, { "epoch": 0.24859712642159323, "grad_norm": 4.890561103820801, "learning_rate": 1.7605967917590514e-05, "loss": 2.858, "step": 7055 }, { "epoch": 0.24863236343455164, "grad_norm": 2.0927257537841797, "learning_rate": 1.7605226923135842e-05, "loss": 2.8454, "step": 7056 }, { "epoch": 0.24866760044751007, "grad_norm": 2.6247336864471436, "learning_rate": 1.7604485829622124e-05, "loss": 4.9646, "step": 7057 }, { "epoch": 0.24870283746046848, "grad_norm": 1.2013039588928223, "learning_rate": 1.7603744637059005e-05, "loss": 0.8824, "step": 7058 }, { "epoch": 0.24873807447342688, "grad_norm": 3.42423152923584, "learning_rate": 1.760300334545615e-05, "loss": 3.4471, "step": 7059 }, { "epoch": 0.2487733114863853, "grad_norm": 2.2011091709136963, "learning_rate": 1.76022619548232e-05, "loss": 0.8676, "step": 7060 }, { "epoch": 0.24880854849934372, "grad_norm": 3.0268726348876953, "learning_rate": 1.7601520465169826e-05, "loss": 3.6136, "step": 7061 }, { "epoch": 0.24884378551230213, "grad_norm": 2.3122026920318604, "learning_rate": 1.7600778876505678e-05, "loss": 3.2933, "step": 7062 }, { "epoch": 0.24887902252526053, "grad_norm": 3.7886335849761963, "learning_rate": 1.7600037188840414e-05, "loss": 3.256, "step": 7063 }, { "epoch": 0.24891425953821894, "grad_norm": 3.7342581748962402, "learning_rate": 1.7599295402183704e-05, "loss": 3.4043, "step": 7064 }, { "epoch": 0.24894949655117735, "grad_norm": 3.044617176055908, "learning_rate": 1.7598553516545197e-05, "loss": 3.9735, "step": 7065 }, { "epoch": 0.24898473356413578, "grad_norm": 1.4700111150741577, "learning_rate": 1.7597811531934566e-05, "loss": 0.8996, "step": 7066 }, { "epoch": 0.24901997057709419, "grad_norm": 5.46886682510376, "learning_rate": 1.7597069448361468e-05, "loss": 5.9643, "step": 7067 }, { "epoch": 0.2490552075900526, "grad_norm": 1.7959221601486206, "learning_rate": 1.7596327265835578e-05, "loss": 1.0025, "step": 7068 }, { "epoch": 0.249090444603011, "grad_norm": 4.811557769775391, "learning_rate": 1.7595584984366554e-05, "loss": 6.572, "step": 7069 }, { "epoch": 0.2491256816159694, "grad_norm": 1.2110871076583862, "learning_rate": 1.7594842603964068e-05, "loss": 0.8709, "step": 7070 }, { "epoch": 0.24916091862892784, "grad_norm": 1.350862741470337, "learning_rate": 1.7594100124637793e-05, "loss": 0.92, "step": 7071 }, { "epoch": 0.24919615564188624, "grad_norm": 2.2816879749298096, "learning_rate": 1.7593357546397392e-05, "loss": 0.921, "step": 7072 }, { "epoch": 0.24923139265484465, "grad_norm": 4.8680925369262695, "learning_rate": 1.7592614869252546e-05, "loss": 2.8703, "step": 7073 }, { "epoch": 0.24926662966780305, "grad_norm": 1.3862813711166382, "learning_rate": 1.7591872093212923e-05, "loss": 0.9063, "step": 7074 }, { "epoch": 0.24930186668076146, "grad_norm": 9.252367973327637, "learning_rate": 1.7591129218288198e-05, "loss": 3.8513, "step": 7075 }, { "epoch": 0.2493371036937199, "grad_norm": 1.8122164011001587, "learning_rate": 1.7590386244488046e-05, "loss": 0.9225, "step": 7076 }, { "epoch": 0.2493723407066783, "grad_norm": 2.5572288036346436, "learning_rate": 1.758964317182215e-05, "loss": 3.0954, "step": 7077 }, { "epoch": 0.2494075777196367, "grad_norm": 3.4509708881378174, "learning_rate": 1.7588900000300185e-05, "loss": 3.2901, "step": 7078 }, { "epoch": 0.2494428147325951, "grad_norm": 1.3066775798797607, "learning_rate": 1.758815672993183e-05, "loss": 1.3428, "step": 7079 }, { "epoch": 0.24947805174555354, "grad_norm": 1.4273104667663574, "learning_rate": 1.7587413360726768e-05, "loss": 0.9114, "step": 7080 }, { "epoch": 0.24951328875851195, "grad_norm": 3.5933165550231934, "learning_rate": 1.7586669892694678e-05, "loss": 4.0671, "step": 7081 }, { "epoch": 0.24954852577147035, "grad_norm": 1.231483817100525, "learning_rate": 1.7585926325845245e-05, "loss": 0.9919, "step": 7082 }, { "epoch": 0.24958376278442876, "grad_norm": 4.655564785003662, "learning_rate": 1.7585182660188155e-05, "loss": 3.3985, "step": 7083 }, { "epoch": 0.24961899979738716, "grad_norm": 1.3627560138702393, "learning_rate": 1.75844388957331e-05, "loss": 0.792, "step": 7084 }, { "epoch": 0.2496542368103456, "grad_norm": 3.623138666152954, "learning_rate": 1.7583695032489757e-05, "loss": 4.0094, "step": 7085 }, { "epoch": 0.249689473823304, "grad_norm": 5.233938217163086, "learning_rate": 1.7582951070467825e-05, "loss": 6.7825, "step": 7086 }, { "epoch": 0.2497247108362624, "grad_norm": 1.0971957445144653, "learning_rate": 1.7582207009676987e-05, "loss": 1.2616, "step": 7087 }, { "epoch": 0.24975994784922081, "grad_norm": 0.9716677069664001, "learning_rate": 1.7581462850126935e-05, "loss": 1.034, "step": 7088 }, { "epoch": 0.24979518486217922, "grad_norm": 3.954808473587036, "learning_rate": 1.7580718591827367e-05, "loss": 3.4876, "step": 7089 }, { "epoch": 0.24983042187513765, "grad_norm": 4.159263610839844, "learning_rate": 1.757997423478797e-05, "loss": 5.4491, "step": 7090 }, { "epoch": 0.24986565888809606, "grad_norm": 6.271335124969482, "learning_rate": 1.7579229779018447e-05, "loss": 3.7604, "step": 7091 }, { "epoch": 0.24990089590105447, "grad_norm": 9.217206001281738, "learning_rate": 1.757848522452849e-05, "loss": 12.2303, "step": 7092 }, { "epoch": 0.24993613291401287, "grad_norm": 3.581193208694458, "learning_rate": 1.7577740571327794e-05, "loss": 2.4572, "step": 7093 }, { "epoch": 0.2499713699269713, "grad_norm": 6.21895694732666, "learning_rate": 1.7576995819426068e-05, "loss": 3.7591, "step": 7094 }, { "epoch": 0.2500066069399297, "grad_norm": 1.0609149932861328, "learning_rate": 1.7576250968832998e-05, "loss": 0.7971, "step": 7095 }, { "epoch": 0.2500418439528881, "grad_norm": 1.2203125953674316, "learning_rate": 1.75755060195583e-05, "loss": 1.0634, "step": 7096 }, { "epoch": 0.2500770809658465, "grad_norm": 3.2740633487701416, "learning_rate": 1.7574760971611667e-05, "loss": 3.704, "step": 7097 }, { "epoch": 0.25011231797880495, "grad_norm": 3.934623956680298, "learning_rate": 1.7574015825002812e-05, "loss": 5.7619, "step": 7098 }, { "epoch": 0.25014755499176333, "grad_norm": 1.5530375242233276, "learning_rate": 1.7573270579741436e-05, "loss": 1.0285, "step": 7099 }, { "epoch": 0.25018279200472177, "grad_norm": 1.9492413997650146, "learning_rate": 1.7572525235837242e-05, "loss": 0.8054, "step": 7100 }, { "epoch": 0.25021802901768014, "grad_norm": 6.949281692504883, "learning_rate": 1.7571779793299948e-05, "loss": 3.2005, "step": 7101 }, { "epoch": 0.2502532660306386, "grad_norm": 1.8501474857330322, "learning_rate": 1.7571034252139253e-05, "loss": 1.1178, "step": 7102 }, { "epoch": 0.250288503043597, "grad_norm": 3.4324357509613037, "learning_rate": 1.7570288612364875e-05, "loss": 2.9033, "step": 7103 }, { "epoch": 0.2503237400565554, "grad_norm": 6.851966381072998, "learning_rate": 1.756954287398652e-05, "loss": 5.7735, "step": 7104 }, { "epoch": 0.2503589770695138, "grad_norm": 1.0196720361709595, "learning_rate": 1.756879703701391e-05, "loss": 0.7987, "step": 7105 }, { "epoch": 0.25039421408247226, "grad_norm": 3.1005589962005615, "learning_rate": 1.756805110145675e-05, "loss": 2.675, "step": 7106 }, { "epoch": 0.25042945109543063, "grad_norm": 1.7713388204574585, "learning_rate": 1.756730506732476e-05, "loss": 1.0376, "step": 7107 }, { "epoch": 0.25046468810838907, "grad_norm": 1.3266470432281494, "learning_rate": 1.756655893462766e-05, "loss": 1.0352, "step": 7108 }, { "epoch": 0.25049992512134744, "grad_norm": 2.4103095531463623, "learning_rate": 1.7565812703375167e-05, "loss": 0.9455, "step": 7109 }, { "epoch": 0.2505351621343059, "grad_norm": 1.8010832071304321, "learning_rate": 1.7565066373577e-05, "loss": 1.1706, "step": 7110 }, { "epoch": 0.2505703991472643, "grad_norm": 2.9798707962036133, "learning_rate": 1.7564319945242876e-05, "loss": 3.6423, "step": 7111 }, { "epoch": 0.2506056361602227, "grad_norm": 3.0092978477478027, "learning_rate": 1.7563573418382524e-05, "loss": 3.3651, "step": 7112 }, { "epoch": 0.2506408731731811, "grad_norm": 1.122390627861023, "learning_rate": 1.7562826793005667e-05, "loss": 0.8881, "step": 7113 }, { "epoch": 0.2506761101861395, "grad_norm": 4.646270751953125, "learning_rate": 1.7562080069122027e-05, "loss": 2.8885, "step": 7114 }, { "epoch": 0.25071134719909793, "grad_norm": 3.327913522720337, "learning_rate": 1.7561333246741325e-05, "loss": 3.3013, "step": 7115 }, { "epoch": 0.25074658421205637, "grad_norm": 0.7382897138595581, "learning_rate": 1.7560586325873298e-05, "loss": 1.2442, "step": 7116 }, { "epoch": 0.25078182122501474, "grad_norm": 2.169440984725952, "learning_rate": 1.7559839306527672e-05, "loss": 3.141, "step": 7117 }, { "epoch": 0.2508170582379732, "grad_norm": 1.6984848976135254, "learning_rate": 1.7559092188714174e-05, "loss": 1.29, "step": 7118 }, { "epoch": 0.25085229525093156, "grad_norm": 2.3212108612060547, "learning_rate": 1.755834497244254e-05, "loss": 3.8574, "step": 7119 }, { "epoch": 0.25088753226389, "grad_norm": 3.9986655712127686, "learning_rate": 1.7557597657722498e-05, "loss": 3.4273, "step": 7120 }, { "epoch": 0.2509227692768484, "grad_norm": 2.4328651428222656, "learning_rate": 1.7556850244563787e-05, "loss": 3.8225, "step": 7121 }, { "epoch": 0.2509580062898068, "grad_norm": 1.2879600524902344, "learning_rate": 1.7556102732976134e-05, "loss": 0.8587, "step": 7122 }, { "epoch": 0.25099324330276523, "grad_norm": 1.7876615524291992, "learning_rate": 1.7555355122969284e-05, "loss": 1.068, "step": 7123 }, { "epoch": 0.2510284803157236, "grad_norm": 1.2803765535354614, "learning_rate": 1.7554607414552966e-05, "loss": 0.8595, "step": 7124 }, { "epoch": 0.25106371732868205, "grad_norm": 4.313019275665283, "learning_rate": 1.755385960773693e-05, "loss": 3.5794, "step": 7125 }, { "epoch": 0.2510989543416405, "grad_norm": 1.4134916067123413, "learning_rate": 1.755311170253091e-05, "loss": 0.8373, "step": 7126 }, { "epoch": 0.25113419135459886, "grad_norm": 1.3041675090789795, "learning_rate": 1.7552363698944646e-05, "loss": 1.0825, "step": 7127 }, { "epoch": 0.2511694283675573, "grad_norm": 1.3421030044555664, "learning_rate": 1.755161559698788e-05, "loss": 1.0501, "step": 7128 }, { "epoch": 0.25120466538051567, "grad_norm": 4.10161828994751, "learning_rate": 1.7550867396670362e-05, "loss": 2.621, "step": 7129 }, { "epoch": 0.2512399023934741, "grad_norm": 4.366252899169922, "learning_rate": 1.7550119098001836e-05, "loss": 3.6883, "step": 7130 }, { "epoch": 0.25127513940643253, "grad_norm": 4.050537109375, "learning_rate": 1.7549370700992046e-05, "loss": 6.855, "step": 7131 }, { "epoch": 0.2513103764193909, "grad_norm": 0.9713522791862488, "learning_rate": 1.754862220565074e-05, "loss": 0.7623, "step": 7132 }, { "epoch": 0.25134561343234935, "grad_norm": 3.678436279296875, "learning_rate": 1.754787361198767e-05, "loss": 3.7043, "step": 7133 }, { "epoch": 0.2513808504453077, "grad_norm": 7.682681560516357, "learning_rate": 1.754712492001258e-05, "loss": 8.7279, "step": 7134 }, { "epoch": 0.25141608745826616, "grad_norm": 4.3225297927856445, "learning_rate": 1.7546376129735232e-05, "loss": 3.4213, "step": 7135 }, { "epoch": 0.2514513244712246, "grad_norm": 2.906362771987915, "learning_rate": 1.754562724116537e-05, "loss": 3.0515, "step": 7136 }, { "epoch": 0.25148656148418297, "grad_norm": 1.1535542011260986, "learning_rate": 1.7544878254312757e-05, "loss": 0.7506, "step": 7137 }, { "epoch": 0.2515217984971414, "grad_norm": 2.123663902282715, "learning_rate": 1.754412916918714e-05, "loss": 3.6408, "step": 7138 }, { "epoch": 0.25155703551009984, "grad_norm": 3.5438172817230225, "learning_rate": 1.754337998579828e-05, "loss": 3.662, "step": 7139 }, { "epoch": 0.2515922725230582, "grad_norm": 5.371511936187744, "learning_rate": 1.7542630704155934e-05, "loss": 6.1511, "step": 7140 }, { "epoch": 0.25162750953601665, "grad_norm": 1.3227702379226685, "learning_rate": 1.7541881324269864e-05, "loss": 0.9857, "step": 7141 }, { "epoch": 0.251662746548975, "grad_norm": 4.029535293579102, "learning_rate": 1.754113184614983e-05, "loss": 5.0781, "step": 7142 }, { "epoch": 0.25169798356193346, "grad_norm": 1.6987297534942627, "learning_rate": 1.754038226980559e-05, "loss": 0.9538, "step": 7143 }, { "epoch": 0.2517332205748919, "grad_norm": 1.6899234056472778, "learning_rate": 1.753963259524691e-05, "loss": 1.0032, "step": 7144 }, { "epoch": 0.25176845758785027, "grad_norm": 0.9838706254959106, "learning_rate": 1.7538882822483557e-05, "loss": 1.188, "step": 7145 }, { "epoch": 0.2518036946008087, "grad_norm": 3.0772922039031982, "learning_rate": 1.753813295152529e-05, "loss": 3.3689, "step": 7146 }, { "epoch": 0.2518389316137671, "grad_norm": 1.6592693328857422, "learning_rate": 1.7537382982381887e-05, "loss": 0.9876, "step": 7147 }, { "epoch": 0.2518741686267255, "grad_norm": 1.2776252031326294, "learning_rate": 1.753663291506311e-05, "loss": 0.8158, "step": 7148 }, { "epoch": 0.25190940563968395, "grad_norm": 1.1365429162979126, "learning_rate": 1.753588274957873e-05, "loss": 0.9722, "step": 7149 }, { "epoch": 0.2519446426526423, "grad_norm": 1.602034091949463, "learning_rate": 1.7535132485938518e-05, "loss": 1.0357, "step": 7150 }, { "epoch": 0.25197987966560076, "grad_norm": 1.5050534009933472, "learning_rate": 1.753438212415224e-05, "loss": 1.008, "step": 7151 }, { "epoch": 0.25201511667855914, "grad_norm": 4.363748073577881, "learning_rate": 1.753363166422968e-05, "loss": 2.9619, "step": 7152 }, { "epoch": 0.25205035369151757, "grad_norm": 1.3818111419677734, "learning_rate": 1.7532881106180605e-05, "loss": 1.2094, "step": 7153 }, { "epoch": 0.252085590704476, "grad_norm": 2.9292900562286377, "learning_rate": 1.7532130450014798e-05, "loss": 3.4296, "step": 7154 }, { "epoch": 0.2521208277174344, "grad_norm": 1.1455992460250854, "learning_rate": 1.753137969574203e-05, "loss": 0.8671, "step": 7155 }, { "epoch": 0.2521560647303928, "grad_norm": 3.2547667026519775, "learning_rate": 1.7530628843372082e-05, "loss": 3.3833, "step": 7156 }, { "epoch": 0.2521913017433512, "grad_norm": 8.261051177978516, "learning_rate": 1.7529877892914733e-05, "loss": 7.3597, "step": 7157 }, { "epoch": 0.2522265387563096, "grad_norm": 1.1514461040496826, "learning_rate": 1.7529126844379768e-05, "loss": 0.8075, "step": 7158 }, { "epoch": 0.25226177576926806, "grad_norm": 1.33035147190094, "learning_rate": 1.752837569777696e-05, "loss": 0.9943, "step": 7159 }, { "epoch": 0.25229701278222644, "grad_norm": 1.4491521120071411, "learning_rate": 1.752762445311611e-05, "loss": 1.0016, "step": 7160 }, { "epoch": 0.25233224979518487, "grad_norm": 1.4843467473983765, "learning_rate": 1.7526873110406987e-05, "loss": 0.964, "step": 7161 }, { "epoch": 0.25236748680814325, "grad_norm": 3.0721688270568848, "learning_rate": 1.752612166965938e-05, "loss": 3.449, "step": 7162 }, { "epoch": 0.2524027238211017, "grad_norm": 5.880751132965088, "learning_rate": 1.7525370130883085e-05, "loss": 3.7474, "step": 7163 }, { "epoch": 0.2524379608340601, "grad_norm": 4.664247989654541, "learning_rate": 1.752461849408788e-05, "loss": 6.6593, "step": 7164 }, { "epoch": 0.2524731978470185, "grad_norm": 3.581239938735962, "learning_rate": 1.7523866759283567e-05, "loss": 3.0509, "step": 7165 }, { "epoch": 0.2525084348599769, "grad_norm": 3.651182174682617, "learning_rate": 1.7523114926479928e-05, "loss": 4.0259, "step": 7166 }, { "epoch": 0.2525436718729353, "grad_norm": 1.44376802444458, "learning_rate": 1.7522362995686758e-05, "loss": 0.826, "step": 7167 }, { "epoch": 0.25257890888589374, "grad_norm": 3.6904795169830322, "learning_rate": 1.7521610966913855e-05, "loss": 3.5246, "step": 7168 }, { "epoch": 0.25261414589885217, "grad_norm": 6.941684722900391, "learning_rate": 1.7520858840171007e-05, "loss": 5.2179, "step": 7169 }, { "epoch": 0.25264938291181055, "grad_norm": 3.134552240371704, "learning_rate": 1.752010661546802e-05, "loss": 2.7776, "step": 7170 }, { "epoch": 0.252684619924769, "grad_norm": 0.9817676544189453, "learning_rate": 1.751935429281468e-05, "loss": 1.018, "step": 7171 }, { "epoch": 0.2527198569377274, "grad_norm": 3.5242249965667725, "learning_rate": 1.7518601872220795e-05, "loss": 3.1032, "step": 7172 }, { "epoch": 0.2527550939506858, "grad_norm": 1.576919674873352, "learning_rate": 1.7517849353696162e-05, "loss": 0.9157, "step": 7173 }, { "epoch": 0.2527903309636442, "grad_norm": 2.8527350425720215, "learning_rate": 1.7517096737250584e-05, "loss": 2.7185, "step": 7174 }, { "epoch": 0.2528255679766026, "grad_norm": 1.4830958843231201, "learning_rate": 1.7516344022893865e-05, "loss": 1.0895, "step": 7175 }, { "epoch": 0.25286080498956104, "grad_norm": 3.7659246921539307, "learning_rate": 1.751559121063581e-05, "loss": 5.7933, "step": 7176 }, { "epoch": 0.25289604200251947, "grad_norm": 5.391792297363281, "learning_rate": 1.751483830048622e-05, "loss": 3.4135, "step": 7177 }, { "epoch": 0.25293127901547785, "grad_norm": 1.6254827976226807, "learning_rate": 1.75140852924549e-05, "loss": 0.9158, "step": 7178 }, { "epoch": 0.2529665160284363, "grad_norm": 1.5994881391525269, "learning_rate": 1.7513332186551666e-05, "loss": 1.1504, "step": 7179 }, { "epoch": 0.25300175304139466, "grad_norm": 10.319168090820312, "learning_rate": 1.751257898278632e-05, "loss": 5.6039, "step": 7180 }, { "epoch": 0.2530369900543531, "grad_norm": 2.1204071044921875, "learning_rate": 1.751182568116868e-05, "loss": 3.9613, "step": 7181 }, { "epoch": 0.25307222706731153, "grad_norm": 2.7065627574920654, "learning_rate": 1.751107228170855e-05, "loss": 3.1216, "step": 7182 }, { "epoch": 0.2531074640802699, "grad_norm": 4.557120323181152, "learning_rate": 1.751031878441575e-05, "loss": 4.4411, "step": 7183 }, { "epoch": 0.25314270109322834, "grad_norm": 1.4927663803100586, "learning_rate": 1.750956518930009e-05, "loss": 0.9566, "step": 7184 }, { "epoch": 0.2531779381061867, "grad_norm": 5.4368391036987305, "learning_rate": 1.7508811496371384e-05, "loss": 6.5985, "step": 7185 }, { "epoch": 0.25321317511914515, "grad_norm": 1.618128776550293, "learning_rate": 1.7508057705639453e-05, "loss": 0.7846, "step": 7186 }, { "epoch": 0.2532484121321036, "grad_norm": 1.615065336227417, "learning_rate": 1.750730381711412e-05, "loss": 0.71, "step": 7187 }, { "epoch": 0.25328364914506196, "grad_norm": 1.065680980682373, "learning_rate": 1.7506549830805192e-05, "loss": 1.0112, "step": 7188 }, { "epoch": 0.2533188861580204, "grad_norm": 3.3865997791290283, "learning_rate": 1.7505795746722495e-05, "loss": 3.6016, "step": 7189 }, { "epoch": 0.2533541231709788, "grad_norm": 4.698020935058594, "learning_rate": 1.7505041564875855e-05, "loss": 3.5082, "step": 7190 }, { "epoch": 0.2533893601839372, "grad_norm": 11.122891426086426, "learning_rate": 1.7504287285275095e-05, "loss": 9.3586, "step": 7191 }, { "epoch": 0.25342459719689564, "grad_norm": 4.795764446258545, "learning_rate": 1.7503532907930036e-05, "loss": 3.5006, "step": 7192 }, { "epoch": 0.253459834209854, "grad_norm": 1.4716076850891113, "learning_rate": 1.7502778432850503e-05, "loss": 0.7856, "step": 7193 }, { "epoch": 0.25349507122281245, "grad_norm": 2.347433567047119, "learning_rate": 1.7502023860046327e-05, "loss": 3.3941, "step": 7194 }, { "epoch": 0.25353030823577083, "grad_norm": 4.235348701477051, "learning_rate": 1.7501269189527335e-05, "loss": 3.7792, "step": 7195 }, { "epoch": 0.25356554524872926, "grad_norm": 1.4485218524932861, "learning_rate": 1.750051442130335e-05, "loss": 0.9894, "step": 7196 }, { "epoch": 0.2536007822616877, "grad_norm": 1.905869960784912, "learning_rate": 1.7499759555384215e-05, "loss": 0.9499, "step": 7197 }, { "epoch": 0.2536360192746461, "grad_norm": 3.447479486465454, "learning_rate": 1.749900459177976e-05, "loss": 3.4634, "step": 7198 }, { "epoch": 0.2536712562876045, "grad_norm": 3.5416057109832764, "learning_rate": 1.749824953049981e-05, "loss": 3.7319, "step": 7199 }, { "epoch": 0.2537064933005629, "grad_norm": 4.496354579925537, "learning_rate": 1.7497494371554204e-05, "loss": 5.4519, "step": 7200 }, { "epoch": 0.2537417303135213, "grad_norm": 4.5247602462768555, "learning_rate": 1.749673911495278e-05, "loss": 5.5748, "step": 7201 }, { "epoch": 0.25377696732647975, "grad_norm": 3.615967273712158, "learning_rate": 1.749598376070538e-05, "loss": 4.3309, "step": 7202 }, { "epoch": 0.25381220433943813, "grad_norm": 1.3086923360824585, "learning_rate": 1.7495228308821827e-05, "loss": 0.9089, "step": 7203 }, { "epoch": 0.25384744135239656, "grad_norm": 1.630920648574829, "learning_rate": 1.7494472759311976e-05, "loss": 0.9938, "step": 7204 }, { "epoch": 0.25388267836535494, "grad_norm": 2.896169900894165, "learning_rate": 1.749371711218566e-05, "loss": 3.6509, "step": 7205 }, { "epoch": 0.2539179153783134, "grad_norm": 1.5623438358306885, "learning_rate": 1.749296136745273e-05, "loss": 1.1448, "step": 7206 }, { "epoch": 0.2539531523912718, "grad_norm": 2.0529749393463135, "learning_rate": 1.7492205525123018e-05, "loss": 0.8583, "step": 7207 }, { "epoch": 0.2539883894042302, "grad_norm": 1.372133493423462, "learning_rate": 1.7491449585206373e-05, "loss": 1.1189, "step": 7208 }, { "epoch": 0.2540236264171886, "grad_norm": 1.6009443998336792, "learning_rate": 1.7490693547712645e-05, "loss": 0.9901, "step": 7209 }, { "epoch": 0.25405886343014705, "grad_norm": 4.088907241821289, "learning_rate": 1.748993741265168e-05, "loss": 3.4791, "step": 7210 }, { "epoch": 0.25409410044310543, "grad_norm": 3.7077817916870117, "learning_rate": 1.748918118003333e-05, "loss": 3.7442, "step": 7211 }, { "epoch": 0.25412933745606386, "grad_norm": 1.4211362600326538, "learning_rate": 1.7488424849867437e-05, "loss": 0.7632, "step": 7212 }, { "epoch": 0.25416457446902224, "grad_norm": 1.247055172920227, "learning_rate": 1.7487668422163855e-05, "loss": 1.1089, "step": 7213 }, { "epoch": 0.2541998114819807, "grad_norm": 5.659896373748779, "learning_rate": 1.7486911896932438e-05, "loss": 6.0746, "step": 7214 }, { "epoch": 0.2542350484949391, "grad_norm": 2.0470449924468994, "learning_rate": 1.7486155274183043e-05, "loss": 0.8933, "step": 7215 }, { "epoch": 0.2542702855078975, "grad_norm": 8.578794479370117, "learning_rate": 1.7485398553925522e-05, "loss": 6.7605, "step": 7216 }, { "epoch": 0.2543055225208559, "grad_norm": 1.17783522605896, "learning_rate": 1.7484641736169732e-05, "loss": 0.7544, "step": 7217 }, { "epoch": 0.2543407595338143, "grad_norm": 2.9636807441711426, "learning_rate": 1.7483884820925525e-05, "loss": 3.1202, "step": 7218 }, { "epoch": 0.25437599654677273, "grad_norm": 4.3343658447265625, "learning_rate": 1.7483127808202767e-05, "loss": 4.1267, "step": 7219 }, { "epoch": 0.25441123355973116, "grad_norm": 2.2076168060302734, "learning_rate": 1.748237069801132e-05, "loss": 1.0054, "step": 7220 }, { "epoch": 0.25444647057268954, "grad_norm": 4.836513519287109, "learning_rate": 1.748161349036104e-05, "loss": 5.8403, "step": 7221 }, { "epoch": 0.254481707585648, "grad_norm": 4.1042094230651855, "learning_rate": 1.748085618526179e-05, "loss": 2.921, "step": 7222 }, { "epoch": 0.25451694459860635, "grad_norm": 1.3488225936889648, "learning_rate": 1.7480098782723434e-05, "loss": 1.4393, "step": 7223 }, { "epoch": 0.2545521816115648, "grad_norm": 3.398705244064331, "learning_rate": 1.747934128275584e-05, "loss": 2.7959, "step": 7224 }, { "epoch": 0.2545874186245232, "grad_norm": 1.3398714065551758, "learning_rate": 1.747858368536887e-05, "loss": 0.7437, "step": 7225 }, { "epoch": 0.2546226556374816, "grad_norm": 6.58212947845459, "learning_rate": 1.74778259905724e-05, "loss": 6.2722, "step": 7226 }, { "epoch": 0.25465789265044003, "grad_norm": 3.9977610111236572, "learning_rate": 1.747706819837629e-05, "loss": 3.3618, "step": 7227 }, { "epoch": 0.2546931296633984, "grad_norm": 1.4390915632247925, "learning_rate": 1.747631030879042e-05, "loss": 1.0813, "step": 7228 }, { "epoch": 0.25472836667635684, "grad_norm": 4.25828742980957, "learning_rate": 1.747555232182465e-05, "loss": 3.0499, "step": 7229 }, { "epoch": 0.2547636036893153, "grad_norm": 1.1026201248168945, "learning_rate": 1.7474794237488863e-05, "loss": 0.9638, "step": 7230 }, { "epoch": 0.25479884070227365, "grad_norm": 2.11291241645813, "learning_rate": 1.7474036055792926e-05, "loss": 0.9912, "step": 7231 }, { "epoch": 0.2548340777152321, "grad_norm": 4.133748531341553, "learning_rate": 1.7473277776746717e-05, "loss": 3.1129, "step": 7232 }, { "epoch": 0.25486931472819047, "grad_norm": 4.154248237609863, "learning_rate": 1.7472519400360112e-05, "loss": 2.4458, "step": 7233 }, { "epoch": 0.2549045517411489, "grad_norm": 4.161622047424316, "learning_rate": 1.747176092664299e-05, "loss": 3.1299, "step": 7234 }, { "epoch": 0.25493978875410733, "grad_norm": 5.5060625076293945, "learning_rate": 1.7471002355605232e-05, "loss": 7.0984, "step": 7235 }, { "epoch": 0.2549750257670657, "grad_norm": 2.710428237915039, "learning_rate": 1.747024368725672e-05, "loss": 3.0263, "step": 7236 }, { "epoch": 0.25501026278002414, "grad_norm": 3.630941867828369, "learning_rate": 1.746948492160733e-05, "loss": 3.422, "step": 7237 }, { "epoch": 0.2550454997929825, "grad_norm": 1.2523396015167236, "learning_rate": 1.7468726058666944e-05, "loss": 0.7565, "step": 7238 }, { "epoch": 0.25508073680594096, "grad_norm": 3.358853340148926, "learning_rate": 1.746796709844545e-05, "loss": 3.0904, "step": 7239 }, { "epoch": 0.2551159738188994, "grad_norm": 2.2345521450042725, "learning_rate": 1.7467208040952738e-05, "loss": 3.9788, "step": 7240 }, { "epoch": 0.25515121083185777, "grad_norm": 0.952559769153595, "learning_rate": 1.7466448886198688e-05, "loss": 0.983, "step": 7241 }, { "epoch": 0.2551864478448162, "grad_norm": 3.832343101501465, "learning_rate": 1.746568963419319e-05, "loss": 3.7891, "step": 7242 }, { "epoch": 0.25522168485777463, "grad_norm": 2.608025550842285, "learning_rate": 1.7464930284946137e-05, "loss": 3.1437, "step": 7243 }, { "epoch": 0.255256921870733, "grad_norm": 1.8628734350204468, "learning_rate": 1.746417083846741e-05, "loss": 1.1566, "step": 7244 }, { "epoch": 0.25529215888369144, "grad_norm": 0.929916262626648, "learning_rate": 1.746341129476691e-05, "loss": 1.0696, "step": 7245 }, { "epoch": 0.2553273958966498, "grad_norm": 1.15700101852417, "learning_rate": 1.7462651653854528e-05, "loss": 0.9787, "step": 7246 }, { "epoch": 0.25536263290960826, "grad_norm": 1.722963571548462, "learning_rate": 1.7461891915740155e-05, "loss": 1.0636, "step": 7247 }, { "epoch": 0.2553978699225667, "grad_norm": 1.1916310787200928, "learning_rate": 1.7461132080433692e-05, "loss": 0.9412, "step": 7248 }, { "epoch": 0.25543310693552507, "grad_norm": 1.723637342453003, "learning_rate": 1.746037214794503e-05, "loss": 0.9207, "step": 7249 }, { "epoch": 0.2554683439484835, "grad_norm": 1.9293783903121948, "learning_rate": 1.7459612118284073e-05, "loss": 1.0742, "step": 7250 }, { "epoch": 0.2555035809614419, "grad_norm": 1.0809448957443237, "learning_rate": 1.7458851991460717e-05, "loss": 0.8768, "step": 7251 }, { "epoch": 0.2555388179744003, "grad_norm": 3.7627031803131104, "learning_rate": 1.7458091767484865e-05, "loss": 3.7065, "step": 7252 }, { "epoch": 0.25557405498735875, "grad_norm": 1.3279571533203125, "learning_rate": 1.7457331446366415e-05, "loss": 0.9529, "step": 7253 }, { "epoch": 0.2556092920003171, "grad_norm": 2.7289719581604004, "learning_rate": 1.7456571028115275e-05, "loss": 3.8977, "step": 7254 }, { "epoch": 0.25564452901327556, "grad_norm": 1.8980257511138916, "learning_rate": 1.745581051274135e-05, "loss": 0.9361, "step": 7255 }, { "epoch": 0.25567976602623393, "grad_norm": 3.8340325355529785, "learning_rate": 1.745504990025454e-05, "loss": 3.7327, "step": 7256 }, { "epoch": 0.25571500303919237, "grad_norm": 4.022204875946045, "learning_rate": 1.7454289190664755e-05, "loss": 6.4283, "step": 7257 }, { "epoch": 0.2557502400521508, "grad_norm": 1.4117510318756104, "learning_rate": 1.7453528383981907e-05, "loss": 0.8879, "step": 7258 }, { "epoch": 0.2557854770651092, "grad_norm": 3.811100482940674, "learning_rate": 1.7452767480215897e-05, "loss": 3.5923, "step": 7259 }, { "epoch": 0.2558207140780676, "grad_norm": 1.83525550365448, "learning_rate": 1.7452006479376648e-05, "loss": 1.0469, "step": 7260 }, { "epoch": 0.255855951091026, "grad_norm": 3.934436559677124, "learning_rate": 1.745124538147406e-05, "loss": 3.7621, "step": 7261 }, { "epoch": 0.2558911881039844, "grad_norm": 4.690822601318359, "learning_rate": 1.7450484186518056e-05, "loss": 6.5984, "step": 7262 }, { "epoch": 0.25592642511694286, "grad_norm": 8.465060234069824, "learning_rate": 1.7449722894518543e-05, "loss": 3.9084, "step": 7263 }, { "epoch": 0.25596166212990124, "grad_norm": 1.7052171230316162, "learning_rate": 1.7448961505485443e-05, "loss": 0.8229, "step": 7264 }, { "epoch": 0.25599689914285967, "grad_norm": 4.941412448883057, "learning_rate": 1.7448200019428666e-05, "loss": 3.6798, "step": 7265 }, { "epoch": 0.25603213615581805, "grad_norm": 1.3491202592849731, "learning_rate": 1.744743843635814e-05, "loss": 1.2111, "step": 7266 }, { "epoch": 0.2560673731687765, "grad_norm": 5.554322719573975, "learning_rate": 1.7446676756283777e-05, "loss": 5.7737, "step": 7267 }, { "epoch": 0.2561026101817349, "grad_norm": 3.7968862056732178, "learning_rate": 1.74459149792155e-05, "loss": 3.4224, "step": 7268 }, { "epoch": 0.2561378471946933, "grad_norm": 1.0867462158203125, "learning_rate": 1.744515310516323e-05, "loss": 0.8739, "step": 7269 }, { "epoch": 0.2561730842076517, "grad_norm": 3.75658917427063, "learning_rate": 1.74443911341369e-05, "loss": 3.1965, "step": 7270 }, { "epoch": 0.2562083212206101, "grad_norm": 2.387559413909912, "learning_rate": 1.7443629066146425e-05, "loss": 0.9721, "step": 7271 }, { "epoch": 0.25624355823356854, "grad_norm": 4.895137786865234, "learning_rate": 1.744286690120173e-05, "loss": 3.5287, "step": 7272 }, { "epoch": 0.25627879524652697, "grad_norm": 0.7353122234344482, "learning_rate": 1.7442104639312748e-05, "loss": 0.9145, "step": 7273 }, { "epoch": 0.25631403225948535, "grad_norm": 7.678900241851807, "learning_rate": 1.74413422804894e-05, "loss": 8.2503, "step": 7274 }, { "epoch": 0.2563492692724438, "grad_norm": 5.487190246582031, "learning_rate": 1.7440579824741628e-05, "loss": 6.7975, "step": 7275 }, { "epoch": 0.2563845062854022, "grad_norm": 1.130662441253662, "learning_rate": 1.743981727207935e-05, "loss": 0.9264, "step": 7276 }, { "epoch": 0.2564197432983606, "grad_norm": 3.1236307621002197, "learning_rate": 1.7439054622512507e-05, "loss": 3.0929, "step": 7277 }, { "epoch": 0.256454980311319, "grad_norm": 1.7102892398834229, "learning_rate": 1.7438291876051033e-05, "loss": 0.9543, "step": 7278 }, { "epoch": 0.2564902173242774, "grad_norm": 4.40526819229126, "learning_rate": 1.7437529032704857e-05, "loss": 3.1315, "step": 7279 }, { "epoch": 0.25652545433723584, "grad_norm": 4.249974250793457, "learning_rate": 1.7436766092483918e-05, "loss": 3.5125, "step": 7280 }, { "epoch": 0.25656069135019427, "grad_norm": 2.333479881286621, "learning_rate": 1.7436003055398152e-05, "loss": 3.6588, "step": 7281 }, { "epoch": 0.25659592836315265, "grad_norm": 5.049411773681641, "learning_rate": 1.74352399214575e-05, "loss": 3.9763, "step": 7282 }, { "epoch": 0.2566311653761111, "grad_norm": 4.264649868011475, "learning_rate": 1.7434476690671903e-05, "loss": 3.4222, "step": 7283 }, { "epoch": 0.25666640238906946, "grad_norm": 2.427333354949951, "learning_rate": 1.7433713363051295e-05, "loss": 2.9136, "step": 7284 }, { "epoch": 0.2567016394020279, "grad_norm": 1.1196937561035156, "learning_rate": 1.743294993860563e-05, "loss": 0.895, "step": 7285 }, { "epoch": 0.2567368764149863, "grad_norm": 8.594534873962402, "learning_rate": 1.743218641734484e-05, "loss": 8.5838, "step": 7286 }, { "epoch": 0.2567721134279447, "grad_norm": 5.970659255981445, "learning_rate": 1.7431422799278877e-05, "loss": 3.7765, "step": 7287 }, { "epoch": 0.25680735044090314, "grad_norm": 7.394329071044922, "learning_rate": 1.7430659084417686e-05, "loss": 3.2361, "step": 7288 }, { "epoch": 0.2568425874538615, "grad_norm": 0.9486382007598877, "learning_rate": 1.7429895272771213e-05, "loss": 0.7652, "step": 7289 }, { "epoch": 0.25687782446681995, "grad_norm": 3.7762277126312256, "learning_rate": 1.742913136434941e-05, "loss": 3.1977, "step": 7290 }, { "epoch": 0.2569130614797784, "grad_norm": 1.207809567451477, "learning_rate": 1.7428367359162224e-05, "loss": 0.83, "step": 7291 }, { "epoch": 0.25694829849273676, "grad_norm": 3.2825047969818115, "learning_rate": 1.7427603257219604e-05, "loss": 3.0365, "step": 7292 }, { "epoch": 0.2569835355056952, "grad_norm": 1.8586363792419434, "learning_rate": 1.742683905853151e-05, "loss": 0.8668, "step": 7293 }, { "epoch": 0.25701877251865357, "grad_norm": 4.387980937957764, "learning_rate": 1.7426074763107887e-05, "loss": 3.831, "step": 7294 }, { "epoch": 0.257054009531612, "grad_norm": 3.1569225788116455, "learning_rate": 1.7425310370958697e-05, "loss": 3.4167, "step": 7295 }, { "epoch": 0.25708924654457044, "grad_norm": 3.9311416149139404, "learning_rate": 1.742454588209389e-05, "loss": 3.1953, "step": 7296 }, { "epoch": 0.2571244835575288, "grad_norm": 1.8967959880828857, "learning_rate": 1.742378129652343e-05, "loss": 1.0199, "step": 7297 }, { "epoch": 0.25715972057048725, "grad_norm": 1.9797877073287964, "learning_rate": 1.7423016614257275e-05, "loss": 0.7828, "step": 7298 }, { "epoch": 0.2571949575834456, "grad_norm": 2.343411922454834, "learning_rate": 1.742225183530538e-05, "loss": 1.0162, "step": 7299 }, { "epoch": 0.25723019459640406, "grad_norm": 0.8990923166275024, "learning_rate": 1.742148695967771e-05, "loss": 1.3082, "step": 7300 }, { "epoch": 0.2572654316093625, "grad_norm": 1.7768443822860718, "learning_rate": 1.7420721987384224e-05, "loss": 1.0235, "step": 7301 }, { "epoch": 0.25730066862232087, "grad_norm": 2.495812177658081, "learning_rate": 1.7419956918434896e-05, "loss": 3.0466, "step": 7302 }, { "epoch": 0.2573359056352793, "grad_norm": 4.862337589263916, "learning_rate": 1.7419191752839678e-05, "loss": 3.5787, "step": 7303 }, { "epoch": 0.2573711426482377, "grad_norm": 1.2117780447006226, "learning_rate": 1.741842649060854e-05, "loss": 0.9572, "step": 7304 }, { "epoch": 0.2574063796611961, "grad_norm": 3.226153612136841, "learning_rate": 1.741766113175146e-05, "loss": 3.5263, "step": 7305 }, { "epoch": 0.25744161667415455, "grad_norm": 5.338033199310303, "learning_rate": 1.7416895676278395e-05, "loss": 5.6629, "step": 7306 }, { "epoch": 0.2574768536871129, "grad_norm": 1.3716318607330322, "learning_rate": 1.741613012419932e-05, "loss": 0.9764, "step": 7307 }, { "epoch": 0.25751209070007136, "grad_norm": 3.7849442958831787, "learning_rate": 1.7415364475524203e-05, "loss": 3.5522, "step": 7308 }, { "epoch": 0.25754732771302974, "grad_norm": 4.228609561920166, "learning_rate": 1.741459873026302e-05, "loss": 5.954, "step": 7309 }, { "epoch": 0.2575825647259882, "grad_norm": 4.965451240539551, "learning_rate": 1.7413832888425747e-05, "loss": 5.2139, "step": 7310 }, { "epoch": 0.2576178017389466, "grad_norm": 6.04259729385376, "learning_rate": 1.7413066950022353e-05, "loss": 3.5329, "step": 7311 }, { "epoch": 0.257653038751905, "grad_norm": 2.4342286586761475, "learning_rate": 1.7412300915062818e-05, "loss": 2.6646, "step": 7312 }, { "epoch": 0.2576882757648634, "grad_norm": 4.90541410446167, "learning_rate": 1.7411534783557122e-05, "loss": 6.2787, "step": 7313 }, { "epoch": 0.25772351277782185, "grad_norm": 1.775153636932373, "learning_rate": 1.7410768555515237e-05, "loss": 1.0717, "step": 7314 }, { "epoch": 0.25775874979078023, "grad_norm": 2.4834060668945312, "learning_rate": 1.741000223094715e-05, "loss": 1.077, "step": 7315 }, { "epoch": 0.25779398680373866, "grad_norm": 5.749081134796143, "learning_rate": 1.7409235809862843e-05, "loss": 4.2864, "step": 7316 }, { "epoch": 0.25782922381669704, "grad_norm": 4.864014625549316, "learning_rate": 1.7408469292272293e-05, "loss": 5.9945, "step": 7317 }, { "epoch": 0.2578644608296555, "grad_norm": 3.95591402053833, "learning_rate": 1.7407702678185486e-05, "loss": 5.8666, "step": 7318 }, { "epoch": 0.2578996978426139, "grad_norm": 2.6006927490234375, "learning_rate": 1.7406935967612407e-05, "loss": 0.9679, "step": 7319 }, { "epoch": 0.2579349348555723, "grad_norm": 4.129010200500488, "learning_rate": 1.7406169160563045e-05, "loss": 3.2932, "step": 7320 }, { "epoch": 0.2579701718685307, "grad_norm": 1.6201651096343994, "learning_rate": 1.7405402257047388e-05, "loss": 0.8483, "step": 7321 }, { "epoch": 0.2580054088814891, "grad_norm": 1.1151046752929688, "learning_rate": 1.7404635257075423e-05, "loss": 0.8997, "step": 7322 }, { "epoch": 0.25804064589444753, "grad_norm": 5.162546634674072, "learning_rate": 1.740386816065714e-05, "loss": 3.2544, "step": 7323 }, { "epoch": 0.25807588290740596, "grad_norm": 4.904140949249268, "learning_rate": 1.740310096780253e-05, "loss": 2.8245, "step": 7324 }, { "epoch": 0.25811111992036434, "grad_norm": 4.288875102996826, "learning_rate": 1.7402333678521588e-05, "loss": 5.0936, "step": 7325 }, { "epoch": 0.2581463569333228, "grad_norm": 10.071358680725098, "learning_rate": 1.7401566292824304e-05, "loss": 8.9503, "step": 7326 }, { "epoch": 0.25818159394628115, "grad_norm": 1.1852928400039673, "learning_rate": 1.740079881072068e-05, "loss": 0.8855, "step": 7327 }, { "epoch": 0.2582168309592396, "grad_norm": 1.0543133020401, "learning_rate": 1.7400031232220707e-05, "loss": 1.3207, "step": 7328 }, { "epoch": 0.258252067972198, "grad_norm": 0.927173912525177, "learning_rate": 1.7399263557334384e-05, "loss": 1.4925, "step": 7329 }, { "epoch": 0.2582873049851564, "grad_norm": 1.5803059339523315, "learning_rate": 1.739849578607171e-05, "loss": 0.7752, "step": 7330 }, { "epoch": 0.25832254199811483, "grad_norm": 1.5000447034835815, "learning_rate": 1.7397727918442692e-05, "loss": 1.0786, "step": 7331 }, { "epoch": 0.2583577790110732, "grad_norm": 9.128645896911621, "learning_rate": 1.739695995445732e-05, "loss": 3.233, "step": 7332 }, { "epoch": 0.25839301602403164, "grad_norm": 1.0728330612182617, "learning_rate": 1.7396191894125604e-05, "loss": 0.9856, "step": 7333 }, { "epoch": 0.2584282530369901, "grad_norm": 3.2364559173583984, "learning_rate": 1.7395423737457546e-05, "loss": 3.6382, "step": 7334 }, { "epoch": 0.25846349004994845, "grad_norm": 1.4519954919815063, "learning_rate": 1.7394655484463152e-05, "loss": 0.9957, "step": 7335 }, { "epoch": 0.2584987270629069, "grad_norm": 1.1039305925369263, "learning_rate": 1.739388713515243e-05, "loss": 1.0629, "step": 7336 }, { "epoch": 0.25853396407586526, "grad_norm": 1.5853537321090698, "learning_rate": 1.7393118689535384e-05, "loss": 1.1721, "step": 7337 }, { "epoch": 0.2585692010888237, "grad_norm": 1.0254322290420532, "learning_rate": 1.7392350147622026e-05, "loss": 1.047, "step": 7338 }, { "epoch": 0.25860443810178213, "grad_norm": 5.666289329528809, "learning_rate": 1.7391581509422365e-05, "loss": 6.4441, "step": 7339 }, { "epoch": 0.2586396751147405, "grad_norm": 1.8772109746932983, "learning_rate": 1.7390812774946414e-05, "loss": 0.947, "step": 7340 }, { "epoch": 0.25867491212769894, "grad_norm": 0.8703340888023376, "learning_rate": 1.7390043944204183e-05, "loss": 0.8728, "step": 7341 }, { "epoch": 0.2587101491406573, "grad_norm": 3.264033794403076, "learning_rate": 1.738927501720569e-05, "loss": 3.0853, "step": 7342 }, { "epoch": 0.25874538615361575, "grad_norm": 1.9006154537200928, "learning_rate": 1.7388505993960948e-05, "loss": 1.1666, "step": 7343 }, { "epoch": 0.2587806231665742, "grad_norm": 0.8410894870758057, "learning_rate": 1.7387736874479977e-05, "loss": 0.9377, "step": 7344 }, { "epoch": 0.25881586017953256, "grad_norm": 1.2357245683670044, "learning_rate": 1.738696765877279e-05, "loss": 0.8599, "step": 7345 }, { "epoch": 0.258851097192491, "grad_norm": 1.099347472190857, "learning_rate": 1.7386198346849408e-05, "loss": 0.9931, "step": 7346 }, { "epoch": 0.25888633420544943, "grad_norm": 3.885848045349121, "learning_rate": 1.7385428938719854e-05, "loss": 2.9666, "step": 7347 }, { "epoch": 0.2589215712184078, "grad_norm": 5.251529693603516, "learning_rate": 1.7384659434394143e-05, "loss": 5.2873, "step": 7348 }, { "epoch": 0.25895680823136624, "grad_norm": 6.725138187408447, "learning_rate": 1.7383889833882305e-05, "loss": 6.8198, "step": 7349 }, { "epoch": 0.2589920452443246, "grad_norm": 3.1756858825683594, "learning_rate": 1.738312013719436e-05, "loss": 3.131, "step": 7350 }, { "epoch": 0.25902728225728305, "grad_norm": 3.915544033050537, "learning_rate": 1.738235034434034e-05, "loss": 2.9244, "step": 7351 }, { "epoch": 0.2590625192702415, "grad_norm": 5.153426647186279, "learning_rate": 1.738158045533026e-05, "loss": 3.1747, "step": 7352 }, { "epoch": 0.25909775628319986, "grad_norm": 1.6869367361068726, "learning_rate": 1.7380810470174157e-05, "loss": 1.2101, "step": 7353 }, { "epoch": 0.2591329932961583, "grad_norm": 1.8817332983016968, "learning_rate": 1.7380040388882058e-05, "loss": 0.9995, "step": 7354 }, { "epoch": 0.2591682303091167, "grad_norm": 4.177927494049072, "learning_rate": 1.737927021146399e-05, "loss": 3.3354, "step": 7355 }, { "epoch": 0.2592034673220751, "grad_norm": 5.422322750091553, "learning_rate": 1.737849993792999e-05, "loss": 6.3738, "step": 7356 }, { "epoch": 0.25923870433503354, "grad_norm": 1.5559203624725342, "learning_rate": 1.737772956829009e-05, "loss": 0.8689, "step": 7357 }, { "epoch": 0.2592739413479919, "grad_norm": 1.3669074773788452, "learning_rate": 1.7376959102554322e-05, "loss": 0.7434, "step": 7358 }, { "epoch": 0.25930917836095035, "grad_norm": 3.159921407699585, "learning_rate": 1.7376188540732717e-05, "loss": 4.256, "step": 7359 }, { "epoch": 0.25934441537390873, "grad_norm": 1.340194821357727, "learning_rate": 1.7375417882835322e-05, "loss": 0.9042, "step": 7360 }, { "epoch": 0.25937965238686717, "grad_norm": 1.8058356046676636, "learning_rate": 1.7374647128872166e-05, "loss": 0.9028, "step": 7361 }, { "epoch": 0.2594148893998256, "grad_norm": 1.1192094087600708, "learning_rate": 1.7373876278853293e-05, "loss": 0.7869, "step": 7362 }, { "epoch": 0.259450126412784, "grad_norm": 4.432722568511963, "learning_rate": 1.7373105332788744e-05, "loss": 4.3259, "step": 7363 }, { "epoch": 0.2594853634257424, "grad_norm": 2.9629011154174805, "learning_rate": 1.737233429068856e-05, "loss": 3.089, "step": 7364 }, { "epoch": 0.2595206004387008, "grad_norm": 4.469459533691406, "learning_rate": 1.7371563152562777e-05, "loss": 5.4117, "step": 7365 }, { "epoch": 0.2595558374516592, "grad_norm": 1.1416696310043335, "learning_rate": 1.7370791918421452e-05, "loss": 1.1278, "step": 7366 }, { "epoch": 0.25959107446461765, "grad_norm": 1.7748050689697266, "learning_rate": 1.7370020588274617e-05, "loss": 1.1638, "step": 7367 }, { "epoch": 0.25962631147757603, "grad_norm": 7.868484973907471, "learning_rate": 1.7369249162132325e-05, "loss": 11.4614, "step": 7368 }, { "epoch": 0.25966154849053447, "grad_norm": 1.1047252416610718, "learning_rate": 1.736847764000463e-05, "loss": 0.9485, "step": 7369 }, { "epoch": 0.25969678550349284, "grad_norm": 2.2974812984466553, "learning_rate": 1.7367706021901575e-05, "loss": 3.3233, "step": 7370 }, { "epoch": 0.2597320225164513, "grad_norm": 7.646052360534668, "learning_rate": 1.7366934307833204e-05, "loss": 10.736, "step": 7371 }, { "epoch": 0.2597672595294097, "grad_norm": 2.749549388885498, "learning_rate": 1.736616249780958e-05, "loss": 3.5361, "step": 7372 }, { "epoch": 0.2598024965423681, "grad_norm": 2.1192548274993896, "learning_rate": 1.736539059184075e-05, "loss": 2.7722, "step": 7373 }, { "epoch": 0.2598377335553265, "grad_norm": 5.106565475463867, "learning_rate": 1.736461858993677e-05, "loss": 3.8382, "step": 7374 }, { "epoch": 0.2598729705682849, "grad_norm": 2.890988826751709, "learning_rate": 1.736384649210769e-05, "loss": 3.9943, "step": 7375 }, { "epoch": 0.25990820758124333, "grad_norm": 2.226729154586792, "learning_rate": 1.7363074298363577e-05, "loss": 0.7798, "step": 7376 }, { "epoch": 0.25994344459420177, "grad_norm": 1.016263484954834, "learning_rate": 1.7362302008714482e-05, "loss": 1.033, "step": 7377 }, { "epoch": 0.25997868160716014, "grad_norm": 1.0543463230133057, "learning_rate": 1.7361529623170466e-05, "loss": 0.8748, "step": 7378 }, { "epoch": 0.2600139186201186, "grad_norm": 1.5633841753005981, "learning_rate": 1.7360757141741586e-05, "loss": 0.9974, "step": 7379 }, { "epoch": 0.260049155633077, "grad_norm": 1.0982286930084229, "learning_rate": 1.735998456443791e-05, "loss": 0.9206, "step": 7380 }, { "epoch": 0.2600843926460354, "grad_norm": 1.7941874265670776, "learning_rate": 1.7359211891269495e-05, "loss": 1.0694, "step": 7381 }, { "epoch": 0.2601196296589938, "grad_norm": 1.1316343545913696, "learning_rate": 1.7358439122246414e-05, "loss": 1.1973, "step": 7382 }, { "epoch": 0.2601548666719522, "grad_norm": 2.3342936038970947, "learning_rate": 1.735766625737872e-05, "loss": 3.0302, "step": 7383 }, { "epoch": 0.26019010368491063, "grad_norm": 2.656677722930908, "learning_rate": 1.7356893296676485e-05, "loss": 2.8933, "step": 7384 }, { "epoch": 0.26022534069786907, "grad_norm": 1.2157976627349854, "learning_rate": 1.7356120240149778e-05, "loss": 1.2022, "step": 7385 }, { "epoch": 0.26026057771082745, "grad_norm": 3.4904420375823975, "learning_rate": 1.735534708780867e-05, "loss": 1.1489, "step": 7386 }, { "epoch": 0.2602958147237859, "grad_norm": 1.1431734561920166, "learning_rate": 1.735457383966323e-05, "loss": 1.1408, "step": 7387 }, { "epoch": 0.26033105173674426, "grad_norm": 1.157391905784607, "learning_rate": 1.735380049572353e-05, "loss": 1.0347, "step": 7388 }, { "epoch": 0.2603662887497027, "grad_norm": 2.810737371444702, "learning_rate": 1.7353027055999637e-05, "loss": 3.4397, "step": 7389 }, { "epoch": 0.2604015257626611, "grad_norm": 6.261516571044922, "learning_rate": 1.7352253520501632e-05, "loss": 2.3068, "step": 7390 }, { "epoch": 0.2604367627756195, "grad_norm": 3.1539342403411865, "learning_rate": 1.735147988923959e-05, "loss": 3.1691, "step": 7391 }, { "epoch": 0.26047199978857793, "grad_norm": 3.0289885997772217, "learning_rate": 1.7350706162223584e-05, "loss": 3.5533, "step": 7392 }, { "epoch": 0.2605072368015363, "grad_norm": 8.396625518798828, "learning_rate": 1.7349932339463696e-05, "loss": 3.8954, "step": 7393 }, { "epoch": 0.26054247381449475, "grad_norm": 1.6984918117523193, "learning_rate": 1.7349158420970003e-05, "loss": 3.1769, "step": 7394 }, { "epoch": 0.2605777108274532, "grad_norm": 3.366482734680176, "learning_rate": 1.7348384406752585e-05, "loss": 3.9418, "step": 7395 }, { "epoch": 0.26061294784041156, "grad_norm": 1.2338151931762695, "learning_rate": 1.734761029682152e-05, "loss": 1.0864, "step": 7396 }, { "epoch": 0.26064818485337, "grad_norm": 4.210331439971924, "learning_rate": 1.73468360911869e-05, "loss": 6.5418, "step": 7397 }, { "epoch": 0.26068342186632837, "grad_norm": 3.223715305328369, "learning_rate": 1.7346061789858802e-05, "loss": 4.0567, "step": 7398 }, { "epoch": 0.2607186588792868, "grad_norm": 2.371826648712158, "learning_rate": 1.7345287392847314e-05, "loss": 3.2883, "step": 7399 }, { "epoch": 0.26075389589224524, "grad_norm": 1.2507340908050537, "learning_rate": 1.734451290016252e-05, "loss": 0.7206, "step": 7400 }, { "epoch": 0.2607891329052036, "grad_norm": 3.365811824798584, "learning_rate": 1.7343738311814514e-05, "loss": 2.9773, "step": 7401 }, { "epoch": 0.26082436991816205, "grad_norm": 2.9737627506256104, "learning_rate": 1.734296362781338e-05, "loss": 3.0731, "step": 7402 }, { "epoch": 0.2608596069311204, "grad_norm": 3.717519998550415, "learning_rate": 1.7342188848169207e-05, "loss": 4.0901, "step": 7403 }, { "epoch": 0.26089484394407886, "grad_norm": 2.1385674476623535, "learning_rate": 1.734141397289209e-05, "loss": 3.186, "step": 7404 }, { "epoch": 0.2609300809570373, "grad_norm": 3.083536386489868, "learning_rate": 1.734063900199212e-05, "loss": 3.4651, "step": 7405 }, { "epoch": 0.26096531796999567, "grad_norm": 2.169404983520508, "learning_rate": 1.7339863935479394e-05, "loss": 0.7748, "step": 7406 }, { "epoch": 0.2610005549829541, "grad_norm": 2.3654181957244873, "learning_rate": 1.7339088773364007e-05, "loss": 3.4817, "step": 7407 }, { "epoch": 0.2610357919959125, "grad_norm": 3.4935595989227295, "learning_rate": 1.733831351565605e-05, "loss": 3.9175, "step": 7408 }, { "epoch": 0.2610710290088709, "grad_norm": 1.3992801904678345, "learning_rate": 1.7337538162365627e-05, "loss": 0.8934, "step": 7409 }, { "epoch": 0.26110626602182935, "grad_norm": 3.168339490890503, "learning_rate": 1.7336762713502834e-05, "loss": 3.0409, "step": 7410 }, { "epoch": 0.2611415030347877, "grad_norm": 1.9675283432006836, "learning_rate": 1.733598716907777e-05, "loss": 1.1093, "step": 7411 }, { "epoch": 0.26117674004774616, "grad_norm": 4.0269856452941895, "learning_rate": 1.733521152910054e-05, "loss": 3.6773, "step": 7412 }, { "epoch": 0.26121197706070454, "grad_norm": 1.1480978727340698, "learning_rate": 1.7334435793581245e-05, "loss": 0.9854, "step": 7413 }, { "epoch": 0.26124721407366297, "grad_norm": 3.1484103202819824, "learning_rate": 1.733365996252999e-05, "loss": 4.0978, "step": 7414 }, { "epoch": 0.2612824510866214, "grad_norm": 6.262365341186523, "learning_rate": 1.733288403595688e-05, "loss": 5.6973, "step": 7415 }, { "epoch": 0.2613176880995798, "grad_norm": 7.974423408508301, "learning_rate": 1.7332108013872023e-05, "loss": 6.2515, "step": 7416 }, { "epoch": 0.2613529251125382, "grad_norm": 2.0545787811279297, "learning_rate": 1.7331331896285524e-05, "loss": 2.8751, "step": 7417 }, { "epoch": 0.26138816212549665, "grad_norm": 1.7383662462234497, "learning_rate": 1.7330555683207492e-05, "loss": 1.2091, "step": 7418 }, { "epoch": 0.261423399138455, "grad_norm": 5.251300811767578, "learning_rate": 1.732977937464804e-05, "loss": 4.1488, "step": 7419 }, { "epoch": 0.26145863615141346, "grad_norm": 3.045621395111084, "learning_rate": 1.7329002970617275e-05, "loss": 3.6681, "step": 7420 }, { "epoch": 0.26149387316437184, "grad_norm": 3.8281924724578857, "learning_rate": 1.7328226471125318e-05, "loss": 6.2388, "step": 7421 }, { "epoch": 0.26152911017733027, "grad_norm": 5.271817684173584, "learning_rate": 1.7327449876182275e-05, "loss": 8.5607, "step": 7422 }, { "epoch": 0.2615643471902887, "grad_norm": 3.589639186859131, "learning_rate": 1.7326673185798263e-05, "loss": 4.9547, "step": 7423 }, { "epoch": 0.2615995842032471, "grad_norm": 1.537086009979248, "learning_rate": 1.7325896399983403e-05, "loss": 1.1292, "step": 7424 }, { "epoch": 0.2616348212162055, "grad_norm": 1.1213617324829102, "learning_rate": 1.7325119518747805e-05, "loss": 0.89, "step": 7425 }, { "epoch": 0.2616700582291639, "grad_norm": 0.9707454442977905, "learning_rate": 1.7324342542101592e-05, "loss": 0.9731, "step": 7426 }, { "epoch": 0.2617052952421223, "grad_norm": 4.087307453155518, "learning_rate": 1.732356547005489e-05, "loss": 3.7344, "step": 7427 }, { "epoch": 0.26174053225508076, "grad_norm": 0.9186748266220093, "learning_rate": 1.732278830261781e-05, "loss": 1.3242, "step": 7428 }, { "epoch": 0.26177576926803914, "grad_norm": 3.086538314819336, "learning_rate": 1.732201103980048e-05, "loss": 3.6654, "step": 7429 }, { "epoch": 0.26181100628099757, "grad_norm": 1.3424052000045776, "learning_rate": 1.7321233681613026e-05, "loss": 0.8833, "step": 7430 }, { "epoch": 0.26184624329395595, "grad_norm": 1.04458749294281, "learning_rate": 1.7320456228065567e-05, "loss": 0.9012, "step": 7431 }, { "epoch": 0.2618814803069144, "grad_norm": 1.8782294988632202, "learning_rate": 1.7319678679168235e-05, "loss": 0.7679, "step": 7432 }, { "epoch": 0.2619167173198728, "grad_norm": 5.182766914367676, "learning_rate": 1.7318901034931158e-05, "loss": 3.1625, "step": 7433 }, { "epoch": 0.2619519543328312, "grad_norm": 5.008943557739258, "learning_rate": 1.731812329536446e-05, "loss": 3.4073, "step": 7434 }, { "epoch": 0.2619871913457896, "grad_norm": 3.3790488243103027, "learning_rate": 1.7317345460478275e-05, "loss": 3.1814, "step": 7435 }, { "epoch": 0.262022428358748, "grad_norm": 0.7920143604278564, "learning_rate": 1.7316567530282733e-05, "loss": 0.865, "step": 7436 }, { "epoch": 0.26205766537170644, "grad_norm": 1.6579675674438477, "learning_rate": 1.7315789504787965e-05, "loss": 1.0325, "step": 7437 }, { "epoch": 0.26209290238466487, "grad_norm": 5.319125175476074, "learning_rate": 1.731501138400411e-05, "loss": 3.8485, "step": 7438 }, { "epoch": 0.26212813939762325, "grad_norm": 11.86490535736084, "learning_rate": 1.7314233167941298e-05, "loss": 5.9364, "step": 7439 }, { "epoch": 0.2621633764105817, "grad_norm": 0.88394695520401, "learning_rate": 1.7313454856609668e-05, "loss": 1.0456, "step": 7440 }, { "epoch": 0.26219861342354006, "grad_norm": 2.7276298999786377, "learning_rate": 1.7312676450019357e-05, "loss": 3.6882, "step": 7441 }, { "epoch": 0.2622338504364985, "grad_norm": 1.264732003211975, "learning_rate": 1.7311897948180504e-05, "loss": 0.6985, "step": 7442 }, { "epoch": 0.26226908744945693, "grad_norm": 3.799621105194092, "learning_rate": 1.7311119351103247e-05, "loss": 4.1002, "step": 7443 }, { "epoch": 0.2623043244624153, "grad_norm": 5.824164867401123, "learning_rate": 1.7310340658797732e-05, "loss": 6.3694, "step": 7444 }, { "epoch": 0.26233956147537374, "grad_norm": 3.263286828994751, "learning_rate": 1.7309561871274096e-05, "loss": 3.0496, "step": 7445 }, { "epoch": 0.2623747984883321, "grad_norm": 4.4925665855407715, "learning_rate": 1.7308782988542486e-05, "loss": 3.0401, "step": 7446 }, { "epoch": 0.26241003550129055, "grad_norm": 6.730453014373779, "learning_rate": 1.730800401061305e-05, "loss": 6.1827, "step": 7447 }, { "epoch": 0.262445272514249, "grad_norm": 3.9053947925567627, "learning_rate": 1.7307224937495925e-05, "loss": 3.7588, "step": 7448 }, { "epoch": 0.26248050952720736, "grad_norm": 4.201854228973389, "learning_rate": 1.730644576920127e-05, "loss": 3.6226, "step": 7449 }, { "epoch": 0.2625157465401658, "grad_norm": 1.4208922386169434, "learning_rate": 1.7305666505739225e-05, "loss": 1.0992, "step": 7450 }, { "epoch": 0.26255098355312423, "grad_norm": 1.162459135055542, "learning_rate": 1.7304887147119945e-05, "loss": 0.7312, "step": 7451 }, { "epoch": 0.2625862205660826, "grad_norm": 2.617143392562866, "learning_rate": 1.730410769335358e-05, "loss": 3.8498, "step": 7452 }, { "epoch": 0.26262145757904104, "grad_norm": 4.466925144195557, "learning_rate": 1.730332814445028e-05, "loss": 3.3287, "step": 7453 }, { "epoch": 0.2626566945919994, "grad_norm": 4.3262858390808105, "learning_rate": 1.73025485004202e-05, "loss": 3.1072, "step": 7454 }, { "epoch": 0.26269193160495785, "grad_norm": 1.661832332611084, "learning_rate": 1.7301768761273497e-05, "loss": 1.0414, "step": 7455 }, { "epoch": 0.2627271686179163, "grad_norm": 1.8375447988510132, "learning_rate": 1.7300988927020325e-05, "loss": 0.7286, "step": 7456 }, { "epoch": 0.26276240563087466, "grad_norm": 1.2545239925384521, "learning_rate": 1.730020899767084e-05, "loss": 0.8888, "step": 7457 }, { "epoch": 0.2627976426438331, "grad_norm": 4.031576156616211, "learning_rate": 1.7299428973235207e-05, "loss": 3.1257, "step": 7458 }, { "epoch": 0.2628328796567915, "grad_norm": 2.2274270057678223, "learning_rate": 1.729864885372358e-05, "loss": 0.991, "step": 7459 }, { "epoch": 0.2628681166697499, "grad_norm": 1.4344775676727295, "learning_rate": 1.7297868639146124e-05, "loss": 1.1063, "step": 7460 }, { "epoch": 0.26290335368270834, "grad_norm": 7.165262699127197, "learning_rate": 1.7297088329512996e-05, "loss": 3.6605, "step": 7461 }, { "epoch": 0.2629385906956667, "grad_norm": 3.97656512260437, "learning_rate": 1.7296307924834365e-05, "loss": 5.2741, "step": 7462 }, { "epoch": 0.26297382770862515, "grad_norm": 4.943676948547363, "learning_rate": 1.7295527425120396e-05, "loss": 3.0133, "step": 7463 }, { "epoch": 0.26300906472158353, "grad_norm": 2.471065044403076, "learning_rate": 1.7294746830381247e-05, "loss": 3.2508, "step": 7464 }, { "epoch": 0.26304430173454196, "grad_norm": 0.9920526146888733, "learning_rate": 1.7293966140627098e-05, "loss": 0.7707, "step": 7465 }, { "epoch": 0.2630795387475004, "grad_norm": 3.809540271759033, "learning_rate": 1.7293185355868107e-05, "loss": 3.5571, "step": 7466 }, { "epoch": 0.2631147757604588, "grad_norm": 6.661957740783691, "learning_rate": 1.729240447611445e-05, "loss": 5.823, "step": 7467 }, { "epoch": 0.2631500127734172, "grad_norm": 0.9692649245262146, "learning_rate": 1.7291623501376293e-05, "loss": 1.0036, "step": 7468 }, { "epoch": 0.2631852497863756, "grad_norm": 1.6541134119033813, "learning_rate": 1.729084243166381e-05, "loss": 1.305, "step": 7469 }, { "epoch": 0.263220486799334, "grad_norm": 2.6170639991760254, "learning_rate": 1.729006126698718e-05, "loss": 3.3029, "step": 7470 }, { "epoch": 0.26325572381229245, "grad_norm": 1.1715495586395264, "learning_rate": 1.7289280007356572e-05, "loss": 0.9016, "step": 7471 }, { "epoch": 0.26329096082525083, "grad_norm": 1.5869923830032349, "learning_rate": 1.7288498652782163e-05, "loss": 0.8869, "step": 7472 }, { "epoch": 0.26332619783820926, "grad_norm": 3.1536710262298584, "learning_rate": 1.7287717203274128e-05, "loss": 3.711, "step": 7473 }, { "epoch": 0.26336143485116764, "grad_norm": 5.338531970977783, "learning_rate": 1.728693565884265e-05, "loss": 3.2139, "step": 7474 }, { "epoch": 0.2633966718641261, "grad_norm": 5.438226222991943, "learning_rate": 1.728615401949791e-05, "loss": 6.1924, "step": 7475 }, { "epoch": 0.2634319088770845, "grad_norm": 8.092710494995117, "learning_rate": 1.728537228525008e-05, "loss": 5.8145, "step": 7476 }, { "epoch": 0.2634671458900429, "grad_norm": 3.0122387409210205, "learning_rate": 1.7284590456109353e-05, "loss": 3.2333, "step": 7477 }, { "epoch": 0.2635023829030013, "grad_norm": 3.6256327629089355, "learning_rate": 1.7283808532085905e-05, "loss": 3.52, "step": 7478 }, { "epoch": 0.2635376199159597, "grad_norm": 5.944387435913086, "learning_rate": 1.7283026513189925e-05, "loss": 3.6461, "step": 7479 }, { "epoch": 0.26357285692891813, "grad_norm": 9.165487289428711, "learning_rate": 1.7282244399431592e-05, "loss": 4.0177, "step": 7480 }, { "epoch": 0.26360809394187656, "grad_norm": 4.976189613342285, "learning_rate": 1.7281462190821107e-05, "loss": 3.5034, "step": 7481 }, { "epoch": 0.26364333095483494, "grad_norm": 8.161396026611328, "learning_rate": 1.7280679887368644e-05, "loss": 3.8878, "step": 7482 }, { "epoch": 0.2636785679677934, "grad_norm": 1.042938470840454, "learning_rate": 1.7279897489084395e-05, "loss": 1.1154, "step": 7483 }, { "epoch": 0.2637138049807518, "grad_norm": 1.1635867357254028, "learning_rate": 1.7279114995978558e-05, "loss": 0.8342, "step": 7484 }, { "epoch": 0.2637490419937102, "grad_norm": 1.261805772781372, "learning_rate": 1.7278332408061322e-05, "loss": 0.7318, "step": 7485 }, { "epoch": 0.2637842790066686, "grad_norm": 3.023761510848999, "learning_rate": 1.7277549725342874e-05, "loss": 3.3398, "step": 7486 }, { "epoch": 0.263819516019627, "grad_norm": 1.4584680795669556, "learning_rate": 1.727676694783342e-05, "loss": 0.8717, "step": 7487 }, { "epoch": 0.26385475303258543, "grad_norm": 5.670164108276367, "learning_rate": 1.7275984075543143e-05, "loss": 2.995, "step": 7488 }, { "epoch": 0.26388999004554387, "grad_norm": 9.801793098449707, "learning_rate": 1.727520110848225e-05, "loss": 3.4845, "step": 7489 }, { "epoch": 0.26392522705850224, "grad_norm": 10.277643203735352, "learning_rate": 1.7274418046660935e-05, "loss": 6.1243, "step": 7490 }, { "epoch": 0.2639604640714607, "grad_norm": 1.5056586265563965, "learning_rate": 1.72736348900894e-05, "loss": 0.9401, "step": 7491 }, { "epoch": 0.26399570108441905, "grad_norm": 3.154783248901367, "learning_rate": 1.7272851638777847e-05, "loss": 3.6472, "step": 7492 }, { "epoch": 0.2640309380973775, "grad_norm": 1.55813467502594, "learning_rate": 1.727206829273647e-05, "loss": 0.8321, "step": 7493 }, { "epoch": 0.2640661751103359, "grad_norm": 2.853736639022827, "learning_rate": 1.727128485197548e-05, "loss": 4.0146, "step": 7494 }, { "epoch": 0.2641014121232943, "grad_norm": 1.346564769744873, "learning_rate": 1.727050131650508e-05, "loss": 0.8442, "step": 7495 }, { "epoch": 0.26413664913625273, "grad_norm": 3.980710506439209, "learning_rate": 1.726971768633547e-05, "loss": 3.3804, "step": 7496 }, { "epoch": 0.2641718861492111, "grad_norm": 1.739660620689392, "learning_rate": 1.726893396147686e-05, "loss": 0.9312, "step": 7497 }, { "epoch": 0.26420712316216954, "grad_norm": 1.084905982017517, "learning_rate": 1.7268150141939464e-05, "loss": 0.8111, "step": 7498 }, { "epoch": 0.264242360175128, "grad_norm": 6.943846225738525, "learning_rate": 1.7267366227733485e-05, "loss": 2.8144, "step": 7499 }, { "epoch": 0.26427759718808636, "grad_norm": 2.065159559249878, "learning_rate": 1.7266582218869135e-05, "loss": 0.7101, "step": 7500 }, { "epoch": 0.2643128342010448, "grad_norm": 1.1361315250396729, "learning_rate": 1.7265798115356625e-05, "loss": 0.9288, "step": 7501 }, { "epoch": 0.26434807121400317, "grad_norm": 1.0710382461547852, "learning_rate": 1.7265013917206168e-05, "loss": 1.3212, "step": 7502 }, { "epoch": 0.2643833082269616, "grad_norm": 1.42022705078125, "learning_rate": 1.7264229624427982e-05, "loss": 0.9961, "step": 7503 }, { "epoch": 0.26441854523992003, "grad_norm": 5.417186260223389, "learning_rate": 1.7263445237032278e-05, "loss": 3.4843, "step": 7504 }, { "epoch": 0.2644537822528784, "grad_norm": 2.449763059616089, "learning_rate": 1.7262660755029276e-05, "loss": 3.0813, "step": 7505 }, { "epoch": 0.26448901926583684, "grad_norm": 1.1281942129135132, "learning_rate": 1.726187617842919e-05, "loss": 1.012, "step": 7506 }, { "epoch": 0.2645242562787952, "grad_norm": 4.556528091430664, "learning_rate": 1.7261091507242243e-05, "loss": 3.5328, "step": 7507 }, { "epoch": 0.26455949329175366, "grad_norm": 0.9725498557090759, "learning_rate": 1.7260306741478657e-05, "loss": 0.7629, "step": 7508 }, { "epoch": 0.2645947303047121, "grad_norm": 4.049869537353516, "learning_rate": 1.725952188114865e-05, "loss": 2.996, "step": 7509 }, { "epoch": 0.26462996731767047, "grad_norm": 1.2466074228286743, "learning_rate": 1.7258736926262444e-05, "loss": 0.8438, "step": 7510 }, { "epoch": 0.2646652043306289, "grad_norm": 1.6277954578399658, "learning_rate": 1.7257951876830264e-05, "loss": 1.3263, "step": 7511 }, { "epoch": 0.2647004413435873, "grad_norm": 2.6016087532043457, "learning_rate": 1.7257166732862337e-05, "loss": 3.1611, "step": 7512 }, { "epoch": 0.2647356783565457, "grad_norm": 1.3028470277786255, "learning_rate": 1.7256381494368894e-05, "loss": 0.9233, "step": 7513 }, { "epoch": 0.26477091536950415, "grad_norm": 1.6171441078186035, "learning_rate": 1.725559616136015e-05, "loss": 1.159, "step": 7514 }, { "epoch": 0.2648061523824625, "grad_norm": 5.475011825561523, "learning_rate": 1.7254810733846347e-05, "loss": 2.9839, "step": 7515 }, { "epoch": 0.26484138939542096, "grad_norm": 1.2725427150726318, "learning_rate": 1.7254025211837705e-05, "loss": 1.0303, "step": 7516 }, { "epoch": 0.26487662640837933, "grad_norm": 0.9690940976142883, "learning_rate": 1.7253239595344465e-05, "loss": 0.9269, "step": 7517 }, { "epoch": 0.26491186342133777, "grad_norm": 11.871789932250977, "learning_rate": 1.7252453884376854e-05, "loss": 3.9286, "step": 7518 }, { "epoch": 0.2649471004342962, "grad_norm": 5.238433837890625, "learning_rate": 1.7251668078945106e-05, "loss": 3.7307, "step": 7519 }, { "epoch": 0.2649823374472546, "grad_norm": 4.859691143035889, "learning_rate": 1.7250882179059464e-05, "loss": 2.624, "step": 7520 }, { "epoch": 0.265017574460213, "grad_norm": 3.9002294540405273, "learning_rate": 1.7250096184730153e-05, "loss": 3.5768, "step": 7521 }, { "epoch": 0.26505281147317145, "grad_norm": 1.2263643741607666, "learning_rate": 1.724931009596741e-05, "loss": 0.9729, "step": 7522 }, { "epoch": 0.2650880484861298, "grad_norm": 8.120265007019043, "learning_rate": 1.724852391278149e-05, "loss": 3.2812, "step": 7523 }, { "epoch": 0.26512328549908826, "grad_norm": 6.324637413024902, "learning_rate": 1.7247737635182616e-05, "loss": 3.8535, "step": 7524 }, { "epoch": 0.26515852251204663, "grad_norm": 9.784055709838867, "learning_rate": 1.7246951263181036e-05, "loss": 5.9192, "step": 7525 }, { "epoch": 0.26519375952500507, "grad_norm": 3.804884910583496, "learning_rate": 1.7246164796786996e-05, "loss": 3.8421, "step": 7526 }, { "epoch": 0.2652289965379635, "grad_norm": 2.699261426925659, "learning_rate": 1.7245378236010737e-05, "loss": 2.4544, "step": 7527 }, { "epoch": 0.2652642335509219, "grad_norm": 1.4919272661209106, "learning_rate": 1.7244591580862503e-05, "loss": 0.9191, "step": 7528 }, { "epoch": 0.2652994705638803, "grad_norm": 4.947016716003418, "learning_rate": 1.7243804831352538e-05, "loss": 6.928, "step": 7529 }, { "epoch": 0.2653347075768387, "grad_norm": 4.285496711730957, "learning_rate": 1.7243017987491092e-05, "loss": 5.821, "step": 7530 }, { "epoch": 0.2653699445897971, "grad_norm": 4.216891765594482, "learning_rate": 1.7242231049288417e-05, "loss": 3.8111, "step": 7531 }, { "epoch": 0.26540518160275556, "grad_norm": 1.8751964569091797, "learning_rate": 1.724144401675476e-05, "loss": 0.936, "step": 7532 }, { "epoch": 0.26544041861571394, "grad_norm": 2.8676259517669678, "learning_rate": 1.7240656889900372e-05, "loss": 1.2032, "step": 7533 }, { "epoch": 0.26547565562867237, "grad_norm": 6.230831146240234, "learning_rate": 1.7239869668735506e-05, "loss": 6.1058, "step": 7534 }, { "epoch": 0.26551089264163075, "grad_norm": 1.095024585723877, "learning_rate": 1.7239082353270415e-05, "loss": 1.2583, "step": 7535 }, { "epoch": 0.2655461296545892, "grad_norm": 3.231919527053833, "learning_rate": 1.7238294943515354e-05, "loss": 3.3346, "step": 7536 }, { "epoch": 0.2655813666675476, "grad_norm": 3.3964502811431885, "learning_rate": 1.723750743948058e-05, "loss": 3.1661, "step": 7537 }, { "epoch": 0.265616603680506, "grad_norm": 5.379337787628174, "learning_rate": 1.7236719841176347e-05, "loss": 6.9088, "step": 7538 }, { "epoch": 0.2656518406934644, "grad_norm": 1.3141100406646729, "learning_rate": 1.7235932148612916e-05, "loss": 0.74, "step": 7539 }, { "epoch": 0.2656870777064228, "grad_norm": 2.081468343734741, "learning_rate": 1.723514436180055e-05, "loss": 2.7455, "step": 7540 }, { "epoch": 0.26572231471938124, "grad_norm": 1.5473952293395996, "learning_rate": 1.723435648074951e-05, "loss": 1.0741, "step": 7541 }, { "epoch": 0.26575755173233967, "grad_norm": 4.310730934143066, "learning_rate": 1.7233568505470048e-05, "loss": 3.1693, "step": 7542 }, { "epoch": 0.26579278874529805, "grad_norm": 2.79180908203125, "learning_rate": 1.723278043597244e-05, "loss": 3.2728, "step": 7543 }, { "epoch": 0.2658280257582565, "grad_norm": 0.8396500945091248, "learning_rate": 1.7231992272266945e-05, "loss": 0.8495, "step": 7544 }, { "epoch": 0.26586326277121486, "grad_norm": 2.420283317565918, "learning_rate": 1.723120401436383e-05, "loss": 0.8017, "step": 7545 }, { "epoch": 0.2658984997841733, "grad_norm": 1.1582105159759521, "learning_rate": 1.7230415662273355e-05, "loss": 0.9422, "step": 7546 }, { "epoch": 0.2659337367971317, "grad_norm": 2.3499562740325928, "learning_rate": 1.72296272160058e-05, "loss": 3.6213, "step": 7547 }, { "epoch": 0.2659689738100901, "grad_norm": 1.0147647857666016, "learning_rate": 1.722883867557143e-05, "loss": 0.9287, "step": 7548 }, { "epoch": 0.26600421082304854, "grad_norm": 0.9656612277030945, "learning_rate": 1.7228050040980516e-05, "loss": 0.7768, "step": 7549 }, { "epoch": 0.2660394478360069, "grad_norm": 3.9425132274627686, "learning_rate": 1.7227261312243328e-05, "loss": 2.5467, "step": 7550 }, { "epoch": 0.26607468484896535, "grad_norm": 6.731571674346924, "learning_rate": 1.722647248937014e-05, "loss": 6.6372, "step": 7551 }, { "epoch": 0.2661099218619238, "grad_norm": 6.5310564041137695, "learning_rate": 1.722568357237123e-05, "loss": 8.3215, "step": 7552 }, { "epoch": 0.26614515887488216, "grad_norm": 5.261713981628418, "learning_rate": 1.722489456125687e-05, "loss": 3.1625, "step": 7553 }, { "epoch": 0.2661803958878406, "grad_norm": 1.673874020576477, "learning_rate": 1.722410545603734e-05, "loss": 1.0545, "step": 7554 }, { "epoch": 0.266215632900799, "grad_norm": 1.296081781387329, "learning_rate": 1.7223316256722912e-05, "loss": 1.0282, "step": 7555 }, { "epoch": 0.2662508699137574, "grad_norm": 3.7043445110321045, "learning_rate": 1.722252696332387e-05, "loss": 3.5714, "step": 7556 }, { "epoch": 0.26628610692671584, "grad_norm": 0.9351409077644348, "learning_rate": 1.7221737575850497e-05, "loss": 1.0394, "step": 7557 }, { "epoch": 0.2663213439396742, "grad_norm": 4.383404731750488, "learning_rate": 1.722094809431307e-05, "loss": 4.0052, "step": 7558 }, { "epoch": 0.26635658095263265, "grad_norm": 1.1892976760864258, "learning_rate": 1.7220158518721876e-05, "loss": 0.6584, "step": 7559 }, { "epoch": 0.2663918179655911, "grad_norm": 1.136171817779541, "learning_rate": 1.7219368849087198e-05, "loss": 0.8718, "step": 7560 }, { "epoch": 0.26642705497854946, "grad_norm": 2.1662344932556152, "learning_rate": 1.721857908541932e-05, "loss": 0.8344, "step": 7561 }, { "epoch": 0.2664622919915079, "grad_norm": 7.300212383270264, "learning_rate": 1.721778922772853e-05, "loss": 6.0812, "step": 7562 }, { "epoch": 0.26649752900446627, "grad_norm": 1.3053486347198486, "learning_rate": 1.7216999276025115e-05, "loss": 0.7919, "step": 7563 }, { "epoch": 0.2665327660174247, "grad_norm": 1.225570797920227, "learning_rate": 1.721620923031937e-05, "loss": 0.866, "step": 7564 }, { "epoch": 0.26656800303038314, "grad_norm": 3.6553432941436768, "learning_rate": 1.7215419090621575e-05, "loss": 2.8917, "step": 7565 }, { "epoch": 0.2666032400433415, "grad_norm": 1.46229887008667, "learning_rate": 1.721462885694203e-05, "loss": 0.7969, "step": 7566 }, { "epoch": 0.26663847705629995, "grad_norm": 3.5153305530548096, "learning_rate": 1.721383852929102e-05, "loss": 3.3209, "step": 7567 }, { "epoch": 0.2666737140692583, "grad_norm": 1.2148431539535522, "learning_rate": 1.721304810767885e-05, "loss": 0.9676, "step": 7568 }, { "epoch": 0.26670895108221676, "grad_norm": 1.360707402229309, "learning_rate": 1.7212257592115807e-05, "loss": 0.8087, "step": 7569 }, { "epoch": 0.2667441880951752, "grad_norm": 1.4041630029678345, "learning_rate": 1.721146698261219e-05, "loss": 0.7653, "step": 7570 }, { "epoch": 0.26677942510813357, "grad_norm": 2.7339069843292236, "learning_rate": 1.72106762791783e-05, "loss": 3.4091, "step": 7571 }, { "epoch": 0.266814662121092, "grad_norm": 1.3809781074523926, "learning_rate": 1.7209885481824428e-05, "loss": 0.6678, "step": 7572 }, { "epoch": 0.2668498991340504, "grad_norm": 1.3861162662506104, "learning_rate": 1.7209094590560883e-05, "loss": 1.1351, "step": 7573 }, { "epoch": 0.2668851361470088, "grad_norm": 2.570958375930786, "learning_rate": 1.7208303605397962e-05, "loss": 3.5054, "step": 7574 }, { "epoch": 0.26692037315996725, "grad_norm": 5.697988986968994, "learning_rate": 1.720751252634597e-05, "loss": 3.3944, "step": 7575 }, { "epoch": 0.26695561017292563, "grad_norm": 5.083930492401123, "learning_rate": 1.7206721353415207e-05, "loss": 6.466, "step": 7576 }, { "epoch": 0.26699084718588406, "grad_norm": 1.4057978391647339, "learning_rate": 1.720593008661598e-05, "loss": 0.9107, "step": 7577 }, { "epoch": 0.26702608419884244, "grad_norm": 1.2567356824874878, "learning_rate": 1.7205138725958596e-05, "loss": 0.9689, "step": 7578 }, { "epoch": 0.2670613212118009, "grad_norm": 1.274505376815796, "learning_rate": 1.7204347271453364e-05, "loss": 0.8673, "step": 7579 }, { "epoch": 0.2670965582247593, "grad_norm": 8.497979164123535, "learning_rate": 1.720355572311059e-05, "loss": 3.1553, "step": 7580 }, { "epoch": 0.2671317952377177, "grad_norm": 1.0968834161758423, "learning_rate": 1.7202764080940584e-05, "loss": 0.8594, "step": 7581 }, { "epoch": 0.2671670322506761, "grad_norm": 3.421618938446045, "learning_rate": 1.7201972344953662e-05, "loss": 3.7175, "step": 7582 }, { "epoch": 0.2672022692636345, "grad_norm": 5.150981903076172, "learning_rate": 1.720118051516013e-05, "loss": 3.8281, "step": 7583 }, { "epoch": 0.26723750627659293, "grad_norm": 1.184883952140808, "learning_rate": 1.7200388591570303e-05, "loss": 1.1888, "step": 7584 }, { "epoch": 0.26727274328955136, "grad_norm": 1.1323626041412354, "learning_rate": 1.71995965741945e-05, "loss": 0.775, "step": 7585 }, { "epoch": 0.26730798030250974, "grad_norm": 1.2374953031539917, "learning_rate": 1.7198804463043032e-05, "loss": 0.7602, "step": 7586 }, { "epoch": 0.2673432173154682, "grad_norm": 4.076559543609619, "learning_rate": 1.7198012258126223e-05, "loss": 4.2987, "step": 7587 }, { "epoch": 0.2673784543284266, "grad_norm": 1.3412553071975708, "learning_rate": 1.7197219959454386e-05, "loss": 1.0967, "step": 7588 }, { "epoch": 0.267413691341385, "grad_norm": 4.400155544281006, "learning_rate": 1.719642756703784e-05, "loss": 3.2136, "step": 7589 }, { "epoch": 0.2674489283543434, "grad_norm": 3.9466192722320557, "learning_rate": 1.7195635080886907e-05, "loss": 3.0858, "step": 7590 }, { "epoch": 0.2674841653673018, "grad_norm": 2.51815128326416, "learning_rate": 1.7194842501011916e-05, "loss": 3.3129, "step": 7591 }, { "epoch": 0.26751940238026023, "grad_norm": 3.5750434398651123, "learning_rate": 1.719404982742318e-05, "loss": 5.939, "step": 7592 }, { "epoch": 0.26755463939321866, "grad_norm": 3.035640001296997, "learning_rate": 1.719325706013103e-05, "loss": 3.949, "step": 7593 }, { "epoch": 0.26758987640617704, "grad_norm": 3.1806862354278564, "learning_rate": 1.7192464199145794e-05, "loss": 5.1455, "step": 7594 }, { "epoch": 0.2676251134191355, "grad_norm": 1.420389175415039, "learning_rate": 1.719167124447779e-05, "loss": 1.2417, "step": 7595 }, { "epoch": 0.26766035043209385, "grad_norm": 1.3157576322555542, "learning_rate": 1.7190878196137355e-05, "loss": 0.9986, "step": 7596 }, { "epoch": 0.2676955874450523, "grad_norm": 2.9079394340515137, "learning_rate": 1.7190085054134816e-05, "loss": 2.8675, "step": 7597 }, { "epoch": 0.2677308244580107, "grad_norm": 1.159990906715393, "learning_rate": 1.7189291818480504e-05, "loss": 0.9316, "step": 7598 }, { "epoch": 0.2677660614709691, "grad_norm": 2.5511374473571777, "learning_rate": 1.7188498489184747e-05, "loss": 0.9858, "step": 7599 }, { "epoch": 0.26780129848392753, "grad_norm": 2.986172914505005, "learning_rate": 1.7187705066257884e-05, "loss": 3.3012, "step": 7600 }, { "epoch": 0.2678365354968859, "grad_norm": 1.5355010032653809, "learning_rate": 1.7186911549710247e-05, "loss": 0.8916, "step": 7601 }, { "epoch": 0.26787177250984434, "grad_norm": 2.062457323074341, "learning_rate": 1.718611793955217e-05, "loss": 1.0034, "step": 7602 }, { "epoch": 0.2679070095228028, "grad_norm": 1.1743348836898804, "learning_rate": 1.718532423579399e-05, "loss": 0.6572, "step": 7603 }, { "epoch": 0.26794224653576115, "grad_norm": 1.164538860321045, "learning_rate": 1.7184530438446053e-05, "loss": 0.9014, "step": 7604 }, { "epoch": 0.2679774835487196, "grad_norm": 3.2672457695007324, "learning_rate": 1.7183736547518687e-05, "loss": 3.4121, "step": 7605 }, { "epoch": 0.26801272056167796, "grad_norm": 1.0654066801071167, "learning_rate": 1.7182942563022237e-05, "loss": 0.9793, "step": 7606 }, { "epoch": 0.2680479575746364, "grad_norm": 1.0347983837127686, "learning_rate": 1.7182148484967048e-05, "loss": 1.1187, "step": 7607 }, { "epoch": 0.26808319458759483, "grad_norm": 2.162602424621582, "learning_rate": 1.7181354313363456e-05, "loss": 3.3729, "step": 7608 }, { "epoch": 0.2681184316005532, "grad_norm": 4.005532264709473, "learning_rate": 1.7180560048221813e-05, "loss": 3.3138, "step": 7609 }, { "epoch": 0.26815366861351164, "grad_norm": 2.985276222229004, "learning_rate": 1.717976568955246e-05, "loss": 3.0557, "step": 7610 }, { "epoch": 0.26818890562647, "grad_norm": 1.7597109079360962, "learning_rate": 1.7178971237365744e-05, "loss": 1.2168, "step": 7611 }, { "epoch": 0.26822414263942845, "grad_norm": 4.328654766082764, "learning_rate": 1.717817669167201e-05, "loss": 5.398, "step": 7612 }, { "epoch": 0.2682593796523869, "grad_norm": 4.879034519195557, "learning_rate": 1.7177382052481616e-05, "loss": 6.2238, "step": 7613 }, { "epoch": 0.26829461666534526, "grad_norm": 1.5886478424072266, "learning_rate": 1.7176587319804904e-05, "loss": 0.8229, "step": 7614 }, { "epoch": 0.2683298536783037, "grad_norm": 5.1415114402771, "learning_rate": 1.7175792493652225e-05, "loss": 6.6308, "step": 7615 }, { "epoch": 0.2683650906912621, "grad_norm": 1.4834944009780884, "learning_rate": 1.717499757403394e-05, "loss": 0.8799, "step": 7616 }, { "epoch": 0.2684003277042205, "grad_norm": 4.483880996704102, "learning_rate": 1.7174202560960395e-05, "loss": 3.7718, "step": 7617 }, { "epoch": 0.26843556471717894, "grad_norm": 1.4302334785461426, "learning_rate": 1.7173407454441944e-05, "loss": 0.9608, "step": 7618 }, { "epoch": 0.2684708017301373, "grad_norm": 1.2118556499481201, "learning_rate": 1.7172612254488953e-05, "loss": 1.061, "step": 7619 }, { "epoch": 0.26850603874309575, "grad_norm": 1.8208556175231934, "learning_rate": 1.7171816961111772e-05, "loss": 1.0216, "step": 7620 }, { "epoch": 0.26854127575605413, "grad_norm": 1.1419813632965088, "learning_rate": 1.7171021574320762e-05, "loss": 1.1967, "step": 7621 }, { "epoch": 0.26857651276901257, "grad_norm": 2.8623855113983154, "learning_rate": 1.7170226094126284e-05, "loss": 3.7495, "step": 7622 }, { "epoch": 0.268611749781971, "grad_norm": 4.134228706359863, "learning_rate": 1.7169430520538694e-05, "loss": 5.353, "step": 7623 }, { "epoch": 0.2686469867949294, "grad_norm": 2.342632293701172, "learning_rate": 1.7168634853568362e-05, "loss": 2.5244, "step": 7624 }, { "epoch": 0.2686822238078878, "grad_norm": 0.7260568737983704, "learning_rate": 1.7167839093225644e-05, "loss": 1.0312, "step": 7625 }, { "epoch": 0.26871746082084624, "grad_norm": 2.5362210273742676, "learning_rate": 1.716704323952091e-05, "loss": 2.8441, "step": 7626 }, { "epoch": 0.2687526978338046, "grad_norm": 2.344420909881592, "learning_rate": 1.7166247292464526e-05, "loss": 3.8107, "step": 7627 }, { "epoch": 0.26878793484676305, "grad_norm": 2.0908493995666504, "learning_rate": 1.7165451252066857e-05, "loss": 0.9311, "step": 7628 }, { "epoch": 0.26882317185972143, "grad_norm": 3.5772173404693604, "learning_rate": 1.7164655118338274e-05, "loss": 3.487, "step": 7629 }, { "epoch": 0.26885840887267987, "grad_norm": 1.9343552589416504, "learning_rate": 1.7163858891289143e-05, "loss": 1.0736, "step": 7630 }, { "epoch": 0.2688936458856383, "grad_norm": 1.4004783630371094, "learning_rate": 1.7163062570929844e-05, "loss": 1.0232, "step": 7631 }, { "epoch": 0.2689288828985967, "grad_norm": 4.225486755371094, "learning_rate": 1.7162266157270735e-05, "loss": 3.6334, "step": 7632 }, { "epoch": 0.2689641199115551, "grad_norm": 1.610440969467163, "learning_rate": 1.71614696503222e-05, "loss": 1.117, "step": 7633 }, { "epoch": 0.2689993569245135, "grad_norm": 3.0552115440368652, "learning_rate": 1.7160673050094612e-05, "loss": 2.9506, "step": 7634 }, { "epoch": 0.2690345939374719, "grad_norm": 3.2029664516448975, "learning_rate": 1.7159876356598342e-05, "loss": 3.9306, "step": 7635 }, { "epoch": 0.26906983095043036, "grad_norm": 4.039170265197754, "learning_rate": 1.7159079569843772e-05, "loss": 3.6866, "step": 7636 }, { "epoch": 0.26910506796338873, "grad_norm": 3.47495698928833, "learning_rate": 1.715828268984128e-05, "loss": 3.4911, "step": 7637 }, { "epoch": 0.26914030497634717, "grad_norm": 3.402021646499634, "learning_rate": 1.7157485716601243e-05, "loss": 5.9559, "step": 7638 }, { "epoch": 0.26917554198930554, "grad_norm": 4.5197649002075195, "learning_rate": 1.7156688650134042e-05, "loss": 3.3923, "step": 7639 }, { "epoch": 0.269210779002264, "grad_norm": 3.4511938095092773, "learning_rate": 1.7155891490450062e-05, "loss": 3.5549, "step": 7640 }, { "epoch": 0.2692460160152224, "grad_norm": 5.3081135749816895, "learning_rate": 1.7155094237559682e-05, "loss": 3.8088, "step": 7641 }, { "epoch": 0.2692812530281808, "grad_norm": 0.977003276348114, "learning_rate": 1.7154296891473288e-05, "loss": 1.3599, "step": 7642 }, { "epoch": 0.2693164900411392, "grad_norm": 5.677283763885498, "learning_rate": 1.7153499452201266e-05, "loss": 3.5351, "step": 7643 }, { "epoch": 0.2693517270540976, "grad_norm": 0.9314908385276794, "learning_rate": 1.7152701919754004e-05, "loss": 1.0612, "step": 7644 }, { "epoch": 0.26938696406705603, "grad_norm": 1.705625057220459, "learning_rate": 1.7151904294141886e-05, "loss": 0.7881, "step": 7645 }, { "epoch": 0.26942220108001447, "grad_norm": 1.395104169845581, "learning_rate": 1.7151106575375304e-05, "loss": 1.028, "step": 7646 }, { "epoch": 0.26945743809297285, "grad_norm": 1.4854179620742798, "learning_rate": 1.715030876346465e-05, "loss": 0.9401, "step": 7647 }, { "epoch": 0.2694926751059313, "grad_norm": 2.142068862915039, "learning_rate": 1.7149510858420312e-05, "loss": 2.7719, "step": 7648 }, { "epoch": 0.26952791211888966, "grad_norm": 7.093265533447266, "learning_rate": 1.7148712860252682e-05, "loss": 5.6518, "step": 7649 }, { "epoch": 0.2695631491318481, "grad_norm": 7.01377010345459, "learning_rate": 1.714791476897216e-05, "loss": 5.2953, "step": 7650 }, { "epoch": 0.2695983861448065, "grad_norm": 5.509644508361816, "learning_rate": 1.7147116584589137e-05, "loss": 6.1248, "step": 7651 }, { "epoch": 0.2696336231577649, "grad_norm": 3.7994441986083984, "learning_rate": 1.714631830711401e-05, "loss": 3.4447, "step": 7652 }, { "epoch": 0.26966886017072333, "grad_norm": 3.113832473754883, "learning_rate": 1.7145519936557176e-05, "loss": 3.4592, "step": 7653 }, { "epoch": 0.2697040971836817, "grad_norm": 1.2998125553131104, "learning_rate": 1.7144721472929035e-05, "loss": 0.8922, "step": 7654 }, { "epoch": 0.26973933419664015, "grad_norm": 1.9399534463882446, "learning_rate": 1.7143922916239985e-05, "loss": 1.1052, "step": 7655 }, { "epoch": 0.2697745712095986, "grad_norm": 6.237707138061523, "learning_rate": 1.7143124266500432e-05, "loss": 3.9538, "step": 7656 }, { "epoch": 0.26980980822255696, "grad_norm": 3.41911244392395, "learning_rate": 1.7142325523720775e-05, "loss": 3.2257, "step": 7657 }, { "epoch": 0.2698450452355154, "grad_norm": 4.294223308563232, "learning_rate": 1.7141526687911418e-05, "loss": 2.7241, "step": 7658 }, { "epoch": 0.2698802822484738, "grad_norm": 1.686974287033081, "learning_rate": 1.7140727759082766e-05, "loss": 0.8646, "step": 7659 }, { "epoch": 0.2699155192614322, "grad_norm": 4.741665840148926, "learning_rate": 1.7139928737245224e-05, "loss": 3.7341, "step": 7660 }, { "epoch": 0.26995075627439064, "grad_norm": 10.486536979675293, "learning_rate": 1.71391296224092e-05, "loss": 6.4004, "step": 7661 }, { "epoch": 0.269985993287349, "grad_norm": 2.5078885555267334, "learning_rate": 1.7138330414585108e-05, "loss": 3.1093, "step": 7662 }, { "epoch": 0.27002123030030745, "grad_norm": 1.3025866746902466, "learning_rate": 1.7137531113783347e-05, "loss": 1.0041, "step": 7663 }, { "epoch": 0.2700564673132659, "grad_norm": 5.119555950164795, "learning_rate": 1.713673172001434e-05, "loss": 5.8591, "step": 7664 }, { "epoch": 0.27009170432622426, "grad_norm": 3.5544211864471436, "learning_rate": 1.7135932233288487e-05, "loss": 4.1096, "step": 7665 }, { "epoch": 0.2701269413391827, "grad_norm": 3.1100876331329346, "learning_rate": 1.7135132653616215e-05, "loss": 3.2367, "step": 7666 }, { "epoch": 0.27016217835214107, "grad_norm": 1.0829209089279175, "learning_rate": 1.7134332981007925e-05, "loss": 0.9338, "step": 7667 }, { "epoch": 0.2701974153650995, "grad_norm": 4.525918483734131, "learning_rate": 1.7133533215474043e-05, "loss": 3.0237, "step": 7668 }, { "epoch": 0.27023265237805794, "grad_norm": 3.79953670501709, "learning_rate": 1.713273335702498e-05, "loss": 3.9259, "step": 7669 }, { "epoch": 0.2702678893910163, "grad_norm": 1.6462843418121338, "learning_rate": 1.7131933405671158e-05, "loss": 0.9085, "step": 7670 }, { "epoch": 0.27030312640397475, "grad_norm": 2.742175340652466, "learning_rate": 1.7131133361422995e-05, "loss": 3.4948, "step": 7671 }, { "epoch": 0.2703383634169331, "grad_norm": 1.7364014387130737, "learning_rate": 1.713033322429091e-05, "loss": 0.9698, "step": 7672 }, { "epoch": 0.27037360042989156, "grad_norm": 1.0024548768997192, "learning_rate": 1.7129532994285326e-05, "loss": 1.1412, "step": 7673 }, { "epoch": 0.27040883744285, "grad_norm": 4.95011568069458, "learning_rate": 1.712873267141667e-05, "loss": 6.068, "step": 7674 }, { "epoch": 0.27044407445580837, "grad_norm": 0.9077582955360413, "learning_rate": 1.712793225569536e-05, "loss": 0.8835, "step": 7675 }, { "epoch": 0.2704793114687668, "grad_norm": 2.9469306468963623, "learning_rate": 1.7127131747131824e-05, "loss": 3.3559, "step": 7676 }, { "epoch": 0.2705145484817252, "grad_norm": 1.6581323146820068, "learning_rate": 1.712633114573649e-05, "loss": 1.1093, "step": 7677 }, { "epoch": 0.2705497854946836, "grad_norm": 2.1885547637939453, "learning_rate": 1.712553045151979e-05, "loss": 0.8741, "step": 7678 }, { "epoch": 0.27058502250764205, "grad_norm": 3.207545518875122, "learning_rate": 1.712472966449214e-05, "loss": 3.3792, "step": 7679 }, { "epoch": 0.2706202595206004, "grad_norm": 2.2865724563598633, "learning_rate": 1.7123928784663984e-05, "loss": 0.8415, "step": 7680 }, { "epoch": 0.27065549653355886, "grad_norm": 4.036205768585205, "learning_rate": 1.7123127812045742e-05, "loss": 3.2845, "step": 7681 }, { "epoch": 0.27069073354651724, "grad_norm": 3.9333372116088867, "learning_rate": 1.7122326746647862e-05, "loss": 2.8796, "step": 7682 }, { "epoch": 0.27072597055947567, "grad_norm": 1.5133686065673828, "learning_rate": 1.7121525588480762e-05, "loss": 0.8293, "step": 7683 }, { "epoch": 0.2707612075724341, "grad_norm": 1.4563547372817993, "learning_rate": 1.7120724337554887e-05, "loss": 1.0204, "step": 7684 }, { "epoch": 0.2707964445853925, "grad_norm": 4.23810338973999, "learning_rate": 1.711992299388067e-05, "loss": 2.3794, "step": 7685 }, { "epoch": 0.2708316815983509, "grad_norm": 5.529513835906982, "learning_rate": 1.711912155746855e-05, "loss": 5.982, "step": 7686 }, { "epoch": 0.2708669186113093, "grad_norm": 3.8681039810180664, "learning_rate": 1.711832002832896e-05, "loss": 2.9592, "step": 7687 }, { "epoch": 0.2709021556242677, "grad_norm": 1.1826590299606323, "learning_rate": 1.711751840647235e-05, "loss": 1.0722, "step": 7688 }, { "epoch": 0.27093739263722616, "grad_norm": 4.487457752227783, "learning_rate": 1.7116716691909158e-05, "loss": 3.6037, "step": 7689 }, { "epoch": 0.27097262965018454, "grad_norm": 4.333584785461426, "learning_rate": 1.7115914884649818e-05, "loss": 3.9499, "step": 7690 }, { "epoch": 0.27100786666314297, "grad_norm": 5.344222068786621, "learning_rate": 1.7115112984704784e-05, "loss": 4.9171, "step": 7691 }, { "epoch": 0.2710431036761014, "grad_norm": 5.31188440322876, "learning_rate": 1.7114310992084497e-05, "loss": 6.9826, "step": 7692 }, { "epoch": 0.2710783406890598, "grad_norm": 4.611670017242432, "learning_rate": 1.7113508906799402e-05, "loss": 4.2859, "step": 7693 }, { "epoch": 0.2711135777020182, "grad_norm": 1.1850669384002686, "learning_rate": 1.7112706728859948e-05, "loss": 1.0915, "step": 7694 }, { "epoch": 0.2711488147149766, "grad_norm": 1.1509829759597778, "learning_rate": 1.7111904458276584e-05, "loss": 1.3119, "step": 7695 }, { "epoch": 0.271184051727935, "grad_norm": 1.2667688131332397, "learning_rate": 1.7111102095059755e-05, "loss": 1.1686, "step": 7696 }, { "epoch": 0.27121928874089346, "grad_norm": 2.7522857189178467, "learning_rate": 1.711029963921992e-05, "loss": 3.4316, "step": 7697 }, { "epoch": 0.27125452575385184, "grad_norm": 4.45460844039917, "learning_rate": 1.7109497090767523e-05, "loss": 3.3429, "step": 7698 }, { "epoch": 0.27128976276681027, "grad_norm": 0.9426041841506958, "learning_rate": 1.7108694449713022e-05, "loss": 0.718, "step": 7699 }, { "epoch": 0.27132499977976865, "grad_norm": 3.382840871810913, "learning_rate": 1.710789171606687e-05, "loss": 3.4654, "step": 7700 }, { "epoch": 0.2713602367927271, "grad_norm": 5.92744779586792, "learning_rate": 1.7107088889839524e-05, "loss": 5.5401, "step": 7701 }, { "epoch": 0.2713954738056855, "grad_norm": 6.239470958709717, "learning_rate": 1.710628597104144e-05, "loss": 5.1095, "step": 7702 }, { "epoch": 0.2714307108186439, "grad_norm": 5.556354522705078, "learning_rate": 1.7105482959683076e-05, "loss": 3.5175, "step": 7703 }, { "epoch": 0.27146594783160233, "grad_norm": 3.0855720043182373, "learning_rate": 1.7104679855774888e-05, "loss": 6.4915, "step": 7704 }, { "epoch": 0.2715011848445607, "grad_norm": 2.6483352184295654, "learning_rate": 1.7103876659327344e-05, "loss": 3.2472, "step": 7705 }, { "epoch": 0.27153642185751914, "grad_norm": 3.9604668617248535, "learning_rate": 1.71030733703509e-05, "loss": 3.7589, "step": 7706 }, { "epoch": 0.2715716588704776, "grad_norm": 2.6791017055511475, "learning_rate": 1.710226998885602e-05, "loss": 3.9693, "step": 7707 }, { "epoch": 0.27160689588343595, "grad_norm": 1.251715064048767, "learning_rate": 1.710146651485317e-05, "loss": 1.0254, "step": 7708 }, { "epoch": 0.2716421328963944, "grad_norm": 3.5483853816986084, "learning_rate": 1.7100662948352816e-05, "loss": 5.8369, "step": 7709 }, { "epoch": 0.27167736990935276, "grad_norm": 1.365407943725586, "learning_rate": 1.7099859289365418e-05, "loss": 1.0026, "step": 7710 }, { "epoch": 0.2717126069223112, "grad_norm": 1.2765616178512573, "learning_rate": 1.709905553790145e-05, "loss": 0.9132, "step": 7711 }, { "epoch": 0.27174784393526963, "grad_norm": 5.1565327644348145, "learning_rate": 1.709825169397138e-05, "loss": 3.1138, "step": 7712 }, { "epoch": 0.271783080948228, "grad_norm": 1.1921262741088867, "learning_rate": 1.7097447757585677e-05, "loss": 1.2308, "step": 7713 }, { "epoch": 0.27181831796118644, "grad_norm": 6.633178234100342, "learning_rate": 1.7096643728754814e-05, "loss": 3.2697, "step": 7714 }, { "epoch": 0.2718535549741448, "grad_norm": 2.1539132595062256, "learning_rate": 1.709583960748926e-05, "loss": 0.8159, "step": 7715 }, { "epoch": 0.27188879198710325, "grad_norm": 1.286666989326477, "learning_rate": 1.7095035393799496e-05, "loss": 1.0471, "step": 7716 }, { "epoch": 0.2719240290000617, "grad_norm": 2.4549219608306885, "learning_rate": 1.709423108769599e-05, "loss": 0.93, "step": 7717 }, { "epoch": 0.27195926601302006, "grad_norm": 3.3880369663238525, "learning_rate": 1.7093426689189217e-05, "loss": 3.382, "step": 7718 }, { "epoch": 0.2719945030259785, "grad_norm": 1.2678927183151245, "learning_rate": 1.709262219828966e-05, "loss": 0.9254, "step": 7719 }, { "epoch": 0.2720297400389369, "grad_norm": 6.128299236297607, "learning_rate": 1.7091817615007795e-05, "loss": 5.8661, "step": 7720 }, { "epoch": 0.2720649770518953, "grad_norm": 6.269692897796631, "learning_rate": 1.7091012939354103e-05, "loss": 6.1133, "step": 7721 }, { "epoch": 0.27210021406485374, "grad_norm": 8.06501293182373, "learning_rate": 1.7090208171339064e-05, "loss": 5.2262, "step": 7722 }, { "epoch": 0.2721354510778121, "grad_norm": 4.73806619644165, "learning_rate": 1.7089403310973162e-05, "loss": 3.2768, "step": 7723 }, { "epoch": 0.27217068809077055, "grad_norm": 1.7144246101379395, "learning_rate": 1.708859835826688e-05, "loss": 0.9744, "step": 7724 }, { "epoch": 0.27220592510372893, "grad_norm": 6.018250465393066, "learning_rate": 1.7087793313230697e-05, "loss": 6.0221, "step": 7725 }, { "epoch": 0.27224116211668736, "grad_norm": 3.527944564819336, "learning_rate": 1.7086988175875104e-05, "loss": 3.3283, "step": 7726 }, { "epoch": 0.2722763991296458, "grad_norm": 2.0905511379241943, "learning_rate": 1.7086182946210588e-05, "loss": 0.7487, "step": 7727 }, { "epoch": 0.2723116361426042, "grad_norm": 6.499960899353027, "learning_rate": 1.7085377624247637e-05, "loss": 6.1722, "step": 7728 }, { "epoch": 0.2723468731555626, "grad_norm": 2.201608657836914, "learning_rate": 1.708457220999674e-05, "loss": 3.5094, "step": 7729 }, { "epoch": 0.27238211016852104, "grad_norm": 3.390425682067871, "learning_rate": 1.708376670346839e-05, "loss": 2.9051, "step": 7730 }, { "epoch": 0.2724173471814794, "grad_norm": 1.108539342880249, "learning_rate": 1.708296110467307e-05, "loss": 1.0269, "step": 7731 }, { "epoch": 0.27245258419443785, "grad_norm": 1.9335638284683228, "learning_rate": 1.7082155413621283e-05, "loss": 0.8918, "step": 7732 }, { "epoch": 0.27248782120739623, "grad_norm": 3.0094099044799805, "learning_rate": 1.708134963032352e-05, "loss": 3.1646, "step": 7733 }, { "epoch": 0.27252305822035466, "grad_norm": 3.9404890537261963, "learning_rate": 1.7080543754790278e-05, "loss": 6.3461, "step": 7734 }, { "epoch": 0.2725582952333131, "grad_norm": 0.9995429515838623, "learning_rate": 1.7079737787032047e-05, "loss": 1.1478, "step": 7735 }, { "epoch": 0.2725935322462715, "grad_norm": 1.2568061351776123, "learning_rate": 1.7078931727059334e-05, "loss": 0.8545, "step": 7736 }, { "epoch": 0.2726287692592299, "grad_norm": 3.760343313217163, "learning_rate": 1.707812557488263e-05, "loss": 3.5942, "step": 7737 }, { "epoch": 0.2726640062721883, "grad_norm": 3.4483017921447754, "learning_rate": 1.707731933051244e-05, "loss": 3.5128, "step": 7738 }, { "epoch": 0.2726992432851467, "grad_norm": 1.4143909215927124, "learning_rate": 1.7076512993959266e-05, "loss": 0.8764, "step": 7739 }, { "epoch": 0.27273448029810515, "grad_norm": 1.4599757194519043, "learning_rate": 1.7075706565233606e-05, "loss": 1.1331, "step": 7740 }, { "epoch": 0.27276971731106353, "grad_norm": 8.604191780090332, "learning_rate": 1.7074900044345968e-05, "loss": 6.5188, "step": 7741 }, { "epoch": 0.27280495432402196, "grad_norm": 3.5368459224700928, "learning_rate": 1.7074093431306856e-05, "loss": 3.514, "step": 7742 }, { "epoch": 0.27284019133698034, "grad_norm": 7.137276649475098, "learning_rate": 1.7073286726126776e-05, "loss": 9.8304, "step": 7743 }, { "epoch": 0.2728754283499388, "grad_norm": 7.987939834594727, "learning_rate": 1.7072479928816234e-05, "loss": 4.904, "step": 7744 }, { "epoch": 0.2729106653628972, "grad_norm": 3.1990506649017334, "learning_rate": 1.7071673039385742e-05, "loss": 3.7798, "step": 7745 }, { "epoch": 0.2729459023758556, "grad_norm": 2.1751065254211426, "learning_rate": 1.7070866057845808e-05, "loss": 0.8916, "step": 7746 }, { "epoch": 0.272981139388814, "grad_norm": 1.408846378326416, "learning_rate": 1.707005898420694e-05, "loss": 1.0636, "step": 7747 }, { "epoch": 0.2730163764017724, "grad_norm": 1.4147818088531494, "learning_rate": 1.706925181847966e-05, "loss": 1.1519, "step": 7748 }, { "epoch": 0.27305161341473083, "grad_norm": 6.081296443939209, "learning_rate": 1.7068444560674467e-05, "loss": 5.6783, "step": 7749 }, { "epoch": 0.27308685042768926, "grad_norm": 1.035746455192566, "learning_rate": 1.7067637210801886e-05, "loss": 1.2249, "step": 7750 }, { "epoch": 0.27312208744064764, "grad_norm": 2.8928959369659424, "learning_rate": 1.706682976887243e-05, "loss": 2.5503, "step": 7751 }, { "epoch": 0.2731573244536061, "grad_norm": 14.205182075500488, "learning_rate": 1.7066022234896615e-05, "loss": 3.3409, "step": 7752 }, { "epoch": 0.27319256146656445, "grad_norm": 4.1990227699279785, "learning_rate": 1.7065214608884964e-05, "loss": 3.7154, "step": 7753 }, { "epoch": 0.2732277984795229, "grad_norm": 1.3473477363586426, "learning_rate": 1.7064406890847988e-05, "loss": 0.7421, "step": 7754 }, { "epoch": 0.2732630354924813, "grad_norm": 1.2069039344787598, "learning_rate": 1.7063599080796214e-05, "loss": 1.177, "step": 7755 }, { "epoch": 0.2732982725054397, "grad_norm": 1.350058913230896, "learning_rate": 1.7062791178740165e-05, "loss": 0.9213, "step": 7756 }, { "epoch": 0.27333350951839813, "grad_norm": 6.14827299118042, "learning_rate": 1.7061983184690358e-05, "loss": 3.1943, "step": 7757 }, { "epoch": 0.2733687465313565, "grad_norm": 1.7631746530532837, "learning_rate": 1.7061175098657322e-05, "loss": 1.1376, "step": 7758 }, { "epoch": 0.27340398354431494, "grad_norm": 3.562757730484009, "learning_rate": 1.706036692065158e-05, "loss": 3.9691, "step": 7759 }, { "epoch": 0.2734392205572734, "grad_norm": 6.711446762084961, "learning_rate": 1.705955865068366e-05, "loss": 6.1433, "step": 7760 }, { "epoch": 0.27347445757023175, "grad_norm": 3.9247231483459473, "learning_rate": 1.705875028876409e-05, "loss": 3.6452, "step": 7761 }, { "epoch": 0.2735096945831902, "grad_norm": 4.61865234375, "learning_rate": 1.7057941834903395e-05, "loss": 2.8083, "step": 7762 }, { "epoch": 0.2735449315961486, "grad_norm": 1.606563925743103, "learning_rate": 1.705713328911211e-05, "loss": 0.9181, "step": 7763 }, { "epoch": 0.273580168609107, "grad_norm": 6.4949517250061035, "learning_rate": 1.7056324651400768e-05, "loss": 4.1637, "step": 7764 }, { "epoch": 0.27361540562206543, "grad_norm": 5.298670768737793, "learning_rate": 1.7055515921779895e-05, "loss": 6.0306, "step": 7765 }, { "epoch": 0.2736506426350238, "grad_norm": 1.438739538192749, "learning_rate": 1.705470710026003e-05, "loss": 0.9473, "step": 7766 }, { "epoch": 0.27368587964798224, "grad_norm": 1.2612199783325195, "learning_rate": 1.7053898186851708e-05, "loss": 0.9372, "step": 7767 }, { "epoch": 0.2737211166609407, "grad_norm": 1.4590901136398315, "learning_rate": 1.705308918156546e-05, "loss": 1.0274, "step": 7768 }, { "epoch": 0.27375635367389906, "grad_norm": 3.353454351425171, "learning_rate": 1.705228008441183e-05, "loss": 2.8779, "step": 7769 }, { "epoch": 0.2737915906868575, "grad_norm": 1.5821377038955688, "learning_rate": 1.7051470895401353e-05, "loss": 1.0564, "step": 7770 }, { "epoch": 0.27382682769981587, "grad_norm": 1.7085870504379272, "learning_rate": 1.7050661614544568e-05, "loss": 1.0593, "step": 7771 }, { "epoch": 0.2738620647127743, "grad_norm": 2.063812732696533, "learning_rate": 1.704985224185202e-05, "loss": 1.1288, "step": 7772 }, { "epoch": 0.27389730172573273, "grad_norm": 3.3878817558288574, "learning_rate": 1.7049042777334245e-05, "loss": 3.8524, "step": 7773 }, { "epoch": 0.2739325387386911, "grad_norm": 2.0530753135681152, "learning_rate": 1.7048233221001793e-05, "loss": 0.9695, "step": 7774 }, { "epoch": 0.27396777575164954, "grad_norm": 1.224611759185791, "learning_rate": 1.7047423572865203e-05, "loss": 1.143, "step": 7775 }, { "epoch": 0.2740030127646079, "grad_norm": 3.7470717430114746, "learning_rate": 1.7046613832935025e-05, "loss": 3.3647, "step": 7776 }, { "epoch": 0.27403824977756636, "grad_norm": 4.897768497467041, "learning_rate": 1.7045804001221804e-05, "loss": 5.5568, "step": 7777 }, { "epoch": 0.2740734867905248, "grad_norm": 3.1915698051452637, "learning_rate": 1.7044994077736087e-05, "loss": 4.1337, "step": 7778 }, { "epoch": 0.27410872380348317, "grad_norm": 1.7311924695968628, "learning_rate": 1.7044184062488427e-05, "loss": 0.9304, "step": 7779 }, { "epoch": 0.2741439608164416, "grad_norm": 3.9586880207061768, "learning_rate": 1.704337395548937e-05, "loss": 3.7196, "step": 7780 }, { "epoch": 0.2741791978294, "grad_norm": 4.1338958740234375, "learning_rate": 1.704256375674947e-05, "loss": 3.3296, "step": 7781 }, { "epoch": 0.2742144348423584, "grad_norm": 1.2747232913970947, "learning_rate": 1.7041753466279286e-05, "loss": 0.9206, "step": 7782 }, { "epoch": 0.27424967185531685, "grad_norm": 1.4601997137069702, "learning_rate": 1.7040943084089358e-05, "loss": 0.8821, "step": 7783 }, { "epoch": 0.2742849088682752, "grad_norm": 5.577575206756592, "learning_rate": 1.7040132610190254e-05, "loss": 6.1693, "step": 7784 }, { "epoch": 0.27432014588123366, "grad_norm": 1.5601131916046143, "learning_rate": 1.7039322044592525e-05, "loss": 0.8544, "step": 7785 }, { "epoch": 0.27435538289419203, "grad_norm": 1.0849493741989136, "learning_rate": 1.703851138730673e-05, "loss": 0.876, "step": 7786 }, { "epoch": 0.27439061990715047, "grad_norm": 7.321214199066162, "learning_rate": 1.7037700638343427e-05, "loss": 5.9045, "step": 7787 }, { "epoch": 0.2744258569201089, "grad_norm": 5.438805103302002, "learning_rate": 1.7036889797713176e-05, "loss": 2.9423, "step": 7788 }, { "epoch": 0.2744610939330673, "grad_norm": 1.66690993309021, "learning_rate": 1.703607886542654e-05, "loss": 1.0429, "step": 7789 }, { "epoch": 0.2744963309460257, "grad_norm": 2.8001232147216797, "learning_rate": 1.703526784149408e-05, "loss": 0.9105, "step": 7790 }, { "epoch": 0.2745315679589841, "grad_norm": 1.8720940351486206, "learning_rate": 1.7034456725926362e-05, "loss": 1.0633, "step": 7791 }, { "epoch": 0.2745668049719425, "grad_norm": 2.446129322052002, "learning_rate": 1.7033645518733946e-05, "loss": 3.0515, "step": 7792 }, { "epoch": 0.27460204198490096, "grad_norm": 5.60662317276001, "learning_rate": 1.7032834219927402e-05, "loss": 5.5084, "step": 7793 }, { "epoch": 0.27463727899785934, "grad_norm": 1.273435354232788, "learning_rate": 1.7032022829517298e-05, "loss": 0.9028, "step": 7794 }, { "epoch": 0.27467251601081777, "grad_norm": 10.217309951782227, "learning_rate": 1.70312113475142e-05, "loss": 6.3254, "step": 7795 }, { "epoch": 0.2747077530237762, "grad_norm": 5.586942195892334, "learning_rate": 1.703039977392868e-05, "loss": 4.0533, "step": 7796 }, { "epoch": 0.2747429900367346, "grad_norm": 4.271275043487549, "learning_rate": 1.7029588108771307e-05, "loss": 3.7418, "step": 7797 }, { "epoch": 0.274778227049693, "grad_norm": 2.0197746753692627, "learning_rate": 1.7028776352052653e-05, "loss": 0.8171, "step": 7798 }, { "epoch": 0.2748134640626514, "grad_norm": 1.3360849618911743, "learning_rate": 1.7027964503783293e-05, "loss": 1.0579, "step": 7799 }, { "epoch": 0.2748487010756098, "grad_norm": 3.9458541870117188, "learning_rate": 1.7027152563973795e-05, "loss": 3.7136, "step": 7800 }, { "epoch": 0.27488393808856826, "grad_norm": 1.317728877067566, "learning_rate": 1.7026340532634744e-05, "loss": 1.081, "step": 7801 }, { "epoch": 0.27491917510152664, "grad_norm": 3.7566449642181396, "learning_rate": 1.702552840977671e-05, "loss": 3.601, "step": 7802 }, { "epoch": 0.27495441211448507, "grad_norm": 1.1146000623703003, "learning_rate": 1.7024716195410274e-05, "loss": 0.7432, "step": 7803 }, { "epoch": 0.27498964912744345, "grad_norm": 2.6671087741851807, "learning_rate": 1.702390388954602e-05, "loss": 3.016, "step": 7804 }, { "epoch": 0.2750248861404019, "grad_norm": 1.366184949874878, "learning_rate": 1.7023091492194517e-05, "loss": 0.9644, "step": 7805 }, { "epoch": 0.2750601231533603, "grad_norm": 6.6655988693237305, "learning_rate": 1.7022279003366357e-05, "loss": 3.9942, "step": 7806 }, { "epoch": 0.2750953601663187, "grad_norm": 2.3351736068725586, "learning_rate": 1.7021466423072113e-05, "loss": 2.7799, "step": 7807 }, { "epoch": 0.2751305971792771, "grad_norm": 2.640033006668091, "learning_rate": 1.702065375132238e-05, "loss": 3.22, "step": 7808 }, { "epoch": 0.2751658341922355, "grad_norm": 5.117120265960693, "learning_rate": 1.7019840988127737e-05, "loss": 2.3247, "step": 7809 }, { "epoch": 0.27520107120519394, "grad_norm": 3.652473211288452, "learning_rate": 1.7019028133498766e-05, "loss": 3.2759, "step": 7810 }, { "epoch": 0.27523630821815237, "grad_norm": 2.919356346130371, "learning_rate": 1.7018215187446067e-05, "loss": 3.4755, "step": 7811 }, { "epoch": 0.27527154523111075, "grad_norm": 2.9852514266967773, "learning_rate": 1.7017402149980214e-05, "loss": 3.8196, "step": 7812 }, { "epoch": 0.2753067822440692, "grad_norm": 6.824577331542969, "learning_rate": 1.7016589021111807e-05, "loss": 6.3739, "step": 7813 }, { "epoch": 0.27534201925702756, "grad_norm": 1.7586686611175537, "learning_rate": 1.7015775800851436e-05, "loss": 1.0217, "step": 7814 }, { "epoch": 0.275377256269986, "grad_norm": 1.190161108970642, "learning_rate": 1.701496248920969e-05, "loss": 0.932, "step": 7815 }, { "epoch": 0.2754124932829444, "grad_norm": 1.5570921897888184, "learning_rate": 1.7014149086197164e-05, "loss": 0.8098, "step": 7816 }, { "epoch": 0.2754477302959028, "grad_norm": 3.1669864654541016, "learning_rate": 1.7013335591824452e-05, "loss": 3.5341, "step": 7817 }, { "epoch": 0.27548296730886124, "grad_norm": 1.2541669607162476, "learning_rate": 1.701252200610215e-05, "loss": 0.8629, "step": 7818 }, { "epoch": 0.2755182043218196, "grad_norm": 4.514847755432129, "learning_rate": 1.701170832904086e-05, "loss": 3.0835, "step": 7819 }, { "epoch": 0.27555344133477805, "grad_norm": 1.1498196125030518, "learning_rate": 1.701089456065117e-05, "loss": 0.757, "step": 7820 }, { "epoch": 0.2755886783477365, "grad_norm": 1.2343220710754395, "learning_rate": 1.701008070094369e-05, "loss": 1.1689, "step": 7821 }, { "epoch": 0.27562391536069486, "grad_norm": 1.9397529363632202, "learning_rate": 1.7009266749929013e-05, "loss": 0.793, "step": 7822 }, { "epoch": 0.2756591523736533, "grad_norm": 3.480491876602173, "learning_rate": 1.7008452707617745e-05, "loss": 3.1727, "step": 7823 }, { "epoch": 0.27569438938661167, "grad_norm": 1.9480481147766113, "learning_rate": 1.7007638574020486e-05, "loss": 1.1269, "step": 7824 }, { "epoch": 0.2757296263995701, "grad_norm": 1.3425588607788086, "learning_rate": 1.7006824349147845e-05, "loss": 0.8785, "step": 7825 }, { "epoch": 0.27576486341252854, "grad_norm": 2.1451992988586426, "learning_rate": 1.7006010033010423e-05, "loss": 2.9652, "step": 7826 }, { "epoch": 0.2758001004254869, "grad_norm": 1.1690481901168823, "learning_rate": 1.7005195625618827e-05, "loss": 0.997, "step": 7827 }, { "epoch": 0.27583533743844535, "grad_norm": 3.584214448928833, "learning_rate": 1.7004381126983667e-05, "loss": 3.2032, "step": 7828 }, { "epoch": 0.2758705744514037, "grad_norm": 1.4524286985397339, "learning_rate": 1.7003566537115546e-05, "loss": 0.9383, "step": 7829 }, { "epoch": 0.27590581146436216, "grad_norm": 1.3802992105484009, "learning_rate": 1.7002751856025085e-05, "loss": 1.1195, "step": 7830 }, { "epoch": 0.2759410484773206, "grad_norm": 3.646937370300293, "learning_rate": 1.7001937083722886e-05, "loss": 3.1711, "step": 7831 }, { "epoch": 0.27597628549027897, "grad_norm": 4.864718437194824, "learning_rate": 1.7001122220219566e-05, "loss": 3.9969, "step": 7832 }, { "epoch": 0.2760115225032374, "grad_norm": 5.404331684112549, "learning_rate": 1.7000307265525736e-05, "loss": 4.9062, "step": 7833 }, { "epoch": 0.27604675951619584, "grad_norm": 4.8470234870910645, "learning_rate": 1.6999492219652015e-05, "loss": 5.151, "step": 7834 }, { "epoch": 0.2760819965291542, "grad_norm": 0.9938691258430481, "learning_rate": 1.6998677082609013e-05, "loss": 0.8249, "step": 7835 }, { "epoch": 0.27611723354211265, "grad_norm": 2.0294134616851807, "learning_rate": 1.699786185440735e-05, "loss": 2.7695, "step": 7836 }, { "epoch": 0.27615247055507103, "grad_norm": 1.5342037677764893, "learning_rate": 1.6997046535057647e-05, "loss": 0.7677, "step": 7837 }, { "epoch": 0.27618770756802946, "grad_norm": 2.1172537803649902, "learning_rate": 1.699623112457052e-05, "loss": 1.0534, "step": 7838 }, { "epoch": 0.2762229445809879, "grad_norm": 3.2362802028656006, "learning_rate": 1.6995415622956596e-05, "loss": 3.1908, "step": 7839 }, { "epoch": 0.2762581815939463, "grad_norm": 5.677807807922363, "learning_rate": 1.6994600030226487e-05, "loss": 3.9135, "step": 7840 }, { "epoch": 0.2762934186069047, "grad_norm": 5.886993408203125, "learning_rate": 1.6993784346390823e-05, "loss": 5.9696, "step": 7841 }, { "epoch": 0.2763286556198631, "grad_norm": 5.182376384735107, "learning_rate": 1.6992968571460228e-05, "loss": 3.5317, "step": 7842 }, { "epoch": 0.2763638926328215, "grad_norm": 2.7653112411499023, "learning_rate": 1.6992152705445324e-05, "loss": 3.4215, "step": 7843 }, { "epoch": 0.27639912964577995, "grad_norm": 4.641713619232178, "learning_rate": 1.6991336748356742e-05, "loss": 5.6774, "step": 7844 }, { "epoch": 0.27643436665873833, "grad_norm": 1.1834218502044678, "learning_rate": 1.6990520700205113e-05, "loss": 1.0141, "step": 7845 }, { "epoch": 0.27646960367169676, "grad_norm": 1.189992904663086, "learning_rate": 1.6989704561001056e-05, "loss": 1.0135, "step": 7846 }, { "epoch": 0.27650484068465514, "grad_norm": 1.3791018724441528, "learning_rate": 1.698888833075521e-05, "loss": 0.8759, "step": 7847 }, { "epoch": 0.2765400776976136, "grad_norm": 1.1027833223342896, "learning_rate": 1.6988072009478198e-05, "loss": 0.7968, "step": 7848 }, { "epoch": 0.276575314710572, "grad_norm": 5.735690116882324, "learning_rate": 1.698725559718066e-05, "loss": 3.5658, "step": 7849 }, { "epoch": 0.2766105517235304, "grad_norm": 1.1902847290039062, "learning_rate": 1.698643909387323e-05, "loss": 1.3068, "step": 7850 }, { "epoch": 0.2766457887364888, "grad_norm": 0.9727261066436768, "learning_rate": 1.698562249956654e-05, "loss": 0.9118, "step": 7851 }, { "epoch": 0.2766810257494472, "grad_norm": 6.3035569190979, "learning_rate": 1.6984805814271225e-05, "loss": 8.9682, "step": 7852 }, { "epoch": 0.27671626276240563, "grad_norm": 1.7313940525054932, "learning_rate": 1.698398903799793e-05, "loss": 0.956, "step": 7853 }, { "epoch": 0.27675149977536406, "grad_norm": 4.7031569480896, "learning_rate": 1.6983172170757282e-05, "loss": 3.2155, "step": 7854 }, { "epoch": 0.27678673678832244, "grad_norm": 2.858426332473755, "learning_rate": 1.6982355212559932e-05, "loss": 3.2526, "step": 7855 }, { "epoch": 0.2768219738012809, "grad_norm": 2.073010206222534, "learning_rate": 1.6981538163416512e-05, "loss": 3.6093, "step": 7856 }, { "epoch": 0.27685721081423925, "grad_norm": 1.409349799156189, "learning_rate": 1.698072102333767e-05, "loss": 0.8448, "step": 7857 }, { "epoch": 0.2768924478271977, "grad_norm": 1.1251789331436157, "learning_rate": 1.6979903792334048e-05, "loss": 0.9173, "step": 7858 }, { "epoch": 0.2769276848401561, "grad_norm": 2.0026912689208984, "learning_rate": 1.6979086470416288e-05, "loss": 2.9214, "step": 7859 }, { "epoch": 0.2769629218531145, "grad_norm": 1.0575964450836182, "learning_rate": 1.697826905759504e-05, "loss": 1.0315, "step": 7860 }, { "epoch": 0.27699815886607293, "grad_norm": 3.1487183570861816, "learning_rate": 1.6977451553880947e-05, "loss": 3.3475, "step": 7861 }, { "epoch": 0.2770333958790313, "grad_norm": 1.6410499811172485, "learning_rate": 1.697663395928466e-05, "loss": 0.9306, "step": 7862 }, { "epoch": 0.27706863289198974, "grad_norm": 1.2611356973648071, "learning_rate": 1.6975816273816825e-05, "loss": 1.0488, "step": 7863 }, { "epoch": 0.2771038699049482, "grad_norm": 3.6302294731140137, "learning_rate": 1.69749984974881e-05, "loss": 3.184, "step": 7864 }, { "epoch": 0.27713910691790655, "grad_norm": 8.988673210144043, "learning_rate": 1.697418063030912e-05, "loss": 5.5063, "step": 7865 }, { "epoch": 0.277174343930865, "grad_norm": 6.364974498748779, "learning_rate": 1.697336267229056e-05, "loss": 4.079, "step": 7866 }, { "epoch": 0.2772095809438234, "grad_norm": 5.20544958114624, "learning_rate": 1.6972544623443057e-05, "loss": 5.6319, "step": 7867 }, { "epoch": 0.2772448179567818, "grad_norm": 0.9326165914535522, "learning_rate": 1.6971726483777274e-05, "loss": 0.7193, "step": 7868 }, { "epoch": 0.27728005496974023, "grad_norm": 1.2336492538452148, "learning_rate": 1.6970908253303867e-05, "loss": 1.0735, "step": 7869 }, { "epoch": 0.2773152919826986, "grad_norm": 2.4588379859924316, "learning_rate": 1.6970089932033488e-05, "loss": 2.8587, "step": 7870 }, { "epoch": 0.27735052899565704, "grad_norm": 2.056260108947754, "learning_rate": 1.6969271519976803e-05, "loss": 3.0662, "step": 7871 }, { "epoch": 0.2773857660086155, "grad_norm": 1.732420563697815, "learning_rate": 1.6968453017144465e-05, "loss": 1.0978, "step": 7872 }, { "epoch": 0.27742100302157385, "grad_norm": 3.277446985244751, "learning_rate": 1.6967634423547143e-05, "loss": 4.0574, "step": 7873 }, { "epoch": 0.2774562400345323, "grad_norm": 1.682615041732788, "learning_rate": 1.696681573919549e-05, "loss": 0.9758, "step": 7874 }, { "epoch": 0.27749147704749066, "grad_norm": 1.7561677694320679, "learning_rate": 1.696599696410018e-05, "loss": 1.1421, "step": 7875 }, { "epoch": 0.2775267140604491, "grad_norm": 1.569543719291687, "learning_rate": 1.6965178098271873e-05, "loss": 0.9742, "step": 7876 }, { "epoch": 0.27756195107340753, "grad_norm": 1.0494067668914795, "learning_rate": 1.696435914172123e-05, "loss": 1.0732, "step": 7877 }, { "epoch": 0.2775971880863659, "grad_norm": 3.711167335510254, "learning_rate": 1.6963540094458926e-05, "loss": 3.133, "step": 7878 }, { "epoch": 0.27763242509932434, "grad_norm": 3.0764029026031494, "learning_rate": 1.6962720956495622e-05, "loss": 3.1702, "step": 7879 }, { "epoch": 0.2776676621122827, "grad_norm": 2.045931100845337, "learning_rate": 1.696190172784199e-05, "loss": 1.1172, "step": 7880 }, { "epoch": 0.27770289912524115, "grad_norm": 2.2637834548950195, "learning_rate": 1.6961082408508703e-05, "loss": 1.2043, "step": 7881 }, { "epoch": 0.2777381361381996, "grad_norm": 1.248789668083191, "learning_rate": 1.696026299850643e-05, "loss": 1.0736, "step": 7882 }, { "epoch": 0.27777337315115797, "grad_norm": 1.6813030242919922, "learning_rate": 1.6959443497845844e-05, "loss": 1.3533, "step": 7883 }, { "epoch": 0.2778086101641164, "grad_norm": 1.1805925369262695, "learning_rate": 1.6958623906537622e-05, "loss": 1.0387, "step": 7884 }, { "epoch": 0.2778438471770748, "grad_norm": 5.694366455078125, "learning_rate": 1.695780422459244e-05, "loss": 5.9889, "step": 7885 }, { "epoch": 0.2778790841900332, "grad_norm": 5.647315502166748, "learning_rate": 1.695698445202097e-05, "loss": 6.7132, "step": 7886 }, { "epoch": 0.27791432120299164, "grad_norm": 1.0537611246109009, "learning_rate": 1.6956164588833886e-05, "loss": 0.9686, "step": 7887 }, { "epoch": 0.27794955821595, "grad_norm": 1.3898664712905884, "learning_rate": 1.695534463504188e-05, "loss": 1.1075, "step": 7888 }, { "epoch": 0.27798479522890845, "grad_norm": 1.242136836051941, "learning_rate": 1.6954524590655618e-05, "loss": 1.2111, "step": 7889 }, { "epoch": 0.27802003224186683, "grad_norm": 3.075324535369873, "learning_rate": 1.6953704455685786e-05, "loss": 3.559, "step": 7890 }, { "epoch": 0.27805526925482527, "grad_norm": 1.7775253057479858, "learning_rate": 1.6952884230143075e-05, "loss": 0.8384, "step": 7891 }, { "epoch": 0.2780905062677837, "grad_norm": 3.398883104324341, "learning_rate": 1.6952063914038155e-05, "loss": 5.331, "step": 7892 }, { "epoch": 0.2781257432807421, "grad_norm": 1.1550018787384033, "learning_rate": 1.695124350738172e-05, "loss": 1.1142, "step": 7893 }, { "epoch": 0.2781609802937005, "grad_norm": 2.7494664192199707, "learning_rate": 1.6950423010184452e-05, "loss": 2.878, "step": 7894 }, { "epoch": 0.2781962173066589, "grad_norm": 3.6187875270843506, "learning_rate": 1.6949602422457035e-05, "loss": 3.2028, "step": 7895 }, { "epoch": 0.2782314543196173, "grad_norm": 3.2142364978790283, "learning_rate": 1.694878174421017e-05, "loss": 3.2431, "step": 7896 }, { "epoch": 0.27826669133257576, "grad_norm": 1.6429407596588135, "learning_rate": 1.6947960975454528e-05, "loss": 0.831, "step": 7897 }, { "epoch": 0.27830192834553413, "grad_norm": 3.5778234004974365, "learning_rate": 1.6947140116200816e-05, "loss": 3.9803, "step": 7898 }, { "epoch": 0.27833716535849257, "grad_norm": 3.8210132122039795, "learning_rate": 1.694631916645972e-05, "loss": 4.2235, "step": 7899 }, { "epoch": 0.27837240237145094, "grad_norm": 4.0708513259887695, "learning_rate": 1.694549812624193e-05, "loss": 3.666, "step": 7900 }, { "epoch": 0.2784076393844094, "grad_norm": 3.942701578140259, "learning_rate": 1.694467699555814e-05, "loss": 4.8, "step": 7901 }, { "epoch": 0.2784428763973678, "grad_norm": 5.221865177154541, "learning_rate": 1.6943855774419048e-05, "loss": 6.3802, "step": 7902 }, { "epoch": 0.2784781134103262, "grad_norm": 3.661680221557617, "learning_rate": 1.6943034462835353e-05, "loss": 3.3409, "step": 7903 }, { "epoch": 0.2785133504232846, "grad_norm": 1.127821683883667, "learning_rate": 1.6942213060817746e-05, "loss": 0.8218, "step": 7904 }, { "epoch": 0.27854858743624306, "grad_norm": 1.8054465055465698, "learning_rate": 1.694139156837693e-05, "loss": 0.9741, "step": 7905 }, { "epoch": 0.27858382444920143, "grad_norm": 1.3719658851623535, "learning_rate": 1.6940569985523607e-05, "loss": 0.9235, "step": 7906 }, { "epoch": 0.27861906146215987, "grad_norm": 3.8308281898498535, "learning_rate": 1.6939748312268475e-05, "loss": 3.5995, "step": 7907 }, { "epoch": 0.27865429847511825, "grad_norm": 2.3234710693359375, "learning_rate": 1.6938926548622236e-05, "loss": 3.3338, "step": 7908 }, { "epoch": 0.2786895354880767, "grad_norm": 1.1589099168777466, "learning_rate": 1.69381046945956e-05, "loss": 1.0244, "step": 7909 }, { "epoch": 0.2787247725010351, "grad_norm": 1.764386534690857, "learning_rate": 1.693728275019926e-05, "loss": 0.9863, "step": 7910 }, { "epoch": 0.2787600095139935, "grad_norm": 5.991752624511719, "learning_rate": 1.693646071544393e-05, "loss": 7.5759, "step": 7911 }, { "epoch": 0.2787952465269519, "grad_norm": 3.2166521549224854, "learning_rate": 1.693563859034032e-05, "loss": 2.8835, "step": 7912 }, { "epoch": 0.2788304835399103, "grad_norm": 1.3783706426620483, "learning_rate": 1.6934816374899127e-05, "loss": 1.033, "step": 7913 }, { "epoch": 0.27886572055286873, "grad_norm": 1.8562017679214478, "learning_rate": 1.6933994069131073e-05, "loss": 0.8107, "step": 7914 }, { "epoch": 0.27890095756582717, "grad_norm": 1.415095567703247, "learning_rate": 1.693317167304686e-05, "loss": 0.9194, "step": 7915 }, { "epoch": 0.27893619457878555, "grad_norm": 2.838683605194092, "learning_rate": 1.6932349186657203e-05, "loss": 3.3896, "step": 7916 }, { "epoch": 0.278971431591744, "grad_norm": 6.012472152709961, "learning_rate": 1.6931526609972816e-05, "loss": 5.5951, "step": 7917 }, { "epoch": 0.27900666860470236, "grad_norm": 5.903464317321777, "learning_rate": 1.693070394300441e-05, "loss": 6.0368, "step": 7918 }, { "epoch": 0.2790419056176608, "grad_norm": 1.0910524129867554, "learning_rate": 1.6929881185762704e-05, "loss": 0.7563, "step": 7919 }, { "epoch": 0.2790771426306192, "grad_norm": 1.439341425895691, "learning_rate": 1.692905833825841e-05, "loss": 1.159, "step": 7920 }, { "epoch": 0.2791123796435776, "grad_norm": 5.2574334144592285, "learning_rate": 1.6928235400502254e-05, "loss": 3.6113, "step": 7921 }, { "epoch": 0.27914761665653604, "grad_norm": 6.574552536010742, "learning_rate": 1.6927412372504947e-05, "loss": 6.1266, "step": 7922 }, { "epoch": 0.2791828536694944, "grad_norm": 1.2190531492233276, "learning_rate": 1.6926589254277212e-05, "loss": 1.2774, "step": 7923 }, { "epoch": 0.27921809068245285, "grad_norm": 2.119443655014038, "learning_rate": 1.6925766045829765e-05, "loss": 0.8913, "step": 7924 }, { "epoch": 0.2792533276954113, "grad_norm": 0.9986267685890198, "learning_rate": 1.692494274717334e-05, "loss": 0.6585, "step": 7925 }, { "epoch": 0.27928856470836966, "grad_norm": 5.669783115386963, "learning_rate": 1.6924119358318648e-05, "loss": 5.8848, "step": 7926 }, { "epoch": 0.2793238017213281, "grad_norm": 3.754469633102417, "learning_rate": 1.6923295879276423e-05, "loss": 3.4372, "step": 7927 }, { "epoch": 0.27935903873428647, "grad_norm": 6.194543838500977, "learning_rate": 1.692247231005739e-05, "loss": 8.967, "step": 7928 }, { "epoch": 0.2793942757472449, "grad_norm": 2.4825658798217773, "learning_rate": 1.6921648650672264e-05, "loss": 3.4453, "step": 7929 }, { "epoch": 0.27942951276020334, "grad_norm": 4.726447105407715, "learning_rate": 1.6920824901131788e-05, "loss": 3.1185, "step": 7930 }, { "epoch": 0.2794647497731617, "grad_norm": 4.950978755950928, "learning_rate": 1.6920001061446687e-05, "loss": 2.881, "step": 7931 }, { "epoch": 0.27949998678612015, "grad_norm": 1.5463054180145264, "learning_rate": 1.691917713162769e-05, "loss": 1.1185, "step": 7932 }, { "epoch": 0.2795352237990785, "grad_norm": 4.095575332641602, "learning_rate": 1.6918353111685528e-05, "loss": 4.0867, "step": 7933 }, { "epoch": 0.27957046081203696, "grad_norm": 1.5869991779327393, "learning_rate": 1.6917529001630936e-05, "loss": 0.9793, "step": 7934 }, { "epoch": 0.2796056978249954, "grad_norm": 5.418867588043213, "learning_rate": 1.6916704801474648e-05, "loss": 4.1674, "step": 7935 }, { "epoch": 0.27964093483795377, "grad_norm": 4.2233662605285645, "learning_rate": 1.69158805112274e-05, "loss": 3.829, "step": 7936 }, { "epoch": 0.2796761718509122, "grad_norm": 1.1435028314590454, "learning_rate": 1.6915056130899926e-05, "loss": 0.9568, "step": 7937 }, { "epoch": 0.27971140886387064, "grad_norm": 2.2624688148498535, "learning_rate": 1.6914231660502966e-05, "loss": 3.2004, "step": 7938 }, { "epoch": 0.279746645876829, "grad_norm": 2.2112863063812256, "learning_rate": 1.6913407100047258e-05, "loss": 3.2929, "step": 7939 }, { "epoch": 0.27978188288978745, "grad_norm": 1.1803864240646362, "learning_rate": 1.6912582449543542e-05, "loss": 0.8778, "step": 7940 }, { "epoch": 0.2798171199027458, "grad_norm": 1.678605079650879, "learning_rate": 1.6911757709002557e-05, "loss": 0.8797, "step": 7941 }, { "epoch": 0.27985235691570426, "grad_norm": 1.1840296983718872, "learning_rate": 1.6910932878435052e-05, "loss": 0.9868, "step": 7942 }, { "epoch": 0.2798875939286627, "grad_norm": 2.7544398307800293, "learning_rate": 1.691010795785176e-05, "loss": 2.951, "step": 7943 }, { "epoch": 0.27992283094162107, "grad_norm": 2.689194679260254, "learning_rate": 1.6909282947263435e-05, "loss": 3.3433, "step": 7944 }, { "epoch": 0.2799580679545795, "grad_norm": 1.3044202327728271, "learning_rate": 1.690845784668082e-05, "loss": 0.7908, "step": 7945 }, { "epoch": 0.2799933049675379, "grad_norm": 8.90505599975586, "learning_rate": 1.690763265611466e-05, "loss": 6.3012, "step": 7946 }, { "epoch": 0.2800285419804963, "grad_norm": 8.489770889282227, "learning_rate": 1.6906807375575708e-05, "loss": 7.6438, "step": 7947 }, { "epoch": 0.28006377899345475, "grad_norm": 1.7461615800857544, "learning_rate": 1.6905982005074704e-05, "loss": 0.8969, "step": 7948 }, { "epoch": 0.2800990160064131, "grad_norm": 1.2442078590393066, "learning_rate": 1.690515654462241e-05, "loss": 0.9784, "step": 7949 }, { "epoch": 0.28013425301937156, "grad_norm": 1.7472496032714844, "learning_rate": 1.6904330994229573e-05, "loss": 0.8939, "step": 7950 }, { "epoch": 0.28016949003232994, "grad_norm": 0.9656159281730652, "learning_rate": 1.690350535390694e-05, "loss": 1.104, "step": 7951 }, { "epoch": 0.28020472704528837, "grad_norm": 1.0485100746154785, "learning_rate": 1.6902679623665276e-05, "loss": 0.8502, "step": 7952 }, { "epoch": 0.2802399640582468, "grad_norm": 6.519137859344482, "learning_rate": 1.6901853803515328e-05, "loss": 6.2162, "step": 7953 }, { "epoch": 0.2802752010712052, "grad_norm": 1.2900996208190918, "learning_rate": 1.6901027893467853e-05, "loss": 0.6297, "step": 7954 }, { "epoch": 0.2803104380841636, "grad_norm": 5.632170677185059, "learning_rate": 1.6900201893533613e-05, "loss": 6.5096, "step": 7955 }, { "epoch": 0.280345675097122, "grad_norm": 1.3431763648986816, "learning_rate": 1.689937580372336e-05, "loss": 1.0771, "step": 7956 }, { "epoch": 0.2803809121100804, "grad_norm": 4.527472972869873, "learning_rate": 1.6898549624047863e-05, "loss": 2.9919, "step": 7957 }, { "epoch": 0.28041614912303886, "grad_norm": 4.594919681549072, "learning_rate": 1.689772335451788e-05, "loss": 3.4521, "step": 7958 }, { "epoch": 0.28045138613599724, "grad_norm": 7.578954696655273, "learning_rate": 1.6896896995144166e-05, "loss": 3.3323, "step": 7959 }, { "epoch": 0.28048662314895567, "grad_norm": 1.5363906621932983, "learning_rate": 1.68960705459375e-05, "loss": 0.9922, "step": 7960 }, { "epoch": 0.28052186016191405, "grad_norm": 1.4905825853347778, "learning_rate": 1.689524400690863e-05, "loss": 0.8714, "step": 7961 }, { "epoch": 0.2805570971748725, "grad_norm": 7.0098557472229, "learning_rate": 1.6894417378068326e-05, "loss": 5.0064, "step": 7962 }, { "epoch": 0.2805923341878309, "grad_norm": 1.3842806816101074, "learning_rate": 1.689359065942736e-05, "loss": 1.1304, "step": 7963 }, { "epoch": 0.2806275712007893, "grad_norm": 3.745006561279297, "learning_rate": 1.68927638509965e-05, "loss": 3.817, "step": 7964 }, { "epoch": 0.2806628082137477, "grad_norm": 1.9350675344467163, "learning_rate": 1.6891936952786507e-05, "loss": 0.948, "step": 7965 }, { "epoch": 0.2806980452267061, "grad_norm": 0.8987715840339661, "learning_rate": 1.6891109964808162e-05, "loss": 1.0053, "step": 7966 }, { "epoch": 0.28073328223966454, "grad_norm": 1.206886887550354, "learning_rate": 1.6890282887072233e-05, "loss": 0.7778, "step": 7967 }, { "epoch": 0.280768519252623, "grad_norm": 4.79189395904541, "learning_rate": 1.6889455719589486e-05, "loss": 6.5842, "step": 7968 }, { "epoch": 0.28080375626558135, "grad_norm": 1.714996099472046, "learning_rate": 1.6888628462370708e-05, "loss": 0.8059, "step": 7969 }, { "epoch": 0.2808389932785398, "grad_norm": 1.0195297002792358, "learning_rate": 1.688780111542666e-05, "loss": 0.8688, "step": 7970 }, { "epoch": 0.2808742302914982, "grad_norm": 0.9747275710105896, "learning_rate": 1.688697367876813e-05, "loss": 1.0098, "step": 7971 }, { "epoch": 0.2809094673044566, "grad_norm": 0.9533060789108276, "learning_rate": 1.6886146152405886e-05, "loss": 0.8861, "step": 7972 }, { "epoch": 0.28094470431741503, "grad_norm": 2.6908583641052246, "learning_rate": 1.6885318536350717e-05, "loss": 3.0645, "step": 7973 }, { "epoch": 0.2809799413303734, "grad_norm": 3.008983850479126, "learning_rate": 1.688449083061339e-05, "loss": 3.6882, "step": 7974 }, { "epoch": 0.28101517834333184, "grad_norm": 1.2464674711227417, "learning_rate": 1.68836630352047e-05, "loss": 1.0459, "step": 7975 }, { "epoch": 0.2810504153562903, "grad_norm": 4.479172229766846, "learning_rate": 1.688283515013542e-05, "loss": 8.9161, "step": 7976 }, { "epoch": 0.28108565236924865, "grad_norm": 2.1118788719177246, "learning_rate": 1.6882007175416333e-05, "loss": 3.9609, "step": 7977 }, { "epoch": 0.2811208893822071, "grad_norm": 4.414670944213867, "learning_rate": 1.6881179111058226e-05, "loss": 5.8202, "step": 7978 }, { "epoch": 0.28115612639516546, "grad_norm": 3.9464664459228516, "learning_rate": 1.6880350957071887e-05, "loss": 5.7341, "step": 7979 }, { "epoch": 0.2811913634081239, "grad_norm": 1.6637758016586304, "learning_rate": 1.6879522713468103e-05, "loss": 1.0893, "step": 7980 }, { "epoch": 0.28122660042108233, "grad_norm": 1.578465461730957, "learning_rate": 1.6878694380257652e-05, "loss": 0.9952, "step": 7981 }, { "epoch": 0.2812618374340407, "grad_norm": 3.213214635848999, "learning_rate": 1.6877865957451333e-05, "loss": 3.2703, "step": 7982 }, { "epoch": 0.28129707444699914, "grad_norm": 3.5571513175964355, "learning_rate": 1.6877037445059936e-05, "loss": 3.2285, "step": 7983 }, { "epoch": 0.2813323114599575, "grad_norm": 1.6533939838409424, "learning_rate": 1.6876208843094252e-05, "loss": 0.8726, "step": 7984 }, { "epoch": 0.28136754847291595, "grad_norm": 0.8978941440582275, "learning_rate": 1.687538015156507e-05, "loss": 0.823, "step": 7985 }, { "epoch": 0.2814027854858744, "grad_norm": 4.5225114822387695, "learning_rate": 1.687455137048318e-05, "loss": 3.2535, "step": 7986 }, { "epoch": 0.28143802249883276, "grad_norm": 1.2795641422271729, "learning_rate": 1.687372249985939e-05, "loss": 0.9723, "step": 7987 }, { "epoch": 0.2814732595117912, "grad_norm": 3.702727794647217, "learning_rate": 1.6872893539704488e-05, "loss": 3.2866, "step": 7988 }, { "epoch": 0.2815084965247496, "grad_norm": 4.720399856567383, "learning_rate": 1.6872064490029267e-05, "loss": 3.572, "step": 7989 }, { "epoch": 0.281543733537708, "grad_norm": 1.387730360031128, "learning_rate": 1.6871235350844538e-05, "loss": 0.9615, "step": 7990 }, { "epoch": 0.28157897055066644, "grad_norm": 6.436264514923096, "learning_rate": 1.6870406122161087e-05, "loss": 6.3196, "step": 7991 }, { "epoch": 0.2816142075636248, "grad_norm": 1.5465145111083984, "learning_rate": 1.6869576803989724e-05, "loss": 0.9153, "step": 7992 }, { "epoch": 0.28164944457658325, "grad_norm": 3.8483128547668457, "learning_rate": 1.6868747396341246e-05, "loss": 3.9451, "step": 7993 }, { "epoch": 0.28168468158954163, "grad_norm": 4.868043899536133, "learning_rate": 1.686791789922646e-05, "loss": 5.9205, "step": 7994 }, { "epoch": 0.28171991860250006, "grad_norm": 2.419816255569458, "learning_rate": 1.686708831265617e-05, "loss": 1.1106, "step": 7995 }, { "epoch": 0.2817551556154585, "grad_norm": 4.579124927520752, "learning_rate": 1.686625863664118e-05, "loss": 3.3476, "step": 7996 }, { "epoch": 0.2817903926284169, "grad_norm": 1.122292160987854, "learning_rate": 1.6865428871192295e-05, "loss": 0.9103, "step": 7997 }, { "epoch": 0.2818256296413753, "grad_norm": 5.808459758758545, "learning_rate": 1.6864599016320324e-05, "loss": 3.4251, "step": 7998 }, { "epoch": 0.2818608666543337, "grad_norm": 5.231362819671631, "learning_rate": 1.6863769072036078e-05, "loss": 3.9883, "step": 7999 }, { "epoch": 0.2818961036672921, "grad_norm": 5.425581932067871, "learning_rate": 1.6862939038350363e-05, "loss": 3.1167, "step": 8000 }, { "epoch": 0.28193134068025055, "grad_norm": 3.423560857772827, "learning_rate": 1.6862108915273995e-05, "loss": 4.157, "step": 8001 }, { "epoch": 0.28196657769320893, "grad_norm": 3.5914411544799805, "learning_rate": 1.6861278702817782e-05, "loss": 2.4796, "step": 8002 }, { "epoch": 0.28200181470616736, "grad_norm": 4.607877731323242, "learning_rate": 1.6860448400992543e-05, "loss": 5.5208, "step": 8003 }, { "epoch": 0.28203705171912574, "grad_norm": 1.7218931913375854, "learning_rate": 1.6859618009809092e-05, "loss": 0.8334, "step": 8004 }, { "epoch": 0.2820722887320842, "grad_norm": 1.1910808086395264, "learning_rate": 1.6858787529278237e-05, "loss": 0.7841, "step": 8005 }, { "epoch": 0.2821075257450426, "grad_norm": 0.9277487397193909, "learning_rate": 1.68579569594108e-05, "loss": 1.0083, "step": 8006 }, { "epoch": 0.282142762758001, "grad_norm": 1.1993141174316406, "learning_rate": 1.6857126300217607e-05, "loss": 1.0443, "step": 8007 }, { "epoch": 0.2821779997709594, "grad_norm": 1.389484167098999, "learning_rate": 1.685629555170946e-05, "loss": 0.9644, "step": 8008 }, { "epoch": 0.28221323678391785, "grad_norm": 1.4083222150802612, "learning_rate": 1.68554647138972e-05, "loss": 1.0137, "step": 8009 }, { "epoch": 0.28224847379687623, "grad_norm": 3.051438093185425, "learning_rate": 1.6854633786791635e-05, "loss": 3.374, "step": 8010 }, { "epoch": 0.28228371080983466, "grad_norm": 1.4287856817245483, "learning_rate": 1.685380277040359e-05, "loss": 0.9769, "step": 8011 }, { "epoch": 0.28231894782279304, "grad_norm": 4.050754070281982, "learning_rate": 1.6852971664743895e-05, "loss": 5.8042, "step": 8012 }, { "epoch": 0.2823541848357515, "grad_norm": 1.7724461555480957, "learning_rate": 1.6852140469823372e-05, "loss": 0.9102, "step": 8013 }, { "epoch": 0.2823894218487099, "grad_norm": 4.588146209716797, "learning_rate": 1.685130918565284e-05, "loss": 4.8755, "step": 8014 }, { "epoch": 0.2824246588616683, "grad_norm": 1.1134675741195679, "learning_rate": 1.6850477812243137e-05, "loss": 0.8951, "step": 8015 }, { "epoch": 0.2824598958746267, "grad_norm": 6.614449977874756, "learning_rate": 1.6849646349605088e-05, "loss": 5.9735, "step": 8016 }, { "epoch": 0.2824951328875851, "grad_norm": 2.846343994140625, "learning_rate": 1.6848814797749523e-05, "loss": 2.9249, "step": 8017 }, { "epoch": 0.28253036990054353, "grad_norm": 1.6768155097961426, "learning_rate": 1.684798315668727e-05, "loss": 0.9233, "step": 8018 }, { "epoch": 0.28256560691350197, "grad_norm": 6.0117363929748535, "learning_rate": 1.684715142642917e-05, "loss": 3.6554, "step": 8019 }, { "epoch": 0.28260084392646034, "grad_norm": 5.58346700668335, "learning_rate": 1.6846319606986043e-05, "loss": 5.7336, "step": 8020 }, { "epoch": 0.2826360809394188, "grad_norm": 4.137393474578857, "learning_rate": 1.6845487698368735e-05, "loss": 3.9124, "step": 8021 }, { "epoch": 0.28267131795237715, "grad_norm": 4.449352264404297, "learning_rate": 1.684465570058808e-05, "loss": 5.0909, "step": 8022 }, { "epoch": 0.2827065549653356, "grad_norm": 2.2548868656158447, "learning_rate": 1.6843823613654912e-05, "loss": 2.8943, "step": 8023 }, { "epoch": 0.282741791978294, "grad_norm": 1.304806113243103, "learning_rate": 1.6842991437580066e-05, "loss": 1.0059, "step": 8024 }, { "epoch": 0.2827770289912524, "grad_norm": 1.615308403968811, "learning_rate": 1.6842159172374387e-05, "loss": 0.8435, "step": 8025 }, { "epoch": 0.28281226600421083, "grad_norm": 3.6576409339904785, "learning_rate": 1.684132681804871e-05, "loss": 3.3542, "step": 8026 }, { "epoch": 0.2828475030171692, "grad_norm": 1.5346555709838867, "learning_rate": 1.6840494374613886e-05, "loss": 0.8536, "step": 8027 }, { "epoch": 0.28288274003012764, "grad_norm": 4.534684181213379, "learning_rate": 1.683966184208075e-05, "loss": 3.4037, "step": 8028 }, { "epoch": 0.2829179770430861, "grad_norm": 5.800376892089844, "learning_rate": 1.6838829220460143e-05, "loss": 5.6022, "step": 8029 }, { "epoch": 0.28295321405604446, "grad_norm": 1.7792507410049438, "learning_rate": 1.6837996509762916e-05, "loss": 0.8258, "step": 8030 }, { "epoch": 0.2829884510690029, "grad_norm": 1.4830400943756104, "learning_rate": 1.6837163709999914e-05, "loss": 1.5067, "step": 8031 }, { "epoch": 0.28302368808196127, "grad_norm": 5.259676456451416, "learning_rate": 1.6836330821181985e-05, "loss": 3.741, "step": 8032 }, { "epoch": 0.2830589250949197, "grad_norm": 3.3752059936523438, "learning_rate": 1.6835497843319975e-05, "loss": 3.6343, "step": 8033 }, { "epoch": 0.28309416210787813, "grad_norm": 5.538804531097412, "learning_rate": 1.6834664776424733e-05, "loss": 5.83, "step": 8034 }, { "epoch": 0.2831293991208365, "grad_norm": 9.881524085998535, "learning_rate": 1.6833831620507114e-05, "loss": 2.8814, "step": 8035 }, { "epoch": 0.28316463613379494, "grad_norm": 5.597273826599121, "learning_rate": 1.683299837557797e-05, "loss": 3.8734, "step": 8036 }, { "epoch": 0.2831998731467533, "grad_norm": 7.583057403564453, "learning_rate": 1.6832165041648148e-05, "loss": 8.776, "step": 8037 }, { "epoch": 0.28323511015971176, "grad_norm": 4.830810546875, "learning_rate": 1.6831331618728505e-05, "loss": 6.0478, "step": 8038 }, { "epoch": 0.2832703471726702, "grad_norm": 3.016721487045288, "learning_rate": 1.68304981068299e-05, "loss": 3.8658, "step": 8039 }, { "epoch": 0.28330558418562857, "grad_norm": 1.8458722829818726, "learning_rate": 1.6829664505963185e-05, "loss": 0.6892, "step": 8040 }, { "epoch": 0.283340821198587, "grad_norm": 1.8770012855529785, "learning_rate": 1.6828830816139222e-05, "loss": 0.8743, "step": 8041 }, { "epoch": 0.28337605821154543, "grad_norm": 1.3132762908935547, "learning_rate": 1.682799703736887e-05, "loss": 1.0324, "step": 8042 }, { "epoch": 0.2834112952245038, "grad_norm": 4.66558837890625, "learning_rate": 1.6827163169662982e-05, "loss": 6.1159, "step": 8043 }, { "epoch": 0.28344653223746225, "grad_norm": 6.768591403961182, "learning_rate": 1.682632921303243e-05, "loss": 5.2772, "step": 8044 }, { "epoch": 0.2834817692504206, "grad_norm": 6.528310775756836, "learning_rate": 1.6825495167488066e-05, "loss": 3.2458, "step": 8045 }, { "epoch": 0.28351700626337906, "grad_norm": 4.898562908172607, "learning_rate": 1.6824661033040758e-05, "loss": 3.639, "step": 8046 }, { "epoch": 0.2835522432763375, "grad_norm": 2.183718204498291, "learning_rate": 1.6823826809701374e-05, "loss": 1.0225, "step": 8047 }, { "epoch": 0.28358748028929587, "grad_norm": 3.7384321689605713, "learning_rate": 1.682299249748077e-05, "loss": 3.3103, "step": 8048 }, { "epoch": 0.2836227173022543, "grad_norm": 2.019315242767334, "learning_rate": 1.6822158096389823e-05, "loss": 1.1262, "step": 8049 }, { "epoch": 0.2836579543152127, "grad_norm": 2.307638168334961, "learning_rate": 1.6821323606439402e-05, "loss": 3.6417, "step": 8050 }, { "epoch": 0.2836931913281711, "grad_norm": 6.628730297088623, "learning_rate": 1.6820489027640368e-05, "loss": 6.3615, "step": 8051 }, { "epoch": 0.28372842834112955, "grad_norm": 2.7044970989227295, "learning_rate": 1.6819654360003597e-05, "loss": 3.7998, "step": 8052 }, { "epoch": 0.2837636653540879, "grad_norm": 2.5174992084503174, "learning_rate": 1.6818819603539958e-05, "loss": 2.9617, "step": 8053 }, { "epoch": 0.28379890236704636, "grad_norm": 5.124081611633301, "learning_rate": 1.6817984758260327e-05, "loss": 4.2909, "step": 8054 }, { "epoch": 0.28383413938000474, "grad_norm": 1.5583726167678833, "learning_rate": 1.6817149824175573e-05, "loss": 1.0589, "step": 8055 }, { "epoch": 0.28386937639296317, "grad_norm": 0.9160099029541016, "learning_rate": 1.6816314801296575e-05, "loss": 1.3463, "step": 8056 }, { "epoch": 0.2839046134059216, "grad_norm": 1.0042297840118408, "learning_rate": 1.681547968963421e-05, "loss": 0.8285, "step": 8057 }, { "epoch": 0.28393985041888, "grad_norm": 1.1577342748641968, "learning_rate": 1.6814644489199352e-05, "loss": 1.0655, "step": 8058 }, { "epoch": 0.2839750874318384, "grad_norm": 1.1837834119796753, "learning_rate": 1.6813809200002883e-05, "loss": 1.0418, "step": 8059 }, { "epoch": 0.2840103244447968, "grad_norm": 2.1292572021484375, "learning_rate": 1.681297382205568e-05, "loss": 2.9384, "step": 8060 }, { "epoch": 0.2840455614577552, "grad_norm": 2.493699789047241, "learning_rate": 1.681213835536863e-05, "loss": 3.0572, "step": 8061 }, { "epoch": 0.28408079847071366, "grad_norm": 2.1145098209381104, "learning_rate": 1.6811302799952605e-05, "loss": 3.0078, "step": 8062 }, { "epoch": 0.28411603548367204, "grad_norm": 3.5297839641571045, "learning_rate": 1.6810467155818494e-05, "loss": 3.1893, "step": 8063 }, { "epoch": 0.28415127249663047, "grad_norm": 1.0106842517852783, "learning_rate": 1.6809631422977182e-05, "loss": 1.0027, "step": 8064 }, { "epoch": 0.28418650950958885, "grad_norm": 3.7844676971435547, "learning_rate": 1.6808795601439553e-05, "loss": 2.5165, "step": 8065 }, { "epoch": 0.2842217465225473, "grad_norm": 3.6040589809417725, "learning_rate": 1.680795969121649e-05, "loss": 3.7094, "step": 8066 }, { "epoch": 0.2842569835355057, "grad_norm": 2.0933454036712646, "learning_rate": 1.680712369231889e-05, "loss": 0.9721, "step": 8067 }, { "epoch": 0.2842922205484641, "grad_norm": 2.5603861808776855, "learning_rate": 1.6806287604757638e-05, "loss": 3.7219, "step": 8068 }, { "epoch": 0.2843274575614225, "grad_norm": 1.537841558456421, "learning_rate": 1.6805451428543617e-05, "loss": 0.9576, "step": 8069 }, { "epoch": 0.2843626945743809, "grad_norm": 1.3519291877746582, "learning_rate": 1.680461516368773e-05, "loss": 1.0335, "step": 8070 }, { "epoch": 0.28439793158733934, "grad_norm": 3.742241859436035, "learning_rate": 1.6803778810200858e-05, "loss": 3.0762, "step": 8071 }, { "epoch": 0.28443316860029777, "grad_norm": 1.4791042804718018, "learning_rate": 1.6802942368093905e-05, "loss": 0.939, "step": 8072 }, { "epoch": 0.28446840561325615, "grad_norm": 1.2470232248306274, "learning_rate": 1.680210583737776e-05, "loss": 0.8715, "step": 8073 }, { "epoch": 0.2845036426262146, "grad_norm": 4.208126544952393, "learning_rate": 1.680126921806332e-05, "loss": 5.5141, "step": 8074 }, { "epoch": 0.284538879639173, "grad_norm": 3.7535226345062256, "learning_rate": 1.6800432510161483e-05, "loss": 3.8578, "step": 8075 }, { "epoch": 0.2845741166521314, "grad_norm": 1.769667625427246, "learning_rate": 1.6799595713683145e-05, "loss": 0.9503, "step": 8076 }, { "epoch": 0.2846093536650898, "grad_norm": 3.2476308345794678, "learning_rate": 1.6798758828639208e-05, "loss": 3.5823, "step": 8077 }, { "epoch": 0.2846445906780482, "grad_norm": 1.2102574110031128, "learning_rate": 1.6797921855040568e-05, "loss": 0.8899, "step": 8078 }, { "epoch": 0.28467982769100664, "grad_norm": 1.3463438749313354, "learning_rate": 1.679708479289813e-05, "loss": 1.0334, "step": 8079 }, { "epoch": 0.28471506470396507, "grad_norm": 4.108516216278076, "learning_rate": 1.6796247642222802e-05, "loss": 3.4007, "step": 8080 }, { "epoch": 0.28475030171692345, "grad_norm": 2.106994867324829, "learning_rate": 1.679541040302548e-05, "loss": 1.1939, "step": 8081 }, { "epoch": 0.2847855387298819, "grad_norm": 1.7665135860443115, "learning_rate": 1.679457307531707e-05, "loss": 2.6461, "step": 8082 }, { "epoch": 0.28482077574284026, "grad_norm": 4.164792537689209, "learning_rate": 1.679373565910848e-05, "loss": 4.3284, "step": 8083 }, { "epoch": 0.2848560127557987, "grad_norm": 5.213602066040039, "learning_rate": 1.679289815441062e-05, "loss": 6.3263, "step": 8084 }, { "epoch": 0.2848912497687571, "grad_norm": 1.1336901187896729, "learning_rate": 1.6792060561234397e-05, "loss": 0.8093, "step": 8085 }, { "epoch": 0.2849264867817155, "grad_norm": 2.670651912689209, "learning_rate": 1.6791222879590718e-05, "loss": 2.9886, "step": 8086 }, { "epoch": 0.28496172379467394, "grad_norm": 1.2322580814361572, "learning_rate": 1.6790385109490498e-05, "loss": 0.8367, "step": 8087 }, { "epoch": 0.2849969608076323, "grad_norm": 4.478555679321289, "learning_rate": 1.6789547250944645e-05, "loss": 3.308, "step": 8088 }, { "epoch": 0.28503219782059075, "grad_norm": 1.3682998418807983, "learning_rate": 1.6788709303964074e-05, "loss": 1.092, "step": 8089 }, { "epoch": 0.2850674348335492, "grad_norm": 5.719285488128662, "learning_rate": 1.67878712685597e-05, "loss": 3.189, "step": 8090 }, { "epoch": 0.28510267184650756, "grad_norm": 1.3267186880111694, "learning_rate": 1.6787033144742437e-05, "loss": 0.8678, "step": 8091 }, { "epoch": 0.285137908859466, "grad_norm": 3.2157440185546875, "learning_rate": 1.6786194932523206e-05, "loss": 3.3692, "step": 8092 }, { "epoch": 0.28517314587242437, "grad_norm": 1.657127022743225, "learning_rate": 1.6785356631912918e-05, "loss": 0.9608, "step": 8093 }, { "epoch": 0.2852083828853828, "grad_norm": 4.83091402053833, "learning_rate": 1.6784518242922498e-05, "loss": 5.8485, "step": 8094 }, { "epoch": 0.28524361989834124, "grad_norm": 2.824795722961426, "learning_rate": 1.6783679765562865e-05, "loss": 3.2105, "step": 8095 }, { "epoch": 0.2852788569112996, "grad_norm": 1.1043657064437866, "learning_rate": 1.6782841199844935e-05, "loss": 1.1379, "step": 8096 }, { "epoch": 0.28531409392425805, "grad_norm": 3.0360703468322754, "learning_rate": 1.6782002545779637e-05, "loss": 3.1386, "step": 8097 }, { "epoch": 0.2853493309372164, "grad_norm": 4.93437385559082, "learning_rate": 1.6781163803377892e-05, "loss": 3.6468, "step": 8098 }, { "epoch": 0.28538456795017486, "grad_norm": 1.2073496580123901, "learning_rate": 1.6780324972650622e-05, "loss": 0.7538, "step": 8099 }, { "epoch": 0.2854198049631333, "grad_norm": 1.253467321395874, "learning_rate": 1.6779486053608755e-05, "loss": 0.5984, "step": 8100 }, { "epoch": 0.2854550419760917, "grad_norm": 3.245706081390381, "learning_rate": 1.6778647046263223e-05, "loss": 0.7557, "step": 8101 }, { "epoch": 0.2854902789890501, "grad_norm": 1.3541468381881714, "learning_rate": 1.6777807950624948e-05, "loss": 0.8022, "step": 8102 }, { "epoch": 0.2855255160020085, "grad_norm": 1.2482482194900513, "learning_rate": 1.6776968766704864e-05, "loss": 1.3233, "step": 8103 }, { "epoch": 0.2855607530149669, "grad_norm": 1.7619423866271973, "learning_rate": 1.6776129494513894e-05, "loss": 1.0518, "step": 8104 }, { "epoch": 0.28559599002792535, "grad_norm": 3.200392961502075, "learning_rate": 1.6775290134062975e-05, "loss": 3.4004, "step": 8105 }, { "epoch": 0.28563122704088373, "grad_norm": 2.443279266357422, "learning_rate": 1.6774450685363038e-05, "loss": 1.0228, "step": 8106 }, { "epoch": 0.28566646405384216, "grad_norm": 6.036070346832275, "learning_rate": 1.6773611148425023e-05, "loss": 3.6919, "step": 8107 }, { "epoch": 0.28570170106680054, "grad_norm": 9.244568824768066, "learning_rate": 1.6772771523259855e-05, "loss": 3.2886, "step": 8108 }, { "epoch": 0.285736938079759, "grad_norm": 7.620248317718506, "learning_rate": 1.6771931809878473e-05, "loss": 5.7156, "step": 8109 }, { "epoch": 0.2857721750927174, "grad_norm": 4.266474723815918, "learning_rate": 1.6771092008291827e-05, "loss": 3.1469, "step": 8110 }, { "epoch": 0.2858074121056758, "grad_norm": 6.540884017944336, "learning_rate": 1.6770252118510836e-05, "loss": 5.814, "step": 8111 }, { "epoch": 0.2858426491186342, "grad_norm": 1.280677080154419, "learning_rate": 1.676941214054645e-05, "loss": 0.9283, "step": 8112 }, { "epoch": 0.28587788613159265, "grad_norm": 5.349303245544434, "learning_rate": 1.676857207440961e-05, "loss": 4.0569, "step": 8113 }, { "epoch": 0.28591312314455103, "grad_norm": 4.0323286056518555, "learning_rate": 1.6767731920111257e-05, "loss": 3.4476, "step": 8114 }, { "epoch": 0.28594836015750946, "grad_norm": 1.1519346237182617, "learning_rate": 1.6766891677662332e-05, "loss": 0.85, "step": 8115 }, { "epoch": 0.28598359717046784, "grad_norm": 1.3562159538269043, "learning_rate": 1.676605134707378e-05, "loss": 0.956, "step": 8116 }, { "epoch": 0.2860188341834263, "grad_norm": 4.425652027130127, "learning_rate": 1.676521092835655e-05, "loss": 7.9402, "step": 8117 }, { "epoch": 0.2860540711963847, "grad_norm": 1.0090299844741821, "learning_rate": 1.6764370421521585e-05, "loss": 1.2166, "step": 8118 }, { "epoch": 0.2860893082093431, "grad_norm": 4.304027557373047, "learning_rate": 1.676352982657983e-05, "loss": 4.1971, "step": 8119 }, { "epoch": 0.2861245452223015, "grad_norm": 2.945815086364746, "learning_rate": 1.6762689143542243e-05, "loss": 2.8488, "step": 8120 }, { "epoch": 0.2861597822352599, "grad_norm": 1.5319125652313232, "learning_rate": 1.6761848372419767e-05, "loss": 1.174, "step": 8121 }, { "epoch": 0.28619501924821833, "grad_norm": 8.14255142211914, "learning_rate": 1.6761007513223355e-05, "loss": 6.1128, "step": 8122 }, { "epoch": 0.28623025626117676, "grad_norm": 2.752987861633301, "learning_rate": 1.676016656596396e-05, "loss": 2.9443, "step": 8123 }, { "epoch": 0.28626549327413514, "grad_norm": 4.914797782897949, "learning_rate": 1.675932553065253e-05, "loss": 6.4589, "step": 8124 }, { "epoch": 0.2863007302870936, "grad_norm": 2.176074266433716, "learning_rate": 1.675848440730003e-05, "loss": 0.9061, "step": 8125 }, { "epoch": 0.28633596730005195, "grad_norm": 3.7611865997314453, "learning_rate": 1.6757643195917404e-05, "loss": 6.4649, "step": 8126 }, { "epoch": 0.2863712043130104, "grad_norm": 3.1220946311950684, "learning_rate": 1.6756801896515614e-05, "loss": 2.9531, "step": 8127 }, { "epoch": 0.2864064413259688, "grad_norm": 4.821173667907715, "learning_rate": 1.6755960509105623e-05, "loss": 6.2656, "step": 8128 }, { "epoch": 0.2864416783389272, "grad_norm": 1.471635341644287, "learning_rate": 1.6755119033698383e-05, "loss": 0.9091, "step": 8129 }, { "epoch": 0.28647691535188563, "grad_norm": 7.254426002502441, "learning_rate": 1.675427747030486e-05, "loss": 6.5961, "step": 8130 }, { "epoch": 0.286512152364844, "grad_norm": 1.8043724298477173, "learning_rate": 1.675343581893601e-05, "loss": 1.0255, "step": 8131 }, { "epoch": 0.28654738937780244, "grad_norm": 6.873076915740967, "learning_rate": 1.67525940796028e-05, "loss": 3.7713, "step": 8132 }, { "epoch": 0.2865826263907609, "grad_norm": 1.194899082183838, "learning_rate": 1.675175225231619e-05, "loss": 0.8172, "step": 8133 }, { "epoch": 0.28661786340371925, "grad_norm": 2.126208782196045, "learning_rate": 1.6750910337087153e-05, "loss": 0.946, "step": 8134 }, { "epoch": 0.2866531004166777, "grad_norm": 1.147605538368225, "learning_rate": 1.6750068333926642e-05, "loss": 0.7511, "step": 8135 }, { "epoch": 0.28668833742963606, "grad_norm": 5.7642340660095215, "learning_rate": 1.6749226242845633e-05, "loss": 6.6525, "step": 8136 }, { "epoch": 0.2867235744425945, "grad_norm": 3.615133047103882, "learning_rate": 1.6748384063855093e-05, "loss": 3.5082, "step": 8137 }, { "epoch": 0.28675881145555293, "grad_norm": 1.5568547248840332, "learning_rate": 1.6747541796965986e-05, "loss": 0.8333, "step": 8138 }, { "epoch": 0.2867940484685113, "grad_norm": 3.4249446392059326, "learning_rate": 1.6746699442189295e-05, "loss": 2.8754, "step": 8139 }, { "epoch": 0.28682928548146974, "grad_norm": 3.2438135147094727, "learning_rate": 1.6745856999535983e-05, "loss": 3.3987, "step": 8140 }, { "epoch": 0.2868645224944281, "grad_norm": 1.1566261053085327, "learning_rate": 1.674501446901702e-05, "loss": 1.2982, "step": 8141 }, { "epoch": 0.28689975950738655, "grad_norm": 3.607276439666748, "learning_rate": 1.674417185064339e-05, "loss": 3.1441, "step": 8142 }, { "epoch": 0.286934996520345, "grad_norm": 1.2917109727859497, "learning_rate": 1.674332914442606e-05, "loss": 0.9611, "step": 8143 }, { "epoch": 0.28697023353330336, "grad_norm": 4.5343017578125, "learning_rate": 1.6742486350376008e-05, "loss": 5.7064, "step": 8144 }, { "epoch": 0.2870054705462618, "grad_norm": 1.5364573001861572, "learning_rate": 1.6741643468504213e-05, "loss": 0.914, "step": 8145 }, { "epoch": 0.28704070755922023, "grad_norm": 5.069891929626465, "learning_rate": 1.674080049882165e-05, "loss": 3.4531, "step": 8146 }, { "epoch": 0.2870759445721786, "grad_norm": 3.1898250579833984, "learning_rate": 1.673995744133931e-05, "loss": 3.0574, "step": 8147 }, { "epoch": 0.28711118158513704, "grad_norm": 1.6902848482131958, "learning_rate": 1.6739114296068158e-05, "loss": 0.7033, "step": 8148 }, { "epoch": 0.2871464185980954, "grad_norm": 1.263906717300415, "learning_rate": 1.6738271063019186e-05, "loss": 1.1355, "step": 8149 }, { "epoch": 0.28718165561105385, "grad_norm": 1.5279473066329956, "learning_rate": 1.6737427742203374e-05, "loss": 1.1028, "step": 8150 }, { "epoch": 0.2872168926240123, "grad_norm": 1.23811674118042, "learning_rate": 1.673658433363171e-05, "loss": 0.9052, "step": 8151 }, { "epoch": 0.28725212963697067, "grad_norm": 5.675217151641846, "learning_rate": 1.6735740837315178e-05, "loss": 3.4567, "step": 8152 }, { "epoch": 0.2872873666499291, "grad_norm": 5.557251930236816, "learning_rate": 1.6734897253264762e-05, "loss": 3.7316, "step": 8153 }, { "epoch": 0.2873226036628875, "grad_norm": 4.651906490325928, "learning_rate": 1.673405358149145e-05, "loss": 3.706, "step": 8154 }, { "epoch": 0.2873578406758459, "grad_norm": 1.172972559928894, "learning_rate": 1.6733209822006233e-05, "loss": 0.7748, "step": 8155 }, { "epoch": 0.28739307768880434, "grad_norm": 6.929162979125977, "learning_rate": 1.67323659748201e-05, "loss": 6.1379, "step": 8156 }, { "epoch": 0.2874283147017627, "grad_norm": 1.6653332710266113, "learning_rate": 1.6731522039944045e-05, "loss": 0.8084, "step": 8157 }, { "epoch": 0.28746355171472115, "grad_norm": 5.110728740692139, "learning_rate": 1.6730678017389057e-05, "loss": 6.8476, "step": 8158 }, { "epoch": 0.28749878872767953, "grad_norm": 1.8675360679626465, "learning_rate": 1.6729833907166127e-05, "loss": 1.0628, "step": 8159 }, { "epoch": 0.28753402574063797, "grad_norm": 8.213935852050781, "learning_rate": 1.6728989709286256e-05, "loss": 3.5315, "step": 8160 }, { "epoch": 0.2875692627535964, "grad_norm": 1.662480115890503, "learning_rate": 1.672814542376044e-05, "loss": 0.8792, "step": 8161 }, { "epoch": 0.2876044997665548, "grad_norm": 1.1333118677139282, "learning_rate": 1.6727301050599666e-05, "loss": 1.0115, "step": 8162 }, { "epoch": 0.2876397367795132, "grad_norm": 7.046194076538086, "learning_rate": 1.6726456589814946e-05, "loss": 3.5213, "step": 8163 }, { "epoch": 0.2876749737924716, "grad_norm": 3.0834286212921143, "learning_rate": 1.6725612041417266e-05, "loss": 3.3821, "step": 8164 }, { "epoch": 0.28771021080543, "grad_norm": 1.1663556098937988, "learning_rate": 1.6724767405417635e-05, "loss": 0.7683, "step": 8165 }, { "epoch": 0.28774544781838846, "grad_norm": 5.268315315246582, "learning_rate": 1.6723922681827053e-05, "loss": 6.4423, "step": 8166 }, { "epoch": 0.28778068483134683, "grad_norm": 12.085624694824219, "learning_rate": 1.672307787065652e-05, "loss": 7.0097, "step": 8167 }, { "epoch": 0.28781592184430527, "grad_norm": 1.6366074085235596, "learning_rate": 1.6722232971917044e-05, "loss": 0.8961, "step": 8168 }, { "epoch": 0.28785115885726364, "grad_norm": 7.452389240264893, "learning_rate": 1.6721387985619626e-05, "loss": 6.5093, "step": 8169 }, { "epoch": 0.2878863958702221, "grad_norm": 6.145853042602539, "learning_rate": 1.6720542911775272e-05, "loss": 3.5957, "step": 8170 }, { "epoch": 0.2879216328831805, "grad_norm": 1.2057429552078247, "learning_rate": 1.6719697750394993e-05, "loss": 1.2562, "step": 8171 }, { "epoch": 0.2879568698961389, "grad_norm": 1.0051082372665405, "learning_rate": 1.6718852501489792e-05, "loss": 1.2356, "step": 8172 }, { "epoch": 0.2879921069090973, "grad_norm": 7.322829246520996, "learning_rate": 1.6718007165070683e-05, "loss": 8.4655, "step": 8173 }, { "epoch": 0.2880273439220557, "grad_norm": 1.7057596445083618, "learning_rate": 1.6717161741148672e-05, "loss": 0.9519, "step": 8174 }, { "epoch": 0.28806258093501413, "grad_norm": 3.7036325931549072, "learning_rate": 1.6716316229734775e-05, "loss": 3.0019, "step": 8175 }, { "epoch": 0.28809781794797257, "grad_norm": 4.425434112548828, "learning_rate": 1.6715470630840006e-05, "loss": 3.4587, "step": 8176 }, { "epoch": 0.28813305496093095, "grad_norm": 5.364171981811523, "learning_rate": 1.6714624944475373e-05, "loss": 3.0588, "step": 8177 }, { "epoch": 0.2881682919738894, "grad_norm": 3.1320602893829346, "learning_rate": 1.67137791706519e-05, "loss": 5.9604, "step": 8178 }, { "epoch": 0.2882035289868478, "grad_norm": 1.3812270164489746, "learning_rate": 1.6712933309380594e-05, "loss": 0.7707, "step": 8179 }, { "epoch": 0.2882387659998062, "grad_norm": 1.3699769973754883, "learning_rate": 1.6712087360672477e-05, "loss": 1.0337, "step": 8180 }, { "epoch": 0.2882740030127646, "grad_norm": 0.992444634437561, "learning_rate": 1.671124132453857e-05, "loss": 0.9627, "step": 8181 }, { "epoch": 0.288309240025723, "grad_norm": 4.033698558807373, "learning_rate": 1.6710395200989885e-05, "loss": 3.8448, "step": 8182 }, { "epoch": 0.28834447703868143, "grad_norm": 3.347693681716919, "learning_rate": 1.670954899003745e-05, "loss": 3.2441, "step": 8183 }, { "epoch": 0.28837971405163987, "grad_norm": 2.645296573638916, "learning_rate": 1.6708702691692285e-05, "loss": 2.9109, "step": 8184 }, { "epoch": 0.28841495106459825, "grad_norm": 6.117947101593018, "learning_rate": 1.6707856305965417e-05, "loss": 4.0508, "step": 8185 }, { "epoch": 0.2884501880775567, "grad_norm": 4.995124340057373, "learning_rate": 1.670700983286786e-05, "loss": 6.444, "step": 8186 }, { "epoch": 0.28848542509051506, "grad_norm": 1.038305401802063, "learning_rate": 1.6706163272410646e-05, "loss": 0.8195, "step": 8187 }, { "epoch": 0.2885206621034735, "grad_norm": 0.9314557313919067, "learning_rate": 1.6705316624604804e-05, "loss": 1.0002, "step": 8188 }, { "epoch": 0.2885558991164319, "grad_norm": 9.343517303466797, "learning_rate": 1.670446988946136e-05, "loss": 8.5656, "step": 8189 }, { "epoch": 0.2885911361293903, "grad_norm": 2.6847875118255615, "learning_rate": 1.6703623066991338e-05, "loss": 1.0111, "step": 8190 }, { "epoch": 0.28862637314234874, "grad_norm": 7.529713153839111, "learning_rate": 1.6702776157205774e-05, "loss": 4.3444, "step": 8191 }, { "epoch": 0.2886616101553071, "grad_norm": 8.897849082946777, "learning_rate": 1.6701929160115695e-05, "loss": 7.3993, "step": 8192 }, { "epoch": 0.28869684716826555, "grad_norm": 1.6518149375915527, "learning_rate": 1.6701082075732134e-05, "loss": 1.0956, "step": 8193 }, { "epoch": 0.288732084181224, "grad_norm": 1.3493252992630005, "learning_rate": 1.6700234904066128e-05, "loss": 0.5824, "step": 8194 }, { "epoch": 0.28876732119418236, "grad_norm": 0.8482406735420227, "learning_rate": 1.6699387645128708e-05, "loss": 0.8588, "step": 8195 }, { "epoch": 0.2888025582071408, "grad_norm": 1.7379428148269653, "learning_rate": 1.6698540298930913e-05, "loss": 0.8593, "step": 8196 }, { "epoch": 0.28883779522009917, "grad_norm": 2.325410842895508, "learning_rate": 1.6697692865483774e-05, "loss": 3.5192, "step": 8197 }, { "epoch": 0.2888730322330576, "grad_norm": 1.2602167129516602, "learning_rate": 1.6696845344798333e-05, "loss": 0.881, "step": 8198 }, { "epoch": 0.28890826924601604, "grad_norm": 1.0141777992248535, "learning_rate": 1.669599773688563e-05, "loss": 1.046, "step": 8199 }, { "epoch": 0.2889435062589744, "grad_norm": 1.0005640983581543, "learning_rate": 1.6695150041756697e-05, "loss": 1.0499, "step": 8200 }, { "epoch": 0.28897874327193285, "grad_norm": 9.910849571228027, "learning_rate": 1.669430225942259e-05, "loss": 3.9721, "step": 8201 }, { "epoch": 0.2890139802848912, "grad_norm": 4.920255184173584, "learning_rate": 1.669345438989434e-05, "loss": 5.4635, "step": 8202 }, { "epoch": 0.28904921729784966, "grad_norm": 4.11003303527832, "learning_rate": 1.6692606433182996e-05, "loss": 3.6798, "step": 8203 }, { "epoch": 0.2890844543108081, "grad_norm": 1.7044382095336914, "learning_rate": 1.66917583892996e-05, "loss": 1.0258, "step": 8204 }, { "epoch": 0.28911969132376647, "grad_norm": 1.2780802249908447, "learning_rate": 1.6690910258255196e-05, "loss": 0.8933, "step": 8205 }, { "epoch": 0.2891549283367249, "grad_norm": 0.9163757562637329, "learning_rate": 1.6690062040060834e-05, "loss": 1.1023, "step": 8206 }, { "epoch": 0.2891901653496833, "grad_norm": 1.003449559211731, "learning_rate": 1.6689213734727564e-05, "loss": 1.06, "step": 8207 }, { "epoch": 0.2892254023626417, "grad_norm": 5.930365562438965, "learning_rate": 1.6688365342266433e-05, "loss": 3.574, "step": 8208 }, { "epoch": 0.28926063937560015, "grad_norm": 6.5283203125, "learning_rate": 1.6687516862688487e-05, "loss": 3.076, "step": 8209 }, { "epoch": 0.2892958763885585, "grad_norm": 1.178066611289978, "learning_rate": 1.6686668296004787e-05, "loss": 0.851, "step": 8210 }, { "epoch": 0.28933111340151696, "grad_norm": 7.544891834259033, "learning_rate": 1.668581964222638e-05, "loss": 3.6668, "step": 8211 }, { "epoch": 0.28936635041447534, "grad_norm": 1.9644227027893066, "learning_rate": 1.6684970901364322e-05, "loss": 2.601, "step": 8212 }, { "epoch": 0.28940158742743377, "grad_norm": 2.08510684967041, "learning_rate": 1.668412207342966e-05, "loss": 1.0222, "step": 8213 }, { "epoch": 0.2894368244403922, "grad_norm": 1.0190479755401611, "learning_rate": 1.6683273158433463e-05, "loss": 0.8366, "step": 8214 }, { "epoch": 0.2894720614533506, "grad_norm": 4.335472106933594, "learning_rate": 1.6682424156386782e-05, "loss": 3.719, "step": 8215 }, { "epoch": 0.289507298466309, "grad_norm": 2.4976632595062256, "learning_rate": 1.6681575067300672e-05, "loss": 3.3501, "step": 8216 }, { "epoch": 0.28954253547926745, "grad_norm": 8.36284351348877, "learning_rate": 1.6680725891186192e-05, "loss": 7.5457, "step": 8217 }, { "epoch": 0.2895777724922258, "grad_norm": 1.1759063005447388, "learning_rate": 1.6679876628054416e-05, "loss": 1.1626, "step": 8218 }, { "epoch": 0.28961300950518426, "grad_norm": 4.167060852050781, "learning_rate": 1.667902727791639e-05, "loss": 3.5848, "step": 8219 }, { "epoch": 0.28964824651814264, "grad_norm": 2.276020050048828, "learning_rate": 1.6678177840783182e-05, "loss": 3.3035, "step": 8220 }, { "epoch": 0.28968348353110107, "grad_norm": 7.1392717361450195, "learning_rate": 1.6677328316665856e-05, "loss": 6.4048, "step": 8221 }, { "epoch": 0.2897187205440595, "grad_norm": 1.3449827432632446, "learning_rate": 1.6676478705575482e-05, "loss": 0.8556, "step": 8222 }, { "epoch": 0.2897539575570179, "grad_norm": 1.41024649143219, "learning_rate": 1.6675629007523122e-05, "loss": 0.7943, "step": 8223 }, { "epoch": 0.2897891945699763, "grad_norm": 1.2243589162826538, "learning_rate": 1.6674779222519842e-05, "loss": 1.0815, "step": 8224 }, { "epoch": 0.2898244315829347, "grad_norm": 2.134343385696411, "learning_rate": 1.6673929350576714e-05, "loss": 0.7459, "step": 8225 }, { "epoch": 0.2898596685958931, "grad_norm": 3.9750523567199707, "learning_rate": 1.6673079391704807e-05, "loss": 3.4108, "step": 8226 }, { "epoch": 0.28989490560885156, "grad_norm": 2.4731950759887695, "learning_rate": 1.6672229345915186e-05, "loss": 3.4989, "step": 8227 }, { "epoch": 0.28993014262180994, "grad_norm": 1.5509148836135864, "learning_rate": 1.6671379213218935e-05, "loss": 0.9567, "step": 8228 }, { "epoch": 0.28996537963476837, "grad_norm": 2.347754716873169, "learning_rate": 1.6670528993627112e-05, "loss": 0.7854, "step": 8229 }, { "epoch": 0.29000061664772675, "grad_norm": 2.647545099258423, "learning_rate": 1.6669678687150807e-05, "loss": 2.8819, "step": 8230 }, { "epoch": 0.2900358536606852, "grad_norm": 1.1600136756896973, "learning_rate": 1.666882829380108e-05, "loss": 1.0015, "step": 8231 }, { "epoch": 0.2900710906736436, "grad_norm": 3.934753179550171, "learning_rate": 1.666797781358902e-05, "loss": 3.7641, "step": 8232 }, { "epoch": 0.290106327686602, "grad_norm": 1.557212233543396, "learning_rate": 1.66671272465257e-05, "loss": 0.8796, "step": 8233 }, { "epoch": 0.29014156469956043, "grad_norm": 1.3237169981002808, "learning_rate": 1.6666276592622198e-05, "loss": 1.005, "step": 8234 }, { "epoch": 0.2901768017125188, "grad_norm": 1.491189956665039, "learning_rate": 1.6665425851889593e-05, "loss": 1.1897, "step": 8235 }, { "epoch": 0.29021203872547724, "grad_norm": 4.788797855377197, "learning_rate": 1.6664575024338968e-05, "loss": 6.232, "step": 8236 }, { "epoch": 0.2902472757384357, "grad_norm": 5.877912998199463, "learning_rate": 1.66637241099814e-05, "loss": 6.2503, "step": 8237 }, { "epoch": 0.29028251275139405, "grad_norm": 1.402687668800354, "learning_rate": 1.666287310882798e-05, "loss": 1.3549, "step": 8238 }, { "epoch": 0.2903177497643525, "grad_norm": 1.998080849647522, "learning_rate": 1.6662022020889792e-05, "loss": 0.9382, "step": 8239 }, { "epoch": 0.29035298677731086, "grad_norm": 2.633958339691162, "learning_rate": 1.6661170846177915e-05, "loss": 3.8003, "step": 8240 }, { "epoch": 0.2903882237902693, "grad_norm": 1.3445465564727783, "learning_rate": 1.666031958470344e-05, "loss": 0.9926, "step": 8241 }, { "epoch": 0.29042346080322773, "grad_norm": 3.370001792907715, "learning_rate": 1.6659468236477455e-05, "loss": 3.0566, "step": 8242 }, { "epoch": 0.2904586978161861, "grad_norm": 8.376468658447266, "learning_rate": 1.6658616801511046e-05, "loss": 3.7555, "step": 8243 }, { "epoch": 0.29049393482914454, "grad_norm": 1.7704213857650757, "learning_rate": 1.6657765279815307e-05, "loss": 1.0334, "step": 8244 }, { "epoch": 0.2905291718421029, "grad_norm": 1.909467339515686, "learning_rate": 1.6656913671401328e-05, "loss": 0.8749, "step": 8245 }, { "epoch": 0.29056440885506135, "grad_norm": 1.364905834197998, "learning_rate": 1.6656061976280195e-05, "loss": 1.1024, "step": 8246 }, { "epoch": 0.2905996458680198, "grad_norm": 3.556431770324707, "learning_rate": 1.6655210194463013e-05, "loss": 2.8475, "step": 8247 }, { "epoch": 0.29063488288097816, "grad_norm": 1.6782350540161133, "learning_rate": 1.6654358325960866e-05, "loss": 0.9795, "step": 8248 }, { "epoch": 0.2906701198939366, "grad_norm": 4.536458492279053, "learning_rate": 1.6653506370784856e-05, "loss": 3.3749, "step": 8249 }, { "epoch": 0.29070535690689503, "grad_norm": 0.9225494265556335, "learning_rate": 1.665265432894608e-05, "loss": 0.9592, "step": 8250 }, { "epoch": 0.2907405939198534, "grad_norm": 5.676928997039795, "learning_rate": 1.665180220045563e-05, "loss": 6.3457, "step": 8251 }, { "epoch": 0.29077583093281184, "grad_norm": 2.3156661987304688, "learning_rate": 1.6650949985324612e-05, "loss": 0.9965, "step": 8252 }, { "epoch": 0.2908110679457702, "grad_norm": 1.5555437803268433, "learning_rate": 1.6650097683564123e-05, "loss": 0.9947, "step": 8253 }, { "epoch": 0.29084630495872865, "grad_norm": 5.9412312507629395, "learning_rate": 1.6649245295185266e-05, "loss": 3.8421, "step": 8254 }, { "epoch": 0.2908815419716871, "grad_norm": 8.627525329589844, "learning_rate": 1.664839282019914e-05, "loss": 3.4382, "step": 8255 }, { "epoch": 0.29091677898464546, "grad_norm": 1.7160667181015015, "learning_rate": 1.664754025861685e-05, "loss": 1.1357, "step": 8256 }, { "epoch": 0.2909520159976039, "grad_norm": 4.452415943145752, "learning_rate": 1.6646687610449502e-05, "loss": 6.7049, "step": 8257 }, { "epoch": 0.2909872530105623, "grad_norm": 1.0974411964416504, "learning_rate": 1.6645834875708203e-05, "loss": 1.058, "step": 8258 }, { "epoch": 0.2910224900235207, "grad_norm": 4.681451320648193, "learning_rate": 1.6644982054404054e-05, "loss": 5.2776, "step": 8259 }, { "epoch": 0.29105772703647914, "grad_norm": 1.28238046169281, "learning_rate": 1.664412914654817e-05, "loss": 0.8067, "step": 8260 }, { "epoch": 0.2910929640494375, "grad_norm": 9.100062370300293, "learning_rate": 1.6643276152151656e-05, "loss": 7.7415, "step": 8261 }, { "epoch": 0.29112820106239595, "grad_norm": 5.718661308288574, "learning_rate": 1.664242307122563e-05, "loss": 3.7772, "step": 8262 }, { "epoch": 0.29116343807535433, "grad_norm": 1.1040070056915283, "learning_rate": 1.664156990378119e-05, "loss": 1.0401, "step": 8263 }, { "epoch": 0.29119867508831276, "grad_norm": 1.980468511581421, "learning_rate": 1.664071664982946e-05, "loss": 0.7892, "step": 8264 }, { "epoch": 0.2912339121012712, "grad_norm": 1.1117390394210815, "learning_rate": 1.6639863309381547e-05, "loss": 1.0078, "step": 8265 }, { "epoch": 0.2912691491142296, "grad_norm": 4.549285411834717, "learning_rate": 1.6639009882448573e-05, "loss": 5.3935, "step": 8266 }, { "epoch": 0.291304386127188, "grad_norm": 1.1562093496322632, "learning_rate": 1.6638156369041642e-05, "loss": 0.8683, "step": 8267 }, { "epoch": 0.2913396231401464, "grad_norm": 1.0752670764923096, "learning_rate": 1.6637302769171886e-05, "loss": 0.9277, "step": 8268 }, { "epoch": 0.2913748601531048, "grad_norm": 4.304111003875732, "learning_rate": 1.6636449082850413e-05, "loss": 5.4838, "step": 8269 }, { "epoch": 0.29141009716606325, "grad_norm": 5.791869163513184, "learning_rate": 1.6635595310088347e-05, "loss": 3.9394, "step": 8270 }, { "epoch": 0.29144533417902163, "grad_norm": 2.2516767978668213, "learning_rate": 1.6634741450896804e-05, "loss": 0.9328, "step": 8271 }, { "epoch": 0.29148057119198006, "grad_norm": 5.472469806671143, "learning_rate": 1.663388750528691e-05, "loss": 5.7382, "step": 8272 }, { "epoch": 0.29151580820493844, "grad_norm": 1.7480387687683105, "learning_rate": 1.6633033473269788e-05, "loss": 1.2478, "step": 8273 }, { "epoch": 0.2915510452178969, "grad_norm": 3.6359121799468994, "learning_rate": 1.6632179354856558e-05, "loss": 3.3802, "step": 8274 }, { "epoch": 0.2915862822308553, "grad_norm": 1.201034665107727, "learning_rate": 1.6631325150058347e-05, "loss": 0.8965, "step": 8275 }, { "epoch": 0.2916215192438137, "grad_norm": 1.1637178659439087, "learning_rate": 1.663047085888628e-05, "loss": 0.7597, "step": 8276 }, { "epoch": 0.2916567562567721, "grad_norm": 1.3852745294570923, "learning_rate": 1.6629616481351488e-05, "loss": 0.9205, "step": 8277 }, { "epoch": 0.2916919932697305, "grad_norm": 1.360727310180664, "learning_rate": 1.6628762017465096e-05, "loss": 0.9544, "step": 8278 }, { "epoch": 0.29172723028268893, "grad_norm": 6.263474941253662, "learning_rate": 1.6627907467238237e-05, "loss": 8.7853, "step": 8279 }, { "epoch": 0.29176246729564737, "grad_norm": 2.7740063667297363, "learning_rate": 1.6627052830682034e-05, "loss": 3.0424, "step": 8280 }, { "epoch": 0.29179770430860574, "grad_norm": 3.5832972526550293, "learning_rate": 1.6626198107807624e-05, "loss": 2.7406, "step": 8281 }, { "epoch": 0.2918329413215642, "grad_norm": 2.560131788253784, "learning_rate": 1.6625343298626144e-05, "loss": 1.0028, "step": 8282 }, { "epoch": 0.2918681783345226, "grad_norm": 3.336606502532959, "learning_rate": 1.6624488403148723e-05, "loss": 3.326, "step": 8283 }, { "epoch": 0.291903415347481, "grad_norm": 0.8721257448196411, "learning_rate": 1.6623633421386492e-05, "loss": 0.9975, "step": 8284 }, { "epoch": 0.2919386523604394, "grad_norm": 1.2140547037124634, "learning_rate": 1.6622778353350598e-05, "loss": 1.0099, "step": 8285 }, { "epoch": 0.2919738893733978, "grad_norm": 5.450159549713135, "learning_rate": 1.6621923199052167e-05, "loss": 5.8073, "step": 8286 }, { "epoch": 0.29200912638635623, "grad_norm": 1.6325929164886475, "learning_rate": 1.6621067958502346e-05, "loss": 0.9214, "step": 8287 }, { "epoch": 0.29204436339931467, "grad_norm": 1.4298847913742065, "learning_rate": 1.6620212631712273e-05, "loss": 0.807, "step": 8288 }, { "epoch": 0.29207960041227304, "grad_norm": 9.274991989135742, "learning_rate": 1.6619357218693083e-05, "loss": 6.8919, "step": 8289 }, { "epoch": 0.2921148374252315, "grad_norm": 6.183311462402344, "learning_rate": 1.6618501719455924e-05, "loss": 3.461, "step": 8290 }, { "epoch": 0.29215007443818986, "grad_norm": 2.2224795818328857, "learning_rate": 1.6617646134011936e-05, "loss": 0.6764, "step": 8291 }, { "epoch": 0.2921853114511483, "grad_norm": 9.5662841796875, "learning_rate": 1.6616790462372266e-05, "loss": 3.8416, "step": 8292 }, { "epoch": 0.2922205484641067, "grad_norm": 2.1518919467926025, "learning_rate": 1.6615934704548054e-05, "loss": 0.7792, "step": 8293 }, { "epoch": 0.2922557854770651, "grad_norm": 1.03912353515625, "learning_rate": 1.6615078860550454e-05, "loss": 1.0477, "step": 8294 }, { "epoch": 0.29229102249002353, "grad_norm": 1.5264216661453247, "learning_rate": 1.6614222930390606e-05, "loss": 1.1751, "step": 8295 }, { "epoch": 0.2923262595029819, "grad_norm": 5.9539361000061035, "learning_rate": 1.6613366914079664e-05, "loss": 3.3464, "step": 8296 }, { "epoch": 0.29236149651594034, "grad_norm": 4.48108434677124, "learning_rate": 1.6612510811628774e-05, "loss": 5.7881, "step": 8297 }, { "epoch": 0.2923967335288988, "grad_norm": 1.1846319437026978, "learning_rate": 1.661165462304909e-05, "loss": 1.079, "step": 8298 }, { "epoch": 0.29243197054185716, "grad_norm": 4.720710277557373, "learning_rate": 1.661079834835176e-05, "loss": 3.1794, "step": 8299 }, { "epoch": 0.2924672075548156, "grad_norm": 5.04609489440918, "learning_rate": 1.660994198754794e-05, "loss": 4.069, "step": 8300 }, { "epoch": 0.29250244456777397, "grad_norm": 1.5688937902450562, "learning_rate": 1.6609085540648786e-05, "loss": 1.3232, "step": 8301 }, { "epoch": 0.2925376815807324, "grad_norm": 5.971279144287109, "learning_rate": 1.6608229007665452e-05, "loss": 3.8446, "step": 8302 }, { "epoch": 0.29257291859369083, "grad_norm": 5.945442199707031, "learning_rate": 1.660737238860909e-05, "loss": 5.8071, "step": 8303 }, { "epoch": 0.2926081556066492, "grad_norm": 1.3605822324752808, "learning_rate": 1.6606515683490864e-05, "loss": 1.1247, "step": 8304 }, { "epoch": 0.29264339261960765, "grad_norm": 7.771316051483154, "learning_rate": 1.6605658892321926e-05, "loss": 8.5164, "step": 8305 }, { "epoch": 0.292678629632566, "grad_norm": 0.8816561698913574, "learning_rate": 1.660480201511344e-05, "loss": 0.8231, "step": 8306 }, { "epoch": 0.29271386664552446, "grad_norm": 4.074199199676514, "learning_rate": 1.660394505187657e-05, "loss": 3.566, "step": 8307 }, { "epoch": 0.2927491036584829, "grad_norm": 3.298612594604492, "learning_rate": 1.660308800262247e-05, "loss": 3.1229, "step": 8308 }, { "epoch": 0.29278434067144127, "grad_norm": 3.7407867908477783, "learning_rate": 1.660223086736231e-05, "loss": 3.1946, "step": 8309 }, { "epoch": 0.2928195776843997, "grad_norm": 1.4872760772705078, "learning_rate": 1.660137364610725e-05, "loss": 1.3232, "step": 8310 }, { "epoch": 0.2928548146973581, "grad_norm": 3.930298328399658, "learning_rate": 1.660051633886846e-05, "loss": 3.6969, "step": 8311 }, { "epoch": 0.2928900517103165, "grad_norm": 1.5409934520721436, "learning_rate": 1.65996589456571e-05, "loss": 0.9116, "step": 8312 }, { "epoch": 0.29292528872327495, "grad_norm": 5.290115833282471, "learning_rate": 1.6598801466484348e-05, "loss": 2.7684, "step": 8313 }, { "epoch": 0.2929605257362333, "grad_norm": 3.8180861473083496, "learning_rate": 1.6597943901361363e-05, "loss": 3.6192, "step": 8314 }, { "epoch": 0.29299576274919176, "grad_norm": 3.8726019859313965, "learning_rate": 1.6597086250299316e-05, "loss": 4.0606, "step": 8315 }, { "epoch": 0.29303099976215014, "grad_norm": 1.4212822914123535, "learning_rate": 1.6596228513309385e-05, "loss": 0.8658, "step": 8316 }, { "epoch": 0.29306623677510857, "grad_norm": 1.2437939643859863, "learning_rate": 1.6595370690402733e-05, "loss": 0.9279, "step": 8317 }, { "epoch": 0.293101473788067, "grad_norm": 1.7341638803482056, "learning_rate": 1.659451278159054e-05, "loss": 0.8389, "step": 8318 }, { "epoch": 0.2931367108010254, "grad_norm": 2.8181049823760986, "learning_rate": 1.6593654786883976e-05, "loss": 3.8644, "step": 8319 }, { "epoch": 0.2931719478139838, "grad_norm": 1.22318696975708, "learning_rate": 1.6592796706294222e-05, "loss": 0.8682, "step": 8320 }, { "epoch": 0.29320718482694225, "grad_norm": 0.9767910242080688, "learning_rate": 1.6591938539832448e-05, "loss": 0.7102, "step": 8321 }, { "epoch": 0.2932424218399006, "grad_norm": 3.204592704772949, "learning_rate": 1.6591080287509836e-05, "loss": 3.9816, "step": 8322 }, { "epoch": 0.29327765885285906, "grad_norm": 3.8265976905822754, "learning_rate": 1.6590221949337563e-05, "loss": 3.3325, "step": 8323 }, { "epoch": 0.29331289586581744, "grad_norm": 1.7410980463027954, "learning_rate": 1.658936352532681e-05, "loss": 0.8617, "step": 8324 }, { "epoch": 0.29334813287877587, "grad_norm": 1.1011029481887817, "learning_rate": 1.658850501548876e-05, "loss": 1.1666, "step": 8325 }, { "epoch": 0.2933833698917343, "grad_norm": 2.071928024291992, "learning_rate": 1.658764641983459e-05, "loss": 1.094, "step": 8326 }, { "epoch": 0.2934186069046927, "grad_norm": 1.744603157043457, "learning_rate": 1.6586787738375486e-05, "loss": 0.9875, "step": 8327 }, { "epoch": 0.2934538439176511, "grad_norm": 4.251791477203369, "learning_rate": 1.6585928971122633e-05, "loss": 3.7928, "step": 8328 }, { "epoch": 0.2934890809306095, "grad_norm": 2.8706743717193604, "learning_rate": 1.6585070118087217e-05, "loss": 4.141, "step": 8329 }, { "epoch": 0.2935243179435679, "grad_norm": 1.5990324020385742, "learning_rate": 1.6584211179280425e-05, "loss": 0.9844, "step": 8330 }, { "epoch": 0.29355955495652636, "grad_norm": 4.34556770324707, "learning_rate": 1.6583352154713444e-05, "loss": 7.4931, "step": 8331 }, { "epoch": 0.29359479196948474, "grad_norm": 7.743198871612549, "learning_rate": 1.6582493044397462e-05, "loss": 9.0516, "step": 8332 }, { "epoch": 0.29363002898244317, "grad_norm": 1.167003870010376, "learning_rate": 1.6581633848343667e-05, "loss": 0.9824, "step": 8333 }, { "epoch": 0.29366526599540155, "grad_norm": 3.523071050643921, "learning_rate": 1.6580774566563257e-05, "loss": 3.1656, "step": 8334 }, { "epoch": 0.29370050300836, "grad_norm": 1.1459146738052368, "learning_rate": 1.6579915199067415e-05, "loss": 1.0544, "step": 8335 }, { "epoch": 0.2937357400213184, "grad_norm": 2.243971109390259, "learning_rate": 1.6579055745867342e-05, "loss": 0.9292, "step": 8336 }, { "epoch": 0.2937709770342768, "grad_norm": 1.740463137626648, "learning_rate": 1.657819620697423e-05, "loss": 0.9178, "step": 8337 }, { "epoch": 0.2938062140472352, "grad_norm": 4.909273147583008, "learning_rate": 1.6577336582399275e-05, "loss": 3.2611, "step": 8338 }, { "epoch": 0.2938414510601936, "grad_norm": 6.3571295738220215, "learning_rate": 1.657647687215367e-05, "loss": 5.8568, "step": 8339 }, { "epoch": 0.29387668807315204, "grad_norm": 2.1216256618499756, "learning_rate": 1.657561707624862e-05, "loss": 1.1401, "step": 8340 }, { "epoch": 0.29391192508611047, "grad_norm": 2.06246280670166, "learning_rate": 1.657475719469532e-05, "loss": 0.9041, "step": 8341 }, { "epoch": 0.29394716209906885, "grad_norm": 1.2922571897506714, "learning_rate": 1.6573897227504968e-05, "loss": 1.2105, "step": 8342 }, { "epoch": 0.2939823991120273, "grad_norm": 4.439757347106934, "learning_rate": 1.6573037174688765e-05, "loss": 3.7396, "step": 8343 }, { "epoch": 0.29401763612498566, "grad_norm": 2.965400457382202, "learning_rate": 1.657217703625792e-05, "loss": 3.1126, "step": 8344 }, { "epoch": 0.2940528731379441, "grad_norm": 5.137584209442139, "learning_rate": 1.6571316812223632e-05, "loss": 4.054, "step": 8345 }, { "epoch": 0.2940881101509025, "grad_norm": 5.219625949859619, "learning_rate": 1.6570456502597102e-05, "loss": 3.7217, "step": 8346 }, { "epoch": 0.2941233471638609, "grad_norm": 4.073866367340088, "learning_rate": 1.656959610738954e-05, "loss": 3.5174, "step": 8347 }, { "epoch": 0.29415858417681934, "grad_norm": 4.975142478942871, "learning_rate": 1.6568735626612154e-05, "loss": 5.2576, "step": 8348 }, { "epoch": 0.2941938211897777, "grad_norm": 6.396603584289551, "learning_rate": 1.6567875060276148e-05, "loss": 6.6838, "step": 8349 }, { "epoch": 0.29422905820273615, "grad_norm": 4.5240702629089355, "learning_rate": 1.6567014408392735e-05, "loss": 3.9583, "step": 8350 }, { "epoch": 0.2942642952156946, "grad_norm": 6.406025409698486, "learning_rate": 1.656615367097312e-05, "loss": 3.368, "step": 8351 }, { "epoch": 0.29429953222865296, "grad_norm": 1.4466532468795776, "learning_rate": 1.656529284802852e-05, "loss": 0.8441, "step": 8352 }, { "epoch": 0.2943347692416114, "grad_norm": 4.3962249755859375, "learning_rate": 1.656443193957014e-05, "loss": 3.7026, "step": 8353 }, { "epoch": 0.2943700062545698, "grad_norm": 1.0630100965499878, "learning_rate": 1.65635709456092e-05, "loss": 0.7707, "step": 8354 }, { "epoch": 0.2944052432675282, "grad_norm": 1.003881573677063, "learning_rate": 1.6562709866156913e-05, "loss": 1.0211, "step": 8355 }, { "epoch": 0.29444048028048664, "grad_norm": 3.4817466735839844, "learning_rate": 1.6561848701224495e-05, "loss": 3.0351, "step": 8356 }, { "epoch": 0.294475717293445, "grad_norm": 5.4688801765441895, "learning_rate": 1.656098745082316e-05, "loss": 3.981, "step": 8357 }, { "epoch": 0.29451095430640345, "grad_norm": 5.118573188781738, "learning_rate": 1.6560126114964128e-05, "loss": 5.9558, "step": 8358 }, { "epoch": 0.2945461913193619, "grad_norm": 3.7706222534179688, "learning_rate": 1.655926469365862e-05, "loss": 3.3983, "step": 8359 }, { "epoch": 0.29458142833232026, "grad_norm": 5.334504127502441, "learning_rate": 1.655840318691785e-05, "loss": 4.3956, "step": 8360 }, { "epoch": 0.2946166653452787, "grad_norm": 4.27299165725708, "learning_rate": 1.6557541594753045e-05, "loss": 3.1735, "step": 8361 }, { "epoch": 0.29465190235823707, "grad_norm": 3.7842659950256348, "learning_rate": 1.6556679917175423e-05, "loss": 4.7234, "step": 8362 }, { "epoch": 0.2946871393711955, "grad_norm": 1.1503868103027344, "learning_rate": 1.6555818154196216e-05, "loss": 1.0563, "step": 8363 }, { "epoch": 0.29472237638415394, "grad_norm": 1.6475872993469238, "learning_rate": 1.6554956305826636e-05, "loss": 0.7935, "step": 8364 }, { "epoch": 0.2947576133971123, "grad_norm": 4.525277137756348, "learning_rate": 1.655409437207792e-05, "loss": 4.1253, "step": 8365 }, { "epoch": 0.29479285041007075, "grad_norm": 4.718638896942139, "learning_rate": 1.6553232352961286e-05, "loss": 4.946, "step": 8366 }, { "epoch": 0.29482808742302913, "grad_norm": 4.270519733428955, "learning_rate": 1.6552370248487966e-05, "loss": 2.9138, "step": 8367 }, { "epoch": 0.29486332443598756, "grad_norm": 4.712353706359863, "learning_rate": 1.6551508058669193e-05, "loss": 3.8621, "step": 8368 }, { "epoch": 0.294898561448946, "grad_norm": 0.9093331098556519, "learning_rate": 1.655064578351619e-05, "loss": 1.0186, "step": 8369 }, { "epoch": 0.2949337984619044, "grad_norm": 1.2766071557998657, "learning_rate": 1.6549783423040193e-05, "loss": 0.772, "step": 8370 }, { "epoch": 0.2949690354748628, "grad_norm": 0.9475164413452148, "learning_rate": 1.654892097725243e-05, "loss": 0.7649, "step": 8371 }, { "epoch": 0.2950042724878212, "grad_norm": 2.324251651763916, "learning_rate": 1.654805844616414e-05, "loss": 0.7779, "step": 8372 }, { "epoch": 0.2950395095007796, "grad_norm": 1.0868127346038818, "learning_rate": 1.654719582978655e-05, "loss": 0.9096, "step": 8373 }, { "epoch": 0.29507474651373805, "grad_norm": 2.881882667541504, "learning_rate": 1.6546333128130906e-05, "loss": 3.6254, "step": 8374 }, { "epoch": 0.29510998352669643, "grad_norm": 4.174623012542725, "learning_rate": 1.6545470341208436e-05, "loss": 3.6416, "step": 8375 }, { "epoch": 0.29514522053965486, "grad_norm": 5.4476518630981445, "learning_rate": 1.654460746903038e-05, "loss": 5.7389, "step": 8376 }, { "epoch": 0.29518045755261324, "grad_norm": 5.975704193115234, "learning_rate": 1.654374451160798e-05, "loss": 5.7412, "step": 8377 }, { "epoch": 0.2952156945655717, "grad_norm": 2.7847957611083984, "learning_rate": 1.6542881468952475e-05, "loss": 0.86, "step": 8378 }, { "epoch": 0.2952509315785301, "grad_norm": 2.8300600051879883, "learning_rate": 1.6542018341075104e-05, "loss": 2.8959, "step": 8379 }, { "epoch": 0.2952861685914885, "grad_norm": 1.2400996685028076, "learning_rate": 1.6541155127987112e-05, "loss": 0.814, "step": 8380 }, { "epoch": 0.2953214056044469, "grad_norm": 1.2400996685028076, "learning_rate": 1.6541155127987112e-05, "loss": 1.1033, "step": 8381 }, { "epoch": 0.2953566426174053, "grad_norm": 1.7689229249954224, "learning_rate": 1.654029182969974e-05, "loss": 1.1541, "step": 8382 }, { "epoch": 0.29539187963036373, "grad_norm": 4.158848762512207, "learning_rate": 1.6539428446224234e-05, "loss": 3.2478, "step": 8383 }, { "epoch": 0.29542711664332216, "grad_norm": 5.5839409828186035, "learning_rate": 1.6538564977571837e-05, "loss": 0.9664, "step": 8384 }, { "epoch": 0.29546235365628054, "grad_norm": 6.5661234855651855, "learning_rate": 1.6537701423753804e-05, "loss": 3.9124, "step": 8385 }, { "epoch": 0.295497590669239, "grad_norm": 9.565437316894531, "learning_rate": 1.653683778478137e-05, "loss": 4.0783, "step": 8386 }, { "epoch": 0.2955328276821974, "grad_norm": 18.57109832763672, "learning_rate": 1.6535974060665795e-05, "loss": 3.8135, "step": 8387 }, { "epoch": 0.2955680646951558, "grad_norm": 14.29507827758789, "learning_rate": 1.6535110251418326e-05, "loss": 3.1586, "step": 8388 }, { "epoch": 0.2956033017081142, "grad_norm": 2.372647285461426, "learning_rate": 1.6534246357050216e-05, "loss": 3.3196, "step": 8389 }, { "epoch": 0.2956385387210726, "grad_norm": 5.788801670074463, "learning_rate": 1.653338237757271e-05, "loss": 2.9732, "step": 8390 }, { "epoch": 0.29567377573403103, "grad_norm": 2.031634569168091, "learning_rate": 1.653251831299707e-05, "loss": 0.8359, "step": 8391 }, { "epoch": 0.29570901274698946, "grad_norm": 3.8494927883148193, "learning_rate": 1.6531654163334546e-05, "loss": 3.6733, "step": 8392 }, { "epoch": 0.29574424975994784, "grad_norm": 1.8495509624481201, "learning_rate": 1.6530789928596393e-05, "loss": 1.0498, "step": 8393 }, { "epoch": 0.2957794867729063, "grad_norm": 2.052111864089966, "learning_rate": 1.652992560879387e-05, "loss": 0.9028, "step": 8394 }, { "epoch": 0.29581472378586465, "grad_norm": 1.6228933334350586, "learning_rate": 1.6529061203938233e-05, "loss": 0.7666, "step": 8395 }, { "epoch": 0.2958499607988231, "grad_norm": 2.989945650100708, "learning_rate": 1.6528196714040742e-05, "loss": 0.8138, "step": 8396 }, { "epoch": 0.2958851978117815, "grad_norm": 1.3861210346221924, "learning_rate": 1.652733213911266e-05, "loss": 1.0274, "step": 8397 }, { "epoch": 0.2959204348247399, "grad_norm": 2.835775375366211, "learning_rate": 1.6526467479165245e-05, "loss": 3.4101, "step": 8398 }, { "epoch": 0.29595567183769833, "grad_norm": 1.6459697484970093, "learning_rate": 1.6525602734209756e-05, "loss": 0.6605, "step": 8399 }, { "epoch": 0.2959909088506567, "grad_norm": 4.156427383422852, "learning_rate": 1.6524737904257466e-05, "loss": 3.5907, "step": 8400 }, { "epoch": 0.29602614586361514, "grad_norm": 6.014216899871826, "learning_rate": 1.652387298931963e-05, "loss": 3.2989, "step": 8401 }, { "epoch": 0.2960613828765736, "grad_norm": 2.8134589195251465, "learning_rate": 1.6523007989407517e-05, "loss": 0.9803, "step": 8402 }, { "epoch": 0.29609661988953195, "grad_norm": 3.045727014541626, "learning_rate": 1.6522142904532397e-05, "loss": 3.6109, "step": 8403 }, { "epoch": 0.2961318569024904, "grad_norm": 1.4535495042800903, "learning_rate": 1.652127773470553e-05, "loss": 0.866, "step": 8404 }, { "epoch": 0.29616709391544876, "grad_norm": 1.8473882675170898, "learning_rate": 1.6520412479938193e-05, "loss": 0.9159, "step": 8405 }, { "epoch": 0.2962023309284072, "grad_norm": 4.2531352043151855, "learning_rate": 1.651954714024165e-05, "loss": 5.2722, "step": 8406 }, { "epoch": 0.29623756794136563, "grad_norm": 1.5581995248794556, "learning_rate": 1.651868171562718e-05, "loss": 1.2111, "step": 8407 }, { "epoch": 0.296272804954324, "grad_norm": 2.3870959281921387, "learning_rate": 1.651781620610605e-05, "loss": 0.8081, "step": 8408 }, { "epoch": 0.29630804196728244, "grad_norm": 6.575413227081299, "learning_rate": 1.651695061168953e-05, "loss": 3.6419, "step": 8409 }, { "epoch": 0.2963432789802408, "grad_norm": 3.01957368850708, "learning_rate": 1.65160849323889e-05, "loss": 1.1501, "step": 8410 }, { "epoch": 0.29637851599319925, "grad_norm": 2.085247755050659, "learning_rate": 1.6515219168215433e-05, "loss": 0.9012, "step": 8411 }, { "epoch": 0.2964137530061577, "grad_norm": 1.9385040998458862, "learning_rate": 1.6514353319180406e-05, "loss": 1.1688, "step": 8412 }, { "epoch": 0.29644899001911607, "grad_norm": 4.275248050689697, "learning_rate": 1.65134873852951e-05, "loss": 2.8194, "step": 8413 }, { "epoch": 0.2964842270320745, "grad_norm": 1.7951934337615967, "learning_rate": 1.6512621366570787e-05, "loss": 0.9231, "step": 8414 }, { "epoch": 0.2965194640450329, "grad_norm": 1.7320902347564697, "learning_rate": 1.6511755263018753e-05, "loss": 1.2259, "step": 8415 }, { "epoch": 0.2965547010579913, "grad_norm": 1.494409441947937, "learning_rate": 1.651088907465028e-05, "loss": 0.893, "step": 8416 }, { "epoch": 0.29658993807094974, "grad_norm": 3.6649553775787354, "learning_rate": 1.6510022801476645e-05, "loss": 5.1736, "step": 8417 }, { "epoch": 0.2966251750839081, "grad_norm": 2.359886884689331, "learning_rate": 1.6509156443509132e-05, "loss": 0.963, "step": 8418 }, { "epoch": 0.29666041209686655, "grad_norm": 2.1536550521850586, "learning_rate": 1.650829000075903e-05, "loss": 0.9971, "step": 8419 }, { "epoch": 0.29669564910982493, "grad_norm": 3.1247408390045166, "learning_rate": 1.650742347323762e-05, "loss": 3.6366, "step": 8420 }, { "epoch": 0.29673088612278337, "grad_norm": 3.8709075450897217, "learning_rate": 1.6506556860956192e-05, "loss": 3.352, "step": 8421 }, { "epoch": 0.2967661231357418, "grad_norm": 4.848031997680664, "learning_rate": 1.6505690163926028e-05, "loss": 3.7379, "step": 8422 }, { "epoch": 0.2968013601487002, "grad_norm": 1.7424285411834717, "learning_rate": 1.6504823382158428e-05, "loss": 1.1493, "step": 8423 }, { "epoch": 0.2968365971616586, "grad_norm": 3.9977829456329346, "learning_rate": 1.650395651566467e-05, "loss": 5.3185, "step": 8424 }, { "epoch": 0.29687183417461704, "grad_norm": 1.3741974830627441, "learning_rate": 1.650308956445605e-05, "loss": 0.9576, "step": 8425 }, { "epoch": 0.2969070711875754, "grad_norm": 1.19340980052948, "learning_rate": 1.650222252854386e-05, "loss": 1.2052, "step": 8426 }, { "epoch": 0.29694230820053386, "grad_norm": 2.337634563446045, "learning_rate": 1.6501355407939397e-05, "loss": 0.9627, "step": 8427 }, { "epoch": 0.29697754521349223, "grad_norm": 2.462486743927002, "learning_rate": 1.6500488202653947e-05, "loss": 0.7733, "step": 8428 }, { "epoch": 0.29701278222645067, "grad_norm": 6.86929178237915, "learning_rate": 1.6499620912698812e-05, "loss": 5.7806, "step": 8429 }, { "epoch": 0.2970480192394091, "grad_norm": 3.457519054412842, "learning_rate": 1.6498753538085286e-05, "loss": 3.5697, "step": 8430 }, { "epoch": 0.2970832562523675, "grad_norm": 5.644266128540039, "learning_rate": 1.6497886078824668e-05, "loss": 3.2131, "step": 8431 }, { "epoch": 0.2971184932653259, "grad_norm": 6.2798686027526855, "learning_rate": 1.6497018534928256e-05, "loss": 4.2093, "step": 8432 }, { "epoch": 0.2971537302782843, "grad_norm": 11.706665992736816, "learning_rate": 1.649615090640735e-05, "loss": 3.6506, "step": 8433 }, { "epoch": 0.2971889672912427, "grad_norm": 1.7879530191421509, "learning_rate": 1.6495283193273252e-05, "loss": 0.831, "step": 8434 }, { "epoch": 0.29722420430420116, "grad_norm": 5.546927452087402, "learning_rate": 1.649441539553726e-05, "loss": 3.8335, "step": 8435 }, { "epoch": 0.29725944131715953, "grad_norm": 8.632643699645996, "learning_rate": 1.6493547513210682e-05, "loss": 3.1499, "step": 8436 }, { "epoch": 0.29729467833011797, "grad_norm": 4.188671588897705, "learning_rate": 1.6492679546304824e-05, "loss": 5.6484, "step": 8437 }, { "epoch": 0.29732991534307635, "grad_norm": 5.743952751159668, "learning_rate": 1.6491811494830987e-05, "loss": 4.9933, "step": 8438 }, { "epoch": 0.2973651523560348, "grad_norm": 4.354948043823242, "learning_rate": 1.6490943358800472e-05, "loss": 2.7399, "step": 8439 }, { "epoch": 0.2974003893689932, "grad_norm": 5.419212341308594, "learning_rate": 1.64900751382246e-05, "loss": 3.7312, "step": 8440 }, { "epoch": 0.2974356263819516, "grad_norm": 4.882862091064453, "learning_rate": 1.648920683311467e-05, "loss": 3.9016, "step": 8441 }, { "epoch": 0.29747086339491, "grad_norm": 2.317051649093628, "learning_rate": 1.6488338443481992e-05, "loss": 1.2036, "step": 8442 }, { "epoch": 0.2975061004078684, "grad_norm": 4.163803577423096, "learning_rate": 1.648746996933788e-05, "loss": 3.5476, "step": 8443 }, { "epoch": 0.29754133742082683, "grad_norm": 5.66461181640625, "learning_rate": 1.6486601410693647e-05, "loss": 3.6956, "step": 8444 }, { "epoch": 0.29757657443378527, "grad_norm": 1.7103140354156494, "learning_rate": 1.6485732767560604e-05, "loss": 0.8031, "step": 8445 }, { "epoch": 0.29761181144674365, "grad_norm": 6.165763854980469, "learning_rate": 1.6484864039950065e-05, "loss": 5.4031, "step": 8446 }, { "epoch": 0.2976470484597021, "grad_norm": 2.645771026611328, "learning_rate": 1.648399522787335e-05, "loss": 3.4927, "step": 8447 }, { "epoch": 0.29768228547266046, "grad_norm": 2.3884401321411133, "learning_rate": 1.6483126331341765e-05, "loss": 2.7951, "step": 8448 }, { "epoch": 0.2977175224856189, "grad_norm": 1.7228580713272095, "learning_rate": 1.6482257350366636e-05, "loss": 1.0859, "step": 8449 }, { "epoch": 0.2977527594985773, "grad_norm": 1.0990691184997559, "learning_rate": 1.648138828495928e-05, "loss": 0.9615, "step": 8450 }, { "epoch": 0.2977879965115357, "grad_norm": 1.1701334714889526, "learning_rate": 1.6480519135131016e-05, "loss": 0.7743, "step": 8451 }, { "epoch": 0.29782323352449414, "grad_norm": 1.5109719038009644, "learning_rate": 1.6479649900893165e-05, "loss": 1.0141, "step": 8452 }, { "epoch": 0.2978584705374525, "grad_norm": 5.196440696716309, "learning_rate": 1.647878058225705e-05, "loss": 3.824, "step": 8453 }, { "epoch": 0.29789370755041095, "grad_norm": 1.8678414821624756, "learning_rate": 1.6477911179233988e-05, "loss": 1.0744, "step": 8454 }, { "epoch": 0.2979289445633694, "grad_norm": 3.9501514434814453, "learning_rate": 1.647704169183531e-05, "loss": 3.4273, "step": 8455 }, { "epoch": 0.29796418157632776, "grad_norm": 1.4995458126068115, "learning_rate": 1.647617212007234e-05, "loss": 1.0122, "step": 8456 }, { "epoch": 0.2979994185892862, "grad_norm": 3.83652400970459, "learning_rate": 1.64753024639564e-05, "loss": 3.0626, "step": 8457 }, { "epoch": 0.2980346556022446, "grad_norm": 6.576342582702637, "learning_rate": 1.6474432723498827e-05, "loss": 5.3094, "step": 8458 }, { "epoch": 0.298069892615203, "grad_norm": 5.220147132873535, "learning_rate": 1.6473562898710936e-05, "loss": 6.0475, "step": 8459 }, { "epoch": 0.29810512962816144, "grad_norm": 2.8708882331848145, "learning_rate": 1.647269298960407e-05, "loss": 3.1948, "step": 8460 }, { "epoch": 0.2981403666411198, "grad_norm": 1.4834774732589722, "learning_rate": 1.6471822996189548e-05, "loss": 0.8851, "step": 8461 }, { "epoch": 0.29817560365407825, "grad_norm": 3.8550312519073486, "learning_rate": 1.647095291847871e-05, "loss": 3.1888, "step": 8462 }, { "epoch": 0.2982108406670367, "grad_norm": 2.4959983825683594, "learning_rate": 1.6470082756482883e-05, "loss": 1.066, "step": 8463 }, { "epoch": 0.29824607767999506, "grad_norm": 3.513420581817627, "learning_rate": 1.6469212510213407e-05, "loss": 3.742, "step": 8464 }, { "epoch": 0.2982813146929535, "grad_norm": 5.849065780639648, "learning_rate": 1.6468342179681615e-05, "loss": 6.7773, "step": 8465 }, { "epoch": 0.29831655170591187, "grad_norm": 2.976867914199829, "learning_rate": 1.646747176489884e-05, "loss": 3.3345, "step": 8466 }, { "epoch": 0.2983517887188703, "grad_norm": 9.197525978088379, "learning_rate": 1.646660126587642e-05, "loss": 3.379, "step": 8467 }, { "epoch": 0.29838702573182874, "grad_norm": 4.33927059173584, "learning_rate": 1.6465730682625698e-05, "loss": 5.5929, "step": 8468 }, { "epoch": 0.2984222627447871, "grad_norm": 1.4713022708892822, "learning_rate": 1.646486001515801e-05, "loss": 1.033, "step": 8469 }, { "epoch": 0.29845749975774555, "grad_norm": 1.6179851293563843, "learning_rate": 1.6463989263484696e-05, "loss": 0.8587, "step": 8470 }, { "epoch": 0.2984927367707039, "grad_norm": 1.5427860021591187, "learning_rate": 1.6463118427617097e-05, "loss": 0.9584, "step": 8471 }, { "epoch": 0.29852797378366236, "grad_norm": 2.4744274616241455, "learning_rate": 1.646224750756656e-05, "loss": 3.5616, "step": 8472 }, { "epoch": 0.2985632107966208, "grad_norm": 5.5714874267578125, "learning_rate": 1.6461376503344425e-05, "loss": 5.4557, "step": 8473 }, { "epoch": 0.29859844780957917, "grad_norm": 3.841517210006714, "learning_rate": 1.6460505414962033e-05, "loss": 3.5648, "step": 8474 }, { "epoch": 0.2986336848225376, "grad_norm": 1.295287013053894, "learning_rate": 1.6459634242430744e-05, "loss": 0.926, "step": 8475 }, { "epoch": 0.298668921835496, "grad_norm": 5.704367637634277, "learning_rate": 1.6458762985761886e-05, "loss": 3.0643, "step": 8476 }, { "epoch": 0.2987041588484544, "grad_norm": 5.156186103820801, "learning_rate": 1.6457891644966826e-05, "loss": 3.7568, "step": 8477 }, { "epoch": 0.29873939586141285, "grad_norm": 1.6416661739349365, "learning_rate": 1.6457020220056903e-05, "loss": 1.0087, "step": 8478 }, { "epoch": 0.2987746328743712, "grad_norm": 2.1343154907226562, "learning_rate": 1.6456148711043466e-05, "loss": 1.0038, "step": 8479 }, { "epoch": 0.29880986988732966, "grad_norm": 6.8952813148498535, "learning_rate": 1.645527711793787e-05, "loss": 3.2187, "step": 8480 }, { "epoch": 0.29884510690028804, "grad_norm": 3.569206714630127, "learning_rate": 1.645440544075147e-05, "loss": 3.3739, "step": 8481 }, { "epoch": 0.29888034391324647, "grad_norm": 3.265315055847168, "learning_rate": 1.645353367949562e-05, "loss": 3.0955, "step": 8482 }, { "epoch": 0.2989155809262049, "grad_norm": 1.619300365447998, "learning_rate": 1.6452661834181664e-05, "loss": 1.0783, "step": 8483 }, { "epoch": 0.2989508179391633, "grad_norm": 3.7115135192871094, "learning_rate": 1.645178990482097e-05, "loss": 3.7238, "step": 8484 }, { "epoch": 0.2989860549521217, "grad_norm": 3.0647497177124023, "learning_rate": 1.6450917891424885e-05, "loss": 2.4622, "step": 8485 }, { "epoch": 0.2990212919650801, "grad_norm": 4.8589396476745605, "learning_rate": 1.645004579400478e-05, "loss": 6.3343, "step": 8486 }, { "epoch": 0.2990565289780385, "grad_norm": 4.440999507904053, "learning_rate": 1.6449173612572006e-05, "loss": 3.7643, "step": 8487 }, { "epoch": 0.29909176599099696, "grad_norm": 1.4681721925735474, "learning_rate": 1.6448301347137923e-05, "loss": 0.9097, "step": 8488 }, { "epoch": 0.29912700300395534, "grad_norm": 1.3673253059387207, "learning_rate": 1.6447428997713896e-05, "loss": 0.7114, "step": 8489 }, { "epoch": 0.29916224001691377, "grad_norm": 1.051418423652649, "learning_rate": 1.644655656431128e-05, "loss": 0.8642, "step": 8490 }, { "epoch": 0.2991974770298722, "grad_norm": 1.5879650115966797, "learning_rate": 1.6445684046941445e-05, "loss": 1.102, "step": 8491 }, { "epoch": 0.2992327140428306, "grad_norm": 5.514323711395264, "learning_rate": 1.6444811445615755e-05, "loss": 5.9539, "step": 8492 }, { "epoch": 0.299267951055789, "grad_norm": 1.4590038061141968, "learning_rate": 1.6443938760345574e-05, "loss": 0.8806, "step": 8493 }, { "epoch": 0.2993031880687474, "grad_norm": 6.2194504737854, "learning_rate": 1.6443065991142274e-05, "loss": 6.3402, "step": 8494 }, { "epoch": 0.29933842508170583, "grad_norm": 1.4365431070327759, "learning_rate": 1.6442193138017213e-05, "loss": 0.901, "step": 8495 }, { "epoch": 0.29937366209466426, "grad_norm": 1.320460319519043, "learning_rate": 1.6441320200981768e-05, "loss": 0.858, "step": 8496 }, { "epoch": 0.29940889910762264, "grad_norm": 1.490483045578003, "learning_rate": 1.644044718004731e-05, "loss": 0.8693, "step": 8497 }, { "epoch": 0.2994441361205811, "grad_norm": 1.5078771114349365, "learning_rate": 1.6439574075225203e-05, "loss": 0.9226, "step": 8498 }, { "epoch": 0.29947937313353945, "grad_norm": 1.3385522365570068, "learning_rate": 1.6438700886526826e-05, "loss": 1.0982, "step": 8499 }, { "epoch": 0.2995146101464979, "grad_norm": 4.457272052764893, "learning_rate": 1.643782761396355e-05, "loss": 3.082, "step": 8500 }, { "epoch": 0.2995498471594563, "grad_norm": 1.1359080076217651, "learning_rate": 1.6436954257546744e-05, "loss": 0.9897, "step": 8501 }, { "epoch": 0.2995850841724147, "grad_norm": 2.0165414810180664, "learning_rate": 1.6436080817287795e-05, "loss": 1.0921, "step": 8502 }, { "epoch": 0.29962032118537313, "grad_norm": 9.622147560119629, "learning_rate": 1.643520729319807e-05, "loss": 8.8778, "step": 8503 }, { "epoch": 0.2996555581983315, "grad_norm": 1.4161053895950317, "learning_rate": 1.6434333685288953e-05, "loss": 0.8594, "step": 8504 }, { "epoch": 0.29969079521128994, "grad_norm": 4.68623685836792, "learning_rate": 1.6433459993571818e-05, "loss": 5.9031, "step": 8505 }, { "epoch": 0.2997260322242484, "grad_norm": 2.19657826423645, "learning_rate": 1.6432586218058045e-05, "loss": 0.9375, "step": 8506 }, { "epoch": 0.29976126923720675, "grad_norm": 1.3388949632644653, "learning_rate": 1.6431712358759023e-05, "loss": 1.0831, "step": 8507 }, { "epoch": 0.2997965062501652, "grad_norm": 1.5256339311599731, "learning_rate": 1.6430838415686124e-05, "loss": 1.002, "step": 8508 }, { "epoch": 0.29983174326312356, "grad_norm": 8.400982856750488, "learning_rate": 1.6429964388850737e-05, "loss": 6.5939, "step": 8509 }, { "epoch": 0.299866980276082, "grad_norm": 7.422930717468262, "learning_rate": 1.6429090278264246e-05, "loss": 6.6162, "step": 8510 }, { "epoch": 0.29990221728904043, "grad_norm": 4.964826583862305, "learning_rate": 1.6428216083938034e-05, "loss": 3.773, "step": 8511 }, { "epoch": 0.2999374543019988, "grad_norm": 5.523885726928711, "learning_rate": 1.6427341805883487e-05, "loss": 3.8031, "step": 8512 }, { "epoch": 0.29997269131495724, "grad_norm": 3.8250229358673096, "learning_rate": 1.6426467444111994e-05, "loss": 2.9576, "step": 8513 }, { "epoch": 0.3000079283279156, "grad_norm": 1.1951228380203247, "learning_rate": 1.6425592998634947e-05, "loss": 1.0161, "step": 8514 }, { "epoch": 0.30004316534087405, "grad_norm": 1.0841883420944214, "learning_rate": 1.6424718469463733e-05, "loss": 1.0849, "step": 8515 }, { "epoch": 0.3000784023538325, "grad_norm": 8.815898895263672, "learning_rate": 1.6423843856609743e-05, "loss": 5.3552, "step": 8516 }, { "epoch": 0.30011363936679086, "grad_norm": 1.493571162223816, "learning_rate": 1.6422969160084362e-05, "loss": 0.8325, "step": 8517 }, { "epoch": 0.3001488763797493, "grad_norm": 2.216438055038452, "learning_rate": 1.6422094379898997e-05, "loss": 1.1493, "step": 8518 }, { "epoch": 0.3001841133927077, "grad_norm": 3.014270782470703, "learning_rate": 1.642121951606503e-05, "loss": 0.9407, "step": 8519 }, { "epoch": 0.3002193504056661, "grad_norm": 1.4342634677886963, "learning_rate": 1.642034456859386e-05, "loss": 0.8768, "step": 8520 }, { "epoch": 0.30025458741862454, "grad_norm": 6.1387410163879395, "learning_rate": 1.641946953749689e-05, "loss": 3.5339, "step": 8521 }, { "epoch": 0.3002898244315829, "grad_norm": 5.896417140960693, "learning_rate": 1.6418594422785504e-05, "loss": 5.1653, "step": 8522 }, { "epoch": 0.30032506144454135, "grad_norm": 3.582230806350708, "learning_rate": 1.6417719224471116e-05, "loss": 2.7238, "step": 8523 }, { "epoch": 0.30036029845749973, "grad_norm": 1.7750049829483032, "learning_rate": 1.6416843942565114e-05, "loss": 0.7858, "step": 8524 }, { "epoch": 0.30039553547045816, "grad_norm": 7.626518726348877, "learning_rate": 1.64159685770789e-05, "loss": 9.0158, "step": 8525 }, { "epoch": 0.3004307724834166, "grad_norm": 1.520544171333313, "learning_rate": 1.6415093128023882e-05, "loss": 1.0621, "step": 8526 }, { "epoch": 0.300466009496375, "grad_norm": 8.412872314453125, "learning_rate": 1.6414217595411455e-05, "loss": 3.6189, "step": 8527 }, { "epoch": 0.3005012465093334, "grad_norm": 8.47866439819336, "learning_rate": 1.641334197925303e-05, "loss": 3.4406, "step": 8528 }, { "epoch": 0.30053648352229184, "grad_norm": 1.1372056007385254, "learning_rate": 1.641246627956001e-05, "loss": 1.0786, "step": 8529 }, { "epoch": 0.3005717205352502, "grad_norm": 5.723965167999268, "learning_rate": 1.6411590496343794e-05, "loss": 6.0041, "step": 8530 }, { "epoch": 0.30060695754820865, "grad_norm": 2.5745561122894287, "learning_rate": 1.64107146296158e-05, "loss": 1.0413, "step": 8531 }, { "epoch": 0.30064219456116703, "grad_norm": 6.739653587341309, "learning_rate": 1.640983867938743e-05, "loss": 5.9912, "step": 8532 }, { "epoch": 0.30067743157412546, "grad_norm": 5.66395902633667, "learning_rate": 1.6408962645670096e-05, "loss": 6.491, "step": 8533 }, { "epoch": 0.3007126685870839, "grad_norm": 1.2632535696029663, "learning_rate": 1.6408086528475204e-05, "loss": 0.9058, "step": 8534 }, { "epoch": 0.3007479056000423, "grad_norm": 2.4031097888946533, "learning_rate": 1.640721032781417e-05, "loss": 2.9903, "step": 8535 }, { "epoch": 0.3007831426130007, "grad_norm": 4.412071228027344, "learning_rate": 1.6406334043698405e-05, "loss": 3.7918, "step": 8536 }, { "epoch": 0.3008183796259591, "grad_norm": 2.345276117324829, "learning_rate": 1.6405457676139323e-05, "loss": 2.3701, "step": 8537 }, { "epoch": 0.3008536166389175, "grad_norm": 1.2346042394638062, "learning_rate": 1.640458122514834e-05, "loss": 0.8905, "step": 8538 }, { "epoch": 0.30088885365187595, "grad_norm": 3.735844850540161, "learning_rate": 1.640370469073687e-05, "loss": 3.5017, "step": 8539 }, { "epoch": 0.30092409066483433, "grad_norm": 1.0126949548721313, "learning_rate": 1.640282807291633e-05, "loss": 1.0945, "step": 8540 }, { "epoch": 0.30095932767779277, "grad_norm": 1.375035047531128, "learning_rate": 1.640195137169814e-05, "loss": 0.9646, "step": 8541 }, { "epoch": 0.30099456469075114, "grad_norm": 5.433242321014404, "learning_rate": 1.6401074587093718e-05, "loss": 5.0155, "step": 8542 }, { "epoch": 0.3010298017037096, "grad_norm": 2.0851001739501953, "learning_rate": 1.6400197719114487e-05, "loss": 0.8005, "step": 8543 }, { "epoch": 0.301065038716668, "grad_norm": 3.2837700843811035, "learning_rate": 1.639932076777186e-05, "loss": 3.72, "step": 8544 }, { "epoch": 0.3011002757296264, "grad_norm": 4.551022052764893, "learning_rate": 1.6398443733077268e-05, "loss": 5.1643, "step": 8545 }, { "epoch": 0.3011355127425848, "grad_norm": 12.109197616577148, "learning_rate": 1.639756661504213e-05, "loss": 2.9926, "step": 8546 }, { "epoch": 0.3011707497555432, "grad_norm": 5.694066047668457, "learning_rate": 1.6396689413677873e-05, "loss": 7.5943, "step": 8547 }, { "epoch": 0.30120598676850163, "grad_norm": 1.1679726839065552, "learning_rate": 1.639581212899592e-05, "loss": 1.1298, "step": 8548 }, { "epoch": 0.30124122378146007, "grad_norm": 4.125857830047607, "learning_rate": 1.63949347610077e-05, "loss": 3.1748, "step": 8549 }, { "epoch": 0.30127646079441844, "grad_norm": 4.685528755187988, "learning_rate": 1.639405730972464e-05, "loss": 5.4804, "step": 8550 }, { "epoch": 0.3013116978073769, "grad_norm": 0.9155448079109192, "learning_rate": 1.6393179775158173e-05, "loss": 0.9209, "step": 8551 }, { "epoch": 0.30134693482033525, "grad_norm": 6.091613292694092, "learning_rate": 1.639230215731972e-05, "loss": 3.5175, "step": 8552 }, { "epoch": 0.3013821718332937, "grad_norm": 4.685067653656006, "learning_rate": 1.639142445622072e-05, "loss": 3.085, "step": 8553 }, { "epoch": 0.3014174088462521, "grad_norm": 5.109794616699219, "learning_rate": 1.63905466718726e-05, "loss": 3.5171, "step": 8554 }, { "epoch": 0.3014526458592105, "grad_norm": 2.977297067642212, "learning_rate": 1.63896688042868e-05, "loss": 3.0248, "step": 8555 }, { "epoch": 0.30148788287216893, "grad_norm": 3.9328060150146484, "learning_rate": 1.6388790853474742e-05, "loss": 3.6388, "step": 8556 }, { "epoch": 0.3015231198851273, "grad_norm": 1.3542463779449463, "learning_rate": 1.6387912819447875e-05, "loss": 0.8577, "step": 8557 }, { "epoch": 0.30155835689808574, "grad_norm": 4.436497688293457, "learning_rate": 1.638703470221763e-05, "loss": 3.6317, "step": 8558 }, { "epoch": 0.3015935939110442, "grad_norm": 5.41873836517334, "learning_rate": 1.6386156501795442e-05, "loss": 3.0993, "step": 8559 }, { "epoch": 0.30162883092400256, "grad_norm": 3.0396108627319336, "learning_rate": 1.6385278218192753e-05, "loss": 3.3554, "step": 8560 }, { "epoch": 0.301664067936961, "grad_norm": 3.8894503116607666, "learning_rate": 1.6384399851421004e-05, "loss": 3.1585, "step": 8561 }, { "epoch": 0.3016993049499194, "grad_norm": 3.0947020053863525, "learning_rate": 1.6383521401491632e-05, "loss": 2.8573, "step": 8562 }, { "epoch": 0.3017345419628778, "grad_norm": 5.931554317474365, "learning_rate": 1.638264286841608e-05, "loss": 4.1824, "step": 8563 }, { "epoch": 0.30176977897583623, "grad_norm": 1.8500778675079346, "learning_rate": 1.638176425220579e-05, "loss": 0.9867, "step": 8564 }, { "epoch": 0.3018050159887946, "grad_norm": 1.5977989435195923, "learning_rate": 1.638088555287221e-05, "loss": 0.9705, "step": 8565 }, { "epoch": 0.30184025300175304, "grad_norm": 5.94376277923584, "learning_rate": 1.6380006770426785e-05, "loss": 8.7146, "step": 8566 }, { "epoch": 0.3018754900147115, "grad_norm": 5.0847859382629395, "learning_rate": 1.6379127904880955e-05, "loss": 5.7238, "step": 8567 }, { "epoch": 0.30191072702766986, "grad_norm": 3.3378260135650635, "learning_rate": 1.6378248956246173e-05, "loss": 3.5349, "step": 8568 }, { "epoch": 0.3019459640406283, "grad_norm": 6.379612445831299, "learning_rate": 1.637736992453389e-05, "loss": 3.5355, "step": 8569 }, { "epoch": 0.30198120105358667, "grad_norm": 5.1277899742126465, "learning_rate": 1.6376490809755546e-05, "loss": 3.274, "step": 8570 }, { "epoch": 0.3020164380665451, "grad_norm": 10.54701042175293, "learning_rate": 1.6375611611922594e-05, "loss": 8.5973, "step": 8571 }, { "epoch": 0.30205167507950353, "grad_norm": 2.9236996173858643, "learning_rate": 1.6374732331046495e-05, "loss": 3.096, "step": 8572 }, { "epoch": 0.3020869120924619, "grad_norm": 3.3456034660339355, "learning_rate": 1.6373852967138693e-05, "loss": 2.991, "step": 8573 }, { "epoch": 0.30212214910542035, "grad_norm": 1.462019443511963, "learning_rate": 1.6372973520210648e-05, "loss": 0.8741, "step": 8574 }, { "epoch": 0.3021573861183787, "grad_norm": 1.9092903137207031, "learning_rate": 1.6372093990273804e-05, "loss": 0.732, "step": 8575 }, { "epoch": 0.30219262313133716, "grad_norm": 1.4892302751541138, "learning_rate": 1.637121437733963e-05, "loss": 1.1001, "step": 8576 }, { "epoch": 0.3022278601442956, "grad_norm": 3.7053160667419434, "learning_rate": 1.6370334681419576e-05, "loss": 3.9781, "step": 8577 }, { "epoch": 0.30226309715725397, "grad_norm": 1.4523078203201294, "learning_rate": 1.6369454902525102e-05, "loss": 0.9079, "step": 8578 }, { "epoch": 0.3022983341702124, "grad_norm": 4.1876654624938965, "learning_rate": 1.6368575040667664e-05, "loss": 5.7121, "step": 8579 }, { "epoch": 0.3023335711831708, "grad_norm": 4.809593200683594, "learning_rate": 1.6367695095858724e-05, "loss": 3.6304, "step": 8580 }, { "epoch": 0.3023688081961292, "grad_norm": 2.870033025741577, "learning_rate": 1.6366815068109745e-05, "loss": 3.289, "step": 8581 }, { "epoch": 0.30240404520908765, "grad_norm": 2.1587986946105957, "learning_rate": 1.636593495743219e-05, "loss": 0.8954, "step": 8582 }, { "epoch": 0.302439282222046, "grad_norm": 1.6310510635375977, "learning_rate": 1.636505476383752e-05, "loss": 1.1068, "step": 8583 }, { "epoch": 0.30247451923500446, "grad_norm": 4.012964725494385, "learning_rate": 1.6364174487337205e-05, "loss": 3.2557, "step": 8584 }, { "epoch": 0.30250975624796284, "grad_norm": 3.579470634460449, "learning_rate": 1.6363294127942704e-05, "loss": 4.1314, "step": 8585 }, { "epoch": 0.30254499326092127, "grad_norm": 1.3172968626022339, "learning_rate": 1.6362413685665486e-05, "loss": 0.9576, "step": 8586 }, { "epoch": 0.3025802302738797, "grad_norm": 1.3894462585449219, "learning_rate": 1.6361533160517015e-05, "loss": 0.952, "step": 8587 }, { "epoch": 0.3026154672868381, "grad_norm": 1.4690572023391724, "learning_rate": 1.6360652552508773e-05, "loss": 1.0692, "step": 8588 }, { "epoch": 0.3026507042997965, "grad_norm": 3.0718634128570557, "learning_rate": 1.6359771861652217e-05, "loss": 3.7625, "step": 8589 }, { "epoch": 0.3026859413127549, "grad_norm": 2.6129424571990967, "learning_rate": 1.635889108795882e-05, "loss": 2.8055, "step": 8590 }, { "epoch": 0.3027211783257133, "grad_norm": 0.9072051048278809, "learning_rate": 1.635801023144006e-05, "loss": 0.7818, "step": 8591 }, { "epoch": 0.30275641533867176, "grad_norm": 1.3755062818527222, "learning_rate": 1.635712929210741e-05, "loss": 1.0386, "step": 8592 }, { "epoch": 0.30279165235163014, "grad_norm": 1.2886244058609009, "learning_rate": 1.6356248269972334e-05, "loss": 0.9533, "step": 8593 }, { "epoch": 0.30282688936458857, "grad_norm": 1.137860894203186, "learning_rate": 1.6355367165046317e-05, "loss": 0.9515, "step": 8594 }, { "epoch": 0.302862126377547, "grad_norm": 3.9456839561462402, "learning_rate": 1.6354485977340834e-05, "loss": 6.178, "step": 8595 }, { "epoch": 0.3028973633905054, "grad_norm": 1.1635944843292236, "learning_rate": 1.635360470686736e-05, "loss": 0.7529, "step": 8596 }, { "epoch": 0.3029326004034638, "grad_norm": 1.5414186716079712, "learning_rate": 1.635272335363738e-05, "loss": 0.9316, "step": 8597 }, { "epoch": 0.3029678374164222, "grad_norm": 6.321290016174316, "learning_rate": 1.6351841917662363e-05, "loss": 5.481, "step": 8598 }, { "epoch": 0.3030030744293806, "grad_norm": 1.234438180923462, "learning_rate": 1.63509603989538e-05, "loss": 0.8781, "step": 8599 }, { "epoch": 0.30303831144233906, "grad_norm": 3.5892603397369385, "learning_rate": 1.6350078797523168e-05, "loss": 5.9253, "step": 8600 }, { "epoch": 0.30307354845529744, "grad_norm": 4.173983573913574, "learning_rate": 1.634919711338195e-05, "loss": 3.8983, "step": 8601 }, { "epoch": 0.30310878546825587, "grad_norm": 3.3833224773406982, "learning_rate": 1.634831534654163e-05, "loss": 3.7507, "step": 8602 }, { "epoch": 0.30314402248121425, "grad_norm": 4.114522457122803, "learning_rate": 1.6347433497013698e-05, "loss": 3.1346, "step": 8603 }, { "epoch": 0.3031792594941727, "grad_norm": 3.627793550491333, "learning_rate": 1.634655156480963e-05, "loss": 3.5287, "step": 8604 }, { "epoch": 0.3032144965071311, "grad_norm": 1.4841948747634888, "learning_rate": 1.6345669549940922e-05, "loss": 1.0045, "step": 8605 }, { "epoch": 0.3032497335200895, "grad_norm": 5.373167514801025, "learning_rate": 1.634478745241906e-05, "loss": 4.4042, "step": 8606 }, { "epoch": 0.3032849705330479, "grad_norm": 1.4130395650863647, "learning_rate": 1.6343905272255537e-05, "loss": 1.2296, "step": 8607 }, { "epoch": 0.3033202075460063, "grad_norm": 3.305124282836914, "learning_rate": 1.6343023009461835e-05, "loss": 3.763, "step": 8608 }, { "epoch": 0.30335544455896474, "grad_norm": 5.070443630218506, "learning_rate": 1.6342140664049453e-05, "loss": 3.641, "step": 8609 }, { "epoch": 0.30339068157192317, "grad_norm": 4.327362060546875, "learning_rate": 1.6341258236029882e-05, "loss": 4.0299, "step": 8610 }, { "epoch": 0.30342591858488155, "grad_norm": 1.5799286365509033, "learning_rate": 1.6340375725414614e-05, "loss": 0.8988, "step": 8611 }, { "epoch": 0.30346115559784, "grad_norm": 1.3919028043746948, "learning_rate": 1.633949313221514e-05, "loss": 1.0426, "step": 8612 }, { "epoch": 0.30349639261079836, "grad_norm": 1.5253314971923828, "learning_rate": 1.6338610456442966e-05, "loss": 0.7174, "step": 8613 }, { "epoch": 0.3035316296237568, "grad_norm": 1.5855478048324585, "learning_rate": 1.6337727698109578e-05, "loss": 0.8561, "step": 8614 }, { "epoch": 0.3035668666367152, "grad_norm": 1.3192744255065918, "learning_rate": 1.6336844857226483e-05, "loss": 1.0245, "step": 8615 }, { "epoch": 0.3036021036496736, "grad_norm": 1.2825360298156738, "learning_rate": 1.633596193380518e-05, "loss": 1.0533, "step": 8616 }, { "epoch": 0.30363734066263204, "grad_norm": 1.0989079475402832, "learning_rate": 1.633507892785716e-05, "loss": 0.7414, "step": 8617 }, { "epoch": 0.3036725776755904, "grad_norm": 5.295858860015869, "learning_rate": 1.633419583939393e-05, "loss": 2.8545, "step": 8618 }, { "epoch": 0.30370781468854885, "grad_norm": 0.9670088291168213, "learning_rate": 1.6333312668426995e-05, "loss": 1.1164, "step": 8619 }, { "epoch": 0.3037430517015073, "grad_norm": 1.3216640949249268, "learning_rate": 1.6332429414967854e-05, "loss": 0.9446, "step": 8620 }, { "epoch": 0.30377828871446566, "grad_norm": 1.9900752305984497, "learning_rate": 1.6331546079028015e-05, "loss": 0.9327, "step": 8621 }, { "epoch": 0.3038135257274241, "grad_norm": 1.7102450132369995, "learning_rate": 1.633066266061898e-05, "loss": 0.9804, "step": 8622 }, { "epoch": 0.30384876274038247, "grad_norm": 2.616846799850464, "learning_rate": 1.632977915975226e-05, "loss": 0.7273, "step": 8623 }, { "epoch": 0.3038839997533409, "grad_norm": 1.8437089920043945, "learning_rate": 1.6328895576439357e-05, "loss": 1.0067, "step": 8624 }, { "epoch": 0.30391923676629934, "grad_norm": 1.6088463068008423, "learning_rate": 1.6328011910691787e-05, "loss": 1.0454, "step": 8625 }, { "epoch": 0.3039544737792577, "grad_norm": 3.571288585662842, "learning_rate": 1.632712816252105e-05, "loss": 3.3985, "step": 8626 }, { "epoch": 0.30398971079221615, "grad_norm": 1.2246147394180298, "learning_rate": 1.6326244331938666e-05, "loss": 0.7468, "step": 8627 }, { "epoch": 0.30402494780517453, "grad_norm": 8.30405044555664, "learning_rate": 1.6325360418956145e-05, "loss": 6.266, "step": 8628 }, { "epoch": 0.30406018481813296, "grad_norm": 3.2336161136627197, "learning_rate": 1.6324476423585e-05, "loss": 2.8925, "step": 8629 }, { "epoch": 0.3040954218310914, "grad_norm": 2.139507532119751, "learning_rate": 1.632359234583674e-05, "loss": 0.8754, "step": 8630 }, { "epoch": 0.3041306588440498, "grad_norm": 6.203794479370117, "learning_rate": 1.6322708185722884e-05, "loss": 6.1609, "step": 8631 }, { "epoch": 0.3041658958570082, "grad_norm": 5.427073001861572, "learning_rate": 1.6321823943254955e-05, "loss": 5.7422, "step": 8632 }, { "epoch": 0.30420113286996664, "grad_norm": 7.746228218078613, "learning_rate": 1.6320939618444455e-05, "loss": 7.4017, "step": 8633 }, { "epoch": 0.304236369882925, "grad_norm": 4.211467742919922, "learning_rate": 1.6320055211302916e-05, "loss": 2.8927, "step": 8634 }, { "epoch": 0.30427160689588345, "grad_norm": 2.2974002361297607, "learning_rate": 1.6319170721841856e-05, "loss": 3.2482, "step": 8635 }, { "epoch": 0.30430684390884183, "grad_norm": 3.62129807472229, "learning_rate": 1.6318286150072786e-05, "loss": 3.1005, "step": 8636 }, { "epoch": 0.30434208092180026, "grad_norm": 3.2964866161346436, "learning_rate": 1.6317401496007237e-05, "loss": 3.3916, "step": 8637 }, { "epoch": 0.3043773179347587, "grad_norm": 1.1315449476242065, "learning_rate": 1.6316516759656727e-05, "loss": 1.1417, "step": 8638 }, { "epoch": 0.3044125549477171, "grad_norm": 4.244359970092773, "learning_rate": 1.6315631941032783e-05, "loss": 5.8688, "step": 8639 }, { "epoch": 0.3044477919606755, "grad_norm": 7.817035675048828, "learning_rate": 1.631474704014693e-05, "loss": 6.7721, "step": 8640 }, { "epoch": 0.3044830289736339, "grad_norm": 2.870046615600586, "learning_rate": 1.631386205701069e-05, "loss": 3.5472, "step": 8641 }, { "epoch": 0.3045182659865923, "grad_norm": 1.9574439525604248, "learning_rate": 1.6312976991635595e-05, "loss": 1.1064, "step": 8642 }, { "epoch": 0.30455350299955075, "grad_norm": 1.3051366806030273, "learning_rate": 1.6312091844033168e-05, "loss": 0.8269, "step": 8643 }, { "epoch": 0.30458874001250913, "grad_norm": 1.2482558488845825, "learning_rate": 1.6311206614214944e-05, "loss": 0.7048, "step": 8644 }, { "epoch": 0.30462397702546756, "grad_norm": 1.3495374917984009, "learning_rate": 1.6310321302192447e-05, "loss": 1.1616, "step": 8645 }, { "epoch": 0.30465921403842594, "grad_norm": 4.990973472595215, "learning_rate": 1.630943590797721e-05, "loss": 5.4793, "step": 8646 }, { "epoch": 0.3046944510513844, "grad_norm": 3.458132028579712, "learning_rate": 1.6308550431580768e-05, "loss": 3.5364, "step": 8647 }, { "epoch": 0.3047296880643428, "grad_norm": 4.865766525268555, "learning_rate": 1.6307664873014654e-05, "loss": 2.7856, "step": 8648 }, { "epoch": 0.3047649250773012, "grad_norm": 1.225811243057251, "learning_rate": 1.63067792322904e-05, "loss": 1.132, "step": 8649 }, { "epoch": 0.3048001620902596, "grad_norm": 4.182303428649902, "learning_rate": 1.6305893509419547e-05, "loss": 3.6799, "step": 8650 }, { "epoch": 0.304835399103218, "grad_norm": 1.3676128387451172, "learning_rate": 1.6305007704413625e-05, "loss": 1.0397, "step": 8651 }, { "epoch": 0.30487063611617643, "grad_norm": 1.7486522197723389, "learning_rate": 1.6304121817284174e-05, "loss": 0.9247, "step": 8652 }, { "epoch": 0.30490587312913486, "grad_norm": 1.509837031364441, "learning_rate": 1.6303235848042735e-05, "loss": 0.91, "step": 8653 }, { "epoch": 0.30494111014209324, "grad_norm": 0.9316834807395935, "learning_rate": 1.6302349796700846e-05, "loss": 0.9732, "step": 8654 }, { "epoch": 0.3049763471550517, "grad_norm": 2.988192319869995, "learning_rate": 1.6301463663270046e-05, "loss": 3.3511, "step": 8655 }, { "epoch": 0.30501158416801005, "grad_norm": 3.229269504547119, "learning_rate": 1.6300577447761882e-05, "loss": 4.0422, "step": 8656 }, { "epoch": 0.3050468211809685, "grad_norm": 5.839419841766357, "learning_rate": 1.6299691150187894e-05, "loss": 8.1688, "step": 8657 }, { "epoch": 0.3050820581939269, "grad_norm": 1.413572072982788, "learning_rate": 1.629880477055962e-05, "loss": 0.9338, "step": 8658 }, { "epoch": 0.3051172952068853, "grad_norm": 1.372654676437378, "learning_rate": 1.6297918308888617e-05, "loss": 1.0587, "step": 8659 }, { "epoch": 0.30515253221984373, "grad_norm": 5.117833614349365, "learning_rate": 1.6297031765186425e-05, "loss": 3.1516, "step": 8660 }, { "epoch": 0.3051877692328021, "grad_norm": 1.3598406314849854, "learning_rate": 1.6296145139464593e-05, "loss": 1.0025, "step": 8661 }, { "epoch": 0.30522300624576054, "grad_norm": 2.7464780807495117, "learning_rate": 1.629525843173467e-05, "loss": 3.0533, "step": 8662 }, { "epoch": 0.305258243258719, "grad_norm": 3.7333455085754395, "learning_rate": 1.62943716420082e-05, "loss": 3.5022, "step": 8663 }, { "epoch": 0.30529348027167735, "grad_norm": 2.691901206970215, "learning_rate": 1.6293484770296738e-05, "loss": 3.5301, "step": 8664 }, { "epoch": 0.3053287172846358, "grad_norm": 9.184146881103516, "learning_rate": 1.6292597816611836e-05, "loss": 6.5334, "step": 8665 }, { "epoch": 0.3053639542975942, "grad_norm": 1.6398943662643433, "learning_rate": 1.6291710780965047e-05, "loss": 0.8827, "step": 8666 }, { "epoch": 0.3053991913105526, "grad_norm": 1.619065761566162, "learning_rate": 1.6290823663367926e-05, "loss": 0.9317, "step": 8667 }, { "epoch": 0.30543442832351103, "grad_norm": 4.770491600036621, "learning_rate": 1.628993646383202e-05, "loss": 2.7064, "step": 8668 }, { "epoch": 0.3054696653364694, "grad_norm": 4.202254772186279, "learning_rate": 1.6289049182368894e-05, "loss": 3.4929, "step": 8669 }, { "epoch": 0.30550490234942784, "grad_norm": 2.917797088623047, "learning_rate": 1.6288161818990103e-05, "loss": 3.1619, "step": 8670 }, { "epoch": 0.3055401393623863, "grad_norm": 4.196552753448486, "learning_rate": 1.6287274373707197e-05, "loss": 3.6556, "step": 8671 }, { "epoch": 0.30557537637534465, "grad_norm": 4.022311687469482, "learning_rate": 1.6286386846531747e-05, "loss": 5.781, "step": 8672 }, { "epoch": 0.3056106133883031, "grad_norm": 4.080284118652344, "learning_rate": 1.6285499237475305e-05, "loss": 5.9747, "step": 8673 }, { "epoch": 0.30564585040126147, "grad_norm": 3.7585010528564453, "learning_rate": 1.6284611546549432e-05, "loss": 3.1638, "step": 8674 }, { "epoch": 0.3056810874142199, "grad_norm": 5.8171706199646, "learning_rate": 1.62837237737657e-05, "loss": 6.5434, "step": 8675 }, { "epoch": 0.30571632442717833, "grad_norm": 1.239724040031433, "learning_rate": 1.628283591913566e-05, "loss": 1.0615, "step": 8676 }, { "epoch": 0.3057515614401367, "grad_norm": 2.076104164123535, "learning_rate": 1.6281947982670887e-05, "loss": 0.8993, "step": 8677 }, { "epoch": 0.30578679845309514, "grad_norm": 1.2986764907836914, "learning_rate": 1.628105996438294e-05, "loss": 1.0485, "step": 8678 }, { "epoch": 0.3058220354660535, "grad_norm": 1.3779276609420776, "learning_rate": 1.628017186428338e-05, "loss": 0.9161, "step": 8679 }, { "epoch": 0.30585727247901195, "grad_norm": 8.03968334197998, "learning_rate": 1.6279283682383788e-05, "loss": 3.5142, "step": 8680 }, { "epoch": 0.3058925094919704, "grad_norm": 22.138961791992188, "learning_rate": 1.6278395418695724e-05, "loss": 3.0682, "step": 8681 }, { "epoch": 0.30592774650492877, "grad_norm": 3.1950864791870117, "learning_rate": 1.627750707323076e-05, "loss": 2.884, "step": 8682 }, { "epoch": 0.3059629835178872, "grad_norm": 9.943095207214355, "learning_rate": 1.6276618646000467e-05, "loss": 6.0186, "step": 8683 }, { "epoch": 0.3059982205308456, "grad_norm": 12.498695373535156, "learning_rate": 1.6275730137016418e-05, "loss": 8.2541, "step": 8684 }, { "epoch": 0.306033457543804, "grad_norm": 3.7125394344329834, "learning_rate": 1.6274841546290183e-05, "loss": 3.0927, "step": 8685 }, { "epoch": 0.30606869455676244, "grad_norm": 1.6250630617141724, "learning_rate": 1.627395287383334e-05, "loss": 0.822, "step": 8686 }, { "epoch": 0.3061039315697208, "grad_norm": 3.9659531116485596, "learning_rate": 1.6273064119657458e-05, "loss": 3.2149, "step": 8687 }, { "epoch": 0.30613916858267926, "grad_norm": 1.2184174060821533, "learning_rate": 1.627217528377412e-05, "loss": 0.9658, "step": 8688 }, { "epoch": 0.30617440559563763, "grad_norm": 4.930677890777588, "learning_rate": 1.6271286366194897e-05, "loss": 5.7564, "step": 8689 }, { "epoch": 0.30620964260859607, "grad_norm": 5.0565690994262695, "learning_rate": 1.627039736693137e-05, "loss": 3.8247, "step": 8690 }, { "epoch": 0.3062448796215545, "grad_norm": 1.1869163513183594, "learning_rate": 1.6269508285995124e-05, "loss": 0.8838, "step": 8691 }, { "epoch": 0.3062801166345129, "grad_norm": 7.78814697265625, "learning_rate": 1.6268619123397732e-05, "loss": 2.6576, "step": 8692 }, { "epoch": 0.3063153536474713, "grad_norm": 5.21753454208374, "learning_rate": 1.6267729879150772e-05, "loss": 3.0713, "step": 8693 }, { "epoch": 0.3063505906604297, "grad_norm": 10.66580581665039, "learning_rate": 1.6266840553265835e-05, "loss": 8.8322, "step": 8694 }, { "epoch": 0.3063858276733881, "grad_norm": 1.962790608406067, "learning_rate": 1.6265951145754505e-05, "loss": 0.7076, "step": 8695 }, { "epoch": 0.30642106468634656, "grad_norm": 6.651484966278076, "learning_rate": 1.626506165662836e-05, "loss": 7.6065, "step": 8696 }, { "epoch": 0.30645630169930493, "grad_norm": 9.509292602539062, "learning_rate": 1.626417208589899e-05, "loss": 9.319, "step": 8697 }, { "epoch": 0.30649153871226337, "grad_norm": 1.6184556484222412, "learning_rate": 1.626328243357798e-05, "loss": 0.9714, "step": 8698 }, { "epoch": 0.3065267757252218, "grad_norm": 1.9036059379577637, "learning_rate": 1.626239269967692e-05, "loss": 0.7895, "step": 8699 }, { "epoch": 0.3065620127381802, "grad_norm": 4.130029678344727, "learning_rate": 1.6261502884207393e-05, "loss": 3.7034, "step": 8700 }, { "epoch": 0.3065972497511386, "grad_norm": 3.1971049308776855, "learning_rate": 1.6260612987180997e-05, "loss": 3.229, "step": 8701 }, { "epoch": 0.306632486764097, "grad_norm": 1.2755241394042969, "learning_rate": 1.625972300860932e-05, "loss": 1.0821, "step": 8702 }, { "epoch": 0.3066677237770554, "grad_norm": 1.5416339635849, "learning_rate": 1.625883294850395e-05, "loss": 1.1329, "step": 8703 }, { "epoch": 0.30670296079001386, "grad_norm": 3.9850964546203613, "learning_rate": 1.6257942806876485e-05, "loss": 3.3106, "step": 8704 }, { "epoch": 0.30673819780297223, "grad_norm": 1.1276928186416626, "learning_rate": 1.6257052583738518e-05, "loss": 1.2864, "step": 8705 }, { "epoch": 0.30677343481593067, "grad_norm": 6.517958641052246, "learning_rate": 1.6256162279101643e-05, "loss": 3.7276, "step": 8706 }, { "epoch": 0.30680867182888905, "grad_norm": 5.305054187774658, "learning_rate": 1.625527189297746e-05, "loss": 3.3012, "step": 8707 }, { "epoch": 0.3068439088418475, "grad_norm": 11.714140892028809, "learning_rate": 1.6254381425377564e-05, "loss": 5.7704, "step": 8708 }, { "epoch": 0.3068791458548059, "grad_norm": 6.124588966369629, "learning_rate": 1.625349087631355e-05, "loss": 4.3775, "step": 8709 }, { "epoch": 0.3069143828677643, "grad_norm": 1.082836389541626, "learning_rate": 1.6252600245797022e-05, "loss": 1.0466, "step": 8710 }, { "epoch": 0.3069496198807227, "grad_norm": 2.8612165451049805, "learning_rate": 1.6251709533839584e-05, "loss": 3.6157, "step": 8711 }, { "epoch": 0.3069848568936811, "grad_norm": 1.4342617988586426, "learning_rate": 1.6250818740452827e-05, "loss": 1.0709, "step": 8712 }, { "epoch": 0.30702009390663954, "grad_norm": 2.4479475021362305, "learning_rate": 1.6249927865648362e-05, "loss": 3.2401, "step": 8713 }, { "epoch": 0.30705533091959797, "grad_norm": 6.518467903137207, "learning_rate": 1.624903690943779e-05, "loss": 6.534, "step": 8714 }, { "epoch": 0.30709056793255635, "grad_norm": 4.060028553009033, "learning_rate": 1.6248145871832718e-05, "loss": 2.9647, "step": 8715 }, { "epoch": 0.3071258049455148, "grad_norm": 1.0172772407531738, "learning_rate": 1.6247254752844747e-05, "loss": 1.0298, "step": 8716 }, { "epoch": 0.30716104195847316, "grad_norm": 5.568291187286377, "learning_rate": 1.6246363552485486e-05, "loss": 5.0552, "step": 8717 }, { "epoch": 0.3071962789714316, "grad_norm": 3.5635364055633545, "learning_rate": 1.6245472270766545e-05, "loss": 4.0144, "step": 8718 }, { "epoch": 0.30723151598439, "grad_norm": 5.318909168243408, "learning_rate": 1.6244580907699536e-05, "loss": 3.393, "step": 8719 }, { "epoch": 0.3072667529973484, "grad_norm": 5.851989269256592, "learning_rate": 1.6243689463296063e-05, "loss": 3.2015, "step": 8720 }, { "epoch": 0.30730199001030684, "grad_norm": 5.088210105895996, "learning_rate": 1.624279793756774e-05, "loss": 3.5541, "step": 8721 }, { "epoch": 0.3073372270232652, "grad_norm": 6.338353633880615, "learning_rate": 1.624190633052618e-05, "loss": 5.8345, "step": 8722 }, { "epoch": 0.30737246403622365, "grad_norm": 18.031768798828125, "learning_rate": 1.624101464218299e-05, "loss": 8.8209, "step": 8723 }, { "epoch": 0.3074077010491821, "grad_norm": 6.258515357971191, "learning_rate": 1.6240122872549795e-05, "loss": 3.7774, "step": 8724 }, { "epoch": 0.30744293806214046, "grad_norm": 0.9345900416374207, "learning_rate": 1.62392310216382e-05, "loss": 1.1516, "step": 8725 }, { "epoch": 0.3074781750750989, "grad_norm": 3.3806447982788086, "learning_rate": 1.623833908945983e-05, "loss": 2.9026, "step": 8726 }, { "epoch": 0.30751341208805727, "grad_norm": 2.0360167026519775, "learning_rate": 1.6237447076026296e-05, "loss": 0.9489, "step": 8727 }, { "epoch": 0.3075486491010157, "grad_norm": 2.314427614212036, "learning_rate": 1.623655498134922e-05, "loss": 2.3719, "step": 8728 }, { "epoch": 0.30758388611397414, "grad_norm": 1.2753649950027466, "learning_rate": 1.623566280544022e-05, "loss": 1.1115, "step": 8729 }, { "epoch": 0.3076191231269325, "grad_norm": 2.2858223915100098, "learning_rate": 1.6234770548310922e-05, "loss": 1.1883, "step": 8730 }, { "epoch": 0.30765436013989095, "grad_norm": 5.742280960083008, "learning_rate": 1.623387820997294e-05, "loss": 6.1125, "step": 8731 }, { "epoch": 0.3076895971528493, "grad_norm": 1.2521135807037354, "learning_rate": 1.62329857904379e-05, "loss": 0.9958, "step": 8732 }, { "epoch": 0.30772483416580776, "grad_norm": 4.110125541687012, "learning_rate": 1.6232093289717426e-05, "loss": 3.6337, "step": 8733 }, { "epoch": 0.3077600711787662, "grad_norm": 5.082820892333984, "learning_rate": 1.623120070782314e-05, "loss": 5.7284, "step": 8734 }, { "epoch": 0.30779530819172457, "grad_norm": 3.4860072135925293, "learning_rate": 1.623030804476668e-05, "loss": 3.0773, "step": 8735 }, { "epoch": 0.307830545204683, "grad_norm": 1.2977794408798218, "learning_rate": 1.6229415300559655e-05, "loss": 1.0938, "step": 8736 }, { "epoch": 0.30786578221764144, "grad_norm": 3.9609274864196777, "learning_rate": 1.622852247521371e-05, "loss": 5.5602, "step": 8737 }, { "epoch": 0.3079010192305998, "grad_norm": 6.6080193519592285, "learning_rate": 1.6227629568740462e-05, "loss": 6.5479, "step": 8738 }, { "epoch": 0.30793625624355825, "grad_norm": 5.073232650756836, "learning_rate": 1.6226736581151545e-05, "loss": 3.8746, "step": 8739 }, { "epoch": 0.3079714932565166, "grad_norm": 1.3106645345687866, "learning_rate": 1.6225843512458593e-05, "loss": 0.8941, "step": 8740 }, { "epoch": 0.30800673026947506, "grad_norm": 7.100404262542725, "learning_rate": 1.6224950362673234e-05, "loss": 3.0758, "step": 8741 }, { "epoch": 0.3080419672824335, "grad_norm": 5.325438976287842, "learning_rate": 1.6224057131807104e-05, "loss": 3.1735, "step": 8742 }, { "epoch": 0.30807720429539187, "grad_norm": 1.5803645849227905, "learning_rate": 1.622316381987184e-05, "loss": 1.0387, "step": 8743 }, { "epoch": 0.3081124413083503, "grad_norm": 4.735512733459473, "learning_rate": 1.622227042687907e-05, "loss": 3.9993, "step": 8744 }, { "epoch": 0.3081476783213087, "grad_norm": 1.4320087432861328, "learning_rate": 1.622137695284044e-05, "loss": 0.9135, "step": 8745 }, { "epoch": 0.3081829153342671, "grad_norm": 1.333829641342163, "learning_rate": 1.622048339776758e-05, "loss": 0.8745, "step": 8746 }, { "epoch": 0.30821815234722555, "grad_norm": 2.1078526973724365, "learning_rate": 1.6219589761672128e-05, "loss": 1.0065, "step": 8747 }, { "epoch": 0.3082533893601839, "grad_norm": 2.66646671295166, "learning_rate": 1.621869604456573e-05, "loss": 3.7744, "step": 8748 }, { "epoch": 0.30828862637314236, "grad_norm": 1.4851999282836914, "learning_rate": 1.6217802246460026e-05, "loss": 1.0035, "step": 8749 }, { "epoch": 0.30832386338610074, "grad_norm": 1.3740739822387695, "learning_rate": 1.6216908367366655e-05, "loss": 0.8335, "step": 8750 }, { "epoch": 0.30835910039905917, "grad_norm": 4.084518909454346, "learning_rate": 1.6216014407297257e-05, "loss": 4.1757, "step": 8751 }, { "epoch": 0.3083943374120176, "grad_norm": 1.7662142515182495, "learning_rate": 1.6215120366263485e-05, "loss": 1.0074, "step": 8752 }, { "epoch": 0.308429574424976, "grad_norm": 2.7331981658935547, "learning_rate": 1.6214226244276976e-05, "loss": 3.8329, "step": 8753 }, { "epoch": 0.3084648114379344, "grad_norm": 4.837780475616455, "learning_rate": 1.6213332041349377e-05, "loss": 5.8156, "step": 8754 }, { "epoch": 0.3085000484508928, "grad_norm": 2.3249361515045166, "learning_rate": 1.6212437757492338e-05, "loss": 1.2166, "step": 8755 }, { "epoch": 0.3085352854638512, "grad_norm": 4.7248358726501465, "learning_rate": 1.621154339271751e-05, "loss": 5.8625, "step": 8756 }, { "epoch": 0.30857052247680966, "grad_norm": 1.2554384469985962, "learning_rate": 1.6210648947036532e-05, "loss": 0.9624, "step": 8757 }, { "epoch": 0.30860575948976804, "grad_norm": 1.4563555717468262, "learning_rate": 1.6209754420461063e-05, "loss": 1.2366, "step": 8758 }, { "epoch": 0.3086409965027265, "grad_norm": 2.993346691131592, "learning_rate": 1.620885981300275e-05, "loss": 3.8681, "step": 8759 }, { "epoch": 0.30867623351568485, "grad_norm": 8.736331939697266, "learning_rate": 1.6207965124673247e-05, "loss": 3.5993, "step": 8760 }, { "epoch": 0.3087114705286433, "grad_norm": 7.469261169433594, "learning_rate": 1.620707035548421e-05, "loss": 6.0186, "step": 8761 }, { "epoch": 0.3087467075416017, "grad_norm": 4.954861164093018, "learning_rate": 1.620617550544729e-05, "loss": 3.4004, "step": 8762 }, { "epoch": 0.3087819445545601, "grad_norm": 2.615959405899048, "learning_rate": 1.620528057457414e-05, "loss": 2.8155, "step": 8763 }, { "epoch": 0.30881718156751853, "grad_norm": 7.845863342285156, "learning_rate": 1.6204385562876425e-05, "loss": 9.5612, "step": 8764 }, { "epoch": 0.3088524185804769, "grad_norm": 2.2307136058807373, "learning_rate": 1.6203490470365796e-05, "loss": 3.3502, "step": 8765 }, { "epoch": 0.30888765559343534, "grad_norm": 1.6775758266448975, "learning_rate": 1.6202595297053915e-05, "loss": 1.2424, "step": 8766 }, { "epoch": 0.3089228926063938, "grad_norm": 1.4458867311477661, "learning_rate": 1.620170004295244e-05, "loss": 0.9678, "step": 8767 }, { "epoch": 0.30895812961935215, "grad_norm": 1.2291967868804932, "learning_rate": 1.620080470807303e-05, "loss": 0.7801, "step": 8768 }, { "epoch": 0.3089933666323106, "grad_norm": 3.7502269744873047, "learning_rate": 1.619990929242735e-05, "loss": 6.4122, "step": 8769 }, { "epoch": 0.309028603645269, "grad_norm": 3.471994400024414, "learning_rate": 1.619901379602706e-05, "loss": 3.3396, "step": 8770 }, { "epoch": 0.3090638406582274, "grad_norm": 1.101001501083374, "learning_rate": 1.6198118218883826e-05, "loss": 1.2206, "step": 8771 }, { "epoch": 0.30909907767118583, "grad_norm": 1.0937585830688477, "learning_rate": 1.6197222561009313e-05, "loss": 1.1378, "step": 8772 }, { "epoch": 0.3091343146841442, "grad_norm": 4.863072872161865, "learning_rate": 1.619632682241519e-05, "loss": 5.8219, "step": 8773 }, { "epoch": 0.30916955169710264, "grad_norm": 5.24934720993042, "learning_rate": 1.6195431003113116e-05, "loss": 3.4208, "step": 8774 }, { "epoch": 0.3092047887100611, "grad_norm": 1.6504807472229004, "learning_rate": 1.6194535103114767e-05, "loss": 0.7706, "step": 8775 }, { "epoch": 0.30924002572301945, "grad_norm": 1.1774245500564575, "learning_rate": 1.619363912243181e-05, "loss": 0.7861, "step": 8776 }, { "epoch": 0.3092752627359779, "grad_norm": 2.655730962753296, "learning_rate": 1.619274306107591e-05, "loss": 2.897, "step": 8777 }, { "epoch": 0.30931049974893626, "grad_norm": 1.45749831199646, "learning_rate": 1.6191846919058746e-05, "loss": 0.9859, "step": 8778 }, { "epoch": 0.3093457367618947, "grad_norm": 1.3259015083312988, "learning_rate": 1.6190950696391988e-05, "loss": 1.078, "step": 8779 }, { "epoch": 0.30938097377485313, "grad_norm": 1.9021074771881104, "learning_rate": 1.6190054393087306e-05, "loss": 0.6999, "step": 8780 }, { "epoch": 0.3094162107878115, "grad_norm": 0.9941698908805847, "learning_rate": 1.6189158009156378e-05, "loss": 1.0595, "step": 8781 }, { "epoch": 0.30945144780076994, "grad_norm": 3.3090333938598633, "learning_rate": 1.618826154461088e-05, "loss": 2.9852, "step": 8782 }, { "epoch": 0.3094866848137283, "grad_norm": 1.0867438316345215, "learning_rate": 1.6187364999462488e-05, "loss": 1.1564, "step": 8783 }, { "epoch": 0.30952192182668675, "grad_norm": 2.8311047554016113, "learning_rate": 1.6186468373722877e-05, "loss": 3.7368, "step": 8784 }, { "epoch": 0.3095571588396452, "grad_norm": 1.841134786605835, "learning_rate": 1.6185571667403724e-05, "loss": 0.892, "step": 8785 }, { "epoch": 0.30959239585260356, "grad_norm": 1.0294195413589478, "learning_rate": 1.6184674880516715e-05, "loss": 1.0307, "step": 8786 }, { "epoch": 0.309627632865562, "grad_norm": 4.5541181564331055, "learning_rate": 1.618377801307353e-05, "loss": 2.7016, "step": 8787 }, { "epoch": 0.3096628698785204, "grad_norm": 1.309768795967102, "learning_rate": 1.6182881065085844e-05, "loss": 1.0839, "step": 8788 }, { "epoch": 0.3096981068914788, "grad_norm": 2.265178680419922, "learning_rate": 1.618198403656535e-05, "loss": 1.16, "step": 8789 }, { "epoch": 0.30973334390443724, "grad_norm": 1.4056785106658936, "learning_rate": 1.6181086927523725e-05, "loss": 1.1032, "step": 8790 }, { "epoch": 0.3097685809173956, "grad_norm": 1.9829522371292114, "learning_rate": 1.6180189737972652e-05, "loss": 1.0702, "step": 8791 }, { "epoch": 0.30980381793035405, "grad_norm": 1.0493462085723877, "learning_rate": 1.6179292467923823e-05, "loss": 0.9824, "step": 8792 }, { "epoch": 0.30983905494331243, "grad_norm": 3.0479354858398438, "learning_rate": 1.6178395117388924e-05, "loss": 3.034, "step": 8793 }, { "epoch": 0.30987429195627086, "grad_norm": 1.160312533378601, "learning_rate": 1.617749768637964e-05, "loss": 0.8141, "step": 8794 }, { "epoch": 0.3099095289692293, "grad_norm": 1.2789019346237183, "learning_rate": 1.617660017490766e-05, "loss": 1.0611, "step": 8795 }, { "epoch": 0.3099447659821877, "grad_norm": 2.247101068496704, "learning_rate": 1.6175702582984677e-05, "loss": 0.8415, "step": 8796 }, { "epoch": 0.3099800029951461, "grad_norm": 1.5303595066070557, "learning_rate": 1.6174804910622385e-05, "loss": 0.9633, "step": 8797 }, { "epoch": 0.3100152400081045, "grad_norm": 1.671656847000122, "learning_rate": 1.6173907157832467e-05, "loss": 0.8874, "step": 8798 }, { "epoch": 0.3100504770210629, "grad_norm": 7.552495002746582, "learning_rate": 1.6173009324626624e-05, "loss": 3.6006, "step": 8799 }, { "epoch": 0.31008571403402135, "grad_norm": 3.0293734073638916, "learning_rate": 1.617211141101655e-05, "loss": 3.6766, "step": 8800 }, { "epoch": 0.31012095104697973, "grad_norm": 1.13909912109375, "learning_rate": 1.6171213417013935e-05, "loss": 1.33, "step": 8801 }, { "epoch": 0.31015618805993816, "grad_norm": 10.536359786987305, "learning_rate": 1.6170315342630483e-05, "loss": 5.3056, "step": 8802 }, { "epoch": 0.3101914250728966, "grad_norm": 4.9260687828063965, "learning_rate": 1.6169417187877886e-05, "loss": 5.4553, "step": 8803 }, { "epoch": 0.310226662085855, "grad_norm": 1.183310866355896, "learning_rate": 1.6168518952767845e-05, "loss": 1.2944, "step": 8804 }, { "epoch": 0.3102618990988134, "grad_norm": 3.777064561843872, "learning_rate": 1.6167620637312057e-05, "loss": 5.5833, "step": 8805 }, { "epoch": 0.3102971361117718, "grad_norm": 1.1556673049926758, "learning_rate": 1.6166722241522225e-05, "loss": 1.129, "step": 8806 }, { "epoch": 0.3103323731247302, "grad_norm": 4.695891380310059, "learning_rate": 1.616582376541005e-05, "loss": 3.5232, "step": 8807 }, { "epoch": 0.31036761013768865, "grad_norm": 1.670235276222229, "learning_rate": 1.6164925208987236e-05, "loss": 1.1169, "step": 8808 }, { "epoch": 0.31040284715064703, "grad_norm": 1.4795268774032593, "learning_rate": 1.6164026572265486e-05, "loss": 1.1068, "step": 8809 }, { "epoch": 0.31043808416360547, "grad_norm": 1.0815762281417847, "learning_rate": 1.6163127855256504e-05, "loss": 1.0726, "step": 8810 }, { "epoch": 0.31047332117656384, "grad_norm": 1.117316722869873, "learning_rate": 1.6162229057971997e-05, "loss": 1.0735, "step": 8811 }, { "epoch": 0.3105085581895223, "grad_norm": 2.132554531097412, "learning_rate": 1.6161330180423672e-05, "loss": 1.04, "step": 8812 }, { "epoch": 0.3105437952024807, "grad_norm": 8.769434928894043, "learning_rate": 1.6160431222623237e-05, "loss": 8.3018, "step": 8813 }, { "epoch": 0.3105790322154391, "grad_norm": 1.1845753192901611, "learning_rate": 1.61595321845824e-05, "loss": 0.8316, "step": 8814 }, { "epoch": 0.3106142692283975, "grad_norm": 0.9056543111801147, "learning_rate": 1.6158633066312873e-05, "loss": 1.2391, "step": 8815 }, { "epoch": 0.3106495062413559, "grad_norm": 2.8451757431030273, "learning_rate": 1.6157733867826362e-05, "loss": 3.3824, "step": 8816 }, { "epoch": 0.31068474325431433, "grad_norm": 1.3087983131408691, "learning_rate": 1.6156834589134587e-05, "loss": 0.8745, "step": 8817 }, { "epoch": 0.31071998026727277, "grad_norm": 1.448423147201538, "learning_rate": 1.6155935230249252e-05, "loss": 0.9555, "step": 8818 }, { "epoch": 0.31075521728023114, "grad_norm": 2.612502098083496, "learning_rate": 1.6155035791182084e-05, "loss": 3.9645, "step": 8819 }, { "epoch": 0.3107904542931896, "grad_norm": 1.3974082469940186, "learning_rate": 1.6154136271944783e-05, "loss": 0.989, "step": 8820 }, { "epoch": 0.31082569130614796, "grad_norm": 1.2065929174423218, "learning_rate": 1.615323667254908e-05, "loss": 1.0193, "step": 8821 }, { "epoch": 0.3108609283191064, "grad_norm": 3.337000608444214, "learning_rate": 1.615233699300668e-05, "loss": 3.5355, "step": 8822 }, { "epoch": 0.3108961653320648, "grad_norm": 2.6358444690704346, "learning_rate": 1.6151437233329314e-05, "loss": 3.6934, "step": 8823 }, { "epoch": 0.3109314023450232, "grad_norm": 5.10784912109375, "learning_rate": 1.615053739352869e-05, "loss": 5.8532, "step": 8824 }, { "epoch": 0.31096663935798163, "grad_norm": 1.2508363723754883, "learning_rate": 1.614963747361653e-05, "loss": 1.2009, "step": 8825 }, { "epoch": 0.31100187637094, "grad_norm": 1.1963533163070679, "learning_rate": 1.6148737473604562e-05, "loss": 1.0939, "step": 8826 }, { "epoch": 0.31103711338389844, "grad_norm": 1.6231403350830078, "learning_rate": 1.6147837393504503e-05, "loss": 0.9445, "step": 8827 }, { "epoch": 0.3110723503968569, "grad_norm": 4.127952575683594, "learning_rate": 1.6146937233328084e-05, "loss": 3.5518, "step": 8828 }, { "epoch": 0.31110758740981526, "grad_norm": 1.2565919160842896, "learning_rate": 1.614603699308702e-05, "loss": 0.8885, "step": 8829 }, { "epoch": 0.3111428244227737, "grad_norm": 1.1959997415542603, "learning_rate": 1.6145136672793046e-05, "loss": 0.8168, "step": 8830 }, { "epoch": 0.31117806143573207, "grad_norm": 3.923530340194702, "learning_rate": 1.6144236272457877e-05, "loss": 3.7268, "step": 8831 }, { "epoch": 0.3112132984486905, "grad_norm": 3.668727159500122, "learning_rate": 1.6143335792093253e-05, "loss": 3.3525, "step": 8832 }, { "epoch": 0.31124853546164893, "grad_norm": 4.9352707862854, "learning_rate": 1.6142435231710893e-05, "loss": 3.9759, "step": 8833 }, { "epoch": 0.3112837724746073, "grad_norm": 3.9300308227539062, "learning_rate": 1.6141534591322538e-05, "loss": 2.9794, "step": 8834 }, { "epoch": 0.31131900948756575, "grad_norm": 2.687246799468994, "learning_rate": 1.614063387093991e-05, "loss": 3.6847, "step": 8835 }, { "epoch": 0.3113542465005241, "grad_norm": 4.527870178222656, "learning_rate": 1.6139733070574742e-05, "loss": 6.0421, "step": 8836 }, { "epoch": 0.31138948351348256, "grad_norm": 5.362805366516113, "learning_rate": 1.6138832190238768e-05, "loss": 2.9497, "step": 8837 }, { "epoch": 0.311424720526441, "grad_norm": 4.35329532623291, "learning_rate": 1.6137931229943727e-05, "loss": 3.391, "step": 8838 }, { "epoch": 0.31145995753939937, "grad_norm": 8.175537109375, "learning_rate": 1.6137030189701347e-05, "loss": 6.6396, "step": 8839 }, { "epoch": 0.3114951945523578, "grad_norm": 3.9079430103302, "learning_rate": 1.613612906952337e-05, "loss": 3.4009, "step": 8840 }, { "epoch": 0.31153043156531623, "grad_norm": 6.53200626373291, "learning_rate": 1.6135227869421524e-05, "loss": 2.8399, "step": 8841 }, { "epoch": 0.3115656685782746, "grad_norm": 5.330334186553955, "learning_rate": 1.6134326589407558e-05, "loss": 0.964, "step": 8842 }, { "epoch": 0.31160090559123305, "grad_norm": 1.4593051671981812, "learning_rate": 1.6133425229493208e-05, "loss": 0.7511, "step": 8843 }, { "epoch": 0.3116361426041914, "grad_norm": 9.039361953735352, "learning_rate": 1.6132523789690207e-05, "loss": 6.4503, "step": 8844 }, { "epoch": 0.31167137961714986, "grad_norm": 1.1027348041534424, "learning_rate": 1.613162227001031e-05, "loss": 1.0859, "step": 8845 }, { "epoch": 0.3117066166301083, "grad_norm": 1.9401929378509521, "learning_rate": 1.613072067046525e-05, "loss": 1.0907, "step": 8846 }, { "epoch": 0.31174185364306667, "grad_norm": 10.18757152557373, "learning_rate": 1.612981899106677e-05, "loss": 5.0561, "step": 8847 }, { "epoch": 0.3117770906560251, "grad_norm": 5.342957019805908, "learning_rate": 1.6128917231826616e-05, "loss": 8.5647, "step": 8848 }, { "epoch": 0.3118123276689835, "grad_norm": 1.431290864944458, "learning_rate": 1.6128015392756536e-05, "loss": 1.1409, "step": 8849 }, { "epoch": 0.3118475646819419, "grad_norm": 1.3640998601913452, "learning_rate": 1.6127113473868275e-05, "loss": 0.8526, "step": 8850 }, { "epoch": 0.31188280169490035, "grad_norm": 1.910997986793518, "learning_rate": 1.6126211475173583e-05, "loss": 1.0099, "step": 8851 }, { "epoch": 0.3119180387078587, "grad_norm": 1.331106185913086, "learning_rate": 1.6125309396684202e-05, "loss": 0.897, "step": 8852 }, { "epoch": 0.31195327572081716, "grad_norm": 1.5959453582763672, "learning_rate": 1.6124407238411887e-05, "loss": 1.02, "step": 8853 }, { "epoch": 0.31198851273377554, "grad_norm": 5.410167217254639, "learning_rate": 1.6123505000368388e-05, "loss": 3.8189, "step": 8854 }, { "epoch": 0.31202374974673397, "grad_norm": 3.6371471881866455, "learning_rate": 1.6122602682565457e-05, "loss": 3.4609, "step": 8855 }, { "epoch": 0.3120589867596924, "grad_norm": 1.0084604024887085, "learning_rate": 1.6121700285014845e-05, "loss": 0.8413, "step": 8856 }, { "epoch": 0.3120942237726508, "grad_norm": 5.16093111038208, "learning_rate": 1.612079780772831e-05, "loss": 3.153, "step": 8857 }, { "epoch": 0.3121294607856092, "grad_norm": 19.217222213745117, "learning_rate": 1.61198952507176e-05, "loss": 7.257, "step": 8858 }, { "epoch": 0.3121646977985676, "grad_norm": 1.1652685403823853, "learning_rate": 1.611899261399448e-05, "loss": 1.2917, "step": 8859 }, { "epoch": 0.312199934811526, "grad_norm": 9.450915336608887, "learning_rate": 1.611808989757069e-05, "loss": 6.3164, "step": 8860 }, { "epoch": 0.31223517182448446, "grad_norm": 1.5533068180084229, "learning_rate": 1.6117187101458013e-05, "loss": 0.9609, "step": 8861 }, { "epoch": 0.31227040883744284, "grad_norm": 4.031784534454346, "learning_rate": 1.6116284225668188e-05, "loss": 7.6261, "step": 8862 }, { "epoch": 0.31230564585040127, "grad_norm": 3.6365463733673096, "learning_rate": 1.6115381270212983e-05, "loss": 6.8122, "step": 8863 }, { "epoch": 0.31234088286335965, "grad_norm": 6.251220703125, "learning_rate": 1.6114478235104162e-05, "loss": 3.5399, "step": 8864 }, { "epoch": 0.3123761198763181, "grad_norm": 1.4285948276519775, "learning_rate": 1.611357512035348e-05, "loss": 0.8281, "step": 8865 }, { "epoch": 0.3124113568892765, "grad_norm": 1.1912903785705566, "learning_rate": 1.61126719259727e-05, "loss": 0.9708, "step": 8866 }, { "epoch": 0.3124465939022349, "grad_norm": 3.303600549697876, "learning_rate": 1.6111768651973595e-05, "loss": 4.1278, "step": 8867 }, { "epoch": 0.3124818309151933, "grad_norm": 1.3036845922470093, "learning_rate": 1.611086529836792e-05, "loss": 1.2438, "step": 8868 }, { "epoch": 0.3125170679281517, "grad_norm": 4.860673904418945, "learning_rate": 1.6109961865167448e-05, "loss": 5.7436, "step": 8869 }, { "epoch": 0.31255230494111014, "grad_norm": 1.6513608694076538, "learning_rate": 1.6109058352383948e-05, "loss": 1.0649, "step": 8870 }, { "epoch": 0.31258754195406857, "grad_norm": 3.796414613723755, "learning_rate": 1.610815476002918e-05, "loss": 2.8533, "step": 8871 }, { "epoch": 0.31262277896702695, "grad_norm": 3.950162887573242, "learning_rate": 1.610725108811492e-05, "loss": 3.7409, "step": 8872 }, { "epoch": 0.3126580159799854, "grad_norm": 1.6366822719573975, "learning_rate": 1.6106347336652937e-05, "loss": 0.9797, "step": 8873 }, { "epoch": 0.3126932529929438, "grad_norm": 1.9988014698028564, "learning_rate": 1.6105443505655e-05, "loss": 0.8952, "step": 8874 }, { "epoch": 0.3127284900059022, "grad_norm": 4.437967777252197, "learning_rate": 1.6104539595132885e-05, "loss": 5.9487, "step": 8875 }, { "epoch": 0.3127637270188606, "grad_norm": 4.89008092880249, "learning_rate": 1.6103635605098367e-05, "loss": 2.8411, "step": 8876 }, { "epoch": 0.312798964031819, "grad_norm": 0.98976731300354, "learning_rate": 1.6102731535563214e-05, "loss": 0.93, "step": 8877 }, { "epoch": 0.31283420104477744, "grad_norm": 1.6733083724975586, "learning_rate": 1.6101827386539206e-05, "loss": 1.0106, "step": 8878 }, { "epoch": 0.31286943805773587, "grad_norm": 1.264260172843933, "learning_rate": 1.610092315803812e-05, "loss": 0.7735, "step": 8879 }, { "epoch": 0.31290467507069425, "grad_norm": 3.2931511402130127, "learning_rate": 1.6100018850071733e-05, "loss": 3.4512, "step": 8880 }, { "epoch": 0.3129399120836527, "grad_norm": 5.45000696182251, "learning_rate": 1.6099114462651823e-05, "loss": 4.166, "step": 8881 }, { "epoch": 0.31297514909661106, "grad_norm": 4.8386688232421875, "learning_rate": 1.6098209995790167e-05, "loss": 4.0334, "step": 8882 }, { "epoch": 0.3130103861095695, "grad_norm": 1.0980017185211182, "learning_rate": 1.6097305449498555e-05, "loss": 0.7765, "step": 8883 }, { "epoch": 0.3130456231225279, "grad_norm": 4.352273464202881, "learning_rate": 1.609640082378876e-05, "loss": 4.1035, "step": 8884 }, { "epoch": 0.3130808601354863, "grad_norm": 4.772514820098877, "learning_rate": 1.6095496118672566e-05, "loss": 3.4193, "step": 8885 }, { "epoch": 0.31311609714844474, "grad_norm": 3.9161062240600586, "learning_rate": 1.609459133416176e-05, "loss": 3.9905, "step": 8886 }, { "epoch": 0.3131513341614031, "grad_norm": 6.5863823890686035, "learning_rate": 1.6093686470268125e-05, "loss": 5.6984, "step": 8887 }, { "epoch": 0.31318657117436155, "grad_norm": 5.39738655090332, "learning_rate": 1.609278152700345e-05, "loss": 5.1355, "step": 8888 }, { "epoch": 0.31322180818732, "grad_norm": 1.380346655845642, "learning_rate": 1.609187650437952e-05, "loss": 0.9805, "step": 8889 }, { "epoch": 0.31325704520027836, "grad_norm": 3.940011978149414, "learning_rate": 1.6090971402408117e-05, "loss": 5.8651, "step": 8890 }, { "epoch": 0.3132922822132368, "grad_norm": 4.295173645019531, "learning_rate": 1.6090066221101038e-05, "loss": 3.7694, "step": 8891 }, { "epoch": 0.3133275192261952, "grad_norm": 4.360815525054932, "learning_rate": 1.6089160960470075e-05, "loss": 4.1304, "step": 8892 }, { "epoch": 0.3133627562391536, "grad_norm": 1.3244216442108154, "learning_rate": 1.608825562052701e-05, "loss": 0.9562, "step": 8893 }, { "epoch": 0.31339799325211204, "grad_norm": 3.856908082962036, "learning_rate": 1.6087350201283642e-05, "loss": 3.6476, "step": 8894 }, { "epoch": 0.3134332302650704, "grad_norm": 6.491287708282471, "learning_rate": 1.608644470275176e-05, "loss": 6.2893, "step": 8895 }, { "epoch": 0.31346846727802885, "grad_norm": 1.0593098402023315, "learning_rate": 1.6085539124943164e-05, "loss": 1.1637, "step": 8896 }, { "epoch": 0.31350370429098723, "grad_norm": 1.768480896949768, "learning_rate": 1.6084633467869645e-05, "loss": 0.8781, "step": 8897 }, { "epoch": 0.31353894130394566, "grad_norm": 3.811349630355835, "learning_rate": 1.6083727731542994e-05, "loss": 5.6266, "step": 8898 }, { "epoch": 0.3135741783169041, "grad_norm": 1.420309066772461, "learning_rate": 1.608282191597502e-05, "loss": 0.7965, "step": 8899 }, { "epoch": 0.3136094153298625, "grad_norm": 4.428309440612793, "learning_rate": 1.608191602117752e-05, "loss": 3.8159, "step": 8900 }, { "epoch": 0.3136446523428209, "grad_norm": 5.042990684509277, "learning_rate": 1.608101004716228e-05, "loss": 3.4306, "step": 8901 }, { "epoch": 0.3136798893557793, "grad_norm": 4.655636787414551, "learning_rate": 1.6080103993941115e-05, "loss": 3.478, "step": 8902 }, { "epoch": 0.3137151263687377, "grad_norm": 3.2225921154022217, "learning_rate": 1.6079197861525822e-05, "loss": 3.2583, "step": 8903 }, { "epoch": 0.31375036338169615, "grad_norm": 1.8681079149246216, "learning_rate": 1.6078291649928198e-05, "loss": 0.7632, "step": 8904 }, { "epoch": 0.31378560039465453, "grad_norm": 1.7442320585250854, "learning_rate": 1.6077385359160054e-05, "loss": 0.9274, "step": 8905 }, { "epoch": 0.31382083740761296, "grad_norm": 3.7791008949279785, "learning_rate": 1.607647898923319e-05, "loss": 3.3779, "step": 8906 }, { "epoch": 0.31385607442057134, "grad_norm": 1.6035882234573364, "learning_rate": 1.607557254015941e-05, "loss": 1.0848, "step": 8907 }, { "epoch": 0.3138913114335298, "grad_norm": 0.9689440131187439, "learning_rate": 1.607466601195053e-05, "loss": 0.7703, "step": 8908 }, { "epoch": 0.3139265484464882, "grad_norm": 5.547242164611816, "learning_rate": 1.607375940461835e-05, "loss": 4.9054, "step": 8909 }, { "epoch": 0.3139617854594466, "grad_norm": 1.6035552024841309, "learning_rate": 1.6072852718174677e-05, "loss": 0.6995, "step": 8910 }, { "epoch": 0.313997022472405, "grad_norm": 5.575188159942627, "learning_rate": 1.607194595263133e-05, "loss": 5.8083, "step": 8911 }, { "epoch": 0.31403225948536345, "grad_norm": 1.6393957138061523, "learning_rate": 1.6071039108000105e-05, "loss": 1.1133, "step": 8912 }, { "epoch": 0.31406749649832183, "grad_norm": 1.2397680282592773, "learning_rate": 1.607013218429283e-05, "loss": 0.8527, "step": 8913 }, { "epoch": 0.31410273351128026, "grad_norm": 1.1468806266784668, "learning_rate": 1.6069225181521306e-05, "loss": 0.9942, "step": 8914 }, { "epoch": 0.31413797052423864, "grad_norm": 1.387628436088562, "learning_rate": 1.6068318099697353e-05, "loss": 0.9486, "step": 8915 }, { "epoch": 0.3141732075371971, "grad_norm": 1.5109336376190186, "learning_rate": 1.606741093883278e-05, "loss": 1.6564, "step": 8916 }, { "epoch": 0.3142084445501555, "grad_norm": 1.2314118146896362, "learning_rate": 1.6066503698939412e-05, "loss": 0.8492, "step": 8917 }, { "epoch": 0.3142436815631139, "grad_norm": 5.623467922210693, "learning_rate": 1.6065596380029057e-05, "loss": 2.9901, "step": 8918 }, { "epoch": 0.3142789185760723, "grad_norm": 1.9595037698745728, "learning_rate": 1.606468898211354e-05, "loss": 1.0568, "step": 8919 }, { "epoch": 0.3143141555890307, "grad_norm": 4.067546844482422, "learning_rate": 1.606378150520467e-05, "loss": 3.5141, "step": 8920 }, { "epoch": 0.31434939260198913, "grad_norm": 1.9655152559280396, "learning_rate": 1.6062873949314277e-05, "loss": 1.0152, "step": 8921 }, { "epoch": 0.31438462961494756, "grad_norm": 7.96320915222168, "learning_rate": 1.6061966314454182e-05, "loss": 3.64, "step": 8922 }, { "epoch": 0.31441986662790594, "grad_norm": 1.3433181047439575, "learning_rate": 1.60610586006362e-05, "loss": 0.8018, "step": 8923 }, { "epoch": 0.3144551036408644, "grad_norm": 1.8607369661331177, "learning_rate": 1.6060150807872153e-05, "loss": 0.7531, "step": 8924 }, { "epoch": 0.31449034065382275, "grad_norm": 5.579258918762207, "learning_rate": 1.6059242936173877e-05, "loss": 3.1303, "step": 8925 }, { "epoch": 0.3145255776667812, "grad_norm": 7.410455703735352, "learning_rate": 1.6058334985553186e-05, "loss": 3.5457, "step": 8926 }, { "epoch": 0.3145608146797396, "grad_norm": 1.1742522716522217, "learning_rate": 1.6057426956021914e-05, "loss": 0.8855, "step": 8927 }, { "epoch": 0.314596051692698, "grad_norm": 3.3857338428497314, "learning_rate": 1.605651884759188e-05, "loss": 0.7781, "step": 8928 }, { "epoch": 0.31463128870565643, "grad_norm": 2.057178020477295, "learning_rate": 1.605561066027492e-05, "loss": 0.9757, "step": 8929 }, { "epoch": 0.3146665257186148, "grad_norm": 5.665741443634033, "learning_rate": 1.605470239408286e-05, "loss": 6.8851, "step": 8930 }, { "epoch": 0.31470176273157324, "grad_norm": 4.866024017333984, "learning_rate": 1.605379404902753e-05, "loss": 3.5983, "step": 8931 }, { "epoch": 0.3147369997445317, "grad_norm": 1.0838656425476074, "learning_rate": 1.605288562512076e-05, "loss": 1.1314, "step": 8932 }, { "epoch": 0.31477223675749005, "grad_norm": 3.705305337905884, "learning_rate": 1.605197712237438e-05, "loss": 4.2335, "step": 8933 }, { "epoch": 0.3148074737704485, "grad_norm": 3.0600438117980957, "learning_rate": 1.6051068540800232e-05, "loss": 3.6007, "step": 8934 }, { "epoch": 0.31484271078340687, "grad_norm": 1.3341068029403687, "learning_rate": 1.6050159880410147e-05, "loss": 0.9679, "step": 8935 }, { "epoch": 0.3148779477963653, "grad_norm": 1.5329222679138184, "learning_rate": 1.6049251141215952e-05, "loss": 0.8212, "step": 8936 }, { "epoch": 0.31491318480932373, "grad_norm": 1.260913372039795, "learning_rate": 1.60483423232295e-05, "loss": 0.9909, "step": 8937 }, { "epoch": 0.3149484218222821, "grad_norm": 1.1121503114700317, "learning_rate": 1.604743342646261e-05, "loss": 0.8391, "step": 8938 }, { "epoch": 0.31498365883524054, "grad_norm": 3.2485620975494385, "learning_rate": 1.6046524450927133e-05, "loss": 3.8014, "step": 8939 }, { "epoch": 0.3150188958481989, "grad_norm": 1.504642128944397, "learning_rate": 1.6045615396634903e-05, "loss": 0.7863, "step": 8940 }, { "epoch": 0.31505413286115735, "grad_norm": 1.4876511096954346, "learning_rate": 1.6044706263597763e-05, "loss": 0.8621, "step": 8941 }, { "epoch": 0.3150893698741158, "grad_norm": 3.0405964851379395, "learning_rate": 1.6043797051827556e-05, "loss": 4.9455, "step": 8942 }, { "epoch": 0.31512460688707417, "grad_norm": 1.4114936590194702, "learning_rate": 1.604288776133612e-05, "loss": 0.9589, "step": 8943 }, { "epoch": 0.3151598439000326, "grad_norm": 3.617560386657715, "learning_rate": 1.6041978392135302e-05, "loss": 3.0429, "step": 8944 }, { "epoch": 0.31519508091299103, "grad_norm": 4.628513336181641, "learning_rate": 1.6041068944236944e-05, "loss": 5.8812, "step": 8945 }, { "epoch": 0.3152303179259494, "grad_norm": 2.5656309127807617, "learning_rate": 1.60401594176529e-05, "loss": 3.5305, "step": 8946 }, { "epoch": 0.31526555493890784, "grad_norm": 5.775937557220459, "learning_rate": 1.6039249812395007e-05, "loss": 5.4691, "step": 8947 }, { "epoch": 0.3153007919518662, "grad_norm": 1.1880083084106445, "learning_rate": 1.6038340128475113e-05, "loss": 0.7986, "step": 8948 }, { "epoch": 0.31533602896482466, "grad_norm": 1.2575608491897583, "learning_rate": 1.6037430365905072e-05, "loss": 0.9861, "step": 8949 }, { "epoch": 0.3153712659777831, "grad_norm": 1.1390224695205688, "learning_rate": 1.603652052469673e-05, "loss": 0.8204, "step": 8950 }, { "epoch": 0.31540650299074147, "grad_norm": 1.377197027206421, "learning_rate": 1.603561060486194e-05, "loss": 0.8506, "step": 8951 }, { "epoch": 0.3154417400036999, "grad_norm": 1.6739670038223267, "learning_rate": 1.603470060641255e-05, "loss": 0.9763, "step": 8952 }, { "epoch": 0.3154769770166583, "grad_norm": 3.7933106422424316, "learning_rate": 1.603379052936042e-05, "loss": 3.7885, "step": 8953 }, { "epoch": 0.3155122140296167, "grad_norm": 5.236859321594238, "learning_rate": 1.60328803737174e-05, "loss": 3.5083, "step": 8954 }, { "epoch": 0.31554745104257514, "grad_norm": 1.5214124917984009, "learning_rate": 1.603197013949534e-05, "loss": 0.9559, "step": 8955 }, { "epoch": 0.3155826880555335, "grad_norm": 6.083425998687744, "learning_rate": 1.603105982670611e-05, "loss": 4.604, "step": 8956 }, { "epoch": 0.31561792506849196, "grad_norm": 8.273496627807617, "learning_rate": 1.603014943536155e-05, "loss": 8.0256, "step": 8957 }, { "epoch": 0.31565316208145033, "grad_norm": 1.3369940519332886, "learning_rate": 1.6029238965473527e-05, "loss": 0.8051, "step": 8958 }, { "epoch": 0.31568839909440877, "grad_norm": 1.8393594026565552, "learning_rate": 1.6028328417053897e-05, "loss": 0.8366, "step": 8959 }, { "epoch": 0.3157236361073672, "grad_norm": 0.7962338924407959, "learning_rate": 1.6027417790114524e-05, "loss": 0.8535, "step": 8960 }, { "epoch": 0.3157588731203256, "grad_norm": 1.5174672603607178, "learning_rate": 1.6026507084667268e-05, "loss": 1.1014, "step": 8961 }, { "epoch": 0.315794110133284, "grad_norm": 2.7182390689849854, "learning_rate": 1.6025596300723988e-05, "loss": 2.8047, "step": 8962 }, { "epoch": 0.3158293471462424, "grad_norm": 0.820388674736023, "learning_rate": 1.6024685438296546e-05, "loss": 0.871, "step": 8963 }, { "epoch": 0.3158645841592008, "grad_norm": 2.329477548599243, "learning_rate": 1.6023774497396808e-05, "loss": 0.9368, "step": 8964 }, { "epoch": 0.31589982117215926, "grad_norm": 3.335968017578125, "learning_rate": 1.6022863478036645e-05, "loss": 3.6197, "step": 8965 }, { "epoch": 0.31593505818511763, "grad_norm": 1.1731700897216797, "learning_rate": 1.6021952380227915e-05, "loss": 0.9396, "step": 8966 }, { "epoch": 0.31597029519807607, "grad_norm": 1.5463931560516357, "learning_rate": 1.6021041203982488e-05, "loss": 0.8943, "step": 8967 }, { "epoch": 0.31600553221103445, "grad_norm": 5.705263614654541, "learning_rate": 1.6020129949312232e-05, "loss": 5.2857, "step": 8968 }, { "epoch": 0.3160407692239929, "grad_norm": 5.967311382293701, "learning_rate": 1.601921861622902e-05, "loss": 3.5603, "step": 8969 }, { "epoch": 0.3160760062369513, "grad_norm": 3.8067102432250977, "learning_rate": 1.6018307204744714e-05, "loss": 3.6598, "step": 8970 }, { "epoch": 0.3161112432499097, "grad_norm": 2.3950183391571045, "learning_rate": 1.6017395714871193e-05, "loss": 1.135, "step": 8971 }, { "epoch": 0.3161464802628681, "grad_norm": 8.171628952026367, "learning_rate": 1.6016484146620327e-05, "loss": 5.0894, "step": 8972 }, { "epoch": 0.3161817172758265, "grad_norm": 4.624574184417725, "learning_rate": 1.6015572500003986e-05, "loss": 3.3473, "step": 8973 }, { "epoch": 0.31621695428878493, "grad_norm": 5.089569568634033, "learning_rate": 1.6014660775034047e-05, "loss": 3.9155, "step": 8974 }, { "epoch": 0.31625219130174337, "grad_norm": 1.5641148090362549, "learning_rate": 1.6013748971722387e-05, "loss": 0.9107, "step": 8975 }, { "epoch": 0.31628742831470175, "grad_norm": 6.431880950927734, "learning_rate": 1.601283709008088e-05, "loss": 8.0005, "step": 8976 }, { "epoch": 0.3163226653276602, "grad_norm": 1.4879509210586548, "learning_rate": 1.6011925130121405e-05, "loss": 0.8497, "step": 8977 }, { "epoch": 0.3163579023406186, "grad_norm": 1.2651687860488892, "learning_rate": 1.601101309185584e-05, "loss": 1.0668, "step": 8978 }, { "epoch": 0.316393139353577, "grad_norm": 1.2457479238510132, "learning_rate": 1.6010100975296058e-05, "loss": 0.767, "step": 8979 }, { "epoch": 0.3164283763665354, "grad_norm": 3.4588518142700195, "learning_rate": 1.600918878045395e-05, "loss": 3.1465, "step": 8980 }, { "epoch": 0.3164636133794938, "grad_norm": 5.6431708335876465, "learning_rate": 1.6008276507341393e-05, "loss": 3.8829, "step": 8981 }, { "epoch": 0.31649885039245224, "grad_norm": 3.990323781967163, "learning_rate": 1.6007364155970272e-05, "loss": 4.1074, "step": 8982 }, { "epoch": 0.31653408740541067, "grad_norm": 1.344948649406433, "learning_rate": 1.6006451726352463e-05, "loss": 0.9845, "step": 8983 }, { "epoch": 0.31656932441836905, "grad_norm": 1.6535086631774902, "learning_rate": 1.6005539218499855e-05, "loss": 0.9322, "step": 8984 }, { "epoch": 0.3166045614313275, "grad_norm": 6.310959815979004, "learning_rate": 1.6004626632424333e-05, "loss": 5.5005, "step": 8985 }, { "epoch": 0.31663979844428586, "grad_norm": 1.2224947214126587, "learning_rate": 1.6003713968137787e-05, "loss": 1.1534, "step": 8986 }, { "epoch": 0.3166750354572443, "grad_norm": 1.5793418884277344, "learning_rate": 1.6002801225652103e-05, "loss": 0.8333, "step": 8987 }, { "epoch": 0.3167102724702027, "grad_norm": 4.644367694854736, "learning_rate": 1.6001888404979165e-05, "loss": 3.2922, "step": 8988 }, { "epoch": 0.3167455094831611, "grad_norm": 3.196021318435669, "learning_rate": 1.6000975506130866e-05, "loss": 3.1325, "step": 8989 }, { "epoch": 0.31678074649611954, "grad_norm": 1.6800600290298462, "learning_rate": 1.60000625291191e-05, "loss": 0.8169, "step": 8990 }, { "epoch": 0.3168159835090779, "grad_norm": 1.2116999626159668, "learning_rate": 1.5999149473955752e-05, "loss": 1.0155, "step": 8991 }, { "epoch": 0.31685122052203635, "grad_norm": 4.846428871154785, "learning_rate": 1.5998236340652718e-05, "loss": 3.7264, "step": 8992 }, { "epoch": 0.3168864575349948, "grad_norm": 1.745078206062317, "learning_rate": 1.5997323129221893e-05, "loss": 0.9441, "step": 8993 }, { "epoch": 0.31692169454795316, "grad_norm": 4.785328388214111, "learning_rate": 1.599640983967517e-05, "loss": 3.3751, "step": 8994 }, { "epoch": 0.3169569315609116, "grad_norm": 1.1266334056854248, "learning_rate": 1.5995496472024447e-05, "loss": 0.9004, "step": 8995 }, { "epoch": 0.31699216857386997, "grad_norm": 3.443263292312622, "learning_rate": 1.5994583026281617e-05, "loss": 3.4749, "step": 8996 }, { "epoch": 0.3170274055868284, "grad_norm": 3.622392416000366, "learning_rate": 1.599366950245858e-05, "loss": 4.9269, "step": 8997 }, { "epoch": 0.31706264259978684, "grad_norm": 2.612342357635498, "learning_rate": 1.5992755900567233e-05, "loss": 2.9847, "step": 8998 }, { "epoch": 0.3170978796127452, "grad_norm": 3.5255789756774902, "learning_rate": 1.599184222061948e-05, "loss": 3.0698, "step": 8999 }, { "epoch": 0.31713311662570365, "grad_norm": 1.1421713829040527, "learning_rate": 1.5990928462627213e-05, "loss": 0.8534, "step": 9000 }, { "epoch": 0.317168353638662, "grad_norm": 4.28450345993042, "learning_rate": 1.5990014626602343e-05, "loss": 4.0961, "step": 9001 }, { "epoch": 0.31720359065162046, "grad_norm": 3.1372876167297363, "learning_rate": 1.5989100712556773e-05, "loss": 3.6633, "step": 9002 }, { "epoch": 0.3172388276645789, "grad_norm": 2.0300216674804688, "learning_rate": 1.5988186720502395e-05, "loss": 0.8396, "step": 9003 }, { "epoch": 0.31727406467753727, "grad_norm": 2.7411768436431885, "learning_rate": 1.598727265045113e-05, "loss": 3.4673, "step": 9004 }, { "epoch": 0.3173093016904957, "grad_norm": 4.2471723556518555, "learning_rate": 1.598635850241487e-05, "loss": 5.5596, "step": 9005 }, { "epoch": 0.3173445387034541, "grad_norm": 4.678719997406006, "learning_rate": 1.5985444276405534e-05, "loss": 2.9148, "step": 9006 }, { "epoch": 0.3173797757164125, "grad_norm": 6.6872358322143555, "learning_rate": 1.5984529972435018e-05, "loss": 3.2397, "step": 9007 }, { "epoch": 0.31741501272937095, "grad_norm": 3.104151964187622, "learning_rate": 1.598361559051524e-05, "loss": 2.9899, "step": 9008 }, { "epoch": 0.3174502497423293, "grad_norm": 1.275699496269226, "learning_rate": 1.5982701130658106e-05, "loss": 1.0015, "step": 9009 }, { "epoch": 0.31748548675528776, "grad_norm": 7.521255970001221, "learning_rate": 1.598178659287553e-05, "loss": 3.4773, "step": 9010 }, { "epoch": 0.31752072376824614, "grad_norm": 1.328999638557434, "learning_rate": 1.5980871977179423e-05, "loss": 0.8725, "step": 9011 }, { "epoch": 0.31755596078120457, "grad_norm": 1.7232481241226196, "learning_rate": 1.597995728358169e-05, "loss": 0.8461, "step": 9012 }, { "epoch": 0.317591197794163, "grad_norm": 5.818004608154297, "learning_rate": 1.5979042512094255e-05, "loss": 3.6516, "step": 9013 }, { "epoch": 0.3176264348071214, "grad_norm": 1.4053384065628052, "learning_rate": 1.597812766272903e-05, "loss": 0.8336, "step": 9014 }, { "epoch": 0.3176616718200798, "grad_norm": 1.6108739376068115, "learning_rate": 1.5977212735497932e-05, "loss": 0.8794, "step": 9015 }, { "epoch": 0.31769690883303825, "grad_norm": 1.7100104093551636, "learning_rate": 1.5976297730412875e-05, "loss": 0.888, "step": 9016 }, { "epoch": 0.3177321458459966, "grad_norm": 4.960547924041748, "learning_rate": 1.5975382647485778e-05, "loss": 3.1357, "step": 9017 }, { "epoch": 0.31776738285895506, "grad_norm": 4.932301998138428, "learning_rate": 1.5974467486728564e-05, "loss": 3.7117, "step": 9018 }, { "epoch": 0.31780261987191344, "grad_norm": 1.5263466835021973, "learning_rate": 1.5973552248153146e-05, "loss": 1.0305, "step": 9019 }, { "epoch": 0.31783785688487187, "grad_norm": 1.5365530252456665, "learning_rate": 1.597263693177145e-05, "loss": 0.8058, "step": 9020 }, { "epoch": 0.3178730938978303, "grad_norm": 1.5866550207138062, "learning_rate": 1.59717215375954e-05, "loss": 0.7952, "step": 9021 }, { "epoch": 0.3179083309107887, "grad_norm": 5.185558319091797, "learning_rate": 1.5970806065636914e-05, "loss": 3.4047, "step": 9022 }, { "epoch": 0.3179435679237471, "grad_norm": 4.188711166381836, "learning_rate": 1.596989051590792e-05, "loss": 3.6957, "step": 9023 }, { "epoch": 0.3179788049367055, "grad_norm": 2.55890154838562, "learning_rate": 1.596897488842034e-05, "loss": 3.3236, "step": 9024 }, { "epoch": 0.31801404194966393, "grad_norm": 2.48610258102417, "learning_rate": 1.5968059183186103e-05, "loss": 2.4995, "step": 9025 }, { "epoch": 0.31804927896262236, "grad_norm": 1.6359764337539673, "learning_rate": 1.5967143400217134e-05, "loss": 0.8047, "step": 9026 }, { "epoch": 0.31808451597558074, "grad_norm": 4.0393781661987305, "learning_rate": 1.596622753952536e-05, "loss": 5.9476, "step": 9027 }, { "epoch": 0.3181197529885392, "grad_norm": 5.7125020027160645, "learning_rate": 1.5965311601122716e-05, "loss": 3.0255, "step": 9028 }, { "epoch": 0.31815499000149755, "grad_norm": 1.2050493955612183, "learning_rate": 1.596439558502113e-05, "loss": 0.924, "step": 9029 }, { "epoch": 0.318190227014456, "grad_norm": 4.587491035461426, "learning_rate": 1.596347949123253e-05, "loss": 3.1199, "step": 9030 }, { "epoch": 0.3182254640274144, "grad_norm": 3.300840139389038, "learning_rate": 1.5962563319768847e-05, "loss": 2.9605, "step": 9031 }, { "epoch": 0.3182607010403728, "grad_norm": 3.1355361938476562, "learning_rate": 1.5961647070642018e-05, "loss": 3.7127, "step": 9032 }, { "epoch": 0.31829593805333123, "grad_norm": 1.086936354637146, "learning_rate": 1.596073074386398e-05, "loss": 0.8835, "step": 9033 }, { "epoch": 0.3183311750662896, "grad_norm": 2.7033298015594482, "learning_rate": 1.595981433944666e-05, "loss": 3.163, "step": 9034 }, { "epoch": 0.31836641207924804, "grad_norm": 2.4928228855133057, "learning_rate": 1.5958897857402002e-05, "loss": 2.9915, "step": 9035 }, { "epoch": 0.3184016490922065, "grad_norm": 1.672191858291626, "learning_rate": 1.595798129774194e-05, "loss": 0.7782, "step": 9036 }, { "epoch": 0.31843688610516485, "grad_norm": 5.683088302612305, "learning_rate": 1.595706466047841e-05, "loss": 3.5417, "step": 9037 }, { "epoch": 0.3184721231181233, "grad_norm": 4.600060939788818, "learning_rate": 1.595614794562336e-05, "loss": 2.9483, "step": 9038 }, { "epoch": 0.31850736013108166, "grad_norm": 7.516275882720947, "learning_rate": 1.5955231153188718e-05, "loss": 6.6521, "step": 9039 }, { "epoch": 0.3185425971440401, "grad_norm": 8.263877868652344, "learning_rate": 1.5954314283186437e-05, "loss": 8.1502, "step": 9040 }, { "epoch": 0.31857783415699853, "grad_norm": 1.3199015855789185, "learning_rate": 1.5953397335628447e-05, "loss": 1.3068, "step": 9041 }, { "epoch": 0.3186130711699569, "grad_norm": 4.325442790985107, "learning_rate": 1.5952480310526706e-05, "loss": 3.647, "step": 9042 }, { "epoch": 0.31864830818291534, "grad_norm": 4.416701793670654, "learning_rate": 1.595156320789315e-05, "loss": 2.9289, "step": 9043 }, { "epoch": 0.3186835451958737, "grad_norm": 1.65653395652771, "learning_rate": 1.595064602773972e-05, "loss": 0.7273, "step": 9044 }, { "epoch": 0.31871878220883215, "grad_norm": 5.014448642730713, "learning_rate": 1.594972877007837e-05, "loss": 3.5818, "step": 9045 }, { "epoch": 0.3187540192217906, "grad_norm": 3.6396353244781494, "learning_rate": 1.5948811434921046e-05, "loss": 3.6482, "step": 9046 }, { "epoch": 0.31878925623474896, "grad_norm": 4.558260440826416, "learning_rate": 1.5947894022279694e-05, "loss": 5.3057, "step": 9047 }, { "epoch": 0.3188244932477074, "grad_norm": 0.9722194671630859, "learning_rate": 1.5946976532166267e-05, "loss": 1.142, "step": 9048 }, { "epoch": 0.31885973026066583, "grad_norm": 2.89371657371521, "learning_rate": 1.594605896459271e-05, "loss": 2.5037, "step": 9049 }, { "epoch": 0.3188949672736242, "grad_norm": 1.8081876039505005, "learning_rate": 1.594514131957098e-05, "loss": 0.8395, "step": 9050 }, { "epoch": 0.31893020428658264, "grad_norm": 4.791382789611816, "learning_rate": 1.5944223597113025e-05, "loss": 3.6553, "step": 9051 }, { "epoch": 0.318965441299541, "grad_norm": 2.75177264213562, "learning_rate": 1.5943305797230802e-05, "loss": 3.0, "step": 9052 }, { "epoch": 0.31900067831249945, "grad_norm": 7.434357166290283, "learning_rate": 1.5942387919936263e-05, "loss": 11.7373, "step": 9053 }, { "epoch": 0.3190359153254579, "grad_norm": 2.0535085201263428, "learning_rate": 1.5941469965241364e-05, "loss": 0.891, "step": 9054 }, { "epoch": 0.31907115233841626, "grad_norm": 1.3515000343322754, "learning_rate": 1.5940551933158062e-05, "loss": 1.0107, "step": 9055 }, { "epoch": 0.3191063893513747, "grad_norm": 2.9091484546661377, "learning_rate": 1.5939633823698314e-05, "loss": 2.9994, "step": 9056 }, { "epoch": 0.3191416263643331, "grad_norm": 0.7687622904777527, "learning_rate": 1.5938715636874083e-05, "loss": 0.7845, "step": 9057 }, { "epoch": 0.3191768633772915, "grad_norm": 2.7493112087249756, "learning_rate": 1.5937797372697323e-05, "loss": 3.7469, "step": 9058 }, { "epoch": 0.31921210039024994, "grad_norm": 0.891662061214447, "learning_rate": 1.5936879031179993e-05, "loss": 1.2707, "step": 9059 }, { "epoch": 0.3192473374032083, "grad_norm": 5.363277435302734, "learning_rate": 1.5935960612334055e-05, "loss": 6.4882, "step": 9060 }, { "epoch": 0.31928257441616675, "grad_norm": 2.5651116371154785, "learning_rate": 1.593504211617148e-05, "loss": 3.0637, "step": 9061 }, { "epoch": 0.31931781142912513, "grad_norm": 3.4665534496307373, "learning_rate": 1.5934123542704225e-05, "loss": 3.3828, "step": 9062 }, { "epoch": 0.31935304844208356, "grad_norm": 4.560170650482178, "learning_rate": 1.593320489194425e-05, "loss": 3.7171, "step": 9063 }, { "epoch": 0.319388285455042, "grad_norm": 2.5162220001220703, "learning_rate": 1.593228616390353e-05, "loss": 3.0945, "step": 9064 }, { "epoch": 0.3194235224680004, "grad_norm": 5.8379597663879395, "learning_rate": 1.593136735859402e-05, "loss": 7.7102, "step": 9065 }, { "epoch": 0.3194587594809588, "grad_norm": 3.3780691623687744, "learning_rate": 1.5930448476027702e-05, "loss": 3.203, "step": 9066 }, { "epoch": 0.3194939964939172, "grad_norm": 3.670888662338257, "learning_rate": 1.5929529516216535e-05, "loss": 3.8795, "step": 9067 }, { "epoch": 0.3195292335068756, "grad_norm": 1.3114699125289917, "learning_rate": 1.592861047917249e-05, "loss": 0.9096, "step": 9068 }, { "epoch": 0.31956447051983405, "grad_norm": 6.363326072692871, "learning_rate": 1.5927691364907542e-05, "loss": 3.3337, "step": 9069 }, { "epoch": 0.31959970753279243, "grad_norm": 4.05754280090332, "learning_rate": 1.5926772173433657e-05, "loss": 3.4774, "step": 9070 }, { "epoch": 0.31963494454575087, "grad_norm": 4.380812168121338, "learning_rate": 1.5925852904762808e-05, "loss": 3.1275, "step": 9071 }, { "epoch": 0.31967018155870924, "grad_norm": 1.2639516592025757, "learning_rate": 1.592493355890697e-05, "loss": 1.0073, "step": 9072 }, { "epoch": 0.3197054185716677, "grad_norm": 1.433393120765686, "learning_rate": 1.592401413587812e-05, "loss": 1.0659, "step": 9073 }, { "epoch": 0.3197406555846261, "grad_norm": 3.9742300510406494, "learning_rate": 1.592309463568823e-05, "loss": 3.2195, "step": 9074 }, { "epoch": 0.3197758925975845, "grad_norm": 2.218757152557373, "learning_rate": 1.5922175058349277e-05, "loss": 3.1699, "step": 9075 }, { "epoch": 0.3198111296105429, "grad_norm": 5.300759315490723, "learning_rate": 1.592125540387324e-05, "loss": 2.555, "step": 9076 }, { "epoch": 0.3198463666235013, "grad_norm": 5.624112606048584, "learning_rate": 1.5920335672272096e-05, "loss": 6.0766, "step": 9077 }, { "epoch": 0.31988160363645973, "grad_norm": 5.2901482582092285, "learning_rate": 1.5919415863557825e-05, "loss": 3.4292, "step": 9078 }, { "epoch": 0.31991684064941817, "grad_norm": 1.924615740776062, "learning_rate": 1.5918495977742413e-05, "loss": 0.8461, "step": 9079 }, { "epoch": 0.31995207766237654, "grad_norm": 5.143081188201904, "learning_rate": 1.591757601483783e-05, "loss": 3.1649, "step": 9080 }, { "epoch": 0.319987314675335, "grad_norm": 6.8287882804870605, "learning_rate": 1.5916655974856074e-05, "loss": 6.1354, "step": 9081 }, { "epoch": 0.3200225516882934, "grad_norm": 3.3817639350891113, "learning_rate": 1.5915735857809114e-05, "loss": 3.7435, "step": 9082 }, { "epoch": 0.3200577887012518, "grad_norm": 0.9672698974609375, "learning_rate": 1.5914815663708946e-05, "loss": 0.7935, "step": 9083 }, { "epoch": 0.3200930257142102, "grad_norm": 7.219888687133789, "learning_rate": 1.5913895392567543e-05, "loss": 3.6526, "step": 9084 }, { "epoch": 0.3201282627271686, "grad_norm": 3.563295364379883, "learning_rate": 1.5912975044396907e-05, "loss": 3.0876, "step": 9085 }, { "epoch": 0.32016349974012703, "grad_norm": 4.853937149047852, "learning_rate": 1.5912054619209013e-05, "loss": 3.4012, "step": 9086 }, { "epoch": 0.32019873675308547, "grad_norm": 4.092740058898926, "learning_rate": 1.5911134117015857e-05, "loss": 3.3645, "step": 9087 }, { "epoch": 0.32023397376604384, "grad_norm": 4.436807155609131, "learning_rate": 1.5910213537829426e-05, "loss": 2.9779, "step": 9088 }, { "epoch": 0.3202692107790023, "grad_norm": 1.1283884048461914, "learning_rate": 1.5909292881661713e-05, "loss": 1.1222, "step": 9089 }, { "epoch": 0.32030444779196066, "grad_norm": 4.389105319976807, "learning_rate": 1.59083721485247e-05, "loss": 3.4048, "step": 9090 }, { "epoch": 0.3203396848049191, "grad_norm": 5.0042290687561035, "learning_rate": 1.5907451338430394e-05, "loss": 5.9987, "step": 9091 }, { "epoch": 0.3203749218178775, "grad_norm": 3.0836548805236816, "learning_rate": 1.590653045139078e-05, "loss": 3.3564, "step": 9092 }, { "epoch": 0.3204101588308359, "grad_norm": 5.680730819702148, "learning_rate": 1.5905609487417852e-05, "loss": 6.2433, "step": 9093 }, { "epoch": 0.32044539584379433, "grad_norm": 1.2061372995376587, "learning_rate": 1.590468844652361e-05, "loss": 0.8658, "step": 9094 }, { "epoch": 0.3204806328567527, "grad_norm": 5.383439064025879, "learning_rate": 1.590376732872005e-05, "loss": 3.3516, "step": 9095 }, { "epoch": 0.32051586986971115, "grad_norm": 4.398301601409912, "learning_rate": 1.5902846134019168e-05, "loss": 2.8166, "step": 9096 }, { "epoch": 0.3205511068826696, "grad_norm": 1.0718753337860107, "learning_rate": 1.590192486243296e-05, "loss": 0.7839, "step": 9097 }, { "epoch": 0.32058634389562796, "grad_norm": 2.4728763103485107, "learning_rate": 1.5901003513973433e-05, "loss": 2.8909, "step": 9098 }, { "epoch": 0.3206215809085864, "grad_norm": 5.163703441619873, "learning_rate": 1.5900082088652584e-05, "loss": 5.6139, "step": 9099 }, { "epoch": 0.32065681792154477, "grad_norm": 1.3515716791152954, "learning_rate": 1.5899160586482413e-05, "loss": 0.9927, "step": 9100 }, { "epoch": 0.3206920549345032, "grad_norm": 3.190833330154419, "learning_rate": 1.5898239007474923e-05, "loss": 2.7222, "step": 9101 }, { "epoch": 0.32072729194746163, "grad_norm": 1.609687328338623, "learning_rate": 1.5897317351642117e-05, "loss": 0.9046, "step": 9102 }, { "epoch": 0.32076252896042, "grad_norm": 5.103118896484375, "learning_rate": 1.5896395618996007e-05, "loss": 2.9516, "step": 9103 }, { "epoch": 0.32079776597337845, "grad_norm": 2.090419292449951, "learning_rate": 1.589547380954859e-05, "loss": 1.0788, "step": 9104 }, { "epoch": 0.3208330029863368, "grad_norm": 1.0155152082443237, "learning_rate": 1.589455192331187e-05, "loss": 0.6958, "step": 9105 }, { "epoch": 0.32086823999929526, "grad_norm": 4.6723408699035645, "learning_rate": 1.5893629960297867e-05, "loss": 3.1391, "step": 9106 }, { "epoch": 0.3209034770122537, "grad_norm": 1.780306100845337, "learning_rate": 1.5892707920518585e-05, "loss": 0.8238, "step": 9107 }, { "epoch": 0.32093871402521207, "grad_norm": 1.2338826656341553, "learning_rate": 1.5891785803986026e-05, "loss": 1.0646, "step": 9108 }, { "epoch": 0.3209739510381705, "grad_norm": 4.526577949523926, "learning_rate": 1.5890863610712208e-05, "loss": 3.2593, "step": 9109 }, { "epoch": 0.3210091880511289, "grad_norm": 1.2995349168777466, "learning_rate": 1.588994134070914e-05, "loss": 0.9005, "step": 9110 }, { "epoch": 0.3210444250640873, "grad_norm": 7.7277655601501465, "learning_rate": 1.588901899398884e-05, "loss": 3.9765, "step": 9111 }, { "epoch": 0.32107966207704575, "grad_norm": 2.252070426940918, "learning_rate": 1.5888096570563315e-05, "loss": 1.1824, "step": 9112 }, { "epoch": 0.3211148990900041, "grad_norm": 1.4454530477523804, "learning_rate": 1.588717407044458e-05, "loss": 0.8568, "step": 9113 }, { "epoch": 0.32115013610296256, "grad_norm": 1.9198070764541626, "learning_rate": 1.5886251493644656e-05, "loss": 0.7619, "step": 9114 }, { "epoch": 0.32118537311592094, "grad_norm": 4.549286842346191, "learning_rate": 1.5885328840175557e-05, "loss": 3.2952, "step": 9115 }, { "epoch": 0.32122061012887937, "grad_norm": 3.924565076828003, "learning_rate": 1.58844061100493e-05, "loss": 3.4792, "step": 9116 }, { "epoch": 0.3212558471418378, "grad_norm": 3.9521005153656006, "learning_rate": 1.58834833032779e-05, "loss": 3.5008, "step": 9117 }, { "epoch": 0.3212910841547962, "grad_norm": 1.8000452518463135, "learning_rate": 1.588256041987338e-05, "loss": 1.0121, "step": 9118 }, { "epoch": 0.3213263211677546, "grad_norm": 4.43837833404541, "learning_rate": 1.588163745984777e-05, "loss": 6.061, "step": 9119 }, { "epoch": 0.32136155818071305, "grad_norm": 1.9132601022720337, "learning_rate": 1.5880714423213073e-05, "loss": 0.889, "step": 9120 }, { "epoch": 0.3213967951936714, "grad_norm": 2.872901201248169, "learning_rate": 1.5879791309981328e-05, "loss": 3.3938, "step": 9121 }, { "epoch": 0.32143203220662986, "grad_norm": 0.9299737215042114, "learning_rate": 1.5878868120164546e-05, "loss": 1.1979, "step": 9122 }, { "epoch": 0.32146726921958824, "grad_norm": 3.646820545196533, "learning_rate": 1.5877944853774763e-05, "loss": 3.453, "step": 9123 }, { "epoch": 0.32150250623254667, "grad_norm": 2.0516088008880615, "learning_rate": 1.5877021510824e-05, "loss": 0.837, "step": 9124 }, { "epoch": 0.3215377432455051, "grad_norm": 4.630493640899658, "learning_rate": 1.587609809132428e-05, "loss": 3.8456, "step": 9125 }, { "epoch": 0.3215729802584635, "grad_norm": 6.3245391845703125, "learning_rate": 1.5875174595287636e-05, "loss": 6.4198, "step": 9126 }, { "epoch": 0.3216082172714219, "grad_norm": 1.336349606513977, "learning_rate": 1.587425102272609e-05, "loss": 1.1374, "step": 9127 }, { "epoch": 0.3216434542843803, "grad_norm": 3.246305227279663, "learning_rate": 1.5873327373651684e-05, "loss": 2.8677, "step": 9128 }, { "epoch": 0.3216786912973387, "grad_norm": 2.1276750564575195, "learning_rate": 1.5872403648076437e-05, "loss": 0.8919, "step": 9129 }, { "epoch": 0.32171392831029716, "grad_norm": 2.9075818061828613, "learning_rate": 1.5871479846012382e-05, "loss": 3.0923, "step": 9130 }, { "epoch": 0.32174916532325554, "grad_norm": 1.5787802934646606, "learning_rate": 1.5870555967471556e-05, "loss": 0.8474, "step": 9131 }, { "epoch": 0.32178440233621397, "grad_norm": 1.5577538013458252, "learning_rate": 1.5869632012465993e-05, "loss": 1.1951, "step": 9132 }, { "epoch": 0.32181963934917235, "grad_norm": 1.0948718786239624, "learning_rate": 1.5868707981007724e-05, "loss": 0.684, "step": 9133 }, { "epoch": 0.3218548763621308, "grad_norm": 4.743138790130615, "learning_rate": 1.5867783873108783e-05, "loss": 5.3641, "step": 9134 }, { "epoch": 0.3218901133750892, "grad_norm": 7.188619613647461, "learning_rate": 1.5866859688781216e-05, "loss": 5.664, "step": 9135 }, { "epoch": 0.3219253503880476, "grad_norm": 2.726168632507324, "learning_rate": 1.586593542803705e-05, "loss": 2.7443, "step": 9136 }, { "epoch": 0.321960587401006, "grad_norm": 4.384655475616455, "learning_rate": 1.5865011090888327e-05, "loss": 3.752, "step": 9137 }, { "epoch": 0.3219958244139644, "grad_norm": 1.3344194889068604, "learning_rate": 1.586408667734709e-05, "loss": 0.8992, "step": 9138 }, { "epoch": 0.32203106142692284, "grad_norm": 3.3935773372650146, "learning_rate": 1.5863162187425372e-05, "loss": 3.9197, "step": 9139 }, { "epoch": 0.32206629843988127, "grad_norm": 3.8898754119873047, "learning_rate": 1.5862237621135227e-05, "loss": 3.9613, "step": 9140 }, { "epoch": 0.32210153545283965, "grad_norm": 1.645003080368042, "learning_rate": 1.5861312978488682e-05, "loss": 0.882, "step": 9141 }, { "epoch": 0.3221367724657981, "grad_norm": 6.9187211990356445, "learning_rate": 1.5860388259497795e-05, "loss": 3.639, "step": 9142 }, { "epoch": 0.32217200947875646, "grad_norm": 3.046675205230713, "learning_rate": 1.5859463464174606e-05, "loss": 3.6366, "step": 9143 }, { "epoch": 0.3222072464917149, "grad_norm": 7.728941440582275, "learning_rate": 1.5858538592531152e-05, "loss": 5.99, "step": 9144 }, { "epoch": 0.3222424835046733, "grad_norm": 1.476227879524231, "learning_rate": 1.5857613644579494e-05, "loss": 0.8608, "step": 9145 }, { "epoch": 0.3222777205176317, "grad_norm": 1.4176620244979858, "learning_rate": 1.5856688620331668e-05, "loss": 0.8375, "step": 9146 }, { "epoch": 0.32231295753059014, "grad_norm": 1.7106608152389526, "learning_rate": 1.585576351979973e-05, "loss": 0.7595, "step": 9147 }, { "epoch": 0.3223481945435485, "grad_norm": 5.93208646774292, "learning_rate": 1.585483834299572e-05, "loss": 5.6629, "step": 9148 }, { "epoch": 0.32238343155650695, "grad_norm": 0.8704789876937866, "learning_rate": 1.5853913089931697e-05, "loss": 0.7521, "step": 9149 }, { "epoch": 0.3224186685694654, "grad_norm": 4.708946704864502, "learning_rate": 1.585298776061971e-05, "loss": 4.8051, "step": 9150 }, { "epoch": 0.32245390558242376, "grad_norm": 0.8712553977966309, "learning_rate": 1.5852062355071817e-05, "loss": 0.8791, "step": 9151 }, { "epoch": 0.3224891425953822, "grad_norm": 1.6512340307235718, "learning_rate": 1.5851136873300062e-05, "loss": 0.8771, "step": 9152 }, { "epoch": 0.32252437960834063, "grad_norm": 3.720327138900757, "learning_rate": 1.5850211315316505e-05, "loss": 3.7829, "step": 9153 }, { "epoch": 0.322559616621299, "grad_norm": 1.489519476890564, "learning_rate": 1.5849285681133203e-05, "loss": 0.8251, "step": 9154 }, { "epoch": 0.32259485363425744, "grad_norm": 1.0745882987976074, "learning_rate": 1.5848359970762204e-05, "loss": 0.8238, "step": 9155 }, { "epoch": 0.3226300906472158, "grad_norm": 1.0735399723052979, "learning_rate": 1.5847434184215575e-05, "loss": 0.8185, "step": 9156 }, { "epoch": 0.32266532766017425, "grad_norm": 1.1294622421264648, "learning_rate": 1.5846508321505375e-05, "loss": 1.0314, "step": 9157 }, { "epoch": 0.3227005646731327, "grad_norm": 3.799893856048584, "learning_rate": 1.584558238264365e-05, "loss": 3.4651, "step": 9158 }, { "epoch": 0.32273580168609106, "grad_norm": 6.750129222869873, "learning_rate": 1.5844656367642478e-05, "loss": 4.04, "step": 9159 }, { "epoch": 0.3227710386990495, "grad_norm": 3.9964938163757324, "learning_rate": 1.5843730276513906e-05, "loss": 3.106, "step": 9160 }, { "epoch": 0.3228062757120079, "grad_norm": 5.4592366218566895, "learning_rate": 1.5842804109270006e-05, "loss": 6.297, "step": 9161 }, { "epoch": 0.3228415127249663, "grad_norm": 1.8853305578231812, "learning_rate": 1.5841877865922838e-05, "loss": 0.8287, "step": 9162 }, { "epoch": 0.32287674973792474, "grad_norm": 7.544528961181641, "learning_rate": 1.5840951546484467e-05, "loss": 5.8838, "step": 9163 }, { "epoch": 0.3229119867508831, "grad_norm": 4.130659580230713, "learning_rate": 1.584002515096696e-05, "loss": 3.1555, "step": 9164 }, { "epoch": 0.32294722376384155, "grad_norm": 5.365231037139893, "learning_rate": 1.583909867938238e-05, "loss": 5.955, "step": 9165 }, { "epoch": 0.32298246077679993, "grad_norm": 3.1624197959899902, "learning_rate": 1.5838172131742792e-05, "loss": 1.0465, "step": 9166 }, { "epoch": 0.32301769778975836, "grad_norm": 10.869022369384766, "learning_rate": 1.5837245508060273e-05, "loss": 3.569, "step": 9167 }, { "epoch": 0.3230529348027168, "grad_norm": 3.7651262283325195, "learning_rate": 1.5836318808346884e-05, "loss": 3.0328, "step": 9168 }, { "epoch": 0.3230881718156752, "grad_norm": 3.55234432220459, "learning_rate": 1.58353920326147e-05, "loss": 2.9764, "step": 9169 }, { "epoch": 0.3231234088286336, "grad_norm": 4.539318561553955, "learning_rate": 1.5834465180875795e-05, "loss": 3.8883, "step": 9170 }, { "epoch": 0.323158645841592, "grad_norm": 1.3861719369888306, "learning_rate": 1.5833538253142234e-05, "loss": 1.0007, "step": 9171 }, { "epoch": 0.3231938828545504, "grad_norm": 4.567380428314209, "learning_rate": 1.58326112494261e-05, "loss": 3.1958, "step": 9172 }, { "epoch": 0.32322911986750885, "grad_norm": 3.830998420715332, "learning_rate": 1.5831684169739456e-05, "loss": 3.336, "step": 9173 }, { "epoch": 0.32326435688046723, "grad_norm": 4.5064616203308105, "learning_rate": 1.583075701409438e-05, "loss": 3.4515, "step": 9174 }, { "epoch": 0.32329959389342566, "grad_norm": 1.53205406665802, "learning_rate": 1.582982978250296e-05, "loss": 0.8884, "step": 9175 }, { "epoch": 0.32333483090638404, "grad_norm": 1.8163264989852905, "learning_rate": 1.5828902474977258e-05, "loss": 0.911, "step": 9176 }, { "epoch": 0.3233700679193425, "grad_norm": 1.1827422380447388, "learning_rate": 1.582797509152936e-05, "loss": 0.9815, "step": 9177 }, { "epoch": 0.3234053049323009, "grad_norm": 1.3173847198486328, "learning_rate": 1.5827047632171346e-05, "loss": 0.6758, "step": 9178 }, { "epoch": 0.3234405419452593, "grad_norm": 2.473400831222534, "learning_rate": 1.5826120096915293e-05, "loss": 1.0171, "step": 9179 }, { "epoch": 0.3234757789582177, "grad_norm": 10.421956062316895, "learning_rate": 1.5825192485773287e-05, "loss": 6.6563, "step": 9180 }, { "epoch": 0.3235110159711761, "grad_norm": 1.5613508224487305, "learning_rate": 1.58242647987574e-05, "loss": 0.9175, "step": 9181 }, { "epoch": 0.32354625298413453, "grad_norm": 1.3025137186050415, "learning_rate": 1.5823337035879725e-05, "loss": 1.1805, "step": 9182 }, { "epoch": 0.32358148999709296, "grad_norm": 2.0442705154418945, "learning_rate": 1.5822409197152347e-05, "loss": 1.2117, "step": 9183 }, { "epoch": 0.32361672701005134, "grad_norm": 3.2712883949279785, "learning_rate": 1.5821481282587344e-05, "loss": 2.7444, "step": 9184 }, { "epoch": 0.3236519640230098, "grad_norm": 5.067145347595215, "learning_rate": 1.5820553292196806e-05, "loss": 4.0312, "step": 9185 }, { "epoch": 0.3236872010359682, "grad_norm": 3.4861433506011963, "learning_rate": 1.581962522599282e-05, "loss": 2.9374, "step": 9186 }, { "epoch": 0.3237224380489266, "grad_norm": 0.9626998901367188, "learning_rate": 1.5818697083987474e-05, "loss": 0.8898, "step": 9187 }, { "epoch": 0.323757675061885, "grad_norm": 1.4842267036437988, "learning_rate": 1.581776886619286e-05, "loss": 1.0744, "step": 9188 }, { "epoch": 0.3237929120748434, "grad_norm": 3.2899131774902344, "learning_rate": 1.5816840572621062e-05, "loss": 3.0435, "step": 9189 }, { "epoch": 0.32382814908780183, "grad_norm": 4.570506572723389, "learning_rate": 1.5815912203284175e-05, "loss": 4.1071, "step": 9190 }, { "epoch": 0.32386338610076026, "grad_norm": 1.5659980773925781, "learning_rate": 1.5814983758194292e-05, "loss": 1.0667, "step": 9191 }, { "epoch": 0.32389862311371864, "grad_norm": 1.5158848762512207, "learning_rate": 1.5814055237363503e-05, "loss": 1.1518, "step": 9192 }, { "epoch": 0.3239338601266771, "grad_norm": 0.9155935049057007, "learning_rate": 1.5813126640803905e-05, "loss": 0.7466, "step": 9193 }, { "epoch": 0.32396909713963545, "grad_norm": 1.2037076950073242, "learning_rate": 1.5812197968527595e-05, "loss": 1.0064, "step": 9194 }, { "epoch": 0.3240043341525939, "grad_norm": 5.979887008666992, "learning_rate": 1.5811269220546662e-05, "loss": 6.1465, "step": 9195 }, { "epoch": 0.3240395711655523, "grad_norm": 1.9677913188934326, "learning_rate": 1.5810340396873208e-05, "loss": 1.0561, "step": 9196 }, { "epoch": 0.3240748081785107, "grad_norm": 5.851032257080078, "learning_rate": 1.580941149751933e-05, "loss": 2.9356, "step": 9197 }, { "epoch": 0.32411004519146913, "grad_norm": 1.1800293922424316, "learning_rate": 1.580848252249713e-05, "loss": 1.2777, "step": 9198 }, { "epoch": 0.3241452822044275, "grad_norm": 8.793334007263184, "learning_rate": 1.5807553471818703e-05, "loss": 8.2218, "step": 9199 }, { "epoch": 0.32418051921738594, "grad_norm": 1.8019732236862183, "learning_rate": 1.5806624345496155e-05, "loss": 0.8647, "step": 9200 }, { "epoch": 0.3242157562303444, "grad_norm": 2.866645097732544, "learning_rate": 1.5805695143541578e-05, "loss": 3.324, "step": 9201 }, { "epoch": 0.32425099324330275, "grad_norm": 7.0794243812561035, "learning_rate": 1.5804765865967087e-05, "loss": 3.3431, "step": 9202 }, { "epoch": 0.3242862302562612, "grad_norm": 1.1647638082504272, "learning_rate": 1.5803836512784782e-05, "loss": 0.8419, "step": 9203 }, { "epoch": 0.32432146726921957, "grad_norm": 3.8552184104919434, "learning_rate": 1.5802907084006766e-05, "loss": 3.2812, "step": 9204 }, { "epoch": 0.324356704282178, "grad_norm": 1.689300537109375, "learning_rate": 1.5801977579645147e-05, "loss": 1.0734, "step": 9205 }, { "epoch": 0.32439194129513643, "grad_norm": 9.456263542175293, "learning_rate": 1.580104799971203e-05, "loss": 5.5915, "step": 9206 }, { "epoch": 0.3244271783080948, "grad_norm": 1.3505939245224, "learning_rate": 1.5800118344219523e-05, "loss": 1.1103, "step": 9207 }, { "epoch": 0.32446241532105324, "grad_norm": 7.100697994232178, "learning_rate": 1.5799188613179737e-05, "loss": 6.6511, "step": 9208 }, { "epoch": 0.3244976523340116, "grad_norm": 8.591947555541992, "learning_rate": 1.579825880660478e-05, "loss": 8.1718, "step": 9209 }, { "epoch": 0.32453288934697005, "grad_norm": 0.9914296865463257, "learning_rate": 1.5797328924506764e-05, "loss": 0.9011, "step": 9210 }, { "epoch": 0.3245681263599285, "grad_norm": 1.2620775699615479, "learning_rate": 1.57963989668978e-05, "loss": 0.9502, "step": 9211 }, { "epoch": 0.32460336337288687, "grad_norm": 5.838936805725098, "learning_rate": 1.5795468933790003e-05, "loss": 5.7605, "step": 9212 }, { "epoch": 0.3246386003858453, "grad_norm": 2.2683191299438477, "learning_rate": 1.5794538825195482e-05, "loss": 2.5569, "step": 9213 }, { "epoch": 0.3246738373988037, "grad_norm": 3.6082797050476074, "learning_rate": 1.5793608641126355e-05, "loss": 2.9593, "step": 9214 }, { "epoch": 0.3247090744117621, "grad_norm": 1.7134050130844116, "learning_rate": 1.579267838159474e-05, "loss": 1.0591, "step": 9215 }, { "epoch": 0.32474431142472054, "grad_norm": 3.8454067707061768, "learning_rate": 1.5791748046612753e-05, "loss": 3.3092, "step": 9216 }, { "epoch": 0.3247795484376789, "grad_norm": 1.8988875150680542, "learning_rate": 1.5790817636192502e-05, "loss": 0.8334, "step": 9217 }, { "epoch": 0.32481478545063736, "grad_norm": 1.8128395080566406, "learning_rate": 1.578988715034612e-05, "loss": 0.6739, "step": 9218 }, { "epoch": 0.32485002246359573, "grad_norm": 4.270351409912109, "learning_rate": 1.578895658908572e-05, "loss": 5.8779, "step": 9219 }, { "epoch": 0.32488525947655417, "grad_norm": 4.548625946044922, "learning_rate": 1.578802595242342e-05, "loss": 3.5407, "step": 9220 }, { "epoch": 0.3249204964895126, "grad_norm": 1.5944511890411377, "learning_rate": 1.5787095240371345e-05, "loss": 0.8892, "step": 9221 }, { "epoch": 0.324955733502471, "grad_norm": 2.8244521617889404, "learning_rate": 1.5786164452941622e-05, "loss": 2.7977, "step": 9222 }, { "epoch": 0.3249909705154294, "grad_norm": 3.670924425125122, "learning_rate": 1.5785233590146363e-05, "loss": 3.5128, "step": 9223 }, { "epoch": 0.32502620752838784, "grad_norm": 1.7583297491073608, "learning_rate": 1.578430265199771e-05, "loss": 0.8431, "step": 9224 }, { "epoch": 0.3250614445413462, "grad_norm": 1.3460302352905273, "learning_rate": 1.5783371638507767e-05, "loss": 1.0232, "step": 9225 }, { "epoch": 0.32509668155430466, "grad_norm": 6.158087253570557, "learning_rate": 1.5782440549688675e-05, "loss": 6.3501, "step": 9226 }, { "epoch": 0.32513191856726303, "grad_norm": 2.0131044387817383, "learning_rate": 1.578150938555256e-05, "loss": 1.0784, "step": 9227 }, { "epoch": 0.32516715558022147, "grad_norm": 2.1385886669158936, "learning_rate": 1.578057814611155e-05, "loss": 1.1139, "step": 9228 }, { "epoch": 0.3252023925931799, "grad_norm": 4.773736953735352, "learning_rate": 1.577964683137777e-05, "loss": 3.6975, "step": 9229 }, { "epoch": 0.3252376296061383, "grad_norm": 4.501582622528076, "learning_rate": 1.5778715441363356e-05, "loss": 2.8718, "step": 9230 }, { "epoch": 0.3252728666190967, "grad_norm": 3.971256732940674, "learning_rate": 1.5777783976080433e-05, "loss": 3.3337, "step": 9231 }, { "epoch": 0.3253081036320551, "grad_norm": 1.3254600763320923, "learning_rate": 1.5776852435541147e-05, "loss": 0.7288, "step": 9232 }, { "epoch": 0.3253433406450135, "grad_norm": 3.079951286315918, "learning_rate": 1.5775920819757613e-05, "loss": 3.6606, "step": 9233 }, { "epoch": 0.32537857765797196, "grad_norm": 1.871262788772583, "learning_rate": 1.577498912874198e-05, "loss": 1.0559, "step": 9234 }, { "epoch": 0.32541381467093033, "grad_norm": 5.145933628082275, "learning_rate": 1.5774057362506375e-05, "loss": 3.8674, "step": 9235 }, { "epoch": 0.32544905168388877, "grad_norm": 1.1278979778289795, "learning_rate": 1.5773125521062937e-05, "loss": 1.3461, "step": 9236 }, { "epoch": 0.32548428869684715, "grad_norm": 1.4681854248046875, "learning_rate": 1.5772193604423807e-05, "loss": 1.1782, "step": 9237 }, { "epoch": 0.3255195257098056, "grad_norm": 4.628835201263428, "learning_rate": 1.5771261612601118e-05, "loss": 3.5755, "step": 9238 }, { "epoch": 0.325554762722764, "grad_norm": 1.7891249656677246, "learning_rate": 1.5770329545607013e-05, "loss": 1.1081, "step": 9239 }, { "epoch": 0.3255899997357224, "grad_norm": 3.9461045265197754, "learning_rate": 1.5769397403453628e-05, "loss": 3.5648, "step": 9240 }, { "epoch": 0.3256252367486808, "grad_norm": 0.9487196803092957, "learning_rate": 1.5768465186153106e-05, "loss": 1.2412, "step": 9241 }, { "epoch": 0.3256604737616392, "grad_norm": 5.773513317108154, "learning_rate": 1.5767532893717594e-05, "loss": 7.1508, "step": 9242 }, { "epoch": 0.32569571077459764, "grad_norm": 1.4539717435836792, "learning_rate": 1.576660052615923e-05, "loss": 1.218, "step": 9243 }, { "epoch": 0.32573094778755607, "grad_norm": 1.090983271598816, "learning_rate": 1.576566808349016e-05, "loss": 0.8894, "step": 9244 }, { "epoch": 0.32576618480051445, "grad_norm": 6.241567611694336, "learning_rate": 1.576473556572253e-05, "loss": 5.2234, "step": 9245 }, { "epoch": 0.3258014218134729, "grad_norm": 3.825869560241699, "learning_rate": 1.576380297286848e-05, "loss": 0.964, "step": 9246 }, { "epoch": 0.32583665882643126, "grad_norm": 0.9976536631584167, "learning_rate": 1.5762870304940164e-05, "loss": 1.1576, "step": 9247 }, { "epoch": 0.3258718958393897, "grad_norm": 1.6612874269485474, "learning_rate": 1.5761937561949728e-05, "loss": 0.8166, "step": 9248 }, { "epoch": 0.3259071328523481, "grad_norm": 5.413204193115234, "learning_rate": 1.5761004743909327e-05, "loss": 3.3944, "step": 9249 }, { "epoch": 0.3259423698653065, "grad_norm": 2.643688201904297, "learning_rate": 1.5760071850831096e-05, "loss": 3.2017, "step": 9250 }, { "epoch": 0.32597760687826494, "grad_norm": 5.983678340911865, "learning_rate": 1.5759138882727204e-05, "loss": 3.3568, "step": 9251 }, { "epoch": 0.3260128438912233, "grad_norm": 1.4045345783233643, "learning_rate": 1.575820583960979e-05, "loss": 1.0263, "step": 9252 }, { "epoch": 0.32604808090418175, "grad_norm": 2.962353467941284, "learning_rate": 1.5757272721491012e-05, "loss": 3.4249, "step": 9253 }, { "epoch": 0.3260833179171402, "grad_norm": 1.8545846939086914, "learning_rate": 1.5756339528383025e-05, "loss": 0.8625, "step": 9254 }, { "epoch": 0.32611855493009856, "grad_norm": 1.2363032102584839, "learning_rate": 1.575540626029798e-05, "loss": 0.8668, "step": 9255 }, { "epoch": 0.326153791943057, "grad_norm": 5.265605926513672, "learning_rate": 1.5754472917248035e-05, "loss": 3.5489, "step": 9256 }, { "epoch": 0.3261890289560154, "grad_norm": 1.8931646347045898, "learning_rate": 1.575353949924535e-05, "loss": 0.8402, "step": 9257 }, { "epoch": 0.3262242659689738, "grad_norm": 3.052119493484497, "learning_rate": 1.575260600630208e-05, "loss": 3.3538, "step": 9258 }, { "epoch": 0.32625950298193224, "grad_norm": 7.603855133056641, "learning_rate": 1.5751672438430383e-05, "loss": 3.2657, "step": 9259 }, { "epoch": 0.3262947399948906, "grad_norm": 10.097686767578125, "learning_rate": 1.5750738795642417e-05, "loss": 6.2125, "step": 9260 }, { "epoch": 0.32632997700784905, "grad_norm": 1.2115367650985718, "learning_rate": 1.574980507795035e-05, "loss": 1.0385, "step": 9261 }, { "epoch": 0.3263652140208075, "grad_norm": 2.1944022178649902, "learning_rate": 1.5748871285366337e-05, "loss": 0.9675, "step": 9262 }, { "epoch": 0.32640045103376586, "grad_norm": 5.668667793273926, "learning_rate": 1.5747937417902543e-05, "loss": 3.7352, "step": 9263 }, { "epoch": 0.3264356880467243, "grad_norm": 2.2578253746032715, "learning_rate": 1.5747003475571136e-05, "loss": 0.9226, "step": 9264 }, { "epoch": 0.32647092505968267, "grad_norm": 1.6379128694534302, "learning_rate": 1.5746069458384273e-05, "loss": 0.8569, "step": 9265 }, { "epoch": 0.3265061620726411, "grad_norm": 5.471342086791992, "learning_rate": 1.574513536635412e-05, "loss": 3.7668, "step": 9266 }, { "epoch": 0.32654139908559954, "grad_norm": 3.4840075969696045, "learning_rate": 1.5744201199492858e-05, "loss": 3.4516, "step": 9267 }, { "epoch": 0.3265766360985579, "grad_norm": 1.14077627658844, "learning_rate": 1.5743266957812635e-05, "loss": 0.9981, "step": 9268 }, { "epoch": 0.32661187311151635, "grad_norm": 1.0327469110488892, "learning_rate": 1.574233264132563e-05, "loss": 0.9296, "step": 9269 }, { "epoch": 0.3266471101244747, "grad_norm": 1.1919387578964233, "learning_rate": 1.574139825004401e-05, "loss": 0.9924, "step": 9270 }, { "epoch": 0.32668234713743316, "grad_norm": 1.1064273118972778, "learning_rate": 1.574046378397995e-05, "loss": 0.6791, "step": 9271 }, { "epoch": 0.3267175841503916, "grad_norm": 3.180332660675049, "learning_rate": 1.5739529243145612e-05, "loss": 3.3511, "step": 9272 }, { "epoch": 0.32675282116334997, "grad_norm": 6.583210468292236, "learning_rate": 1.5738594627553177e-05, "loss": 3.3732, "step": 9273 }, { "epoch": 0.3267880581763084, "grad_norm": 1.4052023887634277, "learning_rate": 1.573765993721482e-05, "loss": 0.6858, "step": 9274 }, { "epoch": 0.3268232951892668, "grad_norm": 8.072808265686035, "learning_rate": 1.573672517214271e-05, "loss": 2.783, "step": 9275 }, { "epoch": 0.3268585322022252, "grad_norm": 5.145166873931885, "learning_rate": 1.5735790332349022e-05, "loss": 4.1141, "step": 9276 }, { "epoch": 0.32689376921518365, "grad_norm": 1.132563591003418, "learning_rate": 1.573485541784593e-05, "loss": 0.7493, "step": 9277 }, { "epoch": 0.326929006228142, "grad_norm": 1.2494889497756958, "learning_rate": 1.5733920428645624e-05, "loss": 1.1492, "step": 9278 }, { "epoch": 0.32696424324110046, "grad_norm": 2.602435350418091, "learning_rate": 1.5732985364760267e-05, "loss": 2.6387, "step": 9279 }, { "epoch": 0.32699948025405884, "grad_norm": 4.861030101776123, "learning_rate": 1.573205022620205e-05, "loss": 3.5753, "step": 9280 }, { "epoch": 0.32703471726701727, "grad_norm": 2.9630095958709717, "learning_rate": 1.573111501298315e-05, "loss": 1.1125, "step": 9281 }, { "epoch": 0.3270699542799757, "grad_norm": 8.25475788116455, "learning_rate": 1.573017972511574e-05, "loss": 7.6025, "step": 9282 }, { "epoch": 0.3271051912929341, "grad_norm": 6.243851184844971, "learning_rate": 1.5729244362612015e-05, "loss": 6.2033, "step": 9283 }, { "epoch": 0.3271404283058925, "grad_norm": 1.3036384582519531, "learning_rate": 1.572830892548415e-05, "loss": 1.1313, "step": 9284 }, { "epoch": 0.3271756653188509, "grad_norm": 1.2937475442886353, "learning_rate": 1.5727373413744333e-05, "loss": 1.1186, "step": 9285 }, { "epoch": 0.32721090233180933, "grad_norm": 2.9118618965148926, "learning_rate": 1.5726437827404748e-05, "loss": 4.1917, "step": 9286 }, { "epoch": 0.32724613934476776, "grad_norm": 1.4398586750030518, "learning_rate": 1.5725502166477576e-05, "loss": 0.9723, "step": 9287 }, { "epoch": 0.32728137635772614, "grad_norm": 7.095956802368164, "learning_rate": 1.5724566430975018e-05, "loss": 3.9144, "step": 9288 }, { "epoch": 0.3273166133706846, "grad_norm": 2.7429749965667725, "learning_rate": 1.5723630620909247e-05, "loss": 3.5251, "step": 9289 }, { "epoch": 0.327351850383643, "grad_norm": 1.4659225940704346, "learning_rate": 1.5722694736292457e-05, "loss": 1.0041, "step": 9290 }, { "epoch": 0.3273870873966014, "grad_norm": 1.429044246673584, "learning_rate": 1.5721758777136843e-05, "loss": 0.9343, "step": 9291 }, { "epoch": 0.3274223244095598, "grad_norm": 1.3003982305526733, "learning_rate": 1.5720822743454588e-05, "loss": 0.7741, "step": 9292 }, { "epoch": 0.3274575614225182, "grad_norm": 4.650475025177002, "learning_rate": 1.5719886635257892e-05, "loss": 3.9847, "step": 9293 }, { "epoch": 0.32749279843547663, "grad_norm": 1.5441203117370605, "learning_rate": 1.5718950452558943e-05, "loss": 1.204, "step": 9294 }, { "epoch": 0.32752803544843506, "grad_norm": 1.2934434413909912, "learning_rate": 1.5718014195369934e-05, "loss": 0.9774, "step": 9295 }, { "epoch": 0.32756327246139344, "grad_norm": 3.62624192237854, "learning_rate": 1.5717077863703063e-05, "loss": 2.9294, "step": 9296 }, { "epoch": 0.3275985094743519, "grad_norm": 12.05385684967041, "learning_rate": 1.5716141457570528e-05, "loss": 3.2856, "step": 9297 }, { "epoch": 0.32763374648731025, "grad_norm": 4.365399360656738, "learning_rate": 1.5715204976984514e-05, "loss": 3.0088, "step": 9298 }, { "epoch": 0.3276689835002687, "grad_norm": 4.620654582977295, "learning_rate": 1.5714268421957234e-05, "loss": 4.8132, "step": 9299 }, { "epoch": 0.3277042205132271, "grad_norm": 1.7914810180664062, "learning_rate": 1.571333179250088e-05, "loss": 0.8839, "step": 9300 }, { "epoch": 0.3277394575261855, "grad_norm": 5.170566558837891, "learning_rate": 1.5712395088627648e-05, "loss": 3.5543, "step": 9301 }, { "epoch": 0.32777469453914393, "grad_norm": 4.295361518859863, "learning_rate": 1.5711458310349743e-05, "loss": 3.1908, "step": 9302 }, { "epoch": 0.3278099315521023, "grad_norm": 1.5246704816818237, "learning_rate": 1.5710521457679368e-05, "loss": 0.7321, "step": 9303 }, { "epoch": 0.32784516856506074, "grad_norm": 3.396635055541992, "learning_rate": 1.5709584530628722e-05, "loss": 3.6968, "step": 9304 }, { "epoch": 0.3278804055780192, "grad_norm": 2.1027300357818604, "learning_rate": 1.570864752921001e-05, "loss": 0.7895, "step": 9305 }, { "epoch": 0.32791564259097755, "grad_norm": 5.979190826416016, "learning_rate": 1.570771045343544e-05, "loss": 3.804, "step": 9306 }, { "epoch": 0.327950879603936, "grad_norm": 1.1258718967437744, "learning_rate": 1.5706773303317213e-05, "loss": 0.8407, "step": 9307 }, { "epoch": 0.32798611661689436, "grad_norm": 4.7603254318237305, "learning_rate": 1.5705836078867532e-05, "loss": 6.8427, "step": 9308 }, { "epoch": 0.3280213536298528, "grad_norm": 2.0891923904418945, "learning_rate": 1.5704898780098612e-05, "loss": 3.1915, "step": 9309 }, { "epoch": 0.32805659064281123, "grad_norm": 3.2868764400482178, "learning_rate": 1.5703961407022663e-05, "loss": 2.7949, "step": 9310 }, { "epoch": 0.3280918276557696, "grad_norm": 1.6914056539535522, "learning_rate": 1.5703023959651885e-05, "loss": 0.8212, "step": 9311 }, { "epoch": 0.32812706466872804, "grad_norm": 2.201319932937622, "learning_rate": 1.5702086437998492e-05, "loss": 0.8122, "step": 9312 }, { "epoch": 0.3281623016816864, "grad_norm": 1.3287373781204224, "learning_rate": 1.5701148842074703e-05, "loss": 1.0974, "step": 9313 }, { "epoch": 0.32819753869464485, "grad_norm": 3.274932384490967, "learning_rate": 1.570021117189272e-05, "loss": 3.3937, "step": 9314 }, { "epoch": 0.3282327757076033, "grad_norm": 1.4242814779281616, "learning_rate": 1.5699273427464762e-05, "loss": 0.8657, "step": 9315 }, { "epoch": 0.32826801272056166, "grad_norm": 3.3979859352111816, "learning_rate": 1.569833560880304e-05, "loss": 3.8398, "step": 9316 }, { "epoch": 0.3283032497335201, "grad_norm": 2.6939215660095215, "learning_rate": 1.5697397715919774e-05, "loss": 2.8651, "step": 9317 }, { "epoch": 0.3283384867464785, "grad_norm": 5.661588668823242, "learning_rate": 1.5696459748827174e-05, "loss": 3.5234, "step": 9318 }, { "epoch": 0.3283737237594369, "grad_norm": 1.81293785572052, "learning_rate": 1.5695521707537463e-05, "loss": 0.9078, "step": 9319 }, { "epoch": 0.32840896077239534, "grad_norm": 6.226576328277588, "learning_rate": 1.5694583592062854e-05, "loss": 3.6494, "step": 9320 }, { "epoch": 0.3284441977853537, "grad_norm": 1.2605836391448975, "learning_rate": 1.5693645402415566e-05, "loss": 0.9903, "step": 9321 }, { "epoch": 0.32847943479831215, "grad_norm": 0.9537296295166016, "learning_rate": 1.5692707138607827e-05, "loss": 0.7579, "step": 9322 }, { "epoch": 0.32851467181127053, "grad_norm": 1.7657923698425293, "learning_rate": 1.569176880065185e-05, "loss": 1.1975, "step": 9323 }, { "epoch": 0.32854990882422896, "grad_norm": 5.202469825744629, "learning_rate": 1.5690830388559855e-05, "loss": 6.6752, "step": 9324 }, { "epoch": 0.3285851458371874, "grad_norm": 5.7381391525268555, "learning_rate": 1.5689891902344076e-05, "loss": 3.8048, "step": 9325 }, { "epoch": 0.3286203828501458, "grad_norm": 2.2219908237457275, "learning_rate": 1.5688953342016725e-05, "loss": 0.9152, "step": 9326 }, { "epoch": 0.3286556198631042, "grad_norm": 4.347884178161621, "learning_rate": 1.5688014707590034e-05, "loss": 3.9673, "step": 9327 }, { "epoch": 0.32869085687606264, "grad_norm": 4.4331207275390625, "learning_rate": 1.5687075999076228e-05, "loss": 3.9553, "step": 9328 }, { "epoch": 0.328726093889021, "grad_norm": 2.590597629547119, "learning_rate": 1.5686137216487533e-05, "loss": 3.5689, "step": 9329 }, { "epoch": 0.32876133090197945, "grad_norm": 2.4863474369049072, "learning_rate": 1.5685198359836176e-05, "loss": 0.8726, "step": 9330 }, { "epoch": 0.32879656791493783, "grad_norm": 5.381743907928467, "learning_rate": 1.5684259429134385e-05, "loss": 4.1825, "step": 9331 }, { "epoch": 0.32883180492789627, "grad_norm": 1.6865776777267456, "learning_rate": 1.568332042439439e-05, "loss": 0.749, "step": 9332 }, { "epoch": 0.3288670419408547, "grad_norm": 3.899447202682495, "learning_rate": 1.5682381345628425e-05, "loss": 3.3512, "step": 9333 }, { "epoch": 0.3289022789538131, "grad_norm": 3.301560878753662, "learning_rate": 1.5681442192848718e-05, "loss": 3.3534, "step": 9334 }, { "epoch": 0.3289375159667715, "grad_norm": 6.518344879150391, "learning_rate": 1.5680502966067503e-05, "loss": 4.0056, "step": 9335 }, { "epoch": 0.3289727529797299, "grad_norm": 0.9062000513076782, "learning_rate": 1.5679563665297014e-05, "loss": 1.0918, "step": 9336 }, { "epoch": 0.3290079899926883, "grad_norm": 1.022241473197937, "learning_rate": 1.5678624290549484e-05, "loss": 0.9806, "step": 9337 }, { "epoch": 0.32904322700564675, "grad_norm": 3.276245594024658, "learning_rate": 1.567768484183715e-05, "loss": 2.8109, "step": 9338 }, { "epoch": 0.32907846401860513, "grad_norm": 3.6673412322998047, "learning_rate": 1.5676745319172245e-05, "loss": 3.1716, "step": 9339 }, { "epoch": 0.32911370103156357, "grad_norm": 3.2042131423950195, "learning_rate": 1.5675805722567012e-05, "loss": 3.377, "step": 9340 }, { "epoch": 0.32914893804452194, "grad_norm": 1.4863853454589844, "learning_rate": 1.5674866052033686e-05, "loss": 0.7225, "step": 9341 }, { "epoch": 0.3291841750574804, "grad_norm": 1.179189920425415, "learning_rate": 1.5673926307584508e-05, "loss": 1.0157, "step": 9342 }, { "epoch": 0.3292194120704388, "grad_norm": 1.108698844909668, "learning_rate": 1.5672986489231714e-05, "loss": 0.978, "step": 9343 }, { "epoch": 0.3292546490833972, "grad_norm": 1.9857314825057983, "learning_rate": 1.567204659698755e-05, "loss": 0.9881, "step": 9344 }, { "epoch": 0.3292898860963556, "grad_norm": 1.1195391416549683, "learning_rate": 1.5671106630864257e-05, "loss": 0.752, "step": 9345 }, { "epoch": 0.329325123109314, "grad_norm": 2.295287609100342, "learning_rate": 1.5670166590874075e-05, "loss": 3.1027, "step": 9346 }, { "epoch": 0.32936036012227243, "grad_norm": 2.5731489658355713, "learning_rate": 1.5669226477029254e-05, "loss": 1.1119, "step": 9347 }, { "epoch": 0.32939559713523087, "grad_norm": 5.852561950683594, "learning_rate": 1.5668286289342035e-05, "loss": 3.2726, "step": 9348 }, { "epoch": 0.32943083414818924, "grad_norm": 1.3543341159820557, "learning_rate": 1.5667346027824667e-05, "loss": 0.834, "step": 9349 }, { "epoch": 0.3294660711611477, "grad_norm": 2.968061923980713, "learning_rate": 1.5666405692489394e-05, "loss": 3.224, "step": 9350 }, { "epoch": 0.32950130817410606, "grad_norm": 1.153971552848816, "learning_rate": 1.5665465283348464e-05, "loss": 0.9551, "step": 9351 }, { "epoch": 0.3295365451870645, "grad_norm": 4.867724895477295, "learning_rate": 1.5664524800414125e-05, "loss": 3.1604, "step": 9352 }, { "epoch": 0.3295717822000229, "grad_norm": 1.2304191589355469, "learning_rate": 1.566358424369863e-05, "loss": 0.9284, "step": 9353 }, { "epoch": 0.3296070192129813, "grad_norm": 5.203736782073975, "learning_rate": 1.566264361321423e-05, "loss": 6.3054, "step": 9354 }, { "epoch": 0.32964225622593973, "grad_norm": 2.803771734237671, "learning_rate": 1.566170290897318e-05, "loss": 3.2026, "step": 9355 }, { "epoch": 0.3296774932388981, "grad_norm": 7.629300594329834, "learning_rate": 1.5660762130987723e-05, "loss": 4.1065, "step": 9356 }, { "epoch": 0.32971273025185655, "grad_norm": 4.064550876617432, "learning_rate": 1.5659821279270116e-05, "loss": 2.3371, "step": 9357 }, { "epoch": 0.329747967264815, "grad_norm": 3.34372615814209, "learning_rate": 1.565888035383262e-05, "loss": 3.2644, "step": 9358 }, { "epoch": 0.32978320427777336, "grad_norm": 0.853990912437439, "learning_rate": 1.565793935468749e-05, "loss": 0.6176, "step": 9359 }, { "epoch": 0.3298184412907318, "grad_norm": 1.2236219644546509, "learning_rate": 1.565699828184697e-05, "loss": 0.6973, "step": 9360 }, { "epoch": 0.3298536783036902, "grad_norm": 1.4964392185211182, "learning_rate": 1.5656057135323335e-05, "loss": 1.2374, "step": 9361 }, { "epoch": 0.3298889153166486, "grad_norm": 3.3767788410186768, "learning_rate": 1.5655115915128833e-05, "loss": 2.9491, "step": 9362 }, { "epoch": 0.32992415232960703, "grad_norm": 3.513474464416504, "learning_rate": 1.5654174621275725e-05, "loss": 3.3815, "step": 9363 }, { "epoch": 0.3299593893425654, "grad_norm": 1.406630516052246, "learning_rate": 1.565323325377627e-05, "loss": 0.871, "step": 9364 }, { "epoch": 0.32999462635552385, "grad_norm": 3.309246063232422, "learning_rate": 1.5652291812642733e-05, "loss": 2.8493, "step": 9365 }, { "epoch": 0.3300298633684823, "grad_norm": 1.474239706993103, "learning_rate": 1.5651350297887376e-05, "loss": 1.0856, "step": 9366 }, { "epoch": 0.33006510038144066, "grad_norm": 4.203390121459961, "learning_rate": 1.565040870952246e-05, "loss": 3.1207, "step": 9367 }, { "epoch": 0.3301003373943991, "grad_norm": 2.128690004348755, "learning_rate": 1.5649467047560256e-05, "loss": 1.0063, "step": 9368 }, { "epoch": 0.33013557440735747, "grad_norm": 5.891946792602539, "learning_rate": 1.564852531201302e-05, "loss": 3.4688, "step": 9369 }, { "epoch": 0.3301708114203159, "grad_norm": 3.7906224727630615, "learning_rate": 1.5647583502893025e-05, "loss": 2.999, "step": 9370 }, { "epoch": 0.33020604843327434, "grad_norm": 1.513727068901062, "learning_rate": 1.564664162021253e-05, "loss": 0.8911, "step": 9371 }, { "epoch": 0.3302412854462327, "grad_norm": 1.9012116193771362, "learning_rate": 1.5645699663983814e-05, "loss": 0.7853, "step": 9372 }, { "epoch": 0.33027652245919115, "grad_norm": 1.7292840480804443, "learning_rate": 1.564475763421914e-05, "loss": 0.9529, "step": 9373 }, { "epoch": 0.3303117594721495, "grad_norm": 4.102146625518799, "learning_rate": 1.5643815530930773e-05, "loss": 5.7414, "step": 9374 }, { "epoch": 0.33034699648510796, "grad_norm": 2.921182870864868, "learning_rate": 1.5642873354130996e-05, "loss": 3.1342, "step": 9375 }, { "epoch": 0.3303822334980664, "grad_norm": 1.6222295761108398, "learning_rate": 1.5641931103832075e-05, "loss": 0.7666, "step": 9376 }, { "epoch": 0.33041747051102477, "grad_norm": 5.876408100128174, "learning_rate": 1.5640988780046277e-05, "loss": 5.4091, "step": 9377 }, { "epoch": 0.3304527075239832, "grad_norm": 7.828518867492676, "learning_rate": 1.564004638278589e-05, "loss": 3.8806, "step": 9378 }, { "epoch": 0.3304879445369416, "grad_norm": 1.017080545425415, "learning_rate": 1.5639103912063172e-05, "loss": 1.091, "step": 9379 }, { "epoch": 0.3305231815499, "grad_norm": 1.5268473625183105, "learning_rate": 1.5638161367890413e-05, "loss": 1.0034, "step": 9380 }, { "epoch": 0.33055841856285845, "grad_norm": 1.8083019256591797, "learning_rate": 1.563721875027988e-05, "loss": 0.7106, "step": 9381 }, { "epoch": 0.3305936555758168, "grad_norm": 4.9573564529418945, "learning_rate": 1.563627605924386e-05, "loss": 2.3522, "step": 9382 }, { "epoch": 0.33062889258877526, "grad_norm": 3.4457931518554688, "learning_rate": 1.563533329479462e-05, "loss": 3.5871, "step": 9383 }, { "epoch": 0.33066412960173364, "grad_norm": 3.354886770248413, "learning_rate": 1.563439045694445e-05, "loss": 3.0529, "step": 9384 }, { "epoch": 0.33069936661469207, "grad_norm": 4.082132339477539, "learning_rate": 1.5633447545705625e-05, "loss": 3.6972, "step": 9385 }, { "epoch": 0.3307346036276505, "grad_norm": 4.063767910003662, "learning_rate": 1.563250456109043e-05, "loss": 3.5432, "step": 9386 }, { "epoch": 0.3307698406406089, "grad_norm": 1.6329952478408813, "learning_rate": 1.5631561503111144e-05, "loss": 1.0645, "step": 9387 }, { "epoch": 0.3308050776535673, "grad_norm": 1.91623055934906, "learning_rate": 1.5630618371780055e-05, "loss": 0.81, "step": 9388 }, { "epoch": 0.3308403146665257, "grad_norm": 3.3311705589294434, "learning_rate": 1.5629675167109438e-05, "loss": 3.7424, "step": 9389 }, { "epoch": 0.3308755516794841, "grad_norm": 1.282901406288147, "learning_rate": 1.562873188911159e-05, "loss": 0.8374, "step": 9390 }, { "epoch": 0.33091078869244256, "grad_norm": 1.3821226358413696, "learning_rate": 1.5627788537798793e-05, "loss": 1.1512, "step": 9391 }, { "epoch": 0.33094602570540094, "grad_norm": 1.36204993724823, "learning_rate": 1.562684511318333e-05, "loss": 0.823, "step": 9392 }, { "epoch": 0.33098126271835937, "grad_norm": 3.807579517364502, "learning_rate": 1.5625901615277494e-05, "loss": 3.7049, "step": 9393 }, { "epoch": 0.3310164997313178, "grad_norm": 4.536350250244141, "learning_rate": 1.5624958044093573e-05, "loss": 3.3618, "step": 9394 }, { "epoch": 0.3310517367442762, "grad_norm": 4.9400835037231445, "learning_rate": 1.562401439964386e-05, "loss": 3.132, "step": 9395 }, { "epoch": 0.3310869737572346, "grad_norm": 1.2168848514556885, "learning_rate": 1.5623070681940636e-05, "loss": 1.2326, "step": 9396 }, { "epoch": 0.331122210770193, "grad_norm": 3.404538869857788, "learning_rate": 1.562212689099621e-05, "loss": 3.1979, "step": 9397 }, { "epoch": 0.3311574477831514, "grad_norm": 1.0127966403961182, "learning_rate": 1.5621183026822856e-05, "loss": 0.9614, "step": 9398 }, { "epoch": 0.33119268479610986, "grad_norm": 1.8576682806015015, "learning_rate": 1.562023908943288e-05, "loss": 0.968, "step": 9399 }, { "epoch": 0.33122792180906824, "grad_norm": 6.4743452072143555, "learning_rate": 1.5619295078838575e-05, "loss": 6.0325, "step": 9400 }, { "epoch": 0.33126315882202667, "grad_norm": 5.783538341522217, "learning_rate": 1.5618350995052235e-05, "loss": 3.4279, "step": 9401 }, { "epoch": 0.33129839583498505, "grad_norm": 3.6129770278930664, "learning_rate": 1.5617406838086155e-05, "loss": 3.1857, "step": 9402 }, { "epoch": 0.3313336328479435, "grad_norm": 6.019942283630371, "learning_rate": 1.561646260795264e-05, "loss": 3.7479, "step": 9403 }, { "epoch": 0.3313688698609019, "grad_norm": 1.6375632286071777, "learning_rate": 1.561551830466398e-05, "loss": 1.1115, "step": 9404 }, { "epoch": 0.3314041068738603, "grad_norm": 6.5191216468811035, "learning_rate": 1.561457392823248e-05, "loss": 3.4005, "step": 9405 }, { "epoch": 0.3314393438868187, "grad_norm": 5.573197841644287, "learning_rate": 1.5613629478670443e-05, "loss": 4.7777, "step": 9406 }, { "epoch": 0.3314745808997771, "grad_norm": 4.756903648376465, "learning_rate": 1.5612684955990163e-05, "loss": 4.1481, "step": 9407 }, { "epoch": 0.33150981791273554, "grad_norm": 5.6294450759887695, "learning_rate": 1.561174036020395e-05, "loss": 3.3574, "step": 9408 }, { "epoch": 0.33154505492569397, "grad_norm": 3.183513641357422, "learning_rate": 1.56107956913241e-05, "loss": 3.1891, "step": 9409 }, { "epoch": 0.33158029193865235, "grad_norm": 2.8692538738250732, "learning_rate": 1.5609850949362924e-05, "loss": 3.2868, "step": 9410 }, { "epoch": 0.3316155289516108, "grad_norm": 4.203895092010498, "learning_rate": 1.5608906134332726e-05, "loss": 5.3844, "step": 9411 }, { "epoch": 0.33165076596456916, "grad_norm": 1.0750497579574585, "learning_rate": 1.5607961246245807e-05, "loss": 0.8095, "step": 9412 }, { "epoch": 0.3316860029775276, "grad_norm": 2.1655235290527344, "learning_rate": 1.560701628511448e-05, "loss": 0.8953, "step": 9413 }, { "epoch": 0.331721239990486, "grad_norm": 1.865217685699463, "learning_rate": 1.560607125095105e-05, "loss": 0.8573, "step": 9414 }, { "epoch": 0.3317564770034444, "grad_norm": 3.6218628883361816, "learning_rate": 1.5605126143767832e-05, "loss": 3.2959, "step": 9415 }, { "epoch": 0.33179171401640284, "grad_norm": 3.8178975582122803, "learning_rate": 1.5604180963577133e-05, "loss": 3.5218, "step": 9416 }, { "epoch": 0.3318269510293612, "grad_norm": 1.518554449081421, "learning_rate": 1.5603235710391258e-05, "loss": 0.7103, "step": 9417 }, { "epoch": 0.33186218804231965, "grad_norm": 5.0095415115356445, "learning_rate": 1.5602290384222528e-05, "loss": 3.6546, "step": 9418 }, { "epoch": 0.3318974250552781, "grad_norm": 4.233001232147217, "learning_rate": 1.560134498508325e-05, "loss": 3.6283, "step": 9419 }, { "epoch": 0.33193266206823646, "grad_norm": 1.9863063097000122, "learning_rate": 1.560039951298574e-05, "loss": 0.9499, "step": 9420 }, { "epoch": 0.3319678990811949, "grad_norm": 1.7107977867126465, "learning_rate": 1.5599453967942313e-05, "loss": 1.0049, "step": 9421 }, { "epoch": 0.3320031360941533, "grad_norm": 4.075400352478027, "learning_rate": 1.5598508349965288e-05, "loss": 3.8611, "step": 9422 }, { "epoch": 0.3320383731071117, "grad_norm": 3.082260847091675, "learning_rate": 1.5597562659066972e-05, "loss": 3.1312, "step": 9423 }, { "epoch": 0.33207361012007014, "grad_norm": 3.7270922660827637, "learning_rate": 1.5596616895259696e-05, "loss": 3.2835, "step": 9424 }, { "epoch": 0.3321088471330285, "grad_norm": 1.783645749092102, "learning_rate": 1.5595671058555768e-05, "loss": 1.0433, "step": 9425 }, { "epoch": 0.33214408414598695, "grad_norm": 4.674806594848633, "learning_rate": 1.559472514896751e-05, "loss": 4.067, "step": 9426 }, { "epoch": 0.33217932115894533, "grad_norm": 2.952479124069214, "learning_rate": 1.5593779166507252e-05, "loss": 3.8511, "step": 9427 }, { "epoch": 0.33221455817190376, "grad_norm": 2.086858034133911, "learning_rate": 1.5592833111187302e-05, "loss": 0.7942, "step": 9428 }, { "epoch": 0.3322497951848622, "grad_norm": 5.676723480224609, "learning_rate": 1.5591886983019988e-05, "loss": 3.3204, "step": 9429 }, { "epoch": 0.3322850321978206, "grad_norm": 1.4358842372894287, "learning_rate": 1.5590940782017635e-05, "loss": 0.9666, "step": 9430 }, { "epoch": 0.332320269210779, "grad_norm": 3.702437400817871, "learning_rate": 1.5589994508192567e-05, "loss": 3.6391, "step": 9431 }, { "epoch": 0.33235550622373744, "grad_norm": 1.2727872133255005, "learning_rate": 1.558904816155711e-05, "loss": 0.8375, "step": 9432 }, { "epoch": 0.3323907432366958, "grad_norm": 1.0838712453842163, "learning_rate": 1.5588101742123583e-05, "loss": 1.0386, "step": 9433 }, { "epoch": 0.33242598024965425, "grad_norm": 5.331846237182617, "learning_rate": 1.5587155249904324e-05, "loss": 7.8524, "step": 9434 }, { "epoch": 0.33246121726261263, "grad_norm": 1.3053332567214966, "learning_rate": 1.5586208684911656e-05, "loss": 0.687, "step": 9435 }, { "epoch": 0.33249645427557106, "grad_norm": 5.161812782287598, "learning_rate": 1.5585262047157906e-05, "loss": 5.5133, "step": 9436 }, { "epoch": 0.3325316912885295, "grad_norm": 3.4323112964630127, "learning_rate": 1.5584315336655407e-05, "loss": 3.0664, "step": 9437 }, { "epoch": 0.3325669283014879, "grad_norm": 1.5342459678649902, "learning_rate": 1.558336855341649e-05, "loss": 0.7736, "step": 9438 }, { "epoch": 0.3326021653144463, "grad_norm": 1.3081930875778198, "learning_rate": 1.5582421697453484e-05, "loss": 1.1957, "step": 9439 }, { "epoch": 0.3326374023274047, "grad_norm": 1.3178189992904663, "learning_rate": 1.558147476877873e-05, "loss": 0.7733, "step": 9440 }, { "epoch": 0.3326726393403631, "grad_norm": 1.5964394807815552, "learning_rate": 1.558052776740455e-05, "loss": 0.9839, "step": 9441 }, { "epoch": 0.33270787635332155, "grad_norm": 1.4727439880371094, "learning_rate": 1.557958069334329e-05, "loss": 0.8955, "step": 9442 }, { "epoch": 0.33274311336627993, "grad_norm": 7.495316505432129, "learning_rate": 1.557863354660728e-05, "loss": 3.5518, "step": 9443 }, { "epoch": 0.33277835037923836, "grad_norm": 2.3001532554626465, "learning_rate": 1.5577686327208855e-05, "loss": 0.6941, "step": 9444 }, { "epoch": 0.33281358739219674, "grad_norm": 5.6880974769592285, "learning_rate": 1.5576739035160354e-05, "loss": 2.9708, "step": 9445 }, { "epoch": 0.3328488244051552, "grad_norm": 5.722536563873291, "learning_rate": 1.557579167047412e-05, "loss": 3.3411, "step": 9446 }, { "epoch": 0.3328840614181136, "grad_norm": 3.4210855960845947, "learning_rate": 1.5574844233162487e-05, "loss": 3.3113, "step": 9447 }, { "epoch": 0.332919298431072, "grad_norm": 8.823991775512695, "learning_rate": 1.55738967232378e-05, "loss": 3.6514, "step": 9448 }, { "epoch": 0.3329545354440304, "grad_norm": 1.5532886981964111, "learning_rate": 1.5572949140712395e-05, "loss": 0.9132, "step": 9449 }, { "epoch": 0.3329897724569888, "grad_norm": 1.1147701740264893, "learning_rate": 1.557200148559862e-05, "loss": 0.9521, "step": 9450 }, { "epoch": 0.33302500946994723, "grad_norm": 2.47766375541687, "learning_rate": 1.5571053757908817e-05, "loss": 3.1439, "step": 9451 }, { "epoch": 0.33306024648290566, "grad_norm": 6.15176248550415, "learning_rate": 1.5570105957655327e-05, "loss": 5.8751, "step": 9452 }, { "epoch": 0.33309548349586404, "grad_norm": 1.0425288677215576, "learning_rate": 1.5569158084850497e-05, "loss": 0.9398, "step": 9453 }, { "epoch": 0.3331307205088225, "grad_norm": 1.8640151023864746, "learning_rate": 1.5568210139506674e-05, "loss": 0.9303, "step": 9454 }, { "epoch": 0.33316595752178085, "grad_norm": 1.2627205848693848, "learning_rate": 1.5567262121636203e-05, "loss": 0.7464, "step": 9455 }, { "epoch": 0.3332011945347393, "grad_norm": 2.671405792236328, "learning_rate": 1.556631403125144e-05, "loss": 3.2913, "step": 9456 }, { "epoch": 0.3332364315476977, "grad_norm": 2.0964608192443848, "learning_rate": 1.5565365868364723e-05, "loss": 1.0149, "step": 9457 }, { "epoch": 0.3332716685606561, "grad_norm": 5.470121383666992, "learning_rate": 1.5564417632988405e-05, "loss": 6.2841, "step": 9458 }, { "epoch": 0.33330690557361453, "grad_norm": 5.678098201751709, "learning_rate": 1.556346932513484e-05, "loss": 3.6444, "step": 9459 }, { "epoch": 0.3333421425865729, "grad_norm": 1.75932776927948, "learning_rate": 1.556252094481638e-05, "loss": 0.8402, "step": 9460 }, { "epoch": 0.33337737959953134, "grad_norm": 1.1464544534683228, "learning_rate": 1.5561572492045373e-05, "loss": 0.8553, "step": 9461 }, { "epoch": 0.3334126166124898, "grad_norm": 5.120190143585205, "learning_rate": 1.556062396683418e-05, "loss": 3.1723, "step": 9462 }, { "epoch": 0.33344785362544815, "grad_norm": 7.99655294418335, "learning_rate": 1.5559675369195147e-05, "loss": 3.2688, "step": 9463 }, { "epoch": 0.3334830906384066, "grad_norm": 1.1675877571105957, "learning_rate": 1.5558726699140637e-05, "loss": 1.0106, "step": 9464 }, { "epoch": 0.333518327651365, "grad_norm": 1.69441556930542, "learning_rate": 1.5557777956683008e-05, "loss": 0.9056, "step": 9465 }, { "epoch": 0.3335535646643234, "grad_norm": 2.856668710708618, "learning_rate": 1.5556829141834606e-05, "loss": 0.9855, "step": 9466 }, { "epoch": 0.33358880167728183, "grad_norm": 1.7741304636001587, "learning_rate": 1.55558802546078e-05, "loss": 0.8724, "step": 9467 }, { "epoch": 0.3336240386902402, "grad_norm": 3.9368414878845215, "learning_rate": 1.5554931295014942e-05, "loss": 2.9323, "step": 9468 }, { "epoch": 0.33365927570319864, "grad_norm": 1.9204803705215454, "learning_rate": 1.5553982263068404e-05, "loss": 1.0693, "step": 9469 }, { "epoch": 0.3336945127161571, "grad_norm": 2.2691798210144043, "learning_rate": 1.5553033158780533e-05, "loss": 0.8478, "step": 9470 }, { "epoch": 0.33372974972911545, "grad_norm": 5.298956394195557, "learning_rate": 1.5552083982163698e-05, "loss": 2.7858, "step": 9471 }, { "epoch": 0.3337649867420739, "grad_norm": 1.6341626644134521, "learning_rate": 1.5551134733230263e-05, "loss": 0.9584, "step": 9472 }, { "epoch": 0.33380022375503227, "grad_norm": 9.39140510559082, "learning_rate": 1.555018541199259e-05, "loss": 8.219, "step": 9473 }, { "epoch": 0.3338354607679907, "grad_norm": 6.2117815017700195, "learning_rate": 1.554923601846305e-05, "loss": 5.6688, "step": 9474 }, { "epoch": 0.33387069778094913, "grad_norm": 8.142960548400879, "learning_rate": 1.5548286552654e-05, "loss": 6.4842, "step": 9475 }, { "epoch": 0.3339059347939075, "grad_norm": 1.6172113418579102, "learning_rate": 1.5547337014577814e-05, "loss": 1.0211, "step": 9476 }, { "epoch": 0.33394117180686594, "grad_norm": 7.230080604553223, "learning_rate": 1.5546387404246852e-05, "loss": 3.8118, "step": 9477 }, { "epoch": 0.3339764088198243, "grad_norm": 5.264935493469238, "learning_rate": 1.554543772167349e-05, "loss": 5.7123, "step": 9478 }, { "epoch": 0.33401164583278276, "grad_norm": 0.9072748422622681, "learning_rate": 1.5544487966870096e-05, "loss": 1.2719, "step": 9479 }, { "epoch": 0.3340468828457412, "grad_norm": 2.8326852321624756, "learning_rate": 1.5543538139849038e-05, "loss": 1.057, "step": 9480 }, { "epoch": 0.33408211985869957, "grad_norm": 4.246123790740967, "learning_rate": 1.554258824062269e-05, "loss": 3.0057, "step": 9481 }, { "epoch": 0.334117356871658, "grad_norm": 5.070047855377197, "learning_rate": 1.5541638269203426e-05, "loss": 3.377, "step": 9482 }, { "epoch": 0.3341525938846164, "grad_norm": 0.9268337488174438, "learning_rate": 1.5540688225603616e-05, "loss": 0.6452, "step": 9483 }, { "epoch": 0.3341878308975748, "grad_norm": 4.73764705657959, "learning_rate": 1.5539738109835637e-05, "loss": 3.5308, "step": 9484 }, { "epoch": 0.33422306791053324, "grad_norm": 7.885714530944824, "learning_rate": 1.553878792191186e-05, "loss": 6.29, "step": 9485 }, { "epoch": 0.3342583049234916, "grad_norm": 1.515176773071289, "learning_rate": 1.5537837661844674e-05, "loss": 1.0279, "step": 9486 }, { "epoch": 0.33429354193645006, "grad_norm": 2.0784363746643066, "learning_rate": 1.553688732964644e-05, "loss": 1.1062, "step": 9487 }, { "epoch": 0.33432877894940843, "grad_norm": 3.7610955238342285, "learning_rate": 1.5535936925329546e-05, "loss": 3.5363, "step": 9488 }, { "epoch": 0.33436401596236687, "grad_norm": 1.257951021194458, "learning_rate": 1.5534986448906365e-05, "loss": 1.0797, "step": 9489 }, { "epoch": 0.3343992529753253, "grad_norm": 1.15724515914917, "learning_rate": 1.553403590038928e-05, "loss": 1.1091, "step": 9490 }, { "epoch": 0.3344344899882837, "grad_norm": 3.6708197593688965, "learning_rate": 1.5533085279790677e-05, "loss": 3.2014, "step": 9491 }, { "epoch": 0.3344697270012421, "grad_norm": 1.2597217559814453, "learning_rate": 1.553213458712293e-05, "loss": 0.7168, "step": 9492 }, { "epoch": 0.3345049640142005, "grad_norm": 6.68149471282959, "learning_rate": 1.5531183822398425e-05, "loss": 3.3157, "step": 9493 }, { "epoch": 0.3345402010271589, "grad_norm": 5.7971062660217285, "learning_rate": 1.5530232985629547e-05, "loss": 5.7777, "step": 9494 }, { "epoch": 0.33457543804011736, "grad_norm": 6.5570220947265625, "learning_rate": 1.552928207682868e-05, "loss": 5.4037, "step": 9495 }, { "epoch": 0.33461067505307573, "grad_norm": 1.512393593788147, "learning_rate": 1.552833109600821e-05, "loss": 0.9051, "step": 9496 }, { "epoch": 0.33464591206603417, "grad_norm": 4.21790075302124, "learning_rate": 1.5527380043180525e-05, "loss": 3.7375, "step": 9497 }, { "epoch": 0.3346811490789926, "grad_norm": 8.747132301330566, "learning_rate": 1.5526428918358008e-05, "loss": 4.026, "step": 9498 }, { "epoch": 0.334716386091951, "grad_norm": 14.260889053344727, "learning_rate": 1.5525477721553052e-05, "loss": 6.6502, "step": 9499 }, { "epoch": 0.3347516231049094, "grad_norm": 2.667747974395752, "learning_rate": 1.5524526452778045e-05, "loss": 3.0337, "step": 9500 }, { "epoch": 0.3347868601178678, "grad_norm": 3.001512050628662, "learning_rate": 1.5523575112045376e-05, "loss": 3.287, "step": 9501 }, { "epoch": 0.3348220971308262, "grad_norm": 5.886066436767578, "learning_rate": 1.552262369936744e-05, "loss": 3.5892, "step": 9502 }, { "epoch": 0.33485733414378466, "grad_norm": 1.142795205116272, "learning_rate": 1.5521672214756623e-05, "loss": 0.8541, "step": 9503 }, { "epoch": 0.33489257115674304, "grad_norm": 5.37729024887085, "learning_rate": 1.5520720658225322e-05, "loss": 3.4548, "step": 9504 }, { "epoch": 0.33492780816970147, "grad_norm": 1.0240354537963867, "learning_rate": 1.5519769029785936e-05, "loss": 1.2086, "step": 9505 }, { "epoch": 0.33496304518265985, "grad_norm": 3.470104217529297, "learning_rate": 1.551881732945085e-05, "loss": 3.2547, "step": 9506 }, { "epoch": 0.3349982821956183, "grad_norm": 5.867452144622803, "learning_rate": 1.551786555723247e-05, "loss": 5.6237, "step": 9507 }, { "epoch": 0.3350335192085767, "grad_norm": 2.3563973903656006, "learning_rate": 1.5516913713143184e-05, "loss": 3.3453, "step": 9508 }, { "epoch": 0.3350687562215351, "grad_norm": 5.20000696182251, "learning_rate": 1.5515961797195397e-05, "loss": 3.4037, "step": 9509 }, { "epoch": 0.3351039932344935, "grad_norm": 4.300683498382568, "learning_rate": 1.5515009809401504e-05, "loss": 5.8298, "step": 9510 }, { "epoch": 0.3351392302474519, "grad_norm": 0.9855855107307434, "learning_rate": 1.5514057749773907e-05, "loss": 0.9382, "step": 9511 }, { "epoch": 0.33517446726041034, "grad_norm": 1.1044294834136963, "learning_rate": 1.5513105618325004e-05, "loss": 0.8589, "step": 9512 }, { "epoch": 0.33520970427336877, "grad_norm": 0.950356125831604, "learning_rate": 1.55121534150672e-05, "loss": 0.7807, "step": 9513 }, { "epoch": 0.33524494128632715, "grad_norm": 6.19652795791626, "learning_rate": 1.5511201140012894e-05, "loss": 4.68, "step": 9514 }, { "epoch": 0.3352801782992856, "grad_norm": 10.144267082214355, "learning_rate": 1.551024879317449e-05, "loss": 3.3979, "step": 9515 }, { "epoch": 0.33531541531224396, "grad_norm": 2.686648368835449, "learning_rate": 1.5509296374564393e-05, "loss": 2.9102, "step": 9516 }, { "epoch": 0.3353506523252024, "grad_norm": 1.87198007106781, "learning_rate": 1.5508343884195013e-05, "loss": 2.644, "step": 9517 }, { "epoch": 0.3353858893381608, "grad_norm": 3.5175347328186035, "learning_rate": 1.5507391322078747e-05, "loss": 3.9853, "step": 9518 }, { "epoch": 0.3354211263511192, "grad_norm": 4.449953556060791, "learning_rate": 1.5506438688228014e-05, "loss": 3.3759, "step": 9519 }, { "epoch": 0.33545636336407764, "grad_norm": 7.866227626800537, "learning_rate": 1.550548598265521e-05, "loss": 3.4412, "step": 9520 }, { "epoch": 0.335491600377036, "grad_norm": 3.518923282623291, "learning_rate": 1.550453320537275e-05, "loss": 3.1945, "step": 9521 }, { "epoch": 0.33552683738999445, "grad_norm": 2.962270736694336, "learning_rate": 1.550358035639305e-05, "loss": 3.0068, "step": 9522 }, { "epoch": 0.3355620744029529, "grad_norm": 7.920713424682617, "learning_rate": 1.550262743572851e-05, "loss": 3.836, "step": 9523 }, { "epoch": 0.33559731141591126, "grad_norm": 1.1946860551834106, "learning_rate": 1.5501674443391546e-05, "loss": 1.3082, "step": 9524 }, { "epoch": 0.3356325484288697, "grad_norm": 1.1149396896362305, "learning_rate": 1.550072137939457e-05, "loss": 0.9335, "step": 9525 }, { "epoch": 0.33566778544182807, "grad_norm": 1.30246102809906, "learning_rate": 1.5499768243750003e-05, "loss": 0.9121, "step": 9526 }, { "epoch": 0.3357030224547865, "grad_norm": 1.5558053255081177, "learning_rate": 1.5498815036470253e-05, "loss": 0.8885, "step": 9527 }, { "epoch": 0.33573825946774494, "grad_norm": 1.377462387084961, "learning_rate": 1.5497861757567734e-05, "loss": 0.9384, "step": 9528 }, { "epoch": 0.3357734964807033, "grad_norm": 4.6814093589782715, "learning_rate": 1.549690840705487e-05, "loss": 3.5612, "step": 9529 }, { "epoch": 0.33580873349366175, "grad_norm": 8.559109687805176, "learning_rate": 1.5495954984944066e-05, "loss": 3.5901, "step": 9530 }, { "epoch": 0.3358439705066201, "grad_norm": 4.369181156158447, "learning_rate": 1.549500149124775e-05, "loss": 2.9878, "step": 9531 }, { "epoch": 0.33587920751957856, "grad_norm": 3.5729153156280518, "learning_rate": 1.5494047925978345e-05, "loss": 4.0208, "step": 9532 }, { "epoch": 0.335914444532537, "grad_norm": 1.2148535251617432, "learning_rate": 1.5493094289148262e-05, "loss": 1.0031, "step": 9533 }, { "epoch": 0.33594968154549537, "grad_norm": 3.4842305183410645, "learning_rate": 1.549214058076993e-05, "loss": 3.0768, "step": 9534 }, { "epoch": 0.3359849185584538, "grad_norm": 1.5940985679626465, "learning_rate": 1.549118680085576e-05, "loss": 1.1336, "step": 9535 }, { "epoch": 0.33602015557141224, "grad_norm": 0.8670622110366821, "learning_rate": 1.5490232949418187e-05, "loss": 0.7418, "step": 9536 }, { "epoch": 0.3360553925843706, "grad_norm": 2.7904410362243652, "learning_rate": 1.548927902646963e-05, "loss": 0.9811, "step": 9537 }, { "epoch": 0.33609062959732905, "grad_norm": 2.456695556640625, "learning_rate": 1.5488325032022518e-05, "loss": 0.7993, "step": 9538 }, { "epoch": 0.3361258666102874, "grad_norm": 5.006896018981934, "learning_rate": 1.5487370966089267e-05, "loss": 4.1774, "step": 9539 }, { "epoch": 0.33616110362324586, "grad_norm": 4.408252239227295, "learning_rate": 1.5486416828682316e-05, "loss": 3.2527, "step": 9540 }, { "epoch": 0.3361963406362043, "grad_norm": 1.1360570192337036, "learning_rate": 1.5485462619814084e-05, "loss": 0.7712, "step": 9541 }, { "epoch": 0.33623157764916267, "grad_norm": 1.5187022686004639, "learning_rate": 1.5484508339497e-05, "loss": 0.8124, "step": 9542 }, { "epoch": 0.3362668146621211, "grad_norm": 3.5161521434783936, "learning_rate": 1.5483553987743504e-05, "loss": 4.65, "step": 9543 }, { "epoch": 0.3363020516750795, "grad_norm": 4.650291919708252, "learning_rate": 1.548259956456601e-05, "loss": 3.2567, "step": 9544 }, { "epoch": 0.3363372886880379, "grad_norm": 4.835629940032959, "learning_rate": 1.5481645069976964e-05, "loss": 2.7737, "step": 9545 }, { "epoch": 0.33637252570099635, "grad_norm": 4.408222198486328, "learning_rate": 1.5480690503988793e-05, "loss": 3.0995, "step": 9546 }, { "epoch": 0.3364077627139547, "grad_norm": 6.3217926025390625, "learning_rate": 1.547973586661393e-05, "loss": 3.3409, "step": 9547 }, { "epoch": 0.33644299972691316, "grad_norm": 1.3175791501998901, "learning_rate": 1.5478781157864803e-05, "loss": 0.8258, "step": 9548 }, { "epoch": 0.33647823673987154, "grad_norm": 4.185262680053711, "learning_rate": 1.547782637775386e-05, "loss": 3.6356, "step": 9549 }, { "epoch": 0.33651347375283, "grad_norm": 3.783184766769409, "learning_rate": 1.5476871526293525e-05, "loss": 3.2017, "step": 9550 }, { "epoch": 0.3365487107657884, "grad_norm": 8.889562606811523, "learning_rate": 1.5475916603496246e-05, "loss": 5.3027, "step": 9551 }, { "epoch": 0.3365839477787468, "grad_norm": 2.1201119422912598, "learning_rate": 1.5474961609374456e-05, "loss": 0.7754, "step": 9552 }, { "epoch": 0.3366191847917052, "grad_norm": 3.198390483856201, "learning_rate": 1.547400654394059e-05, "loss": 3.1448, "step": 9553 }, { "epoch": 0.3366544218046636, "grad_norm": 5.66887903213501, "learning_rate": 1.5473051407207093e-05, "loss": 3.5858, "step": 9554 }, { "epoch": 0.33668965881762203, "grad_norm": 1.2183657884597778, "learning_rate": 1.5472096199186407e-05, "loss": 0.9012, "step": 9555 }, { "epoch": 0.33672489583058046, "grad_norm": 1.3728384971618652, "learning_rate": 1.5471140919890965e-05, "loss": 1.0932, "step": 9556 }, { "epoch": 0.33676013284353884, "grad_norm": 1.5197123289108276, "learning_rate": 1.547018556933322e-05, "loss": 0.9686, "step": 9557 }, { "epoch": 0.3367953698564973, "grad_norm": 1.3020604848861694, "learning_rate": 1.546923014752561e-05, "loss": 1.064, "step": 9558 }, { "epoch": 0.33683060686945565, "grad_norm": 5.536281585693359, "learning_rate": 1.546827465448058e-05, "loss": 5.6059, "step": 9559 }, { "epoch": 0.3368658438824141, "grad_norm": 3.08074688911438, "learning_rate": 1.5467319090210576e-05, "loss": 2.6726, "step": 9560 }, { "epoch": 0.3369010808953725, "grad_norm": 1.2812831401824951, "learning_rate": 1.546636345472805e-05, "loss": 1.0673, "step": 9561 }, { "epoch": 0.3369363179083309, "grad_norm": 1.2537705898284912, "learning_rate": 1.5465407748045436e-05, "loss": 0.8057, "step": 9562 }, { "epoch": 0.33697155492128933, "grad_norm": 3.5656983852386475, "learning_rate": 1.5464451970175193e-05, "loss": 3.5773, "step": 9563 }, { "epoch": 0.3370067919342477, "grad_norm": 2.2170333862304688, "learning_rate": 1.5463496121129768e-05, "loss": 0.9374, "step": 9564 }, { "epoch": 0.33704202894720614, "grad_norm": 2.135864019393921, "learning_rate": 1.546254020092161e-05, "loss": 0.9144, "step": 9565 }, { "epoch": 0.3370772659601646, "grad_norm": 1.2673507928848267, "learning_rate": 1.546158420956317e-05, "loss": 1.0072, "step": 9566 }, { "epoch": 0.33711250297312295, "grad_norm": 1.7484500408172607, "learning_rate": 1.5460628147066897e-05, "loss": 0.8994, "step": 9567 }, { "epoch": 0.3371477399860814, "grad_norm": 2.107842206954956, "learning_rate": 1.545967201344525e-05, "loss": 0.8833, "step": 9568 }, { "epoch": 0.3371829769990398, "grad_norm": 4.098516464233398, "learning_rate": 1.545871580871068e-05, "loss": 3.446, "step": 9569 }, { "epoch": 0.3372182140119982, "grad_norm": 2.7996413707733154, "learning_rate": 1.545775953287564e-05, "loss": 0.7086, "step": 9570 }, { "epoch": 0.33725345102495663, "grad_norm": 5.803427696228027, "learning_rate": 1.5456803185952584e-05, "loss": 3.0543, "step": 9571 }, { "epoch": 0.337288688037915, "grad_norm": 3.984654426574707, "learning_rate": 1.5455846767953976e-05, "loss": 3.5626, "step": 9572 }, { "epoch": 0.33732392505087344, "grad_norm": 1.629693865776062, "learning_rate": 1.5454890278892265e-05, "loss": 1.0512, "step": 9573 }, { "epoch": 0.3373591620638319, "grad_norm": 1.615066409111023, "learning_rate": 1.5453933718779916e-05, "loss": 0.7293, "step": 9574 }, { "epoch": 0.33739439907679025, "grad_norm": 4.117073059082031, "learning_rate": 1.5452977087629383e-05, "loss": 3.7137, "step": 9575 }, { "epoch": 0.3374296360897487, "grad_norm": 7.741898536682129, "learning_rate": 1.5452020385453132e-05, "loss": 5.2592, "step": 9576 }, { "epoch": 0.33746487310270706, "grad_norm": 1.937492847442627, "learning_rate": 1.5451063612263618e-05, "loss": 1.0102, "step": 9577 }, { "epoch": 0.3375001101156655, "grad_norm": 1.883234977722168, "learning_rate": 1.545010676807331e-05, "loss": 1.1194, "step": 9578 }, { "epoch": 0.33753534712862393, "grad_norm": 3.4323770999908447, "learning_rate": 1.5449149852894664e-05, "loss": 3.129, "step": 9579 }, { "epoch": 0.3375705841415823, "grad_norm": 1.1793122291564941, "learning_rate": 1.544819286674015e-05, "loss": 0.8451, "step": 9580 }, { "epoch": 0.33760582115454074, "grad_norm": 2.800496816635132, "learning_rate": 1.544723580962223e-05, "loss": 3.1878, "step": 9581 }, { "epoch": 0.3376410581674991, "grad_norm": 5.640321254730225, "learning_rate": 1.5446278681553363e-05, "loss": 2.8268, "step": 9582 }, { "epoch": 0.33767629518045755, "grad_norm": 1.6364860534667969, "learning_rate": 1.5445321482546027e-05, "loss": 0.829, "step": 9583 }, { "epoch": 0.337711532193416, "grad_norm": 7.4294538497924805, "learning_rate": 1.544436421261269e-05, "loss": 8.1158, "step": 9584 }, { "epoch": 0.33774676920637436, "grad_norm": 5.877633571624756, "learning_rate": 1.5443406871765806e-05, "loss": 3.4226, "step": 9585 }, { "epoch": 0.3377820062193328, "grad_norm": 3.097064971923828, "learning_rate": 1.5442449460017858e-05, "loss": 2.4995, "step": 9586 }, { "epoch": 0.3378172432322912, "grad_norm": 6.916935443878174, "learning_rate": 1.5441491977381313e-05, "loss": 6.1886, "step": 9587 }, { "epoch": 0.3378524802452496, "grad_norm": 6.866634368896484, "learning_rate": 1.5440534423868645e-05, "loss": 6.0583, "step": 9588 }, { "epoch": 0.33788771725820804, "grad_norm": 1.1856961250305176, "learning_rate": 1.543957679949232e-05, "loss": 0.8873, "step": 9589 }, { "epoch": 0.3379229542711664, "grad_norm": 2.1173388957977295, "learning_rate": 1.5438619104264813e-05, "loss": 0.942, "step": 9590 }, { "epoch": 0.33795819128412485, "grad_norm": 1.2310981750488281, "learning_rate": 1.5437661338198605e-05, "loss": 0.7438, "step": 9591 }, { "epoch": 0.33799342829708323, "grad_norm": 5.116223335266113, "learning_rate": 1.543670350130616e-05, "loss": 2.9569, "step": 9592 }, { "epoch": 0.33802866531004166, "grad_norm": 1.5624988079071045, "learning_rate": 1.543574559359996e-05, "loss": 1.1251, "step": 9593 }, { "epoch": 0.3380639023230001, "grad_norm": 2.2021677494049072, "learning_rate": 1.543478761509248e-05, "loss": 1.0497, "step": 9594 }, { "epoch": 0.3380991393359585, "grad_norm": 2.3854293823242188, "learning_rate": 1.5433829565796203e-05, "loss": 3.8849, "step": 9595 }, { "epoch": 0.3381343763489169, "grad_norm": 1.4713307619094849, "learning_rate": 1.54328714457236e-05, "loss": 0.7658, "step": 9596 }, { "epoch": 0.3381696133618753, "grad_norm": 5.077736854553223, "learning_rate": 1.5431913254887158e-05, "loss": 6.3609, "step": 9597 }, { "epoch": 0.3382048503748337, "grad_norm": 2.8761417865753174, "learning_rate": 1.5430954993299348e-05, "loss": 3.2902, "step": 9598 }, { "epoch": 0.33824008738779215, "grad_norm": 1.1114203929901123, "learning_rate": 1.5429996660972662e-05, "loss": 0.676, "step": 9599 }, { "epoch": 0.33827532440075053, "grad_norm": 1.4731481075286865, "learning_rate": 1.542903825791958e-05, "loss": 0.9298, "step": 9600 }, { "epoch": 0.33831056141370897, "grad_norm": 0.9939187169075012, "learning_rate": 1.542807978415258e-05, "loss": 0.8624, "step": 9601 }, { "epoch": 0.3383457984266674, "grad_norm": 1.6673895120620728, "learning_rate": 1.542712123968415e-05, "loss": 0.952, "step": 9602 }, { "epoch": 0.3383810354396258, "grad_norm": 1.342537760734558, "learning_rate": 1.5426162624526778e-05, "loss": 0.7631, "step": 9603 }, { "epoch": 0.3384162724525842, "grad_norm": 3.545091390609741, "learning_rate": 1.5425203938692938e-05, "loss": 3.8744, "step": 9604 }, { "epoch": 0.3384515094655426, "grad_norm": 3.7822604179382324, "learning_rate": 1.5424245182195134e-05, "loss": 6.084, "step": 9605 }, { "epoch": 0.338486746478501, "grad_norm": 3.2667076587677, "learning_rate": 1.542328635504584e-05, "loss": 2.7039, "step": 9606 }, { "epoch": 0.33852198349145945, "grad_norm": 1.4051004648208618, "learning_rate": 1.5422327457257554e-05, "loss": 0.9187, "step": 9607 }, { "epoch": 0.33855722050441783, "grad_norm": 3.0659432411193848, "learning_rate": 1.542136848884276e-05, "loss": 3.0688, "step": 9608 }, { "epoch": 0.33859245751737627, "grad_norm": 6.12666130065918, "learning_rate": 1.542040944981395e-05, "loss": 7.6824, "step": 9609 }, { "epoch": 0.33862769453033464, "grad_norm": 1.8042019605636597, "learning_rate": 1.541945034018362e-05, "loss": 0.8276, "step": 9610 }, { "epoch": 0.3386629315432931, "grad_norm": 6.703479290008545, "learning_rate": 1.541849115996426e-05, "loss": 3.82, "step": 9611 }, { "epoch": 0.3386981685562515, "grad_norm": 1.4990144968032837, "learning_rate": 1.5417531909168358e-05, "loss": 0.9498, "step": 9612 }, { "epoch": 0.3387334055692099, "grad_norm": 5.266093730926514, "learning_rate": 1.5416572587808415e-05, "loss": 5.5878, "step": 9613 }, { "epoch": 0.3387686425821683, "grad_norm": 1.4710936546325684, "learning_rate": 1.5415613195896922e-05, "loss": 0.9271, "step": 9614 }, { "epoch": 0.3388038795951267, "grad_norm": 4.957300186157227, "learning_rate": 1.541465373344638e-05, "loss": 3.4765, "step": 9615 }, { "epoch": 0.33883911660808513, "grad_norm": 1.4606679677963257, "learning_rate": 1.541369420046928e-05, "loss": 0.9578, "step": 9616 }, { "epoch": 0.33887435362104357, "grad_norm": 11.7645902633667, "learning_rate": 1.5412734596978127e-05, "loss": 3.6297, "step": 9617 }, { "epoch": 0.33890959063400194, "grad_norm": 1.9718488454818726, "learning_rate": 1.5411774922985415e-05, "loss": 0.9792, "step": 9618 }, { "epoch": 0.3389448276469604, "grad_norm": 1.491050362586975, "learning_rate": 1.5410815178503648e-05, "loss": 1.0087, "step": 9619 }, { "epoch": 0.33898006465991876, "grad_norm": 1.8178246021270752, "learning_rate": 1.5409855363545318e-05, "loss": 1.1249, "step": 9620 }, { "epoch": 0.3390153016728772, "grad_norm": 1.138069987297058, "learning_rate": 1.540889547812294e-05, "loss": 0.9492, "step": 9621 }, { "epoch": 0.3390505386858356, "grad_norm": 1.4896637201309204, "learning_rate": 1.5407935522249003e-05, "loss": 0.9967, "step": 9622 }, { "epoch": 0.339085775698794, "grad_norm": 13.563164710998535, "learning_rate": 1.5406975495936017e-05, "loss": 6.2905, "step": 9623 }, { "epoch": 0.33912101271175243, "grad_norm": 6.358283042907715, "learning_rate": 1.5406015399196493e-05, "loss": 3.3327, "step": 9624 }, { "epoch": 0.3391562497247108, "grad_norm": 3.807508707046509, "learning_rate": 1.5405055232042924e-05, "loss": 3.0906, "step": 9625 }, { "epoch": 0.33919148673766925, "grad_norm": 2.8463335037231445, "learning_rate": 1.5404094994487825e-05, "loss": 2.8363, "step": 9626 }, { "epoch": 0.3392267237506277, "grad_norm": 1.2612247467041016, "learning_rate": 1.54031346865437e-05, "loss": 0.8259, "step": 9627 }, { "epoch": 0.33926196076358606, "grad_norm": 7.1995649337768555, "learning_rate": 1.5402174308223054e-05, "loss": 4.9534, "step": 9628 }, { "epoch": 0.3392971977765445, "grad_norm": 4.181107044219971, "learning_rate": 1.5401213859538403e-05, "loss": 3.7594, "step": 9629 }, { "epoch": 0.33933243478950287, "grad_norm": 3.373929023742676, "learning_rate": 1.5400253340502254e-05, "loss": 3.254, "step": 9630 }, { "epoch": 0.3393676718024613, "grad_norm": 3.852760076522827, "learning_rate": 1.539929275112711e-05, "loss": 3.7447, "step": 9631 }, { "epoch": 0.33940290881541973, "grad_norm": 1.171393871307373, "learning_rate": 1.53983320914255e-05, "loss": 0.8202, "step": 9632 }, { "epoch": 0.3394381458283781, "grad_norm": 3.897722005844116, "learning_rate": 1.539737136140992e-05, "loss": 3.2044, "step": 9633 }, { "epoch": 0.33947338284133655, "grad_norm": 5.966316223144531, "learning_rate": 1.5396410561092898e-05, "loss": 5.8068, "step": 9634 }, { "epoch": 0.3395086198542949, "grad_norm": 4.288708209991455, "learning_rate": 1.5395449690486936e-05, "loss": 3.859, "step": 9635 }, { "epoch": 0.33954385686725336, "grad_norm": 2.1390175819396973, "learning_rate": 1.5394488749604555e-05, "loss": 0.9177, "step": 9636 }, { "epoch": 0.3395790938802118, "grad_norm": 4.0660529136657715, "learning_rate": 1.539352773845827e-05, "loss": 2.8271, "step": 9637 }, { "epoch": 0.33961433089317017, "grad_norm": 6.691656589508057, "learning_rate": 1.53925666570606e-05, "loss": 3.7658, "step": 9638 }, { "epoch": 0.3396495679061286, "grad_norm": 4.523564338684082, "learning_rate": 1.5391605505424062e-05, "loss": 3.7748, "step": 9639 }, { "epoch": 0.33968480491908704, "grad_norm": 7.426148891448975, "learning_rate": 1.5390644283561177e-05, "loss": 6.1709, "step": 9640 }, { "epoch": 0.3397200419320454, "grad_norm": 1.9530898332595825, "learning_rate": 1.5389682991484462e-05, "loss": 0.8605, "step": 9641 }, { "epoch": 0.33975527894500385, "grad_norm": 7.580346584320068, "learning_rate": 1.5388721629206442e-05, "loss": 5.4787, "step": 9642 }, { "epoch": 0.3397905159579622, "grad_norm": 1.1601330041885376, "learning_rate": 1.5387760196739634e-05, "loss": 0.801, "step": 9643 }, { "epoch": 0.33982575297092066, "grad_norm": 3.0573205947875977, "learning_rate": 1.5386798694096562e-05, "loss": 2.1308, "step": 9644 }, { "epoch": 0.3398609899838791, "grad_norm": 1.353187918663025, "learning_rate": 1.5385837121289753e-05, "loss": 1.0724, "step": 9645 }, { "epoch": 0.33989622699683747, "grad_norm": 1.3607478141784668, "learning_rate": 1.538487547833173e-05, "loss": 0.9013, "step": 9646 }, { "epoch": 0.3399314640097959, "grad_norm": 11.385369300842285, "learning_rate": 1.5383913765235015e-05, "loss": 6.0372, "step": 9647 }, { "epoch": 0.3399667010227543, "grad_norm": 3.8774960041046143, "learning_rate": 1.538295198201214e-05, "loss": 3.3934, "step": 9648 }, { "epoch": 0.3400019380357127, "grad_norm": 3.325690507888794, "learning_rate": 1.5381990128675633e-05, "loss": 2.7627, "step": 9649 }, { "epoch": 0.34003717504867115, "grad_norm": 5.136990547180176, "learning_rate": 1.538102820523801e-05, "loss": 3.5022, "step": 9650 }, { "epoch": 0.3400724120616295, "grad_norm": 4.46668815612793, "learning_rate": 1.538006621171182e-05, "loss": 3.1724, "step": 9651 }, { "epoch": 0.34010764907458796, "grad_norm": 5.780797004699707, "learning_rate": 1.5379104148109573e-05, "loss": 6.0631, "step": 9652 }, { "epoch": 0.34014288608754634, "grad_norm": 1.7751734256744385, "learning_rate": 1.5378142014443812e-05, "loss": 1.0187, "step": 9653 }, { "epoch": 0.34017812310050477, "grad_norm": 3.746551513671875, "learning_rate": 1.5377179810727067e-05, "loss": 3.6648, "step": 9654 }, { "epoch": 0.3402133601134632, "grad_norm": 11.189922332763672, "learning_rate": 1.537621753697187e-05, "loss": 5.6669, "step": 9655 }, { "epoch": 0.3402485971264216, "grad_norm": 1.8757747411727905, "learning_rate": 1.5375255193190755e-05, "loss": 0.7851, "step": 9656 }, { "epoch": 0.34028383413938, "grad_norm": 1.3761674165725708, "learning_rate": 1.537429277939626e-05, "loss": 0.8647, "step": 9657 }, { "epoch": 0.3403190711523384, "grad_norm": 5.917372703552246, "learning_rate": 1.5373330295600908e-05, "loss": 3.4925, "step": 9658 }, { "epoch": 0.3403543081652968, "grad_norm": 6.436709880828857, "learning_rate": 1.537236774181725e-05, "loss": 6.2682, "step": 9659 }, { "epoch": 0.34038954517825526, "grad_norm": 1.3705682754516602, "learning_rate": 1.537140511805782e-05, "loss": 0.7985, "step": 9660 }, { "epoch": 0.34042478219121364, "grad_norm": 2.138476848602295, "learning_rate": 1.537044242433515e-05, "loss": 1.0237, "step": 9661 }, { "epoch": 0.34046001920417207, "grad_norm": 1.6918867826461792, "learning_rate": 1.5369479660661787e-05, "loss": 0.7275, "step": 9662 }, { "epoch": 0.34049525621713045, "grad_norm": 4.567722797393799, "learning_rate": 1.5368516827050265e-05, "loss": 3.4412, "step": 9663 }, { "epoch": 0.3405304932300889, "grad_norm": 1.522629737854004, "learning_rate": 1.5367553923513127e-05, "loss": 0.8232, "step": 9664 }, { "epoch": 0.3405657302430473, "grad_norm": 5.6855974197387695, "learning_rate": 1.5366590950062917e-05, "loss": 6.5192, "step": 9665 }, { "epoch": 0.3406009672560057, "grad_norm": 1.7208565473556519, "learning_rate": 1.5365627906712176e-05, "loss": 0.943, "step": 9666 }, { "epoch": 0.3406362042689641, "grad_norm": 4.741166114807129, "learning_rate": 1.5364664793473445e-05, "loss": 3.8611, "step": 9667 }, { "epoch": 0.3406714412819225, "grad_norm": 4.545846939086914, "learning_rate": 1.5363701610359276e-05, "loss": 3.4305, "step": 9668 }, { "epoch": 0.34070667829488094, "grad_norm": 1.4192513227462769, "learning_rate": 1.5362738357382206e-05, "loss": 0.9105, "step": 9669 }, { "epoch": 0.34074191530783937, "grad_norm": 3.372518539428711, "learning_rate": 1.536177503455479e-05, "loss": 3.5645, "step": 9670 }, { "epoch": 0.34077715232079775, "grad_norm": 3.8283965587615967, "learning_rate": 1.5360811641889567e-05, "loss": 3.1293, "step": 9671 }, { "epoch": 0.3408123893337562, "grad_norm": 2.329462766647339, "learning_rate": 1.5359848179399094e-05, "loss": 3.1508, "step": 9672 }, { "epoch": 0.3408476263467146, "grad_norm": 1.7872977256774902, "learning_rate": 1.5358884647095915e-05, "loss": 0.6099, "step": 9673 }, { "epoch": 0.340882863359673, "grad_norm": 2.129659414291382, "learning_rate": 1.5357921044992582e-05, "loss": 0.9519, "step": 9674 }, { "epoch": 0.3409181003726314, "grad_norm": 4.925720691680908, "learning_rate": 1.535695737310164e-05, "loss": 5.2984, "step": 9675 }, { "epoch": 0.3409533373855898, "grad_norm": 6.004595756530762, "learning_rate": 1.535599363143565e-05, "loss": 5.8469, "step": 9676 }, { "epoch": 0.34098857439854824, "grad_norm": 3.6474673748016357, "learning_rate": 1.5355029820007155e-05, "loss": 5.2412, "step": 9677 }, { "epoch": 0.34102381141150667, "grad_norm": 5.325292110443115, "learning_rate": 1.5354065938828723e-05, "loss": 5.8764, "step": 9678 }, { "epoch": 0.34105904842446505, "grad_norm": 5.1786932945251465, "learning_rate": 1.5353101987912893e-05, "loss": 5.9518, "step": 9679 }, { "epoch": 0.3410942854374235, "grad_norm": 1.053612232208252, "learning_rate": 1.535213796727223e-05, "loss": 0.893, "step": 9680 }, { "epoch": 0.34112952245038186, "grad_norm": 2.3095450401306152, "learning_rate": 1.535117387691929e-05, "loss": 2.6452, "step": 9681 }, { "epoch": 0.3411647594633403, "grad_norm": 2.6154069900512695, "learning_rate": 1.5350209716866625e-05, "loss": 1.0228, "step": 9682 }, { "epoch": 0.34119999647629873, "grad_norm": 4.747991561889648, "learning_rate": 1.5349245487126803e-05, "loss": 3.4465, "step": 9683 }, { "epoch": 0.3412352334892571, "grad_norm": 3.985935688018799, "learning_rate": 1.5348281187712372e-05, "loss": 3.6841, "step": 9684 }, { "epoch": 0.34127047050221554, "grad_norm": 5.8656768798828125, "learning_rate": 1.5347316818635902e-05, "loss": 8.3761, "step": 9685 }, { "epoch": 0.3413057075151739, "grad_norm": 3.2091517448425293, "learning_rate": 1.5346352379909945e-05, "loss": 3.3749, "step": 9686 }, { "epoch": 0.34134094452813235, "grad_norm": 1.3419302701950073, "learning_rate": 1.5345387871547065e-05, "loss": 1.024, "step": 9687 }, { "epoch": 0.3413761815410908, "grad_norm": 1.2661525011062622, "learning_rate": 1.534442329355983e-05, "loss": 0.8901, "step": 9688 }, { "epoch": 0.34141141855404916, "grad_norm": 4.637263298034668, "learning_rate": 1.5343458645960802e-05, "loss": 3.3519, "step": 9689 }, { "epoch": 0.3414466555670076, "grad_norm": 4.078048229217529, "learning_rate": 1.534249392876254e-05, "loss": 3.308, "step": 9690 }, { "epoch": 0.341481892579966, "grad_norm": 1.4930754899978638, "learning_rate": 1.534152914197762e-05, "loss": 1.0661, "step": 9691 }, { "epoch": 0.3415171295929244, "grad_norm": 4.652017593383789, "learning_rate": 1.5340564285618603e-05, "loss": 4.8777, "step": 9692 }, { "epoch": 0.34155236660588284, "grad_norm": 4.026248455047607, "learning_rate": 1.5339599359698052e-05, "loss": 3.6836, "step": 9693 }, { "epoch": 0.3415876036188412, "grad_norm": 6.102459907531738, "learning_rate": 1.533863436422854e-05, "loss": 3.575, "step": 9694 }, { "epoch": 0.34162284063179965, "grad_norm": 6.506194591522217, "learning_rate": 1.5337669299222638e-05, "loss": 5.8529, "step": 9695 }, { "epoch": 0.34165807764475803, "grad_norm": 6.575758457183838, "learning_rate": 1.533670416469291e-05, "loss": 3.4159, "step": 9696 }, { "epoch": 0.34169331465771646, "grad_norm": 1.28402578830719, "learning_rate": 1.5335738960651934e-05, "loss": 0.8632, "step": 9697 }, { "epoch": 0.3417285516706749, "grad_norm": 6.5436201095581055, "learning_rate": 1.5334773687112273e-05, "loss": 3.3921, "step": 9698 }, { "epoch": 0.3417637886836333, "grad_norm": 3.9334700107574463, "learning_rate": 1.533380834408651e-05, "loss": 2.8142, "step": 9699 }, { "epoch": 0.3417990256965917, "grad_norm": 2.0827431678771973, "learning_rate": 1.533284293158721e-05, "loss": 0.8293, "step": 9700 }, { "epoch": 0.3418342627095501, "grad_norm": 5.174749851226807, "learning_rate": 1.5331877449626954e-05, "loss": 5.7651, "step": 9701 }, { "epoch": 0.3418694997225085, "grad_norm": 3.2047581672668457, "learning_rate": 1.5330911898218314e-05, "loss": 3.3545, "step": 9702 }, { "epoch": 0.34190473673546695, "grad_norm": 6.195309638977051, "learning_rate": 1.5329946277373867e-05, "loss": 8.7588, "step": 9703 }, { "epoch": 0.34193997374842533, "grad_norm": 5.256566047668457, "learning_rate": 1.5328980587106193e-05, "loss": 3.6575, "step": 9704 }, { "epoch": 0.34197521076138376, "grad_norm": 1.2763428688049316, "learning_rate": 1.532801482742787e-05, "loss": 1.226, "step": 9705 }, { "epoch": 0.3420104477743422, "grad_norm": 1.391412615776062, "learning_rate": 1.532704899835147e-05, "loss": 0.6906, "step": 9706 }, { "epoch": 0.3420456847873006, "grad_norm": 1.374687910079956, "learning_rate": 1.5326083099889578e-05, "loss": 0.8242, "step": 9707 }, { "epoch": 0.342080921800259, "grad_norm": 2.0331947803497314, "learning_rate": 1.5325117132054778e-05, "loss": 0.8158, "step": 9708 }, { "epoch": 0.3421161588132174, "grad_norm": 3.395843505859375, "learning_rate": 1.5324151094859647e-05, "loss": 3.7567, "step": 9709 }, { "epoch": 0.3421513958261758, "grad_norm": 1.3384017944335938, "learning_rate": 1.5323184988316775e-05, "loss": 1.3722, "step": 9710 }, { "epoch": 0.34218663283913425, "grad_norm": 4.673214912414551, "learning_rate": 1.5322218812438736e-05, "loss": 3.856, "step": 9711 }, { "epoch": 0.34222186985209263, "grad_norm": 5.159557342529297, "learning_rate": 1.532125256723812e-05, "loss": 6.0636, "step": 9712 }, { "epoch": 0.34225710686505106, "grad_norm": 3.159111976623535, "learning_rate": 1.532028625272751e-05, "loss": 2.8387, "step": 9713 }, { "epoch": 0.34229234387800944, "grad_norm": 1.562374472618103, "learning_rate": 1.53193198689195e-05, "loss": 0.9154, "step": 9714 }, { "epoch": 0.3423275808909679, "grad_norm": 3.5428476333618164, "learning_rate": 1.5318353415826665e-05, "loss": 3.7427, "step": 9715 }, { "epoch": 0.3423628179039263, "grad_norm": 4.217818737030029, "learning_rate": 1.53173868934616e-05, "loss": 3.0419, "step": 9716 }, { "epoch": 0.3423980549168847, "grad_norm": 4.982357978820801, "learning_rate": 1.5316420301836893e-05, "loss": 3.0207, "step": 9717 }, { "epoch": 0.3424332919298431, "grad_norm": 1.7685500383377075, "learning_rate": 1.5315453640965136e-05, "loss": 0.986, "step": 9718 }, { "epoch": 0.3424685289428015, "grad_norm": 1.273686170578003, "learning_rate": 1.531448691085892e-05, "loss": 0.8367, "step": 9719 }, { "epoch": 0.34250376595575993, "grad_norm": 3.587232828140259, "learning_rate": 1.531352011153083e-05, "loss": 3.4756, "step": 9720 }, { "epoch": 0.34253900296871836, "grad_norm": 2.0405983924865723, "learning_rate": 1.5312553242993466e-05, "loss": 0.7754, "step": 9721 }, { "epoch": 0.34257423998167674, "grad_norm": 1.0049684047698975, "learning_rate": 1.531158630525942e-05, "loss": 0.8326, "step": 9722 }, { "epoch": 0.3426094769946352, "grad_norm": 1.5303598642349243, "learning_rate": 1.5310619298341285e-05, "loss": 0.8303, "step": 9723 }, { "epoch": 0.34264471400759355, "grad_norm": 3.541011333465576, "learning_rate": 1.5309652222251656e-05, "loss": 3.2336, "step": 9724 }, { "epoch": 0.342679951020552, "grad_norm": 1.2102822065353394, "learning_rate": 1.5308685077003135e-05, "loss": 0.7757, "step": 9725 }, { "epoch": 0.3427151880335104, "grad_norm": 1.972474217414856, "learning_rate": 1.530771786260831e-05, "loss": 0.7885, "step": 9726 }, { "epoch": 0.3427504250464688, "grad_norm": 5.704068183898926, "learning_rate": 1.5306750579079787e-05, "loss": 3.3224, "step": 9727 }, { "epoch": 0.34278566205942723, "grad_norm": 1.6192684173583984, "learning_rate": 1.5305783226430157e-05, "loss": 0.7897, "step": 9728 }, { "epoch": 0.3428208990723856, "grad_norm": 1.5085359811782837, "learning_rate": 1.5304815804672034e-05, "loss": 1.1518, "step": 9729 }, { "epoch": 0.34285613608534404, "grad_norm": 2.955820083618164, "learning_rate": 1.5303848313818e-05, "loss": 0.8073, "step": 9730 }, { "epoch": 0.3428913730983025, "grad_norm": 1.6869088411331177, "learning_rate": 1.5302880753880672e-05, "loss": 0.7826, "step": 9731 }, { "epoch": 0.34292661011126085, "grad_norm": 3.5760228633880615, "learning_rate": 1.5301913124872642e-05, "loss": 2.8022, "step": 9732 }, { "epoch": 0.3429618471242193, "grad_norm": 1.1930949687957764, "learning_rate": 1.530094542680652e-05, "loss": 0.9351, "step": 9733 }, { "epoch": 0.34299708413717767, "grad_norm": 1.1082391738891602, "learning_rate": 1.529997765969491e-05, "loss": 0.9442, "step": 9734 }, { "epoch": 0.3430323211501361, "grad_norm": 8.49905014038086, "learning_rate": 1.5299009823550414e-05, "loss": 4.7182, "step": 9735 }, { "epoch": 0.34306755816309453, "grad_norm": 7.085470199584961, "learning_rate": 1.529804191838564e-05, "loss": 3.6548, "step": 9736 }, { "epoch": 0.3431027951760529, "grad_norm": 4.63291072845459, "learning_rate": 1.52970739442132e-05, "loss": 2.9718, "step": 9737 }, { "epoch": 0.34313803218901134, "grad_norm": 7.853353023529053, "learning_rate": 1.529610590104569e-05, "loss": 2.3303, "step": 9738 }, { "epoch": 0.3431732692019697, "grad_norm": 5.342015266418457, "learning_rate": 1.529513778889573e-05, "loss": 3.4072, "step": 9739 }, { "epoch": 0.34320850621492816, "grad_norm": 1.326264500617981, "learning_rate": 1.5294169607775923e-05, "loss": 0.8542, "step": 9740 }, { "epoch": 0.3432437432278866, "grad_norm": 3.7757272720336914, "learning_rate": 1.5293201357698887e-05, "loss": 3.1887, "step": 9741 }, { "epoch": 0.34327898024084497, "grad_norm": 6.042852401733398, "learning_rate": 1.5292233038677225e-05, "loss": 2.9687, "step": 9742 }, { "epoch": 0.3433142172538034, "grad_norm": 0.9817525148391724, "learning_rate": 1.5291264650723557e-05, "loss": 0.8666, "step": 9743 }, { "epoch": 0.34334945426676183, "grad_norm": 1.760441780090332, "learning_rate": 1.529029619385049e-05, "loss": 0.9249, "step": 9744 }, { "epoch": 0.3433846912797202, "grad_norm": 6.070212364196777, "learning_rate": 1.528932766807064e-05, "loss": 3.2243, "step": 9745 }, { "epoch": 0.34341992829267864, "grad_norm": 1.800689935684204, "learning_rate": 1.5288359073396627e-05, "loss": 0.9082, "step": 9746 }, { "epoch": 0.343455165305637, "grad_norm": 3.1134698390960693, "learning_rate": 1.528739040984106e-05, "loss": 3.1406, "step": 9747 }, { "epoch": 0.34349040231859546, "grad_norm": 5.888528823852539, "learning_rate": 1.5286421677416562e-05, "loss": 2.7119, "step": 9748 }, { "epoch": 0.3435256393315539, "grad_norm": 5.971796989440918, "learning_rate": 1.528545287613575e-05, "loss": 3.2489, "step": 9749 }, { "epoch": 0.34356087634451227, "grad_norm": 0.9710378646850586, "learning_rate": 1.528448400601124e-05, "loss": 0.7636, "step": 9750 }, { "epoch": 0.3435961133574707, "grad_norm": 6.222456932067871, "learning_rate": 1.528351506705565e-05, "loss": 4.3401, "step": 9751 }, { "epoch": 0.3436313503704291, "grad_norm": 1.0910460948944092, "learning_rate": 1.5282546059281604e-05, "loss": 0.9692, "step": 9752 }, { "epoch": 0.3436665873833875, "grad_norm": 4.849890232086182, "learning_rate": 1.5281576982701724e-05, "loss": 3.1076, "step": 9753 }, { "epoch": 0.34370182439634595, "grad_norm": 3.650160074234009, "learning_rate": 1.5280607837328633e-05, "loss": 3.7968, "step": 9754 }, { "epoch": 0.3437370614093043, "grad_norm": 5.43051815032959, "learning_rate": 1.5279638623174946e-05, "loss": 5.063, "step": 9755 }, { "epoch": 0.34377229842226276, "grad_norm": 1.2131344079971313, "learning_rate": 1.52786693402533e-05, "loss": 0.6354, "step": 9756 }, { "epoch": 0.34380753543522113, "grad_norm": 1.1082884073257446, "learning_rate": 1.527769998857631e-05, "loss": 0.8704, "step": 9757 }, { "epoch": 0.34384277244817957, "grad_norm": 7.274131774902344, "learning_rate": 1.5276730568156603e-05, "loss": 3.4738, "step": 9758 }, { "epoch": 0.343878009461138, "grad_norm": 1.4594672918319702, "learning_rate": 1.5275761079006812e-05, "loss": 0.906, "step": 9759 }, { "epoch": 0.3439132464740964, "grad_norm": 3.5307581424713135, "learning_rate": 1.527479152113956e-05, "loss": 3.3939, "step": 9760 }, { "epoch": 0.3439484834870548, "grad_norm": 6.463149070739746, "learning_rate": 1.5273821894567474e-05, "loss": 6.3287, "step": 9761 }, { "epoch": 0.3439837205000132, "grad_norm": 2.2736892700195312, "learning_rate": 1.5272852199303193e-05, "loss": 0.9895, "step": 9762 }, { "epoch": 0.3440189575129716, "grad_norm": 2.254934310913086, "learning_rate": 1.5271882435359332e-05, "loss": 1.3842, "step": 9763 }, { "epoch": 0.34405419452593006, "grad_norm": 1.212018609046936, "learning_rate": 1.5270912602748536e-05, "loss": 1.0564, "step": 9764 }, { "epoch": 0.34408943153888844, "grad_norm": 4.167490482330322, "learning_rate": 1.526994270148343e-05, "loss": 3.2116, "step": 9765 }, { "epoch": 0.34412466855184687, "grad_norm": 2.6282155513763428, "learning_rate": 1.526897273157665e-05, "loss": 2.678, "step": 9766 }, { "epoch": 0.34415990556480525, "grad_norm": 2.136518716812134, "learning_rate": 1.526800269304083e-05, "loss": 0.9096, "step": 9767 }, { "epoch": 0.3441951425777637, "grad_norm": 4.220759391784668, "learning_rate": 1.5267032585888602e-05, "loss": 3.3601, "step": 9768 }, { "epoch": 0.3442303795907221, "grad_norm": 1.7914901971817017, "learning_rate": 1.5266062410132606e-05, "loss": 0.9082, "step": 9769 }, { "epoch": 0.3442656166036805, "grad_norm": 4.084515571594238, "learning_rate": 1.5265092165785475e-05, "loss": 3.8655, "step": 9770 }, { "epoch": 0.3443008536166389, "grad_norm": 2.1049060821533203, "learning_rate": 1.5264121852859845e-05, "loss": 2.6438, "step": 9771 }, { "epoch": 0.3443360906295973, "grad_norm": 5.918384552001953, "learning_rate": 1.526315147136836e-05, "loss": 5.3539, "step": 9772 }, { "epoch": 0.34437132764255574, "grad_norm": 1.0666368007659912, "learning_rate": 1.526218102132366e-05, "loss": 0.6643, "step": 9773 }, { "epoch": 0.34440656465551417, "grad_norm": 1.240821361541748, "learning_rate": 1.5261210502738374e-05, "loss": 0.7557, "step": 9774 }, { "epoch": 0.34444180166847255, "grad_norm": 2.6246016025543213, "learning_rate": 1.5260239915625157e-05, "loss": 0.9076, "step": 9775 }, { "epoch": 0.344477038681431, "grad_norm": 5.938595771789551, "learning_rate": 1.5259269259996645e-05, "loss": 3.2002, "step": 9776 }, { "epoch": 0.3445122756943894, "grad_norm": 2.163724184036255, "learning_rate": 1.5258298535865477e-05, "loss": 1.0654, "step": 9777 }, { "epoch": 0.3445475127073478, "grad_norm": 9.242137908935547, "learning_rate": 1.5257327743244305e-05, "loss": 7.506, "step": 9778 }, { "epoch": 0.3445827497203062, "grad_norm": 3.646768093109131, "learning_rate": 1.5256356882145766e-05, "loss": 2.8858, "step": 9779 }, { "epoch": 0.3446179867332646, "grad_norm": 2.8089234828948975, "learning_rate": 1.5255385952582511e-05, "loss": 2.6489, "step": 9780 }, { "epoch": 0.34465322374622304, "grad_norm": 3.9289777278900146, "learning_rate": 1.5254414954567187e-05, "loss": 3.6567, "step": 9781 }, { "epoch": 0.34468846075918147, "grad_norm": 3.2587532997131348, "learning_rate": 1.5253443888112434e-05, "loss": 3.0234, "step": 9782 }, { "epoch": 0.34472369777213985, "grad_norm": 1.2634046077728271, "learning_rate": 1.525247275323091e-05, "loss": 0.7772, "step": 9783 }, { "epoch": 0.3447589347850983, "grad_norm": 5.459617614746094, "learning_rate": 1.5251501549935254e-05, "loss": 3.6573, "step": 9784 }, { "epoch": 0.34479417179805666, "grad_norm": 6.494335651397705, "learning_rate": 1.5250530278238125e-05, "loss": 3.6432, "step": 9785 }, { "epoch": 0.3448294088110151, "grad_norm": 15.825746536254883, "learning_rate": 1.524955893815217e-05, "loss": 5.4431, "step": 9786 }, { "epoch": 0.3448646458239735, "grad_norm": 4.127247333526611, "learning_rate": 1.5248587529690037e-05, "loss": 3.4194, "step": 9787 }, { "epoch": 0.3448998828369319, "grad_norm": 1.2014557123184204, "learning_rate": 1.5247616052864387e-05, "loss": 0.8296, "step": 9788 }, { "epoch": 0.34493511984989034, "grad_norm": 3.665698289871216, "learning_rate": 1.5246644507687868e-05, "loss": 4.365, "step": 9789 }, { "epoch": 0.3449703568628487, "grad_norm": 1.0941425561904907, "learning_rate": 1.524567289417314e-05, "loss": 0.7596, "step": 9790 }, { "epoch": 0.34500559387580715, "grad_norm": 4.81921911239624, "learning_rate": 1.5244701212332848e-05, "loss": 3.2454, "step": 9791 }, { "epoch": 0.3450408308887656, "grad_norm": 1.2101435661315918, "learning_rate": 1.5243729462179659e-05, "loss": 0.9552, "step": 9792 }, { "epoch": 0.34507606790172396, "grad_norm": 1.0205395221710205, "learning_rate": 1.5242757643726225e-05, "loss": 1.2421, "step": 9793 }, { "epoch": 0.3451113049146824, "grad_norm": 5.557827472686768, "learning_rate": 1.5241785756985206e-05, "loss": 3.7311, "step": 9794 }, { "epoch": 0.34514654192764077, "grad_norm": 2.9300436973571777, "learning_rate": 1.5240813801969258e-05, "loss": 3.3973, "step": 9795 }, { "epoch": 0.3451817789405992, "grad_norm": 1.5895439386367798, "learning_rate": 1.5239841778691042e-05, "loss": 1.2481, "step": 9796 }, { "epoch": 0.34521701595355764, "grad_norm": 0.891252875328064, "learning_rate": 1.523886968716322e-05, "loss": 0.8609, "step": 9797 }, { "epoch": 0.345252252966516, "grad_norm": 5.968738555908203, "learning_rate": 1.5237897527398455e-05, "loss": 4.47, "step": 9798 }, { "epoch": 0.34528748997947445, "grad_norm": 1.2490626573562622, "learning_rate": 1.5236925299409406e-05, "loss": 1.0839, "step": 9799 }, { "epoch": 0.3453227269924328, "grad_norm": 6.632968425750732, "learning_rate": 1.523595300320874e-05, "loss": 3.6894, "step": 9800 }, { "epoch": 0.34535796400539126, "grad_norm": 2.64521861076355, "learning_rate": 1.5234980638809114e-05, "loss": 3.055, "step": 9801 }, { "epoch": 0.3453932010183497, "grad_norm": 1.7968348264694214, "learning_rate": 1.52340082062232e-05, "loss": 0.7973, "step": 9802 }, { "epoch": 0.34542843803130807, "grad_norm": 5.0595011711120605, "learning_rate": 1.5233035705463666e-05, "loss": 3.8478, "step": 9803 }, { "epoch": 0.3454636750442665, "grad_norm": 1.812722086906433, "learning_rate": 1.5232063136543173e-05, "loss": 1.0557, "step": 9804 }, { "epoch": 0.3454989120572249, "grad_norm": 2.654813289642334, "learning_rate": 1.523109049947439e-05, "loss": 0.9993, "step": 9805 }, { "epoch": 0.3455341490701833, "grad_norm": 7.619201183319092, "learning_rate": 1.523011779426999e-05, "loss": 7.9999, "step": 9806 }, { "epoch": 0.34556938608314175, "grad_norm": 3.376688241958618, "learning_rate": 1.5229145020942637e-05, "loss": 3.07, "step": 9807 }, { "epoch": 0.3456046230961001, "grad_norm": 3.6485397815704346, "learning_rate": 1.5228172179505004e-05, "loss": 3.2177, "step": 9808 }, { "epoch": 0.34563986010905856, "grad_norm": 2.249314546585083, "learning_rate": 1.522719926996976e-05, "loss": 0.9141, "step": 9809 }, { "epoch": 0.345675097122017, "grad_norm": 1.5456880331039429, "learning_rate": 1.5226226292349586e-05, "loss": 1.009, "step": 9810 }, { "epoch": 0.34571033413497537, "grad_norm": 4.326352119445801, "learning_rate": 1.5225253246657146e-05, "loss": 6.2345, "step": 9811 }, { "epoch": 0.3457455711479338, "grad_norm": 2.2341115474700928, "learning_rate": 1.5224280132905115e-05, "loss": 0.8345, "step": 9812 }, { "epoch": 0.3457808081608922, "grad_norm": 8.71695613861084, "learning_rate": 1.5223306951106173e-05, "loss": 8.4156, "step": 9813 }, { "epoch": 0.3458160451738506, "grad_norm": 6.364613056182861, "learning_rate": 1.522233370127299e-05, "loss": 3.4439, "step": 9814 }, { "epoch": 0.34585128218680905, "grad_norm": 7.2721686363220215, "learning_rate": 1.5221360383418246e-05, "loss": 5.0723, "step": 9815 }, { "epoch": 0.34588651919976743, "grad_norm": 1.5760539770126343, "learning_rate": 1.522038699755462e-05, "loss": 1.0296, "step": 9816 }, { "epoch": 0.34592175621272586, "grad_norm": 1.5515681505203247, "learning_rate": 1.5219413543694783e-05, "loss": 1.4256, "step": 9817 }, { "epoch": 0.34595699322568424, "grad_norm": 1.5894403457641602, "learning_rate": 1.5218440021851423e-05, "loss": 1.1728, "step": 9818 }, { "epoch": 0.3459922302386427, "grad_norm": 4.622255325317383, "learning_rate": 1.521746643203722e-05, "loss": 3.3355, "step": 9819 }, { "epoch": 0.3460274672516011, "grad_norm": 7.080211639404297, "learning_rate": 1.5216492774264845e-05, "loss": 6.2623, "step": 9820 }, { "epoch": 0.3460627042645595, "grad_norm": 3.704216241836548, "learning_rate": 1.5215519048546996e-05, "loss": 2.9834, "step": 9821 }, { "epoch": 0.3460979412775179, "grad_norm": 4.009006023406982, "learning_rate": 1.5214545254896342e-05, "loss": 5.601, "step": 9822 }, { "epoch": 0.3461331782904763, "grad_norm": 1.3492622375488281, "learning_rate": 1.5213571393325572e-05, "loss": 1.0544, "step": 9823 }, { "epoch": 0.34616841530343473, "grad_norm": 0.7924183011054993, "learning_rate": 1.5212597463847369e-05, "loss": 0.8076, "step": 9824 }, { "epoch": 0.34620365231639316, "grad_norm": 1.257636547088623, "learning_rate": 1.5211623466474424e-05, "loss": 0.9079, "step": 9825 }, { "epoch": 0.34623888932935154, "grad_norm": 1.4547063112258911, "learning_rate": 1.5210649401219418e-05, "loss": 0.8474, "step": 9826 }, { "epoch": 0.34627412634231, "grad_norm": 4.744046211242676, "learning_rate": 1.520967526809504e-05, "loss": 3.3911, "step": 9827 }, { "epoch": 0.34630936335526835, "grad_norm": 4.861782550811768, "learning_rate": 1.5208701067113976e-05, "loss": 3.0099, "step": 9828 }, { "epoch": 0.3463446003682268, "grad_norm": 1.0118523836135864, "learning_rate": 1.5207726798288915e-05, "loss": 0.9603, "step": 9829 }, { "epoch": 0.3463798373811852, "grad_norm": 1.259814977645874, "learning_rate": 1.5206752461632555e-05, "loss": 0.9306, "step": 9830 }, { "epoch": 0.3464150743941436, "grad_norm": 6.147058010101318, "learning_rate": 1.5205778057157576e-05, "loss": 3.6245, "step": 9831 }, { "epoch": 0.34645031140710203, "grad_norm": 3.5034003257751465, "learning_rate": 1.5204803584876682e-05, "loss": 2.737, "step": 9832 }, { "epoch": 0.3464855484200604, "grad_norm": 1.2847775220870972, "learning_rate": 1.520382904480255e-05, "loss": 0.9837, "step": 9833 }, { "epoch": 0.34652078543301884, "grad_norm": 2.0196890830993652, "learning_rate": 1.5202854436947888e-05, "loss": 1.9671, "step": 9834 }, { "epoch": 0.3465560224459773, "grad_norm": 1.6902292966842651, "learning_rate": 1.520187976132538e-05, "loss": 0.7801, "step": 9835 }, { "epoch": 0.34659125945893565, "grad_norm": 2.100722312927246, "learning_rate": 1.5200905017947728e-05, "loss": 1.4053, "step": 9836 }, { "epoch": 0.3466264964718941, "grad_norm": 1.168619155883789, "learning_rate": 1.5199930206827623e-05, "loss": 1.1777, "step": 9837 }, { "epoch": 0.34666173348485246, "grad_norm": 5.284128665924072, "learning_rate": 1.5198955327977767e-05, "loss": 3.3858, "step": 9838 }, { "epoch": 0.3466969704978109, "grad_norm": 7.570098400115967, "learning_rate": 1.5197980381410853e-05, "loss": 3.1908, "step": 9839 }, { "epoch": 0.34673220751076933, "grad_norm": 4.031890392303467, "learning_rate": 1.5197005367139586e-05, "loss": 3.3241, "step": 9840 }, { "epoch": 0.3467674445237277, "grad_norm": 5.989130973815918, "learning_rate": 1.5196030285176661e-05, "loss": 3.9446, "step": 9841 }, { "epoch": 0.34680268153668614, "grad_norm": 3.6049935817718506, "learning_rate": 1.5195055135534777e-05, "loss": 3.4012, "step": 9842 }, { "epoch": 0.3468379185496445, "grad_norm": 3.4790027141571045, "learning_rate": 1.5194079918226642e-05, "loss": 3.2204, "step": 9843 }, { "epoch": 0.34687315556260295, "grad_norm": 1.8368881940841675, "learning_rate": 1.5193104633264952e-05, "loss": 0.7695, "step": 9844 }, { "epoch": 0.3469083925755614, "grad_norm": 4.380210876464844, "learning_rate": 1.5192129280662413e-05, "loss": 4.0468, "step": 9845 }, { "epoch": 0.34694362958851976, "grad_norm": 3.24177622795105, "learning_rate": 1.5191153860431729e-05, "loss": 3.9479, "step": 9846 }, { "epoch": 0.3469788666014782, "grad_norm": 1.237932562828064, "learning_rate": 1.5190178372585606e-05, "loss": 0.9927, "step": 9847 }, { "epoch": 0.34701410361443663, "grad_norm": 9.949136734008789, "learning_rate": 1.5189202817136747e-05, "loss": 2.5839, "step": 9848 }, { "epoch": 0.347049340627395, "grad_norm": 3.741650342941284, "learning_rate": 1.5188227194097863e-05, "loss": 3.3389, "step": 9849 }, { "epoch": 0.34708457764035344, "grad_norm": 1.5751416683197021, "learning_rate": 1.5187251503481654e-05, "loss": 1.1594, "step": 9850 }, { "epoch": 0.3471198146533118, "grad_norm": 1.0857105255126953, "learning_rate": 1.5186275745300837e-05, "loss": 0.7959, "step": 9851 }, { "epoch": 0.34715505166627025, "grad_norm": 1.572450041770935, "learning_rate": 1.5185299919568115e-05, "loss": 0.6797, "step": 9852 }, { "epoch": 0.3471902886792287, "grad_norm": 7.018896579742432, "learning_rate": 1.5184324026296204e-05, "loss": 3.6644, "step": 9853 }, { "epoch": 0.34722552569218706, "grad_norm": 5.364302158355713, "learning_rate": 1.5183348065497812e-05, "loss": 5.1189, "step": 9854 }, { "epoch": 0.3472607627051455, "grad_norm": 4.226381301879883, "learning_rate": 1.5182372037185648e-05, "loss": 3.4706, "step": 9855 }, { "epoch": 0.3472959997181039, "grad_norm": 1.3096853494644165, "learning_rate": 1.5181395941372432e-05, "loss": 1.0309, "step": 9856 }, { "epoch": 0.3473312367310623, "grad_norm": 5.795227527618408, "learning_rate": 1.5180419778070872e-05, "loss": 8.4061, "step": 9857 }, { "epoch": 0.34736647374402074, "grad_norm": 4.564550876617432, "learning_rate": 1.5179443547293685e-05, "loss": 3.7133, "step": 9858 }, { "epoch": 0.3474017107569791, "grad_norm": 1.296661138534546, "learning_rate": 1.5178467249053586e-05, "loss": 0.8476, "step": 9859 }, { "epoch": 0.34743694776993755, "grad_norm": 1.5851587057113647, "learning_rate": 1.517749088336329e-05, "loss": 0.948, "step": 9860 }, { "epoch": 0.34747218478289593, "grad_norm": 5.506128787994385, "learning_rate": 1.5176514450235517e-05, "loss": 3.6154, "step": 9861 }, { "epoch": 0.34750742179585437, "grad_norm": 2.4951131343841553, "learning_rate": 1.5175537949682983e-05, "loss": 3.4462, "step": 9862 }, { "epoch": 0.3475426588088128, "grad_norm": 5.075136184692383, "learning_rate": 1.5174561381718408e-05, "loss": 5.0452, "step": 9863 }, { "epoch": 0.3475778958217712, "grad_norm": 2.139110803604126, "learning_rate": 1.5173584746354512e-05, "loss": 1.0845, "step": 9864 }, { "epoch": 0.3476131328347296, "grad_norm": 5.855177402496338, "learning_rate": 1.5172608043604017e-05, "loss": 6.0479, "step": 9865 }, { "epoch": 0.347648369847688, "grad_norm": 1.349307656288147, "learning_rate": 1.5171631273479643e-05, "loss": 0.669, "step": 9866 }, { "epoch": 0.3476836068606464, "grad_norm": 1.1201379299163818, "learning_rate": 1.5170654435994112e-05, "loss": 1.0497, "step": 9867 }, { "epoch": 0.34771884387360485, "grad_norm": 1.3338440656661987, "learning_rate": 1.5169677531160152e-05, "loss": 0.8719, "step": 9868 }, { "epoch": 0.34775408088656323, "grad_norm": 1.1351596117019653, "learning_rate": 1.5168700558990477e-05, "loss": 0.9184, "step": 9869 }, { "epoch": 0.34778931789952167, "grad_norm": 1.6596711874008179, "learning_rate": 1.5167723519497826e-05, "loss": 0.8391, "step": 9870 }, { "epoch": 0.34782455491248004, "grad_norm": 1.06113862991333, "learning_rate": 1.516674641269491e-05, "loss": 0.8818, "step": 9871 }, { "epoch": 0.3478597919254385, "grad_norm": 9.084142684936523, "learning_rate": 1.516576923859447e-05, "loss": 8.3564, "step": 9872 }, { "epoch": 0.3478950289383969, "grad_norm": 3.4420394897460938, "learning_rate": 1.5164791997209227e-05, "loss": 3.2648, "step": 9873 }, { "epoch": 0.3479302659513553, "grad_norm": 2.7364542484283447, "learning_rate": 1.5163814688551908e-05, "loss": 3.3088, "step": 9874 }, { "epoch": 0.3479655029643137, "grad_norm": 4.793121814727783, "learning_rate": 1.5162837312635246e-05, "loss": 2.9939, "step": 9875 }, { "epoch": 0.3480007399772721, "grad_norm": 0.9511849880218506, "learning_rate": 1.516185986947197e-05, "loss": 0.89, "step": 9876 }, { "epoch": 0.34803597699023053, "grad_norm": 6.6659088134765625, "learning_rate": 1.5160882359074812e-05, "loss": 5.1928, "step": 9877 }, { "epoch": 0.34807121400318897, "grad_norm": 8.231704711914062, "learning_rate": 1.5159904781456506e-05, "loss": 6.6347, "step": 9878 }, { "epoch": 0.34810645101614734, "grad_norm": 9.016559600830078, "learning_rate": 1.5158927136629777e-05, "loss": 9.9531, "step": 9879 }, { "epoch": 0.3481416880291058, "grad_norm": 2.424865484237671, "learning_rate": 1.5157949424607372e-05, "loss": 2.7876, "step": 9880 }, { "epoch": 0.3481769250420642, "grad_norm": 4.817042827606201, "learning_rate": 1.5156971645402014e-05, "loss": 3.7683, "step": 9881 }, { "epoch": 0.3482121620550226, "grad_norm": 1.8914326429367065, "learning_rate": 1.5155993799026445e-05, "loss": 1.0489, "step": 9882 }, { "epoch": 0.348247399067981, "grad_norm": 4.083045959472656, "learning_rate": 1.51550158854934e-05, "loss": 3.3046, "step": 9883 }, { "epoch": 0.3482826360809394, "grad_norm": 2.929989814758301, "learning_rate": 1.5154037904815618e-05, "loss": 0.8272, "step": 9884 }, { "epoch": 0.34831787309389783, "grad_norm": 1.1419987678527832, "learning_rate": 1.5153059857005832e-05, "loss": 1.2644, "step": 9885 }, { "epoch": 0.34835311010685627, "grad_norm": 6.254290580749512, "learning_rate": 1.515208174207679e-05, "loss": 7.792, "step": 9886 }, { "epoch": 0.34838834711981465, "grad_norm": 1.2566556930541992, "learning_rate": 1.5151103560041223e-05, "loss": 1.077, "step": 9887 }, { "epoch": 0.3484235841327731, "grad_norm": 7.054565906524658, "learning_rate": 1.5150125310911876e-05, "loss": 6.1361, "step": 9888 }, { "epoch": 0.34845882114573146, "grad_norm": 3.2056689262390137, "learning_rate": 1.5149146994701493e-05, "loss": 0.9984, "step": 9889 }, { "epoch": 0.3484940581586899, "grad_norm": 0.9355632662773132, "learning_rate": 1.5148168611422812e-05, "loss": 1.2123, "step": 9890 }, { "epoch": 0.3485292951716483, "grad_norm": 3.5656938552856445, "learning_rate": 1.5147190161088582e-05, "loss": 3.2071, "step": 9891 }, { "epoch": 0.3485645321846067, "grad_norm": 1.0076665878295898, "learning_rate": 1.5146211643711544e-05, "loss": 0.8297, "step": 9892 }, { "epoch": 0.34859976919756513, "grad_norm": 1.1059335470199585, "learning_rate": 1.5145233059304443e-05, "loss": 1.1959, "step": 9893 }, { "epoch": 0.3486350062105235, "grad_norm": 1.3362654447555542, "learning_rate": 1.5144254407880024e-05, "loss": 1.0746, "step": 9894 }, { "epoch": 0.34867024322348195, "grad_norm": 5.164617538452148, "learning_rate": 1.514327568945104e-05, "loss": 3.2184, "step": 9895 }, { "epoch": 0.3487054802364404, "grad_norm": 1.402260184288025, "learning_rate": 1.5142296904030231e-05, "loss": 1.2381, "step": 9896 }, { "epoch": 0.34874071724939876, "grad_norm": 5.84577751159668, "learning_rate": 1.5141318051630354e-05, "loss": 7.9137, "step": 9897 }, { "epoch": 0.3487759542623572, "grad_norm": 1.3461639881134033, "learning_rate": 1.514033913226415e-05, "loss": 1.0396, "step": 9898 }, { "epoch": 0.34881119127531557, "grad_norm": 3.291018486022949, "learning_rate": 1.5139360145944377e-05, "loss": 3.0457, "step": 9899 }, { "epoch": 0.348846428288274, "grad_norm": 1.8563361167907715, "learning_rate": 1.5138381092683782e-05, "loss": 0.9319, "step": 9900 }, { "epoch": 0.34888166530123244, "grad_norm": 5.382297039031982, "learning_rate": 1.5137401972495118e-05, "loss": 5.9282, "step": 9901 }, { "epoch": 0.3489169023141908, "grad_norm": 6.337986946105957, "learning_rate": 1.513642278539114e-05, "loss": 6.3621, "step": 9902 }, { "epoch": 0.34895213932714925, "grad_norm": 0.9946587085723877, "learning_rate": 1.5135443531384603e-05, "loss": 0.797, "step": 9903 }, { "epoch": 0.3489873763401076, "grad_norm": 3.7661290168762207, "learning_rate": 1.5134464210488257e-05, "loss": 2.9569, "step": 9904 }, { "epoch": 0.34902261335306606, "grad_norm": 2.009685516357422, "learning_rate": 1.5133484822714861e-05, "loss": 1.0023, "step": 9905 }, { "epoch": 0.3490578503660245, "grad_norm": 3.482520818710327, "learning_rate": 1.5132505368077173e-05, "loss": 3.2157, "step": 9906 }, { "epoch": 0.34909308737898287, "grad_norm": 5.998444557189941, "learning_rate": 1.5131525846587947e-05, "loss": 3.3967, "step": 9907 }, { "epoch": 0.3491283243919413, "grad_norm": 1.4448288679122925, "learning_rate": 1.5130546258259947e-05, "loss": 0.8343, "step": 9908 }, { "epoch": 0.3491635614048997, "grad_norm": 4.603280067443848, "learning_rate": 1.5129566603105926e-05, "loss": 3.3401, "step": 9909 }, { "epoch": 0.3491987984178581, "grad_norm": 1.0330486297607422, "learning_rate": 1.5128586881138645e-05, "loss": 1.0636, "step": 9910 }, { "epoch": 0.34923403543081655, "grad_norm": 0.8929556608200073, "learning_rate": 1.5127607092370865e-05, "loss": 0.7111, "step": 9911 }, { "epoch": 0.3492692724437749, "grad_norm": 1.064985990524292, "learning_rate": 1.5126627236815354e-05, "loss": 0.7787, "step": 9912 }, { "epoch": 0.34930450945673336, "grad_norm": 1.3112778663635254, "learning_rate": 1.5125647314484868e-05, "loss": 1.1131, "step": 9913 }, { "epoch": 0.34933974646969174, "grad_norm": 3.5628480911254883, "learning_rate": 1.5124667325392175e-05, "loss": 2.9634, "step": 9914 }, { "epoch": 0.34937498348265017, "grad_norm": 0.9629765152931213, "learning_rate": 1.5123687269550034e-05, "loss": 0.9507, "step": 9915 }, { "epoch": 0.3494102204956086, "grad_norm": 4.911309242248535, "learning_rate": 1.5122707146971217e-05, "loss": 3.8616, "step": 9916 }, { "epoch": 0.349445457508567, "grad_norm": 1.6550856828689575, "learning_rate": 1.5121726957668483e-05, "loss": 0.74, "step": 9917 }, { "epoch": 0.3494806945215254, "grad_norm": 3.9175972938537598, "learning_rate": 1.5120746701654607e-05, "loss": 3.526, "step": 9918 }, { "epoch": 0.34951593153448385, "grad_norm": 1.4465082883834839, "learning_rate": 1.5119766378942352e-05, "loss": 0.9233, "step": 9919 }, { "epoch": 0.3495511685474422, "grad_norm": 1.245729923248291, "learning_rate": 1.5118785989544485e-05, "loss": 0.8431, "step": 9920 }, { "epoch": 0.34958640556040066, "grad_norm": 4.101963996887207, "learning_rate": 1.511780553347378e-05, "loss": 3.6396, "step": 9921 }, { "epoch": 0.34962164257335904, "grad_norm": 1.343373417854309, "learning_rate": 1.5116825010743008e-05, "loss": 0.9929, "step": 9922 }, { "epoch": 0.34965687958631747, "grad_norm": 6.620779037475586, "learning_rate": 1.5115844421364934e-05, "loss": 3.2315, "step": 9923 }, { "epoch": 0.3496921165992759, "grad_norm": 5.121920108795166, "learning_rate": 1.5114863765352339e-05, "loss": 3.8964, "step": 9924 }, { "epoch": 0.3497273536122343, "grad_norm": 5.369455814361572, "learning_rate": 1.5113883042717987e-05, "loss": 6.1568, "step": 9925 }, { "epoch": 0.3497625906251927, "grad_norm": 6.135120868682861, "learning_rate": 1.5112902253474663e-05, "loss": 6.2156, "step": 9926 }, { "epoch": 0.3497978276381511, "grad_norm": 5.135782241821289, "learning_rate": 1.5111921397635132e-05, "loss": 5.7997, "step": 9927 }, { "epoch": 0.3498330646511095, "grad_norm": 0.9945239424705505, "learning_rate": 1.5110940475212174e-05, "loss": 0.9664, "step": 9928 }, { "epoch": 0.34986830166406796, "grad_norm": 3.3600387573242188, "learning_rate": 1.5109959486218566e-05, "loss": 3.3126, "step": 9929 }, { "epoch": 0.34990353867702634, "grad_norm": 1.051452875137329, "learning_rate": 1.5108978430667085e-05, "loss": 1.1744, "step": 9930 }, { "epoch": 0.34993877568998477, "grad_norm": 1.7876380681991577, "learning_rate": 1.510799730857051e-05, "loss": 0.8553, "step": 9931 }, { "epoch": 0.34997401270294315, "grad_norm": 1.4293540716171265, "learning_rate": 1.5107016119941617e-05, "loss": 0.9793, "step": 9932 }, { "epoch": 0.3500092497159016, "grad_norm": 7.818338394165039, "learning_rate": 1.510603486479319e-05, "loss": 4.1005, "step": 9933 }, { "epoch": 0.35004448672886, "grad_norm": 10.007065773010254, "learning_rate": 1.5105053543138009e-05, "loss": 6.3411, "step": 9934 }, { "epoch": 0.3500797237418184, "grad_norm": 1.501001238822937, "learning_rate": 1.5104072154988855e-05, "loss": 0.8891, "step": 9935 }, { "epoch": 0.3501149607547768, "grad_norm": 1.2164167165756226, "learning_rate": 1.5103090700358508e-05, "loss": 0.8688, "step": 9936 }, { "epoch": 0.3501501977677352, "grad_norm": 1.685562252998352, "learning_rate": 1.5102109179259761e-05, "loss": 1.0178, "step": 9937 }, { "epoch": 0.35018543478069364, "grad_norm": 4.315845012664795, "learning_rate": 1.510112759170539e-05, "loss": 3.2374, "step": 9938 }, { "epoch": 0.35022067179365207, "grad_norm": 0.9727911353111267, "learning_rate": 1.510014593770818e-05, "loss": 1.0815, "step": 9939 }, { "epoch": 0.35025590880661045, "grad_norm": 6.614535331726074, "learning_rate": 1.5099164217280918e-05, "loss": 6.146, "step": 9940 }, { "epoch": 0.3502911458195689, "grad_norm": 1.5539307594299316, "learning_rate": 1.5098182430436397e-05, "loss": 0.84, "step": 9941 }, { "epoch": 0.35032638283252726, "grad_norm": 3.0958001613616943, "learning_rate": 1.5097200577187397e-05, "loss": 2.9322, "step": 9942 }, { "epoch": 0.3503616198454857, "grad_norm": 2.089344024658203, "learning_rate": 1.5096218657546712e-05, "loss": 0.9735, "step": 9943 }, { "epoch": 0.35039685685844413, "grad_norm": 1.1709911823272705, "learning_rate": 1.5095236671527127e-05, "loss": 0.8588, "step": 9944 }, { "epoch": 0.3504320938714025, "grad_norm": 6.887359619140625, "learning_rate": 1.509425461914144e-05, "loss": 6.0639, "step": 9945 }, { "epoch": 0.35046733088436094, "grad_norm": 5.611464023590088, "learning_rate": 1.5093272500402434e-05, "loss": 3.3477, "step": 9946 }, { "epoch": 0.3505025678973193, "grad_norm": 5.139013290405273, "learning_rate": 1.5092290315322906e-05, "loss": 5.4903, "step": 9947 }, { "epoch": 0.35053780491027775, "grad_norm": 5.136908054351807, "learning_rate": 1.5091308063915649e-05, "loss": 3.0675, "step": 9948 }, { "epoch": 0.3505730419232362, "grad_norm": 1.148047924041748, "learning_rate": 1.5090325746193455e-05, "loss": 0.8751, "step": 9949 }, { "epoch": 0.35060827893619456, "grad_norm": 1.0775730609893799, "learning_rate": 1.5089343362169117e-05, "loss": 1.0278, "step": 9950 }, { "epoch": 0.350643515949153, "grad_norm": 1.56417977809906, "learning_rate": 1.5088360911855437e-05, "loss": 0.8253, "step": 9951 }, { "epoch": 0.35067875296211143, "grad_norm": 2.0044093132019043, "learning_rate": 1.5087378395265208e-05, "loss": 1.0513, "step": 9952 }, { "epoch": 0.3507139899750698, "grad_norm": 12.112685203552246, "learning_rate": 1.5086395812411226e-05, "loss": 9.5456, "step": 9953 }, { "epoch": 0.35074922698802824, "grad_norm": 5.598206996917725, "learning_rate": 1.5085413163306292e-05, "loss": 2.8721, "step": 9954 }, { "epoch": 0.3507844640009866, "grad_norm": 1.2971421480178833, "learning_rate": 1.5084430447963203e-05, "loss": 1.0567, "step": 9955 }, { "epoch": 0.35081970101394505, "grad_norm": 3.5166618824005127, "learning_rate": 1.508344766639476e-05, "loss": 3.0791, "step": 9956 }, { "epoch": 0.3508549380269035, "grad_norm": 4.888396263122559, "learning_rate": 1.5082464818613765e-05, "loss": 3.0067, "step": 9957 }, { "epoch": 0.35089017503986186, "grad_norm": 3.921187400817871, "learning_rate": 1.5081481904633015e-05, "loss": 3.3711, "step": 9958 }, { "epoch": 0.3509254120528203, "grad_norm": 1.1235326528549194, "learning_rate": 1.5080498924465316e-05, "loss": 0.8516, "step": 9959 }, { "epoch": 0.3509606490657787, "grad_norm": 0.9952253103256226, "learning_rate": 1.5079515878123476e-05, "loss": 1.0075, "step": 9960 }, { "epoch": 0.3509958860787371, "grad_norm": 4.171586990356445, "learning_rate": 1.5078532765620288e-05, "loss": 2.9856, "step": 9961 }, { "epoch": 0.35103112309169554, "grad_norm": 24.52605628967285, "learning_rate": 1.5077549586968569e-05, "loss": 3.7534, "step": 9962 }, { "epoch": 0.3510663601046539, "grad_norm": 1.3793524503707886, "learning_rate": 1.5076566342181118e-05, "loss": 0.974, "step": 9963 }, { "epoch": 0.35110159711761235, "grad_norm": 15.11178207397461, "learning_rate": 1.5075583031270746e-05, "loss": 5.7197, "step": 9964 }, { "epoch": 0.35113683413057073, "grad_norm": 1.4099397659301758, "learning_rate": 1.5074599654250258e-05, "loss": 1.0748, "step": 9965 }, { "epoch": 0.35117207114352916, "grad_norm": 1.6252589225769043, "learning_rate": 1.5073616211132459e-05, "loss": 1.041, "step": 9966 }, { "epoch": 0.3512073081564876, "grad_norm": 1.383670687675476, "learning_rate": 1.5072632701930165e-05, "loss": 1.0157, "step": 9967 }, { "epoch": 0.351242545169446, "grad_norm": 1.7684619426727295, "learning_rate": 1.5071649126656185e-05, "loss": 0.8504, "step": 9968 }, { "epoch": 0.3512777821824044, "grad_norm": 3.108144760131836, "learning_rate": 1.5070665485323328e-05, "loss": 3.3612, "step": 9969 }, { "epoch": 0.3513130191953628, "grad_norm": 1.9274901151657104, "learning_rate": 1.5069681777944407e-05, "loss": 1.0521, "step": 9970 }, { "epoch": 0.3513482562083212, "grad_norm": 1.1902568340301514, "learning_rate": 1.5068698004532234e-05, "loss": 1.0135, "step": 9971 }, { "epoch": 0.35138349322127965, "grad_norm": 4.331027984619141, "learning_rate": 1.5067714165099627e-05, "loss": 3.0177, "step": 9972 }, { "epoch": 0.35141873023423803, "grad_norm": 1.792042851448059, "learning_rate": 1.5066730259659396e-05, "loss": 0.8169, "step": 9973 }, { "epoch": 0.35145396724719646, "grad_norm": 2.921205997467041, "learning_rate": 1.5065746288224356e-05, "loss": 0.9527, "step": 9974 }, { "epoch": 0.35148920426015484, "grad_norm": 0.9671985507011414, "learning_rate": 1.506476225080733e-05, "loss": 1.2373, "step": 9975 }, { "epoch": 0.3515244412731133, "grad_norm": 1.2015478610992432, "learning_rate": 1.5063778147421127e-05, "loss": 0.87, "step": 9976 }, { "epoch": 0.3515596782860717, "grad_norm": 1.5213801860809326, "learning_rate": 1.5062793978078567e-05, "loss": 1.1031, "step": 9977 }, { "epoch": 0.3515949152990301, "grad_norm": 5.549506187438965, "learning_rate": 1.5061809742792473e-05, "loss": 3.8035, "step": 9978 }, { "epoch": 0.3516301523119885, "grad_norm": 1.3245320320129395, "learning_rate": 1.5060825441575663e-05, "loss": 0.9457, "step": 9979 }, { "epoch": 0.3516653893249469, "grad_norm": 1.7698814868927002, "learning_rate": 1.5059841074440955e-05, "loss": 1.0467, "step": 9980 }, { "epoch": 0.35170062633790533, "grad_norm": 1.064348816871643, "learning_rate": 1.5058856641401176e-05, "loss": 0.9974, "step": 9981 }, { "epoch": 0.35173586335086376, "grad_norm": 2.0084681510925293, "learning_rate": 1.505787214246914e-05, "loss": 0.968, "step": 9982 }, { "epoch": 0.35177110036382214, "grad_norm": 1.422412633895874, "learning_rate": 1.5056887577657679e-05, "loss": 0.8211, "step": 9983 }, { "epoch": 0.3518063373767806, "grad_norm": 1.5248324871063232, "learning_rate": 1.5055902946979611e-05, "loss": 1.0441, "step": 9984 }, { "epoch": 0.351841574389739, "grad_norm": 5.897119522094727, "learning_rate": 1.5054918250447762e-05, "loss": 6.1946, "step": 9985 }, { "epoch": 0.3518768114026974, "grad_norm": 1.0557656288146973, "learning_rate": 1.505393348807496e-05, "loss": 1.1818, "step": 9986 }, { "epoch": 0.3519120484156558, "grad_norm": 1.6289167404174805, "learning_rate": 1.5052948659874034e-05, "loss": 0.9274, "step": 9987 }, { "epoch": 0.3519472854286142, "grad_norm": 5.150247573852539, "learning_rate": 1.5051963765857804e-05, "loss": 3.0175, "step": 9988 }, { "epoch": 0.35198252244157263, "grad_norm": 5.745678901672363, "learning_rate": 1.5050978806039104e-05, "loss": 3.7401, "step": 9989 }, { "epoch": 0.35201775945453107, "grad_norm": 4.442953586578369, "learning_rate": 1.5049993780430761e-05, "loss": 3.2351, "step": 9990 }, { "epoch": 0.35205299646748944, "grad_norm": 4.713006496429443, "learning_rate": 1.5049008689045605e-05, "loss": 5.6633, "step": 9991 }, { "epoch": 0.3520882334804479, "grad_norm": 2.85789155960083, "learning_rate": 1.5048023531896472e-05, "loss": 3.331, "step": 9992 }, { "epoch": 0.35212347049340625, "grad_norm": 6.292925834655762, "learning_rate": 1.5047038308996182e-05, "loss": 3.2966, "step": 9993 }, { "epoch": 0.3521587075063647, "grad_norm": 0.9125822186470032, "learning_rate": 1.5046053020357583e-05, "loss": 0.9572, "step": 9994 }, { "epoch": 0.3521939445193231, "grad_norm": 1.568410873413086, "learning_rate": 1.5045067665993496e-05, "loss": 1.1613, "step": 9995 }, { "epoch": 0.3522291815322815, "grad_norm": 3.880326747894287, "learning_rate": 1.5044082245916763e-05, "loss": 3.3303, "step": 9996 }, { "epoch": 0.35226441854523993, "grad_norm": 1.2628411054611206, "learning_rate": 1.5043096760140212e-05, "loss": 1.1091, "step": 9997 }, { "epoch": 0.3522996555581983, "grad_norm": 1.2524133920669556, "learning_rate": 1.5042111208676687e-05, "loss": 0.7006, "step": 9998 }, { "epoch": 0.35233489257115674, "grad_norm": 6.328972339630127, "learning_rate": 1.5041125591539017e-05, "loss": 3.6586, "step": 9999 }, { "epoch": 0.3523701295841152, "grad_norm": 5.2468037605285645, "learning_rate": 1.504013990874005e-05, "loss": 3.3959, "step": 10000 }, { "epoch": 0.35240536659707355, "grad_norm": 6.2516069412231445, "learning_rate": 1.5039154160292612e-05, "loss": 5.0208, "step": 10001 }, { "epoch": 0.352440603610032, "grad_norm": 3.381458044052124, "learning_rate": 1.5038168346209556e-05, "loss": 3.0595, "step": 10002 }, { "epoch": 0.35247584062299037, "grad_norm": 1.2067172527313232, "learning_rate": 1.503718246650371e-05, "loss": 0.828, "step": 10003 }, { "epoch": 0.3525110776359488, "grad_norm": 1.1671595573425293, "learning_rate": 1.5036196521187925e-05, "loss": 0.7843, "step": 10004 }, { "epoch": 0.35254631464890723, "grad_norm": 4.4594197273254395, "learning_rate": 1.5035210510275035e-05, "loss": 3.1206, "step": 10005 }, { "epoch": 0.3525815516618656, "grad_norm": 5.8375020027160645, "learning_rate": 1.5034224433777888e-05, "loss": 7.9727, "step": 10006 }, { "epoch": 0.35261678867482404, "grad_norm": 1.7219953536987305, "learning_rate": 1.5033238291709329e-05, "loss": 0.6844, "step": 10007 }, { "epoch": 0.3526520256877824, "grad_norm": 2.21779203414917, "learning_rate": 1.5032252084082198e-05, "loss": 2.5177, "step": 10008 }, { "epoch": 0.35268726270074086, "grad_norm": 1.6628193855285645, "learning_rate": 1.503126581090934e-05, "loss": 0.8798, "step": 10009 }, { "epoch": 0.3527224997136993, "grad_norm": 1.0328325033187866, "learning_rate": 1.5030279472203609e-05, "loss": 0.8384, "step": 10010 }, { "epoch": 0.35275773672665767, "grad_norm": 0.9156078696250916, "learning_rate": 1.5029293067977846e-05, "loss": 1.1185, "step": 10011 }, { "epoch": 0.3527929737396161, "grad_norm": 1.0364891290664673, "learning_rate": 1.5028306598244898e-05, "loss": 0.8514, "step": 10012 }, { "epoch": 0.3528282107525745, "grad_norm": 4.959316253662109, "learning_rate": 1.5027320063017617e-05, "loss": 2.5682, "step": 10013 }, { "epoch": 0.3528634477655329, "grad_norm": 5.4613752365112305, "learning_rate": 1.502633346230885e-05, "loss": 2.9309, "step": 10014 }, { "epoch": 0.35289868477849134, "grad_norm": 2.985022783279419, "learning_rate": 1.5025346796131452e-05, "loss": 2.89, "step": 10015 }, { "epoch": 0.3529339217914497, "grad_norm": 2.0166397094726562, "learning_rate": 1.502436006449827e-05, "loss": 0.723, "step": 10016 }, { "epoch": 0.35296915880440816, "grad_norm": 5.472885608673096, "learning_rate": 1.502337326742216e-05, "loss": 3.295, "step": 10017 }, { "epoch": 0.35300439581736653, "grad_norm": 1.078383207321167, "learning_rate": 1.5022386404915971e-05, "loss": 1.0302, "step": 10018 }, { "epoch": 0.35303963283032497, "grad_norm": 1.5040993690490723, "learning_rate": 1.5021399476992563e-05, "loss": 0.9559, "step": 10019 }, { "epoch": 0.3530748698432834, "grad_norm": 6.770705699920654, "learning_rate": 1.5020412483664784e-05, "loss": 5.3897, "step": 10020 }, { "epoch": 0.3531101068562418, "grad_norm": 1.5933314561843872, "learning_rate": 1.5019425424945495e-05, "loss": 1.1793, "step": 10021 }, { "epoch": 0.3531453438692002, "grad_norm": 6.531815052032471, "learning_rate": 1.5018438300847548e-05, "loss": 5.3955, "step": 10022 }, { "epoch": 0.35318058088215865, "grad_norm": 1.4576431512832642, "learning_rate": 1.5017451111383805e-05, "loss": 1.0465, "step": 10023 }, { "epoch": 0.353215817895117, "grad_norm": 2.6412904262542725, "learning_rate": 1.5016463856567121e-05, "loss": 3.4516, "step": 10024 }, { "epoch": 0.35325105490807546, "grad_norm": 2.9011754989624023, "learning_rate": 1.5015476536410356e-05, "loss": 3.2095, "step": 10025 }, { "epoch": 0.35328629192103383, "grad_norm": 8.002055168151855, "learning_rate": 1.5014489150926372e-05, "loss": 5.0909, "step": 10026 }, { "epoch": 0.35332152893399227, "grad_norm": 4.361617565155029, "learning_rate": 1.5013501700128026e-05, "loss": 3.3385, "step": 10027 }, { "epoch": 0.3533567659469507, "grad_norm": 1.1461291313171387, "learning_rate": 1.5012514184028181e-05, "loss": 0.6814, "step": 10028 }, { "epoch": 0.3533920029599091, "grad_norm": 1.4097299575805664, "learning_rate": 1.5011526602639703e-05, "loss": 0.8013, "step": 10029 }, { "epoch": 0.3534272399728675, "grad_norm": 1.2815791368484497, "learning_rate": 1.5010538955975454e-05, "loss": 1.0634, "step": 10030 }, { "epoch": 0.3534624769858259, "grad_norm": 4.098683834075928, "learning_rate": 1.500955124404829e-05, "loss": 3.2595, "step": 10031 }, { "epoch": 0.3534977139987843, "grad_norm": 4.4659271240234375, "learning_rate": 1.500856346687109e-05, "loss": 3.1666, "step": 10032 }, { "epoch": 0.35353295101174276, "grad_norm": 8.200209617614746, "learning_rate": 1.5007575624456712e-05, "loss": 3.3076, "step": 10033 }, { "epoch": 0.35356818802470114, "grad_norm": 5.623754978179932, "learning_rate": 1.5006587716818022e-05, "loss": 4.8831, "step": 10034 }, { "epoch": 0.35360342503765957, "grad_norm": 4.441291332244873, "learning_rate": 1.5005599743967887e-05, "loss": 2.6349, "step": 10035 }, { "epoch": 0.35363866205061795, "grad_norm": 1.0193712711334229, "learning_rate": 1.5004611705919181e-05, "loss": 1.1285, "step": 10036 }, { "epoch": 0.3536738990635764, "grad_norm": 5.5840559005737305, "learning_rate": 1.500362360268477e-05, "loss": 3.1098, "step": 10037 }, { "epoch": 0.3537091360765348, "grad_norm": 5.978540420532227, "learning_rate": 1.500263543427752e-05, "loss": 3.4727, "step": 10038 }, { "epoch": 0.3537443730894932, "grad_norm": 2.6077280044555664, "learning_rate": 1.500164720071031e-05, "loss": 0.9516, "step": 10039 }, { "epoch": 0.3537796101024516, "grad_norm": 7.487827301025391, "learning_rate": 1.5000658901996011e-05, "loss": 3.3331, "step": 10040 }, { "epoch": 0.35381484711541, "grad_norm": 13.561300277709961, "learning_rate": 1.4999670538147486e-05, "loss": 8.6365, "step": 10041 }, { "epoch": 0.35385008412836844, "grad_norm": 3.011629343032837, "learning_rate": 1.499868210917762e-05, "loss": 3.0413, "step": 10042 }, { "epoch": 0.35388532114132687, "grad_norm": 5.242587089538574, "learning_rate": 1.499769361509928e-05, "loss": 3.5872, "step": 10043 }, { "epoch": 0.35392055815428525, "grad_norm": 1.570176362991333, "learning_rate": 1.4996705055925347e-05, "loss": 0.7767, "step": 10044 }, { "epoch": 0.3539557951672437, "grad_norm": 3.498947858810425, "learning_rate": 1.4995716431668691e-05, "loss": 3.0594, "step": 10045 }, { "epoch": 0.35399103218020206, "grad_norm": 4.14881706237793, "learning_rate": 1.4994727742342195e-05, "loss": 3.5085, "step": 10046 }, { "epoch": 0.3540262691931605, "grad_norm": 1.60186767578125, "learning_rate": 1.4993738987958732e-05, "loss": 0.7758, "step": 10047 }, { "epoch": 0.3540615062061189, "grad_norm": 5.904101848602295, "learning_rate": 1.4992750168531182e-05, "loss": 2.4802, "step": 10048 }, { "epoch": 0.3540967432190773, "grad_norm": 4.949439525604248, "learning_rate": 1.499176128407243e-05, "loss": 3.1448, "step": 10049 }, { "epoch": 0.35413198023203574, "grad_norm": 3.2103452682495117, "learning_rate": 1.4990772334595345e-05, "loss": 2.8308, "step": 10050 }, { "epoch": 0.3541672172449941, "grad_norm": 1.4661134481430054, "learning_rate": 1.4989783320112818e-05, "loss": 0.9009, "step": 10051 }, { "epoch": 0.35420245425795255, "grad_norm": 1.9592235088348389, "learning_rate": 1.4988794240637727e-05, "loss": 0.7475, "step": 10052 }, { "epoch": 0.354237691270911, "grad_norm": 6.43533182144165, "learning_rate": 1.4987805096182957e-05, "loss": 5.5211, "step": 10053 }, { "epoch": 0.35427292828386936, "grad_norm": 1.7755006551742554, "learning_rate": 1.4986815886761387e-05, "loss": 0.9239, "step": 10054 }, { "epoch": 0.3543081652968278, "grad_norm": 7.168790340423584, "learning_rate": 1.4985826612385907e-05, "loss": 3.2877, "step": 10055 }, { "epoch": 0.3543434023097862, "grad_norm": 5.494676113128662, "learning_rate": 1.49848372730694e-05, "loss": 3.56, "step": 10056 }, { "epoch": 0.3543786393227446, "grad_norm": 0.9209398031234741, "learning_rate": 1.4983847868824754e-05, "loss": 0.7265, "step": 10057 }, { "epoch": 0.35441387633570304, "grad_norm": 3.711095094680786, "learning_rate": 1.4982858399664853e-05, "loss": 3.5149, "step": 10058 }, { "epoch": 0.3544491133486614, "grad_norm": 1.2942957878112793, "learning_rate": 1.4981868865602587e-05, "loss": 1.0138, "step": 10059 }, { "epoch": 0.35448435036161985, "grad_norm": 3.9295654296875, "learning_rate": 1.4980879266650846e-05, "loss": 3.18, "step": 10060 }, { "epoch": 0.3545195873745783, "grad_norm": 5.111771583557129, "learning_rate": 1.4979889602822515e-05, "loss": 3.5271, "step": 10061 }, { "epoch": 0.35455482438753666, "grad_norm": 6.810529708862305, "learning_rate": 1.4978899874130489e-05, "loss": 4.4906, "step": 10062 }, { "epoch": 0.3545900614004951, "grad_norm": 1.6418019533157349, "learning_rate": 1.4977910080587657e-05, "loss": 0.9738, "step": 10063 }, { "epoch": 0.35462529841345347, "grad_norm": 3.6172802448272705, "learning_rate": 1.4976920222206914e-05, "loss": 3.1233, "step": 10064 }, { "epoch": 0.3546605354264119, "grad_norm": 2.052769899368286, "learning_rate": 1.4975930299001149e-05, "loss": 0.6665, "step": 10065 }, { "epoch": 0.35469577243937034, "grad_norm": 12.03166675567627, "learning_rate": 1.497494031098326e-05, "loss": 3.654, "step": 10066 }, { "epoch": 0.3547310094523287, "grad_norm": 3.430392026901245, "learning_rate": 1.497395025816614e-05, "loss": 2.7795, "step": 10067 }, { "epoch": 0.35476624646528715, "grad_norm": 2.045569896697998, "learning_rate": 1.4972960140562683e-05, "loss": 1.0011, "step": 10068 }, { "epoch": 0.3548014834782455, "grad_norm": 3.428434371948242, "learning_rate": 1.4971969958185786e-05, "loss": 2.5064, "step": 10069 }, { "epoch": 0.35483672049120396, "grad_norm": 3.8881418704986572, "learning_rate": 1.4970979711048349e-05, "loss": 3.2618, "step": 10070 }, { "epoch": 0.3548719575041624, "grad_norm": 1.2617700099945068, "learning_rate": 1.4969989399163267e-05, "loss": 1.0902, "step": 10071 }, { "epoch": 0.35490719451712077, "grad_norm": 3.3348193168640137, "learning_rate": 1.4968999022543437e-05, "loss": 2.8799, "step": 10072 }, { "epoch": 0.3549424315300792, "grad_norm": 1.0985851287841797, "learning_rate": 1.4968008581201767e-05, "loss": 0.8249, "step": 10073 }, { "epoch": 0.3549776685430376, "grad_norm": 2.1740071773529053, "learning_rate": 1.496701807515115e-05, "loss": 0.9823, "step": 10074 }, { "epoch": 0.355012905555996, "grad_norm": 2.1924967765808105, "learning_rate": 1.496602750440449e-05, "loss": 3.4904, "step": 10075 }, { "epoch": 0.35504814256895445, "grad_norm": 5.448268890380859, "learning_rate": 1.4965036868974689e-05, "loss": 3.122, "step": 10076 }, { "epoch": 0.35508337958191283, "grad_norm": 1.6623245477676392, "learning_rate": 1.4964046168874651e-05, "loss": 0.8775, "step": 10077 }, { "epoch": 0.35511861659487126, "grad_norm": 3.0242691040039062, "learning_rate": 1.4963055404117279e-05, "loss": 3.1488, "step": 10078 }, { "epoch": 0.35515385360782964, "grad_norm": 5.404587268829346, "learning_rate": 1.4962064574715478e-05, "loss": 3.4221, "step": 10079 }, { "epoch": 0.3551890906207881, "grad_norm": 2.1982154846191406, "learning_rate": 1.4961073680682154e-05, "loss": 0.9738, "step": 10080 }, { "epoch": 0.3552243276337465, "grad_norm": 7.56168270111084, "learning_rate": 1.4960082722030214e-05, "loss": 4.4546, "step": 10081 }, { "epoch": 0.3552595646467049, "grad_norm": 3.06474232673645, "learning_rate": 1.4959091698772563e-05, "loss": 2.9262, "step": 10082 }, { "epoch": 0.3552948016596633, "grad_norm": 4.080779075622559, "learning_rate": 1.4958100610922113e-05, "loss": 3.2979, "step": 10083 }, { "epoch": 0.3553300386726217, "grad_norm": 4.454233646392822, "learning_rate": 1.4957109458491773e-05, "loss": 3.7924, "step": 10084 }, { "epoch": 0.35536527568558013, "grad_norm": 4.398515701293945, "learning_rate": 1.4956118241494445e-05, "loss": 5.8432, "step": 10085 }, { "epoch": 0.35540051269853856, "grad_norm": 3.6350080966949463, "learning_rate": 1.4955126959943049e-05, "loss": 3.0654, "step": 10086 }, { "epoch": 0.35543574971149694, "grad_norm": 7.2072649002075195, "learning_rate": 1.4954135613850493e-05, "loss": 3.305, "step": 10087 }, { "epoch": 0.3554709867244554, "grad_norm": 6.100956916809082, "learning_rate": 1.4953144203229689e-05, "loss": 6.3811, "step": 10088 }, { "epoch": 0.3555062237374138, "grad_norm": 2.638345241546631, "learning_rate": 1.4952152728093549e-05, "loss": 3.0576, "step": 10089 }, { "epoch": 0.3555414607503722, "grad_norm": 2.436833381652832, "learning_rate": 1.4951161188454992e-05, "loss": 2.3334, "step": 10090 }, { "epoch": 0.3555766977633306, "grad_norm": 3.834577798843384, "learning_rate": 1.4950169584326929e-05, "loss": 3.5158, "step": 10091 }, { "epoch": 0.355611934776289, "grad_norm": 3.6438236236572266, "learning_rate": 1.4949177915722276e-05, "loss": 3.4184, "step": 10092 }, { "epoch": 0.35564717178924743, "grad_norm": 3.505976438522339, "learning_rate": 1.494818618265395e-05, "loss": 2.7332, "step": 10093 }, { "epoch": 0.35568240880220586, "grad_norm": 1.513721227645874, "learning_rate": 1.4947194385134868e-05, "loss": 1.09, "step": 10094 }, { "epoch": 0.35571764581516424, "grad_norm": 6.519976615905762, "learning_rate": 1.4946202523177953e-05, "loss": 3.5864, "step": 10095 }, { "epoch": 0.3557528828281227, "grad_norm": 3.8844082355499268, "learning_rate": 1.4945210596796114e-05, "loss": 3.1251, "step": 10096 }, { "epoch": 0.35578811984108105, "grad_norm": 1.2783628702163696, "learning_rate": 1.4944218606002283e-05, "loss": 1.0372, "step": 10097 }, { "epoch": 0.3558233568540395, "grad_norm": 8.390049934387207, "learning_rate": 1.4943226550809372e-05, "loss": 7.9905, "step": 10098 }, { "epoch": 0.3558585938669979, "grad_norm": 2.752910614013672, "learning_rate": 1.4942234431230305e-05, "loss": 3.0655, "step": 10099 }, { "epoch": 0.3558938308799563, "grad_norm": 1.745120644569397, "learning_rate": 1.4941242247278008e-05, "loss": 0.9387, "step": 10100 }, { "epoch": 0.35592906789291473, "grad_norm": 3.1849899291992188, "learning_rate": 1.4940249998965396e-05, "loss": 3.3128, "step": 10101 }, { "epoch": 0.3559643049058731, "grad_norm": 3.853990077972412, "learning_rate": 1.4939257686305401e-05, "loss": 2.6588, "step": 10102 }, { "epoch": 0.35599954191883154, "grad_norm": 9.314562797546387, "learning_rate": 1.4938265309310948e-05, "loss": 3.5839, "step": 10103 }, { "epoch": 0.35603477893179, "grad_norm": 1.7387197017669678, "learning_rate": 1.4937272867994956e-05, "loss": 0.8256, "step": 10104 }, { "epoch": 0.35607001594474835, "grad_norm": 1.194535732269287, "learning_rate": 1.4936280362370361e-05, "loss": 0.8618, "step": 10105 }, { "epoch": 0.3561052529577068, "grad_norm": 3.1524837017059326, "learning_rate": 1.4935287792450082e-05, "loss": 3.0811, "step": 10106 }, { "epoch": 0.35614048997066516, "grad_norm": 1.105028748512268, "learning_rate": 1.4934295158247052e-05, "loss": 0.8197, "step": 10107 }, { "epoch": 0.3561757269836236, "grad_norm": 6.027410507202148, "learning_rate": 1.4933302459774197e-05, "loss": 3.5095, "step": 10108 }, { "epoch": 0.35621096399658203, "grad_norm": 4.220609664916992, "learning_rate": 1.4932309697044452e-05, "loss": 3.1245, "step": 10109 }, { "epoch": 0.3562462010095404, "grad_norm": 6.788013935089111, "learning_rate": 1.4931316870070745e-05, "loss": 5.3966, "step": 10110 }, { "epoch": 0.35628143802249884, "grad_norm": 1.196964144706726, "learning_rate": 1.4930323978866003e-05, "loss": 0.9735, "step": 10111 }, { "epoch": 0.3563166750354572, "grad_norm": 3.061579942703247, "learning_rate": 1.4929331023443168e-05, "loss": 3.0832, "step": 10112 }, { "epoch": 0.35635191204841565, "grad_norm": 5.183854103088379, "learning_rate": 1.4928338003815166e-05, "loss": 3.3207, "step": 10113 }, { "epoch": 0.3563871490613741, "grad_norm": 6.6584696769714355, "learning_rate": 1.4927344919994938e-05, "loss": 4.046, "step": 10114 }, { "epoch": 0.35642238607433246, "grad_norm": 4.956484794616699, "learning_rate": 1.4926351771995409e-05, "loss": 3.42, "step": 10115 }, { "epoch": 0.3564576230872909, "grad_norm": 1.5478521585464478, "learning_rate": 1.4925358559829524e-05, "loss": 0.7725, "step": 10116 }, { "epoch": 0.3564928601002493, "grad_norm": 1.3789620399475098, "learning_rate": 1.4924365283510213e-05, "loss": 1.0208, "step": 10117 }, { "epoch": 0.3565280971132077, "grad_norm": 3.669847249984741, "learning_rate": 1.4923371943050423e-05, "loss": 3.5579, "step": 10118 }, { "epoch": 0.35656333412616614, "grad_norm": 5.675819396972656, "learning_rate": 1.4922378538463081e-05, "loss": 5.3554, "step": 10119 }, { "epoch": 0.3565985711391245, "grad_norm": 1.7647291421890259, "learning_rate": 1.4921385069761134e-05, "loss": 1.1462, "step": 10120 }, { "epoch": 0.35663380815208295, "grad_norm": 5.9781174659729, "learning_rate": 1.4920391536957517e-05, "loss": 6.1908, "step": 10121 }, { "epoch": 0.35666904516504133, "grad_norm": 4.810697555541992, "learning_rate": 1.4919397940065176e-05, "loss": 3.5206, "step": 10122 }, { "epoch": 0.35670428217799977, "grad_norm": 4.042506694793701, "learning_rate": 1.491840427909705e-05, "loss": 3.8769, "step": 10123 }, { "epoch": 0.3567395191909582, "grad_norm": 1.4345771074295044, "learning_rate": 1.491741055406608e-05, "loss": 0.673, "step": 10124 }, { "epoch": 0.3567747562039166, "grad_norm": 1.378109097480774, "learning_rate": 1.4916416764985213e-05, "loss": 0.8646, "step": 10125 }, { "epoch": 0.356809993216875, "grad_norm": 1.1183079481124878, "learning_rate": 1.4915422911867392e-05, "loss": 0.9773, "step": 10126 }, { "epoch": 0.35684523022983344, "grad_norm": 1.05445396900177, "learning_rate": 1.4914428994725558e-05, "loss": 0.917, "step": 10127 }, { "epoch": 0.3568804672427918, "grad_norm": 6.937618732452393, "learning_rate": 1.4913435013572664e-05, "loss": 3.0726, "step": 10128 }, { "epoch": 0.35691570425575025, "grad_norm": 1.6393173933029175, "learning_rate": 1.4912440968421651e-05, "loss": 0.8761, "step": 10129 }, { "epoch": 0.35695094126870863, "grad_norm": 1.8226304054260254, "learning_rate": 1.491144685928547e-05, "loss": 0.7859, "step": 10130 }, { "epoch": 0.35698617828166707, "grad_norm": 1.299083948135376, "learning_rate": 1.4910452686177067e-05, "loss": 0.9265, "step": 10131 }, { "epoch": 0.3570214152946255, "grad_norm": 2.5524678230285645, "learning_rate": 1.4909458449109391e-05, "loss": 2.7443, "step": 10132 }, { "epoch": 0.3570566523075839, "grad_norm": 5.870178699493408, "learning_rate": 1.4908464148095396e-05, "loss": 5.3174, "step": 10133 }, { "epoch": 0.3570918893205423, "grad_norm": 1.3233749866485596, "learning_rate": 1.4907469783148028e-05, "loss": 1.1693, "step": 10134 }, { "epoch": 0.3571271263335007, "grad_norm": 1.0804709196090698, "learning_rate": 1.4906475354280243e-05, "loss": 0.7728, "step": 10135 }, { "epoch": 0.3571623633464591, "grad_norm": 1.3504197597503662, "learning_rate": 1.490548086150499e-05, "loss": 1.0475, "step": 10136 }, { "epoch": 0.35719760035941756, "grad_norm": 3.7923684120178223, "learning_rate": 1.4904486304835226e-05, "loss": 3.2402, "step": 10137 }, { "epoch": 0.35723283737237593, "grad_norm": 4.37365198135376, "learning_rate": 1.4903491684283902e-05, "loss": 3.7311, "step": 10138 }, { "epoch": 0.35726807438533437, "grad_norm": 5.4380645751953125, "learning_rate": 1.4902496999863973e-05, "loss": 3.4164, "step": 10139 }, { "epoch": 0.35730331139829274, "grad_norm": 4.7606916427612305, "learning_rate": 1.4901502251588398e-05, "loss": 5.5859, "step": 10140 }, { "epoch": 0.3573385484112512, "grad_norm": 1.1372896432876587, "learning_rate": 1.490050743947013e-05, "loss": 1.2824, "step": 10141 }, { "epoch": 0.3573737854242096, "grad_norm": 1.7106083631515503, "learning_rate": 1.4899512563522129e-05, "loss": 1.0183, "step": 10142 }, { "epoch": 0.357409022437168, "grad_norm": 4.074618816375732, "learning_rate": 1.4898517623757354e-05, "loss": 3.4026, "step": 10143 }, { "epoch": 0.3574442594501264, "grad_norm": 1.2469342947006226, "learning_rate": 1.489752262018876e-05, "loss": 1.1668, "step": 10144 }, { "epoch": 0.3574794964630848, "grad_norm": 4.542226791381836, "learning_rate": 1.4896527552829313e-05, "loss": 3.0021, "step": 10145 }, { "epoch": 0.35751473347604323, "grad_norm": 1.3072917461395264, "learning_rate": 1.4895532421691972e-05, "loss": 1.0557, "step": 10146 }, { "epoch": 0.35754997048900167, "grad_norm": 1.1105936765670776, "learning_rate": 1.4894537226789694e-05, "loss": 1.0749, "step": 10147 }, { "epoch": 0.35758520750196005, "grad_norm": 1.2029308080673218, "learning_rate": 1.4893541968135448e-05, "loss": 0.7623, "step": 10148 }, { "epoch": 0.3576204445149185, "grad_norm": 0.9343893527984619, "learning_rate": 1.4892546645742192e-05, "loss": 0.7843, "step": 10149 }, { "epoch": 0.35765568152787686, "grad_norm": 5.607988357543945, "learning_rate": 1.4891551259622894e-05, "loss": 3.7291, "step": 10150 }, { "epoch": 0.3576909185408353, "grad_norm": 1.408101201057434, "learning_rate": 1.4890555809790518e-05, "loss": 0.977, "step": 10151 }, { "epoch": 0.3577261555537937, "grad_norm": 7.5047197341918945, "learning_rate": 1.4889560296258028e-05, "loss": 5.7378, "step": 10152 }, { "epoch": 0.3577613925667521, "grad_norm": 1.5986462831497192, "learning_rate": 1.4888564719038395e-05, "loss": 1.3462, "step": 10153 }, { "epoch": 0.35779662957971053, "grad_norm": 0.8800773024559021, "learning_rate": 1.4887569078144583e-05, "loss": 0.9277, "step": 10154 }, { "epoch": 0.3578318665926689, "grad_norm": 4.033186912536621, "learning_rate": 1.4886573373589558e-05, "loss": 3.6473, "step": 10155 }, { "epoch": 0.35786710360562735, "grad_norm": 1.7144360542297363, "learning_rate": 1.4885577605386296e-05, "loss": 0.9858, "step": 10156 }, { "epoch": 0.3579023406185858, "grad_norm": 1.7959569692611694, "learning_rate": 1.4884581773547763e-05, "loss": 1.3241, "step": 10157 }, { "epoch": 0.35793757763154416, "grad_norm": 2.995389938354492, "learning_rate": 1.4883585878086928e-05, "loss": 3.0849, "step": 10158 }, { "epoch": 0.3579728146445026, "grad_norm": 4.58101749420166, "learning_rate": 1.4882589919016767e-05, "loss": 5.5687, "step": 10159 }, { "epoch": 0.358008051657461, "grad_norm": 1.5706210136413574, "learning_rate": 1.4881593896350252e-05, "loss": 0.961, "step": 10160 }, { "epoch": 0.3580432886704194, "grad_norm": 1.0921759605407715, "learning_rate": 1.488059781010035e-05, "loss": 1.2, "step": 10161 }, { "epoch": 0.35807852568337784, "grad_norm": 4.725221633911133, "learning_rate": 1.4879601660280045e-05, "loss": 3.4835, "step": 10162 }, { "epoch": 0.3581137626963362, "grad_norm": 1.2761567831039429, "learning_rate": 1.4878605446902302e-05, "loss": 0.8814, "step": 10163 }, { "epoch": 0.35814899970929465, "grad_norm": 5.573836803436279, "learning_rate": 1.4877609169980105e-05, "loss": 3.4562, "step": 10164 }, { "epoch": 0.3581842367222531, "grad_norm": 1.0819214582443237, "learning_rate": 1.4876612829526425e-05, "loss": 1.1174, "step": 10165 }, { "epoch": 0.35821947373521146, "grad_norm": 5.505569934844971, "learning_rate": 1.4875616425554243e-05, "loss": 8.0259, "step": 10166 }, { "epoch": 0.3582547107481699, "grad_norm": 1.3416905403137207, "learning_rate": 1.4874619958076534e-05, "loss": 1.0832, "step": 10167 }, { "epoch": 0.35828994776112827, "grad_norm": 4.496041297912598, "learning_rate": 1.4873623427106281e-05, "loss": 3.8146, "step": 10168 }, { "epoch": 0.3583251847740867, "grad_norm": 3.877039670944214, "learning_rate": 1.4872626832656462e-05, "loss": 3.5856, "step": 10169 }, { "epoch": 0.35836042178704514, "grad_norm": 3.1051199436187744, "learning_rate": 1.4871630174740057e-05, "loss": 3.2777, "step": 10170 }, { "epoch": 0.3583956588000035, "grad_norm": 3.7370188236236572, "learning_rate": 1.487063345337005e-05, "loss": 3.1534, "step": 10171 }, { "epoch": 0.35843089581296195, "grad_norm": 1.6418788433074951, "learning_rate": 1.486963666855942e-05, "loss": 0.9817, "step": 10172 }, { "epoch": 0.3584661328259203, "grad_norm": 4.484828948974609, "learning_rate": 1.4868639820321155e-05, "loss": 4.2659, "step": 10173 }, { "epoch": 0.35850136983887876, "grad_norm": 1.5306733846664429, "learning_rate": 1.4867642908668231e-05, "loss": 0.9704, "step": 10174 }, { "epoch": 0.3585366068518372, "grad_norm": 3.8133773803710938, "learning_rate": 1.4866645933613644e-05, "loss": 2.9493, "step": 10175 }, { "epoch": 0.35857184386479557, "grad_norm": 2.7651538848876953, "learning_rate": 1.486564889517037e-05, "loss": 3.2081, "step": 10176 }, { "epoch": 0.358607080877754, "grad_norm": 3.782216787338257, "learning_rate": 1.4864651793351402e-05, "loss": 3.4917, "step": 10177 }, { "epoch": 0.3586423178907124, "grad_norm": 4.703248977661133, "learning_rate": 1.4863654628169724e-05, "loss": 3.3498, "step": 10178 }, { "epoch": 0.3586775549036708, "grad_norm": 3.4341230392456055, "learning_rate": 1.4862657399638323e-05, "loss": 3.0873, "step": 10179 }, { "epoch": 0.35871279191662925, "grad_norm": 3.0947370529174805, "learning_rate": 1.486166010777019e-05, "loss": 3.5391, "step": 10180 }, { "epoch": 0.3587480289295876, "grad_norm": 3.388688087463379, "learning_rate": 1.4860662752578319e-05, "loss": 3.2812, "step": 10181 }, { "epoch": 0.35878326594254606, "grad_norm": 5.532778263092041, "learning_rate": 1.485966533407569e-05, "loss": 5.6553, "step": 10182 }, { "epoch": 0.35881850295550444, "grad_norm": 1.6139942407608032, "learning_rate": 1.4858667852275307e-05, "loss": 0.902, "step": 10183 }, { "epoch": 0.35885373996846287, "grad_norm": 1.094726324081421, "learning_rate": 1.4857670307190152e-05, "loss": 1.0791, "step": 10184 }, { "epoch": 0.3588889769814213, "grad_norm": 4.446621894836426, "learning_rate": 1.4856672698833225e-05, "loss": 3.0408, "step": 10185 }, { "epoch": 0.3589242139943797, "grad_norm": 3.7290687561035156, "learning_rate": 1.4855675027217517e-05, "loss": 3.243, "step": 10186 }, { "epoch": 0.3589594510073381, "grad_norm": 4.42548942565918, "learning_rate": 1.4854677292356023e-05, "loss": 3.2702, "step": 10187 }, { "epoch": 0.3589946880202965, "grad_norm": 4.4061279296875, "learning_rate": 1.4853679494261738e-05, "loss": 6.3037, "step": 10188 }, { "epoch": 0.3590299250332549, "grad_norm": 1.7838380336761475, "learning_rate": 1.4852681632947661e-05, "loss": 0.8493, "step": 10189 }, { "epoch": 0.35906516204621336, "grad_norm": 6.73854923248291, "learning_rate": 1.4851683708426785e-05, "loss": 3.1625, "step": 10190 }, { "epoch": 0.35910039905917174, "grad_norm": 1.5619620084762573, "learning_rate": 1.4850685720712112e-05, "loss": 1.1963, "step": 10191 }, { "epoch": 0.35913563607213017, "grad_norm": 6.509166240692139, "learning_rate": 1.4849687669816643e-05, "loss": 4.0831, "step": 10192 }, { "epoch": 0.3591708730850886, "grad_norm": 1.255760908126831, "learning_rate": 1.4848689555753369e-05, "loss": 0.8563, "step": 10193 }, { "epoch": 0.359206110098047, "grad_norm": 1.814310908317566, "learning_rate": 1.4847691378535301e-05, "loss": 0.8554, "step": 10194 }, { "epoch": 0.3592413471110054, "grad_norm": 2.675980567932129, "learning_rate": 1.4846693138175432e-05, "loss": 2.7484, "step": 10195 }, { "epoch": 0.3592765841239638, "grad_norm": 3.430861711502075, "learning_rate": 1.4845694834686767e-05, "loss": 2.5729, "step": 10196 }, { "epoch": 0.3593118211369222, "grad_norm": 5.109462738037109, "learning_rate": 1.4844696468082312e-05, "loss": 3.6764, "step": 10197 }, { "epoch": 0.35934705814988066, "grad_norm": 1.6725414991378784, "learning_rate": 1.4843698038375064e-05, "loss": 0.8557, "step": 10198 }, { "epoch": 0.35938229516283904, "grad_norm": 1.7555121183395386, "learning_rate": 1.4842699545578036e-05, "loss": 1.0016, "step": 10199 }, { "epoch": 0.35941753217579747, "grad_norm": 6.699163913726807, "learning_rate": 1.4841700989704231e-05, "loss": 5.2111, "step": 10200 }, { "epoch": 0.35945276918875585, "grad_norm": 1.0875917673110962, "learning_rate": 1.484070237076665e-05, "loss": 0.8724, "step": 10201 }, { "epoch": 0.3594880062017143, "grad_norm": 5.354889392852783, "learning_rate": 1.4839703688778307e-05, "loss": 5.201, "step": 10202 }, { "epoch": 0.3595232432146727, "grad_norm": 5.59842586517334, "learning_rate": 1.4838704943752204e-05, "loss": 3.3285, "step": 10203 }, { "epoch": 0.3595584802276311, "grad_norm": 1.1267307996749878, "learning_rate": 1.4837706135701354e-05, "loss": 1.166, "step": 10204 }, { "epoch": 0.3595937172405895, "grad_norm": 1.385443091392517, "learning_rate": 1.4836707264638769e-05, "loss": 0.84, "step": 10205 }, { "epoch": 0.3596289542535479, "grad_norm": 4.199053764343262, "learning_rate": 1.483570833057745e-05, "loss": 2.8663, "step": 10206 }, { "epoch": 0.35966419126650634, "grad_norm": 4.734228610992432, "learning_rate": 1.4834709333530417e-05, "loss": 3.1534, "step": 10207 }, { "epoch": 0.3596994282794648, "grad_norm": 3.3965272903442383, "learning_rate": 1.4833710273510678e-05, "loss": 2.3987, "step": 10208 }, { "epoch": 0.35973466529242315, "grad_norm": 4.025068759918213, "learning_rate": 1.4832711150531248e-05, "loss": 3.6114, "step": 10209 }, { "epoch": 0.3597699023053816, "grad_norm": 0.9776965975761414, "learning_rate": 1.4831711964605138e-05, "loss": 1.0708, "step": 10210 }, { "epoch": 0.35980513931833996, "grad_norm": 5.1784515380859375, "learning_rate": 1.4830712715745366e-05, "loss": 3.7946, "step": 10211 }, { "epoch": 0.3598403763312984, "grad_norm": 6.05475378036499, "learning_rate": 1.4829713403964943e-05, "loss": 3.9974, "step": 10212 }, { "epoch": 0.35987561334425683, "grad_norm": 2.6378719806671143, "learning_rate": 1.4828714029276892e-05, "loss": 2.6305, "step": 10213 }, { "epoch": 0.3599108503572152, "grad_norm": 3.9426209926605225, "learning_rate": 1.4827714591694222e-05, "loss": 3.7289, "step": 10214 }, { "epoch": 0.35994608737017364, "grad_norm": 3.5144524574279785, "learning_rate": 1.4826715091229957e-05, "loss": 4.5842, "step": 10215 }, { "epoch": 0.359981324383132, "grad_norm": 1.3413029909133911, "learning_rate": 1.4825715527897112e-05, "loss": 0.7636, "step": 10216 }, { "epoch": 0.36001656139609045, "grad_norm": 1.3041903972625732, "learning_rate": 1.4824715901708705e-05, "loss": 0.7508, "step": 10217 }, { "epoch": 0.3600517984090489, "grad_norm": 1.2480279207229614, "learning_rate": 1.4823716212677763e-05, "loss": 0.8948, "step": 10218 }, { "epoch": 0.36008703542200726, "grad_norm": 2.001622438430786, "learning_rate": 1.4822716460817304e-05, "loss": 1.0093, "step": 10219 }, { "epoch": 0.3601222724349657, "grad_norm": 6.504123210906982, "learning_rate": 1.4821716646140344e-05, "loss": 2.9161, "step": 10220 }, { "epoch": 0.3601575094479241, "grad_norm": 2.967271566390991, "learning_rate": 1.4820716768659913e-05, "loss": 3.3931, "step": 10221 }, { "epoch": 0.3601927464608825, "grad_norm": 5.877402305603027, "learning_rate": 1.481971682838903e-05, "loss": 5.787, "step": 10222 }, { "epoch": 0.36022798347384094, "grad_norm": 1.7538944482803345, "learning_rate": 1.4818716825340725e-05, "loss": 1.0799, "step": 10223 }, { "epoch": 0.3602632204867993, "grad_norm": 1.0780426263809204, "learning_rate": 1.4817716759528016e-05, "loss": 0.9628, "step": 10224 }, { "epoch": 0.36029845749975775, "grad_norm": 1.8452651500701904, "learning_rate": 1.4816716630963935e-05, "loss": 0.9872, "step": 10225 }, { "epoch": 0.36033369451271613, "grad_norm": 2.0713982582092285, "learning_rate": 1.4815716439661504e-05, "loss": 0.8038, "step": 10226 }, { "epoch": 0.36036893152567456, "grad_norm": 5.116639614105225, "learning_rate": 1.4814716185633755e-05, "loss": 3.7169, "step": 10227 }, { "epoch": 0.360404168538633, "grad_norm": 3.835292339324951, "learning_rate": 1.4813715868893712e-05, "loss": 2.8146, "step": 10228 }, { "epoch": 0.3604394055515914, "grad_norm": 5.424389839172363, "learning_rate": 1.481271548945441e-05, "loss": 3.3082, "step": 10229 }, { "epoch": 0.3604746425645498, "grad_norm": 4.9714765548706055, "learning_rate": 1.4811715047328874e-05, "loss": 5.3474, "step": 10230 }, { "epoch": 0.36050987957750824, "grad_norm": 3.154815196990967, "learning_rate": 1.4810714542530136e-05, "loss": 3.2726, "step": 10231 }, { "epoch": 0.3605451165904666, "grad_norm": 3.27604341506958, "learning_rate": 1.4809713975071229e-05, "loss": 2.8739, "step": 10232 }, { "epoch": 0.36058035360342505, "grad_norm": 4.574913024902344, "learning_rate": 1.4808713344965184e-05, "loss": 5.781, "step": 10233 }, { "epoch": 0.36061559061638343, "grad_norm": 4.056948184967041, "learning_rate": 1.4807712652225034e-05, "loss": 3.4312, "step": 10234 }, { "epoch": 0.36065082762934186, "grad_norm": 1.2634419202804565, "learning_rate": 1.4806711896863816e-05, "loss": 1.1623, "step": 10235 }, { "epoch": 0.3606860646423003, "grad_norm": 1.1667364835739136, "learning_rate": 1.4805711078894564e-05, "loss": 0.7181, "step": 10236 }, { "epoch": 0.3607213016552587, "grad_norm": 1.7704439163208008, "learning_rate": 1.4804710198330311e-05, "loss": 0.9312, "step": 10237 }, { "epoch": 0.3607565386682171, "grad_norm": 1.8667386770248413, "learning_rate": 1.48037092551841e-05, "loss": 1.0348, "step": 10238 }, { "epoch": 0.3607917756811755, "grad_norm": 2.3216898441314697, "learning_rate": 1.480270824946896e-05, "loss": 0.7007, "step": 10239 }, { "epoch": 0.3608270126941339, "grad_norm": 6.751319885253906, "learning_rate": 1.4801707181197935e-05, "loss": 6.1043, "step": 10240 }, { "epoch": 0.36086224970709235, "grad_norm": 1.22044837474823, "learning_rate": 1.4800706050384061e-05, "loss": 0.8104, "step": 10241 }, { "epoch": 0.36089748672005073, "grad_norm": 5.1228227615356445, "learning_rate": 1.479970485704038e-05, "loss": 4.2226, "step": 10242 }, { "epoch": 0.36093272373300916, "grad_norm": 2.6580119132995605, "learning_rate": 1.4798703601179933e-05, "loss": 3.1974, "step": 10243 }, { "epoch": 0.36096796074596754, "grad_norm": 5.503648281097412, "learning_rate": 1.4797702282815757e-05, "loss": 3.3936, "step": 10244 }, { "epoch": 0.361003197758926, "grad_norm": 1.5419546365737915, "learning_rate": 1.4796700901960901e-05, "loss": 1.1601, "step": 10245 }, { "epoch": 0.3610384347718844, "grad_norm": 7.3399152755737305, "learning_rate": 1.4795699458628405e-05, "loss": 5.0619, "step": 10246 }, { "epoch": 0.3610736717848428, "grad_norm": 1.568761944770813, "learning_rate": 1.4794697952831309e-05, "loss": 0.7938, "step": 10247 }, { "epoch": 0.3611089087978012, "grad_norm": 7.665252208709717, "learning_rate": 1.4793696384582665e-05, "loss": 3.2946, "step": 10248 }, { "epoch": 0.3611441458107596, "grad_norm": 1.4371352195739746, "learning_rate": 1.4792694753895513e-05, "loss": 0.8996, "step": 10249 }, { "epoch": 0.36117938282371803, "grad_norm": 1.377347707748413, "learning_rate": 1.4791693060782902e-05, "loss": 0.9028, "step": 10250 }, { "epoch": 0.36121461983667646, "grad_norm": 2.569395065307617, "learning_rate": 1.4790691305257878e-05, "loss": 2.5688, "step": 10251 }, { "epoch": 0.36124985684963484, "grad_norm": 1.8462889194488525, "learning_rate": 1.478968948733349e-05, "loss": 0.9959, "step": 10252 }, { "epoch": 0.3612850938625933, "grad_norm": 1.4430367946624756, "learning_rate": 1.4788687607022784e-05, "loss": 0.9304, "step": 10253 }, { "epoch": 0.36132033087555165, "grad_norm": 1.7618809938430786, "learning_rate": 1.4787685664338816e-05, "loss": 0.9913, "step": 10254 }, { "epoch": 0.3613555678885101, "grad_norm": 1.0892417430877686, "learning_rate": 1.4786683659294629e-05, "loss": 0.8469, "step": 10255 }, { "epoch": 0.3613908049014685, "grad_norm": 9.545145034790039, "learning_rate": 1.478568159190328e-05, "loss": 8.4591, "step": 10256 }, { "epoch": 0.3614260419144269, "grad_norm": 2.121572494506836, "learning_rate": 1.478467946217782e-05, "loss": 0.7915, "step": 10257 }, { "epoch": 0.36146127892738533, "grad_norm": 6.403286457061768, "learning_rate": 1.4783677270131296e-05, "loss": 3.6055, "step": 10258 }, { "epoch": 0.3614965159403437, "grad_norm": 12.154194831848145, "learning_rate": 1.4782675015776771e-05, "loss": 3.7193, "step": 10259 }, { "epoch": 0.36153175295330214, "grad_norm": 1.7075225114822388, "learning_rate": 1.4781672699127293e-05, "loss": 0.7524, "step": 10260 }, { "epoch": 0.3615669899662606, "grad_norm": 4.1309638023376465, "learning_rate": 1.4780670320195915e-05, "loss": 3.5138, "step": 10261 }, { "epoch": 0.36160222697921895, "grad_norm": 1.6270520687103271, "learning_rate": 1.47796678789957e-05, "loss": 0.8879, "step": 10262 }, { "epoch": 0.3616374639921774, "grad_norm": 1.7580928802490234, "learning_rate": 1.4778665375539705e-05, "loss": 0.8504, "step": 10263 }, { "epoch": 0.3616727010051358, "grad_norm": 2.9851226806640625, "learning_rate": 1.4777662809840981e-05, "loss": 3.5824, "step": 10264 }, { "epoch": 0.3617079380180942, "grad_norm": 1.1888645887374878, "learning_rate": 1.4776660181912591e-05, "loss": 0.7446, "step": 10265 }, { "epoch": 0.36174317503105263, "grad_norm": 4.330732822418213, "learning_rate": 1.4775657491767594e-05, "loss": 3.8233, "step": 10266 }, { "epoch": 0.361778412044011, "grad_norm": 3.349468946456909, "learning_rate": 1.4774654739419051e-05, "loss": 2.7184, "step": 10267 }, { "epoch": 0.36181364905696944, "grad_norm": 7.117961883544922, "learning_rate": 1.4773651924880022e-05, "loss": 6.2826, "step": 10268 }, { "epoch": 0.3618488860699279, "grad_norm": 5.240138053894043, "learning_rate": 1.4772649048163566e-05, "loss": 3.7029, "step": 10269 }, { "epoch": 0.36188412308288626, "grad_norm": 1.6612558364868164, "learning_rate": 1.477164610928275e-05, "loss": 0.7992, "step": 10270 }, { "epoch": 0.3619193600958447, "grad_norm": 1.3993611335754395, "learning_rate": 1.4770643108250636e-05, "loss": 0.8484, "step": 10271 }, { "epoch": 0.36195459710880307, "grad_norm": 2.6995043754577637, "learning_rate": 1.4769640045080287e-05, "loss": 2.5449, "step": 10272 }, { "epoch": 0.3619898341217615, "grad_norm": 4.202029228210449, "learning_rate": 1.4768636919784766e-05, "loss": 3.3448, "step": 10273 }, { "epoch": 0.36202507113471993, "grad_norm": 3.143399953842163, "learning_rate": 1.4767633732377147e-05, "loss": 3.5628, "step": 10274 }, { "epoch": 0.3620603081476783, "grad_norm": 2.1150870323181152, "learning_rate": 1.4766630482870487e-05, "loss": 2.5128, "step": 10275 }, { "epoch": 0.36209554516063674, "grad_norm": 3.4916417598724365, "learning_rate": 1.4765627171277863e-05, "loss": 2.8229, "step": 10276 }, { "epoch": 0.3621307821735951, "grad_norm": 0.8865952491760254, "learning_rate": 1.4764623797612333e-05, "loss": 0.9872, "step": 10277 }, { "epoch": 0.36216601918655356, "grad_norm": 4.286468982696533, "learning_rate": 1.4763620361886975e-05, "loss": 3.7676, "step": 10278 }, { "epoch": 0.362201256199512, "grad_norm": 1.2850041389465332, "learning_rate": 1.4762616864114851e-05, "loss": 1.1315, "step": 10279 }, { "epoch": 0.36223649321247037, "grad_norm": 3.1678271293640137, "learning_rate": 1.476161330430904e-05, "loss": 3.1087, "step": 10280 }, { "epoch": 0.3622717302254288, "grad_norm": 0.9891210794448853, "learning_rate": 1.4760609682482605e-05, "loss": 0.8736, "step": 10281 }, { "epoch": 0.3623069672383872, "grad_norm": 9.594735145568848, "learning_rate": 1.4759605998648625e-05, "loss": 3.7947, "step": 10282 }, { "epoch": 0.3623422042513456, "grad_norm": 1.4006787538528442, "learning_rate": 1.4758602252820168e-05, "loss": 0.912, "step": 10283 }, { "epoch": 0.36237744126430405, "grad_norm": 5.088297367095947, "learning_rate": 1.4757598445010314e-05, "loss": 3.5664, "step": 10284 }, { "epoch": 0.3624126782772624, "grad_norm": 5.461989879608154, "learning_rate": 1.475659457523213e-05, "loss": 3.3176, "step": 10285 }, { "epoch": 0.36244791529022086, "grad_norm": 1.318939447402954, "learning_rate": 1.47555906434987e-05, "loss": 0.8852, "step": 10286 }, { "epoch": 0.36248315230317923, "grad_norm": 3.830320358276367, "learning_rate": 1.4754586649823095e-05, "loss": 3.1043, "step": 10287 }, { "epoch": 0.36251838931613767, "grad_norm": 4.192176818847656, "learning_rate": 1.4753582594218395e-05, "loss": 3.7325, "step": 10288 }, { "epoch": 0.3625536263290961, "grad_norm": 1.4283902645111084, "learning_rate": 1.4752578476697673e-05, "loss": 0.925, "step": 10289 }, { "epoch": 0.3625888633420545, "grad_norm": 6.874505043029785, "learning_rate": 1.4751574297274012e-05, "loss": 3.406, "step": 10290 }, { "epoch": 0.3626241003550129, "grad_norm": 6.575316429138184, "learning_rate": 1.4750570055960489e-05, "loss": 3.7481, "step": 10291 }, { "epoch": 0.3626593373679713, "grad_norm": 7.050491809844971, "learning_rate": 1.4749565752770187e-05, "loss": 3.3624, "step": 10292 }, { "epoch": 0.3626945743809297, "grad_norm": 5.055514335632324, "learning_rate": 1.4748561387716187e-05, "loss": 3.4942, "step": 10293 }, { "epoch": 0.36272981139388816, "grad_norm": 1.1838228702545166, "learning_rate": 1.4747556960811568e-05, "loss": 0.9854, "step": 10294 }, { "epoch": 0.36276504840684654, "grad_norm": 4.515137672424316, "learning_rate": 1.474655247206942e-05, "loss": 3.1716, "step": 10295 }, { "epoch": 0.36280028541980497, "grad_norm": 6.301214218139648, "learning_rate": 1.4745547921502813e-05, "loss": 4.0606, "step": 10296 }, { "epoch": 0.3628355224327634, "grad_norm": 1.896410346031189, "learning_rate": 1.4744543309124849e-05, "loss": 0.8919, "step": 10297 }, { "epoch": 0.3628707594457218, "grad_norm": 1.717716932296753, "learning_rate": 1.4743538634948595e-05, "loss": 1.0119, "step": 10298 }, { "epoch": 0.3629059964586802, "grad_norm": 1.7003328800201416, "learning_rate": 1.4742533898987153e-05, "loss": 0.8703, "step": 10299 }, { "epoch": 0.3629412334716386, "grad_norm": 1.3564863204956055, "learning_rate": 1.4741529101253602e-05, "loss": 0.8361, "step": 10300 }, { "epoch": 0.362976470484597, "grad_norm": 1.6134849786758423, "learning_rate": 1.4740524241761028e-05, "loss": 1.0385, "step": 10301 }, { "epoch": 0.36301170749755546, "grad_norm": 1.452060341835022, "learning_rate": 1.4739519320522521e-05, "loss": 1.1218, "step": 10302 }, { "epoch": 0.36304694451051384, "grad_norm": 1.3011921644210815, "learning_rate": 1.4738514337551176e-05, "loss": 0.9886, "step": 10303 }, { "epoch": 0.36308218152347227, "grad_norm": 6.506319046020508, "learning_rate": 1.4737509292860073e-05, "loss": 3.6514, "step": 10304 }, { "epoch": 0.36311741853643065, "grad_norm": 3.1864826679229736, "learning_rate": 1.4736504186462312e-05, "loss": 3.3631, "step": 10305 }, { "epoch": 0.3631526555493891, "grad_norm": 1.8073251247406006, "learning_rate": 1.4735499018370978e-05, "loss": 0.7513, "step": 10306 }, { "epoch": 0.3631878925623475, "grad_norm": 1.6771539449691772, "learning_rate": 1.4734493788599168e-05, "loss": 1.0034, "step": 10307 }, { "epoch": 0.3632231295753059, "grad_norm": 6.016399383544922, "learning_rate": 1.473348849715997e-05, "loss": 5.4088, "step": 10308 }, { "epoch": 0.3632583665882643, "grad_norm": 4.56634521484375, "learning_rate": 1.4732483144066484e-05, "loss": 3.4325, "step": 10309 }, { "epoch": 0.3632936036012227, "grad_norm": 5.852180480957031, "learning_rate": 1.4731477729331802e-05, "loss": 5.8407, "step": 10310 }, { "epoch": 0.36332884061418114, "grad_norm": 3.3404324054718018, "learning_rate": 1.473047225296902e-05, "loss": 3.4055, "step": 10311 }, { "epoch": 0.36336407762713957, "grad_norm": 5.64032506942749, "learning_rate": 1.4729466714991235e-05, "loss": 3.3841, "step": 10312 }, { "epoch": 0.36339931464009795, "grad_norm": 1.6664365530014038, "learning_rate": 1.4728461115411544e-05, "loss": 0.8224, "step": 10313 }, { "epoch": 0.3634345516530564, "grad_norm": 8.594104766845703, "learning_rate": 1.4727455454243042e-05, "loss": 3.4078, "step": 10314 }, { "epoch": 0.36346978866601476, "grad_norm": 4.383820056915283, "learning_rate": 1.4726449731498831e-05, "loss": 3.2157, "step": 10315 }, { "epoch": 0.3635050256789732, "grad_norm": 1.8069183826446533, "learning_rate": 1.4725443947192012e-05, "loss": 0.9604, "step": 10316 }, { "epoch": 0.3635402626919316, "grad_norm": 2.4997997283935547, "learning_rate": 1.4724438101335681e-05, "loss": 0.9791, "step": 10317 }, { "epoch": 0.36357549970489, "grad_norm": 0.8767270445823669, "learning_rate": 1.4723432193942946e-05, "loss": 1.1379, "step": 10318 }, { "epoch": 0.36361073671784844, "grad_norm": 6.405498504638672, "learning_rate": 1.4722426225026902e-05, "loss": 5.2917, "step": 10319 }, { "epoch": 0.3636459737308068, "grad_norm": 5.9803032875061035, "learning_rate": 1.4721420194600655e-05, "loss": 5.2906, "step": 10320 }, { "epoch": 0.36368121074376525, "grad_norm": 1.7168169021606445, "learning_rate": 1.4720414102677309e-05, "loss": 1.0067, "step": 10321 }, { "epoch": 0.3637164477567237, "grad_norm": 1.2850923538208008, "learning_rate": 1.471940794926997e-05, "loss": 0.8473, "step": 10322 }, { "epoch": 0.36375168476968206, "grad_norm": 1.2635265588760376, "learning_rate": 1.4718401734391737e-05, "loss": 0.6102, "step": 10323 }, { "epoch": 0.3637869217826405, "grad_norm": 5.792949676513672, "learning_rate": 1.4717395458055725e-05, "loss": 3.2127, "step": 10324 }, { "epoch": 0.36382215879559887, "grad_norm": 8.105606079101562, "learning_rate": 1.4716389120275031e-05, "loss": 3.349, "step": 10325 }, { "epoch": 0.3638573958085573, "grad_norm": 6.774895191192627, "learning_rate": 1.4715382721062774e-05, "loss": 5.7654, "step": 10326 }, { "epoch": 0.36389263282151574, "grad_norm": 1.1049319505691528, "learning_rate": 1.4714376260432052e-05, "loss": 1.0587, "step": 10327 }, { "epoch": 0.3639278698344741, "grad_norm": 1.1813222169876099, "learning_rate": 1.471336973839598e-05, "loss": 1.1182, "step": 10328 }, { "epoch": 0.36396310684743255, "grad_norm": 4.7940287590026855, "learning_rate": 1.4712363154967668e-05, "loss": 0.8738, "step": 10329 }, { "epoch": 0.3639983438603909, "grad_norm": 1.1953150033950806, "learning_rate": 1.4711356510160223e-05, "loss": 0.8713, "step": 10330 }, { "epoch": 0.36403358087334936, "grad_norm": 1.352034568786621, "learning_rate": 1.4710349803986761e-05, "loss": 0.9839, "step": 10331 }, { "epoch": 0.3640688178863078, "grad_norm": 4.370678424835205, "learning_rate": 1.4709343036460394e-05, "loss": 2.8194, "step": 10332 }, { "epoch": 0.36410405489926617, "grad_norm": 1.0993834733963013, "learning_rate": 1.4708336207594234e-05, "loss": 0.6765, "step": 10333 }, { "epoch": 0.3641392919122246, "grad_norm": 0.775132954120636, "learning_rate": 1.470732931740139e-05, "loss": 0.7696, "step": 10334 }, { "epoch": 0.36417452892518304, "grad_norm": 6.582147121429443, "learning_rate": 1.470632236589499e-05, "loss": 3.3378, "step": 10335 }, { "epoch": 0.3642097659381414, "grad_norm": 13.109588623046875, "learning_rate": 1.4705315353088134e-05, "loss": 3.1373, "step": 10336 }, { "epoch": 0.36424500295109985, "grad_norm": 1.3818403482437134, "learning_rate": 1.4704308278993951e-05, "loss": 0.9746, "step": 10337 }, { "epoch": 0.36428023996405823, "grad_norm": 5.1153154373168945, "learning_rate": 1.4703301143625552e-05, "loss": 3.0487, "step": 10338 }, { "epoch": 0.36431547697701666, "grad_norm": 1.582874059677124, "learning_rate": 1.4702293946996058e-05, "loss": 1.5033, "step": 10339 }, { "epoch": 0.3643507139899751, "grad_norm": 8.03946304321289, "learning_rate": 1.4701286689118583e-05, "loss": 3.7163, "step": 10340 }, { "epoch": 0.3643859510029335, "grad_norm": 5.6348724365234375, "learning_rate": 1.4700279370006255e-05, "loss": 3.4172, "step": 10341 }, { "epoch": 0.3644211880158919, "grad_norm": 4.361040115356445, "learning_rate": 1.4699271989672183e-05, "loss": 3.2085, "step": 10342 }, { "epoch": 0.3644564250288503, "grad_norm": 1.4626191854476929, "learning_rate": 1.4698264548129502e-05, "loss": 1.0949, "step": 10343 }, { "epoch": 0.3644916620418087, "grad_norm": 1.4676553010940552, "learning_rate": 1.469725704539132e-05, "loss": 0.9064, "step": 10344 }, { "epoch": 0.36452689905476715, "grad_norm": 9.850502014160156, "learning_rate": 1.469624948147077e-05, "loss": 3.2741, "step": 10345 }, { "epoch": 0.36456213606772553, "grad_norm": 1.657039761543274, "learning_rate": 1.4695241856380972e-05, "loss": 1.1808, "step": 10346 }, { "epoch": 0.36459737308068396, "grad_norm": 5.578458786010742, "learning_rate": 1.4694234170135052e-05, "loss": 5.9507, "step": 10347 }, { "epoch": 0.36463261009364234, "grad_norm": 1.2497789859771729, "learning_rate": 1.4693226422746132e-05, "loss": 0.94, "step": 10348 }, { "epoch": 0.3646678471066008, "grad_norm": 9.97900390625, "learning_rate": 1.469221861422734e-05, "loss": 5.3528, "step": 10349 }, { "epoch": 0.3647030841195592, "grad_norm": 1.2331016063690186, "learning_rate": 1.4691210744591803e-05, "loss": 0.9131, "step": 10350 }, { "epoch": 0.3647383211325176, "grad_norm": 3.949591636657715, "learning_rate": 1.469020281385265e-05, "loss": 2.8886, "step": 10351 }, { "epoch": 0.364773558145476, "grad_norm": 3.67628812789917, "learning_rate": 1.4689194822023007e-05, "loss": 3.4592, "step": 10352 }, { "epoch": 0.3648087951584344, "grad_norm": 5.097962379455566, "learning_rate": 1.4688186769116004e-05, "loss": 4.6098, "step": 10353 }, { "epoch": 0.36484403217139283, "grad_norm": 2.973310708999634, "learning_rate": 1.4687178655144774e-05, "loss": 3.1143, "step": 10354 }, { "epoch": 0.36487926918435126, "grad_norm": 1.5343114137649536, "learning_rate": 1.4686170480122438e-05, "loss": 1.0083, "step": 10355 }, { "epoch": 0.36491450619730964, "grad_norm": 2.1872758865356445, "learning_rate": 1.4685162244062141e-05, "loss": 0.7897, "step": 10356 }, { "epoch": 0.3649497432102681, "grad_norm": 2.2807066440582275, "learning_rate": 1.4684153946977006e-05, "loss": 0.9247, "step": 10357 }, { "epoch": 0.36498498022322645, "grad_norm": 4.3660054206848145, "learning_rate": 1.4683145588880173e-05, "loss": 3.1946, "step": 10358 }, { "epoch": 0.3650202172361849, "grad_norm": 8.042181968688965, "learning_rate": 1.4682137169784767e-05, "loss": 5.4662, "step": 10359 }, { "epoch": 0.3650554542491433, "grad_norm": 2.7365810871124268, "learning_rate": 1.4681128689703932e-05, "loss": 1.0251, "step": 10360 }, { "epoch": 0.3650906912621017, "grad_norm": 9.387253761291504, "learning_rate": 1.4680120148650796e-05, "loss": 5.7033, "step": 10361 }, { "epoch": 0.36512592827506013, "grad_norm": 0.8969497680664062, "learning_rate": 1.4679111546638503e-05, "loss": 0.8801, "step": 10362 }, { "epoch": 0.3651611652880185, "grad_norm": 6.904600143432617, "learning_rate": 1.4678102883680182e-05, "loss": 3.1746, "step": 10363 }, { "epoch": 0.36519640230097694, "grad_norm": 1.3479937314987183, "learning_rate": 1.4677094159788978e-05, "loss": 1.1592, "step": 10364 }, { "epoch": 0.3652316393139354, "grad_norm": 3.6903347969055176, "learning_rate": 1.4676085374978024e-05, "loss": 3.0257, "step": 10365 }, { "epoch": 0.36526687632689375, "grad_norm": 4.222703456878662, "learning_rate": 1.4675076529260464e-05, "loss": 5.4704, "step": 10366 }, { "epoch": 0.3653021133398522, "grad_norm": 0.9854046106338501, "learning_rate": 1.4674067622649437e-05, "loss": 1.0961, "step": 10367 }, { "epoch": 0.3653373503528106, "grad_norm": 1.3661059141159058, "learning_rate": 1.4673058655158084e-05, "loss": 0.818, "step": 10368 }, { "epoch": 0.365372587365769, "grad_norm": 1.5375745296478271, "learning_rate": 1.4672049626799545e-05, "loss": 0.8917, "step": 10369 }, { "epoch": 0.36540782437872743, "grad_norm": 1.3553682565689087, "learning_rate": 1.4671040537586965e-05, "loss": 0.6985, "step": 10370 }, { "epoch": 0.3654430613916858, "grad_norm": 10.811805725097656, "learning_rate": 1.4670031387533487e-05, "loss": 3.4186, "step": 10371 }, { "epoch": 0.36547829840464424, "grad_norm": 1.0292502641677856, "learning_rate": 1.4669022176652257e-05, "loss": 1.0601, "step": 10372 }, { "epoch": 0.3655135354176027, "grad_norm": 1.6042147874832153, "learning_rate": 1.466801290495642e-05, "loss": 1.038, "step": 10373 }, { "epoch": 0.36554877243056105, "grad_norm": 3.545808792114258, "learning_rate": 1.4667003572459115e-05, "loss": 3.9763, "step": 10374 }, { "epoch": 0.3655840094435195, "grad_norm": 1.145838737487793, "learning_rate": 1.4665994179173498e-05, "loss": 0.9132, "step": 10375 }, { "epoch": 0.36561924645647786, "grad_norm": 1.0547053813934326, "learning_rate": 1.466498472511271e-05, "loss": 0.9814, "step": 10376 }, { "epoch": 0.3656544834694363, "grad_norm": 1.0718055963516235, "learning_rate": 1.4663975210289902e-05, "loss": 1.0128, "step": 10377 }, { "epoch": 0.36568972048239473, "grad_norm": 3.8857383728027344, "learning_rate": 1.4662965634718224e-05, "loss": 3.2008, "step": 10378 }, { "epoch": 0.3657249574953531, "grad_norm": 7.109925270080566, "learning_rate": 1.4661955998410826e-05, "loss": 5.4143, "step": 10379 }, { "epoch": 0.36576019450831154, "grad_norm": 1.1520116329193115, "learning_rate": 1.4660946301380855e-05, "loss": 0.759, "step": 10380 }, { "epoch": 0.3657954315212699, "grad_norm": 1.137558102607727, "learning_rate": 1.4659936543641467e-05, "loss": 0.7092, "step": 10381 }, { "epoch": 0.36583066853422835, "grad_norm": 2.011075496673584, "learning_rate": 1.4658926725205807e-05, "loss": 0.9909, "step": 10382 }, { "epoch": 0.3658659055471868, "grad_norm": 1.1988252401351929, "learning_rate": 1.465791684608704e-05, "loss": 1.0868, "step": 10383 }, { "epoch": 0.36590114256014517, "grad_norm": 1.8053109645843506, "learning_rate": 1.4656906906298308e-05, "loss": 0.8569, "step": 10384 }, { "epoch": 0.3659363795731036, "grad_norm": 6.211877822875977, "learning_rate": 1.4655896905852769e-05, "loss": 2.8057, "step": 10385 }, { "epoch": 0.365971616586062, "grad_norm": 1.3054704666137695, "learning_rate": 1.4654886844763583e-05, "loss": 0.8471, "step": 10386 }, { "epoch": 0.3660068535990204, "grad_norm": 7.107410907745361, "learning_rate": 1.4653876723043903e-05, "loss": 3.5493, "step": 10387 }, { "epoch": 0.36604209061197884, "grad_norm": 7.615004062652588, "learning_rate": 1.4652866540706882e-05, "loss": 5.8747, "step": 10388 }, { "epoch": 0.3660773276249372, "grad_norm": 3.25523042678833, "learning_rate": 1.4651856297765687e-05, "loss": 3.1368, "step": 10389 }, { "epoch": 0.36611256463789565, "grad_norm": 4.908904552459717, "learning_rate": 1.4650845994233467e-05, "loss": 3.4309, "step": 10390 }, { "epoch": 0.36614780165085403, "grad_norm": 1.6055046319961548, "learning_rate": 1.4649835630123389e-05, "loss": 0.8069, "step": 10391 }, { "epoch": 0.36618303866381247, "grad_norm": 1.7924201488494873, "learning_rate": 1.4648825205448607e-05, "loss": 0.8793, "step": 10392 }, { "epoch": 0.3662182756767709, "grad_norm": 5.910682201385498, "learning_rate": 1.4647814720222285e-05, "loss": 3.4749, "step": 10393 }, { "epoch": 0.3662535126897293, "grad_norm": 5.910682201385498, "learning_rate": 1.4647814720222285e-05, "loss": 3.6342, "step": 10394 }, { "epoch": 0.3662887497026877, "grad_norm": 1.4852463006973267, "learning_rate": 1.4646804174457585e-05, "loss": 1.0386, "step": 10395 }, { "epoch": 0.3663239867156461, "grad_norm": 1.4113829135894775, "learning_rate": 1.4645793568167668e-05, "loss": 1.0129, "step": 10396 }, { "epoch": 0.3663592237286045, "grad_norm": 1.316956877708435, "learning_rate": 1.4644782901365698e-05, "loss": 0.9029, "step": 10397 }, { "epoch": 0.36639446074156296, "grad_norm": 6.795673370361328, "learning_rate": 1.4643772174064839e-05, "loss": 5.1407, "step": 10398 }, { "epoch": 0.36642969775452133, "grad_norm": 2.392658233642578, "learning_rate": 1.4642761386278256e-05, "loss": 0.7402, "step": 10399 }, { "epoch": 0.36646493476747977, "grad_norm": 1.0893950462341309, "learning_rate": 1.4641750538019113e-05, "loss": 0.9717, "step": 10400 }, { "epoch": 0.3665001717804382, "grad_norm": 5.452059745788574, "learning_rate": 1.4640739629300582e-05, "loss": 3.881, "step": 10401 }, { "epoch": 0.3665354087933966, "grad_norm": 5.4841203689575195, "learning_rate": 1.4639728660135822e-05, "loss": 2.8958, "step": 10402 }, { "epoch": 0.366570645806355, "grad_norm": 0.8702849745750427, "learning_rate": 1.4638717630538008e-05, "loss": 0.9585, "step": 10403 }, { "epoch": 0.3666058828193134, "grad_norm": 1.3279011249542236, "learning_rate": 1.4637706540520306e-05, "loss": 0.7482, "step": 10404 }, { "epoch": 0.3666411198322718, "grad_norm": 1.2860606908798218, "learning_rate": 1.4636695390095884e-05, "loss": 0.8523, "step": 10405 }, { "epoch": 0.36667635684523026, "grad_norm": 1.541772723197937, "learning_rate": 1.4635684179277914e-05, "loss": 0.9163, "step": 10406 }, { "epoch": 0.36671159385818863, "grad_norm": 3.90446400642395, "learning_rate": 1.463467290807957e-05, "loss": 3.1331, "step": 10407 }, { "epoch": 0.36674683087114707, "grad_norm": 1.093201756477356, "learning_rate": 1.4633661576514016e-05, "loss": 0.8666, "step": 10408 }, { "epoch": 0.36678206788410544, "grad_norm": 1.3891992568969727, "learning_rate": 1.4632650184594435e-05, "loss": 1.122, "step": 10409 }, { "epoch": 0.3668173048970639, "grad_norm": 4.797504425048828, "learning_rate": 1.4631638732333988e-05, "loss": 3.3478, "step": 10410 }, { "epoch": 0.3668525419100223, "grad_norm": 6.273575782775879, "learning_rate": 1.4630627219745863e-05, "loss": 3.4865, "step": 10411 }, { "epoch": 0.3668877789229807, "grad_norm": 5.369051456451416, "learning_rate": 1.462961564684323e-05, "loss": 3.4656, "step": 10412 }, { "epoch": 0.3669230159359391, "grad_norm": 9.758711814880371, "learning_rate": 1.4628604013639258e-05, "loss": 6.2002, "step": 10413 }, { "epoch": 0.3669582529488975, "grad_norm": 1.3904396295547485, "learning_rate": 1.4627592320147131e-05, "loss": 0.8125, "step": 10414 }, { "epoch": 0.36699348996185593, "grad_norm": 4.91854190826416, "learning_rate": 1.4626580566380024e-05, "loss": 3.0717, "step": 10415 }, { "epoch": 0.36702872697481437, "grad_norm": 1.1622837781906128, "learning_rate": 1.4625568752351118e-05, "loss": 0.8473, "step": 10416 }, { "epoch": 0.36706396398777275, "grad_norm": 4.210049152374268, "learning_rate": 1.4624556878073588e-05, "loss": 3.2411, "step": 10417 }, { "epoch": 0.3670992010007312, "grad_norm": 1.678672194480896, "learning_rate": 1.4623544943560615e-05, "loss": 0.9411, "step": 10418 }, { "epoch": 0.36713443801368956, "grad_norm": 4.673572063446045, "learning_rate": 1.4622532948825381e-05, "loss": 3.1987, "step": 10419 }, { "epoch": 0.367169675026648, "grad_norm": 1.8885825872421265, "learning_rate": 1.462152089388107e-05, "loss": 0.8724, "step": 10420 }, { "epoch": 0.3672049120396064, "grad_norm": 1.5675705671310425, "learning_rate": 1.4620508778740853e-05, "loss": 0.9096, "step": 10421 }, { "epoch": 0.3672401490525648, "grad_norm": 6.534097194671631, "learning_rate": 1.4619496603417929e-05, "loss": 2.9696, "step": 10422 }, { "epoch": 0.36727538606552323, "grad_norm": 4.6550140380859375, "learning_rate": 1.461848436792547e-05, "loss": 3.1621, "step": 10423 }, { "epoch": 0.3673106230784816, "grad_norm": 11.598894119262695, "learning_rate": 1.4617472072276661e-05, "loss": 3.6182, "step": 10424 }, { "epoch": 0.36734586009144005, "grad_norm": 3.4480514526367188, "learning_rate": 1.4616459716484691e-05, "loss": 3.0401, "step": 10425 }, { "epoch": 0.3673810971043985, "grad_norm": 2.0045814514160156, "learning_rate": 1.4615447300562749e-05, "loss": 0.9758, "step": 10426 }, { "epoch": 0.36741633411735686, "grad_norm": 4.191074848175049, "learning_rate": 1.4614434824524014e-05, "loss": 3.5082, "step": 10427 }, { "epoch": 0.3674515711303153, "grad_norm": 1.5879048109054565, "learning_rate": 1.461342228838168e-05, "loss": 0.7841, "step": 10428 }, { "epoch": 0.36748680814327367, "grad_norm": 5.078024387359619, "learning_rate": 1.461240969214893e-05, "loss": 3.15, "step": 10429 }, { "epoch": 0.3675220451562321, "grad_norm": 5.942127704620361, "learning_rate": 1.461139703583896e-05, "loss": 2.8711, "step": 10430 }, { "epoch": 0.36755728216919054, "grad_norm": 1.4215748310089111, "learning_rate": 1.4610384319464954e-05, "loss": 0.7622, "step": 10431 }, { "epoch": 0.3675925191821489, "grad_norm": 5.33424186706543, "learning_rate": 1.4609371543040105e-05, "loss": 3.1763, "step": 10432 }, { "epoch": 0.36762775619510735, "grad_norm": 1.0740598440170288, "learning_rate": 1.4608358706577606e-05, "loss": 0.9752, "step": 10433 }, { "epoch": 0.3676629932080657, "grad_norm": 2.819196939468384, "learning_rate": 1.4607345810090647e-05, "loss": 3.2094, "step": 10434 }, { "epoch": 0.36769823022102416, "grad_norm": 1.1878150701522827, "learning_rate": 1.4606332853592421e-05, "loss": 1.0564, "step": 10435 }, { "epoch": 0.3677334672339826, "grad_norm": 5.281088829040527, "learning_rate": 1.4605319837096123e-05, "loss": 3.1603, "step": 10436 }, { "epoch": 0.36776870424694097, "grad_norm": 2.394906520843506, "learning_rate": 1.4604306760614948e-05, "loss": 1.0139, "step": 10437 }, { "epoch": 0.3678039412598994, "grad_norm": 1.4083694219589233, "learning_rate": 1.4603293624162091e-05, "loss": 0.8603, "step": 10438 }, { "epoch": 0.36783917827285784, "grad_norm": 1.2317663431167603, "learning_rate": 1.460228042775075e-05, "loss": 1.357, "step": 10439 }, { "epoch": 0.3678744152858162, "grad_norm": 6.421639442443848, "learning_rate": 1.4601267171394118e-05, "loss": 6.6787, "step": 10440 }, { "epoch": 0.36790965229877465, "grad_norm": 3.1150693893432617, "learning_rate": 1.4600253855105394e-05, "loss": 2.7906, "step": 10441 }, { "epoch": 0.367944889311733, "grad_norm": 1.0326694250106812, "learning_rate": 1.459924047889778e-05, "loss": 0.9659, "step": 10442 }, { "epoch": 0.36798012632469146, "grad_norm": 1.326810359954834, "learning_rate": 1.4598227042784472e-05, "loss": 0.9348, "step": 10443 }, { "epoch": 0.3680153633376499, "grad_norm": 1.5113518238067627, "learning_rate": 1.459721354677867e-05, "loss": 0.9654, "step": 10444 }, { "epoch": 0.36805060035060827, "grad_norm": 4.452908039093018, "learning_rate": 1.4596199990893579e-05, "loss": 3.1786, "step": 10445 }, { "epoch": 0.3680858373635667, "grad_norm": 2.323134660720825, "learning_rate": 1.4595186375142394e-05, "loss": 0.9424, "step": 10446 }, { "epoch": 0.3681210743765251, "grad_norm": 9.782255172729492, "learning_rate": 1.4594172699538324e-05, "loss": 6.4423, "step": 10447 }, { "epoch": 0.3681563113894835, "grad_norm": 6.463655948638916, "learning_rate": 1.4593158964094567e-05, "loss": 3.3588, "step": 10448 }, { "epoch": 0.36819154840244195, "grad_norm": 2.8923940658569336, "learning_rate": 1.4592145168824332e-05, "loss": 2.5507, "step": 10449 }, { "epoch": 0.3682267854154003, "grad_norm": 1.5751233100891113, "learning_rate": 1.4591131313740819e-05, "loss": 1.1285, "step": 10450 }, { "epoch": 0.36826202242835876, "grad_norm": 1.8615449666976929, "learning_rate": 1.4590117398857238e-05, "loss": 0.8164, "step": 10451 }, { "epoch": 0.36829725944131714, "grad_norm": 4.240100860595703, "learning_rate": 1.458910342418679e-05, "loss": 3.9542, "step": 10452 }, { "epoch": 0.36833249645427557, "grad_norm": 1.2490485906600952, "learning_rate": 1.4588089389742685e-05, "loss": 0.8337, "step": 10453 }, { "epoch": 0.368367733467234, "grad_norm": 6.105207920074463, "learning_rate": 1.4587075295538135e-05, "loss": 5.2303, "step": 10454 }, { "epoch": 0.3684029704801924, "grad_norm": 2.02795672416687, "learning_rate": 1.4586061141586344e-05, "loss": 0.8919, "step": 10455 }, { "epoch": 0.3684382074931508, "grad_norm": 1.5641587972640991, "learning_rate": 1.4585046927900522e-05, "loss": 0.9103, "step": 10456 }, { "epoch": 0.3684734445061092, "grad_norm": 0.977938175201416, "learning_rate": 1.4584032654493878e-05, "loss": 1.154, "step": 10457 }, { "epoch": 0.3685086815190676, "grad_norm": 10.890549659729004, "learning_rate": 1.4583018321379626e-05, "loss": 3.7329, "step": 10458 }, { "epoch": 0.36854391853202606, "grad_norm": 1.07062566280365, "learning_rate": 1.4582003928570977e-05, "loss": 1.045, "step": 10459 }, { "epoch": 0.36857915554498444, "grad_norm": 3.1866703033447266, "learning_rate": 1.4580989476081144e-05, "loss": 3.402, "step": 10460 }, { "epoch": 0.36861439255794287, "grad_norm": 0.9596766233444214, "learning_rate": 1.4579974963923338e-05, "loss": 0.8488, "step": 10461 }, { "epoch": 0.36864962957090125, "grad_norm": 1.2854713201522827, "learning_rate": 1.4578960392110775e-05, "loss": 1.0346, "step": 10462 }, { "epoch": 0.3686848665838597, "grad_norm": 5.204065799713135, "learning_rate": 1.457794576065667e-05, "loss": 5.2636, "step": 10463 }, { "epoch": 0.3687201035968181, "grad_norm": 1.6696791648864746, "learning_rate": 1.4576931069574238e-05, "loss": 1.0532, "step": 10464 }, { "epoch": 0.3687553406097765, "grad_norm": 5.956007957458496, "learning_rate": 1.4575916318876697e-05, "loss": 3.4102, "step": 10465 }, { "epoch": 0.3687905776227349, "grad_norm": 1.7063239812850952, "learning_rate": 1.4574901508577263e-05, "loss": 0.7819, "step": 10466 }, { "epoch": 0.3688258146356933, "grad_norm": 2.2584829330444336, "learning_rate": 1.457388663868915e-05, "loss": 0.7685, "step": 10467 }, { "epoch": 0.36886105164865174, "grad_norm": 1.0648629665374756, "learning_rate": 1.4572871709225587e-05, "loss": 0.9967, "step": 10468 }, { "epoch": 0.36889628866161017, "grad_norm": 5.811822891235352, "learning_rate": 1.4571856720199784e-05, "loss": 6.1633, "step": 10469 }, { "epoch": 0.36893152567456855, "grad_norm": 1.1403928995132446, "learning_rate": 1.4570841671624965e-05, "loss": 1.1614, "step": 10470 }, { "epoch": 0.368966762687527, "grad_norm": 7.069975852966309, "learning_rate": 1.456982656351435e-05, "loss": 5.5141, "step": 10471 }, { "epoch": 0.3690019997004854, "grad_norm": 1.173077940940857, "learning_rate": 1.4568811395881163e-05, "loss": 0.9268, "step": 10472 }, { "epoch": 0.3690372367134438, "grad_norm": 1.2036372423171997, "learning_rate": 1.4567796168738628e-05, "loss": 0.8269, "step": 10473 }, { "epoch": 0.36907247372640223, "grad_norm": 3.386296510696411, "learning_rate": 1.4566780882099965e-05, "loss": 2.6555, "step": 10474 }, { "epoch": 0.3691077107393606, "grad_norm": 5.362489223480225, "learning_rate": 1.4565765535978398e-05, "loss": 2.7851, "step": 10475 }, { "epoch": 0.36914294775231904, "grad_norm": 10.757070541381836, "learning_rate": 1.4564750130387153e-05, "loss": 5.7476, "step": 10476 }, { "epoch": 0.3691781847652775, "grad_norm": 4.410133361816406, "learning_rate": 1.4563734665339458e-05, "loss": 2.4557, "step": 10477 }, { "epoch": 0.36921342177823585, "grad_norm": 4.376803398132324, "learning_rate": 1.4562719140848534e-05, "loss": 3.1968, "step": 10478 }, { "epoch": 0.3692486587911943, "grad_norm": 5.7581095695495605, "learning_rate": 1.4561703556927615e-05, "loss": 4.3491, "step": 10479 }, { "epoch": 0.36928389580415266, "grad_norm": 6.6447343826293945, "learning_rate": 1.4560687913589925e-05, "loss": 3.2143, "step": 10480 }, { "epoch": 0.3693191328171111, "grad_norm": 1.2397265434265137, "learning_rate": 1.4559672210848695e-05, "loss": 1.1263, "step": 10481 }, { "epoch": 0.36935436983006953, "grad_norm": 2.9258217811584473, "learning_rate": 1.4558656448717152e-05, "loss": 2.2512, "step": 10482 }, { "epoch": 0.3693896068430279, "grad_norm": 1.3646756410598755, "learning_rate": 1.455764062720853e-05, "loss": 0.9087, "step": 10483 }, { "epoch": 0.36942484385598634, "grad_norm": 3.690810203552246, "learning_rate": 1.4556624746336057e-05, "loss": 3.219, "step": 10484 }, { "epoch": 0.3694600808689447, "grad_norm": 1.252368450164795, "learning_rate": 1.4555608806112969e-05, "loss": 0.8189, "step": 10485 }, { "epoch": 0.36949531788190315, "grad_norm": 1.1280540227890015, "learning_rate": 1.4554592806552492e-05, "loss": 1.0205, "step": 10486 }, { "epoch": 0.3695305548948616, "grad_norm": 12.156963348388672, "learning_rate": 1.4553576747667866e-05, "loss": 5.5019, "step": 10487 }, { "epoch": 0.36956579190781996, "grad_norm": 1.2310947179794312, "learning_rate": 1.4552560629472324e-05, "loss": 0.9578, "step": 10488 }, { "epoch": 0.3696010289207784, "grad_norm": 1.1734133958816528, "learning_rate": 1.4551544451979096e-05, "loss": 0.7409, "step": 10489 }, { "epoch": 0.3696362659337368, "grad_norm": 1.0857499837875366, "learning_rate": 1.4550528215201423e-05, "loss": 1.0845, "step": 10490 }, { "epoch": 0.3696715029466952, "grad_norm": 1.1646426916122437, "learning_rate": 1.4549511919152541e-05, "loss": 1.2009, "step": 10491 }, { "epoch": 0.36970673995965364, "grad_norm": 4.485355377197266, "learning_rate": 1.4548495563845685e-05, "loss": 5.5114, "step": 10492 }, { "epoch": 0.369741976972612, "grad_norm": 1.5503571033477783, "learning_rate": 1.4547479149294099e-05, "loss": 0.7041, "step": 10493 }, { "epoch": 0.36977721398557045, "grad_norm": 1.9402066469192505, "learning_rate": 1.4546462675511015e-05, "loss": 0.7877, "step": 10494 }, { "epoch": 0.36981245099852883, "grad_norm": 1.4513989686965942, "learning_rate": 1.4545446142509675e-05, "loss": 0.8288, "step": 10495 }, { "epoch": 0.36984768801148726, "grad_norm": 1.2684882879257202, "learning_rate": 1.4544429550303323e-05, "loss": 1.0633, "step": 10496 }, { "epoch": 0.3698829250244457, "grad_norm": 7.823393821716309, "learning_rate": 1.4543412898905194e-05, "loss": 5.7081, "step": 10497 }, { "epoch": 0.3699181620374041, "grad_norm": 4.490879535675049, "learning_rate": 1.4542396188328535e-05, "loss": 2.8131, "step": 10498 }, { "epoch": 0.3699533990503625, "grad_norm": 6.3081889152526855, "learning_rate": 1.4541379418586585e-05, "loss": 3.0456, "step": 10499 }, { "epoch": 0.3699886360633209, "grad_norm": 1.7406864166259766, "learning_rate": 1.4540362589692595e-05, "loss": 1.0149, "step": 10500 }, { "epoch": 0.3700238730762793, "grad_norm": 7.832437515258789, "learning_rate": 1.4539345701659798e-05, "loss": 8.5109, "step": 10501 }, { "epoch": 0.37005911008923775, "grad_norm": 6.252984046936035, "learning_rate": 1.4538328754501448e-05, "loss": 3.6804, "step": 10502 }, { "epoch": 0.37009434710219613, "grad_norm": 1.499105453491211, "learning_rate": 1.4537311748230785e-05, "loss": 1.0799, "step": 10503 }, { "epoch": 0.37012958411515456, "grad_norm": 1.5220952033996582, "learning_rate": 1.4536294682861062e-05, "loss": 0.8276, "step": 10504 }, { "epoch": 0.370164821128113, "grad_norm": 4.160373210906982, "learning_rate": 1.453527755840552e-05, "loss": 2.8456, "step": 10505 }, { "epoch": 0.3702000581410714, "grad_norm": 2.099308967590332, "learning_rate": 1.4534260374877414e-05, "loss": 0.7996, "step": 10506 }, { "epoch": 0.3702352951540298, "grad_norm": 5.006701469421387, "learning_rate": 1.4533243132289984e-05, "loss": 3.2903, "step": 10507 }, { "epoch": 0.3702705321669882, "grad_norm": 1.769585132598877, "learning_rate": 1.4532225830656488e-05, "loss": 0.6828, "step": 10508 }, { "epoch": 0.3703057691799466, "grad_norm": 7.700740814208984, "learning_rate": 1.4531208469990172e-05, "loss": 6.1894, "step": 10509 }, { "epoch": 0.37034100619290505, "grad_norm": 1.9274442195892334, "learning_rate": 1.453019105030429e-05, "loss": 1.0822, "step": 10510 }, { "epoch": 0.37037624320586343, "grad_norm": 3.437110662460327, "learning_rate": 1.4529173571612091e-05, "loss": 2.7462, "step": 10511 }, { "epoch": 0.37041148021882186, "grad_norm": 6.0783281326293945, "learning_rate": 1.452815603392683e-05, "loss": 3.6042, "step": 10512 }, { "epoch": 0.37044671723178024, "grad_norm": 6.0783281326293945, "learning_rate": 1.452815603392683e-05, "loss": 5.1678, "step": 10513 }, { "epoch": 0.3704819542447387, "grad_norm": 1.3030383586883545, "learning_rate": 1.452713843726176e-05, "loss": 0.85, "step": 10514 }, { "epoch": 0.3705171912576971, "grad_norm": 1.460493564605713, "learning_rate": 1.4526120781630133e-05, "loss": 1.0957, "step": 10515 }, { "epoch": 0.3705524282706555, "grad_norm": 4.576269626617432, "learning_rate": 1.4525103067045212e-05, "loss": 3.7754, "step": 10516 }, { "epoch": 0.3705876652836139, "grad_norm": 1.5659301280975342, "learning_rate": 1.452408529352024e-05, "loss": 1.1419, "step": 10517 }, { "epoch": 0.3706229022965723, "grad_norm": 1.3115102052688599, "learning_rate": 1.4523067461068485e-05, "loss": 0.6842, "step": 10518 }, { "epoch": 0.37065813930953073, "grad_norm": 1.4159799814224243, "learning_rate": 1.4522049569703198e-05, "loss": 0.9667, "step": 10519 }, { "epoch": 0.37069337632248917, "grad_norm": 2.1379570960998535, "learning_rate": 1.4521031619437642e-05, "loss": 0.863, "step": 10520 }, { "epoch": 0.37072861333544754, "grad_norm": 1.2443890571594238, "learning_rate": 1.4520013610285072e-05, "loss": 1.1491, "step": 10521 }, { "epoch": 0.370763850348406, "grad_norm": 1.2443890571594238, "learning_rate": 1.4520013610285072e-05, "loss": 3.5486, "step": 10522 }, { "epoch": 0.37079908736136435, "grad_norm": 15.074946403503418, "learning_rate": 1.4518995542258749e-05, "loss": 3.5083, "step": 10523 }, { "epoch": 0.3708343243743228, "grad_norm": 15.074946403503418, "learning_rate": 1.4518995542258749e-05, "loss": 3.6542, "step": 10524 }, { "epoch": 0.3708695613872812, "grad_norm": 1.7023570537567139, "learning_rate": 1.4517977415371933e-05, "loss": 1.032, "step": 10525 }, { "epoch": 0.3709047984002396, "grad_norm": 5.946228504180908, "learning_rate": 1.4516959229637887e-05, "loss": 3.6977, "step": 10526 }, { "epoch": 0.37094003541319803, "grad_norm": 3.9112629890441895, "learning_rate": 1.451594098506987e-05, "loss": 2.9545, "step": 10527 }, { "epoch": 0.3709752724261564, "grad_norm": 1.3211191892623901, "learning_rate": 1.451492268168115e-05, "loss": 1.0125, "step": 10528 }, { "epoch": 0.37101050943911484, "grad_norm": 1.1466792821884155, "learning_rate": 1.4513904319484979e-05, "loss": 0.8329, "step": 10529 }, { "epoch": 0.3710457464520733, "grad_norm": 3.6329331398010254, "learning_rate": 1.4512885898494637e-05, "loss": 3.1343, "step": 10530 }, { "epoch": 0.37108098346503166, "grad_norm": 1.6419233083724976, "learning_rate": 1.4511867418723377e-05, "loss": 0.7895, "step": 10531 }, { "epoch": 0.3711162204779901, "grad_norm": 12.930123329162598, "learning_rate": 1.4510848880184471e-05, "loss": 3.4059, "step": 10532 }, { "epoch": 0.37115145749094847, "grad_norm": 3.0552406311035156, "learning_rate": 1.4509830282891184e-05, "loss": 3.4164, "step": 10533 }, { "epoch": 0.3711866945039069, "grad_norm": 0.8926681876182556, "learning_rate": 1.4508811626856783e-05, "loss": 0.8889, "step": 10534 }, { "epoch": 0.37122193151686533, "grad_norm": 1.6754461526870728, "learning_rate": 1.4507792912094536e-05, "loss": 1.0122, "step": 10535 }, { "epoch": 0.3712571685298237, "grad_norm": 1.6953643560409546, "learning_rate": 1.4506774138617715e-05, "loss": 0.8268, "step": 10536 }, { "epoch": 0.37129240554278214, "grad_norm": 0.9849350452423096, "learning_rate": 1.4505755306439583e-05, "loss": 0.7221, "step": 10537 }, { "epoch": 0.3713276425557405, "grad_norm": 7.85754919052124, "learning_rate": 1.4504736415573412e-05, "loss": 2.9605, "step": 10538 }, { "epoch": 0.37136287956869896, "grad_norm": 1.0242459774017334, "learning_rate": 1.4503717466032479e-05, "loss": 0.8262, "step": 10539 }, { "epoch": 0.3713981165816574, "grad_norm": 2.408547878265381, "learning_rate": 1.4502698457830052e-05, "loss": 0.7971, "step": 10540 }, { "epoch": 0.37143335359461577, "grad_norm": 1.4204556941986084, "learning_rate": 1.4501679390979403e-05, "loss": 1.1579, "step": 10541 }, { "epoch": 0.3714685906075742, "grad_norm": 3.124969720840454, "learning_rate": 1.4500660265493806e-05, "loss": 3.3629, "step": 10542 }, { "epoch": 0.37150382762053263, "grad_norm": 1.5417765378952026, "learning_rate": 1.4499641081386537e-05, "loss": 0.9402, "step": 10543 }, { "epoch": 0.371539064633491, "grad_norm": 1.2875514030456543, "learning_rate": 1.449862183867087e-05, "loss": 0.9298, "step": 10544 }, { "epoch": 0.37157430164644945, "grad_norm": 10.555071830749512, "learning_rate": 1.4497602537360082e-05, "loss": 5.3126, "step": 10545 }, { "epoch": 0.3716095386594078, "grad_norm": 3.8560187816619873, "learning_rate": 1.4496583177467444e-05, "loss": 5.4021, "step": 10546 }, { "epoch": 0.37164477567236626, "grad_norm": 7.118328094482422, "learning_rate": 1.4495563759006242e-05, "loss": 3.6667, "step": 10547 }, { "epoch": 0.3716800126853247, "grad_norm": 4.134183883666992, "learning_rate": 1.4494544281989743e-05, "loss": 3.0544, "step": 10548 }, { "epoch": 0.37171524969828307, "grad_norm": 4.019898414611816, "learning_rate": 1.449352474643124e-05, "loss": 5.086, "step": 10549 }, { "epoch": 0.3717504867112415, "grad_norm": 2.3783974647521973, "learning_rate": 1.4492505152344001e-05, "loss": 0.9221, "step": 10550 }, { "epoch": 0.3717857237241999, "grad_norm": 4.8121795654296875, "learning_rate": 1.449148549974131e-05, "loss": 3.2959, "step": 10551 }, { "epoch": 0.3718209607371583, "grad_norm": 1.6043118238449097, "learning_rate": 1.4490465788636448e-05, "loss": 0.7966, "step": 10552 }, { "epoch": 0.37185619775011675, "grad_norm": 1.8840631246566772, "learning_rate": 1.44894460190427e-05, "loss": 0.9019, "step": 10553 }, { "epoch": 0.3718914347630751, "grad_norm": 1.6861528158187866, "learning_rate": 1.4488426190973339e-05, "loss": 1.1009, "step": 10554 }, { "epoch": 0.37192667177603356, "grad_norm": 1.945855736732483, "learning_rate": 1.4487406304441662e-05, "loss": 1.1415, "step": 10555 }, { "epoch": 0.37196190878899194, "grad_norm": 1.783528447151184, "learning_rate": 1.4486386359460944e-05, "loss": 1.1318, "step": 10556 }, { "epoch": 0.37199714580195037, "grad_norm": 1.1719245910644531, "learning_rate": 1.448536635604447e-05, "loss": 0.9018, "step": 10557 }, { "epoch": 0.3720323828149088, "grad_norm": 1.0772744417190552, "learning_rate": 1.4484346294205529e-05, "loss": 0.9417, "step": 10558 }, { "epoch": 0.3720676198278672, "grad_norm": 1.7685096263885498, "learning_rate": 1.4483326173957408e-05, "loss": 0.8669, "step": 10559 }, { "epoch": 0.3721028568408256, "grad_norm": 0.9899843335151672, "learning_rate": 1.448230599531339e-05, "loss": 0.8896, "step": 10560 }, { "epoch": 0.372138093853784, "grad_norm": 1.1597414016723633, "learning_rate": 1.448128575828677e-05, "loss": 1.0491, "step": 10561 }, { "epoch": 0.3721733308667424, "grad_norm": 2.4596855640411377, "learning_rate": 1.4480265462890826e-05, "loss": 1.0342, "step": 10562 }, { "epoch": 0.37220856787970086, "grad_norm": 9.399710655212402, "learning_rate": 1.4479245109138857e-05, "loss": 5.1325, "step": 10563 }, { "epoch": 0.37224380489265924, "grad_norm": 10.644402503967285, "learning_rate": 1.4478224697044151e-05, "loss": 5.9437, "step": 10564 }, { "epoch": 0.37227904190561767, "grad_norm": 1.476502537727356, "learning_rate": 1.4477204226619996e-05, "loss": 1.0966, "step": 10565 }, { "epoch": 0.37231427891857605, "grad_norm": 1.5929375886917114, "learning_rate": 1.4476183697879687e-05, "loss": 1.0603, "step": 10566 }, { "epoch": 0.3723495159315345, "grad_norm": 1.2710713148117065, "learning_rate": 1.4475163110836511e-05, "loss": 0.8647, "step": 10567 }, { "epoch": 0.3723847529444929, "grad_norm": 0.9981525540351868, "learning_rate": 1.4474142465503774e-05, "loss": 0.9212, "step": 10568 }, { "epoch": 0.3724199899574513, "grad_norm": 13.455281257629395, "learning_rate": 1.4473121761894753e-05, "loss": 7.3404, "step": 10569 }, { "epoch": 0.3724552269704097, "grad_norm": 1.431351900100708, "learning_rate": 1.4472101000022756e-05, "loss": 0.9967, "step": 10570 }, { "epoch": 0.3724904639833681, "grad_norm": 1.2740789651870728, "learning_rate": 1.4471080179901072e-05, "loss": 1.0038, "step": 10571 }, { "epoch": 0.37252570099632654, "grad_norm": 1.2150639295578003, "learning_rate": 1.4470059301543001e-05, "loss": 1.026, "step": 10572 }, { "epoch": 0.37256093800928497, "grad_norm": 5.2797980308532715, "learning_rate": 1.4469038364961834e-05, "loss": 3.6823, "step": 10573 }, { "epoch": 0.37259617502224335, "grad_norm": 1.3765392303466797, "learning_rate": 1.4468017370170877e-05, "loss": 0.9152, "step": 10574 }, { "epoch": 0.3726314120352018, "grad_norm": 6.657410621643066, "learning_rate": 1.4466996317183421e-05, "loss": 3.0121, "step": 10575 }, { "epoch": 0.3726666490481602, "grad_norm": 1.2708359956741333, "learning_rate": 1.4465975206012775e-05, "loss": 0.9104, "step": 10576 }, { "epoch": 0.3727018860611186, "grad_norm": 1.0389267206192017, "learning_rate": 1.4464954036672227e-05, "loss": 0.8849, "step": 10577 }, { "epoch": 0.372737123074077, "grad_norm": 1.1220546960830688, "learning_rate": 1.4463932809175087e-05, "loss": 0.9926, "step": 10578 }, { "epoch": 0.3727723600870354, "grad_norm": 3.594273090362549, "learning_rate": 1.4462911523534653e-05, "loss": 2.7325, "step": 10579 }, { "epoch": 0.37280759709999384, "grad_norm": 1.0939985513687134, "learning_rate": 1.4461890179764226e-05, "loss": 0.828, "step": 10580 }, { "epoch": 0.37284283411295227, "grad_norm": 12.659948348999023, "learning_rate": 1.4460868777877113e-05, "loss": 4.3137, "step": 10581 }, { "epoch": 0.37287807112591065, "grad_norm": 1.9020260572433472, "learning_rate": 1.4459847317886615e-05, "loss": 0.8379, "step": 10582 }, { "epoch": 0.3729133081388691, "grad_norm": 3.97782564163208, "learning_rate": 1.4458825799806042e-05, "loss": 3.1537, "step": 10583 }, { "epoch": 0.37294854515182746, "grad_norm": 5.346160888671875, "learning_rate": 1.4457804223648689e-05, "loss": 3.967, "step": 10584 }, { "epoch": 0.3729837821647859, "grad_norm": 12.874044418334961, "learning_rate": 1.445678258942787e-05, "loss": 5.971, "step": 10585 }, { "epoch": 0.3730190191777443, "grad_norm": 10.239158630371094, "learning_rate": 1.445576089715689e-05, "loss": 3.7722, "step": 10586 }, { "epoch": 0.3730542561907027, "grad_norm": 5.127707004547119, "learning_rate": 1.445473914684906e-05, "loss": 3.5892, "step": 10587 }, { "epoch": 0.37308949320366114, "grad_norm": 1.878564476966858, "learning_rate": 1.4453717338517681e-05, "loss": 0.879, "step": 10588 }, { "epoch": 0.3731247302166195, "grad_norm": 4.368464946746826, "learning_rate": 1.4452695472176068e-05, "loss": 3.0332, "step": 10589 }, { "epoch": 0.37315996722957795, "grad_norm": 20.458995819091797, "learning_rate": 1.4451673547837527e-05, "loss": 4.0674, "step": 10590 }, { "epoch": 0.3731952042425364, "grad_norm": 4.877396583557129, "learning_rate": 1.4450651565515377e-05, "loss": 3.3508, "step": 10591 }, { "epoch": 0.37323044125549476, "grad_norm": 17.35704803466797, "learning_rate": 1.4449629525222917e-05, "loss": 8.9443, "step": 10592 }, { "epoch": 0.3732656782684532, "grad_norm": 5.720044136047363, "learning_rate": 1.4448607426973471e-05, "loss": 5.5434, "step": 10593 }, { "epoch": 0.37330091528141157, "grad_norm": 11.692803382873535, "learning_rate": 1.444758527078034e-05, "loss": 3.9411, "step": 10594 }, { "epoch": 0.37333615229437, "grad_norm": 1.4338454008102417, "learning_rate": 1.444656305665685e-05, "loss": 0.9697, "step": 10595 }, { "epoch": 0.37337138930732844, "grad_norm": 15.496726036071777, "learning_rate": 1.4445540784616308e-05, "loss": 3.3479, "step": 10596 }, { "epoch": 0.3734066263202868, "grad_norm": 1.518175721168518, "learning_rate": 1.4444518454672029e-05, "loss": 0.7107, "step": 10597 }, { "epoch": 0.37344186333324525, "grad_norm": 8.897214889526367, "learning_rate": 1.4443496066837334e-05, "loss": 3.3115, "step": 10598 }, { "epoch": 0.3734771003462036, "grad_norm": 26.605697631835938, "learning_rate": 1.4442473621125537e-05, "loss": 3.8438, "step": 10599 }, { "epoch": 0.37351233735916206, "grad_norm": 13.525625228881836, "learning_rate": 1.4441451117549952e-05, "loss": 3.1679, "step": 10600 }, { "epoch": 0.3735475743721205, "grad_norm": 17.54606819152832, "learning_rate": 1.4440428556123903e-05, "loss": 10.623, "step": 10601 }, { "epoch": 0.3735828113850789, "grad_norm": 7.938886642456055, "learning_rate": 1.4439405936860705e-05, "loss": 3.8864, "step": 10602 }, { "epoch": 0.3736180483980373, "grad_norm": 1.4690481424331665, "learning_rate": 1.4438383259773679e-05, "loss": 1.0384, "step": 10603 }, { "epoch": 0.3736532854109957, "grad_norm": 5.721198081970215, "learning_rate": 1.443736052487615e-05, "loss": 2.3939, "step": 10604 }, { "epoch": 0.3736885224239541, "grad_norm": 2.3881635665893555, "learning_rate": 1.4436337732181426e-05, "loss": 0.9038, "step": 10605 }, { "epoch": 0.37372375943691255, "grad_norm": 39.23617935180664, "learning_rate": 1.4435314881702847e-05, "loss": 5.7974, "step": 10606 }, { "epoch": 0.37375899644987093, "grad_norm": 1.207310676574707, "learning_rate": 1.4434291973453723e-05, "loss": 0.988, "step": 10607 }, { "epoch": 0.37379423346282936, "grad_norm": 2.262864828109741, "learning_rate": 1.4433269007447381e-05, "loss": 1.1544, "step": 10608 }, { "epoch": 0.3738294704757878, "grad_norm": 4.935370922088623, "learning_rate": 1.4432245983697146e-05, "loss": 3.1994, "step": 10609 }, { "epoch": 0.3738647074887462, "grad_norm": 55.065914154052734, "learning_rate": 1.4431222902216344e-05, "loss": 3.1075, "step": 10610 }, { "epoch": 0.3738999445017046, "grad_norm": 1.1712759733200073, "learning_rate": 1.4430199763018296e-05, "loss": 0.9067, "step": 10611 }, { "epoch": 0.373935181514663, "grad_norm": 6.645938873291016, "learning_rate": 1.4429176566116334e-05, "loss": 3.0447, "step": 10612 }, { "epoch": 0.3739704185276214, "grad_norm": 2.330868721008301, "learning_rate": 1.442815331152378e-05, "loss": 1.1741, "step": 10613 }, { "epoch": 0.37400565554057985, "grad_norm": 17.102691650390625, "learning_rate": 1.442712999925397e-05, "loss": 0.6548, "step": 10614 }, { "epoch": 0.37404089255353823, "grad_norm": 8.2559814453125, "learning_rate": 1.4426106629320225e-05, "loss": 6.1435, "step": 10615 }, { "epoch": 0.37407612956649666, "grad_norm": 1.139709234237671, "learning_rate": 1.4425083201735877e-05, "loss": 1.2422, "step": 10616 }, { "epoch": 0.37411136657945504, "grad_norm": 6.567777633666992, "learning_rate": 1.4424059716514256e-05, "loss": 3.3767, "step": 10617 }, { "epoch": 0.3741466035924135, "grad_norm": 5.364134311676025, "learning_rate": 1.4423036173668696e-05, "loss": 4.085, "step": 10618 }, { "epoch": 0.3741818406053719, "grad_norm": 2.403998374938965, "learning_rate": 1.4422012573212525e-05, "loss": 0.8791, "step": 10619 }, { "epoch": 0.3742170776183303, "grad_norm": 2.284428596496582, "learning_rate": 1.4420988915159076e-05, "loss": 1.0519, "step": 10620 }, { "epoch": 0.3742523146312887, "grad_norm": 1.297821044921875, "learning_rate": 1.4419965199521686e-05, "loss": 0.9683, "step": 10621 }, { "epoch": 0.3742875516442471, "grad_norm": 1.7389107942581177, "learning_rate": 1.4418941426313684e-05, "loss": 1.2295, "step": 10622 }, { "epoch": 0.37432278865720553, "grad_norm": 6.311460494995117, "learning_rate": 1.4417917595548409e-05, "loss": 3.4722, "step": 10623 }, { "epoch": 0.37435802567016396, "grad_norm": 2.7519021034240723, "learning_rate": 1.441689370723919e-05, "loss": 1.0362, "step": 10624 }, { "epoch": 0.37439326268312234, "grad_norm": 6.22509765625, "learning_rate": 1.4415869761399375e-05, "loss": 2.9974, "step": 10625 }, { "epoch": 0.3744284996960808, "grad_norm": 14.02938175201416, "learning_rate": 1.4414845758042288e-05, "loss": 3.8375, "step": 10626 }, { "epoch": 0.37446373670903915, "grad_norm": 10.270332336425781, "learning_rate": 1.4413821697181275e-05, "loss": 3.5099, "step": 10627 }, { "epoch": 0.3744989737219976, "grad_norm": 4.903720378875732, "learning_rate": 1.4412797578829672e-05, "loss": 2.9244, "step": 10628 }, { "epoch": 0.374534210734956, "grad_norm": 4.770397663116455, "learning_rate": 1.441177340300082e-05, "loss": 4.1366, "step": 10629 }, { "epoch": 0.3745694477479144, "grad_norm": 1.2051525115966797, "learning_rate": 1.4410749169708053e-05, "loss": 1.0308, "step": 10630 }, { "epoch": 0.37460468476087283, "grad_norm": 2.1036059856414795, "learning_rate": 1.4409724878964722e-05, "loss": 1.1951, "step": 10631 }, { "epoch": 0.3746399217738312, "grad_norm": 9.559819221496582, "learning_rate": 1.4408700530784157e-05, "loss": 6.5206, "step": 10632 }, { "epoch": 0.37467515878678964, "grad_norm": 1.5410345792770386, "learning_rate": 1.440767612517971e-05, "loss": 1.2247, "step": 10633 }, { "epoch": 0.3747103957997481, "grad_norm": 3.0008749961853027, "learning_rate": 1.440665166216472e-05, "loss": 3.409, "step": 10634 }, { "epoch": 0.37474563281270645, "grad_norm": 4.6901350021362305, "learning_rate": 1.4405627141752528e-05, "loss": 3.4883, "step": 10635 }, { "epoch": 0.3747808698256649, "grad_norm": 10.642053604125977, "learning_rate": 1.4404602563956483e-05, "loss": 6.5484, "step": 10636 }, { "epoch": 0.37481610683862326, "grad_norm": 5.243148326873779, "learning_rate": 1.4403577928789929e-05, "loss": 5.3979, "step": 10637 }, { "epoch": 0.3748513438515817, "grad_norm": 9.754955291748047, "learning_rate": 1.440255323626621e-05, "loss": 3.1128, "step": 10638 }, { "epoch": 0.37488658086454013, "grad_norm": 21.053865432739258, "learning_rate": 1.4401528486398674e-05, "loss": 4.3097, "step": 10639 }, { "epoch": 0.3749218178774985, "grad_norm": 1.6375010013580322, "learning_rate": 1.440050367920067e-05, "loss": 1.1579, "step": 10640 }, { "epoch": 0.37495705489045694, "grad_norm": 14.887720108032227, "learning_rate": 1.4399478814685547e-05, "loss": 5.0945, "step": 10641 }, { "epoch": 0.3749922919034153, "grad_norm": 20.124284744262695, "learning_rate": 1.4398453892866651e-05, "loss": 6.348, "step": 10642 }, { "epoch": 0.37502752891637375, "grad_norm": 1.8371695280075073, "learning_rate": 1.439742891375733e-05, "loss": 0.9184, "step": 10643 }, { "epoch": 0.3750627659293322, "grad_norm": 1.1311931610107422, "learning_rate": 1.439640387737094e-05, "loss": 1.0339, "step": 10644 }, { "epoch": 0.37509800294229056, "grad_norm": 22.33289337158203, "learning_rate": 1.4395378783720828e-05, "loss": 5.2483, "step": 10645 }, { "epoch": 0.375133239955249, "grad_norm": 1.3826745748519897, "learning_rate": 1.4394353632820348e-05, "loss": 0.8229, "step": 10646 }, { "epoch": 0.37516847696820743, "grad_norm": 1.810214638710022, "learning_rate": 1.4393328424682853e-05, "loss": 0.896, "step": 10647 }, { "epoch": 0.3752037139811658, "grad_norm": 6.011012554168701, "learning_rate": 1.4392303159321694e-05, "loss": 3.4932, "step": 10648 }, { "epoch": 0.37523895099412424, "grad_norm": 1.3694289922714233, "learning_rate": 1.4391277836750227e-05, "loss": 1.0133, "step": 10649 }, { "epoch": 0.3752741880070826, "grad_norm": 5.191201686859131, "learning_rate": 1.439025245698181e-05, "loss": 2.8092, "step": 10650 }, { "epoch": 0.37530942502004105, "grad_norm": 1.1268948316574097, "learning_rate": 1.438922702002979e-05, "loss": 0.8, "step": 10651 }, { "epoch": 0.3753446620329995, "grad_norm": 18.447856903076172, "learning_rate": 1.4388201525907534e-05, "loss": 6.4916, "step": 10652 }, { "epoch": 0.37537989904595787, "grad_norm": 1.8350725173950195, "learning_rate": 1.4387175974628393e-05, "loss": 1.1638, "step": 10653 }, { "epoch": 0.3754151360589163, "grad_norm": 1.9038208723068237, "learning_rate": 1.4386150366205724e-05, "loss": 0.8953, "step": 10654 }, { "epoch": 0.3754503730718747, "grad_norm": 6.712723255157471, "learning_rate": 1.4385124700652889e-05, "loss": 3.6514, "step": 10655 }, { "epoch": 0.3754856100848331, "grad_norm": 21.108911514282227, "learning_rate": 1.4384098977983248e-05, "loss": 3.1736, "step": 10656 }, { "epoch": 0.37552084709779154, "grad_norm": 1.4405349493026733, "learning_rate": 1.4383073198210156e-05, "loss": 0.8927, "step": 10657 }, { "epoch": 0.3755560841107499, "grad_norm": 13.34354019165039, "learning_rate": 1.438204736134698e-05, "loss": 3.5691, "step": 10658 }, { "epoch": 0.37559132112370835, "grad_norm": 1.3840705156326294, "learning_rate": 1.4381021467407077e-05, "loss": 1.0404, "step": 10659 }, { "epoch": 0.37562655813666673, "grad_norm": 1.5396353006362915, "learning_rate": 1.437999551640381e-05, "loss": 1.0619, "step": 10660 }, { "epoch": 0.37566179514962517, "grad_norm": 18.464479446411133, "learning_rate": 1.437896950835055e-05, "loss": 3.0049, "step": 10661 }, { "epoch": 0.3756970321625836, "grad_norm": 1.4593786001205444, "learning_rate": 1.4377943443260647e-05, "loss": 1.0709, "step": 10662 }, { "epoch": 0.375732269175542, "grad_norm": 12.767948150634766, "learning_rate": 1.4376917321147476e-05, "loss": 3.4721, "step": 10663 }, { "epoch": 0.3757675061885004, "grad_norm": 14.991242408752441, "learning_rate": 1.43758911420244e-05, "loss": 4.067, "step": 10664 }, { "epoch": 0.3758027432014588, "grad_norm": 17.307363510131836, "learning_rate": 1.4374864905904785e-05, "loss": 5.487, "step": 10665 }, { "epoch": 0.3758379802144172, "grad_norm": 2.4246888160705566, "learning_rate": 1.4373838612801996e-05, "loss": 2.7188, "step": 10666 }, { "epoch": 0.37587321722737566, "grad_norm": 16.573074340820312, "learning_rate": 1.43728122627294e-05, "loss": 4.2166, "step": 10667 }, { "epoch": 0.37590845424033403, "grad_norm": 18.210002899169922, "learning_rate": 1.437178585570037e-05, "loss": 6.599, "step": 10668 }, { "epoch": 0.37594369125329247, "grad_norm": 1.349717140197754, "learning_rate": 1.4370759391728273e-05, "loss": 1.0521, "step": 10669 }, { "epoch": 0.37597892826625084, "grad_norm": 7.395325183868408, "learning_rate": 1.4369732870826475e-05, "loss": 5.5244, "step": 10670 }, { "epoch": 0.3760141652792093, "grad_norm": 1.216787576675415, "learning_rate": 1.4368706293008354e-05, "loss": 0.927, "step": 10671 }, { "epoch": 0.3760494022921677, "grad_norm": 1.5087789297103882, "learning_rate": 1.4367679658287275e-05, "loss": 1.1274, "step": 10672 }, { "epoch": 0.3760846393051261, "grad_norm": 2.1359949111938477, "learning_rate": 1.4366652966676611e-05, "loss": 1.0808, "step": 10673 }, { "epoch": 0.3761198763180845, "grad_norm": 4.283449649810791, "learning_rate": 1.4365626218189737e-05, "loss": 2.8931, "step": 10674 }, { "epoch": 0.3761551133310429, "grad_norm": 4.882185935974121, "learning_rate": 1.4364599412840025e-05, "loss": 5.4916, "step": 10675 }, { "epoch": 0.37619035034400133, "grad_norm": 1.3039569854736328, "learning_rate": 1.436357255064085e-05, "loss": 0.8377, "step": 10676 }, { "epoch": 0.37622558735695977, "grad_norm": 1.5494787693023682, "learning_rate": 1.4362545631605587e-05, "loss": 1.1256, "step": 10677 }, { "epoch": 0.37626082436991815, "grad_norm": 10.29871940612793, "learning_rate": 1.436151865574761e-05, "loss": 3.638, "step": 10678 }, { "epoch": 0.3762960613828766, "grad_norm": 1.2263426780700684, "learning_rate": 1.4360491623080297e-05, "loss": 1.0505, "step": 10679 }, { "epoch": 0.376331298395835, "grad_norm": 4.375054359436035, "learning_rate": 1.4359464533617028e-05, "loss": 3.8376, "step": 10680 }, { "epoch": 0.3763665354087934, "grad_norm": 3.752326250076294, "learning_rate": 1.4358437387371173e-05, "loss": 3.369, "step": 10681 }, { "epoch": 0.3764017724217518, "grad_norm": 4.582734107971191, "learning_rate": 1.4357410184356123e-05, "loss": 3.5568, "step": 10682 }, { "epoch": 0.3764370094347102, "grad_norm": 7.76741361618042, "learning_rate": 1.4356382924585245e-05, "loss": 4.0037, "step": 10683 }, { "epoch": 0.37647224644766863, "grad_norm": 13.14501667022705, "learning_rate": 1.4355355608071926e-05, "loss": 3.3012, "step": 10684 }, { "epoch": 0.37650748346062707, "grad_norm": 2.950963020324707, "learning_rate": 1.4354328234829545e-05, "loss": 3.2184, "step": 10685 }, { "epoch": 0.37654272047358545, "grad_norm": 5.9899163246154785, "learning_rate": 1.4353300804871484e-05, "loss": 6.0726, "step": 10686 }, { "epoch": 0.3765779574865439, "grad_norm": 5.516660690307617, "learning_rate": 1.4352273318211124e-05, "loss": 4.9935, "step": 10687 }, { "epoch": 0.37661319449950226, "grad_norm": 5.163483142852783, "learning_rate": 1.4351245774861854e-05, "loss": 3.1813, "step": 10688 }, { "epoch": 0.3766484315124607, "grad_norm": 1.0841025114059448, "learning_rate": 1.435021817483705e-05, "loss": 1.0609, "step": 10689 }, { "epoch": 0.3766836685254191, "grad_norm": 10.50275707244873, "learning_rate": 1.4349190518150102e-05, "loss": 5.5402, "step": 10690 }, { "epoch": 0.3767189055383775, "grad_norm": 1.50184166431427, "learning_rate": 1.4348162804814391e-05, "loss": 0.9431, "step": 10691 }, { "epoch": 0.37675414255133594, "grad_norm": 1.6942580938339233, "learning_rate": 1.434713503484331e-05, "loss": 0.8742, "step": 10692 }, { "epoch": 0.3767893795642943, "grad_norm": 1.4191203117370605, "learning_rate": 1.4346107208250237e-05, "loss": 1.3611, "step": 10693 }, { "epoch": 0.37682461657725275, "grad_norm": 1.4060790538787842, "learning_rate": 1.4345079325048565e-05, "loss": 1.1012, "step": 10694 }, { "epoch": 0.3768598535902112, "grad_norm": 4.700819492340088, "learning_rate": 1.4344051385251682e-05, "loss": 3.5646, "step": 10695 }, { "epoch": 0.37689509060316956, "grad_norm": 4.43441104888916, "learning_rate": 1.4343023388872978e-05, "loss": 3.0353, "step": 10696 }, { "epoch": 0.376930327616128, "grad_norm": 3.9069111347198486, "learning_rate": 1.4341995335925837e-05, "loss": 3.433, "step": 10697 }, { "epoch": 0.37696556462908637, "grad_norm": 1.1638329029083252, "learning_rate": 1.4340967226423659e-05, "loss": 0.7644, "step": 10698 }, { "epoch": 0.3770008016420448, "grad_norm": 10.872297286987305, "learning_rate": 1.4339939060379827e-05, "loss": 3.8453, "step": 10699 }, { "epoch": 0.37703603865500324, "grad_norm": 1.0398422479629517, "learning_rate": 1.4338910837807735e-05, "loss": 0.8991, "step": 10700 }, { "epoch": 0.3770712756679616, "grad_norm": 3.9121999740600586, "learning_rate": 1.4337882558720777e-05, "loss": 3.0309, "step": 10701 }, { "epoch": 0.37710651268092005, "grad_norm": 1.2300971746444702, "learning_rate": 1.4336854223132348e-05, "loss": 1.1026, "step": 10702 }, { "epoch": 0.3771417496938784, "grad_norm": 6.825073719024658, "learning_rate": 1.4335825831055838e-05, "loss": 3.0283, "step": 10703 }, { "epoch": 0.37717698670683686, "grad_norm": 16.66377830505371, "learning_rate": 1.4334797382504647e-05, "loss": 8.6743, "step": 10704 }, { "epoch": 0.3772122237197953, "grad_norm": 1.1373944282531738, "learning_rate": 1.4333768877492165e-05, "loss": 0.8281, "step": 10705 }, { "epoch": 0.37724746073275367, "grad_norm": 9.410042762756348, "learning_rate": 1.4332740316031793e-05, "loss": 3.233, "step": 10706 }, { "epoch": 0.3772826977457121, "grad_norm": 8.656877517700195, "learning_rate": 1.4331711698136927e-05, "loss": 4.144, "step": 10707 }, { "epoch": 0.3773179347586705, "grad_norm": 5.8354716300964355, "learning_rate": 1.4330683023820962e-05, "loss": 3.2807, "step": 10708 }, { "epoch": 0.3773531717716289, "grad_norm": 0.9573930501937866, "learning_rate": 1.4329654293097302e-05, "loss": 0.7835, "step": 10709 }, { "epoch": 0.37738840878458735, "grad_norm": 4.2107834815979, "learning_rate": 1.4328625505979341e-05, "loss": 3.1126, "step": 10710 }, { "epoch": 0.3774236457975457, "grad_norm": 1.537591814994812, "learning_rate": 1.4327596662480483e-05, "loss": 0.9128, "step": 10711 }, { "epoch": 0.37745888281050416, "grad_norm": 2.011549711227417, "learning_rate": 1.4326567762614126e-05, "loss": 0.7401, "step": 10712 }, { "epoch": 0.3774941198234626, "grad_norm": 2.611651659011841, "learning_rate": 1.4325538806393673e-05, "loss": 0.9131, "step": 10713 }, { "epoch": 0.37752935683642097, "grad_norm": 0.9667840003967285, "learning_rate": 1.4324509793832527e-05, "loss": 0.8483, "step": 10714 }, { "epoch": 0.3775645938493794, "grad_norm": 1.3193787336349487, "learning_rate": 1.432348072494409e-05, "loss": 0.8503, "step": 10715 }, { "epoch": 0.3775998308623378, "grad_norm": 1.5308482646942139, "learning_rate": 1.4322451599741763e-05, "loss": 1.1172, "step": 10716 }, { "epoch": 0.3776350678752962, "grad_norm": 2.1106343269348145, "learning_rate": 1.4321422418238957e-05, "loss": 0.8382, "step": 10717 }, { "epoch": 0.37767030488825465, "grad_norm": 5.994121074676514, "learning_rate": 1.432039318044907e-05, "loss": 3.75, "step": 10718 }, { "epoch": 0.377705541901213, "grad_norm": 4.924140930175781, "learning_rate": 1.4319363886385516e-05, "loss": 3.786, "step": 10719 }, { "epoch": 0.37774077891417146, "grad_norm": 5.5641703605651855, "learning_rate": 1.4318334536061693e-05, "loss": 3.0222, "step": 10720 }, { "epoch": 0.37777601592712984, "grad_norm": 1.444085717201233, "learning_rate": 1.4317305129491015e-05, "loss": 0.8489, "step": 10721 }, { "epoch": 0.37781125294008827, "grad_norm": 2.0561811923980713, "learning_rate": 1.4316275666686887e-05, "loss": 0.8154, "step": 10722 }, { "epoch": 0.3778464899530467, "grad_norm": 5.379786968231201, "learning_rate": 1.4315246147662717e-05, "loss": 3.4889, "step": 10723 }, { "epoch": 0.3778817269660051, "grad_norm": 3.583686351776123, "learning_rate": 1.4314216572431918e-05, "loss": 3.2595, "step": 10724 }, { "epoch": 0.3779169639789635, "grad_norm": 5.502138614654541, "learning_rate": 1.4313186941007899e-05, "loss": 4.767, "step": 10725 }, { "epoch": 0.3779522009919219, "grad_norm": 1.204689621925354, "learning_rate": 1.4312157253404071e-05, "loss": 0.8914, "step": 10726 }, { "epoch": 0.3779874380048803, "grad_norm": 4.132047176361084, "learning_rate": 1.4311127509633842e-05, "loss": 3.0162, "step": 10727 }, { "epoch": 0.37802267501783876, "grad_norm": 1.9810168743133545, "learning_rate": 1.4310097709710633e-05, "loss": 0.9698, "step": 10728 }, { "epoch": 0.37805791203079714, "grad_norm": 2.522099733352661, "learning_rate": 1.430906785364785e-05, "loss": 0.9483, "step": 10729 }, { "epoch": 0.37809314904375557, "grad_norm": 3.963608503341675, "learning_rate": 1.430803794145891e-05, "loss": 3.1674, "step": 10730 }, { "epoch": 0.37812838605671395, "grad_norm": 1.9677202701568604, "learning_rate": 1.4307007973157225e-05, "loss": 0.9914, "step": 10731 }, { "epoch": 0.3781636230696724, "grad_norm": 5.0417962074279785, "learning_rate": 1.4305977948756216e-05, "loss": 6.0138, "step": 10732 }, { "epoch": 0.3781988600826308, "grad_norm": 6.1043701171875, "learning_rate": 1.4304947868269293e-05, "loss": 3.8651, "step": 10733 }, { "epoch": 0.3782340970955892, "grad_norm": 2.3797974586486816, "learning_rate": 1.4303917731709878e-05, "loss": 0.8054, "step": 10734 }, { "epoch": 0.37826933410854763, "grad_norm": 3.7411839962005615, "learning_rate": 1.4302887539091381e-05, "loss": 3.2471, "step": 10735 }, { "epoch": 0.378304571121506, "grad_norm": 1.5412907600402832, "learning_rate": 1.4301857290427232e-05, "loss": 0.8744, "step": 10736 }, { "epoch": 0.37833980813446444, "grad_norm": 1.1952029466629028, "learning_rate": 1.430082698573084e-05, "loss": 1.0869, "step": 10737 }, { "epoch": 0.3783750451474229, "grad_norm": 5.040562152862549, "learning_rate": 1.429979662501563e-05, "loss": 3.4404, "step": 10738 }, { "epoch": 0.37841028216038125, "grad_norm": 1.2670345306396484, "learning_rate": 1.429876620829502e-05, "loss": 1.4001, "step": 10739 }, { "epoch": 0.3784455191733397, "grad_norm": 1.7131730318069458, "learning_rate": 1.4297735735582433e-05, "loss": 0.9731, "step": 10740 }, { "epoch": 0.37848075618629806, "grad_norm": 4.438293933868408, "learning_rate": 1.4296705206891292e-05, "loss": 3.4136, "step": 10741 }, { "epoch": 0.3785159931992565, "grad_norm": 4.855162143707275, "learning_rate": 1.429567462223502e-05, "loss": 3.3032, "step": 10742 }, { "epoch": 0.37855123021221493, "grad_norm": 1.7630614042282104, "learning_rate": 1.4294643981627034e-05, "loss": 0.8741, "step": 10743 }, { "epoch": 0.3785864672251733, "grad_norm": 1.305914282798767, "learning_rate": 1.4293613285080765e-05, "loss": 1.0045, "step": 10744 }, { "epoch": 0.37862170423813174, "grad_norm": 1.7201064825057983, "learning_rate": 1.4292582532609638e-05, "loss": 0.8307, "step": 10745 }, { "epoch": 0.3786569412510901, "grad_norm": 7.299482822418213, "learning_rate": 1.4291551724227075e-05, "loss": 3.9086, "step": 10746 }, { "epoch": 0.37869217826404855, "grad_norm": 1.310677170753479, "learning_rate": 1.4290520859946506e-05, "loss": 0.7407, "step": 10747 }, { "epoch": 0.378727415277007, "grad_norm": 8.921011924743652, "learning_rate": 1.4289489939781352e-05, "loss": 3.483, "step": 10748 }, { "epoch": 0.37876265228996536, "grad_norm": 9.735416412353516, "learning_rate": 1.4288458963745048e-05, "loss": 6.4481, "step": 10749 }, { "epoch": 0.3787978893029238, "grad_norm": 1.5939167737960815, "learning_rate": 1.428742793185102e-05, "loss": 0.8414, "step": 10750 }, { "epoch": 0.37883312631588223, "grad_norm": 8.504985809326172, "learning_rate": 1.4286396844112698e-05, "loss": 3.2344, "step": 10751 }, { "epoch": 0.3788683633288406, "grad_norm": 9.568028450012207, "learning_rate": 1.428536570054351e-05, "loss": 5.4638, "step": 10752 }, { "epoch": 0.37890360034179904, "grad_norm": 11.160357475280762, "learning_rate": 1.4284334501156889e-05, "loss": 5.9605, "step": 10753 }, { "epoch": 0.3789388373547574, "grad_norm": 1.6410706043243408, "learning_rate": 1.428330324596626e-05, "loss": 0.9953, "step": 10754 }, { "epoch": 0.37897407436771585, "grad_norm": 18.09587287902832, "learning_rate": 1.4282271934985068e-05, "loss": 3.0787, "step": 10755 }, { "epoch": 0.3790093113806743, "grad_norm": 5.0154709815979, "learning_rate": 1.4281240568226736e-05, "loss": 3.1722, "step": 10756 }, { "epoch": 0.37904454839363266, "grad_norm": 4.799598217010498, "learning_rate": 1.4280209145704697e-05, "loss": 3.0631, "step": 10757 }, { "epoch": 0.3790797854065911, "grad_norm": 9.60887336730957, "learning_rate": 1.4279177667432392e-05, "loss": 3.7816, "step": 10758 }, { "epoch": 0.3791150224195495, "grad_norm": 1.153032660484314, "learning_rate": 1.4278146133423251e-05, "loss": 0.8683, "step": 10759 }, { "epoch": 0.3791502594325079, "grad_norm": 1.399544358253479, "learning_rate": 1.4277114543690714e-05, "loss": 0.8633, "step": 10760 }, { "epoch": 0.37918549644546634, "grad_norm": 3.476900815963745, "learning_rate": 1.4276082898248213e-05, "loss": 2.9018, "step": 10761 }, { "epoch": 0.3792207334584247, "grad_norm": 1.2284276485443115, "learning_rate": 1.4275051197109189e-05, "loss": 0.9093, "step": 10762 }, { "epoch": 0.37925597047138315, "grad_norm": 1.5851727724075317, "learning_rate": 1.4274019440287077e-05, "loss": 1.0636, "step": 10763 }, { "epoch": 0.37929120748434153, "grad_norm": 5.1993088722229, "learning_rate": 1.4272987627795321e-05, "loss": 3.4401, "step": 10764 }, { "epoch": 0.37932644449729996, "grad_norm": 8.01624870300293, "learning_rate": 1.4271955759647352e-05, "loss": 3.7616, "step": 10765 }, { "epoch": 0.3793616815102584, "grad_norm": 7.804971694946289, "learning_rate": 1.4270923835856619e-05, "loss": 6.3944, "step": 10766 }, { "epoch": 0.3793969185232168, "grad_norm": 2.2045130729675293, "learning_rate": 1.4269891856436555e-05, "loss": 1.0539, "step": 10767 }, { "epoch": 0.3794321555361752, "grad_norm": 23.64512825012207, "learning_rate": 1.426885982140061e-05, "loss": 3.7507, "step": 10768 }, { "epoch": 0.3794673925491336, "grad_norm": 15.113848686218262, "learning_rate": 1.426782773076222e-05, "loss": 5.7287, "step": 10769 }, { "epoch": 0.379502629562092, "grad_norm": 4.435062408447266, "learning_rate": 1.426679558453483e-05, "loss": 3.9037, "step": 10770 }, { "epoch": 0.37953786657505045, "grad_norm": 1.573401689529419, "learning_rate": 1.4265763382731883e-05, "loss": 0.7954, "step": 10771 }, { "epoch": 0.37957310358800883, "grad_norm": 5.3494038581848145, "learning_rate": 1.4264731125366828e-05, "loss": 3.3672, "step": 10772 }, { "epoch": 0.37960834060096726, "grad_norm": 8.029213905334473, "learning_rate": 1.4263698812453101e-05, "loss": 6.13, "step": 10773 }, { "epoch": 0.37964357761392564, "grad_norm": 6.188083171844482, "learning_rate": 1.4262666444004161e-05, "loss": 3.1183, "step": 10774 }, { "epoch": 0.3796788146268841, "grad_norm": 1.6348527669906616, "learning_rate": 1.4261634020033441e-05, "loss": 0.9599, "step": 10775 }, { "epoch": 0.3797140516398425, "grad_norm": 1.546088695526123, "learning_rate": 1.4260601540554402e-05, "loss": 0.7974, "step": 10776 }, { "epoch": 0.3797492886528009, "grad_norm": 1.1988147497177124, "learning_rate": 1.425956900558048e-05, "loss": 1.013, "step": 10777 }, { "epoch": 0.3797845256657593, "grad_norm": 17.57123374938965, "learning_rate": 1.4258536415125131e-05, "loss": 5.3494, "step": 10778 }, { "epoch": 0.3798197626787177, "grad_norm": 1.383798360824585, "learning_rate": 1.4257503769201802e-05, "loss": 1.0635, "step": 10779 }, { "epoch": 0.37985499969167613, "grad_norm": 18.157133102416992, "learning_rate": 1.4256471067823943e-05, "loss": 3.6632, "step": 10780 }, { "epoch": 0.37989023670463457, "grad_norm": 2.102217435836792, "learning_rate": 1.4255438311005007e-05, "loss": 0.8575, "step": 10781 }, { "epoch": 0.37992547371759294, "grad_norm": 9.478995323181152, "learning_rate": 1.4254405498758447e-05, "loss": 5.9904, "step": 10782 }, { "epoch": 0.3799607107305514, "grad_norm": 3.2829320430755615, "learning_rate": 1.4253372631097713e-05, "loss": 3.1024, "step": 10783 }, { "epoch": 0.3799959477435098, "grad_norm": 8.86417007446289, "learning_rate": 1.4252339708036256e-05, "loss": 3.0175, "step": 10784 }, { "epoch": 0.3800311847564682, "grad_norm": 1.0383012294769287, "learning_rate": 1.4251306729587535e-05, "loss": 1.1949, "step": 10785 }, { "epoch": 0.3800664217694266, "grad_norm": 8.812934875488281, "learning_rate": 1.4250273695764998e-05, "loss": 5.4276, "step": 10786 }, { "epoch": 0.380101658782385, "grad_norm": 8.987570762634277, "learning_rate": 1.4249240606582112e-05, "loss": 3.3267, "step": 10787 }, { "epoch": 0.38013689579534343, "grad_norm": 1.0401721000671387, "learning_rate": 1.4248207462052322e-05, "loss": 0.9636, "step": 10788 }, { "epoch": 0.38017213280830187, "grad_norm": 4.849245071411133, "learning_rate": 1.4247174262189088e-05, "loss": 3.9558, "step": 10789 }, { "epoch": 0.38020736982126024, "grad_norm": 13.029733657836914, "learning_rate": 1.4246141007005867e-05, "loss": 3.6318, "step": 10790 }, { "epoch": 0.3802426068342187, "grad_norm": 1.2859976291656494, "learning_rate": 1.4245107696516124e-05, "loss": 0.7975, "step": 10791 }, { "epoch": 0.38027784384717706, "grad_norm": 7.975731372833252, "learning_rate": 1.4244074330733306e-05, "loss": 3.8206, "step": 10792 }, { "epoch": 0.3803130808601355, "grad_norm": 4.8983378410339355, "learning_rate": 1.4243040909670885e-05, "loss": 2.943, "step": 10793 }, { "epoch": 0.3803483178730939, "grad_norm": 6.043821811676025, "learning_rate": 1.4242007433342311e-05, "loss": 3.8362, "step": 10794 }, { "epoch": 0.3803835548860523, "grad_norm": 7.452822685241699, "learning_rate": 1.4240973901761053e-05, "loss": 6.4157, "step": 10795 }, { "epoch": 0.38041879189901073, "grad_norm": 1.5494245290756226, "learning_rate": 1.4239940314940568e-05, "loss": 1.1663, "step": 10796 }, { "epoch": 0.3804540289119691, "grad_norm": 1.4519718885421753, "learning_rate": 1.423890667289432e-05, "loss": 1.2258, "step": 10797 }, { "epoch": 0.38048926592492754, "grad_norm": 10.598193168640137, "learning_rate": 1.4237872975635774e-05, "loss": 3.0697, "step": 10798 }, { "epoch": 0.380524502937886, "grad_norm": 11.136409759521484, "learning_rate": 1.4236839223178391e-05, "loss": 5.2035, "step": 10799 }, { "epoch": 0.38055973995084436, "grad_norm": 1.9876238107681274, "learning_rate": 1.4235805415535637e-05, "loss": 0.7741, "step": 10800 }, { "epoch": 0.3805949769638028, "grad_norm": 4.113376617431641, "learning_rate": 1.423477155272098e-05, "loss": 3.0739, "step": 10801 }, { "epoch": 0.38063021397676117, "grad_norm": 3.995314359664917, "learning_rate": 1.4233737634747887e-05, "loss": 3.495, "step": 10802 }, { "epoch": 0.3806654509897196, "grad_norm": 22.0662784576416, "learning_rate": 1.4232703661629814e-05, "loss": 8.3375, "step": 10803 }, { "epoch": 0.38070068800267803, "grad_norm": 1.166625738143921, "learning_rate": 1.4231669633380245e-05, "loss": 1.0103, "step": 10804 }, { "epoch": 0.3807359250156364, "grad_norm": 1.7844833135604858, "learning_rate": 1.4230635550012633e-05, "loss": 0.9812, "step": 10805 }, { "epoch": 0.38077116202859485, "grad_norm": 2.2192776203155518, "learning_rate": 1.4229601411540459e-05, "loss": 0.7881, "step": 10806 }, { "epoch": 0.3808063990415532, "grad_norm": 5.68696928024292, "learning_rate": 1.4228567217977184e-05, "loss": 3.4545, "step": 10807 }, { "epoch": 0.38084163605451166, "grad_norm": 1.420994520187378, "learning_rate": 1.4227532969336287e-05, "loss": 0.9387, "step": 10808 }, { "epoch": 0.3808768730674701, "grad_norm": 1.303897500038147, "learning_rate": 1.4226498665631232e-05, "loss": 0.9752, "step": 10809 }, { "epoch": 0.38091211008042847, "grad_norm": 17.17397117614746, "learning_rate": 1.4225464306875494e-05, "loss": 3.3102, "step": 10810 }, { "epoch": 0.3809473470933869, "grad_norm": 2.2549753189086914, "learning_rate": 1.4224429893082544e-05, "loss": 0.7788, "step": 10811 }, { "epoch": 0.3809825841063453, "grad_norm": 14.800678253173828, "learning_rate": 1.4223395424265859e-05, "loss": 2.6862, "step": 10812 }, { "epoch": 0.3810178211193037, "grad_norm": 1.5408169031143188, "learning_rate": 1.4222360900438905e-05, "loss": 0.9574, "step": 10813 }, { "epoch": 0.38105305813226215, "grad_norm": 1.8353761434555054, "learning_rate": 1.422132632161517e-05, "loss": 1.2514, "step": 10814 }, { "epoch": 0.3810882951452205, "grad_norm": 2.667342185974121, "learning_rate": 1.4220291687808116e-05, "loss": 0.7433, "step": 10815 }, { "epoch": 0.38112353215817896, "grad_norm": 0.8944810032844543, "learning_rate": 1.4219256999031228e-05, "loss": 1.1109, "step": 10816 }, { "epoch": 0.3811587691711374, "grad_norm": 18.531553268432617, "learning_rate": 1.4218222255297979e-05, "loss": 11.6511, "step": 10817 }, { "epoch": 0.38119400618409577, "grad_norm": 1.1487427949905396, "learning_rate": 1.4217187456621848e-05, "loss": 0.8479, "step": 10818 }, { "epoch": 0.3812292431970542, "grad_norm": 3.7694592475891113, "learning_rate": 1.4216152603016313e-05, "loss": 3.2587, "step": 10819 }, { "epoch": 0.3812644802100126, "grad_norm": 4.044424533843994, "learning_rate": 1.4215117694494853e-05, "loss": 3.5457, "step": 10820 }, { "epoch": 0.381299717222971, "grad_norm": 2.3569419384002686, "learning_rate": 1.4214082731070951e-05, "loss": 0.8681, "step": 10821 }, { "epoch": 0.38133495423592945, "grad_norm": 8.021970748901367, "learning_rate": 1.421304771275808e-05, "loss": 5.6348, "step": 10822 }, { "epoch": 0.3813701912488878, "grad_norm": 3.2763354778289795, "learning_rate": 1.4212012639569732e-05, "loss": 2.8821, "step": 10823 }, { "epoch": 0.38140542826184626, "grad_norm": 1.456457257270813, "learning_rate": 1.4210977511519377e-05, "loss": 0.9554, "step": 10824 }, { "epoch": 0.38144066527480464, "grad_norm": 3.796743869781494, "learning_rate": 1.4209942328620507e-05, "loss": 3.679, "step": 10825 }, { "epoch": 0.38147590228776307, "grad_norm": 0.9629637598991394, "learning_rate": 1.4208907090886601e-05, "loss": 1.0011, "step": 10826 }, { "epoch": 0.3815111393007215, "grad_norm": 3.875692367553711, "learning_rate": 1.4207871798331144e-05, "loss": 3.0348, "step": 10827 }, { "epoch": 0.3815463763136799, "grad_norm": 1.2723647356033325, "learning_rate": 1.420683645096762e-05, "loss": 1.1164, "step": 10828 }, { "epoch": 0.3815816133266383, "grad_norm": 8.968310356140137, "learning_rate": 1.4205801048809518e-05, "loss": 2.6022, "step": 10829 }, { "epoch": 0.3816168503395967, "grad_norm": 3.748972177505493, "learning_rate": 1.4204765591870317e-05, "loss": 2.8689, "step": 10830 }, { "epoch": 0.3816520873525551, "grad_norm": 1.1681772470474243, "learning_rate": 1.4203730080163513e-05, "loss": 0.9258, "step": 10831 }, { "epoch": 0.38168732436551356, "grad_norm": 3.332172155380249, "learning_rate": 1.4202694513702586e-05, "loss": 2.9201, "step": 10832 }, { "epoch": 0.38172256137847194, "grad_norm": 4.724442005157471, "learning_rate": 1.4201658892501029e-05, "loss": 3.4749, "step": 10833 }, { "epoch": 0.38175779839143037, "grad_norm": 1.8030946254730225, "learning_rate": 1.420062321657233e-05, "loss": 0.9767, "step": 10834 }, { "epoch": 0.38179303540438875, "grad_norm": 23.148725509643555, "learning_rate": 1.4199587485929978e-05, "loss": 3.0666, "step": 10835 }, { "epoch": 0.3818282724173472, "grad_norm": 1.0441824197769165, "learning_rate": 1.4198551700587464e-05, "loss": 1.0118, "step": 10836 }, { "epoch": 0.3818635094303056, "grad_norm": 4.272619247436523, "learning_rate": 1.419751586055828e-05, "loss": 3.206, "step": 10837 }, { "epoch": 0.381898746443264, "grad_norm": 2.097198247909546, "learning_rate": 1.4196479965855916e-05, "loss": 1.2136, "step": 10838 }, { "epoch": 0.3819339834562224, "grad_norm": 10.736981391906738, "learning_rate": 1.4195444016493866e-05, "loss": 3.933, "step": 10839 }, { "epoch": 0.3819692204691808, "grad_norm": 7.747353553771973, "learning_rate": 1.4194408012485626e-05, "loss": 3.3299, "step": 10840 }, { "epoch": 0.38200445748213924, "grad_norm": 1.8638224601745605, "learning_rate": 1.4193371953844684e-05, "loss": 0.8298, "step": 10841 }, { "epoch": 0.38203969449509767, "grad_norm": 4.482132434844971, "learning_rate": 1.4192335840584542e-05, "loss": 3.4531, "step": 10842 }, { "epoch": 0.38207493150805605, "grad_norm": 1.572183609008789, "learning_rate": 1.4191299672718688e-05, "loss": 0.7889, "step": 10843 }, { "epoch": 0.3821101685210145, "grad_norm": 1.1886762380599976, "learning_rate": 1.4190263450260623e-05, "loss": 0.7669, "step": 10844 }, { "epoch": 0.38214540553397286, "grad_norm": 1.105283260345459, "learning_rate": 1.4189227173223844e-05, "loss": 1.1544, "step": 10845 }, { "epoch": 0.3821806425469313, "grad_norm": 1.570574402809143, "learning_rate": 1.4188190841621845e-05, "loss": 0.8, "step": 10846 }, { "epoch": 0.3822158795598897, "grad_norm": 1.4334808588027954, "learning_rate": 1.4187154455468128e-05, "loss": 0.924, "step": 10847 }, { "epoch": 0.3822511165728481, "grad_norm": 1.400528907775879, "learning_rate": 1.418611801477619e-05, "loss": 0.8519, "step": 10848 }, { "epoch": 0.38228635358580654, "grad_norm": 4.444968223571777, "learning_rate": 1.4185081519559534e-05, "loss": 3.6153, "step": 10849 }, { "epoch": 0.3823215905987649, "grad_norm": 6.642810821533203, "learning_rate": 1.4184044969831656e-05, "loss": 2.6115, "step": 10850 }, { "epoch": 0.38235682761172335, "grad_norm": 1.313970923423767, "learning_rate": 1.4183008365606059e-05, "loss": 0.9516, "step": 10851 }, { "epoch": 0.3823920646246818, "grad_norm": 1.4650897979736328, "learning_rate": 1.4181971706896249e-05, "loss": 1.0177, "step": 10852 }, { "epoch": 0.38242730163764016, "grad_norm": 7.158013343811035, "learning_rate": 1.4180934993715721e-05, "loss": 2.7976, "step": 10853 }, { "epoch": 0.3824625386505986, "grad_norm": 7.794042110443115, "learning_rate": 1.4179898226077983e-05, "loss": 3.654, "step": 10854 }, { "epoch": 0.382497775663557, "grad_norm": 4.253039360046387, "learning_rate": 1.4178861403996537e-05, "loss": 3.1839, "step": 10855 }, { "epoch": 0.3825330126765154, "grad_norm": 13.300849914550781, "learning_rate": 1.4177824527484893e-05, "loss": 6.3997, "step": 10856 }, { "epoch": 0.38256824968947384, "grad_norm": 6.984902858734131, "learning_rate": 1.4176787596556547e-05, "loss": 5.6206, "step": 10857 }, { "epoch": 0.3826034867024322, "grad_norm": 0.9195894598960876, "learning_rate": 1.4175750611225016e-05, "loss": 0.7928, "step": 10858 }, { "epoch": 0.38263872371539065, "grad_norm": 0.9063577651977539, "learning_rate": 1.4174713571503796e-05, "loss": 0.8516, "step": 10859 }, { "epoch": 0.3826739607283491, "grad_norm": 1.1557010412216187, "learning_rate": 1.4173676477406404e-05, "loss": 0.8838, "step": 10860 }, { "epoch": 0.38270919774130746, "grad_norm": 1.2592766284942627, "learning_rate": 1.4172639328946343e-05, "loss": 0.9726, "step": 10861 }, { "epoch": 0.3827444347542659, "grad_norm": 6.910470485687256, "learning_rate": 1.4171602126137123e-05, "loss": 5.7066, "step": 10862 }, { "epoch": 0.38277967176722427, "grad_norm": 2.2119827270507812, "learning_rate": 1.4170564868992252e-05, "loss": 0.7271, "step": 10863 }, { "epoch": 0.3828149087801827, "grad_norm": 9.744364738464355, "learning_rate": 1.4169527557525244e-05, "loss": 7.3292, "step": 10864 }, { "epoch": 0.38285014579314114, "grad_norm": 1.2501747608184814, "learning_rate": 1.4168490191749609e-05, "loss": 1.0119, "step": 10865 }, { "epoch": 0.3828853828060995, "grad_norm": 1.8436756134033203, "learning_rate": 1.4167452771678858e-05, "loss": 0.9328, "step": 10866 }, { "epoch": 0.38292061981905795, "grad_norm": 9.116514205932617, "learning_rate": 1.4166415297326503e-05, "loss": 3.2227, "step": 10867 }, { "epoch": 0.38295585683201633, "grad_norm": 3.9469988346099854, "learning_rate": 1.416537776870606e-05, "loss": 3.5276, "step": 10868 }, { "epoch": 0.38299109384497476, "grad_norm": 8.179473876953125, "learning_rate": 1.416434018583104e-05, "loss": 5.7442, "step": 10869 }, { "epoch": 0.3830263308579332, "grad_norm": 1.347286581993103, "learning_rate": 1.4163302548714956e-05, "loss": 1.0377, "step": 10870 }, { "epoch": 0.3830615678708916, "grad_norm": 1.4839648008346558, "learning_rate": 1.4162264857371331e-05, "loss": 1.2264, "step": 10871 }, { "epoch": 0.38309680488385, "grad_norm": 1.0427038669586182, "learning_rate": 1.4161227111813673e-05, "loss": 0.8361, "step": 10872 }, { "epoch": 0.3831320418968084, "grad_norm": 4.582397937774658, "learning_rate": 1.4160189312055502e-05, "loss": 3.4093, "step": 10873 }, { "epoch": 0.3831672789097668, "grad_norm": 1.2422740459442139, "learning_rate": 1.4159151458110337e-05, "loss": 1.0572, "step": 10874 }, { "epoch": 0.38320251592272525, "grad_norm": 5.409600257873535, "learning_rate": 1.4158113549991693e-05, "loss": 4.6555, "step": 10875 }, { "epoch": 0.38323775293568363, "grad_norm": 6.02150821685791, "learning_rate": 1.4157075587713087e-05, "loss": 2.8466, "step": 10876 }, { "epoch": 0.38327298994864206, "grad_norm": 2.5284204483032227, "learning_rate": 1.4156037571288048e-05, "loss": 3.1384, "step": 10877 }, { "epoch": 0.38330822696160044, "grad_norm": 8.935088157653809, "learning_rate": 1.4154999500730084e-05, "loss": 2.7667, "step": 10878 }, { "epoch": 0.3833434639745589, "grad_norm": 4.1535115242004395, "learning_rate": 1.4153961376052729e-05, "loss": 0.6581, "step": 10879 }, { "epoch": 0.3833787009875173, "grad_norm": 7.624721050262451, "learning_rate": 1.4152923197269495e-05, "loss": 3.3579, "step": 10880 }, { "epoch": 0.3834139380004757, "grad_norm": 8.520866394042969, "learning_rate": 1.4151884964393908e-05, "loss": 2.619, "step": 10881 }, { "epoch": 0.3834491750134341, "grad_norm": 4.849951267242432, "learning_rate": 1.4150846677439486e-05, "loss": 2.8755, "step": 10882 }, { "epoch": 0.3834844120263925, "grad_norm": 1.4520113468170166, "learning_rate": 1.4149808336419762e-05, "loss": 0.8178, "step": 10883 }, { "epoch": 0.38351964903935093, "grad_norm": 1.9550448656082153, "learning_rate": 1.4148769941348254e-05, "loss": 0.8721, "step": 10884 }, { "epoch": 0.38355488605230936, "grad_norm": 1.404312014579773, "learning_rate": 1.414773149223849e-05, "loss": 1.0409, "step": 10885 }, { "epoch": 0.38359012306526774, "grad_norm": 1.5305263996124268, "learning_rate": 1.4146692989103994e-05, "loss": 0.7319, "step": 10886 }, { "epoch": 0.3836253600782262, "grad_norm": 3.7259926795959473, "learning_rate": 1.4145654431958294e-05, "loss": 3.3198, "step": 10887 }, { "epoch": 0.3836605970911846, "grad_norm": 4.913064002990723, "learning_rate": 1.414461582081492e-05, "loss": 3.4772, "step": 10888 }, { "epoch": 0.383695834104143, "grad_norm": 8.325993537902832, "learning_rate": 1.4143577155687392e-05, "loss": 3.7546, "step": 10889 }, { "epoch": 0.3837310711171014, "grad_norm": 5.2271809577941895, "learning_rate": 1.4142538436589246e-05, "loss": 3.0601, "step": 10890 }, { "epoch": 0.3837663081300598, "grad_norm": 6.480741500854492, "learning_rate": 1.414149966353401e-05, "loss": 5.8241, "step": 10891 }, { "epoch": 0.38380154514301823, "grad_norm": 1.858069896697998, "learning_rate": 1.4140460836535212e-05, "loss": 1.0577, "step": 10892 }, { "epoch": 0.38383678215597666, "grad_norm": 6.630439758300781, "learning_rate": 1.4139421955606386e-05, "loss": 3.6362, "step": 10893 }, { "epoch": 0.38387201916893504, "grad_norm": 6.598957538604736, "learning_rate": 1.4138383020761057e-05, "loss": 3.5374, "step": 10894 }, { "epoch": 0.3839072561818935, "grad_norm": 8.036176681518555, "learning_rate": 1.4137344032012768e-05, "loss": 3.2989, "step": 10895 }, { "epoch": 0.38394249319485185, "grad_norm": 4.241695404052734, "learning_rate": 1.4136304989375043e-05, "loss": 2.939, "step": 10896 }, { "epoch": 0.3839777302078103, "grad_norm": 0.9583262205123901, "learning_rate": 1.4135265892861418e-05, "loss": 1.0013, "step": 10897 }, { "epoch": 0.3840129672207687, "grad_norm": 1.5648540258407593, "learning_rate": 1.4134226742485431e-05, "loss": 0.8779, "step": 10898 }, { "epoch": 0.3840482042337271, "grad_norm": 5.707594871520996, "learning_rate": 1.4133187538260612e-05, "loss": 3.4349, "step": 10899 }, { "epoch": 0.38408344124668553, "grad_norm": 7.774113178253174, "learning_rate": 1.4132148280200497e-05, "loss": 3.2799, "step": 10900 }, { "epoch": 0.3841186782596439, "grad_norm": 1.5431594848632812, "learning_rate": 1.4131108968318627e-05, "loss": 0.8827, "step": 10901 }, { "epoch": 0.38415391527260234, "grad_norm": 9.092267036437988, "learning_rate": 1.4130069602628535e-05, "loss": 2.9433, "step": 10902 }, { "epoch": 0.3841891522855608, "grad_norm": 9.532814025878906, "learning_rate": 1.4129030183143761e-05, "loss": 3.2079, "step": 10903 }, { "epoch": 0.38422438929851915, "grad_norm": 6.42088508605957, "learning_rate": 1.4127990709877844e-05, "loss": 4.7259, "step": 10904 }, { "epoch": 0.3842596263114776, "grad_norm": 1.3212875127792358, "learning_rate": 1.412695118284432e-05, "loss": 0.7766, "step": 10905 }, { "epoch": 0.38429486332443596, "grad_norm": 7.186335563659668, "learning_rate": 1.4125911602056733e-05, "loss": 3.1761, "step": 10906 }, { "epoch": 0.3843301003373944, "grad_norm": 1.4043619632720947, "learning_rate": 1.4124871967528622e-05, "loss": 0.8864, "step": 10907 }, { "epoch": 0.38436533735035283, "grad_norm": 1.3468377590179443, "learning_rate": 1.4123832279273523e-05, "loss": 0.7063, "step": 10908 }, { "epoch": 0.3844005743633112, "grad_norm": 5.44528865814209, "learning_rate": 1.4122792537304992e-05, "loss": 3.7863, "step": 10909 }, { "epoch": 0.38443581137626964, "grad_norm": 6.802567005157471, "learning_rate": 1.412175274163656e-05, "loss": 4.0628, "step": 10910 }, { "epoch": 0.384471048389228, "grad_norm": 1.8080428838729858, "learning_rate": 1.4120712892281769e-05, "loss": 0.9181, "step": 10911 }, { "epoch": 0.38450628540218645, "grad_norm": 1.0405240058898926, "learning_rate": 1.4119672989254172e-05, "loss": 0.8247, "step": 10912 }, { "epoch": 0.3845415224151449, "grad_norm": 1.3019461631774902, "learning_rate": 1.4118633032567308e-05, "loss": 1.0055, "step": 10913 }, { "epoch": 0.38457675942810327, "grad_norm": 7.746228218078613, "learning_rate": 1.4117593022234724e-05, "loss": 3.4305, "step": 10914 }, { "epoch": 0.3846119964410617, "grad_norm": 0.9400931596755981, "learning_rate": 1.411655295826997e-05, "loss": 1.0131, "step": 10915 }, { "epoch": 0.3846472334540201, "grad_norm": 1.21517813205719, "learning_rate": 1.4115512840686584e-05, "loss": 0.9441, "step": 10916 }, { "epoch": 0.3846824704669785, "grad_norm": 1.0843002796173096, "learning_rate": 1.411447266949812e-05, "loss": 0.9142, "step": 10917 }, { "epoch": 0.38471770747993694, "grad_norm": 7.932275772094727, "learning_rate": 1.4113432444718127e-05, "loss": 5.0682, "step": 10918 }, { "epoch": 0.3847529444928953, "grad_norm": 13.012091636657715, "learning_rate": 1.4112392166360151e-05, "loss": 6.6809, "step": 10919 }, { "epoch": 0.38478818150585375, "grad_norm": 0.845110297203064, "learning_rate": 1.4111351834437745e-05, "loss": 0.8384, "step": 10920 }, { "epoch": 0.38482341851881213, "grad_norm": 4.60506534576416, "learning_rate": 1.4110311448964455e-05, "loss": 3.0813, "step": 10921 }, { "epoch": 0.38485865553177057, "grad_norm": 5.059467315673828, "learning_rate": 1.4109271009953838e-05, "loss": 3.8166, "step": 10922 }, { "epoch": 0.384893892544729, "grad_norm": 9.807147026062012, "learning_rate": 1.4108230517419441e-05, "loss": 3.1283, "step": 10923 }, { "epoch": 0.3849291295576874, "grad_norm": 2.2039177417755127, "learning_rate": 1.4107189971374817e-05, "loss": 0.9377, "step": 10924 }, { "epoch": 0.3849643665706458, "grad_norm": 11.46884536743164, "learning_rate": 1.4106149371833524e-05, "loss": 4.1745, "step": 10925 }, { "epoch": 0.38499960358360424, "grad_norm": 0.9508758187294006, "learning_rate": 1.410510871880911e-05, "loss": 1.2066, "step": 10926 }, { "epoch": 0.3850348405965626, "grad_norm": 1.9127848148345947, "learning_rate": 1.4104068012315132e-05, "loss": 0.8372, "step": 10927 }, { "epoch": 0.38507007760952106, "grad_norm": 1.7802270650863647, "learning_rate": 1.4103027252365146e-05, "loss": 0.8891, "step": 10928 }, { "epoch": 0.38510531462247943, "grad_norm": 3.125378370285034, "learning_rate": 1.4101986438972707e-05, "loss": 3.4672, "step": 10929 }, { "epoch": 0.38514055163543787, "grad_norm": 3.9633121490478516, "learning_rate": 1.4100945572151373e-05, "loss": 3.5098, "step": 10930 }, { "epoch": 0.3851757886483963, "grad_norm": 15.011320114135742, "learning_rate": 1.40999046519147e-05, "loss": 5.5388, "step": 10931 }, { "epoch": 0.3852110256613547, "grad_norm": 10.55396842956543, "learning_rate": 1.4098863678276247e-05, "loss": 2.71, "step": 10932 }, { "epoch": 0.3852462626743131, "grad_norm": 7.717545509338379, "learning_rate": 1.4097822651249572e-05, "loss": 5.8357, "step": 10933 }, { "epoch": 0.3852814996872715, "grad_norm": 4.5131025314331055, "learning_rate": 1.4096781570848234e-05, "loss": 3.2346, "step": 10934 }, { "epoch": 0.3853167367002299, "grad_norm": 8.589024543762207, "learning_rate": 1.4095740437085795e-05, "loss": 5.7521, "step": 10935 }, { "epoch": 0.38535197371318836, "grad_norm": 1.7076159715652466, "learning_rate": 1.4094699249975816e-05, "loss": 0.8884, "step": 10936 }, { "epoch": 0.38538721072614673, "grad_norm": 7.166591167449951, "learning_rate": 1.4093658009531859e-05, "loss": 5.9202, "step": 10937 }, { "epoch": 0.38542244773910517, "grad_norm": 1.669774055480957, "learning_rate": 1.4092616715767482e-05, "loss": 0.9503, "step": 10938 }, { "epoch": 0.38545768475206355, "grad_norm": 8.32889175415039, "learning_rate": 1.4091575368696251e-05, "loss": 6.7958, "step": 10939 }, { "epoch": 0.385492921765022, "grad_norm": 17.225648880004883, "learning_rate": 1.409053396833173e-05, "loss": 7.918, "step": 10940 }, { "epoch": 0.3855281587779804, "grad_norm": 6.762941360473633, "learning_rate": 1.4089492514687484e-05, "loss": 3.7814, "step": 10941 }, { "epoch": 0.3855633957909388, "grad_norm": 1.7466319799423218, "learning_rate": 1.4088451007777075e-05, "loss": 0.8636, "step": 10942 }, { "epoch": 0.3855986328038972, "grad_norm": 4.594016075134277, "learning_rate": 1.4087409447614073e-05, "loss": 3.3852, "step": 10943 }, { "epoch": 0.3856338698168556, "grad_norm": 1.601427435874939, "learning_rate": 1.4086367834212042e-05, "loss": 1.1189, "step": 10944 }, { "epoch": 0.38566910682981403, "grad_norm": 29.77164077758789, "learning_rate": 1.4085326167584552e-05, "loss": 3.6111, "step": 10945 }, { "epoch": 0.38570434384277247, "grad_norm": 11.072300910949707, "learning_rate": 1.4084284447745163e-05, "loss": 3.6703, "step": 10946 }, { "epoch": 0.38573958085573085, "grad_norm": 5.011778354644775, "learning_rate": 1.4083242674707451e-05, "loss": 2.8032, "step": 10947 }, { "epoch": 0.3857748178686893, "grad_norm": 13.74295425415039, "learning_rate": 1.408220084848498e-05, "loss": 3.2267, "step": 10948 }, { "epoch": 0.38581005488164766, "grad_norm": 6.68250846862793, "learning_rate": 1.408115896909133e-05, "loss": 3.2686, "step": 10949 }, { "epoch": 0.3858452918946061, "grad_norm": 14.408296585083008, "learning_rate": 1.408011703654006e-05, "loss": 3.2289, "step": 10950 }, { "epoch": 0.3858805289075645, "grad_norm": 1.0334057807922363, "learning_rate": 1.4079075050844749e-05, "loss": 1.0331, "step": 10951 }, { "epoch": 0.3859157659205229, "grad_norm": 21.254743576049805, "learning_rate": 1.407803301201896e-05, "loss": 3.9349, "step": 10952 }, { "epoch": 0.38595100293348134, "grad_norm": 1.7961310148239136, "learning_rate": 1.407699092007628e-05, "loss": 1.0323, "step": 10953 }, { "epoch": 0.3859862399464397, "grad_norm": 1.3602895736694336, "learning_rate": 1.4075948775030266e-05, "loss": 0.868, "step": 10954 }, { "epoch": 0.38602147695939815, "grad_norm": 45.59990692138672, "learning_rate": 1.4074906576894506e-05, "loss": 5.8589, "step": 10955 }, { "epoch": 0.3860567139723566, "grad_norm": 1.5990492105484009, "learning_rate": 1.4073864325682565e-05, "loss": 1.1379, "step": 10956 }, { "epoch": 0.38609195098531496, "grad_norm": 7.747764587402344, "learning_rate": 1.4072822021408021e-05, "loss": 3.3901, "step": 10957 }, { "epoch": 0.3861271879982734, "grad_norm": 12.717864036560059, "learning_rate": 1.4071779664084455e-05, "loss": 3.8749, "step": 10958 }, { "epoch": 0.3861624250112318, "grad_norm": 12.056775093078613, "learning_rate": 1.4070737253725436e-05, "loss": 2.8628, "step": 10959 }, { "epoch": 0.3861976620241902, "grad_norm": 1.4653925895690918, "learning_rate": 1.406969479034455e-05, "loss": 1.1591, "step": 10960 }, { "epoch": 0.38623289903714864, "grad_norm": 11.147906303405762, "learning_rate": 1.406865227395537e-05, "loss": 6.0327, "step": 10961 }, { "epoch": 0.386268136050107, "grad_norm": 10.875860214233398, "learning_rate": 1.4067609704571474e-05, "loss": 5.6715, "step": 10962 }, { "epoch": 0.38630337306306545, "grad_norm": 7.96838903427124, "learning_rate": 1.4066567082206443e-05, "loss": 5.2966, "step": 10963 }, { "epoch": 0.3863386100760239, "grad_norm": 4.759838104248047, "learning_rate": 1.4065524406873862e-05, "loss": 3.648, "step": 10964 }, { "epoch": 0.38637384708898226, "grad_norm": 4.768996715545654, "learning_rate": 1.4064481678587303e-05, "loss": 2.9465, "step": 10965 }, { "epoch": 0.3864090841019407, "grad_norm": 8.981368064880371, "learning_rate": 1.4063438897360357e-05, "loss": 4.5536, "step": 10966 }, { "epoch": 0.38644432111489907, "grad_norm": 1.6860376596450806, "learning_rate": 1.4062396063206598e-05, "loss": 0.7213, "step": 10967 }, { "epoch": 0.3864795581278575, "grad_norm": 1.9515362977981567, "learning_rate": 1.4061353176139616e-05, "loss": 0.899, "step": 10968 }, { "epoch": 0.38651479514081594, "grad_norm": 1.8710957765579224, "learning_rate": 1.406031023617299e-05, "loss": 0.8402, "step": 10969 }, { "epoch": 0.3865500321537743, "grad_norm": 1.538306474685669, "learning_rate": 1.4059267243320305e-05, "loss": 0.9612, "step": 10970 }, { "epoch": 0.38658526916673275, "grad_norm": 1.6151150465011597, "learning_rate": 1.4058224197595147e-05, "loss": 0.8262, "step": 10971 }, { "epoch": 0.3866205061796911, "grad_norm": 14.883026123046875, "learning_rate": 1.4057181099011105e-05, "loss": 3.3108, "step": 10972 }, { "epoch": 0.38665574319264956, "grad_norm": 6.232085704803467, "learning_rate": 1.4056137947581759e-05, "loss": 3.312, "step": 10973 }, { "epoch": 0.386690980205608, "grad_norm": 9.936317443847656, "learning_rate": 1.40550947433207e-05, "loss": 3.8808, "step": 10974 }, { "epoch": 0.38672621721856637, "grad_norm": 1.1303151845932007, "learning_rate": 1.4054051486241517e-05, "loss": 0.8128, "step": 10975 }, { "epoch": 0.3867614542315248, "grad_norm": 1.694640874862671, "learning_rate": 1.4053008176357797e-05, "loss": 0.8947, "step": 10976 }, { "epoch": 0.3867966912444832, "grad_norm": 11.394364356994629, "learning_rate": 1.4051964813683127e-05, "loss": 3.7447, "step": 10977 }, { "epoch": 0.3868319282574416, "grad_norm": 19.312925338745117, "learning_rate": 1.40509213982311e-05, "loss": 3.8556, "step": 10978 }, { "epoch": 0.38686716527040005, "grad_norm": 1.2778970003128052, "learning_rate": 1.4049877930015306e-05, "loss": 1.0666, "step": 10979 }, { "epoch": 0.3869024022833584, "grad_norm": 15.11251163482666, "learning_rate": 1.4048834409049337e-05, "loss": 3.8016, "step": 10980 }, { "epoch": 0.38693763929631686, "grad_norm": 8.504637718200684, "learning_rate": 1.4047790835346783e-05, "loss": 3.4884, "step": 10981 }, { "epoch": 0.38697287630927524, "grad_norm": 1.3244904279708862, "learning_rate": 1.4046747208921238e-05, "loss": 0.9177, "step": 10982 }, { "epoch": 0.38700811332223367, "grad_norm": 1.9525094032287598, "learning_rate": 1.4045703529786296e-05, "loss": 1.0211, "step": 10983 }, { "epoch": 0.3870433503351921, "grad_norm": 1.4979970455169678, "learning_rate": 1.4044659797955549e-05, "loss": 0.8162, "step": 10984 }, { "epoch": 0.3870785873481505, "grad_norm": 6.126298904418945, "learning_rate": 1.4043616013442593e-05, "loss": 3.7966, "step": 10985 }, { "epoch": 0.3871138243611089, "grad_norm": 9.194466590881348, "learning_rate": 1.4042572176261022e-05, "loss": 3.6189, "step": 10986 }, { "epoch": 0.3871490613740673, "grad_norm": 3.142453193664551, "learning_rate": 1.4041528286424438e-05, "loss": 3.7924, "step": 10987 }, { "epoch": 0.3871842983870257, "grad_norm": 1.0487912893295288, "learning_rate": 1.404048434394643e-05, "loss": 0.8846, "step": 10988 }, { "epoch": 0.38721953539998416, "grad_norm": 1.4727418422698975, "learning_rate": 1.4039440348840598e-05, "loss": 1.1933, "step": 10989 }, { "epoch": 0.38725477241294254, "grad_norm": 1.4727418422698975, "learning_rate": 1.4039440348840598e-05, "loss": 5.7833, "step": 10990 }, { "epoch": 0.38729000942590097, "grad_norm": 1.2032527923583984, "learning_rate": 1.4038396301120543e-05, "loss": 0.8613, "step": 10991 }, { "epoch": 0.3873252464388594, "grad_norm": 8.707468032836914, "learning_rate": 1.4037352200799862e-05, "loss": 5.9848, "step": 10992 }, { "epoch": 0.3873604834518178, "grad_norm": 11.620285034179688, "learning_rate": 1.403630804789215e-05, "loss": 5.4507, "step": 10993 }, { "epoch": 0.3873957204647762, "grad_norm": 5.388837814331055, "learning_rate": 1.4035263842411019e-05, "loss": 3.3921, "step": 10994 }, { "epoch": 0.3874309574777346, "grad_norm": 1.5504823923110962, "learning_rate": 1.4034219584370055e-05, "loss": 0.867, "step": 10995 }, { "epoch": 0.38746619449069303, "grad_norm": 29.02640724182129, "learning_rate": 1.4033175273782873e-05, "loss": 2.9727, "step": 10996 }, { "epoch": 0.38750143150365146, "grad_norm": 6.920773029327393, "learning_rate": 1.4032130910663066e-05, "loss": 3.5051, "step": 10997 }, { "epoch": 0.38753666851660984, "grad_norm": 1.210218906402588, "learning_rate": 1.403108649502424e-05, "loss": 0.9402, "step": 10998 }, { "epoch": 0.3875719055295683, "grad_norm": 1.9160398244857788, "learning_rate": 1.403004202688e-05, "loss": 0.8241, "step": 10999 }, { "epoch": 0.38760714254252665, "grad_norm": 2.0127742290496826, "learning_rate": 1.4028997506243948e-05, "loss": 0.6915, "step": 11000 }, { "epoch": 0.3876423795554851, "grad_norm": 16.987655639648438, "learning_rate": 1.4027952933129691e-05, "loss": 3.7982, "step": 11001 }, { "epoch": 0.3876776165684435, "grad_norm": 1.3880109786987305, "learning_rate": 1.4026908307550837e-05, "loss": 0.8537, "step": 11002 }, { "epoch": 0.3877128535814019, "grad_norm": 1.4099640846252441, "learning_rate": 1.4025863629520986e-05, "loss": 0.9249, "step": 11003 }, { "epoch": 0.38774809059436033, "grad_norm": 13.252307891845703, "learning_rate": 1.402481889905375e-05, "loss": 5.3007, "step": 11004 }, { "epoch": 0.3877833276073187, "grad_norm": 1.36283540725708, "learning_rate": 1.4023774116162732e-05, "loss": 0.932, "step": 11005 }, { "epoch": 0.38781856462027714, "grad_norm": 32.80649185180664, "learning_rate": 1.4022729280861546e-05, "loss": 2.8883, "step": 11006 }, { "epoch": 0.3878538016332356, "grad_norm": 1.4544686079025269, "learning_rate": 1.40216843931638e-05, "loss": 0.7823, "step": 11007 }, { "epoch": 0.38788903864619395, "grad_norm": 6.340048789978027, "learning_rate": 1.40206394530831e-05, "loss": 3.0056, "step": 11008 }, { "epoch": 0.3879242756591524, "grad_norm": 1.6010299921035767, "learning_rate": 1.401959446063306e-05, "loss": 0.9639, "step": 11009 }, { "epoch": 0.38795951267211076, "grad_norm": 22.950489044189453, "learning_rate": 1.401854941582729e-05, "loss": 3.0755, "step": 11010 }, { "epoch": 0.3879947496850692, "grad_norm": 1.74972665309906, "learning_rate": 1.4017504318679402e-05, "loss": 0.6842, "step": 11011 }, { "epoch": 0.38802998669802763, "grad_norm": 1.749386191368103, "learning_rate": 1.4016459169203007e-05, "loss": 1.0485, "step": 11012 }, { "epoch": 0.388065223710986, "grad_norm": 1.7519053220748901, "learning_rate": 1.4015413967411723e-05, "loss": 0.9158, "step": 11013 }, { "epoch": 0.38810046072394444, "grad_norm": 1.1882199048995972, "learning_rate": 1.4014368713319157e-05, "loss": 0.9338, "step": 11014 }, { "epoch": 0.3881356977369028, "grad_norm": 6.1702165603637695, "learning_rate": 1.401332340693893e-05, "loss": 3.166, "step": 11015 }, { "epoch": 0.38817093474986125, "grad_norm": 14.907767295837402, "learning_rate": 1.4012278048284654e-05, "loss": 7.7248, "step": 11016 }, { "epoch": 0.3882061717628197, "grad_norm": 4.1070380210876465, "learning_rate": 1.4011232637369945e-05, "loss": 3.052, "step": 11017 }, { "epoch": 0.38824140877577806, "grad_norm": 4.51221227645874, "learning_rate": 1.4010187174208419e-05, "loss": 4.9818, "step": 11018 }, { "epoch": 0.3882766457887365, "grad_norm": 5.211237907409668, "learning_rate": 1.4009141658813697e-05, "loss": 3.2062, "step": 11019 }, { "epoch": 0.3883118828016949, "grad_norm": 7.519505023956299, "learning_rate": 1.400809609119939e-05, "loss": 3.2344, "step": 11020 }, { "epoch": 0.3883471198146533, "grad_norm": 1.754340410232544, "learning_rate": 1.4007050471379124e-05, "loss": 0.9965, "step": 11021 }, { "epoch": 0.38838235682761174, "grad_norm": 1.5676782131195068, "learning_rate": 1.4006004799366512e-05, "loss": 0.9853, "step": 11022 }, { "epoch": 0.3884175938405701, "grad_norm": 1.543990135192871, "learning_rate": 1.4004959075175182e-05, "loss": 0.8833, "step": 11023 }, { "epoch": 0.38845283085352855, "grad_norm": 7.401891231536865, "learning_rate": 1.4003913298818748e-05, "loss": 3.0475, "step": 11024 }, { "epoch": 0.38848806786648693, "grad_norm": 25.23932647705078, "learning_rate": 1.400286747031083e-05, "loss": 6.0949, "step": 11025 }, { "epoch": 0.38852330487944536, "grad_norm": 13.299479484558105, "learning_rate": 1.4001821589665057e-05, "loss": 5.2062, "step": 11026 }, { "epoch": 0.3885585418924038, "grad_norm": 7.152286529541016, "learning_rate": 1.4000775656895047e-05, "loss": 3.4571, "step": 11027 }, { "epoch": 0.3885937789053622, "grad_norm": 1.4933840036392212, "learning_rate": 1.3999729672014424e-05, "loss": 0.8565, "step": 11028 }, { "epoch": 0.3886290159183206, "grad_norm": 34.56058120727539, "learning_rate": 1.3998683635036812e-05, "loss": 3.5563, "step": 11029 }, { "epoch": 0.38866425293127904, "grad_norm": 4.242497444152832, "learning_rate": 1.399763754597584e-05, "loss": 3.4237, "step": 11030 }, { "epoch": 0.3886994899442374, "grad_norm": 1.279405951499939, "learning_rate": 1.3996591404845124e-05, "loss": 0.9327, "step": 11031 }, { "epoch": 0.38873472695719585, "grad_norm": 5.77582311630249, "learning_rate": 1.3995545211658302e-05, "loss": 2.4775, "step": 11032 }, { "epoch": 0.38876996397015423, "grad_norm": 2.0660619735717773, "learning_rate": 1.399449896642899e-05, "loss": 1.0304, "step": 11033 }, { "epoch": 0.38880520098311266, "grad_norm": 11.756115913391113, "learning_rate": 1.3993452669170822e-05, "loss": 3.0316, "step": 11034 }, { "epoch": 0.3888404379960711, "grad_norm": 22.050033569335938, "learning_rate": 1.3992406319897422e-05, "loss": 8.1383, "step": 11035 }, { "epoch": 0.3888756750090295, "grad_norm": 25.304662704467773, "learning_rate": 1.3991359918622424e-05, "loss": 6.3027, "step": 11036 }, { "epoch": 0.3889109120219879, "grad_norm": 1.5741345882415771, "learning_rate": 1.3990313465359453e-05, "loss": 0.8636, "step": 11037 }, { "epoch": 0.3889461490349463, "grad_norm": 5.42927360534668, "learning_rate": 1.3989266960122143e-05, "loss": 2.8056, "step": 11038 }, { "epoch": 0.3889813860479047, "grad_norm": 8.294632911682129, "learning_rate": 1.398822040292412e-05, "loss": 3.2569, "step": 11039 }, { "epoch": 0.38901662306086315, "grad_norm": 0.9938804507255554, "learning_rate": 1.398717379377902e-05, "loss": 0.8544, "step": 11040 }, { "epoch": 0.38905186007382153, "grad_norm": 1.8338618278503418, "learning_rate": 1.3986127132700471e-05, "loss": 0.9633, "step": 11041 }, { "epoch": 0.38908709708677996, "grad_norm": 1.160952091217041, "learning_rate": 1.3985080419702113e-05, "loss": 0.8246, "step": 11042 }, { "epoch": 0.38912233409973834, "grad_norm": 15.790035247802734, "learning_rate": 1.3984033654797574e-05, "loss": 3.2928, "step": 11043 }, { "epoch": 0.3891575711126968, "grad_norm": 12.929266929626465, "learning_rate": 1.3982986838000486e-05, "loss": 4.121, "step": 11044 }, { "epoch": 0.3891928081256552, "grad_norm": 5.02783203125, "learning_rate": 1.3981939969324487e-05, "loss": 2.925, "step": 11045 }, { "epoch": 0.3892280451386136, "grad_norm": 12.328917503356934, "learning_rate": 1.3980893048783215e-05, "loss": 10.7241, "step": 11046 }, { "epoch": 0.389263282151572, "grad_norm": 7.421319961547852, "learning_rate": 1.3979846076390302e-05, "loss": 3.4712, "step": 11047 }, { "epoch": 0.3892985191645304, "grad_norm": 3.7751572132110596, "learning_rate": 1.3978799052159389e-05, "loss": 3.5101, "step": 11048 }, { "epoch": 0.38933375617748883, "grad_norm": 9.273486137390137, "learning_rate": 1.397775197610411e-05, "loss": 3.5626, "step": 11049 }, { "epoch": 0.38936899319044727, "grad_norm": 1.1634576320648193, "learning_rate": 1.3976704848238103e-05, "loss": 1.416, "step": 11050 }, { "epoch": 0.38940423020340564, "grad_norm": 4.3265275955200195, "learning_rate": 1.3975657668575013e-05, "loss": 0.9024, "step": 11051 }, { "epoch": 0.3894394672163641, "grad_norm": 1.4746705293655396, "learning_rate": 1.3974610437128473e-05, "loss": 1.0575, "step": 11052 }, { "epoch": 0.38947470422932245, "grad_norm": 12.280131340026855, "learning_rate": 1.3973563153912126e-05, "loss": 3.5929, "step": 11053 }, { "epoch": 0.3895099412422809, "grad_norm": 1.530919075012207, "learning_rate": 1.3972515818939614e-05, "loss": 0.9943, "step": 11054 }, { "epoch": 0.3895451782552393, "grad_norm": 1.6823970079421997, "learning_rate": 1.3971468432224575e-05, "loss": 0.807, "step": 11055 }, { "epoch": 0.3895804152681977, "grad_norm": 5.387557029724121, "learning_rate": 1.3970420993780657e-05, "loss": 3.797, "step": 11056 }, { "epoch": 0.38961565228115613, "grad_norm": 5.586008548736572, "learning_rate": 1.3969373503621499e-05, "loss": 3.1645, "step": 11057 }, { "epoch": 0.3896508892941145, "grad_norm": 1.1666014194488525, "learning_rate": 1.3968325961760743e-05, "loss": 1.0671, "step": 11058 }, { "epoch": 0.38968612630707294, "grad_norm": 11.150720596313477, "learning_rate": 1.3967278368212038e-05, "loss": 5.4855, "step": 11059 }, { "epoch": 0.3897213633200314, "grad_norm": 3.6028425693511963, "learning_rate": 1.3966230722989025e-05, "loss": 2.8647, "step": 11060 }, { "epoch": 0.38975660033298976, "grad_norm": 1.8535552024841309, "learning_rate": 1.3965183026105355e-05, "loss": 0.7799, "step": 11061 }, { "epoch": 0.3897918373459482, "grad_norm": 7.243954181671143, "learning_rate": 1.3964135277574668e-05, "loss": 6.0722, "step": 11062 }, { "epoch": 0.3898270743589066, "grad_norm": 1.3420406579971313, "learning_rate": 1.3963087477410615e-05, "loss": 0.8686, "step": 11063 }, { "epoch": 0.389862311371865, "grad_norm": 1.4024194478988647, "learning_rate": 1.3962039625626844e-05, "loss": 1.0863, "step": 11064 }, { "epoch": 0.38989754838482343, "grad_norm": 1.16423499584198, "learning_rate": 1.3960991722237e-05, "loss": 0.9383, "step": 11065 }, { "epoch": 0.3899327853977818, "grad_norm": 15.440693855285645, "learning_rate": 1.3959943767254735e-05, "loss": 3.1905, "step": 11066 }, { "epoch": 0.38996802241074024, "grad_norm": 1.3075889348983765, "learning_rate": 1.3958895760693697e-05, "loss": 0.9965, "step": 11067 }, { "epoch": 0.3900032594236987, "grad_norm": 2.375459671020508, "learning_rate": 1.395784770256754e-05, "loss": 0.925, "step": 11068 }, { "epoch": 0.39003849643665706, "grad_norm": 10.544426918029785, "learning_rate": 1.395679959288991e-05, "loss": 2.9566, "step": 11069 }, { "epoch": 0.3900737334496155, "grad_norm": 6.422154903411865, "learning_rate": 1.3955751431674465e-05, "loss": 3.2171, "step": 11070 }, { "epoch": 0.39010897046257387, "grad_norm": 2.185889482498169, "learning_rate": 1.3954703218934849e-05, "loss": 0.9403, "step": 11071 }, { "epoch": 0.3901442074755323, "grad_norm": 7.252897262573242, "learning_rate": 1.3953654954684725e-05, "loss": 6.5883, "step": 11072 }, { "epoch": 0.39017944448849073, "grad_norm": 1.1011366844177246, "learning_rate": 1.3952606638937737e-05, "loss": 0.9381, "step": 11073 }, { "epoch": 0.3902146815014491, "grad_norm": 1.0892179012298584, "learning_rate": 1.3951558271707546e-05, "loss": 0.7788, "step": 11074 }, { "epoch": 0.39024991851440755, "grad_norm": 9.749381065368652, "learning_rate": 1.3950509853007805e-05, "loss": 3.3004, "step": 11075 }, { "epoch": 0.3902851555273659, "grad_norm": 6.390596389770508, "learning_rate": 1.3949461382852173e-05, "loss": 3.3153, "step": 11076 }, { "epoch": 0.39032039254032436, "grad_norm": 5.39945125579834, "learning_rate": 1.3948412861254296e-05, "loss": 3.1796, "step": 11077 }, { "epoch": 0.3903556295532828, "grad_norm": 12.471121788024902, "learning_rate": 1.3947364288227846e-05, "loss": 3.4157, "step": 11078 }, { "epoch": 0.39039086656624117, "grad_norm": 6.099557876586914, "learning_rate": 1.3946315663786468e-05, "loss": 2.8765, "step": 11079 }, { "epoch": 0.3904261035791996, "grad_norm": 6.8615946769714355, "learning_rate": 1.394526698794383e-05, "loss": 3.0878, "step": 11080 }, { "epoch": 0.390461340592158, "grad_norm": 8.444177627563477, "learning_rate": 1.3944218260713584e-05, "loss": 3.3308, "step": 11081 }, { "epoch": 0.3904965776051164, "grad_norm": 1.3149595260620117, "learning_rate": 1.3943169482109392e-05, "loss": 0.8142, "step": 11082 }, { "epoch": 0.39053181461807485, "grad_norm": 2.2840311527252197, "learning_rate": 1.394212065214492e-05, "loss": 0.8712, "step": 11083 }, { "epoch": 0.3905670516310332, "grad_norm": 1.1738511323928833, "learning_rate": 1.3941071770833818e-05, "loss": 1.1673, "step": 11084 }, { "epoch": 0.39060228864399166, "grad_norm": 17.620742797851562, "learning_rate": 1.394002283818976e-05, "loss": 3.1697, "step": 11085 }, { "epoch": 0.39063752565695004, "grad_norm": 0.8742382526397705, "learning_rate": 1.39389738542264e-05, "loss": 0.9996, "step": 11086 }, { "epoch": 0.39067276266990847, "grad_norm": 15.656391143798828, "learning_rate": 1.3937924818957403e-05, "loss": 6.4761, "step": 11087 }, { "epoch": 0.3907079996828669, "grad_norm": 1.208943486213684, "learning_rate": 1.3936875732396433e-05, "loss": 0.8333, "step": 11088 }, { "epoch": 0.3907432366958253, "grad_norm": 1.1099283695220947, "learning_rate": 1.3935826594557158e-05, "loss": 0.9439, "step": 11089 }, { "epoch": 0.3907784737087837, "grad_norm": 6.254047870635986, "learning_rate": 1.3934777405453236e-05, "loss": 3.0066, "step": 11090 }, { "epoch": 0.3908137107217421, "grad_norm": 1.8575782775878906, "learning_rate": 1.3933728165098343e-05, "loss": 0.9491, "step": 11091 }, { "epoch": 0.3908489477347005, "grad_norm": 1.716057538986206, "learning_rate": 1.3932678873506136e-05, "loss": 0.9202, "step": 11092 }, { "epoch": 0.39088418474765896, "grad_norm": 13.544965744018555, "learning_rate": 1.3931629530690287e-05, "loss": 3.819, "step": 11093 }, { "epoch": 0.39091942176061734, "grad_norm": 2.8796448707580566, "learning_rate": 1.3930580136664462e-05, "loss": 2.7166, "step": 11094 }, { "epoch": 0.39095465877357577, "grad_norm": 9.141377449035645, "learning_rate": 1.3929530691442327e-05, "loss": 3.6948, "step": 11095 }, { "epoch": 0.3909898957865342, "grad_norm": 1.4803358316421509, "learning_rate": 1.3928481195037558e-05, "loss": 0.7499, "step": 11096 }, { "epoch": 0.3910251327994926, "grad_norm": 5.332186698913574, "learning_rate": 1.3927431647463823e-05, "loss": 3.2361, "step": 11097 }, { "epoch": 0.391060369812451, "grad_norm": 8.943921089172363, "learning_rate": 1.3926382048734786e-05, "loss": 2.9872, "step": 11098 }, { "epoch": 0.3910956068254094, "grad_norm": 5.657993316650391, "learning_rate": 1.3925332398864125e-05, "loss": 2.7571, "step": 11099 }, { "epoch": 0.3911308438383678, "grad_norm": 9.386100769042969, "learning_rate": 1.3924282697865508e-05, "loss": 3.2416, "step": 11100 }, { "epoch": 0.39116608085132626, "grad_norm": 22.809675216674805, "learning_rate": 1.392323294575261e-05, "loss": 6.2018, "step": 11101 }, { "epoch": 0.39120131786428464, "grad_norm": 7.241990089416504, "learning_rate": 1.3922183142539103e-05, "loss": 3.2351, "step": 11102 }, { "epoch": 0.39123655487724307, "grad_norm": 13.325286865234375, "learning_rate": 1.392113328823866e-05, "loss": 6.1139, "step": 11103 }, { "epoch": 0.39127179189020145, "grad_norm": 1.5078402757644653, "learning_rate": 1.3920083382864958e-05, "loss": 1.015, "step": 11104 }, { "epoch": 0.3913070289031599, "grad_norm": 33.724952697753906, "learning_rate": 1.391903342643167e-05, "loss": 3.8166, "step": 11105 }, { "epoch": 0.3913422659161183, "grad_norm": 12.48636531829834, "learning_rate": 1.391798341895247e-05, "loss": 3.5432, "step": 11106 }, { "epoch": 0.3913775029290767, "grad_norm": 29.169572830200195, "learning_rate": 1.3916933360441038e-05, "loss": 3.5706, "step": 11107 }, { "epoch": 0.3914127399420351, "grad_norm": 1.8648581504821777, "learning_rate": 1.3915883250911056e-05, "loss": 0.7676, "step": 11108 }, { "epoch": 0.3914479769549935, "grad_norm": 10.003861427307129, "learning_rate": 1.3914833090376187e-05, "loss": 3.5998, "step": 11109 }, { "epoch": 0.39148321396795194, "grad_norm": 1.0110764503479004, "learning_rate": 1.3913782878850123e-05, "loss": 0.9278, "step": 11110 }, { "epoch": 0.39151845098091037, "grad_norm": 4.325700759887695, "learning_rate": 1.391273261634654e-05, "loss": 3.7124, "step": 11111 }, { "epoch": 0.39155368799386875, "grad_norm": 6.203668117523193, "learning_rate": 1.3911682302879112e-05, "loss": 4.9906, "step": 11112 }, { "epoch": 0.3915889250068272, "grad_norm": 2.4969308376312256, "learning_rate": 1.3910631938461528e-05, "loss": 0.7574, "step": 11113 }, { "epoch": 0.39162416201978556, "grad_norm": 1.7136744260787964, "learning_rate": 1.3909581523107463e-05, "loss": 1.0287, "step": 11114 }, { "epoch": 0.391659399032744, "grad_norm": 9.512673377990723, "learning_rate": 1.3908531056830599e-05, "loss": 3.581, "step": 11115 }, { "epoch": 0.3916946360457024, "grad_norm": 23.528629302978516, "learning_rate": 1.3907480539644622e-05, "loss": 3.6933, "step": 11116 }, { "epoch": 0.3917298730586608, "grad_norm": 7.666820526123047, "learning_rate": 1.3906429971563212e-05, "loss": 5.5969, "step": 11117 }, { "epoch": 0.39176511007161924, "grad_norm": 1.6062257289886475, "learning_rate": 1.3905379352600056e-05, "loss": 0.8365, "step": 11118 }, { "epoch": 0.3918003470845776, "grad_norm": 5.437396049499512, "learning_rate": 1.3904328682768835e-05, "loss": 3.3122, "step": 11119 }, { "epoch": 0.39183558409753605, "grad_norm": 16.495140075683594, "learning_rate": 1.3903277962083235e-05, "loss": 5.4542, "step": 11120 }, { "epoch": 0.3918708211104945, "grad_norm": 1.5170388221740723, "learning_rate": 1.3902227190556942e-05, "loss": 1.2834, "step": 11121 }, { "epoch": 0.39190605812345286, "grad_norm": 2.0293588638305664, "learning_rate": 1.3901176368203646e-05, "loss": 1.0977, "step": 11122 }, { "epoch": 0.3919412951364113, "grad_norm": 40.01292419433594, "learning_rate": 1.3900125495037026e-05, "loss": 5.192, "step": 11123 }, { "epoch": 0.39197653214936967, "grad_norm": 8.633697509765625, "learning_rate": 1.3899074571070778e-05, "loss": 3.5908, "step": 11124 }, { "epoch": 0.3920117691623281, "grad_norm": 1.2973648309707642, "learning_rate": 1.3898023596318589e-05, "loss": 1.1556, "step": 11125 }, { "epoch": 0.39204700617528654, "grad_norm": 7.837264537811279, "learning_rate": 1.3896972570794142e-05, "loss": 2.9438, "step": 11126 }, { "epoch": 0.3920822431882449, "grad_norm": 17.85399055480957, "learning_rate": 1.3895921494511132e-05, "loss": 3.7498, "step": 11127 }, { "epoch": 0.39211748020120335, "grad_norm": 4.774271011352539, "learning_rate": 1.3894870367483248e-05, "loss": 3.1744, "step": 11128 }, { "epoch": 0.39215271721416173, "grad_norm": 1.1814841032028198, "learning_rate": 1.3893819189724185e-05, "loss": 1.0236, "step": 11129 }, { "epoch": 0.39218795422712016, "grad_norm": 3.873142957687378, "learning_rate": 1.3892767961247629e-05, "loss": 3.1059, "step": 11130 }, { "epoch": 0.3922231912400786, "grad_norm": 1.006937026977539, "learning_rate": 1.3891716682067272e-05, "loss": 0.9075, "step": 11131 }, { "epoch": 0.392258428253037, "grad_norm": 18.348953247070312, "learning_rate": 1.3890665352196811e-05, "loss": 6.5189, "step": 11132 }, { "epoch": 0.3922936652659954, "grad_norm": 1.3192577362060547, "learning_rate": 1.3889613971649938e-05, "loss": 1.0162, "step": 11133 }, { "epoch": 0.39232890227895384, "grad_norm": 16.997224807739258, "learning_rate": 1.3888562540440348e-05, "loss": 6.6693, "step": 11134 }, { "epoch": 0.3923641392919122, "grad_norm": 7.518590450286865, "learning_rate": 1.3887511058581738e-05, "loss": 3.1357, "step": 11135 }, { "epoch": 0.39239937630487065, "grad_norm": 30.257551193237305, "learning_rate": 1.3886459526087794e-05, "loss": 3.2977, "step": 11136 }, { "epoch": 0.39243461331782903, "grad_norm": 4.960840702056885, "learning_rate": 1.3885407942972227e-05, "loss": 3.3412, "step": 11137 }, { "epoch": 0.39246985033078746, "grad_norm": 1.8320645093917847, "learning_rate": 1.3884356309248724e-05, "loss": 0.9098, "step": 11138 }, { "epoch": 0.3925050873437459, "grad_norm": 1.0060107707977295, "learning_rate": 1.3883304624930984e-05, "loss": 0.6535, "step": 11139 }, { "epoch": 0.3925403243567043, "grad_norm": 8.785036087036133, "learning_rate": 1.388225289003271e-05, "loss": 3.5719, "step": 11140 }, { "epoch": 0.3925755613696627, "grad_norm": 1.9179798364639282, "learning_rate": 1.3881201104567594e-05, "loss": 0.8629, "step": 11141 }, { "epoch": 0.3926107983826211, "grad_norm": 8.816266059875488, "learning_rate": 1.3880149268549341e-05, "loss": 5.702, "step": 11142 }, { "epoch": 0.3926460353955795, "grad_norm": 9.963399887084961, "learning_rate": 1.3879097381991648e-05, "loss": 6.1075, "step": 11143 }, { "epoch": 0.39268127240853795, "grad_norm": 4.1826372146606445, "learning_rate": 1.387804544490822e-05, "loss": 3.0, "step": 11144 }, { "epoch": 0.39271650942149633, "grad_norm": 8.519938468933105, "learning_rate": 1.3876993457312753e-05, "loss": 5.7291, "step": 11145 }, { "epoch": 0.39275174643445476, "grad_norm": 1.2058871984481812, "learning_rate": 1.3875941419218956e-05, "loss": 0.8664, "step": 11146 }, { "epoch": 0.39278698344741314, "grad_norm": 1.409340739250183, "learning_rate": 1.3874889330640524e-05, "loss": 1.1118, "step": 11147 }, { "epoch": 0.3928222204603716, "grad_norm": 1.5009645223617554, "learning_rate": 1.387383719159117e-05, "loss": 0.7777, "step": 11148 }, { "epoch": 0.39285745747333, "grad_norm": 9.562358856201172, "learning_rate": 1.387278500208459e-05, "loss": 5.5145, "step": 11149 }, { "epoch": 0.3928926944862884, "grad_norm": 3.050326108932495, "learning_rate": 1.3871732762134493e-05, "loss": 0.693, "step": 11150 }, { "epoch": 0.3929279314992468, "grad_norm": 1.1005934476852417, "learning_rate": 1.3870680471754583e-05, "loss": 0.9118, "step": 11151 }, { "epoch": 0.3929631685122052, "grad_norm": 26.108829498291016, "learning_rate": 1.3869628130958566e-05, "loss": 3.5793, "step": 11152 }, { "epoch": 0.39299840552516363, "grad_norm": 1.4770662784576416, "learning_rate": 1.3868575739760151e-05, "loss": 0.9128, "step": 11153 }, { "epoch": 0.39303364253812206, "grad_norm": 15.687095642089844, "learning_rate": 1.3867523298173045e-05, "loss": 3.3486, "step": 11154 }, { "epoch": 0.39306887955108044, "grad_norm": 4.073834419250488, "learning_rate": 1.3866470806210951e-05, "loss": 2.9041, "step": 11155 }, { "epoch": 0.3931041165640389, "grad_norm": 8.22252082824707, "learning_rate": 1.3865418263887587e-05, "loss": 3.6508, "step": 11156 }, { "epoch": 0.39313935357699725, "grad_norm": 10.119338035583496, "learning_rate": 1.3864365671216658e-05, "loss": 3.5157, "step": 11157 }, { "epoch": 0.3931745905899557, "grad_norm": 1.5825756788253784, "learning_rate": 1.3863313028211871e-05, "loss": 1.0298, "step": 11158 }, { "epoch": 0.3932098276029141, "grad_norm": 14.982186317443848, "learning_rate": 1.3862260334886939e-05, "loss": 6.5286, "step": 11159 }, { "epoch": 0.3932450646158725, "grad_norm": 7.768136024475098, "learning_rate": 1.3861207591255577e-05, "loss": 2.8554, "step": 11160 }, { "epoch": 0.39328030162883093, "grad_norm": 9.145994186401367, "learning_rate": 1.3860154797331492e-05, "loss": 3.4071, "step": 11161 }, { "epoch": 0.3933155386417893, "grad_norm": 1.2853658199310303, "learning_rate": 1.3859101953128402e-05, "loss": 0.7535, "step": 11162 }, { "epoch": 0.39335077565474774, "grad_norm": 1.3482433557510376, "learning_rate": 1.3858049058660014e-05, "loss": 0.9133, "step": 11163 }, { "epoch": 0.3933860126677062, "grad_norm": 3.761756181716919, "learning_rate": 1.3856996113940048e-05, "loss": 3.0933, "step": 11164 }, { "epoch": 0.39342124968066455, "grad_norm": 5.4123663902282715, "learning_rate": 1.3855943118982218e-05, "loss": 4.0611, "step": 11165 }, { "epoch": 0.393456486693623, "grad_norm": 1.3649518489837646, "learning_rate": 1.3854890073800234e-05, "loss": 0.9021, "step": 11166 }, { "epoch": 0.3934917237065814, "grad_norm": 1.340540885925293, "learning_rate": 1.3853836978407818e-05, "loss": 1.0685, "step": 11167 }, { "epoch": 0.3935269607195398, "grad_norm": 1.506552815437317, "learning_rate": 1.3852783832818681e-05, "loss": 0.9591, "step": 11168 }, { "epoch": 0.39356219773249823, "grad_norm": 2.536301612854004, "learning_rate": 1.3851730637046548e-05, "loss": 1.2328, "step": 11169 }, { "epoch": 0.3935974347454566, "grad_norm": 1.2304702997207642, "learning_rate": 1.3850677391105132e-05, "loss": 0.8741, "step": 11170 }, { "epoch": 0.39363267175841504, "grad_norm": 1.9137675762176514, "learning_rate": 1.384962409500815e-05, "loss": 0.7595, "step": 11171 }, { "epoch": 0.3936679087713735, "grad_norm": 1.2522355318069458, "learning_rate": 1.3848570748769325e-05, "loss": 0.7712, "step": 11172 }, { "epoch": 0.39370314578433185, "grad_norm": 1.7911510467529297, "learning_rate": 1.3847517352402378e-05, "loss": 0.996, "step": 11173 }, { "epoch": 0.3937383827972903, "grad_norm": 1.3433301448822021, "learning_rate": 1.3846463905921023e-05, "loss": 0.895, "step": 11174 }, { "epoch": 0.39377361981024867, "grad_norm": 8.415468215942383, "learning_rate": 1.384541040933899e-05, "loss": 3.5438, "step": 11175 }, { "epoch": 0.3938088568232071, "grad_norm": 2.246063709259033, "learning_rate": 1.3844356862669996e-05, "loss": 0.8348, "step": 11176 }, { "epoch": 0.39384409383616553, "grad_norm": 19.58308982849121, "learning_rate": 1.3843303265927764e-05, "loss": 5.6882, "step": 11177 }, { "epoch": 0.3938793308491239, "grad_norm": 1.8739875555038452, "learning_rate": 1.3842249619126014e-05, "loss": 0.9678, "step": 11178 }, { "epoch": 0.39391456786208234, "grad_norm": 1.080605149269104, "learning_rate": 1.3841195922278476e-05, "loss": 1.0768, "step": 11179 }, { "epoch": 0.3939498048750407, "grad_norm": 10.069085121154785, "learning_rate": 1.3840142175398871e-05, "loss": 5.8907, "step": 11180 }, { "epoch": 0.39398504188799915, "grad_norm": 5.321859359741211, "learning_rate": 1.3839088378500926e-05, "loss": 3.3165, "step": 11181 }, { "epoch": 0.3940202789009576, "grad_norm": 1.9972856044769287, "learning_rate": 1.3838034531598364e-05, "loss": 1.0288, "step": 11182 }, { "epoch": 0.39405551591391597, "grad_norm": 9.921025276184082, "learning_rate": 1.3836980634704917e-05, "loss": 3.4822, "step": 11183 }, { "epoch": 0.3940907529268744, "grad_norm": 20.701444625854492, "learning_rate": 1.3835926687834306e-05, "loss": 6.1206, "step": 11184 }, { "epoch": 0.3941259899398328, "grad_norm": 10.065805435180664, "learning_rate": 1.3834872691000261e-05, "loss": 3.4111, "step": 11185 }, { "epoch": 0.3941612269527912, "grad_norm": 7.961300373077393, "learning_rate": 1.3833818644216513e-05, "loss": 5.2153, "step": 11186 }, { "epoch": 0.39419646396574964, "grad_norm": 12.689834594726562, "learning_rate": 1.3832764547496785e-05, "loss": 3.3841, "step": 11187 }, { "epoch": 0.394231700978708, "grad_norm": 16.279430389404297, "learning_rate": 1.3831710400854813e-05, "loss": 3.5238, "step": 11188 }, { "epoch": 0.39426693799166646, "grad_norm": 13.37366771697998, "learning_rate": 1.3830656204304324e-05, "loss": 5.805, "step": 11189 }, { "epoch": 0.39430217500462483, "grad_norm": 1.4028536081314087, "learning_rate": 1.3829601957859051e-05, "loss": 1.0995, "step": 11190 }, { "epoch": 0.39433741201758327, "grad_norm": 14.882131576538086, "learning_rate": 1.3828547661532723e-05, "loss": 3.0639, "step": 11191 }, { "epoch": 0.3943726490305417, "grad_norm": 1.460943341255188, "learning_rate": 1.3827493315339077e-05, "loss": 0.6969, "step": 11192 }, { "epoch": 0.3944078860435001, "grad_norm": 1.428889513015747, "learning_rate": 1.3826438919291838e-05, "loss": 1.0003, "step": 11193 }, { "epoch": 0.3944431230564585, "grad_norm": 1.2988715171813965, "learning_rate": 1.382538447340475e-05, "loss": 1.1497, "step": 11194 }, { "epoch": 0.3944783600694169, "grad_norm": 1.1882007122039795, "learning_rate": 1.3824329977691537e-05, "loss": 0.8103, "step": 11195 }, { "epoch": 0.3945135970823753, "grad_norm": 1.6738231182098389, "learning_rate": 1.3823275432165946e-05, "loss": 0.702, "step": 11196 }, { "epoch": 0.39454883409533376, "grad_norm": 2.691934585571289, "learning_rate": 1.38222208368417e-05, "loss": 0.7223, "step": 11197 }, { "epoch": 0.39458407110829213, "grad_norm": 1.0457690954208374, "learning_rate": 1.3821166191732543e-05, "loss": 0.8707, "step": 11198 }, { "epoch": 0.39461930812125057, "grad_norm": 19.077659606933594, "learning_rate": 1.3820111496852207e-05, "loss": 3.1079, "step": 11199 }, { "epoch": 0.394654545134209, "grad_norm": 18.008316040039062, "learning_rate": 1.3819056752214435e-05, "loss": 6.3657, "step": 11200 }, { "epoch": 0.3946897821471674, "grad_norm": 7.244303226470947, "learning_rate": 1.3818001957832961e-05, "loss": 3.4885, "step": 11201 }, { "epoch": 0.3947250191601258, "grad_norm": 6.960642337799072, "learning_rate": 1.3816947113721525e-05, "loss": 3.3466, "step": 11202 }, { "epoch": 0.3947602561730842, "grad_norm": 2.1308767795562744, "learning_rate": 1.3815892219893867e-05, "loss": 1.1545, "step": 11203 }, { "epoch": 0.3947954931860426, "grad_norm": 7.783133029937744, "learning_rate": 1.3814837276363727e-05, "loss": 3.2982, "step": 11204 }, { "epoch": 0.39483073019900106, "grad_norm": 12.38023853302002, "learning_rate": 1.3813782283144846e-05, "loss": 5.4676, "step": 11205 }, { "epoch": 0.39486596721195943, "grad_norm": 21.831527709960938, "learning_rate": 1.3812727240250963e-05, "loss": 3.458, "step": 11206 }, { "epoch": 0.39490120422491787, "grad_norm": 1.603759765625, "learning_rate": 1.3811672147695822e-05, "loss": 0.7946, "step": 11207 }, { "epoch": 0.39493644123787625, "grad_norm": 12.269331932067871, "learning_rate": 1.3810617005493167e-05, "loss": 4.0261, "step": 11208 }, { "epoch": 0.3949716782508347, "grad_norm": 18.84273338317871, "learning_rate": 1.380956181365674e-05, "loss": 2.9912, "step": 11209 }, { "epoch": 0.3950069152637931, "grad_norm": 1.3930257558822632, "learning_rate": 1.3808506572200285e-05, "loss": 1.3011, "step": 11210 }, { "epoch": 0.3950421522767515, "grad_norm": 1.5707508325576782, "learning_rate": 1.3807451281137549e-05, "loss": 0.8309, "step": 11211 }, { "epoch": 0.3950773892897099, "grad_norm": 8.352601051330566, "learning_rate": 1.380639594048227e-05, "loss": 3.2822, "step": 11212 }, { "epoch": 0.3951126263026683, "grad_norm": 1.212781548500061, "learning_rate": 1.3805340550248206e-05, "loss": 0.7845, "step": 11213 }, { "epoch": 0.39514786331562674, "grad_norm": 1.8857219219207764, "learning_rate": 1.380428511044909e-05, "loss": 1.1114, "step": 11214 }, { "epoch": 0.39518310032858517, "grad_norm": 13.42357349395752, "learning_rate": 1.380322962109868e-05, "loss": 6.6333, "step": 11215 }, { "epoch": 0.39521833734154355, "grad_norm": 12.177448272705078, "learning_rate": 1.3802174082210718e-05, "loss": 3.4968, "step": 11216 }, { "epoch": 0.395253574354502, "grad_norm": 10.990156173706055, "learning_rate": 1.3801118493798953e-05, "loss": 5.9597, "step": 11217 }, { "epoch": 0.39528881136746036, "grad_norm": 5.671419143676758, "learning_rate": 1.3800062855877136e-05, "loss": 2.5726, "step": 11218 }, { "epoch": 0.3953240483804188, "grad_norm": 5.5700178146362305, "learning_rate": 1.3799007168459019e-05, "loss": 2.8899, "step": 11219 }, { "epoch": 0.3953592853933772, "grad_norm": 1.137758493423462, "learning_rate": 1.3797951431558347e-05, "loss": 0.8458, "step": 11220 }, { "epoch": 0.3953945224063356, "grad_norm": 9.540183067321777, "learning_rate": 1.3796895645188877e-05, "loss": 3.3513, "step": 11221 }, { "epoch": 0.39542975941929404, "grad_norm": 1.556895136833191, "learning_rate": 1.3795839809364355e-05, "loss": 0.7482, "step": 11222 }, { "epoch": 0.3954649964322524, "grad_norm": 2.0019330978393555, "learning_rate": 1.3794783924098536e-05, "loss": 0.9725, "step": 11223 }, { "epoch": 0.39550023344521085, "grad_norm": 6.678134441375732, "learning_rate": 1.3793727989405173e-05, "loss": 3.3765, "step": 11224 }, { "epoch": 0.3955354704581693, "grad_norm": 8.406999588012695, "learning_rate": 1.3792672005298018e-05, "loss": 3.3257, "step": 11225 }, { "epoch": 0.39557070747112766, "grad_norm": 5.134078502655029, "learning_rate": 1.379161597179083e-05, "loss": 2.9163, "step": 11226 }, { "epoch": 0.3956059444840861, "grad_norm": 17.559818267822266, "learning_rate": 1.3790559888897358e-05, "loss": 3.0874, "step": 11227 }, { "epoch": 0.39564118149704447, "grad_norm": 28.447086334228516, "learning_rate": 1.3789503756631364e-05, "loss": 4.0155, "step": 11228 }, { "epoch": 0.3956764185100029, "grad_norm": 6.270000457763672, "learning_rate": 1.3788447575006599e-05, "loss": 3.4659, "step": 11229 }, { "epoch": 0.39571165552296134, "grad_norm": 18.52519416809082, "learning_rate": 1.3787391344036824e-05, "loss": 5.5158, "step": 11230 }, { "epoch": 0.3957468925359197, "grad_norm": 16.74005889892578, "learning_rate": 1.378633506373579e-05, "loss": 7.1114, "step": 11231 }, { "epoch": 0.39578212954887815, "grad_norm": 34.87547302246094, "learning_rate": 1.3785278734117265e-05, "loss": 5.9547, "step": 11232 }, { "epoch": 0.3958173665618365, "grad_norm": 1.3510410785675049, "learning_rate": 1.3784222355194995e-05, "loss": 0.8224, "step": 11233 }, { "epoch": 0.39585260357479496, "grad_norm": 26.522708892822266, "learning_rate": 1.3783165926982754e-05, "loss": 7.6249, "step": 11234 }, { "epoch": 0.3958878405877534, "grad_norm": 1.401604175567627, "learning_rate": 1.3782109449494292e-05, "loss": 0.9775, "step": 11235 }, { "epoch": 0.39592307760071177, "grad_norm": 1.0894883871078491, "learning_rate": 1.3781052922743372e-05, "loss": 1.032, "step": 11236 }, { "epoch": 0.3959583146136702, "grad_norm": 5.217658519744873, "learning_rate": 1.3779996346743756e-05, "loss": 3.2415, "step": 11237 }, { "epoch": 0.39599355162662864, "grad_norm": 7.609328746795654, "learning_rate": 1.377893972150921e-05, "loss": 3.0572, "step": 11238 }, { "epoch": 0.396028788639587, "grad_norm": 3.7280609607696533, "learning_rate": 1.3777883047053486e-05, "loss": 1.7946, "step": 11239 }, { "epoch": 0.39606402565254545, "grad_norm": 7.827577590942383, "learning_rate": 1.377682632339036e-05, "loss": 2.7139, "step": 11240 }, { "epoch": 0.3960992626655038, "grad_norm": 1.4134807586669922, "learning_rate": 1.3775769550533587e-05, "loss": 1.0795, "step": 11241 }, { "epoch": 0.39613449967846226, "grad_norm": 10.559182167053223, "learning_rate": 1.3774712728496936e-05, "loss": 4.5519, "step": 11242 }, { "epoch": 0.3961697366914207, "grad_norm": 2.3376708030700684, "learning_rate": 1.377365585729417e-05, "loss": 1.0384, "step": 11243 }, { "epoch": 0.39620497370437907, "grad_norm": 0.9007918238639832, "learning_rate": 1.3772598936939056e-05, "loss": 1.0938, "step": 11244 }, { "epoch": 0.3962402107173375, "grad_norm": 5.369001388549805, "learning_rate": 1.377154196744536e-05, "loss": 3.26, "step": 11245 }, { "epoch": 0.3962754477302959, "grad_norm": 5.294726848602295, "learning_rate": 1.377048494882685e-05, "loss": 3.5636, "step": 11246 }, { "epoch": 0.3963106847432543, "grad_norm": 1.9350742101669312, "learning_rate": 1.3769427881097293e-05, "loss": 0.8682, "step": 11247 }, { "epoch": 0.39634592175621275, "grad_norm": 1.7210659980773926, "learning_rate": 1.3768370764270456e-05, "loss": 0.6654, "step": 11248 }, { "epoch": 0.3963811587691711, "grad_norm": 1.4200211763381958, "learning_rate": 1.3767313598360112e-05, "loss": 0.9539, "step": 11249 }, { "epoch": 0.39641639578212956, "grad_norm": 1.4168992042541504, "learning_rate": 1.3766256383380028e-05, "loss": 0.9838, "step": 11250 }, { "epoch": 0.39645163279508794, "grad_norm": 5.6941914558410645, "learning_rate": 1.3765199119343975e-05, "loss": 3.3673, "step": 11251 }, { "epoch": 0.39648686980804637, "grad_norm": 3.3104777336120605, "learning_rate": 1.3764141806265722e-05, "loss": 3.3427, "step": 11252 }, { "epoch": 0.3965221068210048, "grad_norm": 1.2244352102279663, "learning_rate": 1.3763084444159048e-05, "loss": 1.5062, "step": 11253 }, { "epoch": 0.3965573438339632, "grad_norm": 9.907620429992676, "learning_rate": 1.3762027033037716e-05, "loss": 3.8121, "step": 11254 }, { "epoch": 0.3965925808469216, "grad_norm": 1.1558083295822144, "learning_rate": 1.37609695729155e-05, "loss": 1.067, "step": 11255 }, { "epoch": 0.39662781785988, "grad_norm": 4.4562225341796875, "learning_rate": 1.375991206380618e-05, "loss": 3.516, "step": 11256 }, { "epoch": 0.3966630548728384, "grad_norm": 6.530771255493164, "learning_rate": 1.3758854505723527e-05, "loss": 2.8738, "step": 11257 }, { "epoch": 0.39669829188579686, "grad_norm": 3.146972894668579, "learning_rate": 1.3757796898681311e-05, "loss": 3.2191, "step": 11258 }, { "epoch": 0.39673352889875524, "grad_norm": 9.579710006713867, "learning_rate": 1.3756739242693315e-05, "loss": 7.4378, "step": 11259 }, { "epoch": 0.3967687659117137, "grad_norm": 10.63296890258789, "learning_rate": 1.375568153777331e-05, "loss": 5.4984, "step": 11260 }, { "epoch": 0.39680400292467205, "grad_norm": 10.508065223693848, "learning_rate": 1.3754623783935078e-05, "loss": 3.4856, "step": 11261 }, { "epoch": 0.3968392399376305, "grad_norm": 5.370915412902832, "learning_rate": 1.375356598119239e-05, "loss": 3.9001, "step": 11262 }, { "epoch": 0.3968744769505889, "grad_norm": 1.0506259202957153, "learning_rate": 1.3752508129559028e-05, "loss": 1.0834, "step": 11263 }, { "epoch": 0.3969097139635473, "grad_norm": 1.308417558670044, "learning_rate": 1.3751450229048768e-05, "loss": 0.9107, "step": 11264 }, { "epoch": 0.39694495097650573, "grad_norm": 9.734201431274414, "learning_rate": 1.3750392279675393e-05, "loss": 6.2087, "step": 11265 }, { "epoch": 0.3969801879894641, "grad_norm": 5.545909404754639, "learning_rate": 1.374933428145268e-05, "loss": 3.2447, "step": 11266 }, { "epoch": 0.39701542500242254, "grad_norm": 1.6346430778503418, "learning_rate": 1.374827623439441e-05, "loss": 0.9508, "step": 11267 }, { "epoch": 0.397050662015381, "grad_norm": 5.707679748535156, "learning_rate": 1.3747218138514365e-05, "loss": 3.3533, "step": 11268 }, { "epoch": 0.39708589902833935, "grad_norm": 1.4579030275344849, "learning_rate": 1.3746159993826327e-05, "loss": 0.8552, "step": 11269 }, { "epoch": 0.3971211360412978, "grad_norm": 8.633578300476074, "learning_rate": 1.3745101800344079e-05, "loss": 3.1362, "step": 11270 }, { "epoch": 0.3971563730542562, "grad_norm": 3.884906768798828, "learning_rate": 1.3744043558081396e-05, "loss": 3.1721, "step": 11271 }, { "epoch": 0.3971916100672146, "grad_norm": 3.4329006671905518, "learning_rate": 1.3742985267052078e-05, "loss": 3.3767, "step": 11272 }, { "epoch": 0.39722684708017303, "grad_norm": 7.992238998413086, "learning_rate": 1.3741926927269896e-05, "loss": 3.5486, "step": 11273 }, { "epoch": 0.3972620840931314, "grad_norm": 9.18364143371582, "learning_rate": 1.374086853874864e-05, "loss": 2.5098, "step": 11274 }, { "epoch": 0.39729732110608984, "grad_norm": 1.4172033071517944, "learning_rate": 1.3739810101502093e-05, "loss": 0.9044, "step": 11275 }, { "epoch": 0.3973325581190483, "grad_norm": 7.889472484588623, "learning_rate": 1.3738751615544046e-05, "loss": 3.708, "step": 11276 }, { "epoch": 0.39736779513200665, "grad_norm": 4.473724842071533, "learning_rate": 1.3737693080888278e-05, "loss": 3.2003, "step": 11277 }, { "epoch": 0.3974030321449651, "grad_norm": 7.231119155883789, "learning_rate": 1.3736634497548586e-05, "loss": 5.7046, "step": 11278 }, { "epoch": 0.39743826915792346, "grad_norm": 1.6931031942367554, "learning_rate": 1.3735575865538751e-05, "loss": 0.8504, "step": 11279 }, { "epoch": 0.3974735061708819, "grad_norm": 6.458872318267822, "learning_rate": 1.3734517184872566e-05, "loss": 2.8691, "step": 11280 }, { "epoch": 0.39750874318384033, "grad_norm": 9.090652465820312, "learning_rate": 1.373345845556382e-05, "loss": 2.8977, "step": 11281 }, { "epoch": 0.3975439801967987, "grad_norm": 1.1100142002105713, "learning_rate": 1.37323996776263e-05, "loss": 1.047, "step": 11282 }, { "epoch": 0.39757921720975714, "grad_norm": 5.645293712615967, "learning_rate": 1.37313408510738e-05, "loss": 3.0911, "step": 11283 }, { "epoch": 0.3976144542227155, "grad_norm": 5.547700881958008, "learning_rate": 1.373028197592011e-05, "loss": 3.5045, "step": 11284 }, { "epoch": 0.39764969123567395, "grad_norm": 3.873826503753662, "learning_rate": 1.3729223052179022e-05, "loss": 3.0479, "step": 11285 }, { "epoch": 0.3976849282486324, "grad_norm": 7.754401683807373, "learning_rate": 1.3728164079864328e-05, "loss": 5.7239, "step": 11286 }, { "epoch": 0.39772016526159076, "grad_norm": 7.434075355529785, "learning_rate": 1.3727105058989822e-05, "loss": 4.13, "step": 11287 }, { "epoch": 0.3977554022745492, "grad_norm": 2.7242045402526855, "learning_rate": 1.3726045989569299e-05, "loss": 3.303, "step": 11288 }, { "epoch": 0.3977906392875076, "grad_norm": 4.5339789390563965, "learning_rate": 1.3724986871616552e-05, "loss": 3.0855, "step": 11289 }, { "epoch": 0.397825876300466, "grad_norm": 1.4899204969406128, "learning_rate": 1.3723927705145374e-05, "loss": 1.143, "step": 11290 }, { "epoch": 0.39786111331342444, "grad_norm": 7.772304534912109, "learning_rate": 1.3722868490169567e-05, "loss": 3.2572, "step": 11291 }, { "epoch": 0.3978963503263828, "grad_norm": 2.097221612930298, "learning_rate": 1.3721809226702923e-05, "loss": 0.8341, "step": 11292 }, { "epoch": 0.39793158733934125, "grad_norm": 12.310519218444824, "learning_rate": 1.372074991475924e-05, "loss": 3.5326, "step": 11293 }, { "epoch": 0.39796682435229963, "grad_norm": 6.307734966278076, "learning_rate": 1.3719690554352311e-05, "loss": 3.2301, "step": 11294 }, { "epoch": 0.39800206136525806, "grad_norm": 1.41530179977417, "learning_rate": 1.3718631145495946e-05, "loss": 0.8879, "step": 11295 }, { "epoch": 0.3980372983782165, "grad_norm": 8.358152389526367, "learning_rate": 1.371757168820393e-05, "loss": 5.694, "step": 11296 }, { "epoch": 0.3980725353911749, "grad_norm": 7.720884799957275, "learning_rate": 1.3716512182490073e-05, "loss": 3.7961, "step": 11297 }, { "epoch": 0.3981077724041333, "grad_norm": 7.285190105438232, "learning_rate": 1.371545262836817e-05, "loss": 2.7643, "step": 11298 }, { "epoch": 0.3981430094170917, "grad_norm": 7.544274806976318, "learning_rate": 1.3714393025852024e-05, "loss": 2.5846, "step": 11299 }, { "epoch": 0.3981782464300501, "grad_norm": 1.375565528869629, "learning_rate": 1.3713333374955436e-05, "loss": 0.6702, "step": 11300 }, { "epoch": 0.39821348344300855, "grad_norm": 1.5077757835388184, "learning_rate": 1.3712273675692207e-05, "loss": 0.8481, "step": 11301 }, { "epoch": 0.39824872045596693, "grad_norm": 1.1243220567703247, "learning_rate": 1.3711213928076142e-05, "loss": 0.9136, "step": 11302 }, { "epoch": 0.39828395746892536, "grad_norm": 0.998925507068634, "learning_rate": 1.3710154132121042e-05, "loss": 0.9083, "step": 11303 }, { "epoch": 0.3983191944818838, "grad_norm": 7.064321041107178, "learning_rate": 1.3709094287840707e-05, "loss": 3.7682, "step": 11304 }, { "epoch": 0.3983544314948422, "grad_norm": 1.8900281190872192, "learning_rate": 1.3708034395248953e-05, "loss": 0.9665, "step": 11305 }, { "epoch": 0.3983896685078006, "grad_norm": 11.032437324523926, "learning_rate": 1.3706974454359574e-05, "loss": 5.3283, "step": 11306 }, { "epoch": 0.398424905520759, "grad_norm": 11.032437324523926, "learning_rate": 1.3706974454359574e-05, "loss": 3.2958, "step": 11307 }, { "epoch": 0.3984601425337174, "grad_norm": 1.6476702690124512, "learning_rate": 1.3705914465186383e-05, "loss": 1.0032, "step": 11308 }, { "epoch": 0.39849537954667585, "grad_norm": 5.17325496673584, "learning_rate": 1.3704854427743186e-05, "loss": 3.5554, "step": 11309 }, { "epoch": 0.39853061655963423, "grad_norm": 0.8733721971511841, "learning_rate": 1.3703794342043786e-05, "loss": 0.8092, "step": 11310 }, { "epoch": 0.39856585357259267, "grad_norm": 0.9913507699966431, "learning_rate": 1.3702734208101993e-05, "loss": 0.7503, "step": 11311 }, { "epoch": 0.39860109058555104, "grad_norm": 7.810795783996582, "learning_rate": 1.3701674025931616e-05, "loss": 3.3574, "step": 11312 }, { "epoch": 0.3986363275985095, "grad_norm": 4.383999347686768, "learning_rate": 1.3700613795546462e-05, "loss": 3.0049, "step": 11313 }, { "epoch": 0.3986715646114679, "grad_norm": 1.1706857681274414, "learning_rate": 1.3699553516960342e-05, "loss": 0.7324, "step": 11314 }, { "epoch": 0.3987068016244263, "grad_norm": 9.108078002929688, "learning_rate": 1.3698493190187068e-05, "loss": 3.1939, "step": 11315 }, { "epoch": 0.3987420386373847, "grad_norm": 5.2478461265563965, "learning_rate": 1.3697432815240448e-05, "loss": 3.6729, "step": 11316 }, { "epoch": 0.3987772756503431, "grad_norm": 9.909257888793945, "learning_rate": 1.3696372392134296e-05, "loss": 3.767, "step": 11317 }, { "epoch": 0.39881251266330153, "grad_norm": 2.133937358856201, "learning_rate": 1.3695311920882417e-05, "loss": 0.8789, "step": 11318 }, { "epoch": 0.39884774967625997, "grad_norm": 1.6775373220443726, "learning_rate": 1.3694251401498638e-05, "loss": 1.1714, "step": 11319 }, { "epoch": 0.39888298668921834, "grad_norm": 6.534048557281494, "learning_rate": 1.3693190833996757e-05, "loss": 3.4879, "step": 11320 }, { "epoch": 0.3989182237021768, "grad_norm": 5.834414482116699, "learning_rate": 1.3692130218390597e-05, "loss": 2.8922, "step": 11321 }, { "epoch": 0.39895346071513516, "grad_norm": 6.989586353302002, "learning_rate": 1.3691069554693973e-05, "loss": 3.3617, "step": 11322 }, { "epoch": 0.3989886977280936, "grad_norm": 1.2487001419067383, "learning_rate": 1.3690008842920697e-05, "loss": 1.3041, "step": 11323 }, { "epoch": 0.399023934741052, "grad_norm": 3.583037853240967, "learning_rate": 1.3688948083084583e-05, "loss": 2.9714, "step": 11324 }, { "epoch": 0.3990591717540104, "grad_norm": 1.6122325658798218, "learning_rate": 1.3687887275199454e-05, "loss": 0.938, "step": 11325 }, { "epoch": 0.39909440876696883, "grad_norm": 3.673590898513794, "learning_rate": 1.368682641927912e-05, "loss": 3.441, "step": 11326 }, { "epoch": 0.3991296457799272, "grad_norm": 1.2850158214569092, "learning_rate": 1.3685765515337406e-05, "loss": 0.924, "step": 11327 }, { "epoch": 0.39916488279288564, "grad_norm": 1.3588464260101318, "learning_rate": 1.3684704563388123e-05, "loss": 0.863, "step": 11328 }, { "epoch": 0.3992001198058441, "grad_norm": 9.37705135345459, "learning_rate": 1.3683643563445095e-05, "loss": 5.0928, "step": 11329 }, { "epoch": 0.39923535681880246, "grad_norm": 5.13537073135376, "learning_rate": 1.368258251552214e-05, "loss": 3.128, "step": 11330 }, { "epoch": 0.3992705938317609, "grad_norm": 1.2987288236618042, "learning_rate": 1.3681521419633078e-05, "loss": 1.0885, "step": 11331 }, { "epoch": 0.39930583084471927, "grad_norm": 2.06591534614563, "learning_rate": 1.3680460275791733e-05, "loss": 0.9991, "step": 11332 }, { "epoch": 0.3993410678576777, "grad_norm": 1.6950771808624268, "learning_rate": 1.3679399084011923e-05, "loss": 0.9317, "step": 11333 }, { "epoch": 0.39937630487063613, "grad_norm": 1.2551918029785156, "learning_rate": 1.3678337844307472e-05, "loss": 0.8063, "step": 11334 }, { "epoch": 0.3994115418835945, "grad_norm": 5.611379146575928, "learning_rate": 1.36772765566922e-05, "loss": 3.4951, "step": 11335 }, { "epoch": 0.39944677889655295, "grad_norm": 5.194058895111084, "learning_rate": 1.3676215221179936e-05, "loss": 3.3534, "step": 11336 }, { "epoch": 0.3994820159095113, "grad_norm": 9.701969146728516, "learning_rate": 1.3675153837784497e-05, "loss": 4.1303, "step": 11337 }, { "epoch": 0.39951725292246976, "grad_norm": 3.3895509243011475, "learning_rate": 1.3674092406519712e-05, "loss": 3.4119, "step": 11338 }, { "epoch": 0.3995524899354282, "grad_norm": 1.305537223815918, "learning_rate": 1.3673030927399406e-05, "loss": 1.0591, "step": 11339 }, { "epoch": 0.39958772694838657, "grad_norm": 15.48573112487793, "learning_rate": 1.3671969400437403e-05, "loss": 8.0616, "step": 11340 }, { "epoch": 0.399622963961345, "grad_norm": 1.138023853302002, "learning_rate": 1.3670907825647533e-05, "loss": 0.9917, "step": 11341 }, { "epoch": 0.39965820097430343, "grad_norm": 7.068296909332275, "learning_rate": 1.366984620304362e-05, "loss": 3.8965, "step": 11342 }, { "epoch": 0.3996934379872618, "grad_norm": 1.8681870698928833, "learning_rate": 1.3668784532639493e-05, "loss": 0.6931, "step": 11343 }, { "epoch": 0.39972867500022025, "grad_norm": 12.7821683883667, "learning_rate": 1.366772281444898e-05, "loss": 4.4183, "step": 11344 }, { "epoch": 0.3997639120131786, "grad_norm": 4.507303237915039, "learning_rate": 1.3666661048485908e-05, "loss": 3.2583, "step": 11345 }, { "epoch": 0.39979914902613706, "grad_norm": 5.245275974273682, "learning_rate": 1.3665599234764111e-05, "loss": 3.1745, "step": 11346 }, { "epoch": 0.3998343860390955, "grad_norm": 5.215296745300293, "learning_rate": 1.3664537373297417e-05, "loss": 3.0758, "step": 11347 }, { "epoch": 0.39986962305205387, "grad_norm": 1.101096510887146, "learning_rate": 1.3663475464099656e-05, "loss": 1.3087, "step": 11348 }, { "epoch": 0.3999048600650123, "grad_norm": 1.1862478256225586, "learning_rate": 1.3662413507184662e-05, "loss": 0.9194, "step": 11349 }, { "epoch": 0.3999400970779707, "grad_norm": 1.8799010515213013, "learning_rate": 1.3661351502566263e-05, "loss": 1.0832, "step": 11350 }, { "epoch": 0.3999753340909291, "grad_norm": 1.2606576681137085, "learning_rate": 1.3660289450258297e-05, "loss": 1.1902, "step": 11351 }, { "epoch": 0.40001057110388755, "grad_norm": 6.917562484741211, "learning_rate": 1.3659227350274594e-05, "loss": 2.8804, "step": 11352 }, { "epoch": 0.4000458081168459, "grad_norm": 1.159103274345398, "learning_rate": 1.3658165202628985e-05, "loss": 1.1865, "step": 11353 }, { "epoch": 0.40008104512980436, "grad_norm": 8.182695388793945, "learning_rate": 1.3657103007335312e-05, "loss": 3.2877, "step": 11354 }, { "epoch": 0.40011628214276274, "grad_norm": 1.4667341709136963, "learning_rate": 1.3656040764407408e-05, "loss": 0.9563, "step": 11355 }, { "epoch": 0.40015151915572117, "grad_norm": 8.014440536499023, "learning_rate": 1.3654978473859105e-05, "loss": 5.2262, "step": 11356 }, { "epoch": 0.4001867561686796, "grad_norm": 1.4981921911239624, "learning_rate": 1.3653916135704241e-05, "loss": 1.0308, "step": 11357 }, { "epoch": 0.400221993181638, "grad_norm": 6.527541637420654, "learning_rate": 1.3652853749956657e-05, "loss": 5.4966, "step": 11358 }, { "epoch": 0.4002572301945964, "grad_norm": 18.873973846435547, "learning_rate": 1.3651791316630183e-05, "loss": 3.7664, "step": 11359 }, { "epoch": 0.4002924672075548, "grad_norm": 1.813425064086914, "learning_rate": 1.3650728835738665e-05, "loss": 0.8815, "step": 11360 }, { "epoch": 0.4003277042205132, "grad_norm": 17.228347778320312, "learning_rate": 1.364966630729594e-05, "loss": 4.2136, "step": 11361 }, { "epoch": 0.40036294123347166, "grad_norm": 9.419495582580566, "learning_rate": 1.3648603731315844e-05, "loss": 3.3312, "step": 11362 }, { "epoch": 0.40039817824643004, "grad_norm": 4.0768723487854, "learning_rate": 1.364754110781222e-05, "loss": 2.8926, "step": 11363 }, { "epoch": 0.40043341525938847, "grad_norm": 4.6313066482543945, "learning_rate": 1.3646478436798906e-05, "loss": 3.204, "step": 11364 }, { "epoch": 0.40046865227234685, "grad_norm": 5.6248579025268555, "learning_rate": 1.3645415718289754e-05, "loss": 3.6402, "step": 11365 }, { "epoch": 0.4005038892853053, "grad_norm": 1.5484265089035034, "learning_rate": 1.3644352952298592e-05, "loss": 0.7272, "step": 11366 }, { "epoch": 0.4005391262982637, "grad_norm": 1.1810816526412964, "learning_rate": 1.3643290138839267e-05, "loss": 0.8129, "step": 11367 }, { "epoch": 0.4005743633112221, "grad_norm": 1.2567137479782104, "learning_rate": 1.3642227277925627e-05, "loss": 0.9105, "step": 11368 }, { "epoch": 0.4006096003241805, "grad_norm": 11.822000503540039, "learning_rate": 1.364116436957151e-05, "loss": 4.8114, "step": 11369 }, { "epoch": 0.4006448373371389, "grad_norm": 0.8144353628158569, "learning_rate": 1.3640101413790768e-05, "loss": 1.0228, "step": 11370 }, { "epoch": 0.40068007435009734, "grad_norm": 12.125638008117676, "learning_rate": 1.3639038410597239e-05, "loss": 5.194, "step": 11371 }, { "epoch": 0.40071531136305577, "grad_norm": 5.7266669273376465, "learning_rate": 1.363797536000477e-05, "loss": 3.1673, "step": 11372 }, { "epoch": 0.40075054837601415, "grad_norm": 1.5058441162109375, "learning_rate": 1.3636912262027209e-05, "loss": 0.9372, "step": 11373 }, { "epoch": 0.4007857853889726, "grad_norm": 1.657389760017395, "learning_rate": 1.3635849116678404e-05, "loss": 0.7731, "step": 11374 }, { "epoch": 0.400821022401931, "grad_norm": 1.1591311693191528, "learning_rate": 1.3634785923972197e-05, "loss": 1.0276, "step": 11375 }, { "epoch": 0.4008562594148894, "grad_norm": 1.080173373222351, "learning_rate": 1.3633722683922445e-05, "loss": 1.0331, "step": 11376 }, { "epoch": 0.4008914964278478, "grad_norm": 44.21908950805664, "learning_rate": 1.3632659396542991e-05, "loss": 3.5789, "step": 11377 }, { "epoch": 0.4009267334408062, "grad_norm": 8.344344139099121, "learning_rate": 1.3631596061847684e-05, "loss": 3.4182, "step": 11378 }, { "epoch": 0.40096197045376464, "grad_norm": 1.4184952974319458, "learning_rate": 1.3630532679850377e-05, "loss": 0.9148, "step": 11379 }, { "epoch": 0.40099720746672307, "grad_norm": 1.4635714292526245, "learning_rate": 1.362946925056492e-05, "loss": 0.9869, "step": 11380 }, { "epoch": 0.40103244447968145, "grad_norm": 0.9069873094558716, "learning_rate": 1.3628405774005162e-05, "loss": 0.7586, "step": 11381 }, { "epoch": 0.4010676814926399, "grad_norm": 14.910326957702637, "learning_rate": 1.3627342250184958e-05, "loss": 3.5095, "step": 11382 }, { "epoch": 0.40110291850559826, "grad_norm": 1.2758548259735107, "learning_rate": 1.3626278679118157e-05, "loss": 0.7763, "step": 11383 }, { "epoch": 0.4011381555185567, "grad_norm": 1.887575387954712, "learning_rate": 1.3625215060818618e-05, "loss": 1.2334, "step": 11384 }, { "epoch": 0.4011733925315151, "grad_norm": 14.762227058410645, "learning_rate": 1.3624151395300188e-05, "loss": 7.9958, "step": 11385 }, { "epoch": 0.4012086295444735, "grad_norm": 17.046659469604492, "learning_rate": 1.3623087682576726e-05, "loss": 3.7486, "step": 11386 }, { "epoch": 0.40124386655743194, "grad_norm": 1.9846609830856323, "learning_rate": 1.3622023922662084e-05, "loss": 1.2213, "step": 11387 }, { "epoch": 0.4012791035703903, "grad_norm": 1.1800395250320435, "learning_rate": 1.3620960115570122e-05, "loss": 0.8966, "step": 11388 }, { "epoch": 0.40131434058334875, "grad_norm": 15.763744354248047, "learning_rate": 1.361989626131469e-05, "loss": 3.6876, "step": 11389 }, { "epoch": 0.4013495775963072, "grad_norm": 64.7420883178711, "learning_rate": 1.3618832359909649e-05, "loss": 2.7439, "step": 11390 }, { "epoch": 0.40138481460926556, "grad_norm": 8.175697326660156, "learning_rate": 1.3617768411368855e-05, "loss": 3.1804, "step": 11391 }, { "epoch": 0.401420051622224, "grad_norm": 19.415483474731445, "learning_rate": 1.3616704415706166e-05, "loss": 7.836, "step": 11392 }, { "epoch": 0.4014552886351824, "grad_norm": 1.691481113433838, "learning_rate": 1.3615640372935444e-05, "loss": 1.2939, "step": 11393 }, { "epoch": 0.4014905256481408, "grad_norm": 1.5325907468795776, "learning_rate": 1.3614576283070541e-05, "loss": 1.0152, "step": 11394 }, { "epoch": 0.40152576266109924, "grad_norm": 1.2210767269134521, "learning_rate": 1.3613512146125327e-05, "loss": 0.7816, "step": 11395 }, { "epoch": 0.4015609996740576, "grad_norm": 1.4914709329605103, "learning_rate": 1.3612447962113654e-05, "loss": 0.9623, "step": 11396 }, { "epoch": 0.40159623668701605, "grad_norm": 9.387298583984375, "learning_rate": 1.3611383731049388e-05, "loss": 3.5604, "step": 11397 }, { "epoch": 0.40163147369997443, "grad_norm": 1.8768821954727173, "learning_rate": 1.3610319452946387e-05, "loss": 1.2188, "step": 11398 }, { "epoch": 0.40166671071293286, "grad_norm": 4.140051364898682, "learning_rate": 1.3609255127818517e-05, "loss": 2.8021, "step": 11399 }, { "epoch": 0.4017019477258913, "grad_norm": 8.633102416992188, "learning_rate": 1.3608190755679638e-05, "loss": 7.5802, "step": 11400 }, { "epoch": 0.4017371847388497, "grad_norm": 7.937446594238281, "learning_rate": 1.3607126336543617e-05, "loss": 2.8948, "step": 11401 }, { "epoch": 0.4017724217518081, "grad_norm": 4.48113489151001, "learning_rate": 1.3606061870424313e-05, "loss": 3.5384, "step": 11402 }, { "epoch": 0.4018076587647665, "grad_norm": 0.9556094408035278, "learning_rate": 1.3604997357335593e-05, "loss": 1.0076, "step": 11403 }, { "epoch": 0.4018428957777249, "grad_norm": 1.0051054954528809, "learning_rate": 1.3603932797291327e-05, "loss": 1.0812, "step": 11404 }, { "epoch": 0.40187813279068335, "grad_norm": 11.644538879394531, "learning_rate": 1.3602868190305376e-05, "loss": 5.157, "step": 11405 }, { "epoch": 0.40191336980364173, "grad_norm": 5.253298282623291, "learning_rate": 1.3601803536391604e-05, "loss": 3.4848, "step": 11406 }, { "epoch": 0.40194860681660016, "grad_norm": 1.0898200273513794, "learning_rate": 1.3600738835563886e-05, "loss": 1.17, "step": 11407 }, { "epoch": 0.4019838438295586, "grad_norm": 1.363399624824524, "learning_rate": 1.3599674087836085e-05, "loss": 1.0178, "step": 11408 }, { "epoch": 0.402019080842517, "grad_norm": 3.744760274887085, "learning_rate": 1.3598609293222071e-05, "loss": 3.5568, "step": 11409 }, { "epoch": 0.4020543178554754, "grad_norm": 3.1680052280426025, "learning_rate": 1.3597544451735712e-05, "loss": 3.0243, "step": 11410 }, { "epoch": 0.4020895548684338, "grad_norm": 7.359568119049072, "learning_rate": 1.3596479563390878e-05, "loss": 6.2042, "step": 11411 }, { "epoch": 0.4021247918813922, "grad_norm": 6.614140033721924, "learning_rate": 1.3595414628201442e-05, "loss": 5.5213, "step": 11412 }, { "epoch": 0.40216002889435065, "grad_norm": 1.3601758480072021, "learning_rate": 1.3594349646181268e-05, "loss": 1.1362, "step": 11413 }, { "epoch": 0.40219526590730903, "grad_norm": 11.731420516967773, "learning_rate": 1.3593284617344238e-05, "loss": 3.2865, "step": 11414 }, { "epoch": 0.40223050292026746, "grad_norm": 1.65241277217865, "learning_rate": 1.3592219541704213e-05, "loss": 0.8598, "step": 11415 }, { "epoch": 0.40226573993322584, "grad_norm": 7.919473648071289, "learning_rate": 1.3591154419275075e-05, "loss": 5.9097, "step": 11416 }, { "epoch": 0.4023009769461843, "grad_norm": 7.221809387207031, "learning_rate": 1.3590089250070691e-05, "loss": 5.9567, "step": 11417 }, { "epoch": 0.4023362139591427, "grad_norm": 1.6917921304702759, "learning_rate": 1.3589024034104937e-05, "loss": 0.9056, "step": 11418 }, { "epoch": 0.4023714509721011, "grad_norm": 1.8299486637115479, "learning_rate": 1.3587958771391686e-05, "loss": 0.9394, "step": 11419 }, { "epoch": 0.4024066879850595, "grad_norm": 7.74230432510376, "learning_rate": 1.358689346194482e-05, "loss": 3.1695, "step": 11420 }, { "epoch": 0.4024419249980179, "grad_norm": 1.0588067770004272, "learning_rate": 1.3585828105778202e-05, "loss": 0.9034, "step": 11421 }, { "epoch": 0.40247716201097633, "grad_norm": 6.551852226257324, "learning_rate": 1.3584762702905724e-05, "loss": 5.1054, "step": 11422 }, { "epoch": 0.40251239902393476, "grad_norm": 1.8810324668884277, "learning_rate": 1.3583697253341253e-05, "loss": 1.0612, "step": 11423 }, { "epoch": 0.40254763603689314, "grad_norm": 1.9781702756881714, "learning_rate": 1.3582631757098668e-05, "loss": 1.1535, "step": 11424 }, { "epoch": 0.4025828730498516, "grad_norm": 1.1818000078201294, "learning_rate": 1.3581566214191846e-05, "loss": 1.1345, "step": 11425 }, { "epoch": 0.40261811006280995, "grad_norm": 1.3770992755889893, "learning_rate": 1.3580500624634669e-05, "loss": 0.857, "step": 11426 }, { "epoch": 0.4026533470757684, "grad_norm": 6.623347282409668, "learning_rate": 1.3579434988441016e-05, "loss": 3.1051, "step": 11427 }, { "epoch": 0.4026885840887268, "grad_norm": 8.290984153747559, "learning_rate": 1.3578369305624765e-05, "loss": 3.4018, "step": 11428 }, { "epoch": 0.4027238211016852, "grad_norm": 1.6405633687973022, "learning_rate": 1.3577303576199801e-05, "loss": 0.8735, "step": 11429 }, { "epoch": 0.40275905811464363, "grad_norm": 6.388003826141357, "learning_rate": 1.357623780018e-05, "loss": 3.6253, "step": 11430 }, { "epoch": 0.402794295127602, "grad_norm": 2.087043046951294, "learning_rate": 1.357517197757925e-05, "loss": 1.0125, "step": 11431 }, { "epoch": 0.40282953214056044, "grad_norm": 1.2216925621032715, "learning_rate": 1.3574106108411422e-05, "loss": 1.0854, "step": 11432 }, { "epoch": 0.4028647691535189, "grad_norm": 6.3060760498046875, "learning_rate": 1.3573040192690412e-05, "loss": 3.5471, "step": 11433 }, { "epoch": 0.40290000616647725, "grad_norm": 5.153471946716309, "learning_rate": 1.3571974230430096e-05, "loss": 3.7118, "step": 11434 }, { "epoch": 0.4029352431794357, "grad_norm": 5.393093109130859, "learning_rate": 1.3570908221644363e-05, "loss": 3.5834, "step": 11435 }, { "epoch": 0.40297048019239406, "grad_norm": 4.983448505401611, "learning_rate": 1.3569842166347096e-05, "loss": 3.0928, "step": 11436 }, { "epoch": 0.4030057172053525, "grad_norm": 1.0837733745574951, "learning_rate": 1.3568776064552177e-05, "loss": 0.9019, "step": 11437 }, { "epoch": 0.40304095421831093, "grad_norm": 1.6836482286453247, "learning_rate": 1.3567709916273496e-05, "loss": 0.9013, "step": 11438 }, { "epoch": 0.4030761912312693, "grad_norm": 5.583499431610107, "learning_rate": 1.356664372152494e-05, "loss": 3.1864, "step": 11439 }, { "epoch": 0.40311142824422774, "grad_norm": 7.515836715698242, "learning_rate": 1.3565577480320395e-05, "loss": 3.1387, "step": 11440 }, { "epoch": 0.4031466652571861, "grad_norm": 5.551714897155762, "learning_rate": 1.356451119267375e-05, "loss": 3.0243, "step": 11441 }, { "epoch": 0.40318190227014455, "grad_norm": 1.8526681661605835, "learning_rate": 1.3563444858598889e-05, "loss": 1.1436, "step": 11442 }, { "epoch": 0.403217139283103, "grad_norm": 4.980816841125488, "learning_rate": 1.356237847810971e-05, "loss": 3.644, "step": 11443 }, { "epoch": 0.40325237629606137, "grad_norm": 9.200495719909668, "learning_rate": 1.3561312051220095e-05, "loss": 8.1305, "step": 11444 }, { "epoch": 0.4032876133090198, "grad_norm": 13.272778511047363, "learning_rate": 1.3560245577943935e-05, "loss": 6.1996, "step": 11445 }, { "epoch": 0.40332285032197823, "grad_norm": 5.209394454956055, "learning_rate": 1.3559179058295125e-05, "loss": 4.7818, "step": 11446 }, { "epoch": 0.4033580873349366, "grad_norm": 5.201747894287109, "learning_rate": 1.3558112492287554e-05, "loss": 3.4474, "step": 11447 }, { "epoch": 0.40339332434789504, "grad_norm": 1.7128093242645264, "learning_rate": 1.3557045879935116e-05, "loss": 0.9737, "step": 11448 }, { "epoch": 0.4034285613608534, "grad_norm": 1.1595561504364014, "learning_rate": 1.3555979221251699e-05, "loss": 0.8194, "step": 11449 }, { "epoch": 0.40346379837381185, "grad_norm": 0.9852051138877869, "learning_rate": 1.3554912516251203e-05, "loss": 1.1545, "step": 11450 }, { "epoch": 0.4034990353867703, "grad_norm": 3.8628318309783936, "learning_rate": 1.3553845764947515e-05, "loss": 3.3836, "step": 11451 }, { "epoch": 0.40353427239972867, "grad_norm": 3.967989444732666, "learning_rate": 1.3552778967354538e-05, "loss": 3.1424, "step": 11452 }, { "epoch": 0.4035695094126871, "grad_norm": 1.3035117387771606, "learning_rate": 1.3551712123486157e-05, "loss": 0.9271, "step": 11453 }, { "epoch": 0.4036047464256455, "grad_norm": 1.6137688159942627, "learning_rate": 1.355064523335628e-05, "loss": 0.7339, "step": 11454 }, { "epoch": 0.4036399834386039, "grad_norm": 3.6818594932556152, "learning_rate": 1.3549578296978792e-05, "loss": 3.3569, "step": 11455 }, { "epoch": 0.40367522045156234, "grad_norm": 17.09050941467285, "learning_rate": 1.3548511314367594e-05, "loss": 5.7766, "step": 11456 }, { "epoch": 0.4037104574645207, "grad_norm": 1.4703963994979858, "learning_rate": 1.3547444285536584e-05, "loss": 0.9249, "step": 11457 }, { "epoch": 0.40374569447747916, "grad_norm": 1.421014666557312, "learning_rate": 1.3546377210499664e-05, "loss": 0.8583, "step": 11458 }, { "epoch": 0.40378093149043753, "grad_norm": 3.6871910095214844, "learning_rate": 1.3545310089270725e-05, "loss": 2.8007, "step": 11459 }, { "epoch": 0.40381616850339597, "grad_norm": 3.505568504333496, "learning_rate": 1.3544242921863673e-05, "loss": 2.9468, "step": 11460 }, { "epoch": 0.4038514055163544, "grad_norm": 8.730182647705078, "learning_rate": 1.3543175708292402e-05, "loss": 3.3959, "step": 11461 }, { "epoch": 0.4038866425293128, "grad_norm": 3.2838857173919678, "learning_rate": 1.3542108448570821e-05, "loss": 3.4644, "step": 11462 }, { "epoch": 0.4039218795422712, "grad_norm": 13.169657707214355, "learning_rate": 1.3541041142712822e-05, "loss": 5.959, "step": 11463 }, { "epoch": 0.4039571165552296, "grad_norm": 11.787766456604004, "learning_rate": 1.3539973790732314e-05, "loss": 5.413, "step": 11464 }, { "epoch": 0.403992353568188, "grad_norm": 2.4020400047302246, "learning_rate": 1.3538906392643197e-05, "loss": 1.0914, "step": 11465 }, { "epoch": 0.40402759058114646, "grad_norm": 2.0226829051971436, "learning_rate": 1.353783894845937e-05, "loss": 0.9401, "step": 11466 }, { "epoch": 0.40406282759410483, "grad_norm": 1.6375277042388916, "learning_rate": 1.3536771458194742e-05, "loss": 1.0809, "step": 11467 }, { "epoch": 0.40409806460706327, "grad_norm": 6.335680961608887, "learning_rate": 1.3535703921863217e-05, "loss": 2.931, "step": 11468 }, { "epoch": 0.40413330162002165, "grad_norm": 1.4205293655395508, "learning_rate": 1.3534636339478697e-05, "loss": 1.034, "step": 11469 }, { "epoch": 0.4041685386329801, "grad_norm": 3.695469379425049, "learning_rate": 1.3533568711055087e-05, "loss": 3.0853, "step": 11470 }, { "epoch": 0.4042037756459385, "grad_norm": 1.1589518785476685, "learning_rate": 1.3532501036606296e-05, "loss": 0.9748, "step": 11471 }, { "epoch": 0.4042390126588969, "grad_norm": 2.529930353164673, "learning_rate": 1.3531433316146227e-05, "loss": 0.914, "step": 11472 }, { "epoch": 0.4042742496718553, "grad_norm": 1.4247338771820068, "learning_rate": 1.3530365549688791e-05, "loss": 0.9286, "step": 11473 }, { "epoch": 0.4043094866848137, "grad_norm": 3.917711019515991, "learning_rate": 1.3529297737247893e-05, "loss": 2.9534, "step": 11474 }, { "epoch": 0.40434472369777213, "grad_norm": 3.5613715648651123, "learning_rate": 1.3528229878837442e-05, "loss": 2.2786, "step": 11475 }, { "epoch": 0.40437996071073057, "grad_norm": 1.015270709991455, "learning_rate": 1.3527161974471348e-05, "loss": 0.9109, "step": 11476 }, { "epoch": 0.40441519772368895, "grad_norm": 7.1453471183776855, "learning_rate": 1.3526094024163523e-05, "loss": 3.9026, "step": 11477 }, { "epoch": 0.4044504347366474, "grad_norm": 6.8727569580078125, "learning_rate": 1.3525026027927868e-05, "loss": 3.1028, "step": 11478 }, { "epoch": 0.4044856717496058, "grad_norm": 7.271147727966309, "learning_rate": 1.3523957985778306e-05, "loss": 3.3467, "step": 11479 }, { "epoch": 0.4045209087625642, "grad_norm": 3.6898438930511475, "learning_rate": 1.3522889897728737e-05, "loss": 2.7979, "step": 11480 }, { "epoch": 0.4045561457755226, "grad_norm": 1.1571128368377686, "learning_rate": 1.3521821763793082e-05, "loss": 1.0785, "step": 11481 }, { "epoch": 0.404591382788481, "grad_norm": 7.4946112632751465, "learning_rate": 1.3520753583985248e-05, "loss": 3.1248, "step": 11482 }, { "epoch": 0.40462661980143944, "grad_norm": 2.194392204284668, "learning_rate": 1.3519685358319153e-05, "loss": 0.9473, "step": 11483 }, { "epoch": 0.40466185681439787, "grad_norm": 1.001281976699829, "learning_rate": 1.3518617086808705e-05, "loss": 0.9984, "step": 11484 }, { "epoch": 0.40469709382735625, "grad_norm": 1.5512408018112183, "learning_rate": 1.3517548769467825e-05, "loss": 0.9338, "step": 11485 }, { "epoch": 0.4047323308403147, "grad_norm": 2.747159957885742, "learning_rate": 1.3516480406310418e-05, "loss": 2.6713, "step": 11486 }, { "epoch": 0.40476756785327306, "grad_norm": 8.814836502075195, "learning_rate": 1.3515411997350412e-05, "loss": 3.6264, "step": 11487 }, { "epoch": 0.4048028048662315, "grad_norm": 13.586715698242188, "learning_rate": 1.3514343542601711e-05, "loss": 3.9505, "step": 11488 }, { "epoch": 0.4048380418791899, "grad_norm": 5.038302421569824, "learning_rate": 1.3513275042078244e-05, "loss": 5.4256, "step": 11489 }, { "epoch": 0.4048732788921483, "grad_norm": 5.147281646728516, "learning_rate": 1.3512206495793918e-05, "loss": 2.8187, "step": 11490 }, { "epoch": 0.40490851590510674, "grad_norm": 6.04327917098999, "learning_rate": 1.3511137903762657e-05, "loss": 3.3612, "step": 11491 }, { "epoch": 0.4049437529180651, "grad_norm": 3.4127047061920166, "learning_rate": 1.3510069265998374e-05, "loss": 2.6697, "step": 11492 }, { "epoch": 0.40497898993102355, "grad_norm": 1.7813801765441895, "learning_rate": 1.3509000582514996e-05, "loss": 0.8376, "step": 11493 }, { "epoch": 0.405014226943982, "grad_norm": 7.788307189941406, "learning_rate": 1.3507931853326434e-05, "loss": 5.9942, "step": 11494 }, { "epoch": 0.40504946395694036, "grad_norm": 3.1586339473724365, "learning_rate": 1.3506863078446615e-05, "loss": 3.0644, "step": 11495 }, { "epoch": 0.4050847009698988, "grad_norm": 1.718416452407837, "learning_rate": 1.3505794257889463e-05, "loss": 0.6924, "step": 11496 }, { "epoch": 0.40511993798285717, "grad_norm": 1.9818110466003418, "learning_rate": 1.3504725391668885e-05, "loss": 0.8656, "step": 11497 }, { "epoch": 0.4051551749958156, "grad_norm": 1.6719363927841187, "learning_rate": 1.3503656479798818e-05, "loss": 1.0055, "step": 11498 }, { "epoch": 0.40519041200877404, "grad_norm": 6.778034210205078, "learning_rate": 1.3502587522293176e-05, "loss": 3.6371, "step": 11499 }, { "epoch": 0.4052256490217324, "grad_norm": 1.5358527898788452, "learning_rate": 1.3501518519165887e-05, "loss": 0.8233, "step": 11500 }, { "epoch": 0.40526088603469085, "grad_norm": 6.374666690826416, "learning_rate": 1.3500449470430873e-05, "loss": 3.0976, "step": 11501 }, { "epoch": 0.4052961230476492, "grad_norm": 16.929672241210938, "learning_rate": 1.349938037610206e-05, "loss": 3.624, "step": 11502 }, { "epoch": 0.40533136006060766, "grad_norm": 5.72470760345459, "learning_rate": 1.3498311236193371e-05, "loss": 3.4858, "step": 11503 }, { "epoch": 0.4053665970735661, "grad_norm": 4.845181941986084, "learning_rate": 1.3497242050718735e-05, "loss": 3.1108, "step": 11504 }, { "epoch": 0.40540183408652447, "grad_norm": 6.338649272918701, "learning_rate": 1.349617281969207e-05, "loss": 4.6743, "step": 11505 }, { "epoch": 0.4054370710994829, "grad_norm": 6.282866954803467, "learning_rate": 1.3495103543127315e-05, "loss": 2.802, "step": 11506 }, { "epoch": 0.4054723081124413, "grad_norm": 7.4464802742004395, "learning_rate": 1.3494034221038385e-05, "loss": 6.0104, "step": 11507 }, { "epoch": 0.4055075451253997, "grad_norm": 1.9399584531784058, "learning_rate": 1.349296485343922e-05, "loss": 0.7264, "step": 11508 }, { "epoch": 0.40554278213835815, "grad_norm": 7.733633995056152, "learning_rate": 1.3491895440343739e-05, "loss": 5.7264, "step": 11509 }, { "epoch": 0.4055780191513165, "grad_norm": 9.81247329711914, "learning_rate": 1.3490825981765877e-05, "loss": 5.3921, "step": 11510 }, { "epoch": 0.40561325616427496, "grad_norm": 1.178041934967041, "learning_rate": 1.348975647771956e-05, "loss": 1.0642, "step": 11511 }, { "epoch": 0.4056484931772334, "grad_norm": 1.3296444416046143, "learning_rate": 1.3488686928218721e-05, "loss": 0.7321, "step": 11512 }, { "epoch": 0.40568373019019177, "grad_norm": 10.963807106018066, "learning_rate": 1.3487617333277293e-05, "loss": 3.4023, "step": 11513 }, { "epoch": 0.4057189672031502, "grad_norm": 16.917869567871094, "learning_rate": 1.3486547692909203e-05, "loss": 5.286, "step": 11514 }, { "epoch": 0.4057542042161086, "grad_norm": 5.735619068145752, "learning_rate": 1.3485478007128383e-05, "loss": 3.8675, "step": 11515 }, { "epoch": 0.405789441229067, "grad_norm": 3.8554067611694336, "learning_rate": 1.3484408275948768e-05, "loss": 3.4418, "step": 11516 }, { "epoch": 0.40582467824202545, "grad_norm": 2.4803435802459717, "learning_rate": 1.3483338499384297e-05, "loss": 0.8929, "step": 11517 }, { "epoch": 0.4058599152549838, "grad_norm": 5.0242390632629395, "learning_rate": 1.3482268677448893e-05, "loss": 2.5545, "step": 11518 }, { "epoch": 0.40589515226794226, "grad_norm": 1.2896803617477417, "learning_rate": 1.3481198810156499e-05, "loss": 0.9194, "step": 11519 }, { "epoch": 0.40593038928090064, "grad_norm": 3.3423736095428467, "learning_rate": 1.3480128897521046e-05, "loss": 3.1536, "step": 11520 }, { "epoch": 0.40596562629385907, "grad_norm": 1.8656156063079834, "learning_rate": 1.3479058939556468e-05, "loss": 0.9114, "step": 11521 }, { "epoch": 0.4060008633068175, "grad_norm": 2.156771421432495, "learning_rate": 1.3477988936276707e-05, "loss": 0.7186, "step": 11522 }, { "epoch": 0.4060361003197759, "grad_norm": 4.799168109893799, "learning_rate": 1.34769188876957e-05, "loss": 3.3243, "step": 11523 }, { "epoch": 0.4060713373327343, "grad_norm": 3.625990629196167, "learning_rate": 1.3475848793827376e-05, "loss": 3.3511, "step": 11524 }, { "epoch": 0.4061065743456927, "grad_norm": 7.84473180770874, "learning_rate": 1.3474778654685683e-05, "loss": 5.8013, "step": 11525 }, { "epoch": 0.40614181135865113, "grad_norm": 7.997030258178711, "learning_rate": 1.347370847028455e-05, "loss": 5.4587, "step": 11526 }, { "epoch": 0.40617704837160956, "grad_norm": 1.2366598844528198, "learning_rate": 1.3472638240637928e-05, "loss": 0.6969, "step": 11527 }, { "epoch": 0.40621228538456794, "grad_norm": 1.4261868000030518, "learning_rate": 1.3471567965759746e-05, "loss": 0.6985, "step": 11528 }, { "epoch": 0.4062475223975264, "grad_norm": 1.7917720079421997, "learning_rate": 1.3470497645663952e-05, "loss": 1.0889, "step": 11529 }, { "epoch": 0.40628275941048475, "grad_norm": 12.367295265197754, "learning_rate": 1.3469427280364483e-05, "loss": 5.4832, "step": 11530 }, { "epoch": 0.4063179964234432, "grad_norm": 1.7399612665176392, "learning_rate": 1.346835686987528e-05, "loss": 1.082, "step": 11531 }, { "epoch": 0.4063532334364016, "grad_norm": 1.1860030889511108, "learning_rate": 1.346728641421029e-05, "loss": 0.9093, "step": 11532 }, { "epoch": 0.40638847044936, "grad_norm": 2.6283226013183594, "learning_rate": 1.346621591338345e-05, "loss": 2.8095, "step": 11533 }, { "epoch": 0.40642370746231843, "grad_norm": 5.52773904800415, "learning_rate": 1.3465145367408707e-05, "loss": 5.4182, "step": 11534 }, { "epoch": 0.4064589444752768, "grad_norm": 5.234413146972656, "learning_rate": 1.3464074776300005e-05, "loss": 3.1233, "step": 11535 }, { "epoch": 0.40649418148823524, "grad_norm": 2.578160285949707, "learning_rate": 1.346300414007129e-05, "loss": 0.9849, "step": 11536 }, { "epoch": 0.4065294185011937, "grad_norm": 1.7571183443069458, "learning_rate": 1.3461933458736501e-05, "loss": 0.9831, "step": 11537 }, { "epoch": 0.40656465551415205, "grad_norm": 1.3121907711029053, "learning_rate": 1.346086273230959e-05, "loss": 0.7514, "step": 11538 }, { "epoch": 0.4065998925271105, "grad_norm": 1.2619836330413818, "learning_rate": 1.34597919608045e-05, "loss": 0.8078, "step": 11539 }, { "epoch": 0.40663512954006886, "grad_norm": 4.608587265014648, "learning_rate": 1.3458721144235179e-05, "loss": 3.3529, "step": 11540 }, { "epoch": 0.4066703665530273, "grad_norm": 3.310624837875366, "learning_rate": 1.3457650282615575e-05, "loss": 3.1005, "step": 11541 }, { "epoch": 0.40670560356598573, "grad_norm": 1.5846024751663208, "learning_rate": 1.3456579375959638e-05, "loss": 1.0168, "step": 11542 }, { "epoch": 0.4067408405789441, "grad_norm": 12.126847267150879, "learning_rate": 1.345550842428131e-05, "loss": 5.6656, "step": 11543 }, { "epoch": 0.40677607759190254, "grad_norm": 1.7339280843734741, "learning_rate": 1.3454437427594547e-05, "loss": 0.949, "step": 11544 }, { "epoch": 0.4068113146048609, "grad_norm": 4.828631401062012, "learning_rate": 1.3453366385913294e-05, "loss": 3.0667, "step": 11545 }, { "epoch": 0.40684655161781935, "grad_norm": 1.2065783739089966, "learning_rate": 1.3452295299251507e-05, "loss": 0.942, "step": 11546 }, { "epoch": 0.4068817886307778, "grad_norm": 4.416387557983398, "learning_rate": 1.3451224167623134e-05, "loss": 3.4246, "step": 11547 }, { "epoch": 0.40691702564373616, "grad_norm": 1.3028631210327148, "learning_rate": 1.3450152991042125e-05, "loss": 1.0087, "step": 11548 }, { "epoch": 0.4069522626566946, "grad_norm": 1.968718409538269, "learning_rate": 1.3449081769522434e-05, "loss": 0.8371, "step": 11549 }, { "epoch": 0.40698749966965303, "grad_norm": 0.9478809833526611, "learning_rate": 1.3448010503078012e-05, "loss": 0.8957, "step": 11550 }, { "epoch": 0.4070227366826114, "grad_norm": 8.653523445129395, "learning_rate": 1.3446939191722816e-05, "loss": 3.0451, "step": 11551 }, { "epoch": 0.40705797369556984, "grad_norm": 1.9059447050094604, "learning_rate": 1.3445867835470798e-05, "loss": 0.8754, "step": 11552 }, { "epoch": 0.4070932107085282, "grad_norm": 1.4966340065002441, "learning_rate": 1.3444796434335911e-05, "loss": 1.156, "step": 11553 }, { "epoch": 0.40712844772148665, "grad_norm": 4.08684778213501, "learning_rate": 1.3443724988332114e-05, "loss": 2.4367, "step": 11554 }, { "epoch": 0.4071636847344451, "grad_norm": 9.764769554138184, "learning_rate": 1.3442653497473362e-05, "loss": 8.2472, "step": 11555 }, { "epoch": 0.40719892174740346, "grad_norm": 1.2748404741287231, "learning_rate": 1.3441581961773604e-05, "loss": 0.8912, "step": 11556 }, { "epoch": 0.4072341587603619, "grad_norm": 2.31305193901062, "learning_rate": 1.3440510381246807e-05, "loss": 1.1438, "step": 11557 }, { "epoch": 0.4072693957733203, "grad_norm": 6.6080756187438965, "learning_rate": 1.3439438755906921e-05, "loss": 2.9912, "step": 11558 }, { "epoch": 0.4073046327862787, "grad_norm": 4.36763334274292, "learning_rate": 1.343836708576791e-05, "loss": 3.1579, "step": 11559 }, { "epoch": 0.40733986979923714, "grad_norm": 10.431256294250488, "learning_rate": 1.3437295370843729e-05, "loss": 5.5139, "step": 11560 }, { "epoch": 0.4073751068121955, "grad_norm": 1.3640142679214478, "learning_rate": 1.3436223611148335e-05, "loss": 0.9323, "step": 11561 }, { "epoch": 0.40741034382515395, "grad_norm": 4.157665729522705, "learning_rate": 1.3435151806695694e-05, "loss": 2.6857, "step": 11562 }, { "epoch": 0.40744558083811233, "grad_norm": 11.967857360839844, "learning_rate": 1.3434079957499764e-05, "loss": 5.677, "step": 11563 }, { "epoch": 0.40748081785107076, "grad_norm": 2.0068585872650146, "learning_rate": 1.34330080635745e-05, "loss": 1.0834, "step": 11564 }, { "epoch": 0.4075160548640292, "grad_norm": 15.937055587768555, "learning_rate": 1.3431936124933876e-05, "loss": 4.8588, "step": 11565 }, { "epoch": 0.4075512918769876, "grad_norm": 2.212301731109619, "learning_rate": 1.3430864141591842e-05, "loss": 0.8107, "step": 11566 }, { "epoch": 0.407586528889946, "grad_norm": 3.842705726623535, "learning_rate": 1.3429792113562366e-05, "loss": 3.4107, "step": 11567 }, { "epoch": 0.4076217659029044, "grad_norm": 3.7977535724639893, "learning_rate": 1.3428720040859411e-05, "loss": 3.2164, "step": 11568 }, { "epoch": 0.4076570029158628, "grad_norm": 8.226066589355469, "learning_rate": 1.3427647923496941e-05, "loss": 5.568, "step": 11569 }, { "epoch": 0.40769223992882125, "grad_norm": 3.806654453277588, "learning_rate": 1.3426575761488918e-05, "loss": 2.8121, "step": 11570 }, { "epoch": 0.40772747694177963, "grad_norm": 5.605442523956299, "learning_rate": 1.3425503554849314e-05, "loss": 3.2747, "step": 11571 }, { "epoch": 0.40776271395473807, "grad_norm": 1.4544247388839722, "learning_rate": 1.3424431303592085e-05, "loss": 0.9298, "step": 11572 }, { "epoch": 0.40779795096769644, "grad_norm": 6.4604997634887695, "learning_rate": 1.3423359007731206e-05, "loss": 3.5203, "step": 11573 }, { "epoch": 0.4078331879806549, "grad_norm": 5.158442497253418, "learning_rate": 1.342228666728064e-05, "loss": 3.4677, "step": 11574 }, { "epoch": 0.4078684249936133, "grad_norm": 1.3254956007003784, "learning_rate": 1.3421214282254349e-05, "loss": 0.7628, "step": 11575 }, { "epoch": 0.4079036620065717, "grad_norm": 5.52496337890625, "learning_rate": 1.342014185266631e-05, "loss": 3.4402, "step": 11576 }, { "epoch": 0.4079388990195301, "grad_norm": 5.041718006134033, "learning_rate": 1.3419069378530488e-05, "loss": 3.0642, "step": 11577 }, { "epoch": 0.4079741360324885, "grad_norm": 7.231421947479248, "learning_rate": 1.3417996859860851e-05, "loss": 3.9682, "step": 11578 }, { "epoch": 0.40800937304544693, "grad_norm": 11.570426940917969, "learning_rate": 1.341692429667137e-05, "loss": 3.4133, "step": 11579 }, { "epoch": 0.40804461005840537, "grad_norm": 1.992856502532959, "learning_rate": 1.3415851688976014e-05, "loss": 0.8042, "step": 11580 }, { "epoch": 0.40807984707136374, "grad_norm": 3.7427148818969727, "learning_rate": 1.3414779036788755e-05, "loss": 3.4967, "step": 11581 }, { "epoch": 0.4081150840843222, "grad_norm": 0.9941467642784119, "learning_rate": 1.3413706340123564e-05, "loss": 1.2231, "step": 11582 }, { "epoch": 0.4081503210972806, "grad_norm": 0.9384065866470337, "learning_rate": 1.3412633598994412e-05, "loss": 0.8992, "step": 11583 }, { "epoch": 0.408185558110239, "grad_norm": 1.251603126525879, "learning_rate": 1.3411560813415276e-05, "loss": 0.9245, "step": 11584 }, { "epoch": 0.4082207951231974, "grad_norm": 8.657963752746582, "learning_rate": 1.3410487983400123e-05, "loss": 3.3922, "step": 11585 }, { "epoch": 0.4082560321361558, "grad_norm": 6.224234104156494, "learning_rate": 1.3409415108962932e-05, "loss": 3.6548, "step": 11586 }, { "epoch": 0.40829126914911423, "grad_norm": 1.984830617904663, "learning_rate": 1.3408342190117671e-05, "loss": 0.991, "step": 11587 }, { "epoch": 0.40832650616207267, "grad_norm": 2.1345598697662354, "learning_rate": 1.3407269226878322e-05, "loss": 1.2632, "step": 11588 }, { "epoch": 0.40836174317503104, "grad_norm": 6.182125568389893, "learning_rate": 1.3406196219258855e-05, "loss": 3.4159, "step": 11589 }, { "epoch": 0.4083969801879895, "grad_norm": 7.139412879943848, "learning_rate": 1.3405123167273248e-05, "loss": 4.9616, "step": 11590 }, { "epoch": 0.40843221720094786, "grad_norm": 7.975956916809082, "learning_rate": 1.3404050070935481e-05, "loss": 3.0506, "step": 11591 }, { "epoch": 0.4084674542139063, "grad_norm": 7.809823989868164, "learning_rate": 1.3402976930259528e-05, "loss": 6.5744, "step": 11592 }, { "epoch": 0.4085026912268647, "grad_norm": 5.073158264160156, "learning_rate": 1.3401903745259366e-05, "loss": 3.0948, "step": 11593 }, { "epoch": 0.4085379282398231, "grad_norm": 1.4985666275024414, "learning_rate": 1.3400830515948972e-05, "loss": 1.0107, "step": 11594 }, { "epoch": 0.40857316525278153, "grad_norm": 2.892247438430786, "learning_rate": 1.3399757242342332e-05, "loss": 3.6324, "step": 11595 }, { "epoch": 0.4086084022657399, "grad_norm": 3.3094749450683594, "learning_rate": 1.3398683924453417e-05, "loss": 2.8973, "step": 11596 }, { "epoch": 0.40864363927869835, "grad_norm": 3.620546817779541, "learning_rate": 1.3397610562296213e-05, "loss": 3.4047, "step": 11597 }, { "epoch": 0.4086788762916568, "grad_norm": 4.417527198791504, "learning_rate": 1.3396537155884699e-05, "loss": 2.9966, "step": 11598 }, { "epoch": 0.40871411330461516, "grad_norm": 6.32259464263916, "learning_rate": 1.3395463705232853e-05, "loss": 3.3729, "step": 11599 }, { "epoch": 0.4087493503175736, "grad_norm": 1.3086683750152588, "learning_rate": 1.3394390210354662e-05, "loss": 0.7157, "step": 11600 }, { "epoch": 0.40878458733053197, "grad_norm": 9.02034854888916, "learning_rate": 1.3393316671264107e-05, "loss": 6.9822, "step": 11601 }, { "epoch": 0.4088198243434904, "grad_norm": 7.629671573638916, "learning_rate": 1.3392243087975167e-05, "loss": 2.6355, "step": 11602 }, { "epoch": 0.40885506135644883, "grad_norm": 1.8409602642059326, "learning_rate": 1.3391169460501834e-05, "loss": 0.7207, "step": 11603 }, { "epoch": 0.4088902983694072, "grad_norm": 10.155949592590332, "learning_rate": 1.3390095788858081e-05, "loss": 3.4326, "step": 11604 }, { "epoch": 0.40892553538236565, "grad_norm": 20.77550506591797, "learning_rate": 1.3389022073057903e-05, "loss": 5.5996, "step": 11605 }, { "epoch": 0.408960772395324, "grad_norm": 3.7552857398986816, "learning_rate": 1.3387948313115277e-05, "loss": 3.1716, "step": 11606 }, { "epoch": 0.40899600940828246, "grad_norm": 12.886919021606445, "learning_rate": 1.3386874509044195e-05, "loss": 3.6423, "step": 11607 }, { "epoch": 0.4090312464212409, "grad_norm": 5.574408531188965, "learning_rate": 1.338580066085864e-05, "loss": 2.8377, "step": 11608 }, { "epoch": 0.40906648343419927, "grad_norm": 1.3102293014526367, "learning_rate": 1.33847267685726e-05, "loss": 1.0439, "step": 11609 }, { "epoch": 0.4091017204471577, "grad_norm": 1.3498977422714233, "learning_rate": 1.3383652832200064e-05, "loss": 0.8495, "step": 11610 }, { "epoch": 0.4091369574601161, "grad_norm": 5.088773727416992, "learning_rate": 1.338257885175502e-05, "loss": 3.028, "step": 11611 }, { "epoch": 0.4091721944730745, "grad_norm": 1.1723690032958984, "learning_rate": 1.3381504827251455e-05, "loss": 1.0028, "step": 11612 }, { "epoch": 0.40920743148603295, "grad_norm": 10.994285583496094, "learning_rate": 1.3380430758703354e-05, "loss": 7.3685, "step": 11613 }, { "epoch": 0.4092426684989913, "grad_norm": 4.406786918640137, "learning_rate": 1.3379356646124717e-05, "loss": 3.2887, "step": 11614 }, { "epoch": 0.40927790551194976, "grad_norm": 4.674252033233643, "learning_rate": 1.3378282489529528e-05, "loss": 2.9526, "step": 11615 }, { "epoch": 0.4093131425249082, "grad_norm": 1.129179835319519, "learning_rate": 1.3377208288931778e-05, "loss": 0.8101, "step": 11616 }, { "epoch": 0.40934837953786657, "grad_norm": 7.609414577484131, "learning_rate": 1.3376134044345462e-05, "loss": 3.4456, "step": 11617 }, { "epoch": 0.409383616550825, "grad_norm": 4.139634132385254, "learning_rate": 1.3375059755784568e-05, "loss": 3.4465, "step": 11618 }, { "epoch": 0.4094188535637834, "grad_norm": 5.0126633644104, "learning_rate": 1.3373985423263092e-05, "loss": 5.0312, "step": 11619 }, { "epoch": 0.4094540905767418, "grad_norm": 1.7026550769805908, "learning_rate": 1.3372911046795027e-05, "loss": 0.8701, "step": 11620 }, { "epoch": 0.40948932758970025, "grad_norm": 6.454401969909668, "learning_rate": 1.3371836626394363e-05, "loss": 3.368, "step": 11621 }, { "epoch": 0.4095245646026586, "grad_norm": 10.048056602478027, "learning_rate": 1.33707621620751e-05, "loss": 3.1989, "step": 11622 }, { "epoch": 0.40955980161561706, "grad_norm": 1.1847394704818726, "learning_rate": 1.3369687653851228e-05, "loss": 0.9291, "step": 11623 }, { "epoch": 0.40959503862857544, "grad_norm": 1.0823073387145996, "learning_rate": 1.3368613101736748e-05, "loss": 0.9353, "step": 11624 }, { "epoch": 0.40963027564153387, "grad_norm": 5.276371479034424, "learning_rate": 1.3367538505745652e-05, "loss": 3.4693, "step": 11625 }, { "epoch": 0.4096655126544923, "grad_norm": 8.560725212097168, "learning_rate": 1.3366463865891936e-05, "loss": 5.9418, "step": 11626 }, { "epoch": 0.4097007496674507, "grad_norm": 12.752039909362793, "learning_rate": 1.3365389182189601e-05, "loss": 8.0611, "step": 11627 }, { "epoch": 0.4097359866804091, "grad_norm": 1.1693572998046875, "learning_rate": 1.3364314454652648e-05, "loss": 0.9524, "step": 11628 }, { "epoch": 0.4097712236933675, "grad_norm": 2.3729684352874756, "learning_rate": 1.3363239683295063e-05, "loss": 0.8482, "step": 11629 }, { "epoch": 0.4098064607063259, "grad_norm": 1.4777276515960693, "learning_rate": 1.3362164868130858e-05, "loss": 0.8859, "step": 11630 }, { "epoch": 0.40984169771928436, "grad_norm": 3.9819064140319824, "learning_rate": 1.3361090009174023e-05, "loss": 3.3068, "step": 11631 }, { "epoch": 0.40987693473224274, "grad_norm": 6.281919956207275, "learning_rate": 1.3360015106438567e-05, "loss": 4.2568, "step": 11632 }, { "epoch": 0.40991217174520117, "grad_norm": 1.3738261461257935, "learning_rate": 1.3358940159938483e-05, "loss": 1.0073, "step": 11633 }, { "epoch": 0.40994740875815955, "grad_norm": 9.160006523132324, "learning_rate": 1.3357865169687774e-05, "loss": 5.5253, "step": 11634 }, { "epoch": 0.409982645771118, "grad_norm": 5.586169242858887, "learning_rate": 1.3356790135700446e-05, "loss": 3.2357, "step": 11635 }, { "epoch": 0.4100178827840764, "grad_norm": 1.5624094009399414, "learning_rate": 1.3355715057990499e-05, "loss": 1.2409, "step": 11636 }, { "epoch": 0.4100531197970348, "grad_norm": 4.350578784942627, "learning_rate": 1.3354639936571935e-05, "loss": 3.2897, "step": 11637 }, { "epoch": 0.4100883568099932, "grad_norm": 8.389632225036621, "learning_rate": 1.335356477145876e-05, "loss": 5.308, "step": 11638 }, { "epoch": 0.4101235938229516, "grad_norm": 1.4526984691619873, "learning_rate": 1.3352489562664976e-05, "loss": 0.9678, "step": 11639 }, { "epoch": 0.41015883083591004, "grad_norm": 5.924980163574219, "learning_rate": 1.3351414310204587e-05, "loss": 3.498, "step": 11640 }, { "epoch": 0.41019406784886847, "grad_norm": 4.1273040771484375, "learning_rate": 1.3350339014091604e-05, "loss": 3.6462, "step": 11641 }, { "epoch": 0.41022930486182685, "grad_norm": 5.974648475646973, "learning_rate": 1.3349263674340026e-05, "loss": 3.5084, "step": 11642 }, { "epoch": 0.4102645418747853, "grad_norm": 1.553336262702942, "learning_rate": 1.334818829096386e-05, "loss": 0.8837, "step": 11643 }, { "epoch": 0.41029977888774366, "grad_norm": 1.327757477760315, "learning_rate": 1.3347112863977117e-05, "loss": 1.0807, "step": 11644 }, { "epoch": 0.4103350159007021, "grad_norm": 1.2529181241989136, "learning_rate": 1.33460373933938e-05, "loss": 0.9947, "step": 11645 }, { "epoch": 0.4103702529136605, "grad_norm": 4.7129974365234375, "learning_rate": 1.3344961879227923e-05, "loss": 2.7784, "step": 11646 }, { "epoch": 0.4104054899266189, "grad_norm": 1.373541235923767, "learning_rate": 1.334388632149349e-05, "loss": 1.1881, "step": 11647 }, { "epoch": 0.41044072693957734, "grad_norm": 15.041089057922363, "learning_rate": 1.3342810720204513e-05, "loss": 7.9117, "step": 11648 }, { "epoch": 0.4104759639525357, "grad_norm": 1.527803897857666, "learning_rate": 1.3341735075375e-05, "loss": 1.2394, "step": 11649 }, { "epoch": 0.41051120096549415, "grad_norm": 5.1357741355896, "learning_rate": 1.3340659387018961e-05, "loss": 5.6285, "step": 11650 }, { "epoch": 0.4105464379784526, "grad_norm": 1.5601305961608887, "learning_rate": 1.333958365515041e-05, "loss": 0.7853, "step": 11651 }, { "epoch": 0.41058167499141096, "grad_norm": 2.05794095993042, "learning_rate": 1.3338507879783353e-05, "loss": 0.9529, "step": 11652 }, { "epoch": 0.4106169120043694, "grad_norm": 1.7140175104141235, "learning_rate": 1.333743206093181e-05, "loss": 0.8658, "step": 11653 }, { "epoch": 0.4106521490173278, "grad_norm": 2.499315023422241, "learning_rate": 1.3336356198609786e-05, "loss": 2.5569, "step": 11654 }, { "epoch": 0.4106873860302862, "grad_norm": 1.0492079257965088, "learning_rate": 1.3335280292831297e-05, "loss": 0.8841, "step": 11655 }, { "epoch": 0.41072262304324464, "grad_norm": 1.8945941925048828, "learning_rate": 1.333420434361036e-05, "loss": 0.9478, "step": 11656 }, { "epoch": 0.410757860056203, "grad_norm": 1.4067113399505615, "learning_rate": 1.3333128350960984e-05, "loss": 0.9359, "step": 11657 }, { "epoch": 0.41079309706916145, "grad_norm": 1.4703681468963623, "learning_rate": 1.3332052314897191e-05, "loss": 0.9197, "step": 11658 }, { "epoch": 0.4108283340821199, "grad_norm": 1.5691789388656616, "learning_rate": 1.3330976235432987e-05, "loss": 0.8314, "step": 11659 }, { "epoch": 0.41086357109507826, "grad_norm": 1.641929268836975, "learning_rate": 1.3329900112582397e-05, "loss": 0.6982, "step": 11660 }, { "epoch": 0.4108988081080367, "grad_norm": 3.4939327239990234, "learning_rate": 1.332882394635943e-05, "loss": 3.7823, "step": 11661 }, { "epoch": 0.4109340451209951, "grad_norm": 1.3003032207489014, "learning_rate": 1.3327747736778112e-05, "loss": 1.1246, "step": 11662 }, { "epoch": 0.4109692821339535, "grad_norm": 1.9671761989593506, "learning_rate": 1.3326671483852453e-05, "loss": 0.9008, "step": 11663 }, { "epoch": 0.41100451914691194, "grad_norm": 1.6488233804702759, "learning_rate": 1.3325595187596474e-05, "loss": 0.8611, "step": 11664 }, { "epoch": 0.4110397561598703, "grad_norm": 4.788035869598389, "learning_rate": 1.3324518848024194e-05, "loss": 3.1516, "step": 11665 }, { "epoch": 0.41107499317282875, "grad_norm": 1.3198370933532715, "learning_rate": 1.3323442465149636e-05, "loss": 0.9045, "step": 11666 }, { "epoch": 0.41111023018578713, "grad_norm": 4.941681861877441, "learning_rate": 1.332236603898681e-05, "loss": 4.4866, "step": 11667 }, { "epoch": 0.41114546719874556, "grad_norm": 6.207740306854248, "learning_rate": 1.3321289569549749e-05, "loss": 5.7647, "step": 11668 }, { "epoch": 0.411180704211704, "grad_norm": 0.9504880905151367, "learning_rate": 1.3320213056852467e-05, "loss": 0.9274, "step": 11669 }, { "epoch": 0.4112159412246624, "grad_norm": 11.174046516418457, "learning_rate": 1.3319136500908987e-05, "loss": 3.4098, "step": 11670 }, { "epoch": 0.4112511782376208, "grad_norm": 1.6608425378799438, "learning_rate": 1.3318059901733329e-05, "loss": 0.8816, "step": 11671 }, { "epoch": 0.4112864152505792, "grad_norm": 6.866544246673584, "learning_rate": 1.331698325933952e-05, "loss": 3.3084, "step": 11672 }, { "epoch": 0.4113216522635376, "grad_norm": 6.846064567565918, "learning_rate": 1.3315906573741582e-05, "loss": 2.7002, "step": 11673 }, { "epoch": 0.41135688927649605, "grad_norm": 6.456997394561768, "learning_rate": 1.3314829844953537e-05, "loss": 2.9356, "step": 11674 }, { "epoch": 0.41139212628945443, "grad_norm": 6.629362106323242, "learning_rate": 1.3313753072989411e-05, "loss": 3.3556, "step": 11675 }, { "epoch": 0.41142736330241286, "grad_norm": 1.2575331926345825, "learning_rate": 1.331267625786323e-05, "loss": 1.2321, "step": 11676 }, { "epoch": 0.41146260031537124, "grad_norm": 6.044445514678955, "learning_rate": 1.331159939958902e-05, "loss": 5.1223, "step": 11677 }, { "epoch": 0.4114978373283297, "grad_norm": 4.6499433517456055, "learning_rate": 1.3310522498180803e-05, "loss": 3.1841, "step": 11678 }, { "epoch": 0.4115330743412881, "grad_norm": 0.9660463333129883, "learning_rate": 1.3309445553652611e-05, "loss": 0.9411, "step": 11679 }, { "epoch": 0.4115683113542465, "grad_norm": 14.498889923095703, "learning_rate": 1.3308368566018466e-05, "loss": 3.1726, "step": 11680 }, { "epoch": 0.4116035483672049, "grad_norm": 6.2543745040893555, "learning_rate": 1.3307291535292404e-05, "loss": 3.3371, "step": 11681 }, { "epoch": 0.4116387853801633, "grad_norm": 2.6796109676361084, "learning_rate": 1.3306214461488445e-05, "loss": 0.884, "step": 11682 }, { "epoch": 0.41167402239312173, "grad_norm": 3.3346614837646484, "learning_rate": 1.3305137344620622e-05, "loss": 2.707, "step": 11683 }, { "epoch": 0.41170925940608016, "grad_norm": 4.18942403793335, "learning_rate": 1.3304060184702963e-05, "loss": 3.3236, "step": 11684 }, { "epoch": 0.41174449641903854, "grad_norm": 6.995692729949951, "learning_rate": 1.3302982981749504e-05, "loss": 3.2684, "step": 11685 }, { "epoch": 0.411779733431997, "grad_norm": 3.83048939704895, "learning_rate": 1.3301905735774266e-05, "loss": 3.3535, "step": 11686 }, { "epoch": 0.4118149704449554, "grad_norm": 1.0093427896499634, "learning_rate": 1.3300828446791287e-05, "loss": 0.8152, "step": 11687 }, { "epoch": 0.4118502074579138, "grad_norm": 1.0091960430145264, "learning_rate": 1.3299751114814595e-05, "loss": 0.815, "step": 11688 }, { "epoch": 0.4118854444708722, "grad_norm": 1.5432510375976562, "learning_rate": 1.3298673739858227e-05, "loss": 0.8953, "step": 11689 }, { "epoch": 0.4119206814838306, "grad_norm": 9.741430282592773, "learning_rate": 1.3297596321936212e-05, "loss": 3.6653, "step": 11690 }, { "epoch": 0.41195591849678903, "grad_norm": 13.72106647491455, "learning_rate": 1.3296518861062586e-05, "loss": 3.2287, "step": 11691 }, { "epoch": 0.41199115550974746, "grad_norm": 1.4326550960540771, "learning_rate": 1.3295441357251382e-05, "loss": 1.0693, "step": 11692 }, { "epoch": 0.41202639252270584, "grad_norm": 8.317733764648438, "learning_rate": 1.3294363810516633e-05, "loss": 3.0606, "step": 11693 }, { "epoch": 0.4120616295356643, "grad_norm": 7.4906816482543945, "learning_rate": 1.3293286220872376e-05, "loss": 3.0072, "step": 11694 }, { "epoch": 0.41209686654862265, "grad_norm": 10.107312202453613, "learning_rate": 1.3292208588332647e-05, "loss": 6.0074, "step": 11695 }, { "epoch": 0.4121321035615811, "grad_norm": 5.158852577209473, "learning_rate": 1.3291130912911485e-05, "loss": 3.199, "step": 11696 }, { "epoch": 0.4121673405745395, "grad_norm": 7.059950828552246, "learning_rate": 1.3290053194622918e-05, "loss": 3.017, "step": 11697 }, { "epoch": 0.4122025775874979, "grad_norm": 14.712640762329102, "learning_rate": 1.3288975433480992e-05, "loss": 5.0485, "step": 11698 }, { "epoch": 0.41223781460045633, "grad_norm": 2.7827908992767334, "learning_rate": 1.328789762949974e-05, "loss": 2.6738, "step": 11699 }, { "epoch": 0.4122730516134147, "grad_norm": 2.540785551071167, "learning_rate": 1.3286819782693205e-05, "loss": 1.0933, "step": 11700 }, { "epoch": 0.41230828862637314, "grad_norm": 14.868088722229004, "learning_rate": 1.3285741893075423e-05, "loss": 3.4161, "step": 11701 }, { "epoch": 0.4123435256393316, "grad_norm": 2.0826046466827393, "learning_rate": 1.3284663960660434e-05, "loss": 1.0478, "step": 11702 }, { "epoch": 0.41237876265228995, "grad_norm": 9.200909614562988, "learning_rate": 1.328358598546228e-05, "loss": 5.2512, "step": 11703 }, { "epoch": 0.4124139996652484, "grad_norm": 4.386285781860352, "learning_rate": 1.3282507967495e-05, "loss": 3.2207, "step": 11704 }, { "epoch": 0.41244923667820677, "grad_norm": 3.638652801513672, "learning_rate": 1.3281429906772632e-05, "loss": 3.0834, "step": 11705 }, { "epoch": 0.4124844736911652, "grad_norm": 15.715863227844238, "learning_rate": 1.3280351803309227e-05, "loss": 5.3808, "step": 11706 }, { "epoch": 0.41251971070412363, "grad_norm": 6.944509506225586, "learning_rate": 1.3279273657118816e-05, "loss": 4.5496, "step": 11707 }, { "epoch": 0.412554947717082, "grad_norm": 1.6872118711471558, "learning_rate": 1.3278195468215452e-05, "loss": 0.973, "step": 11708 }, { "epoch": 0.41259018473004044, "grad_norm": 1.2391784191131592, "learning_rate": 1.3277117236613174e-05, "loss": 0.8382, "step": 11709 }, { "epoch": 0.4126254217429988, "grad_norm": 1.6169097423553467, "learning_rate": 1.3276038962326024e-05, "loss": 0.8458, "step": 11710 }, { "epoch": 0.41266065875595725, "grad_norm": 8.310009002685547, "learning_rate": 1.3274960645368051e-05, "loss": 4.0346, "step": 11711 }, { "epoch": 0.4126958957689157, "grad_norm": 7.648584842681885, "learning_rate": 1.32738822857533e-05, "loss": 3.5596, "step": 11712 }, { "epoch": 0.41273113278187407, "grad_norm": 4.692206382751465, "learning_rate": 1.3272803883495812e-05, "loss": 3.3815, "step": 11713 }, { "epoch": 0.4127663697948325, "grad_norm": 0.7935425639152527, "learning_rate": 1.3271725438609639e-05, "loss": 1.0156, "step": 11714 }, { "epoch": 0.4128016068077909, "grad_norm": 5.769989967346191, "learning_rate": 1.3270646951108826e-05, "loss": 3.4838, "step": 11715 }, { "epoch": 0.4128368438207493, "grad_norm": 6.499983787536621, "learning_rate": 1.3269568421007415e-05, "loss": 5.8177, "step": 11716 }, { "epoch": 0.41287208083370774, "grad_norm": 1.3193594217300415, "learning_rate": 1.3268489848319463e-05, "loss": 1.0425, "step": 11717 }, { "epoch": 0.4129073178466661, "grad_norm": 3.3051650524139404, "learning_rate": 1.3267411233059013e-05, "loss": 3.3781, "step": 11718 }, { "epoch": 0.41294255485962456, "grad_norm": 3.8189663887023926, "learning_rate": 1.3266332575240117e-05, "loss": 3.5335, "step": 11719 }, { "epoch": 0.412977791872583, "grad_norm": 13.465983390808105, "learning_rate": 1.3265253874876822e-05, "loss": 7.5143, "step": 11720 }, { "epoch": 0.41301302888554137, "grad_norm": 1.5263172388076782, "learning_rate": 1.326417513198318e-05, "loss": 0.7895, "step": 11721 }, { "epoch": 0.4130482658984998, "grad_norm": 3.1263368129730225, "learning_rate": 1.3263096346573242e-05, "loss": 2.4045, "step": 11722 }, { "epoch": 0.4130835029114582, "grad_norm": 1.789166808128357, "learning_rate": 1.326201751866106e-05, "loss": 0.8756, "step": 11723 }, { "epoch": 0.4131187399244166, "grad_norm": 6.701084136962891, "learning_rate": 1.3260938648260677e-05, "loss": 3.0659, "step": 11724 }, { "epoch": 0.41315397693737504, "grad_norm": 4.938865661621094, "learning_rate": 1.325985973538616e-05, "loss": 3.3666, "step": 11725 }, { "epoch": 0.4131892139503334, "grad_norm": 3.514828681945801, "learning_rate": 1.325878078005155e-05, "loss": 3.2361, "step": 11726 }, { "epoch": 0.41322445096329186, "grad_norm": 6.717344760894775, "learning_rate": 1.3257701782270908e-05, "loss": 3.4742, "step": 11727 }, { "epoch": 0.41325968797625023, "grad_norm": 7.791308403015137, "learning_rate": 1.3256622742058285e-05, "loss": 2.6765, "step": 11728 }, { "epoch": 0.41329492498920867, "grad_norm": 2.0192999839782715, "learning_rate": 1.3255543659427734e-05, "loss": 0.8063, "step": 11729 }, { "epoch": 0.4133301620021671, "grad_norm": 4.719716548919678, "learning_rate": 1.3254464534393315e-05, "loss": 3.5525, "step": 11730 }, { "epoch": 0.4133653990151255, "grad_norm": 0.9903169274330139, "learning_rate": 1.3253385366969077e-05, "loss": 1.1458, "step": 11731 }, { "epoch": 0.4134006360280839, "grad_norm": 1.3452320098876953, "learning_rate": 1.3252306157169084e-05, "loss": 1.1555, "step": 11732 }, { "epoch": 0.4134358730410423, "grad_norm": 5.223337173461914, "learning_rate": 1.3251226905007388e-05, "loss": 4.3541, "step": 11733 }, { "epoch": 0.4134711100540007, "grad_norm": 1.5240305662155151, "learning_rate": 1.3250147610498045e-05, "loss": 0.8309, "step": 11734 }, { "epoch": 0.41350634706695916, "grad_norm": 1.482489824295044, "learning_rate": 1.324906827365512e-05, "loss": 1.0298, "step": 11735 }, { "epoch": 0.41354158407991753, "grad_norm": 1.1144014596939087, "learning_rate": 1.3247988894492665e-05, "loss": 0.9577, "step": 11736 }, { "epoch": 0.41357682109287597, "grad_norm": 7.027919292449951, "learning_rate": 1.3246909473024738e-05, "loss": 3.2101, "step": 11737 }, { "epoch": 0.41361205810583435, "grad_norm": 9.309654235839844, "learning_rate": 1.3245830009265405e-05, "loss": 3.5208, "step": 11738 }, { "epoch": 0.4136472951187928, "grad_norm": 1.3770688772201538, "learning_rate": 1.3244750503228722e-05, "loss": 0.9358, "step": 11739 }, { "epoch": 0.4136825321317512, "grad_norm": 1.1666165590286255, "learning_rate": 1.3243670954928749e-05, "loss": 0.9471, "step": 11740 }, { "epoch": 0.4137177691447096, "grad_norm": 1.5994725227355957, "learning_rate": 1.3242591364379548e-05, "loss": 0.6648, "step": 11741 }, { "epoch": 0.413753006157668, "grad_norm": 5.445045471191406, "learning_rate": 1.3241511731595188e-05, "loss": 3.697, "step": 11742 }, { "epoch": 0.4137882431706264, "grad_norm": 4.469274044036865, "learning_rate": 1.3240432056589716e-05, "loss": 3.884, "step": 11743 }, { "epoch": 0.41382348018358484, "grad_norm": 1.3730655908584595, "learning_rate": 1.3239352339377209e-05, "loss": 1.009, "step": 11744 }, { "epoch": 0.41385871719654327, "grad_norm": 1.184205412864685, "learning_rate": 1.3238272579971723e-05, "loss": 0.8973, "step": 11745 }, { "epoch": 0.41389395420950165, "grad_norm": 6.622937202453613, "learning_rate": 1.3237192778387326e-05, "loss": 2.9324, "step": 11746 }, { "epoch": 0.4139291912224601, "grad_norm": 4.6559295654296875, "learning_rate": 1.3236112934638078e-05, "loss": 3.1677, "step": 11747 }, { "epoch": 0.41396442823541846, "grad_norm": 1.9701550006866455, "learning_rate": 1.3235033048738049e-05, "loss": 1.1262, "step": 11748 }, { "epoch": 0.4139996652483769, "grad_norm": 0.8358315229415894, "learning_rate": 1.3233953120701301e-05, "loss": 0.8074, "step": 11749 }, { "epoch": 0.4140349022613353, "grad_norm": 1.854514479637146, "learning_rate": 1.3232873150541903e-05, "loss": 0.9083, "step": 11750 }, { "epoch": 0.4140701392742937, "grad_norm": 4.634392738342285, "learning_rate": 1.3231793138273919e-05, "loss": 2.8611, "step": 11751 }, { "epoch": 0.41410537628725214, "grad_norm": 1.6125587224960327, "learning_rate": 1.3230713083911418e-05, "loss": 1.1805, "step": 11752 }, { "epoch": 0.4141406133002105, "grad_norm": 4.604494571685791, "learning_rate": 1.3229632987468468e-05, "loss": 3.3198, "step": 11753 }, { "epoch": 0.41417585031316895, "grad_norm": 7.579935073852539, "learning_rate": 1.3228552848959137e-05, "loss": 3.8783, "step": 11754 }, { "epoch": 0.4142110873261274, "grad_norm": 0.9789791703224182, "learning_rate": 1.3227472668397497e-05, "loss": 0.8663, "step": 11755 }, { "epoch": 0.41424632433908576, "grad_norm": 3.3447165489196777, "learning_rate": 1.3226392445797607e-05, "loss": 3.5083, "step": 11756 }, { "epoch": 0.4142815613520442, "grad_norm": 2.29390287399292, "learning_rate": 1.322531218117355e-05, "loss": 1.0034, "step": 11757 }, { "epoch": 0.4143167983650026, "grad_norm": 8.251449584960938, "learning_rate": 1.3224231874539391e-05, "loss": 3.8754, "step": 11758 }, { "epoch": 0.414352035377961, "grad_norm": 3.691159963607788, "learning_rate": 1.32231515259092e-05, "loss": 3.5264, "step": 11759 }, { "epoch": 0.41438727239091944, "grad_norm": 4.460087776184082, "learning_rate": 1.3222071135297048e-05, "loss": 3.1805, "step": 11760 }, { "epoch": 0.4144225094038778, "grad_norm": 3.38394832611084, "learning_rate": 1.3220990702717012e-05, "loss": 3.1197, "step": 11761 }, { "epoch": 0.41445774641683625, "grad_norm": 1.9382132291793823, "learning_rate": 1.3219910228183159e-05, "loss": 1.0032, "step": 11762 }, { "epoch": 0.4144929834297947, "grad_norm": 4.350151538848877, "learning_rate": 1.3218829711709568e-05, "loss": 3.5468, "step": 11763 }, { "epoch": 0.41452822044275306, "grad_norm": 8.849020004272461, "learning_rate": 1.3217749153310307e-05, "loss": 8.1158, "step": 11764 }, { "epoch": 0.4145634574557115, "grad_norm": 1.056890845298767, "learning_rate": 1.3216668552999457e-05, "loss": 0.9351, "step": 11765 }, { "epoch": 0.41459869446866987, "grad_norm": 5.268514633178711, "learning_rate": 1.3215587910791088e-05, "loss": 3.1212, "step": 11766 }, { "epoch": 0.4146339314816283, "grad_norm": 4.633944034576416, "learning_rate": 1.3214507226699276e-05, "loss": 3.5619, "step": 11767 }, { "epoch": 0.41466916849458674, "grad_norm": 4.988337993621826, "learning_rate": 1.3213426500738096e-05, "loss": 2.6319, "step": 11768 }, { "epoch": 0.4147044055075451, "grad_norm": 6.466914176940918, "learning_rate": 1.321234573292163e-05, "loss": 2.911, "step": 11769 }, { "epoch": 0.41473964252050355, "grad_norm": 1.08694326877594, "learning_rate": 1.3211264923263949e-05, "loss": 0.9765, "step": 11770 }, { "epoch": 0.4147748795334619, "grad_norm": 7.250988006591797, "learning_rate": 1.3210184071779134e-05, "loss": 6.7618, "step": 11771 }, { "epoch": 0.41481011654642036, "grad_norm": 1.073058009147644, "learning_rate": 1.3209103178481264e-05, "loss": 1.0428, "step": 11772 }, { "epoch": 0.4148453535593788, "grad_norm": 5.0687761306762695, "learning_rate": 1.3208022243384416e-05, "loss": 3.3637, "step": 11773 }, { "epoch": 0.41488059057233717, "grad_norm": 1.126613736152649, "learning_rate": 1.3206941266502668e-05, "loss": 0.7883, "step": 11774 }, { "epoch": 0.4149158275852956, "grad_norm": 1.8609870672225952, "learning_rate": 1.3205860247850105e-05, "loss": 0.8294, "step": 11775 }, { "epoch": 0.414951064598254, "grad_norm": 5.716135501861572, "learning_rate": 1.3204779187440799e-05, "loss": 2.7883, "step": 11776 }, { "epoch": 0.4149863016112124, "grad_norm": 3.8903870582580566, "learning_rate": 1.3203698085288838e-05, "loss": 3.1629, "step": 11777 }, { "epoch": 0.41502153862417085, "grad_norm": 5.128055095672607, "learning_rate": 1.3202616941408305e-05, "loss": 3.1871, "step": 11778 }, { "epoch": 0.4150567756371292, "grad_norm": 4.02897834777832, "learning_rate": 1.3201535755813274e-05, "loss": 3.5548, "step": 11779 }, { "epoch": 0.41509201265008766, "grad_norm": 6.8625078201293945, "learning_rate": 1.3200454528517831e-05, "loss": 5.043, "step": 11780 }, { "epoch": 0.41512724966304604, "grad_norm": 1.595898985862732, "learning_rate": 1.3199373259536064e-05, "loss": 0.8426, "step": 11781 }, { "epoch": 0.41516248667600447, "grad_norm": 4.080220699310303, "learning_rate": 1.3198291948882051e-05, "loss": 3.1141, "step": 11782 }, { "epoch": 0.4151977236889629, "grad_norm": 7.050160884857178, "learning_rate": 1.3197210596569876e-05, "loss": 5.8811, "step": 11783 }, { "epoch": 0.4152329607019213, "grad_norm": 1.3235325813293457, "learning_rate": 1.319612920261363e-05, "loss": 0.8924, "step": 11784 }, { "epoch": 0.4152681977148797, "grad_norm": 5.174333572387695, "learning_rate": 1.319504776702739e-05, "loss": 3.0063, "step": 11785 }, { "epoch": 0.4153034347278381, "grad_norm": 1.6253763437271118, "learning_rate": 1.3193966289825248e-05, "loss": 0.9816, "step": 11786 }, { "epoch": 0.41533867174079653, "grad_norm": 3.8274247646331787, "learning_rate": 1.3192884771021288e-05, "loss": 2.8568, "step": 11787 }, { "epoch": 0.41537390875375496, "grad_norm": 4.755245208740234, "learning_rate": 1.3191803210629597e-05, "loss": 2.8528, "step": 11788 }, { "epoch": 0.41540914576671334, "grad_norm": 1.2856285572052002, "learning_rate": 1.3190721608664264e-05, "loss": 0.8239, "step": 11789 }, { "epoch": 0.4154443827796718, "grad_norm": 104.63188934326172, "learning_rate": 1.3189639965139374e-05, "loss": 3.4083, "step": 11790 }, { "epoch": 0.4154796197926302, "grad_norm": 1.066835641860962, "learning_rate": 1.3188558280069014e-05, "loss": 0.8892, "step": 11791 }, { "epoch": 0.4155148568055886, "grad_norm": 1.4256904125213623, "learning_rate": 1.3187476553467282e-05, "loss": 1.0158, "step": 11792 }, { "epoch": 0.415550093818547, "grad_norm": 1.2137813568115234, "learning_rate": 1.318639478534826e-05, "loss": 0.791, "step": 11793 }, { "epoch": 0.4155853308315054, "grad_norm": 2.829152822494507, "learning_rate": 1.318531297572604e-05, "loss": 3.2139, "step": 11794 }, { "epoch": 0.41562056784446383, "grad_norm": 8.180044174194336, "learning_rate": 1.3184231124614711e-05, "loss": 2.9794, "step": 11795 }, { "epoch": 0.41565580485742226, "grad_norm": 7.439951419830322, "learning_rate": 1.3183149232028367e-05, "loss": 5.375, "step": 11796 }, { "epoch": 0.41569104187038064, "grad_norm": 1.8185924291610718, "learning_rate": 1.3182067297981099e-05, "loss": 1.1983, "step": 11797 }, { "epoch": 0.4157262788833391, "grad_norm": 18.787370681762695, "learning_rate": 1.3180985322486999e-05, "loss": 5.9716, "step": 11798 }, { "epoch": 0.41576151589629745, "grad_norm": 1.1978158950805664, "learning_rate": 1.317990330556016e-05, "loss": 0.9139, "step": 11799 }, { "epoch": 0.4157967529092559, "grad_norm": 1.055939793586731, "learning_rate": 1.3178821247214674e-05, "loss": 0.855, "step": 11800 }, { "epoch": 0.4158319899222143, "grad_norm": 6.1616597175598145, "learning_rate": 1.317773914746464e-05, "loss": 2.7797, "step": 11801 }, { "epoch": 0.4158672269351727, "grad_norm": 1.5559369325637817, "learning_rate": 1.3176657006324146e-05, "loss": 0.9334, "step": 11802 }, { "epoch": 0.41590246394813113, "grad_norm": 1.607029676437378, "learning_rate": 1.317557482380729e-05, "loss": 1.0006, "step": 11803 }, { "epoch": 0.4159377009610895, "grad_norm": 6.774272441864014, "learning_rate": 1.3174492599928168e-05, "loss": 2.7281, "step": 11804 }, { "epoch": 0.41597293797404794, "grad_norm": 1.798048496246338, "learning_rate": 1.3173410334700875e-05, "loss": 0.8993, "step": 11805 }, { "epoch": 0.4160081749870064, "grad_norm": 1.116576075553894, "learning_rate": 1.317232802813951e-05, "loss": 0.7488, "step": 11806 }, { "epoch": 0.41604341199996475, "grad_norm": 0.997885525226593, "learning_rate": 1.3171245680258165e-05, "loss": 0.7075, "step": 11807 }, { "epoch": 0.4160786490129232, "grad_norm": 1.2735127210617065, "learning_rate": 1.3170163291070943e-05, "loss": 0.7286, "step": 11808 }, { "epoch": 0.41611388602588156, "grad_norm": 4.797351837158203, "learning_rate": 1.3169080860591943e-05, "loss": 3.4916, "step": 11809 }, { "epoch": 0.41614912303884, "grad_norm": 4.110494613647461, "learning_rate": 1.3167998388835255e-05, "loss": 3.2461, "step": 11810 }, { "epoch": 0.41618436005179843, "grad_norm": 2.5947697162628174, "learning_rate": 1.3166915875814992e-05, "loss": 0.7791, "step": 11811 }, { "epoch": 0.4162195970647568, "grad_norm": 5.165349006652832, "learning_rate": 1.316583332154524e-05, "loss": 3.3609, "step": 11812 }, { "epoch": 0.41625483407771524, "grad_norm": 5.754498481750488, "learning_rate": 1.3164750726040108e-05, "loss": 3.2126, "step": 11813 }, { "epoch": 0.4162900710906736, "grad_norm": 7.744683265686035, "learning_rate": 1.3163668089313695e-05, "loss": 4.0012, "step": 11814 }, { "epoch": 0.41632530810363205, "grad_norm": 1.5344483852386475, "learning_rate": 1.31625854113801e-05, "loss": 0.8523, "step": 11815 }, { "epoch": 0.4163605451165905, "grad_norm": 1.554675817489624, "learning_rate": 1.3161502692253429e-05, "loss": 0.8428, "step": 11816 }, { "epoch": 0.41639578212954886, "grad_norm": 2.451166868209839, "learning_rate": 1.3160419931947784e-05, "loss": 1.0016, "step": 11817 }, { "epoch": 0.4164310191425073, "grad_norm": 1.0017541646957397, "learning_rate": 1.3159337130477263e-05, "loss": 1.0984, "step": 11818 }, { "epoch": 0.4164662561554657, "grad_norm": 12.010035514831543, "learning_rate": 1.3158254287855977e-05, "loss": 5.4209, "step": 11819 }, { "epoch": 0.4165014931684241, "grad_norm": 5.726729869842529, "learning_rate": 1.3157171404098029e-05, "loss": 2.5736, "step": 11820 }, { "epoch": 0.41653673018138254, "grad_norm": 11.638117790222168, "learning_rate": 1.3156088479217513e-05, "loss": 3.4085, "step": 11821 }, { "epoch": 0.4165719671943409, "grad_norm": 5.919540882110596, "learning_rate": 1.315500551322855e-05, "loss": 5.2429, "step": 11822 }, { "epoch": 0.41660720420729935, "grad_norm": 1.384113073348999, "learning_rate": 1.3153922506145235e-05, "loss": 0.8591, "step": 11823 }, { "epoch": 0.4166424412202578, "grad_norm": 1.6404591798782349, "learning_rate": 1.3152839457981679e-05, "loss": 0.9638, "step": 11824 }, { "epoch": 0.41667767823321616, "grad_norm": 1.23036527633667, "learning_rate": 1.3151756368751985e-05, "loss": 0.8892, "step": 11825 }, { "epoch": 0.4167129152461746, "grad_norm": 4.8194499015808105, "learning_rate": 1.3150673238470263e-05, "loss": 5.2915, "step": 11826 }, { "epoch": 0.416748152259133, "grad_norm": 1.3307923078536987, "learning_rate": 1.3149590067150624e-05, "loss": 0.897, "step": 11827 }, { "epoch": 0.4167833892720914, "grad_norm": 4.3587727546691895, "learning_rate": 1.3148506854807172e-05, "loss": 3.1293, "step": 11828 }, { "epoch": 0.41681862628504984, "grad_norm": 1.233799695968628, "learning_rate": 1.3147423601454015e-05, "loss": 0.9366, "step": 11829 }, { "epoch": 0.4168538632980082, "grad_norm": 1.250514030456543, "learning_rate": 1.3146340307105267e-05, "loss": 1.0016, "step": 11830 }, { "epoch": 0.41688910031096665, "grad_norm": 1.1943700313568115, "learning_rate": 1.3145256971775036e-05, "loss": 0.9307, "step": 11831 }, { "epoch": 0.41692433732392503, "grad_norm": 1.0205720663070679, "learning_rate": 1.314417359547743e-05, "loss": 1.1698, "step": 11832 }, { "epoch": 0.41695957433688347, "grad_norm": 5.230285167694092, "learning_rate": 1.3143090178226565e-05, "loss": 3.1471, "step": 11833 }, { "epoch": 0.4169948113498419, "grad_norm": 5.824563503265381, "learning_rate": 1.3142006720036549e-05, "loss": 3.2289, "step": 11834 }, { "epoch": 0.4170300483628003, "grad_norm": 9.14360237121582, "learning_rate": 1.3140923220921493e-05, "loss": 3.3784, "step": 11835 }, { "epoch": 0.4170652853757587, "grad_norm": 2.0258069038391113, "learning_rate": 1.3139839680895514e-05, "loss": 0.9262, "step": 11836 }, { "epoch": 0.4171005223887171, "grad_norm": 1.2181720733642578, "learning_rate": 1.3138756099972722e-05, "loss": 0.7131, "step": 11837 }, { "epoch": 0.4171357594016755, "grad_norm": 0.8545265793800354, "learning_rate": 1.3137672478167234e-05, "loss": 0.735, "step": 11838 }, { "epoch": 0.41717099641463395, "grad_norm": 5.190419673919678, "learning_rate": 1.3136588815493163e-05, "loss": 3.3024, "step": 11839 }, { "epoch": 0.41720623342759233, "grad_norm": 1.5045477151870728, "learning_rate": 1.313550511196462e-05, "loss": 0.7684, "step": 11840 }, { "epoch": 0.41724147044055077, "grad_norm": 1.3643046617507935, "learning_rate": 1.3134421367595726e-05, "loss": 0.7404, "step": 11841 }, { "epoch": 0.41727670745350914, "grad_norm": 8.212021827697754, "learning_rate": 1.3133337582400594e-05, "loss": 5.6422, "step": 11842 }, { "epoch": 0.4173119444664676, "grad_norm": 1.3067383766174316, "learning_rate": 1.313225375639334e-05, "loss": 0.7713, "step": 11843 }, { "epoch": 0.417347181479426, "grad_norm": 7.120114326477051, "learning_rate": 1.3131169889588082e-05, "loss": 3.372, "step": 11844 }, { "epoch": 0.4173824184923844, "grad_norm": 16.043516159057617, "learning_rate": 1.3130085981998937e-05, "loss": 8.3864, "step": 11845 }, { "epoch": 0.4174176555053428, "grad_norm": 11.22260570526123, "learning_rate": 1.3129002033640023e-05, "loss": 5.9178, "step": 11846 }, { "epoch": 0.4174528925183012, "grad_norm": 2.5188138484954834, "learning_rate": 1.3127918044525464e-05, "loss": 1.1971, "step": 11847 }, { "epoch": 0.41748812953125963, "grad_norm": 1.5890603065490723, "learning_rate": 1.3126834014669365e-05, "loss": 0.9703, "step": 11848 }, { "epoch": 0.41752336654421807, "grad_norm": 4.164116382598877, "learning_rate": 1.3125749944085863e-05, "loss": 3.1629, "step": 11849 }, { "epoch": 0.41755860355717644, "grad_norm": 1.607818841934204, "learning_rate": 1.3124665832789066e-05, "loss": 0.9064, "step": 11850 }, { "epoch": 0.4175938405701349, "grad_norm": 43.56708526611328, "learning_rate": 1.3123581680793098e-05, "loss": 5.928, "step": 11851 }, { "epoch": 0.41762907758309326, "grad_norm": 9.604461669921875, "learning_rate": 1.3122497488112082e-05, "loss": 5.3585, "step": 11852 }, { "epoch": 0.4176643145960517, "grad_norm": 1.3845261335372925, "learning_rate": 1.3121413254760138e-05, "loss": 0.8182, "step": 11853 }, { "epoch": 0.4176995516090101, "grad_norm": 4.2959160804748535, "learning_rate": 1.312032898075139e-05, "loss": 3.0001, "step": 11854 }, { "epoch": 0.4177347886219685, "grad_norm": 4.143740177154541, "learning_rate": 1.3119244666099957e-05, "loss": 2.9749, "step": 11855 }, { "epoch": 0.41777002563492693, "grad_norm": 1.548909068107605, "learning_rate": 1.3118160310819965e-05, "loss": 0.7832, "step": 11856 }, { "epoch": 0.4178052626478853, "grad_norm": 1.0301491022109985, "learning_rate": 1.3117075914925538e-05, "loss": 0.8978, "step": 11857 }, { "epoch": 0.41784049966084374, "grad_norm": 1.1187630891799927, "learning_rate": 1.3115991478430805e-05, "loss": 1.1805, "step": 11858 }, { "epoch": 0.4178757366738022, "grad_norm": 3.9180729389190674, "learning_rate": 1.3114907001349879e-05, "loss": 2.9453, "step": 11859 }, { "epoch": 0.41791097368676056, "grad_norm": 11.85024356842041, "learning_rate": 1.3113822483696899e-05, "loss": 3.1849, "step": 11860 }, { "epoch": 0.417946210699719, "grad_norm": 11.889747619628906, "learning_rate": 1.311273792548598e-05, "loss": 6.0264, "step": 11861 }, { "epoch": 0.4179814477126774, "grad_norm": 1.8646671772003174, "learning_rate": 1.3111653326731255e-05, "loss": 0.7807, "step": 11862 }, { "epoch": 0.4180166847256358, "grad_norm": 1.201869010925293, "learning_rate": 1.3110568687446848e-05, "loss": 1.1895, "step": 11863 }, { "epoch": 0.41805192173859423, "grad_norm": 1.2123016119003296, "learning_rate": 1.3109484007646888e-05, "loss": 0.9041, "step": 11864 }, { "epoch": 0.4180871587515526, "grad_norm": 0.9847707152366638, "learning_rate": 1.3108399287345504e-05, "loss": 0.9028, "step": 11865 }, { "epoch": 0.41812239576451105, "grad_norm": 1.891660451889038, "learning_rate": 1.3107314526556823e-05, "loss": 1.0486, "step": 11866 }, { "epoch": 0.4181576327774695, "grad_norm": 1.5031850337982178, "learning_rate": 1.310622972529497e-05, "loss": 1.0809, "step": 11867 }, { "epoch": 0.41819286979042786, "grad_norm": 2.599011182785034, "learning_rate": 1.3105144883574085e-05, "loss": 0.8466, "step": 11868 }, { "epoch": 0.4182281068033863, "grad_norm": 6.467948913574219, "learning_rate": 1.310406000140829e-05, "loss": 4.1824, "step": 11869 }, { "epoch": 0.41826334381634467, "grad_norm": 51.20038986206055, "learning_rate": 1.3102975078811718e-05, "loss": 3.9973, "step": 11870 }, { "epoch": 0.4182985808293031, "grad_norm": 7.828527927398682, "learning_rate": 1.31018901157985e-05, "loss": 3.3598, "step": 11871 }, { "epoch": 0.41833381784226153, "grad_norm": 24.97281837463379, "learning_rate": 1.3100805112382768e-05, "loss": 5.4127, "step": 11872 }, { "epoch": 0.4183690548552199, "grad_norm": 24.699533462524414, "learning_rate": 1.3099720068578656e-05, "loss": 3.7522, "step": 11873 }, { "epoch": 0.41840429186817835, "grad_norm": 1.6054644584655762, "learning_rate": 1.3098634984400293e-05, "loss": 0.959, "step": 11874 }, { "epoch": 0.4184395288811367, "grad_norm": 9.495438575744629, "learning_rate": 1.3097549859861815e-05, "loss": 3.7628, "step": 11875 }, { "epoch": 0.41847476589409516, "grad_norm": 5.397059440612793, "learning_rate": 1.3096464694977357e-05, "loss": 2.4567, "step": 11876 }, { "epoch": 0.4185100029070536, "grad_norm": 14.99561595916748, "learning_rate": 1.3095379489761053e-05, "loss": 5.6116, "step": 11877 }, { "epoch": 0.41854523992001197, "grad_norm": 4.604128360748291, "learning_rate": 1.3094294244227035e-05, "loss": 2.9511, "step": 11878 }, { "epoch": 0.4185804769329704, "grad_norm": 3.000648021697998, "learning_rate": 1.309320895838944e-05, "loss": 1.1119, "step": 11879 }, { "epoch": 0.4186157139459288, "grad_norm": 3.0614089965820312, "learning_rate": 1.3092123632262403e-05, "loss": 0.8597, "step": 11880 }, { "epoch": 0.4186509509588872, "grad_norm": 16.650222778320312, "learning_rate": 1.3091038265860067e-05, "loss": 5.4076, "step": 11881 }, { "epoch": 0.41868618797184565, "grad_norm": 8.167202949523926, "learning_rate": 1.308995285919656e-05, "loss": 3.3862, "step": 11882 }, { "epoch": 0.418721424984804, "grad_norm": 1.0199010372161865, "learning_rate": 1.3088867412286024e-05, "loss": 0.9811, "step": 11883 }, { "epoch": 0.41875666199776246, "grad_norm": 12.811788558959961, "learning_rate": 1.3087781925142595e-05, "loss": 5.8617, "step": 11884 }, { "epoch": 0.41879189901072084, "grad_norm": 21.3895263671875, "learning_rate": 1.3086696397780418e-05, "loss": 3.3136, "step": 11885 }, { "epoch": 0.41882713602367927, "grad_norm": 5.28989839553833, "learning_rate": 1.308561083021362e-05, "loss": 3.0595, "step": 11886 }, { "epoch": 0.4188623730366377, "grad_norm": 1.773113489151001, "learning_rate": 1.3084525222456354e-05, "loss": 1.0381, "step": 11887 }, { "epoch": 0.4188976100495961, "grad_norm": 1.7832648754119873, "learning_rate": 1.3083439574522752e-05, "loss": 0.9148, "step": 11888 }, { "epoch": 0.4189328470625545, "grad_norm": 1.8109768629074097, "learning_rate": 1.3082353886426957e-05, "loss": 1.0101, "step": 11889 }, { "epoch": 0.4189680840755129, "grad_norm": 1.1974537372589111, "learning_rate": 1.3081268158183109e-05, "loss": 0.9859, "step": 11890 }, { "epoch": 0.4190033210884713, "grad_norm": 1.3189585208892822, "learning_rate": 1.3080182389805351e-05, "loss": 1.1244, "step": 11891 }, { "epoch": 0.41903855810142976, "grad_norm": 9.10437297821045, "learning_rate": 1.3079096581307825e-05, "loss": 5.2649, "step": 11892 }, { "epoch": 0.41907379511438814, "grad_norm": 1.5111520290374756, "learning_rate": 1.3078010732704673e-05, "loss": 0.9132, "step": 11893 }, { "epoch": 0.41910903212734657, "grad_norm": 1.2953380346298218, "learning_rate": 1.3076924844010042e-05, "loss": 0.7669, "step": 11894 }, { "epoch": 0.419144269140305, "grad_norm": 3.6686346530914307, "learning_rate": 1.307583891523807e-05, "loss": 3.1152, "step": 11895 }, { "epoch": 0.4191795061532634, "grad_norm": 11.91445541381836, "learning_rate": 1.3074752946402906e-05, "loss": 5.8181, "step": 11896 }, { "epoch": 0.4192147431662218, "grad_norm": 6.3837995529174805, "learning_rate": 1.307366693751869e-05, "loss": 3.9548, "step": 11897 }, { "epoch": 0.4192499801791802, "grad_norm": 9.593385696411133, "learning_rate": 1.3072580888599577e-05, "loss": 4.8889, "step": 11898 }, { "epoch": 0.4192852171921386, "grad_norm": 1.938215732574463, "learning_rate": 1.3071494799659701e-05, "loss": 1.1074, "step": 11899 }, { "epoch": 0.41932045420509706, "grad_norm": 4.58364200592041, "learning_rate": 1.3070408670713217e-05, "loss": 2.8261, "step": 11900 }, { "epoch": 0.41935569121805544, "grad_norm": 1.2872118949890137, "learning_rate": 1.3069322501774267e-05, "loss": 1.0499, "step": 11901 }, { "epoch": 0.41939092823101387, "grad_norm": 6.2253828048706055, "learning_rate": 1.3068236292857001e-05, "loss": 2.9148, "step": 11902 }, { "epoch": 0.41942616524397225, "grad_norm": 7.7762980461120605, "learning_rate": 1.3067150043975568e-05, "loss": 5.4459, "step": 11903 }, { "epoch": 0.4194614022569307, "grad_norm": 1.1724417209625244, "learning_rate": 1.3066063755144113e-05, "loss": 0.8939, "step": 11904 }, { "epoch": 0.4194966392698891, "grad_norm": 1.0074855089187622, "learning_rate": 1.3064977426376789e-05, "loss": 0.9768, "step": 11905 }, { "epoch": 0.4195318762828475, "grad_norm": 82.892822265625, "learning_rate": 1.3063891057687744e-05, "loss": 3.2244, "step": 11906 }, { "epoch": 0.4195671132958059, "grad_norm": 5.777791500091553, "learning_rate": 1.3062804649091123e-05, "loss": 3.4086, "step": 11907 }, { "epoch": 0.4196023503087643, "grad_norm": 40.53087615966797, "learning_rate": 1.3061718200601085e-05, "loss": 2.9703, "step": 11908 }, { "epoch": 0.41963758732172274, "grad_norm": 1.0882800817489624, "learning_rate": 1.3060631712231777e-05, "loss": 0.7383, "step": 11909 }, { "epoch": 0.41967282433468117, "grad_norm": 1.422463297843933, "learning_rate": 1.3059545183997351e-05, "loss": 1.1389, "step": 11910 }, { "epoch": 0.41970806134763955, "grad_norm": 12.56614875793457, "learning_rate": 1.305845861591196e-05, "loss": 5.4864, "step": 11911 }, { "epoch": 0.419743298360598, "grad_norm": 9.6939115524292, "learning_rate": 1.3057372007989755e-05, "loss": 4.8452, "step": 11912 }, { "epoch": 0.41977853537355636, "grad_norm": 3.826918601989746, "learning_rate": 1.305628536024489e-05, "loss": 2.5547, "step": 11913 }, { "epoch": 0.4198137723865148, "grad_norm": 4.069167137145996, "learning_rate": 1.3055198672691518e-05, "loss": 3.1439, "step": 11914 }, { "epoch": 0.4198490093994732, "grad_norm": 70.92704772949219, "learning_rate": 1.3054111945343797e-05, "loss": 5.2052, "step": 11915 }, { "epoch": 0.4198842464124316, "grad_norm": 1.997428297996521, "learning_rate": 1.3053025178215875e-05, "loss": 1.005, "step": 11916 }, { "epoch": 0.41991948342539004, "grad_norm": 8.227042198181152, "learning_rate": 1.3051938371321916e-05, "loss": 3.4498, "step": 11917 }, { "epoch": 0.4199547204383484, "grad_norm": 6.994395732879639, "learning_rate": 1.3050851524676065e-05, "loss": 3.058, "step": 11918 }, { "epoch": 0.41998995745130685, "grad_norm": 1.402398705482483, "learning_rate": 1.3049764638292489e-05, "loss": 1.0148, "step": 11919 }, { "epoch": 0.4200251944642653, "grad_norm": 6.574413776397705, "learning_rate": 1.3048677712185337e-05, "loss": 2.8822, "step": 11920 }, { "epoch": 0.42006043147722366, "grad_norm": 9.573902130126953, "learning_rate": 1.3047590746368773e-05, "loss": 3.5782, "step": 11921 }, { "epoch": 0.4200956684901821, "grad_norm": 1.2190154790878296, "learning_rate": 1.3046503740856948e-05, "loss": 1.0801, "step": 11922 }, { "epoch": 0.4201309055031405, "grad_norm": 7.323479652404785, "learning_rate": 1.3045416695664029e-05, "loss": 3.2009, "step": 11923 }, { "epoch": 0.4201661425160989, "grad_norm": 1.8121845722198486, "learning_rate": 1.304432961080416e-05, "loss": 1.2161, "step": 11924 }, { "epoch": 0.42020137952905734, "grad_norm": 7.216464996337891, "learning_rate": 1.304324248629152e-05, "loss": 2.5202, "step": 11925 }, { "epoch": 0.4202366165420157, "grad_norm": 7.066014766693115, "learning_rate": 1.304215532214025e-05, "loss": 3.1281, "step": 11926 }, { "epoch": 0.42027185355497415, "grad_norm": 4.897197246551514, "learning_rate": 1.3041068118364528e-05, "loss": 3.2376, "step": 11927 }, { "epoch": 0.42030709056793253, "grad_norm": 8.78206729888916, "learning_rate": 1.3039980874978501e-05, "loss": 3.4291, "step": 11928 }, { "epoch": 0.42034232758089096, "grad_norm": 3.9070470333099365, "learning_rate": 1.3038893591996339e-05, "loss": 1.0865, "step": 11929 }, { "epoch": 0.4203775645938494, "grad_norm": 1.9267923831939697, "learning_rate": 1.3037806269432199e-05, "loss": 0.9153, "step": 11930 }, { "epoch": 0.4204128016068078, "grad_norm": 2.5898168087005615, "learning_rate": 1.3036718907300245e-05, "loss": 0.9867, "step": 11931 }, { "epoch": 0.4204480386197662, "grad_norm": 2.3482563495635986, "learning_rate": 1.3035631505614641e-05, "loss": 1.009, "step": 11932 }, { "epoch": 0.42048327563272464, "grad_norm": 4.50426721572876, "learning_rate": 1.3034544064389551e-05, "loss": 3.2695, "step": 11933 }, { "epoch": 0.420518512645683, "grad_norm": 2.1024646759033203, "learning_rate": 1.3033456583639134e-05, "loss": 0.836, "step": 11934 }, { "epoch": 0.42055374965864145, "grad_norm": 8.94986343383789, "learning_rate": 1.3032369063377566e-05, "loss": 5.5453, "step": 11935 }, { "epoch": 0.42058898667159983, "grad_norm": 11.941231727600098, "learning_rate": 1.3031281503619e-05, "loss": 7.6569, "step": 11936 }, { "epoch": 0.42062422368455826, "grad_norm": 4.448258399963379, "learning_rate": 1.3030193904377607e-05, "loss": 3.3789, "step": 11937 }, { "epoch": 0.4206594606975167, "grad_norm": 5.684335231781006, "learning_rate": 1.302910626566755e-05, "loss": 2.8135, "step": 11938 }, { "epoch": 0.4206946977104751, "grad_norm": 0.8854876160621643, "learning_rate": 1.3028018587503e-05, "loss": 1.0063, "step": 11939 }, { "epoch": 0.4207299347234335, "grad_norm": 1.2674427032470703, "learning_rate": 1.3026930869898122e-05, "loss": 0.9974, "step": 11940 }, { "epoch": 0.4207651717363919, "grad_norm": 1.178863763809204, "learning_rate": 1.3025843112867085e-05, "loss": 0.8753, "step": 11941 }, { "epoch": 0.4208004087493503, "grad_norm": 8.646707534790039, "learning_rate": 1.3024755316424056e-05, "loss": 7.9196, "step": 11942 }, { "epoch": 0.42083564576230875, "grad_norm": 11.97336196899414, "learning_rate": 1.30236674805832e-05, "loss": 6.259, "step": 11943 }, { "epoch": 0.42087088277526713, "grad_norm": 1.2655367851257324, "learning_rate": 1.3022579605358697e-05, "loss": 1.0053, "step": 11944 }, { "epoch": 0.42090611978822556, "grad_norm": 13.289244651794434, "learning_rate": 1.3021491690764702e-05, "loss": 8.8182, "step": 11945 }, { "epoch": 0.42094135680118394, "grad_norm": 5.6745781898498535, "learning_rate": 1.3020403736815399e-05, "loss": 4.4241, "step": 11946 }, { "epoch": 0.4209765938141424, "grad_norm": 1.658636450767517, "learning_rate": 1.3019315743524947e-05, "loss": 0.9947, "step": 11947 }, { "epoch": 0.4210118308271008, "grad_norm": 6.105731964111328, "learning_rate": 1.3018227710907526e-05, "loss": 3.8475, "step": 11948 }, { "epoch": 0.4210470678400592, "grad_norm": 1.852245807647705, "learning_rate": 1.3017139638977306e-05, "loss": 0.9594, "step": 11949 }, { "epoch": 0.4210823048530176, "grad_norm": 0.9793496131896973, "learning_rate": 1.3016051527748457e-05, "loss": 0.8368, "step": 11950 }, { "epoch": 0.421117541865976, "grad_norm": 29.270837783813477, "learning_rate": 1.301496337723515e-05, "loss": 3.1649, "step": 11951 }, { "epoch": 0.42115277887893443, "grad_norm": 2.7427268028259277, "learning_rate": 1.3013875187451563e-05, "loss": 3.3461, "step": 11952 }, { "epoch": 0.42118801589189286, "grad_norm": 1.713456392288208, "learning_rate": 1.3012786958411868e-05, "loss": 0.7579, "step": 11953 }, { "epoch": 0.42122325290485124, "grad_norm": 1.656022548675537, "learning_rate": 1.3011698690130239e-05, "loss": 0.9588, "step": 11954 }, { "epoch": 0.4212584899178097, "grad_norm": 1.1058602333068848, "learning_rate": 1.301061038262085e-05, "loss": 0.8679, "step": 11955 }, { "epoch": 0.42129372693076805, "grad_norm": 3.7977712154388428, "learning_rate": 1.300952203589788e-05, "loss": 3.7286, "step": 11956 }, { "epoch": 0.4213289639437265, "grad_norm": 10.729963302612305, "learning_rate": 1.3008433649975498e-05, "loss": 5.9459, "step": 11957 }, { "epoch": 0.4213642009566849, "grad_norm": 2.2452476024627686, "learning_rate": 1.3007345224867887e-05, "loss": 1.0978, "step": 11958 }, { "epoch": 0.4213994379696433, "grad_norm": 1.673062801361084, "learning_rate": 1.3006256760589219e-05, "loss": 1.0885, "step": 11959 }, { "epoch": 0.42143467498260173, "grad_norm": 1.6500163078308105, "learning_rate": 1.3005168257153678e-05, "loss": 0.8113, "step": 11960 }, { "epoch": 0.4214699119955601, "grad_norm": 1.4433026313781738, "learning_rate": 1.3004079714575435e-05, "loss": 0.7761, "step": 11961 }, { "epoch": 0.42150514900851854, "grad_norm": 7.763062953948975, "learning_rate": 1.3002991132868669e-05, "loss": 5.0746, "step": 11962 }, { "epoch": 0.421540386021477, "grad_norm": 15.14836311340332, "learning_rate": 1.3001902512047567e-05, "loss": 3.2763, "step": 11963 }, { "epoch": 0.42157562303443535, "grad_norm": 1.0395078659057617, "learning_rate": 1.3000813852126295e-05, "loss": 1.0549, "step": 11964 }, { "epoch": 0.4216108600473938, "grad_norm": 6.598300457000732, "learning_rate": 1.2999725153119047e-05, "loss": 3.8124, "step": 11965 }, { "epoch": 0.4216460970603522, "grad_norm": 2.037421941757202, "learning_rate": 1.2998636415039994e-05, "loss": 0.8948, "step": 11966 }, { "epoch": 0.4216813340733106, "grad_norm": 10.416411399841309, "learning_rate": 1.2997547637903319e-05, "loss": 3.7064, "step": 11967 }, { "epoch": 0.42171657108626903, "grad_norm": 1.8980463743209839, "learning_rate": 1.2996458821723203e-05, "loss": 0.9883, "step": 11968 }, { "epoch": 0.4217518080992274, "grad_norm": 4.866113662719727, "learning_rate": 1.2995369966513834e-05, "loss": 5.154, "step": 11969 }, { "epoch": 0.42178704511218584, "grad_norm": 1.2549327611923218, "learning_rate": 1.2994281072289383e-05, "loss": 0.7072, "step": 11970 }, { "epoch": 0.4218222821251443, "grad_norm": 1.1356371641159058, "learning_rate": 1.2993192139064046e-05, "loss": 1.0037, "step": 11971 }, { "epoch": 0.42185751913810265, "grad_norm": 1.9692869186401367, "learning_rate": 1.2992103166851994e-05, "loss": 0.9808, "step": 11972 }, { "epoch": 0.4218927561510611, "grad_norm": 1.38277268409729, "learning_rate": 1.2991014155667423e-05, "loss": 0.9293, "step": 11973 }, { "epoch": 0.42192799316401947, "grad_norm": 2.1580374240875244, "learning_rate": 1.2989925105524507e-05, "loss": 1.4073, "step": 11974 }, { "epoch": 0.4219632301769779, "grad_norm": 1.3116517066955566, "learning_rate": 1.2988836016437439e-05, "loss": 0.6782, "step": 11975 }, { "epoch": 0.42199846718993633, "grad_norm": 1.8904656171798706, "learning_rate": 1.29877468884204e-05, "loss": 0.8621, "step": 11976 }, { "epoch": 0.4220337042028947, "grad_norm": 1.778632402420044, "learning_rate": 1.2986657721487577e-05, "loss": 0.8177, "step": 11977 }, { "epoch": 0.42206894121585314, "grad_norm": 5.37869119644165, "learning_rate": 1.2985568515653157e-05, "loss": 3.5522, "step": 11978 }, { "epoch": 0.4221041782288115, "grad_norm": 7.848970890045166, "learning_rate": 1.2984479270931325e-05, "loss": 6.1757, "step": 11979 }, { "epoch": 0.42213941524176996, "grad_norm": 6.80164098739624, "learning_rate": 1.2983389987336272e-05, "loss": 3.2124, "step": 11980 }, { "epoch": 0.4221746522547284, "grad_norm": 7.820616245269775, "learning_rate": 1.2982300664882184e-05, "loss": 3.2311, "step": 11981 }, { "epoch": 0.42220988926768677, "grad_norm": 7.096263885498047, "learning_rate": 1.2981211303583253e-05, "loss": 4.8713, "step": 11982 }, { "epoch": 0.4222451262806452, "grad_norm": 5.739904880523682, "learning_rate": 1.298012190345366e-05, "loss": 4.071, "step": 11983 }, { "epoch": 0.4222803632936036, "grad_norm": 6.618044853210449, "learning_rate": 1.2979032464507605e-05, "loss": 3.1514, "step": 11984 }, { "epoch": 0.422315600306562, "grad_norm": 2.216845750808716, "learning_rate": 1.2977942986759268e-05, "loss": 1.1288, "step": 11985 }, { "epoch": 0.42235083731952044, "grad_norm": 7.902429103851318, "learning_rate": 1.2976853470222845e-05, "loss": 3.5637, "step": 11986 }, { "epoch": 0.4223860743324788, "grad_norm": 3.7403407096862793, "learning_rate": 1.2975763914912527e-05, "loss": 3.3805, "step": 11987 }, { "epoch": 0.42242131134543726, "grad_norm": 8.650525093078613, "learning_rate": 1.2974674320842508e-05, "loss": 3.6189, "step": 11988 }, { "epoch": 0.42245654835839563, "grad_norm": 2.3957388401031494, "learning_rate": 1.2973584688026972e-05, "loss": 1.0261, "step": 11989 }, { "epoch": 0.42249178537135407, "grad_norm": 1.1547883749008179, "learning_rate": 1.297249501648012e-05, "loss": 0.9887, "step": 11990 }, { "epoch": 0.4225270223843125, "grad_norm": 7.608021259307861, "learning_rate": 1.2971405306216139e-05, "loss": 2.8073, "step": 11991 }, { "epoch": 0.4225622593972709, "grad_norm": 5.555563449859619, "learning_rate": 1.2970315557249228e-05, "loss": 3.5133, "step": 11992 }, { "epoch": 0.4225974964102293, "grad_norm": 6.215681076049805, "learning_rate": 1.2969225769593578e-05, "loss": 3.1685, "step": 11993 }, { "epoch": 0.4226327334231877, "grad_norm": 5.137460231781006, "learning_rate": 1.2968135943263382e-05, "loss": 3.0327, "step": 11994 }, { "epoch": 0.4226679704361461, "grad_norm": 5.547117710113525, "learning_rate": 1.296704607827284e-05, "loss": 3.607, "step": 11995 }, { "epoch": 0.42270320744910456, "grad_norm": 10.731036186218262, "learning_rate": 1.2965956174636143e-05, "loss": 3.7102, "step": 11996 }, { "epoch": 0.42273844446206293, "grad_norm": 4.061404228210449, "learning_rate": 1.296486623236749e-05, "loss": 2.6042, "step": 11997 }, { "epoch": 0.42277368147502137, "grad_norm": 11.962862014770508, "learning_rate": 1.2963776251481077e-05, "loss": 5.6063, "step": 11998 }, { "epoch": 0.4228089184879798, "grad_norm": 5.896179676055908, "learning_rate": 1.29626862319911e-05, "loss": 3.3465, "step": 11999 }, { "epoch": 0.4228441555009382, "grad_norm": 3.2176108360290527, "learning_rate": 1.2961596173911757e-05, "loss": 3.1722, "step": 12000 }, { "epoch": 0.4228793925138966, "grad_norm": 1.1579540967941284, "learning_rate": 1.2960506077257249e-05, "loss": 1.0913, "step": 12001 }, { "epoch": 0.422914629526855, "grad_norm": 8.022298812866211, "learning_rate": 1.295941594204177e-05, "loss": 5.6025, "step": 12002 }, { "epoch": 0.4229498665398134, "grad_norm": 60.91880416870117, "learning_rate": 1.2958325768279523e-05, "loss": 8.5195, "step": 12003 }, { "epoch": 0.42298510355277186, "grad_norm": 4.42491340637207, "learning_rate": 1.2957235555984704e-05, "loss": 3.3171, "step": 12004 }, { "epoch": 0.42302034056573024, "grad_norm": 13.50173568725586, "learning_rate": 1.2956145305171516e-05, "loss": 2.9608, "step": 12005 }, { "epoch": 0.42305557757868867, "grad_norm": 2.804680824279785, "learning_rate": 1.2955055015854159e-05, "loss": 1.0353, "step": 12006 }, { "epoch": 0.42309081459164705, "grad_norm": 1.359700322151184, "learning_rate": 1.2953964688046835e-05, "loss": 1.001, "step": 12007 }, { "epoch": 0.4231260516046055, "grad_norm": 4.6413068771362305, "learning_rate": 1.2952874321763742e-05, "loss": 2.9784, "step": 12008 }, { "epoch": 0.4231612886175639, "grad_norm": 5.203223705291748, "learning_rate": 1.2951783917019089e-05, "loss": 2.9325, "step": 12009 }, { "epoch": 0.4231965256305223, "grad_norm": 2.0642709732055664, "learning_rate": 1.295069347382707e-05, "loss": 0.8121, "step": 12010 }, { "epoch": 0.4232317626434807, "grad_norm": 6.912400245666504, "learning_rate": 1.2949602992201897e-05, "loss": 3.4512, "step": 12011 }, { "epoch": 0.4232669996564391, "grad_norm": 10.30289363861084, "learning_rate": 1.2948512472157767e-05, "loss": 5.4556, "step": 12012 }, { "epoch": 0.42330223666939754, "grad_norm": 1.9043245315551758, "learning_rate": 1.2947421913708884e-05, "loss": 0.7836, "step": 12013 }, { "epoch": 0.42333747368235597, "grad_norm": 16.244539260864258, "learning_rate": 1.2946331316869459e-05, "loss": 3.3471, "step": 12014 }, { "epoch": 0.42337271069531435, "grad_norm": 5.941574573516846, "learning_rate": 1.294524068165369e-05, "loss": 3.2747, "step": 12015 }, { "epoch": 0.4234079477082728, "grad_norm": 1.5356519222259521, "learning_rate": 1.2944150008075789e-05, "loss": 0.9808, "step": 12016 }, { "epoch": 0.42344318472123116, "grad_norm": 1.6313127279281616, "learning_rate": 1.2943059296149958e-05, "loss": 1.0313, "step": 12017 }, { "epoch": 0.4234784217341896, "grad_norm": 1.3427910804748535, "learning_rate": 1.2941968545890405e-05, "loss": 0.7475, "step": 12018 }, { "epoch": 0.423513658747148, "grad_norm": 6.402644634246826, "learning_rate": 1.2940877757311333e-05, "loss": 4.5455, "step": 12019 }, { "epoch": 0.4235488957601064, "grad_norm": 9.246337890625, "learning_rate": 1.293978693042696e-05, "loss": 3.4779, "step": 12020 }, { "epoch": 0.42358413277306484, "grad_norm": 5.899450302124023, "learning_rate": 1.2938696065251484e-05, "loss": 4.9298, "step": 12021 }, { "epoch": 0.4236193697860232, "grad_norm": 7.42885684967041, "learning_rate": 1.293760516179912e-05, "loss": 3.9086, "step": 12022 }, { "epoch": 0.42365460679898165, "grad_norm": 1.144703984260559, "learning_rate": 1.2936514220084073e-05, "loss": 0.7312, "step": 12023 }, { "epoch": 0.4236898438119401, "grad_norm": 1.307236909866333, "learning_rate": 1.2935423240120555e-05, "loss": 0.9941, "step": 12024 }, { "epoch": 0.42372508082489846, "grad_norm": 1.2599914073944092, "learning_rate": 1.2934332221922774e-05, "loss": 0.9411, "step": 12025 }, { "epoch": 0.4237603178378569, "grad_norm": 9.850789070129395, "learning_rate": 1.2933241165504941e-05, "loss": 5.2232, "step": 12026 }, { "epoch": 0.42379555485081527, "grad_norm": 1.2193628549575806, "learning_rate": 1.2932150070881272e-05, "loss": 1.2385, "step": 12027 }, { "epoch": 0.4238307918637737, "grad_norm": 1.7107270956039429, "learning_rate": 1.2931058938065974e-05, "loss": 1.1628, "step": 12028 }, { "epoch": 0.42386602887673214, "grad_norm": 1.3855589628219604, "learning_rate": 1.2929967767073257e-05, "loss": 0.7937, "step": 12029 }, { "epoch": 0.4239012658896905, "grad_norm": 72.43801879882812, "learning_rate": 1.2928876557917338e-05, "loss": 3.2217, "step": 12030 }, { "epoch": 0.42393650290264895, "grad_norm": 1.3647615909576416, "learning_rate": 1.2927785310612429e-05, "loss": 0.6522, "step": 12031 }, { "epoch": 0.4239717399156073, "grad_norm": 1.6457663774490356, "learning_rate": 1.2926694025172743e-05, "loss": 0.9212, "step": 12032 }, { "epoch": 0.42400697692856576, "grad_norm": 1.7515995502471924, "learning_rate": 1.2925602701612497e-05, "loss": 0.8915, "step": 12033 }, { "epoch": 0.4240422139415242, "grad_norm": 11.147064208984375, "learning_rate": 1.29245113399459e-05, "loss": 3.4281, "step": 12034 }, { "epoch": 0.42407745095448257, "grad_norm": 1.5539743900299072, "learning_rate": 1.2923419940187172e-05, "loss": 0.829, "step": 12035 }, { "epoch": 0.424112687967441, "grad_norm": 1.7542438507080078, "learning_rate": 1.2922328502350524e-05, "loss": 1.0503, "step": 12036 }, { "epoch": 0.42414792498039944, "grad_norm": 4.680191993713379, "learning_rate": 1.2921237026450178e-05, "loss": 3.0143, "step": 12037 }, { "epoch": 0.4241831619933578, "grad_norm": 19.053194046020508, "learning_rate": 1.2920145512500346e-05, "loss": 5.5175, "step": 12038 }, { "epoch": 0.42421839900631625, "grad_norm": 1.9926090240478516, "learning_rate": 1.2919053960515248e-05, "loss": 0.83, "step": 12039 }, { "epoch": 0.4242536360192746, "grad_norm": 1.509163498878479, "learning_rate": 1.2917962370509098e-05, "loss": 0.8035, "step": 12040 }, { "epoch": 0.42428887303223306, "grad_norm": 9.734257698059082, "learning_rate": 1.2916870742496121e-05, "loss": 5.4891, "step": 12041 }, { "epoch": 0.4243241100451915, "grad_norm": 23.118379592895508, "learning_rate": 1.291577907649053e-05, "loss": 6.2972, "step": 12042 }, { "epoch": 0.42435934705814987, "grad_norm": 2.078880786895752, "learning_rate": 1.2914687372506543e-05, "loss": 0.8998, "step": 12043 }, { "epoch": 0.4243945840711083, "grad_norm": 35.08610153198242, "learning_rate": 1.2913595630558383e-05, "loss": 3.0627, "step": 12044 }, { "epoch": 0.4244298210840667, "grad_norm": 1.0677648782730103, "learning_rate": 1.2912503850660265e-05, "loss": 0.9327, "step": 12045 }, { "epoch": 0.4244650580970251, "grad_norm": 1.0733027458190918, "learning_rate": 1.2911412032826419e-05, "loss": 0.8274, "step": 12046 }, { "epoch": 0.42450029510998355, "grad_norm": 2.887535333633423, "learning_rate": 1.2910320177071059e-05, "loss": 0.7884, "step": 12047 }, { "epoch": 0.4245355321229419, "grad_norm": 2.2184510231018066, "learning_rate": 1.2909228283408403e-05, "loss": 1.0366, "step": 12048 }, { "epoch": 0.42457076913590036, "grad_norm": 1.9792282581329346, "learning_rate": 1.2908136351852685e-05, "loss": 0.8643, "step": 12049 }, { "epoch": 0.42460600614885874, "grad_norm": 4.990980625152588, "learning_rate": 1.2907044382418117e-05, "loss": 3.1352, "step": 12050 }, { "epoch": 0.4246412431618172, "grad_norm": 6.348422527313232, "learning_rate": 1.2905952375118927e-05, "loss": 2.8555, "step": 12051 }, { "epoch": 0.4246764801747756, "grad_norm": 0.9385572075843811, "learning_rate": 1.2904860329969333e-05, "loss": 0.905, "step": 12052 }, { "epoch": 0.424711717187734, "grad_norm": 5.080289363861084, "learning_rate": 1.2903768246983568e-05, "loss": 2.7721, "step": 12053 }, { "epoch": 0.4247469542006924, "grad_norm": 4.55698823928833, "learning_rate": 1.2902676126175848e-05, "loss": 3.0901, "step": 12054 }, { "epoch": 0.4247821912136508, "grad_norm": 20.14344596862793, "learning_rate": 1.2901583967560405e-05, "loss": 5.5378, "step": 12055 }, { "epoch": 0.42481742822660923, "grad_norm": 1.6279147863388062, "learning_rate": 1.290049177115146e-05, "loss": 0.9691, "step": 12056 }, { "epoch": 0.42485266523956766, "grad_norm": 13.63371753692627, "learning_rate": 1.289939953696324e-05, "loss": 2.363, "step": 12057 }, { "epoch": 0.42488790225252604, "grad_norm": 3.537529468536377, "learning_rate": 1.2898307265009972e-05, "loss": 0.7773, "step": 12058 }, { "epoch": 0.4249231392654845, "grad_norm": 4.67883825302124, "learning_rate": 1.2897214955305881e-05, "loss": 3.1229, "step": 12059 }, { "epoch": 0.42495837627844285, "grad_norm": 1.993402123451233, "learning_rate": 1.28961226078652e-05, "loss": 0.988, "step": 12060 }, { "epoch": 0.4249936132914013, "grad_norm": 10.632491111755371, "learning_rate": 1.289503022270215e-05, "loss": 3.0761, "step": 12061 }, { "epoch": 0.4250288503043597, "grad_norm": 7.600780963897705, "learning_rate": 1.2893937799830964e-05, "loss": 5.7723, "step": 12062 }, { "epoch": 0.4250640873173181, "grad_norm": 1.1255924701690674, "learning_rate": 1.289284533926587e-05, "loss": 0.9721, "step": 12063 }, { "epoch": 0.42509932433027653, "grad_norm": 11.704219818115234, "learning_rate": 1.2891752841021095e-05, "loss": 3.3807, "step": 12064 }, { "epoch": 0.4251345613432349, "grad_norm": 1.3783252239227295, "learning_rate": 1.2890660305110872e-05, "loss": 0.9046, "step": 12065 }, { "epoch": 0.42516979835619334, "grad_norm": 30.019405364990234, "learning_rate": 1.288956773154943e-05, "loss": 7.9531, "step": 12066 }, { "epoch": 0.4252050353691518, "grad_norm": 1.1509143114089966, "learning_rate": 1.2888475120350998e-05, "loss": 1.1706, "step": 12067 }, { "epoch": 0.42524027238211015, "grad_norm": 2.3730242252349854, "learning_rate": 1.2887382471529815e-05, "loss": 1.1354, "step": 12068 }, { "epoch": 0.4252755093950686, "grad_norm": 5.1189422607421875, "learning_rate": 1.2886289785100106e-05, "loss": 3.2445, "step": 12069 }, { "epoch": 0.425310746408027, "grad_norm": 1.0787746906280518, "learning_rate": 1.2885197061076102e-05, "loss": 1.0547, "step": 12070 }, { "epoch": 0.4253459834209854, "grad_norm": 2.5489532947540283, "learning_rate": 1.288410429947204e-05, "loss": 1.0343, "step": 12071 }, { "epoch": 0.42538122043394383, "grad_norm": 4.269850254058838, "learning_rate": 1.2883011500302152e-05, "loss": 2.82, "step": 12072 }, { "epoch": 0.4254164574469022, "grad_norm": 5.736845016479492, "learning_rate": 1.288191866358067e-05, "loss": 3.087, "step": 12073 }, { "epoch": 0.42545169445986064, "grad_norm": 0.8388820290565491, "learning_rate": 1.2880825789321833e-05, "loss": 0.8765, "step": 12074 }, { "epoch": 0.4254869314728191, "grad_norm": 1.6316542625427246, "learning_rate": 1.2879732877539874e-05, "loss": 0.7485, "step": 12075 }, { "epoch": 0.42552216848577745, "grad_norm": 34.39154815673828, "learning_rate": 1.2878639928249027e-05, "loss": 3.6426, "step": 12076 }, { "epoch": 0.4255574054987359, "grad_norm": 10.363870620727539, "learning_rate": 1.2877546941463528e-05, "loss": 5.8987, "step": 12077 }, { "epoch": 0.42559264251169426, "grad_norm": 9.009453773498535, "learning_rate": 1.2876453917197612e-05, "loss": 3.0494, "step": 12078 }, { "epoch": 0.4256278795246527, "grad_norm": 6.273680210113525, "learning_rate": 1.287536085546552e-05, "loss": 3.2637, "step": 12079 }, { "epoch": 0.42566311653761113, "grad_norm": 4.422746658325195, "learning_rate": 1.2874267756281484e-05, "loss": 2.9291, "step": 12080 }, { "epoch": 0.4256983535505695, "grad_norm": 7.4584150314331055, "learning_rate": 1.2873174619659742e-05, "loss": 3.4825, "step": 12081 }, { "epoch": 0.42573359056352794, "grad_norm": 7.038222312927246, "learning_rate": 1.287208144561454e-05, "loss": 3.0568, "step": 12082 }, { "epoch": 0.4257688275764863, "grad_norm": 1.704659104347229, "learning_rate": 1.2870988234160108e-05, "loss": 1.003, "step": 12083 }, { "epoch": 0.42580406458944475, "grad_norm": 2.3016345500946045, "learning_rate": 1.2869894985310692e-05, "loss": 1.1682, "step": 12084 }, { "epoch": 0.4258393016024032, "grad_norm": 1.182955265045166, "learning_rate": 1.2868801699080527e-05, "loss": 0.9613, "step": 12085 }, { "epoch": 0.42587453861536156, "grad_norm": 47.60118103027344, "learning_rate": 1.286770837548385e-05, "loss": 3.1908, "step": 12086 }, { "epoch": 0.42590977562832, "grad_norm": 7.63771915435791, "learning_rate": 1.2866615014534912e-05, "loss": 3.5962, "step": 12087 }, { "epoch": 0.4259450126412784, "grad_norm": 3.321946620941162, "learning_rate": 1.2865521616247945e-05, "loss": 2.9045, "step": 12088 }, { "epoch": 0.4259802496542368, "grad_norm": 3.507338047027588, "learning_rate": 1.2864428180637194e-05, "loss": 2.9859, "step": 12089 }, { "epoch": 0.42601548666719524, "grad_norm": 2.0111818313598633, "learning_rate": 1.28633347077169e-05, "loss": 1.0096, "step": 12090 }, { "epoch": 0.4260507236801536, "grad_norm": 1.2658571004867554, "learning_rate": 1.2862241197501311e-05, "loss": 0.9185, "step": 12091 }, { "epoch": 0.42608596069311205, "grad_norm": 6.674522876739502, "learning_rate": 1.2861147650004662e-05, "loss": 2.9912, "step": 12092 }, { "epoch": 0.42612119770607043, "grad_norm": 9.050829887390137, "learning_rate": 1.2860054065241205e-05, "loss": 6.1313, "step": 12093 }, { "epoch": 0.42615643471902886, "grad_norm": 1.4214680194854736, "learning_rate": 1.2858960443225174e-05, "loss": 0.9654, "step": 12094 }, { "epoch": 0.4261916717319873, "grad_norm": 3.769834518432617, "learning_rate": 1.2857866783970822e-05, "loss": 2.9967, "step": 12095 }, { "epoch": 0.4262269087449457, "grad_norm": 5.907123565673828, "learning_rate": 1.2856773087492391e-05, "loss": 2.7486, "step": 12096 }, { "epoch": 0.4262621457579041, "grad_norm": 7.539651393890381, "learning_rate": 1.2855679353804124e-05, "loss": 5.1442, "step": 12097 }, { "epoch": 0.4262973827708625, "grad_norm": 5.492518901824951, "learning_rate": 1.2854585582920274e-05, "loss": 3.1007, "step": 12098 }, { "epoch": 0.4263326197838209, "grad_norm": 21.871660232543945, "learning_rate": 1.285349177485508e-05, "loss": 2.8744, "step": 12099 }, { "epoch": 0.42636785679677935, "grad_norm": 6.938889026641846, "learning_rate": 1.2852397929622792e-05, "loss": 3.3048, "step": 12100 }, { "epoch": 0.42640309380973773, "grad_norm": 4.65501594543457, "learning_rate": 1.2851304047237658e-05, "loss": 3.0146, "step": 12101 }, { "epoch": 0.42643833082269617, "grad_norm": 1.655648112297058, "learning_rate": 1.2850210127713927e-05, "loss": 1.0174, "step": 12102 }, { "epoch": 0.4264735678356546, "grad_norm": 11.230833053588867, "learning_rate": 1.2849116171065845e-05, "loss": 6.8954, "step": 12103 }, { "epoch": 0.426508804848613, "grad_norm": 5.039300918579102, "learning_rate": 1.2848022177307663e-05, "loss": 3.2435, "step": 12104 }, { "epoch": 0.4265440418615714, "grad_norm": 2.258734703063965, "learning_rate": 1.2846928146453629e-05, "loss": 0.8798, "step": 12105 }, { "epoch": 0.4265792788745298, "grad_norm": 2.0292348861694336, "learning_rate": 1.2845834078517994e-05, "loss": 0.8983, "step": 12106 }, { "epoch": 0.4266145158874882, "grad_norm": 11.551526069641113, "learning_rate": 1.2844739973515003e-05, "loss": 3.208, "step": 12107 }, { "epoch": 0.42664975290044665, "grad_norm": 6.354508876800537, "learning_rate": 1.2843645831458917e-05, "loss": 2.843, "step": 12108 }, { "epoch": 0.42668498991340503, "grad_norm": 1.5226168632507324, "learning_rate": 1.2842551652363979e-05, "loss": 0.893, "step": 12109 }, { "epoch": 0.42672022692636347, "grad_norm": 1.513687014579773, "learning_rate": 1.2841457436244444e-05, "loss": 0.9312, "step": 12110 }, { "epoch": 0.42675546393932184, "grad_norm": 5.374491214752197, "learning_rate": 1.2840363183114564e-05, "loss": 3.4591, "step": 12111 }, { "epoch": 0.4267907009522803, "grad_norm": 13.907451629638672, "learning_rate": 1.2839268892988594e-05, "loss": 3.7592, "step": 12112 }, { "epoch": 0.4268259379652387, "grad_norm": 0.891154944896698, "learning_rate": 1.2838174565880781e-05, "loss": 0.8301, "step": 12113 }, { "epoch": 0.4268611749781971, "grad_norm": 7.459696292877197, "learning_rate": 1.2837080201805388e-05, "loss": 5.8765, "step": 12114 }, { "epoch": 0.4268964119911555, "grad_norm": 8.383260726928711, "learning_rate": 1.283598580077666e-05, "loss": 3.2833, "step": 12115 }, { "epoch": 0.4269316490041139, "grad_norm": 17.290359497070312, "learning_rate": 1.2834891362808857e-05, "loss": 5.0925, "step": 12116 }, { "epoch": 0.42696688601707233, "grad_norm": 1.2573785781860352, "learning_rate": 1.283379688791623e-05, "loss": 0.9636, "step": 12117 }, { "epoch": 0.42700212303003077, "grad_norm": 1.7847706079483032, "learning_rate": 1.283270237611304e-05, "loss": 1.0416, "step": 12118 }, { "epoch": 0.42703736004298914, "grad_norm": 1.6362882852554321, "learning_rate": 1.2831607827413542e-05, "loss": 0.8733, "step": 12119 }, { "epoch": 0.4270725970559476, "grad_norm": 1.0737065076828003, "learning_rate": 1.283051324183199e-05, "loss": 0.9185, "step": 12120 }, { "epoch": 0.42710783406890596, "grad_norm": 9.261435508728027, "learning_rate": 1.2829418619382641e-05, "loss": 3.1805, "step": 12121 }, { "epoch": 0.4271430710818644, "grad_norm": 5.6400465965271, "learning_rate": 1.2828323960079757e-05, "loss": 3.5647, "step": 12122 }, { "epoch": 0.4271783080948228, "grad_norm": 5.949159622192383, "learning_rate": 1.2827229263937593e-05, "loss": 4.8903, "step": 12123 }, { "epoch": 0.4272135451077812, "grad_norm": 4.817262649536133, "learning_rate": 1.2826134530970405e-05, "loss": 3.2633, "step": 12124 }, { "epoch": 0.42724878212073963, "grad_norm": 4.225367069244385, "learning_rate": 1.2825039761192457e-05, "loss": 3.2867, "step": 12125 }, { "epoch": 0.427284019133698, "grad_norm": 1.0912960767745972, "learning_rate": 1.2823944954618002e-05, "loss": 0.9955, "step": 12126 }, { "epoch": 0.42731925614665645, "grad_norm": 1.8058171272277832, "learning_rate": 1.2822850111261312e-05, "loss": 0.9415, "step": 12127 }, { "epoch": 0.4273544931596149, "grad_norm": 4.106834411621094, "learning_rate": 1.2821755231136635e-05, "loss": 3.6301, "step": 12128 }, { "epoch": 0.42738973017257326, "grad_norm": 7.736016750335693, "learning_rate": 1.2820660314258236e-05, "loss": 4.879, "step": 12129 }, { "epoch": 0.4274249671855317, "grad_norm": 1.2558236122131348, "learning_rate": 1.2819565360640379e-05, "loss": 0.9873, "step": 12130 }, { "epoch": 0.42746020419849007, "grad_norm": 2.6542980670928955, "learning_rate": 1.2818470370297325e-05, "loss": 0.9683, "step": 12131 }, { "epoch": 0.4274954412114485, "grad_norm": 1.833990454673767, "learning_rate": 1.2817375343243331e-05, "loss": 1.049, "step": 12132 }, { "epoch": 0.42753067822440693, "grad_norm": 1.2193440198898315, "learning_rate": 1.2816280279492668e-05, "loss": 1.1128, "step": 12133 }, { "epoch": 0.4275659152373653, "grad_norm": 1.2830684185028076, "learning_rate": 1.2815185179059595e-05, "loss": 1.0779, "step": 12134 }, { "epoch": 0.42760115225032375, "grad_norm": 1.6495686769485474, "learning_rate": 1.2814090041958376e-05, "loss": 0.9676, "step": 12135 }, { "epoch": 0.4276363892632821, "grad_norm": 1.2026629447937012, "learning_rate": 1.2812994868203275e-05, "loss": 0.9516, "step": 12136 }, { "epoch": 0.42767162627624056, "grad_norm": 1.6206284761428833, "learning_rate": 1.2811899657808557e-05, "loss": 1.0324, "step": 12137 }, { "epoch": 0.427706863289199, "grad_norm": 1.2567893266677856, "learning_rate": 1.281080441078849e-05, "loss": 0.6407, "step": 12138 }, { "epoch": 0.42774210030215737, "grad_norm": 6.588374614715576, "learning_rate": 1.2809709127157336e-05, "loss": 3.4205, "step": 12139 }, { "epoch": 0.4277773373151158, "grad_norm": 1.6813586950302124, "learning_rate": 1.2808613806929361e-05, "loss": 0.9753, "step": 12140 }, { "epoch": 0.42781257432807424, "grad_norm": 14.145906448364258, "learning_rate": 1.2807518450118836e-05, "loss": 3.4227, "step": 12141 }, { "epoch": 0.4278478113410326, "grad_norm": 1.144434928894043, "learning_rate": 1.280642305674003e-05, "loss": 0.9862, "step": 12142 }, { "epoch": 0.42788304835399105, "grad_norm": 6.245482444763184, "learning_rate": 1.2805327626807198e-05, "loss": 5.2941, "step": 12143 }, { "epoch": 0.4279182853669494, "grad_norm": 7.9231367111206055, "learning_rate": 1.2804232160334622e-05, "loss": 5.4355, "step": 12144 }, { "epoch": 0.42795352237990786, "grad_norm": 5.448605060577393, "learning_rate": 1.2803136657336559e-05, "loss": 3.1809, "step": 12145 }, { "epoch": 0.4279887593928663, "grad_norm": 1.0196950435638428, "learning_rate": 1.2802041117827292e-05, "loss": 0.5897, "step": 12146 }, { "epoch": 0.42802399640582467, "grad_norm": 1.2163994312286377, "learning_rate": 1.2800945541821077e-05, "loss": 0.8313, "step": 12147 }, { "epoch": 0.4280592334187831, "grad_norm": 41.15309143066406, "learning_rate": 1.2799849929332191e-05, "loss": 3.2355, "step": 12148 }, { "epoch": 0.4280944704317415, "grad_norm": 1.5118293762207031, "learning_rate": 1.2798754280374901e-05, "loss": 0.7724, "step": 12149 }, { "epoch": 0.4281297074446999, "grad_norm": 1.8819918632507324, "learning_rate": 1.2797658594963485e-05, "loss": 0.7117, "step": 12150 }, { "epoch": 0.42816494445765835, "grad_norm": 3.686521530151367, "learning_rate": 1.2796562873112204e-05, "loss": 2.8505, "step": 12151 }, { "epoch": 0.4282001814706167, "grad_norm": 1.0136643648147583, "learning_rate": 1.2795467114835338e-05, "loss": 1.1051, "step": 12152 }, { "epoch": 0.42823541848357516, "grad_norm": 1.5916827917099, "learning_rate": 1.2794371320147154e-05, "loss": 0.9363, "step": 12153 }, { "epoch": 0.42827065549653354, "grad_norm": 1.6913195848464966, "learning_rate": 1.2793275489061931e-05, "loss": 0.7666, "step": 12154 }, { "epoch": 0.42830589250949197, "grad_norm": 12.815010070800781, "learning_rate": 1.279217962159394e-05, "loss": 3.4641, "step": 12155 }, { "epoch": 0.4283411295224504, "grad_norm": 2.0565595626831055, "learning_rate": 1.279108371775745e-05, "loss": 1.0868, "step": 12156 }, { "epoch": 0.4283763665354088, "grad_norm": 6.587215900421143, "learning_rate": 1.2789987777566739e-05, "loss": 3.5608, "step": 12157 }, { "epoch": 0.4284116035483672, "grad_norm": 2.9733622074127197, "learning_rate": 1.2788891801036086e-05, "loss": 0.8586, "step": 12158 }, { "epoch": 0.4284468405613256, "grad_norm": 1.2026746273040771, "learning_rate": 1.2787795788179757e-05, "loss": 1.2696, "step": 12159 }, { "epoch": 0.428482077574284, "grad_norm": 1.1202670335769653, "learning_rate": 1.2786699739012036e-05, "loss": 0.9905, "step": 12160 }, { "epoch": 0.42851731458724246, "grad_norm": 6.6798624992370605, "learning_rate": 1.2785603653547196e-05, "loss": 3.6915, "step": 12161 }, { "epoch": 0.42855255160020084, "grad_norm": 4.220729827880859, "learning_rate": 1.2784507531799512e-05, "loss": 2.9786, "step": 12162 }, { "epoch": 0.42858778861315927, "grad_norm": 7.822731018066406, "learning_rate": 1.2783411373783265e-05, "loss": 5.8418, "step": 12163 }, { "epoch": 0.42862302562611765, "grad_norm": 1.1208844184875488, "learning_rate": 1.2782315179512725e-05, "loss": 1.0043, "step": 12164 }, { "epoch": 0.4286582626390761, "grad_norm": 3.5386176109313965, "learning_rate": 1.2781218949002182e-05, "loss": 3.271, "step": 12165 }, { "epoch": 0.4286934996520345, "grad_norm": 1.5046857595443726, "learning_rate": 1.2780122682265905e-05, "loss": 0.8877, "step": 12166 }, { "epoch": 0.4287287366649929, "grad_norm": 1.5494352579116821, "learning_rate": 1.2779026379318176e-05, "loss": 0.7432, "step": 12167 }, { "epoch": 0.4287639736779513, "grad_norm": 1.5073318481445312, "learning_rate": 1.2777930040173274e-05, "loss": 1.0351, "step": 12168 }, { "epoch": 0.4287992106909097, "grad_norm": 1.068878412246704, "learning_rate": 1.2776833664845482e-05, "loss": 1.0914, "step": 12169 }, { "epoch": 0.42883444770386814, "grad_norm": 3.4087436199188232, "learning_rate": 1.2775737253349072e-05, "loss": 3.1989, "step": 12170 }, { "epoch": 0.42886968471682657, "grad_norm": 5.191170692443848, "learning_rate": 1.2774640805698336e-05, "loss": 3.4434, "step": 12171 }, { "epoch": 0.42890492172978495, "grad_norm": 1.1391527652740479, "learning_rate": 1.2773544321907547e-05, "loss": 0.9745, "step": 12172 }, { "epoch": 0.4289401587427434, "grad_norm": 1.1812857389450073, "learning_rate": 1.2772447801990993e-05, "loss": 1.128, "step": 12173 }, { "epoch": 0.4289753957557018, "grad_norm": 5.97927188873291, "learning_rate": 1.277135124596295e-05, "loss": 3.2122, "step": 12174 }, { "epoch": 0.4290106327686602, "grad_norm": 1.3718918561935425, "learning_rate": 1.2770254653837708e-05, "loss": 1.1957, "step": 12175 }, { "epoch": 0.4290458697816186, "grad_norm": 4.392268180847168, "learning_rate": 1.2769158025629543e-05, "loss": 3.1137, "step": 12176 }, { "epoch": 0.429081106794577, "grad_norm": 1.8752230405807495, "learning_rate": 1.2768061361352741e-05, "loss": 0.8971, "step": 12177 }, { "epoch": 0.42911634380753544, "grad_norm": 0.9687107801437378, "learning_rate": 1.276696466102159e-05, "loss": 0.8984, "step": 12178 }, { "epoch": 0.42915158082049387, "grad_norm": 1.3716959953308105, "learning_rate": 1.276586792465037e-05, "loss": 0.877, "step": 12179 }, { "epoch": 0.42918681783345225, "grad_norm": 3.0081934928894043, "learning_rate": 1.2764771152253368e-05, "loss": 2.9702, "step": 12180 }, { "epoch": 0.4292220548464107, "grad_norm": 1.1688365936279297, "learning_rate": 1.2763674343844871e-05, "loss": 0.7438, "step": 12181 }, { "epoch": 0.42925729185936906, "grad_norm": 1.5689594745635986, "learning_rate": 1.2762577499439166e-05, "loss": 0.9436, "step": 12182 }, { "epoch": 0.4292925288723275, "grad_norm": 9.230145454406738, "learning_rate": 1.276148061905053e-05, "loss": 7.9323, "step": 12183 }, { "epoch": 0.42932776588528593, "grad_norm": 1.7491447925567627, "learning_rate": 1.2760383702693264e-05, "loss": 0.9388, "step": 12184 }, { "epoch": 0.4293630028982443, "grad_norm": 10.729866027832031, "learning_rate": 1.2759286750381645e-05, "loss": 3.4175, "step": 12185 }, { "epoch": 0.42939823991120274, "grad_norm": 1.5907782316207886, "learning_rate": 1.2758189762129964e-05, "loss": 0.9327, "step": 12186 }, { "epoch": 0.4294334769241611, "grad_norm": 1.4316424131393433, "learning_rate": 1.2757092737952515e-05, "loss": 0.8723, "step": 12187 }, { "epoch": 0.42946871393711955, "grad_norm": 10.478582382202148, "learning_rate": 1.2755995677863577e-05, "loss": 3.235, "step": 12188 }, { "epoch": 0.429503950950078, "grad_norm": 1.2773795127868652, "learning_rate": 1.2754898581877443e-05, "loss": 1.0091, "step": 12189 }, { "epoch": 0.42953918796303636, "grad_norm": 7.277721405029297, "learning_rate": 1.2753801450008409e-05, "loss": 6.0334, "step": 12190 }, { "epoch": 0.4295744249759948, "grad_norm": 0.921821117401123, "learning_rate": 1.2752704282270757e-05, "loss": 1.2202, "step": 12191 }, { "epoch": 0.4296096619889532, "grad_norm": 5.225455284118652, "learning_rate": 1.2751607078678786e-05, "loss": 7.7798, "step": 12192 }, { "epoch": 0.4296448990019116, "grad_norm": 2.6560516357421875, "learning_rate": 1.2750509839246777e-05, "loss": 3.046, "step": 12193 }, { "epoch": 0.42968013601487004, "grad_norm": 1.4055235385894775, "learning_rate": 1.2749412563989027e-05, "loss": 1.0565, "step": 12194 }, { "epoch": 0.4297153730278284, "grad_norm": 27.54376792907715, "learning_rate": 1.2748315252919829e-05, "loss": 3.4737, "step": 12195 }, { "epoch": 0.42975061004078685, "grad_norm": 7.750766754150391, "learning_rate": 1.2747217906053475e-05, "loss": 5.4091, "step": 12196 }, { "epoch": 0.42978584705374523, "grad_norm": 1.2218148708343506, "learning_rate": 1.274612052340426e-05, "loss": 1.1554, "step": 12197 }, { "epoch": 0.42982108406670366, "grad_norm": 1.059657335281372, "learning_rate": 1.2745023104986473e-05, "loss": 1.0513, "step": 12198 }, { "epoch": 0.4298563210796621, "grad_norm": 1.82327401638031, "learning_rate": 1.2743925650814411e-05, "loss": 0.8986, "step": 12199 }, { "epoch": 0.4298915580926205, "grad_norm": 1.8559234142303467, "learning_rate": 1.2742828160902368e-05, "loss": 1.1399, "step": 12200 }, { "epoch": 0.4299267951055789, "grad_norm": 1.1747864484786987, "learning_rate": 1.274173063526464e-05, "loss": 1.0172, "step": 12201 }, { "epoch": 0.4299620321185373, "grad_norm": 3.560542583465576, "learning_rate": 1.2740633073915515e-05, "loss": 2.6761, "step": 12202 }, { "epoch": 0.4299972691314957, "grad_norm": 8.787449836730957, "learning_rate": 1.2739535476869303e-05, "loss": 3.6039, "step": 12203 }, { "epoch": 0.43003250614445415, "grad_norm": 1.4272109270095825, "learning_rate": 1.273843784414029e-05, "loss": 0.8243, "step": 12204 }, { "epoch": 0.43006774315741253, "grad_norm": 15.917983055114746, "learning_rate": 1.2737340175742774e-05, "loss": 4.9447, "step": 12205 }, { "epoch": 0.43010298017037096, "grad_norm": 0.890693187713623, "learning_rate": 1.2736242471691054e-05, "loss": 1.1219, "step": 12206 }, { "epoch": 0.4301382171833294, "grad_norm": 1.5800987482070923, "learning_rate": 1.2735144731999427e-05, "loss": 1.0343, "step": 12207 }, { "epoch": 0.4301734541962878, "grad_norm": 1.044643521308899, "learning_rate": 1.2734046956682194e-05, "loss": 0.63, "step": 12208 }, { "epoch": 0.4302086912092462, "grad_norm": 6.495516777038574, "learning_rate": 1.2732949145753649e-05, "loss": 4.9478, "step": 12209 }, { "epoch": 0.4302439282222046, "grad_norm": 4.332582950592041, "learning_rate": 1.273185129922809e-05, "loss": 3.139, "step": 12210 }, { "epoch": 0.430279165235163, "grad_norm": 5.125439167022705, "learning_rate": 1.2730753417119828e-05, "loss": 3.2067, "step": 12211 }, { "epoch": 0.43031440224812145, "grad_norm": 7.172553539276123, "learning_rate": 1.2729655499443148e-05, "loss": 2.6928, "step": 12212 }, { "epoch": 0.43034963926107983, "grad_norm": 1.3379414081573486, "learning_rate": 1.2728557546212358e-05, "loss": 0.8881, "step": 12213 }, { "epoch": 0.43038487627403826, "grad_norm": 2.463033676147461, "learning_rate": 1.2727459557441762e-05, "loss": 1.0149, "step": 12214 }, { "epoch": 0.43042011328699664, "grad_norm": 5.642643928527832, "learning_rate": 1.2726361533145655e-05, "loss": 3.5154, "step": 12215 }, { "epoch": 0.4304553502999551, "grad_norm": 0.943107008934021, "learning_rate": 1.2725263473338342e-05, "loss": 0.9502, "step": 12216 }, { "epoch": 0.4304905873129135, "grad_norm": 10.591251373291016, "learning_rate": 1.2724165378034125e-05, "loss": 3.3649, "step": 12217 }, { "epoch": 0.4305258243258719, "grad_norm": 8.70412540435791, "learning_rate": 1.2723067247247308e-05, "loss": 3.2658, "step": 12218 }, { "epoch": 0.4305610613388303, "grad_norm": 1.109393835067749, "learning_rate": 1.272196908099219e-05, "loss": 1.0161, "step": 12219 }, { "epoch": 0.4305962983517887, "grad_norm": 4.081793785095215, "learning_rate": 1.2720870879283085e-05, "loss": 3.3553, "step": 12220 }, { "epoch": 0.43063153536474713, "grad_norm": 10.351593971252441, "learning_rate": 1.2719772642134283e-05, "loss": 8.3299, "step": 12221 }, { "epoch": 0.43066677237770556, "grad_norm": 6.03851842880249, "learning_rate": 1.27186743695601e-05, "loss": 2.9249, "step": 12222 }, { "epoch": 0.43070200939066394, "grad_norm": 36.52555847167969, "learning_rate": 1.2717576061574838e-05, "loss": 5.2762, "step": 12223 }, { "epoch": 0.4307372464036224, "grad_norm": 1.4112666845321655, "learning_rate": 1.2716477718192798e-05, "loss": 1.1775, "step": 12224 }, { "epoch": 0.43077248341658075, "grad_norm": 7.268242835998535, "learning_rate": 1.271537933942829e-05, "loss": 3.1503, "step": 12225 }, { "epoch": 0.4308077204295392, "grad_norm": 4.287947654724121, "learning_rate": 1.2714280925295624e-05, "loss": 3.0348, "step": 12226 }, { "epoch": 0.4308429574424976, "grad_norm": 5.076045989990234, "learning_rate": 1.2713182475809101e-05, "loss": 3.0276, "step": 12227 }, { "epoch": 0.430878194455456, "grad_norm": 4.351247787475586, "learning_rate": 1.271208399098303e-05, "loss": 3.0862, "step": 12228 }, { "epoch": 0.43091343146841443, "grad_norm": 0.9579623937606812, "learning_rate": 1.2710985470831719e-05, "loss": 0.9825, "step": 12229 }, { "epoch": 0.4309486684813728, "grad_norm": 4.161073684692383, "learning_rate": 1.2709886915369478e-05, "loss": 2.9269, "step": 12230 }, { "epoch": 0.43098390549433124, "grad_norm": 12.716434478759766, "learning_rate": 1.2708788324610615e-05, "loss": 4.8723, "step": 12231 }, { "epoch": 0.4310191425072897, "grad_norm": 8.429636001586914, "learning_rate": 1.2707689698569439e-05, "loss": 2.5194, "step": 12232 }, { "epoch": 0.43105437952024805, "grad_norm": 1.5769705772399902, "learning_rate": 1.270659103726026e-05, "loss": 1.1658, "step": 12233 }, { "epoch": 0.4310896165332065, "grad_norm": 47.83310317993164, "learning_rate": 1.2705492340697385e-05, "loss": 3.7616, "step": 12234 }, { "epoch": 0.43112485354616487, "grad_norm": 1.1419203281402588, "learning_rate": 1.2704393608895132e-05, "loss": 0.8087, "step": 12235 }, { "epoch": 0.4311600905591233, "grad_norm": 1.2292802333831787, "learning_rate": 1.2703294841867804e-05, "loss": 1.0873, "step": 12236 }, { "epoch": 0.43119532757208173, "grad_norm": 1.0298062562942505, "learning_rate": 1.2702196039629718e-05, "loss": 0.9685, "step": 12237 }, { "epoch": 0.4312305645850401, "grad_norm": 1.1089686155319214, "learning_rate": 1.2701097202195185e-05, "loss": 0.8924, "step": 12238 }, { "epoch": 0.43126580159799854, "grad_norm": 1.0156973600387573, "learning_rate": 1.2699998329578516e-05, "loss": 1.0462, "step": 12239 }, { "epoch": 0.4313010386109569, "grad_norm": 9.62623119354248, "learning_rate": 1.2698899421794026e-05, "loss": 2.2197, "step": 12240 }, { "epoch": 0.43133627562391536, "grad_norm": 8.552310943603516, "learning_rate": 1.2697800478856027e-05, "loss": 5.8055, "step": 12241 }, { "epoch": 0.4313715126368738, "grad_norm": 4.5649261474609375, "learning_rate": 1.2696701500778832e-05, "loss": 2.4906, "step": 12242 }, { "epoch": 0.43140674964983217, "grad_norm": 1.1062684059143066, "learning_rate": 1.2695602487576757e-05, "loss": 1.0331, "step": 12243 }, { "epoch": 0.4314419866627906, "grad_norm": 5.625481128692627, "learning_rate": 1.2694503439264114e-05, "loss": 5.2005, "step": 12244 }, { "epoch": 0.43147722367574903, "grad_norm": 5.921546936035156, "learning_rate": 1.2693404355855223e-05, "loss": 3.0971, "step": 12245 }, { "epoch": 0.4315124606887074, "grad_norm": 4.787866115570068, "learning_rate": 1.2692305237364398e-05, "loss": 3.4482, "step": 12246 }, { "epoch": 0.43154769770166584, "grad_norm": 2.0050132274627686, "learning_rate": 1.2691206083805954e-05, "loss": 0.9997, "step": 12247 }, { "epoch": 0.4315829347146242, "grad_norm": 5.441212177276611, "learning_rate": 1.2690106895194206e-05, "loss": 3.6099, "step": 12248 }, { "epoch": 0.43161817172758266, "grad_norm": 1.910654067993164, "learning_rate": 1.2689007671543476e-05, "loss": 0.7474, "step": 12249 }, { "epoch": 0.4316534087405411, "grad_norm": 1.5957999229431152, "learning_rate": 1.2687908412868075e-05, "loss": 0.8097, "step": 12250 }, { "epoch": 0.43168864575349947, "grad_norm": 1.5142385959625244, "learning_rate": 1.2686809119182326e-05, "loss": 0.9624, "step": 12251 }, { "epoch": 0.4317238827664579, "grad_norm": 6.253934383392334, "learning_rate": 1.2685709790500546e-05, "loss": 2.6406, "step": 12252 }, { "epoch": 0.4317591197794163, "grad_norm": 1.121474266052246, "learning_rate": 1.2684610426837054e-05, "loss": 0.842, "step": 12253 }, { "epoch": 0.4317943567923747, "grad_norm": 15.123774528503418, "learning_rate": 1.268351102820617e-05, "loss": 3.3141, "step": 12254 }, { "epoch": 0.43182959380533315, "grad_norm": 7.560991287231445, "learning_rate": 1.2682411594622215e-05, "loss": 3.627, "step": 12255 }, { "epoch": 0.4318648308182915, "grad_norm": 10.601320266723633, "learning_rate": 1.26813121260995e-05, "loss": 3.5522, "step": 12256 }, { "epoch": 0.43190006783124996, "grad_norm": 1.1283270120620728, "learning_rate": 1.2680212622652361e-05, "loss": 1.111, "step": 12257 }, { "epoch": 0.43193530484420833, "grad_norm": 5.70640230178833, "learning_rate": 1.2679113084295107e-05, "loss": 3.3504, "step": 12258 }, { "epoch": 0.43197054185716677, "grad_norm": 32.18118667602539, "learning_rate": 1.2678013511042064e-05, "loss": 2.4796, "step": 12259 }, { "epoch": 0.4320057788701252, "grad_norm": 2.130143642425537, "learning_rate": 1.2676913902907555e-05, "loss": 0.8471, "step": 12260 }, { "epoch": 0.4320410158830836, "grad_norm": 49.6653938293457, "learning_rate": 1.26758142599059e-05, "loss": 3.9796, "step": 12261 }, { "epoch": 0.432076252896042, "grad_norm": 1.1275677680969238, "learning_rate": 1.2674714582051423e-05, "loss": 1.1925, "step": 12262 }, { "epoch": 0.4321114899090004, "grad_norm": 3.8339006900787354, "learning_rate": 1.267361486935845e-05, "loss": 2.8227, "step": 12263 }, { "epoch": 0.4321467269219588, "grad_norm": 2.5840322971343994, "learning_rate": 1.2672515121841302e-05, "loss": 1.1169, "step": 12264 }, { "epoch": 0.43218196393491726, "grad_norm": 1.5399945974349976, "learning_rate": 1.2671415339514303e-05, "loss": 0.7063, "step": 12265 }, { "epoch": 0.43221720094787563, "grad_norm": 1.3883200883865356, "learning_rate": 1.2670315522391782e-05, "loss": 0.7794, "step": 12266 }, { "epoch": 0.43225243796083407, "grad_norm": 1.6497881412506104, "learning_rate": 1.2669215670488057e-05, "loss": 0.9425, "step": 12267 }, { "epoch": 0.43228767497379245, "grad_norm": 11.563355445861816, "learning_rate": 1.2668115783817463e-05, "loss": 7.8168, "step": 12268 }, { "epoch": 0.4323229119867509, "grad_norm": 14.888330459594727, "learning_rate": 1.2667015862394319e-05, "loss": 5.7634, "step": 12269 }, { "epoch": 0.4323581489997093, "grad_norm": 5.517493724822998, "learning_rate": 1.2665915906232954e-05, "loss": 5.4009, "step": 12270 }, { "epoch": 0.4323933860126677, "grad_norm": 5.194879055023193, "learning_rate": 1.2664815915347693e-05, "loss": 2.7793, "step": 12271 }, { "epoch": 0.4324286230256261, "grad_norm": 1.182214617729187, "learning_rate": 1.2663715889752864e-05, "loss": 0.8973, "step": 12272 }, { "epoch": 0.4324638600385845, "grad_norm": 1.8667641878128052, "learning_rate": 1.26626158294628e-05, "loss": 1.0653, "step": 12273 }, { "epoch": 0.43249909705154294, "grad_norm": 8.512438774108887, "learning_rate": 1.2661515734491823e-05, "loss": 3.7091, "step": 12274 }, { "epoch": 0.43253433406450137, "grad_norm": 4.650157928466797, "learning_rate": 1.2660415604854263e-05, "loss": 2.9175, "step": 12275 }, { "epoch": 0.43256957107745975, "grad_norm": 1.1008896827697754, "learning_rate": 1.2659315440564455e-05, "loss": 0.666, "step": 12276 }, { "epoch": 0.4326048080904182, "grad_norm": 7.8503737449646, "learning_rate": 1.2658215241636724e-05, "loss": 3.3608, "step": 12277 }, { "epoch": 0.4326400451033766, "grad_norm": 11.275334358215332, "learning_rate": 1.26571150080854e-05, "loss": 3.2601, "step": 12278 }, { "epoch": 0.432675282116335, "grad_norm": 4.285036563873291, "learning_rate": 1.2656014739924813e-05, "loss": 3.1953, "step": 12279 }, { "epoch": 0.4327105191292934, "grad_norm": 1.6767421960830688, "learning_rate": 1.2654914437169296e-05, "loss": 1.0274, "step": 12280 }, { "epoch": 0.4327457561422518, "grad_norm": 7.942537307739258, "learning_rate": 1.2653814099833181e-05, "loss": 3.1959, "step": 12281 }, { "epoch": 0.43278099315521024, "grad_norm": 1.4751287698745728, "learning_rate": 1.2652713727930798e-05, "loss": 0.8325, "step": 12282 }, { "epoch": 0.43281623016816867, "grad_norm": 20.6278018951416, "learning_rate": 1.2651613321476483e-05, "loss": 5.0533, "step": 12283 }, { "epoch": 0.43285146718112705, "grad_norm": 2.0050017833709717, "learning_rate": 1.2650512880484567e-05, "loss": 0.8247, "step": 12284 }, { "epoch": 0.4328867041940855, "grad_norm": 48.572383880615234, "learning_rate": 1.2649412404969384e-05, "loss": 2.8459, "step": 12285 }, { "epoch": 0.43292194120704386, "grad_norm": 1.5857734680175781, "learning_rate": 1.2648311894945261e-05, "loss": 0.8729, "step": 12286 }, { "epoch": 0.4329571782200023, "grad_norm": 0.9629765152931213, "learning_rate": 1.2647211350426544e-05, "loss": 0.7649, "step": 12287 }, { "epoch": 0.4329924152329607, "grad_norm": 2.5446386337280273, "learning_rate": 1.2646110771427562e-05, "loss": 0.8548, "step": 12288 }, { "epoch": 0.4330276522459191, "grad_norm": 16.984161376953125, "learning_rate": 1.2645010157962647e-05, "loss": 3.4322, "step": 12289 }, { "epoch": 0.43306288925887754, "grad_norm": 5.695333480834961, "learning_rate": 1.264390951004614e-05, "loss": 3.8174, "step": 12290 }, { "epoch": 0.4330981262718359, "grad_norm": 1.3626424074172974, "learning_rate": 1.2642808827692372e-05, "loss": 0.708, "step": 12291 }, { "epoch": 0.43313336328479435, "grad_norm": 1.4715458154678345, "learning_rate": 1.2641708110915686e-05, "loss": 0.8376, "step": 12292 }, { "epoch": 0.4331686002977528, "grad_norm": 1.3753716945648193, "learning_rate": 1.2640607359730415e-05, "loss": 0.9076, "step": 12293 }, { "epoch": 0.43320383731071116, "grad_norm": 1.488961100578308, "learning_rate": 1.2639506574150893e-05, "loss": 0.9123, "step": 12294 }, { "epoch": 0.4332390743236696, "grad_norm": 1.5381118059158325, "learning_rate": 1.2638405754191468e-05, "loss": 0.9044, "step": 12295 }, { "epoch": 0.43327431133662797, "grad_norm": 1.332572102546692, "learning_rate": 1.2637304899866467e-05, "loss": 1.0996, "step": 12296 }, { "epoch": 0.4333095483495864, "grad_norm": 1.314577341079712, "learning_rate": 1.2636204011190234e-05, "loss": 0.9254, "step": 12297 }, { "epoch": 0.43334478536254484, "grad_norm": 3.206561803817749, "learning_rate": 1.2635103088177108e-05, "loss": 3.1294, "step": 12298 }, { "epoch": 0.4333800223755032, "grad_norm": 22.00928497314453, "learning_rate": 1.263400213084143e-05, "loss": 5.9848, "step": 12299 }, { "epoch": 0.43341525938846165, "grad_norm": 3.9817819595336914, "learning_rate": 1.2632901139197537e-05, "loss": 2.9249, "step": 12300 }, { "epoch": 0.43345049640142, "grad_norm": 2.0621085166931152, "learning_rate": 1.2631800113259774e-05, "loss": 0.8711, "step": 12301 }, { "epoch": 0.43348573341437846, "grad_norm": 5.582346439361572, "learning_rate": 1.2630699053042477e-05, "loss": 2.8252, "step": 12302 }, { "epoch": 0.4335209704273369, "grad_norm": 1.130829095840454, "learning_rate": 1.2629597958559988e-05, "loss": 1.1532, "step": 12303 }, { "epoch": 0.43355620744029527, "grad_norm": 4.94288969039917, "learning_rate": 1.2628496829826657e-05, "loss": 3.3929, "step": 12304 }, { "epoch": 0.4335914444532537, "grad_norm": 11.70991325378418, "learning_rate": 1.2627395666856815e-05, "loss": 3.5759, "step": 12305 }, { "epoch": 0.4336266814662121, "grad_norm": 2.471560001373291, "learning_rate": 1.2626294469664812e-05, "loss": 0.9794, "step": 12306 }, { "epoch": 0.4336619184791705, "grad_norm": 5.114617824554443, "learning_rate": 1.2625193238264988e-05, "loss": 2.8562, "step": 12307 }, { "epoch": 0.43369715549212895, "grad_norm": 1.6537083387374878, "learning_rate": 1.2624091972671686e-05, "loss": 0.9219, "step": 12308 }, { "epoch": 0.4337323925050873, "grad_norm": 1.5755409002304077, "learning_rate": 1.2622990672899254e-05, "loss": 0.9322, "step": 12309 }, { "epoch": 0.43376762951804576, "grad_norm": 1.3185063600540161, "learning_rate": 1.2621889338962032e-05, "loss": 1.0004, "step": 12310 }, { "epoch": 0.4338028665310042, "grad_norm": 1.2720850706100464, "learning_rate": 1.262078797087437e-05, "loss": 1.1201, "step": 12311 }, { "epoch": 0.43383810354396257, "grad_norm": 8.157567024230957, "learning_rate": 1.2619686568650611e-05, "loss": 3.0779, "step": 12312 }, { "epoch": 0.433873340556921, "grad_norm": 4.80421781539917, "learning_rate": 1.2618585132305098e-05, "loss": 3.5118, "step": 12313 }, { "epoch": 0.4339085775698794, "grad_norm": 6.070481777191162, "learning_rate": 1.2617483661852183e-05, "loss": 3.474, "step": 12314 }, { "epoch": 0.4339438145828378, "grad_norm": 1.227662205696106, "learning_rate": 1.2616382157306211e-05, "loss": 1.1063, "step": 12315 }, { "epoch": 0.43397905159579625, "grad_norm": 1.0989153385162354, "learning_rate": 1.2615280618681524e-05, "loss": 0.9247, "step": 12316 }, { "epoch": 0.43401428860875463, "grad_norm": 9.10123062133789, "learning_rate": 1.2614179045992474e-05, "loss": 5.1671, "step": 12317 }, { "epoch": 0.43404952562171306, "grad_norm": 7.038218975067139, "learning_rate": 1.2613077439253412e-05, "loss": 2.4078, "step": 12318 }, { "epoch": 0.43408476263467144, "grad_norm": 6.169441223144531, "learning_rate": 1.2611975798478681e-05, "loss": 5.5041, "step": 12319 }, { "epoch": 0.4341199996476299, "grad_norm": 1.0001518726348877, "learning_rate": 1.2610874123682635e-05, "loss": 1.0016, "step": 12320 }, { "epoch": 0.4341552366605883, "grad_norm": 8.691362380981445, "learning_rate": 1.2609772414879617e-05, "loss": 3.2644, "step": 12321 }, { "epoch": 0.4341904736735467, "grad_norm": 3.5463500022888184, "learning_rate": 1.2608670672083985e-05, "loss": 3.2112, "step": 12322 }, { "epoch": 0.4342257106865051, "grad_norm": 1.1172997951507568, "learning_rate": 1.2607568895310085e-05, "loss": 1.2572, "step": 12323 }, { "epoch": 0.4342609476994635, "grad_norm": 11.924277305603027, "learning_rate": 1.2606467084572266e-05, "loss": 3.0894, "step": 12324 }, { "epoch": 0.43429618471242193, "grad_norm": 13.818333625793457, "learning_rate": 1.2605365239884881e-05, "loss": 3.5014, "step": 12325 }, { "epoch": 0.43433142172538036, "grad_norm": 1.2206945419311523, "learning_rate": 1.260426336126228e-05, "loss": 0.6967, "step": 12326 }, { "epoch": 0.43436665873833874, "grad_norm": 8.359006881713867, "learning_rate": 1.260316144871882e-05, "loss": 3.3115, "step": 12327 }, { "epoch": 0.4344018957512972, "grad_norm": 6.051272869110107, "learning_rate": 1.2602059502268852e-05, "loss": 3.5132, "step": 12328 }, { "epoch": 0.43443713276425555, "grad_norm": 10.385842323303223, "learning_rate": 1.2600957521926727e-05, "loss": 3.4268, "step": 12329 }, { "epoch": 0.434472369777214, "grad_norm": 1.9210094213485718, "learning_rate": 1.2599855507706798e-05, "loss": 0.8809, "step": 12330 }, { "epoch": 0.4345076067901724, "grad_norm": 20.710683822631836, "learning_rate": 1.259875345962342e-05, "loss": 3.1344, "step": 12331 }, { "epoch": 0.4345428438031308, "grad_norm": 2.5046091079711914, "learning_rate": 1.2597651377690945e-05, "loss": 0.9379, "step": 12332 }, { "epoch": 0.43457808081608923, "grad_norm": 4.910887241363525, "learning_rate": 1.2596549261923734e-05, "loss": 3.2548, "step": 12333 }, { "epoch": 0.4346133178290476, "grad_norm": 1.2008243799209595, "learning_rate": 1.2595447112336134e-05, "loss": 1.0914, "step": 12334 }, { "epoch": 0.43464855484200604, "grad_norm": 10.374677658081055, "learning_rate": 1.2594344928942506e-05, "loss": 2.8753, "step": 12335 }, { "epoch": 0.4346837918549645, "grad_norm": 1.4025874137878418, "learning_rate": 1.2593242711757205e-05, "loss": 0.7684, "step": 12336 }, { "epoch": 0.43471902886792285, "grad_norm": 1.1531727313995361, "learning_rate": 1.2592140460794589e-05, "loss": 0.7452, "step": 12337 }, { "epoch": 0.4347542658808813, "grad_norm": 4.550042152404785, "learning_rate": 1.2591038176069011e-05, "loss": 3.2817, "step": 12338 }, { "epoch": 0.43478950289383966, "grad_norm": 1.638821005821228, "learning_rate": 1.258993585759483e-05, "loss": 0.9924, "step": 12339 }, { "epoch": 0.4348247399067981, "grad_norm": 11.707910537719727, "learning_rate": 1.2588833505386408e-05, "loss": 3.4291, "step": 12340 }, { "epoch": 0.43485997691975653, "grad_norm": 5.190701007843018, "learning_rate": 1.2587731119458098e-05, "loss": 3.3426, "step": 12341 }, { "epoch": 0.4348952139327149, "grad_norm": 2.9296445846557617, "learning_rate": 1.2586628699824261e-05, "loss": 0.9571, "step": 12342 }, { "epoch": 0.43493045094567334, "grad_norm": 3.8018581867218018, "learning_rate": 1.2585526246499254e-05, "loss": 3.4653, "step": 12343 }, { "epoch": 0.4349656879586317, "grad_norm": 2.8416171073913574, "learning_rate": 1.2584423759497441e-05, "loss": 0.9521, "step": 12344 }, { "epoch": 0.43500092497159015, "grad_norm": 5.509604454040527, "learning_rate": 1.2583321238833177e-05, "loss": 2.939, "step": 12345 }, { "epoch": 0.4350361619845486, "grad_norm": 13.770487785339355, "learning_rate": 1.2582218684520825e-05, "loss": 5.4241, "step": 12346 }, { "epoch": 0.43507139899750696, "grad_norm": 12.242050170898438, "learning_rate": 1.2581116096574749e-05, "loss": 3.4852, "step": 12347 }, { "epoch": 0.4351066360104654, "grad_norm": 4.864894390106201, "learning_rate": 1.2580013475009303e-05, "loss": 3.0786, "step": 12348 }, { "epoch": 0.43514187302342383, "grad_norm": 11.300769805908203, "learning_rate": 1.2578910819838854e-05, "loss": 5.0811, "step": 12349 }, { "epoch": 0.4351771100363822, "grad_norm": 6.05109977722168, "learning_rate": 1.2577808131077766e-05, "loss": 2.9353, "step": 12350 }, { "epoch": 0.43521234704934064, "grad_norm": 16.640966415405273, "learning_rate": 1.2576705408740394e-05, "loss": 3.6899, "step": 12351 }, { "epoch": 0.435247584062299, "grad_norm": 1.6611047983169556, "learning_rate": 1.257560265284111e-05, "loss": 0.8864, "step": 12352 }, { "epoch": 0.43528282107525745, "grad_norm": 7.758324146270752, "learning_rate": 1.257449986339427e-05, "loss": 3.0144, "step": 12353 }, { "epoch": 0.4353180580882159, "grad_norm": 4.818902492523193, "learning_rate": 1.2573397040414243e-05, "loss": 2.7488, "step": 12354 }, { "epoch": 0.43535329510117426, "grad_norm": 1.7323040962219238, "learning_rate": 1.257229418391539e-05, "loss": 0.7345, "step": 12355 }, { "epoch": 0.4353885321141327, "grad_norm": 6.344490051269531, "learning_rate": 1.2571191293912079e-05, "loss": 5.1701, "step": 12356 }, { "epoch": 0.4354237691270911, "grad_norm": 1.5224783420562744, "learning_rate": 1.2570088370418674e-05, "loss": 0.661, "step": 12357 }, { "epoch": 0.4354590061400495, "grad_norm": 0.9573078155517578, "learning_rate": 1.2568985413449541e-05, "loss": 0.7986, "step": 12358 }, { "epoch": 0.43549424315300794, "grad_norm": 4.8192138671875, "learning_rate": 1.2567882423019045e-05, "loss": 3.4315, "step": 12359 }, { "epoch": 0.4355294801659663, "grad_norm": 1.1922043561935425, "learning_rate": 1.2566779399141553e-05, "loss": 0.74, "step": 12360 }, { "epoch": 0.43556471717892475, "grad_norm": 5.412435054779053, "learning_rate": 1.2565676341831433e-05, "loss": 2.902, "step": 12361 }, { "epoch": 0.43559995419188313, "grad_norm": 5.69755744934082, "learning_rate": 1.2564573251103052e-05, "loss": 2.8783, "step": 12362 }, { "epoch": 0.43563519120484157, "grad_norm": 1.659993290901184, "learning_rate": 1.256347012697078e-05, "loss": 0.8468, "step": 12363 }, { "epoch": 0.4356704282178, "grad_norm": 2.5549755096435547, "learning_rate": 1.2562366969448978e-05, "loss": 1.0347, "step": 12364 }, { "epoch": 0.4357056652307584, "grad_norm": 10.673378944396973, "learning_rate": 1.2561263778552024e-05, "loss": 3.5441, "step": 12365 }, { "epoch": 0.4357409022437168, "grad_norm": 11.105247497558594, "learning_rate": 1.256016055429428e-05, "loss": 7.3944, "step": 12366 }, { "epoch": 0.4357761392566752, "grad_norm": 1.4774723052978516, "learning_rate": 1.2559057296690122e-05, "loss": 0.7979, "step": 12367 }, { "epoch": 0.4358113762696336, "grad_norm": 7.072542667388916, "learning_rate": 1.2557954005753913e-05, "loss": 5.7734, "step": 12368 }, { "epoch": 0.43584661328259205, "grad_norm": 1.0986607074737549, "learning_rate": 1.2556850681500032e-05, "loss": 0.9826, "step": 12369 }, { "epoch": 0.43588185029555043, "grad_norm": 4.266400337219238, "learning_rate": 1.2555747323942839e-05, "loss": 3.3259, "step": 12370 }, { "epoch": 0.43591708730850887, "grad_norm": 33.93312454223633, "learning_rate": 1.2554643933096717e-05, "loss": 3.7557, "step": 12371 }, { "epoch": 0.43595232432146724, "grad_norm": 5.7294230461120605, "learning_rate": 1.2553540508976027e-05, "loss": 2.9225, "step": 12372 }, { "epoch": 0.4359875613344257, "grad_norm": 6.723012447357178, "learning_rate": 1.2552437051595155e-05, "loss": 3.2427, "step": 12373 }, { "epoch": 0.4360227983473841, "grad_norm": 3.0400502681732178, "learning_rate": 1.2551333560968457e-05, "loss": 1.1861, "step": 12374 }, { "epoch": 0.4360580353603425, "grad_norm": 1.072055697441101, "learning_rate": 1.2550230037110319e-05, "loss": 0.7745, "step": 12375 }, { "epoch": 0.4360932723733009, "grad_norm": 2.3195035457611084, "learning_rate": 1.2549126480035107e-05, "loss": 0.7609, "step": 12376 }, { "epoch": 0.4361285093862593, "grad_norm": 13.868542671203613, "learning_rate": 1.25480228897572e-05, "loss": 4.7561, "step": 12377 }, { "epoch": 0.43616374639921773, "grad_norm": 1.9759904146194458, "learning_rate": 1.2546919266290969e-05, "loss": 0.8921, "step": 12378 }, { "epoch": 0.43619898341217617, "grad_norm": 17.214677810668945, "learning_rate": 1.254581560965079e-05, "loss": 2.8538, "step": 12379 }, { "epoch": 0.43623422042513454, "grad_norm": 5.893038272857666, "learning_rate": 1.2544711919851042e-05, "loss": 5.4427, "step": 12380 }, { "epoch": 0.436269457438093, "grad_norm": 2.7268083095550537, "learning_rate": 1.2543608196906092e-05, "loss": 1.0786, "step": 12381 }, { "epoch": 0.4363046944510514, "grad_norm": 1.879703164100647, "learning_rate": 1.2542504440830325e-05, "loss": 0.7489, "step": 12382 }, { "epoch": 0.4363399314640098, "grad_norm": 1.3416712284088135, "learning_rate": 1.254140065163811e-05, "loss": 0.9544, "step": 12383 }, { "epoch": 0.4363751684769682, "grad_norm": 8.662657737731934, "learning_rate": 1.2540296829343831e-05, "loss": 3.0312, "step": 12384 }, { "epoch": 0.4364104054899266, "grad_norm": 1.4466941356658936, "learning_rate": 1.2539192973961863e-05, "loss": 1.0359, "step": 12385 }, { "epoch": 0.43644564250288503, "grad_norm": 33.04090881347656, "learning_rate": 1.253808908550658e-05, "loss": 6.3046, "step": 12386 }, { "epoch": 0.43648087951584347, "grad_norm": 1.6803947687149048, "learning_rate": 1.2536985163992363e-05, "loss": 0.8588, "step": 12387 }, { "epoch": 0.43651611652880185, "grad_norm": 1.1926387548446655, "learning_rate": 1.2535881209433595e-05, "loss": 0.7403, "step": 12388 }, { "epoch": 0.4365513535417603, "grad_norm": 1.3150722980499268, "learning_rate": 1.2534777221844647e-05, "loss": 0.7987, "step": 12389 }, { "epoch": 0.43658659055471866, "grad_norm": 1.2131181955337524, "learning_rate": 1.2533673201239908e-05, "loss": 0.8511, "step": 12390 }, { "epoch": 0.4366218275676771, "grad_norm": 18.952089309692383, "learning_rate": 1.2532569147633746e-05, "loss": 5.9068, "step": 12391 }, { "epoch": 0.4366570645806355, "grad_norm": 1.5352400541305542, "learning_rate": 1.2531465061040555e-05, "loss": 0.8811, "step": 12392 }, { "epoch": 0.4366923015935939, "grad_norm": 13.825718879699707, "learning_rate": 1.2530360941474707e-05, "loss": 5.3509, "step": 12393 }, { "epoch": 0.43672753860655233, "grad_norm": 9.267485618591309, "learning_rate": 1.2529256788950584e-05, "loss": 3.0772, "step": 12394 }, { "epoch": 0.4367627756195107, "grad_norm": 5.616399765014648, "learning_rate": 1.2528152603482569e-05, "loss": 3.4396, "step": 12395 }, { "epoch": 0.43679801263246915, "grad_norm": 3.5904593467712402, "learning_rate": 1.2527048385085044e-05, "loss": 2.8689, "step": 12396 }, { "epoch": 0.4368332496454276, "grad_norm": 1.5224844217300415, "learning_rate": 1.2525944133772391e-05, "loss": 1.0429, "step": 12397 }, { "epoch": 0.43686848665838596, "grad_norm": 1.3841047286987305, "learning_rate": 1.2524839849558998e-05, "loss": 1.0567, "step": 12398 }, { "epoch": 0.4369037236713444, "grad_norm": 1.4950289726257324, "learning_rate": 1.252373553245924e-05, "loss": 0.7883, "step": 12399 }, { "epoch": 0.43693896068430277, "grad_norm": 1.2124598026275635, "learning_rate": 1.252263118248751e-05, "loss": 0.8943, "step": 12400 }, { "epoch": 0.4369741976972612, "grad_norm": 1.4451271295547485, "learning_rate": 1.2521526799658184e-05, "loss": 0.7356, "step": 12401 }, { "epoch": 0.43700943471021964, "grad_norm": 1.7093199491500854, "learning_rate": 1.252042238398565e-05, "loss": 0.9593, "step": 12402 }, { "epoch": 0.437044671723178, "grad_norm": 12.274943351745605, "learning_rate": 1.2519317935484294e-05, "loss": 5.9001, "step": 12403 }, { "epoch": 0.43707990873613645, "grad_norm": 1.2293007373809814, "learning_rate": 1.2518213454168503e-05, "loss": 0.6998, "step": 12404 }, { "epoch": 0.4371151457490948, "grad_norm": 6.4994635581970215, "learning_rate": 1.2517108940052658e-05, "loss": 3.4481, "step": 12405 }, { "epoch": 0.43715038276205326, "grad_norm": 11.866957664489746, "learning_rate": 1.2516004393151148e-05, "loss": 4.9485, "step": 12406 }, { "epoch": 0.4371856197750117, "grad_norm": 1.0437334775924683, "learning_rate": 1.2514899813478366e-05, "loss": 1.2377, "step": 12407 }, { "epoch": 0.43722085678797007, "grad_norm": 2.6792194843292236, "learning_rate": 1.2513795201048686e-05, "loss": 1.0483, "step": 12408 }, { "epoch": 0.4372560938009285, "grad_norm": 33.10273742675781, "learning_rate": 1.251269055587651e-05, "loss": 5.5257, "step": 12409 }, { "epoch": 0.4372913308138869, "grad_norm": 1.5223795175552368, "learning_rate": 1.2511585877976214e-05, "loss": 0.8795, "step": 12410 }, { "epoch": 0.4373265678268453, "grad_norm": 4.014518737792969, "learning_rate": 1.2510481167362195e-05, "loss": 3.1699, "step": 12411 }, { "epoch": 0.43736180483980375, "grad_norm": 4.658275604248047, "learning_rate": 1.250937642404884e-05, "loss": 2.3472, "step": 12412 }, { "epoch": 0.4373970418527621, "grad_norm": 26.405418395996094, "learning_rate": 1.2508271648050535e-05, "loss": 10.6211, "step": 12413 }, { "epoch": 0.43743227886572056, "grad_norm": 15.140278816223145, "learning_rate": 1.2507166839381673e-05, "loss": 6.1774, "step": 12414 }, { "epoch": 0.437467515878679, "grad_norm": 8.724507331848145, "learning_rate": 1.2506061998056644e-05, "loss": 3.5438, "step": 12415 }, { "epoch": 0.43750275289163737, "grad_norm": 9.084307670593262, "learning_rate": 1.2504957124089838e-05, "loss": 3.0365, "step": 12416 }, { "epoch": 0.4375379899045958, "grad_norm": 1.0192588567733765, "learning_rate": 1.2503852217495646e-05, "loss": 0.9616, "step": 12417 }, { "epoch": 0.4375732269175542, "grad_norm": 1.490717887878418, "learning_rate": 1.2502747278288458e-05, "loss": 1.1508, "step": 12418 }, { "epoch": 0.4376084639305126, "grad_norm": 1.4243589639663696, "learning_rate": 1.2501642306482674e-05, "loss": 0.6046, "step": 12419 }, { "epoch": 0.43764370094347105, "grad_norm": 1.4856966733932495, "learning_rate": 1.2500537302092677e-05, "loss": 1.046, "step": 12420 }, { "epoch": 0.4376789379564294, "grad_norm": 6.309030532836914, "learning_rate": 1.2499432265132861e-05, "loss": 3.2802, "step": 12421 }, { "epoch": 0.43771417496938786, "grad_norm": 17.753292083740234, "learning_rate": 1.2498327195617623e-05, "loss": 7.6517, "step": 12422 }, { "epoch": 0.43774941198234624, "grad_norm": 1.4305412769317627, "learning_rate": 1.2497222093561355e-05, "loss": 1.0386, "step": 12423 }, { "epoch": 0.43778464899530467, "grad_norm": 1.6181628704071045, "learning_rate": 1.249611695897845e-05, "loss": 0.8077, "step": 12424 }, { "epoch": 0.4378198860082631, "grad_norm": 4.065464973449707, "learning_rate": 1.2495011791883304e-05, "loss": 2.3202, "step": 12425 }, { "epoch": 0.4378551230212215, "grad_norm": 13.262752532958984, "learning_rate": 1.2493906592290313e-05, "loss": 5.075, "step": 12426 }, { "epoch": 0.4378903600341799, "grad_norm": 1.1591368913650513, "learning_rate": 1.249280136021387e-05, "loss": 0.7901, "step": 12427 }, { "epoch": 0.4379255970471383, "grad_norm": 1.1407867670059204, "learning_rate": 1.2491696095668373e-05, "loss": 0.8135, "step": 12428 }, { "epoch": 0.4379608340600967, "grad_norm": 1.1309189796447754, "learning_rate": 1.2490590798668214e-05, "loss": 1.2362, "step": 12429 }, { "epoch": 0.43799607107305516, "grad_norm": 1.4938688278198242, "learning_rate": 1.2489485469227798e-05, "loss": 1.2267, "step": 12430 }, { "epoch": 0.43803130808601354, "grad_norm": 1.5944499969482422, "learning_rate": 1.2488380107361513e-05, "loss": 0.9765, "step": 12431 }, { "epoch": 0.43806654509897197, "grad_norm": 6.868760585784912, "learning_rate": 1.2487274713083759e-05, "loss": 3.3243, "step": 12432 }, { "epoch": 0.43810178211193035, "grad_norm": 6.954991817474365, "learning_rate": 1.2486169286408937e-05, "loss": 2.6203, "step": 12433 }, { "epoch": 0.4381370191248888, "grad_norm": 1.865916132926941, "learning_rate": 1.2485063827351444e-05, "loss": 1.1117, "step": 12434 }, { "epoch": 0.4381722561378472, "grad_norm": 1.175927996635437, "learning_rate": 1.2483958335925674e-05, "loss": 1.1862, "step": 12435 }, { "epoch": 0.4382074931508056, "grad_norm": 1.215611457824707, "learning_rate": 1.2482852812146036e-05, "loss": 1.0334, "step": 12436 }, { "epoch": 0.438242730163764, "grad_norm": 1.7788331508636475, "learning_rate": 1.248174725602692e-05, "loss": 0.7713, "step": 12437 }, { "epoch": 0.4382779671767224, "grad_norm": 9.076712608337402, "learning_rate": 1.2480641667582732e-05, "loss": 2.8764, "step": 12438 }, { "epoch": 0.43831320418968084, "grad_norm": 46.62179183959961, "learning_rate": 1.2479536046827869e-05, "loss": 5.8988, "step": 12439 }, { "epoch": 0.43834844120263927, "grad_norm": 1.0812582969665527, "learning_rate": 1.2478430393776733e-05, "loss": 1.2143, "step": 12440 }, { "epoch": 0.43838367821559765, "grad_norm": 1.1541982889175415, "learning_rate": 1.2477324708443726e-05, "loss": 1.1543, "step": 12441 }, { "epoch": 0.4384189152285561, "grad_norm": 1.2412114143371582, "learning_rate": 1.2476218990843248e-05, "loss": 1.147, "step": 12442 }, { "epoch": 0.43845415224151446, "grad_norm": 6.72080135345459, "learning_rate": 1.2475113240989702e-05, "loss": 3.5125, "step": 12443 }, { "epoch": 0.4384893892544729, "grad_norm": 9.824628829956055, "learning_rate": 1.2474007458897492e-05, "loss": 6.4168, "step": 12444 }, { "epoch": 0.43852462626743133, "grad_norm": 1.4068140983581543, "learning_rate": 1.2472901644581019e-05, "loss": 0.8224, "step": 12445 }, { "epoch": 0.4385598632803897, "grad_norm": 5.811402797698975, "learning_rate": 1.247179579805469e-05, "loss": 3.1008, "step": 12446 }, { "epoch": 0.43859510029334814, "grad_norm": 1.290330171585083, "learning_rate": 1.2470689919332904e-05, "loss": 0.8039, "step": 12447 }, { "epoch": 0.4386303373063065, "grad_norm": 16.722400665283203, "learning_rate": 1.2469584008430064e-05, "loss": 2.8484, "step": 12448 }, { "epoch": 0.43866557431926495, "grad_norm": 6.1109232902526855, "learning_rate": 1.2468478065360583e-05, "loss": 5.6952, "step": 12449 }, { "epoch": 0.4387008113322234, "grad_norm": 96.9982681274414, "learning_rate": 1.2467372090138856e-05, "loss": 3.9117, "step": 12450 }, { "epoch": 0.43873604834518176, "grad_norm": 12.13008975982666, "learning_rate": 1.2466266082779297e-05, "loss": 2.795, "step": 12451 }, { "epoch": 0.4387712853581402, "grad_norm": 7.39126443862915, "learning_rate": 1.2465160043296305e-05, "loss": 3.2028, "step": 12452 }, { "epoch": 0.43880652237109863, "grad_norm": 1.171652913093567, "learning_rate": 1.246405397170429e-05, "loss": 0.9524, "step": 12453 }, { "epoch": 0.438841759384057, "grad_norm": 9.720264434814453, "learning_rate": 1.246294786801766e-05, "loss": 6.3395, "step": 12454 }, { "epoch": 0.43887699639701544, "grad_norm": 1.1069642305374146, "learning_rate": 1.2461841732250822e-05, "loss": 1.1147, "step": 12455 }, { "epoch": 0.4389122334099738, "grad_norm": 12.996687889099121, "learning_rate": 1.2460735564418177e-05, "loss": 5.6474, "step": 12456 }, { "epoch": 0.43894747042293225, "grad_norm": 1.634874701499939, "learning_rate": 1.245962936453414e-05, "loss": 0.7578, "step": 12457 }, { "epoch": 0.4389827074358907, "grad_norm": 2.330770254135132, "learning_rate": 1.2458523132613119e-05, "loss": 0.8219, "step": 12458 }, { "epoch": 0.43901794444884906, "grad_norm": 9.142078399658203, "learning_rate": 1.2457416868669519e-05, "loss": 6.1172, "step": 12459 }, { "epoch": 0.4390531814618075, "grad_norm": 1.321884036064148, "learning_rate": 1.245631057271775e-05, "loss": 0.8542, "step": 12460 }, { "epoch": 0.4390884184747659, "grad_norm": 2.862377405166626, "learning_rate": 1.2455204244772223e-05, "loss": 0.8198, "step": 12461 }, { "epoch": 0.4391236554877243, "grad_norm": 1.655945897102356, "learning_rate": 1.245409788484735e-05, "loss": 0.7965, "step": 12462 }, { "epoch": 0.43915889250068274, "grad_norm": 0.9915703535079956, "learning_rate": 1.2452991492957539e-05, "loss": 1.1565, "step": 12463 }, { "epoch": 0.4391941295136411, "grad_norm": 1.4208468198776245, "learning_rate": 1.24518850691172e-05, "loss": 0.6831, "step": 12464 }, { "epoch": 0.43922936652659955, "grad_norm": 1.0477205514907837, "learning_rate": 1.2450778613340747e-05, "loss": 0.8628, "step": 12465 }, { "epoch": 0.43926460353955793, "grad_norm": 70.38004302978516, "learning_rate": 1.2449672125642593e-05, "loss": 2.5329, "step": 12466 }, { "epoch": 0.43929984055251636, "grad_norm": 6.122734546661377, "learning_rate": 1.2448565606037141e-05, "loss": 3.6227, "step": 12467 }, { "epoch": 0.4393350775654748, "grad_norm": 8.246452331542969, "learning_rate": 1.2447459054538817e-05, "loss": 3.2134, "step": 12468 }, { "epoch": 0.4393703145784332, "grad_norm": 1.739588737487793, "learning_rate": 1.2446352471162025e-05, "loss": 0.9326, "step": 12469 }, { "epoch": 0.4394055515913916, "grad_norm": 5.130894660949707, "learning_rate": 1.2445245855921178e-05, "loss": 2.7764, "step": 12470 }, { "epoch": 0.43944078860435, "grad_norm": 6.635869026184082, "learning_rate": 1.2444139208830693e-05, "loss": 5.5582, "step": 12471 }, { "epoch": 0.4394760256173084, "grad_norm": 1.2085328102111816, "learning_rate": 1.2443032529904984e-05, "loss": 0.7967, "step": 12472 }, { "epoch": 0.43951126263026685, "grad_norm": 8.579649925231934, "learning_rate": 1.2441925819158465e-05, "loss": 7.6095, "step": 12473 }, { "epoch": 0.43954649964322523, "grad_norm": 0.9361922144889832, "learning_rate": 1.2440819076605553e-05, "loss": 0.695, "step": 12474 }, { "epoch": 0.43958173665618366, "grad_norm": 2.1561012268066406, "learning_rate": 1.2439712302260658e-05, "loss": 0.8272, "step": 12475 }, { "epoch": 0.43961697366914204, "grad_norm": 1.6562286615371704, "learning_rate": 1.2438605496138204e-05, "loss": 1.0314, "step": 12476 }, { "epoch": 0.4396522106821005, "grad_norm": 8.925877571105957, "learning_rate": 1.2437498658252599e-05, "loss": 3.3844, "step": 12477 }, { "epoch": 0.4396874476950589, "grad_norm": 6.464143753051758, "learning_rate": 1.2436391788618266e-05, "loss": 5.2269, "step": 12478 }, { "epoch": 0.4397226847080173, "grad_norm": 1.2948225736618042, "learning_rate": 1.2435284887249616e-05, "loss": 0.8934, "step": 12479 }, { "epoch": 0.4397579217209757, "grad_norm": 1.4143214225769043, "learning_rate": 1.2434177954161072e-05, "loss": 0.9779, "step": 12480 }, { "epoch": 0.4397931587339341, "grad_norm": 2.849231243133545, "learning_rate": 1.2433070989367052e-05, "loss": 3.2619, "step": 12481 }, { "epoch": 0.43982839574689253, "grad_norm": 11.32047176361084, "learning_rate": 1.2431963992881971e-05, "loss": 3.6765, "step": 12482 }, { "epoch": 0.43986363275985096, "grad_norm": 1.5108944177627563, "learning_rate": 1.2430856964720248e-05, "loss": 0.8621, "step": 12483 }, { "epoch": 0.43989886977280934, "grad_norm": 13.181821823120117, "learning_rate": 1.2429749904896305e-05, "loss": 3.273, "step": 12484 }, { "epoch": 0.4399341067857678, "grad_norm": 13.010974884033203, "learning_rate": 1.242864281342456e-05, "loss": 2.8409, "step": 12485 }, { "epoch": 0.4399693437987262, "grad_norm": 6.701216697692871, "learning_rate": 1.242753569031943e-05, "loss": 3.673, "step": 12486 }, { "epoch": 0.4400045808116846, "grad_norm": 2.382111072540283, "learning_rate": 1.2426428535595343e-05, "loss": 1.1576, "step": 12487 }, { "epoch": 0.440039817824643, "grad_norm": 6.660083293914795, "learning_rate": 1.2425321349266713e-05, "loss": 3.3538, "step": 12488 }, { "epoch": 0.4400750548376014, "grad_norm": 1.6624767780303955, "learning_rate": 1.2424214131347965e-05, "loss": 0.9587, "step": 12489 }, { "epoch": 0.44011029185055983, "grad_norm": 1.2009087800979614, "learning_rate": 1.2423106881853517e-05, "loss": 1.1221, "step": 12490 }, { "epoch": 0.44014552886351826, "grad_norm": 4.048017978668213, "learning_rate": 1.2421999600797794e-05, "loss": 2.5731, "step": 12491 }, { "epoch": 0.44018076587647664, "grad_norm": 4.638913631439209, "learning_rate": 1.2420892288195216e-05, "loss": 3.3643, "step": 12492 }, { "epoch": 0.4402160028894351, "grad_norm": 1.8208110332489014, "learning_rate": 1.2419784944060211e-05, "loss": 0.9422, "step": 12493 }, { "epoch": 0.44025123990239345, "grad_norm": 1.051828384399414, "learning_rate": 1.2418677568407196e-05, "loss": 0.7897, "step": 12494 }, { "epoch": 0.4402864769153519, "grad_norm": 1.0847265720367432, "learning_rate": 1.24175701612506e-05, "loss": 1.2863, "step": 12495 }, { "epoch": 0.4403217139283103, "grad_norm": 9.447965621948242, "learning_rate": 1.2416462722604842e-05, "loss": 2.7782, "step": 12496 }, { "epoch": 0.4403569509412687, "grad_norm": 10.22441291809082, "learning_rate": 1.2415355252484352e-05, "loss": 5.2179, "step": 12497 }, { "epoch": 0.44039218795422713, "grad_norm": 10.528676986694336, "learning_rate": 1.241424775090355e-05, "loss": 6.2141, "step": 12498 }, { "epoch": 0.4404274249671855, "grad_norm": 9.987411499023438, "learning_rate": 1.2413140217876865e-05, "loss": 3.4184, "step": 12499 }, { "epoch": 0.44046266198014394, "grad_norm": 17.017772674560547, "learning_rate": 1.241203265341872e-05, "loss": 6.0255, "step": 12500 }, { "epoch": 0.4404978989931024, "grad_norm": 29.285335540771484, "learning_rate": 1.2410925057543544e-05, "loss": 4.2715, "step": 12501 }, { "epoch": 0.44053313600606075, "grad_norm": 1.111765742301941, "learning_rate": 1.240981743026576e-05, "loss": 1.079, "step": 12502 }, { "epoch": 0.4405683730190192, "grad_norm": 8.019392967224121, "learning_rate": 1.2408709771599798e-05, "loss": 5.2403, "step": 12503 }, { "epoch": 0.44060361003197757, "grad_norm": 5.620748996734619, "learning_rate": 1.2407602081560086e-05, "loss": 3.0302, "step": 12504 }, { "epoch": 0.440638847044936, "grad_norm": 4.994743824005127, "learning_rate": 1.2406494360161046e-05, "loss": 3.7565, "step": 12505 }, { "epoch": 0.44067408405789443, "grad_norm": 5.161260604858398, "learning_rate": 1.2405386607417116e-05, "loss": 5.1025, "step": 12506 }, { "epoch": 0.4407093210708528, "grad_norm": 11.355037689208984, "learning_rate": 1.2404278823342715e-05, "loss": 6.1264, "step": 12507 }, { "epoch": 0.44074455808381124, "grad_norm": 5.673516750335693, "learning_rate": 1.2403171007952276e-05, "loss": 3.2503, "step": 12508 }, { "epoch": 0.4407797950967696, "grad_norm": 15.808673858642578, "learning_rate": 1.2402063161260231e-05, "loss": 6.5002, "step": 12509 }, { "epoch": 0.44081503210972806, "grad_norm": 1.4676368236541748, "learning_rate": 1.2400955283281007e-05, "loss": 0.8396, "step": 12510 }, { "epoch": 0.4408502691226865, "grad_norm": 1.3515559434890747, "learning_rate": 1.2399847374029032e-05, "loss": 0.8343, "step": 12511 }, { "epoch": 0.44088550613564487, "grad_norm": 1.075443148612976, "learning_rate": 1.2398739433518742e-05, "loss": 0.9705, "step": 12512 }, { "epoch": 0.4409207431486033, "grad_norm": 7.2238383293151855, "learning_rate": 1.2397631461764563e-05, "loss": 3.0896, "step": 12513 }, { "epoch": 0.4409559801615617, "grad_norm": 1.8587300777435303, "learning_rate": 1.2396523458780934e-05, "loss": 0.9604, "step": 12514 }, { "epoch": 0.4409912171745201, "grad_norm": 8.126031875610352, "learning_rate": 1.2395415424582276e-05, "loss": 5.6797, "step": 12515 }, { "epoch": 0.44102645418747854, "grad_norm": 8.423430442810059, "learning_rate": 1.2394307359183029e-05, "loss": 6.4248, "step": 12516 }, { "epoch": 0.4410616912004369, "grad_norm": 1.150201678276062, "learning_rate": 1.2393199262597624e-05, "loss": 0.8783, "step": 12517 }, { "epoch": 0.44109692821339536, "grad_norm": 4.921455383300781, "learning_rate": 1.2392091134840491e-05, "loss": 3.2135, "step": 12518 }, { "epoch": 0.4411321652263538, "grad_norm": 0.8523688912391663, "learning_rate": 1.2390982975926069e-05, "loss": 0.6696, "step": 12519 }, { "epoch": 0.44116740223931217, "grad_norm": 1.8793596029281616, "learning_rate": 1.2389874785868787e-05, "loss": 1.2403, "step": 12520 }, { "epoch": 0.4412026392522706, "grad_norm": 3.074380874633789, "learning_rate": 1.2388766564683084e-05, "loss": 2.6928, "step": 12521 }, { "epoch": 0.441237876265229, "grad_norm": 10.634817123413086, "learning_rate": 1.2387658312383391e-05, "loss": 5.9813, "step": 12522 }, { "epoch": 0.4412731132781874, "grad_norm": 5.329981327056885, "learning_rate": 1.2386550028984145e-05, "loss": 2.7287, "step": 12523 }, { "epoch": 0.44130835029114585, "grad_norm": 1.7350255250930786, "learning_rate": 1.238544171449978e-05, "loss": 1.0516, "step": 12524 }, { "epoch": 0.4413435873041042, "grad_norm": 1.6444981098175049, "learning_rate": 1.2384333368944734e-05, "loss": 1.0053, "step": 12525 }, { "epoch": 0.44137882431706266, "grad_norm": 2.1977856159210205, "learning_rate": 1.238322499233344e-05, "loss": 0.7387, "step": 12526 }, { "epoch": 0.44141406133002103, "grad_norm": 5.8451642990112305, "learning_rate": 1.2382116584680342e-05, "loss": 3.8582, "step": 12527 }, { "epoch": 0.44144929834297947, "grad_norm": 1.9982662200927734, "learning_rate": 1.2381008145999866e-05, "loss": 0.9386, "step": 12528 }, { "epoch": 0.4414845353559379, "grad_norm": 2.5683159828186035, "learning_rate": 1.2379899676306459e-05, "loss": 0.9876, "step": 12529 }, { "epoch": 0.4415197723688963, "grad_norm": 2.1959478855133057, "learning_rate": 1.2378791175614554e-05, "loss": 0.9973, "step": 12530 }, { "epoch": 0.4415550093818547, "grad_norm": 4.277263164520264, "learning_rate": 1.2377682643938593e-05, "loss": 3.3946, "step": 12531 }, { "epoch": 0.4415902463948131, "grad_norm": 9.46450424194336, "learning_rate": 1.237657408129301e-05, "loss": 3.2671, "step": 12532 }, { "epoch": 0.4416254834077715, "grad_norm": 7.685690879821777, "learning_rate": 1.2375465487692248e-05, "loss": 6.4864, "step": 12533 }, { "epoch": 0.44166072042072996, "grad_norm": 6.036724090576172, "learning_rate": 1.2374356863150747e-05, "loss": 3.1723, "step": 12534 }, { "epoch": 0.44169595743368834, "grad_norm": 5.659098148345947, "learning_rate": 1.2373248207682944e-05, "loss": 5.2134, "step": 12535 }, { "epoch": 0.44173119444664677, "grad_norm": 0.9660853147506714, "learning_rate": 1.2372139521303281e-05, "loss": 0.9295, "step": 12536 }, { "epoch": 0.44176643145960515, "grad_norm": 4.7491021156311035, "learning_rate": 1.2371030804026195e-05, "loss": 2.7889, "step": 12537 }, { "epoch": 0.4418016684725636, "grad_norm": 1.34053635597229, "learning_rate": 1.2369922055866138e-05, "loss": 0.9215, "step": 12538 }, { "epoch": 0.441836905485522, "grad_norm": 5.992857933044434, "learning_rate": 1.236881327683754e-05, "loss": 3.2834, "step": 12539 }, { "epoch": 0.4418721424984804, "grad_norm": 2.269015073776245, "learning_rate": 1.2367704466954847e-05, "loss": 0.8175, "step": 12540 }, { "epoch": 0.4419073795114388, "grad_norm": 5.469919204711914, "learning_rate": 1.23665956262325e-05, "loss": 4.68, "step": 12541 }, { "epoch": 0.4419426165243972, "grad_norm": 9.578274726867676, "learning_rate": 1.236548675468495e-05, "loss": 5.9959, "step": 12542 }, { "epoch": 0.44197785353735564, "grad_norm": 5.39728307723999, "learning_rate": 1.2364377852326626e-05, "loss": 3.5462, "step": 12543 }, { "epoch": 0.44201309055031407, "grad_norm": 1.359054684638977, "learning_rate": 1.2363268919171985e-05, "loss": 0.8452, "step": 12544 }, { "epoch": 0.44204832756327245, "grad_norm": 1.363161325454712, "learning_rate": 1.2362159955235465e-05, "loss": 1.1932, "step": 12545 }, { "epoch": 0.4420835645762309, "grad_norm": 10.835156440734863, "learning_rate": 1.2361050960531507e-05, "loss": 3.5867, "step": 12546 }, { "epoch": 0.44211880158918926, "grad_norm": 3.303719997406006, "learning_rate": 1.2359941935074562e-05, "loss": 3.5016, "step": 12547 }, { "epoch": 0.4421540386021477, "grad_norm": 11.397113800048828, "learning_rate": 1.2358832878879072e-05, "loss": 3.1493, "step": 12548 }, { "epoch": 0.4421892756151061, "grad_norm": 1.9533737897872925, "learning_rate": 1.2357723791959483e-05, "loss": 0.9244, "step": 12549 }, { "epoch": 0.4422245126280645, "grad_norm": 1.1148223876953125, "learning_rate": 1.2356614674330242e-05, "loss": 1.0407, "step": 12550 }, { "epoch": 0.44225974964102294, "grad_norm": 1.475974678993225, "learning_rate": 1.2355505526005792e-05, "loss": 1.0015, "step": 12551 }, { "epoch": 0.4422949866539813, "grad_norm": 18.784767150878906, "learning_rate": 1.2354396347000588e-05, "loss": 2.9329, "step": 12552 }, { "epoch": 0.44233022366693975, "grad_norm": 1.7582931518554688, "learning_rate": 1.2353287137329069e-05, "loss": 0.7319, "step": 12553 }, { "epoch": 0.4423654606798982, "grad_norm": 0.9639822244644165, "learning_rate": 1.2352177897005684e-05, "loss": 0.9212, "step": 12554 }, { "epoch": 0.44240069769285656, "grad_norm": 10.186542510986328, "learning_rate": 1.2351068626044884e-05, "loss": 3.6336, "step": 12555 }, { "epoch": 0.442435934705815, "grad_norm": 3.941377878189087, "learning_rate": 1.2349959324461113e-05, "loss": 3.051, "step": 12556 }, { "epoch": 0.4424711717187734, "grad_norm": 1.9409726858139038, "learning_rate": 1.2348849992268825e-05, "loss": 0.8278, "step": 12557 }, { "epoch": 0.4425064087317318, "grad_norm": 7.6257429122924805, "learning_rate": 1.2347740629482467e-05, "loss": 3.352, "step": 12558 }, { "epoch": 0.44254164574469024, "grad_norm": 1.272795557975769, "learning_rate": 1.2346631236116485e-05, "loss": 0.8819, "step": 12559 }, { "epoch": 0.4425768827576486, "grad_norm": 9.699382781982422, "learning_rate": 1.2345521812185337e-05, "loss": 3.1052, "step": 12560 }, { "epoch": 0.44261211977060705, "grad_norm": 6.567225456237793, "learning_rate": 1.2344412357703467e-05, "loss": 3.4425, "step": 12561 }, { "epoch": 0.4426473567835655, "grad_norm": 4.163601398468018, "learning_rate": 1.2343302872685326e-05, "loss": 3.1466, "step": 12562 }, { "epoch": 0.44268259379652386, "grad_norm": 3.982593059539795, "learning_rate": 1.2342193357145365e-05, "loss": 2.8107, "step": 12563 }, { "epoch": 0.4427178308094823, "grad_norm": 1.6002079248428345, "learning_rate": 1.234108381109804e-05, "loss": 0.8495, "step": 12564 }, { "epoch": 0.44275306782244067, "grad_norm": 1.6406848430633545, "learning_rate": 1.2339974234557798e-05, "loss": 0.9821, "step": 12565 }, { "epoch": 0.4427883048353991, "grad_norm": 1.1408921480178833, "learning_rate": 1.2338864627539095e-05, "loss": 0.9802, "step": 12566 }, { "epoch": 0.44282354184835754, "grad_norm": 8.330063819885254, "learning_rate": 1.2337754990056383e-05, "loss": 3.2203, "step": 12567 }, { "epoch": 0.4428587788613159, "grad_norm": 3.2080001831054688, "learning_rate": 1.2336645322124112e-05, "loss": 2.3694, "step": 12568 }, { "epoch": 0.44289401587427435, "grad_norm": 8.596959114074707, "learning_rate": 1.2335535623756743e-05, "loss": 5.486, "step": 12569 }, { "epoch": 0.4429292528872327, "grad_norm": 3.251210927963257, "learning_rate": 1.2334425894968719e-05, "loss": 2.8525, "step": 12570 }, { "epoch": 0.44296448990019116, "grad_norm": 5.494112491607666, "learning_rate": 1.2333316135774504e-05, "loss": 3.5067, "step": 12571 }, { "epoch": 0.4429997269131496, "grad_norm": 2.093170642852783, "learning_rate": 1.2332206346188547e-05, "loss": 0.855, "step": 12572 }, { "epoch": 0.44303496392610797, "grad_norm": 1.0521844625473022, "learning_rate": 1.2331096526225307e-05, "loss": 0.9743, "step": 12573 }, { "epoch": 0.4430702009390664, "grad_norm": 7.858604431152344, "learning_rate": 1.2329986675899238e-05, "loss": 5.5356, "step": 12574 }, { "epoch": 0.4431054379520248, "grad_norm": 1.4169217348098755, "learning_rate": 1.2328876795224793e-05, "loss": 0.9493, "step": 12575 }, { "epoch": 0.4431406749649832, "grad_norm": 1.5514540672302246, "learning_rate": 1.2327766884216434e-05, "loss": 0.9662, "step": 12576 }, { "epoch": 0.44317591197794165, "grad_norm": 15.689486503601074, "learning_rate": 1.2326656942888615e-05, "loss": 3.2566, "step": 12577 }, { "epoch": 0.44321114899090003, "grad_norm": 4.749606132507324, "learning_rate": 1.232554697125579e-05, "loss": 3.2147, "step": 12578 }, { "epoch": 0.44324638600385846, "grad_norm": 5.692201614379883, "learning_rate": 1.2324436969332423e-05, "loss": 3.1584, "step": 12579 }, { "epoch": 0.44328162301681684, "grad_norm": 1.3542630672454834, "learning_rate": 1.2323326937132967e-05, "loss": 1.1073, "step": 12580 }, { "epoch": 0.4433168600297753, "grad_norm": 3.180695056915283, "learning_rate": 1.2322216874671882e-05, "loss": 2.6398, "step": 12581 }, { "epoch": 0.4433520970427337, "grad_norm": 8.706536293029785, "learning_rate": 1.2321106781963625e-05, "loss": 3.2204, "step": 12582 }, { "epoch": 0.4433873340556921, "grad_norm": 1.3696279525756836, "learning_rate": 1.2319996659022658e-05, "loss": 0.8702, "step": 12583 }, { "epoch": 0.4434225710686505, "grad_norm": 11.982719421386719, "learning_rate": 1.2318886505863436e-05, "loss": 5.457, "step": 12584 }, { "epoch": 0.4434578080816089, "grad_norm": 1.5716702938079834, "learning_rate": 1.2317776322500426e-05, "loss": 1.0839, "step": 12585 }, { "epoch": 0.44349304509456733, "grad_norm": 1.4175176620483398, "learning_rate": 1.231666610894808e-05, "loss": 0.8755, "step": 12586 }, { "epoch": 0.44352828210752576, "grad_norm": 0.9369708299636841, "learning_rate": 1.2315555865220867e-05, "loss": 0.7148, "step": 12587 }, { "epoch": 0.44356351912048414, "grad_norm": 6.570001125335693, "learning_rate": 1.2314445591333243e-05, "loss": 3.1285, "step": 12588 }, { "epoch": 0.4435987561334426, "grad_norm": 2.6306564807891846, "learning_rate": 1.2313335287299668e-05, "loss": 3.1368, "step": 12589 }, { "epoch": 0.443633993146401, "grad_norm": 6.511308670043945, "learning_rate": 1.2312224953134608e-05, "loss": 4.7824, "step": 12590 }, { "epoch": 0.4436692301593594, "grad_norm": 4.949074745178223, "learning_rate": 1.2311114588852522e-05, "loss": 3.3674, "step": 12591 }, { "epoch": 0.4437044671723178, "grad_norm": 13.078186988830566, "learning_rate": 1.2310004194467877e-05, "loss": 2.8229, "step": 12592 }, { "epoch": 0.4437397041852762, "grad_norm": 2.13075852394104, "learning_rate": 1.2308893769995132e-05, "loss": 0.8382, "step": 12593 }, { "epoch": 0.44377494119823463, "grad_norm": 3.7770631313323975, "learning_rate": 1.2307783315448749e-05, "loss": 3.0013, "step": 12594 }, { "epoch": 0.44381017821119306, "grad_norm": 1.6695424318313599, "learning_rate": 1.2306672830843195e-05, "loss": 1.0528, "step": 12595 }, { "epoch": 0.44384541522415144, "grad_norm": 9.391565322875977, "learning_rate": 1.2305562316192938e-05, "loss": 5.4754, "step": 12596 }, { "epoch": 0.4438806522371099, "grad_norm": 1.0421931743621826, "learning_rate": 1.2304451771512433e-05, "loss": 0.8644, "step": 12597 }, { "epoch": 0.44391588925006825, "grad_norm": 1.2377241849899292, "learning_rate": 1.2303341196816154e-05, "loss": 0.5761, "step": 12598 }, { "epoch": 0.4439511262630267, "grad_norm": 6.590122699737549, "learning_rate": 1.2302230592118557e-05, "loss": 3.985, "step": 12599 }, { "epoch": 0.4439863632759851, "grad_norm": 0.9655013680458069, "learning_rate": 1.230111995743412e-05, "loss": 0.8271, "step": 12600 }, { "epoch": 0.4440216002889435, "grad_norm": 19.325414657592773, "learning_rate": 1.23000092927773e-05, "loss": 3.0183, "step": 12601 }, { "epoch": 0.44405683730190193, "grad_norm": 6.274941444396973, "learning_rate": 1.2298898598162563e-05, "loss": 3.2372, "step": 12602 }, { "epoch": 0.4440920743148603, "grad_norm": 6.174764633178711, "learning_rate": 1.2297787873604381e-05, "loss": 3.5048, "step": 12603 }, { "epoch": 0.44412731132781874, "grad_norm": 1.1373112201690674, "learning_rate": 1.229667711911722e-05, "loss": 0.935, "step": 12604 }, { "epoch": 0.4441625483407772, "grad_norm": 8.742526054382324, "learning_rate": 1.2295566334715545e-05, "loss": 5.8229, "step": 12605 }, { "epoch": 0.44419778535373555, "grad_norm": 1.8449817895889282, "learning_rate": 1.2294455520413828e-05, "loss": 0.9398, "step": 12606 }, { "epoch": 0.444233022366694, "grad_norm": 1.1939326524734497, "learning_rate": 1.2293344676226538e-05, "loss": 1.0952, "step": 12607 }, { "epoch": 0.44426825937965236, "grad_norm": 0.8780450820922852, "learning_rate": 1.2292233802168137e-05, "loss": 0.8994, "step": 12608 }, { "epoch": 0.4443034963926108, "grad_norm": 6.133490562438965, "learning_rate": 1.2291122898253101e-05, "loss": 2.9165, "step": 12609 }, { "epoch": 0.44433873340556923, "grad_norm": 1.0839537382125854, "learning_rate": 1.2290011964495894e-05, "loss": 1.3511, "step": 12610 }, { "epoch": 0.4443739704185276, "grad_norm": 13.710150718688965, "learning_rate": 1.2288901000910994e-05, "loss": 5.3125, "step": 12611 }, { "epoch": 0.44440920743148604, "grad_norm": 1.5368115901947021, "learning_rate": 1.2287790007512866e-05, "loss": 1.0045, "step": 12612 }, { "epoch": 0.4444444444444444, "grad_norm": 0.9801442623138428, "learning_rate": 1.228667898431598e-05, "loss": 1.0842, "step": 12613 }, { "epoch": 0.44447968145740285, "grad_norm": 8.641571044921875, "learning_rate": 1.228556793133481e-05, "loss": 5.7332, "step": 12614 }, { "epoch": 0.4445149184703613, "grad_norm": 2.4933671951293945, "learning_rate": 1.2284456848583828e-05, "loss": 2.3846, "step": 12615 }, { "epoch": 0.44455015548331966, "grad_norm": 1.1293140649795532, "learning_rate": 1.22833457360775e-05, "loss": 1.212, "step": 12616 }, { "epoch": 0.4445853924962781, "grad_norm": 1.5086809396743774, "learning_rate": 1.2282234593830307e-05, "loss": 0.853, "step": 12617 }, { "epoch": 0.4446206295092365, "grad_norm": 0.9627047181129456, "learning_rate": 1.2281123421856716e-05, "loss": 1.1071, "step": 12618 }, { "epoch": 0.4446558665221949, "grad_norm": 1.8649882078170776, "learning_rate": 1.2280012220171202e-05, "loss": 1.8076, "step": 12619 }, { "epoch": 0.44469110353515334, "grad_norm": 6.583646774291992, "learning_rate": 1.227890098878824e-05, "loss": 5.3005, "step": 12620 }, { "epoch": 0.4447263405481117, "grad_norm": 4.477213382720947, "learning_rate": 1.22777897277223e-05, "loss": 2.7368, "step": 12621 }, { "epoch": 0.44476157756107015, "grad_norm": 1.5061763525009155, "learning_rate": 1.227667843698786e-05, "loss": 0.8131, "step": 12622 }, { "epoch": 0.4447968145740286, "grad_norm": 1.8883365392684937, "learning_rate": 1.2275567116599394e-05, "loss": 0.7012, "step": 12623 }, { "epoch": 0.44483205158698697, "grad_norm": 51.083377838134766, "learning_rate": 1.2274455766571374e-05, "loss": 2.6653, "step": 12624 }, { "epoch": 0.4448672885999454, "grad_norm": 4.367396354675293, "learning_rate": 1.227334438691828e-05, "loss": 2.8345, "step": 12625 }, { "epoch": 0.4449025256129038, "grad_norm": 1.0931061506271362, "learning_rate": 1.2272232977654586e-05, "loss": 1.01, "step": 12626 }, { "epoch": 0.4449377626258622, "grad_norm": 0.9823723435401917, "learning_rate": 1.2271121538794765e-05, "loss": 1.094, "step": 12627 }, { "epoch": 0.44497299963882064, "grad_norm": 1.3496997356414795, "learning_rate": 1.2270010070353301e-05, "loss": 1.1148, "step": 12628 }, { "epoch": 0.445008236651779, "grad_norm": 4.610360145568848, "learning_rate": 1.2268898572344664e-05, "loss": 3.4376, "step": 12629 }, { "epoch": 0.44504347366473745, "grad_norm": 1.0616693496704102, "learning_rate": 1.2267787044783334e-05, "loss": 0.8676, "step": 12630 }, { "epoch": 0.44507871067769583, "grad_norm": 1.069016933441162, "learning_rate": 1.2266675487683792e-05, "loss": 0.8376, "step": 12631 }, { "epoch": 0.44511394769065427, "grad_norm": 2.601682424545288, "learning_rate": 1.2265563901060509e-05, "loss": 0.7919, "step": 12632 }, { "epoch": 0.4451491847036127, "grad_norm": 5.2513532638549805, "learning_rate": 1.226445228492797e-05, "loss": 3.1649, "step": 12633 }, { "epoch": 0.4451844217165711, "grad_norm": 8.511152267456055, "learning_rate": 1.2263340639300652e-05, "loss": 3.5608, "step": 12634 }, { "epoch": 0.4452196587295295, "grad_norm": 10.695240020751953, "learning_rate": 1.2262228964193031e-05, "loss": 3.0688, "step": 12635 }, { "epoch": 0.4452548957424879, "grad_norm": 7.89973258972168, "learning_rate": 1.2261117259619593e-05, "loss": 5.2901, "step": 12636 }, { "epoch": 0.4452901327554463, "grad_norm": 5.215230464935303, "learning_rate": 1.226000552559481e-05, "loss": 3.2092, "step": 12637 }, { "epoch": 0.44532536976840476, "grad_norm": 7.546919822692871, "learning_rate": 1.2258893762133173e-05, "loss": 3.6703, "step": 12638 }, { "epoch": 0.44536060678136313, "grad_norm": 13.852201461791992, "learning_rate": 1.2257781969249154e-05, "loss": 2.9246, "step": 12639 }, { "epoch": 0.44539584379432157, "grad_norm": 8.653223991394043, "learning_rate": 1.2256670146957237e-05, "loss": 2.9042, "step": 12640 }, { "epoch": 0.44543108080727994, "grad_norm": 31.350975036621094, "learning_rate": 1.2255558295271903e-05, "loss": 6.0472, "step": 12641 }, { "epoch": 0.4454663178202384, "grad_norm": 2.3824353218078613, "learning_rate": 1.2254446414207637e-05, "loss": 1.1536, "step": 12642 }, { "epoch": 0.4455015548331968, "grad_norm": 1.3924851417541504, "learning_rate": 1.2253334503778918e-05, "loss": 1.2159, "step": 12643 }, { "epoch": 0.4455367918461552, "grad_norm": 6.4851298332214355, "learning_rate": 1.225222256400023e-05, "loss": 2.9916, "step": 12644 }, { "epoch": 0.4455720288591136, "grad_norm": 1.1956548690795898, "learning_rate": 1.2251110594886056e-05, "loss": 0.9692, "step": 12645 }, { "epoch": 0.445607265872072, "grad_norm": 0.9412292838096619, "learning_rate": 1.224999859645088e-05, "loss": 0.9566, "step": 12646 }, { "epoch": 0.44564250288503043, "grad_norm": 4.581099987030029, "learning_rate": 1.2248886568709189e-05, "loss": 3.2505, "step": 12647 }, { "epoch": 0.44567773989798887, "grad_norm": 5.187403202056885, "learning_rate": 1.2247774511675458e-05, "loss": 3.707, "step": 12648 }, { "epoch": 0.44571297691094725, "grad_norm": 2.7316129207611084, "learning_rate": 1.2246662425364181e-05, "loss": 0.9507, "step": 12649 }, { "epoch": 0.4457482139239057, "grad_norm": 3.3223679065704346, "learning_rate": 1.224555030978984e-05, "loss": 3.0013, "step": 12650 }, { "epoch": 0.44578345093686406, "grad_norm": 2.157102108001709, "learning_rate": 1.224443816496692e-05, "loss": 0.8374, "step": 12651 }, { "epoch": 0.4458186879498225, "grad_norm": 9.873786926269531, "learning_rate": 1.2243325990909906e-05, "loss": 9.42, "step": 12652 }, { "epoch": 0.4458539249627809, "grad_norm": 1.394594669342041, "learning_rate": 1.2242213787633288e-05, "loss": 0.8224, "step": 12653 }, { "epoch": 0.4458891619757393, "grad_norm": 0.9343805313110352, "learning_rate": 1.2241101555151546e-05, "loss": 0.7261, "step": 12654 }, { "epoch": 0.44592439898869773, "grad_norm": 1.6776036024093628, "learning_rate": 1.2239989293479174e-05, "loss": 0.8369, "step": 12655 }, { "epoch": 0.4459596360016561, "grad_norm": 1.301016092300415, "learning_rate": 1.2238877002630653e-05, "loss": 0.9251, "step": 12656 }, { "epoch": 0.44599487301461455, "grad_norm": 1.9063392877578735, "learning_rate": 1.2237764682620476e-05, "loss": 0.8997, "step": 12657 }, { "epoch": 0.446030110027573, "grad_norm": 18.154191970825195, "learning_rate": 1.2236652333463129e-05, "loss": 5.7929, "step": 12658 }, { "epoch": 0.44606534704053136, "grad_norm": 1.2212477922439575, "learning_rate": 1.22355399551731e-05, "loss": 0.8354, "step": 12659 }, { "epoch": 0.4461005840534898, "grad_norm": 6.031785488128662, "learning_rate": 1.2234427547764877e-05, "loss": 3.0315, "step": 12660 }, { "epoch": 0.4461358210664482, "grad_norm": 3.669987440109253, "learning_rate": 1.2233315111252953e-05, "loss": 2.8577, "step": 12661 }, { "epoch": 0.4461710580794066, "grad_norm": 1.5699422359466553, "learning_rate": 1.2232202645651814e-05, "loss": 0.7535, "step": 12662 }, { "epoch": 0.44620629509236504, "grad_norm": 1.2850412130355835, "learning_rate": 1.223109015097595e-05, "loss": 0.9251, "step": 12663 }, { "epoch": 0.4462415321053234, "grad_norm": 1.6933249235153198, "learning_rate": 1.2229977627239853e-05, "loss": 0.8614, "step": 12664 }, { "epoch": 0.44627676911828185, "grad_norm": 8.061471939086914, "learning_rate": 1.2228865074458015e-05, "loss": 3.2698, "step": 12665 }, { "epoch": 0.4463120061312403, "grad_norm": 1.8427209854125977, "learning_rate": 1.2227752492644927e-05, "loss": 0.9262, "step": 12666 }, { "epoch": 0.44634724314419866, "grad_norm": 3.596423387527466, "learning_rate": 1.2226639881815072e-05, "loss": 3.6383, "step": 12667 }, { "epoch": 0.4463824801571571, "grad_norm": 4.022284507751465, "learning_rate": 1.2225527241982955e-05, "loss": 3.7436, "step": 12668 }, { "epoch": 0.44641771717011547, "grad_norm": 0.9591077566146851, "learning_rate": 1.2224414573163061e-05, "loss": 0.8388, "step": 12669 }, { "epoch": 0.4464529541830739, "grad_norm": 4.897217750549316, "learning_rate": 1.2223301875369884e-05, "loss": 3.6032, "step": 12670 }, { "epoch": 0.44648819119603234, "grad_norm": 3.3768317699432373, "learning_rate": 1.2222189148617915e-05, "loss": 3.1762, "step": 12671 }, { "epoch": 0.4465234282089907, "grad_norm": 1.2970013618469238, "learning_rate": 1.2221076392921652e-05, "loss": 0.9303, "step": 12672 }, { "epoch": 0.44655866522194915, "grad_norm": 1.4380719661712646, "learning_rate": 1.2219963608295585e-05, "loss": 0.9279, "step": 12673 }, { "epoch": 0.4465939022349075, "grad_norm": 0.980915904045105, "learning_rate": 1.2218850794754208e-05, "loss": 0.9903, "step": 12674 }, { "epoch": 0.44662913924786596, "grad_norm": 1.297025442123413, "learning_rate": 1.2217737952312017e-05, "loss": 0.8863, "step": 12675 }, { "epoch": 0.4466643762608244, "grad_norm": 1.3770618438720703, "learning_rate": 1.2216625080983508e-05, "loss": 0.8969, "step": 12676 }, { "epoch": 0.44669961327378277, "grad_norm": 1.6323162317276, "learning_rate": 1.2215512180783175e-05, "loss": 0.9016, "step": 12677 }, { "epoch": 0.4467348502867412, "grad_norm": 11.46345329284668, "learning_rate": 1.2214399251725514e-05, "loss": 5.7971, "step": 12678 }, { "epoch": 0.4467700872996996, "grad_norm": 1.3927946090698242, "learning_rate": 1.221328629382502e-05, "loss": 1.0312, "step": 12679 }, { "epoch": 0.446805324312658, "grad_norm": 4.835305690765381, "learning_rate": 1.2212173307096189e-05, "loss": 3.0936, "step": 12680 }, { "epoch": 0.44684056132561645, "grad_norm": 1.0543855428695679, "learning_rate": 1.2211060291553517e-05, "loss": 1.0896, "step": 12681 }, { "epoch": 0.4468757983385748, "grad_norm": 5.3887763023376465, "learning_rate": 1.2209947247211507e-05, "loss": 3.0152, "step": 12682 }, { "epoch": 0.44691103535153326, "grad_norm": 10.68034553527832, "learning_rate": 1.2208834174084651e-05, "loss": 3.6709, "step": 12683 }, { "epoch": 0.44694627236449164, "grad_norm": 16.509857177734375, "learning_rate": 1.2207721072187449e-05, "loss": 3.635, "step": 12684 }, { "epoch": 0.44698150937745007, "grad_norm": 2.7423040866851807, "learning_rate": 1.22066079415344e-05, "loss": 0.772, "step": 12685 }, { "epoch": 0.4470167463904085, "grad_norm": 2.7019081115722656, "learning_rate": 1.2205494782139997e-05, "loss": 3.3419, "step": 12686 }, { "epoch": 0.4470519834033669, "grad_norm": 5.985245227813721, "learning_rate": 1.220438159401875e-05, "loss": 2.9143, "step": 12687 }, { "epoch": 0.4470872204163253, "grad_norm": 6.821437835693359, "learning_rate": 1.2203268377185148e-05, "loss": 2.7628, "step": 12688 }, { "epoch": 0.4471224574292837, "grad_norm": 7.483639717102051, "learning_rate": 1.2202155131653694e-05, "loss": 3.7268, "step": 12689 }, { "epoch": 0.4471576944422421, "grad_norm": 1.789575219154358, "learning_rate": 1.220104185743889e-05, "loss": 0.9421, "step": 12690 }, { "epoch": 0.44719293145520056, "grad_norm": 40.620914459228516, "learning_rate": 1.2199928554555237e-05, "loss": 5.4519, "step": 12691 }, { "epoch": 0.44722816846815894, "grad_norm": 1.5309737920761108, "learning_rate": 1.2198815223017233e-05, "loss": 0.8045, "step": 12692 }, { "epoch": 0.44726340548111737, "grad_norm": 7.845858097076416, "learning_rate": 1.2197701862839381e-05, "loss": 5.5638, "step": 12693 }, { "epoch": 0.4472986424940758, "grad_norm": 2.1470489501953125, "learning_rate": 1.219658847403618e-05, "loss": 0.8607, "step": 12694 }, { "epoch": 0.4473338795070342, "grad_norm": 1.5178171396255493, "learning_rate": 1.2195475056622139e-05, "loss": 1.0519, "step": 12695 }, { "epoch": 0.4473691165199926, "grad_norm": 3.681678533554077, "learning_rate": 1.2194361610611753e-05, "loss": 3.0127, "step": 12696 }, { "epoch": 0.447404353532951, "grad_norm": 1.2782979011535645, "learning_rate": 1.2193248136019527e-05, "loss": 0.9402, "step": 12697 }, { "epoch": 0.4474395905459094, "grad_norm": 5.547874450683594, "learning_rate": 1.2192134632859965e-05, "loss": 3.3272, "step": 12698 }, { "epoch": 0.44747482755886786, "grad_norm": 9.52474594116211, "learning_rate": 1.2191021101147572e-05, "loss": 3.0681, "step": 12699 }, { "epoch": 0.44751006457182624, "grad_norm": 2.836876630783081, "learning_rate": 1.2189907540896847e-05, "loss": 2.5439, "step": 12700 }, { "epoch": 0.44754530158478467, "grad_norm": 6.244723320007324, "learning_rate": 1.2188793952122297e-05, "loss": 3.8566, "step": 12701 }, { "epoch": 0.44758053859774305, "grad_norm": 2.1230292320251465, "learning_rate": 1.218768033483843e-05, "loss": 1.0071, "step": 12702 }, { "epoch": 0.4476157756107015, "grad_norm": 0.893727719783783, "learning_rate": 1.2186566689059749e-05, "loss": 1.0466, "step": 12703 }, { "epoch": 0.4476510126236599, "grad_norm": 11.164839744567871, "learning_rate": 1.2185453014800754e-05, "loss": 5.3638, "step": 12704 }, { "epoch": 0.4476862496366183, "grad_norm": 12.916336059570312, "learning_rate": 1.2184339312075957e-05, "loss": 3.3743, "step": 12705 }, { "epoch": 0.4477214866495767, "grad_norm": 0.8845587372779846, "learning_rate": 1.2183225580899863e-05, "loss": 1.1178, "step": 12706 }, { "epoch": 0.4477567236625351, "grad_norm": 5.686328411102295, "learning_rate": 1.2182111821286978e-05, "loss": 3.2022, "step": 12707 }, { "epoch": 0.44779196067549354, "grad_norm": 8.386073112487793, "learning_rate": 1.2180998033251805e-05, "loss": 3.2576, "step": 12708 }, { "epoch": 0.447827197688452, "grad_norm": 2.34055233001709, "learning_rate": 1.2179884216808857e-05, "loss": 0.8843, "step": 12709 }, { "epoch": 0.44786243470141035, "grad_norm": 5.5626935958862305, "learning_rate": 1.217877037197264e-05, "loss": 3.449, "step": 12710 }, { "epoch": 0.4478976717143688, "grad_norm": 6.546090602874756, "learning_rate": 1.2177656498757662e-05, "loss": 5.8296, "step": 12711 }, { "epoch": 0.44793290872732716, "grad_norm": 1.9785053730010986, "learning_rate": 1.217654259717843e-05, "loss": 1.0199, "step": 12712 }, { "epoch": 0.4479681457402856, "grad_norm": 9.308172225952148, "learning_rate": 1.2175428667249452e-05, "loss": 5.374, "step": 12713 }, { "epoch": 0.44800338275324403, "grad_norm": 4.641201496124268, "learning_rate": 1.2174314708985242e-05, "loss": 5.501, "step": 12714 }, { "epoch": 0.4480386197662024, "grad_norm": 11.355494499206543, "learning_rate": 1.2173200722400305e-05, "loss": 3.5278, "step": 12715 }, { "epoch": 0.44807385677916084, "grad_norm": 6.613666534423828, "learning_rate": 1.217208670750915e-05, "loss": 2.9131, "step": 12716 }, { "epoch": 0.4481090937921192, "grad_norm": 11.172562599182129, "learning_rate": 1.2170972664326288e-05, "loss": 3.4489, "step": 12717 }, { "epoch": 0.44814433080507765, "grad_norm": 1.4179329872131348, "learning_rate": 1.2169858592866233e-05, "loss": 0.981, "step": 12718 }, { "epoch": 0.4481795678180361, "grad_norm": 1.2579855918884277, "learning_rate": 1.2168744493143491e-05, "loss": 0.6502, "step": 12719 }, { "epoch": 0.44821480483099446, "grad_norm": 1.7345856428146362, "learning_rate": 1.216763036517258e-05, "loss": 1.0214, "step": 12720 }, { "epoch": 0.4482500418439529, "grad_norm": 8.756422996520996, "learning_rate": 1.2166516208968002e-05, "loss": 3.2332, "step": 12721 }, { "epoch": 0.4482852788569113, "grad_norm": 5.242745876312256, "learning_rate": 1.216540202454428e-05, "loss": 3.3778, "step": 12722 }, { "epoch": 0.4483205158698697, "grad_norm": 12.036446571350098, "learning_rate": 1.2164287811915916e-05, "loss": 3.0267, "step": 12723 }, { "epoch": 0.44835575288282814, "grad_norm": 16.778247833251953, "learning_rate": 1.2163173571097428e-05, "loss": 3.3927, "step": 12724 }, { "epoch": 0.4483909898957865, "grad_norm": 3.9130163192749023, "learning_rate": 1.2162059302103331e-05, "loss": 2.6537, "step": 12725 }, { "epoch": 0.44842622690874495, "grad_norm": 1.376375436782837, "learning_rate": 1.2160945004948135e-05, "loss": 1.0848, "step": 12726 }, { "epoch": 0.4484614639217034, "grad_norm": 1.3851009607315063, "learning_rate": 1.2159830679646356e-05, "loss": 0.8547, "step": 12727 }, { "epoch": 0.44849670093466176, "grad_norm": 1.2080907821655273, "learning_rate": 1.2158716326212505e-05, "loss": 1.146, "step": 12728 }, { "epoch": 0.4485319379476202, "grad_norm": 23.0056209564209, "learning_rate": 1.2157601944661102e-05, "loss": 3.2068, "step": 12729 }, { "epoch": 0.4485671749605786, "grad_norm": 5.95257568359375, "learning_rate": 1.2156487535006656e-05, "loss": 3.4487, "step": 12730 }, { "epoch": 0.448602411973537, "grad_norm": 10.199026107788086, "learning_rate": 1.2155373097263687e-05, "loss": 3.3464, "step": 12731 }, { "epoch": 0.44863764898649544, "grad_norm": 5.228093147277832, "learning_rate": 1.2154258631446708e-05, "loss": 3.6476, "step": 12732 }, { "epoch": 0.4486728859994538, "grad_norm": 12.336585998535156, "learning_rate": 1.2153144137570237e-05, "loss": 3.042, "step": 12733 }, { "epoch": 0.44870812301241225, "grad_norm": 11.168864250183105, "learning_rate": 1.2152029615648789e-05, "loss": 3.9573, "step": 12734 }, { "epoch": 0.44874336002537063, "grad_norm": 4.675954341888428, "learning_rate": 1.215091506569688e-05, "loss": 2.8793, "step": 12735 }, { "epoch": 0.44877859703832906, "grad_norm": 1.3852297067642212, "learning_rate": 1.2149800487729028e-05, "loss": 0.9973, "step": 12736 }, { "epoch": 0.4488138340512875, "grad_norm": 0.7889742851257324, "learning_rate": 1.2148685881759751e-05, "loss": 0.9589, "step": 12737 }, { "epoch": 0.4488490710642459, "grad_norm": 3.9511773586273193, "learning_rate": 1.2147571247803569e-05, "loss": 2.5554, "step": 12738 }, { "epoch": 0.4488843080772043, "grad_norm": 12.733335494995117, "learning_rate": 1.2146456585874998e-05, "loss": 2.9813, "step": 12739 }, { "epoch": 0.4489195450901627, "grad_norm": 2.60434889793396, "learning_rate": 1.2145341895988552e-05, "loss": 0.8445, "step": 12740 }, { "epoch": 0.4489547821031211, "grad_norm": 5.580162525177002, "learning_rate": 1.2144227178158761e-05, "loss": 2.7254, "step": 12741 }, { "epoch": 0.44899001911607955, "grad_norm": 4.275728702545166, "learning_rate": 1.2143112432400133e-05, "loss": 2.9482, "step": 12742 }, { "epoch": 0.44902525612903793, "grad_norm": 1.970104455947876, "learning_rate": 1.2141997658727194e-05, "loss": 0.8869, "step": 12743 }, { "epoch": 0.44906049314199636, "grad_norm": 15.125993728637695, "learning_rate": 1.214088285715446e-05, "loss": 8.1299, "step": 12744 }, { "epoch": 0.44909573015495474, "grad_norm": 1.752501130104065, "learning_rate": 1.2139768027696459e-05, "loss": 0.8608, "step": 12745 }, { "epoch": 0.4491309671679132, "grad_norm": 11.750956535339355, "learning_rate": 1.2138653170367705e-05, "loss": 5.787, "step": 12746 }, { "epoch": 0.4491662041808716, "grad_norm": 1.9024145603179932, "learning_rate": 1.213753828518272e-05, "loss": 0.9032, "step": 12747 }, { "epoch": 0.44920144119383, "grad_norm": 3.7236685752868652, "learning_rate": 1.2136423372156025e-05, "loss": 2.9559, "step": 12748 }, { "epoch": 0.4492366782067884, "grad_norm": 6.769397735595703, "learning_rate": 1.2135308431302147e-05, "loss": 5.5651, "step": 12749 }, { "epoch": 0.4492719152197468, "grad_norm": 8.785526275634766, "learning_rate": 1.2134193462635603e-05, "loss": 2.6921, "step": 12750 }, { "epoch": 0.44930715223270523, "grad_norm": 5.394495010375977, "learning_rate": 1.2133078466170916e-05, "loss": 3.2682, "step": 12751 }, { "epoch": 0.44934238924566366, "grad_norm": 3.612276077270508, "learning_rate": 1.2131963441922614e-05, "loss": 2.7885, "step": 12752 }, { "epoch": 0.44937762625862204, "grad_norm": 4.969374179840088, "learning_rate": 1.2130848389905214e-05, "loss": 3.2583, "step": 12753 }, { "epoch": 0.4494128632715805, "grad_norm": 15.477245330810547, "learning_rate": 1.2129733310133243e-05, "loss": 5.4433, "step": 12754 }, { "epoch": 0.44944810028453885, "grad_norm": 2.198014974594116, "learning_rate": 1.2128618202621225e-05, "loss": 1.1371, "step": 12755 }, { "epoch": 0.4494833372974973, "grad_norm": 1.1333317756652832, "learning_rate": 1.2127503067383682e-05, "loss": 1.0164, "step": 12756 }, { "epoch": 0.4495185743104557, "grad_norm": 5.9482951164245605, "learning_rate": 1.212638790443514e-05, "loss": 4.1693, "step": 12757 }, { "epoch": 0.4495538113234141, "grad_norm": 1.0988200902938843, "learning_rate": 1.2125272713790129e-05, "loss": 0.663, "step": 12758 }, { "epoch": 0.44958904833637253, "grad_norm": 2.025761604309082, "learning_rate": 1.2124157495463165e-05, "loss": 0.9272, "step": 12759 }, { "epoch": 0.4496242853493309, "grad_norm": 1.3817336559295654, "learning_rate": 1.2123042249468781e-05, "loss": 0.8167, "step": 12760 }, { "epoch": 0.44965952236228934, "grad_norm": 7.293359279632568, "learning_rate": 1.2121926975821503e-05, "loss": 4.9012, "step": 12761 }, { "epoch": 0.4496947593752478, "grad_norm": 2.0870373249053955, "learning_rate": 1.2120811674535856e-05, "loss": 0.83, "step": 12762 }, { "epoch": 0.44972999638820615, "grad_norm": 7.0114593505859375, "learning_rate": 1.2119696345626364e-05, "loss": 3.1874, "step": 12763 }, { "epoch": 0.4497652334011646, "grad_norm": 1.2066707611083984, "learning_rate": 1.2118580989107559e-05, "loss": 0.774, "step": 12764 }, { "epoch": 0.449800470414123, "grad_norm": 1.2679740190505981, "learning_rate": 1.2117465604993965e-05, "loss": 0.9593, "step": 12765 }, { "epoch": 0.4498357074270814, "grad_norm": 6.8103227615356445, "learning_rate": 1.2116350193300113e-05, "loss": 5.5432, "step": 12766 }, { "epoch": 0.44987094444003983, "grad_norm": 1.207704782485962, "learning_rate": 1.2115234754040527e-05, "loss": 0.9973, "step": 12767 }, { "epoch": 0.4499061814529982, "grad_norm": 2.0876567363739014, "learning_rate": 1.2114119287229745e-05, "loss": 0.7425, "step": 12768 }, { "epoch": 0.44994141846595664, "grad_norm": 1.2660638093948364, "learning_rate": 1.2113003792882287e-05, "loss": 0.9149, "step": 12769 }, { "epoch": 0.4499766554789151, "grad_norm": 1.299935221672058, "learning_rate": 1.2111888271012685e-05, "loss": 0.7897, "step": 12770 }, { "epoch": 0.45001189249187346, "grad_norm": 2.0040669441223145, "learning_rate": 1.2110772721635468e-05, "loss": 0.8873, "step": 12771 }, { "epoch": 0.4500471295048319, "grad_norm": 6.742909908294678, "learning_rate": 1.2109657144765167e-05, "loss": 3.6217, "step": 12772 }, { "epoch": 0.45008236651779027, "grad_norm": 11.18929386138916, "learning_rate": 1.2108541540416319e-05, "loss": 2.9424, "step": 12773 }, { "epoch": 0.4501176035307487, "grad_norm": 8.036105155944824, "learning_rate": 1.2107425908603442e-05, "loss": 7.0888, "step": 12774 }, { "epoch": 0.45015284054370713, "grad_norm": 3.409503936767578, "learning_rate": 1.2106310249341077e-05, "loss": 3.2361, "step": 12775 }, { "epoch": 0.4501880775566655, "grad_norm": 1.2144696712493896, "learning_rate": 1.2105194562643754e-05, "loss": 0.8608, "step": 12776 }, { "epoch": 0.45022331456962394, "grad_norm": 7.470377445220947, "learning_rate": 1.2104078848526003e-05, "loss": 3.2723, "step": 12777 }, { "epoch": 0.4502585515825823, "grad_norm": 9.044934272766113, "learning_rate": 1.2102963107002355e-05, "loss": 3.6372, "step": 12778 }, { "epoch": 0.45029378859554076, "grad_norm": 4.862640380859375, "learning_rate": 1.2101847338087346e-05, "loss": 2.6031, "step": 12779 }, { "epoch": 0.4503290256084992, "grad_norm": 1.8480585813522339, "learning_rate": 1.210073154179551e-05, "loss": 0.8445, "step": 12780 }, { "epoch": 0.45036426262145757, "grad_norm": 1.1482734680175781, "learning_rate": 1.2099615718141373e-05, "loss": 0.8578, "step": 12781 }, { "epoch": 0.450399499634416, "grad_norm": 4.5803937911987305, "learning_rate": 1.2098499867139477e-05, "loss": 2.958, "step": 12782 }, { "epoch": 0.4504347366473744, "grad_norm": 4.9006829261779785, "learning_rate": 1.2097383988804352e-05, "loss": 2.9481, "step": 12783 }, { "epoch": 0.4504699736603328, "grad_norm": 10.426933288574219, "learning_rate": 1.2096268083150535e-05, "loss": 4.9753, "step": 12784 }, { "epoch": 0.45050521067329125, "grad_norm": 5.819761276245117, "learning_rate": 1.2095152150192558e-05, "loss": 3.169, "step": 12785 }, { "epoch": 0.4505404476862496, "grad_norm": 1.5581934452056885, "learning_rate": 1.2094036189944955e-05, "loss": 1.1508, "step": 12786 }, { "epoch": 0.45057568469920806, "grad_norm": 3.9298553466796875, "learning_rate": 1.2092920202422267e-05, "loss": 3.982, "step": 12787 }, { "epoch": 0.45061092171216643, "grad_norm": 5.158254146575928, "learning_rate": 1.2091804187639029e-05, "loss": 3.1073, "step": 12788 }, { "epoch": 0.45064615872512487, "grad_norm": 7.1268415451049805, "learning_rate": 1.2090688145609768e-05, "loss": 4.8805, "step": 12789 }, { "epoch": 0.4506813957380833, "grad_norm": 5.960379600524902, "learning_rate": 1.2089572076349035e-05, "loss": 2.7915, "step": 12790 }, { "epoch": 0.4507166327510417, "grad_norm": 2.395131826400757, "learning_rate": 1.2088455979871354e-05, "loss": 1.1258, "step": 12791 }, { "epoch": 0.4507518697640001, "grad_norm": 8.686654090881348, "learning_rate": 1.2087339856191271e-05, "loss": 2.87, "step": 12792 }, { "epoch": 0.4507871067769585, "grad_norm": 0.9689276814460754, "learning_rate": 1.208622370532332e-05, "loss": 1.1345, "step": 12793 }, { "epoch": 0.4508223437899169, "grad_norm": 2.490159034729004, "learning_rate": 1.2085107527282038e-05, "loss": 1.2919, "step": 12794 }, { "epoch": 0.45085758080287536, "grad_norm": 2.5211355686187744, "learning_rate": 1.2083991322081964e-05, "loss": 1.0661, "step": 12795 }, { "epoch": 0.45089281781583374, "grad_norm": 2.0082106590270996, "learning_rate": 1.2082875089737643e-05, "loss": 0.9051, "step": 12796 }, { "epoch": 0.45092805482879217, "grad_norm": 2.1532812118530273, "learning_rate": 1.2081758830263604e-05, "loss": 0.9736, "step": 12797 }, { "epoch": 0.4509632918417506, "grad_norm": 3.8597466945648193, "learning_rate": 1.2080642543674393e-05, "loss": 2.6922, "step": 12798 }, { "epoch": 0.450998528854709, "grad_norm": 8.2821626663208, "learning_rate": 1.2079526229984546e-05, "loss": 5.7226, "step": 12799 }, { "epoch": 0.4510337658676674, "grad_norm": 6.355107307434082, "learning_rate": 1.2078409889208607e-05, "loss": 2.5874, "step": 12800 }, { "epoch": 0.4510690028806258, "grad_norm": 12.799471855163574, "learning_rate": 1.2077293521361113e-05, "loss": 4.6777, "step": 12801 }, { "epoch": 0.4511042398935842, "grad_norm": 1.4916515350341797, "learning_rate": 1.2076177126456605e-05, "loss": 0.9516, "step": 12802 }, { "epoch": 0.45113947690654266, "grad_norm": 1.6481356620788574, "learning_rate": 1.2075060704509626e-05, "loss": 0.9676, "step": 12803 }, { "epoch": 0.45117471391950104, "grad_norm": 9.794415473937988, "learning_rate": 1.2073944255534719e-05, "loss": 5.8972, "step": 12804 }, { "epoch": 0.45120995093245947, "grad_norm": 1.5782344341278076, "learning_rate": 1.2072827779546423e-05, "loss": 0.8695, "step": 12805 }, { "epoch": 0.45124518794541785, "grad_norm": 6.7204108238220215, "learning_rate": 1.2071711276559281e-05, "loss": 3.7183, "step": 12806 }, { "epoch": 0.4512804249583763, "grad_norm": 6.068042278289795, "learning_rate": 1.2070594746587838e-05, "loss": 2.942, "step": 12807 }, { "epoch": 0.4513156619713347, "grad_norm": 4.376129150390625, "learning_rate": 1.206947818964663e-05, "loss": 3.216, "step": 12808 }, { "epoch": 0.4513508989842931, "grad_norm": 3.4616150856018066, "learning_rate": 1.206836160575021e-05, "loss": 2.6717, "step": 12809 }, { "epoch": 0.4513861359972515, "grad_norm": 7.588601112365723, "learning_rate": 1.206724499491311e-05, "loss": 5.5385, "step": 12810 }, { "epoch": 0.4514213730102099, "grad_norm": 1.8341012001037598, "learning_rate": 1.2066128357149887e-05, "loss": 1.0866, "step": 12811 }, { "epoch": 0.45145661002316834, "grad_norm": 6.012998580932617, "learning_rate": 1.2065011692475076e-05, "loss": 3.2223, "step": 12812 }, { "epoch": 0.45149184703612677, "grad_norm": 8.789831161499023, "learning_rate": 1.2063895000903228e-05, "loss": 3.67, "step": 12813 }, { "epoch": 0.45152708404908515, "grad_norm": 1.3582916259765625, "learning_rate": 1.206277828244888e-05, "loss": 0.9107, "step": 12814 }, { "epoch": 0.4515623210620436, "grad_norm": 16.690311431884766, "learning_rate": 1.2061661537126585e-05, "loss": 5.1142, "step": 12815 }, { "epoch": 0.45159755807500196, "grad_norm": 1.2418057918548584, "learning_rate": 1.2060544764950884e-05, "loss": 1.2553, "step": 12816 }, { "epoch": 0.4516327950879604, "grad_norm": 7.77502965927124, "learning_rate": 1.2059427965936328e-05, "loss": 5.3845, "step": 12817 }, { "epoch": 0.4516680321009188, "grad_norm": 5.10621976852417, "learning_rate": 1.2058311140097457e-05, "loss": 4.4554, "step": 12818 }, { "epoch": 0.4517032691138772, "grad_norm": 8.348835945129395, "learning_rate": 1.2057194287448824e-05, "loss": 2.9226, "step": 12819 }, { "epoch": 0.45173850612683564, "grad_norm": 4.785747528076172, "learning_rate": 1.2056077408004971e-05, "loss": 3.1299, "step": 12820 }, { "epoch": 0.451773743139794, "grad_norm": 1.5960478782653809, "learning_rate": 1.2054960501780449e-05, "loss": 0.9535, "step": 12821 }, { "epoch": 0.45180898015275245, "grad_norm": 7.822993755340576, "learning_rate": 1.2053843568789804e-05, "loss": 3.8159, "step": 12822 }, { "epoch": 0.4518442171657109, "grad_norm": 18.891077041625977, "learning_rate": 1.2052726609047586e-05, "loss": 7.5935, "step": 12823 }, { "epoch": 0.45187945417866926, "grad_norm": 2.737621307373047, "learning_rate": 1.2051609622568342e-05, "loss": 3.1523, "step": 12824 }, { "epoch": 0.4519146911916277, "grad_norm": 2.5861499309539795, "learning_rate": 1.205049260936662e-05, "loss": 3.0665, "step": 12825 }, { "epoch": 0.45194992820458607, "grad_norm": 8.369418144226074, "learning_rate": 1.2049375569456974e-05, "loss": 5.465, "step": 12826 }, { "epoch": 0.4519851652175445, "grad_norm": 1.5861965417861938, "learning_rate": 1.2048258502853945e-05, "loss": 0.6576, "step": 12827 }, { "epoch": 0.45202040223050294, "grad_norm": 1.9700038433074951, "learning_rate": 1.2047141409572095e-05, "loss": 1.0152, "step": 12828 }, { "epoch": 0.4520556392434613, "grad_norm": 1.3056323528289795, "learning_rate": 1.2046024289625959e-05, "loss": 0.8861, "step": 12829 }, { "epoch": 0.45209087625641975, "grad_norm": 1.1027748584747314, "learning_rate": 1.2044907143030102e-05, "loss": 0.9405, "step": 12830 }, { "epoch": 0.4521261132693782, "grad_norm": 1.3953577280044556, "learning_rate": 1.204378996979907e-05, "loss": 1.0044, "step": 12831 }, { "epoch": 0.45216135028233656, "grad_norm": 8.693909645080566, "learning_rate": 1.2042672769947407e-05, "loss": 3.55, "step": 12832 }, { "epoch": 0.452196587295295, "grad_norm": 8.588059425354004, "learning_rate": 1.2041555543489675e-05, "loss": 5.5413, "step": 12833 }, { "epoch": 0.45223182430825337, "grad_norm": 4.8573784828186035, "learning_rate": 1.2040438290440423e-05, "loss": 3.1509, "step": 12834 }, { "epoch": 0.4522670613212118, "grad_norm": 6.809389591217041, "learning_rate": 1.20393210108142e-05, "loss": 2.8737, "step": 12835 }, { "epoch": 0.45230229833417024, "grad_norm": 4.039734840393066, "learning_rate": 1.2038203704625564e-05, "loss": 3.4442, "step": 12836 }, { "epoch": 0.4523375353471286, "grad_norm": 4.775275707244873, "learning_rate": 1.2037086371889062e-05, "loss": 3.0769, "step": 12837 }, { "epoch": 0.45237277236008705, "grad_norm": 10.669319152832031, "learning_rate": 1.2035969012619253e-05, "loss": 3.6951, "step": 12838 }, { "epoch": 0.4524080093730454, "grad_norm": 1.1256983280181885, "learning_rate": 1.2034851626830686e-05, "loss": 0.7468, "step": 12839 }, { "epoch": 0.45244324638600386, "grad_norm": 1.1172459125518799, "learning_rate": 1.203373421453792e-05, "loss": 0.8526, "step": 12840 }, { "epoch": 0.4524784833989623, "grad_norm": 5.631953239440918, "learning_rate": 1.2032616775755504e-05, "loss": 2.9542, "step": 12841 }, { "epoch": 0.4525137204119207, "grad_norm": 1.9309182167053223, "learning_rate": 1.2031499310497996e-05, "loss": 0.8036, "step": 12842 }, { "epoch": 0.4525489574248791, "grad_norm": 4.131162643432617, "learning_rate": 1.2030381818779952e-05, "loss": 3.2392, "step": 12843 }, { "epoch": 0.4525841944378375, "grad_norm": 1.5601166486740112, "learning_rate": 1.2029264300615927e-05, "loss": 0.9742, "step": 12844 }, { "epoch": 0.4526194314507959, "grad_norm": 4.441697120666504, "learning_rate": 1.2028146756020473e-05, "loss": 2.6806, "step": 12845 }, { "epoch": 0.45265466846375435, "grad_norm": 1.2414599657058716, "learning_rate": 1.202702918500815e-05, "loss": 1.1429, "step": 12846 }, { "epoch": 0.45268990547671273, "grad_norm": 12.506209373474121, "learning_rate": 1.2025911587593518e-05, "loss": 7.8682, "step": 12847 }, { "epoch": 0.45272514248967116, "grad_norm": 1.218691349029541, "learning_rate": 1.2024793963791121e-05, "loss": 1.056, "step": 12848 }, { "epoch": 0.45276037950262954, "grad_norm": 9.434309959411621, "learning_rate": 1.2023676313615533e-05, "loss": 3.0226, "step": 12849 }, { "epoch": 0.452795616515588, "grad_norm": 4.751672267913818, "learning_rate": 1.20225586370813e-05, "loss": 3.1462, "step": 12850 }, { "epoch": 0.4528308535285464, "grad_norm": 3.5308077335357666, "learning_rate": 1.2021440934202984e-05, "loss": 3.2086, "step": 12851 }, { "epoch": 0.4528660905415048, "grad_norm": 0.8323675990104675, "learning_rate": 1.202032320499514e-05, "loss": 0.836, "step": 12852 }, { "epoch": 0.4529013275544632, "grad_norm": 8.575757026672363, "learning_rate": 1.2019205449472332e-05, "loss": 3.6278, "step": 12853 }, { "epoch": 0.4529365645674216, "grad_norm": 1.5493195056915283, "learning_rate": 1.2018087667649111e-05, "loss": 1.0585, "step": 12854 }, { "epoch": 0.45297180158038003, "grad_norm": 1.4325114488601685, "learning_rate": 1.2016969859540047e-05, "loss": 0.9688, "step": 12855 }, { "epoch": 0.45300703859333846, "grad_norm": 5.685748100280762, "learning_rate": 1.2015852025159689e-05, "loss": 5.1171, "step": 12856 }, { "epoch": 0.45304227560629684, "grad_norm": 1.629084825515747, "learning_rate": 1.2014734164522606e-05, "loss": 1.0977, "step": 12857 }, { "epoch": 0.4530775126192553, "grad_norm": 3.741553544998169, "learning_rate": 1.201361627764335e-05, "loss": 3.3146, "step": 12858 }, { "epoch": 0.45311274963221365, "grad_norm": 3.863584041595459, "learning_rate": 1.2012498364536487e-05, "loss": 2.6945, "step": 12859 }, { "epoch": 0.4531479866451721, "grad_norm": 7.16544771194458, "learning_rate": 1.2011380425216574e-05, "loss": 6.3924, "step": 12860 }, { "epoch": 0.4531832236581305, "grad_norm": 10.807083129882812, "learning_rate": 1.201026245969818e-05, "loss": 5.6939, "step": 12861 }, { "epoch": 0.4532184606710889, "grad_norm": 6.916772365570068, "learning_rate": 1.2009144467995856e-05, "loss": 3.7935, "step": 12862 }, { "epoch": 0.45325369768404733, "grad_norm": 1.5922871828079224, "learning_rate": 1.2008026450124174e-05, "loss": 0.914, "step": 12863 }, { "epoch": 0.4532889346970057, "grad_norm": 4.804394245147705, "learning_rate": 1.2006908406097689e-05, "loss": 3.5728, "step": 12864 }, { "epoch": 0.45332417170996414, "grad_norm": 1.445129156112671, "learning_rate": 1.2005790335930967e-05, "loss": 0.951, "step": 12865 }, { "epoch": 0.4533594087229226, "grad_norm": 9.887616157531738, "learning_rate": 1.200467223963857e-05, "loss": 2.9366, "step": 12866 }, { "epoch": 0.45339464573588095, "grad_norm": 1.0796126127243042, "learning_rate": 1.2003554117235061e-05, "loss": 1.3072, "step": 12867 }, { "epoch": 0.4534298827488394, "grad_norm": 1.7130485773086548, "learning_rate": 1.2002435968735007e-05, "loss": 1.1943, "step": 12868 }, { "epoch": 0.4534651197617978, "grad_norm": 5.8403191566467285, "learning_rate": 1.2001317794152966e-05, "loss": 4.4942, "step": 12869 }, { "epoch": 0.4535003567747562, "grad_norm": 1.5719135999679565, "learning_rate": 1.2000199593503507e-05, "loss": 0.7911, "step": 12870 }, { "epoch": 0.45353559378771463, "grad_norm": 7.355525016784668, "learning_rate": 1.1999081366801195e-05, "loss": 5.6652, "step": 12871 }, { "epoch": 0.453570830800673, "grad_norm": 6.514468193054199, "learning_rate": 1.1997963114060594e-05, "loss": 3.3421, "step": 12872 }, { "epoch": 0.45360606781363144, "grad_norm": 5.799647331237793, "learning_rate": 1.1996844835296266e-05, "loss": 3.2868, "step": 12873 }, { "epoch": 0.4536413048265899, "grad_norm": 6.938753128051758, "learning_rate": 1.1995726530522782e-05, "loss": 5.3959, "step": 12874 }, { "epoch": 0.45367654183954825, "grad_norm": 3.2246413230895996, "learning_rate": 1.1994608199754703e-05, "loss": 3.1295, "step": 12875 }, { "epoch": 0.4537117788525067, "grad_norm": 12.319838523864746, "learning_rate": 1.1993489843006603e-05, "loss": 3.2474, "step": 12876 }, { "epoch": 0.45374701586546506, "grad_norm": 1.444854974746704, "learning_rate": 1.199237146029304e-05, "loss": 0.9666, "step": 12877 }, { "epoch": 0.4537822528784235, "grad_norm": 1.3080681562423706, "learning_rate": 1.1991253051628586e-05, "loss": 0.6877, "step": 12878 }, { "epoch": 0.45381748989138193, "grad_norm": 6.3999104499816895, "learning_rate": 1.1990134617027807e-05, "loss": 4.5816, "step": 12879 }, { "epoch": 0.4538527269043403, "grad_norm": 1.171898603439331, "learning_rate": 1.1989016156505273e-05, "loss": 1.1172, "step": 12880 }, { "epoch": 0.45388796391729874, "grad_norm": 1.7812312841415405, "learning_rate": 1.1987897670075546e-05, "loss": 0.9802, "step": 12881 }, { "epoch": 0.4539232009302571, "grad_norm": 1.8904471397399902, "learning_rate": 1.1986779157753203e-05, "loss": 0.8672, "step": 12882 }, { "epoch": 0.45395843794321555, "grad_norm": 6.038005828857422, "learning_rate": 1.1985660619552807e-05, "loss": 3.1246, "step": 12883 }, { "epoch": 0.453993674956174, "grad_norm": 2.2123823165893555, "learning_rate": 1.1984542055488929e-05, "loss": 0.9564, "step": 12884 }, { "epoch": 0.45402891196913236, "grad_norm": 12.694196701049805, "learning_rate": 1.1983423465576136e-05, "loss": 6.4763, "step": 12885 }, { "epoch": 0.4540641489820908, "grad_norm": 6.087677478790283, "learning_rate": 1.1982304849829001e-05, "loss": 3.4285, "step": 12886 }, { "epoch": 0.4540993859950492, "grad_norm": 1.244570016860962, "learning_rate": 1.1981186208262093e-05, "loss": 1.2601, "step": 12887 }, { "epoch": 0.4541346230080076, "grad_norm": 5.397217750549316, "learning_rate": 1.1980067540889982e-05, "loss": 2.8731, "step": 12888 }, { "epoch": 0.45416986002096604, "grad_norm": 5.720608234405518, "learning_rate": 1.1978948847727236e-05, "loss": 3.0136, "step": 12889 }, { "epoch": 0.4542050970339244, "grad_norm": 1.3195743560791016, "learning_rate": 1.1977830128788432e-05, "loss": 1.1001, "step": 12890 }, { "epoch": 0.45424033404688285, "grad_norm": 3.744155168533325, "learning_rate": 1.197671138408814e-05, "loss": 2.3426, "step": 12891 }, { "epoch": 0.45427557105984123, "grad_norm": 15.335393905639648, "learning_rate": 1.1975592613640929e-05, "loss": 3.6599, "step": 12892 }, { "epoch": 0.45431080807279967, "grad_norm": 1.3793810606002808, "learning_rate": 1.1974473817461373e-05, "loss": 0.9909, "step": 12893 }, { "epoch": 0.4543460450857581, "grad_norm": 1.3511451482772827, "learning_rate": 1.197335499556404e-05, "loss": 1.2125, "step": 12894 }, { "epoch": 0.4543812820987165, "grad_norm": 1.529687523841858, "learning_rate": 1.1972236147963514e-05, "loss": 1.0537, "step": 12895 }, { "epoch": 0.4544165191116749, "grad_norm": 1.5809274911880493, "learning_rate": 1.1971117274674358e-05, "loss": 0.7843, "step": 12896 }, { "epoch": 0.4544517561246333, "grad_norm": 5.066341876983643, "learning_rate": 1.1969998375711147e-05, "loss": 5.0073, "step": 12897 }, { "epoch": 0.4544869931375917, "grad_norm": 1.1810832023620605, "learning_rate": 1.1968879451088457e-05, "loss": 0.9986, "step": 12898 }, { "epoch": 0.45452223015055015, "grad_norm": 0.9938077330589294, "learning_rate": 1.1967760500820864e-05, "loss": 0.8616, "step": 12899 }, { "epoch": 0.45455746716350853, "grad_norm": 6.248518466949463, "learning_rate": 1.1966641524922935e-05, "loss": 3.1604, "step": 12900 }, { "epoch": 0.45459270417646697, "grad_norm": 4.598728179931641, "learning_rate": 1.1965522523409254e-05, "loss": 2.9718, "step": 12901 }, { "epoch": 0.4546279411894254, "grad_norm": 1.574889898300171, "learning_rate": 1.1964403496294388e-05, "loss": 0.9489, "step": 12902 }, { "epoch": 0.4546631782023838, "grad_norm": 2.4052934646606445, "learning_rate": 1.196328444359292e-05, "loss": 0.9413, "step": 12903 }, { "epoch": 0.4546984152153422, "grad_norm": 7.607938289642334, "learning_rate": 1.196216536531942e-05, "loss": 3.7673, "step": 12904 }, { "epoch": 0.4547336522283006, "grad_norm": 1.2178664207458496, "learning_rate": 1.1961046261488465e-05, "loss": 1.1559, "step": 12905 }, { "epoch": 0.454768889241259, "grad_norm": 5.8985981941223145, "learning_rate": 1.1959927132114635e-05, "loss": 3.1357, "step": 12906 }, { "epoch": 0.45480412625421746, "grad_norm": 3.6055526733398438, "learning_rate": 1.1958807977212503e-05, "loss": 2.7093, "step": 12907 }, { "epoch": 0.45483936326717583, "grad_norm": 4.55246639251709, "learning_rate": 1.1957688796796648e-05, "loss": 3.2657, "step": 12908 }, { "epoch": 0.45487460028013427, "grad_norm": 9.972569465637207, "learning_rate": 1.1956569590881646e-05, "loss": 5.7334, "step": 12909 }, { "epoch": 0.45490983729309264, "grad_norm": 3.2161200046539307, "learning_rate": 1.1955450359482076e-05, "loss": 2.4402, "step": 12910 }, { "epoch": 0.4549450743060511, "grad_norm": 20.150014877319336, "learning_rate": 1.1954331102612517e-05, "loss": 3.1829, "step": 12911 }, { "epoch": 0.4549803113190095, "grad_norm": 1.1074506044387817, "learning_rate": 1.1953211820287548e-05, "loss": 0.849, "step": 12912 }, { "epoch": 0.4550155483319679, "grad_norm": 9.192481994628906, "learning_rate": 1.1952092512521742e-05, "loss": 5.9945, "step": 12913 }, { "epoch": 0.4550507853449263, "grad_norm": 1.0338356494903564, "learning_rate": 1.1950973179329686e-05, "loss": 0.9157, "step": 12914 }, { "epoch": 0.4550860223578847, "grad_norm": 6.4602131843566895, "learning_rate": 1.1949853820725954e-05, "loss": 3.2726, "step": 12915 }, { "epoch": 0.45512125937084313, "grad_norm": 7.060700416564941, "learning_rate": 1.1948734436725128e-05, "loss": 3.1657, "step": 12916 }, { "epoch": 0.45515649638380157, "grad_norm": 1.2364519834518433, "learning_rate": 1.1947615027341788e-05, "loss": 1.1141, "step": 12917 }, { "epoch": 0.45519173339675995, "grad_norm": 1.0013574361801147, "learning_rate": 1.1946495592590514e-05, "loss": 0.9281, "step": 12918 }, { "epoch": 0.4552269704097184, "grad_norm": 6.386325836181641, "learning_rate": 1.1945376132485889e-05, "loss": 5.3059, "step": 12919 }, { "epoch": 0.45526220742267676, "grad_norm": 11.177458763122559, "learning_rate": 1.1944256647042492e-05, "loss": 5.67, "step": 12920 }, { "epoch": 0.4552974444356352, "grad_norm": 2.9455714225769043, "learning_rate": 1.1943137136274898e-05, "loss": 2.4862, "step": 12921 }, { "epoch": 0.4553326814485936, "grad_norm": 1.0869083404541016, "learning_rate": 1.1942017600197704e-05, "loss": 0.9846, "step": 12922 }, { "epoch": 0.455367918461552, "grad_norm": 6.7949323654174805, "learning_rate": 1.1940898038825477e-05, "loss": 5.552, "step": 12923 }, { "epoch": 0.45540315547451043, "grad_norm": 1.4235304594039917, "learning_rate": 1.193977845217281e-05, "loss": 0.7496, "step": 12924 }, { "epoch": 0.4554383924874688, "grad_norm": 6.053178787231445, "learning_rate": 1.1938658840254278e-05, "loss": 3.5024, "step": 12925 }, { "epoch": 0.45547362950042725, "grad_norm": 1.4245857000350952, "learning_rate": 1.193753920308447e-05, "loss": 0.8723, "step": 12926 }, { "epoch": 0.4555088665133857, "grad_norm": 5.745840072631836, "learning_rate": 1.1936419540677966e-05, "loss": 3.1223, "step": 12927 }, { "epoch": 0.45554410352634406, "grad_norm": 1.3771181106567383, "learning_rate": 1.193529985304935e-05, "loss": 0.9955, "step": 12928 }, { "epoch": 0.4555793405393025, "grad_norm": 8.25905990600586, "learning_rate": 1.193418014021321e-05, "loss": 3.8981, "step": 12929 }, { "epoch": 0.45561457755226087, "grad_norm": 2.0513317584991455, "learning_rate": 1.1933060402184125e-05, "loss": 0.7474, "step": 12930 }, { "epoch": 0.4556498145652193, "grad_norm": 5.9601664543151855, "learning_rate": 1.1931940638976684e-05, "loss": 2.6932, "step": 12931 }, { "epoch": 0.45568505157817774, "grad_norm": 9.115747451782227, "learning_rate": 1.1930820850605467e-05, "loss": 2.9545, "step": 12932 }, { "epoch": 0.4557202885911361, "grad_norm": 2.1171860694885254, "learning_rate": 1.1929701037085066e-05, "loss": 1.0468, "step": 12933 }, { "epoch": 0.45575552560409455, "grad_norm": 2.0434134006500244, "learning_rate": 1.1928581198430061e-05, "loss": 0.9015, "step": 12934 }, { "epoch": 0.4557907626170529, "grad_norm": 5.339193344116211, "learning_rate": 1.1927461334655039e-05, "loss": 2.7437, "step": 12935 }, { "epoch": 0.45582599963001136, "grad_norm": 7.144775390625, "learning_rate": 1.192634144577459e-05, "loss": 3.1767, "step": 12936 }, { "epoch": 0.4558612366429698, "grad_norm": 4.550444602966309, "learning_rate": 1.1925221531803296e-05, "loss": 3.1811, "step": 12937 }, { "epoch": 0.45589647365592817, "grad_norm": 6.327531337738037, "learning_rate": 1.1924101592755748e-05, "loss": 3.0504, "step": 12938 }, { "epoch": 0.4559317106688866, "grad_norm": 1.3687554597854614, "learning_rate": 1.1922981628646532e-05, "loss": 0.8779, "step": 12939 }, { "epoch": 0.45596694768184504, "grad_norm": 1.1924374103546143, "learning_rate": 1.1921861639490233e-05, "loss": 0.9624, "step": 12940 }, { "epoch": 0.4560021846948034, "grad_norm": 3.3062562942504883, "learning_rate": 1.1920741625301446e-05, "loss": 2.9127, "step": 12941 }, { "epoch": 0.45603742170776185, "grad_norm": 5.5667901039123535, "learning_rate": 1.1919621586094752e-05, "loss": 3.5819, "step": 12942 }, { "epoch": 0.4560726587207202, "grad_norm": 3.759092092514038, "learning_rate": 1.1918501521884741e-05, "loss": 3.0208, "step": 12943 }, { "epoch": 0.45610789573367866, "grad_norm": 1.6834652423858643, "learning_rate": 1.1917381432686004e-05, "loss": 0.9639, "step": 12944 }, { "epoch": 0.4561431327466371, "grad_norm": 7.941773414611816, "learning_rate": 1.191626131851313e-05, "loss": 3.4252, "step": 12945 }, { "epoch": 0.45617836975959547, "grad_norm": 1.725211501121521, "learning_rate": 1.1915141179380707e-05, "loss": 0.8389, "step": 12946 }, { "epoch": 0.4562136067725539, "grad_norm": 9.01289176940918, "learning_rate": 1.1914021015303328e-05, "loss": 3.3161, "step": 12947 }, { "epoch": 0.4562488437855123, "grad_norm": 0.9736882448196411, "learning_rate": 1.191290082629558e-05, "loss": 1.0102, "step": 12948 }, { "epoch": 0.4562840807984707, "grad_norm": 6.3289875984191895, "learning_rate": 1.1911780612372056e-05, "loss": 3.6354, "step": 12949 }, { "epoch": 0.45631931781142915, "grad_norm": 4.518286228179932, "learning_rate": 1.1910660373547349e-05, "loss": 3.292, "step": 12950 }, { "epoch": 0.4563545548243875, "grad_norm": 5.348483085632324, "learning_rate": 1.1909540109836042e-05, "loss": 2.887, "step": 12951 }, { "epoch": 0.45638979183734596, "grad_norm": 1.5830273628234863, "learning_rate": 1.1908419821252735e-05, "loss": 0.9866, "step": 12952 }, { "epoch": 0.45642502885030434, "grad_norm": 7.20939302444458, "learning_rate": 1.1907299507812016e-05, "loss": 3.2289, "step": 12953 }, { "epoch": 0.45646026586326277, "grad_norm": 1.2779244184494019, "learning_rate": 1.1906179169528478e-05, "loss": 0.8344, "step": 12954 }, { "epoch": 0.4564955028762212, "grad_norm": 1.0829308032989502, "learning_rate": 1.1905058806416712e-05, "loss": 0.9314, "step": 12955 }, { "epoch": 0.4565307398891796, "grad_norm": 11.567089080810547, "learning_rate": 1.1903938418491314e-05, "loss": 3.3185, "step": 12956 }, { "epoch": 0.456565976902138, "grad_norm": 1.0882662534713745, "learning_rate": 1.1902818005766875e-05, "loss": 1.129, "step": 12957 }, { "epoch": 0.4566012139150964, "grad_norm": 13.353754997253418, "learning_rate": 1.190169756825799e-05, "loss": 5.6394, "step": 12958 }, { "epoch": 0.4566364509280548, "grad_norm": 13.747659683227539, "learning_rate": 1.190057710597925e-05, "loss": 3.8233, "step": 12959 }, { "epoch": 0.45667168794101326, "grad_norm": 1.3039792776107788, "learning_rate": 1.1899456618945253e-05, "loss": 1.146, "step": 12960 }, { "epoch": 0.45670692495397164, "grad_norm": 2.3524177074432373, "learning_rate": 1.1898336107170592e-05, "loss": 0.8874, "step": 12961 }, { "epoch": 0.45674216196693007, "grad_norm": 6.536332607269287, "learning_rate": 1.189721557066986e-05, "loss": 3.1611, "step": 12962 }, { "epoch": 0.45677739897988845, "grad_norm": 1.682275652885437, "learning_rate": 1.1896095009457654e-05, "loss": 0.7946, "step": 12963 }, { "epoch": 0.4568126359928469, "grad_norm": 1.1967387199401855, "learning_rate": 1.1894974423548567e-05, "loss": 1.0872, "step": 12964 }, { "epoch": 0.4568478730058053, "grad_norm": 1.3913403749465942, "learning_rate": 1.18938538129572e-05, "loss": 1.0612, "step": 12965 }, { "epoch": 0.4568831100187637, "grad_norm": 13.1758394241333, "learning_rate": 1.1892733177698143e-05, "loss": 5.2703, "step": 12966 }, { "epoch": 0.4569183470317221, "grad_norm": 10.52318286895752, "learning_rate": 1.1891612517785997e-05, "loss": 3.6327, "step": 12967 }, { "epoch": 0.4569535840446805, "grad_norm": 4.11875581741333, "learning_rate": 1.1890491833235356e-05, "loss": 3.3309, "step": 12968 }, { "epoch": 0.45698882105763894, "grad_norm": 4.3274431228637695, "learning_rate": 1.188937112406082e-05, "loss": 3.1399, "step": 12969 }, { "epoch": 0.45702405807059737, "grad_norm": 1.2796452045440674, "learning_rate": 1.188825039027698e-05, "loss": 0.8719, "step": 12970 }, { "epoch": 0.45705929508355575, "grad_norm": 1.7207111120224, "learning_rate": 1.1887129631898445e-05, "loss": 0.6934, "step": 12971 }, { "epoch": 0.4570945320965142, "grad_norm": 4.375392913818359, "learning_rate": 1.1886008848939801e-05, "loss": 3.2428, "step": 12972 }, { "epoch": 0.4571297691094726, "grad_norm": 14.339094161987305, "learning_rate": 1.1884888041415655e-05, "loss": 2.9637, "step": 12973 }, { "epoch": 0.457165006122431, "grad_norm": 1.4704127311706543, "learning_rate": 1.1883767209340597e-05, "loss": 0.8283, "step": 12974 }, { "epoch": 0.45720024313538943, "grad_norm": 1.599029541015625, "learning_rate": 1.1882646352729235e-05, "loss": 0.6998, "step": 12975 }, { "epoch": 0.4572354801483478, "grad_norm": 13.867473602294922, "learning_rate": 1.1881525471596164e-05, "loss": 5.5718, "step": 12976 }, { "epoch": 0.45727071716130624, "grad_norm": 6.379948139190674, "learning_rate": 1.1880404565955985e-05, "loss": 3.3753, "step": 12977 }, { "epoch": 0.4573059541742647, "grad_norm": 1.9308745861053467, "learning_rate": 1.1879283635823293e-05, "loss": 0.8855, "step": 12978 }, { "epoch": 0.45734119118722305, "grad_norm": 1.1164143085479736, "learning_rate": 1.1878162681212697e-05, "loss": 0.8087, "step": 12979 }, { "epoch": 0.4573764282001815, "grad_norm": 1.6448583602905273, "learning_rate": 1.1877041702138792e-05, "loss": 1.1542, "step": 12980 }, { "epoch": 0.45741166521313986, "grad_norm": 2.1040592193603516, "learning_rate": 1.1875920698616179e-05, "loss": 0.9736, "step": 12981 }, { "epoch": 0.4574469022260983, "grad_norm": 1.2873873710632324, "learning_rate": 1.1874799670659458e-05, "loss": 1.012, "step": 12982 }, { "epoch": 0.45748213923905673, "grad_norm": 1.094382643699646, "learning_rate": 1.1873678618283236e-05, "loss": 1.0437, "step": 12983 }, { "epoch": 0.4575173762520151, "grad_norm": 7.87684440612793, "learning_rate": 1.1872557541502108e-05, "loss": 5.3996, "step": 12984 }, { "epoch": 0.45755261326497354, "grad_norm": 5.8682355880737305, "learning_rate": 1.1871436440330679e-05, "loss": 5.2984, "step": 12985 }, { "epoch": 0.4575878502779319, "grad_norm": 7.564645290374756, "learning_rate": 1.1870315314783554e-05, "loss": 3.1854, "step": 12986 }, { "epoch": 0.45762308729089035, "grad_norm": 4.375140190124512, "learning_rate": 1.1869194164875334e-05, "loss": 2.9347, "step": 12987 }, { "epoch": 0.4576583243038488, "grad_norm": 1.9059672355651855, "learning_rate": 1.1868072990620623e-05, "loss": 0.8208, "step": 12988 }, { "epoch": 0.45769356131680716, "grad_norm": 8.745353698730469, "learning_rate": 1.186695179203402e-05, "loss": 7.4488, "step": 12989 }, { "epoch": 0.4577287983297656, "grad_norm": 1.246822476387024, "learning_rate": 1.1865830569130138e-05, "loss": 0.8973, "step": 12990 }, { "epoch": 0.457764035342724, "grad_norm": 7.11838436126709, "learning_rate": 1.1864709321923567e-05, "loss": 4.8019, "step": 12991 }, { "epoch": 0.4577992723556824, "grad_norm": 1.0283303260803223, "learning_rate": 1.1863588050428925e-05, "loss": 0.9279, "step": 12992 }, { "epoch": 0.45783450936864084, "grad_norm": 20.184167861938477, "learning_rate": 1.186246675466081e-05, "loss": 3.4548, "step": 12993 }, { "epoch": 0.4578697463815992, "grad_norm": 4.562374114990234, "learning_rate": 1.186134543463383e-05, "loss": 3.3472, "step": 12994 }, { "epoch": 0.45790498339455765, "grad_norm": 3.3532986640930176, "learning_rate": 1.1860224090362584e-05, "loss": 3.0991, "step": 12995 }, { "epoch": 0.45794022040751603, "grad_norm": 6.378759860992432, "learning_rate": 1.1859102721861688e-05, "loss": 3.2643, "step": 12996 }, { "epoch": 0.45797545742047446, "grad_norm": 1.1014759540557861, "learning_rate": 1.1857981329145736e-05, "loss": 0.9214, "step": 12997 }, { "epoch": 0.4580106944334329, "grad_norm": 4.516565322875977, "learning_rate": 1.1856859912229345e-05, "loss": 5.0885, "step": 12998 }, { "epoch": 0.4580459314463913, "grad_norm": 1.8001261949539185, "learning_rate": 1.1855738471127118e-05, "loss": 1.1044, "step": 12999 }, { "epoch": 0.4580811684593497, "grad_norm": 1.6735624074935913, "learning_rate": 1.185461700585366e-05, "loss": 1.1472, "step": 13000 }, { "epoch": 0.4581164054723081, "grad_norm": 1.285277009010315, "learning_rate": 1.1853495516423576e-05, "loss": 0.9523, "step": 13001 }, { "epoch": 0.4581516424852665, "grad_norm": 10.13851547241211, "learning_rate": 1.185237400285148e-05, "loss": 4.9407, "step": 13002 }, { "epoch": 0.45818687949822495, "grad_norm": 1.199234127998352, "learning_rate": 1.1851252465151975e-05, "loss": 0.9661, "step": 13003 }, { "epoch": 0.45822211651118333, "grad_norm": 3.512615919113159, "learning_rate": 1.1850130903339671e-05, "loss": 3.379, "step": 13004 }, { "epoch": 0.45825735352414176, "grad_norm": 1.356895089149475, "learning_rate": 1.1849009317429176e-05, "loss": 1.2018, "step": 13005 }, { "epoch": 0.4582925905371002, "grad_norm": 8.738309860229492, "learning_rate": 1.1847887707435099e-05, "loss": 2.7565, "step": 13006 }, { "epoch": 0.4583278275500586, "grad_norm": 7.2969441413879395, "learning_rate": 1.1846766073372054e-05, "loss": 5.6065, "step": 13007 }, { "epoch": 0.458363064563017, "grad_norm": 5.377574920654297, "learning_rate": 1.1845644415254637e-05, "loss": 3.5817, "step": 13008 }, { "epoch": 0.4583983015759754, "grad_norm": 5.542119979858398, "learning_rate": 1.1844522733097475e-05, "loss": 3.5804, "step": 13009 }, { "epoch": 0.4584335385889338, "grad_norm": 6.145172119140625, "learning_rate": 1.1843401026915165e-05, "loss": 2.5517, "step": 13010 }, { "epoch": 0.45846877560189225, "grad_norm": 1.318709373474121, "learning_rate": 1.1842279296722324e-05, "loss": 0.8011, "step": 13011 }, { "epoch": 0.45850401261485063, "grad_norm": 1.1508049964904785, "learning_rate": 1.1841157542533556e-05, "loss": 0.9995, "step": 13012 }, { "epoch": 0.45853924962780906, "grad_norm": 5.930150985717773, "learning_rate": 1.184003576436348e-05, "loss": 3.5191, "step": 13013 }, { "epoch": 0.45857448664076744, "grad_norm": 4.805660247802734, "learning_rate": 1.1838913962226704e-05, "loss": 3.0282, "step": 13014 }, { "epoch": 0.4586097236537259, "grad_norm": 4.91530704498291, "learning_rate": 1.1837792136137841e-05, "loss": 3.3783, "step": 13015 }, { "epoch": 0.4586449606666843, "grad_norm": 4.47419548034668, "learning_rate": 1.1836670286111496e-05, "loss": 3.1493, "step": 13016 }, { "epoch": 0.4586801976796427, "grad_norm": 3.4308395385742188, "learning_rate": 1.1835548412162292e-05, "loss": 3.4119, "step": 13017 }, { "epoch": 0.4587154346926011, "grad_norm": 2.3528082370758057, "learning_rate": 1.183442651430483e-05, "loss": 0.8333, "step": 13018 }, { "epoch": 0.4587506717055595, "grad_norm": 2.1819560527801514, "learning_rate": 1.1833304592553736e-05, "loss": 0.9018, "step": 13019 }, { "epoch": 0.45878590871851793, "grad_norm": 1.6509112119674683, "learning_rate": 1.1832182646923611e-05, "loss": 0.8203, "step": 13020 }, { "epoch": 0.45882114573147637, "grad_norm": 9.828285217285156, "learning_rate": 1.1831060677429075e-05, "loss": 2.8363, "step": 13021 }, { "epoch": 0.45885638274443474, "grad_norm": 1.5824517011642456, "learning_rate": 1.1829938684084742e-05, "loss": 0.8225, "step": 13022 }, { "epoch": 0.4588916197573932, "grad_norm": 1.5438165664672852, "learning_rate": 1.1828816666905224e-05, "loss": 0.9689, "step": 13023 }, { "epoch": 0.45892685677035155, "grad_norm": 3.565443277359009, "learning_rate": 1.1827694625905136e-05, "loss": 2.5853, "step": 13024 }, { "epoch": 0.45896209378331, "grad_norm": 28.78523063659668, "learning_rate": 1.1826572561099093e-05, "loss": 3.2108, "step": 13025 }, { "epoch": 0.4589973307962684, "grad_norm": 1.3329814672470093, "learning_rate": 1.1825450472501709e-05, "loss": 0.9003, "step": 13026 }, { "epoch": 0.4590325678092268, "grad_norm": 7.47274923324585, "learning_rate": 1.1824328360127599e-05, "loss": 3.149, "step": 13027 }, { "epoch": 0.45906780482218523, "grad_norm": 6.360299110412598, "learning_rate": 1.1823206223991382e-05, "loss": 5.3939, "step": 13028 }, { "epoch": 0.4591030418351436, "grad_norm": 22.525808334350586, "learning_rate": 1.1822084064107668e-05, "loss": 3.7532, "step": 13029 }, { "epoch": 0.45913827884810204, "grad_norm": 1.1983433961868286, "learning_rate": 1.182096188049108e-05, "loss": 1.0248, "step": 13030 }, { "epoch": 0.4591735158610605, "grad_norm": 1.6637459993362427, "learning_rate": 1.1819839673156231e-05, "loss": 0.9886, "step": 13031 }, { "epoch": 0.45920875287401886, "grad_norm": 10.19444465637207, "learning_rate": 1.1818717442117738e-05, "loss": 5.2638, "step": 13032 }, { "epoch": 0.4592439898869773, "grad_norm": 1.4445760250091553, "learning_rate": 1.1817595187390219e-05, "loss": 0.8567, "step": 13033 }, { "epoch": 0.45927922689993567, "grad_norm": 2.123932123184204, "learning_rate": 1.1816472908988292e-05, "loss": 0.7687, "step": 13034 }, { "epoch": 0.4593144639128941, "grad_norm": 2.411653995513916, "learning_rate": 1.181535060692657e-05, "loss": 0.8714, "step": 13035 }, { "epoch": 0.45934970092585253, "grad_norm": 1.0847238302230835, "learning_rate": 1.1814228281219681e-05, "loss": 0.8291, "step": 13036 }, { "epoch": 0.4593849379388109, "grad_norm": 3.3655426502227783, "learning_rate": 1.1813105931882232e-05, "loss": 2.7073, "step": 13037 }, { "epoch": 0.45942017495176934, "grad_norm": 10.520769119262695, "learning_rate": 1.1811983558928853e-05, "loss": 7.5886, "step": 13038 }, { "epoch": 0.4594554119647277, "grad_norm": 1.436323881149292, "learning_rate": 1.1810861162374155e-05, "loss": 0.7809, "step": 13039 }, { "epoch": 0.45949064897768616, "grad_norm": 14.39291000366211, "learning_rate": 1.1809738742232758e-05, "loss": 3.3933, "step": 13040 }, { "epoch": 0.4595258859906446, "grad_norm": 12.969619750976562, "learning_rate": 1.1808616298519284e-05, "loss": 7.6022, "step": 13041 }, { "epoch": 0.45956112300360297, "grad_norm": 6.75434684753418, "learning_rate": 1.1807493831248354e-05, "loss": 5.7611, "step": 13042 }, { "epoch": 0.4595963600165614, "grad_norm": 19.64647102355957, "learning_rate": 1.1806371340434583e-05, "loss": 5.7113, "step": 13043 }, { "epoch": 0.45963159702951983, "grad_norm": 1.1269655227661133, "learning_rate": 1.1805248826092598e-05, "loss": 0.9909, "step": 13044 }, { "epoch": 0.4596668340424782, "grad_norm": 6.89022970199585, "learning_rate": 1.1804126288237018e-05, "loss": 4.9788, "step": 13045 }, { "epoch": 0.45970207105543665, "grad_norm": 2.044034242630005, "learning_rate": 1.1803003726882462e-05, "loss": 0.9342, "step": 13046 }, { "epoch": 0.459737308068395, "grad_norm": 1.3190395832061768, "learning_rate": 1.1801881142043552e-05, "loss": 1.0293, "step": 13047 }, { "epoch": 0.45977254508135346, "grad_norm": 2.0394928455352783, "learning_rate": 1.180075853373491e-05, "loss": 0.9958, "step": 13048 }, { "epoch": 0.4598077820943119, "grad_norm": 1.4579042196273804, "learning_rate": 1.179963590197116e-05, "loss": 0.8513, "step": 13049 }, { "epoch": 0.45984301910727027, "grad_norm": 1.1758370399475098, "learning_rate": 1.1798513246766925e-05, "loss": 1.1035, "step": 13050 }, { "epoch": 0.4598782561202287, "grad_norm": 16.61248779296875, "learning_rate": 1.1797390568136823e-05, "loss": 3.2395, "step": 13051 }, { "epoch": 0.4599134931331871, "grad_norm": 9.270047187805176, "learning_rate": 1.179626786609548e-05, "loss": 3.2798, "step": 13052 }, { "epoch": 0.4599487301461455, "grad_norm": 5.517600059509277, "learning_rate": 1.1795145140657522e-05, "loss": 3.2861, "step": 13053 }, { "epoch": 0.45998396715910395, "grad_norm": 2.5537161827087402, "learning_rate": 1.1794022391837564e-05, "loss": 1.0449, "step": 13054 }, { "epoch": 0.4600192041720623, "grad_norm": 6.147225856781006, "learning_rate": 1.1792899619650245e-05, "loss": 2.912, "step": 13055 }, { "epoch": 0.46005444118502076, "grad_norm": 1.3229193687438965, "learning_rate": 1.1791776824110173e-05, "loss": 0.9727, "step": 13056 }, { "epoch": 0.46008967819797914, "grad_norm": 7.120674133300781, "learning_rate": 1.1790654005231983e-05, "loss": 5.0733, "step": 13057 }, { "epoch": 0.46012491521093757, "grad_norm": 1.1345703601837158, "learning_rate": 1.1789531163030294e-05, "loss": 1.2669, "step": 13058 }, { "epoch": 0.460160152223896, "grad_norm": 4.35668420791626, "learning_rate": 1.1788408297519736e-05, "loss": 3.0101, "step": 13059 }, { "epoch": 0.4601953892368544, "grad_norm": 8.899667739868164, "learning_rate": 1.178728540871493e-05, "loss": 4.5464, "step": 13060 }, { "epoch": 0.4602306262498128, "grad_norm": 10.437848091125488, "learning_rate": 1.1786162496630506e-05, "loss": 3.0126, "step": 13061 }, { "epoch": 0.4602658632627712, "grad_norm": 20.09676170349121, "learning_rate": 1.1785039561281085e-05, "loss": 2.8718, "step": 13062 }, { "epoch": 0.4603011002757296, "grad_norm": 4.566854000091553, "learning_rate": 1.17839166026813e-05, "loss": 2.9015, "step": 13063 }, { "epoch": 0.46033633728868806, "grad_norm": 4.806646347045898, "learning_rate": 1.178279362084577e-05, "loss": 3.5157, "step": 13064 }, { "epoch": 0.46037157430164644, "grad_norm": 2.324918270111084, "learning_rate": 1.178167061578913e-05, "loss": 0.7523, "step": 13065 }, { "epoch": 0.46040681131460487, "grad_norm": 3.212261915206909, "learning_rate": 1.1780547587525998e-05, "loss": 3.1491, "step": 13066 }, { "epoch": 0.46044204832756325, "grad_norm": 5.929786682128906, "learning_rate": 1.1779424536071011e-05, "loss": 3.2803, "step": 13067 }, { "epoch": 0.4604772853405217, "grad_norm": 8.138887405395508, "learning_rate": 1.1778301461438792e-05, "loss": 5.1119, "step": 13068 }, { "epoch": 0.4605125223534801, "grad_norm": 4.678337574005127, "learning_rate": 1.1777178363643971e-05, "loss": 2.9195, "step": 13069 }, { "epoch": 0.4605477593664385, "grad_norm": 7.030489444732666, "learning_rate": 1.1776055242701175e-05, "loss": 2.7224, "step": 13070 }, { "epoch": 0.4605829963793969, "grad_norm": 8.731285095214844, "learning_rate": 1.1774932098625032e-05, "loss": 3.2764, "step": 13071 }, { "epoch": 0.4606182333923553, "grad_norm": 16.18901252746582, "learning_rate": 1.1773808931430173e-05, "loss": 3.3942, "step": 13072 }, { "epoch": 0.46065347040531374, "grad_norm": 2.9471240043640137, "learning_rate": 1.1772685741131224e-05, "loss": 0.795, "step": 13073 }, { "epoch": 0.46068870741827217, "grad_norm": 1.0872163772583008, "learning_rate": 1.1771562527742821e-05, "loss": 1.2182, "step": 13074 }, { "epoch": 0.46072394443123055, "grad_norm": 6.933578968048096, "learning_rate": 1.1770439291279589e-05, "loss": 2.8879, "step": 13075 }, { "epoch": 0.460759181444189, "grad_norm": 1.1073673963546753, "learning_rate": 1.1769316031756161e-05, "loss": 0.8194, "step": 13076 }, { "epoch": 0.4607944184571474, "grad_norm": 3.4279391765594482, "learning_rate": 1.1768192749187166e-05, "loss": 2.7131, "step": 13077 }, { "epoch": 0.4608296554701058, "grad_norm": 8.420045852661133, "learning_rate": 1.1767069443587234e-05, "loss": 2.6832, "step": 13078 }, { "epoch": 0.4608648924830642, "grad_norm": 1.5002248287200928, "learning_rate": 1.1765946114970998e-05, "loss": 0.9062, "step": 13079 }, { "epoch": 0.4609001294960226, "grad_norm": 6.1653361320495605, "learning_rate": 1.176482276335309e-05, "loss": 2.8454, "step": 13080 }, { "epoch": 0.46093536650898104, "grad_norm": 8.496986389160156, "learning_rate": 1.1763699388748137e-05, "loss": 3.3899, "step": 13081 }, { "epoch": 0.46097060352193947, "grad_norm": 11.86632251739502, "learning_rate": 1.1762575991170782e-05, "loss": 5.5352, "step": 13082 }, { "epoch": 0.46100584053489785, "grad_norm": 5.983958721160889, "learning_rate": 1.1761452570635644e-05, "loss": 2.9691, "step": 13083 }, { "epoch": 0.4610410775478563, "grad_norm": 7.097707748413086, "learning_rate": 1.1760329127157365e-05, "loss": 3.2005, "step": 13084 }, { "epoch": 0.46107631456081466, "grad_norm": 1.486502766609192, "learning_rate": 1.1759205660750573e-05, "loss": 0.8181, "step": 13085 }, { "epoch": 0.4611115515737731, "grad_norm": 6.8606133460998535, "learning_rate": 1.1758082171429903e-05, "loss": 4.9046, "step": 13086 }, { "epoch": 0.4611467885867315, "grad_norm": 1.2085784673690796, "learning_rate": 1.1756958659209987e-05, "loss": 1.1511, "step": 13087 }, { "epoch": 0.4611820255996899, "grad_norm": 6.815593242645264, "learning_rate": 1.1755835124105464e-05, "loss": 3.2223, "step": 13088 }, { "epoch": 0.46121726261264834, "grad_norm": 2.610403060913086, "learning_rate": 1.1754711566130963e-05, "loss": 2.8052, "step": 13089 }, { "epoch": 0.4612524996256067, "grad_norm": 3.6854779720306396, "learning_rate": 1.1753587985301123e-05, "loss": 2.7062, "step": 13090 }, { "epoch": 0.46128773663856515, "grad_norm": 4.95382833480835, "learning_rate": 1.1752464381630572e-05, "loss": 2.8881, "step": 13091 }, { "epoch": 0.4613229736515236, "grad_norm": 4.732333183288574, "learning_rate": 1.1751340755133953e-05, "loss": 2.8131, "step": 13092 }, { "epoch": 0.46135821066448196, "grad_norm": 1.3317655324935913, "learning_rate": 1.1750217105825899e-05, "loss": 0.7258, "step": 13093 }, { "epoch": 0.4613934476774404, "grad_norm": 1.6779261827468872, "learning_rate": 1.1749093433721037e-05, "loss": 0.7775, "step": 13094 }, { "epoch": 0.46142868469039877, "grad_norm": 41.9369010925293, "learning_rate": 1.1747969738834018e-05, "loss": 5.255, "step": 13095 }, { "epoch": 0.4614639217033572, "grad_norm": 5.01364278793335, "learning_rate": 1.1746846021179466e-05, "loss": 2.7983, "step": 13096 }, { "epoch": 0.46149915871631564, "grad_norm": 8.699069023132324, "learning_rate": 1.1745722280772023e-05, "loss": 2.4873, "step": 13097 }, { "epoch": 0.461534395729274, "grad_norm": 1.4727158546447754, "learning_rate": 1.1744598517626324e-05, "loss": 0.7126, "step": 13098 }, { "epoch": 0.46156963274223245, "grad_norm": 2.156919479370117, "learning_rate": 1.1743474731757012e-05, "loss": 0.9554, "step": 13099 }, { "epoch": 0.4616048697551908, "grad_norm": 1.9867067337036133, "learning_rate": 1.1742350923178712e-05, "loss": 1.2449, "step": 13100 }, { "epoch": 0.46164010676814926, "grad_norm": 8.562432289123535, "learning_rate": 1.1741227091906076e-05, "loss": 3.239, "step": 13101 }, { "epoch": 0.4616753437811077, "grad_norm": 7.377821445465088, "learning_rate": 1.1740103237953728e-05, "loss": 3.3989, "step": 13102 }, { "epoch": 0.46171058079406607, "grad_norm": 3.5410215854644775, "learning_rate": 1.1738979361336322e-05, "loss": 2.868, "step": 13103 }, { "epoch": 0.4617458178070245, "grad_norm": 5.376431465148926, "learning_rate": 1.1737855462068484e-05, "loss": 3.386, "step": 13104 }, { "epoch": 0.4617810548199829, "grad_norm": 6.318586826324463, "learning_rate": 1.1736731540164856e-05, "loss": 3.7607, "step": 13105 }, { "epoch": 0.4618162918329413, "grad_norm": 1.2562347650527954, "learning_rate": 1.173560759564008e-05, "loss": 0.6573, "step": 13106 }, { "epoch": 0.46185152884589975, "grad_norm": 7.216982841491699, "learning_rate": 1.1734483628508795e-05, "loss": 3.563, "step": 13107 }, { "epoch": 0.46188676585885813, "grad_norm": 2.2905638217926025, "learning_rate": 1.1733359638785638e-05, "loss": 0.9329, "step": 13108 }, { "epoch": 0.46192200287181656, "grad_norm": 16.597944259643555, "learning_rate": 1.1732235626485251e-05, "loss": 5.4965, "step": 13109 }, { "epoch": 0.461957239884775, "grad_norm": 6.556042671203613, "learning_rate": 1.1731111591622275e-05, "loss": 3.347, "step": 13110 }, { "epoch": 0.4619924768977334, "grad_norm": 1.3530902862548828, "learning_rate": 1.1729987534211351e-05, "loss": 1.0703, "step": 13111 }, { "epoch": 0.4620277139106918, "grad_norm": 4.670529842376709, "learning_rate": 1.1728863454267118e-05, "loss": 2.7019, "step": 13112 }, { "epoch": 0.4620629509236502, "grad_norm": 5.856513023376465, "learning_rate": 1.1727739351804218e-05, "loss": 3.1614, "step": 13113 }, { "epoch": 0.4620981879366086, "grad_norm": 6.8094868659973145, "learning_rate": 1.1726615226837295e-05, "loss": 3.0472, "step": 13114 }, { "epoch": 0.46213342494956705, "grad_norm": 1.4205855131149292, "learning_rate": 1.1725491079380985e-05, "loss": 0.7092, "step": 13115 }, { "epoch": 0.46216866196252543, "grad_norm": 11.540188789367676, "learning_rate": 1.1724366909449935e-05, "loss": 5.6837, "step": 13116 }, { "epoch": 0.46220389897548386, "grad_norm": 10.76835823059082, "learning_rate": 1.1723242717058789e-05, "loss": 7.595, "step": 13117 }, { "epoch": 0.46223913598844224, "grad_norm": 1.4232549667358398, "learning_rate": 1.1722118502222184e-05, "loss": 0.952, "step": 13118 }, { "epoch": 0.4622743730014007, "grad_norm": 9.662333488464355, "learning_rate": 1.1720994264954768e-05, "loss": 5.504, "step": 13119 }, { "epoch": 0.4623096100143591, "grad_norm": 4.306710720062256, "learning_rate": 1.1719870005271183e-05, "loss": 3.0443, "step": 13120 }, { "epoch": 0.4623448470273175, "grad_norm": 13.905508041381836, "learning_rate": 1.171874572318607e-05, "loss": 7.9607, "step": 13121 }, { "epoch": 0.4623800840402759, "grad_norm": 5.592897891998291, "learning_rate": 1.1717621418714078e-05, "loss": 2.5286, "step": 13122 }, { "epoch": 0.4624153210532343, "grad_norm": 3.864036798477173, "learning_rate": 1.1716497091869845e-05, "loss": 3.1324, "step": 13123 }, { "epoch": 0.46245055806619273, "grad_norm": 3.314589023590088, "learning_rate": 1.171537274266802e-05, "loss": 2.9529, "step": 13124 }, { "epoch": 0.46248579507915116, "grad_norm": 6.52235221862793, "learning_rate": 1.1714248371123247e-05, "loss": 3.4291, "step": 13125 }, { "epoch": 0.46252103209210954, "grad_norm": 11.395000457763672, "learning_rate": 1.171312397725017e-05, "loss": 5.4019, "step": 13126 }, { "epoch": 0.462556269105068, "grad_norm": 9.164623260498047, "learning_rate": 1.1711999561063435e-05, "loss": 5.6718, "step": 13127 }, { "epoch": 0.46259150611802635, "grad_norm": 2.099727153778076, "learning_rate": 1.1710875122577687e-05, "loss": 0.7197, "step": 13128 }, { "epoch": 0.4626267431309848, "grad_norm": 5.803989410400391, "learning_rate": 1.1709750661807573e-05, "loss": 3.6772, "step": 13129 }, { "epoch": 0.4626619801439432, "grad_norm": 1.4388693571090698, "learning_rate": 1.170862617876774e-05, "loss": 0.7531, "step": 13130 }, { "epoch": 0.4626972171569016, "grad_norm": 4.535473823547363, "learning_rate": 1.1707501673472834e-05, "loss": 3.1971, "step": 13131 }, { "epoch": 0.46273245416986003, "grad_norm": 1.806955099105835, "learning_rate": 1.1706377145937498e-05, "loss": 0.8477, "step": 13132 }, { "epoch": 0.4627676911828184, "grad_norm": 0.9680859446525574, "learning_rate": 1.1705252596176386e-05, "loss": 0.917, "step": 13133 }, { "epoch": 0.46280292819577684, "grad_norm": 11.661630630493164, "learning_rate": 1.170412802420414e-05, "loss": 3.0803, "step": 13134 }, { "epoch": 0.4628381652087353, "grad_norm": 2.901510715484619, "learning_rate": 1.170300343003541e-05, "loss": 2.4543, "step": 13135 }, { "epoch": 0.46287340222169365, "grad_norm": 0.9329636693000793, "learning_rate": 1.1701878813684839e-05, "loss": 0.9499, "step": 13136 }, { "epoch": 0.4629086392346521, "grad_norm": 10.29334831237793, "learning_rate": 1.1700754175167083e-05, "loss": 3.211, "step": 13137 }, { "epoch": 0.46294387624761046, "grad_norm": 7.555207252502441, "learning_rate": 1.1699629514496789e-05, "loss": 5.9596, "step": 13138 }, { "epoch": 0.4629791132605689, "grad_norm": 1.2433702945709229, "learning_rate": 1.1698504831688604e-05, "loss": 1.1667, "step": 13139 }, { "epoch": 0.46301435027352733, "grad_norm": 1.3368356227874756, "learning_rate": 1.1697380126757174e-05, "loss": 1.2635, "step": 13140 }, { "epoch": 0.4630495872864857, "grad_norm": 6.732743740081787, "learning_rate": 1.1696255399717156e-05, "loss": 4.9432, "step": 13141 }, { "epoch": 0.46308482429944414, "grad_norm": 9.308911323547363, "learning_rate": 1.169513065058319e-05, "loss": 2.4682, "step": 13142 }, { "epoch": 0.4631200613124025, "grad_norm": 1.7894330024719238, "learning_rate": 1.1694005879369937e-05, "loss": 1.0388, "step": 13143 }, { "epoch": 0.46315529832536095, "grad_norm": 1.6504708528518677, "learning_rate": 1.1692881086092037e-05, "loss": 0.9654, "step": 13144 }, { "epoch": 0.4631905353383194, "grad_norm": 4.562735557556152, "learning_rate": 1.1691756270764148e-05, "loss": 3.4535, "step": 13145 }, { "epoch": 0.46322577235127776, "grad_norm": 6.250023365020752, "learning_rate": 1.1690631433400917e-05, "loss": 5.2391, "step": 13146 }, { "epoch": 0.4632610093642362, "grad_norm": 12.431605339050293, "learning_rate": 1.1689506574016995e-05, "loss": 3.6914, "step": 13147 }, { "epoch": 0.46329624637719463, "grad_norm": 4.750227451324463, "learning_rate": 1.1688381692627036e-05, "loss": 3.3601, "step": 13148 }, { "epoch": 0.463331483390153, "grad_norm": 6.9960126876831055, "learning_rate": 1.1687256789245688e-05, "loss": 5.4977, "step": 13149 }, { "epoch": 0.46336672040311144, "grad_norm": 1.3156490325927734, "learning_rate": 1.1686131863887612e-05, "loss": 0.9058, "step": 13150 }, { "epoch": 0.4634019574160698, "grad_norm": 5.102877140045166, "learning_rate": 1.1685006916567448e-05, "loss": 5.3182, "step": 13151 }, { "epoch": 0.46343719442902825, "grad_norm": 8.97118091583252, "learning_rate": 1.1683881947299855e-05, "loss": 4.8882, "step": 13152 }, { "epoch": 0.4634724314419867, "grad_norm": 3.4671056270599365, "learning_rate": 1.1682756956099489e-05, "loss": 2.9145, "step": 13153 }, { "epoch": 0.46350766845494507, "grad_norm": 4.104032516479492, "learning_rate": 1.1681631942980994e-05, "loss": 3.0682, "step": 13154 }, { "epoch": 0.4635429054679035, "grad_norm": 7.105488300323486, "learning_rate": 1.168050690795903e-05, "loss": 5.2208, "step": 13155 }, { "epoch": 0.4635781424808619, "grad_norm": 6.734366416931152, "learning_rate": 1.1679381851048251e-05, "loss": 3.0114, "step": 13156 }, { "epoch": 0.4636133794938203, "grad_norm": 6.540205478668213, "learning_rate": 1.1678256772263309e-05, "loss": 3.3234, "step": 13157 }, { "epoch": 0.46364861650677874, "grad_norm": 7.341106414794922, "learning_rate": 1.167713167161886e-05, "loss": 4.7274, "step": 13158 }, { "epoch": 0.4636838535197371, "grad_norm": 11.548736572265625, "learning_rate": 1.1676006549129553e-05, "loss": 5.2605, "step": 13159 }, { "epoch": 0.46371909053269555, "grad_norm": 6.182748317718506, "learning_rate": 1.1674881404810053e-05, "loss": 3.1311, "step": 13160 }, { "epoch": 0.46375432754565393, "grad_norm": 1.011171579360962, "learning_rate": 1.1673756238675007e-05, "loss": 1.0037, "step": 13161 }, { "epoch": 0.46378956455861237, "grad_norm": 5.201155662536621, "learning_rate": 1.1672631050739071e-05, "loss": 3.6814, "step": 13162 }, { "epoch": 0.4638248015715708, "grad_norm": 1.1518898010253906, "learning_rate": 1.1671505841016903e-05, "loss": 0.8448, "step": 13163 }, { "epoch": 0.4638600385845292, "grad_norm": 1.113538384437561, "learning_rate": 1.167038060952316e-05, "loss": 0.8516, "step": 13164 }, { "epoch": 0.4638952755974876, "grad_norm": 4.072871685028076, "learning_rate": 1.1669255356272495e-05, "loss": 1.0087, "step": 13165 }, { "epoch": 0.463930512610446, "grad_norm": 8.495455741882324, "learning_rate": 1.1668130081279567e-05, "loss": 3.0363, "step": 13166 }, { "epoch": 0.4639657496234044, "grad_norm": 6.749729633331299, "learning_rate": 1.1667004784559032e-05, "loss": 2.9867, "step": 13167 }, { "epoch": 0.46400098663636286, "grad_norm": 3.736351251602173, "learning_rate": 1.1665879466125552e-05, "loss": 2.4986, "step": 13168 }, { "epoch": 0.46403622364932123, "grad_norm": 12.170574188232422, "learning_rate": 1.1664754125993776e-05, "loss": 3.0572, "step": 13169 }, { "epoch": 0.46407146066227967, "grad_norm": 2.7816004753112793, "learning_rate": 1.1663628764178362e-05, "loss": 0.9467, "step": 13170 }, { "epoch": 0.46410669767523804, "grad_norm": 1.1891651153564453, "learning_rate": 1.1662503380693976e-05, "loss": 0.9903, "step": 13171 }, { "epoch": 0.4641419346881965, "grad_norm": 1.677180290222168, "learning_rate": 1.1661377975555271e-05, "loss": 0.9282, "step": 13172 }, { "epoch": 0.4641771717011549, "grad_norm": 1.2751327753067017, "learning_rate": 1.1660252548776905e-05, "loss": 0.9693, "step": 13173 }, { "epoch": 0.4642124087141133, "grad_norm": 6.237636566162109, "learning_rate": 1.1659127100373537e-05, "loss": 3.4884, "step": 13174 }, { "epoch": 0.4642476457270717, "grad_norm": 22.74219512939453, "learning_rate": 1.1658001630359829e-05, "loss": 4.5489, "step": 13175 }, { "epoch": 0.4642828827400301, "grad_norm": 1.9965914487838745, "learning_rate": 1.1656876138750439e-05, "loss": 0.8763, "step": 13176 }, { "epoch": 0.46431811975298853, "grad_norm": 28.881345748901367, "learning_rate": 1.1655750625560028e-05, "loss": 5.4316, "step": 13177 }, { "epoch": 0.46435335676594697, "grad_norm": 0.9703781008720398, "learning_rate": 1.1654625090803248e-05, "loss": 0.8216, "step": 13178 }, { "epoch": 0.46438859377890535, "grad_norm": 1.5301945209503174, "learning_rate": 1.1653499534494773e-05, "loss": 0.9648, "step": 13179 }, { "epoch": 0.4644238307918638, "grad_norm": 19.623905181884766, "learning_rate": 1.1652373956649253e-05, "loss": 5.897, "step": 13180 }, { "epoch": 0.4644590678048222, "grad_norm": 1.1329529285430908, "learning_rate": 1.1651248357281349e-05, "loss": 1.0333, "step": 13181 }, { "epoch": 0.4644943048177806, "grad_norm": 1.1592769622802734, "learning_rate": 1.1650122736405727e-05, "loss": 0.9284, "step": 13182 }, { "epoch": 0.464529541830739, "grad_norm": 6.733973503112793, "learning_rate": 1.1648997094037044e-05, "loss": 3.4464, "step": 13183 }, { "epoch": 0.4645647788436974, "grad_norm": 5.363095283508301, "learning_rate": 1.1647871430189968e-05, "loss": 3.3934, "step": 13184 }, { "epoch": 0.46460001585665583, "grad_norm": 4.923282146453857, "learning_rate": 1.1646745744879156e-05, "loss": 3.131, "step": 13185 }, { "epoch": 0.46463525286961427, "grad_norm": 4.0233635902404785, "learning_rate": 1.1645620038119265e-05, "loss": 3.3032, "step": 13186 }, { "epoch": 0.46467048988257265, "grad_norm": 7.284640312194824, "learning_rate": 1.164449430992497e-05, "loss": 3.061, "step": 13187 }, { "epoch": 0.4647057268955311, "grad_norm": 6.8615264892578125, "learning_rate": 1.1643368560310926e-05, "loss": 3.0868, "step": 13188 }, { "epoch": 0.46474096390848946, "grad_norm": 1.5064646005630493, "learning_rate": 1.1642242789291795e-05, "loss": 0.8283, "step": 13189 }, { "epoch": 0.4647762009214479, "grad_norm": 8.60623836517334, "learning_rate": 1.1641116996882242e-05, "loss": 4.9556, "step": 13190 }, { "epoch": 0.4648114379344063, "grad_norm": 6.183159351348877, "learning_rate": 1.1639991183096933e-05, "loss": 5.2259, "step": 13191 }, { "epoch": 0.4648466749473647, "grad_norm": 1.1269510984420776, "learning_rate": 1.163886534795053e-05, "loss": 1.1897, "step": 13192 }, { "epoch": 0.46488191196032314, "grad_norm": 5.786351203918457, "learning_rate": 1.1637739491457697e-05, "loss": 3.1675, "step": 13193 }, { "epoch": 0.4649171489732815, "grad_norm": 1.9654383659362793, "learning_rate": 1.1636613613633097e-05, "loss": 0.8957, "step": 13194 }, { "epoch": 0.46495238598623995, "grad_norm": 10.715091705322266, "learning_rate": 1.1635487714491398e-05, "loss": 3.3173, "step": 13195 }, { "epoch": 0.4649876229991984, "grad_norm": 1.17910635471344, "learning_rate": 1.1634361794047264e-05, "loss": 0.9615, "step": 13196 }, { "epoch": 0.46502286001215676, "grad_norm": 1.1518030166625977, "learning_rate": 1.1633235852315354e-05, "loss": 0.8413, "step": 13197 }, { "epoch": 0.4650580970251152, "grad_norm": 3.9442896842956543, "learning_rate": 1.1632109889310345e-05, "loss": 2.5584, "step": 13198 }, { "epoch": 0.46509333403807357, "grad_norm": 4.468230724334717, "learning_rate": 1.1630983905046896e-05, "loss": 2.723, "step": 13199 }, { "epoch": 0.465128571051032, "grad_norm": 1.7889373302459717, "learning_rate": 1.1629857899539672e-05, "loss": 0.9202, "step": 13200 }, { "epoch": 0.46516380806399044, "grad_norm": 6.872321605682373, "learning_rate": 1.1628731872803343e-05, "loss": 2.7583, "step": 13201 }, { "epoch": 0.4651990450769488, "grad_norm": 1.3237485885620117, "learning_rate": 1.162760582485257e-05, "loss": 0.9274, "step": 13202 }, { "epoch": 0.46523428208990725, "grad_norm": 6.265046119689941, "learning_rate": 1.1626479755702027e-05, "loss": 3.4532, "step": 13203 }, { "epoch": 0.4652695191028656, "grad_norm": 1.442606806755066, "learning_rate": 1.1625353665366381e-05, "loss": 0.7587, "step": 13204 }, { "epoch": 0.46530475611582406, "grad_norm": 9.500694274902344, "learning_rate": 1.162422755386029e-05, "loss": 7.2779, "step": 13205 }, { "epoch": 0.4653399931287825, "grad_norm": 4.90692663192749, "learning_rate": 1.1623101421198432e-05, "loss": 3.5629, "step": 13206 }, { "epoch": 0.46537523014174087, "grad_norm": 6.474205017089844, "learning_rate": 1.162197526739547e-05, "loss": 3.2493, "step": 13207 }, { "epoch": 0.4654104671546993, "grad_norm": 1.7882684469223022, "learning_rate": 1.1620849092466075e-05, "loss": 0.9871, "step": 13208 }, { "epoch": 0.4654457041676577, "grad_norm": 1.7630631923675537, "learning_rate": 1.1619722896424914e-05, "loss": 0.9302, "step": 13209 }, { "epoch": 0.4654809411806161, "grad_norm": 0.8034396171569824, "learning_rate": 1.1618596679286654e-05, "loss": 0.8424, "step": 13210 }, { "epoch": 0.46551617819357455, "grad_norm": 1.1905933618545532, "learning_rate": 1.1617470441065968e-05, "loss": 0.9898, "step": 13211 }, { "epoch": 0.4655514152065329, "grad_norm": 1.6051992177963257, "learning_rate": 1.161634418177752e-05, "loss": 0.7869, "step": 13212 }, { "epoch": 0.46558665221949136, "grad_norm": 2.734733819961548, "learning_rate": 1.1615217901435986e-05, "loss": 0.8323, "step": 13213 }, { "epoch": 0.4656218892324498, "grad_norm": 5.500189304351807, "learning_rate": 1.1614091600056034e-05, "loss": 3.1048, "step": 13214 }, { "epoch": 0.46565712624540817, "grad_norm": 9.96145248413086, "learning_rate": 1.1612965277652335e-05, "loss": 3.4193, "step": 13215 }, { "epoch": 0.4656923632583666, "grad_norm": 6.722842216491699, "learning_rate": 1.1611838934239551e-05, "loss": 3.2832, "step": 13216 }, { "epoch": 0.465727600271325, "grad_norm": 9.513612747192383, "learning_rate": 1.1610712569832366e-05, "loss": 5.6709, "step": 13217 }, { "epoch": 0.4657628372842834, "grad_norm": 11.39028263092041, "learning_rate": 1.1609586184445442e-05, "loss": 4.8725, "step": 13218 }, { "epoch": 0.46579807429724185, "grad_norm": 1.2437355518341064, "learning_rate": 1.1608459778093453e-05, "loss": 0.9309, "step": 13219 }, { "epoch": 0.4658333113102002, "grad_norm": 1.0367696285247803, "learning_rate": 1.160733335079107e-05, "loss": 0.8965, "step": 13220 }, { "epoch": 0.46586854832315866, "grad_norm": 3.8403663635253906, "learning_rate": 1.1606206902552966e-05, "loss": 3.3465, "step": 13221 }, { "epoch": 0.46590378533611704, "grad_norm": 5.393975734710693, "learning_rate": 1.1605080433393812e-05, "loss": 2.9939, "step": 13222 }, { "epoch": 0.46593902234907547, "grad_norm": 3.289368152618408, "learning_rate": 1.1603953943328283e-05, "loss": 2.7195, "step": 13223 }, { "epoch": 0.4659742593620339, "grad_norm": 1.0420966148376465, "learning_rate": 1.1602827432371047e-05, "loss": 1.1684, "step": 13224 }, { "epoch": 0.4660094963749923, "grad_norm": 20.290294647216797, "learning_rate": 1.160170090053678e-05, "loss": 3.066, "step": 13225 }, { "epoch": 0.4660447333879507, "grad_norm": 1.106201171875, "learning_rate": 1.1600574347840157e-05, "loss": 1.0067, "step": 13226 }, { "epoch": 0.4660799704009091, "grad_norm": 7.402954578399658, "learning_rate": 1.1599447774295846e-05, "loss": 3.702, "step": 13227 }, { "epoch": 0.4661152074138675, "grad_norm": 6.987366676330566, "learning_rate": 1.1598321179918525e-05, "loss": 3.1526, "step": 13228 }, { "epoch": 0.46615044442682596, "grad_norm": 19.867721557617188, "learning_rate": 1.159719456472287e-05, "loss": 2.9377, "step": 13229 }, { "epoch": 0.46618568143978434, "grad_norm": 15.91910457611084, "learning_rate": 1.1596067928723546e-05, "loss": 4.7444, "step": 13230 }, { "epoch": 0.46622091845274277, "grad_norm": 6.10935640335083, "learning_rate": 1.1594941271935239e-05, "loss": 3.4419, "step": 13231 }, { "epoch": 0.46625615546570115, "grad_norm": 1.6299571990966797, "learning_rate": 1.1593814594372617e-05, "loss": 1.0987, "step": 13232 }, { "epoch": 0.4662913924786596, "grad_norm": 1.65402090549469, "learning_rate": 1.1592687896050357e-05, "loss": 0.8284, "step": 13233 }, { "epoch": 0.466326629491618, "grad_norm": 6.436931610107422, "learning_rate": 1.1591561176983137e-05, "loss": 4.675, "step": 13234 }, { "epoch": 0.4663618665045764, "grad_norm": 3.1181817054748535, "learning_rate": 1.1590434437185625e-05, "loss": 2.9628, "step": 13235 }, { "epoch": 0.46639710351753483, "grad_norm": 2.3774876594543457, "learning_rate": 1.1589307676672507e-05, "loss": 0.8711, "step": 13236 }, { "epoch": 0.4664323405304932, "grad_norm": 1.0421870946884155, "learning_rate": 1.158818089545845e-05, "loss": 0.9596, "step": 13237 }, { "epoch": 0.46646757754345164, "grad_norm": 11.904168128967285, "learning_rate": 1.1587054093558137e-05, "loss": 6.0513, "step": 13238 }, { "epoch": 0.4665028145564101, "grad_norm": 5.599725246429443, "learning_rate": 1.158592727098624e-05, "loss": 3.157, "step": 13239 }, { "epoch": 0.46653805156936845, "grad_norm": 1.608945608139038, "learning_rate": 1.1584800427757442e-05, "loss": 0.8589, "step": 13240 }, { "epoch": 0.4665732885823269, "grad_norm": 1.0639238357543945, "learning_rate": 1.1583673563886413e-05, "loss": 0.8085, "step": 13241 }, { "epoch": 0.46660852559528526, "grad_norm": 7.51070499420166, "learning_rate": 1.1582546679387836e-05, "loss": 3.131, "step": 13242 }, { "epoch": 0.4666437626082437, "grad_norm": 6.2599029541015625, "learning_rate": 1.1581419774276386e-05, "loss": 2.9732, "step": 13243 }, { "epoch": 0.46667899962120213, "grad_norm": 4.631826877593994, "learning_rate": 1.1580292848566742e-05, "loss": 3.055, "step": 13244 }, { "epoch": 0.4667142366341605, "grad_norm": 3.9994149208068848, "learning_rate": 1.1579165902273583e-05, "loss": 3.2171, "step": 13245 }, { "epoch": 0.46674947364711894, "grad_norm": 7.644943714141846, "learning_rate": 1.1578038935411586e-05, "loss": 3.2393, "step": 13246 }, { "epoch": 0.4667847106600773, "grad_norm": 3.863436698913574, "learning_rate": 1.157691194799543e-05, "loss": 1.116, "step": 13247 }, { "epoch": 0.46681994767303575, "grad_norm": 1.4475486278533936, "learning_rate": 1.1575784940039799e-05, "loss": 0.9019, "step": 13248 }, { "epoch": 0.4668551846859942, "grad_norm": 8.364002227783203, "learning_rate": 1.1574657911559363e-05, "loss": 4.3323, "step": 13249 }, { "epoch": 0.46689042169895256, "grad_norm": 1.194445013999939, "learning_rate": 1.1573530862568811e-05, "loss": 0.806, "step": 13250 }, { "epoch": 0.466925658711911, "grad_norm": 5.219081401824951, "learning_rate": 1.1572403793082818e-05, "loss": 2.836, "step": 13251 }, { "epoch": 0.46696089572486943, "grad_norm": 1.5927696228027344, "learning_rate": 1.1571276703116066e-05, "loss": 0.8068, "step": 13252 }, { "epoch": 0.4669961327378278, "grad_norm": 1.232082724571228, "learning_rate": 1.1570149592683234e-05, "loss": 0.6211, "step": 13253 }, { "epoch": 0.46703136975078624, "grad_norm": 1.004906415939331, "learning_rate": 1.1569022461799003e-05, "loss": 1.2356, "step": 13254 }, { "epoch": 0.4670666067637446, "grad_norm": 19.617128372192383, "learning_rate": 1.1567895310478056e-05, "loss": 4.8573, "step": 13255 }, { "epoch": 0.46710184377670305, "grad_norm": 3.372168779373169, "learning_rate": 1.1566768138735069e-05, "loss": 3.103, "step": 13256 }, { "epoch": 0.4671370807896615, "grad_norm": 1.5298123359680176, "learning_rate": 1.1565640946584732e-05, "loss": 1.207, "step": 13257 }, { "epoch": 0.46717231780261986, "grad_norm": 5.585853576660156, "learning_rate": 1.1564513734041723e-05, "loss": 2.4907, "step": 13258 }, { "epoch": 0.4672075548155783, "grad_norm": 4.209167003631592, "learning_rate": 1.1563386501120721e-05, "loss": 3.1269, "step": 13259 }, { "epoch": 0.4672427918285367, "grad_norm": 1.1668652296066284, "learning_rate": 1.156225924783641e-05, "loss": 0.853, "step": 13260 }, { "epoch": 0.4672780288414951, "grad_norm": 1.5830085277557373, "learning_rate": 1.1561131974203477e-05, "loss": 0.7841, "step": 13261 }, { "epoch": 0.46731326585445354, "grad_norm": 1.2368974685668945, "learning_rate": 1.1560004680236594e-05, "loss": 0.8527, "step": 13262 }, { "epoch": 0.4673485028674119, "grad_norm": 1.7376006841659546, "learning_rate": 1.1558877365950459e-05, "loss": 1.0379, "step": 13263 }, { "epoch": 0.46738373988037035, "grad_norm": 15.245024681091309, "learning_rate": 1.155775003135974e-05, "loss": 7.8687, "step": 13264 }, { "epoch": 0.46741897689332873, "grad_norm": 1.2326679229736328, "learning_rate": 1.1556622676479137e-05, "loss": 1.0371, "step": 13265 }, { "epoch": 0.46745421390628716, "grad_norm": 7.654696941375732, "learning_rate": 1.155549530132332e-05, "loss": 2.7865, "step": 13266 }, { "epoch": 0.4674894509192456, "grad_norm": 2.569571018218994, "learning_rate": 1.1554367905906977e-05, "loss": 0.8243, "step": 13267 }, { "epoch": 0.467524687932204, "grad_norm": 1.7341097593307495, "learning_rate": 1.15532404902448e-05, "loss": 0.9763, "step": 13268 }, { "epoch": 0.4675599249451624, "grad_norm": 1.3498538732528687, "learning_rate": 1.1552113054351464e-05, "loss": 0.8992, "step": 13269 }, { "epoch": 0.4675951619581208, "grad_norm": 1.4272688627243042, "learning_rate": 1.1550985598241658e-05, "loss": 0.9648, "step": 13270 }, { "epoch": 0.4676303989710792, "grad_norm": 4.544154644012451, "learning_rate": 1.1549858121930067e-05, "loss": 2.7044, "step": 13271 }, { "epoch": 0.46766563598403765, "grad_norm": 18.75043487548828, "learning_rate": 1.1548730625431377e-05, "loss": 5.3883, "step": 13272 }, { "epoch": 0.46770087299699603, "grad_norm": 3.9928643703460693, "learning_rate": 1.1547603108760273e-05, "loss": 3.0836, "step": 13273 }, { "epoch": 0.46773611000995446, "grad_norm": 31.22264289855957, "learning_rate": 1.1546475571931442e-05, "loss": 3.2293, "step": 13274 }, { "epoch": 0.46777134702291284, "grad_norm": 1.987749695777893, "learning_rate": 1.1545348014959568e-05, "loss": 0.9292, "step": 13275 }, { "epoch": 0.4678065840358713, "grad_norm": 1.4122639894485474, "learning_rate": 1.1544220437859342e-05, "loss": 0.9817, "step": 13276 }, { "epoch": 0.4678418210488297, "grad_norm": 7.405008792877197, "learning_rate": 1.1543092840645444e-05, "loss": 2.5877, "step": 13277 }, { "epoch": 0.4678770580617881, "grad_norm": 6.8737688064575195, "learning_rate": 1.154196522333257e-05, "loss": 3.3849, "step": 13278 }, { "epoch": 0.4679122950747465, "grad_norm": 5.92650842666626, "learning_rate": 1.1540837585935398e-05, "loss": 3.2132, "step": 13279 }, { "epoch": 0.4679475320877049, "grad_norm": 6.509490013122559, "learning_rate": 1.1539709928468625e-05, "loss": 3.3576, "step": 13280 }, { "epoch": 0.46798276910066333, "grad_norm": 1.5131069421768188, "learning_rate": 1.1538582250946928e-05, "loss": 0.6502, "step": 13281 }, { "epoch": 0.46801800611362177, "grad_norm": 16.920122146606445, "learning_rate": 1.1537454553385009e-05, "loss": 2.9291, "step": 13282 }, { "epoch": 0.46805324312658014, "grad_norm": 2.7214109897613525, "learning_rate": 1.1536326835797539e-05, "loss": 0.8283, "step": 13283 }, { "epoch": 0.4680884801395386, "grad_norm": 1.4379912614822388, "learning_rate": 1.1535199098199226e-05, "loss": 1.0619, "step": 13284 }, { "epoch": 0.468123717152497, "grad_norm": 1.2130495309829712, "learning_rate": 1.1534071340604744e-05, "loss": 1.0691, "step": 13285 }, { "epoch": 0.4681589541654554, "grad_norm": 7.556614875793457, "learning_rate": 1.1532943563028787e-05, "loss": 3.3519, "step": 13286 }, { "epoch": 0.4681941911784138, "grad_norm": 5.514894485473633, "learning_rate": 1.1531815765486046e-05, "loss": 5.55, "step": 13287 }, { "epoch": 0.4682294281913722, "grad_norm": 5.422609806060791, "learning_rate": 1.153068794799121e-05, "loss": 4.6805, "step": 13288 }, { "epoch": 0.46826466520433063, "grad_norm": 18.694021224975586, "learning_rate": 1.1529560110558968e-05, "loss": 3.3172, "step": 13289 }, { "epoch": 0.46829990221728907, "grad_norm": 1.3706045150756836, "learning_rate": 1.152843225320401e-05, "loss": 0.9175, "step": 13290 }, { "epoch": 0.46833513923024744, "grad_norm": 2.0584816932678223, "learning_rate": 1.1527304375941028e-05, "loss": 0.91, "step": 13291 }, { "epoch": 0.4683703762432059, "grad_norm": 1.1238272190093994, "learning_rate": 1.1526176478784712e-05, "loss": 0.7681, "step": 13292 }, { "epoch": 0.46840561325616425, "grad_norm": 1.954399585723877, "learning_rate": 1.1525048561749754e-05, "loss": 0.8856, "step": 13293 }, { "epoch": 0.4684408502691227, "grad_norm": 2.004352569580078, "learning_rate": 1.152392062485084e-05, "loss": 0.8049, "step": 13294 }, { "epoch": 0.4684760872820811, "grad_norm": 1.826593041419983, "learning_rate": 1.152279266810267e-05, "loss": 0.7947, "step": 13295 }, { "epoch": 0.4685113242950395, "grad_norm": 4.0875372886657715, "learning_rate": 1.1521664691519931e-05, "loss": 2.9665, "step": 13296 }, { "epoch": 0.46854656130799793, "grad_norm": 2.0960330963134766, "learning_rate": 1.1520536695117315e-05, "loss": 0.9154, "step": 13297 }, { "epoch": 0.4685817983209563, "grad_norm": 1.005806803703308, "learning_rate": 1.1519408678909513e-05, "loss": 0.8153, "step": 13298 }, { "epoch": 0.46861703533391474, "grad_norm": 1.3899714946746826, "learning_rate": 1.1518280642911224e-05, "loss": 0.9536, "step": 13299 }, { "epoch": 0.4686522723468732, "grad_norm": 1.5406767129898071, "learning_rate": 1.1517152587137131e-05, "loss": 0.7766, "step": 13300 }, { "epoch": 0.46868750935983156, "grad_norm": 1.1247260570526123, "learning_rate": 1.1516024511601937e-05, "loss": 1.2078, "step": 13301 }, { "epoch": 0.46872274637279, "grad_norm": 1.7475513219833374, "learning_rate": 1.1514896416320327e-05, "loss": 0.9694, "step": 13302 }, { "epoch": 0.46875798338574837, "grad_norm": 1.324395775794983, "learning_rate": 1.1513768301307e-05, "loss": 0.7943, "step": 13303 }, { "epoch": 0.4687932203987068, "grad_norm": 7.437224388122559, "learning_rate": 1.1512640166576647e-05, "loss": 3.0671, "step": 13304 }, { "epoch": 0.46882845741166523, "grad_norm": 9.363999366760254, "learning_rate": 1.1511512012143963e-05, "loss": 3.4902, "step": 13305 }, { "epoch": 0.4688636944246236, "grad_norm": 3.186798095703125, "learning_rate": 1.1510383838023645e-05, "loss": 2.5353, "step": 13306 }, { "epoch": 0.46889893143758204, "grad_norm": 1.4786510467529297, "learning_rate": 1.1509255644230382e-05, "loss": 0.8103, "step": 13307 }, { "epoch": 0.4689341684505404, "grad_norm": 1.6269385814666748, "learning_rate": 1.1508127430778874e-05, "loss": 0.8875, "step": 13308 }, { "epoch": 0.46896940546349886, "grad_norm": 0.9856446981430054, "learning_rate": 1.1506999197683813e-05, "loss": 0.8097, "step": 13309 }, { "epoch": 0.4690046424764573, "grad_norm": 19.820236206054688, "learning_rate": 1.1505870944959896e-05, "loss": 5.3847, "step": 13310 }, { "epoch": 0.46903987948941567, "grad_norm": 1.5496197938919067, "learning_rate": 1.1504742672621818e-05, "loss": 0.7502, "step": 13311 }, { "epoch": 0.4690751165023741, "grad_norm": 5.085241794586182, "learning_rate": 1.1503614380684277e-05, "loss": 2.8754, "step": 13312 }, { "epoch": 0.4691103535153325, "grad_norm": 1.1957783699035645, "learning_rate": 1.1502486069161962e-05, "loss": 0.9032, "step": 13313 }, { "epoch": 0.4691455905282909, "grad_norm": 6.694576740264893, "learning_rate": 1.150135773806958e-05, "loss": 4.5309, "step": 13314 }, { "epoch": 0.46918082754124935, "grad_norm": 6.9561381340026855, "learning_rate": 1.150022938742182e-05, "loss": 2.5739, "step": 13315 }, { "epoch": 0.4692160645542077, "grad_norm": 8.345254898071289, "learning_rate": 1.149910101723338e-05, "loss": 5.1846, "step": 13316 }, { "epoch": 0.46925130156716616, "grad_norm": 15.751334190368652, "learning_rate": 1.149797262751896e-05, "loss": 2.781, "step": 13317 }, { "epoch": 0.4692865385801246, "grad_norm": 0.9520118832588196, "learning_rate": 1.1496844218293257e-05, "loss": 1.0985, "step": 13318 }, { "epoch": 0.46932177559308297, "grad_norm": 22.286569595336914, "learning_rate": 1.1495715789570962e-05, "loss": 5.1016, "step": 13319 }, { "epoch": 0.4693570126060414, "grad_norm": 7.044591903686523, "learning_rate": 1.1494587341366784e-05, "loss": 5.9159, "step": 13320 }, { "epoch": 0.4693922496189998, "grad_norm": 3.093543529510498, "learning_rate": 1.1493458873695411e-05, "loss": 1.1521, "step": 13321 }, { "epoch": 0.4694274866319582, "grad_norm": 1.2387923002243042, "learning_rate": 1.1492330386571552e-05, "loss": 0.7156, "step": 13322 }, { "epoch": 0.46946272364491665, "grad_norm": 7.953478813171387, "learning_rate": 1.1491201880009895e-05, "loss": 2.9528, "step": 13323 }, { "epoch": 0.469497960657875, "grad_norm": 7.346299648284912, "learning_rate": 1.1490073354025143e-05, "loss": 2.925, "step": 13324 }, { "epoch": 0.46953319767083346, "grad_norm": 13.395124435424805, "learning_rate": 1.1488944808631999e-05, "loss": 5.312, "step": 13325 }, { "epoch": 0.46956843468379184, "grad_norm": 13.04189395904541, "learning_rate": 1.1487816243845159e-05, "loss": 4.5894, "step": 13326 }, { "epoch": 0.46960367169675027, "grad_norm": 13.770657539367676, "learning_rate": 1.148668765967932e-05, "loss": 7.0734, "step": 13327 }, { "epoch": 0.4696389087097087, "grad_norm": 1.2638869285583496, "learning_rate": 1.1485559056149187e-05, "loss": 0.8809, "step": 13328 }, { "epoch": 0.4696741457226671, "grad_norm": 1.2661161422729492, "learning_rate": 1.1484430433269455e-05, "loss": 0.7726, "step": 13329 }, { "epoch": 0.4697093827356255, "grad_norm": 0.9586359858512878, "learning_rate": 1.148330179105483e-05, "loss": 0.7, "step": 13330 }, { "epoch": 0.4697446197485839, "grad_norm": 1.7161364555358887, "learning_rate": 1.148217312952001e-05, "loss": 0.8742, "step": 13331 }, { "epoch": 0.4697798567615423, "grad_norm": 0.984902024269104, "learning_rate": 1.1481044448679698e-05, "loss": 1.4519, "step": 13332 }, { "epoch": 0.46981509377450076, "grad_norm": 1.4860683679580688, "learning_rate": 1.1479915748548592e-05, "loss": 1.0363, "step": 13333 }, { "epoch": 0.46985033078745914, "grad_norm": 4.679792881011963, "learning_rate": 1.1478787029141393e-05, "loss": 2.7786, "step": 13334 }, { "epoch": 0.46988556780041757, "grad_norm": 8.511317253112793, "learning_rate": 1.1477658290472806e-05, "loss": 5.0881, "step": 13335 }, { "epoch": 0.46992080481337595, "grad_norm": 20.549652099609375, "learning_rate": 1.1476529532557529e-05, "loss": 6.9437, "step": 13336 }, { "epoch": 0.4699560418263344, "grad_norm": 6.724545001983643, "learning_rate": 1.1475400755410271e-05, "loss": 3.0647, "step": 13337 }, { "epoch": 0.4699912788392928, "grad_norm": 8.050110816955566, "learning_rate": 1.1474271959045728e-05, "loss": 2.8907, "step": 13338 }, { "epoch": 0.4700265158522512, "grad_norm": 7.294077396392822, "learning_rate": 1.1473143143478609e-05, "loss": 2.6542, "step": 13339 }, { "epoch": 0.4700617528652096, "grad_norm": 32.406959533691406, "learning_rate": 1.1472014308723607e-05, "loss": 3.8567, "step": 13340 }, { "epoch": 0.470096989878168, "grad_norm": 1.401168942451477, "learning_rate": 1.1470885454795435e-05, "loss": 0.8991, "step": 13341 }, { "epoch": 0.47013222689112644, "grad_norm": 6.327846050262451, "learning_rate": 1.146975658170879e-05, "loss": 3.5976, "step": 13342 }, { "epoch": 0.47016746390408487, "grad_norm": 14.979256629943848, "learning_rate": 1.146862768947838e-05, "loss": 3.336, "step": 13343 }, { "epoch": 0.47020270091704325, "grad_norm": 7.027661323547363, "learning_rate": 1.1467498778118907e-05, "loss": 3.2285, "step": 13344 }, { "epoch": 0.4702379379300017, "grad_norm": 6.918109893798828, "learning_rate": 1.1466369847645075e-05, "loss": 2.9133, "step": 13345 }, { "epoch": 0.47027317494296006, "grad_norm": 1.2839807271957397, "learning_rate": 1.1465240898071588e-05, "loss": 0.8607, "step": 13346 }, { "epoch": 0.4703084119559185, "grad_norm": 2.114736557006836, "learning_rate": 1.1464111929413155e-05, "loss": 0.7242, "step": 13347 }, { "epoch": 0.4703436489688769, "grad_norm": 4.0009260177612305, "learning_rate": 1.1462982941684472e-05, "loss": 3.2433, "step": 13348 }, { "epoch": 0.4703788859818353, "grad_norm": 5.8358235359191895, "learning_rate": 1.1461853934900255e-05, "loss": 2.7803, "step": 13349 }, { "epoch": 0.47041412299479374, "grad_norm": 11.181198120117188, "learning_rate": 1.1460724909075203e-05, "loss": 5.4305, "step": 13350 }, { "epoch": 0.4704493600077521, "grad_norm": 1.5701442956924438, "learning_rate": 1.1459595864224022e-05, "loss": 1.011, "step": 13351 }, { "epoch": 0.47048459702071055, "grad_norm": 1.1956171989440918, "learning_rate": 1.145846680036142e-05, "loss": 0.9798, "step": 13352 }, { "epoch": 0.470519834033669, "grad_norm": 5.088256359100342, "learning_rate": 1.1457337717502102e-05, "loss": 2.6185, "step": 13353 }, { "epoch": 0.47055507104662736, "grad_norm": 1.911258339881897, "learning_rate": 1.1456208615660774e-05, "loss": 0.8501, "step": 13354 }, { "epoch": 0.4705903080595858, "grad_norm": 1.3036253452301025, "learning_rate": 1.1455079494852144e-05, "loss": 0.8806, "step": 13355 }, { "epoch": 0.4706255450725442, "grad_norm": 1.0951544046401978, "learning_rate": 1.1453950355090917e-05, "loss": 0.7843, "step": 13356 }, { "epoch": 0.4706607820855026, "grad_norm": 5.249883651733398, "learning_rate": 1.1452821196391801e-05, "loss": 3.3199, "step": 13357 }, { "epoch": 0.47069601909846104, "grad_norm": 1.401979684829712, "learning_rate": 1.1451692018769507e-05, "loss": 1.0038, "step": 13358 }, { "epoch": 0.4707312561114194, "grad_norm": 7.780002593994141, "learning_rate": 1.1450562822238736e-05, "loss": 5.4634, "step": 13359 }, { "epoch": 0.47076649312437785, "grad_norm": 1.0800789594650269, "learning_rate": 1.1449433606814201e-05, "loss": 1.0623, "step": 13360 }, { "epoch": 0.4708017301373363, "grad_norm": 3.1397712230682373, "learning_rate": 1.144830437251061e-05, "loss": 2.9037, "step": 13361 }, { "epoch": 0.47083696715029466, "grad_norm": 3.024127244949341, "learning_rate": 1.1447175119342668e-05, "loss": 1.2059, "step": 13362 }, { "epoch": 0.4708722041632531, "grad_norm": 5.9611287117004395, "learning_rate": 1.1446045847325087e-05, "loss": 2.4943, "step": 13363 }, { "epoch": 0.47090744117621147, "grad_norm": 33.06524658203125, "learning_rate": 1.1444916556472574e-05, "loss": 7.2771, "step": 13364 }, { "epoch": 0.4709426781891699, "grad_norm": 5.8318328857421875, "learning_rate": 1.1443787246799837e-05, "loss": 2.9721, "step": 13365 }, { "epoch": 0.47097791520212834, "grad_norm": 13.492185592651367, "learning_rate": 1.1442657918321592e-05, "loss": 3.3001, "step": 13366 }, { "epoch": 0.4710131522150867, "grad_norm": 7.838754177093506, "learning_rate": 1.1441528571052539e-05, "loss": 3.6053, "step": 13367 }, { "epoch": 0.47104838922804515, "grad_norm": 20.200641632080078, "learning_rate": 1.1440399205007395e-05, "loss": 7.2814, "step": 13368 }, { "epoch": 0.47108362624100353, "grad_norm": 4.408144474029541, "learning_rate": 1.1439269820200869e-05, "loss": 3.1006, "step": 13369 }, { "epoch": 0.47111886325396196, "grad_norm": 4.561007976531982, "learning_rate": 1.1438140416647668e-05, "loss": 2.6653, "step": 13370 }, { "epoch": 0.4711541002669204, "grad_norm": 3.2558248043060303, "learning_rate": 1.1437010994362505e-05, "loss": 3.0391, "step": 13371 }, { "epoch": 0.4711893372798788, "grad_norm": 6.570993423461914, "learning_rate": 1.1435881553360094e-05, "loss": 3.3728, "step": 13372 }, { "epoch": 0.4712245742928372, "grad_norm": 7.9108476638793945, "learning_rate": 1.143475209365514e-05, "loss": 5.077, "step": 13373 }, { "epoch": 0.4712598113057956, "grad_norm": 6.839019775390625, "learning_rate": 1.1433622615262356e-05, "loss": 3.3546, "step": 13374 }, { "epoch": 0.471295048318754, "grad_norm": 9.34634780883789, "learning_rate": 1.1432493118196455e-05, "loss": 3.2586, "step": 13375 }, { "epoch": 0.47133028533171245, "grad_norm": 1.5926064252853394, "learning_rate": 1.1431363602472149e-05, "loss": 0.8294, "step": 13376 }, { "epoch": 0.47136552234467083, "grad_norm": 5.504833221435547, "learning_rate": 1.1430234068104155e-05, "loss": 2.4591, "step": 13377 }, { "epoch": 0.47140075935762926, "grad_norm": 1.9025605916976929, "learning_rate": 1.1429104515107172e-05, "loss": 0.7187, "step": 13378 }, { "epoch": 0.47143599637058764, "grad_norm": 3.5085713863372803, "learning_rate": 1.1427974943495928e-05, "loss": 2.672, "step": 13379 }, { "epoch": 0.4714712333835461, "grad_norm": 6.360789775848389, "learning_rate": 1.1426845353285123e-05, "loss": 2.9488, "step": 13380 }, { "epoch": 0.4715064703965045, "grad_norm": 6.6629533767700195, "learning_rate": 1.142571574448948e-05, "loss": 3.4717, "step": 13381 }, { "epoch": 0.4715417074094629, "grad_norm": 2.148538112640381, "learning_rate": 1.1424586117123703e-05, "loss": 0.8033, "step": 13382 }, { "epoch": 0.4715769444224213, "grad_norm": 5.920726299285889, "learning_rate": 1.1423456471202514e-05, "loss": 3.0417, "step": 13383 }, { "epoch": 0.4716121814353797, "grad_norm": 5.001676082611084, "learning_rate": 1.142232680674062e-05, "loss": 3.2054, "step": 13384 }, { "epoch": 0.47164741844833813, "grad_norm": 1.3789939880371094, "learning_rate": 1.1421197123752744e-05, "loss": 1.0927, "step": 13385 }, { "epoch": 0.47168265546129656, "grad_norm": 1.2802391052246094, "learning_rate": 1.1420067422253587e-05, "loss": 1.0075, "step": 13386 }, { "epoch": 0.47171789247425494, "grad_norm": 1.2365832328796387, "learning_rate": 1.1418937702257876e-05, "loss": 1.0397, "step": 13387 }, { "epoch": 0.4717531294872134, "grad_norm": 1.4311062097549438, "learning_rate": 1.1417807963780318e-05, "loss": 0.8437, "step": 13388 }, { "epoch": 0.4717883665001718, "grad_norm": 1.1364161968231201, "learning_rate": 1.1416678206835632e-05, "loss": 0.9057, "step": 13389 }, { "epoch": 0.4718236035131302, "grad_norm": 25.145906448364258, "learning_rate": 1.1415548431438533e-05, "loss": 5.5698, "step": 13390 }, { "epoch": 0.4718588405260886, "grad_norm": 8.149650573730469, "learning_rate": 1.1414418637603732e-05, "loss": 3.0763, "step": 13391 }, { "epoch": 0.471894077539047, "grad_norm": 10.927009582519531, "learning_rate": 1.141328882534595e-05, "loss": 7.9819, "step": 13392 }, { "epoch": 0.47192931455200543, "grad_norm": 4.086466312408447, "learning_rate": 1.14121589946799e-05, "loss": 2.9944, "step": 13393 }, { "epoch": 0.47196455156496386, "grad_norm": 12.408170700073242, "learning_rate": 1.1411029145620302e-05, "loss": 5.1899, "step": 13394 }, { "epoch": 0.47199978857792224, "grad_norm": 11.072809219360352, "learning_rate": 1.1409899278181866e-05, "loss": 4.7191, "step": 13395 }, { "epoch": 0.4720350255908807, "grad_norm": 8.145694732666016, "learning_rate": 1.1408769392379317e-05, "loss": 6.9308, "step": 13396 }, { "epoch": 0.47207026260383905, "grad_norm": 3.4344449043273926, "learning_rate": 1.1407639488227362e-05, "loss": 2.9008, "step": 13397 }, { "epoch": 0.4721054996167975, "grad_norm": 5.991207122802734, "learning_rate": 1.1406509565740728e-05, "loss": 2.9458, "step": 13398 }, { "epoch": 0.4721407366297559, "grad_norm": 1.2330905199050903, "learning_rate": 1.1405379624934126e-05, "loss": 1.1605, "step": 13399 }, { "epoch": 0.4721759736427143, "grad_norm": 1.1831971406936646, "learning_rate": 1.1404249665822275e-05, "loss": 0.7887, "step": 13400 }, { "epoch": 0.47221121065567273, "grad_norm": 1.0019952058792114, "learning_rate": 1.1403119688419893e-05, "loss": 1.0492, "step": 13401 }, { "epoch": 0.4722464476686311, "grad_norm": 10.519232749938965, "learning_rate": 1.1401989692741697e-05, "loss": 5.3721, "step": 13402 }, { "epoch": 0.47228168468158954, "grad_norm": 1.4063938856124878, "learning_rate": 1.140085967880241e-05, "loss": 0.9807, "step": 13403 }, { "epoch": 0.472316921694548, "grad_norm": 1.366197943687439, "learning_rate": 1.1399729646616746e-05, "loss": 1.0273, "step": 13404 }, { "epoch": 0.47235215870750635, "grad_norm": 9.476280212402344, "learning_rate": 1.1398599596199422e-05, "loss": 5.5182, "step": 13405 }, { "epoch": 0.4723873957204648, "grad_norm": 2.03413462638855, "learning_rate": 1.1397469527565166e-05, "loss": 0.9379, "step": 13406 }, { "epoch": 0.47242263273342316, "grad_norm": 1.3369746208190918, "learning_rate": 1.1396339440728688e-05, "loss": 0.8083, "step": 13407 }, { "epoch": 0.4724578697463816, "grad_norm": 1.9323009252548218, "learning_rate": 1.1395209335704709e-05, "loss": 0.886, "step": 13408 }, { "epoch": 0.47249310675934003, "grad_norm": 6.72907829284668, "learning_rate": 1.1394079212507953e-05, "loss": 2.9283, "step": 13409 }, { "epoch": 0.4725283437722984, "grad_norm": 1.4332891702651978, "learning_rate": 1.1392949071153137e-05, "loss": 1.2029, "step": 13410 }, { "epoch": 0.47256358078525684, "grad_norm": 5.616357326507568, "learning_rate": 1.1391818911654982e-05, "loss": 2.8434, "step": 13411 }, { "epoch": 0.4725988177982152, "grad_norm": 1.8606234788894653, "learning_rate": 1.1390688734028208e-05, "loss": 0.7769, "step": 13412 }, { "epoch": 0.47263405481117365, "grad_norm": 1.4865642786026, "learning_rate": 1.1389558538287538e-05, "loss": 1.0406, "step": 13413 }, { "epoch": 0.4726692918241321, "grad_norm": 1.6530894041061401, "learning_rate": 1.1388428324447687e-05, "loss": 0.9901, "step": 13414 }, { "epoch": 0.47270452883709047, "grad_norm": 1.2800157070159912, "learning_rate": 1.1387298092523387e-05, "loss": 1.006, "step": 13415 }, { "epoch": 0.4727397658500489, "grad_norm": 1.2750234603881836, "learning_rate": 1.1386167842529345e-05, "loss": 0.9829, "step": 13416 }, { "epoch": 0.4727750028630073, "grad_norm": 4.28350305557251, "learning_rate": 1.1385037574480296e-05, "loss": 2.8321, "step": 13417 }, { "epoch": 0.4728102398759657, "grad_norm": 4.066495418548584, "learning_rate": 1.1383907288390955e-05, "loss": 2.9475, "step": 13418 }, { "epoch": 0.47284547688892414, "grad_norm": 11.46770191192627, "learning_rate": 1.1382776984276042e-05, "loss": 3.7316, "step": 13419 }, { "epoch": 0.4728807139018825, "grad_norm": 4.848984241485596, "learning_rate": 1.1381646662150285e-05, "loss": 3.2964, "step": 13420 }, { "epoch": 0.47291595091484095, "grad_norm": 8.734017372131348, "learning_rate": 1.1380516322028404e-05, "loss": 3.2023, "step": 13421 }, { "epoch": 0.4729511879277994, "grad_norm": 1.51955246925354, "learning_rate": 1.1379385963925122e-05, "loss": 0.976, "step": 13422 }, { "epoch": 0.47298642494075777, "grad_norm": 2.226740598678589, "learning_rate": 1.1378255587855164e-05, "loss": 0.9615, "step": 13423 }, { "epoch": 0.4730216619537162, "grad_norm": 0.9858365058898926, "learning_rate": 1.1377125193833248e-05, "loss": 1.35, "step": 13424 }, { "epoch": 0.4730568989666746, "grad_norm": 17.774742126464844, "learning_rate": 1.1375994781874104e-05, "loss": 3.0009, "step": 13425 }, { "epoch": 0.473092135979633, "grad_norm": 5.244336128234863, "learning_rate": 1.1374864351992453e-05, "loss": 4.6661, "step": 13426 }, { "epoch": 0.47312737299259144, "grad_norm": 3.279755115509033, "learning_rate": 1.1373733904203015e-05, "loss": 2.9785, "step": 13427 }, { "epoch": 0.4731626100055498, "grad_norm": 1.5289076566696167, "learning_rate": 1.137260343852052e-05, "loss": 0.8945, "step": 13428 }, { "epoch": 0.47319784701850826, "grad_norm": 4.063310146331787, "learning_rate": 1.1371472954959691e-05, "loss": 2.5074, "step": 13429 }, { "epoch": 0.47323308403146663, "grad_norm": 5.442110538482666, "learning_rate": 1.137034245353525e-05, "loss": 5.2687, "step": 13430 }, { "epoch": 0.47326832104442507, "grad_norm": 12.137840270996094, "learning_rate": 1.1369211934261926e-05, "loss": 3.3501, "step": 13431 }, { "epoch": 0.4733035580573835, "grad_norm": 1.254675269126892, "learning_rate": 1.136808139715444e-05, "loss": 0.8324, "step": 13432 }, { "epoch": 0.4733387950703419, "grad_norm": 17.974620819091797, "learning_rate": 1.1366950842227522e-05, "loss": 3.1325, "step": 13433 }, { "epoch": 0.4733740320833003, "grad_norm": 8.661408424377441, "learning_rate": 1.1365820269495896e-05, "loss": 5.1935, "step": 13434 }, { "epoch": 0.4734092690962587, "grad_norm": 1.5869742631912231, "learning_rate": 1.1364689678974283e-05, "loss": 1.0572, "step": 13435 }, { "epoch": 0.4734445061092171, "grad_norm": 5.649849891662598, "learning_rate": 1.1363559070677417e-05, "loss": 5.3784, "step": 13436 }, { "epoch": 0.47347974312217556, "grad_norm": 6.211889266967773, "learning_rate": 1.1362428444620018e-05, "loss": 3.3191, "step": 13437 }, { "epoch": 0.47351498013513393, "grad_norm": 6.616042613983154, "learning_rate": 1.1361297800816815e-05, "loss": 2.7641, "step": 13438 }, { "epoch": 0.47355021714809237, "grad_norm": 7.998701572418213, "learning_rate": 1.1360167139282537e-05, "loss": 3.9585, "step": 13439 }, { "epoch": 0.47358545416105075, "grad_norm": 1.3419194221496582, "learning_rate": 1.1359036460031907e-05, "loss": 1.0472, "step": 13440 }, { "epoch": 0.4736206911740092, "grad_norm": 1.0814744234085083, "learning_rate": 1.1357905763079653e-05, "loss": 0.9663, "step": 13441 }, { "epoch": 0.4736559281869676, "grad_norm": 4.760093688964844, "learning_rate": 1.1356775048440507e-05, "loss": 2.6012, "step": 13442 }, { "epoch": 0.473691165199926, "grad_norm": 9.12138843536377, "learning_rate": 1.1355644316129188e-05, "loss": 5.3527, "step": 13443 }, { "epoch": 0.4737264022128844, "grad_norm": 7.184081077575684, "learning_rate": 1.1354513566160434e-05, "loss": 5.7125, "step": 13444 }, { "epoch": 0.4737616392258428, "grad_norm": 14.9723482131958, "learning_rate": 1.1353382798548967e-05, "loss": 3.7824, "step": 13445 }, { "epoch": 0.47379687623880123, "grad_norm": 1.025651216506958, "learning_rate": 1.1352252013309515e-05, "loss": 0.943, "step": 13446 }, { "epoch": 0.47383211325175967, "grad_norm": 1.208382248878479, "learning_rate": 1.1351121210456808e-05, "loss": 1.0203, "step": 13447 }, { "epoch": 0.47386735026471805, "grad_norm": 24.98995590209961, "learning_rate": 1.1349990390005576e-05, "loss": 2.868, "step": 13448 }, { "epoch": 0.4739025872776765, "grad_norm": 5.073128700256348, "learning_rate": 1.1348859551970548e-05, "loss": 3.7573, "step": 13449 }, { "epoch": 0.47393782429063486, "grad_norm": 6.768165111541748, "learning_rate": 1.1347728696366452e-05, "loss": 2.8816, "step": 13450 }, { "epoch": 0.4739730613035933, "grad_norm": 2.0003767013549805, "learning_rate": 1.1346597823208017e-05, "loss": 1.0541, "step": 13451 }, { "epoch": 0.4740082983165517, "grad_norm": 9.428092002868652, "learning_rate": 1.1345466932509975e-05, "loss": 5.7142, "step": 13452 }, { "epoch": 0.4740435353295101, "grad_norm": 40.24091720581055, "learning_rate": 1.1344336024287056e-05, "loss": 5.3863, "step": 13453 }, { "epoch": 0.47407877234246854, "grad_norm": 2.1299967765808105, "learning_rate": 1.1343205098553986e-05, "loss": 0.9922, "step": 13454 }, { "epoch": 0.4741140093554269, "grad_norm": 12.51383113861084, "learning_rate": 1.1342074155325502e-05, "loss": 4.892, "step": 13455 }, { "epoch": 0.47414924636838535, "grad_norm": 1.9715876579284668, "learning_rate": 1.1340943194616327e-05, "loss": 0.9804, "step": 13456 }, { "epoch": 0.4741844833813438, "grad_norm": 5.329116344451904, "learning_rate": 1.1339812216441201e-05, "loss": 3.1368, "step": 13457 }, { "epoch": 0.47421972039430216, "grad_norm": 7.587287425994873, "learning_rate": 1.1338681220814847e-05, "loss": 3.3637, "step": 13458 }, { "epoch": 0.4742549574072606, "grad_norm": 5.21053409576416, "learning_rate": 1.1337550207752e-05, "loss": 3.0339, "step": 13459 }, { "epoch": 0.474290194420219, "grad_norm": 3.1458709239959717, "learning_rate": 1.133641917726739e-05, "loss": 2.8055, "step": 13460 }, { "epoch": 0.4743254314331774, "grad_norm": 2.3578298091888428, "learning_rate": 1.1335288129375755e-05, "loss": 0.834, "step": 13461 }, { "epoch": 0.47436066844613584, "grad_norm": 3.535827398300171, "learning_rate": 1.1334157064091815e-05, "loss": 2.7209, "step": 13462 }, { "epoch": 0.4743959054590942, "grad_norm": 8.212068557739258, "learning_rate": 1.1333025981430316e-05, "loss": 3.3174, "step": 13463 }, { "epoch": 0.47443114247205265, "grad_norm": 8.989166259765625, "learning_rate": 1.133189488140598e-05, "loss": 3.0871, "step": 13464 }, { "epoch": 0.4744663794850111, "grad_norm": 1.300966739654541, "learning_rate": 1.1330763764033543e-05, "loss": 0.9483, "step": 13465 }, { "epoch": 0.47450161649796946, "grad_norm": 4.599660396575928, "learning_rate": 1.132963262932774e-05, "loss": 3.4946, "step": 13466 }, { "epoch": 0.4745368535109279, "grad_norm": 22.120908737182617, "learning_rate": 1.1328501477303302e-05, "loss": 5.922, "step": 13467 }, { "epoch": 0.47457209052388627, "grad_norm": 1.2196060419082642, "learning_rate": 1.1327370307974963e-05, "loss": 0.8101, "step": 13468 }, { "epoch": 0.4746073275368447, "grad_norm": 15.995087623596191, "learning_rate": 1.1326239121357456e-05, "loss": 4.6074, "step": 13469 }, { "epoch": 0.47464256454980314, "grad_norm": 1.2474150657653809, "learning_rate": 1.1325107917465516e-05, "loss": 0.9079, "step": 13470 }, { "epoch": 0.4746778015627615, "grad_norm": 8.893747329711914, "learning_rate": 1.1323976696313878e-05, "loss": 3.4014, "step": 13471 }, { "epoch": 0.47471303857571995, "grad_norm": 5.106251239776611, "learning_rate": 1.1322845457917276e-05, "loss": 2.6795, "step": 13472 }, { "epoch": 0.4747482755886783, "grad_norm": 1.7020877599716187, "learning_rate": 1.1321714202290442e-05, "loss": 0.8978, "step": 13473 }, { "epoch": 0.47478351260163676, "grad_norm": 1.3492916822433472, "learning_rate": 1.132058292944811e-05, "loss": 0.8476, "step": 13474 }, { "epoch": 0.4748187496145952, "grad_norm": 1.5480290651321411, "learning_rate": 1.1319451639405017e-05, "loss": 1.0437, "step": 13475 }, { "epoch": 0.47485398662755357, "grad_norm": 42.705623626708984, "learning_rate": 1.1318320332175901e-05, "loss": 3.4379, "step": 13476 }, { "epoch": 0.474889223640512, "grad_norm": 1.2153445482254028, "learning_rate": 1.1317189007775494e-05, "loss": 0.8632, "step": 13477 }, { "epoch": 0.4749244606534704, "grad_norm": 1.2634482383728027, "learning_rate": 1.1316057666218529e-05, "loss": 0.7969, "step": 13478 }, { "epoch": 0.4749596976664288, "grad_norm": 1.5721588134765625, "learning_rate": 1.1314926307519748e-05, "loss": 0.895, "step": 13479 }, { "epoch": 0.47499493467938725, "grad_norm": 1.4171444177627563, "learning_rate": 1.1313794931693889e-05, "loss": 0.8241, "step": 13480 }, { "epoch": 0.4750301716923456, "grad_norm": 7.042574405670166, "learning_rate": 1.1312663538755675e-05, "loss": 3.2246, "step": 13481 }, { "epoch": 0.47506540870530406, "grad_norm": 1.5719847679138184, "learning_rate": 1.1311532128719857e-05, "loss": 1.375, "step": 13482 }, { "epoch": 0.47510064571826244, "grad_norm": 9.796751976013184, "learning_rate": 1.1310400701601162e-05, "loss": 3.558, "step": 13483 }, { "epoch": 0.47513588273122087, "grad_norm": 10.663731575012207, "learning_rate": 1.1309269257414336e-05, "loss": 3.1774, "step": 13484 }, { "epoch": 0.4751711197441793, "grad_norm": 6.285039901733398, "learning_rate": 1.1308137796174108e-05, "loss": 3.0603, "step": 13485 }, { "epoch": 0.4752063567571377, "grad_norm": 7.763967037200928, "learning_rate": 1.130700631789522e-05, "loss": 6.1589, "step": 13486 }, { "epoch": 0.4752415937700961, "grad_norm": 10.634733200073242, "learning_rate": 1.1305874822592406e-05, "loss": 6.3117, "step": 13487 }, { "epoch": 0.4752768307830545, "grad_norm": 6.699540615081787, "learning_rate": 1.1304743310280411e-05, "loss": 4.5365, "step": 13488 }, { "epoch": 0.4753120677960129, "grad_norm": 2.3668923377990723, "learning_rate": 1.1303611780973962e-05, "loss": 0.7658, "step": 13489 }, { "epoch": 0.47534730480897136, "grad_norm": 6.078759670257568, "learning_rate": 1.1302480234687808e-05, "loss": 3.5461, "step": 13490 }, { "epoch": 0.47538254182192974, "grad_norm": 7.723932266235352, "learning_rate": 1.130134867143668e-05, "loss": 3.6518, "step": 13491 }, { "epoch": 0.47541777883488817, "grad_norm": 7.987955093383789, "learning_rate": 1.1300217091235322e-05, "loss": 4.3139, "step": 13492 }, { "epoch": 0.4754530158478466, "grad_norm": 36.21770477294922, "learning_rate": 1.1299085494098468e-05, "loss": 5.8231, "step": 13493 }, { "epoch": 0.475488252860805, "grad_norm": 3.199310541152954, "learning_rate": 1.1297953880040861e-05, "loss": 2.1469, "step": 13494 }, { "epoch": 0.4755234898737634, "grad_norm": 4.634385585784912, "learning_rate": 1.129682224907724e-05, "loss": 2.8887, "step": 13495 }, { "epoch": 0.4755587268867218, "grad_norm": 1.0687730312347412, "learning_rate": 1.1295690601222345e-05, "loss": 0.8259, "step": 13496 }, { "epoch": 0.4755939638996802, "grad_norm": 5.233795642852783, "learning_rate": 1.1294558936490914e-05, "loss": 3.5635, "step": 13497 }, { "epoch": 0.47562920091263866, "grad_norm": 3.974550247192383, "learning_rate": 1.1293427254897686e-05, "loss": 3.1572, "step": 13498 }, { "epoch": 0.47566443792559704, "grad_norm": 4.66811466217041, "learning_rate": 1.1292295556457407e-05, "loss": 2.9927, "step": 13499 }, { "epoch": 0.4756996749385555, "grad_norm": 5.402266502380371, "learning_rate": 1.1291163841184808e-05, "loss": 3.4479, "step": 13500 }, { "epoch": 0.47573491195151385, "grad_norm": 9.388202667236328, "learning_rate": 1.1290032109094645e-05, "loss": 7.7147, "step": 13501 }, { "epoch": 0.4757701489644723, "grad_norm": 4.885872840881348, "learning_rate": 1.128890036020164e-05, "loss": 3.4617, "step": 13502 }, { "epoch": 0.4758053859774307, "grad_norm": 4.002029895782471, "learning_rate": 1.128776859452055e-05, "loss": 0.9636, "step": 13503 }, { "epoch": 0.4758406229903891, "grad_norm": 1.4427918195724487, "learning_rate": 1.1286636812066107e-05, "loss": 0.9134, "step": 13504 }, { "epoch": 0.47587586000334753, "grad_norm": 4.280287742614746, "learning_rate": 1.1285505012853055e-05, "loss": 3.6889, "step": 13505 }, { "epoch": 0.4759110970163059, "grad_norm": 8.6510648727417, "learning_rate": 1.1284373196896137e-05, "loss": 7.2856, "step": 13506 }, { "epoch": 0.47594633402926434, "grad_norm": 3.8609442710876465, "learning_rate": 1.1283241364210098e-05, "loss": 2.8497, "step": 13507 }, { "epoch": 0.4759815710422228, "grad_norm": 2.41475510597229, "learning_rate": 1.1282109514809672e-05, "loss": 0.8906, "step": 13508 }, { "epoch": 0.47601680805518115, "grad_norm": 3.9250314235687256, "learning_rate": 1.1280977648709608e-05, "loss": 2.7756, "step": 13509 }, { "epoch": 0.4760520450681396, "grad_norm": 4.953522205352783, "learning_rate": 1.1279845765924646e-05, "loss": 3.1854, "step": 13510 }, { "epoch": 0.47608728208109796, "grad_norm": 1.8189247846603394, "learning_rate": 1.1278713866469533e-05, "loss": 0.7164, "step": 13511 }, { "epoch": 0.4761225190940564, "grad_norm": 9.82861042022705, "learning_rate": 1.1277581950359007e-05, "loss": 3.2895, "step": 13512 }, { "epoch": 0.47615775610701483, "grad_norm": 0.9400641322135925, "learning_rate": 1.1276450017607812e-05, "loss": 0.5424, "step": 13513 }, { "epoch": 0.4761929931199732, "grad_norm": 1.2842880487442017, "learning_rate": 1.1275318068230694e-05, "loss": 0.8792, "step": 13514 }, { "epoch": 0.47622823013293164, "grad_norm": 7.66074275970459, "learning_rate": 1.1274186102242396e-05, "loss": 3.3437, "step": 13515 }, { "epoch": 0.47626346714589, "grad_norm": 8.993509292602539, "learning_rate": 1.127305411965766e-05, "loss": 3.327, "step": 13516 }, { "epoch": 0.47629870415884845, "grad_norm": 4.810579776763916, "learning_rate": 1.1271922120491234e-05, "loss": 2.9438, "step": 13517 }, { "epoch": 0.4763339411718069, "grad_norm": 8.346535682678223, "learning_rate": 1.1270790104757864e-05, "loss": 3.3491, "step": 13518 }, { "epoch": 0.47636917818476526, "grad_norm": 14.384476661682129, "learning_rate": 1.1269658072472283e-05, "loss": 6.9825, "step": 13519 }, { "epoch": 0.4764044151977237, "grad_norm": 3.7641568183898926, "learning_rate": 1.1268526023649248e-05, "loss": 3.0598, "step": 13520 }, { "epoch": 0.4764396522106821, "grad_norm": 1.3242923021316528, "learning_rate": 1.1267393958303499e-05, "loss": 0.8431, "step": 13521 }, { "epoch": 0.4764748892236405, "grad_norm": 8.887686729431152, "learning_rate": 1.1266261876449784e-05, "loss": 5.2974, "step": 13522 }, { "epoch": 0.47651012623659894, "grad_norm": 6.306596279144287, "learning_rate": 1.1265129778102845e-05, "loss": 3.0985, "step": 13523 }, { "epoch": 0.4765453632495573, "grad_norm": 1.3256356716156006, "learning_rate": 1.1263997663277433e-05, "loss": 0.7785, "step": 13524 }, { "epoch": 0.47658060026251575, "grad_norm": 7.5377583503723145, "learning_rate": 1.1262865531988285e-05, "loss": 5.0392, "step": 13525 }, { "epoch": 0.4766158372754742, "grad_norm": 4.799155235290527, "learning_rate": 1.1261733384250157e-05, "loss": 3.3204, "step": 13526 }, { "epoch": 0.47665107428843256, "grad_norm": 3.6239607334136963, "learning_rate": 1.1260601220077789e-05, "loss": 2.5999, "step": 13527 }, { "epoch": 0.476686311301391, "grad_norm": 7.088210105895996, "learning_rate": 1.1259469039485931e-05, "loss": 2.7524, "step": 13528 }, { "epoch": 0.4767215483143494, "grad_norm": 8.005060195922852, "learning_rate": 1.1258336842489324e-05, "loss": 3.2564, "step": 13529 }, { "epoch": 0.4767567853273078, "grad_norm": 4.373584270477295, "learning_rate": 1.1257204629102724e-05, "loss": 2.8037, "step": 13530 }, { "epoch": 0.47679202234026624, "grad_norm": 4.349547386169434, "learning_rate": 1.1256072399340873e-05, "loss": 2.3626, "step": 13531 }, { "epoch": 0.4768272593532246, "grad_norm": 9.779715538024902, "learning_rate": 1.1254940153218517e-05, "loss": 2.9125, "step": 13532 }, { "epoch": 0.47686249636618305, "grad_norm": 3.2690494060516357, "learning_rate": 1.1253807890750408e-05, "loss": 0.7402, "step": 13533 }, { "epoch": 0.47689773337914143, "grad_norm": 1.3767411708831787, "learning_rate": 1.1252675611951289e-05, "loss": 0.9501, "step": 13534 }, { "epoch": 0.47693297039209986, "grad_norm": 12.344329833984375, "learning_rate": 1.1251543316835911e-05, "loss": 2.7167, "step": 13535 }, { "epoch": 0.4769682074050583, "grad_norm": 3.0909764766693115, "learning_rate": 1.1250411005419024e-05, "loss": 0.8657, "step": 13536 }, { "epoch": 0.4770034444180167, "grad_norm": 7.587392330169678, "learning_rate": 1.1249278677715371e-05, "loss": 3.3451, "step": 13537 }, { "epoch": 0.4770386814309751, "grad_norm": 15.65914535522461, "learning_rate": 1.1248146333739706e-05, "loss": 3.1833, "step": 13538 }, { "epoch": 0.4770739184439335, "grad_norm": 9.268933296203613, "learning_rate": 1.1247013973506779e-05, "loss": 3.6059, "step": 13539 }, { "epoch": 0.4771091554568919, "grad_norm": 3.020867109298706, "learning_rate": 1.124588159703133e-05, "loss": 1.039, "step": 13540 }, { "epoch": 0.47714439246985035, "grad_norm": 4.893120288848877, "learning_rate": 1.124474920432812e-05, "loss": 3.4688, "step": 13541 }, { "epoch": 0.47717962948280873, "grad_norm": 4.230764865875244, "learning_rate": 1.1243616795411892e-05, "loss": 2.6887, "step": 13542 }, { "epoch": 0.47721486649576716, "grad_norm": 1.7417654991149902, "learning_rate": 1.1242484370297394e-05, "loss": 0.6913, "step": 13543 }, { "epoch": 0.47725010350872554, "grad_norm": 1.2396210432052612, "learning_rate": 1.124135192899938e-05, "loss": 0.8547, "step": 13544 }, { "epoch": 0.477285340521684, "grad_norm": 2.1253864765167236, "learning_rate": 1.1240219471532603e-05, "loss": 1.0965, "step": 13545 }, { "epoch": 0.4773205775346424, "grad_norm": 1.7697877883911133, "learning_rate": 1.1239086997911803e-05, "loss": 0.9433, "step": 13546 }, { "epoch": 0.4773558145476008, "grad_norm": 10.214788436889648, "learning_rate": 1.123795450815174e-05, "loss": 5.0034, "step": 13547 }, { "epoch": 0.4773910515605592, "grad_norm": 0.8255372047424316, "learning_rate": 1.123682200226716e-05, "loss": 0.629, "step": 13548 }, { "epoch": 0.4774262885735176, "grad_norm": 8.37366771697998, "learning_rate": 1.123568948027282e-05, "loss": 4.9102, "step": 13549 }, { "epoch": 0.47746152558647603, "grad_norm": 1.339390516281128, "learning_rate": 1.1234556942183465e-05, "loss": 1.0251, "step": 13550 }, { "epoch": 0.47749676259943447, "grad_norm": 0.9620997309684753, "learning_rate": 1.1233424388013847e-05, "loss": 0.7534, "step": 13551 }, { "epoch": 0.47753199961239284, "grad_norm": 6.003520488739014, "learning_rate": 1.1232291817778723e-05, "loss": 2.9795, "step": 13552 }, { "epoch": 0.4775672366253513, "grad_norm": 1.8114306926727295, "learning_rate": 1.1231159231492837e-05, "loss": 0.8919, "step": 13553 }, { "epoch": 0.47760247363830965, "grad_norm": 6.011003017425537, "learning_rate": 1.1230026629170947e-05, "loss": 3.1875, "step": 13554 }, { "epoch": 0.4776377106512681, "grad_norm": 10.350001335144043, "learning_rate": 1.1228894010827802e-05, "loss": 7.3902, "step": 13555 }, { "epoch": 0.4776729476642265, "grad_norm": 4.068069934844971, "learning_rate": 1.1227761376478157e-05, "loss": 2.949, "step": 13556 }, { "epoch": 0.4777081846771849, "grad_norm": 4.904843330383301, "learning_rate": 1.1226628726136763e-05, "loss": 3.1383, "step": 13557 }, { "epoch": 0.47774342169014333, "grad_norm": 4.740001201629639, "learning_rate": 1.1225496059818377e-05, "loss": 3.356, "step": 13558 }, { "epoch": 0.4777786587031017, "grad_norm": 3.047606945037842, "learning_rate": 1.1224363377537744e-05, "loss": 2.8231, "step": 13559 }, { "epoch": 0.47781389571606014, "grad_norm": 0.8769275546073914, "learning_rate": 1.1223230679309627e-05, "loss": 0.7784, "step": 13560 }, { "epoch": 0.4778491327290186, "grad_norm": 17.38772964477539, "learning_rate": 1.122209796514877e-05, "loss": 3.3336, "step": 13561 }, { "epoch": 0.47788436974197696, "grad_norm": 1.092615008354187, "learning_rate": 1.1220965235069934e-05, "loss": 0.8952, "step": 13562 }, { "epoch": 0.4779196067549354, "grad_norm": 6.115655899047852, "learning_rate": 1.1219832489087869e-05, "loss": 5.4949, "step": 13563 }, { "epoch": 0.4779548437678938, "grad_norm": 22.313873291015625, "learning_rate": 1.1218699727217332e-05, "loss": 4.5857, "step": 13564 }, { "epoch": 0.4779900807808522, "grad_norm": 1.2040586471557617, "learning_rate": 1.1217566949473072e-05, "loss": 0.8799, "step": 13565 }, { "epoch": 0.47802531779381063, "grad_norm": 6.024569511413574, "learning_rate": 1.1216434155869852e-05, "loss": 3.39, "step": 13566 }, { "epoch": 0.478060554806769, "grad_norm": 1.397051215171814, "learning_rate": 1.121530134642242e-05, "loss": 0.9554, "step": 13567 }, { "epoch": 0.47809579181972744, "grad_norm": 20.850156784057617, "learning_rate": 1.1214168521145534e-05, "loss": 3.3312, "step": 13568 }, { "epoch": 0.4781310288326859, "grad_norm": 14.13221549987793, "learning_rate": 1.1213035680053949e-05, "loss": 7.998, "step": 13569 }, { "epoch": 0.47816626584564426, "grad_norm": 3.4200968742370605, "learning_rate": 1.1211902823162418e-05, "loss": 2.8339, "step": 13570 }, { "epoch": 0.4782015028586027, "grad_norm": 1.2286204099655151, "learning_rate": 1.1210769950485699e-05, "loss": 1.0668, "step": 13571 }, { "epoch": 0.47823673987156107, "grad_norm": 8.958088874816895, "learning_rate": 1.1209637062038548e-05, "loss": 3.3328, "step": 13572 }, { "epoch": 0.4782719768845195, "grad_norm": 1.7056783437728882, "learning_rate": 1.1208504157835718e-05, "loss": 0.8459, "step": 13573 }, { "epoch": 0.47830721389747793, "grad_norm": 3.70051908493042, "learning_rate": 1.1207371237891969e-05, "loss": 3.1487, "step": 13574 }, { "epoch": 0.4783424509104363, "grad_norm": 2.175673246383667, "learning_rate": 1.1206238302222054e-05, "loss": 1.2953, "step": 13575 }, { "epoch": 0.47837768792339475, "grad_norm": 11.08649730682373, "learning_rate": 1.1205105350840733e-05, "loss": 6.5163, "step": 13576 }, { "epoch": 0.4784129249363531, "grad_norm": 9.646316528320312, "learning_rate": 1.1203972383762764e-05, "loss": 5.1726, "step": 13577 }, { "epoch": 0.47844816194931156, "grad_norm": 4.152552604675293, "learning_rate": 1.1202839401002896e-05, "loss": 2.9944, "step": 13578 }, { "epoch": 0.47848339896227, "grad_norm": 1.624634027481079, "learning_rate": 1.1201706402575896e-05, "loss": 0.8918, "step": 13579 }, { "epoch": 0.47851863597522837, "grad_norm": 1.187048077583313, "learning_rate": 1.1200573388496517e-05, "loss": 0.973, "step": 13580 }, { "epoch": 0.4785538729881868, "grad_norm": 2.7165963649749756, "learning_rate": 1.1199440358779512e-05, "loss": 1.2947, "step": 13581 }, { "epoch": 0.4785891100011452, "grad_norm": 5.832449436187744, "learning_rate": 1.1198307313439646e-05, "loss": 3.6997, "step": 13582 }, { "epoch": 0.4786243470141036, "grad_norm": 0.9834335446357727, "learning_rate": 1.1197174252491676e-05, "loss": 1.1067, "step": 13583 }, { "epoch": 0.47865958402706205, "grad_norm": 2.055088996887207, "learning_rate": 1.1196041175950359e-05, "loss": 0.9581, "step": 13584 }, { "epoch": 0.4786948210400204, "grad_norm": 3.231278896331787, "learning_rate": 1.1194908083830453e-05, "loss": 2.3416, "step": 13585 }, { "epoch": 0.47873005805297886, "grad_norm": 6.766650676727295, "learning_rate": 1.1193774976146715e-05, "loss": 4.426, "step": 13586 }, { "epoch": 0.47876529506593724, "grad_norm": 1.0949254035949707, "learning_rate": 1.1192641852913908e-05, "loss": 0.889, "step": 13587 }, { "epoch": 0.47880053207889567, "grad_norm": 11.9750337600708, "learning_rate": 1.1191508714146788e-05, "loss": 3.0628, "step": 13588 }, { "epoch": 0.4788357690918541, "grad_norm": 1.1315977573394775, "learning_rate": 1.1190375559860117e-05, "loss": 0.923, "step": 13589 }, { "epoch": 0.4788710061048125, "grad_norm": 5.3194475173950195, "learning_rate": 1.1189242390068648e-05, "loss": 3.1949, "step": 13590 }, { "epoch": 0.4789062431177709, "grad_norm": 1.3615776300430298, "learning_rate": 1.1188109204787149e-05, "loss": 1.0245, "step": 13591 }, { "epoch": 0.4789414801307293, "grad_norm": 2.1293606758117676, "learning_rate": 1.1186976004030375e-05, "loss": 1.1034, "step": 13592 }, { "epoch": 0.4789767171436877, "grad_norm": 2.741607189178467, "learning_rate": 1.1185842787813088e-05, "loss": 1.0236, "step": 13593 }, { "epoch": 0.47901195415664616, "grad_norm": 3.4902236461639404, "learning_rate": 1.1184709556150049e-05, "loss": 2.7895, "step": 13594 }, { "epoch": 0.47904719116960454, "grad_norm": 7.285216808319092, "learning_rate": 1.1183576309056014e-05, "loss": 4.8012, "step": 13595 }, { "epoch": 0.47908242818256297, "grad_norm": 1.6196672916412354, "learning_rate": 1.118244304654575e-05, "loss": 0.8034, "step": 13596 }, { "epoch": 0.4791176651955214, "grad_norm": 3.5764505863189697, "learning_rate": 1.1181309768634012e-05, "loss": 2.8725, "step": 13597 }, { "epoch": 0.4791529022084798, "grad_norm": 2.612208604812622, "learning_rate": 1.1180176475335567e-05, "loss": 0.7738, "step": 13598 }, { "epoch": 0.4791881392214382, "grad_norm": 3.7595772743225098, "learning_rate": 1.1179043166665171e-05, "loss": 3.3591, "step": 13599 }, { "epoch": 0.4792233762343966, "grad_norm": 4.883491039276123, "learning_rate": 1.1177909842637586e-05, "loss": 3.3537, "step": 13600 }, { "epoch": 0.479258613247355, "grad_norm": 3.3470511436462402, "learning_rate": 1.1176776503267579e-05, "loss": 3.0262, "step": 13601 }, { "epoch": 0.47929385026031346, "grad_norm": 9.157991409301758, "learning_rate": 1.1175643148569905e-05, "loss": 3.5, "step": 13602 }, { "epoch": 0.47932908727327184, "grad_norm": 1.1722301244735718, "learning_rate": 1.117450977855933e-05, "loss": 0.8596, "step": 13603 }, { "epoch": 0.47936432428623027, "grad_norm": 1.115877628326416, "learning_rate": 1.1173376393250619e-05, "loss": 0.8395, "step": 13604 }, { "epoch": 0.47939956129918865, "grad_norm": 1.3281798362731934, "learning_rate": 1.1172242992658524e-05, "loss": 0.787, "step": 13605 }, { "epoch": 0.4794347983121471, "grad_norm": 1.3281798362731934, "learning_rate": 1.1172242992658524e-05, "loss": 3.1617, "step": 13606 }, { "epoch": 0.4794700353251055, "grad_norm": 6.339447975158691, "learning_rate": 1.117110957679782e-05, "loss": 3.3121, "step": 13607 }, { "epoch": 0.4795052723380639, "grad_norm": 12.700398445129395, "learning_rate": 1.1169976145683264e-05, "loss": 5.0781, "step": 13608 }, { "epoch": 0.4795405093510223, "grad_norm": 2.113797187805176, "learning_rate": 1.116884269932962e-05, "loss": 0.8417, "step": 13609 }, { "epoch": 0.4795757463639807, "grad_norm": 1.4581457376480103, "learning_rate": 1.1167709237751648e-05, "loss": 0.852, "step": 13610 }, { "epoch": 0.47961098337693914, "grad_norm": 1.237447738647461, "learning_rate": 1.1166575760964116e-05, "loss": 0.8393, "step": 13611 }, { "epoch": 0.47964622038989757, "grad_norm": 2.025085926055908, "learning_rate": 1.1165442268981787e-05, "loss": 0.9865, "step": 13612 }, { "epoch": 0.47968145740285595, "grad_norm": 3.573622941970825, "learning_rate": 1.1164308761819423e-05, "loss": 3.1578, "step": 13613 }, { "epoch": 0.4797166944158144, "grad_norm": 1.2024269104003906, "learning_rate": 1.1163175239491788e-05, "loss": 0.9855, "step": 13614 }, { "epoch": 0.47975193142877276, "grad_norm": 4.853854179382324, "learning_rate": 1.1162041702013649e-05, "loss": 3.3822, "step": 13615 }, { "epoch": 0.4797871684417312, "grad_norm": 2.084805488586426, "learning_rate": 1.1160908149399772e-05, "loss": 0.875, "step": 13616 }, { "epoch": 0.4798224054546896, "grad_norm": 3.255004405975342, "learning_rate": 1.1159774581664911e-05, "loss": 2.914, "step": 13617 }, { "epoch": 0.479857642467648, "grad_norm": 2.4031612873077393, "learning_rate": 1.1158640998823845e-05, "loss": 0.8639, "step": 13618 }, { "epoch": 0.47989287948060644, "grad_norm": 2.867781400680542, "learning_rate": 1.115750740089133e-05, "loss": 3.0276, "step": 13619 }, { "epoch": 0.4799281164935648, "grad_norm": 5.493002891540527, "learning_rate": 1.1156373787882133e-05, "loss": 3.4222, "step": 13620 }, { "epoch": 0.47996335350652325, "grad_norm": 4.744209289550781, "learning_rate": 1.115524015981102e-05, "loss": 3.1657, "step": 13621 }, { "epoch": 0.4799985905194817, "grad_norm": 6.498272895812988, "learning_rate": 1.1154106516692758e-05, "loss": 5.2108, "step": 13622 }, { "epoch": 0.48003382753244006, "grad_norm": 1.5137392282485962, "learning_rate": 1.115297285854211e-05, "loss": 1.1543, "step": 13623 }, { "epoch": 0.4800690645453985, "grad_norm": 2.4205386638641357, "learning_rate": 1.1151839185373847e-05, "loss": 2.2801, "step": 13624 }, { "epoch": 0.48010430155835687, "grad_norm": 11.096582412719727, "learning_rate": 1.1150705497202728e-05, "loss": 6.8483, "step": 13625 }, { "epoch": 0.4801395385713153, "grad_norm": 3.8997766971588135, "learning_rate": 1.1149571794043525e-05, "loss": 3.1033, "step": 13626 }, { "epoch": 0.48017477558427374, "grad_norm": 1.434492826461792, "learning_rate": 1.1148438075911003e-05, "loss": 0.8574, "step": 13627 }, { "epoch": 0.4802100125972321, "grad_norm": 1.1931238174438477, "learning_rate": 1.114730434281993e-05, "loss": 0.6919, "step": 13628 }, { "epoch": 0.48024524961019055, "grad_norm": 1.4374439716339111, "learning_rate": 1.1146170594785068e-05, "loss": 1.0147, "step": 13629 }, { "epoch": 0.480280486623149, "grad_norm": 5.123417377471924, "learning_rate": 1.114503683182119e-05, "loss": 4.822, "step": 13630 }, { "epoch": 0.48031572363610736, "grad_norm": 16.234891891479492, "learning_rate": 1.1143903053943061e-05, "loss": 2.9447, "step": 13631 }, { "epoch": 0.4803509606490658, "grad_norm": 7.426626682281494, "learning_rate": 1.1142769261165451e-05, "loss": 5.6539, "step": 13632 }, { "epoch": 0.4803861976620242, "grad_norm": 7.08326530456543, "learning_rate": 1.1141635453503123e-05, "loss": 5.2031, "step": 13633 }, { "epoch": 0.4804214346749826, "grad_norm": 1.01764976978302, "learning_rate": 1.1140501630970852e-05, "loss": 0.9469, "step": 13634 }, { "epoch": 0.48045667168794104, "grad_norm": 6.910699844360352, "learning_rate": 1.11393677935834e-05, "loss": 3.2807, "step": 13635 }, { "epoch": 0.4804919087008994, "grad_norm": 5.084510326385498, "learning_rate": 1.1138233941355536e-05, "loss": 4.9796, "step": 13636 }, { "epoch": 0.48052714571385785, "grad_norm": 1.5982120037078857, "learning_rate": 1.113710007430203e-05, "loss": 0.9263, "step": 13637 }, { "epoch": 0.48056238272681623, "grad_norm": 3.806076765060425, "learning_rate": 1.1135966192437653e-05, "loss": 2.9487, "step": 13638 }, { "epoch": 0.48059761973977466, "grad_norm": 3.1129603385925293, "learning_rate": 1.113483229577717e-05, "loss": 2.4749, "step": 13639 }, { "epoch": 0.4806328567527331, "grad_norm": 6.28737211227417, "learning_rate": 1.113369838433535e-05, "loss": 3.3081, "step": 13640 }, { "epoch": 0.4806680937656915, "grad_norm": 7.470504283905029, "learning_rate": 1.1132564458126967e-05, "loss": 2.8235, "step": 13641 }, { "epoch": 0.4807033307786499, "grad_norm": 1.9564510583877563, "learning_rate": 1.1131430517166786e-05, "loss": 0.8139, "step": 13642 }, { "epoch": 0.4807385677916083, "grad_norm": 3.6410646438598633, "learning_rate": 1.113029656146958e-05, "loss": 2.8836, "step": 13643 }, { "epoch": 0.4807738048045667, "grad_norm": 5.929222583770752, "learning_rate": 1.1129162591050113e-05, "loss": 5.0377, "step": 13644 }, { "epoch": 0.48080904181752515, "grad_norm": 7.988402843475342, "learning_rate": 1.1128028605923165e-05, "loss": 5.1355, "step": 13645 }, { "epoch": 0.48084427883048353, "grad_norm": 1.4075270891189575, "learning_rate": 1.1126894606103498e-05, "loss": 1.1421, "step": 13646 }, { "epoch": 0.48087951584344196, "grad_norm": 8.522134780883789, "learning_rate": 1.1125760591605885e-05, "loss": 5.2132, "step": 13647 }, { "epoch": 0.48091475285640034, "grad_norm": 2.2838287353515625, "learning_rate": 1.1124626562445098e-05, "loss": 1.0759, "step": 13648 }, { "epoch": 0.4809499898693588, "grad_norm": 2.2548656463623047, "learning_rate": 1.1123492518635906e-05, "loss": 0.7851, "step": 13649 }, { "epoch": 0.4809852268823172, "grad_norm": 6.482716083526611, "learning_rate": 1.112235846019308e-05, "loss": 4.2998, "step": 13650 }, { "epoch": 0.4810204638952756, "grad_norm": 1.0103483200073242, "learning_rate": 1.1121224387131398e-05, "loss": 0.7124, "step": 13651 }, { "epoch": 0.481055700908234, "grad_norm": 12.512402534484863, "learning_rate": 1.1120090299465618e-05, "loss": 4.2754, "step": 13652 }, { "epoch": 0.4810909379211924, "grad_norm": 17.143991470336914, "learning_rate": 1.1118956197210522e-05, "loss": 3.7605, "step": 13653 }, { "epoch": 0.48112617493415083, "grad_norm": 4.817492961883545, "learning_rate": 1.111782208038088e-05, "loss": 2.8106, "step": 13654 }, { "epoch": 0.48116141194710926, "grad_norm": 1.329444408416748, "learning_rate": 1.1116687948991461e-05, "loss": 0.7583, "step": 13655 }, { "epoch": 0.48119664896006764, "grad_norm": 3.9026970863342285, "learning_rate": 1.111555380305704e-05, "loss": 2.7668, "step": 13656 }, { "epoch": 0.4812318859730261, "grad_norm": 1.2345807552337646, "learning_rate": 1.111441964259239e-05, "loss": 0.9067, "step": 13657 }, { "epoch": 0.48126712298598445, "grad_norm": 3.6639583110809326, "learning_rate": 1.111328546761228e-05, "loss": 2.932, "step": 13658 }, { "epoch": 0.4813023599989429, "grad_norm": 2.917113780975342, "learning_rate": 1.1112151278131484e-05, "loss": 2.7173, "step": 13659 }, { "epoch": 0.4813375970119013, "grad_norm": 8.047598838806152, "learning_rate": 1.1111017074164777e-05, "loss": 5.9896, "step": 13660 }, { "epoch": 0.4813728340248597, "grad_norm": 3.7954113483428955, "learning_rate": 1.1109882855726935e-05, "loss": 2.9599, "step": 13661 }, { "epoch": 0.48140807103781813, "grad_norm": 6.451779842376709, "learning_rate": 1.1108748622832725e-05, "loss": 3.2656, "step": 13662 }, { "epoch": 0.4814433080507765, "grad_norm": 1.211426019668579, "learning_rate": 1.110761437549692e-05, "loss": 0.9842, "step": 13663 }, { "epoch": 0.48147854506373494, "grad_norm": 10.183989524841309, "learning_rate": 1.1106480113734302e-05, "loss": 5.7451, "step": 13664 }, { "epoch": 0.4815137820766934, "grad_norm": 3.8543782234191895, "learning_rate": 1.1105345837559636e-05, "loss": 2.1316, "step": 13665 }, { "epoch": 0.48154901908965175, "grad_norm": 2.6472933292388916, "learning_rate": 1.11042115469877e-05, "loss": 0.8175, "step": 13666 }, { "epoch": 0.4815842561026102, "grad_norm": 1.123127818107605, "learning_rate": 1.1103077242033267e-05, "loss": 0.8679, "step": 13667 }, { "epoch": 0.4816194931155686, "grad_norm": 1.4983876943588257, "learning_rate": 1.1101942922711114e-05, "loss": 0.9079, "step": 13668 }, { "epoch": 0.481654730128527, "grad_norm": 1.8345931768417358, "learning_rate": 1.1100808589036014e-05, "loss": 0.8861, "step": 13669 }, { "epoch": 0.48168996714148543, "grad_norm": 8.371312141418457, "learning_rate": 1.1099674241022743e-05, "loss": 5.1352, "step": 13670 }, { "epoch": 0.4817252041544438, "grad_norm": 5.426936149597168, "learning_rate": 1.109853987868607e-05, "loss": 3.3067, "step": 13671 }, { "epoch": 0.48176044116740224, "grad_norm": 7.761175632476807, "learning_rate": 1.1097405502040782e-05, "loss": 5.2573, "step": 13672 }, { "epoch": 0.4817956781803607, "grad_norm": 1.41243577003479, "learning_rate": 1.1096271111101644e-05, "loss": 1.1434, "step": 13673 }, { "epoch": 0.48183091519331905, "grad_norm": 1.0017549991607666, "learning_rate": 1.1095136705883433e-05, "loss": 0.8295, "step": 13674 }, { "epoch": 0.4818661522062775, "grad_norm": 32.27690887451172, "learning_rate": 1.1094002286400931e-05, "loss": 6.0807, "step": 13675 }, { "epoch": 0.48190138921923587, "grad_norm": 7.073572158813477, "learning_rate": 1.1092867852668908e-05, "loss": 5.7085, "step": 13676 }, { "epoch": 0.4819366262321943, "grad_norm": 4.2859206199646, "learning_rate": 1.109173340470214e-05, "loss": 2.9648, "step": 13677 }, { "epoch": 0.48197186324515273, "grad_norm": 4.008100509643555, "learning_rate": 1.1090598942515408e-05, "loss": 0.8997, "step": 13678 }, { "epoch": 0.4820071002581111, "grad_norm": 5.646526336669922, "learning_rate": 1.1089464466123485e-05, "loss": 3.5288, "step": 13679 }, { "epoch": 0.48204233727106954, "grad_norm": 7.217013359069824, "learning_rate": 1.108832997554115e-05, "loss": 3.3397, "step": 13680 }, { "epoch": 0.4820775742840279, "grad_norm": 1.584633708000183, "learning_rate": 1.1087195470783178e-05, "loss": 0.8615, "step": 13681 }, { "epoch": 0.48211281129698635, "grad_norm": 1.861796259880066, "learning_rate": 1.1086060951864343e-05, "loss": 0.7215, "step": 13682 }, { "epoch": 0.4821480483099448, "grad_norm": 6.157980918884277, "learning_rate": 1.108492641879943e-05, "loss": 3.0333, "step": 13683 }, { "epoch": 0.48218328532290317, "grad_norm": 5.545621395111084, "learning_rate": 1.108379187160321e-05, "loss": 3.1138, "step": 13684 }, { "epoch": 0.4822185223358616, "grad_norm": 2.9708571434020996, "learning_rate": 1.1082657310290466e-05, "loss": 2.6733, "step": 13685 }, { "epoch": 0.48225375934882, "grad_norm": 10.280194282531738, "learning_rate": 1.1081522734875969e-05, "loss": 5.5836, "step": 13686 }, { "epoch": 0.4822889963617784, "grad_norm": 1.225711464881897, "learning_rate": 1.1080388145374503e-05, "loss": 0.9713, "step": 13687 }, { "epoch": 0.48232423337473684, "grad_norm": 4.493772506713867, "learning_rate": 1.1079253541800842e-05, "loss": 3.5024, "step": 13688 }, { "epoch": 0.4823594703876952, "grad_norm": 3.354321241378784, "learning_rate": 1.107811892416977e-05, "loss": 2.9079, "step": 13689 }, { "epoch": 0.48239470740065366, "grad_norm": 2.6968612670898438, "learning_rate": 1.1076984292496056e-05, "loss": 2.4225, "step": 13690 }, { "epoch": 0.48242994441361203, "grad_norm": 8.80734634399414, "learning_rate": 1.1075849646794488e-05, "loss": 3.1923, "step": 13691 }, { "epoch": 0.48246518142657047, "grad_norm": 1.039526104927063, "learning_rate": 1.1074714987079839e-05, "loss": 1.1867, "step": 13692 }, { "epoch": 0.4825004184395289, "grad_norm": 4.810908794403076, "learning_rate": 1.1073580313366892e-05, "loss": 2.9903, "step": 13693 }, { "epoch": 0.4825356554524873, "grad_norm": 1.3591023683547974, "learning_rate": 1.1072445625670425e-05, "loss": 0.9715, "step": 13694 }, { "epoch": 0.4825708924654457, "grad_norm": 1.2287503480911255, "learning_rate": 1.1071310924005214e-05, "loss": 1.2356, "step": 13695 }, { "epoch": 0.4826061294784041, "grad_norm": 5.402595043182373, "learning_rate": 1.1070176208386045e-05, "loss": 3.3508, "step": 13696 }, { "epoch": 0.4826413664913625, "grad_norm": 2.5040807723999023, "learning_rate": 1.1069041478827694e-05, "loss": 0.953, "step": 13697 }, { "epoch": 0.48267660350432096, "grad_norm": 1.412406086921692, "learning_rate": 1.106790673534494e-05, "loss": 1.0166, "step": 13698 }, { "epoch": 0.48271184051727933, "grad_norm": 5.84859037399292, "learning_rate": 1.1066771977952566e-05, "loss": 3.3013, "step": 13699 }, { "epoch": 0.48274707753023777, "grad_norm": 5.217444896697998, "learning_rate": 1.1065637206665354e-05, "loss": 2.8499, "step": 13700 }, { "epoch": 0.4827823145431962, "grad_norm": 1.523193359375, "learning_rate": 1.1064502421498076e-05, "loss": 1.0222, "step": 13701 }, { "epoch": 0.4828175515561546, "grad_norm": 9.760246276855469, "learning_rate": 1.1063367622465524e-05, "loss": 2.1508, "step": 13702 }, { "epoch": 0.482852788569113, "grad_norm": 1.05363130569458, "learning_rate": 1.1062232809582468e-05, "loss": 0.8233, "step": 13703 }, { "epoch": 0.4828880255820714, "grad_norm": 0.9829039573669434, "learning_rate": 1.1061097982863698e-05, "loss": 1.0625, "step": 13704 }, { "epoch": 0.4829232625950298, "grad_norm": 1.5572789907455444, "learning_rate": 1.1059963142323991e-05, "loss": 0.8255, "step": 13705 }, { "epoch": 0.48295849960798826, "grad_norm": 1.5581127405166626, "learning_rate": 1.1058828287978127e-05, "loss": 0.7819, "step": 13706 }, { "epoch": 0.48299373662094663, "grad_norm": 12.291118621826172, "learning_rate": 1.1057693419840891e-05, "loss": 3.6114, "step": 13707 }, { "epoch": 0.48302897363390507, "grad_norm": 1.0212500095367432, "learning_rate": 1.1056558537927064e-05, "loss": 0.9207, "step": 13708 }, { "epoch": 0.48306421064686345, "grad_norm": 3.3977317810058594, "learning_rate": 1.1055423642251427e-05, "loss": 3.538, "step": 13709 }, { "epoch": 0.4830994476598219, "grad_norm": 4.991691589355469, "learning_rate": 1.1054288732828765e-05, "loss": 3.4514, "step": 13710 }, { "epoch": 0.4831346846727803, "grad_norm": 1.2189695835113525, "learning_rate": 1.1053153809673852e-05, "loss": 0.9364, "step": 13711 }, { "epoch": 0.4831699216857387, "grad_norm": 2.5248193740844727, "learning_rate": 1.1052018872801481e-05, "loss": 2.4288, "step": 13712 }, { "epoch": 0.4832051586986971, "grad_norm": 1.0714441537857056, "learning_rate": 1.105088392222643e-05, "loss": 1.1263, "step": 13713 }, { "epoch": 0.4832403957116555, "grad_norm": 4.252325057983398, "learning_rate": 1.104974895796348e-05, "loss": 3.2275, "step": 13714 }, { "epoch": 0.48327563272461393, "grad_norm": 8.144291877746582, "learning_rate": 1.1048613980027418e-05, "loss": 5.1647, "step": 13715 }, { "epoch": 0.48331086973757237, "grad_norm": 1.9893467426300049, "learning_rate": 1.1047478988433025e-05, "loss": 1.0392, "step": 13716 }, { "epoch": 0.48334610675053075, "grad_norm": 6.855242729187012, "learning_rate": 1.1046343983195083e-05, "loss": 2.8346, "step": 13717 }, { "epoch": 0.4833813437634892, "grad_norm": 1.1232960224151611, "learning_rate": 1.1045208964328378e-05, "loss": 0.7448, "step": 13718 }, { "epoch": 0.48341658077644756, "grad_norm": 38.955963134765625, "learning_rate": 1.1044073931847696e-05, "loss": 3.026, "step": 13719 }, { "epoch": 0.483451817789406, "grad_norm": 1.6302355527877808, "learning_rate": 1.1042938885767813e-05, "loss": 0.7205, "step": 13720 }, { "epoch": 0.4834870548023644, "grad_norm": 1.352381944656372, "learning_rate": 1.1041803826103525e-05, "loss": 0.7909, "step": 13721 }, { "epoch": 0.4835222918153228, "grad_norm": 4.635409355163574, "learning_rate": 1.1040668752869604e-05, "loss": 3.7005, "step": 13722 }, { "epoch": 0.48355752882828124, "grad_norm": 8.275120735168457, "learning_rate": 1.1039533666080842e-05, "loss": 4.9827, "step": 13723 }, { "epoch": 0.4835927658412396, "grad_norm": 3.7184932231903076, "learning_rate": 1.1038398565752022e-05, "loss": 2.5191, "step": 13724 }, { "epoch": 0.48362800285419805, "grad_norm": 2.448002338409424, "learning_rate": 1.1037263451897926e-05, "loss": 2.8212, "step": 13725 }, { "epoch": 0.4836632398671565, "grad_norm": 2.109799385070801, "learning_rate": 1.1036128324533344e-05, "loss": 0.957, "step": 13726 }, { "epoch": 0.48369847688011486, "grad_norm": 11.326355934143066, "learning_rate": 1.1034993183673058e-05, "loss": 5.0925, "step": 13727 }, { "epoch": 0.4837337138930733, "grad_norm": 4.663186550140381, "learning_rate": 1.1033858029331853e-05, "loss": 4.6305, "step": 13728 }, { "epoch": 0.48376895090603167, "grad_norm": 7.318153381347656, "learning_rate": 1.1032722861524516e-05, "loss": 5.2966, "step": 13729 }, { "epoch": 0.4838041879189901, "grad_norm": 4.224973201751709, "learning_rate": 1.103158768026583e-05, "loss": 3.2732, "step": 13730 }, { "epoch": 0.48383942493194854, "grad_norm": 8.729077339172363, "learning_rate": 1.1030452485570586e-05, "loss": 5.1644, "step": 13731 }, { "epoch": 0.4838746619449069, "grad_norm": 4.263499736785889, "learning_rate": 1.1029317277453566e-05, "loss": 2.5699, "step": 13732 }, { "epoch": 0.48390989895786535, "grad_norm": 7.8155436515808105, "learning_rate": 1.102818205592956e-05, "loss": 4.7378, "step": 13733 }, { "epoch": 0.4839451359708238, "grad_norm": 0.9283852577209473, "learning_rate": 1.1027046821013348e-05, "loss": 1.2198, "step": 13734 }, { "epoch": 0.48398037298378216, "grad_norm": 1.0723631381988525, "learning_rate": 1.102591157271972e-05, "loss": 0.915, "step": 13735 }, { "epoch": 0.4840156099967406, "grad_norm": 3.454192876815796, "learning_rate": 1.1024776311063466e-05, "loss": 2.6006, "step": 13736 }, { "epoch": 0.48405084700969897, "grad_norm": 4.449784755706787, "learning_rate": 1.102364103605937e-05, "loss": 2.82, "step": 13737 }, { "epoch": 0.4840860840226574, "grad_norm": 4.82192325592041, "learning_rate": 1.1022505747722219e-05, "loss": 5.3773, "step": 13738 }, { "epoch": 0.48412132103561584, "grad_norm": 7.457801818847656, "learning_rate": 1.1021370446066797e-05, "loss": 4.8847, "step": 13739 }, { "epoch": 0.4841565580485742, "grad_norm": 5.718437671661377, "learning_rate": 1.1020235131107899e-05, "loss": 3.5594, "step": 13740 }, { "epoch": 0.48419179506153265, "grad_norm": 1.0134553909301758, "learning_rate": 1.1019099802860303e-05, "loss": 0.8343, "step": 13741 }, { "epoch": 0.484227032074491, "grad_norm": 2.051600456237793, "learning_rate": 1.1017964461338809e-05, "loss": 0.8898, "step": 13742 }, { "epoch": 0.48426226908744946, "grad_norm": 5.512228012084961, "learning_rate": 1.1016829106558193e-05, "loss": 3.386, "step": 13743 }, { "epoch": 0.4842975061004079, "grad_norm": 2.2143940925598145, "learning_rate": 1.101569373853325e-05, "loss": 0.9526, "step": 13744 }, { "epoch": 0.48433274311336627, "grad_norm": 1.1695387363433838, "learning_rate": 1.1014558357278768e-05, "loss": 0.8371, "step": 13745 }, { "epoch": 0.4843679801263247, "grad_norm": 3.5063295364379883, "learning_rate": 1.1013422962809534e-05, "loss": 3.3161, "step": 13746 }, { "epoch": 0.4844032171392831, "grad_norm": 2.054718017578125, "learning_rate": 1.1012287555140336e-05, "loss": 1.0161, "step": 13747 }, { "epoch": 0.4844384541522415, "grad_norm": 1.8140062093734741, "learning_rate": 1.1011152134285965e-05, "loss": 0.9261, "step": 13748 }, { "epoch": 0.48447369116519995, "grad_norm": 1.4388775825500488, "learning_rate": 1.1010016700261206e-05, "loss": 0.9287, "step": 13749 }, { "epoch": 0.4845089281781583, "grad_norm": 1.9179871082305908, "learning_rate": 1.1008881253080855e-05, "loss": 0.9516, "step": 13750 }, { "epoch": 0.48454416519111676, "grad_norm": 2.174973487854004, "learning_rate": 1.1007745792759692e-05, "loss": 0.8763, "step": 13751 }, { "epoch": 0.48457940220407514, "grad_norm": 1.1440246105194092, "learning_rate": 1.1006610319312513e-05, "loss": 0.9585, "step": 13752 }, { "epoch": 0.48461463921703357, "grad_norm": 1.3797168731689453, "learning_rate": 1.1005474832754108e-05, "loss": 0.9153, "step": 13753 }, { "epoch": 0.484649876229992, "grad_norm": 1.6654133796691895, "learning_rate": 1.1004339333099265e-05, "loss": 0.6858, "step": 13754 }, { "epoch": 0.4846851132429504, "grad_norm": 6.431890487670898, "learning_rate": 1.1003203820362773e-05, "loss": 3.202, "step": 13755 }, { "epoch": 0.4847203502559088, "grad_norm": 8.22154712677002, "learning_rate": 1.1002068294559424e-05, "loss": 5.1935, "step": 13756 }, { "epoch": 0.4847555872688672, "grad_norm": 6.739893436431885, "learning_rate": 1.1000932755704008e-05, "loss": 3.3442, "step": 13757 }, { "epoch": 0.4847908242818256, "grad_norm": 5.0988874435424805, "learning_rate": 1.0999797203811316e-05, "loss": 3.1726, "step": 13758 }, { "epoch": 0.48482606129478406, "grad_norm": 2.015437364578247, "learning_rate": 1.0998661638896139e-05, "loss": 0.9262, "step": 13759 }, { "epoch": 0.48486129830774244, "grad_norm": 1.8836084604263306, "learning_rate": 1.0997526060973264e-05, "loss": 0.8574, "step": 13760 }, { "epoch": 0.48489653532070087, "grad_norm": 12.392623901367188, "learning_rate": 1.099639047005749e-05, "loss": 3.2504, "step": 13761 }, { "epoch": 0.48493177233365925, "grad_norm": 1.446403980255127, "learning_rate": 1.0995254866163598e-05, "loss": 0.8194, "step": 13762 }, { "epoch": 0.4849670093466177, "grad_norm": 1.7245584726333618, "learning_rate": 1.0994119249306385e-05, "loss": 0.9501, "step": 13763 }, { "epoch": 0.4850022463595761, "grad_norm": 7.100967884063721, "learning_rate": 1.0992983619500646e-05, "loss": 5.1489, "step": 13764 }, { "epoch": 0.4850374833725345, "grad_norm": 12.73105525970459, "learning_rate": 1.0991847976761168e-05, "loss": 6.9745, "step": 13765 }, { "epoch": 0.48507272038549293, "grad_norm": 4.047303199768066, "learning_rate": 1.0990712321102739e-05, "loss": 2.9912, "step": 13766 }, { "epoch": 0.4851079573984513, "grad_norm": 33.51314163208008, "learning_rate": 1.098957665254016e-05, "loss": 5.5814, "step": 13767 }, { "epoch": 0.48514319441140974, "grad_norm": 3.8108556270599365, "learning_rate": 1.0988440971088216e-05, "loss": 2.7702, "step": 13768 }, { "epoch": 0.4851784314243682, "grad_norm": 2.21821928024292, "learning_rate": 1.0987305276761706e-05, "loss": 0.9027, "step": 13769 }, { "epoch": 0.48521366843732655, "grad_norm": 1.828956961631775, "learning_rate": 1.0986169569575418e-05, "loss": 0.9296, "step": 13770 }, { "epoch": 0.485248905450285, "grad_norm": 4.893457412719727, "learning_rate": 1.0985033849544145e-05, "loss": 2.2642, "step": 13771 }, { "epoch": 0.4852841424632434, "grad_norm": 1.2620848417282104, "learning_rate": 1.0983898116682681e-05, "loss": 0.9082, "step": 13772 }, { "epoch": 0.4853193794762018, "grad_norm": 1.0315806865692139, "learning_rate": 1.0982762371005817e-05, "loss": 1.2841, "step": 13773 }, { "epoch": 0.48535461648916023, "grad_norm": 16.7401065826416, "learning_rate": 1.0981626612528352e-05, "loss": 4.9043, "step": 13774 }, { "epoch": 0.4853898535021186, "grad_norm": 1.2386305332183838, "learning_rate": 1.098049084126507e-05, "loss": 1.099, "step": 13775 }, { "epoch": 0.48542509051507704, "grad_norm": 4.121029376983643, "learning_rate": 1.0979355057230774e-05, "loss": 2.9969, "step": 13776 }, { "epoch": 0.4854603275280355, "grad_norm": 26.420751571655273, "learning_rate": 1.097821926044025e-05, "loss": 5.8081, "step": 13777 }, { "epoch": 0.48549556454099385, "grad_norm": 1.1611531972885132, "learning_rate": 1.0977083450908302e-05, "loss": 0.9214, "step": 13778 }, { "epoch": 0.4855308015539523, "grad_norm": 1.3725477457046509, "learning_rate": 1.097594762864971e-05, "loss": 0.9246, "step": 13779 }, { "epoch": 0.48556603856691066, "grad_norm": 3.6672897338867188, "learning_rate": 1.0974811793679281e-05, "loss": 3.0207, "step": 13780 }, { "epoch": 0.4856012755798691, "grad_norm": 1.50191330909729, "learning_rate": 1.0973675946011803e-05, "loss": 1.1285, "step": 13781 }, { "epoch": 0.48563651259282753, "grad_norm": 0.9591001272201538, "learning_rate": 1.0972540085662071e-05, "loss": 1.3088, "step": 13782 }, { "epoch": 0.4856717496057859, "grad_norm": 1.7340896129608154, "learning_rate": 1.0971404212644881e-05, "loss": 0.84, "step": 13783 }, { "epoch": 0.48570698661874434, "grad_norm": 8.612662315368652, "learning_rate": 1.0970268326975026e-05, "loss": 5.2003, "step": 13784 }, { "epoch": 0.4857422236317027, "grad_norm": 6.508815288543701, "learning_rate": 1.0969132428667304e-05, "loss": 5.1027, "step": 13785 }, { "epoch": 0.48577746064466115, "grad_norm": 1.5919212102890015, "learning_rate": 1.096799651773651e-05, "loss": 1.0219, "step": 13786 }, { "epoch": 0.4858126976576196, "grad_norm": 5.9687676429748535, "learning_rate": 1.0966860594197435e-05, "loss": 3.1974, "step": 13787 }, { "epoch": 0.48584793467057796, "grad_norm": 1.3753124475479126, "learning_rate": 1.096572465806488e-05, "loss": 0.9613, "step": 13788 }, { "epoch": 0.4858831716835364, "grad_norm": 8.365299224853516, "learning_rate": 1.0964588709353636e-05, "loss": 5.3327, "step": 13789 }, { "epoch": 0.4859184086964948, "grad_norm": 3.842090606689453, "learning_rate": 1.0963452748078502e-05, "loss": 0.986, "step": 13790 }, { "epoch": 0.4859536457094532, "grad_norm": 1.3947169780731201, "learning_rate": 1.0962316774254274e-05, "loss": 0.7102, "step": 13791 }, { "epoch": 0.48598888272241164, "grad_norm": 11.763649940490723, "learning_rate": 1.096118078789575e-05, "loss": 5.9026, "step": 13792 }, { "epoch": 0.48602411973537, "grad_norm": 1.3117297887802124, "learning_rate": 1.0960044789017718e-05, "loss": 0.9391, "step": 13793 }, { "epoch": 0.48605935674832845, "grad_norm": 5.719181060791016, "learning_rate": 1.0958908777634985e-05, "loss": 3.1738, "step": 13794 }, { "epoch": 0.48609459376128683, "grad_norm": 2.2779345512390137, "learning_rate": 1.095777275376234e-05, "loss": 0.782, "step": 13795 }, { "epoch": 0.48612983077424526, "grad_norm": 1.433830976486206, "learning_rate": 1.0956636717414587e-05, "loss": 1.1115, "step": 13796 }, { "epoch": 0.4861650677872037, "grad_norm": 7.064773082733154, "learning_rate": 1.0955500668606513e-05, "loss": 3.4364, "step": 13797 }, { "epoch": 0.4862003048001621, "grad_norm": 1.4534612894058228, "learning_rate": 1.0954364607352924e-05, "loss": 0.9037, "step": 13798 }, { "epoch": 0.4862355418131205, "grad_norm": 1.2112709283828735, "learning_rate": 1.0953228533668616e-05, "loss": 1.1474, "step": 13799 }, { "epoch": 0.4862707788260789, "grad_norm": 10.577634811401367, "learning_rate": 1.0952092447568384e-05, "loss": 3.1187, "step": 13800 }, { "epoch": 0.4863060158390373, "grad_norm": 4.988076210021973, "learning_rate": 1.0950956349067026e-05, "loss": 3.1594, "step": 13801 }, { "epoch": 0.48634125285199575, "grad_norm": 7.353003978729248, "learning_rate": 1.0949820238179339e-05, "loss": 5.2236, "step": 13802 }, { "epoch": 0.48637648986495413, "grad_norm": 1.1698530912399292, "learning_rate": 1.0948684114920123e-05, "loss": 0.9195, "step": 13803 }, { "epoch": 0.48641172687791256, "grad_norm": 14.160053253173828, "learning_rate": 1.0947547979304176e-05, "loss": 8.1209, "step": 13804 }, { "epoch": 0.486446963890871, "grad_norm": 12.142027854919434, "learning_rate": 1.09464118313463e-05, "loss": 3.7462, "step": 13805 }, { "epoch": 0.4864822009038294, "grad_norm": 6.894593238830566, "learning_rate": 1.0945275671061282e-05, "loss": 2.6704, "step": 13806 }, { "epoch": 0.4865174379167878, "grad_norm": 9.304924964904785, "learning_rate": 1.0944139498463935e-05, "loss": 3.0394, "step": 13807 }, { "epoch": 0.4865526749297462, "grad_norm": 1.270267128944397, "learning_rate": 1.0943003313569047e-05, "loss": 0.7372, "step": 13808 }, { "epoch": 0.4865879119427046, "grad_norm": 1.6070058345794678, "learning_rate": 1.094186711639142e-05, "loss": 0.8144, "step": 13809 }, { "epoch": 0.48662314895566305, "grad_norm": 3.404757022857666, "learning_rate": 1.0940730906945855e-05, "loss": 0.8745, "step": 13810 }, { "epoch": 0.48665838596862143, "grad_norm": 4.589013576507568, "learning_rate": 1.0939594685247152e-05, "loss": 3.7098, "step": 13811 }, { "epoch": 0.48669362298157987, "grad_norm": 3.618513584136963, "learning_rate": 1.0938458451310104e-05, "loss": 3.2151, "step": 13812 }, { "epoch": 0.48672885999453824, "grad_norm": 1.4248716831207275, "learning_rate": 1.0937322205149521e-05, "loss": 0.7918, "step": 13813 }, { "epoch": 0.4867640970074967, "grad_norm": 3.7755651473999023, "learning_rate": 1.0936185946780192e-05, "loss": 3.4658, "step": 13814 }, { "epoch": 0.4867993340204551, "grad_norm": 3.2687063217163086, "learning_rate": 1.0935049676216924e-05, "loss": 0.8023, "step": 13815 }, { "epoch": 0.4868345710334135, "grad_norm": 3.692993640899658, "learning_rate": 1.0933913393474513e-05, "loss": 2.6174, "step": 13816 }, { "epoch": 0.4868698080463719, "grad_norm": 6.294998645782471, "learning_rate": 1.0932777098567762e-05, "loss": 4.8311, "step": 13817 }, { "epoch": 0.4869050450593303, "grad_norm": 1.0642048120498657, "learning_rate": 1.093164079151147e-05, "loss": 1.015, "step": 13818 }, { "epoch": 0.48694028207228873, "grad_norm": 8.838662147521973, "learning_rate": 1.093050447232044e-05, "loss": 4.9666, "step": 13819 }, { "epoch": 0.48697551908524717, "grad_norm": 7.659955978393555, "learning_rate": 1.092936814100947e-05, "loss": 5.6435, "step": 13820 }, { "epoch": 0.48701075609820554, "grad_norm": 4.567893028259277, "learning_rate": 1.092823179759336e-05, "loss": 2.4955, "step": 13821 }, { "epoch": 0.487045993111164, "grad_norm": 3.0759005546569824, "learning_rate": 1.0927095442086914e-05, "loss": 3.3374, "step": 13822 }, { "epoch": 0.48708123012412236, "grad_norm": 3.658252477645874, "learning_rate": 1.0925959074504931e-05, "loss": 2.5985, "step": 13823 }, { "epoch": 0.4871164671370808, "grad_norm": 1.1379321813583374, "learning_rate": 1.0924822694862216e-05, "loss": 0.9242, "step": 13824 }, { "epoch": 0.4871517041500392, "grad_norm": 9.948532104492188, "learning_rate": 1.0923686303173561e-05, "loss": 5.7164, "step": 13825 }, { "epoch": 0.4871869411629976, "grad_norm": 4.378251552581787, "learning_rate": 1.0922549899453783e-05, "loss": 4.9875, "step": 13826 }, { "epoch": 0.48722217817595603, "grad_norm": 8.947017669677734, "learning_rate": 1.0921413483717669e-05, "loss": 3.0264, "step": 13827 }, { "epoch": 0.4872574151889144, "grad_norm": 1.6594748497009277, "learning_rate": 1.0920277055980028e-05, "loss": 1.059, "step": 13828 }, { "epoch": 0.48729265220187284, "grad_norm": 3.4291698932647705, "learning_rate": 1.0919140616255663e-05, "loss": 2.5821, "step": 13829 }, { "epoch": 0.4873278892148313, "grad_norm": 1.6297104358673096, "learning_rate": 1.0918004164559373e-05, "loss": 0.9621, "step": 13830 }, { "epoch": 0.48736312622778966, "grad_norm": 1.45170259475708, "learning_rate": 1.0916867700905959e-05, "loss": 0.8227, "step": 13831 }, { "epoch": 0.4873983632407481, "grad_norm": 1.340090036392212, "learning_rate": 1.0915731225310233e-05, "loss": 0.8406, "step": 13832 }, { "epoch": 0.48743360025370647, "grad_norm": 5.872740268707275, "learning_rate": 1.0914594737786985e-05, "loss": 3.6941, "step": 13833 }, { "epoch": 0.4874688372666649, "grad_norm": 44.13584899902344, "learning_rate": 1.091345823835103e-05, "loss": 3.3056, "step": 13834 }, { "epoch": 0.48750407427962333, "grad_norm": 1.5260251760482788, "learning_rate": 1.0912321727017159e-05, "loss": 0.7402, "step": 13835 }, { "epoch": 0.4875393112925817, "grad_norm": 1.6322846412658691, "learning_rate": 1.0911185203800185e-05, "loss": 1.0366, "step": 13836 }, { "epoch": 0.48757454830554015, "grad_norm": 2.80218243598938, "learning_rate": 1.0910048668714906e-05, "loss": 3.069, "step": 13837 }, { "epoch": 0.4876097853184985, "grad_norm": 1.3651915788650513, "learning_rate": 1.0908912121776127e-05, "loss": 0.7561, "step": 13838 }, { "epoch": 0.48764502233145696, "grad_norm": 3.4724984169006348, "learning_rate": 1.0907775562998652e-05, "loss": 2.4447, "step": 13839 }, { "epoch": 0.4876802593444154, "grad_norm": 9.808378219604492, "learning_rate": 1.0906638992397285e-05, "loss": 5.383, "step": 13840 }, { "epoch": 0.48771549635737377, "grad_norm": 1.7525572776794434, "learning_rate": 1.090550240998683e-05, "loss": 1.0044, "step": 13841 }, { "epoch": 0.4877507333703322, "grad_norm": 4.439604759216309, "learning_rate": 1.090436581578209e-05, "loss": 3.145, "step": 13842 }, { "epoch": 0.48778597038329063, "grad_norm": 1.126556158065796, "learning_rate": 1.0903229209797874e-05, "loss": 0.8139, "step": 13843 }, { "epoch": 0.487821207396249, "grad_norm": 1.2860511541366577, "learning_rate": 1.0902092592048976e-05, "loss": 0.9414, "step": 13844 }, { "epoch": 0.48785644440920745, "grad_norm": 1.37253737449646, "learning_rate": 1.0900955962550213e-05, "loss": 0.8722, "step": 13845 }, { "epoch": 0.4878916814221658, "grad_norm": 1.9750304222106934, "learning_rate": 1.089981932131638e-05, "loss": 0.9196, "step": 13846 }, { "epoch": 0.48792691843512426, "grad_norm": 8.986907958984375, "learning_rate": 1.0898682668362286e-05, "loss": 3.2937, "step": 13847 }, { "epoch": 0.4879621554480827, "grad_norm": 3.6736905574798584, "learning_rate": 1.0897546003702736e-05, "loss": 2.886, "step": 13848 }, { "epoch": 0.48799739246104107, "grad_norm": 1.012104868888855, "learning_rate": 1.0896409327352534e-05, "loss": 1.022, "step": 13849 }, { "epoch": 0.4880326294739995, "grad_norm": 1.0988879203796387, "learning_rate": 1.0895272639326491e-05, "loss": 0.7226, "step": 13850 }, { "epoch": 0.4880678664869579, "grad_norm": 3.9298598766326904, "learning_rate": 1.0894135939639405e-05, "loss": 3.4213, "step": 13851 }, { "epoch": 0.4881031034999163, "grad_norm": 5.2119293212890625, "learning_rate": 1.0892999228306082e-05, "loss": 3.1338, "step": 13852 }, { "epoch": 0.48813834051287475, "grad_norm": 2.105522632598877, "learning_rate": 1.0891862505341334e-05, "loss": 0.8258, "step": 13853 }, { "epoch": 0.4881735775258331, "grad_norm": 2.549384117126465, "learning_rate": 1.0890725770759962e-05, "loss": 2.9993, "step": 13854 }, { "epoch": 0.48820881453879156, "grad_norm": 3.0985171794891357, "learning_rate": 1.0889589024576773e-05, "loss": 3.3222, "step": 13855 }, { "epoch": 0.48824405155174994, "grad_norm": 5.881160736083984, "learning_rate": 1.0888452266806572e-05, "loss": 2.7607, "step": 13856 }, { "epoch": 0.48827928856470837, "grad_norm": 1.6274646520614624, "learning_rate": 1.088731549746417e-05, "loss": 0.8601, "step": 13857 }, { "epoch": 0.4883145255776668, "grad_norm": 1.5096611976623535, "learning_rate": 1.0886178716564368e-05, "loss": 0.8606, "step": 13858 }, { "epoch": 0.4883497625906252, "grad_norm": 1.2895480394363403, "learning_rate": 1.0885041924121977e-05, "loss": 0.8719, "step": 13859 }, { "epoch": 0.4883849996035836, "grad_norm": 1.1625010967254639, "learning_rate": 1.0883905120151803e-05, "loss": 1.2209, "step": 13860 }, { "epoch": 0.488420236616542, "grad_norm": 8.963303565979004, "learning_rate": 1.088276830466865e-05, "loss": 4.9021, "step": 13861 }, { "epoch": 0.4884554736295004, "grad_norm": 3.068826675415039, "learning_rate": 1.0881631477687332e-05, "loss": 2.6119, "step": 13862 }, { "epoch": 0.48849071064245886, "grad_norm": 5.068811416625977, "learning_rate": 1.0880494639222645e-05, "loss": 3.1145, "step": 13863 }, { "epoch": 0.48852594765541724, "grad_norm": 1.2858113050460815, "learning_rate": 1.0879357789289409e-05, "loss": 0.7866, "step": 13864 }, { "epoch": 0.48856118466837567, "grad_norm": 5.027441024780273, "learning_rate": 1.087822092790242e-05, "loss": 2.6723, "step": 13865 }, { "epoch": 0.48859642168133405, "grad_norm": 4.4906721115112305, "learning_rate": 1.0877084055076497e-05, "loss": 3.4386, "step": 13866 }, { "epoch": 0.4886316586942925, "grad_norm": 1.3253200054168701, "learning_rate": 1.087594717082644e-05, "loss": 1.0527, "step": 13867 }, { "epoch": 0.4886668957072509, "grad_norm": 5.619740009307861, "learning_rate": 1.0874810275167058e-05, "loss": 3.7696, "step": 13868 }, { "epoch": 0.4887021327202093, "grad_norm": 7.091053009033203, "learning_rate": 1.0873673368113162e-05, "loss": 6.4215, "step": 13869 }, { "epoch": 0.4887373697331677, "grad_norm": 3.516510248184204, "learning_rate": 1.0872536449679563e-05, "loss": 2.9701, "step": 13870 }, { "epoch": 0.4887726067461261, "grad_norm": 6.943324565887451, "learning_rate": 1.0871399519881059e-05, "loss": 5.4165, "step": 13871 }, { "epoch": 0.48880784375908454, "grad_norm": 6.1687703132629395, "learning_rate": 1.087026257873247e-05, "loss": 4.9477, "step": 13872 }, { "epoch": 0.48884308077204297, "grad_norm": 1.0346134901046753, "learning_rate": 1.0869125626248598e-05, "loss": 0.8593, "step": 13873 }, { "epoch": 0.48887831778500135, "grad_norm": 4.9042181968688965, "learning_rate": 1.0867988662444255e-05, "loss": 3.0502, "step": 13874 }, { "epoch": 0.4889135547979598, "grad_norm": 5.633683204650879, "learning_rate": 1.0866851687334248e-05, "loss": 3.4483, "step": 13875 }, { "epoch": 0.4889487918109182, "grad_norm": 6.5775251388549805, "learning_rate": 1.0865714700933387e-05, "loss": 2.9214, "step": 13876 }, { "epoch": 0.4889840288238766, "grad_norm": 1.42656672000885, "learning_rate": 1.0864577703256482e-05, "loss": 0.8428, "step": 13877 }, { "epoch": 0.489019265836835, "grad_norm": 1.3196343183517456, "learning_rate": 1.0863440694318344e-05, "loss": 0.808, "step": 13878 }, { "epoch": 0.4890545028497934, "grad_norm": 3.6408159732818604, "learning_rate": 1.0862303674133781e-05, "loss": 2.9618, "step": 13879 }, { "epoch": 0.48908973986275184, "grad_norm": 1.9323363304138184, "learning_rate": 1.08611666427176e-05, "loss": 1.0519, "step": 13880 }, { "epoch": 0.48912497687571027, "grad_norm": 1.2222387790679932, "learning_rate": 1.0860029600084619e-05, "loss": 0.7863, "step": 13881 }, { "epoch": 0.48916021388866865, "grad_norm": 1.1164175271987915, "learning_rate": 1.0858892546249635e-05, "loss": 0.9313, "step": 13882 }, { "epoch": 0.4891954509016271, "grad_norm": 4.07636833190918, "learning_rate": 1.0857755481227474e-05, "loss": 3.0936, "step": 13883 }, { "epoch": 0.48923068791458546, "grad_norm": 9.220125198364258, "learning_rate": 1.0856618405032935e-05, "loss": 4.8314, "step": 13884 }, { "epoch": 0.4892659249275439, "grad_norm": 4.989102840423584, "learning_rate": 1.0855481317680831e-05, "loss": 2.5121, "step": 13885 }, { "epoch": 0.4893011619405023, "grad_norm": 12.765751838684082, "learning_rate": 1.0854344219185975e-05, "loss": 5.7534, "step": 13886 }, { "epoch": 0.4893363989534607, "grad_norm": 5.20992374420166, "learning_rate": 1.0853207109563177e-05, "loss": 3.114, "step": 13887 }, { "epoch": 0.48937163596641914, "grad_norm": 1.4613462686538696, "learning_rate": 1.0852069988827248e-05, "loss": 0.9881, "step": 13888 }, { "epoch": 0.4894068729793775, "grad_norm": 1.497129201889038, "learning_rate": 1.0850932856993e-05, "loss": 0.8127, "step": 13889 }, { "epoch": 0.48944210999233595, "grad_norm": 1.6454675197601318, "learning_rate": 1.084979571407524e-05, "loss": 0.9099, "step": 13890 }, { "epoch": 0.4894773470052944, "grad_norm": 2.200504779815674, "learning_rate": 1.0848658560088785e-05, "loss": 0.6798, "step": 13891 }, { "epoch": 0.48951258401825276, "grad_norm": 1.0327552556991577, "learning_rate": 1.0847521395048443e-05, "loss": 1.1678, "step": 13892 }, { "epoch": 0.4895478210312112, "grad_norm": 1.4889867305755615, "learning_rate": 1.0846384218969028e-05, "loss": 1.0998, "step": 13893 }, { "epoch": 0.4895830580441696, "grad_norm": 1.4799290895462036, "learning_rate": 1.0845247031865348e-05, "loss": 1.0929, "step": 13894 }, { "epoch": 0.489618295057128, "grad_norm": 24.91497802734375, "learning_rate": 1.084410983375222e-05, "loss": 7.1783, "step": 13895 }, { "epoch": 0.48965353207008644, "grad_norm": 1.0362964868545532, "learning_rate": 1.0842972624644455e-05, "loss": 1.1127, "step": 13896 }, { "epoch": 0.4896887690830448, "grad_norm": 9.978713035583496, "learning_rate": 1.0841835404556862e-05, "loss": 5.6649, "step": 13897 }, { "epoch": 0.48972400609600325, "grad_norm": 1.3720484972000122, "learning_rate": 1.0840698173504256e-05, "loss": 0.7924, "step": 13898 }, { "epoch": 0.48975924310896163, "grad_norm": 1.8349906206130981, "learning_rate": 1.0839560931501449e-05, "loss": 1.087, "step": 13899 }, { "epoch": 0.48979448012192006, "grad_norm": 1.294930338859558, "learning_rate": 1.0838423678563257e-05, "loss": 0.9444, "step": 13900 }, { "epoch": 0.4898297171348785, "grad_norm": 8.001587867736816, "learning_rate": 1.0837286414704487e-05, "loss": 4.8269, "step": 13901 }, { "epoch": 0.4898649541478369, "grad_norm": 4.346065044403076, "learning_rate": 1.0836149139939955e-05, "loss": 3.2398, "step": 13902 }, { "epoch": 0.4899001911607953, "grad_norm": 1.462874412536621, "learning_rate": 1.0835011854284471e-05, "loss": 1.0282, "step": 13903 }, { "epoch": 0.4899354281737537, "grad_norm": 3.2806668281555176, "learning_rate": 1.0833874557752858e-05, "loss": 2.5646, "step": 13904 }, { "epoch": 0.4899706651867121, "grad_norm": 6.565463542938232, "learning_rate": 1.0832737250359919e-05, "loss": 3.3192, "step": 13905 }, { "epoch": 0.49000590219967055, "grad_norm": 4.040198802947998, "learning_rate": 1.0831599932120472e-05, "loss": 4.4954, "step": 13906 }, { "epoch": 0.49004113921262893, "grad_norm": 1.5861587524414062, "learning_rate": 1.0830462603049327e-05, "loss": 0.959, "step": 13907 }, { "epoch": 0.49007637622558736, "grad_norm": 6.333445072174072, "learning_rate": 1.0829325263161305e-05, "loss": 5.1422, "step": 13908 }, { "epoch": 0.4901116132385458, "grad_norm": 5.16196870803833, "learning_rate": 1.0828187912471211e-05, "loss": 3.5147, "step": 13909 }, { "epoch": 0.4901468502515042, "grad_norm": 1.2282118797302246, "learning_rate": 1.082705055099387e-05, "loss": 0.97, "step": 13910 }, { "epoch": 0.4901820872644626, "grad_norm": 2.0821681022644043, "learning_rate": 1.0825913178744086e-05, "loss": 0.8936, "step": 13911 }, { "epoch": 0.490217324277421, "grad_norm": 7.307789325714111, "learning_rate": 1.082477579573668e-05, "loss": 3.436, "step": 13912 }, { "epoch": 0.4902525612903794, "grad_norm": 9.734658241271973, "learning_rate": 1.0823638401986461e-05, "loss": 5.5359, "step": 13913 }, { "epoch": 0.49028779830333785, "grad_norm": 1.143967628479004, "learning_rate": 1.0822500997508249e-05, "loss": 0.7771, "step": 13914 }, { "epoch": 0.49032303531629623, "grad_norm": 1.0690128803253174, "learning_rate": 1.0821363582316855e-05, "loss": 1.001, "step": 13915 }, { "epoch": 0.49035827232925466, "grad_norm": 4.760937213897705, "learning_rate": 1.0820226156427098e-05, "loss": 3.0669, "step": 13916 }, { "epoch": 0.49039350934221304, "grad_norm": 5.731876850128174, "learning_rate": 1.081908871985379e-05, "loss": 7.5271, "step": 13917 }, { "epoch": 0.4904287463551715, "grad_norm": 5.319046497344971, "learning_rate": 1.0817951272611747e-05, "loss": 2.9472, "step": 13918 }, { "epoch": 0.4904639833681299, "grad_norm": 8.312202453613281, "learning_rate": 1.0816813814715786e-05, "loss": 5.0552, "step": 13919 }, { "epoch": 0.4904992203810883, "grad_norm": 8.754931449890137, "learning_rate": 1.0815676346180717e-05, "loss": 3.2555, "step": 13920 }, { "epoch": 0.4905344573940467, "grad_norm": 3.5163962841033936, "learning_rate": 1.0814538867021365e-05, "loss": 2.6834, "step": 13921 }, { "epoch": 0.4905696944070051, "grad_norm": 3.699108123779297, "learning_rate": 1.0813401377252535e-05, "loss": 2.6428, "step": 13922 }, { "epoch": 0.49060493141996353, "grad_norm": 1.9283771514892578, "learning_rate": 1.0812263876889053e-05, "loss": 1.1418, "step": 13923 }, { "epoch": 0.49064016843292196, "grad_norm": 5.528260707855225, "learning_rate": 1.0811126365945728e-05, "loss": 3.419, "step": 13924 }, { "epoch": 0.49067540544588034, "grad_norm": 3.048222064971924, "learning_rate": 1.080998884443738e-05, "loss": 3.2212, "step": 13925 }, { "epoch": 0.4907106424588388, "grad_norm": 5.773005485534668, "learning_rate": 1.0808851312378822e-05, "loss": 3.3663, "step": 13926 }, { "epoch": 0.49074587947179715, "grad_norm": 1.9996305704116821, "learning_rate": 1.0807713769784878e-05, "loss": 0.8019, "step": 13927 }, { "epoch": 0.4907811164847556, "grad_norm": 1.206106185913086, "learning_rate": 1.0806576216670352e-05, "loss": 1.0248, "step": 13928 }, { "epoch": 0.490816353497714, "grad_norm": 4.608087062835693, "learning_rate": 1.0805438653050075e-05, "loss": 2.6779, "step": 13929 }, { "epoch": 0.4908515905106724, "grad_norm": 1.597949504852295, "learning_rate": 1.080430107893885e-05, "loss": 1.0223, "step": 13930 }, { "epoch": 0.49088682752363083, "grad_norm": 4.006103515625, "learning_rate": 1.0803163494351507e-05, "loss": 2.7368, "step": 13931 }, { "epoch": 0.4909220645365892, "grad_norm": 9.80892562866211, "learning_rate": 1.0802025899302855e-05, "loss": 2.8424, "step": 13932 }, { "epoch": 0.49095730154954764, "grad_norm": 1.489122748374939, "learning_rate": 1.0800888293807712e-05, "loss": 0.9323, "step": 13933 }, { "epoch": 0.4909925385625061, "grad_norm": 5.520280361175537, "learning_rate": 1.0799750677880897e-05, "loss": 2.764, "step": 13934 }, { "epoch": 0.49102777557546445, "grad_norm": 15.085233688354492, "learning_rate": 1.079861305153723e-05, "loss": 2.7944, "step": 13935 }, { "epoch": 0.4910630125884229, "grad_norm": 1.3278532028198242, "learning_rate": 1.0797475414791526e-05, "loss": 0.9106, "step": 13936 }, { "epoch": 0.49109824960138126, "grad_norm": 1.3686854839324951, "learning_rate": 1.0796337767658601e-05, "loss": 0.8631, "step": 13937 }, { "epoch": 0.4911334866143397, "grad_norm": 8.079498291015625, "learning_rate": 1.0795200110153278e-05, "loss": 2.1999, "step": 13938 }, { "epoch": 0.49116872362729813, "grad_norm": 6.110268592834473, "learning_rate": 1.079406244229037e-05, "loss": 5.119, "step": 13939 }, { "epoch": 0.4912039606402565, "grad_norm": 1.3378804922103882, "learning_rate": 1.0792924764084698e-05, "loss": 0.8951, "step": 13940 }, { "epoch": 0.49123919765321494, "grad_norm": 4.901769161224365, "learning_rate": 1.0791787075551082e-05, "loss": 3.7172, "step": 13941 }, { "epoch": 0.4912744346661733, "grad_norm": 4.905928134918213, "learning_rate": 1.0790649376704334e-05, "loss": 2.8937, "step": 13942 }, { "epoch": 0.49130967167913175, "grad_norm": 1.1493977308273315, "learning_rate": 1.078951166755928e-05, "loss": 0.9886, "step": 13943 }, { "epoch": 0.4913449086920902, "grad_norm": 3.1583495140075684, "learning_rate": 1.0788373948130735e-05, "loss": 2.2348, "step": 13944 }, { "epoch": 0.49138014570504857, "grad_norm": 9.611603736877441, "learning_rate": 1.078723621843352e-05, "loss": 2.7344, "step": 13945 }, { "epoch": 0.491415382718007, "grad_norm": 4.431898593902588, "learning_rate": 1.0786098478482454e-05, "loss": 3.1482, "step": 13946 }, { "epoch": 0.49145061973096543, "grad_norm": 1.9438530206680298, "learning_rate": 1.078496072829235e-05, "loss": 0.8724, "step": 13947 }, { "epoch": 0.4914858567439238, "grad_norm": 1.7004162073135376, "learning_rate": 1.0783822967878039e-05, "loss": 1.0466, "step": 13948 }, { "epoch": 0.49152109375688224, "grad_norm": 5.099184036254883, "learning_rate": 1.078268519725433e-05, "loss": 2.6245, "step": 13949 }, { "epoch": 0.4915563307698406, "grad_norm": 2.065565586090088, "learning_rate": 1.078154741643605e-05, "loss": 0.8209, "step": 13950 }, { "epoch": 0.49159156778279905, "grad_norm": 1.0828946828842163, "learning_rate": 1.0780409625438012e-05, "loss": 1.0192, "step": 13951 }, { "epoch": 0.4916268047957575, "grad_norm": 1.4759172201156616, "learning_rate": 1.077927182427504e-05, "loss": 1.018, "step": 13952 }, { "epoch": 0.49166204180871587, "grad_norm": 1.2445240020751953, "learning_rate": 1.0778134012961954e-05, "loss": 1.1927, "step": 13953 }, { "epoch": 0.4916972788216743, "grad_norm": 2.5992977619171143, "learning_rate": 1.0776996191513576e-05, "loss": 0.8572, "step": 13954 }, { "epoch": 0.4917325158346327, "grad_norm": 1.242523193359375, "learning_rate": 1.0775858359944718e-05, "loss": 1.0064, "step": 13955 }, { "epoch": 0.4917677528475911, "grad_norm": 1.457639455795288, "learning_rate": 1.0774720518270209e-05, "loss": 0.9183, "step": 13956 }, { "epoch": 0.49180298986054954, "grad_norm": 4.960226535797119, "learning_rate": 1.0773582666504865e-05, "loss": 2.8456, "step": 13957 }, { "epoch": 0.4918382268735079, "grad_norm": 4.351638317108154, "learning_rate": 1.0772444804663511e-05, "loss": 3.2063, "step": 13958 }, { "epoch": 0.49187346388646636, "grad_norm": 5.249059677124023, "learning_rate": 1.0771306932760963e-05, "loss": 3.1293, "step": 13959 }, { "epoch": 0.49190870089942473, "grad_norm": 4.908087730407715, "learning_rate": 1.0770169050812045e-05, "loss": 2.6237, "step": 13960 }, { "epoch": 0.49194393791238317, "grad_norm": 4.977560997009277, "learning_rate": 1.0769031158831576e-05, "loss": 2.622, "step": 13961 }, { "epoch": 0.4919791749253416, "grad_norm": 8.27295207977295, "learning_rate": 1.0767893256834379e-05, "loss": 4.869, "step": 13962 }, { "epoch": 0.4920144119383, "grad_norm": 6.117827415466309, "learning_rate": 1.0766755344835273e-05, "loss": 5.6428, "step": 13963 }, { "epoch": 0.4920496489512584, "grad_norm": 1.4953563213348389, "learning_rate": 1.0765617422849082e-05, "loss": 0.8691, "step": 13964 }, { "epoch": 0.4920848859642168, "grad_norm": 1.597455620765686, "learning_rate": 1.0764479490890629e-05, "loss": 1.0316, "step": 13965 }, { "epoch": 0.4921201229771752, "grad_norm": 4.810290336608887, "learning_rate": 1.0763341548974728e-05, "loss": 3.4676, "step": 13966 }, { "epoch": 0.49215535999013366, "grad_norm": 6.164805889129639, "learning_rate": 1.0762203597116212e-05, "loss": 5.25, "step": 13967 }, { "epoch": 0.49219059700309203, "grad_norm": 9.815990447998047, "learning_rate": 1.076106563532989e-05, "loss": 5.6846, "step": 13968 }, { "epoch": 0.49222583401605047, "grad_norm": 3.3661582469940186, "learning_rate": 1.0759927663630598e-05, "loss": 2.934, "step": 13969 }, { "epoch": 0.49226107102900885, "grad_norm": 1.2419058084487915, "learning_rate": 1.0758789682033148e-05, "loss": 0.8173, "step": 13970 }, { "epoch": 0.4922963080419673, "grad_norm": 1.587895393371582, "learning_rate": 1.0757651690552367e-05, "loss": 0.6836, "step": 13971 }, { "epoch": 0.4923315450549257, "grad_norm": 8.685460090637207, "learning_rate": 1.0756513689203074e-05, "loss": 4.959, "step": 13972 }, { "epoch": 0.4923667820678841, "grad_norm": 5.470230579376221, "learning_rate": 1.0755375678000098e-05, "loss": 2.6222, "step": 13973 }, { "epoch": 0.4924020190808425, "grad_norm": 2.1662769317626953, "learning_rate": 1.0754237656958253e-05, "loss": 0.8898, "step": 13974 }, { "epoch": 0.4924372560938009, "grad_norm": 11.01533031463623, "learning_rate": 1.075309962609237e-05, "loss": 2.8292, "step": 13975 }, { "epoch": 0.49247249310675933, "grad_norm": 4.72981071472168, "learning_rate": 1.0751961585417263e-05, "loss": 3.1941, "step": 13976 }, { "epoch": 0.49250773011971777, "grad_norm": 5.360821723937988, "learning_rate": 1.0750823534947768e-05, "loss": 3.0038, "step": 13977 }, { "epoch": 0.49254296713267615, "grad_norm": 1.1911191940307617, "learning_rate": 1.0749685474698697e-05, "loss": 1.0151, "step": 13978 }, { "epoch": 0.4925782041456346, "grad_norm": 1.147238850593567, "learning_rate": 1.0748547404684877e-05, "loss": 0.9882, "step": 13979 }, { "epoch": 0.492613441158593, "grad_norm": 1.11627197265625, "learning_rate": 1.0747409324921132e-05, "loss": 0.9863, "step": 13980 }, { "epoch": 0.4926486781715514, "grad_norm": 1.524556040763855, "learning_rate": 1.0746271235422286e-05, "loss": 0.8753, "step": 13981 }, { "epoch": 0.4926839151845098, "grad_norm": 7.656929969787598, "learning_rate": 1.074513313620316e-05, "loss": 3.4673, "step": 13982 }, { "epoch": 0.4927191521974682, "grad_norm": 12.538552284240723, "learning_rate": 1.0743995027278581e-05, "loss": 5.4199, "step": 13983 }, { "epoch": 0.49275438921042664, "grad_norm": 1.579764485359192, "learning_rate": 1.0742856908663374e-05, "loss": 0.7821, "step": 13984 }, { "epoch": 0.49278962622338507, "grad_norm": 1.480975866317749, "learning_rate": 1.0741718780372359e-05, "loss": 0.7777, "step": 13985 }, { "epoch": 0.49282486323634345, "grad_norm": 4.3761067390441895, "learning_rate": 1.0740580642420364e-05, "loss": 2.8814, "step": 13986 }, { "epoch": 0.4928601002493019, "grad_norm": 0.8154239654541016, "learning_rate": 1.073944249482221e-05, "loss": 0.9725, "step": 13987 }, { "epoch": 0.49289533726226026, "grad_norm": 4.027829170227051, "learning_rate": 1.0738304337592724e-05, "loss": 3.0439, "step": 13988 }, { "epoch": 0.4929305742752187, "grad_norm": 1.4905242919921875, "learning_rate": 1.0737166170746731e-05, "loss": 0.9846, "step": 13989 }, { "epoch": 0.4929658112881771, "grad_norm": 12.712788581848145, "learning_rate": 1.0736027994299053e-05, "loss": 5.1685, "step": 13990 }, { "epoch": 0.4930010483011355, "grad_norm": 1.6468324661254883, "learning_rate": 1.073488980826452e-05, "loss": 0.8811, "step": 13991 }, { "epoch": 0.49303628531409394, "grad_norm": 1.6284353733062744, "learning_rate": 1.0733751612657953e-05, "loss": 0.9809, "step": 13992 }, { "epoch": 0.4930715223270523, "grad_norm": 24.125675201416016, "learning_rate": 1.0732613407494173e-05, "loss": 7.594, "step": 13993 }, { "epoch": 0.49310675934001075, "grad_norm": 0.9679938554763794, "learning_rate": 1.0731475192788015e-05, "loss": 1.042, "step": 13994 }, { "epoch": 0.4931419963529692, "grad_norm": 2.6017863750457764, "learning_rate": 1.0730336968554294e-05, "loss": 0.9984, "step": 13995 }, { "epoch": 0.49317723336592756, "grad_norm": 1.2235169410705566, "learning_rate": 1.0729198734807846e-05, "loss": 0.7887, "step": 13996 }, { "epoch": 0.493212470378886, "grad_norm": 4.432601451873779, "learning_rate": 1.072806049156349e-05, "loss": 2.69, "step": 13997 }, { "epoch": 0.49324770739184437, "grad_norm": 3.2967398166656494, "learning_rate": 1.0726922238836053e-05, "loss": 3.2744, "step": 13998 }, { "epoch": 0.4932829444048028, "grad_norm": 1.374266505241394, "learning_rate": 1.072578397664036e-05, "loss": 0.8929, "step": 13999 }, { "epoch": 0.49331818141776124, "grad_norm": 7.233217239379883, "learning_rate": 1.0724645704991242e-05, "loss": 5.5035, "step": 14000 }, { "epoch": 0.4933534184307196, "grad_norm": 3.9547479152679443, "learning_rate": 1.072350742390352e-05, "loss": 2.6377, "step": 14001 }, { "epoch": 0.49338865544367805, "grad_norm": 7.037841320037842, "learning_rate": 1.0722369133392018e-05, "loss": 5.0823, "step": 14002 }, { "epoch": 0.4934238924566364, "grad_norm": 3.8342363834381104, "learning_rate": 1.0721230833471569e-05, "loss": 3.1047, "step": 14003 }, { "epoch": 0.49345912946959486, "grad_norm": 5.798559188842773, "learning_rate": 1.0720092524156995e-05, "loss": 2.5391, "step": 14004 }, { "epoch": 0.4934943664825533, "grad_norm": 5.7044358253479, "learning_rate": 1.0718954205463128e-05, "loss": 2.9379, "step": 14005 }, { "epoch": 0.49352960349551167, "grad_norm": 2.795048475265503, "learning_rate": 1.0717815877404788e-05, "loss": 0.8415, "step": 14006 }, { "epoch": 0.4935648405084701, "grad_norm": 5.3679704666137695, "learning_rate": 1.0716677539996806e-05, "loss": 2.9273, "step": 14007 }, { "epoch": 0.4936000775214285, "grad_norm": 2.947699785232544, "learning_rate": 1.0715539193254007e-05, "loss": 2.2776, "step": 14008 }, { "epoch": 0.4936353145343869, "grad_norm": 8.234570503234863, "learning_rate": 1.071440083719122e-05, "loss": 2.5577, "step": 14009 }, { "epoch": 0.49367055154734535, "grad_norm": 11.157559394836426, "learning_rate": 1.071326247182327e-05, "loss": 3.3729, "step": 14010 }, { "epoch": 0.4937057885603037, "grad_norm": 6.326112747192383, "learning_rate": 1.0712124097164987e-05, "loss": 5.5811, "step": 14011 }, { "epoch": 0.49374102557326216, "grad_norm": 1.1753355264663696, "learning_rate": 1.0710985713231191e-05, "loss": 0.9863, "step": 14012 }, { "epoch": 0.4937762625862206, "grad_norm": 1.0211365222930908, "learning_rate": 1.070984732003672e-05, "loss": 0.9849, "step": 14013 }, { "epoch": 0.49381149959917897, "grad_norm": 5.301026344299316, "learning_rate": 1.0708708917596397e-05, "loss": 3.2675, "step": 14014 }, { "epoch": 0.4938467366121374, "grad_norm": 0.8886890411376953, "learning_rate": 1.0707570505925051e-05, "loss": 0.9148, "step": 14015 }, { "epoch": 0.4938819736250958, "grad_norm": 0.977495551109314, "learning_rate": 1.0706432085037507e-05, "loss": 0.8532, "step": 14016 }, { "epoch": 0.4939172106380542, "grad_norm": 4.772822380065918, "learning_rate": 1.0705293654948599e-05, "loss": 2.3059, "step": 14017 }, { "epoch": 0.49395244765101265, "grad_norm": 4.6719160079956055, "learning_rate": 1.0704155215673147e-05, "loss": 3.0417, "step": 14018 }, { "epoch": 0.493987684663971, "grad_norm": 1.2399436235427856, "learning_rate": 1.0703016767225985e-05, "loss": 0.8157, "step": 14019 }, { "epoch": 0.49402292167692946, "grad_norm": 2.1584603786468506, "learning_rate": 1.070187830962194e-05, "loss": 0.8136, "step": 14020 }, { "epoch": 0.49405815868988784, "grad_norm": 2.0875308513641357, "learning_rate": 1.070073984287584e-05, "loss": 0.7289, "step": 14021 }, { "epoch": 0.49409339570284627, "grad_norm": 1.4909032583236694, "learning_rate": 1.0699601367002516e-05, "loss": 1.1722, "step": 14022 }, { "epoch": 0.4941286327158047, "grad_norm": 13.931599617004395, "learning_rate": 1.0698462882016794e-05, "loss": 3.4787, "step": 14023 }, { "epoch": 0.4941638697287631, "grad_norm": 1.2243119478225708, "learning_rate": 1.0697324387933505e-05, "loss": 0.8566, "step": 14024 }, { "epoch": 0.4941991067417215, "grad_norm": 1.269242763519287, "learning_rate": 1.0696185884767474e-05, "loss": 0.8019, "step": 14025 }, { "epoch": 0.4942343437546799, "grad_norm": 4.306800842285156, "learning_rate": 1.0695047372533537e-05, "loss": 2.1871, "step": 14026 }, { "epoch": 0.49426958076763833, "grad_norm": 6.981380939483643, "learning_rate": 1.0693908851246519e-05, "loss": 3.2798, "step": 14027 }, { "epoch": 0.49430481778059676, "grad_norm": 1.3826802968978882, "learning_rate": 1.069277032092125e-05, "loss": 0.7579, "step": 14028 }, { "epoch": 0.49434005479355514, "grad_norm": 1.3814526796340942, "learning_rate": 1.0691631781572556e-05, "loss": 0.8022, "step": 14029 }, { "epoch": 0.4943752918065136, "grad_norm": 2.8877358436584473, "learning_rate": 1.0690493233215272e-05, "loss": 2.8414, "step": 14030 }, { "epoch": 0.49441052881947195, "grad_norm": 3.418529987335205, "learning_rate": 1.0689354675864225e-05, "loss": 2.877, "step": 14031 }, { "epoch": 0.4944457658324304, "grad_norm": 1.0144027471542358, "learning_rate": 1.0688216109534248e-05, "loss": 1.1445, "step": 14032 }, { "epoch": 0.4944810028453888, "grad_norm": 5.509897232055664, "learning_rate": 1.0687077534240165e-05, "loss": 5.776, "step": 14033 }, { "epoch": 0.4945162398583472, "grad_norm": 3.1049060821533203, "learning_rate": 1.0685938949996813e-05, "loss": 1.017, "step": 14034 }, { "epoch": 0.49455147687130563, "grad_norm": 2.0182175636291504, "learning_rate": 1.0684800356819015e-05, "loss": 1.1051, "step": 14035 }, { "epoch": 0.494586713884264, "grad_norm": 4.070767879486084, "learning_rate": 1.0683661754721608e-05, "loss": 3.2844, "step": 14036 }, { "epoch": 0.49462195089722244, "grad_norm": 1.117130994796753, "learning_rate": 1.0682523143719419e-05, "loss": 1.2652, "step": 14037 }, { "epoch": 0.4946571879101809, "grad_norm": 1.6323513984680176, "learning_rate": 1.0681384523827279e-05, "loss": 1.0874, "step": 14038 }, { "epoch": 0.49469242492313925, "grad_norm": 1.7589383125305176, "learning_rate": 1.0680245895060018e-05, "loss": 0.919, "step": 14039 }, { "epoch": 0.4947276619360977, "grad_norm": 1.276233196258545, "learning_rate": 1.0679107257432469e-05, "loss": 0.8902, "step": 14040 }, { "epoch": 0.49476289894905606, "grad_norm": 7.159460067749023, "learning_rate": 1.067796861095946e-05, "loss": 5.6207, "step": 14041 }, { "epoch": 0.4947981359620145, "grad_norm": 0.9821751117706299, "learning_rate": 1.0676829955655822e-05, "loss": 0.9496, "step": 14042 }, { "epoch": 0.49483337297497293, "grad_norm": 7.227525234222412, "learning_rate": 1.0675691291536391e-05, "loss": 5.7568, "step": 14043 }, { "epoch": 0.4948686099879313, "grad_norm": 2.3752996921539307, "learning_rate": 1.0674552618615995e-05, "loss": 0.8686, "step": 14044 }, { "epoch": 0.49490384700088974, "grad_norm": 1.9601792097091675, "learning_rate": 1.0673413936909464e-05, "loss": 1.0251, "step": 14045 }, { "epoch": 0.4949390840138481, "grad_norm": 1.2310259342193604, "learning_rate": 1.0672275246431632e-05, "loss": 0.8789, "step": 14046 }, { "epoch": 0.49497432102680655, "grad_norm": 5.972388744354248, "learning_rate": 1.0671136547197326e-05, "loss": 5.1277, "step": 14047 }, { "epoch": 0.495009558039765, "grad_norm": 4.844724178314209, "learning_rate": 1.0669997839221382e-05, "loss": 3.4487, "step": 14048 }, { "epoch": 0.49504479505272336, "grad_norm": 5.885207653045654, "learning_rate": 1.0668859122518632e-05, "loss": 5.7262, "step": 14049 }, { "epoch": 0.4950800320656818, "grad_norm": 1.1958465576171875, "learning_rate": 1.0667720397103907e-05, "loss": 1.1727, "step": 14050 }, { "epoch": 0.49511526907864023, "grad_norm": 1.6890244483947754, "learning_rate": 1.066658166299204e-05, "loss": 0.8228, "step": 14051 }, { "epoch": 0.4951505060915986, "grad_norm": 1.270698070526123, "learning_rate": 1.0665442920197857e-05, "loss": 0.9896, "step": 14052 }, { "epoch": 0.49518574310455704, "grad_norm": 3.2730934619903564, "learning_rate": 1.06643041687362e-05, "loss": 2.7964, "step": 14053 }, { "epoch": 0.4952209801175154, "grad_norm": 2.8902242183685303, "learning_rate": 1.0663165408621895e-05, "loss": 2.8131, "step": 14054 }, { "epoch": 0.49525621713047385, "grad_norm": 0.9885086417198181, "learning_rate": 1.0662026639869776e-05, "loss": 0.7333, "step": 14055 }, { "epoch": 0.4952914541434323, "grad_norm": 8.33594799041748, "learning_rate": 1.0660887862494676e-05, "loss": 4.5725, "step": 14056 }, { "epoch": 0.49532669115639066, "grad_norm": 0.9952625632286072, "learning_rate": 1.0659749076511425e-05, "loss": 1.3294, "step": 14057 }, { "epoch": 0.4953619281693491, "grad_norm": 19.91439437866211, "learning_rate": 1.065861028193486e-05, "loss": 3.7682, "step": 14058 }, { "epoch": 0.4953971651823075, "grad_norm": 2.7664895057678223, "learning_rate": 1.0657471478779813e-05, "loss": 3.2007, "step": 14059 }, { "epoch": 0.4954324021952659, "grad_norm": 2.3075079917907715, "learning_rate": 1.0656332667061115e-05, "loss": 1.0052, "step": 14060 }, { "epoch": 0.49546763920822434, "grad_norm": 4.274442195892334, "learning_rate": 1.0655193846793601e-05, "loss": 2.9618, "step": 14061 }, { "epoch": 0.4955028762211827, "grad_norm": 6.360972881317139, "learning_rate": 1.0654055017992107e-05, "loss": 3.3271, "step": 14062 }, { "epoch": 0.49553811323414115, "grad_norm": 1.3115484714508057, "learning_rate": 1.0652916180671455e-05, "loss": 0.9799, "step": 14063 }, { "epoch": 0.49557335024709953, "grad_norm": 2.2548677921295166, "learning_rate": 1.0651777334846494e-05, "loss": 1.0289, "step": 14064 }, { "epoch": 0.49560858726005796, "grad_norm": 1.26166570186615, "learning_rate": 1.0650638480532047e-05, "loss": 0.9643, "step": 14065 }, { "epoch": 0.4956438242730164, "grad_norm": 6.710322856903076, "learning_rate": 1.0649499617742954e-05, "loss": 3.4258, "step": 14066 }, { "epoch": 0.4956790612859748, "grad_norm": 1.8393137454986572, "learning_rate": 1.0648360746494041e-05, "loss": 0.8107, "step": 14067 }, { "epoch": 0.4957142982989332, "grad_norm": 14.693465232849121, "learning_rate": 1.0647221866800148e-05, "loss": 3.4151, "step": 14068 }, { "epoch": 0.4957495353118916, "grad_norm": 1.339196801185608, "learning_rate": 1.064608297867611e-05, "loss": 1.0002, "step": 14069 }, { "epoch": 0.49578477232485, "grad_norm": 11.06746768951416, "learning_rate": 1.0644944082136759e-05, "loss": 3.3126, "step": 14070 }, { "epoch": 0.49582000933780845, "grad_norm": 1.1172077655792236, "learning_rate": 1.0643805177196924e-05, "loss": 1.1223, "step": 14071 }, { "epoch": 0.49585524635076683, "grad_norm": 6.288723468780518, "learning_rate": 1.064266626387145e-05, "loss": 3.5035, "step": 14072 }, { "epoch": 0.49589048336372527, "grad_norm": 1.4162013530731201, "learning_rate": 1.0641527342175166e-05, "loss": 0.9663, "step": 14073 }, { "epoch": 0.49592572037668364, "grad_norm": 6.0414652824401855, "learning_rate": 1.0640388412122904e-05, "loss": 3.1832, "step": 14074 }, { "epoch": 0.4959609573896421, "grad_norm": 13.571515083312988, "learning_rate": 1.0639249473729502e-05, "loss": 3.6584, "step": 14075 }, { "epoch": 0.4959961944026005, "grad_norm": 0.8573833107948303, "learning_rate": 1.0638110527009795e-05, "loss": 0.9229, "step": 14076 }, { "epoch": 0.4960314314155589, "grad_norm": 1.6081781387329102, "learning_rate": 1.0636971571978616e-05, "loss": 1.276, "step": 14077 }, { "epoch": 0.4960666684285173, "grad_norm": 3.462862968444824, "learning_rate": 1.0635832608650803e-05, "loss": 2.9078, "step": 14078 }, { "epoch": 0.4961019054414757, "grad_norm": 1.6273866891860962, "learning_rate": 1.0634693637041187e-05, "loss": 1.0042, "step": 14079 }, { "epoch": 0.49613714245443413, "grad_norm": 5.862022399902344, "learning_rate": 1.0633554657164607e-05, "loss": 4.9459, "step": 14080 }, { "epoch": 0.49617237946739257, "grad_norm": 4.596334934234619, "learning_rate": 1.0632415669035898e-05, "loss": 3.6748, "step": 14081 }, { "epoch": 0.49620761648035094, "grad_norm": 1.274043083190918, "learning_rate": 1.063127667266989e-05, "loss": 1.0992, "step": 14082 }, { "epoch": 0.4962428534933094, "grad_norm": 4.581240177154541, "learning_rate": 1.0630137668081428e-05, "loss": 3.2902, "step": 14083 }, { "epoch": 0.4962780905062678, "grad_norm": 0.9887449145317078, "learning_rate": 1.0628998655285341e-05, "loss": 0.9372, "step": 14084 }, { "epoch": 0.4963133275192262, "grad_norm": 11.002429962158203, "learning_rate": 1.0627859634296465e-05, "loss": 5.6297, "step": 14085 }, { "epoch": 0.4963485645321846, "grad_norm": 1.1393828392028809, "learning_rate": 1.0626720605129637e-05, "loss": 0.9396, "step": 14086 }, { "epoch": 0.496383801545143, "grad_norm": 5.732470989227295, "learning_rate": 1.0625581567799695e-05, "loss": 3.2667, "step": 14087 }, { "epoch": 0.49641903855810143, "grad_norm": 3.665313482284546, "learning_rate": 1.0624442522321472e-05, "loss": 3.1263, "step": 14088 }, { "epoch": 0.49645427557105987, "grad_norm": 6.478874206542969, "learning_rate": 1.0623303468709808e-05, "loss": 5.429, "step": 14089 }, { "epoch": 0.49648951258401824, "grad_norm": 2.6907668113708496, "learning_rate": 1.0622164406979534e-05, "loss": 3.0801, "step": 14090 }, { "epoch": 0.4965247495969767, "grad_norm": 2.806990146636963, "learning_rate": 1.0621025337145492e-05, "loss": 1.0412, "step": 14091 }, { "epoch": 0.49655998660993506, "grad_norm": 4.579677581787109, "learning_rate": 1.0619886259222514e-05, "loss": 3.7213, "step": 14092 }, { "epoch": 0.4965952236228935, "grad_norm": 1.8902788162231445, "learning_rate": 1.0618747173225438e-05, "loss": 0.8296, "step": 14093 }, { "epoch": 0.4966304606358519, "grad_norm": 1.4989898204803467, "learning_rate": 1.0617608079169103e-05, "loss": 0.8623, "step": 14094 }, { "epoch": 0.4966656976488103, "grad_norm": 1.3321701288223267, "learning_rate": 1.0616468977068344e-05, "loss": 0.7943, "step": 14095 }, { "epoch": 0.49670093466176873, "grad_norm": 1.1861094236373901, "learning_rate": 1.0615329866937995e-05, "loss": 0.7635, "step": 14096 }, { "epoch": 0.4967361716747271, "grad_norm": 6.524302005767822, "learning_rate": 1.06141907487929e-05, "loss": 6.8217, "step": 14097 }, { "epoch": 0.49677140868768555, "grad_norm": 14.429781913757324, "learning_rate": 1.061305162264789e-05, "loss": 5.9216, "step": 14098 }, { "epoch": 0.496806645700644, "grad_norm": 7.868647575378418, "learning_rate": 1.0611912488517808e-05, "loss": 3.5613, "step": 14099 }, { "epoch": 0.49684188271360236, "grad_norm": 6.157867908477783, "learning_rate": 1.0610773346417485e-05, "loss": 5.6704, "step": 14100 }, { "epoch": 0.4968771197265608, "grad_norm": 16.547016143798828, "learning_rate": 1.0609634196361762e-05, "loss": 3.2396, "step": 14101 }, { "epoch": 0.49691235673951917, "grad_norm": 11.409307479858398, "learning_rate": 1.0608495038365475e-05, "loss": 7.5969, "step": 14102 }, { "epoch": 0.4969475937524776, "grad_norm": 5.688292980194092, "learning_rate": 1.0607355872443463e-05, "loss": 3.2804, "step": 14103 }, { "epoch": 0.49698283076543603, "grad_norm": 1.674904465675354, "learning_rate": 1.0606216698610563e-05, "loss": 0.7466, "step": 14104 }, { "epoch": 0.4970180677783944, "grad_norm": 6.678242206573486, "learning_rate": 1.0605077516881612e-05, "loss": 5.9408, "step": 14105 }, { "epoch": 0.49705330479135285, "grad_norm": 8.20940113067627, "learning_rate": 1.0603938327271453e-05, "loss": 4.791, "step": 14106 }, { "epoch": 0.4970885418043112, "grad_norm": 10.560194969177246, "learning_rate": 1.0602799129794916e-05, "loss": 2.9673, "step": 14107 }, { "epoch": 0.49712377881726966, "grad_norm": 1.9171053171157837, "learning_rate": 1.060165992446685e-05, "loss": 1.2039, "step": 14108 }, { "epoch": 0.4971590158302281, "grad_norm": 6.7746262550354, "learning_rate": 1.060052071130208e-05, "loss": 5.5952, "step": 14109 }, { "epoch": 0.49719425284318647, "grad_norm": 1.0327409505844116, "learning_rate": 1.0599381490315454e-05, "loss": 0.7675, "step": 14110 }, { "epoch": 0.4972294898561449, "grad_norm": 1.7507915496826172, "learning_rate": 1.059824226152181e-05, "loss": 0.7518, "step": 14111 }, { "epoch": 0.4972647268691033, "grad_norm": 3.3773138523101807, "learning_rate": 1.0597103024935981e-05, "loss": 2.954, "step": 14112 }, { "epoch": 0.4972999638820617, "grad_norm": 5.5601019859313965, "learning_rate": 1.0595963780572811e-05, "loss": 5.7262, "step": 14113 }, { "epoch": 0.49733520089502015, "grad_norm": 1.4980294704437256, "learning_rate": 1.0594824528447133e-05, "loss": 1.0937, "step": 14114 }, { "epoch": 0.4973704379079785, "grad_norm": 5.8100996017456055, "learning_rate": 1.0593685268573794e-05, "loss": 3.5135, "step": 14115 }, { "epoch": 0.49740567492093696, "grad_norm": 1.254833459854126, "learning_rate": 1.059254600096763e-05, "loss": 1.0141, "step": 14116 }, { "epoch": 0.4974409119338954, "grad_norm": 4.378107070922852, "learning_rate": 1.0591406725643474e-05, "loss": 4.1815, "step": 14117 }, { "epoch": 0.49747614894685377, "grad_norm": 7.250990390777588, "learning_rate": 1.0590267442616173e-05, "loss": 4.9581, "step": 14118 }, { "epoch": 0.4975113859598122, "grad_norm": 6.129569053649902, "learning_rate": 1.0589128151900563e-05, "loss": 3.3784, "step": 14119 }, { "epoch": 0.4975466229727706, "grad_norm": 5.00813627243042, "learning_rate": 1.0587988853511484e-05, "loss": 3.001, "step": 14120 }, { "epoch": 0.497581859985729, "grad_norm": 0.9972643852233887, "learning_rate": 1.0586849547463776e-05, "loss": 0.9589, "step": 14121 }, { "epoch": 0.49761709699868745, "grad_norm": 1.1348302364349365, "learning_rate": 1.0585710233772276e-05, "loss": 0.9644, "step": 14122 }, { "epoch": 0.4976523340116458, "grad_norm": 4.442795276641846, "learning_rate": 1.0584570912451826e-05, "loss": 3.2754, "step": 14123 }, { "epoch": 0.49768757102460426, "grad_norm": 1.4138840436935425, "learning_rate": 1.0583431583517264e-05, "loss": 1.0219, "step": 14124 }, { "epoch": 0.49772280803756264, "grad_norm": 1.2070534229278564, "learning_rate": 1.0582292246983434e-05, "loss": 0.8767, "step": 14125 }, { "epoch": 0.49775804505052107, "grad_norm": 10.380183219909668, "learning_rate": 1.0581152902865173e-05, "loss": 5.3253, "step": 14126 }, { "epoch": 0.4977932820634795, "grad_norm": 1.3503130674362183, "learning_rate": 1.0580013551177321e-05, "loss": 1.2297, "step": 14127 }, { "epoch": 0.4978285190764379, "grad_norm": 1.1895664930343628, "learning_rate": 1.0578874191934717e-05, "loss": 0.8963, "step": 14128 }, { "epoch": 0.4978637560893963, "grad_norm": 1.2451450824737549, "learning_rate": 1.0577734825152206e-05, "loss": 0.918, "step": 14129 }, { "epoch": 0.4978989931023547, "grad_norm": 1.8841559886932373, "learning_rate": 1.0576595450844622e-05, "loss": 0.899, "step": 14130 }, { "epoch": 0.4979342301153131, "grad_norm": 1.600394368171692, "learning_rate": 1.0575456069026812e-05, "loss": 0.8695, "step": 14131 }, { "epoch": 0.49796946712827156, "grad_norm": 3.1379218101501465, "learning_rate": 1.0574316679713611e-05, "loss": 2.3482, "step": 14132 }, { "epoch": 0.49800470414122994, "grad_norm": 16.251758575439453, "learning_rate": 1.0573177282919864e-05, "loss": 2.9536, "step": 14133 }, { "epoch": 0.49803994115418837, "grad_norm": 5.177811145782471, "learning_rate": 1.0572037878660408e-05, "loss": 3.175, "step": 14134 }, { "epoch": 0.49807517816714675, "grad_norm": 9.508828163146973, "learning_rate": 1.0570898466950087e-05, "loss": 9.6705, "step": 14135 }, { "epoch": 0.4981104151801052, "grad_norm": 1.4240087270736694, "learning_rate": 1.0569759047803739e-05, "loss": 0.9334, "step": 14136 }, { "epoch": 0.4981456521930636, "grad_norm": 5.562371253967285, "learning_rate": 1.0568619621236211e-05, "loss": 3.3394, "step": 14137 }, { "epoch": 0.498180889206022, "grad_norm": 7.401174068450928, "learning_rate": 1.0567480187262336e-05, "loss": 3.1361, "step": 14138 }, { "epoch": 0.4982161262189804, "grad_norm": 2.003917932510376, "learning_rate": 1.0566340745896961e-05, "loss": 1.0381, "step": 14139 }, { "epoch": 0.4982513632319388, "grad_norm": 1.7075111865997314, "learning_rate": 1.0565201297154926e-05, "loss": 0.7718, "step": 14140 }, { "epoch": 0.49828660024489724, "grad_norm": 6.127976894378662, "learning_rate": 1.056406184105107e-05, "loss": 2.8923, "step": 14141 }, { "epoch": 0.49832183725785567, "grad_norm": 4.379056453704834, "learning_rate": 1.0562922377600237e-05, "loss": 3.4554, "step": 14142 }, { "epoch": 0.49835707427081405, "grad_norm": 5.719613552093506, "learning_rate": 1.0561782906817268e-05, "loss": 3.4462, "step": 14143 }, { "epoch": 0.4983923112837725, "grad_norm": 1.0389697551727295, "learning_rate": 1.0560643428717008e-05, "loss": 0.9908, "step": 14144 }, { "epoch": 0.49842754829673086, "grad_norm": 1.8299877643585205, "learning_rate": 1.0559503943314294e-05, "loss": 1.3418, "step": 14145 }, { "epoch": 0.4984627853096893, "grad_norm": 1.4925105571746826, "learning_rate": 1.055836445062397e-05, "loss": 0.8881, "step": 14146 }, { "epoch": 0.4984980223226477, "grad_norm": 1.5765066146850586, "learning_rate": 1.0557224950660876e-05, "loss": 1.2213, "step": 14147 }, { "epoch": 0.4985332593356061, "grad_norm": 4.085042953491211, "learning_rate": 1.055608544343986e-05, "loss": 2.614, "step": 14148 }, { "epoch": 0.49856849634856454, "grad_norm": 1.0038552284240723, "learning_rate": 1.0554945928975757e-05, "loss": 1.0521, "step": 14149 }, { "epoch": 0.4986037333615229, "grad_norm": 2.119680166244507, "learning_rate": 1.0553806407283416e-05, "loss": 0.8219, "step": 14150 }, { "epoch": 0.49863897037448135, "grad_norm": 2.0326199531555176, "learning_rate": 1.0552666878377673e-05, "loss": 0.9594, "step": 14151 }, { "epoch": 0.4986742073874398, "grad_norm": 14.738727569580078, "learning_rate": 1.0551527342273374e-05, "loss": 3.1186, "step": 14152 }, { "epoch": 0.49870944440039816, "grad_norm": 1.008424162864685, "learning_rate": 1.0550387798985362e-05, "loss": 0.8379, "step": 14153 }, { "epoch": 0.4987446814133566, "grad_norm": 3.27825927734375, "learning_rate": 1.054924824852848e-05, "loss": 2.4046, "step": 14154 }, { "epoch": 0.498779918426315, "grad_norm": 1.9140982627868652, "learning_rate": 1.0548108690917565e-05, "loss": 0.8759, "step": 14155 }, { "epoch": 0.4988151554392734, "grad_norm": 11.384137153625488, "learning_rate": 1.054696912616747e-05, "loss": 3.1261, "step": 14156 }, { "epoch": 0.49885039245223184, "grad_norm": 8.516485214233398, "learning_rate": 1.054582955429303e-05, "loss": 4.6213, "step": 14157 }, { "epoch": 0.4988856294651902, "grad_norm": 3.0314435958862305, "learning_rate": 1.054468997530909e-05, "loss": 0.9843, "step": 14158 }, { "epoch": 0.49892086647814865, "grad_norm": 2.5545284748077393, "learning_rate": 1.0543550389230493e-05, "loss": 1.1037, "step": 14159 }, { "epoch": 0.4989561034911071, "grad_norm": 6.674264430999756, "learning_rate": 1.0542410796072084e-05, "loss": 3.2256, "step": 14160 }, { "epoch": 0.49899134050406546, "grad_norm": 1.430814504623413, "learning_rate": 1.0541271195848706e-05, "loss": 1.0899, "step": 14161 }, { "epoch": 0.4990265775170239, "grad_norm": 1.585970401763916, "learning_rate": 1.05401315885752e-05, "loss": 1.1886, "step": 14162 }, { "epoch": 0.4990618145299823, "grad_norm": 6.089288234710693, "learning_rate": 1.0538991974266412e-05, "loss": 3.1201, "step": 14163 }, { "epoch": 0.4990970515429407, "grad_norm": 20.019563674926758, "learning_rate": 1.0537852352937186e-05, "loss": 5.4691, "step": 14164 }, { "epoch": 0.49913228855589914, "grad_norm": 4.460477352142334, "learning_rate": 1.0536712724602364e-05, "loss": 3.066, "step": 14165 }, { "epoch": 0.4991675255688575, "grad_norm": 1.112931251525879, "learning_rate": 1.0535573089276787e-05, "loss": 0.7987, "step": 14166 }, { "epoch": 0.49920276258181595, "grad_norm": 1.189918875694275, "learning_rate": 1.0534433446975307e-05, "loss": 1.0838, "step": 14167 }, { "epoch": 0.49923799959477433, "grad_norm": 1.1273667812347412, "learning_rate": 1.053329379771276e-05, "loss": 1.0272, "step": 14168 }, { "epoch": 0.49927323660773276, "grad_norm": 6.963191986083984, "learning_rate": 1.0532154141503999e-05, "loss": 3.5628, "step": 14169 }, { "epoch": 0.4993084736206912, "grad_norm": 6.232819080352783, "learning_rate": 1.0531014478363857e-05, "loss": 3.1715, "step": 14170 }, { "epoch": 0.4993437106336496, "grad_norm": 1.8849974870681763, "learning_rate": 1.0529874808307185e-05, "loss": 0.9269, "step": 14171 }, { "epoch": 0.499378947646608, "grad_norm": 6.564675331115723, "learning_rate": 1.0528735131348827e-05, "loss": 2.6952, "step": 14172 }, { "epoch": 0.4994141846595664, "grad_norm": 1.0724914073944092, "learning_rate": 1.0527595447503627e-05, "loss": 1.0356, "step": 14173 }, { "epoch": 0.4994494216725248, "grad_norm": 1.255311131477356, "learning_rate": 1.0526455756786427e-05, "loss": 1.0018, "step": 14174 }, { "epoch": 0.49948465868548325, "grad_norm": 1.4277384281158447, "learning_rate": 1.0525316059212076e-05, "loss": 0.9454, "step": 14175 }, { "epoch": 0.49951989569844163, "grad_norm": 1.791223406791687, "learning_rate": 1.0524176354795413e-05, "loss": 1.0713, "step": 14176 }, { "epoch": 0.49955513271140006, "grad_norm": 1.3408337831497192, "learning_rate": 1.052303664355129e-05, "loss": 1.2081, "step": 14177 }, { "epoch": 0.49959036972435844, "grad_norm": 2.1645596027374268, "learning_rate": 1.0521896925494545e-05, "loss": 0.75, "step": 14178 }, { "epoch": 0.4996256067373169, "grad_norm": 1.401382327079773, "learning_rate": 1.0520757200640028e-05, "loss": 1.1313, "step": 14179 }, { "epoch": 0.4996608437502753, "grad_norm": 4.755080223083496, "learning_rate": 1.0519617469002583e-05, "loss": 3.3211, "step": 14180 }, { "epoch": 0.4996960807632337, "grad_norm": 5.514292240142822, "learning_rate": 1.0518477730597052e-05, "loss": 3.5798, "step": 14181 }, { "epoch": 0.4997313177761921, "grad_norm": 14.373275756835938, "learning_rate": 1.0517337985438283e-05, "loss": 2.827, "step": 14182 }, { "epoch": 0.4997665547891505, "grad_norm": 5.9138407707214355, "learning_rate": 1.0516198233541121e-05, "loss": 3.5596, "step": 14183 }, { "epoch": 0.49980179180210893, "grad_norm": 22.05267333984375, "learning_rate": 1.0515058474920415e-05, "loss": 7.6009, "step": 14184 }, { "epoch": 0.49983702881506736, "grad_norm": 6.142730712890625, "learning_rate": 1.0513918709591e-05, "loss": 5.4514, "step": 14185 }, { "epoch": 0.49987226582802574, "grad_norm": 2.3510138988494873, "learning_rate": 1.0512778937567734e-05, "loss": 0.9118, "step": 14186 }, { "epoch": 0.4999075028409842, "grad_norm": 11.455521583557129, "learning_rate": 1.0511639158865452e-05, "loss": 3.7844, "step": 14187 }, { "epoch": 0.4999427398539426, "grad_norm": 1.2891713380813599, "learning_rate": 1.051049937349901e-05, "loss": 0.7863, "step": 14188 }, { "epoch": 0.499977976866901, "grad_norm": 5.605447769165039, "learning_rate": 1.0509359581483247e-05, "loss": 5.245, "step": 14189 }, { "epoch": 0.5000132138798594, "grad_norm": 1.377671480178833, "learning_rate": 1.0508219782833009e-05, "loss": 0.7297, "step": 14190 }, { "epoch": 0.5000484508928178, "grad_norm": 1.1426002979278564, "learning_rate": 1.0507079977563144e-05, "loss": 0.9255, "step": 14191 }, { "epoch": 0.5000836879057762, "grad_norm": 1.3377708196640015, "learning_rate": 1.05059401656885e-05, "loss": 1.0869, "step": 14192 }, { "epoch": 0.5001189249187347, "grad_norm": 1.2849230766296387, "learning_rate": 1.0504800347223916e-05, "loss": 0.9886, "step": 14193 }, { "epoch": 0.500154161931693, "grad_norm": 1.3672579526901245, "learning_rate": 1.050366052218425e-05, "loss": 1.4299, "step": 14194 }, { "epoch": 0.5001893989446514, "grad_norm": 16.21377182006836, "learning_rate": 1.0502520690584335e-05, "loss": 4.537, "step": 14195 }, { "epoch": 0.5002246359576099, "grad_norm": 1.1348017454147339, "learning_rate": 1.0501380852439028e-05, "loss": 0.9982, "step": 14196 }, { "epoch": 0.5002598729705683, "grad_norm": 1.1165281534194946, "learning_rate": 1.050024100776317e-05, "loss": 0.8639, "step": 14197 }, { "epoch": 0.5002951099835267, "grad_norm": 3.8920845985412598, "learning_rate": 1.049910115657161e-05, "loss": 3.3606, "step": 14198 }, { "epoch": 0.5003303469964852, "grad_norm": 5.799983978271484, "learning_rate": 1.049796129887919e-05, "loss": 3.2008, "step": 14199 }, { "epoch": 0.5003655840094435, "grad_norm": 1.0698778629302979, "learning_rate": 1.0496821434700766e-05, "loss": 0.8358, "step": 14200 }, { "epoch": 0.5004008210224019, "grad_norm": 1.5138955116271973, "learning_rate": 1.0495681564051178e-05, "loss": 0.7223, "step": 14201 }, { "epoch": 0.5004360580353603, "grad_norm": 4.623783111572266, "learning_rate": 1.0494541686945272e-05, "loss": 3.1157, "step": 14202 }, { "epoch": 0.5004712950483188, "grad_norm": 7.197026252746582, "learning_rate": 1.0493401803397897e-05, "loss": 5.5555, "step": 14203 }, { "epoch": 0.5005065320612772, "grad_norm": 4.9982523918151855, "learning_rate": 1.0492261913423904e-05, "loss": 3.1937, "step": 14204 }, { "epoch": 0.5005417690742355, "grad_norm": 1.7241371870040894, "learning_rate": 1.0491122017038138e-05, "loss": 0.8733, "step": 14205 }, { "epoch": 0.500577006087194, "grad_norm": 7.589977741241455, "learning_rate": 1.0489982114255441e-05, "loss": 5.0666, "step": 14206 }, { "epoch": 0.5006122431001524, "grad_norm": 6.607818126678467, "learning_rate": 1.0488842205090668e-05, "loss": 3.041, "step": 14207 }, { "epoch": 0.5006474801131108, "grad_norm": 1.2131755352020264, "learning_rate": 1.0487702289558664e-05, "loss": 0.8812, "step": 14208 }, { "epoch": 0.5006827171260693, "grad_norm": 6.297287940979004, "learning_rate": 1.0486562367674271e-05, "loss": 3.0461, "step": 14209 }, { "epoch": 0.5007179541390276, "grad_norm": 6.551669120788574, "learning_rate": 1.0485422439452345e-05, "loss": 3.2934, "step": 14210 }, { "epoch": 0.500753191151986, "grad_norm": 11.051400184631348, "learning_rate": 1.0484282504907729e-05, "loss": 5.2181, "step": 14211 }, { "epoch": 0.5007884281649445, "grad_norm": 1.65911066532135, "learning_rate": 1.0483142564055268e-05, "loss": 1.0937, "step": 14212 }, { "epoch": 0.5008236651779029, "grad_norm": 4.513627529144287, "learning_rate": 1.048200261690982e-05, "loss": 4.3555, "step": 14213 }, { "epoch": 0.5008589021908613, "grad_norm": 1.253294587135315, "learning_rate": 1.048086266348622e-05, "loss": 0.8866, "step": 14214 }, { "epoch": 0.5008941392038196, "grad_norm": 5.129399299621582, "learning_rate": 1.0479722703799326e-05, "loss": 2.7829, "step": 14215 }, { "epoch": 0.5009293762167781, "grad_norm": 1.629634141921997, "learning_rate": 1.0478582737863982e-05, "loss": 0.996, "step": 14216 }, { "epoch": 0.5009646132297365, "grad_norm": 2.0492098331451416, "learning_rate": 1.0477442765695039e-05, "loss": 0.7651, "step": 14217 }, { "epoch": 0.5009998502426949, "grad_norm": 3.607698678970337, "learning_rate": 1.047630278730734e-05, "loss": 3.2697, "step": 14218 }, { "epoch": 0.5010350872556534, "grad_norm": 1.1654222011566162, "learning_rate": 1.0475162802715737e-05, "loss": 1.2395, "step": 14219 }, { "epoch": 0.5010703242686118, "grad_norm": 3.579688787460327, "learning_rate": 1.047402281193508e-05, "loss": 3.061, "step": 14220 }, { "epoch": 0.5011055612815701, "grad_norm": 6.989461898803711, "learning_rate": 1.0472882814980214e-05, "loss": 3.2616, "step": 14221 }, { "epoch": 0.5011407982945286, "grad_norm": 1.052232027053833, "learning_rate": 1.0471742811865987e-05, "loss": 1.1171, "step": 14222 }, { "epoch": 0.501176035307487, "grad_norm": 4.637919902801514, "learning_rate": 1.0470602802607255e-05, "loss": 2.9342, "step": 14223 }, { "epoch": 0.5012112723204454, "grad_norm": 7.712132930755615, "learning_rate": 1.046946278721886e-05, "loss": 5.5629, "step": 14224 }, { "epoch": 0.5012465093334038, "grad_norm": 5.694477081298828, "learning_rate": 1.0468322765715648e-05, "loss": 2.7886, "step": 14225 }, { "epoch": 0.5012817463463622, "grad_norm": 1.1213816404342651, "learning_rate": 1.0467182738112479e-05, "loss": 0.9174, "step": 14226 }, { "epoch": 0.5013169833593206, "grad_norm": 15.095193862915039, "learning_rate": 1.0466042704424191e-05, "loss": 2.9653, "step": 14227 }, { "epoch": 0.501352220372279, "grad_norm": 33.58190155029297, "learning_rate": 1.0464902664665636e-05, "loss": 2.9965, "step": 14228 }, { "epoch": 0.5013874573852375, "grad_norm": 0.8959887027740479, "learning_rate": 1.0463762618851667e-05, "loss": 1.0979, "step": 14229 }, { "epoch": 0.5014226943981959, "grad_norm": 3.1209280490875244, "learning_rate": 1.0462622566997133e-05, "loss": 2.5193, "step": 14230 }, { "epoch": 0.5014579314111542, "grad_norm": 9.908574104309082, "learning_rate": 1.0461482509116877e-05, "loss": 2.8665, "step": 14231 }, { "epoch": 0.5014931684241127, "grad_norm": 5.139084815979004, "learning_rate": 1.0460342445225756e-05, "loss": 3.1535, "step": 14232 }, { "epoch": 0.5015284054370711, "grad_norm": 1.1780765056610107, "learning_rate": 1.045920237533861e-05, "loss": 0.8836, "step": 14233 }, { "epoch": 0.5015636424500295, "grad_norm": 3.9211275577545166, "learning_rate": 1.0458062299470301e-05, "loss": 2.7584, "step": 14234 }, { "epoch": 0.5015988794629879, "grad_norm": 3.528508186340332, "learning_rate": 1.0456922217635671e-05, "loss": 3.0868, "step": 14235 }, { "epoch": 0.5016341164759464, "grad_norm": 3.005434274673462, "learning_rate": 1.045578212984957e-05, "loss": 2.89, "step": 14236 }, { "epoch": 0.5016693534889047, "grad_norm": 3.244724988937378, "learning_rate": 1.045464203612685e-05, "loss": 2.3882, "step": 14237 }, { "epoch": 0.5017045905018631, "grad_norm": 2.0042150020599365, "learning_rate": 1.0453501936482358e-05, "loss": 1.0369, "step": 14238 }, { "epoch": 0.5017398275148216, "grad_norm": 1.291702389717102, "learning_rate": 1.0452361830930944e-05, "loss": 0.8728, "step": 14239 }, { "epoch": 0.50177506452778, "grad_norm": 4.849856853485107, "learning_rate": 1.0451221719487461e-05, "loss": 2.8462, "step": 14240 }, { "epoch": 0.5018103015407384, "grad_norm": 7.987967014312744, "learning_rate": 1.0450081602166757e-05, "loss": 5.5136, "step": 14241 }, { "epoch": 0.5018455385536968, "grad_norm": 1.512007236480713, "learning_rate": 1.0448941478983686e-05, "loss": 0.868, "step": 14242 }, { "epoch": 0.5018807755666552, "grad_norm": 1.2410211563110352, "learning_rate": 1.0447801349953093e-05, "loss": 0.8374, "step": 14243 }, { "epoch": 0.5019160125796136, "grad_norm": 1.8917292356491089, "learning_rate": 1.0446661215089829e-05, "loss": 1.0312, "step": 14244 }, { "epoch": 0.5019512495925721, "grad_norm": 1.5228708982467651, "learning_rate": 1.0445521074408749e-05, "loss": 1.0073, "step": 14245 }, { "epoch": 0.5019864866055305, "grad_norm": 1.5908015966415405, "learning_rate": 1.0444380927924695e-05, "loss": 1.0839, "step": 14246 }, { "epoch": 0.5020217236184888, "grad_norm": 6.446753978729248, "learning_rate": 1.0443240775652528e-05, "loss": 3.2113, "step": 14247 }, { "epoch": 0.5020569606314472, "grad_norm": 9.411747932434082, "learning_rate": 1.0442100617607092e-05, "loss": 2.9043, "step": 14248 }, { "epoch": 0.5020921976444057, "grad_norm": 2.7896945476531982, "learning_rate": 1.0440960453803237e-05, "loss": 1.0711, "step": 14249 }, { "epoch": 0.5021274346573641, "grad_norm": 6.102477550506592, "learning_rate": 1.0439820284255817e-05, "loss": 3.2705, "step": 14250 }, { "epoch": 0.5021626716703225, "grad_norm": 1.586295485496521, "learning_rate": 1.0438680108979683e-05, "loss": 0.9627, "step": 14251 }, { "epoch": 0.502197908683281, "grad_norm": 13.0330810546875, "learning_rate": 1.0437539927989681e-05, "loss": 4.8681, "step": 14252 }, { "epoch": 0.5022331456962393, "grad_norm": 1.7376068830490112, "learning_rate": 1.0436399741300672e-05, "loss": 0.9566, "step": 14253 }, { "epoch": 0.5022683827091977, "grad_norm": 5.386181354522705, "learning_rate": 1.0435259548927497e-05, "loss": 3.2475, "step": 14254 }, { "epoch": 0.5023036197221562, "grad_norm": 1.737152338027954, "learning_rate": 1.0434119350885012e-05, "loss": 1.0633, "step": 14255 }, { "epoch": 0.5023388567351146, "grad_norm": 10.400885581970215, "learning_rate": 1.0432979147188067e-05, "loss": 3.0909, "step": 14256 }, { "epoch": 0.502374093748073, "grad_norm": 1.5897767543792725, "learning_rate": 1.0431838937851515e-05, "loss": 0.8028, "step": 14257 }, { "epoch": 0.5024093307610313, "grad_norm": 1.965344786643982, "learning_rate": 1.04306987228902e-05, "loss": 1.1338, "step": 14258 }, { "epoch": 0.5024445677739898, "grad_norm": 7.416628837585449, "learning_rate": 1.0429558502318985e-05, "loss": 7.346, "step": 14259 }, { "epoch": 0.5024798047869482, "grad_norm": 8.216873168945312, "learning_rate": 1.0428418276152711e-05, "loss": 3.3916, "step": 14260 }, { "epoch": 0.5025150417999066, "grad_norm": 1.0482181310653687, "learning_rate": 1.042727804440624e-05, "loss": 0.9893, "step": 14261 }, { "epoch": 0.5025502788128651, "grad_norm": 4.929145812988281, "learning_rate": 1.0426137807094413e-05, "loss": 2.8119, "step": 14262 }, { "epoch": 0.5025855158258234, "grad_norm": 4.851727485656738, "learning_rate": 1.0424997564232091e-05, "loss": 3.3055, "step": 14263 }, { "epoch": 0.5026207528387818, "grad_norm": 1.234113335609436, "learning_rate": 1.0423857315834118e-05, "loss": 0.8769, "step": 14264 }, { "epoch": 0.5026559898517403, "grad_norm": 3.175840139389038, "learning_rate": 1.0422717061915353e-05, "loss": 3.0862, "step": 14265 }, { "epoch": 0.5026912268646987, "grad_norm": 12.165388107299805, "learning_rate": 1.042157680249064e-05, "loss": 3.379, "step": 14266 }, { "epoch": 0.5027264638776571, "grad_norm": 1.0539851188659668, "learning_rate": 1.0420436537574838e-05, "loss": 1.0058, "step": 14267 }, { "epoch": 0.5027617008906154, "grad_norm": 4.523126125335693, "learning_rate": 1.0419296267182797e-05, "loss": 3.1288, "step": 14268 }, { "epoch": 0.5027969379035739, "grad_norm": 9.122249603271484, "learning_rate": 1.0418155991329369e-05, "loss": 5.3255, "step": 14269 }, { "epoch": 0.5028321749165323, "grad_norm": 6.623316287994385, "learning_rate": 1.0417015710029408e-05, "loss": 3.2478, "step": 14270 }, { "epoch": 0.5028674119294907, "grad_norm": 5.383681774139404, "learning_rate": 1.041587542329776e-05, "loss": 3.0215, "step": 14271 }, { "epoch": 0.5029026489424492, "grad_norm": 0.8287683725357056, "learning_rate": 1.0414735131149282e-05, "loss": 1.0418, "step": 14272 }, { "epoch": 0.5029378859554076, "grad_norm": 5.451388835906982, "learning_rate": 1.0413594833598828e-05, "loss": 3.225, "step": 14273 }, { "epoch": 0.5029731229683659, "grad_norm": 5.722011566162109, "learning_rate": 1.0412454530661247e-05, "loss": 5.7926, "step": 14274 }, { "epoch": 0.5030083599813244, "grad_norm": 3.7543675899505615, "learning_rate": 1.0411314222351395e-05, "loss": 2.8888, "step": 14275 }, { "epoch": 0.5030435969942828, "grad_norm": 4.028902053833008, "learning_rate": 1.041017390868412e-05, "loss": 3.3244, "step": 14276 }, { "epoch": 0.5030788340072412, "grad_norm": 1.2022768259048462, "learning_rate": 1.040903358967428e-05, "loss": 0.8454, "step": 14277 }, { "epoch": 0.5031140710201997, "grad_norm": 4.314349174499512, "learning_rate": 1.0407893265336726e-05, "loss": 3.0124, "step": 14278 }, { "epoch": 0.503149308033158, "grad_norm": 4.249244689941406, "learning_rate": 1.0406752935686308e-05, "loss": 3.3506, "step": 14279 }, { "epoch": 0.5031845450461164, "grad_norm": 5.384983539581299, "learning_rate": 1.0405612600737884e-05, "loss": 2.5375, "step": 14280 }, { "epoch": 0.5032197820590748, "grad_norm": 5.359556198120117, "learning_rate": 1.0404472260506303e-05, "loss": 3.5607, "step": 14281 }, { "epoch": 0.5032550190720333, "grad_norm": 7.087229251861572, "learning_rate": 1.0403331915006419e-05, "loss": 3.0653, "step": 14282 }, { "epoch": 0.5032902560849917, "grad_norm": 5.503868103027344, "learning_rate": 1.0402191564253084e-05, "loss": 2.8085, "step": 14283 }, { "epoch": 0.50332549309795, "grad_norm": 4.199734687805176, "learning_rate": 1.0401051208261153e-05, "loss": 3.0299, "step": 14284 }, { "epoch": 0.5033607301109085, "grad_norm": 1.1589350700378418, "learning_rate": 1.039991084704548e-05, "loss": 0.7986, "step": 14285 }, { "epoch": 0.5033959671238669, "grad_norm": 3.2519402503967285, "learning_rate": 1.0398770480620917e-05, "loss": 3.0223, "step": 14286 }, { "epoch": 0.5034312041368253, "grad_norm": 1.2970103025436401, "learning_rate": 1.0397630109002318e-05, "loss": 0.8859, "step": 14287 }, { "epoch": 0.5034664411497838, "grad_norm": 6.828129291534424, "learning_rate": 1.0396489732204536e-05, "loss": 5.2793, "step": 14288 }, { "epoch": 0.5035016781627422, "grad_norm": 1.5823047161102295, "learning_rate": 1.0395349350242426e-05, "loss": 1.0155, "step": 14289 }, { "epoch": 0.5035369151757005, "grad_norm": 1.1814684867858887, "learning_rate": 1.0394208963130837e-05, "loss": 0.8441, "step": 14290 }, { "epoch": 0.5035721521886589, "grad_norm": 2.7546231746673584, "learning_rate": 1.039306857088463e-05, "loss": 1.0378, "step": 14291 }, { "epoch": 0.5036073892016174, "grad_norm": 1.6905943155288696, "learning_rate": 1.0391928173518653e-05, "loss": 0.8197, "step": 14292 }, { "epoch": 0.5036426262145758, "grad_norm": 2.069291591644287, "learning_rate": 1.0390787771047761e-05, "loss": 0.9098, "step": 14293 }, { "epoch": 0.5036778632275342, "grad_norm": 7.01664400100708, "learning_rate": 1.0389647363486808e-05, "loss": 2.8995, "step": 14294 }, { "epoch": 0.5037131002404927, "grad_norm": 4.532363414764404, "learning_rate": 1.038850695085065e-05, "loss": 5.3266, "step": 14295 }, { "epoch": 0.503748337253451, "grad_norm": 1.7132012844085693, "learning_rate": 1.0387366533154137e-05, "loss": 0.7364, "step": 14296 }, { "epoch": 0.5037835742664094, "grad_norm": 24.77471351623535, "learning_rate": 1.038622611041213e-05, "loss": 2.9858, "step": 14297 }, { "epoch": 0.5038188112793679, "grad_norm": 1.2168325185775757, "learning_rate": 1.0385085682639474e-05, "loss": 0.9874, "step": 14298 }, { "epoch": 0.5038540482923263, "grad_norm": 1.07725191116333, "learning_rate": 1.0383945249851033e-05, "loss": 0.9928, "step": 14299 }, { "epoch": 0.5038892853052847, "grad_norm": 1.7491427659988403, "learning_rate": 1.0382804812061653e-05, "loss": 0.6567, "step": 14300 }, { "epoch": 0.503924522318243, "grad_norm": 2.09429669380188, "learning_rate": 1.0381664369286191e-05, "loss": 0.8693, "step": 14301 }, { "epoch": 0.5039597593312015, "grad_norm": 1.8147027492523193, "learning_rate": 1.0380523921539503e-05, "loss": 0.8143, "step": 14302 }, { "epoch": 0.5039949963441599, "grad_norm": 1.0721759796142578, "learning_rate": 1.037938346883644e-05, "loss": 0.931, "step": 14303 }, { "epoch": 0.5040302333571183, "grad_norm": 7.096301078796387, "learning_rate": 1.0378243011191863e-05, "loss": 2.9099, "step": 14304 }, { "epoch": 0.5040654703700768, "grad_norm": 2.733379364013672, "learning_rate": 1.037710254862062e-05, "loss": 0.8319, "step": 14305 }, { "epoch": 0.5041007073830351, "grad_norm": 1.1794601678848267, "learning_rate": 1.037596208113757e-05, "loss": 0.7814, "step": 14306 }, { "epoch": 0.5041359443959935, "grad_norm": 7.588767051696777, "learning_rate": 1.0374821608757563e-05, "loss": 7.0427, "step": 14307 }, { "epoch": 0.504171181408952, "grad_norm": 4.034278869628906, "learning_rate": 1.0373681131495463e-05, "loss": 2.9273, "step": 14308 }, { "epoch": 0.5042064184219104, "grad_norm": 3.503535509109497, "learning_rate": 1.0372540649366112e-05, "loss": 2.9996, "step": 14309 }, { "epoch": 0.5042416554348688, "grad_norm": 5.101656436920166, "learning_rate": 1.0371400162384375e-05, "loss": 2.7731, "step": 14310 }, { "epoch": 0.5042768924478273, "grad_norm": 3.5408294200897217, "learning_rate": 1.0370259670565104e-05, "loss": 2.9104, "step": 14311 }, { "epoch": 0.5043121294607856, "grad_norm": 5.811540603637695, "learning_rate": 1.0369119173923153e-05, "loss": 3.2279, "step": 14312 }, { "epoch": 0.504347366473744, "grad_norm": 9.200268745422363, "learning_rate": 1.0367978672473375e-05, "loss": 5.262, "step": 14313 }, { "epoch": 0.5043826034867024, "grad_norm": 1.9567394256591797, "learning_rate": 1.036683816623063e-05, "loss": 0.9003, "step": 14314 }, { "epoch": 0.5044178404996609, "grad_norm": 1.8836448192596436, "learning_rate": 1.0365697655209771e-05, "loss": 0.9594, "step": 14315 }, { "epoch": 0.5044530775126193, "grad_norm": 1.2158125638961792, "learning_rate": 1.0364557139425656e-05, "loss": 0.8435, "step": 14316 }, { "epoch": 0.5044883145255776, "grad_norm": 4.87724494934082, "learning_rate": 1.0363416618893134e-05, "loss": 4.6528, "step": 14317 }, { "epoch": 0.5045235515385361, "grad_norm": 5.22485876083374, "learning_rate": 1.0362276093627068e-05, "loss": 3.4573, "step": 14318 }, { "epoch": 0.5045587885514945, "grad_norm": 2.003241777420044, "learning_rate": 1.0361135563642307e-05, "loss": 1.153, "step": 14319 }, { "epoch": 0.5045940255644529, "grad_norm": 1.2717413902282715, "learning_rate": 1.035999502895371e-05, "loss": 0.9714, "step": 14320 }, { "epoch": 0.5046292625774114, "grad_norm": 6.694680213928223, "learning_rate": 1.0358854489576132e-05, "loss": 3.5105, "step": 14321 }, { "epoch": 0.5046644995903697, "grad_norm": 1.3269801139831543, "learning_rate": 1.0357713945524425e-05, "loss": 1.2208, "step": 14322 }, { "epoch": 0.5046997366033281, "grad_norm": 5.9722700119018555, "learning_rate": 1.0356573396813451e-05, "loss": 2.516, "step": 14323 }, { "epoch": 0.5047349736162865, "grad_norm": 1.691484808921814, "learning_rate": 1.0355432843458067e-05, "loss": 1.0642, "step": 14324 }, { "epoch": 0.504770210629245, "grad_norm": 1.4885088205337524, "learning_rate": 1.035429228547312e-05, "loss": 0.9015, "step": 14325 }, { "epoch": 0.5048054476422034, "grad_norm": 3.7738661766052246, "learning_rate": 1.035315172287347e-05, "loss": 3.0961, "step": 14326 }, { "epoch": 0.5048406846551617, "grad_norm": 3.9894161224365234, "learning_rate": 1.035201115567398e-05, "loss": 2.4654, "step": 14327 }, { "epoch": 0.5048759216681202, "grad_norm": 13.024625778198242, "learning_rate": 1.0350870583889495e-05, "loss": 3.0208, "step": 14328 }, { "epoch": 0.5049111586810786, "grad_norm": 3.0456390380859375, "learning_rate": 1.0349730007534877e-05, "loss": 3.0767, "step": 14329 }, { "epoch": 0.504946395694037, "grad_norm": 13.920735359191895, "learning_rate": 1.0348589426624982e-05, "loss": 2.7091, "step": 14330 }, { "epoch": 0.5049816327069955, "grad_norm": 4.470091819763184, "learning_rate": 1.0347448841174664e-05, "loss": 3.3293, "step": 14331 }, { "epoch": 0.5050168697199539, "grad_norm": 0.9339510798454285, "learning_rate": 1.034630825119878e-05, "loss": 0.7922, "step": 14332 }, { "epoch": 0.5050521067329122, "grad_norm": 2.575814962387085, "learning_rate": 1.034516765671219e-05, "loss": 0.9035, "step": 14333 }, { "epoch": 0.5050873437458706, "grad_norm": 1.0601253509521484, "learning_rate": 1.0344027057729744e-05, "loss": 1.0207, "step": 14334 }, { "epoch": 0.5051225807588291, "grad_norm": 1.1528587341308594, "learning_rate": 1.0342886454266306e-05, "loss": 0.93, "step": 14335 }, { "epoch": 0.5051578177717875, "grad_norm": 7.11469030380249, "learning_rate": 1.0341745846336721e-05, "loss": 5.1651, "step": 14336 }, { "epoch": 0.5051930547847459, "grad_norm": 4.2790045738220215, "learning_rate": 1.0340605233955861e-05, "loss": 2.9827, "step": 14337 }, { "epoch": 0.5052282917977043, "grad_norm": 2.845287561416626, "learning_rate": 1.0339464617138568e-05, "loss": 2.7492, "step": 14338 }, { "epoch": 0.5052635288106627, "grad_norm": 1.4431532621383667, "learning_rate": 1.0338323995899708e-05, "loss": 0.875, "step": 14339 }, { "epoch": 0.5052987658236211, "grad_norm": 2.0651378631591797, "learning_rate": 1.0337183370254136e-05, "loss": 0.963, "step": 14340 }, { "epoch": 0.5053340028365796, "grad_norm": 1.3606529235839844, "learning_rate": 1.0336042740216706e-05, "loss": 0.8835, "step": 14341 }, { "epoch": 0.505369239849538, "grad_norm": 6.239525318145752, "learning_rate": 1.0334902105802276e-05, "loss": 5.7873, "step": 14342 }, { "epoch": 0.5054044768624963, "grad_norm": 3.219923734664917, "learning_rate": 1.0333761467025703e-05, "loss": 2.3788, "step": 14343 }, { "epoch": 0.5054397138754548, "grad_norm": 2.032317876815796, "learning_rate": 1.0332620823901844e-05, "loss": 0.9036, "step": 14344 }, { "epoch": 0.5054749508884132, "grad_norm": 4.872641086578369, "learning_rate": 1.0331480176445558e-05, "loss": 2.7433, "step": 14345 }, { "epoch": 0.5055101879013716, "grad_norm": 4.0535454750061035, "learning_rate": 1.0330339524671705e-05, "loss": 3.8882, "step": 14346 }, { "epoch": 0.50554542491433, "grad_norm": 3.571056842803955, "learning_rate": 1.032919886859513e-05, "loss": 3.6042, "step": 14347 }, { "epoch": 0.5055806619272885, "grad_norm": 1.4888019561767578, "learning_rate": 1.0328058208230702e-05, "loss": 1.1102, "step": 14348 }, { "epoch": 0.5056158989402468, "grad_norm": 6.034604072570801, "learning_rate": 1.0326917543593271e-05, "loss": 2.9268, "step": 14349 }, { "epoch": 0.5056511359532052, "grad_norm": 1.9107792377471924, "learning_rate": 1.0325776874697699e-05, "loss": 0.7778, "step": 14350 }, { "epoch": 0.5056863729661637, "grad_norm": 3.208066463470459, "learning_rate": 1.0324636201558842e-05, "loss": 2.5111, "step": 14351 }, { "epoch": 0.5057216099791221, "grad_norm": 1.1480954885482788, "learning_rate": 1.0323495524191559e-05, "loss": 0.6969, "step": 14352 }, { "epoch": 0.5057568469920805, "grad_norm": 2.89528751373291, "learning_rate": 1.0322354842610702e-05, "loss": 2.3127, "step": 14353 }, { "epoch": 0.5057920840050389, "grad_norm": 1.8575853109359741, "learning_rate": 1.0321214156831137e-05, "loss": 0.848, "step": 14354 }, { "epoch": 0.5058273210179973, "grad_norm": 1.2070293426513672, "learning_rate": 1.032007346686771e-05, "loss": 0.9032, "step": 14355 }, { "epoch": 0.5058625580309557, "grad_norm": 1.4224079847335815, "learning_rate": 1.0318932772735291e-05, "loss": 0.9546, "step": 14356 }, { "epoch": 0.5058977950439141, "grad_norm": 8.058816909790039, "learning_rate": 1.0317792074448729e-05, "loss": 2.7422, "step": 14357 }, { "epoch": 0.5059330320568726, "grad_norm": 1.555515170097351, "learning_rate": 1.0316651372022886e-05, "loss": 0.7882, "step": 14358 }, { "epoch": 0.505968269069831, "grad_norm": 1.4147242307662964, "learning_rate": 1.0315510665472616e-05, "loss": 1.003, "step": 14359 }, { "epoch": 0.5060035060827893, "grad_norm": 3.9157135486602783, "learning_rate": 1.0314369954812781e-05, "loss": 2.9887, "step": 14360 }, { "epoch": 0.5060387430957478, "grad_norm": 14.975432395935059, "learning_rate": 1.0313229240058239e-05, "loss": 2.9377, "step": 14361 }, { "epoch": 0.5060739801087062, "grad_norm": 1.3439515829086304, "learning_rate": 1.0312088521223845e-05, "loss": 0.8653, "step": 14362 }, { "epoch": 0.5061092171216646, "grad_norm": 1.241258978843689, "learning_rate": 1.0310947798324458e-05, "loss": 0.9379, "step": 14363 }, { "epoch": 0.5061444541346231, "grad_norm": 5.222277641296387, "learning_rate": 1.0309807071374935e-05, "loss": 2.7364, "step": 14364 }, { "epoch": 0.5061796911475814, "grad_norm": 4.9281439781188965, "learning_rate": 1.030866634039014e-05, "loss": 2.7575, "step": 14365 }, { "epoch": 0.5062149281605398, "grad_norm": 4.101578712463379, "learning_rate": 1.0307525605384921e-05, "loss": 2.4735, "step": 14366 }, { "epoch": 0.5062501651734982, "grad_norm": 2.953381299972534, "learning_rate": 1.0306384866374146e-05, "loss": 2.341, "step": 14367 }, { "epoch": 0.5062854021864567, "grad_norm": 1.0290790796279907, "learning_rate": 1.0305244123372666e-05, "loss": 0.7892, "step": 14368 }, { "epoch": 0.5063206391994151, "grad_norm": 4.635192394256592, "learning_rate": 1.0304103376395345e-05, "loss": 4.9871, "step": 14369 }, { "epoch": 0.5063558762123734, "grad_norm": 9.170808792114258, "learning_rate": 1.0302962625457038e-05, "loss": 5.4155, "step": 14370 }, { "epoch": 0.5063911132253319, "grad_norm": 1.1373634338378906, "learning_rate": 1.0301821870572604e-05, "loss": 0.7947, "step": 14371 }, { "epoch": 0.5064263502382903, "grad_norm": 2.0151405334472656, "learning_rate": 1.0300681111756901e-05, "loss": 0.9882, "step": 14372 }, { "epoch": 0.5064615872512487, "grad_norm": 3.4014410972595215, "learning_rate": 1.0299540349024791e-05, "loss": 2.4182, "step": 14373 }, { "epoch": 0.5064968242642072, "grad_norm": 6.844539642333984, "learning_rate": 1.0298399582391125e-05, "loss": 3.1437, "step": 14374 }, { "epoch": 0.5065320612771655, "grad_norm": 1.0037833452224731, "learning_rate": 1.0297258811870772e-05, "loss": 1.0862, "step": 14375 }, { "epoch": 0.5065672982901239, "grad_norm": 1.4234906435012817, "learning_rate": 1.0296118037478578e-05, "loss": 0.8134, "step": 14376 }, { "epoch": 0.5066025353030824, "grad_norm": 3.9738569259643555, "learning_rate": 1.0294977259229415e-05, "loss": 2.7096, "step": 14377 }, { "epoch": 0.5066377723160408, "grad_norm": 3.502967357635498, "learning_rate": 1.0293836477138135e-05, "loss": 3.0021, "step": 14378 }, { "epoch": 0.5066730093289992, "grad_norm": 5.834810256958008, "learning_rate": 1.0292695691219594e-05, "loss": 4.525, "step": 14379 }, { "epoch": 0.5067082463419575, "grad_norm": 4.567699432373047, "learning_rate": 1.0291554901488655e-05, "loss": 3.1591, "step": 14380 }, { "epoch": 0.506743483354916, "grad_norm": 6.713141918182373, "learning_rate": 1.0290414107960179e-05, "loss": 5.5019, "step": 14381 }, { "epoch": 0.5067787203678744, "grad_norm": 7.308163642883301, "learning_rate": 1.0289273310649021e-05, "loss": 3.7726, "step": 14382 }, { "epoch": 0.5068139573808328, "grad_norm": 4.274272441864014, "learning_rate": 1.0288132509570042e-05, "loss": 2.9313, "step": 14383 }, { "epoch": 0.5068491943937913, "grad_norm": 1.597733974456787, "learning_rate": 1.02869917047381e-05, "loss": 0.755, "step": 14384 }, { "epoch": 0.5068844314067497, "grad_norm": 5.700599670410156, "learning_rate": 1.0285850896168052e-05, "loss": 4.9369, "step": 14385 }, { "epoch": 0.506919668419708, "grad_norm": 1.2364764213562012, "learning_rate": 1.0284710083874765e-05, "loss": 0.9637, "step": 14386 }, { "epoch": 0.5069549054326665, "grad_norm": 5.179368019104004, "learning_rate": 1.0283569267873087e-05, "loss": 2.8965, "step": 14387 }, { "epoch": 0.5069901424456249, "grad_norm": 1.4170053005218506, "learning_rate": 1.0282428448177885e-05, "loss": 0.8136, "step": 14388 }, { "epoch": 0.5070253794585833, "grad_norm": 1.6445255279541016, "learning_rate": 1.028128762480402e-05, "loss": 1.0902, "step": 14389 }, { "epoch": 0.5070606164715417, "grad_norm": 2.324655771255493, "learning_rate": 1.0280146797766344e-05, "loss": 0.753, "step": 14390 }, { "epoch": 0.5070958534845001, "grad_norm": 5.754016876220703, "learning_rate": 1.027900596707972e-05, "loss": 2.9512, "step": 14391 }, { "epoch": 0.5071310904974585, "grad_norm": 1.569861650466919, "learning_rate": 1.0277865132759012e-05, "loss": 0.7996, "step": 14392 }, { "epoch": 0.5071663275104169, "grad_norm": 5.22941255569458, "learning_rate": 1.0276724294819069e-05, "loss": 2.9313, "step": 14393 }, { "epoch": 0.5072015645233754, "grad_norm": 7.067366123199463, "learning_rate": 1.0275583453274763e-05, "loss": 5.0257, "step": 14394 }, { "epoch": 0.5072368015363338, "grad_norm": 34.31807327270508, "learning_rate": 1.0274442608140943e-05, "loss": 3.0394, "step": 14395 }, { "epoch": 0.5072720385492921, "grad_norm": 30.95977020263672, "learning_rate": 1.0273301759432476e-05, "loss": 3.344, "step": 14396 }, { "epoch": 0.5073072755622506, "grad_norm": 6.705629825592041, "learning_rate": 1.0272160907164216e-05, "loss": 3.3048, "step": 14397 }, { "epoch": 0.507342512575209, "grad_norm": 5.269925594329834, "learning_rate": 1.0271020051351029e-05, "loss": 2.9079, "step": 14398 }, { "epoch": 0.5073777495881674, "grad_norm": 1.5093295574188232, "learning_rate": 1.0269879192007769e-05, "loss": 0.9141, "step": 14399 }, { "epoch": 0.5074129866011258, "grad_norm": 1.0216736793518066, "learning_rate": 1.0268738329149296e-05, "loss": 0.7244, "step": 14400 }, { "epoch": 0.5074482236140843, "grad_norm": 1.4587490558624268, "learning_rate": 1.0267597462790476e-05, "loss": 0.8359, "step": 14401 }, { "epoch": 0.5074834606270426, "grad_norm": 2.5218801498413086, "learning_rate": 1.0266456592946164e-05, "loss": 0.7739, "step": 14402 }, { "epoch": 0.507518697640001, "grad_norm": 2.415929079055786, "learning_rate": 1.0265315719631222e-05, "loss": 2.7187, "step": 14403 }, { "epoch": 0.5075539346529595, "grad_norm": 5.714524745941162, "learning_rate": 1.0264174842860507e-05, "loss": 4.6846, "step": 14404 }, { "epoch": 0.5075891716659179, "grad_norm": 1.29810631275177, "learning_rate": 1.026303396264888e-05, "loss": 0.8066, "step": 14405 }, { "epoch": 0.5076244086788763, "grad_norm": 7.9445037841796875, "learning_rate": 1.0261893079011203e-05, "loss": 5.4937, "step": 14406 }, { "epoch": 0.5076596456918347, "grad_norm": 2.0554182529449463, "learning_rate": 1.0260752191962335e-05, "loss": 0.7374, "step": 14407 }, { "epoch": 0.5076948827047931, "grad_norm": 1.1201682090759277, "learning_rate": 1.0259611301517138e-05, "loss": 0.8222, "step": 14408 }, { "epoch": 0.5077301197177515, "grad_norm": 0.9663760662078857, "learning_rate": 1.025847040769047e-05, "loss": 1.1737, "step": 14409 }, { "epoch": 0.5077653567307099, "grad_norm": 6.095882415771484, "learning_rate": 1.0257329510497191e-05, "loss": 5.5113, "step": 14410 }, { "epoch": 0.5078005937436684, "grad_norm": 1.2847483158111572, "learning_rate": 1.0256188609952166e-05, "loss": 1.0049, "step": 14411 }, { "epoch": 0.5078358307566267, "grad_norm": 1.9156181812286377, "learning_rate": 1.0255047706070247e-05, "loss": 0.9096, "step": 14412 }, { "epoch": 0.5078710677695851, "grad_norm": 4.22566032409668, "learning_rate": 1.0253906798866302e-05, "loss": 2.6482, "step": 14413 }, { "epoch": 0.5079063047825436, "grad_norm": 1.0529932975769043, "learning_rate": 1.0252765888355186e-05, "loss": 1.1784, "step": 14414 }, { "epoch": 0.507941541795502, "grad_norm": 3.239952325820923, "learning_rate": 1.0251624974551766e-05, "loss": 3.3459, "step": 14415 }, { "epoch": 0.5079767788084604, "grad_norm": 4.615278720855713, "learning_rate": 1.0250484057470895e-05, "loss": 2.4284, "step": 14416 }, { "epoch": 0.5080120158214189, "grad_norm": 2.003124237060547, "learning_rate": 1.0249343137127439e-05, "loss": 0.9763, "step": 14417 }, { "epoch": 0.5080472528343772, "grad_norm": 1.2573176622390747, "learning_rate": 1.0248202213536257e-05, "loss": 0.7696, "step": 14418 }, { "epoch": 0.5080824898473356, "grad_norm": 4.969417095184326, "learning_rate": 1.0247061286712212e-05, "loss": 4.9879, "step": 14419 }, { "epoch": 0.5081177268602941, "grad_norm": 1.3713834285736084, "learning_rate": 1.0245920356670156e-05, "loss": 1.065, "step": 14420 }, { "epoch": 0.5081529638732525, "grad_norm": 5.469782829284668, "learning_rate": 1.0244779423424962e-05, "loss": 6.3321, "step": 14421 }, { "epoch": 0.5081882008862109, "grad_norm": 1.433056354522705, "learning_rate": 1.0243638486991481e-05, "loss": 0.8236, "step": 14422 }, { "epoch": 0.5082234378991692, "grad_norm": 6.224735736846924, "learning_rate": 1.024249754738458e-05, "loss": 5.2626, "step": 14423 }, { "epoch": 0.5082586749121277, "grad_norm": 6.434048175811768, "learning_rate": 1.0241356604619117e-05, "loss": 3.5253, "step": 14424 }, { "epoch": 0.5082939119250861, "grad_norm": 5.184652805328369, "learning_rate": 1.024021565870995e-05, "loss": 3.5389, "step": 14425 }, { "epoch": 0.5083291489380445, "grad_norm": 5.427766799926758, "learning_rate": 1.0239074709671947e-05, "loss": 3.038, "step": 14426 }, { "epoch": 0.508364385951003, "grad_norm": 1.2681705951690674, "learning_rate": 1.0237933757519967e-05, "loss": 1.1528, "step": 14427 }, { "epoch": 0.5083996229639614, "grad_norm": 1.5415046215057373, "learning_rate": 1.0236792802268867e-05, "loss": 0.8846, "step": 14428 }, { "epoch": 0.5084348599769197, "grad_norm": 1.348810076713562, "learning_rate": 1.023565184393351e-05, "loss": 0.9852, "step": 14429 }, { "epoch": 0.5084700969898782, "grad_norm": 2.057330846786499, "learning_rate": 1.0234510882528761e-05, "loss": 0.9646, "step": 14430 }, { "epoch": 0.5085053340028366, "grad_norm": 1.5203009843826294, "learning_rate": 1.0233369918069473e-05, "loss": 0.7539, "step": 14431 }, { "epoch": 0.508540571015795, "grad_norm": 2.3277244567871094, "learning_rate": 1.0232228950570518e-05, "loss": 1.2534, "step": 14432 }, { "epoch": 0.5085758080287534, "grad_norm": 1.8287402391433716, "learning_rate": 1.0231087980046743e-05, "loss": 0.8204, "step": 14433 }, { "epoch": 0.5086110450417118, "grad_norm": 1.298865795135498, "learning_rate": 1.0229947006513026e-05, "loss": 0.9568, "step": 14434 }, { "epoch": 0.5086462820546702, "grad_norm": 5.002750396728516, "learning_rate": 1.0228806029984217e-05, "loss": 3.3518, "step": 14435 }, { "epoch": 0.5086815190676286, "grad_norm": 1.3854360580444336, "learning_rate": 1.0227665050475179e-05, "loss": 0.9294, "step": 14436 }, { "epoch": 0.5087167560805871, "grad_norm": 5.71954345703125, "learning_rate": 1.0226524068000776e-05, "loss": 3.4298, "step": 14437 }, { "epoch": 0.5087519930935455, "grad_norm": 1.6159847974777222, "learning_rate": 1.022538308257587e-05, "loss": 0.9035, "step": 14438 }, { "epoch": 0.5087872301065038, "grad_norm": 1.1288385391235352, "learning_rate": 1.0224242094215315e-05, "loss": 1.0507, "step": 14439 }, { "epoch": 0.5088224671194623, "grad_norm": 6.292333126068115, "learning_rate": 1.0223101102933983e-05, "loss": 2.9844, "step": 14440 }, { "epoch": 0.5088577041324207, "grad_norm": 8.112204551696777, "learning_rate": 1.0221960108746728e-05, "loss": 5.9901, "step": 14441 }, { "epoch": 0.5088929411453791, "grad_norm": 1.2153106927871704, "learning_rate": 1.0220819111668418e-05, "loss": 1.451, "step": 14442 }, { "epoch": 0.5089281781583375, "grad_norm": 4.342495918273926, "learning_rate": 1.021967811171391e-05, "loss": 2.5442, "step": 14443 }, { "epoch": 0.508963415171296, "grad_norm": 5.208501815795898, "learning_rate": 1.0218537108898064e-05, "loss": 4.7198, "step": 14444 }, { "epoch": 0.5089986521842543, "grad_norm": 7.4843525886535645, "learning_rate": 1.0217396103235744e-05, "loss": 5.9094, "step": 14445 }, { "epoch": 0.5090338891972127, "grad_norm": 4.45363187789917, "learning_rate": 1.0216255094741814e-05, "loss": 2.7997, "step": 14446 }, { "epoch": 0.5090691262101712, "grad_norm": 6.846858024597168, "learning_rate": 1.0215114083431132e-05, "loss": 3.1534, "step": 14447 }, { "epoch": 0.5091043632231296, "grad_norm": 1.5693542957305908, "learning_rate": 1.0213973069318563e-05, "loss": 0.8531, "step": 14448 }, { "epoch": 0.509139600236088, "grad_norm": 1.276831030845642, "learning_rate": 1.0212832052418968e-05, "loss": 0.8717, "step": 14449 }, { "epoch": 0.5091748372490464, "grad_norm": 1.7214446067810059, "learning_rate": 1.0211691032747208e-05, "loss": 0.9577, "step": 14450 }, { "epoch": 0.5092100742620048, "grad_norm": 1.4633597135543823, "learning_rate": 1.0210550010318147e-05, "loss": 1.0853, "step": 14451 }, { "epoch": 0.5092453112749632, "grad_norm": 2.8508284091949463, "learning_rate": 1.0209408985146642e-05, "loss": 1.0904, "step": 14452 }, { "epoch": 0.5092805482879217, "grad_norm": 1.471816062927246, "learning_rate": 1.0208267957247561e-05, "loss": 0.9982, "step": 14453 }, { "epoch": 0.5093157853008801, "grad_norm": 1.1798490285873413, "learning_rate": 1.0207126926635762e-05, "loss": 1.0292, "step": 14454 }, { "epoch": 0.5093510223138384, "grad_norm": 1.6701334714889526, "learning_rate": 1.0205985893326109e-05, "loss": 0.9755, "step": 14455 }, { "epoch": 0.5093862593267968, "grad_norm": 2.7030351161956787, "learning_rate": 1.0204844857333462e-05, "loss": 1.0371, "step": 14456 }, { "epoch": 0.5094214963397553, "grad_norm": 1.5845361948013306, "learning_rate": 1.020370381867269e-05, "loss": 0.9297, "step": 14457 }, { "epoch": 0.5094567333527137, "grad_norm": 4.666733264923096, "learning_rate": 1.0202562777358642e-05, "loss": 2.684, "step": 14458 }, { "epoch": 0.5094919703656721, "grad_norm": 1.1075516939163208, "learning_rate": 1.0201421733406194e-05, "loss": 0.8431, "step": 14459 }, { "epoch": 0.5095272073786306, "grad_norm": 1.589920997619629, "learning_rate": 1.0200280686830197e-05, "loss": 0.9643, "step": 14460 }, { "epoch": 0.5095624443915889, "grad_norm": 4.991813659667969, "learning_rate": 1.0199139637645524e-05, "loss": 4.488, "step": 14461 }, { "epoch": 0.5095976814045473, "grad_norm": 3.732318878173828, "learning_rate": 1.0197998585867029e-05, "loss": 2.7989, "step": 14462 }, { "epoch": 0.5096329184175058, "grad_norm": 1.2593426704406738, "learning_rate": 1.0196857531509576e-05, "loss": 0.8426, "step": 14463 }, { "epoch": 0.5096681554304642, "grad_norm": 1.2050660848617554, "learning_rate": 1.019571647458803e-05, "loss": 1.1133, "step": 14464 }, { "epoch": 0.5097033924434226, "grad_norm": 1.0469471216201782, "learning_rate": 1.0194575415117253e-05, "loss": 0.9266, "step": 14465 }, { "epoch": 0.5097386294563809, "grad_norm": 1.0362427234649658, "learning_rate": 1.0193434353112106e-05, "loss": 0.9165, "step": 14466 }, { "epoch": 0.5097738664693394, "grad_norm": 1.7664748430252075, "learning_rate": 1.0192293288587452e-05, "loss": 0.9063, "step": 14467 }, { "epoch": 0.5098091034822978, "grad_norm": 11.911670684814453, "learning_rate": 1.019115222155815e-05, "loss": 3.3306, "step": 14468 }, { "epoch": 0.5098443404952562, "grad_norm": 1.5563671588897705, "learning_rate": 1.0190011152039069e-05, "loss": 0.8601, "step": 14469 }, { "epoch": 0.5098795775082147, "grad_norm": 13.175888061523438, "learning_rate": 1.018887008004507e-05, "loss": 2.774, "step": 14470 }, { "epoch": 0.509914814521173, "grad_norm": 5.407647609710693, "learning_rate": 1.0187729005591012e-05, "loss": 3.0417, "step": 14471 }, { "epoch": 0.5099500515341314, "grad_norm": 1.4643645286560059, "learning_rate": 1.0186587928691763e-05, "loss": 0.9956, "step": 14472 }, { "epoch": 0.5099852885470899, "grad_norm": 1.5472897291183472, "learning_rate": 1.0185446849362178e-05, "loss": 0.7425, "step": 14473 }, { "epoch": 0.5100205255600483, "grad_norm": 1.0747867822647095, "learning_rate": 1.0184305767617126e-05, "loss": 0.9364, "step": 14474 }, { "epoch": 0.5100557625730067, "grad_norm": 1.8793919086456299, "learning_rate": 1.0183164683471467e-05, "loss": 0.7598, "step": 14475 }, { "epoch": 0.510090999585965, "grad_norm": 1.392400860786438, "learning_rate": 1.0182023596940067e-05, "loss": 0.8195, "step": 14476 }, { "epoch": 0.5101262365989235, "grad_norm": 1.5376960039138794, "learning_rate": 1.0180882508037783e-05, "loss": 0.7167, "step": 14477 }, { "epoch": 0.5101614736118819, "grad_norm": 4.437353610992432, "learning_rate": 1.0179741416779484e-05, "loss": 2.8904, "step": 14478 }, { "epoch": 0.5101967106248403, "grad_norm": 7.983026504516602, "learning_rate": 1.0178600323180028e-05, "loss": 5.4393, "step": 14479 }, { "epoch": 0.5102319476377988, "grad_norm": 1.5923919677734375, "learning_rate": 1.0177459227254283e-05, "loss": 0.8599, "step": 14480 }, { "epoch": 0.5102671846507572, "grad_norm": 8.700422286987305, "learning_rate": 1.017631812901711e-05, "loss": 2.9593, "step": 14481 }, { "epoch": 0.5103024216637155, "grad_norm": 1.6266595125198364, "learning_rate": 1.0175177028483366e-05, "loss": 1.231, "step": 14482 }, { "epoch": 0.510337658676674, "grad_norm": 4.799373626708984, "learning_rate": 1.017403592566792e-05, "loss": 3.3008, "step": 14483 }, { "epoch": 0.5103728956896324, "grad_norm": 1.4388660192489624, "learning_rate": 1.0172894820585635e-05, "loss": 0.7693, "step": 14484 }, { "epoch": 0.5104081327025908, "grad_norm": 7.495140075683594, "learning_rate": 1.0171753713251374e-05, "loss": 4.9512, "step": 14485 }, { "epoch": 0.5104433697155493, "grad_norm": 4.546133518218994, "learning_rate": 1.0170612603679997e-05, "loss": 3.4014, "step": 14486 }, { "epoch": 0.5104786067285076, "grad_norm": 3.4128336906433105, "learning_rate": 1.0169471491886372e-05, "loss": 2.7952, "step": 14487 }, { "epoch": 0.510513843741466, "grad_norm": 6.121387004852295, "learning_rate": 1.0168330377885356e-05, "loss": 3.4819, "step": 14488 }, { "epoch": 0.5105490807544244, "grad_norm": 4.164939880371094, "learning_rate": 1.0167189261691818e-05, "loss": 3.4369, "step": 14489 }, { "epoch": 0.5105843177673829, "grad_norm": 4.36669921875, "learning_rate": 1.0166048143320614e-05, "loss": 4.4997, "step": 14490 }, { "epoch": 0.5106195547803413, "grad_norm": 1.0326820611953735, "learning_rate": 1.0164907022786618e-05, "loss": 0.7046, "step": 14491 }, { "epoch": 0.5106547917932996, "grad_norm": 3.1131906509399414, "learning_rate": 1.0163765900104684e-05, "loss": 3.0374, "step": 14492 }, { "epoch": 0.5106900288062581, "grad_norm": 3.8082151412963867, "learning_rate": 1.0162624775289677e-05, "loss": 2.8802, "step": 14493 }, { "epoch": 0.5107252658192165, "grad_norm": 1.6201869249343872, "learning_rate": 1.0161483648356465e-05, "loss": 0.8925, "step": 14494 }, { "epoch": 0.5107605028321749, "grad_norm": 3.154627561569214, "learning_rate": 1.0160342519319906e-05, "loss": 2.9602, "step": 14495 }, { "epoch": 0.5107957398451334, "grad_norm": 1.1908658742904663, "learning_rate": 1.0159201388194863e-05, "loss": 1.1679, "step": 14496 }, { "epoch": 0.5108309768580918, "grad_norm": 3.458914041519165, "learning_rate": 1.0158060254996207e-05, "loss": 3.2016, "step": 14497 }, { "epoch": 0.5108662138710501, "grad_norm": 1.7788995504379272, "learning_rate": 1.0156919119738791e-05, "loss": 1.0163, "step": 14498 }, { "epoch": 0.5109014508840085, "grad_norm": 1.0740562677383423, "learning_rate": 1.0155777982437487e-05, "loss": 0.7225, "step": 14499 }, { "epoch": 0.510936687896967, "grad_norm": 3.1493799686431885, "learning_rate": 1.0154636843107152e-05, "loss": 3.1396, "step": 14500 }, { "epoch": 0.5109719249099254, "grad_norm": 4.431644916534424, "learning_rate": 1.0153495701762655e-05, "loss": 3.1939, "step": 14501 }, { "epoch": 0.5110071619228838, "grad_norm": 1.7910332679748535, "learning_rate": 1.0152354558418856e-05, "loss": 1.2807, "step": 14502 }, { "epoch": 0.5110423989358422, "grad_norm": 8.130220413208008, "learning_rate": 1.0151213413090617e-05, "loss": 7.5732, "step": 14503 }, { "epoch": 0.5110776359488006, "grad_norm": 4.457345485687256, "learning_rate": 1.0150072265792806e-05, "loss": 3.2759, "step": 14504 }, { "epoch": 0.511112872961759, "grad_norm": 1.2989312410354614, "learning_rate": 1.0148931116540284e-05, "loss": 1.2538, "step": 14505 }, { "epoch": 0.5111481099747175, "grad_norm": 4.138485908508301, "learning_rate": 1.0147789965347917e-05, "loss": 3.4051, "step": 14506 }, { "epoch": 0.5111833469876759, "grad_norm": 2.4538683891296387, "learning_rate": 1.0146648812230566e-05, "loss": 2.9675, "step": 14507 }, { "epoch": 0.5112185840006342, "grad_norm": 3.0948948860168457, "learning_rate": 1.0145507657203098e-05, "loss": 2.8469, "step": 14508 }, { "epoch": 0.5112538210135926, "grad_norm": 1.0907807350158691, "learning_rate": 1.0144366500280367e-05, "loss": 0.7249, "step": 14509 }, { "epoch": 0.5112890580265511, "grad_norm": 5.563296794891357, "learning_rate": 1.0143225341477251e-05, "loss": 7.8039, "step": 14510 }, { "epoch": 0.5113242950395095, "grad_norm": 2.5000734329223633, "learning_rate": 1.0142084180808605e-05, "loss": 1.0194, "step": 14511 }, { "epoch": 0.5113595320524679, "grad_norm": 5.459576606750488, "learning_rate": 1.0140943018289294e-05, "loss": 2.7154, "step": 14512 }, { "epoch": 0.5113947690654264, "grad_norm": 1.589088797569275, "learning_rate": 1.0139801853934182e-05, "loss": 0.9572, "step": 14513 }, { "epoch": 0.5114300060783847, "grad_norm": 3.6332452297210693, "learning_rate": 1.0138660687758131e-05, "loss": 3.015, "step": 14514 }, { "epoch": 0.5114652430913431, "grad_norm": 0.9038736820220947, "learning_rate": 1.013751951977601e-05, "loss": 1.1208, "step": 14515 }, { "epoch": 0.5115004801043016, "grad_norm": 2.818146228790283, "learning_rate": 1.013637835000268e-05, "loss": 2.9799, "step": 14516 }, { "epoch": 0.51153571711726, "grad_norm": 24.035268783569336, "learning_rate": 1.0135237178453002e-05, "loss": 3.2995, "step": 14517 }, { "epoch": 0.5115709541302184, "grad_norm": 1.3265380859375, "learning_rate": 1.0134096005141845e-05, "loss": 1.1378, "step": 14518 }, { "epoch": 0.5116061911431768, "grad_norm": 5.788397312164307, "learning_rate": 1.0132954830084067e-05, "loss": 4.865, "step": 14519 }, { "epoch": 0.5116414281561352, "grad_norm": 5.204978942871094, "learning_rate": 1.0131813653294537e-05, "loss": 2.8835, "step": 14520 }, { "epoch": 0.5116766651690936, "grad_norm": 1.1367661952972412, "learning_rate": 1.013067247478812e-05, "loss": 0.8989, "step": 14521 }, { "epoch": 0.511711902182052, "grad_norm": 4.8464250564575195, "learning_rate": 1.0129531294579675e-05, "loss": 3.2417, "step": 14522 }, { "epoch": 0.5117471391950105, "grad_norm": 6.61510705947876, "learning_rate": 1.0128390112684068e-05, "loss": 5.2604, "step": 14523 }, { "epoch": 0.5117823762079688, "grad_norm": 1.0640536546707153, "learning_rate": 1.0127248929116164e-05, "loss": 0.7621, "step": 14524 }, { "epoch": 0.5118176132209272, "grad_norm": 6.965216159820557, "learning_rate": 1.0126107743890828e-05, "loss": 5.2566, "step": 14525 }, { "epoch": 0.5118528502338857, "grad_norm": 5.632867813110352, "learning_rate": 1.012496655702292e-05, "loss": 3.3318, "step": 14526 }, { "epoch": 0.5118880872468441, "grad_norm": 1.4390825033187866, "learning_rate": 1.012382536852731e-05, "loss": 0.8209, "step": 14527 }, { "epoch": 0.5119233242598025, "grad_norm": 3.059748649597168, "learning_rate": 1.0122684178418854e-05, "loss": 2.9668, "step": 14528 }, { "epoch": 0.511958561272761, "grad_norm": 1.2511183023452759, "learning_rate": 1.0121542986712425e-05, "loss": 1.1332, "step": 14529 }, { "epoch": 0.5119937982857193, "grad_norm": 6.817562580108643, "learning_rate": 1.0120401793422883e-05, "loss": 2.9737, "step": 14530 }, { "epoch": 0.5120290352986777, "grad_norm": 3.6990883350372314, "learning_rate": 1.011926059856509e-05, "loss": 2.7454, "step": 14531 }, { "epoch": 0.5120642723116361, "grad_norm": 4.470005989074707, "learning_rate": 1.0118119402153912e-05, "loss": 2.9572, "step": 14532 }, { "epoch": 0.5120995093245946, "grad_norm": 1.6822365522384644, "learning_rate": 1.0116978204204215e-05, "loss": 1.0517, "step": 14533 }, { "epoch": 0.512134746337553, "grad_norm": 14.024616241455078, "learning_rate": 1.0115837004730864e-05, "loss": 5.226, "step": 14534 }, { "epoch": 0.5121699833505113, "grad_norm": 1.3444935083389282, "learning_rate": 1.011469580374872e-05, "loss": 1.2717, "step": 14535 }, { "epoch": 0.5122052203634698, "grad_norm": 1.0797157287597656, "learning_rate": 1.0113554601272646e-05, "loss": 1.1042, "step": 14536 }, { "epoch": 0.5122404573764282, "grad_norm": 1.6414893865585327, "learning_rate": 1.0112413397317513e-05, "loss": 0.7952, "step": 14537 }, { "epoch": 0.5122756943893866, "grad_norm": 2.8558340072631836, "learning_rate": 1.011127219189818e-05, "loss": 1.7142, "step": 14538 }, { "epoch": 0.5123109314023451, "grad_norm": 8.215206146240234, "learning_rate": 1.011013098502951e-05, "loss": 4.949, "step": 14539 }, { "epoch": 0.5123461684153034, "grad_norm": 0.9586382508277893, "learning_rate": 1.0108989776726373e-05, "loss": 0.7942, "step": 14540 }, { "epoch": 0.5123814054282618, "grad_norm": 15.823570251464844, "learning_rate": 1.010784856700363e-05, "loss": 2.5749, "step": 14541 }, { "epoch": 0.5124166424412202, "grad_norm": 1.5277459621429443, "learning_rate": 1.0106707355876143e-05, "loss": 0.8858, "step": 14542 }, { "epoch": 0.5124518794541787, "grad_norm": 1.413369059562683, "learning_rate": 1.010556614335878e-05, "loss": 0.688, "step": 14543 }, { "epoch": 0.5124871164671371, "grad_norm": 7.200511455535889, "learning_rate": 1.0104424929466405e-05, "loss": 5.5344, "step": 14544 }, { "epoch": 0.5125223534800954, "grad_norm": 0.9840661287307739, "learning_rate": 1.010328371421388e-05, "loss": 0.983, "step": 14545 }, { "epoch": 0.5125575904930539, "grad_norm": 1.1839874982833862, "learning_rate": 1.0102142497616077e-05, "loss": 0.7519, "step": 14546 }, { "epoch": 0.5125928275060123, "grad_norm": 1.2072721719741821, "learning_rate": 1.010100127968785e-05, "loss": 0.9937, "step": 14547 }, { "epoch": 0.5126280645189707, "grad_norm": 2.9217896461486816, "learning_rate": 1.0099860060444071e-05, "loss": 1.128, "step": 14548 }, { "epoch": 0.5126633015319292, "grad_norm": 4.740454196929932, "learning_rate": 1.00987188398996e-05, "loss": 3.0847, "step": 14549 }, { "epoch": 0.5126985385448876, "grad_norm": 1.6538783311843872, "learning_rate": 1.0097577618069305e-05, "loss": 0.9267, "step": 14550 }, { "epoch": 0.5127337755578459, "grad_norm": 3.867110013961792, "learning_rate": 1.009643639496805e-05, "loss": 2.7867, "step": 14551 }, { "epoch": 0.5127690125708044, "grad_norm": 1.6364755630493164, "learning_rate": 1.0095295170610695e-05, "loss": 1.0769, "step": 14552 }, { "epoch": 0.5128042495837628, "grad_norm": 5.566164970397949, "learning_rate": 1.009415394501211e-05, "loss": 3.1366, "step": 14553 }, { "epoch": 0.5128394865967212, "grad_norm": 0.90641850233078, "learning_rate": 1.0093012718187163e-05, "loss": 0.7823, "step": 14554 }, { "epoch": 0.5128747236096796, "grad_norm": 0.9486638903617859, "learning_rate": 1.0091871490150707e-05, "loss": 1.2469, "step": 14555 }, { "epoch": 0.512909960622638, "grad_norm": 2.9839515686035156, "learning_rate": 1.0090730260917617e-05, "loss": 0.8782, "step": 14556 }, { "epoch": 0.5129451976355964, "grad_norm": 0.8333720564842224, "learning_rate": 1.0089589030502751e-05, "loss": 1.0036, "step": 14557 }, { "epoch": 0.5129804346485548, "grad_norm": 54.919288635253906, "learning_rate": 1.0088447798920978e-05, "loss": 5.2568, "step": 14558 }, { "epoch": 0.5130156716615133, "grad_norm": 2.0746397972106934, "learning_rate": 1.0087306566187158e-05, "loss": 0.9327, "step": 14559 }, { "epoch": 0.5130509086744717, "grad_norm": 6.603916645050049, "learning_rate": 1.0086165332316163e-05, "loss": 5.1668, "step": 14560 }, { "epoch": 0.51308614568743, "grad_norm": 4.5995283126831055, "learning_rate": 1.008502409732285e-05, "loss": 2.9754, "step": 14561 }, { "epoch": 0.5131213827003885, "grad_norm": 7.930303573608398, "learning_rate": 1.0083882861222092e-05, "loss": 3.218, "step": 14562 }, { "epoch": 0.5131566197133469, "grad_norm": 6.232983112335205, "learning_rate": 1.0082741624028745e-05, "loss": 3.381, "step": 14563 }, { "epoch": 0.5131918567263053, "grad_norm": 1.7970938682556152, "learning_rate": 1.008160038575768e-05, "loss": 1.0652, "step": 14564 }, { "epoch": 0.5132270937392637, "grad_norm": 1.1208407878875732, "learning_rate": 1.0080459146423757e-05, "loss": 0.8986, "step": 14565 }, { "epoch": 0.5132623307522222, "grad_norm": 3.0964393615722656, "learning_rate": 1.0079317906041845e-05, "loss": 0.7579, "step": 14566 }, { "epoch": 0.5132975677651805, "grad_norm": 3.9669272899627686, "learning_rate": 1.0078176664626808e-05, "loss": 3.372, "step": 14567 }, { "epoch": 0.5133328047781389, "grad_norm": 1.393813133239746, "learning_rate": 1.0077035422193509e-05, "loss": 1.1336, "step": 14568 }, { "epoch": 0.5133680417910974, "grad_norm": 4.421877861022949, "learning_rate": 1.0075894178756813e-05, "loss": 3.4887, "step": 14569 }, { "epoch": 0.5134032788040558, "grad_norm": 1.153394103050232, "learning_rate": 1.0074752934331586e-05, "loss": 0.8106, "step": 14570 }, { "epoch": 0.5134385158170142, "grad_norm": 4.578080654144287, "learning_rate": 1.0073611688932694e-05, "loss": 3.6617, "step": 14571 }, { "epoch": 0.5134737528299727, "grad_norm": 1.237559199333191, "learning_rate": 1.0072470442575e-05, "loss": 0.876, "step": 14572 }, { "epoch": 0.513508989842931, "grad_norm": 5.067732334136963, "learning_rate": 1.0071329195273369e-05, "loss": 5.0587, "step": 14573 }, { "epoch": 0.5135442268558894, "grad_norm": 3.3890724182128906, "learning_rate": 1.0070187947042664e-05, "loss": 3.2336, "step": 14574 }, { "epoch": 0.5135794638688478, "grad_norm": 12.133780479431152, "learning_rate": 1.0069046697897757e-05, "loss": 2.9146, "step": 14575 }, { "epoch": 0.5136147008818063, "grad_norm": 4.191497802734375, "learning_rate": 1.0067905447853505e-05, "loss": 3.1866, "step": 14576 }, { "epoch": 0.5136499378947647, "grad_norm": 7.530580997467041, "learning_rate": 1.0066764196924777e-05, "loss": 3.4473, "step": 14577 }, { "epoch": 0.513685174907723, "grad_norm": 1.0309776067733765, "learning_rate": 1.0065622945126434e-05, "loss": 0.7697, "step": 14578 }, { "epoch": 0.5137204119206815, "grad_norm": 24.791152954101562, "learning_rate": 1.0064481692473344e-05, "loss": 2.9731, "step": 14579 }, { "epoch": 0.5137556489336399, "grad_norm": 5.19220495223999, "learning_rate": 1.0063340438980373e-05, "loss": 3.5569, "step": 14580 }, { "epoch": 0.5137908859465983, "grad_norm": 2.9641335010528564, "learning_rate": 1.0062199184662389e-05, "loss": 2.96, "step": 14581 }, { "epoch": 0.5138261229595568, "grad_norm": 6.022119522094727, "learning_rate": 1.0061057929534246e-05, "loss": 2.959, "step": 14582 }, { "epoch": 0.5138613599725151, "grad_norm": 1.512925148010254, "learning_rate": 1.0059916673610819e-05, "loss": 1.1068, "step": 14583 }, { "epoch": 0.5138965969854735, "grad_norm": 0.9318622350692749, "learning_rate": 1.005877541690697e-05, "loss": 1.1028, "step": 14584 }, { "epoch": 0.513931833998432, "grad_norm": 6.1568121910095215, "learning_rate": 1.0057634159437562e-05, "loss": 3.3301, "step": 14585 }, { "epoch": 0.5139670710113904, "grad_norm": 12.963346481323242, "learning_rate": 1.0056492901217463e-05, "loss": 3.3401, "step": 14586 }, { "epoch": 0.5140023080243488, "grad_norm": 2.0576171875, "learning_rate": 1.0055351642261534e-05, "loss": 1.0044, "step": 14587 }, { "epoch": 0.5140375450373071, "grad_norm": 5.329389572143555, "learning_rate": 1.0054210382584647e-05, "loss": 3.1382, "step": 14588 }, { "epoch": 0.5140727820502656, "grad_norm": 4.92050838470459, "learning_rate": 1.0053069122201659e-05, "loss": 5.4752, "step": 14589 }, { "epoch": 0.514108019063224, "grad_norm": 3.1310713291168213, "learning_rate": 1.0051927861127442e-05, "loss": 3.3212, "step": 14590 }, { "epoch": 0.5141432560761824, "grad_norm": 1.3599708080291748, "learning_rate": 1.0050786599376857e-05, "loss": 1.1321, "step": 14591 }, { "epoch": 0.5141784930891409, "grad_norm": 1.7010959386825562, "learning_rate": 1.0049645336964771e-05, "loss": 0.7561, "step": 14592 }, { "epoch": 0.5142137301020993, "grad_norm": 4.96511697769165, "learning_rate": 1.0048504073906045e-05, "loss": 3.1116, "step": 14593 }, { "epoch": 0.5142489671150576, "grad_norm": 1.6542346477508545, "learning_rate": 1.0047362810215551e-05, "loss": 0.9239, "step": 14594 }, { "epoch": 0.5142842041280161, "grad_norm": 22.131839752197266, "learning_rate": 1.0046221545908147e-05, "loss": 7.6887, "step": 14595 }, { "epoch": 0.5143194411409745, "grad_norm": 0.8784547448158264, "learning_rate": 1.0045080280998705e-05, "loss": 0.8704, "step": 14596 }, { "epoch": 0.5143546781539329, "grad_norm": 3.8360066413879395, "learning_rate": 1.0043939015502084e-05, "loss": 3.338, "step": 14597 }, { "epoch": 0.5143899151668913, "grad_norm": 0.998275101184845, "learning_rate": 1.004279774943315e-05, "loss": 1.0064, "step": 14598 }, { "epoch": 0.5144251521798497, "grad_norm": 1.1563974618911743, "learning_rate": 1.0041656482806775e-05, "loss": 0.9041, "step": 14599 }, { "epoch": 0.5144603891928081, "grad_norm": 10.65659236907959, "learning_rate": 1.0040515215637818e-05, "loss": 5.2351, "step": 14600 }, { "epoch": 0.5144956262057665, "grad_norm": 12.1597318649292, "learning_rate": 1.003937394794114e-05, "loss": 2.5681, "step": 14601 }, { "epoch": 0.514530863218725, "grad_norm": 3.593146800994873, "learning_rate": 1.0038232679731617e-05, "loss": 2.4922, "step": 14602 }, { "epoch": 0.5145661002316834, "grad_norm": 7.353244304656982, "learning_rate": 1.0037091411024108e-05, "loss": 5.4896, "step": 14603 }, { "epoch": 0.5146013372446417, "grad_norm": 1.131817102432251, "learning_rate": 1.0035950141833475e-05, "loss": 1.0856, "step": 14604 }, { "epoch": 0.5146365742576002, "grad_norm": 4.783444881439209, "learning_rate": 1.0034808872174589e-05, "loss": 2.9564, "step": 14605 }, { "epoch": 0.5146718112705586, "grad_norm": 1.371108889579773, "learning_rate": 1.0033667602062311e-05, "loss": 1.1232, "step": 14606 }, { "epoch": 0.514707048283517, "grad_norm": 6.65207576751709, "learning_rate": 1.0032526331511511e-05, "loss": 4.8159, "step": 14607 }, { "epoch": 0.5147422852964754, "grad_norm": 1.1884181499481201, "learning_rate": 1.003138506053705e-05, "loss": 0.8731, "step": 14608 }, { "epoch": 0.5147775223094339, "grad_norm": 2.056790351867676, "learning_rate": 1.0030243789153795e-05, "loss": 0.9542, "step": 14609 }, { "epoch": 0.5148127593223922, "grad_norm": 4.954875469207764, "learning_rate": 1.002910251737661e-05, "loss": 4.7048, "step": 14610 }, { "epoch": 0.5148479963353506, "grad_norm": 6.71278190612793, "learning_rate": 1.0027961245220363e-05, "loss": 6.5523, "step": 14611 }, { "epoch": 0.5148832333483091, "grad_norm": 1.3520783185958862, "learning_rate": 1.0026819972699915e-05, "loss": 0.6707, "step": 14612 }, { "epoch": 0.5149184703612675, "grad_norm": 3.496581554412842, "learning_rate": 1.0025678699830135e-05, "loss": 3.046, "step": 14613 }, { "epoch": 0.5149537073742259, "grad_norm": 1.2909129858016968, "learning_rate": 1.0024537426625884e-05, "loss": 0.8102, "step": 14614 }, { "epoch": 0.5149889443871843, "grad_norm": 0.9143957495689392, "learning_rate": 1.0023396153102034e-05, "loss": 0.7742, "step": 14615 }, { "epoch": 0.5150241814001427, "grad_norm": 0.9194294214248657, "learning_rate": 1.0022254879273443e-05, "loss": 0.7793, "step": 14616 }, { "epoch": 0.5150594184131011, "grad_norm": 1.9897403717041016, "learning_rate": 1.0021113605154981e-05, "loss": 1.1859, "step": 14617 }, { "epoch": 0.5150946554260595, "grad_norm": 5.213234901428223, "learning_rate": 1.0019972330761511e-05, "loss": 4.658, "step": 14618 }, { "epoch": 0.515129892439018, "grad_norm": 4.666385173797607, "learning_rate": 1.00188310561079e-05, "loss": 4.8054, "step": 14619 }, { "epoch": 0.5151651294519763, "grad_norm": 2.0714845657348633, "learning_rate": 1.001768978120901e-05, "loss": 1.022, "step": 14620 }, { "epoch": 0.5152003664649347, "grad_norm": 1.2683765888214111, "learning_rate": 1.001654850607971e-05, "loss": 0.9733, "step": 14621 }, { "epoch": 0.5152356034778932, "grad_norm": 3.686521053314209, "learning_rate": 1.0015407230734861e-05, "loss": 2.8983, "step": 14622 }, { "epoch": 0.5152708404908516, "grad_norm": 4.561354637145996, "learning_rate": 1.0014265955189335e-05, "loss": 3.1232, "step": 14623 }, { "epoch": 0.51530607750381, "grad_norm": 3.8029162883758545, "learning_rate": 1.001312467945799e-05, "loss": 3.3271, "step": 14624 }, { "epoch": 0.5153413145167685, "grad_norm": 2.6366629600524902, "learning_rate": 1.0011983403555697e-05, "loss": 1.0637, "step": 14625 }, { "epoch": 0.5153765515297268, "grad_norm": 5.139429092407227, "learning_rate": 1.0010842127497316e-05, "loss": 3.3266, "step": 14626 }, { "epoch": 0.5154117885426852, "grad_norm": 5.685807704925537, "learning_rate": 1.0009700851297714e-05, "loss": 3.3809, "step": 14627 }, { "epoch": 0.5154470255556437, "grad_norm": 2.4092071056365967, "learning_rate": 1.0008559574971759e-05, "loss": 2.1436, "step": 14628 }, { "epoch": 0.5154822625686021, "grad_norm": 0.9823794364929199, "learning_rate": 1.0007418298534315e-05, "loss": 0.8689, "step": 14629 }, { "epoch": 0.5155174995815605, "grad_norm": 10.004326820373535, "learning_rate": 1.0006277022000247e-05, "loss": 5.4708, "step": 14630 }, { "epoch": 0.5155527365945188, "grad_norm": 3.643258810043335, "learning_rate": 1.0005135745384417e-05, "loss": 2.8596, "step": 14631 }, { "epoch": 0.5155879736074773, "grad_norm": 2.754178524017334, "learning_rate": 1.0003994468701698e-05, "loss": 0.7545, "step": 14632 }, { "epoch": 0.5156232106204357, "grad_norm": 4.518738746643066, "learning_rate": 1.0002853191966946e-05, "loss": 2.6771, "step": 14633 }, { "epoch": 0.5156584476333941, "grad_norm": 1.249528169631958, "learning_rate": 1.0001711915195033e-05, "loss": 1.1877, "step": 14634 }, { "epoch": 0.5156936846463526, "grad_norm": 4.7086052894592285, "learning_rate": 1.0000570638400824e-05, "loss": 2.9806, "step": 14635 }, { "epoch": 0.515728921659311, "grad_norm": 1.5092058181762695, "learning_rate": 9.999429361599179e-06, "loss": 0.7403, "step": 14636 }, { "epoch": 0.5157641586722693, "grad_norm": 2.085799217224121, "learning_rate": 9.998288084804965e-06, "loss": 0.8262, "step": 14637 }, { "epoch": 0.5157993956852278, "grad_norm": 6.493227958679199, "learning_rate": 9.997146808033057e-06, "loss": 3.1146, "step": 14638 }, { "epoch": 0.5158346326981862, "grad_norm": 4.988609313964844, "learning_rate": 9.996005531298303e-06, "loss": 4.3661, "step": 14639 }, { "epoch": 0.5158698697111446, "grad_norm": 3.800037384033203, "learning_rate": 9.994864254615583e-06, "loss": 3.0594, "step": 14640 }, { "epoch": 0.515905106724103, "grad_norm": 2.045478582382202, "learning_rate": 9.993722977999758e-06, "loss": 1.0078, "step": 14641 }, { "epoch": 0.5159403437370614, "grad_norm": 4.126773357391357, "learning_rate": 9.992581701465688e-06, "loss": 2.6147, "step": 14642 }, { "epoch": 0.5159755807500198, "grad_norm": 1.1590924263000488, "learning_rate": 9.991440425028241e-06, "loss": 0.7544, "step": 14643 }, { "epoch": 0.5160108177629782, "grad_norm": 1.2887215614318848, "learning_rate": 9.990299148702287e-06, "loss": 0.923, "step": 14644 }, { "epoch": 0.5160460547759367, "grad_norm": 5.200810432434082, "learning_rate": 9.989157872502687e-06, "loss": 3.1945, "step": 14645 }, { "epoch": 0.5160812917888951, "grad_norm": 4.29536771774292, "learning_rate": 9.988016596444306e-06, "loss": 2.9383, "step": 14646 }, { "epoch": 0.5161165288018534, "grad_norm": 1.1033170223236084, "learning_rate": 9.986875320542015e-06, "loss": 0.8562, "step": 14647 }, { "epoch": 0.5161517658148119, "grad_norm": 4.659749507904053, "learning_rate": 9.985734044810668e-06, "loss": 3.3782, "step": 14648 }, { "epoch": 0.5161870028277703, "grad_norm": 1.3302454948425293, "learning_rate": 9.984592769265139e-06, "loss": 0.8553, "step": 14649 }, { "epoch": 0.5162222398407287, "grad_norm": 9.155915260314941, "learning_rate": 9.983451493920295e-06, "loss": 5.7532, "step": 14650 }, { "epoch": 0.5162574768536871, "grad_norm": 4.711287498474121, "learning_rate": 9.982310218790994e-06, "loss": 3.1058, "step": 14651 }, { "epoch": 0.5162927138666455, "grad_norm": 5.30338716506958, "learning_rate": 9.981168943892103e-06, "loss": 3.005, "step": 14652 }, { "epoch": 0.5163279508796039, "grad_norm": 1.241398572921753, "learning_rate": 9.980027669238494e-06, "loss": 0.8454, "step": 14653 }, { "epoch": 0.5163631878925623, "grad_norm": 6.6912760734558105, "learning_rate": 9.978886394845022e-06, "loss": 5.3076, "step": 14654 }, { "epoch": 0.5163984249055208, "grad_norm": 1.6304161548614502, "learning_rate": 9.977745120726558e-06, "loss": 0.8853, "step": 14655 }, { "epoch": 0.5164336619184792, "grad_norm": 1.1622035503387451, "learning_rate": 9.976603846897966e-06, "loss": 0.9359, "step": 14656 }, { "epoch": 0.5164688989314375, "grad_norm": 1.2062132358551025, "learning_rate": 9.975462573374118e-06, "loss": 0.7985, "step": 14657 }, { "epoch": 0.516504135944396, "grad_norm": 1.808391809463501, "learning_rate": 9.974321300169867e-06, "loss": 0.9547, "step": 14658 }, { "epoch": 0.5165393729573544, "grad_norm": 0.7750352621078491, "learning_rate": 9.973180027300087e-06, "loss": 0.7152, "step": 14659 }, { "epoch": 0.5165746099703128, "grad_norm": 1.1209484338760376, "learning_rate": 9.972038754779642e-06, "loss": 1.2607, "step": 14660 }, { "epoch": 0.5166098469832713, "grad_norm": 3.5263466835021973, "learning_rate": 9.970897482623393e-06, "loss": 2.6282, "step": 14661 }, { "epoch": 0.5166450839962297, "grad_norm": 3.8104605674743652, "learning_rate": 9.969756210846206e-06, "loss": 3.1863, "step": 14662 }, { "epoch": 0.516680321009188, "grad_norm": 1.5510481595993042, "learning_rate": 9.968614939462952e-06, "loss": 1.1671, "step": 14663 }, { "epoch": 0.5167155580221464, "grad_norm": 1.1601982116699219, "learning_rate": 9.967473668488492e-06, "loss": 1.1028, "step": 14664 }, { "epoch": 0.5167507950351049, "grad_norm": 3.3932337760925293, "learning_rate": 9.966332397937692e-06, "loss": 3.1539, "step": 14665 }, { "epoch": 0.5167860320480633, "grad_norm": 6.394647121429443, "learning_rate": 9.965191127825416e-06, "loss": 5.8145, "step": 14666 }, { "epoch": 0.5168212690610217, "grad_norm": 1.8819828033447266, "learning_rate": 9.964049858166528e-06, "loss": 1.0608, "step": 14667 }, { "epoch": 0.5168565060739801, "grad_norm": 4.020730018615723, "learning_rate": 9.962908588975897e-06, "loss": 3.1948, "step": 14668 }, { "epoch": 0.5168917430869385, "grad_norm": 1.6613162755966187, "learning_rate": 9.961767320268388e-06, "loss": 0.9298, "step": 14669 }, { "epoch": 0.5169269800998969, "grad_norm": 1.1897398233413696, "learning_rate": 9.960626052058862e-06, "loss": 0.7075, "step": 14670 }, { "epoch": 0.5169622171128554, "grad_norm": 4.459661483764648, "learning_rate": 9.959484784362185e-06, "loss": 3.1554, "step": 14671 }, { "epoch": 0.5169974541258138, "grad_norm": 1.4517154693603516, "learning_rate": 9.958343517193227e-06, "loss": 0.7736, "step": 14672 }, { "epoch": 0.5170326911387721, "grad_norm": 1.560928463935852, "learning_rate": 9.957202250566851e-06, "loss": 0.9846, "step": 14673 }, { "epoch": 0.5170679281517305, "grad_norm": 1.1672112941741943, "learning_rate": 9.95606098449792e-06, "loss": 0.9119, "step": 14674 }, { "epoch": 0.517103165164689, "grad_norm": 6.379639625549316, "learning_rate": 9.954919719001297e-06, "loss": 3.1869, "step": 14675 }, { "epoch": 0.5171384021776474, "grad_norm": 1.4002183675765991, "learning_rate": 9.953778454091857e-06, "loss": 0.8694, "step": 14676 }, { "epoch": 0.5171736391906058, "grad_norm": 1.5805178880691528, "learning_rate": 9.952637189784452e-06, "loss": 0.7645, "step": 14677 }, { "epoch": 0.5172088762035643, "grad_norm": 1.2927460670471191, "learning_rate": 9.951495926093956e-06, "loss": 0.8127, "step": 14678 }, { "epoch": 0.5172441132165226, "grad_norm": 2.5851786136627197, "learning_rate": 9.950354663035234e-06, "loss": 0.7653, "step": 14679 }, { "epoch": 0.517279350229481, "grad_norm": 1.612611174583435, "learning_rate": 9.949213400623147e-06, "loss": 0.8013, "step": 14680 }, { "epoch": 0.5173145872424395, "grad_norm": 5.07456111907959, "learning_rate": 9.94807213887256e-06, "loss": 3.328, "step": 14681 }, { "epoch": 0.5173498242553979, "grad_norm": 4.149472713470459, "learning_rate": 9.946930877798343e-06, "loss": 3.2608, "step": 14682 }, { "epoch": 0.5173850612683563, "grad_norm": 3.2836263179779053, "learning_rate": 9.945789617415357e-06, "loss": 3.3755, "step": 14683 }, { "epoch": 0.5174202982813146, "grad_norm": 4.1508612632751465, "learning_rate": 9.944648357738467e-06, "loss": 2.6911, "step": 14684 }, { "epoch": 0.5174555352942731, "grad_norm": 4.1508612632751465, "learning_rate": 9.944648357738467e-06, "loss": 3.0446, "step": 14685 }, { "epoch": 0.5174907723072315, "grad_norm": 1.7448135614395142, "learning_rate": 9.943507098782544e-06, "loss": 0.8494, "step": 14686 }, { "epoch": 0.5175260093201899, "grad_norm": 9.149195671081543, "learning_rate": 9.942365840562441e-06, "loss": 3.4687, "step": 14687 }, { "epoch": 0.5175612463331484, "grad_norm": 4.035409450531006, "learning_rate": 9.941224583093034e-06, "loss": 3.2266, "step": 14688 }, { "epoch": 0.5175964833461068, "grad_norm": 14.020915031433105, "learning_rate": 9.940083326389184e-06, "loss": 3.1691, "step": 14689 }, { "epoch": 0.5176317203590651, "grad_norm": 1.7268108129501343, "learning_rate": 9.938942070465757e-06, "loss": 1.0969, "step": 14690 }, { "epoch": 0.5176669573720236, "grad_norm": 1.6892485618591309, "learning_rate": 9.937800815337615e-06, "loss": 0.8222, "step": 14691 }, { "epoch": 0.517702194384982, "grad_norm": 3.4912068843841553, "learning_rate": 9.936659561019627e-06, "loss": 2.4626, "step": 14692 }, { "epoch": 0.5177374313979404, "grad_norm": 4.4312896728515625, "learning_rate": 9.935518307526658e-06, "loss": 3.0901, "step": 14693 }, { "epoch": 0.5177726684108989, "grad_norm": 5.870833396911621, "learning_rate": 9.934377054873569e-06, "loss": 7.6171, "step": 14694 }, { "epoch": 0.5178079054238572, "grad_norm": 6.516109943389893, "learning_rate": 9.933235803075225e-06, "loss": 3.5973, "step": 14695 }, { "epoch": 0.5178431424368156, "grad_norm": 4.052240371704102, "learning_rate": 9.9320945521465e-06, "loss": 3.609, "step": 14696 }, { "epoch": 0.517878379449774, "grad_norm": 1.1181737184524536, "learning_rate": 9.930953302102246e-06, "loss": 1.072, "step": 14697 }, { "epoch": 0.5179136164627325, "grad_norm": 3.2637202739715576, "learning_rate": 9.929812052957336e-06, "loss": 2.5416, "step": 14698 }, { "epoch": 0.5179488534756909, "grad_norm": 3.779839515686035, "learning_rate": 9.928670804726635e-06, "loss": 3.0918, "step": 14699 }, { "epoch": 0.5179840904886492, "grad_norm": 4.795876979827881, "learning_rate": 9.927529557425004e-06, "loss": 3.5222, "step": 14700 }, { "epoch": 0.5180193275016077, "grad_norm": 3.1999645233154297, "learning_rate": 9.926388311067306e-06, "loss": 3.0757, "step": 14701 }, { "epoch": 0.5180545645145661, "grad_norm": 5.804141521453857, "learning_rate": 9.925247065668416e-06, "loss": 5.0573, "step": 14702 }, { "epoch": 0.5180898015275245, "grad_norm": 1.9063169956207275, "learning_rate": 9.924105821243188e-06, "loss": 0.916, "step": 14703 }, { "epoch": 0.518125038540483, "grad_norm": 3.9873883724212646, "learning_rate": 9.922964577806494e-06, "loss": 3.2506, "step": 14704 }, { "epoch": 0.5181602755534414, "grad_norm": 1.1208751201629639, "learning_rate": 9.921823335373197e-06, "loss": 0.7336, "step": 14705 }, { "epoch": 0.5181955125663997, "grad_norm": 0.9703019857406616, "learning_rate": 9.920682093958157e-06, "loss": 0.9296, "step": 14706 }, { "epoch": 0.5182307495793581, "grad_norm": 1.5974537134170532, "learning_rate": 9.919540853576244e-06, "loss": 0.8329, "step": 14707 }, { "epoch": 0.5182659865923166, "grad_norm": 14.377031326293945, "learning_rate": 9.918399614242325e-06, "loss": 3.9086, "step": 14708 }, { "epoch": 0.518301223605275, "grad_norm": 3.249938488006592, "learning_rate": 9.91725837597126e-06, "loss": 2.8274, "step": 14709 }, { "epoch": 0.5183364606182334, "grad_norm": 4.253471851348877, "learning_rate": 9.916117138777912e-06, "loss": 3.0386, "step": 14710 }, { "epoch": 0.5183716976311918, "grad_norm": 1.6921619176864624, "learning_rate": 9.91497590267715e-06, "loss": 0.829, "step": 14711 }, { "epoch": 0.5184069346441502, "grad_norm": 1.408216118812561, "learning_rate": 9.913834667683842e-06, "loss": 0.8669, "step": 14712 }, { "epoch": 0.5184421716571086, "grad_norm": 6.048059463500977, "learning_rate": 9.912693433812845e-06, "loss": 3.3805, "step": 14713 }, { "epoch": 0.5184774086700671, "grad_norm": 1.2323601245880127, "learning_rate": 9.911552201079027e-06, "loss": 0.9699, "step": 14714 }, { "epoch": 0.5185126456830255, "grad_norm": 1.4471040964126587, "learning_rate": 9.910410969497255e-06, "loss": 0.7339, "step": 14715 }, { "epoch": 0.5185478826959838, "grad_norm": 2.5020692348480225, "learning_rate": 9.909269739082388e-06, "loss": 0.8271, "step": 14716 }, { "epoch": 0.5185831197089422, "grad_norm": 4.539916038513184, "learning_rate": 9.908128509849295e-06, "loss": 3.0227, "step": 14717 }, { "epoch": 0.5186183567219007, "grad_norm": 1.4257354736328125, "learning_rate": 9.906987281812843e-06, "loss": 0.84, "step": 14718 }, { "epoch": 0.5186535937348591, "grad_norm": 2.2145509719848633, "learning_rate": 9.905846054987892e-06, "loss": 1.0796, "step": 14719 }, { "epoch": 0.5186888307478175, "grad_norm": 3.2497973442077637, "learning_rate": 9.904704829389305e-06, "loss": 2.6521, "step": 14720 }, { "epoch": 0.518724067760776, "grad_norm": 8.00011157989502, "learning_rate": 9.903563605031955e-06, "loss": 5.456, "step": 14721 }, { "epoch": 0.5187593047737343, "grad_norm": 0.981857180595398, "learning_rate": 9.902422381930697e-06, "loss": 0.8242, "step": 14722 }, { "epoch": 0.5187945417866927, "grad_norm": 8.016616821289062, "learning_rate": 9.901281160100402e-06, "loss": 5.4599, "step": 14723 }, { "epoch": 0.5188297787996512, "grad_norm": 4.784143924713135, "learning_rate": 9.900139939555934e-06, "loss": 3.0055, "step": 14724 }, { "epoch": 0.5188650158126096, "grad_norm": 1.2275018692016602, "learning_rate": 9.898998720312154e-06, "loss": 1.0276, "step": 14725 }, { "epoch": 0.518900252825568, "grad_norm": 4.9661784172058105, "learning_rate": 9.897857502383926e-06, "loss": 3.2793, "step": 14726 }, { "epoch": 0.5189354898385264, "grad_norm": 3.6536335945129395, "learning_rate": 9.896716285786121e-06, "loss": 3.4008, "step": 14727 }, { "epoch": 0.5189707268514848, "grad_norm": 1.6478146314620972, "learning_rate": 9.895575070533598e-06, "loss": 0.9927, "step": 14728 }, { "epoch": 0.5190059638644432, "grad_norm": 5.028219223022461, "learning_rate": 9.894433856641223e-06, "loss": 3.2603, "step": 14729 }, { "epoch": 0.5190412008774016, "grad_norm": 5.9693403244018555, "learning_rate": 9.893292644123859e-06, "loss": 2.9354, "step": 14730 }, { "epoch": 0.5190764378903601, "grad_norm": 3.7719380855560303, "learning_rate": 9.892151432996376e-06, "loss": 2.6496, "step": 14731 }, { "epoch": 0.5191116749033184, "grad_norm": 4.916289329528809, "learning_rate": 9.89101022327363e-06, "loss": 4.6469, "step": 14732 }, { "epoch": 0.5191469119162768, "grad_norm": 4.8657636642456055, "learning_rate": 9.889869014970493e-06, "loss": 2.803, "step": 14733 }, { "epoch": 0.5191821489292353, "grad_norm": 5.004621982574463, "learning_rate": 9.888727808101825e-06, "loss": 3.1159, "step": 14734 }, { "epoch": 0.5192173859421937, "grad_norm": 10.528779983520508, "learning_rate": 9.88758660268249e-06, "loss": 7.2182, "step": 14735 }, { "epoch": 0.5192526229551521, "grad_norm": 7.1805195808410645, "learning_rate": 9.886445398727355e-06, "loss": 3.3865, "step": 14736 }, { "epoch": 0.5192878599681106, "grad_norm": 1.269963026046753, "learning_rate": 9.885304196251284e-06, "loss": 0.9326, "step": 14737 }, { "epoch": 0.5193230969810689, "grad_norm": 1.274635910987854, "learning_rate": 9.88416299526914e-06, "loss": 0.9764, "step": 14738 }, { "epoch": 0.5193583339940273, "grad_norm": 6.914662837982178, "learning_rate": 9.883021795795785e-06, "loss": 4.8747, "step": 14739 }, { "epoch": 0.5193935710069857, "grad_norm": 1.2162922620773315, "learning_rate": 9.881880597846091e-06, "loss": 1.0791, "step": 14740 }, { "epoch": 0.5194288080199442, "grad_norm": 5.695186138153076, "learning_rate": 9.880739401434913e-06, "loss": 3.021, "step": 14741 }, { "epoch": 0.5194640450329026, "grad_norm": 1.2248401641845703, "learning_rate": 9.879598206577122e-06, "loss": 0.8062, "step": 14742 }, { "epoch": 0.5194992820458609, "grad_norm": 4.777493476867676, "learning_rate": 9.87845701328758e-06, "loss": 5.4468, "step": 14743 }, { "epoch": 0.5195345190588194, "grad_norm": 1.972880244255066, "learning_rate": 9.87731582158115e-06, "loss": 1.0058, "step": 14744 }, { "epoch": 0.5195697560717778, "grad_norm": 1.2631183862686157, "learning_rate": 9.876174631472694e-06, "loss": 0.8764, "step": 14745 }, { "epoch": 0.5196049930847362, "grad_norm": 1.123611330986023, "learning_rate": 9.87503344297708e-06, "loss": 1.0316, "step": 14746 }, { "epoch": 0.5196402300976947, "grad_norm": 11.497624397277832, "learning_rate": 9.873892256109177e-06, "loss": 5.3758, "step": 14747 }, { "epoch": 0.519675467110653, "grad_norm": 5.145022392272949, "learning_rate": 9.87275107088384e-06, "loss": 3.3054, "step": 14748 }, { "epoch": 0.5197107041236114, "grad_norm": 8.951851844787598, "learning_rate": 9.871609887315932e-06, "loss": 3.3654, "step": 14749 }, { "epoch": 0.5197459411365698, "grad_norm": 1.4608699083328247, "learning_rate": 9.870468705420328e-06, "loss": 0.8806, "step": 14750 }, { "epoch": 0.5197811781495283, "grad_norm": 3.1652064323425293, "learning_rate": 9.869327525211884e-06, "loss": 2.636, "step": 14751 }, { "epoch": 0.5198164151624867, "grad_norm": 5.61535120010376, "learning_rate": 9.868186346705464e-06, "loss": 3.2329, "step": 14752 }, { "epoch": 0.519851652175445, "grad_norm": 1.3997613191604614, "learning_rate": 9.867045169915936e-06, "loss": 1.0084, "step": 14753 }, { "epoch": 0.5198868891884035, "grad_norm": 1.0975104570388794, "learning_rate": 9.865903994858158e-06, "loss": 0.931, "step": 14754 }, { "epoch": 0.5199221262013619, "grad_norm": 6.428347587585449, "learning_rate": 9.864762821547e-06, "loss": 3.101, "step": 14755 }, { "epoch": 0.5199573632143203, "grad_norm": 1.396970272064209, "learning_rate": 9.863621649997326e-06, "loss": 0.9167, "step": 14756 }, { "epoch": 0.5199926002272788, "grad_norm": 4.3465986251831055, "learning_rate": 9.862480480223994e-06, "loss": 3.0315, "step": 14757 }, { "epoch": 0.5200278372402372, "grad_norm": 5.217595100402832, "learning_rate": 9.861339312241869e-06, "loss": 5.0597, "step": 14758 }, { "epoch": 0.5200630742531955, "grad_norm": 6.2993927001953125, "learning_rate": 9.860198146065825e-06, "loss": 3.455, "step": 14759 }, { "epoch": 0.520098311266154, "grad_norm": 4.075021266937256, "learning_rate": 9.85905698171071e-06, "loss": 2.7779, "step": 14760 }, { "epoch": 0.5201335482791124, "grad_norm": 1.633485198020935, "learning_rate": 9.857915819191398e-06, "loss": 0.8569, "step": 14761 }, { "epoch": 0.5201687852920708, "grad_norm": 1.364939570426941, "learning_rate": 9.856774658522754e-06, "loss": 0.7226, "step": 14762 }, { "epoch": 0.5202040223050292, "grad_norm": 3.395159959793091, "learning_rate": 9.855633499719635e-06, "loss": 3.036, "step": 14763 }, { "epoch": 0.5202392593179876, "grad_norm": 2.8080101013183594, "learning_rate": 9.854492342796907e-06, "loss": 0.7811, "step": 14764 }, { "epoch": 0.520274496330946, "grad_norm": 6.853890895843506, "learning_rate": 9.853351187769436e-06, "loss": 2.8973, "step": 14765 }, { "epoch": 0.5203097333439044, "grad_norm": 1.4728657007217407, "learning_rate": 9.852210034652087e-06, "loss": 0.8271, "step": 14766 }, { "epoch": 0.5203449703568629, "grad_norm": 1.3770763874053955, "learning_rate": 9.851068883459718e-06, "loss": 1.0567, "step": 14767 }, { "epoch": 0.5203802073698213, "grad_norm": 13.762791633605957, "learning_rate": 9.849927734207194e-06, "loss": 2.7118, "step": 14768 }, { "epoch": 0.5204154443827796, "grad_norm": 1.0443620681762695, "learning_rate": 9.848786586909386e-06, "loss": 0.6993, "step": 14769 }, { "epoch": 0.5204506813957381, "grad_norm": 5.517846584320068, "learning_rate": 9.847645441581149e-06, "loss": 4.9874, "step": 14770 }, { "epoch": 0.5204859184086965, "grad_norm": 1.2110469341278076, "learning_rate": 9.846504298237349e-06, "loss": 0.8173, "step": 14771 }, { "epoch": 0.5205211554216549, "grad_norm": 4.6894707679748535, "learning_rate": 9.845363156892853e-06, "loss": 3.1865, "step": 14772 }, { "epoch": 0.5205563924346133, "grad_norm": 1.5858256816864014, "learning_rate": 9.844222017562517e-06, "loss": 0.7923, "step": 14773 }, { "epoch": 0.5205916294475718, "grad_norm": 1.5360671281814575, "learning_rate": 9.84308088026121e-06, "loss": 0.8091, "step": 14774 }, { "epoch": 0.5206268664605301, "grad_norm": 4.021053314208984, "learning_rate": 9.841939745003798e-06, "loss": 3.4169, "step": 14775 }, { "epoch": 0.5206621034734885, "grad_norm": 1.39369797706604, "learning_rate": 9.840798611805139e-06, "loss": 0.8111, "step": 14776 }, { "epoch": 0.520697340486447, "grad_norm": 0.9626370668411255, "learning_rate": 9.839657480680095e-06, "loss": 1.1234, "step": 14777 }, { "epoch": 0.5207325774994054, "grad_norm": 3.118251323699951, "learning_rate": 9.83851635164354e-06, "loss": 0.9949, "step": 14778 }, { "epoch": 0.5207678145123638, "grad_norm": 7.0167365074157715, "learning_rate": 9.837375224710324e-06, "loss": 3.2874, "step": 14779 }, { "epoch": 0.5208030515253222, "grad_norm": 5.122473239898682, "learning_rate": 9.836234099895319e-06, "loss": 3.2929, "step": 14780 }, { "epoch": 0.5208382885382806, "grad_norm": 6.544384479522705, "learning_rate": 9.835092977213387e-06, "loss": 2.9832, "step": 14781 }, { "epoch": 0.520873525551239, "grad_norm": 2.0019025802612305, "learning_rate": 9.833951856679387e-06, "loss": 0.817, "step": 14782 }, { "epoch": 0.5209087625641974, "grad_norm": 3.9802141189575195, "learning_rate": 9.832810738308185e-06, "loss": 3.2039, "step": 14783 }, { "epoch": 0.5209439995771559, "grad_norm": 2.4661099910736084, "learning_rate": 9.831669622114645e-06, "loss": 0.935, "step": 14784 }, { "epoch": 0.5209792365901142, "grad_norm": 1.4998165369033813, "learning_rate": 9.830528508113632e-06, "loss": 0.7789, "step": 14785 }, { "epoch": 0.5210144736030726, "grad_norm": 2.9456002712249756, "learning_rate": 9.829387396320005e-06, "loss": 1.0751, "step": 14786 }, { "epoch": 0.5210497106160311, "grad_norm": 1.0555698871612549, "learning_rate": 9.828246286748628e-06, "loss": 1.0298, "step": 14787 }, { "epoch": 0.5210849476289895, "grad_norm": 1.152645468711853, "learning_rate": 9.827105179414367e-06, "loss": 0.6537, "step": 14788 }, { "epoch": 0.5211201846419479, "grad_norm": 7.727374076843262, "learning_rate": 9.825964074332081e-06, "loss": 2.7926, "step": 14789 }, { "epoch": 0.5211554216549064, "grad_norm": 2.5503342151641846, "learning_rate": 9.824822971516636e-06, "loss": 1.1413, "step": 14790 }, { "epoch": 0.5211906586678647, "grad_norm": 5.628607273101807, "learning_rate": 9.823681870982897e-06, "loss": 2.7389, "step": 14791 }, { "epoch": 0.5212258956808231, "grad_norm": 1.0047835111618042, "learning_rate": 9.822540772745719e-06, "loss": 0.8495, "step": 14792 }, { "epoch": 0.5212611326937816, "grad_norm": 2.602527379989624, "learning_rate": 9.821399676819973e-06, "loss": 0.9384, "step": 14793 }, { "epoch": 0.52129636970674, "grad_norm": 6.888021945953369, "learning_rate": 9.820258583220519e-06, "loss": 2.9993, "step": 14794 }, { "epoch": 0.5213316067196984, "grad_norm": 1.064587116241455, "learning_rate": 9.81911749196222e-06, "loss": 0.7933, "step": 14795 }, { "epoch": 0.5213668437326567, "grad_norm": 1.4654629230499268, "learning_rate": 9.817976403059936e-06, "loss": 1.2037, "step": 14796 }, { "epoch": 0.5214020807456152, "grad_norm": 6.0565924644470215, "learning_rate": 9.816835316528538e-06, "loss": 4.8793, "step": 14797 }, { "epoch": 0.5214373177585736, "grad_norm": 1.070459246635437, "learning_rate": 9.815694232382878e-06, "loss": 1.1903, "step": 14798 }, { "epoch": 0.521472554771532, "grad_norm": 6.462881565093994, "learning_rate": 9.814553150637826e-06, "loss": 2.8999, "step": 14799 }, { "epoch": 0.5215077917844905, "grad_norm": 1.2959444522857666, "learning_rate": 9.813412071308244e-06, "loss": 1.0748, "step": 14800 }, { "epoch": 0.5215430287974488, "grad_norm": 1.3972461223602295, "learning_rate": 9.812270994408993e-06, "loss": 0.7356, "step": 14801 }, { "epoch": 0.5215782658104072, "grad_norm": 4.831198215484619, "learning_rate": 9.811129919954931e-06, "loss": 5.6185, "step": 14802 }, { "epoch": 0.5216135028233657, "grad_norm": 1.6384371519088745, "learning_rate": 9.809988847960931e-06, "loss": 1.0009, "step": 14803 }, { "epoch": 0.5216487398363241, "grad_norm": 6.649521350860596, "learning_rate": 9.808847778441853e-06, "loss": 3.2265, "step": 14804 }, { "epoch": 0.5216839768492825, "grad_norm": 4.7097673416137695, "learning_rate": 9.807706711412553e-06, "loss": 3.1582, "step": 14805 }, { "epoch": 0.5217192138622408, "grad_norm": 0.8748592138290405, "learning_rate": 9.806565646887896e-06, "loss": 0.8854, "step": 14806 }, { "epoch": 0.5217544508751993, "grad_norm": 1.0278607606887817, "learning_rate": 9.80542458488275e-06, "loss": 0.7331, "step": 14807 }, { "epoch": 0.5217896878881577, "grad_norm": 1.2225511074066162, "learning_rate": 9.804283525411971e-06, "loss": 0.7722, "step": 14808 }, { "epoch": 0.5218249249011161, "grad_norm": 10.301742553710938, "learning_rate": 9.803142468490426e-06, "loss": 2.9394, "step": 14809 }, { "epoch": 0.5218601619140746, "grad_norm": 8.6731538772583, "learning_rate": 9.802001414132976e-06, "loss": 2.7769, "step": 14810 }, { "epoch": 0.521895398927033, "grad_norm": 6.088810443878174, "learning_rate": 9.80086036235448e-06, "loss": 5.384, "step": 14811 }, { "epoch": 0.5219306359399913, "grad_norm": 0.9224230051040649, "learning_rate": 9.799719313169803e-06, "loss": 1.0944, "step": 14812 }, { "epoch": 0.5219658729529498, "grad_norm": 1.6118561029434204, "learning_rate": 9.798578266593811e-06, "loss": 0.78, "step": 14813 }, { "epoch": 0.5220011099659082, "grad_norm": 6.445191860198975, "learning_rate": 9.79743722264136e-06, "loss": 5.486, "step": 14814 }, { "epoch": 0.5220363469788666, "grad_norm": 1.4729095697402954, "learning_rate": 9.796296181327313e-06, "loss": 1.0494, "step": 14815 }, { "epoch": 0.522071583991825, "grad_norm": 1.782181978225708, "learning_rate": 9.795155142666541e-06, "loss": 0.935, "step": 14816 }, { "epoch": 0.5221068210047835, "grad_norm": 5.797970771789551, "learning_rate": 9.794014106673894e-06, "loss": 2.9697, "step": 14817 }, { "epoch": 0.5221420580177418, "grad_norm": 1.3269758224487305, "learning_rate": 9.79287307336424e-06, "loss": 0.71, "step": 14818 }, { "epoch": 0.5221772950307002, "grad_norm": 0.9961037039756775, "learning_rate": 9.791732042752439e-06, "loss": 1.1894, "step": 14819 }, { "epoch": 0.5222125320436587, "grad_norm": 4.561619281768799, "learning_rate": 9.79059101485336e-06, "loss": 2.6454, "step": 14820 }, { "epoch": 0.5222477690566171, "grad_norm": 1.7811845541000366, "learning_rate": 9.789449989681855e-06, "loss": 0.9638, "step": 14821 }, { "epoch": 0.5222830060695755, "grad_norm": 5.2524542808532715, "learning_rate": 9.788308967252792e-06, "loss": 3.3265, "step": 14822 }, { "epoch": 0.5223182430825339, "grad_norm": 5.838130950927734, "learning_rate": 9.787167947581035e-06, "loss": 2.8027, "step": 14823 }, { "epoch": 0.5223534800954923, "grad_norm": 2.3881900310516357, "learning_rate": 9.786026930681439e-06, "loss": 0.8611, "step": 14824 }, { "epoch": 0.5223887171084507, "grad_norm": 5.389630317687988, "learning_rate": 9.784885916568868e-06, "loss": 2.8287, "step": 14825 }, { "epoch": 0.5224239541214091, "grad_norm": 3.2840566635131836, "learning_rate": 9.78374490525819e-06, "loss": 2.6283, "step": 14826 }, { "epoch": 0.5224591911343676, "grad_norm": 1.1582887172698975, "learning_rate": 9.782603896764259e-06, "loss": 1.0847, "step": 14827 }, { "epoch": 0.5224944281473259, "grad_norm": 3.6174817085266113, "learning_rate": 9.78146289110194e-06, "loss": 3.3157, "step": 14828 }, { "epoch": 0.5225296651602843, "grad_norm": 7.242424964904785, "learning_rate": 9.780321888286097e-06, "loss": 4.8835, "step": 14829 }, { "epoch": 0.5225649021732428, "grad_norm": 2.9850828647613525, "learning_rate": 9.779180888331585e-06, "loss": 3.3087, "step": 14830 }, { "epoch": 0.5226001391862012, "grad_norm": 6.17368745803833, "learning_rate": 9.778039891253273e-06, "loss": 2.8109, "step": 14831 }, { "epoch": 0.5226353761991596, "grad_norm": 5.518560886383057, "learning_rate": 9.77689889706602e-06, "loss": 5.1698, "step": 14832 }, { "epoch": 0.522670613212118, "grad_norm": 4.267583847045898, "learning_rate": 9.775757905784687e-06, "loss": 2.7014, "step": 14833 }, { "epoch": 0.5227058502250764, "grad_norm": 2.4623727798461914, "learning_rate": 9.774616917424133e-06, "loss": 2.5725, "step": 14834 }, { "epoch": 0.5227410872380348, "grad_norm": 1.08924400806427, "learning_rate": 9.77347593199923e-06, "loss": 0.9335, "step": 14835 }, { "epoch": 0.5227763242509933, "grad_norm": 4.256348609924316, "learning_rate": 9.772334949524823e-06, "loss": 2.8249, "step": 14836 }, { "epoch": 0.5228115612639517, "grad_norm": 4.545199394226074, "learning_rate": 9.771193970015787e-06, "loss": 2.7133, "step": 14837 }, { "epoch": 0.52284679827691, "grad_norm": 3.597476005554199, "learning_rate": 9.770052993486975e-06, "loss": 2.8334, "step": 14838 }, { "epoch": 0.5228820352898684, "grad_norm": 7.0915937423706055, "learning_rate": 9.768912019953259e-06, "loss": 5.9034, "step": 14839 }, { "epoch": 0.5229172723028269, "grad_norm": 6.417726516723633, "learning_rate": 9.767771049429486e-06, "loss": 2.6073, "step": 14840 }, { "epoch": 0.5229525093157853, "grad_norm": 4.3080315589904785, "learning_rate": 9.766630081930528e-06, "loss": 2.9306, "step": 14841 }, { "epoch": 0.5229877463287437, "grad_norm": 1.1503844261169434, "learning_rate": 9.765489117471244e-06, "loss": 1.0538, "step": 14842 }, { "epoch": 0.5230229833417022, "grad_norm": 4.389695644378662, "learning_rate": 9.764348156066493e-06, "loss": 3.1475, "step": 14843 }, { "epoch": 0.5230582203546605, "grad_norm": 5.809800624847412, "learning_rate": 9.763207197731135e-06, "loss": 3.2924, "step": 14844 }, { "epoch": 0.5230934573676189, "grad_norm": 1.4068105220794678, "learning_rate": 9.762066242480037e-06, "loss": 1.1135, "step": 14845 }, { "epoch": 0.5231286943805774, "grad_norm": 1.4182747602462769, "learning_rate": 9.760925290328054e-06, "loss": 0.8298, "step": 14846 }, { "epoch": 0.5231639313935358, "grad_norm": 3.2866930961608887, "learning_rate": 9.759784341290051e-06, "loss": 2.6267, "step": 14847 }, { "epoch": 0.5231991684064942, "grad_norm": 1.020349383354187, "learning_rate": 9.75864339538089e-06, "loss": 0.888, "step": 14848 }, { "epoch": 0.5232344054194525, "grad_norm": 7.189998149871826, "learning_rate": 9.757502452615423e-06, "loss": 3.3747, "step": 14849 }, { "epoch": 0.523269642432411, "grad_norm": 1.3919074535369873, "learning_rate": 9.75636151300852e-06, "loss": 0.8546, "step": 14850 }, { "epoch": 0.5233048794453694, "grad_norm": 4.390908241271973, "learning_rate": 9.755220576575043e-06, "loss": 2.814, "step": 14851 }, { "epoch": 0.5233401164583278, "grad_norm": 8.23220157623291, "learning_rate": 9.754079643329848e-06, "loss": 6.0337, "step": 14852 }, { "epoch": 0.5233753534712863, "grad_norm": 1.576796293258667, "learning_rate": 9.752938713287792e-06, "loss": 1.2764, "step": 14853 }, { "epoch": 0.5234105904842447, "grad_norm": 5.214077949523926, "learning_rate": 9.751797786463748e-06, "loss": 3.3168, "step": 14854 }, { "epoch": 0.523445827497203, "grad_norm": 0.8866961598396301, "learning_rate": 9.750656862872563e-06, "loss": 0.7846, "step": 14855 }, { "epoch": 0.5234810645101615, "grad_norm": 1.0779746770858765, "learning_rate": 9.749515942529107e-06, "loss": 0.7845, "step": 14856 }, { "epoch": 0.5235163015231199, "grad_norm": 1.6072416305541992, "learning_rate": 9.748375025448234e-06, "loss": 0.7203, "step": 14857 }, { "epoch": 0.5235515385360783, "grad_norm": 5.593657970428467, "learning_rate": 9.747234111644816e-06, "loss": 5.2843, "step": 14858 }, { "epoch": 0.5235867755490367, "grad_norm": 6.35692024230957, "learning_rate": 9.7460932011337e-06, "loss": 5.1413, "step": 14859 }, { "epoch": 0.5236220125619951, "grad_norm": 4.91215181350708, "learning_rate": 9.744952293929755e-06, "loss": 4.1969, "step": 14860 }, { "epoch": 0.5236572495749535, "grad_norm": 2.2843332290649414, "learning_rate": 9.743811390047839e-06, "loss": 0.908, "step": 14861 }, { "epoch": 0.5236924865879119, "grad_norm": 1.153464674949646, "learning_rate": 9.74267048950281e-06, "loss": 0.8054, "step": 14862 }, { "epoch": 0.5237277236008704, "grad_norm": 7.101963043212891, "learning_rate": 9.74152959230953e-06, "loss": 2.9643, "step": 14863 }, { "epoch": 0.5237629606138288, "grad_norm": 2.1041111946105957, "learning_rate": 9.740388698482865e-06, "loss": 0.8154, "step": 14864 }, { "epoch": 0.5237981976267871, "grad_norm": 1.9734262228012085, "learning_rate": 9.739247808037668e-06, "loss": 1.0995, "step": 14865 }, { "epoch": 0.5238334346397456, "grad_norm": 3.1574134826660156, "learning_rate": 9.738106920988799e-06, "loss": 2.8272, "step": 14866 }, { "epoch": 0.523868671652704, "grad_norm": 12.191451072692871, "learning_rate": 9.736966037351126e-06, "loss": 3.5168, "step": 14867 }, { "epoch": 0.5239039086656624, "grad_norm": 4.95574426651001, "learning_rate": 9.735825157139497e-06, "loss": 5.3584, "step": 14868 }, { "epoch": 0.5239391456786209, "grad_norm": 7.898538589477539, "learning_rate": 9.734684280368783e-06, "loss": 5.4762, "step": 14869 }, { "epoch": 0.5239743826915793, "grad_norm": 3.1323018074035645, "learning_rate": 9.73354340705384e-06, "loss": 2.6783, "step": 14870 }, { "epoch": 0.5240096197045376, "grad_norm": 1.560568928718567, "learning_rate": 9.732402537209528e-06, "loss": 0.9407, "step": 14871 }, { "epoch": 0.524044856717496, "grad_norm": 1.3065718412399292, "learning_rate": 9.731261670850705e-06, "loss": 0.9467, "step": 14872 }, { "epoch": 0.5240800937304545, "grad_norm": 1.1405460834503174, "learning_rate": 9.730120807992236e-06, "loss": 0.8917, "step": 14873 }, { "epoch": 0.5241153307434129, "grad_norm": 4.348145008087158, "learning_rate": 9.728979948648975e-06, "loss": 2.904, "step": 14874 }, { "epoch": 0.5241505677563713, "grad_norm": 2.0644724369049072, "learning_rate": 9.727839092835785e-06, "loss": 0.9021, "step": 14875 }, { "epoch": 0.5241858047693297, "grad_norm": 2.1063392162323, "learning_rate": 9.726698240567527e-06, "loss": 0.6539, "step": 14876 }, { "epoch": 0.5242210417822881, "grad_norm": 5.169435977935791, "learning_rate": 9.72555739185906e-06, "loss": 3.4558, "step": 14877 }, { "epoch": 0.5242562787952465, "grad_norm": 1.5079587697982788, "learning_rate": 9.724416546725239e-06, "loss": 1.0381, "step": 14878 }, { "epoch": 0.524291515808205, "grad_norm": 11.596755981445312, "learning_rate": 9.723275705180931e-06, "loss": 2.9295, "step": 14879 }, { "epoch": 0.5243267528211634, "grad_norm": 1.6481704711914062, "learning_rate": 9.722134867240993e-06, "loss": 1.0116, "step": 14880 }, { "epoch": 0.5243619898341217, "grad_norm": 0.9433703422546387, "learning_rate": 9.720994032920282e-06, "loss": 1.2316, "step": 14881 }, { "epoch": 0.5243972268470801, "grad_norm": 2.2471563816070557, "learning_rate": 9.719853202233656e-06, "loss": 2.6173, "step": 14882 }, { "epoch": 0.5244324638600386, "grad_norm": 10.441946983337402, "learning_rate": 9.718712375195987e-06, "loss": 3.4803, "step": 14883 }, { "epoch": 0.524467700872997, "grad_norm": 2.9897496700286865, "learning_rate": 9.717571551822116e-06, "loss": 3.1921, "step": 14884 }, { "epoch": 0.5245029378859554, "grad_norm": 1.9239201545715332, "learning_rate": 9.716430732126915e-06, "loss": 0.7055, "step": 14885 }, { "epoch": 0.5245381748989139, "grad_norm": 11.192684173583984, "learning_rate": 9.715289916125242e-06, "loss": 4.9895, "step": 14886 }, { "epoch": 0.5245734119118722, "grad_norm": 1.7964580059051514, "learning_rate": 9.71414910383195e-06, "loss": 0.7666, "step": 14887 }, { "epoch": 0.5246086489248306, "grad_norm": 1.6779191493988037, "learning_rate": 9.713008295261903e-06, "loss": 0.7744, "step": 14888 }, { "epoch": 0.5246438859377891, "grad_norm": 6.256261348724365, "learning_rate": 9.711867490429961e-06, "loss": 5.2391, "step": 14889 }, { "epoch": 0.5246791229507475, "grad_norm": 0.8045135140419006, "learning_rate": 9.710726689350982e-06, "loss": 0.7438, "step": 14890 }, { "epoch": 0.5247143599637059, "grad_norm": 4.408333778381348, "learning_rate": 9.709585892039823e-06, "loss": 2.7675, "step": 14891 }, { "epoch": 0.5247495969766642, "grad_norm": 1.8440827131271362, "learning_rate": 9.708445098511344e-06, "loss": 0.9173, "step": 14892 }, { "epoch": 0.5247848339896227, "grad_norm": 3.133928060531616, "learning_rate": 9.707304308780408e-06, "loss": 3.0056, "step": 14893 }, { "epoch": 0.5248200710025811, "grad_norm": 7.114940643310547, "learning_rate": 9.706163522861869e-06, "loss": 3.2455, "step": 14894 }, { "epoch": 0.5248553080155395, "grad_norm": 7.899099349975586, "learning_rate": 9.705022740770588e-06, "loss": 3.1634, "step": 14895 }, { "epoch": 0.524890545028498, "grad_norm": 1.290778636932373, "learning_rate": 9.703881962521424e-06, "loss": 1.0138, "step": 14896 }, { "epoch": 0.5249257820414563, "grad_norm": 1.9942049980163574, "learning_rate": 9.702741188129233e-06, "loss": 1.0489, "step": 14897 }, { "epoch": 0.5249610190544147, "grad_norm": 1.4101330041885376, "learning_rate": 9.701600417608877e-06, "loss": 0.916, "step": 14898 }, { "epoch": 0.5249962560673732, "grad_norm": 8.677477836608887, "learning_rate": 9.700459650975214e-06, "loss": 3.3112, "step": 14899 }, { "epoch": 0.5250314930803316, "grad_norm": 4.714353561401367, "learning_rate": 9.699318888243102e-06, "loss": 3.5154, "step": 14900 }, { "epoch": 0.52506673009329, "grad_norm": 1.4080135822296143, "learning_rate": 9.698178129427396e-06, "loss": 0.7958, "step": 14901 }, { "epoch": 0.5251019671062485, "grad_norm": 6.0877556800842285, "learning_rate": 9.697037374542967e-06, "loss": 2.9942, "step": 14902 }, { "epoch": 0.5251372041192068, "grad_norm": 6.54805326461792, "learning_rate": 9.695896623604657e-06, "loss": 3.0491, "step": 14903 }, { "epoch": 0.5251724411321652, "grad_norm": 8.843456268310547, "learning_rate": 9.694755876627336e-06, "loss": 2.902, "step": 14904 }, { "epoch": 0.5252076781451236, "grad_norm": 3.1866445541381836, "learning_rate": 9.693615133625859e-06, "loss": 2.5921, "step": 14905 }, { "epoch": 0.5252429151580821, "grad_norm": 3.165543794631958, "learning_rate": 9.692474394615082e-06, "loss": 2.6107, "step": 14906 }, { "epoch": 0.5252781521710405, "grad_norm": 1.806322693824768, "learning_rate": 9.691333659609862e-06, "loss": 1.125, "step": 14907 }, { "epoch": 0.5253133891839988, "grad_norm": 4.743503093719482, "learning_rate": 9.690192928625067e-06, "loss": 4.5066, "step": 14908 }, { "epoch": 0.5253486261969573, "grad_norm": 4.743503093719482, "learning_rate": 9.690192928625067e-06, "loss": 5.3153, "step": 14909 }, { "epoch": 0.5253838632099157, "grad_norm": 4.649029731750488, "learning_rate": 9.689052201675545e-06, "loss": 3.2258, "step": 14910 }, { "epoch": 0.5254191002228741, "grad_norm": 3.7028443813323975, "learning_rate": 9.687911478776159e-06, "loss": 3.1099, "step": 14911 }, { "epoch": 0.5254543372358326, "grad_norm": 5.758273124694824, "learning_rate": 9.686770759941761e-06, "loss": 2.9847, "step": 14912 }, { "epoch": 0.525489574248791, "grad_norm": 1.8291261196136475, "learning_rate": 9.68563004518722e-06, "loss": 1.0589, "step": 14913 }, { "epoch": 0.5255248112617493, "grad_norm": 3.0226898193359375, "learning_rate": 9.684489334527385e-06, "loss": 0.8994, "step": 14914 }, { "epoch": 0.5255600482747077, "grad_norm": 6.355788230895996, "learning_rate": 9.683348627977117e-06, "loss": 4.2448, "step": 14915 }, { "epoch": 0.5255952852876662, "grad_norm": 4.994376182556152, "learning_rate": 9.682207925551276e-06, "loss": 3.1827, "step": 14916 }, { "epoch": 0.5256305223006246, "grad_norm": 4.994376182556152, "learning_rate": 9.682207925551276e-06, "loss": 4.2288, "step": 14917 }, { "epoch": 0.525665759313583, "grad_norm": 1.2951757907867432, "learning_rate": 9.681067227264712e-06, "loss": 1.0418, "step": 14918 }, { "epoch": 0.5257009963265414, "grad_norm": 0.9091987609863281, "learning_rate": 9.679926533132291e-06, "loss": 0.9549, "step": 14919 }, { "epoch": 0.5257362333394998, "grad_norm": 1.2916860580444336, "learning_rate": 9.67878584316887e-06, "loss": 0.7127, "step": 14920 }, { "epoch": 0.5257714703524582, "grad_norm": 0.9810687899589539, "learning_rate": 9.677645157389301e-06, "loss": 1.2791, "step": 14921 }, { "epoch": 0.5258067073654167, "grad_norm": 1.5209730863571167, "learning_rate": 9.676504475808443e-06, "loss": 0.8897, "step": 14922 }, { "epoch": 0.5258419443783751, "grad_norm": 11.889168739318848, "learning_rate": 9.675363798441162e-06, "loss": 3.0299, "step": 14923 }, { "epoch": 0.5258771813913334, "grad_norm": 2.138789176940918, "learning_rate": 9.674223125302303e-06, "loss": 1.1174, "step": 14924 }, { "epoch": 0.5259124184042918, "grad_norm": 7.848493576049805, "learning_rate": 9.67308245640673e-06, "loss": 4.5187, "step": 14925 }, { "epoch": 0.5259476554172503, "grad_norm": 1.7308244705200195, "learning_rate": 9.671941791769304e-06, "loss": 0.9283, "step": 14926 }, { "epoch": 0.5259828924302087, "grad_norm": 1.3168376684188843, "learning_rate": 9.670801131404875e-06, "loss": 1.0447, "step": 14927 }, { "epoch": 0.5260181294431671, "grad_norm": 3.8502678871154785, "learning_rate": 9.6696604753283e-06, "loss": 3.0146, "step": 14928 }, { "epoch": 0.5260533664561255, "grad_norm": 8.087068557739258, "learning_rate": 9.668519823554443e-06, "loss": 8.058, "step": 14929 }, { "epoch": 0.5260886034690839, "grad_norm": 4.018180847167969, "learning_rate": 9.667379176098157e-06, "loss": 2.8724, "step": 14930 }, { "epoch": 0.5261238404820423, "grad_norm": 9.374689102172852, "learning_rate": 9.666238532974299e-06, "loss": 5.6253, "step": 14931 }, { "epoch": 0.5261590774950008, "grad_norm": 1.1183191537857056, "learning_rate": 9.665097894197726e-06, "loss": 1.0384, "step": 14932 }, { "epoch": 0.5261943145079592, "grad_norm": 1.1509569883346558, "learning_rate": 9.663957259783297e-06, "loss": 0.8348, "step": 14933 }, { "epoch": 0.5262295515209175, "grad_norm": 1.1200181245803833, "learning_rate": 9.662816629745868e-06, "loss": 0.9174, "step": 14934 }, { "epoch": 0.526264788533876, "grad_norm": 2.968867778778076, "learning_rate": 9.661676004100294e-06, "loss": 0.8574, "step": 14935 }, { "epoch": 0.5263000255468344, "grad_norm": 2.1603825092315674, "learning_rate": 9.660535382861436e-06, "loss": 0.8964, "step": 14936 }, { "epoch": 0.5263352625597928, "grad_norm": 5.915347099304199, "learning_rate": 9.659394766044144e-06, "loss": 3.7396, "step": 14937 }, { "epoch": 0.5263704995727512, "grad_norm": 1.247908115386963, "learning_rate": 9.658254153663279e-06, "loss": 0.7265, "step": 14938 }, { "epoch": 0.5264057365857097, "grad_norm": 6.412903785705566, "learning_rate": 9.6571135457337e-06, "loss": 3.2807, "step": 14939 }, { "epoch": 0.526440973598668, "grad_norm": 8.55811595916748, "learning_rate": 9.65597294227026e-06, "loss": 4.111, "step": 14940 }, { "epoch": 0.5264762106116264, "grad_norm": 5.365174770355225, "learning_rate": 9.654832343287812e-06, "loss": 3.0033, "step": 14941 }, { "epoch": 0.5265114476245849, "grad_norm": 1.492876648902893, "learning_rate": 9.653691748801224e-06, "loss": 1.0201, "step": 14942 }, { "epoch": 0.5265466846375433, "grad_norm": 1.2873852252960205, "learning_rate": 9.652551158825338e-06, "loss": 0.9056, "step": 14943 }, { "epoch": 0.5265819216505017, "grad_norm": 4.660391807556152, "learning_rate": 9.651410573375021e-06, "loss": 2.5472, "step": 14944 }, { "epoch": 0.5266171586634601, "grad_norm": 2.9665565490722656, "learning_rate": 9.650269992465126e-06, "loss": 2.7507, "step": 14945 }, { "epoch": 0.5266523956764185, "grad_norm": 2.642430543899536, "learning_rate": 9.649129416110509e-06, "loss": 2.8795, "step": 14946 }, { "epoch": 0.5266876326893769, "grad_norm": 3.9572389125823975, "learning_rate": 9.647988844326023e-06, "loss": 2.8693, "step": 14947 }, { "epoch": 0.5267228697023353, "grad_norm": 3.097158908843994, "learning_rate": 9.646848277126531e-06, "loss": 2.6966, "step": 14948 }, { "epoch": 0.5267581067152938, "grad_norm": 4.349148273468018, "learning_rate": 9.645707714526883e-06, "loss": 4.734, "step": 14949 }, { "epoch": 0.5267933437282522, "grad_norm": 1.2948347330093384, "learning_rate": 9.644567156541937e-06, "loss": 0.9991, "step": 14950 }, { "epoch": 0.5268285807412105, "grad_norm": 4.442108154296875, "learning_rate": 9.643426603186547e-06, "loss": 2.9567, "step": 14951 }, { "epoch": 0.526863817754169, "grad_norm": 3.75590181350708, "learning_rate": 9.642286054475576e-06, "loss": 0.8285, "step": 14952 }, { "epoch": 0.5268990547671274, "grad_norm": 6.729316234588623, "learning_rate": 9.641145510423873e-06, "loss": 2.6599, "step": 14953 }, { "epoch": 0.5269342917800858, "grad_norm": 0.9942330718040466, "learning_rate": 9.640004971046294e-06, "loss": 1.0758, "step": 14954 }, { "epoch": 0.5269695287930443, "grad_norm": 4.735429286956787, "learning_rate": 9.638864436357698e-06, "loss": 2.6479, "step": 14955 }, { "epoch": 0.5270047658060026, "grad_norm": 1.3667837381362915, "learning_rate": 9.637723906372935e-06, "loss": 0.957, "step": 14956 }, { "epoch": 0.527040002818961, "grad_norm": 9.416862487792969, "learning_rate": 9.636583381106868e-06, "loss": 2.3577, "step": 14957 }, { "epoch": 0.5270752398319194, "grad_norm": 7.8359808921813965, "learning_rate": 9.635442860574349e-06, "loss": 2.8748, "step": 14958 }, { "epoch": 0.5271104768448779, "grad_norm": 4.636030197143555, "learning_rate": 9.634302344790232e-06, "loss": 3.051, "step": 14959 }, { "epoch": 0.5271457138578363, "grad_norm": 5.365165710449219, "learning_rate": 9.63316183376937e-06, "loss": 3.5308, "step": 14960 }, { "epoch": 0.5271809508707946, "grad_norm": 1.7878676652908325, "learning_rate": 9.63202132752663e-06, "loss": 0.8557, "step": 14961 }, { "epoch": 0.5272161878837531, "grad_norm": 7.061310291290283, "learning_rate": 9.630880826076852e-06, "loss": 6.2891, "step": 14962 }, { "epoch": 0.5272514248967115, "grad_norm": 3.8610599040985107, "learning_rate": 9.6297403294349e-06, "loss": 2.7928, "step": 14963 }, { "epoch": 0.5272866619096699, "grad_norm": 5.309040069580078, "learning_rate": 9.628599837615628e-06, "loss": 3.1452, "step": 14964 }, { "epoch": 0.5273218989226284, "grad_norm": 3.7660927772521973, "learning_rate": 9.627459350633891e-06, "loss": 2.5526, "step": 14965 }, { "epoch": 0.5273571359355868, "grad_norm": 1.0323357582092285, "learning_rate": 9.62631886850454e-06, "loss": 1.2976, "step": 14966 }, { "epoch": 0.5273923729485451, "grad_norm": 9.546407699584961, "learning_rate": 9.625178391242435e-06, "loss": 5.8581, "step": 14967 }, { "epoch": 0.5274276099615036, "grad_norm": 4.398406505584717, "learning_rate": 9.624037918862433e-06, "loss": 3.0931, "step": 14968 }, { "epoch": 0.527462846974462, "grad_norm": 1.3273769617080688, "learning_rate": 9.622897451379383e-06, "loss": 0.8423, "step": 14969 }, { "epoch": 0.5274980839874204, "grad_norm": 3.1040396690368652, "learning_rate": 9.621756988808137e-06, "loss": 2.8053, "step": 14970 }, { "epoch": 0.5275333210003788, "grad_norm": 3.520601987838745, "learning_rate": 9.620616531163561e-06, "loss": 3.5851, "step": 14971 }, { "epoch": 0.5275685580133372, "grad_norm": 1.7881187200546265, "learning_rate": 9.6194760784605e-06, "loss": 0.9613, "step": 14972 }, { "epoch": 0.5276037950262956, "grad_norm": 5.172154426574707, "learning_rate": 9.618335630713812e-06, "loss": 2.7452, "step": 14973 }, { "epoch": 0.527639032039254, "grad_norm": 2.6720778942108154, "learning_rate": 9.617195187938354e-06, "loss": 2.4574, "step": 14974 }, { "epoch": 0.5276742690522125, "grad_norm": 2.060255765914917, "learning_rate": 9.616054750148972e-06, "loss": 0.9452, "step": 14975 }, { "epoch": 0.5277095060651709, "grad_norm": 7.041426181793213, "learning_rate": 9.614914317360526e-06, "loss": 4.8792, "step": 14976 }, { "epoch": 0.5277447430781292, "grad_norm": 9.031492233276367, "learning_rate": 9.613773889587875e-06, "loss": 2.7163, "step": 14977 }, { "epoch": 0.5277799800910877, "grad_norm": 6.232468128204346, "learning_rate": 9.612633466845865e-06, "loss": 3.4605, "step": 14978 }, { "epoch": 0.5278152171040461, "grad_norm": 2.642904281616211, "learning_rate": 9.611493049149351e-06, "loss": 2.6928, "step": 14979 }, { "epoch": 0.5278504541170045, "grad_norm": 0.9460288882255554, "learning_rate": 9.610352636513197e-06, "loss": 0.8825, "step": 14980 }, { "epoch": 0.5278856911299629, "grad_norm": 4.603534698486328, "learning_rate": 9.609212228952242e-06, "loss": 2.9502, "step": 14981 }, { "epoch": 0.5279209281429214, "grad_norm": 1.20616614818573, "learning_rate": 9.60807182648135e-06, "loss": 0.8793, "step": 14982 }, { "epoch": 0.5279561651558797, "grad_norm": 21.764039993286133, "learning_rate": 9.606931429115376e-06, "loss": 5.1931, "step": 14983 }, { "epoch": 0.5279914021688381, "grad_norm": 3.653078079223633, "learning_rate": 9.605791036869166e-06, "loss": 3.2722, "step": 14984 }, { "epoch": 0.5280266391817966, "grad_norm": 1.6351079940795898, "learning_rate": 9.604650649757578e-06, "loss": 0.8833, "step": 14985 }, { "epoch": 0.528061876194755, "grad_norm": 1.317569375038147, "learning_rate": 9.603510267795466e-06, "loss": 0.7989, "step": 14986 }, { "epoch": 0.5280971132077134, "grad_norm": 5.429270267486572, "learning_rate": 9.602369890997686e-06, "loss": 3.9238, "step": 14987 }, { "epoch": 0.5281323502206718, "grad_norm": 9.100739479064941, "learning_rate": 9.601229519379087e-06, "loss": 4.7244, "step": 14988 }, { "epoch": 0.5281675872336302, "grad_norm": 1.6893285512924194, "learning_rate": 9.60008915295452e-06, "loss": 0.9281, "step": 14989 }, { "epoch": 0.5282028242465886, "grad_norm": 1.2498550415039062, "learning_rate": 9.598948791738848e-06, "loss": 0.8108, "step": 14990 }, { "epoch": 0.528238061259547, "grad_norm": 1.2186574935913086, "learning_rate": 9.59780843574692e-06, "loss": 0.8614, "step": 14991 }, { "epoch": 0.5282732982725055, "grad_norm": 1.2780370712280273, "learning_rate": 9.596668084993586e-06, "loss": 1.2005, "step": 14992 }, { "epoch": 0.5283085352854638, "grad_norm": 3.6549229621887207, "learning_rate": 9.595527739493702e-06, "loss": 2.87, "step": 14993 }, { "epoch": 0.5283437722984222, "grad_norm": 6.845510959625244, "learning_rate": 9.594387399262119e-06, "loss": 6.8611, "step": 14994 }, { "epoch": 0.5283790093113807, "grad_norm": 3.2987961769104004, "learning_rate": 9.593247064313694e-06, "loss": 2.6133, "step": 14995 }, { "epoch": 0.5284142463243391, "grad_norm": 1.4165701866149902, "learning_rate": 9.592106734663279e-06, "loss": 0.8755, "step": 14996 }, { "epoch": 0.5284494833372975, "grad_norm": 3.3069891929626465, "learning_rate": 9.590966410325723e-06, "loss": 2.9079, "step": 14997 }, { "epoch": 0.528484720350256, "grad_norm": 3.2452898025512695, "learning_rate": 9.58982609131588e-06, "loss": 3.0014, "step": 14998 }, { "epoch": 0.5285199573632143, "grad_norm": 5.872589111328125, "learning_rate": 9.58868577764861e-06, "loss": 3.0246, "step": 14999 }, { "epoch": 0.5285551943761727, "grad_norm": 1.0914318561553955, "learning_rate": 9.587545469338755e-06, "loss": 0.763, "step": 15000 }, { "epoch": 0.5285904313891311, "grad_norm": 31.3741397857666, "learning_rate": 9.586405166401175e-06, "loss": 5.3084, "step": 15001 }, { "epoch": 0.5286256684020896, "grad_norm": 5.197569370269775, "learning_rate": 9.585264868850721e-06, "loss": 3.2656, "step": 15002 }, { "epoch": 0.528660905415048, "grad_norm": 2.268986463546753, "learning_rate": 9.584124576702246e-06, "loss": 0.9847, "step": 15003 }, { "epoch": 0.5286961424280063, "grad_norm": 10.28769588470459, "learning_rate": 9.582984289970595e-06, "loss": 3.6705, "step": 15004 }, { "epoch": 0.5287313794409648, "grad_norm": 6.071534156799316, "learning_rate": 9.581844008670631e-06, "loss": 2.6293, "step": 15005 }, { "epoch": 0.5287666164539232, "grad_norm": 5.2791619300842285, "learning_rate": 9.580703732817205e-06, "loss": 2.8021, "step": 15006 }, { "epoch": 0.5288018534668816, "grad_norm": 8.6264066696167, "learning_rate": 9.579563462425163e-06, "loss": 7.1428, "step": 15007 }, { "epoch": 0.5288370904798401, "grad_norm": 13.24733829498291, "learning_rate": 9.57842319750936e-06, "loss": 5.2058, "step": 15008 }, { "epoch": 0.5288723274927984, "grad_norm": 1.9986571073532104, "learning_rate": 9.57728293808465e-06, "loss": 0.8599, "step": 15009 }, { "epoch": 0.5289075645057568, "grad_norm": 1.1859593391418457, "learning_rate": 9.576142684165884e-06, "loss": 0.9061, "step": 15010 }, { "epoch": 0.5289428015187153, "grad_norm": 4.531250476837158, "learning_rate": 9.575002435767912e-06, "loss": 3.4107, "step": 15011 }, { "epoch": 0.5289780385316737, "grad_norm": 1.4528095722198486, "learning_rate": 9.57386219290559e-06, "loss": 0.7379, "step": 15012 }, { "epoch": 0.5290132755446321, "grad_norm": 10.44192886352539, "learning_rate": 9.572721955593762e-06, "loss": 5.4515, "step": 15013 }, { "epoch": 0.5290485125575904, "grad_norm": 2.9740214347839355, "learning_rate": 9.571581723847289e-06, "loss": 2.4474, "step": 15014 }, { "epoch": 0.5290837495705489, "grad_norm": 4.327117443084717, "learning_rate": 9.57044149768102e-06, "loss": 3.3015, "step": 15015 }, { "epoch": 0.5291189865835073, "grad_norm": 52.375144958496094, "learning_rate": 9.569301277109803e-06, "loss": 5.4477, "step": 15016 }, { "epoch": 0.5291542235964657, "grad_norm": 1.2226942777633667, "learning_rate": 9.568161062148489e-06, "loss": 0.8339, "step": 15017 }, { "epoch": 0.5291894606094242, "grad_norm": 7.945554733276367, "learning_rate": 9.56702085281194e-06, "loss": 5.6266, "step": 15018 }, { "epoch": 0.5292246976223826, "grad_norm": 5.563025951385498, "learning_rate": 9.565880649114991e-06, "loss": 5.2313, "step": 15019 }, { "epoch": 0.5292599346353409, "grad_norm": 1.1055676937103271, "learning_rate": 9.564740451072505e-06, "loss": 0.6567, "step": 15020 }, { "epoch": 0.5292951716482994, "grad_norm": 5.00057315826416, "learning_rate": 9.563600258699333e-06, "loss": 3.0836, "step": 15021 }, { "epoch": 0.5293304086612578, "grad_norm": 2.4185173511505127, "learning_rate": 9.56246007201032e-06, "loss": 0.9894, "step": 15022 }, { "epoch": 0.5293656456742162, "grad_norm": 4.802945613861084, "learning_rate": 9.561319891020318e-06, "loss": 2.9767, "step": 15023 }, { "epoch": 0.5294008826871746, "grad_norm": 7.396271228790283, "learning_rate": 9.560179715744183e-06, "loss": 4.556, "step": 15024 }, { "epoch": 0.529436119700133, "grad_norm": 6.693800449371338, "learning_rate": 9.559039546196766e-06, "loss": 3.6744, "step": 15025 }, { "epoch": 0.5294713567130914, "grad_norm": 0.9211235642433167, "learning_rate": 9.557899382392912e-06, "loss": 0.7857, "step": 15026 }, { "epoch": 0.5295065937260498, "grad_norm": 1.0111443996429443, "learning_rate": 9.556759224347473e-06, "loss": 1.0666, "step": 15027 }, { "epoch": 0.5295418307390083, "grad_norm": 1.1993037462234497, "learning_rate": 9.555619072075307e-06, "loss": 0.7016, "step": 15028 }, { "epoch": 0.5295770677519667, "grad_norm": 1.1699719429016113, "learning_rate": 9.554478925591255e-06, "loss": 0.8241, "step": 15029 }, { "epoch": 0.529612304764925, "grad_norm": 3.241475820541382, "learning_rate": 9.553338784910173e-06, "loss": 2.9694, "step": 15030 }, { "epoch": 0.5296475417778835, "grad_norm": 12.672796249389648, "learning_rate": 9.552198650046912e-06, "loss": 6.1686, "step": 15031 }, { "epoch": 0.5296827787908419, "grad_norm": 2.0932633876800537, "learning_rate": 9.55105852101632e-06, "loss": 0.9702, "step": 15032 }, { "epoch": 0.5297180158038003, "grad_norm": 2.852435827255249, "learning_rate": 9.549918397833243e-06, "loss": 2.6512, "step": 15033 }, { "epoch": 0.5297532528167587, "grad_norm": 3.49971079826355, "learning_rate": 9.548778280512542e-06, "loss": 2.9967, "step": 15034 }, { "epoch": 0.5297884898297172, "grad_norm": 3.535203218460083, "learning_rate": 9.54763816906906e-06, "loss": 0.9097, "step": 15035 }, { "epoch": 0.5298237268426755, "grad_norm": 4.555724143981934, "learning_rate": 9.546498063517646e-06, "loss": 2.7416, "step": 15036 }, { "epoch": 0.5298589638556339, "grad_norm": 1.4274914264678955, "learning_rate": 9.545357963873156e-06, "loss": 1.1456, "step": 15037 }, { "epoch": 0.5298942008685924, "grad_norm": 4.66254186630249, "learning_rate": 9.544217870150432e-06, "loss": 3.0647, "step": 15038 }, { "epoch": 0.5299294378815508, "grad_norm": 5.095966339111328, "learning_rate": 9.543077782364332e-06, "loss": 3.0903, "step": 15039 }, { "epoch": 0.5299646748945092, "grad_norm": 6.1684346199035645, "learning_rate": 9.5419377005297e-06, "loss": 3.0403, "step": 15040 }, { "epoch": 0.5299999119074676, "grad_norm": 1.4325636625289917, "learning_rate": 9.540797624661391e-06, "loss": 0.9048, "step": 15041 }, { "epoch": 0.530035148920426, "grad_norm": 5.140499591827393, "learning_rate": 9.539657554774246e-06, "loss": 3.4431, "step": 15042 }, { "epoch": 0.5300703859333844, "grad_norm": 6.0461812019348145, "learning_rate": 9.538517490883123e-06, "loss": 2.8284, "step": 15043 }, { "epoch": 0.5301056229463429, "grad_norm": 4.6956000328063965, "learning_rate": 9.537377433002872e-06, "loss": 3.0537, "step": 15044 }, { "epoch": 0.5301408599593013, "grad_norm": 2.1306095123291016, "learning_rate": 9.536237381148336e-06, "loss": 0.7558, "step": 15045 }, { "epoch": 0.5301760969722596, "grad_norm": 4.765504837036133, "learning_rate": 9.535097335334363e-06, "loss": 3.6941, "step": 15046 }, { "epoch": 0.530211333985218, "grad_norm": 1.5963308811187744, "learning_rate": 9.533957295575816e-06, "loss": 1.1038, "step": 15047 }, { "epoch": 0.5302465709981765, "grad_norm": 1.1314021348953247, "learning_rate": 9.532817261887525e-06, "loss": 1.0115, "step": 15048 }, { "epoch": 0.5302818080111349, "grad_norm": 5.553425312042236, "learning_rate": 9.531677234284352e-06, "loss": 3.5304, "step": 15049 }, { "epoch": 0.5303170450240933, "grad_norm": 3.037493944168091, "learning_rate": 9.530537212781145e-06, "loss": 2.6523, "step": 15050 }, { "epoch": 0.5303522820370518, "grad_norm": 1.4751713275909424, "learning_rate": 9.529397197392749e-06, "loss": 0.7815, "step": 15051 }, { "epoch": 0.5303875190500101, "grad_norm": 0.9808251261711121, "learning_rate": 9.52825718813401e-06, "loss": 0.9685, "step": 15052 }, { "epoch": 0.5304227560629685, "grad_norm": 1.1886398792266846, "learning_rate": 9.527117185019789e-06, "loss": 0.7758, "step": 15053 }, { "epoch": 0.530457993075927, "grad_norm": 1.1580229997634888, "learning_rate": 9.525977188064923e-06, "loss": 0.7124, "step": 15054 }, { "epoch": 0.5304932300888854, "grad_norm": 3.1942975521087646, "learning_rate": 9.524837197284265e-06, "loss": 2.8172, "step": 15055 }, { "epoch": 0.5305284671018438, "grad_norm": 5.014629364013672, "learning_rate": 9.523697212692665e-06, "loss": 3.0593, "step": 15056 }, { "epoch": 0.5305637041148021, "grad_norm": 1.2111340761184692, "learning_rate": 9.522557234304965e-06, "loss": 0.7763, "step": 15057 }, { "epoch": 0.5305989411277606, "grad_norm": 3.849153757095337, "learning_rate": 9.52141726213602e-06, "loss": 2.6588, "step": 15058 }, { "epoch": 0.530634178140719, "grad_norm": 4.680314064025879, "learning_rate": 9.520277296200676e-06, "loss": 3.9813, "step": 15059 }, { "epoch": 0.5306694151536774, "grad_norm": 7.103225231170654, "learning_rate": 9.519137336513782e-06, "loss": 3.097, "step": 15060 }, { "epoch": 0.5307046521666359, "grad_norm": 1.5984313488006592, "learning_rate": 9.517997383090184e-06, "loss": 0.8708, "step": 15061 }, { "epoch": 0.5307398891795942, "grad_norm": 7.966615676879883, "learning_rate": 9.516857435944732e-06, "loss": 5.0242, "step": 15062 }, { "epoch": 0.5307751261925526, "grad_norm": 1.3658252954483032, "learning_rate": 9.515717495092276e-06, "loss": 1.0691, "step": 15063 }, { "epoch": 0.5308103632055111, "grad_norm": 5.877202033996582, "learning_rate": 9.51457756054766e-06, "loss": 5.2145, "step": 15064 }, { "epoch": 0.5308456002184695, "grad_norm": 1.6381844282150269, "learning_rate": 9.513437632325729e-06, "loss": 1.0995, "step": 15065 }, { "epoch": 0.5308808372314279, "grad_norm": 8.79601764678955, "learning_rate": 9.512297710441343e-06, "loss": 3.2862, "step": 15066 }, { "epoch": 0.5309160742443862, "grad_norm": 10.11768627166748, "learning_rate": 9.511157794909334e-06, "loss": 7.9658, "step": 15067 }, { "epoch": 0.5309513112573447, "grad_norm": 1.1257268190383911, "learning_rate": 9.510017885744559e-06, "loss": 0.6715, "step": 15068 }, { "epoch": 0.5309865482703031, "grad_norm": 1.2445738315582275, "learning_rate": 9.508877982961866e-06, "loss": 0.7779, "step": 15069 }, { "epoch": 0.5310217852832615, "grad_norm": 1.6778925657272339, "learning_rate": 9.507738086576098e-06, "loss": 0.8059, "step": 15070 }, { "epoch": 0.53105702229622, "grad_norm": 2.7544572353363037, "learning_rate": 9.506598196602101e-06, "loss": 0.9876, "step": 15071 }, { "epoch": 0.5310922593091784, "grad_norm": 1.5526082515716553, "learning_rate": 9.50545831305473e-06, "loss": 0.7929, "step": 15072 }, { "epoch": 0.5311274963221367, "grad_norm": 6.658034801483154, "learning_rate": 9.504318435948828e-06, "loss": 2.5417, "step": 15073 }, { "epoch": 0.5311627333350952, "grad_norm": 1.2271788120269775, "learning_rate": 9.503178565299239e-06, "loss": 0.6531, "step": 15074 }, { "epoch": 0.5311979703480536, "grad_norm": 67.83509826660156, "learning_rate": 9.502038701120814e-06, "loss": 5.2091, "step": 15075 }, { "epoch": 0.531233207361012, "grad_norm": 1.2928993701934814, "learning_rate": 9.500898843428394e-06, "loss": 1.1701, "step": 15076 }, { "epoch": 0.5312684443739705, "grad_norm": 1.5034078359603882, "learning_rate": 9.499758992236834e-06, "loss": 0.8714, "step": 15077 }, { "epoch": 0.5313036813869288, "grad_norm": 4.061500072479248, "learning_rate": 9.498619147560976e-06, "loss": 3.4353, "step": 15078 }, { "epoch": 0.5313389183998872, "grad_norm": 2.892300605773926, "learning_rate": 9.49747930941567e-06, "loss": 1.0108, "step": 15079 }, { "epoch": 0.5313741554128456, "grad_norm": 11.313464164733887, "learning_rate": 9.496339477815755e-06, "loss": 5.0455, "step": 15080 }, { "epoch": 0.5314093924258041, "grad_norm": 9.504987716674805, "learning_rate": 9.495199652776086e-06, "loss": 6.9632, "step": 15081 }, { "epoch": 0.5314446294387625, "grad_norm": 4.434656143188477, "learning_rate": 9.494059834311506e-06, "loss": 3.1133, "step": 15082 }, { "epoch": 0.5314798664517209, "grad_norm": 1.5275704860687256, "learning_rate": 9.492920022436859e-06, "loss": 1.1279, "step": 15083 }, { "epoch": 0.5315151034646793, "grad_norm": 4.00307559967041, "learning_rate": 9.491780217166991e-06, "loss": 2.7524, "step": 15084 }, { "epoch": 0.5315503404776377, "grad_norm": 3.92909836769104, "learning_rate": 9.49064041851676e-06, "loss": 2.8641, "step": 15085 }, { "epoch": 0.5315855774905961, "grad_norm": 19.847543716430664, "learning_rate": 9.489500626500992e-06, "loss": 3.1813, "step": 15086 }, { "epoch": 0.5316208145035546, "grad_norm": 1.27983820438385, "learning_rate": 9.488360841134548e-06, "loss": 0.7211, "step": 15087 }, { "epoch": 0.531656051516513, "grad_norm": 4.468421936035156, "learning_rate": 9.48722106243227e-06, "loss": 5.4255, "step": 15088 }, { "epoch": 0.5316912885294713, "grad_norm": 7.383250713348389, "learning_rate": 9.486081290409002e-06, "loss": 5.2123, "step": 15089 }, { "epoch": 0.5317265255424297, "grad_norm": 1.6133785247802734, "learning_rate": 9.484941525079588e-06, "loss": 0.7701, "step": 15090 }, { "epoch": 0.5317617625553882, "grad_norm": 4.295303821563721, "learning_rate": 9.48380176645888e-06, "loss": 3.5675, "step": 15091 }, { "epoch": 0.5317969995683466, "grad_norm": 0.864655077457428, "learning_rate": 9.482662014561719e-06, "loss": 0.732, "step": 15092 }, { "epoch": 0.531832236581305, "grad_norm": 1.0587797164916992, "learning_rate": 9.481522269402951e-06, "loss": 0.8205, "step": 15093 }, { "epoch": 0.5318674735942635, "grad_norm": 1.48735511302948, "learning_rate": 9.480382530997418e-06, "loss": 0.9161, "step": 15094 }, { "epoch": 0.5319027106072218, "grad_norm": 6.509936809539795, "learning_rate": 9.479242799359973e-06, "loss": 3.5406, "step": 15095 }, { "epoch": 0.5319379476201802, "grad_norm": 4.829355239868164, "learning_rate": 9.478103074505457e-06, "loss": 2.4808, "step": 15096 }, { "epoch": 0.5319731846331387, "grad_norm": 3.454245090484619, "learning_rate": 9.476963356448713e-06, "loss": 1.1674, "step": 15097 }, { "epoch": 0.5320084216460971, "grad_norm": 1.097501277923584, "learning_rate": 9.47582364520459e-06, "loss": 0.7285, "step": 15098 }, { "epoch": 0.5320436586590555, "grad_norm": 1.413289189338684, "learning_rate": 9.474683940787927e-06, "loss": 0.9308, "step": 15099 }, { "epoch": 0.5320788956720138, "grad_norm": 3.626044750213623, "learning_rate": 9.473544243213575e-06, "loss": 3.2719, "step": 15100 }, { "epoch": 0.5321141326849723, "grad_norm": 3.6257598400115967, "learning_rate": 9.472404552496376e-06, "loss": 3.1084, "step": 15101 }, { "epoch": 0.5321493696979307, "grad_norm": 2.4818215370178223, "learning_rate": 9.471264868651176e-06, "loss": 0.6902, "step": 15102 }, { "epoch": 0.5321846067108891, "grad_norm": 4.96703577041626, "learning_rate": 9.470125191692815e-06, "loss": 4.9469, "step": 15103 }, { "epoch": 0.5322198437238476, "grad_norm": 1.193281888961792, "learning_rate": 9.468985521636148e-06, "loss": 0.946, "step": 15104 }, { "epoch": 0.5322550807368059, "grad_norm": 7.456015110015869, "learning_rate": 9.467845858496006e-06, "loss": 3.0678, "step": 15105 }, { "epoch": 0.5322903177497643, "grad_norm": 1.5845539569854736, "learning_rate": 9.46670620228724e-06, "loss": 0.7826, "step": 15106 }, { "epoch": 0.5323255547627228, "grad_norm": 5.316145420074463, "learning_rate": 9.465566553024696e-06, "loss": 4.8976, "step": 15107 }, { "epoch": 0.5323607917756812, "grad_norm": 5.254111289978027, "learning_rate": 9.464426910723214e-06, "loss": 2.6164, "step": 15108 }, { "epoch": 0.5323960287886396, "grad_norm": 1.6752139329910278, "learning_rate": 9.463287275397639e-06, "loss": 0.912, "step": 15109 }, { "epoch": 0.532431265801598, "grad_norm": 1.6187775135040283, "learning_rate": 9.462147647062818e-06, "loss": 0.9972, "step": 15110 }, { "epoch": 0.5324665028145564, "grad_norm": 1.3257943391799927, "learning_rate": 9.461008025733591e-06, "loss": 0.9353, "step": 15111 }, { "epoch": 0.5325017398275148, "grad_norm": 9.11674690246582, "learning_rate": 9.459868411424804e-06, "loss": 5.6424, "step": 15112 }, { "epoch": 0.5325369768404732, "grad_norm": 21.64848518371582, "learning_rate": 9.458728804151296e-06, "loss": 3.2733, "step": 15113 }, { "epoch": 0.5325722138534317, "grad_norm": 2.657966136932373, "learning_rate": 9.45758920392792e-06, "loss": 0.7941, "step": 15114 }, { "epoch": 0.53260745086639, "grad_norm": 2.807180404663086, "learning_rate": 9.45644961076951e-06, "loss": 1.0581, "step": 15115 }, { "epoch": 0.5326426878793484, "grad_norm": 0.8760673403739929, "learning_rate": 9.455310024690914e-06, "loss": 1.159, "step": 15116 }, { "epoch": 0.5326779248923069, "grad_norm": 8.30209732055664, "learning_rate": 9.454170445706976e-06, "loss": 4.7879, "step": 15117 }, { "epoch": 0.5327131619052653, "grad_norm": 3.184574604034424, "learning_rate": 9.453030873832534e-06, "loss": 2.5021, "step": 15118 }, { "epoch": 0.5327483989182237, "grad_norm": 5.074282646179199, "learning_rate": 9.451891309082436e-06, "loss": 2.7749, "step": 15119 }, { "epoch": 0.5327836359311822, "grad_norm": 5.212418556213379, "learning_rate": 9.450751751471526e-06, "loss": 2.7289, "step": 15120 }, { "epoch": 0.5328188729441405, "grad_norm": 1.8767931461334229, "learning_rate": 9.449612201014642e-06, "loss": 0.8176, "step": 15121 }, { "epoch": 0.5328541099570989, "grad_norm": 1.7736319303512573, "learning_rate": 9.448472657726626e-06, "loss": 0.6608, "step": 15122 }, { "epoch": 0.5328893469700573, "grad_norm": 6.4703474044799805, "learning_rate": 9.447333121622332e-06, "loss": 3.1978, "step": 15123 }, { "epoch": 0.5329245839830158, "grad_norm": 4.5588154792785645, "learning_rate": 9.446193592716587e-06, "loss": 3.0331, "step": 15124 }, { "epoch": 0.5329598209959742, "grad_norm": 4.262763023376465, "learning_rate": 9.445054071024244e-06, "loss": 2.63, "step": 15125 }, { "epoch": 0.5329950580089325, "grad_norm": 6.420073509216309, "learning_rate": 9.443914556560145e-06, "loss": 3.2114, "step": 15126 }, { "epoch": 0.533030295021891, "grad_norm": 4.656804084777832, "learning_rate": 9.442775049339126e-06, "loss": 2.9801, "step": 15127 }, { "epoch": 0.5330655320348494, "grad_norm": 5.87251615524292, "learning_rate": 9.441635549376032e-06, "loss": 2.7772, "step": 15128 }, { "epoch": 0.5331007690478078, "grad_norm": 1.4210872650146484, "learning_rate": 9.44049605668571e-06, "loss": 1.0304, "step": 15129 }, { "epoch": 0.5331360060607663, "grad_norm": 1.3688966035842896, "learning_rate": 9.439356571282995e-06, "loss": 1.0592, "step": 15130 }, { "epoch": 0.5331712430737247, "grad_norm": 1.4815962314605713, "learning_rate": 9.438217093182734e-06, "loss": 0.8481, "step": 15131 }, { "epoch": 0.533206480086683, "grad_norm": 5.677876949310303, "learning_rate": 9.437077622399764e-06, "loss": 2.5547, "step": 15132 }, { "epoch": 0.5332417170996414, "grad_norm": 1.2133350372314453, "learning_rate": 9.435938158948932e-06, "loss": 1.2506, "step": 15133 }, { "epoch": 0.5332769541125999, "grad_norm": 8.35986614227295, "learning_rate": 9.434798702845078e-06, "loss": 3.1669, "step": 15134 }, { "epoch": 0.5333121911255583, "grad_norm": 5.623571872711182, "learning_rate": 9.433659254103042e-06, "loss": 5.6053, "step": 15135 }, { "epoch": 0.5333474281385167, "grad_norm": 6.555689811706543, "learning_rate": 9.432519812737668e-06, "loss": 5.1492, "step": 15136 }, { "epoch": 0.5333826651514751, "grad_norm": 1.4104386568069458, "learning_rate": 9.431380378763792e-06, "loss": 1.0863, "step": 15137 }, { "epoch": 0.5334179021644335, "grad_norm": 1.1803390979766846, "learning_rate": 9.430240952196261e-06, "loss": 1.3574, "step": 15138 }, { "epoch": 0.5334531391773919, "grad_norm": 6.609861373901367, "learning_rate": 9.429101533049916e-06, "loss": 3.0748, "step": 15139 }, { "epoch": 0.5334883761903504, "grad_norm": 6.017730712890625, "learning_rate": 9.427962121339595e-06, "loss": 2.5497, "step": 15140 }, { "epoch": 0.5335236132033088, "grad_norm": 2.480720281600952, "learning_rate": 9.426822717080137e-06, "loss": 1.2512, "step": 15141 }, { "epoch": 0.5335588502162671, "grad_norm": 1.0854781866073608, "learning_rate": 9.425683320286394e-06, "loss": 1.1548, "step": 15142 }, { "epoch": 0.5335940872292256, "grad_norm": 6.441556453704834, "learning_rate": 9.424543930973191e-06, "loss": 5.669, "step": 15143 }, { "epoch": 0.533629324242184, "grad_norm": 1.286116600036621, "learning_rate": 9.42340454915538e-06, "loss": 0.8367, "step": 15144 }, { "epoch": 0.5336645612551424, "grad_norm": 4.4334588050842285, "learning_rate": 9.4222651748478e-06, "loss": 3.2008, "step": 15145 }, { "epoch": 0.5336997982681008, "grad_norm": 2.9505906105041504, "learning_rate": 9.421125808065285e-06, "loss": 2.4726, "step": 15146 }, { "epoch": 0.5337350352810593, "grad_norm": 3.396179437637329, "learning_rate": 9.41998644882268e-06, "loss": 1.1234, "step": 15147 }, { "epoch": 0.5337702722940176, "grad_norm": 9.923454284667969, "learning_rate": 9.418847097134832e-06, "loss": 3.0575, "step": 15148 }, { "epoch": 0.533805509306976, "grad_norm": 3.9833076000213623, "learning_rate": 9.417707753016568e-06, "loss": 2.5838, "step": 15149 }, { "epoch": 0.5338407463199345, "grad_norm": 9.372200965881348, "learning_rate": 9.416568416482737e-06, "loss": 3.082, "step": 15150 }, { "epoch": 0.5338759833328929, "grad_norm": 5.706882953643799, "learning_rate": 9.415429087548174e-06, "loss": 2.7108, "step": 15151 }, { "epoch": 0.5339112203458513, "grad_norm": 1.1273361444473267, "learning_rate": 9.414289766227728e-06, "loss": 0.8233, "step": 15152 }, { "epoch": 0.5339464573588097, "grad_norm": 1.0564606189727783, "learning_rate": 9.413150452536229e-06, "loss": 0.8538, "step": 15153 }, { "epoch": 0.5339816943717681, "grad_norm": 1.9031381607055664, "learning_rate": 9.412011146488518e-06, "loss": 0.7561, "step": 15154 }, { "epoch": 0.5340169313847265, "grad_norm": 1.0585339069366455, "learning_rate": 9.41087184809944e-06, "loss": 1.12, "step": 15155 }, { "epoch": 0.5340521683976849, "grad_norm": 4.210237979888916, "learning_rate": 9.409732557383829e-06, "loss": 3.2618, "step": 15156 }, { "epoch": 0.5340874054106434, "grad_norm": 1.1760283708572388, "learning_rate": 9.408593274356527e-06, "loss": 1.0576, "step": 15157 }, { "epoch": 0.5341226424236017, "grad_norm": 5.973382949829102, "learning_rate": 9.407453999032375e-06, "loss": 2.769, "step": 15158 }, { "epoch": 0.5341578794365601, "grad_norm": 5.928293228149414, "learning_rate": 9.406314731426208e-06, "loss": 5.0071, "step": 15159 }, { "epoch": 0.5341931164495186, "grad_norm": 3.9107422828674316, "learning_rate": 9.405175471552865e-06, "loss": 3.3025, "step": 15160 }, { "epoch": 0.534228353462477, "grad_norm": 11.754813194274902, "learning_rate": 9.404036219427195e-06, "loss": 5.2322, "step": 15161 }, { "epoch": 0.5342635904754354, "grad_norm": 13.938386917114258, "learning_rate": 9.402896975064022e-06, "loss": 4.9797, "step": 15162 }, { "epoch": 0.5342988274883939, "grad_norm": 4.305731773376465, "learning_rate": 9.401757738478193e-06, "loss": 2.9189, "step": 15163 }, { "epoch": 0.5343340645013522, "grad_norm": 3.270551919937134, "learning_rate": 9.40061850968455e-06, "loss": 3.0104, "step": 15164 }, { "epoch": 0.5343693015143106, "grad_norm": 4.180139541625977, "learning_rate": 9.399479288697924e-06, "loss": 2.8267, "step": 15165 }, { "epoch": 0.534404538527269, "grad_norm": 0.958753228187561, "learning_rate": 9.398340075533154e-06, "loss": 1.1641, "step": 15166 }, { "epoch": 0.5344397755402275, "grad_norm": 6.380086898803711, "learning_rate": 9.397200870205084e-06, "loss": 5.8446, "step": 15167 }, { "epoch": 0.5344750125531859, "grad_norm": 1.2174631357192993, "learning_rate": 9.396061672728552e-06, "loss": 0.8682, "step": 15168 }, { "epoch": 0.5345102495661442, "grad_norm": 1.3211429119110107, "learning_rate": 9.39492248311839e-06, "loss": 0.8193, "step": 15169 }, { "epoch": 0.5345454865791027, "grad_norm": 4.289778709411621, "learning_rate": 9.393783301389439e-06, "loss": 0.7078, "step": 15170 }, { "epoch": 0.5345807235920611, "grad_norm": 2.39963960647583, "learning_rate": 9.39264412755654e-06, "loss": 0.7241, "step": 15171 }, { "epoch": 0.5346159606050195, "grad_norm": 12.480352401733398, "learning_rate": 9.391504961634529e-06, "loss": 6.9712, "step": 15172 }, { "epoch": 0.534651197617978, "grad_norm": 3.747459650039673, "learning_rate": 9.390365803638242e-06, "loss": 2.8983, "step": 15173 }, { "epoch": 0.5346864346309363, "grad_norm": 1.9046931266784668, "learning_rate": 9.38922665358252e-06, "loss": 1.0297, "step": 15174 }, { "epoch": 0.5347216716438947, "grad_norm": 1.8495643138885498, "learning_rate": 9.388087511482196e-06, "loss": 0.6688, "step": 15175 }, { "epoch": 0.5347569086568532, "grad_norm": 1.7127422094345093, "learning_rate": 9.38694837735211e-06, "loss": 0.7466, "step": 15176 }, { "epoch": 0.5347921456698116, "grad_norm": 3.4664061069488525, "learning_rate": 9.385809251207103e-06, "loss": 0.668, "step": 15177 }, { "epoch": 0.53482738268277, "grad_norm": 9.035918235778809, "learning_rate": 9.384670133062006e-06, "loss": 5.6228, "step": 15178 }, { "epoch": 0.5348626196957283, "grad_norm": 2.0004138946533203, "learning_rate": 9.38353102293166e-06, "loss": 0.7846, "step": 15179 }, { "epoch": 0.5348978567086868, "grad_norm": 2.938403844833374, "learning_rate": 9.382391920830902e-06, "loss": 3.0124, "step": 15180 }, { "epoch": 0.5349330937216452, "grad_norm": 19.513294219970703, "learning_rate": 9.381252826774563e-06, "loss": 2.9023, "step": 15181 }, { "epoch": 0.5349683307346036, "grad_norm": 4.897764205932617, "learning_rate": 9.38011374077749e-06, "loss": 3.3629, "step": 15182 }, { "epoch": 0.5350035677475621, "grad_norm": 1.379518747329712, "learning_rate": 9.378974662854513e-06, "loss": 0.7403, "step": 15183 }, { "epoch": 0.5350388047605205, "grad_norm": 2.6120262145996094, "learning_rate": 9.37783559302047e-06, "loss": 2.8452, "step": 15184 }, { "epoch": 0.5350740417734788, "grad_norm": 22.532133102416992, "learning_rate": 9.376696531290195e-06, "loss": 3.0358, "step": 15185 }, { "epoch": 0.5351092787864373, "grad_norm": 6.1793107986450195, "learning_rate": 9.375557477678528e-06, "loss": 3.318, "step": 15186 }, { "epoch": 0.5351445157993957, "grad_norm": 3.954078197479248, "learning_rate": 9.374418432200308e-06, "loss": 3.094, "step": 15187 }, { "epoch": 0.5351797528123541, "grad_norm": 1.6341639757156372, "learning_rate": 9.373279394870365e-06, "loss": 0.9733, "step": 15188 }, { "epoch": 0.5352149898253125, "grad_norm": 4.179924011230469, "learning_rate": 9.372140365703536e-06, "loss": 3.5605, "step": 15189 }, { "epoch": 0.535250226838271, "grad_norm": 2.7855801582336426, "learning_rate": 9.371001344714666e-06, "loss": 0.9712, "step": 15190 }, { "epoch": 0.5352854638512293, "grad_norm": 3.9334640502929688, "learning_rate": 9.369862331918574e-06, "loss": 2.7771, "step": 15191 }, { "epoch": 0.5353207008641877, "grad_norm": 1.3796124458312988, "learning_rate": 9.368723327330111e-06, "loss": 1.0903, "step": 15192 }, { "epoch": 0.5353559378771462, "grad_norm": 1.5262292623519897, "learning_rate": 9.367584330964107e-06, "loss": 0.8657, "step": 15193 }, { "epoch": 0.5353911748901046, "grad_norm": 4.643296241760254, "learning_rate": 9.366445342835397e-06, "loss": 3.6971, "step": 15194 }, { "epoch": 0.535426411903063, "grad_norm": 1.1017664670944214, "learning_rate": 9.365306362958813e-06, "loss": 0.9961, "step": 15195 }, { "epoch": 0.5354616489160214, "grad_norm": 20.928661346435547, "learning_rate": 9.3641673913492e-06, "loss": 3.4812, "step": 15196 }, { "epoch": 0.5354968859289798, "grad_norm": 1.711763858795166, "learning_rate": 9.363028428021386e-06, "loss": 0.7938, "step": 15197 }, { "epoch": 0.5355321229419382, "grad_norm": 1.6400054693222046, "learning_rate": 9.361889472990207e-06, "loss": 0.9214, "step": 15198 }, { "epoch": 0.5355673599548966, "grad_norm": 6.324028015136719, "learning_rate": 9.360750526270503e-06, "loss": 3.3043, "step": 15199 }, { "epoch": 0.5356025969678551, "grad_norm": 1.4047788381576538, "learning_rate": 9.3596115878771e-06, "loss": 0.6861, "step": 15200 }, { "epoch": 0.5356378339808134, "grad_norm": 4.502621173858643, "learning_rate": 9.358472657824837e-06, "loss": 3.1689, "step": 15201 }, { "epoch": 0.5356730709937718, "grad_norm": 4.646291255950928, "learning_rate": 9.357333736128553e-06, "loss": 3.0605, "step": 15202 }, { "epoch": 0.5357083080067303, "grad_norm": 8.918757438659668, "learning_rate": 9.356194822803077e-06, "loss": 2.7328, "step": 15203 }, { "epoch": 0.5357435450196887, "grad_norm": 2.0610909461975098, "learning_rate": 9.355055917863243e-06, "loss": 2.0207, "step": 15204 }, { "epoch": 0.5357787820326471, "grad_norm": 4.338530540466309, "learning_rate": 9.353917021323892e-06, "loss": 2.6463, "step": 15205 }, { "epoch": 0.5358140190456055, "grad_norm": 3.1941077709198, "learning_rate": 9.352778133199854e-06, "loss": 2.7529, "step": 15206 }, { "epoch": 0.5358492560585639, "grad_norm": 2.222895860671997, "learning_rate": 9.351639253505962e-06, "loss": 0.8319, "step": 15207 }, { "epoch": 0.5358844930715223, "grad_norm": 4.275622367858887, "learning_rate": 9.350500382257048e-06, "loss": 2.7038, "step": 15208 }, { "epoch": 0.5359197300844807, "grad_norm": 5.064227104187012, "learning_rate": 9.349361519467958e-06, "loss": 3.2202, "step": 15209 }, { "epoch": 0.5359549670974392, "grad_norm": 4.315766334533691, "learning_rate": 9.348222665153508e-06, "loss": 3.3887, "step": 15210 }, { "epoch": 0.5359902041103976, "grad_norm": 2.6929585933685303, "learning_rate": 9.347083819328543e-06, "loss": 2.5744, "step": 15211 }, { "epoch": 0.5360254411233559, "grad_norm": 3.6095430850982666, "learning_rate": 9.3459449820079e-06, "loss": 3.0912, "step": 15212 }, { "epoch": 0.5360606781363144, "grad_norm": 5.3035888671875, "learning_rate": 9.344806153206402e-06, "loss": 4.8463, "step": 15213 }, { "epoch": 0.5360959151492728, "grad_norm": 2.4079322814941406, "learning_rate": 9.343667332938887e-06, "loss": 0.8102, "step": 15214 }, { "epoch": 0.5361311521622312, "grad_norm": 6.216043472290039, "learning_rate": 9.34252852122019e-06, "loss": 4.8062, "step": 15215 }, { "epoch": 0.5361663891751897, "grad_norm": 9.161953926086426, "learning_rate": 9.341389718065142e-06, "loss": 5.1769, "step": 15216 }, { "epoch": 0.536201626188148, "grad_norm": 7.917092800140381, "learning_rate": 9.340250923488578e-06, "loss": 6.8722, "step": 15217 }, { "epoch": 0.5362368632011064, "grad_norm": 10.538174629211426, "learning_rate": 9.339112137505331e-06, "loss": 5.3911, "step": 15218 }, { "epoch": 0.5362721002140649, "grad_norm": 0.9424600005149841, "learning_rate": 9.337973360130227e-06, "loss": 1.0367, "step": 15219 }, { "epoch": 0.5363073372270233, "grad_norm": 1.325020670890808, "learning_rate": 9.336834591378107e-06, "loss": 0.8184, "step": 15220 }, { "epoch": 0.5363425742399817, "grad_norm": 3.071591854095459, "learning_rate": 9.335695831263805e-06, "loss": 2.4877, "step": 15221 }, { "epoch": 0.53637781125294, "grad_norm": 3.6543750762939453, "learning_rate": 9.334557079802146e-06, "loss": 3.284, "step": 15222 }, { "epoch": 0.5364130482658985, "grad_norm": 4.422588348388672, "learning_rate": 9.333418337007964e-06, "loss": 3.5703, "step": 15223 }, { "epoch": 0.5364482852788569, "grad_norm": 1.2768357992172241, "learning_rate": 9.332279602896095e-06, "loss": 1.2623, "step": 15224 }, { "epoch": 0.5364835222918153, "grad_norm": 4.9934611320495605, "learning_rate": 9.33114087748137e-06, "loss": 4.5721, "step": 15225 }, { "epoch": 0.5365187593047738, "grad_norm": 3.645038366317749, "learning_rate": 9.330002160778621e-06, "loss": 2.4196, "step": 15226 }, { "epoch": 0.5365539963177322, "grad_norm": 1.656311273574829, "learning_rate": 9.328863452802675e-06, "loss": 0.8076, "step": 15227 }, { "epoch": 0.5365892333306905, "grad_norm": 1.3809820413589478, "learning_rate": 9.327724753568375e-06, "loss": 1.1833, "step": 15228 }, { "epoch": 0.536624470343649, "grad_norm": 2.5962536334991455, "learning_rate": 9.326586063090539e-06, "loss": 0.798, "step": 15229 }, { "epoch": 0.5366597073566074, "grad_norm": 0.765507161617279, "learning_rate": 9.325447381384007e-06, "loss": 0.7525, "step": 15230 }, { "epoch": 0.5366949443695658, "grad_norm": 1.097687840461731, "learning_rate": 9.324308708463612e-06, "loss": 1.0148, "step": 15231 }, { "epoch": 0.5367301813825242, "grad_norm": 12.74608039855957, "learning_rate": 9.32317004434418e-06, "loss": 5.1487, "step": 15232 }, { "epoch": 0.5367654183954826, "grad_norm": 1.0050653219223022, "learning_rate": 9.322031389040542e-06, "loss": 1.0698, "step": 15233 }, { "epoch": 0.536800655408441, "grad_norm": 2.424337148666382, "learning_rate": 9.320892742567536e-06, "loss": 0.8011, "step": 15234 }, { "epoch": 0.5368358924213994, "grad_norm": 1.5990997552871704, "learning_rate": 9.319754104939985e-06, "loss": 0.8015, "step": 15235 }, { "epoch": 0.5368711294343579, "grad_norm": 1.7272330522537231, "learning_rate": 9.318615476172724e-06, "loss": 0.6164, "step": 15236 }, { "epoch": 0.5369063664473163, "grad_norm": 1.2715028524398804, "learning_rate": 9.317476856280586e-06, "loss": 0.8619, "step": 15237 }, { "epoch": 0.5369416034602746, "grad_norm": 1.184174656867981, "learning_rate": 9.316338245278395e-06, "loss": 1.0441, "step": 15238 }, { "epoch": 0.5369768404732331, "grad_norm": 1.089118480682373, "learning_rate": 9.315199643180986e-06, "loss": 0.9716, "step": 15239 }, { "epoch": 0.5370120774861915, "grad_norm": 1.4243665933609009, "learning_rate": 9.31406105000319e-06, "loss": 0.8305, "step": 15240 }, { "epoch": 0.5370473144991499, "grad_norm": 1.427249550819397, "learning_rate": 9.312922465759838e-06, "loss": 0.6881, "step": 15241 }, { "epoch": 0.5370825515121083, "grad_norm": 1.8143924474716187, "learning_rate": 9.311783890465754e-06, "loss": 1.0499, "step": 15242 }, { "epoch": 0.5371177885250668, "grad_norm": 7.053700923919678, "learning_rate": 9.310645324135775e-06, "loss": 5.9606, "step": 15243 }, { "epoch": 0.5371530255380251, "grad_norm": 35.090267181396484, "learning_rate": 9.309506766784732e-06, "loss": 5.8866, "step": 15244 }, { "epoch": 0.5371882625509835, "grad_norm": 17.370365142822266, "learning_rate": 9.308368218427447e-06, "loss": 5.5764, "step": 15245 }, { "epoch": 0.537223499563942, "grad_norm": 5.281847953796387, "learning_rate": 9.307229679078752e-06, "loss": 3.1265, "step": 15246 }, { "epoch": 0.5372587365769004, "grad_norm": 2.01196026802063, "learning_rate": 9.306091148753486e-06, "loss": 0.8237, "step": 15247 }, { "epoch": 0.5372939735898588, "grad_norm": 5.705165386199951, "learning_rate": 9.304952627466464e-06, "loss": 4.4814, "step": 15248 }, { "epoch": 0.5373292106028172, "grad_norm": 3.8617563247680664, "learning_rate": 9.303814115232526e-06, "loss": 2.3332, "step": 15249 }, { "epoch": 0.5373644476157756, "grad_norm": 1.3615314960479736, "learning_rate": 9.302675612066499e-06, "loss": 0.831, "step": 15250 }, { "epoch": 0.537399684628734, "grad_norm": 6.631632328033447, "learning_rate": 9.30153711798321e-06, "loss": 3.0994, "step": 15251 }, { "epoch": 0.5374349216416925, "grad_norm": 1.3788561820983887, "learning_rate": 9.300398632997486e-06, "loss": 1.0156, "step": 15252 }, { "epoch": 0.5374701586546509, "grad_norm": 5.1330885887146, "learning_rate": 9.299260157124163e-06, "loss": 3.4433, "step": 15253 }, { "epoch": 0.5375053956676092, "grad_norm": 3.4427711963653564, "learning_rate": 9.298121690378062e-06, "loss": 3.7707, "step": 15254 }, { "epoch": 0.5375406326805676, "grad_norm": 5.684151649475098, "learning_rate": 9.296983232774016e-06, "loss": 2.9306, "step": 15255 }, { "epoch": 0.5375758696935261, "grad_norm": 1.7857494354248047, "learning_rate": 9.295844784326858e-06, "loss": 0.8878, "step": 15256 }, { "epoch": 0.5376111067064845, "grad_norm": 2.611733913421631, "learning_rate": 9.294706345051404e-06, "loss": 3.8147, "step": 15257 }, { "epoch": 0.5376463437194429, "grad_norm": 1.7156304121017456, "learning_rate": 9.293567914962494e-06, "loss": 0.8623, "step": 15258 }, { "epoch": 0.5376815807324014, "grad_norm": 1.0767513513565063, "learning_rate": 9.29242949407495e-06, "loss": 0.829, "step": 15259 }, { "epoch": 0.5377168177453597, "grad_norm": 3.262516736984253, "learning_rate": 9.291291082403606e-06, "loss": 2.9407, "step": 15260 }, { "epoch": 0.5377520547583181, "grad_norm": 1.477100133895874, "learning_rate": 9.290152679963281e-06, "loss": 1.1614, "step": 15261 }, { "epoch": 0.5377872917712766, "grad_norm": 1.1519756317138672, "learning_rate": 9.28901428676881e-06, "loss": 0.778, "step": 15262 }, { "epoch": 0.537822528784235, "grad_norm": 37.587242126464844, "learning_rate": 9.28787590283502e-06, "loss": 3.6311, "step": 15263 }, { "epoch": 0.5378577657971934, "grad_norm": 3.671264886856079, "learning_rate": 9.286737528176735e-06, "loss": 3.0291, "step": 15264 }, { "epoch": 0.5378930028101517, "grad_norm": 1.7439793348312378, "learning_rate": 9.285599162808782e-06, "loss": 0.8482, "step": 15265 }, { "epoch": 0.5379282398231102, "grad_norm": 7.232016086578369, "learning_rate": 9.284460806745998e-06, "loss": 5.8514, "step": 15266 }, { "epoch": 0.5379634768360686, "grad_norm": 3.9756062030792236, "learning_rate": 9.283322460003197e-06, "loss": 2.8759, "step": 15267 }, { "epoch": 0.537998713849027, "grad_norm": 1.3983067274093628, "learning_rate": 9.282184122595214e-06, "loss": 0.7787, "step": 15268 }, { "epoch": 0.5380339508619855, "grad_norm": 4.068305492401123, "learning_rate": 9.281045794536875e-06, "loss": 2.8705, "step": 15269 }, { "epoch": 0.5380691878749438, "grad_norm": 1.5238189697265625, "learning_rate": 9.279907475843006e-06, "loss": 0.7339, "step": 15270 }, { "epoch": 0.5381044248879022, "grad_norm": 1.2326254844665527, "learning_rate": 9.278769166528433e-06, "loss": 0.7439, "step": 15271 }, { "epoch": 0.5381396619008607, "grad_norm": 1.4308524131774902, "learning_rate": 9.277630866607983e-06, "loss": 0.8788, "step": 15272 }, { "epoch": 0.5381748989138191, "grad_norm": 0.9321812391281128, "learning_rate": 9.276492576096486e-06, "loss": 0.9043, "step": 15273 }, { "epoch": 0.5382101359267775, "grad_norm": 1.5209007263183594, "learning_rate": 9.275354295008762e-06, "loss": 0.9274, "step": 15274 }, { "epoch": 0.5382453729397358, "grad_norm": 1.179280400276184, "learning_rate": 9.274216023359643e-06, "loss": 0.707, "step": 15275 }, { "epoch": 0.5382806099526943, "grad_norm": 4.844753265380859, "learning_rate": 9.27307776116395e-06, "loss": 3.1385, "step": 15276 }, { "epoch": 0.5383158469656527, "grad_norm": 1.368442177772522, "learning_rate": 9.271939508436512e-06, "loss": 1.0345, "step": 15277 }, { "epoch": 0.5383510839786111, "grad_norm": 15.135926246643066, "learning_rate": 9.270801265192157e-06, "loss": 5.4973, "step": 15278 }, { "epoch": 0.5383863209915696, "grad_norm": 1.5065540075302124, "learning_rate": 9.26966303144571e-06, "loss": 0.819, "step": 15279 }, { "epoch": 0.538421558004528, "grad_norm": 1.7992503643035889, "learning_rate": 9.268524807211989e-06, "loss": 1.0684, "step": 15280 }, { "epoch": 0.5384567950174863, "grad_norm": 12.07469654083252, "learning_rate": 9.267386592505829e-06, "loss": 7.077, "step": 15281 }, { "epoch": 0.5384920320304448, "grad_norm": 1.4791339635849, "learning_rate": 9.266248387342054e-06, "loss": 0.8238, "step": 15282 }, { "epoch": 0.5385272690434032, "grad_norm": 4.161824703216553, "learning_rate": 9.265110191735485e-06, "loss": 2.7916, "step": 15283 }, { "epoch": 0.5385625060563616, "grad_norm": 9.600191116333008, "learning_rate": 9.263972005700947e-06, "loss": 5.2567, "step": 15284 }, { "epoch": 0.5385977430693201, "grad_norm": 1.7061762809753418, "learning_rate": 9.262833829253274e-06, "loss": 0.9629, "step": 15285 }, { "epoch": 0.5386329800822784, "grad_norm": 1.191473126411438, "learning_rate": 9.261695662407277e-06, "loss": 0.9425, "step": 15286 }, { "epoch": 0.5386682170952368, "grad_norm": 4.070258617401123, "learning_rate": 9.260557505177792e-06, "loss": 3.0829, "step": 15287 }, { "epoch": 0.5387034541081952, "grad_norm": 1.5380414724349976, "learning_rate": 9.259419357579641e-06, "loss": 0.8501, "step": 15288 }, { "epoch": 0.5387386911211537, "grad_norm": 6.424943447113037, "learning_rate": 9.258281219627644e-06, "loss": 3.1369, "step": 15289 }, { "epoch": 0.5387739281341121, "grad_norm": 1.6049792766571045, "learning_rate": 9.257143091336627e-06, "loss": 0.9462, "step": 15290 }, { "epoch": 0.5388091651470704, "grad_norm": 1.5890847444534302, "learning_rate": 9.25600497272142e-06, "loss": 0.7973, "step": 15291 }, { "epoch": 0.5388444021600289, "grad_norm": 10.576577186584473, "learning_rate": 9.254866863796843e-06, "loss": 3.2266, "step": 15292 }, { "epoch": 0.5388796391729873, "grad_norm": 1.1130667924880981, "learning_rate": 9.253728764577718e-06, "loss": 0.7382, "step": 15293 }, { "epoch": 0.5389148761859457, "grad_norm": 1.248537540435791, "learning_rate": 9.252590675078873e-06, "loss": 1.083, "step": 15294 }, { "epoch": 0.5389501131989042, "grad_norm": 1.1059062480926514, "learning_rate": 9.251452595315126e-06, "loss": 0.9169, "step": 15295 }, { "epoch": 0.5389853502118626, "grad_norm": 1.584962248802185, "learning_rate": 9.250314525301306e-06, "loss": 0.9711, "step": 15296 }, { "epoch": 0.5390205872248209, "grad_norm": 3.055879592895508, "learning_rate": 9.249176465052236e-06, "loss": 3.0433, "step": 15297 }, { "epoch": 0.5390558242377793, "grad_norm": 5.1728196144104, "learning_rate": 9.248038414582738e-06, "loss": 3.2216, "step": 15298 }, { "epoch": 0.5390910612507378, "grad_norm": 7.178133964538574, "learning_rate": 9.246900373907634e-06, "loss": 5.1245, "step": 15299 }, { "epoch": 0.5391262982636962, "grad_norm": 1.328956961631775, "learning_rate": 9.245762343041748e-06, "loss": 0.9296, "step": 15300 }, { "epoch": 0.5391615352766546, "grad_norm": 4.050919055938721, "learning_rate": 9.244624321999907e-06, "loss": 2.7602, "step": 15301 }, { "epoch": 0.539196772289613, "grad_norm": 5.02921724319458, "learning_rate": 9.243486310796928e-06, "loss": 5.6368, "step": 15302 }, { "epoch": 0.5392320093025714, "grad_norm": 3.8143882751464844, "learning_rate": 9.242348309447633e-06, "loss": 3.233, "step": 15303 }, { "epoch": 0.5392672463155298, "grad_norm": 20.37710952758789, "learning_rate": 9.241210317966857e-06, "loss": 5.9498, "step": 15304 }, { "epoch": 0.5393024833284883, "grad_norm": 5.7735185623168945, "learning_rate": 9.240072336369406e-06, "loss": 5.4373, "step": 15305 }, { "epoch": 0.5393377203414467, "grad_norm": 5.68397855758667, "learning_rate": 9.23893436467011e-06, "loss": 5.6739, "step": 15306 }, { "epoch": 0.539372957354405, "grad_norm": 4.6755876541137695, "learning_rate": 9.237796402883794e-06, "loss": 2.9963, "step": 15307 }, { "epoch": 0.5394081943673634, "grad_norm": 4.835494041442871, "learning_rate": 9.236658451025276e-06, "loss": 2.9795, "step": 15308 }, { "epoch": 0.5394434313803219, "grad_norm": 1.9752726554870605, "learning_rate": 9.235520509109374e-06, "loss": 0.9368, "step": 15309 }, { "epoch": 0.5394786683932803, "grad_norm": 6.9703497886657715, "learning_rate": 9.234382577150922e-06, "loss": 2.8992, "step": 15310 }, { "epoch": 0.5395139054062387, "grad_norm": 3.3278191089630127, "learning_rate": 9.23324465516473e-06, "loss": 2.884, "step": 15311 }, { "epoch": 0.5395491424191972, "grad_norm": 16.35795783996582, "learning_rate": 9.232106743165626e-06, "loss": 3.306, "step": 15312 }, { "epoch": 0.5395843794321555, "grad_norm": 1.7483747005462646, "learning_rate": 9.230968841168426e-06, "loss": 0.8963, "step": 15313 }, { "epoch": 0.5396196164451139, "grad_norm": 4.223227500915527, "learning_rate": 9.22983094918796e-06, "loss": 4.9, "step": 15314 }, { "epoch": 0.5396548534580724, "grad_norm": 2.117154121398926, "learning_rate": 9.22869306723904e-06, "loss": 1.1204, "step": 15315 }, { "epoch": 0.5396900904710308, "grad_norm": 6.995985984802246, "learning_rate": 9.227555195336492e-06, "loss": 3.4832, "step": 15316 }, { "epoch": 0.5397253274839892, "grad_norm": 1.3477578163146973, "learning_rate": 9.226417333495139e-06, "loss": 0.8463, "step": 15317 }, { "epoch": 0.5397605644969476, "grad_norm": 1.611397624015808, "learning_rate": 9.225279481729793e-06, "loss": 0.9338, "step": 15318 }, { "epoch": 0.539795801509906, "grad_norm": 1.1276062726974487, "learning_rate": 9.224141640055284e-06, "loss": 0.7714, "step": 15319 }, { "epoch": 0.5398310385228644, "grad_norm": 1.530202865600586, "learning_rate": 9.22300380848643e-06, "loss": 0.9408, "step": 15320 }, { "epoch": 0.5398662755358228, "grad_norm": 5.067382335662842, "learning_rate": 9.221865987038049e-06, "loss": 2.6702, "step": 15321 }, { "epoch": 0.5399015125487813, "grad_norm": 1.3559436798095703, "learning_rate": 9.22072817572496e-06, "loss": 1.0765, "step": 15322 }, { "epoch": 0.5399367495617396, "grad_norm": 9.585716247558594, "learning_rate": 9.219590374561993e-06, "loss": 5.5265, "step": 15323 }, { "epoch": 0.539971986574698, "grad_norm": 9.118500709533691, "learning_rate": 9.218452583563954e-06, "loss": 3.0415, "step": 15324 }, { "epoch": 0.5400072235876565, "grad_norm": 1.18899667263031, "learning_rate": 9.217314802745671e-06, "loss": 1.178, "step": 15325 }, { "epoch": 0.5400424606006149, "grad_norm": 1.8298002481460571, "learning_rate": 9.216177032121966e-06, "loss": 1.0119, "step": 15326 }, { "epoch": 0.5400776976135733, "grad_norm": 36.9577751159668, "learning_rate": 9.215039271707652e-06, "loss": 2.8321, "step": 15327 }, { "epoch": 0.5401129346265318, "grad_norm": 1.9409701824188232, "learning_rate": 9.213901521517548e-06, "loss": 0.936, "step": 15328 }, { "epoch": 0.5401481716394901, "grad_norm": 6.277939796447754, "learning_rate": 9.212763781566482e-06, "loss": 5.2067, "step": 15329 }, { "epoch": 0.5401834086524485, "grad_norm": 1.4322859048843384, "learning_rate": 9.211626051869268e-06, "loss": 0.936, "step": 15330 }, { "epoch": 0.5402186456654069, "grad_norm": 4.129067897796631, "learning_rate": 9.210488332440723e-06, "loss": 2.5855, "step": 15331 }, { "epoch": 0.5402538826783654, "grad_norm": 1.1284300088882446, "learning_rate": 9.209350623295666e-06, "loss": 0.7417, "step": 15332 }, { "epoch": 0.5402891196913238, "grad_norm": 1.9205982685089111, "learning_rate": 9.208212924448923e-06, "loss": 0.9228, "step": 15333 }, { "epoch": 0.5403243567042821, "grad_norm": 1.4044189453125, "learning_rate": 9.207075235915305e-06, "loss": 0.7414, "step": 15334 }, { "epoch": 0.5403595937172406, "grad_norm": 3.5885162353515625, "learning_rate": 9.205937557709632e-06, "loss": 2.7185, "step": 15335 }, { "epoch": 0.540394830730199, "grad_norm": 7.076998710632324, "learning_rate": 9.204799889846727e-06, "loss": 3.159, "step": 15336 }, { "epoch": 0.5404300677431574, "grad_norm": 2.70914888381958, "learning_rate": 9.2036622323414e-06, "loss": 1.1051, "step": 15337 }, { "epoch": 0.5404653047561159, "grad_norm": 7.143411159515381, "learning_rate": 9.202524585208476e-06, "loss": 3.1019, "step": 15338 }, { "epoch": 0.5405005417690742, "grad_norm": 1.3107550144195557, "learning_rate": 9.201386948462772e-06, "loss": 1.0528, "step": 15339 }, { "epoch": 0.5405357787820326, "grad_norm": 7.377230167388916, "learning_rate": 9.200249322119104e-06, "loss": 3.1654, "step": 15340 }, { "epoch": 0.540571015794991, "grad_norm": 1.7967368364334106, "learning_rate": 9.19911170619229e-06, "loss": 0.6654, "step": 15341 }, { "epoch": 0.5406062528079495, "grad_norm": 1.6125283241271973, "learning_rate": 9.19797410069715e-06, "loss": 0.8373, "step": 15342 }, { "epoch": 0.5406414898209079, "grad_norm": 1.005469560623169, "learning_rate": 9.196836505648495e-06, "loss": 1.0012, "step": 15343 }, { "epoch": 0.5406767268338663, "grad_norm": 4.986631393432617, "learning_rate": 9.19569892106115e-06, "loss": 2.6293, "step": 15344 }, { "epoch": 0.5407119638468247, "grad_norm": 7.353647232055664, "learning_rate": 9.19456134694993e-06, "loss": 3.3193, "step": 15345 }, { "epoch": 0.5407472008597831, "grad_norm": 3.885101795196533, "learning_rate": 9.19342378332965e-06, "loss": 2.7797, "step": 15346 }, { "epoch": 0.5407824378727415, "grad_norm": 1.2478270530700684, "learning_rate": 9.192286230215125e-06, "loss": 0.9045, "step": 15347 }, { "epoch": 0.5408176748857, "grad_norm": 6.351255416870117, "learning_rate": 9.191148687621183e-06, "loss": 3.8261, "step": 15348 }, { "epoch": 0.5408529118986584, "grad_norm": 1.2085334062576294, "learning_rate": 9.190011155562622e-06, "loss": 0.7864, "step": 15349 }, { "epoch": 0.5408881489116167, "grad_norm": 5.964612007141113, "learning_rate": 9.188873634054274e-06, "loss": 3.1626, "step": 15350 }, { "epoch": 0.5409233859245752, "grad_norm": 6.449562072753906, "learning_rate": 9.187736123110947e-06, "loss": 5.2009, "step": 15351 }, { "epoch": 0.5409586229375336, "grad_norm": 5.410470008850098, "learning_rate": 9.186598622747468e-06, "loss": 3.147, "step": 15352 }, { "epoch": 0.540993859950492, "grad_norm": 5.410470008850098, "learning_rate": 9.186598622747468e-06, "loss": 3.4303, "step": 15353 }, { "epoch": 0.5410290969634504, "grad_norm": 1.212544560432434, "learning_rate": 9.185461132978639e-06, "loss": 0.7459, "step": 15354 }, { "epoch": 0.5410643339764089, "grad_norm": 1.7428832054138184, "learning_rate": 9.184323653819283e-06, "loss": 1.0937, "step": 15355 }, { "epoch": 0.5410995709893672, "grad_norm": 14.18539047241211, "learning_rate": 9.183186185284219e-06, "loss": 5.308, "step": 15356 }, { "epoch": 0.5411348080023256, "grad_norm": 2.327709913253784, "learning_rate": 9.182048727388257e-06, "loss": 0.7743, "step": 15357 }, { "epoch": 0.5411700450152841, "grad_norm": 9.573830604553223, "learning_rate": 9.18091128014621e-06, "loss": 5.31, "step": 15358 }, { "epoch": 0.5412052820282425, "grad_norm": 5.689017295837402, "learning_rate": 9.179773843572905e-06, "loss": 3.03, "step": 15359 }, { "epoch": 0.5412405190412009, "grad_norm": 5.254258155822754, "learning_rate": 9.178636417683146e-06, "loss": 5.4959, "step": 15360 }, { "epoch": 0.5412757560541593, "grad_norm": 1.469033122062683, "learning_rate": 9.177499002491754e-06, "loss": 0.7924, "step": 15361 }, { "epoch": 0.5413109930671177, "grad_norm": 6.54274845123291, "learning_rate": 9.176361598013544e-06, "loss": 5.3576, "step": 15362 }, { "epoch": 0.5413462300800761, "grad_norm": 33.054290771484375, "learning_rate": 9.175224204263324e-06, "loss": 3.1175, "step": 15363 }, { "epoch": 0.5413814670930345, "grad_norm": 1.6845614910125732, "learning_rate": 9.174086821255916e-06, "loss": 1.0756, "step": 15364 }, { "epoch": 0.541416704105993, "grad_norm": 3.8473286628723145, "learning_rate": 9.172949449006135e-06, "loss": 2.8606, "step": 15365 }, { "epoch": 0.5414519411189513, "grad_norm": 1.5526679754257202, "learning_rate": 9.17181208752879e-06, "loss": 0.946, "step": 15366 }, { "epoch": 0.5414871781319097, "grad_norm": 6.048658847808838, "learning_rate": 9.170674736838696e-06, "loss": 3.2835, "step": 15367 }, { "epoch": 0.5415224151448682, "grad_norm": 3.3674323558807373, "learning_rate": 9.169537396950678e-06, "loss": 0.6296, "step": 15368 }, { "epoch": 0.5415576521578266, "grad_norm": 4.466602802276611, "learning_rate": 9.168400067879533e-06, "loss": 2.5682, "step": 15369 }, { "epoch": 0.541592889170785, "grad_norm": 3.209280252456665, "learning_rate": 9.167262749640085e-06, "loss": 2.472, "step": 15370 }, { "epoch": 0.5416281261837435, "grad_norm": 3.96401309967041, "learning_rate": 9.166125442247144e-06, "loss": 2.8746, "step": 15371 }, { "epoch": 0.5416633631967018, "grad_norm": 2.256990909576416, "learning_rate": 9.16498814571553e-06, "loss": 0.7695, "step": 15372 }, { "epoch": 0.5416986002096602, "grad_norm": 24.031803131103516, "learning_rate": 9.163850860060047e-06, "loss": 2.5549, "step": 15373 }, { "epoch": 0.5417338372226186, "grad_norm": 2.7658064365386963, "learning_rate": 9.162713585295515e-06, "loss": 2.2769, "step": 15374 }, { "epoch": 0.5417690742355771, "grad_norm": 5.1841278076171875, "learning_rate": 9.161576321436747e-06, "loss": 5.1421, "step": 15375 }, { "epoch": 0.5418043112485355, "grad_norm": 3.4379303455352783, "learning_rate": 9.160439068498553e-06, "loss": 2.3861, "step": 15376 }, { "epoch": 0.5418395482614938, "grad_norm": 5.395273208618164, "learning_rate": 9.159301826495744e-06, "loss": 5.0105, "step": 15377 }, { "epoch": 0.5418747852744523, "grad_norm": 3.0854172706604004, "learning_rate": 9.158164595443142e-06, "loss": 2.4989, "step": 15378 }, { "epoch": 0.5419100222874107, "grad_norm": 1.2827465534210205, "learning_rate": 9.157027375355548e-06, "loss": 0.8943, "step": 15379 }, { "epoch": 0.5419452593003691, "grad_norm": 1.125152587890625, "learning_rate": 9.155890166247783e-06, "loss": 0.6376, "step": 15380 }, { "epoch": 0.5419804963133276, "grad_norm": 1.363959789276123, "learning_rate": 9.154752968134655e-06, "loss": 1.1222, "step": 15381 }, { "epoch": 0.5420157333262859, "grad_norm": 1.2110469341278076, "learning_rate": 9.153615781030974e-06, "loss": 0.8764, "step": 15382 }, { "epoch": 0.5420509703392443, "grad_norm": 4.8005194664001465, "learning_rate": 9.152478604951558e-06, "loss": 5.8736, "step": 15383 }, { "epoch": 0.5420862073522028, "grad_norm": 3.9619202613830566, "learning_rate": 9.151341439911219e-06, "loss": 3.0269, "step": 15384 }, { "epoch": 0.5421214443651612, "grad_norm": 13.640351295471191, "learning_rate": 9.150204285924764e-06, "loss": 4.9119, "step": 15385 }, { "epoch": 0.5421566813781196, "grad_norm": 6.771047115325928, "learning_rate": 9.149067143007002e-06, "loss": 4.9334, "step": 15386 }, { "epoch": 0.5421919183910779, "grad_norm": 6.258583068847656, "learning_rate": 9.147930011172754e-06, "loss": 5.458, "step": 15387 }, { "epoch": 0.5422271554040364, "grad_norm": 5.843844890594482, "learning_rate": 9.146792890436825e-06, "loss": 3.4029, "step": 15388 }, { "epoch": 0.5422623924169948, "grad_norm": 4.46674919128418, "learning_rate": 9.145655780814027e-06, "loss": 3.1254, "step": 15389 }, { "epoch": 0.5422976294299532, "grad_norm": 10.029172897338867, "learning_rate": 9.144518682319169e-06, "loss": 7.0868, "step": 15390 }, { "epoch": 0.5423328664429117, "grad_norm": 1.5771344900131226, "learning_rate": 9.14338159496707e-06, "loss": 0.9498, "step": 15391 }, { "epoch": 0.54236810345587, "grad_norm": 4.388433933258057, "learning_rate": 9.142244518772529e-06, "loss": 2.3683, "step": 15392 }, { "epoch": 0.5424033404688284, "grad_norm": 9.266006469726562, "learning_rate": 9.141107453750365e-06, "loss": 3.5547, "step": 15393 }, { "epoch": 0.5424385774817869, "grad_norm": 2.487577199935913, "learning_rate": 9.139970399915388e-06, "loss": 0.8, "step": 15394 }, { "epoch": 0.5424738144947453, "grad_norm": 5.430992126464844, "learning_rate": 9.138833357282404e-06, "loss": 3.1933, "step": 15395 }, { "epoch": 0.5425090515077037, "grad_norm": 11.171091079711914, "learning_rate": 9.13769632586622e-06, "loss": 2.8463, "step": 15396 }, { "epoch": 0.542544288520662, "grad_norm": 5.257277011871338, "learning_rate": 9.136559305681657e-06, "loss": 3.0331, "step": 15397 }, { "epoch": 0.5425795255336205, "grad_norm": 8.13863754272461, "learning_rate": 9.13542229674352e-06, "loss": 5.498, "step": 15398 }, { "epoch": 0.5426147625465789, "grad_norm": 5.0154128074646, "learning_rate": 9.134285299066614e-06, "loss": 3.1791, "step": 15399 }, { "epoch": 0.5426499995595373, "grad_norm": 3.0753331184387207, "learning_rate": 9.133148312665757e-06, "loss": 2.781, "step": 15400 }, { "epoch": 0.5426852365724958, "grad_norm": 3.4620120525360107, "learning_rate": 9.132011337555748e-06, "loss": 2.8519, "step": 15401 }, { "epoch": 0.5427204735854542, "grad_norm": 2.3018739223480225, "learning_rate": 9.130874373751404e-06, "loss": 0.8407, "step": 15402 }, { "epoch": 0.5427557105984125, "grad_norm": 6.408244609832764, "learning_rate": 9.129737421267535e-06, "loss": 5.1302, "step": 15403 }, { "epoch": 0.542790947611371, "grad_norm": 3.162473678588867, "learning_rate": 9.128600480118943e-06, "loss": 3.191, "step": 15404 }, { "epoch": 0.5428261846243294, "grad_norm": 6.071486473083496, "learning_rate": 9.12746355032044e-06, "loss": 3.2716, "step": 15405 }, { "epoch": 0.5428614216372878, "grad_norm": 5.761983394622803, "learning_rate": 9.126326631886837e-06, "loss": 4.7175, "step": 15406 }, { "epoch": 0.5428966586502462, "grad_norm": 4.835844039916992, "learning_rate": 9.125189724832944e-06, "loss": 5.1863, "step": 15407 }, { "epoch": 0.5429318956632047, "grad_norm": 2.3401503562927246, "learning_rate": 9.124052829173564e-06, "loss": 0.9137, "step": 15408 }, { "epoch": 0.542967132676163, "grad_norm": 6.616344928741455, "learning_rate": 9.122915944923505e-06, "loss": 5.0431, "step": 15409 }, { "epoch": 0.5430023696891214, "grad_norm": 2.865110158920288, "learning_rate": 9.121779072097583e-06, "loss": 2.9537, "step": 15410 }, { "epoch": 0.5430376067020799, "grad_norm": 3.6573781967163086, "learning_rate": 9.120642210710595e-06, "loss": 3.0274, "step": 15411 }, { "epoch": 0.5430728437150383, "grad_norm": 2.2647039890289307, "learning_rate": 9.119505360777357e-06, "loss": 0.8797, "step": 15412 }, { "epoch": 0.5431080807279967, "grad_norm": 7.280066013336182, "learning_rate": 9.118368522312675e-06, "loss": 5.3755, "step": 15413 }, { "epoch": 0.5431433177409551, "grad_norm": 1.2723780870437622, "learning_rate": 9.117231695331352e-06, "loss": 1.0194, "step": 15414 }, { "epoch": 0.5431785547539135, "grad_norm": 4.846107482910156, "learning_rate": 9.116094879848199e-06, "loss": 5.196, "step": 15415 }, { "epoch": 0.5432137917668719, "grad_norm": 6.248170852661133, "learning_rate": 9.114958075878025e-06, "loss": 2.4114, "step": 15416 }, { "epoch": 0.5432490287798303, "grad_norm": 1.4111007452011108, "learning_rate": 9.113821283435635e-06, "loss": 0.8484, "step": 15417 }, { "epoch": 0.5432842657927888, "grad_norm": 7.900753021240234, "learning_rate": 9.112684502535833e-06, "loss": 5.7879, "step": 15418 }, { "epoch": 0.5433195028057471, "grad_norm": 4.546868801116943, "learning_rate": 9.111547733193433e-06, "loss": 3.6124, "step": 15419 }, { "epoch": 0.5433547398187055, "grad_norm": 7.984790802001953, "learning_rate": 9.11041097542323e-06, "loss": 5.534, "step": 15420 }, { "epoch": 0.543389976831664, "grad_norm": 4.604796409606934, "learning_rate": 9.109274229240041e-06, "loss": 3.2866, "step": 15421 }, { "epoch": 0.5434252138446224, "grad_norm": 2.1752519607543945, "learning_rate": 9.10813749465867e-06, "loss": 1.2343, "step": 15422 }, { "epoch": 0.5434604508575808, "grad_norm": 1.6619045734405518, "learning_rate": 9.107000771693921e-06, "loss": 0.7609, "step": 15423 }, { "epoch": 0.5434956878705393, "grad_norm": 0.999576985836029, "learning_rate": 9.105864060360598e-06, "loss": 0.7666, "step": 15424 }, { "epoch": 0.5435309248834976, "grad_norm": 7.091386795043945, "learning_rate": 9.10472736067351e-06, "loss": 5.1674, "step": 15425 }, { "epoch": 0.543566161896456, "grad_norm": 2.348712921142578, "learning_rate": 9.103590672647467e-06, "loss": 1.0039, "step": 15426 }, { "epoch": 0.5436013989094145, "grad_norm": 4.666414737701416, "learning_rate": 9.102453996297265e-06, "loss": 3.1353, "step": 15427 }, { "epoch": 0.5436366359223729, "grad_norm": 1.1744835376739502, "learning_rate": 9.101317331637714e-06, "loss": 0.9737, "step": 15428 }, { "epoch": 0.5436718729353313, "grad_norm": 1.497511863708496, "learning_rate": 9.100180678683626e-06, "loss": 0.9052, "step": 15429 }, { "epoch": 0.5437071099482896, "grad_norm": 1.1362385749816895, "learning_rate": 9.09904403744979e-06, "loss": 0.8481, "step": 15430 }, { "epoch": 0.5437423469612481, "grad_norm": 6.027854919433594, "learning_rate": 9.097907407951024e-06, "loss": 5.0148, "step": 15431 }, { "epoch": 0.5437775839742065, "grad_norm": 2.6435253620147705, "learning_rate": 9.096770790202133e-06, "loss": 0.9653, "step": 15432 }, { "epoch": 0.5438128209871649, "grad_norm": 4.191865921020508, "learning_rate": 9.095634184217912e-06, "loss": 3.1251, "step": 15433 }, { "epoch": 0.5438480580001234, "grad_norm": 4.20283317565918, "learning_rate": 9.09449759001317e-06, "loss": 2.8296, "step": 15434 }, { "epoch": 0.5438832950130817, "grad_norm": 5.7162628173828125, "learning_rate": 9.093361007602717e-06, "loss": 3.1369, "step": 15435 }, { "epoch": 0.5439185320260401, "grad_norm": 6.42234992980957, "learning_rate": 9.09222443700135e-06, "loss": 3.0549, "step": 15436 }, { "epoch": 0.5439537690389986, "grad_norm": 1.0468610525131226, "learning_rate": 9.091087878223875e-06, "loss": 0.9999, "step": 15437 }, { "epoch": 0.543989006051957, "grad_norm": 1.3042560815811157, "learning_rate": 9.0899513312851e-06, "loss": 0.7469, "step": 15438 }, { "epoch": 0.5440242430649154, "grad_norm": 1.6609266996383667, "learning_rate": 9.088814796199818e-06, "loss": 1.0798, "step": 15439 }, { "epoch": 0.5440594800778737, "grad_norm": 1.1561872959136963, "learning_rate": 9.087678272982843e-06, "loss": 0.8401, "step": 15440 }, { "epoch": 0.5440947170908322, "grad_norm": 5.861200332641602, "learning_rate": 9.086541761648976e-06, "loss": 4.5883, "step": 15441 }, { "epoch": 0.5441299541037906, "grad_norm": 4.170168876647949, "learning_rate": 9.085405262213018e-06, "loss": 3.2219, "step": 15442 }, { "epoch": 0.544165191116749, "grad_norm": 7.966849327087402, "learning_rate": 9.08426877468977e-06, "loss": 2.9948, "step": 15443 }, { "epoch": 0.5442004281297075, "grad_norm": 0.9600743651390076, "learning_rate": 9.08313229909404e-06, "loss": 0.7749, "step": 15444 }, { "epoch": 0.5442356651426659, "grad_norm": 0.9210410118103027, "learning_rate": 9.08199583544063e-06, "loss": 0.7776, "step": 15445 }, { "epoch": 0.5442709021556242, "grad_norm": 3.0899603366851807, "learning_rate": 9.08085938374434e-06, "loss": 3.0695, "step": 15446 }, { "epoch": 0.5443061391685827, "grad_norm": 1.4822067022323608, "learning_rate": 9.079722944019972e-06, "loss": 1.0707, "step": 15447 }, { "epoch": 0.5443413761815411, "grad_norm": 5.207711219787598, "learning_rate": 9.078586516282334e-06, "loss": 3.0917, "step": 15448 }, { "epoch": 0.5443766131944995, "grad_norm": 8.3616943359375, "learning_rate": 9.07745010054622e-06, "loss": 4.9909, "step": 15449 }, { "epoch": 0.5444118502074579, "grad_norm": 6.581691741943359, "learning_rate": 9.076313696826437e-06, "loss": 3.4068, "step": 15450 }, { "epoch": 0.5444470872204163, "grad_norm": 1.8653032779693604, "learning_rate": 9.07517730513779e-06, "loss": 0.9446, "step": 15451 }, { "epoch": 0.5444823242333747, "grad_norm": 9.639348983764648, "learning_rate": 9.07404092549507e-06, "loss": 9.4014, "step": 15452 }, { "epoch": 0.5445175612463331, "grad_norm": 2.472191095352173, "learning_rate": 9.072904557913087e-06, "loss": 2.7777, "step": 15453 }, { "epoch": 0.5445527982592916, "grad_norm": 6.402626037597656, "learning_rate": 9.071768202406642e-06, "loss": 5.4549, "step": 15454 }, { "epoch": 0.54458803527225, "grad_norm": 38.583160400390625, "learning_rate": 9.070631858990532e-06, "loss": 5.8587, "step": 15455 }, { "epoch": 0.5446232722852083, "grad_norm": 3.027132272720337, "learning_rate": 9.069495527679562e-06, "loss": 2.7495, "step": 15456 }, { "epoch": 0.5446585092981668, "grad_norm": 2.241089105606079, "learning_rate": 9.068359208488533e-06, "loss": 0.9718, "step": 15457 }, { "epoch": 0.5446937463111252, "grad_norm": 3.1543304920196533, "learning_rate": 9.06722290143224e-06, "loss": 1.2285, "step": 15458 }, { "epoch": 0.5447289833240836, "grad_norm": 3.3330612182617188, "learning_rate": 9.06608660652549e-06, "loss": 0.7105, "step": 15459 }, { "epoch": 0.5447642203370421, "grad_norm": 5.934630393981934, "learning_rate": 9.06495032378308e-06, "loss": 4.7855, "step": 15460 }, { "epoch": 0.5447994573500005, "grad_norm": 4.8185505867004395, "learning_rate": 9.063814053219813e-06, "loss": 3.0686, "step": 15461 }, { "epoch": 0.5448346943629588, "grad_norm": 3.6549251079559326, "learning_rate": 9.062677794850482e-06, "loss": 2.5319, "step": 15462 }, { "epoch": 0.5448699313759172, "grad_norm": 2.102142095565796, "learning_rate": 9.061541548689896e-06, "loss": 0.9128, "step": 15463 }, { "epoch": 0.5449051683888757, "grad_norm": 4.141654968261719, "learning_rate": 9.060405314752853e-06, "loss": 2.9676, "step": 15464 }, { "epoch": 0.5449404054018341, "grad_norm": 4.684807300567627, "learning_rate": 9.059269093054149e-06, "loss": 4.6409, "step": 15465 }, { "epoch": 0.5449756424147925, "grad_norm": 1.2997928857803345, "learning_rate": 9.05813288360858e-06, "loss": 0.711, "step": 15466 }, { "epoch": 0.545010879427751, "grad_norm": 1.068090558052063, "learning_rate": 9.05699668643096e-06, "loss": 0.9754, "step": 15467 }, { "epoch": 0.5450461164407093, "grad_norm": 1.1562591791152954, "learning_rate": 9.055860501536068e-06, "loss": 1.3633, "step": 15468 }, { "epoch": 0.5450813534536677, "grad_norm": 4.42954683303833, "learning_rate": 9.054724328938718e-06, "loss": 2.8239, "step": 15469 }, { "epoch": 0.5451165904666262, "grad_norm": 3.5897297859191895, "learning_rate": 9.053588168653706e-06, "loss": 3.3475, "step": 15470 }, { "epoch": 0.5451518274795846, "grad_norm": 29.518815994262695, "learning_rate": 9.052452020695826e-06, "loss": 3.258, "step": 15471 }, { "epoch": 0.545187064492543, "grad_norm": 7.955965518951416, "learning_rate": 9.051315885079877e-06, "loss": 3.1721, "step": 15472 }, { "epoch": 0.5452223015055013, "grad_norm": 20.388328552246094, "learning_rate": 9.050179761820665e-06, "loss": 3.5851, "step": 15473 }, { "epoch": 0.5452575385184598, "grad_norm": 1.9052234888076782, "learning_rate": 9.049043650932979e-06, "loss": 0.8799, "step": 15474 }, { "epoch": 0.5452927755314182, "grad_norm": 3.620858669281006, "learning_rate": 9.047907552431621e-06, "loss": 3.1417, "step": 15475 }, { "epoch": 0.5453280125443766, "grad_norm": 1.1295539140701294, "learning_rate": 9.04677146633139e-06, "loss": 1.0288, "step": 15476 }, { "epoch": 0.5453632495573351, "grad_norm": 5.2283244132995605, "learning_rate": 9.045635392647077e-06, "loss": 3.4232, "step": 15477 }, { "epoch": 0.5453984865702934, "grad_norm": 4.233792781829834, "learning_rate": 9.044499331393489e-06, "loss": 3.3009, "step": 15478 }, { "epoch": 0.5454337235832518, "grad_norm": 1.1937906742095947, "learning_rate": 9.043363282585418e-06, "loss": 0.9666, "step": 15479 }, { "epoch": 0.5454689605962103, "grad_norm": 1.1981549263000488, "learning_rate": 9.042227246237663e-06, "loss": 1.0031, "step": 15480 }, { "epoch": 0.5455041976091687, "grad_norm": 6.9479660987854, "learning_rate": 9.041091222365016e-06, "loss": 3.3245, "step": 15481 }, { "epoch": 0.5455394346221271, "grad_norm": 4.45908260345459, "learning_rate": 9.039955210982284e-06, "loss": 2.8621, "step": 15482 }, { "epoch": 0.5455746716350854, "grad_norm": 2.596060037612915, "learning_rate": 9.038819212104256e-06, "loss": 0.9353, "step": 15483 }, { "epoch": 0.5456099086480439, "grad_norm": 6.187264919281006, "learning_rate": 9.037683225745727e-06, "loss": 3.2293, "step": 15484 }, { "epoch": 0.5456451456610023, "grad_norm": 4.3574018478393555, "learning_rate": 9.0365472519215e-06, "loss": 4.7634, "step": 15485 }, { "epoch": 0.5456803826739607, "grad_norm": 5.145599365234375, "learning_rate": 9.035411290646367e-06, "loss": 3.2408, "step": 15486 }, { "epoch": 0.5457156196869192, "grad_norm": 1.7169865369796753, "learning_rate": 9.034275341935123e-06, "loss": 0.82, "step": 15487 }, { "epoch": 0.5457508566998776, "grad_norm": 2.1682686805725098, "learning_rate": 9.033139405802567e-06, "loss": 1.1155, "step": 15488 }, { "epoch": 0.5457860937128359, "grad_norm": 6.3604302406311035, "learning_rate": 9.032003482263495e-06, "loss": 5.1664, "step": 15489 }, { "epoch": 0.5458213307257944, "grad_norm": 5.552687168121338, "learning_rate": 9.030867571332698e-06, "loss": 3.4141, "step": 15490 }, { "epoch": 0.5458565677387528, "grad_norm": 8.718358993530273, "learning_rate": 9.029731673024974e-06, "loss": 4.9199, "step": 15491 }, { "epoch": 0.5458918047517112, "grad_norm": 1.4065380096435547, "learning_rate": 9.028595787355122e-06, "loss": 1.0979, "step": 15492 }, { "epoch": 0.5459270417646697, "grad_norm": 2.280117988586426, "learning_rate": 9.027459914337932e-06, "loss": 1.0473, "step": 15493 }, { "epoch": 0.545962278777628, "grad_norm": 3.824714183807373, "learning_rate": 9.0263240539882e-06, "loss": 3.1489, "step": 15494 }, { "epoch": 0.5459975157905864, "grad_norm": 2.363490343093872, "learning_rate": 9.025188206320722e-06, "loss": 0.9036, "step": 15495 }, { "epoch": 0.5460327528035448, "grad_norm": 6.3245673179626465, "learning_rate": 9.024052371350292e-06, "loss": 4.9849, "step": 15496 }, { "epoch": 0.5460679898165033, "grad_norm": 6.625250816345215, "learning_rate": 9.022916549091702e-06, "loss": 3.0032, "step": 15497 }, { "epoch": 0.5461032268294617, "grad_norm": 6.660531997680664, "learning_rate": 9.02178073955975e-06, "loss": 2.6716, "step": 15498 }, { "epoch": 0.54613846384242, "grad_norm": 3.7263505458831787, "learning_rate": 9.02064494276923e-06, "loss": 3.0398, "step": 15499 }, { "epoch": 0.5461737008553785, "grad_norm": 3.2190768718719482, "learning_rate": 9.019509158734932e-06, "loss": 0.8319, "step": 15500 }, { "epoch": 0.5462089378683369, "grad_norm": 1.1760691404342651, "learning_rate": 9.01837338747165e-06, "loss": 0.8467, "step": 15501 }, { "epoch": 0.5462441748812953, "grad_norm": 1.015005350112915, "learning_rate": 9.017237628994185e-06, "loss": 0.8688, "step": 15502 }, { "epoch": 0.5462794118942538, "grad_norm": 2.747830390930176, "learning_rate": 9.016101883317322e-06, "loss": 1.0794, "step": 15503 }, { "epoch": 0.5463146489072122, "grad_norm": 1.776726484298706, "learning_rate": 9.014966150455858e-06, "loss": 0.8211, "step": 15504 }, { "epoch": 0.5463498859201705, "grad_norm": 0.9894757866859436, "learning_rate": 9.013830430424587e-06, "loss": 0.9249, "step": 15505 }, { "epoch": 0.5463851229331289, "grad_norm": 4.5948805809021, "learning_rate": 9.012694723238296e-06, "loss": 2.8766, "step": 15506 }, { "epoch": 0.5464203599460874, "grad_norm": 6.103562355041504, "learning_rate": 9.011559028911786e-06, "loss": 4.1183, "step": 15507 }, { "epoch": 0.5464555969590458, "grad_norm": 1.5233925580978394, "learning_rate": 9.010423347459844e-06, "loss": 0.969, "step": 15508 }, { "epoch": 0.5464908339720042, "grad_norm": 3.7660155296325684, "learning_rate": 9.009287678897264e-06, "loss": 3.0226, "step": 15509 }, { "epoch": 0.5465260709849626, "grad_norm": 68.42993927001953, "learning_rate": 9.008152023238837e-06, "loss": 5.0862, "step": 15510 }, { "epoch": 0.546561307997921, "grad_norm": 1.0755425691604614, "learning_rate": 9.00701638049936e-06, "loss": 0.883, "step": 15511 }, { "epoch": 0.5465965450108794, "grad_norm": 1.3755087852478027, "learning_rate": 9.005880750693617e-06, "loss": 1.1288, "step": 15512 }, { "epoch": 0.5466317820238379, "grad_norm": 4.871565341949463, "learning_rate": 9.004745133836405e-06, "loss": 2.7584, "step": 15513 }, { "epoch": 0.5466670190367963, "grad_norm": 4.071077823638916, "learning_rate": 9.003609529942512e-06, "loss": 2.8937, "step": 15514 }, { "epoch": 0.5467022560497546, "grad_norm": 1.2557919025421143, "learning_rate": 9.00247393902674e-06, "loss": 0.9955, "step": 15515 }, { "epoch": 0.546737493062713, "grad_norm": 10.885479927062988, "learning_rate": 9.001338361103863e-06, "loss": 3.4518, "step": 15516 }, { "epoch": 0.5467727300756715, "grad_norm": 4.01617431640625, "learning_rate": 9.000202796188684e-06, "loss": 3.5392, "step": 15517 }, { "epoch": 0.5468079670886299, "grad_norm": 5.17629337310791, "learning_rate": 8.999067244295994e-06, "loss": 4.9571, "step": 15518 }, { "epoch": 0.5468432041015883, "grad_norm": 29.009143829345703, "learning_rate": 8.997931705440577e-06, "loss": 3.8773, "step": 15519 }, { "epoch": 0.5468784411145468, "grad_norm": 1.4544353485107422, "learning_rate": 8.996796179637227e-06, "loss": 0.9498, "step": 15520 }, { "epoch": 0.5469136781275051, "grad_norm": 5.64267110824585, "learning_rate": 8.995660666900737e-06, "loss": 4.6772, "step": 15521 }, { "epoch": 0.5469489151404635, "grad_norm": 6.290316581726074, "learning_rate": 8.994525167245895e-06, "loss": 3.2765, "step": 15522 }, { "epoch": 0.546984152153422, "grad_norm": 4.891120433807373, "learning_rate": 8.993389680687488e-06, "loss": 3.3592, "step": 15523 }, { "epoch": 0.5470193891663804, "grad_norm": 1.7723246812820435, "learning_rate": 8.992254207240313e-06, "loss": 1.1464, "step": 15524 }, { "epoch": 0.5470546261793388, "grad_norm": 4.549907207489014, "learning_rate": 8.99111874691915e-06, "loss": 3.2004, "step": 15525 }, { "epoch": 0.5470898631922972, "grad_norm": 1.144250750541687, "learning_rate": 8.989983299738796e-06, "loss": 1.0384, "step": 15526 }, { "epoch": 0.5471251002052556, "grad_norm": 1.4846230745315552, "learning_rate": 8.98884786571404e-06, "loss": 0.9565, "step": 15527 }, { "epoch": 0.547160337218214, "grad_norm": 1.4066078662872314, "learning_rate": 8.987712444859667e-06, "loss": 1.026, "step": 15528 }, { "epoch": 0.5471955742311724, "grad_norm": 1.2727879285812378, "learning_rate": 8.986577037190467e-06, "loss": 0.9544, "step": 15529 }, { "epoch": 0.5472308112441309, "grad_norm": 5.427886009216309, "learning_rate": 8.985441642721237e-06, "loss": 3.4558, "step": 15530 }, { "epoch": 0.5472660482570892, "grad_norm": 5.081087112426758, "learning_rate": 8.984306261466751e-06, "loss": 2.8478, "step": 15531 }, { "epoch": 0.5473012852700476, "grad_norm": 11.571441650390625, "learning_rate": 8.983170893441808e-06, "loss": 5.8859, "step": 15532 }, { "epoch": 0.5473365222830061, "grad_norm": 4.407556056976318, "learning_rate": 8.982035538661193e-06, "loss": 2.9764, "step": 15533 }, { "epoch": 0.5473717592959645, "grad_norm": 2.00594425201416, "learning_rate": 8.980900197139699e-06, "loss": 0.8381, "step": 15534 }, { "epoch": 0.5474069963089229, "grad_norm": 1.1459741592407227, "learning_rate": 8.979764868892104e-06, "loss": 0.823, "step": 15535 }, { "epoch": 0.5474422333218814, "grad_norm": 4.013062477111816, "learning_rate": 8.978629553933205e-06, "loss": 3.1147, "step": 15536 }, { "epoch": 0.5474774703348397, "grad_norm": 11.053115844726562, "learning_rate": 8.977494252277786e-06, "loss": 3.0985, "step": 15537 }, { "epoch": 0.5475127073477981, "grad_norm": 10.492278099060059, "learning_rate": 8.976358963940634e-06, "loss": 5.5468, "step": 15538 }, { "epoch": 0.5475479443607565, "grad_norm": 4.861824989318848, "learning_rate": 8.975223688936536e-06, "loss": 2.8421, "step": 15539 }, { "epoch": 0.547583181373715, "grad_norm": 1.2242913246154785, "learning_rate": 8.974088427280281e-06, "loss": 0.7271, "step": 15540 }, { "epoch": 0.5476184183866734, "grad_norm": 5.021175384521484, "learning_rate": 8.972953178986656e-06, "loss": 4.3519, "step": 15541 }, { "epoch": 0.5476536553996317, "grad_norm": 9.460658073425293, "learning_rate": 8.971817944070445e-06, "loss": 4.894, "step": 15542 }, { "epoch": 0.5476888924125902, "grad_norm": 2.2468152046203613, "learning_rate": 8.970682722546437e-06, "loss": 1.093, "step": 15543 }, { "epoch": 0.5477241294255486, "grad_norm": 2.9719395637512207, "learning_rate": 8.969547514429416e-06, "loss": 3.0565, "step": 15544 }, { "epoch": 0.547759366438507, "grad_norm": 1.6336674690246582, "learning_rate": 8.968412319734172e-06, "loss": 0.9904, "step": 15545 }, { "epoch": 0.5477946034514655, "grad_norm": 1.0466159582138062, "learning_rate": 8.967277138475489e-06, "loss": 1.2001, "step": 15546 }, { "epoch": 0.5478298404644238, "grad_norm": 1.4973938465118408, "learning_rate": 8.966141970668152e-06, "loss": 1.2127, "step": 15547 }, { "epoch": 0.5478650774773822, "grad_norm": 1.0392183065414429, "learning_rate": 8.965006816326944e-06, "loss": 0.9889, "step": 15548 }, { "epoch": 0.5479003144903406, "grad_norm": 5.22922420501709, "learning_rate": 8.963871675466662e-06, "loss": 4.9999, "step": 15549 }, { "epoch": 0.5479355515032991, "grad_norm": 1.179841160774231, "learning_rate": 8.962736548102077e-06, "loss": 0.8607, "step": 15550 }, { "epoch": 0.5479707885162575, "grad_norm": 4.8562912940979, "learning_rate": 8.961601434247981e-06, "loss": 4.977, "step": 15551 }, { "epoch": 0.5480060255292158, "grad_norm": 1.118006944656372, "learning_rate": 8.960466333919158e-06, "loss": 1.3723, "step": 15552 }, { "epoch": 0.5480412625421743, "grad_norm": 5.829901695251465, "learning_rate": 8.959331247130401e-06, "loss": 3.2674, "step": 15553 }, { "epoch": 0.5480764995551327, "grad_norm": 4.4104228019714355, "learning_rate": 8.958196173896478e-06, "loss": 3.0668, "step": 15554 }, { "epoch": 0.5481117365680911, "grad_norm": 1.278076171875, "learning_rate": 8.957061114232186e-06, "loss": 1.1376, "step": 15555 }, { "epoch": 0.5481469735810496, "grad_norm": 1.3943413496017456, "learning_rate": 8.955926068152308e-06, "loss": 0.9827, "step": 15556 }, { "epoch": 0.548182210594008, "grad_norm": 1.4543389081954956, "learning_rate": 8.954791035671623e-06, "loss": 0.642, "step": 15557 }, { "epoch": 0.5482174476069663, "grad_norm": 3.371107816696167, "learning_rate": 8.953656016804917e-06, "loss": 2.8446, "step": 15558 }, { "epoch": 0.5482526846199248, "grad_norm": 1.651667594909668, "learning_rate": 8.952521011566979e-06, "loss": 0.7482, "step": 15559 }, { "epoch": 0.5482879216328832, "grad_norm": 2.6214308738708496, "learning_rate": 8.951386019972585e-06, "loss": 2.3484, "step": 15560 }, { "epoch": 0.5483231586458416, "grad_norm": 1.7462437152862549, "learning_rate": 8.950251042036521e-06, "loss": 0.7688, "step": 15561 }, { "epoch": 0.5483583956588, "grad_norm": 4.205202579498291, "learning_rate": 8.949116077773575e-06, "loss": 2.5733, "step": 15562 }, { "epoch": 0.5483936326717584, "grad_norm": 5.041943550109863, "learning_rate": 8.947981127198522e-06, "loss": 3.2544, "step": 15563 }, { "epoch": 0.5484288696847168, "grad_norm": 2.2760884761810303, "learning_rate": 8.94684619032615e-06, "loss": 0.8506, "step": 15564 }, { "epoch": 0.5484641066976752, "grad_norm": 0.9663840532302856, "learning_rate": 8.945711267171242e-06, "loss": 1.0243, "step": 15565 }, { "epoch": 0.5484993437106337, "grad_norm": 2.38897967338562, "learning_rate": 8.944576357748578e-06, "loss": 0.8566, "step": 15566 }, { "epoch": 0.5485345807235921, "grad_norm": 1.277939796447754, "learning_rate": 8.943441462072938e-06, "loss": 0.7672, "step": 15567 }, { "epoch": 0.5485698177365504, "grad_norm": 1.8614990711212158, "learning_rate": 8.942306580159114e-06, "loss": 0.8216, "step": 15568 }, { "epoch": 0.5486050547495089, "grad_norm": 1.360870599746704, "learning_rate": 8.941171712021876e-06, "loss": 1.0632, "step": 15569 }, { "epoch": 0.5486402917624673, "grad_norm": 1.3001309633255005, "learning_rate": 8.940036857676014e-06, "loss": 0.9796, "step": 15570 }, { "epoch": 0.5486755287754257, "grad_norm": 1.2090246677398682, "learning_rate": 8.938902017136302e-06, "loss": 0.8704, "step": 15571 }, { "epoch": 0.5487107657883841, "grad_norm": 1.3998397588729858, "learning_rate": 8.937767190417537e-06, "loss": 0.7562, "step": 15572 }, { "epoch": 0.5487460028013426, "grad_norm": 1.005759596824646, "learning_rate": 8.93663237753448e-06, "loss": 1.0984, "step": 15573 }, { "epoch": 0.5487812398143009, "grad_norm": 1.95266854763031, "learning_rate": 8.935497578501925e-06, "loss": 1.0061, "step": 15574 }, { "epoch": 0.5488164768272593, "grad_norm": 0.9111654758453369, "learning_rate": 8.934362793334653e-06, "loss": 0.7797, "step": 15575 }, { "epoch": 0.5488517138402178, "grad_norm": 1.2955199480056763, "learning_rate": 8.933228022047437e-06, "loss": 1.2147, "step": 15576 }, { "epoch": 0.5488869508531762, "grad_norm": 1.498319149017334, "learning_rate": 8.93209326465506e-06, "loss": 1.2521, "step": 15577 }, { "epoch": 0.5489221878661346, "grad_norm": 4.28712797164917, "learning_rate": 8.93095852117231e-06, "loss": 2.9088, "step": 15578 }, { "epoch": 0.548957424879093, "grad_norm": 1.258297085762024, "learning_rate": 8.929823791613957e-06, "loss": 0.9712, "step": 15579 }, { "epoch": 0.5489926618920514, "grad_norm": 3.8234689235687256, "learning_rate": 8.928689075994788e-06, "loss": 3.0559, "step": 15580 }, { "epoch": 0.5490278989050098, "grad_norm": 1.7383674383163452, "learning_rate": 8.927554374329582e-06, "loss": 1.042, "step": 15581 }, { "epoch": 0.5490631359179682, "grad_norm": 3.882058620452881, "learning_rate": 8.926419686633112e-06, "loss": 3.0933, "step": 15582 }, { "epoch": 0.5490983729309267, "grad_norm": 2.0751953125, "learning_rate": 8.925285012920163e-06, "loss": 1.0486, "step": 15583 }, { "epoch": 0.549133609943885, "grad_norm": 5.356316089630127, "learning_rate": 8.924150353205517e-06, "loss": 3.0458, "step": 15584 }, { "epoch": 0.5491688469568434, "grad_norm": 1.1905587911605835, "learning_rate": 8.923015707503948e-06, "loss": 0.9121, "step": 15585 }, { "epoch": 0.5492040839698019, "grad_norm": 5.3243513107299805, "learning_rate": 8.921881075830235e-06, "loss": 2.866, "step": 15586 }, { "epoch": 0.5492393209827603, "grad_norm": 2.011233329772949, "learning_rate": 8.92074645819916e-06, "loss": 2.5353, "step": 15587 }, { "epoch": 0.5492745579957187, "grad_norm": 2.3953895568847656, "learning_rate": 8.9196118546255e-06, "loss": 0.5945, "step": 15588 }, { "epoch": 0.5493097950086772, "grad_norm": 1.0924519300460815, "learning_rate": 8.918477265124033e-06, "loss": 0.8354, "step": 15589 }, { "epoch": 0.5493450320216355, "grad_norm": 1.3791300058364868, "learning_rate": 8.917342689709535e-06, "loss": 0.8734, "step": 15590 }, { "epoch": 0.5493802690345939, "grad_norm": 4.818009853363037, "learning_rate": 8.916208128396793e-06, "loss": 2.8797, "step": 15591 }, { "epoch": 0.5494155060475524, "grad_norm": 7.282834053039551, "learning_rate": 8.91507358120057e-06, "loss": 5.104, "step": 15592 }, { "epoch": 0.5494507430605108, "grad_norm": 4.6102986335754395, "learning_rate": 8.913939048135655e-06, "loss": 3.0961, "step": 15593 }, { "epoch": 0.5494859800734692, "grad_norm": 3.9244463443756104, "learning_rate": 8.912804529216827e-06, "loss": 3.2972, "step": 15594 }, { "epoch": 0.5495212170864275, "grad_norm": 7.257391929626465, "learning_rate": 8.911670024458853e-06, "loss": 2.8022, "step": 15595 }, { "epoch": 0.549556454099386, "grad_norm": 1.1942180395126343, "learning_rate": 8.910535533876514e-06, "loss": 1.1819, "step": 15596 }, { "epoch": 0.5495916911123444, "grad_norm": 2.068772315979004, "learning_rate": 8.909401057484593e-06, "loss": 1.085, "step": 15597 }, { "epoch": 0.5496269281253028, "grad_norm": 2.753244638442993, "learning_rate": 8.908266595297862e-06, "loss": 2.2042, "step": 15598 }, { "epoch": 0.5496621651382613, "grad_norm": 5.839825630187988, "learning_rate": 8.907132147331096e-06, "loss": 3.2938, "step": 15599 }, { "epoch": 0.5496974021512196, "grad_norm": 3.607607364654541, "learning_rate": 8.905997713599074e-06, "loss": 2.9787, "step": 15600 }, { "epoch": 0.549732639164178, "grad_norm": 5.512181758880615, "learning_rate": 8.904863294116569e-06, "loss": 3.0436, "step": 15601 }, { "epoch": 0.5497678761771365, "grad_norm": 1.4439221620559692, "learning_rate": 8.90372888889836e-06, "loss": 0.8421, "step": 15602 }, { "epoch": 0.5498031131900949, "grad_norm": 1.1840862035751343, "learning_rate": 8.902594497959223e-06, "loss": 0.9594, "step": 15603 }, { "epoch": 0.5498383502030533, "grad_norm": 4.438760280609131, "learning_rate": 8.901460121313932e-06, "loss": 3.4777, "step": 15604 }, { "epoch": 0.5498735872160117, "grad_norm": 4.961489200592041, "learning_rate": 8.900325758977259e-06, "loss": 2.8141, "step": 15605 }, { "epoch": 0.5499088242289701, "grad_norm": 3.389559745788574, "learning_rate": 8.899191410963987e-06, "loss": 3.1604, "step": 15606 }, { "epoch": 0.5499440612419285, "grad_norm": 2.2324378490448, "learning_rate": 8.898057077288888e-06, "loss": 1.0035, "step": 15607 }, { "epoch": 0.5499792982548869, "grad_norm": 0.993123471736908, "learning_rate": 8.896922757966735e-06, "loss": 0.8182, "step": 15608 }, { "epoch": 0.5500145352678454, "grad_norm": 1.3535747528076172, "learning_rate": 8.895788453012301e-06, "loss": 0.963, "step": 15609 }, { "epoch": 0.5500497722808038, "grad_norm": 6.21498966217041, "learning_rate": 8.89465416244037e-06, "loss": 5.2393, "step": 15610 }, { "epoch": 0.5500850092937621, "grad_norm": 3.906428813934326, "learning_rate": 8.893519886265703e-06, "loss": 2.6598, "step": 15611 }, { "epoch": 0.5501202463067206, "grad_norm": 3.5832014083862305, "learning_rate": 8.892385624503081e-06, "loss": 4.7008, "step": 15612 }, { "epoch": 0.550155483319679, "grad_norm": 3.7645180225372314, "learning_rate": 8.89125137716728e-06, "loss": 3.0802, "step": 15613 }, { "epoch": 0.5501907203326374, "grad_norm": 1.7989628314971924, "learning_rate": 8.89011714427307e-06, "loss": 0.9327, "step": 15614 }, { "epoch": 0.5502259573455958, "grad_norm": 4.898663520812988, "learning_rate": 8.888982925835222e-06, "loss": 5.2773, "step": 15615 }, { "epoch": 0.5502611943585543, "grad_norm": 1.658756971359253, "learning_rate": 8.887848721868518e-06, "loss": 0.8539, "step": 15616 }, { "epoch": 0.5502964313715126, "grad_norm": 2.095001220703125, "learning_rate": 8.886714532387723e-06, "loss": 0.8187, "step": 15617 }, { "epoch": 0.550331668384471, "grad_norm": 2.7273709774017334, "learning_rate": 8.885580357407614e-06, "loss": 2.5742, "step": 15618 }, { "epoch": 0.5503669053974295, "grad_norm": 12.774979591369629, "learning_rate": 8.884446196942965e-06, "loss": 5.0575, "step": 15619 }, { "epoch": 0.5504021424103879, "grad_norm": 6.483046054840088, "learning_rate": 8.88331205100854e-06, "loss": 2.6748, "step": 15620 }, { "epoch": 0.5504373794233463, "grad_norm": 2.849905014038086, "learning_rate": 8.882177919619124e-06, "loss": 2.8467, "step": 15621 }, { "epoch": 0.5504726164363047, "grad_norm": 1.4767385721206665, "learning_rate": 8.881043802789482e-06, "loss": 0.8651, "step": 15622 }, { "epoch": 0.5505078534492631, "grad_norm": 8.36115837097168, "learning_rate": 8.879909700534387e-06, "loss": 3.048, "step": 15623 }, { "epoch": 0.5505430904622215, "grad_norm": 1.1688427925109863, "learning_rate": 8.878775612868606e-06, "loss": 1.0671, "step": 15624 }, { "epoch": 0.5505783274751799, "grad_norm": 2.5634138584136963, "learning_rate": 8.87764153980692e-06, "loss": 2.5399, "step": 15625 }, { "epoch": 0.5506135644881384, "grad_norm": 3.9126102924346924, "learning_rate": 8.876507481364095e-06, "loss": 2.5353, "step": 15626 }, { "epoch": 0.5506488015010967, "grad_norm": 5.546586990356445, "learning_rate": 8.875373437554904e-06, "loss": 3.4197, "step": 15627 }, { "epoch": 0.5506840385140551, "grad_norm": 1.6975853443145752, "learning_rate": 8.874239408394114e-06, "loss": 0.9083, "step": 15628 }, { "epoch": 0.5507192755270136, "grad_norm": 2.803173303604126, "learning_rate": 8.873105393896506e-06, "loss": 2.7319, "step": 15629 }, { "epoch": 0.550754512539972, "grad_norm": 1.3135632276535034, "learning_rate": 8.871971394076837e-06, "loss": 0.7538, "step": 15630 }, { "epoch": 0.5507897495529304, "grad_norm": 1.201499581336975, "learning_rate": 8.870837408949886e-06, "loss": 0.7777, "step": 15631 }, { "epoch": 0.5508249865658889, "grad_norm": 1.2701853513717651, "learning_rate": 8.869703438530425e-06, "loss": 0.8631, "step": 15632 }, { "epoch": 0.5508602235788472, "grad_norm": 0.851351797580719, "learning_rate": 8.868569482833219e-06, "loss": 0.926, "step": 15633 }, { "epoch": 0.5508954605918056, "grad_norm": 1.5139074325561523, "learning_rate": 8.867435541873034e-06, "loss": 0.9208, "step": 15634 }, { "epoch": 0.5509306976047641, "grad_norm": 9.088237762451172, "learning_rate": 8.866301615664654e-06, "loss": 4.564, "step": 15635 }, { "epoch": 0.5509659346177225, "grad_norm": 3.8425419330596924, "learning_rate": 8.865167704222834e-06, "loss": 2.5164, "step": 15636 }, { "epoch": 0.5510011716306809, "grad_norm": 4.994855880737305, "learning_rate": 8.86403380756235e-06, "loss": 3.0147, "step": 15637 }, { "epoch": 0.5510364086436392, "grad_norm": 4.172482490539551, "learning_rate": 8.862899925697975e-06, "loss": 3.5179, "step": 15638 }, { "epoch": 0.5510716456565977, "grad_norm": 1.2767525911331177, "learning_rate": 8.861766058644466e-06, "loss": 0.9491, "step": 15639 }, { "epoch": 0.5511068826695561, "grad_norm": 4.297550201416016, "learning_rate": 8.860632206416603e-06, "loss": 3.0458, "step": 15640 }, { "epoch": 0.5511421196825145, "grad_norm": 3.141145706176758, "learning_rate": 8.859498369029153e-06, "loss": 0.7481, "step": 15641 }, { "epoch": 0.551177356695473, "grad_norm": 1.7527785301208496, "learning_rate": 8.858364546496879e-06, "loss": 0.8419, "step": 15642 }, { "epoch": 0.5512125937084313, "grad_norm": 7.28505277633667, "learning_rate": 8.857230738834552e-06, "loss": 3.4614, "step": 15643 }, { "epoch": 0.5512478307213897, "grad_norm": 5.628776550292969, "learning_rate": 8.856096946056939e-06, "loss": 3.0829, "step": 15644 }, { "epoch": 0.5512830677343482, "grad_norm": 1.6369436979293823, "learning_rate": 8.854963168178811e-06, "loss": 1.1188, "step": 15645 }, { "epoch": 0.5513183047473066, "grad_norm": 1.4387503862380981, "learning_rate": 8.853829405214935e-06, "loss": 1.0769, "step": 15646 }, { "epoch": 0.551353541760265, "grad_norm": 2.385364294052124, "learning_rate": 8.852695657180076e-06, "loss": 0.9894, "step": 15647 }, { "epoch": 0.5513887787732233, "grad_norm": 2.9740521907806396, "learning_rate": 8.851561924089002e-06, "loss": 0.6958, "step": 15648 }, { "epoch": 0.5514240157861818, "grad_norm": 1.7188389301300049, "learning_rate": 8.850428205956477e-06, "loss": 0.9788, "step": 15649 }, { "epoch": 0.5514592527991402, "grad_norm": 5.170938968658447, "learning_rate": 8.849294502797274e-06, "loss": 6.4166, "step": 15650 }, { "epoch": 0.5514944898120986, "grad_norm": 10.220317840576172, "learning_rate": 8.84816081462616e-06, "loss": 3.0552, "step": 15651 }, { "epoch": 0.5515297268250571, "grad_norm": 1.149523138999939, "learning_rate": 8.847027141457893e-06, "loss": 0.9945, "step": 15652 }, { "epoch": 0.5515649638380155, "grad_norm": 4.6155009269714355, "learning_rate": 8.845893483307242e-06, "loss": 3.3993, "step": 15653 }, { "epoch": 0.5516002008509738, "grad_norm": 1.1353625059127808, "learning_rate": 8.844759840188984e-06, "loss": 0.9793, "step": 15654 }, { "epoch": 0.5516354378639323, "grad_norm": 2.964052438735962, "learning_rate": 8.843626212117869e-06, "loss": 3.1564, "step": 15655 }, { "epoch": 0.5516706748768907, "grad_norm": 7.244533538818359, "learning_rate": 8.842492599108672e-06, "loss": 5.6166, "step": 15656 }, { "epoch": 0.5517059118898491, "grad_norm": 3.8365159034729004, "learning_rate": 8.84135900117616e-06, "loss": 2.6965, "step": 15657 }, { "epoch": 0.5517411489028075, "grad_norm": 1.3651982545852661, "learning_rate": 8.84022541833509e-06, "loss": 1.0936, "step": 15658 }, { "epoch": 0.5517763859157659, "grad_norm": 5.344969272613525, "learning_rate": 8.839091850600232e-06, "loss": 5.3784, "step": 15659 }, { "epoch": 0.5518116229287243, "grad_norm": 2.0152957439422607, "learning_rate": 8.837958297986351e-06, "loss": 0.8709, "step": 15660 }, { "epoch": 0.5518468599416827, "grad_norm": 3.897641897201538, "learning_rate": 8.836824760508214e-06, "loss": 2.6092, "step": 15661 }, { "epoch": 0.5518820969546412, "grad_norm": 1.263319730758667, "learning_rate": 8.83569123818058e-06, "loss": 0.9738, "step": 15662 }, { "epoch": 0.5519173339675996, "grad_norm": 3.273512601852417, "learning_rate": 8.834557731018215e-06, "loss": 1.0817, "step": 15663 }, { "epoch": 0.5519525709805579, "grad_norm": 5.587802410125732, "learning_rate": 8.833424239035886e-06, "loss": 4.2591, "step": 15664 }, { "epoch": 0.5519878079935164, "grad_norm": 1.393261194229126, "learning_rate": 8.832290762248354e-06, "loss": 1.124, "step": 15665 }, { "epoch": 0.5520230450064748, "grad_norm": 1.2515754699707031, "learning_rate": 8.831157300670384e-06, "loss": 0.9545, "step": 15666 }, { "epoch": 0.5520582820194332, "grad_norm": 4.121696472167969, "learning_rate": 8.830023854316741e-06, "loss": 2.8769, "step": 15667 }, { "epoch": 0.5520935190323917, "grad_norm": 4.6045074462890625, "learning_rate": 8.828890423202182e-06, "loss": 2.841, "step": 15668 }, { "epoch": 0.55212875604535, "grad_norm": 5.128659725189209, "learning_rate": 8.827757007341476e-06, "loss": 4.1951, "step": 15669 }, { "epoch": 0.5521639930583084, "grad_norm": 0.9765064716339111, "learning_rate": 8.826623606749388e-06, "loss": 1.0002, "step": 15670 }, { "epoch": 0.5521992300712668, "grad_norm": 2.0108258724212646, "learning_rate": 8.825490221440672e-06, "loss": 0.8787, "step": 15671 }, { "epoch": 0.5522344670842253, "grad_norm": 1.0066137313842773, "learning_rate": 8.824356851430096e-06, "loss": 0.6039, "step": 15672 }, { "epoch": 0.5522697040971837, "grad_norm": 1.6401848793029785, "learning_rate": 8.823223496732428e-06, "loss": 0.8205, "step": 15673 }, { "epoch": 0.5523049411101421, "grad_norm": 2.2905733585357666, "learning_rate": 8.822090157362415e-06, "loss": 2.1553, "step": 15674 }, { "epoch": 0.5523401781231005, "grad_norm": 4.385238170623779, "learning_rate": 8.820956833334832e-06, "loss": 3.364, "step": 15675 }, { "epoch": 0.5523754151360589, "grad_norm": 3.894212007522583, "learning_rate": 8.819823524664438e-06, "loss": 2.9945, "step": 15676 }, { "epoch": 0.5524106521490173, "grad_norm": 4.17099666595459, "learning_rate": 8.81869023136599e-06, "loss": 4.4949, "step": 15677 }, { "epoch": 0.5524458891619758, "grad_norm": 6.964938163757324, "learning_rate": 8.817556953454251e-06, "loss": 5.459, "step": 15678 }, { "epoch": 0.5524811261749342, "grad_norm": 16.43741798400879, "learning_rate": 8.816423690943988e-06, "loss": 3.2959, "step": 15679 }, { "epoch": 0.5525163631878925, "grad_norm": 1.7753148078918457, "learning_rate": 8.815290443849955e-06, "loss": 1.0379, "step": 15680 }, { "epoch": 0.5525516002008509, "grad_norm": 5.712280750274658, "learning_rate": 8.814157212186915e-06, "loss": 5.232, "step": 15681 }, { "epoch": 0.5525868372138094, "grad_norm": 1.9531235694885254, "learning_rate": 8.813023995969625e-06, "loss": 0.8195, "step": 15682 }, { "epoch": 0.5526220742267678, "grad_norm": 0.9045756459236145, "learning_rate": 8.811890795212854e-06, "loss": 0.8479, "step": 15683 }, { "epoch": 0.5526573112397262, "grad_norm": 1.6462218761444092, "learning_rate": 8.810757609931354e-06, "loss": 1.0726, "step": 15684 }, { "epoch": 0.5526925482526847, "grad_norm": 3.0931406021118164, "learning_rate": 8.809624440139888e-06, "loss": 3.0187, "step": 15685 }, { "epoch": 0.552727785265643, "grad_norm": 5.855481147766113, "learning_rate": 8.808491285853217e-06, "loss": 5.6727, "step": 15686 }, { "epoch": 0.5527630222786014, "grad_norm": 4.235279560089111, "learning_rate": 8.807358147086095e-06, "loss": 2.7445, "step": 15687 }, { "epoch": 0.5527982592915599, "grad_norm": 3.8725228309631348, "learning_rate": 8.806225023853287e-06, "loss": 2.4346, "step": 15688 }, { "epoch": 0.5528334963045183, "grad_norm": 1.7006103992462158, "learning_rate": 8.805091916169552e-06, "loss": 1.057, "step": 15689 }, { "epoch": 0.5528687333174767, "grad_norm": 1.0761151313781738, "learning_rate": 8.803958824049646e-06, "loss": 0.925, "step": 15690 }, { "epoch": 0.552903970330435, "grad_norm": 4.036915302276611, "learning_rate": 8.802825747508324e-06, "loss": 3.4407, "step": 15691 }, { "epoch": 0.5529392073433935, "grad_norm": 6.430813312530518, "learning_rate": 8.801692686560359e-06, "loss": 5.588, "step": 15692 }, { "epoch": 0.5529744443563519, "grad_norm": 1.3944050073623657, "learning_rate": 8.80055964122049e-06, "loss": 0.8904, "step": 15693 }, { "epoch": 0.5530096813693103, "grad_norm": 3.898895502090454, "learning_rate": 8.799426611503488e-06, "loss": 2.3833, "step": 15694 }, { "epoch": 0.5530449183822688, "grad_norm": 5.672348499298096, "learning_rate": 8.798293597424109e-06, "loss": 3.0277, "step": 15695 }, { "epoch": 0.5530801553952271, "grad_norm": 1.1636033058166504, "learning_rate": 8.797160598997108e-06, "loss": 1.0602, "step": 15696 }, { "epoch": 0.5531153924081855, "grad_norm": 3.1539671421051025, "learning_rate": 8.796027616237239e-06, "loss": 2.9307, "step": 15697 }, { "epoch": 0.553150629421144, "grad_norm": 1.4049571752548218, "learning_rate": 8.794894649159267e-06, "loss": 0.7286, "step": 15698 }, { "epoch": 0.5531858664341024, "grad_norm": 3.278477191925049, "learning_rate": 8.793761697777948e-06, "loss": 2.9623, "step": 15699 }, { "epoch": 0.5532211034470608, "grad_norm": 2.9226233959198, "learning_rate": 8.792628762108034e-06, "loss": 2.9889, "step": 15700 }, { "epoch": 0.5532563404600193, "grad_norm": 1.289180874824524, "learning_rate": 8.791495842164283e-06, "loss": 0.8814, "step": 15701 }, { "epoch": 0.5532915774729776, "grad_norm": 1.6463701725006104, "learning_rate": 8.790362937961456e-06, "loss": 0.8772, "step": 15702 }, { "epoch": 0.553326814485936, "grad_norm": 4.532435417175293, "learning_rate": 8.789230049514303e-06, "loss": 3.7258, "step": 15703 }, { "epoch": 0.5533620514988944, "grad_norm": 6.488966941833496, "learning_rate": 8.788097176837584e-06, "loss": 5.2679, "step": 15704 }, { "epoch": 0.5533972885118529, "grad_norm": 3.330165386199951, "learning_rate": 8.786964319946056e-06, "loss": 2.8223, "step": 15705 }, { "epoch": 0.5534325255248113, "grad_norm": 1.0262666940689087, "learning_rate": 8.785831478854467e-06, "loss": 0.8425, "step": 15706 }, { "epoch": 0.5534677625377696, "grad_norm": 1.3845237493515015, "learning_rate": 8.784698653577582e-06, "loss": 0.743, "step": 15707 }, { "epoch": 0.5535029995507281, "grad_norm": 1.0300166606903076, "learning_rate": 8.783565844130151e-06, "loss": 0.8639, "step": 15708 }, { "epoch": 0.5535382365636865, "grad_norm": 1.2941536903381348, "learning_rate": 8.78243305052693e-06, "loss": 0.796, "step": 15709 }, { "epoch": 0.5535734735766449, "grad_norm": 4.352117538452148, "learning_rate": 8.78130027278267e-06, "loss": 4.9402, "step": 15710 }, { "epoch": 0.5536087105896034, "grad_norm": 1.1734209060668945, "learning_rate": 8.780167510912135e-06, "loss": 0.6219, "step": 15711 }, { "epoch": 0.5536439476025617, "grad_norm": 1.3482547998428345, "learning_rate": 8.779034764930069e-06, "loss": 0.8261, "step": 15712 }, { "epoch": 0.5536791846155201, "grad_norm": 0.9241886138916016, "learning_rate": 8.777902034851232e-06, "loss": 1.059, "step": 15713 }, { "epoch": 0.5537144216284785, "grad_norm": 5.5113911628723145, "learning_rate": 8.77676932069038e-06, "loss": 4.4695, "step": 15714 }, { "epoch": 0.553749658641437, "grad_norm": 3.648700714111328, "learning_rate": 8.77563662246226e-06, "loss": 2.7795, "step": 15715 }, { "epoch": 0.5537848956543954, "grad_norm": 5.970643997192383, "learning_rate": 8.774503940181625e-06, "loss": 3.2113, "step": 15716 }, { "epoch": 0.5538201326673537, "grad_norm": 2.635338306427002, "learning_rate": 8.773371273863237e-06, "loss": 1.0934, "step": 15717 }, { "epoch": 0.5538553696803122, "grad_norm": 4.38990592956543, "learning_rate": 8.772238623521846e-06, "loss": 3.0715, "step": 15718 }, { "epoch": 0.5538906066932706, "grad_norm": 1.76427161693573, "learning_rate": 8.771105989172201e-06, "loss": 0.9494, "step": 15719 }, { "epoch": 0.553925843706229, "grad_norm": 1.2951385974884033, "learning_rate": 8.769973370829054e-06, "loss": 0.9914, "step": 15720 }, { "epoch": 0.5539610807191875, "grad_norm": 36.102989196777344, "learning_rate": 8.768840768507166e-06, "loss": 4.6902, "step": 15721 }, { "epoch": 0.5539963177321459, "grad_norm": 1.2998765707015991, "learning_rate": 8.767708182221282e-06, "loss": 0.898, "step": 15722 }, { "epoch": 0.5540315547451042, "grad_norm": 4.77347469329834, "learning_rate": 8.766575611986156e-06, "loss": 3.155, "step": 15723 }, { "epoch": 0.5540667917580626, "grad_norm": 1.1444791555404663, "learning_rate": 8.765443057816541e-06, "loss": 0.9927, "step": 15724 }, { "epoch": 0.5541020287710211, "grad_norm": 5.749847888946533, "learning_rate": 8.764310519727183e-06, "loss": 5.4201, "step": 15725 }, { "epoch": 0.5541372657839795, "grad_norm": 2.6598119735717773, "learning_rate": 8.763177997732841e-06, "loss": 2.6495, "step": 15726 }, { "epoch": 0.5541725027969379, "grad_norm": 2.783998966217041, "learning_rate": 8.762045491848263e-06, "loss": 1.0384, "step": 15727 }, { "epoch": 0.5542077398098963, "grad_norm": 1.3798736333847046, "learning_rate": 8.7609130020882e-06, "loss": 1.0578, "step": 15728 }, { "epoch": 0.5542429768228547, "grad_norm": 1.1978336572647095, "learning_rate": 8.7597805284674e-06, "loss": 0.7787, "step": 15729 }, { "epoch": 0.5542782138358131, "grad_norm": 3.7122206687927246, "learning_rate": 8.758648071000625e-06, "loss": 3.0891, "step": 15730 }, { "epoch": 0.5543134508487716, "grad_norm": 1.6136246919631958, "learning_rate": 8.757515629702609e-06, "loss": 0.7843, "step": 15731 }, { "epoch": 0.55434868786173, "grad_norm": 1.6247457265853882, "learning_rate": 8.756383204588113e-06, "loss": 0.9451, "step": 15732 }, { "epoch": 0.5543839248746883, "grad_norm": 5.226813316345215, "learning_rate": 8.75525079567188e-06, "loss": 2.4293, "step": 15733 }, { "epoch": 0.5544191618876468, "grad_norm": 5.557923793792725, "learning_rate": 8.754118402968671e-06, "loss": 3.2927, "step": 15734 }, { "epoch": 0.5544543989006052, "grad_norm": 1.1634026765823364, "learning_rate": 8.752986026493225e-06, "loss": 0.9631, "step": 15735 }, { "epoch": 0.5544896359135636, "grad_norm": 7.030352592468262, "learning_rate": 8.751853666260294e-06, "loss": 5.0374, "step": 15736 }, { "epoch": 0.554524872926522, "grad_norm": 3.6235480308532715, "learning_rate": 8.750721322284632e-06, "loss": 3.3401, "step": 15737 }, { "epoch": 0.5545601099394805, "grad_norm": 1.0956592559814453, "learning_rate": 8.74958899458098e-06, "loss": 0.9531, "step": 15738 }, { "epoch": 0.5545953469524388, "grad_norm": 32.93299102783203, "learning_rate": 8.748456683164089e-06, "loss": 9.8881, "step": 15739 }, { "epoch": 0.5546305839653972, "grad_norm": 3.8356361389160156, "learning_rate": 8.747324388048713e-06, "loss": 3.2237, "step": 15740 }, { "epoch": 0.5546658209783557, "grad_norm": 4.711119651794434, "learning_rate": 8.746192109249595e-06, "loss": 3.1718, "step": 15741 }, { "epoch": 0.5547010579913141, "grad_norm": 6.686392307281494, "learning_rate": 8.745059846781485e-06, "loss": 5.204, "step": 15742 }, { "epoch": 0.5547362950042725, "grad_norm": 4.114997863769531, "learning_rate": 8.743927600659132e-06, "loss": 2.9131, "step": 15743 }, { "epoch": 0.554771532017231, "grad_norm": 2.468078136444092, "learning_rate": 8.742795370897278e-06, "loss": 0.7877, "step": 15744 }, { "epoch": 0.5548067690301893, "grad_norm": 1.179102897644043, "learning_rate": 8.741663157510676e-06, "loss": 1.0403, "step": 15745 }, { "epoch": 0.5548420060431477, "grad_norm": 1.1012500524520874, "learning_rate": 8.740530960514074e-06, "loss": 0.8801, "step": 15746 }, { "epoch": 0.5548772430561061, "grad_norm": 9.545223236083984, "learning_rate": 8.739398779922214e-06, "loss": 3.3286, "step": 15747 }, { "epoch": 0.5549124800690646, "grad_norm": 6.975803852081299, "learning_rate": 8.738266615749845e-06, "loss": 5.7464, "step": 15748 }, { "epoch": 0.554947717082023, "grad_norm": 90.04816436767578, "learning_rate": 8.737134468011719e-06, "loss": 5.791, "step": 15749 }, { "epoch": 0.5549829540949813, "grad_norm": 5.20267915725708, "learning_rate": 8.73600233672257e-06, "loss": 3.276, "step": 15750 }, { "epoch": 0.5550181911079398, "grad_norm": 2.1787595748901367, "learning_rate": 8.734870221897156e-06, "loss": 0.8568, "step": 15751 }, { "epoch": 0.5550534281208982, "grad_norm": 3.9985687732696533, "learning_rate": 8.733738123550214e-06, "loss": 2.5537, "step": 15752 }, { "epoch": 0.5550886651338566, "grad_norm": 92.57942962646484, "learning_rate": 8.732606041696504e-06, "loss": 5.0437, "step": 15753 }, { "epoch": 0.5551239021468151, "grad_norm": 1.408620834350586, "learning_rate": 8.731473976350754e-06, "loss": 0.7337, "step": 15754 }, { "epoch": 0.5551591391597734, "grad_norm": 12.39826488494873, "learning_rate": 8.730341927527719e-06, "loss": 6.9179, "step": 15755 }, { "epoch": 0.5551943761727318, "grad_norm": 3.976710081100464, "learning_rate": 8.729209895242142e-06, "loss": 2.5971, "step": 15756 }, { "epoch": 0.5552296131856902, "grad_norm": 29.29963493347168, "learning_rate": 8.728077879508769e-06, "loss": 6.64, "step": 15757 }, { "epoch": 0.5552648501986487, "grad_norm": 1.02353835105896, "learning_rate": 8.72694588034234e-06, "loss": 1.1163, "step": 15758 }, { "epoch": 0.5553000872116071, "grad_norm": 1.7465916872024536, "learning_rate": 8.725813897757606e-06, "loss": 0.9067, "step": 15759 }, { "epoch": 0.5553353242245654, "grad_norm": 18.308544158935547, "learning_rate": 8.72468193176931e-06, "loss": 3.0524, "step": 15760 }, { "epoch": 0.5553705612375239, "grad_norm": 38.22103500366211, "learning_rate": 8.723549982392191e-06, "loss": 3.2534, "step": 15761 }, { "epoch": 0.5554057982504823, "grad_norm": 14.339707374572754, "learning_rate": 8.722418049641e-06, "loss": 3.0707, "step": 15762 }, { "epoch": 0.5554410352634407, "grad_norm": 1.1134859323501587, "learning_rate": 8.72128613353047e-06, "loss": 1.0419, "step": 15763 }, { "epoch": 0.5554762722763992, "grad_norm": 11.81836986541748, "learning_rate": 8.720154234075355e-06, "loss": 5.7811, "step": 15764 }, { "epoch": 0.5555115092893576, "grad_norm": 8.276741027832031, "learning_rate": 8.719022351290395e-06, "loss": 2.6501, "step": 15765 }, { "epoch": 0.5555467463023159, "grad_norm": 19.220348358154297, "learning_rate": 8.717890485190333e-06, "loss": 5.5942, "step": 15766 }, { "epoch": 0.5555819833152744, "grad_norm": 29.474430084228516, "learning_rate": 8.716758635789906e-06, "loss": 3.7503, "step": 15767 }, { "epoch": 0.5556172203282328, "grad_norm": 1.8360732793807983, "learning_rate": 8.715626803103867e-06, "loss": 0.9085, "step": 15768 }, { "epoch": 0.5556524573411912, "grad_norm": 11.138704299926758, "learning_rate": 8.714494987146948e-06, "loss": 5.5952, "step": 15769 }, { "epoch": 0.5556876943541496, "grad_norm": 1.9724737405776978, "learning_rate": 8.713363187933896e-06, "loss": 0.9908, "step": 15770 }, { "epoch": 0.555722931367108, "grad_norm": 16.468093872070312, "learning_rate": 8.712231405479452e-06, "loss": 5.9496, "step": 15771 }, { "epoch": 0.5557581683800664, "grad_norm": 5.364832401275635, "learning_rate": 8.711099639798364e-06, "loss": 3.103, "step": 15772 }, { "epoch": 0.5557934053930248, "grad_norm": 8.020282745361328, "learning_rate": 8.70996789090536e-06, "loss": 3.411, "step": 15773 }, { "epoch": 0.5558286424059833, "grad_norm": 1.3055732250213623, "learning_rate": 8.70883615881519e-06, "loss": 0.9485, "step": 15774 }, { "epoch": 0.5558638794189417, "grad_norm": 13.011590003967285, "learning_rate": 8.707704443542598e-06, "loss": 6.5028, "step": 15775 }, { "epoch": 0.5558991164319, "grad_norm": 5.9821367263793945, "learning_rate": 8.706572745102316e-06, "loss": 3.1648, "step": 15776 }, { "epoch": 0.5559343534448585, "grad_norm": 1.2234470844268799, "learning_rate": 8.705441063509087e-06, "loss": 0.8688, "step": 15777 }, { "epoch": 0.5559695904578169, "grad_norm": 4.2874226570129395, "learning_rate": 8.704309398777659e-06, "loss": 3.5334, "step": 15778 }, { "epoch": 0.5560048274707753, "grad_norm": 1.631170392036438, "learning_rate": 8.703177750922762e-06, "loss": 0.7617, "step": 15779 }, { "epoch": 0.5560400644837337, "grad_norm": 1.204978108406067, "learning_rate": 8.70204611995914e-06, "loss": 0.7148, "step": 15780 }, { "epoch": 0.5560753014966922, "grad_norm": 1.8392170667648315, "learning_rate": 8.700914505901536e-06, "loss": 0.696, "step": 15781 }, { "epoch": 0.5561105385096505, "grad_norm": 4.709309101104736, "learning_rate": 8.699782908764683e-06, "loss": 3.051, "step": 15782 }, { "epoch": 0.5561457755226089, "grad_norm": 1.4996227025985718, "learning_rate": 8.698651328563324e-06, "loss": 1.1393, "step": 15783 }, { "epoch": 0.5561810125355674, "grad_norm": 1.8563164472579956, "learning_rate": 8.697519765312197e-06, "loss": 0.9563, "step": 15784 }, { "epoch": 0.5562162495485258, "grad_norm": 1.3426607847213745, "learning_rate": 8.696388219026041e-06, "loss": 0.8061, "step": 15785 }, { "epoch": 0.5562514865614842, "grad_norm": 1.5507314205169678, "learning_rate": 8.695256689719594e-06, "loss": 0.9478, "step": 15786 }, { "epoch": 0.5562867235744426, "grad_norm": 1.1732584238052368, "learning_rate": 8.694125177407599e-06, "loss": 0.8803, "step": 15787 }, { "epoch": 0.556321960587401, "grad_norm": 15.107514381408691, "learning_rate": 8.692993682104784e-06, "loss": 5.3323, "step": 15788 }, { "epoch": 0.5563571976003594, "grad_norm": 2.5708446502685547, "learning_rate": 8.691862203825894e-06, "loss": 0.8426, "step": 15789 }, { "epoch": 0.5563924346133178, "grad_norm": 1.1188368797302246, "learning_rate": 8.690730742585667e-06, "loss": 0.7665, "step": 15790 }, { "epoch": 0.5564276716262763, "grad_norm": 4.693960666656494, "learning_rate": 8.68959929839884e-06, "loss": 3.0235, "step": 15791 }, { "epoch": 0.5564629086392346, "grad_norm": 1.8212908506393433, "learning_rate": 8.688467871280146e-06, "loss": 0.6433, "step": 15792 }, { "epoch": 0.556498145652193, "grad_norm": 0.867228627204895, "learning_rate": 8.687336461244325e-06, "loss": 0.9988, "step": 15793 }, { "epoch": 0.5565333826651515, "grad_norm": 4.357036590576172, "learning_rate": 8.686205068306118e-06, "loss": 2.8852, "step": 15794 }, { "epoch": 0.5565686196781099, "grad_norm": 1.666565179824829, "learning_rate": 8.685073692480253e-06, "loss": 1.1933, "step": 15795 }, { "epoch": 0.5566038566910683, "grad_norm": 2.296692132949829, "learning_rate": 8.68394233378147e-06, "loss": 1.0892, "step": 15796 }, { "epoch": 0.5566390937040268, "grad_norm": 3.6293787956237793, "learning_rate": 8.682810992224511e-06, "loss": 2.8224, "step": 15797 }, { "epoch": 0.5566743307169851, "grad_norm": 1.7607190608978271, "learning_rate": 8.681679667824104e-06, "loss": 0.7813, "step": 15798 }, { "epoch": 0.5567095677299435, "grad_norm": 1.839406967163086, "learning_rate": 8.680548360594984e-06, "loss": 0.861, "step": 15799 }, { "epoch": 0.5567448047429019, "grad_norm": 4.364367485046387, "learning_rate": 8.679417070551895e-06, "loss": 3.2078, "step": 15800 }, { "epoch": 0.5567800417558604, "grad_norm": 1.863008737564087, "learning_rate": 8.678285797709562e-06, "loss": 0.831, "step": 15801 }, { "epoch": 0.5568152787688188, "grad_norm": 1.3557533025741577, "learning_rate": 8.677154542082727e-06, "loss": 0.9727, "step": 15802 }, { "epoch": 0.5568505157817771, "grad_norm": 4.224435806274414, "learning_rate": 8.676023303686125e-06, "loss": 2.7137, "step": 15803 }, { "epoch": 0.5568857527947356, "grad_norm": 1.941758632659912, "learning_rate": 8.674892082534486e-06, "loss": 0.7993, "step": 15804 }, { "epoch": 0.556920989807694, "grad_norm": 7.172891139984131, "learning_rate": 8.673760878642547e-06, "loss": 3.0577, "step": 15805 }, { "epoch": 0.5569562268206524, "grad_norm": 2.3296291828155518, "learning_rate": 8.672629692025037e-06, "loss": 0.7901, "step": 15806 }, { "epoch": 0.5569914638336109, "grad_norm": 1.3152849674224854, "learning_rate": 8.671498522696701e-06, "loss": 1.0819, "step": 15807 }, { "epoch": 0.5570267008465692, "grad_norm": 1.4879944324493408, "learning_rate": 8.670367370672262e-06, "loss": 1.1103, "step": 15808 }, { "epoch": 0.5570619378595276, "grad_norm": 8.181199073791504, "learning_rate": 8.669236235966459e-06, "loss": 7.3246, "step": 15809 }, { "epoch": 0.5570971748724861, "grad_norm": 1.3652523756027222, "learning_rate": 8.668105118594026e-06, "loss": 0.9207, "step": 15810 }, { "epoch": 0.5571324118854445, "grad_norm": 1.3750941753387451, "learning_rate": 8.666974018569687e-06, "loss": 1.2031, "step": 15811 }, { "epoch": 0.5571676488984029, "grad_norm": 7.488258361816406, "learning_rate": 8.665842935908183e-06, "loss": 3.0442, "step": 15812 }, { "epoch": 0.5572028859113612, "grad_norm": 1.496668815612793, "learning_rate": 8.66471187062425e-06, "loss": 0.8119, "step": 15813 }, { "epoch": 0.5572381229243197, "grad_norm": 9.590590476989746, "learning_rate": 8.663580822732612e-06, "loss": 3.2739, "step": 15814 }, { "epoch": 0.5572733599372781, "grad_norm": 1.319146990776062, "learning_rate": 8.662449792248e-06, "loss": 0.9287, "step": 15815 }, { "epoch": 0.5573085969502365, "grad_norm": 1.1178309917449951, "learning_rate": 8.661318779185156e-06, "loss": 0.929, "step": 15816 }, { "epoch": 0.557343833963195, "grad_norm": 1.240429162979126, "learning_rate": 8.660187783558802e-06, "loss": 0.6758, "step": 15817 }, { "epoch": 0.5573790709761534, "grad_norm": 1.2651572227478027, "learning_rate": 8.659056805383673e-06, "loss": 0.9419, "step": 15818 }, { "epoch": 0.5574143079891117, "grad_norm": 1.2410705089569092, "learning_rate": 8.657925844674503e-06, "loss": 0.9089, "step": 15819 }, { "epoch": 0.5574495450020702, "grad_norm": 6.012125015258789, "learning_rate": 8.656794901446017e-06, "loss": 5.1265, "step": 15820 }, { "epoch": 0.5574847820150286, "grad_norm": 2.3656015396118164, "learning_rate": 8.655663975712945e-06, "loss": 0.9408, "step": 15821 }, { "epoch": 0.557520019027987, "grad_norm": 1.6480435132980347, "learning_rate": 8.654533067490027e-06, "loss": 1.1421, "step": 15822 }, { "epoch": 0.5575552560409454, "grad_norm": 4.536592960357666, "learning_rate": 8.653402176791985e-06, "loss": 3.3797, "step": 15823 }, { "epoch": 0.5575904930539038, "grad_norm": 0.9314783811569214, "learning_rate": 8.652271303633552e-06, "loss": 0.6648, "step": 15824 }, { "epoch": 0.5576257300668622, "grad_norm": 1.1389080286026, "learning_rate": 8.651140448029453e-06, "loss": 0.8237, "step": 15825 }, { "epoch": 0.5576609670798206, "grad_norm": 4.670940399169922, "learning_rate": 8.650009609994426e-06, "loss": 3.9131, "step": 15826 }, { "epoch": 0.5576962040927791, "grad_norm": 4.06890869140625, "learning_rate": 8.648878789543194e-06, "loss": 2.5598, "step": 15827 }, { "epoch": 0.5577314411057375, "grad_norm": 7.468679904937744, "learning_rate": 8.647747986690488e-06, "loss": 5.6562, "step": 15828 }, { "epoch": 0.5577666781186958, "grad_norm": 0.8964110612869263, "learning_rate": 8.646617201451038e-06, "loss": 0.8882, "step": 15829 }, { "epoch": 0.5578019151316543, "grad_norm": 1.988752007484436, "learning_rate": 8.645486433839568e-06, "loss": 0.9797, "step": 15830 }, { "epoch": 0.5578371521446127, "grad_norm": 7.491117000579834, "learning_rate": 8.644355683870812e-06, "loss": 5.5153, "step": 15831 }, { "epoch": 0.5578723891575711, "grad_norm": 1.6067866086959839, "learning_rate": 8.643224951559498e-06, "loss": 1.1402, "step": 15832 }, { "epoch": 0.5579076261705295, "grad_norm": 2.12977933883667, "learning_rate": 8.64209423692035e-06, "loss": 1.0716, "step": 15833 }, { "epoch": 0.557942863183488, "grad_norm": 2.061694622039795, "learning_rate": 8.640963539968094e-06, "loss": 0.7823, "step": 15834 }, { "epoch": 0.5579781001964463, "grad_norm": 8.92520809173584, "learning_rate": 8.639832860717468e-06, "loss": 5.2701, "step": 15835 }, { "epoch": 0.5580133372094047, "grad_norm": 1.8971501588821411, "learning_rate": 8.638702199183187e-06, "loss": 0.7567, "step": 15836 }, { "epoch": 0.5580485742223632, "grad_norm": 1.783644437789917, "learning_rate": 8.637571555379984e-06, "loss": 0.9516, "step": 15837 }, { "epoch": 0.5580838112353216, "grad_norm": 1.4211573600769043, "learning_rate": 8.636440929322588e-06, "loss": 0.8472, "step": 15838 }, { "epoch": 0.55811904824828, "grad_norm": 5.976417541503906, "learning_rate": 8.63531032102572e-06, "loss": 5.7512, "step": 15839 }, { "epoch": 0.5581542852612384, "grad_norm": 1.4298659563064575, "learning_rate": 8.634179730504106e-06, "loss": 0.7288, "step": 15840 }, { "epoch": 0.5581895222741968, "grad_norm": 1.3014097213745117, "learning_rate": 8.633049157772481e-06, "loss": 0.838, "step": 15841 }, { "epoch": 0.5582247592871552, "grad_norm": 21.138713836669922, "learning_rate": 8.631918602845561e-06, "loss": 5.0095, "step": 15842 }, { "epoch": 0.5582599963001137, "grad_norm": 14.286582946777344, "learning_rate": 8.630788065738077e-06, "loss": 5.5349, "step": 15843 }, { "epoch": 0.5582952333130721, "grad_norm": 5.843650817871094, "learning_rate": 8.62965754646475e-06, "loss": 2.765, "step": 15844 }, { "epoch": 0.5583304703260304, "grad_norm": 3.758619546890259, "learning_rate": 8.628527045040312e-06, "loss": 4.2359, "step": 15845 }, { "epoch": 0.5583657073389888, "grad_norm": 4.928225517272949, "learning_rate": 8.627396561479483e-06, "loss": 3.1305, "step": 15846 }, { "epoch": 0.5584009443519473, "grad_norm": 5.930572986602783, "learning_rate": 8.626266095796986e-06, "loss": 3.2519, "step": 15847 }, { "epoch": 0.5584361813649057, "grad_norm": 5.1706695556640625, "learning_rate": 8.625135648007553e-06, "loss": 3.0979, "step": 15848 }, { "epoch": 0.5584714183778641, "grad_norm": 7.126402378082275, "learning_rate": 8.624005218125898e-06, "loss": 3.3283, "step": 15849 }, { "epoch": 0.5585066553908226, "grad_norm": 5.955796718597412, "learning_rate": 8.622874806166752e-06, "loss": 4.7798, "step": 15850 }, { "epoch": 0.5585418924037809, "grad_norm": 1.3102604150772095, "learning_rate": 8.62174441214484e-06, "loss": 0.9107, "step": 15851 }, { "epoch": 0.5585771294167393, "grad_norm": 5.670228481292725, "learning_rate": 8.620614036074881e-06, "loss": 3.2608, "step": 15852 }, { "epoch": 0.5586123664296978, "grad_norm": 2.6112682819366455, "learning_rate": 8.619483677971596e-06, "loss": 0.8774, "step": 15853 }, { "epoch": 0.5586476034426562, "grad_norm": 4.9059014320373535, "learning_rate": 8.61835333784972e-06, "loss": 5.2185, "step": 15854 }, { "epoch": 0.5586828404556146, "grad_norm": 4.4752631187438965, "learning_rate": 8.61722301572396e-06, "loss": 3.5929, "step": 15855 }, { "epoch": 0.5587180774685729, "grad_norm": 1.0882911682128906, "learning_rate": 8.61609271160905e-06, "loss": 0.7441, "step": 15856 }, { "epoch": 0.5587533144815314, "grad_norm": 4.732077598571777, "learning_rate": 8.614962425519709e-06, "loss": 3.4614, "step": 15857 }, { "epoch": 0.5587885514944898, "grad_norm": 5.00331974029541, "learning_rate": 8.613832157470658e-06, "loss": 5.2101, "step": 15858 }, { "epoch": 0.5588237885074482, "grad_norm": 2.2578976154327393, "learning_rate": 8.612701907476618e-06, "loss": 0.9876, "step": 15859 }, { "epoch": 0.5588590255204067, "grad_norm": 1.4226261377334595, "learning_rate": 8.611571675552314e-06, "loss": 0.7482, "step": 15860 }, { "epoch": 0.558894262533365, "grad_norm": 1.8838164806365967, "learning_rate": 8.610441461712466e-06, "loss": 0.9755, "step": 15861 }, { "epoch": 0.5589294995463234, "grad_norm": 5.849233150482178, "learning_rate": 8.609311265971795e-06, "loss": 2.9957, "step": 15862 }, { "epoch": 0.5589647365592819, "grad_norm": 5.349681377410889, "learning_rate": 8.60818108834502e-06, "loss": 5.4754, "step": 15863 }, { "epoch": 0.5589999735722403, "grad_norm": 1.8001368045806885, "learning_rate": 8.607050928846866e-06, "loss": 0.8278, "step": 15864 }, { "epoch": 0.5590352105851987, "grad_norm": 2.145636796951294, "learning_rate": 8.60592078749205e-06, "loss": 0.8415, "step": 15865 }, { "epoch": 0.559070447598157, "grad_norm": 1.214697241783142, "learning_rate": 8.604790664295295e-06, "loss": 0.898, "step": 15866 }, { "epoch": 0.5591056846111155, "grad_norm": 2.8328776359558105, "learning_rate": 8.603660559271319e-06, "loss": 3.0415, "step": 15867 }, { "epoch": 0.5591409216240739, "grad_norm": 5.880095958709717, "learning_rate": 8.602530472434839e-06, "loss": 4.6114, "step": 15868 }, { "epoch": 0.5591761586370323, "grad_norm": 1.3769350051879883, "learning_rate": 8.601400403800578e-06, "loss": 0.916, "step": 15869 }, { "epoch": 0.5592113956499908, "grad_norm": 2.182239532470703, "learning_rate": 8.60027035338326e-06, "loss": 0.8804, "step": 15870 }, { "epoch": 0.5592466326629492, "grad_norm": 11.747634887695312, "learning_rate": 8.599140321197594e-06, "loss": 5.2505, "step": 15871 }, { "epoch": 0.5592818696759075, "grad_norm": 0.9724463224411011, "learning_rate": 8.598010307258303e-06, "loss": 1.0073, "step": 15872 }, { "epoch": 0.559317106688866, "grad_norm": 0.9326531291007996, "learning_rate": 8.596880311580112e-06, "loss": 1.0341, "step": 15873 }, { "epoch": 0.5593523437018244, "grad_norm": 4.2679266929626465, "learning_rate": 8.595750334177728e-06, "loss": 2.8728, "step": 15874 }, { "epoch": 0.5593875807147828, "grad_norm": 4.123842239379883, "learning_rate": 8.594620375065878e-06, "loss": 3.0329, "step": 15875 }, { "epoch": 0.5594228177277413, "grad_norm": 1.7871320247650146, "learning_rate": 8.593490434259277e-06, "loss": 0.9915, "step": 15876 }, { "epoch": 0.5594580547406997, "grad_norm": 3.2109310626983643, "learning_rate": 8.592360511772641e-06, "loss": 2.8906, "step": 15877 }, { "epoch": 0.559493291753658, "grad_norm": 7.7974934577941895, "learning_rate": 8.591230607620686e-06, "loss": 3.3655, "step": 15878 }, { "epoch": 0.5595285287666164, "grad_norm": 2.53446888923645, "learning_rate": 8.590100721818133e-06, "loss": 1.0657, "step": 15879 }, { "epoch": 0.5595637657795749, "grad_norm": 1.636070728302002, "learning_rate": 8.588970854379701e-06, "loss": 1.0352, "step": 15880 }, { "epoch": 0.5595990027925333, "grad_norm": 1.9752756357192993, "learning_rate": 8.587841005320102e-06, "loss": 0.9407, "step": 15881 }, { "epoch": 0.5596342398054917, "grad_norm": 7.1158766746521, "learning_rate": 8.58671117465405e-06, "loss": 6.8513, "step": 15882 }, { "epoch": 0.5596694768184501, "grad_norm": 1.4840610027313232, "learning_rate": 8.58558136239627e-06, "loss": 0.9423, "step": 15883 }, { "epoch": 0.5597047138314085, "grad_norm": 3.0816285610198975, "learning_rate": 8.584451568561472e-06, "loss": 0.9468, "step": 15884 }, { "epoch": 0.5597399508443669, "grad_norm": 1.0636314153671265, "learning_rate": 8.58332179316437e-06, "loss": 1.0083, "step": 15885 }, { "epoch": 0.5597751878573254, "grad_norm": 4.76178503036499, "learning_rate": 8.582192036219687e-06, "loss": 2.8658, "step": 15886 }, { "epoch": 0.5598104248702838, "grad_norm": 5.4486494064331055, "learning_rate": 8.581062297742126e-06, "loss": 2.543, "step": 15887 }, { "epoch": 0.5598456618832421, "grad_norm": 4.837399959564209, "learning_rate": 8.579932577746414e-06, "loss": 2.9393, "step": 15888 }, { "epoch": 0.5598808988962005, "grad_norm": 8.173172950744629, "learning_rate": 8.578802876247261e-06, "loss": 5.0694, "step": 15889 }, { "epoch": 0.559916135909159, "grad_norm": 3.6977908611297607, "learning_rate": 8.577673193259382e-06, "loss": 0.9625, "step": 15890 }, { "epoch": 0.5599513729221174, "grad_norm": 5.341900825500488, "learning_rate": 8.576543528797488e-06, "loss": 3.125, "step": 15891 }, { "epoch": 0.5599866099350758, "grad_norm": 1.1366695165634155, "learning_rate": 8.575413882876302e-06, "loss": 1.3398, "step": 15892 }, { "epoch": 0.5600218469480343, "grad_norm": 14.427437782287598, "learning_rate": 8.574284255510524e-06, "loss": 3.4922, "step": 15893 }, { "epoch": 0.5600570839609926, "grad_norm": 3.635638475418091, "learning_rate": 8.573154646714878e-06, "loss": 2.885, "step": 15894 }, { "epoch": 0.560092320973951, "grad_norm": 1.125637173652649, "learning_rate": 8.572025056504077e-06, "loss": 1.1656, "step": 15895 }, { "epoch": 0.5601275579869095, "grad_norm": 1.4868310689926147, "learning_rate": 8.57089548489283e-06, "loss": 0.8152, "step": 15896 }, { "epoch": 0.5601627949998679, "grad_norm": 1.895381212234497, "learning_rate": 8.569765931895848e-06, "loss": 0.8773, "step": 15897 }, { "epoch": 0.5601980320128263, "grad_norm": 1.057325005531311, "learning_rate": 8.568636397527849e-06, "loss": 0.8009, "step": 15898 }, { "epoch": 0.5602332690257846, "grad_norm": 8.312463760375977, "learning_rate": 8.567506881803547e-06, "loss": 4.6272, "step": 15899 }, { "epoch": 0.5602685060387431, "grad_norm": 5.983716011047363, "learning_rate": 8.566377384737647e-06, "loss": 5.1791, "step": 15900 }, { "epoch": 0.5603037430517015, "grad_norm": 1.2767184972763062, "learning_rate": 8.565247906344862e-06, "loss": 0.9084, "step": 15901 }, { "epoch": 0.5603389800646599, "grad_norm": 1.5487756729125977, "learning_rate": 8.564118446639911e-06, "loss": 0.9312, "step": 15902 }, { "epoch": 0.5603742170776184, "grad_norm": 1.6073760986328125, "learning_rate": 8.562989005637496e-06, "loss": 0.9293, "step": 15903 }, { "epoch": 0.5604094540905767, "grad_norm": 1.5328729152679443, "learning_rate": 8.561859583352334e-06, "loss": 0.8352, "step": 15904 }, { "epoch": 0.5604446911035351, "grad_norm": 6.587606430053711, "learning_rate": 8.560730179799136e-06, "loss": 4.0031, "step": 15905 }, { "epoch": 0.5604799281164936, "grad_norm": 1.4279282093048096, "learning_rate": 8.559600794992607e-06, "loss": 0.9798, "step": 15906 }, { "epoch": 0.560515165129452, "grad_norm": 5.6081929206848145, "learning_rate": 8.558471428947461e-06, "loss": 4.8894, "step": 15907 }, { "epoch": 0.5605504021424104, "grad_norm": 1.8814724683761597, "learning_rate": 8.557342081678413e-06, "loss": 0.9453, "step": 15908 }, { "epoch": 0.5605856391553689, "grad_norm": 5.696958541870117, "learning_rate": 8.556212753200164e-06, "loss": 3.0674, "step": 15909 }, { "epoch": 0.5606208761683272, "grad_norm": 6.67974328994751, "learning_rate": 8.555083443527428e-06, "loss": 2.9195, "step": 15910 }, { "epoch": 0.5606561131812856, "grad_norm": 6.004271984100342, "learning_rate": 8.553954152674918e-06, "loss": 3.2716, "step": 15911 }, { "epoch": 0.560691350194244, "grad_norm": 2.8809075355529785, "learning_rate": 8.552824880657334e-06, "loss": 0.8681, "step": 15912 }, { "epoch": 0.5607265872072025, "grad_norm": 1.291407823562622, "learning_rate": 8.551695627489393e-06, "loss": 0.9916, "step": 15913 }, { "epoch": 0.5607618242201609, "grad_norm": 5.241656303405762, "learning_rate": 8.550566393185804e-06, "loss": 5.0604, "step": 15914 }, { "epoch": 0.5607970612331192, "grad_norm": 10.873777389526367, "learning_rate": 8.549437177761268e-06, "loss": 4.6188, "step": 15915 }, { "epoch": 0.5608322982460777, "grad_norm": 1.6153849363327026, "learning_rate": 8.548307981230496e-06, "loss": 1.0343, "step": 15916 }, { "epoch": 0.5608675352590361, "grad_norm": 1.4419759511947632, "learning_rate": 8.547178803608199e-06, "loss": 0.9595, "step": 15917 }, { "epoch": 0.5609027722719945, "grad_norm": 6.698879718780518, "learning_rate": 8.546049644909087e-06, "loss": 3.104, "step": 15918 }, { "epoch": 0.560938009284953, "grad_norm": 5.685988903045654, "learning_rate": 8.544920505147861e-06, "loss": 4.792, "step": 15919 }, { "epoch": 0.5609732462979113, "grad_norm": 2.0294971466064453, "learning_rate": 8.543791384339228e-06, "loss": 1.0327, "step": 15920 }, { "epoch": 0.5610084833108697, "grad_norm": 5.963497161865234, "learning_rate": 8.542662282497902e-06, "loss": 5.4408, "step": 15921 }, { "epoch": 0.5610437203238281, "grad_norm": 1.254128098487854, "learning_rate": 8.541533199638583e-06, "loss": 0.9595, "step": 15922 }, { "epoch": 0.5610789573367866, "grad_norm": 5.1876420974731445, "learning_rate": 8.540404135775981e-06, "loss": 2.8471, "step": 15923 }, { "epoch": 0.561114194349745, "grad_norm": 1.1108949184417725, "learning_rate": 8.5392750909248e-06, "loss": 0.8913, "step": 15924 }, { "epoch": 0.5611494313627033, "grad_norm": 1.3500878810882568, "learning_rate": 8.538146065099747e-06, "loss": 0.7993, "step": 15925 }, { "epoch": 0.5611846683756618, "grad_norm": 4.162587642669678, "learning_rate": 8.537017058315528e-06, "loss": 3.1504, "step": 15926 }, { "epoch": 0.5612199053886202, "grad_norm": 8.692127227783203, "learning_rate": 8.535888070586852e-06, "loss": 5.7142, "step": 15927 }, { "epoch": 0.5612551424015786, "grad_norm": 2.96049165725708, "learning_rate": 8.534759101928416e-06, "loss": 0.967, "step": 15928 }, { "epoch": 0.5612903794145371, "grad_norm": 1.258259892463684, "learning_rate": 8.533630152354926e-06, "loss": 0.8658, "step": 15929 }, { "epoch": 0.5613256164274955, "grad_norm": 5.426548957824707, "learning_rate": 8.532501221881098e-06, "loss": 3.0535, "step": 15930 }, { "epoch": 0.5613608534404538, "grad_norm": 6.174755573272705, "learning_rate": 8.531372310521622e-06, "loss": 5.1486, "step": 15931 }, { "epoch": 0.5613960904534122, "grad_norm": 1.919302225112915, "learning_rate": 8.530243418291213e-06, "loss": 0.9589, "step": 15932 }, { "epoch": 0.5614313274663707, "grad_norm": 2.9729678630828857, "learning_rate": 8.529114545204567e-06, "loss": 0.8162, "step": 15933 }, { "epoch": 0.5614665644793291, "grad_norm": 1.6880096197128296, "learning_rate": 8.527985691276396e-06, "loss": 0.9322, "step": 15934 }, { "epoch": 0.5615018014922875, "grad_norm": 1.1844676733016968, "learning_rate": 8.526856856521394e-06, "loss": 0.6566, "step": 15935 }, { "epoch": 0.561537038505246, "grad_norm": 1.3485029935836792, "learning_rate": 8.525728040954272e-06, "loss": 0.782, "step": 15936 }, { "epoch": 0.5615722755182043, "grad_norm": 1.2304795980453491, "learning_rate": 8.524599244589732e-06, "loss": 0.9756, "step": 15937 }, { "epoch": 0.5616075125311627, "grad_norm": 2.103013515472412, "learning_rate": 8.523470467442473e-06, "loss": 2.1777, "step": 15938 }, { "epoch": 0.5616427495441212, "grad_norm": 3.3477776050567627, "learning_rate": 8.522341709527196e-06, "loss": 3.4648, "step": 15939 }, { "epoch": 0.5616779865570796, "grad_norm": 123.99456024169922, "learning_rate": 8.52121297085861e-06, "loss": 3.2774, "step": 15940 }, { "epoch": 0.561713223570038, "grad_norm": 1.2017102241516113, "learning_rate": 8.520084251451413e-06, "loss": 0.8887, "step": 15941 }, { "epoch": 0.5617484605829964, "grad_norm": 4.684961795806885, "learning_rate": 8.518955551320306e-06, "loss": 2.655, "step": 15942 }, { "epoch": 0.5617836975959548, "grad_norm": 4.0337066650390625, "learning_rate": 8.517826870479994e-06, "loss": 3.0152, "step": 15943 }, { "epoch": 0.5618189346089132, "grad_norm": 1.894105076789856, "learning_rate": 8.516698208945171e-06, "loss": 0.8613, "step": 15944 }, { "epoch": 0.5618541716218716, "grad_norm": 3.516225576400757, "learning_rate": 8.515569566730547e-06, "loss": 3.2017, "step": 15945 }, { "epoch": 0.5618894086348301, "grad_norm": 1.655656099319458, "learning_rate": 8.514440943850818e-06, "loss": 0.937, "step": 15946 }, { "epoch": 0.5619246456477884, "grad_norm": 6.351780891418457, "learning_rate": 8.513312340320684e-06, "loss": 3.4565, "step": 15947 }, { "epoch": 0.5619598826607468, "grad_norm": 9.45040225982666, "learning_rate": 8.512183756154843e-06, "loss": 5.0647, "step": 15948 }, { "epoch": 0.5619951196737053, "grad_norm": 4.93228006362915, "learning_rate": 8.511055191368006e-06, "loss": 3.0161, "step": 15949 }, { "epoch": 0.5620303566866637, "grad_norm": 7.1566667556762695, "learning_rate": 8.509926645974858e-06, "loss": 3.5147, "step": 15950 }, { "epoch": 0.5620655936996221, "grad_norm": 4.462547302246094, "learning_rate": 8.508798119990108e-06, "loss": 2.9226, "step": 15951 }, { "epoch": 0.5621008307125805, "grad_norm": 2.8147475719451904, "learning_rate": 8.507669613428452e-06, "loss": 3.0091, "step": 15952 }, { "epoch": 0.5621360677255389, "grad_norm": 3.587151050567627, "learning_rate": 8.50654112630459e-06, "loss": 2.8946, "step": 15953 }, { "epoch": 0.5621713047384973, "grad_norm": 3.9074156284332275, "learning_rate": 8.505412658633217e-06, "loss": 2.999, "step": 15954 }, { "epoch": 0.5622065417514557, "grad_norm": 4.672226905822754, "learning_rate": 8.504284210429038e-06, "loss": 3.2707, "step": 15955 }, { "epoch": 0.5622417787644142, "grad_norm": 7.675459384918213, "learning_rate": 8.503155781706748e-06, "loss": 5.1484, "step": 15956 }, { "epoch": 0.5622770157773725, "grad_norm": 1.2721537351608276, "learning_rate": 8.502027372481043e-06, "loss": 0.8043, "step": 15957 }, { "epoch": 0.5623122527903309, "grad_norm": 2.4300577640533447, "learning_rate": 8.500898982766619e-06, "loss": 0.8336, "step": 15958 }, { "epoch": 0.5623474898032894, "grad_norm": 1.7933223247528076, "learning_rate": 8.499770612578184e-06, "loss": 1.1322, "step": 15959 }, { "epoch": 0.5623827268162478, "grad_norm": 3.99749755859375, "learning_rate": 8.498642261930423e-06, "loss": 2.9083, "step": 15960 }, { "epoch": 0.5624179638292062, "grad_norm": 1.5871365070343018, "learning_rate": 8.497513930838038e-06, "loss": 0.8044, "step": 15961 }, { "epoch": 0.5624532008421647, "grad_norm": 5.357926368713379, "learning_rate": 8.496385619315728e-06, "loss": 3.2655, "step": 15962 }, { "epoch": 0.562488437855123, "grad_norm": 5.497339725494385, "learning_rate": 8.495257327378185e-06, "loss": 2.9941, "step": 15963 }, { "epoch": 0.5625236748680814, "grad_norm": 5.014674186706543, "learning_rate": 8.494129055040106e-06, "loss": 0.972, "step": 15964 }, { "epoch": 0.5625589118810398, "grad_norm": 1.509167194366455, "learning_rate": 8.49300080231619e-06, "loss": 0.9996, "step": 15965 }, { "epoch": 0.5625941488939983, "grad_norm": 4.47278356552124, "learning_rate": 8.491872569221129e-06, "loss": 3.425, "step": 15966 }, { "epoch": 0.5626293859069567, "grad_norm": 1.0189242362976074, "learning_rate": 8.49074435576962e-06, "loss": 0.9304, "step": 15967 }, { "epoch": 0.562664622919915, "grad_norm": 2.2540647983551025, "learning_rate": 8.48961616197636e-06, "loss": 0.9218, "step": 15968 }, { "epoch": 0.5626998599328735, "grad_norm": 14.863430976867676, "learning_rate": 8.488487987856038e-06, "loss": 5.2294, "step": 15969 }, { "epoch": 0.5627350969458319, "grad_norm": 1.1818134784698486, "learning_rate": 8.487359833423354e-06, "loss": 0.8803, "step": 15970 }, { "epoch": 0.5627703339587903, "grad_norm": 3.6415209770202637, "learning_rate": 8.486231698693002e-06, "loss": 0.8373, "step": 15971 }, { "epoch": 0.5628055709717488, "grad_norm": 4.028013229370117, "learning_rate": 8.485103583679676e-06, "loss": 2.8433, "step": 15972 }, { "epoch": 0.5628408079847071, "grad_norm": 9.249838829040527, "learning_rate": 8.483975488398066e-06, "loss": 3.4423, "step": 15973 }, { "epoch": 0.5628760449976655, "grad_norm": 1.3571490049362183, "learning_rate": 8.482847412862869e-06, "loss": 0.7609, "step": 15974 }, { "epoch": 0.562911282010624, "grad_norm": 4.108963966369629, "learning_rate": 8.48171935708878e-06, "loss": 2.9647, "step": 15975 }, { "epoch": 0.5629465190235824, "grad_norm": 5.519519805908203, "learning_rate": 8.480591321090488e-06, "loss": 2.6586, "step": 15976 }, { "epoch": 0.5629817560365408, "grad_norm": 9.216403007507324, "learning_rate": 8.479463304882685e-06, "loss": 3.6255, "step": 15977 }, { "epoch": 0.5630169930494991, "grad_norm": 1.2319587469100952, "learning_rate": 8.478335308480074e-06, "loss": 0.9001, "step": 15978 }, { "epoch": 0.5630522300624576, "grad_norm": 2.0315299034118652, "learning_rate": 8.477207331897331e-06, "loss": 1.2103, "step": 15979 }, { "epoch": 0.563087467075416, "grad_norm": 1.1953701972961426, "learning_rate": 8.47607937514916e-06, "loss": 0.9716, "step": 15980 }, { "epoch": 0.5631227040883744, "grad_norm": 1.23697030544281, "learning_rate": 8.474951438250251e-06, "loss": 0.9083, "step": 15981 }, { "epoch": 0.5631579411013329, "grad_norm": 4.523537635803223, "learning_rate": 8.473823521215291e-06, "loss": 2.8924, "step": 15982 }, { "epoch": 0.5631931781142913, "grad_norm": 7.233031749725342, "learning_rate": 8.472695624058973e-06, "loss": 6.1098, "step": 15983 }, { "epoch": 0.5632284151272496, "grad_norm": 13.046398162841797, "learning_rate": 8.471567746795993e-06, "loss": 2.6896, "step": 15984 }, { "epoch": 0.5632636521402081, "grad_norm": 1.290737509727478, "learning_rate": 8.470439889441035e-06, "loss": 0.9939, "step": 15985 }, { "epoch": 0.5632988891531665, "grad_norm": 3.6827127933502197, "learning_rate": 8.469312052008794e-06, "loss": 2.8455, "step": 15986 }, { "epoch": 0.5633341261661249, "grad_norm": 5.6211323738098145, "learning_rate": 8.468184234513959e-06, "loss": 3.4801, "step": 15987 }, { "epoch": 0.5633693631790833, "grad_norm": 1.4230624437332153, "learning_rate": 8.467056436971214e-06, "loss": 0.9032, "step": 15988 }, { "epoch": 0.5634046001920417, "grad_norm": 3.540266513824463, "learning_rate": 8.465928659395259e-06, "loss": 3.22, "step": 15989 }, { "epoch": 0.5634398372050001, "grad_norm": 1.3885310888290405, "learning_rate": 8.464800901800779e-06, "loss": 0.7589, "step": 15990 }, { "epoch": 0.5634750742179585, "grad_norm": 1.4790315628051758, "learning_rate": 8.463673164202462e-06, "loss": 1.3511, "step": 15991 }, { "epoch": 0.563510311230917, "grad_norm": 19.215620040893555, "learning_rate": 8.462545446614995e-06, "loss": 2.9558, "step": 15992 }, { "epoch": 0.5635455482438754, "grad_norm": 9.372187614440918, "learning_rate": 8.461417749053072e-06, "loss": 3.0844, "step": 15993 }, { "epoch": 0.5635807852568337, "grad_norm": 6.63362455368042, "learning_rate": 8.46029007153138e-06, "loss": 3.5331, "step": 15994 }, { "epoch": 0.5636160222697922, "grad_norm": 10.900934219360352, "learning_rate": 8.459162414064604e-06, "loss": 5.1658, "step": 15995 }, { "epoch": 0.5636512592827506, "grad_norm": 1.159490942955017, "learning_rate": 8.458034776667432e-06, "loss": 0.7725, "step": 15996 }, { "epoch": 0.563686496295709, "grad_norm": 12.170955657958984, "learning_rate": 8.45690715935456e-06, "loss": 3.3432, "step": 15997 }, { "epoch": 0.5637217333086674, "grad_norm": 4.600131511688232, "learning_rate": 8.455779562140663e-06, "loss": 3.1872, "step": 15998 }, { "epoch": 0.5637569703216259, "grad_norm": 3.85996413230896, "learning_rate": 8.454651985040433e-06, "loss": 2.5924, "step": 15999 }, { "epoch": 0.5637922073345842, "grad_norm": 1.230232834815979, "learning_rate": 8.453524428068563e-06, "loss": 0.8964, "step": 16000 }, { "epoch": 0.5638274443475426, "grad_norm": 4.083792686462402, "learning_rate": 8.452396891239732e-06, "loss": 2.5536, "step": 16001 }, { "epoch": 0.5638626813605011, "grad_norm": 1.284751057624817, "learning_rate": 8.451269374568625e-06, "loss": 0.9638, "step": 16002 }, { "epoch": 0.5638979183734595, "grad_norm": 1.5747878551483154, "learning_rate": 8.450141878069938e-06, "loss": 0.8491, "step": 16003 }, { "epoch": 0.5639331553864179, "grad_norm": 1.0866063833236694, "learning_rate": 8.449014401758347e-06, "loss": 0.6555, "step": 16004 }, { "epoch": 0.5639683923993764, "grad_norm": 16.58873176574707, "learning_rate": 8.44788694564854e-06, "loss": 6.0962, "step": 16005 }, { "epoch": 0.5640036294123347, "grad_norm": 1.053554892539978, "learning_rate": 8.446759509755202e-06, "loss": 0.9855, "step": 16006 }, { "epoch": 0.5640388664252931, "grad_norm": 1.2541848421096802, "learning_rate": 8.445632094093024e-06, "loss": 1.017, "step": 16007 }, { "epoch": 0.5640741034382515, "grad_norm": 1.585218906402588, "learning_rate": 8.444504698676683e-06, "loss": 0.931, "step": 16008 }, { "epoch": 0.56410934045121, "grad_norm": 3.2273435592651367, "learning_rate": 8.443377323520866e-06, "loss": 2.423, "step": 16009 }, { "epoch": 0.5641445774641684, "grad_norm": 6.825433254241943, "learning_rate": 8.442249968640262e-06, "loss": 5.5911, "step": 16010 }, { "epoch": 0.5641798144771267, "grad_norm": 7.658723831176758, "learning_rate": 8.441122634049545e-06, "loss": 5.9868, "step": 16011 }, { "epoch": 0.5642150514900852, "grad_norm": 4.234168529510498, "learning_rate": 8.439995319763405e-06, "loss": 2.8168, "step": 16012 }, { "epoch": 0.5642502885030436, "grad_norm": 1.4723095893859863, "learning_rate": 8.43886802579653e-06, "loss": 0.6735, "step": 16013 }, { "epoch": 0.564285525516002, "grad_norm": 6.914557456970215, "learning_rate": 8.437740752163592e-06, "loss": 2.8566, "step": 16014 }, { "epoch": 0.5643207625289605, "grad_norm": 6.672532558441162, "learning_rate": 8.43661349887928e-06, "loss": 3.3066, "step": 16015 }, { "epoch": 0.5643559995419188, "grad_norm": 4.376091957092285, "learning_rate": 8.435486265958282e-06, "loss": 3.2812, "step": 16016 }, { "epoch": 0.5643912365548772, "grad_norm": 1.5595706701278687, "learning_rate": 8.43435905341527e-06, "loss": 0.9733, "step": 16017 }, { "epoch": 0.5644264735678357, "grad_norm": 1.0237237215042114, "learning_rate": 8.43323186126493e-06, "loss": 0.9669, "step": 16018 }, { "epoch": 0.5644617105807941, "grad_norm": 4.373929500579834, "learning_rate": 8.432104689521949e-06, "loss": 3.3918, "step": 16019 }, { "epoch": 0.5644969475937525, "grad_norm": 2.500741481781006, "learning_rate": 8.430977538201e-06, "loss": 2.9101, "step": 16020 }, { "epoch": 0.5645321846067108, "grad_norm": 1.022104024887085, "learning_rate": 8.429850407316767e-06, "loss": 1.3514, "step": 16021 }, { "epoch": 0.5645674216196693, "grad_norm": 3.468853235244751, "learning_rate": 8.428723296883939e-06, "loss": 3.2169, "step": 16022 }, { "epoch": 0.5646026586326277, "grad_norm": 4.542963981628418, "learning_rate": 8.427596206917184e-06, "loss": 2.66, "step": 16023 }, { "epoch": 0.5646378956455861, "grad_norm": 4.569720268249512, "learning_rate": 8.426469137431192e-06, "loss": 2.8934, "step": 16024 }, { "epoch": 0.5646731326585446, "grad_norm": 6.5088887214660645, "learning_rate": 8.425342088440635e-06, "loss": 3.1705, "step": 16025 }, { "epoch": 0.564708369671503, "grad_norm": 6.863347053527832, "learning_rate": 8.424215059960206e-06, "loss": 4.5568, "step": 16026 }, { "epoch": 0.5647436066844613, "grad_norm": 3.4052343368530273, "learning_rate": 8.423088052004572e-06, "loss": 3.2407, "step": 16027 }, { "epoch": 0.5647788436974198, "grad_norm": 3.4606146812438965, "learning_rate": 8.421961064588417e-06, "loss": 3.0651, "step": 16028 }, { "epoch": 0.5648140807103782, "grad_norm": 1.271087646484375, "learning_rate": 8.420834097726422e-06, "loss": 0.7384, "step": 16029 }, { "epoch": 0.5648493177233366, "grad_norm": 10.172642707824707, "learning_rate": 8.419707151433261e-06, "loss": 3.5161, "step": 16030 }, { "epoch": 0.564884554736295, "grad_norm": 4.373618125915527, "learning_rate": 8.418580225723616e-06, "loss": 3.2635, "step": 16031 }, { "epoch": 0.5649197917492534, "grad_norm": 1.135838508605957, "learning_rate": 8.417453320612167e-06, "loss": 1.0519, "step": 16032 }, { "epoch": 0.5649550287622118, "grad_norm": 3.8459932804107666, "learning_rate": 8.41632643611359e-06, "loss": 2.5154, "step": 16033 }, { "epoch": 0.5649902657751702, "grad_norm": 6.563310623168945, "learning_rate": 8.41519957224256e-06, "loss": 2.8925, "step": 16034 }, { "epoch": 0.5650255027881287, "grad_norm": 1.0984771251678467, "learning_rate": 8.414072729013764e-06, "loss": 1.0237, "step": 16035 }, { "epoch": 0.5650607398010871, "grad_norm": 1.087355375289917, "learning_rate": 8.412945906441866e-06, "loss": 0.8932, "step": 16036 }, { "epoch": 0.5650959768140454, "grad_norm": 3.1861765384674072, "learning_rate": 8.411819104541552e-06, "loss": 3.1405, "step": 16037 }, { "epoch": 0.5651312138270039, "grad_norm": 1.0572665929794312, "learning_rate": 8.410692323327498e-06, "loss": 0.957, "step": 16038 }, { "epoch": 0.5651664508399623, "grad_norm": 4.533376693725586, "learning_rate": 8.409565562814377e-06, "loss": 3.5146, "step": 16039 }, { "epoch": 0.5652016878529207, "grad_norm": 4.515148639678955, "learning_rate": 8.408438823016866e-06, "loss": 3.0334, "step": 16040 }, { "epoch": 0.5652369248658791, "grad_norm": 4.297812461853027, "learning_rate": 8.407312103949645e-06, "loss": 3.3494, "step": 16041 }, { "epoch": 0.5652721618788376, "grad_norm": 3.2912681102752686, "learning_rate": 8.406185405627385e-06, "loss": 3.0275, "step": 16042 }, { "epoch": 0.5653073988917959, "grad_norm": 2.082106828689575, "learning_rate": 8.405058728064763e-06, "loss": 0.8754, "step": 16043 }, { "epoch": 0.5653426359047543, "grad_norm": 11.537254333496094, "learning_rate": 8.403932071276452e-06, "loss": 4.8266, "step": 16044 }, { "epoch": 0.5653778729177128, "grad_norm": 8.457352638244629, "learning_rate": 8.402805435277135e-06, "loss": 8.3071, "step": 16045 }, { "epoch": 0.5654131099306712, "grad_norm": 2.0085322856903076, "learning_rate": 8.401678820081476e-06, "loss": 0.8291, "step": 16046 }, { "epoch": 0.5654483469436296, "grad_norm": 0.975009560585022, "learning_rate": 8.400552225704156e-06, "loss": 0.8175, "step": 16047 }, { "epoch": 0.565483583956588, "grad_norm": 1.025967001914978, "learning_rate": 8.399425652159848e-06, "loss": 1.1823, "step": 16048 }, { "epoch": 0.5655188209695464, "grad_norm": 1.1312181949615479, "learning_rate": 8.398299099463221e-06, "loss": 0.7839, "step": 16049 }, { "epoch": 0.5655540579825048, "grad_norm": 4.728481292724609, "learning_rate": 8.397172567628955e-06, "loss": 3.0869, "step": 16050 }, { "epoch": 0.5655892949954633, "grad_norm": 0.9638043642044067, "learning_rate": 8.396046056671722e-06, "loss": 0.9017, "step": 16051 }, { "epoch": 0.5656245320084217, "grad_norm": 5.390205383300781, "learning_rate": 8.39491956660619e-06, "loss": 5.0997, "step": 16052 }, { "epoch": 0.56565976902138, "grad_norm": 6.365561008453369, "learning_rate": 8.393793097447034e-06, "loss": 3.8311, "step": 16053 }, { "epoch": 0.5656950060343384, "grad_norm": 4.755992412567139, "learning_rate": 8.392666649208935e-06, "loss": 2.5333, "step": 16054 }, { "epoch": 0.5657302430472969, "grad_norm": 5.466028690338135, "learning_rate": 8.39154022190655e-06, "loss": 3.3493, "step": 16055 }, { "epoch": 0.5657654800602553, "grad_norm": 9.267115592956543, "learning_rate": 8.390413815554561e-06, "loss": 5.2303, "step": 16056 }, { "epoch": 0.5658007170732137, "grad_norm": 7.261538505554199, "learning_rate": 8.389287430167638e-06, "loss": 3.1231, "step": 16057 }, { "epoch": 0.5658359540861722, "grad_norm": 5.108530044555664, "learning_rate": 8.388161065760452e-06, "loss": 3.3514, "step": 16058 }, { "epoch": 0.5658711910991305, "grad_norm": 4.655511379241943, "learning_rate": 8.387034722347669e-06, "loss": 3.5538, "step": 16059 }, { "epoch": 0.5659064281120889, "grad_norm": 2.7812819480895996, "learning_rate": 8.38590839994397e-06, "loss": 0.9931, "step": 16060 }, { "epoch": 0.5659416651250474, "grad_norm": 4.978321552276611, "learning_rate": 8.384782098564015e-06, "loss": 2.8174, "step": 16061 }, { "epoch": 0.5659769021380058, "grad_norm": 4.250336647033691, "learning_rate": 8.383655818222481e-06, "loss": 3.526, "step": 16062 }, { "epoch": 0.5660121391509642, "grad_norm": 4.99479866027832, "learning_rate": 8.382529558934034e-06, "loss": 4.1087, "step": 16063 }, { "epoch": 0.5660473761639225, "grad_norm": 3.219101667404175, "learning_rate": 8.381403320713348e-06, "loss": 2.5509, "step": 16064 }, { "epoch": 0.566082613176881, "grad_norm": 4.837749481201172, "learning_rate": 8.38027710357509e-06, "loss": 3.1255, "step": 16065 }, { "epoch": 0.5661178501898394, "grad_norm": 1.1583484411239624, "learning_rate": 8.379150907533926e-06, "loss": 1.1126, "step": 16066 }, { "epoch": 0.5661530872027978, "grad_norm": 1.3464667797088623, "learning_rate": 8.378024732604535e-06, "loss": 0.7507, "step": 16067 }, { "epoch": 0.5661883242157563, "grad_norm": 4.197670936584473, "learning_rate": 8.37689857880157e-06, "loss": 3.7886, "step": 16068 }, { "epoch": 0.5662235612287146, "grad_norm": 3.4962079524993896, "learning_rate": 8.375772446139711e-06, "loss": 3.085, "step": 16069 }, { "epoch": 0.566258798241673, "grad_norm": 4.692408084869385, "learning_rate": 8.374646334633626e-06, "loss": 2.7858, "step": 16070 }, { "epoch": 0.5662940352546315, "grad_norm": 1.9399081468582153, "learning_rate": 8.373520244297975e-06, "loss": 0.9236, "step": 16071 }, { "epoch": 0.5663292722675899, "grad_norm": 9.909590721130371, "learning_rate": 8.37239417514743e-06, "loss": 2.9034, "step": 16072 }, { "epoch": 0.5663645092805483, "grad_norm": 45.40212631225586, "learning_rate": 8.371268127196664e-06, "loss": 2.5774, "step": 16073 }, { "epoch": 0.5663997462935066, "grad_norm": 5.0838212966918945, "learning_rate": 8.37014210046033e-06, "loss": 4.8661, "step": 16074 }, { "epoch": 0.5664349833064651, "grad_norm": 5.127656936645508, "learning_rate": 8.369016094953106e-06, "loss": 4.5324, "step": 16075 }, { "epoch": 0.5664702203194235, "grad_norm": 7.821997165679932, "learning_rate": 8.367890110689658e-06, "loss": 3.2531, "step": 16076 }, { "epoch": 0.5665054573323819, "grad_norm": 21.96147346496582, "learning_rate": 8.366764147684647e-06, "loss": 5.809, "step": 16077 }, { "epoch": 0.5665406943453404, "grad_norm": 2.191667079925537, "learning_rate": 8.36563820595274e-06, "loss": 0.78, "step": 16078 }, { "epoch": 0.5665759313582988, "grad_norm": 3.377136707305908, "learning_rate": 8.364512285508603e-06, "loss": 3.1378, "step": 16079 }, { "epoch": 0.5666111683712571, "grad_norm": 0.8664025068283081, "learning_rate": 8.363386386366905e-06, "loss": 0.8238, "step": 16080 }, { "epoch": 0.5666464053842156, "grad_norm": 3.3734216690063477, "learning_rate": 8.362260508542305e-06, "loss": 3.0204, "step": 16081 }, { "epoch": 0.566681642397174, "grad_norm": 3.9406418800354004, "learning_rate": 8.361134652049471e-06, "loss": 2.7564, "step": 16082 }, { "epoch": 0.5667168794101324, "grad_norm": 1.3981975317001343, "learning_rate": 8.360008816903069e-06, "loss": 0.8273, "step": 16083 }, { "epoch": 0.5667521164230909, "grad_norm": 2.4784324169158936, "learning_rate": 8.35888300311776e-06, "loss": 2.6424, "step": 16084 }, { "epoch": 0.5667873534360492, "grad_norm": 1.407859444618225, "learning_rate": 8.357757210708208e-06, "loss": 0.825, "step": 16085 }, { "epoch": 0.5668225904490076, "grad_norm": 1.254880428314209, "learning_rate": 8.35663143968908e-06, "loss": 1.0142, "step": 16086 }, { "epoch": 0.566857827461966, "grad_norm": 1.61818265914917, "learning_rate": 8.355505690075032e-06, "loss": 0.8389, "step": 16087 }, { "epoch": 0.5668930644749245, "grad_norm": 1.1105271577835083, "learning_rate": 8.354379961880733e-06, "loss": 1.2172, "step": 16088 }, { "epoch": 0.5669283014878829, "grad_norm": 0.9499982595443726, "learning_rate": 8.353254255120851e-06, "loss": 0.8911, "step": 16089 }, { "epoch": 0.5669635385008412, "grad_norm": 1.39656662940979, "learning_rate": 8.352128569810037e-06, "loss": 0.9036, "step": 16090 }, { "epoch": 0.5669987755137997, "grad_norm": 2.0812740325927734, "learning_rate": 8.351002905962956e-06, "loss": 0.9679, "step": 16091 }, { "epoch": 0.5670340125267581, "grad_norm": 1.1542034149169922, "learning_rate": 8.34987726359428e-06, "loss": 0.8077, "step": 16092 }, { "epoch": 0.5670692495397165, "grad_norm": 2.922666311264038, "learning_rate": 8.348751642718654e-06, "loss": 3.1481, "step": 16093 }, { "epoch": 0.567104486552675, "grad_norm": 3.10760498046875, "learning_rate": 8.347626043350752e-06, "loss": 2.6776, "step": 16094 }, { "epoch": 0.5671397235656334, "grad_norm": 5.814879894256592, "learning_rate": 8.346500465505232e-06, "loss": 5.2297, "step": 16095 }, { "epoch": 0.5671749605785917, "grad_norm": 1.1698427200317383, "learning_rate": 8.345374909196753e-06, "loss": 1.1014, "step": 16096 }, { "epoch": 0.5672101975915501, "grad_norm": 3.370244026184082, "learning_rate": 8.344249374439975e-06, "loss": 3.2999, "step": 16097 }, { "epoch": 0.5672454346045086, "grad_norm": 1.4708445072174072, "learning_rate": 8.343123861249563e-06, "loss": 0.7898, "step": 16098 }, { "epoch": 0.567280671617467, "grad_norm": 1.8650519847869873, "learning_rate": 8.341998369640173e-06, "loss": 0.8945, "step": 16099 }, { "epoch": 0.5673159086304254, "grad_norm": 4.748635292053223, "learning_rate": 8.340872899626465e-06, "loss": 3.2637, "step": 16100 }, { "epoch": 0.5673511456433838, "grad_norm": 14.981785774230957, "learning_rate": 8.339747451223095e-06, "loss": 6.0573, "step": 16101 }, { "epoch": 0.5673863826563422, "grad_norm": 3.2128448486328125, "learning_rate": 8.338622024444732e-06, "loss": 3.088, "step": 16102 }, { "epoch": 0.5674216196693006, "grad_norm": 5.761301517486572, "learning_rate": 8.337496619306027e-06, "loss": 5.2775, "step": 16103 }, { "epoch": 0.5674568566822591, "grad_norm": 1.3648149967193604, "learning_rate": 8.33637123582164e-06, "loss": 1.0822, "step": 16104 }, { "epoch": 0.5674920936952175, "grad_norm": 10.750929832458496, "learning_rate": 8.33524587400623e-06, "loss": 7.9567, "step": 16105 }, { "epoch": 0.5675273307081758, "grad_norm": 4.092633247375488, "learning_rate": 8.334120533874452e-06, "loss": 3.2438, "step": 16106 }, { "epoch": 0.5675625677211342, "grad_norm": 1.9929208755493164, "learning_rate": 8.332995215440968e-06, "loss": 0.8924, "step": 16107 }, { "epoch": 0.5675978047340927, "grad_norm": 9.460271835327148, "learning_rate": 8.331869918720435e-06, "loss": 5.9604, "step": 16108 }, { "epoch": 0.5676330417470511, "grad_norm": 3.263909339904785, "learning_rate": 8.330744643727508e-06, "loss": 0.7997, "step": 16109 }, { "epoch": 0.5676682787600095, "grad_norm": 9.59859561920166, "learning_rate": 8.329619390476843e-06, "loss": 4.9611, "step": 16110 }, { "epoch": 0.567703515772968, "grad_norm": 1.6972936391830444, "learning_rate": 8.328494158983102e-06, "loss": 0.7335, "step": 16111 }, { "epoch": 0.5677387527859263, "grad_norm": 5.172227382659912, "learning_rate": 8.327368949260932e-06, "loss": 2.7925, "step": 16112 }, { "epoch": 0.5677739897988847, "grad_norm": 4.975461006164551, "learning_rate": 8.326243761324996e-06, "loss": 2.6511, "step": 16113 }, { "epoch": 0.5678092268118432, "grad_norm": 3.0354342460632324, "learning_rate": 8.325118595189952e-06, "loss": 2.5841, "step": 16114 }, { "epoch": 0.5678444638248016, "grad_norm": 1.3719758987426758, "learning_rate": 8.323993450870448e-06, "loss": 0.8431, "step": 16115 }, { "epoch": 0.56787970083776, "grad_norm": 4.925502300262451, "learning_rate": 8.322868328381144e-06, "loss": 2.9986, "step": 16116 }, { "epoch": 0.5679149378507184, "grad_norm": 11.098908424377441, "learning_rate": 8.321743227736693e-06, "loss": 3.1601, "step": 16117 }, { "epoch": 0.5679501748636768, "grad_norm": 1.4365726709365845, "learning_rate": 8.320618148951752e-06, "loss": 1.0392, "step": 16118 }, { "epoch": 0.5679854118766352, "grad_norm": 1.1876068115234375, "learning_rate": 8.319493092040971e-06, "loss": 1.0643, "step": 16119 }, { "epoch": 0.5680206488895936, "grad_norm": 4.470085620880127, "learning_rate": 8.318368057019006e-06, "loss": 2.6728, "step": 16120 }, { "epoch": 0.5680558859025521, "grad_norm": 1.4745895862579346, "learning_rate": 8.317243043900518e-06, "loss": 0.8258, "step": 16121 }, { "epoch": 0.5680911229155104, "grad_norm": 1.4255732297897339, "learning_rate": 8.316118052700146e-06, "loss": 1.0044, "step": 16122 }, { "epoch": 0.5681263599284688, "grad_norm": 11.60169506072998, "learning_rate": 8.314993083432554e-06, "loss": 3.4672, "step": 16123 }, { "epoch": 0.5681615969414273, "grad_norm": 1.6419683694839478, "learning_rate": 8.313868136112393e-06, "loss": 0.8576, "step": 16124 }, { "epoch": 0.5681968339543857, "grad_norm": 4.689891338348389, "learning_rate": 8.312743210754313e-06, "loss": 2.919, "step": 16125 }, { "epoch": 0.5682320709673441, "grad_norm": 2.722595453262329, "learning_rate": 8.311618307372965e-06, "loss": 3.1754, "step": 16126 }, { "epoch": 0.5682673079803026, "grad_norm": 1.8554158210754395, "learning_rate": 8.310493425983008e-06, "loss": 1.0691, "step": 16127 }, { "epoch": 0.5683025449932609, "grad_norm": 1.580675482749939, "learning_rate": 8.309368566599088e-06, "loss": 0.9096, "step": 16128 }, { "epoch": 0.5683377820062193, "grad_norm": 1.3253518342971802, "learning_rate": 8.308243729235857e-06, "loss": 0.7716, "step": 16129 }, { "epoch": 0.5683730190191777, "grad_norm": 4.491563320159912, "learning_rate": 8.307118913907968e-06, "loss": 2.4962, "step": 16130 }, { "epoch": 0.5684082560321362, "grad_norm": 2.98433256149292, "learning_rate": 8.305994120630068e-06, "loss": 2.6726, "step": 16131 }, { "epoch": 0.5684434930450946, "grad_norm": 4.250982761383057, "learning_rate": 8.304869349416811e-06, "loss": 3.5982, "step": 16132 }, { "epoch": 0.5684787300580529, "grad_norm": 7.76693868637085, "learning_rate": 8.30374460028285e-06, "loss": 3.2065, "step": 16133 }, { "epoch": 0.5685139670710114, "grad_norm": 6.9945220947265625, "learning_rate": 8.30261987324283e-06, "loss": 3.669, "step": 16134 }, { "epoch": 0.5685492040839698, "grad_norm": 3.4199821949005127, "learning_rate": 8.3014951683114e-06, "loss": 3.2919, "step": 16135 }, { "epoch": 0.5685844410969282, "grad_norm": 5.041635990142822, "learning_rate": 8.300370485503213e-06, "loss": 3.1943, "step": 16136 }, { "epoch": 0.5686196781098867, "grad_norm": 4.3948750495910645, "learning_rate": 8.299245824832918e-06, "loss": 3.2761, "step": 16137 }, { "epoch": 0.568654915122845, "grad_norm": 1.548291802406311, "learning_rate": 8.298121186315163e-06, "loss": 0.8838, "step": 16138 }, { "epoch": 0.5686901521358034, "grad_norm": 7.0768351554870605, "learning_rate": 8.296996569964592e-06, "loss": 2.6552, "step": 16139 }, { "epoch": 0.5687253891487618, "grad_norm": 3.8166704177856445, "learning_rate": 8.295871975795866e-06, "loss": 3.0828, "step": 16140 }, { "epoch": 0.5687606261617203, "grad_norm": 9.432373046875, "learning_rate": 8.294747403823617e-06, "loss": 5.0572, "step": 16141 }, { "epoch": 0.5687958631746787, "grad_norm": 6.440566539764404, "learning_rate": 8.293622854062504e-06, "loss": 4.5884, "step": 16142 }, { "epoch": 0.568831100187637, "grad_norm": 2.7974698543548584, "learning_rate": 8.292498326527171e-06, "loss": 2.6279, "step": 16143 }, { "epoch": 0.5688663372005955, "grad_norm": 2.585217237472534, "learning_rate": 8.291373821232264e-06, "loss": 0.8823, "step": 16144 }, { "epoch": 0.5689015742135539, "grad_norm": 5.383573532104492, "learning_rate": 8.290249338192427e-06, "loss": 5.57, "step": 16145 }, { "epoch": 0.5689368112265123, "grad_norm": 1.1577504873275757, "learning_rate": 8.289124877422315e-06, "loss": 0.6374, "step": 16146 }, { "epoch": 0.5689720482394708, "grad_norm": 1.3264458179473877, "learning_rate": 8.288000438936569e-06, "loss": 0.8938, "step": 16147 }, { "epoch": 0.5690072852524292, "grad_norm": 1.6610618829727173, "learning_rate": 8.286876022749834e-06, "loss": 1.1322, "step": 16148 }, { "epoch": 0.5690425222653875, "grad_norm": 3.8854243755340576, "learning_rate": 8.285751628876758e-06, "loss": 2.4996, "step": 16149 }, { "epoch": 0.569077759278346, "grad_norm": 2.928378105163574, "learning_rate": 8.284627257331983e-06, "loss": 3.0156, "step": 16150 }, { "epoch": 0.5691129962913044, "grad_norm": 1.8847944736480713, "learning_rate": 8.283502908130157e-06, "loss": 1.1432, "step": 16151 }, { "epoch": 0.5691482333042628, "grad_norm": 1.042418360710144, "learning_rate": 8.282378581285925e-06, "loss": 0.6086, "step": 16152 }, { "epoch": 0.5691834703172212, "grad_norm": 1.276366114616394, "learning_rate": 8.281254276813934e-06, "loss": 0.9683, "step": 16153 }, { "epoch": 0.5692187073301797, "grad_norm": 2.373821496963501, "learning_rate": 8.28012999472882e-06, "loss": 0.7512, "step": 16154 }, { "epoch": 0.569253944343138, "grad_norm": 3.5593278408050537, "learning_rate": 8.279005735045232e-06, "loss": 3.0169, "step": 16155 }, { "epoch": 0.5692891813560964, "grad_norm": 6.893791675567627, "learning_rate": 8.277881497777818e-06, "loss": 5.3455, "step": 16156 }, { "epoch": 0.5693244183690549, "grad_norm": 10.498836517333984, "learning_rate": 8.276757282941214e-06, "loss": 5.625, "step": 16157 }, { "epoch": 0.5693596553820133, "grad_norm": 3.2096498012542725, "learning_rate": 8.275633090550064e-06, "loss": 2.83, "step": 16158 }, { "epoch": 0.5693948923949717, "grad_norm": 3.8131144046783447, "learning_rate": 8.27450892061902e-06, "loss": 3.0197, "step": 16159 }, { "epoch": 0.5694301294079301, "grad_norm": 5.517795562744141, "learning_rate": 8.273384773162708e-06, "loss": 2.9296, "step": 16160 }, { "epoch": 0.5694653664208885, "grad_norm": 1.489479899406433, "learning_rate": 8.272260648195784e-06, "loss": 0.7842, "step": 16161 }, { "epoch": 0.5695006034338469, "grad_norm": 11.136953353881836, "learning_rate": 8.271136545732885e-06, "loss": 7.5939, "step": 16162 }, { "epoch": 0.5695358404468053, "grad_norm": 6.914123058319092, "learning_rate": 8.270012465788652e-06, "loss": 7.3016, "step": 16163 }, { "epoch": 0.5695710774597638, "grad_norm": 3.956670045852661, "learning_rate": 8.268888408377725e-06, "loss": 3.1856, "step": 16164 }, { "epoch": 0.5696063144727221, "grad_norm": 1.0909713506698608, "learning_rate": 8.267764373514752e-06, "loss": 0.7282, "step": 16165 }, { "epoch": 0.5696415514856805, "grad_norm": 4.217411994934082, "learning_rate": 8.266640361214364e-06, "loss": 3.2262, "step": 16166 }, { "epoch": 0.569676788498639, "grad_norm": 6.174971103668213, "learning_rate": 8.26551637149121e-06, "loss": 4.8038, "step": 16167 }, { "epoch": 0.5697120255115974, "grad_norm": 1.8533844947814941, "learning_rate": 8.264392404359924e-06, "loss": 0.8283, "step": 16168 }, { "epoch": 0.5697472625245558, "grad_norm": 1.7087419033050537, "learning_rate": 8.263268459835146e-06, "loss": 0.9809, "step": 16169 }, { "epoch": 0.5697824995375143, "grad_norm": 1.4852427244186401, "learning_rate": 8.26214453793152e-06, "loss": 0.9053, "step": 16170 }, { "epoch": 0.5698177365504726, "grad_norm": 7.086819171905518, "learning_rate": 8.261020638663682e-06, "loss": 2.0825, "step": 16171 }, { "epoch": 0.569852973563431, "grad_norm": 2.5806288719177246, "learning_rate": 8.259896762046274e-06, "loss": 0.972, "step": 16172 }, { "epoch": 0.5698882105763894, "grad_norm": 1.0511863231658936, "learning_rate": 8.258772908093927e-06, "loss": 0.8093, "step": 16173 }, { "epoch": 0.5699234475893479, "grad_norm": 4.628970623016357, "learning_rate": 8.257649076821288e-06, "loss": 3.2354, "step": 16174 }, { "epoch": 0.5699586846023063, "grad_norm": 2.4869794845581055, "learning_rate": 8.256525268242995e-06, "loss": 1.1161, "step": 16175 }, { "epoch": 0.5699939216152646, "grad_norm": 1.6399884223937988, "learning_rate": 8.255401482373678e-06, "loss": 0.8814, "step": 16176 }, { "epoch": 0.5700291586282231, "grad_norm": 4.163149356842041, "learning_rate": 8.254277719227978e-06, "loss": 3.2869, "step": 16177 }, { "epoch": 0.5700643956411815, "grad_norm": 2.0119946002960205, "learning_rate": 8.25315397882054e-06, "loss": 0.6761, "step": 16178 }, { "epoch": 0.5700996326541399, "grad_norm": 1.3454036712646484, "learning_rate": 8.252030261165986e-06, "loss": 0.7018, "step": 16179 }, { "epoch": 0.5701348696670984, "grad_norm": 1.382590413093567, "learning_rate": 8.250906566278963e-06, "loss": 0.6243, "step": 16180 }, { "epoch": 0.5701701066800567, "grad_norm": 1.325903296470642, "learning_rate": 8.249782894174108e-06, "loss": 0.7184, "step": 16181 }, { "epoch": 0.5702053436930151, "grad_norm": 5.4407782554626465, "learning_rate": 8.24865924486605e-06, "loss": 5.9178, "step": 16182 }, { "epoch": 0.5702405807059736, "grad_norm": 3.532249927520752, "learning_rate": 8.247535618369428e-06, "loss": 2.842, "step": 16183 }, { "epoch": 0.570275817718932, "grad_norm": 8.198448181152344, "learning_rate": 8.246412014698882e-06, "loss": 5.1046, "step": 16184 }, { "epoch": 0.5703110547318904, "grad_norm": 6.3489580154418945, "learning_rate": 8.245288433869038e-06, "loss": 5.8362, "step": 16185 }, { "epoch": 0.5703462917448487, "grad_norm": 4.027195930480957, "learning_rate": 8.24416487589454e-06, "loss": 3.1773, "step": 16186 }, { "epoch": 0.5703815287578072, "grad_norm": 8.890419960021973, "learning_rate": 8.243041340790016e-06, "loss": 5.4522, "step": 16187 }, { "epoch": 0.5704167657707656, "grad_norm": 1.040442705154419, "learning_rate": 8.2419178285701e-06, "loss": 1.0204, "step": 16188 }, { "epoch": 0.570452002783724, "grad_norm": 1.4010502099990845, "learning_rate": 8.240794339249432e-06, "loss": 0.7145, "step": 16189 }, { "epoch": 0.5704872397966825, "grad_norm": 1.1668370962142944, "learning_rate": 8.23967087284264e-06, "loss": 0.9077, "step": 16190 }, { "epoch": 0.5705224768096409, "grad_norm": 1.024789571762085, "learning_rate": 8.238547429364361e-06, "loss": 0.5635, "step": 16191 }, { "epoch": 0.5705577138225992, "grad_norm": 1.2965010404586792, "learning_rate": 8.237424008829221e-06, "loss": 1.0515, "step": 16192 }, { "epoch": 0.5705929508355577, "grad_norm": 1.0651233196258545, "learning_rate": 8.236300611251863e-06, "loss": 1.0439, "step": 16193 }, { "epoch": 0.5706281878485161, "grad_norm": 3.302424907684326, "learning_rate": 8.235177236646913e-06, "loss": 2.5189, "step": 16194 }, { "epoch": 0.5706634248614745, "grad_norm": 1.2898181676864624, "learning_rate": 8.234053885029005e-06, "loss": 1.0086, "step": 16195 }, { "epoch": 0.5706986618744329, "grad_norm": 7.74549674987793, "learning_rate": 8.232930556412766e-06, "loss": 3.1364, "step": 16196 }, { "epoch": 0.5707338988873913, "grad_norm": 4.618804454803467, "learning_rate": 8.23180725081284e-06, "loss": 2.6224, "step": 16197 }, { "epoch": 0.5707691359003497, "grad_norm": 4.805073261260986, "learning_rate": 8.23068396824384e-06, "loss": 2.8405, "step": 16198 }, { "epoch": 0.5708043729133081, "grad_norm": 5.898482799530029, "learning_rate": 8.229560708720413e-06, "loss": 2.8378, "step": 16199 }, { "epoch": 0.5708396099262666, "grad_norm": 4.730830669403076, "learning_rate": 8.228437472257182e-06, "loss": 4.8576, "step": 16200 }, { "epoch": 0.570874846939225, "grad_norm": 5.486391067504883, "learning_rate": 8.227314258868778e-06, "loss": 2.6592, "step": 16201 }, { "epoch": 0.5709100839521833, "grad_norm": 2.2037224769592285, "learning_rate": 8.226191068569829e-06, "loss": 0.845, "step": 16202 }, { "epoch": 0.5709453209651418, "grad_norm": 1.582170844078064, "learning_rate": 8.225067901374973e-06, "loss": 0.7169, "step": 16203 }, { "epoch": 0.5709805579781002, "grad_norm": 17.214353561401367, "learning_rate": 8.223944757298829e-06, "loss": 7.8362, "step": 16204 }, { "epoch": 0.5710157949910586, "grad_norm": 3.9795281887054443, "learning_rate": 8.222821636356034e-06, "loss": 3.0781, "step": 16205 }, { "epoch": 0.571051032004017, "grad_norm": 4.414933204650879, "learning_rate": 8.221698538561213e-06, "loss": 2.7448, "step": 16206 }, { "epoch": 0.5710862690169755, "grad_norm": 3.4397153854370117, "learning_rate": 8.220575463928992e-06, "loss": 2.9558, "step": 16207 }, { "epoch": 0.5711215060299338, "grad_norm": 3.0089733600616455, "learning_rate": 8.219452412474003e-06, "loss": 2.6363, "step": 16208 }, { "epoch": 0.5711567430428922, "grad_norm": 6.229884624481201, "learning_rate": 8.218329384210875e-06, "loss": 5.1292, "step": 16209 }, { "epoch": 0.5711919800558507, "grad_norm": 1.0943495035171509, "learning_rate": 8.217206379154234e-06, "loss": 0.7955, "step": 16210 }, { "epoch": 0.5712272170688091, "grad_norm": 1.001713514328003, "learning_rate": 8.216083397318704e-06, "loss": 1.2001, "step": 16211 }, { "epoch": 0.5712624540817675, "grad_norm": 6.526747226715088, "learning_rate": 8.214960438718917e-06, "loss": 3.13, "step": 16212 }, { "epoch": 0.571297691094726, "grad_norm": 1.3434253931045532, "learning_rate": 8.213837503369499e-06, "loss": 0.8862, "step": 16213 }, { "epoch": 0.5713329281076843, "grad_norm": 1.4008699655532837, "learning_rate": 8.212714591285074e-06, "loss": 0.8636, "step": 16214 }, { "epoch": 0.5713681651206427, "grad_norm": 3.7783918380737305, "learning_rate": 8.211591702480266e-06, "loss": 3.1479, "step": 16215 }, { "epoch": 0.5714034021336011, "grad_norm": 1.820619821548462, "learning_rate": 8.210468836969711e-06, "loss": 1.0648, "step": 16216 }, { "epoch": 0.5714386391465596, "grad_norm": 15.191678047180176, "learning_rate": 8.209345994768022e-06, "loss": 3.165, "step": 16217 }, { "epoch": 0.571473876159518, "grad_norm": 3.7049670219421387, "learning_rate": 8.208223175889829e-06, "loss": 2.8888, "step": 16218 }, { "epoch": 0.5715091131724763, "grad_norm": 7.502945423126221, "learning_rate": 8.20710038034976e-06, "loss": 2.9193, "step": 16219 }, { "epoch": 0.5715443501854348, "grad_norm": 2.5813560485839844, "learning_rate": 8.205977608162437e-06, "loss": 1.0573, "step": 16220 }, { "epoch": 0.5715795871983932, "grad_norm": 2.1940879821777344, "learning_rate": 8.20485485934248e-06, "loss": 1.0017, "step": 16221 }, { "epoch": 0.5716148242113516, "grad_norm": 1.4010357856750488, "learning_rate": 8.203732133904523e-06, "loss": 1.2086, "step": 16222 }, { "epoch": 0.5716500612243101, "grad_norm": 0.9772736430168152, "learning_rate": 8.20260943186318e-06, "loss": 1.0039, "step": 16223 }, { "epoch": 0.5716852982372684, "grad_norm": 9.278470039367676, "learning_rate": 8.201486753233079e-06, "loss": 2.7085, "step": 16224 }, { "epoch": 0.5717205352502268, "grad_norm": 4.268360614776611, "learning_rate": 8.20036409802884e-06, "loss": 3.2076, "step": 16225 }, { "epoch": 0.5717557722631853, "grad_norm": 8.191638946533203, "learning_rate": 8.199241466265091e-06, "loss": 5.7336, "step": 16226 }, { "epoch": 0.5717910092761437, "grad_norm": 1.5387942790985107, "learning_rate": 8.198118857956451e-06, "loss": 0.8593, "step": 16227 }, { "epoch": 0.5718262462891021, "grad_norm": 0.9208574891090393, "learning_rate": 8.196996273117541e-06, "loss": 0.9261, "step": 16228 }, { "epoch": 0.5718614833020604, "grad_norm": 0.8802427649497986, "learning_rate": 8.195873711762987e-06, "loss": 0.8048, "step": 16229 }, { "epoch": 0.5718967203150189, "grad_norm": 4.6144890785217285, "learning_rate": 8.194751173907403e-06, "loss": 3.0821, "step": 16230 }, { "epoch": 0.5719319573279773, "grad_norm": 3.800868511199951, "learning_rate": 8.193628659565419e-06, "loss": 2.7509, "step": 16231 }, { "epoch": 0.5719671943409357, "grad_norm": 1.7780896425247192, "learning_rate": 8.19250616875165e-06, "loss": 0.7083, "step": 16232 }, { "epoch": 0.5720024313538942, "grad_norm": 6.666990756988525, "learning_rate": 8.19138370148072e-06, "loss": 3.0465, "step": 16233 }, { "epoch": 0.5720376683668525, "grad_norm": 1.3824862241744995, "learning_rate": 8.190261257767244e-06, "loss": 0.9985, "step": 16234 }, { "epoch": 0.5720729053798109, "grad_norm": 1.7728252410888672, "learning_rate": 8.189138837625852e-06, "loss": 0.8239, "step": 16235 }, { "epoch": 0.5721081423927694, "grad_norm": 16.099824905395508, "learning_rate": 8.18801644107115e-06, "loss": 3.3779, "step": 16236 }, { "epoch": 0.5721433794057278, "grad_norm": 3.9042882919311523, "learning_rate": 8.186894068117768e-06, "loss": 2.4572, "step": 16237 }, { "epoch": 0.5721786164186862, "grad_norm": 10.756723403930664, "learning_rate": 8.185771718780324e-06, "loss": 3.4438, "step": 16238 }, { "epoch": 0.5722138534316445, "grad_norm": 2.4661943912506104, "learning_rate": 8.184649393073432e-06, "loss": 0.8298, "step": 16239 }, { "epoch": 0.572249090444603, "grad_norm": 6.56673002243042, "learning_rate": 8.183527091011711e-06, "loss": 2.6936, "step": 16240 }, { "epoch": 0.5722843274575614, "grad_norm": 3.084568500518799, "learning_rate": 8.182404812609785e-06, "loss": 0.9966, "step": 16241 }, { "epoch": 0.5723195644705198, "grad_norm": 1.1539578437805176, "learning_rate": 8.181282557882265e-06, "loss": 0.7557, "step": 16242 }, { "epoch": 0.5723548014834783, "grad_norm": 1.3186382055282593, "learning_rate": 8.180160326843772e-06, "loss": 0.9374, "step": 16243 }, { "epoch": 0.5723900384964367, "grad_norm": 1.2062954902648926, "learning_rate": 8.17903811950892e-06, "loss": 0.7199, "step": 16244 }, { "epoch": 0.572425275509395, "grad_norm": 7.683018684387207, "learning_rate": 8.177915935892334e-06, "loss": 2.721, "step": 16245 }, { "epoch": 0.5724605125223535, "grad_norm": 1.1322029829025269, "learning_rate": 8.176793776008622e-06, "loss": 1.0574, "step": 16246 }, { "epoch": 0.5724957495353119, "grad_norm": 1.0903912782669067, "learning_rate": 8.175671639872403e-06, "loss": 0.9385, "step": 16247 }, { "epoch": 0.5725309865482703, "grad_norm": 1.1821246147155762, "learning_rate": 8.174549527498296e-06, "loss": 0.8082, "step": 16248 }, { "epoch": 0.5725662235612287, "grad_norm": 6.5694580078125, "learning_rate": 8.173427438900909e-06, "loss": 3.2071, "step": 16249 }, { "epoch": 0.5726014605741871, "grad_norm": 5.120858192443848, "learning_rate": 8.172305374094866e-06, "loss": 2.9897, "step": 16250 }, { "epoch": 0.5726366975871455, "grad_norm": 1.057166337966919, "learning_rate": 8.171183333094779e-06, "loss": 0.9171, "step": 16251 }, { "epoch": 0.5726719346001039, "grad_norm": 2.0125386714935303, "learning_rate": 8.170061315915261e-06, "loss": 0.8838, "step": 16252 }, { "epoch": 0.5727071716130624, "grad_norm": 1.0932507514953613, "learning_rate": 8.168939322570923e-06, "loss": 1.2026, "step": 16253 }, { "epoch": 0.5727424086260208, "grad_norm": 7.655388832092285, "learning_rate": 8.167817353076392e-06, "loss": 5.2892, "step": 16254 }, { "epoch": 0.5727776456389791, "grad_norm": 1.7014093399047852, "learning_rate": 8.166695407446267e-06, "loss": 1.074, "step": 16255 }, { "epoch": 0.5728128826519376, "grad_norm": 2.1992359161376953, "learning_rate": 8.16557348569517e-06, "loss": 0.6607, "step": 16256 }, { "epoch": 0.572848119664896, "grad_norm": 4.8094048500061035, "learning_rate": 8.164451587837713e-06, "loss": 3.1154, "step": 16257 }, { "epoch": 0.5728833566778544, "grad_norm": 2.869732141494751, "learning_rate": 8.163329713888505e-06, "loss": 0.9645, "step": 16258 }, { "epoch": 0.5729185936908129, "grad_norm": 1.9869272708892822, "learning_rate": 8.16220786386216e-06, "loss": 0.9841, "step": 16259 }, { "epoch": 0.5729538307037713, "grad_norm": 1.1038272380828857, "learning_rate": 8.161086037773298e-06, "loss": 0.7278, "step": 16260 }, { "epoch": 0.5729890677167296, "grad_norm": 1.0199891328811646, "learning_rate": 8.159964235636522e-06, "loss": 1.1821, "step": 16261 }, { "epoch": 0.573024304729688, "grad_norm": 1.0369915962219238, "learning_rate": 8.158842457466446e-06, "loss": 0.6818, "step": 16262 }, { "epoch": 0.5730595417426465, "grad_norm": 6.936855316162109, "learning_rate": 8.157720703277678e-06, "loss": 3.5765, "step": 16263 }, { "epoch": 0.5730947787556049, "grad_norm": 2.535548210144043, "learning_rate": 8.156598973084837e-06, "loss": 2.1991, "step": 16264 }, { "epoch": 0.5731300157685633, "grad_norm": 4.403383731842041, "learning_rate": 8.155477266902527e-06, "loss": 3.2322, "step": 16265 }, { "epoch": 0.5731652527815218, "grad_norm": 5.020127773284912, "learning_rate": 8.154355584745361e-06, "loss": 2.5945, "step": 16266 }, { "epoch": 0.5732004897944801, "grad_norm": 7.4032464027404785, "learning_rate": 8.153233926627951e-06, "loss": 3.5682, "step": 16267 }, { "epoch": 0.5732357268074385, "grad_norm": 1.1369332075119019, "learning_rate": 8.152112292564903e-06, "loss": 1.038, "step": 16268 }, { "epoch": 0.573270963820397, "grad_norm": 2.108591318130493, "learning_rate": 8.150990682570824e-06, "loss": 0.8188, "step": 16269 }, { "epoch": 0.5733062008333554, "grad_norm": 13.221007347106934, "learning_rate": 8.149869096660332e-06, "loss": 5.2865, "step": 16270 }, { "epoch": 0.5733414378463138, "grad_norm": 4.8057379722595215, "learning_rate": 8.148747534848028e-06, "loss": 2.4853, "step": 16271 }, { "epoch": 0.5733766748592721, "grad_norm": 0.957806408405304, "learning_rate": 8.147625997148524e-06, "loss": 0.9382, "step": 16272 }, { "epoch": 0.5734119118722306, "grad_norm": 1.164025068283081, "learning_rate": 8.146504483576429e-06, "loss": 0.8583, "step": 16273 }, { "epoch": 0.573447148885189, "grad_norm": 1.4343677759170532, "learning_rate": 8.145382994146345e-06, "loss": 0.8973, "step": 16274 }, { "epoch": 0.5734823858981474, "grad_norm": 6.021832466125488, "learning_rate": 8.144261528872886e-06, "loss": 3.2142, "step": 16275 }, { "epoch": 0.5735176229111059, "grad_norm": 1.033320426940918, "learning_rate": 8.143140087770658e-06, "loss": 1.0101, "step": 16276 }, { "epoch": 0.5735528599240642, "grad_norm": 7.2187581062316895, "learning_rate": 8.142018670854265e-06, "loss": 5.1085, "step": 16277 }, { "epoch": 0.5735880969370226, "grad_norm": 1.1561822891235352, "learning_rate": 8.140897278138315e-06, "loss": 1.1344, "step": 16278 }, { "epoch": 0.5736233339499811, "grad_norm": 1.8260600566864014, "learning_rate": 8.139775909637421e-06, "loss": 0.8223, "step": 16279 }, { "epoch": 0.5736585709629395, "grad_norm": 4.158740043640137, "learning_rate": 8.138654565366176e-06, "loss": 3.2523, "step": 16280 }, { "epoch": 0.5736938079758979, "grad_norm": 1.9041517972946167, "learning_rate": 8.137533245339193e-06, "loss": 0.7581, "step": 16281 }, { "epoch": 0.5737290449888562, "grad_norm": 5.005270957946777, "learning_rate": 8.136411949571075e-06, "loss": 5.1187, "step": 16282 }, { "epoch": 0.5737642820018147, "grad_norm": 3.771333932876587, "learning_rate": 8.135290678076436e-06, "loss": 3.0126, "step": 16283 }, { "epoch": 0.5737995190147731, "grad_norm": 0.9610435962677002, "learning_rate": 8.134169430869868e-06, "loss": 0.7586, "step": 16284 }, { "epoch": 0.5738347560277315, "grad_norm": 4.464406967163086, "learning_rate": 8.133048207965982e-06, "loss": 3.2471, "step": 16285 }, { "epoch": 0.57386999304069, "grad_norm": 5.621157646179199, "learning_rate": 8.131927009379382e-06, "loss": 4.9775, "step": 16286 }, { "epoch": 0.5739052300536484, "grad_norm": 1.8699325323104858, "learning_rate": 8.130805835124669e-06, "loss": 0.9155, "step": 16287 }, { "epoch": 0.5739404670666067, "grad_norm": 5.832091808319092, "learning_rate": 8.129684685216446e-06, "loss": 4.4395, "step": 16288 }, { "epoch": 0.5739757040795652, "grad_norm": 1.2151051759719849, "learning_rate": 8.128563559669323e-06, "loss": 0.7225, "step": 16289 }, { "epoch": 0.5740109410925236, "grad_norm": 1.5714613199234009, "learning_rate": 8.127442458497897e-06, "loss": 1.0228, "step": 16290 }, { "epoch": 0.574046178105482, "grad_norm": 2.0460407733917236, "learning_rate": 8.126321381716768e-06, "loss": 0.8893, "step": 16291 }, { "epoch": 0.5740814151184405, "grad_norm": 5.731200695037842, "learning_rate": 8.125200329340547e-06, "loss": 3.772, "step": 16292 }, { "epoch": 0.5741166521313988, "grad_norm": 1.148518681526184, "learning_rate": 8.124079301383825e-06, "loss": 0.7685, "step": 16293 }, { "epoch": 0.5741518891443572, "grad_norm": 7.85646390914917, "learning_rate": 8.122958297861211e-06, "loss": 2.7683, "step": 16294 }, { "epoch": 0.5741871261573156, "grad_norm": 1.6833813190460205, "learning_rate": 8.121837318787306e-06, "loss": 1.0292, "step": 16295 }, { "epoch": 0.5742223631702741, "grad_norm": 1.4202431440353394, "learning_rate": 8.120716364176709e-06, "loss": 1.0742, "step": 16296 }, { "epoch": 0.5742576001832325, "grad_norm": 0.9780586361885071, "learning_rate": 8.119595434044018e-06, "loss": 0.8429, "step": 16297 }, { "epoch": 0.5742928371961908, "grad_norm": 1.480093002319336, "learning_rate": 8.118474528403838e-06, "loss": 1.0407, "step": 16298 }, { "epoch": 0.5743280742091493, "grad_norm": 3.3829941749572754, "learning_rate": 8.117353647270766e-06, "loss": 3.1118, "step": 16299 }, { "epoch": 0.5743633112221077, "grad_norm": 3.8233113288879395, "learning_rate": 8.116232790659404e-06, "loss": 3.2221, "step": 16300 }, { "epoch": 0.5743985482350661, "grad_norm": 3.6547257900238037, "learning_rate": 8.115111958584347e-06, "loss": 2.1642, "step": 16301 }, { "epoch": 0.5744337852480246, "grad_norm": 5.692656993865967, "learning_rate": 8.113991151060202e-06, "loss": 3.0808, "step": 16302 }, { "epoch": 0.574469022260983, "grad_norm": 3.751523017883301, "learning_rate": 8.112870368101558e-06, "loss": 3.4502, "step": 16303 }, { "epoch": 0.5745042592739413, "grad_norm": 4.536850929260254, "learning_rate": 8.11174960972302e-06, "loss": 3.1868, "step": 16304 }, { "epoch": 0.5745394962868997, "grad_norm": 1.5774028301239014, "learning_rate": 8.110628875939183e-06, "loss": 1.1471, "step": 16305 }, { "epoch": 0.5745747332998582, "grad_norm": 1.1943931579589844, "learning_rate": 8.109508166764647e-06, "loss": 0.8444, "step": 16306 }, { "epoch": 0.5746099703128166, "grad_norm": 1.2423683404922485, "learning_rate": 8.108387482214005e-06, "loss": 0.9234, "step": 16307 }, { "epoch": 0.574645207325775, "grad_norm": 16.02474021911621, "learning_rate": 8.10726682230186e-06, "loss": 1.9332, "step": 16308 }, { "epoch": 0.5746804443387334, "grad_norm": 5.160621166229248, "learning_rate": 8.106146187042804e-06, "loss": 3.0308, "step": 16309 }, { "epoch": 0.5747156813516918, "grad_norm": 7.259384632110596, "learning_rate": 8.105025576451435e-06, "loss": 2.97, "step": 16310 }, { "epoch": 0.5747509183646502, "grad_norm": 1.1229422092437744, "learning_rate": 8.103904990542351e-06, "loss": 0.7892, "step": 16311 }, { "epoch": 0.5747861553776087, "grad_norm": 1.1527025699615479, "learning_rate": 8.102784429330145e-06, "loss": 0.6896, "step": 16312 }, { "epoch": 0.5748213923905671, "grad_norm": 1.062546730041504, "learning_rate": 8.101663892829411e-06, "loss": 0.8001, "step": 16313 }, { "epoch": 0.5748566294035254, "grad_norm": 2.9008240699768066, "learning_rate": 8.10054338105475e-06, "loss": 2.7267, "step": 16314 }, { "epoch": 0.5748918664164838, "grad_norm": 3.993468999862671, "learning_rate": 8.099422894020752e-06, "loss": 2.9108, "step": 16315 }, { "epoch": 0.5749271034294423, "grad_norm": 3.539341926574707, "learning_rate": 8.098302431742013e-06, "loss": 3.1277, "step": 16316 }, { "epoch": 0.5749623404424007, "grad_norm": 44.98651885986328, "learning_rate": 8.097181994233127e-06, "loss": 3.0988, "step": 16317 }, { "epoch": 0.5749975774553591, "grad_norm": 8.543097496032715, "learning_rate": 8.09606158150869e-06, "loss": 5.1031, "step": 16318 }, { "epoch": 0.5750328144683176, "grad_norm": 4.531136989593506, "learning_rate": 8.09494119358329e-06, "loss": 3.0123, "step": 16319 }, { "epoch": 0.5750680514812759, "grad_norm": 3.728484869003296, "learning_rate": 8.093820830471524e-06, "loss": 2.5562, "step": 16320 }, { "epoch": 0.5751032884942343, "grad_norm": 1.5553051233291626, "learning_rate": 8.092700492187989e-06, "loss": 0.7533, "step": 16321 }, { "epoch": 0.5751385255071928, "grad_norm": 2.489258289337158, "learning_rate": 8.091580178747266e-06, "loss": 0.9494, "step": 16322 }, { "epoch": 0.5751737625201512, "grad_norm": 15.629277229309082, "learning_rate": 8.09045989016396e-06, "loss": 7.6258, "step": 16323 }, { "epoch": 0.5752089995331096, "grad_norm": 1.2412160634994507, "learning_rate": 8.089339626452656e-06, "loss": 0.8976, "step": 16324 }, { "epoch": 0.575244236546068, "grad_norm": 1.5560386180877686, "learning_rate": 8.088219387627947e-06, "loss": 0.9657, "step": 16325 }, { "epoch": 0.5752794735590264, "grad_norm": 4.216660976409912, "learning_rate": 8.087099173704421e-06, "loss": 3.1515, "step": 16326 }, { "epoch": 0.5753147105719848, "grad_norm": 1.2694833278656006, "learning_rate": 8.085978984696674e-06, "loss": 0.9876, "step": 16327 }, { "epoch": 0.5753499475849432, "grad_norm": 3.9234795570373535, "learning_rate": 8.084858820619295e-06, "loss": 4.0518, "step": 16328 }, { "epoch": 0.5753851845979017, "grad_norm": 1.0736370086669922, "learning_rate": 8.083738681486872e-06, "loss": 1.1021, "step": 16329 }, { "epoch": 0.57542042161086, "grad_norm": 2.0662753582000732, "learning_rate": 8.082618567314e-06, "loss": 0.91, "step": 16330 }, { "epoch": 0.5754556586238184, "grad_norm": 1.0560048818588257, "learning_rate": 8.081498478115262e-06, "loss": 0.862, "step": 16331 }, { "epoch": 0.5754908956367769, "grad_norm": 2.638556480407715, "learning_rate": 8.080378413905251e-06, "loss": 0.7798, "step": 16332 }, { "epoch": 0.5755261326497353, "grad_norm": 7.0538530349731445, "learning_rate": 8.079258374698559e-06, "loss": 6.471, "step": 16333 }, { "epoch": 0.5755613696626937, "grad_norm": 7.495184421539307, "learning_rate": 8.078138360509769e-06, "loss": 4.0815, "step": 16334 }, { "epoch": 0.5755966066756522, "grad_norm": 1.357340931892395, "learning_rate": 8.077018371353469e-06, "loss": 0.9724, "step": 16335 }, { "epoch": 0.5756318436886105, "grad_norm": 8.110888481140137, "learning_rate": 8.075898407244252e-06, "loss": 4.6634, "step": 16336 }, { "epoch": 0.5756670807015689, "grad_norm": 1.119368076324463, "learning_rate": 8.074778468196707e-06, "loss": 1.1131, "step": 16337 }, { "epoch": 0.5757023177145273, "grad_norm": 1.7184590101242065, "learning_rate": 8.073658554225413e-06, "loss": 1.0409, "step": 16338 }, { "epoch": 0.5757375547274858, "grad_norm": 1.2820758819580078, "learning_rate": 8.072538665344961e-06, "loss": 1.0312, "step": 16339 }, { "epoch": 0.5757727917404442, "grad_norm": 8.267738342285156, "learning_rate": 8.071418801569944e-06, "loss": 5.8123, "step": 16340 }, { "epoch": 0.5758080287534025, "grad_norm": 4.4970011711120605, "learning_rate": 8.070298962914938e-06, "loss": 5.1929, "step": 16341 }, { "epoch": 0.575843265766361, "grad_norm": 5.923201560974121, "learning_rate": 8.069179149394535e-06, "loss": 5.2408, "step": 16342 }, { "epoch": 0.5758785027793194, "grad_norm": 9.969606399536133, "learning_rate": 8.068059361023322e-06, "loss": 3.2826, "step": 16343 }, { "epoch": 0.5759137397922778, "grad_norm": 2.456913709640503, "learning_rate": 8.066939597815878e-06, "loss": 2.6796, "step": 16344 }, { "epoch": 0.5759489768052363, "grad_norm": 0.9644672274589539, "learning_rate": 8.065819859786792e-06, "loss": 0.7043, "step": 16345 }, { "epoch": 0.5759842138181946, "grad_norm": 2.578425168991089, "learning_rate": 8.064700146950651e-06, "loss": 2.9201, "step": 16346 }, { "epoch": 0.576019450831153, "grad_norm": 4.049433708190918, "learning_rate": 8.063580459322037e-06, "loss": 3.1789, "step": 16347 }, { "epoch": 0.5760546878441114, "grad_norm": 6.3952860832214355, "learning_rate": 8.062460796915532e-06, "loss": 5.235, "step": 16348 }, { "epoch": 0.5760899248570699, "grad_norm": 8.024953842163086, "learning_rate": 8.061341159745725e-06, "loss": 3.6532, "step": 16349 }, { "epoch": 0.5761251618700283, "grad_norm": 1.6516661643981934, "learning_rate": 8.060221547827194e-06, "loss": 0.8862, "step": 16350 }, { "epoch": 0.5761603988829866, "grad_norm": 5.016360759735107, "learning_rate": 8.059101961174525e-06, "loss": 4.2373, "step": 16351 }, { "epoch": 0.5761956358959451, "grad_norm": 6.1060309410095215, "learning_rate": 8.0579823998023e-06, "loss": 4.6969, "step": 16352 }, { "epoch": 0.5762308729089035, "grad_norm": 5.676056861877441, "learning_rate": 8.056862863725104e-06, "loss": 3.329, "step": 16353 }, { "epoch": 0.5762661099218619, "grad_norm": 3.7642695903778076, "learning_rate": 8.055743352957512e-06, "loss": 2.7732, "step": 16354 }, { "epoch": 0.5763013469348204, "grad_norm": 1.0435786247253418, "learning_rate": 8.054623867514113e-06, "loss": 0.8085, "step": 16355 }, { "epoch": 0.5763365839477788, "grad_norm": 0.9414821267127991, "learning_rate": 8.053504407409488e-06, "loss": 0.8925, "step": 16356 }, { "epoch": 0.5763718209607371, "grad_norm": 1.0886929035186768, "learning_rate": 8.052384972658214e-06, "loss": 0.7345, "step": 16357 }, { "epoch": 0.5764070579736956, "grad_norm": 1.1746376752853394, "learning_rate": 8.051265563274872e-06, "loss": 1.0766, "step": 16358 }, { "epoch": 0.576442294986654, "grad_norm": 1.2737406492233276, "learning_rate": 8.05014617927405e-06, "loss": 0.8875, "step": 16359 }, { "epoch": 0.5764775319996124, "grad_norm": 1.1775234937667847, "learning_rate": 8.049026820670315e-06, "loss": 1.0597, "step": 16360 }, { "epoch": 0.5765127690125708, "grad_norm": 1.2963603734970093, "learning_rate": 8.04790748747826e-06, "loss": 0.7789, "step": 16361 }, { "epoch": 0.5765480060255292, "grad_norm": 6.03078556060791, "learning_rate": 8.046788179712457e-06, "loss": 2.9103, "step": 16362 }, { "epoch": 0.5765832430384876, "grad_norm": 3.315753221511841, "learning_rate": 8.045668897387486e-06, "loss": 0.9186, "step": 16363 }, { "epoch": 0.576618480051446, "grad_norm": 4.823299407958984, "learning_rate": 8.044549640517925e-06, "loss": 2.7817, "step": 16364 }, { "epoch": 0.5766537170644045, "grad_norm": 5.493121147155762, "learning_rate": 8.043430409118357e-06, "loss": 2.7251, "step": 16365 }, { "epoch": 0.5766889540773629, "grad_norm": 10.94846248626709, "learning_rate": 8.042311203203356e-06, "loss": 3.5999, "step": 16366 }, { "epoch": 0.5767241910903212, "grad_norm": 4.730111122131348, "learning_rate": 8.0411920227875e-06, "loss": 1.067, "step": 16367 }, { "epoch": 0.5767594281032797, "grad_norm": 1.3687514066696167, "learning_rate": 8.040072867885372e-06, "loss": 0.881, "step": 16368 }, { "epoch": 0.5767946651162381, "grad_norm": 6.8780364990234375, "learning_rate": 8.038953738511537e-06, "loss": 3.239, "step": 16369 }, { "epoch": 0.5768299021291965, "grad_norm": 4.934865951538086, "learning_rate": 8.037834634680584e-06, "loss": 3.2079, "step": 16370 }, { "epoch": 0.5768651391421549, "grad_norm": 7.767095565795898, "learning_rate": 8.036715556407083e-06, "loss": 5.4965, "step": 16371 }, { "epoch": 0.5769003761551134, "grad_norm": 4.763613224029541, "learning_rate": 8.035596503705615e-06, "loss": 4.1252, "step": 16372 }, { "epoch": 0.5769356131680717, "grad_norm": 32.703922271728516, "learning_rate": 8.034477476590748e-06, "loss": 2.756, "step": 16373 }, { "epoch": 0.5769708501810301, "grad_norm": 7.860046863555908, "learning_rate": 8.033358475077065e-06, "loss": 6.8979, "step": 16374 }, { "epoch": 0.5770060871939886, "grad_norm": 5.195614814758301, "learning_rate": 8.03223949917914e-06, "loss": 5.0968, "step": 16375 }, { "epoch": 0.577041324206947, "grad_norm": 2.2051808834075928, "learning_rate": 8.031120548911544e-06, "loss": 0.9466, "step": 16376 }, { "epoch": 0.5770765612199054, "grad_norm": 1.274790644645691, "learning_rate": 8.030001624288853e-06, "loss": 0.8216, "step": 16377 }, { "epoch": 0.5771117982328638, "grad_norm": 3.3632123470306396, "learning_rate": 8.028882725325645e-06, "loss": 2.2868, "step": 16378 }, { "epoch": 0.5771470352458222, "grad_norm": 6.5070719718933105, "learning_rate": 8.027763852036488e-06, "loss": 5.0934, "step": 16379 }, { "epoch": 0.5771822722587806, "grad_norm": 5.1515631675720215, "learning_rate": 8.026645004435957e-06, "loss": 3.0608, "step": 16380 }, { "epoch": 0.577217509271739, "grad_norm": 2.2039926052093506, "learning_rate": 8.025526182538632e-06, "loss": 1.0292, "step": 16381 }, { "epoch": 0.5772527462846975, "grad_norm": 1.8711216449737549, "learning_rate": 8.024407386359074e-06, "loss": 0.9616, "step": 16382 }, { "epoch": 0.5772879832976558, "grad_norm": 1.646182894706726, "learning_rate": 8.023288615911861e-06, "loss": 0.8389, "step": 16383 }, { "epoch": 0.5773232203106142, "grad_norm": 2.634477138519287, "learning_rate": 8.02216987121157e-06, "loss": 2.7482, "step": 16384 }, { "epoch": 0.5773584573235727, "grad_norm": 1.5471247434616089, "learning_rate": 8.021051152272766e-06, "loss": 1.0921, "step": 16385 }, { "epoch": 0.5773936943365311, "grad_norm": 8.616257667541504, "learning_rate": 8.019932459110023e-06, "loss": 10.1304, "step": 16386 }, { "epoch": 0.5774289313494895, "grad_norm": 0.9665212035179138, "learning_rate": 8.018813791737912e-06, "loss": 0.7513, "step": 16387 }, { "epoch": 0.577464168362448, "grad_norm": 4.322688579559326, "learning_rate": 8.017695150171002e-06, "loss": 2.9644, "step": 16388 }, { "epoch": 0.5774994053754063, "grad_norm": 5.760797500610352, "learning_rate": 8.016576534423867e-06, "loss": 3.254, "step": 16389 }, { "epoch": 0.5775346423883647, "grad_norm": 3.6992337703704834, "learning_rate": 8.015457944511075e-06, "loss": 3.1202, "step": 16390 }, { "epoch": 0.5775698794013232, "grad_norm": 12.710639953613281, "learning_rate": 8.014339380447198e-06, "loss": 4.3639, "step": 16391 }, { "epoch": 0.5776051164142816, "grad_norm": 4.000575065612793, "learning_rate": 8.013220842246799e-06, "loss": 3.1059, "step": 16392 }, { "epoch": 0.57764035342724, "grad_norm": 1.4331881999969482, "learning_rate": 8.012102329924454e-06, "loss": 0.6715, "step": 16393 }, { "epoch": 0.5776755904401983, "grad_norm": 1.2089766263961792, "learning_rate": 8.010983843494732e-06, "loss": 0.964, "step": 16394 }, { "epoch": 0.5777108274531568, "grad_norm": 1.4522939920425415, "learning_rate": 8.009865382972195e-06, "loss": 0.9886, "step": 16395 }, { "epoch": 0.5777460644661152, "grad_norm": 0.9485310912132263, "learning_rate": 8.008746948371416e-06, "loss": 0.9839, "step": 16396 }, { "epoch": 0.5777813014790736, "grad_norm": 8.65326976776123, "learning_rate": 8.007628539706964e-06, "loss": 3.6343, "step": 16397 }, { "epoch": 0.5778165384920321, "grad_norm": 4.177578926086426, "learning_rate": 8.0065101569934e-06, "loss": 3.0553, "step": 16398 }, { "epoch": 0.5778517755049905, "grad_norm": 1.5926542282104492, "learning_rate": 8.005391800245297e-06, "loss": 1.0671, "step": 16399 }, { "epoch": 0.5778870125179488, "grad_norm": 1.1687679290771484, "learning_rate": 8.004273469477222e-06, "loss": 0.8257, "step": 16400 }, { "epoch": 0.5779222495309073, "grad_norm": 6.65951681137085, "learning_rate": 8.003155164703737e-06, "loss": 3.1607, "step": 16401 }, { "epoch": 0.5779574865438657, "grad_norm": 2.293797254562378, "learning_rate": 8.002036885939407e-06, "loss": 2.8039, "step": 16402 }, { "epoch": 0.5779927235568241, "grad_norm": 0.9657493829727173, "learning_rate": 8.000918633198808e-06, "loss": 1.0205, "step": 16403 }, { "epoch": 0.5780279605697825, "grad_norm": 27.2861385345459, "learning_rate": 7.999800406496494e-06, "loss": 3.0311, "step": 16404 }, { "epoch": 0.5780631975827409, "grad_norm": 1.3126156330108643, "learning_rate": 7.998682205847037e-06, "loss": 0.8531, "step": 16405 }, { "epoch": 0.5780984345956993, "grad_norm": 25.058326721191406, "learning_rate": 7.997564031265e-06, "loss": 6.4837, "step": 16406 }, { "epoch": 0.5781336716086577, "grad_norm": 5.663881301879883, "learning_rate": 7.99644588276494e-06, "loss": 3.369, "step": 16407 }, { "epoch": 0.5781689086216162, "grad_norm": 4.080422401428223, "learning_rate": 7.995327760361433e-06, "loss": 2.785, "step": 16408 }, { "epoch": 0.5782041456345746, "grad_norm": 3.53365421295166, "learning_rate": 7.994209664069037e-06, "loss": 2.8536, "step": 16409 }, { "epoch": 0.5782393826475329, "grad_norm": 1.5988566875457764, "learning_rate": 7.993091593902316e-06, "loss": 0.7996, "step": 16410 }, { "epoch": 0.5782746196604914, "grad_norm": 6.240988254547119, "learning_rate": 7.991973549875829e-06, "loss": 4.8095, "step": 16411 }, { "epoch": 0.5783098566734498, "grad_norm": 4.766180515289307, "learning_rate": 7.990855532004145e-06, "loss": 2.9999, "step": 16412 }, { "epoch": 0.5783450936864082, "grad_norm": 7.775813102722168, "learning_rate": 7.989737540301826e-06, "loss": 3.7668, "step": 16413 }, { "epoch": 0.5783803306993666, "grad_norm": 5.182733058929443, "learning_rate": 7.988619574783428e-06, "loss": 4.5548, "step": 16414 }, { "epoch": 0.578415567712325, "grad_norm": 5.13050651550293, "learning_rate": 7.987501635463518e-06, "loss": 4.9943, "step": 16415 }, { "epoch": 0.5784508047252834, "grad_norm": 0.9320746660232544, "learning_rate": 7.986383722356655e-06, "loss": 0.8989, "step": 16416 }, { "epoch": 0.5784860417382418, "grad_norm": 4.537327289581299, "learning_rate": 7.985265835477397e-06, "loss": 3.5682, "step": 16417 }, { "epoch": 0.5785212787512003, "grad_norm": 1.331068992614746, "learning_rate": 7.984147974840313e-06, "loss": 0.7956, "step": 16418 }, { "epoch": 0.5785565157641587, "grad_norm": 5.1312575340271, "learning_rate": 7.983030140459957e-06, "loss": 3.4143, "step": 16419 }, { "epoch": 0.578591752777117, "grad_norm": 2.782513380050659, "learning_rate": 7.981912332350892e-06, "loss": 2.3991, "step": 16420 }, { "epoch": 0.5786269897900755, "grad_norm": 1.3754262924194336, "learning_rate": 7.98079455052767e-06, "loss": 0.7298, "step": 16421 }, { "epoch": 0.5786622268030339, "grad_norm": 1.2437862157821655, "learning_rate": 7.979676795004862e-06, "loss": 0.8877, "step": 16422 }, { "epoch": 0.5786974638159923, "grad_norm": 1.2426623106002808, "learning_rate": 7.978559065797021e-06, "loss": 0.8966, "step": 16423 }, { "epoch": 0.5787327008289507, "grad_norm": 6.442774295806885, "learning_rate": 7.977441362918703e-06, "loss": 4.561, "step": 16424 }, { "epoch": 0.5787679378419092, "grad_norm": 1.115365743637085, "learning_rate": 7.976323686384466e-06, "loss": 0.9381, "step": 16425 }, { "epoch": 0.5788031748548675, "grad_norm": 4.101889610290527, "learning_rate": 7.97520603620888e-06, "loss": 2.241, "step": 16426 }, { "epoch": 0.5788384118678259, "grad_norm": 1.0588489770889282, "learning_rate": 7.974088412406487e-06, "loss": 0.9687, "step": 16427 }, { "epoch": 0.5788736488807844, "grad_norm": 5.098881244659424, "learning_rate": 7.97297081499185e-06, "loss": 4.6106, "step": 16428 }, { "epoch": 0.5789088858937428, "grad_norm": 1.2508692741394043, "learning_rate": 7.97185324397953e-06, "loss": 0.882, "step": 16429 }, { "epoch": 0.5789441229067012, "grad_norm": 5.169722080230713, "learning_rate": 7.970735699384078e-06, "loss": 4.6726, "step": 16430 }, { "epoch": 0.5789793599196597, "grad_norm": 8.546053886413574, "learning_rate": 7.96961818122005e-06, "loss": 5.0246, "step": 16431 }, { "epoch": 0.579014596932618, "grad_norm": 4.414614677429199, "learning_rate": 7.968500689502005e-06, "loss": 3.2163, "step": 16432 }, { "epoch": 0.5790498339455764, "grad_norm": 1.292343020439148, "learning_rate": 7.967383224244498e-06, "loss": 0.81, "step": 16433 }, { "epoch": 0.5790850709585349, "grad_norm": 23.131425857543945, "learning_rate": 7.966265785462085e-06, "loss": 5.0879, "step": 16434 }, { "epoch": 0.5791203079714933, "grad_norm": 1.0570695400238037, "learning_rate": 7.965148373169319e-06, "loss": 1.0527, "step": 16435 }, { "epoch": 0.5791555449844517, "grad_norm": 2.2456436157226562, "learning_rate": 7.964030987380749e-06, "loss": 1.0794, "step": 16436 }, { "epoch": 0.57919078199741, "grad_norm": 1.6439790725708008, "learning_rate": 7.96291362811094e-06, "loss": 0.7576, "step": 16437 }, { "epoch": 0.5792260190103685, "grad_norm": 6.939592361450195, "learning_rate": 7.96179629537444e-06, "loss": 5.7222, "step": 16438 }, { "epoch": 0.5792612560233269, "grad_norm": 5.842763423919678, "learning_rate": 7.960678989185802e-06, "loss": 5.1673, "step": 16439 }, { "epoch": 0.5792964930362853, "grad_norm": 3.767160654067993, "learning_rate": 7.959561709559579e-06, "loss": 2.9808, "step": 16440 }, { "epoch": 0.5793317300492438, "grad_norm": 2.3111729621887207, "learning_rate": 7.95844445651033e-06, "loss": 1.0591, "step": 16441 }, { "epoch": 0.5793669670622021, "grad_norm": 5.346735000610352, "learning_rate": 7.957327230052595e-06, "loss": 3.051, "step": 16442 }, { "epoch": 0.5794022040751605, "grad_norm": 0.6939183473587036, "learning_rate": 7.956210030200935e-06, "loss": 0.8922, "step": 16443 }, { "epoch": 0.579437441088119, "grad_norm": 5.576465606689453, "learning_rate": 7.955092856969897e-06, "loss": 3.7128, "step": 16444 }, { "epoch": 0.5794726781010774, "grad_norm": 19.71095848083496, "learning_rate": 7.953975710374043e-06, "loss": 5.2526, "step": 16445 }, { "epoch": 0.5795079151140358, "grad_norm": 1.311408281326294, "learning_rate": 7.95285859042791e-06, "loss": 1.225, "step": 16446 }, { "epoch": 0.5795431521269941, "grad_norm": 10.283531188964844, "learning_rate": 7.951741497146055e-06, "loss": 5.2201, "step": 16447 }, { "epoch": 0.5795783891399526, "grad_norm": 0.9527390599250793, "learning_rate": 7.95062443054303e-06, "loss": 0.7562, "step": 16448 }, { "epoch": 0.579613626152911, "grad_norm": 5.783959865570068, "learning_rate": 7.949507390633382e-06, "loss": 2.8853, "step": 16449 }, { "epoch": 0.5796488631658694, "grad_norm": 4.361507415771484, "learning_rate": 7.94839037743166e-06, "loss": 2.824, "step": 16450 }, { "epoch": 0.5796841001788279, "grad_norm": 1.3107095956802368, "learning_rate": 7.947273390952416e-06, "loss": 1.0381, "step": 16451 }, { "epoch": 0.5797193371917863, "grad_norm": 2.0799171924591064, "learning_rate": 7.946156431210199e-06, "loss": 0.7631, "step": 16452 }, { "epoch": 0.5797545742047446, "grad_norm": 1.175710678100586, "learning_rate": 7.945039498219553e-06, "loss": 0.7857, "step": 16453 }, { "epoch": 0.5797898112177031, "grad_norm": 1.5868574380874634, "learning_rate": 7.943922591995034e-06, "loss": 0.8766, "step": 16454 }, { "epoch": 0.5798250482306615, "grad_norm": 4.503478527069092, "learning_rate": 7.94280571255118e-06, "loss": 3.4559, "step": 16455 }, { "epoch": 0.5798602852436199, "grad_norm": 3.0191798210144043, "learning_rate": 7.941688859902545e-06, "loss": 2.8008, "step": 16456 }, { "epoch": 0.5798955222565783, "grad_norm": 1.1848355531692505, "learning_rate": 7.940572034063677e-06, "loss": 1.1396, "step": 16457 }, { "epoch": 0.5799307592695367, "grad_norm": 4.712426662445068, "learning_rate": 7.939455235049118e-06, "loss": 2.6317, "step": 16458 }, { "epoch": 0.5799659962824951, "grad_norm": 2.3019979000091553, "learning_rate": 7.938338462873416e-06, "loss": 0.7554, "step": 16459 }, { "epoch": 0.5800012332954535, "grad_norm": 4.805258274078369, "learning_rate": 7.937221717551125e-06, "loss": 3.5226, "step": 16460 }, { "epoch": 0.580036470308412, "grad_norm": 1.1549514532089233, "learning_rate": 7.936104999096777e-06, "loss": 0.7975, "step": 16461 }, { "epoch": 0.5800717073213704, "grad_norm": 1.4491389989852905, "learning_rate": 7.934988307524925e-06, "loss": 0.7085, "step": 16462 }, { "epoch": 0.5801069443343287, "grad_norm": 1.108370065689087, "learning_rate": 7.933871642850113e-06, "loss": 0.7731, "step": 16463 }, { "epoch": 0.5801421813472872, "grad_norm": 1.4761786460876465, "learning_rate": 7.932755005086892e-06, "loss": 1.0458, "step": 16464 }, { "epoch": 0.5801774183602456, "grad_norm": 4.4573774337768555, "learning_rate": 7.931638394249794e-06, "loss": 3.1845, "step": 16465 }, { "epoch": 0.580212655373204, "grad_norm": 3.9367516040802, "learning_rate": 7.930521810353372e-06, "loss": 2.6776, "step": 16466 }, { "epoch": 0.5802478923861625, "grad_norm": 5.2920684814453125, "learning_rate": 7.929405253412167e-06, "loss": 3.1666, "step": 16467 }, { "epoch": 0.5802831293991209, "grad_norm": 0.9759703278541565, "learning_rate": 7.928288723440722e-06, "loss": 0.8548, "step": 16468 }, { "epoch": 0.5803183664120792, "grad_norm": 1.2157366275787354, "learning_rate": 7.927172220453577e-06, "loss": 1.0596, "step": 16469 }, { "epoch": 0.5803536034250376, "grad_norm": 6.074169635772705, "learning_rate": 7.926055744465284e-06, "loss": 3.452, "step": 16470 }, { "epoch": 0.5803888404379961, "grad_norm": 5.934717178344727, "learning_rate": 7.924939295490375e-06, "loss": 4.3511, "step": 16471 }, { "epoch": 0.5804240774509545, "grad_norm": 5.35943603515625, "learning_rate": 7.923822873543397e-06, "loss": 2.5569, "step": 16472 }, { "epoch": 0.5804593144639129, "grad_norm": 3.5422983169555664, "learning_rate": 7.922706478638894e-06, "loss": 2.2961, "step": 16473 }, { "epoch": 0.5804945514768713, "grad_norm": 3.8767738342285156, "learning_rate": 7.921590110791398e-06, "loss": 3.4217, "step": 16474 }, { "epoch": 0.5805297884898297, "grad_norm": 6.898200988769531, "learning_rate": 7.920473770015456e-06, "loss": 5.3009, "step": 16475 }, { "epoch": 0.5805650255027881, "grad_norm": 6.7267746925354, "learning_rate": 7.919357456325612e-06, "loss": 2.7919, "step": 16476 }, { "epoch": 0.5806002625157466, "grad_norm": 6.7524027824401855, "learning_rate": 7.9182411697364e-06, "loss": 4.5111, "step": 16477 }, { "epoch": 0.580635499528705, "grad_norm": 3.6492600440979004, "learning_rate": 7.917124910262359e-06, "loss": 3.0278, "step": 16478 }, { "epoch": 0.5806707365416633, "grad_norm": 5.369684219360352, "learning_rate": 7.916008677918039e-06, "loss": 3.0145, "step": 16479 }, { "epoch": 0.5807059735546217, "grad_norm": 1.0644599199295044, "learning_rate": 7.914892472717965e-06, "loss": 1.0366, "step": 16480 }, { "epoch": 0.5807412105675802, "grad_norm": 1.2738398313522339, "learning_rate": 7.913776294676684e-06, "loss": 0.9032, "step": 16481 }, { "epoch": 0.5807764475805386, "grad_norm": 18.048444747924805, "learning_rate": 7.91266014380873e-06, "loss": 4.9058, "step": 16482 }, { "epoch": 0.580811684593497, "grad_norm": 6.721794128417969, "learning_rate": 7.91154402012865e-06, "loss": 3.1351, "step": 16483 }, { "epoch": 0.5808469216064555, "grad_norm": 4.456277370452881, "learning_rate": 7.910427923650969e-06, "loss": 2.604, "step": 16484 }, { "epoch": 0.5808821586194138, "grad_norm": 4.022440433502197, "learning_rate": 7.909311854390232e-06, "loss": 2.836, "step": 16485 }, { "epoch": 0.5809173956323722, "grad_norm": 1.1696555614471436, "learning_rate": 7.908195812360976e-06, "loss": 1.0015, "step": 16486 }, { "epoch": 0.5809526326453307, "grad_norm": 3.5691938400268555, "learning_rate": 7.907079797577736e-06, "loss": 2.335, "step": 16487 }, { "epoch": 0.5809878696582891, "grad_norm": 8.5831937789917, "learning_rate": 7.905963810055044e-06, "loss": 3.1102, "step": 16488 }, { "epoch": 0.5810231066712475, "grad_norm": 2.8470442295074463, "learning_rate": 7.904847849807446e-06, "loss": 2.8529, "step": 16489 }, { "epoch": 0.5810583436842058, "grad_norm": 4.457712173461914, "learning_rate": 7.903731916849469e-06, "loss": 3.1476, "step": 16490 }, { "epoch": 0.5810935806971643, "grad_norm": 1.3365325927734375, "learning_rate": 7.90261601119565e-06, "loss": 1.0212, "step": 16491 }, { "epoch": 0.5811288177101227, "grad_norm": 2.3440299034118652, "learning_rate": 7.901500132860527e-06, "loss": 1.02, "step": 16492 }, { "epoch": 0.5811640547230811, "grad_norm": 5.492095470428467, "learning_rate": 7.900384281858629e-06, "loss": 5.7375, "step": 16493 }, { "epoch": 0.5811992917360396, "grad_norm": 2.0465445518493652, "learning_rate": 7.899268458204494e-06, "loss": 0.815, "step": 16494 }, { "epoch": 0.581234528748998, "grad_norm": 5.000392436981201, "learning_rate": 7.898152661912657e-06, "loss": 2.859, "step": 16495 }, { "epoch": 0.5812697657619563, "grad_norm": 1.425611972808838, "learning_rate": 7.897036892997648e-06, "loss": 0.7569, "step": 16496 }, { "epoch": 0.5813050027749148, "grad_norm": 2.8128998279571533, "learning_rate": 7.895921151473998e-06, "loss": 2.5581, "step": 16497 }, { "epoch": 0.5813402397878732, "grad_norm": 1.9797922372817993, "learning_rate": 7.894805437356248e-06, "loss": 0.7956, "step": 16498 }, { "epoch": 0.5813754768008316, "grad_norm": 1.0424890518188477, "learning_rate": 7.893689750658925e-06, "loss": 0.8167, "step": 16499 }, { "epoch": 0.5814107138137901, "grad_norm": 1.0238707065582275, "learning_rate": 7.89257409139656e-06, "loss": 0.9218, "step": 16500 }, { "epoch": 0.5814459508267484, "grad_norm": 5.811520099639893, "learning_rate": 7.891458459583683e-06, "loss": 3.0987, "step": 16501 }, { "epoch": 0.5814811878397068, "grad_norm": 1.548172116279602, "learning_rate": 7.890342855234835e-06, "loss": 0.8672, "step": 16502 }, { "epoch": 0.5815164248526652, "grad_norm": 1.0972129106521606, "learning_rate": 7.889227278364533e-06, "loss": 0.8839, "step": 16503 }, { "epoch": 0.5815516618656237, "grad_norm": 1.4641691446304321, "learning_rate": 7.888111728987317e-06, "loss": 1.1198, "step": 16504 }, { "epoch": 0.5815868988785821, "grad_norm": 1.1057233810424805, "learning_rate": 7.886996207117718e-06, "loss": 0.9716, "step": 16505 }, { "epoch": 0.5816221358915404, "grad_norm": 2.8743550777435303, "learning_rate": 7.88588071277026e-06, "loss": 2.2334, "step": 16506 }, { "epoch": 0.5816573729044989, "grad_norm": 4.798880100250244, "learning_rate": 7.884765245959471e-06, "loss": 3.2445, "step": 16507 }, { "epoch": 0.5816926099174573, "grad_norm": 1.6644032001495361, "learning_rate": 7.88364980669989e-06, "loss": 0.911, "step": 16508 }, { "epoch": 0.5817278469304157, "grad_norm": 1.0235443115234375, "learning_rate": 7.882534395006038e-06, "loss": 1.0524, "step": 16509 }, { "epoch": 0.5817630839433742, "grad_norm": 8.571882247924805, "learning_rate": 7.881419010892446e-06, "loss": 5.1389, "step": 16510 }, { "epoch": 0.5817983209563325, "grad_norm": 0.9976853132247925, "learning_rate": 7.880303654373643e-06, "loss": 0.8111, "step": 16511 }, { "epoch": 0.5818335579692909, "grad_norm": 1.4083467721939087, "learning_rate": 7.87918832546415e-06, "loss": 0.7407, "step": 16512 }, { "epoch": 0.5818687949822493, "grad_norm": 4.0157790184021, "learning_rate": 7.878073024178499e-06, "loss": 3.1849, "step": 16513 }, { "epoch": 0.5819040319952078, "grad_norm": 4.497529029846191, "learning_rate": 7.876957750531222e-06, "loss": 2.7373, "step": 16514 }, { "epoch": 0.5819392690081662, "grad_norm": 5.235144138336182, "learning_rate": 7.875842504536839e-06, "loss": 5.5989, "step": 16515 }, { "epoch": 0.5819745060211245, "grad_norm": 1.281416654586792, "learning_rate": 7.874727286209875e-06, "loss": 1.1942, "step": 16516 }, { "epoch": 0.582009743034083, "grad_norm": 3.8549251556396484, "learning_rate": 7.87361209556486e-06, "loss": 2.7004, "step": 16517 }, { "epoch": 0.5820449800470414, "grad_norm": 0.9165918827056885, "learning_rate": 7.872496932616321e-06, "loss": 0.7957, "step": 16518 }, { "epoch": 0.5820802170599998, "grad_norm": 1.698927640914917, "learning_rate": 7.871381797378778e-06, "loss": 1.1539, "step": 16519 }, { "epoch": 0.5821154540729583, "grad_norm": 1.8652300834655762, "learning_rate": 7.870266689866758e-06, "loss": 0.9881, "step": 16520 }, { "epoch": 0.5821506910859167, "grad_norm": 6.7617645263671875, "learning_rate": 7.869151610094791e-06, "loss": 3.255, "step": 16521 }, { "epoch": 0.582185928098875, "grad_norm": 6.693383693695068, "learning_rate": 7.868036558077388e-06, "loss": 2.8552, "step": 16522 }, { "epoch": 0.5822211651118334, "grad_norm": 2.29910945892334, "learning_rate": 7.866921533829084e-06, "loss": 2.2216, "step": 16523 }, { "epoch": 0.5822564021247919, "grad_norm": 7.318210601806641, "learning_rate": 7.8658065373644e-06, "loss": 2.7025, "step": 16524 }, { "epoch": 0.5822916391377503, "grad_norm": 1.4139317274093628, "learning_rate": 7.864691568697856e-06, "loss": 1.0262, "step": 16525 }, { "epoch": 0.5823268761507087, "grad_norm": 1.2179052829742432, "learning_rate": 7.863576627843975e-06, "loss": 1.1423, "step": 16526 }, { "epoch": 0.5823621131636671, "grad_norm": 0.8913506269454956, "learning_rate": 7.862461714817284e-06, "loss": 0.8822, "step": 16527 }, { "epoch": 0.5823973501766255, "grad_norm": 1.3263206481933594, "learning_rate": 7.8613468296323e-06, "loss": 0.7646, "step": 16528 }, { "epoch": 0.5824325871895839, "grad_norm": 1.7148188352584839, "learning_rate": 7.860231972303545e-06, "loss": 0.8995, "step": 16529 }, { "epoch": 0.5824678242025424, "grad_norm": 1.0570883750915527, "learning_rate": 7.859117142845543e-06, "loss": 0.9988, "step": 16530 }, { "epoch": 0.5825030612155008, "grad_norm": 2.013087511062622, "learning_rate": 7.858002341272809e-06, "loss": 0.9445, "step": 16531 }, { "epoch": 0.5825382982284592, "grad_norm": 3.860774278640747, "learning_rate": 7.85688756759987e-06, "loss": 2.2944, "step": 16532 }, { "epoch": 0.5825735352414176, "grad_norm": 4.125715732574463, "learning_rate": 7.855772821841246e-06, "loss": 3.1048, "step": 16533 }, { "epoch": 0.582608772254376, "grad_norm": 4.103662014007568, "learning_rate": 7.85465810401145e-06, "loss": 0.9025, "step": 16534 }, { "epoch": 0.5826440092673344, "grad_norm": 6.733957290649414, "learning_rate": 7.853543414125004e-06, "loss": 3.6579, "step": 16535 }, { "epoch": 0.5826792462802928, "grad_norm": 1.2436244487762451, "learning_rate": 7.852428752196433e-06, "loss": 0.8762, "step": 16536 }, { "epoch": 0.5827144832932513, "grad_norm": 4.781322002410889, "learning_rate": 7.85131411824025e-06, "loss": 3.3605, "step": 16537 }, { "epoch": 0.5827497203062096, "grad_norm": 3.986013889312744, "learning_rate": 7.850199512270973e-06, "loss": 2.7151, "step": 16538 }, { "epoch": 0.582784957319168, "grad_norm": 7.024099349975586, "learning_rate": 7.84908493430312e-06, "loss": 4.9465, "step": 16539 }, { "epoch": 0.5828201943321265, "grad_norm": 1.1739253997802734, "learning_rate": 7.847970384351217e-06, "loss": 0.9269, "step": 16540 }, { "epoch": 0.5828554313450849, "grad_norm": 34.30646514892578, "learning_rate": 7.846855862429765e-06, "loss": 3.0924, "step": 16541 }, { "epoch": 0.5828906683580433, "grad_norm": 4.198955059051514, "learning_rate": 7.845741368553294e-06, "loss": 3.0683, "step": 16542 }, { "epoch": 0.5829259053710018, "grad_norm": 4.250304222106934, "learning_rate": 7.844626902736316e-06, "loss": 2.9113, "step": 16543 }, { "epoch": 0.5829611423839601, "grad_norm": 1.574175238609314, "learning_rate": 7.843512464993348e-06, "loss": 0.7601, "step": 16544 }, { "epoch": 0.5829963793969185, "grad_norm": 3.417234182357788, "learning_rate": 7.8423980553389e-06, "loss": 2.8659, "step": 16545 }, { "epoch": 0.5830316164098769, "grad_norm": 6.544731140136719, "learning_rate": 7.841283673787497e-06, "loss": 5.0124, "step": 16546 }, { "epoch": 0.5830668534228354, "grad_norm": 3.637608766555786, "learning_rate": 7.840169320353648e-06, "loss": 2.6671, "step": 16547 }, { "epoch": 0.5831020904357938, "grad_norm": 4.587361812591553, "learning_rate": 7.839054995051868e-06, "loss": 2.6014, "step": 16548 }, { "epoch": 0.5831373274487521, "grad_norm": 7.558318138122559, "learning_rate": 7.837940697896674e-06, "loss": 5.6041, "step": 16549 }, { "epoch": 0.5831725644617106, "grad_norm": 5.6214447021484375, "learning_rate": 7.836826428902574e-06, "loss": 2.3786, "step": 16550 }, { "epoch": 0.583207801474669, "grad_norm": 1.7105841636657715, "learning_rate": 7.835712188084087e-06, "loss": 0.8267, "step": 16551 }, { "epoch": 0.5832430384876274, "grad_norm": 3.1765687465667725, "learning_rate": 7.834597975455727e-06, "loss": 2.2418, "step": 16552 }, { "epoch": 0.5832782755005859, "grad_norm": 8.933304786682129, "learning_rate": 7.833483791032002e-06, "loss": 3.2048, "step": 16553 }, { "epoch": 0.5833135125135442, "grad_norm": 1.9943076372146606, "learning_rate": 7.832369634827423e-06, "loss": 0.894, "step": 16554 }, { "epoch": 0.5833487495265026, "grad_norm": 6.891759872436523, "learning_rate": 7.831255506856509e-06, "loss": 3.4577, "step": 16555 }, { "epoch": 0.583383986539461, "grad_norm": 1.1528881788253784, "learning_rate": 7.83014140713377e-06, "loss": 1.1248, "step": 16556 }, { "epoch": 0.5834192235524195, "grad_norm": 1.2611043453216553, "learning_rate": 7.829027335673715e-06, "loss": 0.8077, "step": 16557 }, { "epoch": 0.5834544605653779, "grad_norm": 1.610955834388733, "learning_rate": 7.827913292490852e-06, "loss": 0.7813, "step": 16558 }, { "epoch": 0.5834896975783362, "grad_norm": 8.952445030212402, "learning_rate": 7.826799277599702e-06, "loss": 2.7918, "step": 16559 }, { "epoch": 0.5835249345912947, "grad_norm": 2.644211769104004, "learning_rate": 7.82568529101476e-06, "loss": 2.7577, "step": 16560 }, { "epoch": 0.5835601716042531, "grad_norm": 1.4401203393936157, "learning_rate": 7.824571332750548e-06, "loss": 1.0728, "step": 16561 }, { "epoch": 0.5835954086172115, "grad_norm": 5.879584312438965, "learning_rate": 7.823457402821575e-06, "loss": 2.8295, "step": 16562 }, { "epoch": 0.58363064563017, "grad_norm": 2.3482956886291504, "learning_rate": 7.822343501242342e-06, "loss": 2.9011, "step": 16563 }, { "epoch": 0.5836658826431284, "grad_norm": 1.666153907775879, "learning_rate": 7.821229628027361e-06, "loss": 0.8378, "step": 16564 }, { "epoch": 0.5837011196560867, "grad_norm": 4.545166969299316, "learning_rate": 7.820115783191145e-06, "loss": 3.1639, "step": 16565 }, { "epoch": 0.5837363566690452, "grad_norm": 1.2675762176513672, "learning_rate": 7.819001966748197e-06, "loss": 0.7958, "step": 16566 }, { "epoch": 0.5837715936820036, "grad_norm": 1.29167902469635, "learning_rate": 7.817888178713028e-06, "loss": 0.8684, "step": 16567 }, { "epoch": 0.583806830694962, "grad_norm": 2.70094895362854, "learning_rate": 7.816774419100142e-06, "loss": 0.844, "step": 16568 }, { "epoch": 0.5838420677079204, "grad_norm": 4.075169086456299, "learning_rate": 7.815660687924045e-06, "loss": 5.3576, "step": 16569 }, { "epoch": 0.5838773047208788, "grad_norm": 1.5760934352874756, "learning_rate": 7.814546985199248e-06, "loss": 0.6486, "step": 16570 }, { "epoch": 0.5839125417338372, "grad_norm": 3.4816861152648926, "learning_rate": 7.813433310940254e-06, "loss": 3.1204, "step": 16571 }, { "epoch": 0.5839477787467956, "grad_norm": 4.895631790161133, "learning_rate": 7.812319665161573e-06, "loss": 2.5378, "step": 16572 }, { "epoch": 0.5839830157597541, "grad_norm": 11.456982612609863, "learning_rate": 7.811206047877706e-06, "loss": 7.794, "step": 16573 }, { "epoch": 0.5840182527727125, "grad_norm": 3.6256814002990723, "learning_rate": 7.810092459103155e-06, "loss": 2.9787, "step": 16574 }, { "epoch": 0.5840534897856708, "grad_norm": 1.566689372062683, "learning_rate": 7.808978898852433e-06, "loss": 0.7931, "step": 16575 }, { "epoch": 0.5840887267986293, "grad_norm": 2.8448991775512695, "learning_rate": 7.807865367140037e-06, "loss": 2.7458, "step": 16576 }, { "epoch": 0.5841239638115877, "grad_norm": 1.2014261484146118, "learning_rate": 7.806751863980476e-06, "loss": 1.0795, "step": 16577 }, { "epoch": 0.5841592008245461, "grad_norm": 27.130382537841797, "learning_rate": 7.805638389388253e-06, "loss": 5.2613, "step": 16578 }, { "epoch": 0.5841944378375045, "grad_norm": 5.741908550262451, "learning_rate": 7.804524943377863e-06, "loss": 3.1712, "step": 16579 }, { "epoch": 0.584229674850463, "grad_norm": 5.634159088134766, "learning_rate": 7.80341152596382e-06, "loss": 3.3685, "step": 16580 }, { "epoch": 0.5842649118634213, "grad_norm": 1.1008434295654297, "learning_rate": 7.802298137160624e-06, "loss": 0.7765, "step": 16581 }, { "epoch": 0.5843001488763797, "grad_norm": 2.9455816745758057, "learning_rate": 7.80118477698277e-06, "loss": 3.3734, "step": 16582 }, { "epoch": 0.5843353858893382, "grad_norm": 4.916268348693848, "learning_rate": 7.800071445444763e-06, "loss": 3.1752, "step": 16583 }, { "epoch": 0.5843706229022966, "grad_norm": 4.962145805358887, "learning_rate": 7.798958142561115e-06, "loss": 0.943, "step": 16584 }, { "epoch": 0.584405859915255, "grad_norm": 0.9793183207511902, "learning_rate": 7.797844868346308e-06, "loss": 0.8114, "step": 16585 }, { "epoch": 0.5844410969282134, "grad_norm": 4.115386486053467, "learning_rate": 7.796731622814855e-06, "loss": 3.2989, "step": 16586 }, { "epoch": 0.5844763339411718, "grad_norm": 3.644373893737793, "learning_rate": 7.795618405981256e-06, "loss": 2.6888, "step": 16587 }, { "epoch": 0.5845115709541302, "grad_norm": 4.004243850708008, "learning_rate": 7.794505217860005e-06, "loss": 2.5739, "step": 16588 }, { "epoch": 0.5845468079670886, "grad_norm": 10.912145614624023, "learning_rate": 7.793392058465603e-06, "loss": 3.305, "step": 16589 }, { "epoch": 0.5845820449800471, "grad_norm": 9.709250450134277, "learning_rate": 7.792278927812551e-06, "loss": 2.8521, "step": 16590 }, { "epoch": 0.5846172819930054, "grad_norm": 1.0529955625534058, "learning_rate": 7.791165825915352e-06, "loss": 0.8345, "step": 16591 }, { "epoch": 0.5846525190059638, "grad_norm": 11.61038875579834, "learning_rate": 7.790052752788494e-06, "loss": 7.0933, "step": 16592 }, { "epoch": 0.5846877560189223, "grad_norm": 5.941863536834717, "learning_rate": 7.788939708446483e-06, "loss": 2.5525, "step": 16593 }, { "epoch": 0.5847229930318807, "grad_norm": 6.4375810623168945, "learning_rate": 7.787826692903815e-06, "loss": 3.3428, "step": 16594 }, { "epoch": 0.5847582300448391, "grad_norm": 2.5007402896881104, "learning_rate": 7.786713706174985e-06, "loss": 0.9564, "step": 16595 }, { "epoch": 0.5847934670577976, "grad_norm": 0.9950478672981262, "learning_rate": 7.785600748274488e-06, "loss": 0.9716, "step": 16596 }, { "epoch": 0.5848287040707559, "grad_norm": 4.452295303344727, "learning_rate": 7.78448781921683e-06, "loss": 3.0036, "step": 16597 }, { "epoch": 0.5848639410837143, "grad_norm": 8.304767608642578, "learning_rate": 7.783374919016493e-06, "loss": 5.3967, "step": 16598 }, { "epoch": 0.5848991780966728, "grad_norm": 3.615096092224121, "learning_rate": 7.782262047687985e-06, "loss": 3.5577, "step": 16599 }, { "epoch": 0.5849344151096312, "grad_norm": 2.800762414932251, "learning_rate": 7.781149205245795e-06, "loss": 0.7646, "step": 16600 }, { "epoch": 0.5849696521225896, "grad_norm": 3.1389265060424805, "learning_rate": 7.780036391704418e-06, "loss": 3.1316, "step": 16601 }, { "epoch": 0.5850048891355479, "grad_norm": 2.403414011001587, "learning_rate": 7.77892360707835e-06, "loss": 0.8327, "step": 16602 }, { "epoch": 0.5850401261485064, "grad_norm": 6.567077159881592, "learning_rate": 7.77781085138209e-06, "loss": 3.2727, "step": 16603 }, { "epoch": 0.5850753631614648, "grad_norm": 6.62812614440918, "learning_rate": 7.776698124630118e-06, "loss": 7.8686, "step": 16604 }, { "epoch": 0.5851106001744232, "grad_norm": 3.877864360809326, "learning_rate": 7.77558542683694e-06, "loss": 0.9309, "step": 16605 }, { "epoch": 0.5851458371873817, "grad_norm": 2.699758291244507, "learning_rate": 7.774472758017049e-06, "loss": 2.8214, "step": 16606 }, { "epoch": 0.58518107420034, "grad_norm": 8.783312797546387, "learning_rate": 7.773360118184931e-06, "loss": 3.0563, "step": 16607 }, { "epoch": 0.5852163112132984, "grad_norm": 3.172616720199585, "learning_rate": 7.772247507355077e-06, "loss": 2.9998, "step": 16608 }, { "epoch": 0.5852515482262569, "grad_norm": 1.0871238708496094, "learning_rate": 7.771134925541987e-06, "loss": 1.0818, "step": 16609 }, { "epoch": 0.5852867852392153, "grad_norm": 1.3228989839553833, "learning_rate": 7.77002237276015e-06, "loss": 1.017, "step": 16610 }, { "epoch": 0.5853220222521737, "grad_norm": 1.1692776679992676, "learning_rate": 7.768909849024053e-06, "loss": 0.9525, "step": 16611 }, { "epoch": 0.585357259265132, "grad_norm": 5.611279010772705, "learning_rate": 7.767797354348187e-06, "loss": 0.7422, "step": 16612 }, { "epoch": 0.5853924962780905, "grad_norm": 0.9697017073631287, "learning_rate": 7.76668488874705e-06, "loss": 0.8281, "step": 16613 }, { "epoch": 0.5854277332910489, "grad_norm": 5.741782188415527, "learning_rate": 7.765572452235125e-06, "loss": 3.0007, "step": 16614 }, { "epoch": 0.5854629703040073, "grad_norm": 1.7900292873382568, "learning_rate": 7.764460044826903e-06, "loss": 0.8469, "step": 16615 }, { "epoch": 0.5854982073169658, "grad_norm": 3.505002021789551, "learning_rate": 7.763347666536876e-06, "loss": 2.951, "step": 16616 }, { "epoch": 0.5855334443299242, "grad_norm": 1.1057109832763672, "learning_rate": 7.762235317379526e-06, "loss": 0.9827, "step": 16617 }, { "epoch": 0.5855686813428825, "grad_norm": 5.2097578048706055, "learning_rate": 7.761122997369348e-06, "loss": 3.1821, "step": 16618 }, { "epoch": 0.585603918355841, "grad_norm": 1.3424984216690063, "learning_rate": 7.760010706520831e-06, "loss": 1.1096, "step": 16619 }, { "epoch": 0.5856391553687994, "grad_norm": 1.3161402940750122, "learning_rate": 7.758898444848457e-06, "loss": 1.0461, "step": 16620 }, { "epoch": 0.5856743923817578, "grad_norm": 1.6468969583511353, "learning_rate": 7.757786212366714e-06, "loss": 0.7976, "step": 16621 }, { "epoch": 0.5857096293947162, "grad_norm": 1.1856390237808228, "learning_rate": 7.756674009090099e-06, "loss": 0.9283, "step": 16622 }, { "epoch": 0.5857448664076746, "grad_norm": 2.450660467147827, "learning_rate": 7.755561835033083e-06, "loss": 0.988, "step": 16623 }, { "epoch": 0.585780103420633, "grad_norm": 2.1019670963287354, "learning_rate": 7.754449690210161e-06, "loss": 1.0289, "step": 16624 }, { "epoch": 0.5858153404335914, "grad_norm": 5.337421417236328, "learning_rate": 7.753337574635822e-06, "loss": 3.1222, "step": 16625 }, { "epoch": 0.5858505774465499, "grad_norm": 4.257758617401123, "learning_rate": 7.752225488324545e-06, "loss": 0.7951, "step": 16626 }, { "epoch": 0.5858858144595083, "grad_norm": 1.9546430110931396, "learning_rate": 7.751113431290815e-06, "loss": 0.945, "step": 16627 }, { "epoch": 0.5859210514724666, "grad_norm": 2.406399726867676, "learning_rate": 7.75000140354912e-06, "loss": 0.7729, "step": 16628 }, { "epoch": 0.5859562884854251, "grad_norm": 0.8021903038024902, "learning_rate": 7.748889405113945e-06, "loss": 0.7454, "step": 16629 }, { "epoch": 0.5859915254983835, "grad_norm": 3.281860113143921, "learning_rate": 7.747777435999772e-06, "loss": 2.327, "step": 16630 }, { "epoch": 0.5860267625113419, "grad_norm": 2.1756296157836914, "learning_rate": 7.746665496221083e-06, "loss": 0.8761, "step": 16631 }, { "epoch": 0.5860619995243003, "grad_norm": 1.1791369915008545, "learning_rate": 7.745553585792368e-06, "loss": 1.5653, "step": 16632 }, { "epoch": 0.5860972365372588, "grad_norm": 1.3189550638198853, "learning_rate": 7.744441704728099e-06, "loss": 0.6606, "step": 16633 }, { "epoch": 0.5861324735502171, "grad_norm": 3.495161771774292, "learning_rate": 7.743329853042767e-06, "loss": 0.7101, "step": 16634 }, { "epoch": 0.5861677105631755, "grad_norm": 3.143338441848755, "learning_rate": 7.742218030750853e-06, "loss": 2.6453, "step": 16635 }, { "epoch": 0.586202947576134, "grad_norm": 4.433863162994385, "learning_rate": 7.74110623786683e-06, "loss": 3.255, "step": 16636 }, { "epoch": 0.5862381845890924, "grad_norm": 1.221314549446106, "learning_rate": 7.73999447440519e-06, "loss": 0.7527, "step": 16637 }, { "epoch": 0.5862734216020508, "grad_norm": 5.420474052429199, "learning_rate": 7.738882740380412e-06, "loss": 4.88, "step": 16638 }, { "epoch": 0.5863086586150092, "grad_norm": 4.29414701461792, "learning_rate": 7.737771035806972e-06, "loss": 3.1581, "step": 16639 }, { "epoch": 0.5863438956279676, "grad_norm": 1.6323308944702148, "learning_rate": 7.73665936069935e-06, "loss": 0.5972, "step": 16640 }, { "epoch": 0.586379132640926, "grad_norm": 1.0018070936203003, "learning_rate": 7.735547715072035e-06, "loss": 1.1523, "step": 16641 }, { "epoch": 0.5864143696538845, "grad_norm": 0.9170602560043335, "learning_rate": 7.734436098939493e-06, "loss": 1.0241, "step": 16642 }, { "epoch": 0.5864496066668429, "grad_norm": 7.529542922973633, "learning_rate": 7.733324512316211e-06, "loss": 5.7448, "step": 16643 }, { "epoch": 0.5864848436798012, "grad_norm": 1.5347439050674438, "learning_rate": 7.732212955216664e-06, "loss": 0.9029, "step": 16644 }, { "epoch": 0.5865200806927596, "grad_norm": 1.1158190965652466, "learning_rate": 7.73110142765534e-06, "loss": 0.9383, "step": 16645 }, { "epoch": 0.5865553177057181, "grad_norm": 27.9220027923584, "learning_rate": 7.7299899296467e-06, "loss": 2.9502, "step": 16646 }, { "epoch": 0.5865905547186765, "grad_norm": 4.757026672363281, "learning_rate": 7.728878461205235e-06, "loss": 3.2414, "step": 16647 }, { "epoch": 0.5866257917316349, "grad_norm": 42.14824295043945, "learning_rate": 7.727767022345417e-06, "loss": 4.9952, "step": 16648 }, { "epoch": 0.5866610287445934, "grad_norm": 1.5768810510635376, "learning_rate": 7.726655613081722e-06, "loss": 0.8507, "step": 16649 }, { "epoch": 0.5866962657575517, "grad_norm": 1.1745933294296265, "learning_rate": 7.725544233428626e-06, "loss": 1.0533, "step": 16650 }, { "epoch": 0.5867315027705101, "grad_norm": 1.5498355627059937, "learning_rate": 7.72443288340061e-06, "loss": 0.8154, "step": 16651 }, { "epoch": 0.5867667397834686, "grad_norm": 5.933260917663574, "learning_rate": 7.723321563012143e-06, "loss": 3.3852, "step": 16652 }, { "epoch": 0.586801976796427, "grad_norm": 1.030977487564087, "learning_rate": 7.722210272277702e-06, "loss": 0.8525, "step": 16653 }, { "epoch": 0.5868372138093854, "grad_norm": 1.9547759294509888, "learning_rate": 7.721099011211765e-06, "loss": 0.9138, "step": 16654 }, { "epoch": 0.5868724508223437, "grad_norm": 2.5985562801361084, "learning_rate": 7.7199877798288e-06, "loss": 3.0412, "step": 16655 }, { "epoch": 0.5869076878353022, "grad_norm": 1.5583839416503906, "learning_rate": 7.718876578143286e-06, "loss": 0.8521, "step": 16656 }, { "epoch": 0.5869429248482606, "grad_norm": 8.12394905090332, "learning_rate": 7.717765406169696e-06, "loss": 2.8352, "step": 16657 }, { "epoch": 0.586978161861219, "grad_norm": 2.608009099960327, "learning_rate": 7.716654263922503e-06, "loss": 2.7174, "step": 16658 }, { "epoch": 0.5870133988741775, "grad_norm": 4.100790977478027, "learning_rate": 7.715543151416176e-06, "loss": 3.2814, "step": 16659 }, { "epoch": 0.5870486358871359, "grad_norm": 1.3843443393707275, "learning_rate": 7.714432068665195e-06, "loss": 0.7725, "step": 16660 }, { "epoch": 0.5870838729000942, "grad_norm": 4.444530963897705, "learning_rate": 7.713321015684023e-06, "loss": 3.4003, "step": 16661 }, { "epoch": 0.5871191099130527, "grad_norm": 4.748152732849121, "learning_rate": 7.712209992487137e-06, "loss": 2.9267, "step": 16662 }, { "epoch": 0.5871543469260111, "grad_norm": 4.813482284545898, "learning_rate": 7.711098999089006e-06, "loss": 3.4227, "step": 16663 }, { "epoch": 0.5871895839389695, "grad_norm": 1.0761672258377075, "learning_rate": 7.709988035504107e-06, "loss": 1.0834, "step": 16664 }, { "epoch": 0.5872248209519279, "grad_norm": 10.0958833694458, "learning_rate": 7.708877101746902e-06, "loss": 5.6503, "step": 16665 }, { "epoch": 0.5872600579648863, "grad_norm": 6.419818878173828, "learning_rate": 7.707766197831865e-06, "loss": 5.7542, "step": 16666 }, { "epoch": 0.5872952949778447, "grad_norm": 1.0226091146469116, "learning_rate": 7.706655323773467e-06, "loss": 0.8311, "step": 16667 }, { "epoch": 0.5873305319908031, "grad_norm": 6.014123916625977, "learning_rate": 7.705544479586175e-06, "loss": 2.9215, "step": 16668 }, { "epoch": 0.5873657690037616, "grad_norm": 1.5217822790145874, "learning_rate": 7.704433665284455e-06, "loss": 0.9885, "step": 16669 }, { "epoch": 0.58740100601672, "grad_norm": 4.381580829620361, "learning_rate": 7.703322880882782e-06, "loss": 3.9326, "step": 16670 }, { "epoch": 0.5874362430296783, "grad_norm": 5.359240531921387, "learning_rate": 7.70221212639562e-06, "loss": 5.8058, "step": 16671 }, { "epoch": 0.5874714800426368, "grad_norm": 4.325717449188232, "learning_rate": 7.701101401837439e-06, "loss": 2.6354, "step": 16672 }, { "epoch": 0.5875067170555952, "grad_norm": 3.3752644062042236, "learning_rate": 7.699990707222707e-06, "loss": 3.6454, "step": 16673 }, { "epoch": 0.5875419540685536, "grad_norm": 5.171133518218994, "learning_rate": 7.698880042565883e-06, "loss": 3.1566, "step": 16674 }, { "epoch": 0.5875771910815121, "grad_norm": 3.7386348247528076, "learning_rate": 7.697769407881443e-06, "loss": 3.0672, "step": 16675 }, { "epoch": 0.5876124280944705, "grad_norm": 2.5634164810180664, "learning_rate": 7.696658803183851e-06, "loss": 2.6079, "step": 16676 }, { "epoch": 0.5876476651074288, "grad_norm": 1.1701858043670654, "learning_rate": 7.69554822848757e-06, "loss": 0.811, "step": 16677 }, { "epoch": 0.5876829021203872, "grad_norm": 5.680184364318848, "learning_rate": 7.694437683807064e-06, "loss": 5.7016, "step": 16678 }, { "epoch": 0.5877181391333457, "grad_norm": 1.2078325748443604, "learning_rate": 7.693327169156808e-06, "loss": 1.0809, "step": 16679 }, { "epoch": 0.5877533761463041, "grad_norm": 1.5212479829788208, "learning_rate": 7.692216684551253e-06, "loss": 0.9172, "step": 16680 }, { "epoch": 0.5877886131592625, "grad_norm": 4.4858479499816895, "learning_rate": 7.691106230004873e-06, "loss": 2.523, "step": 16681 }, { "epoch": 0.5878238501722209, "grad_norm": 2.711838960647583, "learning_rate": 7.689995805532125e-06, "loss": 3.1556, "step": 16682 }, { "epoch": 0.5878590871851793, "grad_norm": 4.434664726257324, "learning_rate": 7.688885411147481e-06, "loss": 3.1636, "step": 16683 }, { "epoch": 0.5878943241981377, "grad_norm": 7.004672050476074, "learning_rate": 7.687775046865394e-06, "loss": 3.4396, "step": 16684 }, { "epoch": 0.5879295612110962, "grad_norm": 1.0725157260894775, "learning_rate": 7.686664712700334e-06, "loss": 0.8799, "step": 16685 }, { "epoch": 0.5879647982240546, "grad_norm": 1.2097454071044922, "learning_rate": 7.685554408666762e-06, "loss": 1.0166, "step": 16686 }, { "epoch": 0.5880000352370129, "grad_norm": 3.2538974285125732, "learning_rate": 7.684444134779136e-06, "loss": 2.6232, "step": 16687 }, { "epoch": 0.5880352722499713, "grad_norm": 3.240600824356079, "learning_rate": 7.68333389105192e-06, "loss": 3.1157, "step": 16688 }, { "epoch": 0.5880705092629298, "grad_norm": 6.0344462394714355, "learning_rate": 7.682223677499578e-06, "loss": 5.2652, "step": 16689 }, { "epoch": 0.5881057462758882, "grad_norm": 0.9571640491485596, "learning_rate": 7.681113494136565e-06, "loss": 1.0724, "step": 16690 }, { "epoch": 0.5881409832888466, "grad_norm": 10.740875244140625, "learning_rate": 7.680003340977346e-06, "loss": 2.5503, "step": 16691 }, { "epoch": 0.588176220301805, "grad_norm": 4.89198112487793, "learning_rate": 7.67889321803638e-06, "loss": 2.6804, "step": 16692 }, { "epoch": 0.5882114573147634, "grad_norm": 4.937920093536377, "learning_rate": 7.677783125328123e-06, "loss": 5.5775, "step": 16693 }, { "epoch": 0.5882466943277218, "grad_norm": 1.284355878829956, "learning_rate": 7.676673062867036e-06, "loss": 1.0327, "step": 16694 }, { "epoch": 0.5882819313406803, "grad_norm": 7.222254276275635, "learning_rate": 7.67556303066758e-06, "loss": 5.4928, "step": 16695 }, { "epoch": 0.5883171683536387, "grad_norm": 7.959338665008545, "learning_rate": 7.674453028744213e-06, "loss": 2.9106, "step": 16696 }, { "epoch": 0.588352405366597, "grad_norm": 4.681266784667969, "learning_rate": 7.673343057111387e-06, "loss": 5.1854, "step": 16697 }, { "epoch": 0.5883876423795554, "grad_norm": 1.2921019792556763, "learning_rate": 7.67223311578357e-06, "loss": 1.088, "step": 16698 }, { "epoch": 0.5884228793925139, "grad_norm": 2.6428937911987305, "learning_rate": 7.671123204775209e-06, "loss": 0.9175, "step": 16699 }, { "epoch": 0.5884581164054723, "grad_norm": 3.624211549758911, "learning_rate": 7.670013324100765e-06, "loss": 2.7655, "step": 16700 }, { "epoch": 0.5884933534184307, "grad_norm": 4.370492458343506, "learning_rate": 7.668903473774694e-06, "loss": 3.1398, "step": 16701 }, { "epoch": 0.5885285904313892, "grad_norm": 1.8235780000686646, "learning_rate": 7.667793653811458e-06, "loss": 2.3231, "step": 16702 }, { "epoch": 0.5885638274443475, "grad_norm": 14.269579887390137, "learning_rate": 7.666683864225498e-06, "loss": 2.6925, "step": 16703 }, { "epoch": 0.5885990644573059, "grad_norm": 5.334963321685791, "learning_rate": 7.665574105031283e-06, "loss": 3.295, "step": 16704 }, { "epoch": 0.5886343014702644, "grad_norm": 12.648612022399902, "learning_rate": 7.664464376243264e-06, "loss": 4.8898, "step": 16705 }, { "epoch": 0.5886695384832228, "grad_norm": 4.468443393707275, "learning_rate": 7.66335467787589e-06, "loss": 3.0125, "step": 16706 }, { "epoch": 0.5887047754961812, "grad_norm": 1.3405195474624634, "learning_rate": 7.662245009943619e-06, "loss": 1.1404, "step": 16707 }, { "epoch": 0.5887400125091397, "grad_norm": 1.8430155515670776, "learning_rate": 7.661135372460908e-06, "loss": 0.9049, "step": 16708 }, { "epoch": 0.588775249522098, "grad_norm": 7.785821914672852, "learning_rate": 7.660025765442205e-06, "loss": 7.3508, "step": 16709 }, { "epoch": 0.5888104865350564, "grad_norm": 6.793999671936035, "learning_rate": 7.658916188901963e-06, "loss": 2.8677, "step": 16710 }, { "epoch": 0.5888457235480148, "grad_norm": 1.217528223991394, "learning_rate": 7.65780664285464e-06, "loss": 0.7767, "step": 16711 }, { "epoch": 0.5888809605609733, "grad_norm": 1.0441131591796875, "learning_rate": 7.656697127314679e-06, "loss": 0.7497, "step": 16712 }, { "epoch": 0.5889161975739317, "grad_norm": 3.204925060272217, "learning_rate": 7.655587642296538e-06, "loss": 3.0575, "step": 16713 }, { "epoch": 0.58895143458689, "grad_norm": 1.6555516719818115, "learning_rate": 7.65447818781467e-06, "loss": 0.8378, "step": 16714 }, { "epoch": 0.5889866715998485, "grad_norm": 1.3582247495651245, "learning_rate": 7.653368763883517e-06, "loss": 0.9568, "step": 16715 }, { "epoch": 0.5890219086128069, "grad_norm": 5.450733661651611, "learning_rate": 7.652259370517535e-06, "loss": 2.7876, "step": 16716 }, { "epoch": 0.5890571456257653, "grad_norm": 2.8472390174865723, "learning_rate": 7.651150007731177e-06, "loss": 2.9761, "step": 16717 }, { "epoch": 0.5890923826387238, "grad_norm": 3.740736246109009, "learning_rate": 7.650040675538888e-06, "loss": 3.0918, "step": 16718 }, { "epoch": 0.5891276196516821, "grad_norm": 1.1509618759155273, "learning_rate": 7.64893137395512e-06, "loss": 0.9108, "step": 16719 }, { "epoch": 0.5891628566646405, "grad_norm": 3.3626863956451416, "learning_rate": 7.647822102994319e-06, "loss": 2.798, "step": 16720 }, { "epoch": 0.5891980936775989, "grad_norm": 3.992471694946289, "learning_rate": 7.646712862670936e-06, "loss": 2.7115, "step": 16721 }, { "epoch": 0.5892333306905574, "grad_norm": 4.7910614013671875, "learning_rate": 7.645603652999415e-06, "loss": 3.4656, "step": 16722 }, { "epoch": 0.5892685677035158, "grad_norm": 1.1999948024749756, "learning_rate": 7.644494473994207e-06, "loss": 0.8941, "step": 16723 }, { "epoch": 0.5893038047164741, "grad_norm": 1.416689157485962, "learning_rate": 7.643385325669761e-06, "loss": 0.8203, "step": 16724 }, { "epoch": 0.5893390417294326, "grad_norm": 5.664568901062012, "learning_rate": 7.64227620804052e-06, "loss": 4.7437, "step": 16725 }, { "epoch": 0.589374278742391, "grad_norm": 4.330299377441406, "learning_rate": 7.64116712112093e-06, "loss": 3.3347, "step": 16726 }, { "epoch": 0.5894095157553494, "grad_norm": 0.905672550201416, "learning_rate": 7.640058064925441e-06, "loss": 1.0413, "step": 16727 }, { "epoch": 0.5894447527683079, "grad_norm": 1.1436994075775146, "learning_rate": 7.638949039468494e-06, "loss": 0.9333, "step": 16728 }, { "epoch": 0.5894799897812663, "grad_norm": 3.7524685859680176, "learning_rate": 7.63784004476454e-06, "loss": 3.279, "step": 16729 }, { "epoch": 0.5895152267942246, "grad_norm": 1.9074865579605103, "learning_rate": 7.63673108082802e-06, "loss": 0.774, "step": 16730 }, { "epoch": 0.589550463807183, "grad_norm": 1.2222981452941895, "learning_rate": 7.635622147673376e-06, "loss": 0.9981, "step": 16731 }, { "epoch": 0.5895857008201415, "grad_norm": 2.271073818206787, "learning_rate": 7.634513245315053e-06, "loss": 0.8509, "step": 16732 }, { "epoch": 0.5896209378330999, "grad_norm": 5.271721839904785, "learning_rate": 7.633404373767501e-06, "loss": 5.3456, "step": 16733 }, { "epoch": 0.5896561748460583, "grad_norm": 5.969484806060791, "learning_rate": 7.632295533045157e-06, "loss": 4.3366, "step": 16734 }, { "epoch": 0.5896914118590167, "grad_norm": 5.85452127456665, "learning_rate": 7.631186723162465e-06, "loss": 7.391, "step": 16735 }, { "epoch": 0.5897266488719751, "grad_norm": 5.121200084686279, "learning_rate": 7.630077944133864e-06, "loss": 4.9179, "step": 16736 }, { "epoch": 0.5897618858849335, "grad_norm": 4.743022441864014, "learning_rate": 7.628969195973805e-06, "loss": 2.6914, "step": 16737 }, { "epoch": 0.589797122897892, "grad_norm": 3.0764098167419434, "learning_rate": 7.627860478696722e-06, "loss": 2.467, "step": 16738 }, { "epoch": 0.5898323599108504, "grad_norm": 1.6780592203140259, "learning_rate": 7.62675179231706e-06, "loss": 1.3179, "step": 16739 }, { "epoch": 0.5898675969238087, "grad_norm": 3.6288821697235107, "learning_rate": 7.625643136849258e-06, "loss": 2.7954, "step": 16740 }, { "epoch": 0.5899028339367672, "grad_norm": 1.1887463331222534, "learning_rate": 7.624534512307753e-06, "loss": 0.9125, "step": 16741 }, { "epoch": 0.5899380709497256, "grad_norm": 1.293868064880371, "learning_rate": 7.623425918706991e-06, "loss": 0.7615, "step": 16742 }, { "epoch": 0.589973307962684, "grad_norm": 3.1022486686706543, "learning_rate": 7.6223173560614104e-06, "loss": 2.6593, "step": 16743 }, { "epoch": 0.5900085449756424, "grad_norm": 3.168290376663208, "learning_rate": 7.621208824385449e-06, "loss": 3.0579, "step": 16744 }, { "epoch": 0.5900437819886009, "grad_norm": 1.373860239982605, "learning_rate": 7.620100323693543e-06, "loss": 0.681, "step": 16745 }, { "epoch": 0.5900790190015592, "grad_norm": 4.124445915222168, "learning_rate": 7.618991854000137e-06, "loss": 2.9505, "step": 16746 }, { "epoch": 0.5901142560145176, "grad_norm": 1.9661232233047485, "learning_rate": 7.617883415319663e-06, "loss": 0.7621, "step": 16747 }, { "epoch": 0.5901494930274761, "grad_norm": 1.4467283487319946, "learning_rate": 7.61677500766656e-06, "loss": 1.0556, "step": 16748 }, { "epoch": 0.5901847300404345, "grad_norm": 2.691746234893799, "learning_rate": 7.615666631055269e-06, "loss": 0.937, "step": 16749 }, { "epoch": 0.5902199670533929, "grad_norm": 9.718496322631836, "learning_rate": 7.614558285500223e-06, "loss": 3.0582, "step": 16750 }, { "epoch": 0.5902552040663513, "grad_norm": 4.399882793426514, "learning_rate": 7.613449971015856e-06, "loss": 3.2437, "step": 16751 }, { "epoch": 0.5902904410793097, "grad_norm": 1.1773191690444946, "learning_rate": 7.612341687616611e-06, "loss": 0.945, "step": 16752 }, { "epoch": 0.5903256780922681, "grad_norm": 1.1599687337875366, "learning_rate": 7.6112334353169186e-06, "loss": 0.5628, "step": 16753 }, { "epoch": 0.5903609151052265, "grad_norm": 4.797297477722168, "learning_rate": 7.610125214131214e-06, "loss": 3.1744, "step": 16754 }, { "epoch": 0.590396152118185, "grad_norm": 5.290040969848633, "learning_rate": 7.609017024073933e-06, "loss": 3.2277, "step": 16755 }, { "epoch": 0.5904313891311433, "grad_norm": 3.2560954093933105, "learning_rate": 7.6079088651595115e-06, "loss": 3.1609, "step": 16756 }, { "epoch": 0.5904666261441017, "grad_norm": 2.4936466217041016, "learning_rate": 7.60680073740238e-06, "loss": 3.0719, "step": 16757 }, { "epoch": 0.5905018631570602, "grad_norm": 3.146078586578369, "learning_rate": 7.605692640816974e-06, "loss": 2.5695, "step": 16758 }, { "epoch": 0.5905371001700186, "grad_norm": 6.42816686630249, "learning_rate": 7.604584575417729e-06, "loss": 2.8364, "step": 16759 }, { "epoch": 0.590572337182977, "grad_norm": 3.0174717903137207, "learning_rate": 7.60347654121907e-06, "loss": 3.1598, "step": 16760 }, { "epoch": 0.5906075741959355, "grad_norm": 13.157570838928223, "learning_rate": 7.602368538235437e-06, "loss": 3.492, "step": 16761 }, { "epoch": 0.5906428112088938, "grad_norm": 5.213133335113525, "learning_rate": 7.601260566481261e-06, "loss": 4.9435, "step": 16762 }, { "epoch": 0.5906780482218522, "grad_norm": 2.5796470642089844, "learning_rate": 7.60015262597097e-06, "loss": 1.0759, "step": 16763 }, { "epoch": 0.5907132852348106, "grad_norm": 4.117455959320068, "learning_rate": 7.599044716718995e-06, "loss": 3.2737, "step": 16764 }, { "epoch": 0.5907485222477691, "grad_norm": 4.7499494552612305, "learning_rate": 7.597936838739775e-06, "loss": 2.9069, "step": 16765 }, { "epoch": 0.5907837592607275, "grad_norm": 1.5113657712936401, "learning_rate": 7.596828992047725e-06, "loss": 0.8372, "step": 16766 }, { "epoch": 0.5908189962736858, "grad_norm": 7.36107063293457, "learning_rate": 7.595721176657287e-06, "loss": 5.3475, "step": 16767 }, { "epoch": 0.5908542332866443, "grad_norm": 4.122912883758545, "learning_rate": 7.5946133925828895e-06, "loss": 3.3678, "step": 16768 }, { "epoch": 0.5908894702996027, "grad_norm": 2.5317764282226562, "learning_rate": 7.5935056398389565e-06, "loss": 2.7388, "step": 16769 }, { "epoch": 0.5909247073125611, "grad_norm": 6.54238224029541, "learning_rate": 7.5923979184399165e-06, "loss": 2.3412, "step": 16770 }, { "epoch": 0.5909599443255196, "grad_norm": 11.04384994506836, "learning_rate": 7.591290228400203e-06, "loss": 5.3527, "step": 16771 }, { "epoch": 0.590995181338478, "grad_norm": 1.4167840480804443, "learning_rate": 7.590182569734243e-06, "loss": 0.8744, "step": 16772 }, { "epoch": 0.5910304183514363, "grad_norm": 1.2535799741744995, "learning_rate": 7.58907494245646e-06, "loss": 0.91, "step": 16773 }, { "epoch": 0.5910656553643948, "grad_norm": 4.965532302856445, "learning_rate": 7.5879673465812805e-06, "loss": 2.663, "step": 16774 }, { "epoch": 0.5911008923773532, "grad_norm": 6.3704023361206055, "learning_rate": 7.5868597821231384e-06, "loss": 7.1123, "step": 16775 }, { "epoch": 0.5911361293903116, "grad_norm": 1.3224132061004639, "learning_rate": 7.585752249096452e-06, "loss": 1.0996, "step": 16776 }, { "epoch": 0.59117136640327, "grad_norm": 5.473446846008301, "learning_rate": 7.584644747515651e-06, "loss": 3.1965, "step": 16777 }, { "epoch": 0.5912066034162284, "grad_norm": 2.77154278755188, "learning_rate": 7.583537277395161e-06, "loss": 0.8516, "step": 16778 }, { "epoch": 0.5912418404291868, "grad_norm": 2.767935276031494, "learning_rate": 7.5824298387494036e-06, "loss": 2.8556, "step": 16779 }, { "epoch": 0.5912770774421452, "grad_norm": 1.888715386390686, "learning_rate": 7.581322431592806e-06, "loss": 0.7192, "step": 16780 }, { "epoch": 0.5913123144551037, "grad_norm": 2.418555736541748, "learning_rate": 7.580215055939793e-06, "loss": 0.6299, "step": 16781 }, { "epoch": 0.5913475514680621, "grad_norm": 6.089168548583984, "learning_rate": 7.579107711804786e-06, "loss": 6.2106, "step": 16782 }, { "epoch": 0.5913827884810204, "grad_norm": 5.866827964782715, "learning_rate": 7.5780003992022075e-06, "loss": 3.7819, "step": 16783 }, { "epoch": 0.5914180254939789, "grad_norm": 6.5586347579956055, "learning_rate": 7.5768931181464885e-06, "loss": 6.4457, "step": 16784 }, { "epoch": 0.5914532625069373, "grad_norm": 1.3631762266159058, "learning_rate": 7.5757858686520395e-06, "loss": 0.8863, "step": 16785 }, { "epoch": 0.5914884995198957, "grad_norm": 5.046847820281982, "learning_rate": 7.574678650733289e-06, "loss": 2.9012, "step": 16786 }, { "epoch": 0.5915237365328541, "grad_norm": 3.597736120223999, "learning_rate": 7.573571464404662e-06, "loss": 2.8191, "step": 16787 }, { "epoch": 0.5915589735458125, "grad_norm": 1.089632511138916, "learning_rate": 7.572464309680573e-06, "loss": 0.6282, "step": 16788 }, { "epoch": 0.5915942105587709, "grad_norm": 1.2595173120498657, "learning_rate": 7.571357186575443e-06, "loss": 0.8382, "step": 16789 }, { "epoch": 0.5916294475717293, "grad_norm": 4.924615859985352, "learning_rate": 7.570250095103697e-06, "loss": 3.4716, "step": 16790 }, { "epoch": 0.5916646845846878, "grad_norm": 1.2147923707962036, "learning_rate": 7.569143035279756e-06, "loss": 1.0739, "step": 16791 }, { "epoch": 0.5916999215976462, "grad_norm": 1.9743542671203613, "learning_rate": 7.568036007118033e-06, "loss": 1.0621, "step": 16792 }, { "epoch": 0.5917351586106046, "grad_norm": 3.835186004638672, "learning_rate": 7.56692901063295e-06, "loss": 2.5889, "step": 16793 }, { "epoch": 0.591770395623563, "grad_norm": 5.086119651794434, "learning_rate": 7.565822045838931e-06, "loss": 3.311, "step": 16794 }, { "epoch": 0.5918056326365214, "grad_norm": 2.1215991973876953, "learning_rate": 7.564715112750387e-06, "loss": 0.8367, "step": 16795 }, { "epoch": 0.5918408696494798, "grad_norm": 4.543814182281494, "learning_rate": 7.563608211381739e-06, "loss": 2.6485, "step": 16796 }, { "epoch": 0.5918761066624382, "grad_norm": 6.03373384475708, "learning_rate": 7.5625013417474055e-06, "loss": 5.7247, "step": 16797 }, { "epoch": 0.5919113436753967, "grad_norm": 5.498546123504639, "learning_rate": 7.5613945038617995e-06, "loss": 5.7235, "step": 16798 }, { "epoch": 0.591946580688355, "grad_norm": 1.0785813331604004, "learning_rate": 7.560287697739344e-06, "loss": 1.0511, "step": 16799 }, { "epoch": 0.5919818177013134, "grad_norm": 1.0897800922393799, "learning_rate": 7.559180923394452e-06, "loss": 0.8163, "step": 16800 }, { "epoch": 0.5920170547142719, "grad_norm": 2.300743341445923, "learning_rate": 7.558074180841538e-06, "loss": 0.7937, "step": 16801 }, { "epoch": 0.5920522917272303, "grad_norm": 6.186944961547852, "learning_rate": 7.5569674700950165e-06, "loss": 3.2553, "step": 16802 }, { "epoch": 0.5920875287401887, "grad_norm": 1.5464485883712769, "learning_rate": 7.555860791169311e-06, "loss": 1.1341, "step": 16803 }, { "epoch": 0.5921227657531472, "grad_norm": 3.7246830463409424, "learning_rate": 7.5547541440788255e-06, "loss": 2.7574, "step": 16804 }, { "epoch": 0.5921580027661055, "grad_norm": 1.434913992881775, "learning_rate": 7.5536475288379795e-06, "loss": 0.9716, "step": 16805 }, { "epoch": 0.5921932397790639, "grad_norm": 2.0248141288757324, "learning_rate": 7.552540945461189e-06, "loss": 0.8451, "step": 16806 }, { "epoch": 0.5922284767920223, "grad_norm": 6.452784538269043, "learning_rate": 7.551434393962861e-06, "loss": 5.5738, "step": 16807 }, { "epoch": 0.5922637138049808, "grad_norm": 1.4389663934707642, "learning_rate": 7.55032787435741e-06, "loss": 0.7765, "step": 16808 }, { "epoch": 0.5922989508179392, "grad_norm": 2.1863486766815186, "learning_rate": 7.549221386659253e-06, "loss": 1.069, "step": 16809 }, { "epoch": 0.5923341878308975, "grad_norm": 6.548658847808838, "learning_rate": 7.548114930882802e-06, "loss": 4.9762, "step": 16810 }, { "epoch": 0.592369424843856, "grad_norm": 1.3579844236373901, "learning_rate": 7.547008507042464e-06, "loss": 0.8697, "step": 16811 }, { "epoch": 0.5924046618568144, "grad_norm": 1.605358600616455, "learning_rate": 7.54590211515265e-06, "loss": 0.9488, "step": 16812 }, { "epoch": 0.5924398988697728, "grad_norm": 1.6392138004302979, "learning_rate": 7.544795755227778e-06, "loss": 0.7184, "step": 16813 }, { "epoch": 0.5924751358827313, "grad_norm": 1.3222780227661133, "learning_rate": 7.543689427282253e-06, "loss": 0.7264, "step": 16814 }, { "epoch": 0.5925103728956896, "grad_norm": 3.2016067504882812, "learning_rate": 7.5425831313304855e-06, "loss": 0.7016, "step": 16815 }, { "epoch": 0.592545609908648, "grad_norm": 1.159008502960205, "learning_rate": 7.541476867386888e-06, "loss": 1.0807, "step": 16816 }, { "epoch": 0.5925808469216065, "grad_norm": 4.293093681335449, "learning_rate": 7.540370635465862e-06, "loss": 3.1336, "step": 16817 }, { "epoch": 0.5926160839345649, "grad_norm": 3.378981113433838, "learning_rate": 7.539264435581824e-06, "loss": 2.9742, "step": 16818 }, { "epoch": 0.5926513209475233, "grad_norm": 5.726954460144043, "learning_rate": 7.538158267749184e-06, "loss": 3.257, "step": 16819 }, { "epoch": 0.5926865579604816, "grad_norm": 5.647912502288818, "learning_rate": 7.537052131982343e-06, "loss": 3.3614, "step": 16820 }, { "epoch": 0.5927217949734401, "grad_norm": 3.2607922554016113, "learning_rate": 7.53594602829571e-06, "loss": 1.0642, "step": 16821 }, { "epoch": 0.5927570319863985, "grad_norm": 5.266204833984375, "learning_rate": 7.534839956703699e-06, "loss": 3.0698, "step": 16822 }, { "epoch": 0.5927922689993569, "grad_norm": 2.9200100898742676, "learning_rate": 7.533733917220707e-06, "loss": 0.8369, "step": 16823 }, { "epoch": 0.5928275060123154, "grad_norm": 5.001333236694336, "learning_rate": 7.532627909861145e-06, "loss": 3.5897, "step": 16824 }, { "epoch": 0.5928627430252738, "grad_norm": 0.6901278495788574, "learning_rate": 7.5315219346394234e-06, "loss": 0.9126, "step": 16825 }, { "epoch": 0.5928979800382321, "grad_norm": 5.433087348937988, "learning_rate": 7.53041599156994e-06, "loss": 3.1462, "step": 16826 }, { "epoch": 0.5929332170511906, "grad_norm": 2.4412739276885986, "learning_rate": 7.5293100806671e-06, "loss": 1.0111, "step": 16827 }, { "epoch": 0.592968454064149, "grad_norm": 1.0037293434143066, "learning_rate": 7.528204201945313e-06, "loss": 0.8604, "step": 16828 }, { "epoch": 0.5930036910771074, "grad_norm": 2.2376041412353516, "learning_rate": 7.527098355418983e-06, "loss": 2.8353, "step": 16829 }, { "epoch": 0.5930389280900658, "grad_norm": 1.1896294355392456, "learning_rate": 7.52599254110251e-06, "loss": 0.9493, "step": 16830 }, { "epoch": 0.5930741651030242, "grad_norm": 1.1510505676269531, "learning_rate": 7.524886759010298e-06, "loss": 0.7692, "step": 16831 }, { "epoch": 0.5931094021159826, "grad_norm": 4.500489234924316, "learning_rate": 7.523781009156754e-06, "loss": 2.8222, "step": 16832 }, { "epoch": 0.593144639128941, "grad_norm": 1.0253890752792358, "learning_rate": 7.522675291556277e-06, "loss": 0.9551, "step": 16833 }, { "epoch": 0.5931798761418995, "grad_norm": 6.8437418937683105, "learning_rate": 7.5215696062232705e-06, "loss": 6.2384, "step": 16834 }, { "epoch": 0.5932151131548579, "grad_norm": 6.758711814880371, "learning_rate": 7.520463953172137e-06, "loss": 3.6016, "step": 16835 }, { "epoch": 0.5932503501678162, "grad_norm": 5.603881359100342, "learning_rate": 7.519358332417272e-06, "loss": 2.6084, "step": 16836 }, { "epoch": 0.5932855871807747, "grad_norm": 3.948085308074951, "learning_rate": 7.518252743973082e-06, "loss": 3.15, "step": 16837 }, { "epoch": 0.5933208241937331, "grad_norm": 2.3923802375793457, "learning_rate": 7.517147187853969e-06, "loss": 2.615, "step": 16838 }, { "epoch": 0.5933560612066915, "grad_norm": 1.1878429651260376, "learning_rate": 7.516041664074327e-06, "loss": 1.1352, "step": 16839 }, { "epoch": 0.5933912982196499, "grad_norm": 3.4898502826690674, "learning_rate": 7.514936172648558e-06, "loss": 2.6791, "step": 16840 }, { "epoch": 0.5934265352326084, "grad_norm": 5.610536098480225, "learning_rate": 7.513830713591068e-06, "loss": 3.1007, "step": 16841 }, { "epoch": 0.5934617722455667, "grad_norm": 7.719180107116699, "learning_rate": 7.512725286916244e-06, "loss": 4.5913, "step": 16842 }, { "epoch": 0.5934970092585251, "grad_norm": 3.8977067470550537, "learning_rate": 7.511619892638491e-06, "loss": 2.8787, "step": 16843 }, { "epoch": 0.5935322462714836, "grad_norm": 5.482645511627197, "learning_rate": 7.510514530772204e-06, "loss": 3.3961, "step": 16844 }, { "epoch": 0.593567483284442, "grad_norm": 6.328798770904541, "learning_rate": 7.509409201331787e-06, "loss": 5.0518, "step": 16845 }, { "epoch": 0.5936027202974004, "grad_norm": 2.274153470993042, "learning_rate": 7.508303904331629e-06, "loss": 0.8906, "step": 16846 }, { "epoch": 0.5936379573103588, "grad_norm": 0.8298598527908325, "learning_rate": 7.507198639786131e-06, "loss": 1.0022, "step": 16847 }, { "epoch": 0.5936731943233172, "grad_norm": 0.9726420044898987, "learning_rate": 7.50609340770969e-06, "loss": 0.7508, "step": 16848 }, { "epoch": 0.5937084313362756, "grad_norm": 1.462231993675232, "learning_rate": 7.504988208116698e-06, "loss": 1.0755, "step": 16849 }, { "epoch": 0.5937436683492341, "grad_norm": 1.6251105070114136, "learning_rate": 7.50388304102155e-06, "loss": 1.0796, "step": 16850 }, { "epoch": 0.5937789053621925, "grad_norm": 3.853482723236084, "learning_rate": 7.502777906438648e-06, "loss": 3.0688, "step": 16851 }, { "epoch": 0.5938141423751508, "grad_norm": 1.6521985530853271, "learning_rate": 7.501672804382381e-06, "loss": 0.7672, "step": 16852 }, { "epoch": 0.5938493793881092, "grad_norm": 5.534282207489014, "learning_rate": 7.500567734867142e-06, "loss": 3.106, "step": 16853 }, { "epoch": 0.5938846164010677, "grad_norm": 3.2440950870513916, "learning_rate": 7.4994626979073295e-06, "loss": 3.4167, "step": 16854 }, { "epoch": 0.5939198534140261, "grad_norm": 1.411354660987854, "learning_rate": 7.498357693517329e-06, "loss": 1.3385, "step": 16855 }, { "epoch": 0.5939550904269845, "grad_norm": 2.0321362018585205, "learning_rate": 7.497252721711541e-06, "loss": 0.9494, "step": 16856 }, { "epoch": 0.593990327439943, "grad_norm": 3.460399866104126, "learning_rate": 7.496147782504357e-06, "loss": 4.1741, "step": 16857 }, { "epoch": 0.5940255644529013, "grad_norm": 3.977196216583252, "learning_rate": 7.495042875910166e-06, "loss": 2.767, "step": 16858 }, { "epoch": 0.5940608014658597, "grad_norm": 4.224954128265381, "learning_rate": 7.493938001943357e-06, "loss": 4.0861, "step": 16859 }, { "epoch": 0.5940960384788182, "grad_norm": 1.3326483964920044, "learning_rate": 7.492833160618332e-06, "loss": 1.0545, "step": 16860 }, { "epoch": 0.5941312754917766, "grad_norm": 1.0564483404159546, "learning_rate": 7.491728351949468e-06, "loss": 0.9832, "step": 16861 }, { "epoch": 0.594166512504735, "grad_norm": 4.520407676696777, "learning_rate": 7.490623575951163e-06, "loss": 4.2645, "step": 16862 }, { "epoch": 0.5942017495176933, "grad_norm": 1.3724373579025269, "learning_rate": 7.4895188326378045e-06, "loss": 1.1267, "step": 16863 }, { "epoch": 0.5942369865306518, "grad_norm": 1.520956039428711, "learning_rate": 7.4884141220237884e-06, "loss": 0.9063, "step": 16864 }, { "epoch": 0.5942722235436102, "grad_norm": 3.206542491912842, "learning_rate": 7.487309444123494e-06, "loss": 3.1195, "step": 16865 }, { "epoch": 0.5943074605565686, "grad_norm": 0.9683027863502502, "learning_rate": 7.486204798951313e-06, "loss": 0.5903, "step": 16866 }, { "epoch": 0.5943426975695271, "grad_norm": 2.086080312728882, "learning_rate": 7.485100186521639e-06, "loss": 0.9323, "step": 16867 }, { "epoch": 0.5943779345824854, "grad_norm": 4.885565280914307, "learning_rate": 7.4839956068488526e-06, "loss": 4.7796, "step": 16868 }, { "epoch": 0.5944131715954438, "grad_norm": 2.2536818981170654, "learning_rate": 7.482891059947343e-06, "loss": 0.7639, "step": 16869 }, { "epoch": 0.5944484086084023, "grad_norm": 17.467376708984375, "learning_rate": 7.481786545831501e-06, "loss": 5.5929, "step": 16870 }, { "epoch": 0.5944836456213607, "grad_norm": 1.5467084646224976, "learning_rate": 7.480682064515707e-06, "loss": 0.7949, "step": 16871 }, { "epoch": 0.5945188826343191, "grad_norm": 1.0533188581466675, "learning_rate": 7.479577616014353e-06, "loss": 0.8536, "step": 16872 }, { "epoch": 0.5945541196472774, "grad_norm": 3.7692949771881104, "learning_rate": 7.47847320034182e-06, "loss": 3.1601, "step": 16873 }, { "epoch": 0.5945893566602359, "grad_norm": 3.4741454124450684, "learning_rate": 7.4773688175124935e-06, "loss": 2.2575, "step": 16874 }, { "epoch": 0.5946245936731943, "grad_norm": 4.371335029602051, "learning_rate": 7.47626446754076e-06, "loss": 2.9422, "step": 16875 }, { "epoch": 0.5946598306861527, "grad_norm": 4.0210490226745605, "learning_rate": 7.4751601504410065e-06, "loss": 4.4875, "step": 16876 }, { "epoch": 0.5946950676991112, "grad_norm": 1.317540168762207, "learning_rate": 7.47405586622761e-06, "loss": 0.7941, "step": 16877 }, { "epoch": 0.5947303047120696, "grad_norm": 6.329266548156738, "learning_rate": 7.472951614914959e-06, "loss": 5.1723, "step": 16878 }, { "epoch": 0.5947655417250279, "grad_norm": 1.6407197713851929, "learning_rate": 7.471847396517438e-06, "loss": 1.3782, "step": 16879 }, { "epoch": 0.5948007787379864, "grad_norm": 5.671454906463623, "learning_rate": 7.470743211049421e-06, "loss": 5.1894, "step": 16880 }, { "epoch": 0.5948360157509448, "grad_norm": 1.5860668420791626, "learning_rate": 7.469639058525297e-06, "loss": 0.9533, "step": 16881 }, { "epoch": 0.5948712527639032, "grad_norm": 1.2718942165374756, "learning_rate": 7.468534938959449e-06, "loss": 0.975, "step": 16882 }, { "epoch": 0.5949064897768617, "grad_norm": 3.6002542972564697, "learning_rate": 7.467430852366255e-06, "loss": 2.7903, "step": 16883 }, { "epoch": 0.59494172678982, "grad_norm": 4.060214519500732, "learning_rate": 7.466326798760095e-06, "loss": 3.3414, "step": 16884 }, { "epoch": 0.5949769638027784, "grad_norm": 1.3664062023162842, "learning_rate": 7.465222778155352e-06, "loss": 1.1059, "step": 16885 }, { "epoch": 0.5950122008157368, "grad_norm": 7.266522407531738, "learning_rate": 7.464118790566408e-06, "loss": 5.5146, "step": 16886 }, { "epoch": 0.5950474378286953, "grad_norm": 1.8397597074508667, "learning_rate": 7.463014836007638e-06, "loss": 0.7304, "step": 16887 }, { "epoch": 0.5950826748416537, "grad_norm": 1.554398775100708, "learning_rate": 7.461910914493421e-06, "loss": 0.8802, "step": 16888 }, { "epoch": 0.595117911854612, "grad_norm": 4.823786735534668, "learning_rate": 7.4608070260381415e-06, "loss": 5.3651, "step": 16889 }, { "epoch": 0.5951531488675705, "grad_norm": 1.5736316442489624, "learning_rate": 7.459703170656172e-06, "loss": 0.8708, "step": 16890 }, { "epoch": 0.5951883858805289, "grad_norm": 3.641572952270508, "learning_rate": 7.45859934836189e-06, "loss": 2.9425, "step": 16891 }, { "epoch": 0.5952236228934873, "grad_norm": 2.518080711364746, "learning_rate": 7.457495559169681e-06, "loss": 2.7091, "step": 16892 }, { "epoch": 0.5952588599064458, "grad_norm": 5.717798709869385, "learning_rate": 7.456391803093912e-06, "loss": 4.9089, "step": 16893 }, { "epoch": 0.5952940969194042, "grad_norm": 6.307401657104492, "learning_rate": 7.455288080148961e-06, "loss": 5.6582, "step": 16894 }, { "epoch": 0.5953293339323625, "grad_norm": 1.1729536056518555, "learning_rate": 7.454184390349213e-06, "loss": 0.9435, "step": 16895 }, { "epoch": 0.5953645709453209, "grad_norm": 3.1731977462768555, "learning_rate": 7.453080733709034e-06, "loss": 2.8323, "step": 16896 }, { "epoch": 0.5953998079582794, "grad_norm": 5.958014488220215, "learning_rate": 7.451977110242803e-06, "loss": 4.6175, "step": 16897 }, { "epoch": 0.5954350449712378, "grad_norm": 1.0131901502609253, "learning_rate": 7.450873519964897e-06, "loss": 1.1361, "step": 16898 }, { "epoch": 0.5954702819841962, "grad_norm": 1.1022419929504395, "learning_rate": 7.449769962889684e-06, "loss": 1.1335, "step": 16899 }, { "epoch": 0.5955055189971546, "grad_norm": 3.161611318588257, "learning_rate": 7.448666439031545e-06, "loss": 3.0315, "step": 16900 }, { "epoch": 0.595540756010113, "grad_norm": 6.165548324584961, "learning_rate": 7.44756294840485e-06, "loss": 5.5879, "step": 16901 }, { "epoch": 0.5955759930230714, "grad_norm": 2.727327346801758, "learning_rate": 7.446459491023974e-06, "loss": 0.9274, "step": 16902 }, { "epoch": 0.5956112300360299, "grad_norm": 1.6601439714431763, "learning_rate": 7.445356066903284e-06, "loss": 0.7929, "step": 16903 }, { "epoch": 0.5956464670489883, "grad_norm": 1.2289570569992065, "learning_rate": 7.444252676057161e-06, "loss": 0.9653, "step": 16904 }, { "epoch": 0.5956817040619466, "grad_norm": 5.113221645355225, "learning_rate": 7.443149318499973e-06, "loss": 5.283, "step": 16905 }, { "epoch": 0.595716941074905, "grad_norm": 5.340070724487305, "learning_rate": 7.442045994246089e-06, "loss": 5.3159, "step": 16906 }, { "epoch": 0.5957521780878635, "grad_norm": 6.291414260864258, "learning_rate": 7.440942703309879e-06, "loss": 5.929, "step": 16907 }, { "epoch": 0.5957874151008219, "grad_norm": 2.722123146057129, "learning_rate": 7.4398394457057234e-06, "loss": 2.9697, "step": 16908 }, { "epoch": 0.5958226521137803, "grad_norm": 1.8761013746261597, "learning_rate": 7.438736221447978e-06, "loss": 0.729, "step": 16909 }, { "epoch": 0.5958578891267388, "grad_norm": 4.347021579742432, "learning_rate": 7.437633030551023e-06, "loss": 3.2354, "step": 16910 }, { "epoch": 0.5958931261396971, "grad_norm": 3.0150463581085205, "learning_rate": 7.436529873029225e-06, "loss": 2.5163, "step": 16911 }, { "epoch": 0.5959283631526555, "grad_norm": 3.7718257904052734, "learning_rate": 7.43542674889695e-06, "loss": 0.9127, "step": 16912 }, { "epoch": 0.595963600165614, "grad_norm": 1.2470686435699463, "learning_rate": 7.434323658168567e-06, "loss": 0.8129, "step": 16913 }, { "epoch": 0.5959988371785724, "grad_norm": 2.7399837970733643, "learning_rate": 7.433220600858449e-06, "loss": 2.7832, "step": 16914 }, { "epoch": 0.5960340741915308, "grad_norm": 4.935359954833984, "learning_rate": 7.432117576980957e-06, "loss": 3.0423, "step": 16915 }, { "epoch": 0.5960693112044892, "grad_norm": 2.3988020420074463, "learning_rate": 7.431014586550462e-06, "loss": 0.8304, "step": 16916 }, { "epoch": 0.5961045482174476, "grad_norm": 3.5971102714538574, "learning_rate": 7.429911629581326e-06, "loss": 2.9852, "step": 16917 }, { "epoch": 0.596139785230406, "grad_norm": 2.7687084674835205, "learning_rate": 7.428808706087922e-06, "loss": 0.9239, "step": 16918 }, { "epoch": 0.5961750222433644, "grad_norm": 7.401571273803711, "learning_rate": 7.427705816084612e-06, "loss": 4.6949, "step": 16919 }, { "epoch": 0.5962102592563229, "grad_norm": 4.859325885772705, "learning_rate": 7.42660295958576e-06, "loss": 2.8513, "step": 16920 }, { "epoch": 0.5962454962692812, "grad_norm": 8.106160163879395, "learning_rate": 7.425500136605734e-06, "loss": 5.4874, "step": 16921 }, { "epoch": 0.5962807332822396, "grad_norm": 3.7954514026641846, "learning_rate": 7.4243973471588935e-06, "loss": 4.5739, "step": 16922 }, { "epoch": 0.5963159702951981, "grad_norm": 1.1457014083862305, "learning_rate": 7.423294591259606e-06, "loss": 0.7209, "step": 16923 }, { "epoch": 0.5963512073081565, "grad_norm": 2.0239005088806152, "learning_rate": 7.42219186892224e-06, "loss": 1.2999, "step": 16924 }, { "epoch": 0.5963864443211149, "grad_norm": 3.885702133178711, "learning_rate": 7.421089180161148e-06, "loss": 2.7864, "step": 16925 }, { "epoch": 0.5964216813340734, "grad_norm": 1.7756644487380981, "learning_rate": 7.419986524990698e-06, "loss": 1.1007, "step": 16926 }, { "epoch": 0.5964569183470317, "grad_norm": 1.1525861024856567, "learning_rate": 7.418883903425257e-06, "loss": 0.9997, "step": 16927 }, { "epoch": 0.5964921553599901, "grad_norm": 4.307831287384033, "learning_rate": 7.4177813154791755e-06, "loss": 3.0592, "step": 16928 }, { "epoch": 0.5965273923729485, "grad_norm": 2.277550220489502, "learning_rate": 7.416678761166825e-06, "loss": 2.7205, "step": 16929 }, { "epoch": 0.596562629385907, "grad_norm": 1.2268284559249878, "learning_rate": 7.415576240502564e-06, "loss": 0.6574, "step": 16930 }, { "epoch": 0.5965978663988654, "grad_norm": 1.7286735773086548, "learning_rate": 7.414473753500748e-06, "loss": 0.8088, "step": 16931 }, { "epoch": 0.5966331034118237, "grad_norm": 5.595335960388184, "learning_rate": 7.41337130017574e-06, "loss": 3.5003, "step": 16932 }, { "epoch": 0.5966683404247822, "grad_norm": 1.0227802991867065, "learning_rate": 7.412268880541906e-06, "loss": 0.9614, "step": 16933 }, { "epoch": 0.5967035774377406, "grad_norm": 0.7748078107833862, "learning_rate": 7.411166494613595e-06, "loss": 0.8279, "step": 16934 }, { "epoch": 0.596738814450699, "grad_norm": 1.5233005285263062, "learning_rate": 7.410064142405171e-06, "loss": 0.6501, "step": 16935 }, { "epoch": 0.5967740514636575, "grad_norm": 7.792613506317139, "learning_rate": 7.408961823930989e-06, "loss": 3.0522, "step": 16936 }, { "epoch": 0.5968092884766159, "grad_norm": 1.3497354984283447, "learning_rate": 7.407859539205415e-06, "loss": 0.9682, "step": 16937 }, { "epoch": 0.5968445254895742, "grad_norm": 3.7308406829833984, "learning_rate": 7.406757288242796e-06, "loss": 3.1241, "step": 16938 }, { "epoch": 0.5968797625025326, "grad_norm": 2.4479331970214844, "learning_rate": 7.405655071057496e-06, "loss": 0.829, "step": 16939 }, { "epoch": 0.5969149995154911, "grad_norm": 6.181369781494141, "learning_rate": 7.404552887663871e-06, "loss": 3.3061, "step": 16940 }, { "epoch": 0.5969502365284495, "grad_norm": 4.610713481903076, "learning_rate": 7.4034507380762696e-06, "loss": 2.756, "step": 16941 }, { "epoch": 0.5969854735414079, "grad_norm": 6.058234214782715, "learning_rate": 7.402348622309057e-06, "loss": 5.2211, "step": 16942 }, { "epoch": 0.5970207105543663, "grad_norm": 1.4404773712158203, "learning_rate": 7.401246540376585e-06, "loss": 0.9206, "step": 16943 }, { "epoch": 0.5970559475673247, "grad_norm": 38.919925689697266, "learning_rate": 7.400144492293206e-06, "loss": 3.0465, "step": 16944 }, { "epoch": 0.5970911845802831, "grad_norm": 4.5875678062438965, "learning_rate": 7.399042478073275e-06, "loss": 2.9167, "step": 16945 }, { "epoch": 0.5971264215932416, "grad_norm": 1.411095142364502, "learning_rate": 7.397940497731153e-06, "loss": 1.084, "step": 16946 }, { "epoch": 0.5971616586062, "grad_norm": 5.355322360992432, "learning_rate": 7.39683855128118e-06, "loss": 2.5863, "step": 16947 }, { "epoch": 0.5971968956191583, "grad_norm": 2.8704373836517334, "learning_rate": 7.3957366387377194e-06, "loss": 0.9865, "step": 16948 }, { "epoch": 0.5972321326321168, "grad_norm": 5.260283946990967, "learning_rate": 7.394634760115122e-06, "loss": 2.9824, "step": 16949 }, { "epoch": 0.5972673696450752, "grad_norm": 1.0500941276550293, "learning_rate": 7.3935329154277395e-06, "loss": 0.8585, "step": 16950 }, { "epoch": 0.5973026066580336, "grad_norm": 4.838894844055176, "learning_rate": 7.392431104689919e-06, "loss": 2.4279, "step": 16951 }, { "epoch": 0.597337843670992, "grad_norm": 5.408505439758301, "learning_rate": 7.39132932791602e-06, "loss": 3.3941, "step": 16952 }, { "epoch": 0.5973730806839505, "grad_norm": 1.3780696392059326, "learning_rate": 7.390227585120384e-06, "loss": 1.0377, "step": 16953 }, { "epoch": 0.5974083176969088, "grad_norm": 1.8215796947479248, "learning_rate": 7.389125876317369e-06, "loss": 0.9035, "step": 16954 }, { "epoch": 0.5974435547098672, "grad_norm": 1.8867731094360352, "learning_rate": 7.38802420152132e-06, "loss": 0.7987, "step": 16955 }, { "epoch": 0.5974787917228257, "grad_norm": 1.272759199142456, "learning_rate": 7.386922560746591e-06, "loss": 0.901, "step": 16956 }, { "epoch": 0.5975140287357841, "grad_norm": 1.2735142707824707, "learning_rate": 7.385820954007528e-06, "loss": 0.8162, "step": 16957 }, { "epoch": 0.5975492657487425, "grad_norm": 1.422581672668457, "learning_rate": 7.384719381318479e-06, "loss": 0.6725, "step": 16958 }, { "epoch": 0.5975845027617009, "grad_norm": 1.1910653114318848, "learning_rate": 7.383617842693795e-06, "loss": 0.9361, "step": 16959 }, { "epoch": 0.5976197397746593, "grad_norm": 27.750761032104492, "learning_rate": 7.382516338147818e-06, "loss": 5.7293, "step": 16960 }, { "epoch": 0.5976549767876177, "grad_norm": 1.3658816814422607, "learning_rate": 7.3814148676949024e-06, "loss": 0.7221, "step": 16961 }, { "epoch": 0.5976902138005761, "grad_norm": 3.314338207244873, "learning_rate": 7.380313431349393e-06, "loss": 0.8951, "step": 16962 }, { "epoch": 0.5977254508135346, "grad_norm": 4.510630130767822, "learning_rate": 7.379212029125633e-06, "loss": 2.3274, "step": 16963 }, { "epoch": 0.5977606878264929, "grad_norm": 4.786966323852539, "learning_rate": 7.378110661037967e-06, "loss": 3.293, "step": 16964 }, { "epoch": 0.5977959248394513, "grad_norm": 1.0152850151062012, "learning_rate": 7.3770093271007515e-06, "loss": 1.3509, "step": 16965 }, { "epoch": 0.5978311618524098, "grad_norm": 0.9025702476501465, "learning_rate": 7.3759080273283154e-06, "loss": 1.0224, "step": 16966 }, { "epoch": 0.5978663988653682, "grad_norm": 1.323289394378662, "learning_rate": 7.374806761735016e-06, "loss": 0.8272, "step": 16967 }, { "epoch": 0.5979016358783266, "grad_norm": 4.094712734222412, "learning_rate": 7.3737055303351936e-06, "loss": 3.0002, "step": 16968 }, { "epoch": 0.597936872891285, "grad_norm": 2.7942190170288086, "learning_rate": 7.37260433314319e-06, "loss": 2.6926, "step": 16969 }, { "epoch": 0.5979721099042434, "grad_norm": 3.4992146492004395, "learning_rate": 7.371503170173346e-06, "loss": 3.1404, "step": 16970 }, { "epoch": 0.5980073469172018, "grad_norm": 4.0988264083862305, "learning_rate": 7.370402041440012e-06, "loss": 3.2392, "step": 16971 }, { "epoch": 0.5980425839301602, "grad_norm": 3.651110887527466, "learning_rate": 7.3693009469575264e-06, "loss": 2.764, "step": 16972 }, { "epoch": 0.5980778209431187, "grad_norm": 7.545522689819336, "learning_rate": 7.3681998867402295e-06, "loss": 3.3807, "step": 16973 }, { "epoch": 0.598113057956077, "grad_norm": 6.823958396911621, "learning_rate": 7.367098860802463e-06, "loss": 5.1031, "step": 16974 }, { "epoch": 0.5981482949690354, "grad_norm": 28.291889190673828, "learning_rate": 7.365997869158572e-06, "loss": 2.9325, "step": 16975 }, { "epoch": 0.5981835319819939, "grad_norm": 8.673707962036133, "learning_rate": 7.364896911822894e-06, "loss": 6.7584, "step": 16976 }, { "epoch": 0.5982187689949523, "grad_norm": 2.389197587966919, "learning_rate": 7.3637959888097685e-06, "loss": 2.7857, "step": 16977 }, { "epoch": 0.5982540060079107, "grad_norm": 1.4689754247665405, "learning_rate": 7.362695100133539e-06, "loss": 0.9559, "step": 16978 }, { "epoch": 0.5982892430208692, "grad_norm": 0.9660224318504333, "learning_rate": 7.361594245808536e-06, "loss": 1.2399, "step": 16979 }, { "epoch": 0.5983244800338275, "grad_norm": 3.947094678878784, "learning_rate": 7.360493425849107e-06, "loss": 2.7288, "step": 16980 }, { "epoch": 0.5983597170467859, "grad_norm": 1.3666424751281738, "learning_rate": 7.359392640269589e-06, "loss": 0.8346, "step": 16981 }, { "epoch": 0.5983949540597444, "grad_norm": 1.0687731504440308, "learning_rate": 7.358291889084317e-06, "loss": 0.8583, "step": 16982 }, { "epoch": 0.5984301910727028, "grad_norm": 3.92897891998291, "learning_rate": 7.357191172307627e-06, "loss": 2.6538, "step": 16983 }, { "epoch": 0.5984654280856612, "grad_norm": 6.65969705581665, "learning_rate": 7.356090489953865e-06, "loss": 4.9209, "step": 16984 }, { "epoch": 0.5985006650986195, "grad_norm": 1.0886778831481934, "learning_rate": 7.354989842037354e-06, "loss": 1.0344, "step": 16985 }, { "epoch": 0.598535902111578, "grad_norm": 4.390848159790039, "learning_rate": 7.353889228572441e-06, "loss": 5.0666, "step": 16986 }, { "epoch": 0.5985711391245364, "grad_norm": 0.9280232787132263, "learning_rate": 7.352788649573459e-06, "loss": 0.6038, "step": 16987 }, { "epoch": 0.5986063761374948, "grad_norm": 1.2206984758377075, "learning_rate": 7.351688105054741e-06, "loss": 1.0885, "step": 16988 }, { "epoch": 0.5986416131504533, "grad_norm": 0.8997414112091064, "learning_rate": 7.350587595030619e-06, "loss": 0.9646, "step": 16989 }, { "epoch": 0.5986768501634117, "grad_norm": 1.0498337745666504, "learning_rate": 7.349487119515434e-06, "loss": 1.0108, "step": 16990 }, { "epoch": 0.59871208717637, "grad_norm": 3.4793734550476074, "learning_rate": 7.348386678523519e-06, "loss": 2.8484, "step": 16991 }, { "epoch": 0.5987473241893285, "grad_norm": 7.0284013748168945, "learning_rate": 7.347286272069203e-06, "loss": 5.0392, "step": 16992 }, { "epoch": 0.5987825612022869, "grad_norm": 1.2648853063583374, "learning_rate": 7.346185900166819e-06, "loss": 0.9402, "step": 16993 }, { "epoch": 0.5988177982152453, "grad_norm": 1.462424635887146, "learning_rate": 7.345085562830706e-06, "loss": 0.9017, "step": 16994 }, { "epoch": 0.5988530352282037, "grad_norm": 1.3877688646316528, "learning_rate": 7.34398526007519e-06, "loss": 0.7936, "step": 16995 }, { "epoch": 0.5988882722411621, "grad_norm": 5.48963737487793, "learning_rate": 7.342884991914604e-06, "loss": 2.7931, "step": 16996 }, { "epoch": 0.5989235092541205, "grad_norm": 8.01915168762207, "learning_rate": 7.341784758363282e-06, "loss": 4.6853, "step": 16997 }, { "epoch": 0.5989587462670789, "grad_norm": 4.666849136352539, "learning_rate": 7.340684559435547e-06, "loss": 5.0646, "step": 16998 }, { "epoch": 0.5989939832800374, "grad_norm": 2.312020778656006, "learning_rate": 7.3395843951457365e-06, "loss": 0.8703, "step": 16999 }, { "epoch": 0.5990292202929958, "grad_norm": 1.2911951541900635, "learning_rate": 7.338484265508181e-06, "loss": 0.901, "step": 17000 }, { "epoch": 0.5990644573059541, "grad_norm": 0.8229383230209351, "learning_rate": 7.3373841705372055e-06, "loss": 1.0349, "step": 17001 }, { "epoch": 0.5990996943189126, "grad_norm": 4.670854568481445, "learning_rate": 7.336284110247136e-06, "loss": 3.1067, "step": 17002 }, { "epoch": 0.599134931331871, "grad_norm": 1.6975975036621094, "learning_rate": 7.335184084652314e-06, "loss": 1.2671, "step": 17003 }, { "epoch": 0.5991701683448294, "grad_norm": 2.2634198665618896, "learning_rate": 7.334084093767051e-06, "loss": 1.0191, "step": 17004 }, { "epoch": 0.5992054053577878, "grad_norm": 3.6141819953918457, "learning_rate": 7.332984137605684e-06, "loss": 2.5619, "step": 17005 }, { "epoch": 0.5992406423707463, "grad_norm": 3.967560291290283, "learning_rate": 7.331884216182541e-06, "loss": 2.9457, "step": 17006 }, { "epoch": 0.5992758793837046, "grad_norm": 1.2547709941864014, "learning_rate": 7.330784329511944e-06, "loss": 0.744, "step": 17007 }, { "epoch": 0.599311116396663, "grad_norm": 5.303253173828125, "learning_rate": 7.329684477608219e-06, "loss": 2.5245, "step": 17008 }, { "epoch": 0.5993463534096215, "grad_norm": 6.591689586639404, "learning_rate": 7.328584660485696e-06, "loss": 4.6466, "step": 17009 }, { "epoch": 0.5993815904225799, "grad_norm": 4.414709091186523, "learning_rate": 7.327484878158699e-06, "loss": 2.9515, "step": 17010 }, { "epoch": 0.5994168274355383, "grad_norm": 1.192650556564331, "learning_rate": 7.326385130641551e-06, "loss": 0.8335, "step": 17011 }, { "epoch": 0.5994520644484967, "grad_norm": 1.4820655584335327, "learning_rate": 7.325285417948575e-06, "loss": 0.6113, "step": 17012 }, { "epoch": 0.5994873014614551, "grad_norm": 13.831055641174316, "learning_rate": 7.324185740094102e-06, "loss": 5.4367, "step": 17013 }, { "epoch": 0.5995225384744135, "grad_norm": 8.916605949401855, "learning_rate": 7.323086097092448e-06, "loss": 6.1373, "step": 17014 }, { "epoch": 0.5995577754873719, "grad_norm": 2.2995643615722656, "learning_rate": 7.321986488957937e-06, "loss": 0.8709, "step": 17015 }, { "epoch": 0.5995930125003304, "grad_norm": 3.914480686187744, "learning_rate": 7.320886915704898e-06, "loss": 3.7174, "step": 17016 }, { "epoch": 0.5996282495132887, "grad_norm": 5.266054630279541, "learning_rate": 7.319787377347643e-06, "loss": 5.539, "step": 17017 }, { "epoch": 0.5996634865262471, "grad_norm": 4.390064239501953, "learning_rate": 7.3186878739005e-06, "loss": 2.988, "step": 17018 }, { "epoch": 0.5996987235392056, "grad_norm": 2.1353790760040283, "learning_rate": 7.317588405377791e-06, "loss": 0.9912, "step": 17019 }, { "epoch": 0.599733960552164, "grad_norm": 6.154698848724365, "learning_rate": 7.3164889717938334e-06, "loss": 4.5885, "step": 17020 }, { "epoch": 0.5997691975651224, "grad_norm": 3.9234259128570557, "learning_rate": 7.315389573162947e-06, "loss": 3.316, "step": 17021 }, { "epoch": 0.5998044345780809, "grad_norm": 3.5240421295166016, "learning_rate": 7.314290209499458e-06, "loss": 3.1257, "step": 17022 }, { "epoch": 0.5998396715910392, "grad_norm": 0.9729987382888794, "learning_rate": 7.313190880817675e-06, "loss": 0.7216, "step": 17023 }, { "epoch": 0.5998749086039976, "grad_norm": 1.6698591709136963, "learning_rate": 7.312091587131928e-06, "loss": 0.666, "step": 17024 }, { "epoch": 0.5999101456169561, "grad_norm": 3.5931034088134766, "learning_rate": 7.31099232845653e-06, "loss": 2.5835, "step": 17025 }, { "epoch": 0.5999453826299145, "grad_norm": 5.291471481323242, "learning_rate": 7.309893104805799e-06, "loss": 4.4642, "step": 17026 }, { "epoch": 0.5999806196428729, "grad_norm": 1.835376262664795, "learning_rate": 7.308793916194048e-06, "loss": 0.9558, "step": 17027 }, { "epoch": 0.6000158566558312, "grad_norm": 4.849117755889893, "learning_rate": 7.307694762635604e-06, "loss": 3.244, "step": 17028 }, { "epoch": 0.6000510936687897, "grad_norm": 5.142545700073242, "learning_rate": 7.3065956441447795e-06, "loss": 2.6535, "step": 17029 }, { "epoch": 0.6000863306817481, "grad_norm": 1.1873302459716797, "learning_rate": 7.305496560735888e-06, "loss": 0.8937, "step": 17030 }, { "epoch": 0.6001215676947065, "grad_norm": 3.5924625396728516, "learning_rate": 7.304397512423245e-06, "loss": 3.175, "step": 17031 }, { "epoch": 0.600156804707665, "grad_norm": 0.7907223105430603, "learning_rate": 7.303298499221172e-06, "loss": 0.7263, "step": 17032 }, { "epoch": 0.6001920417206233, "grad_norm": 2.002934694290161, "learning_rate": 7.3021995211439765e-06, "loss": 0.8919, "step": 17033 }, { "epoch": 0.6002272787335817, "grad_norm": 1.2014880180358887, "learning_rate": 7.301100578205978e-06, "loss": 0.8447, "step": 17034 }, { "epoch": 0.6002625157465402, "grad_norm": 6.248047828674316, "learning_rate": 7.300001670421489e-06, "loss": 2.9078, "step": 17035 }, { "epoch": 0.6002977527594986, "grad_norm": 5.040388107299805, "learning_rate": 7.298902797804816e-06, "loss": 2.5533, "step": 17036 }, { "epoch": 0.600332989772457, "grad_norm": 1.0810637474060059, "learning_rate": 7.297803960370284e-06, "loss": 0.8861, "step": 17037 }, { "epoch": 0.6003682267854153, "grad_norm": 1.1166986227035522, "learning_rate": 7.296705158132199e-06, "loss": 0.9267, "step": 17038 }, { "epoch": 0.6004034637983738, "grad_norm": 6.294376850128174, "learning_rate": 7.295606391104872e-06, "loss": 5.4473, "step": 17039 }, { "epoch": 0.6004387008113322, "grad_norm": 4.974393844604492, "learning_rate": 7.294507659302616e-06, "loss": 3.0314, "step": 17040 }, { "epoch": 0.6004739378242906, "grad_norm": 6.023075580596924, "learning_rate": 7.2934089627397465e-06, "loss": 4.9827, "step": 17041 }, { "epoch": 0.6005091748372491, "grad_norm": 4.873607635498047, "learning_rate": 7.292310301430565e-06, "loss": 3.0024, "step": 17042 }, { "epoch": 0.6005444118502075, "grad_norm": 1.5556881427764893, "learning_rate": 7.291211675389388e-06, "loss": 0.9771, "step": 17043 }, { "epoch": 0.6005796488631658, "grad_norm": 1.0783264636993408, "learning_rate": 7.290113084630526e-06, "loss": 0.6584, "step": 17044 }, { "epoch": 0.6006148858761243, "grad_norm": 1.2057039737701416, "learning_rate": 7.289014529168286e-06, "loss": 0.9932, "step": 17045 }, { "epoch": 0.6006501228890827, "grad_norm": 1.7328568696975708, "learning_rate": 7.287916009016973e-06, "loss": 1.0524, "step": 17046 }, { "epoch": 0.6006853599020411, "grad_norm": 9.543330192565918, "learning_rate": 7.286817524190901e-06, "loss": 3.1339, "step": 17047 }, { "epoch": 0.6007205969149995, "grad_norm": 5.07565975189209, "learning_rate": 7.285719074704381e-06, "loss": 3.3901, "step": 17048 }, { "epoch": 0.600755833927958, "grad_norm": 5.124457836151123, "learning_rate": 7.2846206605717105e-06, "loss": 2.274, "step": 17049 }, { "epoch": 0.6007910709409163, "grad_norm": 1.3789913654327393, "learning_rate": 7.2835222818072026e-06, "loss": 1.0965, "step": 17050 }, { "epoch": 0.6008263079538747, "grad_norm": 1.1582552194595337, "learning_rate": 7.2824239384251675e-06, "loss": 0.8728, "step": 17051 }, { "epoch": 0.6008615449668332, "grad_norm": 1.2903505563735962, "learning_rate": 7.2813256304399016e-06, "loss": 0.8279, "step": 17052 }, { "epoch": 0.6008967819797916, "grad_norm": 3.8423080444335938, "learning_rate": 7.280227357865717e-06, "loss": 2.9436, "step": 17053 }, { "epoch": 0.60093201899275, "grad_norm": 1.1136858463287354, "learning_rate": 7.27912912071692e-06, "loss": 0.907, "step": 17054 }, { "epoch": 0.6009672560057084, "grad_norm": 1.3917770385742188, "learning_rate": 7.27803091900781e-06, "loss": 0.7077, "step": 17055 }, { "epoch": 0.6010024930186668, "grad_norm": 3.2899117469787598, "learning_rate": 7.276932752752692e-06, "loss": 2.8646, "step": 17056 }, { "epoch": 0.6010377300316252, "grad_norm": 1.709078073501587, "learning_rate": 7.2758346219658784e-06, "loss": 0.9858, "step": 17057 }, { "epoch": 0.6010729670445837, "grad_norm": 1.5519731044769287, "learning_rate": 7.274736526661661e-06, "loss": 0.7758, "step": 17058 }, { "epoch": 0.6011082040575421, "grad_norm": 2.589989185333252, "learning_rate": 7.2736384668543485e-06, "loss": 0.8076, "step": 17059 }, { "epoch": 0.6011434410705004, "grad_norm": 4.880417823791504, "learning_rate": 7.272540442558243e-06, "loss": 3.0933, "step": 17060 }, { "epoch": 0.6011786780834588, "grad_norm": 1.2944480180740356, "learning_rate": 7.2714424537876425e-06, "loss": 0.7939, "step": 17061 }, { "epoch": 0.6012139150964173, "grad_norm": 6.51862096786499, "learning_rate": 7.2703445005568544e-06, "loss": 5.086, "step": 17062 }, { "epoch": 0.6012491521093757, "grad_norm": 5.35152006149292, "learning_rate": 7.269246582880176e-06, "loss": 3.3164, "step": 17063 }, { "epoch": 0.6012843891223341, "grad_norm": 1.2586452960968018, "learning_rate": 7.268148700771911e-06, "loss": 0.8376, "step": 17064 }, { "epoch": 0.6013196261352926, "grad_norm": 1.5564818382263184, "learning_rate": 7.267050854246354e-06, "loss": 0.9467, "step": 17065 }, { "epoch": 0.6013548631482509, "grad_norm": 3.6592061519622803, "learning_rate": 7.265953043317808e-06, "loss": 2.9493, "step": 17066 }, { "epoch": 0.6013901001612093, "grad_norm": 1.4473240375518799, "learning_rate": 7.264855268000575e-06, "loss": 0.6678, "step": 17067 }, { "epoch": 0.6014253371741678, "grad_norm": 6.212142467498779, "learning_rate": 7.2637575283089475e-06, "loss": 3.0238, "step": 17068 }, { "epoch": 0.6014605741871262, "grad_norm": 3.5153887271881104, "learning_rate": 7.262659824257227e-06, "loss": 2.9809, "step": 17069 }, { "epoch": 0.6014958112000846, "grad_norm": 4.915975093841553, "learning_rate": 7.261562155859715e-06, "loss": 2.8322, "step": 17070 }, { "epoch": 0.6015310482130429, "grad_norm": 1.140272617340088, "learning_rate": 7.260464523130699e-06, "loss": 0.9805, "step": 17071 }, { "epoch": 0.6015662852260014, "grad_norm": 3.833214044570923, "learning_rate": 7.259366926084483e-06, "loss": 3.1479, "step": 17072 }, { "epoch": 0.6016015222389598, "grad_norm": 1.395822525024414, "learning_rate": 7.258269364735367e-06, "loss": 1.0815, "step": 17073 }, { "epoch": 0.6016367592519182, "grad_norm": 4.2365312576293945, "learning_rate": 7.257171839097636e-06, "loss": 3.417, "step": 17074 }, { "epoch": 0.6016719962648767, "grad_norm": 4.719814300537109, "learning_rate": 7.256074349185592e-06, "loss": 4.7793, "step": 17075 }, { "epoch": 0.601707233277835, "grad_norm": 5.262876987457275, "learning_rate": 7.25497689501353e-06, "loss": 4.9183, "step": 17076 }, { "epoch": 0.6017424702907934, "grad_norm": 3.003718614578247, "learning_rate": 7.253879476595744e-06, "loss": 2.5129, "step": 17077 }, { "epoch": 0.6017777073037519, "grad_norm": 1.919766902923584, "learning_rate": 7.252782093946527e-06, "loss": 1.0677, "step": 17078 }, { "epoch": 0.6018129443167103, "grad_norm": 1.2361136674880981, "learning_rate": 7.2516847470801744e-06, "loss": 0.7307, "step": 17079 }, { "epoch": 0.6018481813296687, "grad_norm": 5.4307050704956055, "learning_rate": 7.250587436010975e-06, "loss": 6.5628, "step": 17080 }, { "epoch": 0.601883418342627, "grad_norm": 4.858793258666992, "learning_rate": 7.249490160753226e-06, "loss": 3.0317, "step": 17081 }, { "epoch": 0.6019186553555855, "grad_norm": 1.0961956977844238, "learning_rate": 7.248392921321219e-06, "loss": 0.9218, "step": 17082 }, { "epoch": 0.6019538923685439, "grad_norm": 3.078244686126709, "learning_rate": 7.247295717729246e-06, "loss": 3.0638, "step": 17083 }, { "epoch": 0.6019891293815023, "grad_norm": 1.2149131298065186, "learning_rate": 7.246198549991592e-06, "loss": 0.8245, "step": 17084 }, { "epoch": 0.6020243663944608, "grad_norm": 1.0309067964553833, "learning_rate": 7.245101418122556e-06, "loss": 1.0279, "step": 17085 }, { "epoch": 0.6020596034074192, "grad_norm": 0.9547691941261292, "learning_rate": 7.244004322136426e-06, "loss": 0.6226, "step": 17086 }, { "epoch": 0.6020948404203775, "grad_norm": 3.499544620513916, "learning_rate": 7.24290726204749e-06, "loss": 3.111, "step": 17087 }, { "epoch": 0.602130077433336, "grad_norm": 8.693784713745117, "learning_rate": 7.2418102378700354e-06, "loss": 3.4497, "step": 17088 }, { "epoch": 0.6021653144462944, "grad_norm": 1.0384382009506226, "learning_rate": 7.2407132496183595e-06, "loss": 1.2045, "step": 17089 }, { "epoch": 0.6022005514592528, "grad_norm": 1.0873900651931763, "learning_rate": 7.23961629730674e-06, "loss": 1.0447, "step": 17090 }, { "epoch": 0.6022357884722113, "grad_norm": 2.9533884525299072, "learning_rate": 7.23851938094947e-06, "loss": 2.815, "step": 17091 }, { "epoch": 0.6022710254851696, "grad_norm": 0.9038052558898926, "learning_rate": 7.23742250056084e-06, "loss": 1.0993, "step": 17092 }, { "epoch": 0.602306262498128, "grad_norm": 2.732673406600952, "learning_rate": 7.236325656155132e-06, "loss": 1.0687, "step": 17093 }, { "epoch": 0.6023414995110864, "grad_norm": 1.244372010231018, "learning_rate": 7.2352288477466315e-06, "loss": 0.7351, "step": 17094 }, { "epoch": 0.6023767365240449, "grad_norm": 3.9461777210235596, "learning_rate": 7.234132075349633e-06, "loss": 3.3004, "step": 17095 }, { "epoch": 0.6024119735370033, "grad_norm": 4.483983039855957, "learning_rate": 7.233035338978414e-06, "loss": 2.8359, "step": 17096 }, { "epoch": 0.6024472105499616, "grad_norm": 3.760820150375366, "learning_rate": 7.23193863864726e-06, "loss": 2.7877, "step": 17097 }, { "epoch": 0.6024824475629201, "grad_norm": 10.381964683532715, "learning_rate": 7.230841974370462e-06, "loss": 4.788, "step": 17098 }, { "epoch": 0.6025176845758785, "grad_norm": 5.700990676879883, "learning_rate": 7.229745346162296e-06, "loss": 2.9093, "step": 17099 }, { "epoch": 0.6025529215888369, "grad_norm": 4.210046291351318, "learning_rate": 7.228648754037051e-06, "loss": 2.9792, "step": 17100 }, { "epoch": 0.6025881586017954, "grad_norm": 4.367655277252197, "learning_rate": 7.227552198009009e-06, "loss": 3.9747, "step": 17101 }, { "epoch": 0.6026233956147538, "grad_norm": 9.230080604553223, "learning_rate": 7.226455678092456e-06, "loss": 7.0373, "step": 17102 }, { "epoch": 0.6026586326277121, "grad_norm": 3.4253759384155273, "learning_rate": 7.225359194301665e-06, "loss": 2.9842, "step": 17103 }, { "epoch": 0.6026938696406705, "grad_norm": 1.6407510042190552, "learning_rate": 7.224262746650928e-06, "loss": 0.9043, "step": 17104 }, { "epoch": 0.602729106653629, "grad_norm": 2.182222366333008, "learning_rate": 7.2231663351545235e-06, "loss": 0.82, "step": 17105 }, { "epoch": 0.6027643436665874, "grad_norm": 1.5145009756088257, "learning_rate": 7.2220699598267285e-06, "loss": 0.8054, "step": 17106 }, { "epoch": 0.6027995806795458, "grad_norm": 5.498722553253174, "learning_rate": 7.220973620681825e-06, "loss": 2.7497, "step": 17107 }, { "epoch": 0.6028348176925042, "grad_norm": 1.3144786357879639, "learning_rate": 7.2198773177341e-06, "loss": 1.094, "step": 17108 }, { "epoch": 0.6028700547054626, "grad_norm": 5.244979381561279, "learning_rate": 7.218781050997821e-06, "loss": 3.5229, "step": 17109 }, { "epoch": 0.602905291718421, "grad_norm": 1.7020010948181152, "learning_rate": 7.217684820487274e-06, "loss": 0.7646, "step": 17110 }, { "epoch": 0.6029405287313795, "grad_norm": 4.440914154052734, "learning_rate": 7.216588626216741e-06, "loss": 3.5439, "step": 17111 }, { "epoch": 0.6029757657443379, "grad_norm": 1.852043628692627, "learning_rate": 7.215492468200491e-06, "loss": 0.8242, "step": 17112 }, { "epoch": 0.6030110027572962, "grad_norm": 7.24239444732666, "learning_rate": 7.214396346452805e-06, "loss": 7.7617, "step": 17113 }, { "epoch": 0.6030462397702546, "grad_norm": 4.830329895019531, "learning_rate": 7.213300260987968e-06, "loss": 3.134, "step": 17114 }, { "epoch": 0.6030814767832131, "grad_norm": 4.791012763977051, "learning_rate": 7.212204211820245e-06, "loss": 2.4699, "step": 17115 }, { "epoch": 0.6031167137961715, "grad_norm": 0.9971742630004883, "learning_rate": 7.211108198963919e-06, "loss": 0.8622, "step": 17116 }, { "epoch": 0.6031519508091299, "grad_norm": 13.008347511291504, "learning_rate": 7.210012222433265e-06, "loss": 4.9658, "step": 17117 }, { "epoch": 0.6031871878220884, "grad_norm": 2.1204233169555664, "learning_rate": 7.208916282242553e-06, "loss": 0.8708, "step": 17118 }, { "epoch": 0.6032224248350467, "grad_norm": 1.1431043148040771, "learning_rate": 7.207820378406064e-06, "loss": 1.1352, "step": 17119 }, { "epoch": 0.6032576618480051, "grad_norm": 2.10046124458313, "learning_rate": 7.2067245109380705e-06, "loss": 0.8944, "step": 17120 }, { "epoch": 0.6032928988609636, "grad_norm": 1.438855528831482, "learning_rate": 7.205628679852848e-06, "loss": 0.8059, "step": 17121 }, { "epoch": 0.603328135873922, "grad_norm": 1.6747561693191528, "learning_rate": 7.204532885164665e-06, "loss": 0.8205, "step": 17122 }, { "epoch": 0.6033633728868804, "grad_norm": 4.274397373199463, "learning_rate": 7.203437126887798e-06, "loss": 3.2674, "step": 17123 }, { "epoch": 0.6033986098998388, "grad_norm": 3.644974708557129, "learning_rate": 7.202341405036521e-06, "loss": 3.8677, "step": 17124 }, { "epoch": 0.6034338469127972, "grad_norm": 4.502880573272705, "learning_rate": 7.2012457196251e-06, "loss": 3.0035, "step": 17125 }, { "epoch": 0.6034690839257556, "grad_norm": 5.476222991943359, "learning_rate": 7.2001500706678106e-06, "loss": 2.3387, "step": 17126 }, { "epoch": 0.603504320938714, "grad_norm": 1.2011559009552002, "learning_rate": 7.199054458178929e-06, "loss": 0.6663, "step": 17127 }, { "epoch": 0.6035395579516725, "grad_norm": 1.0320892333984375, "learning_rate": 7.1979588821727135e-06, "loss": 0.9677, "step": 17128 }, { "epoch": 0.6035747949646308, "grad_norm": 2.3801183700561523, "learning_rate": 7.19686334266344e-06, "loss": 2.6481, "step": 17129 }, { "epoch": 0.6036100319775892, "grad_norm": 4.371452331542969, "learning_rate": 7.195767839665384e-06, "loss": 3.7526, "step": 17130 }, { "epoch": 0.6036452689905477, "grad_norm": 0.8167800307273865, "learning_rate": 7.1946723731928046e-06, "loss": 0.6981, "step": 17131 }, { "epoch": 0.6036805060035061, "grad_norm": 39.37136459350586, "learning_rate": 7.193576943259975e-06, "loss": 2.9949, "step": 17132 }, { "epoch": 0.6037157430164645, "grad_norm": 11.836702346801758, "learning_rate": 7.192481549881165e-06, "loss": 3.4821, "step": 17133 }, { "epoch": 0.603750980029423, "grad_norm": 3.8438360691070557, "learning_rate": 7.19138619307064e-06, "loss": 0.6827, "step": 17134 }, { "epoch": 0.6037862170423813, "grad_norm": 1.9186062812805176, "learning_rate": 7.190290872842667e-06, "loss": 0.6396, "step": 17135 }, { "epoch": 0.6038214540553397, "grad_norm": 3.845899820327759, "learning_rate": 7.189195589211511e-06, "loss": 2.5492, "step": 17136 }, { "epoch": 0.6038566910682981, "grad_norm": 9.894539833068848, "learning_rate": 7.188100342191444e-06, "loss": 7.0854, "step": 17137 }, { "epoch": 0.6038919280812566, "grad_norm": 1.6442264318466187, "learning_rate": 7.1870051317967286e-06, "loss": 0.8625, "step": 17138 }, { "epoch": 0.603927165094215, "grad_norm": 4.740469932556152, "learning_rate": 7.185909958041628e-06, "loss": 4.8603, "step": 17139 }, { "epoch": 0.6039624021071733, "grad_norm": 1.386154055595398, "learning_rate": 7.184814820940411e-06, "loss": 0.8258, "step": 17140 }, { "epoch": 0.6039976391201318, "grad_norm": 1.160380482673645, "learning_rate": 7.183719720507334e-06, "loss": 0.9978, "step": 17141 }, { "epoch": 0.6040328761330902, "grad_norm": 1.0387380123138428, "learning_rate": 7.182624656756671e-06, "loss": 1.0079, "step": 17142 }, { "epoch": 0.6040681131460486, "grad_norm": 2.5055675506591797, "learning_rate": 7.181529629702681e-06, "loss": 1.2314, "step": 17143 }, { "epoch": 0.6041033501590071, "grad_norm": 1.3642910718917847, "learning_rate": 7.1804346393596255e-06, "loss": 1.0076, "step": 17144 }, { "epoch": 0.6041385871719654, "grad_norm": 4.373058319091797, "learning_rate": 7.179339685741764e-06, "loss": 3.2689, "step": 17145 }, { "epoch": 0.6041738241849238, "grad_norm": 3.6954243183135986, "learning_rate": 7.178244768863371e-06, "loss": 3.5523, "step": 17146 }, { "epoch": 0.6042090611978822, "grad_norm": 1.6322375535964966, "learning_rate": 7.177149888738692e-06, "loss": 0.9991, "step": 17147 }, { "epoch": 0.6042442982108407, "grad_norm": 5.792183876037598, "learning_rate": 7.1760550453819975e-06, "loss": 3.5833, "step": 17148 }, { "epoch": 0.6042795352237991, "grad_norm": 3.843297004699707, "learning_rate": 7.1749602388075475e-06, "loss": 2.509, "step": 17149 }, { "epoch": 0.6043147722367574, "grad_norm": 3.3567910194396973, "learning_rate": 7.1738654690296e-06, "loss": 3.0845, "step": 17150 }, { "epoch": 0.6043500092497159, "grad_norm": 3.8842556476593018, "learning_rate": 7.17277073606241e-06, "loss": 2.7431, "step": 17151 }, { "epoch": 0.6043852462626743, "grad_norm": 3.8736987113952637, "learning_rate": 7.1716760399202465e-06, "loss": 2.7258, "step": 17152 }, { "epoch": 0.6044204832756327, "grad_norm": 4.625690460205078, "learning_rate": 7.170581380617361e-06, "loss": 2.8516, "step": 17153 }, { "epoch": 0.6044557202885912, "grad_norm": 4.879698753356934, "learning_rate": 7.169486758168014e-06, "loss": 5.3535, "step": 17154 }, { "epoch": 0.6044909573015496, "grad_norm": 1.0719972848892212, "learning_rate": 7.168392172586459e-06, "loss": 1.089, "step": 17155 }, { "epoch": 0.6045261943145079, "grad_norm": 5.076480388641357, "learning_rate": 7.167297623886962e-06, "loss": 3.3694, "step": 17156 }, { "epoch": 0.6045614313274664, "grad_norm": 1.2609463930130005, "learning_rate": 7.166203112083772e-06, "loss": 0.9666, "step": 17157 }, { "epoch": 0.6045966683404248, "grad_norm": 2.2794077396392822, "learning_rate": 7.1651086371911475e-06, "loss": 0.845, "step": 17158 }, { "epoch": 0.6046319053533832, "grad_norm": 0.9574462175369263, "learning_rate": 7.1640141992233455e-06, "loss": 0.8923, "step": 17159 }, { "epoch": 0.6046671423663416, "grad_norm": 2.3844501972198486, "learning_rate": 7.1629197981946165e-06, "loss": 0.9839, "step": 17160 }, { "epoch": 0.6047023793793, "grad_norm": 1.5730217695236206, "learning_rate": 7.161825434119219e-06, "loss": 0.9141, "step": 17161 }, { "epoch": 0.6047376163922584, "grad_norm": 1.632532000541687, "learning_rate": 7.1607311070114106e-06, "loss": 1.1478, "step": 17162 }, { "epoch": 0.6047728534052168, "grad_norm": 6.8781914710998535, "learning_rate": 7.159636816885438e-06, "loss": 3.2654, "step": 17163 }, { "epoch": 0.6048080904181753, "grad_norm": 2.768779993057251, "learning_rate": 7.1585425637555564e-06, "loss": 0.7719, "step": 17164 }, { "epoch": 0.6048433274311337, "grad_norm": 1.997296690940857, "learning_rate": 7.157448347636026e-06, "loss": 0.9172, "step": 17165 }, { "epoch": 0.604878564444092, "grad_norm": 0.9568131566047668, "learning_rate": 7.156354168541087e-06, "loss": 0.8704, "step": 17166 }, { "epoch": 0.6049138014570505, "grad_norm": 1.8361612558364868, "learning_rate": 7.155260026484998e-06, "loss": 0.7596, "step": 17167 }, { "epoch": 0.6049490384700089, "grad_norm": 5.775777339935303, "learning_rate": 7.154165921482012e-06, "loss": 7.0683, "step": 17168 }, { "epoch": 0.6049842754829673, "grad_norm": 5.4153242111206055, "learning_rate": 7.1530718535463765e-06, "loss": 3.4999, "step": 17169 }, { "epoch": 0.6050195124959257, "grad_norm": 1.3207694292068481, "learning_rate": 7.151977822692339e-06, "loss": 1.0857, "step": 17170 }, { "epoch": 0.6050547495088842, "grad_norm": 1.0411211252212524, "learning_rate": 7.1508838289341585e-06, "loss": 1.0999, "step": 17171 }, { "epoch": 0.6050899865218425, "grad_norm": 1.8833978176116943, "learning_rate": 7.149789872286076e-06, "loss": 2.5994, "step": 17172 }, { "epoch": 0.6051252235348009, "grad_norm": 1.487989068031311, "learning_rate": 7.148695952762344e-06, "loss": 1.0427, "step": 17173 }, { "epoch": 0.6051604605477594, "grad_norm": 1.4219409227371216, "learning_rate": 7.147602070377208e-06, "loss": 0.7999, "step": 17174 }, { "epoch": 0.6051956975607178, "grad_norm": 1.3484333753585815, "learning_rate": 7.146508225144924e-06, "loss": 0.5711, "step": 17175 }, { "epoch": 0.6052309345736762, "grad_norm": 4.897206783294678, "learning_rate": 7.14541441707973e-06, "loss": 3.0396, "step": 17176 }, { "epoch": 0.6052661715866346, "grad_norm": 2.2862792015075684, "learning_rate": 7.144320646195877e-06, "loss": 0.8824, "step": 17177 }, { "epoch": 0.605301408599593, "grad_norm": 1.8527082204818726, "learning_rate": 7.143226912507615e-06, "loss": 1.0692, "step": 17178 }, { "epoch": 0.6053366456125514, "grad_norm": 4.039455890655518, "learning_rate": 7.14213321602918e-06, "loss": 3.1398, "step": 17179 }, { "epoch": 0.6053718826255098, "grad_norm": 8.267654418945312, "learning_rate": 7.141039556774827e-06, "loss": 5.9614, "step": 17180 }, { "epoch": 0.6054071196384683, "grad_norm": 3.6181132793426514, "learning_rate": 7.139945934758801e-06, "loss": 3.0274, "step": 17181 }, { "epoch": 0.6054423566514266, "grad_norm": 1.5265542268753052, "learning_rate": 7.13885234999534e-06, "loss": 0.8291, "step": 17182 }, { "epoch": 0.605477593664385, "grad_norm": 1.2256850004196167, "learning_rate": 7.13775880249869e-06, "loss": 0.9195, "step": 17183 }, { "epoch": 0.6055128306773435, "grad_norm": 3.752039670944214, "learning_rate": 7.136665292283103e-06, "loss": 3.174, "step": 17184 }, { "epoch": 0.6055480676903019, "grad_norm": 3.3242006301879883, "learning_rate": 7.135571819362807e-06, "loss": 2.8298, "step": 17185 }, { "epoch": 0.6055833047032603, "grad_norm": 5.139764785766602, "learning_rate": 7.134478383752057e-06, "loss": 5.509, "step": 17186 }, { "epoch": 0.6056185417162188, "grad_norm": 1.3595890998840332, "learning_rate": 7.1333849854650925e-06, "loss": 1.0505, "step": 17187 }, { "epoch": 0.6056537787291771, "grad_norm": 2.640958070755005, "learning_rate": 7.132291624516152e-06, "loss": 0.7594, "step": 17188 }, { "epoch": 0.6056890157421355, "grad_norm": 1.2489453554153442, "learning_rate": 7.131198300919478e-06, "loss": 1.0063, "step": 17189 }, { "epoch": 0.605724252755094, "grad_norm": 6.379782199859619, "learning_rate": 7.130105014689311e-06, "loss": 4.9767, "step": 17190 }, { "epoch": 0.6057594897680524, "grad_norm": 0.9029676914215088, "learning_rate": 7.129011765839894e-06, "loss": 0.8203, "step": 17191 }, { "epoch": 0.6057947267810108, "grad_norm": 4.038215637207031, "learning_rate": 7.127918554385462e-06, "loss": 3.4672, "step": 17192 }, { "epoch": 0.6058299637939691, "grad_norm": 1.7729833126068115, "learning_rate": 7.126825380340257e-06, "loss": 0.8144, "step": 17193 }, { "epoch": 0.6058652008069276, "grad_norm": 1.3238542079925537, "learning_rate": 7.12573224371852e-06, "loss": 1.0874, "step": 17194 }, { "epoch": 0.605900437819886, "grad_norm": 4.152118682861328, "learning_rate": 7.124639144534485e-06, "loss": 2.7685, "step": 17195 }, { "epoch": 0.6059356748328444, "grad_norm": 4.502940654754639, "learning_rate": 7.123546082802391e-06, "loss": 4.6256, "step": 17196 }, { "epoch": 0.6059709118458029, "grad_norm": 1.3388525247573853, "learning_rate": 7.122453058536477e-06, "loss": 0.8106, "step": 17197 }, { "epoch": 0.6060061488587613, "grad_norm": 5.643052577972412, "learning_rate": 7.1213600717509775e-06, "loss": 2.8019, "step": 17198 }, { "epoch": 0.6060413858717196, "grad_norm": 1.542981743812561, "learning_rate": 7.120267122460127e-06, "loss": 0.7207, "step": 17199 }, { "epoch": 0.6060766228846781, "grad_norm": 1.1876808404922485, "learning_rate": 7.119174210678168e-06, "loss": 0.5935, "step": 17200 }, { "epoch": 0.6061118598976365, "grad_norm": 6.6066508293151855, "learning_rate": 7.118081336419331e-06, "loss": 2.6923, "step": 17201 }, { "epoch": 0.6061470969105949, "grad_norm": 5.344023704528809, "learning_rate": 7.1169884996978515e-06, "loss": 2.7986, "step": 17202 }, { "epoch": 0.6061823339235533, "grad_norm": 1.0489321947097778, "learning_rate": 7.115895700527966e-06, "loss": 0.7854, "step": 17203 }, { "epoch": 0.6062175709365117, "grad_norm": 0.9821867942810059, "learning_rate": 7.1148029389239015e-06, "loss": 0.8221, "step": 17204 }, { "epoch": 0.6062528079494701, "grad_norm": 1.4967764616012573, "learning_rate": 7.113710214899898e-06, "loss": 1.2305, "step": 17205 }, { "epoch": 0.6062880449624285, "grad_norm": 1.1531707048416138, "learning_rate": 7.112617528470189e-06, "loss": 1.002, "step": 17206 }, { "epoch": 0.606323281975387, "grad_norm": 3.9753012657165527, "learning_rate": 7.111524879649003e-06, "loss": 2.9763, "step": 17207 }, { "epoch": 0.6063585189883454, "grad_norm": 4.752046585083008, "learning_rate": 7.110432268450571e-06, "loss": 5.408, "step": 17208 }, { "epoch": 0.6063937560013037, "grad_norm": 1.7027058601379395, "learning_rate": 7.1093396948891305e-06, "loss": 0.8698, "step": 17209 }, { "epoch": 0.6064289930142622, "grad_norm": 9.202550888061523, "learning_rate": 7.1082471589789085e-06, "loss": 5.6016, "step": 17210 }, { "epoch": 0.6064642300272206, "grad_norm": 6.426397800445557, "learning_rate": 7.107154660734134e-06, "loss": 4.8591, "step": 17211 }, { "epoch": 0.606499467040179, "grad_norm": 1.1796807050704956, "learning_rate": 7.106062200169037e-06, "loss": 1.1539, "step": 17212 }, { "epoch": 0.6065347040531374, "grad_norm": 3.4235713481903076, "learning_rate": 7.104969777297855e-06, "loss": 3.4057, "step": 17213 }, { "epoch": 0.6065699410660959, "grad_norm": 5.007631778717041, "learning_rate": 7.103877392134803e-06, "loss": 4.676, "step": 17214 }, { "epoch": 0.6066051780790542, "grad_norm": 21.63555908203125, "learning_rate": 7.102785044694119e-06, "loss": 3.1298, "step": 17215 }, { "epoch": 0.6066404150920126, "grad_norm": 5.595263481140137, "learning_rate": 7.101692734990031e-06, "loss": 2.9434, "step": 17216 }, { "epoch": 0.6066756521049711, "grad_norm": 4.966416358947754, "learning_rate": 7.1006004630367645e-06, "loss": 3.2918, "step": 17217 }, { "epoch": 0.6067108891179295, "grad_norm": 7.090332508087158, "learning_rate": 7.099508228848542e-06, "loss": 5.2232, "step": 17218 }, { "epoch": 0.6067461261308879, "grad_norm": 1.6762410402297974, "learning_rate": 7.098416032439599e-06, "loss": 0.7865, "step": 17219 }, { "epoch": 0.6067813631438463, "grad_norm": 1.5137454271316528, "learning_rate": 7.097323873824153e-06, "loss": 1.0062, "step": 17220 }, { "epoch": 0.6068166001568047, "grad_norm": 1.0824105739593506, "learning_rate": 7.096231753016436e-06, "loss": 0.7404, "step": 17221 }, { "epoch": 0.6068518371697631, "grad_norm": 2.794527769088745, "learning_rate": 7.09513967003067e-06, "loss": 0.8535, "step": 17222 }, { "epoch": 0.6068870741827215, "grad_norm": 6.330441474914551, "learning_rate": 7.094047624881077e-06, "loss": 5.0298, "step": 17223 }, { "epoch": 0.60692231119568, "grad_norm": 0.9443989396095276, "learning_rate": 7.092955617581887e-06, "loss": 0.8664, "step": 17224 }, { "epoch": 0.6069575482086383, "grad_norm": 5.083337306976318, "learning_rate": 7.091863648147321e-06, "loss": 3.2835, "step": 17225 }, { "epoch": 0.6069927852215967, "grad_norm": 6.7723307609558105, "learning_rate": 7.090771716591599e-06, "loss": 5.1693, "step": 17226 }, { "epoch": 0.6070280222345552, "grad_norm": 1.0706284046173096, "learning_rate": 7.089679822928945e-06, "loss": 1.065, "step": 17227 }, { "epoch": 0.6070632592475136, "grad_norm": 2.180605411529541, "learning_rate": 7.088587967173584e-06, "loss": 1.1457, "step": 17228 }, { "epoch": 0.607098496260472, "grad_norm": 4.606034278869629, "learning_rate": 7.087496149339736e-06, "loss": 3.1474, "step": 17229 }, { "epoch": 0.6071337332734305, "grad_norm": 0.9919105172157288, "learning_rate": 7.0864043694416215e-06, "loss": 0.715, "step": 17230 }, { "epoch": 0.6071689702863888, "grad_norm": 4.0596418380737305, "learning_rate": 7.0853126274934584e-06, "loss": 3.0185, "step": 17231 }, { "epoch": 0.6072042072993472, "grad_norm": 1.5777487754821777, "learning_rate": 7.084220923509476e-06, "loss": 0.7897, "step": 17232 }, { "epoch": 0.6072394443123057, "grad_norm": 4.442471981048584, "learning_rate": 7.08312925750388e-06, "loss": 3.0024, "step": 17233 }, { "epoch": 0.6072746813252641, "grad_norm": 3.0469210147857666, "learning_rate": 7.082037629490901e-06, "loss": 3.2175, "step": 17234 }, { "epoch": 0.6073099183382225, "grad_norm": 1.8813807964324951, "learning_rate": 7.080946039484755e-06, "loss": 0.8977, "step": 17235 }, { "epoch": 0.6073451553511808, "grad_norm": 2.7385430335998535, "learning_rate": 7.079854487499657e-06, "loss": 2.6999, "step": 17236 }, { "epoch": 0.6073803923641393, "grad_norm": 4.782470703125, "learning_rate": 7.078762973549823e-06, "loss": 2.9737, "step": 17237 }, { "epoch": 0.6074156293770977, "grad_norm": 4.985039710998535, "learning_rate": 7.077671497649479e-06, "loss": 5.6541, "step": 17238 }, { "epoch": 0.6074508663900561, "grad_norm": 2.606377363204956, "learning_rate": 7.076580059812832e-06, "loss": 0.7385, "step": 17239 }, { "epoch": 0.6074861034030146, "grad_norm": 5.261284828186035, "learning_rate": 7.075488660054104e-06, "loss": 4.5524, "step": 17240 }, { "epoch": 0.6075213404159729, "grad_norm": 1.01321542263031, "learning_rate": 7.07439729838751e-06, "loss": 0.6409, "step": 17241 }, { "epoch": 0.6075565774289313, "grad_norm": 2.7749531269073486, "learning_rate": 7.0733059748272585e-06, "loss": 3.1058, "step": 17242 }, { "epoch": 0.6075918144418898, "grad_norm": 3.7230112552642822, "learning_rate": 7.0722146893875735e-06, "loss": 3.2546, "step": 17243 }, { "epoch": 0.6076270514548482, "grad_norm": 5.293024063110352, "learning_rate": 7.071123442082665e-06, "loss": 3.2657, "step": 17244 }, { "epoch": 0.6076622884678066, "grad_norm": 1.328007459640503, "learning_rate": 7.0700322329267465e-06, "loss": 1.0011, "step": 17245 }, { "epoch": 0.6076975254807649, "grad_norm": 2.0240914821624756, "learning_rate": 7.068941061934029e-06, "loss": 0.7651, "step": 17246 }, { "epoch": 0.6077327624937234, "grad_norm": 2.9020771980285645, "learning_rate": 7.06784992911873e-06, "loss": 2.9122, "step": 17247 }, { "epoch": 0.6077679995066818, "grad_norm": 1.0652974843978882, "learning_rate": 7.06675883449506e-06, "loss": 0.8245, "step": 17248 }, { "epoch": 0.6078032365196402, "grad_norm": 1.273232340812683, "learning_rate": 7.06566777807723e-06, "loss": 0.769, "step": 17249 }, { "epoch": 0.6078384735325987, "grad_norm": 2.498511791229248, "learning_rate": 7.0645767598794465e-06, "loss": 0.6859, "step": 17250 }, { "epoch": 0.6078737105455571, "grad_norm": 3.668375015258789, "learning_rate": 7.063485779915932e-06, "loss": 2.6955, "step": 17251 }, { "epoch": 0.6079089475585154, "grad_norm": 1.7770482301712036, "learning_rate": 7.062394838200882e-06, "loss": 1.0136, "step": 17252 }, { "epoch": 0.6079441845714739, "grad_norm": 5.627606391906738, "learning_rate": 7.061303934748516e-06, "loss": 4.8893, "step": 17253 }, { "epoch": 0.6079794215844323, "grad_norm": 3.369931221008301, "learning_rate": 7.060213069573043e-06, "loss": 3.044, "step": 17254 }, { "epoch": 0.6080146585973907, "grad_norm": 2.4360158443450928, "learning_rate": 7.059122242688667e-06, "loss": 0.8634, "step": 17255 }, { "epoch": 0.6080498956103491, "grad_norm": 15.708168029785156, "learning_rate": 7.0580314541095976e-06, "loss": 5.3129, "step": 17256 }, { "epoch": 0.6080851326233075, "grad_norm": 1.3222861289978027, "learning_rate": 7.056940703850046e-06, "loss": 1.0096, "step": 17257 }, { "epoch": 0.6081203696362659, "grad_norm": 1.409584879875183, "learning_rate": 7.055849991924214e-06, "loss": 0.8691, "step": 17258 }, { "epoch": 0.6081556066492243, "grad_norm": 5.296816349029541, "learning_rate": 7.054759318346312e-06, "loss": 3.1355, "step": 17259 }, { "epoch": 0.6081908436621828, "grad_norm": 1.635699987411499, "learning_rate": 7.053668683130547e-06, "loss": 0.8277, "step": 17260 }, { "epoch": 0.6082260806751412, "grad_norm": 1.5954959392547607, "learning_rate": 7.052578086291117e-06, "loss": 0.9637, "step": 17261 }, { "epoch": 0.6082613176880995, "grad_norm": 2.0315732955932617, "learning_rate": 7.0514875278422365e-06, "loss": 1.709, "step": 17262 }, { "epoch": 0.608296554701058, "grad_norm": 1.0614113807678223, "learning_rate": 7.050397007798107e-06, "loss": 0.984, "step": 17263 }, { "epoch": 0.6083317917140164, "grad_norm": 5.933357238769531, "learning_rate": 7.049306526172933e-06, "loss": 4.8528, "step": 17264 }, { "epoch": 0.6083670287269748, "grad_norm": 40.954795837402344, "learning_rate": 7.048216082980914e-06, "loss": 3.1982, "step": 17265 }, { "epoch": 0.6084022657399333, "grad_norm": 7.625871181488037, "learning_rate": 7.047125678236259e-06, "loss": 5.363, "step": 17266 }, { "epoch": 0.6084375027528917, "grad_norm": 6.097087860107422, "learning_rate": 7.0460353119531686e-06, "loss": 5.0472, "step": 17267 }, { "epoch": 0.60847273976585, "grad_norm": 4.836888313293457, "learning_rate": 7.044944984145844e-06, "loss": 5.3395, "step": 17268 }, { "epoch": 0.6085079767788084, "grad_norm": 3.6574037075042725, "learning_rate": 7.043854694828484e-06, "loss": 2.4782, "step": 17269 }, { "epoch": 0.6085432137917669, "grad_norm": 3.371607542037964, "learning_rate": 7.0427644440153e-06, "loss": 3.132, "step": 17270 }, { "epoch": 0.6085784508047253, "grad_norm": 7.217452526092529, "learning_rate": 7.04167423172048e-06, "loss": 5.6617, "step": 17271 }, { "epoch": 0.6086136878176837, "grad_norm": 3.8014450073242188, "learning_rate": 7.0405840579582315e-06, "loss": 4.6385, "step": 17272 }, { "epoch": 0.6086489248306421, "grad_norm": 2.517051935195923, "learning_rate": 7.039493922742755e-06, "loss": 0.7681, "step": 17273 }, { "epoch": 0.6086841618436005, "grad_norm": 1.5868483781814575, "learning_rate": 7.038403826088245e-06, "loss": 0.8903, "step": 17274 }, { "epoch": 0.6087193988565589, "grad_norm": 3.940199375152588, "learning_rate": 7.037313768008901e-06, "loss": 2.6222, "step": 17275 }, { "epoch": 0.6087546358695174, "grad_norm": 1.401789665222168, "learning_rate": 7.036223748518927e-06, "loss": 0.7269, "step": 17276 }, { "epoch": 0.6087898728824758, "grad_norm": 1.5721354484558105, "learning_rate": 7.0351337676325125e-06, "loss": 0.8806, "step": 17277 }, { "epoch": 0.6088251098954341, "grad_norm": 5.040294170379639, "learning_rate": 7.034043825363859e-06, "loss": 3.2589, "step": 17278 }, { "epoch": 0.6088603469083925, "grad_norm": 1.8761107921600342, "learning_rate": 7.0329539217271645e-06, "loss": 1.0993, "step": 17279 }, { "epoch": 0.608895583921351, "grad_norm": 1.2761905193328857, "learning_rate": 7.031864056736619e-06, "loss": 0.5809, "step": 17280 }, { "epoch": 0.6089308209343094, "grad_norm": 5.395557403564453, "learning_rate": 7.0307742304064255e-06, "loss": 2.9, "step": 17281 }, { "epoch": 0.6089660579472678, "grad_norm": 1.193516492843628, "learning_rate": 7.029684442750775e-06, "loss": 0.7925, "step": 17282 }, { "epoch": 0.6090012949602263, "grad_norm": 2.5726399421691895, "learning_rate": 7.028594693783863e-06, "loss": 1.2854, "step": 17283 }, { "epoch": 0.6090365319731846, "grad_norm": 2.5424389839172363, "learning_rate": 7.027504983519883e-06, "loss": 2.4266, "step": 17284 }, { "epoch": 0.609071768986143, "grad_norm": 1.6736807823181152, "learning_rate": 7.0264153119730294e-06, "loss": 0.9735, "step": 17285 }, { "epoch": 0.6091070059991015, "grad_norm": 1.1911083459854126, "learning_rate": 7.025325679157497e-06, "loss": 0.7399, "step": 17286 }, { "epoch": 0.6091422430120599, "grad_norm": 4.081936836242676, "learning_rate": 7.024236085087475e-06, "loss": 2.5789, "step": 17287 }, { "epoch": 0.6091774800250183, "grad_norm": 5.295891284942627, "learning_rate": 7.023146529777155e-06, "loss": 3.0613, "step": 17288 }, { "epoch": 0.6092127170379766, "grad_norm": 4.318723201751709, "learning_rate": 7.022057013240738e-06, "loss": 3.2104, "step": 17289 }, { "epoch": 0.6092479540509351, "grad_norm": 3.2107300758361816, "learning_rate": 7.020967535492399e-06, "loss": 3.1864, "step": 17290 }, { "epoch": 0.6092831910638935, "grad_norm": 5.96823787689209, "learning_rate": 7.019878096546341e-06, "loss": 5.4822, "step": 17291 }, { "epoch": 0.6093184280768519, "grad_norm": 1.6854933500289917, "learning_rate": 7.018788696416751e-06, "loss": 1.1596, "step": 17292 }, { "epoch": 0.6093536650898104, "grad_norm": 0.7608400583267212, "learning_rate": 7.017699335117818e-06, "loss": 0.8032, "step": 17293 }, { "epoch": 0.6093889021027687, "grad_norm": 1.0802130699157715, "learning_rate": 7.0166100126637284e-06, "loss": 0.9697, "step": 17294 }, { "epoch": 0.6094241391157271, "grad_norm": 15.494024276733398, "learning_rate": 7.015520729068677e-06, "loss": 5.7633, "step": 17295 }, { "epoch": 0.6094593761286856, "grad_norm": 4.054077625274658, "learning_rate": 7.014431484346846e-06, "loss": 2.946, "step": 17296 }, { "epoch": 0.609494613141644, "grad_norm": 1.7808032035827637, "learning_rate": 7.013342278512425e-06, "loss": 0.8726, "step": 17297 }, { "epoch": 0.6095298501546024, "grad_norm": 4.648444652557373, "learning_rate": 7.012253111579606e-06, "loss": 2.7884, "step": 17298 }, { "epoch": 0.6095650871675609, "grad_norm": 1.6240736246109009, "learning_rate": 7.011163983562563e-06, "loss": 0.8205, "step": 17299 }, { "epoch": 0.6096003241805192, "grad_norm": 1.0622987747192383, "learning_rate": 7.010074894475494e-06, "loss": 0.89, "step": 17300 }, { "epoch": 0.6096355611934776, "grad_norm": 3.347517967224121, "learning_rate": 7.00898584433258e-06, "loss": 2.4803, "step": 17301 }, { "epoch": 0.609670798206436, "grad_norm": 1.6881169080734253, "learning_rate": 7.007896833148008e-06, "loss": 0.7772, "step": 17302 }, { "epoch": 0.6097060352193945, "grad_norm": 1.4343990087509155, "learning_rate": 7.006807860935957e-06, "loss": 0.9251, "step": 17303 }, { "epoch": 0.6097412722323529, "grad_norm": 0.8665816187858582, "learning_rate": 7.005718927710616e-06, "loss": 0.9414, "step": 17304 }, { "epoch": 0.6097765092453112, "grad_norm": 2.5240378379821777, "learning_rate": 7.0046300334861705e-06, "loss": 3.1105, "step": 17305 }, { "epoch": 0.6098117462582697, "grad_norm": 4.604873180389404, "learning_rate": 7.003541178276799e-06, "loss": 2.6216, "step": 17306 }, { "epoch": 0.6098469832712281, "grad_norm": 3.5367624759674072, "learning_rate": 7.002452362096682e-06, "loss": 3.2093, "step": 17307 }, { "epoch": 0.6098822202841865, "grad_norm": 0.9690883755683899, "learning_rate": 7.001363584960012e-06, "loss": 0.878, "step": 17308 }, { "epoch": 0.609917457297145, "grad_norm": 9.303153038024902, "learning_rate": 7.000274846880957e-06, "loss": 2.5401, "step": 17309 }, { "epoch": 0.6099526943101033, "grad_norm": 4.681835651397705, "learning_rate": 6.999186147873706e-06, "loss": 3.1941, "step": 17310 }, { "epoch": 0.6099879313230617, "grad_norm": 1.2485027313232422, "learning_rate": 6.998097487952438e-06, "loss": 0.997, "step": 17311 }, { "epoch": 0.6100231683360201, "grad_norm": 3.526505947113037, "learning_rate": 6.997008867131333e-06, "loss": 2.5575, "step": 17312 }, { "epoch": 0.6100584053489786, "grad_norm": 2.8393447399139404, "learning_rate": 6.995920285424567e-06, "loss": 2.4357, "step": 17313 }, { "epoch": 0.610093642361937, "grad_norm": 3.9857869148254395, "learning_rate": 6.994831742846326e-06, "loss": 3.0945, "step": 17314 }, { "epoch": 0.6101288793748953, "grad_norm": 3.482013463973999, "learning_rate": 6.993743239410783e-06, "loss": 2.4182, "step": 17315 }, { "epoch": 0.6101641163878538, "grad_norm": 2.2384955883026123, "learning_rate": 6.992654775132118e-06, "loss": 0.7853, "step": 17316 }, { "epoch": 0.6101993534008122, "grad_norm": 3.096876621246338, "learning_rate": 6.991566350024507e-06, "loss": 2.4408, "step": 17317 }, { "epoch": 0.6102345904137706, "grad_norm": 1.1457406282424927, "learning_rate": 6.9904779641021246e-06, "loss": 0.7515, "step": 17318 }, { "epoch": 0.6102698274267291, "grad_norm": 5.5787763595581055, "learning_rate": 6.989389617379152e-06, "loss": 2.889, "step": 17319 }, { "epoch": 0.6103050644396875, "grad_norm": 3.923164129257202, "learning_rate": 6.988301309869764e-06, "loss": 3.3372, "step": 17320 }, { "epoch": 0.6103403014526458, "grad_norm": 6.064996242523193, "learning_rate": 6.9872130415881365e-06, "loss": 4.8467, "step": 17321 }, { "epoch": 0.6103755384656042, "grad_norm": 1.6959104537963867, "learning_rate": 6.986124812548438e-06, "loss": 0.9339, "step": 17322 }, { "epoch": 0.6104107754785627, "grad_norm": 3.9700167179107666, "learning_rate": 6.985036622764851e-06, "loss": 0.8123, "step": 17323 }, { "epoch": 0.6104460124915211, "grad_norm": 5.887316703796387, "learning_rate": 6.9839484722515474e-06, "loss": 5.0305, "step": 17324 }, { "epoch": 0.6104812495044795, "grad_norm": 5.620492935180664, "learning_rate": 6.982860361022699e-06, "loss": 5.0662, "step": 17325 }, { "epoch": 0.610516486517438, "grad_norm": 2.3740360736846924, "learning_rate": 6.981772289092474e-06, "loss": 0.8069, "step": 17326 }, { "epoch": 0.6105517235303963, "grad_norm": 1.105076551437378, "learning_rate": 6.980684256475056e-06, "loss": 1.0392, "step": 17327 }, { "epoch": 0.6105869605433547, "grad_norm": 5.59970235824585, "learning_rate": 6.979596263184605e-06, "loss": 5.3376, "step": 17328 }, { "epoch": 0.6106221975563132, "grad_norm": 1.8390798568725586, "learning_rate": 6.978508309235299e-06, "loss": 0.9923, "step": 17329 }, { "epoch": 0.6106574345692716, "grad_norm": 5.8779449462890625, "learning_rate": 6.97742039464131e-06, "loss": 2.6721, "step": 17330 }, { "epoch": 0.61069267158223, "grad_norm": 4.033367156982422, "learning_rate": 6.9763325194168015e-06, "loss": 3.2215, "step": 17331 }, { "epoch": 0.6107279085951884, "grad_norm": 2.320481777191162, "learning_rate": 6.975244683575946e-06, "loss": 1.0662, "step": 17332 }, { "epoch": 0.6107631456081468, "grad_norm": 4.291075706481934, "learning_rate": 6.974156887132919e-06, "loss": 2.3095, "step": 17333 }, { "epoch": 0.6107983826211052, "grad_norm": 5.242997646331787, "learning_rate": 6.97306913010188e-06, "loss": 2.8215, "step": 17334 }, { "epoch": 0.6108336196340636, "grad_norm": 6.023749828338623, "learning_rate": 6.971981412497002e-06, "loss": 4.8585, "step": 17335 }, { "epoch": 0.6108688566470221, "grad_norm": 9.276244163513184, "learning_rate": 6.97089373433245e-06, "loss": 4.5873, "step": 17336 }, { "epoch": 0.6109040936599804, "grad_norm": 1.1199842691421509, "learning_rate": 6.969806095622398e-06, "loss": 0.9982, "step": 17337 }, { "epoch": 0.6109393306729388, "grad_norm": 1.1809990406036377, "learning_rate": 6.968718496381004e-06, "loss": 0.9483, "step": 17338 }, { "epoch": 0.6109745676858973, "grad_norm": 4.060452461242676, "learning_rate": 6.967630936622438e-06, "loss": 4.6514, "step": 17339 }, { "epoch": 0.6110098046988557, "grad_norm": 2.500288486480713, "learning_rate": 6.966543416360867e-06, "loss": 2.9018, "step": 17340 }, { "epoch": 0.6110450417118141, "grad_norm": 3.7833259105682373, "learning_rate": 6.965455935610451e-06, "loss": 2.5071, "step": 17341 }, { "epoch": 0.6110802787247726, "grad_norm": 6.122189998626709, "learning_rate": 6.964368494385361e-06, "loss": 4.5382, "step": 17342 }, { "epoch": 0.6111155157377309, "grad_norm": 7.23016357421875, "learning_rate": 6.963281092699759e-06, "loss": 5.7607, "step": 17343 }, { "epoch": 0.6111507527506893, "grad_norm": 3.479492664337158, "learning_rate": 6.962193730567804e-06, "loss": 2.9262, "step": 17344 }, { "epoch": 0.6111859897636477, "grad_norm": 5.046010971069336, "learning_rate": 6.961106408003664e-06, "loss": 3.3155, "step": 17345 }, { "epoch": 0.6112212267766062, "grad_norm": 34.75667190551758, "learning_rate": 6.960019125021502e-06, "loss": 5.135, "step": 17346 }, { "epoch": 0.6112564637895646, "grad_norm": 1.4799392223358154, "learning_rate": 6.958931881635476e-06, "loss": 0.9125, "step": 17347 }, { "epoch": 0.6112917008025229, "grad_norm": 4.2424211502075195, "learning_rate": 6.9578446778597485e-06, "loss": 0.7018, "step": 17348 }, { "epoch": 0.6113269378154814, "grad_norm": 1.7603033781051636, "learning_rate": 6.956757513708485e-06, "loss": 1.0277, "step": 17349 }, { "epoch": 0.6113621748284398, "grad_norm": 1.1970847845077515, "learning_rate": 6.955670389195841e-06, "loss": 1.0594, "step": 17350 }, { "epoch": 0.6113974118413982, "grad_norm": 0.8619364500045776, "learning_rate": 6.9545833043359755e-06, "loss": 0.9933, "step": 17351 }, { "epoch": 0.6114326488543567, "grad_norm": 4.420737266540527, "learning_rate": 6.953496259143054e-06, "loss": 3.2027, "step": 17352 }, { "epoch": 0.611467885867315, "grad_norm": 1.1057208776474, "learning_rate": 6.952409253631232e-06, "loss": 0.8455, "step": 17353 }, { "epoch": 0.6115031228802734, "grad_norm": 7.618278980255127, "learning_rate": 6.951322287814664e-06, "loss": 7.949, "step": 17354 }, { "epoch": 0.6115383598932318, "grad_norm": 1.4347593784332275, "learning_rate": 6.950235361707511e-06, "loss": 1.0183, "step": 17355 }, { "epoch": 0.6115735969061903, "grad_norm": 5.103239059448242, "learning_rate": 6.949148475323939e-06, "loss": 6.8998, "step": 17356 }, { "epoch": 0.6116088339191487, "grad_norm": 2.1910574436187744, "learning_rate": 6.9480616286780886e-06, "loss": 2.2902, "step": 17357 }, { "epoch": 0.611644070932107, "grad_norm": 1.206154465675354, "learning_rate": 6.946974821784127e-06, "loss": 0.8626, "step": 17358 }, { "epoch": 0.6116793079450655, "grad_norm": 1.214013934135437, "learning_rate": 6.945888054656209e-06, "loss": 1.1962, "step": 17359 }, { "epoch": 0.6117145449580239, "grad_norm": 1.60706627368927, "learning_rate": 6.944801327308485e-06, "loss": 0.8065, "step": 17360 }, { "epoch": 0.6117497819709823, "grad_norm": 1.169999599456787, "learning_rate": 6.9437146397551125e-06, "loss": 0.8015, "step": 17361 }, { "epoch": 0.6117850189839408, "grad_norm": 1.1458885669708252, "learning_rate": 6.942627992010249e-06, "loss": 0.8823, "step": 17362 }, { "epoch": 0.6118202559968992, "grad_norm": 6.217249870300293, "learning_rate": 6.941541384088044e-06, "loss": 5.1111, "step": 17363 }, { "epoch": 0.6118554930098575, "grad_norm": 1.1351513862609863, "learning_rate": 6.9404548160026515e-06, "loss": 1.1082, "step": 17364 }, { "epoch": 0.611890730022816, "grad_norm": 2.1089653968811035, "learning_rate": 6.939368287768227e-06, "loss": 0.9787, "step": 17365 }, { "epoch": 0.6119259670357744, "grad_norm": 1.3767811059951782, "learning_rate": 6.9382817993989185e-06, "loss": 0.9801, "step": 17366 }, { "epoch": 0.6119612040487328, "grad_norm": 0.9772698879241943, "learning_rate": 6.937195350908879e-06, "loss": 0.7136, "step": 17367 }, { "epoch": 0.6119964410616912, "grad_norm": 4.216664791107178, "learning_rate": 6.936108942312263e-06, "loss": 2.7398, "step": 17368 }, { "epoch": 0.6120316780746496, "grad_norm": 2.0970253944396973, "learning_rate": 6.935022573623216e-06, "loss": 0.7873, "step": 17369 }, { "epoch": 0.612066915087608, "grad_norm": 2.2352075576782227, "learning_rate": 6.933936244855887e-06, "loss": 0.8386, "step": 17370 }, { "epoch": 0.6121021521005664, "grad_norm": 4.9419708251953125, "learning_rate": 6.932849956024438e-06, "loss": 3.0514, "step": 17371 }, { "epoch": 0.6121373891135249, "grad_norm": 1.0605915784835815, "learning_rate": 6.931763707143001e-06, "loss": 0.6838, "step": 17372 }, { "epoch": 0.6121726261264833, "grad_norm": 1.5667251348495483, "learning_rate": 6.930677498225735e-06, "loss": 0.6854, "step": 17373 }, { "epoch": 0.6122078631394416, "grad_norm": 1.0418094396591187, "learning_rate": 6.929591329286784e-06, "loss": 0.7273, "step": 17374 }, { "epoch": 0.6122431001524001, "grad_norm": 3.6496999263763428, "learning_rate": 6.928505200340303e-06, "loss": 2.5975, "step": 17375 }, { "epoch": 0.6122783371653585, "grad_norm": 4.653679847717285, "learning_rate": 6.927419111400427e-06, "loss": 2.4257, "step": 17376 }, { "epoch": 0.6123135741783169, "grad_norm": 4.37910270690918, "learning_rate": 6.9263330624813095e-06, "loss": 2.6769, "step": 17377 }, { "epoch": 0.6123488111912753, "grad_norm": 6.036082744598389, "learning_rate": 6.925247053597099e-06, "loss": 3.1125, "step": 17378 }, { "epoch": 0.6123840482042338, "grad_norm": 1.8527586460113525, "learning_rate": 6.924161084761934e-06, "loss": 0.7392, "step": 17379 }, { "epoch": 0.6124192852171921, "grad_norm": 5.163527488708496, "learning_rate": 6.92307515598996e-06, "loss": 2.9716, "step": 17380 }, { "epoch": 0.6124545222301505, "grad_norm": 1.626037359237671, "learning_rate": 6.921989267295329e-06, "loss": 0.8651, "step": 17381 }, { "epoch": 0.612489759243109, "grad_norm": 4.571915149688721, "learning_rate": 6.920903418692179e-06, "loss": 4.8123, "step": 17382 }, { "epoch": 0.6125249962560674, "grad_norm": 1.3952763080596924, "learning_rate": 6.919817610194653e-06, "loss": 0.8949, "step": 17383 }, { "epoch": 0.6125602332690258, "grad_norm": 4.145780563354492, "learning_rate": 6.918731841816897e-06, "loss": 2.9167, "step": 17384 }, { "epoch": 0.6125954702819842, "grad_norm": 4.4524688720703125, "learning_rate": 6.917646113573047e-06, "loss": 2.7392, "step": 17385 }, { "epoch": 0.6126307072949426, "grad_norm": 1.2378911972045898, "learning_rate": 6.916560425477251e-06, "loss": 0.7626, "step": 17386 }, { "epoch": 0.612665944307901, "grad_norm": 1.3210093975067139, "learning_rate": 6.9154747775436516e-06, "loss": 0.8629, "step": 17387 }, { "epoch": 0.6127011813208594, "grad_norm": 8.114293098449707, "learning_rate": 6.914389169786382e-06, "loss": 2.8844, "step": 17388 }, { "epoch": 0.6127364183338179, "grad_norm": 1.6434392929077148, "learning_rate": 6.913303602219585e-06, "loss": 0.8991, "step": 17389 }, { "epoch": 0.6127716553467762, "grad_norm": 5.37001895904541, "learning_rate": 6.912218074857409e-06, "loss": 2.5092, "step": 17390 }, { "epoch": 0.6128068923597346, "grad_norm": 3.878288984298706, "learning_rate": 6.9111325877139805e-06, "loss": 2.4929, "step": 17391 }, { "epoch": 0.6128421293726931, "grad_norm": 6.786642074584961, "learning_rate": 6.910047140803443e-06, "loss": 3.0082, "step": 17392 }, { "epoch": 0.6128773663856515, "grad_norm": 1.2103569507598877, "learning_rate": 6.908961734139934e-06, "loss": 0.9739, "step": 17393 }, { "epoch": 0.6129126033986099, "grad_norm": 1.7149757146835327, "learning_rate": 6.9078763677375996e-06, "loss": 1.234, "step": 17394 }, { "epoch": 0.6129478404115684, "grad_norm": 1.6683846712112427, "learning_rate": 6.906791041610561e-06, "loss": 0.7034, "step": 17395 }, { "epoch": 0.6129830774245267, "grad_norm": 5.648796081542969, "learning_rate": 6.905705755772967e-06, "loss": 5.0717, "step": 17396 }, { "epoch": 0.6130183144374851, "grad_norm": 2.142897367477417, "learning_rate": 6.9046205102389505e-06, "loss": 1.2676, "step": 17397 }, { "epoch": 0.6130535514504436, "grad_norm": 5.482088088989258, "learning_rate": 6.903535305022645e-06, "loss": 3.4961, "step": 17398 }, { "epoch": 0.613088788463402, "grad_norm": 36.79518127441406, "learning_rate": 6.902450140138184e-06, "loss": 3.8315, "step": 17399 }, { "epoch": 0.6131240254763604, "grad_norm": 1.4377909898757935, "learning_rate": 6.901365015599709e-06, "loss": 0.8893, "step": 17400 }, { "epoch": 0.6131592624893187, "grad_norm": 0.841813325881958, "learning_rate": 6.900279931421347e-06, "loss": 0.8617, "step": 17401 }, { "epoch": 0.6131944995022772, "grad_norm": 3.7681643962860107, "learning_rate": 6.899194887617234e-06, "loss": 2.848, "step": 17402 }, { "epoch": 0.6132297365152356, "grad_norm": 3.583836555480957, "learning_rate": 6.898109884201505e-06, "loss": 3.1033, "step": 17403 }, { "epoch": 0.613264973528194, "grad_norm": 1.4838780164718628, "learning_rate": 6.897024921188285e-06, "loss": 0.8952, "step": 17404 }, { "epoch": 0.6133002105411525, "grad_norm": 3.7111666202545166, "learning_rate": 6.8959399985917134e-06, "loss": 2.7433, "step": 17405 }, { "epoch": 0.6133354475541108, "grad_norm": 10.166955947875977, "learning_rate": 6.8948551164259204e-06, "loss": 5.1662, "step": 17406 }, { "epoch": 0.6133706845670692, "grad_norm": 5.843210697174072, "learning_rate": 6.893770274705032e-06, "loss": 2.7923, "step": 17407 }, { "epoch": 0.6134059215800277, "grad_norm": 1.9550964832305908, "learning_rate": 6.892685473443181e-06, "loss": 0.803, "step": 17408 }, { "epoch": 0.6134411585929861, "grad_norm": 4.565337181091309, "learning_rate": 6.891600712654498e-06, "loss": 3.1849, "step": 17409 }, { "epoch": 0.6134763956059445, "grad_norm": 1.8414472341537476, "learning_rate": 6.890515992353115e-06, "loss": 0.9826, "step": 17410 }, { "epoch": 0.6135116326189028, "grad_norm": 1.2025926113128662, "learning_rate": 6.889431312553155e-06, "loss": 1.0051, "step": 17411 }, { "epoch": 0.6135468696318613, "grad_norm": 0.9801995158195496, "learning_rate": 6.888346673268747e-06, "loss": 0.9071, "step": 17412 }, { "epoch": 0.6135821066448197, "grad_norm": 10.077570915222168, "learning_rate": 6.887262074514025e-06, "loss": 4.4235, "step": 17413 }, { "epoch": 0.6136173436577781, "grad_norm": 1.5726103782653809, "learning_rate": 6.886177516303104e-06, "loss": 0.79, "step": 17414 }, { "epoch": 0.6136525806707366, "grad_norm": 1.4104315042495728, "learning_rate": 6.88509299865012e-06, "loss": 0.8895, "step": 17415 }, { "epoch": 0.613687817683695, "grad_norm": 0.8577379584312439, "learning_rate": 6.884008521569201e-06, "loss": 0.8132, "step": 17416 }, { "epoch": 0.6137230546966533, "grad_norm": 3.2891314029693604, "learning_rate": 6.8829240850744625e-06, "loss": 0.8711, "step": 17417 }, { "epoch": 0.6137582917096118, "grad_norm": 1.7951852083206177, "learning_rate": 6.8818396891800345e-06, "loss": 1.1766, "step": 17418 }, { "epoch": 0.6137935287225702, "grad_norm": 4.789584636688232, "learning_rate": 6.880755333900047e-06, "loss": 4.987, "step": 17419 }, { "epoch": 0.6138287657355286, "grad_norm": 2.7699756622314453, "learning_rate": 6.879671019248615e-06, "loss": 3.0038, "step": 17420 }, { "epoch": 0.613864002748487, "grad_norm": 1.0789412260055542, "learning_rate": 6.878586745239865e-06, "loss": 0.8768, "step": 17421 }, { "epoch": 0.6138992397614454, "grad_norm": 1.3511576652526855, "learning_rate": 6.877502511887924e-06, "loss": 0.9923, "step": 17422 }, { "epoch": 0.6139344767744038, "grad_norm": 1.818129062652588, "learning_rate": 6.876418319206903e-06, "loss": 0.9169, "step": 17423 }, { "epoch": 0.6139697137873622, "grad_norm": 1.6442209482192993, "learning_rate": 6.875334167210937e-06, "loss": 0.6991, "step": 17424 }, { "epoch": 0.6140049508003207, "grad_norm": 4.269176959991455, "learning_rate": 6.874250055914143e-06, "loss": 2.9081, "step": 17425 }, { "epoch": 0.6140401878132791, "grad_norm": 1.2987511157989502, "learning_rate": 6.873165985330636e-06, "loss": 0.9966, "step": 17426 }, { "epoch": 0.6140754248262374, "grad_norm": 1.3110225200653076, "learning_rate": 6.87208195547454e-06, "loss": 1.0225, "step": 17427 }, { "epoch": 0.6141106618391959, "grad_norm": 1.9686510562896729, "learning_rate": 6.870997966359977e-06, "loss": 2.4347, "step": 17428 }, { "epoch": 0.6141458988521543, "grad_norm": 4.00609827041626, "learning_rate": 6.869914018001065e-06, "loss": 3.0865, "step": 17429 }, { "epoch": 0.6141811358651127, "grad_norm": 4.335239887237549, "learning_rate": 6.868830110411922e-06, "loss": 2.5831, "step": 17430 }, { "epoch": 0.6142163728780711, "grad_norm": 1.5677191019058228, "learning_rate": 6.867746243606661e-06, "loss": 0.7119, "step": 17431 }, { "epoch": 0.6142516098910296, "grad_norm": 2.054391384124756, "learning_rate": 6.8666624175994115e-06, "loss": 0.9812, "step": 17432 }, { "epoch": 0.6142868469039879, "grad_norm": 3.21832013130188, "learning_rate": 6.8655786324042765e-06, "loss": 2.7995, "step": 17433 }, { "epoch": 0.6143220839169463, "grad_norm": 5.035463809967041, "learning_rate": 6.864494888035382e-06, "loss": 3.0822, "step": 17434 }, { "epoch": 0.6143573209299048, "grad_norm": 1.2929012775421143, "learning_rate": 6.863411184506841e-06, "loss": 0.7464, "step": 17435 }, { "epoch": 0.6143925579428632, "grad_norm": 8.586665153503418, "learning_rate": 6.862327521832768e-06, "loss": 6.6183, "step": 17436 }, { "epoch": 0.6144277949558216, "grad_norm": 1.4369243383407593, "learning_rate": 6.861243900027277e-06, "loss": 0.9967, "step": 17437 }, { "epoch": 0.61446303196878, "grad_norm": 5.001270771026611, "learning_rate": 6.860160319104489e-06, "loss": 5.2357, "step": 17438 }, { "epoch": 0.6144982689817384, "grad_norm": 5.745979309082031, "learning_rate": 6.859076779078509e-06, "loss": 5.313, "step": 17439 }, { "epoch": 0.6145335059946968, "grad_norm": 1.448654055595398, "learning_rate": 6.857993279963455e-06, "loss": 0.8131, "step": 17440 }, { "epoch": 0.6145687430076553, "grad_norm": 7.236717224121094, "learning_rate": 6.8569098217734405e-06, "loss": 3.5922, "step": 17441 }, { "epoch": 0.6146039800206137, "grad_norm": 1.2190821170806885, "learning_rate": 6.855826404522572e-06, "loss": 0.7786, "step": 17442 }, { "epoch": 0.614639217033572, "grad_norm": 4.057336807250977, "learning_rate": 6.8547430282249665e-06, "loss": 2.5783, "step": 17443 }, { "epoch": 0.6146744540465304, "grad_norm": 6.106607437133789, "learning_rate": 6.853659692894736e-06, "loss": 3.1081, "step": 17444 }, { "epoch": 0.6147096910594889, "grad_norm": 5.4010515213012695, "learning_rate": 6.852576398545987e-06, "loss": 3.4006, "step": 17445 }, { "epoch": 0.6147449280724473, "grad_norm": 5.704502582550049, "learning_rate": 6.85149314519283e-06, "loss": 3.0077, "step": 17446 }, { "epoch": 0.6147801650854057, "grad_norm": 5.429331302642822, "learning_rate": 6.850409932849378e-06, "loss": 5.0207, "step": 17447 }, { "epoch": 0.6148154020983642, "grad_norm": 3.902707099914551, "learning_rate": 6.8493267615297376e-06, "loss": 3.2854, "step": 17448 }, { "epoch": 0.6148506391113225, "grad_norm": 4.736166477203369, "learning_rate": 6.8482436312480175e-06, "loss": 4.8194, "step": 17449 }, { "epoch": 0.6148858761242809, "grad_norm": 5.072999000549316, "learning_rate": 6.847160542018323e-06, "loss": 5.1433, "step": 17450 }, { "epoch": 0.6149211131372394, "grad_norm": 1.037579894065857, "learning_rate": 6.8460774938547706e-06, "loss": 1.0355, "step": 17451 }, { "epoch": 0.6149563501501978, "grad_norm": 1.1769003868103027, "learning_rate": 6.844994486771453e-06, "loss": 0.9637, "step": 17452 }, { "epoch": 0.6149915871631562, "grad_norm": 3.020798683166504, "learning_rate": 6.843911520782487e-06, "loss": 2.9855, "step": 17453 }, { "epoch": 0.6150268241761145, "grad_norm": 1.8196353912353516, "learning_rate": 6.842828595901978e-06, "loss": 0.8609, "step": 17454 }, { "epoch": 0.615062061189073, "grad_norm": 1.2451750040054321, "learning_rate": 6.841745712144026e-06, "loss": 0.7095, "step": 17455 }, { "epoch": 0.6150972982020314, "grad_norm": 4.4608564376831055, "learning_rate": 6.8406628695227365e-06, "loss": 3.2771, "step": 17456 }, { "epoch": 0.6151325352149898, "grad_norm": 1.4195702075958252, "learning_rate": 6.83958006805222e-06, "loss": 0.7444, "step": 17457 }, { "epoch": 0.6151677722279483, "grad_norm": 4.788619041442871, "learning_rate": 6.8384973077465735e-06, "loss": 3.3735, "step": 17458 }, { "epoch": 0.6152030092409067, "grad_norm": 7.6540207862854, "learning_rate": 6.837414588619902e-06, "loss": 2.923, "step": 17459 }, { "epoch": 0.615238246253865, "grad_norm": 1.1383708715438843, "learning_rate": 6.836331910686311e-06, "loss": 0.9213, "step": 17460 }, { "epoch": 0.6152734832668235, "grad_norm": 1.2670023441314697, "learning_rate": 6.835249273959894e-06, "loss": 0.9104, "step": 17461 }, { "epoch": 0.6153087202797819, "grad_norm": 3.529855251312256, "learning_rate": 6.834166678454762e-06, "loss": 2.9625, "step": 17462 }, { "epoch": 0.6153439572927403, "grad_norm": 0.9925833940505981, "learning_rate": 6.833084124185014e-06, "loss": 0.6631, "step": 17463 }, { "epoch": 0.6153791943056987, "grad_norm": 2.9027788639068604, "learning_rate": 6.832001611164747e-06, "loss": 0.951, "step": 17464 }, { "epoch": 0.6154144313186571, "grad_norm": 6.112791538238525, "learning_rate": 6.830919139408061e-06, "loss": 3.0779, "step": 17465 }, { "epoch": 0.6154496683316155, "grad_norm": 7.377790451049805, "learning_rate": 6.829836708929057e-06, "loss": 5.3826, "step": 17466 }, { "epoch": 0.6154849053445739, "grad_norm": 8.293359756469727, "learning_rate": 6.828754319741837e-06, "loss": 4.877, "step": 17467 }, { "epoch": 0.6155201423575324, "grad_norm": 3.4019696712493896, "learning_rate": 6.827671971860495e-06, "loss": 2.9371, "step": 17468 }, { "epoch": 0.6155553793704908, "grad_norm": 10.233854293823242, "learning_rate": 6.826589665299125e-06, "loss": 4.8677, "step": 17469 }, { "epoch": 0.6155906163834491, "grad_norm": 2.6717920303344727, "learning_rate": 6.825507400071836e-06, "loss": 2.7845, "step": 17470 }, { "epoch": 0.6156258533964076, "grad_norm": 1.4351091384887695, "learning_rate": 6.824425176192713e-06, "loss": 0.6645, "step": 17471 }, { "epoch": 0.615661090409366, "grad_norm": 1.912916898727417, "learning_rate": 6.823342993675857e-06, "loss": 0.7743, "step": 17472 }, { "epoch": 0.6156963274223244, "grad_norm": 0.9492259621620178, "learning_rate": 6.822260852535365e-06, "loss": 1.0743, "step": 17473 }, { "epoch": 0.6157315644352829, "grad_norm": 3.3385791778564453, "learning_rate": 6.821178752785328e-06, "loss": 2.6385, "step": 17474 }, { "epoch": 0.6157668014482413, "grad_norm": 1.3331224918365479, "learning_rate": 6.8200966944398416e-06, "loss": 0.7858, "step": 17475 }, { "epoch": 0.6158020384611996, "grad_norm": 2.808723211288452, "learning_rate": 6.819014677513004e-06, "loss": 2.8388, "step": 17476 }, { "epoch": 0.615837275474158, "grad_norm": 1.1741193532943726, "learning_rate": 6.817932702018903e-06, "loss": 0.8677, "step": 17477 }, { "epoch": 0.6158725124871165, "grad_norm": 3.60068678855896, "learning_rate": 6.8168507679716354e-06, "loss": 2.8187, "step": 17478 }, { "epoch": 0.6159077495000749, "grad_norm": 1.1741480827331543, "learning_rate": 6.815768875385293e-06, "loss": 1.0506, "step": 17479 }, { "epoch": 0.6159429865130333, "grad_norm": 1.3535411357879639, "learning_rate": 6.814687024273964e-06, "loss": 0.699, "step": 17480 }, { "epoch": 0.6159782235259917, "grad_norm": 4.035964488983154, "learning_rate": 6.813605214651743e-06, "loss": 3.0903, "step": 17481 }, { "epoch": 0.6160134605389501, "grad_norm": 5.021096229553223, "learning_rate": 6.8125234465327195e-06, "loss": 2.9951, "step": 17482 }, { "epoch": 0.6160486975519085, "grad_norm": 6.0432844161987305, "learning_rate": 6.8114417199309866e-06, "loss": 3.1038, "step": 17483 }, { "epoch": 0.616083934564867, "grad_norm": 2.1908302307128906, "learning_rate": 6.810360034860628e-06, "loss": 0.7715, "step": 17484 }, { "epoch": 0.6161191715778254, "grad_norm": 7.938329696655273, "learning_rate": 6.809278391335739e-06, "loss": 5.2777, "step": 17485 }, { "epoch": 0.6161544085907837, "grad_norm": 5.6091203689575195, "learning_rate": 6.808196789370406e-06, "loss": 3.4021, "step": 17486 }, { "epoch": 0.6161896456037421, "grad_norm": 3.541353464126587, "learning_rate": 6.807115228978715e-06, "loss": 0.9292, "step": 17487 }, { "epoch": 0.6162248826167006, "grad_norm": 5.224154472351074, "learning_rate": 6.806033710174752e-06, "loss": 5.2617, "step": 17488 }, { "epoch": 0.616260119629659, "grad_norm": 3.282749891281128, "learning_rate": 6.804952232972613e-06, "loss": 2.6546, "step": 17489 }, { "epoch": 0.6162953566426174, "grad_norm": 4.765364170074463, "learning_rate": 6.803870797386372e-06, "loss": 4.8677, "step": 17490 }, { "epoch": 0.6163305936555759, "grad_norm": 4.715886116027832, "learning_rate": 6.8027894034301234e-06, "loss": 3.6481, "step": 17491 }, { "epoch": 0.6163658306685342, "grad_norm": 1.196161150932312, "learning_rate": 6.801708051117952e-06, "loss": 0.725, "step": 17492 }, { "epoch": 0.6164010676814926, "grad_norm": 4.586812496185303, "learning_rate": 6.80062674046394e-06, "loss": 2.6212, "step": 17493 }, { "epoch": 0.6164363046944511, "grad_norm": 2.5878870487213135, "learning_rate": 6.799545471482169e-06, "loss": 0.8086, "step": 17494 }, { "epoch": 0.6164715417074095, "grad_norm": 3.7559337615966797, "learning_rate": 6.798464244186729e-06, "loss": 2.8154, "step": 17495 }, { "epoch": 0.6165067787203679, "grad_norm": 6.242964267730713, "learning_rate": 6.7973830585917e-06, "loss": 5.0359, "step": 17496 }, { "epoch": 0.6165420157333262, "grad_norm": 1.1950043439865112, "learning_rate": 6.796301914711163e-06, "loss": 1.0595, "step": 17497 }, { "epoch": 0.6165772527462847, "grad_norm": 1.4164981842041016, "learning_rate": 6.795220812559205e-06, "loss": 1.0622, "step": 17498 }, { "epoch": 0.6166124897592431, "grad_norm": 4.482004165649414, "learning_rate": 6.794139752149899e-06, "loss": 3.0234, "step": 17499 }, { "epoch": 0.6166477267722015, "grad_norm": 1.141208529472351, "learning_rate": 6.793058733497333e-06, "loss": 0.7372, "step": 17500 }, { "epoch": 0.61668296378516, "grad_norm": 4.731905937194824, "learning_rate": 6.791977756615587e-06, "loss": 3.2536, "step": 17501 }, { "epoch": 0.6167182007981183, "grad_norm": 2.311244487762451, "learning_rate": 6.79089682151874e-06, "loss": 0.7601, "step": 17502 }, { "epoch": 0.6167534378110767, "grad_norm": 4.845560073852539, "learning_rate": 6.789815928220869e-06, "loss": 4.3765, "step": 17503 }, { "epoch": 0.6167886748240352, "grad_norm": 6.94791841506958, "learning_rate": 6.788735076736052e-06, "loss": 5.5597, "step": 17504 }, { "epoch": 0.6168239118369936, "grad_norm": 1.5367040634155273, "learning_rate": 6.787654267078374e-06, "loss": 1.1775, "step": 17505 }, { "epoch": 0.616859148849952, "grad_norm": 0.9216181635856628, "learning_rate": 6.786573499261906e-06, "loss": 0.8714, "step": 17506 }, { "epoch": 0.6168943858629105, "grad_norm": 5.61507511138916, "learning_rate": 6.7854927733007285e-06, "loss": 3.6803, "step": 17507 }, { "epoch": 0.6169296228758688, "grad_norm": 3.840740203857422, "learning_rate": 6.784412089208918e-06, "loss": 2.4716, "step": 17508 }, { "epoch": 0.6169648598888272, "grad_norm": 2.076555013656616, "learning_rate": 6.783331447000545e-06, "loss": 0.9401, "step": 17509 }, { "epoch": 0.6170000969017856, "grad_norm": 8.17323112487793, "learning_rate": 6.782250846689693e-06, "loss": 7.6852, "step": 17510 }, { "epoch": 0.6170353339147441, "grad_norm": 7.373654842376709, "learning_rate": 6.781170288290436e-06, "loss": 3.3574, "step": 17511 }, { "epoch": 0.6170705709277025, "grad_norm": 3.8831498622894287, "learning_rate": 6.780089771816844e-06, "loss": 0.7568, "step": 17512 }, { "epoch": 0.6171058079406608, "grad_norm": 1.6103945970535278, "learning_rate": 6.77900929728299e-06, "loss": 0.9052, "step": 17513 }, { "epoch": 0.6171410449536193, "grad_norm": 1.979495644569397, "learning_rate": 6.7779288647029565e-06, "loss": 0.8469, "step": 17514 }, { "epoch": 0.6171762819665777, "grad_norm": 2.8580265045166016, "learning_rate": 6.776848474090804e-06, "loss": 2.8854, "step": 17515 }, { "epoch": 0.6172115189795361, "grad_norm": 3.1372132301330566, "learning_rate": 6.775768125460612e-06, "loss": 2.7646, "step": 17516 }, { "epoch": 0.6172467559924946, "grad_norm": 1.827926516532898, "learning_rate": 6.774687818826453e-06, "loss": 1.1165, "step": 17517 }, { "epoch": 0.617281993005453, "grad_norm": 2.5903515815734863, "learning_rate": 6.773607554202395e-06, "loss": 2.6967, "step": 17518 }, { "epoch": 0.6173172300184113, "grad_norm": 7.551126956939697, "learning_rate": 6.772527331602507e-06, "loss": 7.5572, "step": 17519 }, { "epoch": 0.6173524670313697, "grad_norm": 7.043665409088135, "learning_rate": 6.771447151040864e-06, "loss": 2.788, "step": 17520 }, { "epoch": 0.6173877040443282, "grad_norm": 4.577453136444092, "learning_rate": 6.770367012531536e-06, "loss": 3.115, "step": 17521 }, { "epoch": 0.6174229410572866, "grad_norm": 4.044814586639404, "learning_rate": 6.7692869160885845e-06, "loss": 3.0127, "step": 17522 }, { "epoch": 0.617458178070245, "grad_norm": 1.4372392892837524, "learning_rate": 6.768206861726082e-06, "loss": 0.8995, "step": 17523 }, { "epoch": 0.6174934150832034, "grad_norm": 1.3360862731933594, "learning_rate": 6.767126849458101e-06, "loss": 0.9674, "step": 17524 }, { "epoch": 0.6175286520961618, "grad_norm": 1.1486643552780151, "learning_rate": 6.766046879298702e-06, "loss": 1.1294, "step": 17525 }, { "epoch": 0.6175638891091202, "grad_norm": 1.5718603134155273, "learning_rate": 6.764966951261954e-06, "loss": 0.7288, "step": 17526 }, { "epoch": 0.6175991261220787, "grad_norm": 3.0722815990448, "learning_rate": 6.763887065361927e-06, "loss": 2.8256, "step": 17527 }, { "epoch": 0.6176343631350371, "grad_norm": 4.8639326095581055, "learning_rate": 6.762807221612677e-06, "loss": 2.3928, "step": 17528 }, { "epoch": 0.6176696001479954, "grad_norm": 0.8887619972229004, "learning_rate": 6.761727420028279e-06, "loss": 1.3058, "step": 17529 }, { "epoch": 0.6177048371609538, "grad_norm": 0.9343754053115845, "learning_rate": 6.760647660622795e-06, "loss": 0.8371, "step": 17530 }, { "epoch": 0.6177400741739123, "grad_norm": 4.336442470550537, "learning_rate": 6.7595679434102866e-06, "loss": 3.306, "step": 17531 }, { "epoch": 0.6177753111868707, "grad_norm": 2.443004846572876, "learning_rate": 6.758488268404816e-06, "loss": 0.7771, "step": 17532 }, { "epoch": 0.6178105481998291, "grad_norm": 7.225245952606201, "learning_rate": 6.757408635620455e-06, "loss": 5.8042, "step": 17533 }, { "epoch": 0.6178457852127875, "grad_norm": 15.914131164550781, "learning_rate": 6.756329045071253e-06, "loss": 6.6386, "step": 17534 }, { "epoch": 0.6178810222257459, "grad_norm": 6.767190456390381, "learning_rate": 6.755249496771281e-06, "loss": 4.8993, "step": 17535 }, { "epoch": 0.6179162592387043, "grad_norm": 1.0751532316207886, "learning_rate": 6.754169990734599e-06, "loss": 1.0065, "step": 17536 }, { "epoch": 0.6179514962516628, "grad_norm": 5.924740314483643, "learning_rate": 6.753090526975266e-06, "loss": 4.6563, "step": 17537 }, { "epoch": 0.6179867332646212, "grad_norm": 1.086053490638733, "learning_rate": 6.752011105507338e-06, "loss": 0.9242, "step": 17538 }, { "epoch": 0.6180219702775795, "grad_norm": 9.344927787780762, "learning_rate": 6.750931726344882e-06, "loss": 3.1334, "step": 17539 }, { "epoch": 0.618057207290538, "grad_norm": 1.502110242843628, "learning_rate": 6.749852389501956e-06, "loss": 0.8697, "step": 17540 }, { "epoch": 0.6180924443034964, "grad_norm": 1.486223578453064, "learning_rate": 6.7487730949926156e-06, "loss": 1.0081, "step": 17541 }, { "epoch": 0.6181276813164548, "grad_norm": 1.4216923713684082, "learning_rate": 6.747693842830916e-06, "loss": 0.6809, "step": 17542 }, { "epoch": 0.6181629183294132, "grad_norm": 0.8025431036949158, "learning_rate": 6.7466146330309234e-06, "loss": 0.7782, "step": 17543 }, { "epoch": 0.6181981553423717, "grad_norm": 3.725560426712036, "learning_rate": 6.7455354656066885e-06, "loss": 2.3514, "step": 17544 }, { "epoch": 0.61823339235533, "grad_norm": 3.2975516319274902, "learning_rate": 6.744456340572267e-06, "loss": 2.8028, "step": 17545 }, { "epoch": 0.6182686293682884, "grad_norm": 6.618969917297363, "learning_rate": 6.74337725794172e-06, "loss": 3.2998, "step": 17546 }, { "epoch": 0.6183038663812469, "grad_norm": 1.6991827487945557, "learning_rate": 6.742298217729094e-06, "loss": 0.6681, "step": 17547 }, { "epoch": 0.6183391033942053, "grad_norm": 3.741051197052002, "learning_rate": 6.741219219948452e-06, "loss": 2.6268, "step": 17548 }, { "epoch": 0.6183743404071637, "grad_norm": 2.582754373550415, "learning_rate": 6.740140264613846e-06, "loss": 2.8774, "step": 17549 }, { "epoch": 0.6184095774201221, "grad_norm": 2.0374269485473633, "learning_rate": 6.739061351739325e-06, "loss": 0.8846, "step": 17550 }, { "epoch": 0.6184448144330805, "grad_norm": 0.9858195781707764, "learning_rate": 6.737982481338945e-06, "loss": 1.1991, "step": 17551 }, { "epoch": 0.6184800514460389, "grad_norm": 3.6236209869384766, "learning_rate": 6.7369036534267626e-06, "loss": 3.0328, "step": 17552 }, { "epoch": 0.6185152884589973, "grad_norm": 0.9583480358123779, "learning_rate": 6.735824868016823e-06, "loss": 0.8897, "step": 17553 }, { "epoch": 0.6185505254719558, "grad_norm": 0.8390605449676514, "learning_rate": 6.734746125123179e-06, "loss": 0.7714, "step": 17554 }, { "epoch": 0.6185857624849141, "grad_norm": 3.0559353828430176, "learning_rate": 6.733667424759882e-06, "loss": 2.6557, "step": 17555 }, { "epoch": 0.6186209994978725, "grad_norm": 2.2965140342712402, "learning_rate": 6.73258876694099e-06, "loss": 1.0875, "step": 17556 }, { "epoch": 0.618656236510831, "grad_norm": 6.367354869842529, "learning_rate": 6.731510151680538e-06, "loss": 7.0086, "step": 17557 }, { "epoch": 0.6186914735237894, "grad_norm": 2.8418312072753906, "learning_rate": 6.7304315789925845e-06, "loss": 3.0805, "step": 17558 }, { "epoch": 0.6187267105367478, "grad_norm": 5.381515979766846, "learning_rate": 6.72935304889118e-06, "loss": 3.2845, "step": 17559 }, { "epoch": 0.6187619475497063, "grad_norm": 3.304306745529175, "learning_rate": 6.728274561390364e-06, "loss": 2.9026, "step": 17560 }, { "epoch": 0.6187971845626646, "grad_norm": 3.7495920658111572, "learning_rate": 6.727196116504189e-06, "loss": 2.3917, "step": 17561 }, { "epoch": 0.618832421575623, "grad_norm": 1.331229329109192, "learning_rate": 6.726117714246705e-06, "loss": 0.7607, "step": 17562 }, { "epoch": 0.6188676585885814, "grad_norm": 6.038943290710449, "learning_rate": 6.725039354631951e-06, "loss": 5.3104, "step": 17563 }, { "epoch": 0.6189028956015399, "grad_norm": 11.575811386108398, "learning_rate": 6.7239610376739785e-06, "loss": 7.3122, "step": 17564 }, { "epoch": 0.6189381326144983, "grad_norm": 4.135802268981934, "learning_rate": 6.722882763386832e-06, "loss": 3.1603, "step": 17565 }, { "epoch": 0.6189733696274566, "grad_norm": 5.045031547546387, "learning_rate": 6.72180453178455e-06, "loss": 5.3563, "step": 17566 }, { "epoch": 0.6190086066404151, "grad_norm": 3.7450361251831055, "learning_rate": 6.720726342881185e-06, "loss": 2.9831, "step": 17567 }, { "epoch": 0.6190438436533735, "grad_norm": 4.277322769165039, "learning_rate": 6.719648196690779e-06, "loss": 1.979, "step": 17568 }, { "epoch": 0.6190790806663319, "grad_norm": 1.5923879146575928, "learning_rate": 6.718570093227371e-06, "loss": 0.8895, "step": 17569 }, { "epoch": 0.6191143176792904, "grad_norm": 6.859611511230469, "learning_rate": 6.717492032505003e-06, "loss": 5.0094, "step": 17570 }, { "epoch": 0.6191495546922487, "grad_norm": 1.7665235996246338, "learning_rate": 6.716414014537726e-06, "loss": 0.8901, "step": 17571 }, { "epoch": 0.6191847917052071, "grad_norm": 6.271398067474365, "learning_rate": 6.715336039339569e-06, "loss": 4.6137, "step": 17572 }, { "epoch": 0.6192200287181656, "grad_norm": 3.7641172409057617, "learning_rate": 6.714258106924579e-06, "loss": 2.6852, "step": 17573 }, { "epoch": 0.619255265731124, "grad_norm": 1.0096946954727173, "learning_rate": 6.713180217306795e-06, "loss": 1.0081, "step": 17574 }, { "epoch": 0.6192905027440824, "grad_norm": 1.3348824977874756, "learning_rate": 6.712102370500263e-06, "loss": 0.8292, "step": 17575 }, { "epoch": 0.6193257397570407, "grad_norm": 1.37516450881958, "learning_rate": 6.711024566519011e-06, "loss": 0.8932, "step": 17576 }, { "epoch": 0.6193609767699992, "grad_norm": 8.886408805847168, "learning_rate": 6.709946805377083e-06, "loss": 4.865, "step": 17577 }, { "epoch": 0.6193962137829576, "grad_norm": 4.3143768310546875, "learning_rate": 6.708869087088521e-06, "loss": 3.2862, "step": 17578 }, { "epoch": 0.619431450795916, "grad_norm": 9.68786334991455, "learning_rate": 6.707791411667355e-06, "loss": 4.731, "step": 17579 }, { "epoch": 0.6194666878088745, "grad_norm": 3.7196085453033447, "learning_rate": 6.706713779127625e-06, "loss": 2.6203, "step": 17580 }, { "epoch": 0.6195019248218329, "grad_norm": 1.5711653232574463, "learning_rate": 6.70563618948337e-06, "loss": 1.0383, "step": 17581 }, { "epoch": 0.6195371618347912, "grad_norm": 1.0421830415725708, "learning_rate": 6.704558642748622e-06, "loss": 1.1038, "step": 17582 }, { "epoch": 0.6195723988477497, "grad_norm": 9.141804695129395, "learning_rate": 6.703481138937417e-06, "loss": 5.2948, "step": 17583 }, { "epoch": 0.6196076358607081, "grad_norm": 3.3061017990112305, "learning_rate": 6.702403678063791e-06, "loss": 2.5078, "step": 17584 }, { "epoch": 0.6196428728736665, "grad_norm": 5.736839771270752, "learning_rate": 6.701326260141775e-06, "loss": 2.8925, "step": 17585 }, { "epoch": 0.6196781098866249, "grad_norm": 1.361677646636963, "learning_rate": 6.700248885185406e-06, "loss": 0.8645, "step": 17586 }, { "epoch": 0.6197133468995834, "grad_norm": 5.402080535888672, "learning_rate": 6.699171553208719e-06, "loss": 3.1674, "step": 17587 }, { "epoch": 0.6197485839125417, "grad_norm": 3.8558530807495117, "learning_rate": 6.698094264225738e-06, "loss": 0.768, "step": 17588 }, { "epoch": 0.6197838209255001, "grad_norm": 6.467503547668457, "learning_rate": 6.6970170182505e-06, "loss": 3.2214, "step": 17589 }, { "epoch": 0.6198190579384586, "grad_norm": 1.036085605621338, "learning_rate": 6.6959398152970416e-06, "loss": 0.7249, "step": 17590 }, { "epoch": 0.619854294951417, "grad_norm": 3.7091617584228516, "learning_rate": 6.694862655379382e-06, "loss": 3.0828, "step": 17591 }, { "epoch": 0.6198895319643754, "grad_norm": 3.265472650527954, "learning_rate": 6.693785538511559e-06, "loss": 2.5903, "step": 17592 }, { "epoch": 0.6199247689773338, "grad_norm": 1.900320053100586, "learning_rate": 6.692708464707598e-06, "loss": 0.9238, "step": 17593 }, { "epoch": 0.6199600059902922, "grad_norm": 6.952297687530518, "learning_rate": 6.691631433981538e-06, "loss": 4.9592, "step": 17594 }, { "epoch": 0.6199952430032506, "grad_norm": 1.2964236736297607, "learning_rate": 6.690554446347393e-06, "loss": 0.5979, "step": 17595 }, { "epoch": 0.620030480016209, "grad_norm": 1.263109803199768, "learning_rate": 6.689477501819199e-06, "loss": 0.7138, "step": 17596 }, { "epoch": 0.6200657170291675, "grad_norm": 1.3327878713607788, "learning_rate": 6.688400600410986e-06, "loss": 0.6377, "step": 17597 }, { "epoch": 0.6201009540421258, "grad_norm": 1.2836498022079468, "learning_rate": 6.687323742136774e-06, "loss": 1.1142, "step": 17598 }, { "epoch": 0.6201361910550842, "grad_norm": 1.3887354135513306, "learning_rate": 6.686246927010591e-06, "loss": 0.9279, "step": 17599 }, { "epoch": 0.6201714280680427, "grad_norm": 3.563643217086792, "learning_rate": 6.685170155046468e-06, "loss": 2.9445, "step": 17600 }, { "epoch": 0.6202066650810011, "grad_norm": 1.1028332710266113, "learning_rate": 6.684093426258422e-06, "loss": 0.9646, "step": 17601 }, { "epoch": 0.6202419020939595, "grad_norm": 4.867433547973633, "learning_rate": 6.683016740660484e-06, "loss": 4.7052, "step": 17602 }, { "epoch": 0.620277139106918, "grad_norm": 4.024231910705566, "learning_rate": 6.681940098266676e-06, "loss": 3.5966, "step": 17603 }, { "epoch": 0.6203123761198763, "grad_norm": 1.1007007360458374, "learning_rate": 6.680863499091017e-06, "loss": 1.0552, "step": 17604 }, { "epoch": 0.6203476131328347, "grad_norm": 0.9778441190719604, "learning_rate": 6.679786943147536e-06, "loss": 0.9353, "step": 17605 }, { "epoch": 0.6203828501457932, "grad_norm": 1.2946317195892334, "learning_rate": 6.678710430450255e-06, "loss": 0.8313, "step": 17606 }, { "epoch": 0.6204180871587516, "grad_norm": 2.158553123474121, "learning_rate": 6.677633961013193e-06, "loss": 1.0381, "step": 17607 }, { "epoch": 0.62045332417171, "grad_norm": 1.3006482124328613, "learning_rate": 6.676557534850368e-06, "loss": 0.9206, "step": 17608 }, { "epoch": 0.6204885611846683, "grad_norm": 4.783717155456543, "learning_rate": 6.675481151975806e-06, "loss": 4.7453, "step": 17609 }, { "epoch": 0.6205237981976268, "grad_norm": 7.089656352996826, "learning_rate": 6.6744048124035274e-06, "loss": 9.2171, "step": 17610 }, { "epoch": 0.6205590352105852, "grad_norm": 4.909121036529541, "learning_rate": 6.67332851614755e-06, "loss": 2.9329, "step": 17611 }, { "epoch": 0.6205942722235436, "grad_norm": 1.232442855834961, "learning_rate": 6.6722522632218885e-06, "loss": 0.7019, "step": 17612 }, { "epoch": 0.6206295092365021, "grad_norm": 6.212102890014648, "learning_rate": 6.671176053640572e-06, "loss": 3.3147, "step": 17613 }, { "epoch": 0.6206647462494604, "grad_norm": 1.2133301496505737, "learning_rate": 6.670099887417605e-06, "loss": 0.9493, "step": 17614 }, { "epoch": 0.6206999832624188, "grad_norm": 1.243223786354065, "learning_rate": 6.669023764567014e-06, "loss": 1.0056, "step": 17615 }, { "epoch": 0.6207352202753773, "grad_norm": 1.0805630683898926, "learning_rate": 6.667947685102814e-06, "loss": 0.7585, "step": 17616 }, { "epoch": 0.6207704572883357, "grad_norm": 10.781734466552734, "learning_rate": 6.666871649039018e-06, "loss": 5.1915, "step": 17617 }, { "epoch": 0.6208056943012941, "grad_norm": 3.18839168548584, "learning_rate": 6.665795656389642e-06, "loss": 3.0336, "step": 17618 }, { "epoch": 0.6208409313142524, "grad_norm": 4.998705863952637, "learning_rate": 6.664719707168706e-06, "loss": 4.8638, "step": 17619 }, { "epoch": 0.6208761683272109, "grad_norm": 4.294942855834961, "learning_rate": 6.663643801390217e-06, "loss": 2.8061, "step": 17620 }, { "epoch": 0.6209114053401693, "grad_norm": 3.7819132804870605, "learning_rate": 6.662567939068194e-06, "loss": 2.8434, "step": 17621 }, { "epoch": 0.6209466423531277, "grad_norm": 3.7060117721557617, "learning_rate": 6.661492120216651e-06, "loss": 3.0913, "step": 17622 }, { "epoch": 0.6209818793660862, "grad_norm": 1.2179147005081177, "learning_rate": 6.660416344849594e-06, "loss": 0.7602, "step": 17623 }, { "epoch": 0.6210171163790446, "grad_norm": 1.1470115184783936, "learning_rate": 6.65934061298104e-06, "loss": 0.6739, "step": 17624 }, { "epoch": 0.6210523533920029, "grad_norm": 1.4156479835510254, "learning_rate": 6.658264924625004e-06, "loss": 0.915, "step": 17625 }, { "epoch": 0.6210875904049614, "grad_norm": 2.2278850078582764, "learning_rate": 6.65718927979549e-06, "loss": 0.9111, "step": 17626 }, { "epoch": 0.6211228274179198, "grad_norm": 4.302063941955566, "learning_rate": 6.656113678506511e-06, "loss": 0.9778, "step": 17627 }, { "epoch": 0.6211580644308782, "grad_norm": 0.9977445006370544, "learning_rate": 6.655038120772077e-06, "loss": 0.8668, "step": 17628 }, { "epoch": 0.6211933014438366, "grad_norm": 2.016618251800537, "learning_rate": 6.653962606606201e-06, "loss": 1.029, "step": 17629 }, { "epoch": 0.621228538456795, "grad_norm": 7.577256679534912, "learning_rate": 6.652887136022886e-06, "loss": 4.7809, "step": 17630 }, { "epoch": 0.6212637754697534, "grad_norm": 1.5203509330749512, "learning_rate": 6.65181170903614e-06, "loss": 0.6207, "step": 17631 }, { "epoch": 0.6212990124827118, "grad_norm": 0.9663668870925903, "learning_rate": 6.650736325659981e-06, "loss": 1.0134, "step": 17632 }, { "epoch": 0.6213342494956703, "grad_norm": 0.984045684337616, "learning_rate": 6.649660985908399e-06, "loss": 0.9328, "step": 17633 }, { "epoch": 0.6213694865086287, "grad_norm": 4.249575138092041, "learning_rate": 6.6485856897954125e-06, "loss": 3.5137, "step": 17634 }, { "epoch": 0.621404723521587, "grad_norm": 2.926579475402832, "learning_rate": 6.647510437335028e-06, "loss": 3.0724, "step": 17635 }, { "epoch": 0.6214399605345455, "grad_norm": 6.079693794250488, "learning_rate": 6.646435228541244e-06, "loss": 5.117, "step": 17636 }, { "epoch": 0.6214751975475039, "grad_norm": 3.857245683670044, "learning_rate": 6.6453600634280655e-06, "loss": 2.5376, "step": 17637 }, { "epoch": 0.6215104345604623, "grad_norm": 1.1213725805282593, "learning_rate": 6.6442849420095036e-06, "loss": 0.8767, "step": 17638 }, { "epoch": 0.6215456715734207, "grad_norm": 1.400577425956726, "learning_rate": 6.643209864299556e-06, "loss": 0.7167, "step": 17639 }, { "epoch": 0.6215809085863792, "grad_norm": 1.4506455659866333, "learning_rate": 6.642134830312228e-06, "loss": 0.8792, "step": 17640 }, { "epoch": 0.6216161455993375, "grad_norm": 1.4021893739700317, "learning_rate": 6.641059840061523e-06, "loss": 0.861, "step": 17641 }, { "epoch": 0.6216513826122959, "grad_norm": 0.9899793267250061, "learning_rate": 6.639984893561437e-06, "loss": 1.0388, "step": 17642 }, { "epoch": 0.6216866196252544, "grad_norm": 4.434035778045654, "learning_rate": 6.638909990825979e-06, "loss": 3.0035, "step": 17643 }, { "epoch": 0.6217218566382128, "grad_norm": 4.369715213775635, "learning_rate": 6.637835131869148e-06, "loss": 2.6171, "step": 17644 }, { "epoch": 0.6217570936511712, "grad_norm": 4.024897575378418, "learning_rate": 6.63676031670494e-06, "loss": 3.079, "step": 17645 }, { "epoch": 0.6217923306641296, "grad_norm": 1.964613437652588, "learning_rate": 6.635685545347357e-06, "loss": 1.074, "step": 17646 }, { "epoch": 0.621827567677088, "grad_norm": 1.2987748384475708, "learning_rate": 6.6346108178103985e-06, "loss": 0.9037, "step": 17647 }, { "epoch": 0.6218628046900464, "grad_norm": 1.177810788154602, "learning_rate": 6.633536134108065e-06, "loss": 0.7455, "step": 17648 }, { "epoch": 0.6218980417030049, "grad_norm": 1.5345022678375244, "learning_rate": 6.632461494254352e-06, "loss": 0.9132, "step": 17649 }, { "epoch": 0.6219332787159633, "grad_norm": 1.263736367225647, "learning_rate": 6.631386898263255e-06, "loss": 1.0736, "step": 17650 }, { "epoch": 0.6219685157289216, "grad_norm": 1.88512122631073, "learning_rate": 6.630312346148775e-06, "loss": 0.7277, "step": 17651 }, { "epoch": 0.62200375274188, "grad_norm": 1.2372506856918335, "learning_rate": 6.629237837924902e-06, "loss": 1.1185, "step": 17652 }, { "epoch": 0.6220389897548385, "grad_norm": 6.742319107055664, "learning_rate": 6.628163373605638e-06, "loss": 3.2847, "step": 17653 }, { "epoch": 0.6220742267677969, "grad_norm": 2.0678274631500244, "learning_rate": 6.627088953204979e-06, "loss": 0.7919, "step": 17654 }, { "epoch": 0.6221094637807553, "grad_norm": 1.0791853666305542, "learning_rate": 6.626014576736912e-06, "loss": 0.7287, "step": 17655 }, { "epoch": 0.6221447007937138, "grad_norm": 4.714371681213379, "learning_rate": 6.624940244215432e-06, "loss": 3.2134, "step": 17656 }, { "epoch": 0.6221799378066721, "grad_norm": 1.7338716983795166, "learning_rate": 6.623865955654542e-06, "loss": 0.7459, "step": 17657 }, { "epoch": 0.6222151748196305, "grad_norm": 3.5456607341766357, "learning_rate": 6.622791711068224e-06, "loss": 3.2368, "step": 17658 }, { "epoch": 0.622250411832589, "grad_norm": 3.352677822113037, "learning_rate": 6.621717510470474e-06, "loss": 2.294, "step": 17659 }, { "epoch": 0.6222856488455474, "grad_norm": 3.2636754512786865, "learning_rate": 6.620643353875286e-06, "loss": 3.1901, "step": 17660 }, { "epoch": 0.6223208858585058, "grad_norm": 1.2638545036315918, "learning_rate": 6.6195692412966485e-06, "loss": 0.8411, "step": 17661 }, { "epoch": 0.6223561228714641, "grad_norm": 1.685323715209961, "learning_rate": 6.618495172748549e-06, "loss": 0.872, "step": 17662 }, { "epoch": 0.6223913598844226, "grad_norm": 1.6267071962356567, "learning_rate": 6.617421148244984e-06, "loss": 1.0245, "step": 17663 }, { "epoch": 0.622426596897381, "grad_norm": 4.679083347320557, "learning_rate": 6.616347167799939e-06, "loss": 3.25, "step": 17664 }, { "epoch": 0.6224618339103394, "grad_norm": 6.670947551727295, "learning_rate": 6.615273231427402e-06, "loss": 5.1237, "step": 17665 }, { "epoch": 0.6224970709232979, "grad_norm": 2.5378072261810303, "learning_rate": 6.614199339141361e-06, "loss": 2.8586, "step": 17666 }, { "epoch": 0.6225323079362562, "grad_norm": 4.672435283660889, "learning_rate": 6.613125490955807e-06, "loss": 3.347, "step": 17667 }, { "epoch": 0.6225675449492146, "grad_norm": 1.3333529233932495, "learning_rate": 6.612051686884724e-06, "loss": 0.869, "step": 17668 }, { "epoch": 0.6226027819621731, "grad_norm": 2.6282076835632324, "learning_rate": 6.6109779269421e-06, "loss": 2.0413, "step": 17669 }, { "epoch": 0.6226380189751315, "grad_norm": 4.697638511657715, "learning_rate": 6.609904211141922e-06, "loss": 3.1638, "step": 17670 }, { "epoch": 0.6226732559880899, "grad_norm": 1.3330488204956055, "learning_rate": 6.60883053949817e-06, "loss": 0.7718, "step": 17671 }, { "epoch": 0.6227084930010482, "grad_norm": 1.6327804327011108, "learning_rate": 6.607756912024832e-06, "loss": 0.7467, "step": 17672 }, { "epoch": 0.6227437300140067, "grad_norm": 3.662702798843384, "learning_rate": 6.606683328735897e-06, "loss": 2.8787, "step": 17673 }, { "epoch": 0.6227789670269651, "grad_norm": 1.1289323568344116, "learning_rate": 6.60560978964534e-06, "loss": 0.7424, "step": 17674 }, { "epoch": 0.6228142040399235, "grad_norm": 6.990084648132324, "learning_rate": 6.6045362947671465e-06, "loss": 5.4709, "step": 17675 }, { "epoch": 0.622849441052882, "grad_norm": 4.011154651641846, "learning_rate": 6.603462844115307e-06, "loss": 3.2456, "step": 17676 }, { "epoch": 0.6228846780658404, "grad_norm": 3.529512405395508, "learning_rate": 6.602389437703789e-06, "loss": 3.0505, "step": 17677 }, { "epoch": 0.6229199150787987, "grad_norm": 4.257999897003174, "learning_rate": 6.601316075546584e-06, "loss": 1.9812, "step": 17678 }, { "epoch": 0.6229551520917572, "grad_norm": 5.876646995544434, "learning_rate": 6.6002427576576736e-06, "loss": 3.5485, "step": 17679 }, { "epoch": 0.6229903891047156, "grad_norm": 2.277984380722046, "learning_rate": 6.5991694840510305e-06, "loss": 0.7318, "step": 17680 }, { "epoch": 0.623025626117674, "grad_norm": 3.5773983001708984, "learning_rate": 6.598096254740636e-06, "loss": 2.9635, "step": 17681 }, { "epoch": 0.6230608631306325, "grad_norm": 4.485123157501221, "learning_rate": 6.597023069740474e-06, "loss": 5.1793, "step": 17682 }, { "epoch": 0.6230961001435908, "grad_norm": 2.6459624767303467, "learning_rate": 6.595949929064523e-06, "loss": 2.7343, "step": 17683 }, { "epoch": 0.6231313371565492, "grad_norm": 1.0866734981536865, "learning_rate": 6.594876832726753e-06, "loss": 1.1596, "step": 17684 }, { "epoch": 0.6231665741695076, "grad_norm": 1.1158109903335571, "learning_rate": 6.593803780741146e-06, "loss": 0.9203, "step": 17685 }, { "epoch": 0.6232018111824661, "grad_norm": 1.069144606590271, "learning_rate": 6.5927307731216815e-06, "loss": 0.7543, "step": 17686 }, { "epoch": 0.6232370481954245, "grad_norm": 4.539741516113281, "learning_rate": 6.591657809882331e-06, "loss": 3.1777, "step": 17687 }, { "epoch": 0.6232722852083828, "grad_norm": 1.472023606300354, "learning_rate": 6.5905848910370725e-06, "loss": 0.8735, "step": 17688 }, { "epoch": 0.6233075222213413, "grad_norm": 4.256010055541992, "learning_rate": 6.589512016599882e-06, "loss": 2.9707, "step": 17689 }, { "epoch": 0.6233427592342997, "grad_norm": 1.6261377334594727, "learning_rate": 6.588439186584727e-06, "loss": 0.9416, "step": 17690 }, { "epoch": 0.6233779962472581, "grad_norm": 1.056829571723938, "learning_rate": 6.587366401005588e-06, "loss": 0.888, "step": 17691 }, { "epoch": 0.6234132332602166, "grad_norm": 3.612145185470581, "learning_rate": 6.586293659876439e-06, "loss": 2.938, "step": 17692 }, { "epoch": 0.623448470273175, "grad_norm": 5.456039905548096, "learning_rate": 6.585220963211248e-06, "loss": 2.8304, "step": 17693 }, { "epoch": 0.6234837072861333, "grad_norm": 4.956912517547607, "learning_rate": 6.584148311023988e-06, "loss": 5.1996, "step": 17694 }, { "epoch": 0.6235189442990917, "grad_norm": 3.898641347885132, "learning_rate": 6.583075703328636e-06, "loss": 2.3607, "step": 17695 }, { "epoch": 0.6235541813120502, "grad_norm": 8.218265533447266, "learning_rate": 6.582003140139153e-06, "loss": 7.1884, "step": 17696 }, { "epoch": 0.6235894183250086, "grad_norm": 1.6227566003799438, "learning_rate": 6.5809306214695146e-06, "loss": 0.7791, "step": 17697 }, { "epoch": 0.623624655337967, "grad_norm": 1.0660525560379028, "learning_rate": 6.579858147333694e-06, "loss": 1.0544, "step": 17698 }, { "epoch": 0.6236598923509254, "grad_norm": 2.3582980632781982, "learning_rate": 6.578785717745655e-06, "loss": 0.9811, "step": 17699 }, { "epoch": 0.6236951293638838, "grad_norm": 6.555629730224609, "learning_rate": 6.577713332719364e-06, "loss": 3.1398, "step": 17700 }, { "epoch": 0.6237303663768422, "grad_norm": 1.4639642238616943, "learning_rate": 6.576640992268797e-06, "loss": 0.7501, "step": 17701 }, { "epoch": 0.6237656033898007, "grad_norm": 2.1211633682250977, "learning_rate": 6.575568696407917e-06, "loss": 0.8517, "step": 17702 }, { "epoch": 0.6238008404027591, "grad_norm": 2.1673130989074707, "learning_rate": 6.574496445150691e-06, "loss": 1.014, "step": 17703 }, { "epoch": 0.6238360774157174, "grad_norm": 3.638253927230835, "learning_rate": 6.5734242385110815e-06, "loss": 2.5763, "step": 17704 }, { "epoch": 0.6238713144286758, "grad_norm": 1.9093064069747925, "learning_rate": 6.572352076503062e-06, "loss": 0.6473, "step": 17705 }, { "epoch": 0.6239065514416343, "grad_norm": 1.4009376764297485, "learning_rate": 6.571279959140593e-06, "loss": 0.808, "step": 17706 }, { "epoch": 0.6239417884545927, "grad_norm": 3.6876120567321777, "learning_rate": 6.570207886437638e-06, "loss": 0.6494, "step": 17707 }, { "epoch": 0.6239770254675511, "grad_norm": 4.237249851226807, "learning_rate": 6.569135858408163e-06, "loss": 2.8125, "step": 17708 }, { "epoch": 0.6240122624805096, "grad_norm": 4.287484645843506, "learning_rate": 6.568063875066129e-06, "loss": 4.5333, "step": 17709 }, { "epoch": 0.6240474994934679, "grad_norm": 4.044960975646973, "learning_rate": 6.5669919364255e-06, "loss": 2.6293, "step": 17710 }, { "epoch": 0.6240827365064263, "grad_norm": 5.672971725463867, "learning_rate": 6.565920042500241e-06, "loss": 3.0088, "step": 17711 }, { "epoch": 0.6241179735193848, "grad_norm": 1.6278266906738281, "learning_rate": 6.564848193304309e-06, "loss": 0.7987, "step": 17712 }, { "epoch": 0.6241532105323432, "grad_norm": 5.45184326171875, "learning_rate": 6.563776388851665e-06, "loss": 6.9595, "step": 17713 }, { "epoch": 0.6241884475453016, "grad_norm": 1.0579019784927368, "learning_rate": 6.562704629156278e-06, "loss": 1.0011, "step": 17714 }, { "epoch": 0.62422368455826, "grad_norm": 1.1522743701934814, "learning_rate": 6.5616329142320935e-06, "loss": 0.7969, "step": 17715 }, { "epoch": 0.6242589215712184, "grad_norm": 7.195172309875488, "learning_rate": 6.5605612440930795e-06, "loss": 4.9611, "step": 17716 }, { "epoch": 0.6242941585841768, "grad_norm": 3.508669853210449, "learning_rate": 6.5594896187531985e-06, "loss": 3.0894, "step": 17717 }, { "epoch": 0.6243293955971352, "grad_norm": 6.611788749694824, "learning_rate": 6.558418038226399e-06, "loss": 5.7554, "step": 17718 }, { "epoch": 0.6243646326100937, "grad_norm": 5.455493927001953, "learning_rate": 6.557346502526642e-06, "loss": 5.3071, "step": 17719 }, { "epoch": 0.624399869623052, "grad_norm": 6.357633590698242, "learning_rate": 6.5562750116678875e-06, "loss": 5.4143, "step": 17720 }, { "epoch": 0.6244351066360104, "grad_norm": 1.3277682065963745, "learning_rate": 6.555203565664091e-06, "loss": 1.0685, "step": 17721 }, { "epoch": 0.6244703436489689, "grad_norm": 14.551033973693848, "learning_rate": 6.554132164529206e-06, "loss": 7.0813, "step": 17722 }, { "epoch": 0.6245055806619273, "grad_norm": 2.2992136478424072, "learning_rate": 6.553060808277185e-06, "loss": 0.9075, "step": 17723 }, { "epoch": 0.6245408176748857, "grad_norm": 3.440200090408325, "learning_rate": 6.5519894969219896e-06, "loss": 2.8311, "step": 17724 }, { "epoch": 0.6245760546878442, "grad_norm": 1.0227749347686768, "learning_rate": 6.5509182304775696e-06, "loss": 1.2343, "step": 17725 }, { "epoch": 0.6246112917008025, "grad_norm": 6.1159868240356445, "learning_rate": 6.549847008957879e-06, "loss": 4.5234, "step": 17726 }, { "epoch": 0.6246465287137609, "grad_norm": 4.242842674255371, "learning_rate": 6.548775832376872e-06, "loss": 2.9934, "step": 17727 }, { "epoch": 0.6246817657267193, "grad_norm": 6.806243896484375, "learning_rate": 6.5477047007484964e-06, "loss": 3.0715, "step": 17728 }, { "epoch": 0.6247170027396778, "grad_norm": 1.022821068763733, "learning_rate": 6.546633614086707e-06, "loss": 0.6837, "step": 17729 }, { "epoch": 0.6247522397526362, "grad_norm": 0.9322742819786072, "learning_rate": 6.5455625724054575e-06, "loss": 0.8073, "step": 17730 }, { "epoch": 0.6247874767655945, "grad_norm": 1.2181613445281982, "learning_rate": 6.544491575718694e-06, "loss": 0.9539, "step": 17731 }, { "epoch": 0.624822713778553, "grad_norm": 4.498186111450195, "learning_rate": 6.543420624040366e-06, "loss": 3.1464, "step": 17732 }, { "epoch": 0.6248579507915114, "grad_norm": 2.3073768615722656, "learning_rate": 6.542349717384431e-06, "loss": 0.8028, "step": 17733 }, { "epoch": 0.6248931878044698, "grad_norm": 5.35852575302124, "learning_rate": 6.541278855764823e-06, "loss": 2.816, "step": 17734 }, { "epoch": 0.6249284248174283, "grad_norm": 5.192331314086914, "learning_rate": 6.5402080391955025e-06, "loss": 5.0863, "step": 17735 }, { "epoch": 0.6249636618303867, "grad_norm": 1.9761641025543213, "learning_rate": 6.539137267690414e-06, "loss": 0.8632, "step": 17736 }, { "epoch": 0.624998898843345, "grad_norm": 3.7482094764709473, "learning_rate": 6.538066541263503e-06, "loss": 2.7482, "step": 17737 }, { "epoch": 0.6250341358563034, "grad_norm": 1.4754544496536255, "learning_rate": 6.536995859928713e-06, "loss": 0.9351, "step": 17738 }, { "epoch": 0.6250693728692619, "grad_norm": 0.9149352312088013, "learning_rate": 6.535925223699995e-06, "loss": 1.0342, "step": 17739 }, { "epoch": 0.6251046098822203, "grad_norm": 1.8504140377044678, "learning_rate": 6.534854632591294e-06, "loss": 0.9753, "step": 17740 }, { "epoch": 0.6251398468951787, "grad_norm": 5.278398513793945, "learning_rate": 6.533784086616553e-06, "loss": 3.1845, "step": 17741 }, { "epoch": 0.6251750839081371, "grad_norm": 3.4888241291046143, "learning_rate": 6.5327135857897115e-06, "loss": 3.3099, "step": 17742 }, { "epoch": 0.6252103209210955, "grad_norm": 1.9835679531097412, "learning_rate": 6.531643130124722e-06, "loss": 0.9671, "step": 17743 }, { "epoch": 0.6252455579340539, "grad_norm": 1.1924794912338257, "learning_rate": 6.530572719635519e-06, "loss": 0.9673, "step": 17744 }, { "epoch": 0.6252807949470124, "grad_norm": 1.8613197803497314, "learning_rate": 6.52950235433605e-06, "loss": 1.0545, "step": 17745 }, { "epoch": 0.6253160319599708, "grad_norm": 1.3896898031234741, "learning_rate": 6.5284320342402575e-06, "loss": 0.7452, "step": 17746 }, { "epoch": 0.6253512689729291, "grad_norm": 3.736968517303467, "learning_rate": 6.527361759362074e-06, "loss": 3.0784, "step": 17747 }, { "epoch": 0.6253865059858876, "grad_norm": 4.969117164611816, "learning_rate": 6.5262915297154495e-06, "loss": 5.3729, "step": 17748 }, { "epoch": 0.625421742998846, "grad_norm": 1.1084176301956177, "learning_rate": 6.525221345314322e-06, "loss": 0.8163, "step": 17749 }, { "epoch": 0.6254569800118044, "grad_norm": 3.341291666030884, "learning_rate": 6.5241512061726266e-06, "loss": 2.9133, "step": 17750 }, { "epoch": 0.6254922170247628, "grad_norm": 2.6664326190948486, "learning_rate": 6.523081112304304e-06, "loss": 0.8541, "step": 17751 }, { "epoch": 0.6255274540377213, "grad_norm": 2.5027663707733154, "learning_rate": 6.522011063723297e-06, "loss": 1.1107, "step": 17752 }, { "epoch": 0.6255626910506796, "grad_norm": 6.8360395431518555, "learning_rate": 6.520941060443533e-06, "loss": 2.9351, "step": 17753 }, { "epoch": 0.625597928063638, "grad_norm": 1.4111363887786865, "learning_rate": 6.519871102478958e-06, "loss": 0.896, "step": 17754 }, { "epoch": 0.6256331650765965, "grad_norm": 4.594118595123291, "learning_rate": 6.518801189843502e-06, "loss": 2.7633, "step": 17755 }, { "epoch": 0.6256684020895549, "grad_norm": 4.6118316650390625, "learning_rate": 6.51773132255111e-06, "loss": 3.1248, "step": 17756 }, { "epoch": 0.6257036391025133, "grad_norm": 5.587399482727051, "learning_rate": 6.516661500615706e-06, "loss": 5.0708, "step": 17757 }, { "epoch": 0.6257388761154717, "grad_norm": 1.3225195407867432, "learning_rate": 6.515591724051231e-06, "loss": 0.9539, "step": 17758 }, { "epoch": 0.6257741131284301, "grad_norm": 1.7527910470962524, "learning_rate": 6.514521992871621e-06, "loss": 0.964, "step": 17759 }, { "epoch": 0.6258093501413885, "grad_norm": 6.345778465270996, "learning_rate": 6.513452307090801e-06, "loss": 5.1579, "step": 17760 }, { "epoch": 0.6258445871543469, "grad_norm": 2.450406312942505, "learning_rate": 6.51238266672271e-06, "loss": 2.6127, "step": 17761 }, { "epoch": 0.6258798241673054, "grad_norm": 1.9991042613983154, "learning_rate": 6.5113130717812815e-06, "loss": 1.0141, "step": 17762 }, { "epoch": 0.6259150611802637, "grad_norm": 2.3629941940307617, "learning_rate": 6.5102435222804425e-06, "loss": 1.1322, "step": 17763 }, { "epoch": 0.6259502981932221, "grad_norm": 6.463447093963623, "learning_rate": 6.509174018234126e-06, "loss": 3.0316, "step": 17764 }, { "epoch": 0.6259855352061806, "grad_norm": 8.608585357666016, "learning_rate": 6.508104559656266e-06, "loss": 5.7759, "step": 17765 }, { "epoch": 0.626020772219139, "grad_norm": 1.4516993761062622, "learning_rate": 6.507035146560783e-06, "loss": 0.7806, "step": 17766 }, { "epoch": 0.6260560092320974, "grad_norm": 1.135852336883545, "learning_rate": 6.5059657789616155e-06, "loss": 1.0289, "step": 17767 }, { "epoch": 0.6260912462450559, "grad_norm": 6.06099271774292, "learning_rate": 6.5048964568726905e-06, "loss": 4.6047, "step": 17768 }, { "epoch": 0.6261264832580142, "grad_norm": 4.759147644042969, "learning_rate": 6.503827180307932e-06, "loss": 4.5929, "step": 17769 }, { "epoch": 0.6261617202709726, "grad_norm": 2.9469799995422363, "learning_rate": 6.502757949281269e-06, "loss": 2.741, "step": 17770 }, { "epoch": 0.626196957283931, "grad_norm": 7.232760906219482, "learning_rate": 6.501688763806633e-06, "loss": 3.1693, "step": 17771 }, { "epoch": 0.6262321942968895, "grad_norm": 3.053983688354492, "learning_rate": 6.500619623897943e-06, "loss": 2.8832, "step": 17772 }, { "epoch": 0.6262674313098479, "grad_norm": 4.084592342376709, "learning_rate": 6.499550529569127e-06, "loss": 4.8927, "step": 17773 }, { "epoch": 0.6263026683228062, "grad_norm": 6.184239387512207, "learning_rate": 6.498481480834111e-06, "loss": 4.9848, "step": 17774 }, { "epoch": 0.6263379053357647, "grad_norm": 1.3820323944091797, "learning_rate": 6.497412477706827e-06, "loss": 1.0734, "step": 17775 }, { "epoch": 0.6263731423487231, "grad_norm": 1.216219425201416, "learning_rate": 6.496343520201185e-06, "loss": 0.9351, "step": 17776 }, { "epoch": 0.6264083793616815, "grad_norm": 5.450376033782959, "learning_rate": 6.495274608331116e-06, "loss": 5.5757, "step": 17777 }, { "epoch": 0.62644361637464, "grad_norm": 5.180008888244629, "learning_rate": 6.494205742110544e-06, "loss": 5.2584, "step": 17778 }, { "epoch": 0.6264788533875983, "grad_norm": 3.7650961875915527, "learning_rate": 6.493136921553387e-06, "loss": 2.5904, "step": 17779 }, { "epoch": 0.6265140904005567, "grad_norm": 1.2985246181488037, "learning_rate": 6.492068146673566e-06, "loss": 0.8124, "step": 17780 }, { "epoch": 0.6265493274135152, "grad_norm": 1.86366605758667, "learning_rate": 6.490999417485009e-06, "loss": 0.7152, "step": 17781 }, { "epoch": 0.6265845644264736, "grad_norm": 1.3069809675216675, "learning_rate": 6.489930734001628e-06, "loss": 0.9399, "step": 17782 }, { "epoch": 0.626619801439432, "grad_norm": 1.9823716878890991, "learning_rate": 6.4888620962373475e-06, "loss": 0.6801, "step": 17783 }, { "epoch": 0.6266550384523903, "grad_norm": 3.5543291568756104, "learning_rate": 6.487793504206088e-06, "loss": 3.6957, "step": 17784 }, { "epoch": 0.6266902754653488, "grad_norm": 2.0631375312805176, "learning_rate": 6.48672495792176e-06, "loss": 1.0483, "step": 17785 }, { "epoch": 0.6267255124783072, "grad_norm": 1.578589916229248, "learning_rate": 6.48565645739829e-06, "loss": 0.8821, "step": 17786 }, { "epoch": 0.6267607494912656, "grad_norm": 5.005820274353027, "learning_rate": 6.484588002649593e-06, "loss": 3.2688, "step": 17787 }, { "epoch": 0.6267959865042241, "grad_norm": 3.2909657955169678, "learning_rate": 6.483519593689585e-06, "loss": 2.9756, "step": 17788 }, { "epoch": 0.6268312235171825, "grad_norm": 1.0644317865371704, "learning_rate": 6.482451230532179e-06, "loss": 0.9027, "step": 17789 }, { "epoch": 0.6268664605301408, "grad_norm": 1.1275135278701782, "learning_rate": 6.4813829131912995e-06, "loss": 0.9367, "step": 17790 }, { "epoch": 0.6269016975430993, "grad_norm": 2.4893746376037598, "learning_rate": 6.480314641680851e-06, "loss": 0.9804, "step": 17791 }, { "epoch": 0.6269369345560577, "grad_norm": 2.805797576904297, "learning_rate": 6.479246416014753e-06, "loss": 2.7607, "step": 17792 }, { "epoch": 0.6269721715690161, "grad_norm": 3.7891461849212646, "learning_rate": 6.478178236206918e-06, "loss": 3.2587, "step": 17793 }, { "epoch": 0.6270074085819745, "grad_norm": 4.796695709228516, "learning_rate": 6.4771101022712655e-06, "loss": 2.6243, "step": 17794 }, { "epoch": 0.627042645594933, "grad_norm": 2.6651766300201416, "learning_rate": 6.476042014221697e-06, "loss": 1.0121, "step": 17795 }, { "epoch": 0.6270778826078913, "grad_norm": 15.902071952819824, "learning_rate": 6.4749739720721315e-06, "loss": 2.9698, "step": 17796 }, { "epoch": 0.6271131196208497, "grad_norm": 4.055643558502197, "learning_rate": 6.473905975836483e-06, "loss": 2.911, "step": 17797 }, { "epoch": 0.6271483566338082, "grad_norm": 1.9105846881866455, "learning_rate": 6.472838025528655e-06, "loss": 0.6576, "step": 17798 }, { "epoch": 0.6271835936467666, "grad_norm": 2.530949115753174, "learning_rate": 6.471770121162558e-06, "loss": 0.9947, "step": 17799 }, { "epoch": 0.627218830659725, "grad_norm": 4.088836669921875, "learning_rate": 6.470702262752111e-06, "loss": 2.6455, "step": 17800 }, { "epoch": 0.6272540676726834, "grad_norm": 6.256904602050781, "learning_rate": 6.469634450311213e-06, "loss": 3.094, "step": 17801 }, { "epoch": 0.6272893046856418, "grad_norm": 5.048011779785156, "learning_rate": 6.468566683853777e-06, "loss": 3.2165, "step": 17802 }, { "epoch": 0.6273245416986002, "grad_norm": 3.989487886428833, "learning_rate": 6.46749896339371e-06, "loss": 4.5974, "step": 17803 }, { "epoch": 0.6273597787115586, "grad_norm": 5.253726959228516, "learning_rate": 6.466431288944916e-06, "loss": 3.231, "step": 17804 }, { "epoch": 0.6273950157245171, "grad_norm": 12.556475639343262, "learning_rate": 6.465363660521306e-06, "loss": 7.486, "step": 17805 }, { "epoch": 0.6274302527374754, "grad_norm": 2.774853467941284, "learning_rate": 6.464296078136787e-06, "loss": 2.8357, "step": 17806 }, { "epoch": 0.6274654897504338, "grad_norm": 4.994633674621582, "learning_rate": 6.46322854180526e-06, "loss": 3.0411, "step": 17807 }, { "epoch": 0.6275007267633923, "grad_norm": 3.2177884578704834, "learning_rate": 6.4621610515406315e-06, "loss": 3.1064, "step": 17808 }, { "epoch": 0.6275359637763507, "grad_norm": 5.488150119781494, "learning_rate": 6.4610936073568085e-06, "loss": 3.141, "step": 17809 }, { "epoch": 0.6275712007893091, "grad_norm": 18.69074249267578, "learning_rate": 6.460026209267688e-06, "loss": 3.2195, "step": 17810 }, { "epoch": 0.6276064378022675, "grad_norm": 4.3706254959106445, "learning_rate": 6.4589588572871785e-06, "loss": 2.854, "step": 17811 }, { "epoch": 0.6276416748152259, "grad_norm": 8.368457794189453, "learning_rate": 6.4578915514291806e-06, "loss": 5.3142, "step": 17812 }, { "epoch": 0.6276769118281843, "grad_norm": 2.399264335632324, "learning_rate": 6.456824291707599e-06, "loss": 0.7755, "step": 17813 }, { "epoch": 0.6277121488411427, "grad_norm": 3.813924789428711, "learning_rate": 6.455757078136329e-06, "loss": 3.2227, "step": 17814 }, { "epoch": 0.6277473858541012, "grad_norm": 3.6894640922546387, "learning_rate": 6.454689910729277e-06, "loss": 2.8836, "step": 17815 }, { "epoch": 0.6277826228670595, "grad_norm": 1.0641469955444336, "learning_rate": 6.45362278950034e-06, "loss": 0.8646, "step": 17816 }, { "epoch": 0.6278178598800179, "grad_norm": 4.4761786460876465, "learning_rate": 6.452555714463418e-06, "loss": 3.1224, "step": 17817 }, { "epoch": 0.6278530968929764, "grad_norm": 3.677117109298706, "learning_rate": 6.451488685632408e-06, "loss": 5.0332, "step": 17818 }, { "epoch": 0.6278883339059348, "grad_norm": 1.9294121265411377, "learning_rate": 6.4504217030212145e-06, "loss": 0.8883, "step": 17819 }, { "epoch": 0.6279235709188932, "grad_norm": 7.407010555267334, "learning_rate": 6.449354766643725e-06, "loss": 3.1503, "step": 17820 }, { "epoch": 0.6279588079318517, "grad_norm": 1.5312178134918213, "learning_rate": 6.4482878765138434e-06, "loss": 0.9375, "step": 17821 }, { "epoch": 0.62799404494481, "grad_norm": 1.1154475212097168, "learning_rate": 6.447221032645468e-06, "loss": 0.9237, "step": 17822 }, { "epoch": 0.6280292819577684, "grad_norm": 2.8318192958831787, "learning_rate": 6.4461542350524876e-06, "loss": 2.79, "step": 17823 }, { "epoch": 0.6280645189707269, "grad_norm": 3.271998643875122, "learning_rate": 6.4450874837488e-06, "loss": 2.2621, "step": 17824 }, { "epoch": 0.6280997559836853, "grad_norm": 1.6589021682739258, "learning_rate": 6.444020778748303e-06, "loss": 0.9682, "step": 17825 }, { "epoch": 0.6281349929966437, "grad_norm": 1.050269603729248, "learning_rate": 6.442954120064888e-06, "loss": 0.9666, "step": 17826 }, { "epoch": 0.628170230009602, "grad_norm": 1.1469560861587524, "learning_rate": 6.4418875077124485e-06, "loss": 0.8258, "step": 17827 }, { "epoch": 0.6282054670225605, "grad_norm": 2.6442031860351562, "learning_rate": 6.440820941704875e-06, "loss": 0.9432, "step": 17828 }, { "epoch": 0.6282407040355189, "grad_norm": 0.8978802561759949, "learning_rate": 6.4397544220560674e-06, "loss": 0.9264, "step": 17829 }, { "epoch": 0.6282759410484773, "grad_norm": 4.5167460441589355, "learning_rate": 6.43868794877991e-06, "loss": 4.7891, "step": 17830 }, { "epoch": 0.6283111780614358, "grad_norm": 6.556020259857178, "learning_rate": 6.437621521890294e-06, "loss": 5.173, "step": 17831 }, { "epoch": 0.6283464150743941, "grad_norm": 3.5726799964904785, "learning_rate": 6.436555141401113e-06, "loss": 3.2751, "step": 17832 }, { "epoch": 0.6283816520873525, "grad_norm": 1.6170454025268555, "learning_rate": 6.435488807326254e-06, "loss": 0.7317, "step": 17833 }, { "epoch": 0.628416889100311, "grad_norm": 4.491982460021973, "learning_rate": 6.4344225196796065e-06, "loss": 5.0624, "step": 17834 }, { "epoch": 0.6284521261132694, "grad_norm": 1.756561279296875, "learning_rate": 6.433356278475063e-06, "loss": 0.8903, "step": 17835 }, { "epoch": 0.6284873631262278, "grad_norm": 1.5377297401428223, "learning_rate": 6.432290083726507e-06, "loss": 0.8295, "step": 17836 }, { "epoch": 0.6285226001391861, "grad_norm": 1.026918888092041, "learning_rate": 6.431223935447823e-06, "loss": 0.9463, "step": 17837 }, { "epoch": 0.6285578371521446, "grad_norm": 3.677701234817505, "learning_rate": 6.4301578336529105e-06, "loss": 2.6991, "step": 17838 }, { "epoch": 0.628593074165103, "grad_norm": 3.2575926780700684, "learning_rate": 6.429091778355639e-06, "loss": 0.9694, "step": 17839 }, { "epoch": 0.6286283111780614, "grad_norm": 3.725219964981079, "learning_rate": 6.428025769569906e-06, "loss": 2.9564, "step": 17840 }, { "epoch": 0.6286635481910199, "grad_norm": 3.5750317573547363, "learning_rate": 6.4269598073095914e-06, "loss": 2.5329, "step": 17841 }, { "epoch": 0.6286987852039783, "grad_norm": 0.9001880288124084, "learning_rate": 6.42589389158858e-06, "loss": 1.003, "step": 17842 }, { "epoch": 0.6287340222169366, "grad_norm": 2.801710605621338, "learning_rate": 6.424828022420755e-06, "loss": 0.7535, "step": 17843 }, { "epoch": 0.6287692592298951, "grad_norm": 1.3554151058197021, "learning_rate": 6.4237621998200024e-06, "loss": 0.9812, "step": 17844 }, { "epoch": 0.6288044962428535, "grad_norm": 1.0160235166549683, "learning_rate": 6.422696423800202e-06, "loss": 0.7238, "step": 17845 }, { "epoch": 0.6288397332558119, "grad_norm": 6.016749858856201, "learning_rate": 6.421630694375236e-06, "loss": 5.6087, "step": 17846 }, { "epoch": 0.6288749702687703, "grad_norm": 17.4970703125, "learning_rate": 6.4205650115589835e-06, "loss": 2.5462, "step": 17847 }, { "epoch": 0.6289102072817288, "grad_norm": 1.0822702646255493, "learning_rate": 6.419499375365332e-06, "loss": 0.705, "step": 17848 }, { "epoch": 0.6289454442946871, "grad_norm": 1.07893967628479, "learning_rate": 6.418433785808156e-06, "loss": 1.0701, "step": 17849 }, { "epoch": 0.6289806813076455, "grad_norm": 1.3118772506713867, "learning_rate": 6.417368242901337e-06, "loss": 1.0065, "step": 17850 }, { "epoch": 0.629015918320604, "grad_norm": 1.2162359952926636, "learning_rate": 6.416302746658754e-06, "loss": 0.935, "step": 17851 }, { "epoch": 0.6290511553335624, "grad_norm": 3.717118263244629, "learning_rate": 6.415237297094279e-06, "loss": 4.549, "step": 17852 }, { "epoch": 0.6290863923465208, "grad_norm": 6.2616729736328125, "learning_rate": 6.414171894221797e-06, "loss": 3.2267, "step": 17853 }, { "epoch": 0.6291216293594792, "grad_norm": 2.840593099594116, "learning_rate": 6.413106538055185e-06, "loss": 2.6675, "step": 17854 }, { "epoch": 0.6291568663724376, "grad_norm": 5.343278884887695, "learning_rate": 6.412041228608317e-06, "loss": 2.6692, "step": 17855 }, { "epoch": 0.629192103385396, "grad_norm": 5.616418838500977, "learning_rate": 6.410975965895066e-06, "loss": 3.0308, "step": 17856 }, { "epoch": 0.6292273403983545, "grad_norm": 3.5176894664764404, "learning_rate": 6.409910749929316e-06, "loss": 2.7623, "step": 17857 }, { "epoch": 0.6292625774113129, "grad_norm": 1.248254418373108, "learning_rate": 6.40884558072493e-06, "loss": 0.8897, "step": 17858 }, { "epoch": 0.6292978144242712, "grad_norm": 1.2243446111679077, "learning_rate": 6.407780458295788e-06, "loss": 0.9037, "step": 17859 }, { "epoch": 0.6293330514372296, "grad_norm": 5.745357036590576, "learning_rate": 6.406715382655767e-06, "loss": 5.0545, "step": 17860 }, { "epoch": 0.6293682884501881, "grad_norm": 2.2608466148376465, "learning_rate": 6.405650353818733e-06, "loss": 0.7352, "step": 17861 }, { "epoch": 0.6294035254631465, "grad_norm": 4.852903842926025, "learning_rate": 6.40458537179856e-06, "loss": 2.1984, "step": 17862 }, { "epoch": 0.6294387624761049, "grad_norm": 2.842388391494751, "learning_rate": 6.403520436609123e-06, "loss": 0.7608, "step": 17863 }, { "epoch": 0.6294739994890634, "grad_norm": 4.0551958084106445, "learning_rate": 6.402455548264291e-06, "loss": 2.9914, "step": 17864 }, { "epoch": 0.6295092365020217, "grad_norm": 4.850154399871826, "learning_rate": 6.401390706777931e-06, "loss": 3.1335, "step": 17865 }, { "epoch": 0.6295444735149801, "grad_norm": 4.326788425445557, "learning_rate": 6.400325912163915e-06, "loss": 2.5609, "step": 17866 }, { "epoch": 0.6295797105279386, "grad_norm": 4.222649097442627, "learning_rate": 6.399261164436116e-06, "loss": 2.9578, "step": 17867 }, { "epoch": 0.629614947540897, "grad_norm": 3.7789199352264404, "learning_rate": 6.398196463608397e-06, "loss": 2.785, "step": 17868 }, { "epoch": 0.6296501845538554, "grad_norm": 7.067897319793701, "learning_rate": 6.397131809694628e-06, "loss": 4.5564, "step": 17869 }, { "epoch": 0.6296854215668137, "grad_norm": 4.880972862243652, "learning_rate": 6.396067202708678e-06, "loss": 3.2943, "step": 17870 }, { "epoch": 0.6297206585797722, "grad_norm": 1.1936349868774414, "learning_rate": 6.395002642664407e-06, "loss": 0.8689, "step": 17871 }, { "epoch": 0.6297558955927306, "grad_norm": 4.043163776397705, "learning_rate": 6.39393812957569e-06, "loss": 2.156, "step": 17872 }, { "epoch": 0.629791132605689, "grad_norm": 0.9976171255111694, "learning_rate": 6.392873663456388e-06, "loss": 1.0501, "step": 17873 }, { "epoch": 0.6298263696186475, "grad_norm": 2.889655590057373, "learning_rate": 6.391809244320365e-06, "loss": 2.7927, "step": 17874 }, { "epoch": 0.6298616066316058, "grad_norm": 5.728270530700684, "learning_rate": 6.390744872181485e-06, "loss": 2.6704, "step": 17875 }, { "epoch": 0.6298968436445642, "grad_norm": 4.475739479064941, "learning_rate": 6.389680547053617e-06, "loss": 2.8682, "step": 17876 }, { "epoch": 0.6299320806575227, "grad_norm": 2.0118303298950195, "learning_rate": 6.3886162689506135e-06, "loss": 1.0759, "step": 17877 }, { "epoch": 0.6299673176704811, "grad_norm": 8.468804359436035, "learning_rate": 6.387552037886346e-06, "loss": 4.6874, "step": 17878 }, { "epoch": 0.6300025546834395, "grad_norm": 3.203299045562744, "learning_rate": 6.386487853874676e-06, "loss": 3.075, "step": 17879 }, { "epoch": 0.6300377916963978, "grad_norm": 4.391237735748291, "learning_rate": 6.385423716929459e-06, "loss": 2.833, "step": 17880 }, { "epoch": 0.6300730287093563, "grad_norm": 1.5863218307495117, "learning_rate": 6.384359627064558e-06, "loss": 0.8295, "step": 17881 }, { "epoch": 0.6301082657223147, "grad_norm": 5.459733009338379, "learning_rate": 6.383295584293836e-06, "loss": 5.2404, "step": 17882 }, { "epoch": 0.6301435027352731, "grad_norm": 1.061315894126892, "learning_rate": 6.382231588631149e-06, "loss": 0.9109, "step": 17883 }, { "epoch": 0.6301787397482316, "grad_norm": 4.47465181350708, "learning_rate": 6.381167640090356e-06, "loss": 3.1307, "step": 17884 }, { "epoch": 0.63021397676119, "grad_norm": 2.103804349899292, "learning_rate": 6.380103738685312e-06, "loss": 0.8115, "step": 17885 }, { "epoch": 0.6302492137741483, "grad_norm": 5.906819820404053, "learning_rate": 6.379039884429883e-06, "loss": 2.656, "step": 17886 }, { "epoch": 0.6302844507871068, "grad_norm": 4.538823127746582, "learning_rate": 6.377976077337919e-06, "loss": 3.5033, "step": 17887 }, { "epoch": 0.6303196878000652, "grad_norm": 7.208857536315918, "learning_rate": 6.376912317423277e-06, "loss": 3.1111, "step": 17888 }, { "epoch": 0.6303549248130236, "grad_norm": 1.2368735074996948, "learning_rate": 6.375848604699817e-06, "loss": 0.8967, "step": 17889 }, { "epoch": 0.6303901618259821, "grad_norm": 4.583374977111816, "learning_rate": 6.374784939181385e-06, "loss": 2.8392, "step": 17890 }, { "epoch": 0.6304253988389404, "grad_norm": 1.9967864751815796, "learning_rate": 6.373721320881844e-06, "loss": 1.8757, "step": 17891 }, { "epoch": 0.6304606358518988, "grad_norm": 4.005180835723877, "learning_rate": 6.3726577498150455e-06, "loss": 3.5479, "step": 17892 }, { "epoch": 0.6304958728648572, "grad_norm": 4.259688377380371, "learning_rate": 6.371594225994842e-06, "loss": 2.9694, "step": 17893 }, { "epoch": 0.6305311098778157, "grad_norm": 6.309142112731934, "learning_rate": 6.370530749435082e-06, "loss": 4.4039, "step": 17894 }, { "epoch": 0.6305663468907741, "grad_norm": 1.8677459955215454, "learning_rate": 6.369467320149628e-06, "loss": 0.752, "step": 17895 }, { "epoch": 0.6306015839037324, "grad_norm": 2.553715467453003, "learning_rate": 6.368403938152319e-06, "loss": 2.9077, "step": 17896 }, { "epoch": 0.6306368209166909, "grad_norm": 0.8536930680274963, "learning_rate": 6.367340603457013e-06, "loss": 1.1248, "step": 17897 }, { "epoch": 0.6306720579296493, "grad_norm": 5.142810344696045, "learning_rate": 6.36627731607756e-06, "loss": 3.1817, "step": 17898 }, { "epoch": 0.6307072949426077, "grad_norm": 3.235541582107544, "learning_rate": 6.365214076027806e-06, "loss": 2.6189, "step": 17899 }, { "epoch": 0.6307425319555662, "grad_norm": 4.786983489990234, "learning_rate": 6.3641508833215996e-06, "loss": 3.1908, "step": 17900 }, { "epoch": 0.6307777689685246, "grad_norm": 2.8731906414031982, "learning_rate": 6.363087737972793e-06, "loss": 1.1023, "step": 17901 }, { "epoch": 0.6308130059814829, "grad_norm": 1.1519041061401367, "learning_rate": 6.3620246399952344e-06, "loss": 1.2568, "step": 17902 }, { "epoch": 0.6308482429944413, "grad_norm": 1.9982548952102661, "learning_rate": 6.360961589402765e-06, "loss": 0.988, "step": 17903 }, { "epoch": 0.6308834800073998, "grad_norm": 5.196329116821289, "learning_rate": 6.3598985862092345e-06, "loss": 3.5168, "step": 17904 }, { "epoch": 0.6309187170203582, "grad_norm": 1.6706750392913818, "learning_rate": 6.35883563042849e-06, "loss": 0.8714, "step": 17905 }, { "epoch": 0.6309539540333166, "grad_norm": 4.530325412750244, "learning_rate": 6.357772722074376e-06, "loss": 3.1791, "step": 17906 }, { "epoch": 0.630989191046275, "grad_norm": 5.35599422454834, "learning_rate": 6.356709861160734e-06, "loss": 3.1379, "step": 17907 }, { "epoch": 0.6310244280592334, "grad_norm": 1.965240240097046, "learning_rate": 6.355647047701414e-06, "loss": 1.0855, "step": 17908 }, { "epoch": 0.6310596650721918, "grad_norm": 3.589350461959839, "learning_rate": 6.35458428171025e-06, "loss": 3.0036, "step": 17909 }, { "epoch": 0.6310949020851503, "grad_norm": 9.937376976013184, "learning_rate": 6.353521563201093e-06, "loss": 2.9504, "step": 17910 }, { "epoch": 0.6311301390981087, "grad_norm": 3.7442047595977783, "learning_rate": 6.352458892187783e-06, "loss": 2.9595, "step": 17911 }, { "epoch": 0.631165376111067, "grad_norm": 2.711851119995117, "learning_rate": 6.35139626868416e-06, "loss": 2.5683, "step": 17912 }, { "epoch": 0.6312006131240254, "grad_norm": 14.253039360046387, "learning_rate": 6.350333692704063e-06, "loss": 3.1413, "step": 17913 }, { "epoch": 0.6312358501369839, "grad_norm": 1.1079864501953125, "learning_rate": 6.349271164261339e-06, "loss": 1.1252, "step": 17914 }, { "epoch": 0.6312710871499423, "grad_norm": 1.098902940750122, "learning_rate": 6.348208683369819e-06, "loss": 0.9513, "step": 17915 }, { "epoch": 0.6313063241629007, "grad_norm": 5.760491371154785, "learning_rate": 6.3471462500433465e-06, "loss": 5.4045, "step": 17916 }, { "epoch": 0.6313415611758592, "grad_norm": 1.3661303520202637, "learning_rate": 6.346083864295762e-06, "loss": 0.7747, "step": 17917 }, { "epoch": 0.6313767981888175, "grad_norm": 1.1204768419265747, "learning_rate": 6.3450215261408995e-06, "loss": 0.7572, "step": 17918 }, { "epoch": 0.6314120352017759, "grad_norm": 5.795668601989746, "learning_rate": 6.343959235592594e-06, "loss": 5.4622, "step": 17919 }, { "epoch": 0.6314472722147344, "grad_norm": 1.5897592306137085, "learning_rate": 6.342896992664689e-06, "loss": 0.815, "step": 17920 }, { "epoch": 0.6314825092276928, "grad_norm": 5.139483451843262, "learning_rate": 6.341834797371016e-06, "loss": 3.1905, "step": 17921 }, { "epoch": 0.6315177462406512, "grad_norm": 4.243089199066162, "learning_rate": 6.340772649725411e-06, "loss": 2.9786, "step": 17922 }, { "epoch": 0.6315529832536096, "grad_norm": 7.382236003875732, "learning_rate": 6.339710549741704e-06, "loss": 3.0297, "step": 17923 }, { "epoch": 0.631588220266568, "grad_norm": 3.138632297515869, "learning_rate": 6.338648497433739e-06, "loss": 3.0356, "step": 17924 }, { "epoch": 0.6316234572795264, "grad_norm": 6.172389030456543, "learning_rate": 6.337586492815342e-06, "loss": 2.8395, "step": 17925 }, { "epoch": 0.6316586942924848, "grad_norm": 1.570504069328308, "learning_rate": 6.336524535900346e-06, "loss": 1.1773, "step": 17926 }, { "epoch": 0.6316939313054433, "grad_norm": 3.5156326293945312, "learning_rate": 6.335462626702588e-06, "loss": 2.8037, "step": 17927 }, { "epoch": 0.6317291683184016, "grad_norm": 7.832324028015137, "learning_rate": 6.334400765235891e-06, "loss": 4.8124, "step": 17928 }, { "epoch": 0.63176440533136, "grad_norm": 1.1968592405319214, "learning_rate": 6.333338951514094e-06, "loss": 0.8485, "step": 17929 }, { "epoch": 0.6317996423443185, "grad_norm": 1.5753790140151978, "learning_rate": 6.3322771855510255e-06, "loss": 0.8095, "step": 17930 }, { "epoch": 0.6318348793572769, "grad_norm": 1.2039633989334106, "learning_rate": 6.3312154673605116e-06, "loss": 0.7299, "step": 17931 }, { "epoch": 0.6318701163702353, "grad_norm": 2.536867380142212, "learning_rate": 6.330153796956382e-06, "loss": 3.1865, "step": 17932 }, { "epoch": 0.6319053533831938, "grad_norm": 7.151152610778809, "learning_rate": 6.329092174352472e-06, "loss": 4.8218, "step": 17933 }, { "epoch": 0.6319405903961521, "grad_norm": 1.1791127920150757, "learning_rate": 6.328030599562599e-06, "loss": 0.9843, "step": 17934 }, { "epoch": 0.6319758274091105, "grad_norm": 0.9305340647697449, "learning_rate": 6.326969072600597e-06, "loss": 1.0351, "step": 17935 }, { "epoch": 0.6320110644220689, "grad_norm": 1.7574776411056519, "learning_rate": 6.325907593480292e-06, "loss": 0.6988, "step": 17936 }, { "epoch": 0.6320463014350274, "grad_norm": 1.3679488897323608, "learning_rate": 6.3248461622155076e-06, "loss": 0.848, "step": 17937 }, { "epoch": 0.6320815384479858, "grad_norm": 2.031665086746216, "learning_rate": 6.323784778820067e-06, "loss": 2.4954, "step": 17938 }, { "epoch": 0.6321167754609441, "grad_norm": 2.4915988445281982, "learning_rate": 6.322723443307801e-06, "loss": 2.74, "step": 17939 }, { "epoch": 0.6321520124739026, "grad_norm": 16.22096824645996, "learning_rate": 6.321662155692532e-06, "loss": 4.7163, "step": 17940 }, { "epoch": 0.632187249486861, "grad_norm": 1.2354774475097656, "learning_rate": 6.32060091598808e-06, "loss": 0.734, "step": 17941 }, { "epoch": 0.6322224864998194, "grad_norm": 1.2799931764602661, "learning_rate": 6.3195397242082675e-06, "loss": 0.813, "step": 17942 }, { "epoch": 0.6322577235127779, "grad_norm": 2.9345455169677734, "learning_rate": 6.318478580366922e-06, "loss": 2.8634, "step": 17943 }, { "epoch": 0.6322929605257362, "grad_norm": 1.4134502410888672, "learning_rate": 6.3174174844778614e-06, "loss": 0.7848, "step": 17944 }, { "epoch": 0.6323281975386946, "grad_norm": 5.9241743087768555, "learning_rate": 6.316356436554907e-06, "loss": 3.2968, "step": 17945 }, { "epoch": 0.632363434551653, "grad_norm": 0.97328782081604, "learning_rate": 6.315295436611882e-06, "loss": 1.0793, "step": 17946 }, { "epoch": 0.6323986715646115, "grad_norm": 1.5658326148986816, "learning_rate": 6.314234484662598e-06, "loss": 0.8663, "step": 17947 }, { "epoch": 0.6324339085775699, "grad_norm": 4.260051727294922, "learning_rate": 6.313173580720881e-06, "loss": 2.8958, "step": 17948 }, { "epoch": 0.6324691455905282, "grad_norm": 2.0649209022521973, "learning_rate": 6.312112724800551e-06, "loss": 0.9389, "step": 17949 }, { "epoch": 0.6325043826034867, "grad_norm": 1.2261059284210205, "learning_rate": 6.311051916915421e-06, "loss": 0.817, "step": 17950 }, { "epoch": 0.6325396196164451, "grad_norm": 4.754776954650879, "learning_rate": 6.309991157079306e-06, "loss": 3.3487, "step": 17951 }, { "epoch": 0.6325748566294035, "grad_norm": 4.598262786865234, "learning_rate": 6.308930445306033e-06, "loss": 2.3308, "step": 17952 }, { "epoch": 0.632610093642362, "grad_norm": 1.1980456113815308, "learning_rate": 6.307869781609404e-06, "loss": 0.8883, "step": 17953 }, { "epoch": 0.6326453306553204, "grad_norm": 3.7100517749786377, "learning_rate": 6.306809166003245e-06, "loss": 3.2322, "step": 17954 }, { "epoch": 0.6326805676682787, "grad_norm": 1.1224522590637207, "learning_rate": 6.305748598501369e-06, "loss": 0.937, "step": 17955 }, { "epoch": 0.6327158046812372, "grad_norm": 1.6894069910049438, "learning_rate": 6.304688079117585e-06, "loss": 0.8083, "step": 17956 }, { "epoch": 0.6327510416941956, "grad_norm": 1.5716383457183838, "learning_rate": 6.3036276078657075e-06, "loss": 0.864, "step": 17957 }, { "epoch": 0.632786278707154, "grad_norm": 5.617770195007324, "learning_rate": 6.302567184759554e-06, "loss": 4.0089, "step": 17958 }, { "epoch": 0.6328215157201124, "grad_norm": 0.9991961717605591, "learning_rate": 6.301506809812936e-06, "loss": 0.9458, "step": 17959 }, { "epoch": 0.6328567527330708, "grad_norm": 1.2710002660751343, "learning_rate": 6.30044648303966e-06, "loss": 0.6715, "step": 17960 }, { "epoch": 0.6328919897460292, "grad_norm": 2.487642526626587, "learning_rate": 6.299386204453539e-06, "loss": 0.8536, "step": 17961 }, { "epoch": 0.6329272267589876, "grad_norm": 16.458276748657227, "learning_rate": 6.298325974068387e-06, "loss": 5.1463, "step": 17962 }, { "epoch": 0.6329624637719461, "grad_norm": 6.086555004119873, "learning_rate": 6.297265791898008e-06, "loss": 3.4614, "step": 17963 }, { "epoch": 0.6329977007849045, "grad_norm": 3.80785870552063, "learning_rate": 6.296205657956216e-06, "loss": 2.8158, "step": 17964 }, { "epoch": 0.6330329377978628, "grad_norm": 2.7405834197998047, "learning_rate": 6.295145572256818e-06, "loss": 0.7946, "step": 17965 }, { "epoch": 0.6330681748108213, "grad_norm": 1.3045482635498047, "learning_rate": 6.294085534813619e-06, "loss": 0.8782, "step": 17966 }, { "epoch": 0.6331034118237797, "grad_norm": 5.288857936859131, "learning_rate": 6.293025545640425e-06, "loss": 4.7679, "step": 17967 }, { "epoch": 0.6331386488367381, "grad_norm": 6.992429733276367, "learning_rate": 6.291965604751052e-06, "loss": 2.7337, "step": 17968 }, { "epoch": 0.6331738858496965, "grad_norm": 1.118189811706543, "learning_rate": 6.2909057121592945e-06, "loss": 1.2038, "step": 17969 }, { "epoch": 0.633209122862655, "grad_norm": 5.641335964202881, "learning_rate": 6.2898458678789635e-06, "loss": 6.875, "step": 17970 }, { "epoch": 0.6332443598756133, "grad_norm": 5.408361911773682, "learning_rate": 6.2887860719238646e-06, "loss": 3.2548, "step": 17971 }, { "epoch": 0.6332795968885717, "grad_norm": 3.303051471710205, "learning_rate": 6.287726324307796e-06, "loss": 2.8653, "step": 17972 }, { "epoch": 0.6333148339015302, "grad_norm": 1.0947822332382202, "learning_rate": 6.2866666250445665e-06, "loss": 0.9984, "step": 17973 }, { "epoch": 0.6333500709144886, "grad_norm": 8.11725902557373, "learning_rate": 6.285606974147977e-06, "loss": 7.2055, "step": 17974 }, { "epoch": 0.633385307927447, "grad_norm": 0.9502100348472595, "learning_rate": 6.284547371631833e-06, "loss": 0.9786, "step": 17975 }, { "epoch": 0.6334205449404054, "grad_norm": 1.3948514461517334, "learning_rate": 6.283487817509928e-06, "loss": 1.0936, "step": 17976 }, { "epoch": 0.6334557819533638, "grad_norm": 1.784894347190857, "learning_rate": 6.28242831179607e-06, "loss": 0.5877, "step": 17977 }, { "epoch": 0.6334910189663222, "grad_norm": 1.8360321521759033, "learning_rate": 6.281368854504059e-06, "loss": 0.8418, "step": 17978 }, { "epoch": 0.6335262559792806, "grad_norm": 2.5158979892730713, "learning_rate": 6.28030944564769e-06, "loss": 1.0401, "step": 17979 }, { "epoch": 0.6335614929922391, "grad_norm": 5.25177526473999, "learning_rate": 6.279250085240763e-06, "loss": 2.4098, "step": 17980 }, { "epoch": 0.6335967300051975, "grad_norm": 6.809379577636719, "learning_rate": 6.278190773297081e-06, "loss": 6.8173, "step": 17981 }, { "epoch": 0.6336319670181558, "grad_norm": 7.922146320343018, "learning_rate": 6.277131509830434e-06, "loss": 7.1521, "step": 17982 }, { "epoch": 0.6336672040311143, "grad_norm": 3.6691436767578125, "learning_rate": 6.276072294854626e-06, "loss": 3.181, "step": 17983 }, { "epoch": 0.6337024410440727, "grad_norm": 2.9066460132598877, "learning_rate": 6.275013128383451e-06, "loss": 2.2215, "step": 17984 }, { "epoch": 0.6337376780570311, "grad_norm": 2.7904796600341797, "learning_rate": 6.2739540104307055e-06, "loss": 2.8995, "step": 17985 }, { "epoch": 0.6337729150699896, "grad_norm": 3.795384407043457, "learning_rate": 6.2728949410101795e-06, "loss": 2.6893, "step": 17986 }, { "epoch": 0.6338081520829479, "grad_norm": 3.7133359909057617, "learning_rate": 6.271835920135675e-06, "loss": 2.9304, "step": 17987 }, { "epoch": 0.6338433890959063, "grad_norm": 6.9140448570251465, "learning_rate": 6.270776947820981e-06, "loss": 5.5155, "step": 17988 }, { "epoch": 0.6338786261088648, "grad_norm": 2.672628879547119, "learning_rate": 6.269718024079893e-06, "loss": 2.6468, "step": 17989 }, { "epoch": 0.6339138631218232, "grad_norm": 0.954024076461792, "learning_rate": 6.268659148926204e-06, "loss": 0.8936, "step": 17990 }, { "epoch": 0.6339491001347816, "grad_norm": 1.121593713760376, "learning_rate": 6.267600322373702e-06, "loss": 0.8213, "step": 17991 }, { "epoch": 0.6339843371477399, "grad_norm": 4.2265944480896, "learning_rate": 6.266541544436182e-06, "loss": 3.1376, "step": 17992 }, { "epoch": 0.6340195741606984, "grad_norm": 1.256559133529663, "learning_rate": 6.265482815127435e-06, "loss": 1.0463, "step": 17993 }, { "epoch": 0.6340548111736568, "grad_norm": 1.0122822523117065, "learning_rate": 6.264424134461251e-06, "loss": 0.7596, "step": 17994 }, { "epoch": 0.6340900481866152, "grad_norm": 4.330949306488037, "learning_rate": 6.263365502451415e-06, "loss": 2.8322, "step": 17995 }, { "epoch": 0.6341252851995737, "grad_norm": 1.4071012735366821, "learning_rate": 6.262306919111722e-06, "loss": 1.0674, "step": 17996 }, { "epoch": 0.634160522212532, "grad_norm": 6.289210319519043, "learning_rate": 6.261248384455959e-06, "loss": 4.3976, "step": 17997 }, { "epoch": 0.6341957592254904, "grad_norm": 1.8071361780166626, "learning_rate": 6.260189898497911e-06, "loss": 1.1542, "step": 17998 }, { "epoch": 0.6342309962384489, "grad_norm": 1.12715482711792, "learning_rate": 6.259131461251362e-06, "loss": 0.9253, "step": 17999 }, { "epoch": 0.6342662332514073, "grad_norm": 0.9446585178375244, "learning_rate": 6.2580730727301085e-06, "loss": 0.7875, "step": 18000 }, { "epoch": 0.6343014702643657, "grad_norm": 1.543677806854248, "learning_rate": 6.257014732947925e-06, "loss": 0.7987, "step": 18001 }, { "epoch": 0.634336707277324, "grad_norm": 4.558265686035156, "learning_rate": 6.255956441918603e-06, "loss": 4.8366, "step": 18002 }, { "epoch": 0.6343719442902825, "grad_norm": 3.3714215755462646, "learning_rate": 6.254898199655928e-06, "loss": 3.1157, "step": 18003 }, { "epoch": 0.6344071813032409, "grad_norm": 3.568817615509033, "learning_rate": 6.253840006173677e-06, "loss": 2.9685, "step": 18004 }, { "epoch": 0.6344424183161993, "grad_norm": 1.3683863878250122, "learning_rate": 6.252781861485636e-06, "loss": 0.6841, "step": 18005 }, { "epoch": 0.6344776553291578, "grad_norm": 4.253195762634277, "learning_rate": 6.251723765605593e-06, "loss": 2.934, "step": 18006 }, { "epoch": 0.6345128923421162, "grad_norm": 3.3019325733184814, "learning_rate": 6.250665718547324e-06, "loss": 2.7939, "step": 18007 }, { "epoch": 0.6345481293550745, "grad_norm": 2.9464313983917236, "learning_rate": 6.2496077203246105e-06, "loss": 2.4545, "step": 18008 }, { "epoch": 0.634583366368033, "grad_norm": 5.295986652374268, "learning_rate": 6.248549770951235e-06, "loss": 3.3377, "step": 18009 }, { "epoch": 0.6346186033809914, "grad_norm": 4.753384590148926, "learning_rate": 6.247491870440976e-06, "loss": 2.6853, "step": 18010 }, { "epoch": 0.6346538403939498, "grad_norm": 5.1155877113342285, "learning_rate": 6.2464340188076124e-06, "loss": 5.3625, "step": 18011 }, { "epoch": 0.6346890774069082, "grad_norm": 3.0164287090301514, "learning_rate": 6.245376216064925e-06, "loss": 0.8164, "step": 18012 }, { "epoch": 0.6347243144198667, "grad_norm": 1.3051363229751587, "learning_rate": 6.244318462226693e-06, "loss": 0.9783, "step": 18013 }, { "epoch": 0.634759551432825, "grad_norm": 5.921405792236328, "learning_rate": 6.243260757306686e-06, "loss": 4.693, "step": 18014 }, { "epoch": 0.6347947884457834, "grad_norm": 4.777769565582275, "learning_rate": 6.242203101318689e-06, "loss": 4.8217, "step": 18015 }, { "epoch": 0.6348300254587419, "grad_norm": 1.232640027999878, "learning_rate": 6.241145494276478e-06, "loss": 0.861, "step": 18016 }, { "epoch": 0.6348652624717003, "grad_norm": 3.1611409187316895, "learning_rate": 6.240087936193824e-06, "loss": 3.0652, "step": 18017 }, { "epoch": 0.6349004994846587, "grad_norm": 4.354402542114258, "learning_rate": 6.2390304270845e-06, "loss": 3.1641, "step": 18018 }, { "epoch": 0.6349357364976171, "grad_norm": 2.931032657623291, "learning_rate": 6.237972966962291e-06, "loss": 0.823, "step": 18019 }, { "epoch": 0.6349709735105755, "grad_norm": 1.2990329265594482, "learning_rate": 6.2369155558409565e-06, "loss": 0.737, "step": 18020 }, { "epoch": 0.6350062105235339, "grad_norm": 1.9030706882476807, "learning_rate": 6.235858193734279e-06, "loss": 1.1012, "step": 18021 }, { "epoch": 0.6350414475364923, "grad_norm": 4.575558662414551, "learning_rate": 6.234800880656029e-06, "loss": 3.2919, "step": 18022 }, { "epoch": 0.6350766845494508, "grad_norm": 4.611262798309326, "learning_rate": 6.233743616619976e-06, "loss": 2.9394, "step": 18023 }, { "epoch": 0.6351119215624091, "grad_norm": 3.961540937423706, "learning_rate": 6.232686401639889e-06, "loss": 2.887, "step": 18024 }, { "epoch": 0.6351471585753675, "grad_norm": 1.758086919784546, "learning_rate": 6.231629235729546e-06, "loss": 1.0794, "step": 18025 }, { "epoch": 0.635182395588326, "grad_norm": 1.0556151866912842, "learning_rate": 6.230572118902711e-06, "loss": 0.9788, "step": 18026 }, { "epoch": 0.6352176326012844, "grad_norm": 9.311480522155762, "learning_rate": 6.229515051173154e-06, "loss": 7.7202, "step": 18027 }, { "epoch": 0.6352528696142428, "grad_norm": 5.393751621246338, "learning_rate": 6.2284580325546405e-06, "loss": 3.1749, "step": 18028 }, { "epoch": 0.6352881066272013, "grad_norm": 1.0987337827682495, "learning_rate": 6.2274010630609474e-06, "loss": 0.8111, "step": 18029 }, { "epoch": 0.6353233436401596, "grad_norm": 3.0079965591430664, "learning_rate": 6.226344142705834e-06, "loss": 3.1355, "step": 18030 }, { "epoch": 0.635358580653118, "grad_norm": 4.308621883392334, "learning_rate": 6.225287271503067e-06, "loss": 2.7827, "step": 18031 }, { "epoch": 0.6353938176660765, "grad_norm": 10.228650093078613, "learning_rate": 6.224230449466417e-06, "loss": 3.4259, "step": 18032 }, { "epoch": 0.6354290546790349, "grad_norm": 4.759636402130127, "learning_rate": 6.223173676609643e-06, "loss": 2.9338, "step": 18033 }, { "epoch": 0.6354642916919933, "grad_norm": 4.022433757781982, "learning_rate": 6.222116952946514e-06, "loss": 2.6362, "step": 18034 }, { "epoch": 0.6354995287049516, "grad_norm": 7.562189102172852, "learning_rate": 6.221060278490796e-06, "loss": 5.3356, "step": 18035 }, { "epoch": 0.6355347657179101, "grad_norm": 1.6446794271469116, "learning_rate": 6.2200036532562454e-06, "loss": 0.8353, "step": 18036 }, { "epoch": 0.6355700027308685, "grad_norm": 5.444246768951416, "learning_rate": 6.218947077256628e-06, "loss": 3.3597, "step": 18037 }, { "epoch": 0.6356052397438269, "grad_norm": 0.9840380549430847, "learning_rate": 6.217890550505713e-06, "loss": 1.2278, "step": 18038 }, { "epoch": 0.6356404767567854, "grad_norm": 3.012395143508911, "learning_rate": 6.216834073017248e-06, "loss": 3.2833, "step": 18039 }, { "epoch": 0.6356757137697437, "grad_norm": 1.1043267250061035, "learning_rate": 6.2157776448050045e-06, "loss": 0.6961, "step": 18040 }, { "epoch": 0.6357109507827021, "grad_norm": 1.0335429906845093, "learning_rate": 6.214721265882742e-06, "loss": 0.7863, "step": 18041 }, { "epoch": 0.6357461877956606, "grad_norm": 1.209997534751892, "learning_rate": 6.213664936264212e-06, "loss": 1.0983, "step": 18042 }, { "epoch": 0.635781424808619, "grad_norm": 1.20372474193573, "learning_rate": 6.212608655963179e-06, "loss": 0.7127, "step": 18043 }, { "epoch": 0.6358166618215774, "grad_norm": 3.2628989219665527, "learning_rate": 6.211552424993404e-06, "loss": 2.8991, "step": 18044 }, { "epoch": 0.6358518988345357, "grad_norm": 3.1668901443481445, "learning_rate": 6.210496243368639e-06, "loss": 2.6357, "step": 18045 }, { "epoch": 0.6358871358474942, "grad_norm": 5.0242486000061035, "learning_rate": 6.209440111102642e-06, "loss": 2.673, "step": 18046 }, { "epoch": 0.6359223728604526, "grad_norm": 5.120606422424316, "learning_rate": 6.208384028209171e-06, "loss": 5.0862, "step": 18047 }, { "epoch": 0.635957609873411, "grad_norm": 4.5260539054870605, "learning_rate": 6.207327994701984e-06, "loss": 3.0772, "step": 18048 }, { "epoch": 0.6359928468863695, "grad_norm": 4.374147891998291, "learning_rate": 6.206272010594831e-06, "loss": 2.4864, "step": 18049 }, { "epoch": 0.6360280838993279, "grad_norm": 4.504182815551758, "learning_rate": 6.205216075901468e-06, "loss": 2.1518, "step": 18050 }, { "epoch": 0.6360633209122862, "grad_norm": 1.2235883474349976, "learning_rate": 6.20416019063565e-06, "loss": 1.0173, "step": 18051 }, { "epoch": 0.6360985579252447, "grad_norm": 2.558793544769287, "learning_rate": 6.203104354811127e-06, "loss": 3.0761, "step": 18052 }, { "epoch": 0.6361337949382031, "grad_norm": 4.447798728942871, "learning_rate": 6.202048568441654e-06, "loss": 2.7691, "step": 18053 }, { "epoch": 0.6361690319511615, "grad_norm": 2.2668025493621826, "learning_rate": 6.200992831540984e-06, "loss": 0.8942, "step": 18054 }, { "epoch": 0.6362042689641199, "grad_norm": 3.2539944648742676, "learning_rate": 6.199937144122865e-06, "loss": 3.0643, "step": 18055 }, { "epoch": 0.6362395059770783, "grad_norm": 1.1886506080627441, "learning_rate": 6.198881506201046e-06, "loss": 0.8938, "step": 18056 }, { "epoch": 0.6362747429900367, "grad_norm": 1.2519279718399048, "learning_rate": 6.197825917789287e-06, "loss": 1.0516, "step": 18057 }, { "epoch": 0.6363099800029951, "grad_norm": 1.6368368864059448, "learning_rate": 6.196770378901323e-06, "loss": 1.0089, "step": 18058 }, { "epoch": 0.6363452170159536, "grad_norm": 3.1930179595947266, "learning_rate": 6.195714889550912e-06, "loss": 0.6695, "step": 18059 }, { "epoch": 0.636380454028912, "grad_norm": 0.7383262515068054, "learning_rate": 6.194659449751801e-06, "loss": 0.6602, "step": 18060 }, { "epoch": 0.6364156910418703, "grad_norm": 1.548505187034607, "learning_rate": 6.193604059517733e-06, "loss": 0.9667, "step": 18061 }, { "epoch": 0.6364509280548288, "grad_norm": 2.398937463760376, "learning_rate": 6.192548718862454e-06, "loss": 3.1106, "step": 18062 }, { "epoch": 0.6364861650677872, "grad_norm": 1.359999418258667, "learning_rate": 6.191493427799718e-06, "loss": 1.0584, "step": 18063 }, { "epoch": 0.6365214020807456, "grad_norm": 1.2308937311172485, "learning_rate": 6.190438186343263e-06, "loss": 1.0597, "step": 18064 }, { "epoch": 0.6365566390937041, "grad_norm": 2.882967472076416, "learning_rate": 6.189382994506835e-06, "loss": 3.0896, "step": 18065 }, { "epoch": 0.6365918761066625, "grad_norm": 4.7698163986206055, "learning_rate": 6.1883278523041775e-06, "loss": 2.8073, "step": 18066 }, { "epoch": 0.6366271131196208, "grad_norm": 1.886254906654358, "learning_rate": 6.18727275974904e-06, "loss": 0.8674, "step": 18067 }, { "epoch": 0.6366623501325792, "grad_norm": 1.9159510135650635, "learning_rate": 6.186217716855158e-06, "loss": 0.967, "step": 18068 }, { "epoch": 0.6366975871455377, "grad_norm": 0.8338581919670105, "learning_rate": 6.185162723636275e-06, "loss": 0.7554, "step": 18069 }, { "epoch": 0.6367328241584961, "grad_norm": 5.024890422821045, "learning_rate": 6.184107780106136e-06, "loss": 4.6452, "step": 18070 }, { "epoch": 0.6367680611714545, "grad_norm": 5.122786998748779, "learning_rate": 6.183052886278478e-06, "loss": 4.3854, "step": 18071 }, { "epoch": 0.636803298184413, "grad_norm": 4.055861949920654, "learning_rate": 6.181998042167041e-06, "loss": 3.0614, "step": 18072 }, { "epoch": 0.6368385351973713, "grad_norm": 5.203192234039307, "learning_rate": 6.180943247785569e-06, "loss": 3.0427, "step": 18073 }, { "epoch": 0.6368737722103297, "grad_norm": 4.486349105834961, "learning_rate": 6.179888503147795e-06, "loss": 3.4133, "step": 18074 }, { "epoch": 0.6369090092232882, "grad_norm": 2.8658432960510254, "learning_rate": 6.178833808267459e-06, "loss": 2.8038, "step": 18075 }, { "epoch": 0.6369442462362466, "grad_norm": 1.4165315628051758, "learning_rate": 6.177779163158304e-06, "loss": 0.9427, "step": 18076 }, { "epoch": 0.636979483249205, "grad_norm": 2.4704763889312744, "learning_rate": 6.1767245678340585e-06, "loss": 0.9135, "step": 18077 }, { "epoch": 0.6370147202621633, "grad_norm": 4.855730056762695, "learning_rate": 6.175670022308462e-06, "loss": 3.4105, "step": 18078 }, { "epoch": 0.6370499572751218, "grad_norm": 3.785778045654297, "learning_rate": 6.174615526595253e-06, "loss": 3.2044, "step": 18079 }, { "epoch": 0.6370851942880802, "grad_norm": 2.845595359802246, "learning_rate": 6.1735610807081635e-06, "loss": 2.8729, "step": 18080 }, { "epoch": 0.6371204313010386, "grad_norm": 5.2562336921691895, "learning_rate": 6.172506684660926e-06, "loss": 4.7942, "step": 18081 }, { "epoch": 0.6371556683139971, "grad_norm": 1.7654982805252075, "learning_rate": 6.17145233846728e-06, "loss": 0.7637, "step": 18082 }, { "epoch": 0.6371909053269554, "grad_norm": 1.6163146495819092, "learning_rate": 6.170398042140953e-06, "loss": 0.7451, "step": 18083 }, { "epoch": 0.6372261423399138, "grad_norm": 3.0162932872772217, "learning_rate": 6.169343795695679e-06, "loss": 2.4119, "step": 18084 }, { "epoch": 0.6372613793528723, "grad_norm": 1.2990199327468872, "learning_rate": 6.168289599145187e-06, "loss": 0.8584, "step": 18085 }, { "epoch": 0.6372966163658307, "grad_norm": 6.392269134521484, "learning_rate": 6.1672354525032175e-06, "loss": 3.2182, "step": 18086 }, { "epoch": 0.6373318533787891, "grad_norm": 11.746374130249023, "learning_rate": 6.166181355783492e-06, "loss": 5.2896, "step": 18087 }, { "epoch": 0.6373670903917474, "grad_norm": 2.7857372760772705, "learning_rate": 6.165127308999741e-06, "loss": 3.4045, "step": 18088 }, { "epoch": 0.6374023274047059, "grad_norm": 1.9722800254821777, "learning_rate": 6.164073312165699e-06, "loss": 0.8186, "step": 18089 }, { "epoch": 0.6374375644176643, "grad_norm": 4.353124618530273, "learning_rate": 6.163019365295085e-06, "loss": 4.723, "step": 18090 }, { "epoch": 0.6374728014306227, "grad_norm": 1.1755398511886597, "learning_rate": 6.161965468401636e-06, "loss": 1.4157, "step": 18091 }, { "epoch": 0.6375080384435812, "grad_norm": 7.760361671447754, "learning_rate": 6.160911621499077e-06, "loss": 4.9653, "step": 18092 }, { "epoch": 0.6375432754565395, "grad_norm": 1.4076553583145142, "learning_rate": 6.159857824601132e-06, "loss": 0.9721, "step": 18093 }, { "epoch": 0.6375785124694979, "grad_norm": 6.720585823059082, "learning_rate": 6.158804077721525e-06, "loss": 4.5052, "step": 18094 }, { "epoch": 0.6376137494824564, "grad_norm": 1.546317219734192, "learning_rate": 6.15775038087399e-06, "loss": 0.8226, "step": 18095 }, { "epoch": 0.6376489864954148, "grad_norm": 5.098783016204834, "learning_rate": 6.156696734072241e-06, "loss": 4.8832, "step": 18096 }, { "epoch": 0.6376842235083732, "grad_norm": 4.746749401092529, "learning_rate": 6.1556431373300075e-06, "loss": 3.2717, "step": 18097 }, { "epoch": 0.6377194605213317, "grad_norm": 0.7889241576194763, "learning_rate": 6.154589590661014e-06, "loss": 1.1282, "step": 18098 }, { "epoch": 0.63775469753429, "grad_norm": 3.997206211090088, "learning_rate": 6.15353609407898e-06, "loss": 2.479, "step": 18099 }, { "epoch": 0.6377899345472484, "grad_norm": 1.3098814487457275, "learning_rate": 6.152482647597624e-06, "loss": 0.8578, "step": 18100 }, { "epoch": 0.6378251715602068, "grad_norm": 5.7410759925842285, "learning_rate": 6.151429251230676e-06, "loss": 5.607, "step": 18101 }, { "epoch": 0.6378604085731653, "grad_norm": 5.7477898597717285, "learning_rate": 6.150375904991852e-06, "loss": 4.9659, "step": 18102 }, { "epoch": 0.6378956455861237, "grad_norm": 3.323629856109619, "learning_rate": 6.149322608894872e-06, "loss": 2.6436, "step": 18103 }, { "epoch": 0.637930882599082, "grad_norm": 1.3507691621780396, "learning_rate": 6.148269362953454e-06, "loss": 1.1271, "step": 18104 }, { "epoch": 0.6379661196120405, "grad_norm": 1.9560165405273438, "learning_rate": 6.14721616718132e-06, "loss": 1.1056, "step": 18105 }, { "epoch": 0.6380013566249989, "grad_norm": 1.8553446531295776, "learning_rate": 6.146163021592186e-06, "loss": 0.9194, "step": 18106 }, { "epoch": 0.6380365936379573, "grad_norm": 1.3854202032089233, "learning_rate": 6.145109926199768e-06, "loss": 0.8916, "step": 18107 }, { "epoch": 0.6380718306509158, "grad_norm": 6.211848735809326, "learning_rate": 6.1440568810177885e-06, "loss": 4.8097, "step": 18108 }, { "epoch": 0.6381070676638742, "grad_norm": 4.688534259796143, "learning_rate": 6.143003886059953e-06, "loss": 2.6911, "step": 18109 }, { "epoch": 0.6381423046768325, "grad_norm": 4.59518575668335, "learning_rate": 6.141950941339987e-06, "loss": 5.0876, "step": 18110 }, { "epoch": 0.6381775416897909, "grad_norm": 1.8175064325332642, "learning_rate": 6.140898046871602e-06, "loss": 0.6729, "step": 18111 }, { "epoch": 0.6382127787027494, "grad_norm": 1.2587683200836182, "learning_rate": 6.139845202668509e-06, "loss": 0.5778, "step": 18112 }, { "epoch": 0.6382480157157078, "grad_norm": 6.310168266296387, "learning_rate": 6.138792408744425e-06, "loss": 4.91, "step": 18113 }, { "epoch": 0.6382832527286662, "grad_norm": 3.315598487854004, "learning_rate": 6.137739665113065e-06, "loss": 2.6557, "step": 18114 }, { "epoch": 0.6383184897416246, "grad_norm": 1.0214416980743408, "learning_rate": 6.136686971788133e-06, "loss": 0.9743, "step": 18115 }, { "epoch": 0.638353726754583, "grad_norm": 5.190445899963379, "learning_rate": 6.135634328783346e-06, "loss": 4.9623, "step": 18116 }, { "epoch": 0.6383889637675414, "grad_norm": 4.1954169273376465, "learning_rate": 6.134581736112417e-06, "loss": 2.4537, "step": 18117 }, { "epoch": 0.6384242007804999, "grad_norm": 1.1059566736221313, "learning_rate": 6.133529193789052e-06, "loss": 0.9037, "step": 18118 }, { "epoch": 0.6384594377934583, "grad_norm": 2.6196365356445312, "learning_rate": 6.132476701826957e-06, "loss": 2.9257, "step": 18119 }, { "epoch": 0.6384946748064166, "grad_norm": 2.5417590141296387, "learning_rate": 6.131424260239851e-06, "loss": 0.9288, "step": 18120 }, { "epoch": 0.638529911819375, "grad_norm": 3.1277902126312256, "learning_rate": 6.130371869041437e-06, "loss": 2.7892, "step": 18121 }, { "epoch": 0.6385651488323335, "grad_norm": 2.790415048599243, "learning_rate": 6.129319528245421e-06, "loss": 0.9612, "step": 18122 }, { "epoch": 0.6386003858452919, "grad_norm": 1.639682412147522, "learning_rate": 6.128267237865508e-06, "loss": 0.8939, "step": 18123 }, { "epoch": 0.6386356228582503, "grad_norm": 0.7626342177391052, "learning_rate": 6.127214997915413e-06, "loss": 0.8687, "step": 18124 }, { "epoch": 0.6386708598712088, "grad_norm": 4.6380510330200195, "learning_rate": 6.126162808408833e-06, "loss": 3.2633, "step": 18125 }, { "epoch": 0.6387060968841671, "grad_norm": 4.754288196563721, "learning_rate": 6.125110669359475e-06, "loss": 2.9736, "step": 18126 }, { "epoch": 0.6387413338971255, "grad_norm": 1.264269471168518, "learning_rate": 6.124058580781048e-06, "loss": 1.2124, "step": 18127 }, { "epoch": 0.638776570910084, "grad_norm": 0.8322306871414185, "learning_rate": 6.1230065426872485e-06, "loss": 0.8717, "step": 18128 }, { "epoch": 0.6388118079230424, "grad_norm": 2.9443862438201904, "learning_rate": 6.1219545550917815e-06, "loss": 3.0712, "step": 18129 }, { "epoch": 0.6388470449360008, "grad_norm": 6.295992374420166, "learning_rate": 6.120902618008354e-06, "loss": 4.7437, "step": 18130 }, { "epoch": 0.6388822819489592, "grad_norm": 4.357635498046875, "learning_rate": 6.119850731450663e-06, "loss": 5.0011, "step": 18131 }, { "epoch": 0.6389175189619176, "grad_norm": 4.216884613037109, "learning_rate": 6.118798895432408e-06, "loss": 2.9374, "step": 18132 }, { "epoch": 0.638952755974876, "grad_norm": 8.317116737365723, "learning_rate": 6.117747109967296e-06, "loss": 5.0887, "step": 18133 }, { "epoch": 0.6389879929878344, "grad_norm": 6.378798007965088, "learning_rate": 6.116695375069017e-06, "loss": 2.7968, "step": 18134 }, { "epoch": 0.6390232300007929, "grad_norm": 3.3282179832458496, "learning_rate": 6.115643690751279e-06, "loss": 2.6777, "step": 18135 }, { "epoch": 0.6390584670137512, "grad_norm": 4.647929668426514, "learning_rate": 6.114592057027777e-06, "loss": 3.4498, "step": 18136 }, { "epoch": 0.6390937040267096, "grad_norm": 1.3374221324920654, "learning_rate": 6.113540473912208e-06, "loss": 0.9424, "step": 18137 }, { "epoch": 0.6391289410396681, "grad_norm": 1.1971646547317505, "learning_rate": 6.1124889414182665e-06, "loss": 0.9365, "step": 18138 }, { "epoch": 0.6391641780526265, "grad_norm": 1.7606912851333618, "learning_rate": 6.1114374595596525e-06, "loss": 0.7657, "step": 18139 }, { "epoch": 0.6391994150655849, "grad_norm": 1.450932264328003, "learning_rate": 6.110386028350065e-06, "loss": 1.0341, "step": 18140 }, { "epoch": 0.6392346520785434, "grad_norm": 3.9287333488464355, "learning_rate": 6.1093346478031915e-06, "loss": 3.21, "step": 18141 }, { "epoch": 0.6392698890915017, "grad_norm": 5.292864799499512, "learning_rate": 6.108283317932728e-06, "loss": 3.0553, "step": 18142 }, { "epoch": 0.6393051261044601, "grad_norm": 1.0852129459381104, "learning_rate": 6.107232038752378e-06, "loss": 0.7049, "step": 18143 }, { "epoch": 0.6393403631174185, "grad_norm": 3.4557385444641113, "learning_rate": 6.106180810275818e-06, "loss": 3.0216, "step": 18144 }, { "epoch": 0.639375600130377, "grad_norm": 4.351437568664551, "learning_rate": 6.105129632516752e-06, "loss": 4.8419, "step": 18145 }, { "epoch": 0.6394108371433354, "grad_norm": 1.0320135354995728, "learning_rate": 6.104078505488871e-06, "loss": 1.2, "step": 18146 }, { "epoch": 0.6394460741562937, "grad_norm": 0.9366640448570251, "learning_rate": 6.103027429205861e-06, "loss": 1.1049, "step": 18147 }, { "epoch": 0.6394813111692522, "grad_norm": 5.342625141143799, "learning_rate": 6.101976403681414e-06, "loss": 5.1511, "step": 18148 }, { "epoch": 0.6395165481822106, "grad_norm": 5.928982257843018, "learning_rate": 6.100925428929223e-06, "loss": 4.6249, "step": 18149 }, { "epoch": 0.639551785195169, "grad_norm": 1.2581851482391357, "learning_rate": 6.0998745049629745e-06, "loss": 0.6439, "step": 18150 }, { "epoch": 0.6395870222081275, "grad_norm": 1.2927831411361694, "learning_rate": 6.098823631796358e-06, "loss": 0.9245, "step": 18151 }, { "epoch": 0.6396222592210858, "grad_norm": 1.3412210941314697, "learning_rate": 6.097772809443061e-06, "loss": 0.6519, "step": 18152 }, { "epoch": 0.6396574962340442, "grad_norm": 2.9591548442840576, "learning_rate": 6.096722037916767e-06, "loss": 0.8271, "step": 18153 }, { "epoch": 0.6396927332470026, "grad_norm": 0.8966507911682129, "learning_rate": 6.095671317231168e-06, "loss": 0.9209, "step": 18154 }, { "epoch": 0.6397279702599611, "grad_norm": 2.5752196311950684, "learning_rate": 6.0946206473999496e-06, "loss": 2.2045, "step": 18155 }, { "epoch": 0.6397632072729195, "grad_norm": 4.058794021606445, "learning_rate": 6.093570028436791e-06, "loss": 2.8964, "step": 18156 }, { "epoch": 0.6397984442858778, "grad_norm": 1.4818145036697388, "learning_rate": 6.09251946035538e-06, "loss": 0.9931, "step": 18157 }, { "epoch": 0.6398336812988363, "grad_norm": 5.509145259857178, "learning_rate": 6.091468943169402e-06, "loss": 5.3801, "step": 18158 }, { "epoch": 0.6398689183117947, "grad_norm": 3.69158673286438, "learning_rate": 6.090418476892541e-06, "loss": 3.6712, "step": 18159 }, { "epoch": 0.6399041553247531, "grad_norm": 5.857550621032715, "learning_rate": 6.089368061538476e-06, "loss": 4.5457, "step": 18160 }, { "epoch": 0.6399393923377116, "grad_norm": 3.642185688018799, "learning_rate": 6.088317697120887e-06, "loss": 2.9923, "step": 18161 }, { "epoch": 0.63997462935067, "grad_norm": 4.356273174285889, "learning_rate": 6.087267383653466e-06, "loss": 2.8483, "step": 18162 }, { "epoch": 0.6400098663636283, "grad_norm": 1.4044703245162964, "learning_rate": 6.086217121149877e-06, "loss": 1.0458, "step": 18163 }, { "epoch": 0.6400451033765868, "grad_norm": 4.236495494842529, "learning_rate": 6.085166909623813e-06, "loss": 3.0575, "step": 18164 }, { "epoch": 0.6400803403895452, "grad_norm": 2.6526803970336914, "learning_rate": 6.084116749088951e-06, "loss": 2.9639, "step": 18165 }, { "epoch": 0.6401155774025036, "grad_norm": 1.008303165435791, "learning_rate": 6.083066639558963e-06, "loss": 0.9163, "step": 18166 }, { "epoch": 0.640150814415462, "grad_norm": 6.772192478179932, "learning_rate": 6.08201658104753e-06, "loss": 2.781, "step": 18167 }, { "epoch": 0.6401860514284204, "grad_norm": 4.355175971984863, "learning_rate": 6.080966573568334e-06, "loss": 3.1334, "step": 18168 }, { "epoch": 0.6402212884413788, "grad_norm": 1.2277635335922241, "learning_rate": 6.079916617135046e-06, "loss": 0.9746, "step": 18169 }, { "epoch": 0.6402565254543372, "grad_norm": 1.2648792266845703, "learning_rate": 6.0788667117613424e-06, "loss": 0.7308, "step": 18170 }, { "epoch": 0.6402917624672957, "grad_norm": 2.856565237045288, "learning_rate": 6.077816857460903e-06, "loss": 3.2863, "step": 18171 }, { "epoch": 0.6403269994802541, "grad_norm": 3.484267234802246, "learning_rate": 6.076767054247393e-06, "loss": 2.8786, "step": 18172 }, { "epoch": 0.6403622364932124, "grad_norm": 3.3100638389587402, "learning_rate": 6.075717302134494e-06, "loss": 2.6382, "step": 18173 }, { "epoch": 0.6403974735061709, "grad_norm": 2.959625482559204, "learning_rate": 6.074667601135878e-06, "loss": 2.9915, "step": 18174 }, { "epoch": 0.6404327105191293, "grad_norm": 4.067081928253174, "learning_rate": 6.073617951265218e-06, "loss": 2.6055, "step": 18175 }, { "epoch": 0.6404679475320877, "grad_norm": 3.679516553878784, "learning_rate": 6.072568352536181e-06, "loss": 3.4399, "step": 18176 }, { "epoch": 0.6405031845450461, "grad_norm": 4.862331390380859, "learning_rate": 6.0715188049624414e-06, "loss": 0.699, "step": 18177 }, { "epoch": 0.6405384215580046, "grad_norm": 1.156342625617981, "learning_rate": 6.0704693085576736e-06, "loss": 1.0697, "step": 18178 }, { "epoch": 0.6405736585709629, "grad_norm": 5.063096046447754, "learning_rate": 6.0694198633355415e-06, "loss": 4.6727, "step": 18179 }, { "epoch": 0.6406088955839213, "grad_norm": 1.8405343294143677, "learning_rate": 6.068370469309715e-06, "loss": 0.7144, "step": 18180 }, { "epoch": 0.6406441325968798, "grad_norm": 1.3837451934814453, "learning_rate": 6.067321126493868e-06, "loss": 0.7821, "step": 18181 }, { "epoch": 0.6406793696098382, "grad_norm": 2.1186254024505615, "learning_rate": 6.066271834901661e-06, "loss": 0.8514, "step": 18182 }, { "epoch": 0.6407146066227966, "grad_norm": 4.002279758453369, "learning_rate": 6.065222594546763e-06, "loss": 3.1416, "step": 18183 }, { "epoch": 0.640749843635755, "grad_norm": 0.939855694770813, "learning_rate": 6.064173405442846e-06, "loss": 0.8438, "step": 18184 }, { "epoch": 0.6407850806487134, "grad_norm": 0.8396545052528381, "learning_rate": 6.06312426760357e-06, "loss": 1.1768, "step": 18185 }, { "epoch": 0.6408203176616718, "grad_norm": 3.271789312362671, "learning_rate": 6.0620751810425994e-06, "loss": 2.8687, "step": 18186 }, { "epoch": 0.6408555546746302, "grad_norm": 3.666563034057617, "learning_rate": 6.061026145773604e-06, "loss": 2.8881, "step": 18187 }, { "epoch": 0.6408907916875887, "grad_norm": 3.6862030029296875, "learning_rate": 6.059977161810245e-06, "loss": 2.7186, "step": 18188 }, { "epoch": 0.640926028700547, "grad_norm": 4.028059959411621, "learning_rate": 6.058928229166183e-06, "loss": 4.4748, "step": 18189 }, { "epoch": 0.6409612657135054, "grad_norm": 2.7681992053985596, "learning_rate": 6.057879347855087e-06, "loss": 2.22, "step": 18190 }, { "epoch": 0.6409965027264639, "grad_norm": 2.801551580429077, "learning_rate": 6.056830517890608e-06, "loss": 2.5111, "step": 18191 }, { "epoch": 0.6410317397394223, "grad_norm": 4.616788864135742, "learning_rate": 6.055781739286418e-06, "loss": 3.0058, "step": 18192 }, { "epoch": 0.6410669767523807, "grad_norm": 6.633090496063232, "learning_rate": 6.054733012056173e-06, "loss": 4.9366, "step": 18193 }, { "epoch": 0.6411022137653392, "grad_norm": 3.042417526245117, "learning_rate": 6.0536843362135346e-06, "loss": 2.5673, "step": 18194 }, { "epoch": 0.6411374507782975, "grad_norm": 4.562692642211914, "learning_rate": 6.052635711772157e-06, "loss": 2.8373, "step": 18195 }, { "epoch": 0.6411726877912559, "grad_norm": 3.8500797748565674, "learning_rate": 6.051587138745703e-06, "loss": 2.661, "step": 18196 }, { "epoch": 0.6412079248042144, "grad_norm": 3.8958141803741455, "learning_rate": 6.050538617147832e-06, "loss": 2.9001, "step": 18197 }, { "epoch": 0.6412431618171728, "grad_norm": 1.6433991193771362, "learning_rate": 6.0494901469921965e-06, "loss": 0.8711, "step": 18198 }, { "epoch": 0.6412783988301312, "grad_norm": 4.857601165771484, "learning_rate": 6.0484417282924535e-06, "loss": 3.2091, "step": 18199 }, { "epoch": 0.6413136358430895, "grad_norm": 4.662984848022461, "learning_rate": 6.047393361062266e-06, "loss": 2.7173, "step": 18200 }, { "epoch": 0.641348872856048, "grad_norm": 1.222583532333374, "learning_rate": 6.046345045315278e-06, "loss": 0.7661, "step": 18201 }, { "epoch": 0.6413841098690064, "grad_norm": 1.3226683139801025, "learning_rate": 6.045296781065151e-06, "loss": 0.8151, "step": 18202 }, { "epoch": 0.6414193468819648, "grad_norm": 3.6438474655151367, "learning_rate": 6.044248568325539e-06, "loss": 3.0481, "step": 18203 }, { "epoch": 0.6414545838949233, "grad_norm": 1.1420745849609375, "learning_rate": 6.043200407110092e-06, "loss": 0.9547, "step": 18204 }, { "epoch": 0.6414898209078816, "grad_norm": 4.743288040161133, "learning_rate": 6.04215229743246e-06, "loss": 2.9678, "step": 18205 }, { "epoch": 0.64152505792084, "grad_norm": 2.5072999000549316, "learning_rate": 6.0411042393063035e-06, "loss": 0.9681, "step": 18206 }, { "epoch": 0.6415602949337985, "grad_norm": 2.969533920288086, "learning_rate": 6.0400562327452675e-06, "loss": 2.1889, "step": 18207 }, { "epoch": 0.6415955319467569, "grad_norm": 4.000550270080566, "learning_rate": 6.039008277763002e-06, "loss": 2.7481, "step": 18208 }, { "epoch": 0.6416307689597153, "grad_norm": 4.424300193786621, "learning_rate": 6.037960374373162e-06, "loss": 2.6346, "step": 18209 }, { "epoch": 0.6416660059726736, "grad_norm": 3.533101797103882, "learning_rate": 6.036912522589387e-06, "loss": 3.288, "step": 18210 }, { "epoch": 0.6417012429856321, "grad_norm": 1.8020570278167725, "learning_rate": 6.035864722425333e-06, "loss": 0.7735, "step": 18211 }, { "epoch": 0.6417364799985905, "grad_norm": 4.875016212463379, "learning_rate": 6.034816973894647e-06, "loss": 2.4311, "step": 18212 }, { "epoch": 0.6417717170115489, "grad_norm": 5.600376129150391, "learning_rate": 6.033769277010977e-06, "loss": 4.884, "step": 18213 }, { "epoch": 0.6418069540245074, "grad_norm": 6.076886177062988, "learning_rate": 6.0327216317879635e-06, "loss": 7.1615, "step": 18214 }, { "epoch": 0.6418421910374658, "grad_norm": 3.8973944187164307, "learning_rate": 6.0316740382392566e-06, "loss": 3.0626, "step": 18215 }, { "epoch": 0.6418774280504241, "grad_norm": 1.4242204427719116, "learning_rate": 6.030626496378504e-06, "loss": 0.9811, "step": 18216 }, { "epoch": 0.6419126650633826, "grad_norm": 1.1260638236999512, "learning_rate": 6.029579006219346e-06, "loss": 0.8691, "step": 18217 }, { "epoch": 0.641947902076341, "grad_norm": 7.379639148712158, "learning_rate": 6.028531567775424e-06, "loss": 5.7153, "step": 18218 }, { "epoch": 0.6419831390892994, "grad_norm": 8.708712577819824, "learning_rate": 6.0274841810603905e-06, "loss": 7.3553, "step": 18219 }, { "epoch": 0.6420183761022578, "grad_norm": 5.329339981079102, "learning_rate": 6.026436846087875e-06, "loss": 2.5433, "step": 18220 }, { "epoch": 0.6420536131152162, "grad_norm": 2.340087652206421, "learning_rate": 6.025389562871528e-06, "loss": 0.6874, "step": 18221 }, { "epoch": 0.6420888501281746, "grad_norm": 4.427449703216553, "learning_rate": 6.024342331424991e-06, "loss": 3.003, "step": 18222 }, { "epoch": 0.642124087141133, "grad_norm": 1.5368213653564453, "learning_rate": 6.023295151761899e-06, "loss": 1.157, "step": 18223 }, { "epoch": 0.6421593241540915, "grad_norm": 2.312903881072998, "learning_rate": 6.0222480238958914e-06, "loss": 0.8429, "step": 18224 }, { "epoch": 0.6421945611670499, "grad_norm": 4.900276184082031, "learning_rate": 6.021200947840615e-06, "loss": 2.9388, "step": 18225 }, { "epoch": 0.6422297981800082, "grad_norm": 10.14102554321289, "learning_rate": 6.0201539236097005e-06, "loss": 5.389, "step": 18226 }, { "epoch": 0.6422650351929667, "grad_norm": 3.878484010696411, "learning_rate": 6.019106951216789e-06, "loss": 2.5696, "step": 18227 }, { "epoch": 0.6423002722059251, "grad_norm": 6.57135009765625, "learning_rate": 6.018060030675517e-06, "loss": 5.5941, "step": 18228 }, { "epoch": 0.6423355092188835, "grad_norm": 1.2633706331253052, "learning_rate": 6.017013161999518e-06, "loss": 0.6431, "step": 18229 }, { "epoch": 0.6423707462318419, "grad_norm": 1.1925036907196045, "learning_rate": 6.015966345202431e-06, "loss": 0.8104, "step": 18230 }, { "epoch": 0.6424059832448004, "grad_norm": 4.720335483551025, "learning_rate": 6.01491958029789e-06, "loss": 5.3767, "step": 18231 }, { "epoch": 0.6424412202577587, "grad_norm": 1.4961564540863037, "learning_rate": 6.01387286729953e-06, "loss": 0.9101, "step": 18232 }, { "epoch": 0.6424764572707171, "grad_norm": 1.0688427686691284, "learning_rate": 6.012826206220982e-06, "loss": 0.8424, "step": 18233 }, { "epoch": 0.6425116942836756, "grad_norm": 1.6429623365402222, "learning_rate": 6.011779597075881e-06, "loss": 0.86, "step": 18234 }, { "epoch": 0.642546931296634, "grad_norm": 3.919621467590332, "learning_rate": 6.010733039877861e-06, "loss": 3.0135, "step": 18235 }, { "epoch": 0.6425821683095924, "grad_norm": 3.123950242996216, "learning_rate": 6.009686534640548e-06, "loss": 2.7973, "step": 18236 }, { "epoch": 0.6426174053225508, "grad_norm": 6.7472243309021, "learning_rate": 6.008640081377577e-06, "loss": 5.1684, "step": 18237 }, { "epoch": 0.6426526423355092, "grad_norm": 3.6720423698425293, "learning_rate": 6.0075936801025814e-06, "loss": 3.0893, "step": 18238 }, { "epoch": 0.6426878793484676, "grad_norm": 3.9624762535095215, "learning_rate": 6.006547330829182e-06, "loss": 2.9644, "step": 18239 }, { "epoch": 0.6427231163614261, "grad_norm": 1.2837682962417603, "learning_rate": 6.005501033571013e-06, "loss": 0.8403, "step": 18240 }, { "epoch": 0.6427583533743845, "grad_norm": 4.084066390991211, "learning_rate": 6.004454788341705e-06, "loss": 5.8223, "step": 18241 }, { "epoch": 0.6427935903873429, "grad_norm": 6.151668071746826, "learning_rate": 6.003408595154878e-06, "loss": 4.7134, "step": 18242 }, { "epoch": 0.6428288274003012, "grad_norm": 6.307146072387695, "learning_rate": 6.002362454024164e-06, "loss": 7.489, "step": 18243 }, { "epoch": 0.6428640644132597, "grad_norm": 1.1205320358276367, "learning_rate": 6.00131636496319e-06, "loss": 0.897, "step": 18244 }, { "epoch": 0.6428993014262181, "grad_norm": 5.003749370574951, "learning_rate": 6.00027032798558e-06, "loss": 4.6379, "step": 18245 }, { "epoch": 0.6429345384391765, "grad_norm": 1.722274661064148, "learning_rate": 5.999224343104957e-06, "loss": 0.6515, "step": 18246 }, { "epoch": 0.642969775452135, "grad_norm": 1.4449825286865234, "learning_rate": 5.998178410334944e-06, "loss": 0.8456, "step": 18247 }, { "epoch": 0.6430050124650933, "grad_norm": 1.6224247217178345, "learning_rate": 5.997132529689172e-06, "loss": 0.7892, "step": 18248 }, { "epoch": 0.6430402494780517, "grad_norm": 1.2549409866333008, "learning_rate": 5.996086701181257e-06, "loss": 1.1076, "step": 18249 }, { "epoch": 0.6430754864910102, "grad_norm": 1.1075817346572876, "learning_rate": 5.995040924824822e-06, "loss": 0.9315, "step": 18250 }, { "epoch": 0.6431107235039686, "grad_norm": 1.2812899351119995, "learning_rate": 5.993995200633491e-06, "loss": 0.6312, "step": 18251 }, { "epoch": 0.643145960516927, "grad_norm": 1.1317938566207886, "learning_rate": 5.9929495286208785e-06, "loss": 1.2265, "step": 18252 }, { "epoch": 0.6431811975298853, "grad_norm": 1.4322328567504883, "learning_rate": 5.991903908800612e-06, "loss": 0.7647, "step": 18253 }, { "epoch": 0.6432164345428438, "grad_norm": 3.425962448120117, "learning_rate": 5.9908583411863075e-06, "loss": 2.981, "step": 18254 }, { "epoch": 0.6432516715558022, "grad_norm": 3.819596767425537, "learning_rate": 5.989812825791584e-06, "loss": 4.1979, "step": 18255 }, { "epoch": 0.6432869085687606, "grad_norm": 0.9429121613502502, "learning_rate": 5.988767362630057e-06, "loss": 1.1097, "step": 18256 }, { "epoch": 0.6433221455817191, "grad_norm": 5.160093784332275, "learning_rate": 5.987721951715351e-06, "loss": 2.9783, "step": 18257 }, { "epoch": 0.6433573825946775, "grad_norm": 1.3583488464355469, "learning_rate": 5.986676593061072e-06, "loss": 0.9921, "step": 18258 }, { "epoch": 0.6433926196076358, "grad_norm": 4.690882205963135, "learning_rate": 5.985631286680844e-06, "loss": 3.0715, "step": 18259 }, { "epoch": 0.6434278566205943, "grad_norm": 5.705071926116943, "learning_rate": 5.984586032588282e-06, "loss": 3.1407, "step": 18260 }, { "epoch": 0.6434630936335527, "grad_norm": 1.3060321807861328, "learning_rate": 5.983540830796996e-06, "loss": 0.8081, "step": 18261 }, { "epoch": 0.6434983306465111, "grad_norm": 1.5936192274093628, "learning_rate": 5.982495681320599e-06, "loss": 0.9595, "step": 18262 }, { "epoch": 0.6435335676594695, "grad_norm": 1.2318763732910156, "learning_rate": 5.981450584172715e-06, "loss": 0.7984, "step": 18263 }, { "epoch": 0.6435688046724279, "grad_norm": 1.5385017395019531, "learning_rate": 5.980405539366943e-06, "loss": 0.8357, "step": 18264 }, { "epoch": 0.6436040416853863, "grad_norm": 2.9176270961761475, "learning_rate": 5.9793605469169035e-06, "loss": 2.835, "step": 18265 }, { "epoch": 0.6436392786983447, "grad_norm": 2.445080518722534, "learning_rate": 5.978315606836203e-06, "loss": 0.8565, "step": 18266 }, { "epoch": 0.6436745157113032, "grad_norm": 1.472622275352478, "learning_rate": 5.977270719138456e-06, "loss": 0.8074, "step": 18267 }, { "epoch": 0.6437097527242616, "grad_norm": 1.5784560441970825, "learning_rate": 5.97622588383727e-06, "loss": 0.8946, "step": 18268 }, { "epoch": 0.6437449897372199, "grad_norm": 0.9564048051834106, "learning_rate": 5.975181100946253e-06, "loss": 1.0548, "step": 18269 }, { "epoch": 0.6437802267501784, "grad_norm": 3.7281229496002197, "learning_rate": 5.974136370479018e-06, "loss": 2.8027, "step": 18270 }, { "epoch": 0.6438154637631368, "grad_norm": 1.3266005516052246, "learning_rate": 5.973091692449167e-06, "loss": 0.9432, "step": 18271 }, { "epoch": 0.6438507007760952, "grad_norm": 3.6129486560821533, "learning_rate": 5.9720470668703086e-06, "loss": 3.2551, "step": 18272 }, { "epoch": 0.6438859377890537, "grad_norm": 6.622525215148926, "learning_rate": 5.971002493756054e-06, "loss": 4.7306, "step": 18273 }, { "epoch": 0.643921174802012, "grad_norm": 3.091174364089966, "learning_rate": 5.969957973120003e-06, "loss": 2.6119, "step": 18274 }, { "epoch": 0.6439564118149704, "grad_norm": 3.5883288383483887, "learning_rate": 5.968913504975762e-06, "loss": 2.9499, "step": 18275 }, { "epoch": 0.6439916488279288, "grad_norm": 1.2214257717132568, "learning_rate": 5.96786908933694e-06, "loss": 0.9355, "step": 18276 }, { "epoch": 0.6440268858408873, "grad_norm": 3.73876690864563, "learning_rate": 5.9668247262171304e-06, "loss": 2.7268, "step": 18277 }, { "epoch": 0.6440621228538457, "grad_norm": 2.6881487369537354, "learning_rate": 5.965780415629945e-06, "loss": 0.9345, "step": 18278 }, { "epoch": 0.644097359866804, "grad_norm": 5.212965965270996, "learning_rate": 5.9647361575889865e-06, "loss": 4.4107, "step": 18279 }, { "epoch": 0.6441325968797625, "grad_norm": 3.5548079013824463, "learning_rate": 5.963691952107851e-06, "loss": 3.1389, "step": 18280 }, { "epoch": 0.6441678338927209, "grad_norm": 5.433385848999023, "learning_rate": 5.96264779920014e-06, "loss": 2.7773, "step": 18281 }, { "epoch": 0.6442030709056793, "grad_norm": 1.7981146574020386, "learning_rate": 5.961603698879459e-06, "loss": 0.8672, "step": 18282 }, { "epoch": 0.6442383079186378, "grad_norm": 2.6219379901885986, "learning_rate": 5.9605596511594036e-06, "loss": 0.6118, "step": 18283 }, { "epoch": 0.6442735449315962, "grad_norm": 2.5200040340423584, "learning_rate": 5.959515656053573e-06, "loss": 2.707, "step": 18284 }, { "epoch": 0.6443087819445545, "grad_norm": 1.4814835786819458, "learning_rate": 5.958471713575562e-06, "loss": 0.8146, "step": 18285 }, { "epoch": 0.6443440189575129, "grad_norm": 2.862476348876953, "learning_rate": 5.957427823738979e-06, "loss": 3.1493, "step": 18286 }, { "epoch": 0.6443792559704714, "grad_norm": 2.2526021003723145, "learning_rate": 5.956383986557408e-06, "loss": 1.0374, "step": 18287 }, { "epoch": 0.6444144929834298, "grad_norm": 3.7584729194641113, "learning_rate": 5.955340202044453e-06, "loss": 2.8091, "step": 18288 }, { "epoch": 0.6444497299963882, "grad_norm": 1.1560373306274414, "learning_rate": 5.954296470213707e-06, "loss": 0.9532, "step": 18289 }, { "epoch": 0.6444849670093467, "grad_norm": 1.297682285308838, "learning_rate": 5.953252791078765e-06, "loss": 0.8322, "step": 18290 }, { "epoch": 0.644520204022305, "grad_norm": 1.4130011796951294, "learning_rate": 5.9522091646532176e-06, "loss": 1.0328, "step": 18291 }, { "epoch": 0.6445554410352634, "grad_norm": 1.1077197790145874, "learning_rate": 5.951165590950666e-06, "loss": 0.9672, "step": 18292 }, { "epoch": 0.6445906780482219, "grad_norm": 7.677926540374756, "learning_rate": 5.9501220699846965e-06, "loss": 3.0892, "step": 18293 }, { "epoch": 0.6446259150611803, "grad_norm": 3.225750684738159, "learning_rate": 5.949078601768903e-06, "loss": 2.6729, "step": 18294 }, { "epoch": 0.6446611520741387, "grad_norm": 3.994957685470581, "learning_rate": 5.948035186316878e-06, "loss": 3.5431, "step": 18295 }, { "epoch": 0.644696389087097, "grad_norm": 3.601395606994629, "learning_rate": 5.946991823642207e-06, "loss": 3.3408, "step": 18296 }, { "epoch": 0.6447316261000555, "grad_norm": 1.1026570796966553, "learning_rate": 5.945948513758486e-06, "loss": 0.7334, "step": 18297 }, { "epoch": 0.6447668631130139, "grad_norm": 4.703022003173828, "learning_rate": 5.944905256679303e-06, "loss": 3.0897, "step": 18298 }, { "epoch": 0.6448021001259723, "grad_norm": 1.6895111799240112, "learning_rate": 5.943862052418245e-06, "loss": 0.9802, "step": 18299 }, { "epoch": 0.6448373371389308, "grad_norm": 5.223789215087891, "learning_rate": 5.942818900988898e-06, "loss": 4.8992, "step": 18300 }, { "epoch": 0.6448725741518891, "grad_norm": 5.479739665985107, "learning_rate": 5.941775802404857e-06, "loss": 4.9599, "step": 18301 }, { "epoch": 0.6449078111648475, "grad_norm": 4.0328264236450195, "learning_rate": 5.940732756679698e-06, "loss": 0.8342, "step": 18302 }, { "epoch": 0.644943048177806, "grad_norm": 2.8440310955047607, "learning_rate": 5.939689763827014e-06, "loss": 2.9074, "step": 18303 }, { "epoch": 0.6449782851907644, "grad_norm": 1.2825356721878052, "learning_rate": 5.938646823860386e-06, "loss": 1.0426, "step": 18304 }, { "epoch": 0.6450135222037228, "grad_norm": 1.2605445384979248, "learning_rate": 5.937603936793406e-06, "loss": 0.9642, "step": 18305 }, { "epoch": 0.6450487592166813, "grad_norm": 3.0874414443969727, "learning_rate": 5.936561102639645e-06, "loss": 3.0843, "step": 18306 }, { "epoch": 0.6450839962296396, "grad_norm": 1.0644925832748413, "learning_rate": 5.9355183214126965e-06, "loss": 1.0116, "step": 18307 }, { "epoch": 0.645119233242598, "grad_norm": 5.891138553619385, "learning_rate": 5.934475593126141e-06, "loss": 7.1324, "step": 18308 }, { "epoch": 0.6451544702555564, "grad_norm": 1.1326713562011719, "learning_rate": 5.933432917793558e-06, "loss": 0.9973, "step": 18309 }, { "epoch": 0.6451897072685149, "grad_norm": 3.2466177940368652, "learning_rate": 5.932390295428526e-06, "loss": 3.026, "step": 18310 }, { "epoch": 0.6452249442814733, "grad_norm": 4.74871301651001, "learning_rate": 5.9313477260446335e-06, "loss": 5.6902, "step": 18311 }, { "epoch": 0.6452601812944316, "grad_norm": 5.09525728225708, "learning_rate": 5.930305209655452e-06, "loss": 7.1389, "step": 18312 }, { "epoch": 0.6452954183073901, "grad_norm": 3.2204811573028564, "learning_rate": 5.929262746274564e-06, "loss": 2.5438, "step": 18313 }, { "epoch": 0.6453306553203485, "grad_norm": 3.23574161529541, "learning_rate": 5.928220335915551e-06, "loss": 3.4859, "step": 18314 }, { "epoch": 0.6453658923333069, "grad_norm": 1.3401634693145752, "learning_rate": 5.9271779785919805e-06, "loss": 0.7074, "step": 18315 }, { "epoch": 0.6454011293462654, "grad_norm": 1.0780616998672485, "learning_rate": 5.9261356743174395e-06, "loss": 0.9135, "step": 18316 }, { "epoch": 0.6454363663592237, "grad_norm": 1.0487574338912964, "learning_rate": 5.9250934231055e-06, "loss": 0.9909, "step": 18317 }, { "epoch": 0.6454716033721821, "grad_norm": 1.1932339668273926, "learning_rate": 5.9240512249697365e-06, "loss": 1.0479, "step": 18318 }, { "epoch": 0.6455068403851405, "grad_norm": 1.5970560312271118, "learning_rate": 5.923009079923724e-06, "loss": 0.976, "step": 18319 }, { "epoch": 0.645542077398099, "grad_norm": 5.684813022613525, "learning_rate": 5.921966987981038e-06, "loss": 5.2554, "step": 18320 }, { "epoch": 0.6455773144110574, "grad_norm": 2.36120343208313, "learning_rate": 5.920924949155255e-06, "loss": 2.0808, "step": 18321 }, { "epoch": 0.6456125514240157, "grad_norm": 1.2497113943099976, "learning_rate": 5.919882963459942e-06, "loss": 0.6506, "step": 18322 }, { "epoch": 0.6456477884369742, "grad_norm": 1.7315553426742554, "learning_rate": 5.918841030908671e-06, "loss": 0.7968, "step": 18323 }, { "epoch": 0.6456830254499326, "grad_norm": 1.6435084342956543, "learning_rate": 5.91779915151502e-06, "loss": 0.9811, "step": 18324 }, { "epoch": 0.645718262462891, "grad_norm": 3.8391237258911133, "learning_rate": 5.916757325292551e-06, "loss": 0.9405, "step": 18325 }, { "epoch": 0.6457534994758495, "grad_norm": 1.1513632535934448, "learning_rate": 5.91571555225484e-06, "loss": 1.0294, "step": 18326 }, { "epoch": 0.6457887364888079, "grad_norm": 6.196882247924805, "learning_rate": 5.914673832415454e-06, "loss": 5.2346, "step": 18327 }, { "epoch": 0.6458239735017662, "grad_norm": 2.0799856185913086, "learning_rate": 5.913632165787961e-06, "loss": 0.9484, "step": 18328 }, { "epoch": 0.6458592105147246, "grad_norm": 5.7584381103515625, "learning_rate": 5.912590552385927e-06, "loss": 5.1465, "step": 18329 }, { "epoch": 0.6458944475276831, "grad_norm": 0.9511262774467468, "learning_rate": 5.911548992222926e-06, "loss": 1.1006, "step": 18330 }, { "epoch": 0.6459296845406415, "grad_norm": 1.6406911611557007, "learning_rate": 5.910507485312519e-06, "loss": 0.8337, "step": 18331 }, { "epoch": 0.6459649215535999, "grad_norm": 3.963033437728882, "learning_rate": 5.909466031668272e-06, "loss": 2.9727, "step": 18332 }, { "epoch": 0.6460001585665583, "grad_norm": 2.3801534175872803, "learning_rate": 5.908424631303753e-06, "loss": 2.3438, "step": 18333 }, { "epoch": 0.6460353955795167, "grad_norm": 7.928863525390625, "learning_rate": 5.907383284232522e-06, "loss": 3.8176, "step": 18334 }, { "epoch": 0.6460706325924751, "grad_norm": 4.688148498535156, "learning_rate": 5.9063419904681455e-06, "loss": 3.4312, "step": 18335 }, { "epoch": 0.6461058696054336, "grad_norm": 3.876964807510376, "learning_rate": 5.9053007500241874e-06, "loss": 2.795, "step": 18336 }, { "epoch": 0.646141106618392, "grad_norm": 3.904573917388916, "learning_rate": 5.904259562914208e-06, "loss": 4.09, "step": 18337 }, { "epoch": 0.6461763436313503, "grad_norm": 4.213022708892822, "learning_rate": 5.903218429151767e-06, "loss": 3.0929, "step": 18338 }, { "epoch": 0.6462115806443088, "grad_norm": 1.0505436658859253, "learning_rate": 5.90217734875043e-06, "loss": 0.9202, "step": 18339 }, { "epoch": 0.6462468176572672, "grad_norm": 1.3638741970062256, "learning_rate": 5.9011363217237575e-06, "loss": 0.7826, "step": 18340 }, { "epoch": 0.6462820546702256, "grad_norm": 0.863287627696991, "learning_rate": 5.900095348085303e-06, "loss": 0.8965, "step": 18341 }, { "epoch": 0.646317291683184, "grad_norm": 4.855027198791504, "learning_rate": 5.899054427848627e-06, "loss": 2.7852, "step": 18342 }, { "epoch": 0.6463525286961425, "grad_norm": 1.4803142547607422, "learning_rate": 5.898013561027296e-06, "loss": 0.9954, "step": 18343 }, { "epoch": 0.6463877657091008, "grad_norm": 4.451463222503662, "learning_rate": 5.8969727476348545e-06, "loss": 5.0697, "step": 18344 }, { "epoch": 0.6464230027220592, "grad_norm": 0.9541531205177307, "learning_rate": 5.895931987684868e-06, "loss": 0.8965, "step": 18345 }, { "epoch": 0.6464582397350177, "grad_norm": 3.5971856117248535, "learning_rate": 5.894891281190893e-06, "loss": 2.9423, "step": 18346 }, { "epoch": 0.6464934767479761, "grad_norm": 1.5869932174682617, "learning_rate": 5.89385062816648e-06, "loss": 0.9707, "step": 18347 }, { "epoch": 0.6465287137609345, "grad_norm": 3.4717323780059814, "learning_rate": 5.892810028625182e-06, "loss": 2.4075, "step": 18348 }, { "epoch": 0.646563950773893, "grad_norm": 4.949849605560303, "learning_rate": 5.891769482580562e-06, "loss": 5.246, "step": 18349 }, { "epoch": 0.6465991877868513, "grad_norm": 1.4547423124313354, "learning_rate": 5.890728990046164e-06, "loss": 0.6698, "step": 18350 }, { "epoch": 0.6466344247998097, "grad_norm": 0.9721702933311462, "learning_rate": 5.889688551035546e-06, "loss": 0.8967, "step": 18351 }, { "epoch": 0.6466696618127681, "grad_norm": 3.08170747756958, "learning_rate": 5.8886481655622596e-06, "loss": 2.2609, "step": 18352 }, { "epoch": 0.6467048988257266, "grad_norm": 4.292920112609863, "learning_rate": 5.8876078336398505e-06, "loss": 3.5235, "step": 18353 }, { "epoch": 0.646740135838685, "grad_norm": 5.304681777954102, "learning_rate": 5.8865675552818755e-06, "loss": 3.2697, "step": 18354 }, { "epoch": 0.6467753728516433, "grad_norm": 2.401904344558716, "learning_rate": 5.885527330501884e-06, "loss": 2.0734, "step": 18355 }, { "epoch": 0.6468106098646018, "grad_norm": 6.108802318572998, "learning_rate": 5.8844871593134215e-06, "loss": 4.8864, "step": 18356 }, { "epoch": 0.6468458468775602, "grad_norm": 1.006044864654541, "learning_rate": 5.883447041730035e-06, "loss": 0.9482, "step": 18357 }, { "epoch": 0.6468810838905186, "grad_norm": 1.3269602060317993, "learning_rate": 5.882406977765278e-06, "loss": 0.7273, "step": 18358 }, { "epoch": 0.6469163209034771, "grad_norm": 0.9999664425849915, "learning_rate": 5.881366967432696e-06, "loss": 1.0301, "step": 18359 }, { "epoch": 0.6469515579164354, "grad_norm": 4.736318588256836, "learning_rate": 5.880327010745832e-06, "loss": 6.8643, "step": 18360 }, { "epoch": 0.6469867949293938, "grad_norm": 1.4112963676452637, "learning_rate": 5.879287107718232e-06, "loss": 0.8945, "step": 18361 }, { "epoch": 0.6470220319423522, "grad_norm": 0.9547730088233948, "learning_rate": 5.8782472583634474e-06, "loss": 1.0604, "step": 18362 }, { "epoch": 0.6470572689553107, "grad_norm": 1.649921178817749, "learning_rate": 5.877207462695011e-06, "loss": 0.797, "step": 18363 }, { "epoch": 0.6470925059682691, "grad_norm": 2.7091219425201416, "learning_rate": 5.876167720726476e-06, "loss": 2.48, "step": 18364 }, { "epoch": 0.6471277429812274, "grad_norm": 6.1654767990112305, "learning_rate": 5.875128032471383e-06, "loss": 4.8359, "step": 18365 }, { "epoch": 0.6471629799941859, "grad_norm": 1.9120709896087646, "learning_rate": 5.8740883979432715e-06, "loss": 0.8562, "step": 18366 }, { "epoch": 0.6471982170071443, "grad_norm": 2.993500232696533, "learning_rate": 5.873048817155681e-06, "loss": 2.3512, "step": 18367 }, { "epoch": 0.6472334540201027, "grad_norm": 3.4213716983795166, "learning_rate": 5.87200929012216e-06, "loss": 3.0628, "step": 18368 }, { "epoch": 0.6472686910330612, "grad_norm": 2.5782926082611084, "learning_rate": 5.870969816856243e-06, "loss": 0.8925, "step": 18369 }, { "epoch": 0.6473039280460195, "grad_norm": 1.8603068590164185, "learning_rate": 5.8699303973714675e-06, "loss": 0.8525, "step": 18370 }, { "epoch": 0.6473391650589779, "grad_norm": 1.391165852546692, "learning_rate": 5.868891031681379e-06, "loss": 0.8833, "step": 18371 }, { "epoch": 0.6473744020719364, "grad_norm": 4.141734600067139, "learning_rate": 5.867851719799504e-06, "loss": 2.6759, "step": 18372 }, { "epoch": 0.6474096390848948, "grad_norm": 2.0482404232025146, "learning_rate": 5.866812461739392e-06, "loss": 0.9674, "step": 18373 }, { "epoch": 0.6474448760978532, "grad_norm": 3.178746461868286, "learning_rate": 5.865773257514574e-06, "loss": 2.5315, "step": 18374 }, { "epoch": 0.6474801131108116, "grad_norm": 3.148591995239258, "learning_rate": 5.864734107138586e-06, "loss": 2.9524, "step": 18375 }, { "epoch": 0.64751535012377, "grad_norm": 1.1316155195236206, "learning_rate": 5.863695010624959e-06, "loss": 0.6863, "step": 18376 }, { "epoch": 0.6475505871367284, "grad_norm": 3.403735876083374, "learning_rate": 5.8626559679872345e-06, "loss": 3.0713, "step": 18377 }, { "epoch": 0.6475858241496868, "grad_norm": 4.28926420211792, "learning_rate": 5.861616979238944e-06, "loss": 3.1064, "step": 18378 }, { "epoch": 0.6476210611626453, "grad_norm": 1.3275654315948486, "learning_rate": 5.8605780443936186e-06, "loss": 0.6933, "step": 18379 }, { "epoch": 0.6476562981756037, "grad_norm": 4.608315467834473, "learning_rate": 5.859539163464789e-06, "loss": 3.1407, "step": 18380 }, { "epoch": 0.647691535188562, "grad_norm": 3.7535641193389893, "learning_rate": 5.8585003364659954e-06, "loss": 2.7779, "step": 18381 }, { "epoch": 0.6477267722015205, "grad_norm": 1.1101926565170288, "learning_rate": 5.857461563410756e-06, "loss": 0.908, "step": 18382 }, { "epoch": 0.6477620092144789, "grad_norm": 5.627599716186523, "learning_rate": 5.85642284431261e-06, "loss": 2.7134, "step": 18383 }, { "epoch": 0.6477972462274373, "grad_norm": 4.945760250091553, "learning_rate": 5.855384179185086e-06, "loss": 5.2009, "step": 18384 }, { "epoch": 0.6478324832403957, "grad_norm": 3.9950735569000244, "learning_rate": 5.854345568041708e-06, "loss": 2.9505, "step": 18385 }, { "epoch": 0.6478677202533542, "grad_norm": 3.878751277923584, "learning_rate": 5.853307010896007e-06, "loss": 2.8971, "step": 18386 }, { "epoch": 0.6479029572663125, "grad_norm": 3.5545108318328857, "learning_rate": 5.852268507761514e-06, "loss": 3.0709, "step": 18387 }, { "epoch": 0.6479381942792709, "grad_norm": 5.334402561187744, "learning_rate": 5.851230058651749e-06, "loss": 3.0272, "step": 18388 }, { "epoch": 0.6479734312922294, "grad_norm": 1.3682230710983276, "learning_rate": 5.85019166358024e-06, "loss": 1.0201, "step": 18389 }, { "epoch": 0.6480086683051878, "grad_norm": 2.6185407638549805, "learning_rate": 5.849153322560517e-06, "loss": 3.0815, "step": 18390 }, { "epoch": 0.6480439053181462, "grad_norm": 6.20954704284668, "learning_rate": 5.8481150356060965e-06, "loss": 5.4162, "step": 18391 }, { "epoch": 0.6480791423311046, "grad_norm": 2.848301887512207, "learning_rate": 5.847076802730509e-06, "loss": 2.0942, "step": 18392 }, { "epoch": 0.648114379344063, "grad_norm": 1.2430108785629272, "learning_rate": 5.846038623947275e-06, "loss": 0.7724, "step": 18393 }, { "epoch": 0.6481496163570214, "grad_norm": 4.646132469177246, "learning_rate": 5.845000499269917e-06, "loss": 4.2319, "step": 18394 }, { "epoch": 0.6481848533699798, "grad_norm": 1.0514639616012573, "learning_rate": 5.843962428711956e-06, "loss": 0.6697, "step": 18395 }, { "epoch": 0.6482200903829383, "grad_norm": 5.225450038909912, "learning_rate": 5.8429244122869125e-06, "loss": 6.9816, "step": 18396 }, { "epoch": 0.6482553273958966, "grad_norm": 2.737849473953247, "learning_rate": 5.841886450008312e-06, "loss": 0.9178, "step": 18397 }, { "epoch": 0.648290564408855, "grad_norm": 6.130531311035156, "learning_rate": 5.840848541889667e-06, "loss": 5.1528, "step": 18398 }, { "epoch": 0.6483258014218135, "grad_norm": 4.410112380981445, "learning_rate": 5.839810687944498e-06, "loss": 4.3177, "step": 18399 }, { "epoch": 0.6483610384347719, "grad_norm": 1.244206190109253, "learning_rate": 5.838772888186333e-06, "loss": 0.9221, "step": 18400 }, { "epoch": 0.6483962754477303, "grad_norm": 2.6059842109680176, "learning_rate": 5.837735142628672e-06, "loss": 2.8528, "step": 18401 }, { "epoch": 0.6484315124606888, "grad_norm": 6.819274425506592, "learning_rate": 5.836697451285045e-06, "loss": 4.7831, "step": 18402 }, { "epoch": 0.6484667494736471, "grad_norm": 2.1215851306915283, "learning_rate": 5.835659814168966e-06, "loss": 0.7227, "step": 18403 }, { "epoch": 0.6485019864866055, "grad_norm": 1.0199813842773438, "learning_rate": 5.834622231293942e-06, "loss": 0.9182, "step": 18404 }, { "epoch": 0.648537223499564, "grad_norm": 2.4305050373077393, "learning_rate": 5.833584702673498e-06, "loss": 2.0652, "step": 18405 }, { "epoch": 0.6485724605125224, "grad_norm": 2.2056987285614014, "learning_rate": 5.832547228321147e-06, "loss": 0.6978, "step": 18406 }, { "epoch": 0.6486076975254808, "grad_norm": 0.8639589548110962, "learning_rate": 5.831509808250395e-06, "loss": 1.0115, "step": 18407 }, { "epoch": 0.6486429345384391, "grad_norm": 1.0025231838226318, "learning_rate": 5.830472442474756e-06, "loss": 1.2826, "step": 18408 }, { "epoch": 0.6486781715513976, "grad_norm": 0.9244048595428467, "learning_rate": 5.829435131007752e-06, "loss": 0.8728, "step": 18409 }, { "epoch": 0.648713408564356, "grad_norm": 4.841215133666992, "learning_rate": 5.828397873862884e-06, "loss": 4.8683, "step": 18410 }, { "epoch": 0.6487486455773144, "grad_norm": 0.9848536849021912, "learning_rate": 5.8273606710536626e-06, "loss": 0.7727, "step": 18411 }, { "epoch": 0.6487838825902729, "grad_norm": 5.830624103546143, "learning_rate": 5.826323522593597e-06, "loss": 3.2613, "step": 18412 }, { "epoch": 0.6488191196032312, "grad_norm": 9.178187370300293, "learning_rate": 5.825286428496207e-06, "loss": 5.0505, "step": 18413 }, { "epoch": 0.6488543566161896, "grad_norm": 0.9218095541000366, "learning_rate": 5.8242493887749895e-06, "loss": 0.7938, "step": 18414 }, { "epoch": 0.6488895936291481, "grad_norm": 5.006187438964844, "learning_rate": 5.823212403443455e-06, "loss": 2.4642, "step": 18415 }, { "epoch": 0.6489248306421065, "grad_norm": 2.9974193572998047, "learning_rate": 5.822175472515113e-06, "loss": 2.9396, "step": 18416 }, { "epoch": 0.6489600676550649, "grad_norm": 1.1833199262619019, "learning_rate": 5.821138596003465e-06, "loss": 0.9623, "step": 18417 }, { "epoch": 0.6489953046680232, "grad_norm": 0.9980507493019104, "learning_rate": 5.82010177392202e-06, "loss": 1.1211, "step": 18418 }, { "epoch": 0.6490305416809817, "grad_norm": 3.2934157848358154, "learning_rate": 5.8190650062842814e-06, "loss": 3.2802, "step": 18419 }, { "epoch": 0.6490657786939401, "grad_norm": 6.405581951141357, "learning_rate": 5.818028293103755e-06, "loss": 5.5982, "step": 18420 }, { "epoch": 0.6491010157068985, "grad_norm": 0.9819667935371399, "learning_rate": 5.816991634393942e-06, "loss": 0.8307, "step": 18421 }, { "epoch": 0.649136252719857, "grad_norm": 1.4636099338531494, "learning_rate": 5.81595503016835e-06, "loss": 0.8327, "step": 18422 }, { "epoch": 0.6491714897328154, "grad_norm": 1.096745252609253, "learning_rate": 5.814918480440468e-06, "loss": 1.2064, "step": 18423 }, { "epoch": 0.6492067267457737, "grad_norm": 3.4880917072296143, "learning_rate": 5.81388198522381e-06, "loss": 4.5772, "step": 18424 }, { "epoch": 0.6492419637587322, "grad_norm": 2.7561116218566895, "learning_rate": 5.812845544531878e-06, "loss": 0.9237, "step": 18425 }, { "epoch": 0.6492772007716906, "grad_norm": 4.169785976409912, "learning_rate": 5.81180915837816e-06, "loss": 2.5676, "step": 18426 }, { "epoch": 0.649312437784649, "grad_norm": 1.3921988010406494, "learning_rate": 5.810772826776157e-06, "loss": 0.8937, "step": 18427 }, { "epoch": 0.6493476747976074, "grad_norm": 2.6430752277374268, "learning_rate": 5.809736549739382e-06, "loss": 2.1575, "step": 18428 }, { "epoch": 0.6493829118105658, "grad_norm": 1.4252445697784424, "learning_rate": 5.808700327281316e-06, "loss": 0.9036, "step": 18429 }, { "epoch": 0.6494181488235242, "grad_norm": 8.056808471679688, "learning_rate": 5.8076641594154625e-06, "loss": 5.1677, "step": 18430 }, { "epoch": 0.6494533858364826, "grad_norm": 5.719935417175293, "learning_rate": 5.8066280461553136e-06, "loss": 4.8666, "step": 18431 }, { "epoch": 0.6494886228494411, "grad_norm": 1.045660138130188, "learning_rate": 5.805591987514378e-06, "loss": 0.9156, "step": 18432 }, { "epoch": 0.6495238598623995, "grad_norm": 1.3552578687667847, "learning_rate": 5.804555983506135e-06, "loss": 0.8616, "step": 18433 }, { "epoch": 0.6495590968753578, "grad_norm": 0.8374524712562561, "learning_rate": 5.8035200341440854e-06, "loss": 0.8028, "step": 18434 }, { "epoch": 0.6495943338883163, "grad_norm": 3.4658477306365967, "learning_rate": 5.802484139441722e-06, "loss": 2.9133, "step": 18435 }, { "epoch": 0.6496295709012747, "grad_norm": 3.290837049484253, "learning_rate": 5.801448299412537e-06, "loss": 2.882, "step": 18436 }, { "epoch": 0.6496648079142331, "grad_norm": 1.316414713859558, "learning_rate": 5.800412514070024e-06, "loss": 0.8583, "step": 18437 }, { "epoch": 0.6497000449271915, "grad_norm": 1.4293607473373413, "learning_rate": 5.799376783427673e-06, "loss": 0.6896, "step": 18438 }, { "epoch": 0.64973528194015, "grad_norm": 0.9388307929039001, "learning_rate": 5.798341107498973e-06, "loss": 0.8531, "step": 18439 }, { "epoch": 0.6497705189531083, "grad_norm": 0.8574258089065552, "learning_rate": 5.7973054862974155e-06, "loss": 0.7654, "step": 18440 }, { "epoch": 0.6498057559660667, "grad_norm": 5.359281063079834, "learning_rate": 5.796269919836494e-06, "loss": 2.0852, "step": 18441 }, { "epoch": 0.6498409929790252, "grad_norm": 2.591109037399292, "learning_rate": 5.7952344081296844e-06, "loss": 0.9197, "step": 18442 }, { "epoch": 0.6498762299919836, "grad_norm": 1.092260479927063, "learning_rate": 5.794198951190484e-06, "loss": 0.6862, "step": 18443 }, { "epoch": 0.649911467004942, "grad_norm": 4.782685279846191, "learning_rate": 5.793163549032384e-06, "loss": 2.9702, "step": 18444 }, { "epoch": 0.6499467040179004, "grad_norm": 9.403504371643066, "learning_rate": 5.7921282016688605e-06, "loss": 3.2924, "step": 18445 }, { "epoch": 0.6499819410308588, "grad_norm": 5.367691516876221, "learning_rate": 5.791092909113399e-06, "loss": 2.9545, "step": 18446 }, { "epoch": 0.6500171780438172, "grad_norm": 1.0894807577133179, "learning_rate": 5.790057671379496e-06, "loss": 1.0032, "step": 18447 }, { "epoch": 0.6500524150567757, "grad_norm": 1.5418691635131836, "learning_rate": 5.789022488480625e-06, "loss": 0.8781, "step": 18448 }, { "epoch": 0.6500876520697341, "grad_norm": 1.6115086078643799, "learning_rate": 5.787987360430273e-06, "loss": 0.9095, "step": 18449 }, { "epoch": 0.6501228890826924, "grad_norm": 1.7759002447128296, "learning_rate": 5.786952287241917e-06, "loss": 0.8719, "step": 18450 }, { "epoch": 0.6501581260956508, "grad_norm": 1.5292896032333374, "learning_rate": 5.785917268929052e-06, "loss": 0.946, "step": 18451 }, { "epoch": 0.6501933631086093, "grad_norm": 4.026628494262695, "learning_rate": 5.784882305505148e-06, "loss": 2.8926, "step": 18452 }, { "epoch": 0.6502286001215677, "grad_norm": 0.9918579459190369, "learning_rate": 5.783847396983688e-06, "loss": 0.5921, "step": 18453 }, { "epoch": 0.6502638371345261, "grad_norm": 1.2940213680267334, "learning_rate": 5.782812543378154e-06, "loss": 0.9779, "step": 18454 }, { "epoch": 0.6502990741474846, "grad_norm": 1.5355411767959595, "learning_rate": 5.781777744702024e-06, "loss": 0.7481, "step": 18455 }, { "epoch": 0.6503343111604429, "grad_norm": 1.083134651184082, "learning_rate": 5.780743000968775e-06, "loss": 0.963, "step": 18456 }, { "epoch": 0.6503695481734013, "grad_norm": 0.8689036965370178, "learning_rate": 5.779708312191885e-06, "loss": 0.8156, "step": 18457 }, { "epoch": 0.6504047851863598, "grad_norm": 0.8514080047607422, "learning_rate": 5.778673678384834e-06, "loss": 0.8069, "step": 18458 }, { "epoch": 0.6504400221993182, "grad_norm": 3.5052168369293213, "learning_rate": 5.777639099561094e-06, "loss": 2.773, "step": 18459 }, { "epoch": 0.6504752592122766, "grad_norm": 4.8473663330078125, "learning_rate": 5.776604575734148e-06, "loss": 5.4719, "step": 18460 }, { "epoch": 0.6505104962252349, "grad_norm": 1.1154038906097412, "learning_rate": 5.775570106917458e-06, "loss": 1.1451, "step": 18461 }, { "epoch": 0.6505457332381934, "grad_norm": 3.4729530811309814, "learning_rate": 5.774535693124508e-06, "loss": 2.2455, "step": 18462 }, { "epoch": 0.6505809702511518, "grad_norm": 1.175458550453186, "learning_rate": 5.773501334368774e-06, "loss": 0.9678, "step": 18463 }, { "epoch": 0.6506162072641102, "grad_norm": 2.8452415466308594, "learning_rate": 5.7724670306637175e-06, "loss": 2.9901, "step": 18464 }, { "epoch": 0.6506514442770687, "grad_norm": 14.733871459960938, "learning_rate": 5.771432782022814e-06, "loss": 5.1476, "step": 18465 }, { "epoch": 0.650686681290027, "grad_norm": 4.9677348136901855, "learning_rate": 5.770398588459542e-06, "loss": 4.9983, "step": 18466 }, { "epoch": 0.6507219183029854, "grad_norm": 2.5159289836883545, "learning_rate": 5.769364449987369e-06, "loss": 0.9085, "step": 18467 }, { "epoch": 0.6507571553159439, "grad_norm": 5.808477878570557, "learning_rate": 5.768330366619761e-06, "loss": 4.859, "step": 18468 }, { "epoch": 0.6507923923289023, "grad_norm": 2.474113702774048, "learning_rate": 5.767296338370183e-06, "loss": 2.8435, "step": 18469 }, { "epoch": 0.6508276293418607, "grad_norm": 1.3438351154327393, "learning_rate": 5.7662623652521184e-06, "loss": 0.8693, "step": 18470 }, { "epoch": 0.650862866354819, "grad_norm": 3.494875192642212, "learning_rate": 5.765228447279022e-06, "loss": 2.6627, "step": 18471 }, { "epoch": 0.6508981033677775, "grad_norm": 4.077419281005859, "learning_rate": 5.764194584464364e-06, "loss": 3.352, "step": 18472 }, { "epoch": 0.6509333403807359, "grad_norm": 1.5636940002441406, "learning_rate": 5.763160776821611e-06, "loss": 0.9331, "step": 18473 }, { "epoch": 0.6509685773936943, "grad_norm": 0.8174044489860535, "learning_rate": 5.762127024364229e-06, "loss": 0.7629, "step": 18474 }, { "epoch": 0.6510038144066528, "grad_norm": 2.2518558502197266, "learning_rate": 5.7610933271056824e-06, "loss": 0.9433, "step": 18475 }, { "epoch": 0.6510390514196112, "grad_norm": 6.139550685882568, "learning_rate": 5.760059685059435e-06, "loss": 4.9805, "step": 18476 }, { "epoch": 0.6510742884325695, "grad_norm": 1.997946858406067, "learning_rate": 5.75902609823895e-06, "loss": 0.9234, "step": 18477 }, { "epoch": 0.651109525445528, "grad_norm": 5.653560161590576, "learning_rate": 5.757992566657691e-06, "loss": 3.4687, "step": 18478 }, { "epoch": 0.6511447624584864, "grad_norm": 1.1117953062057495, "learning_rate": 5.756959090329123e-06, "loss": 0.9832, "step": 18479 }, { "epoch": 0.6511799994714448, "grad_norm": 5.72857141494751, "learning_rate": 5.755925669266695e-06, "loss": 3.4798, "step": 18480 }, { "epoch": 0.6512152364844033, "grad_norm": 1.232147455215454, "learning_rate": 5.754892303483879e-06, "loss": 0.9491, "step": 18481 }, { "epoch": 0.6512504734973616, "grad_norm": 8.260144233703613, "learning_rate": 5.753858992994137e-06, "loss": 5.4262, "step": 18482 }, { "epoch": 0.65128571051032, "grad_norm": 1.1915727853775024, "learning_rate": 5.752825737810918e-06, "loss": 1.0104, "step": 18483 }, { "epoch": 0.6513209475232784, "grad_norm": 1.3434075117111206, "learning_rate": 5.75179253794768e-06, "loss": 0.7984, "step": 18484 }, { "epoch": 0.6513561845362369, "grad_norm": 1.1489733457565308, "learning_rate": 5.750759393417889e-06, "loss": 0.902, "step": 18485 }, { "epoch": 0.6513914215491953, "grad_norm": 2.247925043106079, "learning_rate": 5.749726304235004e-06, "loss": 0.8373, "step": 18486 }, { "epoch": 0.6514266585621536, "grad_norm": 5.367922782897949, "learning_rate": 5.748693270412471e-06, "loss": 3.1722, "step": 18487 }, { "epoch": 0.6514618955751121, "grad_norm": 6.191176414489746, "learning_rate": 5.747660291963744e-06, "loss": 6.0366, "step": 18488 }, { "epoch": 0.6514971325880705, "grad_norm": 1.4021506309509277, "learning_rate": 5.746627368902294e-06, "loss": 0.9428, "step": 18489 }, { "epoch": 0.6515323696010289, "grad_norm": 1.6420373916625977, "learning_rate": 5.745594501241557e-06, "loss": 0.8567, "step": 18490 }, { "epoch": 0.6515676066139874, "grad_norm": 3.6794304847717285, "learning_rate": 5.744561688994996e-06, "loss": 4.7947, "step": 18491 }, { "epoch": 0.6516028436269458, "grad_norm": 1.9316730499267578, "learning_rate": 5.743528932176059e-06, "loss": 0.9841, "step": 18492 }, { "epoch": 0.6516380806399041, "grad_norm": 3.172626495361328, "learning_rate": 5.742496230798202e-06, "loss": 3.0944, "step": 18493 }, { "epoch": 0.6516733176528625, "grad_norm": 3.979144334793091, "learning_rate": 5.741463584874873e-06, "loss": 3.0475, "step": 18494 }, { "epoch": 0.651708554665821, "grad_norm": 4.784096717834473, "learning_rate": 5.740430994419523e-06, "loss": 4.6612, "step": 18495 }, { "epoch": 0.6517437916787794, "grad_norm": 3.9802520275115967, "learning_rate": 5.739398459445602e-06, "loss": 3.1094, "step": 18496 }, { "epoch": 0.6517790286917378, "grad_norm": 1.1916786432266235, "learning_rate": 5.738365979966559e-06, "loss": 0.6998, "step": 18497 }, { "epoch": 0.6518142657046962, "grad_norm": 4.05224609375, "learning_rate": 5.737333555995846e-06, "loss": 5.2565, "step": 18498 }, { "epoch": 0.6518495027176546, "grad_norm": 4.504818916320801, "learning_rate": 5.736301187546898e-06, "loss": 4.6301, "step": 18499 }, { "epoch": 0.651884739730613, "grad_norm": 1.7977803945541382, "learning_rate": 5.735268874633175e-06, "loss": 0.7369, "step": 18500 }, { "epoch": 0.6519199767435715, "grad_norm": 6.453733444213867, "learning_rate": 5.734236617268121e-06, "loss": 2.9275, "step": 18501 }, { "epoch": 0.6519552137565299, "grad_norm": 1.1922789812088013, "learning_rate": 5.733204415465175e-06, "loss": 0.7946, "step": 18502 }, { "epoch": 0.6519904507694883, "grad_norm": 4.691100597381592, "learning_rate": 5.732172269237782e-06, "loss": 2.6753, "step": 18503 }, { "epoch": 0.6520256877824466, "grad_norm": 1.2902413606643677, "learning_rate": 5.731140178599392e-06, "loss": 0.8906, "step": 18504 }, { "epoch": 0.6520609247954051, "grad_norm": 2.4277842044830322, "learning_rate": 5.730108143563449e-06, "loss": 2.319, "step": 18505 }, { "epoch": 0.6520961618083635, "grad_norm": 0.9444353580474854, "learning_rate": 5.729076164143386e-06, "loss": 0.9607, "step": 18506 }, { "epoch": 0.6521313988213219, "grad_norm": 3.321039915084839, "learning_rate": 5.728044240352648e-06, "loss": 3.2368, "step": 18507 }, { "epoch": 0.6521666358342804, "grad_norm": 1.1117140054702759, "learning_rate": 5.727012372204685e-06, "loss": 1.0248, "step": 18508 }, { "epoch": 0.6522018728472387, "grad_norm": 6.238398551940918, "learning_rate": 5.725980559712926e-06, "loss": 5.328, "step": 18509 }, { "epoch": 0.6522371098601971, "grad_norm": 5.177398204803467, "learning_rate": 5.724948802890815e-06, "loss": 4.7001, "step": 18510 }, { "epoch": 0.6522723468731556, "grad_norm": 2.6466259956359863, "learning_rate": 5.7239171017517905e-06, "loss": 2.6815, "step": 18511 }, { "epoch": 0.652307583886114, "grad_norm": 4.555815696716309, "learning_rate": 5.7228854563092895e-06, "loss": 3.4174, "step": 18512 }, { "epoch": 0.6523428208990724, "grad_norm": 4.39714241027832, "learning_rate": 5.721853866576751e-06, "loss": 2.6872, "step": 18513 }, { "epoch": 0.6523780579120309, "grad_norm": 1.1492528915405273, "learning_rate": 5.720822332567611e-06, "loss": 1.1181, "step": 18514 }, { "epoch": 0.6524132949249892, "grad_norm": 0.9088982939720154, "learning_rate": 5.719790854295304e-06, "loss": 0.8671, "step": 18515 }, { "epoch": 0.6524485319379476, "grad_norm": 1.7370582818984985, "learning_rate": 5.718759431773268e-06, "loss": 0.7731, "step": 18516 }, { "epoch": 0.652483768950906, "grad_norm": 1.4328149557113647, "learning_rate": 5.717728065014939e-06, "loss": 0.9757, "step": 18517 }, { "epoch": 0.6525190059638645, "grad_norm": 5.144509315490723, "learning_rate": 5.7166967540337396e-06, "loss": 3.2473, "step": 18518 }, { "epoch": 0.6525542429768229, "grad_norm": 5.319338321685791, "learning_rate": 5.715665498843115e-06, "loss": 2.6117, "step": 18519 }, { "epoch": 0.6525894799897812, "grad_norm": 1.1441776752471924, "learning_rate": 5.714634299456492e-06, "loss": 1.2563, "step": 18520 }, { "epoch": 0.6526247170027397, "grad_norm": 2.706751823425293, "learning_rate": 5.713603155887307e-06, "loss": 3.4932, "step": 18521 }, { "epoch": 0.6526599540156981, "grad_norm": 5.194153785705566, "learning_rate": 5.7125720681489804e-06, "loss": 5.2262, "step": 18522 }, { "epoch": 0.6526951910286565, "grad_norm": 2.5541269779205322, "learning_rate": 5.711541036254951e-06, "loss": 2.9798, "step": 18523 }, { "epoch": 0.652730428041615, "grad_norm": 2.355475664138794, "learning_rate": 5.71051006021865e-06, "loss": 0.9196, "step": 18524 }, { "epoch": 0.6527656650545733, "grad_norm": 5.487065315246582, "learning_rate": 5.7094791400535e-06, "loss": 6.4685, "step": 18525 }, { "epoch": 0.6528009020675317, "grad_norm": 1.7305326461791992, "learning_rate": 5.7084482757729245e-06, "loss": 0.865, "step": 18526 }, { "epoch": 0.6528361390804901, "grad_norm": 3.0167043209075928, "learning_rate": 5.707417467390366e-06, "loss": 2.5984, "step": 18527 }, { "epoch": 0.6528713760934486, "grad_norm": 1.2000664472579956, "learning_rate": 5.7063867149192366e-06, "loss": 1.2518, "step": 18528 }, { "epoch": 0.652906613106407, "grad_norm": 1.2448688745498657, "learning_rate": 5.705356018372968e-06, "loss": 1.084, "step": 18529 }, { "epoch": 0.6529418501193653, "grad_norm": 4.417893886566162, "learning_rate": 5.704325377764986e-06, "loss": 3.3297, "step": 18530 }, { "epoch": 0.6529770871323238, "grad_norm": 1.1452409029006958, "learning_rate": 5.70329479310871e-06, "loss": 0.9503, "step": 18531 }, { "epoch": 0.6530123241452822, "grad_norm": 1.9232025146484375, "learning_rate": 5.702264264417567e-06, "loss": 0.8055, "step": 18532 }, { "epoch": 0.6530475611582406, "grad_norm": 1.4498670101165771, "learning_rate": 5.701233791704981e-06, "loss": 0.8008, "step": 18533 }, { "epoch": 0.6530827981711991, "grad_norm": 4.613658905029297, "learning_rate": 5.700203374984372e-06, "loss": 3.0567, "step": 18534 }, { "epoch": 0.6531180351841575, "grad_norm": 6.194761276245117, "learning_rate": 5.699173014269162e-06, "loss": 5.1371, "step": 18535 }, { "epoch": 0.6531532721971158, "grad_norm": 4.0321173667907715, "learning_rate": 5.698142709572775e-06, "loss": 1.0095, "step": 18536 }, { "epoch": 0.6531885092100742, "grad_norm": 5.2955169677734375, "learning_rate": 5.697112460908619e-06, "loss": 5.3666, "step": 18537 }, { "epoch": 0.6532237462230327, "grad_norm": 1.0546842813491821, "learning_rate": 5.696082268290127e-06, "loss": 0.7729, "step": 18538 }, { "epoch": 0.6532589832359911, "grad_norm": 3.645207405090332, "learning_rate": 5.695052131730709e-06, "loss": 2.6093, "step": 18539 }, { "epoch": 0.6532942202489495, "grad_norm": 3.0693202018737793, "learning_rate": 5.69402205124379e-06, "loss": 2.997, "step": 18540 }, { "epoch": 0.6533294572619079, "grad_norm": 2.414020538330078, "learning_rate": 5.692992026842774e-06, "loss": 0.798, "step": 18541 }, { "epoch": 0.6533646942748663, "grad_norm": 5.351477146148682, "learning_rate": 5.6919620585410915e-06, "loss": 2.9155, "step": 18542 }, { "epoch": 0.6533999312878247, "grad_norm": 1.3949249982833862, "learning_rate": 5.690932146352155e-06, "loss": 0.9126, "step": 18543 }, { "epoch": 0.6534351683007832, "grad_norm": 3.7610435485839844, "learning_rate": 5.689902290289371e-06, "loss": 3.3418, "step": 18544 }, { "epoch": 0.6534704053137416, "grad_norm": 1.361672043800354, "learning_rate": 5.688872490366155e-06, "loss": 0.8577, "step": 18545 }, { "epoch": 0.6535056423266999, "grad_norm": 6.437358856201172, "learning_rate": 5.687842746595933e-06, "loss": 3.2165, "step": 18546 }, { "epoch": 0.6535408793396584, "grad_norm": 4.3703155517578125, "learning_rate": 5.686813058992104e-06, "loss": 5.6444, "step": 18547 }, { "epoch": 0.6535761163526168, "grad_norm": 3.227538585662842, "learning_rate": 5.685783427568085e-06, "loss": 2.7723, "step": 18548 }, { "epoch": 0.6536113533655752, "grad_norm": 3.8744804859161377, "learning_rate": 5.684753852337286e-06, "loss": 0.7639, "step": 18549 }, { "epoch": 0.6536465903785336, "grad_norm": 3.3901989459991455, "learning_rate": 5.683724333313116e-06, "loss": 2.8225, "step": 18550 }, { "epoch": 0.653681827391492, "grad_norm": 4.107635021209717, "learning_rate": 5.682694870508989e-06, "loss": 5.1568, "step": 18551 }, { "epoch": 0.6537170644044504, "grad_norm": 1.4672664403915405, "learning_rate": 5.681665463938309e-06, "loss": 0.8202, "step": 18552 }, { "epoch": 0.6537523014174088, "grad_norm": 2.89469838142395, "learning_rate": 5.680636113614488e-06, "loss": 3.2152, "step": 18553 }, { "epoch": 0.6537875384303673, "grad_norm": 1.4719393253326416, "learning_rate": 5.6796068195509315e-06, "loss": 1.1109, "step": 18554 }, { "epoch": 0.6538227754433257, "grad_norm": 6.521927833557129, "learning_rate": 5.67857758176105e-06, "loss": 4.6561, "step": 18555 }, { "epoch": 0.653858012456284, "grad_norm": 3.073575973510742, "learning_rate": 5.677548400258238e-06, "loss": 2.9005, "step": 18556 }, { "epoch": 0.6538932494692425, "grad_norm": 1.4097280502319336, "learning_rate": 5.676519275055914e-06, "loss": 1.1204, "step": 18557 }, { "epoch": 0.6539284864822009, "grad_norm": 5.044844150543213, "learning_rate": 5.675490206167476e-06, "loss": 4.6153, "step": 18558 }, { "epoch": 0.6539637234951593, "grad_norm": 0.9892658591270447, "learning_rate": 5.674461193606332e-06, "loss": 0.7538, "step": 18559 }, { "epoch": 0.6539989605081177, "grad_norm": 3.0774712562561035, "learning_rate": 5.673432237385875e-06, "loss": 3.2479, "step": 18560 }, { "epoch": 0.6540341975210762, "grad_norm": 6.960439205169678, "learning_rate": 5.672403337519519e-06, "loss": 7.4754, "step": 18561 }, { "epoch": 0.6540694345340345, "grad_norm": 1.5050963163375854, "learning_rate": 5.671374494020663e-06, "loss": 0.7774, "step": 18562 }, { "epoch": 0.6541046715469929, "grad_norm": 6.32145357131958, "learning_rate": 5.670345706902703e-06, "loss": 7.1193, "step": 18563 }, { "epoch": 0.6541399085599514, "grad_norm": 2.1128883361816406, "learning_rate": 5.669316976179038e-06, "loss": 1.126, "step": 18564 }, { "epoch": 0.6541751455729098, "grad_norm": 4.84998893737793, "learning_rate": 5.668288301863078e-06, "loss": 3.9164, "step": 18565 }, { "epoch": 0.6542103825858682, "grad_norm": 1.0438978672027588, "learning_rate": 5.667259683968211e-06, "loss": 0.4948, "step": 18566 }, { "epoch": 0.6542456195988267, "grad_norm": 1.2659860849380493, "learning_rate": 5.666231122507837e-06, "loss": 0.7316, "step": 18567 }, { "epoch": 0.654280856611785, "grad_norm": 2.3816044330596924, "learning_rate": 5.665202617495357e-06, "loss": 0.7104, "step": 18568 }, { "epoch": 0.6543160936247434, "grad_norm": 6.264956951141357, "learning_rate": 5.664174168944164e-06, "loss": 5.5205, "step": 18569 }, { "epoch": 0.6543513306377018, "grad_norm": 6.3385233879089355, "learning_rate": 5.663145776867656e-06, "loss": 4.7889, "step": 18570 }, { "epoch": 0.6543865676506603, "grad_norm": 0.9382907152175903, "learning_rate": 5.662117441279225e-06, "loss": 1.019, "step": 18571 }, { "epoch": 0.6544218046636187, "grad_norm": 27.601037979125977, "learning_rate": 5.661089162192267e-06, "loss": 3.3858, "step": 18572 }, { "epoch": 0.654457041676577, "grad_norm": 2.6790390014648438, "learning_rate": 5.660060939620176e-06, "loss": 2.439, "step": 18573 }, { "epoch": 0.6544922786895355, "grad_norm": 2.9320335388183594, "learning_rate": 5.659032773576347e-06, "loss": 3.386, "step": 18574 }, { "epoch": 0.6545275157024939, "grad_norm": 1.0797730684280396, "learning_rate": 5.658004664074163e-06, "loss": 1.1186, "step": 18575 }, { "epoch": 0.6545627527154523, "grad_norm": 1.347595453262329, "learning_rate": 5.656976611127024e-06, "loss": 0.7601, "step": 18576 }, { "epoch": 0.6545979897284108, "grad_norm": 1.4892112016677856, "learning_rate": 5.655948614748319e-06, "loss": 0.8494, "step": 18577 }, { "epoch": 0.6546332267413691, "grad_norm": 6.323366641998291, "learning_rate": 5.654920674951438e-06, "loss": 2.815, "step": 18578 }, { "epoch": 0.6546684637543275, "grad_norm": 3.0480942726135254, "learning_rate": 5.653892791749763e-06, "loss": 2.5406, "step": 18579 }, { "epoch": 0.654703700767286, "grad_norm": 3.3997654914855957, "learning_rate": 5.652864965156692e-06, "loss": 2.6734, "step": 18580 }, { "epoch": 0.6547389377802444, "grad_norm": 6.594940662384033, "learning_rate": 5.651837195185612e-06, "loss": 2.3166, "step": 18581 }, { "epoch": 0.6547741747932028, "grad_norm": 1.7319726943969727, "learning_rate": 5.650809481849903e-06, "loss": 1.0683, "step": 18582 }, { "epoch": 0.6548094118061611, "grad_norm": 1.6458868980407715, "learning_rate": 5.64978182516295e-06, "loss": 1.1275, "step": 18583 }, { "epoch": 0.6548446488191196, "grad_norm": 5.203217506408691, "learning_rate": 5.64875422513815e-06, "loss": 4.9956, "step": 18584 }, { "epoch": 0.654879885832078, "grad_norm": 3.9403204917907715, "learning_rate": 5.647726681788877e-06, "loss": 0.8352, "step": 18585 }, { "epoch": 0.6549151228450364, "grad_norm": 1.6943013668060303, "learning_rate": 5.64669919512852e-06, "loss": 1.1427, "step": 18586 }, { "epoch": 0.6549503598579949, "grad_norm": 1.346221923828125, "learning_rate": 5.6456717651704586e-06, "loss": 0.7524, "step": 18587 }, { "epoch": 0.6549855968709533, "grad_norm": 1.5159952640533447, "learning_rate": 5.644644391928078e-06, "loss": 0.77, "step": 18588 }, { "epoch": 0.6550208338839116, "grad_norm": 2.992189884185791, "learning_rate": 5.643617075414758e-06, "loss": 2.99, "step": 18589 }, { "epoch": 0.6550560708968701, "grad_norm": 3.685868978500366, "learning_rate": 5.642589815643882e-06, "loss": 2.786, "step": 18590 }, { "epoch": 0.6550913079098285, "grad_norm": 1.4443562030792236, "learning_rate": 5.641562612628827e-06, "loss": 0.6352, "step": 18591 }, { "epoch": 0.6551265449227869, "grad_norm": 1.0821105241775513, "learning_rate": 5.640535466382975e-06, "loss": 0.8657, "step": 18592 }, { "epoch": 0.6551617819357453, "grad_norm": 15.81167221069336, "learning_rate": 5.639508376919705e-06, "loss": 2.8043, "step": 18593 }, { "epoch": 0.6551970189487037, "grad_norm": 3.271836996078491, "learning_rate": 5.638481344252391e-06, "loss": 2.9886, "step": 18594 }, { "epoch": 0.6552322559616621, "grad_norm": 9.879096984863281, "learning_rate": 5.637454368394415e-06, "loss": 5.5943, "step": 18595 }, { "epoch": 0.6552674929746205, "grad_norm": 1.5823837518692017, "learning_rate": 5.636427449359152e-06, "loss": 0.9892, "step": 18596 }, { "epoch": 0.655302729987579, "grad_norm": 1.3627883195877075, "learning_rate": 5.635400587159981e-06, "loss": 0.7591, "step": 18597 }, { "epoch": 0.6553379670005374, "grad_norm": 1.1781249046325684, "learning_rate": 5.6343737818102635e-06, "loss": 0.9763, "step": 18598 }, { "epoch": 0.6553732040134957, "grad_norm": 1.2462576627731323, "learning_rate": 5.633347033323389e-06, "loss": 1.333, "step": 18599 }, { "epoch": 0.6554084410264542, "grad_norm": 11.889388084411621, "learning_rate": 5.632320341712729e-06, "loss": 5.0986, "step": 18600 }, { "epoch": 0.6554436780394126, "grad_norm": 1.6946526765823364, "learning_rate": 5.631293706991651e-06, "loss": 0.8413, "step": 18601 }, { "epoch": 0.655478915052371, "grad_norm": 4.043195724487305, "learning_rate": 5.630267129173523e-06, "loss": 2.8081, "step": 18602 }, { "epoch": 0.6555141520653294, "grad_norm": 16.636749267578125, "learning_rate": 5.629240608271731e-06, "loss": 2.9305, "step": 18603 }, { "epoch": 0.6555493890782879, "grad_norm": 1.6783461570739746, "learning_rate": 5.628214144299633e-06, "loss": 0.8803, "step": 18604 }, { "epoch": 0.6555846260912462, "grad_norm": 1.0470024347305298, "learning_rate": 5.627187737270601e-06, "loss": 0.8614, "step": 18605 }, { "epoch": 0.6556198631042046, "grad_norm": 1.2199070453643799, "learning_rate": 5.626161387198008e-06, "loss": 1.0114, "step": 18606 }, { "epoch": 0.6556551001171631, "grad_norm": 4.486076831817627, "learning_rate": 5.625135094095218e-06, "loss": 2.9251, "step": 18607 }, { "epoch": 0.6556903371301215, "grad_norm": 5.609832763671875, "learning_rate": 5.624108857975602e-06, "loss": 3.2692, "step": 18608 }, { "epoch": 0.6557255741430799, "grad_norm": 1.40882408618927, "learning_rate": 5.623082678852525e-06, "loss": 1.0643, "step": 18609 }, { "epoch": 0.6557608111560383, "grad_norm": 1.1282217502593994, "learning_rate": 5.622056556739355e-06, "loss": 0.934, "step": 18610 }, { "epoch": 0.6557960481689967, "grad_norm": 3.7384157180786133, "learning_rate": 5.621030491649454e-06, "loss": 3.0374, "step": 18611 }, { "epoch": 0.6558312851819551, "grad_norm": 2.8305513858795166, "learning_rate": 5.620004483596189e-06, "loss": 3.4905, "step": 18612 }, { "epoch": 0.6558665221949136, "grad_norm": 1.173348069190979, "learning_rate": 5.618978532592924e-06, "loss": 0.7737, "step": 18613 }, { "epoch": 0.655901759207872, "grad_norm": 4.141885280609131, "learning_rate": 5.6179526386530215e-06, "loss": 3.073, "step": 18614 }, { "epoch": 0.6559369962208303, "grad_norm": 6.992356777191162, "learning_rate": 5.616926801789845e-06, "loss": 3.0658, "step": 18615 }, { "epoch": 0.6559722332337887, "grad_norm": 0.9838780760765076, "learning_rate": 5.615901022016759e-06, "loss": 0.9154, "step": 18616 }, { "epoch": 0.6560074702467472, "grad_norm": 1.6603496074676514, "learning_rate": 5.614875299347111e-06, "loss": 0.9369, "step": 18617 }, { "epoch": 0.6560427072597056, "grad_norm": 3.384593963623047, "learning_rate": 5.613849633794276e-06, "loss": 2.1939, "step": 18618 }, { "epoch": 0.656077944272664, "grad_norm": 1.4555855989456177, "learning_rate": 5.612824025371613e-06, "loss": 0.8974, "step": 18619 }, { "epoch": 0.6561131812856225, "grad_norm": 3.0664966106414795, "learning_rate": 5.61179847409247e-06, "loss": 2.5571, "step": 18620 }, { "epoch": 0.6561484182985808, "grad_norm": 1.2895359992980957, "learning_rate": 5.6107729799702085e-06, "loss": 0.7161, "step": 18621 }, { "epoch": 0.6561836553115392, "grad_norm": 2.107515811920166, "learning_rate": 5.609747543018194e-06, "loss": 0.839, "step": 18622 }, { "epoch": 0.6562188923244977, "grad_norm": 8.545401573181152, "learning_rate": 5.608722163249776e-06, "loss": 5.3776, "step": 18623 }, { "epoch": 0.6562541293374561, "grad_norm": 5.602674961090088, "learning_rate": 5.607696840678309e-06, "loss": 5.1791, "step": 18624 }, { "epoch": 0.6562893663504145, "grad_norm": 4.604748725891113, "learning_rate": 5.606671575317151e-06, "loss": 3.0356, "step": 18625 }, { "epoch": 0.6563246033633728, "grad_norm": 5.154394149780273, "learning_rate": 5.605646367179654e-06, "loss": 3.1581, "step": 18626 }, { "epoch": 0.6563598403763313, "grad_norm": 3.7626001834869385, "learning_rate": 5.604621216279173e-06, "loss": 2.8937, "step": 18627 }, { "epoch": 0.6563950773892897, "grad_norm": 7.1168951988220215, "learning_rate": 5.603596122629063e-06, "loss": 5.3798, "step": 18628 }, { "epoch": 0.6564303144022481, "grad_norm": 0.9833725690841675, "learning_rate": 5.602571086242672e-06, "loss": 0.9465, "step": 18629 }, { "epoch": 0.6564655514152066, "grad_norm": 3.3462767601013184, "learning_rate": 5.601546107133352e-06, "loss": 2.7914, "step": 18630 }, { "epoch": 0.656500788428165, "grad_norm": 1.545885682106018, "learning_rate": 5.600521185314456e-06, "loss": 0.6655, "step": 18631 }, { "epoch": 0.6565360254411233, "grad_norm": 0.9844536781311035, "learning_rate": 5.59949632079933e-06, "loss": 1.0794, "step": 18632 }, { "epoch": 0.6565712624540818, "grad_norm": 1.2126717567443848, "learning_rate": 5.598471513601326e-06, "loss": 0.8748, "step": 18633 }, { "epoch": 0.6566064994670402, "grad_norm": 1.5426937341690063, "learning_rate": 5.597446763733792e-06, "loss": 0.7985, "step": 18634 }, { "epoch": 0.6566417364799986, "grad_norm": 4.809416770935059, "learning_rate": 5.596422071210077e-06, "loss": 5.9415, "step": 18635 }, { "epoch": 0.656676973492957, "grad_norm": 1.2180067300796509, "learning_rate": 5.5953974360435174e-06, "loss": 0.7764, "step": 18636 }, { "epoch": 0.6567122105059154, "grad_norm": 4.889868259429932, "learning_rate": 5.594372858247472e-06, "loss": 3.3304, "step": 18637 }, { "epoch": 0.6567474475188738, "grad_norm": 1.2506011724472046, "learning_rate": 5.593348337835285e-06, "loss": 1.0491, "step": 18638 }, { "epoch": 0.6567826845318322, "grad_norm": 3.7844700813293457, "learning_rate": 5.592323874820294e-06, "loss": 2.8647, "step": 18639 }, { "epoch": 0.6568179215447907, "grad_norm": 6.092862129211426, "learning_rate": 5.591299469215842e-06, "loss": 2.5927, "step": 18640 }, { "epoch": 0.6568531585577491, "grad_norm": 1.1990529298782349, "learning_rate": 5.5902751210352835e-06, "loss": 1.1696, "step": 18641 }, { "epoch": 0.6568883955707074, "grad_norm": 0.7792985439300537, "learning_rate": 5.589250830291949e-06, "loss": 0.8921, "step": 18642 }, { "epoch": 0.6569236325836659, "grad_norm": 5.1150383949279785, "learning_rate": 5.588226596999184e-06, "loss": 3.2997, "step": 18643 }, { "epoch": 0.6569588695966243, "grad_norm": 1.1853870153427124, "learning_rate": 5.5872024211703305e-06, "loss": 0.8981, "step": 18644 }, { "epoch": 0.6569941066095827, "grad_norm": 2.4965384006500244, "learning_rate": 5.586178302818728e-06, "loss": 1.0954, "step": 18645 }, { "epoch": 0.6570293436225411, "grad_norm": 1.0480867624282837, "learning_rate": 5.585154241957713e-06, "loss": 1.1006, "step": 18646 }, { "epoch": 0.6570645806354996, "grad_norm": 1.6515830755233765, "learning_rate": 5.584130238600629e-06, "loss": 0.9388, "step": 18647 }, { "epoch": 0.6570998176484579, "grad_norm": 4.370319366455078, "learning_rate": 5.5831062927608096e-06, "loss": 2.7683, "step": 18648 }, { "epoch": 0.6571350546614163, "grad_norm": 4.648461818695068, "learning_rate": 5.582082404451594e-06, "loss": 4.722, "step": 18649 }, { "epoch": 0.6571702916743748, "grad_norm": 6.232800483703613, "learning_rate": 5.581058573686318e-06, "loss": 0.8536, "step": 18650 }, { "epoch": 0.6572055286873332, "grad_norm": 4.822461128234863, "learning_rate": 5.580034800478315e-06, "loss": 3.0483, "step": 18651 }, { "epoch": 0.6572407657002916, "grad_norm": 1.0209707021713257, "learning_rate": 5.579011084840924e-06, "loss": 0.6784, "step": 18652 }, { "epoch": 0.65727600271325, "grad_norm": 1.311578631401062, "learning_rate": 5.577987426787477e-06, "loss": 0.7841, "step": 18653 }, { "epoch": 0.6573112397262084, "grad_norm": 3.1691949367523193, "learning_rate": 5.576963826331309e-06, "loss": 2.7787, "step": 18654 }, { "epoch": 0.6573464767391668, "grad_norm": 2.542924404144287, "learning_rate": 5.575940283485744e-06, "loss": 0.7996, "step": 18655 }, { "epoch": 0.6573817137521253, "grad_norm": 2.6059811115264893, "learning_rate": 5.574916798264124e-06, "loss": 2.5378, "step": 18656 }, { "epoch": 0.6574169507650837, "grad_norm": 1.7232731580734253, "learning_rate": 5.573893370679781e-06, "loss": 0.9296, "step": 18657 }, { "epoch": 0.657452187778042, "grad_norm": 5.233950138092041, "learning_rate": 5.572870000746035e-06, "loss": 3.4042, "step": 18658 }, { "epoch": 0.6574874247910004, "grad_norm": 4.913528919219971, "learning_rate": 5.571846688476222e-06, "loss": 3.2043, "step": 18659 }, { "epoch": 0.6575226618039589, "grad_norm": 8.177964210510254, "learning_rate": 5.5708234338836695e-06, "loss": 6.6294, "step": 18660 }, { "epoch": 0.6575578988169173, "grad_norm": 2.564913034439087, "learning_rate": 5.569800236981709e-06, "loss": 2.9418, "step": 18661 }, { "epoch": 0.6575931358298757, "grad_norm": 1.4721996784210205, "learning_rate": 5.568777097783661e-06, "loss": 0.7515, "step": 18662 }, { "epoch": 0.6576283728428342, "grad_norm": 0.901282787322998, "learning_rate": 5.567754016302858e-06, "loss": 0.7337, "step": 18663 }, { "epoch": 0.6576636098557925, "grad_norm": 2.322974920272827, "learning_rate": 5.5667309925526214e-06, "loss": 2.5532, "step": 18664 }, { "epoch": 0.6576988468687509, "grad_norm": 3.6591994762420654, "learning_rate": 5.5657080265462795e-06, "loss": 3.115, "step": 18665 }, { "epoch": 0.6577340838817094, "grad_norm": 3.3539538383483887, "learning_rate": 5.564685118297155e-06, "loss": 2.9464, "step": 18666 }, { "epoch": 0.6577693208946678, "grad_norm": 6.303050994873047, "learning_rate": 5.563662267818573e-06, "loss": 4.6416, "step": 18667 }, { "epoch": 0.6578045579076262, "grad_norm": 2.569622755050659, "learning_rate": 5.562639475123854e-06, "loss": 2.7396, "step": 18668 }, { "epoch": 0.6578397949205845, "grad_norm": 0.8764735460281372, "learning_rate": 5.561616740226321e-06, "loss": 0.6909, "step": 18669 }, { "epoch": 0.657875031933543, "grad_norm": 21.460256576538086, "learning_rate": 5.5605940631392996e-06, "loss": 2.9629, "step": 18670 }, { "epoch": 0.6579102689465014, "grad_norm": 1.7323108911514282, "learning_rate": 5.559571443876098e-06, "loss": 1.1562, "step": 18671 }, { "epoch": 0.6579455059594598, "grad_norm": 4.1455254554748535, "learning_rate": 5.5585488824500475e-06, "loss": 2.7909, "step": 18672 }, { "epoch": 0.6579807429724183, "grad_norm": 1.0212020874023438, "learning_rate": 5.5575263788744695e-06, "loss": 1.1077, "step": 18673 }, { "epoch": 0.6580159799853766, "grad_norm": 18.576133728027344, "learning_rate": 5.55650393316267e-06, "loss": 4.6588, "step": 18674 }, { "epoch": 0.658051216998335, "grad_norm": 1.104629635810852, "learning_rate": 5.55548154532797e-06, "loss": 0.8626, "step": 18675 }, { "epoch": 0.6580864540112935, "grad_norm": 4.5489349365234375, "learning_rate": 5.5544592153836965e-06, "loss": 5.4856, "step": 18676 }, { "epoch": 0.6581216910242519, "grad_norm": 1.14803147315979, "learning_rate": 5.553436943343154e-06, "loss": 0.7978, "step": 18677 }, { "epoch": 0.6581569280372103, "grad_norm": 3.8890676498413086, "learning_rate": 5.552414729219662e-06, "loss": 4.9671, "step": 18678 }, { "epoch": 0.6581921650501686, "grad_norm": 4.655867576599121, "learning_rate": 5.551392573026535e-06, "loss": 5.398, "step": 18679 }, { "epoch": 0.6582274020631271, "grad_norm": 3.017380952835083, "learning_rate": 5.550370474777086e-06, "loss": 2.7087, "step": 18680 }, { "epoch": 0.6582626390760855, "grad_norm": 1.7300915718078613, "learning_rate": 5.549348434484628e-06, "loss": 0.7319, "step": 18681 }, { "epoch": 0.6582978760890439, "grad_norm": 4.248936653137207, "learning_rate": 5.548326452162473e-06, "loss": 3.1621, "step": 18682 }, { "epoch": 0.6583331131020024, "grad_norm": 4.899859428405762, "learning_rate": 5.547304527823935e-06, "loss": 3.2474, "step": 18683 }, { "epoch": 0.6583683501149608, "grad_norm": 1.352128267288208, "learning_rate": 5.54628266148232e-06, "loss": 0.6986, "step": 18684 }, { "epoch": 0.6584035871279191, "grad_norm": 5.201448917388916, "learning_rate": 5.545260853150944e-06, "loss": 2.8443, "step": 18685 }, { "epoch": 0.6584388241408776, "grad_norm": 1.4284480810165405, "learning_rate": 5.544239102843111e-06, "loss": 0.7419, "step": 18686 }, { "epoch": 0.658474061153836, "grad_norm": 3.1588478088378906, "learning_rate": 5.54321741057213e-06, "loss": 2.5304, "step": 18687 }, { "epoch": 0.6585092981667944, "grad_norm": 1.7528343200683594, "learning_rate": 5.542195776351312e-06, "loss": 1.0944, "step": 18688 }, { "epoch": 0.6585445351797529, "grad_norm": 6.077995300292969, "learning_rate": 5.541174200193966e-06, "loss": 3.4394, "step": 18689 }, { "epoch": 0.6585797721927112, "grad_norm": 3.332064628601074, "learning_rate": 5.540152682113385e-06, "loss": 2.8411, "step": 18690 }, { "epoch": 0.6586150092056696, "grad_norm": 1.9864226579666138, "learning_rate": 5.539131222122887e-06, "loss": 0.8091, "step": 18691 }, { "epoch": 0.658650246218628, "grad_norm": 1.1001166105270386, "learning_rate": 5.5381098202357776e-06, "loss": 0.7266, "step": 18692 }, { "epoch": 0.6586854832315865, "grad_norm": 5.454890727996826, "learning_rate": 5.537088476465352e-06, "loss": 4.915, "step": 18693 }, { "epoch": 0.6587207202445449, "grad_norm": 10.675078392028809, "learning_rate": 5.536067190824914e-06, "loss": 3.1683, "step": 18694 }, { "epoch": 0.6587559572575032, "grad_norm": 5.582509517669678, "learning_rate": 5.535045963327777e-06, "loss": 4.9648, "step": 18695 }, { "epoch": 0.6587911942704617, "grad_norm": 3.994515895843506, "learning_rate": 5.534024793987232e-06, "loss": 3.1053, "step": 18696 }, { "epoch": 0.6588264312834201, "grad_norm": 7.098285675048828, "learning_rate": 5.533003682816581e-06, "loss": 5.484, "step": 18697 }, { "epoch": 0.6588616682963785, "grad_norm": 1.263228416442871, "learning_rate": 5.531982629829127e-06, "loss": 0.9297, "step": 18698 }, { "epoch": 0.658896905309337, "grad_norm": 3.740251302719116, "learning_rate": 5.530961635038168e-06, "loss": 2.6069, "step": 18699 }, { "epoch": 0.6589321423222954, "grad_norm": 4.362308502197266, "learning_rate": 5.529940698457003e-06, "loss": 3.177, "step": 18700 }, { "epoch": 0.6589673793352537, "grad_norm": 1.299448847770691, "learning_rate": 5.528919820098932e-06, "loss": 0.8719, "step": 18701 }, { "epoch": 0.6590026163482121, "grad_norm": 1.8907886743545532, "learning_rate": 5.5278989999772485e-06, "loss": 0.9362, "step": 18702 }, { "epoch": 0.6590378533611706, "grad_norm": 1.4483435153961182, "learning_rate": 5.5268782381052485e-06, "loss": 0.8785, "step": 18703 }, { "epoch": 0.659073090374129, "grad_norm": 5.024356365203857, "learning_rate": 5.52585753449623e-06, "loss": 2.945, "step": 18704 }, { "epoch": 0.6591083273870874, "grad_norm": 0.829547643661499, "learning_rate": 5.524836889163489e-06, "loss": 0.9049, "step": 18705 }, { "epoch": 0.6591435644000458, "grad_norm": 8.418097496032715, "learning_rate": 5.523816302120315e-06, "loss": 7.9621, "step": 18706 }, { "epoch": 0.6591788014130042, "grad_norm": 1.810964822769165, "learning_rate": 5.522795773380005e-06, "loss": 0.8423, "step": 18707 }, { "epoch": 0.6592140384259626, "grad_norm": 1.3764575719833374, "learning_rate": 5.5217753029558544e-06, "loss": 0.9415, "step": 18708 }, { "epoch": 0.6592492754389211, "grad_norm": 1.372151494026184, "learning_rate": 5.520754890861143e-06, "loss": 1.0301, "step": 18709 }, { "epoch": 0.6592845124518795, "grad_norm": 3.7618448734283447, "learning_rate": 5.519734537109173e-06, "loss": 3.3323, "step": 18710 }, { "epoch": 0.6593197494648378, "grad_norm": 0.9637882113456726, "learning_rate": 5.518714241713237e-06, "loss": 0.6597, "step": 18711 }, { "epoch": 0.6593549864777962, "grad_norm": 1.4073106050491333, "learning_rate": 5.517694004686614e-06, "loss": 0.7333, "step": 18712 }, { "epoch": 0.6593902234907547, "grad_norm": 1.744328260421753, "learning_rate": 5.516673826042594e-06, "loss": 0.9492, "step": 18713 }, { "epoch": 0.6594254605037131, "grad_norm": 4.429995536804199, "learning_rate": 5.515653705794475e-06, "loss": 2.8949, "step": 18714 }, { "epoch": 0.6594606975166715, "grad_norm": 1.0421974658966064, "learning_rate": 5.5146336439555335e-06, "loss": 0.8516, "step": 18715 }, { "epoch": 0.65949593452963, "grad_norm": 3.1665408611297607, "learning_rate": 5.5136136405390615e-06, "loss": 3.123, "step": 18716 }, { "epoch": 0.6595311715425883, "grad_norm": 1.2896678447723389, "learning_rate": 5.512593695558344e-06, "loss": 0.7126, "step": 18717 }, { "epoch": 0.6595664085555467, "grad_norm": 1.141306757926941, "learning_rate": 5.511573809026662e-06, "loss": 0.8597, "step": 18718 }, { "epoch": 0.6596016455685052, "grad_norm": 1.2214531898498535, "learning_rate": 5.510553980957306e-06, "loss": 1.0408, "step": 18719 }, { "epoch": 0.6596368825814636, "grad_norm": 5.7695183753967285, "learning_rate": 5.509534211363555e-06, "loss": 4.7557, "step": 18720 }, { "epoch": 0.659672119594422, "grad_norm": 1.554472804069519, "learning_rate": 5.508514500258693e-06, "loss": 0.8609, "step": 18721 }, { "epoch": 0.6597073566073804, "grad_norm": 1.2020988464355469, "learning_rate": 5.507494847656003e-06, "loss": 0.785, "step": 18722 }, { "epoch": 0.6597425936203388, "grad_norm": 1.487890601158142, "learning_rate": 5.506475253568764e-06, "loss": 0.8734, "step": 18723 }, { "epoch": 0.6597778306332972, "grad_norm": 1.2497352361679077, "learning_rate": 5.5054557180102564e-06, "loss": 0.9392, "step": 18724 }, { "epoch": 0.6598130676462556, "grad_norm": 0.9223514199256897, "learning_rate": 5.504436240993761e-06, "loss": 0.9444, "step": 18725 }, { "epoch": 0.6598483046592141, "grad_norm": 3.7263169288635254, "learning_rate": 5.503416822532557e-06, "loss": 2.2952, "step": 18726 }, { "epoch": 0.6598835416721724, "grad_norm": 5.058456897735596, "learning_rate": 5.502397462639924e-06, "loss": 4.3947, "step": 18727 }, { "epoch": 0.6599187786851308, "grad_norm": 1.1143244504928589, "learning_rate": 5.501378161329129e-06, "loss": 0.5532, "step": 18728 }, { "epoch": 0.6599540156980893, "grad_norm": 5.553747653961182, "learning_rate": 5.500358918613462e-06, "loss": 5.3583, "step": 18729 }, { "epoch": 0.6599892527110477, "grad_norm": 1.2646321058273315, "learning_rate": 5.499339734506197e-06, "loss": 0.738, "step": 18730 }, { "epoch": 0.6600244897240061, "grad_norm": 5.273525714874268, "learning_rate": 5.4983206090206024e-06, "loss": 5.1941, "step": 18731 }, { "epoch": 0.6600597267369646, "grad_norm": 7.2625885009765625, "learning_rate": 5.4973015421699484e-06, "loss": 4.7419, "step": 18732 }, { "epoch": 0.6600949637499229, "grad_norm": 1.2220687866210938, "learning_rate": 5.496282533967525e-06, "loss": 1.1549, "step": 18733 }, { "epoch": 0.6601302007628813, "grad_norm": 3.7436485290527344, "learning_rate": 5.495263584426591e-06, "loss": 2.5681, "step": 18734 }, { "epoch": 0.6601654377758397, "grad_norm": 0.9668826460838318, "learning_rate": 5.494244693560423e-06, "loss": 0.843, "step": 18735 }, { "epoch": 0.6602006747887982, "grad_norm": 1.5303360223770142, "learning_rate": 5.493225861382292e-06, "loss": 0.6225, "step": 18736 }, { "epoch": 0.6602359118017566, "grad_norm": 3.361064910888672, "learning_rate": 5.492207087905468e-06, "loss": 2.6228, "step": 18737 }, { "epoch": 0.6602711488147149, "grad_norm": 4.210919380187988, "learning_rate": 5.491188373143219e-06, "loss": 5.1928, "step": 18738 }, { "epoch": 0.6603063858276734, "grad_norm": 5.135347843170166, "learning_rate": 5.490169717108817e-06, "loss": 4.9749, "step": 18739 }, { "epoch": 0.6603416228406318, "grad_norm": 0.9882996082305908, "learning_rate": 5.48915111981553e-06, "loss": 0.8732, "step": 18740 }, { "epoch": 0.6603768598535902, "grad_norm": 1.0871654748916626, "learning_rate": 5.4881325812766235e-06, "loss": 0.9645, "step": 18741 }, { "epoch": 0.6604120968665487, "grad_norm": 2.41218638420105, "learning_rate": 5.487114101505367e-06, "loss": 0.8308, "step": 18742 }, { "epoch": 0.660447333879507, "grad_norm": 6.6334547996521, "learning_rate": 5.48609568051502e-06, "loss": 5.4018, "step": 18743 }, { "epoch": 0.6604825708924654, "grad_norm": 4.370026588439941, "learning_rate": 5.485077318318854e-06, "loss": 2.6007, "step": 18744 }, { "epoch": 0.6605178079054238, "grad_norm": 1.905889630317688, "learning_rate": 5.484059014930132e-06, "loss": 0.8827, "step": 18745 }, { "epoch": 0.6605530449183823, "grad_norm": 3.9574878215789795, "learning_rate": 5.483040770362118e-06, "loss": 3.0904, "step": 18746 }, { "epoch": 0.6605882819313407, "grad_norm": 2.3238415718078613, "learning_rate": 5.482022584628068e-06, "loss": 0.8769, "step": 18747 }, { "epoch": 0.660623518944299, "grad_norm": 1.7124091386795044, "learning_rate": 5.481004457741252e-06, "loss": 0.8224, "step": 18748 }, { "epoch": 0.6606587559572575, "grad_norm": 3.42392897605896, "learning_rate": 5.479986389714933e-06, "loss": 3.0238, "step": 18749 }, { "epoch": 0.6606939929702159, "grad_norm": 3.2911369800567627, "learning_rate": 5.478968380562362e-06, "loss": 2.5748, "step": 18750 }, { "epoch": 0.6607292299831743, "grad_norm": 1.2741296291351318, "learning_rate": 5.477950430296801e-06, "loss": 0.8784, "step": 18751 }, { "epoch": 0.6607644669961328, "grad_norm": 0.9998449087142944, "learning_rate": 5.476932538931519e-06, "loss": 0.9094, "step": 18752 }, { "epoch": 0.6607997040090912, "grad_norm": 5.2098798751831055, "learning_rate": 5.475914706479762e-06, "loss": 5.1419, "step": 18753 }, { "epoch": 0.6608349410220495, "grad_norm": 0.9876176118850708, "learning_rate": 5.474896932954794e-06, "loss": 0.896, "step": 18754 }, { "epoch": 0.660870178035008, "grad_norm": 0.9291267991065979, "learning_rate": 5.473879218369869e-06, "loss": 1.1154, "step": 18755 }, { "epoch": 0.6609054150479664, "grad_norm": 3.839752197265625, "learning_rate": 5.472861562738243e-06, "loss": 3.0181, "step": 18756 }, { "epoch": 0.6609406520609248, "grad_norm": 1.3144506216049194, "learning_rate": 5.471843966073174e-06, "loss": 0.9404, "step": 18757 }, { "epoch": 0.6609758890738832, "grad_norm": 1.2824186086654663, "learning_rate": 5.470826428387912e-06, "loss": 1.0086, "step": 18758 }, { "epoch": 0.6610111260868416, "grad_norm": 1.9392210245132446, "learning_rate": 5.469808949695713e-06, "loss": 1.0672, "step": 18759 }, { "epoch": 0.6610463630998, "grad_norm": 2.8168795108795166, "learning_rate": 5.4687915300098294e-06, "loss": 2.6284, "step": 18760 }, { "epoch": 0.6610816001127584, "grad_norm": 1.5104107856750488, "learning_rate": 5.4677741693435136e-06, "loss": 0.9453, "step": 18761 }, { "epoch": 0.6611168371257169, "grad_norm": 4.297476768493652, "learning_rate": 5.466756867710017e-06, "loss": 2.8399, "step": 18762 }, { "epoch": 0.6611520741386753, "grad_norm": 5.776416301727295, "learning_rate": 5.46573962512259e-06, "loss": 5.1125, "step": 18763 }, { "epoch": 0.6611873111516336, "grad_norm": 4.267692565917969, "learning_rate": 5.464722441594481e-06, "loss": 2.8827, "step": 18764 }, { "epoch": 0.6612225481645921, "grad_norm": 5.303689479827881, "learning_rate": 5.463705317138944e-06, "loss": 4.7089, "step": 18765 }, { "epoch": 0.6612577851775505, "grad_norm": 3.3348915576934814, "learning_rate": 5.462688251769214e-06, "loss": 2.4898, "step": 18766 }, { "epoch": 0.6612930221905089, "grad_norm": 1.4034857749938965, "learning_rate": 5.461671245498555e-06, "loss": 0.8191, "step": 18767 }, { "epoch": 0.6613282592034673, "grad_norm": 6.5419182777404785, "learning_rate": 5.460654298340207e-06, "loss": 4.7603, "step": 18768 }, { "epoch": 0.6613634962164258, "grad_norm": 3.6802892684936523, "learning_rate": 5.459637410307412e-06, "loss": 2.8555, "step": 18769 }, { "epoch": 0.6613987332293841, "grad_norm": 2.2786550521850586, "learning_rate": 5.4586205814134145e-06, "loss": 1.0547, "step": 18770 }, { "epoch": 0.6614339702423425, "grad_norm": 8.073001861572266, "learning_rate": 5.457603811671469e-06, "loss": 7.1152, "step": 18771 }, { "epoch": 0.661469207255301, "grad_norm": 1.9116617441177368, "learning_rate": 5.45658710109481e-06, "loss": 1.0054, "step": 18772 }, { "epoch": 0.6615044442682594, "grad_norm": 2.183403730392456, "learning_rate": 5.455570449696681e-06, "loss": 0.734, "step": 18773 }, { "epoch": 0.6615396812812178, "grad_norm": 2.452192544937134, "learning_rate": 5.454553857490328e-06, "loss": 2.9152, "step": 18774 }, { "epoch": 0.6615749182941763, "grad_norm": 2.751753568649292, "learning_rate": 5.4535373244889886e-06, "loss": 0.8529, "step": 18775 }, { "epoch": 0.6616101553071346, "grad_norm": 1.300178050994873, "learning_rate": 5.4525208507059045e-06, "loss": 1.1679, "step": 18776 }, { "epoch": 0.661645392320093, "grad_norm": 3.8632545471191406, "learning_rate": 5.451504436154316e-06, "loss": 2.9676, "step": 18777 }, { "epoch": 0.6616806293330514, "grad_norm": 3.769320011138916, "learning_rate": 5.450488080847461e-06, "loss": 4.1751, "step": 18778 }, { "epoch": 0.6617158663460099, "grad_norm": 10.392022132873535, "learning_rate": 5.449471784798579e-06, "loss": 2.806, "step": 18779 }, { "epoch": 0.6617511033589683, "grad_norm": 1.777613878250122, "learning_rate": 5.448455548020907e-06, "loss": 0.9018, "step": 18780 }, { "epoch": 0.6617863403719266, "grad_norm": 2.700528621673584, "learning_rate": 5.44743937052768e-06, "loss": 2.998, "step": 18781 }, { "epoch": 0.6618215773848851, "grad_norm": 4.689688682556152, "learning_rate": 5.446423252332136e-06, "loss": 2.8327, "step": 18782 }, { "epoch": 0.6618568143978435, "grad_norm": 3.8682491779327393, "learning_rate": 5.445407193447509e-06, "loss": 2.762, "step": 18783 }, { "epoch": 0.6618920514108019, "grad_norm": 7.670090675354004, "learning_rate": 5.444391193887038e-06, "loss": 6.795, "step": 18784 }, { "epoch": 0.6619272884237604, "grad_norm": 1.3069162368774414, "learning_rate": 5.443375253663943e-06, "loss": 1.0048, "step": 18785 }, { "epoch": 0.6619625254367187, "grad_norm": 1.7564175128936768, "learning_rate": 5.44235937279147e-06, "loss": 1.0174, "step": 18786 }, { "epoch": 0.6619977624496771, "grad_norm": 5.072279453277588, "learning_rate": 5.441343551282851e-06, "loss": 2.9591, "step": 18787 }, { "epoch": 0.6620329994626356, "grad_norm": 6.326255798339844, "learning_rate": 5.44032778915131e-06, "loss": 3.0615, "step": 18788 }, { "epoch": 0.662068236475594, "grad_norm": 1.7902227640151978, "learning_rate": 5.439312086410075e-06, "loss": 2.409, "step": 18789 }, { "epoch": 0.6621034734885524, "grad_norm": 5.2428483963012695, "learning_rate": 5.438296443072388e-06, "loss": 3.3832, "step": 18790 }, { "epoch": 0.6621387105015107, "grad_norm": 3.3453171253204346, "learning_rate": 5.4372808591514695e-06, "loss": 2.7077, "step": 18791 }, { "epoch": 0.6621739475144692, "grad_norm": 2.4539716243743896, "learning_rate": 5.436265334660547e-06, "loss": 0.8557, "step": 18792 }, { "epoch": 0.6622091845274276, "grad_norm": 3.641615152359009, "learning_rate": 5.435249869612851e-06, "loss": 3.176, "step": 18793 }, { "epoch": 0.662244421540386, "grad_norm": 3.3129823207855225, "learning_rate": 5.434234464021606e-06, "loss": 3.2971, "step": 18794 }, { "epoch": 0.6622796585533445, "grad_norm": 1.0036550760269165, "learning_rate": 5.433219117900039e-06, "loss": 0.6935, "step": 18795 }, { "epoch": 0.6623148955663029, "grad_norm": 4.867922306060791, "learning_rate": 5.432203831261376e-06, "loss": 2.8858, "step": 18796 }, { "epoch": 0.6623501325792612, "grad_norm": 4.449568748474121, "learning_rate": 5.431188604118837e-06, "loss": 3.4706, "step": 18797 }, { "epoch": 0.6623853695922197, "grad_norm": 2.7670845985412598, "learning_rate": 5.4301734364856505e-06, "loss": 2.6833, "step": 18798 }, { "epoch": 0.6624206066051781, "grad_norm": 1.537194013595581, "learning_rate": 5.4291583283750375e-06, "loss": 1.0089, "step": 18799 }, { "epoch": 0.6624558436181365, "grad_norm": 3.391146421432495, "learning_rate": 5.428143279800219e-06, "loss": 2.6232, "step": 18800 }, { "epoch": 0.6624910806310949, "grad_norm": 0.9608175754547119, "learning_rate": 5.427128290774416e-06, "loss": 0.6905, "step": 18801 }, { "epoch": 0.6625263176440533, "grad_norm": 3.0494418144226074, "learning_rate": 5.426113361310849e-06, "loss": 2.9582, "step": 18802 }, { "epoch": 0.6625615546570117, "grad_norm": 3.368751049041748, "learning_rate": 5.4250984914227445e-06, "loss": 2.588, "step": 18803 }, { "epoch": 0.6625967916699701, "grad_norm": 3.638749122619629, "learning_rate": 5.424083681123304e-06, "loss": 2.7167, "step": 18804 }, { "epoch": 0.6626320286829286, "grad_norm": 4.383744239807129, "learning_rate": 5.423068930425762e-06, "loss": 2.3857, "step": 18805 }, { "epoch": 0.662667265695887, "grad_norm": 1.0097873210906982, "learning_rate": 5.422054239343335e-06, "loss": 0.9089, "step": 18806 }, { "epoch": 0.6627025027088453, "grad_norm": 1.3796526193618774, "learning_rate": 5.421039607889229e-06, "loss": 0.8198, "step": 18807 }, { "epoch": 0.6627377397218038, "grad_norm": 2.13796067237854, "learning_rate": 5.420025036076662e-06, "loss": 2.7327, "step": 18808 }, { "epoch": 0.6627729767347622, "grad_norm": 1.409753680229187, "learning_rate": 5.419010523918861e-06, "loss": 1.1583, "step": 18809 }, { "epoch": 0.6628082137477206, "grad_norm": 4.001274585723877, "learning_rate": 5.417996071429026e-06, "loss": 2.9548, "step": 18810 }, { "epoch": 0.662843450760679, "grad_norm": 1.167351245880127, "learning_rate": 5.416981678620377e-06, "loss": 0.7823, "step": 18811 }, { "epoch": 0.6628786877736375, "grad_norm": 1.1954588890075684, "learning_rate": 5.415967345506122e-06, "loss": 0.7641, "step": 18812 }, { "epoch": 0.6629139247865958, "grad_norm": 2.128262996673584, "learning_rate": 5.414953072099482e-06, "loss": 0.8771, "step": 18813 }, { "epoch": 0.6629491617995542, "grad_norm": 2.8117125034332275, "learning_rate": 5.41393885841366e-06, "loss": 2.8749, "step": 18814 }, { "epoch": 0.6629843988125127, "grad_norm": 2.835247755050659, "learning_rate": 5.4129247044618686e-06, "loss": 2.6256, "step": 18815 }, { "epoch": 0.6630196358254711, "grad_norm": 2.9503960609436035, "learning_rate": 5.411910610257316e-06, "loss": 2.3864, "step": 18816 }, { "epoch": 0.6630548728384295, "grad_norm": 4.78928279876709, "learning_rate": 5.410896575813213e-06, "loss": 3.2612, "step": 18817 }, { "epoch": 0.6630901098513879, "grad_norm": 3.705777645111084, "learning_rate": 5.409882601142766e-06, "loss": 3.0487, "step": 18818 }, { "epoch": 0.6631253468643463, "grad_norm": 1.0107156038284302, "learning_rate": 5.408868686259183e-06, "loss": 0.8993, "step": 18819 }, { "epoch": 0.6631605838773047, "grad_norm": 4.275957107543945, "learning_rate": 5.407854831175672e-06, "loss": 2.7119, "step": 18820 }, { "epoch": 0.6631958208902631, "grad_norm": 5.37501859664917, "learning_rate": 5.4068410359054345e-06, "loss": 3.0974, "step": 18821 }, { "epoch": 0.6632310579032216, "grad_norm": 1.4880130290985107, "learning_rate": 5.405827300461682e-06, "loss": 0.7952, "step": 18822 }, { "epoch": 0.6632662949161799, "grad_norm": 5.7275824546813965, "learning_rate": 5.404813624857606e-06, "loss": 5.5548, "step": 18823 }, { "epoch": 0.6633015319291383, "grad_norm": 2.5676777362823486, "learning_rate": 5.403800009106424e-06, "loss": 2.4573, "step": 18824 }, { "epoch": 0.6633367689420968, "grad_norm": 3.6992218494415283, "learning_rate": 5.402786453221334e-06, "loss": 3.3884, "step": 18825 }, { "epoch": 0.6633720059550552, "grad_norm": 5.686279773712158, "learning_rate": 5.401772957215532e-06, "loss": 5.3181, "step": 18826 }, { "epoch": 0.6634072429680136, "grad_norm": 3.201646566390991, "learning_rate": 5.40075952110222e-06, "loss": 3.0726, "step": 18827 }, { "epoch": 0.663442479980972, "grad_norm": 1.5765111446380615, "learning_rate": 5.399746144894609e-06, "loss": 0.8587, "step": 18828 }, { "epoch": 0.6634777169939304, "grad_norm": 4.1734185218811035, "learning_rate": 5.398732828605886e-06, "loss": 3.2798, "step": 18829 }, { "epoch": 0.6635129540068888, "grad_norm": 0.9167117476463318, "learning_rate": 5.397719572249253e-06, "loss": 0.8439, "step": 18830 }, { "epoch": 0.6635481910198473, "grad_norm": 1.2285730838775635, "learning_rate": 5.396706375837908e-06, "loss": 0.9973, "step": 18831 }, { "epoch": 0.6635834280328057, "grad_norm": 2.5640666484832764, "learning_rate": 5.395693239385054e-06, "loss": 2.7575, "step": 18832 }, { "epoch": 0.6636186650457641, "grad_norm": 3.9308412075042725, "learning_rate": 5.394680162903879e-06, "loss": 2.6422, "step": 18833 }, { "epoch": 0.6636539020587224, "grad_norm": 4.89121150970459, "learning_rate": 5.393667146407582e-06, "loss": 3.209, "step": 18834 }, { "epoch": 0.6636891390716809, "grad_norm": 6.134122371673584, "learning_rate": 5.3926541899093566e-06, "loss": 5.7138, "step": 18835 }, { "epoch": 0.6637243760846393, "grad_norm": 0.7137956619262695, "learning_rate": 5.391641293422396e-06, "loss": 0.7061, "step": 18836 }, { "epoch": 0.6637596130975977, "grad_norm": 4.212110996246338, "learning_rate": 5.390628456959896e-06, "loss": 2.2504, "step": 18837 }, { "epoch": 0.6637948501105562, "grad_norm": 4.6462273597717285, "learning_rate": 5.389615680535047e-06, "loss": 2.2812, "step": 18838 }, { "epoch": 0.6638300871235145, "grad_norm": 1.0953216552734375, "learning_rate": 5.388602964161043e-06, "loss": 0.9411, "step": 18839 }, { "epoch": 0.6638653241364729, "grad_norm": 1.2853704690933228, "learning_rate": 5.3875903078510695e-06, "loss": 0.9423, "step": 18840 }, { "epoch": 0.6639005611494314, "grad_norm": 1.6115539073944092, "learning_rate": 5.386577711618326e-06, "loss": 0.9269, "step": 18841 }, { "epoch": 0.6639357981623898, "grad_norm": 6.061176776885986, "learning_rate": 5.385565175475986e-06, "loss": 4.2005, "step": 18842 }, { "epoch": 0.6639710351753482, "grad_norm": 1.2601196765899658, "learning_rate": 5.384552699437253e-06, "loss": 0.8715, "step": 18843 }, { "epoch": 0.6640062721883065, "grad_norm": 3.9024245738983154, "learning_rate": 5.383540283515312e-06, "loss": 3.1238, "step": 18844 }, { "epoch": 0.664041509201265, "grad_norm": 1.3180145025253296, "learning_rate": 5.382527927723343e-06, "loss": 0.9911, "step": 18845 }, { "epoch": 0.6640767462142234, "grad_norm": 2.9972987174987793, "learning_rate": 5.381515632074532e-06, "loss": 3.2602, "step": 18846 }, { "epoch": 0.6641119832271818, "grad_norm": 5.605334758758545, "learning_rate": 5.380503396582076e-06, "loss": 4.2941, "step": 18847 }, { "epoch": 0.6641472202401403, "grad_norm": 2.16984224319458, "learning_rate": 5.379491221259147e-06, "loss": 1.015, "step": 18848 }, { "epoch": 0.6641824572530987, "grad_norm": 1.3761955499649048, "learning_rate": 5.378479106118936e-06, "loss": 0.7604, "step": 18849 }, { "epoch": 0.664217694266057, "grad_norm": 1.6727240085601807, "learning_rate": 5.3774670511746165e-06, "loss": 0.8722, "step": 18850 }, { "epoch": 0.6642529312790155, "grad_norm": 1.6727240085601807, "learning_rate": 5.3774670511746165e-06, "loss": 5.2603, "step": 18851 }, { "epoch": 0.6642881682919739, "grad_norm": 3.9269447326660156, "learning_rate": 5.376455056439387e-06, "loss": 2.8217, "step": 18852 }, { "epoch": 0.6643234053049323, "grad_norm": 1.3147180080413818, "learning_rate": 5.3754431219264155e-06, "loss": 0.9587, "step": 18853 }, { "epoch": 0.6643586423178907, "grad_norm": 2.464661121368408, "learning_rate": 5.374431247648886e-06, "loss": 1.1107, "step": 18854 }, { "epoch": 0.6643938793308491, "grad_norm": 2.023289203643799, "learning_rate": 5.373419433619977e-06, "loss": 0.7887, "step": 18855 }, { "epoch": 0.6644291163438075, "grad_norm": 1.0075311660766602, "learning_rate": 5.3724076798528716e-06, "loss": 0.8265, "step": 18856 }, { "epoch": 0.6644643533567659, "grad_norm": 9.772688865661621, "learning_rate": 5.371395986360745e-06, "loss": 6.7504, "step": 18857 }, { "epoch": 0.6644995903697244, "grad_norm": 4.00703239440918, "learning_rate": 5.3703843531567745e-06, "loss": 3.2917, "step": 18858 }, { "epoch": 0.6645348273826828, "grad_norm": 4.5020904541015625, "learning_rate": 5.369372780254138e-06, "loss": 4.5207, "step": 18859 }, { "epoch": 0.6645700643956411, "grad_norm": 1.5197728872299194, "learning_rate": 5.368361267666011e-06, "loss": 0.9701, "step": 18860 }, { "epoch": 0.6646053014085996, "grad_norm": 5.347396373748779, "learning_rate": 5.367349815405572e-06, "loss": 7.8598, "step": 18861 }, { "epoch": 0.664640538421558, "grad_norm": 5.522266864776611, "learning_rate": 5.366338423485985e-06, "loss": 4.9905, "step": 18862 }, { "epoch": 0.6646757754345164, "grad_norm": 3.2717459201812744, "learning_rate": 5.365327091920434e-06, "loss": 3.0299, "step": 18863 }, { "epoch": 0.6647110124474749, "grad_norm": 4.018315315246582, "learning_rate": 5.36431582072209e-06, "loss": 3.631, "step": 18864 }, { "epoch": 0.6647462494604333, "grad_norm": 1.3821070194244385, "learning_rate": 5.3633046099041205e-06, "loss": 1.1026, "step": 18865 }, { "epoch": 0.6647814864733916, "grad_norm": 1.2827272415161133, "learning_rate": 5.362293459479695e-06, "loss": 0.981, "step": 18866 }, { "epoch": 0.66481672348635, "grad_norm": 3.0109145641326904, "learning_rate": 5.361282369461996e-06, "loss": 2.3302, "step": 18867 }, { "epoch": 0.6648519604993085, "grad_norm": 2.953265905380249, "learning_rate": 5.360271339864181e-06, "loss": 2.9919, "step": 18868 }, { "epoch": 0.6648871975122669, "grad_norm": 21.756593704223633, "learning_rate": 5.359260370699423e-06, "loss": 5.0501, "step": 18869 }, { "epoch": 0.6649224345252253, "grad_norm": 1.0989508628845215, "learning_rate": 5.358249461980886e-06, "loss": 1.0333, "step": 18870 }, { "epoch": 0.6649576715381837, "grad_norm": 2.6699819564819336, "learning_rate": 5.357238613721747e-06, "loss": 2.9311, "step": 18871 }, { "epoch": 0.6649929085511421, "grad_norm": 4.441648006439209, "learning_rate": 5.356227825935165e-06, "loss": 4.8751, "step": 18872 }, { "epoch": 0.6650281455641005, "grad_norm": 5.783633232116699, "learning_rate": 5.355217098634306e-06, "loss": 5.0202, "step": 18873 }, { "epoch": 0.665063382577059, "grad_norm": 3.3123810291290283, "learning_rate": 5.354206431832336e-06, "loss": 3.3311, "step": 18874 }, { "epoch": 0.6650986195900174, "grad_norm": 2.1032958030700684, "learning_rate": 5.353195825542419e-06, "loss": 2.65, "step": 18875 }, { "epoch": 0.6651338566029757, "grad_norm": 0.8606825470924377, "learning_rate": 5.3521852797777175e-06, "loss": 0.6409, "step": 18876 }, { "epoch": 0.6651690936159341, "grad_norm": 3.5155439376831055, "learning_rate": 5.351174794551396e-06, "loss": 3.1585, "step": 18877 }, { "epoch": 0.6652043306288926, "grad_norm": 3.8898544311523438, "learning_rate": 5.350164369876615e-06, "loss": 5.4023, "step": 18878 }, { "epoch": 0.665239567641851, "grad_norm": 1.7860441207885742, "learning_rate": 5.349154005766535e-06, "loss": 0.7535, "step": 18879 }, { "epoch": 0.6652748046548094, "grad_norm": 1.6893234252929688, "learning_rate": 5.34814370223432e-06, "loss": 0.8833, "step": 18880 }, { "epoch": 0.6653100416677679, "grad_norm": 2.180114507675171, "learning_rate": 5.347133459293118e-06, "loss": 0.7231, "step": 18881 }, { "epoch": 0.6653452786807262, "grad_norm": 4.35331916809082, "learning_rate": 5.346123276956099e-06, "loss": 3.0065, "step": 18882 }, { "epoch": 0.6653805156936846, "grad_norm": 1.3211380243301392, "learning_rate": 5.345113155236422e-06, "loss": 0.9492, "step": 18883 }, { "epoch": 0.6654157527066431, "grad_norm": 2.1018338203430176, "learning_rate": 5.344103094147235e-06, "loss": 0.8165, "step": 18884 }, { "epoch": 0.6654509897196015, "grad_norm": 4.756684303283691, "learning_rate": 5.343093093701695e-06, "loss": 3.3569, "step": 18885 }, { "epoch": 0.6654862267325599, "grad_norm": 1.4684696197509766, "learning_rate": 5.342083153912963e-06, "loss": 0.9731, "step": 18886 }, { "epoch": 0.6655214637455182, "grad_norm": 2.229356288909912, "learning_rate": 5.341073274794195e-06, "loss": 0.9821, "step": 18887 }, { "epoch": 0.6655567007584767, "grad_norm": 1.3307164907455444, "learning_rate": 5.340063456358539e-06, "loss": 1.213, "step": 18888 }, { "epoch": 0.6655919377714351, "grad_norm": 1.8737386465072632, "learning_rate": 5.339053698619145e-06, "loss": 0.9419, "step": 18889 }, { "epoch": 0.6656271747843935, "grad_norm": 4.783186912536621, "learning_rate": 5.338044001589178e-06, "loss": 2.9242, "step": 18890 }, { "epoch": 0.665662411797352, "grad_norm": 1.2949053049087524, "learning_rate": 5.337034365281779e-06, "loss": 0.6599, "step": 18891 }, { "epoch": 0.6656976488103103, "grad_norm": 4.704491138458252, "learning_rate": 5.336024789710099e-06, "loss": 6.3987, "step": 18892 }, { "epoch": 0.6657328858232687, "grad_norm": 1.059159517288208, "learning_rate": 5.3350152748872915e-06, "loss": 0.7205, "step": 18893 }, { "epoch": 0.6657681228362272, "grad_norm": 1.6029247045516968, "learning_rate": 5.3340058208265056e-06, "loss": 0.9587, "step": 18894 }, { "epoch": 0.6658033598491856, "grad_norm": 3.2797110080718994, "learning_rate": 5.332996427540887e-06, "loss": 2.9803, "step": 18895 }, { "epoch": 0.665838596862144, "grad_norm": 6.602704048156738, "learning_rate": 5.331987095043585e-06, "loss": 4.7679, "step": 18896 }, { "epoch": 0.6658738338751025, "grad_norm": 8.11467170715332, "learning_rate": 5.330977823347745e-06, "loss": 3.0584, "step": 18897 }, { "epoch": 0.6659090708880608, "grad_norm": 0.9033846259117126, "learning_rate": 5.329968612466514e-06, "loss": 0.906, "step": 18898 }, { "epoch": 0.6659443079010192, "grad_norm": 5.074606895446777, "learning_rate": 5.328959462413039e-06, "loss": 4.6038, "step": 18899 }, { "epoch": 0.6659795449139776, "grad_norm": 5.425937652587891, "learning_rate": 5.327950373200457e-06, "loss": 4.7598, "step": 18900 }, { "epoch": 0.6660147819269361, "grad_norm": 1.383767008781433, "learning_rate": 5.326941344841918e-06, "loss": 0.7353, "step": 18901 }, { "epoch": 0.6660500189398945, "grad_norm": 3.479497194290161, "learning_rate": 5.325932377350569e-06, "loss": 2.7571, "step": 18902 }, { "epoch": 0.6660852559528528, "grad_norm": 8.96697998046875, "learning_rate": 5.3249234707395405e-06, "loss": 5.3906, "step": 18903 }, { "epoch": 0.6661204929658113, "grad_norm": 3.0333471298217773, "learning_rate": 5.323914625021976e-06, "loss": 2.9106, "step": 18904 }, { "epoch": 0.6661557299787697, "grad_norm": 4.72669792175293, "learning_rate": 5.322905840211023e-06, "loss": 3.3021, "step": 18905 }, { "epoch": 0.6661909669917281, "grad_norm": 1.0516064167022705, "learning_rate": 5.321897116319823e-06, "loss": 1.0425, "step": 18906 }, { "epoch": 0.6662262040046866, "grad_norm": 1.019737720489502, "learning_rate": 5.320888453361502e-06, "loss": 0.9054, "step": 18907 }, { "epoch": 0.666261441017645, "grad_norm": 1.237615704536438, "learning_rate": 5.319879851349203e-06, "loss": 0.8741, "step": 18908 }, { "epoch": 0.6662966780306033, "grad_norm": 2.5624008178710938, "learning_rate": 5.318871310296072e-06, "loss": 2.9775, "step": 18909 }, { "epoch": 0.6663319150435617, "grad_norm": 6.83899450302124, "learning_rate": 5.317862830215235e-06, "loss": 2.839, "step": 18910 }, { "epoch": 0.6663671520565202, "grad_norm": 4.693056106567383, "learning_rate": 5.316854411119832e-06, "loss": 4.4102, "step": 18911 }, { "epoch": 0.6664023890694786, "grad_norm": 3.621943473815918, "learning_rate": 5.315846053022995e-06, "loss": 2.9038, "step": 18912 }, { "epoch": 0.666437626082437, "grad_norm": 1.0623626708984375, "learning_rate": 5.3148377559378605e-06, "loss": 0.7698, "step": 18913 }, { "epoch": 0.6664728630953954, "grad_norm": 7.267268657684326, "learning_rate": 5.313829519877562e-06, "loss": 5.3971, "step": 18914 }, { "epoch": 0.6665081001083538, "grad_norm": 4.1394782066345215, "learning_rate": 5.31282134485523e-06, "loss": 2.9899, "step": 18915 }, { "epoch": 0.6665433371213122, "grad_norm": 3.398965358734131, "learning_rate": 5.311813230883998e-06, "loss": 3.0957, "step": 18916 }, { "epoch": 0.6665785741342707, "grad_norm": 1.3422527313232422, "learning_rate": 5.3108051779769946e-06, "loss": 0.8539, "step": 18917 }, { "epoch": 0.6666138111472291, "grad_norm": 1.234552025794983, "learning_rate": 5.309797186147354e-06, "loss": 0.6448, "step": 18918 }, { "epoch": 0.6666490481601874, "grad_norm": 4.617318630218506, "learning_rate": 5.308789255408197e-06, "loss": 3.4369, "step": 18919 }, { "epoch": 0.6666842851731458, "grad_norm": 1.460208535194397, "learning_rate": 5.30778138577266e-06, "loss": 1.0226, "step": 18920 }, { "epoch": 0.6667195221861043, "grad_norm": 2.9862825870513916, "learning_rate": 5.306773577253872e-06, "loss": 2.5515, "step": 18921 }, { "epoch": 0.6667547591990627, "grad_norm": 1.2812261581420898, "learning_rate": 5.3057658298649525e-06, "loss": 0.9999, "step": 18922 }, { "epoch": 0.6667899962120211, "grad_norm": 4.216817855834961, "learning_rate": 5.304758143619028e-06, "loss": 2.8174, "step": 18923 }, { "epoch": 0.6668252332249796, "grad_norm": 0.9282050132751465, "learning_rate": 5.303750518529229e-06, "loss": 0.8476, "step": 18924 }, { "epoch": 0.6668604702379379, "grad_norm": 3.535186767578125, "learning_rate": 5.302742954608683e-06, "loss": 2.8966, "step": 18925 }, { "epoch": 0.6668957072508963, "grad_norm": 2.480952262878418, "learning_rate": 5.301735451870505e-06, "loss": 2.7795, "step": 18926 }, { "epoch": 0.6669309442638548, "grad_norm": 2.976452112197876, "learning_rate": 5.300728010327816e-06, "loss": 3.7646, "step": 18927 }, { "epoch": 0.6669661812768132, "grad_norm": 1.3665740489959717, "learning_rate": 5.29972062999375e-06, "loss": 0.9876, "step": 18928 }, { "epoch": 0.6670014182897716, "grad_norm": 2.6459243297576904, "learning_rate": 5.298713310881419e-06, "loss": 2.9957, "step": 18929 }, { "epoch": 0.66703665530273, "grad_norm": 2.2106709480285645, "learning_rate": 5.297706053003946e-06, "loss": 1.002, "step": 18930 }, { "epoch": 0.6670718923156884, "grad_norm": 4.078781604766846, "learning_rate": 5.29669885637445e-06, "loss": 4.8739, "step": 18931 }, { "epoch": 0.6671071293286468, "grad_norm": 1.9451881647109985, "learning_rate": 5.295691721006051e-06, "loss": 0.8614, "step": 18932 }, { "epoch": 0.6671423663416052, "grad_norm": 5.0158586502075195, "learning_rate": 5.294684646911867e-06, "loss": 5.1395, "step": 18933 }, { "epoch": 0.6671776033545637, "grad_norm": 5.6605610847473145, "learning_rate": 5.293677634105014e-06, "loss": 2.9118, "step": 18934 }, { "epoch": 0.667212840367522, "grad_norm": 2.2215189933776855, "learning_rate": 5.29267068259861e-06, "loss": 2.7544, "step": 18935 }, { "epoch": 0.6672480773804804, "grad_norm": 1.7292048931121826, "learning_rate": 5.29166379240577e-06, "loss": 0.7207, "step": 18936 }, { "epoch": 0.6672833143934389, "grad_norm": 1.3240145444869995, "learning_rate": 5.290656963539612e-06, "loss": 0.7862, "step": 18937 }, { "epoch": 0.6673185514063973, "grad_norm": 4.058612823486328, "learning_rate": 5.28965019601324e-06, "loss": 3.4129, "step": 18938 }, { "epoch": 0.6673537884193557, "grad_norm": 3.730379819869995, "learning_rate": 5.2886434898397774e-06, "loss": 3.0082, "step": 18939 }, { "epoch": 0.6673890254323142, "grad_norm": 8.333255767822266, "learning_rate": 5.2876368450323335e-06, "loss": 5.1347, "step": 18940 }, { "epoch": 0.6674242624452725, "grad_norm": 1.1780537366867065, "learning_rate": 5.286630261604024e-06, "loss": 0.824, "step": 18941 }, { "epoch": 0.6674594994582309, "grad_norm": 1.3007423877716064, "learning_rate": 5.285623739567949e-06, "loss": 0.7472, "step": 18942 }, { "epoch": 0.6674947364711893, "grad_norm": 3.397892475128174, "learning_rate": 5.284617278937227e-06, "loss": 2.8811, "step": 18943 }, { "epoch": 0.6675299734841478, "grad_norm": 5.3508830070495605, "learning_rate": 5.283610879724972e-06, "loss": 5.1903, "step": 18944 }, { "epoch": 0.6675652104971062, "grad_norm": 1.1808147430419922, "learning_rate": 5.28260454194428e-06, "loss": 1.1093, "step": 18945 }, { "epoch": 0.6676004475100645, "grad_norm": 1.5465677976608276, "learning_rate": 5.281598265608261e-06, "loss": 0.8377, "step": 18946 }, { "epoch": 0.667635684523023, "grad_norm": 1.2142095565795898, "learning_rate": 5.2805920507300345e-06, "loss": 0.9368, "step": 18947 }, { "epoch": 0.6676709215359814, "grad_norm": 5.564097881317139, "learning_rate": 5.279585897322694e-06, "loss": 2.8139, "step": 18948 }, { "epoch": 0.6677061585489398, "grad_norm": 4.493297576904297, "learning_rate": 5.278579805399347e-06, "loss": 2.6443, "step": 18949 }, { "epoch": 0.6677413955618983, "grad_norm": 1.4399853944778442, "learning_rate": 5.277573774973101e-06, "loss": 0.9335, "step": 18950 }, { "epoch": 0.6677766325748566, "grad_norm": 4.359673500061035, "learning_rate": 5.276567806057058e-06, "loss": 2.8168, "step": 18951 }, { "epoch": 0.667811869587815, "grad_norm": 1.129511833190918, "learning_rate": 5.27556189866432e-06, "loss": 0.7956, "step": 18952 }, { "epoch": 0.6678471066007734, "grad_norm": 2.4250385761260986, "learning_rate": 5.274556052807991e-06, "loss": 0.9879, "step": 18953 }, { "epoch": 0.6678823436137319, "grad_norm": 2.978120803833008, "learning_rate": 5.273550268501171e-06, "loss": 2.6339, "step": 18954 }, { "epoch": 0.6679175806266903, "grad_norm": 1.7172528505325317, "learning_rate": 5.272544545756959e-06, "loss": 0.9572, "step": 18955 }, { "epoch": 0.6679528176396486, "grad_norm": 1.0529857873916626, "learning_rate": 5.271538884588463e-06, "loss": 1.1778, "step": 18956 }, { "epoch": 0.6679880546526071, "grad_norm": 4.961987018585205, "learning_rate": 5.270533285008766e-06, "loss": 3.2068, "step": 18957 }, { "epoch": 0.6680232916655655, "grad_norm": 4.934623718261719, "learning_rate": 5.26952774703098e-06, "loss": 2.8618, "step": 18958 }, { "epoch": 0.6680585286785239, "grad_norm": 1.6467735767364502, "learning_rate": 5.268522270668197e-06, "loss": 1.0113, "step": 18959 }, { "epoch": 0.6680937656914824, "grad_norm": 1.200219988822937, "learning_rate": 5.267516855933519e-06, "loss": 0.9787, "step": 18960 }, { "epoch": 0.6681290027044408, "grad_norm": 4.795826435089111, "learning_rate": 5.266511502840028e-06, "loss": 2.8898, "step": 18961 }, { "epoch": 0.6681642397173991, "grad_norm": 0.9288893342018127, "learning_rate": 5.2655062114008325e-06, "loss": 0.7205, "step": 18962 }, { "epoch": 0.6681994767303576, "grad_norm": 1.1444324254989624, "learning_rate": 5.264500981629026e-06, "loss": 0.8554, "step": 18963 }, { "epoch": 0.668234713743316, "grad_norm": 2.6911706924438477, "learning_rate": 5.263495813537693e-06, "loss": 2.8718, "step": 18964 }, { "epoch": 0.6682699507562744, "grad_norm": 2.766951560974121, "learning_rate": 5.262490707139929e-06, "loss": 2.947, "step": 18965 }, { "epoch": 0.6683051877692328, "grad_norm": 3.9881324768066406, "learning_rate": 5.261485662448829e-06, "loss": 4.5264, "step": 18966 }, { "epoch": 0.6683404247821912, "grad_norm": 1.004611849784851, "learning_rate": 5.26048067947748e-06, "loss": 0.6808, "step": 18967 }, { "epoch": 0.6683756617951496, "grad_norm": 3.977670192718506, "learning_rate": 5.259475758238974e-06, "loss": 2.7291, "step": 18968 }, { "epoch": 0.668410898808108, "grad_norm": 1.6586856842041016, "learning_rate": 5.258470898746402e-06, "loss": 0.7792, "step": 18969 }, { "epoch": 0.6684461358210665, "grad_norm": 5.246018886566162, "learning_rate": 5.257466101012849e-06, "loss": 4.6605, "step": 18970 }, { "epoch": 0.6684813728340249, "grad_norm": 1.2127636671066284, "learning_rate": 5.256461365051404e-06, "loss": 0.9165, "step": 18971 }, { "epoch": 0.6685166098469832, "grad_norm": 1.503142237663269, "learning_rate": 5.255456690875155e-06, "loss": 0.9181, "step": 18972 }, { "epoch": 0.6685518468599417, "grad_norm": 5.1060991287231445, "learning_rate": 5.254452078497187e-06, "loss": 3.0162, "step": 18973 }, { "epoch": 0.6685870838729001, "grad_norm": 4.797863960266113, "learning_rate": 5.253447527930584e-06, "loss": 2.5735, "step": 18974 }, { "epoch": 0.6686223208858585, "grad_norm": 2.520989418029785, "learning_rate": 5.252443039188435e-06, "loss": 2.7778, "step": 18975 }, { "epoch": 0.6686575578988169, "grad_norm": 2.496659755706787, "learning_rate": 5.251438612283818e-06, "loss": 0.9316, "step": 18976 }, { "epoch": 0.6686927949117754, "grad_norm": 2.4700074195861816, "learning_rate": 5.2504342472298144e-06, "loss": 0.5948, "step": 18977 }, { "epoch": 0.6687280319247337, "grad_norm": 5.113286018371582, "learning_rate": 5.249429944039511e-06, "loss": 5.4639, "step": 18978 }, { "epoch": 0.6687632689376921, "grad_norm": 4.6545491218566895, "learning_rate": 5.248425702725994e-06, "loss": 5.003, "step": 18979 }, { "epoch": 0.6687985059506506, "grad_norm": 1.656599998474121, "learning_rate": 5.247421523302332e-06, "loss": 1.04, "step": 18980 }, { "epoch": 0.668833742963609, "grad_norm": 1.0689481496810913, "learning_rate": 5.246417405781608e-06, "loss": 1.125, "step": 18981 }, { "epoch": 0.6688689799765674, "grad_norm": 5.53952693939209, "learning_rate": 5.245413350176909e-06, "loss": 5.0215, "step": 18982 }, { "epoch": 0.6689042169895258, "grad_norm": 1.6118911504745483, "learning_rate": 5.2444093565013035e-06, "loss": 0.6349, "step": 18983 }, { "epoch": 0.6689394540024842, "grad_norm": 2.181767225265503, "learning_rate": 5.243405424767871e-06, "loss": 2.8389, "step": 18984 }, { "epoch": 0.6689746910154426, "grad_norm": 3.462181806564331, "learning_rate": 5.242401554989689e-06, "loss": 2.7485, "step": 18985 }, { "epoch": 0.669009928028401, "grad_norm": 5.296095371246338, "learning_rate": 5.2413977471798335e-06, "loss": 2.5816, "step": 18986 }, { "epoch": 0.6690451650413595, "grad_norm": 6.246471405029297, "learning_rate": 5.2403940013513785e-06, "loss": 5.7656, "step": 18987 }, { "epoch": 0.6690804020543178, "grad_norm": 3.2482850551605225, "learning_rate": 5.239390317517398e-06, "loss": 2.5073, "step": 18988 }, { "epoch": 0.6691156390672762, "grad_norm": 4.252011775970459, "learning_rate": 5.238386695690965e-06, "loss": 2.6612, "step": 18989 }, { "epoch": 0.6691508760802347, "grad_norm": 1.1862385272979736, "learning_rate": 5.23738313588515e-06, "loss": 0.8601, "step": 18990 }, { "epoch": 0.6691861130931931, "grad_norm": 3.698136568069458, "learning_rate": 5.236379638113033e-06, "loss": 2.8885, "step": 18991 }, { "epoch": 0.6692213501061515, "grad_norm": 1.1805682182312012, "learning_rate": 5.235376202387669e-06, "loss": 0.9852, "step": 18992 }, { "epoch": 0.66925658711911, "grad_norm": 1.6622889041900635, "learning_rate": 5.234372828722141e-06, "loss": 0.905, "step": 18993 }, { "epoch": 0.6692918241320683, "grad_norm": 1.3168644905090332, "learning_rate": 5.2333695171295164e-06, "loss": 0.9593, "step": 18994 }, { "epoch": 0.6693270611450267, "grad_norm": 1.1071196794509888, "learning_rate": 5.232366267622858e-06, "loss": 0.9301, "step": 18995 }, { "epoch": 0.6693622981579852, "grad_norm": 5.290136337280273, "learning_rate": 5.231363080215232e-06, "loss": 4.7897, "step": 18996 }, { "epoch": 0.6693975351709436, "grad_norm": 6.0597686767578125, "learning_rate": 5.2303599549197144e-06, "loss": 5.4435, "step": 18997 }, { "epoch": 0.669432772183902, "grad_norm": 1.5077431201934814, "learning_rate": 5.22935689174937e-06, "loss": 0.9818, "step": 18998 }, { "epoch": 0.6694680091968603, "grad_norm": 3.757354259490967, "learning_rate": 5.228353890717255e-06, "loss": 2.9405, "step": 18999 }, { "epoch": 0.6695032462098188, "grad_norm": 4.047852993011475, "learning_rate": 5.227350951836434e-06, "loss": 2.9969, "step": 19000 }, { "epoch": 0.6695384832227772, "grad_norm": 1.1573125123977661, "learning_rate": 5.226348075119983e-06, "loss": 0.9521, "step": 19001 }, { "epoch": 0.6695737202357356, "grad_norm": 3.347321033477783, "learning_rate": 5.225345260580953e-06, "loss": 3.0706, "step": 19002 }, { "epoch": 0.6696089572486941, "grad_norm": 4.47609281539917, "learning_rate": 5.2243425082324095e-06, "loss": 3.5691, "step": 19003 }, { "epoch": 0.6696441942616524, "grad_norm": 0.8520947694778442, "learning_rate": 5.223339818087412e-06, "loss": 0.7129, "step": 19004 }, { "epoch": 0.6696794312746108, "grad_norm": 3.236698627471924, "learning_rate": 5.2223371901590215e-06, "loss": 2.9724, "step": 19005 }, { "epoch": 0.6697146682875693, "grad_norm": 2.6132044792175293, "learning_rate": 5.221334624460299e-06, "loss": 2.7259, "step": 19006 }, { "epoch": 0.6697499053005277, "grad_norm": 3.037019729614258, "learning_rate": 5.220332121004301e-06, "loss": 3.5138, "step": 19007 }, { "epoch": 0.6697851423134861, "grad_norm": 1.384002685546875, "learning_rate": 5.219329679804086e-06, "loss": 0.9081, "step": 19008 }, { "epoch": 0.6698203793264444, "grad_norm": 2.1718780994415283, "learning_rate": 5.218327300872712e-06, "loss": 2.5049, "step": 19009 }, { "epoch": 0.6698556163394029, "grad_norm": 4.801389217376709, "learning_rate": 5.217324984223236e-06, "loss": 4.7549, "step": 19010 }, { "epoch": 0.6698908533523613, "grad_norm": 3.526102066040039, "learning_rate": 5.216322729868706e-06, "loss": 2.6924, "step": 19011 }, { "epoch": 0.6699260903653197, "grad_norm": 1.2851296663284302, "learning_rate": 5.215320537822184e-06, "loss": 0.7489, "step": 19012 }, { "epoch": 0.6699613273782782, "grad_norm": 7.849478244781494, "learning_rate": 5.214318408096721e-06, "loss": 4.7562, "step": 19013 }, { "epoch": 0.6699965643912366, "grad_norm": 0.9651573896408081, "learning_rate": 5.213316340705374e-06, "loss": 0.805, "step": 19014 }, { "epoch": 0.6700318014041949, "grad_norm": 1.2290760278701782, "learning_rate": 5.212314335661185e-06, "loss": 1.0678, "step": 19015 }, { "epoch": 0.6700670384171534, "grad_norm": 2.8783059120178223, "learning_rate": 5.211312392977215e-06, "loss": 2.9089, "step": 19016 }, { "epoch": 0.6701022754301118, "grad_norm": 3.700528144836426, "learning_rate": 5.210310512666514e-06, "loss": 2.8686, "step": 19017 }, { "epoch": 0.6701375124430702, "grad_norm": 8.112039566040039, "learning_rate": 5.209308694742126e-06, "loss": 3.3605, "step": 19018 }, { "epoch": 0.6701727494560286, "grad_norm": 1.1439622640609741, "learning_rate": 5.208306939217098e-06, "loss": 0.6912, "step": 19019 }, { "epoch": 0.670207986468987, "grad_norm": 2.5237412452697754, "learning_rate": 5.2073052461044915e-06, "loss": 2.673, "step": 19020 }, { "epoch": 0.6702432234819454, "grad_norm": 5.091150760650635, "learning_rate": 5.20630361541734e-06, "loss": 5.3146, "step": 19021 }, { "epoch": 0.6702784604949038, "grad_norm": 0.9555259943008423, "learning_rate": 5.2053020471686945e-06, "loss": 0.8227, "step": 19022 }, { "epoch": 0.6703136975078623, "grad_norm": 1.6407054662704468, "learning_rate": 5.2043005413716e-06, "loss": 1.0066, "step": 19023 }, { "epoch": 0.6703489345208207, "grad_norm": 1.4990450143814087, "learning_rate": 5.203299098039102e-06, "loss": 1.0187, "step": 19024 }, { "epoch": 0.670384171533779, "grad_norm": 7.330928802490234, "learning_rate": 5.202297717184245e-06, "loss": 4.6539, "step": 19025 }, { "epoch": 0.6704194085467375, "grad_norm": 3.4882924556732178, "learning_rate": 5.2012963988200705e-06, "loss": 3.4583, "step": 19026 }, { "epoch": 0.6704546455596959, "grad_norm": 3.2806851863861084, "learning_rate": 5.200295142959622e-06, "loss": 3.0484, "step": 19027 }, { "epoch": 0.6704898825726543, "grad_norm": 1.6595056056976318, "learning_rate": 5.19929394961594e-06, "loss": 0.7471, "step": 19028 }, { "epoch": 0.6705251195856127, "grad_norm": 1.179232120513916, "learning_rate": 5.198292818802071e-06, "loss": 1.0766, "step": 19029 }, { "epoch": 0.6705603565985712, "grad_norm": 1.3290425539016724, "learning_rate": 5.1972917505310415e-06, "loss": 0.836, "step": 19030 }, { "epoch": 0.6705955936115295, "grad_norm": 3.1496312618255615, "learning_rate": 5.196290744815902e-06, "loss": 3.0882, "step": 19031 }, { "epoch": 0.6706308306244879, "grad_norm": 2.9826226234436035, "learning_rate": 5.195289801669687e-06, "loss": 3.2159, "step": 19032 }, { "epoch": 0.6706660676374464, "grad_norm": 5.621828079223633, "learning_rate": 5.194288921105441e-06, "loss": 5.4428, "step": 19033 }, { "epoch": 0.6707013046504048, "grad_norm": 3.6858999729156494, "learning_rate": 5.193288103136182e-06, "loss": 2.8918, "step": 19034 }, { "epoch": 0.6707365416633632, "grad_norm": 4.07338809967041, "learning_rate": 5.192287347774965e-06, "loss": 3.6106, "step": 19035 }, { "epoch": 0.6707717786763217, "grad_norm": 0.8966887593269348, "learning_rate": 5.191286655034821e-06, "loss": 0.724, "step": 19036 }, { "epoch": 0.67080701568928, "grad_norm": 1.5511387586593628, "learning_rate": 5.190286024928777e-06, "loss": 0.9565, "step": 19037 }, { "epoch": 0.6708422527022384, "grad_norm": 5.296377182006836, "learning_rate": 5.189285457469865e-06, "loss": 5.1531, "step": 19038 }, { "epoch": 0.6708774897151969, "grad_norm": 2.983905792236328, "learning_rate": 5.188284952671131e-06, "loss": 2.5779, "step": 19039 }, { "epoch": 0.6709127267281553, "grad_norm": 7.468128204345703, "learning_rate": 5.187284510545595e-06, "loss": 3.1772, "step": 19040 }, { "epoch": 0.6709479637411137, "grad_norm": 3.6217615604400635, "learning_rate": 5.186284131106291e-06, "loss": 2.1874, "step": 19041 }, { "epoch": 0.670983200754072, "grad_norm": 1.3289811611175537, "learning_rate": 5.185283814366249e-06, "loss": 0.7423, "step": 19042 }, { "epoch": 0.6710184377670305, "grad_norm": 6.093389511108398, "learning_rate": 5.184283560338499e-06, "loss": 6.6088, "step": 19043 }, { "epoch": 0.6710536747799889, "grad_norm": 1.290578842163086, "learning_rate": 5.183283369036069e-06, "loss": 0.7422, "step": 19044 }, { "epoch": 0.6710889117929473, "grad_norm": 2.9027223587036133, "learning_rate": 5.182283240471987e-06, "loss": 2.7685, "step": 19045 }, { "epoch": 0.6711241488059058, "grad_norm": 2.8817906379699707, "learning_rate": 5.18128317465928e-06, "loss": 2.9342, "step": 19046 }, { "epoch": 0.6711593858188641, "grad_norm": 1.1171174049377441, "learning_rate": 5.180283171610971e-06, "loss": 0.893, "step": 19047 }, { "epoch": 0.6711946228318225, "grad_norm": 0.7673472762107849, "learning_rate": 5.1792832313400934e-06, "loss": 0.6672, "step": 19048 }, { "epoch": 0.671229859844781, "grad_norm": 1.215921401977539, "learning_rate": 5.178283353859659e-06, "loss": 0.9276, "step": 19049 }, { "epoch": 0.6712650968577394, "grad_norm": 4.4722490310668945, "learning_rate": 5.1772835391827e-06, "loss": 3.1943, "step": 19050 }, { "epoch": 0.6713003338706978, "grad_norm": 1.1237177848815918, "learning_rate": 5.176283787322238e-06, "loss": 0.8611, "step": 19051 }, { "epoch": 0.6713355708836561, "grad_norm": 1.6159605979919434, "learning_rate": 5.175284098291298e-06, "loss": 0.683, "step": 19052 }, { "epoch": 0.6713708078966146, "grad_norm": 3.393998861312866, "learning_rate": 5.174284472102888e-06, "loss": 2.7739, "step": 19053 }, { "epoch": 0.671406044909573, "grad_norm": 42.52260971069336, "learning_rate": 5.173284908770044e-06, "loss": 4.3439, "step": 19054 }, { "epoch": 0.6714412819225314, "grad_norm": 2.5460257530212402, "learning_rate": 5.1722854083057815e-06, "loss": 2.6532, "step": 19055 }, { "epoch": 0.6714765189354899, "grad_norm": 3.945446729660034, "learning_rate": 5.1712859707231125e-06, "loss": 5.279, "step": 19056 }, { "epoch": 0.6715117559484483, "grad_norm": 3.676481008529663, "learning_rate": 5.170286596035055e-06, "loss": 2.6825, "step": 19057 }, { "epoch": 0.6715469929614066, "grad_norm": 2.8010077476501465, "learning_rate": 5.169287284254638e-06, "loss": 3.0741, "step": 19058 }, { "epoch": 0.6715822299743651, "grad_norm": 5.525374889373779, "learning_rate": 5.168288035394865e-06, "loss": 5.362, "step": 19059 }, { "epoch": 0.6716174669873235, "grad_norm": 5.733251571655273, "learning_rate": 5.167288849468756e-06, "loss": 2.9414, "step": 19060 }, { "epoch": 0.6716527040002819, "grad_norm": 2.507009506225586, "learning_rate": 5.166289726489325e-06, "loss": 2.7484, "step": 19061 }, { "epoch": 0.6716879410132403, "grad_norm": 1.389795184135437, "learning_rate": 5.165290666469587e-06, "loss": 0.8943, "step": 19062 }, { "epoch": 0.6717231780261987, "grad_norm": 3.8578593730926514, "learning_rate": 5.164291669422552e-06, "loss": 3.0086, "step": 19063 }, { "epoch": 0.6717584150391571, "grad_norm": 1.6434859037399292, "learning_rate": 5.1632927353612365e-06, "loss": 0.7737, "step": 19064 }, { "epoch": 0.6717936520521155, "grad_norm": 3.3602592945098877, "learning_rate": 5.162293864298647e-06, "loss": 3.3315, "step": 19065 }, { "epoch": 0.671828889065074, "grad_norm": 3.5551598072052, "learning_rate": 5.161295056247797e-06, "loss": 3.772, "step": 19066 }, { "epoch": 0.6718641260780324, "grad_norm": 1.4611786603927612, "learning_rate": 5.1602963112217e-06, "loss": 1.0062, "step": 19067 }, { "epoch": 0.6718993630909907, "grad_norm": 2.0487747192382812, "learning_rate": 5.1592976292333504e-06, "loss": 0.9479, "step": 19068 }, { "epoch": 0.6719346001039492, "grad_norm": 1.1149455308914185, "learning_rate": 5.158299010295771e-06, "loss": 0.732, "step": 19069 }, { "epoch": 0.6719698371169076, "grad_norm": 4.332531929016113, "learning_rate": 5.157300454421964e-06, "loss": 2.7849, "step": 19070 }, { "epoch": 0.672005074129866, "grad_norm": 1.2902194261550903, "learning_rate": 5.156301961624939e-06, "loss": 0.9166, "step": 19071 }, { "epoch": 0.6720403111428245, "grad_norm": 3.681518077850342, "learning_rate": 5.1553035319176905e-06, "loss": 2.3838, "step": 19072 }, { "epoch": 0.6720755481557829, "grad_norm": 1.3279328346252441, "learning_rate": 5.1543051653132336e-06, "loss": 0.8003, "step": 19073 }, { "epoch": 0.6721107851687412, "grad_norm": 1.6210874319076538, "learning_rate": 5.153306861824573e-06, "loss": 0.9261, "step": 19074 }, { "epoch": 0.6721460221816996, "grad_norm": 1.5509657859802246, "learning_rate": 5.152308621464705e-06, "loss": 0.7934, "step": 19075 }, { "epoch": 0.6721812591946581, "grad_norm": 0.9985475540161133, "learning_rate": 5.151310444246629e-06, "loss": 1.1431, "step": 19076 }, { "epoch": 0.6722164962076165, "grad_norm": 1.4753037691116333, "learning_rate": 5.150312330183362e-06, "loss": 0.8222, "step": 19077 }, { "epoch": 0.6722517332205749, "grad_norm": 7.37712287902832, "learning_rate": 5.149314279287889e-06, "loss": 4.6572, "step": 19078 }, { "epoch": 0.6722869702335333, "grad_norm": 1.239885687828064, "learning_rate": 5.148316291573216e-06, "loss": 0.9712, "step": 19079 }, { "epoch": 0.6723222072464917, "grad_norm": 1.3358601331710815, "learning_rate": 5.147318367052343e-06, "loss": 0.6303, "step": 19080 }, { "epoch": 0.6723574442594501, "grad_norm": 11.840633392333984, "learning_rate": 5.146320505738265e-06, "loss": 3.2182, "step": 19081 }, { "epoch": 0.6723926812724086, "grad_norm": 2.9341635704040527, "learning_rate": 5.1453227076439805e-06, "loss": 2.9823, "step": 19082 }, { "epoch": 0.672427918285367, "grad_norm": 5.6906609535217285, "learning_rate": 5.144324972782486e-06, "loss": 2.5244, "step": 19083 }, { "epoch": 0.6724631552983253, "grad_norm": 1.185687780380249, "learning_rate": 5.143327301166778e-06, "loss": 0.8568, "step": 19084 }, { "epoch": 0.6724983923112837, "grad_norm": 2.5823891162872314, "learning_rate": 5.142329692809849e-06, "loss": 0.8052, "step": 19085 }, { "epoch": 0.6725336293242422, "grad_norm": 4.891453742980957, "learning_rate": 5.141332147724696e-06, "loss": 3.096, "step": 19086 }, { "epoch": 0.6725688663372006, "grad_norm": 1.3709514141082764, "learning_rate": 5.14033466592431e-06, "loss": 0.8373, "step": 19087 }, { "epoch": 0.672604103350159, "grad_norm": 2.680751085281372, "learning_rate": 5.139337247421684e-06, "loss": 3.0272, "step": 19088 }, { "epoch": 0.6726393403631175, "grad_norm": 1.4058141708374023, "learning_rate": 5.13833989222981e-06, "loss": 0.9308, "step": 19089 }, { "epoch": 0.6726745773760758, "grad_norm": 3.2779366970062256, "learning_rate": 5.137342600361681e-06, "loss": 0.56, "step": 19090 }, { "epoch": 0.6727098143890342, "grad_norm": 1.8225021362304688, "learning_rate": 5.1363453718302784e-06, "loss": 0.7349, "step": 19091 }, { "epoch": 0.6727450514019927, "grad_norm": 1.3702208995819092, "learning_rate": 5.135348206648598e-06, "loss": 1.008, "step": 19092 }, { "epoch": 0.6727802884149511, "grad_norm": 4.432314872741699, "learning_rate": 5.134351104829633e-06, "loss": 2.9579, "step": 19093 }, { "epoch": 0.6728155254279095, "grad_norm": 1.7470157146453857, "learning_rate": 5.13335406638636e-06, "loss": 0.8691, "step": 19094 }, { "epoch": 0.6728507624408678, "grad_norm": 4.709120273590088, "learning_rate": 5.1323570913317655e-06, "loss": 3.2727, "step": 19095 }, { "epoch": 0.6728859994538263, "grad_norm": 2.3688981533050537, "learning_rate": 5.13136017967885e-06, "loss": 2.6847, "step": 19096 }, { "epoch": 0.6729212364667847, "grad_norm": 5.8892292976379395, "learning_rate": 5.130363331440582e-06, "loss": 3.045, "step": 19097 }, { "epoch": 0.6729564734797431, "grad_norm": 0.9185275435447693, "learning_rate": 5.129366546629953e-06, "loss": 1.2273, "step": 19098 }, { "epoch": 0.6729917104927016, "grad_norm": 16.532833099365234, "learning_rate": 5.128369825259944e-06, "loss": 3.136, "step": 19099 }, { "epoch": 0.67302694750566, "grad_norm": 9.305211067199707, "learning_rate": 5.1273731673435405e-06, "loss": 2.3507, "step": 19100 }, { "epoch": 0.6730621845186183, "grad_norm": 1.287016749382019, "learning_rate": 5.126376572893721e-06, "loss": 0.814, "step": 19101 }, { "epoch": 0.6730974215315768, "grad_norm": 2.0456550121307373, "learning_rate": 5.125380041923467e-06, "loss": 2.2909, "step": 19102 }, { "epoch": 0.6731326585445352, "grad_norm": 2.6760339736938477, "learning_rate": 5.124383574445759e-06, "loss": 2.2199, "step": 19103 }, { "epoch": 0.6731678955574936, "grad_norm": 1.2181932926177979, "learning_rate": 5.123387170473577e-06, "loss": 1.0428, "step": 19104 }, { "epoch": 0.6732031325704521, "grad_norm": 3.1823878288269043, "learning_rate": 5.122390830019899e-06, "loss": 3.0968, "step": 19105 }, { "epoch": 0.6732383695834104, "grad_norm": 1.7920981645584106, "learning_rate": 5.121394553097701e-06, "loss": 1.0232, "step": 19106 }, { "epoch": 0.6732736065963688, "grad_norm": 1.2404327392578125, "learning_rate": 5.120398339719959e-06, "loss": 0.8909, "step": 19107 }, { "epoch": 0.6733088436093272, "grad_norm": 1.0672253370285034, "learning_rate": 5.119402189899651e-06, "loss": 0.7839, "step": 19108 }, { "epoch": 0.6733440806222857, "grad_norm": 3.060070753097534, "learning_rate": 5.118406103649755e-06, "loss": 1.9738, "step": 19109 }, { "epoch": 0.6733793176352441, "grad_norm": 2.948484182357788, "learning_rate": 5.117410080983234e-06, "loss": 3.4854, "step": 19110 }, { "epoch": 0.6734145546482024, "grad_norm": 1.4125458002090454, "learning_rate": 5.116414121913072e-06, "loss": 0.9491, "step": 19111 }, { "epoch": 0.6734497916611609, "grad_norm": 3.5079195499420166, "learning_rate": 5.115418226452242e-06, "loss": 3.0443, "step": 19112 }, { "epoch": 0.6734850286741193, "grad_norm": 2.4729583263397217, "learning_rate": 5.114422394613709e-06, "loss": 2.8403, "step": 19113 }, { "epoch": 0.6735202656870777, "grad_norm": 2.8666269779205322, "learning_rate": 5.113426626410441e-06, "loss": 2.5705, "step": 19114 }, { "epoch": 0.6735555027000362, "grad_norm": 1.1469992399215698, "learning_rate": 5.112430921855422e-06, "loss": 0.8812, "step": 19115 }, { "epoch": 0.6735907397129945, "grad_norm": 1.4217609167099, "learning_rate": 5.111435280961609e-06, "loss": 0.9864, "step": 19116 }, { "epoch": 0.6736259767259529, "grad_norm": 4.420795917510986, "learning_rate": 5.110439703741975e-06, "loss": 3.3676, "step": 19117 }, { "epoch": 0.6736612137389113, "grad_norm": 6.200117588043213, "learning_rate": 5.109444190209486e-06, "loss": 7.6725, "step": 19118 }, { "epoch": 0.6736964507518698, "grad_norm": 3.8452184200286865, "learning_rate": 5.1084487403771095e-06, "loss": 3.1206, "step": 19119 }, { "epoch": 0.6737316877648282, "grad_norm": 3.3110949993133545, "learning_rate": 5.107453354257812e-06, "loss": 2.8167, "step": 19120 }, { "epoch": 0.6737669247777865, "grad_norm": 1.193610668182373, "learning_rate": 5.106458031864556e-06, "loss": 1.1024, "step": 19121 }, { "epoch": 0.673802161790745, "grad_norm": 1.0936940908432007, "learning_rate": 5.105462773210308e-06, "loss": 1.4391, "step": 19122 }, { "epoch": 0.6738373988037034, "grad_norm": 1.2896918058395386, "learning_rate": 5.104467578308032e-06, "loss": 0.7609, "step": 19123 }, { "epoch": 0.6738726358166618, "grad_norm": 5.29108190536499, "learning_rate": 5.103472447170689e-06, "loss": 5.2673, "step": 19124 }, { "epoch": 0.6739078728296203, "grad_norm": 1.6903117895126343, "learning_rate": 5.10247737981124e-06, "loss": 0.7355, "step": 19125 }, { "epoch": 0.6739431098425787, "grad_norm": 1.2808634042739868, "learning_rate": 5.101482376242648e-06, "loss": 1.0651, "step": 19126 }, { "epoch": 0.673978346855537, "grad_norm": 5.889373779296875, "learning_rate": 5.1004874364778726e-06, "loss": 2.7317, "step": 19127 }, { "epoch": 0.6740135838684954, "grad_norm": 3.880878210067749, "learning_rate": 5.099492560529875e-06, "loss": 2.8042, "step": 19128 }, { "epoch": 0.6740488208814539, "grad_norm": 3.3649559020996094, "learning_rate": 5.098497748411604e-06, "loss": 2.7357, "step": 19129 }, { "epoch": 0.6740840578944123, "grad_norm": 2.7887744903564453, "learning_rate": 5.097503000136027e-06, "loss": 2.5608, "step": 19130 }, { "epoch": 0.6741192949073707, "grad_norm": 4.339052677154541, "learning_rate": 5.096508315716102e-06, "loss": 3.2176, "step": 19131 }, { "epoch": 0.6741545319203291, "grad_norm": 3.5228893756866455, "learning_rate": 5.0955136951647775e-06, "loss": 2.9184, "step": 19132 }, { "epoch": 0.6741897689332875, "grad_norm": 1.161185383796692, "learning_rate": 5.0945191384950084e-06, "loss": 0.7662, "step": 19133 }, { "epoch": 0.6742250059462459, "grad_norm": 1.1444188356399536, "learning_rate": 5.09352464571976e-06, "loss": 1.1505, "step": 19134 }, { "epoch": 0.6742602429592044, "grad_norm": 1.252763032913208, "learning_rate": 5.092530216851974e-06, "loss": 0.9027, "step": 19135 }, { "epoch": 0.6742954799721628, "grad_norm": 3.701115369796753, "learning_rate": 5.0915358519046075e-06, "loss": 3.0652, "step": 19136 }, { "epoch": 0.6743307169851211, "grad_norm": 4.183012008666992, "learning_rate": 5.090541550890611e-06, "loss": 3.2676, "step": 19137 }, { "epoch": 0.6743659539980796, "grad_norm": 3.0388426780700684, "learning_rate": 5.089547313822936e-06, "loss": 2.4115, "step": 19138 }, { "epoch": 0.674401191011038, "grad_norm": 1.4153045415878296, "learning_rate": 5.088553140714534e-06, "loss": 0.8565, "step": 19139 }, { "epoch": 0.6744364280239964, "grad_norm": 1.980677604675293, "learning_rate": 5.087559031578351e-06, "loss": 0.8392, "step": 19140 }, { "epoch": 0.6744716650369548, "grad_norm": 1.7548232078552246, "learning_rate": 5.086564986427338e-06, "loss": 0.8196, "step": 19141 }, { "epoch": 0.6745069020499133, "grad_norm": 4.180817604064941, "learning_rate": 5.0855710052744436e-06, "loss": 2.8718, "step": 19142 }, { "epoch": 0.6745421390628716, "grad_norm": 3.017764091491699, "learning_rate": 5.084577088132612e-06, "loss": 2.7865, "step": 19143 }, { "epoch": 0.67457737607583, "grad_norm": 5.283233642578125, "learning_rate": 5.0835832350147885e-06, "loss": 3.1097, "step": 19144 }, { "epoch": 0.6746126130887885, "grad_norm": 2.8441965579986572, "learning_rate": 5.082589445933921e-06, "loss": 2.5382, "step": 19145 }, { "epoch": 0.6746478501017469, "grad_norm": 2.4573843479156494, "learning_rate": 5.081595720902952e-06, "loss": 2.7985, "step": 19146 }, { "epoch": 0.6746830871147053, "grad_norm": 3.8771164417266846, "learning_rate": 5.08060205993483e-06, "loss": 3.337, "step": 19147 }, { "epoch": 0.6747183241276637, "grad_norm": 2.281754493713379, "learning_rate": 5.0796084630424824e-06, "loss": 0.8193, "step": 19148 }, { "epoch": 0.6747535611406221, "grad_norm": 1.2309962511062622, "learning_rate": 5.078614930238867e-06, "loss": 1.2167, "step": 19149 }, { "epoch": 0.6747887981535805, "grad_norm": 1.126521110534668, "learning_rate": 5.077621461536923e-06, "loss": 0.7895, "step": 19150 }, { "epoch": 0.6748240351665389, "grad_norm": 4.455150604248047, "learning_rate": 5.076628056949583e-06, "loss": 3.3859, "step": 19151 }, { "epoch": 0.6748592721794974, "grad_norm": 1.4476661682128906, "learning_rate": 5.075634716489785e-06, "loss": 0.558, "step": 19152 }, { "epoch": 0.6748945091924557, "grad_norm": 5.705733776092529, "learning_rate": 5.07464144017048e-06, "loss": 5.1427, "step": 19153 }, { "epoch": 0.6749297462054141, "grad_norm": 4.704233169555664, "learning_rate": 5.073648228004595e-06, "loss": 4.8317, "step": 19154 }, { "epoch": 0.6749649832183726, "grad_norm": 1.1768559217453003, "learning_rate": 5.072655080005068e-06, "loss": 0.6696, "step": 19155 }, { "epoch": 0.675000220231331, "grad_norm": 2.8004038333892822, "learning_rate": 5.071661996184836e-06, "loss": 3.0069, "step": 19156 }, { "epoch": 0.6750354572442894, "grad_norm": 0.8986152410507202, "learning_rate": 5.070668976556835e-06, "loss": 1.1125, "step": 19157 }, { "epoch": 0.6750706942572479, "grad_norm": 2.131566047668457, "learning_rate": 5.069676021133998e-06, "loss": 0.9129, "step": 19158 }, { "epoch": 0.6751059312702062, "grad_norm": 3.3689730167388916, "learning_rate": 5.068683129929259e-06, "loss": 2.8322, "step": 19159 }, { "epoch": 0.6751411682831646, "grad_norm": 6.4129533767700195, "learning_rate": 5.06769030295555e-06, "loss": 5.1084, "step": 19160 }, { "epoch": 0.675176405296123, "grad_norm": 1.0676286220550537, "learning_rate": 5.066697540225805e-06, "loss": 1.0479, "step": 19161 }, { "epoch": 0.6752116423090815, "grad_norm": 3.986271381378174, "learning_rate": 5.065704841752951e-06, "loss": 3.2574, "step": 19162 }, { "epoch": 0.6752468793220399, "grad_norm": 2.5431456565856934, "learning_rate": 5.06471220754992e-06, "loss": 2.7134, "step": 19163 }, { "epoch": 0.6752821163349982, "grad_norm": 2.5232326984405518, "learning_rate": 5.063719637629641e-06, "loss": 2.681, "step": 19164 }, { "epoch": 0.6753173533479567, "grad_norm": 1.1237293481826782, "learning_rate": 5.062727132005043e-06, "loss": 0.9151, "step": 19165 }, { "epoch": 0.6753525903609151, "grad_norm": 3.493398904800415, "learning_rate": 5.0617346906890575e-06, "loss": 3.0523, "step": 19166 }, { "epoch": 0.6753878273738735, "grad_norm": 3.282905340194702, "learning_rate": 5.0607423136945985e-06, "loss": 2.8263, "step": 19167 }, { "epoch": 0.675423064386832, "grad_norm": 2.71618390083313, "learning_rate": 5.059750001034604e-06, "loss": 2.8016, "step": 19168 }, { "epoch": 0.6754583013997904, "grad_norm": 3.2378578186035156, "learning_rate": 5.058757752721999e-06, "loss": 2.9861, "step": 19169 }, { "epoch": 0.6754935384127487, "grad_norm": 1.1968797445297241, "learning_rate": 5.057765568769699e-06, "loss": 1.0956, "step": 19170 }, { "epoch": 0.6755287754257072, "grad_norm": 1.3218625783920288, "learning_rate": 5.056773449190628e-06, "loss": 0.7849, "step": 19171 }, { "epoch": 0.6755640124386656, "grad_norm": 0.9215607643127441, "learning_rate": 5.0557813939977215e-06, "loss": 0.7172, "step": 19172 }, { "epoch": 0.675599249451624, "grad_norm": 3.9882419109344482, "learning_rate": 5.054789403203888e-06, "loss": 3.1319, "step": 19173 }, { "epoch": 0.6756344864645824, "grad_norm": 5.403225421905518, "learning_rate": 5.053797476822052e-06, "loss": 4.9797, "step": 19174 }, { "epoch": 0.6756697234775408, "grad_norm": 2.5755093097686768, "learning_rate": 5.052805614865134e-06, "loss": 2.831, "step": 19175 }, { "epoch": 0.6757049604904992, "grad_norm": 5.18494176864624, "learning_rate": 5.051813817346053e-06, "loss": 3.3034, "step": 19176 }, { "epoch": 0.6757401975034576, "grad_norm": 5.555219650268555, "learning_rate": 5.050822084277727e-06, "loss": 2.8214, "step": 19177 }, { "epoch": 0.6757754345164161, "grad_norm": 2.0785610675811768, "learning_rate": 5.049830415673075e-06, "loss": 1.0412, "step": 19178 }, { "epoch": 0.6758106715293745, "grad_norm": 7.523789882659912, "learning_rate": 5.048838811545012e-06, "loss": 3.3216, "step": 19179 }, { "epoch": 0.6758459085423328, "grad_norm": 2.884078025817871, "learning_rate": 5.047847271906453e-06, "loss": 2.9968, "step": 19180 }, { "epoch": 0.6758811455552913, "grad_norm": 2.8427274227142334, "learning_rate": 5.046855796770315e-06, "loss": 2.878, "step": 19181 }, { "epoch": 0.6759163825682497, "grad_norm": 3.79590106010437, "learning_rate": 5.04586438614951e-06, "loss": 3.0996, "step": 19182 }, { "epoch": 0.6759516195812081, "grad_norm": 5.774974346160889, "learning_rate": 5.044873040056955e-06, "loss": 5.579, "step": 19183 }, { "epoch": 0.6759868565941665, "grad_norm": 1.2927675247192383, "learning_rate": 5.043881758505558e-06, "loss": 0.7637, "step": 19184 }, { "epoch": 0.676022093607125, "grad_norm": 1.3738203048706055, "learning_rate": 5.042890541508237e-06, "loss": 0.6631, "step": 19185 }, { "epoch": 0.6760573306200833, "grad_norm": 3.982374429702759, "learning_rate": 5.041899389077889e-06, "loss": 1.1523, "step": 19186 }, { "epoch": 0.6760925676330417, "grad_norm": 4.900463104248047, "learning_rate": 5.0409083012274375e-06, "loss": 4.6773, "step": 19187 }, { "epoch": 0.6761278046460002, "grad_norm": 4.991797924041748, "learning_rate": 5.0399172779697914e-06, "loss": 2.8691, "step": 19188 }, { "epoch": 0.6761630416589586, "grad_norm": 3.7602694034576416, "learning_rate": 5.0389263193178496e-06, "loss": 3.0974, "step": 19189 }, { "epoch": 0.676198278671917, "grad_norm": 1.455460548400879, "learning_rate": 5.0379354252845215e-06, "loss": 0.7737, "step": 19190 }, { "epoch": 0.6762335156848754, "grad_norm": 4.615467071533203, "learning_rate": 5.0369445958827245e-06, "loss": 3.0856, "step": 19191 }, { "epoch": 0.6762687526978338, "grad_norm": 28.846555709838867, "learning_rate": 5.035953831125353e-06, "loss": 4.8329, "step": 19192 }, { "epoch": 0.6763039897107922, "grad_norm": 4.550044059753418, "learning_rate": 5.034963131025315e-06, "loss": 2.732, "step": 19193 }, { "epoch": 0.6763392267237506, "grad_norm": 0.9597529172897339, "learning_rate": 5.0339724955955136e-06, "loss": 0.8657, "step": 19194 }, { "epoch": 0.6763744637367091, "grad_norm": 2.9458911418914795, "learning_rate": 5.032981924848852e-06, "loss": 0.7024, "step": 19195 }, { "epoch": 0.6764097007496674, "grad_norm": 4.670785427093506, "learning_rate": 5.0319914187982365e-06, "loss": 3.5233, "step": 19196 }, { "epoch": 0.6764449377626258, "grad_norm": 1.0987097024917603, "learning_rate": 5.031000977456564e-06, "loss": 1.091, "step": 19197 }, { "epoch": 0.6764801747755843, "grad_norm": 3.991403102874756, "learning_rate": 5.030010600836737e-06, "loss": 2.6898, "step": 19198 }, { "epoch": 0.6765154117885427, "grad_norm": 1.0609270334243774, "learning_rate": 5.029020288951654e-06, "loss": 0.7573, "step": 19199 }, { "epoch": 0.6765506488015011, "grad_norm": 1.3696101903915405, "learning_rate": 5.0280300418142155e-06, "loss": 0.745, "step": 19200 }, { "epoch": 0.6765858858144596, "grad_norm": 1.737633228302002, "learning_rate": 5.027039859437319e-06, "loss": 0.8462, "step": 19201 }, { "epoch": 0.6766211228274179, "grad_norm": 3.023381471633911, "learning_rate": 5.026049741833863e-06, "loss": 2.7572, "step": 19202 }, { "epoch": 0.6766563598403763, "grad_norm": 3.2255115509033203, "learning_rate": 5.025059689016741e-06, "loss": 2.7683, "step": 19203 }, { "epoch": 0.6766915968533348, "grad_norm": 1.240193247795105, "learning_rate": 5.024069700998855e-06, "loss": 0.7271, "step": 19204 }, { "epoch": 0.6767268338662932, "grad_norm": 1.3169784545898438, "learning_rate": 5.023079777793087e-06, "loss": 0.787, "step": 19205 }, { "epoch": 0.6767620708792516, "grad_norm": 1.490073800086975, "learning_rate": 5.0220899194123425e-06, "loss": 0.9349, "step": 19206 }, { "epoch": 0.6767973078922099, "grad_norm": 1.6537809371948242, "learning_rate": 5.021100125869516e-06, "loss": 1.1948, "step": 19207 }, { "epoch": 0.6768325449051684, "grad_norm": 3.6491379737854004, "learning_rate": 5.020110397177489e-06, "loss": 3.2966, "step": 19208 }, { "epoch": 0.6768677819181268, "grad_norm": 2.245835781097412, "learning_rate": 5.019120733349157e-06, "loss": 0.8436, "step": 19209 }, { "epoch": 0.6769030189310852, "grad_norm": 0.9863584637641907, "learning_rate": 5.018131134397417e-06, "loss": 1.2272, "step": 19210 }, { "epoch": 0.6769382559440437, "grad_norm": 6.909917831420898, "learning_rate": 5.017141600335151e-06, "loss": 3.6074, "step": 19211 }, { "epoch": 0.676973492957002, "grad_norm": 1.460394263267517, "learning_rate": 5.016152131175249e-06, "loss": 0.6513, "step": 19212 }, { "epoch": 0.6770087299699604, "grad_norm": 3.051163673400879, "learning_rate": 5.0151627269306024e-06, "loss": 3.0775, "step": 19213 }, { "epoch": 0.6770439669829189, "grad_norm": 5.8630266189575195, "learning_rate": 5.014173387614095e-06, "loss": 5.1707, "step": 19214 }, { "epoch": 0.6770792039958773, "grad_norm": 3.453422784805298, "learning_rate": 5.013184113238614e-06, "loss": 2.6531, "step": 19215 }, { "epoch": 0.6771144410088357, "grad_norm": 8.008441925048828, "learning_rate": 5.012194903817047e-06, "loss": 5.2844, "step": 19216 }, { "epoch": 0.677149678021794, "grad_norm": 1.0495765209197998, "learning_rate": 5.011205759362274e-06, "loss": 0.8952, "step": 19217 }, { "epoch": 0.6771849150347525, "grad_norm": 2.050640106201172, "learning_rate": 5.010216679887183e-06, "loss": 0.7786, "step": 19218 }, { "epoch": 0.6772201520477109, "grad_norm": 1.867234230041504, "learning_rate": 5.009227665404656e-06, "loss": 0.9094, "step": 19219 }, { "epoch": 0.6772553890606693, "grad_norm": 5.942520618438721, "learning_rate": 5.008238715927574e-06, "loss": 3.0993, "step": 19220 }, { "epoch": 0.6772906260736278, "grad_norm": 1.3636493682861328, "learning_rate": 5.007249831468818e-06, "loss": 0.862, "step": 19221 }, { "epoch": 0.6773258630865862, "grad_norm": 3.6337685585021973, "learning_rate": 5.006261012041269e-06, "loss": 3.127, "step": 19222 }, { "epoch": 0.6773611000995445, "grad_norm": 5.286961555480957, "learning_rate": 5.00527225765781e-06, "loss": 4.6508, "step": 19223 }, { "epoch": 0.677396337112503, "grad_norm": 1.2462917566299438, "learning_rate": 5.0042835683313095e-06, "loss": 0.9698, "step": 19224 }, { "epoch": 0.6774315741254614, "grad_norm": 1.3671302795410156, "learning_rate": 5.003294944074654e-06, "loss": 0.6366, "step": 19225 }, { "epoch": 0.6774668111384198, "grad_norm": 3.8307809829711914, "learning_rate": 5.002306384900724e-06, "loss": 3.2957, "step": 19226 }, { "epoch": 0.6775020481513782, "grad_norm": 3.50020432472229, "learning_rate": 5.0013178908223845e-06, "loss": 2.746, "step": 19227 }, { "epoch": 0.6775372851643366, "grad_norm": 4.563049793243408, "learning_rate": 5.0003294618525135e-06, "loss": 3.1304, "step": 19228 }, { "epoch": 0.677572522177295, "grad_norm": 1.1194887161254883, "learning_rate": 4.999341098003996e-06, "loss": 0.9564, "step": 19229 }, { "epoch": 0.6776077591902534, "grad_norm": 5.140832424163818, "learning_rate": 4.998352799289693e-06, "loss": 5.1758, "step": 19230 }, { "epoch": 0.6776429962032119, "grad_norm": 2.0679197311401367, "learning_rate": 4.997364565722481e-06, "loss": 2.6195, "step": 19231 }, { "epoch": 0.6776782332161703, "grad_norm": 3.7182247638702393, "learning_rate": 4.99637639731523e-06, "loss": 3.1126, "step": 19232 }, { "epoch": 0.6777134702291286, "grad_norm": 2.5339853763580322, "learning_rate": 4.995388294080822e-06, "loss": 0.9701, "step": 19233 }, { "epoch": 0.6777487072420871, "grad_norm": 4.709712982177734, "learning_rate": 4.9944002560321145e-06, "loss": 4.6263, "step": 19234 }, { "epoch": 0.6777839442550455, "grad_norm": 5.586520195007324, "learning_rate": 4.993412283181982e-06, "loss": 3.0299, "step": 19235 }, { "epoch": 0.6778191812680039, "grad_norm": 4.261791229248047, "learning_rate": 4.992424375543291e-06, "loss": 4.7132, "step": 19236 }, { "epoch": 0.6778544182809623, "grad_norm": 3.251877784729004, "learning_rate": 4.991436533128911e-06, "loss": 2.6182, "step": 19237 }, { "epoch": 0.6778896552939208, "grad_norm": 5.38463830947876, "learning_rate": 4.990448755951709e-06, "loss": 3.0856, "step": 19238 }, { "epoch": 0.6779248923068791, "grad_norm": 6.216385841369629, "learning_rate": 4.9894610440245495e-06, "loss": 3.0831, "step": 19239 }, { "epoch": 0.6779601293198375, "grad_norm": 5.0008864402771, "learning_rate": 4.988473397360297e-06, "loss": 3.1486, "step": 19240 }, { "epoch": 0.677995366332796, "grad_norm": 4.821496486663818, "learning_rate": 4.987485815971819e-06, "loss": 4.9431, "step": 19241 }, { "epoch": 0.6780306033457544, "grad_norm": 6.0665059089660645, "learning_rate": 4.986498299871978e-06, "loss": 4.7925, "step": 19242 }, { "epoch": 0.6780658403587128, "grad_norm": 1.0080771446228027, "learning_rate": 4.985510849073629e-06, "loss": 0.7215, "step": 19243 }, { "epoch": 0.6781010773716712, "grad_norm": 1.2710808515548706, "learning_rate": 4.984523463589645e-06, "loss": 0.7845, "step": 19244 }, { "epoch": 0.6781363143846296, "grad_norm": 2.4405062198638916, "learning_rate": 4.983536143432883e-06, "loss": 2.9969, "step": 19245 }, { "epoch": 0.678171551397588, "grad_norm": 7.359620094299316, "learning_rate": 4.9825488886162e-06, "loss": 3.0275, "step": 19246 }, { "epoch": 0.6782067884105465, "grad_norm": 3.2001757621765137, "learning_rate": 4.9815616991524515e-06, "loss": 3.2695, "step": 19247 }, { "epoch": 0.6782420254235049, "grad_norm": 4.902711391448975, "learning_rate": 4.98057457505451e-06, "loss": 3.1403, "step": 19248 }, { "epoch": 0.6782772624364632, "grad_norm": 3.4136648178100586, "learning_rate": 4.979587516335219e-06, "loss": 3.097, "step": 19249 }, { "epoch": 0.6783124994494216, "grad_norm": 4.2645487785339355, "learning_rate": 4.978600523007441e-06, "loss": 2.9522, "step": 19250 }, { "epoch": 0.6783477364623801, "grad_norm": 1.091688871383667, "learning_rate": 4.977613595084026e-06, "loss": 0.759, "step": 19251 }, { "epoch": 0.6783829734753385, "grad_norm": 7.61771821975708, "learning_rate": 4.976626732577842e-06, "loss": 4.9541, "step": 19252 }, { "epoch": 0.6784182104882969, "grad_norm": 1.3071330785751343, "learning_rate": 4.975639935501732e-06, "loss": 1.1752, "step": 19253 }, { "epoch": 0.6784534475012554, "grad_norm": 2.7004075050354004, "learning_rate": 4.9746532038685516e-06, "loss": 3.0594, "step": 19254 }, { "epoch": 0.6784886845142137, "grad_norm": 4.867295265197754, "learning_rate": 4.973666537691152e-06, "loss": 4.4967, "step": 19255 }, { "epoch": 0.6785239215271721, "grad_norm": 2.258794069290161, "learning_rate": 4.972679936982386e-06, "loss": 0.9992, "step": 19256 }, { "epoch": 0.6785591585401306, "grad_norm": 3.295600175857544, "learning_rate": 4.971693401755106e-06, "loss": 2.9746, "step": 19257 }, { "epoch": 0.678594395553089, "grad_norm": 1.6837477684020996, "learning_rate": 4.970706932022158e-06, "loss": 1.0717, "step": 19258 }, { "epoch": 0.6786296325660474, "grad_norm": 6.586562633514404, "learning_rate": 4.9697205277963936e-06, "loss": 6.7217, "step": 19259 }, { "epoch": 0.6786648695790057, "grad_norm": 4.569746971130371, "learning_rate": 4.968734189090659e-06, "loss": 2.8378, "step": 19260 }, { "epoch": 0.6787001065919642, "grad_norm": 3.4519968032836914, "learning_rate": 4.967747915917808e-06, "loss": 2.6443, "step": 19261 }, { "epoch": 0.6787353436049226, "grad_norm": 3.8648197650909424, "learning_rate": 4.966761708290673e-06, "loss": 4.7885, "step": 19262 }, { "epoch": 0.678770580617881, "grad_norm": 9.014274597167969, "learning_rate": 4.965775566222112e-06, "loss": 4.6052, "step": 19263 }, { "epoch": 0.6788058176308395, "grad_norm": 0.8588084578514099, "learning_rate": 4.964789489724969e-06, "loss": 0.8523, "step": 19264 }, { "epoch": 0.6788410546437978, "grad_norm": 4.808574199676514, "learning_rate": 4.96380347881208e-06, "loss": 5.4808, "step": 19265 }, { "epoch": 0.6788762916567562, "grad_norm": 3.615708351135254, "learning_rate": 4.9628175334962894e-06, "loss": 3.627, "step": 19266 }, { "epoch": 0.6789115286697147, "grad_norm": 1.5544687509536743, "learning_rate": 4.961831653790449e-06, "loss": 0.8986, "step": 19267 }, { "epoch": 0.6789467656826731, "grad_norm": 3.734158992767334, "learning_rate": 4.96084583970739e-06, "loss": 4.3377, "step": 19268 }, { "epoch": 0.6789820026956315, "grad_norm": 3.8076744079589844, "learning_rate": 4.959860091259955e-06, "loss": 2.9012, "step": 19269 }, { "epoch": 0.6790172397085898, "grad_norm": 1.4220693111419678, "learning_rate": 4.958874408460984e-06, "loss": 0.8209, "step": 19270 }, { "epoch": 0.6790524767215483, "grad_norm": 1.2585821151733398, "learning_rate": 4.957888791323318e-06, "loss": 0.8036, "step": 19271 }, { "epoch": 0.6790877137345067, "grad_norm": 2.989532709121704, "learning_rate": 4.9569032398597905e-06, "loss": 2.7762, "step": 19272 }, { "epoch": 0.6791229507474651, "grad_norm": 7.700921058654785, "learning_rate": 4.955917754083241e-06, "loss": 3.1033, "step": 19273 }, { "epoch": 0.6791581877604236, "grad_norm": 1.1723418235778809, "learning_rate": 4.954932334006506e-06, "loss": 0.9525, "step": 19274 }, { "epoch": 0.679193424773382, "grad_norm": 1.3350623846054077, "learning_rate": 4.95394697964242e-06, "loss": 1.0232, "step": 19275 }, { "epoch": 0.6792286617863403, "grad_norm": 1.4417647123336792, "learning_rate": 4.952961691003817e-06, "loss": 0.828, "step": 19276 }, { "epoch": 0.6792638987992988, "grad_norm": 4.6341376304626465, "learning_rate": 4.951976468103531e-06, "loss": 2.9537, "step": 19277 }, { "epoch": 0.6792991358122572, "grad_norm": 2.357841968536377, "learning_rate": 4.950991310954395e-06, "loss": 0.7289, "step": 19278 }, { "epoch": 0.6793343728252156, "grad_norm": 3.624908447265625, "learning_rate": 4.95000621956924e-06, "loss": 3.2111, "step": 19279 }, { "epoch": 0.6793696098381741, "grad_norm": 1.1118806600570679, "learning_rate": 4.949021193960901e-06, "loss": 0.9124, "step": 19280 }, { "epoch": 0.6794048468511324, "grad_norm": 1.1886403560638428, "learning_rate": 4.948036234142196e-06, "loss": 1.0109, "step": 19281 }, { "epoch": 0.6794400838640908, "grad_norm": 1.2726603746414185, "learning_rate": 4.947051340125967e-06, "loss": 0.7879, "step": 19282 }, { "epoch": 0.6794753208770492, "grad_norm": 5.125973701477051, "learning_rate": 4.9460665119250414e-06, "loss": 4.9043, "step": 19283 }, { "epoch": 0.6795105578900077, "grad_norm": 2.0007851123809814, "learning_rate": 4.945081749552241e-06, "loss": 2.6143, "step": 19284 }, { "epoch": 0.6795457949029661, "grad_norm": 2.301750898361206, "learning_rate": 4.944097053020392e-06, "loss": 0.8907, "step": 19285 }, { "epoch": 0.6795810319159244, "grad_norm": 5.635221004486084, "learning_rate": 4.9431124223423255e-06, "loss": 5.1361, "step": 19286 }, { "epoch": 0.6796162689288829, "grad_norm": 3.73633074760437, "learning_rate": 4.942127857530863e-06, "loss": 4.2602, "step": 19287 }, { "epoch": 0.6796515059418413, "grad_norm": 2.3714852333068848, "learning_rate": 4.9411433585988285e-06, "loss": 0.7476, "step": 19288 }, { "epoch": 0.6796867429547997, "grad_norm": 3.9653100967407227, "learning_rate": 4.940158925559047e-06, "loss": 2.5733, "step": 19289 }, { "epoch": 0.6797219799677582, "grad_norm": 3.9392030239105225, "learning_rate": 4.93917455842434e-06, "loss": 3.2022, "step": 19290 }, { "epoch": 0.6797572169807166, "grad_norm": 4.711658000946045, "learning_rate": 4.938190257207529e-06, "loss": 5.172, "step": 19291 }, { "epoch": 0.6797924539936749, "grad_norm": 1.564184546470642, "learning_rate": 4.937206021921434e-06, "loss": 0.6722, "step": 19292 }, { "epoch": 0.6798276910066333, "grad_norm": 2.620242118835449, "learning_rate": 4.936221852578876e-06, "loss": 2.9766, "step": 19293 }, { "epoch": 0.6798629280195918, "grad_norm": 3.239072322845459, "learning_rate": 4.935237749192674e-06, "loss": 2.8643, "step": 19294 }, { "epoch": 0.6798981650325502, "grad_norm": 1.1198680400848389, "learning_rate": 4.934253711775645e-06, "loss": 0.6962, "step": 19295 }, { "epoch": 0.6799334020455086, "grad_norm": 1.3440784215927124, "learning_rate": 4.93326974034061e-06, "loss": 0.9125, "step": 19296 }, { "epoch": 0.679968639058467, "grad_norm": 2.064249277114868, "learning_rate": 4.932285834900374e-06, "loss": 0.8406, "step": 19297 }, { "epoch": 0.6800038760714254, "grad_norm": 1.3459173440933228, "learning_rate": 4.931301995467766e-06, "loss": 0.7905, "step": 19298 }, { "epoch": 0.6800391130843838, "grad_norm": 0.8184226751327515, "learning_rate": 4.930318222055598e-06, "loss": 0.7217, "step": 19299 }, { "epoch": 0.6800743500973423, "grad_norm": 1.1548351049423218, "learning_rate": 4.929334514676677e-06, "loss": 1.0369, "step": 19300 }, { "epoch": 0.6801095871103007, "grad_norm": 5.4724955558776855, "learning_rate": 4.928350873343817e-06, "loss": 3.4327, "step": 19301 }, { "epoch": 0.680144824123259, "grad_norm": 4.888408184051514, "learning_rate": 4.92736729806984e-06, "loss": 3.2023, "step": 19302 }, { "epoch": 0.6801800611362174, "grad_norm": 3.0744285583496094, "learning_rate": 4.926383788867545e-06, "loss": 2.7182, "step": 19303 }, { "epoch": 0.6802152981491759, "grad_norm": 1.589653491973877, "learning_rate": 4.925400345749749e-06, "loss": 0.7986, "step": 19304 }, { "epoch": 0.6802505351621343, "grad_norm": 4.084537982940674, "learning_rate": 4.924416968729255e-06, "loss": 2.5292, "step": 19305 }, { "epoch": 0.6802857721750927, "grad_norm": 5.372291088104248, "learning_rate": 4.923433657818885e-06, "loss": 4.9258, "step": 19306 }, { "epoch": 0.6803210091880512, "grad_norm": 4.908165454864502, "learning_rate": 4.9224504130314346e-06, "loss": 4.8224, "step": 19307 }, { "epoch": 0.6803562462010095, "grad_norm": 4.654219627380371, "learning_rate": 4.921467234379714e-06, "loss": 3.2803, "step": 19308 }, { "epoch": 0.6803914832139679, "grad_norm": 3.8316879272460938, "learning_rate": 4.920484121876528e-06, "loss": 2.0411, "step": 19309 }, { "epoch": 0.6804267202269264, "grad_norm": 6.611757278442383, "learning_rate": 4.919501075534685e-06, "loss": 4.9705, "step": 19310 }, { "epoch": 0.6804619572398848, "grad_norm": 4.955198764801025, "learning_rate": 4.918518095366988e-06, "loss": 2.912, "step": 19311 }, { "epoch": 0.6804971942528432, "grad_norm": 5.160293102264404, "learning_rate": 4.917535181386239e-06, "loss": 4.9145, "step": 19312 }, { "epoch": 0.6805324312658017, "grad_norm": 3.6661055088043213, "learning_rate": 4.916552333605243e-06, "loss": 4.1755, "step": 19313 }, { "epoch": 0.68056766827876, "grad_norm": 1.633583903312683, "learning_rate": 4.915569552036798e-06, "loss": 0.7976, "step": 19314 }, { "epoch": 0.6806029052917184, "grad_norm": 4.29604959487915, "learning_rate": 4.914586836693714e-06, "loss": 3.1121, "step": 19315 }, { "epoch": 0.6806381423046768, "grad_norm": 4.892409324645996, "learning_rate": 4.913604187588775e-06, "loss": 5.0484, "step": 19316 }, { "epoch": 0.6806733793176353, "grad_norm": 40.82551956176758, "learning_rate": 4.912621604734793e-06, "loss": 3.3791, "step": 19317 }, { "epoch": 0.6807086163305937, "grad_norm": 1.5610994100570679, "learning_rate": 4.9116390881445665e-06, "loss": 1.1151, "step": 19318 }, { "epoch": 0.680743853343552, "grad_norm": 5.614017486572266, "learning_rate": 4.910656637830886e-06, "loss": 3.1345, "step": 19319 }, { "epoch": 0.6807790903565105, "grad_norm": 2.315887689590454, "learning_rate": 4.909674253806547e-06, "loss": 0.8862, "step": 19320 }, { "epoch": 0.6808143273694689, "grad_norm": 7.094054222106934, "learning_rate": 4.9086919360843565e-06, "loss": 4.8137, "step": 19321 }, { "epoch": 0.6808495643824273, "grad_norm": 11.532041549682617, "learning_rate": 4.907709684677099e-06, "loss": 2.356, "step": 19322 }, { "epoch": 0.6808848013953858, "grad_norm": 1.7051265239715576, "learning_rate": 4.90672749959757e-06, "loss": 0.9458, "step": 19323 }, { "epoch": 0.6809200384083441, "grad_norm": 4.206980228424072, "learning_rate": 4.905745380858562e-06, "loss": 5.7741, "step": 19324 }, { "epoch": 0.6809552754213025, "grad_norm": 5.30345344543457, "learning_rate": 4.904763328472876e-06, "loss": 3.5307, "step": 19325 }, { "epoch": 0.6809905124342609, "grad_norm": 2.974141836166382, "learning_rate": 4.903781342453293e-06, "loss": 3.1751, "step": 19326 }, { "epoch": 0.6810257494472194, "grad_norm": 1.2688570022583008, "learning_rate": 4.902799422812607e-06, "loss": 1.0421, "step": 19327 }, { "epoch": 0.6810609864601778, "grad_norm": 1.0533965826034546, "learning_rate": 4.901817569563608e-06, "loss": 1.0018, "step": 19328 }, { "epoch": 0.6810962234731361, "grad_norm": 1.0031999349594116, "learning_rate": 4.9008357827190845e-06, "loss": 1.0512, "step": 19329 }, { "epoch": 0.6811314604860946, "grad_norm": 5.9405317306518555, "learning_rate": 4.899854062291825e-06, "loss": 3.1617, "step": 19330 }, { "epoch": 0.681166697499053, "grad_norm": 4.142174243927002, "learning_rate": 4.898872408294615e-06, "loss": 3.2834, "step": 19331 }, { "epoch": 0.6812019345120114, "grad_norm": 1.1141839027404785, "learning_rate": 4.897890820740242e-06, "loss": 0.6635, "step": 19332 }, { "epoch": 0.6812371715249699, "grad_norm": 3.0991103649139404, "learning_rate": 4.8969092996414915e-06, "loss": 2.6454, "step": 19333 }, { "epoch": 0.6812724085379283, "grad_norm": 3.613295555114746, "learning_rate": 4.8959278450111504e-06, "loss": 3.2226, "step": 19334 }, { "epoch": 0.6813076455508866, "grad_norm": 1.4438295364379883, "learning_rate": 4.894946456861992e-06, "loss": 0.7812, "step": 19335 }, { "epoch": 0.681342882563845, "grad_norm": 3.8571434020996094, "learning_rate": 4.893965135206811e-06, "loss": 3.4507, "step": 19336 }, { "epoch": 0.6813781195768035, "grad_norm": 6.579952239990234, "learning_rate": 4.892983880058387e-06, "loss": 2.4214, "step": 19337 }, { "epoch": 0.6814133565897619, "grad_norm": 1.0337141752243042, "learning_rate": 4.892002691429495e-06, "loss": 0.8374, "step": 19338 }, { "epoch": 0.6814485936027203, "grad_norm": 6.7646965980529785, "learning_rate": 4.891021569332915e-06, "loss": 5.0342, "step": 19339 }, { "epoch": 0.6814838306156787, "grad_norm": 0.9259514212608337, "learning_rate": 4.8900405137814375e-06, "loss": 0.8517, "step": 19340 }, { "epoch": 0.6815190676286371, "grad_norm": 4.652165412902832, "learning_rate": 4.889059524787829e-06, "loss": 4.221, "step": 19341 }, { "epoch": 0.6815543046415955, "grad_norm": 2.4160022735595703, "learning_rate": 4.888078602364872e-06, "loss": 1.0175, "step": 19342 }, { "epoch": 0.681589541654554, "grad_norm": 3.407453775405884, "learning_rate": 4.887097746525338e-06, "loss": 3.1701, "step": 19343 }, { "epoch": 0.6816247786675124, "grad_norm": 3.5902082920074463, "learning_rate": 4.8861169572820155e-06, "loss": 2.6162, "step": 19344 }, { "epoch": 0.6816600156804707, "grad_norm": 1.3432010412216187, "learning_rate": 4.885136234647666e-06, "loss": 0.7863, "step": 19345 }, { "epoch": 0.6816952526934292, "grad_norm": 0.9612700343132019, "learning_rate": 4.884155578635069e-06, "loss": 0.7814, "step": 19346 }, { "epoch": 0.6817304897063876, "grad_norm": 1.2592116594314575, "learning_rate": 4.883174989256998e-06, "loss": 0.9381, "step": 19347 }, { "epoch": 0.681765726719346, "grad_norm": 5.252457141876221, "learning_rate": 4.882194466526223e-06, "loss": 2.98, "step": 19348 }, { "epoch": 0.6818009637323044, "grad_norm": 3.8438587188720703, "learning_rate": 4.881214010455518e-06, "loss": 3.1668, "step": 19349 }, { "epoch": 0.6818362007452629, "grad_norm": 1.0503143072128296, "learning_rate": 4.880233621057653e-06, "loss": 0.8648, "step": 19350 }, { "epoch": 0.6818714377582212, "grad_norm": 4.970189094543457, "learning_rate": 4.879253298345397e-06, "loss": 3.2613, "step": 19351 }, { "epoch": 0.6819066747711796, "grad_norm": 8.355098724365234, "learning_rate": 4.878273042331518e-06, "loss": 3.214, "step": 19352 }, { "epoch": 0.6819419117841381, "grad_norm": 2.4886960983276367, "learning_rate": 4.87729285302879e-06, "loss": 2.5539, "step": 19353 }, { "epoch": 0.6819771487970965, "grad_norm": 4.948042392730713, "learning_rate": 4.876312730449967e-06, "loss": 3.0441, "step": 19354 }, { "epoch": 0.6820123858100549, "grad_norm": 3.093569755554199, "learning_rate": 4.875332674607828e-06, "loss": 2.3822, "step": 19355 }, { "epoch": 0.6820476228230133, "grad_norm": 4.785622596740723, "learning_rate": 4.874352685515137e-06, "loss": 5.104, "step": 19356 }, { "epoch": 0.6820828598359717, "grad_norm": 5.702849388122559, "learning_rate": 4.873372763184651e-06, "loss": 3.3795, "step": 19357 }, { "epoch": 0.6821180968489301, "grad_norm": 3.7030160427093506, "learning_rate": 4.872392907629133e-06, "loss": 2.9229, "step": 19358 }, { "epoch": 0.6821533338618885, "grad_norm": 3.7281386852264404, "learning_rate": 4.8714131188613554e-06, "loss": 3.4571, "step": 19359 }, { "epoch": 0.682188570874847, "grad_norm": 1.7801393270492554, "learning_rate": 4.87043339689408e-06, "loss": 0.8972, "step": 19360 }, { "epoch": 0.6822238078878053, "grad_norm": 2.625772476196289, "learning_rate": 4.869453741740058e-06, "loss": 2.5736, "step": 19361 }, { "epoch": 0.6822590449007637, "grad_norm": 1.5344432592391968, "learning_rate": 4.868474153412051e-06, "loss": 1.0169, "step": 19362 }, { "epoch": 0.6822942819137222, "grad_norm": 3.8132729530334473, "learning_rate": 4.867494631922829e-06, "loss": 3.1093, "step": 19363 }, { "epoch": 0.6823295189266806, "grad_norm": 1.760623574256897, "learning_rate": 4.86651517728514e-06, "loss": 0.9733, "step": 19364 }, { "epoch": 0.682364755939639, "grad_norm": 1.263737678527832, "learning_rate": 4.865535789511745e-06, "loss": 1.2669, "step": 19365 }, { "epoch": 0.6823999929525975, "grad_norm": 2.5188794136047363, "learning_rate": 4.8645564686154e-06, "loss": 3.3481, "step": 19366 }, { "epoch": 0.6824352299655558, "grad_norm": 3.523566484451294, "learning_rate": 4.863577214608861e-06, "loss": 2.6477, "step": 19367 }, { "epoch": 0.6824704669785142, "grad_norm": 1.1959787607192993, "learning_rate": 4.862598027504882e-06, "loss": 0.869, "step": 19368 }, { "epoch": 0.6825057039914726, "grad_norm": 3.6060116291046143, "learning_rate": 4.8616189073162205e-06, "loss": 0.9897, "step": 19369 }, { "epoch": 0.6825409410044311, "grad_norm": 3.1308634281158447, "learning_rate": 4.860639854055625e-06, "loss": 2.0616, "step": 19370 }, { "epoch": 0.6825761780173895, "grad_norm": 3.5307812690734863, "learning_rate": 4.859660867735851e-06, "loss": 2.0937, "step": 19371 }, { "epoch": 0.6826114150303478, "grad_norm": 1.4042692184448242, "learning_rate": 4.8586819483696536e-06, "loss": 1.0459, "step": 19372 }, { "epoch": 0.6826466520433063, "grad_norm": 1.6299067735671997, "learning_rate": 4.8577030959697706e-06, "loss": 0.9382, "step": 19373 }, { "epoch": 0.6826818890562647, "grad_norm": 3.7059900760650635, "learning_rate": 4.856724310548963e-06, "loss": 3.3778, "step": 19374 }, { "epoch": 0.6827171260692231, "grad_norm": 1.4814902544021606, "learning_rate": 4.855745592119981e-06, "loss": 0.7146, "step": 19375 }, { "epoch": 0.6827523630821816, "grad_norm": 1.6472903490066528, "learning_rate": 4.8547669406955636e-06, "loss": 0.7511, "step": 19376 }, { "epoch": 0.68278760009514, "grad_norm": 1.1396688222885132, "learning_rate": 4.853788356288458e-06, "loss": 0.9085, "step": 19377 }, { "epoch": 0.6828228371080983, "grad_norm": 5.015659809112549, "learning_rate": 4.852809838911418e-06, "loss": 3.4789, "step": 19378 }, { "epoch": 0.6828580741210568, "grad_norm": 3.437800884246826, "learning_rate": 4.851831388577191e-06, "loss": 2.938, "step": 19379 }, { "epoch": 0.6828933111340152, "grad_norm": 5.228105068206787, "learning_rate": 4.850853005298511e-06, "loss": 3.2842, "step": 19380 }, { "epoch": 0.6829285481469736, "grad_norm": 4.30763053894043, "learning_rate": 4.8498746890881235e-06, "loss": 3.3354, "step": 19381 }, { "epoch": 0.682963785159932, "grad_norm": 1.5130103826522827, "learning_rate": 4.84889643995878e-06, "loss": 1.0325, "step": 19382 }, { "epoch": 0.6829990221728904, "grad_norm": 1.209252119064331, "learning_rate": 4.8479182579232145e-06, "loss": 0.6299, "step": 19383 }, { "epoch": 0.6830342591858488, "grad_norm": 5.969370365142822, "learning_rate": 4.84694014299417e-06, "loss": 2.7862, "step": 19384 }, { "epoch": 0.6830694961988072, "grad_norm": 1.0283921957015991, "learning_rate": 4.845962095184386e-06, "loss": 0.8843, "step": 19385 }, { "epoch": 0.6831047332117657, "grad_norm": 1.0198274850845337, "learning_rate": 4.844984114506601e-06, "loss": 0.9336, "step": 19386 }, { "epoch": 0.6831399702247241, "grad_norm": 5.769680023193359, "learning_rate": 4.844006200973557e-06, "loss": 3.2695, "step": 19387 }, { "epoch": 0.6831752072376824, "grad_norm": 3.2828292846679688, "learning_rate": 4.843028354597988e-06, "loss": 2.985, "step": 19388 }, { "epoch": 0.6832104442506409, "grad_norm": 4.883892059326172, "learning_rate": 4.842050575392631e-06, "loss": 2.7954, "step": 19389 }, { "epoch": 0.6832456812635993, "grad_norm": 0.9964795112609863, "learning_rate": 4.841072863370222e-06, "loss": 1.0971, "step": 19390 }, { "epoch": 0.6832809182765577, "grad_norm": 11.69233512878418, "learning_rate": 4.8400952185435e-06, "loss": 5.7795, "step": 19391 }, { "epoch": 0.6833161552895161, "grad_norm": 3.2117111682891846, "learning_rate": 4.839117640925188e-06, "loss": 2.7303, "step": 19392 }, { "epoch": 0.6833513923024745, "grad_norm": 0.9835559129714966, "learning_rate": 4.838140130528031e-06, "loss": 0.7247, "step": 19393 }, { "epoch": 0.6833866293154329, "grad_norm": 5.0061259269714355, "learning_rate": 4.8371626873647585e-06, "loss": 3.1483, "step": 19394 }, { "epoch": 0.6834218663283913, "grad_norm": 2.9927830696105957, "learning_rate": 4.836185311448096e-06, "loss": 3.0132, "step": 19395 }, { "epoch": 0.6834571033413498, "grad_norm": 8.905858993530273, "learning_rate": 4.835208002790775e-06, "loss": 3.4758, "step": 19396 }, { "epoch": 0.6834923403543082, "grad_norm": 1.4606075286865234, "learning_rate": 4.834230761405529e-06, "loss": 0.9268, "step": 19397 }, { "epoch": 0.6835275773672665, "grad_norm": 1.10585618019104, "learning_rate": 4.833253587305091e-06, "loss": 0.7092, "step": 19398 }, { "epoch": 0.683562814380225, "grad_norm": 1.0094457864761353, "learning_rate": 4.832276480502179e-06, "loss": 0.727, "step": 19399 }, { "epoch": 0.6835980513931834, "grad_norm": 1.8899427652359009, "learning_rate": 4.8312994410095204e-06, "loss": 0.7982, "step": 19400 }, { "epoch": 0.6836332884061418, "grad_norm": 2.9100348949432373, "learning_rate": 4.830322468839853e-06, "loss": 3.2092, "step": 19401 }, { "epoch": 0.6836685254191002, "grad_norm": 1.0166929960250854, "learning_rate": 4.82934556400589e-06, "loss": 1.0372, "step": 19402 }, { "epoch": 0.6837037624320587, "grad_norm": 1.1354467868804932, "learning_rate": 4.828368726520358e-06, "loss": 1.1751, "step": 19403 }, { "epoch": 0.683738999445017, "grad_norm": 3.4411659240722656, "learning_rate": 4.827391956395985e-06, "loss": 3.1789, "step": 19404 }, { "epoch": 0.6837742364579754, "grad_norm": 5.779587268829346, "learning_rate": 4.826415253645489e-06, "loss": 4.6602, "step": 19405 }, { "epoch": 0.6838094734709339, "grad_norm": 0.8801896572113037, "learning_rate": 4.825438618281592e-06, "loss": 1.0494, "step": 19406 }, { "epoch": 0.6838447104838923, "grad_norm": 1.749258041381836, "learning_rate": 4.8244620503170185e-06, "loss": 0.8611, "step": 19407 }, { "epoch": 0.6838799474968507, "grad_norm": 1.341711401939392, "learning_rate": 4.823485549764485e-06, "loss": 1.0336, "step": 19408 }, { "epoch": 0.6839151845098091, "grad_norm": 2.0047404766082764, "learning_rate": 4.8225091166367126e-06, "loss": 1.0793, "step": 19409 }, { "epoch": 0.6839504215227675, "grad_norm": 2.121750831604004, "learning_rate": 4.82153275094642e-06, "loss": 0.9335, "step": 19410 }, { "epoch": 0.6839856585357259, "grad_norm": 1.4510245323181152, "learning_rate": 4.820556452706318e-06, "loss": 1.1066, "step": 19411 }, { "epoch": 0.6840208955486844, "grad_norm": 3.8100154399871826, "learning_rate": 4.81958022192913e-06, "loss": 3.2736, "step": 19412 }, { "epoch": 0.6840561325616428, "grad_norm": 2.7959351539611816, "learning_rate": 4.818604058627573e-06, "loss": 2.7174, "step": 19413 }, { "epoch": 0.6840913695746011, "grad_norm": 14.372147560119629, "learning_rate": 4.817627962814355e-06, "loss": 2.5916, "step": 19414 }, { "epoch": 0.6841266065875595, "grad_norm": 16.078380584716797, "learning_rate": 4.81665193450219e-06, "loss": 3.1354, "step": 19415 }, { "epoch": 0.684161843600518, "grad_norm": 6.871374130249023, "learning_rate": 4.815675973703796e-06, "loss": 3.5352, "step": 19416 }, { "epoch": 0.6841970806134764, "grad_norm": 3.623445749282837, "learning_rate": 4.814700080431888e-06, "loss": 3.2142, "step": 19417 }, { "epoch": 0.6842323176264348, "grad_norm": 4.127107620239258, "learning_rate": 4.813724254699168e-06, "loss": 3.2911, "step": 19418 }, { "epoch": 0.6842675546393933, "grad_norm": 5.529921054840088, "learning_rate": 4.8127484965183455e-06, "loss": 4.9665, "step": 19419 }, { "epoch": 0.6843027916523516, "grad_norm": 1.3559033870697021, "learning_rate": 4.811772805902143e-06, "loss": 0.9434, "step": 19420 }, { "epoch": 0.68433802866531, "grad_norm": 1.1644660234451294, "learning_rate": 4.810797182863256e-06, "loss": 0.8572, "step": 19421 }, { "epoch": 0.6843732656782685, "grad_norm": 3.240579605102539, "learning_rate": 4.809821627414397e-06, "loss": 3.2294, "step": 19422 }, { "epoch": 0.6844085026912269, "grad_norm": 0.9516647458076477, "learning_rate": 4.808846139568273e-06, "loss": 0.7277, "step": 19423 }, { "epoch": 0.6844437397041853, "grad_norm": 1.1679999828338623, "learning_rate": 4.807870719337589e-06, "loss": 0.7884, "step": 19424 }, { "epoch": 0.6844789767171436, "grad_norm": 3.839263439178467, "learning_rate": 4.80689536673505e-06, "loss": 2.918, "step": 19425 }, { "epoch": 0.6845142137301021, "grad_norm": 3.5734140872955322, "learning_rate": 4.805920081773361e-06, "loss": 2.729, "step": 19426 }, { "epoch": 0.6845494507430605, "grad_norm": 1.0655548572540283, "learning_rate": 4.804944864465224e-06, "loss": 0.7136, "step": 19427 }, { "epoch": 0.6845846877560189, "grad_norm": 6.090733528137207, "learning_rate": 4.803969714823341e-06, "loss": 5.3495, "step": 19428 }, { "epoch": 0.6846199247689774, "grad_norm": 7.299376487731934, "learning_rate": 4.802994632860419e-06, "loss": 5.4034, "step": 19429 }, { "epoch": 0.6846551617819358, "grad_norm": 3.940394163131714, "learning_rate": 4.802019618589147e-06, "loss": 3.0365, "step": 19430 }, { "epoch": 0.6846903987948941, "grad_norm": 1.944834589958191, "learning_rate": 4.801044672022235e-06, "loss": 0.8545, "step": 19431 }, { "epoch": 0.6847256358078526, "grad_norm": 2.903582811355591, "learning_rate": 4.800069793172377e-06, "loss": 2.1159, "step": 19432 }, { "epoch": 0.684760872820811, "grad_norm": 3.2543649673461914, "learning_rate": 4.799094982052278e-06, "loss": 3.2384, "step": 19433 }, { "epoch": 0.6847961098337694, "grad_norm": 2.809741258621216, "learning_rate": 4.798120238674621e-06, "loss": 2.4235, "step": 19434 }, { "epoch": 0.6848313468467278, "grad_norm": 0.9350866675376892, "learning_rate": 4.797145563052115e-06, "loss": 0.9838, "step": 19435 }, { "epoch": 0.6848665838596862, "grad_norm": 5.022607803344727, "learning_rate": 4.796170955197453e-06, "loss": 3.5402, "step": 19436 }, { "epoch": 0.6849018208726446, "grad_norm": 3.0819149017333984, "learning_rate": 4.795196415123324e-06, "loss": 2.7332, "step": 19437 }, { "epoch": 0.684937057885603, "grad_norm": 1.3682522773742676, "learning_rate": 4.794221942842422e-06, "loss": 1.0516, "step": 19438 }, { "epoch": 0.6849722948985615, "grad_norm": 2.1743552684783936, "learning_rate": 4.793247538367448e-06, "loss": 0.8718, "step": 19439 }, { "epoch": 0.6850075319115199, "grad_norm": 3.9110870361328125, "learning_rate": 4.792273201711086e-06, "loss": 3.5735, "step": 19440 }, { "epoch": 0.6850427689244782, "grad_norm": 1.1292765140533447, "learning_rate": 4.791298932886028e-06, "loss": 0.8741, "step": 19441 }, { "epoch": 0.6850780059374367, "grad_norm": 6.240963459014893, "learning_rate": 4.790324731904965e-06, "loss": 3.4182, "step": 19442 }, { "epoch": 0.6851132429503951, "grad_norm": 1.201483964920044, "learning_rate": 4.789350598780586e-06, "loss": 1.0405, "step": 19443 }, { "epoch": 0.6851484799633535, "grad_norm": 1.1953774690628052, "learning_rate": 4.78837653352558e-06, "loss": 0.9265, "step": 19444 }, { "epoch": 0.6851837169763119, "grad_norm": 1.6644511222839355, "learning_rate": 4.787402536152632e-06, "loss": 0.7843, "step": 19445 }, { "epoch": 0.6852189539892704, "grad_norm": 6.2660136222839355, "learning_rate": 4.78642860667443e-06, "loss": 2.7885, "step": 19446 }, { "epoch": 0.6852541910022287, "grad_norm": 2.981905221939087, "learning_rate": 4.7854547451036605e-06, "loss": 2.9151, "step": 19447 }, { "epoch": 0.6852894280151871, "grad_norm": 2.7664217948913574, "learning_rate": 4.784480951453011e-06, "loss": 2.544, "step": 19448 }, { "epoch": 0.6853246650281456, "grad_norm": 2.727518081665039, "learning_rate": 4.783507225735154e-06, "loss": 2.473, "step": 19449 }, { "epoch": 0.685359902041104, "grad_norm": 11.858482360839844, "learning_rate": 4.782533567962782e-06, "loss": 5.5767, "step": 19450 }, { "epoch": 0.6853951390540624, "grad_norm": 1.259614109992981, "learning_rate": 4.781559978148577e-06, "loss": 0.9238, "step": 19451 }, { "epoch": 0.6854303760670208, "grad_norm": 1.1724517345428467, "learning_rate": 4.78058645630522e-06, "loss": 0.9368, "step": 19452 }, { "epoch": 0.6854656130799792, "grad_norm": 5.557875633239746, "learning_rate": 4.779613002445382e-06, "loss": 3.1151, "step": 19453 }, { "epoch": 0.6855008500929376, "grad_norm": 4.57785701751709, "learning_rate": 4.778639616581755e-06, "loss": 3.0582, "step": 19454 }, { "epoch": 0.6855360871058961, "grad_norm": 1.1018707752227783, "learning_rate": 4.777666298727015e-06, "loss": 0.9544, "step": 19455 }, { "epoch": 0.6855713241188545, "grad_norm": 1.3266537189483643, "learning_rate": 4.776693048893832e-06, "loss": 1.186, "step": 19456 }, { "epoch": 0.6856065611318128, "grad_norm": 3.327486038208008, "learning_rate": 4.775719867094885e-06, "loss": 2.664, "step": 19457 }, { "epoch": 0.6856417981447712, "grad_norm": 2.504363536834717, "learning_rate": 4.7747467533428585e-06, "loss": 2.9083, "step": 19458 }, { "epoch": 0.6856770351577297, "grad_norm": 3.2983760833740234, "learning_rate": 4.7737737076504175e-06, "loss": 2.9056, "step": 19459 }, { "epoch": 0.6857122721706881, "grad_norm": 0.9834989905357361, "learning_rate": 4.77280073003024e-06, "loss": 0.7932, "step": 19460 }, { "epoch": 0.6857475091836465, "grad_norm": 1.0705063343048096, "learning_rate": 4.771827820495e-06, "loss": 0.5968, "step": 19461 }, { "epoch": 0.685782746196605, "grad_norm": 2.653031587600708, "learning_rate": 4.770854979057367e-06, "loss": 2.7494, "step": 19462 }, { "epoch": 0.6858179832095633, "grad_norm": 1.3308804035186768, "learning_rate": 4.769882205730014e-06, "loss": 0.928, "step": 19463 }, { "epoch": 0.6858532202225217, "grad_norm": 1.1334034204483032, "learning_rate": 4.768909500525612e-06, "loss": 0.8962, "step": 19464 }, { "epoch": 0.6858884572354802, "grad_norm": 13.994743347167969, "learning_rate": 4.76793686345683e-06, "loss": 3.4918, "step": 19465 }, { "epoch": 0.6859236942484386, "grad_norm": 1.7742058038711548, "learning_rate": 4.766964294536337e-06, "loss": 1.9611, "step": 19466 }, { "epoch": 0.685958931261397, "grad_norm": 3.6271188259124756, "learning_rate": 4.765991793776804e-06, "loss": 2.6147, "step": 19467 }, { "epoch": 0.6859941682743553, "grad_norm": 2.9683837890625, "learning_rate": 4.765019361190888e-06, "loss": 2.827, "step": 19468 }, { "epoch": 0.6860294052873138, "grad_norm": 1.435787320137024, "learning_rate": 4.764046996791265e-06, "loss": 0.9632, "step": 19469 }, { "epoch": 0.6860646423002722, "grad_norm": 1.143437385559082, "learning_rate": 4.763074700590597e-06, "loss": 0.9764, "step": 19470 }, { "epoch": 0.6860998793132306, "grad_norm": 1.1509838104248047, "learning_rate": 4.762102472601551e-06, "loss": 1.0573, "step": 19471 }, { "epoch": 0.6861351163261891, "grad_norm": 3.817152500152588, "learning_rate": 4.76113031283678e-06, "loss": 2.5445, "step": 19472 }, { "epoch": 0.6861703533391474, "grad_norm": 1.5423744916915894, "learning_rate": 4.760158221308958e-06, "loss": 1.0236, "step": 19473 }, { "epoch": 0.6862055903521058, "grad_norm": 3.437145233154297, "learning_rate": 4.759186198030747e-06, "loss": 3.0258, "step": 19474 }, { "epoch": 0.6862408273650643, "grad_norm": 5.385245323181152, "learning_rate": 4.758214243014799e-06, "loss": 4.5633, "step": 19475 }, { "epoch": 0.6862760643780227, "grad_norm": 1.6722649335861206, "learning_rate": 4.7572423562737745e-06, "loss": 0.8239, "step": 19476 }, { "epoch": 0.6863113013909811, "grad_norm": 4.536595344543457, "learning_rate": 4.756270537820344e-06, "loss": 5.1343, "step": 19477 }, { "epoch": 0.6863465384039394, "grad_norm": 1.0808160305023193, "learning_rate": 4.755298787667153e-06, "loss": 0.8195, "step": 19478 }, { "epoch": 0.6863817754168979, "grad_norm": 6.228704452514648, "learning_rate": 4.754327105826863e-06, "loss": 4.4939, "step": 19479 }, { "epoch": 0.6864170124298563, "grad_norm": 1.1721324920654297, "learning_rate": 4.7533554923121325e-06, "loss": 0.9131, "step": 19480 }, { "epoch": 0.6864522494428147, "grad_norm": 5.595917224884033, "learning_rate": 4.752383947135615e-06, "loss": 3.2906, "step": 19481 }, { "epoch": 0.6864874864557732, "grad_norm": 1.8116120100021362, "learning_rate": 4.751412470309964e-06, "loss": 0.9628, "step": 19482 }, { "epoch": 0.6865227234687316, "grad_norm": 1.4152405261993408, "learning_rate": 4.7504410618478334e-06, "loss": 0.97, "step": 19483 }, { "epoch": 0.6865579604816899, "grad_norm": 1.7652229070663452, "learning_rate": 4.749469721761878e-06, "loss": 0.8387, "step": 19484 }, { "epoch": 0.6865931974946484, "grad_norm": 5.320981502532959, "learning_rate": 4.748498450064748e-06, "loss": 4.4112, "step": 19485 }, { "epoch": 0.6866284345076068, "grad_norm": 1.1371785402297974, "learning_rate": 4.747527246769098e-06, "loss": 1.0396, "step": 19486 }, { "epoch": 0.6866636715205652, "grad_norm": 4.373649597167969, "learning_rate": 4.746556111887568e-06, "loss": 4.0787, "step": 19487 }, { "epoch": 0.6866989085335237, "grad_norm": 1.3125344514846802, "learning_rate": 4.745585045432817e-06, "loss": 0.7084, "step": 19488 }, { "epoch": 0.686734145546482, "grad_norm": 3.168698787689209, "learning_rate": 4.74461404741749e-06, "loss": 2.6572, "step": 19489 }, { "epoch": 0.6867693825594404, "grad_norm": 3.1483492851257324, "learning_rate": 4.743643117854237e-06, "loss": 2.4493, "step": 19490 }, { "epoch": 0.6868046195723988, "grad_norm": 6.848592758178711, "learning_rate": 4.742672256755697e-06, "loss": 3.3876, "step": 19491 }, { "epoch": 0.6868398565853573, "grad_norm": 1.026423454284668, "learning_rate": 4.7417014641345225e-06, "loss": 1.0988, "step": 19492 }, { "epoch": 0.6868750935983157, "grad_norm": 1.0195344686508179, "learning_rate": 4.7407307400033605e-06, "loss": 1.0235, "step": 19493 }, { "epoch": 0.686910330611274, "grad_norm": 1.810992956161499, "learning_rate": 4.739760084374847e-06, "loss": 0.6511, "step": 19494 }, { "epoch": 0.6869455676242325, "grad_norm": 1.3431912660598755, "learning_rate": 4.7387894972616245e-06, "loss": 0.9289, "step": 19495 }, { "epoch": 0.6869808046371909, "grad_norm": 6.386590480804443, "learning_rate": 4.737818978676346e-06, "loss": 4.9288, "step": 19496 }, { "epoch": 0.6870160416501493, "grad_norm": 7.7390289306640625, "learning_rate": 4.736848528631643e-06, "loss": 5.0374, "step": 19497 }, { "epoch": 0.6870512786631078, "grad_norm": 2.6081178188323975, "learning_rate": 4.735878147140157e-06, "loss": 2.8782, "step": 19498 }, { "epoch": 0.6870865156760662, "grad_norm": 1.7343060970306396, "learning_rate": 4.734907834214529e-06, "loss": 0.7344, "step": 19499 }, { "epoch": 0.6871217526890245, "grad_norm": 5.487366676330566, "learning_rate": 4.733937589867398e-06, "loss": 3.3039, "step": 19500 }, { "epoch": 0.6871569897019829, "grad_norm": 2.28059983253479, "learning_rate": 4.7329674141114e-06, "loss": 2.5839, "step": 19501 }, { "epoch": 0.6871922267149414, "grad_norm": 1.1931999921798706, "learning_rate": 4.731997306959173e-06, "loss": 0.9632, "step": 19502 }, { "epoch": 0.6872274637278998, "grad_norm": 1.2395234107971191, "learning_rate": 4.731027268423352e-06, "loss": 0.8888, "step": 19503 }, { "epoch": 0.6872627007408582, "grad_norm": 2.0050127506256104, "learning_rate": 4.73005729851657e-06, "loss": 0.7838, "step": 19504 }, { "epoch": 0.6872979377538166, "grad_norm": 4.852070331573486, "learning_rate": 4.729087397251465e-06, "loss": 4.0867, "step": 19505 }, { "epoch": 0.687333174766775, "grad_norm": 5.429441928863525, "learning_rate": 4.728117564640667e-06, "loss": 3.2944, "step": 19506 }, { "epoch": 0.6873684117797334, "grad_norm": 4.329361438751221, "learning_rate": 4.727147800696811e-06, "loss": 3.0329, "step": 19507 }, { "epoch": 0.6874036487926919, "grad_norm": 3.330892562866211, "learning_rate": 4.726178105432525e-06, "loss": 2.7822, "step": 19508 }, { "epoch": 0.6874388858056503, "grad_norm": 1.6986231803894043, "learning_rate": 4.725208478860445e-06, "loss": 1.1123, "step": 19509 }, { "epoch": 0.6874741228186086, "grad_norm": 0.9849306344985962, "learning_rate": 4.7242389209931895e-06, "loss": 0.9663, "step": 19510 }, { "epoch": 0.687509359831567, "grad_norm": 1.056153416633606, "learning_rate": 4.7232694318433966e-06, "loss": 1.0067, "step": 19511 }, { "epoch": 0.6875445968445255, "grad_norm": 1.1576768159866333, "learning_rate": 4.722300011423696e-06, "loss": 0.8219, "step": 19512 }, { "epoch": 0.6875798338574839, "grad_norm": 1.0301477909088135, "learning_rate": 4.7213306597467065e-06, "loss": 1.0845, "step": 19513 }, { "epoch": 0.6876150708704423, "grad_norm": 0.8776478171348572, "learning_rate": 4.7203613768250535e-06, "loss": 1.0167, "step": 19514 }, { "epoch": 0.6876503078834008, "grad_norm": 3.0952258110046387, "learning_rate": 4.719392162671374e-06, "loss": 2.918, "step": 19515 }, { "epoch": 0.6876855448963591, "grad_norm": 1.1063261032104492, "learning_rate": 4.7184230172982805e-06, "loss": 0.7678, "step": 19516 }, { "epoch": 0.6877207819093175, "grad_norm": 4.226132392883301, "learning_rate": 4.7174539407183995e-06, "loss": 2.7964, "step": 19517 }, { "epoch": 0.687756018922276, "grad_norm": 2.6605632305145264, "learning_rate": 4.716484932944354e-06, "loss": 0.9682, "step": 19518 }, { "epoch": 0.6877912559352344, "grad_norm": 1.204978585243225, "learning_rate": 4.715515993988765e-06, "loss": 0.9428, "step": 19519 }, { "epoch": 0.6878264929481928, "grad_norm": 1.1291882991790771, "learning_rate": 4.714547123864253e-06, "loss": 0.7694, "step": 19520 }, { "epoch": 0.6878617299611512, "grad_norm": 1.259463906288147, "learning_rate": 4.713578322583439e-06, "loss": 0.8546, "step": 19521 }, { "epoch": 0.6878969669741096, "grad_norm": 2.7126612663269043, "learning_rate": 4.71260959015894e-06, "loss": 1.101, "step": 19522 }, { "epoch": 0.687932203987068, "grad_norm": 1.5406389236450195, "learning_rate": 4.711640926603375e-06, "loss": 1.0972, "step": 19523 }, { "epoch": 0.6879674410000264, "grad_norm": 1.261497974395752, "learning_rate": 4.7106723319293605e-06, "loss": 1.0366, "step": 19524 }, { "epoch": 0.6880026780129849, "grad_norm": 1.3764742612838745, "learning_rate": 4.709703806149513e-06, "loss": 0.9862, "step": 19525 }, { "epoch": 0.6880379150259432, "grad_norm": 3.7646608352661133, "learning_rate": 4.708735349276447e-06, "loss": 2.6904, "step": 19526 }, { "epoch": 0.6880731520389016, "grad_norm": 4.627975940704346, "learning_rate": 4.707766961322776e-06, "loss": 4.8923, "step": 19527 }, { "epoch": 0.6881083890518601, "grad_norm": 1.4336462020874023, "learning_rate": 4.706798642301119e-06, "loss": 0.8687, "step": 19528 }, { "epoch": 0.6881436260648185, "grad_norm": 1.6201003789901733, "learning_rate": 4.705830392224077e-06, "loss": 0.632, "step": 19529 }, { "epoch": 0.6881788630777769, "grad_norm": 2.921704053878784, "learning_rate": 4.704862211104272e-06, "loss": 3.1137, "step": 19530 }, { "epoch": 0.6882141000907354, "grad_norm": 4.342672824859619, "learning_rate": 4.7038940989543134e-06, "loss": 4.5666, "step": 19531 }, { "epoch": 0.6882493371036937, "grad_norm": 3.13100266456604, "learning_rate": 4.702926055786807e-06, "loss": 4.494, "step": 19532 }, { "epoch": 0.6882845741166521, "grad_norm": 1.2517368793487549, "learning_rate": 4.70195808161436e-06, "loss": 0.9951, "step": 19533 }, { "epoch": 0.6883198111296105, "grad_norm": 2.797740936279297, "learning_rate": 4.700990176449589e-06, "loss": 3.1582, "step": 19534 }, { "epoch": 0.688355048142569, "grad_norm": 2.445481300354004, "learning_rate": 4.700022340305094e-06, "loss": 0.7105, "step": 19535 }, { "epoch": 0.6883902851555274, "grad_norm": 3.8776817321777344, "learning_rate": 4.6990545731934825e-06, "loss": 2.9688, "step": 19536 }, { "epoch": 0.6884255221684857, "grad_norm": 1.389724850654602, "learning_rate": 4.6980868751273615e-06, "loss": 0.9117, "step": 19537 }, { "epoch": 0.6884607591814442, "grad_norm": 2.612197160720825, "learning_rate": 4.6971192461193326e-06, "loss": 2.6914, "step": 19538 }, { "epoch": 0.6884959961944026, "grad_norm": 12.457027435302734, "learning_rate": 4.696151686182002e-06, "loss": 4.9481, "step": 19539 }, { "epoch": 0.688531233207361, "grad_norm": 4.221597194671631, "learning_rate": 4.695184195327971e-06, "loss": 4.8166, "step": 19540 }, { "epoch": 0.6885664702203195, "grad_norm": 1.056630253791809, "learning_rate": 4.694216773569842e-06, "loss": 0.7474, "step": 19541 }, { "epoch": 0.6886017072332778, "grad_norm": 4.885460376739502, "learning_rate": 4.693249420920215e-06, "loss": 5.3111, "step": 19542 }, { "epoch": 0.6886369442462362, "grad_norm": 3.190267324447632, "learning_rate": 4.692282137391691e-06, "loss": 2.8573, "step": 19543 }, { "epoch": 0.6886721812591946, "grad_norm": 2.0318539142608643, "learning_rate": 4.691314922996867e-06, "loss": 0.8895, "step": 19544 }, { "epoch": 0.6887074182721531, "grad_norm": 1.3116581439971924, "learning_rate": 4.690347777748343e-06, "loss": 0.7202, "step": 19545 }, { "epoch": 0.6887426552851115, "grad_norm": 1.21300208568573, "learning_rate": 4.689380701658715e-06, "loss": 0.724, "step": 19546 }, { "epoch": 0.6887778922980698, "grad_norm": 1.3491255044937134, "learning_rate": 4.688413694740584e-06, "loss": 0.8551, "step": 19547 }, { "epoch": 0.6888131293110283, "grad_norm": 1.1592381000518799, "learning_rate": 4.687446757006534e-06, "loss": 0.7839, "step": 19548 }, { "epoch": 0.6888483663239867, "grad_norm": 1.4698519706726074, "learning_rate": 4.6864798884691705e-06, "loss": 0.956, "step": 19549 }, { "epoch": 0.6888836033369451, "grad_norm": 2.902108907699585, "learning_rate": 4.6855130891410875e-06, "loss": 1.025, "step": 19550 }, { "epoch": 0.6889188403499036, "grad_norm": 5.366255283355713, "learning_rate": 4.684546359034868e-06, "loss": 3.661, "step": 19551 }, { "epoch": 0.688954077362862, "grad_norm": 3.385329484939575, "learning_rate": 4.6835796981631065e-06, "loss": 2.7233, "step": 19552 }, { "epoch": 0.6889893143758203, "grad_norm": 1.2998405694961548, "learning_rate": 4.6826131065384035e-06, "loss": 1.0299, "step": 19553 }, { "epoch": 0.6890245513887788, "grad_norm": 3.8919215202331543, "learning_rate": 4.681646584173341e-06, "loss": 2.6309, "step": 19554 }, { "epoch": 0.6890597884017372, "grad_norm": 2.646322250366211, "learning_rate": 4.680680131080506e-06, "loss": 2.9908, "step": 19555 }, { "epoch": 0.6890950254146956, "grad_norm": 3.5022952556610107, "learning_rate": 4.679713747272491e-06, "loss": 2.9478, "step": 19556 }, { "epoch": 0.689130262427654, "grad_norm": 1.2077250480651855, "learning_rate": 4.678747432761882e-06, "loss": 0.8428, "step": 19557 }, { "epoch": 0.6891654994406125, "grad_norm": 2.485625982284546, "learning_rate": 4.677781187561265e-06, "loss": 1.0807, "step": 19558 }, { "epoch": 0.6892007364535708, "grad_norm": 6.924465656280518, "learning_rate": 4.676815011683228e-06, "loss": 3.5629, "step": 19559 }, { "epoch": 0.6892359734665292, "grad_norm": 4.002251625061035, "learning_rate": 4.6758489051403525e-06, "loss": 2.9915, "step": 19560 }, { "epoch": 0.6892712104794877, "grad_norm": 6.44556188583374, "learning_rate": 4.674882867945223e-06, "loss": 3.244, "step": 19561 }, { "epoch": 0.6893064474924461, "grad_norm": 4.749683380126953, "learning_rate": 4.673916900110422e-06, "loss": 4.4353, "step": 19562 }, { "epoch": 0.6893416845054045, "grad_norm": 4.096682548522949, "learning_rate": 4.672951001648532e-06, "loss": 2.7068, "step": 19563 }, { "epoch": 0.6893769215183629, "grad_norm": 1.1430281400680542, "learning_rate": 4.671985172572135e-06, "loss": 0.8585, "step": 19564 }, { "epoch": 0.6894121585313213, "grad_norm": 4.937448501586914, "learning_rate": 4.671019412893808e-06, "loss": 3.0815, "step": 19565 }, { "epoch": 0.6894473955442797, "grad_norm": 1.3763058185577393, "learning_rate": 4.670053722626137e-06, "loss": 0.8401, "step": 19566 }, { "epoch": 0.6894826325572381, "grad_norm": 3.537665367126465, "learning_rate": 4.669088101781686e-06, "loss": 2.5605, "step": 19567 }, { "epoch": 0.6895178695701966, "grad_norm": 0.819512128829956, "learning_rate": 4.668122550373046e-06, "loss": 0.714, "step": 19568 }, { "epoch": 0.6895531065831549, "grad_norm": 4.551994800567627, "learning_rate": 4.667157068412794e-06, "loss": 2.713, "step": 19569 }, { "epoch": 0.6895883435961133, "grad_norm": 1.539533257484436, "learning_rate": 4.666191655913496e-06, "loss": 0.7808, "step": 19570 }, { "epoch": 0.6896235806090718, "grad_norm": 2.2591817378997803, "learning_rate": 4.665226312887727e-06, "loss": 2.6941, "step": 19571 }, { "epoch": 0.6896588176220302, "grad_norm": 1.221016526222229, "learning_rate": 4.664261039348072e-06, "loss": 1.021, "step": 19572 }, { "epoch": 0.6896940546349886, "grad_norm": 1.600536823272705, "learning_rate": 4.663295835307094e-06, "loss": 0.8532, "step": 19573 }, { "epoch": 0.689729291647947, "grad_norm": 1.0448147058486938, "learning_rate": 4.662330700777366e-06, "loss": 0.6919, "step": 19574 }, { "epoch": 0.6897645286609054, "grad_norm": 2.352776288986206, "learning_rate": 4.661365635771462e-06, "loss": 2.8252, "step": 19575 }, { "epoch": 0.6897997656738638, "grad_norm": 1.3743340969085693, "learning_rate": 4.6604006403019495e-06, "loss": 1.2106, "step": 19576 }, { "epoch": 0.6898350026868222, "grad_norm": 6.115122318267822, "learning_rate": 4.659435714381401e-06, "loss": 5.3, "step": 19577 }, { "epoch": 0.6898702396997807, "grad_norm": 4.822808742523193, "learning_rate": 4.658470858022381e-06, "loss": 4.6673, "step": 19578 }, { "epoch": 0.689905476712739, "grad_norm": 3.0250661373138428, "learning_rate": 4.657506071237458e-06, "loss": 3.0328, "step": 19579 }, { "epoch": 0.6899407137256974, "grad_norm": 1.8134675025939941, "learning_rate": 4.656541354039199e-06, "loss": 0.7816, "step": 19580 }, { "epoch": 0.6899759507386559, "grad_norm": 1.3485361337661743, "learning_rate": 4.65557670644017e-06, "loss": 0.935, "step": 19581 }, { "epoch": 0.6900111877516143, "grad_norm": 3.620389223098755, "learning_rate": 4.654612128452936e-06, "loss": 4.3429, "step": 19582 }, { "epoch": 0.6900464247645727, "grad_norm": 1.3161460161209106, "learning_rate": 4.653647620090058e-06, "loss": 0.8746, "step": 19583 }, { "epoch": 0.6900816617775312, "grad_norm": 1.5200265645980835, "learning_rate": 4.652683181364102e-06, "loss": 1.0385, "step": 19584 }, { "epoch": 0.6901168987904895, "grad_norm": 2.673234462738037, "learning_rate": 4.651718812287632e-06, "loss": 3.0299, "step": 19585 }, { "epoch": 0.6901521358034479, "grad_norm": 4.324657917022705, "learning_rate": 4.650754512873199e-06, "loss": 3.3007, "step": 19586 }, { "epoch": 0.6901873728164064, "grad_norm": 2.6085519790649414, "learning_rate": 4.649790283133373e-06, "loss": 2.1651, "step": 19587 }, { "epoch": 0.6902226098293648, "grad_norm": 1.2371574640274048, "learning_rate": 4.648826123080714e-06, "loss": 0.9516, "step": 19588 }, { "epoch": 0.6902578468423232, "grad_norm": 1.2228808403015137, "learning_rate": 4.647862032727773e-06, "loss": 0.7894, "step": 19589 }, { "epoch": 0.6902930838552815, "grad_norm": 4.544802665710449, "learning_rate": 4.64689801208711e-06, "loss": 2.7198, "step": 19590 }, { "epoch": 0.69032832086824, "grad_norm": 3.6885154247283936, "learning_rate": 4.645934061171282e-06, "loss": 3.3871, "step": 19591 }, { "epoch": 0.6903635578811984, "grad_norm": 6.357540607452393, "learning_rate": 4.644970179992846e-06, "loss": 5.1798, "step": 19592 }, { "epoch": 0.6903987948941568, "grad_norm": 1.3874235153198242, "learning_rate": 4.6440063685643545e-06, "loss": 0.7514, "step": 19593 }, { "epoch": 0.6904340319071153, "grad_norm": 5.207400321960449, "learning_rate": 4.643042626898363e-06, "loss": 2.833, "step": 19594 }, { "epoch": 0.6904692689200737, "grad_norm": 1.1224299669265747, "learning_rate": 4.642078955007423e-06, "loss": 1.064, "step": 19595 }, { "epoch": 0.690504505933032, "grad_norm": 1.2094823122024536, "learning_rate": 4.6411153529040875e-06, "loss": 0.814, "step": 19596 }, { "epoch": 0.6905397429459905, "grad_norm": 3.2078442573547363, "learning_rate": 4.640151820600907e-06, "loss": 2.6828, "step": 19597 }, { "epoch": 0.6905749799589489, "grad_norm": 0.939694881439209, "learning_rate": 4.639188358110433e-06, "loss": 0.8392, "step": 19598 }, { "epoch": 0.6906102169719073, "grad_norm": 1.5701911449432373, "learning_rate": 4.638224965445211e-06, "loss": 1.1058, "step": 19599 }, { "epoch": 0.6906454539848657, "grad_norm": 4.675471305847168, "learning_rate": 4.637261642617794e-06, "loss": 3.1962, "step": 19600 }, { "epoch": 0.6906806909978241, "grad_norm": 1.4293453693389893, "learning_rate": 4.6362983896407296e-06, "loss": 0.8027, "step": 19601 }, { "epoch": 0.6907159280107825, "grad_norm": 1.139161467552185, "learning_rate": 4.635335206526556e-06, "loss": 1.0581, "step": 19602 }, { "epoch": 0.6907511650237409, "grad_norm": 5.898914813995361, "learning_rate": 4.634372093287826e-06, "loss": 6.7486, "step": 19603 }, { "epoch": 0.6907864020366994, "grad_norm": 6.261618614196777, "learning_rate": 4.633409049937089e-06, "loss": 5.1718, "step": 19604 }, { "epoch": 0.6908216390496578, "grad_norm": 7.270510196685791, "learning_rate": 4.6324460764868785e-06, "loss": 6.8813, "step": 19605 }, { "epoch": 0.6908568760626161, "grad_norm": 1.1158719062805176, "learning_rate": 4.6314831729497365e-06, "loss": 0.7235, "step": 19606 }, { "epoch": 0.6908921130755746, "grad_norm": 0.9655440449714661, "learning_rate": 4.6305203393382196e-06, "loss": 0.6982, "step": 19607 }, { "epoch": 0.690927350088533, "grad_norm": 3.6517772674560547, "learning_rate": 4.629557575664854e-06, "loss": 2.8192, "step": 19608 }, { "epoch": 0.6909625871014914, "grad_norm": 0.8622295260429382, "learning_rate": 4.628594881942186e-06, "loss": 0.756, "step": 19609 }, { "epoch": 0.6909978241144498, "grad_norm": 3.9214649200439453, "learning_rate": 4.627632258182753e-06, "loss": 2.8931, "step": 19610 }, { "epoch": 0.6910330611274083, "grad_norm": 1.3999063968658447, "learning_rate": 4.6266697043990934e-06, "loss": 0.7753, "step": 19611 }, { "epoch": 0.6910682981403666, "grad_norm": 0.9434711933135986, "learning_rate": 4.625707220603746e-06, "loss": 0.8545, "step": 19612 }, { "epoch": 0.691103535153325, "grad_norm": 3.499004602432251, "learning_rate": 4.6247448068092486e-06, "loss": 3.2272, "step": 19613 }, { "epoch": 0.6911387721662835, "grad_norm": 7.963979244232178, "learning_rate": 4.623782463028133e-06, "loss": 3.7603, "step": 19614 }, { "epoch": 0.6911740091792419, "grad_norm": 3.6108620166778564, "learning_rate": 4.6228201892729355e-06, "loss": 2.8567, "step": 19615 }, { "epoch": 0.6912092461922003, "grad_norm": 2.528637170791626, "learning_rate": 4.62185798555619e-06, "loss": 0.7976, "step": 19616 }, { "epoch": 0.6912444832051587, "grad_norm": 3.178102970123291, "learning_rate": 4.62089585189043e-06, "loss": 2.5362, "step": 19617 }, { "epoch": 0.6912797202181171, "grad_norm": 1.0353538990020752, "learning_rate": 4.619933788288186e-06, "loss": 0.9785, "step": 19618 }, { "epoch": 0.6913149572310755, "grad_norm": 1.0868573188781738, "learning_rate": 4.618971794761989e-06, "loss": 0.7549, "step": 19619 }, { "epoch": 0.691350194244034, "grad_norm": 3.2349722385406494, "learning_rate": 4.618009871324376e-06, "loss": 2.6895, "step": 19620 }, { "epoch": 0.6913854312569924, "grad_norm": 2.1948227882385254, "learning_rate": 4.617048017987861e-06, "loss": 0.738, "step": 19621 }, { "epoch": 0.6914206682699507, "grad_norm": 1.045543909072876, "learning_rate": 4.616086234764985e-06, "loss": 0.844, "step": 19622 }, { "epoch": 0.6914559052829091, "grad_norm": 1.1866121292114258, "learning_rate": 4.615124521668275e-06, "loss": 0.7474, "step": 19623 }, { "epoch": 0.6914911422958676, "grad_norm": 3.6300179958343506, "learning_rate": 4.614162878710251e-06, "loss": 2.7071, "step": 19624 }, { "epoch": 0.691526379308826, "grad_norm": 1.2872365713119507, "learning_rate": 4.613201305903438e-06, "loss": 1.0932, "step": 19625 }, { "epoch": 0.6915616163217844, "grad_norm": 1.1417995691299438, "learning_rate": 4.612239803260371e-06, "loss": 1.0264, "step": 19626 }, { "epoch": 0.6915968533347429, "grad_norm": 2.0494134426116943, "learning_rate": 4.611278370793563e-06, "loss": 1.1031, "step": 19627 }, { "epoch": 0.6916320903477012, "grad_norm": 5.260972499847412, "learning_rate": 4.61031700851554e-06, "loss": 5.1244, "step": 19628 }, { "epoch": 0.6916673273606596, "grad_norm": 1.0794832706451416, "learning_rate": 4.609355716438826e-06, "loss": 0.6856, "step": 19629 }, { "epoch": 0.6917025643736181, "grad_norm": 1.0309151411056519, "learning_rate": 4.6083944945759396e-06, "loss": 0.9928, "step": 19630 }, { "epoch": 0.6917378013865765, "grad_norm": 3.805009126663208, "learning_rate": 4.607433342939402e-06, "loss": 2.744, "step": 19631 }, { "epoch": 0.6917730383995349, "grad_norm": 3.2317683696746826, "learning_rate": 4.6064722615417326e-06, "loss": 4.6388, "step": 19632 }, { "epoch": 0.6918082754124932, "grad_norm": 4.954229354858398, "learning_rate": 4.605511250395448e-06, "loss": 3.0421, "step": 19633 }, { "epoch": 0.6918435124254517, "grad_norm": 1.0601847171783447, "learning_rate": 4.604550309513067e-06, "loss": 0.84, "step": 19634 }, { "epoch": 0.6918787494384101, "grad_norm": 1.0300145149230957, "learning_rate": 4.603589438907106e-06, "loss": 0.6602, "step": 19635 }, { "epoch": 0.6919139864513685, "grad_norm": 1.390453815460205, "learning_rate": 4.602628638590078e-06, "loss": 0.7981, "step": 19636 }, { "epoch": 0.691949223464327, "grad_norm": 1.2388689517974854, "learning_rate": 4.601667908574502e-06, "loss": 1.1549, "step": 19637 }, { "epoch": 0.6919844604772853, "grad_norm": 3.299248695373535, "learning_rate": 4.600707248872888e-06, "loss": 2.9816, "step": 19638 }, { "epoch": 0.6920196974902437, "grad_norm": 1.8933086395263672, "learning_rate": 4.599746659497754e-06, "loss": 0.9572, "step": 19639 }, { "epoch": 0.6920549345032022, "grad_norm": 1.6958861351013184, "learning_rate": 4.598786140461598e-06, "loss": 0.9956, "step": 19640 }, { "epoch": 0.6920901715161606, "grad_norm": 1.8157140016555786, "learning_rate": 4.597825691776947e-06, "loss": 0.8402, "step": 19641 }, { "epoch": 0.692125408529119, "grad_norm": 3.2107930183410645, "learning_rate": 4.596865313456307e-06, "loss": 2.0658, "step": 19642 }, { "epoch": 0.6921606455420773, "grad_norm": 1.0853986740112305, "learning_rate": 4.595905005512181e-06, "loss": 0.7114, "step": 19643 }, { "epoch": 0.6921958825550358, "grad_norm": 2.625471353530884, "learning_rate": 4.594944767957077e-06, "loss": 2.7896, "step": 19644 }, { "epoch": 0.6922311195679942, "grad_norm": 1.7707586288452148, "learning_rate": 4.593984600803513e-06, "loss": 0.7548, "step": 19645 }, { "epoch": 0.6922663565809526, "grad_norm": 6.4749603271484375, "learning_rate": 4.5930245040639856e-06, "loss": 2.8973, "step": 19646 }, { "epoch": 0.6923015935939111, "grad_norm": 1.37868332862854, "learning_rate": 4.592064477751002e-06, "loss": 1.0648, "step": 19647 }, { "epoch": 0.6923368306068695, "grad_norm": 9.035202026367188, "learning_rate": 4.591104521877067e-06, "loss": 6.9905, "step": 19648 }, { "epoch": 0.6923720676198278, "grad_norm": 3.1649513244628906, "learning_rate": 4.590144636454684e-06, "loss": 3.2076, "step": 19649 }, { "epoch": 0.6924073046327863, "grad_norm": 1.5971429347991943, "learning_rate": 4.589184821496358e-06, "loss": 0.8421, "step": 19650 }, { "epoch": 0.6924425416457447, "grad_norm": 4.025147438049316, "learning_rate": 4.588225077014587e-06, "loss": 3.168, "step": 19651 }, { "epoch": 0.6924777786587031, "grad_norm": 2.521271228790283, "learning_rate": 4.5872654030218755e-06, "loss": 0.9009, "step": 19652 }, { "epoch": 0.6925130156716615, "grad_norm": 0.9640065431594849, "learning_rate": 4.586305799530721e-06, "loss": 0.9471, "step": 19653 }, { "epoch": 0.69254825268462, "grad_norm": 6.885505199432373, "learning_rate": 4.585346266553623e-06, "loss": 6.5917, "step": 19654 }, { "epoch": 0.6925834896975783, "grad_norm": 4.120974063873291, "learning_rate": 4.584386804103079e-06, "loss": 2.7945, "step": 19655 }, { "epoch": 0.6926187267105367, "grad_norm": 3.770881414413452, "learning_rate": 4.583427412191588e-06, "loss": 3.2874, "step": 19656 }, { "epoch": 0.6926539637234952, "grad_norm": 4.374365329742432, "learning_rate": 4.582468090831645e-06, "loss": 3.3068, "step": 19657 }, { "epoch": 0.6926892007364536, "grad_norm": 6.29444694519043, "learning_rate": 4.581508840035747e-06, "loss": 3.0447, "step": 19658 }, { "epoch": 0.692724437749412, "grad_norm": 1.3654111623764038, "learning_rate": 4.58054965981638e-06, "loss": 0.8064, "step": 19659 }, { "epoch": 0.6927596747623704, "grad_norm": 4.408524990081787, "learning_rate": 4.579590550186048e-06, "loss": 3.4323, "step": 19660 }, { "epoch": 0.6927949117753288, "grad_norm": 1.0616984367370605, "learning_rate": 4.578631511157244e-06, "loss": 0.7958, "step": 19661 }, { "epoch": 0.6928301487882872, "grad_norm": 5.852322101593018, "learning_rate": 4.577672542742451e-06, "loss": 4.6926, "step": 19662 }, { "epoch": 0.6928653858012457, "grad_norm": 3.5872652530670166, "learning_rate": 4.576713644954159e-06, "loss": 2.7854, "step": 19663 }, { "epoch": 0.6929006228142041, "grad_norm": 0.9636393785476685, "learning_rate": 4.575754817804871e-06, "loss": 0.8504, "step": 19664 }, { "epoch": 0.6929358598271624, "grad_norm": 1.0691661834716797, "learning_rate": 4.5747960613070645e-06, "loss": 0.8502, "step": 19665 }, { "epoch": 0.6929710968401208, "grad_norm": 3.8976097106933594, "learning_rate": 4.573837375473229e-06, "loss": 3.3465, "step": 19666 }, { "epoch": 0.6930063338530793, "grad_norm": 0.7834271192550659, "learning_rate": 4.572878760315853e-06, "loss": 0.561, "step": 19667 }, { "epoch": 0.6930415708660377, "grad_norm": 4.6086907386779785, "learning_rate": 4.571920215847423e-06, "loss": 3.1302, "step": 19668 }, { "epoch": 0.6930768078789961, "grad_norm": 1.8228018283843994, "learning_rate": 4.570961742080423e-06, "loss": 1.0168, "step": 19669 }, { "epoch": 0.6931120448919545, "grad_norm": 3.045074939727783, "learning_rate": 4.570003339027338e-06, "loss": 2.326, "step": 19670 }, { "epoch": 0.6931472819049129, "grad_norm": 0.915057897567749, "learning_rate": 4.569045006700651e-06, "loss": 0.8973, "step": 19671 }, { "epoch": 0.6931825189178713, "grad_norm": 3.519958734512329, "learning_rate": 4.568086745112845e-06, "loss": 2.9134, "step": 19672 }, { "epoch": 0.6932177559308298, "grad_norm": 5.832251071929932, "learning_rate": 4.567128554276401e-06, "loss": 3.2687, "step": 19673 }, { "epoch": 0.6932529929437882, "grad_norm": 5.339918613433838, "learning_rate": 4.566170434203799e-06, "loss": 4.9469, "step": 19674 }, { "epoch": 0.6932882299567465, "grad_norm": 1.0196056365966797, "learning_rate": 4.56521238490752e-06, "loss": 0.7145, "step": 19675 }, { "epoch": 0.6933234669697049, "grad_norm": 3.5775623321533203, "learning_rate": 4.564254406400042e-06, "loss": 2.9113, "step": 19676 }, { "epoch": 0.6933587039826634, "grad_norm": 3.0618209838867188, "learning_rate": 4.563296498693847e-06, "loss": 2.7938, "step": 19677 }, { "epoch": 0.6933939409956218, "grad_norm": 4.948843002319336, "learning_rate": 4.5623386618014e-06, "loss": 3.2634, "step": 19678 }, { "epoch": 0.6934291780085802, "grad_norm": 1.1232930421829224, "learning_rate": 4.561380895735187e-06, "loss": 0.9527, "step": 19679 }, { "epoch": 0.6934644150215387, "grad_norm": 3.30865740776062, "learning_rate": 4.560423200507685e-06, "loss": 2.6973, "step": 19680 }, { "epoch": 0.693499652034497, "grad_norm": 3.8292155265808105, "learning_rate": 4.55946557613136e-06, "loss": 3.0085, "step": 19681 }, { "epoch": 0.6935348890474554, "grad_norm": 3.203800916671753, "learning_rate": 4.558508022618685e-06, "loss": 0.6784, "step": 19682 }, { "epoch": 0.6935701260604139, "grad_norm": 1.3559494018554688, "learning_rate": 4.557550539982144e-06, "loss": 0.5874, "step": 19683 }, { "epoch": 0.6936053630733723, "grad_norm": 4.513883590698242, "learning_rate": 4.556593128234197e-06, "loss": 3.2192, "step": 19684 }, { "epoch": 0.6936406000863307, "grad_norm": 4.612298488616943, "learning_rate": 4.555635787387318e-06, "loss": 2.8184, "step": 19685 }, { "epoch": 0.693675837099289, "grad_norm": 4.311649322509766, "learning_rate": 4.5546785174539754e-06, "loss": 3.2039, "step": 19686 }, { "epoch": 0.6937110741122475, "grad_norm": 2.779064893722534, "learning_rate": 4.5537213184466385e-06, "loss": 2.6039, "step": 19687 }, { "epoch": 0.6937463111252059, "grad_norm": 5.156973361968994, "learning_rate": 4.5527641903777764e-06, "loss": 2.8877, "step": 19688 }, { "epoch": 0.6937815481381643, "grad_norm": 1.046474814414978, "learning_rate": 4.551807133259854e-06, "loss": 1.0847, "step": 19689 }, { "epoch": 0.6938167851511228, "grad_norm": 1.046474814414978, "learning_rate": 4.551807133259854e-06, "loss": 4.4394, "step": 19690 }, { "epoch": 0.6938520221640812, "grad_norm": 0.8918147087097168, "learning_rate": 4.550850147105338e-06, "loss": 0.8694, "step": 19691 }, { "epoch": 0.6938872591770395, "grad_norm": 8.870742797851562, "learning_rate": 4.549893231926693e-06, "loss": 4.7406, "step": 19692 }, { "epoch": 0.693922496189998, "grad_norm": 1.1975090503692627, "learning_rate": 4.548936387736382e-06, "loss": 0.7464, "step": 19693 }, { "epoch": 0.6939577332029564, "grad_norm": 0.9155622124671936, "learning_rate": 4.54797961454687e-06, "loss": 0.7549, "step": 19694 }, { "epoch": 0.6939929702159148, "grad_norm": 3.2080063819885254, "learning_rate": 4.547022912370617e-06, "loss": 2.6487, "step": 19695 }, { "epoch": 0.6940282072288733, "grad_norm": 6.582421779632568, "learning_rate": 4.546066281220086e-06, "loss": 5.4692, "step": 19696 }, { "epoch": 0.6940634442418316, "grad_norm": 7.401474475860596, "learning_rate": 4.54510972110774e-06, "loss": 7.3024, "step": 19697 }, { "epoch": 0.69409868125479, "grad_norm": 1.6187422275543213, "learning_rate": 4.544153232046027e-06, "loss": 0.9923, "step": 19698 }, { "epoch": 0.6941339182677484, "grad_norm": 2.9683165550231934, "learning_rate": 4.543196814047417e-06, "loss": 2.6168, "step": 19699 }, { "epoch": 0.6941691552807069, "grad_norm": 2.046389579772949, "learning_rate": 4.542240467124367e-06, "loss": 0.835, "step": 19700 }, { "epoch": 0.6942043922936653, "grad_norm": 1.624199628829956, "learning_rate": 4.541284191289326e-06, "loss": 0.9377, "step": 19701 }, { "epoch": 0.6942396293066236, "grad_norm": 2.361117124557495, "learning_rate": 4.5403279865547516e-06, "loss": 0.9151, "step": 19702 }, { "epoch": 0.6942748663195821, "grad_norm": 2.8970298767089844, "learning_rate": 4.539371852933107e-06, "loss": 2.9366, "step": 19703 }, { "epoch": 0.6943101033325405, "grad_norm": 1.0203312635421753, "learning_rate": 4.5384157904368356e-06, "loss": 0.9158, "step": 19704 }, { "epoch": 0.6943453403454989, "grad_norm": 5.239936351776123, "learning_rate": 4.537459799078394e-06, "loss": 2.9555, "step": 19705 }, { "epoch": 0.6943805773584574, "grad_norm": 2.0259640216827393, "learning_rate": 4.536503878870235e-06, "loss": 1.1217, "step": 19706 }, { "epoch": 0.6944158143714158, "grad_norm": 5.415520668029785, "learning_rate": 4.535548029824809e-06, "loss": 2.6201, "step": 19707 }, { "epoch": 0.6944510513843741, "grad_norm": 4.74150276184082, "learning_rate": 4.534592251954566e-06, "loss": 3.0784, "step": 19708 }, { "epoch": 0.6944862883973325, "grad_norm": 1.056233286857605, "learning_rate": 4.533636545271954e-06, "loss": 0.792, "step": 19709 }, { "epoch": 0.694521525410291, "grad_norm": 1.1979960203170776, "learning_rate": 4.5326809097894245e-06, "loss": 1.0086, "step": 19710 }, { "epoch": 0.6945567624232494, "grad_norm": 1.114709496498108, "learning_rate": 4.531725345519421e-06, "loss": 1.028, "step": 19711 }, { "epoch": 0.6945919994362078, "grad_norm": 3.625758647918701, "learning_rate": 4.530769852474391e-06, "loss": 3.0634, "step": 19712 }, { "epoch": 0.6946272364491662, "grad_norm": 2.270037889480591, "learning_rate": 4.529814430666781e-06, "loss": 0.6752, "step": 19713 }, { "epoch": 0.6946624734621246, "grad_norm": 7.309194087982178, "learning_rate": 4.528859080109035e-06, "loss": 3.3479, "step": 19714 }, { "epoch": 0.694697710475083, "grad_norm": 1.5561368465423584, "learning_rate": 4.527903800813597e-06, "loss": 0.8185, "step": 19715 }, { "epoch": 0.6947329474880415, "grad_norm": 1.0127204656600952, "learning_rate": 4.526948592792912e-06, "loss": 0.6731, "step": 19716 }, { "epoch": 0.6947681845009999, "grad_norm": 1.069549322128296, "learning_rate": 4.525993456059411e-06, "loss": 0.7946, "step": 19717 }, { "epoch": 0.6948034215139582, "grad_norm": 5.694787502288818, "learning_rate": 4.525038390625546e-06, "loss": 3.0726, "step": 19718 }, { "epoch": 0.6948386585269166, "grad_norm": 1.9099154472351074, "learning_rate": 4.5240833965037575e-06, "loss": 0.8796, "step": 19719 }, { "epoch": 0.6948738955398751, "grad_norm": 0.9892815947532654, "learning_rate": 4.523128473706477e-06, "loss": 0.8796, "step": 19720 }, { "epoch": 0.6949091325528335, "grad_norm": 5.2515130043029785, "learning_rate": 4.522173622246142e-06, "loss": 3.4466, "step": 19721 }, { "epoch": 0.6949443695657919, "grad_norm": 3.3370089530944824, "learning_rate": 4.5212188421351996e-06, "loss": 2.9749, "step": 19722 }, { "epoch": 0.6949796065787504, "grad_norm": 2.1657354831695557, "learning_rate": 4.520264133386078e-06, "loss": 0.9085, "step": 19723 }, { "epoch": 0.6950148435917087, "grad_norm": 3.2858986854553223, "learning_rate": 4.519309496011213e-06, "loss": 2.9094, "step": 19724 }, { "epoch": 0.6950500806046671, "grad_norm": 2.362330198287964, "learning_rate": 4.5183549300230355e-06, "loss": 0.753, "step": 19725 }, { "epoch": 0.6950853176176256, "grad_norm": 2.4987456798553467, "learning_rate": 4.517400435433992e-06, "loss": 2.5191, "step": 19726 }, { "epoch": 0.695120554630584, "grad_norm": 2.3633270263671875, "learning_rate": 4.516446012256501e-06, "loss": 0.7963, "step": 19727 }, { "epoch": 0.6951557916435424, "grad_norm": 7.978983402252197, "learning_rate": 4.515491660503e-06, "loss": 5.5325, "step": 19728 }, { "epoch": 0.6951910286565008, "grad_norm": 2.9055418968200684, "learning_rate": 4.51453738018592e-06, "loss": 3.2451, "step": 19729 }, { "epoch": 0.6952262656694592, "grad_norm": 14.503904342651367, "learning_rate": 4.513583171317687e-06, "loss": 3.0686, "step": 19730 }, { "epoch": 0.6952615026824176, "grad_norm": 1.3570300340652466, "learning_rate": 4.512629033910734e-06, "loss": 0.9252, "step": 19731 }, { "epoch": 0.695296739695376, "grad_norm": 3.2526559829711914, "learning_rate": 4.511674967977485e-06, "loss": 2.9605, "step": 19732 }, { "epoch": 0.6953319767083345, "grad_norm": 4.534245014190674, "learning_rate": 4.51072097353037e-06, "loss": 2.9056, "step": 19733 }, { "epoch": 0.6953672137212928, "grad_norm": 5.078657150268555, "learning_rate": 4.5097670505818135e-06, "loss": 3.4756, "step": 19734 }, { "epoch": 0.6954024507342512, "grad_norm": 3.689105987548828, "learning_rate": 4.508813199144244e-06, "loss": 4.6529, "step": 19735 }, { "epoch": 0.6954376877472097, "grad_norm": 3.7900238037109375, "learning_rate": 4.5078594192300736e-06, "loss": 2.6463, "step": 19736 }, { "epoch": 0.6954729247601681, "grad_norm": 0.806670069694519, "learning_rate": 4.506905710851739e-06, "loss": 0.83, "step": 19737 }, { "epoch": 0.6955081617731265, "grad_norm": 6.330793857574463, "learning_rate": 4.50595207402166e-06, "loss": 6.6452, "step": 19738 }, { "epoch": 0.695543398786085, "grad_norm": 1.1955686807632446, "learning_rate": 4.504998508752252e-06, "loss": 0.7745, "step": 19739 }, { "epoch": 0.6955786357990433, "grad_norm": 1.8316271305084229, "learning_rate": 4.504045015055935e-06, "loss": 0.9498, "step": 19740 }, { "epoch": 0.6956138728120017, "grad_norm": 3.318784236907959, "learning_rate": 4.503091592945138e-06, "loss": 1.0959, "step": 19741 }, { "epoch": 0.6956491098249601, "grad_norm": 4.32312536239624, "learning_rate": 4.50213824243227e-06, "loss": 4.7625, "step": 19742 }, { "epoch": 0.6956843468379186, "grad_norm": 4.752070903778076, "learning_rate": 4.501184963529752e-06, "loss": 4.0947, "step": 19743 }, { "epoch": 0.695719583850877, "grad_norm": 3.220343828201294, "learning_rate": 4.500231756249997e-06, "loss": 2.7109, "step": 19744 }, { "epoch": 0.6957548208638353, "grad_norm": 5.932055473327637, "learning_rate": 4.4992786206054306e-06, "loss": 4.2396, "step": 19745 }, { "epoch": 0.6957900578767938, "grad_norm": 1.1769604682922363, "learning_rate": 4.498325556608457e-06, "loss": 0.9311, "step": 19746 }, { "epoch": 0.6958252948897522, "grad_norm": 1.3384097814559937, "learning_rate": 4.497372564271494e-06, "loss": 0.9115, "step": 19747 }, { "epoch": 0.6958605319027106, "grad_norm": 1.2194066047668457, "learning_rate": 4.496419643606954e-06, "loss": 1.3208, "step": 19748 }, { "epoch": 0.6958957689156691, "grad_norm": 1.1178783178329468, "learning_rate": 4.49546679462725e-06, "loss": 0.8592, "step": 19749 }, { "epoch": 0.6959310059286274, "grad_norm": 3.7222516536712646, "learning_rate": 4.494514017344792e-06, "loss": 1.0647, "step": 19750 }, { "epoch": 0.6959662429415858, "grad_norm": 1.9765844345092773, "learning_rate": 4.49356131177199e-06, "loss": 0.9376, "step": 19751 }, { "epoch": 0.6960014799545442, "grad_norm": 1.0267884731292725, "learning_rate": 4.492608677921253e-06, "loss": 0.9847, "step": 19752 }, { "epoch": 0.6960367169675027, "grad_norm": 3.8291678428649902, "learning_rate": 4.4916561158049895e-06, "loss": 3.1109, "step": 19753 }, { "epoch": 0.6960719539804611, "grad_norm": 2.5848429203033447, "learning_rate": 4.49070362543561e-06, "loss": 2.6599, "step": 19754 }, { "epoch": 0.6961071909934194, "grad_norm": 10.193388938903809, "learning_rate": 4.489751206825512e-06, "loss": 4.8302, "step": 19755 }, { "epoch": 0.6961424280063779, "grad_norm": 2.7056310176849365, "learning_rate": 4.488798859987109e-06, "loss": 2.4189, "step": 19756 }, { "epoch": 0.6961776650193363, "grad_norm": 1.2246603965759277, "learning_rate": 4.487846584932806e-06, "loss": 1.0765, "step": 19757 }, { "epoch": 0.6962129020322947, "grad_norm": 3.6904516220092773, "learning_rate": 4.486894381675e-06, "loss": 2.8891, "step": 19758 }, { "epoch": 0.6962481390452532, "grad_norm": 1.4934197664260864, "learning_rate": 4.485942250226094e-06, "loss": 0.936, "step": 19759 }, { "epoch": 0.6962833760582116, "grad_norm": 4.1311354637146, "learning_rate": 4.4849901905984995e-06, "loss": 2.6309, "step": 19760 }, { "epoch": 0.6963186130711699, "grad_norm": 5.263419151306152, "learning_rate": 4.484038202804606e-06, "loss": 4.7198, "step": 19761 }, { "epoch": 0.6963538500841284, "grad_norm": 4.397795677185059, "learning_rate": 4.483086286856818e-06, "loss": 3.2479, "step": 19762 }, { "epoch": 0.6963890870970868, "grad_norm": 1.2413966655731201, "learning_rate": 4.4821344427675305e-06, "loss": 0.8431, "step": 19763 }, { "epoch": 0.6964243241100452, "grad_norm": 2.880315065383911, "learning_rate": 4.481182670549152e-06, "loss": 2.6088, "step": 19764 }, { "epoch": 0.6964595611230036, "grad_norm": 1.6832174062728882, "learning_rate": 4.4802309702140675e-06, "loss": 0.7015, "step": 19765 }, { "epoch": 0.696494798135962, "grad_norm": 4.660118579864502, "learning_rate": 4.479279341774679e-06, "loss": 2.6009, "step": 19766 }, { "epoch": 0.6965300351489204, "grad_norm": 3.2590672969818115, "learning_rate": 4.478327785243381e-06, "loss": 2.3362, "step": 19767 }, { "epoch": 0.6965652721618788, "grad_norm": 4.460403919219971, "learning_rate": 4.477376300632565e-06, "loss": 2.9478, "step": 19768 }, { "epoch": 0.6966005091748373, "grad_norm": 3.0969371795654297, "learning_rate": 4.476424887954627e-06, "loss": 2.7472, "step": 19769 }, { "epoch": 0.6966357461877957, "grad_norm": 9.3555326461792, "learning_rate": 4.475473547221958e-06, "loss": 3.3715, "step": 19770 }, { "epoch": 0.696670983200754, "grad_norm": 3.2589926719665527, "learning_rate": 4.47452227844695e-06, "loss": 2.8537, "step": 19771 }, { "epoch": 0.6967062202137125, "grad_norm": 8.47890853881836, "learning_rate": 4.473571081641993e-06, "loss": 5.1303, "step": 19772 }, { "epoch": 0.6967414572266709, "grad_norm": 3.520045280456543, "learning_rate": 4.472619956819481e-06, "loss": 3.312, "step": 19773 }, { "epoch": 0.6967766942396293, "grad_norm": 3.7706174850463867, "learning_rate": 4.47166890399179e-06, "loss": 3.3249, "step": 19774 }, { "epoch": 0.6968119312525877, "grad_norm": 7.278682708740234, "learning_rate": 4.47071792317132e-06, "loss": 5.3632, "step": 19775 }, { "epoch": 0.6968471682655462, "grad_norm": 1.181211233139038, "learning_rate": 4.4697670143704574e-06, "loss": 0.7293, "step": 19776 }, { "epoch": 0.6968824052785045, "grad_norm": 1.0001707077026367, "learning_rate": 4.468816177601579e-06, "loss": 0.9379, "step": 19777 }, { "epoch": 0.6969176422914629, "grad_norm": 1.0752480030059814, "learning_rate": 4.467865412877072e-06, "loss": 1.0067, "step": 19778 }, { "epoch": 0.6969528793044214, "grad_norm": 1.5809712409973145, "learning_rate": 4.466914720209325e-06, "loss": 1.2484, "step": 19779 }, { "epoch": 0.6969881163173798, "grad_norm": 2.1653435230255127, "learning_rate": 4.465964099610723e-06, "loss": 0.918, "step": 19780 }, { "epoch": 0.6970233533303382, "grad_norm": 3.0437376499176025, "learning_rate": 4.46501355109364e-06, "loss": 2.9523, "step": 19781 }, { "epoch": 0.6970585903432966, "grad_norm": 4.667918682098389, "learning_rate": 4.464063074670457e-06, "loss": 3.0141, "step": 19782 }, { "epoch": 0.697093827356255, "grad_norm": 5.111334323883057, "learning_rate": 4.463112670353564e-06, "loss": 5.4076, "step": 19783 }, { "epoch": 0.6971290643692134, "grad_norm": 6.8296732902526855, "learning_rate": 4.462162338155331e-06, "loss": 4.9425, "step": 19784 }, { "epoch": 0.6971643013821718, "grad_norm": 2.881274700164795, "learning_rate": 4.461212078088139e-06, "loss": 2.8788, "step": 19785 }, { "epoch": 0.6971995383951303, "grad_norm": 1.5177363157272339, "learning_rate": 4.460261890164365e-06, "loss": 0.8148, "step": 19786 }, { "epoch": 0.6972347754080886, "grad_norm": 2.715416193008423, "learning_rate": 4.4593117743963855e-06, "loss": 2.2355, "step": 19787 }, { "epoch": 0.697270012421047, "grad_norm": 5.2556257247924805, "learning_rate": 4.4583617307965754e-06, "loss": 6.3943, "step": 19788 }, { "epoch": 0.6973052494340055, "grad_norm": 1.224481463432312, "learning_rate": 4.457411759377311e-06, "loss": 1.0434, "step": 19789 }, { "epoch": 0.6973404864469639, "grad_norm": 1.195119857788086, "learning_rate": 4.456461860150964e-06, "loss": 0.7504, "step": 19790 }, { "epoch": 0.6973757234599223, "grad_norm": 3.407444477081299, "learning_rate": 4.455512033129906e-06, "loss": 2.421, "step": 19791 }, { "epoch": 0.6974109604728808, "grad_norm": 1.3372373580932617, "learning_rate": 4.4545622783265145e-06, "loss": 0.7111, "step": 19792 }, { "epoch": 0.6974461974858391, "grad_norm": 1.6617804765701294, "learning_rate": 4.453612595753149e-06, "loss": 0.991, "step": 19793 }, { "epoch": 0.6974814344987975, "grad_norm": 1.4290742874145508, "learning_rate": 4.45266298542219e-06, "loss": 0.888, "step": 19794 }, { "epoch": 0.697516671511756, "grad_norm": 3.969475269317627, "learning_rate": 4.451713447346003e-06, "loss": 2.7875, "step": 19795 }, { "epoch": 0.6975519085247144, "grad_norm": 4.194002628326416, "learning_rate": 4.450763981536954e-06, "loss": 3.538, "step": 19796 }, { "epoch": 0.6975871455376728, "grad_norm": 4.916006088256836, "learning_rate": 4.4498145880074074e-06, "loss": 3.6145, "step": 19797 }, { "epoch": 0.6976223825506311, "grad_norm": 3.6600213050842285, "learning_rate": 4.448865266769735e-06, "loss": 3.2315, "step": 19798 }, { "epoch": 0.6976576195635896, "grad_norm": 1.459743857383728, "learning_rate": 4.447916017836305e-06, "loss": 0.9451, "step": 19799 }, { "epoch": 0.697692856576548, "grad_norm": 1.4031943082809448, "learning_rate": 4.446966841219471e-06, "loss": 0.8477, "step": 19800 }, { "epoch": 0.6977280935895064, "grad_norm": 1.2911477088928223, "learning_rate": 4.4460177369315984e-06, "loss": 0.8679, "step": 19801 }, { "epoch": 0.6977633306024649, "grad_norm": 1.0093486309051514, "learning_rate": 4.4450687049850595e-06, "loss": 0.9003, "step": 19802 }, { "epoch": 0.6977985676154232, "grad_norm": 4.991230010986328, "learning_rate": 4.444119745392204e-06, "loss": 5.6378, "step": 19803 }, { "epoch": 0.6978338046283816, "grad_norm": 2.336510419845581, "learning_rate": 4.443170858165397e-06, "loss": 2.7414, "step": 19804 }, { "epoch": 0.6978690416413401, "grad_norm": 2.7095730304718018, "learning_rate": 4.442222043316997e-06, "loss": 3.005, "step": 19805 }, { "epoch": 0.6979042786542985, "grad_norm": 4.536082744598389, "learning_rate": 4.4412733008593635e-06, "loss": 2.7259, "step": 19806 }, { "epoch": 0.6979395156672569, "grad_norm": 8.34624195098877, "learning_rate": 4.4403246308048535e-06, "loss": 5.0793, "step": 19807 }, { "epoch": 0.6979747526802152, "grad_norm": 1.3745499849319458, "learning_rate": 4.439376033165822e-06, "loss": 0.6454, "step": 19808 }, { "epoch": 0.6980099896931737, "grad_norm": 2.7184598445892334, "learning_rate": 4.438427507954627e-06, "loss": 2.2239, "step": 19809 }, { "epoch": 0.6980452267061321, "grad_norm": 4.78683614730835, "learning_rate": 4.4374790551836226e-06, "loss": 3.0014, "step": 19810 }, { "epoch": 0.6980804637190905, "grad_norm": 4.170346736907959, "learning_rate": 4.436530674865165e-06, "loss": 2.747, "step": 19811 }, { "epoch": 0.698115700732049, "grad_norm": 1.3906813859939575, "learning_rate": 4.435582367011596e-06, "loss": 1.0275, "step": 19812 }, { "epoch": 0.6981509377450074, "grad_norm": 0.9532615542411804, "learning_rate": 4.434634131635281e-06, "loss": 1.0109, "step": 19813 }, { "epoch": 0.6981861747579657, "grad_norm": 5.464015483856201, "learning_rate": 4.4336859687485666e-06, "loss": 4.8401, "step": 19814 }, { "epoch": 0.6982214117709242, "grad_norm": 1.080457091331482, "learning_rate": 4.4327378783638e-06, "loss": 0.7563, "step": 19815 }, { "epoch": 0.6982566487838826, "grad_norm": 1.0523146390914917, "learning_rate": 4.431789860493326e-06, "loss": 0.8829, "step": 19816 }, { "epoch": 0.698291885796841, "grad_norm": 4.0455451011657715, "learning_rate": 4.430841915149504e-06, "loss": 2.9364, "step": 19817 }, { "epoch": 0.6983271228097994, "grad_norm": 0.9882156848907471, "learning_rate": 4.429894042344678e-06, "loss": 0.9303, "step": 19818 }, { "epoch": 0.6983623598227578, "grad_norm": 3.756302833557129, "learning_rate": 4.428946242091189e-06, "loss": 4.2681, "step": 19819 }, { "epoch": 0.6983975968357162, "grad_norm": 1.2302528619766235, "learning_rate": 4.42799851440138e-06, "loss": 0.9391, "step": 19820 }, { "epoch": 0.6984328338486746, "grad_norm": 1.9252665042877197, "learning_rate": 4.427050859287608e-06, "loss": 1.0033, "step": 19821 }, { "epoch": 0.6984680708616331, "grad_norm": 1.1736252307891846, "learning_rate": 4.426103276762205e-06, "loss": 0.8195, "step": 19822 }, { "epoch": 0.6985033078745915, "grad_norm": 0.9885103702545166, "learning_rate": 4.425155766837516e-06, "loss": 0.6593, "step": 19823 }, { "epoch": 0.6985385448875499, "grad_norm": 4.330965518951416, "learning_rate": 4.424208329525883e-06, "loss": 4.4587, "step": 19824 }, { "epoch": 0.6985737819005083, "grad_norm": 1.2177129983901978, "learning_rate": 4.4232609648396484e-06, "loss": 0.842, "step": 19825 }, { "epoch": 0.6986090189134667, "grad_norm": 2.945305585861206, "learning_rate": 4.422313672791149e-06, "loss": 2.9257, "step": 19826 }, { "epoch": 0.6986442559264251, "grad_norm": 5.524040222167969, "learning_rate": 4.421366453392725e-06, "loss": 4.5742, "step": 19827 }, { "epoch": 0.6986794929393835, "grad_norm": 3.813304901123047, "learning_rate": 4.420419306656713e-06, "loss": 4.083, "step": 19828 }, { "epoch": 0.698714729952342, "grad_norm": 6.973010540008545, "learning_rate": 4.41947223259545e-06, "loss": 7.1081, "step": 19829 }, { "epoch": 0.6987499669653003, "grad_norm": 3.9536664485931396, "learning_rate": 4.418525231221277e-06, "loss": 2.6572, "step": 19830 }, { "epoch": 0.6987852039782587, "grad_norm": 1.3917534351348877, "learning_rate": 4.417578302546514e-06, "loss": 0.8057, "step": 19831 }, { "epoch": 0.6988204409912172, "grad_norm": 3.385222911834717, "learning_rate": 4.416631446583512e-06, "loss": 3.2789, "step": 19832 }, { "epoch": 0.6988556780041756, "grad_norm": 0.9768375754356384, "learning_rate": 4.4156846633445974e-06, "loss": 0.6427, "step": 19833 }, { "epoch": 0.698890915017134, "grad_norm": 1.5185633897781372, "learning_rate": 4.414737952842099e-06, "loss": 1.0166, "step": 19834 }, { "epoch": 0.6989261520300925, "grad_norm": 1.3220583200454712, "learning_rate": 4.413791315088346e-06, "loss": 0.6412, "step": 19835 }, { "epoch": 0.6989613890430508, "grad_norm": 4.758223056793213, "learning_rate": 4.412844750095676e-06, "loss": 3.2246, "step": 19836 }, { "epoch": 0.6989966260560092, "grad_norm": 0.9818735718727112, "learning_rate": 4.411898257876419e-06, "loss": 1.0805, "step": 19837 }, { "epoch": 0.6990318630689677, "grad_norm": 4.216704845428467, "learning_rate": 4.410951838442896e-06, "loss": 2.7852, "step": 19838 }, { "epoch": 0.6990671000819261, "grad_norm": 1.2194921970367432, "learning_rate": 4.410005491807433e-06, "loss": 0.8092, "step": 19839 }, { "epoch": 0.6991023370948845, "grad_norm": 4.980133533477783, "learning_rate": 4.409059217982368e-06, "loss": 2.7495, "step": 19840 }, { "epoch": 0.6991375741078428, "grad_norm": 3.0916526317596436, "learning_rate": 4.4081130169800146e-06, "loss": 3.2472, "step": 19841 }, { "epoch": 0.6991728111208013, "grad_norm": 4.8168625831604, "learning_rate": 4.407166888812702e-06, "loss": 2.9085, "step": 19842 }, { "epoch": 0.6992080481337597, "grad_norm": 3.293368339538574, "learning_rate": 4.406220833492753e-06, "loss": 2.7064, "step": 19843 }, { "epoch": 0.6992432851467181, "grad_norm": 3.297492504119873, "learning_rate": 4.405274851032489e-06, "loss": 3.0612, "step": 19844 }, { "epoch": 0.6992785221596766, "grad_norm": 1.2573744058609009, "learning_rate": 4.404328941444235e-06, "loss": 0.8265, "step": 19845 }, { "epoch": 0.6993137591726349, "grad_norm": 4.795062065124512, "learning_rate": 4.403383104740307e-06, "loss": 3.3469, "step": 19846 }, { "epoch": 0.6993489961855933, "grad_norm": 5.246625900268555, "learning_rate": 4.402437340933029e-06, "loss": 2.4242, "step": 19847 }, { "epoch": 0.6993842331985518, "grad_norm": 4.404265403747559, "learning_rate": 4.401491650034716e-06, "loss": 2.4083, "step": 19848 }, { "epoch": 0.6994194702115102, "grad_norm": 4.965662479400635, "learning_rate": 4.400546032057692e-06, "loss": 4.6753, "step": 19849 }, { "epoch": 0.6994547072244686, "grad_norm": 7.116086006164551, "learning_rate": 4.3996004870142614e-06, "loss": 7.3053, "step": 19850 }, { "epoch": 0.6994899442374269, "grad_norm": 4.35141134262085, "learning_rate": 4.398655014916752e-06, "loss": 4.5559, "step": 19851 }, { "epoch": 0.6995251812503854, "grad_norm": 3.5621397495269775, "learning_rate": 4.397709615777475e-06, "loss": 2.6508, "step": 19852 }, { "epoch": 0.6995604182633438, "grad_norm": 1.491742491722107, "learning_rate": 4.396764289608746e-06, "loss": 1.0398, "step": 19853 }, { "epoch": 0.6995956552763022, "grad_norm": 1.3150163888931274, "learning_rate": 4.395819036422869e-06, "loss": 0.7625, "step": 19854 }, { "epoch": 0.6996308922892607, "grad_norm": 2.869861364364624, "learning_rate": 4.394873856232167e-06, "loss": 2.9402, "step": 19855 }, { "epoch": 0.699666129302219, "grad_norm": 5.212666034698486, "learning_rate": 4.393928749048951e-06, "loss": 5.0654, "step": 19856 }, { "epoch": 0.6997013663151774, "grad_norm": 0.819798469543457, "learning_rate": 4.392983714885522e-06, "loss": 0.9694, "step": 19857 }, { "epoch": 0.6997366033281359, "grad_norm": 2.2114365100860596, "learning_rate": 4.392038753754193e-06, "loss": 0.72, "step": 19858 }, { "epoch": 0.6997718403410943, "grad_norm": 4.120842456817627, "learning_rate": 4.391093865667279e-06, "loss": 2.7536, "step": 19859 }, { "epoch": 0.6998070773540527, "grad_norm": 2.2689762115478516, "learning_rate": 4.390149050637079e-06, "loss": 0.8709, "step": 19860 }, { "epoch": 0.699842314367011, "grad_norm": 0.9724084734916687, "learning_rate": 4.389204308675902e-06, "loss": 0.7222, "step": 19861 }, { "epoch": 0.6998775513799695, "grad_norm": 4.737201690673828, "learning_rate": 4.388259639796054e-06, "loss": 2.6588, "step": 19862 }, { "epoch": 0.6999127883929279, "grad_norm": 4.490013122558594, "learning_rate": 4.387315044009838e-06, "loss": 4.663, "step": 19863 }, { "epoch": 0.6999480254058863, "grad_norm": 5.039369106292725, "learning_rate": 4.3863705213295595e-06, "loss": 3.7628, "step": 19864 }, { "epoch": 0.6999832624188448, "grad_norm": 3.8028299808502197, "learning_rate": 4.385426071767519e-06, "loss": 2.5598, "step": 19865 }, { "epoch": 0.7000184994318032, "grad_norm": 1.2493896484375, "learning_rate": 4.38448169533602e-06, "loss": 1.0392, "step": 19866 }, { "epoch": 0.7000537364447615, "grad_norm": 4.664955139160156, "learning_rate": 4.3835373920473624e-06, "loss": 3.0564, "step": 19867 }, { "epoch": 0.70008897345772, "grad_norm": 3.3456406593322754, "learning_rate": 4.382593161913848e-06, "loss": 2.7575, "step": 19868 }, { "epoch": 0.7001242104706784, "grad_norm": 1.0970391035079956, "learning_rate": 4.381649004947768e-06, "loss": 0.8172, "step": 19869 }, { "epoch": 0.7001594474836368, "grad_norm": 0.8647969365119934, "learning_rate": 4.380704921161427e-06, "loss": 0.6786, "step": 19870 }, { "epoch": 0.7001946844965953, "grad_norm": 3.9369750022888184, "learning_rate": 4.3797609105671214e-06, "loss": 3.1798, "step": 19871 }, { "epoch": 0.7002299215095537, "grad_norm": 1.0926868915557861, "learning_rate": 4.378816973177149e-06, "loss": 1.0312, "step": 19872 }, { "epoch": 0.700265158522512, "grad_norm": 2.7697577476501465, "learning_rate": 4.377873109003794e-06, "loss": 3.1045, "step": 19873 }, { "epoch": 0.7003003955354704, "grad_norm": 5.132143974304199, "learning_rate": 4.376929318059362e-06, "loss": 5.3193, "step": 19874 }, { "epoch": 0.7003356325484289, "grad_norm": 3.8763718605041504, "learning_rate": 4.375985600356146e-06, "loss": 3.2145, "step": 19875 }, { "epoch": 0.7003708695613873, "grad_norm": 2.374495506286621, "learning_rate": 4.37504195590643e-06, "loss": 0.6211, "step": 19876 }, { "epoch": 0.7004061065743457, "grad_norm": 16.22773551940918, "learning_rate": 4.374098384722505e-06, "loss": 3.073, "step": 19877 }, { "epoch": 0.7004413435873041, "grad_norm": 0.9110894799232483, "learning_rate": 4.373154886816672e-06, "loss": 0.8572, "step": 19878 }, { "epoch": 0.7004765806002625, "grad_norm": 2.8796114921569824, "learning_rate": 4.37221146220121e-06, "loss": 1.0616, "step": 19879 }, { "epoch": 0.7005118176132209, "grad_norm": 1.268554925918579, "learning_rate": 4.371268110888411e-06, "loss": 0.8338, "step": 19880 }, { "epoch": 0.7005470546261794, "grad_norm": 5.0062174797058105, "learning_rate": 4.370324832890563e-06, "loss": 3.3063, "step": 19881 }, { "epoch": 0.7005822916391378, "grad_norm": 1.2468057870864868, "learning_rate": 4.369381628219949e-06, "loss": 1.1313, "step": 19882 }, { "epoch": 0.7006175286520961, "grad_norm": 3.857738733291626, "learning_rate": 4.368438496888857e-06, "loss": 2.687, "step": 19883 }, { "epoch": 0.7006527656650545, "grad_norm": 0.9616125822067261, "learning_rate": 4.367495438909572e-06, "loss": 1.1457, "step": 19884 }, { "epoch": 0.700688002678013, "grad_norm": 4.656718730926514, "learning_rate": 4.366552454294375e-06, "loss": 3.3793, "step": 19885 }, { "epoch": 0.7007232396909714, "grad_norm": 4.487571716308594, "learning_rate": 4.3656095430555516e-06, "loss": 5.0891, "step": 19886 }, { "epoch": 0.7007584767039298, "grad_norm": 5.867691516876221, "learning_rate": 4.364666705205384e-06, "loss": 7.3193, "step": 19887 }, { "epoch": 0.7007937137168883, "grad_norm": 5.840888023376465, "learning_rate": 4.363723940756142e-06, "loss": 6.9403, "step": 19888 }, { "epoch": 0.7008289507298466, "grad_norm": 0.8925896286964417, "learning_rate": 4.3627812497201195e-06, "loss": 0.6082, "step": 19889 }, { "epoch": 0.700864187742805, "grad_norm": 3.519813299179077, "learning_rate": 4.361838632109588e-06, "loss": 2.9699, "step": 19890 }, { "epoch": 0.7008994247557635, "grad_norm": 1.3837112188339233, "learning_rate": 4.36089608793683e-06, "loss": 0.8923, "step": 19891 }, { "epoch": 0.7009346617687219, "grad_norm": 1.0754886865615845, "learning_rate": 4.3599536172141125e-06, "loss": 0.7096, "step": 19892 }, { "epoch": 0.7009698987816803, "grad_norm": 1.2628620862960815, "learning_rate": 4.35901121995372e-06, "loss": 0.7624, "step": 19893 }, { "epoch": 0.7010051357946386, "grad_norm": 1.2137994766235352, "learning_rate": 4.35806889616793e-06, "loss": 0.7538, "step": 19894 }, { "epoch": 0.7010403728075971, "grad_norm": 4.4823808670043945, "learning_rate": 4.357126645869007e-06, "loss": 4.4487, "step": 19895 }, { "epoch": 0.7010756098205555, "grad_norm": 1.4444420337677002, "learning_rate": 4.356184469069228e-06, "loss": 0.8656, "step": 19896 }, { "epoch": 0.7011108468335139, "grad_norm": 1.6129909753799438, "learning_rate": 4.355242365780865e-06, "loss": 0.9972, "step": 19897 }, { "epoch": 0.7011460838464724, "grad_norm": 4.224483966827393, "learning_rate": 4.35430033601619e-06, "loss": 3.1287, "step": 19898 }, { "epoch": 0.7011813208594307, "grad_norm": 5.874544620513916, "learning_rate": 4.353358379787471e-06, "loss": 5.7297, "step": 19899 }, { "epoch": 0.7012165578723891, "grad_norm": 3.178290367126465, "learning_rate": 4.3524164971069795e-06, "loss": 0.7816, "step": 19900 }, { "epoch": 0.7012517948853476, "grad_norm": 2.641069173812866, "learning_rate": 4.351474687986983e-06, "loss": 0.8598, "step": 19901 }, { "epoch": 0.701287031898306, "grad_norm": 1.0083608627319336, "learning_rate": 4.350532952439747e-06, "loss": 1.0447, "step": 19902 }, { "epoch": 0.7013222689112644, "grad_norm": 1.5920000076293945, "learning_rate": 4.34959129047754e-06, "loss": 0.8184, "step": 19903 }, { "epoch": 0.7013575059242229, "grad_norm": 6.133429050445557, "learning_rate": 4.348649702112625e-06, "loss": 3.6109, "step": 19904 }, { "epoch": 0.7013927429371812, "grad_norm": 4.139997959136963, "learning_rate": 4.3477081873572675e-06, "loss": 5.3943, "step": 19905 }, { "epoch": 0.7014279799501396, "grad_norm": 1.1404200792312622, "learning_rate": 4.346766746223735e-06, "loss": 0.9519, "step": 19906 }, { "epoch": 0.701463216963098, "grad_norm": 1.6325550079345703, "learning_rate": 4.345825378724282e-06, "loss": 0.889, "step": 19907 }, { "epoch": 0.7014984539760565, "grad_norm": 2.7649805545806885, "learning_rate": 4.344884084871171e-06, "loss": 2.587, "step": 19908 }, { "epoch": 0.7015336909890149, "grad_norm": 1.0278565883636475, "learning_rate": 4.343942864676667e-06, "loss": 0.9646, "step": 19909 }, { "epoch": 0.7015689280019732, "grad_norm": 2.9831624031066895, "learning_rate": 4.343001718153033e-06, "loss": 2.6955, "step": 19910 }, { "epoch": 0.7016041650149317, "grad_norm": 7.742787837982178, "learning_rate": 4.342060645312517e-06, "loss": 4.7551, "step": 19911 }, { "epoch": 0.7016394020278901, "grad_norm": 1.6903584003448486, "learning_rate": 4.341119646167379e-06, "loss": 0.6372, "step": 19912 }, { "epoch": 0.7016746390408485, "grad_norm": 1.035820484161377, "learning_rate": 4.340178720729885e-06, "loss": 1.0261, "step": 19913 }, { "epoch": 0.701709876053807, "grad_norm": 3.5171244144439697, "learning_rate": 4.339237869012282e-06, "loss": 2.941, "step": 19914 }, { "epoch": 0.7017451130667653, "grad_norm": 3.6603829860687256, "learning_rate": 4.338297091026825e-06, "loss": 3.1561, "step": 19915 }, { "epoch": 0.7017803500797237, "grad_norm": 3.169811725616455, "learning_rate": 4.337356386785771e-06, "loss": 2.9576, "step": 19916 }, { "epoch": 0.7018155870926821, "grad_norm": 3.6384878158569336, "learning_rate": 4.336415756301371e-06, "loss": 2.8764, "step": 19917 }, { "epoch": 0.7018508241056406, "grad_norm": 2.177619218826294, "learning_rate": 4.335475199585877e-06, "loss": 2.8735, "step": 19918 }, { "epoch": 0.701886061118599, "grad_norm": 0.8965924978256226, "learning_rate": 4.334534716651539e-06, "loss": 0.8554, "step": 19919 }, { "epoch": 0.7019212981315573, "grad_norm": 4.298486709594727, "learning_rate": 4.33359430751061e-06, "loss": 2.9275, "step": 19920 }, { "epoch": 0.7019565351445158, "grad_norm": 0.9223489165306091, "learning_rate": 4.332653972175336e-06, "loss": 0.8931, "step": 19921 }, { "epoch": 0.7019917721574742, "grad_norm": 2.608827829360962, "learning_rate": 4.33171371065797e-06, "loss": 2.9698, "step": 19922 }, { "epoch": 0.7020270091704326, "grad_norm": 2.738236427307129, "learning_rate": 4.330773522970747e-06, "loss": 1.9931, "step": 19923 }, { "epoch": 0.7020622461833911, "grad_norm": 1.5473191738128662, "learning_rate": 4.329833409125924e-06, "loss": 1.0048, "step": 19924 }, { "epoch": 0.7020974831963495, "grad_norm": 1.3371248245239258, "learning_rate": 4.328893369135745e-06, "loss": 0.7565, "step": 19925 }, { "epoch": 0.7021327202093078, "grad_norm": 3.4211952686309814, "learning_rate": 4.327953403012455e-06, "loss": 2.8807, "step": 19926 }, { "epoch": 0.7021679572222662, "grad_norm": 1.31026291847229, "learning_rate": 4.327013510768286e-06, "loss": 0.8863, "step": 19927 }, { "epoch": 0.7022031942352247, "grad_norm": 1.39634370803833, "learning_rate": 4.326073692415494e-06, "loss": 0.7729, "step": 19928 }, { "epoch": 0.7022384312481831, "grad_norm": 0.991188108921051, "learning_rate": 4.325133947966318e-06, "loss": 0.6803, "step": 19929 }, { "epoch": 0.7022736682611415, "grad_norm": 4.796958923339844, "learning_rate": 4.324194277432992e-06, "loss": 2.1951, "step": 19930 }, { "epoch": 0.7023089052741, "grad_norm": 1.1551787853240967, "learning_rate": 4.323254680827753e-06, "loss": 1.1301, "step": 19931 }, { "epoch": 0.7023441422870583, "grad_norm": 2.6398682594299316, "learning_rate": 4.322315158162854e-06, "loss": 2.597, "step": 19932 }, { "epoch": 0.7023793793000167, "grad_norm": 1.0478944778442383, "learning_rate": 4.32137570945052e-06, "loss": 0.8885, "step": 19933 }, { "epoch": 0.7024146163129752, "grad_norm": 1.0569679737091064, "learning_rate": 4.320436334702989e-06, "loss": 0.7218, "step": 19934 }, { "epoch": 0.7024498533259336, "grad_norm": 1.410117745399475, "learning_rate": 4.319497033932499e-06, "loss": 0.7795, "step": 19935 }, { "epoch": 0.702485090338892, "grad_norm": 1.108232855796814, "learning_rate": 4.318557807151284e-06, "loss": 1.1261, "step": 19936 }, { "epoch": 0.7025203273518504, "grad_norm": 1.7214328050613403, "learning_rate": 4.317618654371577e-06, "loss": 0.7807, "step": 19937 }, { "epoch": 0.7025555643648088, "grad_norm": 6.141653537750244, "learning_rate": 4.3166795756056114e-06, "loss": 5.1683, "step": 19938 }, { "epoch": 0.7025908013777672, "grad_norm": 2.1165592670440674, "learning_rate": 4.315740570865617e-06, "loss": 0.862, "step": 19939 }, { "epoch": 0.7026260383907256, "grad_norm": 1.2195512056350708, "learning_rate": 4.314801640163826e-06, "loss": 0.9728, "step": 19940 }, { "epoch": 0.7026612754036841, "grad_norm": 8.991312026977539, "learning_rate": 4.313862783512472e-06, "loss": 2.8606, "step": 19941 }, { "epoch": 0.7026965124166424, "grad_norm": 4.693901538848877, "learning_rate": 4.312924000923773e-06, "loss": 5.1719, "step": 19942 }, { "epoch": 0.7027317494296008, "grad_norm": 1.337369680404663, "learning_rate": 4.311985292409966e-06, "loss": 1.141, "step": 19943 }, { "epoch": 0.7027669864425593, "grad_norm": 2.6699113845825195, "learning_rate": 4.311046657983274e-06, "loss": 2.9581, "step": 19944 }, { "epoch": 0.7028022234555177, "grad_norm": 4.350414752960205, "learning_rate": 4.310108097655929e-06, "loss": 2.9, "step": 19945 }, { "epoch": 0.7028374604684761, "grad_norm": 1.259865164756775, "learning_rate": 4.309169611440144e-06, "loss": 0.7892, "step": 19946 }, { "epoch": 0.7028726974814345, "grad_norm": 4.485861301422119, "learning_rate": 4.308231199348153e-06, "loss": 2.7587, "step": 19947 }, { "epoch": 0.7029079344943929, "grad_norm": 5.9686784744262695, "learning_rate": 4.3072928613921785e-06, "loss": 5.0343, "step": 19948 }, { "epoch": 0.7029431715073513, "grad_norm": 1.2777068614959717, "learning_rate": 4.306354597584437e-06, "loss": 0.884, "step": 19949 }, { "epoch": 0.7029784085203097, "grad_norm": 1.1766084432601929, "learning_rate": 4.305416407937147e-06, "loss": 0.8431, "step": 19950 }, { "epoch": 0.7030136455332682, "grad_norm": 1.4192626476287842, "learning_rate": 4.304478292462543e-06, "loss": 0.8293, "step": 19951 }, { "epoch": 0.7030488825462266, "grad_norm": 4.244843482971191, "learning_rate": 4.303540251172829e-06, "loss": 2.5808, "step": 19952 }, { "epoch": 0.7030841195591849, "grad_norm": 2.550724506378174, "learning_rate": 4.3026022840802305e-06, "loss": 2.9836, "step": 19953 }, { "epoch": 0.7031193565721434, "grad_norm": 2.7827863693237305, "learning_rate": 4.301664391196961e-06, "loss": 2.8345, "step": 19954 }, { "epoch": 0.7031545935851018, "grad_norm": 1.47697913646698, "learning_rate": 4.3007265725352405e-06, "loss": 0.6966, "step": 19955 }, { "epoch": 0.7031898305980602, "grad_norm": 4.0123610496521, "learning_rate": 4.299788828107282e-06, "loss": 1.0792, "step": 19956 }, { "epoch": 0.7032250676110187, "grad_norm": 1.8912503719329834, "learning_rate": 4.2988511579253e-06, "loss": 0.9476, "step": 19957 }, { "epoch": 0.703260304623977, "grad_norm": 7.608402729034424, "learning_rate": 4.2979135620015075e-06, "loss": 2.8867, "step": 19958 }, { "epoch": 0.7032955416369354, "grad_norm": 5.321897506713867, "learning_rate": 4.296976040348118e-06, "loss": 3.4386, "step": 19959 }, { "epoch": 0.7033307786498938, "grad_norm": 4.304455757141113, "learning_rate": 4.296038592977345e-06, "loss": 4.3821, "step": 19960 }, { "epoch": 0.7033660156628523, "grad_norm": 6.377869606018066, "learning_rate": 4.295101219901389e-06, "loss": 5.2515, "step": 19961 }, { "epoch": 0.7034012526758107, "grad_norm": 4.282380104064941, "learning_rate": 4.294163921132469e-06, "loss": 4.9096, "step": 19962 }, { "epoch": 0.703436489688769, "grad_norm": 6.144049644470215, "learning_rate": 4.293226696682792e-06, "loss": 3.2398, "step": 19963 }, { "epoch": 0.7034717267017275, "grad_norm": 2.0563297271728516, "learning_rate": 4.2922895465645665e-06, "loss": 0.7056, "step": 19964 }, { "epoch": 0.7035069637146859, "grad_norm": 3.395146131515503, "learning_rate": 4.291352470789991e-06, "loss": 2.9775, "step": 19965 }, { "epoch": 0.7035422007276443, "grad_norm": 1.8431421518325806, "learning_rate": 4.29041546937128e-06, "loss": 0.7858, "step": 19966 }, { "epoch": 0.7035774377406028, "grad_norm": 1.0041991472244263, "learning_rate": 4.289478542320637e-06, "loss": 1.1217, "step": 19967 }, { "epoch": 0.7036126747535612, "grad_norm": 3.2540125846862793, "learning_rate": 4.288541689650261e-06, "loss": 3.0756, "step": 19968 }, { "epoch": 0.7036479117665195, "grad_norm": 1.4934921264648438, "learning_rate": 4.287604911372353e-06, "loss": 0.7507, "step": 19969 }, { "epoch": 0.703683148779478, "grad_norm": 1.157703161239624, "learning_rate": 4.286668207499126e-06, "loss": 0.9598, "step": 19970 }, { "epoch": 0.7037183857924364, "grad_norm": 4.251433372497559, "learning_rate": 4.2857315780427695e-06, "loss": 2.9003, "step": 19971 }, { "epoch": 0.7037536228053948, "grad_norm": 3.7014636993408203, "learning_rate": 4.2847950230154875e-06, "loss": 3.0755, "step": 19972 }, { "epoch": 0.7037888598183532, "grad_norm": 2.5381572246551514, "learning_rate": 4.283858542429479e-06, "loss": 2.9521, "step": 19973 }, { "epoch": 0.7038240968313116, "grad_norm": 7.8854498863220215, "learning_rate": 4.28292213629694e-06, "loss": 6.8852, "step": 19974 }, { "epoch": 0.70385933384427, "grad_norm": 1.1145564317703247, "learning_rate": 4.281985804630069e-06, "loss": 1.1099, "step": 19975 }, { "epoch": 0.7038945708572284, "grad_norm": 0.9885667562484741, "learning_rate": 4.281049547441061e-06, "loss": 0.8855, "step": 19976 }, { "epoch": 0.7039298078701869, "grad_norm": 1.1690665483474731, "learning_rate": 4.28011336474211e-06, "loss": 0.8815, "step": 19977 }, { "epoch": 0.7039650448831453, "grad_norm": 3.1532602310180664, "learning_rate": 4.279177256545412e-06, "loss": 3.0545, "step": 19978 }, { "epoch": 0.7040002818961036, "grad_norm": 3.6440179347991943, "learning_rate": 4.278241222863163e-06, "loss": 0.6235, "step": 19979 }, { "epoch": 0.7040355189090621, "grad_norm": 3.016857862472534, "learning_rate": 4.277305263707544e-06, "loss": 0.6413, "step": 19980 }, { "epoch": 0.7040707559220205, "grad_norm": 21.915681838989258, "learning_rate": 4.276369379090756e-06, "loss": 2.7685, "step": 19981 }, { "epoch": 0.7041059929349789, "grad_norm": 2.288606643676758, "learning_rate": 4.275433569024985e-06, "loss": 2.7056, "step": 19982 }, { "epoch": 0.7041412299479373, "grad_norm": 5.550734996795654, "learning_rate": 4.274497833522425e-06, "loss": 4.2617, "step": 19983 }, { "epoch": 0.7041764669608958, "grad_norm": 1.6660699844360352, "learning_rate": 4.273562172595254e-06, "loss": 0.8837, "step": 19984 }, { "epoch": 0.7042117039738541, "grad_norm": 4.765303611755371, "learning_rate": 4.2726265862556685e-06, "loss": 2.7447, "step": 19985 }, { "epoch": 0.7042469409868125, "grad_norm": 1.9330394268035889, "learning_rate": 4.271691074515854e-06, "loss": 0.9144, "step": 19986 }, { "epoch": 0.704282177999771, "grad_norm": 1.4330252408981323, "learning_rate": 4.270755637387989e-06, "loss": 0.8937, "step": 19987 }, { "epoch": 0.7043174150127294, "grad_norm": 1.1118404865264893, "learning_rate": 4.269820274884259e-06, "loss": 0.8815, "step": 19988 }, { "epoch": 0.7043526520256878, "grad_norm": 3.545308828353882, "learning_rate": 4.268884987016856e-06, "loss": 2.9911, "step": 19989 }, { "epoch": 0.7043878890386462, "grad_norm": 1.119351863861084, "learning_rate": 4.267949773797954e-06, "loss": 1.0795, "step": 19990 }, { "epoch": 0.7044231260516046, "grad_norm": 4.29395866394043, "learning_rate": 4.267014635239735e-06, "loss": 4.734, "step": 19991 }, { "epoch": 0.704458363064563, "grad_norm": 3.0069119930267334, "learning_rate": 4.266079571354381e-06, "loss": 2.8939, "step": 19992 }, { "epoch": 0.7044936000775214, "grad_norm": 0.9384123682975769, "learning_rate": 4.265144582154071e-06, "loss": 0.9699, "step": 19993 }, { "epoch": 0.7045288370904799, "grad_norm": 1.1877918243408203, "learning_rate": 4.264209667650984e-06, "loss": 0.6677, "step": 19994 }, { "epoch": 0.7045640741034382, "grad_norm": 1.1723229885101318, "learning_rate": 4.263274827857295e-06, "loss": 0.7282, "step": 19995 }, { "epoch": 0.7045993111163966, "grad_norm": 1.1560564041137695, "learning_rate": 4.262340062785184e-06, "loss": 0.9372, "step": 19996 }, { "epoch": 0.7046345481293551, "grad_norm": 1.0247689485549927, "learning_rate": 4.2614053724468226e-06, "loss": 1.0919, "step": 19997 }, { "epoch": 0.7046697851423135, "grad_norm": 3.320754051208496, "learning_rate": 4.260470756854389e-06, "loss": 3.0177, "step": 19998 }, { "epoch": 0.7047050221552719, "grad_norm": 1.4153894186019897, "learning_rate": 4.259536216020055e-06, "loss": 0.9614, "step": 19999 }, { "epoch": 0.7047402591682304, "grad_norm": 1.0446807146072388, "learning_rate": 4.2586017499559915e-06, "loss": 0.8802, "step": 20000 }, { "epoch": 0.7047754961811887, "grad_norm": 1.2715340852737427, "learning_rate": 4.257667358674372e-06, "loss": 0.9663, "step": 20001 }, { "epoch": 0.7048107331941471, "grad_norm": 4.44944953918457, "learning_rate": 4.256733042187371e-06, "loss": 2.5318, "step": 20002 }, { "epoch": 0.7048459702071056, "grad_norm": 1.2091728448867798, "learning_rate": 4.255798800507146e-06, "loss": 0.6995, "step": 20003 }, { "epoch": 0.704881207220064, "grad_norm": 0.9772961735725403, "learning_rate": 4.254864633645877e-06, "loss": 0.9715, "step": 20004 }, { "epoch": 0.7049164442330224, "grad_norm": 1.1974716186523438, "learning_rate": 4.253930541615732e-06, "loss": 0.8997, "step": 20005 }, { "epoch": 0.7049516812459807, "grad_norm": 3.199261426925659, "learning_rate": 4.252996524428868e-06, "loss": 2.3487, "step": 20006 }, { "epoch": 0.7049869182589392, "grad_norm": 6.164889335632324, "learning_rate": 4.252062582097455e-06, "loss": 2.8032, "step": 20007 }, { "epoch": 0.7050221552718976, "grad_norm": 3.259986400604248, "learning_rate": 4.251128714633666e-06, "loss": 3.4113, "step": 20008 }, { "epoch": 0.705057392284856, "grad_norm": 4.046184062957764, "learning_rate": 4.250194922049653e-06, "loss": 2.9847, "step": 20009 }, { "epoch": 0.7050926292978145, "grad_norm": 1.5449151992797852, "learning_rate": 4.249261204357584e-06, "loss": 0.9168, "step": 20010 }, { "epoch": 0.7051278663107728, "grad_norm": 1.0587079524993896, "learning_rate": 4.24832756156962e-06, "loss": 0.7281, "step": 20011 }, { "epoch": 0.7051631033237312, "grad_norm": 2.7812654972076416, "learning_rate": 4.247393993697924e-06, "loss": 0.9563, "step": 20012 }, { "epoch": 0.7051983403366897, "grad_norm": 2.4353318214416504, "learning_rate": 4.246460500754652e-06, "loss": 2.6029, "step": 20013 }, { "epoch": 0.7052335773496481, "grad_norm": 2.935309648513794, "learning_rate": 4.245527082751965e-06, "loss": 2.5093, "step": 20014 }, { "epoch": 0.7052688143626065, "grad_norm": 2.5218911170959473, "learning_rate": 4.244593739702022e-06, "loss": 1.0074, "step": 20015 }, { "epoch": 0.7053040513755648, "grad_norm": 1.1134812831878662, "learning_rate": 4.243660471616978e-06, "loss": 0.6898, "step": 20016 }, { "epoch": 0.7053392883885233, "grad_norm": 0.8851204514503479, "learning_rate": 4.242727278508989e-06, "loss": 0.6093, "step": 20017 }, { "epoch": 0.7053745254014817, "grad_norm": 5.392335891723633, "learning_rate": 4.241794160390211e-06, "loss": 2.9311, "step": 20018 }, { "epoch": 0.7054097624144401, "grad_norm": 2.2807693481445312, "learning_rate": 4.240861117272799e-06, "loss": 0.7502, "step": 20019 }, { "epoch": 0.7054449994273986, "grad_norm": 3.7414710521698, "learning_rate": 4.239928149168903e-06, "loss": 2.9467, "step": 20020 }, { "epoch": 0.705480236440357, "grad_norm": 4.686147212982178, "learning_rate": 4.23899525609068e-06, "loss": 3.1088, "step": 20021 }, { "epoch": 0.7055154734533153, "grad_norm": 6.157679080963135, "learning_rate": 4.238062438050271e-06, "loss": 3.1395, "step": 20022 }, { "epoch": 0.7055507104662738, "grad_norm": 3.119210958480835, "learning_rate": 4.237129695059836e-06, "loss": 2.9036, "step": 20023 }, { "epoch": 0.7055859474792322, "grad_norm": 3.8997933864593506, "learning_rate": 4.236197027131524e-06, "loss": 2.6624, "step": 20024 }, { "epoch": 0.7056211844921906, "grad_norm": 2.2446460723876953, "learning_rate": 4.235264434277477e-06, "loss": 0.9086, "step": 20025 }, { "epoch": 0.705656421505149, "grad_norm": 5.579392433166504, "learning_rate": 4.234331916509841e-06, "loss": 5.3557, "step": 20026 }, { "epoch": 0.7056916585181074, "grad_norm": 1.3707020282745361, "learning_rate": 4.2333994738407735e-06, "loss": 0.6359, "step": 20027 }, { "epoch": 0.7057268955310658, "grad_norm": 5.405124664306641, "learning_rate": 4.23246710628241e-06, "loss": 2.5482, "step": 20028 }, { "epoch": 0.7057621325440242, "grad_norm": 3.7580788135528564, "learning_rate": 4.231534813846896e-06, "loss": 3.1545, "step": 20029 }, { "epoch": 0.7057973695569827, "grad_norm": 24.337373733520508, "learning_rate": 4.230602596546376e-06, "loss": 2.5461, "step": 20030 }, { "epoch": 0.7058326065699411, "grad_norm": 1.433332920074463, "learning_rate": 4.229670454392991e-06, "loss": 0.9145, "step": 20031 }, { "epoch": 0.7058678435828994, "grad_norm": 7.714199542999268, "learning_rate": 4.228738387398885e-06, "loss": 6.5657, "step": 20032 }, { "epoch": 0.7059030805958579, "grad_norm": 1.0265324115753174, "learning_rate": 4.227806395576196e-06, "loss": 0.8072, "step": 20033 }, { "epoch": 0.7059383176088163, "grad_norm": 4.42432165145874, "learning_rate": 4.226874478937064e-06, "loss": 4.6502, "step": 20034 }, { "epoch": 0.7059735546217747, "grad_norm": 4.083748817443848, "learning_rate": 4.225942637493627e-06, "loss": 2.9611, "step": 20035 }, { "epoch": 0.7060087916347331, "grad_norm": 3.78900408744812, "learning_rate": 4.225010871258023e-06, "loss": 0.7695, "step": 20036 }, { "epoch": 0.7060440286476916, "grad_norm": 3.15384578704834, "learning_rate": 4.224079180242387e-06, "loss": 2.5736, "step": 20037 }, { "epoch": 0.7060792656606499, "grad_norm": 2.729132890701294, "learning_rate": 4.223147564458857e-06, "loss": 2.6073, "step": 20038 }, { "epoch": 0.7061145026736083, "grad_norm": 3.247396230697632, "learning_rate": 4.2222160239195654e-06, "loss": 2.7828, "step": 20039 }, { "epoch": 0.7061497396865668, "grad_norm": 1.2488048076629639, "learning_rate": 4.221284558636649e-06, "loss": 0.778, "step": 20040 }, { "epoch": 0.7061849766995252, "grad_norm": 1.0734500885009766, "learning_rate": 4.220353168622231e-06, "loss": 0.6915, "step": 20041 }, { "epoch": 0.7062202137124836, "grad_norm": 4.115890979766846, "learning_rate": 4.219421853888452e-06, "loss": 3.2911, "step": 20042 }, { "epoch": 0.706255450725442, "grad_norm": 3.132610321044922, "learning_rate": 4.218490614447444e-06, "loss": 0.8811, "step": 20043 }, { "epoch": 0.7062906877384004, "grad_norm": 4.13413667678833, "learning_rate": 4.217559450311328e-06, "loss": 5.2918, "step": 20044 }, { "epoch": 0.7063259247513588, "grad_norm": 0.966086745262146, "learning_rate": 4.216628361492233e-06, "loss": 0.9704, "step": 20045 }, { "epoch": 0.7063611617643173, "grad_norm": 4.858600616455078, "learning_rate": 4.215697348002297e-06, "loss": 3.152, "step": 20046 }, { "epoch": 0.7063963987772757, "grad_norm": 5.135886192321777, "learning_rate": 4.2147664098536376e-06, "loss": 5.389, "step": 20047 }, { "epoch": 0.706431635790234, "grad_norm": 1.6030930280685425, "learning_rate": 4.213835547058382e-06, "loss": 0.9435, "step": 20048 }, { "epoch": 0.7064668728031924, "grad_norm": 4.016361236572266, "learning_rate": 4.212904759628657e-06, "loss": 2.4737, "step": 20049 }, { "epoch": 0.7065021098161509, "grad_norm": 10.437467575073242, "learning_rate": 4.211974047576582e-06, "loss": 3.2183, "step": 20050 }, { "epoch": 0.7065373468291093, "grad_norm": 2.887796640396118, "learning_rate": 4.211043410914284e-06, "loss": 2.7378, "step": 20051 }, { "epoch": 0.7065725838420677, "grad_norm": 2.201946496963501, "learning_rate": 4.210112849653882e-06, "loss": 0.8542, "step": 20052 }, { "epoch": 0.7066078208550262, "grad_norm": 1.4291462898254395, "learning_rate": 4.209182363807498e-06, "loss": 0.7747, "step": 20053 }, { "epoch": 0.7066430578679845, "grad_norm": 4.505331039428711, "learning_rate": 4.208251953387251e-06, "loss": 4.7442, "step": 20054 }, { "epoch": 0.7066782948809429, "grad_norm": 1.0584080219268799, "learning_rate": 4.207321618405261e-06, "loss": 0.9379, "step": 20055 }, { "epoch": 0.7067135318939014, "grad_norm": 4.609860897064209, "learning_rate": 4.206391358873645e-06, "loss": 3.4068, "step": 20056 }, { "epoch": 0.7067487689068598, "grad_norm": 12.610366821289062, "learning_rate": 4.2054611748045184e-06, "loss": 5.1781, "step": 20057 }, { "epoch": 0.7067840059198182, "grad_norm": 4.524061679840088, "learning_rate": 4.204531066209998e-06, "loss": 4.8862, "step": 20058 }, { "epoch": 0.7068192429327765, "grad_norm": 2.6429660320281982, "learning_rate": 4.2036010331022035e-06, "loss": 2.997, "step": 20059 }, { "epoch": 0.706854479945735, "grad_norm": 3.683748722076416, "learning_rate": 4.2026710754932356e-06, "loss": 3.0003, "step": 20060 }, { "epoch": 0.7068897169586934, "grad_norm": 3.6959121227264404, "learning_rate": 4.201741193395219e-06, "loss": 3.3372, "step": 20061 }, { "epoch": 0.7069249539716518, "grad_norm": 1.0995814800262451, "learning_rate": 4.200811386820266e-06, "loss": 0.8341, "step": 20062 }, { "epoch": 0.7069601909846103, "grad_norm": 1.5181864500045776, "learning_rate": 4.19988165578048e-06, "loss": 0.8175, "step": 20063 }, { "epoch": 0.7069954279975686, "grad_norm": 4.5941596031188965, "learning_rate": 4.19895200028797e-06, "loss": 3.048, "step": 20064 }, { "epoch": 0.707030665010527, "grad_norm": 3.402665615081787, "learning_rate": 4.198022420354857e-06, "loss": 2.8321, "step": 20065 }, { "epoch": 0.7070659020234855, "grad_norm": 1.3117656707763672, "learning_rate": 4.1970929159932374e-06, "loss": 0.9708, "step": 20066 }, { "epoch": 0.7071011390364439, "grad_norm": 1.2860361337661743, "learning_rate": 4.196163487215221e-06, "loss": 0.7889, "step": 20067 }, { "epoch": 0.7071363760494023, "grad_norm": 1.1459451913833618, "learning_rate": 4.195234134032915e-06, "loss": 0.8211, "step": 20068 }, { "epoch": 0.7071716130623606, "grad_norm": 5.3999409675598145, "learning_rate": 4.194304856458424e-06, "loss": 3.0957, "step": 20069 }, { "epoch": 0.7072068500753191, "grad_norm": 1.3569544553756714, "learning_rate": 4.193375654503852e-06, "loss": 0.8451, "step": 20070 }, { "epoch": 0.7072420870882775, "grad_norm": 3.1697652339935303, "learning_rate": 4.192446528181301e-06, "loss": 2.818, "step": 20071 }, { "epoch": 0.7072773241012359, "grad_norm": 3.074366807937622, "learning_rate": 4.191517477502874e-06, "loss": 2.9219, "step": 20072 }, { "epoch": 0.7073125611141944, "grad_norm": 3.3122496604919434, "learning_rate": 4.190588502480671e-06, "loss": 3.064, "step": 20073 }, { "epoch": 0.7073477981271528, "grad_norm": 3.4168362617492676, "learning_rate": 4.189659603126794e-06, "loss": 2.3814, "step": 20074 }, { "epoch": 0.7073830351401111, "grad_norm": 5.92197847366333, "learning_rate": 4.188730779453341e-06, "loss": 3.1506, "step": 20075 }, { "epoch": 0.7074182721530696, "grad_norm": 5.257883071899414, "learning_rate": 4.187802031472409e-06, "loss": 2.7198, "step": 20076 }, { "epoch": 0.707453509166028, "grad_norm": 5.314043045043945, "learning_rate": 4.186873359196095e-06, "loss": 3.0944, "step": 20077 }, { "epoch": 0.7074887461789864, "grad_norm": 4.063651084899902, "learning_rate": 4.185944762636502e-06, "loss": 3.1899, "step": 20078 }, { "epoch": 0.7075239831919449, "grad_norm": 4.8685503005981445, "learning_rate": 4.185016241805709e-06, "loss": 2.7618, "step": 20079 }, { "epoch": 0.7075592202049032, "grad_norm": 2.861393690109253, "learning_rate": 4.184087796715826e-06, "loss": 2.8987, "step": 20080 }, { "epoch": 0.7075944572178616, "grad_norm": 4.925932884216309, "learning_rate": 4.183159427378943e-06, "loss": 2.7312, "step": 20081 }, { "epoch": 0.70762969423082, "grad_norm": 3.4695706367492676, "learning_rate": 4.182231133807147e-06, "loss": 3.0186, "step": 20082 }, { "epoch": 0.7076649312437785, "grad_norm": 0.8195905089378357, "learning_rate": 4.181302916012526e-06, "loss": 0.7235, "step": 20083 }, { "epoch": 0.7077001682567369, "grad_norm": 5.248266220092773, "learning_rate": 4.1803747740071845e-06, "loss": 5.1264, "step": 20084 }, { "epoch": 0.7077354052696953, "grad_norm": 2.5404860973358154, "learning_rate": 4.179446707803198e-06, "loss": 2.1183, "step": 20085 }, { "epoch": 0.7077706422826537, "grad_norm": 3.4084572792053223, "learning_rate": 4.17851871741266e-06, "loss": 0.6579, "step": 20086 }, { "epoch": 0.7078058792956121, "grad_norm": 4.611446857452393, "learning_rate": 4.177590802847657e-06, "loss": 2.9045, "step": 20087 }, { "epoch": 0.7078411163085705, "grad_norm": 6.660131454467773, "learning_rate": 4.176662964120276e-06, "loss": 3.2201, "step": 20088 }, { "epoch": 0.707876353321529, "grad_norm": 4.1321539878845215, "learning_rate": 4.175735201242601e-06, "loss": 3.1002, "step": 20089 }, { "epoch": 0.7079115903344874, "grad_norm": 4.569525718688965, "learning_rate": 4.174807514226718e-06, "loss": 5.4102, "step": 20090 }, { "epoch": 0.7079468273474457, "grad_norm": 3.190070629119873, "learning_rate": 4.173879903084708e-06, "loss": 3.4936, "step": 20091 }, { "epoch": 0.7079820643604041, "grad_norm": 1.1236006021499634, "learning_rate": 4.172952367828655e-06, "loss": 1.3424, "step": 20092 }, { "epoch": 0.7080173013733626, "grad_norm": 4.2585062980651855, "learning_rate": 4.17202490847064e-06, "loss": 2.7703, "step": 20093 }, { "epoch": 0.708052538386321, "grad_norm": 6.553628444671631, "learning_rate": 4.171097525022743e-06, "loss": 3.1165, "step": 20094 }, { "epoch": 0.7080877753992794, "grad_norm": 2.835625648498535, "learning_rate": 4.170170217497042e-06, "loss": 2.359, "step": 20095 }, { "epoch": 0.7081230124122379, "grad_norm": 1.361255168914795, "learning_rate": 4.169242985905617e-06, "loss": 0.8397, "step": 20096 }, { "epoch": 0.7081582494251962, "grad_norm": 3.8014848232269287, "learning_rate": 4.168315830260549e-06, "loss": 2.9645, "step": 20097 }, { "epoch": 0.7081934864381546, "grad_norm": 5.301586627960205, "learning_rate": 4.167388750573902e-06, "loss": 5.9471, "step": 20098 }, { "epoch": 0.7082287234511131, "grad_norm": 1.2778172492980957, "learning_rate": 4.166461746857764e-06, "loss": 0.8608, "step": 20099 }, { "epoch": 0.7082639604640715, "grad_norm": 4.935421466827393, "learning_rate": 4.1655348191242085e-06, "loss": 2.8863, "step": 20100 }, { "epoch": 0.7082991974770299, "grad_norm": 7.577879428863525, "learning_rate": 4.1646079673853006e-06, "loss": 4.8506, "step": 20101 }, { "epoch": 0.7083344344899882, "grad_norm": 0.9743044376373291, "learning_rate": 4.163681191653114e-06, "loss": 1.0524, "step": 20102 }, { "epoch": 0.7083696715029467, "grad_norm": 3.911092758178711, "learning_rate": 4.162754491939731e-06, "loss": 2.794, "step": 20103 }, { "epoch": 0.7084049085159051, "grad_norm": 1.0853617191314697, "learning_rate": 4.16182786825721e-06, "loss": 0.9392, "step": 20104 }, { "epoch": 0.7084401455288635, "grad_norm": 1.3822691440582275, "learning_rate": 4.160901320617625e-06, "loss": 1.0485, "step": 20105 }, { "epoch": 0.708475382541822, "grad_norm": 1.112653374671936, "learning_rate": 4.159974849033045e-06, "loss": 1.0083, "step": 20106 }, { "epoch": 0.7085106195547803, "grad_norm": 2.596296548843384, "learning_rate": 4.159048453515535e-06, "loss": 3.0572, "step": 20107 }, { "epoch": 0.7085458565677387, "grad_norm": 4.178374290466309, "learning_rate": 4.158122134077164e-06, "loss": 2.9294, "step": 20108 }, { "epoch": 0.7085810935806972, "grad_norm": 1.366331934928894, "learning_rate": 4.157195890729996e-06, "loss": 0.9929, "step": 20109 }, { "epoch": 0.7086163305936556, "grad_norm": 3.825733184814453, "learning_rate": 4.156269723486095e-06, "loss": 2.9842, "step": 20110 }, { "epoch": 0.708651567606614, "grad_norm": 3.8795833587646484, "learning_rate": 4.1553436323575256e-06, "loss": 3.1493, "step": 20111 }, { "epoch": 0.7086868046195725, "grad_norm": 1.3352447748184204, "learning_rate": 4.154417617356351e-06, "loss": 0.9033, "step": 20112 }, { "epoch": 0.7087220416325308, "grad_norm": 3.7433784008026123, "learning_rate": 4.15349167849463e-06, "loss": 2.5105, "step": 20113 }, { "epoch": 0.7087572786454892, "grad_norm": 1.0237109661102295, "learning_rate": 4.152565815784426e-06, "loss": 0.8383, "step": 20114 }, { "epoch": 0.7087925156584476, "grad_norm": 3.0336620807647705, "learning_rate": 4.151640029237796e-06, "loss": 2.3868, "step": 20115 }, { "epoch": 0.7088277526714061, "grad_norm": 1.2367507219314575, "learning_rate": 4.150714318866804e-06, "loss": 0.7352, "step": 20116 }, { "epoch": 0.7088629896843645, "grad_norm": 1.2649109363555908, "learning_rate": 4.149788684683496e-06, "loss": 0.8383, "step": 20117 }, { "epoch": 0.7088982266973228, "grad_norm": 0.9919140338897705, "learning_rate": 4.148863126699938e-06, "loss": 0.6827, "step": 20118 }, { "epoch": 0.7089334637102813, "grad_norm": 1.0365002155303955, "learning_rate": 4.1479376449281875e-06, "loss": 0.7026, "step": 20119 }, { "epoch": 0.7089687007232397, "grad_norm": 2.363819122314453, "learning_rate": 4.14701223938029e-06, "loss": 1.7645, "step": 20120 }, { "epoch": 0.7090039377361981, "grad_norm": 4.320711135864258, "learning_rate": 4.146086910068301e-06, "loss": 3.5891, "step": 20121 }, { "epoch": 0.7090391747491566, "grad_norm": 4.309304237365723, "learning_rate": 4.145161657004283e-06, "loss": 3.1995, "step": 20122 }, { "epoch": 0.7090744117621149, "grad_norm": 4.50403356552124, "learning_rate": 4.144236480200277e-06, "loss": 2.6425, "step": 20123 }, { "epoch": 0.7091096487750733, "grad_norm": 1.7398279905319214, "learning_rate": 4.143311379668336e-06, "loss": 0.826, "step": 20124 }, { "epoch": 0.7091448857880317, "grad_norm": 6.695784568786621, "learning_rate": 4.1423863554205104e-06, "loss": 2.7584, "step": 20125 }, { "epoch": 0.7091801228009902, "grad_norm": 6.243353366851807, "learning_rate": 4.141461407468848e-06, "loss": 3.7814, "step": 20126 }, { "epoch": 0.7092153598139486, "grad_norm": 2.890169382095337, "learning_rate": 4.140536535825398e-06, "loss": 3.3987, "step": 20127 }, { "epoch": 0.7092505968269069, "grad_norm": 0.9549965262413025, "learning_rate": 4.139611740502207e-06, "loss": 0.7662, "step": 20128 }, { "epoch": 0.7092858338398654, "grad_norm": 3.175183057785034, "learning_rate": 4.1386870215113175e-06, "loss": 2.7905, "step": 20129 }, { "epoch": 0.7093210708528238, "grad_norm": 1.0737528800964355, "learning_rate": 4.137762378864777e-06, "loss": 0.9583, "step": 20130 }, { "epoch": 0.7093563078657822, "grad_norm": 4.253481388092041, "learning_rate": 4.1368378125746275e-06, "loss": 4.8277, "step": 20131 }, { "epoch": 0.7093915448787407, "grad_norm": 2.3113646507263184, "learning_rate": 4.135913322652914e-06, "loss": 1.0037, "step": 20132 }, { "epoch": 0.709426781891699, "grad_norm": 2.0786333084106445, "learning_rate": 4.134988909111674e-06, "loss": 1.9144, "step": 20133 }, { "epoch": 0.7094620189046574, "grad_norm": 5.947328567504883, "learning_rate": 4.134064571962952e-06, "loss": 3.0821, "step": 20134 }, { "epoch": 0.7094972559176158, "grad_norm": 1.244933843612671, "learning_rate": 4.133140311218791e-06, "loss": 0.9098, "step": 20135 }, { "epoch": 0.7095324929305743, "grad_norm": 2.245016098022461, "learning_rate": 4.132216126891215e-06, "loss": 2.7678, "step": 20136 }, { "epoch": 0.7095677299435327, "grad_norm": 5.546409606933594, "learning_rate": 4.1312920189922764e-06, "loss": 2.6779, "step": 20137 }, { "epoch": 0.709602966956491, "grad_norm": 1.1868937015533447, "learning_rate": 4.1303679875340105e-06, "loss": 0.6789, "step": 20138 }, { "epoch": 0.7096382039694495, "grad_norm": 2.3992068767547607, "learning_rate": 4.1294440325284455e-06, "loss": 2.7065, "step": 20139 }, { "epoch": 0.7096734409824079, "grad_norm": 1.5334858894348145, "learning_rate": 4.128520153987616e-06, "loss": 1.0532, "step": 20140 }, { "epoch": 0.7097086779953663, "grad_norm": 11.549304008483887, "learning_rate": 4.1275963519235675e-06, "loss": 3.086, "step": 20141 }, { "epoch": 0.7097439150083248, "grad_norm": 1.7673760652542114, "learning_rate": 4.1266726263483205e-06, "loss": 0.6901, "step": 20142 }, { "epoch": 0.7097791520212832, "grad_norm": 1.065631628036499, "learning_rate": 4.12574897727391e-06, "loss": 0.7723, "step": 20143 }, { "epoch": 0.7098143890342415, "grad_norm": 1.3889776468276978, "learning_rate": 4.124825404712363e-06, "loss": 0.9369, "step": 20144 }, { "epoch": 0.7098496260472, "grad_norm": 3.8630073070526123, "learning_rate": 4.123901908675722e-06, "loss": 3.574, "step": 20145 }, { "epoch": 0.7098848630601584, "grad_norm": 3.5303149223327637, "learning_rate": 4.122978489176004e-06, "loss": 2.887, "step": 20146 }, { "epoch": 0.7099201000731168, "grad_norm": 5.334456443786621, "learning_rate": 4.122055146225239e-06, "loss": 2.5894, "step": 20147 }, { "epoch": 0.7099553370860752, "grad_norm": 3.099829912185669, "learning_rate": 4.121131879835455e-06, "loss": 2.2424, "step": 20148 }, { "epoch": 0.7099905740990337, "grad_norm": 6.41057825088501, "learning_rate": 4.120208690018677e-06, "loss": 5.4962, "step": 20149 }, { "epoch": 0.710025811111992, "grad_norm": 4.8758087158203125, "learning_rate": 4.119285576786929e-06, "loss": 2.9317, "step": 20150 }, { "epoch": 0.7100610481249504, "grad_norm": 8.241413116455078, "learning_rate": 4.118362540152235e-06, "loss": 5.3166, "step": 20151 }, { "epoch": 0.7100962851379089, "grad_norm": 4.694029808044434, "learning_rate": 4.11743958012662e-06, "loss": 2.4048, "step": 20152 }, { "epoch": 0.7101315221508673, "grad_norm": 4.914709091186523, "learning_rate": 4.116516696722102e-06, "loss": 4.7207, "step": 20153 }, { "epoch": 0.7101667591638257, "grad_norm": 4.651692867279053, "learning_rate": 4.1155938899507074e-06, "loss": 2.8007, "step": 20154 }, { "epoch": 0.7102019961767841, "grad_norm": 1.5694338083267212, "learning_rate": 4.114671159824446e-06, "loss": 0.8959, "step": 20155 }, { "epoch": 0.7102372331897425, "grad_norm": 0.8984068632125854, "learning_rate": 4.113748506355345e-06, "loss": 0.8282, "step": 20156 }, { "epoch": 0.7102724702027009, "grad_norm": 1.8913137912750244, "learning_rate": 4.1128259295554226e-06, "loss": 0.8075, "step": 20157 }, { "epoch": 0.7103077072156593, "grad_norm": 1.2241061925888062, "learning_rate": 4.1119034294366895e-06, "loss": 0.8019, "step": 20158 }, { "epoch": 0.7103429442286178, "grad_norm": 1.2362083196640015, "learning_rate": 4.110981006011161e-06, "loss": 0.6621, "step": 20159 }, { "epoch": 0.7103781812415761, "grad_norm": 2.090888500213623, "learning_rate": 4.110058659290862e-06, "loss": 0.7809, "step": 20160 }, { "epoch": 0.7104134182545345, "grad_norm": 2.1761739253997803, "learning_rate": 4.1091363892877955e-06, "loss": 1.0964, "step": 20161 }, { "epoch": 0.710448655267493, "grad_norm": 1.289913296699524, "learning_rate": 4.108214196013978e-06, "loss": 1.1256, "step": 20162 }, { "epoch": 0.7104838922804514, "grad_norm": 1.060837745666504, "learning_rate": 4.107292079481417e-06, "loss": 0.6015, "step": 20163 }, { "epoch": 0.7105191292934098, "grad_norm": 1.2982378005981445, "learning_rate": 4.106370039702136e-06, "loss": 0.7406, "step": 20164 }, { "epoch": 0.7105543663063683, "grad_norm": 4.962023735046387, "learning_rate": 4.10544807668813e-06, "loss": 3.1877, "step": 20165 }, { "epoch": 0.7105896033193266, "grad_norm": 1.8720816373825073, "learning_rate": 4.104526190451416e-06, "loss": 0.7718, "step": 20166 }, { "epoch": 0.710624840332285, "grad_norm": 1.7744044065475464, "learning_rate": 4.103604381003997e-06, "loss": 0.7939, "step": 20167 }, { "epoch": 0.7106600773452434, "grad_norm": 4.090576648712158, "learning_rate": 4.102682648357883e-06, "loss": 5.2827, "step": 20168 }, { "epoch": 0.7106953143582019, "grad_norm": 5.1602373123168945, "learning_rate": 4.101760992525079e-06, "loss": 2.9791, "step": 20169 }, { "epoch": 0.7107305513711603, "grad_norm": 4.250265598297119, "learning_rate": 4.10083941351759e-06, "loss": 4.9076, "step": 20170 }, { "epoch": 0.7107657883841186, "grad_norm": 3.084155559539795, "learning_rate": 4.099917911347419e-06, "loss": 3.5947, "step": 20171 }, { "epoch": 0.7108010253970771, "grad_norm": 3.437042474746704, "learning_rate": 4.098996486026567e-06, "loss": 2.734, "step": 20172 }, { "epoch": 0.7108362624100355, "grad_norm": 3.550999164581299, "learning_rate": 4.0980751375670425e-06, "loss": 5.1576, "step": 20173 }, { "epoch": 0.7108714994229939, "grad_norm": 3.986877679824829, "learning_rate": 4.097153865980833e-06, "loss": 3.1232, "step": 20174 }, { "epoch": 0.7109067364359524, "grad_norm": 1.7502338886260986, "learning_rate": 4.096232671279951e-06, "loss": 1.0942, "step": 20175 }, { "epoch": 0.7109419734489107, "grad_norm": 5.380919456481934, "learning_rate": 4.095311553476393e-06, "loss": 4.387, "step": 20176 }, { "epoch": 0.7109772104618691, "grad_norm": 1.1941441297531128, "learning_rate": 4.094390512582152e-06, "loss": 0.7682, "step": 20177 }, { "epoch": 0.7110124474748276, "grad_norm": 0.9672217965126038, "learning_rate": 4.093469548609222e-06, "loss": 0.9804, "step": 20178 }, { "epoch": 0.711047684487786, "grad_norm": 3.097689151763916, "learning_rate": 4.092548661569611e-06, "loss": 3.1928, "step": 20179 }, { "epoch": 0.7110829215007444, "grad_norm": 3.8696274757385254, "learning_rate": 4.091627851475303e-06, "loss": 2.8652, "step": 20180 }, { "epoch": 0.7111181585137027, "grad_norm": 1.3381023406982422, "learning_rate": 4.090707118338294e-06, "loss": 1.006, "step": 20181 }, { "epoch": 0.7111533955266612, "grad_norm": 1.4666744470596313, "learning_rate": 4.089786462170574e-06, "loss": 1.0308, "step": 20182 }, { "epoch": 0.7111886325396196, "grad_norm": 9.123337745666504, "learning_rate": 4.088865882984147e-06, "loss": 5.2007, "step": 20183 }, { "epoch": 0.711223869552578, "grad_norm": 6.621835708618164, "learning_rate": 4.087945380790988e-06, "loss": 4.9116, "step": 20184 }, { "epoch": 0.7112591065655365, "grad_norm": 1.2002841234207153, "learning_rate": 4.087024955603096e-06, "loss": 0.8025, "step": 20185 }, { "epoch": 0.7112943435784949, "grad_norm": 1.2082968950271606, "learning_rate": 4.086104607432457e-06, "loss": 0.7582, "step": 20186 }, { "epoch": 0.7113295805914532, "grad_norm": 4.797231674194336, "learning_rate": 4.0851843362910585e-06, "loss": 4.8092, "step": 20187 }, { "epoch": 0.7113648176044117, "grad_norm": 5.117405891418457, "learning_rate": 4.084264142190887e-06, "loss": 2.9872, "step": 20188 }, { "epoch": 0.7114000546173701, "grad_norm": 1.1082360744476318, "learning_rate": 4.083344025143929e-06, "loss": 1.0153, "step": 20189 }, { "epoch": 0.7114352916303285, "grad_norm": 3.446051836013794, "learning_rate": 4.082423985162168e-06, "loss": 2.8911, "step": 20190 }, { "epoch": 0.7114705286432869, "grad_norm": 1.2107608318328857, "learning_rate": 4.08150402225759e-06, "loss": 0.8661, "step": 20191 }, { "epoch": 0.7115057656562453, "grad_norm": 3.6698014736175537, "learning_rate": 4.080584136442179e-06, "loss": 2.4949, "step": 20192 }, { "epoch": 0.7115410026692037, "grad_norm": 3.7724361419677734, "learning_rate": 4.079664327727905e-06, "loss": 3.226, "step": 20193 }, { "epoch": 0.7115762396821621, "grad_norm": 5.045708179473877, "learning_rate": 4.078744596126763e-06, "loss": 5.1794, "step": 20194 }, { "epoch": 0.7116114766951206, "grad_norm": 3.472475051879883, "learning_rate": 4.077824941650729e-06, "loss": 2.7896, "step": 20195 }, { "epoch": 0.711646713708079, "grad_norm": 4.755342483520508, "learning_rate": 4.076905364311776e-06, "loss": 2.6938, "step": 20196 }, { "epoch": 0.7116819507210373, "grad_norm": 0.9410973191261292, "learning_rate": 4.075985864121882e-06, "loss": 0.987, "step": 20197 }, { "epoch": 0.7117171877339958, "grad_norm": 5.618770122528076, "learning_rate": 4.07506644109303e-06, "loss": 5.0633, "step": 20198 }, { "epoch": 0.7117524247469542, "grad_norm": 3.1329894065856934, "learning_rate": 4.074147095237196e-06, "loss": 0.7507, "step": 20199 }, { "epoch": 0.7117876617599126, "grad_norm": 6.968745231628418, "learning_rate": 4.073227826566348e-06, "loss": 3.27, "step": 20200 }, { "epoch": 0.711822898772871, "grad_norm": 1.2040811777114868, "learning_rate": 4.072308635092461e-06, "loss": 1.0242, "step": 20201 }, { "epoch": 0.7118581357858295, "grad_norm": 6.753559589385986, "learning_rate": 4.071389520827511e-06, "loss": 3.3037, "step": 20202 }, { "epoch": 0.7118933727987878, "grad_norm": 4.348690032958984, "learning_rate": 4.0704704837834665e-06, "loss": 4.801, "step": 20203 }, { "epoch": 0.7119286098117462, "grad_norm": 3.510124444961548, "learning_rate": 4.069551523972299e-06, "loss": 2.8165, "step": 20204 }, { "epoch": 0.7119638468247047, "grad_norm": 0.9354941248893738, "learning_rate": 4.06863264140598e-06, "loss": 0.9688, "step": 20205 }, { "epoch": 0.7119990838376631, "grad_norm": 5.181520938873291, "learning_rate": 4.067713836096474e-06, "loss": 2.9065, "step": 20206 }, { "epoch": 0.7120343208506215, "grad_norm": 1.3106192350387573, "learning_rate": 4.066795108055752e-06, "loss": 0.9349, "step": 20207 }, { "epoch": 0.71206955786358, "grad_norm": 7.434866428375244, "learning_rate": 4.06587645729578e-06, "loss": 5.5917, "step": 20208 }, { "epoch": 0.7121047948765383, "grad_norm": 0.9239770174026489, "learning_rate": 4.064957883828522e-06, "loss": 0.7778, "step": 20209 }, { "epoch": 0.7121400318894967, "grad_norm": 14.93123722076416, "learning_rate": 4.064039387665944e-06, "loss": 1.1244, "step": 20210 }, { "epoch": 0.7121752689024552, "grad_norm": 1.1351231336593628, "learning_rate": 4.063120968820013e-06, "loss": 0.8911, "step": 20211 }, { "epoch": 0.7122105059154136, "grad_norm": 2.089492082595825, "learning_rate": 4.062202627302684e-06, "loss": 0.7711, "step": 20212 }, { "epoch": 0.712245742928372, "grad_norm": 1.154729962348938, "learning_rate": 4.061284363125919e-06, "loss": 0.8746, "step": 20213 }, { "epoch": 0.7122809799413303, "grad_norm": 1.1025303602218628, "learning_rate": 4.060366176301688e-06, "loss": 0.8378, "step": 20214 }, { "epoch": 0.7123162169542888, "grad_norm": 0.9849153757095337, "learning_rate": 4.0594480668419416e-06, "loss": 0.6751, "step": 20215 }, { "epoch": 0.7123514539672472, "grad_norm": 1.4581317901611328, "learning_rate": 4.058530034758639e-06, "loss": 0.8917, "step": 20216 }, { "epoch": 0.7123866909802056, "grad_norm": 1.3048368692398071, "learning_rate": 4.057612080063737e-06, "loss": 0.926, "step": 20217 }, { "epoch": 0.7124219279931641, "grad_norm": 4.228998184204102, "learning_rate": 4.056694202769201e-06, "loss": 2.6394, "step": 20218 }, { "epoch": 0.7124571650061224, "grad_norm": 3.860044002532959, "learning_rate": 4.055776402886978e-06, "loss": 3.1357, "step": 20219 }, { "epoch": 0.7124924020190808, "grad_norm": 2.6784682273864746, "learning_rate": 4.054858680429024e-06, "loss": 2.1411, "step": 20220 }, { "epoch": 0.7125276390320393, "grad_norm": 1.8044666051864624, "learning_rate": 4.0539410354072926e-06, "loss": 0.7298, "step": 20221 }, { "epoch": 0.7125628760449977, "grad_norm": 3.212975263595581, "learning_rate": 4.053023467833737e-06, "loss": 2.8354, "step": 20222 }, { "epoch": 0.7125981130579561, "grad_norm": 3.1425766944885254, "learning_rate": 4.052105977720307e-06, "loss": 3.6697, "step": 20223 }, { "epoch": 0.7126333500709144, "grad_norm": 2.48413348197937, "learning_rate": 4.051188565078956e-06, "loss": 0.6828, "step": 20224 }, { "epoch": 0.7126685870838729, "grad_norm": 2.116971731185913, "learning_rate": 4.050271229921631e-06, "loss": 0.8281, "step": 20225 }, { "epoch": 0.7127038240968313, "grad_norm": 2.5765607357025146, "learning_rate": 4.0493539722602806e-06, "loss": 3.0937, "step": 20226 }, { "epoch": 0.7127390611097897, "grad_norm": 2.8211488723754883, "learning_rate": 4.048436792106857e-06, "loss": 1.0688, "step": 20227 }, { "epoch": 0.7127742981227482, "grad_norm": 0.8511940836906433, "learning_rate": 4.047519689473295e-06, "loss": 0.7447, "step": 20228 }, { "epoch": 0.7128095351357066, "grad_norm": 1.0684514045715332, "learning_rate": 4.046602664371551e-06, "loss": 0.8016, "step": 20229 }, { "epoch": 0.7128447721486649, "grad_norm": 3.452739953994751, "learning_rate": 4.045685716813569e-06, "loss": 2.9728, "step": 20230 }, { "epoch": 0.7128800091616234, "grad_norm": 1.6190286874771118, "learning_rate": 4.044768846811284e-06, "loss": 0.7481, "step": 20231 }, { "epoch": 0.7129152461745818, "grad_norm": 3.8947322368621826, "learning_rate": 4.043852054376643e-06, "loss": 4.7046, "step": 20232 }, { "epoch": 0.7129504831875402, "grad_norm": 2.7627267837524414, "learning_rate": 4.042935339521592e-06, "loss": 3.1217, "step": 20233 }, { "epoch": 0.7129857202004986, "grad_norm": 5.001183986663818, "learning_rate": 4.0420187022580646e-06, "loss": 2.3177, "step": 20234 }, { "epoch": 0.713020957213457, "grad_norm": 2.2015459537506104, "learning_rate": 4.041102142598002e-06, "loss": 0.9517, "step": 20235 }, { "epoch": 0.7130561942264154, "grad_norm": 3.7875428199768066, "learning_rate": 4.0401856605533395e-06, "loss": 2.8192, "step": 20236 }, { "epoch": 0.7130914312393738, "grad_norm": 2.7595016956329346, "learning_rate": 4.039269256136026e-06, "loss": 2.5772, "step": 20237 }, { "epoch": 0.7131266682523323, "grad_norm": 4.575747489929199, "learning_rate": 4.038352929357986e-06, "loss": 4.9661, "step": 20238 }, { "epoch": 0.7131619052652907, "grad_norm": 4.013069152832031, "learning_rate": 4.037436680231157e-06, "loss": 2.7309, "step": 20239 }, { "epoch": 0.713197142278249, "grad_norm": 3.775665283203125, "learning_rate": 4.036520508767477e-06, "loss": 3.0143, "step": 20240 }, { "epoch": 0.7132323792912075, "grad_norm": 0.9807435870170593, "learning_rate": 4.035604414978875e-06, "loss": 0.943, "step": 20241 }, { "epoch": 0.7132676163041659, "grad_norm": 1.2005473375320435, "learning_rate": 4.034688398877286e-06, "loss": 0.6456, "step": 20242 }, { "epoch": 0.7133028533171243, "grad_norm": 1.1373809576034546, "learning_rate": 4.033772460474641e-06, "loss": 0.9627, "step": 20243 }, { "epoch": 0.7133380903300827, "grad_norm": 1.6586636304855347, "learning_rate": 4.032856599782868e-06, "loss": 0.5906, "step": 20244 }, { "epoch": 0.7133733273430412, "grad_norm": 3.2843053340911865, "learning_rate": 4.0319408168139e-06, "loss": 2.5252, "step": 20245 }, { "epoch": 0.7134085643559995, "grad_norm": 1.1855882406234741, "learning_rate": 4.031025111579665e-06, "loss": 0.8867, "step": 20246 }, { "epoch": 0.7134438013689579, "grad_norm": 3.7370522022247314, "learning_rate": 4.030109484092082e-06, "loss": 3.0524, "step": 20247 }, { "epoch": 0.7134790383819164, "grad_norm": 1.8975754976272583, "learning_rate": 4.029193934363087e-06, "loss": 0.9454, "step": 20248 }, { "epoch": 0.7135142753948748, "grad_norm": 3.088815450668335, "learning_rate": 4.028278462404605e-06, "loss": 2.3156, "step": 20249 }, { "epoch": 0.7135495124078332, "grad_norm": 6.997194766998291, "learning_rate": 4.027363068228552e-06, "loss": 5.2457, "step": 20250 }, { "epoch": 0.7135847494207916, "grad_norm": 4.324065685272217, "learning_rate": 4.026447751846854e-06, "loss": 2.8078, "step": 20251 }, { "epoch": 0.71361998643375, "grad_norm": 3.0348854064941406, "learning_rate": 4.025532513271441e-06, "loss": 2.5602, "step": 20252 }, { "epoch": 0.7136552234467084, "grad_norm": 2.725080966949463, "learning_rate": 4.024617352514225e-06, "loss": 3.0658, "step": 20253 }, { "epoch": 0.7136904604596669, "grad_norm": 8.895256042480469, "learning_rate": 4.023702269587128e-06, "loss": 3.1152, "step": 20254 }, { "epoch": 0.7137256974726253, "grad_norm": 1.5434679985046387, "learning_rate": 4.022787264502068e-06, "loss": 0.9369, "step": 20255 }, { "epoch": 0.7137609344855836, "grad_norm": 4.396332263946533, "learning_rate": 4.021872337270972e-06, "loss": 2.5428, "step": 20256 }, { "epoch": 0.713796171498542, "grad_norm": 1.5247180461883545, "learning_rate": 4.020957487905748e-06, "loss": 0.9726, "step": 20257 }, { "epoch": 0.7138314085115005, "grad_norm": 2.7151663303375244, "learning_rate": 4.020042716418312e-06, "loss": 2.6514, "step": 20258 }, { "epoch": 0.7138666455244589, "grad_norm": 3.6632766723632812, "learning_rate": 4.019128022820583e-06, "loss": 0.8908, "step": 20259 }, { "epoch": 0.7139018825374173, "grad_norm": 2.6141576766967773, "learning_rate": 4.018213407124473e-06, "loss": 2.7661, "step": 20260 }, { "epoch": 0.7139371195503758, "grad_norm": 2.7313461303710938, "learning_rate": 4.017298869341894e-06, "loss": 2.7445, "step": 20261 }, { "epoch": 0.7139723565633341, "grad_norm": 1.5002025365829468, "learning_rate": 4.016384409484762e-06, "loss": 0.6856, "step": 20262 }, { "epoch": 0.7140075935762925, "grad_norm": 1.0824733972549438, "learning_rate": 4.0154700275649826e-06, "loss": 1.0121, "step": 20263 }, { "epoch": 0.714042830589251, "grad_norm": 1.4884710311889648, "learning_rate": 4.01455572359447e-06, "loss": 0.9899, "step": 20264 }, { "epoch": 0.7140780676022094, "grad_norm": 1.2020442485809326, "learning_rate": 4.013641497585134e-06, "loss": 0.7769, "step": 20265 }, { "epoch": 0.7141133046151678, "grad_norm": 1.4393370151519775, "learning_rate": 4.012727349548873e-06, "loss": 0.8882, "step": 20266 }, { "epoch": 0.7141485416281261, "grad_norm": 2.0428199768066406, "learning_rate": 4.011813279497605e-06, "loss": 0.9751, "step": 20267 }, { "epoch": 0.7141837786410846, "grad_norm": 1.223457932472229, "learning_rate": 4.010899287443235e-06, "loss": 0.8118, "step": 20268 }, { "epoch": 0.714219015654043, "grad_norm": 1.0160770416259766, "learning_rate": 4.009985373397662e-06, "loss": 0.8227, "step": 20269 }, { "epoch": 0.7142542526670014, "grad_norm": 1.189238429069519, "learning_rate": 4.009071537372788e-06, "loss": 0.8314, "step": 20270 }, { "epoch": 0.7142894896799599, "grad_norm": 4.088260173797607, "learning_rate": 4.008157779380525e-06, "loss": 2.9893, "step": 20271 }, { "epoch": 0.7143247266929182, "grad_norm": 1.631237268447876, "learning_rate": 4.007244099432772e-06, "loss": 0.7536, "step": 20272 }, { "epoch": 0.7143599637058766, "grad_norm": 2.166257619857788, "learning_rate": 4.006330497541426e-06, "loss": 0.8918, "step": 20273 }, { "epoch": 0.7143952007188351, "grad_norm": 1.5449409484863281, "learning_rate": 4.005416973718384e-06, "loss": 1.2989, "step": 20274 }, { "epoch": 0.7144304377317935, "grad_norm": 4.416757106781006, "learning_rate": 4.004503527975558e-06, "loss": 3.1865, "step": 20275 }, { "epoch": 0.7144656747447519, "grad_norm": 0.9165359735488892, "learning_rate": 4.003590160324833e-06, "loss": 1.0013, "step": 20276 }, { "epoch": 0.7145009117577102, "grad_norm": 1.16740882396698, "learning_rate": 4.002676870778109e-06, "loss": 0.9276, "step": 20277 }, { "epoch": 0.7145361487706687, "grad_norm": 1.2514797449111938, "learning_rate": 4.001763659347284e-06, "loss": 0.956, "step": 20278 }, { "epoch": 0.7145713857836271, "grad_norm": 2.683260917663574, "learning_rate": 4.000850526044251e-06, "loss": 2.3924, "step": 20279 }, { "epoch": 0.7146066227965855, "grad_norm": 1.5083867311477661, "learning_rate": 3.9999374708809035e-06, "loss": 0.9824, "step": 20280 }, { "epoch": 0.714641859809544, "grad_norm": 1.2516942024230957, "learning_rate": 3.999024493869136e-06, "loss": 0.8443, "step": 20281 }, { "epoch": 0.7146770968225024, "grad_norm": 4.96334981918335, "learning_rate": 3.998111595020837e-06, "loss": 4.3208, "step": 20282 }, { "epoch": 0.7147123338354607, "grad_norm": 4.2437357902526855, "learning_rate": 3.9971987743479e-06, "loss": 2.7875, "step": 20283 }, { "epoch": 0.7147475708484192, "grad_norm": 4.5090227127075195, "learning_rate": 3.996286031862218e-06, "loss": 2.8811, "step": 20284 }, { "epoch": 0.7147828078613776, "grad_norm": 3.616910219192505, "learning_rate": 3.995373367575667e-06, "loss": 3.147, "step": 20285 }, { "epoch": 0.714818044874336, "grad_norm": 5.022408962249756, "learning_rate": 3.994460781500147e-06, "loss": 4.7495, "step": 20286 }, { "epoch": 0.7148532818872945, "grad_norm": 8.253527641296387, "learning_rate": 3.993548273647544e-06, "loss": 4.8311, "step": 20287 }, { "epoch": 0.7148885189002528, "grad_norm": 4.717407703399658, "learning_rate": 3.992635844029736e-06, "loss": 5.087, "step": 20288 }, { "epoch": 0.7149237559132112, "grad_norm": 1.206838607788086, "learning_rate": 3.991723492658607e-06, "loss": 0.9752, "step": 20289 }, { "epoch": 0.7149589929261696, "grad_norm": 6.28311014175415, "learning_rate": 3.990811219546049e-06, "loss": 2.63, "step": 20290 }, { "epoch": 0.7149942299391281, "grad_norm": 2.440011501312256, "learning_rate": 3.989899024703944e-06, "loss": 1.1072, "step": 20291 }, { "epoch": 0.7150294669520865, "grad_norm": 1.2371406555175781, "learning_rate": 3.988986908144166e-06, "loss": 1.2374, "step": 20292 }, { "epoch": 0.7150647039650448, "grad_norm": 5.390044689178467, "learning_rate": 3.988074869878595e-06, "loss": 5.5224, "step": 20293 }, { "epoch": 0.7150999409780033, "grad_norm": 3.1684117317199707, "learning_rate": 3.987162909919122e-06, "loss": 3.3464, "step": 20294 }, { "epoch": 0.7151351779909617, "grad_norm": 3.7802276611328125, "learning_rate": 3.986251028277616e-06, "loss": 3.043, "step": 20295 }, { "epoch": 0.7151704150039201, "grad_norm": 1.2429933547973633, "learning_rate": 3.985339224965955e-06, "loss": 0.8096, "step": 20296 }, { "epoch": 0.7152056520168786, "grad_norm": 2.7530617713928223, "learning_rate": 3.984427499996016e-06, "loss": 2.7921, "step": 20297 }, { "epoch": 0.715240889029837, "grad_norm": 2.0649683475494385, "learning_rate": 3.983515853379677e-06, "loss": 0.8546, "step": 20298 }, { "epoch": 0.7152761260427953, "grad_norm": 4.5509257316589355, "learning_rate": 3.982604285128809e-06, "loss": 4.7645, "step": 20299 }, { "epoch": 0.7153113630557537, "grad_norm": 0.912800133228302, "learning_rate": 3.981692795255288e-06, "loss": 1.1334, "step": 20300 }, { "epoch": 0.7153466000687122, "grad_norm": 3.4685442447662354, "learning_rate": 3.980781383770984e-06, "loss": 2.8356, "step": 20301 }, { "epoch": 0.7153818370816706, "grad_norm": 2.4614410400390625, "learning_rate": 3.979870050687769e-06, "loss": 2.8373, "step": 20302 }, { "epoch": 0.715417074094629, "grad_norm": 5.772610664367676, "learning_rate": 3.978958796017517e-06, "loss": 6.6498, "step": 20303 }, { "epoch": 0.7154523111075874, "grad_norm": 5.98753023147583, "learning_rate": 3.978047619772087e-06, "loss": 2.9973, "step": 20304 }, { "epoch": 0.7154875481205458, "grad_norm": 1.0352495908737183, "learning_rate": 3.977136521963357e-06, "loss": 0.8531, "step": 20305 }, { "epoch": 0.7155227851335042, "grad_norm": 5.187468528747559, "learning_rate": 3.976225502603195e-06, "loss": 5.12, "step": 20306 }, { "epoch": 0.7155580221464627, "grad_norm": 2.704759120941162, "learning_rate": 3.97531456170346e-06, "loss": 2.3761, "step": 20307 }, { "epoch": 0.7155932591594211, "grad_norm": 5.187048435211182, "learning_rate": 3.9744036992760146e-06, "loss": 2.1729, "step": 20308 }, { "epoch": 0.7156284961723794, "grad_norm": 1.0480629205703735, "learning_rate": 3.973492915332734e-06, "loss": 0.8766, "step": 20309 }, { "epoch": 0.7156637331853378, "grad_norm": 7.051554203033447, "learning_rate": 3.972582209885479e-06, "loss": 4.905, "step": 20310 }, { "epoch": 0.7156989701982963, "grad_norm": 5.486911296844482, "learning_rate": 3.971671582946105e-06, "loss": 5.4527, "step": 20311 }, { "epoch": 0.7157342072112547, "grad_norm": 1.3106611967086792, "learning_rate": 3.9707610345264725e-06, "loss": 0.7042, "step": 20312 }, { "epoch": 0.7157694442242131, "grad_norm": 3.3124940395355225, "learning_rate": 3.969850564638453e-06, "loss": 2.6707, "step": 20313 }, { "epoch": 0.7158046812371716, "grad_norm": 1.438957929611206, "learning_rate": 3.968940173293895e-06, "loss": 0.7094, "step": 20314 }, { "epoch": 0.7158399182501299, "grad_norm": 1.1492228507995605, "learning_rate": 3.968029860504658e-06, "loss": 1.0169, "step": 20315 }, { "epoch": 0.7158751552630883, "grad_norm": 3.2729759216308594, "learning_rate": 3.967119626282602e-06, "loss": 2.8585, "step": 20316 }, { "epoch": 0.7159103922760468, "grad_norm": 2.8732662200927734, "learning_rate": 3.96620947063958e-06, "loss": 2.9137, "step": 20317 }, { "epoch": 0.7159456292890052, "grad_norm": 1.9734903573989868, "learning_rate": 3.965299393587449e-06, "loss": 0.7403, "step": 20318 }, { "epoch": 0.7159808663019636, "grad_norm": 2.605349540710449, "learning_rate": 3.964389395138063e-06, "loss": 3.0386, "step": 20319 }, { "epoch": 0.716016103314922, "grad_norm": 3.6371219158172607, "learning_rate": 3.963479475303272e-06, "loss": 3.3412, "step": 20320 }, { "epoch": 0.7160513403278804, "grad_norm": 4.55154275894165, "learning_rate": 3.962569634094932e-06, "loss": 2.9161, "step": 20321 }, { "epoch": 0.7160865773408388, "grad_norm": 1.2211192846298218, "learning_rate": 3.961659871524893e-06, "loss": 0.9039, "step": 20322 }, { "epoch": 0.7161218143537972, "grad_norm": 3.5319201946258545, "learning_rate": 3.960750187604997e-06, "loss": 2.0523, "step": 20323 }, { "epoch": 0.7161570513667557, "grad_norm": 0.9742681980133057, "learning_rate": 3.959840582347103e-06, "loss": 0.5415, "step": 20324 }, { "epoch": 0.716192288379714, "grad_norm": 3.423725128173828, "learning_rate": 3.958931055763058e-06, "loss": 2.7626, "step": 20325 }, { "epoch": 0.7162275253926724, "grad_norm": 7.272803783416748, "learning_rate": 3.958021607864701e-06, "loss": 5.188, "step": 20326 }, { "epoch": 0.7162627624056309, "grad_norm": 3.565615653991699, "learning_rate": 3.95711223866388e-06, "loss": 2.9037, "step": 20327 }, { "epoch": 0.7162979994185893, "grad_norm": 4.062987327575684, "learning_rate": 3.956202948172445e-06, "loss": 2.7841, "step": 20328 }, { "epoch": 0.7163332364315477, "grad_norm": 1.4914301633834839, "learning_rate": 3.95529373640224e-06, "loss": 0.7943, "step": 20329 }, { "epoch": 0.7163684734445062, "grad_norm": 2.6085643768310547, "learning_rate": 3.954384603365102e-06, "loss": 0.9056, "step": 20330 }, { "epoch": 0.7164037104574645, "grad_norm": 4.827327251434326, "learning_rate": 3.953475549072868e-06, "loss": 3.1346, "step": 20331 }, { "epoch": 0.7164389474704229, "grad_norm": 2.64442777633667, "learning_rate": 3.952566573537394e-06, "loss": 0.7808, "step": 20332 }, { "epoch": 0.7164741844833813, "grad_norm": 3.9770591259002686, "learning_rate": 3.951657676770507e-06, "loss": 3.179, "step": 20333 }, { "epoch": 0.7165094214963398, "grad_norm": 1.354814052581787, "learning_rate": 3.950748858784049e-06, "loss": 0.7306, "step": 20334 }, { "epoch": 0.7165446585092982, "grad_norm": 2.5694522857666016, "learning_rate": 3.949840119589859e-06, "loss": 2.9842, "step": 20335 }, { "epoch": 0.7165798955222565, "grad_norm": 9.788375854492188, "learning_rate": 3.9489314591997705e-06, "loss": 7.774, "step": 20336 }, { "epoch": 0.716615132535215, "grad_norm": 1.2002222537994385, "learning_rate": 3.948022877625621e-06, "loss": 1.026, "step": 20337 }, { "epoch": 0.7166503695481734, "grad_norm": 1.4587682485580444, "learning_rate": 3.947114374879244e-06, "loss": 0.7604, "step": 20338 }, { "epoch": 0.7166856065611318, "grad_norm": 3.334455728530884, "learning_rate": 3.946205950972474e-06, "loss": 2.612, "step": 20339 }, { "epoch": 0.7167208435740903, "grad_norm": 1.4294776916503906, "learning_rate": 3.9452976059171425e-06, "loss": 0.7308, "step": 20340 }, { "epoch": 0.7167560805870486, "grad_norm": 4.238361358642578, "learning_rate": 3.944389339725084e-06, "loss": 3.3194, "step": 20341 }, { "epoch": 0.716791317600007, "grad_norm": 4.315934658050537, "learning_rate": 3.94348115240812e-06, "loss": 3.118, "step": 20342 }, { "epoch": 0.7168265546129654, "grad_norm": 1.6171876192092896, "learning_rate": 3.9425730439780875e-06, "loss": 0.8106, "step": 20343 }, { "epoch": 0.7168617916259239, "grad_norm": 3.470142126083374, "learning_rate": 3.941665014446813e-06, "loss": 3.1794, "step": 20344 }, { "epoch": 0.7168970286388823, "grad_norm": 6.3902058601379395, "learning_rate": 3.9407570638261276e-06, "loss": 2.7448, "step": 20345 }, { "epoch": 0.7169322656518407, "grad_norm": 5.36558723449707, "learning_rate": 3.939849192127845e-06, "loss": 5.17, "step": 20346 }, { "epoch": 0.7169675026647991, "grad_norm": 3.2442538738250732, "learning_rate": 3.938941399363803e-06, "loss": 2.6227, "step": 20347 }, { "epoch": 0.7170027396777575, "grad_norm": 6.95051908493042, "learning_rate": 3.938033685545824e-06, "loss": 4.6834, "step": 20348 }, { "epoch": 0.7170379766907159, "grad_norm": 4.23100471496582, "learning_rate": 3.937126050685726e-06, "loss": 2.7011, "step": 20349 }, { "epoch": 0.7170732137036744, "grad_norm": 3.2059779167175293, "learning_rate": 3.936218494795329e-06, "loss": 3.0506, "step": 20350 }, { "epoch": 0.7171084507166328, "grad_norm": 4.554233551025391, "learning_rate": 3.9353110178864664e-06, "loss": 4.6471, "step": 20351 }, { "epoch": 0.7171436877295911, "grad_norm": 2.883887767791748, "learning_rate": 3.934403619970947e-06, "loss": 3.2679, "step": 20352 }, { "epoch": 0.7171789247425496, "grad_norm": 4.7079949378967285, "learning_rate": 3.933496301060592e-06, "loss": 2.8361, "step": 20353 }, { "epoch": 0.717214161755508, "grad_norm": 3.055845022201538, "learning_rate": 3.9325890611672215e-06, "loss": 2.5878, "step": 20354 }, { "epoch": 0.7172493987684664, "grad_norm": 1.0612579584121704, "learning_rate": 3.931681900302651e-06, "loss": 1.0376, "step": 20355 }, { "epoch": 0.7172846357814248, "grad_norm": 8.332301139831543, "learning_rate": 3.930774818478696e-06, "loss": 4.9504, "step": 20356 }, { "epoch": 0.7173198727943833, "grad_norm": 2.220872640609741, "learning_rate": 3.929867815707173e-06, "loss": 1.1365, "step": 20357 }, { "epoch": 0.7173551098073416, "grad_norm": 1.281571865081787, "learning_rate": 3.928960891999894e-06, "loss": 0.8804, "step": 20358 }, { "epoch": 0.7173903468203, "grad_norm": 2.917933225631714, "learning_rate": 3.928054047368673e-06, "loss": 2.5182, "step": 20359 }, { "epoch": 0.7174255838332585, "grad_norm": 6.155808925628662, "learning_rate": 3.927147281825325e-06, "loss": 5.0516, "step": 20360 }, { "epoch": 0.7174608208462169, "grad_norm": 2.1616384983062744, "learning_rate": 3.926240595381651e-06, "loss": 0.8938, "step": 20361 }, { "epoch": 0.7174960578591753, "grad_norm": 4.492408275604248, "learning_rate": 3.92533398804947e-06, "loss": 3.209, "step": 20362 }, { "epoch": 0.7175312948721337, "grad_norm": 1.2514655590057373, "learning_rate": 3.924427459840588e-06, "loss": 0.7003, "step": 20363 }, { "epoch": 0.7175665318850921, "grad_norm": 1.1409064531326294, "learning_rate": 3.923521010766814e-06, "loss": 1.1248, "step": 20364 }, { "epoch": 0.7176017688980505, "grad_norm": 2.245732545852661, "learning_rate": 3.922614640839948e-06, "loss": 0.9808, "step": 20365 }, { "epoch": 0.7176370059110089, "grad_norm": 4.116143226623535, "learning_rate": 3.921708350071802e-06, "loss": 3.1105, "step": 20366 }, { "epoch": 0.7176722429239674, "grad_norm": 3.203420400619507, "learning_rate": 3.920802138474185e-06, "loss": 2.0838, "step": 20367 }, { "epoch": 0.7177074799369257, "grad_norm": 1.1823519468307495, "learning_rate": 3.919896006058889e-06, "loss": 0.9321, "step": 20368 }, { "epoch": 0.7177427169498841, "grad_norm": 2.5819523334503174, "learning_rate": 3.918989952837718e-06, "loss": 2.2709, "step": 20369 }, { "epoch": 0.7177779539628426, "grad_norm": 6.28311824798584, "learning_rate": 3.918083978822486e-06, "loss": 4.211, "step": 20370 }, { "epoch": 0.717813190975801, "grad_norm": 4.656942844390869, "learning_rate": 3.917178084024981e-06, "loss": 2.7288, "step": 20371 }, { "epoch": 0.7178484279887594, "grad_norm": 1.6073821783065796, "learning_rate": 3.916272268457006e-06, "loss": 0.9323, "step": 20372 }, { "epoch": 0.7178836650017179, "grad_norm": 3.0390334129333496, "learning_rate": 3.91536653213036e-06, "loss": 3.0253, "step": 20373 }, { "epoch": 0.7179189020146762, "grad_norm": 6.321186065673828, "learning_rate": 3.914460875056839e-06, "loss": 4.994, "step": 20374 }, { "epoch": 0.7179541390276346, "grad_norm": 3.691441774368286, "learning_rate": 3.913555297248242e-06, "loss": 3.0146, "step": 20375 }, { "epoch": 0.717989376040593, "grad_norm": 1.1376205682754517, "learning_rate": 3.912649798716361e-06, "loss": 0.802, "step": 20376 }, { "epoch": 0.7180246130535515, "grad_norm": 1.004593014717102, "learning_rate": 3.911744379472993e-06, "loss": 0.8478, "step": 20377 }, { "epoch": 0.7180598500665099, "grad_norm": 1.1531997919082642, "learning_rate": 3.910839039529929e-06, "loss": 0.947, "step": 20378 }, { "epoch": 0.7180950870794682, "grad_norm": 3.652070999145508, "learning_rate": 3.909933778898965e-06, "loss": 2.9393, "step": 20379 }, { "epoch": 0.7181303240924267, "grad_norm": 1.13809335231781, "learning_rate": 3.909028597591884e-06, "loss": 0.9215, "step": 20380 }, { "epoch": 0.7181655611053851, "grad_norm": 4.620161533355713, "learning_rate": 3.908123495620484e-06, "loss": 4.976, "step": 20381 }, { "epoch": 0.7182007981183435, "grad_norm": 1.1886987686157227, "learning_rate": 3.907218472996553e-06, "loss": 0.8551, "step": 20382 }, { "epoch": 0.718236035131302, "grad_norm": 5.426540374755859, "learning_rate": 3.9063135297318784e-06, "loss": 3.431, "step": 20383 }, { "epoch": 0.7182712721442603, "grad_norm": 1.0767827033996582, "learning_rate": 3.90540866583824e-06, "loss": 0.9014, "step": 20384 }, { "epoch": 0.7183065091572187, "grad_norm": 1.8000259399414062, "learning_rate": 3.904503881327435e-06, "loss": 0.9497, "step": 20385 }, { "epoch": 0.7183417461701772, "grad_norm": 4.548223972320557, "learning_rate": 3.9035991762112455e-06, "loss": 5.0634, "step": 20386 }, { "epoch": 0.7183769831831356, "grad_norm": 3.844224691390991, "learning_rate": 3.90269455050145e-06, "loss": 4.7129, "step": 20387 }, { "epoch": 0.718412220196094, "grad_norm": 1.0897817611694336, "learning_rate": 3.9017900042098315e-06, "loss": 0.7927, "step": 20388 }, { "epoch": 0.7184474572090523, "grad_norm": 1.1859822273254395, "learning_rate": 3.900885537348181e-06, "loss": 1.1851, "step": 20389 }, { "epoch": 0.7184826942220108, "grad_norm": 2.4300146102905273, "learning_rate": 3.899981149928271e-06, "loss": 0.7398, "step": 20390 }, { "epoch": 0.7185179312349692, "grad_norm": 2.022268772125244, "learning_rate": 3.899076841961883e-06, "loss": 0.9296, "step": 20391 }, { "epoch": 0.7185531682479276, "grad_norm": 2.037799119949341, "learning_rate": 3.898172613460797e-06, "loss": 0.6883, "step": 20392 }, { "epoch": 0.7185884052608861, "grad_norm": 5.304250717163086, "learning_rate": 3.897268464436789e-06, "loss": 3.6058, "step": 20393 }, { "epoch": 0.7186236422738445, "grad_norm": 6.195444107055664, "learning_rate": 3.896364394901637e-06, "loss": 5.8549, "step": 20394 }, { "epoch": 0.7186588792868028, "grad_norm": 5.670198917388916, "learning_rate": 3.895460404867116e-06, "loss": 2.7453, "step": 20395 }, { "epoch": 0.7186941162997613, "grad_norm": 3.38474440574646, "learning_rate": 3.894556494345002e-06, "loss": 2.6767, "step": 20396 }, { "epoch": 0.7187293533127197, "grad_norm": 2.11871075630188, "learning_rate": 3.893652663347066e-06, "loss": 0.8084, "step": 20397 }, { "epoch": 0.7187645903256781, "grad_norm": 4.092187404632568, "learning_rate": 3.892748911885086e-06, "loss": 4.8662, "step": 20398 }, { "epoch": 0.7187998273386365, "grad_norm": 1.0231529474258423, "learning_rate": 3.891845239970822e-06, "loss": 0.8645, "step": 20399 }, { "epoch": 0.7188350643515949, "grad_norm": 1.8686444759368896, "learning_rate": 3.890941647616056e-06, "loss": 0.8781, "step": 20400 }, { "epoch": 0.7188703013645533, "grad_norm": 1.463206171989441, "learning_rate": 3.8900381348325526e-06, "loss": 0.7681, "step": 20401 }, { "epoch": 0.7189055383775117, "grad_norm": 3.258005380630493, "learning_rate": 3.889134701632083e-06, "loss": 2.5968, "step": 20402 }, { "epoch": 0.7189407753904702, "grad_norm": 2.8837716579437256, "learning_rate": 3.888231348026408e-06, "loss": 2.641, "step": 20403 }, { "epoch": 0.7189760124034286, "grad_norm": 3.3960649967193604, "learning_rate": 3.8873280740273e-06, "loss": 2.9555, "step": 20404 }, { "epoch": 0.7190112494163869, "grad_norm": 3.462470769882202, "learning_rate": 3.886424879646527e-06, "loss": 3.0667, "step": 20405 }, { "epoch": 0.7190464864293454, "grad_norm": 6.569980144500732, "learning_rate": 3.885521764895843e-06, "loss": 5.262, "step": 20406 }, { "epoch": 0.7190817234423038, "grad_norm": 3.1761670112609863, "learning_rate": 3.884618729787015e-06, "loss": 2.6512, "step": 20407 }, { "epoch": 0.7191169604552622, "grad_norm": 4.681097507476807, "learning_rate": 3.883715774331815e-06, "loss": 3.1771, "step": 20408 }, { "epoch": 0.7191521974682206, "grad_norm": 3.287980556488037, "learning_rate": 3.882812898541991e-06, "loss": 3.1142, "step": 20409 }, { "epoch": 0.719187434481179, "grad_norm": 6.80821418762207, "learning_rate": 3.881910102429309e-06, "loss": 4.1133, "step": 20410 }, { "epoch": 0.7192226714941374, "grad_norm": 1.301056146621704, "learning_rate": 3.881007386005526e-06, "loss": 0.8736, "step": 20411 }, { "epoch": 0.7192579085070958, "grad_norm": 8.332849502563477, "learning_rate": 3.880104749282403e-06, "loss": 5.1904, "step": 20412 }, { "epoch": 0.7192931455200543, "grad_norm": 3.489354133605957, "learning_rate": 3.879202192271695e-06, "loss": 2.7185, "step": 20413 }, { "epoch": 0.7193283825330127, "grad_norm": 4.331324100494385, "learning_rate": 3.878299714985156e-06, "loss": 3.3824, "step": 20414 }, { "epoch": 0.7193636195459711, "grad_norm": 4.374390125274658, "learning_rate": 3.8773973174345455e-06, "loss": 4.877, "step": 20415 }, { "epoch": 0.7193988565589295, "grad_norm": 4.03442907333374, "learning_rate": 3.876494999631614e-06, "loss": 2.651, "step": 20416 }, { "epoch": 0.7194340935718879, "grad_norm": 1.059216022491455, "learning_rate": 3.875592761588115e-06, "loss": 1.0247, "step": 20417 }, { "epoch": 0.7194693305848463, "grad_norm": 3.528088092803955, "learning_rate": 3.8746906033158e-06, "loss": 3.478, "step": 20418 }, { "epoch": 0.7195045675978048, "grad_norm": 1.113682508468628, "learning_rate": 3.873788524826421e-06, "loss": 1.2353, "step": 20419 }, { "epoch": 0.7195398046107632, "grad_norm": 1.1309196949005127, "learning_rate": 3.872886526131726e-06, "loss": 1.0929, "step": 20420 }, { "epoch": 0.7195750416237215, "grad_norm": 3.5954723358154297, "learning_rate": 3.871984607243469e-06, "loss": 3.5172, "step": 20421 }, { "epoch": 0.7196102786366799, "grad_norm": 1.685930848121643, "learning_rate": 3.871082768173385e-06, "loss": 0.8712, "step": 20422 }, { "epoch": 0.7196455156496384, "grad_norm": 3.165194034576416, "learning_rate": 3.8701810089332325e-06, "loss": 3.2393, "step": 20423 }, { "epoch": 0.7196807526625968, "grad_norm": 3.1294331550598145, "learning_rate": 3.869279329534757e-06, "loss": 2.7255, "step": 20424 }, { "epoch": 0.7197159896755552, "grad_norm": 1.1306108236312866, "learning_rate": 3.868377729989695e-06, "loss": 0.989, "step": 20425 }, { "epoch": 0.7197512266885137, "grad_norm": 5.937462329864502, "learning_rate": 3.867476210309791e-06, "loss": 3.6875, "step": 20426 }, { "epoch": 0.719786463701472, "grad_norm": 1.1354553699493408, "learning_rate": 3.866574770506798e-06, "loss": 0.936, "step": 20427 }, { "epoch": 0.7198217007144304, "grad_norm": 3.1088132858276367, "learning_rate": 3.865673410592445e-06, "loss": 3.0491, "step": 20428 }, { "epoch": 0.7198569377273889, "grad_norm": 5.472142219543457, "learning_rate": 3.864772130578478e-06, "loss": 2.8513, "step": 20429 }, { "epoch": 0.7198921747403473, "grad_norm": 1.7126833200454712, "learning_rate": 3.863870930476636e-06, "loss": 1.0061, "step": 20430 }, { "epoch": 0.7199274117533057, "grad_norm": 2.3440065383911133, "learning_rate": 3.862969810298657e-06, "loss": 3.1276, "step": 20431 }, { "epoch": 0.719962648766264, "grad_norm": 4.127840518951416, "learning_rate": 3.862068770056277e-06, "loss": 2.9537, "step": 20432 }, { "epoch": 0.7199978857792225, "grad_norm": 3.852630376815796, "learning_rate": 3.861167809761233e-06, "loss": 3.716, "step": 20433 }, { "epoch": 0.7200331227921809, "grad_norm": 4.047766208648682, "learning_rate": 3.86026692942526e-06, "loss": 3.0989, "step": 20434 }, { "epoch": 0.7200683598051393, "grad_norm": 3.6436679363250732, "learning_rate": 3.859366129060094e-06, "loss": 2.8429, "step": 20435 }, { "epoch": 0.7201035968180978, "grad_norm": 1.30465829372406, "learning_rate": 3.858465408677466e-06, "loss": 0.9669, "step": 20436 }, { "epoch": 0.7201388338310561, "grad_norm": 1.269256830215454, "learning_rate": 3.857564768289107e-06, "loss": 0.6261, "step": 20437 }, { "epoch": 0.7201740708440145, "grad_norm": 5.154746055603027, "learning_rate": 3.85666420790675e-06, "loss": 2.9955, "step": 20438 }, { "epoch": 0.720209307856973, "grad_norm": 2.969101667404175, "learning_rate": 3.855763727542125e-06, "loss": 2.7052, "step": 20439 }, { "epoch": 0.7202445448699314, "grad_norm": 1.1471537351608276, "learning_rate": 3.854863327206962e-06, "loss": 0.7863, "step": 20440 }, { "epoch": 0.7202797818828898, "grad_norm": 4.062792778015137, "learning_rate": 3.853963006912981e-06, "loss": 2.7433, "step": 20441 }, { "epoch": 0.7203150188958481, "grad_norm": 0.8099066615104675, "learning_rate": 3.853062766671918e-06, "loss": 0.6555, "step": 20442 }, { "epoch": 0.7203502559088066, "grad_norm": 1.327619194984436, "learning_rate": 3.852162606495498e-06, "loss": 0.9915, "step": 20443 }, { "epoch": 0.720385492921765, "grad_norm": 4.342833518981934, "learning_rate": 3.851262526395442e-06, "loss": 2.2008, "step": 20444 }, { "epoch": 0.7204207299347234, "grad_norm": 8.158007621765137, "learning_rate": 3.850362526383469e-06, "loss": 2.917, "step": 20445 }, { "epoch": 0.7204559669476819, "grad_norm": 1.2944813966751099, "learning_rate": 3.849462606471316e-06, "loss": 0.9283, "step": 20446 }, { "epoch": 0.7204912039606403, "grad_norm": 1.4871724843978882, "learning_rate": 3.848562766670691e-06, "loss": 0.8618, "step": 20447 }, { "epoch": 0.7205264409735986, "grad_norm": 3.4391305446624756, "learning_rate": 3.847663006993321e-06, "loss": 2.3737, "step": 20448 }, { "epoch": 0.7205616779865571, "grad_norm": 4.291377544403076, "learning_rate": 3.846763327450923e-06, "loss": 3.0164, "step": 20449 }, { "epoch": 0.7205969149995155, "grad_norm": 1.3142640590667725, "learning_rate": 3.845863728055217e-06, "loss": 0.9684, "step": 20450 }, { "epoch": 0.7206321520124739, "grad_norm": 1.31354558467865, "learning_rate": 3.84496420881792e-06, "loss": 0.8709, "step": 20451 }, { "epoch": 0.7206673890254323, "grad_norm": 1.9423962831497192, "learning_rate": 3.844064769750748e-06, "loss": 0.9821, "step": 20452 }, { "epoch": 0.7207026260383907, "grad_norm": 4.2822089195251465, "learning_rate": 3.843165410865416e-06, "loss": 4.3458, "step": 20453 }, { "epoch": 0.7207378630513491, "grad_norm": 1.6672298908233643, "learning_rate": 3.84226613217364e-06, "loss": 0.921, "step": 20454 }, { "epoch": 0.7207731000643075, "grad_norm": 3.1876866817474365, "learning_rate": 3.8413669336871305e-06, "loss": 2.4306, "step": 20455 }, { "epoch": 0.720808337077266, "grad_norm": 2.980353832244873, "learning_rate": 3.840467815417602e-06, "loss": 3.123, "step": 20456 }, { "epoch": 0.7208435740902244, "grad_norm": 4.744766712188721, "learning_rate": 3.8395687773767655e-06, "loss": 2.9838, "step": 20457 }, { "epoch": 0.7208788111031827, "grad_norm": 1.8568185567855835, "learning_rate": 3.83866981957633e-06, "loss": 0.8697, "step": 20458 }, { "epoch": 0.7209140481161412, "grad_norm": 3.488416910171509, "learning_rate": 3.837770942028008e-06, "loss": 3.0492, "step": 20459 }, { "epoch": 0.7209492851290996, "grad_norm": 4.625082015991211, "learning_rate": 3.836872144743496e-06, "loss": 4.8039, "step": 20460 }, { "epoch": 0.720984522142058, "grad_norm": 2.0081634521484375, "learning_rate": 3.835973427734515e-06, "loss": 0.8429, "step": 20461 }, { "epoch": 0.7210197591550165, "grad_norm": 5.231252670288086, "learning_rate": 3.835074791012769e-06, "loss": 3.0143, "step": 20462 }, { "epoch": 0.7210549961679749, "grad_norm": 1.6086666584014893, "learning_rate": 3.834176234589954e-06, "loss": 0.7131, "step": 20463 }, { "epoch": 0.7210902331809332, "grad_norm": 6.340967178344727, "learning_rate": 3.833277758477776e-06, "loss": 2.8531, "step": 20464 }, { "epoch": 0.7211254701938916, "grad_norm": 1.0877103805541992, "learning_rate": 3.832379362687948e-06, "loss": 0.8858, "step": 20465 }, { "epoch": 0.7211607072068501, "grad_norm": 3.2829394340515137, "learning_rate": 3.831481047232161e-06, "loss": 2.7543, "step": 20466 }, { "epoch": 0.7211959442198085, "grad_norm": 2.8082382678985596, "learning_rate": 3.8305828121221186e-06, "loss": 2.9347, "step": 20467 }, { "epoch": 0.7212311812327669, "grad_norm": 1.545129656791687, "learning_rate": 3.829684657369521e-06, "loss": 0.9903, "step": 20468 }, { "epoch": 0.7212664182457253, "grad_norm": 3.7947545051574707, "learning_rate": 3.828786582986067e-06, "loss": 2.817, "step": 20469 }, { "epoch": 0.7213016552586837, "grad_norm": 5.817199230194092, "learning_rate": 3.8278885889834534e-06, "loss": 4.6136, "step": 20470 }, { "epoch": 0.7213368922716421, "grad_norm": 2.9755983352661133, "learning_rate": 3.826990675373377e-06, "loss": 0.5936, "step": 20471 }, { "epoch": 0.7213721292846006, "grad_norm": 10.227463722229004, "learning_rate": 3.826092842167534e-06, "loss": 5.976, "step": 20472 }, { "epoch": 0.721407366297559, "grad_norm": 2.786158561706543, "learning_rate": 3.825195089377619e-06, "loss": 2.7876, "step": 20473 }, { "epoch": 0.7214426033105173, "grad_norm": 3.0728981494903564, "learning_rate": 3.824297417015324e-06, "loss": 2.6909, "step": 20474 }, { "epoch": 0.7214778403234757, "grad_norm": 1.8829809427261353, "learning_rate": 3.823399825092341e-06, "loss": 0.7628, "step": 20475 }, { "epoch": 0.7215130773364342, "grad_norm": 3.2385623455047607, "learning_rate": 3.822502313620362e-06, "loss": 2.7361, "step": 20476 }, { "epoch": 0.7215483143493926, "grad_norm": 3.6846604347229004, "learning_rate": 3.821604882611079e-06, "loss": 2.9125, "step": 20477 }, { "epoch": 0.721583551362351, "grad_norm": 1.7119693756103516, "learning_rate": 3.820707532076181e-06, "loss": 0.6736, "step": 20478 }, { "epoch": 0.7216187883753095, "grad_norm": 4.783674240112305, "learning_rate": 3.819810262027348e-06, "loss": 6.9845, "step": 20479 }, { "epoch": 0.7216540253882678, "grad_norm": 4.226673126220703, "learning_rate": 3.818913072476277e-06, "loss": 3.4844, "step": 20480 }, { "epoch": 0.7216892624012262, "grad_norm": 3.8150112628936768, "learning_rate": 3.818015963434653e-06, "loss": 3.0086, "step": 20481 }, { "epoch": 0.7217244994141847, "grad_norm": 2.937162160873413, "learning_rate": 3.817118934914157e-06, "loss": 0.8079, "step": 20482 }, { "epoch": 0.7217597364271431, "grad_norm": 5.875143051147461, "learning_rate": 3.81622198692647e-06, "loss": 4.4925, "step": 20483 }, { "epoch": 0.7217949734401015, "grad_norm": 3.203746795654297, "learning_rate": 3.815325119483287e-06, "loss": 2.6952, "step": 20484 }, { "epoch": 0.7218302104530598, "grad_norm": 4.1437764167785645, "learning_rate": 3.8144283325962783e-06, "loss": 3.181, "step": 20485 }, { "epoch": 0.7218654474660183, "grad_norm": 7.06322717666626, "learning_rate": 3.813531626277128e-06, "loss": 5.0827, "step": 20486 }, { "epoch": 0.7219006844789767, "grad_norm": 3.236382484436035, "learning_rate": 3.812635000537517e-06, "loss": 2.919, "step": 20487 }, { "epoch": 0.7219359214919351, "grad_norm": 3.2221341133117676, "learning_rate": 3.8117384553891224e-06, "loss": 2.8697, "step": 20488 }, { "epoch": 0.7219711585048936, "grad_norm": 1.418203592300415, "learning_rate": 3.8108419908436233e-06, "loss": 0.9417, "step": 20489 }, { "epoch": 0.722006395517852, "grad_norm": 2.9818716049194336, "learning_rate": 3.809945606912696e-06, "loss": 3.0231, "step": 20490 }, { "epoch": 0.7220416325308103, "grad_norm": 3.5848937034606934, "learning_rate": 3.8090493036080143e-06, "loss": 2.3563, "step": 20491 }, { "epoch": 0.7220768695437688, "grad_norm": 22.802021026611328, "learning_rate": 3.808153080941256e-06, "loss": 5.2277, "step": 20492 }, { "epoch": 0.7221121065567272, "grad_norm": 1.1917330026626587, "learning_rate": 3.807256938924091e-06, "loss": 1.3526, "step": 20493 }, { "epoch": 0.7221473435696856, "grad_norm": 0.8522911071777344, "learning_rate": 3.806360877568194e-06, "loss": 0.7503, "step": 20494 }, { "epoch": 0.7221825805826441, "grad_norm": 1.3424088954925537, "learning_rate": 3.805464896885235e-06, "loss": 0.9633, "step": 20495 }, { "epoch": 0.7222178175956024, "grad_norm": 0.87506103515625, "learning_rate": 3.8045689968868847e-06, "loss": 0.6898, "step": 20496 }, { "epoch": 0.7222530546085608, "grad_norm": 3.1332075595855713, "learning_rate": 3.803673177584816e-06, "loss": 2.9607, "step": 20497 }, { "epoch": 0.7222882916215192, "grad_norm": 1.5537158250808716, "learning_rate": 3.8027774389906867e-06, "loss": 0.858, "step": 20498 }, { "epoch": 0.7223235286344777, "grad_norm": 1.2491177320480347, "learning_rate": 3.8018817811161746e-06, "loss": 1.0265, "step": 20499 }, { "epoch": 0.7223587656474361, "grad_norm": 3.2944841384887695, "learning_rate": 3.800986203972944e-06, "loss": 0.8088, "step": 20500 }, { "epoch": 0.7223940026603944, "grad_norm": 0.8810132145881653, "learning_rate": 3.800090707572656e-06, "loss": 0.7879, "step": 20501 }, { "epoch": 0.7224292396733529, "grad_norm": 1.7453813552856445, "learning_rate": 3.7991952919269704e-06, "loss": 0.7503, "step": 20502 }, { "epoch": 0.7224644766863113, "grad_norm": 1.0931918621063232, "learning_rate": 3.798299957047564e-06, "loss": 0.9412, "step": 20503 }, { "epoch": 0.7224997136992697, "grad_norm": 3.0384864807128906, "learning_rate": 3.797404702946088e-06, "loss": 2.56, "step": 20504 }, { "epoch": 0.7225349507122282, "grad_norm": 1.687219262123108, "learning_rate": 3.796509529634206e-06, "loss": 0.9783, "step": 20505 }, { "epoch": 0.7225701877251866, "grad_norm": 1.1701701879501343, "learning_rate": 3.795614437123576e-06, "loss": 0.959, "step": 20506 }, { "epoch": 0.7226054247381449, "grad_norm": 1.0910491943359375, "learning_rate": 3.794719425425859e-06, "loss": 0.777, "step": 20507 }, { "epoch": 0.7226406617511033, "grad_norm": 6.881234169006348, "learning_rate": 3.7938244945527124e-06, "loss": 3.5306, "step": 20508 }, { "epoch": 0.7226758987640618, "grad_norm": 3.8584816455841064, "learning_rate": 3.7929296445157915e-06, "loss": 3.0879, "step": 20509 }, { "epoch": 0.7227111357770202, "grad_norm": 0.9750163555145264, "learning_rate": 3.7920348753267534e-06, "loss": 0.9105, "step": 20510 }, { "epoch": 0.7227463727899786, "grad_norm": 5.4012250900268555, "learning_rate": 3.791140186997252e-06, "loss": 5.5571, "step": 20511 }, { "epoch": 0.722781609802937, "grad_norm": 4.296968460083008, "learning_rate": 3.790245579538939e-06, "loss": 4.8156, "step": 20512 }, { "epoch": 0.7228168468158954, "grad_norm": 7.990161895751953, "learning_rate": 3.7893510529634703e-06, "loss": 5.2211, "step": 20513 }, { "epoch": 0.7228520838288538, "grad_norm": 3.6946518421173096, "learning_rate": 3.7884566072824936e-06, "loss": 2.8439, "step": 20514 }, { "epoch": 0.7228873208418123, "grad_norm": 1.0586981773376465, "learning_rate": 3.787562242507662e-06, "loss": 0.7208, "step": 20515 }, { "epoch": 0.7229225578547707, "grad_norm": 0.9125272035598755, "learning_rate": 3.786667958650626e-06, "loss": 0.7979, "step": 20516 }, { "epoch": 0.722957794867729, "grad_norm": 3.565974473953247, "learning_rate": 3.7857737557230256e-06, "loss": 2.7842, "step": 20517 }, { "epoch": 0.7229930318806874, "grad_norm": 5.749906539916992, "learning_rate": 3.7848796337365156e-06, "loss": 3.2796, "step": 20518 }, { "epoch": 0.7230282688936459, "grad_norm": 4.951615810394287, "learning_rate": 3.7839855927027446e-06, "loss": 4.6696, "step": 20519 }, { "epoch": 0.7230635059066043, "grad_norm": 1.4068886041641235, "learning_rate": 3.783091632633349e-06, "loss": 1.004, "step": 20520 }, { "epoch": 0.7230987429195627, "grad_norm": 1.4316327571868896, "learning_rate": 3.782197753539977e-06, "loss": 0.7946, "step": 20521 }, { "epoch": 0.7231339799325212, "grad_norm": 3.558886766433716, "learning_rate": 3.781303955434271e-06, "loss": 2.81, "step": 20522 }, { "epoch": 0.7231692169454795, "grad_norm": 4.8403000831604, "learning_rate": 3.780410238327874e-06, "loss": 3.1088, "step": 20523 }, { "epoch": 0.7232044539584379, "grad_norm": 6.360695838928223, "learning_rate": 3.779516602232425e-06, "loss": 6.4367, "step": 20524 }, { "epoch": 0.7232396909713964, "grad_norm": 3.017886161804199, "learning_rate": 3.7786230471595653e-06, "loss": 2.392, "step": 20525 }, { "epoch": 0.7232749279843548, "grad_norm": 3.1970009803771973, "learning_rate": 3.777729573120933e-06, "loss": 3.0471, "step": 20526 }, { "epoch": 0.7233101649973132, "grad_norm": 8.272468566894531, "learning_rate": 3.776836180128165e-06, "loss": 3.2496, "step": 20527 }, { "epoch": 0.7233454020102716, "grad_norm": 1.7415467500686646, "learning_rate": 3.775942868192898e-06, "loss": 0.806, "step": 20528 }, { "epoch": 0.72338063902323, "grad_norm": 4.649087905883789, "learning_rate": 3.7750496373267686e-06, "loss": 3.2929, "step": 20529 }, { "epoch": 0.7234158760361884, "grad_norm": 4.1647233963012695, "learning_rate": 3.7741564875414104e-06, "loss": 2.7989, "step": 20530 }, { "epoch": 0.7234511130491468, "grad_norm": 1.521028995513916, "learning_rate": 3.7732634188484572e-06, "loss": 0.8895, "step": 20531 }, { "epoch": 0.7234863500621053, "grad_norm": 2.942251682281494, "learning_rate": 3.7723704312595445e-06, "loss": 2.0785, "step": 20532 }, { "epoch": 0.7235215870750636, "grad_norm": 1.8311882019042969, "learning_rate": 3.7714775247862935e-06, "loss": 0.8769, "step": 20533 }, { "epoch": 0.723556824088022, "grad_norm": 5.688077926635742, "learning_rate": 3.770584699440344e-06, "loss": 2.6743, "step": 20534 }, { "epoch": 0.7235920611009805, "grad_norm": 1.4632549285888672, "learning_rate": 3.7696919552333265e-06, "loss": 0.7888, "step": 20535 }, { "epoch": 0.7236272981139389, "grad_norm": 1.7113569974899292, "learning_rate": 3.768799292176861e-06, "loss": 2.2981, "step": 20536 }, { "epoch": 0.7236625351268973, "grad_norm": 2.775960683822632, "learning_rate": 3.7679067102825752e-06, "loss": 2.2996, "step": 20537 }, { "epoch": 0.7236977721398558, "grad_norm": 6.346614837646484, "learning_rate": 3.7670142095621053e-06, "loss": 4.9784, "step": 20538 }, { "epoch": 0.7237330091528141, "grad_norm": 1.446677803993225, "learning_rate": 3.766121790027065e-06, "loss": 0.9011, "step": 20539 }, { "epoch": 0.7237682461657725, "grad_norm": 7.249956130981445, "learning_rate": 3.7652294516890844e-06, "loss": 5.3334, "step": 20540 }, { "epoch": 0.7238034831787309, "grad_norm": 1.0518680810928345, "learning_rate": 3.764337194559783e-06, "loss": 0.8733, "step": 20541 }, { "epoch": 0.7238387201916894, "grad_norm": 4.502711296081543, "learning_rate": 3.7634450186507833e-06, "loss": 2.9162, "step": 20542 }, { "epoch": 0.7238739572046478, "grad_norm": 3.0054752826690674, "learning_rate": 3.7625529239737078e-06, "loss": 2.7757, "step": 20543 }, { "epoch": 0.7239091942176061, "grad_norm": 5.581301689147949, "learning_rate": 3.7616609105401747e-06, "loss": 3.3158, "step": 20544 }, { "epoch": 0.7239444312305646, "grad_norm": 3.673414707183838, "learning_rate": 3.7607689783618016e-06, "loss": 3.1966, "step": 20545 }, { "epoch": 0.723979668243523, "grad_norm": 9.982139587402344, "learning_rate": 3.759877127450209e-06, "loss": 3.1186, "step": 20546 }, { "epoch": 0.7240149052564814, "grad_norm": 3.6543543338775635, "learning_rate": 3.758985357817011e-06, "loss": 3.0422, "step": 20547 }, { "epoch": 0.7240501422694399, "grad_norm": 1.9175150394439697, "learning_rate": 3.7580936694738246e-06, "loss": 0.9088, "step": 20548 }, { "epoch": 0.7240853792823982, "grad_norm": 1.0831841230392456, "learning_rate": 3.7572020624322615e-06, "loss": 0.9192, "step": 20549 }, { "epoch": 0.7241206162953566, "grad_norm": 1.3312170505523682, "learning_rate": 3.7563105367039378e-06, "loss": 0.6326, "step": 20550 }, { "epoch": 0.724155853308315, "grad_norm": 5.106521129608154, "learning_rate": 3.7554190923004687e-06, "loss": 2.5831, "step": 20551 }, { "epoch": 0.7241910903212735, "grad_norm": 3.686549186706543, "learning_rate": 3.7545277292334536e-06, "loss": 2.2387, "step": 20552 }, { "epoch": 0.7242263273342319, "grad_norm": 1.2459290027618408, "learning_rate": 3.7536364475145138e-06, "loss": 0.8121, "step": 20553 }, { "epoch": 0.7242615643471902, "grad_norm": 3.042628049850464, "learning_rate": 3.752745247155258e-06, "loss": 2.7249, "step": 20554 }, { "epoch": 0.7242968013601487, "grad_norm": 0.9793987274169922, "learning_rate": 3.7518541281672895e-06, "loss": 0.712, "step": 20555 }, { "epoch": 0.7243320383731071, "grad_norm": 1.0514224767684937, "learning_rate": 3.7509630905622117e-06, "loss": 0.9059, "step": 20556 }, { "epoch": 0.7243672753860655, "grad_norm": 6.389773368835449, "learning_rate": 3.7500721343516434e-06, "loss": 2.4874, "step": 20557 }, { "epoch": 0.724402512399024, "grad_norm": 1.7592414617538452, "learning_rate": 3.7491812595471777e-06, "loss": 0.8717, "step": 20558 }, { "epoch": 0.7244377494119824, "grad_norm": 1.1861222982406616, "learning_rate": 3.7482904661604226e-06, "loss": 1.0833, "step": 20559 }, { "epoch": 0.7244729864249407, "grad_norm": 1.0959261655807495, "learning_rate": 3.7473997542029805e-06, "loss": 0.7577, "step": 20560 }, { "epoch": 0.7245082234378992, "grad_norm": 4.952012062072754, "learning_rate": 3.7465091236864527e-06, "loss": 5.0071, "step": 20561 }, { "epoch": 0.7245434604508576, "grad_norm": 6.037342071533203, "learning_rate": 3.7456185746224405e-06, "loss": 4.8875, "step": 20562 }, { "epoch": 0.724578697463816, "grad_norm": 2.8769376277923584, "learning_rate": 3.744728107022543e-06, "loss": 2.2616, "step": 20563 }, { "epoch": 0.7246139344767744, "grad_norm": 3.9742767810821533, "learning_rate": 3.7438377208983577e-06, "loss": 2.6882, "step": 20564 }, { "epoch": 0.7246491714897328, "grad_norm": 2.015568494796753, "learning_rate": 3.7429474162614844e-06, "loss": 0.8395, "step": 20565 }, { "epoch": 0.7246844085026912, "grad_norm": 5.147274017333984, "learning_rate": 3.742057193123517e-06, "loss": 3.4783, "step": 20566 }, { "epoch": 0.7247196455156496, "grad_norm": 1.5869196653366089, "learning_rate": 3.741167051496053e-06, "loss": 1.0334, "step": 20567 }, { "epoch": 0.7247548825286081, "grad_norm": 2.903280019760132, "learning_rate": 3.7402769913906843e-06, "loss": 3.0941, "step": 20568 }, { "epoch": 0.7247901195415665, "grad_norm": 0.9588490724563599, "learning_rate": 3.739387012819006e-06, "loss": 0.8215, "step": 20569 }, { "epoch": 0.7248253565545248, "grad_norm": 3.6463329792022705, "learning_rate": 3.738497115792612e-06, "loss": 2.8446, "step": 20570 }, { "epoch": 0.7248605935674833, "grad_norm": 3.0419063568115234, "learning_rate": 3.737607300323084e-06, "loss": 2.8013, "step": 20571 }, { "epoch": 0.7248958305804417, "grad_norm": 1.8380770683288574, "learning_rate": 3.7367175664220224e-06, "loss": 1.0377, "step": 20572 }, { "epoch": 0.7249310675934001, "grad_norm": 8.428437232971191, "learning_rate": 3.7358279141010156e-06, "loss": 4.8234, "step": 20573 }, { "epoch": 0.7249663046063585, "grad_norm": 1.8734105825424194, "learning_rate": 3.7349383433716445e-06, "loss": 0.9299, "step": 20574 }, { "epoch": 0.725001541619317, "grad_norm": 2.94767427444458, "learning_rate": 3.7340488542454965e-06, "loss": 2.8843, "step": 20575 }, { "epoch": 0.7250367786322753, "grad_norm": 1.4837806224822998, "learning_rate": 3.733159446734167e-06, "loss": 0.6578, "step": 20576 }, { "epoch": 0.7250720156452337, "grad_norm": 1.4457826614379883, "learning_rate": 3.732270120849231e-06, "loss": 0.7238, "step": 20577 }, { "epoch": 0.7251072526581922, "grad_norm": 3.8759989738464355, "learning_rate": 3.731380876602274e-06, "loss": 2.8347, "step": 20578 }, { "epoch": 0.7251424896711506, "grad_norm": 1.970219373703003, "learning_rate": 3.73049171400488e-06, "loss": 1.1395, "step": 20579 }, { "epoch": 0.725177726684109, "grad_norm": 4.7803802490234375, "learning_rate": 3.729602633068631e-06, "loss": 4.6455, "step": 20580 }, { "epoch": 0.7252129636970674, "grad_norm": 1.0503959655761719, "learning_rate": 3.7287136338051046e-06, "loss": 0.9899, "step": 20581 }, { "epoch": 0.7252482007100258, "grad_norm": 6.014061450958252, "learning_rate": 3.727824716225883e-06, "loss": 4.1501, "step": 20582 }, { "epoch": 0.7252834377229842, "grad_norm": 3.5939252376556396, "learning_rate": 3.726935880342544e-06, "loss": 2.5265, "step": 20583 }, { "epoch": 0.7253186747359426, "grad_norm": 1.01682448387146, "learning_rate": 3.726047126166663e-06, "loss": 0.8807, "step": 20584 }, { "epoch": 0.7253539117489011, "grad_norm": 4.516378879547119, "learning_rate": 3.7251584537098186e-06, "loss": 4.1348, "step": 20585 }, { "epoch": 0.7253891487618594, "grad_norm": 5.325130462646484, "learning_rate": 3.7242698629835838e-06, "loss": 3.1235, "step": 20586 }, { "epoch": 0.7254243857748178, "grad_norm": 1.184165358543396, "learning_rate": 3.7233813539995334e-06, "loss": 1.1018, "step": 20587 }, { "epoch": 0.7254596227877763, "grad_norm": 0.7313258051872253, "learning_rate": 3.72249292676924e-06, "loss": 0.7528, "step": 20588 }, { "epoch": 0.7254948598007347, "grad_norm": 1.5253691673278809, "learning_rate": 3.72160458130428e-06, "loss": 1.1766, "step": 20589 }, { "epoch": 0.7255300968136931, "grad_norm": 4.553045749664307, "learning_rate": 3.720716317616213e-06, "loss": 3.0145, "step": 20590 }, { "epoch": 0.7255653338266516, "grad_norm": 1.2101342678070068, "learning_rate": 3.71982813571662e-06, "loss": 0.6816, "step": 20591 }, { "epoch": 0.7256005708396099, "grad_norm": 3.5220282077789307, "learning_rate": 3.718940035617068e-06, "loss": 3.1597, "step": 20592 }, { "epoch": 0.7256358078525683, "grad_norm": 4.214145660400391, "learning_rate": 3.718052017329119e-06, "loss": 3.1875, "step": 20593 }, { "epoch": 0.7256710448655268, "grad_norm": 1.7390820980072021, "learning_rate": 3.717164080864338e-06, "loss": 0.9073, "step": 20594 }, { "epoch": 0.7257062818784852, "grad_norm": 1.3466784954071045, "learning_rate": 3.7162762262343044e-06, "loss": 0.7248, "step": 20595 }, { "epoch": 0.7257415188914436, "grad_norm": 5.470798492431641, "learning_rate": 3.715388453450568e-06, "loss": 4.5343, "step": 20596 }, { "epoch": 0.7257767559044019, "grad_norm": 4.282473087310791, "learning_rate": 3.7145007625246986e-06, "loss": 4.9097, "step": 20597 }, { "epoch": 0.7258119929173604, "grad_norm": 8.445928573608398, "learning_rate": 3.7136131534682573e-06, "loss": 6.9785, "step": 20598 }, { "epoch": 0.7258472299303188, "grad_norm": 3.9107253551483154, "learning_rate": 3.7127256262928045e-06, "loss": 3.0959, "step": 20599 }, { "epoch": 0.7258824669432772, "grad_norm": 1.7448424100875854, "learning_rate": 3.7118381810099023e-06, "loss": 0.9332, "step": 20600 }, { "epoch": 0.7259177039562357, "grad_norm": 4.572019100189209, "learning_rate": 3.710950817631108e-06, "loss": 3.289, "step": 20601 }, { "epoch": 0.725952940969194, "grad_norm": 3.2257118225097656, "learning_rate": 3.7100635361679816e-06, "loss": 3.1211, "step": 20602 }, { "epoch": 0.7259881779821524, "grad_norm": 3.569669485092163, "learning_rate": 3.7091763366320777e-06, "loss": 4.9342, "step": 20603 }, { "epoch": 0.7260234149951109, "grad_norm": 4.227556228637695, "learning_rate": 3.708289219034954e-06, "loss": 3.0849, "step": 20604 }, { "epoch": 0.7260586520080693, "grad_norm": 5.264559268951416, "learning_rate": 3.7074021833881645e-06, "loss": 5.0854, "step": 20605 }, { "epoch": 0.7260938890210277, "grad_norm": 1.0704014301300049, "learning_rate": 3.7065152297032623e-06, "loss": 0.7702, "step": 20606 }, { "epoch": 0.726129126033986, "grad_norm": 4.408141136169434, "learning_rate": 3.705628357991802e-06, "loss": 2.5894, "step": 20607 }, { "epoch": 0.7261643630469445, "grad_norm": 4.878468036651611, "learning_rate": 3.7047415682653376e-06, "loss": 3.568, "step": 20608 }, { "epoch": 0.7261996000599029, "grad_norm": 1.163474440574646, "learning_rate": 3.7038548605354086e-06, "loss": 0.9158, "step": 20609 }, { "epoch": 0.7262348370728613, "grad_norm": 4.634936809539795, "learning_rate": 3.7029682348135753e-06, "loss": 3.0641, "step": 20610 }, { "epoch": 0.7262700740858198, "grad_norm": 11.017404556274414, "learning_rate": 3.702081691111387e-06, "loss": 2.6372, "step": 20611 }, { "epoch": 0.7263053110987782, "grad_norm": 4.812758445739746, "learning_rate": 3.701195229440383e-06, "loss": 3.404, "step": 20612 }, { "epoch": 0.7263405481117365, "grad_norm": 3.122166395187378, "learning_rate": 3.70030884981211e-06, "loss": 4.5528, "step": 20613 }, { "epoch": 0.726375785124695, "grad_norm": 1.3233402967453003, "learning_rate": 3.6994225522381244e-06, "loss": 1.1937, "step": 20614 }, { "epoch": 0.7264110221376534, "grad_norm": 3.885556697845459, "learning_rate": 3.6985363367299585e-06, "loss": 2.9208, "step": 20615 }, { "epoch": 0.7264462591506118, "grad_norm": 1.4466545581817627, "learning_rate": 3.697650203299159e-06, "loss": 1.0162, "step": 20616 }, { "epoch": 0.7264814961635702, "grad_norm": 3.3334925174713135, "learning_rate": 3.6967641519572662e-06, "loss": 3.2603, "step": 20617 }, { "epoch": 0.7265167331765287, "grad_norm": 0.9803867340087891, "learning_rate": 3.6958781827158287e-06, "loss": 0.9191, "step": 20618 }, { "epoch": 0.726551970189487, "grad_norm": 3.2776734828948975, "learning_rate": 3.6949922955863783e-06, "loss": 3.1713, "step": 20619 }, { "epoch": 0.7265872072024454, "grad_norm": 3.8187267780303955, "learning_rate": 3.694106490580457e-06, "loss": 2.8604, "step": 20620 }, { "epoch": 0.7266224442154039, "grad_norm": 3.397378921508789, "learning_rate": 3.693220767709601e-06, "loss": 2.9967, "step": 20621 }, { "epoch": 0.7266576812283623, "grad_norm": 1.240676760673523, "learning_rate": 3.692335126985348e-06, "loss": 0.8005, "step": 20622 }, { "epoch": 0.7266929182413207, "grad_norm": 7.204624176025391, "learning_rate": 3.691449568419233e-06, "loss": 6.7176, "step": 20623 }, { "epoch": 0.7267281552542791, "grad_norm": 1.1318832635879517, "learning_rate": 3.6905640920227916e-06, "loss": 0.9088, "step": 20624 }, { "epoch": 0.7267633922672375, "grad_norm": 3.911867618560791, "learning_rate": 3.6896786978075563e-06, "loss": 3.1374, "step": 20625 }, { "epoch": 0.7267986292801959, "grad_norm": 3.911867618560791, "learning_rate": 3.6896786978075563e-06, "loss": 2.9665, "step": 20626 }, { "epoch": 0.7268338662931544, "grad_norm": 1.7192323207855225, "learning_rate": 3.68879338578506e-06, "loss": 0.8372, "step": 20627 }, { "epoch": 0.7268691033061128, "grad_norm": 1.6376323699951172, "learning_rate": 3.687908155966837e-06, "loss": 0.908, "step": 20628 }, { "epoch": 0.7269043403190711, "grad_norm": 1.4643751382827759, "learning_rate": 3.6870230083644076e-06, "loss": 0.7421, "step": 20629 }, { "epoch": 0.7269395773320295, "grad_norm": 3.449612617492676, "learning_rate": 3.6861379429893107e-06, "loss": 2.1599, "step": 20630 }, { "epoch": 0.726974814344988, "grad_norm": 1.483784556388855, "learning_rate": 3.685252959853075e-06, "loss": 0.9502, "step": 20631 }, { "epoch": 0.7270100513579464, "grad_norm": 6.736733436584473, "learning_rate": 3.6843680589672204e-06, "loss": 6.7046, "step": 20632 }, { "epoch": 0.7270452883709048, "grad_norm": 3.3325583934783936, "learning_rate": 3.683483240343273e-06, "loss": 2.9387, "step": 20633 }, { "epoch": 0.7270805253838633, "grad_norm": 1.3364924192428589, "learning_rate": 3.6825985039927682e-06, "loss": 0.9936, "step": 20634 }, { "epoch": 0.7271157623968216, "grad_norm": 1.0485535860061646, "learning_rate": 3.6817138499272186e-06, "loss": 1.1634, "step": 20635 }, { "epoch": 0.72715099940978, "grad_norm": 3.754915714263916, "learning_rate": 3.6808292781581513e-06, "loss": 2.9105, "step": 20636 }, { "epoch": 0.7271862364227385, "grad_norm": 3.5440332889556885, "learning_rate": 3.679944788697084e-06, "loss": 3.2253, "step": 20637 }, { "epoch": 0.7272214734356969, "grad_norm": 1.290331482887268, "learning_rate": 3.679060381555547e-06, "loss": 0.6767, "step": 20638 }, { "epoch": 0.7272567104486553, "grad_norm": 1.2235981225967407, "learning_rate": 3.6781760567450508e-06, "loss": 1.1583, "step": 20639 }, { "epoch": 0.7272919474616136, "grad_norm": 3.0089375972747803, "learning_rate": 3.6772918142771175e-06, "loss": 2.6815, "step": 20640 }, { "epoch": 0.7273271844745721, "grad_norm": 1.2021313905715942, "learning_rate": 3.676407654163263e-06, "loss": 0.8603, "step": 20641 }, { "epoch": 0.7273624214875305, "grad_norm": 2.9704337120056152, "learning_rate": 3.6755235764150044e-06, "loss": 2.6114, "step": 20642 }, { "epoch": 0.7273976585004889, "grad_norm": 2.6130106449127197, "learning_rate": 3.6746395810438575e-06, "loss": 3.1627, "step": 20643 }, { "epoch": 0.7274328955134474, "grad_norm": 4.216840744018555, "learning_rate": 3.673755668061334e-06, "loss": 3.2307, "step": 20644 }, { "epoch": 0.7274681325264057, "grad_norm": 1.8223872184753418, "learning_rate": 3.672871837478951e-06, "loss": 0.7678, "step": 20645 }, { "epoch": 0.7275033695393641, "grad_norm": 1.1040111780166626, "learning_rate": 3.671988089308216e-06, "loss": 0.7997, "step": 20646 }, { "epoch": 0.7275386065523226, "grad_norm": 1.8093370199203491, "learning_rate": 3.6711044235606472e-06, "loss": 0.8575, "step": 20647 }, { "epoch": 0.727573843565281, "grad_norm": 1.163871169090271, "learning_rate": 3.6702208402477426e-06, "loss": 1.1787, "step": 20648 }, { "epoch": 0.7276090805782394, "grad_norm": 1.2777281999588013, "learning_rate": 3.6693373393810206e-06, "loss": 0.9349, "step": 20649 }, { "epoch": 0.7276443175911977, "grad_norm": 5.086383819580078, "learning_rate": 3.668453920971989e-06, "loss": 4.5996, "step": 20650 }, { "epoch": 0.7276795546041562, "grad_norm": 6.489931583404541, "learning_rate": 3.6675705850321497e-06, "loss": 7.8272, "step": 20651 }, { "epoch": 0.7277147916171146, "grad_norm": 1.8692708015441895, "learning_rate": 3.666687331573006e-06, "loss": 0.9383, "step": 20652 }, { "epoch": 0.727750028630073, "grad_norm": 1.3431583642959595, "learning_rate": 3.6658041606060736e-06, "loss": 0.6666, "step": 20653 }, { "epoch": 0.7277852656430315, "grad_norm": 4.83924674987793, "learning_rate": 3.6649210721428454e-06, "loss": 4.9576, "step": 20654 }, { "epoch": 0.7278205026559899, "grad_norm": 1.4170161485671997, "learning_rate": 3.664038066194827e-06, "loss": 0.9368, "step": 20655 }, { "epoch": 0.7278557396689482, "grad_norm": 1.07260000705719, "learning_rate": 3.6631551427735156e-06, "loss": 0.6706, "step": 20656 }, { "epoch": 0.7278909766819067, "grad_norm": 3.226529598236084, "learning_rate": 3.6622723018904238e-06, "loss": 3.3066, "step": 20657 }, { "epoch": 0.7279262136948651, "grad_norm": 4.916079044342041, "learning_rate": 3.661389543557039e-06, "loss": 5.1392, "step": 20658 }, { "epoch": 0.7279614507078235, "grad_norm": 15.375253677368164, "learning_rate": 3.660506867784862e-06, "loss": 8.5494, "step": 20659 }, { "epoch": 0.7279966877207819, "grad_norm": 6.923789978027344, "learning_rate": 3.6596242745853915e-06, "loss": 4.1911, "step": 20660 }, { "epoch": 0.7280319247337403, "grad_norm": 4.510317325592041, "learning_rate": 3.658741763970122e-06, "loss": 4.9704, "step": 20661 }, { "epoch": 0.7280671617466987, "grad_norm": 1.6174200773239136, "learning_rate": 3.6578593359505486e-06, "loss": 0.8134, "step": 20662 }, { "epoch": 0.7281023987596571, "grad_norm": 3.144591808319092, "learning_rate": 3.656976990538166e-06, "loss": 2.6335, "step": 20663 }, { "epoch": 0.7281376357726156, "grad_norm": 1.2448009252548218, "learning_rate": 3.6560947277444645e-06, "loss": 0.7922, "step": 20664 }, { "epoch": 0.728172872785574, "grad_norm": 2.082476854324341, "learning_rate": 3.6552125475809384e-06, "loss": 0.9168, "step": 20665 }, { "epoch": 0.7282081097985323, "grad_norm": 1.1725337505340576, "learning_rate": 3.6543304500590806e-06, "loss": 0.8251, "step": 20666 }, { "epoch": 0.7282433468114908, "grad_norm": 4.741247653961182, "learning_rate": 3.65344843519037e-06, "loss": 4.9652, "step": 20667 }, { "epoch": 0.7282785838244492, "grad_norm": 5.043947219848633, "learning_rate": 3.6525665029863057e-06, "loss": 2.9398, "step": 20668 }, { "epoch": 0.7283138208374076, "grad_norm": 1.2645622491836548, "learning_rate": 3.651684653458374e-06, "loss": 0.7904, "step": 20669 }, { "epoch": 0.7283490578503661, "grad_norm": 3.3062477111816406, "learning_rate": 3.6508028866180545e-06, "loss": 3.981, "step": 20670 }, { "epoch": 0.7283842948633245, "grad_norm": 5.301288604736328, "learning_rate": 3.649921202476834e-06, "loss": 3.1923, "step": 20671 }, { "epoch": 0.7284195318762828, "grad_norm": 1.0352730751037598, "learning_rate": 3.649039601046205e-06, "loss": 0.8736, "step": 20672 }, { "epoch": 0.7284547688892412, "grad_norm": 4.727250099182129, "learning_rate": 3.6481580823376407e-06, "loss": 4.2177, "step": 20673 }, { "epoch": 0.7284900059021997, "grad_norm": 1.3275309801101685, "learning_rate": 3.6472766463626262e-06, "loss": 0.9794, "step": 20674 }, { "epoch": 0.7285252429151581, "grad_norm": 1.404636025428772, "learning_rate": 3.6463952931326395e-06, "loss": 0.6501, "step": 20675 }, { "epoch": 0.7285604799281165, "grad_norm": 3.1603987216949463, "learning_rate": 3.6455140226591703e-06, "loss": 3.1603, "step": 20676 }, { "epoch": 0.728595716941075, "grad_norm": 1.0428082942962646, "learning_rate": 3.6446328349536864e-06, "loss": 0.8668, "step": 20677 }, { "epoch": 0.7286309539540333, "grad_norm": 1.80083429813385, "learning_rate": 3.6437517300276702e-06, "loss": 0.7685, "step": 20678 }, { "epoch": 0.7286661909669917, "grad_norm": 5.455737113952637, "learning_rate": 3.6428707078925974e-06, "loss": 3.2155, "step": 20679 }, { "epoch": 0.7287014279799502, "grad_norm": 0.9738062620162964, "learning_rate": 3.6419897685599427e-06, "loss": 0.8753, "step": 20680 }, { "epoch": 0.7287366649929086, "grad_norm": 5.937668800354004, "learning_rate": 3.6411089120411814e-06, "loss": 3.3524, "step": 20681 }, { "epoch": 0.728771902005867, "grad_norm": 0.9508028626441956, "learning_rate": 3.6402281383477868e-06, "loss": 0.762, "step": 20682 }, { "epoch": 0.7288071390188253, "grad_norm": 9.242121696472168, "learning_rate": 3.6393474474912303e-06, "loss": 4.8471, "step": 20683 }, { "epoch": 0.7288423760317838, "grad_norm": 0.9606651067733765, "learning_rate": 3.6384668394829838e-06, "loss": 0.667, "step": 20684 }, { "epoch": 0.7288776130447422, "grad_norm": 1.716921329498291, "learning_rate": 3.6375863143345202e-06, "loss": 0.8593, "step": 20685 }, { "epoch": 0.7289128500577006, "grad_norm": 3.5904319286346436, "learning_rate": 3.6367058720572988e-06, "loss": 2.9044, "step": 20686 }, { "epoch": 0.7289480870706591, "grad_norm": 8.619160652160645, "learning_rate": 3.635825512662797e-06, "loss": 5.2724, "step": 20687 }, { "epoch": 0.7289833240836174, "grad_norm": 2.9854636192321777, "learning_rate": 3.6349452361624824e-06, "loss": 2.5527, "step": 20688 }, { "epoch": 0.7290185610965758, "grad_norm": 1.3140301704406738, "learning_rate": 3.6340650425678137e-06, "loss": 0.8249, "step": 20689 }, { "epoch": 0.7290537981095343, "grad_norm": 1.048925757408142, "learning_rate": 3.6331849318902544e-06, "loss": 0.8133, "step": 20690 }, { "epoch": 0.7290890351224927, "grad_norm": 25.9361572265625, "learning_rate": 3.632304904141276e-06, "loss": 4.7396, "step": 20691 }, { "epoch": 0.7291242721354511, "grad_norm": 0.9899535179138184, "learning_rate": 3.6314249593323415e-06, "loss": 0.8529, "step": 20692 }, { "epoch": 0.7291595091484094, "grad_norm": 1.151452660560608, "learning_rate": 3.6305450974749045e-06, "loss": 0.8455, "step": 20693 }, { "epoch": 0.7291947461613679, "grad_norm": 2.8244099617004395, "learning_rate": 3.6296653185804254e-06, "loss": 3.1398, "step": 20694 }, { "epoch": 0.7292299831743263, "grad_norm": 3.2663793563842773, "learning_rate": 3.6287856226603736e-06, "loss": 3.5158, "step": 20695 }, { "epoch": 0.7292652201872847, "grad_norm": 5.000687599182129, "learning_rate": 3.6279060097261975e-06, "loss": 4.8382, "step": 20696 }, { "epoch": 0.7293004572002432, "grad_norm": 11.689985275268555, "learning_rate": 3.6270264797893574e-06, "loss": 3.3461, "step": 20697 }, { "epoch": 0.7293356942132015, "grad_norm": 2.8841426372528076, "learning_rate": 3.626147032861308e-06, "loss": 2.6835, "step": 20698 }, { "epoch": 0.7293709312261599, "grad_norm": 3.1406455039978027, "learning_rate": 3.625267668953507e-06, "loss": 2.6953, "step": 20699 }, { "epoch": 0.7294061682391184, "grad_norm": 1.2400332689285278, "learning_rate": 3.6243883880774055e-06, "loss": 0.7271, "step": 20700 }, { "epoch": 0.7294414052520768, "grad_norm": 4.473021507263184, "learning_rate": 3.6235091902444573e-06, "loss": 4.8742, "step": 20701 }, { "epoch": 0.7294766422650352, "grad_norm": 1.708120346069336, "learning_rate": 3.622630075466115e-06, "loss": 0.6977, "step": 20702 }, { "epoch": 0.7295118792779937, "grad_norm": 4.778226375579834, "learning_rate": 3.621751043753828e-06, "loss": 4.8754, "step": 20703 }, { "epoch": 0.729547116290952, "grad_norm": 23.518178939819336, "learning_rate": 3.62087209511905e-06, "loss": 5.9481, "step": 20704 }, { "epoch": 0.7295823533039104, "grad_norm": 3.9590394496917725, "learning_rate": 3.6199932295732177e-06, "loss": 2.976, "step": 20705 }, { "epoch": 0.7296175903168688, "grad_norm": 5.924077987670898, "learning_rate": 3.6191144471277904e-06, "loss": 6.7037, "step": 20706 }, { "epoch": 0.7296528273298273, "grad_norm": 1.888572335243225, "learning_rate": 3.6182357477942134e-06, "loss": 2.051, "step": 20707 }, { "epoch": 0.7296880643427857, "grad_norm": 1.6328932046890259, "learning_rate": 3.617357131583925e-06, "loss": 0.9391, "step": 20708 }, { "epoch": 0.729723301355744, "grad_norm": 3.7644801139831543, "learning_rate": 3.6164785985083706e-06, "loss": 4.9808, "step": 20709 }, { "epoch": 0.7297585383687025, "grad_norm": 3.3843514919281006, "learning_rate": 3.615600148578998e-06, "loss": 2.4732, "step": 20710 }, { "epoch": 0.7297937753816609, "grad_norm": 1.9389649629592896, "learning_rate": 3.61472178180725e-06, "loss": 0.8138, "step": 20711 }, { "epoch": 0.7298290123946193, "grad_norm": 2.272963523864746, "learning_rate": 3.613843498204561e-06, "loss": 2.8319, "step": 20712 }, { "epoch": 0.7298642494075778, "grad_norm": 9.335155487060547, "learning_rate": 3.6129652977823702e-06, "loss": 3.477, "step": 20713 }, { "epoch": 0.7298994864205361, "grad_norm": 3.152590036392212, "learning_rate": 3.612087180552127e-06, "loss": 2.2969, "step": 20714 }, { "epoch": 0.7299347234334945, "grad_norm": 1.0082552433013916, "learning_rate": 3.6112091465252597e-06, "loss": 0.5701, "step": 20715 }, { "epoch": 0.7299699604464529, "grad_norm": 1.4229762554168701, "learning_rate": 3.6103311957132057e-06, "loss": 0.9811, "step": 20716 }, { "epoch": 0.7300051974594114, "grad_norm": 4.911684036254883, "learning_rate": 3.6094533281274022e-06, "loss": 2.4319, "step": 20717 }, { "epoch": 0.7300404344723698, "grad_norm": 5.357486724853516, "learning_rate": 3.6085755437792834e-06, "loss": 2.6596, "step": 20718 }, { "epoch": 0.7300756714853281, "grad_norm": 3.531599283218384, "learning_rate": 3.607697842680282e-06, "loss": 3.1956, "step": 20719 }, { "epoch": 0.7301109084982866, "grad_norm": 1.0938084125518799, "learning_rate": 3.6068202248418305e-06, "loss": 0.7964, "step": 20720 }, { "epoch": 0.730146145511245, "grad_norm": 3.865959405899048, "learning_rate": 3.6059426902753604e-06, "loss": 3.0767, "step": 20721 }, { "epoch": 0.7301813825242034, "grad_norm": 0.7282947897911072, "learning_rate": 3.6050652389923003e-06, "loss": 0.8046, "step": 20722 }, { "epoch": 0.7302166195371619, "grad_norm": 1.9913556575775146, "learning_rate": 3.6041878710040834e-06, "loss": 0.968, "step": 20723 }, { "epoch": 0.7302518565501203, "grad_norm": 2.8035049438476562, "learning_rate": 3.6033105863221287e-06, "loss": 2.4652, "step": 20724 }, { "epoch": 0.7302870935630786, "grad_norm": 9.567638397216797, "learning_rate": 3.6024333849578706e-06, "loss": 5.2213, "step": 20725 }, { "epoch": 0.730322330576037, "grad_norm": 3.5863378047943115, "learning_rate": 3.6015562669227364e-06, "loss": 2.8267, "step": 20726 }, { "epoch": 0.7303575675889955, "grad_norm": 1.3963028192520142, "learning_rate": 3.600679232228144e-06, "loss": 0.7544, "step": 20727 }, { "epoch": 0.7303928046019539, "grad_norm": 3.823556423187256, "learning_rate": 3.599802280885515e-06, "loss": 3.0739, "step": 20728 }, { "epoch": 0.7304280416149123, "grad_norm": 4.399441719055176, "learning_rate": 3.598925412906281e-06, "loss": 4.7812, "step": 20729 }, { "epoch": 0.7304632786278707, "grad_norm": 10.210453033447266, "learning_rate": 3.5980486283018613e-06, "loss": 2.9269, "step": 20730 }, { "epoch": 0.7304985156408291, "grad_norm": 3.385981559753418, "learning_rate": 3.597171927083671e-06, "loss": 2.9273, "step": 20731 }, { "epoch": 0.7305337526537875, "grad_norm": 1.5858180522918701, "learning_rate": 3.596295309263129e-06, "loss": 0.6808, "step": 20732 }, { "epoch": 0.730568989666746, "grad_norm": 1.0689277648925781, "learning_rate": 3.5954187748516624e-06, "loss": 1.04, "step": 20733 }, { "epoch": 0.7306042266797044, "grad_norm": 5.578851222991943, "learning_rate": 3.594542323860679e-06, "loss": 5.2589, "step": 20734 }, { "epoch": 0.7306394636926627, "grad_norm": 4.044247627258301, "learning_rate": 3.5936659563015973e-06, "loss": 2.9123, "step": 20735 }, { "epoch": 0.7306747007056212, "grad_norm": 5.938787460327148, "learning_rate": 3.5927896721858335e-06, "loss": 4.978, "step": 20736 }, { "epoch": 0.7307099377185796, "grad_norm": 6.179537296295166, "learning_rate": 3.5919134715247993e-06, "loss": 5.4054, "step": 20737 }, { "epoch": 0.730745174731538, "grad_norm": 1.1897319555282593, "learning_rate": 3.5910373543299094e-06, "loss": 0.7419, "step": 20738 }, { "epoch": 0.7307804117444964, "grad_norm": 4.243498802185059, "learning_rate": 3.5901613206125728e-06, "loss": 2.6054, "step": 20739 }, { "epoch": 0.7308156487574549, "grad_norm": 7.637845993041992, "learning_rate": 3.5892853703842025e-06, "loss": 3.7094, "step": 20740 }, { "epoch": 0.7308508857704132, "grad_norm": 5.777255535125732, "learning_rate": 3.588409503656207e-06, "loss": 5.6205, "step": 20741 }, { "epoch": 0.7308861227833716, "grad_norm": 4.922970771789551, "learning_rate": 3.5875337204399975e-06, "loss": 3.161, "step": 20742 }, { "epoch": 0.7309213597963301, "grad_norm": 1.6426241397857666, "learning_rate": 3.5866580207469713e-06, "loss": 0.7171, "step": 20743 }, { "epoch": 0.7309565968092885, "grad_norm": 4.833972930908203, "learning_rate": 3.585782404588545e-06, "loss": 4.694, "step": 20744 }, { "epoch": 0.7309918338222469, "grad_norm": 3.876532554626465, "learning_rate": 3.5849068719761235e-06, "loss": 2.8556, "step": 20745 }, { "epoch": 0.7310270708352054, "grad_norm": 7.881927013397217, "learning_rate": 3.5840314229211036e-06, "loss": 2.835, "step": 20746 }, { "epoch": 0.7310623078481637, "grad_norm": 1.3704102039337158, "learning_rate": 3.583156057434888e-06, "loss": 0.8284, "step": 20747 }, { "epoch": 0.7310975448611221, "grad_norm": 0.9128052592277527, "learning_rate": 3.582280775528886e-06, "loss": 0.757, "step": 20748 }, { "epoch": 0.7311327818740805, "grad_norm": 4.95941686630249, "learning_rate": 3.5814055772144975e-06, "loss": 3.4955, "step": 20749 }, { "epoch": 0.731168018887039, "grad_norm": 1.0028656721115112, "learning_rate": 3.5805304625031147e-06, "loss": 0.823, "step": 20750 }, { "epoch": 0.7312032558999974, "grad_norm": 5.034238338470459, "learning_rate": 3.5796554314061383e-06, "loss": 2.6835, "step": 20751 }, { "epoch": 0.7312384929129557, "grad_norm": 0.8577761650085449, "learning_rate": 3.578780483934974e-06, "loss": 0.9201, "step": 20752 }, { "epoch": 0.7312737299259142, "grad_norm": 9.130331039428711, "learning_rate": 3.5779056201010075e-06, "loss": 3.4025, "step": 20753 }, { "epoch": 0.7313089669388726, "grad_norm": 3.0690948963165283, "learning_rate": 3.5770308399156396e-06, "loss": 3.3528, "step": 20754 }, { "epoch": 0.731344203951831, "grad_norm": 3.924740791320801, "learning_rate": 3.5761561433902624e-06, "loss": 3.2888, "step": 20755 }, { "epoch": 0.7313794409647895, "grad_norm": 3.214116096496582, "learning_rate": 3.5752815305362697e-06, "loss": 2.847, "step": 20756 }, { "epoch": 0.7314146779777478, "grad_norm": 5.385311126708984, "learning_rate": 3.5744070013650546e-06, "loss": 2.8132, "step": 20757 }, { "epoch": 0.7314499149907062, "grad_norm": 2.499885320663452, "learning_rate": 3.573532555888005e-06, "loss": 1.0605, "step": 20758 }, { "epoch": 0.7314851520036646, "grad_norm": 3.7307863235473633, "learning_rate": 3.572658194116514e-06, "loss": 2.8347, "step": 20759 }, { "epoch": 0.7315203890166231, "grad_norm": 1.1950807571411133, "learning_rate": 3.5717839160619684e-06, "loss": 0.8867, "step": 20760 }, { "epoch": 0.7315556260295815, "grad_norm": 4.179291725158691, "learning_rate": 3.570909721735759e-06, "loss": 4.9895, "step": 20761 }, { "epoch": 0.7315908630425398, "grad_norm": 3.3684980869293213, "learning_rate": 3.5700356111492627e-06, "loss": 2.6217, "step": 20762 }, { "epoch": 0.7316261000554983, "grad_norm": 1.1692606210708618, "learning_rate": 3.569161584313876e-06, "loss": 0.6808, "step": 20763 }, { "epoch": 0.7316613370684567, "grad_norm": 1.3527239561080933, "learning_rate": 3.5682876412409772e-06, "loss": 0.9267, "step": 20764 }, { "epoch": 0.7316965740814151, "grad_norm": 3.6474616527557373, "learning_rate": 3.5674137819419564e-06, "loss": 2.8931, "step": 20765 }, { "epoch": 0.7317318110943736, "grad_norm": 1.0164496898651123, "learning_rate": 3.566540006428182e-06, "loss": 0.9839, "step": 20766 }, { "epoch": 0.731767048107332, "grad_norm": 1.2802149057388306, "learning_rate": 3.565666314711048e-06, "loss": 0.937, "step": 20767 }, { "epoch": 0.7318022851202903, "grad_norm": 1.1486104726791382, "learning_rate": 3.564792706801933e-06, "loss": 1.0293, "step": 20768 }, { "epoch": 0.7318375221332488, "grad_norm": 8.844080924987793, "learning_rate": 3.563919182712209e-06, "loss": 4.6343, "step": 20769 }, { "epoch": 0.7318727591462072, "grad_norm": 3.325502395629883, "learning_rate": 3.563045742453254e-06, "loss": 3.1506, "step": 20770 }, { "epoch": 0.7319079961591656, "grad_norm": 1.4775545597076416, "learning_rate": 3.5621723860364554e-06, "loss": 0.8388, "step": 20771 }, { "epoch": 0.731943233172124, "grad_norm": 3.4173262119293213, "learning_rate": 3.5612991134731778e-06, "loss": 2.6642, "step": 20772 }, { "epoch": 0.7319784701850824, "grad_norm": 3.9171743392944336, "learning_rate": 3.5604259247747995e-06, "loss": 5.2194, "step": 20773 }, { "epoch": 0.7320137071980408, "grad_norm": 0.7339069247245789, "learning_rate": 3.559552819952694e-06, "loss": 0.6454, "step": 20774 }, { "epoch": 0.7320489442109992, "grad_norm": 9.32046127319336, "learning_rate": 3.5586797990182332e-06, "loss": 5.3029, "step": 20775 }, { "epoch": 0.7320841812239577, "grad_norm": 1.2748385667800903, "learning_rate": 3.5578068619827876e-06, "loss": 0.7462, "step": 20776 }, { "epoch": 0.7321194182369161, "grad_norm": 4.361721515655518, "learning_rate": 3.5569340088577296e-06, "loss": 2.8552, "step": 20777 }, { "epoch": 0.7321546552498744, "grad_norm": 3.9770379066467285, "learning_rate": 3.556061239654426e-06, "loss": 3.1074, "step": 20778 }, { "epoch": 0.7321898922628329, "grad_norm": 2.636152505874634, "learning_rate": 3.555188554384247e-06, "loss": 1.1684, "step": 20779 }, { "epoch": 0.7322251292757913, "grad_norm": 1.0519208908081055, "learning_rate": 3.554315953058559e-06, "loss": 0.6791, "step": 20780 }, { "epoch": 0.7322603662887497, "grad_norm": 3.9310946464538574, "learning_rate": 3.5534434356887215e-06, "loss": 2.8415, "step": 20781 }, { "epoch": 0.7322956033017081, "grad_norm": 3.659853458404541, "learning_rate": 3.5525710022861084e-06, "loss": 4.5216, "step": 20782 }, { "epoch": 0.7323308403146666, "grad_norm": 1.4372318983078003, "learning_rate": 3.551698652862078e-06, "loss": 0.6335, "step": 20783 }, { "epoch": 0.7323660773276249, "grad_norm": 5.551060676574707, "learning_rate": 3.550826387427998e-06, "loss": 3.1079, "step": 20784 }, { "epoch": 0.7324013143405833, "grad_norm": 3.1276943683624268, "learning_rate": 3.5499542059952198e-06, "loss": 2.8949, "step": 20785 }, { "epoch": 0.7324365513535418, "grad_norm": 1.1942507028579712, "learning_rate": 3.5490821085751116e-06, "loss": 0.9815, "step": 20786 }, { "epoch": 0.7324717883665002, "grad_norm": 0.9469272494316101, "learning_rate": 3.5482100951790344e-06, "loss": 0.8509, "step": 20787 }, { "epoch": 0.7325070253794586, "grad_norm": 1.5150879621505737, "learning_rate": 3.547338165818339e-06, "loss": 0.9191, "step": 20788 }, { "epoch": 0.732542262392417, "grad_norm": 5.176866054534912, "learning_rate": 3.5464663205043836e-06, "loss": 3.1884, "step": 20789 }, { "epoch": 0.7325774994053754, "grad_norm": 2.9590625762939453, "learning_rate": 3.545594559248533e-06, "loss": 2.9567, "step": 20790 }, { "epoch": 0.7326127364183338, "grad_norm": 0.9795133471488953, "learning_rate": 3.5447228820621315e-06, "loss": 0.8546, "step": 20791 }, { "epoch": 0.7326479734312922, "grad_norm": 4.680698871612549, "learning_rate": 3.543851288956537e-06, "loss": 3.5423, "step": 20792 }, { "epoch": 0.7326832104442507, "grad_norm": 4.447275638580322, "learning_rate": 3.542979779943101e-06, "loss": 5.0224, "step": 20793 }, { "epoch": 0.732718447457209, "grad_norm": 3.6207451820373535, "learning_rate": 3.542108355033177e-06, "loss": 3.1167, "step": 20794 }, { "epoch": 0.7327536844701674, "grad_norm": 17.884504318237305, "learning_rate": 3.5412370142381135e-06, "loss": 3.109, "step": 20795 }, { "epoch": 0.7327889214831259, "grad_norm": 6.161310195922852, "learning_rate": 3.5403657575692606e-06, "loss": 4.7996, "step": 20796 }, { "epoch": 0.7328241584960843, "grad_norm": 1.2148383855819702, "learning_rate": 3.539494585037967e-06, "loss": 0.8279, "step": 20797 }, { "epoch": 0.7328593955090427, "grad_norm": 2.9226486682891846, "learning_rate": 3.538623496655579e-06, "loss": 2.0793, "step": 20798 }, { "epoch": 0.7328946325220012, "grad_norm": 4.368692398071289, "learning_rate": 3.5377524924334474e-06, "loss": 2.4634, "step": 20799 }, { "epoch": 0.7329298695349595, "grad_norm": 4.039892673492432, "learning_rate": 3.5368815723829042e-06, "loss": 4.7912, "step": 20800 }, { "epoch": 0.7329651065479179, "grad_norm": 1.2551885843276978, "learning_rate": 3.536010736515306e-06, "loss": 0.7871, "step": 20801 }, { "epoch": 0.7330003435608764, "grad_norm": 3.624882459640503, "learning_rate": 3.535139984841992e-06, "loss": 3.4184, "step": 20802 }, { "epoch": 0.7330355805738348, "grad_norm": 2.904637336730957, "learning_rate": 3.5342693173743057e-06, "loss": 2.7328, "step": 20803 }, { "epoch": 0.7330708175867932, "grad_norm": 1.0862860679626465, "learning_rate": 3.5333987341235785e-06, "loss": 0.8021, "step": 20804 }, { "epoch": 0.7331060545997515, "grad_norm": 3.9781715869903564, "learning_rate": 3.5325282351011604e-06, "loss": 2.6287, "step": 20805 }, { "epoch": 0.73314129161271, "grad_norm": 2.468093156814575, "learning_rate": 3.5316578203183894e-06, "loss": 2.5519, "step": 20806 }, { "epoch": 0.7331765286256684, "grad_norm": 2.916788339614868, "learning_rate": 3.530787489786596e-06, "loss": 2.9558, "step": 20807 }, { "epoch": 0.7332117656386268, "grad_norm": 1.2935179471969604, "learning_rate": 3.5299172435171147e-06, "loss": 0.7979, "step": 20808 }, { "epoch": 0.7332470026515853, "grad_norm": 2.3853514194488525, "learning_rate": 3.529047081521293e-06, "loss": 2.4441, "step": 20809 }, { "epoch": 0.7332822396645436, "grad_norm": 1.4611707925796509, "learning_rate": 3.5281770038104547e-06, "loss": 0.8595, "step": 20810 }, { "epoch": 0.733317476677502, "grad_norm": 4.32355260848999, "learning_rate": 3.527307010395934e-06, "loss": 2.9193, "step": 20811 }, { "epoch": 0.7333527136904605, "grad_norm": 3.2958271503448486, "learning_rate": 3.5264371012890653e-06, "loss": 2.282, "step": 20812 }, { "epoch": 0.7333879507034189, "grad_norm": 3.476579427719116, "learning_rate": 3.525567276501177e-06, "loss": 3.1175, "step": 20813 }, { "epoch": 0.7334231877163773, "grad_norm": 1.2731258869171143, "learning_rate": 3.5246975360435996e-06, "loss": 0.6867, "step": 20814 }, { "epoch": 0.7334584247293356, "grad_norm": 1.1648736000061035, "learning_rate": 3.5238278799276625e-06, "loss": 0.7972, "step": 20815 }, { "epoch": 0.7334936617422941, "grad_norm": 1.1761983633041382, "learning_rate": 3.522958308164691e-06, "loss": 0.6998, "step": 20816 }, { "epoch": 0.7335288987552525, "grad_norm": 4.963457107543945, "learning_rate": 3.522088820766014e-06, "loss": 6.2176, "step": 20817 }, { "epoch": 0.7335641357682109, "grad_norm": 6.544497489929199, "learning_rate": 3.521219417742957e-06, "loss": 5.046, "step": 20818 }, { "epoch": 0.7335993727811694, "grad_norm": 3.800147533416748, "learning_rate": 3.5203500991068373e-06, "loss": 2.8411, "step": 20819 }, { "epoch": 0.7336346097941278, "grad_norm": 3.7923192977905273, "learning_rate": 3.519480864868985e-06, "loss": 4.301, "step": 20820 }, { "epoch": 0.7336698468070861, "grad_norm": 5.3363471031188965, "learning_rate": 3.5186117150407206e-06, "loss": 5.0135, "step": 20821 }, { "epoch": 0.7337050838200446, "grad_norm": 0.9582157135009766, "learning_rate": 3.517742649633368e-06, "loss": 0.8549, "step": 20822 }, { "epoch": 0.733740320833003, "grad_norm": 4.660693645477295, "learning_rate": 3.516873668658235e-06, "loss": 4.8245, "step": 20823 }, { "epoch": 0.7337755578459614, "grad_norm": 2.180741548538208, "learning_rate": 3.516004772126652e-06, "loss": 0.9769, "step": 20824 }, { "epoch": 0.7338107948589198, "grad_norm": 4.097277641296387, "learning_rate": 3.515135960049937e-06, "loss": 4.2001, "step": 20825 }, { "epoch": 0.7338460318718782, "grad_norm": 0.8740531206130981, "learning_rate": 3.514267232439399e-06, "loss": 0.8349, "step": 20826 }, { "epoch": 0.7338812688848366, "grad_norm": 1.3142811059951782, "learning_rate": 3.513398589306355e-06, "loss": 0.8568, "step": 20827 }, { "epoch": 0.733916505897795, "grad_norm": 1.1269341707229614, "learning_rate": 3.512530030662121e-06, "loss": 0.7512, "step": 20828 }, { "epoch": 0.7339517429107535, "grad_norm": 1.1865099668502808, "learning_rate": 3.511661556518011e-06, "loss": 0.7583, "step": 20829 }, { "epoch": 0.7339869799237119, "grad_norm": 9.180916786193848, "learning_rate": 3.5107931668853344e-06, "loss": 5.0459, "step": 20830 }, { "epoch": 0.7340222169366702, "grad_norm": 3.0965728759765625, "learning_rate": 3.5099248617754044e-06, "loss": 2.8444, "step": 20831 }, { "epoch": 0.7340574539496287, "grad_norm": 3.250664472579956, "learning_rate": 3.5090566411995285e-06, "loss": 2.7558, "step": 20832 }, { "epoch": 0.7340926909625871, "grad_norm": 1.3779927492141724, "learning_rate": 3.508188505169018e-06, "loss": 1.1122, "step": 20833 }, { "epoch": 0.7341279279755455, "grad_norm": 3.2953600883483887, "learning_rate": 3.507320453695179e-06, "loss": 2.6199, "step": 20834 }, { "epoch": 0.7341631649885039, "grad_norm": 2.872666835784912, "learning_rate": 3.506452486789317e-06, "loss": 2.5162, "step": 20835 }, { "epoch": 0.7341984020014624, "grad_norm": 5.176486492156982, "learning_rate": 3.5055846044627396e-06, "loss": 2.8613, "step": 20836 }, { "epoch": 0.7342336390144207, "grad_norm": 4.500028610229492, "learning_rate": 3.504716806726749e-06, "loss": 3.1217, "step": 20837 }, { "epoch": 0.7342688760273791, "grad_norm": 1.3716377019882202, "learning_rate": 3.503849093592654e-06, "loss": 0.8231, "step": 20838 }, { "epoch": 0.7343041130403376, "grad_norm": 1.3798612356185913, "learning_rate": 3.5029814650717443e-06, "loss": 0.7767, "step": 20839 }, { "epoch": 0.734339350053296, "grad_norm": 6.029638290405273, "learning_rate": 3.502113921175332e-06, "loss": 6.581, "step": 20840 }, { "epoch": 0.7343745870662544, "grad_norm": 2.3109936714172363, "learning_rate": 3.5012464619147167e-06, "loss": 0.8825, "step": 20841 }, { "epoch": 0.7344098240792128, "grad_norm": 4.061249732971191, "learning_rate": 3.5003790873011922e-06, "loss": 2.9951, "step": 20842 }, { "epoch": 0.7344450610921712, "grad_norm": 5.014894008636475, "learning_rate": 3.4995117973460537e-06, "loss": 3.2532, "step": 20843 }, { "epoch": 0.7344802981051296, "grad_norm": 3.5494818687438965, "learning_rate": 3.4986445920606084e-06, "loss": 2.9473, "step": 20844 }, { "epoch": 0.7345155351180881, "grad_norm": 6.167048931121826, "learning_rate": 3.497777471456142e-06, "loss": 5.0656, "step": 20845 }, { "epoch": 0.7345507721310465, "grad_norm": 6.633947849273682, "learning_rate": 3.4969104355439533e-06, "loss": 4.9717, "step": 20846 }, { "epoch": 0.7345860091440048, "grad_norm": 1.0672255754470825, "learning_rate": 3.496043484335333e-06, "loss": 0.7969, "step": 20847 }, { "epoch": 0.7346212461569632, "grad_norm": 3.9006364345550537, "learning_rate": 3.4951766178415757e-06, "loss": 2.9197, "step": 20848 }, { "epoch": 0.7346564831699217, "grad_norm": 1.8012129068374634, "learning_rate": 3.4943098360739725e-06, "loss": 2.3406, "step": 20849 }, { "epoch": 0.7346917201828801, "grad_norm": 3.856281042098999, "learning_rate": 3.4934431390438104e-06, "loss": 2.6868, "step": 20850 }, { "epoch": 0.7347269571958385, "grad_norm": 4.010284900665283, "learning_rate": 3.4925765267623813e-06, "loss": 2.7742, "step": 20851 }, { "epoch": 0.734762194208797, "grad_norm": 7.763437747955322, "learning_rate": 3.491709999240972e-06, "loss": 7.1603, "step": 20852 }, { "epoch": 0.7347974312217553, "grad_norm": 3.2045748233795166, "learning_rate": 3.4908435564908727e-06, "loss": 2.9823, "step": 20853 }, { "epoch": 0.7348326682347137, "grad_norm": 2.860013484954834, "learning_rate": 3.4899771985233576e-06, "loss": 3.265, "step": 20854 }, { "epoch": 0.7348679052476722, "grad_norm": 1.274793267250061, "learning_rate": 3.4891109253497224e-06, "loss": 0.9145, "step": 20855 }, { "epoch": 0.7349031422606306, "grad_norm": 1.342689037322998, "learning_rate": 3.4882447369812466e-06, "loss": 0.9477, "step": 20856 }, { "epoch": 0.734938379273589, "grad_norm": 1.6856954097747803, "learning_rate": 3.487378633429216e-06, "loss": 0.838, "step": 20857 }, { "epoch": 0.7349736162865473, "grad_norm": 4.165821552276611, "learning_rate": 3.4865126147049023e-06, "loss": 2.8332, "step": 20858 }, { "epoch": 0.7350088532995058, "grad_norm": 3.8284292221069336, "learning_rate": 3.4856466808195934e-06, "loss": 2.6089, "step": 20859 }, { "epoch": 0.7350440903124642, "grad_norm": 6.001513481140137, "learning_rate": 3.484780831784571e-06, "loss": 3.5406, "step": 20860 }, { "epoch": 0.7350793273254226, "grad_norm": 5.609755992889404, "learning_rate": 3.4839150676111045e-06, "loss": 3.3708, "step": 20861 }, { "epoch": 0.7351145643383811, "grad_norm": 1.3011456727981567, "learning_rate": 3.4830493883104697e-06, "loss": 1.0156, "step": 20862 }, { "epoch": 0.7351498013513394, "grad_norm": 2.9317522048950195, "learning_rate": 3.482183793893955e-06, "loss": 2.8002, "step": 20863 }, { "epoch": 0.7351850383642978, "grad_norm": 4.851472854614258, "learning_rate": 3.4813182843728223e-06, "loss": 4.532, "step": 20864 }, { "epoch": 0.7352202753772563, "grad_norm": 1.0027930736541748, "learning_rate": 3.4804528597583496e-06, "loss": 0.6569, "step": 20865 }, { "epoch": 0.7352555123902147, "grad_norm": 4.338350296020508, "learning_rate": 3.479587520061809e-06, "loss": 3.0926, "step": 20866 }, { "epoch": 0.7352907494031731, "grad_norm": 1.3317265510559082, "learning_rate": 3.478722265294472e-06, "loss": 0.7049, "step": 20867 }, { "epoch": 0.7353259864161314, "grad_norm": 1.1317116022109985, "learning_rate": 3.477857095467607e-06, "loss": 0.6657, "step": 20868 }, { "epoch": 0.7353612234290899, "grad_norm": 3.601879119873047, "learning_rate": 3.4769920105924847e-06, "loss": 2.9717, "step": 20869 }, { "epoch": 0.7353964604420483, "grad_norm": 1.3485636711120605, "learning_rate": 3.4761270106803724e-06, "loss": 0.9344, "step": 20870 }, { "epoch": 0.7354316974550067, "grad_norm": 0.880186140537262, "learning_rate": 3.4752620957425364e-06, "loss": 0.768, "step": 20871 }, { "epoch": 0.7354669344679652, "grad_norm": 4.723521709442139, "learning_rate": 3.474397265790247e-06, "loss": 3.5388, "step": 20872 }, { "epoch": 0.7355021714809236, "grad_norm": 1.1010938882827759, "learning_rate": 3.473532520834757e-06, "loss": 0.9279, "step": 20873 }, { "epoch": 0.7355374084938819, "grad_norm": 17.985822677612305, "learning_rate": 3.472667860887341e-06, "loss": 2.544, "step": 20874 }, { "epoch": 0.7355726455068404, "grad_norm": 25.65854835510254, "learning_rate": 3.4718032859592575e-06, "loss": 2.9027, "step": 20875 }, { "epoch": 0.7356078825197988, "grad_norm": 2.883211851119995, "learning_rate": 3.47093879606177e-06, "loss": 2.9719, "step": 20876 }, { "epoch": 0.7356431195327572, "grad_norm": 7.03028678894043, "learning_rate": 3.470074391206132e-06, "loss": 2.949, "step": 20877 }, { "epoch": 0.7356783565457157, "grad_norm": 14.389287948608398, "learning_rate": 3.469210071403608e-06, "loss": 4.676, "step": 20878 }, { "epoch": 0.735713593558674, "grad_norm": 0.7603312134742737, "learning_rate": 3.468345836665459e-06, "loss": 0.8054, "step": 20879 }, { "epoch": 0.7357488305716324, "grad_norm": 6.70092248916626, "learning_rate": 3.4674816870029337e-06, "loss": 5.3082, "step": 20880 }, { "epoch": 0.7357840675845908, "grad_norm": 4.488877296447754, "learning_rate": 3.4666176224272885e-06, "loss": 2.885, "step": 20881 }, { "epoch": 0.7358193045975493, "grad_norm": 1.0376019477844238, "learning_rate": 3.4657536429497883e-06, "loss": 0.8516, "step": 20882 }, { "epoch": 0.7358545416105077, "grad_norm": 1.4659099578857422, "learning_rate": 3.4648897485816746e-06, "loss": 0.9227, "step": 20883 }, { "epoch": 0.735889778623466, "grad_norm": 1.1524157524108887, "learning_rate": 3.4640259393342057e-06, "loss": 0.8437, "step": 20884 }, { "epoch": 0.7359250156364245, "grad_norm": 3.182481050491333, "learning_rate": 3.4631622152186306e-06, "loss": 2.8514, "step": 20885 }, { "epoch": 0.7359602526493829, "grad_norm": 2.234447956085205, "learning_rate": 3.4622985762462e-06, "loss": 2.1087, "step": 20886 }, { "epoch": 0.7359954896623413, "grad_norm": 1.7896618843078613, "learning_rate": 3.4614350224281644e-06, "loss": 1.1356, "step": 20887 }, { "epoch": 0.7360307266752998, "grad_norm": 11.23212718963623, "learning_rate": 3.46057155377577e-06, "loss": 4.9712, "step": 20888 }, { "epoch": 0.7360659636882582, "grad_norm": 1.1668109893798828, "learning_rate": 3.4597081703002633e-06, "loss": 1.0266, "step": 20889 }, { "epoch": 0.7361012007012165, "grad_norm": 4.691789627075195, "learning_rate": 3.458844872012892e-06, "loss": 2.7539, "step": 20890 }, { "epoch": 0.7361364377141749, "grad_norm": 4.473639011383057, "learning_rate": 3.4579816589249015e-06, "loss": 4.959, "step": 20891 }, { "epoch": 0.7361716747271334, "grad_norm": 1.0503078699111938, "learning_rate": 3.4571185310475273e-06, "loss": 1.0033, "step": 20892 }, { "epoch": 0.7362069117400918, "grad_norm": 2.044700860977173, "learning_rate": 3.4562554883920206e-06, "loss": 0.8262, "step": 20893 }, { "epoch": 0.7362421487530502, "grad_norm": 5.836465835571289, "learning_rate": 3.45539253096962e-06, "loss": 2.855, "step": 20894 }, { "epoch": 0.7362773857660087, "grad_norm": 0.932793140411377, "learning_rate": 3.4545296587915687e-06, "loss": 0.9657, "step": 20895 }, { "epoch": 0.736312622778967, "grad_norm": 4.148293972015381, "learning_rate": 3.4536668718690967e-06, "loss": 2.9021, "step": 20896 }, { "epoch": 0.7363478597919254, "grad_norm": 0.9852578639984131, "learning_rate": 3.4528041702134496e-06, "loss": 0.9628, "step": 20897 }, { "epoch": 0.7363830968048839, "grad_norm": 1.2615153789520264, "learning_rate": 3.451941553835866e-06, "loss": 0.8924, "step": 20898 }, { "epoch": 0.7364183338178423, "grad_norm": 3.8926315307617188, "learning_rate": 3.451079022747574e-06, "loss": 2.7078, "step": 20899 }, { "epoch": 0.7364535708308007, "grad_norm": 2.9789366722106934, "learning_rate": 3.450216576959808e-06, "loss": 2.8433, "step": 20900 }, { "epoch": 0.736488807843759, "grad_norm": 1.2458407878875732, "learning_rate": 3.4493542164838133e-06, "loss": 0.611, "step": 20901 }, { "epoch": 0.7365240448567175, "grad_norm": 18.564157485961914, "learning_rate": 3.4484919413308117e-06, "loss": 2.7161, "step": 20902 }, { "epoch": 0.7365592818696759, "grad_norm": 1.5535014867782593, "learning_rate": 3.447629751512035e-06, "loss": 0.9512, "step": 20903 }, { "epoch": 0.7365945188826343, "grad_norm": 2.2553281784057617, "learning_rate": 3.446767647038717e-06, "loss": 1.0922, "step": 20904 }, { "epoch": 0.7366297558955928, "grad_norm": 2.943821430206299, "learning_rate": 3.445905627922085e-06, "loss": 2.5874, "step": 20905 }, { "epoch": 0.7366649929085511, "grad_norm": 1.0529032945632935, "learning_rate": 3.4450436941733656e-06, "loss": 0.6461, "step": 20906 }, { "epoch": 0.7367002299215095, "grad_norm": 1.3578619956970215, "learning_rate": 3.4441818458037878e-06, "loss": 0.6878, "step": 20907 }, { "epoch": 0.736735466934468, "grad_norm": 1.0173141956329346, "learning_rate": 3.443320082824577e-06, "loss": 0.7424, "step": 20908 }, { "epoch": 0.7367707039474264, "grad_norm": 6.754522800445557, "learning_rate": 3.442458405246958e-06, "loss": 3.0559, "step": 20909 }, { "epoch": 0.7368059409603848, "grad_norm": 2.7715353965759277, "learning_rate": 3.441596813082152e-06, "loss": 0.8278, "step": 20910 }, { "epoch": 0.7368411779733433, "grad_norm": 2.0568888187408447, "learning_rate": 3.440735306341384e-06, "loss": 3.0759, "step": 20911 }, { "epoch": 0.7368764149863016, "grad_norm": 10.646973609924316, "learning_rate": 3.439873885035874e-06, "loss": 3.584, "step": 20912 }, { "epoch": 0.73691165199926, "grad_norm": 6.756459712982178, "learning_rate": 3.439012549176841e-06, "loss": 5.0029, "step": 20913 }, { "epoch": 0.7369468890122184, "grad_norm": 1.3821604251861572, "learning_rate": 3.4381512987755104e-06, "loss": 0.8709, "step": 20914 }, { "epoch": 0.7369821260251769, "grad_norm": 1.1225470304489136, "learning_rate": 3.4372901338430877e-06, "loss": 0.8163, "step": 20915 }, { "epoch": 0.7370173630381353, "grad_norm": 1.6204557418823242, "learning_rate": 3.4364290543908e-06, "loss": 0.9582, "step": 20916 }, { "epoch": 0.7370526000510936, "grad_norm": 1.8499727249145508, "learning_rate": 3.4355680604298637e-06, "loss": 0.8228, "step": 20917 }, { "epoch": 0.7370878370640521, "grad_norm": 2.9955270290374756, "learning_rate": 3.434707151971486e-06, "loss": 3.2241, "step": 20918 }, { "epoch": 0.7371230740770105, "grad_norm": 4.900275230407715, "learning_rate": 3.4338463290268796e-06, "loss": 2.4412, "step": 20919 }, { "epoch": 0.7371583110899689, "grad_norm": 1.0818803310394287, "learning_rate": 3.4329855916072696e-06, "loss": 0.8562, "step": 20920 }, { "epoch": 0.7371935481029274, "grad_norm": 5.325018405914307, "learning_rate": 3.432124939723854e-06, "loss": 2.8958, "step": 20921 }, { "epoch": 0.7372287851158857, "grad_norm": 4.067811965942383, "learning_rate": 3.431264373387849e-06, "loss": 3.0451, "step": 20922 }, { "epoch": 0.7372640221288441, "grad_norm": 2.1117944717407227, "learning_rate": 3.430403892610461e-06, "loss": 1.0277, "step": 20923 }, { "epoch": 0.7372992591418025, "grad_norm": 1.1869820356369019, "learning_rate": 3.4295434974028998e-06, "loss": 0.7678, "step": 20924 }, { "epoch": 0.737334496154761, "grad_norm": 1.274458408355713, "learning_rate": 3.4286831877763718e-06, "loss": 0.9322, "step": 20925 }, { "epoch": 0.7373697331677194, "grad_norm": 3.9008400440216064, "learning_rate": 3.4278229637420813e-06, "loss": 2.5646, "step": 20926 }, { "epoch": 0.7374049701806777, "grad_norm": 1.0711323022842407, "learning_rate": 3.426962825311235e-06, "loss": 0.7406, "step": 20927 }, { "epoch": 0.7374402071936362, "grad_norm": 1.261022925376892, "learning_rate": 3.4261027724950344e-06, "loss": 0.7278, "step": 20928 }, { "epoch": 0.7374754442065946, "grad_norm": 2.5101494789123535, "learning_rate": 3.425242805304683e-06, "loss": 0.9889, "step": 20929 }, { "epoch": 0.737510681219553, "grad_norm": 1.6814470291137695, "learning_rate": 3.4243829237513815e-06, "loss": 0.801, "step": 20930 }, { "epoch": 0.7375459182325115, "grad_norm": 3.958616256713867, "learning_rate": 3.42352312784633e-06, "loss": 5.1832, "step": 20931 }, { "epoch": 0.7375811552454699, "grad_norm": 1.1589951515197754, "learning_rate": 3.422663417600727e-06, "loss": 1.1837, "step": 20932 }, { "epoch": 0.7376163922584282, "grad_norm": 2.5929529666900635, "learning_rate": 3.421803793025774e-06, "loss": 0.8443, "step": 20933 }, { "epoch": 0.7376516292713866, "grad_norm": 1.0885117053985596, "learning_rate": 3.420944254132659e-06, "loss": 1.0567, "step": 20934 }, { "epoch": 0.7376868662843451, "grad_norm": 2.5260298252105713, "learning_rate": 3.420084800932586e-06, "loss": 2.8908, "step": 20935 }, { "epoch": 0.7377221032973035, "grad_norm": 1.8521806001663208, "learning_rate": 3.4192254334367493e-06, "loss": 0.6066, "step": 20936 }, { "epoch": 0.7377573403102619, "grad_norm": 6.912240505218506, "learning_rate": 3.418366151656337e-06, "loss": 5.1563, "step": 20937 }, { "epoch": 0.7377925773232203, "grad_norm": 0.865305483341217, "learning_rate": 3.417506955602541e-06, "loss": 0.7172, "step": 20938 }, { "epoch": 0.7378278143361787, "grad_norm": 3.780341386795044, "learning_rate": 3.4166478452865604e-06, "loss": 2.7207, "step": 20939 }, { "epoch": 0.7378630513491371, "grad_norm": 2.2426953315734863, "learning_rate": 3.415788820719578e-06, "loss": 0.6638, "step": 20940 }, { "epoch": 0.7378982883620956, "grad_norm": 1.792729377746582, "learning_rate": 3.414929881912785e-06, "loss": 1.168, "step": 20941 }, { "epoch": 0.737933525375054, "grad_norm": 0.8629108667373657, "learning_rate": 3.414071028877368e-06, "loss": 0.7911, "step": 20942 }, { "epoch": 0.7379687623880123, "grad_norm": 7.090115547180176, "learning_rate": 3.4132122616245166e-06, "loss": 2.7956, "step": 20943 }, { "epoch": 0.7380039994009708, "grad_norm": 1.6157724857330322, "learning_rate": 3.4123535801654127e-06, "loss": 0.7312, "step": 20944 }, { "epoch": 0.7380392364139292, "grad_norm": 2.5094892978668213, "learning_rate": 3.4114949845112444e-06, "loss": 0.9892, "step": 20945 }, { "epoch": 0.7380744734268876, "grad_norm": 1.5351181030273438, "learning_rate": 3.4106364746731913e-06, "loss": 0.8526, "step": 20946 }, { "epoch": 0.738109710439846, "grad_norm": 1.7682058811187744, "learning_rate": 3.409778050662439e-06, "loss": 0.7192, "step": 20947 }, { "epoch": 0.7381449474528045, "grad_norm": 1.248922348022461, "learning_rate": 3.408919712490166e-06, "loss": 0.8454, "step": 20948 }, { "epoch": 0.7381801844657628, "grad_norm": 5.401991367340088, "learning_rate": 3.4080614601675533e-06, "loss": 6.9932, "step": 20949 }, { "epoch": 0.7382154214787212, "grad_norm": 6.263096809387207, "learning_rate": 3.407203293705781e-06, "loss": 4.237, "step": 20950 }, { "epoch": 0.7382506584916797, "grad_norm": 1.0973976850509644, "learning_rate": 3.4063452131160248e-06, "loss": 0.9969, "step": 20951 }, { "epoch": 0.7382858955046381, "grad_norm": 0.8005090951919556, "learning_rate": 3.4054872184094645e-06, "loss": 0.9346, "step": 20952 }, { "epoch": 0.7383211325175965, "grad_norm": 3.9408211708068848, "learning_rate": 3.4046293095972682e-06, "loss": 3.2092, "step": 20953 }, { "epoch": 0.738356369530555, "grad_norm": 1.4055323600769043, "learning_rate": 3.403771486690617e-06, "loss": 0.8721, "step": 20954 }, { "epoch": 0.7383916065435133, "grad_norm": 3.532926082611084, "learning_rate": 3.402913749700687e-06, "loss": 3.0816, "step": 20955 }, { "epoch": 0.7384268435564717, "grad_norm": 1.2418007850646973, "learning_rate": 3.402056098638642e-06, "loss": 0.8069, "step": 20956 }, { "epoch": 0.7384620805694301, "grad_norm": 7.783145427703857, "learning_rate": 3.4011985335156527e-06, "loss": 4.9777, "step": 20957 }, { "epoch": 0.7384973175823886, "grad_norm": 1.5949208736419678, "learning_rate": 3.4003410543429005e-06, "loss": 0.788, "step": 20958 }, { "epoch": 0.738532554595347, "grad_norm": 1.6366093158721924, "learning_rate": 3.399483661131544e-06, "loss": 0.9031, "step": 20959 }, { "epoch": 0.7385677916083053, "grad_norm": 1.467966079711914, "learning_rate": 3.398626353892752e-06, "loss": 0.9759, "step": 20960 }, { "epoch": 0.7386030286212638, "grad_norm": 27.68771743774414, "learning_rate": 3.397769132637694e-06, "loss": 5.2325, "step": 20961 }, { "epoch": 0.7386382656342222, "grad_norm": 2.833765745162964, "learning_rate": 3.3969119973775334e-06, "loss": 2.927, "step": 20962 }, { "epoch": 0.7386735026471806, "grad_norm": 3.097270965576172, "learning_rate": 3.396054948123435e-06, "loss": 2.7521, "step": 20963 }, { "epoch": 0.7387087396601391, "grad_norm": 1.2905100584030151, "learning_rate": 3.3951979848865623e-06, "loss": 0.7959, "step": 20964 }, { "epoch": 0.7387439766730974, "grad_norm": 1.1900612115859985, "learning_rate": 3.3943411076780774e-06, "loss": 1.1859, "step": 20965 }, { "epoch": 0.7387792136860558, "grad_norm": 3.1016387939453125, "learning_rate": 3.3934843165091414e-06, "loss": 2.8833, "step": 20966 }, { "epoch": 0.7388144506990142, "grad_norm": 4.379891872406006, "learning_rate": 3.3926276113909127e-06, "loss": 2.46, "step": 20967 }, { "epoch": 0.7388496877119727, "grad_norm": 3.5594496726989746, "learning_rate": 3.391770992334551e-06, "loss": 2.499, "step": 20968 }, { "epoch": 0.7388849247249311, "grad_norm": 3.3451457023620605, "learning_rate": 3.3909144593512146e-06, "loss": 1.0991, "step": 20969 }, { "epoch": 0.7389201617378894, "grad_norm": 1.1910444498062134, "learning_rate": 3.3900580124520586e-06, "loss": 1.0642, "step": 20970 }, { "epoch": 0.7389553987508479, "grad_norm": 1.484124779701233, "learning_rate": 3.389201651648243e-06, "loss": 0.9217, "step": 20971 }, { "epoch": 0.7389906357638063, "grad_norm": 3.599799394607544, "learning_rate": 3.388345376950911e-06, "loss": 3.1666, "step": 20972 }, { "epoch": 0.7390258727767647, "grad_norm": 5.027444362640381, "learning_rate": 3.3874891883712257e-06, "loss": 4.2721, "step": 20973 }, { "epoch": 0.7390611097897232, "grad_norm": 2.6116793155670166, "learning_rate": 3.3866330859203402e-06, "loss": 1.0268, "step": 20974 }, { "epoch": 0.7390963468026815, "grad_norm": 14.893739700317383, "learning_rate": 3.3857770696093974e-06, "loss": 0.7561, "step": 20975 }, { "epoch": 0.7391315838156399, "grad_norm": 5.3410491943359375, "learning_rate": 3.3849211394495463e-06, "loss": 2.5353, "step": 20976 }, { "epoch": 0.7391668208285984, "grad_norm": 3.3156187534332275, "learning_rate": 3.3840652954519483e-06, "loss": 2.6133, "step": 20977 }, { "epoch": 0.7392020578415568, "grad_norm": 3.523195266723633, "learning_rate": 3.3832095376277385e-06, "loss": 3.09, "step": 20978 }, { "epoch": 0.7392372948545152, "grad_norm": 1.4004210233688354, "learning_rate": 3.382353865988067e-06, "loss": 0.7208, "step": 20979 }, { "epoch": 0.7392725318674735, "grad_norm": 1.2261581420898438, "learning_rate": 3.38149828054408e-06, "loss": 1.1811, "step": 20980 }, { "epoch": 0.739307768880432, "grad_norm": 6.383528232574463, "learning_rate": 3.380642781306921e-06, "loss": 2.8997, "step": 20981 }, { "epoch": 0.7393430058933904, "grad_norm": 0.9203280806541443, "learning_rate": 3.3797873682877326e-06, "loss": 0.8562, "step": 20982 }, { "epoch": 0.7393782429063488, "grad_norm": 6.066143989562988, "learning_rate": 3.378932041497657e-06, "loss": 4.7146, "step": 20983 }, { "epoch": 0.7394134799193073, "grad_norm": 1.136674165725708, "learning_rate": 3.378076800947835e-06, "loss": 0.819, "step": 20984 }, { "epoch": 0.7394487169322657, "grad_norm": 2.1695759296417236, "learning_rate": 3.3772216466494058e-06, "loss": 2.1931, "step": 20985 }, { "epoch": 0.739483953945224, "grad_norm": 1.4025782346725464, "learning_rate": 3.376366578613508e-06, "loss": 1.0716, "step": 20986 }, { "epoch": 0.7395191909581825, "grad_norm": 0.9921898245811462, "learning_rate": 3.3755115968512807e-06, "loss": 0.9326, "step": 20987 }, { "epoch": 0.7395544279711409, "grad_norm": 1.3983291387557983, "learning_rate": 3.3746567013738575e-06, "loss": 0.7932, "step": 20988 }, { "epoch": 0.7395896649840993, "grad_norm": 1.4914956092834473, "learning_rate": 3.3738018921923755e-06, "loss": 0.8024, "step": 20989 }, { "epoch": 0.7396249019970577, "grad_norm": 3.0720722675323486, "learning_rate": 3.3729471693179705e-06, "loss": 2.6936, "step": 20990 }, { "epoch": 0.7396601390100161, "grad_norm": 5.482826232910156, "learning_rate": 3.372092532761766e-06, "loss": 2.898, "step": 20991 }, { "epoch": 0.7396953760229745, "grad_norm": 4.738382816314697, "learning_rate": 3.3712379825349038e-06, "loss": 3.5717, "step": 20992 }, { "epoch": 0.7397306130359329, "grad_norm": 1.3037880659103394, "learning_rate": 3.3703835186485156e-06, "loss": 0.8737, "step": 20993 }, { "epoch": 0.7397658500488914, "grad_norm": 1.1314632892608643, "learning_rate": 3.369529141113722e-06, "loss": 0.9469, "step": 20994 }, { "epoch": 0.7398010870618498, "grad_norm": 3.556718587875366, "learning_rate": 3.3686748499416535e-06, "loss": 4.546, "step": 20995 }, { "epoch": 0.7398363240748081, "grad_norm": 2.5435853004455566, "learning_rate": 3.367820645143446e-06, "loss": 2.3891, "step": 20996 }, { "epoch": 0.7398715610877666, "grad_norm": 1.0491149425506592, "learning_rate": 3.3669665267302165e-06, "loss": 1.021, "step": 20997 }, { "epoch": 0.739906798100725, "grad_norm": 1.1156727075576782, "learning_rate": 3.366112494713093e-06, "loss": 0.9008, "step": 20998 }, { "epoch": 0.7399420351136834, "grad_norm": 2.2648634910583496, "learning_rate": 3.3652585491031988e-06, "loss": 0.6824, "step": 20999 }, { "epoch": 0.7399772721266418, "grad_norm": 3.2133922576904297, "learning_rate": 3.364404689911657e-06, "loss": 3.3081, "step": 21000 }, { "epoch": 0.7400125091396003, "grad_norm": 2.0783920288085938, "learning_rate": 3.3635509171495894e-06, "loss": 1.0242, "step": 21001 }, { "epoch": 0.7400477461525586, "grad_norm": 1.222126841545105, "learning_rate": 3.362697230828117e-06, "loss": 0.9426, "step": 21002 }, { "epoch": 0.740082983165517, "grad_norm": 2.2315874099731445, "learning_rate": 3.361843630958358e-06, "loss": 1.131, "step": 21003 }, { "epoch": 0.7401182201784755, "grad_norm": 2.847856283187866, "learning_rate": 3.3609901175514313e-06, "loss": 3.1507, "step": 21004 }, { "epoch": 0.7401534571914339, "grad_norm": 8.083069801330566, "learning_rate": 3.360136690618454e-06, "loss": 5.189, "step": 21005 }, { "epoch": 0.7401886942043923, "grad_norm": 1.049659252166748, "learning_rate": 3.3592833501705426e-06, "loss": 0.8592, "step": 21006 }, { "epoch": 0.7402239312173508, "grad_norm": 0.9926193952560425, "learning_rate": 3.358430096218811e-06, "loss": 0.928, "step": 21007 }, { "epoch": 0.7402591682303091, "grad_norm": 1.3650555610656738, "learning_rate": 3.357576928774374e-06, "loss": 1.2221, "step": 21008 }, { "epoch": 0.7402944052432675, "grad_norm": 1.2375179529190063, "learning_rate": 3.3567238478483465e-06, "loss": 0.901, "step": 21009 }, { "epoch": 0.740329642256226, "grad_norm": 3.6720938682556152, "learning_rate": 3.35587085345183e-06, "loss": 3.0344, "step": 21010 }, { "epoch": 0.7403648792691844, "grad_norm": 1.550134539604187, "learning_rate": 3.3550179455959464e-06, "loss": 1.0537, "step": 21011 }, { "epoch": 0.7404001162821428, "grad_norm": 1.69136381149292, "learning_rate": 3.3541651242918026e-06, "loss": 0.7033, "step": 21012 }, { "epoch": 0.7404353532951011, "grad_norm": 2.646777391433716, "learning_rate": 3.3533123895505016e-06, "loss": 2.7434, "step": 21013 }, { "epoch": 0.7404705903080596, "grad_norm": 4.205752372741699, "learning_rate": 3.3524597413831508e-06, "loss": 2.7799, "step": 21014 }, { "epoch": 0.740505827321018, "grad_norm": 4.259535312652588, "learning_rate": 3.3516071798008643e-06, "loss": 2.9039, "step": 21015 }, { "epoch": 0.7405410643339764, "grad_norm": 1.1721519231796265, "learning_rate": 3.3507547048147383e-06, "loss": 0.9622, "step": 21016 }, { "epoch": 0.7405763013469349, "grad_norm": 1.8550608158111572, "learning_rate": 3.349902316435879e-06, "loss": 1.0601, "step": 21017 }, { "epoch": 0.7406115383598932, "grad_norm": 1.5179308652877808, "learning_rate": 3.34905001467539e-06, "loss": 0.8987, "step": 21018 }, { "epoch": 0.7406467753728516, "grad_norm": 1.4797013998031616, "learning_rate": 3.3481977995443704e-06, "loss": 0.7369, "step": 21019 }, { "epoch": 0.7406820123858101, "grad_norm": 5.846994876861572, "learning_rate": 3.3473456710539233e-06, "loss": 4.7009, "step": 21020 }, { "epoch": 0.7407172493987685, "grad_norm": 5.260209083557129, "learning_rate": 3.3464936292151452e-06, "loss": 2.7009, "step": 21021 }, { "epoch": 0.7407524864117269, "grad_norm": 1.5104496479034424, "learning_rate": 3.345641674039135e-06, "loss": 0.7079, "step": 21022 }, { "epoch": 0.7407877234246852, "grad_norm": 2.662238836288452, "learning_rate": 3.34478980553699e-06, "loss": 2.5886, "step": 21023 }, { "epoch": 0.7408229604376437, "grad_norm": 1.4850670099258423, "learning_rate": 3.343938023719806e-06, "loss": 0.9847, "step": 21024 }, { "epoch": 0.7408581974506021, "grad_norm": 1.9143807888031006, "learning_rate": 3.343086328598676e-06, "loss": 0.6469, "step": 21025 }, { "epoch": 0.7408934344635605, "grad_norm": 3.479431390762329, "learning_rate": 3.342234720184695e-06, "loss": 3.1422, "step": 21026 }, { "epoch": 0.740928671476519, "grad_norm": 5.979597568511963, "learning_rate": 3.3413831984889554e-06, "loss": 5.3548, "step": 21027 }, { "epoch": 0.7409639084894774, "grad_norm": 1.7975088357925415, "learning_rate": 3.340531763522551e-06, "loss": 0.7162, "step": 21028 }, { "epoch": 0.7409991455024357, "grad_norm": 3.8416106700897217, "learning_rate": 3.3396804152965612e-06, "loss": 5.4739, "step": 21029 }, { "epoch": 0.7410343825153942, "grad_norm": 2.0934512615203857, "learning_rate": 3.3388291538220864e-06, "loss": 0.7999, "step": 21030 }, { "epoch": 0.7410696195283526, "grad_norm": 0.9470532536506653, "learning_rate": 3.337977979110213e-06, "loss": 1.0411, "step": 21031 }, { "epoch": 0.741104856541311, "grad_norm": 1.4844833612442017, "learning_rate": 3.3371268911720223e-06, "loss": 1.0583, "step": 21032 }, { "epoch": 0.7411400935542694, "grad_norm": 3.8591504096984863, "learning_rate": 3.336275890018599e-06, "loss": 2.9612, "step": 21033 }, { "epoch": 0.7411753305672278, "grad_norm": 5.380969524383545, "learning_rate": 3.335424975661038e-06, "loss": 4.8689, "step": 21034 }, { "epoch": 0.7412105675801862, "grad_norm": 4.081583023071289, "learning_rate": 3.334574148110412e-06, "loss": 3.1064, "step": 21035 }, { "epoch": 0.7412458045931446, "grad_norm": 3.5358450412750244, "learning_rate": 3.333723407377807e-06, "loss": 2.5385, "step": 21036 }, { "epoch": 0.7412810416061031, "grad_norm": 2.490770101547241, "learning_rate": 3.3328727534743e-06, "loss": 2.3276, "step": 21037 }, { "epoch": 0.7413162786190615, "grad_norm": 1.276374340057373, "learning_rate": 3.3320221864109815e-06, "loss": 1.2256, "step": 21038 }, { "epoch": 0.7413515156320198, "grad_norm": 3.2728030681610107, "learning_rate": 3.3311717061989202e-06, "loss": 2.9724, "step": 21039 }, { "epoch": 0.7413867526449783, "grad_norm": 1.1729003190994263, "learning_rate": 3.3303213128491975e-06, "loss": 0.7876, "step": 21040 }, { "epoch": 0.7414219896579367, "grad_norm": 4.791574478149414, "learning_rate": 3.3294710063728886e-06, "loss": 2.8064, "step": 21041 }, { "epoch": 0.7414572266708951, "grad_norm": 1.1261212825775146, "learning_rate": 3.3286207867810694e-06, "loss": 0.75, "step": 21042 }, { "epoch": 0.7414924636838535, "grad_norm": 2.807684898376465, "learning_rate": 3.3277706540848145e-06, "loss": 2.4349, "step": 21043 }, { "epoch": 0.741527700696812, "grad_norm": 3.9480478763580322, "learning_rate": 3.3269206082951966e-06, "loss": 3.4484, "step": 21044 }, { "epoch": 0.7415629377097703, "grad_norm": 1.0654191970825195, "learning_rate": 3.3260706494232873e-06, "loss": 0.8253, "step": 21045 }, { "epoch": 0.7415981747227287, "grad_norm": 1.946361780166626, "learning_rate": 3.325220777480159e-06, "loss": 0.9672, "step": 21046 }, { "epoch": 0.7416334117356872, "grad_norm": 3.0672738552093506, "learning_rate": 3.324370992476883e-06, "loss": 2.6718, "step": 21047 }, { "epoch": 0.7416686487486456, "grad_norm": 0.8792442679405212, "learning_rate": 3.323521294424519e-06, "loss": 0.8298, "step": 21048 }, { "epoch": 0.741703885761604, "grad_norm": 4.423359394073486, "learning_rate": 3.3226716833341433e-06, "loss": 2.6642, "step": 21049 }, { "epoch": 0.7417391227745624, "grad_norm": 1.2159028053283691, "learning_rate": 3.321822159216823e-06, "loss": 1.0149, "step": 21050 }, { "epoch": 0.7417743597875208, "grad_norm": 1.0866038799285889, "learning_rate": 3.320972722083616e-06, "loss": 1.0006, "step": 21051 }, { "epoch": 0.7418095968004792, "grad_norm": 7.418279647827148, "learning_rate": 3.3201233719455873e-06, "loss": 2.9788, "step": 21052 }, { "epoch": 0.7418448338134377, "grad_norm": 3.101261854171753, "learning_rate": 3.319274108813809e-06, "loss": 3.1852, "step": 21053 }, { "epoch": 0.7418800708263961, "grad_norm": 5.845250606536865, "learning_rate": 3.3184249326993334e-06, "loss": 5.0377, "step": 21054 }, { "epoch": 0.7419153078393544, "grad_norm": 1.2683311700820923, "learning_rate": 3.3175758436132243e-06, "loss": 0.7444, "step": 21055 }, { "epoch": 0.7419505448523128, "grad_norm": 7.387120723724365, "learning_rate": 3.3167268415665365e-06, "loss": 6.8251, "step": 21056 }, { "epoch": 0.7419857818652713, "grad_norm": 1.3520790338516235, "learning_rate": 3.3158779265703412e-06, "loss": 0.7856, "step": 21057 }, { "epoch": 0.7420210188782297, "grad_norm": 1.8984582424163818, "learning_rate": 3.3150290986356825e-06, "loss": 0.9001, "step": 21058 }, { "epoch": 0.7420562558911881, "grad_norm": 1.1051523685455322, "learning_rate": 3.314180357773622e-06, "loss": 0.8756, "step": 21059 }, { "epoch": 0.7420914929041466, "grad_norm": 5.111873626708984, "learning_rate": 3.313331703995214e-06, "loss": 3.1451, "step": 21060 }, { "epoch": 0.7421267299171049, "grad_norm": 4.9129719734191895, "learning_rate": 3.312483137311513e-06, "loss": 2.8849, "step": 21061 }, { "epoch": 0.7421619669300633, "grad_norm": 9.710933685302734, "learning_rate": 3.31163465773357e-06, "loss": 2.3966, "step": 21062 }, { "epoch": 0.7421972039430218, "grad_norm": 1.2014228105545044, "learning_rate": 3.3107862652724377e-06, "loss": 0.8541, "step": 21063 }, { "epoch": 0.7422324409559802, "grad_norm": 3.8925139904022217, "learning_rate": 3.309937959939167e-06, "loss": 2.355, "step": 21064 }, { "epoch": 0.7422676779689386, "grad_norm": 1.292789340019226, "learning_rate": 3.3090897417448056e-06, "loss": 0.667, "step": 21065 }, { "epoch": 0.7423029149818969, "grad_norm": 2.8852884769439697, "learning_rate": 3.3082416107004068e-06, "loss": 3.1777, "step": 21066 }, { "epoch": 0.7423381519948554, "grad_norm": 2.287330389022827, "learning_rate": 3.3073935668170065e-06, "loss": 2.7215, "step": 21067 }, { "epoch": 0.7423733890078138, "grad_norm": 1.321791172027588, "learning_rate": 3.3065456101056602e-06, "loss": 0.6752, "step": 21068 }, { "epoch": 0.7424086260207722, "grad_norm": 3.3548924922943115, "learning_rate": 3.3056977405774137e-06, "loss": 2.5668, "step": 21069 }, { "epoch": 0.7424438630337307, "grad_norm": 5.488738059997559, "learning_rate": 3.3048499582433037e-06, "loss": 5.1193, "step": 21070 }, { "epoch": 0.742479100046689, "grad_norm": 3.9634363651275635, "learning_rate": 3.304002263114372e-06, "loss": 3.6546, "step": 21071 }, { "epoch": 0.7425143370596474, "grad_norm": 3.171389102935791, "learning_rate": 3.3031546552016713e-06, "loss": 1.0127, "step": 21072 }, { "epoch": 0.7425495740726059, "grad_norm": 5.1685638427734375, "learning_rate": 3.3023071345162305e-06, "loss": 5.7013, "step": 21073 }, { "epoch": 0.7425848110855643, "grad_norm": 1.3854858875274658, "learning_rate": 3.301459701069092e-06, "loss": 0.9446, "step": 21074 }, { "epoch": 0.7426200480985227, "grad_norm": 4.749687194824219, "learning_rate": 3.300612354871291e-06, "loss": 3.1737, "step": 21075 }, { "epoch": 0.742655285111481, "grad_norm": 1.4028297662734985, "learning_rate": 3.2997650959338744e-06, "loss": 0.8139, "step": 21076 }, { "epoch": 0.7426905221244395, "grad_norm": 1.320025086402893, "learning_rate": 3.298917924267867e-06, "loss": 0.9566, "step": 21077 }, { "epoch": 0.7427257591373979, "grad_norm": 1.7617806196212769, "learning_rate": 3.2980708398843088e-06, "loss": 0.7417, "step": 21078 }, { "epoch": 0.7427609961503563, "grad_norm": 1.3583301305770874, "learning_rate": 3.297223842794229e-06, "loss": 1.0044, "step": 21079 }, { "epoch": 0.7427962331633148, "grad_norm": 3.3470115661621094, "learning_rate": 3.2963769330086647e-06, "loss": 2.6204, "step": 21080 }, { "epoch": 0.7428314701762732, "grad_norm": 1.2844487428665161, "learning_rate": 3.2955301105386437e-06, "loss": 1.0744, "step": 21081 }, { "epoch": 0.7428667071892315, "grad_norm": 1.3554946184158325, "learning_rate": 3.294683375395198e-06, "loss": 0.8175, "step": 21082 }, { "epoch": 0.74290194420219, "grad_norm": 1.0726715326309204, "learning_rate": 3.2938367275893546e-06, "loss": 0.9266, "step": 21083 }, { "epoch": 0.7429371812151484, "grad_norm": 1.18583083152771, "learning_rate": 3.292990167132142e-06, "loss": 0.9051, "step": 21084 }, { "epoch": 0.7429724182281068, "grad_norm": 4.772561073303223, "learning_rate": 3.2921436940345895e-06, "loss": 5.967, "step": 21085 }, { "epoch": 0.7430076552410653, "grad_norm": 3.131406545639038, "learning_rate": 3.291297308307715e-06, "loss": 2.8986, "step": 21086 }, { "epoch": 0.7430428922540236, "grad_norm": 3.2544658184051514, "learning_rate": 3.2904510099625498e-06, "loss": 2.9017, "step": 21087 }, { "epoch": 0.743078129266982, "grad_norm": 1.2880558967590332, "learning_rate": 3.2896047990101178e-06, "loss": 0.926, "step": 21088 }, { "epoch": 0.7431133662799404, "grad_norm": 3.92431378364563, "learning_rate": 3.2887586754614355e-06, "loss": 2.843, "step": 21089 }, { "epoch": 0.7431486032928989, "grad_norm": 4.04873514175415, "learning_rate": 3.2879126393275228e-06, "loss": 3.895, "step": 21090 }, { "epoch": 0.7431838403058573, "grad_norm": 2.67083477973938, "learning_rate": 3.2870666906194093e-06, "loss": 2.5193, "step": 21091 }, { "epoch": 0.7432190773188156, "grad_norm": 5.286798000335693, "learning_rate": 3.2862208293481048e-06, "loss": 7.155, "step": 21092 }, { "epoch": 0.7432543143317741, "grad_norm": 4.736953258514404, "learning_rate": 3.285375055524628e-06, "loss": 4.4331, "step": 21093 }, { "epoch": 0.7432895513447325, "grad_norm": 1.4952846765518188, "learning_rate": 3.2845293691599933e-06, "loss": 0.9269, "step": 21094 }, { "epoch": 0.7433247883576909, "grad_norm": 2.044145107269287, "learning_rate": 3.2836837702652257e-06, "loss": 0.8327, "step": 21095 }, { "epoch": 0.7433600253706494, "grad_norm": 1.1359120607376099, "learning_rate": 3.2828382588513297e-06, "loss": 0.8459, "step": 21096 }, { "epoch": 0.7433952623836078, "grad_norm": 1.051896572113037, "learning_rate": 3.2819928349293206e-06, "loss": 0.8422, "step": 21097 }, { "epoch": 0.7434304993965661, "grad_norm": 4.090312957763672, "learning_rate": 3.281147498510211e-06, "loss": 2.6614, "step": 21098 }, { "epoch": 0.7434657364095245, "grad_norm": 1.0460259914398193, "learning_rate": 3.280302249605011e-06, "loss": 0.9145, "step": 21099 }, { "epoch": 0.743500973422483, "grad_norm": 3.320927858352661, "learning_rate": 3.27945708822473e-06, "loss": 3.3055, "step": 21100 }, { "epoch": 0.7435362104354414, "grad_norm": 1.4040124416351318, "learning_rate": 3.2786120143803767e-06, "loss": 0.7996, "step": 21101 }, { "epoch": 0.7435714474483998, "grad_norm": 3.2349441051483154, "learning_rate": 3.2777670280829576e-06, "loss": 3.259, "step": 21102 }, { "epoch": 0.7436066844613582, "grad_norm": 1.3246119022369385, "learning_rate": 3.2769221293434795e-06, "loss": 0.585, "step": 21103 }, { "epoch": 0.7436419214743166, "grad_norm": 2.0038578510284424, "learning_rate": 3.27607731817295e-06, "loss": 1.1157, "step": 21104 }, { "epoch": 0.743677158487275, "grad_norm": 1.2464264631271362, "learning_rate": 3.275232594582364e-06, "loss": 0.9822, "step": 21105 }, { "epoch": 0.7437123955002335, "grad_norm": 1.6265259981155396, "learning_rate": 3.2743879585827333e-06, "loss": 0.7932, "step": 21106 }, { "epoch": 0.7437476325131919, "grad_norm": 3.2363734245300293, "learning_rate": 3.2735434101850583e-06, "loss": 2.446, "step": 21107 }, { "epoch": 0.7437828695261502, "grad_norm": 1.1102060079574585, "learning_rate": 3.272698949400335e-06, "loss": 0.9786, "step": 21108 }, { "epoch": 0.7438181065391086, "grad_norm": 1.2702162265777588, "learning_rate": 3.2718545762395616e-06, "loss": 0.7347, "step": 21109 }, { "epoch": 0.7438533435520671, "grad_norm": 5.466851711273193, "learning_rate": 3.271010290713742e-06, "loss": 5.2517, "step": 21110 }, { "epoch": 0.7438885805650255, "grad_norm": 3.363880157470703, "learning_rate": 3.270166092833874e-06, "loss": 2.622, "step": 21111 }, { "epoch": 0.7439238175779839, "grad_norm": 2.5105085372924805, "learning_rate": 3.2693219826109468e-06, "loss": 2.5946, "step": 21112 }, { "epoch": 0.7439590545909424, "grad_norm": 3.2360386848449707, "learning_rate": 3.268477960055955e-06, "loss": 3.2178, "step": 21113 }, { "epoch": 0.7439942916039007, "grad_norm": 21.866289138793945, "learning_rate": 3.267634025179901e-06, "loss": 2.6379, "step": 21114 }, { "epoch": 0.7440295286168591, "grad_norm": 4.335658550262451, "learning_rate": 3.2667901779937696e-06, "loss": 2.8548, "step": 21115 }, { "epoch": 0.7440647656298176, "grad_norm": 1.4445780515670776, "learning_rate": 3.265946418508552e-06, "loss": 0.9502, "step": 21116 }, { "epoch": 0.744100002642776, "grad_norm": 4.812493324279785, "learning_rate": 3.2651027467352414e-06, "loss": 3.3734, "step": 21117 }, { "epoch": 0.7441352396557344, "grad_norm": 1.4576410055160522, "learning_rate": 3.2642591626848254e-06, "loss": 0.991, "step": 21118 }, { "epoch": 0.7441704766686928, "grad_norm": 1.410439133644104, "learning_rate": 3.26341566636829e-06, "loss": 0.678, "step": 21119 }, { "epoch": 0.7442057136816512, "grad_norm": 0.9216834902763367, "learning_rate": 3.2625722577966255e-06, "loss": 0.6973, "step": 21120 }, { "epoch": 0.7442409506946096, "grad_norm": 1.4524978399276733, "learning_rate": 3.261728936980816e-06, "loss": 0.7162, "step": 21121 }, { "epoch": 0.744276187707568, "grad_norm": 1.1557341814041138, "learning_rate": 3.260885703931844e-06, "loss": 0.7666, "step": 21122 }, { "epoch": 0.7443114247205265, "grad_norm": 1.32395601272583, "learning_rate": 3.2600425586606977e-06, "loss": 0.7901, "step": 21123 }, { "epoch": 0.7443466617334848, "grad_norm": 2.092832088470459, "learning_rate": 3.2591995011783494e-06, "loss": 0.9555, "step": 21124 }, { "epoch": 0.7443818987464432, "grad_norm": 5.23991584777832, "learning_rate": 3.2583565314957887e-06, "loss": 0.9766, "step": 21125 }, { "epoch": 0.7444171357594017, "grad_norm": 0.9259029030799866, "learning_rate": 3.257513649623997e-06, "loss": 1.0218, "step": 21126 }, { "epoch": 0.7444523727723601, "grad_norm": 3.3865692615509033, "learning_rate": 3.2566708555739456e-06, "loss": 3.0854, "step": 21127 }, { "epoch": 0.7444876097853185, "grad_norm": 2.210658073425293, "learning_rate": 3.2558281493566125e-06, "loss": 2.9968, "step": 21128 }, { "epoch": 0.744522846798277, "grad_norm": 3.26177978515625, "learning_rate": 3.254985530982979e-06, "loss": 3.1545, "step": 21129 }, { "epoch": 0.7445580838112353, "grad_norm": 0.7779905200004578, "learning_rate": 3.2541430004640216e-06, "loss": 0.7283, "step": 21130 }, { "epoch": 0.7445933208241937, "grad_norm": 7.066875457763672, "learning_rate": 3.253300557810708e-06, "loss": 5.1236, "step": 21131 }, { "epoch": 0.7446285578371521, "grad_norm": 2.664177417755127, "learning_rate": 3.2524582030340133e-06, "loss": 2.5478, "step": 21132 }, { "epoch": 0.7446637948501106, "grad_norm": 1.7420035600662231, "learning_rate": 3.251615936144912e-06, "loss": 0.9995, "step": 21133 }, { "epoch": 0.744699031863069, "grad_norm": 1.151262640953064, "learning_rate": 3.2507737571543708e-06, "loss": 0.7236, "step": 21134 }, { "epoch": 0.7447342688760273, "grad_norm": 1.1276534795761108, "learning_rate": 3.2499316660733614e-06, "loss": 1.0267, "step": 21135 }, { "epoch": 0.7447695058889858, "grad_norm": 3.2883801460266113, "learning_rate": 3.249089662912852e-06, "loss": 2.9777, "step": 21136 }, { "epoch": 0.7448047429019442, "grad_norm": 1.9095877408981323, "learning_rate": 3.2482477476838103e-06, "loss": 0.7799, "step": 21137 }, { "epoch": 0.7448399799149026, "grad_norm": 3.2578141689300537, "learning_rate": 3.2474059203972008e-06, "loss": 2.5956, "step": 21138 }, { "epoch": 0.7448752169278611, "grad_norm": 5.015048980712891, "learning_rate": 3.2465641810639903e-06, "loss": 2.5809, "step": 21139 }, { "epoch": 0.7449104539408195, "grad_norm": 3.05068302154541, "learning_rate": 3.2457225296951413e-06, "loss": 2.9703, "step": 21140 }, { "epoch": 0.7449456909537778, "grad_norm": 4.256862163543701, "learning_rate": 3.2448809663016156e-06, "loss": 3.0068, "step": 21141 }, { "epoch": 0.7449809279667362, "grad_norm": 3.137685775756836, "learning_rate": 3.2440394908943805e-06, "loss": 2.8601, "step": 21142 }, { "epoch": 0.7450161649796947, "grad_norm": 2.7684741020202637, "learning_rate": 3.2431981034843874e-06, "loss": 3.1195, "step": 21143 }, { "epoch": 0.7450514019926531, "grad_norm": 1.5427218675613403, "learning_rate": 3.242356804082597e-06, "loss": 0.8284, "step": 21144 }, { "epoch": 0.7450866390056115, "grad_norm": 1.2805125713348389, "learning_rate": 3.241515592699976e-06, "loss": 1.1223, "step": 21145 }, { "epoch": 0.7451218760185699, "grad_norm": 0.9572580456733704, "learning_rate": 3.240674469347472e-06, "loss": 0.7556, "step": 21146 }, { "epoch": 0.7451571130315283, "grad_norm": 4.925874710083008, "learning_rate": 3.239833434036045e-06, "loss": 3.2792, "step": 21147 }, { "epoch": 0.7451923500444867, "grad_norm": 2.2178027629852295, "learning_rate": 3.2389924867766444e-06, "loss": 2.9197, "step": 21148 }, { "epoch": 0.7452275870574452, "grad_norm": 1.668158769607544, "learning_rate": 3.238151627580235e-06, "loss": 0.8358, "step": 21149 }, { "epoch": 0.7452628240704036, "grad_norm": 4.16357421875, "learning_rate": 3.237310856457758e-06, "loss": 4.8835, "step": 21150 }, { "epoch": 0.7452980610833619, "grad_norm": 1.603163480758667, "learning_rate": 3.23647017342017e-06, "loss": 0.9241, "step": 21151 }, { "epoch": 0.7453332980963204, "grad_norm": 11.218500137329102, "learning_rate": 3.2356295784784177e-06, "loss": 2.954, "step": 21152 }, { "epoch": 0.7453685351092788, "grad_norm": 3.8493213653564453, "learning_rate": 3.2347890716434526e-06, "loss": 2.9348, "step": 21153 }, { "epoch": 0.7454037721222372, "grad_norm": 2.476219654083252, "learning_rate": 3.233948652926222e-06, "loss": 0.6933, "step": 21154 }, { "epoch": 0.7454390091351956, "grad_norm": 2.7837798595428467, "learning_rate": 3.2331083223376713e-06, "loss": 0.6581, "step": 21155 }, { "epoch": 0.745474246148154, "grad_norm": 6.59421968460083, "learning_rate": 3.2322680798887475e-06, "loss": 4.5862, "step": 21156 }, { "epoch": 0.7455094831611124, "grad_norm": 3.795607328414917, "learning_rate": 3.2314279255903935e-06, "loss": 3.4425, "step": 21157 }, { "epoch": 0.7455447201740708, "grad_norm": 3.455761671066284, "learning_rate": 3.2305878594535554e-06, "loss": 4.2366, "step": 21158 }, { "epoch": 0.7455799571870293, "grad_norm": 3.256748676300049, "learning_rate": 3.229747881489168e-06, "loss": 2.9993, "step": 21159 }, { "epoch": 0.7456151941999877, "grad_norm": 1.7518751621246338, "learning_rate": 3.228907991708179e-06, "loss": 0.6578, "step": 21160 }, { "epoch": 0.745650431212946, "grad_norm": 2.31825590133667, "learning_rate": 3.2280681901215294e-06, "loss": 0.7791, "step": 21161 }, { "epoch": 0.7456856682259045, "grad_norm": 4.117037296295166, "learning_rate": 3.2272284767401507e-06, "loss": 3.3447, "step": 21162 }, { "epoch": 0.7457209052388629, "grad_norm": 2.4333572387695312, "learning_rate": 3.226388851574981e-06, "loss": 3.2063, "step": 21163 }, { "epoch": 0.7457561422518213, "grad_norm": 4.743281841278076, "learning_rate": 3.225549314636965e-06, "loss": 3.1674, "step": 21164 }, { "epoch": 0.7457913792647797, "grad_norm": 4.078580379486084, "learning_rate": 3.2247098659370293e-06, "loss": 2.9213, "step": 21165 }, { "epoch": 0.7458266162777382, "grad_norm": 3.332022190093994, "learning_rate": 3.2238705054861107e-06, "loss": 2.9912, "step": 21166 }, { "epoch": 0.7458618532906965, "grad_norm": 2.5855767726898193, "learning_rate": 3.2230312332951385e-06, "loss": 0.9861, "step": 21167 }, { "epoch": 0.7458970903036549, "grad_norm": 6.576940536499023, "learning_rate": 3.222192049375055e-06, "loss": 3.4285, "step": 21168 }, { "epoch": 0.7459323273166134, "grad_norm": 3.4934942722320557, "learning_rate": 3.2213529537367794e-06, "loss": 2.7125, "step": 21169 }, { "epoch": 0.7459675643295718, "grad_norm": 2.341840982437134, "learning_rate": 3.220513946391245e-06, "loss": 0.9836, "step": 21170 }, { "epoch": 0.7460028013425302, "grad_norm": 3.6293134689331055, "learning_rate": 3.2196750273493803e-06, "loss": 3.2584, "step": 21171 }, { "epoch": 0.7460380383554887, "grad_norm": 1.6797192096710205, "learning_rate": 3.2188361966221127e-06, "loss": 0.811, "step": 21172 }, { "epoch": 0.746073275368447, "grad_norm": 3.523678779602051, "learning_rate": 3.2179974542203664e-06, "loss": 4.1712, "step": 21173 }, { "epoch": 0.7461085123814054, "grad_norm": 3.32818865776062, "learning_rate": 3.217158800155068e-06, "loss": 3.4271, "step": 21174 }, { "epoch": 0.7461437493943638, "grad_norm": 4.008873462677002, "learning_rate": 3.2163202344371392e-06, "loss": 3.0167, "step": 21175 }, { "epoch": 0.7461789864073223, "grad_norm": 3.084899663925171, "learning_rate": 3.215481757077504e-06, "loss": 2.2828, "step": 21176 }, { "epoch": 0.7462142234202807, "grad_norm": 3.2474749088287354, "learning_rate": 3.214643368087086e-06, "loss": 3.1203, "step": 21177 }, { "epoch": 0.746249460433239, "grad_norm": 1.278365135192871, "learning_rate": 3.213805067476796e-06, "loss": 0.8735, "step": 21178 }, { "epoch": 0.7462846974461975, "grad_norm": 1.399672031402588, "learning_rate": 3.2129668552575634e-06, "loss": 0.6523, "step": 21179 }, { "epoch": 0.7463199344591559, "grad_norm": 2.783163547515869, "learning_rate": 3.212128731440305e-06, "loss": 1.8572, "step": 21180 }, { "epoch": 0.7463551714721143, "grad_norm": 2.7767720222473145, "learning_rate": 3.2112906960359313e-06, "loss": 2.6916, "step": 21181 }, { "epoch": 0.7463904084850728, "grad_norm": 4.7453083992004395, "learning_rate": 3.210452749055357e-06, "loss": 3.3585, "step": 21182 }, { "epoch": 0.7464256454980311, "grad_norm": 1.4925652742385864, "learning_rate": 3.209614890509505e-06, "loss": 1.1017, "step": 21183 }, { "epoch": 0.7464608825109895, "grad_norm": 1.3698718547821045, "learning_rate": 3.2087771204092864e-06, "loss": 0.9301, "step": 21184 }, { "epoch": 0.746496119523948, "grad_norm": 6.375796794891357, "learning_rate": 3.207939438765607e-06, "loss": 5.2609, "step": 21185 }, { "epoch": 0.7465313565369064, "grad_norm": 1.7564661502838135, "learning_rate": 3.20710184558938e-06, "loss": 0.85, "step": 21186 }, { "epoch": 0.7465665935498648, "grad_norm": 1.682356357574463, "learning_rate": 3.2062643408915216e-06, "loss": 0.8447, "step": 21187 }, { "epoch": 0.7466018305628231, "grad_norm": 0.8702467679977417, "learning_rate": 3.205426924682934e-06, "loss": 0.7714, "step": 21188 }, { "epoch": 0.7466370675757816, "grad_norm": 4.709803104400635, "learning_rate": 3.2045895969745243e-06, "loss": 3.1759, "step": 21189 }, { "epoch": 0.74667230458874, "grad_norm": 4.8450846672058105, "learning_rate": 3.2037523577772023e-06, "loss": 3.3049, "step": 21190 }, { "epoch": 0.7467075416016984, "grad_norm": 4.910796642303467, "learning_rate": 3.202915207101871e-06, "loss": 2.3964, "step": 21191 }, { "epoch": 0.7467427786146569, "grad_norm": 2.436903476715088, "learning_rate": 3.202078144959434e-06, "loss": 0.8047, "step": 21192 }, { "epoch": 0.7467780156276153, "grad_norm": 4.090433120727539, "learning_rate": 3.201241171360796e-06, "loss": 3.3418, "step": 21193 }, { "epoch": 0.7468132526405736, "grad_norm": 1.993626594543457, "learning_rate": 3.2004042863168585e-06, "loss": 0.698, "step": 21194 }, { "epoch": 0.7468484896535321, "grad_norm": 1.4181323051452637, "learning_rate": 3.19956748983852e-06, "loss": 1.0535, "step": 21195 }, { "epoch": 0.7468837266664905, "grad_norm": 3.7485342025756836, "learning_rate": 3.198730781936685e-06, "loss": 3.5081, "step": 21196 }, { "epoch": 0.7469189636794489, "grad_norm": 3.3337290287017822, "learning_rate": 3.1978941626222415e-06, "loss": 2.927, "step": 21197 }, { "epoch": 0.7469542006924073, "grad_norm": 2.8351516723632812, "learning_rate": 3.1970576319060953e-06, "loss": 1.0735, "step": 21198 }, { "epoch": 0.7469894377053657, "grad_norm": 3.0118675231933594, "learning_rate": 3.1962211897991448e-06, "loss": 3.0139, "step": 21199 }, { "epoch": 0.7470246747183241, "grad_norm": 1.2269233465194702, "learning_rate": 3.195384836312275e-06, "loss": 1.1455, "step": 21200 }, { "epoch": 0.7470599117312825, "grad_norm": 1.5530390739440918, "learning_rate": 3.194548571456383e-06, "loss": 0.895, "step": 21201 }, { "epoch": 0.747095148744241, "grad_norm": 5.457483291625977, "learning_rate": 3.1937123952423655e-06, "loss": 5.0795, "step": 21202 }, { "epoch": 0.7471303857571994, "grad_norm": 1.3031566143035889, "learning_rate": 3.1928763076811133e-06, "loss": 0.9148, "step": 21203 }, { "epoch": 0.7471656227701577, "grad_norm": 3.688415765762329, "learning_rate": 3.1920403087835116e-06, "loss": 2.8602, "step": 21204 }, { "epoch": 0.7472008597831162, "grad_norm": 1.7875158786773682, "learning_rate": 3.191204398560449e-06, "loss": 0.8632, "step": 21205 }, { "epoch": 0.7472360967960746, "grad_norm": 5.63530969619751, "learning_rate": 3.190368577022823e-06, "loss": 5.17, "step": 21206 }, { "epoch": 0.747271333809033, "grad_norm": 0.9092466235160828, "learning_rate": 3.1895328441815097e-06, "loss": 1.1407, "step": 21207 }, { "epoch": 0.7473065708219914, "grad_norm": 2.146876096725464, "learning_rate": 3.188697200047399e-06, "loss": 2.5512, "step": 21208 }, { "epoch": 0.7473418078349499, "grad_norm": 3.7667455673217773, "learning_rate": 3.187861644631376e-06, "loss": 3.1429, "step": 21209 }, { "epoch": 0.7473770448479082, "grad_norm": 11.422469139099121, "learning_rate": 3.187026177944321e-06, "loss": 5.1652, "step": 21210 }, { "epoch": 0.7474122818608666, "grad_norm": 3.9480879306793213, "learning_rate": 3.1861907999971186e-06, "loss": 3.0224, "step": 21211 }, { "epoch": 0.7474475188738251, "grad_norm": 3.4646081924438477, "learning_rate": 3.1853555108006496e-06, "loss": 2.8575, "step": 21212 }, { "epoch": 0.7474827558867835, "grad_norm": 1.349846363067627, "learning_rate": 3.184520310365793e-06, "loss": 0.8466, "step": 21213 }, { "epoch": 0.7475179928997419, "grad_norm": 4.671499252319336, "learning_rate": 3.1836851987034266e-06, "loss": 3.6994, "step": 21214 }, { "epoch": 0.7475532299127003, "grad_norm": 1.5253140926361084, "learning_rate": 3.1828501758244323e-06, "loss": 0.9994, "step": 21215 }, { "epoch": 0.7475884669256587, "grad_norm": 2.606131076812744, "learning_rate": 3.1820152417396765e-06, "loss": 0.8837, "step": 21216 }, { "epoch": 0.7476237039386171, "grad_norm": 0.7786887884140015, "learning_rate": 3.181180396460044e-06, "loss": 0.7539, "step": 21217 }, { "epoch": 0.7476589409515756, "grad_norm": 3.115950107574463, "learning_rate": 3.1803456399964084e-06, "loss": 3.255, "step": 21218 }, { "epoch": 0.747694177964534, "grad_norm": 4.115999221801758, "learning_rate": 3.1795109723596364e-06, "loss": 5.2588, "step": 21219 }, { "epoch": 0.7477294149774923, "grad_norm": 3.904500961303711, "learning_rate": 3.1786763935606003e-06, "loss": 0.8645, "step": 21220 }, { "epoch": 0.7477646519904507, "grad_norm": 3.9572525024414062, "learning_rate": 3.177841903610175e-06, "loss": 3.101, "step": 21221 }, { "epoch": 0.7477998890034092, "grad_norm": 4.173479080200195, "learning_rate": 3.1770075025192313e-06, "loss": 3.0574, "step": 21222 }, { "epoch": 0.7478351260163676, "grad_norm": 1.4845538139343262, "learning_rate": 3.1761731902986313e-06, "loss": 0.9734, "step": 21223 }, { "epoch": 0.747870363029326, "grad_norm": 3.722687005996704, "learning_rate": 3.1753389669592415e-06, "loss": 3.3612, "step": 21224 }, { "epoch": 0.7479056000422845, "grad_norm": 1.2042196989059448, "learning_rate": 3.174504832511938e-06, "loss": 0.8454, "step": 21225 }, { "epoch": 0.7479408370552428, "grad_norm": 2.0160398483276367, "learning_rate": 3.173670786967574e-06, "loss": 0.9505, "step": 21226 }, { "epoch": 0.7479760740682012, "grad_norm": 3.297426700592041, "learning_rate": 3.1728368303370182e-06, "loss": 2.938, "step": 21227 }, { "epoch": 0.7480113110811597, "grad_norm": 4.289209365844727, "learning_rate": 3.172002962631132e-06, "loss": 4.7313, "step": 21228 }, { "epoch": 0.7480465480941181, "grad_norm": 4.706963539123535, "learning_rate": 3.1711691838607785e-06, "loss": 3.1137, "step": 21229 }, { "epoch": 0.7480817851070765, "grad_norm": 1.304227352142334, "learning_rate": 3.1703354940368148e-06, "loss": 0.7501, "step": 21230 }, { "epoch": 0.7481170221200348, "grad_norm": 2.0709869861602783, "learning_rate": 3.1695018931701006e-06, "loss": 0.8364, "step": 21231 }, { "epoch": 0.7481522591329933, "grad_norm": 2.955225944519043, "learning_rate": 3.168668381271496e-06, "loss": 2.6986, "step": 21232 }, { "epoch": 0.7481874961459517, "grad_norm": 0.9511698484420776, "learning_rate": 3.167834958351854e-06, "loss": 0.6471, "step": 21233 }, { "epoch": 0.7482227331589101, "grad_norm": 2.701909303665161, "learning_rate": 3.167001624422037e-06, "loss": 2.4602, "step": 21234 }, { "epoch": 0.7482579701718686, "grad_norm": 2.259885787963867, "learning_rate": 3.1661683794928864e-06, "loss": 0.677, "step": 21235 }, { "epoch": 0.748293207184827, "grad_norm": 1.0259379148483276, "learning_rate": 3.1653352235752675e-06, "loss": 1.0073, "step": 21236 }, { "epoch": 0.7483284441977853, "grad_norm": 4.249759674072266, "learning_rate": 3.1645021566800304e-06, "loss": 2.9328, "step": 21237 }, { "epoch": 0.7483636812107438, "grad_norm": 1.6308863162994385, "learning_rate": 3.163669178818021e-06, "loss": 0.8566, "step": 21238 }, { "epoch": 0.7483989182237022, "grad_norm": 0.9678855538368225, "learning_rate": 3.1628362900000866e-06, "loss": 0.812, "step": 21239 }, { "epoch": 0.7484341552366606, "grad_norm": 2.2813773155212402, "learning_rate": 3.1620034902370842e-06, "loss": 0.9411, "step": 21240 }, { "epoch": 0.748469392249619, "grad_norm": 4.356629848480225, "learning_rate": 3.1611707795398614e-06, "loss": 5.1278, "step": 21241 }, { "epoch": 0.7485046292625774, "grad_norm": 5.0903167724609375, "learning_rate": 3.160338157919257e-06, "loss": 4.7045, "step": 21242 }, { "epoch": 0.7485398662755358, "grad_norm": 5.166206359863281, "learning_rate": 3.1595056253861157e-06, "loss": 4.7723, "step": 21243 }, { "epoch": 0.7485751032884942, "grad_norm": 4.137970924377441, "learning_rate": 3.1586731819512916e-06, "loss": 3.2284, "step": 21244 }, { "epoch": 0.7486103403014527, "grad_norm": 2.856642723083496, "learning_rate": 3.157840827625617e-06, "loss": 2.8044, "step": 21245 }, { "epoch": 0.7486455773144111, "grad_norm": 3.9572205543518066, "learning_rate": 3.1570085624199377e-06, "loss": 5.0061, "step": 21246 }, { "epoch": 0.7486808143273694, "grad_norm": 1.0065144300460815, "learning_rate": 3.1561763863450936e-06, "loss": 0.775, "step": 21247 }, { "epoch": 0.7487160513403279, "grad_norm": 6.934474945068359, "learning_rate": 3.155344299411923e-06, "loss": 3.0857, "step": 21248 }, { "epoch": 0.7487512883532863, "grad_norm": 4.713164806365967, "learning_rate": 3.154512301631266e-06, "loss": 2.9927, "step": 21249 }, { "epoch": 0.7487865253662447, "grad_norm": 1.2768698930740356, "learning_rate": 3.153680393013957e-06, "loss": 0.8421, "step": 21250 }, { "epoch": 0.7488217623792031, "grad_norm": 1.3993899822235107, "learning_rate": 3.1528485735708337e-06, "loss": 0.966, "step": 21251 }, { "epoch": 0.7488569993921615, "grad_norm": 5.5611467361450195, "learning_rate": 3.1520168433127295e-06, "loss": 4.9348, "step": 21252 }, { "epoch": 0.7488922364051199, "grad_norm": 3.626491069793701, "learning_rate": 3.1511852022504828e-06, "loss": 3.0943, "step": 21253 }, { "epoch": 0.7489274734180783, "grad_norm": 0.989440381526947, "learning_rate": 3.1503536503949127e-06, "loss": 0.7737, "step": 21254 }, { "epoch": 0.7489627104310368, "grad_norm": 3.965115547180176, "learning_rate": 3.149522187756864e-06, "loss": 4.2581, "step": 21255 }, { "epoch": 0.7489979474439952, "grad_norm": 1.936128854751587, "learning_rate": 3.1486908143471606e-06, "loss": 0.9708, "step": 21256 }, { "epoch": 0.7490331844569535, "grad_norm": 1.7277114391326904, "learning_rate": 3.1478595301766347e-06, "loss": 0.6561, "step": 21257 }, { "epoch": 0.749068421469912, "grad_norm": 0.9479044675827026, "learning_rate": 3.147028335256106e-06, "loss": 0.639, "step": 21258 }, { "epoch": 0.7491036584828704, "grad_norm": 3.568185806274414, "learning_rate": 3.1461972295964082e-06, "loss": 2.6233, "step": 21259 }, { "epoch": 0.7491388954958288, "grad_norm": 1.3090931177139282, "learning_rate": 3.1453662132083686e-06, "loss": 1.1257, "step": 21260 }, { "epoch": 0.7491741325087873, "grad_norm": 1.9332494735717773, "learning_rate": 3.144535286102803e-06, "loss": 0.8546, "step": 21261 }, { "epoch": 0.7492093695217457, "grad_norm": 3.672361373901367, "learning_rate": 3.1437044482905366e-06, "loss": 2.7786, "step": 21262 }, { "epoch": 0.749244606534704, "grad_norm": 1.0151129961013794, "learning_rate": 3.1428736997823984e-06, "loss": 0.7441, "step": 21263 }, { "epoch": 0.7492798435476624, "grad_norm": 1.619027018547058, "learning_rate": 3.142043040589201e-06, "loss": 0.9129, "step": 21264 }, { "epoch": 0.7493150805606209, "grad_norm": 4.636424541473389, "learning_rate": 3.1412124707217663e-06, "loss": 4.78, "step": 21265 }, { "epoch": 0.7493503175735793, "grad_norm": 1.4944075345993042, "learning_rate": 3.1403819901909137e-06, "loss": 1.0059, "step": 21266 }, { "epoch": 0.7493855545865377, "grad_norm": 1.0829960107803345, "learning_rate": 3.139551599007459e-06, "loss": 0.7216, "step": 21267 }, { "epoch": 0.7494207915994961, "grad_norm": 1.2710238695144653, "learning_rate": 3.1387212971822177e-06, "loss": 0.9347, "step": 21268 }, { "epoch": 0.7494560286124545, "grad_norm": 0.88849276304245, "learning_rate": 3.137891084726007e-06, "loss": 0.7717, "step": 21269 }, { "epoch": 0.7494912656254129, "grad_norm": 4.111162185668945, "learning_rate": 3.137060961649637e-06, "loss": 2.8409, "step": 21270 }, { "epoch": 0.7495265026383714, "grad_norm": 2.148031234741211, "learning_rate": 3.1362309279639237e-06, "loss": 0.6842, "step": 21271 }, { "epoch": 0.7495617396513298, "grad_norm": 2.556368350982666, "learning_rate": 3.1354009836796808e-06, "loss": 2.8517, "step": 21272 }, { "epoch": 0.7495969766642882, "grad_norm": 1.3163564205169678, "learning_rate": 3.1345711288077064e-06, "loss": 1.1724, "step": 21273 }, { "epoch": 0.7496322136772465, "grad_norm": 5.745412826538086, "learning_rate": 3.133741363358822e-06, "loss": 6.0475, "step": 21274 }, { "epoch": 0.749667450690205, "grad_norm": 4.733509540557861, "learning_rate": 3.1329116873438304e-06, "loss": 3.5496, "step": 21275 }, { "epoch": 0.7497026877031634, "grad_norm": 0.8416821360588074, "learning_rate": 3.1320821007735413e-06, "loss": 0.9404, "step": 21276 }, { "epoch": 0.7497379247161218, "grad_norm": 4.247461795806885, "learning_rate": 3.131252603658752e-06, "loss": 2.903, "step": 21277 }, { "epoch": 0.7497731617290803, "grad_norm": 2.1049246788024902, "learning_rate": 3.130423196010276e-06, "loss": 1.1031, "step": 21278 }, { "epoch": 0.7498083987420386, "grad_norm": 3.1490676403045654, "learning_rate": 3.1295938778389156e-06, "loss": 3.172, "step": 21279 }, { "epoch": 0.749843635754997, "grad_norm": 6.474460601806641, "learning_rate": 3.128764649155467e-06, "loss": 4.5496, "step": 21280 }, { "epoch": 0.7498788727679555, "grad_norm": 2.193561553955078, "learning_rate": 3.1279355099707308e-06, "loss": 0.9466, "step": 21281 }, { "epoch": 0.7499141097809139, "grad_norm": 1.313612461090088, "learning_rate": 3.127106460295517e-06, "loss": 0.7983, "step": 21282 }, { "epoch": 0.7499493467938723, "grad_norm": 2.128967761993408, "learning_rate": 3.126277500140613e-06, "loss": 1.0741, "step": 21283 }, { "epoch": 0.7499845838068306, "grad_norm": 4.018016815185547, "learning_rate": 3.125448629516821e-06, "loss": 3.2026, "step": 21284 }, { "epoch": 0.7500198208197891, "grad_norm": 3.1638600826263428, "learning_rate": 3.1246198484349354e-06, "loss": 2.52, "step": 21285 }, { "epoch": 0.7500550578327475, "grad_norm": 4.553194999694824, "learning_rate": 3.123791156905752e-06, "loss": 4.3377, "step": 21286 }, { "epoch": 0.7500902948457059, "grad_norm": 1.4686002731323242, "learning_rate": 3.1229625549400654e-06, "loss": 0.8365, "step": 21287 }, { "epoch": 0.7501255318586644, "grad_norm": 6.046433448791504, "learning_rate": 3.122134042548668e-06, "loss": 6.3337, "step": 21288 }, { "epoch": 0.7501607688716228, "grad_norm": 3.152862071990967, "learning_rate": 3.1213056197423497e-06, "loss": 2.9919, "step": 21289 }, { "epoch": 0.7501960058845811, "grad_norm": 1.147377848625183, "learning_rate": 3.1204772865319024e-06, "loss": 0.7087, "step": 21290 }, { "epoch": 0.7502312428975396, "grad_norm": 1.8288110494613647, "learning_rate": 3.119649042928117e-06, "loss": 1.0183, "step": 21291 }, { "epoch": 0.750266479910498, "grad_norm": 2.8476808071136475, "learning_rate": 3.1188208889417737e-06, "loss": 2.4548, "step": 21292 }, { "epoch": 0.7503017169234564, "grad_norm": 1.1454414129257202, "learning_rate": 3.117992824583668e-06, "loss": 0.8424, "step": 21293 }, { "epoch": 0.7503369539364149, "grad_norm": 2.74692964553833, "learning_rate": 3.1171648498645825e-06, "loss": 0.8916, "step": 21294 }, { "epoch": 0.7503721909493732, "grad_norm": 7.069122791290283, "learning_rate": 3.1163369647953046e-06, "loss": 4.1703, "step": 21295 }, { "epoch": 0.7504074279623316, "grad_norm": 1.5391061305999756, "learning_rate": 3.115509169386608e-06, "loss": 0.8386, "step": 21296 }, { "epoch": 0.75044266497529, "grad_norm": 1.1636099815368652, "learning_rate": 3.114681463649285e-06, "loss": 0.8038, "step": 21297 }, { "epoch": 0.7504779019882485, "grad_norm": 2.107957601547241, "learning_rate": 3.113853847594116e-06, "loss": 2.4063, "step": 21298 }, { "epoch": 0.7505131390012069, "grad_norm": 4.498339653015137, "learning_rate": 3.1130263212318746e-06, "loss": 2.7987, "step": 21299 }, { "epoch": 0.7505483760141652, "grad_norm": 3.911529302597046, "learning_rate": 3.1121988845733386e-06, "loss": 4.4912, "step": 21300 }, { "epoch": 0.7505836130271237, "grad_norm": 3.3283979892730713, "learning_rate": 3.1113715376292976e-06, "loss": 2.0102, "step": 21301 }, { "epoch": 0.7506188500400821, "grad_norm": 4.4903411865234375, "learning_rate": 3.1105442804105156e-06, "loss": 5.0175, "step": 21302 }, { "epoch": 0.7506540870530405, "grad_norm": 4.469829082489014, "learning_rate": 3.1097171129277716e-06, "loss": 2.9304, "step": 21303 }, { "epoch": 0.750689324065999, "grad_norm": 1.0902595520019531, "learning_rate": 3.1088900351918406e-06, "loss": 0.9579, "step": 21304 }, { "epoch": 0.7507245610789574, "grad_norm": 3.5473382472991943, "learning_rate": 3.1080630472134945e-06, "loss": 4.7622, "step": 21305 }, { "epoch": 0.7507597980919157, "grad_norm": 1.0412594079971313, "learning_rate": 3.107236149003504e-06, "loss": 0.8663, "step": 21306 }, { "epoch": 0.7507950351048741, "grad_norm": 1.1970386505126953, "learning_rate": 3.1064093405726426e-06, "loss": 1.1708, "step": 21307 }, { "epoch": 0.7508302721178326, "grad_norm": 1.2912635803222656, "learning_rate": 3.1055826219316764e-06, "loss": 1.0204, "step": 21308 }, { "epoch": 0.750865509130791, "grad_norm": 5.575351238250732, "learning_rate": 3.104755993091375e-06, "loss": 5.2087, "step": 21309 }, { "epoch": 0.7509007461437494, "grad_norm": 3.480834722518921, "learning_rate": 3.103929454062509e-06, "loss": 2.9268, "step": 21310 }, { "epoch": 0.7509359831567078, "grad_norm": 1.533273458480835, "learning_rate": 3.103103004855832e-06, "loss": 0.8767, "step": 21311 }, { "epoch": 0.7509712201696662, "grad_norm": 1.0496971607208252, "learning_rate": 3.1022766454821216e-06, "loss": 0.9206, "step": 21312 }, { "epoch": 0.7510064571826246, "grad_norm": 3.3574132919311523, "learning_rate": 3.1014503759521366e-06, "loss": 2.5041, "step": 21313 }, { "epoch": 0.7510416941955831, "grad_norm": 3.436793088912964, "learning_rate": 3.1006241962766415e-06, "loss": 2.7037, "step": 21314 }, { "epoch": 0.7510769312085415, "grad_norm": 3.72562313079834, "learning_rate": 3.0997981064663895e-06, "loss": 2.8387, "step": 21315 }, { "epoch": 0.7511121682214998, "grad_norm": 1.4826622009277344, "learning_rate": 3.0989721065321486e-06, "loss": 0.7071, "step": 21316 }, { "epoch": 0.7511474052344582, "grad_norm": 2.564241409301758, "learning_rate": 3.0981461964846793e-06, "loss": 2.8635, "step": 21317 }, { "epoch": 0.7511826422474167, "grad_norm": 5.7490458488464355, "learning_rate": 3.0973203763347303e-06, "loss": 2.582, "step": 21318 }, { "epoch": 0.7512178792603751, "grad_norm": 2.677001953125, "learning_rate": 3.096494646093061e-06, "loss": 2.6125, "step": 21319 }, { "epoch": 0.7512531162733335, "grad_norm": 3.0263867378234863, "learning_rate": 3.095669005770434e-06, "loss": 3.1703, "step": 21320 }, { "epoch": 0.751288353286292, "grad_norm": 4.5708909034729, "learning_rate": 3.094843455377594e-06, "loss": 2.8525, "step": 21321 }, { "epoch": 0.7513235902992503, "grad_norm": 1.4885281324386597, "learning_rate": 3.094017994925298e-06, "loss": 0.6888, "step": 21322 }, { "epoch": 0.7513588273122087, "grad_norm": 2.927276134490967, "learning_rate": 3.0931926244242972e-06, "loss": 2.7087, "step": 21323 }, { "epoch": 0.7513940643251672, "grad_norm": 6.791018009185791, "learning_rate": 3.092367343885343e-06, "loss": 5.3013, "step": 21324 }, { "epoch": 0.7514293013381256, "grad_norm": 1.0268429517745972, "learning_rate": 3.0915421533191826e-06, "loss": 0.8075, "step": 21325 }, { "epoch": 0.751464538351084, "grad_norm": 1.570176601409912, "learning_rate": 3.090717052736567e-06, "loss": 0.9171, "step": 21326 }, { "epoch": 0.7514997753640424, "grad_norm": 6.262390613555908, "learning_rate": 3.0898920421482413e-06, "loss": 5.4164, "step": 21327 }, { "epoch": 0.7515350123770008, "grad_norm": 5.009958267211914, "learning_rate": 3.0890671215649513e-06, "loss": 6.5528, "step": 21328 }, { "epoch": 0.7515702493899592, "grad_norm": 1.289643406867981, "learning_rate": 3.088242290997444e-06, "loss": 1.2503, "step": 21329 }, { "epoch": 0.7516054864029176, "grad_norm": 1.38153076171875, "learning_rate": 3.0874175504564595e-06, "loss": 0.9295, "step": 21330 }, { "epoch": 0.7516407234158761, "grad_norm": 5.44016695022583, "learning_rate": 3.0865928999527426e-06, "loss": 5.1421, "step": 21331 }, { "epoch": 0.7516759604288344, "grad_norm": 4.571765422821045, "learning_rate": 3.085768339497035e-06, "loss": 4.6027, "step": 21332 }, { "epoch": 0.7517111974417928, "grad_norm": 1.0376845598220825, "learning_rate": 3.084943869100078e-06, "loss": 0.9085, "step": 21333 }, { "epoch": 0.7517464344547513, "grad_norm": 1.1108086109161377, "learning_rate": 3.0841194887726e-06, "loss": 1.0965, "step": 21334 }, { "epoch": 0.7517816714677097, "grad_norm": 3.346349000930786, "learning_rate": 3.0832951985253514e-06, "loss": 2.2866, "step": 21335 }, { "epoch": 0.7518169084806681, "grad_norm": 2.906074047088623, "learning_rate": 3.0824709983690673e-06, "loss": 0.7581, "step": 21336 }, { "epoch": 0.7518521454936266, "grad_norm": 2.07684326171875, "learning_rate": 3.0816468883144755e-06, "loss": 2.763, "step": 21337 }, { "epoch": 0.7518873825065849, "grad_norm": 1.6264634132385254, "learning_rate": 3.0808228683723105e-06, "loss": 1.1362, "step": 21338 }, { "epoch": 0.7519226195195433, "grad_norm": 3.0777781009674072, "learning_rate": 3.079998938553317e-06, "loss": 2.9792, "step": 21339 }, { "epoch": 0.7519578565325017, "grad_norm": 1.4179097414016724, "learning_rate": 3.0791750988682145e-06, "loss": 0.9735, "step": 21340 }, { "epoch": 0.7519930935454602, "grad_norm": 3.1237196922302246, "learning_rate": 3.0783513493277383e-06, "loss": 2.6841, "step": 21341 }, { "epoch": 0.7520283305584186, "grad_norm": 4.118913650512695, "learning_rate": 3.077527689942618e-06, "loss": 2.8213, "step": 21342 }, { "epoch": 0.7520635675713769, "grad_norm": 4.426970481872559, "learning_rate": 3.0767041207235803e-06, "loss": 5.2062, "step": 21343 }, { "epoch": 0.7520988045843354, "grad_norm": 4.116177082061768, "learning_rate": 3.0758806416813538e-06, "loss": 2.959, "step": 21344 }, { "epoch": 0.7521340415972938, "grad_norm": 1.5063310861587524, "learning_rate": 3.075057252826663e-06, "loss": 0.8459, "step": 21345 }, { "epoch": 0.7521692786102522, "grad_norm": 3.1809821128845215, "learning_rate": 3.074233954170235e-06, "loss": 2.5621, "step": 21346 }, { "epoch": 0.7522045156232107, "grad_norm": 5.63734769821167, "learning_rate": 3.0734107457227914e-06, "loss": 8.738, "step": 21347 }, { "epoch": 0.752239752636169, "grad_norm": 1.149742841720581, "learning_rate": 3.072587627495055e-06, "loss": 1.043, "step": 21348 }, { "epoch": 0.7522749896491274, "grad_norm": 1.433846354484558, "learning_rate": 3.0717645994977475e-06, "loss": 0.9189, "step": 21349 }, { "epoch": 0.7523102266620858, "grad_norm": 3.317251443862915, "learning_rate": 3.0709416617415887e-06, "loss": 2.7785, "step": 21350 }, { "epoch": 0.7523454636750443, "grad_norm": 3.4686975479125977, "learning_rate": 3.0701188142372974e-06, "loss": 2.8372, "step": 21351 }, { "epoch": 0.7523807006880027, "grad_norm": 0.7458406090736389, "learning_rate": 3.0692960569955933e-06, "loss": 0.8091, "step": 21352 }, { "epoch": 0.752415937700961, "grad_norm": 2.1220357418060303, "learning_rate": 3.0684733900271857e-06, "loss": 0.9469, "step": 21353 }, { "epoch": 0.7524511747139195, "grad_norm": 82.62684631347656, "learning_rate": 3.0676508133427985e-06, "loss": 6.0873, "step": 21354 }, { "epoch": 0.7524864117268779, "grad_norm": 8.146010398864746, "learning_rate": 3.066828326953145e-06, "loss": 5.0928, "step": 21355 }, { "epoch": 0.7525216487398363, "grad_norm": 3.454608678817749, "learning_rate": 3.0660059308689326e-06, "loss": 3.0145, "step": 21356 }, { "epoch": 0.7525568857527948, "grad_norm": 1.3908644914627075, "learning_rate": 3.065183625100873e-06, "loss": 0.8668, "step": 21357 }, { "epoch": 0.7525921227657532, "grad_norm": 1.237438678741455, "learning_rate": 3.064361409659686e-06, "loss": 0.861, "step": 21358 }, { "epoch": 0.7526273597787115, "grad_norm": 0.936567485332489, "learning_rate": 3.063539284556073e-06, "loss": 0.6944, "step": 21359 }, { "epoch": 0.75266259679167, "grad_norm": 5.4516754150390625, "learning_rate": 3.0627172498007427e-06, "loss": 4.6056, "step": 21360 }, { "epoch": 0.7526978338046284, "grad_norm": 2.6940598487854004, "learning_rate": 3.0618953054044054e-06, "loss": 2.4478, "step": 21361 }, { "epoch": 0.7527330708175868, "grad_norm": 1.1033943891525269, "learning_rate": 3.061073451377765e-06, "loss": 0.7851, "step": 21362 }, { "epoch": 0.7527683078305452, "grad_norm": 3.5024993419647217, "learning_rate": 3.060251687731527e-06, "loss": 3.1784, "step": 21363 }, { "epoch": 0.7528035448435036, "grad_norm": 3.2617101669311523, "learning_rate": 3.0594300144763943e-06, "loss": 2.8735, "step": 21364 }, { "epoch": 0.752838781856462, "grad_norm": 1.0134648084640503, "learning_rate": 3.0586084316230703e-06, "loss": 1.1608, "step": 21365 }, { "epoch": 0.7528740188694204, "grad_norm": 1.898226022720337, "learning_rate": 3.0577869391822556e-06, "loss": 0.9454, "step": 21366 }, { "epoch": 0.7529092558823789, "grad_norm": 0.7827395796775818, "learning_rate": 3.05696553716465e-06, "loss": 0.8634, "step": 21367 }, { "epoch": 0.7529444928953373, "grad_norm": 2.334141969680786, "learning_rate": 3.0561442255809524e-06, "loss": 2.5199, "step": 21368 }, { "epoch": 0.7529797299082956, "grad_norm": 2.0787558555603027, "learning_rate": 3.0553230044418626e-06, "loss": 0.9073, "step": 21369 }, { "epoch": 0.7530149669212541, "grad_norm": 4.090106964111328, "learning_rate": 3.0545018737580734e-06, "loss": 2.856, "step": 21370 }, { "epoch": 0.7530502039342125, "grad_norm": 3.9182474613189697, "learning_rate": 3.053680833540286e-06, "loss": 2.9436, "step": 21371 }, { "epoch": 0.7530854409471709, "grad_norm": 3.8584346771240234, "learning_rate": 3.052859883799183e-06, "loss": 4.3777, "step": 21372 }, { "epoch": 0.7531206779601293, "grad_norm": 2.724977731704712, "learning_rate": 3.052039024545469e-06, "loss": 2.9928, "step": 21373 }, { "epoch": 0.7531559149730878, "grad_norm": 4.4171013832092285, "learning_rate": 3.0512182557898353e-06, "loss": 2.4172, "step": 21374 }, { "epoch": 0.7531911519860461, "grad_norm": 4.890232563018799, "learning_rate": 3.050397577542965e-06, "loss": 3.2241, "step": 21375 }, { "epoch": 0.7532263889990045, "grad_norm": 8.815336227416992, "learning_rate": 3.049576989815549e-06, "loss": 7.4313, "step": 21376 }, { "epoch": 0.753261626011963, "grad_norm": 0.8729561567306519, "learning_rate": 3.0487564926182846e-06, "loss": 0.6842, "step": 21377 }, { "epoch": 0.7532968630249214, "grad_norm": 0.804428219795227, "learning_rate": 3.047936085961848e-06, "loss": 0.9322, "step": 21378 }, { "epoch": 0.7533321000378798, "grad_norm": 3.200190782546997, "learning_rate": 3.0471157698569296e-06, "loss": 0.8155, "step": 21379 }, { "epoch": 0.7533673370508382, "grad_norm": 4.5620856285095215, "learning_rate": 3.0462955443142152e-06, "loss": 2.8164, "step": 21380 }, { "epoch": 0.7534025740637966, "grad_norm": 1.0360668897628784, "learning_rate": 3.045475409344386e-06, "loss": 0.8513, "step": 21381 }, { "epoch": 0.753437811076755, "grad_norm": 1.1932624578475952, "learning_rate": 3.0446553649581257e-06, "loss": 0.8996, "step": 21382 }, { "epoch": 0.7534730480897134, "grad_norm": 2.81821346282959, "learning_rate": 3.043835411166116e-06, "loss": 2.774, "step": 21383 }, { "epoch": 0.7535082851026719, "grad_norm": 1.4190491437911987, "learning_rate": 3.0430155479790347e-06, "loss": 0.8527, "step": 21384 }, { "epoch": 0.7535435221156302, "grad_norm": 0.8443043828010559, "learning_rate": 3.0421957754075638e-06, "loss": 0.8166, "step": 21385 }, { "epoch": 0.7535787591285886, "grad_norm": 1.1504268646240234, "learning_rate": 3.041376093462378e-06, "loss": 0.8809, "step": 21386 }, { "epoch": 0.7536139961415471, "grad_norm": 1.4666807651519775, "learning_rate": 3.040556502154155e-06, "loss": 0.8244, "step": 21387 }, { "epoch": 0.7536492331545055, "grad_norm": 3.63051176071167, "learning_rate": 3.03973700149357e-06, "loss": 2.8929, "step": 21388 }, { "epoch": 0.7536844701674639, "grad_norm": 2.2556581497192383, "learning_rate": 3.038917591491297e-06, "loss": 1.09, "step": 21389 }, { "epoch": 0.7537197071804224, "grad_norm": 3.8073110580444336, "learning_rate": 3.0380982721580134e-06, "loss": 2.6328, "step": 21390 }, { "epoch": 0.7537549441933807, "grad_norm": 5.569900035858154, "learning_rate": 3.0372790435043798e-06, "loss": 5.2419, "step": 21391 }, { "epoch": 0.7537901812063391, "grad_norm": 1.343185305595398, "learning_rate": 3.036459905541076e-06, "loss": 1.0156, "step": 21392 }, { "epoch": 0.7538254182192976, "grad_norm": 1.0651662349700928, "learning_rate": 3.035640858278772e-06, "loss": 1.3826, "step": 21393 }, { "epoch": 0.753860655232256, "grad_norm": 1.5470257997512817, "learning_rate": 3.034821901728131e-06, "loss": 0.8733, "step": 21394 }, { "epoch": 0.7538958922452144, "grad_norm": 1.7184745073318481, "learning_rate": 3.0340030358998185e-06, "loss": 0.794, "step": 21395 }, { "epoch": 0.7539311292581727, "grad_norm": 2.399712085723877, "learning_rate": 3.033184260804509e-06, "loss": 0.7632, "step": 21396 }, { "epoch": 0.7539663662711312, "grad_norm": 1.7784091234207153, "learning_rate": 3.0323655764528603e-06, "loss": 0.7079, "step": 21397 }, { "epoch": 0.7540016032840896, "grad_norm": 4.303527355194092, "learning_rate": 3.031546982855537e-06, "loss": 2.8693, "step": 21398 }, { "epoch": 0.754036840297048, "grad_norm": 4.419039726257324, "learning_rate": 3.0307284800232007e-06, "loss": 2.9806, "step": 21399 }, { "epoch": 0.7540720773100065, "grad_norm": 3.0952723026275635, "learning_rate": 3.029910067966515e-06, "loss": 2.7721, "step": 21400 }, { "epoch": 0.7541073143229649, "grad_norm": 4.819399356842041, "learning_rate": 3.0290917466961377e-06, "loss": 3.0341, "step": 21401 }, { "epoch": 0.7541425513359232, "grad_norm": 1.039050817489624, "learning_rate": 3.0282735162227294e-06, "loss": 0.8948, "step": 21402 }, { "epoch": 0.7541777883488817, "grad_norm": 6.114231586456299, "learning_rate": 3.027455376556946e-06, "loss": 3.5538, "step": 21403 }, { "epoch": 0.7542130253618401, "grad_norm": 1.721160888671875, "learning_rate": 3.0266373277094442e-06, "loss": 0.7836, "step": 21404 }, { "epoch": 0.7542482623747985, "grad_norm": 1.4085289239883423, "learning_rate": 3.02581936969088e-06, "loss": 0.755, "step": 21405 }, { "epoch": 0.7542834993877569, "grad_norm": 2.7775187492370605, "learning_rate": 3.025001502511906e-06, "loss": 3.0011, "step": 21406 }, { "epoch": 0.7543187364007153, "grad_norm": 1.0227473974227905, "learning_rate": 3.024183726183176e-06, "loss": 0.8006, "step": 21407 }, { "epoch": 0.7543539734136737, "grad_norm": 4.386433124542236, "learning_rate": 3.0233660407153432e-06, "loss": 2.6102, "step": 21408 }, { "epoch": 0.7543892104266321, "grad_norm": 3.137075424194336, "learning_rate": 3.022548446119058e-06, "loss": 2.6363, "step": 21409 }, { "epoch": 0.7544244474395906, "grad_norm": 3.5956943035125732, "learning_rate": 3.0217309424049622e-06, "loss": 4.1828, "step": 21410 }, { "epoch": 0.754459684452549, "grad_norm": 5.712434768676758, "learning_rate": 3.0209135295837133e-06, "loss": 6.6291, "step": 21411 }, { "epoch": 0.7544949214655073, "grad_norm": 4.9336838722229, "learning_rate": 3.0200962076659568e-06, "loss": 5.1477, "step": 21412 }, { "epoch": 0.7545301584784658, "grad_norm": 2.9586074352264404, "learning_rate": 3.0192789766623345e-06, "loss": 2.618, "step": 21413 }, { "epoch": 0.7545653954914242, "grad_norm": 2.274372100830078, "learning_rate": 3.0184618365834884e-06, "loss": 2.6023, "step": 21414 }, { "epoch": 0.7546006325043826, "grad_norm": 4.311603546142578, "learning_rate": 3.0176447874400726e-06, "loss": 5.127, "step": 21415 }, { "epoch": 0.754635869517341, "grad_norm": 2.9817659854888916, "learning_rate": 3.01682782924272e-06, "loss": 2.7701, "step": 21416 }, { "epoch": 0.7546711065302995, "grad_norm": 0.8876693844795227, "learning_rate": 3.016010962002075e-06, "loss": 0.7091, "step": 21417 }, { "epoch": 0.7547063435432578, "grad_norm": 1.6578679084777832, "learning_rate": 3.015194185728776e-06, "loss": 0.9565, "step": 21418 }, { "epoch": 0.7547415805562162, "grad_norm": 1.2777451276779175, "learning_rate": 3.014377500433463e-06, "loss": 0.7598, "step": 21419 }, { "epoch": 0.7547768175691747, "grad_norm": 2.4644951820373535, "learning_rate": 3.013560906126772e-06, "loss": 2.9159, "step": 21420 }, { "epoch": 0.7548120545821331, "grad_norm": 6.018304824829102, "learning_rate": 3.012744402819341e-06, "loss": 3.2797, "step": 21421 }, { "epoch": 0.7548472915950915, "grad_norm": 1.0319724082946777, "learning_rate": 3.0119279905218045e-06, "loss": 1.0125, "step": 21422 }, { "epoch": 0.7548825286080499, "grad_norm": 1.0544365644454956, "learning_rate": 3.011111669244795e-06, "loss": 0.6551, "step": 21423 }, { "epoch": 0.7549177656210083, "grad_norm": 2.2859840393066406, "learning_rate": 3.010295438998947e-06, "loss": 2.8626, "step": 21424 }, { "epoch": 0.7549530026339667, "grad_norm": 1.0658491849899292, "learning_rate": 3.0094792997948906e-06, "loss": 1.0439, "step": 21425 }, { "epoch": 0.7549882396469252, "grad_norm": 1.4403704404830933, "learning_rate": 3.008663251643258e-06, "loss": 0.786, "step": 21426 }, { "epoch": 0.7550234766598836, "grad_norm": 3.9202768802642822, "learning_rate": 3.0078472945546754e-06, "loss": 3.9699, "step": 21427 }, { "epoch": 0.7550587136728419, "grad_norm": 5.190011978149414, "learning_rate": 3.007031428539777e-06, "loss": 4.4994, "step": 21428 }, { "epoch": 0.7550939506858003, "grad_norm": 3.7241909503936768, "learning_rate": 3.006215653609178e-06, "loss": 2.8369, "step": 21429 }, { "epoch": 0.7551291876987588, "grad_norm": 3.0753047466278076, "learning_rate": 3.005399969773515e-06, "loss": 2.6862, "step": 21430 }, { "epoch": 0.7551644247117172, "grad_norm": 4.311507701873779, "learning_rate": 3.0045843770434104e-06, "loss": 2.8882, "step": 21431 }, { "epoch": 0.7551996617246756, "grad_norm": 4.163470268249512, "learning_rate": 3.0037688754294823e-06, "loss": 2.754, "step": 21432 }, { "epoch": 0.755234898737634, "grad_norm": 1.0580087900161743, "learning_rate": 3.002953464942353e-06, "loss": 1.1182, "step": 21433 }, { "epoch": 0.7552701357505924, "grad_norm": 1.329647183418274, "learning_rate": 3.0021381455926522e-06, "loss": 0.9421, "step": 21434 }, { "epoch": 0.7553053727635508, "grad_norm": 6.1915059089660645, "learning_rate": 3.001322917390991e-06, "loss": 5.0453, "step": 21435 }, { "epoch": 0.7553406097765093, "grad_norm": 3.6170575618743896, "learning_rate": 3.0005077803479897e-06, "loss": 2.9126, "step": 21436 }, { "epoch": 0.7553758467894677, "grad_norm": 1.996851921081543, "learning_rate": 2.9996927344742664e-06, "loss": 0.6788, "step": 21437 }, { "epoch": 0.755411083802426, "grad_norm": 1.320921540260315, "learning_rate": 2.998877779780437e-06, "loss": 0.8481, "step": 21438 }, { "epoch": 0.7554463208153844, "grad_norm": 0.9753356575965881, "learning_rate": 2.9980629162771156e-06, "loss": 0.8658, "step": 21439 }, { "epoch": 0.7554815578283429, "grad_norm": 4.537599086761475, "learning_rate": 2.997248143974917e-06, "loss": 3.0096, "step": 21440 }, { "epoch": 0.7555167948413013, "grad_norm": 3.043362617492676, "learning_rate": 2.996433462884454e-06, "loss": 2.6738, "step": 21441 }, { "epoch": 0.7555520318542597, "grad_norm": 1.3240764141082764, "learning_rate": 2.9956188730163362e-06, "loss": 0.7168, "step": 21442 }, { "epoch": 0.7555872688672182, "grad_norm": 3.1093103885650635, "learning_rate": 2.994804374381175e-06, "loss": 2.9317, "step": 21443 }, { "epoch": 0.7556225058801765, "grad_norm": 3.6814510822296143, "learning_rate": 2.99398996698958e-06, "loss": 2.9382, "step": 21444 }, { "epoch": 0.7556577428931349, "grad_norm": 1.312394618988037, "learning_rate": 2.993175650852157e-06, "loss": 0.9102, "step": 21445 }, { "epoch": 0.7556929799060934, "grad_norm": 2.8288745880126953, "learning_rate": 2.992361425979514e-06, "loss": 3.0979, "step": 21446 }, { "epoch": 0.7557282169190518, "grad_norm": 11.094372749328613, "learning_rate": 2.9915472923822597e-06, "loss": 4.9617, "step": 21447 }, { "epoch": 0.7557634539320102, "grad_norm": 1.0947070121765137, "learning_rate": 2.9907332500709917e-06, "loss": 0.8278, "step": 21448 }, { "epoch": 0.7557986909449685, "grad_norm": 1.2546180486679077, "learning_rate": 2.9899192990563118e-06, "loss": 0.8273, "step": 21449 }, { "epoch": 0.755833927957927, "grad_norm": 6.407475471496582, "learning_rate": 2.989105439348833e-06, "loss": 2.9153, "step": 21450 }, { "epoch": 0.7558691649708854, "grad_norm": 12.509045600891113, "learning_rate": 2.9882916709591457e-06, "loss": 3.0489, "step": 21451 }, { "epoch": 0.7559044019838438, "grad_norm": 4.113334655761719, "learning_rate": 2.9874779938978527e-06, "loss": 4.7215, "step": 21452 }, { "epoch": 0.7559396389968023, "grad_norm": 4.018625736236572, "learning_rate": 2.986664408175551e-06, "loss": 4.0641, "step": 21453 }, { "epoch": 0.7559748760097607, "grad_norm": 2.2428629398345947, "learning_rate": 2.98585091380284e-06, "loss": 0.9928, "step": 21454 }, { "epoch": 0.756010113022719, "grad_norm": 3.5082569122314453, "learning_rate": 2.985037510790314e-06, "loss": 2.9321, "step": 21455 }, { "epoch": 0.7560453500356775, "grad_norm": 3.8567440509796143, "learning_rate": 2.984224199148568e-06, "loss": 2.7894, "step": 21456 }, { "epoch": 0.7560805870486359, "grad_norm": 5.745720863342285, "learning_rate": 2.983410978888195e-06, "loss": 7.9479, "step": 21457 }, { "epoch": 0.7561158240615943, "grad_norm": 3.398733615875244, "learning_rate": 2.982597850019787e-06, "loss": 2.4835, "step": 21458 }, { "epoch": 0.7561510610745527, "grad_norm": 2.183764934539795, "learning_rate": 2.9817848125539375e-06, "loss": 0.7594, "step": 21459 }, { "epoch": 0.7561862980875111, "grad_norm": 3.7886838912963867, "learning_rate": 2.9809718665012344e-06, "loss": 2.4829, "step": 21460 }, { "epoch": 0.7562215351004695, "grad_norm": 3.8345866203308105, "learning_rate": 2.980159011872267e-06, "loss": 3.0556, "step": 21461 }, { "epoch": 0.7562567721134279, "grad_norm": 6.170181751251221, "learning_rate": 2.9793462486776213e-06, "loss": 4.6765, "step": 21462 }, { "epoch": 0.7562920091263864, "grad_norm": 1.091619610786438, "learning_rate": 2.9785335769278888e-06, "loss": 0.8384, "step": 21463 }, { "epoch": 0.7563272461393448, "grad_norm": 1.185126543045044, "learning_rate": 2.9777209966336453e-06, "loss": 0.9321, "step": 21464 }, { "epoch": 0.7563624831523031, "grad_norm": 2.1914305686950684, "learning_rate": 2.9769085078054837e-06, "loss": 0.7729, "step": 21465 }, { "epoch": 0.7563977201652616, "grad_norm": 2.87921404838562, "learning_rate": 2.9760961104539854e-06, "loss": 2.1431, "step": 21466 }, { "epoch": 0.75643295717822, "grad_norm": 5.317502498626709, "learning_rate": 2.975283804589727e-06, "loss": 3.0482, "step": 21467 }, { "epoch": 0.7564681941911784, "grad_norm": 2.814755439758301, "learning_rate": 2.974471590223289e-06, "loss": 2.5367, "step": 21468 }, { "epoch": 0.7565034312041369, "grad_norm": 1.1530671119689941, "learning_rate": 2.9736594673652597e-06, "loss": 0.9142, "step": 21469 }, { "epoch": 0.7565386682170953, "grad_norm": 1.2571241855621338, "learning_rate": 2.9728474360262083e-06, "loss": 1.0515, "step": 21470 }, { "epoch": 0.7565739052300536, "grad_norm": 3.2997677326202393, "learning_rate": 2.972035496216713e-06, "loss": 2.0988, "step": 21471 }, { "epoch": 0.756609142243012, "grad_norm": 4.8437700271606445, "learning_rate": 2.9712236479473512e-06, "loss": 4.9181, "step": 21472 }, { "epoch": 0.7566443792559705, "grad_norm": 3.5273492336273193, "learning_rate": 2.9704118912286972e-06, "loss": 3.1507, "step": 21473 }, { "epoch": 0.7566796162689289, "grad_norm": 1.0334289073944092, "learning_rate": 2.9696002260713232e-06, "loss": 1.0961, "step": 21474 }, { "epoch": 0.7567148532818873, "grad_norm": 1.0050384998321533, "learning_rate": 2.9687886524858024e-06, "loss": 0.955, "step": 21475 }, { "epoch": 0.7567500902948457, "grad_norm": 3.54897141456604, "learning_rate": 2.9679771704827044e-06, "loss": 2.8605, "step": 21476 }, { "epoch": 0.7567853273078041, "grad_norm": 3.319322347640991, "learning_rate": 2.9671657800725993e-06, "loss": 2.8931, "step": 21477 }, { "epoch": 0.7568205643207625, "grad_norm": 2.72760272026062, "learning_rate": 2.966354481266056e-06, "loss": 3.2015, "step": 21478 }, { "epoch": 0.756855801333721, "grad_norm": 5.099769115447998, "learning_rate": 2.9655432740736413e-06, "loss": 3.2111, "step": 21479 }, { "epoch": 0.7568910383466794, "grad_norm": 0.9838951826095581, "learning_rate": 2.964732158505922e-06, "loss": 0.8071, "step": 21480 }, { "epoch": 0.7569262753596377, "grad_norm": 6.883358478546143, "learning_rate": 2.963921134573462e-06, "loss": 2.9255, "step": 21481 }, { "epoch": 0.7569615123725961, "grad_norm": 3.5206496715545654, "learning_rate": 2.9631102022868287e-06, "loss": 2.5965, "step": 21482 }, { "epoch": 0.7569967493855546, "grad_norm": 2.990962266921997, "learning_rate": 2.9622993616565756e-06, "loss": 2.7721, "step": 21483 }, { "epoch": 0.757031986398513, "grad_norm": 6.418529987335205, "learning_rate": 2.9614886126932716e-06, "loss": 3.3, "step": 21484 }, { "epoch": 0.7570672234114714, "grad_norm": 1.7007391452789307, "learning_rate": 2.9606779554074792e-06, "loss": 0.9165, "step": 21485 }, { "epoch": 0.7571024604244299, "grad_norm": 2.949397087097168, "learning_rate": 2.959867389809751e-06, "loss": 2.7455, "step": 21486 }, { "epoch": 0.7571376974373882, "grad_norm": 1.3308199644088745, "learning_rate": 2.959056915910642e-06, "loss": 0.808, "step": 21487 }, { "epoch": 0.7571729344503466, "grad_norm": 5.33384370803833, "learning_rate": 2.95824653372072e-06, "loss": 4.8253, "step": 21488 }, { "epoch": 0.7572081714633051, "grad_norm": 4.912447452545166, "learning_rate": 2.957436243250531e-06, "loss": 2.9037, "step": 21489 }, { "epoch": 0.7572434084762635, "grad_norm": 2.3692634105682373, "learning_rate": 2.956626044510632e-06, "loss": 0.9508, "step": 21490 }, { "epoch": 0.7572786454892219, "grad_norm": 5.829580307006836, "learning_rate": 2.955815937511576e-06, "loss": 2.625, "step": 21491 }, { "epoch": 0.7573138825021802, "grad_norm": 1.0151327848434448, "learning_rate": 2.955005922263915e-06, "loss": 1.0964, "step": 21492 }, { "epoch": 0.7573491195151387, "grad_norm": 1.8624792098999023, "learning_rate": 2.954195998778199e-06, "loss": 0.8066, "step": 21493 }, { "epoch": 0.7573843565280971, "grad_norm": 4.269789218902588, "learning_rate": 2.953386167064978e-06, "loss": 2.9606, "step": 21494 }, { "epoch": 0.7574195935410555, "grad_norm": 2.6024863719940186, "learning_rate": 2.952576427134799e-06, "loss": 2.8853, "step": 21495 }, { "epoch": 0.757454830554014, "grad_norm": 3.894084930419922, "learning_rate": 2.9517667789982095e-06, "loss": 2.8692, "step": 21496 }, { "epoch": 0.7574900675669723, "grad_norm": 1.06206214427948, "learning_rate": 2.9509572226657567e-06, "loss": 0.9776, "step": 21497 }, { "epoch": 0.7575253045799307, "grad_norm": 1.3737900257110596, "learning_rate": 2.950147758147983e-06, "loss": 0.9845, "step": 21498 }, { "epoch": 0.7575605415928892, "grad_norm": 3.5390090942382812, "learning_rate": 2.9493383854554336e-06, "loss": 3.2647, "step": 21499 }, { "epoch": 0.7575957786058476, "grad_norm": 1.0972907543182373, "learning_rate": 2.9485291045986485e-06, "loss": 0.9771, "step": 21500 }, { "epoch": 0.757631015618806, "grad_norm": 6.9987263679504395, "learning_rate": 2.9477199155881743e-06, "loss": 5.3566, "step": 21501 }, { "epoch": 0.7576662526317645, "grad_norm": 4.108140468597412, "learning_rate": 2.9469108184345408e-06, "loss": 3.1591, "step": 21502 }, { "epoch": 0.7577014896447228, "grad_norm": 3.503392457962036, "learning_rate": 2.9461018131482945e-06, "loss": 2.6586, "step": 21503 }, { "epoch": 0.7577367266576812, "grad_norm": 1.172965168952942, "learning_rate": 2.9452928997399732e-06, "loss": 0.7993, "step": 21504 }, { "epoch": 0.7577719636706396, "grad_norm": 1.4298961162567139, "learning_rate": 2.944484078220108e-06, "loss": 0.807, "step": 21505 }, { "epoch": 0.7578072006835981, "grad_norm": 3.421962022781372, "learning_rate": 2.9436753485992333e-06, "loss": 3.0074, "step": 21506 }, { "epoch": 0.7578424376965565, "grad_norm": 1.0603952407836914, "learning_rate": 2.942866710887893e-06, "loss": 0.7214, "step": 21507 }, { "epoch": 0.7578776747095148, "grad_norm": 6.982495307922363, "learning_rate": 2.9420581650966083e-06, "loss": 3.0619, "step": 21508 }, { "epoch": 0.7579129117224733, "grad_norm": 3.775834321975708, "learning_rate": 2.9412497112359164e-06, "loss": 3.0354, "step": 21509 }, { "epoch": 0.7579481487354317, "grad_norm": 1.4936376810073853, "learning_rate": 2.9404413493163454e-06, "loss": 0.6657, "step": 21510 }, { "epoch": 0.7579833857483901, "grad_norm": 1.6042544841766357, "learning_rate": 2.939633079348424e-06, "loss": 0.8514, "step": 21511 }, { "epoch": 0.7580186227613486, "grad_norm": 3.7549304962158203, "learning_rate": 2.938824901342683e-06, "loss": 2.5161, "step": 21512 }, { "epoch": 0.758053859774307, "grad_norm": 1.0278433561325073, "learning_rate": 2.9380168153096454e-06, "loss": 0.9121, "step": 21513 }, { "epoch": 0.7580890967872653, "grad_norm": 1.0788440704345703, "learning_rate": 2.9372088212598394e-06, "loss": 0.6906, "step": 21514 }, { "epoch": 0.7581243338002237, "grad_norm": 8.068950653076172, "learning_rate": 2.936400919203788e-06, "loss": 5.4416, "step": 21515 }, { "epoch": 0.7581595708131822, "grad_norm": 1.1537494659423828, "learning_rate": 2.935593109152014e-06, "loss": 1.0145, "step": 21516 }, { "epoch": 0.7581948078261406, "grad_norm": 1.0278470516204834, "learning_rate": 2.9347853911150405e-06, "loss": 0.9027, "step": 21517 }, { "epoch": 0.758230044839099, "grad_norm": 3.347040891647339, "learning_rate": 2.933977765103386e-06, "loss": 2.6393, "step": 21518 }, { "epoch": 0.7582652818520574, "grad_norm": 1.2052183151245117, "learning_rate": 2.9331702311275723e-06, "loss": 1.0524, "step": 21519 }, { "epoch": 0.7583005188650158, "grad_norm": 1.1638460159301758, "learning_rate": 2.93236278919812e-06, "loss": 0.9854, "step": 21520 }, { "epoch": 0.7583357558779742, "grad_norm": 7.624972343444824, "learning_rate": 2.9315554393255354e-06, "loss": 5.5143, "step": 21521 }, { "epoch": 0.7583709928909327, "grad_norm": 4.929596424102783, "learning_rate": 2.930748181520344e-06, "loss": 4.1465, "step": 21522 }, { "epoch": 0.7584062299038911, "grad_norm": 2.42327880859375, "learning_rate": 2.929941015793063e-06, "loss": 2.9571, "step": 21523 }, { "epoch": 0.7584414669168494, "grad_norm": 5.2822957038879395, "learning_rate": 2.9291339421541966e-06, "loss": 4.3522, "step": 21524 }, { "epoch": 0.7584767039298078, "grad_norm": 2.247986316680908, "learning_rate": 2.9283269606142585e-06, "loss": 0.7381, "step": 21525 }, { "epoch": 0.7585119409427663, "grad_norm": 3.9352169036865234, "learning_rate": 2.927520071183768e-06, "loss": 3.0769, "step": 21526 }, { "epoch": 0.7585471779557247, "grad_norm": 1.4404306411743164, "learning_rate": 2.9267132738732275e-06, "loss": 0.6525, "step": 21527 }, { "epoch": 0.7585824149686831, "grad_norm": 3.50622820854187, "learning_rate": 2.9259065686931465e-06, "loss": 2.9449, "step": 21528 }, { "epoch": 0.7586176519816415, "grad_norm": 0.9429997205734253, "learning_rate": 2.9250999556540315e-06, "loss": 0.8001, "step": 21529 }, { "epoch": 0.7586528889945999, "grad_norm": 0.9963518381118774, "learning_rate": 2.9242934347663965e-06, "loss": 0.7168, "step": 21530 }, { "epoch": 0.7586881260075583, "grad_norm": 1.2298569679260254, "learning_rate": 2.9234870060407373e-06, "loss": 0.9807, "step": 21531 }, { "epoch": 0.7587233630205168, "grad_norm": 4.740841865539551, "learning_rate": 2.922680669487562e-06, "loss": 2.6914, "step": 21532 }, { "epoch": 0.7587586000334752, "grad_norm": 0.9892904758453369, "learning_rate": 2.9218744251173715e-06, "loss": 0.9452, "step": 21533 }, { "epoch": 0.7587938370464336, "grad_norm": 4.360891819000244, "learning_rate": 2.921068272940669e-06, "loss": 3.4091, "step": 21534 }, { "epoch": 0.758829074059392, "grad_norm": 1.0009400844573975, "learning_rate": 2.9202622129679536e-06, "loss": 0.5925, "step": 21535 }, { "epoch": 0.7588643110723504, "grad_norm": 4.411675930023193, "learning_rate": 2.919456245209725e-06, "loss": 2.8358, "step": 21536 }, { "epoch": 0.7588995480853088, "grad_norm": 1.3465354442596436, "learning_rate": 2.918650369676481e-06, "loss": 0.8489, "step": 21537 }, { "epoch": 0.7589347850982672, "grad_norm": 5.231393337249756, "learning_rate": 2.9178445863787173e-06, "loss": 3.3304, "step": 21538 }, { "epoch": 0.7589700221112257, "grad_norm": 1.285725474357605, "learning_rate": 2.9170388953269335e-06, "loss": 0.929, "step": 21539 }, { "epoch": 0.759005259124184, "grad_norm": 3.014662742614746, "learning_rate": 2.916233296531614e-06, "loss": 2.8899, "step": 21540 }, { "epoch": 0.7590404961371424, "grad_norm": 1.4317783117294312, "learning_rate": 2.9154277900032614e-06, "loss": 1.0756, "step": 21541 }, { "epoch": 0.7590757331501009, "grad_norm": 5.312386512756348, "learning_rate": 2.9146223757523672e-06, "loss": 2.9914, "step": 21542 }, { "epoch": 0.7591109701630593, "grad_norm": 2.88550066947937, "learning_rate": 2.913817053789416e-06, "loss": 2.5838, "step": 21543 }, { "epoch": 0.7591462071760177, "grad_norm": 1.206696629524231, "learning_rate": 2.913011824124897e-06, "loss": 0.6242, "step": 21544 }, { "epoch": 0.7591814441889762, "grad_norm": 1.0742391347885132, "learning_rate": 2.9122066867693077e-06, "loss": 0.9046, "step": 21545 }, { "epoch": 0.7592166812019345, "grad_norm": 1.181404709815979, "learning_rate": 2.911401641733127e-06, "loss": 0.6647, "step": 21546 }, { "epoch": 0.7592519182148929, "grad_norm": 9.873516082763672, "learning_rate": 2.9105966890268424e-06, "loss": 3.1641, "step": 21547 }, { "epoch": 0.7592871552278513, "grad_norm": 1.1535942554473877, "learning_rate": 2.909791828660935e-06, "loss": 1.0815, "step": 21548 }, { "epoch": 0.7593223922408098, "grad_norm": 9.567000389099121, "learning_rate": 2.908987060645899e-06, "loss": 3.3463, "step": 21549 }, { "epoch": 0.7593576292537682, "grad_norm": 1.1017097234725952, "learning_rate": 2.9081823849922076e-06, "loss": 0.9901, "step": 21550 }, { "epoch": 0.7593928662667265, "grad_norm": 1.0737518072128296, "learning_rate": 2.9073778017103427e-06, "loss": 0.6837, "step": 21551 }, { "epoch": 0.759428103279685, "grad_norm": 5.477869987487793, "learning_rate": 2.9065733108107865e-06, "loss": 2.5053, "step": 21552 }, { "epoch": 0.7594633402926434, "grad_norm": 1.2007631063461304, "learning_rate": 2.9057689123040155e-06, "loss": 0.8297, "step": 21553 }, { "epoch": 0.7594985773056018, "grad_norm": 2.909614086151123, "learning_rate": 2.9049646062005087e-06, "loss": 3.0612, "step": 21554 }, { "epoch": 0.7595338143185603, "grad_norm": 2.427203893661499, "learning_rate": 2.904160392510741e-06, "loss": 2.7463, "step": 21555 }, { "epoch": 0.7595690513315186, "grad_norm": 4.9493536949157715, "learning_rate": 2.9033562712451892e-06, "loss": 2.8695, "step": 21556 }, { "epoch": 0.759604288344477, "grad_norm": 1.3295477628707886, "learning_rate": 2.902552242414325e-06, "loss": 0.8152, "step": 21557 }, { "epoch": 0.7596395253574354, "grad_norm": 4.700571537017822, "learning_rate": 2.9017483060286257e-06, "loss": 3.4598, "step": 21558 }, { "epoch": 0.7596747623703939, "grad_norm": 2.1264350414276123, "learning_rate": 2.9009444620985516e-06, "loss": 0.7824, "step": 21559 }, { "epoch": 0.7597099993833523, "grad_norm": 5.350836277008057, "learning_rate": 2.9001407106345836e-06, "loss": 5.2079, "step": 21560 }, { "epoch": 0.7597452363963106, "grad_norm": 4.552835941314697, "learning_rate": 2.899337051647191e-06, "loss": 2.7267, "step": 21561 }, { "epoch": 0.7597804734092691, "grad_norm": 1.277372121810913, "learning_rate": 2.8985334851468337e-06, "loss": 0.9798, "step": 21562 }, { "epoch": 0.7598157104222275, "grad_norm": 4.328713417053223, "learning_rate": 2.89773001114398e-06, "loss": 2.8841, "step": 21563 }, { "epoch": 0.7598509474351859, "grad_norm": 4.90262508392334, "learning_rate": 2.896926629649104e-06, "loss": 3.0978, "step": 21564 }, { "epoch": 0.7598861844481444, "grad_norm": 3.787191152572632, "learning_rate": 2.89612334067266e-06, "loss": 3.6015, "step": 21565 }, { "epoch": 0.7599214214611028, "grad_norm": 5.436776638031006, "learning_rate": 2.8953201442251145e-06, "loss": 5.2681, "step": 21566 }, { "epoch": 0.7599566584740611, "grad_norm": 5.28593111038208, "learning_rate": 2.8945170403169254e-06, "loss": 3.6233, "step": 21567 }, { "epoch": 0.7599918954870196, "grad_norm": 1.0837053060531616, "learning_rate": 2.8937140289585643e-06, "loss": 0.8375, "step": 21568 }, { "epoch": 0.760027132499978, "grad_norm": 1.0666749477386475, "learning_rate": 2.892911110160479e-06, "loss": 0.69, "step": 21569 }, { "epoch": 0.7600623695129364, "grad_norm": 3.9435555934906006, "learning_rate": 2.892108283933133e-06, "loss": 5.2407, "step": 21570 }, { "epoch": 0.7600976065258948, "grad_norm": 4.4268999099731445, "learning_rate": 2.89130555028698e-06, "loss": 2.8887, "step": 21571 }, { "epoch": 0.7601328435388532, "grad_norm": 3.6818604469299316, "learning_rate": 2.8905029092324787e-06, "loss": 3.3751, "step": 21572 }, { "epoch": 0.7601680805518116, "grad_norm": 8.333561897277832, "learning_rate": 2.889700360780083e-06, "loss": 6.6118, "step": 21573 }, { "epoch": 0.76020331756477, "grad_norm": 8.329902648925781, "learning_rate": 2.8888979049402455e-06, "loss": 3.227, "step": 21574 }, { "epoch": 0.7602385545777285, "grad_norm": 1.7433470487594604, "learning_rate": 2.8880955417234192e-06, "loss": 0.7696, "step": 21575 }, { "epoch": 0.7602737915906869, "grad_norm": 4.438653469085693, "learning_rate": 2.8872932711400536e-06, "loss": 2.9719, "step": 21576 }, { "epoch": 0.7603090286036452, "grad_norm": 5.036303520202637, "learning_rate": 2.886491093200603e-06, "loss": 3.4383, "step": 21577 }, { "epoch": 0.7603442656166037, "grad_norm": 5.344481468200684, "learning_rate": 2.885689007915505e-06, "loss": 4.9842, "step": 21578 }, { "epoch": 0.7603795026295621, "grad_norm": 3.2966244220733643, "learning_rate": 2.8848870152952178e-06, "loss": 2.8777, "step": 21579 }, { "epoch": 0.7604147396425205, "grad_norm": 0.8391402363777161, "learning_rate": 2.884085115350186e-06, "loss": 0.9114, "step": 21580 }, { "epoch": 0.7604499766554789, "grad_norm": 4.373669147491455, "learning_rate": 2.8832833080908485e-06, "loss": 3.618, "step": 21581 }, { "epoch": 0.7604852136684374, "grad_norm": 3.863074779510498, "learning_rate": 2.8824815935276507e-06, "loss": 2.5507, "step": 21582 }, { "epoch": 0.7605204506813957, "grad_norm": 1.4278488159179688, "learning_rate": 2.881679971671042e-06, "loss": 1.1853, "step": 21583 }, { "epoch": 0.7605556876943541, "grad_norm": 2.0342869758605957, "learning_rate": 2.880878442531455e-06, "loss": 0.9913, "step": 21584 }, { "epoch": 0.7605909247073126, "grad_norm": 0.7677744626998901, "learning_rate": 2.880077006119334e-06, "loss": 0.9238, "step": 21585 }, { "epoch": 0.760626161720271, "grad_norm": 1.0689672231674194, "learning_rate": 2.8792756624451134e-06, "loss": 0.8806, "step": 21586 }, { "epoch": 0.7606613987332294, "grad_norm": 3.3312911987304688, "learning_rate": 2.878474411519241e-06, "loss": 4.2997, "step": 21587 }, { "epoch": 0.7606966357461878, "grad_norm": 1.2896031141281128, "learning_rate": 2.8776732533521424e-06, "loss": 0.9055, "step": 21588 }, { "epoch": 0.7607318727591462, "grad_norm": 3.9718146324157715, "learning_rate": 2.8768721879542584e-06, "loss": 2.664, "step": 21589 }, { "epoch": 0.7607671097721046, "grad_norm": 1.6591802835464478, "learning_rate": 2.8760712153360203e-06, "loss": 0.9982, "step": 21590 }, { "epoch": 0.760802346785063, "grad_norm": 0.7418540716171265, "learning_rate": 2.8752703355078626e-06, "loss": 0.7385, "step": 21591 }, { "epoch": 0.7608375837980215, "grad_norm": 1.5398399829864502, "learning_rate": 2.8744695484802154e-06, "loss": 0.9585, "step": 21592 }, { "epoch": 0.7608728208109798, "grad_norm": 2.617349624633789, "learning_rate": 2.8736688542635107e-06, "loss": 0.8632, "step": 21593 }, { "epoch": 0.7609080578239382, "grad_norm": 13.627673149108887, "learning_rate": 2.872868252868177e-06, "loss": 2.9294, "step": 21594 }, { "epoch": 0.7609432948368967, "grad_norm": 1.0716184377670288, "learning_rate": 2.872067744304642e-06, "loss": 0.7373, "step": 21595 }, { "epoch": 0.7609785318498551, "grad_norm": 4.028143405914307, "learning_rate": 2.871267328583336e-06, "loss": 5.0148, "step": 21596 }, { "epoch": 0.7610137688628135, "grad_norm": 1.0099693536758423, "learning_rate": 2.8704670057146742e-06, "loss": 1.0743, "step": 21597 }, { "epoch": 0.761049005875772, "grad_norm": 1.3223501443862915, "learning_rate": 2.869666775709091e-06, "loss": 1.0707, "step": 21598 }, { "epoch": 0.7610842428887303, "grad_norm": 1.6322757005691528, "learning_rate": 2.8688666385770104e-06, "loss": 0.88, "step": 21599 }, { "epoch": 0.7611194799016887, "grad_norm": 1.2957875728607178, "learning_rate": 2.8680665943288464e-06, "loss": 1.0571, "step": 21600 }, { "epoch": 0.7611547169146472, "grad_norm": 0.8538879752159119, "learning_rate": 2.867266642975021e-06, "loss": 0.8539, "step": 21601 }, { "epoch": 0.7611899539276056, "grad_norm": 3.1515769958496094, "learning_rate": 2.8664667845259585e-06, "loss": 2.9756, "step": 21602 }, { "epoch": 0.761225190940564, "grad_norm": 1.0032899379730225, "learning_rate": 2.8656670189920775e-06, "loss": 0.9426, "step": 21603 }, { "epoch": 0.7612604279535223, "grad_norm": 7.782149314880371, "learning_rate": 2.8648673463837904e-06, "loss": 4.8626, "step": 21604 }, { "epoch": 0.7612956649664808, "grad_norm": 1.325487732887268, "learning_rate": 2.8640677667115115e-06, "loss": 0.8879, "step": 21605 }, { "epoch": 0.7613309019794392, "grad_norm": 5.336071014404297, "learning_rate": 2.8632682799856647e-06, "loss": 3.4724, "step": 21606 }, { "epoch": 0.7613661389923976, "grad_norm": 2.930361270904541, "learning_rate": 2.862468886216655e-06, "loss": 2.7221, "step": 21607 }, { "epoch": 0.7614013760053561, "grad_norm": 1.4199872016906738, "learning_rate": 2.861669585414897e-06, "loss": 0.7388, "step": 21608 }, { "epoch": 0.7614366130183144, "grad_norm": 1.0786126852035522, "learning_rate": 2.860870377590801e-06, "loss": 0.7588, "step": 21609 }, { "epoch": 0.7614718500312728, "grad_norm": 1.3743246793746948, "learning_rate": 2.860071262754779e-06, "loss": 1.0091, "step": 21610 }, { "epoch": 0.7615070870442313, "grad_norm": 49.58292770385742, "learning_rate": 2.859272240917238e-06, "loss": 3.0833, "step": 21611 }, { "epoch": 0.7615423240571897, "grad_norm": 6.641088008880615, "learning_rate": 2.8584733120885857e-06, "loss": 6.0316, "step": 21612 }, { "epoch": 0.7615775610701481, "grad_norm": 1.6078770160675049, "learning_rate": 2.8576744762792273e-06, "loss": 1.0126, "step": 21613 }, { "epoch": 0.7616127980831064, "grad_norm": 3.5342352390289307, "learning_rate": 2.85687573349957e-06, "loss": 2.3507, "step": 21614 }, { "epoch": 0.7616480350960649, "grad_norm": 1.0502104759216309, "learning_rate": 2.856077083760018e-06, "loss": 1.0256, "step": 21615 }, { "epoch": 0.7616832721090233, "grad_norm": 5.939503192901611, "learning_rate": 2.855278527070967e-06, "loss": 3.5845, "step": 21616 }, { "epoch": 0.7617185091219817, "grad_norm": 1.2750657796859741, "learning_rate": 2.854480063442826e-06, "loss": 0.7637, "step": 21617 }, { "epoch": 0.7617537461349402, "grad_norm": 1.1256622076034546, "learning_rate": 2.8536816928859943e-06, "loss": 0.7237, "step": 21618 }, { "epoch": 0.7617889831478986, "grad_norm": 5.2613043785095215, "learning_rate": 2.8528834154108676e-06, "loss": 2.5285, "step": 21619 }, { "epoch": 0.7618242201608569, "grad_norm": 3.6764464378356934, "learning_rate": 2.8520852310278404e-06, "loss": 3.3349, "step": 21620 }, { "epoch": 0.7618594571738154, "grad_norm": 3.3530850410461426, "learning_rate": 2.8512871397473174e-06, "loss": 3.0229, "step": 21621 }, { "epoch": 0.7618946941867738, "grad_norm": 4.536489486694336, "learning_rate": 2.8504891415796932e-06, "loss": 3.8845, "step": 21622 }, { "epoch": 0.7619299311997322, "grad_norm": 1.7007951736450195, "learning_rate": 2.8496912365353544e-06, "loss": 0.7862, "step": 21623 }, { "epoch": 0.7619651682126906, "grad_norm": 4.477169513702393, "learning_rate": 2.848893424624696e-06, "loss": 4.8863, "step": 21624 }, { "epoch": 0.762000405225649, "grad_norm": 6.369141101837158, "learning_rate": 2.848095705858117e-06, "loss": 5.3644, "step": 21625 }, { "epoch": 0.7620356422386074, "grad_norm": 1.6179763078689575, "learning_rate": 2.8472980802460005e-06, "loss": 0.8039, "step": 21626 }, { "epoch": 0.7620708792515658, "grad_norm": 3.206033706665039, "learning_rate": 2.8465005477987374e-06, "loss": 3.0266, "step": 21627 }, { "epoch": 0.7621061162645243, "grad_norm": 5.487853527069092, "learning_rate": 2.845703108526715e-06, "loss": 3.0078, "step": 21628 }, { "epoch": 0.7621413532774827, "grad_norm": 0.9222224950790405, "learning_rate": 2.8449057624403208e-06, "loss": 0.8542, "step": 21629 }, { "epoch": 0.762176590290441, "grad_norm": 6.896393299102783, "learning_rate": 2.844108509549941e-06, "loss": 5.0852, "step": 21630 }, { "epoch": 0.7622118273033995, "grad_norm": 1.6367595195770264, "learning_rate": 2.8433113498659603e-06, "loss": 0.9228, "step": 21631 }, { "epoch": 0.7622470643163579, "grad_norm": 10.159076690673828, "learning_rate": 2.8425142833987595e-06, "loss": 6.6984, "step": 21632 }, { "epoch": 0.7622823013293163, "grad_norm": 1.0465221405029297, "learning_rate": 2.8417173101587227e-06, "loss": 0.8162, "step": 21633 }, { "epoch": 0.7623175383422748, "grad_norm": 0.9490911960601807, "learning_rate": 2.8409204301562322e-06, "loss": 0.9628, "step": 21634 }, { "epoch": 0.7623527753552332, "grad_norm": 3.488509178161621, "learning_rate": 2.840123643401659e-06, "loss": 3.0662, "step": 21635 }, { "epoch": 0.7623880123681915, "grad_norm": 3.994396686553955, "learning_rate": 2.839326949905391e-06, "loss": 3.2743, "step": 21636 }, { "epoch": 0.7624232493811499, "grad_norm": 1.197326898574829, "learning_rate": 2.8385303496778037e-06, "loss": 0.9025, "step": 21637 }, { "epoch": 0.7624584863941084, "grad_norm": 3.2599380016326904, "learning_rate": 2.837733842729269e-06, "loss": 2.7605, "step": 21638 }, { "epoch": 0.7624937234070668, "grad_norm": 2.684433937072754, "learning_rate": 2.836937429070159e-06, "loss": 2.6089, "step": 21639 }, { "epoch": 0.7625289604200252, "grad_norm": 1.6229389905929565, "learning_rate": 2.8361411087108558e-06, "loss": 0.7578, "step": 21640 }, { "epoch": 0.7625641974329836, "grad_norm": 5.821047306060791, "learning_rate": 2.835344881661729e-06, "loss": 4.9785, "step": 21641 }, { "epoch": 0.762599434445942, "grad_norm": 1.458918571472168, "learning_rate": 2.8345487479331457e-06, "loss": 0.9123, "step": 21642 }, { "epoch": 0.7626346714589004, "grad_norm": 1.1447184085845947, "learning_rate": 2.8337527075354743e-06, "loss": 0.995, "step": 21643 }, { "epoch": 0.7626699084718589, "grad_norm": 1.2388252019882202, "learning_rate": 2.832956760479092e-06, "loss": 1.0911, "step": 21644 }, { "epoch": 0.7627051454848173, "grad_norm": 1.2281091213226318, "learning_rate": 2.8321609067743584e-06, "loss": 0.7195, "step": 21645 }, { "epoch": 0.7627403824977756, "grad_norm": 2.816102981567383, "learning_rate": 2.8313651464316426e-06, "loss": 2.5642, "step": 21646 }, { "epoch": 0.762775619510734, "grad_norm": 1.012041687965393, "learning_rate": 2.830569479461308e-06, "loss": 0.8564, "step": 21647 }, { "epoch": 0.7628108565236925, "grad_norm": 1.0750535726547241, "learning_rate": 2.8297739058737197e-06, "loss": 0.9967, "step": 21648 }, { "epoch": 0.7628460935366509, "grad_norm": 4.069238662719727, "learning_rate": 2.8289784256792394e-06, "loss": 4.8415, "step": 21649 }, { "epoch": 0.7628813305496093, "grad_norm": 3.8239405155181885, "learning_rate": 2.828183038888228e-06, "loss": 4.9784, "step": 21650 }, { "epoch": 0.7629165675625678, "grad_norm": 4.88222074508667, "learning_rate": 2.8273877455110477e-06, "loss": 2.9564, "step": 21651 }, { "epoch": 0.7629518045755261, "grad_norm": 1.156990647315979, "learning_rate": 2.8265925455580547e-06, "loss": 0.8558, "step": 21652 }, { "epoch": 0.7629870415884845, "grad_norm": 3.6598238945007324, "learning_rate": 2.8257974390396105e-06, "loss": 3.2252, "step": 21653 }, { "epoch": 0.763022278601443, "grad_norm": 5.170620441436768, "learning_rate": 2.8250024259660625e-06, "loss": 3.4075, "step": 21654 }, { "epoch": 0.7630575156144014, "grad_norm": 4.540202617645264, "learning_rate": 2.8242075063477747e-06, "loss": 2.7215, "step": 21655 }, { "epoch": 0.7630927526273598, "grad_norm": 2.5524532794952393, "learning_rate": 2.823412680195101e-06, "loss": 3.1466, "step": 21656 }, { "epoch": 0.7631279896403181, "grad_norm": 1.8149915933609009, "learning_rate": 2.822617947518389e-06, "loss": 0.9628, "step": 21657 }, { "epoch": 0.7631632266532766, "grad_norm": 4.043614864349365, "learning_rate": 2.821823308327989e-06, "loss": 2.9497, "step": 21658 }, { "epoch": 0.763198463666235, "grad_norm": 1.021010398864746, "learning_rate": 2.821028762634258e-06, "loss": 0.9388, "step": 21659 }, { "epoch": 0.7632337006791934, "grad_norm": 4.813673973083496, "learning_rate": 2.820234310447545e-06, "loss": 3.0645, "step": 21660 }, { "epoch": 0.7632689376921519, "grad_norm": 1.6857441663742065, "learning_rate": 2.819439951778191e-06, "loss": 1.1615, "step": 21661 }, { "epoch": 0.7633041747051102, "grad_norm": 4.478763103485107, "learning_rate": 2.818645686636543e-06, "loss": 0.7697, "step": 21662 }, { "epoch": 0.7633394117180686, "grad_norm": 0.9541675448417664, "learning_rate": 2.8178515150329565e-06, "loss": 0.927, "step": 21663 }, { "epoch": 0.7633746487310271, "grad_norm": 7.586774826049805, "learning_rate": 2.817057436977766e-06, "loss": 6.3324, "step": 21664 }, { "epoch": 0.7634098857439855, "grad_norm": 1.2732019424438477, "learning_rate": 2.816263452481317e-06, "loss": 1.123, "step": 21665 }, { "epoch": 0.7634451227569439, "grad_norm": 4.548957347869873, "learning_rate": 2.815469561553952e-06, "loss": 4.1794, "step": 21666 }, { "epoch": 0.7634803597699023, "grad_norm": 1.0216256380081177, "learning_rate": 2.814675764206011e-06, "loss": 0.686, "step": 21667 }, { "epoch": 0.7635155967828607, "grad_norm": 2.623434066772461, "learning_rate": 2.8138820604478323e-06, "loss": 2.0794, "step": 21668 }, { "epoch": 0.7635508337958191, "grad_norm": 0.9813080430030823, "learning_rate": 2.813088450289756e-06, "loss": 0.7151, "step": 21669 }, { "epoch": 0.7635860708087775, "grad_norm": 5.1636810302734375, "learning_rate": 2.812294933742118e-06, "loss": 2.715, "step": 21670 }, { "epoch": 0.763621307821736, "grad_norm": 6.123357772827148, "learning_rate": 2.8115015108152543e-06, "loss": 5.3193, "step": 21671 }, { "epoch": 0.7636565448346944, "grad_norm": 1.3473618030548096, "learning_rate": 2.8107081815195026e-06, "loss": 0.8751, "step": 21672 }, { "epoch": 0.7636917818476527, "grad_norm": 1.7143183946609497, "learning_rate": 2.809914945865185e-06, "loss": 0.9087, "step": 21673 }, { "epoch": 0.7637270188606112, "grad_norm": 4.824372291564941, "learning_rate": 2.8091218038626454e-06, "loss": 3.2209, "step": 21674 }, { "epoch": 0.7637622558735696, "grad_norm": 4.058988571166992, "learning_rate": 2.8083287555222107e-06, "loss": 2.8538, "step": 21675 }, { "epoch": 0.763797492886528, "grad_norm": 2.832760810852051, "learning_rate": 2.807535800854212e-06, "loss": 3.3211, "step": 21676 }, { "epoch": 0.7638327298994865, "grad_norm": 2.9567036628723145, "learning_rate": 2.8067429398689696e-06, "loss": 0.9215, "step": 21677 }, { "epoch": 0.7638679669124449, "grad_norm": 1.2490345239639282, "learning_rate": 2.80595017257682e-06, "loss": 0.655, "step": 21678 }, { "epoch": 0.7639032039254032, "grad_norm": 3.3679590225219727, "learning_rate": 2.8051574989880894e-06, "loss": 3.2654, "step": 21679 }, { "epoch": 0.7639384409383616, "grad_norm": 5.7804741859436035, "learning_rate": 2.804364919113095e-06, "loss": 4.902, "step": 21680 }, { "epoch": 0.7639736779513201, "grad_norm": 6.643808841705322, "learning_rate": 2.8035724329621618e-06, "loss": 2.6509, "step": 21681 }, { "epoch": 0.7640089149642785, "grad_norm": 3.615729331970215, "learning_rate": 2.8027800405456198e-06, "loss": 4.1998, "step": 21682 }, { "epoch": 0.7640441519772369, "grad_norm": 1.805247187614441, "learning_rate": 2.8019877418737817e-06, "loss": 0.9681, "step": 21683 }, { "epoch": 0.7640793889901953, "grad_norm": 1.2608346939086914, "learning_rate": 2.80119553695697e-06, "loss": 0.7825, "step": 21684 }, { "epoch": 0.7641146260031537, "grad_norm": 2.8151581287384033, "learning_rate": 2.8004034258055037e-06, "loss": 2.7554, "step": 21685 }, { "epoch": 0.7641498630161121, "grad_norm": 4.8592352867126465, "learning_rate": 2.7996114084296987e-06, "loss": 5.1981, "step": 21686 }, { "epoch": 0.7641851000290706, "grad_norm": 4.742523193359375, "learning_rate": 2.7988194848398744e-06, "loss": 2.8929, "step": 21687 }, { "epoch": 0.764220337042029, "grad_norm": 2.7208914756774902, "learning_rate": 2.7980276550463414e-06, "loss": 3.0676, "step": 21688 }, { "epoch": 0.7642555740549873, "grad_norm": 4.40386962890625, "learning_rate": 2.797235919059417e-06, "loss": 3.3811, "step": 21689 }, { "epoch": 0.7642908110679457, "grad_norm": 3.3190181255340576, "learning_rate": 2.796444276889412e-06, "loss": 2.8733, "step": 21690 }, { "epoch": 0.7643260480809042, "grad_norm": 4.711496829986572, "learning_rate": 2.7956527285466406e-06, "loss": 4.7784, "step": 21691 }, { "epoch": 0.7643612850938626, "grad_norm": 13.168679237365723, "learning_rate": 2.794861274041405e-06, "loss": 3.6399, "step": 21692 }, { "epoch": 0.764396522106821, "grad_norm": 6.108905792236328, "learning_rate": 2.794069913384021e-06, "loss": 4.577, "step": 21693 }, { "epoch": 0.7644317591197795, "grad_norm": 3.7719364166259766, "learning_rate": 2.7932786465847948e-06, "loss": 0.98, "step": 21694 }, { "epoch": 0.7644669961327378, "grad_norm": 3.965521812438965, "learning_rate": 2.7924874736540354e-06, "loss": 2.8424, "step": 21695 }, { "epoch": 0.7645022331456962, "grad_norm": 2.103898048400879, "learning_rate": 2.791696394602038e-06, "loss": 2.6204, "step": 21696 }, { "epoch": 0.7645374701586547, "grad_norm": 0.9368506669998169, "learning_rate": 2.7909054094391166e-06, "loss": 0.7345, "step": 21697 }, { "epoch": 0.7645727071716131, "grad_norm": 4.525773525238037, "learning_rate": 2.7901145181755727e-06, "loss": 2.7282, "step": 21698 }, { "epoch": 0.7646079441845715, "grad_norm": 4.699651718139648, "learning_rate": 2.789323720821704e-06, "loss": 6.6274, "step": 21699 }, { "epoch": 0.7646431811975298, "grad_norm": 3.6594982147216797, "learning_rate": 2.788533017387809e-06, "loss": 3.0791, "step": 21700 }, { "epoch": 0.7646784182104883, "grad_norm": 1.699379563331604, "learning_rate": 2.7877424078841963e-06, "loss": 0.6954, "step": 21701 }, { "epoch": 0.7647136552234467, "grad_norm": 4.50679349899292, "learning_rate": 2.7869518923211535e-06, "loss": 6.8576, "step": 21702 }, { "epoch": 0.7647488922364051, "grad_norm": 4.8890790939331055, "learning_rate": 2.7861614707089813e-06, "loss": 4.9069, "step": 21703 }, { "epoch": 0.7647841292493636, "grad_norm": 4.159041881561279, "learning_rate": 2.7853711430579757e-06, "loss": 5.0518, "step": 21704 }, { "epoch": 0.7648193662623219, "grad_norm": 3.024353265762329, "learning_rate": 2.7845809093784294e-06, "loss": 2.5317, "step": 21705 }, { "epoch": 0.7648546032752803, "grad_norm": 2.0356929302215576, "learning_rate": 2.7837907696806352e-06, "loss": 0.9846, "step": 21706 }, { "epoch": 0.7648898402882388, "grad_norm": 2.452299118041992, "learning_rate": 2.7830007239748867e-06, "loss": 1.0113, "step": 21707 }, { "epoch": 0.7649250773011972, "grad_norm": 1.135362148284912, "learning_rate": 2.782210772271472e-06, "loss": 0.7369, "step": 21708 }, { "epoch": 0.7649603143141556, "grad_norm": 1.2073099613189697, "learning_rate": 2.7814209145806813e-06, "loss": 0.9067, "step": 21709 }, { "epoch": 0.764995551327114, "grad_norm": 2.67611026763916, "learning_rate": 2.780631150912807e-06, "loss": 2.4971, "step": 21710 }, { "epoch": 0.7650307883400724, "grad_norm": 2.077125072479248, "learning_rate": 2.779841481278125e-06, "loss": 0.9764, "step": 21711 }, { "epoch": 0.7650660253530308, "grad_norm": 1.0732797384262085, "learning_rate": 2.779051905686929e-06, "loss": 0.8367, "step": 21712 }, { "epoch": 0.7651012623659892, "grad_norm": 1.2858713865280151, "learning_rate": 2.7782624241495036e-06, "loss": 0.8077, "step": 21713 }, { "epoch": 0.7651364993789477, "grad_norm": 1.3537883758544922, "learning_rate": 2.7774730366761327e-06, "loss": 1.0284, "step": 21714 }, { "epoch": 0.765171736391906, "grad_norm": 0.9995283484458923, "learning_rate": 2.7766837432770886e-06, "loss": 0.6634, "step": 21715 }, { "epoch": 0.7652069734048644, "grad_norm": 1.0952562093734741, "learning_rate": 2.775894543962663e-06, "loss": 0.6911, "step": 21716 }, { "epoch": 0.7652422104178229, "grad_norm": 1.4193776845932007, "learning_rate": 2.7751054387431333e-06, "loss": 0.8959, "step": 21717 }, { "epoch": 0.7652774474307813, "grad_norm": 3.01796293258667, "learning_rate": 2.7743164276287727e-06, "loss": 0.927, "step": 21718 }, { "epoch": 0.7653126844437397, "grad_norm": 3.447443962097168, "learning_rate": 2.7735275106298578e-06, "loss": 3.1809, "step": 21719 }, { "epoch": 0.7653479214566982, "grad_norm": 6.706490993499756, "learning_rate": 2.772738687756674e-06, "loss": 4.9094, "step": 21720 }, { "epoch": 0.7653831584696565, "grad_norm": 3.4375593662261963, "learning_rate": 2.7719499590194866e-06, "loss": 2.8496, "step": 21721 }, { "epoch": 0.7654183954826149, "grad_norm": 3.458432674407959, "learning_rate": 2.7711613244285717e-06, "loss": 2.9035, "step": 21722 }, { "epoch": 0.7654536324955733, "grad_norm": 1.2456656694412231, "learning_rate": 2.770372783994201e-06, "loss": 0.8524, "step": 21723 }, { "epoch": 0.7654888695085318, "grad_norm": 1.1940373182296753, "learning_rate": 2.769584337726645e-06, "loss": 0.7335, "step": 21724 }, { "epoch": 0.7655241065214902, "grad_norm": 2.2374250888824463, "learning_rate": 2.7687959856361758e-06, "loss": 0.9183, "step": 21725 }, { "epoch": 0.7655593435344485, "grad_norm": 3.261699676513672, "learning_rate": 2.7680077277330585e-06, "loss": 3.5399, "step": 21726 }, { "epoch": 0.765594580547407, "grad_norm": 1.4240936040878296, "learning_rate": 2.7672195640275622e-06, "loss": 0.9071, "step": 21727 }, { "epoch": 0.7656298175603654, "grad_norm": 5.419236660003662, "learning_rate": 2.766431494529952e-06, "loss": 5.0893, "step": 21728 }, { "epoch": 0.7656650545733238, "grad_norm": 2.9137165546417236, "learning_rate": 2.765643519250497e-06, "loss": 2.7108, "step": 21729 }, { "epoch": 0.7657002915862823, "grad_norm": 4.850419521331787, "learning_rate": 2.7648556381994494e-06, "loss": 2.7881, "step": 21730 }, { "epoch": 0.7657355285992407, "grad_norm": 1.5060442686080933, "learning_rate": 2.7640678513870834e-06, "loss": 1.2058, "step": 21731 }, { "epoch": 0.765770765612199, "grad_norm": 5.594270706176758, "learning_rate": 2.7632801588236537e-06, "loss": 4.5136, "step": 21732 }, { "epoch": 0.7658060026251574, "grad_norm": 1.1783366203308105, "learning_rate": 2.762492560519425e-06, "loss": 0.9283, "step": 21733 }, { "epoch": 0.7658412396381159, "grad_norm": 5.068961143493652, "learning_rate": 2.761705056484647e-06, "loss": 0.7469, "step": 21734 }, { "epoch": 0.7658764766510743, "grad_norm": 1.001997470855713, "learning_rate": 2.7609176467295863e-06, "loss": 0.9579, "step": 21735 }, { "epoch": 0.7659117136640327, "grad_norm": 1.3455880880355835, "learning_rate": 2.760130331264498e-06, "loss": 0.9619, "step": 21736 }, { "epoch": 0.7659469506769911, "grad_norm": 4.429168701171875, "learning_rate": 2.7593431100996304e-06, "loss": 2.5456, "step": 21737 }, { "epoch": 0.7659821876899495, "grad_norm": 2.237971782684326, "learning_rate": 2.758555983245239e-06, "loss": 0.7287, "step": 21738 }, { "epoch": 0.7660174247029079, "grad_norm": 2.717928647994995, "learning_rate": 2.7577689507115847e-06, "loss": 2.7072, "step": 21739 }, { "epoch": 0.7660526617158664, "grad_norm": 0.9159141182899475, "learning_rate": 2.756982012508909e-06, "loss": 0.8075, "step": 21740 }, { "epoch": 0.7660878987288248, "grad_norm": 6.206621170043945, "learning_rate": 2.756195168647464e-06, "loss": 5.851, "step": 21741 }, { "epoch": 0.7661231357417831, "grad_norm": 12.115177154541016, "learning_rate": 2.7554084191375006e-06, "loss": 2.755, "step": 21742 }, { "epoch": 0.7661583727547416, "grad_norm": 3.4138495922088623, "learning_rate": 2.754621763989266e-06, "loss": 3.6376, "step": 21743 }, { "epoch": 0.7661936097677, "grad_norm": 4.267955780029297, "learning_rate": 2.7538352032130043e-06, "loss": 4.7017, "step": 21744 }, { "epoch": 0.7662288467806584, "grad_norm": 2.9558069705963135, "learning_rate": 2.753048736818963e-06, "loss": 2.6241, "step": 21745 }, { "epoch": 0.7662640837936168, "grad_norm": 1.4396827220916748, "learning_rate": 2.7522623648173853e-06, "loss": 0.7336, "step": 21746 }, { "epoch": 0.7662993208065753, "grad_norm": 3.0685219764709473, "learning_rate": 2.751476087218513e-06, "loss": 2.9075, "step": 21747 }, { "epoch": 0.7663345578195336, "grad_norm": 6.216687202453613, "learning_rate": 2.7506899040325883e-06, "loss": 7.0263, "step": 21748 }, { "epoch": 0.766369794832492, "grad_norm": 0.8966318964958191, "learning_rate": 2.749903815269851e-06, "loss": 0.9248, "step": 21749 }, { "epoch": 0.7664050318454505, "grad_norm": 5.285506725311279, "learning_rate": 2.7491178209405402e-06, "loss": 6.4945, "step": 21750 }, { "epoch": 0.7664402688584089, "grad_norm": 4.3990631103515625, "learning_rate": 2.748331921054893e-06, "loss": 5.0366, "step": 21751 }, { "epoch": 0.7664755058713673, "grad_norm": 4.264256477355957, "learning_rate": 2.74754611562315e-06, "loss": 4.0271, "step": 21752 }, { "epoch": 0.7665107428843257, "grad_norm": 9.129383087158203, "learning_rate": 2.7467604046555354e-06, "loss": 5.413, "step": 21753 }, { "epoch": 0.7665459798972841, "grad_norm": 1.5627933740615845, "learning_rate": 2.7459747881622933e-06, "loss": 0.9578, "step": 21754 }, { "epoch": 0.7665812169102425, "grad_norm": 2.9847583770751953, "learning_rate": 2.7451892661536584e-06, "loss": 3.3062, "step": 21755 }, { "epoch": 0.7666164539232009, "grad_norm": 4.216163158416748, "learning_rate": 2.744403838639853e-06, "loss": 5.1043, "step": 21756 }, { "epoch": 0.7666516909361594, "grad_norm": 0.9789855480194092, "learning_rate": 2.743618505631113e-06, "loss": 1.0838, "step": 21757 }, { "epoch": 0.7666869279491177, "grad_norm": 3.5248570442199707, "learning_rate": 2.7428332671376643e-06, "loss": 2.4074, "step": 21758 }, { "epoch": 0.7667221649620761, "grad_norm": 1.506688117980957, "learning_rate": 2.742048123169738e-06, "loss": 1.0008, "step": 21759 }, { "epoch": 0.7667574019750346, "grad_norm": 1.5515376329421997, "learning_rate": 2.741263073737559e-06, "loss": 1.22, "step": 21760 }, { "epoch": 0.766792638987993, "grad_norm": 1.6271536350250244, "learning_rate": 2.740478118851354e-06, "loss": 0.6469, "step": 21761 }, { "epoch": 0.7668278760009514, "grad_norm": 5.689609527587891, "learning_rate": 2.7396932585213455e-06, "loss": 3.1925, "step": 21762 }, { "epoch": 0.7668631130139099, "grad_norm": 1.7284247875213623, "learning_rate": 2.738908492757757e-06, "loss": 0.771, "step": 21763 }, { "epoch": 0.7668983500268682, "grad_norm": 1.1513901948928833, "learning_rate": 2.7381238215708104e-06, "loss": 0.9272, "step": 21764 }, { "epoch": 0.7669335870398266, "grad_norm": 1.3259226083755493, "learning_rate": 2.737339244970726e-06, "loss": 0.9805, "step": 21765 }, { "epoch": 0.766968824052785, "grad_norm": 4.699222564697266, "learning_rate": 2.7365547629677235e-06, "loss": 3.2887, "step": 21766 }, { "epoch": 0.7670040610657435, "grad_norm": 3.2648191452026367, "learning_rate": 2.73577037557202e-06, "loss": 2.7474, "step": 21767 }, { "epoch": 0.7670392980787019, "grad_norm": 3.756296157836914, "learning_rate": 2.7349860827938357e-06, "loss": 2.9697, "step": 21768 }, { "epoch": 0.7670745350916602, "grad_norm": 0.970956563949585, "learning_rate": 2.734201884643377e-06, "loss": 0.87, "step": 21769 }, { "epoch": 0.7671097721046187, "grad_norm": 1.2143899202346802, "learning_rate": 2.7334177811308683e-06, "loss": 0.9721, "step": 21770 }, { "epoch": 0.7671450091175771, "grad_norm": 1.2921613454818726, "learning_rate": 2.7326337722665206e-06, "loss": 0.889, "step": 21771 }, { "epoch": 0.7671802461305355, "grad_norm": 1.0741360187530518, "learning_rate": 2.7318498580605413e-06, "loss": 1.0224, "step": 21772 }, { "epoch": 0.767215483143494, "grad_norm": 3.926734209060669, "learning_rate": 2.7310660385231392e-06, "loss": 3.2526, "step": 21773 }, { "epoch": 0.7672507201564523, "grad_norm": 3.8316738605499268, "learning_rate": 2.730282313664535e-06, "loss": 4.8803, "step": 21774 }, { "epoch": 0.7672859571694107, "grad_norm": 4.597710609436035, "learning_rate": 2.7294986834949267e-06, "loss": 4.9641, "step": 21775 }, { "epoch": 0.7673211941823692, "grad_norm": 4.09345006942749, "learning_rate": 2.7287151480245234e-06, "loss": 5.0234, "step": 21776 }, { "epoch": 0.7673564311953276, "grad_norm": 1.4614380598068237, "learning_rate": 2.7279317072635324e-06, "loss": 0.658, "step": 21777 }, { "epoch": 0.767391668208286, "grad_norm": 6.187455654144287, "learning_rate": 2.727148361222156e-06, "loss": 3.2672, "step": 21778 }, { "epoch": 0.7674269052212443, "grad_norm": 2.041334629058838, "learning_rate": 2.7263651099106005e-06, "loss": 0.9062, "step": 21779 }, { "epoch": 0.7674621422342028, "grad_norm": 3.3722774982452393, "learning_rate": 2.7255819533390638e-06, "loss": 2.6762, "step": 21780 }, { "epoch": 0.7674973792471612, "grad_norm": 1.285851001739502, "learning_rate": 2.72479889151775e-06, "loss": 0.9018, "step": 21781 }, { "epoch": 0.7675326162601196, "grad_norm": 1.3179584741592407, "learning_rate": 2.7240159244568576e-06, "loss": 1.0042, "step": 21782 }, { "epoch": 0.7675678532730781, "grad_norm": 5.427783966064453, "learning_rate": 2.7232330521665873e-06, "loss": 4.5704, "step": 21783 }, { "epoch": 0.7676030902860365, "grad_norm": 3.791907787322998, "learning_rate": 2.722450274657127e-06, "loss": 3.4175, "step": 21784 }, { "epoch": 0.7676383272989948, "grad_norm": 1.72042977809906, "learning_rate": 2.7216675919386815e-06, "loss": 0.9171, "step": 21785 }, { "epoch": 0.7676735643119533, "grad_norm": 6.944509506225586, "learning_rate": 2.7208850040214442e-06, "loss": 2.8715, "step": 21786 }, { "epoch": 0.7677088013249117, "grad_norm": 5.078197956085205, "learning_rate": 2.7201025109156087e-06, "loss": 4.4406, "step": 21787 }, { "epoch": 0.7677440383378701, "grad_norm": 2.5632143020629883, "learning_rate": 2.7193201126313594e-06, "loss": 2.6773, "step": 21788 }, { "epoch": 0.7677792753508285, "grad_norm": 1.0268996953964233, "learning_rate": 2.718537809178896e-06, "loss": 0.9337, "step": 21789 }, { "epoch": 0.767814512363787, "grad_norm": 4.730264186859131, "learning_rate": 2.717755600568409e-06, "loss": 3.5802, "step": 21790 }, { "epoch": 0.7678497493767453, "grad_norm": 1.0976660251617432, "learning_rate": 2.7169734868100796e-06, "loss": 0.7702, "step": 21791 }, { "epoch": 0.7678849863897037, "grad_norm": 1.2159061431884766, "learning_rate": 2.7161914679140955e-06, "loss": 0.8131, "step": 21792 }, { "epoch": 0.7679202234026622, "grad_norm": 3.1530416011810303, "learning_rate": 2.7154095438906504e-06, "loss": 4.966, "step": 21793 }, { "epoch": 0.7679554604156206, "grad_norm": 4.094188690185547, "learning_rate": 2.7146277147499212e-06, "loss": 4.9563, "step": 21794 }, { "epoch": 0.767990697428579, "grad_norm": 1.5095975399017334, "learning_rate": 2.7138459805020945e-06, "loss": 1.0905, "step": 21795 }, { "epoch": 0.7680259344415374, "grad_norm": 3.1628806591033936, "learning_rate": 2.713064341157352e-06, "loss": 2.3253, "step": 21796 }, { "epoch": 0.7680611714544958, "grad_norm": 7.994472026824951, "learning_rate": 2.7122827967258735e-06, "loss": 3.1582, "step": 21797 }, { "epoch": 0.7680964084674542, "grad_norm": 12.308982849121094, "learning_rate": 2.711501347217841e-06, "loss": 5.3082, "step": 21798 }, { "epoch": 0.7681316454804126, "grad_norm": 3.697946548461914, "learning_rate": 2.7107199926434323e-06, "loss": 2.6517, "step": 21799 }, { "epoch": 0.7681668824933711, "grad_norm": 8.628327369689941, "learning_rate": 2.7099387330128235e-06, "loss": 7.4649, "step": 21800 }, { "epoch": 0.7682021195063294, "grad_norm": 1.2571642398834229, "learning_rate": 2.7091575683361917e-06, "loss": 0.8309, "step": 21801 }, { "epoch": 0.7682373565192878, "grad_norm": 3.8081164360046387, "learning_rate": 2.708376498623714e-06, "loss": 3.1091, "step": 21802 }, { "epoch": 0.7682725935322463, "grad_norm": 1.0461639165878296, "learning_rate": 2.7075955238855544e-06, "loss": 1.0337, "step": 21803 }, { "epoch": 0.7683078305452047, "grad_norm": 2.945920467376709, "learning_rate": 2.7068146441318956e-06, "loss": 2.7334, "step": 21804 }, { "epoch": 0.7683430675581631, "grad_norm": 1.4848233461380005, "learning_rate": 2.706033859372905e-06, "loss": 0.9063, "step": 21805 }, { "epoch": 0.7683783045711216, "grad_norm": 5.499495029449463, "learning_rate": 2.7052531696187555e-06, "loss": 2.8989, "step": 21806 }, { "epoch": 0.7684135415840799, "grad_norm": 3.705108642578125, "learning_rate": 2.7044725748796077e-06, "loss": 2.3525, "step": 21807 }, { "epoch": 0.7684487785970383, "grad_norm": 1.1250180006027222, "learning_rate": 2.703692075165636e-06, "loss": 1.1216, "step": 21808 }, { "epoch": 0.7684840156099968, "grad_norm": 4.649644374847412, "learning_rate": 2.7029116704870073e-06, "loss": 3.2388, "step": 21809 }, { "epoch": 0.7685192526229552, "grad_norm": 4.284852027893066, "learning_rate": 2.7021313608538814e-06, "loss": 2.7466, "step": 21810 }, { "epoch": 0.7685544896359136, "grad_norm": 4.1630706787109375, "learning_rate": 2.7013511462764196e-06, "loss": 3.9575, "step": 21811 }, { "epoch": 0.7685897266488719, "grad_norm": 1.0299056768417358, "learning_rate": 2.7005710267647955e-06, "loss": 0.938, "step": 21812 }, { "epoch": 0.7686249636618304, "grad_norm": 1.2756062746047974, "learning_rate": 2.699791002329162e-06, "loss": 0.7691, "step": 21813 }, { "epoch": 0.7686602006747888, "grad_norm": 6.147345542907715, "learning_rate": 2.6990110729796794e-06, "loss": 5.0667, "step": 21814 }, { "epoch": 0.7686954376877472, "grad_norm": 5.353527069091797, "learning_rate": 2.6982312387265064e-06, "loss": 4.4545, "step": 21815 }, { "epoch": 0.7687306747007057, "grad_norm": 4.27371883392334, "learning_rate": 2.697451499579803e-06, "loss": 4.6788, "step": 21816 }, { "epoch": 0.768765911713664, "grad_norm": 1.184363603591919, "learning_rate": 2.696671855549724e-06, "loss": 0.6645, "step": 21817 }, { "epoch": 0.7688011487266224, "grad_norm": 1.323756456375122, "learning_rate": 2.695892306646424e-06, "loss": 0.828, "step": 21818 }, { "epoch": 0.7688363857395809, "grad_norm": 3.784715175628662, "learning_rate": 2.6951128528800574e-06, "loss": 3.4889, "step": 21819 }, { "epoch": 0.7688716227525393, "grad_norm": 4.107000827789307, "learning_rate": 2.694333494260777e-06, "loss": 0.8708, "step": 21820 }, { "epoch": 0.7689068597654977, "grad_norm": 1.6504857540130615, "learning_rate": 2.693554230798732e-06, "loss": 0.8646, "step": 21821 }, { "epoch": 0.768942096778456, "grad_norm": 3.7358248233795166, "learning_rate": 2.692775062504075e-06, "loss": 4.5126, "step": 21822 }, { "epoch": 0.7689773337914145, "grad_norm": 3.4867513179779053, "learning_rate": 2.691995989386953e-06, "loss": 3.0453, "step": 21823 }, { "epoch": 0.7690125708043729, "grad_norm": 1.2222853899002075, "learning_rate": 2.6912170114575144e-06, "loss": 0.8221, "step": 21824 }, { "epoch": 0.7690478078173313, "grad_norm": 3.388728618621826, "learning_rate": 2.690438128725907e-06, "loss": 3.0028, "step": 21825 }, { "epoch": 0.7690830448302898, "grad_norm": 4.2171311378479, "learning_rate": 2.68965934120227e-06, "loss": 3.1988, "step": 21826 }, { "epoch": 0.7691182818432482, "grad_norm": 3.3677210807800293, "learning_rate": 2.6888806488967532e-06, "loss": 2.858, "step": 21827 }, { "epoch": 0.7691535188562065, "grad_norm": 1.3611037731170654, "learning_rate": 2.6881020518195002e-06, "loss": 0.9844, "step": 21828 }, { "epoch": 0.769188755869165, "grad_norm": 1.9914897680282593, "learning_rate": 2.687323549980647e-06, "loss": 0.655, "step": 21829 }, { "epoch": 0.7692239928821234, "grad_norm": 7.291301727294922, "learning_rate": 2.6865451433903323e-06, "loss": 2.9663, "step": 21830 }, { "epoch": 0.7692592298950818, "grad_norm": 1.5858956575393677, "learning_rate": 2.685766832058705e-06, "loss": 1.1, "step": 21831 }, { "epoch": 0.7692944669080402, "grad_norm": 3.023155450820923, "learning_rate": 2.684988615995894e-06, "loss": 2.8708, "step": 21832 }, { "epoch": 0.7693297039209986, "grad_norm": 0.9844511151313782, "learning_rate": 2.6842104952120375e-06, "loss": 0.6672, "step": 21833 }, { "epoch": 0.769364940933957, "grad_norm": 1.592711091041565, "learning_rate": 2.683432469717271e-06, "loss": 0.7867, "step": 21834 }, { "epoch": 0.7694001779469154, "grad_norm": 2.3936569690704346, "learning_rate": 2.6826545395217284e-06, "loss": 0.8596, "step": 21835 }, { "epoch": 0.7694354149598739, "grad_norm": 28.357219696044922, "learning_rate": 2.681876704635543e-06, "loss": 4.9924, "step": 21836 }, { "epoch": 0.7694706519728323, "grad_norm": 9.863436698913574, "learning_rate": 2.6810989650688456e-06, "loss": 2.9211, "step": 21837 }, { "epoch": 0.7695058889857906, "grad_norm": 1.0022594928741455, "learning_rate": 2.6803213208317667e-06, "loss": 0.779, "step": 21838 }, { "epoch": 0.7695411259987491, "grad_norm": 3.4200279712677, "learning_rate": 2.679543771934434e-06, "loss": 2.4821, "step": 21839 }, { "epoch": 0.7695763630117075, "grad_norm": 1.2730677127838135, "learning_rate": 2.6787663183869773e-06, "loss": 1.0334, "step": 21840 }, { "epoch": 0.7696116000246659, "grad_norm": 1.3611342906951904, "learning_rate": 2.6779889601995214e-06, "loss": 0.7934, "step": 21841 }, { "epoch": 0.7696468370376243, "grad_norm": 2.859483480453491, "learning_rate": 2.6772116973821914e-06, "loss": 2.7747, "step": 21842 }, { "epoch": 0.7696820740505828, "grad_norm": 1.7618889808654785, "learning_rate": 2.676434529945113e-06, "loss": 0.7598, "step": 21843 }, { "epoch": 0.7697173110635411, "grad_norm": 1.1060250997543335, "learning_rate": 2.675657457898411e-06, "loss": 1.1688, "step": 21844 }, { "epoch": 0.7697525480764995, "grad_norm": 1.2476881742477417, "learning_rate": 2.6748804812521954e-06, "loss": 0.8439, "step": 21845 }, { "epoch": 0.769787785089458, "grad_norm": 1.165860652923584, "learning_rate": 2.6741036000166e-06, "loss": 0.7125, "step": 21846 }, { "epoch": 0.7698230221024164, "grad_norm": 3.4228289127349854, "learning_rate": 2.67332681420174e-06, "loss": 2.4712, "step": 21847 }, { "epoch": 0.7698582591153748, "grad_norm": 4.472154140472412, "learning_rate": 2.6725501238177286e-06, "loss": 3.2816, "step": 21848 }, { "epoch": 0.7698934961283332, "grad_norm": 1.4098607301712036, "learning_rate": 2.6717735288746827e-06, "loss": 0.787, "step": 21849 }, { "epoch": 0.7699287331412916, "grad_norm": 4.911669731140137, "learning_rate": 2.670997029382726e-06, "loss": 4.2645, "step": 21850 }, { "epoch": 0.76996397015425, "grad_norm": 1.0730901956558228, "learning_rate": 2.6702206253519626e-06, "loss": 0.9089, "step": 21851 }, { "epoch": 0.7699992071672085, "grad_norm": 4.317840576171875, "learning_rate": 2.66944431679251e-06, "loss": 5.1868, "step": 21852 }, { "epoch": 0.7700344441801669, "grad_norm": 2.823087453842163, "learning_rate": 2.668668103714479e-06, "loss": 2.4132, "step": 21853 }, { "epoch": 0.7700696811931252, "grad_norm": 5.207018852233887, "learning_rate": 2.6678919861279793e-06, "loss": 2.5208, "step": 21854 }, { "epoch": 0.7701049182060836, "grad_norm": 4.569581508636475, "learning_rate": 2.667115964043121e-06, "loss": 2.8497, "step": 21855 }, { "epoch": 0.7701401552190421, "grad_norm": 2.9752397537231445, "learning_rate": 2.666340037470011e-06, "loss": 2.5708, "step": 21856 }, { "epoch": 0.7701753922320005, "grad_norm": 4.580031394958496, "learning_rate": 2.665564206418757e-06, "loss": 4.7873, "step": 21857 }, { "epoch": 0.7702106292449589, "grad_norm": 1.2572741508483887, "learning_rate": 2.664788470899463e-06, "loss": 0.915, "step": 21858 }, { "epoch": 0.7702458662579174, "grad_norm": 4.868519306182861, "learning_rate": 2.6640128309222334e-06, "loss": 2.931, "step": 21859 }, { "epoch": 0.7702811032708757, "grad_norm": 1.1601587533950806, "learning_rate": 2.663237286497171e-06, "loss": 0.7562, "step": 21860 }, { "epoch": 0.7703163402838341, "grad_norm": 5.140082359313965, "learning_rate": 2.6624618376343767e-06, "loss": 2.8914, "step": 21861 }, { "epoch": 0.7703515772967926, "grad_norm": 4.961682319641113, "learning_rate": 2.6616864843439525e-06, "loss": 4.9162, "step": 21862 }, { "epoch": 0.770386814309751, "grad_norm": 1.067758560180664, "learning_rate": 2.6609112266360003e-06, "loss": 1.1645, "step": 21863 }, { "epoch": 0.7704220513227094, "grad_norm": 0.9968632459640503, "learning_rate": 2.6601360645206077e-06, "loss": 0.7158, "step": 21864 }, { "epoch": 0.7704572883356677, "grad_norm": 3.201028823852539, "learning_rate": 2.6593609980078805e-06, "loss": 3.0628, "step": 21865 }, { "epoch": 0.7704925253486262, "grad_norm": 5.905301570892334, "learning_rate": 2.6585860271079146e-06, "loss": 5.3852, "step": 21866 }, { "epoch": 0.7705277623615846, "grad_norm": 4.0102009773254395, "learning_rate": 2.657811151830797e-06, "loss": 2.8357, "step": 21867 }, { "epoch": 0.770562999374543, "grad_norm": 1.352530837059021, "learning_rate": 2.657036372186622e-06, "loss": 0.9176, "step": 21868 }, { "epoch": 0.7705982363875015, "grad_norm": 3.178615093231201, "learning_rate": 2.65626168818549e-06, "loss": 2.5839, "step": 21869 }, { "epoch": 0.7706334734004598, "grad_norm": 4.957679748535156, "learning_rate": 2.6554870998374815e-06, "loss": 6.6318, "step": 21870 }, { "epoch": 0.7706687104134182, "grad_norm": 1.657213568687439, "learning_rate": 2.65471260715269e-06, "loss": 0.884, "step": 21871 }, { "epoch": 0.7707039474263767, "grad_norm": 8.914300918579102, "learning_rate": 2.6539382101412013e-06, "loss": 5.81, "step": 21872 }, { "epoch": 0.7707391844393351, "grad_norm": 0.9241706132888794, "learning_rate": 2.6531639088131036e-06, "loss": 0.6769, "step": 21873 }, { "epoch": 0.7707744214522935, "grad_norm": 1.0198357105255127, "learning_rate": 2.652389703178482e-06, "loss": 0.6758, "step": 21874 }, { "epoch": 0.7708096584652518, "grad_norm": 0.9378443956375122, "learning_rate": 2.65161559324742e-06, "loss": 0.7773, "step": 21875 }, { "epoch": 0.7708448954782103, "grad_norm": 5.33309268951416, "learning_rate": 2.65084157903e-06, "loss": 4.8019, "step": 21876 }, { "epoch": 0.7708801324911687, "grad_norm": 1.6314959526062012, "learning_rate": 2.6500676605363064e-06, "loss": 0.7616, "step": 21877 }, { "epoch": 0.7709153695041271, "grad_norm": 2.7977614402770996, "learning_rate": 2.6492938377764164e-06, "loss": 2.4004, "step": 21878 }, { "epoch": 0.7709506065170856, "grad_norm": 1.0917677879333496, "learning_rate": 2.648520110760412e-06, "loss": 0.9095, "step": 21879 }, { "epoch": 0.770985843530044, "grad_norm": 5.034469127655029, "learning_rate": 2.6477464794983686e-06, "loss": 3.1213, "step": 21880 }, { "epoch": 0.7710210805430023, "grad_norm": 2.263561248779297, "learning_rate": 2.6469729440003633e-06, "loss": 2.4915, "step": 21881 }, { "epoch": 0.7710563175559608, "grad_norm": 2.3416497707366943, "learning_rate": 2.6461995042764767e-06, "loss": 0.8353, "step": 21882 }, { "epoch": 0.7710915545689192, "grad_norm": 1.9421628713607788, "learning_rate": 2.6454261603367703e-06, "loss": 0.8597, "step": 21883 }, { "epoch": 0.7711267915818776, "grad_norm": 5.86774206161499, "learning_rate": 2.644652912191329e-06, "loss": 4.0661, "step": 21884 }, { "epoch": 0.7711620285948361, "grad_norm": 2.1661276817321777, "learning_rate": 2.6438797598502243e-06, "loss": 0.7772, "step": 21885 }, { "epoch": 0.7711972656077944, "grad_norm": 1.1146492958068848, "learning_rate": 2.643106703323518e-06, "loss": 0.9005, "step": 21886 }, { "epoch": 0.7712325026207528, "grad_norm": 3.1336188316345215, "learning_rate": 2.6423337426212825e-06, "loss": 2.9158, "step": 21887 }, { "epoch": 0.7712677396337112, "grad_norm": 5.795633792877197, "learning_rate": 2.6415608777535918e-06, "loss": 4.6532, "step": 21888 }, { "epoch": 0.7713029766466697, "grad_norm": 3.1497340202331543, "learning_rate": 2.6407881087305065e-06, "loss": 2.4027, "step": 21889 }, { "epoch": 0.7713382136596281, "grad_norm": 5.346787929534912, "learning_rate": 2.6400154355620922e-06, "loss": 2.8525, "step": 21890 }, { "epoch": 0.7713734506725864, "grad_norm": 5.932474136352539, "learning_rate": 2.6392428582584153e-06, "loss": 4.1504, "step": 21891 }, { "epoch": 0.7714086876855449, "grad_norm": 3.9532887935638428, "learning_rate": 2.6384703768295374e-06, "loss": 3.3563, "step": 21892 }, { "epoch": 0.7714439246985033, "grad_norm": 1.6165817975997925, "learning_rate": 2.6376979912855206e-06, "loss": 0.8566, "step": 21893 }, { "epoch": 0.7714791617114617, "grad_norm": 1.233717679977417, "learning_rate": 2.6369257016364246e-06, "loss": 0.8525, "step": 21894 }, { "epoch": 0.7715143987244202, "grad_norm": 5.173638820648193, "learning_rate": 2.6361535078923105e-06, "loss": 5.1835, "step": 21895 }, { "epoch": 0.7715496357373786, "grad_norm": 1.9588286876678467, "learning_rate": 2.6353814100632335e-06, "loss": 0.9821, "step": 21896 }, { "epoch": 0.7715848727503369, "grad_norm": 3.009174108505249, "learning_rate": 2.634609408159253e-06, "loss": 3.2591, "step": 21897 }, { "epoch": 0.7716201097632953, "grad_norm": 4.0758562088012695, "learning_rate": 2.6338375021904227e-06, "loss": 3.5084, "step": 21898 }, { "epoch": 0.7716553467762538, "grad_norm": 4.672928810119629, "learning_rate": 2.6330656921667976e-06, "loss": 5.0404, "step": 21899 }, { "epoch": 0.7716905837892122, "grad_norm": 2.310394048690796, "learning_rate": 2.632293978098429e-06, "loss": 2.0725, "step": 21900 }, { "epoch": 0.7717258208021706, "grad_norm": 0.8188763856887817, "learning_rate": 2.6315223599953743e-06, "loss": 0.5823, "step": 21901 }, { "epoch": 0.771761057815129, "grad_norm": 1.5019760131835938, "learning_rate": 2.6307508378676727e-06, "loss": 0.7976, "step": 21902 }, { "epoch": 0.7717962948280874, "grad_norm": 0.9458659291267395, "learning_rate": 2.629979411725384e-06, "loss": 0.8257, "step": 21903 }, { "epoch": 0.7718315318410458, "grad_norm": 0.9737187027931213, "learning_rate": 2.6292080815785537e-06, "loss": 0.6163, "step": 21904 }, { "epoch": 0.7718667688540043, "grad_norm": 1.7234463691711426, "learning_rate": 2.6284368474372257e-06, "loss": 0.7325, "step": 21905 }, { "epoch": 0.7719020058669627, "grad_norm": 3.1945369243621826, "learning_rate": 2.627665709311442e-06, "loss": 2.873, "step": 21906 }, { "epoch": 0.771937242879921, "grad_norm": 1.5187127590179443, "learning_rate": 2.6268946672112594e-06, "loss": 0.9065, "step": 21907 }, { "epoch": 0.7719724798928794, "grad_norm": 2.4306633472442627, "learning_rate": 2.6261237211467083e-06, "loss": 2.5007, "step": 21908 }, { "epoch": 0.7720077169058379, "grad_norm": 2.380330801010132, "learning_rate": 2.6253528711278365e-06, "loss": 2.3295, "step": 21909 }, { "epoch": 0.7720429539187963, "grad_norm": 4.308595657348633, "learning_rate": 2.624582117164681e-06, "loss": 5.0629, "step": 21910 }, { "epoch": 0.7720781909317547, "grad_norm": 3.277013063430786, "learning_rate": 2.6238114592672846e-06, "loss": 0.8016, "step": 21911 }, { "epoch": 0.7721134279447132, "grad_norm": 6.048562526702881, "learning_rate": 2.6230408974456835e-06, "loss": 3.2125, "step": 21912 }, { "epoch": 0.7721486649576715, "grad_norm": 6.639017105102539, "learning_rate": 2.6222704317099134e-06, "loss": 5.6731, "step": 21913 }, { "epoch": 0.7721839019706299, "grad_norm": 0.7938767075538635, "learning_rate": 2.6215000620700113e-06, "loss": 0.7378, "step": 21914 }, { "epoch": 0.7722191389835884, "grad_norm": 2.3243157863616943, "learning_rate": 2.6207297885360107e-06, "loss": 0.8209, "step": 21915 }, { "epoch": 0.7722543759965468, "grad_norm": 2.843350887298584, "learning_rate": 2.6199596111179447e-06, "loss": 2.9739, "step": 21916 }, { "epoch": 0.7722896130095052, "grad_norm": 2.7946865558624268, "learning_rate": 2.619189529825844e-06, "loss": 2.6643, "step": 21917 }, { "epoch": 0.7723248500224636, "grad_norm": 1.0238333940505981, "learning_rate": 2.6184195446697413e-06, "loss": 0.7945, "step": 21918 }, { "epoch": 0.772360087035422, "grad_norm": 1.7617801427841187, "learning_rate": 2.6176496556596633e-06, "loss": 0.9292, "step": 21919 }, { "epoch": 0.7723953240483804, "grad_norm": 3.787569761276245, "learning_rate": 2.6168798628056413e-06, "loss": 2.8188, "step": 21920 }, { "epoch": 0.7724305610613388, "grad_norm": 1.4125341176986694, "learning_rate": 2.6161101661176944e-06, "loss": 0.9259, "step": 21921 }, { "epoch": 0.7724657980742973, "grad_norm": 4.246138572692871, "learning_rate": 2.6153405656058563e-06, "loss": 2.9307, "step": 21922 }, { "epoch": 0.7725010350872556, "grad_norm": 0.7926246523857117, "learning_rate": 2.614571061280151e-06, "loss": 0.7482, "step": 21923 }, { "epoch": 0.772536272100214, "grad_norm": 3.7740795612335205, "learning_rate": 2.613801653150595e-06, "loss": 3.2641, "step": 21924 }, { "epoch": 0.7725715091131725, "grad_norm": 4.327142715454102, "learning_rate": 2.6130323412272108e-06, "loss": 2.8768, "step": 21925 }, { "epoch": 0.7726067461261309, "grad_norm": 4.811450481414795, "learning_rate": 2.612263125520027e-06, "loss": 2.9195, "step": 21926 }, { "epoch": 0.7726419831390893, "grad_norm": 5.597357273101807, "learning_rate": 2.6114940060390537e-06, "loss": 4.7633, "step": 21927 }, { "epoch": 0.7726772201520478, "grad_norm": 1.2537729740142822, "learning_rate": 2.610724982794313e-06, "loss": 0.8802, "step": 21928 }, { "epoch": 0.7727124571650061, "grad_norm": 1.3341097831726074, "learning_rate": 2.609956055795819e-06, "loss": 0.7309, "step": 21929 }, { "epoch": 0.7727476941779645, "grad_norm": 0.8571672439575195, "learning_rate": 2.6091872250535897e-06, "loss": 0.8133, "step": 21930 }, { "epoch": 0.7727829311909229, "grad_norm": 1.5647984743118286, "learning_rate": 2.608418490577639e-06, "loss": 0.9406, "step": 21931 }, { "epoch": 0.7728181682038814, "grad_norm": 1.3332693576812744, "learning_rate": 2.607649852377978e-06, "loss": 0.7189, "step": 21932 }, { "epoch": 0.7728534052168398, "grad_norm": 6.519789695739746, "learning_rate": 2.6068813104646197e-06, "loss": 2.4782, "step": 21933 }, { "epoch": 0.7728886422297981, "grad_norm": 4.217646598815918, "learning_rate": 2.606112864847574e-06, "loss": 2.7897, "step": 21934 }, { "epoch": 0.7729238792427566, "grad_norm": 1.4282506704330444, "learning_rate": 2.605344515536851e-06, "loss": 1.0794, "step": 21935 }, { "epoch": 0.772959116255715, "grad_norm": 2.0046725273132324, "learning_rate": 2.6045762625424552e-06, "loss": 0.9916, "step": 21936 }, { "epoch": 0.7729943532686734, "grad_norm": 4.03709602355957, "learning_rate": 2.603808105874398e-06, "loss": 3.2865, "step": 21937 }, { "epoch": 0.7730295902816319, "grad_norm": 1.3678206205368042, "learning_rate": 2.6030400455426807e-06, "loss": 0.9591, "step": 21938 }, { "epoch": 0.7730648272945903, "grad_norm": 2.4758243560791016, "learning_rate": 2.6022720815573133e-06, "loss": 2.4529, "step": 21939 }, { "epoch": 0.7731000643075486, "grad_norm": 6.48415470123291, "learning_rate": 2.6015042139282875e-06, "loss": 5.2624, "step": 21940 }, { "epoch": 0.773135301320507, "grad_norm": 2.5701277256011963, "learning_rate": 2.6007364426656144e-06, "loss": 2.4984, "step": 21941 }, { "epoch": 0.7731705383334655, "grad_norm": 1.0963153839111328, "learning_rate": 2.599968767779296e-06, "loss": 0.9199, "step": 21942 }, { "epoch": 0.7732057753464239, "grad_norm": 0.7545276880264282, "learning_rate": 2.5992011892793235e-06, "loss": 0.9785, "step": 21943 }, { "epoch": 0.7732410123593823, "grad_norm": 2.9691972732543945, "learning_rate": 2.598433707175695e-06, "loss": 3.2177, "step": 21944 }, { "epoch": 0.7732762493723407, "grad_norm": 5.733715534210205, "learning_rate": 2.597666321478416e-06, "loss": 3.0643, "step": 21945 }, { "epoch": 0.7733114863852991, "grad_norm": 0.9167474508285522, "learning_rate": 2.596899032197473e-06, "loss": 0.9472, "step": 21946 }, { "epoch": 0.7733467233982575, "grad_norm": 5.923259735107422, "learning_rate": 2.5961318393428637e-06, "loss": 7.3147, "step": 21947 }, { "epoch": 0.773381960411216, "grad_norm": 0.8948968052864075, "learning_rate": 2.5953647429245775e-06, "loss": 0.6615, "step": 21948 }, { "epoch": 0.7734171974241744, "grad_norm": 2.146810531616211, "learning_rate": 2.5945977429526137e-06, "loss": 0.8955, "step": 21949 }, { "epoch": 0.7734524344371327, "grad_norm": 3.31792950630188, "learning_rate": 2.593830839436955e-06, "loss": 2.4573, "step": 21950 }, { "epoch": 0.7734876714500912, "grad_norm": 7.7098469734191895, "learning_rate": 2.593064032387593e-06, "loss": 6.9057, "step": 21951 }, { "epoch": 0.7735229084630496, "grad_norm": 3.0937960147857666, "learning_rate": 2.5922973218145163e-06, "loss": 2.9152, "step": 21952 }, { "epoch": 0.773558145476008, "grad_norm": 1.6208446025848389, "learning_rate": 2.5915307077277095e-06, "loss": 0.8026, "step": 21953 }, { "epoch": 0.7735933824889664, "grad_norm": 1.88522207736969, "learning_rate": 2.5907641901371594e-06, "loss": 1.021, "step": 21954 }, { "epoch": 0.7736286195019249, "grad_norm": 3.8843491077423096, "learning_rate": 2.58999776905285e-06, "loss": 4.0939, "step": 21955 }, { "epoch": 0.7736638565148832, "grad_norm": 1.0977367162704468, "learning_rate": 2.589231444484763e-06, "loss": 0.7359, "step": 21956 }, { "epoch": 0.7736990935278416, "grad_norm": 2.7197961807250977, "learning_rate": 2.5884652164428804e-06, "loss": 2.5754, "step": 21957 }, { "epoch": 0.7737343305408001, "grad_norm": 1.1407580375671387, "learning_rate": 2.587699084937185e-06, "loss": 1.0717, "step": 21958 }, { "epoch": 0.7737695675537585, "grad_norm": 5.660708427429199, "learning_rate": 2.586933049977648e-06, "loss": 5.151, "step": 21959 }, { "epoch": 0.7738048045667169, "grad_norm": 4.510324478149414, "learning_rate": 2.586167111574255e-06, "loss": 0.8953, "step": 21960 }, { "epoch": 0.7738400415796753, "grad_norm": 12.525035858154297, "learning_rate": 2.5854012697369823e-06, "loss": 2.8244, "step": 21961 }, { "epoch": 0.7738752785926337, "grad_norm": 5.3815789222717285, "learning_rate": 2.5846355244758003e-06, "loss": 4.7846, "step": 21962 }, { "epoch": 0.7739105156055921, "grad_norm": 4.750017166137695, "learning_rate": 2.583869875800682e-06, "loss": 3.4243, "step": 21963 }, { "epoch": 0.7739457526185505, "grad_norm": 1.7964650392532349, "learning_rate": 2.583104323721609e-06, "loss": 0.9451, "step": 21964 }, { "epoch": 0.773980989631509, "grad_norm": 1.114372968673706, "learning_rate": 2.582338868248544e-06, "loss": 0.8504, "step": 21965 }, { "epoch": 0.7740162266444673, "grad_norm": 4.644189357757568, "learning_rate": 2.5815735093914596e-06, "loss": 2.701, "step": 21966 }, { "epoch": 0.7740514636574257, "grad_norm": 5.911581039428711, "learning_rate": 2.5808082471603225e-06, "loss": 4.7433, "step": 21967 }, { "epoch": 0.7740867006703842, "grad_norm": 2.9254953861236572, "learning_rate": 2.58004308156511e-06, "loss": 3.2971, "step": 21968 }, { "epoch": 0.7741219376833426, "grad_norm": 1.3836710453033447, "learning_rate": 2.5792780126157767e-06, "loss": 0.9635, "step": 21969 }, { "epoch": 0.774157174696301, "grad_norm": 1.109277606010437, "learning_rate": 2.5785130403222946e-06, "loss": 0.9026, "step": 21970 }, { "epoch": 0.7741924117092595, "grad_norm": 2.366025924682617, "learning_rate": 2.5777481646946243e-06, "loss": 0.6899, "step": 21971 }, { "epoch": 0.7742276487222178, "grad_norm": 5.629636764526367, "learning_rate": 2.5769833857427296e-06, "loss": 2.7429, "step": 21972 }, { "epoch": 0.7742628857351762, "grad_norm": 2.55437970161438, "learning_rate": 2.576218703476572e-06, "loss": 0.7974, "step": 21973 }, { "epoch": 0.7742981227481346, "grad_norm": 4.385952472686768, "learning_rate": 2.575454117906111e-06, "loss": 2.7066, "step": 21974 }, { "epoch": 0.7743333597610931, "grad_norm": 5.22998571395874, "learning_rate": 2.574689629041307e-06, "loss": 4.738, "step": 21975 }, { "epoch": 0.7743685967740515, "grad_norm": 0.833550214767456, "learning_rate": 2.573925236892115e-06, "loss": 0.8534, "step": 21976 }, { "epoch": 0.7744038337870098, "grad_norm": 4.347108840942383, "learning_rate": 2.5731609414684967e-06, "loss": 2.9739, "step": 21977 }, { "epoch": 0.7744390707999683, "grad_norm": 2.5955698490142822, "learning_rate": 2.5723967427803964e-06, "loss": 2.6118, "step": 21978 }, { "epoch": 0.7744743078129267, "grad_norm": 2.9185335636138916, "learning_rate": 2.571632640837779e-06, "loss": 3.0643, "step": 21979 }, { "epoch": 0.7745095448258851, "grad_norm": 2.165422201156616, "learning_rate": 2.5708686356505942e-06, "loss": 0.7483, "step": 21980 }, { "epoch": 0.7745447818388436, "grad_norm": 3.678124189376831, "learning_rate": 2.570104727228789e-06, "loss": 2.6317, "step": 21981 }, { "epoch": 0.7745800188518019, "grad_norm": 1.5479016304016113, "learning_rate": 2.5693409155823135e-06, "loss": 0.999, "step": 21982 }, { "epoch": 0.7746152558647603, "grad_norm": 1.7270008325576782, "learning_rate": 2.568577200721125e-06, "loss": 0.8762, "step": 21983 }, { "epoch": 0.7746504928777188, "grad_norm": 4.27484655380249, "learning_rate": 2.567813582655163e-06, "loss": 2.5346, "step": 21984 }, { "epoch": 0.7746857298906772, "grad_norm": 2.6518056392669678, "learning_rate": 2.567050061394375e-06, "loss": 2.3914, "step": 21985 }, { "epoch": 0.7747209669036356, "grad_norm": 0.966463029384613, "learning_rate": 2.566286636948703e-06, "loss": 1.1546, "step": 21986 }, { "epoch": 0.7747562039165939, "grad_norm": 1.0408686399459839, "learning_rate": 2.5655233093281016e-06, "loss": 0.8128, "step": 21987 }, { "epoch": 0.7747914409295524, "grad_norm": 1.0336068868637085, "learning_rate": 2.564760078542502e-06, "loss": 0.9848, "step": 21988 }, { "epoch": 0.7748266779425108, "grad_norm": 4.509195327758789, "learning_rate": 2.563996944601851e-06, "loss": 4.7619, "step": 21989 }, { "epoch": 0.7748619149554692, "grad_norm": 1.039473295211792, "learning_rate": 2.5632339075160863e-06, "loss": 0.7547, "step": 21990 }, { "epoch": 0.7748971519684277, "grad_norm": 2.834155797958374, "learning_rate": 2.5624709672951464e-06, "loss": 2.7374, "step": 21991 }, { "epoch": 0.7749323889813861, "grad_norm": 0.9742562174797058, "learning_rate": 2.5617081239489703e-06, "loss": 0.7576, "step": 21992 }, { "epoch": 0.7749676259943444, "grad_norm": 4.034585475921631, "learning_rate": 2.560945377487494e-06, "loss": 4.718, "step": 21993 }, { "epoch": 0.7750028630073029, "grad_norm": 3.704277515411377, "learning_rate": 2.5601827279206505e-06, "loss": 2.9149, "step": 21994 }, { "epoch": 0.7750381000202613, "grad_norm": 1.4600342512130737, "learning_rate": 2.5594201752583747e-06, "loss": 1.0552, "step": 21995 }, { "epoch": 0.7750733370332197, "grad_norm": 0.8530190587043762, "learning_rate": 2.5586577195106023e-06, "loss": 0.6679, "step": 21996 }, { "epoch": 0.7751085740461781, "grad_norm": 1.2379590272903442, "learning_rate": 2.557895360687256e-06, "loss": 0.6777, "step": 21997 }, { "epoch": 0.7751438110591365, "grad_norm": 2.8189096450805664, "learning_rate": 2.5571330987982723e-06, "loss": 2.5805, "step": 21998 }, { "epoch": 0.7751790480720949, "grad_norm": 2.0170397758483887, "learning_rate": 2.5563709338535815e-06, "loss": 0.6143, "step": 21999 }, { "epoch": 0.7752142850850533, "grad_norm": 3.7277677059173584, "learning_rate": 2.555608865863104e-06, "loss": 2.9231, "step": 22000 }, { "epoch": 0.7752495220980118, "grad_norm": 1.0637117624282837, "learning_rate": 2.5548468948367666e-06, "loss": 0.8863, "step": 22001 }, { "epoch": 0.7752847591109702, "grad_norm": 5.4170308113098145, "learning_rate": 2.5540850207845025e-06, "loss": 2.8409, "step": 22002 }, { "epoch": 0.7753199961239285, "grad_norm": 2.406965494155884, "learning_rate": 2.553323243716227e-06, "loss": 2.781, "step": 22003 }, { "epoch": 0.775355233136887, "grad_norm": 1.194812536239624, "learning_rate": 2.5525615636418646e-06, "loss": 1.0492, "step": 22004 }, { "epoch": 0.7753904701498454, "grad_norm": 5.225827693939209, "learning_rate": 2.5517999805713333e-06, "loss": 5.0022, "step": 22005 }, { "epoch": 0.7754257071628038, "grad_norm": 4.273256301879883, "learning_rate": 2.5510384945145627e-06, "loss": 2.9009, "step": 22006 }, { "epoch": 0.7754609441757622, "grad_norm": 4.227302074432373, "learning_rate": 2.5502771054814614e-06, "loss": 4.0852, "step": 22007 }, { "epoch": 0.7754961811887207, "grad_norm": 4.714666843414307, "learning_rate": 2.5495158134819486e-06, "loss": 3.0755, "step": 22008 }, { "epoch": 0.775531418201679, "grad_norm": 2.5578019618988037, "learning_rate": 2.5487546185259426e-06, "loss": 2.7436, "step": 22009 }, { "epoch": 0.7755666552146374, "grad_norm": 6.876488208770752, "learning_rate": 2.5479935206233563e-06, "loss": 5.0259, "step": 22010 }, { "epoch": 0.7756018922275959, "grad_norm": 1.4772688150405884, "learning_rate": 2.547232519784103e-06, "loss": 0.8628, "step": 22011 }, { "epoch": 0.7756371292405543, "grad_norm": 3.60060453414917, "learning_rate": 2.546471616018097e-06, "loss": 2.8087, "step": 22012 }, { "epoch": 0.7756723662535127, "grad_norm": 3.775800943374634, "learning_rate": 2.545710809335247e-06, "loss": 2.8056, "step": 22013 }, { "epoch": 0.7757076032664711, "grad_norm": 1.5815708637237549, "learning_rate": 2.5449500997454624e-06, "loss": 0.974, "step": 22014 }, { "epoch": 0.7757428402794295, "grad_norm": 3.815042018890381, "learning_rate": 2.5441894872586557e-06, "loss": 0.5633, "step": 22015 }, { "epoch": 0.7757780772923879, "grad_norm": 3.912245750427246, "learning_rate": 2.543428971884725e-06, "loss": 3.1649, "step": 22016 }, { "epoch": 0.7758133143053464, "grad_norm": 1.308727741241455, "learning_rate": 2.542668553633585e-06, "loss": 0.7802, "step": 22017 }, { "epoch": 0.7758485513183048, "grad_norm": 3.062704086303711, "learning_rate": 2.54190823251514e-06, "loss": 2.9497, "step": 22018 }, { "epoch": 0.7758837883312631, "grad_norm": 4.497646331787109, "learning_rate": 2.541148008539287e-06, "loss": 5.0829, "step": 22019 }, { "epoch": 0.7759190253442215, "grad_norm": 0.7320609092712402, "learning_rate": 2.5403878817159277e-06, "loss": 0.8957, "step": 22020 }, { "epoch": 0.77595426235718, "grad_norm": 4.868218421936035, "learning_rate": 2.53962785205497e-06, "loss": 6.2654, "step": 22021 }, { "epoch": 0.7759894993701384, "grad_norm": 1.0931798219680786, "learning_rate": 2.5388679195663124e-06, "loss": 0.6987, "step": 22022 }, { "epoch": 0.7760247363830968, "grad_norm": 1.1287755966186523, "learning_rate": 2.538108084259848e-06, "loss": 1.0381, "step": 22023 }, { "epoch": 0.7760599733960553, "grad_norm": 3.1247010231018066, "learning_rate": 2.537348346145474e-06, "loss": 2.8964, "step": 22024 }, { "epoch": 0.7760952104090136, "grad_norm": 5.282299518585205, "learning_rate": 2.536588705233094e-06, "loss": 4.5343, "step": 22025 }, { "epoch": 0.776130447421972, "grad_norm": 1.9191871881484985, "learning_rate": 2.5358291615325924e-06, "loss": 2.7508, "step": 22026 }, { "epoch": 0.7761656844349305, "grad_norm": 3.0516650676727295, "learning_rate": 2.5350697150538684e-06, "loss": 2.6206, "step": 22027 }, { "epoch": 0.7762009214478889, "grad_norm": 1.120721459388733, "learning_rate": 2.5343103658068114e-06, "loss": 0.85, "step": 22028 }, { "epoch": 0.7762361584608473, "grad_norm": 5.3392486572265625, "learning_rate": 2.5335511138013137e-06, "loss": 5.2192, "step": 22029 }, { "epoch": 0.7762713954738056, "grad_norm": 1.7954446077346802, "learning_rate": 2.5327919590472626e-06, "loss": 0.9125, "step": 22030 }, { "epoch": 0.7763066324867641, "grad_norm": 6.7505388259887695, "learning_rate": 2.5320329015545485e-06, "loss": 3.1423, "step": 22031 }, { "epoch": 0.7763418694997225, "grad_norm": 3.5871083736419678, "learning_rate": 2.5312739413330568e-06, "loss": 4.2518, "step": 22032 }, { "epoch": 0.7763771065126809, "grad_norm": 2.118488073348999, "learning_rate": 2.5305150783926735e-06, "loss": 0.8961, "step": 22033 }, { "epoch": 0.7764123435256394, "grad_norm": 9.257155418395996, "learning_rate": 2.5297563127432844e-06, "loss": 4.9507, "step": 22034 }, { "epoch": 0.7764475805385977, "grad_norm": 1.2768667936325073, "learning_rate": 2.5289976443947662e-06, "loss": 0.6583, "step": 22035 }, { "epoch": 0.7764828175515561, "grad_norm": 6.392040729522705, "learning_rate": 2.5282390733570083e-06, "loss": 4.5138, "step": 22036 }, { "epoch": 0.7765180545645146, "grad_norm": 5.340403079986572, "learning_rate": 2.527480599639891e-06, "loss": 4.8388, "step": 22037 }, { "epoch": 0.776553291577473, "grad_norm": 6.331844806671143, "learning_rate": 2.5267222232532884e-06, "loss": 7.0098, "step": 22038 }, { "epoch": 0.7765885285904314, "grad_norm": 3.9870429039001465, "learning_rate": 2.5259639442070772e-06, "loss": 2.7891, "step": 22039 }, { "epoch": 0.7766237656033897, "grad_norm": 1.763289213180542, "learning_rate": 2.5252057625111403e-06, "loss": 0.8711, "step": 22040 }, { "epoch": 0.7766590026163482, "grad_norm": 1.7752655744552612, "learning_rate": 2.5244476781753535e-06, "loss": 0.7538, "step": 22041 }, { "epoch": 0.7766942396293066, "grad_norm": 6.1091156005859375, "learning_rate": 2.5236896912095852e-06, "loss": 5.0433, "step": 22042 }, { "epoch": 0.776729476642265, "grad_norm": 4.022854328155518, "learning_rate": 2.522931801623708e-06, "loss": 3.3846, "step": 22043 }, { "epoch": 0.7767647136552235, "grad_norm": 3.172677755355835, "learning_rate": 2.522174009427603e-06, "loss": 2.9041, "step": 22044 }, { "epoch": 0.7767999506681819, "grad_norm": 1.0523185729980469, "learning_rate": 2.521416314631131e-06, "loss": 0.9047, "step": 22045 }, { "epoch": 0.7768351876811402, "grad_norm": 1.611743450164795, "learning_rate": 2.520658717244163e-06, "loss": 0.7335, "step": 22046 }, { "epoch": 0.7768704246940987, "grad_norm": 1.4306364059448242, "learning_rate": 2.5199012172765693e-06, "loss": 0.9498, "step": 22047 }, { "epoch": 0.7769056617070571, "grad_norm": 5.764530658721924, "learning_rate": 2.5191438147382143e-06, "loss": 4.8944, "step": 22048 }, { "epoch": 0.7769408987200155, "grad_norm": 2.9698901176452637, "learning_rate": 2.5183865096389635e-06, "loss": 2.9795, "step": 22049 }, { "epoch": 0.7769761357329739, "grad_norm": 1.195215106010437, "learning_rate": 2.517629301988682e-06, "loss": 0.7541, "step": 22050 }, { "epoch": 0.7770113727459323, "grad_norm": 3.477221965789795, "learning_rate": 2.516872191797233e-06, "loss": 2.533, "step": 22051 }, { "epoch": 0.7770466097588907, "grad_norm": 4.368900299072266, "learning_rate": 2.5161151790744754e-06, "loss": 5.4122, "step": 22052 }, { "epoch": 0.7770818467718491, "grad_norm": 6.642825603485107, "learning_rate": 2.5153582638302744e-06, "loss": 5.4007, "step": 22053 }, { "epoch": 0.7771170837848076, "grad_norm": 0.8712427616119385, "learning_rate": 2.5146014460744796e-06, "loss": 0.8679, "step": 22054 }, { "epoch": 0.777152320797766, "grad_norm": 3.8135147094726562, "learning_rate": 2.5138447258169572e-06, "loss": 3.0727, "step": 22055 }, { "epoch": 0.7771875578107243, "grad_norm": 3.8056960105895996, "learning_rate": 2.5130881030675634e-06, "loss": 3.2467, "step": 22056 }, { "epoch": 0.7772227948236828, "grad_norm": 1.0496584177017212, "learning_rate": 2.512331577836149e-06, "loss": 0.7772, "step": 22057 }, { "epoch": 0.7772580318366412, "grad_norm": 7.025930404663086, "learning_rate": 2.511575150132566e-06, "loss": 3.1204, "step": 22058 }, { "epoch": 0.7772932688495996, "grad_norm": 4.985445499420166, "learning_rate": 2.510818819966673e-06, "loss": 3.39, "step": 22059 }, { "epoch": 0.7773285058625581, "grad_norm": 1.1295002698898315, "learning_rate": 2.510062587348322e-06, "loss": 0.979, "step": 22060 }, { "epoch": 0.7773637428755165, "grad_norm": 0.9912006258964539, "learning_rate": 2.5093064522873567e-06, "loss": 0.9036, "step": 22061 }, { "epoch": 0.7773989798884748, "grad_norm": 3.5997300148010254, "learning_rate": 2.508550414793629e-06, "loss": 3.2089, "step": 22062 }, { "epoch": 0.7774342169014332, "grad_norm": 1.1666772365570068, "learning_rate": 2.507794474876988e-06, "loss": 0.9596, "step": 22063 }, { "epoch": 0.7774694539143917, "grad_norm": 9.910632133483887, "learning_rate": 2.5070386325472763e-06, "loss": 6.7702, "step": 22064 }, { "epoch": 0.7775046909273501, "grad_norm": 1.6818108558654785, "learning_rate": 2.506282887814342e-06, "loss": 0.828, "step": 22065 }, { "epoch": 0.7775399279403085, "grad_norm": 1.261269211769104, "learning_rate": 2.505527240688027e-06, "loss": 1.008, "step": 22066 }, { "epoch": 0.777575164953267, "grad_norm": 2.2184195518493652, "learning_rate": 2.504771691178175e-06, "loss": 1.0794, "step": 22067 }, { "epoch": 0.7776104019662253, "grad_norm": 4.12846040725708, "learning_rate": 2.5040162392946254e-06, "loss": 2.4589, "step": 22068 }, { "epoch": 0.7776456389791837, "grad_norm": 1.179187536239624, "learning_rate": 2.5032608850472197e-06, "loss": 0.9605, "step": 22069 }, { "epoch": 0.7776808759921422, "grad_norm": 7.420553207397461, "learning_rate": 2.502505628445796e-06, "loss": 3.5215, "step": 22070 }, { "epoch": 0.7777161130051006, "grad_norm": 1.1285324096679688, "learning_rate": 2.5017504695001917e-06, "loss": 0.8881, "step": 22071 }, { "epoch": 0.777751350018059, "grad_norm": 2.397744655609131, "learning_rate": 2.500995408220246e-06, "loss": 2.5898, "step": 22072 }, { "epoch": 0.7777865870310173, "grad_norm": 3.8746047019958496, "learning_rate": 2.500240444615787e-06, "loss": 3.1101, "step": 22073 }, { "epoch": 0.7778218240439758, "grad_norm": 2.6621222496032715, "learning_rate": 2.4994855786966477e-06, "loss": 2.8569, "step": 22074 }, { "epoch": 0.7778570610569342, "grad_norm": 1.1365596055984497, "learning_rate": 2.4987308104726703e-06, "loss": 0.9542, "step": 22075 }, { "epoch": 0.7778922980698926, "grad_norm": 0.8434750437736511, "learning_rate": 2.4979761399536774e-06, "loss": 0.8593, "step": 22076 }, { "epoch": 0.7779275350828511, "grad_norm": 25.016942977905273, "learning_rate": 2.4972215671495005e-06, "loss": 2.7036, "step": 22077 }, { "epoch": 0.7779627720958094, "grad_norm": 2.6883490085601807, "learning_rate": 2.4964670920699664e-06, "loss": 2.407, "step": 22078 }, { "epoch": 0.7779980091087678, "grad_norm": 4.479471206665039, "learning_rate": 2.4957127147249084e-06, "loss": 3.186, "step": 22079 }, { "epoch": 0.7780332461217263, "grad_norm": 6.090377330780029, "learning_rate": 2.494958435124146e-06, "loss": 4.6782, "step": 22080 }, { "epoch": 0.7780684831346847, "grad_norm": 3.9498507976531982, "learning_rate": 2.494204253277507e-06, "loss": 3.0097, "step": 22081 }, { "epoch": 0.7781037201476431, "grad_norm": 3.570667266845703, "learning_rate": 2.493450169194812e-06, "loss": 2.7327, "step": 22082 }, { "epoch": 0.7781389571606014, "grad_norm": 1.4757163524627686, "learning_rate": 2.4926961828858853e-06, "loss": 1.0591, "step": 22083 }, { "epoch": 0.7781741941735599, "grad_norm": 0.6592480540275574, "learning_rate": 2.4919422943605477e-06, "loss": 0.7913, "step": 22084 }, { "epoch": 0.7782094311865183, "grad_norm": 1.0715343952178955, "learning_rate": 2.4911885036286178e-06, "loss": 0.9203, "step": 22085 }, { "epoch": 0.7782446681994767, "grad_norm": 9.293996810913086, "learning_rate": 2.4904348106999132e-06, "loss": 3.5108, "step": 22086 }, { "epoch": 0.7782799052124352, "grad_norm": 5.090691566467285, "learning_rate": 2.489681215584253e-06, "loss": 4.9802, "step": 22087 }, { "epoch": 0.7783151422253936, "grad_norm": 6.623345375061035, "learning_rate": 2.488927718291454e-06, "loss": 3.5349, "step": 22088 }, { "epoch": 0.7783503792383519, "grad_norm": 1.4256343841552734, "learning_rate": 2.488174318831322e-06, "loss": 0.7547, "step": 22089 }, { "epoch": 0.7783856162513104, "grad_norm": 4.013413906097412, "learning_rate": 2.48742101721368e-06, "loss": 2.64, "step": 22090 }, { "epoch": 0.7784208532642688, "grad_norm": 1.6077324151992798, "learning_rate": 2.486667813448339e-06, "loss": 0.7644, "step": 22091 }, { "epoch": 0.7784560902772272, "grad_norm": 1.192212462425232, "learning_rate": 2.485914707545104e-06, "loss": 1.1116, "step": 22092 }, { "epoch": 0.7784913272901857, "grad_norm": 1.085044264793396, "learning_rate": 2.485161699513784e-06, "loss": 0.9778, "step": 22093 }, { "epoch": 0.778526564303144, "grad_norm": 1.2226213216781616, "learning_rate": 2.4844087893641923e-06, "loss": 0.8986, "step": 22094 }, { "epoch": 0.7785618013161024, "grad_norm": 1.0740480422973633, "learning_rate": 2.4836559771061374e-06, "loss": 0.9529, "step": 22095 }, { "epoch": 0.7785970383290608, "grad_norm": 0.8854149580001831, "learning_rate": 2.4829032627494176e-06, "loss": 1.0718, "step": 22096 }, { "epoch": 0.7786322753420193, "grad_norm": 2.260274887084961, "learning_rate": 2.482150646303837e-06, "loss": 2.7749, "step": 22097 }, { "epoch": 0.7786675123549777, "grad_norm": 1.2324718236923218, "learning_rate": 2.481398127779208e-06, "loss": 0.7749, "step": 22098 }, { "epoch": 0.778702749367936, "grad_norm": 4.806285381317139, "learning_rate": 2.480645707185323e-06, "loss": 5.5637, "step": 22099 }, { "epoch": 0.7787379863808945, "grad_norm": 4.275031089782715, "learning_rate": 2.479893384531985e-06, "loss": 2.7711, "step": 22100 }, { "epoch": 0.7787732233938529, "grad_norm": 1.220826268196106, "learning_rate": 2.479141159828995e-06, "loss": 0.7514, "step": 22101 }, { "epoch": 0.7788084604068113, "grad_norm": 5.94734001159668, "learning_rate": 2.478389033086148e-06, "loss": 4.9912, "step": 22102 }, { "epoch": 0.7788436974197698, "grad_norm": 7.385219097137451, "learning_rate": 2.4776370043132423e-06, "loss": 5.4773, "step": 22103 }, { "epoch": 0.7788789344327282, "grad_norm": 2.7770094871520996, "learning_rate": 2.4768850735200735e-06, "loss": 3.176, "step": 22104 }, { "epoch": 0.7789141714456865, "grad_norm": 1.6741899251937866, "learning_rate": 2.476133240716434e-06, "loss": 0.8049, "step": 22105 }, { "epoch": 0.7789494084586449, "grad_norm": 4.352237224578857, "learning_rate": 2.4753815059121187e-06, "loss": 4.9938, "step": 22106 }, { "epoch": 0.7789846454716034, "grad_norm": 2.5351672172546387, "learning_rate": 2.47462986911692e-06, "loss": 2.9859, "step": 22107 }, { "epoch": 0.7790198824845618, "grad_norm": 1.4161736965179443, "learning_rate": 2.47387833034062e-06, "loss": 0.9286, "step": 22108 }, { "epoch": 0.7790551194975202, "grad_norm": 1.0306520462036133, "learning_rate": 2.473126889593016e-06, "loss": 0.7856, "step": 22109 }, { "epoch": 0.7790903565104786, "grad_norm": 5.5914082527160645, "learning_rate": 2.472375546883896e-06, "loss": 3.0613, "step": 22110 }, { "epoch": 0.779125593523437, "grad_norm": 6.102265357971191, "learning_rate": 2.4716243022230403e-06, "loss": 2.2167, "step": 22111 }, { "epoch": 0.7791608305363954, "grad_norm": 5.374413013458252, "learning_rate": 2.470873155620235e-06, "loss": 4.6648, "step": 22112 }, { "epoch": 0.7791960675493539, "grad_norm": 3.5935728549957275, "learning_rate": 2.4701221070852675e-06, "loss": 2.9696, "step": 22113 }, { "epoch": 0.7792313045623123, "grad_norm": 1.0420047044754028, "learning_rate": 2.4693711566279233e-06, "loss": 0.9739, "step": 22114 }, { "epoch": 0.7792665415752706, "grad_norm": 1.9686576128005981, "learning_rate": 2.468620304257975e-06, "loss": 0.7656, "step": 22115 }, { "epoch": 0.779301778588229, "grad_norm": 3.367152214050293, "learning_rate": 2.4678695499852036e-06, "loss": 2.6835, "step": 22116 }, { "epoch": 0.7793370156011875, "grad_norm": 3.9708194732666016, "learning_rate": 2.4671188938193967e-06, "loss": 4.1038, "step": 22117 }, { "epoch": 0.7793722526141459, "grad_norm": 5.6643452644348145, "learning_rate": 2.466368335770323e-06, "loss": 3.417, "step": 22118 }, { "epoch": 0.7794074896271043, "grad_norm": 3.096493721008301, "learning_rate": 2.4656178758477613e-06, "loss": 2.8732, "step": 22119 }, { "epoch": 0.7794427266400628, "grad_norm": 3.095717430114746, "learning_rate": 2.4648675140614865e-06, "loss": 2.6831, "step": 22120 }, { "epoch": 0.7794779636530211, "grad_norm": 4.609238147735596, "learning_rate": 2.4641172504212718e-06, "loss": 2.9012, "step": 22121 }, { "epoch": 0.7795132006659795, "grad_norm": 2.278601884841919, "learning_rate": 2.463367084936891e-06, "loss": 2.2856, "step": 22122 }, { "epoch": 0.779548437678938, "grad_norm": 1.4599636793136597, "learning_rate": 2.462617017618113e-06, "loss": 0.8069, "step": 22123 }, { "epoch": 0.7795836746918964, "grad_norm": 5.500956058502197, "learning_rate": 2.4618670484747086e-06, "loss": 5.058, "step": 22124 }, { "epoch": 0.7796189117048548, "grad_norm": 1.0994513034820557, "learning_rate": 2.4611171775164457e-06, "loss": 0.703, "step": 22125 }, { "epoch": 0.7796541487178132, "grad_norm": 1.3408374786376953, "learning_rate": 2.460367404753096e-06, "loss": 0.6532, "step": 22126 }, { "epoch": 0.7796893857307716, "grad_norm": 1.0061531066894531, "learning_rate": 2.4596177301944134e-06, "loss": 0.7618, "step": 22127 }, { "epoch": 0.77972462274373, "grad_norm": 5.617111682891846, "learning_rate": 2.4588681538501745e-06, "loss": 3.0664, "step": 22128 }, { "epoch": 0.7797598597566884, "grad_norm": 3.513892412185669, "learning_rate": 2.4581186757301423e-06, "loss": 3.2449, "step": 22129 }, { "epoch": 0.7797950967696469, "grad_norm": 3.357987880706787, "learning_rate": 2.457369295844071e-06, "loss": 2.9995, "step": 22130 }, { "epoch": 0.7798303337826052, "grad_norm": 1.7544798851013184, "learning_rate": 2.456620014201723e-06, "loss": 0.7807, "step": 22131 }, { "epoch": 0.7798655707955636, "grad_norm": 3.763577699661255, "learning_rate": 2.455870830812862e-06, "loss": 2.8592, "step": 22132 }, { "epoch": 0.7799008078085221, "grad_norm": 9.758176803588867, "learning_rate": 2.455121745687248e-06, "loss": 3.1439, "step": 22133 }, { "epoch": 0.7799360448214805, "grad_norm": 1.1671836376190186, "learning_rate": 2.454372758834632e-06, "loss": 0.8152, "step": 22134 }, { "epoch": 0.7799712818344389, "grad_norm": 1.6096158027648926, "learning_rate": 2.4536238702647686e-06, "loss": 0.8385, "step": 22135 }, { "epoch": 0.7800065188473974, "grad_norm": 1.3295539617538452, "learning_rate": 2.4528750799874212e-06, "loss": 0.9044, "step": 22136 }, { "epoch": 0.7800417558603557, "grad_norm": 1.2610118389129639, "learning_rate": 2.4521263880123347e-06, "loss": 0.855, "step": 22137 }, { "epoch": 0.7800769928733141, "grad_norm": 2.4973278045654297, "learning_rate": 2.4513777943492622e-06, "loss": 3.0645, "step": 22138 }, { "epoch": 0.7801122298862725, "grad_norm": 3.2769107818603516, "learning_rate": 2.4506292990079574e-06, "loss": 2.5696, "step": 22139 }, { "epoch": 0.780147466899231, "grad_norm": 4.641809940338135, "learning_rate": 2.449880901998166e-06, "loss": 3.1695, "step": 22140 }, { "epoch": 0.7801827039121894, "grad_norm": 2.0465922355651855, "learning_rate": 2.449132603329638e-06, "loss": 1.1245, "step": 22141 }, { "epoch": 0.7802179409251477, "grad_norm": 3.500746011734009, "learning_rate": 2.4483844030121206e-06, "loss": 3.0502, "step": 22142 }, { "epoch": 0.7802531779381062, "grad_norm": 1.3188841342926025, "learning_rate": 2.4476363010553574e-06, "loss": 1.0234, "step": 22143 }, { "epoch": 0.7802884149510646, "grad_norm": 2.426621198654175, "learning_rate": 2.4468882974690932e-06, "loss": 0.8208, "step": 22144 }, { "epoch": 0.780323651964023, "grad_norm": 5.93417501449585, "learning_rate": 2.4461403922630746e-06, "loss": 6.4945, "step": 22145 }, { "epoch": 0.7803588889769815, "grad_norm": 2.090545892715454, "learning_rate": 2.445392585447034e-06, "loss": 0.8102, "step": 22146 }, { "epoch": 0.7803941259899398, "grad_norm": 4.519524097442627, "learning_rate": 2.4446448770307197e-06, "loss": 2.5076, "step": 22147 }, { "epoch": 0.7804293630028982, "grad_norm": 3.8183131217956543, "learning_rate": 2.4438972670238715e-06, "loss": 2.7244, "step": 22148 }, { "epoch": 0.7804646000158566, "grad_norm": 2.178769826889038, "learning_rate": 2.4431497554362204e-06, "loss": 0.8604, "step": 22149 }, { "epoch": 0.7804998370288151, "grad_norm": 1.1248180866241455, "learning_rate": 2.4424023422775036e-06, "loss": 0.9067, "step": 22150 }, { "epoch": 0.7805350740417735, "grad_norm": 1.1608049869537354, "learning_rate": 2.4416550275574615e-06, "loss": 1.0119, "step": 22151 }, { "epoch": 0.7805703110547318, "grad_norm": 4.293360233306885, "learning_rate": 2.4409078112858286e-06, "loss": 2.9747, "step": 22152 }, { "epoch": 0.7806055480676903, "grad_norm": 1.2218401432037354, "learning_rate": 2.4401606934723322e-06, "loss": 0.7788, "step": 22153 }, { "epoch": 0.7806407850806487, "grad_norm": 6.510761737823486, "learning_rate": 2.4394136741267027e-06, "loss": 5.4108, "step": 22154 }, { "epoch": 0.7806760220936071, "grad_norm": 0.7843003273010254, "learning_rate": 2.4386667532586784e-06, "loss": 0.6325, "step": 22155 }, { "epoch": 0.7807112591065656, "grad_norm": 3.110382318496704, "learning_rate": 2.43791993087798e-06, "loss": 2.8847, "step": 22156 }, { "epoch": 0.780746496119524, "grad_norm": 3.4306907653808594, "learning_rate": 2.4371732069943376e-06, "loss": 2.5874, "step": 22157 }, { "epoch": 0.7807817331324823, "grad_norm": 1.1754770278930664, "learning_rate": 2.436426581617478e-06, "loss": 1.0697, "step": 22158 }, { "epoch": 0.7808169701454408, "grad_norm": 3.5283865928649902, "learning_rate": 2.4356800547571247e-06, "loss": 4.0182, "step": 22159 }, { "epoch": 0.7808522071583992, "grad_norm": 7.769755840301514, "learning_rate": 2.4349336264230027e-06, "loss": 3.6252, "step": 22160 }, { "epoch": 0.7808874441713576, "grad_norm": 4.637844085693359, "learning_rate": 2.4341872966248346e-06, "loss": 2.7466, "step": 22161 }, { "epoch": 0.780922681184316, "grad_norm": 1.1521148681640625, "learning_rate": 2.4334410653723405e-06, "loss": 0.7777, "step": 22162 }, { "epoch": 0.7809579181972744, "grad_norm": 1.3644514083862305, "learning_rate": 2.4326949326752393e-06, "loss": 0.8107, "step": 22163 }, { "epoch": 0.7809931552102328, "grad_norm": 1.6369434595108032, "learning_rate": 2.431948898543255e-06, "loss": 1.0166, "step": 22164 }, { "epoch": 0.7810283922231912, "grad_norm": 1.0427660942077637, "learning_rate": 2.4312029629860934e-06, "loss": 1.0703, "step": 22165 }, { "epoch": 0.7810636292361497, "grad_norm": 3.987715244293213, "learning_rate": 2.43045712601348e-06, "loss": 2.9365, "step": 22166 }, { "epoch": 0.7810988662491081, "grad_norm": 2.2536027431488037, "learning_rate": 2.4297113876351275e-06, "loss": 0.8121, "step": 22167 }, { "epoch": 0.7811341032620664, "grad_norm": 0.9994810223579407, "learning_rate": 2.428965747860751e-06, "loss": 0.7888, "step": 22168 }, { "epoch": 0.7811693402750249, "grad_norm": 4.95305871963501, "learning_rate": 2.428220206700054e-06, "loss": 2.8407, "step": 22169 }, { "epoch": 0.7812045772879833, "grad_norm": 2.821167230606079, "learning_rate": 2.427474764162756e-06, "loss": 2.8767, "step": 22170 }, { "epoch": 0.7812398143009417, "grad_norm": 2.931462287902832, "learning_rate": 2.426729420258568e-06, "loss": 3.3754, "step": 22171 }, { "epoch": 0.7812750513139001, "grad_norm": 1.0928994417190552, "learning_rate": 2.42598417499719e-06, "loss": 1.0404, "step": 22172 }, { "epoch": 0.7813102883268586, "grad_norm": 3.147824764251709, "learning_rate": 2.4252390283883297e-06, "loss": 2.935, "step": 22173 }, { "epoch": 0.7813455253398169, "grad_norm": 0.8278436064720154, "learning_rate": 2.4244939804417023e-06, "loss": 0.8488, "step": 22174 }, { "epoch": 0.7813807623527753, "grad_norm": 4.354174613952637, "learning_rate": 2.423749031167003e-06, "loss": 4.1651, "step": 22175 }, { "epoch": 0.7814159993657338, "grad_norm": 2.832669734954834, "learning_rate": 2.423004180573938e-06, "loss": 2.8121, "step": 22176 }, { "epoch": 0.7814512363786922, "grad_norm": 1.7011864185333252, "learning_rate": 2.4222594286722078e-06, "loss": 0.6223, "step": 22177 }, { "epoch": 0.7814864733916506, "grad_norm": 2.9649765491485596, "learning_rate": 2.421514775471514e-06, "loss": 3.0304, "step": 22178 }, { "epoch": 0.781521710404609, "grad_norm": 4.548192977905273, "learning_rate": 2.4207702209815563e-06, "loss": 2.2211, "step": 22179 }, { "epoch": 0.7815569474175674, "grad_norm": 1.2189799547195435, "learning_rate": 2.420025765212031e-06, "loss": 0.9504, "step": 22180 }, { "epoch": 0.7815921844305258, "grad_norm": 1.3343837261199951, "learning_rate": 2.4192814081726357e-06, "loss": 0.7294, "step": 22181 }, { "epoch": 0.7816274214434842, "grad_norm": 2.3512935638427734, "learning_rate": 2.418537149873066e-06, "loss": 2.2555, "step": 22182 }, { "epoch": 0.7816626584564427, "grad_norm": 1.089647889137268, "learning_rate": 2.417792990323019e-06, "loss": 1.0038, "step": 22183 }, { "epoch": 0.781697895469401, "grad_norm": 2.181889772415161, "learning_rate": 2.4170489295321775e-06, "loss": 0.7875, "step": 22184 }, { "epoch": 0.7817331324823594, "grad_norm": 2.238755226135254, "learning_rate": 2.4163049675102433e-06, "loss": 0.7797, "step": 22185 }, { "epoch": 0.7817683694953179, "grad_norm": 1.2732182741165161, "learning_rate": 2.4155611042669024e-06, "loss": 0.703, "step": 22186 }, { "epoch": 0.7818036065082763, "grad_norm": 9.563278198242188, "learning_rate": 2.414817339811847e-06, "loss": 4.6814, "step": 22187 }, { "epoch": 0.7818388435212347, "grad_norm": 1.167748212814331, "learning_rate": 2.4140736741547566e-06, "loss": 1.11, "step": 22188 }, { "epoch": 0.7818740805341932, "grad_norm": 1.0365183353424072, "learning_rate": 2.413330107305325e-06, "loss": 0.957, "step": 22189 }, { "epoch": 0.7819093175471515, "grad_norm": 1.3276599645614624, "learning_rate": 2.4125866392732388e-06, "loss": 0.7224, "step": 22190 }, { "epoch": 0.7819445545601099, "grad_norm": 1.1008813381195068, "learning_rate": 2.4118432700681747e-06, "loss": 0.9696, "step": 22191 }, { "epoch": 0.7819797915730684, "grad_norm": 1.4616540670394897, "learning_rate": 2.4110999996998163e-06, "loss": 0.8433, "step": 22192 }, { "epoch": 0.7820150285860268, "grad_norm": 3.7414441108703613, "learning_rate": 2.4103568281778523e-06, "loss": 3.1938, "step": 22193 }, { "epoch": 0.7820502655989852, "grad_norm": 3.3761940002441406, "learning_rate": 2.4096137555119546e-06, "loss": 2.332, "step": 22194 }, { "epoch": 0.7820855026119435, "grad_norm": 3.588449716567993, "learning_rate": 2.408870781711805e-06, "loss": 2.753, "step": 22195 }, { "epoch": 0.782120739624902, "grad_norm": 6.797513484954834, "learning_rate": 2.40812790678708e-06, "loss": 5.3046, "step": 22196 }, { "epoch": 0.7821559766378604, "grad_norm": 4.443263053894043, "learning_rate": 2.407385130747457e-06, "loss": 3.1787, "step": 22197 }, { "epoch": 0.7821912136508188, "grad_norm": 1.116775631904602, "learning_rate": 2.406642453602609e-06, "loss": 1.1318, "step": 22198 }, { "epoch": 0.7822264506637773, "grad_norm": 6.205083847045898, "learning_rate": 2.4058998753622097e-06, "loss": 5.1255, "step": 22199 }, { "epoch": 0.7822616876767357, "grad_norm": 1.1549839973449707, "learning_rate": 2.405157396035933e-06, "loss": 1.3171, "step": 22200 }, { "epoch": 0.782296924689694, "grad_norm": 1.3347117900848389, "learning_rate": 2.404415015633448e-06, "loss": 0.9324, "step": 22201 }, { "epoch": 0.7823321617026525, "grad_norm": 1.1372318267822266, "learning_rate": 2.4036727341644272e-06, "loss": 0.9389, "step": 22202 }, { "epoch": 0.7823673987156109, "grad_norm": 6.6792731285095215, "learning_rate": 2.4029305516385327e-06, "loss": 5.8911, "step": 22203 }, { "epoch": 0.7824026357285693, "grad_norm": 3.0959770679473877, "learning_rate": 2.4021884680654373e-06, "loss": 2.457, "step": 22204 }, { "epoch": 0.7824378727415277, "grad_norm": 1.064197301864624, "learning_rate": 2.4014464834548048e-06, "loss": 1.1489, "step": 22205 }, { "epoch": 0.7824731097544861, "grad_norm": 3.7245562076568604, "learning_rate": 2.4007045978163024e-06, "loss": 2.9865, "step": 22206 }, { "epoch": 0.7825083467674445, "grad_norm": 4.1201276779174805, "learning_rate": 2.399962811159585e-06, "loss": 2.2569, "step": 22207 }, { "epoch": 0.7825435837804029, "grad_norm": 5.6061859130859375, "learning_rate": 2.399221123494323e-06, "loss": 4.5943, "step": 22208 }, { "epoch": 0.7825788207933614, "grad_norm": 4.234320640563965, "learning_rate": 2.398479534830177e-06, "loss": 2.8968, "step": 22209 }, { "epoch": 0.7826140578063198, "grad_norm": 3.840820074081421, "learning_rate": 2.3977380451768005e-06, "loss": 3.0387, "step": 22210 }, { "epoch": 0.7826492948192781, "grad_norm": 3.0543346405029297, "learning_rate": 2.3969966545438515e-06, "loss": 3.8119, "step": 22211 }, { "epoch": 0.7826845318322366, "grad_norm": 1.321620225906372, "learning_rate": 2.396255362940996e-06, "loss": 0.7188, "step": 22212 }, { "epoch": 0.782719768845195, "grad_norm": 0.9388036727905273, "learning_rate": 2.3955141703778793e-06, "loss": 0.7661, "step": 22213 }, { "epoch": 0.7827550058581534, "grad_norm": 3.9493350982666016, "learning_rate": 2.3947730768641588e-06, "loss": 4.3707, "step": 22214 }, { "epoch": 0.7827902428711118, "grad_norm": 4.13464879989624, "learning_rate": 2.3940320824094887e-06, "loss": 4.8872, "step": 22215 }, { "epoch": 0.7828254798840703, "grad_norm": 1.0553817749023438, "learning_rate": 2.3932911870235186e-06, "loss": 0.6694, "step": 22216 }, { "epoch": 0.7828607168970286, "grad_norm": 1.2857019901275635, "learning_rate": 2.3925503907159e-06, "loss": 0.7395, "step": 22217 }, { "epoch": 0.782895953909987, "grad_norm": 3.8239457607269287, "learning_rate": 2.391809693496282e-06, "loss": 4.7291, "step": 22218 }, { "epoch": 0.7829311909229455, "grad_norm": 1.1339774131774902, "learning_rate": 2.391069095374312e-06, "loss": 0.7927, "step": 22219 }, { "epoch": 0.7829664279359039, "grad_norm": 0.8630178570747375, "learning_rate": 2.3903285963596356e-06, "loss": 0.7431, "step": 22220 }, { "epoch": 0.7830016649488623, "grad_norm": 3.4710912704467773, "learning_rate": 2.389588196461902e-06, "loss": 2.7354, "step": 22221 }, { "epoch": 0.7830369019618207, "grad_norm": 6.410550594329834, "learning_rate": 2.3888478956907457e-06, "loss": 6.6593, "step": 22222 }, { "epoch": 0.7830721389747791, "grad_norm": 1.7447091341018677, "learning_rate": 2.3881076940558167e-06, "loss": 0.6994, "step": 22223 }, { "epoch": 0.7831073759877375, "grad_norm": 3.3104817867279053, "learning_rate": 2.387367591566756e-06, "loss": 2.3903, "step": 22224 }, { "epoch": 0.783142613000696, "grad_norm": 4.224648475646973, "learning_rate": 2.3866275882332045e-06, "loss": 2.8822, "step": 22225 }, { "epoch": 0.7831778500136544, "grad_norm": 1.4167994260787964, "learning_rate": 2.385887684064794e-06, "loss": 0.9833, "step": 22226 }, { "epoch": 0.7832130870266127, "grad_norm": 4.2098212242126465, "learning_rate": 2.385147879071167e-06, "loss": 2.8445, "step": 22227 }, { "epoch": 0.7832483240395711, "grad_norm": 2.7345151901245117, "learning_rate": 2.384408173261963e-06, "loss": 2.9963, "step": 22228 }, { "epoch": 0.7832835610525296, "grad_norm": 1.0407394170761108, "learning_rate": 2.3836685666468094e-06, "loss": 0.755, "step": 22229 }, { "epoch": 0.783318798065488, "grad_norm": 1.1546136140823364, "learning_rate": 2.3829290592353406e-06, "loss": 0.659, "step": 22230 }, { "epoch": 0.7833540350784464, "grad_norm": 5.153935432434082, "learning_rate": 2.3821896510371954e-06, "loss": 2.747, "step": 22231 }, { "epoch": 0.7833892720914049, "grad_norm": 7.332991600036621, "learning_rate": 2.3814503420619995e-06, "loss": 2.6916, "step": 22232 }, { "epoch": 0.7834245091043632, "grad_norm": 3.5801844596862793, "learning_rate": 2.3807111323193823e-06, "loss": 2.8689, "step": 22233 }, { "epoch": 0.7834597461173216, "grad_norm": 2.9244418144226074, "learning_rate": 2.3799720218189726e-06, "loss": 2.5594, "step": 22234 }, { "epoch": 0.7834949831302801, "grad_norm": 5.218806266784668, "learning_rate": 2.379233010570399e-06, "loss": 6.0157, "step": 22235 }, { "epoch": 0.7835302201432385, "grad_norm": 4.302206993103027, "learning_rate": 2.378494098583285e-06, "loss": 2.9762, "step": 22236 }, { "epoch": 0.7835654571561969, "grad_norm": 4.942309856414795, "learning_rate": 2.377755285867257e-06, "loss": 4.2377, "step": 22237 }, { "epoch": 0.7836006941691552, "grad_norm": 7.44042444229126, "learning_rate": 2.377016572431937e-06, "loss": 5.4862, "step": 22238 }, { "epoch": 0.7836359311821137, "grad_norm": 3.259401559829712, "learning_rate": 2.3762779582869467e-06, "loss": 2.3064, "step": 22239 }, { "epoch": 0.7836711681950721, "grad_norm": 4.8570709228515625, "learning_rate": 2.3755394434419077e-06, "loss": 4.8415, "step": 22240 }, { "epoch": 0.7837064052080305, "grad_norm": 1.212761402130127, "learning_rate": 2.3748010279064393e-06, "loss": 1.114, "step": 22241 }, { "epoch": 0.783741642220989, "grad_norm": 1.328037977218628, "learning_rate": 2.3740627116901584e-06, "loss": 0.6349, "step": 22242 }, { "epoch": 0.7837768792339473, "grad_norm": 6.3599090576171875, "learning_rate": 2.3733244948026814e-06, "loss": 5.0463, "step": 22243 }, { "epoch": 0.7838121162469057, "grad_norm": 4.23650598526001, "learning_rate": 2.3725863772536275e-06, "loss": 3.1231, "step": 22244 }, { "epoch": 0.7838473532598642, "grad_norm": 3.080219030380249, "learning_rate": 2.371848359052602e-06, "loss": 2.7553, "step": 22245 }, { "epoch": 0.7838825902728226, "grad_norm": 3.7975971698760986, "learning_rate": 2.3711104402092254e-06, "loss": 3.1356, "step": 22246 }, { "epoch": 0.783917827285781, "grad_norm": 0.9869764447212219, "learning_rate": 2.3703726207331112e-06, "loss": 1.1382, "step": 22247 }, { "epoch": 0.7839530642987393, "grad_norm": 6.297752857208252, "learning_rate": 2.3696349006338624e-06, "loss": 3.3253, "step": 22248 }, { "epoch": 0.7839883013116978, "grad_norm": 5.250089645385742, "learning_rate": 2.3688972799210876e-06, "loss": 3.5897, "step": 22249 }, { "epoch": 0.7840235383246562, "grad_norm": 3.1848740577697754, "learning_rate": 2.3681597586044037e-06, "loss": 3.536, "step": 22250 }, { "epoch": 0.7840587753376146, "grad_norm": 1.1551494598388672, "learning_rate": 2.367422336693408e-06, "loss": 0.963, "step": 22251 }, { "epoch": 0.7840940123505731, "grad_norm": 0.906036913394928, "learning_rate": 2.3666850141977096e-06, "loss": 0.7976, "step": 22252 }, { "epoch": 0.7841292493635315, "grad_norm": 2.7859930992126465, "learning_rate": 2.36594779112691e-06, "loss": 2.63, "step": 22253 }, { "epoch": 0.7841644863764898, "grad_norm": 1.7905840873718262, "learning_rate": 2.365210667490613e-06, "loss": 1.0376, "step": 22254 }, { "epoch": 0.7841997233894483, "grad_norm": 1.652233600616455, "learning_rate": 2.3644736432984205e-06, "loss": 0.8117, "step": 22255 }, { "epoch": 0.7842349604024067, "grad_norm": 1.0902836322784424, "learning_rate": 2.36373671855993e-06, "loss": 0.9208, "step": 22256 }, { "epoch": 0.7842701974153651, "grad_norm": 1.1385643482208252, "learning_rate": 2.3629998932847433e-06, "loss": 0.9035, "step": 22257 }, { "epoch": 0.7843054344283235, "grad_norm": 1.202512264251709, "learning_rate": 2.3622631674824536e-06, "loss": 1.1929, "step": 22258 }, { "epoch": 0.784340671441282, "grad_norm": 6.262712001800537, "learning_rate": 2.361526541162661e-06, "loss": 3.0215, "step": 22259 }, { "epoch": 0.7843759084542403, "grad_norm": 1.058159589767456, "learning_rate": 2.360790014334957e-06, "loss": 0.9778, "step": 22260 }, { "epoch": 0.7844111454671987, "grad_norm": 3.2281494140625, "learning_rate": 2.3600535870089357e-06, "loss": 2.9026, "step": 22261 }, { "epoch": 0.7844463824801572, "grad_norm": 3.295804738998413, "learning_rate": 2.3593172591941917e-06, "loss": 3.0367, "step": 22262 }, { "epoch": 0.7844816194931156, "grad_norm": 5.919328212738037, "learning_rate": 2.3585810309003145e-06, "loss": 3.1849, "step": 22263 }, { "epoch": 0.784516856506074, "grad_norm": 1.8379877805709839, "learning_rate": 2.3578449021368874e-06, "loss": 0.9805, "step": 22264 }, { "epoch": 0.7845520935190324, "grad_norm": 1.6139647960662842, "learning_rate": 2.357108872913507e-06, "loss": 1.1736, "step": 22265 }, { "epoch": 0.7845873305319908, "grad_norm": 5.793694972991943, "learning_rate": 2.356372943239761e-06, "loss": 5.3491, "step": 22266 }, { "epoch": 0.7846225675449492, "grad_norm": 4.093243598937988, "learning_rate": 2.3556371131252264e-06, "loss": 4.6042, "step": 22267 }, { "epoch": 0.7846578045579077, "grad_norm": 4.1988372802734375, "learning_rate": 2.354901382579491e-06, "loss": 4.2978, "step": 22268 }, { "epoch": 0.7846930415708661, "grad_norm": 1.1948565244674683, "learning_rate": 2.354165751612143e-06, "loss": 0.7262, "step": 22269 }, { "epoch": 0.7847282785838244, "grad_norm": 5.297008037567139, "learning_rate": 2.3534302202327584e-06, "loss": 4.9409, "step": 22270 }, { "epoch": 0.7847635155967828, "grad_norm": 5.473340034484863, "learning_rate": 2.352694788450919e-06, "loss": 4.7984, "step": 22271 }, { "epoch": 0.7847987526097413, "grad_norm": 5.215998649597168, "learning_rate": 2.3519594562762026e-06, "loss": 4.973, "step": 22272 }, { "epoch": 0.7848339896226997, "grad_norm": 0.9301957488059998, "learning_rate": 2.3512242237181903e-06, "loss": 0.9365, "step": 22273 }, { "epoch": 0.7848692266356581, "grad_norm": 3.5850555896759033, "learning_rate": 2.3504890907864563e-06, "loss": 3.0515, "step": 22274 }, { "epoch": 0.7849044636486165, "grad_norm": 2.7440829277038574, "learning_rate": 2.3497540574905765e-06, "loss": 3.0733, "step": 22275 }, { "epoch": 0.7849397006615749, "grad_norm": 3.1845862865448, "learning_rate": 2.349019123840124e-06, "loss": 2.0952, "step": 22276 }, { "epoch": 0.7849749376745333, "grad_norm": 3.442904233932495, "learning_rate": 2.3482842898446724e-06, "loss": 4.2987, "step": 22277 }, { "epoch": 0.7850101746874918, "grad_norm": 4.601625919342041, "learning_rate": 2.347549555513792e-06, "loss": 3.3474, "step": 22278 }, { "epoch": 0.7850454117004502, "grad_norm": 5.616323471069336, "learning_rate": 2.3468149208570534e-06, "loss": 3.354, "step": 22279 }, { "epoch": 0.7850806487134085, "grad_norm": 1.436671257019043, "learning_rate": 2.3460803858840264e-06, "loss": 0.9989, "step": 22280 }, { "epoch": 0.7851158857263669, "grad_norm": 5.584552764892578, "learning_rate": 2.3453459506042765e-06, "loss": 6.3415, "step": 22281 }, { "epoch": 0.7851511227393254, "grad_norm": 1.5369925498962402, "learning_rate": 2.3446116150273734e-06, "loss": 0.8895, "step": 22282 }, { "epoch": 0.7851863597522838, "grad_norm": 2.8485593795776367, "learning_rate": 2.343877379162874e-06, "loss": 2.8153, "step": 22283 }, { "epoch": 0.7852215967652422, "grad_norm": 4.064391613006592, "learning_rate": 2.3431432430203495e-06, "loss": 3.2042, "step": 22284 }, { "epoch": 0.7852568337782007, "grad_norm": 4.325857162475586, "learning_rate": 2.342409206609363e-06, "loss": 2.4875, "step": 22285 }, { "epoch": 0.785292070791159, "grad_norm": 1.184202790260315, "learning_rate": 2.3416752699394683e-06, "loss": 0.705, "step": 22286 }, { "epoch": 0.7853273078041174, "grad_norm": 2.5373780727386475, "learning_rate": 2.340941433020226e-06, "loss": 2.5133, "step": 22287 }, { "epoch": 0.7853625448170759, "grad_norm": 1.1234019994735718, "learning_rate": 2.340207695861203e-06, "loss": 0.9387, "step": 22288 }, { "epoch": 0.7853977818300343, "grad_norm": 1.2378300428390503, "learning_rate": 2.3394740584719476e-06, "loss": 0.9801, "step": 22289 }, { "epoch": 0.7854330188429927, "grad_norm": 2.2968084812164307, "learning_rate": 2.3387405208620184e-06, "loss": 0.7739, "step": 22290 }, { "epoch": 0.785468255855951, "grad_norm": 0.9374126195907593, "learning_rate": 2.3380070830409696e-06, "loss": 0.7079, "step": 22291 }, { "epoch": 0.7855034928689095, "grad_norm": 3.257404327392578, "learning_rate": 2.3372737450183548e-06, "loss": 3.0647, "step": 22292 }, { "epoch": 0.7855387298818679, "grad_norm": 1.1654748916625977, "learning_rate": 2.336540506803726e-06, "loss": 0.7674, "step": 22293 }, { "epoch": 0.7855739668948263, "grad_norm": 1.620606541633606, "learning_rate": 2.3358073684066328e-06, "loss": 0.7998, "step": 22294 }, { "epoch": 0.7856092039077848, "grad_norm": 6.020612716674805, "learning_rate": 2.3350743298366253e-06, "loss": 5.7146, "step": 22295 }, { "epoch": 0.7856444409207431, "grad_norm": 6.5038981437683105, "learning_rate": 2.3343413911032497e-06, "loss": 5.13, "step": 22296 }, { "epoch": 0.7856796779337015, "grad_norm": 5.954604148864746, "learning_rate": 2.3336085522160557e-06, "loss": 5.1789, "step": 22297 }, { "epoch": 0.78571491494666, "grad_norm": 4.0219316482543945, "learning_rate": 2.3328758131845854e-06, "loss": 3.0994, "step": 22298 }, { "epoch": 0.7857501519596184, "grad_norm": 5.197520732879639, "learning_rate": 2.3321431740183855e-06, "loss": 2.8056, "step": 22299 }, { "epoch": 0.7857853889725768, "grad_norm": 0.8797054290771484, "learning_rate": 2.331410634726996e-06, "loss": 0.8517, "step": 22300 }, { "epoch": 0.7858206259855353, "grad_norm": 5.144171714782715, "learning_rate": 2.3306781953199643e-06, "loss": 5.1806, "step": 22301 }, { "epoch": 0.7858558629984936, "grad_norm": 3.572829246520996, "learning_rate": 2.3299458558068187e-06, "loss": 2.4698, "step": 22302 }, { "epoch": 0.785891100011452, "grad_norm": 9.083890914916992, "learning_rate": 2.329213616197109e-06, "loss": 2.548, "step": 22303 }, { "epoch": 0.7859263370244104, "grad_norm": 5.073260307312012, "learning_rate": 2.3284814765003715e-06, "loss": 4.5252, "step": 22304 }, { "epoch": 0.7859615740373689, "grad_norm": 5.646954536437988, "learning_rate": 2.3277494367261376e-06, "loss": 2.8706, "step": 22305 }, { "epoch": 0.7859968110503273, "grad_norm": 1.1722819805145264, "learning_rate": 2.3270174968839423e-06, "loss": 1.1409, "step": 22306 }, { "epoch": 0.7860320480632856, "grad_norm": 1.3584394454956055, "learning_rate": 2.3262856569833258e-06, "loss": 0.7151, "step": 22307 }, { "epoch": 0.7860672850762441, "grad_norm": 3.2840399742126465, "learning_rate": 2.3255539170338147e-06, "loss": 2.8702, "step": 22308 }, { "epoch": 0.7861025220892025, "grad_norm": 3.174631118774414, "learning_rate": 2.32482227704494e-06, "loss": 2.5149, "step": 22309 }, { "epoch": 0.7861377591021609, "grad_norm": 5.037970066070557, "learning_rate": 2.3240907370262344e-06, "loss": 5.229, "step": 22310 }, { "epoch": 0.7861729961151194, "grad_norm": 4.249553680419922, "learning_rate": 2.323359296987223e-06, "loss": 4.3615, "step": 22311 }, { "epoch": 0.7862082331280777, "grad_norm": 1.3195793628692627, "learning_rate": 2.3226279569374353e-06, "loss": 0.7688, "step": 22312 }, { "epoch": 0.7862434701410361, "grad_norm": 0.8901321887969971, "learning_rate": 2.3218967168863972e-06, "loss": 0.9586, "step": 22313 }, { "epoch": 0.7862787071539945, "grad_norm": 1.1281185150146484, "learning_rate": 2.3211655768436312e-06, "loss": 0.9943, "step": 22314 }, { "epoch": 0.786313944166953, "grad_norm": 1.5197134017944336, "learning_rate": 2.320434536818662e-06, "loss": 0.805, "step": 22315 }, { "epoch": 0.7863491811799114, "grad_norm": 1.784825086593628, "learning_rate": 2.319703596821011e-06, "loss": 0.6584, "step": 22316 }, { "epoch": 0.7863844181928697, "grad_norm": 4.788491725921631, "learning_rate": 2.3189727568602005e-06, "loss": 4.7985, "step": 22317 }, { "epoch": 0.7864196552058282, "grad_norm": 3.4730262756347656, "learning_rate": 2.3182420169457466e-06, "loss": 2.8502, "step": 22318 }, { "epoch": 0.7864548922187866, "grad_norm": 5.014289379119873, "learning_rate": 2.3175113770871693e-06, "loss": 3.0062, "step": 22319 }, { "epoch": 0.786490129231745, "grad_norm": 2.42334246635437, "learning_rate": 2.316780837293988e-06, "loss": 0.8983, "step": 22320 }, { "epoch": 0.7865253662447035, "grad_norm": 0.9628473520278931, "learning_rate": 2.316050397575709e-06, "loss": 1.0369, "step": 22321 }, { "epoch": 0.7865606032576619, "grad_norm": 0.9374674558639526, "learning_rate": 2.3153200579418556e-06, "loss": 0.8277, "step": 22322 }, { "epoch": 0.7865958402706202, "grad_norm": 3.24353289604187, "learning_rate": 2.31458981840194e-06, "loss": 2.8616, "step": 22323 }, { "epoch": 0.7866310772835786, "grad_norm": 4.642093181610107, "learning_rate": 2.3138596789654677e-06, "loss": 4.4553, "step": 22324 }, { "epoch": 0.7866663142965371, "grad_norm": 2.9963576793670654, "learning_rate": 2.31312963964195e-06, "loss": 2.1702, "step": 22325 }, { "epoch": 0.7867015513094955, "grad_norm": 1.396614909172058, "learning_rate": 2.3123997004409025e-06, "loss": 1.0152, "step": 22326 }, { "epoch": 0.7867367883224539, "grad_norm": 7.089485168457031, "learning_rate": 2.311669861371826e-06, "loss": 3.2733, "step": 22327 }, { "epoch": 0.7867720253354124, "grad_norm": 6.710535526275635, "learning_rate": 2.3109401224442296e-06, "loss": 3.1572, "step": 22328 }, { "epoch": 0.7868072623483707, "grad_norm": 4.41563606262207, "learning_rate": 2.3102104836676165e-06, "loss": 4.4289, "step": 22329 }, { "epoch": 0.7868424993613291, "grad_norm": 3.6363275051116943, "learning_rate": 2.3094809450514923e-06, "loss": 2.8415, "step": 22330 }, { "epoch": 0.7868777363742876, "grad_norm": 2.946406602859497, "learning_rate": 2.3087515066053565e-06, "loss": 2.9676, "step": 22331 }, { "epoch": 0.786912973387246, "grad_norm": 1.2880516052246094, "learning_rate": 2.3080221683387138e-06, "loss": 0.8686, "step": 22332 }, { "epoch": 0.7869482104002044, "grad_norm": 1.677922248840332, "learning_rate": 2.3072929302610614e-06, "loss": 1.0654, "step": 22333 }, { "epoch": 0.7869834474131628, "grad_norm": 6.711081504821777, "learning_rate": 2.3065637923818996e-06, "loss": 4.6714, "step": 22334 }, { "epoch": 0.7870186844261212, "grad_norm": 1.5205823183059692, "learning_rate": 2.305834754710723e-06, "loss": 0.984, "step": 22335 }, { "epoch": 0.7870539214390796, "grad_norm": 5.309000492095947, "learning_rate": 2.3051058172570296e-06, "loss": 2.9942, "step": 22336 }, { "epoch": 0.787089158452038, "grad_norm": 6.2851996421813965, "learning_rate": 2.304376980030313e-06, "loss": 5.2459, "step": 22337 }, { "epoch": 0.7871243954649965, "grad_norm": 1.0702862739562988, "learning_rate": 2.3036482430400664e-06, "loss": 0.9433, "step": 22338 }, { "epoch": 0.7871596324779548, "grad_norm": 5.144563674926758, "learning_rate": 2.302919606295785e-06, "loss": 3.212, "step": 22339 }, { "epoch": 0.7871948694909132, "grad_norm": 0.9450212121009827, "learning_rate": 2.3021910698069495e-06, "loss": 0.7315, "step": 22340 }, { "epoch": 0.7872301065038717, "grad_norm": 4.639621734619141, "learning_rate": 2.301462633583059e-06, "loss": 4.5555, "step": 22341 }, { "epoch": 0.7872653435168301, "grad_norm": 1.3184823989868164, "learning_rate": 2.3007342976336023e-06, "loss": 0.7503, "step": 22342 }, { "epoch": 0.7873005805297885, "grad_norm": 4.635016441345215, "learning_rate": 2.3000060619680586e-06, "loss": 3.4105, "step": 22343 }, { "epoch": 0.787335817542747, "grad_norm": 1.253382682800293, "learning_rate": 2.299277926595913e-06, "loss": 0.9626, "step": 22344 }, { "epoch": 0.7873710545557053, "grad_norm": 20.596973419189453, "learning_rate": 2.29854989152666e-06, "loss": 3.3065, "step": 22345 }, { "epoch": 0.7874062915686637, "grad_norm": 2.626352548599243, "learning_rate": 2.2978219567697724e-06, "loss": 2.718, "step": 22346 }, { "epoch": 0.7874415285816221, "grad_norm": 3.168360471725464, "learning_rate": 2.297094122334735e-06, "loss": 2.6901, "step": 22347 }, { "epoch": 0.7874767655945806, "grad_norm": 1.0138143301010132, "learning_rate": 2.2963663882310284e-06, "loss": 0.8621, "step": 22348 }, { "epoch": 0.787512002607539, "grad_norm": 4.7618913650512695, "learning_rate": 2.2956387544681304e-06, "loss": 4.8822, "step": 22349 }, { "epoch": 0.7875472396204973, "grad_norm": 3.584550142288208, "learning_rate": 2.294911221055519e-06, "loss": 2.7772, "step": 22350 }, { "epoch": 0.7875824766334558, "grad_norm": 2.870384693145752, "learning_rate": 2.2941837880026707e-06, "loss": 2.988, "step": 22351 }, { "epoch": 0.7876177136464142, "grad_norm": 1.7149944305419922, "learning_rate": 2.29345645531906e-06, "loss": 0.9506, "step": 22352 }, { "epoch": 0.7876529506593726, "grad_norm": 1.0266011953353882, "learning_rate": 2.2927292230141606e-06, "loss": 0.8829, "step": 22353 }, { "epoch": 0.7876881876723311, "grad_norm": 1.114141583442688, "learning_rate": 2.292002091097445e-06, "loss": 0.7507, "step": 22354 }, { "epoch": 0.7877234246852894, "grad_norm": 1.9187490940093994, "learning_rate": 2.2912750595783827e-06, "loss": 0.9837, "step": 22355 }, { "epoch": 0.7877586616982478, "grad_norm": 5.170555591583252, "learning_rate": 2.2905481284664467e-06, "loss": 4.705, "step": 22356 }, { "epoch": 0.7877938987112062, "grad_norm": 1.9003881216049194, "learning_rate": 2.289821297771102e-06, "loss": 0.8785, "step": 22357 }, { "epoch": 0.7878291357241647, "grad_norm": 1.2777281999588013, "learning_rate": 2.2890945675018206e-06, "loss": 0.8292, "step": 22358 }, { "epoch": 0.7878643727371231, "grad_norm": 2.892669200897217, "learning_rate": 2.288367937668058e-06, "loss": 0.7022, "step": 22359 }, { "epoch": 0.7878996097500814, "grad_norm": 6.801629066467285, "learning_rate": 2.2876414082792886e-06, "loss": 5.0994, "step": 22360 }, { "epoch": 0.7879348467630399, "grad_norm": 3.209887742996216, "learning_rate": 2.2869149793449765e-06, "loss": 5.3574, "step": 22361 }, { "epoch": 0.7879700837759983, "grad_norm": 4.387744426727295, "learning_rate": 2.2861886508745746e-06, "loss": 3.0978, "step": 22362 }, { "epoch": 0.7880053207889567, "grad_norm": 1.05951726436615, "learning_rate": 2.285462422877546e-06, "loss": 1.1029, "step": 22363 }, { "epoch": 0.7880405578019152, "grad_norm": 2.4888644218444824, "learning_rate": 2.284736295363358e-06, "loss": 2.2339, "step": 22364 }, { "epoch": 0.7880757948148736, "grad_norm": 2.1614933013916016, "learning_rate": 2.284010268341459e-06, "loss": 2.5633, "step": 22365 }, { "epoch": 0.7881110318278319, "grad_norm": 8.878990173339844, "learning_rate": 2.283284341821309e-06, "loss": 3.3994, "step": 22366 }, { "epoch": 0.7881462688407904, "grad_norm": 2.3458974361419678, "learning_rate": 2.2825585158123642e-06, "loss": 2.7026, "step": 22367 }, { "epoch": 0.7881815058537488, "grad_norm": 4.82597017288208, "learning_rate": 2.2818327903240766e-06, "loss": 4.9941, "step": 22368 }, { "epoch": 0.7882167428667072, "grad_norm": 6.123423099517822, "learning_rate": 2.281107165365901e-06, "loss": 4.8714, "step": 22369 }, { "epoch": 0.7882519798796656, "grad_norm": 1.889001727104187, "learning_rate": 2.280381640947288e-06, "loss": 0.7331, "step": 22370 }, { "epoch": 0.788287216892624, "grad_norm": 1.227595329284668, "learning_rate": 2.279656217077687e-06, "loss": 0.9973, "step": 22371 }, { "epoch": 0.7883224539055824, "grad_norm": 1.5398728847503662, "learning_rate": 2.278930893766548e-06, "loss": 0.6387, "step": 22372 }, { "epoch": 0.7883576909185408, "grad_norm": 0.8937748670578003, "learning_rate": 2.278205671023316e-06, "loss": 0.9375, "step": 22373 }, { "epoch": 0.7883929279314993, "grad_norm": 1.171408772468567, "learning_rate": 2.2774805488574404e-06, "loss": 0.7211, "step": 22374 }, { "epoch": 0.7884281649444577, "grad_norm": 2.3365018367767334, "learning_rate": 2.2767555272783626e-06, "loss": 2.2944, "step": 22375 }, { "epoch": 0.788463401957416, "grad_norm": 1.4223432540893555, "learning_rate": 2.2760306062955295e-06, "loss": 0.979, "step": 22376 }, { "epoch": 0.7884986389703745, "grad_norm": 3.684929609298706, "learning_rate": 2.2753057859183834e-06, "loss": 2.9256, "step": 22377 }, { "epoch": 0.7885338759833329, "grad_norm": 1.6951284408569336, "learning_rate": 2.2745810661563605e-06, "loss": 1.041, "step": 22378 }, { "epoch": 0.7885691129962913, "grad_norm": 4.4079742431640625, "learning_rate": 2.2738564470189016e-06, "loss": 5.2427, "step": 22379 }, { "epoch": 0.7886043500092497, "grad_norm": 5.763249397277832, "learning_rate": 2.2731319285154506e-06, "loss": 5.2618, "step": 22380 }, { "epoch": 0.7886395870222082, "grad_norm": 7.218061923980713, "learning_rate": 2.2724075106554387e-06, "loss": 8.6268, "step": 22381 }, { "epoch": 0.7886748240351665, "grad_norm": 2.9985694885253906, "learning_rate": 2.271683193448304e-06, "loss": 2.717, "step": 22382 }, { "epoch": 0.7887100610481249, "grad_norm": 5.226414203643799, "learning_rate": 2.270958976903479e-06, "loss": 0.9537, "step": 22383 }, { "epoch": 0.7887452980610834, "grad_norm": 3.60148549079895, "learning_rate": 2.2702348610303983e-06, "loss": 4.9111, "step": 22384 }, { "epoch": 0.7887805350740418, "grad_norm": 1.8080939054489136, "learning_rate": 2.269510845838493e-06, "loss": 0.6657, "step": 22385 }, { "epoch": 0.7888157720870002, "grad_norm": 2.8439345359802246, "learning_rate": 2.2687869313371936e-06, "loss": 0.9741, "step": 22386 }, { "epoch": 0.7888510090999586, "grad_norm": 0.9681040048599243, "learning_rate": 2.268063117535929e-06, "loss": 0.8703, "step": 22387 }, { "epoch": 0.788886246112917, "grad_norm": 1.0512527227401733, "learning_rate": 2.267339404444128e-06, "loss": 0.7483, "step": 22388 }, { "epoch": 0.7889214831258754, "grad_norm": 1.5708814859390259, "learning_rate": 2.266615792071216e-06, "loss": 0.8732, "step": 22389 }, { "epoch": 0.7889567201388338, "grad_norm": 4.471501350402832, "learning_rate": 2.265892280426618e-06, "loss": 3.111, "step": 22390 }, { "epoch": 0.7889919571517923, "grad_norm": 2.47896671295166, "learning_rate": 2.265168869519758e-06, "loss": 0.8028, "step": 22391 }, { "epoch": 0.7890271941647506, "grad_norm": 4.730491638183594, "learning_rate": 2.264445559360059e-06, "loss": 5.042, "step": 22392 }, { "epoch": 0.789062431177709, "grad_norm": 6.077097415924072, "learning_rate": 2.263722349956945e-06, "loss": 3.3803, "step": 22393 }, { "epoch": 0.7890976681906675, "grad_norm": 1.860100269317627, "learning_rate": 2.2629992413198277e-06, "loss": 0.874, "step": 22394 }, { "epoch": 0.7891329052036259, "grad_norm": 1.1014739274978638, "learning_rate": 2.2622762334581326e-06, "loss": 1.1408, "step": 22395 }, { "epoch": 0.7891681422165843, "grad_norm": 1.0458790063858032, "learning_rate": 2.2615533263812783e-06, "loss": 0.6599, "step": 22396 }, { "epoch": 0.7892033792295428, "grad_norm": 4.82619571685791, "learning_rate": 2.2608305200986745e-06, "loss": 5.4762, "step": 22397 }, { "epoch": 0.7892386162425011, "grad_norm": 3.533689022064209, "learning_rate": 2.2601078146197375e-06, "loss": 3.6264, "step": 22398 }, { "epoch": 0.7892738532554595, "grad_norm": 3.3549981117248535, "learning_rate": 2.259385209953886e-06, "loss": 2.4974, "step": 22399 }, { "epoch": 0.789309090268418, "grad_norm": 4.032942295074463, "learning_rate": 2.258662706110527e-06, "loss": 3.1866, "step": 22400 }, { "epoch": 0.7893443272813764, "grad_norm": 4.678252696990967, "learning_rate": 2.257940303099071e-06, "loss": 2.6519, "step": 22401 }, { "epoch": 0.7893795642943348, "grad_norm": 3.9095993041992188, "learning_rate": 2.257218000928929e-06, "loss": 3.107, "step": 22402 }, { "epoch": 0.7894148013072931, "grad_norm": 1.0455334186553955, "learning_rate": 2.2564957996095093e-06, "loss": 0.8777, "step": 22403 }, { "epoch": 0.7894500383202516, "grad_norm": 4.4095635414123535, "learning_rate": 2.2557736991502175e-06, "loss": 4.6151, "step": 22404 }, { "epoch": 0.78948527533321, "grad_norm": 5.974982261657715, "learning_rate": 2.255051699560461e-06, "loss": 2.715, "step": 22405 }, { "epoch": 0.7895205123461684, "grad_norm": 1.2188903093338013, "learning_rate": 2.2543298008496417e-06, "loss": 0.7198, "step": 22406 }, { "epoch": 0.7895557493591269, "grad_norm": 5.077315807342529, "learning_rate": 2.2536080030271633e-06, "loss": 6.659, "step": 22407 }, { "epoch": 0.7895909863720852, "grad_norm": 1.1364200115203857, "learning_rate": 2.2528863061024276e-06, "loss": 1.0361, "step": 22408 }, { "epoch": 0.7896262233850436, "grad_norm": 0.8527932167053223, "learning_rate": 2.2521647100848345e-06, "loss": 0.9415, "step": 22409 }, { "epoch": 0.7896614603980021, "grad_norm": 3.8531723022460938, "learning_rate": 2.2514432149837826e-06, "loss": 3.1839, "step": 22410 }, { "epoch": 0.7896966974109605, "grad_norm": 3.4772491455078125, "learning_rate": 2.250721820808671e-06, "loss": 3.1405, "step": 22411 }, { "epoch": 0.7897319344239189, "grad_norm": 0.9680846929550171, "learning_rate": 2.2500005275688965e-06, "loss": 0.9012, "step": 22412 }, { "epoch": 0.7897671714368772, "grad_norm": 1.3664690256118774, "learning_rate": 2.2492793352738473e-06, "loss": 0.8148, "step": 22413 }, { "epoch": 0.7898024084498357, "grad_norm": 1.3106898069381714, "learning_rate": 2.2485582439329244e-06, "loss": 1.5837, "step": 22414 }, { "epoch": 0.7898376454627941, "grad_norm": 4.603273868560791, "learning_rate": 2.2478372535555217e-06, "loss": 3.0379, "step": 22415 }, { "epoch": 0.7898728824757525, "grad_norm": 4.486194610595703, "learning_rate": 2.247116364151022e-06, "loss": 3.2595, "step": 22416 }, { "epoch": 0.789908119488711, "grad_norm": 3.1189417839050293, "learning_rate": 2.246395575728817e-06, "loss": 2.4771, "step": 22417 }, { "epoch": 0.7899433565016694, "grad_norm": 2.6363790035247803, "learning_rate": 2.2456748882983025e-06, "loss": 2.9823, "step": 22418 }, { "epoch": 0.7899785935146277, "grad_norm": 1.2152553796768188, "learning_rate": 2.244954301868858e-06, "loss": 0.8605, "step": 22419 }, { "epoch": 0.7900138305275862, "grad_norm": 1.1180667877197266, "learning_rate": 2.244233816449871e-06, "loss": 0.7955, "step": 22420 }, { "epoch": 0.7900490675405446, "grad_norm": 5.541865348815918, "learning_rate": 2.243513432050727e-06, "loss": 3.657, "step": 22421 }, { "epoch": 0.790084304553503, "grad_norm": 0.8378959894180298, "learning_rate": 2.2427931486808087e-06, "loss": 0.796, "step": 22422 }, { "epoch": 0.7901195415664614, "grad_norm": 1.0752757787704468, "learning_rate": 2.2420729663494977e-06, "loss": 0.7503, "step": 22423 }, { "epoch": 0.7901547785794198, "grad_norm": 1.2658871412277222, "learning_rate": 2.2413528850661735e-06, "loss": 0.8659, "step": 22424 }, { "epoch": 0.7901900155923782, "grad_norm": 1.0216357707977295, "learning_rate": 2.2406329048402174e-06, "loss": 0.9505, "step": 22425 }, { "epoch": 0.7902252526053366, "grad_norm": 9.399277687072754, "learning_rate": 2.2399130256810065e-06, "loss": 4.5498, "step": 22426 }, { "epoch": 0.7902604896182951, "grad_norm": 6.78986120223999, "learning_rate": 2.239193247597916e-06, "loss": 5.2504, "step": 22427 }, { "epoch": 0.7902957266312535, "grad_norm": 4.256997585296631, "learning_rate": 2.2384735706003214e-06, "loss": 5.0206, "step": 22428 }, { "epoch": 0.7903309636442118, "grad_norm": 3.09783935546875, "learning_rate": 2.237753994697598e-06, "loss": 3.0569, "step": 22429 }, { "epoch": 0.7903662006571703, "grad_norm": 4.729599475860596, "learning_rate": 2.237034519899118e-06, "loss": 4.1948, "step": 22430 }, { "epoch": 0.7904014376701287, "grad_norm": 3.617128610610962, "learning_rate": 2.236315146214254e-06, "loss": 2.5228, "step": 22431 }, { "epoch": 0.7904366746830871, "grad_norm": 1.6162866353988647, "learning_rate": 2.23559587365237e-06, "loss": 0.9242, "step": 22432 }, { "epoch": 0.7904719116960456, "grad_norm": 1.2505565881729126, "learning_rate": 2.2348767022228404e-06, "loss": 0.784, "step": 22433 }, { "epoch": 0.790507148709004, "grad_norm": 4.2408623695373535, "learning_rate": 2.2341576319350345e-06, "loss": 4.4625, "step": 22434 }, { "epoch": 0.7905423857219623, "grad_norm": 3.151676654815674, "learning_rate": 2.2334386627983107e-06, "loss": 2.3683, "step": 22435 }, { "epoch": 0.7905776227349207, "grad_norm": 1.0523591041564941, "learning_rate": 2.232719794822036e-06, "loss": 0.809, "step": 22436 }, { "epoch": 0.7906128597478792, "grad_norm": 1.041451096534729, "learning_rate": 2.2320010280155802e-06, "loss": 1.0117, "step": 22437 }, { "epoch": 0.7906480967608376, "grad_norm": 8.231603622436523, "learning_rate": 2.231282362388297e-06, "loss": 2.7551, "step": 22438 }, { "epoch": 0.790683333773796, "grad_norm": 3.7421276569366455, "learning_rate": 2.230563797949552e-06, "loss": 3.0996, "step": 22439 }, { "epoch": 0.7907185707867544, "grad_norm": 1.0101488828659058, "learning_rate": 2.2298453347087003e-06, "loss": 1.1277, "step": 22440 }, { "epoch": 0.7907538077997128, "grad_norm": 4.915873050689697, "learning_rate": 2.2291269726751085e-06, "loss": 5.1836, "step": 22441 }, { "epoch": 0.7907890448126712, "grad_norm": 20.452133178710938, "learning_rate": 2.2284087118581245e-06, "loss": 2.6846, "step": 22442 }, { "epoch": 0.7908242818256297, "grad_norm": 3.5018317699432373, "learning_rate": 2.2276905522671067e-06, "loss": 3.0808, "step": 22443 }, { "epoch": 0.7908595188385881, "grad_norm": 2.2817156314849854, "learning_rate": 2.22697249391141e-06, "loss": 0.674, "step": 22444 }, { "epoch": 0.7908947558515464, "grad_norm": 3.3131070137023926, "learning_rate": 2.2262545368003864e-06, "loss": 0.8778, "step": 22445 }, { "epoch": 0.7909299928645048, "grad_norm": 3.545424222946167, "learning_rate": 2.225536680943388e-06, "loss": 3.1192, "step": 22446 }, { "epoch": 0.7909652298774633, "grad_norm": 4.07384729385376, "learning_rate": 2.2248189263497646e-06, "loss": 2.982, "step": 22447 }, { "epoch": 0.7910004668904217, "grad_norm": 7.102238178253174, "learning_rate": 2.2241012730288646e-06, "loss": 2.8782, "step": 22448 }, { "epoch": 0.7910357039033801, "grad_norm": 3.735358953475952, "learning_rate": 2.2233837209900367e-06, "loss": 2.9399, "step": 22449 }, { "epoch": 0.7910709409163386, "grad_norm": 1.4325684309005737, "learning_rate": 2.222666270242628e-06, "loss": 1.1777, "step": 22450 }, { "epoch": 0.7911061779292969, "grad_norm": 4.721086502075195, "learning_rate": 2.221948920795978e-06, "loss": 5.1325, "step": 22451 }, { "epoch": 0.7911414149422553, "grad_norm": 3.7899632453918457, "learning_rate": 2.221231672659435e-06, "loss": 3.2401, "step": 22452 }, { "epoch": 0.7911766519552138, "grad_norm": 4.909945487976074, "learning_rate": 2.2205145258423444e-06, "loss": 4.9134, "step": 22453 }, { "epoch": 0.7912118889681722, "grad_norm": 1.2749437093734741, "learning_rate": 2.21979748035404e-06, "loss": 0.676, "step": 22454 }, { "epoch": 0.7912471259811306, "grad_norm": 1.5047749280929565, "learning_rate": 2.2190805362038615e-06, "loss": 0.7501, "step": 22455 }, { "epoch": 0.7912823629940889, "grad_norm": 0.985966682434082, "learning_rate": 2.2183636934011554e-06, "loss": 0.8682, "step": 22456 }, { "epoch": 0.7913176000070474, "grad_norm": 4.358902454376221, "learning_rate": 2.217646951955251e-06, "loss": 4.7819, "step": 22457 }, { "epoch": 0.7913528370200058, "grad_norm": 5.330896377563477, "learning_rate": 2.216930311875486e-06, "loss": 4.3587, "step": 22458 }, { "epoch": 0.7913880740329642, "grad_norm": 0.8965469002723694, "learning_rate": 2.216213773171193e-06, "loss": 0.5711, "step": 22459 }, { "epoch": 0.7914233110459227, "grad_norm": 1.0402421951293945, "learning_rate": 2.2154973358517128e-06, "loss": 0.6972, "step": 22460 }, { "epoch": 0.791458548058881, "grad_norm": 7.022422790527344, "learning_rate": 2.214780999926368e-06, "loss": 2.9652, "step": 22461 }, { "epoch": 0.7914937850718394, "grad_norm": 2.3589868545532227, "learning_rate": 2.214064765404492e-06, "loss": 2.3853, "step": 22462 }, { "epoch": 0.7915290220847979, "grad_norm": 4.695982456207275, "learning_rate": 2.2133486322954145e-06, "loss": 4.353, "step": 22463 }, { "epoch": 0.7915642590977563, "grad_norm": 1.1748299598693848, "learning_rate": 2.2126326006084643e-06, "loss": 0.8227, "step": 22464 }, { "epoch": 0.7915994961107147, "grad_norm": 1.5723228454589844, "learning_rate": 2.211916670352965e-06, "loss": 0.7942, "step": 22465 }, { "epoch": 0.791634733123673, "grad_norm": 1.7029136419296265, "learning_rate": 2.2112008415382425e-06, "loss": 0.8719, "step": 22466 }, { "epoch": 0.7916699701366315, "grad_norm": 5.8054351806640625, "learning_rate": 2.210485114173623e-06, "loss": 4.7541, "step": 22467 }, { "epoch": 0.7917052071495899, "grad_norm": 2.446246862411499, "learning_rate": 2.2097694882684263e-06, "loss": 2.5231, "step": 22468 }, { "epoch": 0.7917404441625483, "grad_norm": 1.3514885902404785, "learning_rate": 2.2090539638319762e-06, "loss": 0.7333, "step": 22469 }, { "epoch": 0.7917756811755068, "grad_norm": 1.578772783279419, "learning_rate": 2.208338540873586e-06, "loss": 0.6561, "step": 22470 }, { "epoch": 0.7918109181884652, "grad_norm": 1.2390164136886597, "learning_rate": 2.2076232194025815e-06, "loss": 0.8294, "step": 22471 }, { "epoch": 0.7918461552014235, "grad_norm": 4.830935478210449, "learning_rate": 2.2069079994282792e-06, "loss": 4.3913, "step": 22472 }, { "epoch": 0.791881392214382, "grad_norm": 6.478458404541016, "learning_rate": 2.206192880959991e-06, "loss": 7.2764, "step": 22473 }, { "epoch": 0.7919166292273404, "grad_norm": 4.993150234222412, "learning_rate": 2.205477864007031e-06, "loss": 5.1307, "step": 22474 }, { "epoch": 0.7919518662402988, "grad_norm": 3.5734593868255615, "learning_rate": 2.204762948578719e-06, "loss": 3.0307, "step": 22475 }, { "epoch": 0.7919871032532573, "grad_norm": 1.0535478591918945, "learning_rate": 2.20404813468436e-06, "loss": 0.8487, "step": 22476 }, { "epoch": 0.7920223402662157, "grad_norm": 2.545266628265381, "learning_rate": 2.203333422333268e-06, "loss": 2.751, "step": 22477 }, { "epoch": 0.792057577279174, "grad_norm": 1.2398196458816528, "learning_rate": 2.202618811534748e-06, "loss": 0.7781, "step": 22478 }, { "epoch": 0.7920928142921324, "grad_norm": 1.1811147928237915, "learning_rate": 2.201904302298117e-06, "loss": 0.6597, "step": 22479 }, { "epoch": 0.7921280513050909, "grad_norm": 0.9985816478729248, "learning_rate": 2.201189894632673e-06, "loss": 0.9119, "step": 22480 }, { "epoch": 0.7921632883180493, "grad_norm": 1.2472916841506958, "learning_rate": 2.2004755885477223e-06, "loss": 0.9698, "step": 22481 }, { "epoch": 0.7921985253310077, "grad_norm": 3.3164241313934326, "learning_rate": 2.199761384052572e-06, "loss": 3.0891, "step": 22482 }, { "epoch": 0.7922337623439661, "grad_norm": 1.2233662605285645, "learning_rate": 2.199047281156523e-06, "loss": 0.8593, "step": 22483 }, { "epoch": 0.7922689993569245, "grad_norm": 2.3790581226348877, "learning_rate": 2.1983332798688782e-06, "loss": 2.8098, "step": 22484 }, { "epoch": 0.7923042363698829, "grad_norm": 1.1116548776626587, "learning_rate": 2.1976193801989354e-06, "loss": 0.7717, "step": 22485 }, { "epoch": 0.7923394733828414, "grad_norm": 3.728839635848999, "learning_rate": 2.1969055821559936e-06, "loss": 3.3614, "step": 22486 }, { "epoch": 0.7923747103957998, "grad_norm": 1.6710518598556519, "learning_rate": 2.19619188574935e-06, "loss": 0.6368, "step": 22487 }, { "epoch": 0.7924099474087581, "grad_norm": 3.802915096282959, "learning_rate": 2.1954782909883056e-06, "loss": 2.8729, "step": 22488 }, { "epoch": 0.7924451844217165, "grad_norm": 2.281848430633545, "learning_rate": 2.1947647978821427e-06, "loss": 1.1224, "step": 22489 }, { "epoch": 0.792480421434675, "grad_norm": 1.5835812091827393, "learning_rate": 2.1940514064401674e-06, "loss": 0.6969, "step": 22490 }, { "epoch": 0.7925156584476334, "grad_norm": 6.9912309646606445, "learning_rate": 2.1933381166716683e-06, "loss": 2.6798, "step": 22491 }, { "epoch": 0.7925508954605918, "grad_norm": 0.9889402985572815, "learning_rate": 2.1926249285859323e-06, "loss": 0.8412, "step": 22492 }, { "epoch": 0.7925861324735503, "grad_norm": 2.819570302963257, "learning_rate": 2.191911842192249e-06, "loss": 2.997, "step": 22493 }, { "epoch": 0.7926213694865086, "grad_norm": 1.2742670774459839, "learning_rate": 2.191198857499913e-06, "loss": 0.8748, "step": 22494 }, { "epoch": 0.792656606499467, "grad_norm": 2.3387930393218994, "learning_rate": 2.190485974518205e-06, "loss": 0.7339, "step": 22495 }, { "epoch": 0.7926918435124255, "grad_norm": 2.649467945098877, "learning_rate": 2.1897731932564115e-06, "loss": 0.9687, "step": 22496 }, { "epoch": 0.7927270805253839, "grad_norm": 2.0906221866607666, "learning_rate": 2.189060513723814e-06, "loss": 0.7981, "step": 22497 }, { "epoch": 0.7927623175383423, "grad_norm": 1.6338541507720947, "learning_rate": 2.1883479359297044e-06, "loss": 0.8568, "step": 22498 }, { "epoch": 0.7927975545513006, "grad_norm": 2.567892551422119, "learning_rate": 2.1876354598833537e-06, "loss": 2.9733, "step": 22499 }, { "epoch": 0.7928327915642591, "grad_norm": 3.1653330326080322, "learning_rate": 2.1869230855940483e-06, "loss": 2.6987, "step": 22500 }, { "epoch": 0.7928680285772175, "grad_norm": 1.3880136013031006, "learning_rate": 2.1862108130710636e-06, "loss": 1.0081, "step": 22501 }, { "epoch": 0.7929032655901759, "grad_norm": 1.2151044607162476, "learning_rate": 2.185498642323678e-06, "loss": 0.8112, "step": 22502 }, { "epoch": 0.7929385026031344, "grad_norm": 4.323856353759766, "learning_rate": 2.1847865733611695e-06, "loss": 2.8945, "step": 22503 }, { "epoch": 0.7929737396160927, "grad_norm": 3.3734285831451416, "learning_rate": 2.1840746061928097e-06, "loss": 3.0226, "step": 22504 }, { "epoch": 0.7930089766290511, "grad_norm": 1.2617805004119873, "learning_rate": 2.1833627408278744e-06, "loss": 1.0525, "step": 22505 }, { "epoch": 0.7930442136420096, "grad_norm": 2.3827524185180664, "learning_rate": 2.1826509772756353e-06, "loss": 2.7406, "step": 22506 }, { "epoch": 0.793079450654968, "grad_norm": 3.8157124519348145, "learning_rate": 2.181939315545366e-06, "loss": 2.4261, "step": 22507 }, { "epoch": 0.7931146876679264, "grad_norm": 3.9003348350524902, "learning_rate": 2.181227755646328e-06, "loss": 4.8634, "step": 22508 }, { "epoch": 0.7931499246808849, "grad_norm": 1.6359206438064575, "learning_rate": 2.180516297587796e-06, "loss": 0.9463, "step": 22509 }, { "epoch": 0.7931851616938432, "grad_norm": 2.3065648078918457, "learning_rate": 2.179804941379038e-06, "loss": 0.9967, "step": 22510 }, { "epoch": 0.7932203987068016, "grad_norm": 3.249870538711548, "learning_rate": 2.1790936870293146e-06, "loss": 3.0495, "step": 22511 }, { "epoch": 0.79325563571976, "grad_norm": 1.3442655801773071, "learning_rate": 2.1783825345478894e-06, "loss": 0.9885, "step": 22512 }, { "epoch": 0.7932908727327185, "grad_norm": 1.7745622396469116, "learning_rate": 2.1776714839440304e-06, "loss": 0.9753, "step": 22513 }, { "epoch": 0.7933261097456769, "grad_norm": 1.3094533681869507, "learning_rate": 2.1769605352269995e-06, "loss": 0.7341, "step": 22514 }, { "epoch": 0.7933613467586352, "grad_norm": 9.126338005065918, "learning_rate": 2.1762496884060513e-06, "loss": 4.6104, "step": 22515 }, { "epoch": 0.7933965837715937, "grad_norm": 3.729397773742676, "learning_rate": 2.175538943490445e-06, "loss": 2.9101, "step": 22516 }, { "epoch": 0.7934318207845521, "grad_norm": 1.23753023147583, "learning_rate": 2.1748283004894466e-06, "loss": 0.7458, "step": 22517 }, { "epoch": 0.7934670577975105, "grad_norm": 2.7654130458831787, "learning_rate": 2.1741177594123033e-06, "loss": 2.7775, "step": 22518 }, { "epoch": 0.793502294810469, "grad_norm": 2.920313596725464, "learning_rate": 2.173407320268273e-06, "loss": 2.7911, "step": 22519 }, { "epoch": 0.7935375318234273, "grad_norm": 1.0066603422164917, "learning_rate": 2.172696983066609e-06, "loss": 0.9076, "step": 22520 }, { "epoch": 0.7935727688363857, "grad_norm": 2.2848052978515625, "learning_rate": 2.1719867478165635e-06, "loss": 0.7351, "step": 22521 }, { "epoch": 0.7936080058493441, "grad_norm": 4.9936017990112305, "learning_rate": 2.171276614527389e-06, "loss": 5.0232, "step": 22522 }, { "epoch": 0.7936432428623026, "grad_norm": 1.5057343244552612, "learning_rate": 2.1705665832083324e-06, "loss": 0.791, "step": 22523 }, { "epoch": 0.793678479875261, "grad_norm": 4.025954246520996, "learning_rate": 2.1698566538686437e-06, "loss": 2.6681, "step": 22524 }, { "epoch": 0.7937137168882193, "grad_norm": 0.8062583804130554, "learning_rate": 2.169146826517571e-06, "loss": 0.8589, "step": 22525 }, { "epoch": 0.7937489539011778, "grad_norm": 2.7005531787872314, "learning_rate": 2.168437101164359e-06, "loss": 2.364, "step": 22526 }, { "epoch": 0.7937841909141362, "grad_norm": 4.241428852081299, "learning_rate": 2.1677274778182466e-06, "loss": 3.0722, "step": 22527 }, { "epoch": 0.7938194279270946, "grad_norm": 1.058201551437378, "learning_rate": 2.1670179564884853e-06, "loss": 0.8709, "step": 22528 }, { "epoch": 0.7938546649400531, "grad_norm": 3.0577611923217773, "learning_rate": 2.166308537184314e-06, "loss": 2.8858, "step": 22529 }, { "epoch": 0.7938899019530115, "grad_norm": 1.1329845190048218, "learning_rate": 2.16559921991497e-06, "loss": 0.9164, "step": 22530 }, { "epoch": 0.7939251389659698, "grad_norm": 3.1398355960845947, "learning_rate": 2.1648900046896915e-06, "loss": 3.0401, "step": 22531 }, { "epoch": 0.7939603759789282, "grad_norm": 1.2501916885375977, "learning_rate": 2.1641808915177207e-06, "loss": 0.7434, "step": 22532 }, { "epoch": 0.7939956129918867, "grad_norm": 1.2501916885375977, "learning_rate": 2.1641808915177207e-06, "loss": 2.7252, "step": 22533 }, { "epoch": 0.7940308500048451, "grad_norm": 1.9640640020370483, "learning_rate": 2.1634718804082955e-06, "loss": 0.964, "step": 22534 }, { "epoch": 0.7940660870178035, "grad_norm": 1.0349514484405518, "learning_rate": 2.1627629713706446e-06, "loss": 0.7874, "step": 22535 }, { "epoch": 0.794101324030762, "grad_norm": 6.923484802246094, "learning_rate": 2.1620541644140002e-06, "loss": 7.5628, "step": 22536 }, { "epoch": 0.7941365610437203, "grad_norm": 1.5668824911117554, "learning_rate": 2.1613454595476057e-06, "loss": 0.9658, "step": 22537 }, { "epoch": 0.7941717980566787, "grad_norm": 4.4474005699157715, "learning_rate": 2.1606368567806814e-06, "loss": 3.074, "step": 22538 }, { "epoch": 0.7942070350696372, "grad_norm": 0.9852759838104248, "learning_rate": 2.1599283561224604e-06, "loss": 0.8124, "step": 22539 }, { "epoch": 0.7942422720825956, "grad_norm": 0.9011654853820801, "learning_rate": 2.159219957582173e-06, "loss": 0.8826, "step": 22540 }, { "epoch": 0.794277509095554, "grad_norm": 4.9935688972473145, "learning_rate": 2.158511661169043e-06, "loss": 5.5119, "step": 22541 }, { "epoch": 0.7943127461085124, "grad_norm": 1.9813159704208374, "learning_rate": 2.1578034668922975e-06, "loss": 1.0334, "step": 22542 }, { "epoch": 0.7943479831214708, "grad_norm": 5.801140308380127, "learning_rate": 2.15709537476116e-06, "loss": 3.1749, "step": 22543 }, { "epoch": 0.7943832201344292, "grad_norm": 1.3459047079086304, "learning_rate": 2.156387384784856e-06, "loss": 1.0565, "step": 22544 }, { "epoch": 0.7944184571473876, "grad_norm": 2.4904227256774902, "learning_rate": 2.155679496972605e-06, "loss": 2.8304, "step": 22545 }, { "epoch": 0.7944536941603461, "grad_norm": 2.9327197074890137, "learning_rate": 2.15497171133363e-06, "loss": 2.7254, "step": 22546 }, { "epoch": 0.7944889311733044, "grad_norm": 3.893326759338379, "learning_rate": 2.154264027877142e-06, "loss": 3.0891, "step": 22547 }, { "epoch": 0.7945241681862628, "grad_norm": 6.519801616668701, "learning_rate": 2.153556446612367e-06, "loss": 5.0628, "step": 22548 }, { "epoch": 0.7945594051992213, "grad_norm": 3.678997039794922, "learning_rate": 2.152848967548522e-06, "loss": 3.2534, "step": 22549 }, { "epoch": 0.7945946422121797, "grad_norm": 1.3061580657958984, "learning_rate": 2.152141590694816e-06, "loss": 0.8808, "step": 22550 }, { "epoch": 0.7946298792251381, "grad_norm": 1.2176059484481812, "learning_rate": 2.1514343160604636e-06, "loss": 0.7495, "step": 22551 }, { "epoch": 0.7946651162380965, "grad_norm": 7.492985248565674, "learning_rate": 2.1507271436546806e-06, "loss": 6.534, "step": 22552 }, { "epoch": 0.7947003532510549, "grad_norm": 5.879065036773682, "learning_rate": 2.1500200734866793e-06, "loss": 3.5416, "step": 22553 }, { "epoch": 0.7947355902640133, "grad_norm": 4.567521095275879, "learning_rate": 2.1493131055656633e-06, "loss": 3.9324, "step": 22554 }, { "epoch": 0.7947708272769717, "grad_norm": 1.3255647420883179, "learning_rate": 2.1486062399008423e-06, "loss": 0.9051, "step": 22555 }, { "epoch": 0.7948060642899302, "grad_norm": 1.046423316001892, "learning_rate": 2.1478994765014304e-06, "loss": 0.5459, "step": 22556 }, { "epoch": 0.7948413013028885, "grad_norm": 3.5189759731292725, "learning_rate": 2.1471928153766254e-06, "loss": 2.9814, "step": 22557 }, { "epoch": 0.7948765383158469, "grad_norm": 0.9381722211837769, "learning_rate": 2.146486256535634e-06, "loss": 0.7675, "step": 22558 }, { "epoch": 0.7949117753288054, "grad_norm": 2.1982977390289307, "learning_rate": 2.1457797999876594e-06, "loss": 0.7209, "step": 22559 }, { "epoch": 0.7949470123417638, "grad_norm": 4.044241905212402, "learning_rate": 2.145073445741903e-06, "loss": 2.9825, "step": 22560 }, { "epoch": 0.7949822493547222, "grad_norm": 1.0212129354476929, "learning_rate": 2.144367193807566e-06, "loss": 0.9523, "step": 22561 }, { "epoch": 0.7950174863676807, "grad_norm": 5.054717540740967, "learning_rate": 2.1436610441938478e-06, "loss": 3.1921, "step": 22562 }, { "epoch": 0.795052723380639, "grad_norm": 1.110159993171692, "learning_rate": 2.142954996909944e-06, "loss": 1.0346, "step": 22563 }, { "epoch": 0.7950879603935974, "grad_norm": 3.474283218383789, "learning_rate": 2.1422490519650528e-06, "loss": 2.97, "step": 22564 }, { "epoch": 0.7951231974065558, "grad_norm": 3.8245151042938232, "learning_rate": 2.141543209368371e-06, "loss": 3.2681, "step": 22565 }, { "epoch": 0.7951584344195143, "grad_norm": 1.0811792612075806, "learning_rate": 2.1408374691290846e-06, "loss": 0.9826, "step": 22566 }, { "epoch": 0.7951936714324727, "grad_norm": 3.890632390975952, "learning_rate": 2.1401318312563935e-06, "loss": 2.3086, "step": 22567 }, { "epoch": 0.795228908445431, "grad_norm": 0.9504929780960083, "learning_rate": 2.1394262957594904e-06, "loss": 0.9609, "step": 22568 }, { "epoch": 0.7952641454583895, "grad_norm": 1.3176378011703491, "learning_rate": 2.1387208626475576e-06, "loss": 0.8121, "step": 22569 }, { "epoch": 0.7952993824713479, "grad_norm": 4.452112197875977, "learning_rate": 2.1380155319297836e-06, "loss": 3.0714, "step": 22570 }, { "epoch": 0.7953346194843063, "grad_norm": 5.96500301361084, "learning_rate": 2.137310303615362e-06, "loss": 6.8434, "step": 22571 }, { "epoch": 0.7953698564972648, "grad_norm": 2.750671148300171, "learning_rate": 2.1366051777134788e-06, "loss": 1.0528, "step": 22572 }, { "epoch": 0.7954050935102231, "grad_norm": 3.714553117752075, "learning_rate": 2.1359001542333115e-06, "loss": 2.9396, "step": 22573 }, { "epoch": 0.7954403305231815, "grad_norm": 5.646120548248291, "learning_rate": 2.135195233184043e-06, "loss": 2.6288, "step": 22574 }, { "epoch": 0.79547556753614, "grad_norm": 1.5487838983535767, "learning_rate": 2.1344904145748645e-06, "loss": 0.7637, "step": 22575 }, { "epoch": 0.7955108045490984, "grad_norm": 1.7061817646026611, "learning_rate": 2.1337856984149473e-06, "loss": 0.9182, "step": 22576 }, { "epoch": 0.7955460415620568, "grad_norm": 1.2450530529022217, "learning_rate": 2.133081084713473e-06, "loss": 0.7841, "step": 22577 }, { "epoch": 0.7955812785750151, "grad_norm": 4.984751224517822, "learning_rate": 2.1323765734796207e-06, "loss": 5.1127, "step": 22578 }, { "epoch": 0.7956165155879736, "grad_norm": 3.5071218013763428, "learning_rate": 2.1316721647225645e-06, "loss": 4.6004, "step": 22579 }, { "epoch": 0.795651752600932, "grad_norm": 1.213621735572815, "learning_rate": 2.1309678584514814e-06, "loss": 0.6673, "step": 22580 }, { "epoch": 0.7956869896138904, "grad_norm": 4.154074192047119, "learning_rate": 2.130263654675545e-06, "loss": 2.8418, "step": 22581 }, { "epoch": 0.7957222266268489, "grad_norm": 1.2071887254714966, "learning_rate": 2.129559553403926e-06, "loss": 1.146, "step": 22582 }, { "epoch": 0.7957574636398073, "grad_norm": 3.060354232788086, "learning_rate": 2.1288555546457965e-06, "loss": 2.4761, "step": 22583 }, { "epoch": 0.7957927006527656, "grad_norm": 4.273438453674316, "learning_rate": 2.128151658410329e-06, "loss": 4.8825, "step": 22584 }, { "epoch": 0.7958279376657241, "grad_norm": 3.608827590942383, "learning_rate": 2.1274478647066823e-06, "loss": 3.2463, "step": 22585 }, { "epoch": 0.7958631746786825, "grad_norm": 5.815035343170166, "learning_rate": 2.1267441735440344e-06, "loss": 6.7697, "step": 22586 }, { "epoch": 0.7958984116916409, "grad_norm": 3.5011327266693115, "learning_rate": 2.1260405849315447e-06, "loss": 2.9019, "step": 22587 }, { "epoch": 0.7959336487045993, "grad_norm": 1.6925991773605347, "learning_rate": 2.125337098878384e-06, "loss": 0.9129, "step": 22588 }, { "epoch": 0.7959688857175578, "grad_norm": 1.6646053791046143, "learning_rate": 2.1246337153937034e-06, "loss": 1.2736, "step": 22589 }, { "epoch": 0.7960041227305161, "grad_norm": 1.0628433227539062, "learning_rate": 2.123930434486676e-06, "loss": 0.9038, "step": 22590 }, { "epoch": 0.7960393597434745, "grad_norm": 0.800437867641449, "learning_rate": 2.1232272561664602e-06, "loss": 0.8507, "step": 22591 }, { "epoch": 0.796074596756433, "grad_norm": 3.238417863845825, "learning_rate": 2.1225241804422093e-06, "loss": 2.8049, "step": 22592 }, { "epoch": 0.7961098337693914, "grad_norm": 3.3982179164886475, "learning_rate": 2.1218212073230826e-06, "loss": 2.3528, "step": 22593 }, { "epoch": 0.7961450707823498, "grad_norm": 7.500579833984375, "learning_rate": 2.121118336818243e-06, "loss": 3.0664, "step": 22594 }, { "epoch": 0.7961803077953082, "grad_norm": 3.198153257369995, "learning_rate": 2.1204155689368377e-06, "loss": 0.6472, "step": 22595 }, { "epoch": 0.7962155448082666, "grad_norm": 5.155540466308594, "learning_rate": 2.1197129036880247e-06, "loss": 4.3538, "step": 22596 }, { "epoch": 0.796250781821225, "grad_norm": 5.750617504119873, "learning_rate": 2.1190103410809536e-06, "loss": 2.8239, "step": 22597 }, { "epoch": 0.7962860188341834, "grad_norm": 0.9158186316490173, "learning_rate": 2.118307881124777e-06, "loss": 0.7939, "step": 22598 }, { "epoch": 0.7963212558471419, "grad_norm": 1.0994317531585693, "learning_rate": 2.117605523828644e-06, "loss": 0.6761, "step": 22599 }, { "epoch": 0.7963564928601002, "grad_norm": 5.28450870513916, "learning_rate": 2.1169032692017043e-06, "loss": 2.7206, "step": 22600 }, { "epoch": 0.7963917298730586, "grad_norm": 1.1439218521118164, "learning_rate": 2.1162011172531026e-06, "loss": 0.7348, "step": 22601 }, { "epoch": 0.7964269668860171, "grad_norm": 2.600752353668213, "learning_rate": 2.115499067991986e-06, "loss": 3.1314, "step": 22602 }, { "epoch": 0.7964622038989755, "grad_norm": 1.0199552774429321, "learning_rate": 2.114797121427501e-06, "loss": 0.9311, "step": 22603 }, { "epoch": 0.7964974409119339, "grad_norm": 6.073203086853027, "learning_rate": 2.1140952775687817e-06, "loss": 3.2625, "step": 22604 }, { "epoch": 0.7965326779248924, "grad_norm": 1.4235135316848755, "learning_rate": 2.113393536424979e-06, "loss": 0.7057, "step": 22605 }, { "epoch": 0.7965679149378507, "grad_norm": 1.3664411306381226, "learning_rate": 2.1126918980052314e-06, "loss": 0.9658, "step": 22606 }, { "epoch": 0.7966031519508091, "grad_norm": 1.1713671684265137, "learning_rate": 2.1119903623186778e-06, "loss": 1.0706, "step": 22607 }, { "epoch": 0.7966383889637676, "grad_norm": 3.171565055847168, "learning_rate": 2.11128892937445e-06, "loss": 2.8877, "step": 22608 }, { "epoch": 0.796673625976726, "grad_norm": 5.457366466522217, "learning_rate": 2.1105875991816904e-06, "loss": 4.8585, "step": 22609 }, { "epoch": 0.7967088629896844, "grad_norm": 5.019738674163818, "learning_rate": 2.1098863717495353e-06, "loss": 5.4318, "step": 22610 }, { "epoch": 0.7967441000026427, "grad_norm": 1.1505868434906006, "learning_rate": 2.109185247087113e-06, "loss": 0.7417, "step": 22611 }, { "epoch": 0.7967793370156012, "grad_norm": 4.4722137451171875, "learning_rate": 2.1084842252035552e-06, "loss": 4.9814, "step": 22612 }, { "epoch": 0.7968145740285596, "grad_norm": 1.1595780849456787, "learning_rate": 2.107783306108001e-06, "loss": 1.0909, "step": 22613 }, { "epoch": 0.796849811041518, "grad_norm": 5.749635219573975, "learning_rate": 2.107082489809571e-06, "loss": 4.8241, "step": 22614 }, { "epoch": 0.7968850480544765, "grad_norm": 12.052881240844727, "learning_rate": 2.106381776317398e-06, "loss": 2.2747, "step": 22615 }, { "epoch": 0.7969202850674348, "grad_norm": 0.9428819417953491, "learning_rate": 2.1056811656406074e-06, "loss": 0.6165, "step": 22616 }, { "epoch": 0.7969555220803932, "grad_norm": 5.631814479827881, "learning_rate": 2.1049806577883257e-06, "loss": 6.8862, "step": 22617 }, { "epoch": 0.7969907590933517, "grad_norm": 2.5976462364196777, "learning_rate": 2.104280252769676e-06, "loss": 2.8053, "step": 22618 }, { "epoch": 0.7970259961063101, "grad_norm": 1.3429248332977295, "learning_rate": 2.1035799505937815e-06, "loss": 0.7792, "step": 22619 }, { "epoch": 0.7970612331192685, "grad_norm": 6.205041408538818, "learning_rate": 2.1028797512697653e-06, "loss": 4.7809, "step": 22620 }, { "epoch": 0.7970964701322268, "grad_norm": 1.3084628582000732, "learning_rate": 2.102179654806745e-06, "loss": 0.6965, "step": 22621 }, { "epoch": 0.7971317071451853, "grad_norm": 1.673498511314392, "learning_rate": 2.1014796612138444e-06, "loss": 0.8261, "step": 22622 }, { "epoch": 0.7971669441581437, "grad_norm": 4.926129341125488, "learning_rate": 2.100779770500171e-06, "loss": 5.4407, "step": 22623 }, { "epoch": 0.7972021811711021, "grad_norm": 5.118544101715088, "learning_rate": 2.10007998267485e-06, "loss": 5.6725, "step": 22624 }, { "epoch": 0.7972374181840606, "grad_norm": 1.102871060371399, "learning_rate": 2.0993802977469934e-06, "loss": 0.7767, "step": 22625 }, { "epoch": 0.797272655197019, "grad_norm": 3.0749664306640625, "learning_rate": 2.0986807157257183e-06, "loss": 0.9615, "step": 22626 }, { "epoch": 0.7973078922099773, "grad_norm": 1.3189445734024048, "learning_rate": 2.097981236620128e-06, "loss": 0.7245, "step": 22627 }, { "epoch": 0.7973431292229358, "grad_norm": 1.1547387838363647, "learning_rate": 2.097281860439341e-06, "loss": 0.6462, "step": 22628 }, { "epoch": 0.7973783662358942, "grad_norm": 3.164950132369995, "learning_rate": 2.096582587192467e-06, "loss": 2.659, "step": 22629 }, { "epoch": 0.7974136032488526, "grad_norm": 4.803704261779785, "learning_rate": 2.095883416888609e-06, "loss": 2.8387, "step": 22630 }, { "epoch": 0.797448840261811, "grad_norm": 2.5860157012939453, "learning_rate": 2.095184349536873e-06, "loss": 2.902, "step": 22631 }, { "epoch": 0.7974840772747694, "grad_norm": 5.030038833618164, "learning_rate": 2.0944853851463743e-06, "loss": 2.9199, "step": 22632 }, { "epoch": 0.7975193142877278, "grad_norm": 5.349267959594727, "learning_rate": 2.0937865237262068e-06, "loss": 6.4196, "step": 22633 }, { "epoch": 0.7975545513006862, "grad_norm": 4.283128261566162, "learning_rate": 2.0930877652854775e-06, "loss": 5.0505, "step": 22634 }, { "epoch": 0.7975897883136447, "grad_norm": 4.144399642944336, "learning_rate": 2.092389109833287e-06, "loss": 2.5077, "step": 22635 }, { "epoch": 0.7976250253266031, "grad_norm": 2.506760597229004, "learning_rate": 2.091690557378736e-06, "loss": 2.4589, "step": 22636 }, { "epoch": 0.7976602623395614, "grad_norm": 0.9837708473205566, "learning_rate": 2.0909921079309216e-06, "loss": 0.9448, "step": 22637 }, { "epoch": 0.7976954993525199, "grad_norm": 2.851121664047241, "learning_rate": 2.0902937614989437e-06, "loss": 2.5407, "step": 22638 }, { "epoch": 0.7977307363654783, "grad_norm": 5.41312837600708, "learning_rate": 2.089595518091897e-06, "loss": 3.5585, "step": 22639 }, { "epoch": 0.7977659733784367, "grad_norm": 4.9293131828308105, "learning_rate": 2.0888973777188757e-06, "loss": 5.0979, "step": 22640 }, { "epoch": 0.7978012103913952, "grad_norm": 5.213503837585449, "learning_rate": 2.0881993403889765e-06, "loss": 3.0567, "step": 22641 }, { "epoch": 0.7978364474043536, "grad_norm": 5.239825248718262, "learning_rate": 2.087501406111284e-06, "loss": 4.8638, "step": 22642 }, { "epoch": 0.7978716844173119, "grad_norm": 1.341447353363037, "learning_rate": 2.0868035748948955e-06, "loss": 0.9037, "step": 22643 }, { "epoch": 0.7979069214302703, "grad_norm": 1.2935327291488647, "learning_rate": 2.086105846748899e-06, "loss": 0.6649, "step": 22644 }, { "epoch": 0.7979421584432288, "grad_norm": 5.803707122802734, "learning_rate": 2.0854082216823846e-06, "loss": 6.8355, "step": 22645 }, { "epoch": 0.7979773954561872, "grad_norm": 1.686448335647583, "learning_rate": 2.084710699704431e-06, "loss": 0.6406, "step": 22646 }, { "epoch": 0.7980126324691456, "grad_norm": 1.7915796041488647, "learning_rate": 2.084013280824131e-06, "loss": 0.9425, "step": 22647 }, { "epoch": 0.798047869482104, "grad_norm": 5.156874179840088, "learning_rate": 2.0833159650505685e-06, "loss": 3.0209, "step": 22648 }, { "epoch": 0.7980831064950624, "grad_norm": 2.7749311923980713, "learning_rate": 2.082618752392822e-06, "loss": 2.454, "step": 22649 }, { "epoch": 0.7981183435080208, "grad_norm": 3.2997655868530273, "learning_rate": 2.081921642859972e-06, "loss": 2.8596, "step": 22650 }, { "epoch": 0.7981535805209793, "grad_norm": 4.555384635925293, "learning_rate": 2.0812246364611066e-06, "loss": 2.8095, "step": 22651 }, { "epoch": 0.7981888175339377, "grad_norm": 2.330306053161621, "learning_rate": 2.0805277332052966e-06, "loss": 0.7525, "step": 22652 }, { "epoch": 0.798224054546896, "grad_norm": 0.8905330300331116, "learning_rate": 2.079830933101621e-06, "loss": 0.7099, "step": 22653 }, { "epoch": 0.7982592915598544, "grad_norm": 2.9365134239196777, "learning_rate": 2.0791342361591573e-06, "loss": 2.546, "step": 22654 }, { "epoch": 0.7982945285728129, "grad_norm": 0.9649789333343506, "learning_rate": 2.078437642386978e-06, "loss": 0.8629, "step": 22655 }, { "epoch": 0.7983297655857713, "grad_norm": 4.2765278816223145, "learning_rate": 2.077741151794158e-06, "loss": 4.9564, "step": 22656 }, { "epoch": 0.7983650025987297, "grad_norm": 4.577566146850586, "learning_rate": 2.0770447643897693e-06, "loss": 2.9493, "step": 22657 }, { "epoch": 0.7984002396116882, "grad_norm": 1.2616336345672607, "learning_rate": 2.0763484801828816e-06, "loss": 0.6878, "step": 22658 }, { "epoch": 0.7984354766246465, "grad_norm": 1.5017907619476318, "learning_rate": 2.0756522991825633e-06, "loss": 0.635, "step": 22659 }, { "epoch": 0.7984707136376049, "grad_norm": 3.867469310760498, "learning_rate": 2.0749562213978835e-06, "loss": 3.0222, "step": 22660 }, { "epoch": 0.7985059506505634, "grad_norm": 2.90317964553833, "learning_rate": 2.0742602468379093e-06, "loss": 2.9062, "step": 22661 }, { "epoch": 0.7985411876635218, "grad_norm": 4.119962215423584, "learning_rate": 2.073564375511704e-06, "loss": 3.4028, "step": 22662 }, { "epoch": 0.7985764246764802, "grad_norm": 1.902500033378601, "learning_rate": 2.072868607428332e-06, "loss": 0.8093, "step": 22663 }, { "epoch": 0.7986116616894385, "grad_norm": 3.3226568698883057, "learning_rate": 2.0721729425968594e-06, "loss": 2.7672, "step": 22664 }, { "epoch": 0.798646898702397, "grad_norm": 1.014924168586731, "learning_rate": 2.071477381026339e-06, "loss": 0.9045, "step": 22665 }, { "epoch": 0.7986821357153554, "grad_norm": 2.7481000423431396, "learning_rate": 2.0707819227258386e-06, "loss": 2.6743, "step": 22666 }, { "epoch": 0.7987173727283138, "grad_norm": 5.530528545379639, "learning_rate": 2.070086567704416e-06, "loss": 3.7114, "step": 22667 }, { "epoch": 0.7987526097412723, "grad_norm": 1.0181655883789062, "learning_rate": 2.0693913159711233e-06, "loss": 0.7242, "step": 22668 }, { "epoch": 0.7987878467542306, "grad_norm": 1.64262056350708, "learning_rate": 2.068696167535016e-06, "loss": 0.9908, "step": 22669 }, { "epoch": 0.798823083767189, "grad_norm": 4.159261226654053, "learning_rate": 2.0680011224051566e-06, "loss": 4.0694, "step": 22670 }, { "epoch": 0.7988583207801475, "grad_norm": 4.100793838500977, "learning_rate": 2.0673061805905904e-06, "loss": 2.6428, "step": 22671 }, { "epoch": 0.7988935577931059, "grad_norm": 8.600225448608398, "learning_rate": 2.066611342100372e-06, "loss": 2.5558, "step": 22672 }, { "epoch": 0.7989287948060643, "grad_norm": 4.546121597290039, "learning_rate": 2.065916606943551e-06, "loss": 4.5871, "step": 22673 }, { "epoch": 0.7989640318190226, "grad_norm": 1.7246675491333008, "learning_rate": 2.0652219751291767e-06, "loss": 0.7764, "step": 22674 }, { "epoch": 0.7989992688319811, "grad_norm": 1.177782416343689, "learning_rate": 2.064527446666298e-06, "loss": 0.8969, "step": 22675 }, { "epoch": 0.7990345058449395, "grad_norm": 4.7332329750061035, "learning_rate": 2.063833021563959e-06, "loss": 5.0384, "step": 22676 }, { "epoch": 0.7990697428578979, "grad_norm": 1.2072385549545288, "learning_rate": 2.0631386998312062e-06, "loss": 0.7771, "step": 22677 }, { "epoch": 0.7991049798708564, "grad_norm": 5.3407673835754395, "learning_rate": 2.0624444814770827e-06, "loss": 5.3242, "step": 22678 }, { "epoch": 0.7991402168838148, "grad_norm": 3.736772060394287, "learning_rate": 2.061750366510631e-06, "loss": 2.5403, "step": 22679 }, { "epoch": 0.7991754538967731, "grad_norm": 7.360430717468262, "learning_rate": 2.0610563549408913e-06, "loss": 5.3834, "step": 22680 }, { "epoch": 0.7992106909097316, "grad_norm": 5.336611747741699, "learning_rate": 2.0603624467769047e-06, "loss": 2.9943, "step": 22681 }, { "epoch": 0.79924592792269, "grad_norm": 3.489718198776245, "learning_rate": 2.059668642027709e-06, "loss": 3.2474, "step": 22682 }, { "epoch": 0.7992811649356484, "grad_norm": 6.435970306396484, "learning_rate": 2.058974940702343e-06, "loss": 2.8664, "step": 22683 }, { "epoch": 0.7993164019486069, "grad_norm": 5.356417655944824, "learning_rate": 2.0582813428098335e-06, "loss": 5.6403, "step": 22684 }, { "epoch": 0.7993516389615652, "grad_norm": 2.9429805278778076, "learning_rate": 2.057587848359225e-06, "loss": 3.3447, "step": 22685 }, { "epoch": 0.7993868759745236, "grad_norm": 1.031984806060791, "learning_rate": 2.0568944573595484e-06, "loss": 0.9386, "step": 22686 }, { "epoch": 0.799422112987482, "grad_norm": 2.9741525650024414, "learning_rate": 2.0562011698198316e-06, "loss": 2.4339, "step": 22687 }, { "epoch": 0.7994573500004405, "grad_norm": 5.5340375900268555, "learning_rate": 2.0555079857491055e-06, "loss": 2.1744, "step": 22688 }, { "epoch": 0.7994925870133989, "grad_norm": 3.275649070739746, "learning_rate": 2.0548149051564004e-06, "loss": 2.7593, "step": 22689 }, { "epoch": 0.7995278240263572, "grad_norm": 1.2583199739456177, "learning_rate": 2.054121928050743e-06, "loss": 0.8729, "step": 22690 }, { "epoch": 0.7995630610393157, "grad_norm": 2.6832306385040283, "learning_rate": 2.0534290544411596e-06, "loss": 2.5866, "step": 22691 }, { "epoch": 0.7995982980522741, "grad_norm": 3.036114454269409, "learning_rate": 2.052736284336675e-06, "loss": 2.8874, "step": 22692 }, { "epoch": 0.7996335350652325, "grad_norm": 1.102465271949768, "learning_rate": 2.052043617746312e-06, "loss": 0.8946, "step": 22693 }, { "epoch": 0.799668772078191, "grad_norm": 5.730294227600098, "learning_rate": 2.051351054679094e-06, "loss": 5.3753, "step": 22694 }, { "epoch": 0.7997040090911494, "grad_norm": 1.7624859809875488, "learning_rate": 2.0506585951440404e-06, "loss": 1.0971, "step": 22695 }, { "epoch": 0.7997392461041077, "grad_norm": 64.18367767333984, "learning_rate": 2.0499662391501716e-06, "loss": 5.2255, "step": 22696 }, { "epoch": 0.7997744831170661, "grad_norm": 3.435288429260254, "learning_rate": 2.049273986706506e-06, "loss": 2.2965, "step": 22697 }, { "epoch": 0.7998097201300246, "grad_norm": 1.7182124853134155, "learning_rate": 2.048581837822059e-06, "loss": 0.9604, "step": 22698 }, { "epoch": 0.799844957142983, "grad_norm": 6.863185882568359, "learning_rate": 2.0478897925058482e-06, "loss": 3.3419, "step": 22699 }, { "epoch": 0.7998801941559414, "grad_norm": 1.3904083967208862, "learning_rate": 2.047197850766881e-06, "loss": 1.068, "step": 22700 }, { "epoch": 0.7999154311688998, "grad_norm": 22.154621124267578, "learning_rate": 2.0465060126141777e-06, "loss": 4.8863, "step": 22701 }, { "epoch": 0.7999506681818582, "grad_norm": 8.432475090026855, "learning_rate": 2.045814278056749e-06, "loss": 7.6405, "step": 22702 }, { "epoch": 0.7999859051948166, "grad_norm": 1.1126184463500977, "learning_rate": 2.0451226471036012e-06, "loss": 0.992, "step": 22703 }, { "epoch": 0.8000211422077751, "grad_norm": 6.491226673126221, "learning_rate": 2.04443111976374e-06, "loss": 9.0864, "step": 22704 }, { "epoch": 0.8000563792207335, "grad_norm": 2.265706777572632, "learning_rate": 2.0437396960461832e-06, "loss": 0.756, "step": 22705 }, { "epoch": 0.8000916162336918, "grad_norm": 1.1313029527664185, "learning_rate": 2.0430483759599263e-06, "loss": 0.8922, "step": 22706 }, { "epoch": 0.8001268532466502, "grad_norm": 12.618844985961914, "learning_rate": 2.042357159513979e-06, "loss": 2.9659, "step": 22707 }, { "epoch": 0.8001620902596087, "grad_norm": 2.8384358882904053, "learning_rate": 2.0416660467173435e-06, "loss": 3.0881, "step": 22708 }, { "epoch": 0.8001973272725671, "grad_norm": 5.599916934967041, "learning_rate": 2.0409750375790203e-06, "loss": 5.355, "step": 22709 }, { "epoch": 0.8002325642855255, "grad_norm": 2.484802722930908, "learning_rate": 2.0402841321080126e-06, "loss": 2.7735, "step": 22710 }, { "epoch": 0.800267801298484, "grad_norm": 1.011979341506958, "learning_rate": 2.0395933303133163e-06, "loss": 0.653, "step": 22711 }, { "epoch": 0.8003030383114423, "grad_norm": 0.9940326809883118, "learning_rate": 2.038902632203932e-06, "loss": 1.2982, "step": 22712 }, { "epoch": 0.8003382753244007, "grad_norm": 3.224500894546509, "learning_rate": 2.0382120377888547e-06, "loss": 3.3428, "step": 22713 }, { "epoch": 0.8003735123373592, "grad_norm": 5.909052848815918, "learning_rate": 2.0375215470770795e-06, "loss": 4.7478, "step": 22714 }, { "epoch": 0.8004087493503176, "grad_norm": 3.950101613998413, "learning_rate": 2.0368311600776e-06, "loss": 2.7554, "step": 22715 }, { "epoch": 0.800443986363276, "grad_norm": 6.481320858001709, "learning_rate": 2.0361408767994106e-06, "loss": 3.1811, "step": 22716 }, { "epoch": 0.8004792233762344, "grad_norm": 3.8389618396759033, "learning_rate": 2.0354506972514998e-06, "loss": 4.9024, "step": 22717 }, { "epoch": 0.8005144603891928, "grad_norm": 4.628937244415283, "learning_rate": 2.0347606214428608e-06, "loss": 3.0386, "step": 22718 }, { "epoch": 0.8005496974021512, "grad_norm": 4.987138271331787, "learning_rate": 2.0340706493824734e-06, "loss": 2.8981, "step": 22719 }, { "epoch": 0.8005849344151096, "grad_norm": 1.8543369770050049, "learning_rate": 2.033380781079334e-06, "loss": 0.8004, "step": 22720 }, { "epoch": 0.8006201714280681, "grad_norm": 7.264986515045166, "learning_rate": 2.0326910165424273e-06, "loss": 4.8742, "step": 22721 }, { "epoch": 0.8006554084410265, "grad_norm": 4.043729782104492, "learning_rate": 2.032001355780733e-06, "loss": 2.9036, "step": 22722 }, { "epoch": 0.8006906454539848, "grad_norm": 4.226921558380127, "learning_rate": 2.0313117988032318e-06, "loss": 2.9311, "step": 22723 }, { "epoch": 0.8007258824669433, "grad_norm": 1.098456621170044, "learning_rate": 2.0306223456189157e-06, "loss": 1.2599, "step": 22724 }, { "epoch": 0.8007611194799017, "grad_norm": 0.9800295829772949, "learning_rate": 2.0299329962367555e-06, "loss": 0.8768, "step": 22725 }, { "epoch": 0.8007963564928601, "grad_norm": 5.778082370758057, "learning_rate": 2.0292437506657337e-06, "loss": 5.4062, "step": 22726 }, { "epoch": 0.8008315935058186, "grad_norm": 1.751834750175476, "learning_rate": 2.0285546089148265e-06, "loss": 0.8899, "step": 22727 }, { "epoch": 0.8008668305187769, "grad_norm": 3.3355553150177, "learning_rate": 2.027865570993012e-06, "loss": 3.3224, "step": 22728 }, { "epoch": 0.8009020675317353, "grad_norm": 1.1041814088821411, "learning_rate": 2.027176636909264e-06, "loss": 1.0259, "step": 22729 }, { "epoch": 0.8009373045446937, "grad_norm": 1.1184691190719604, "learning_rate": 2.0264878066725556e-06, "loss": 0.8939, "step": 22730 }, { "epoch": 0.8009725415576522, "grad_norm": 1.837868332862854, "learning_rate": 2.0257990802918594e-06, "loss": 1.1228, "step": 22731 }, { "epoch": 0.8010077785706106, "grad_norm": 3.451840877532959, "learning_rate": 2.0251104577761448e-06, "loss": 2.8775, "step": 22732 }, { "epoch": 0.8010430155835689, "grad_norm": 5.763401985168457, "learning_rate": 2.024421939134383e-06, "loss": 5.0456, "step": 22733 }, { "epoch": 0.8010782525965274, "grad_norm": 3.0940816402435303, "learning_rate": 2.0237335243755416e-06, "loss": 2.4521, "step": 22734 }, { "epoch": 0.8011134896094858, "grad_norm": 1.8705036640167236, "learning_rate": 2.0230452135085865e-06, "loss": 0.8014, "step": 22735 }, { "epoch": 0.8011487266224442, "grad_norm": 2.6487784385681152, "learning_rate": 2.022357006542485e-06, "loss": 2.7112, "step": 22736 }, { "epoch": 0.8011839636354027, "grad_norm": 1.1458145380020142, "learning_rate": 2.021668903486201e-06, "loss": 0.9216, "step": 22737 }, { "epoch": 0.801219200648361, "grad_norm": 0.913302481174469, "learning_rate": 2.0209809043486904e-06, "loss": 0.6877, "step": 22738 }, { "epoch": 0.8012544376613194, "grad_norm": 1.3680859804153442, "learning_rate": 2.0202930091389226e-06, "loss": 0.9318, "step": 22739 }, { "epoch": 0.8012896746742778, "grad_norm": 4.089729309082031, "learning_rate": 2.0196052178658586e-06, "loss": 3.28, "step": 22740 }, { "epoch": 0.8013249116872363, "grad_norm": 9.359517097473145, "learning_rate": 2.01891753053845e-06, "loss": 6.6369, "step": 22741 }, { "epoch": 0.8013601487001947, "grad_norm": 3.176204204559326, "learning_rate": 2.018229947165654e-06, "loss": 2.6327, "step": 22742 }, { "epoch": 0.801395385713153, "grad_norm": 4.077749729156494, "learning_rate": 2.017542467756435e-06, "loss": 3.0826, "step": 22743 }, { "epoch": 0.8014306227261115, "grad_norm": 1.0525403022766113, "learning_rate": 2.0168550923197404e-06, "loss": 0.6675, "step": 22744 }, { "epoch": 0.8014658597390699, "grad_norm": 2.5250818729400635, "learning_rate": 2.016167820864524e-06, "loss": 2.2383, "step": 22745 }, { "epoch": 0.8015010967520283, "grad_norm": 3.0951218605041504, "learning_rate": 2.0154806533997394e-06, "loss": 2.7234, "step": 22746 }, { "epoch": 0.8015363337649868, "grad_norm": 1.7109118700027466, "learning_rate": 2.0147935899343363e-06, "loss": 1.1153, "step": 22747 }, { "epoch": 0.8015715707779452, "grad_norm": 3.445770025253296, "learning_rate": 2.0141066304772637e-06, "loss": 2.7693, "step": 22748 }, { "epoch": 0.8016068077909035, "grad_norm": 5.046731948852539, "learning_rate": 2.0134197750374697e-06, "loss": 4.8278, "step": 22749 }, { "epoch": 0.801642044803862, "grad_norm": 4.052493572235107, "learning_rate": 2.0127330236238998e-06, "loss": 2.8047, "step": 22750 }, { "epoch": 0.8016772818168204, "grad_norm": 6.298375606536865, "learning_rate": 2.0120463762455e-06, "loss": 3.3364, "step": 22751 }, { "epoch": 0.8017125188297788, "grad_norm": 5.07299280166626, "learning_rate": 2.011359832911214e-06, "loss": 4.4038, "step": 22752 }, { "epoch": 0.8017477558427372, "grad_norm": 0.8272849917411804, "learning_rate": 2.0106733936299837e-06, "loss": 0.8026, "step": 22753 }, { "epoch": 0.8017829928556957, "grad_norm": 1.5007213354110718, "learning_rate": 2.0099870584107497e-06, "loss": 0.9368, "step": 22754 }, { "epoch": 0.801818229868654, "grad_norm": 2.94169020652771, "learning_rate": 2.0093008272624524e-06, "loss": 2.6161, "step": 22755 }, { "epoch": 0.8018534668816124, "grad_norm": 2.7388811111450195, "learning_rate": 2.0086147001940325e-06, "loss": 2.8531, "step": 22756 }, { "epoch": 0.8018887038945709, "grad_norm": 1.4393436908721924, "learning_rate": 2.007928677214418e-06, "loss": 0.606, "step": 22757 }, { "epoch": 0.8019239409075293, "grad_norm": 4.02114200592041, "learning_rate": 2.007242758332554e-06, "loss": 2.5611, "step": 22758 }, { "epoch": 0.8019591779204877, "grad_norm": 1.5891025066375732, "learning_rate": 2.006556943557374e-06, "loss": 0.8086, "step": 22759 }, { "epoch": 0.8019944149334461, "grad_norm": 4.034277439117432, "learning_rate": 2.005871232897805e-06, "loss": 2.6828, "step": 22760 }, { "epoch": 0.8020296519464045, "grad_norm": 1.2911444902420044, "learning_rate": 2.005185626362778e-06, "loss": 0.8483, "step": 22761 }, { "epoch": 0.8020648889593629, "grad_norm": 4.151024341583252, "learning_rate": 2.0045001239612326e-06, "loss": 2.6546, "step": 22762 }, { "epoch": 0.8021001259723213, "grad_norm": 3.23596453666687, "learning_rate": 2.003814725702089e-06, "loss": 2.9935, "step": 22763 }, { "epoch": 0.8021353629852798, "grad_norm": 0.9316726326942444, "learning_rate": 2.003129431594277e-06, "loss": 0.8307, "step": 22764 }, { "epoch": 0.8021705999982381, "grad_norm": 2.5857796669006348, "learning_rate": 2.0024442416467227e-06, "loss": 2.8573, "step": 22765 }, { "epoch": 0.8022058370111965, "grad_norm": 5.258224010467529, "learning_rate": 2.0017591558683513e-06, "loss": 4.0545, "step": 22766 }, { "epoch": 0.802241074024155, "grad_norm": 3.539205551147461, "learning_rate": 2.001074174268085e-06, "loss": 3.0703, "step": 22767 }, { "epoch": 0.8022763110371134, "grad_norm": 10.245441436767578, "learning_rate": 2.0003892968548465e-06, "loss": 5.5243, "step": 22768 }, { "epoch": 0.8023115480500718, "grad_norm": 7.268245220184326, "learning_rate": 1.999704523637558e-06, "loss": 5.4031, "step": 22769 }, { "epoch": 0.8023467850630303, "grad_norm": 1.6278369426727295, "learning_rate": 1.999019854625135e-06, "loss": 0.9792, "step": 22770 }, { "epoch": 0.8023820220759886, "grad_norm": 3.4424400329589844, "learning_rate": 1.9983352898264983e-06, "loss": 2.8139, "step": 22771 }, { "epoch": 0.802417259088947, "grad_norm": 1.432629942893982, "learning_rate": 1.997650829250565e-06, "loss": 1.1384, "step": 22772 }, { "epoch": 0.8024524961019054, "grad_norm": 4.2330145835876465, "learning_rate": 1.9969664729062478e-06, "loss": 5.2333, "step": 22773 }, { "epoch": 0.8024877331148639, "grad_norm": 0.9989662766456604, "learning_rate": 1.9962822208024614e-06, "loss": 0.899, "step": 22774 }, { "epoch": 0.8025229701278223, "grad_norm": 24.708358764648438, "learning_rate": 1.995598072948122e-06, "loss": 3.1597, "step": 22775 }, { "epoch": 0.8025582071407806, "grad_norm": 3.4432759284973145, "learning_rate": 1.9949140293521308e-06, "loss": 2.9009, "step": 22776 }, { "epoch": 0.8025934441537391, "grad_norm": 5.648686408996582, "learning_rate": 1.9942300900234077e-06, "loss": 4.0158, "step": 22777 }, { "epoch": 0.8026286811666975, "grad_norm": 1.2307895421981812, "learning_rate": 1.9935462549708606e-06, "loss": 0.9303, "step": 22778 }, { "epoch": 0.8026639181796559, "grad_norm": 3.501431465148926, "learning_rate": 1.9928625242033896e-06, "loss": 2.7231, "step": 22779 }, { "epoch": 0.8026991551926144, "grad_norm": 1.1058422327041626, "learning_rate": 1.9921788977299016e-06, "loss": 0.7995, "step": 22780 }, { "epoch": 0.8027343922055727, "grad_norm": 4.625960350036621, "learning_rate": 1.991495375559309e-06, "loss": 4.2254, "step": 22781 }, { "epoch": 0.8027696292185311, "grad_norm": 4.05695104598999, "learning_rate": 1.990811957700507e-06, "loss": 2.5792, "step": 22782 }, { "epoch": 0.8028048662314896, "grad_norm": 1.6962518692016602, "learning_rate": 1.9901286441624003e-06, "loss": 0.7466, "step": 22783 }, { "epoch": 0.802840103244448, "grad_norm": 1.235176682472229, "learning_rate": 1.9894454349538872e-06, "loss": 0.7702, "step": 22784 }, { "epoch": 0.8028753402574064, "grad_norm": 1.7400808334350586, "learning_rate": 1.9887623300838675e-06, "loss": 0.806, "step": 22785 }, { "epoch": 0.8029105772703647, "grad_norm": 2.660627841949463, "learning_rate": 1.988079329561239e-06, "loss": 2.4112, "step": 22786 }, { "epoch": 0.8029458142833232, "grad_norm": 5.85211706161499, "learning_rate": 1.987396433394898e-06, "loss": 5.4763, "step": 22787 }, { "epoch": 0.8029810512962816, "grad_norm": 1.0015596151351929, "learning_rate": 1.9867136415937394e-06, "loss": 1.4911, "step": 22788 }, { "epoch": 0.80301628830924, "grad_norm": 1.9946919679641724, "learning_rate": 1.9860309541666554e-06, "loss": 0.9543, "step": 22789 }, { "epoch": 0.8030515253221985, "grad_norm": 2.6912033557891846, "learning_rate": 1.9853483711225395e-06, "loss": 2.7404, "step": 22790 }, { "epoch": 0.8030867623351569, "grad_norm": 3.4976818561553955, "learning_rate": 1.984665892470282e-06, "loss": 3.1853, "step": 22791 }, { "epoch": 0.8031219993481152, "grad_norm": 4.408421516418457, "learning_rate": 1.9839835182187726e-06, "loss": 3.3985, "step": 22792 }, { "epoch": 0.8031572363610737, "grad_norm": 7.085491180419922, "learning_rate": 1.9833012483768988e-06, "loss": 4.7214, "step": 22793 }, { "epoch": 0.8031924733740321, "grad_norm": 1.191060185432434, "learning_rate": 1.9826190829535497e-06, "loss": 1.0327, "step": 22794 }, { "epoch": 0.8032277103869905, "grad_norm": 4.521582126617432, "learning_rate": 1.9819370219576037e-06, "loss": 2.739, "step": 22795 }, { "epoch": 0.8032629473999489, "grad_norm": 1.819649338722229, "learning_rate": 1.9812550653979513e-06, "loss": 0.948, "step": 22796 }, { "epoch": 0.8032981844129073, "grad_norm": 4.754632949829102, "learning_rate": 1.9805732132834764e-06, "loss": 3.1391, "step": 22797 }, { "epoch": 0.8033334214258657, "grad_norm": 2.4344637393951416, "learning_rate": 1.979891465623054e-06, "loss": 3.02, "step": 22798 }, { "epoch": 0.8033686584388241, "grad_norm": 1.7850173711776733, "learning_rate": 1.979209822425564e-06, "loss": 0.8658, "step": 22799 }, { "epoch": 0.8034038954517826, "grad_norm": 0.7563028931617737, "learning_rate": 1.978528283699893e-06, "loss": 0.6907, "step": 22800 }, { "epoch": 0.803439132464741, "grad_norm": 1.5168602466583252, "learning_rate": 1.9778468494549096e-06, "loss": 0.9817, "step": 22801 }, { "epoch": 0.8034743694776993, "grad_norm": 1.2303274869918823, "learning_rate": 1.9771655196994945e-06, "loss": 1.0432, "step": 22802 }, { "epoch": 0.8035096064906578, "grad_norm": 1.0088093280792236, "learning_rate": 1.976484294442519e-06, "loss": 1.1787, "step": 22803 }, { "epoch": 0.8035448435036162, "grad_norm": 4.243621349334717, "learning_rate": 1.9758031736928583e-06, "loss": 3.1686, "step": 22804 }, { "epoch": 0.8035800805165746, "grad_norm": 9.000781059265137, "learning_rate": 1.9751221574593826e-06, "loss": 4.8147, "step": 22805 }, { "epoch": 0.803615317529533, "grad_norm": 5.006643295288086, "learning_rate": 1.9744412457509643e-06, "loss": 4.898, "step": 22806 }, { "epoch": 0.8036505545424915, "grad_norm": 1.3428629636764526, "learning_rate": 1.9737604385764698e-06, "loss": 0.8589, "step": 22807 }, { "epoch": 0.8036857915554498, "grad_norm": 1.11793053150177, "learning_rate": 1.9730797359447683e-06, "loss": 0.6169, "step": 22808 }, { "epoch": 0.8037210285684082, "grad_norm": 1.7078044414520264, "learning_rate": 1.9723991378647267e-06, "loss": 0.8605, "step": 22809 }, { "epoch": 0.8037562655813667, "grad_norm": 4.1636481285095215, "learning_rate": 1.9717186443452074e-06, "loss": 4.3163, "step": 22810 }, { "epoch": 0.8037915025943251, "grad_norm": 3.7355475425720215, "learning_rate": 1.9710382553950767e-06, "loss": 3.4276, "step": 22811 }, { "epoch": 0.8038267396072835, "grad_norm": 2.96317982673645, "learning_rate": 1.9703579710231946e-06, "loss": 2.634, "step": 22812 }, { "epoch": 0.803861976620242, "grad_norm": 6.541738033294678, "learning_rate": 1.9696777912384258e-06, "loss": 5.0231, "step": 22813 }, { "epoch": 0.8038972136332003, "grad_norm": 3.78896164894104, "learning_rate": 1.9689977160496208e-06, "loss": 2.8219, "step": 22814 }, { "epoch": 0.8039324506461587, "grad_norm": 1.4932141304016113, "learning_rate": 1.9683177454656476e-06, "loss": 0.9104, "step": 22815 }, { "epoch": 0.8039676876591172, "grad_norm": 7.4889817237854, "learning_rate": 1.96763787949536e-06, "loss": 3.3754, "step": 22816 }, { "epoch": 0.8040029246720756, "grad_norm": 2.8841843605041504, "learning_rate": 1.9669581181476106e-06, "loss": 2.839, "step": 22817 }, { "epoch": 0.804038161685034, "grad_norm": 3.829528331756592, "learning_rate": 1.9662784614312514e-06, "loss": 2.2132, "step": 22818 }, { "epoch": 0.8040733986979923, "grad_norm": 4.882403373718262, "learning_rate": 1.965598909355144e-06, "loss": 5.0, "step": 22819 }, { "epoch": 0.8041086357109508, "grad_norm": 3.27852201461792, "learning_rate": 1.964919461928132e-06, "loss": 2.6614, "step": 22820 }, { "epoch": 0.8041438727239092, "grad_norm": 4.870680332183838, "learning_rate": 1.964240119159068e-06, "loss": 4.8924, "step": 22821 }, { "epoch": 0.8041791097368676, "grad_norm": 4.059686660766602, "learning_rate": 1.9635608810567986e-06, "loss": 2.8269, "step": 22822 }, { "epoch": 0.8042143467498261, "grad_norm": 1.1829502582550049, "learning_rate": 1.9628817476301733e-06, "loss": 0.7806, "step": 22823 }, { "epoch": 0.8042495837627844, "grad_norm": 1.150669813156128, "learning_rate": 1.9622027188880365e-06, "loss": 1.0237, "step": 22824 }, { "epoch": 0.8042848207757428, "grad_norm": 7.210608005523682, "learning_rate": 1.9615237948392342e-06, "loss": 5.3339, "step": 22825 }, { "epoch": 0.8043200577887013, "grad_norm": 4.162729740142822, "learning_rate": 1.9608449754926074e-06, "loss": 2.7562, "step": 22826 }, { "epoch": 0.8043552948016597, "grad_norm": 2.2348504066467285, "learning_rate": 1.9601662608569994e-06, "loss": 1.1523, "step": 22827 }, { "epoch": 0.8043905318146181, "grad_norm": 0.9163604378700256, "learning_rate": 1.9594876509412496e-06, "loss": 0.6648, "step": 22828 }, { "epoch": 0.8044257688275764, "grad_norm": 5.18290376663208, "learning_rate": 1.958809145754197e-06, "loss": 3.1855, "step": 22829 }, { "epoch": 0.8044610058405349, "grad_norm": 1.169493317604065, "learning_rate": 1.9581307453046807e-06, "loss": 1.156, "step": 22830 }, { "epoch": 0.8044962428534933, "grad_norm": 4.054906368255615, "learning_rate": 1.9574524496015345e-06, "loss": 3.2539, "step": 22831 }, { "epoch": 0.8045314798664517, "grad_norm": 4.079555034637451, "learning_rate": 1.9567742586535978e-06, "loss": 2.8659, "step": 22832 }, { "epoch": 0.8045667168794102, "grad_norm": 3.2434487342834473, "learning_rate": 1.9560961724696957e-06, "loss": 2.8773, "step": 22833 }, { "epoch": 0.8046019538923685, "grad_norm": 2.60628080368042, "learning_rate": 1.955418191058668e-06, "loss": 2.8783, "step": 22834 }, { "epoch": 0.8046371909053269, "grad_norm": 1.240726351737976, "learning_rate": 1.9547403144293466e-06, "loss": 0.728, "step": 22835 }, { "epoch": 0.8046724279182854, "grad_norm": 3.5832931995391846, "learning_rate": 1.9540625425905534e-06, "loss": 3.1155, "step": 22836 }, { "epoch": 0.8047076649312438, "grad_norm": 5.253668308258057, "learning_rate": 1.9533848755511187e-06, "loss": 2.8814, "step": 22837 }, { "epoch": 0.8047429019442022, "grad_norm": 4.005971431732178, "learning_rate": 1.9527073133198758e-06, "loss": 3.1392, "step": 22838 }, { "epoch": 0.8047781389571605, "grad_norm": 5.4178619384765625, "learning_rate": 1.952029855905644e-06, "loss": 2.8898, "step": 22839 }, { "epoch": 0.804813375970119, "grad_norm": 1.1841477155685425, "learning_rate": 1.9513525033172463e-06, "loss": 1.1611, "step": 22840 }, { "epoch": 0.8048486129830774, "grad_norm": 1.075717568397522, "learning_rate": 1.950675255563509e-06, "loss": 0.8541, "step": 22841 }, { "epoch": 0.8048838499960358, "grad_norm": 4.495126724243164, "learning_rate": 1.9499981126532508e-06, "loss": 2.7085, "step": 22842 }, { "epoch": 0.8049190870089943, "grad_norm": 1.4712047576904297, "learning_rate": 1.949321074595294e-06, "loss": 0.7617, "step": 22843 }, { "epoch": 0.8049543240219527, "grad_norm": 3.8563737869262695, "learning_rate": 1.9486441413984547e-06, "loss": 3.1613, "step": 22844 }, { "epoch": 0.804989561034911, "grad_norm": 1.354841947555542, "learning_rate": 1.9479673130715503e-06, "loss": 0.8881, "step": 22845 }, { "epoch": 0.8050247980478695, "grad_norm": 6.550892353057861, "learning_rate": 1.9472905896233985e-06, "loss": 2.6331, "step": 22846 }, { "epoch": 0.8050600350608279, "grad_norm": 2.2841625213623047, "learning_rate": 1.946613971062813e-06, "loss": 0.6628, "step": 22847 }, { "epoch": 0.8050952720737863, "grad_norm": 1.265499472618103, "learning_rate": 1.9459374573986045e-06, "loss": 0.6658, "step": 22848 }, { "epoch": 0.8051305090867447, "grad_norm": 4.712457656860352, "learning_rate": 1.9452610486395886e-06, "loss": 4.8099, "step": 22849 }, { "epoch": 0.8051657460997032, "grad_norm": 1.0523920059204102, "learning_rate": 1.944584744794572e-06, "loss": 0.7099, "step": 22850 }, { "epoch": 0.8052009831126615, "grad_norm": 0.9992762804031372, "learning_rate": 1.9439085458723682e-06, "loss": 0.8468, "step": 22851 }, { "epoch": 0.8052362201256199, "grad_norm": 0.8096052408218384, "learning_rate": 1.943232451881777e-06, "loss": 0.7121, "step": 22852 }, { "epoch": 0.8052714571385784, "grad_norm": 4.030966281890869, "learning_rate": 1.942556462831612e-06, "loss": 2.8559, "step": 22853 }, { "epoch": 0.8053066941515368, "grad_norm": 3.0239932537078857, "learning_rate": 1.9418805787306774e-06, "loss": 0.9171, "step": 22854 }, { "epoch": 0.8053419311644952, "grad_norm": 0.9558642506599426, "learning_rate": 1.9412047995877726e-06, "loss": 0.9896, "step": 22855 }, { "epoch": 0.8053771681774536, "grad_norm": 1.202433705329895, "learning_rate": 1.9405291254116986e-06, "loss": 0.9067, "step": 22856 }, { "epoch": 0.805412405190412, "grad_norm": 3.3196563720703125, "learning_rate": 1.939853556211265e-06, "loss": 2.7678, "step": 22857 }, { "epoch": 0.8054476422033704, "grad_norm": 6.1124091148376465, "learning_rate": 1.939178091995263e-06, "loss": 4.2368, "step": 22858 }, { "epoch": 0.8054828792163289, "grad_norm": 4.851268291473389, "learning_rate": 1.9385027327724935e-06, "loss": 4.6278, "step": 22859 }, { "epoch": 0.8055181162292873, "grad_norm": 0.8610866665840149, "learning_rate": 1.937827478551749e-06, "loss": 0.9396, "step": 22860 }, { "epoch": 0.8055533532422456, "grad_norm": 2.1841113567352295, "learning_rate": 1.937152329341835e-06, "loss": 2.2142, "step": 22861 }, { "epoch": 0.805588590255204, "grad_norm": 6.344409942626953, "learning_rate": 1.9364772851515356e-06, "loss": 4.4597, "step": 22862 }, { "epoch": 0.8056238272681625, "grad_norm": 3.7587289810180664, "learning_rate": 1.9358023459896467e-06, "loss": 3.1607, "step": 22863 }, { "epoch": 0.8056590642811209, "grad_norm": 6.5151686668396, "learning_rate": 1.9351275118649603e-06, "loss": 3.3591, "step": 22864 }, { "epoch": 0.8056943012940793, "grad_norm": 5.540228843688965, "learning_rate": 1.9344527827862646e-06, "loss": 3.3629, "step": 22865 }, { "epoch": 0.8057295383070378, "grad_norm": 5.0850348472595215, "learning_rate": 1.9337781587623485e-06, "loss": 4.8111, "step": 22866 }, { "epoch": 0.8057647753199961, "grad_norm": 3.035998582839966, "learning_rate": 1.933103639802e-06, "loss": 2.8177, "step": 22867 }, { "epoch": 0.8058000123329545, "grad_norm": 3.6036176681518555, "learning_rate": 1.9324292259140044e-06, "loss": 2.8027, "step": 22868 }, { "epoch": 0.805835249345913, "grad_norm": 4.443436622619629, "learning_rate": 1.9317549171071447e-06, "loss": 2.4878, "step": 22869 }, { "epoch": 0.8058704863588714, "grad_norm": 4.088804721832275, "learning_rate": 1.931080713390209e-06, "loss": 4.79, "step": 22870 }, { "epoch": 0.8059057233718298, "grad_norm": 1.23827064037323, "learning_rate": 1.930406614771968e-06, "loss": 0.9022, "step": 22871 }, { "epoch": 0.8059409603847881, "grad_norm": 4.154186248779297, "learning_rate": 1.9297326212612122e-06, "loss": 4.8571, "step": 22872 }, { "epoch": 0.8059761973977466, "grad_norm": 1.5084202289581299, "learning_rate": 1.92905873286672e-06, "loss": 0.9228, "step": 22873 }, { "epoch": 0.806011434410705, "grad_norm": 4.095597743988037, "learning_rate": 1.928384949597263e-06, "loss": 3.004, "step": 22874 }, { "epoch": 0.8060466714236634, "grad_norm": 7.241376876831055, "learning_rate": 1.9277112714616165e-06, "loss": 2.9414, "step": 22875 }, { "epoch": 0.8060819084366219, "grad_norm": 0.9578415751457214, "learning_rate": 1.9270376984685656e-06, "loss": 0.6956, "step": 22876 }, { "epoch": 0.8061171454495802, "grad_norm": 4.709328651428223, "learning_rate": 1.9263642306268738e-06, "loss": 4.3435, "step": 22877 }, { "epoch": 0.8061523824625386, "grad_norm": 3.2299211025238037, "learning_rate": 1.9256908679453155e-06, "loss": 2.3317, "step": 22878 }, { "epoch": 0.8061876194754971, "grad_norm": 1.0142148733139038, "learning_rate": 1.9250176104326614e-06, "loss": 0.7864, "step": 22879 }, { "epoch": 0.8062228564884555, "grad_norm": 2.4482202529907227, "learning_rate": 1.9243444580976854e-06, "loss": 2.2099, "step": 22880 }, { "epoch": 0.8062580935014139, "grad_norm": 7.564966678619385, "learning_rate": 1.9236714109491493e-06, "loss": 3.3545, "step": 22881 }, { "epoch": 0.8062933305143722, "grad_norm": 4.2546162605285645, "learning_rate": 1.9229984689958215e-06, "loss": 3.1687, "step": 22882 }, { "epoch": 0.8063285675273307, "grad_norm": 1.4969414472579956, "learning_rate": 1.922325632246468e-06, "loss": 0.752, "step": 22883 }, { "epoch": 0.8063638045402891, "grad_norm": 5.562400817871094, "learning_rate": 1.9216529007098517e-06, "loss": 3.0084, "step": 22884 }, { "epoch": 0.8063990415532475, "grad_norm": 4.351430892944336, "learning_rate": 1.9209802743947357e-06, "loss": 2.6807, "step": 22885 }, { "epoch": 0.806434278566206, "grad_norm": 6.062699317932129, "learning_rate": 1.9203077533098804e-06, "loss": 3.3952, "step": 22886 }, { "epoch": 0.8064695155791644, "grad_norm": 1.1083648204803467, "learning_rate": 1.9196353374640466e-06, "loss": 0.8723, "step": 22887 }, { "epoch": 0.8065047525921227, "grad_norm": 3.003527879714966, "learning_rate": 1.918963026865991e-06, "loss": 2.7056, "step": 22888 }, { "epoch": 0.8065399896050812, "grad_norm": 1.4337122440338135, "learning_rate": 1.9182908215244735e-06, "loss": 0.8551, "step": 22889 }, { "epoch": 0.8065752266180396, "grad_norm": 4.043867111206055, "learning_rate": 1.917618721448243e-06, "loss": 4.7435, "step": 22890 }, { "epoch": 0.806610463630998, "grad_norm": 1.3856555223464966, "learning_rate": 1.9169467266460605e-06, "loss": 0.9232, "step": 22891 }, { "epoch": 0.8066457006439565, "grad_norm": 1.2286335229873657, "learning_rate": 1.916274837126679e-06, "loss": 0.6567, "step": 22892 }, { "epoch": 0.8066809376569148, "grad_norm": 3.5372488498687744, "learning_rate": 1.9156030528988455e-06, "loss": 2.772, "step": 22893 }, { "epoch": 0.8067161746698732, "grad_norm": 4.027212619781494, "learning_rate": 1.9149313739713095e-06, "loss": 2.9561, "step": 22894 }, { "epoch": 0.8067514116828316, "grad_norm": 3.8630871772766113, "learning_rate": 1.914259800352827e-06, "loss": 2.4913, "step": 22895 }, { "epoch": 0.8067866486957901, "grad_norm": 1.4496686458587646, "learning_rate": 1.9135883320521376e-06, "loss": 1.0107, "step": 22896 }, { "epoch": 0.8068218857087485, "grad_norm": 2.148991823196411, "learning_rate": 1.91291696907799e-06, "loss": 0.751, "step": 22897 }, { "epoch": 0.8068571227217068, "grad_norm": 2.4966893196105957, "learning_rate": 1.912245711439127e-06, "loss": 2.4006, "step": 22898 }, { "epoch": 0.8068923597346653, "grad_norm": 3.5115363597869873, "learning_rate": 1.911574559144299e-06, "loss": 2.4778, "step": 22899 }, { "epoch": 0.8069275967476237, "grad_norm": 3.5115363597869873, "learning_rate": 1.911574559144299e-06, "loss": 3.0209, "step": 22900 }, { "epoch": 0.8069628337605821, "grad_norm": 3.256013870239258, "learning_rate": 1.91090351220224e-06, "loss": 2.61, "step": 22901 }, { "epoch": 0.8069980707735406, "grad_norm": 3.11738920211792, "learning_rate": 1.9102325706216927e-06, "loss": 2.704, "step": 22902 }, { "epoch": 0.807033307786499, "grad_norm": 5.758680820465088, "learning_rate": 1.909561734411398e-06, "loss": 6.886, "step": 22903 }, { "epoch": 0.8070685447994573, "grad_norm": 7.304947853088379, "learning_rate": 1.908891003580091e-06, "loss": 3.2223, "step": 22904 }, { "epoch": 0.8071037818124157, "grad_norm": 1.1342370510101318, "learning_rate": 1.9082203781365093e-06, "loss": 0.7676, "step": 22905 }, { "epoch": 0.8071390188253742, "grad_norm": 1.1834099292755127, "learning_rate": 1.9075498580893893e-06, "loss": 0.7338, "step": 22906 }, { "epoch": 0.8071742558383326, "grad_norm": 4.363394260406494, "learning_rate": 1.9068794434474614e-06, "loss": 2.9914, "step": 22907 }, { "epoch": 0.807209492851291, "grad_norm": 5.664098739624023, "learning_rate": 1.906209134219461e-06, "loss": 4.8413, "step": 22908 }, { "epoch": 0.8072447298642494, "grad_norm": 0.8654316067695618, "learning_rate": 1.9055389304141193e-06, "loss": 0.6653, "step": 22909 }, { "epoch": 0.8072799668772078, "grad_norm": 4.384472370147705, "learning_rate": 1.9048688320401597e-06, "loss": 3.258, "step": 22910 }, { "epoch": 0.8073152038901662, "grad_norm": 4.391758441925049, "learning_rate": 1.904198839106317e-06, "loss": 4.2642, "step": 22911 }, { "epoch": 0.8073504409031247, "grad_norm": 3.5390918254852295, "learning_rate": 1.9035289516213173e-06, "loss": 3.056, "step": 22912 }, { "epoch": 0.8073856779160831, "grad_norm": 1.0689761638641357, "learning_rate": 1.9028591695938826e-06, "loss": 0.9837, "step": 22913 }, { "epoch": 0.8074209149290414, "grad_norm": 3.2659859657287598, "learning_rate": 1.9021894930327366e-06, "loss": 2.3993, "step": 22914 }, { "epoch": 0.8074561519419998, "grad_norm": 5.296316623687744, "learning_rate": 1.9015199219466085e-06, "loss": 5.3614, "step": 22915 }, { "epoch": 0.8074913889549583, "grad_norm": 4.186898231506348, "learning_rate": 1.900850456344212e-06, "loss": 4.8641, "step": 22916 }, { "epoch": 0.8075266259679167, "grad_norm": 1.3999539613723755, "learning_rate": 1.9001810962342703e-06, "loss": 0.8306, "step": 22917 }, { "epoch": 0.8075618629808751, "grad_norm": 5.227004051208496, "learning_rate": 1.8995118416254998e-06, "loss": 5.2633, "step": 22918 }, { "epoch": 0.8075970999938336, "grad_norm": 4.472835063934326, "learning_rate": 1.8988426925266235e-06, "loss": 2.7552, "step": 22919 }, { "epoch": 0.8076323370067919, "grad_norm": 4.969320774078369, "learning_rate": 1.898173648946351e-06, "loss": 2.9611, "step": 22920 }, { "epoch": 0.8076675740197503, "grad_norm": 1.3202190399169922, "learning_rate": 1.8975047108933987e-06, "loss": 1.0416, "step": 22921 }, { "epoch": 0.8077028110327088, "grad_norm": 3.5849223136901855, "learning_rate": 1.8968358783764796e-06, "loss": 2.705, "step": 22922 }, { "epoch": 0.8077380480456672, "grad_norm": 1.2198654413223267, "learning_rate": 1.8961671514043057e-06, "loss": 0.803, "step": 22923 }, { "epoch": 0.8077732850586256, "grad_norm": 0.8667076230049133, "learning_rate": 1.8954985299855855e-06, "loss": 0.6786, "step": 22924 }, { "epoch": 0.807808522071584, "grad_norm": 1.1880226135253906, "learning_rate": 1.894830014129031e-06, "loss": 0.6769, "step": 22925 }, { "epoch": 0.8078437590845424, "grad_norm": 3.126504421234131, "learning_rate": 1.8941616038433475e-06, "loss": 2.9291, "step": 22926 }, { "epoch": 0.8078789960975008, "grad_norm": 0.9691339135169983, "learning_rate": 1.8934932991372413e-06, "loss": 0.734, "step": 22927 }, { "epoch": 0.8079142331104592, "grad_norm": 1.0098806619644165, "learning_rate": 1.8928251000194198e-06, "loss": 0.7333, "step": 22928 }, { "epoch": 0.8079494701234177, "grad_norm": 4.59263801574707, "learning_rate": 1.8921570064985794e-06, "loss": 4.5442, "step": 22929 }, { "epoch": 0.807984707136376, "grad_norm": 4.049613952636719, "learning_rate": 1.8914890185834289e-06, "loss": 3.8892, "step": 22930 }, { "epoch": 0.8080199441493344, "grad_norm": 1.4001165628433228, "learning_rate": 1.8908211362826701e-06, "loss": 1.2386, "step": 22931 }, { "epoch": 0.8080551811622929, "grad_norm": 4.068437576293945, "learning_rate": 1.8901533596049959e-06, "loss": 4.4034, "step": 22932 }, { "epoch": 0.8080904181752513, "grad_norm": 5.980659008026123, "learning_rate": 1.889485688559105e-06, "loss": 4.5977, "step": 22933 }, { "epoch": 0.8081256551882097, "grad_norm": 4.256535530090332, "learning_rate": 1.8888181231536984e-06, "loss": 2.7014, "step": 22934 }, { "epoch": 0.8081608922011682, "grad_norm": 9.7366943359375, "learning_rate": 1.8881506633974711e-06, "loss": 4.2391, "step": 22935 }, { "epoch": 0.8081961292141265, "grad_norm": 0.9389842748641968, "learning_rate": 1.887483309299113e-06, "loss": 0.8168, "step": 22936 }, { "epoch": 0.8082313662270849, "grad_norm": 4.357661724090576, "learning_rate": 1.8868160608673146e-06, "loss": 2.4557, "step": 22937 }, { "epoch": 0.8082666032400433, "grad_norm": 4.5989227294921875, "learning_rate": 1.886148918110776e-06, "loss": 3.0208, "step": 22938 }, { "epoch": 0.8083018402530018, "grad_norm": 0.9845666885375977, "learning_rate": 1.8854818810381792e-06, "loss": 0.6403, "step": 22939 }, { "epoch": 0.8083370772659602, "grad_norm": 0.865778923034668, "learning_rate": 1.8848149496582148e-06, "loss": 1.139, "step": 22940 }, { "epoch": 0.8083723142789185, "grad_norm": 3.7989487648010254, "learning_rate": 1.884148123979569e-06, "loss": 4.4706, "step": 22941 }, { "epoch": 0.808407551291877, "grad_norm": 4.323718547821045, "learning_rate": 1.8834814040109273e-06, "loss": 3.1978, "step": 22942 }, { "epoch": 0.8084427883048354, "grad_norm": 15.130425453186035, "learning_rate": 1.882814789760974e-06, "loss": 2.9985, "step": 22943 }, { "epoch": 0.8084780253177938, "grad_norm": 4.376277923583984, "learning_rate": 1.8821482812383918e-06, "loss": 2.9711, "step": 22944 }, { "epoch": 0.8085132623307523, "grad_norm": 0.9241415858268738, "learning_rate": 1.8814818784518629e-06, "loss": 0.785, "step": 22945 }, { "epoch": 0.8085484993437106, "grad_norm": 17.44839859008789, "learning_rate": 1.880815581410067e-06, "loss": 3.4078, "step": 22946 }, { "epoch": 0.808583736356669, "grad_norm": 1.0197774171829224, "learning_rate": 1.880149390121685e-06, "loss": 0.7514, "step": 22947 }, { "epoch": 0.8086189733696274, "grad_norm": 3.011322498321533, "learning_rate": 1.8794833045953853e-06, "loss": 2.6094, "step": 22948 }, { "epoch": 0.8086542103825859, "grad_norm": 1.176165223121643, "learning_rate": 1.8788173248398534e-06, "loss": 0.8584, "step": 22949 }, { "epoch": 0.8086894473955443, "grad_norm": 1.4511268138885498, "learning_rate": 1.8781514508637633e-06, "loss": 0.7402, "step": 22950 }, { "epoch": 0.8087246844085026, "grad_norm": 8.82007884979248, "learning_rate": 1.8774856826757815e-06, "loss": 9.3836, "step": 22951 }, { "epoch": 0.8087599214214611, "grad_norm": 5.735105037689209, "learning_rate": 1.876820020284581e-06, "loss": 5.0393, "step": 22952 }, { "epoch": 0.8087951584344195, "grad_norm": 4.799253940582275, "learning_rate": 1.8761544636988371e-06, "loss": 4.6153, "step": 22953 }, { "epoch": 0.8088303954473779, "grad_norm": 6.579282283782959, "learning_rate": 1.8754890129272184e-06, "loss": 5.4415, "step": 22954 }, { "epoch": 0.8088656324603364, "grad_norm": 5.219445705413818, "learning_rate": 1.8748236679783883e-06, "loss": 3.0261, "step": 22955 }, { "epoch": 0.8089008694732948, "grad_norm": 4.283161163330078, "learning_rate": 1.8741584288610125e-06, "loss": 4.5173, "step": 22956 }, { "epoch": 0.8089361064862531, "grad_norm": 3.4596409797668457, "learning_rate": 1.8734932955837627e-06, "loss": 3.0438, "step": 22957 }, { "epoch": 0.8089713434992116, "grad_norm": 1.2673403024673462, "learning_rate": 1.8728282681552956e-06, "loss": 0.7722, "step": 22958 }, { "epoch": 0.80900658051217, "grad_norm": 12.65184211730957, "learning_rate": 1.8721633465842747e-06, "loss": 2.8794, "step": 22959 }, { "epoch": 0.8090418175251284, "grad_norm": 3.73026442527771, "learning_rate": 1.871498530879363e-06, "loss": 2.8031, "step": 22960 }, { "epoch": 0.8090770545380868, "grad_norm": 1.2678080797195435, "learning_rate": 1.8708338210492172e-06, "loss": 0.5809, "step": 22961 }, { "epoch": 0.8091122915510452, "grad_norm": 5.012516498565674, "learning_rate": 1.8701692171024977e-06, "loss": 2.7837, "step": 22962 }, { "epoch": 0.8091475285640036, "grad_norm": 0.9158727526664734, "learning_rate": 1.8695047190478588e-06, "loss": 1.1439, "step": 22963 }, { "epoch": 0.809182765576962, "grad_norm": 4.980392932891846, "learning_rate": 1.868840326893956e-06, "loss": 2.4409, "step": 22964 }, { "epoch": 0.8092180025899205, "grad_norm": 5.076459884643555, "learning_rate": 1.8681760406494442e-06, "loss": 2.8047, "step": 22965 }, { "epoch": 0.8092532396028789, "grad_norm": 3.781322717666626, "learning_rate": 1.867511860322978e-06, "loss": 2.4191, "step": 22966 }, { "epoch": 0.8092884766158372, "grad_norm": 15.630568504333496, "learning_rate": 1.8668477859232003e-06, "loss": 4.7609, "step": 22967 }, { "epoch": 0.8093237136287957, "grad_norm": 1.2501897811889648, "learning_rate": 1.8661838174587699e-06, "loss": 0.8601, "step": 22968 }, { "epoch": 0.8093589506417541, "grad_norm": 4.7234086990356445, "learning_rate": 1.8655199549383329e-06, "loss": 4.2753, "step": 22969 }, { "epoch": 0.8093941876547125, "grad_norm": 4.218266010284424, "learning_rate": 1.8648561983705317e-06, "loss": 2.9274, "step": 22970 }, { "epoch": 0.8094294246676709, "grad_norm": 4.882857322692871, "learning_rate": 1.8641925477640132e-06, "loss": 3.573, "step": 22971 }, { "epoch": 0.8094646616806294, "grad_norm": 1.290370225906372, "learning_rate": 1.8635290031274256e-06, "loss": 0.829, "step": 22972 }, { "epoch": 0.8094998986935877, "grad_norm": 1.5397546291351318, "learning_rate": 1.8628655644694104e-06, "loss": 1.1633, "step": 22973 }, { "epoch": 0.8095351357065461, "grad_norm": 4.049288749694824, "learning_rate": 1.8622022317986067e-06, "loss": 2.8259, "step": 22974 }, { "epoch": 0.8095703727195046, "grad_norm": 5.116827011108398, "learning_rate": 1.8615390051236526e-06, "loss": 3.0564, "step": 22975 }, { "epoch": 0.809605609732463, "grad_norm": 0.9260827898979187, "learning_rate": 1.8608758844531938e-06, "loss": 0.892, "step": 22976 }, { "epoch": 0.8096408467454214, "grad_norm": 3.9029150009155273, "learning_rate": 1.8602128697958622e-06, "loss": 3.3496, "step": 22977 }, { "epoch": 0.8096760837583798, "grad_norm": 0.9729259014129639, "learning_rate": 1.859549961160294e-06, "loss": 0.602, "step": 22978 }, { "epoch": 0.8097113207713382, "grad_norm": 4.474096775054932, "learning_rate": 1.858887158555125e-06, "loss": 3.1179, "step": 22979 }, { "epoch": 0.8097465577842966, "grad_norm": 1.282772421836853, "learning_rate": 1.8582244619889867e-06, "loss": 1.0823, "step": 22980 }, { "epoch": 0.809781794797255, "grad_norm": 1.1837778091430664, "learning_rate": 1.8575618714705123e-06, "loss": 0.9109, "step": 22981 }, { "epoch": 0.8098170318102135, "grad_norm": 3.8363189697265625, "learning_rate": 1.8568993870083329e-06, "loss": 3.3974, "step": 22982 }, { "epoch": 0.8098522688231719, "grad_norm": 1.5048083066940308, "learning_rate": 1.8562370086110749e-06, "loss": 0.7282, "step": 22983 }, { "epoch": 0.8098875058361302, "grad_norm": 0.8251479268074036, "learning_rate": 1.8555747362873677e-06, "loss": 0.7746, "step": 22984 }, { "epoch": 0.8099227428490887, "grad_norm": 1.9874839782714844, "learning_rate": 1.8549125700458392e-06, "loss": 0.9673, "step": 22985 }, { "epoch": 0.8099579798620471, "grad_norm": 3.964445114135742, "learning_rate": 1.8542505098951068e-06, "loss": 3.0764, "step": 22986 }, { "epoch": 0.8099932168750055, "grad_norm": 8.237889289855957, "learning_rate": 1.8535885558438028e-06, "loss": 4.1694, "step": 22987 }, { "epoch": 0.810028453887964, "grad_norm": 2.8141109943389893, "learning_rate": 1.8529267079005475e-06, "loss": 2.8411, "step": 22988 }, { "epoch": 0.8100636909009223, "grad_norm": 0.9108269810676575, "learning_rate": 1.8522649660739566e-06, "loss": 0.8071, "step": 22989 }, { "epoch": 0.8100989279138807, "grad_norm": 4.1251325607299805, "learning_rate": 1.85160333037265e-06, "loss": 5.0488, "step": 22990 }, { "epoch": 0.8101341649268392, "grad_norm": 1.2941011190414429, "learning_rate": 1.8509418008052504e-06, "loss": 0.6935, "step": 22991 }, { "epoch": 0.8101694019397976, "grad_norm": 1.6508822441101074, "learning_rate": 1.8502803773803745e-06, "loss": 0.8976, "step": 22992 }, { "epoch": 0.810204638952756, "grad_norm": 5.96970272064209, "learning_rate": 1.8496190601066333e-06, "loss": 5.2533, "step": 22993 }, { "epoch": 0.8102398759657143, "grad_norm": 1.6560962200164795, "learning_rate": 1.8489578489926406e-06, "loss": 0.6987, "step": 22994 }, { "epoch": 0.8102751129786728, "grad_norm": 3.793231964111328, "learning_rate": 1.8482967440470112e-06, "loss": 2.6395, "step": 22995 }, { "epoch": 0.8103103499916312, "grad_norm": 4.1454267501831055, "learning_rate": 1.8476357452783544e-06, "loss": 2.7234, "step": 22996 }, { "epoch": 0.8103455870045896, "grad_norm": 7.508594989776611, "learning_rate": 1.8469748526952801e-06, "loss": 3.0516, "step": 22997 }, { "epoch": 0.8103808240175481, "grad_norm": 0.7999280691146851, "learning_rate": 1.8463140663063972e-06, "loss": 0.9775, "step": 22998 }, { "epoch": 0.8104160610305065, "grad_norm": 4.74873161315918, "learning_rate": 1.845653386120313e-06, "loss": 2.9147, "step": 22999 }, { "epoch": 0.8104512980434648, "grad_norm": 3.297762393951416, "learning_rate": 1.8449928121456317e-06, "loss": 2.9703, "step": 23000 }, { "epoch": 0.8104865350564233, "grad_norm": 1.2190709114074707, "learning_rate": 1.8443323443909577e-06, "loss": 0.9034, "step": 23001 }, { "epoch": 0.8105217720693817, "grad_norm": 3.862337350845337, "learning_rate": 1.843671982864893e-06, "loss": 2.8585, "step": 23002 }, { "epoch": 0.8105570090823401, "grad_norm": 3.972562313079834, "learning_rate": 1.8430117275760407e-06, "loss": 2.6972, "step": 23003 }, { "epoch": 0.8105922460952985, "grad_norm": 4.561047077178955, "learning_rate": 1.8423515785330027e-06, "loss": 3.0171, "step": 23004 }, { "epoch": 0.8106274831082569, "grad_norm": 1.278509259223938, "learning_rate": 1.841691535744371e-06, "loss": 0.6878, "step": 23005 }, { "epoch": 0.8106627201212153, "grad_norm": 1.228147268295288, "learning_rate": 1.8410315992187433e-06, "loss": 1.0324, "step": 23006 }, { "epoch": 0.8106979571341737, "grad_norm": 1.497890830039978, "learning_rate": 1.8403717689647217e-06, "loss": 0.7477, "step": 23007 }, { "epoch": 0.8107331941471322, "grad_norm": 1.090535044670105, "learning_rate": 1.8397120449909e-06, "loss": 0.8541, "step": 23008 }, { "epoch": 0.8107684311600906, "grad_norm": 4.480315685272217, "learning_rate": 1.8390524273058653e-06, "loss": 2.5288, "step": 23009 }, { "epoch": 0.8108036681730489, "grad_norm": 3.382371664047241, "learning_rate": 1.838392915918209e-06, "loss": 3.4006, "step": 23010 }, { "epoch": 0.8108389051860074, "grad_norm": 0.9327142834663391, "learning_rate": 1.8377335108365312e-06, "loss": 0.678, "step": 23011 }, { "epoch": 0.8108741421989658, "grad_norm": 1.4047863483428955, "learning_rate": 1.8370742120694108e-06, "loss": 0.8565, "step": 23012 }, { "epoch": 0.8109093792119242, "grad_norm": 1.074199914932251, "learning_rate": 1.8364150196254382e-06, "loss": 1.1599, "step": 23013 }, { "epoch": 0.8109446162248826, "grad_norm": 3.9582979679107666, "learning_rate": 1.8357559335132003e-06, "loss": 3.2139, "step": 23014 }, { "epoch": 0.810979853237841, "grad_norm": 1.4360181093215942, "learning_rate": 1.835096953741281e-06, "loss": 1.0035, "step": 23015 }, { "epoch": 0.8110150902507994, "grad_norm": 4.507673263549805, "learning_rate": 1.8344380803182648e-06, "loss": 3.0518, "step": 23016 }, { "epoch": 0.8110503272637578, "grad_norm": 9.285937309265137, "learning_rate": 1.8337793132527315e-06, "loss": 3.4167, "step": 23017 }, { "epoch": 0.8110855642767163, "grad_norm": 3.90458083152771, "learning_rate": 1.8331206525532642e-06, "loss": 2.942, "step": 23018 }, { "epoch": 0.8111208012896747, "grad_norm": 8.163100242614746, "learning_rate": 1.832462098228439e-06, "loss": 3.2759, "step": 23019 }, { "epoch": 0.811156038302633, "grad_norm": 2.2894647121429443, "learning_rate": 1.8318036502868398e-06, "loss": 2.362, "step": 23020 }, { "epoch": 0.8111912753155915, "grad_norm": 7.54555606842041, "learning_rate": 1.8311453087370323e-06, "loss": 6.8994, "step": 23021 }, { "epoch": 0.8112265123285499, "grad_norm": 1.6441924571990967, "learning_rate": 1.8304870735876002e-06, "loss": 1.0931, "step": 23022 }, { "epoch": 0.8112617493415083, "grad_norm": 1.4635316133499146, "learning_rate": 1.8298289448471174e-06, "loss": 0.8948, "step": 23023 }, { "epoch": 0.8112969863544668, "grad_norm": 8.477564811706543, "learning_rate": 1.8291709225241505e-06, "loss": 5.2794, "step": 23024 }, { "epoch": 0.8113322233674252, "grad_norm": 1.2883449792861938, "learning_rate": 1.8285130066272716e-06, "loss": 0.8797, "step": 23025 }, { "epoch": 0.8113674603803835, "grad_norm": 1.2613656520843506, "learning_rate": 1.8278551971650527e-06, "loss": 0.9211, "step": 23026 }, { "epoch": 0.8114026973933419, "grad_norm": 3.9608213901519775, "learning_rate": 1.8271974941460646e-06, "loss": 3.2929, "step": 23027 }, { "epoch": 0.8114379344063004, "grad_norm": 4.399479866027832, "learning_rate": 1.8265398975788674e-06, "loss": 2.4718, "step": 23028 }, { "epoch": 0.8114731714192588, "grad_norm": 7.842363357543945, "learning_rate": 1.825882407472026e-06, "loss": 3.2194, "step": 23029 }, { "epoch": 0.8115084084322172, "grad_norm": 5.70798921585083, "learning_rate": 1.8252250238341129e-06, "loss": 2.931, "step": 23030 }, { "epoch": 0.8115436454451757, "grad_norm": 1.1898114681243896, "learning_rate": 1.8245677466736821e-06, "loss": 0.5659, "step": 23031 }, { "epoch": 0.811578882458134, "grad_norm": 4.216958999633789, "learning_rate": 1.8239105759992991e-06, "loss": 2.1439, "step": 23032 }, { "epoch": 0.8116141194710924, "grad_norm": 7.392083168029785, "learning_rate": 1.8232535118195215e-06, "loss": 5.152, "step": 23033 }, { "epoch": 0.8116493564840509, "grad_norm": 3.3977160453796387, "learning_rate": 1.8225965541429081e-06, "loss": 2.764, "step": 23034 }, { "epoch": 0.8116845934970093, "grad_norm": 1.2767891883850098, "learning_rate": 1.8219397029780172e-06, "loss": 0.7275, "step": 23035 }, { "epoch": 0.8117198305099677, "grad_norm": 1.6819816827774048, "learning_rate": 1.8212829583334024e-06, "loss": 0.7679, "step": 23036 }, { "epoch": 0.811755067522926, "grad_norm": 1.3493410348892212, "learning_rate": 1.8206263202176188e-06, "loss": 0.8656, "step": 23037 }, { "epoch": 0.8117903045358845, "grad_norm": 1.9347574710845947, "learning_rate": 1.8199697886392198e-06, "loss": 0.7912, "step": 23038 }, { "epoch": 0.8118255415488429, "grad_norm": 3.481743812561035, "learning_rate": 1.8193133636067595e-06, "loss": 2.7533, "step": 23039 }, { "epoch": 0.8118607785618013, "grad_norm": 3.8545565605163574, "learning_rate": 1.8186570451287788e-06, "loss": 2.8425, "step": 23040 }, { "epoch": 0.8118960155747598, "grad_norm": 7.938698768615723, "learning_rate": 1.8180008332138343e-06, "loss": 3.0894, "step": 23041 }, { "epoch": 0.8119312525877181, "grad_norm": 4.418581008911133, "learning_rate": 1.8173447278704747e-06, "loss": 2.9201, "step": 23042 }, { "epoch": 0.8119664896006765, "grad_norm": 3.377436399459839, "learning_rate": 1.8166887291072388e-06, "loss": 2.3459, "step": 23043 }, { "epoch": 0.812001726613635, "grad_norm": 5.087085723876953, "learning_rate": 1.816032836932673e-06, "loss": 3.4951, "step": 23044 }, { "epoch": 0.8120369636265934, "grad_norm": 3.9874112606048584, "learning_rate": 1.8153770513553236e-06, "loss": 3.299, "step": 23045 }, { "epoch": 0.8120722006395518, "grad_norm": 1.0096875429153442, "learning_rate": 1.8147213723837343e-06, "loss": 0.6894, "step": 23046 }, { "epoch": 0.8121074376525101, "grad_norm": 4.2518439292907715, "learning_rate": 1.8140658000264388e-06, "loss": 3.1018, "step": 23047 }, { "epoch": 0.8121426746654686, "grad_norm": 1.1058205366134644, "learning_rate": 1.8134103342919762e-06, "loss": 0.7776, "step": 23048 }, { "epoch": 0.812177911678427, "grad_norm": 1.2901328802108765, "learning_rate": 1.8127549751888918e-06, "loss": 1.037, "step": 23049 }, { "epoch": 0.8122131486913854, "grad_norm": 1.3682796955108643, "learning_rate": 1.8120997227257141e-06, "loss": 0.931, "step": 23050 }, { "epoch": 0.8122483857043439, "grad_norm": 1.1185964345932007, "learning_rate": 1.8114445769109812e-06, "loss": 0.7275, "step": 23051 }, { "epoch": 0.8122836227173023, "grad_norm": 4.015613079071045, "learning_rate": 1.8107895377532248e-06, "loss": 4.0634, "step": 23052 }, { "epoch": 0.8123188597302606, "grad_norm": 5.707907676696777, "learning_rate": 1.8101346052609782e-06, "loss": 3.0239, "step": 23053 }, { "epoch": 0.8123540967432191, "grad_norm": 5.756191730499268, "learning_rate": 1.809479779442772e-06, "loss": 5.5944, "step": 23054 }, { "epoch": 0.8123893337561775, "grad_norm": 1.6988935470581055, "learning_rate": 1.8088250603071344e-06, "loss": 0.7566, "step": 23055 }, { "epoch": 0.8124245707691359, "grad_norm": 0.9232789278030396, "learning_rate": 1.808170447862594e-06, "loss": 0.8735, "step": 23056 }, { "epoch": 0.8124598077820943, "grad_norm": 1.89383065700531, "learning_rate": 1.8075159421176769e-06, "loss": 2.6458, "step": 23057 }, { "epoch": 0.8124950447950527, "grad_norm": 1.8083810806274414, "learning_rate": 1.806861543080911e-06, "loss": 0.8534, "step": 23058 }, { "epoch": 0.8125302818080111, "grad_norm": 3.0773978233337402, "learning_rate": 1.806207250760812e-06, "loss": 2.5228, "step": 23059 }, { "epoch": 0.8125655188209695, "grad_norm": 3.4573745727539062, "learning_rate": 1.8055530651659092e-06, "loss": 1.0708, "step": 23060 }, { "epoch": 0.812600755833928, "grad_norm": 1.5401360988616943, "learning_rate": 1.804898986304725e-06, "loss": 0.9939, "step": 23061 }, { "epoch": 0.8126359928468864, "grad_norm": 5.369460105895996, "learning_rate": 1.8042450141857726e-06, "loss": 5.0024, "step": 23062 }, { "epoch": 0.8126712298598447, "grad_norm": 1.396663784980774, "learning_rate": 1.8035911488175707e-06, "loss": 0.7482, "step": 23063 }, { "epoch": 0.8127064668728032, "grad_norm": 3.177234649658203, "learning_rate": 1.80293739020864e-06, "loss": 2.8339, "step": 23064 }, { "epoch": 0.8127417038857616, "grad_norm": 3.2571609020233154, "learning_rate": 1.8022837383674962e-06, "loss": 2.68, "step": 23065 }, { "epoch": 0.81277694089872, "grad_norm": 3.302415370941162, "learning_rate": 1.8016301933026493e-06, "loss": 0.743, "step": 23066 }, { "epoch": 0.8128121779116785, "grad_norm": 2.0106658935546875, "learning_rate": 1.80097675502261e-06, "loss": 1.0105, "step": 23067 }, { "epoch": 0.8128474149246369, "grad_norm": 5.303804874420166, "learning_rate": 1.800323423535898e-06, "loss": 2.7162, "step": 23068 }, { "epoch": 0.8128826519375952, "grad_norm": 1.494740605354309, "learning_rate": 1.7996701988510157e-06, "loss": 0.9696, "step": 23069 }, { "epoch": 0.8129178889505536, "grad_norm": 5.320158004760742, "learning_rate": 1.799017080976474e-06, "loss": 5.0851, "step": 23070 }, { "epoch": 0.8129531259635121, "grad_norm": 1.571524739265442, "learning_rate": 1.7983640699207795e-06, "loss": 0.7928, "step": 23071 }, { "epoch": 0.8129883629764705, "grad_norm": 8.641905784606934, "learning_rate": 1.7977111656924373e-06, "loss": 6.8398, "step": 23072 }, { "epoch": 0.8130235999894289, "grad_norm": 2.9723997116088867, "learning_rate": 1.797058368299951e-06, "loss": 2.8414, "step": 23073 }, { "epoch": 0.8130588370023873, "grad_norm": 1.303078055381775, "learning_rate": 1.7964056777518258e-06, "loss": 0.8529, "step": 23074 }, { "epoch": 0.8130940740153457, "grad_norm": 1.4094821214675903, "learning_rate": 1.7957530940565616e-06, "loss": 0.7008, "step": 23075 }, { "epoch": 0.8131293110283041, "grad_norm": 3.006049394607544, "learning_rate": 1.795100617222657e-06, "loss": 2.7873, "step": 23076 }, { "epoch": 0.8131645480412626, "grad_norm": 5.297328948974609, "learning_rate": 1.7944482472586156e-06, "loss": 3.3196, "step": 23077 }, { "epoch": 0.813199785054221, "grad_norm": 4.37567138671875, "learning_rate": 1.7937959841729257e-06, "loss": 4.4829, "step": 23078 }, { "epoch": 0.8132350220671793, "grad_norm": 4.659286022186279, "learning_rate": 1.793143827974091e-06, "loss": 4.5567, "step": 23079 }, { "epoch": 0.8132702590801377, "grad_norm": 1.2291781902313232, "learning_rate": 1.7924917786706043e-06, "loss": 0.9634, "step": 23080 }, { "epoch": 0.8133054960930962, "grad_norm": 3.620361089706421, "learning_rate": 1.7918398362709589e-06, "loss": 2.4988, "step": 23081 }, { "epoch": 0.8133407331060546, "grad_norm": 1.2818061113357544, "learning_rate": 1.7911880007836401e-06, "loss": 0.9738, "step": 23082 }, { "epoch": 0.813375970119013, "grad_norm": 3.479896068572998, "learning_rate": 1.790536272217146e-06, "loss": 2.6341, "step": 23083 }, { "epoch": 0.8134112071319715, "grad_norm": 1.1789268255233765, "learning_rate": 1.789884650579966e-06, "loss": 1.0271, "step": 23084 }, { "epoch": 0.8134464441449298, "grad_norm": 5.0667877197265625, "learning_rate": 1.7892331358805803e-06, "loss": 2.8787, "step": 23085 }, { "epoch": 0.8134816811578882, "grad_norm": 36.531105041503906, "learning_rate": 1.7885817281274775e-06, "loss": 3.0852, "step": 23086 }, { "epoch": 0.8135169181708467, "grad_norm": 4.06082010269165, "learning_rate": 1.7879304273291488e-06, "loss": 2.8972, "step": 23087 }, { "epoch": 0.8135521551838051, "grad_norm": 3.1285464763641357, "learning_rate": 1.7872792334940692e-06, "loss": 2.6485, "step": 23088 }, { "epoch": 0.8135873921967635, "grad_norm": 2.165886402130127, "learning_rate": 1.7866281466307234e-06, "loss": 0.8668, "step": 23089 }, { "epoch": 0.8136226292097218, "grad_norm": 7.929114818572998, "learning_rate": 1.7859771667475922e-06, "loss": 7.4673, "step": 23090 }, { "epoch": 0.8136578662226803, "grad_norm": 1.4380254745483398, "learning_rate": 1.7853262938531558e-06, "loss": 0.7094, "step": 23091 }, { "epoch": 0.8136931032356387, "grad_norm": 1.8637118339538574, "learning_rate": 1.7846755279558903e-06, "loss": 0.7115, "step": 23092 }, { "epoch": 0.8137283402485971, "grad_norm": 7.337295055389404, "learning_rate": 1.7840248690642715e-06, "loss": 5.5559, "step": 23093 }, { "epoch": 0.8137635772615556, "grad_norm": 1.0521584749221802, "learning_rate": 1.7833743171867768e-06, "loss": 0.7306, "step": 23094 }, { "epoch": 0.813798814274514, "grad_norm": 5.549114227294922, "learning_rate": 1.7827238723318773e-06, "loss": 5.3149, "step": 23095 }, { "epoch": 0.8138340512874723, "grad_norm": 2.9965386390686035, "learning_rate": 1.7820735345080486e-06, "loss": 1.0006, "step": 23096 }, { "epoch": 0.8138692883004308, "grad_norm": 2.584470510482788, "learning_rate": 1.7814233037237549e-06, "loss": 2.3803, "step": 23097 }, { "epoch": 0.8139045253133892, "grad_norm": 2.1454358100891113, "learning_rate": 1.7807731799874716e-06, "loss": 0.6208, "step": 23098 }, { "epoch": 0.8139397623263476, "grad_norm": 3.2806832790374756, "learning_rate": 1.7801231633076644e-06, "loss": 3.0531, "step": 23099 }, { "epoch": 0.8139749993393061, "grad_norm": 4.870945930480957, "learning_rate": 1.7794732536928017e-06, "loss": 4.6588, "step": 23100 }, { "epoch": 0.8140102363522644, "grad_norm": 8.00719165802002, "learning_rate": 1.7788234511513437e-06, "loss": 2.7127, "step": 23101 }, { "epoch": 0.8140454733652228, "grad_norm": 4.088541030883789, "learning_rate": 1.7781737556917588e-06, "loss": 2.9693, "step": 23102 }, { "epoch": 0.8140807103781812, "grad_norm": 3.3516194820404053, "learning_rate": 1.7775241673225108e-06, "loss": 3.0585, "step": 23103 }, { "epoch": 0.8141159473911397, "grad_norm": 1.1987392902374268, "learning_rate": 1.7768746860520548e-06, "loss": 0.843, "step": 23104 }, { "epoch": 0.8141511844040981, "grad_norm": 5.7905988693237305, "learning_rate": 1.7762253118888517e-06, "loss": 4.5679, "step": 23105 }, { "epoch": 0.8141864214170564, "grad_norm": 4.095556735992432, "learning_rate": 1.7755760448413662e-06, "loss": 2.1931, "step": 23106 }, { "epoch": 0.8142216584300149, "grad_norm": 1.6739261150360107, "learning_rate": 1.7749268849180478e-06, "loss": 0.9332, "step": 23107 }, { "epoch": 0.8142568954429733, "grad_norm": 5.174335479736328, "learning_rate": 1.7742778321273545e-06, "loss": 2.9228, "step": 23108 }, { "epoch": 0.8142921324559317, "grad_norm": 4.021670341491699, "learning_rate": 1.7736288864777406e-06, "loss": 2.8505, "step": 23109 }, { "epoch": 0.8143273694688902, "grad_norm": 8.630509376525879, "learning_rate": 1.7729800479776583e-06, "loss": 5.2608, "step": 23110 }, { "epoch": 0.8143626064818485, "grad_norm": 4.4098591804504395, "learning_rate": 1.7723313166355583e-06, "loss": 2.8043, "step": 23111 }, { "epoch": 0.8143978434948069, "grad_norm": 2.207040786743164, "learning_rate": 1.7716826924598917e-06, "loss": 0.7241, "step": 23112 }, { "epoch": 0.8144330805077653, "grad_norm": 2.976196765899658, "learning_rate": 1.7710341754591055e-06, "loss": 2.4374, "step": 23113 }, { "epoch": 0.8144683175207238, "grad_norm": 5.985952377319336, "learning_rate": 1.770385765641648e-06, "loss": 5.1739, "step": 23114 }, { "epoch": 0.8145035545336822, "grad_norm": 4.47517204284668, "learning_rate": 1.7697374630159657e-06, "loss": 2.7392, "step": 23115 }, { "epoch": 0.8145387915466406, "grad_norm": 4.014225482940674, "learning_rate": 1.769089267590497e-06, "loss": 2.4408, "step": 23116 }, { "epoch": 0.814574028559599, "grad_norm": 0.9998860955238342, "learning_rate": 1.768441179373691e-06, "loss": 0.8661, "step": 23117 }, { "epoch": 0.8146092655725574, "grad_norm": 4.552295684814453, "learning_rate": 1.7677931983739883e-06, "loss": 2.9488, "step": 23118 }, { "epoch": 0.8146445025855158, "grad_norm": 2.218123435974121, "learning_rate": 1.7671453245998294e-06, "loss": 0.9185, "step": 23119 }, { "epoch": 0.8146797395984743, "grad_norm": 2.778809070587158, "learning_rate": 1.7664975580596466e-06, "loss": 2.5962, "step": 23120 }, { "epoch": 0.8147149766114327, "grad_norm": 1.2570029497146606, "learning_rate": 1.7658498987618833e-06, "loss": 0.7959, "step": 23121 }, { "epoch": 0.814750213624391, "grad_norm": 3.335076093673706, "learning_rate": 1.765202346714978e-06, "loss": 1.9611, "step": 23122 }, { "epoch": 0.8147854506373494, "grad_norm": 1.1456328630447388, "learning_rate": 1.7645549019273577e-06, "loss": 0.8162, "step": 23123 }, { "epoch": 0.8148206876503079, "grad_norm": 5.472574234008789, "learning_rate": 1.7639075644074566e-06, "loss": 4.5805, "step": 23124 }, { "epoch": 0.8148559246632663, "grad_norm": 1.72178053855896, "learning_rate": 1.7632603341637133e-06, "loss": 0.8583, "step": 23125 }, { "epoch": 0.8148911616762247, "grad_norm": 4.727935791015625, "learning_rate": 1.7626132112045513e-06, "loss": 2.9635, "step": 23126 }, { "epoch": 0.8149263986891832, "grad_norm": 7.9940266609191895, "learning_rate": 1.7619661955384016e-06, "loss": 7.0646, "step": 23127 }, { "epoch": 0.8149616357021415, "grad_norm": 2.9178080558776855, "learning_rate": 1.7613192871736917e-06, "loss": 2.8307, "step": 23128 }, { "epoch": 0.8149968727150999, "grad_norm": 1.8623337745666504, "learning_rate": 1.7606724861188473e-06, "loss": 0.9271, "step": 23129 }, { "epoch": 0.8150321097280584, "grad_norm": 1.999468207359314, "learning_rate": 1.7600257923822927e-06, "loss": 0.8596, "step": 23130 }, { "epoch": 0.8150673467410168, "grad_norm": 2.7438225746154785, "learning_rate": 1.7593792059724524e-06, "loss": 2.9562, "step": 23131 }, { "epoch": 0.8151025837539752, "grad_norm": 2.9316396713256836, "learning_rate": 1.7587327268977473e-06, "loss": 3.0866, "step": 23132 }, { "epoch": 0.8151378207669336, "grad_norm": 0.8566991686820984, "learning_rate": 1.7580863551665984e-06, "loss": 0.9292, "step": 23133 }, { "epoch": 0.815173057779892, "grad_norm": 4.513739109039307, "learning_rate": 1.7574400907874245e-06, "loss": 5.6082, "step": 23134 }, { "epoch": 0.8152082947928504, "grad_norm": 1.291448712348938, "learning_rate": 1.7567939337686434e-06, "loss": 0.8848, "step": 23135 }, { "epoch": 0.8152435318058088, "grad_norm": 1.1194560527801514, "learning_rate": 1.7561478841186709e-06, "loss": 1.067, "step": 23136 }, { "epoch": 0.8152787688187673, "grad_norm": 2.743652582168579, "learning_rate": 1.755501941845923e-06, "loss": 2.5342, "step": 23137 }, { "epoch": 0.8153140058317256, "grad_norm": 3.982675075531006, "learning_rate": 1.7548561069588143e-06, "loss": 2.5705, "step": 23138 }, { "epoch": 0.815349242844684, "grad_norm": 2.24006724357605, "learning_rate": 1.7542103794657505e-06, "loss": 2.7793, "step": 23139 }, { "epoch": 0.8153844798576425, "grad_norm": 3.1038622856140137, "learning_rate": 1.7535647593751492e-06, "loss": 2.9548, "step": 23140 }, { "epoch": 0.8154197168706009, "grad_norm": 5.215834140777588, "learning_rate": 1.7529192466954203e-06, "loss": 4.1748, "step": 23141 }, { "epoch": 0.8154549538835593, "grad_norm": 3.965456247329712, "learning_rate": 1.752273841434966e-06, "loss": 2.633, "step": 23142 }, { "epoch": 0.8154901908965178, "grad_norm": 1.3036280870437622, "learning_rate": 1.751628543602193e-06, "loss": 0.9428, "step": 23143 }, { "epoch": 0.8155254279094761, "grad_norm": 1.1120433807373047, "learning_rate": 1.7509833532055132e-06, "loss": 1.1422, "step": 23144 }, { "epoch": 0.8155606649224345, "grad_norm": 1.2197906970977783, "learning_rate": 1.7503382702533244e-06, "loss": 0.7738, "step": 23145 }, { "epoch": 0.8155959019353929, "grad_norm": 11.323989868164062, "learning_rate": 1.7496932947540302e-06, "loss": 5.1785, "step": 23146 }, { "epoch": 0.8156311389483514, "grad_norm": 7.5149312019348145, "learning_rate": 1.7490484267160313e-06, "loss": 2.8602, "step": 23147 }, { "epoch": 0.8156663759613098, "grad_norm": 1.3160638809204102, "learning_rate": 1.748403666147729e-06, "loss": 0.7022, "step": 23148 }, { "epoch": 0.8157016129742681, "grad_norm": 3.009000301361084, "learning_rate": 1.7477590130575183e-06, "loss": 2.874, "step": 23149 }, { "epoch": 0.8157368499872266, "grad_norm": 4.210899829864502, "learning_rate": 1.7471144674537988e-06, "loss": 4.8981, "step": 23150 }, { "epoch": 0.815772087000185, "grad_norm": 1.896186351776123, "learning_rate": 1.7464700293449643e-06, "loss": 0.721, "step": 23151 }, { "epoch": 0.8158073240131434, "grad_norm": 3.5738584995269775, "learning_rate": 1.7458256987394084e-06, "loss": 4.6929, "step": 23152 }, { "epoch": 0.8158425610261019, "grad_norm": 1.054425597190857, "learning_rate": 1.7451814756455243e-06, "loss": 1.0685, "step": 23153 }, { "epoch": 0.8158777980390602, "grad_norm": 3.0125410556793213, "learning_rate": 1.744537360071704e-06, "loss": 2.8664, "step": 23154 }, { "epoch": 0.8159130350520186, "grad_norm": 7.650835037231445, "learning_rate": 1.7438933520263357e-06, "loss": 5.2096, "step": 23155 }, { "epoch": 0.815948272064977, "grad_norm": 1.1304876804351807, "learning_rate": 1.7432494515178078e-06, "loss": 0.844, "step": 23156 }, { "epoch": 0.8159835090779355, "grad_norm": 1.029521107673645, "learning_rate": 1.7426056585545104e-06, "loss": 0.7745, "step": 23157 }, { "epoch": 0.8160187460908939, "grad_norm": 4.357435703277588, "learning_rate": 1.74196197314482e-06, "loss": 3.1724, "step": 23158 }, { "epoch": 0.8160539831038522, "grad_norm": 1.1542346477508545, "learning_rate": 1.7413183952971301e-06, "loss": 0.7125, "step": 23159 }, { "epoch": 0.8160892201168107, "grad_norm": 2.905478000640869, "learning_rate": 1.740674925019823e-06, "loss": 2.1135, "step": 23160 }, { "epoch": 0.8161244571297691, "grad_norm": 4.724579811096191, "learning_rate": 1.740031562321275e-06, "loss": 2.7373, "step": 23161 }, { "epoch": 0.8161596941427275, "grad_norm": 4.387088775634766, "learning_rate": 1.739388307209865e-06, "loss": 3.1401, "step": 23162 }, { "epoch": 0.816194931155686, "grad_norm": 0.9718416333198547, "learning_rate": 1.7387451596939785e-06, "loss": 0.6256, "step": 23163 }, { "epoch": 0.8162301681686444, "grad_norm": 1.5939308404922485, "learning_rate": 1.738102119781987e-06, "loss": 0.9815, "step": 23164 }, { "epoch": 0.8162654051816027, "grad_norm": 3.2189524173736572, "learning_rate": 1.7374591874822677e-06, "loss": 3.4611, "step": 23165 }, { "epoch": 0.8163006421945612, "grad_norm": 1.172046422958374, "learning_rate": 1.7368163628031953e-06, "loss": 0.8691, "step": 23166 }, { "epoch": 0.8163358792075196, "grad_norm": 1.0541085004806519, "learning_rate": 1.7361736457531419e-06, "loss": 0.6873, "step": 23167 }, { "epoch": 0.816371116220478, "grad_norm": 3.128054141998291, "learning_rate": 1.7355310363404798e-06, "loss": 2.7644, "step": 23168 }, { "epoch": 0.8164063532334364, "grad_norm": 1.379149079322815, "learning_rate": 1.734888534573579e-06, "loss": 0.7697, "step": 23169 }, { "epoch": 0.8164415902463948, "grad_norm": 2.76489520072937, "learning_rate": 1.7342461404608068e-06, "loss": 0.6238, "step": 23170 }, { "epoch": 0.8164768272593532, "grad_norm": 6.775177955627441, "learning_rate": 1.7336038540105327e-06, "loss": 5.4832, "step": 23171 }, { "epoch": 0.8165120642723116, "grad_norm": 4.755316257476807, "learning_rate": 1.7329616752311207e-06, "loss": 2.6443, "step": 23172 }, { "epoch": 0.8165473012852701, "grad_norm": 1.0405333042144775, "learning_rate": 1.7323196041309353e-06, "loss": 0.5997, "step": 23173 }, { "epoch": 0.8165825382982285, "grad_norm": 1.6702905893325806, "learning_rate": 1.7316776407183412e-06, "loss": 0.9014, "step": 23174 }, { "epoch": 0.8166177753111868, "grad_norm": 3.637047529220581, "learning_rate": 1.7310357850016978e-06, "loss": 2.5168, "step": 23175 }, { "epoch": 0.8166530123241453, "grad_norm": 6.407128810882568, "learning_rate": 1.7303940369893702e-06, "loss": 3.211, "step": 23176 }, { "epoch": 0.8166882493371037, "grad_norm": 0.7860403060913086, "learning_rate": 1.729752396689709e-06, "loss": 0.8844, "step": 23177 }, { "epoch": 0.8167234863500621, "grad_norm": 5.989867687225342, "learning_rate": 1.7291108641110777e-06, "loss": 4.9443, "step": 23178 }, { "epoch": 0.8167587233630205, "grad_norm": 7.049134254455566, "learning_rate": 1.728469439261834e-06, "loss": 4.7475, "step": 23179 }, { "epoch": 0.816793960375979, "grad_norm": 1.6036810874938965, "learning_rate": 1.727828122150327e-06, "loss": 0.8818, "step": 23180 }, { "epoch": 0.8168291973889373, "grad_norm": 3.6122593879699707, "learning_rate": 1.7271869127849096e-06, "loss": 3.0523, "step": 23181 }, { "epoch": 0.8168644344018957, "grad_norm": 2.268343687057495, "learning_rate": 1.726545811173942e-06, "loss": 0.8286, "step": 23182 }, { "epoch": 0.8168996714148542, "grad_norm": 5.395438194274902, "learning_rate": 1.7259048173257664e-06, "loss": 4.8023, "step": 23183 }, { "epoch": 0.8169349084278126, "grad_norm": 3.351529836654663, "learning_rate": 1.7252639312487341e-06, "loss": 0.8521, "step": 23184 }, { "epoch": 0.816970145440771, "grad_norm": 2.971282958984375, "learning_rate": 1.7246231529511937e-06, "loss": 2.8226, "step": 23185 }, { "epoch": 0.8170053824537294, "grad_norm": 3.332451105117798, "learning_rate": 1.7239824824414908e-06, "loss": 2.9271, "step": 23186 }, { "epoch": 0.8170406194666878, "grad_norm": 3.244657039642334, "learning_rate": 1.72334191972797e-06, "loss": 2.8892, "step": 23187 }, { "epoch": 0.8170758564796462, "grad_norm": 2.467916488647461, "learning_rate": 1.7227014648189744e-06, "loss": 2.8755, "step": 23188 }, { "epoch": 0.8171110934926046, "grad_norm": 1.446998119354248, "learning_rate": 1.7220611177228474e-06, "loss": 0.8052, "step": 23189 }, { "epoch": 0.8171463305055631, "grad_norm": 1.4140490293502808, "learning_rate": 1.721420878447928e-06, "loss": 0.9329, "step": 23190 }, { "epoch": 0.8171815675185214, "grad_norm": 6.2510881423950195, "learning_rate": 1.7207807470025572e-06, "loss": 4.9906, "step": 23191 }, { "epoch": 0.8172168045314798, "grad_norm": 1.5154913663864136, "learning_rate": 1.7201407233950717e-06, "loss": 1.0002, "step": 23192 }, { "epoch": 0.8172520415444383, "grad_norm": 3.374330759048462, "learning_rate": 1.719500807633807e-06, "loss": 3.1194, "step": 23193 }, { "epoch": 0.8172872785573967, "grad_norm": 0.9770318865776062, "learning_rate": 1.7188609997271e-06, "loss": 0.7907, "step": 23194 }, { "epoch": 0.8173225155703551, "grad_norm": 7.215816497802734, "learning_rate": 1.7182212996832847e-06, "loss": 5.269, "step": 23195 }, { "epoch": 0.8173577525833136, "grad_norm": 4.594951629638672, "learning_rate": 1.7175817075106881e-06, "loss": 2.3488, "step": 23196 }, { "epoch": 0.8173929895962719, "grad_norm": 1.1249483823776245, "learning_rate": 1.716942223217647e-06, "loss": 0.8108, "step": 23197 }, { "epoch": 0.8174282266092303, "grad_norm": 2.3237016201019287, "learning_rate": 1.7163028468124909e-06, "loss": 2.8233, "step": 23198 }, { "epoch": 0.8174634636221888, "grad_norm": 3.962606430053711, "learning_rate": 1.7156635783035425e-06, "loss": 5.1757, "step": 23199 }, { "epoch": 0.8174987006351472, "grad_norm": 3.9182088375091553, "learning_rate": 1.7150244176991294e-06, "loss": 2.2375, "step": 23200 }, { "epoch": 0.8175339376481056, "grad_norm": 3.3040452003479004, "learning_rate": 1.7143853650075837e-06, "loss": 2.7806, "step": 23201 }, { "epoch": 0.8175691746610639, "grad_norm": 3.3248913288116455, "learning_rate": 1.7137464202372211e-06, "loss": 2.9584, "step": 23202 }, { "epoch": 0.8176044116740224, "grad_norm": 5.201516151428223, "learning_rate": 1.7131075833963661e-06, "loss": 3.0731, "step": 23203 }, { "epoch": 0.8176396486869808, "grad_norm": 4.475726127624512, "learning_rate": 1.712468854493342e-06, "loss": 2.8777, "step": 23204 }, { "epoch": 0.8176748856999392, "grad_norm": 1.2622677087783813, "learning_rate": 1.7118302335364655e-06, "loss": 0.8912, "step": 23205 }, { "epoch": 0.8177101227128977, "grad_norm": 3.165344715118408, "learning_rate": 1.7111917205340567e-06, "loss": 2.6357, "step": 23206 }, { "epoch": 0.817745359725856, "grad_norm": 8.277894973754883, "learning_rate": 1.7105533154944308e-06, "loss": 3.1715, "step": 23207 }, { "epoch": 0.8177805967388144, "grad_norm": 1.10576331615448, "learning_rate": 1.709915018425905e-06, "loss": 0.7849, "step": 23208 }, { "epoch": 0.8178158337517729, "grad_norm": 1.5231540203094482, "learning_rate": 1.709276829336791e-06, "loss": 0.9923, "step": 23209 }, { "epoch": 0.8178510707647313, "grad_norm": 1.2172846794128418, "learning_rate": 1.7086387482354027e-06, "loss": 1.0511, "step": 23210 }, { "epoch": 0.8178863077776897, "grad_norm": 1.0890623331069946, "learning_rate": 1.7080007751300508e-06, "loss": 0.7742, "step": 23211 }, { "epoch": 0.817921544790648, "grad_norm": 3.9079365730285645, "learning_rate": 1.7073629100290457e-06, "loss": 2.8967, "step": 23212 }, { "epoch": 0.8179567818036065, "grad_norm": 5.219621181488037, "learning_rate": 1.7067251529406948e-06, "loss": 4.7246, "step": 23213 }, { "epoch": 0.8179920188165649, "grad_norm": 10.307284355163574, "learning_rate": 1.7060875038733072e-06, "loss": 4.3192, "step": 23214 }, { "epoch": 0.8180272558295233, "grad_norm": 1.4595832824707031, "learning_rate": 1.7054499628351806e-06, "loss": 0.9045, "step": 23215 }, { "epoch": 0.8180624928424818, "grad_norm": 0.8798251152038574, "learning_rate": 1.7048125298346286e-06, "loss": 0.956, "step": 23216 }, { "epoch": 0.8180977298554402, "grad_norm": 4.99702787399292, "learning_rate": 1.704175204879952e-06, "loss": 2.7472, "step": 23217 }, { "epoch": 0.8181329668683985, "grad_norm": 5.055918216705322, "learning_rate": 1.7035379879794478e-06, "loss": 2.7326, "step": 23218 }, { "epoch": 0.818168203881357, "grad_norm": 2.7899646759033203, "learning_rate": 1.702900879141416e-06, "loss": 0.6933, "step": 23219 }, { "epoch": 0.8182034408943154, "grad_norm": 3.930223226547241, "learning_rate": 1.7022638783741607e-06, "loss": 2.9111, "step": 23220 }, { "epoch": 0.8182386779072738, "grad_norm": 1.4515618085861206, "learning_rate": 1.7016269856859746e-06, "loss": 0.7274, "step": 23221 }, { "epoch": 0.8182739149202322, "grad_norm": 5.259538650512695, "learning_rate": 1.700990201085153e-06, "loss": 4.6675, "step": 23222 }, { "epoch": 0.8183091519331906, "grad_norm": 1.197989821434021, "learning_rate": 1.7003535245799919e-06, "loss": 0.9213, "step": 23223 }, { "epoch": 0.818344388946149, "grad_norm": 1.3655381202697754, "learning_rate": 1.6997169561787819e-06, "loss": 0.8623, "step": 23224 }, { "epoch": 0.8183796259591074, "grad_norm": 1.3209089040756226, "learning_rate": 1.6990804958898177e-06, "loss": 0.9156, "step": 23225 }, { "epoch": 0.8184148629720659, "grad_norm": 5.113448619842529, "learning_rate": 1.698444143721386e-06, "loss": 2.9251, "step": 23226 }, { "epoch": 0.8184500999850243, "grad_norm": 4.035817623138428, "learning_rate": 1.697807899681777e-06, "loss": 2.8439, "step": 23227 }, { "epoch": 0.8184853369979826, "grad_norm": 5.61490535736084, "learning_rate": 1.6971717637792783e-06, "loss": 5.0914, "step": 23228 }, { "epoch": 0.8185205740109411, "grad_norm": 3.8767621517181396, "learning_rate": 1.6965357360221745e-06, "loss": 3.5552, "step": 23229 }, { "epoch": 0.8185558110238995, "grad_norm": 2.8304131031036377, "learning_rate": 1.6958998164187501e-06, "loss": 0.8066, "step": 23230 }, { "epoch": 0.8185910480368579, "grad_norm": 1.143790364265442, "learning_rate": 1.6952640049772894e-06, "loss": 0.8483, "step": 23231 }, { "epoch": 0.8186262850498164, "grad_norm": 6.397582530975342, "learning_rate": 1.6946283017060728e-06, "loss": 6.9765, "step": 23232 }, { "epoch": 0.8186615220627748, "grad_norm": 3.856579065322876, "learning_rate": 1.693992706613382e-06, "loss": 2.4257, "step": 23233 }, { "epoch": 0.8186967590757331, "grad_norm": 1.1106938123703003, "learning_rate": 1.6933572197074899e-06, "loss": 0.8551, "step": 23234 }, { "epoch": 0.8187319960886915, "grad_norm": 3.9471356868743896, "learning_rate": 1.6927218409966805e-06, "loss": 2.8673, "step": 23235 }, { "epoch": 0.81876723310165, "grad_norm": 1.9283185005187988, "learning_rate": 1.6920865704892298e-06, "loss": 2.2796, "step": 23236 }, { "epoch": 0.8188024701146084, "grad_norm": 0.9615034461021423, "learning_rate": 1.6914514081934064e-06, "loss": 0.97, "step": 23237 }, { "epoch": 0.8188377071275668, "grad_norm": 4.9430670738220215, "learning_rate": 1.690816354117485e-06, "loss": 3.4108, "step": 23238 }, { "epoch": 0.8188729441405252, "grad_norm": 5.034846305847168, "learning_rate": 1.6901814082697442e-06, "loss": 4.5791, "step": 23239 }, { "epoch": 0.8189081811534836, "grad_norm": 1.5027531385421753, "learning_rate": 1.6895465706584468e-06, "loss": 0.8572, "step": 23240 }, { "epoch": 0.818943418166442, "grad_norm": 1.2182871103286743, "learning_rate": 1.6889118412918637e-06, "loss": 1.1488, "step": 23241 }, { "epoch": 0.8189786551794005, "grad_norm": 2.851868152618408, "learning_rate": 1.6882772201782616e-06, "loss": 2.5968, "step": 23242 }, { "epoch": 0.8190138921923589, "grad_norm": 1.0209730863571167, "learning_rate": 1.6876427073259084e-06, "loss": 1.1593, "step": 23243 }, { "epoch": 0.8190491292053173, "grad_norm": 3.1615095138549805, "learning_rate": 1.6870083027430672e-06, "loss": 2.8896, "step": 23244 }, { "epoch": 0.8190843662182756, "grad_norm": 3.7439379692077637, "learning_rate": 1.686374006438003e-06, "loss": 2.7526, "step": 23245 }, { "epoch": 0.8191196032312341, "grad_norm": 5.251092910766602, "learning_rate": 1.6857398184189756e-06, "loss": 5.0889, "step": 23246 }, { "epoch": 0.8191548402441925, "grad_norm": 2.4133923053741455, "learning_rate": 1.685105738694246e-06, "loss": 2.7869, "step": 23247 }, { "epoch": 0.8191900772571509, "grad_norm": 3.476771116256714, "learning_rate": 1.6844717672720734e-06, "loss": 2.947, "step": 23248 }, { "epoch": 0.8192253142701094, "grad_norm": 3.054518222808838, "learning_rate": 1.6838379041607145e-06, "loss": 2.4338, "step": 23249 }, { "epoch": 0.8192605512830677, "grad_norm": 2.0210683345794678, "learning_rate": 1.683204149368427e-06, "loss": 0.8549, "step": 23250 }, { "epoch": 0.8192957882960261, "grad_norm": 4.690201282501221, "learning_rate": 1.6825705029034645e-06, "loss": 2.7894, "step": 23251 }, { "epoch": 0.8193310253089846, "grad_norm": 2.898956537246704, "learning_rate": 1.6819369647740835e-06, "loss": 2.4151, "step": 23252 }, { "epoch": 0.819366262321943, "grad_norm": 1.626891016960144, "learning_rate": 1.6813035349885288e-06, "loss": 0.9081, "step": 23253 }, { "epoch": 0.8194014993349014, "grad_norm": 0.991412878036499, "learning_rate": 1.6806702135550568e-06, "loss": 0.928, "step": 23254 }, { "epoch": 0.8194367363478597, "grad_norm": 1.3275444507598877, "learning_rate": 1.6800370004819166e-06, "loss": 0.7263, "step": 23255 }, { "epoch": 0.8194719733608182, "grad_norm": 4.6598029136657715, "learning_rate": 1.6794038957773529e-06, "loss": 4.4463, "step": 23256 }, { "epoch": 0.8195072103737766, "grad_norm": 1.8557251691818237, "learning_rate": 1.678770899449611e-06, "loss": 0.8867, "step": 23257 }, { "epoch": 0.819542447386735, "grad_norm": 0.938205361366272, "learning_rate": 1.6781380115069422e-06, "loss": 0.88, "step": 23258 }, { "epoch": 0.8195776843996935, "grad_norm": 16.613624572753906, "learning_rate": 1.6775052319575846e-06, "loss": 2.791, "step": 23259 }, { "epoch": 0.8196129214126519, "grad_norm": 2.963590383529663, "learning_rate": 1.6768725608097802e-06, "loss": 2.9682, "step": 23260 }, { "epoch": 0.8196481584256102, "grad_norm": 1.3406627178192139, "learning_rate": 1.6762399980717714e-06, "loss": 0.8078, "step": 23261 }, { "epoch": 0.8196833954385687, "grad_norm": 1.1462713479995728, "learning_rate": 1.675607543751797e-06, "loss": 0.7925, "step": 23262 }, { "epoch": 0.8197186324515271, "grad_norm": 1.1294664144515991, "learning_rate": 1.6749751978580952e-06, "loss": 0.7217, "step": 23263 }, { "epoch": 0.8197538694644855, "grad_norm": 3.930663824081421, "learning_rate": 1.6743429603989015e-06, "loss": 3.0856, "step": 23264 }, { "epoch": 0.8197891064774439, "grad_norm": 5.734065532684326, "learning_rate": 1.6737108313824502e-06, "loss": 2.4075, "step": 23265 }, { "epoch": 0.8198243434904023, "grad_norm": 0.7439548373222351, "learning_rate": 1.6730788108169772e-06, "loss": 0.8456, "step": 23266 }, { "epoch": 0.8198595805033607, "grad_norm": 1.7291738986968994, "learning_rate": 1.6724468987107133e-06, "loss": 0.7679, "step": 23267 }, { "epoch": 0.8198948175163191, "grad_norm": 4.103090763092041, "learning_rate": 1.6718150950718892e-06, "loss": 4.9225, "step": 23268 }, { "epoch": 0.8199300545292776, "grad_norm": 1.4997191429138184, "learning_rate": 1.6711833999087334e-06, "loss": 1.0613, "step": 23269 }, { "epoch": 0.819965291542236, "grad_norm": 3.7488691806793213, "learning_rate": 1.6705518132294762e-06, "loss": 2.9889, "step": 23270 }, { "epoch": 0.8200005285551943, "grad_norm": 7.417044162750244, "learning_rate": 1.6699203350423443e-06, "loss": 5.3953, "step": 23271 }, { "epoch": 0.8200357655681528, "grad_norm": 4.482786178588867, "learning_rate": 1.6692889653555556e-06, "loss": 4.9854, "step": 23272 }, { "epoch": 0.8200710025811112, "grad_norm": 0.9114459753036499, "learning_rate": 1.6686577041773421e-06, "loss": 1.0105, "step": 23273 }, { "epoch": 0.8201062395940696, "grad_norm": 4.751440525054932, "learning_rate": 1.6680265515159267e-06, "loss": 4.9688, "step": 23274 }, { "epoch": 0.8201414766070281, "grad_norm": 4.703347206115723, "learning_rate": 1.6673955073795233e-06, "loss": 4.597, "step": 23275 }, { "epoch": 0.8201767136199865, "grad_norm": 5.839388370513916, "learning_rate": 1.6667645717763525e-06, "loss": 2.8063, "step": 23276 }, { "epoch": 0.8202119506329448, "grad_norm": 4.072858810424805, "learning_rate": 1.6661337447146398e-06, "loss": 2.4674, "step": 23277 }, { "epoch": 0.8202471876459032, "grad_norm": 4.04490327835083, "learning_rate": 1.665503026202595e-06, "loss": 2.8492, "step": 23278 }, { "epoch": 0.8202824246588617, "grad_norm": 5.481987476348877, "learning_rate": 1.664872416248433e-06, "loss": 2.9684, "step": 23279 }, { "epoch": 0.8203176616718201, "grad_norm": 1.2873848676681519, "learning_rate": 1.6642419148603683e-06, "loss": 0.85, "step": 23280 }, { "epoch": 0.8203528986847785, "grad_norm": 1.321578860282898, "learning_rate": 1.6636115220466198e-06, "loss": 0.761, "step": 23281 }, { "epoch": 0.8203881356977369, "grad_norm": 4.251269340515137, "learning_rate": 1.6629812378153897e-06, "loss": 3.2828, "step": 23282 }, { "epoch": 0.8204233727106953, "grad_norm": 2.2421953678131104, "learning_rate": 1.6623510621748917e-06, "loss": 0.9865, "step": 23283 }, { "epoch": 0.8204586097236537, "grad_norm": 3.9198689460754395, "learning_rate": 1.6617209951333325e-06, "loss": 4.3905, "step": 23284 }, { "epoch": 0.8204938467366122, "grad_norm": 1.1260225772857666, "learning_rate": 1.6610910366989197e-06, "loss": 0.7087, "step": 23285 }, { "epoch": 0.8205290837495706, "grad_norm": 9.050140380859375, "learning_rate": 1.660461186879858e-06, "loss": 4.6004, "step": 23286 }, { "epoch": 0.8205643207625289, "grad_norm": 1.7360025644302368, "learning_rate": 1.6598314456843522e-06, "loss": 0.8604, "step": 23287 }, { "epoch": 0.8205995577754873, "grad_norm": 1.3740261793136597, "learning_rate": 1.6592018131206044e-06, "loss": 1.0277, "step": 23288 }, { "epoch": 0.8206347947884458, "grad_norm": 2.767817735671997, "learning_rate": 1.6585722891968147e-06, "loss": 2.8994, "step": 23289 }, { "epoch": 0.8206700318014042, "grad_norm": 2.3618786334991455, "learning_rate": 1.6579428739211867e-06, "loss": 2.6072, "step": 23290 }, { "epoch": 0.8207052688143626, "grad_norm": 1.116749882698059, "learning_rate": 1.6573135673019103e-06, "loss": 0.9576, "step": 23291 }, { "epoch": 0.820740505827321, "grad_norm": 1.1724392175674438, "learning_rate": 1.656684369347189e-06, "loss": 0.8211, "step": 23292 }, { "epoch": 0.8207757428402794, "grad_norm": 1.1229050159454346, "learning_rate": 1.6560552800652197e-06, "loss": 0.9914, "step": 23293 }, { "epoch": 0.8208109798532378, "grad_norm": 5.974658489227295, "learning_rate": 1.65542629946419e-06, "loss": 4.7332, "step": 23294 }, { "epoch": 0.8208462168661963, "grad_norm": 2.2939958572387695, "learning_rate": 1.6547974275522948e-06, "loss": 0.9009, "step": 23295 }, { "epoch": 0.8208814538791547, "grad_norm": 4.290991306304932, "learning_rate": 1.6541686643377296e-06, "loss": 4.3332, "step": 23296 }, { "epoch": 0.820916690892113, "grad_norm": 1.0303305387496948, "learning_rate": 1.6535400098286791e-06, "loss": 0.9692, "step": 23297 }, { "epoch": 0.8209519279050714, "grad_norm": 1.184285283088684, "learning_rate": 1.6529114640333322e-06, "loss": 0.9297, "step": 23298 }, { "epoch": 0.8209871649180299, "grad_norm": 4.4945268630981445, "learning_rate": 1.6522830269598777e-06, "loss": 5.2171, "step": 23299 }, { "epoch": 0.8210224019309883, "grad_norm": 4.641103744506836, "learning_rate": 1.6516546986164994e-06, "loss": 4.4602, "step": 23300 }, { "epoch": 0.8210576389439467, "grad_norm": 1.0387277603149414, "learning_rate": 1.6510264790113818e-06, "loss": 0.7204, "step": 23301 }, { "epoch": 0.8210928759569052, "grad_norm": 6.48966121673584, "learning_rate": 1.6503983681527092e-06, "loss": 9.0638, "step": 23302 }, { "epoch": 0.8211281129698635, "grad_norm": 9.620450019836426, "learning_rate": 1.6497703660486596e-06, "loss": 2.7691, "step": 23303 }, { "epoch": 0.8211633499828219, "grad_norm": 1.4492907524108887, "learning_rate": 1.6491424727074156e-06, "loss": 0.8714, "step": 23304 }, { "epoch": 0.8211985869957804, "grad_norm": 0.9113817811012268, "learning_rate": 1.648514688137155e-06, "loss": 1.0609, "step": 23305 }, { "epoch": 0.8212338240087388, "grad_norm": 2.7576491832733154, "learning_rate": 1.6478870123460545e-06, "loss": 2.8281, "step": 23306 }, { "epoch": 0.8212690610216972, "grad_norm": 4.276280403137207, "learning_rate": 1.6472594453422886e-06, "loss": 3.0088, "step": 23307 }, { "epoch": 0.8213042980346557, "grad_norm": 2.360567569732666, "learning_rate": 1.6466319871340331e-06, "loss": 2.5449, "step": 23308 }, { "epoch": 0.821339535047614, "grad_norm": 1.449913501739502, "learning_rate": 1.6460046377294614e-06, "loss": 0.8895, "step": 23309 }, { "epoch": 0.8213747720605724, "grad_norm": 2.231684684753418, "learning_rate": 1.6453773971367427e-06, "loss": 2.578, "step": 23310 }, { "epoch": 0.8214100090735308, "grad_norm": 4.119232177734375, "learning_rate": 1.6447502653640436e-06, "loss": 4.7794, "step": 23311 }, { "epoch": 0.8214452460864893, "grad_norm": 0.9818999767303467, "learning_rate": 1.6441232424195419e-06, "loss": 1.1255, "step": 23312 }, { "epoch": 0.8214804830994477, "grad_norm": 5.053558349609375, "learning_rate": 1.6434963283113969e-06, "loss": 2.9422, "step": 23313 }, { "epoch": 0.821515720112406, "grad_norm": 3.9185545444488525, "learning_rate": 1.6428695230477776e-06, "loss": 2.8082, "step": 23314 }, { "epoch": 0.8215509571253645, "grad_norm": 1.3712365627288818, "learning_rate": 1.6422428266368462e-06, "loss": 0.7856, "step": 23315 }, { "epoch": 0.8215861941383229, "grad_norm": 2.8346521854400635, "learning_rate": 1.6416162390867673e-06, "loss": 3.0076, "step": 23316 }, { "epoch": 0.8216214311512813, "grad_norm": 2.975703239440918, "learning_rate": 1.6409897604057012e-06, "loss": 3.2689, "step": 23317 }, { "epoch": 0.8216566681642398, "grad_norm": 3.315927267074585, "learning_rate": 1.640363390601808e-06, "loss": 2.5026, "step": 23318 }, { "epoch": 0.8216919051771981, "grad_norm": 1.2238291501998901, "learning_rate": 1.6397371296832465e-06, "loss": 0.7866, "step": 23319 }, { "epoch": 0.8217271421901565, "grad_norm": 3.0561437606811523, "learning_rate": 1.6391109776581736e-06, "loss": 3.0362, "step": 23320 }, { "epoch": 0.8217623792031149, "grad_norm": 3.8175697326660156, "learning_rate": 1.6384849345347463e-06, "loss": 0.7094, "step": 23321 }, { "epoch": 0.8217976162160734, "grad_norm": 2.141996145248413, "learning_rate": 1.637859000321117e-06, "loss": 0.8443, "step": 23322 }, { "epoch": 0.8218328532290318, "grad_norm": 3.078449010848999, "learning_rate": 1.63723317502544e-06, "loss": 2.5996, "step": 23323 }, { "epoch": 0.8218680902419901, "grad_norm": 1.4130891561508179, "learning_rate": 1.6366074586558655e-06, "loss": 1.0466, "step": 23324 }, { "epoch": 0.8219033272549486, "grad_norm": 0.9000744223594666, "learning_rate": 1.6359818512205473e-06, "loss": 1.1028, "step": 23325 }, { "epoch": 0.821938564267907, "grad_norm": 2.812817096710205, "learning_rate": 1.6353563527276251e-06, "loss": 2.6559, "step": 23326 }, { "epoch": 0.8219738012808654, "grad_norm": 4.422623157501221, "learning_rate": 1.6347309631852548e-06, "loss": 4.934, "step": 23327 }, { "epoch": 0.8220090382938239, "grad_norm": 5.812335014343262, "learning_rate": 1.6341056826015823e-06, "loss": 6.6322, "step": 23328 }, { "epoch": 0.8220442753067823, "grad_norm": 3.0765371322631836, "learning_rate": 1.6334805109847452e-06, "loss": 3.0771, "step": 23329 }, { "epoch": 0.8220795123197406, "grad_norm": 1.2248820066452026, "learning_rate": 1.6328554483428883e-06, "loss": 0.7413, "step": 23330 }, { "epoch": 0.822114749332699, "grad_norm": 1.2055221796035767, "learning_rate": 1.6322304946841605e-06, "loss": 0.8251, "step": 23331 }, { "epoch": 0.8221499863456575, "grad_norm": 7.348846912384033, "learning_rate": 1.6316056500166933e-06, "loss": 4.4691, "step": 23332 }, { "epoch": 0.8221852233586159, "grad_norm": 4.655500411987305, "learning_rate": 1.6309809143486287e-06, "loss": 2.3569, "step": 23333 }, { "epoch": 0.8222204603715743, "grad_norm": 0.9796566963195801, "learning_rate": 1.630356287688103e-06, "loss": 1.2296, "step": 23334 }, { "epoch": 0.8222556973845327, "grad_norm": 5.502994060516357, "learning_rate": 1.6297317700432536e-06, "loss": 5.0065, "step": 23335 }, { "epoch": 0.8222909343974911, "grad_norm": 4.204957485198975, "learning_rate": 1.629107361422214e-06, "loss": 2.3896, "step": 23336 }, { "epoch": 0.8223261714104495, "grad_norm": 3.0968852043151855, "learning_rate": 1.628483061833117e-06, "loss": 2.8251, "step": 23337 }, { "epoch": 0.822361408423408, "grad_norm": 1.24972403049469, "learning_rate": 1.6278588712840947e-06, "loss": 1.0826, "step": 23338 }, { "epoch": 0.8223966454363664, "grad_norm": 7.136420249938965, "learning_rate": 1.6272347897832774e-06, "loss": 2.9226, "step": 23339 }, { "epoch": 0.8224318824493247, "grad_norm": 1.1044081449508667, "learning_rate": 1.6266108173387928e-06, "loss": 0.8673, "step": 23340 }, { "epoch": 0.8224671194622832, "grad_norm": 7.524779796600342, "learning_rate": 1.625986953958769e-06, "loss": 3.3323, "step": 23341 }, { "epoch": 0.8225023564752416, "grad_norm": 2.235438585281372, "learning_rate": 1.6253631996513331e-06, "loss": 0.8588, "step": 23342 }, { "epoch": 0.8225375934882, "grad_norm": 1.1474945545196533, "learning_rate": 1.624739554424607e-06, "loss": 0.855, "step": 23343 }, { "epoch": 0.8225728305011584, "grad_norm": 6.073083400726318, "learning_rate": 1.6241160182867178e-06, "loss": 3.5532, "step": 23344 }, { "epoch": 0.8226080675141169, "grad_norm": 1.1844590902328491, "learning_rate": 1.6234925912457788e-06, "loss": 0.7306, "step": 23345 }, { "epoch": 0.8226433045270752, "grad_norm": 1.1456836462020874, "learning_rate": 1.6228692733099194e-06, "loss": 0.7629, "step": 23346 }, { "epoch": 0.8226785415400336, "grad_norm": 2.4852871894836426, "learning_rate": 1.6222460644872563e-06, "loss": 2.3653, "step": 23347 }, { "epoch": 0.8227137785529921, "grad_norm": 4.563925743103027, "learning_rate": 1.6216229647859038e-06, "loss": 2.6081, "step": 23348 }, { "epoch": 0.8227490155659505, "grad_norm": 1.0250872373580933, "learning_rate": 1.620999974213976e-06, "loss": 1.0611, "step": 23349 }, { "epoch": 0.8227842525789089, "grad_norm": 1.2639824151992798, "learning_rate": 1.6203770927795959e-06, "loss": 0.7458, "step": 23350 }, { "epoch": 0.8228194895918673, "grad_norm": 3.3302242755889893, "learning_rate": 1.6197543204908694e-06, "loss": 2.5329, "step": 23351 }, { "epoch": 0.8228547266048257, "grad_norm": 1.1411656141281128, "learning_rate": 1.6191316573559102e-06, "loss": 0.8191, "step": 23352 }, { "epoch": 0.8228899636177841, "grad_norm": 1.3921431303024292, "learning_rate": 1.618509103382827e-06, "loss": 0.7913, "step": 23353 }, { "epoch": 0.8229252006307425, "grad_norm": 1.2960588932037354, "learning_rate": 1.6178866585797349e-06, "loss": 0.861, "step": 23354 }, { "epoch": 0.822960437643701, "grad_norm": 1.7382481098175049, "learning_rate": 1.6172643229547335e-06, "loss": 0.6329, "step": 23355 }, { "epoch": 0.8229956746566593, "grad_norm": 3.3511040210723877, "learning_rate": 1.6166420965159334e-06, "loss": 2.9395, "step": 23356 }, { "epoch": 0.8230309116696177, "grad_norm": 7.336299896240234, "learning_rate": 1.616019979271437e-06, "loss": 7.2352, "step": 23357 }, { "epoch": 0.8230661486825762, "grad_norm": 3.7899563312530518, "learning_rate": 1.6153979712293488e-06, "loss": 3.0605, "step": 23358 }, { "epoch": 0.8231013856955346, "grad_norm": 1.3661459684371948, "learning_rate": 1.614776072397769e-06, "loss": 1.1652, "step": 23359 }, { "epoch": 0.823136622708493, "grad_norm": 1.256030797958374, "learning_rate": 1.6141542827848e-06, "loss": 0.8776, "step": 23360 }, { "epoch": 0.8231718597214515, "grad_norm": 1.979272484779358, "learning_rate": 1.6135326023985398e-06, "loss": 0.7198, "step": 23361 }, { "epoch": 0.8232070967344098, "grad_norm": 3.8881473541259766, "learning_rate": 1.6129110312470852e-06, "loss": 3.09, "step": 23362 }, { "epoch": 0.8232423337473682, "grad_norm": 1.1271439790725708, "learning_rate": 1.6122895693385353e-06, "loss": 0.9042, "step": 23363 }, { "epoch": 0.8232775707603266, "grad_norm": 1.971737265586853, "learning_rate": 1.611668216680977e-06, "loss": 0.9677, "step": 23364 }, { "epoch": 0.8233128077732851, "grad_norm": 2.5396196842193604, "learning_rate": 1.611046973282512e-06, "loss": 2.8184, "step": 23365 }, { "epoch": 0.8233480447862435, "grad_norm": 5.023652076721191, "learning_rate": 1.6104258391512296e-06, "loss": 4.8162, "step": 23366 }, { "epoch": 0.8233832817992018, "grad_norm": 4.748095989227295, "learning_rate": 1.6098048142952182e-06, "loss": 4.7655, "step": 23367 }, { "epoch": 0.8234185188121603, "grad_norm": 1.1579992771148682, "learning_rate": 1.609183898722565e-06, "loss": 0.6395, "step": 23368 }, { "epoch": 0.8234537558251187, "grad_norm": 1.3387993574142456, "learning_rate": 1.6085630924413652e-06, "loss": 1.2076, "step": 23369 }, { "epoch": 0.8234889928380771, "grad_norm": 1.1561452150344849, "learning_rate": 1.6079423954596973e-06, "loss": 0.9618, "step": 23370 }, { "epoch": 0.8235242298510356, "grad_norm": 3.6656956672668457, "learning_rate": 1.60732180778565e-06, "loss": 3.2859, "step": 23371 }, { "epoch": 0.823559466863994, "grad_norm": 5.707005977630615, "learning_rate": 1.6067013294273015e-06, "loss": 4.1009, "step": 23372 }, { "epoch": 0.8235947038769523, "grad_norm": 1.2845447063446045, "learning_rate": 1.6060809603927418e-06, "loss": 0.7477, "step": 23373 }, { "epoch": 0.8236299408899108, "grad_norm": 4.392851829528809, "learning_rate": 1.6054607006900458e-06, "loss": 3.0693, "step": 23374 }, { "epoch": 0.8236651779028692, "grad_norm": 4.074159145355225, "learning_rate": 1.6048405503272923e-06, "loss": 2.8809, "step": 23375 }, { "epoch": 0.8237004149158276, "grad_norm": 1.3553706407546997, "learning_rate": 1.6042205093125607e-06, "loss": 0.6387, "step": 23376 }, { "epoch": 0.823735651928786, "grad_norm": 3.283057451248169, "learning_rate": 1.6036005776539264e-06, "loss": 2.7661, "step": 23377 }, { "epoch": 0.8237708889417444, "grad_norm": 1.0458455085754395, "learning_rate": 1.6029807553594633e-06, "loss": 0.9612, "step": 23378 }, { "epoch": 0.8238061259547028, "grad_norm": 1.5190389156341553, "learning_rate": 1.602361042437246e-06, "loss": 0.7847, "step": 23379 }, { "epoch": 0.8238413629676612, "grad_norm": 2.7601945400238037, "learning_rate": 1.6017414388953467e-06, "loss": 0.7556, "step": 23380 }, { "epoch": 0.8238765999806197, "grad_norm": 0.9293733239173889, "learning_rate": 1.6011219447418336e-06, "loss": 0.5467, "step": 23381 }, { "epoch": 0.8239118369935781, "grad_norm": 3.0814359188079834, "learning_rate": 1.60050255998478e-06, "loss": 2.7298, "step": 23382 }, { "epoch": 0.8239470740065364, "grad_norm": 1.4590542316436768, "learning_rate": 1.5998832846322454e-06, "loss": 0.8323, "step": 23383 }, { "epoch": 0.8239823110194949, "grad_norm": 4.057697296142578, "learning_rate": 1.5992641186923041e-06, "loss": 3.0182, "step": 23384 }, { "epoch": 0.8240175480324533, "grad_norm": 0.8003936409950256, "learning_rate": 1.5986450621730198e-06, "loss": 0.6996, "step": 23385 }, { "epoch": 0.8240527850454117, "grad_norm": 0.9918831586837769, "learning_rate": 1.5980261150824506e-06, "loss": 0.8311, "step": 23386 }, { "epoch": 0.8240880220583701, "grad_norm": 0.9431636333465576, "learning_rate": 1.5974072774286598e-06, "loss": 0.5789, "step": 23387 }, { "epoch": 0.8241232590713286, "grad_norm": 5.814202308654785, "learning_rate": 1.5967885492197144e-06, "loss": 4.8453, "step": 23388 }, { "epoch": 0.8241584960842869, "grad_norm": 4.813756465911865, "learning_rate": 1.5961699304636657e-06, "loss": 2.7343, "step": 23389 }, { "epoch": 0.8241937330972453, "grad_norm": 0.8655018210411072, "learning_rate": 1.595551421168574e-06, "loss": 1.0378, "step": 23390 }, { "epoch": 0.8242289701102038, "grad_norm": 2.5999317169189453, "learning_rate": 1.594933021342494e-06, "loss": 2.7513, "step": 23391 }, { "epoch": 0.8242642071231622, "grad_norm": 1.2742239236831665, "learning_rate": 1.5943147309934858e-06, "loss": 0.8126, "step": 23392 }, { "epoch": 0.8242994441361206, "grad_norm": 5.818105220794678, "learning_rate": 1.5936965501295976e-06, "loss": 4.9042, "step": 23393 }, { "epoch": 0.824334681149079, "grad_norm": 1.082187294960022, "learning_rate": 1.5930784787588828e-06, "loss": 1.1686, "step": 23394 }, { "epoch": 0.8243699181620374, "grad_norm": 0.8373277187347412, "learning_rate": 1.5924605168893902e-06, "loss": 1.0032, "step": 23395 }, { "epoch": 0.8244051551749958, "grad_norm": 5.973605155944824, "learning_rate": 1.5918426645291719e-06, "loss": 4.289, "step": 23396 }, { "epoch": 0.8244403921879542, "grad_norm": 3.9122793674468994, "learning_rate": 1.5912249216862729e-06, "loss": 2.9287, "step": 23397 }, { "epoch": 0.8244756292009127, "grad_norm": 6.254510879516602, "learning_rate": 1.5906072883687406e-06, "loss": 5.268, "step": 23398 }, { "epoch": 0.824510866213871, "grad_norm": 1.243506908416748, "learning_rate": 1.5899897645846195e-06, "loss": 0.8219, "step": 23399 }, { "epoch": 0.8245461032268294, "grad_norm": 1.0820636749267578, "learning_rate": 1.5893723503419534e-06, "loss": 0.9073, "step": 23400 }, { "epoch": 0.8245813402397879, "grad_norm": 1.1846225261688232, "learning_rate": 1.5887550456487855e-06, "loss": 0.6912, "step": 23401 }, { "epoch": 0.8246165772527463, "grad_norm": 1.78728449344635, "learning_rate": 1.5881378505131505e-06, "loss": 0.8769, "step": 23402 }, { "epoch": 0.8246518142657047, "grad_norm": 1.234183669090271, "learning_rate": 1.5875207649430935e-06, "loss": 0.923, "step": 23403 }, { "epoch": 0.8246870512786632, "grad_norm": 3.829296827316284, "learning_rate": 1.586903788946652e-06, "loss": 2.8423, "step": 23404 }, { "epoch": 0.8247222882916215, "grad_norm": 5.748762130737305, "learning_rate": 1.586286922531859e-06, "loss": 4.6271, "step": 23405 }, { "epoch": 0.8247575253045799, "grad_norm": 1.1536484956741333, "learning_rate": 1.5856701657067476e-06, "loss": 0.7325, "step": 23406 }, { "epoch": 0.8247927623175384, "grad_norm": 2.219606637954712, "learning_rate": 1.5850535184793593e-06, "loss": 0.8301, "step": 23407 }, { "epoch": 0.8248279993304968, "grad_norm": 3.908504009246826, "learning_rate": 1.5844369808577175e-06, "loss": 3.3574, "step": 23408 }, { "epoch": 0.8248632363434552, "grad_norm": 5.832359790802002, "learning_rate": 1.5838205528498574e-06, "loss": 4.962, "step": 23409 }, { "epoch": 0.8248984733564135, "grad_norm": 1.2899354696273804, "learning_rate": 1.5832042344638033e-06, "loss": 0.8659, "step": 23410 }, { "epoch": 0.824933710369372, "grad_norm": 4.370452404022217, "learning_rate": 1.582588025707591e-06, "loss": 2.789, "step": 23411 }, { "epoch": 0.8249689473823304, "grad_norm": 1.1273983716964722, "learning_rate": 1.5819719265892396e-06, "loss": 0.8618, "step": 23412 }, { "epoch": 0.8250041843952888, "grad_norm": 0.9559870958328247, "learning_rate": 1.5813559371167764e-06, "loss": 1.0228, "step": 23413 }, { "epoch": 0.8250394214082473, "grad_norm": 0.8640231490135193, "learning_rate": 1.5807400572982235e-06, "loss": 0.6934, "step": 23414 }, { "epoch": 0.8250746584212056, "grad_norm": 0.9763893485069275, "learning_rate": 1.5801242871416045e-06, "loss": 1.0793, "step": 23415 }, { "epoch": 0.825109895434164, "grad_norm": 7.243161678314209, "learning_rate": 1.5795086266549397e-06, "loss": 4.9381, "step": 23416 }, { "epoch": 0.8251451324471225, "grad_norm": 3.8332526683807373, "learning_rate": 1.5788930758462473e-06, "loss": 3.3638, "step": 23417 }, { "epoch": 0.8251803694600809, "grad_norm": 3.2474122047424316, "learning_rate": 1.578277634723544e-06, "loss": 2.7681, "step": 23418 }, { "epoch": 0.8252156064730393, "grad_norm": 1.3180241584777832, "learning_rate": 1.577662303294849e-06, "loss": 0.9172, "step": 23419 }, { "epoch": 0.8252508434859976, "grad_norm": 3.725083589553833, "learning_rate": 1.577047081568177e-06, "loss": 2.9474, "step": 23420 }, { "epoch": 0.8252860804989561, "grad_norm": 0.9414101243019104, "learning_rate": 1.576431969551535e-06, "loss": 0.9696, "step": 23421 }, { "epoch": 0.8253213175119145, "grad_norm": 3.056448221206665, "learning_rate": 1.5758169672529423e-06, "loss": 2.7427, "step": 23422 }, { "epoch": 0.8253565545248729, "grad_norm": 1.132368803024292, "learning_rate": 1.5752020746804087e-06, "loss": 0.7911, "step": 23423 }, { "epoch": 0.8253917915378314, "grad_norm": 1.147417426109314, "learning_rate": 1.5745872918419381e-06, "loss": 0.8488, "step": 23424 }, { "epoch": 0.8254270285507898, "grad_norm": 13.872086524963379, "learning_rate": 1.5739726187455396e-06, "loss": 3.172, "step": 23425 }, { "epoch": 0.8254622655637481, "grad_norm": 1.0943934917449951, "learning_rate": 1.5733580553992234e-06, "loss": 0.8148, "step": 23426 }, { "epoch": 0.8254975025767066, "grad_norm": 1.0402512550354004, "learning_rate": 1.5727436018109943e-06, "loss": 0.7828, "step": 23427 }, { "epoch": 0.825532739589665, "grad_norm": 3.5356082916259766, "learning_rate": 1.5721292579888513e-06, "loss": 3.0911, "step": 23428 }, { "epoch": 0.8255679766026234, "grad_norm": 1.1066570281982422, "learning_rate": 1.5715150239407951e-06, "loss": 0.9555, "step": 23429 }, { "epoch": 0.8256032136155818, "grad_norm": 0.9217120409011841, "learning_rate": 1.5709008996748344e-06, "loss": 0.8564, "step": 23430 }, { "epoch": 0.8256384506285402, "grad_norm": 2.835388422012329, "learning_rate": 1.5702868851989617e-06, "loss": 3.0632, "step": 23431 }, { "epoch": 0.8256736876414986, "grad_norm": 5.61583137512207, "learning_rate": 1.5696729805211751e-06, "loss": 4.5287, "step": 23432 }, { "epoch": 0.825708924654457, "grad_norm": 7.436511039733887, "learning_rate": 1.5690591856494718e-06, "loss": 5.7734, "step": 23433 }, { "epoch": 0.8257441616674155, "grad_norm": 6.122801780700684, "learning_rate": 1.5684455005918465e-06, "loss": 5.0661, "step": 23434 }, { "epoch": 0.8257793986803739, "grad_norm": 0.8816125988960266, "learning_rate": 1.5678319253562923e-06, "loss": 0.6621, "step": 23435 }, { "epoch": 0.8258146356933322, "grad_norm": 3.1099979877471924, "learning_rate": 1.5672184599508023e-06, "loss": 2.5217, "step": 23436 }, { "epoch": 0.8258498727062907, "grad_norm": 3.7130212783813477, "learning_rate": 1.5666051043833653e-06, "loss": 4.4524, "step": 23437 }, { "epoch": 0.8258851097192491, "grad_norm": 4.05467414855957, "learning_rate": 1.5659918586619715e-06, "loss": 4.1062, "step": 23438 }, { "epoch": 0.8259203467322075, "grad_norm": 4.1388068199157715, "learning_rate": 1.5653787227946105e-06, "loss": 2.4827, "step": 23439 }, { "epoch": 0.825955583745166, "grad_norm": 2.9085848331451416, "learning_rate": 1.564765696789261e-06, "loss": 3.1708, "step": 23440 }, { "epoch": 0.8259908207581244, "grad_norm": 2.961808681488037, "learning_rate": 1.5641527806539147e-06, "loss": 1.9822, "step": 23441 }, { "epoch": 0.8260260577710827, "grad_norm": 1.8388371467590332, "learning_rate": 1.5635399743965562e-06, "loss": 0.8716, "step": 23442 }, { "epoch": 0.8260612947840411, "grad_norm": 1.3070709705352783, "learning_rate": 1.5629272780251615e-06, "loss": 0.7155, "step": 23443 }, { "epoch": 0.8260965317969996, "grad_norm": 1.029160976409912, "learning_rate": 1.5623146915477104e-06, "loss": 1.0802, "step": 23444 }, { "epoch": 0.826131768809958, "grad_norm": 3.3435513973236084, "learning_rate": 1.5617022149721895e-06, "loss": 2.7196, "step": 23445 }, { "epoch": 0.8261670058229164, "grad_norm": 0.9655979871749878, "learning_rate": 1.5610898483065728e-06, "loss": 0.856, "step": 23446 }, { "epoch": 0.8262022428358748, "grad_norm": 1.4485126733779907, "learning_rate": 1.5604775915588343e-06, "loss": 0.9767, "step": 23447 }, { "epoch": 0.8262374798488332, "grad_norm": 2.1137051582336426, "learning_rate": 1.5598654447369478e-06, "loss": 1.0667, "step": 23448 }, { "epoch": 0.8262727168617916, "grad_norm": 6.801572322845459, "learning_rate": 1.559253407848893e-06, "loss": 4.8897, "step": 23449 }, { "epoch": 0.8263079538747501, "grad_norm": 3.324125289916992, "learning_rate": 1.5586414809026363e-06, "loss": 3.2184, "step": 23450 }, { "epoch": 0.8263431908877085, "grad_norm": 5.423824787139893, "learning_rate": 1.5580296639061487e-06, "loss": 3.5099, "step": 23451 }, { "epoch": 0.8263784279006668, "grad_norm": 3.5882351398468018, "learning_rate": 1.5574179568674007e-06, "loss": 3.3606, "step": 23452 }, { "epoch": 0.8264136649136252, "grad_norm": 1.1844730377197266, "learning_rate": 1.5568063597943582e-06, "loss": 0.8931, "step": 23453 }, { "epoch": 0.8264489019265837, "grad_norm": 5.639872074127197, "learning_rate": 1.5561948726949893e-06, "loss": 5.4061, "step": 23454 }, { "epoch": 0.8264841389395421, "grad_norm": 0.848828911781311, "learning_rate": 1.5555834955772586e-06, "loss": 0.8189, "step": 23455 }, { "epoch": 0.8265193759525005, "grad_norm": 3.7542216777801514, "learning_rate": 1.5549722284491276e-06, "loss": 2.9883, "step": 23456 }, { "epoch": 0.826554612965459, "grad_norm": 3.067451000213623, "learning_rate": 1.5543610713185585e-06, "loss": 2.9099, "step": 23457 }, { "epoch": 0.8265898499784173, "grad_norm": 3.697507381439209, "learning_rate": 1.5537500241935155e-06, "loss": 4.7184, "step": 23458 }, { "epoch": 0.8266250869913757, "grad_norm": 1.0763829946517944, "learning_rate": 1.5531390870819486e-06, "loss": 0.8695, "step": 23459 }, { "epoch": 0.8266603240043342, "grad_norm": 3.4614319801330566, "learning_rate": 1.5525282599918246e-06, "loss": 2.7303, "step": 23460 }, { "epoch": 0.8266955610172926, "grad_norm": 4.301884651184082, "learning_rate": 1.5519175429310974e-06, "loss": 3.2863, "step": 23461 }, { "epoch": 0.826730798030251, "grad_norm": 1.5903773307800293, "learning_rate": 1.5513069359077181e-06, "loss": 0.8893, "step": 23462 }, { "epoch": 0.8267660350432093, "grad_norm": 4.192723751068115, "learning_rate": 1.550696438929641e-06, "loss": 3.4105, "step": 23463 }, { "epoch": 0.8268012720561678, "grad_norm": 3.6344101428985596, "learning_rate": 1.55008605200482e-06, "loss": 2.9052, "step": 23464 }, { "epoch": 0.8268365090691262, "grad_norm": 3.453169107437134, "learning_rate": 1.549475775141207e-06, "loss": 2.5336, "step": 23465 }, { "epoch": 0.8268717460820846, "grad_norm": 4.263736724853516, "learning_rate": 1.5488656083467468e-06, "loss": 2.8723, "step": 23466 }, { "epoch": 0.8269069830950431, "grad_norm": 3.981041431427002, "learning_rate": 1.5482555516293862e-06, "loss": 2.8053, "step": 23467 }, { "epoch": 0.8269422201080014, "grad_norm": 4.059556007385254, "learning_rate": 1.547645604997078e-06, "loss": 4.9277, "step": 23468 }, { "epoch": 0.8269774571209598, "grad_norm": 1.1280605792999268, "learning_rate": 1.5470357684577598e-06, "loss": 0.8066, "step": 23469 }, { "epoch": 0.8270126941339183, "grad_norm": 1.1276742219924927, "learning_rate": 1.5464260420193777e-06, "loss": 0.7089, "step": 23470 }, { "epoch": 0.8270479311468767, "grad_norm": 1.7177497148513794, "learning_rate": 1.5458164256898732e-06, "loss": 1.0059, "step": 23471 }, { "epoch": 0.8270831681598351, "grad_norm": 9.5181245803833, "learning_rate": 1.5452069194771869e-06, "loss": 4.9559, "step": 23472 }, { "epoch": 0.8271184051727934, "grad_norm": 3.887328624725342, "learning_rate": 1.5445975233892574e-06, "loss": 3.3803, "step": 23473 }, { "epoch": 0.8271536421857519, "grad_norm": 1.1587334871292114, "learning_rate": 1.543988237434022e-06, "loss": 1.1169, "step": 23474 }, { "epoch": 0.8271888791987103, "grad_norm": 3.1262288093566895, "learning_rate": 1.5433790616194166e-06, "loss": 2.1648, "step": 23475 }, { "epoch": 0.8272241162116687, "grad_norm": 1.3375647068023682, "learning_rate": 1.542769995953376e-06, "loss": 0.9282, "step": 23476 }, { "epoch": 0.8272593532246272, "grad_norm": 1.5535144805908203, "learning_rate": 1.542161040443837e-06, "loss": 0.8835, "step": 23477 }, { "epoch": 0.8272945902375856, "grad_norm": 1.6009602546691895, "learning_rate": 1.5415521950987222e-06, "loss": 1.0255, "step": 23478 }, { "epoch": 0.8273298272505439, "grad_norm": 4.461945056915283, "learning_rate": 1.5409434599259698e-06, "loss": 2.6555, "step": 23479 }, { "epoch": 0.8273650642635024, "grad_norm": 5.054724216461182, "learning_rate": 1.540334834933509e-06, "loss": 2.756, "step": 23480 }, { "epoch": 0.8274003012764608, "grad_norm": 3.549699068069458, "learning_rate": 1.5397263201292623e-06, "loss": 3.28, "step": 23481 }, { "epoch": 0.8274355382894192, "grad_norm": 1.090999960899353, "learning_rate": 1.5391179155211555e-06, "loss": 1.056, "step": 23482 }, { "epoch": 0.8274707753023777, "grad_norm": 5.272793769836426, "learning_rate": 1.538509621117118e-06, "loss": 4.7515, "step": 23483 }, { "epoch": 0.827506012315336, "grad_norm": 2.1847214698791504, "learning_rate": 1.5379014369250733e-06, "loss": 2.5727, "step": 23484 }, { "epoch": 0.8275412493282944, "grad_norm": 3.0762102603912354, "learning_rate": 1.5372933629529374e-06, "loss": 3.0896, "step": 23485 }, { "epoch": 0.8275764863412528, "grad_norm": 4.987418174743652, "learning_rate": 1.5366853992086305e-06, "loss": 2.9096, "step": 23486 }, { "epoch": 0.8276117233542113, "grad_norm": 3.4229354858398438, "learning_rate": 1.5360775457000799e-06, "loss": 2.3228, "step": 23487 }, { "epoch": 0.8276469603671697, "grad_norm": 5.256289958953857, "learning_rate": 1.5354698024351944e-06, "loss": 3.1965, "step": 23488 }, { "epoch": 0.827682197380128, "grad_norm": 1.170220136642456, "learning_rate": 1.5348621694218936e-06, "loss": 0.8451, "step": 23489 }, { "epoch": 0.8277174343930865, "grad_norm": 5.311406135559082, "learning_rate": 1.5342546466680908e-06, "loss": 5.9587, "step": 23490 }, { "epoch": 0.8277526714060449, "grad_norm": 3.6849262714385986, "learning_rate": 1.5336472341816998e-06, "loss": 2.4961, "step": 23491 }, { "epoch": 0.8277879084190033, "grad_norm": 10.180534362792969, "learning_rate": 1.5330399319706323e-06, "loss": 4.7411, "step": 23492 }, { "epoch": 0.8278231454319618, "grad_norm": 4.555865287780762, "learning_rate": 1.5324327400427986e-06, "loss": 3.0959, "step": 23493 }, { "epoch": 0.8278583824449202, "grad_norm": 1.4347882270812988, "learning_rate": 1.5318256584061063e-06, "loss": 0.7823, "step": 23494 }, { "epoch": 0.8278936194578785, "grad_norm": 1.2988628149032593, "learning_rate": 1.531218687068463e-06, "loss": 0.7077, "step": 23495 }, { "epoch": 0.8279288564708369, "grad_norm": 3.476167917251587, "learning_rate": 1.5306118260377778e-06, "loss": 2.7158, "step": 23496 }, { "epoch": 0.8279640934837954, "grad_norm": 0.9114460945129395, "learning_rate": 1.5300050753219486e-06, "loss": 0.8042, "step": 23497 }, { "epoch": 0.8279993304967538, "grad_norm": 4.915186405181885, "learning_rate": 1.5293984349288838e-06, "loss": 5.2032, "step": 23498 }, { "epoch": 0.8280345675097122, "grad_norm": 2.870323896408081, "learning_rate": 1.5287919048664835e-06, "loss": 3.0099, "step": 23499 }, { "epoch": 0.8280698045226706, "grad_norm": 0.8945288062095642, "learning_rate": 1.5281854851426493e-06, "loss": 0.8094, "step": 23500 }, { "epoch": 0.828105041535629, "grad_norm": 3.114542245864868, "learning_rate": 1.5275791757652747e-06, "loss": 2.9396, "step": 23501 }, { "epoch": 0.8281402785485874, "grad_norm": 2.822666645050049, "learning_rate": 1.5269729767422626e-06, "loss": 2.6851, "step": 23502 }, { "epoch": 0.8281755155615459, "grad_norm": 2.791959762573242, "learning_rate": 1.5263668880815086e-06, "loss": 3.2666, "step": 23503 }, { "epoch": 0.8282107525745043, "grad_norm": 6.441599369049072, "learning_rate": 1.5257609097909033e-06, "loss": 8.3795, "step": 23504 }, { "epoch": 0.8282459895874626, "grad_norm": 5.103252410888672, "learning_rate": 1.525155041878339e-06, "loss": 3.3091, "step": 23505 }, { "epoch": 0.828281226600421, "grad_norm": 1.0694180727005005, "learning_rate": 1.5245492843517152e-06, "loss": 0.8228, "step": 23506 }, { "epoch": 0.8283164636133795, "grad_norm": 2.7323789596557617, "learning_rate": 1.523943637218913e-06, "loss": 2.237, "step": 23507 }, { "epoch": 0.8283517006263379, "grad_norm": 1.0746474266052246, "learning_rate": 1.5233381004878256e-06, "loss": 0.9319, "step": 23508 }, { "epoch": 0.8283869376392963, "grad_norm": 1.4730989933013916, "learning_rate": 1.5227326741663395e-06, "loss": 0.8902, "step": 23509 }, { "epoch": 0.8284221746522548, "grad_norm": 3.882345676422119, "learning_rate": 1.5221273582623398e-06, "loss": 2.6903, "step": 23510 }, { "epoch": 0.8284574116652131, "grad_norm": 2.510653257369995, "learning_rate": 1.5215221527837099e-06, "loss": 2.2986, "step": 23511 }, { "epoch": 0.8284926486781715, "grad_norm": 1.2481038570404053, "learning_rate": 1.5209170577383358e-06, "loss": 0.8801, "step": 23512 }, { "epoch": 0.82852788569113, "grad_norm": 2.723268508911133, "learning_rate": 1.5203120731340958e-06, "loss": 2.815, "step": 23513 }, { "epoch": 0.8285631227040884, "grad_norm": 1.3686577081680298, "learning_rate": 1.5197071989788714e-06, "loss": 0.7386, "step": 23514 }, { "epoch": 0.8285983597170468, "grad_norm": 6.694034576416016, "learning_rate": 1.519102435280544e-06, "loss": 5.5848, "step": 23515 }, { "epoch": 0.8286335967300053, "grad_norm": 2.5668158531188965, "learning_rate": 1.5184977820469825e-06, "loss": 2.545, "step": 23516 }, { "epoch": 0.8286688337429636, "grad_norm": 7.084469318389893, "learning_rate": 1.5178932392860702e-06, "loss": 4.8608, "step": 23517 }, { "epoch": 0.828704070755922, "grad_norm": 2.487064838409424, "learning_rate": 1.517288807005679e-06, "loss": 2.8056, "step": 23518 }, { "epoch": 0.8287393077688804, "grad_norm": 1.1608531475067139, "learning_rate": 1.5166844852136841e-06, "loss": 0.7292, "step": 23519 }, { "epoch": 0.8287745447818389, "grad_norm": 3.831277370452881, "learning_rate": 1.5160802739179492e-06, "loss": 2.9066, "step": 23520 }, { "epoch": 0.8288097817947973, "grad_norm": 1.6903032064437866, "learning_rate": 1.5154761731263523e-06, "loss": 0.7153, "step": 23521 }, { "epoch": 0.8288450188077556, "grad_norm": 1.1957063674926758, "learning_rate": 1.5148721828467606e-06, "loss": 0.7417, "step": 23522 }, { "epoch": 0.8288802558207141, "grad_norm": 1.7653924226760864, "learning_rate": 1.5142683030870375e-06, "loss": 0.7583, "step": 23523 }, { "epoch": 0.8289154928336725, "grad_norm": 2.007246732711792, "learning_rate": 1.5136645338550482e-06, "loss": 0.9229, "step": 23524 }, { "epoch": 0.8289507298466309, "grad_norm": 1.2795425653457642, "learning_rate": 1.513060875158665e-06, "loss": 1.1389, "step": 23525 }, { "epoch": 0.8289859668595894, "grad_norm": 1.166965365409851, "learning_rate": 1.5124573270057418e-06, "loss": 0.7067, "step": 23526 }, { "epoch": 0.8290212038725477, "grad_norm": 3.1305394172668457, "learning_rate": 1.5118538894041423e-06, "loss": 2.304, "step": 23527 }, { "epoch": 0.8290564408855061, "grad_norm": 3.282339096069336, "learning_rate": 1.511250562361728e-06, "loss": 2.606, "step": 23528 }, { "epoch": 0.8290916778984645, "grad_norm": 1.124745488166809, "learning_rate": 1.5106473458863568e-06, "loss": 0.6934, "step": 23529 }, { "epoch": 0.829126914911423, "grad_norm": 1.1434773206710815, "learning_rate": 1.510044239985885e-06, "loss": 0.8093, "step": 23530 }, { "epoch": 0.8291621519243814, "grad_norm": 4.067975997924805, "learning_rate": 1.5094412446681684e-06, "loss": 3.129, "step": 23531 }, { "epoch": 0.8291973889373397, "grad_norm": 7.284512996673584, "learning_rate": 1.5088383599410606e-06, "loss": 3.2686, "step": 23532 }, { "epoch": 0.8292326259502982, "grad_norm": 1.0379129648208618, "learning_rate": 1.5082355858124154e-06, "loss": 0.8574, "step": 23533 }, { "epoch": 0.8292678629632566, "grad_norm": 3.393893003463745, "learning_rate": 1.5076329222900855e-06, "loss": 2.7065, "step": 23534 }, { "epoch": 0.829303099976215, "grad_norm": 6.763015270233154, "learning_rate": 1.5070303693819133e-06, "loss": 2.8231, "step": 23535 }, { "epoch": 0.8293383369891735, "grad_norm": 1.1206467151641846, "learning_rate": 1.506427927095756e-06, "loss": 0.9347, "step": 23536 }, { "epoch": 0.8293735740021319, "grad_norm": 4.5871100425720215, "learning_rate": 1.5058255954394562e-06, "loss": 4.9402, "step": 23537 }, { "epoch": 0.8294088110150902, "grad_norm": 1.068920612335205, "learning_rate": 1.5052233744208622e-06, "loss": 0.9797, "step": 23538 }, { "epoch": 0.8294440480280486, "grad_norm": 2.8913111686706543, "learning_rate": 1.5046212640478108e-06, "loss": 2.956, "step": 23539 }, { "epoch": 0.8294792850410071, "grad_norm": 4.871947765350342, "learning_rate": 1.5040192643281525e-06, "loss": 3.2456, "step": 23540 }, { "epoch": 0.8295145220539655, "grad_norm": 8.038538932800293, "learning_rate": 1.5034173752697278e-06, "loss": 2.839, "step": 23541 }, { "epoch": 0.8295497590669239, "grad_norm": 1.1469756364822388, "learning_rate": 1.5028155968803714e-06, "loss": 0.757, "step": 23542 }, { "epoch": 0.8295849960798823, "grad_norm": 5.34940242767334, "learning_rate": 1.5022139291679215e-06, "loss": 3.6274, "step": 23543 }, { "epoch": 0.8296202330928407, "grad_norm": 2.787440061569214, "learning_rate": 1.501612372140223e-06, "loss": 2.5859, "step": 23544 }, { "epoch": 0.8296554701057991, "grad_norm": 4.100719928741455, "learning_rate": 1.5010109258051032e-06, "loss": 2.9364, "step": 23545 }, { "epoch": 0.8296907071187576, "grad_norm": 0.9300795197486877, "learning_rate": 1.5004095901703997e-06, "loss": 0.7994, "step": 23546 }, { "epoch": 0.829725944131716, "grad_norm": 2.3253793716430664, "learning_rate": 1.499808365243942e-06, "loss": 0.7409, "step": 23547 }, { "epoch": 0.8297611811446743, "grad_norm": 5.391641139984131, "learning_rate": 1.4992072510335653e-06, "loss": 4.6498, "step": 23548 }, { "epoch": 0.8297964181576328, "grad_norm": 4.608232021331787, "learning_rate": 1.4986062475470952e-06, "loss": 5.2902, "step": 23549 }, { "epoch": 0.8298316551705912, "grad_norm": 5.181941032409668, "learning_rate": 1.4980053547923623e-06, "loss": 2.7017, "step": 23550 }, { "epoch": 0.8298668921835496, "grad_norm": 1.2946840524673462, "learning_rate": 1.4974045727771935e-06, "loss": 0.6309, "step": 23551 }, { "epoch": 0.829902129196508, "grad_norm": 0.9557110071182251, "learning_rate": 1.4968039015094116e-06, "loss": 1.1152, "step": 23552 }, { "epoch": 0.8299373662094665, "grad_norm": 1.16377854347229, "learning_rate": 1.4962033409968434e-06, "loss": 0.676, "step": 23553 }, { "epoch": 0.8299726032224248, "grad_norm": 0.994657576084137, "learning_rate": 1.4956028912473086e-06, "loss": 1.0095, "step": 23554 }, { "epoch": 0.8300078402353832, "grad_norm": 4.679629325866699, "learning_rate": 1.4950025522686306e-06, "loss": 2.9112, "step": 23555 }, { "epoch": 0.8300430772483417, "grad_norm": 3.7418229579925537, "learning_rate": 1.4944023240686268e-06, "loss": 2.6579, "step": 23556 }, { "epoch": 0.8300783142613001, "grad_norm": 1.1784589290618896, "learning_rate": 1.4938022066551194e-06, "loss": 0.6765, "step": 23557 }, { "epoch": 0.8301135512742585, "grad_norm": 1.0314449071884155, "learning_rate": 1.493202200035917e-06, "loss": 0.7923, "step": 23558 }, { "epoch": 0.8301487882872169, "grad_norm": 2.481750011444092, "learning_rate": 1.492602304218841e-06, "loss": 0.9051, "step": 23559 }, { "epoch": 0.8301840253001753, "grad_norm": 4.294014930725098, "learning_rate": 1.492002519211706e-06, "loss": 4.4576, "step": 23560 }, { "epoch": 0.8302192623131337, "grad_norm": 2.643195629119873, "learning_rate": 1.4914028450223207e-06, "loss": 2.7046, "step": 23561 }, { "epoch": 0.8302544993260921, "grad_norm": 4.5566301345825195, "learning_rate": 1.490803281658494e-06, "loss": 3.0275, "step": 23562 }, { "epoch": 0.8302897363390506, "grad_norm": 2.4588794708251953, "learning_rate": 1.490203829128044e-06, "loss": 2.206, "step": 23563 }, { "epoch": 0.8303249733520089, "grad_norm": 2.4267842769622803, "learning_rate": 1.4896044874387694e-06, "loss": 0.8373, "step": 23564 }, { "epoch": 0.8303602103649673, "grad_norm": 3.026388168334961, "learning_rate": 1.4890052565984814e-06, "loss": 2.2765, "step": 23565 }, { "epoch": 0.8303954473779258, "grad_norm": 4.170498847961426, "learning_rate": 1.4884061366149839e-06, "loss": 4.673, "step": 23566 }, { "epoch": 0.8304306843908842, "grad_norm": 0.9259750843048096, "learning_rate": 1.4878071274960804e-06, "loss": 0.7975, "step": 23567 }, { "epoch": 0.8304659214038426, "grad_norm": 5.409329414367676, "learning_rate": 1.4872082292495738e-06, "loss": 4.8987, "step": 23568 }, { "epoch": 0.830501158416801, "grad_norm": 1.0777028799057007, "learning_rate": 1.4866094418832645e-06, "loss": 0.7287, "step": 23569 }, { "epoch": 0.8305363954297594, "grad_norm": 5.439825534820557, "learning_rate": 1.4860107654049504e-06, "loss": 5.0375, "step": 23570 }, { "epoch": 0.8305716324427178, "grad_norm": 1.5187406539916992, "learning_rate": 1.4854121998224314e-06, "loss": 1.1379, "step": 23571 }, { "epoch": 0.8306068694556762, "grad_norm": 3.284273862838745, "learning_rate": 1.484813745143503e-06, "loss": 2.7614, "step": 23572 }, { "epoch": 0.8306421064686347, "grad_norm": 3.3192648887634277, "learning_rate": 1.4842154013759613e-06, "loss": 2.5641, "step": 23573 }, { "epoch": 0.8306773434815931, "grad_norm": 4.879654884338379, "learning_rate": 1.4836171685275968e-06, "loss": 5.4152, "step": 23574 }, { "epoch": 0.8307125804945514, "grad_norm": 3.464125394821167, "learning_rate": 1.4830190466062044e-06, "loss": 2.6888, "step": 23575 }, { "epoch": 0.8307478175075099, "grad_norm": 1.9630861282348633, "learning_rate": 1.4824210356195768e-06, "loss": 0.8373, "step": 23576 }, { "epoch": 0.8307830545204683, "grad_norm": 4.601267337799072, "learning_rate": 1.4818231355754953e-06, "loss": 5.1715, "step": 23577 }, { "epoch": 0.8308182915334267, "grad_norm": 1.141602635383606, "learning_rate": 1.481225346481754e-06, "loss": 0.8897, "step": 23578 }, { "epoch": 0.8308535285463852, "grad_norm": 10.960697174072266, "learning_rate": 1.4806276683461407e-06, "loss": 7.5719, "step": 23579 }, { "epoch": 0.8308887655593435, "grad_norm": 3.6357839107513428, "learning_rate": 1.480030101176434e-06, "loss": 2.8639, "step": 23580 }, { "epoch": 0.8309240025723019, "grad_norm": 6.059629917144775, "learning_rate": 1.4794326449804197e-06, "loss": 4.3186, "step": 23581 }, { "epoch": 0.8309592395852604, "grad_norm": 5.512031555175781, "learning_rate": 1.478835299765884e-06, "loss": 2.732, "step": 23582 }, { "epoch": 0.8309944765982188, "grad_norm": 3.112776756286621, "learning_rate": 1.4782380655406014e-06, "loss": 2.4714, "step": 23583 }, { "epoch": 0.8310297136111772, "grad_norm": 1.5986074209213257, "learning_rate": 1.477640942312355e-06, "loss": 0.9314, "step": 23584 }, { "epoch": 0.8310649506241355, "grad_norm": 5.232603549957275, "learning_rate": 1.4770439300889195e-06, "loss": 4.105, "step": 23585 }, { "epoch": 0.831100187637094, "grad_norm": 0.8486392498016357, "learning_rate": 1.476447028878073e-06, "loss": 0.8524, "step": 23586 }, { "epoch": 0.8311354246500524, "grad_norm": 12.156522750854492, "learning_rate": 1.4758502386875894e-06, "loss": 5.2759, "step": 23587 }, { "epoch": 0.8311706616630108, "grad_norm": 1.9105557203292847, "learning_rate": 1.4752535595252427e-06, "loss": 0.6912, "step": 23588 }, { "epoch": 0.8312058986759693, "grad_norm": 3.106973648071289, "learning_rate": 1.474656991398804e-06, "loss": 1.9623, "step": 23589 }, { "epoch": 0.8312411356889277, "grad_norm": 3.100257396697998, "learning_rate": 1.474060534316044e-06, "loss": 2.1095, "step": 23590 }, { "epoch": 0.831276372701886, "grad_norm": 5.604452610015869, "learning_rate": 1.4734641882847312e-06, "loss": 5.3536, "step": 23591 }, { "epoch": 0.8313116097148445, "grad_norm": 5.547293663024902, "learning_rate": 1.4728679533126344e-06, "loss": 3.1007, "step": 23592 }, { "epoch": 0.8313468467278029, "grad_norm": 1.4859408140182495, "learning_rate": 1.4722718294075178e-06, "loss": 0.8916, "step": 23593 }, { "epoch": 0.8313820837407613, "grad_norm": 2.0765788555145264, "learning_rate": 1.4716758165771473e-06, "loss": 0.8074, "step": 23594 }, { "epoch": 0.8314173207537197, "grad_norm": 0.9719076156616211, "learning_rate": 1.4710799148292876e-06, "loss": 0.7707, "step": 23595 }, { "epoch": 0.8314525577666781, "grad_norm": 4.686905860900879, "learning_rate": 1.4704841241716937e-06, "loss": 4.9337, "step": 23596 }, { "epoch": 0.8314877947796365, "grad_norm": 2.2577321529388428, "learning_rate": 1.469888444612133e-06, "loss": 2.486, "step": 23597 }, { "epoch": 0.8315230317925949, "grad_norm": 1.7659568786621094, "learning_rate": 1.4692928761583647e-06, "loss": 0.6193, "step": 23598 }, { "epoch": 0.8315582688055534, "grad_norm": 1.525956630706787, "learning_rate": 1.4686974188181402e-06, "loss": 1.0318, "step": 23599 }, { "epoch": 0.8315935058185118, "grad_norm": 2.6802990436553955, "learning_rate": 1.4681020725992168e-06, "loss": 0.8763, "step": 23600 }, { "epoch": 0.8316287428314701, "grad_norm": 4.714765548706055, "learning_rate": 1.467506837509356e-06, "loss": 4.3668, "step": 23601 }, { "epoch": 0.8316639798444286, "grad_norm": 4.231626033782959, "learning_rate": 1.4669117135563015e-06, "loss": 2.4807, "step": 23602 }, { "epoch": 0.831699216857387, "grad_norm": 5.09288215637207, "learning_rate": 1.4663167007478097e-06, "loss": 2.9442, "step": 23603 }, { "epoch": 0.8317344538703454, "grad_norm": 1.209049105644226, "learning_rate": 1.4657217990916305e-06, "loss": 0.7834, "step": 23604 }, { "epoch": 0.8317696908833038, "grad_norm": 11.315235137939453, "learning_rate": 1.4651270085955117e-06, "loss": 7.2218, "step": 23605 }, { "epoch": 0.8318049278962623, "grad_norm": 3.1585566997528076, "learning_rate": 1.4645323292671997e-06, "loss": 2.5146, "step": 23606 }, { "epoch": 0.8318401649092206, "grad_norm": 15.047350883483887, "learning_rate": 1.4639377611144424e-06, "loss": 6.6136, "step": 23607 }, { "epoch": 0.831875401922179, "grad_norm": 1.265371322631836, "learning_rate": 1.463343304144983e-06, "loss": 0.8615, "step": 23608 }, { "epoch": 0.8319106389351375, "grad_norm": 4.82305383682251, "learning_rate": 1.4627489583665644e-06, "loss": 3.0578, "step": 23609 }, { "epoch": 0.8319458759480959, "grad_norm": 3.9920010566711426, "learning_rate": 1.4621547237869283e-06, "loss": 2.5516, "step": 23610 }, { "epoch": 0.8319811129610543, "grad_norm": 3.8256826400756836, "learning_rate": 1.4615606004138138e-06, "loss": 4.3915, "step": 23611 }, { "epoch": 0.8320163499740127, "grad_norm": 1.7653383016586304, "learning_rate": 1.4609665882549607e-06, "loss": 0.928, "step": 23612 }, { "epoch": 0.8320515869869711, "grad_norm": 1.1702520847320557, "learning_rate": 1.4603726873181047e-06, "loss": 1.1608, "step": 23613 }, { "epoch": 0.8320868239999295, "grad_norm": 3.419414758682251, "learning_rate": 1.459778897610985e-06, "loss": 2.8794, "step": 23614 }, { "epoch": 0.832122061012888, "grad_norm": 4.287243843078613, "learning_rate": 1.4591852191413315e-06, "loss": 3.2057, "step": 23615 }, { "epoch": 0.8321572980258464, "grad_norm": 2.8334083557128906, "learning_rate": 1.4585916519168753e-06, "loss": 2.8505, "step": 23616 }, { "epoch": 0.8321925350388047, "grad_norm": 1.2516846656799316, "learning_rate": 1.4579981959453548e-06, "loss": 0.8612, "step": 23617 }, { "epoch": 0.8322277720517631, "grad_norm": 3.1549391746520996, "learning_rate": 1.457404851234495e-06, "loss": 2.9976, "step": 23618 }, { "epoch": 0.8322630090647216, "grad_norm": 9.174129486083984, "learning_rate": 1.4568116177920244e-06, "loss": 7.0606, "step": 23619 }, { "epoch": 0.83229824607768, "grad_norm": 1.6153414249420166, "learning_rate": 1.4562184956256708e-06, "loss": 0.8186, "step": 23620 }, { "epoch": 0.8323334830906384, "grad_norm": 0.96136075258255, "learning_rate": 1.4556254847431594e-06, "loss": 0.8981, "step": 23621 }, { "epoch": 0.8323687201035969, "grad_norm": 1.1080631017684937, "learning_rate": 1.4550325851522152e-06, "loss": 1.2176, "step": 23622 }, { "epoch": 0.8324039571165552, "grad_norm": 1.0181519985198975, "learning_rate": 1.4544397968605594e-06, "loss": 0.6712, "step": 23623 }, { "epoch": 0.8324391941295136, "grad_norm": 1.5243103504180908, "learning_rate": 1.4538471198759141e-06, "loss": 0.9182, "step": 23624 }, { "epoch": 0.8324744311424721, "grad_norm": 1.1391026973724365, "learning_rate": 1.4532545542059984e-06, "loss": 0.7386, "step": 23625 }, { "epoch": 0.8325096681554305, "grad_norm": 1.109972596168518, "learning_rate": 1.4526620998585305e-06, "loss": 0.9139, "step": 23626 }, { "epoch": 0.8325449051683889, "grad_norm": 10.335923194885254, "learning_rate": 1.4520697568412278e-06, "loss": 2.4984, "step": 23627 }, { "epoch": 0.8325801421813472, "grad_norm": 5.204598426818848, "learning_rate": 1.4514775251618052e-06, "loss": 5.0172, "step": 23628 }, { "epoch": 0.8326153791943057, "grad_norm": 3.0386509895324707, "learning_rate": 1.4508854048279774e-06, "loss": 3.1221, "step": 23629 }, { "epoch": 0.8326506162072641, "grad_norm": 0.8905532360076904, "learning_rate": 1.4502933958474575e-06, "loss": 0.7012, "step": 23630 }, { "epoch": 0.8326858532202225, "grad_norm": 0.8413159847259521, "learning_rate": 1.4497014982279512e-06, "loss": 0.7402, "step": 23631 }, { "epoch": 0.832721090233181, "grad_norm": 4.163570404052734, "learning_rate": 1.4491097119771747e-06, "loss": 5.1581, "step": 23632 }, { "epoch": 0.8327563272461393, "grad_norm": 5.264125823974609, "learning_rate": 1.4485180371028352e-06, "loss": 2.7602, "step": 23633 }, { "epoch": 0.8327915642590977, "grad_norm": 2.7702372074127197, "learning_rate": 1.4479264736126354e-06, "loss": 2.7724, "step": 23634 }, { "epoch": 0.8328268012720562, "grad_norm": 5.1012864112854, "learning_rate": 1.44733502151428e-06, "loss": 3.2698, "step": 23635 }, { "epoch": 0.8328620382850146, "grad_norm": 1.18574857711792, "learning_rate": 1.44674368081548e-06, "loss": 0.7947, "step": 23636 }, { "epoch": 0.832897275297973, "grad_norm": 0.9926720857620239, "learning_rate": 1.4461524515239312e-06, "loss": 0.928, "step": 23637 }, { "epoch": 0.8329325123109314, "grad_norm": 1.3071268796920776, "learning_rate": 1.4455613336473362e-06, "loss": 0.7843, "step": 23638 }, { "epoch": 0.8329677493238898, "grad_norm": 6.291407585144043, "learning_rate": 1.4449703271933945e-06, "loss": 5.0894, "step": 23639 }, { "epoch": 0.8330029863368482, "grad_norm": 3.3571879863739014, "learning_rate": 1.4443794321698034e-06, "loss": 3.0932, "step": 23640 }, { "epoch": 0.8330382233498066, "grad_norm": 1.040582537651062, "learning_rate": 1.4437886485842613e-06, "loss": 0.7762, "step": 23641 }, { "epoch": 0.8330734603627651, "grad_norm": 2.9957244396209717, "learning_rate": 1.4431979764444603e-06, "loss": 2.8213, "step": 23642 }, { "epoch": 0.8331086973757235, "grad_norm": 5.421429634094238, "learning_rate": 1.442607415758097e-06, "loss": 4.8121, "step": 23643 }, { "epoch": 0.8331439343886818, "grad_norm": 2.3718249797821045, "learning_rate": 1.4420169665328621e-06, "loss": 1.0796, "step": 23644 }, { "epoch": 0.8331791714016403, "grad_norm": 2.017573595046997, "learning_rate": 1.4414266287764456e-06, "loss": 0.7854, "step": 23645 }, { "epoch": 0.8332144084145987, "grad_norm": 1.004072666168213, "learning_rate": 1.440836402496537e-06, "loss": 0.926, "step": 23646 }, { "epoch": 0.8332496454275571, "grad_norm": 2.715762138366699, "learning_rate": 1.440246287700826e-06, "loss": 2.7367, "step": 23647 }, { "epoch": 0.8332848824405156, "grad_norm": 2.3094899654388428, "learning_rate": 1.4396562843969963e-06, "loss": 2.5669, "step": 23648 }, { "epoch": 0.833320119453474, "grad_norm": 2.865340232849121, "learning_rate": 1.4390663925927363e-06, "loss": 2.3307, "step": 23649 }, { "epoch": 0.8333553564664323, "grad_norm": 4.392439365386963, "learning_rate": 1.4384766122957229e-06, "loss": 2.922, "step": 23650 }, { "epoch": 0.8333905934793907, "grad_norm": 0.9537295699119568, "learning_rate": 1.4378869435136444e-06, "loss": 0.9567, "step": 23651 }, { "epoch": 0.8334258304923492, "grad_norm": 26.67595100402832, "learning_rate": 1.4372973862541817e-06, "loss": 3.0953, "step": 23652 }, { "epoch": 0.8334610675053076, "grad_norm": 4.030035972595215, "learning_rate": 1.4367079405250083e-06, "loss": 2.8537, "step": 23653 }, { "epoch": 0.833496304518266, "grad_norm": 1.2046574354171753, "learning_rate": 1.4361186063338028e-06, "loss": 0.7506, "step": 23654 }, { "epoch": 0.8335315415312244, "grad_norm": 3.924243688583374, "learning_rate": 1.4355293836882478e-06, "loss": 3.6844, "step": 23655 }, { "epoch": 0.8335667785441828, "grad_norm": 4.043109893798828, "learning_rate": 1.4349402725960115e-06, "loss": 2.4172, "step": 23656 }, { "epoch": 0.8336020155571412, "grad_norm": 1.4896429777145386, "learning_rate": 1.4343512730647679e-06, "loss": 0.9678, "step": 23657 }, { "epoch": 0.8336372525700997, "grad_norm": 4.786540508270264, "learning_rate": 1.433762385102192e-06, "loss": 3.36, "step": 23658 }, { "epoch": 0.8336724895830581, "grad_norm": 1.2047258615493774, "learning_rate": 1.4331736087159509e-06, "loss": 0.7451, "step": 23659 }, { "epoch": 0.8337077265960164, "grad_norm": 3.5027756690979004, "learning_rate": 1.4325849439137152e-06, "loss": 3.1261, "step": 23660 }, { "epoch": 0.8337429636089748, "grad_norm": 2.336432456970215, "learning_rate": 1.431996390703152e-06, "loss": 2.1226, "step": 23661 }, { "epoch": 0.8337782006219333, "grad_norm": 7.431774139404297, "learning_rate": 1.4314079490919265e-06, "loss": 6.7497, "step": 23662 }, { "epoch": 0.8338134376348917, "grad_norm": 1.093961477279663, "learning_rate": 1.4308196190877044e-06, "loss": 0.8839, "step": 23663 }, { "epoch": 0.8338486746478501, "grad_norm": 4.246042728424072, "learning_rate": 1.4302314006981488e-06, "loss": 4.7203, "step": 23664 }, { "epoch": 0.8338839116608086, "grad_norm": 1.6938825845718384, "learning_rate": 1.4296432939309202e-06, "loss": 1.0325, "step": 23665 }, { "epoch": 0.8339191486737669, "grad_norm": 4.170046806335449, "learning_rate": 1.42905529879368e-06, "loss": 2.7039, "step": 23666 }, { "epoch": 0.8339543856867253, "grad_norm": 5.157416820526123, "learning_rate": 1.4284674152940858e-06, "loss": 3.3972, "step": 23667 }, { "epoch": 0.8339896226996838, "grad_norm": 1.2590640783309937, "learning_rate": 1.4278796434397978e-06, "loss": 0.8096, "step": 23668 }, { "epoch": 0.8340248597126422, "grad_norm": 4.187922954559326, "learning_rate": 1.4272919832384647e-06, "loss": 3.1228, "step": 23669 }, { "epoch": 0.8340600967256006, "grad_norm": 2.186722755432129, "learning_rate": 1.4267044346977477e-06, "loss": 2.5884, "step": 23670 }, { "epoch": 0.8340953337385589, "grad_norm": 3.964292049407959, "learning_rate": 1.4261169978252997e-06, "loss": 3.4984, "step": 23671 }, { "epoch": 0.8341305707515174, "grad_norm": 3.3009164333343506, "learning_rate": 1.4255296726287682e-06, "loss": 2.9519, "step": 23672 }, { "epoch": 0.8341658077644758, "grad_norm": 1.817144751548767, "learning_rate": 1.4249424591158024e-06, "loss": 0.7419, "step": 23673 }, { "epoch": 0.8342010447774342, "grad_norm": 1.5408151149749756, "learning_rate": 1.4243553572940583e-06, "loss": 0.7496, "step": 23674 }, { "epoch": 0.8342362817903927, "grad_norm": 4.812411308288574, "learning_rate": 1.4237683671711743e-06, "loss": 4.8685, "step": 23675 }, { "epoch": 0.834271518803351, "grad_norm": 8.35343074798584, "learning_rate": 1.4231814887548012e-06, "loss": 5.8374, "step": 23676 }, { "epoch": 0.8343067558163094, "grad_norm": 3.7350945472717285, "learning_rate": 1.4225947220525815e-06, "loss": 2.7867, "step": 23677 }, { "epoch": 0.8343419928292679, "grad_norm": 1.3181551694869995, "learning_rate": 1.4220080670721581e-06, "loss": 0.9168, "step": 23678 }, { "epoch": 0.8343772298422263, "grad_norm": 5.797211170196533, "learning_rate": 1.4214215238211714e-06, "loss": 4.9609, "step": 23679 }, { "epoch": 0.8344124668551847, "grad_norm": 1.702034831047058, "learning_rate": 1.4208350923072634e-06, "loss": 0.7952, "step": 23680 }, { "epoch": 0.834447703868143, "grad_norm": 1.4341152906417847, "learning_rate": 1.4202487725380697e-06, "loss": 1.1041, "step": 23681 }, { "epoch": 0.8344829408811015, "grad_norm": 5.797520637512207, "learning_rate": 1.4196625645212303e-06, "loss": 4.9185, "step": 23682 }, { "epoch": 0.8345181778940599, "grad_norm": 1.2720032930374146, "learning_rate": 1.4190764682643777e-06, "loss": 0.9629, "step": 23683 }, { "epoch": 0.8345534149070183, "grad_norm": 2.329432487487793, "learning_rate": 1.418490483775148e-06, "loss": 2.3106, "step": 23684 }, { "epoch": 0.8345886519199768, "grad_norm": 1.25088632106781, "learning_rate": 1.4179046110611716e-06, "loss": 0.8567, "step": 23685 }, { "epoch": 0.8346238889329352, "grad_norm": 5.40861701965332, "learning_rate": 1.4173188501300815e-06, "loss": 2.7346, "step": 23686 }, { "epoch": 0.8346591259458935, "grad_norm": 1.581567406654358, "learning_rate": 1.4167332009895097e-06, "loss": 0.6935, "step": 23687 }, { "epoch": 0.834694362958852, "grad_norm": 6.100990295410156, "learning_rate": 1.416147663647076e-06, "loss": 2.9133, "step": 23688 }, { "epoch": 0.8347295999718104, "grad_norm": 1.1027469635009766, "learning_rate": 1.415562238110415e-06, "loss": 1.0884, "step": 23689 }, { "epoch": 0.8347648369847688, "grad_norm": 1.0521626472473145, "learning_rate": 1.4149769243871514e-06, "loss": 0.8452, "step": 23690 }, { "epoch": 0.8348000739977273, "grad_norm": 4.3199334144592285, "learning_rate": 1.4143917224849046e-06, "loss": 4.5067, "step": 23691 }, { "epoch": 0.8348353110106856, "grad_norm": 3.8155672550201416, "learning_rate": 1.4138066324112964e-06, "loss": 3.12, "step": 23692 }, { "epoch": 0.834870548023644, "grad_norm": 1.3852540254592896, "learning_rate": 1.4132216541739563e-06, "loss": 0.8648, "step": 23693 }, { "epoch": 0.8349057850366024, "grad_norm": 1.2208688259124756, "learning_rate": 1.4126367877804947e-06, "loss": 0.8899, "step": 23694 }, { "epoch": 0.8349410220495609, "grad_norm": 8.234976768493652, "learning_rate": 1.4120520332385323e-06, "loss": 2.9057, "step": 23695 }, { "epoch": 0.8349762590625193, "grad_norm": 0.8392673134803772, "learning_rate": 1.4114673905556875e-06, "loss": 0.8688, "step": 23696 }, { "epoch": 0.8350114960754776, "grad_norm": 1.3056448698043823, "learning_rate": 1.410882859739573e-06, "loss": 1.2246, "step": 23697 }, { "epoch": 0.8350467330884361, "grad_norm": 4.634881019592285, "learning_rate": 1.4102984407978037e-06, "loss": 3.0668, "step": 23698 }, { "epoch": 0.8350819701013945, "grad_norm": 1.268688440322876, "learning_rate": 1.4097141337379916e-06, "loss": 0.9827, "step": 23699 }, { "epoch": 0.8351172071143529, "grad_norm": 5.027810573577881, "learning_rate": 1.409129938567747e-06, "loss": 4.6943, "step": 23700 }, { "epoch": 0.8351524441273114, "grad_norm": 1.8042383193969727, "learning_rate": 1.4085458552946785e-06, "loss": 0.719, "step": 23701 }, { "epoch": 0.8351876811402698, "grad_norm": 1.0852032899856567, "learning_rate": 1.4079618839263953e-06, "loss": 0.9952, "step": 23702 }, { "epoch": 0.8352229181532281, "grad_norm": 3.7601990699768066, "learning_rate": 1.4073780244705037e-06, "loss": 5.0642, "step": 23703 }, { "epoch": 0.8352581551661865, "grad_norm": 5.233160018920898, "learning_rate": 1.4067942769346066e-06, "loss": 5.1963, "step": 23704 }, { "epoch": 0.835293392179145, "grad_norm": 3.155916690826416, "learning_rate": 1.4062106413263088e-06, "loss": 2.7277, "step": 23705 }, { "epoch": 0.8353286291921034, "grad_norm": 3.669635534286499, "learning_rate": 1.4056271176532154e-06, "loss": 3.0686, "step": 23706 }, { "epoch": 0.8353638662050618, "grad_norm": 1.0131348371505737, "learning_rate": 1.4050437059229183e-06, "loss": 1.0857, "step": 23707 }, { "epoch": 0.8353991032180202, "grad_norm": 1.5465303659439087, "learning_rate": 1.4044604061430234e-06, "loss": 0.935, "step": 23708 }, { "epoch": 0.8354343402309786, "grad_norm": 3.215559244155884, "learning_rate": 1.4038772183211303e-06, "loss": 2.4348, "step": 23709 }, { "epoch": 0.835469577243937, "grad_norm": 1.11042058467865, "learning_rate": 1.4032941424648272e-06, "loss": 0.7693, "step": 23710 }, { "epoch": 0.8355048142568955, "grad_norm": 0.7764989137649536, "learning_rate": 1.4027111785817126e-06, "loss": 0.7372, "step": 23711 }, { "epoch": 0.8355400512698539, "grad_norm": 1.293410301208496, "learning_rate": 1.4021283266793839e-06, "loss": 0.6779, "step": 23712 }, { "epoch": 0.8355752882828122, "grad_norm": 3.972900152206421, "learning_rate": 1.401545586765426e-06, "loss": 3.2457, "step": 23713 }, { "epoch": 0.8356105252957706, "grad_norm": 1.0941932201385498, "learning_rate": 1.4009629588474326e-06, "loss": 0.8007, "step": 23714 }, { "epoch": 0.8356457623087291, "grad_norm": 4.465131759643555, "learning_rate": 1.4003804429329915e-06, "loss": 4.7411, "step": 23715 }, { "epoch": 0.8356809993216875, "grad_norm": 4.281974792480469, "learning_rate": 1.3997980390296918e-06, "loss": 3.836, "step": 23716 }, { "epoch": 0.8357162363346459, "grad_norm": 1.1114767789840698, "learning_rate": 1.3992157471451174e-06, "loss": 0.7783, "step": 23717 }, { "epoch": 0.8357514733476044, "grad_norm": 4.21045446395874, "learning_rate": 1.3986335672868534e-06, "loss": 4.5523, "step": 23718 }, { "epoch": 0.8357867103605627, "grad_norm": 1.1210283041000366, "learning_rate": 1.3980514994624827e-06, "loss": 0.6539, "step": 23719 }, { "epoch": 0.8358219473735211, "grad_norm": 3.5903313159942627, "learning_rate": 1.3974695436795871e-06, "loss": 3.2871, "step": 23720 }, { "epoch": 0.8358571843864796, "grad_norm": 1.2036080360412598, "learning_rate": 1.396887699945747e-06, "loss": 0.7679, "step": 23721 }, { "epoch": 0.835892421399438, "grad_norm": 3.2108733654022217, "learning_rate": 1.3963059682685399e-06, "loss": 3.024, "step": 23722 }, { "epoch": 0.8359276584123964, "grad_norm": 5.364743232727051, "learning_rate": 1.3957243486555428e-06, "loss": 5.3808, "step": 23723 }, { "epoch": 0.8359628954253548, "grad_norm": 18.23314094543457, "learning_rate": 1.395142841114333e-06, "loss": 2.8251, "step": 23724 }, { "epoch": 0.8359981324383132, "grad_norm": 1.0811160802841187, "learning_rate": 1.394561445652486e-06, "loss": 0.8899, "step": 23725 }, { "epoch": 0.8360333694512716, "grad_norm": 5.057388782501221, "learning_rate": 1.3939801622775673e-06, "loss": 4.4168, "step": 23726 }, { "epoch": 0.83606860646423, "grad_norm": 1.0228291749954224, "learning_rate": 1.3933989909971558e-06, "loss": 0.9233, "step": 23727 }, { "epoch": 0.8361038434771885, "grad_norm": 1.1150178909301758, "learning_rate": 1.3928179318188218e-06, "loss": 0.7399, "step": 23728 }, { "epoch": 0.8361390804901468, "grad_norm": 1.1558867692947388, "learning_rate": 1.3922369847501272e-06, "loss": 0.8588, "step": 23729 }, { "epoch": 0.8361743175031052, "grad_norm": 2.978652238845825, "learning_rate": 1.3916561497986403e-06, "loss": 2.571, "step": 23730 }, { "epoch": 0.8362095545160637, "grad_norm": 1.0212016105651855, "learning_rate": 1.391075426971934e-06, "loss": 1.0152, "step": 23731 }, { "epoch": 0.8362447915290221, "grad_norm": 5.44041633605957, "learning_rate": 1.3904948162775633e-06, "loss": 4.1733, "step": 23732 }, { "epoch": 0.8362800285419805, "grad_norm": 5.113409519195557, "learning_rate": 1.3899143177230957e-06, "loss": 5.7377, "step": 23733 }, { "epoch": 0.836315265554939, "grad_norm": 1.5811504125595093, "learning_rate": 1.3893339313160891e-06, "loss": 0.9408, "step": 23734 }, { "epoch": 0.8363505025678973, "grad_norm": 1.4160524606704712, "learning_rate": 1.3887536570641059e-06, "loss": 1.1857, "step": 23735 }, { "epoch": 0.8363857395808557, "grad_norm": 3.480163335800171, "learning_rate": 1.3881734949747038e-06, "loss": 2.7467, "step": 23736 }, { "epoch": 0.8364209765938141, "grad_norm": 10.354308128356934, "learning_rate": 1.3875934450554374e-06, "loss": 2.7283, "step": 23737 }, { "epoch": 0.8364562136067726, "grad_norm": 0.9859454035758972, "learning_rate": 1.3870135073138647e-06, "loss": 0.7708, "step": 23738 }, { "epoch": 0.836491450619731, "grad_norm": 3.3963615894317627, "learning_rate": 1.3864336817575374e-06, "loss": 2.8801, "step": 23739 }, { "epoch": 0.8365266876326893, "grad_norm": 3.84769344329834, "learning_rate": 1.3858539683940097e-06, "loss": 3.1053, "step": 23740 }, { "epoch": 0.8365619246456478, "grad_norm": 5.722720623016357, "learning_rate": 1.38527436723083e-06, "loss": 6.0898, "step": 23741 }, { "epoch": 0.8365971616586062, "grad_norm": 3.6264805793762207, "learning_rate": 1.3846948782755499e-06, "loss": 2.5705, "step": 23742 }, { "epoch": 0.8366323986715646, "grad_norm": 4.841343879699707, "learning_rate": 1.3841155015357155e-06, "loss": 2.6881, "step": 23743 }, { "epoch": 0.8366676356845231, "grad_norm": 4.438056945800781, "learning_rate": 1.3835362370188776e-06, "loss": 4.9168, "step": 23744 }, { "epoch": 0.8367028726974814, "grad_norm": 3.1887824535369873, "learning_rate": 1.3829570847325724e-06, "loss": 2.8638, "step": 23745 }, { "epoch": 0.8367381097104398, "grad_norm": 7.311418056488037, "learning_rate": 1.3823780446843514e-06, "loss": 5.2725, "step": 23746 }, { "epoch": 0.8367733467233982, "grad_norm": 1.357362985610962, "learning_rate": 1.3817991168817569e-06, "loss": 0.7086, "step": 23747 }, { "epoch": 0.8368085837363567, "grad_norm": 3.968829393386841, "learning_rate": 1.3812203013323233e-06, "loss": 2.7476, "step": 23748 }, { "epoch": 0.8368438207493151, "grad_norm": 3.74603271484375, "learning_rate": 1.3806415980435916e-06, "loss": 2.864, "step": 23749 }, { "epoch": 0.8368790577622734, "grad_norm": 2.5167040824890137, "learning_rate": 1.3800630070231046e-06, "loss": 2.078, "step": 23750 }, { "epoch": 0.8369142947752319, "grad_norm": 3.1540846824645996, "learning_rate": 1.3794845282783942e-06, "loss": 2.6644, "step": 23751 }, { "epoch": 0.8369495317881903, "grad_norm": 1.4701403379440308, "learning_rate": 1.3789061618169953e-06, "loss": 0.8001, "step": 23752 }, { "epoch": 0.8369847688011487, "grad_norm": 4.251919269561768, "learning_rate": 1.3783279076464406e-06, "loss": 4.9715, "step": 23753 }, { "epoch": 0.8370200058141072, "grad_norm": 4.719156265258789, "learning_rate": 1.3777497657742645e-06, "loss": 2.7839, "step": 23754 }, { "epoch": 0.8370552428270656, "grad_norm": 5.674561977386475, "learning_rate": 1.3771717362079962e-06, "loss": 2.8125, "step": 23755 }, { "epoch": 0.8370904798400239, "grad_norm": 2.2627906799316406, "learning_rate": 1.3765938189551632e-06, "loss": 0.7449, "step": 23756 }, { "epoch": 0.8371257168529824, "grad_norm": 1.1621953248977661, "learning_rate": 1.3760160140232948e-06, "loss": 0.6389, "step": 23757 }, { "epoch": 0.8371609538659408, "grad_norm": 2.1793925762176514, "learning_rate": 1.3754383214199164e-06, "loss": 0.9659, "step": 23758 }, { "epoch": 0.8371961908788992, "grad_norm": 3.0178325176239014, "learning_rate": 1.3748607411525527e-06, "loss": 3.1739, "step": 23759 }, { "epoch": 0.8372314278918576, "grad_norm": 3.9397027492523193, "learning_rate": 1.374283273228726e-06, "loss": 2.6501, "step": 23760 }, { "epoch": 0.837266664904816, "grad_norm": 1.684474229812622, "learning_rate": 1.3737059176559587e-06, "loss": 0.9171, "step": 23761 }, { "epoch": 0.8373019019177744, "grad_norm": 1.1142683029174805, "learning_rate": 1.3731286744417705e-06, "loss": 1.1346, "step": 23762 }, { "epoch": 0.8373371389307328, "grad_norm": 3.023522138595581, "learning_rate": 1.372551543593682e-06, "loss": 3.2611, "step": 23763 }, { "epoch": 0.8373723759436913, "grad_norm": 8.518898963928223, "learning_rate": 1.3719745251192052e-06, "loss": 4.6887, "step": 23764 }, { "epoch": 0.8374076129566497, "grad_norm": 10.748414039611816, "learning_rate": 1.3713976190258604e-06, "loss": 7.4545, "step": 23765 }, { "epoch": 0.837442849969608, "grad_norm": 4.72423791885376, "learning_rate": 1.370820825321164e-06, "loss": 5.1302, "step": 23766 }, { "epoch": 0.8374780869825665, "grad_norm": 1.8568601608276367, "learning_rate": 1.3702441440126223e-06, "loss": 0.701, "step": 23767 }, { "epoch": 0.8375133239955249, "grad_norm": 2.1352035999298096, "learning_rate": 1.369667575107748e-06, "loss": 1.0294, "step": 23768 }, { "epoch": 0.8375485610084833, "grad_norm": 4.685373306274414, "learning_rate": 1.3690911186140577e-06, "loss": 4.6501, "step": 23769 }, { "epoch": 0.8375837980214417, "grad_norm": 0.956798791885376, "learning_rate": 1.3685147745390525e-06, "loss": 0.7783, "step": 23770 }, { "epoch": 0.8376190350344002, "grad_norm": 0.9147884249687195, "learning_rate": 1.3679385428902425e-06, "loss": 0.802, "step": 23771 }, { "epoch": 0.8376542720473585, "grad_norm": 1.3106642961502075, "learning_rate": 1.3673624236751293e-06, "loss": 0.8427, "step": 23772 }, { "epoch": 0.8376895090603169, "grad_norm": 1.443102478981018, "learning_rate": 1.3667864169012247e-06, "loss": 1.0237, "step": 23773 }, { "epoch": 0.8377247460732754, "grad_norm": 1.8082873821258545, "learning_rate": 1.3662105225760247e-06, "loss": 0.8139, "step": 23774 }, { "epoch": 0.8377599830862338, "grad_norm": 3.3664307594299316, "learning_rate": 1.3656347407070325e-06, "loss": 3.1331, "step": 23775 }, { "epoch": 0.8377952200991922, "grad_norm": 1.091805100440979, "learning_rate": 1.3650590713017476e-06, "loss": 0.7655, "step": 23776 }, { "epoch": 0.8378304571121507, "grad_norm": 3.293944835662842, "learning_rate": 1.3644835143676681e-06, "loss": 2.5761, "step": 23777 }, { "epoch": 0.837865694125109, "grad_norm": 1.019051432609558, "learning_rate": 1.3639080699122908e-06, "loss": 1.1814, "step": 23778 }, { "epoch": 0.8379009311380674, "grad_norm": 1.8639500141143799, "learning_rate": 1.3633327379431117e-06, "loss": 0.7673, "step": 23779 }, { "epoch": 0.8379361681510258, "grad_norm": 4.286108493804932, "learning_rate": 1.3627575184676223e-06, "loss": 2.4162, "step": 23780 }, { "epoch": 0.8379714051639843, "grad_norm": 4.05135440826416, "learning_rate": 1.362182411493318e-06, "loss": 3.0839, "step": 23781 }, { "epoch": 0.8380066421769427, "grad_norm": 3.369232416152954, "learning_rate": 1.3616074170276894e-06, "loss": 4.952, "step": 23782 }, { "epoch": 0.838041879189901, "grad_norm": 4.815995693206787, "learning_rate": 1.3610325350782205e-06, "loss": 4.36, "step": 23783 }, { "epoch": 0.8380771162028595, "grad_norm": 1.0113072395324707, "learning_rate": 1.3604577656524055e-06, "loss": 0.9866, "step": 23784 }, { "epoch": 0.8381123532158179, "grad_norm": 5.441813945770264, "learning_rate": 1.3598831087577303e-06, "loss": 2.7382, "step": 23785 }, { "epoch": 0.8381475902287763, "grad_norm": 0.8291784524917603, "learning_rate": 1.3593085644016757e-06, "loss": 0.7576, "step": 23786 }, { "epoch": 0.8381828272417348, "grad_norm": 4.4882731437683105, "learning_rate": 1.358734132591727e-06, "loss": 2.8957, "step": 23787 }, { "epoch": 0.8382180642546931, "grad_norm": 3.5637898445129395, "learning_rate": 1.3581598133353702e-06, "loss": 3.2438, "step": 23788 }, { "epoch": 0.8382533012676515, "grad_norm": 1.2165560722351074, "learning_rate": 1.3575856066400806e-06, "loss": 0.9638, "step": 23789 }, { "epoch": 0.83828853828061, "grad_norm": 1.4972807168960571, "learning_rate": 1.3570115125133387e-06, "loss": 0.8005, "step": 23790 }, { "epoch": 0.8383237752935684, "grad_norm": 3.405803918838501, "learning_rate": 1.356437530962622e-06, "loss": 2.8675, "step": 23791 }, { "epoch": 0.8383590123065268, "grad_norm": 4.934432506561279, "learning_rate": 1.3558636619954113e-06, "loss": 3.1504, "step": 23792 }, { "epoch": 0.8383942493194851, "grad_norm": 0.9270363450050354, "learning_rate": 1.3552899056191748e-06, "loss": 0.7156, "step": 23793 }, { "epoch": 0.8384294863324436, "grad_norm": 1.1201047897338867, "learning_rate": 1.3547162618413889e-06, "loss": 0.8954, "step": 23794 }, { "epoch": 0.838464723345402, "grad_norm": 4.661647319793701, "learning_rate": 1.3541427306695244e-06, "loss": 3.372, "step": 23795 }, { "epoch": 0.8384999603583604, "grad_norm": 3.550778388977051, "learning_rate": 1.3535693121110527e-06, "loss": 2.5458, "step": 23796 }, { "epoch": 0.8385351973713189, "grad_norm": 3.142233371734619, "learning_rate": 1.3529960061734426e-06, "loss": 2.9376, "step": 23797 }, { "epoch": 0.8385704343842773, "grad_norm": 4.410864353179932, "learning_rate": 1.3524228128641603e-06, "loss": 2.6255, "step": 23798 }, { "epoch": 0.8386056713972356, "grad_norm": 2.9244213104248047, "learning_rate": 1.351849732190672e-06, "loss": 3.1238, "step": 23799 }, { "epoch": 0.8386409084101941, "grad_norm": 0.9224763512611389, "learning_rate": 1.351276764160443e-06, "loss": 0.6233, "step": 23800 }, { "epoch": 0.8386761454231525, "grad_norm": 1.0412402153015137, "learning_rate": 1.3507039087809382e-06, "loss": 1.1821, "step": 23801 }, { "epoch": 0.8387113824361109, "grad_norm": 1.65201735496521, "learning_rate": 1.3501311660596116e-06, "loss": 0.7338, "step": 23802 }, { "epoch": 0.8387466194490693, "grad_norm": 3.1512327194213867, "learning_rate": 1.349558536003931e-06, "loss": 2.5883, "step": 23803 }, { "epoch": 0.8387818564620277, "grad_norm": 1.268510341644287, "learning_rate": 1.3489860186213544e-06, "loss": 0.6732, "step": 23804 }, { "epoch": 0.8388170934749861, "grad_norm": 1.440808653831482, "learning_rate": 1.3484136139193349e-06, "loss": 0.621, "step": 23805 }, { "epoch": 0.8388523304879445, "grad_norm": 1.495230793952942, "learning_rate": 1.3478413219053288e-06, "loss": 0.6673, "step": 23806 }, { "epoch": 0.838887567500903, "grad_norm": 5.527683258056641, "learning_rate": 1.3472691425867944e-06, "loss": 2.9744, "step": 23807 }, { "epoch": 0.8389228045138614, "grad_norm": 2.179823875427246, "learning_rate": 1.3466970759711807e-06, "loss": 2.8023, "step": 23808 }, { "epoch": 0.8389580415268197, "grad_norm": 2.944058418273926, "learning_rate": 1.3461251220659387e-06, "loss": 0.6995, "step": 23809 }, { "epoch": 0.8389932785397782, "grad_norm": 3.252194404602051, "learning_rate": 1.3455532808785176e-06, "loss": 2.5445, "step": 23810 }, { "epoch": 0.8390285155527366, "grad_norm": 4.374547958374023, "learning_rate": 1.3449815524163722e-06, "loss": 2.8806, "step": 23811 }, { "epoch": 0.839063752565695, "grad_norm": 2.7570414543151855, "learning_rate": 1.3444099366869423e-06, "loss": 2.3076, "step": 23812 }, { "epoch": 0.8390989895786534, "grad_norm": 2.1803417205810547, "learning_rate": 1.3438384336976751e-06, "loss": 1.1131, "step": 23813 }, { "epoch": 0.8391342265916119, "grad_norm": 7.394811630249023, "learning_rate": 1.3432670434560147e-06, "loss": 4.8095, "step": 23814 }, { "epoch": 0.8391694636045702, "grad_norm": 5.919139862060547, "learning_rate": 1.342695765969404e-06, "loss": 4.9024, "step": 23815 }, { "epoch": 0.8392047006175286, "grad_norm": 6.9554619789123535, "learning_rate": 1.3421246012452838e-06, "loss": 3.0955, "step": 23816 }, { "epoch": 0.8392399376304871, "grad_norm": 3.6876189708709717, "learning_rate": 1.3415535492910936e-06, "loss": 2.828, "step": 23817 }, { "epoch": 0.8392751746434455, "grad_norm": 4.473631381988525, "learning_rate": 1.340982610114272e-06, "loss": 5.1091, "step": 23818 }, { "epoch": 0.8393104116564039, "grad_norm": 1.7744877338409424, "learning_rate": 1.3404117837222542e-06, "loss": 0.877, "step": 23819 }, { "epoch": 0.8393456486693623, "grad_norm": 4.507968425750732, "learning_rate": 1.3398410701224773e-06, "loss": 2.6987, "step": 23820 }, { "epoch": 0.8393808856823207, "grad_norm": 16.78060531616211, "learning_rate": 1.3392704693223701e-06, "loss": 3.6444, "step": 23821 }, { "epoch": 0.8394161226952791, "grad_norm": 1.5556975603103638, "learning_rate": 1.3386999813293711e-06, "loss": 0.851, "step": 23822 }, { "epoch": 0.8394513597082376, "grad_norm": 5.148254871368408, "learning_rate": 1.3381296061509097e-06, "loss": 4.8055, "step": 23823 }, { "epoch": 0.839486596721196, "grad_norm": 5.428715705871582, "learning_rate": 1.3375593437944112e-06, "loss": 5.4218, "step": 23824 }, { "epoch": 0.8395218337341543, "grad_norm": 8.07925033569336, "learning_rate": 1.3369891942673031e-06, "loss": 4.6845, "step": 23825 }, { "epoch": 0.8395570707471127, "grad_norm": 3.028838872909546, "learning_rate": 1.3364191575770191e-06, "loss": 2.8141, "step": 23826 }, { "epoch": 0.8395923077600712, "grad_norm": 3.266852378845215, "learning_rate": 1.335849233730977e-06, "loss": 3.1287, "step": 23827 }, { "epoch": 0.8396275447730296, "grad_norm": 6.265598297119141, "learning_rate": 1.3352794227366017e-06, "loss": 5.6368, "step": 23828 }, { "epoch": 0.839662781785988, "grad_norm": 3.50730299949646, "learning_rate": 1.334709724601314e-06, "loss": 2.5282, "step": 23829 }, { "epoch": 0.8396980187989465, "grad_norm": 1.7255656719207764, "learning_rate": 1.3341401393325404e-06, "loss": 0.9908, "step": 23830 }, { "epoch": 0.8397332558119048, "grad_norm": 3.658684492111206, "learning_rate": 1.3335706669376925e-06, "loss": 2.6548, "step": 23831 }, { "epoch": 0.8397684928248632, "grad_norm": 2.1624999046325684, "learning_rate": 1.3330013074241909e-06, "loss": 2.0579, "step": 23832 }, { "epoch": 0.8398037298378217, "grad_norm": 1.163913607597351, "learning_rate": 1.3324320607994513e-06, "loss": 1.0716, "step": 23833 }, { "epoch": 0.8398389668507801, "grad_norm": 2.7312333583831787, "learning_rate": 1.3318629270708882e-06, "loss": 2.479, "step": 23834 }, { "epoch": 0.8398742038637385, "grad_norm": 3.3522093296051025, "learning_rate": 1.3312939062459153e-06, "loss": 2.5699, "step": 23835 }, { "epoch": 0.8399094408766968, "grad_norm": 4.587559223175049, "learning_rate": 1.3307249983319425e-06, "loss": 4.9252, "step": 23836 }, { "epoch": 0.8399446778896553, "grad_norm": 0.9215252995491028, "learning_rate": 1.3301562033363814e-06, "loss": 1.032, "step": 23837 }, { "epoch": 0.8399799149026137, "grad_norm": 3.7741708755493164, "learning_rate": 1.3295875212666397e-06, "loss": 3.1097, "step": 23838 }, { "epoch": 0.8400151519155721, "grad_norm": 7.848482131958008, "learning_rate": 1.3290189521301278e-06, "loss": 2.9373, "step": 23839 }, { "epoch": 0.8400503889285306, "grad_norm": 6.455053329467773, "learning_rate": 1.3284504959342436e-06, "loss": 3.2919, "step": 23840 }, { "epoch": 0.840085625941489, "grad_norm": 1.3170340061187744, "learning_rate": 1.3278821526863983e-06, "loss": 1.0262, "step": 23841 }, { "epoch": 0.8401208629544473, "grad_norm": 3.8910348415374756, "learning_rate": 1.3273139223939958e-06, "loss": 2.5738, "step": 23842 }, { "epoch": 0.8401560999674058, "grad_norm": 2.439133405685425, "learning_rate": 1.326745805064431e-06, "loss": 2.4534, "step": 23843 }, { "epoch": 0.8401913369803642, "grad_norm": 3.765659809112549, "learning_rate": 1.3261778007051052e-06, "loss": 2.5978, "step": 23844 }, { "epoch": 0.8402265739933226, "grad_norm": 3.4081146717071533, "learning_rate": 1.3256099093234197e-06, "loss": 2.7888, "step": 23845 }, { "epoch": 0.840261811006281, "grad_norm": 3.958038568496704, "learning_rate": 1.3250421309267736e-06, "loss": 2.6034, "step": 23846 }, { "epoch": 0.8402970480192394, "grad_norm": 1.186750054359436, "learning_rate": 1.324474465522556e-06, "loss": 0.8903, "step": 23847 }, { "epoch": 0.8403322850321978, "grad_norm": 1.228767991065979, "learning_rate": 1.3239069131181614e-06, "loss": 0.9026, "step": 23848 }, { "epoch": 0.8403675220451562, "grad_norm": 6.872649669647217, "learning_rate": 1.3233394737209892e-06, "loss": 7.0896, "step": 23849 }, { "epoch": 0.8404027590581147, "grad_norm": 4.723219394683838, "learning_rate": 1.3227721473384224e-06, "loss": 3.3571, "step": 23850 }, { "epoch": 0.8404379960710731, "grad_norm": 1.183332085609436, "learning_rate": 1.3222049339778542e-06, "loss": 0.9372, "step": 23851 }, { "epoch": 0.8404732330840314, "grad_norm": 3.7257871627807617, "learning_rate": 1.3216378336466719e-06, "loss": 2.5616, "step": 23852 }, { "epoch": 0.8405084700969899, "grad_norm": 1.3548948764801025, "learning_rate": 1.3210708463522626e-06, "loss": 0.7373, "step": 23853 }, { "epoch": 0.8405437071099483, "grad_norm": 2.6127636432647705, "learning_rate": 1.3205039721020108e-06, "loss": 0.9004, "step": 23854 }, { "epoch": 0.8405789441229067, "grad_norm": 1.802437663078308, "learning_rate": 1.3199372109033015e-06, "loss": 0.762, "step": 23855 }, { "epoch": 0.8406141811358651, "grad_norm": 1.361654281616211, "learning_rate": 1.3193705627635144e-06, "loss": 0.7698, "step": 23856 }, { "epoch": 0.8406494181488235, "grad_norm": 5.796951770782471, "learning_rate": 1.3188040276900317e-06, "loss": 3.5191, "step": 23857 }, { "epoch": 0.8406846551617819, "grad_norm": 0.6370079517364502, "learning_rate": 1.3182376056902345e-06, "loss": 0.7488, "step": 23858 }, { "epoch": 0.8407198921747403, "grad_norm": 1.0458295345306396, "learning_rate": 1.3176712967714945e-06, "loss": 1.0148, "step": 23859 }, { "epoch": 0.8407551291876988, "grad_norm": 0.7611309289932251, "learning_rate": 1.3171051009411929e-06, "loss": 0.6752, "step": 23860 }, { "epoch": 0.8407903662006572, "grad_norm": 4.133496284484863, "learning_rate": 1.3165390182067062e-06, "loss": 3.1135, "step": 23861 }, { "epoch": 0.8408256032136155, "grad_norm": 2.9456658363342285, "learning_rate": 1.315973048575403e-06, "loss": 1.0822, "step": 23862 }, { "epoch": 0.840860840226574, "grad_norm": 7.498083591461182, "learning_rate": 1.3154071920546541e-06, "loss": 3.2949, "step": 23863 }, { "epoch": 0.8408960772395324, "grad_norm": 7.389756679534912, "learning_rate": 1.3148414486518358e-06, "loss": 5.1405, "step": 23864 }, { "epoch": 0.8409313142524908, "grad_norm": 3.3051164150238037, "learning_rate": 1.3142758183743153e-06, "loss": 3.0014, "step": 23865 }, { "epoch": 0.8409665512654493, "grad_norm": 1.2611732482910156, "learning_rate": 1.3137103012294572e-06, "loss": 0.7188, "step": 23866 }, { "epoch": 0.8410017882784077, "grad_norm": 5.421111106872559, "learning_rate": 1.3131448972246262e-06, "loss": 3.3397, "step": 23867 }, { "epoch": 0.841037025291366, "grad_norm": 4.433513641357422, "learning_rate": 1.3125796063671948e-06, "loss": 3.5012, "step": 23868 }, { "epoch": 0.8410722623043244, "grad_norm": 1.1488505601882935, "learning_rate": 1.3120144286645176e-06, "loss": 0.8434, "step": 23869 }, { "epoch": 0.8411074993172829, "grad_norm": 9.137964248657227, "learning_rate": 1.311449364123959e-06, "loss": 2.8847, "step": 23870 }, { "epoch": 0.8411427363302413, "grad_norm": 1.2766145467758179, "learning_rate": 1.3108844127528797e-06, "loss": 1.0743, "step": 23871 }, { "epoch": 0.8411779733431997, "grad_norm": 1.1115832328796387, "learning_rate": 1.3103195745586373e-06, "loss": 0.901, "step": 23872 }, { "epoch": 0.8412132103561581, "grad_norm": 1.3142096996307373, "learning_rate": 1.309754849548589e-06, "loss": 0.6868, "step": 23873 }, { "epoch": 0.8412484473691165, "grad_norm": 3.7184090614318848, "learning_rate": 1.3091902377300913e-06, "loss": 2.7112, "step": 23874 }, { "epoch": 0.8412836843820749, "grad_norm": 1.4708831310272217, "learning_rate": 1.3086257391104983e-06, "loss": 0.8769, "step": 23875 }, { "epoch": 0.8413189213950334, "grad_norm": 3.9943339824676514, "learning_rate": 1.308061353697161e-06, "loss": 3.1514, "step": 23876 }, { "epoch": 0.8413541584079918, "grad_norm": 3.2167227268218994, "learning_rate": 1.3074970814974342e-06, "loss": 2.8441, "step": 23877 }, { "epoch": 0.8413893954209501, "grad_norm": 1.639852523803711, "learning_rate": 1.306932922518661e-06, "loss": 1.0053, "step": 23878 }, { "epoch": 0.8414246324339085, "grad_norm": 3.5166683197021484, "learning_rate": 1.3063688767681971e-06, "loss": 2.5466, "step": 23879 }, { "epoch": 0.841459869446867, "grad_norm": 4.082133769989014, "learning_rate": 1.3058049442533871e-06, "loss": 2.5586, "step": 23880 }, { "epoch": 0.8414951064598254, "grad_norm": 2.713059425354004, "learning_rate": 1.3052411249815722e-06, "loss": 2.7528, "step": 23881 }, { "epoch": 0.8415303434727838, "grad_norm": 1.249776840209961, "learning_rate": 1.3046774189600985e-06, "loss": 0.7046, "step": 23882 }, { "epoch": 0.8415655804857423, "grad_norm": 1.5516899824142456, "learning_rate": 1.3041138261963104e-06, "loss": 0.8719, "step": 23883 }, { "epoch": 0.8416008174987006, "grad_norm": 1.0790553092956543, "learning_rate": 1.3035503466975496e-06, "loss": 0.8951, "step": 23884 }, { "epoch": 0.841636054511659, "grad_norm": 3.3630964756011963, "learning_rate": 1.3029869804711515e-06, "loss": 2.9749, "step": 23885 }, { "epoch": 0.8416712915246175, "grad_norm": 3.9671342372894287, "learning_rate": 1.3024237275244523e-06, "loss": 4.7653, "step": 23886 }, { "epoch": 0.8417065285375759, "grad_norm": 1.8500583171844482, "learning_rate": 1.3018605878647982e-06, "loss": 0.8611, "step": 23887 }, { "epoch": 0.8417417655505343, "grad_norm": 1.4062267541885376, "learning_rate": 1.301297561499515e-06, "loss": 0.9221, "step": 23888 }, { "epoch": 0.8417770025634926, "grad_norm": 1.006005883216858, "learning_rate": 1.3007346484359384e-06, "loss": 0.8923, "step": 23889 }, { "epoch": 0.8418122395764511, "grad_norm": 6.081780910491943, "learning_rate": 1.3001718486814018e-06, "loss": 4.7179, "step": 23890 }, { "epoch": 0.8418474765894095, "grad_norm": 2.5816214084625244, "learning_rate": 1.299609162243235e-06, "loss": 2.5336, "step": 23891 }, { "epoch": 0.8418827136023679, "grad_norm": 4.761343002319336, "learning_rate": 1.2990465891287663e-06, "loss": 3.2116, "step": 23892 }, { "epoch": 0.8419179506153264, "grad_norm": 4.137990951538086, "learning_rate": 1.2984841293453254e-06, "loss": 4.5848, "step": 23893 }, { "epoch": 0.8419531876282847, "grad_norm": 3.632568359375, "learning_rate": 1.2979217829002367e-06, "loss": 2.9078, "step": 23894 }, { "epoch": 0.8419884246412431, "grad_norm": 4.245673179626465, "learning_rate": 1.2973595498008252e-06, "loss": 2.9321, "step": 23895 }, { "epoch": 0.8420236616542016, "grad_norm": 4.40200662612915, "learning_rate": 1.2967974300544162e-06, "loss": 5.2915, "step": 23896 }, { "epoch": 0.84205889866716, "grad_norm": 3.786104440689087, "learning_rate": 1.296235423668325e-06, "loss": 4.2565, "step": 23897 }, { "epoch": 0.8420941356801184, "grad_norm": 2.938447952270508, "learning_rate": 1.295673530649878e-06, "loss": 2.5256, "step": 23898 }, { "epoch": 0.8421293726930769, "grad_norm": 1.3135331869125366, "learning_rate": 1.2951117510063937e-06, "loss": 0.8666, "step": 23899 }, { "epoch": 0.8421646097060352, "grad_norm": 2.1421022415161133, "learning_rate": 1.2945500847451864e-06, "loss": 2.6198, "step": 23900 }, { "epoch": 0.8421998467189936, "grad_norm": 4.576263427734375, "learning_rate": 1.2939885318735712e-06, "loss": 6.9441, "step": 23901 }, { "epoch": 0.842235083731952, "grad_norm": 1.0261595249176025, "learning_rate": 1.2934270923988657e-06, "loss": 0.6871, "step": 23902 }, { "epoch": 0.8422703207449105, "grad_norm": 1.7237330675125122, "learning_rate": 1.292865766328384e-06, "loss": 1.1305, "step": 23903 }, { "epoch": 0.8423055577578689, "grad_norm": 2.3836231231689453, "learning_rate": 1.2923045536694324e-06, "loss": 0.938, "step": 23904 }, { "epoch": 0.8423407947708272, "grad_norm": 4.477309226989746, "learning_rate": 1.2917434544293205e-06, "loss": 2.8304, "step": 23905 }, { "epoch": 0.8423760317837857, "grad_norm": 3.284090757369995, "learning_rate": 1.2911824686153652e-06, "loss": 3.0666, "step": 23906 }, { "epoch": 0.8424112687967441, "grad_norm": 1.2167210578918457, "learning_rate": 1.2906215962348644e-06, "loss": 1.0237, "step": 23907 }, { "epoch": 0.8424465058097025, "grad_norm": 4.508981704711914, "learning_rate": 1.2900608372951263e-06, "loss": 2.668, "step": 23908 }, { "epoch": 0.842481742822661, "grad_norm": 0.7906919717788696, "learning_rate": 1.2895001918034555e-06, "loss": 0.9411, "step": 23909 }, { "epoch": 0.8425169798356194, "grad_norm": 4.705386638641357, "learning_rate": 1.2889396597671544e-06, "loss": 4.5062, "step": 23910 }, { "epoch": 0.8425522168485777, "grad_norm": 5.597592830657959, "learning_rate": 1.2883792411935248e-06, "loss": 5.2488, "step": 23911 }, { "epoch": 0.8425874538615361, "grad_norm": 4.262453079223633, "learning_rate": 1.2878189360898641e-06, "loss": 2.9738, "step": 23912 }, { "epoch": 0.8426226908744946, "grad_norm": 1.0953255891799927, "learning_rate": 1.287258744463472e-06, "loss": 0.8506, "step": 23913 }, { "epoch": 0.842657927887453, "grad_norm": 1.3400298357009888, "learning_rate": 1.2866986663216442e-06, "loss": 0.689, "step": 23914 }, { "epoch": 0.8426931649004114, "grad_norm": 1.279241681098938, "learning_rate": 1.2861387016716787e-06, "loss": 0.9659, "step": 23915 }, { "epoch": 0.8427284019133698, "grad_norm": 1.3586453199386597, "learning_rate": 1.2855788505208621e-06, "loss": 0.7588, "step": 23916 }, { "epoch": 0.8427636389263282, "grad_norm": 5.44785737991333, "learning_rate": 1.2850191128764932e-06, "loss": 5.2687, "step": 23917 }, { "epoch": 0.8427988759392866, "grad_norm": 1.0878616571426392, "learning_rate": 1.2844594887458605e-06, "loss": 1.0674, "step": 23918 }, { "epoch": 0.8428341129522451, "grad_norm": 5.388010501861572, "learning_rate": 1.2838999781362548e-06, "loss": 3.1704, "step": 23919 }, { "epoch": 0.8428693499652035, "grad_norm": 3.7804510593414307, "learning_rate": 1.2833405810549582e-06, "loss": 2.3686, "step": 23920 }, { "epoch": 0.8429045869781618, "grad_norm": 1.0975165367126465, "learning_rate": 1.2827812975092636e-06, "loss": 0.9316, "step": 23921 }, { "epoch": 0.8429398239911202, "grad_norm": 7.193370342254639, "learning_rate": 1.282222127506455e-06, "loss": 4.9473, "step": 23922 }, { "epoch": 0.8429750610040787, "grad_norm": 0.9053157567977905, "learning_rate": 1.2816630710538102e-06, "loss": 1.0031, "step": 23923 }, { "epoch": 0.8430102980170371, "grad_norm": 5.306347846984863, "learning_rate": 1.2811041281586157e-06, "loss": 5.2064, "step": 23924 }, { "epoch": 0.8430455350299955, "grad_norm": 5.634690284729004, "learning_rate": 1.2805452988281507e-06, "loss": 4.7195, "step": 23925 }, { "epoch": 0.843080772042954, "grad_norm": 1.6426894664764404, "learning_rate": 1.2799865830696934e-06, "loss": 0.7742, "step": 23926 }, { "epoch": 0.8431160090559123, "grad_norm": 2.4414100646972656, "learning_rate": 1.2794279808905207e-06, "loss": 2.6939, "step": 23927 }, { "epoch": 0.8431512460688707, "grad_norm": 0.9816379547119141, "learning_rate": 1.2788694922979106e-06, "loss": 1.1896, "step": 23928 }, { "epoch": 0.8431864830818292, "grad_norm": 1.4436192512512207, "learning_rate": 1.2783111172991357e-06, "loss": 0.9481, "step": 23929 }, { "epoch": 0.8432217200947876, "grad_norm": 11.54176139831543, "learning_rate": 1.2777528559014707e-06, "loss": 2.848, "step": 23930 }, { "epoch": 0.843256957107746, "grad_norm": 3.5978877544403076, "learning_rate": 1.277194708112184e-06, "loss": 3.3774, "step": 23931 }, { "epoch": 0.8432921941207044, "grad_norm": 1.8686931133270264, "learning_rate": 1.2766366739385482e-06, "loss": 0.9444, "step": 23932 }, { "epoch": 0.8433274311336628, "grad_norm": 1.2049181461334229, "learning_rate": 1.2760787533878304e-06, "loss": 1.1457, "step": 23933 }, { "epoch": 0.8433626681466212, "grad_norm": 3.4423282146453857, "learning_rate": 1.2755209464673014e-06, "loss": 3.0813, "step": 23934 }, { "epoch": 0.8433979051595796, "grad_norm": 7.4848809242248535, "learning_rate": 1.2749632531842204e-06, "loss": 5.2739, "step": 23935 }, { "epoch": 0.8434331421725381, "grad_norm": 2.477583885192871, "learning_rate": 1.2744056735458521e-06, "loss": 2.6361, "step": 23936 }, { "epoch": 0.8434683791854964, "grad_norm": 1.2602336406707764, "learning_rate": 1.2738482075594638e-06, "loss": 1.253, "step": 23937 }, { "epoch": 0.8435036161984548, "grad_norm": 1.2113070487976074, "learning_rate": 1.2732908552323165e-06, "loss": 0.9797, "step": 23938 }, { "epoch": 0.8435388532114133, "grad_norm": 1.4525564908981323, "learning_rate": 1.2727336165716654e-06, "loss": 0.8346, "step": 23939 }, { "epoch": 0.8435740902243717, "grad_norm": 1.5280791521072388, "learning_rate": 1.2721764915847679e-06, "loss": 1.177, "step": 23940 }, { "epoch": 0.8436093272373301, "grad_norm": 5.146205425262451, "learning_rate": 1.2716194802788873e-06, "loss": 5.0488, "step": 23941 }, { "epoch": 0.8436445642502886, "grad_norm": 1.4133955240249634, "learning_rate": 1.2710625826612733e-06, "loss": 1.0685, "step": 23942 }, { "epoch": 0.8436798012632469, "grad_norm": 1.3555132150650024, "learning_rate": 1.270505798739181e-06, "loss": 0.8625, "step": 23943 }, { "epoch": 0.8437150382762053, "grad_norm": 3.4029996395111084, "learning_rate": 1.269949128519863e-06, "loss": 2.5885, "step": 23944 }, { "epoch": 0.8437502752891637, "grad_norm": 0.8766750693321228, "learning_rate": 1.2693925720105693e-06, "loss": 0.6948, "step": 23945 }, { "epoch": 0.8437855123021222, "grad_norm": 1.1446141004562378, "learning_rate": 1.2688361292185492e-06, "loss": 0.967, "step": 23946 }, { "epoch": 0.8438207493150806, "grad_norm": 3.7040436267852783, "learning_rate": 1.268279800151051e-06, "loss": 2.5905, "step": 23947 }, { "epoch": 0.8438559863280389, "grad_norm": 4.623743057250977, "learning_rate": 1.2677235848153202e-06, "loss": 3.1052, "step": 23948 }, { "epoch": 0.8438912233409974, "grad_norm": 1.2868577241897583, "learning_rate": 1.267167483218602e-06, "loss": 0.9035, "step": 23949 }, { "epoch": 0.8439264603539558, "grad_norm": 3.240436553955078, "learning_rate": 1.2666114953681419e-06, "loss": 2.9322, "step": 23950 }, { "epoch": 0.8439616973669142, "grad_norm": 8.74678897857666, "learning_rate": 1.266055621271176e-06, "loss": 4.7094, "step": 23951 }, { "epoch": 0.8439969343798727, "grad_norm": 7.291079521179199, "learning_rate": 1.26549986093495e-06, "loss": 5.2934, "step": 23952 }, { "epoch": 0.844032171392831, "grad_norm": 5.714992046356201, "learning_rate": 1.2649442143667024e-06, "loss": 4.7918, "step": 23953 }, { "epoch": 0.8440674084057894, "grad_norm": 7.276212215423584, "learning_rate": 1.2643886815736673e-06, "loss": 5.2522, "step": 23954 }, { "epoch": 0.8441026454187478, "grad_norm": 6.874399662017822, "learning_rate": 1.263833262563081e-06, "loss": 5.8976, "step": 23955 }, { "epoch": 0.8441378824317063, "grad_norm": 2.7496492862701416, "learning_rate": 1.2632779573421806e-06, "loss": 3.0163, "step": 23956 }, { "epoch": 0.8441731194446647, "grad_norm": 8.127028465270996, "learning_rate": 1.2627227659181995e-06, "loss": 3.002, "step": 23957 }, { "epoch": 0.844208356457623, "grad_norm": 3.9649717807769775, "learning_rate": 1.2621676882983657e-06, "loss": 2.9587, "step": 23958 }, { "epoch": 0.8442435934705815, "grad_norm": 1.1434845924377441, "learning_rate": 1.261612724489909e-06, "loss": 0.6055, "step": 23959 }, { "epoch": 0.8442788304835399, "grad_norm": 0.7596133947372437, "learning_rate": 1.2610578745000634e-06, "loss": 0.7732, "step": 23960 }, { "epoch": 0.8443140674964983, "grad_norm": 2.3541359901428223, "learning_rate": 1.26050313833605e-06, "loss": 0.8885, "step": 23961 }, { "epoch": 0.8443493045094568, "grad_norm": 3.969865560531616, "learning_rate": 1.2599485160050962e-06, "loss": 4.9558, "step": 23962 }, { "epoch": 0.8443845415224152, "grad_norm": 2.50777530670166, "learning_rate": 1.2593940075144274e-06, "loss": 2.437, "step": 23963 }, { "epoch": 0.8444197785353735, "grad_norm": 1.341519832611084, "learning_rate": 1.2588396128712643e-06, "loss": 0.799, "step": 23964 }, { "epoch": 0.844455015548332, "grad_norm": 6.8298821449279785, "learning_rate": 1.2582853320828282e-06, "loss": 7.2694, "step": 23965 }, { "epoch": 0.8444902525612904, "grad_norm": 1.2430132627487183, "learning_rate": 1.2577311651563395e-06, "loss": 0.9066, "step": 23966 }, { "epoch": 0.8445254895742488, "grad_norm": 2.6278624534606934, "learning_rate": 1.2571771120990162e-06, "loss": 2.7375, "step": 23967 }, { "epoch": 0.8445607265872072, "grad_norm": 1.2985756397247314, "learning_rate": 1.2566231729180744e-06, "loss": 0.8471, "step": 23968 }, { "epoch": 0.8445959636001656, "grad_norm": 1.8084083795547485, "learning_rate": 1.256069347620732e-06, "loss": 0.7517, "step": 23969 }, { "epoch": 0.844631200613124, "grad_norm": 3.0056872367858887, "learning_rate": 1.2555156362141963e-06, "loss": 2.3036, "step": 23970 }, { "epoch": 0.8446664376260824, "grad_norm": 3.2289493083953857, "learning_rate": 1.254962038705685e-06, "loss": 2.5575, "step": 23971 }, { "epoch": 0.8447016746390409, "grad_norm": 1.4119408130645752, "learning_rate": 1.2544085551024065e-06, "loss": 0.9054, "step": 23972 }, { "epoch": 0.8447369116519993, "grad_norm": 1.8355326652526855, "learning_rate": 1.2538551854115732e-06, "loss": 0.7504, "step": 23973 }, { "epoch": 0.8447721486649576, "grad_norm": 4.027858734130859, "learning_rate": 1.253301929640386e-06, "loss": 2.7324, "step": 23974 }, { "epoch": 0.8448073856779161, "grad_norm": 1.1331017017364502, "learning_rate": 1.2527487877960576e-06, "loss": 0.8251, "step": 23975 }, { "epoch": 0.8448426226908745, "grad_norm": 0.9752628207206726, "learning_rate": 1.2521957598857915e-06, "loss": 0.7815, "step": 23976 }, { "epoch": 0.8448778597038329, "grad_norm": 1.168068528175354, "learning_rate": 1.2516428459167883e-06, "loss": 0.6707, "step": 23977 }, { "epoch": 0.8449130967167913, "grad_norm": 1.1158325672149658, "learning_rate": 1.2510900458962504e-06, "loss": 0.8324, "step": 23978 }, { "epoch": 0.8449483337297498, "grad_norm": 2.0592310428619385, "learning_rate": 1.250537359831382e-06, "loss": 0.7669, "step": 23979 }, { "epoch": 0.8449835707427081, "grad_norm": 3.873467206954956, "learning_rate": 1.249984787729378e-06, "loss": 2.399, "step": 23980 }, { "epoch": 0.8450188077556665, "grad_norm": 1.0097682476043701, "learning_rate": 1.2494323295974353e-06, "loss": 1.0042, "step": 23981 }, { "epoch": 0.845054044768625, "grad_norm": 1.0536847114562988, "learning_rate": 1.2488799854427525e-06, "loss": 0.8886, "step": 23982 }, { "epoch": 0.8450892817815834, "grad_norm": 4.602229118347168, "learning_rate": 1.2483277552725214e-06, "loss": 4.7762, "step": 23983 }, { "epoch": 0.8451245187945418, "grad_norm": 1.2376700639724731, "learning_rate": 1.2477756390939376e-06, "loss": 0.6681, "step": 23984 }, { "epoch": 0.8451597558075002, "grad_norm": 3.347076416015625, "learning_rate": 1.2472236369141898e-06, "loss": 3.0249, "step": 23985 }, { "epoch": 0.8451949928204586, "grad_norm": 4.917965412139893, "learning_rate": 1.2466717487404689e-06, "loss": 4.335, "step": 23986 }, { "epoch": 0.845230229833417, "grad_norm": 3.033968210220337, "learning_rate": 1.2461199745799646e-06, "loss": 2.2801, "step": 23987 }, { "epoch": 0.8452654668463754, "grad_norm": 2.2749571800231934, "learning_rate": 1.2455683144398634e-06, "loss": 0.8639, "step": 23988 }, { "epoch": 0.8453007038593339, "grad_norm": 0.909091591835022, "learning_rate": 1.2450167683273473e-06, "loss": 1.0857, "step": 23989 }, { "epoch": 0.8453359408722922, "grad_norm": 4.643585681915283, "learning_rate": 1.244465336249604e-06, "loss": 3.1939, "step": 23990 }, { "epoch": 0.8453711778852506, "grad_norm": 3.4691805839538574, "learning_rate": 1.2439140182138154e-06, "loss": 2.7595, "step": 23991 }, { "epoch": 0.8454064148982091, "grad_norm": 3.1998846530914307, "learning_rate": 1.2433628142271637e-06, "loss": 2.5091, "step": 23992 }, { "epoch": 0.8454416519111675, "grad_norm": 1.1686195135116577, "learning_rate": 1.242811724296823e-06, "loss": 0.7482, "step": 23993 }, { "epoch": 0.8454768889241259, "grad_norm": 4.845870018005371, "learning_rate": 1.2422607484299775e-06, "loss": 3.14, "step": 23994 }, { "epoch": 0.8455121259370844, "grad_norm": 0.9629552960395813, "learning_rate": 1.241709886633803e-06, "loss": 1.0078, "step": 23995 }, { "epoch": 0.8455473629500427, "grad_norm": 2.3739659786224365, "learning_rate": 1.241159138915471e-06, "loss": 2.5294, "step": 23996 }, { "epoch": 0.8455825999630011, "grad_norm": 4.180646896362305, "learning_rate": 1.2406085052821537e-06, "loss": 2.8146, "step": 23997 }, { "epoch": 0.8456178369759596, "grad_norm": 5.9012250900268555, "learning_rate": 1.2400579857410322e-06, "loss": 5.2992, "step": 23998 }, { "epoch": 0.845653073988918, "grad_norm": 4.4221906661987305, "learning_rate": 1.2395075802992685e-06, "loss": 3.073, "step": 23999 }, { "epoch": 0.8456883110018764, "grad_norm": 0.8683682084083557, "learning_rate": 1.2389572889640344e-06, "loss": 0.988, "step": 24000 }, { "epoch": 0.8457235480148347, "grad_norm": 3.1014909744262695, "learning_rate": 1.2384071117424978e-06, "loss": 2.7989, "step": 24001 }, { "epoch": 0.8457587850277932, "grad_norm": 6.191281795501709, "learning_rate": 1.237857048641824e-06, "loss": 5.0681, "step": 24002 }, { "epoch": 0.8457940220407516, "grad_norm": 1.6772969961166382, "learning_rate": 1.2373070996691783e-06, "loss": 0.8419, "step": 24003 }, { "epoch": 0.84582925905371, "grad_norm": 5.003681659698486, "learning_rate": 1.2367572648317249e-06, "loss": 6.5357, "step": 24004 }, { "epoch": 0.8458644960666685, "grad_norm": 5.197945594787598, "learning_rate": 1.2362075441366238e-06, "loss": 6.835, "step": 24005 }, { "epoch": 0.8458997330796268, "grad_norm": 7.906796932220459, "learning_rate": 1.2356579375910349e-06, "loss": 5.2685, "step": 24006 }, { "epoch": 0.8459349700925852, "grad_norm": 5.5286054611206055, "learning_rate": 1.2351084452021212e-06, "loss": 5.1068, "step": 24007 }, { "epoch": 0.8459702071055437, "grad_norm": 12.672111511230469, "learning_rate": 1.2345590669770313e-06, "loss": 3.6805, "step": 24008 }, { "epoch": 0.8460054441185021, "grad_norm": 3.8628721237182617, "learning_rate": 1.2340098029229285e-06, "loss": 3.3486, "step": 24009 }, { "epoch": 0.8460406811314605, "grad_norm": 5.016303062438965, "learning_rate": 1.233460653046965e-06, "loss": 3.0509, "step": 24010 }, { "epoch": 0.8460759181444188, "grad_norm": 1.021606683731079, "learning_rate": 1.2329116173562949e-06, "loss": 0.7229, "step": 24011 }, { "epoch": 0.8461111551573773, "grad_norm": 0.9796959161758423, "learning_rate": 1.2323626958580636e-06, "loss": 0.8456, "step": 24012 }, { "epoch": 0.8461463921703357, "grad_norm": 1.9215165376663208, "learning_rate": 1.2318138885594278e-06, "loss": 0.8799, "step": 24013 }, { "epoch": 0.8461816291832941, "grad_norm": 3.5906283855438232, "learning_rate": 1.2312651954675336e-06, "loss": 2.7767, "step": 24014 }, { "epoch": 0.8462168661962526, "grad_norm": 5.193286418914795, "learning_rate": 1.2307166165895257e-06, "loss": 2.7676, "step": 24015 }, { "epoch": 0.846252103209211, "grad_norm": 1.5593329668045044, "learning_rate": 1.2301681519325493e-06, "loss": 0.8046, "step": 24016 }, { "epoch": 0.8462873402221693, "grad_norm": 3.666975736618042, "learning_rate": 1.2296198015037542e-06, "loss": 2.9334, "step": 24017 }, { "epoch": 0.8463225772351278, "grad_norm": 4.776176452636719, "learning_rate": 1.2290715653102759e-06, "loss": 4.6492, "step": 24018 }, { "epoch": 0.8463578142480862, "grad_norm": 1.2861677408218384, "learning_rate": 1.2285234433592574e-06, "loss": 0.6856, "step": 24019 }, { "epoch": 0.8463930512610446, "grad_norm": 2.9533803462982178, "learning_rate": 1.227975435657839e-06, "loss": 2.8644, "step": 24020 }, { "epoch": 0.846428288274003, "grad_norm": 3.163360834121704, "learning_rate": 1.2274275422131565e-06, "loss": 2.447, "step": 24021 }, { "epoch": 0.8464635252869614, "grad_norm": 1.6641947031021118, "learning_rate": 1.2268797630323493e-06, "loss": 0.7603, "step": 24022 }, { "epoch": 0.8464987622999198, "grad_norm": 1.2424384355545044, "learning_rate": 1.2263320981225502e-06, "loss": 0.7056, "step": 24023 }, { "epoch": 0.8465339993128782, "grad_norm": 1.9264639616012573, "learning_rate": 1.2257845474908926e-06, "loss": 0.6862, "step": 24024 }, { "epoch": 0.8465692363258367, "grad_norm": 5.2088141441345215, "learning_rate": 1.2252371111445093e-06, "loss": 2.8159, "step": 24025 }, { "epoch": 0.8466044733387951, "grad_norm": 4.170497894287109, "learning_rate": 1.224689789090533e-06, "loss": 2.9604, "step": 24026 }, { "epoch": 0.8466397103517534, "grad_norm": 0.9602330327033997, "learning_rate": 1.2241425813360851e-06, "loss": 0.784, "step": 24027 }, { "epoch": 0.8466749473647119, "grad_norm": 3.1818559169769287, "learning_rate": 1.2235954878882994e-06, "loss": 2.8418, "step": 24028 }, { "epoch": 0.8467101843776703, "grad_norm": 1.220198631286621, "learning_rate": 1.223048508754302e-06, "loss": 0.7036, "step": 24029 }, { "epoch": 0.8467454213906287, "grad_norm": 1.0370814800262451, "learning_rate": 1.2225016439412163e-06, "loss": 0.7961, "step": 24030 }, { "epoch": 0.8467806584035872, "grad_norm": 5.242066383361816, "learning_rate": 1.221954893456162e-06, "loss": 5.462, "step": 24031 }, { "epoch": 0.8468158954165456, "grad_norm": 5.191157341003418, "learning_rate": 1.2214082573062647e-06, "loss": 4.7949, "step": 24032 }, { "epoch": 0.8468511324295039, "grad_norm": 1.0154110193252563, "learning_rate": 1.2208617354986452e-06, "loss": 0.851, "step": 24033 }, { "epoch": 0.8468863694424623, "grad_norm": 2.8613924980163574, "learning_rate": 1.2203153280404178e-06, "loss": 1.97, "step": 24034 }, { "epoch": 0.8469216064554208, "grad_norm": 1.2171992063522339, "learning_rate": 1.2197690349386993e-06, "loss": 0.9777, "step": 24035 }, { "epoch": 0.8469568434683792, "grad_norm": 3.258249282836914, "learning_rate": 1.2192228562006126e-06, "loss": 2.7028, "step": 24036 }, { "epoch": 0.8469920804813376, "grad_norm": 0.937326967716217, "learning_rate": 1.2186767918332632e-06, "loss": 0.611, "step": 24037 }, { "epoch": 0.847027317494296, "grad_norm": 1.2899266481399536, "learning_rate": 1.2181308418437687e-06, "loss": 1.0452, "step": 24038 }, { "epoch": 0.8470625545072544, "grad_norm": 2.5009076595306396, "learning_rate": 1.2175850062392381e-06, "loss": 2.27, "step": 24039 }, { "epoch": 0.8470977915202128, "grad_norm": 2.9485015869140625, "learning_rate": 1.217039285026781e-06, "loss": 2.9934, "step": 24040 }, { "epoch": 0.8471330285331713, "grad_norm": 2.979623556137085, "learning_rate": 1.2164936782135062e-06, "loss": 2.7754, "step": 24041 }, { "epoch": 0.8471682655461297, "grad_norm": 2.111074447631836, "learning_rate": 1.2159481858065203e-06, "loss": 2.6767, "step": 24042 }, { "epoch": 0.847203502559088, "grad_norm": 1.2813621759414673, "learning_rate": 1.2154028078129275e-06, "loss": 1.0574, "step": 24043 }, { "epoch": 0.8472387395720464, "grad_norm": 3.77193284034729, "learning_rate": 1.2148575442398325e-06, "loss": 4.7279, "step": 24044 }, { "epoch": 0.8472739765850049, "grad_norm": 5.400562286376953, "learning_rate": 1.214312395094337e-06, "loss": 3.1511, "step": 24045 }, { "epoch": 0.8473092135979633, "grad_norm": 0.9865167737007141, "learning_rate": 1.2137673603835421e-06, "loss": 0.6957, "step": 24046 }, { "epoch": 0.8473444506109217, "grad_norm": 1.3033732175827026, "learning_rate": 1.2132224401145466e-06, "loss": 0.9063, "step": 24047 }, { "epoch": 0.8473796876238802, "grad_norm": 1.0176312923431396, "learning_rate": 1.212677634294448e-06, "loss": 0.7008, "step": 24048 }, { "epoch": 0.8474149246368385, "grad_norm": 1.6743226051330566, "learning_rate": 1.2121329429303442e-06, "loss": 0.9035, "step": 24049 }, { "epoch": 0.8474501616497969, "grad_norm": 2.1263349056243896, "learning_rate": 1.2115883660293238e-06, "loss": 2.4885, "step": 24050 }, { "epoch": 0.8474853986627554, "grad_norm": 2.996751546859741, "learning_rate": 1.2110439035984866e-06, "loss": 2.99, "step": 24051 }, { "epoch": 0.8475206356757138, "grad_norm": 1.1244996786117554, "learning_rate": 1.210499555644925e-06, "loss": 0.7707, "step": 24052 }, { "epoch": 0.8475558726886722, "grad_norm": 2.2842798233032227, "learning_rate": 1.2099553221757231e-06, "loss": 2.5322, "step": 24053 }, { "epoch": 0.8475911097016305, "grad_norm": 1.1325680017471313, "learning_rate": 1.209411203197971e-06, "loss": 0.7676, "step": 24054 }, { "epoch": 0.847626346714589, "grad_norm": 12.679577827453613, "learning_rate": 1.2088671987187627e-06, "loss": 5.0751, "step": 24055 }, { "epoch": 0.8476615837275474, "grad_norm": 3.9370951652526855, "learning_rate": 1.2083233087451762e-06, "loss": 4.7266, "step": 24056 }, { "epoch": 0.8476968207405058, "grad_norm": 2.2993412017822266, "learning_rate": 1.207779533284299e-06, "loss": 0.6344, "step": 24057 }, { "epoch": 0.8477320577534643, "grad_norm": 4.444892883300781, "learning_rate": 1.2072358723432133e-06, "loss": 2.5268, "step": 24058 }, { "epoch": 0.8477672947664227, "grad_norm": 1.7986127138137817, "learning_rate": 1.2066923259289997e-06, "loss": 1.0506, "step": 24059 }, { "epoch": 0.847802531779381, "grad_norm": 1.3635307550430298, "learning_rate": 1.2061488940487386e-06, "loss": 0.7414, "step": 24060 }, { "epoch": 0.8478377687923395, "grad_norm": 5.438211441040039, "learning_rate": 1.2056055767095087e-06, "loss": 3.1168, "step": 24061 }, { "epoch": 0.8478730058052979, "grad_norm": 4.984470844268799, "learning_rate": 1.2050623739183865e-06, "loss": 2.4086, "step": 24062 }, { "epoch": 0.8479082428182563, "grad_norm": 1.1180671453475952, "learning_rate": 1.2045192856824473e-06, "loss": 0.8062, "step": 24063 }, { "epoch": 0.8479434798312147, "grad_norm": 2.6276941299438477, "learning_rate": 1.2039763120087643e-06, "loss": 2.5813, "step": 24064 }, { "epoch": 0.8479787168441731, "grad_norm": 1.124821662902832, "learning_rate": 1.2034334529044101e-06, "loss": 0.9582, "step": 24065 }, { "epoch": 0.8480139538571315, "grad_norm": 3.154052495956421, "learning_rate": 1.2028907083764563e-06, "loss": 2.6304, "step": 24066 }, { "epoch": 0.8480491908700899, "grad_norm": 1.572766661643982, "learning_rate": 1.2023480784319718e-06, "loss": 1.144, "step": 24067 }, { "epoch": 0.8480844278830484, "grad_norm": 2.932323455810547, "learning_rate": 1.2018055630780257e-06, "loss": 0.9551, "step": 24068 }, { "epoch": 0.8481196648960068, "grad_norm": 1.3842852115631104, "learning_rate": 1.2012631623216786e-06, "loss": 0.9132, "step": 24069 }, { "epoch": 0.8481549019089651, "grad_norm": 15.828237533569336, "learning_rate": 1.2007208761700018e-06, "loss": 2.818, "step": 24070 }, { "epoch": 0.8481901389219236, "grad_norm": 1.2600477933883667, "learning_rate": 1.2001787046300585e-06, "loss": 0.8156, "step": 24071 }, { "epoch": 0.848225375934882, "grad_norm": 1.3659368753433228, "learning_rate": 1.1996366477089073e-06, "loss": 0.7119, "step": 24072 }, { "epoch": 0.8482606129478404, "grad_norm": 2.967383861541748, "learning_rate": 1.199094705413606e-06, "loss": 2.7603, "step": 24073 }, { "epoch": 0.8482958499607989, "grad_norm": 4.04658842086792, "learning_rate": 1.1985528777512224e-06, "loss": 5.2299, "step": 24074 }, { "epoch": 0.8483310869737573, "grad_norm": 2.5678436756134033, "learning_rate": 1.1980111647288073e-06, "loss": 2.9515, "step": 24075 }, { "epoch": 0.8483663239867156, "grad_norm": 1.0844073295593262, "learning_rate": 1.1974695663534163e-06, "loss": 0.7948, "step": 24076 }, { "epoch": 0.848401560999674, "grad_norm": 3.9142088890075684, "learning_rate": 1.1969280826321072e-06, "loss": 3.0265, "step": 24077 }, { "epoch": 0.8484367980126325, "grad_norm": 1.4492614269256592, "learning_rate": 1.1963867135719299e-06, "loss": 1.0133, "step": 24078 }, { "epoch": 0.8484720350255909, "grad_norm": 1.3164026737213135, "learning_rate": 1.1958454591799374e-06, "loss": 1.1713, "step": 24079 }, { "epoch": 0.8485072720385493, "grad_norm": 5.7605299949646, "learning_rate": 1.1953043194631798e-06, "loss": 7.8916, "step": 24080 }, { "epoch": 0.8485425090515077, "grad_norm": 1.013218641281128, "learning_rate": 1.1947632944287046e-06, "loss": 0.8101, "step": 24081 }, { "epoch": 0.8485777460644661, "grad_norm": 1.275620937347412, "learning_rate": 1.1942223840835588e-06, "loss": 0.9993, "step": 24082 }, { "epoch": 0.8486129830774245, "grad_norm": 5.794699192047119, "learning_rate": 1.1936815884347886e-06, "loss": 3.0812, "step": 24083 }, { "epoch": 0.848648220090383, "grad_norm": 1.2654662132263184, "learning_rate": 1.1931409074894363e-06, "loss": 0.7696, "step": 24084 }, { "epoch": 0.8486834571033414, "grad_norm": 1.1206607818603516, "learning_rate": 1.1926003412545473e-06, "loss": 0.925, "step": 24085 }, { "epoch": 0.8487186941162997, "grad_norm": 1.2716541290283203, "learning_rate": 1.1920598897371593e-06, "loss": 0.5546, "step": 24086 }, { "epoch": 0.8487539311292581, "grad_norm": 1.7699259519577026, "learning_rate": 1.1915195529443157e-06, "loss": 1.1732, "step": 24087 }, { "epoch": 0.8487891681422166, "grad_norm": 1.0830894708633423, "learning_rate": 1.1909793308830476e-06, "loss": 0.7132, "step": 24088 }, { "epoch": 0.848824405155175, "grad_norm": 2.0122437477111816, "learning_rate": 1.190439223560399e-06, "loss": 0.8679, "step": 24089 }, { "epoch": 0.8488596421681334, "grad_norm": 3.0054686069488525, "learning_rate": 1.1898992309834023e-06, "loss": 2.7685, "step": 24090 }, { "epoch": 0.8488948791810919, "grad_norm": 3.750714063644409, "learning_rate": 1.1893593531590885e-06, "loss": 2.7184, "step": 24091 }, { "epoch": 0.8489301161940502, "grad_norm": 1.225225806236267, "learning_rate": 1.18881959009449e-06, "loss": 0.9253, "step": 24092 }, { "epoch": 0.8489653532070086, "grad_norm": 2.9987354278564453, "learning_rate": 1.1882799417966428e-06, "loss": 2.5856, "step": 24093 }, { "epoch": 0.8490005902199671, "grad_norm": 2.270188331604004, "learning_rate": 1.1877404082725696e-06, "loss": 0.8896, "step": 24094 }, { "epoch": 0.8490358272329255, "grad_norm": 5.921977519989014, "learning_rate": 1.1872009895293002e-06, "loss": 5.2913, "step": 24095 }, { "epoch": 0.8490710642458839, "grad_norm": 0.9732044339179993, "learning_rate": 1.18666168557386e-06, "loss": 1.0808, "step": 24096 }, { "epoch": 0.8491063012588422, "grad_norm": 3.120310068130493, "learning_rate": 1.1861224964132744e-06, "loss": 2.7356, "step": 24097 }, { "epoch": 0.8491415382718007, "grad_norm": 3.892179012298584, "learning_rate": 1.1855834220545658e-06, "loss": 3.2088, "step": 24098 }, { "epoch": 0.8491767752847591, "grad_norm": 1.198021411895752, "learning_rate": 1.1850444625047563e-06, "loss": 0.9216, "step": 24099 }, { "epoch": 0.8492120122977175, "grad_norm": 4.181865215301514, "learning_rate": 1.1845056177708648e-06, "loss": 4.3985, "step": 24100 }, { "epoch": 0.849247249310676, "grad_norm": 4.339890956878662, "learning_rate": 1.1839668878599109e-06, "loss": 3.0478, "step": 24101 }, { "epoch": 0.8492824863236343, "grad_norm": 2.17169451713562, "learning_rate": 1.1834282727789114e-06, "loss": 2.1983, "step": 24102 }, { "epoch": 0.8493177233365927, "grad_norm": 1.0287425518035889, "learning_rate": 1.1828897725348809e-06, "loss": 0.7928, "step": 24103 }, { "epoch": 0.8493529603495512, "grad_norm": 1.071250557899475, "learning_rate": 1.1823513871348358e-06, "loss": 0.6905, "step": 24104 }, { "epoch": 0.8493881973625096, "grad_norm": 1.2246731519699097, "learning_rate": 1.181813116585786e-06, "loss": 0.7603, "step": 24105 }, { "epoch": 0.849423434375468, "grad_norm": 7.784976959228516, "learning_rate": 1.181274960894746e-06, "loss": 5.124, "step": 24106 }, { "epoch": 0.8494586713884265, "grad_norm": 1.1080656051635742, "learning_rate": 1.180736920068718e-06, "loss": 0.8585, "step": 24107 }, { "epoch": 0.8494939084013848, "grad_norm": 1.2382237911224365, "learning_rate": 1.1801989941147184e-06, "loss": 0.9306, "step": 24108 }, { "epoch": 0.8495291454143432, "grad_norm": 3.002671957015991, "learning_rate": 1.179661183039752e-06, "loss": 2.7865, "step": 24109 }, { "epoch": 0.8495643824273016, "grad_norm": 2.2415006160736084, "learning_rate": 1.1791234868508195e-06, "loss": 2.4398, "step": 24110 }, { "epoch": 0.8495996194402601, "grad_norm": 3.9989261627197266, "learning_rate": 1.1785859055549254e-06, "loss": 2.8674, "step": 24111 }, { "epoch": 0.8496348564532185, "grad_norm": 2.8400330543518066, "learning_rate": 1.1780484391590786e-06, "loss": 2.8545, "step": 24112 }, { "epoch": 0.8496700934661768, "grad_norm": 2.932812213897705, "learning_rate": 1.1775110876702712e-06, "loss": 2.6841, "step": 24113 }, { "epoch": 0.8497053304791353, "grad_norm": 4.518774032592773, "learning_rate": 1.1769738510955064e-06, "loss": 3.2274, "step": 24114 }, { "epoch": 0.8497405674920937, "grad_norm": 2.8007819652557373, "learning_rate": 1.176436729441781e-06, "loss": 2.5886, "step": 24115 }, { "epoch": 0.8497758045050521, "grad_norm": 1.0640767812728882, "learning_rate": 1.1758997227160906e-06, "loss": 1.1065, "step": 24116 }, { "epoch": 0.8498110415180106, "grad_norm": 1.1496362686157227, "learning_rate": 1.1753628309254316e-06, "loss": 0.9091, "step": 24117 }, { "epoch": 0.849846278530969, "grad_norm": 5.334359645843506, "learning_rate": 1.1748260540767952e-06, "loss": 3.0146, "step": 24118 }, { "epoch": 0.8498815155439273, "grad_norm": 1.7995437383651733, "learning_rate": 1.1742893921771737e-06, "loss": 1.1198, "step": 24119 }, { "epoch": 0.8499167525568857, "grad_norm": 1.2447755336761475, "learning_rate": 1.1737528452335567e-06, "loss": 0.8658, "step": 24120 }, { "epoch": 0.8499519895698442, "grad_norm": 3.1010355949401855, "learning_rate": 1.1732164132529333e-06, "loss": 3.1042, "step": 24121 }, { "epoch": 0.8499872265828026, "grad_norm": 1.1900737285614014, "learning_rate": 1.1726800962422902e-06, "loss": 1.0025, "step": 24122 }, { "epoch": 0.850022463595761, "grad_norm": 1.098388910293579, "learning_rate": 1.1721438942086138e-06, "loss": 1.0414, "step": 24123 }, { "epoch": 0.8500577006087194, "grad_norm": 2.7117369174957275, "learning_rate": 1.1716078071588866e-06, "loss": 0.986, "step": 24124 }, { "epoch": 0.8500929376216778, "grad_norm": 1.2566989660263062, "learning_rate": 1.1710718351000961e-06, "loss": 0.8284, "step": 24125 }, { "epoch": 0.8501281746346362, "grad_norm": 5.245670318603516, "learning_rate": 1.1705359780392133e-06, "loss": 3.2539, "step": 24126 }, { "epoch": 0.8501634116475947, "grad_norm": 2.9001123905181885, "learning_rate": 1.1700002359832275e-06, "loss": 2.297, "step": 24127 }, { "epoch": 0.8501986486605531, "grad_norm": 1.0658279657363892, "learning_rate": 1.169464608939116e-06, "loss": 1.0196, "step": 24128 }, { "epoch": 0.8502338856735114, "grad_norm": 3.735182285308838, "learning_rate": 1.1689290969138501e-06, "loss": 4.3245, "step": 24129 }, { "epoch": 0.8502691226864698, "grad_norm": 3.136798143386841, "learning_rate": 1.1683936999144051e-06, "loss": 2.9908, "step": 24130 }, { "epoch": 0.8503043596994283, "grad_norm": 3.0563290119171143, "learning_rate": 1.1678584179477614e-06, "loss": 2.4746, "step": 24131 }, { "epoch": 0.8503395967123867, "grad_norm": 3.9374852180480957, "learning_rate": 1.1673232510208855e-06, "loss": 4.8031, "step": 24132 }, { "epoch": 0.8503748337253451, "grad_norm": 1.645061731338501, "learning_rate": 1.1667881991407481e-06, "loss": 0.7736, "step": 24133 }, { "epoch": 0.8504100707383035, "grad_norm": 4.26238489151001, "learning_rate": 1.1662532623143208e-06, "loss": 2.96, "step": 24134 }, { "epoch": 0.8504453077512619, "grad_norm": 1.3977018594741821, "learning_rate": 1.16571844054857e-06, "loss": 0.5807, "step": 24135 }, { "epoch": 0.8504805447642203, "grad_norm": 3.9012255668640137, "learning_rate": 1.1651837338504612e-06, "loss": 2.635, "step": 24136 }, { "epoch": 0.8505157817771788, "grad_norm": 1.8535572290420532, "learning_rate": 1.164649142226959e-06, "loss": 0.985, "step": 24137 }, { "epoch": 0.8505510187901372, "grad_norm": 5.029356002807617, "learning_rate": 1.164114665685029e-06, "loss": 2.7929, "step": 24138 }, { "epoch": 0.8505862558030955, "grad_norm": 3.0324039459228516, "learning_rate": 1.1635803042316295e-06, "loss": 2.5499, "step": 24139 }, { "epoch": 0.850621492816054, "grad_norm": 2.6253013610839844, "learning_rate": 1.1630460578737223e-06, "loss": 0.8985, "step": 24140 }, { "epoch": 0.8506567298290124, "grad_norm": 1.5908926725387573, "learning_rate": 1.162511926618266e-06, "loss": 0.7645, "step": 24141 }, { "epoch": 0.8506919668419708, "grad_norm": 3.118121385574341, "learning_rate": 1.1619779104722185e-06, "loss": 2.789, "step": 24142 }, { "epoch": 0.8507272038549292, "grad_norm": 6.420340061187744, "learning_rate": 1.1614440094425327e-06, "loss": 5.3281, "step": 24143 }, { "epoch": 0.8507624408678877, "grad_norm": 1.060033917427063, "learning_rate": 1.160910223536168e-06, "loss": 0.7124, "step": 24144 }, { "epoch": 0.850797677880846, "grad_norm": 4.302529335021973, "learning_rate": 1.1603765527600685e-06, "loss": 3.2474, "step": 24145 }, { "epoch": 0.8508329148938044, "grad_norm": 1.2770825624465942, "learning_rate": 1.1598429971211934e-06, "loss": 0.7904, "step": 24146 }, { "epoch": 0.8508681519067629, "grad_norm": 1.4378491640090942, "learning_rate": 1.1593095566264912e-06, "loss": 0.754, "step": 24147 }, { "epoch": 0.8509033889197213, "grad_norm": 4.608020782470703, "learning_rate": 1.1587762312829055e-06, "loss": 3.1299, "step": 24148 }, { "epoch": 0.8509386259326797, "grad_norm": 1.1109418869018555, "learning_rate": 1.1582430210973839e-06, "loss": 1.0433, "step": 24149 }, { "epoch": 0.8509738629456381, "grad_norm": 1.2640312910079956, "learning_rate": 1.1577099260768775e-06, "loss": 0.9361, "step": 24150 }, { "epoch": 0.8510090999585965, "grad_norm": 1.4610307216644287, "learning_rate": 1.1571769462283234e-06, "loss": 0.7954, "step": 24151 }, { "epoch": 0.8510443369715549, "grad_norm": 1.2265968322753906, "learning_rate": 1.1566440815586654e-06, "loss": 0.7949, "step": 24152 }, { "epoch": 0.8510795739845133, "grad_norm": 1.7731413841247559, "learning_rate": 1.1561113320748452e-06, "loss": 1.0436, "step": 24153 }, { "epoch": 0.8511148109974718, "grad_norm": 4.496140003204346, "learning_rate": 1.1555786977838025e-06, "loss": 2.8137, "step": 24154 }, { "epoch": 0.8511500480104301, "grad_norm": 2.3297269344329834, "learning_rate": 1.1550461786924727e-06, "loss": 0.8892, "step": 24155 }, { "epoch": 0.8511852850233885, "grad_norm": 1.2636178731918335, "learning_rate": 1.1545137748077928e-06, "loss": 1.1945, "step": 24156 }, { "epoch": 0.851220522036347, "grad_norm": 0.9963608384132385, "learning_rate": 1.1539814861366983e-06, "loss": 0.7967, "step": 24157 }, { "epoch": 0.8512557590493054, "grad_norm": 2.9916281700134277, "learning_rate": 1.1534493126861212e-06, "loss": 3.2738, "step": 24158 }, { "epoch": 0.8512909960622638, "grad_norm": 6.704701900482178, "learning_rate": 1.152917254462994e-06, "loss": 5.236, "step": 24159 }, { "epoch": 0.8513262330752223, "grad_norm": 1.2765275239944458, "learning_rate": 1.1523853114742467e-06, "loss": 0.8959, "step": 24160 }, { "epoch": 0.8513614700881806, "grad_norm": 1.2382538318634033, "learning_rate": 1.1518534837268081e-06, "loss": 0.8593, "step": 24161 }, { "epoch": 0.851396707101139, "grad_norm": 3.6888861656188965, "learning_rate": 1.1513217712276037e-06, "loss": 3.575, "step": 24162 }, { "epoch": 0.8514319441140974, "grad_norm": 1.0149790048599243, "learning_rate": 1.1507901739835635e-06, "loss": 1.0997, "step": 24163 }, { "epoch": 0.8514671811270559, "grad_norm": 1.1453444957733154, "learning_rate": 1.1502586920016046e-06, "loss": 0.609, "step": 24164 }, { "epoch": 0.8515024181400143, "grad_norm": 3.251833438873291, "learning_rate": 1.1497273252886543e-06, "loss": 2.7714, "step": 24165 }, { "epoch": 0.8515376551529726, "grad_norm": 1.2297242879867554, "learning_rate": 1.1491960738516361e-06, "loss": 1.0551, "step": 24166 }, { "epoch": 0.8515728921659311, "grad_norm": 1.8886573314666748, "learning_rate": 1.1486649376974635e-06, "loss": 0.8262, "step": 24167 }, { "epoch": 0.8516081291788895, "grad_norm": 2.6342055797576904, "learning_rate": 1.1481339168330563e-06, "loss": 3.007, "step": 24168 }, { "epoch": 0.8516433661918479, "grad_norm": 1.0218915939331055, "learning_rate": 1.1476030112653359e-06, "loss": 0.6127, "step": 24169 }, { "epoch": 0.8516786032048064, "grad_norm": 0.9213566780090332, "learning_rate": 1.147072221001212e-06, "loss": 0.8977, "step": 24170 }, { "epoch": 0.8517138402177648, "grad_norm": 3.073261022567749, "learning_rate": 1.1465415460476004e-06, "loss": 2.9861, "step": 24171 }, { "epoch": 0.8517490772307231, "grad_norm": 4.559511661529541, "learning_rate": 1.1460109864114134e-06, "loss": 5.1898, "step": 24172 }, { "epoch": 0.8517843142436816, "grad_norm": 0.9830952286720276, "learning_rate": 1.1454805420995596e-06, "loss": 0.831, "step": 24173 }, { "epoch": 0.85181955125664, "grad_norm": 3.5277299880981445, "learning_rate": 1.1449502131189505e-06, "loss": 3.0505, "step": 24174 }, { "epoch": 0.8518547882695984, "grad_norm": 1.6110225915908813, "learning_rate": 1.1444199994764937e-06, "loss": 0.8128, "step": 24175 }, { "epoch": 0.8518900252825568, "grad_norm": 3.1661698818206787, "learning_rate": 1.1438899011790927e-06, "loss": 4.5155, "step": 24176 }, { "epoch": 0.8519252622955152, "grad_norm": 2.650693416595459, "learning_rate": 1.1433599182336553e-06, "loss": 2.8775, "step": 24177 }, { "epoch": 0.8519604993084736, "grad_norm": 4.242926597595215, "learning_rate": 1.1428300506470824e-06, "loss": 2.9994, "step": 24178 }, { "epoch": 0.851995736321432, "grad_norm": 4.025580406188965, "learning_rate": 1.1423002984262766e-06, "loss": 3.2676, "step": 24179 }, { "epoch": 0.8520309733343905, "grad_norm": 1.0939955711364746, "learning_rate": 1.1417706615781377e-06, "loss": 0.7792, "step": 24180 }, { "epoch": 0.8520662103473489, "grad_norm": 0.9661078453063965, "learning_rate": 1.1412411401095636e-06, "loss": 1.1232, "step": 24181 }, { "epoch": 0.8521014473603072, "grad_norm": 7.636331081390381, "learning_rate": 1.1407117340274542e-06, "loss": 4.9807, "step": 24182 }, { "epoch": 0.8521366843732657, "grad_norm": 3.6473116874694824, "learning_rate": 1.1401824433387e-06, "loss": 4.4233, "step": 24183 }, { "epoch": 0.8521719213862241, "grad_norm": 4.316575527191162, "learning_rate": 1.1396532680501994e-06, "loss": 4.0531, "step": 24184 }, { "epoch": 0.8522071583991825, "grad_norm": 5.485607624053955, "learning_rate": 1.1391242081688458e-06, "loss": 2.5029, "step": 24185 }, { "epoch": 0.8522423954121409, "grad_norm": 0.8183934092521667, "learning_rate": 1.1385952637015252e-06, "loss": 0.7478, "step": 24186 }, { "epoch": 0.8522776324250994, "grad_norm": 1.9883989095687866, "learning_rate": 1.1380664346551295e-06, "loss": 0.8536, "step": 24187 }, { "epoch": 0.8523128694380577, "grad_norm": 0.9577586054801941, "learning_rate": 1.1375377210365502e-06, "loss": 1.2914, "step": 24188 }, { "epoch": 0.8523481064510161, "grad_norm": 4.661844730377197, "learning_rate": 1.1370091228526692e-06, "loss": 4.8662, "step": 24189 }, { "epoch": 0.8523833434639746, "grad_norm": 3.8492071628570557, "learning_rate": 1.1364806401103734e-06, "loss": 2.9816, "step": 24190 }, { "epoch": 0.852418580476933, "grad_norm": 1.1046555042266846, "learning_rate": 1.1359522728165451e-06, "loss": 0.8534, "step": 24191 }, { "epoch": 0.8524538174898914, "grad_norm": 2.7040398120880127, "learning_rate": 1.1354240209780699e-06, "loss": 2.6803, "step": 24192 }, { "epoch": 0.8524890545028498, "grad_norm": 1.1717959642410278, "learning_rate": 1.1348958846018254e-06, "loss": 0.8758, "step": 24193 }, { "epoch": 0.8525242915158082, "grad_norm": 1.0718152523040771, "learning_rate": 1.1343678636946908e-06, "loss": 0.9779, "step": 24194 }, { "epoch": 0.8525595285287666, "grad_norm": 3.3593599796295166, "learning_rate": 1.133839958263544e-06, "loss": 2.8015, "step": 24195 }, { "epoch": 0.852594765541725, "grad_norm": 1.7318382263183594, "learning_rate": 1.1333121683152615e-06, "loss": 1.0183, "step": 24196 }, { "epoch": 0.8526300025546835, "grad_norm": 5.102470397949219, "learning_rate": 1.1327844938567168e-06, "loss": 4.71, "step": 24197 }, { "epoch": 0.8526652395676418, "grad_norm": 2.5210273265838623, "learning_rate": 1.1322569348947832e-06, "loss": 2.6179, "step": 24198 }, { "epoch": 0.8527004765806002, "grad_norm": 1.1657480001449585, "learning_rate": 1.1317294914363342e-06, "loss": 1.1157, "step": 24199 }, { "epoch": 0.8527357135935587, "grad_norm": 2.9743807315826416, "learning_rate": 1.1312021634882363e-06, "loss": 1.0625, "step": 24200 }, { "epoch": 0.8527709506065171, "grad_norm": 4.047652721405029, "learning_rate": 1.1306749510573633e-06, "loss": 2.8317, "step": 24201 }, { "epoch": 0.8528061876194755, "grad_norm": 2.731458902359009, "learning_rate": 1.130147854150575e-06, "loss": 2.3045, "step": 24202 }, { "epoch": 0.852841424632434, "grad_norm": 1.1042027473449707, "learning_rate": 1.1296208727747415e-06, "loss": 0.7664, "step": 24203 }, { "epoch": 0.8528766616453923, "grad_norm": 1.3866150379180908, "learning_rate": 1.1290940069367284e-06, "loss": 0.781, "step": 24204 }, { "epoch": 0.8529118986583507, "grad_norm": 1.274537444114685, "learning_rate": 1.1285672566433947e-06, "loss": 1.0037, "step": 24205 }, { "epoch": 0.8529471356713092, "grad_norm": 1.0377042293548584, "learning_rate": 1.1280406219015993e-06, "loss": 0.9291, "step": 24206 }, { "epoch": 0.8529823726842676, "grad_norm": 4.201216220855713, "learning_rate": 1.1275141027182102e-06, "loss": 2.7095, "step": 24207 }, { "epoch": 0.853017609697226, "grad_norm": 3.7161452770233154, "learning_rate": 1.1269876991000773e-06, "loss": 2.5268, "step": 24208 }, { "epoch": 0.8530528467101843, "grad_norm": 0.9747800230979919, "learning_rate": 1.1264614110540606e-06, "loss": 1.2644, "step": 24209 }, { "epoch": 0.8530880837231428, "grad_norm": 3.368093490600586, "learning_rate": 1.1259352385870114e-06, "loss": 2.9198, "step": 24210 }, { "epoch": 0.8531233207361012, "grad_norm": 1.7086079120635986, "learning_rate": 1.12540918170579e-06, "loss": 0.9878, "step": 24211 }, { "epoch": 0.8531585577490596, "grad_norm": 1.018186092376709, "learning_rate": 1.1248832404172427e-06, "loss": 1.0004, "step": 24212 }, { "epoch": 0.8531937947620181, "grad_norm": 3.4956986904144287, "learning_rate": 1.124357414728222e-06, "loss": 2.817, "step": 24213 }, { "epoch": 0.8532290317749764, "grad_norm": 1.1576383113861084, "learning_rate": 1.1238317046455771e-06, "loss": 0.9137, "step": 24214 }, { "epoch": 0.8532642687879348, "grad_norm": 2.9576172828674316, "learning_rate": 1.1233061101761545e-06, "loss": 2.6149, "step": 24215 }, { "epoch": 0.8532995058008933, "grad_norm": 5.730461597442627, "learning_rate": 1.1227806313267997e-06, "loss": 4.9262, "step": 24216 }, { "epoch": 0.8533347428138517, "grad_norm": 1.8275983333587646, "learning_rate": 1.1222552681043587e-06, "loss": 0.8003, "step": 24217 }, { "epoch": 0.8533699798268101, "grad_norm": 0.9904049634933472, "learning_rate": 1.1217300205156733e-06, "loss": 0.7271, "step": 24218 }, { "epoch": 0.8534052168397684, "grad_norm": 2.710801124572754, "learning_rate": 1.1212048885675852e-06, "loss": 2.8272, "step": 24219 }, { "epoch": 0.8534404538527269, "grad_norm": 1.4856834411621094, "learning_rate": 1.1206798722669365e-06, "loss": 0.9894, "step": 24220 }, { "epoch": 0.8534756908656853, "grad_norm": 1.2742502689361572, "learning_rate": 1.1201549716205606e-06, "loss": 0.6958, "step": 24221 }, { "epoch": 0.8535109278786437, "grad_norm": 1.0393273830413818, "learning_rate": 1.1196301866352976e-06, "loss": 0.9407, "step": 24222 }, { "epoch": 0.8535461648916022, "grad_norm": 3.9456539154052734, "learning_rate": 1.1191055173179855e-06, "loss": 2.6647, "step": 24223 }, { "epoch": 0.8535814019045606, "grad_norm": 1.1108638048171997, "learning_rate": 1.118580963675453e-06, "loss": 0.7771, "step": 24224 }, { "epoch": 0.8536166389175189, "grad_norm": 4.104806900024414, "learning_rate": 1.1180565257145338e-06, "loss": 3.1354, "step": 24225 }, { "epoch": 0.8536518759304774, "grad_norm": 3.8409512042999268, "learning_rate": 1.1175322034420622e-06, "loss": 2.7319, "step": 24226 }, { "epoch": 0.8536871129434358, "grad_norm": 1.0468623638153076, "learning_rate": 1.117007996864864e-06, "loss": 1.0269, "step": 24227 }, { "epoch": 0.8537223499563942, "grad_norm": 3.749945640563965, "learning_rate": 1.1164839059897681e-06, "loss": 3.0527, "step": 24228 }, { "epoch": 0.8537575869693526, "grad_norm": 1.1393407583236694, "learning_rate": 1.1159599308236003e-06, "loss": 0.8299, "step": 24229 }, { "epoch": 0.853792823982311, "grad_norm": 5.108153343200684, "learning_rate": 1.115436071373187e-06, "loss": 4.9771, "step": 24230 }, { "epoch": 0.8538280609952694, "grad_norm": 5.006130218505859, "learning_rate": 1.11491232764535e-06, "loss": 2.8173, "step": 24231 }, { "epoch": 0.8538632980082278, "grad_norm": 1.2446637153625488, "learning_rate": 1.114388699646911e-06, "loss": 0.9355, "step": 24232 }, { "epoch": 0.8538985350211863, "grad_norm": 0.7797626256942749, "learning_rate": 1.1138651873846928e-06, "loss": 0.7442, "step": 24233 }, { "epoch": 0.8539337720341447, "grad_norm": 5.11806583404541, "learning_rate": 1.113341790865511e-06, "loss": 4.4214, "step": 24234 }, { "epoch": 0.853969009047103, "grad_norm": 1.3496836423873901, "learning_rate": 1.1128185100961852e-06, "loss": 0.9058, "step": 24235 }, { "epoch": 0.8540042460600615, "grad_norm": 5.312305450439453, "learning_rate": 1.1122953450835294e-06, "loss": 4.7784, "step": 24236 }, { "epoch": 0.8540394830730199, "grad_norm": 1.5914762020111084, "learning_rate": 1.111772295834359e-06, "loss": 0.802, "step": 24237 }, { "epoch": 0.8540747200859783, "grad_norm": 6.684813022613525, "learning_rate": 1.1112493623554876e-06, "loss": 3.2155, "step": 24238 }, { "epoch": 0.8541099570989368, "grad_norm": 2.9559311866760254, "learning_rate": 1.1107265446537275e-06, "loss": 2.6951, "step": 24239 }, { "epoch": 0.8541451941118952, "grad_norm": 1.1210302114486694, "learning_rate": 1.110203842735883e-06, "loss": 0.8911, "step": 24240 }, { "epoch": 0.8541804311248535, "grad_norm": 4.470909118652344, "learning_rate": 1.1096812566087644e-06, "loss": 3.2552, "step": 24241 }, { "epoch": 0.8542156681378119, "grad_norm": 5.676196575164795, "learning_rate": 1.1091587862791853e-06, "loss": 3.4896, "step": 24242 }, { "epoch": 0.8542509051507704, "grad_norm": 3.806643009185791, "learning_rate": 1.108636431753942e-06, "loss": 2.8211, "step": 24243 }, { "epoch": 0.8542861421637288, "grad_norm": 3.628687858581543, "learning_rate": 1.108114193039842e-06, "loss": 3.2092, "step": 24244 }, { "epoch": 0.8543213791766872, "grad_norm": 1.2563778162002563, "learning_rate": 1.107592070143687e-06, "loss": 0.8684, "step": 24245 }, { "epoch": 0.8543566161896456, "grad_norm": 1.348413109779358, "learning_rate": 1.1070700630722787e-06, "loss": 0.9334, "step": 24246 }, { "epoch": 0.854391853202604, "grad_norm": 0.9335436224937439, "learning_rate": 1.1065481718324155e-06, "loss": 0.594, "step": 24247 }, { "epoch": 0.8544270902155624, "grad_norm": 3.5135908126831055, "learning_rate": 1.1060263964308949e-06, "loss": 2.9613, "step": 24248 }, { "epoch": 0.8544623272285209, "grad_norm": 3.21321439743042, "learning_rate": 1.1055047368745142e-06, "loss": 2.7725, "step": 24249 }, { "epoch": 0.8544975642414793, "grad_norm": 1.377353549003601, "learning_rate": 1.1049831931700661e-06, "loss": 0.6445, "step": 24250 }, { "epoch": 0.8545328012544376, "grad_norm": 3.231821060180664, "learning_rate": 1.1044617653243461e-06, "loss": 2.4162, "step": 24251 }, { "epoch": 0.854568038267396, "grad_norm": 0.9008983373641968, "learning_rate": 1.1039404533441434e-06, "loss": 0.7287, "step": 24252 }, { "epoch": 0.8546032752803545, "grad_norm": 3.8012897968292236, "learning_rate": 1.1034192572362502e-06, "loss": 3.0438, "step": 24253 }, { "epoch": 0.8546385122933129, "grad_norm": 1.325361967086792, "learning_rate": 1.1028981770074553e-06, "loss": 0.7079, "step": 24254 }, { "epoch": 0.8546737493062713, "grad_norm": 1.0254923105239868, "learning_rate": 1.102377212664546e-06, "loss": 1.1978, "step": 24255 }, { "epoch": 0.8547089863192298, "grad_norm": 1.0210235118865967, "learning_rate": 1.1018563642143022e-06, "loss": 0.9008, "step": 24256 }, { "epoch": 0.8547442233321881, "grad_norm": 0.8930448293685913, "learning_rate": 1.101335631663516e-06, "loss": 0.969, "step": 24257 }, { "epoch": 0.8547794603451465, "grad_norm": 4.2748284339904785, "learning_rate": 1.1008150150189679e-06, "loss": 4.6797, "step": 24258 }, { "epoch": 0.854814697358105, "grad_norm": 1.175075888633728, "learning_rate": 1.1002945142874367e-06, "loss": 0.9624, "step": 24259 }, { "epoch": 0.8548499343710634, "grad_norm": 1.5137752294540405, "learning_rate": 1.0997741294757004e-06, "loss": 0.8257, "step": 24260 }, { "epoch": 0.8548851713840218, "grad_norm": 2.3865272998809814, "learning_rate": 1.0992538605905435e-06, "loss": 2.4569, "step": 24261 }, { "epoch": 0.8549204083969801, "grad_norm": 1.5706701278686523, "learning_rate": 1.0987337076387371e-06, "loss": 1.0812, "step": 24262 }, { "epoch": 0.8549556454099386, "grad_norm": 4.713156223297119, "learning_rate": 1.0982136706270584e-06, "loss": 2.6005, "step": 24263 }, { "epoch": 0.854990882422897, "grad_norm": 2.8101119995117188, "learning_rate": 1.0976937495622774e-06, "loss": 0.8515, "step": 24264 }, { "epoch": 0.8550261194358554, "grad_norm": 1.2909250259399414, "learning_rate": 1.0971739444511742e-06, "loss": 0.8248, "step": 24265 }, { "epoch": 0.8550613564488139, "grad_norm": 5.377814292907715, "learning_rate": 1.0966542553005122e-06, "loss": 3.1201, "step": 24266 }, { "epoch": 0.8550965934617722, "grad_norm": 5.400846004486084, "learning_rate": 1.0961346821170627e-06, "loss": 5.1346, "step": 24267 }, { "epoch": 0.8551318304747306, "grad_norm": 3.495216131210327, "learning_rate": 1.0956152249075936e-06, "loss": 2.5262, "step": 24268 }, { "epoch": 0.8551670674876891, "grad_norm": 8.143250465393066, "learning_rate": 1.0950958836788705e-06, "loss": 2.1153, "step": 24269 }, { "epoch": 0.8552023045006475, "grad_norm": 1.3874958753585815, "learning_rate": 1.0945766584376572e-06, "loss": 0.8117, "step": 24270 }, { "epoch": 0.8552375415136059, "grad_norm": 5.943840980529785, "learning_rate": 1.0940575491907168e-06, "loss": 2.9178, "step": 24271 }, { "epoch": 0.8552727785265642, "grad_norm": 3.9910426139831543, "learning_rate": 1.0935385559448119e-06, "loss": 2.5926, "step": 24272 }, { "epoch": 0.8553080155395227, "grad_norm": 3.3964059352874756, "learning_rate": 1.0930196787067016e-06, "loss": 3.2092, "step": 24273 }, { "epoch": 0.8553432525524811, "grad_norm": 4.670825958251953, "learning_rate": 1.092500917483147e-06, "loss": 2.5139, "step": 24274 }, { "epoch": 0.8553784895654395, "grad_norm": 2.7689709663391113, "learning_rate": 1.091982272280898e-06, "loss": 2.4011, "step": 24275 }, { "epoch": 0.855413726578398, "grad_norm": 1.447843074798584, "learning_rate": 1.0914637431067176e-06, "loss": 0.9487, "step": 24276 }, { "epoch": 0.8554489635913564, "grad_norm": 2.0423574447631836, "learning_rate": 1.0909453299673578e-06, "loss": 0.6862, "step": 24277 }, { "epoch": 0.8554842006043147, "grad_norm": 3.8863189220428467, "learning_rate": 1.0904270328695687e-06, "loss": 2.9527, "step": 24278 }, { "epoch": 0.8555194376172732, "grad_norm": 1.18528413772583, "learning_rate": 1.0899088518200996e-06, "loss": 0.9171, "step": 24279 }, { "epoch": 0.8555546746302316, "grad_norm": 1.6559250354766846, "learning_rate": 1.089390786825707e-06, "loss": 2.3328, "step": 24280 }, { "epoch": 0.85558991164319, "grad_norm": 3.5328524112701416, "learning_rate": 1.0888728378931324e-06, "loss": 2.8544, "step": 24281 }, { "epoch": 0.8556251486561485, "grad_norm": 6.632867336273193, "learning_rate": 1.0883550050291237e-06, "loss": 2.8134, "step": 24282 }, { "epoch": 0.8556603856691068, "grad_norm": 6.6520256996154785, "learning_rate": 1.0878372882404254e-06, "loss": 4.4903, "step": 24283 }, { "epoch": 0.8556956226820652, "grad_norm": 3.3024933338165283, "learning_rate": 1.0873196875337844e-06, "loss": 2.7098, "step": 24284 }, { "epoch": 0.8557308596950236, "grad_norm": 3.8728041648864746, "learning_rate": 1.0868022029159387e-06, "loss": 3.0659, "step": 24285 }, { "epoch": 0.8557660967079821, "grad_norm": 6.761275768280029, "learning_rate": 1.0862848343936284e-06, "loss": 4.7511, "step": 24286 }, { "epoch": 0.8558013337209405, "grad_norm": 3.2995259761810303, "learning_rate": 1.0857675819735946e-06, "loss": 2.455, "step": 24287 }, { "epoch": 0.8558365707338988, "grad_norm": 18.77596664428711, "learning_rate": 1.0852504456625735e-06, "loss": 2.1665, "step": 24288 }, { "epoch": 0.8558718077468573, "grad_norm": 1.6500812768936157, "learning_rate": 1.0847334254673002e-06, "loss": 0.6975, "step": 24289 }, { "epoch": 0.8559070447598157, "grad_norm": 3.4408185482025146, "learning_rate": 1.0842165213945111e-06, "loss": 2.7894, "step": 24290 }, { "epoch": 0.8559422817727741, "grad_norm": 3.2715179920196533, "learning_rate": 1.0836997334509359e-06, "loss": 3.0713, "step": 24291 }, { "epoch": 0.8559775187857326, "grad_norm": 1.2054051160812378, "learning_rate": 1.083183061643308e-06, "loss": 0.7593, "step": 24292 }, { "epoch": 0.856012755798691, "grad_norm": 1.6567730903625488, "learning_rate": 1.0826665059783592e-06, "loss": 0.8487, "step": 24293 }, { "epoch": 0.8560479928116493, "grad_norm": 7.518689155578613, "learning_rate": 1.0821500664628105e-06, "loss": 4.9562, "step": 24294 }, { "epoch": 0.8560832298246077, "grad_norm": 3.814216136932373, "learning_rate": 1.0816337431033953e-06, "loss": 2.8272, "step": 24295 }, { "epoch": 0.8561184668375662, "grad_norm": 2.6791372299194336, "learning_rate": 1.0811175359068382e-06, "loss": 1.0069, "step": 24296 }, { "epoch": 0.8561537038505246, "grad_norm": 2.965545177459717, "learning_rate": 1.0806014448798597e-06, "loss": 2.3147, "step": 24297 }, { "epoch": 0.856188940863483, "grad_norm": 2.965545177459717, "learning_rate": 1.0806014448798597e-06, "loss": 2.7164, "step": 24298 }, { "epoch": 0.8562241778764415, "grad_norm": 3.667600154876709, "learning_rate": 1.0800854700291818e-06, "loss": 2.791, "step": 24299 }, { "epoch": 0.8562594148893998, "grad_norm": 1.0037593841552734, "learning_rate": 1.0795696113615295e-06, "loss": 0.7502, "step": 24300 }, { "epoch": 0.8562946519023582, "grad_norm": 3.9332082271575928, "learning_rate": 1.0790538688836183e-06, "loss": 2.656, "step": 24301 }, { "epoch": 0.8563298889153167, "grad_norm": 4.142557621002197, "learning_rate": 1.0785382426021673e-06, "loss": 3.3792, "step": 24302 }, { "epoch": 0.8563651259282751, "grad_norm": 5.204930305480957, "learning_rate": 1.078022732523889e-06, "loss": 3.0776, "step": 24303 }, { "epoch": 0.8564003629412335, "grad_norm": 3.972254514694214, "learning_rate": 1.0775073386555057e-06, "loss": 5.337, "step": 24304 }, { "epoch": 0.8564355999541918, "grad_norm": 2.3045897483825684, "learning_rate": 1.0769920610037243e-06, "loss": 2.8664, "step": 24305 }, { "epoch": 0.8564708369671503, "grad_norm": 4.450252532958984, "learning_rate": 1.0764768995752573e-06, "loss": 2.5918, "step": 24306 }, { "epoch": 0.8565060739801087, "grad_norm": 0.9675558805465698, "learning_rate": 1.0759618543768158e-06, "loss": 0.9294, "step": 24307 }, { "epoch": 0.8565413109930671, "grad_norm": 1.1064519882202148, "learning_rate": 1.0754469254151067e-06, "loss": 0.7378, "step": 24308 }, { "epoch": 0.8565765480060256, "grad_norm": 4.271491527557373, "learning_rate": 1.0749321126968393e-06, "loss": 3.1918, "step": 24309 }, { "epoch": 0.8566117850189839, "grad_norm": 6.932583332061768, "learning_rate": 1.074417416228718e-06, "loss": 5.3337, "step": 24310 }, { "epoch": 0.8566470220319423, "grad_norm": 2.849661111831665, "learning_rate": 1.0739028360174465e-06, "loss": 2.8616, "step": 24311 }, { "epoch": 0.8566822590449008, "grad_norm": 2.4357340335845947, "learning_rate": 1.073388372069728e-06, "loss": 2.9639, "step": 24312 }, { "epoch": 0.8567174960578592, "grad_norm": 0.9993183016777039, "learning_rate": 1.0728740243922643e-06, "loss": 1.1554, "step": 24313 }, { "epoch": 0.8567527330708176, "grad_norm": 1.4079142808914185, "learning_rate": 1.0723597929917495e-06, "loss": 0.8877, "step": 24314 }, { "epoch": 0.856787970083776, "grad_norm": 4.543672561645508, "learning_rate": 1.0718456778748887e-06, "loss": 3.2818, "step": 24315 }, { "epoch": 0.8568232070967344, "grad_norm": 2.238112211227417, "learning_rate": 1.0713316790483762e-06, "loss": 0.6101, "step": 24316 }, { "epoch": 0.8568584441096928, "grad_norm": 3.317812204360962, "learning_rate": 1.0708177965189048e-06, "loss": 2.4266, "step": 24317 }, { "epoch": 0.8568936811226512, "grad_norm": 1.0556684732437134, "learning_rate": 1.0703040302931655e-06, "loss": 0.8348, "step": 24318 }, { "epoch": 0.8569289181356097, "grad_norm": 1.4697930812835693, "learning_rate": 1.0697903803778586e-06, "loss": 0.8919, "step": 24319 }, { "epoch": 0.856964155148568, "grad_norm": 1.2078386545181274, "learning_rate": 1.0692768467796677e-06, "loss": 0.9749, "step": 24320 }, { "epoch": 0.8569993921615264, "grad_norm": 1.244710087776184, "learning_rate": 1.068763429505284e-06, "loss": 0.6579, "step": 24321 }, { "epoch": 0.8570346291744849, "grad_norm": 1.1153430938720703, "learning_rate": 1.0682501285613911e-06, "loss": 0.7861, "step": 24322 }, { "epoch": 0.8570698661874433, "grad_norm": 0.8634878993034363, "learning_rate": 1.0677369439546826e-06, "loss": 0.6514, "step": 24323 }, { "epoch": 0.8571051032004017, "grad_norm": 17.59600830078125, "learning_rate": 1.0672238756918362e-06, "loss": 5.233, "step": 24324 }, { "epoch": 0.8571403402133602, "grad_norm": 2.4549617767333984, "learning_rate": 1.0667109237795359e-06, "loss": 1.0752, "step": 24325 }, { "epoch": 0.8571755772263185, "grad_norm": 2.3132078647613525, "learning_rate": 1.066198088224465e-06, "loss": 2.2102, "step": 24326 }, { "epoch": 0.8572108142392769, "grad_norm": 4.572910308837891, "learning_rate": 1.0656853690333013e-06, "loss": 2.4922, "step": 24327 }, { "epoch": 0.8572460512522353, "grad_norm": 1.026242971420288, "learning_rate": 1.065172766212723e-06, "loss": 0.783, "step": 24328 }, { "epoch": 0.8572812882651938, "grad_norm": 1.1133493185043335, "learning_rate": 1.0646602797694095e-06, "loss": 0.8914, "step": 24329 }, { "epoch": 0.8573165252781522, "grad_norm": 4.586532115936279, "learning_rate": 1.0641479097100327e-06, "loss": 2.7796, "step": 24330 }, { "epoch": 0.8573517622911105, "grad_norm": 3.906106948852539, "learning_rate": 1.0636356560412675e-06, "loss": 2.8664, "step": 24331 }, { "epoch": 0.857386999304069, "grad_norm": 4.673696517944336, "learning_rate": 1.0631235187697886e-06, "loss": 4.82, "step": 24332 }, { "epoch": 0.8574222363170274, "grad_norm": 1.4718871116638184, "learning_rate": 1.0626114979022594e-06, "loss": 1.06, "step": 24333 }, { "epoch": 0.8574574733299858, "grad_norm": 5.94757604598999, "learning_rate": 1.062099593445357e-06, "loss": 3.1834, "step": 24334 }, { "epoch": 0.8574927103429443, "grad_norm": 1.4473949670791626, "learning_rate": 1.061587805405747e-06, "loss": 0.9971, "step": 24335 }, { "epoch": 0.8575279473559027, "grad_norm": 7.3340983390808105, "learning_rate": 1.0610761337900933e-06, "loss": 4.9981, "step": 24336 }, { "epoch": 0.857563184368861, "grad_norm": 0.8840674757957458, "learning_rate": 1.0605645786050588e-06, "loss": 0.8669, "step": 24337 }, { "epoch": 0.8575984213818194, "grad_norm": 3.0997748374938965, "learning_rate": 1.060053139857312e-06, "loss": 3.3144, "step": 24338 }, { "epoch": 0.8576336583947779, "grad_norm": 1.072938084602356, "learning_rate": 1.0595418175535144e-06, "loss": 0.9747, "step": 24339 }, { "epoch": 0.8576688954077363, "grad_norm": 2.323880195617676, "learning_rate": 1.0590306117003202e-06, "loss": 0.7278, "step": 24340 }, { "epoch": 0.8577041324206947, "grad_norm": 1.3285655975341797, "learning_rate": 1.0585195223043898e-06, "loss": 0.8086, "step": 24341 }, { "epoch": 0.8577393694336531, "grad_norm": 2.3019940853118896, "learning_rate": 1.0580085493723858e-06, "loss": 2.8641, "step": 24342 }, { "epoch": 0.8577746064466115, "grad_norm": 1.5632320642471313, "learning_rate": 1.057497692910956e-06, "loss": 1.123, "step": 24343 }, { "epoch": 0.8578098434595699, "grad_norm": 1.2369580268859863, "learning_rate": 1.0569869529267596e-06, "loss": 0.9131, "step": 24344 }, { "epoch": 0.8578450804725284, "grad_norm": 5.621363162994385, "learning_rate": 1.0564763294264468e-06, "loss": 3.1312, "step": 24345 }, { "epoch": 0.8578803174854868, "grad_norm": 1.229589581489563, "learning_rate": 1.055965822416668e-06, "loss": 1.0037, "step": 24346 }, { "epoch": 0.8579155544984451, "grad_norm": 4.243902683258057, "learning_rate": 1.0554554319040754e-06, "loss": 2.9687, "step": 24347 }, { "epoch": 0.8579507915114036, "grad_norm": 5.786107540130615, "learning_rate": 1.054945157895314e-06, "loss": 3.3689, "step": 24348 }, { "epoch": 0.857986028524362, "grad_norm": 3.2471821308135986, "learning_rate": 1.0544350003970316e-06, "loss": 2.8274, "step": 24349 }, { "epoch": 0.8580212655373204, "grad_norm": 1.029533863067627, "learning_rate": 1.0539249594158728e-06, "loss": 0.9369, "step": 24350 }, { "epoch": 0.8580565025502788, "grad_norm": 2.8559041023254395, "learning_rate": 1.0534150349584826e-06, "loss": 2.4456, "step": 24351 }, { "epoch": 0.8580917395632373, "grad_norm": 1.289538860321045, "learning_rate": 1.0529052270314977e-06, "loss": 1.1481, "step": 24352 }, { "epoch": 0.8581269765761956, "grad_norm": 4.591545581817627, "learning_rate": 1.0523955356415638e-06, "loss": 4.5834, "step": 24353 }, { "epoch": 0.858162213589154, "grad_norm": 3.190279960632324, "learning_rate": 1.0518859607953192e-06, "loss": 2.8206, "step": 24354 }, { "epoch": 0.8581974506021125, "grad_norm": 3.4455151557922363, "learning_rate": 1.0513765024993982e-06, "loss": 4.9404, "step": 24355 }, { "epoch": 0.8582326876150709, "grad_norm": 6.820628643035889, "learning_rate": 1.050867160760437e-06, "loss": 3.1664, "step": 24356 }, { "epoch": 0.8582679246280293, "grad_norm": 4.486536979675293, "learning_rate": 1.0503579355850725e-06, "loss": 3.3566, "step": 24357 }, { "epoch": 0.8583031616409877, "grad_norm": 4.394885063171387, "learning_rate": 1.0498488269799368e-06, "loss": 2.6828, "step": 24358 }, { "epoch": 0.8583383986539461, "grad_norm": 5.744024276733398, "learning_rate": 1.0493398349516592e-06, "loss": 4.585, "step": 24359 }, { "epoch": 0.8583736356669045, "grad_norm": 2.6751773357391357, "learning_rate": 1.0488309595068679e-06, "loss": 2.6507, "step": 24360 }, { "epoch": 0.8584088726798629, "grad_norm": 3.401902437210083, "learning_rate": 1.0483222006521976e-06, "loss": 2.7968, "step": 24361 }, { "epoch": 0.8584441096928214, "grad_norm": 1.354504108428955, "learning_rate": 1.0478135583942695e-06, "loss": 0.9006, "step": 24362 }, { "epoch": 0.8584793467057797, "grad_norm": 0.8444731831550598, "learning_rate": 1.0473050327397094e-06, "loss": 0.8342, "step": 24363 }, { "epoch": 0.8585145837187381, "grad_norm": 2.309189558029175, "learning_rate": 1.0467966236951422e-06, "loss": 2.3251, "step": 24364 }, { "epoch": 0.8585498207316966, "grad_norm": 8.95493221282959, "learning_rate": 1.0462883312671889e-06, "loss": 7.3989, "step": 24365 }, { "epoch": 0.858585057744655, "grad_norm": 2.9309825897216797, "learning_rate": 1.0457801554624714e-06, "loss": 3.0515, "step": 24366 }, { "epoch": 0.8586202947576134, "grad_norm": 5.568182945251465, "learning_rate": 1.0452720962876073e-06, "loss": 4.832, "step": 24367 }, { "epoch": 0.8586555317705719, "grad_norm": 3.8326010704040527, "learning_rate": 1.0447641537492149e-06, "loss": 3.1208, "step": 24368 }, { "epoch": 0.8586907687835302, "grad_norm": 2.922870397567749, "learning_rate": 1.0442563278539108e-06, "loss": 2.6008, "step": 24369 }, { "epoch": 0.8587260057964886, "grad_norm": 2.62400221824646, "learning_rate": 1.04374861860831e-06, "loss": 2.7089, "step": 24370 }, { "epoch": 0.858761242809447, "grad_norm": 6.046113967895508, "learning_rate": 1.0432410260190207e-06, "loss": 3.2391, "step": 24371 }, { "epoch": 0.8587964798224055, "grad_norm": 1.275176763534546, "learning_rate": 1.0427335500926594e-06, "loss": 0.872, "step": 24372 }, { "epoch": 0.8588317168353639, "grad_norm": 2.9172017574310303, "learning_rate": 1.0422261908358367e-06, "loss": 2.4213, "step": 24373 }, { "epoch": 0.8588669538483222, "grad_norm": 5.11386251449585, "learning_rate": 1.041718948255157e-06, "loss": 2.9044, "step": 24374 }, { "epoch": 0.8589021908612807, "grad_norm": 6.315155982971191, "learning_rate": 1.0412118223572277e-06, "loss": 5.001, "step": 24375 }, { "epoch": 0.8589374278742391, "grad_norm": 1.9613590240478516, "learning_rate": 1.0407048131486575e-06, "loss": 0.8106, "step": 24376 }, { "epoch": 0.8589726648871975, "grad_norm": 3.523876428604126, "learning_rate": 1.040197920636049e-06, "loss": 2.8149, "step": 24377 }, { "epoch": 0.859007901900156, "grad_norm": 1.9362282752990723, "learning_rate": 1.0396911448260028e-06, "loss": 1.0157, "step": 24378 }, { "epoch": 0.8590431389131143, "grad_norm": 1.0390338897705078, "learning_rate": 1.0391844857251187e-06, "loss": 0.7568, "step": 24379 }, { "epoch": 0.8590783759260727, "grad_norm": 1.2265911102294922, "learning_rate": 1.0386779433400029e-06, "loss": 0.6678, "step": 24380 }, { "epoch": 0.8591136129390312, "grad_norm": 1.020567774772644, "learning_rate": 1.0381715176772456e-06, "loss": 0.6681, "step": 24381 }, { "epoch": 0.8591488499519896, "grad_norm": 6.512418270111084, "learning_rate": 1.0376652087434457e-06, "loss": 2.9691, "step": 24382 }, { "epoch": 0.859184086964948, "grad_norm": 1.631242036819458, "learning_rate": 1.0371590165451982e-06, "loss": 0.9551, "step": 24383 }, { "epoch": 0.8592193239779063, "grad_norm": 4.97868013381958, "learning_rate": 1.0366529410890969e-06, "loss": 5.0133, "step": 24384 }, { "epoch": 0.8592545609908648, "grad_norm": 1.0645874738693237, "learning_rate": 1.0361469823817316e-06, "loss": 0.8962, "step": 24385 }, { "epoch": 0.8592897980038232, "grad_norm": 1.4412442445755005, "learning_rate": 1.035641140429694e-06, "loss": 1.036, "step": 24386 }, { "epoch": 0.8593250350167816, "grad_norm": 0.9724048972129822, "learning_rate": 1.0351354152395732e-06, "loss": 0.9113, "step": 24387 }, { "epoch": 0.8593602720297401, "grad_norm": 4.669565200805664, "learning_rate": 1.034629806817955e-06, "loss": 4.7863, "step": 24388 }, { "epoch": 0.8593955090426985, "grad_norm": 3.0326600074768066, "learning_rate": 1.0341243151714276e-06, "loss": 3.0567, "step": 24389 }, { "epoch": 0.8594307460556568, "grad_norm": 1.976104974746704, "learning_rate": 1.0336189403065688e-06, "loss": 1.0375, "step": 24390 }, { "epoch": 0.8594659830686153, "grad_norm": 4.51386833190918, "learning_rate": 1.0331136822299682e-06, "loss": 3.391, "step": 24391 }, { "epoch": 0.8595012200815737, "grad_norm": 1.293965220451355, "learning_rate": 1.0326085409482033e-06, "loss": 0.9486, "step": 24392 }, { "epoch": 0.8595364570945321, "grad_norm": 1.2501716613769531, "learning_rate": 1.0321035164678573e-06, "loss": 0.6664, "step": 24393 }, { "epoch": 0.8595716941074905, "grad_norm": 5.955606460571289, "learning_rate": 1.0315986087955e-06, "loss": 2.6084, "step": 24394 }, { "epoch": 0.859606931120449, "grad_norm": 1.43000328540802, "learning_rate": 1.0310938179377162e-06, "loss": 1.0313, "step": 24395 }, { "epoch": 0.8596421681334073, "grad_norm": 1.5829583406448364, "learning_rate": 1.0305891439010796e-06, "loss": 0.8002, "step": 24396 }, { "epoch": 0.8596774051463657, "grad_norm": 0.9383837580680847, "learning_rate": 1.0300845866921605e-06, "loss": 0.7985, "step": 24397 }, { "epoch": 0.8597126421593242, "grad_norm": 1.1468119621276855, "learning_rate": 1.0295801463175292e-06, "loss": 0.9258, "step": 24398 }, { "epoch": 0.8597478791722826, "grad_norm": 1.0659762620925903, "learning_rate": 1.0290758227837638e-06, "loss": 0.636, "step": 24399 }, { "epoch": 0.859783116185241, "grad_norm": 2.337940216064453, "learning_rate": 1.028571616097427e-06, "loss": 0.9411, "step": 24400 }, { "epoch": 0.8598183531981994, "grad_norm": 4.822609901428223, "learning_rate": 1.0280675262650863e-06, "loss": 3.0016, "step": 24401 }, { "epoch": 0.8598535902111578, "grad_norm": 3.9179720878601074, "learning_rate": 1.0275635532933092e-06, "loss": 2.7936, "step": 24402 }, { "epoch": 0.8598888272241162, "grad_norm": 5.018969535827637, "learning_rate": 1.0270596971886604e-06, "loss": 4.7258, "step": 24403 }, { "epoch": 0.8599240642370746, "grad_norm": 1.0541061162948608, "learning_rate": 1.0265559579577023e-06, "loss": 0.7204, "step": 24404 }, { "epoch": 0.8599593012500331, "grad_norm": 3.1300442218780518, "learning_rate": 1.026052335606995e-06, "loss": 2.2933, "step": 24405 }, { "epoch": 0.8599945382629914, "grad_norm": 0.820732593536377, "learning_rate": 1.0255488301430993e-06, "loss": 0.8024, "step": 24406 }, { "epoch": 0.8600297752759498, "grad_norm": 2.6747450828552246, "learning_rate": 1.025045441572574e-06, "loss": 2.8053, "step": 24407 }, { "epoch": 0.8600650122889083, "grad_norm": 0.8675774931907654, "learning_rate": 1.0245421699019764e-06, "loss": 0.7658, "step": 24408 }, { "epoch": 0.8601002493018667, "grad_norm": 6.792649269104004, "learning_rate": 1.0240390151378565e-06, "loss": 3.4378, "step": 24409 }, { "epoch": 0.8601354863148251, "grad_norm": 1.5952844619750977, "learning_rate": 1.0235359772867736e-06, "loss": 0.8619, "step": 24410 }, { "epoch": 0.8601707233277835, "grad_norm": 2.938971519470215, "learning_rate": 1.023033056355278e-06, "loss": 2.0217, "step": 24411 }, { "epoch": 0.8602059603407419, "grad_norm": 1.071553111076355, "learning_rate": 1.0225302523499224e-06, "loss": 0.6613, "step": 24412 }, { "epoch": 0.8602411973537003, "grad_norm": 3.2245864868164062, "learning_rate": 1.0220275652772505e-06, "loss": 2.7725, "step": 24413 }, { "epoch": 0.8602764343666588, "grad_norm": 3.136414051055908, "learning_rate": 1.0215249951438144e-06, "loss": 3.5958, "step": 24414 }, { "epoch": 0.8603116713796172, "grad_norm": 3.176567792892456, "learning_rate": 1.0210225419561604e-06, "loss": 3.1635, "step": 24415 }, { "epoch": 0.8603469083925755, "grad_norm": 3.2377820014953613, "learning_rate": 1.0205202057208307e-06, "loss": 1.9268, "step": 24416 }, { "epoch": 0.8603821454055339, "grad_norm": 1.090585708618164, "learning_rate": 1.020017986444367e-06, "loss": 0.8785, "step": 24417 }, { "epoch": 0.8604173824184924, "grad_norm": 2.827481508255005, "learning_rate": 1.0195158841333164e-06, "loss": 2.5153, "step": 24418 }, { "epoch": 0.8604526194314508, "grad_norm": 1.1452007293701172, "learning_rate": 1.0190138987942134e-06, "loss": 0.7872, "step": 24419 }, { "epoch": 0.8604878564444092, "grad_norm": 1.119499921798706, "learning_rate": 1.0185120304335982e-06, "loss": 1.0486, "step": 24420 }, { "epoch": 0.8605230934573677, "grad_norm": 0.9495428800582886, "learning_rate": 1.0180102790580082e-06, "loss": 1.112, "step": 24421 }, { "epoch": 0.860558330470326, "grad_norm": 5.182185173034668, "learning_rate": 1.017508644673978e-06, "loss": 2.5929, "step": 24422 }, { "epoch": 0.8605935674832844, "grad_norm": 3.5999536514282227, "learning_rate": 1.0170071272880421e-06, "loss": 3.1358, "step": 24423 }, { "epoch": 0.8606288044962429, "grad_norm": 1.7186665534973145, "learning_rate": 1.0165057269067324e-06, "loss": 0.7495, "step": 24424 }, { "epoch": 0.8606640415092013, "grad_norm": 12.017815589904785, "learning_rate": 1.0160044435365812e-06, "loss": 4.3169, "step": 24425 }, { "epoch": 0.8606992785221597, "grad_norm": 2.1022937297821045, "learning_rate": 1.0155032771841155e-06, "loss": 0.8266, "step": 24426 }, { "epoch": 0.860734515535118, "grad_norm": 2.8959624767303467, "learning_rate": 1.0150022278558658e-06, "loss": 3.123, "step": 24427 }, { "epoch": 0.8607697525480765, "grad_norm": 0.9678441286087036, "learning_rate": 1.0145012955583532e-06, "loss": 1.0842, "step": 24428 }, { "epoch": 0.8608049895610349, "grad_norm": 3.054649591445923, "learning_rate": 1.0140004802981073e-06, "loss": 3.0073, "step": 24429 }, { "epoch": 0.8608402265739933, "grad_norm": 4.9432783126831055, "learning_rate": 1.0134997820816495e-06, "loss": 3.3414, "step": 24430 }, { "epoch": 0.8608754635869518, "grad_norm": 3.200718879699707, "learning_rate": 1.0129992009155043e-06, "loss": 2.7285, "step": 24431 }, { "epoch": 0.8609107005999102, "grad_norm": 7.285894870758057, "learning_rate": 1.0124987368061855e-06, "loss": 4.6635, "step": 24432 }, { "epoch": 0.8609459376128685, "grad_norm": 3.074692964553833, "learning_rate": 1.011998389760217e-06, "loss": 2.9269, "step": 24433 }, { "epoch": 0.860981174625827, "grad_norm": 2.6634483337402344, "learning_rate": 1.0114981597841168e-06, "loss": 2.8267, "step": 24434 }, { "epoch": 0.8610164116387854, "grad_norm": 1.8366674184799194, "learning_rate": 1.010998046884395e-06, "loss": 0.8102, "step": 24435 }, { "epoch": 0.8610516486517438, "grad_norm": 4.564911842346191, "learning_rate": 1.0104980510675677e-06, "loss": 3.0598, "step": 24436 }, { "epoch": 0.8610868856647022, "grad_norm": 1.2697674036026, "learning_rate": 1.009998172340152e-06, "loss": 1.1384, "step": 24437 }, { "epoch": 0.8611221226776606, "grad_norm": 1.1026993989944458, "learning_rate": 1.0094984107086536e-06, "loss": 1.1204, "step": 24438 }, { "epoch": 0.861157359690619, "grad_norm": 3.8881266117095947, "learning_rate": 1.008998766179583e-06, "loss": 3.053, "step": 24439 }, { "epoch": 0.8611925967035774, "grad_norm": 1.2324575185775757, "learning_rate": 1.0084992387594495e-06, "loss": 0.8921, "step": 24440 }, { "epoch": 0.8612278337165359, "grad_norm": 1.1718508005142212, "learning_rate": 1.0079998284547577e-06, "loss": 0.9563, "step": 24441 }, { "epoch": 0.8612630707294943, "grad_norm": 1.173464298248291, "learning_rate": 1.0075005352720147e-06, "loss": 0.8113, "step": 24442 }, { "epoch": 0.8612983077424526, "grad_norm": 1.1996105909347534, "learning_rate": 1.007001359217722e-06, "loss": 1.0227, "step": 24443 }, { "epoch": 0.8613335447554111, "grad_norm": 2.6669328212738037, "learning_rate": 1.006502300298382e-06, "loss": 2.7871, "step": 24444 }, { "epoch": 0.8613687817683695, "grad_norm": 1.3857415914535522, "learning_rate": 1.0060033585204953e-06, "loss": 1.0349, "step": 24445 }, { "epoch": 0.8614040187813279, "grad_norm": 1.0502092838287354, "learning_rate": 1.0055045338905634e-06, "loss": 0.7102, "step": 24446 }, { "epoch": 0.8614392557942864, "grad_norm": 4.201552391052246, "learning_rate": 1.0050058264150752e-06, "loss": 2.7215, "step": 24447 }, { "epoch": 0.8614744928072448, "grad_norm": 1.1674410104751587, "learning_rate": 1.0045072361005348e-06, "loss": 0.8602, "step": 24448 }, { "epoch": 0.8615097298202031, "grad_norm": 3.0050973892211914, "learning_rate": 1.0040087629534335e-06, "loss": 1.0063, "step": 24449 }, { "epoch": 0.8615449668331615, "grad_norm": 1.2789220809936523, "learning_rate": 1.0035104069802647e-06, "loss": 0.9793, "step": 24450 }, { "epoch": 0.86158020384612, "grad_norm": 2.029606580734253, "learning_rate": 1.003012168187516e-06, "loss": 1.0864, "step": 24451 }, { "epoch": 0.8616154408590784, "grad_norm": 2.650998115539551, "learning_rate": 1.002514046581682e-06, "loss": 3.0622, "step": 24452 }, { "epoch": 0.8616506778720368, "grad_norm": 1.1816881895065308, "learning_rate": 1.0020160421692504e-06, "loss": 0.6547, "step": 24453 }, { "epoch": 0.8616859148849952, "grad_norm": 1.2944586277008057, "learning_rate": 1.0015181549567043e-06, "loss": 0.7086, "step": 24454 }, { "epoch": 0.8617211518979536, "grad_norm": 1.0986509323120117, "learning_rate": 1.0010203849505273e-06, "loss": 0.8273, "step": 24455 }, { "epoch": 0.861756388910912, "grad_norm": 3.2260539531707764, "learning_rate": 1.0005227321572109e-06, "loss": 2.5276, "step": 24456 }, { "epoch": 0.8617916259238705, "grad_norm": 2.776984214782715, "learning_rate": 1.0000251965832308e-06, "loss": 2.3953, "step": 24457 }, { "epoch": 0.8618268629368289, "grad_norm": 7.2730326652526855, "learning_rate": 9.995277782350676e-07, "loss": 2.7731, "step": 24458 }, { "epoch": 0.8618620999497872, "grad_norm": 5.967110633850098, "learning_rate": 9.990304771192027e-07, "loss": 4.6472, "step": 24459 }, { "epoch": 0.8618973369627456, "grad_norm": 5.035961151123047, "learning_rate": 9.98533293242112e-07, "loss": 5.3316, "step": 24460 }, { "epoch": 0.8619325739757041, "grad_norm": 2.4302587509155273, "learning_rate": 9.980362266102717e-07, "loss": 2.1495, "step": 24461 }, { "epoch": 0.8619678109886625, "grad_norm": 1.670485258102417, "learning_rate": 9.975392772301563e-07, "loss": 0.5795, "step": 24462 }, { "epoch": 0.8620030480016209, "grad_norm": 1.128616452217102, "learning_rate": 9.970424451082384e-07, "loss": 0.6241, "step": 24463 }, { "epoch": 0.8620382850145794, "grad_norm": 1.1275923252105713, "learning_rate": 9.965457302509896e-07, "loss": 0.9796, "step": 24464 }, { "epoch": 0.8620735220275377, "grad_norm": 1.5090309381484985, "learning_rate": 9.960491326648792e-07, "loss": 0.7819, "step": 24465 }, { "epoch": 0.8621087590404961, "grad_norm": 4.887805938720703, "learning_rate": 9.955526523563762e-07, "loss": 3.11, "step": 24466 }, { "epoch": 0.8621439960534546, "grad_norm": 6.345694065093994, "learning_rate": 9.95056289331947e-07, "loss": 3.3674, "step": 24467 }, { "epoch": 0.862179233066413, "grad_norm": 3.2912850379943848, "learning_rate": 9.945600435980563e-07, "loss": 2.9, "step": 24468 }, { "epoch": 0.8622144700793714, "grad_norm": 2.8164987564086914, "learning_rate": 9.940639151611698e-07, "loss": 2.7463, "step": 24469 }, { "epoch": 0.8622497070923297, "grad_norm": 10.520636558532715, "learning_rate": 9.935679040277447e-07, "loss": 5.0055, "step": 24470 }, { "epoch": 0.8622849441052882, "grad_norm": 1.266798496246338, "learning_rate": 9.930720102042467e-07, "loss": 0.8117, "step": 24471 }, { "epoch": 0.8623201811182466, "grad_norm": 3.154151678085327, "learning_rate": 9.925762336971345e-07, "loss": 2.9938, "step": 24472 }, { "epoch": 0.862355418131205, "grad_norm": 1.1135835647583008, "learning_rate": 9.920805745128637e-07, "loss": 0.877, "step": 24473 }, { "epoch": 0.8623906551441635, "grad_norm": 1.144227385520935, "learning_rate": 9.915850326578868e-07, "loss": 0.8666, "step": 24474 }, { "epoch": 0.8624258921571218, "grad_norm": 4.420280456542969, "learning_rate": 9.910896081386679e-07, "loss": 5.095, "step": 24475 }, { "epoch": 0.8624611291700802, "grad_norm": 6.071223735809326, "learning_rate": 9.905943009616515e-07, "loss": 4.5803, "step": 24476 }, { "epoch": 0.8624963661830387, "grad_norm": 1.2701624631881714, "learning_rate": 9.900991111332924e-07, "loss": 0.8176, "step": 24477 }, { "epoch": 0.8625316031959971, "grad_norm": 1.6755627393722534, "learning_rate": 9.896040386600403e-07, "loss": 1.0541, "step": 24478 }, { "epoch": 0.8625668402089555, "grad_norm": 2.5258023738861084, "learning_rate": 9.89109083548343e-07, "loss": 2.2367, "step": 24479 }, { "epoch": 0.8626020772219138, "grad_norm": 3.320845603942871, "learning_rate": 9.886142458046477e-07, "loss": 2.4743, "step": 24480 }, { "epoch": 0.8626373142348723, "grad_norm": 5.666204929351807, "learning_rate": 9.881195254353992e-07, "loss": 3.1182, "step": 24481 }, { "epoch": 0.8626725512478307, "grad_norm": 1.5575069189071655, "learning_rate": 9.876249224470425e-07, "loss": 1.1314, "step": 24482 }, { "epoch": 0.8627077882607891, "grad_norm": 1.2934165000915527, "learning_rate": 9.871304368460188e-07, "loss": 0.6951, "step": 24483 }, { "epoch": 0.8627430252737476, "grad_norm": 3.489081859588623, "learning_rate": 9.8663606863877e-07, "loss": 3.1546, "step": 24484 }, { "epoch": 0.862778262286706, "grad_norm": 2.8450708389282227, "learning_rate": 9.86141817831734e-07, "loss": 1.2203, "step": 24485 }, { "epoch": 0.8628134992996643, "grad_norm": 2.643383026123047, "learning_rate": 9.856476844313489e-07, "loss": 2.5649, "step": 24486 }, { "epoch": 0.8628487363126228, "grad_norm": 2.4213688373565674, "learning_rate": 9.85153668444051e-07, "loss": 0.9328, "step": 24487 }, { "epoch": 0.8628839733255812, "grad_norm": 3.7935004234313965, "learning_rate": 9.846597698762771e-07, "loss": 3.0398, "step": 24488 }, { "epoch": 0.8629192103385396, "grad_norm": 1.2848373651504517, "learning_rate": 9.841659887344535e-07, "loss": 1.0036, "step": 24489 }, { "epoch": 0.8629544473514981, "grad_norm": 1.9921742677688599, "learning_rate": 9.836723250250191e-07, "loss": 2.4525, "step": 24490 }, { "epoch": 0.8629896843644564, "grad_norm": 3.4269402027130127, "learning_rate": 9.831787787544023e-07, "loss": 2.8689, "step": 24491 }, { "epoch": 0.8630249213774148, "grad_norm": 1.2259774208068848, "learning_rate": 9.82685349929029e-07, "loss": 0.6665, "step": 24492 }, { "epoch": 0.8630601583903732, "grad_norm": 2.650420665740967, "learning_rate": 9.821920385553263e-07, "loss": 2.5014, "step": 24493 }, { "epoch": 0.8630953954033317, "grad_norm": 11.782527923583984, "learning_rate": 9.816988446397234e-07, "loss": 5.6472, "step": 24494 }, { "epoch": 0.8631306324162901, "grad_norm": 2.6869890689849854, "learning_rate": 9.812057681886412e-07, "loss": 3.0483, "step": 24495 }, { "epoch": 0.8631658694292484, "grad_norm": 5.067850112915039, "learning_rate": 9.807128092085016e-07, "loss": 5.7697, "step": 24496 }, { "epoch": 0.8632011064422069, "grad_norm": 1.0112853050231934, "learning_rate": 9.802199677057278e-07, "loss": 0.7203, "step": 24497 }, { "epoch": 0.8632363434551653, "grad_norm": 2.811570644378662, "learning_rate": 9.797272436867366e-07, "loss": 2.641, "step": 24498 }, { "epoch": 0.8632715804681237, "grad_norm": 1.7010749578475952, "learning_rate": 9.792346371579476e-07, "loss": 0.8064, "step": 24499 }, { "epoch": 0.8633068174810822, "grad_norm": 4.0863752365112305, "learning_rate": 9.787421481257764e-07, "loss": 4.6909, "step": 24500 }, { "epoch": 0.8633420544940406, "grad_norm": 4.767584323883057, "learning_rate": 9.78249776596638e-07, "loss": 4.4184, "step": 24501 }, { "epoch": 0.8633772915069989, "grad_norm": 4.828559875488281, "learning_rate": 9.777575225769453e-07, "loss": 5.3492, "step": 24502 }, { "epoch": 0.8634125285199573, "grad_norm": 3.461427927017212, "learning_rate": 9.772653860731108e-07, "loss": 3.0283, "step": 24503 }, { "epoch": 0.8634477655329158, "grad_norm": 1.2519038915634155, "learning_rate": 9.767733670915424e-07, "loss": 1.0195, "step": 24504 }, { "epoch": 0.8634830025458742, "grad_norm": 2.4837424755096436, "learning_rate": 9.762814656386521e-07, "loss": 2.9577, "step": 24505 }, { "epoch": 0.8635182395588326, "grad_norm": 0.9386342167854309, "learning_rate": 9.757896817208434e-07, "loss": 1.1296, "step": 24506 }, { "epoch": 0.863553476571791, "grad_norm": 4.832935810089111, "learning_rate": 9.752980153445267e-07, "loss": 3.3933, "step": 24507 }, { "epoch": 0.8635887135847494, "grad_norm": 3.5286457538604736, "learning_rate": 9.748064665160994e-07, "loss": 4.194, "step": 24508 }, { "epoch": 0.8636239505977078, "grad_norm": 5.730057239532471, "learning_rate": 9.743150352419685e-07, "loss": 5.4513, "step": 24509 }, { "epoch": 0.8636591876106663, "grad_norm": 4.6237311363220215, "learning_rate": 9.738237215285361e-07, "loss": 3.1512, "step": 24510 }, { "epoch": 0.8636944246236247, "grad_norm": 2.3608100414276123, "learning_rate": 9.73332525382198e-07, "loss": 0.9976, "step": 24511 }, { "epoch": 0.863729661636583, "grad_norm": 0.9986171126365662, "learning_rate": 9.728414468093506e-07, "loss": 0.9626, "step": 24512 }, { "epoch": 0.8637648986495414, "grad_norm": 3.2651093006134033, "learning_rate": 9.72350485816398e-07, "loss": 2.8283, "step": 24513 }, { "epoch": 0.8638001356624999, "grad_norm": 3.304422616958618, "learning_rate": 9.718596424097281e-07, "loss": 2.8671, "step": 24514 }, { "epoch": 0.8638353726754583, "grad_norm": 2.7422282695770264, "learning_rate": 9.713689165957363e-07, "loss": 2.9425, "step": 24515 }, { "epoch": 0.8638706096884167, "grad_norm": 2.8724260330200195, "learning_rate": 9.70878308380815e-07, "loss": 2.8268, "step": 24516 }, { "epoch": 0.8639058467013752, "grad_norm": 1.3276548385620117, "learning_rate": 9.703878177713533e-07, "loss": 1.0202, "step": 24517 }, { "epoch": 0.8639410837143335, "grad_norm": 3.108700752258301, "learning_rate": 9.698974447737397e-07, "loss": 2.5526, "step": 24518 }, { "epoch": 0.8639763207272919, "grad_norm": 1.129056453704834, "learning_rate": 9.694071893943634e-07, "loss": 0.753, "step": 24519 }, { "epoch": 0.8640115577402504, "grad_norm": 1.0114067792892456, "learning_rate": 9.689170516396084e-07, "loss": 0.9647, "step": 24520 }, { "epoch": 0.8640467947532088, "grad_norm": 1.3080909252166748, "learning_rate": 9.684270315158583e-07, "loss": 0.6694, "step": 24521 }, { "epoch": 0.8640820317661672, "grad_norm": 1.1104228496551514, "learning_rate": 9.679371290294981e-07, "loss": 0.7667, "step": 24522 }, { "epoch": 0.8641172687791256, "grad_norm": 5.4293532371521, "learning_rate": 9.67447344186906e-07, "loss": 5.1051, "step": 24523 }, { "epoch": 0.864152505792084, "grad_norm": 0.8769849538803101, "learning_rate": 9.669576769944633e-07, "loss": 0.9124, "step": 24524 }, { "epoch": 0.8641877428050424, "grad_norm": 3.1697824001312256, "learning_rate": 9.664681274585475e-07, "loss": 2.9584, "step": 24525 }, { "epoch": 0.8642229798180008, "grad_norm": 3.6112220287323, "learning_rate": 9.659786955855376e-07, "loss": 4.5106, "step": 24526 }, { "epoch": 0.8642582168309593, "grad_norm": 1.0539809465408325, "learning_rate": 9.654893813818012e-07, "loss": 1.2001, "step": 24527 }, { "epoch": 0.8642934538439176, "grad_norm": 3.7031052112579346, "learning_rate": 9.650001848537183e-07, "loss": 2.6586, "step": 24528 }, { "epoch": 0.864328690856876, "grad_norm": 4.283527851104736, "learning_rate": 9.645111060076617e-07, "loss": 2.4492, "step": 24529 }, { "epoch": 0.8643639278698345, "grad_norm": 1.7623145580291748, "learning_rate": 9.640221448499964e-07, "loss": 0.6911, "step": 24530 }, { "epoch": 0.8643991648827929, "grad_norm": 0.9094802141189575, "learning_rate": 9.635333013870917e-07, "loss": 0.8669, "step": 24531 }, { "epoch": 0.8644344018957513, "grad_norm": 3.6693246364593506, "learning_rate": 9.630445756253203e-07, "loss": 2.9698, "step": 24532 }, { "epoch": 0.8644696389087098, "grad_norm": 3.186001777648926, "learning_rate": 9.625559675710427e-07, "loss": 0.6534, "step": 24533 }, { "epoch": 0.8645048759216681, "grad_norm": 4.487429618835449, "learning_rate": 9.620674772306248e-07, "loss": 3.0662, "step": 24534 }, { "epoch": 0.8645401129346265, "grad_norm": 7.498418807983398, "learning_rate": 9.615791046104294e-07, "loss": 8.4764, "step": 24535 }, { "epoch": 0.8645753499475849, "grad_norm": 1.0737805366516113, "learning_rate": 9.61090849716818e-07, "loss": 0.8678, "step": 24536 }, { "epoch": 0.8646105869605434, "grad_norm": 6.4150848388671875, "learning_rate": 9.60602712556149e-07, "loss": 4.9591, "step": 24537 }, { "epoch": 0.8646458239735018, "grad_norm": 2.8816373348236084, "learning_rate": 9.601146931347815e-07, "loss": 2.9245, "step": 24538 }, { "epoch": 0.8646810609864601, "grad_norm": 1.0450137853622437, "learning_rate": 9.596267914590707e-07, "loss": 0.6684, "step": 24539 }, { "epoch": 0.8647162979994186, "grad_norm": 1.5763461589813232, "learning_rate": 9.591390075353735e-07, "loss": 0.9568, "step": 24540 }, { "epoch": 0.864751535012377, "grad_norm": 1.2175301313400269, "learning_rate": 9.586513413700427e-07, "loss": 1.1947, "step": 24541 }, { "epoch": 0.8647867720253354, "grad_norm": 1.5189028978347778, "learning_rate": 9.581637929694287e-07, "loss": 0.8843, "step": 24542 }, { "epoch": 0.8648220090382939, "grad_norm": 2.8157527446746826, "learning_rate": 9.576763623398832e-07, "loss": 2.8807, "step": 24543 }, { "epoch": 0.8648572460512522, "grad_norm": 18.882097244262695, "learning_rate": 9.571890494877556e-07, "loss": 2.997, "step": 24544 }, { "epoch": 0.8648924830642106, "grad_norm": 8.02839469909668, "learning_rate": 9.56701854419394e-07, "loss": 3.31, "step": 24545 }, { "epoch": 0.864927720077169, "grad_norm": 1.3906505107879639, "learning_rate": 9.562147771411413e-07, "loss": 0.7721, "step": 24546 }, { "epoch": 0.8649629570901275, "grad_norm": 3.6730875968933105, "learning_rate": 9.557278176593398e-07, "loss": 2.9569, "step": 24547 }, { "epoch": 0.8649981941030859, "grad_norm": 1.170472264289856, "learning_rate": 9.552409759803405e-07, "loss": 1.0871, "step": 24548 }, { "epoch": 0.8650334311160442, "grad_norm": 1.653967022895813, "learning_rate": 9.54754252110478e-07, "loss": 0.7419, "step": 24549 }, { "epoch": 0.8650686681290027, "grad_norm": 17.893980026245117, "learning_rate": 9.54267646056093e-07, "loss": 2.5292, "step": 24550 }, { "epoch": 0.8651039051419611, "grad_norm": 4.252853870391846, "learning_rate": 9.537811578235234e-07, "loss": 2.9247, "step": 24551 }, { "epoch": 0.8651391421549195, "grad_norm": 5.597041606903076, "learning_rate": 9.532947874191079e-07, "loss": 4.6129, "step": 24552 }, { "epoch": 0.865174379167878, "grad_norm": 6.161646366119385, "learning_rate": 9.528085348491788e-07, "loss": 2.8313, "step": 24553 }, { "epoch": 0.8652096161808364, "grad_norm": 1.0729786157608032, "learning_rate": 9.523224001200725e-07, "loss": 0.8817, "step": 24554 }, { "epoch": 0.8652448531937947, "grad_norm": 5.653758525848389, "learning_rate": 9.518363832381183e-07, "loss": 4.6175, "step": 24555 }, { "epoch": 0.8652800902067532, "grad_norm": 3.2407307624816895, "learning_rate": 9.513504842096477e-07, "loss": 2.1985, "step": 24556 }, { "epoch": 0.8653153272197116, "grad_norm": 4.545494079589844, "learning_rate": 9.508647030409901e-07, "loss": 4.2593, "step": 24557 }, { "epoch": 0.86535056423267, "grad_norm": 3.3231306076049805, "learning_rate": 9.503790397384727e-07, "loss": 2.7047, "step": 24558 }, { "epoch": 0.8653858012456284, "grad_norm": 3.3966963291168213, "learning_rate": 9.498934943084204e-07, "loss": 2.8398, "step": 24559 }, { "epoch": 0.8654210382585868, "grad_norm": 2.645106554031372, "learning_rate": 9.494080667571593e-07, "loss": 2.7149, "step": 24560 }, { "epoch": 0.8654562752715452, "grad_norm": 0.9269329309463501, "learning_rate": 9.48922757091012e-07, "loss": 0.9492, "step": 24561 }, { "epoch": 0.8654915122845036, "grad_norm": 6.417209625244141, "learning_rate": 9.484375653162947e-07, "loss": 5.2266, "step": 24562 }, { "epoch": 0.8655267492974621, "grad_norm": 4.960114479064941, "learning_rate": 9.479524914393334e-07, "loss": 3.3757, "step": 24563 }, { "epoch": 0.8655619863104205, "grad_norm": 5.005548477172852, "learning_rate": 9.474675354664453e-07, "loss": 5.1907, "step": 24564 }, { "epoch": 0.8655972233233789, "grad_norm": 1.4249635934829712, "learning_rate": 9.469826974039442e-07, "loss": 0.6721, "step": 24565 }, { "epoch": 0.8656324603363373, "grad_norm": 1.4820888042449951, "learning_rate": 9.464979772581451e-07, "loss": 1.0934, "step": 24566 }, { "epoch": 0.8656676973492957, "grad_norm": 4.057193756103516, "learning_rate": 9.460133750353662e-07, "loss": 3.0014, "step": 24567 }, { "epoch": 0.8657029343622541, "grad_norm": 1.3609744310379028, "learning_rate": 9.455288907419146e-07, "loss": 0.7404, "step": 24568 }, { "epoch": 0.8657381713752125, "grad_norm": 3.057058095932007, "learning_rate": 9.450445243841033e-07, "loss": 2.4489, "step": 24569 }, { "epoch": 0.865773408388171, "grad_norm": 1.0992279052734375, "learning_rate": 9.445602759682393e-07, "loss": 0.8648, "step": 24570 }, { "epoch": 0.8658086454011293, "grad_norm": 3.062574863433838, "learning_rate": 9.44076145500632e-07, "loss": 2.6692, "step": 24571 }, { "epoch": 0.8658438824140877, "grad_norm": 4.670673847198486, "learning_rate": 9.435921329875863e-07, "loss": 5.071, "step": 24572 }, { "epoch": 0.8658791194270462, "grad_norm": 6.076564788818359, "learning_rate": 9.431082384354062e-07, "loss": 2.1634, "step": 24573 }, { "epoch": 0.8659143564400046, "grad_norm": 1.111008882522583, "learning_rate": 9.426244618503954e-07, "loss": 0.768, "step": 24574 }, { "epoch": 0.865949593452963, "grad_norm": 4.461453437805176, "learning_rate": 9.421408032388557e-07, "loss": 2.9705, "step": 24575 }, { "epoch": 0.8659848304659215, "grad_norm": 2.0464086532592773, "learning_rate": 9.416572626070841e-07, "loss": 2.5327, "step": 24576 }, { "epoch": 0.8660200674788798, "grad_norm": 3.593273878097534, "learning_rate": 9.411738399613812e-07, "loss": 2.8384, "step": 24577 }, { "epoch": 0.8660553044918382, "grad_norm": 1.3491159677505493, "learning_rate": 9.406905353080431e-07, "loss": 0.8494, "step": 24578 }, { "epoch": 0.8660905415047966, "grad_norm": 4.777735233306885, "learning_rate": 9.402073486533647e-07, "loss": 2.9342, "step": 24579 }, { "epoch": 0.8661257785177551, "grad_norm": 4.345947742462158, "learning_rate": 9.39724280003641e-07, "loss": 2.7551, "step": 24580 }, { "epoch": 0.8661610155307135, "grad_norm": 1.3522437810897827, "learning_rate": 9.392413293651592e-07, "loss": 1.0356, "step": 24581 }, { "epoch": 0.8661962525436718, "grad_norm": 2.909651279449463, "learning_rate": 9.387584967442153e-07, "loss": 3.0715, "step": 24582 }, { "epoch": 0.8662314895566303, "grad_norm": 1.1787225008010864, "learning_rate": 9.382757821470989e-07, "loss": 0.8549, "step": 24583 }, { "epoch": 0.8662667265695887, "grad_norm": 5.326894760131836, "learning_rate": 9.377931855800915e-07, "loss": 4.831, "step": 24584 }, { "epoch": 0.8663019635825471, "grad_norm": 3.10188627243042, "learning_rate": 9.373107070494803e-07, "loss": 2.8589, "step": 24585 }, { "epoch": 0.8663372005955056, "grad_norm": 2.420835018157959, "learning_rate": 9.368283465615557e-07, "loss": 2.6675, "step": 24586 }, { "epoch": 0.8663724376084639, "grad_norm": 1.1786755323410034, "learning_rate": 9.363461041225952e-07, "loss": 0.7774, "step": 24587 }, { "epoch": 0.8664076746214223, "grad_norm": 0.9101313352584839, "learning_rate": 9.358639797388802e-07, "loss": 0.9739, "step": 24588 }, { "epoch": 0.8664429116343808, "grad_norm": 1.19887375831604, "learning_rate": 9.353819734166913e-07, "loss": 0.9307, "step": 24589 }, { "epoch": 0.8664781486473392, "grad_norm": 1.518634557723999, "learning_rate": 9.349000851623081e-07, "loss": 1.013, "step": 24590 }, { "epoch": 0.8665133856602976, "grad_norm": 3.732131004333496, "learning_rate": 9.344183149820052e-07, "loss": 3.4501, "step": 24591 }, { "epoch": 0.8665486226732559, "grad_norm": 2.5439951419830322, "learning_rate": 9.33936662882059e-07, "loss": 2.671, "step": 24592 }, { "epoch": 0.8665838596862144, "grad_norm": 1.0934561491012573, "learning_rate": 9.334551288687421e-07, "loss": 0.8997, "step": 24593 }, { "epoch": 0.8666190966991728, "grad_norm": 3.7672348022460938, "learning_rate": 9.329737129483263e-07, "loss": 2.7011, "step": 24594 }, { "epoch": 0.8666543337121312, "grad_norm": 0.7750113010406494, "learning_rate": 9.324924151270843e-07, "loss": 0.6937, "step": 24595 }, { "epoch": 0.8666895707250897, "grad_norm": 6.071338176727295, "learning_rate": 9.320112354112821e-07, "loss": 5.9308, "step": 24596 }, { "epoch": 0.866724807738048, "grad_norm": 3.647794485092163, "learning_rate": 9.315301738071903e-07, "loss": 2.6443, "step": 24597 }, { "epoch": 0.8667600447510064, "grad_norm": 2.9107859134674072, "learning_rate": 9.310492303210717e-07, "loss": 2.9196, "step": 24598 }, { "epoch": 0.8667952817639649, "grad_norm": 5.582028865814209, "learning_rate": 9.305684049591945e-07, "loss": 6.3394, "step": 24599 }, { "epoch": 0.8668305187769233, "grad_norm": 0.7948172688484192, "learning_rate": 9.300876977278162e-07, "loss": 0.8084, "step": 24600 }, { "epoch": 0.8668657557898817, "grad_norm": 5.8429951667785645, "learning_rate": 9.296071086332015e-07, "loss": 3.4145, "step": 24601 }, { "epoch": 0.86690099280284, "grad_norm": 1.05898916721344, "learning_rate": 9.291266376816122e-07, "loss": 0.9362, "step": 24602 }, { "epoch": 0.8669362298157985, "grad_norm": 2.5778729915618896, "learning_rate": 9.286462848793021e-07, "loss": 2.3671, "step": 24603 }, { "epoch": 0.8669714668287569, "grad_norm": 2.8953075408935547, "learning_rate": 9.281660502325273e-07, "loss": 3.0421, "step": 24604 }, { "epoch": 0.8670067038417153, "grad_norm": 2.935195207595825, "learning_rate": 9.276859337475485e-07, "loss": 2.773, "step": 24605 }, { "epoch": 0.8670419408546738, "grad_norm": 4.329037189483643, "learning_rate": 9.272059354306151e-07, "loss": 2.8179, "step": 24606 }, { "epoch": 0.8670771778676322, "grad_norm": 3.135479211807251, "learning_rate": 9.267260552879798e-07, "loss": 2.5301, "step": 24607 }, { "epoch": 0.8671124148805905, "grad_norm": 4.186887741088867, "learning_rate": 9.262462933258943e-07, "loss": 4.8335, "step": 24608 }, { "epoch": 0.867147651893549, "grad_norm": 3.927335739135742, "learning_rate": 9.257666495506056e-07, "loss": 4.7339, "step": 24609 }, { "epoch": 0.8671828889065074, "grad_norm": 2.776106119155884, "learning_rate": 9.252871239683614e-07, "loss": 2.7585, "step": 24610 }, { "epoch": 0.8672181259194658, "grad_norm": 1.51531183719635, "learning_rate": 9.248077165854097e-07, "loss": 1.1755, "step": 24611 }, { "epoch": 0.8672533629324242, "grad_norm": 0.9502837061882019, "learning_rate": 9.243284274079933e-07, "loss": 0.8705, "step": 24612 }, { "epoch": 0.8672885999453827, "grad_norm": 1.0467090606689453, "learning_rate": 9.238492564423551e-07, "loss": 0.8084, "step": 24613 }, { "epoch": 0.867323836958341, "grad_norm": 1.1267303228378296, "learning_rate": 9.233702036947356e-07, "loss": 0.8831, "step": 24614 }, { "epoch": 0.8673590739712994, "grad_norm": 3.047863721847534, "learning_rate": 9.228912691713765e-07, "loss": 2.8171, "step": 24615 }, { "epoch": 0.8673943109842579, "grad_norm": 3.473752021789551, "learning_rate": 9.224124528785139e-07, "loss": 2.5423, "step": 24616 }, { "epoch": 0.8674295479972163, "grad_norm": 8.398557662963867, "learning_rate": 9.219337548223862e-07, "loss": 4.9165, "step": 24617 }, { "epoch": 0.8674647850101747, "grad_norm": 0.7964876294136047, "learning_rate": 9.214551750092293e-07, "loss": 0.8203, "step": 24618 }, { "epoch": 0.8675000220231331, "grad_norm": 5.167464733123779, "learning_rate": 9.209767134452718e-07, "loss": 2.4577, "step": 24619 }, { "epoch": 0.8675352590360915, "grad_norm": 1.057450532913208, "learning_rate": 9.20498370136751e-07, "loss": 0.6494, "step": 24620 }, { "epoch": 0.8675704960490499, "grad_norm": 1.3926482200622559, "learning_rate": 9.20020145089896e-07, "loss": 0.9689, "step": 24621 }, { "epoch": 0.8676057330620084, "grad_norm": 3.449453830718994, "learning_rate": 9.195420383109343e-07, "loss": 2.5343, "step": 24622 }, { "epoch": 0.8676409700749668, "grad_norm": 5.6115899085998535, "learning_rate": 9.190640498060932e-07, "loss": 4.6349, "step": 24623 }, { "epoch": 0.8676762070879251, "grad_norm": 1.2111510038375854, "learning_rate": 9.18586179581602e-07, "loss": 0.8833, "step": 24624 }, { "epoch": 0.8677114441008835, "grad_norm": 4.466387748718262, "learning_rate": 9.181084276436813e-07, "loss": 3.1606, "step": 24625 }, { "epoch": 0.867746681113842, "grad_norm": 1.869336485862732, "learning_rate": 9.176307939985552e-07, "loss": 0.86, "step": 24626 }, { "epoch": 0.8677819181268004, "grad_norm": 1.0674290657043457, "learning_rate": 9.17153278652445e-07, "loss": 0.6523, "step": 24627 }, { "epoch": 0.8678171551397588, "grad_norm": 5.815131187438965, "learning_rate": 9.166758816115696e-07, "loss": 4.9189, "step": 24628 }, { "epoch": 0.8678523921527173, "grad_norm": 1.1759097576141357, "learning_rate": 9.16198602882149e-07, "loss": 0.856, "step": 24629 }, { "epoch": 0.8678876291656756, "grad_norm": 5.649778842926025, "learning_rate": 9.157214424703976e-07, "loss": 4.6106, "step": 24630 }, { "epoch": 0.867922866178634, "grad_norm": 0.9093621373176575, "learning_rate": 9.152444003825322e-07, "loss": 0.9227, "step": 24631 }, { "epoch": 0.8679581031915925, "grad_norm": 5.160057067871094, "learning_rate": 9.147674766247649e-07, "loss": 3.4168, "step": 24632 }, { "epoch": 0.8679933402045509, "grad_norm": 7.189404487609863, "learning_rate": 9.142906712033095e-07, "loss": 4.2347, "step": 24633 }, { "epoch": 0.8680285772175093, "grad_norm": 1.175477147102356, "learning_rate": 9.138139841243754e-07, "loss": 0.9635, "step": 24634 }, { "epoch": 0.8680638142304676, "grad_norm": 1.2912554740905762, "learning_rate": 9.13337415394171e-07, "loss": 1.105, "step": 24635 }, { "epoch": 0.8680990512434261, "grad_norm": 4.026507377624512, "learning_rate": 9.128609650189047e-07, "loss": 2.8106, "step": 24636 }, { "epoch": 0.8681342882563845, "grad_norm": 4.615684509277344, "learning_rate": 9.123846330047826e-07, "loss": 5.198, "step": 24637 }, { "epoch": 0.8681695252693429, "grad_norm": 1.0357346534729004, "learning_rate": 9.119084193580064e-07, "loss": 0.8998, "step": 24638 }, { "epoch": 0.8682047622823014, "grad_norm": 1.317389726638794, "learning_rate": 9.114323240847811e-07, "loss": 0.8595, "step": 24639 }, { "epoch": 0.8682399992952597, "grad_norm": 1.0029641389846802, "learning_rate": 9.109563471913097e-07, "loss": 0.9022, "step": 24640 }, { "epoch": 0.8682752363082181, "grad_norm": 3.262657642364502, "learning_rate": 9.104804886837882e-07, "loss": 0.7078, "step": 24641 }, { "epoch": 0.8683104733211766, "grad_norm": 3.895580291748047, "learning_rate": 9.100047485684138e-07, "loss": 3.2876, "step": 24642 }, { "epoch": 0.868345710334135, "grad_norm": 1.8373632431030273, "learning_rate": 9.095291268513884e-07, "loss": 0.7558, "step": 24643 }, { "epoch": 0.8683809473470934, "grad_norm": 1.5205401182174683, "learning_rate": 9.090536235389036e-07, "loss": 0.8713, "step": 24644 }, { "epoch": 0.8684161843600517, "grad_norm": 2.555053234100342, "learning_rate": 9.085782386371522e-07, "loss": 0.8828, "step": 24645 }, { "epoch": 0.8684514213730102, "grad_norm": 6.529934406280518, "learning_rate": 9.081029721523282e-07, "loss": 3.0682, "step": 24646 }, { "epoch": 0.8684866583859686, "grad_norm": 2.6785078048706055, "learning_rate": 9.076278240906211e-07, "loss": 2.4486, "step": 24647 }, { "epoch": 0.868521895398927, "grad_norm": 1.106460690498352, "learning_rate": 9.071527944582192e-07, "loss": 0.9212, "step": 24648 }, { "epoch": 0.8685571324118855, "grad_norm": 1.2751379013061523, "learning_rate": 9.06677883261311e-07, "loss": 0.9348, "step": 24649 }, { "epoch": 0.8685923694248439, "grad_norm": 1.7329694032669067, "learning_rate": 9.062030905060814e-07, "loss": 0.9622, "step": 24650 }, { "epoch": 0.8686276064378022, "grad_norm": 1.8644533157348633, "learning_rate": 9.057284161987146e-07, "loss": 0.8926, "step": 24651 }, { "epoch": 0.8686628434507607, "grad_norm": 1.151302456855774, "learning_rate": 9.052538603453942e-07, "loss": 0.7446, "step": 24652 }, { "epoch": 0.8686980804637191, "grad_norm": 3.6864781379699707, "learning_rate": 9.047794229522999e-07, "loss": 2.4538, "step": 24653 }, { "epoch": 0.8687333174766775, "grad_norm": 1.5178782939910889, "learning_rate": 9.043051040256135e-07, "loss": 0.9801, "step": 24654 }, { "epoch": 0.868768554489636, "grad_norm": 3.2862377166748047, "learning_rate": 9.038309035715109e-07, "loss": 2.9761, "step": 24655 }, { "epoch": 0.8688037915025943, "grad_norm": 3.8632140159606934, "learning_rate": 9.033568215961708e-07, "loss": 3.249, "step": 24656 }, { "epoch": 0.8688390285155527, "grad_norm": 5.2219319343566895, "learning_rate": 9.028828581057636e-07, "loss": 4.8555, "step": 24657 }, { "epoch": 0.8688742655285111, "grad_norm": 3.1471476554870605, "learning_rate": 9.024090131064678e-07, "loss": 2.8941, "step": 24658 }, { "epoch": 0.8689095025414696, "grad_norm": 2.002157211303711, "learning_rate": 9.019352866044561e-07, "loss": 0.6615, "step": 24659 }, { "epoch": 0.868944739554428, "grad_norm": 4.35582160949707, "learning_rate": 9.014616786058927e-07, "loss": 3.2036, "step": 24660 }, { "epoch": 0.8689799765673863, "grad_norm": 1.1864211559295654, "learning_rate": 9.009881891169492e-07, "loss": 1.0443, "step": 24661 }, { "epoch": 0.8690152135803448, "grad_norm": 1.0580142736434937, "learning_rate": 9.005148181437973e-07, "loss": 0.8718, "step": 24662 }, { "epoch": 0.8690504505933032, "grad_norm": 4.314872741699219, "learning_rate": 9.000415656925965e-07, "loss": 3.1179, "step": 24663 }, { "epoch": 0.8690856876062616, "grad_norm": 1.2838644981384277, "learning_rate": 8.995684317695142e-07, "loss": 0.929, "step": 24664 }, { "epoch": 0.8691209246192201, "grad_norm": 2.937913417816162, "learning_rate": 8.99095416380712e-07, "loss": 2.8895, "step": 24665 }, { "epoch": 0.8691561616321785, "grad_norm": 5.243342876434326, "learning_rate": 8.986225195323506e-07, "loss": 2.8624, "step": 24666 }, { "epoch": 0.8691913986451368, "grad_norm": 3.585991621017456, "learning_rate": 8.981497412305917e-07, "loss": 2.6668, "step": 24667 }, { "epoch": 0.8692266356580952, "grad_norm": 2.852499008178711, "learning_rate": 8.976770814815905e-07, "loss": 2.8757, "step": 24668 }, { "epoch": 0.8692618726710537, "grad_norm": 1.0888910293579102, "learning_rate": 8.972045402915052e-07, "loss": 1.1328, "step": 24669 }, { "epoch": 0.8692971096840121, "grad_norm": 5.034414291381836, "learning_rate": 8.967321176664901e-07, "loss": 3.8567, "step": 24670 }, { "epoch": 0.8693323466969705, "grad_norm": 7.245591640472412, "learning_rate": 8.962598136126988e-07, "loss": 4.4665, "step": 24671 }, { "epoch": 0.869367583709929, "grad_norm": 0.978714108467102, "learning_rate": 8.957876281362831e-07, "loss": 0.7207, "step": 24672 }, { "epoch": 0.8694028207228873, "grad_norm": 1.4409492015838623, "learning_rate": 8.953155612433939e-07, "loss": 0.737, "step": 24673 }, { "epoch": 0.8694380577358457, "grad_norm": 3.842991352081299, "learning_rate": 8.948436129401783e-07, "loss": 3.0701, "step": 24674 }, { "epoch": 0.8694732947488042, "grad_norm": 2.794445276260376, "learning_rate": 8.94371783232787e-07, "loss": 2.3634, "step": 24675 }, { "epoch": 0.8695085317617626, "grad_norm": 2.9493300914764404, "learning_rate": 8.939000721273583e-07, "loss": 2.435, "step": 24676 }, { "epoch": 0.869543768774721, "grad_norm": 1.2544317245483398, "learning_rate": 8.934284796300441e-07, "loss": 0.8494, "step": 24677 }, { "epoch": 0.8695790057876793, "grad_norm": 3.7237560749053955, "learning_rate": 8.92957005746985e-07, "loss": 4.6951, "step": 24678 }, { "epoch": 0.8696142428006378, "grad_norm": 0.8836169838905334, "learning_rate": 8.924856504843205e-07, "loss": 0.6784, "step": 24679 }, { "epoch": 0.8696494798135962, "grad_norm": 4.060762882232666, "learning_rate": 8.920144138481868e-07, "loss": 4.8144, "step": 24680 }, { "epoch": 0.8696847168265546, "grad_norm": 5.531142234802246, "learning_rate": 8.915432958447301e-07, "loss": 5.0156, "step": 24681 }, { "epoch": 0.8697199538395131, "grad_norm": 1.1857101917266846, "learning_rate": 8.910722964800811e-07, "loss": 0.9495, "step": 24682 }, { "epoch": 0.8697551908524714, "grad_norm": 4.755704402923584, "learning_rate": 8.906014157603749e-07, "loss": 2.5503, "step": 24683 }, { "epoch": 0.8697904278654298, "grad_norm": 2.3950536251068115, "learning_rate": 8.901306536917431e-07, "loss": 2.5313, "step": 24684 }, { "epoch": 0.8698256648783883, "grad_norm": 1.3522815704345703, "learning_rate": 8.896600102803243e-07, "loss": 0.7726, "step": 24685 }, { "epoch": 0.8698609018913467, "grad_norm": 4.853440761566162, "learning_rate": 8.891894855322436e-07, "loss": 2.9146, "step": 24686 }, { "epoch": 0.8698961389043051, "grad_norm": 1.7192306518554688, "learning_rate": 8.887190794536294e-07, "loss": 0.9864, "step": 24687 }, { "epoch": 0.8699313759172634, "grad_norm": 6.628043174743652, "learning_rate": 8.882487920506088e-07, "loss": 3.1899, "step": 24688 }, { "epoch": 0.8699666129302219, "grad_norm": 1.0925779342651367, "learning_rate": 8.877786233293095e-07, "loss": 1.126, "step": 24689 }, { "epoch": 0.8700018499431803, "grad_norm": 0.9255868792533875, "learning_rate": 8.873085732958541e-07, "loss": 0.7882, "step": 24690 }, { "epoch": 0.8700370869561387, "grad_norm": 5.993507385253906, "learning_rate": 8.868386419563657e-07, "loss": 4.9894, "step": 24691 }, { "epoch": 0.8700723239690972, "grad_norm": 4.286248207092285, "learning_rate": 8.863688293169648e-07, "loss": 2.4364, "step": 24692 }, { "epoch": 0.8701075609820556, "grad_norm": 0.9507619738578796, "learning_rate": 8.858991353837698e-07, "loss": 0.958, "step": 24693 }, { "epoch": 0.8701427979950139, "grad_norm": 1.2997950315475464, "learning_rate": 8.854295601629026e-07, "loss": 0.9641, "step": 24694 }, { "epoch": 0.8701780350079724, "grad_norm": 1.227910041809082, "learning_rate": 8.849601036604716e-07, "loss": 0.7272, "step": 24695 }, { "epoch": 0.8702132720209308, "grad_norm": 1.3244154453277588, "learning_rate": 8.844907658825985e-07, "loss": 0.6199, "step": 24696 }, { "epoch": 0.8702485090338892, "grad_norm": 2.930821657180786, "learning_rate": 8.840215468353963e-07, "loss": 2.8596, "step": 24697 }, { "epoch": 0.8702837460468477, "grad_norm": 6.216341972351074, "learning_rate": 8.835524465249734e-07, "loss": 4.8983, "step": 24698 }, { "epoch": 0.870318983059806, "grad_norm": 1.3343167304992676, "learning_rate": 8.830834649574382e-07, "loss": 0.8129, "step": 24699 }, { "epoch": 0.8703542200727644, "grad_norm": 0.9940195083618164, "learning_rate": 8.826146021389059e-07, "loss": 0.9673, "step": 24700 }, { "epoch": 0.8703894570857228, "grad_norm": 4.088086128234863, "learning_rate": 8.821458580754793e-07, "loss": 4.66, "step": 24701 }, { "epoch": 0.8704246940986813, "grad_norm": 3.7332417964935303, "learning_rate": 8.816772327732626e-07, "loss": 3.0151, "step": 24702 }, { "epoch": 0.8704599311116397, "grad_norm": 4.489217281341553, "learning_rate": 8.812087262383606e-07, "loss": 3.0243, "step": 24703 }, { "epoch": 0.870495168124598, "grad_norm": 3.05442476272583, "learning_rate": 8.807403384768798e-07, "loss": 2.4055, "step": 24704 }, { "epoch": 0.8705304051375565, "grad_norm": 3.3995590209960938, "learning_rate": 8.802720694949162e-07, "loss": 2.3381, "step": 24705 }, { "epoch": 0.8705656421505149, "grad_norm": 4.615240573883057, "learning_rate": 8.798039192985708e-07, "loss": 3.0503, "step": 24706 }, { "epoch": 0.8706008791634733, "grad_norm": 1.5882186889648438, "learning_rate": 8.793358878939406e-07, "loss": 1.0162, "step": 24707 }, { "epoch": 0.8706361161764318, "grad_norm": 1.7335582971572876, "learning_rate": 8.788679752871221e-07, "loss": 0.9547, "step": 24708 }, { "epoch": 0.8706713531893902, "grad_norm": 1.3056695461273193, "learning_rate": 8.784001814842102e-07, "loss": 0.8287, "step": 24709 }, { "epoch": 0.8707065902023485, "grad_norm": 1.108436942100525, "learning_rate": 8.77932506491298e-07, "loss": 0.7802, "step": 24710 }, { "epoch": 0.8707418272153069, "grad_norm": 2.737457513809204, "learning_rate": 8.774649503144772e-07, "loss": 3.0925, "step": 24711 }, { "epoch": 0.8707770642282654, "grad_norm": 0.9460933804512024, "learning_rate": 8.769975129598374e-07, "loss": 0.4946, "step": 24712 }, { "epoch": 0.8708123012412238, "grad_norm": 7.8412184715271, "learning_rate": 8.765301944334681e-07, "loss": 2.9062, "step": 24713 }, { "epoch": 0.8708475382541822, "grad_norm": 4.353003978729248, "learning_rate": 8.760629947414523e-07, "loss": 4.739, "step": 24714 }, { "epoch": 0.8708827752671406, "grad_norm": 6.263458251953125, "learning_rate": 8.755959138898795e-07, "loss": 6.778, "step": 24715 }, { "epoch": 0.870918012280099, "grad_norm": 0.7905840277671814, "learning_rate": 8.751289518848338e-07, "loss": 0.7939, "step": 24716 }, { "epoch": 0.8709532492930574, "grad_norm": 1.0636273622512817, "learning_rate": 8.746621087323948e-07, "loss": 0.6942, "step": 24717 }, { "epoch": 0.8709884863060159, "grad_norm": 6.432676315307617, "learning_rate": 8.74195384438642e-07, "loss": 6.9984, "step": 24718 }, { "epoch": 0.8710237233189743, "grad_norm": 2.824638843536377, "learning_rate": 8.737287790096605e-07, "loss": 1.0301, "step": 24719 }, { "epoch": 0.8710589603319326, "grad_norm": 3.830420970916748, "learning_rate": 8.732622924515222e-07, "loss": 4.0253, "step": 24720 }, { "epoch": 0.871094197344891, "grad_norm": 0.8989851474761963, "learning_rate": 8.727959247703055e-07, "loss": 0.946, "step": 24721 }, { "epoch": 0.8711294343578495, "grad_norm": 3.4008448123931885, "learning_rate": 8.723296759720823e-07, "loss": 2.5209, "step": 24722 }, { "epoch": 0.8711646713708079, "grad_norm": 4.6571197509765625, "learning_rate": 8.71863546062931e-07, "loss": 4.6386, "step": 24723 }, { "epoch": 0.8711999083837663, "grad_norm": 2.666761636734009, "learning_rate": 8.71397535048919e-07, "loss": 2.7791, "step": 24724 }, { "epoch": 0.8712351453967248, "grad_norm": 7.245334148406982, "learning_rate": 8.709316429361158e-07, "loss": 6.9228, "step": 24725 }, { "epoch": 0.8712703824096831, "grad_norm": 4.374128341674805, "learning_rate": 8.704658697305913e-07, "loss": 3.4865, "step": 24726 }, { "epoch": 0.8713056194226415, "grad_norm": 6.602663040161133, "learning_rate": 8.700002154384113e-07, "loss": 6.5819, "step": 24727 }, { "epoch": 0.8713408564356, "grad_norm": 6.710937023162842, "learning_rate": 8.695346800656423e-07, "loss": 5.0368, "step": 24728 }, { "epoch": 0.8713760934485584, "grad_norm": 3.526636838912964, "learning_rate": 8.690692636183462e-07, "loss": 2.578, "step": 24729 }, { "epoch": 0.8714113304615168, "grad_norm": 2.532254695892334, "learning_rate": 8.686039661025869e-07, "loss": 2.4251, "step": 24730 }, { "epoch": 0.8714465674744752, "grad_norm": 1.400463581085205, "learning_rate": 8.68138787524424e-07, "loss": 0.8881, "step": 24731 }, { "epoch": 0.8714818044874336, "grad_norm": 3.0404229164123535, "learning_rate": 8.676737278899183e-07, "loss": 2.1543, "step": 24732 }, { "epoch": 0.871517041500392, "grad_norm": 5.52439022064209, "learning_rate": 8.672087872051216e-07, "loss": 3.5615, "step": 24733 }, { "epoch": 0.8715522785133504, "grad_norm": 3.644392728805542, "learning_rate": 8.667439654760956e-07, "loss": 3.1555, "step": 24734 }, { "epoch": 0.8715875155263089, "grad_norm": 5.782275199890137, "learning_rate": 8.662792627088957e-07, "loss": 2.9077, "step": 24735 }, { "epoch": 0.8716227525392672, "grad_norm": 4.16356897354126, "learning_rate": 8.65814678909569e-07, "loss": 2.5516, "step": 24736 }, { "epoch": 0.8716579895522256, "grad_norm": 4.632294654846191, "learning_rate": 8.653502140841674e-07, "loss": 3.0209, "step": 24737 }, { "epoch": 0.8716932265651841, "grad_norm": 4.586246967315674, "learning_rate": 8.648858682387473e-07, "loss": 4.5344, "step": 24738 }, { "epoch": 0.8717284635781425, "grad_norm": 3.165295124053955, "learning_rate": 8.644216413793505e-07, "loss": 2.0804, "step": 24739 }, { "epoch": 0.8717637005911009, "grad_norm": 3.0655229091644287, "learning_rate": 8.639575335120254e-07, "loss": 2.6672, "step": 24740 }, { "epoch": 0.8717989376040594, "grad_norm": 3.3924641609191895, "learning_rate": 8.63493544642815e-07, "loss": 2.5202, "step": 24741 }, { "epoch": 0.8718341746170177, "grad_norm": 1.1076035499572754, "learning_rate": 8.630296747777688e-07, "loss": 1.2452, "step": 24742 }, { "epoch": 0.8718694116299761, "grad_norm": 8.521435737609863, "learning_rate": 8.625659239229233e-07, "loss": 8.4992, "step": 24743 }, { "epoch": 0.8719046486429345, "grad_norm": 1.5298064947128296, "learning_rate": 8.621022920843203e-07, "loss": 0.8933, "step": 24744 }, { "epoch": 0.871939885655893, "grad_norm": 6.0316619873046875, "learning_rate": 8.616387792679992e-07, "loss": 5.0935, "step": 24745 }, { "epoch": 0.8719751226688514, "grad_norm": 5.796063423156738, "learning_rate": 8.611753854799976e-07, "loss": 3.1194, "step": 24746 }, { "epoch": 0.8720103596818097, "grad_norm": 1.3189365863800049, "learning_rate": 8.607121107263505e-07, "loss": 0.7974, "step": 24747 }, { "epoch": 0.8720455966947682, "grad_norm": 1.3644944429397583, "learning_rate": 8.602489550130911e-07, "loss": 1.1246, "step": 24748 }, { "epoch": 0.8720808337077266, "grad_norm": 3.3455300331115723, "learning_rate": 8.597859183462542e-07, "loss": 2.7051, "step": 24749 }, { "epoch": 0.872116070720685, "grad_norm": 1.2385390996932983, "learning_rate": 8.593230007318698e-07, "loss": 0.7044, "step": 24750 }, { "epoch": 0.8721513077336435, "grad_norm": 3.7256314754486084, "learning_rate": 8.588602021759695e-07, "loss": 2.7933, "step": 24751 }, { "epoch": 0.8721865447466018, "grad_norm": 4.740683555603027, "learning_rate": 8.583975226845764e-07, "loss": 3.3757, "step": 24752 }, { "epoch": 0.8722217817595602, "grad_norm": 6.636411666870117, "learning_rate": 8.579349622637212e-07, "loss": 4.7976, "step": 24753 }, { "epoch": 0.8722570187725186, "grad_norm": 2.5025699138641357, "learning_rate": 8.574725209194301e-07, "loss": 2.7515, "step": 24754 }, { "epoch": 0.8722922557854771, "grad_norm": 0.9984577298164368, "learning_rate": 8.570101986577218e-07, "loss": 0.9593, "step": 24755 }, { "epoch": 0.8723274927984355, "grad_norm": 1.2711060047149658, "learning_rate": 8.565479954846178e-07, "loss": 0.9929, "step": 24756 }, { "epoch": 0.8723627298113938, "grad_norm": 4.949714183807373, "learning_rate": 8.560859114061426e-07, "loss": 5.1803, "step": 24757 }, { "epoch": 0.8723979668243523, "grad_norm": 4.949714183807373, "learning_rate": 8.560859114061426e-07, "loss": 2.5296, "step": 24758 }, { "epoch": 0.8724332038373107, "grad_norm": 2.627236843109131, "learning_rate": 8.556239464283156e-07, "loss": 2.8469, "step": 24759 }, { "epoch": 0.8724684408502691, "grad_norm": 2.2879438400268555, "learning_rate": 8.551621005571497e-07, "loss": 2.5196, "step": 24760 }, { "epoch": 0.8725036778632276, "grad_norm": 1.547912836074829, "learning_rate": 8.547003737986615e-07, "loss": 0.7572, "step": 24761 }, { "epoch": 0.872538914876186, "grad_norm": 5.029482841491699, "learning_rate": 8.54238766158868e-07, "loss": 2.7901, "step": 24762 }, { "epoch": 0.8725741518891443, "grad_norm": 2.079960584640503, "learning_rate": 8.53777277643778e-07, "loss": 0.8499, "step": 24763 }, { "epoch": 0.8726093889021028, "grad_norm": 1.5320452451705933, "learning_rate": 8.533159082594044e-07, "loss": 0.6248, "step": 24764 }, { "epoch": 0.8726446259150612, "grad_norm": 3.502159833908081, "learning_rate": 8.528546580117558e-07, "loss": 2.8717, "step": 24765 }, { "epoch": 0.8726798629280196, "grad_norm": 6.100521564483643, "learning_rate": 8.523935269068407e-07, "loss": 4.8925, "step": 24766 }, { "epoch": 0.872715099940978, "grad_norm": 2.171112060546875, "learning_rate": 8.519325149506653e-07, "loss": 0.813, "step": 24767 }, { "epoch": 0.8727503369539364, "grad_norm": 2.0207605361938477, "learning_rate": 8.514716221492336e-07, "loss": 0.8521, "step": 24768 }, { "epoch": 0.8727855739668948, "grad_norm": 6.587234973907471, "learning_rate": 8.5101084850855e-07, "loss": 4.6024, "step": 24769 }, { "epoch": 0.8728208109798532, "grad_norm": 2.8939101696014404, "learning_rate": 8.505501940346151e-07, "loss": 2.4571, "step": 24770 }, { "epoch": 0.8728560479928117, "grad_norm": 1.1272883415222168, "learning_rate": 8.500896587334307e-07, "loss": 1.1801, "step": 24771 }, { "epoch": 0.8728912850057701, "grad_norm": 6.220097064971924, "learning_rate": 8.496292426109909e-07, "loss": 2.9867, "step": 24772 }, { "epoch": 0.8729265220187284, "grad_norm": 4.7338151931762695, "learning_rate": 8.491689456732977e-07, "loss": 3.1878, "step": 24773 }, { "epoch": 0.8729617590316869, "grad_norm": 1.2677456140518188, "learning_rate": 8.487087679263461e-07, "loss": 0.7605, "step": 24774 }, { "epoch": 0.8729969960446453, "grad_norm": 3.9210662841796875, "learning_rate": 8.482487093761272e-07, "loss": 2.7345, "step": 24775 }, { "epoch": 0.8730322330576037, "grad_norm": 1.1064772605895996, "learning_rate": 8.477887700286324e-07, "loss": 0.892, "step": 24776 }, { "epoch": 0.8730674700705621, "grad_norm": 2.0574517250061035, "learning_rate": 8.473289498898573e-07, "loss": 0.7634, "step": 24777 }, { "epoch": 0.8731027070835206, "grad_norm": 1.6153526306152344, "learning_rate": 8.468692489657892e-07, "loss": 0.8412, "step": 24778 }, { "epoch": 0.8731379440964789, "grad_norm": 1.2414121627807617, "learning_rate": 8.464096672624133e-07, "loss": 1.0662, "step": 24779 }, { "epoch": 0.8731731811094373, "grad_norm": 1.1988134384155273, "learning_rate": 8.459502047857159e-07, "loss": 0.8863, "step": 24780 }, { "epoch": 0.8732084181223958, "grad_norm": 5.146242618560791, "learning_rate": 8.454908615416879e-07, "loss": 2.7927, "step": 24781 }, { "epoch": 0.8732436551353542, "grad_norm": 1.1837873458862305, "learning_rate": 8.450316375363044e-07, "loss": 0.7444, "step": 24782 }, { "epoch": 0.8732788921483126, "grad_norm": 5.030573844909668, "learning_rate": 8.445725327755505e-07, "loss": 2.4692, "step": 24783 }, { "epoch": 0.873314129161271, "grad_norm": 4.541416168212891, "learning_rate": 8.44113547265405e-07, "loss": 0.8433, "step": 24784 }, { "epoch": 0.8733493661742294, "grad_norm": 1.2504173517227173, "learning_rate": 8.436546810118473e-07, "loss": 1.0253, "step": 24785 }, { "epoch": 0.8733846031871878, "grad_norm": 1.6490815877914429, "learning_rate": 8.431959340208529e-07, "loss": 0.8635, "step": 24786 }, { "epoch": 0.8734198402001462, "grad_norm": 3.166935920715332, "learning_rate": 8.42737306298399e-07, "loss": 0.8914, "step": 24787 }, { "epoch": 0.8734550772131047, "grad_norm": 1.2265211343765259, "learning_rate": 8.422787978504576e-07, "loss": 1.1212, "step": 24788 }, { "epoch": 0.873490314226063, "grad_norm": 6.019961357116699, "learning_rate": 8.418204086830018e-07, "loss": 2.2976, "step": 24789 }, { "epoch": 0.8735255512390214, "grad_norm": 1.486566424369812, "learning_rate": 8.413621388020033e-07, "loss": 0.9152, "step": 24790 }, { "epoch": 0.8735607882519799, "grad_norm": 4.471142768859863, "learning_rate": 8.409039882134251e-07, "loss": 4.9303, "step": 24791 }, { "epoch": 0.8735960252649383, "grad_norm": 4.900449275970459, "learning_rate": 8.404459569232415e-07, "loss": 3.0828, "step": 24792 }, { "epoch": 0.8736312622778967, "grad_norm": 5.653690338134766, "learning_rate": 8.399880449374187e-07, "loss": 5.8007, "step": 24793 }, { "epoch": 0.8736664992908552, "grad_norm": 2.3158364295959473, "learning_rate": 8.395302522619153e-07, "loss": 0.8693, "step": 24794 }, { "epoch": 0.8737017363038135, "grad_norm": 1.4959537982940674, "learning_rate": 8.390725789026966e-07, "loss": 0.6774, "step": 24795 }, { "epoch": 0.8737369733167719, "grad_norm": 6.425935745239258, "learning_rate": 8.386150248657255e-07, "loss": 2.9658, "step": 24796 }, { "epoch": 0.8737722103297304, "grad_norm": 1.1283072233200073, "learning_rate": 8.381575901569627e-07, "loss": 0.7565, "step": 24797 }, { "epoch": 0.8738074473426888, "grad_norm": 1.0598094463348389, "learning_rate": 8.377002747823626e-07, "loss": 0.7617, "step": 24798 }, { "epoch": 0.8738426843556472, "grad_norm": 1.1542526483535767, "learning_rate": 8.372430787478814e-07, "loss": 0.8255, "step": 24799 }, { "epoch": 0.8738779213686055, "grad_norm": 2.402115821838379, "learning_rate": 8.367860020594809e-07, "loss": 0.9546, "step": 24800 }, { "epoch": 0.873913158381564, "grad_norm": 1.6529675722122192, "learning_rate": 8.363290447231065e-07, "loss": 0.6909, "step": 24801 }, { "epoch": 0.8739483953945224, "grad_norm": 1.0759602785110474, "learning_rate": 8.358722067447144e-07, "loss": 0.7959, "step": 24802 }, { "epoch": 0.8739836324074808, "grad_norm": 4.014743804931641, "learning_rate": 8.354154881302545e-07, "loss": 2.9266, "step": 24803 }, { "epoch": 0.8740188694204393, "grad_norm": 12.723361015319824, "learning_rate": 8.349588888856752e-07, "loss": 6.9773, "step": 24804 }, { "epoch": 0.8740541064333976, "grad_norm": 5.455953598022461, "learning_rate": 8.345024090169229e-07, "loss": 2.8761, "step": 24805 }, { "epoch": 0.874089343446356, "grad_norm": 4.128818035125732, "learning_rate": 8.34046048529945e-07, "loss": 2.6249, "step": 24806 }, { "epoch": 0.8741245804593145, "grad_norm": 0.876013457775116, "learning_rate": 8.335898074306847e-07, "loss": 0.9053, "step": 24807 }, { "epoch": 0.8741598174722729, "grad_norm": 8.325935363769531, "learning_rate": 8.331336857250849e-07, "loss": 4.6715, "step": 24808 }, { "epoch": 0.8741950544852313, "grad_norm": 1.0165692567825317, "learning_rate": 8.326776834190875e-07, "loss": 1.1063, "step": 24809 }, { "epoch": 0.8742302914981896, "grad_norm": 1.1365643739700317, "learning_rate": 8.322218005186289e-07, "loss": 0.6481, "step": 24810 }, { "epoch": 0.8742655285111481, "grad_norm": 7.257049083709717, "learning_rate": 8.317660370296498e-07, "loss": 5.2605, "step": 24811 }, { "epoch": 0.8743007655241065, "grad_norm": 1.2113444805145264, "learning_rate": 8.313103929580857e-07, "loss": 1.0243, "step": 24812 }, { "epoch": 0.8743360025370649, "grad_norm": 1.2178730964660645, "learning_rate": 8.308548683098738e-07, "loss": 1.1366, "step": 24813 }, { "epoch": 0.8743712395500234, "grad_norm": 3.0336403846740723, "learning_rate": 8.303994630909418e-07, "loss": 0.7801, "step": 24814 }, { "epoch": 0.8744064765629818, "grad_norm": 1.0253843069076538, "learning_rate": 8.299441773072269e-07, "loss": 0.8293, "step": 24815 }, { "epoch": 0.8744417135759401, "grad_norm": 3.0370595455169678, "learning_rate": 8.294890109646591e-07, "loss": 2.9953, "step": 24816 }, { "epoch": 0.8744769505888986, "grad_norm": 8.947446823120117, "learning_rate": 8.290339640691624e-07, "loss": 4.9469, "step": 24817 }, { "epoch": 0.874512187601857, "grad_norm": 6.44490909576416, "learning_rate": 8.285790366266655e-07, "loss": 8.9897, "step": 24818 }, { "epoch": 0.8745474246148154, "grad_norm": 4.622852802276611, "learning_rate": 8.28124228643098e-07, "loss": 3.4183, "step": 24819 }, { "epoch": 0.8745826616277738, "grad_norm": 1.2807488441467285, "learning_rate": 8.276695401243784e-07, "loss": 0.9298, "step": 24820 }, { "epoch": 0.8746178986407322, "grad_norm": 1.1334583759307861, "learning_rate": 8.272149710764321e-07, "loss": 0.8354, "step": 24821 }, { "epoch": 0.8746531356536906, "grad_norm": 3.508150815963745, "learning_rate": 8.267605215051799e-07, "loss": 3.1708, "step": 24822 }, { "epoch": 0.874688372666649, "grad_norm": 6.8070597648620605, "learning_rate": 8.263061914165393e-07, "loss": 5.3566, "step": 24823 }, { "epoch": 0.8747236096796075, "grad_norm": 3.3082940578460693, "learning_rate": 8.258519808164301e-07, "loss": 2.6828, "step": 24824 }, { "epoch": 0.8747588466925659, "grad_norm": 5.416910171508789, "learning_rate": 8.253978897107661e-07, "loss": 5.0857, "step": 24825 }, { "epoch": 0.8747940837055243, "grad_norm": 7.670494079589844, "learning_rate": 8.249439181054642e-07, "loss": 5.3231, "step": 24826 }, { "epoch": 0.8748293207184827, "grad_norm": 4.01006555557251, "learning_rate": 8.244900660064359e-07, "loss": 3.9991, "step": 24827 }, { "epoch": 0.8748645577314411, "grad_norm": 4.786919593811035, "learning_rate": 8.240363334195955e-07, "loss": 3.1179, "step": 24828 }, { "epoch": 0.8748997947443995, "grad_norm": 4.319292068481445, "learning_rate": 8.235827203508473e-07, "loss": 1.061, "step": 24829 }, { "epoch": 0.874935031757358, "grad_norm": 1.7505968809127808, "learning_rate": 8.231292268061041e-07, "loss": 0.8483, "step": 24830 }, { "epoch": 0.8749702687703164, "grad_norm": 1.451941967010498, "learning_rate": 8.226758527912726e-07, "loss": 0.9042, "step": 24831 }, { "epoch": 0.8750055057832747, "grad_norm": 6.427565574645996, "learning_rate": 8.22222598312259e-07, "loss": 4.5427, "step": 24832 }, { "epoch": 0.8750407427962331, "grad_norm": 1.1076055765151978, "learning_rate": 8.217694633749606e-07, "loss": 0.8648, "step": 24833 }, { "epoch": 0.8750759798091916, "grad_norm": 3.1008248329162598, "learning_rate": 8.213164479852864e-07, "loss": 2.542, "step": 24834 }, { "epoch": 0.87511121682215, "grad_norm": 3.516634702682495, "learning_rate": 8.208635521491359e-07, "loss": 2.3851, "step": 24835 }, { "epoch": 0.8751464538351084, "grad_norm": 1.2180025577545166, "learning_rate": 8.204107758724055e-07, "loss": 0.8357, "step": 24836 }, { "epoch": 0.8751816908480669, "grad_norm": 1.4347854852676392, "learning_rate": 8.199581191609917e-07, "loss": 0.7039, "step": 24837 }, { "epoch": 0.8752169278610252, "grad_norm": 0.9254647493362427, "learning_rate": 8.195055820207976e-07, "loss": 0.7724, "step": 24838 }, { "epoch": 0.8752521648739836, "grad_norm": 1.383983850479126, "learning_rate": 8.190531644577093e-07, "loss": 0.7581, "step": 24839 }, { "epoch": 0.8752874018869421, "grad_norm": 5.75377893447876, "learning_rate": 8.186008664776246e-07, "loss": 4.8662, "step": 24840 }, { "epoch": 0.8753226388999005, "grad_norm": 2.21970796585083, "learning_rate": 8.18148688086432e-07, "loss": 2.3106, "step": 24841 }, { "epoch": 0.8753578759128589, "grad_norm": 1.7699778079986572, "learning_rate": 8.176966292900224e-07, "loss": 0.9419, "step": 24842 }, { "epoch": 0.8753931129258172, "grad_norm": 4.285924911499023, "learning_rate": 8.172446900942843e-07, "loss": 2.8609, "step": 24843 }, { "epoch": 0.8754283499387757, "grad_norm": 2.9319300651550293, "learning_rate": 8.167928705051042e-07, "loss": 2.9599, "step": 24844 }, { "epoch": 0.8754635869517341, "grad_norm": 3.136796474456787, "learning_rate": 8.163411705283664e-07, "loss": 2.377, "step": 24845 }, { "epoch": 0.8754988239646925, "grad_norm": 1.3586596250534058, "learning_rate": 8.158895901699537e-07, "loss": 0.9236, "step": 24846 }, { "epoch": 0.875534060977651, "grad_norm": 1.0440479516983032, "learning_rate": 8.154381294357516e-07, "loss": 0.9968, "step": 24847 }, { "epoch": 0.8755692979906093, "grad_norm": 4.319094181060791, "learning_rate": 8.149867883316342e-07, "loss": 4.7047, "step": 24848 }, { "epoch": 0.8756045350035677, "grad_norm": 5.095163822174072, "learning_rate": 8.145355668634858e-07, "loss": 4.1842, "step": 24849 }, { "epoch": 0.8756397720165262, "grad_norm": 1.540149211883545, "learning_rate": 8.140844650371815e-07, "loss": 1.1175, "step": 24850 }, { "epoch": 0.8756750090294846, "grad_norm": 4.677753448486328, "learning_rate": 8.13633482858599e-07, "loss": 2.6711, "step": 24851 }, { "epoch": 0.875710246042443, "grad_norm": 2.05497670173645, "learning_rate": 8.131826203336079e-07, "loss": 0.874, "step": 24852 }, { "epoch": 0.8757454830554013, "grad_norm": 1.633425235748291, "learning_rate": 8.127318774680826e-07, "loss": 0.8643, "step": 24853 }, { "epoch": 0.8757807200683598, "grad_norm": 1.1993252038955688, "learning_rate": 8.122812542678982e-07, "loss": 0.875, "step": 24854 }, { "epoch": 0.8758159570813182, "grad_norm": 3.011394739151001, "learning_rate": 8.118307507389189e-07, "loss": 3.1401, "step": 24855 }, { "epoch": 0.8758511940942766, "grad_norm": 4.126911163330078, "learning_rate": 8.113803668870146e-07, "loss": 5.1554, "step": 24856 }, { "epoch": 0.8758864311072351, "grad_norm": 3.899113893508911, "learning_rate": 8.109301027180516e-07, "loss": 3.0329, "step": 24857 }, { "epoch": 0.8759216681201935, "grad_norm": 1.7286828756332397, "learning_rate": 8.104799582378941e-07, "loss": 0.7049, "step": 24858 }, { "epoch": 0.8759569051331518, "grad_norm": 1.139479398727417, "learning_rate": 8.100299334524053e-07, "loss": 0.682, "step": 24859 }, { "epoch": 0.8759921421461103, "grad_norm": 3.9152700901031494, "learning_rate": 8.095800283674482e-07, "loss": 3.5497, "step": 24860 }, { "epoch": 0.8760273791590687, "grad_norm": 4.488125801086426, "learning_rate": 8.091302429888814e-07, "loss": 2.8568, "step": 24861 }, { "epoch": 0.8760626161720271, "grad_norm": 5.029369831085205, "learning_rate": 8.086805773225648e-07, "loss": 4.6212, "step": 24862 }, { "epoch": 0.8760978531849855, "grad_norm": 4.660384654998779, "learning_rate": 8.082310313743546e-07, "loss": 3.0936, "step": 24863 }, { "epoch": 0.8761330901979439, "grad_norm": 3.3552138805389404, "learning_rate": 8.077816051501042e-07, "loss": 3.0838, "step": 24864 }, { "epoch": 0.8761683272109023, "grad_norm": 1.2694708108901978, "learning_rate": 8.073322986556697e-07, "loss": 0.8343, "step": 24865 }, { "epoch": 0.8762035642238607, "grad_norm": 2.0059096813201904, "learning_rate": 8.068831118969056e-07, "loss": 0.9708, "step": 24866 }, { "epoch": 0.8762388012368192, "grad_norm": 1.6296230554580688, "learning_rate": 8.064340448796582e-07, "loss": 0.9953, "step": 24867 }, { "epoch": 0.8762740382497776, "grad_norm": 7.619668483734131, "learning_rate": 8.059850976097761e-07, "loss": 5.5082, "step": 24868 }, { "epoch": 0.8763092752627359, "grad_norm": 33.86717987060547, "learning_rate": 8.055362700931113e-07, "loss": 4.5933, "step": 24869 }, { "epoch": 0.8763445122756944, "grad_norm": 2.7960915565490723, "learning_rate": 8.050875623355092e-07, "loss": 0.7809, "step": 24870 }, { "epoch": 0.8763797492886528, "grad_norm": 4.666399955749512, "learning_rate": 8.046389743428118e-07, "loss": 4.6792, "step": 24871 }, { "epoch": 0.8764149863016112, "grad_norm": 0.7824236750602722, "learning_rate": 8.04190506120861e-07, "loss": 0.9984, "step": 24872 }, { "epoch": 0.8764502233145697, "grad_norm": 2.8442909717559814, "learning_rate": 8.037421576755033e-07, "loss": 2.2741, "step": 24873 }, { "epoch": 0.876485460327528, "grad_norm": 3.3238747119903564, "learning_rate": 8.032939290125741e-07, "loss": 3.0136, "step": 24874 }, { "epoch": 0.8765206973404864, "grad_norm": 3.64907169342041, "learning_rate": 8.02845820137913e-07, "loss": 3.2085, "step": 24875 }, { "epoch": 0.8765559343534448, "grad_norm": 4.4735565185546875, "learning_rate": 8.023978310573566e-07, "loss": 5.0921, "step": 24876 }, { "epoch": 0.8765911713664033, "grad_norm": 3.404541492462158, "learning_rate": 8.019499617767401e-07, "loss": 2.8023, "step": 24877 }, { "epoch": 0.8766264083793617, "grad_norm": 2.302243947982788, "learning_rate": 8.015022123018968e-07, "loss": 1.0067, "step": 24878 }, { "epoch": 0.87666164539232, "grad_norm": 1.1076829433441162, "learning_rate": 8.010545826386584e-07, "loss": 1.1457, "step": 24879 }, { "epoch": 0.8766968824052785, "grad_norm": 3.2151939868927, "learning_rate": 8.006070727928561e-07, "loss": 2.8647, "step": 24880 }, { "epoch": 0.8767321194182369, "grad_norm": 3.4363389015197754, "learning_rate": 8.001596827703184e-07, "loss": 2.7823, "step": 24881 }, { "epoch": 0.8767673564311953, "grad_norm": 3.7278707027435303, "learning_rate": 7.997124125768751e-07, "loss": 2.9722, "step": 24882 }, { "epoch": 0.8768025934441538, "grad_norm": 7.268178939819336, "learning_rate": 7.992652622183461e-07, "loss": 4.6796, "step": 24883 }, { "epoch": 0.8768378304571122, "grad_norm": 1.987475872039795, "learning_rate": 7.988182317005599e-07, "loss": 0.6436, "step": 24884 }, { "epoch": 0.8768730674700705, "grad_norm": 6.76682710647583, "learning_rate": 7.983713210293387e-07, "loss": 4.5905, "step": 24885 }, { "epoch": 0.8769083044830289, "grad_norm": 6.161757946014404, "learning_rate": 7.979245302105054e-07, "loss": 3.0836, "step": 24886 }, { "epoch": 0.8769435414959874, "grad_norm": 2.8975930213928223, "learning_rate": 7.974778592498733e-07, "loss": 2.2135, "step": 24887 }, { "epoch": 0.8769787785089458, "grad_norm": 1.367856740951538, "learning_rate": 7.970313081532666e-07, "loss": 0.9328, "step": 24888 }, { "epoch": 0.8770140155219042, "grad_norm": 3.851132392883301, "learning_rate": 7.965848769265005e-07, "loss": 3.0887, "step": 24889 }, { "epoch": 0.8770492525348627, "grad_norm": 1.193905234336853, "learning_rate": 7.961385655753861e-07, "loss": 0.8995, "step": 24890 }, { "epoch": 0.877084489547821, "grad_norm": 4.743964195251465, "learning_rate": 7.956923741057388e-07, "loss": 4.2844, "step": 24891 }, { "epoch": 0.8771197265607794, "grad_norm": 1.343416690826416, "learning_rate": 7.952463025233747e-07, "loss": 0.9796, "step": 24892 }, { "epoch": 0.8771549635737379, "grad_norm": 6.082038402557373, "learning_rate": 7.948003508340974e-07, "loss": 3.1103, "step": 24893 }, { "epoch": 0.8771902005866963, "grad_norm": 1.3210608959197998, "learning_rate": 7.943545190437174e-07, "loss": 0.8085, "step": 24894 }, { "epoch": 0.8772254375996547, "grad_norm": 4.040872097015381, "learning_rate": 7.939088071580436e-07, "loss": 2.5987, "step": 24895 }, { "epoch": 0.877260674612613, "grad_norm": 6.0508551597595215, "learning_rate": 7.934632151828803e-07, "loss": 4.1083, "step": 24896 }, { "epoch": 0.8772959116255715, "grad_norm": 2.059396505355835, "learning_rate": 7.930177431240305e-07, "loss": 0.7465, "step": 24897 }, { "epoch": 0.8773311486385299, "grad_norm": 1.1090463399887085, "learning_rate": 7.925723909872973e-07, "loss": 0.9474, "step": 24898 }, { "epoch": 0.8773663856514883, "grad_norm": 1.0640387535095215, "learning_rate": 7.921271587784829e-07, "loss": 0.9515, "step": 24899 }, { "epoch": 0.8774016226644468, "grad_norm": 2.271284818649292, "learning_rate": 7.916820465033837e-07, "loss": 0.7441, "step": 24900 }, { "epoch": 0.8774368596774051, "grad_norm": 8.848127365112305, "learning_rate": 7.912370541678016e-07, "loss": 4.9903, "step": 24901 }, { "epoch": 0.8774720966903635, "grad_norm": 4.882328033447266, "learning_rate": 7.907921817775254e-07, "loss": 2.9166, "step": 24902 }, { "epoch": 0.877507333703322, "grad_norm": 4.399173736572266, "learning_rate": 7.903474293383562e-07, "loss": 4.796, "step": 24903 }, { "epoch": 0.8775425707162804, "grad_norm": 1.0579204559326172, "learning_rate": 7.899027968560857e-07, "loss": 0.8712, "step": 24904 }, { "epoch": 0.8775778077292388, "grad_norm": 1.249281406402588, "learning_rate": 7.89458284336505e-07, "loss": 0.8346, "step": 24905 }, { "epoch": 0.8776130447421973, "grad_norm": 4.942139148712158, "learning_rate": 7.890138917854007e-07, "loss": 4.6274, "step": 24906 }, { "epoch": 0.8776482817551556, "grad_norm": 5.239228248596191, "learning_rate": 7.885696192085656e-07, "loss": 4.7001, "step": 24907 }, { "epoch": 0.877683518768114, "grad_norm": 1.2034947872161865, "learning_rate": 7.881254666117855e-07, "loss": 0.6852, "step": 24908 }, { "epoch": 0.8777187557810724, "grad_norm": 2.9577345848083496, "learning_rate": 7.876814340008432e-07, "loss": 2.9698, "step": 24909 }, { "epoch": 0.8777539927940309, "grad_norm": 3.11674427986145, "learning_rate": 7.87237521381522e-07, "loss": 0.8919, "step": 24910 }, { "epoch": 0.8777892298069893, "grad_norm": 1.3677667379379272, "learning_rate": 7.867937287596094e-07, "loss": 1.0409, "step": 24911 }, { "epoch": 0.8778244668199476, "grad_norm": 3.796234369277954, "learning_rate": 7.863500561408799e-07, "loss": 2.2513, "step": 24912 }, { "epoch": 0.8778597038329061, "grad_norm": 0.9197430610656738, "learning_rate": 7.859065035311142e-07, "loss": 0.786, "step": 24913 }, { "epoch": 0.8778949408458645, "grad_norm": 85.59578704833984, "learning_rate": 7.854630709360911e-07, "loss": 2.9982, "step": 24914 }, { "epoch": 0.8779301778588229, "grad_norm": 0.7961505055427551, "learning_rate": 7.850197583615848e-07, "loss": 0.6983, "step": 24915 }, { "epoch": 0.8779654148717814, "grad_norm": 0.9008117318153381, "learning_rate": 7.845765658133697e-07, "loss": 0.8664, "step": 24916 }, { "epoch": 0.8780006518847397, "grad_norm": 1.049760341644287, "learning_rate": 7.841334932972178e-07, "loss": 0.9928, "step": 24917 }, { "epoch": 0.8780358888976981, "grad_norm": 1.3771004676818848, "learning_rate": 7.836905408189021e-07, "loss": 0.8776, "step": 24918 }, { "epoch": 0.8780711259106565, "grad_norm": 1.1004244089126587, "learning_rate": 7.832477083841916e-07, "loss": 0.728, "step": 24919 }, { "epoch": 0.878106362923615, "grad_norm": 1.1841888427734375, "learning_rate": 7.828049959988549e-07, "loss": 0.8582, "step": 24920 }, { "epoch": 0.8781415999365734, "grad_norm": 1.1814374923706055, "learning_rate": 7.823624036686528e-07, "loss": 1.0189, "step": 24921 }, { "epoch": 0.8781768369495317, "grad_norm": 1.0389012098312378, "learning_rate": 7.819199313993575e-07, "loss": 1.0761, "step": 24922 }, { "epoch": 0.8782120739624902, "grad_norm": 1.0025103092193604, "learning_rate": 7.814775791967288e-07, "loss": 0.8503, "step": 24923 }, { "epoch": 0.8782473109754486, "grad_norm": 1.1667826175689697, "learning_rate": 7.810353470665311e-07, "loss": 0.8181, "step": 24924 }, { "epoch": 0.878282547988407, "grad_norm": 5.1462016105651855, "learning_rate": 7.805932350145185e-07, "loss": 5.0044, "step": 24925 }, { "epoch": 0.8783177850013655, "grad_norm": 0.9782009720802307, "learning_rate": 7.801512430464553e-07, "loss": 0.9275, "step": 24926 }, { "epoch": 0.8783530220143239, "grad_norm": 1.3231106996536255, "learning_rate": 7.797093711680981e-07, "loss": 0.8173, "step": 24927 }, { "epoch": 0.8783882590272822, "grad_norm": 3.207446575164795, "learning_rate": 7.792676193851989e-07, "loss": 2.8658, "step": 24928 }, { "epoch": 0.8784234960402406, "grad_norm": 3.513542413711548, "learning_rate": 7.78825987703512e-07, "loss": 3.0744, "step": 24929 }, { "epoch": 0.8784587330531991, "grad_norm": 6.871156215667725, "learning_rate": 7.78384476128794e-07, "loss": 5.1914, "step": 24930 }, { "epoch": 0.8784939700661575, "grad_norm": 1.3618378639221191, "learning_rate": 7.779430846667924e-07, "loss": 0.734, "step": 24931 }, { "epoch": 0.8785292070791159, "grad_norm": 4.677286624908447, "learning_rate": 7.775018133232559e-07, "loss": 4.014, "step": 24932 }, { "epoch": 0.8785644440920743, "grad_norm": 5.138099193572998, "learning_rate": 7.770606621039333e-07, "loss": 5.0256, "step": 24933 }, { "epoch": 0.8785996811050327, "grad_norm": 7.764570713043213, "learning_rate": 7.766196310145702e-07, "loss": 6.3747, "step": 24934 }, { "epoch": 0.8786349181179911, "grad_norm": 3.794400930404663, "learning_rate": 7.761787200609117e-07, "loss": 3.9807, "step": 24935 }, { "epoch": 0.8786701551309496, "grad_norm": 5.025764465332031, "learning_rate": 7.757379292487e-07, "loss": 3.0357, "step": 24936 }, { "epoch": 0.878705392143908, "grad_norm": 7.43776798248291, "learning_rate": 7.752972585836771e-07, "loss": 2.6513, "step": 24937 }, { "epoch": 0.8787406291568663, "grad_norm": 1.080794095993042, "learning_rate": 7.748567080715819e-07, "loss": 0.6916, "step": 24938 }, { "epoch": 0.8787758661698248, "grad_norm": 0.9746726155281067, "learning_rate": 7.744162777181552e-07, "loss": 0.9268, "step": 24939 }, { "epoch": 0.8788111031827832, "grad_norm": 2.448810577392578, "learning_rate": 7.73975967529128e-07, "loss": 2.5288, "step": 24940 }, { "epoch": 0.8788463401957416, "grad_norm": 2.4656355381011963, "learning_rate": 7.735357775102415e-07, "loss": 1.0125, "step": 24941 }, { "epoch": 0.8788815772087, "grad_norm": 5.29393196105957, "learning_rate": 7.730957076672263e-07, "loss": 3.066, "step": 24942 }, { "epoch": 0.8789168142216585, "grad_norm": 1.0542738437652588, "learning_rate": 7.726557580058158e-07, "loss": 0.7431, "step": 24943 }, { "epoch": 0.8789520512346168, "grad_norm": 1.3169198036193848, "learning_rate": 7.722159285317365e-07, "loss": 0.6997, "step": 24944 }, { "epoch": 0.8789872882475752, "grad_norm": 4.4983367919921875, "learning_rate": 7.717762192507228e-07, "loss": 2.8641, "step": 24945 }, { "epoch": 0.8790225252605337, "grad_norm": 1.1432509422302246, "learning_rate": 7.713366301684999e-07, "loss": 0.9694, "step": 24946 }, { "epoch": 0.8790577622734921, "grad_norm": 1.230187177658081, "learning_rate": 7.708971612907911e-07, "loss": 0.9032, "step": 24947 }, { "epoch": 0.8790929992864505, "grad_norm": 4.264861106872559, "learning_rate": 7.704578126233219e-07, "loss": 3.0618, "step": 24948 }, { "epoch": 0.879128236299409, "grad_norm": 1.16204833984375, "learning_rate": 7.700185841718178e-07, "loss": 0.7113, "step": 24949 }, { "epoch": 0.8791634733123673, "grad_norm": 6.474520206451416, "learning_rate": 7.695794759419961e-07, "loss": 4.4822, "step": 24950 }, { "epoch": 0.8791987103253257, "grad_norm": 4.916584014892578, "learning_rate": 7.69140487939577e-07, "loss": 5.1621, "step": 24951 }, { "epoch": 0.8792339473382841, "grad_norm": 0.9381728768348694, "learning_rate": 7.68701620170279e-07, "loss": 1.0906, "step": 24952 }, { "epoch": 0.8792691843512426, "grad_norm": 1.3059622049331665, "learning_rate": 7.682628726398178e-07, "loss": 0.8496, "step": 24953 }, { "epoch": 0.879304421364201, "grad_norm": 2.146148443222046, "learning_rate": 7.678242453539098e-07, "loss": 1.8972, "step": 24954 }, { "epoch": 0.8793396583771593, "grad_norm": 1.155670166015625, "learning_rate": 7.67385738318266e-07, "loss": 0.8565, "step": 24955 }, { "epoch": 0.8793748953901178, "grad_norm": 2.0040764808654785, "learning_rate": 7.669473515385995e-07, "loss": 2.2972, "step": 24956 }, { "epoch": 0.8794101324030762, "grad_norm": 2.4517951011657715, "learning_rate": 7.665090850206203e-07, "loss": 2.9664, "step": 24957 }, { "epoch": 0.8794453694160346, "grad_norm": 26.388694763183594, "learning_rate": 7.660709387700349e-07, "loss": 4.9312, "step": 24958 }, { "epoch": 0.8794806064289931, "grad_norm": 1.1587871313095093, "learning_rate": 7.656329127925533e-07, "loss": 0.9293, "step": 24959 }, { "epoch": 0.8795158434419514, "grad_norm": 3.4316492080688477, "learning_rate": 7.651950070938786e-07, "loss": 2.8806, "step": 24960 }, { "epoch": 0.8795510804549098, "grad_norm": 3.9702417850494385, "learning_rate": 7.64757221679715e-07, "loss": 2.8305, "step": 24961 }, { "epoch": 0.8795863174678682, "grad_norm": 2.859081268310547, "learning_rate": 7.64319556555767e-07, "loss": 2.518, "step": 24962 }, { "epoch": 0.8796215544808267, "grad_norm": 3.7716708183288574, "learning_rate": 7.638820117277301e-07, "loss": 3.3077, "step": 24963 }, { "epoch": 0.8796567914937851, "grad_norm": 1.1041409969329834, "learning_rate": 7.634445872013085e-07, "loss": 1.0237, "step": 24964 }, { "epoch": 0.8796920285067434, "grad_norm": 4.1238112449646, "learning_rate": 7.630072829821989e-07, "loss": 2.8372, "step": 24965 }, { "epoch": 0.8797272655197019, "grad_norm": 1.0122051239013672, "learning_rate": 7.625700990760942e-07, "loss": 0.9061, "step": 24966 }, { "epoch": 0.8797625025326603, "grad_norm": 0.9987159371376038, "learning_rate": 7.621330354886892e-07, "loss": 0.8091, "step": 24967 }, { "epoch": 0.8797977395456187, "grad_norm": 2.6978535652160645, "learning_rate": 7.616960922256811e-07, "loss": 2.2485, "step": 24968 }, { "epoch": 0.8798329765585772, "grad_norm": 5.122323989868164, "learning_rate": 7.612592692927567e-07, "loss": 4.9265, "step": 24969 }, { "epoch": 0.8798682135715356, "grad_norm": 1.4266360998153687, "learning_rate": 7.608225666956071e-07, "loss": 0.812, "step": 24970 }, { "epoch": 0.8799034505844939, "grad_norm": 1.1698870658874512, "learning_rate": 7.603859844399197e-07, "loss": 0.839, "step": 24971 }, { "epoch": 0.8799386875974524, "grad_norm": 3.83668851852417, "learning_rate": 7.599495225313836e-07, "loss": 3.0697, "step": 24972 }, { "epoch": 0.8799739246104108, "grad_norm": 1.537094235420227, "learning_rate": 7.595131809756806e-07, "loss": 0.7297, "step": 24973 }, { "epoch": 0.8800091616233692, "grad_norm": 1.061466932296753, "learning_rate": 7.590769597784953e-07, "loss": 0.9582, "step": 24974 }, { "epoch": 0.8800443986363276, "grad_norm": 1.1162657737731934, "learning_rate": 7.586408589455108e-07, "loss": 0.8008, "step": 24975 }, { "epoch": 0.880079635649286, "grad_norm": 0.9808439612388611, "learning_rate": 7.582048784824059e-07, "loss": 0.8495, "step": 24976 }, { "epoch": 0.8801148726622444, "grad_norm": 3.6144227981567383, "learning_rate": 7.577690183948594e-07, "loss": 2.9666, "step": 24977 }, { "epoch": 0.8801501096752028, "grad_norm": 4.469106197357178, "learning_rate": 7.57333278688549e-07, "loss": 4.6127, "step": 24978 }, { "epoch": 0.8801853466881613, "grad_norm": 4.602641582489014, "learning_rate": 7.56897659369149e-07, "loss": 4.9109, "step": 24979 }, { "epoch": 0.8802205837011197, "grad_norm": 1.1241519451141357, "learning_rate": 7.56462160442335e-07, "loss": 0.8713, "step": 24980 }, { "epoch": 0.880255820714078, "grad_norm": 4.538400650024414, "learning_rate": 7.560267819137812e-07, "loss": 4.8257, "step": 24981 }, { "epoch": 0.8802910577270365, "grad_norm": 1.5026862621307373, "learning_rate": 7.55591523789152e-07, "loss": 0.7655, "step": 24982 }, { "epoch": 0.8803262947399949, "grad_norm": 1.023398756980896, "learning_rate": 7.551563860741229e-07, "loss": 0.7347, "step": 24983 }, { "epoch": 0.8803615317529533, "grad_norm": 1.0023750066757202, "learning_rate": 7.547213687743604e-07, "loss": 0.6764, "step": 24984 }, { "epoch": 0.8803967687659117, "grad_norm": 1.0627515316009521, "learning_rate": 7.542864718955289e-07, "loss": 0.9219, "step": 24985 }, { "epoch": 0.8804320057788702, "grad_norm": 4.33639669418335, "learning_rate": 7.538516954432907e-07, "loss": 2.8238, "step": 24986 }, { "epoch": 0.8804672427918285, "grad_norm": 4.0730414390563965, "learning_rate": 7.534170394233165e-07, "loss": 2.7249, "step": 24987 }, { "epoch": 0.8805024798047869, "grad_norm": 1.4787201881408691, "learning_rate": 7.529825038412619e-07, "loss": 0.7172, "step": 24988 }, { "epoch": 0.8805377168177454, "grad_norm": 3.613084554672241, "learning_rate": 7.52548088702787e-07, "loss": 4.4473, "step": 24989 }, { "epoch": 0.8805729538307038, "grad_norm": 1.0767875909805298, "learning_rate": 7.521137940135515e-07, "loss": 0.6036, "step": 24990 }, { "epoch": 0.8806081908436622, "grad_norm": 3.5514273643493652, "learning_rate": 7.51679619779212e-07, "loss": 2.6013, "step": 24991 }, { "epoch": 0.8806434278566206, "grad_norm": 1.1686919927597046, "learning_rate": 7.512455660054241e-07, "loss": 0.6213, "step": 24992 }, { "epoch": 0.880678664869579, "grad_norm": 1.041053295135498, "learning_rate": 7.508116326978399e-07, "loss": 0.8218, "step": 24993 }, { "epoch": 0.8807139018825374, "grad_norm": 3.2242188453674316, "learning_rate": 7.503778198621125e-07, "loss": 2.6976, "step": 24994 }, { "epoch": 0.8807491388954958, "grad_norm": 1.0300085544586182, "learning_rate": 7.499441275038922e-07, "loss": 0.9491, "step": 24995 }, { "epoch": 0.8807843759084543, "grad_norm": 1.2564408779144287, "learning_rate": 7.495105556288284e-07, "loss": 0.8957, "step": 24996 }, { "epoch": 0.8808196129214126, "grad_norm": 4.206805229187012, "learning_rate": 7.490771042425671e-07, "loss": 4.789, "step": 24997 }, { "epoch": 0.880854849934371, "grad_norm": 5.838010311126709, "learning_rate": 7.486437733507557e-07, "loss": 4.3825, "step": 24998 }, { "epoch": 0.8808900869473295, "grad_norm": 5.126920223236084, "learning_rate": 7.482105629590375e-07, "loss": 3.1028, "step": 24999 }, { "epoch": 0.8809253239602879, "grad_norm": 3.8837709426879883, "learning_rate": 7.477774730730569e-07, "loss": 2.989, "step": 25000 }, { "epoch": 0.8809605609732463, "grad_norm": 2.258115530014038, "learning_rate": 7.473445036984495e-07, "loss": 0.7609, "step": 25001 }, { "epoch": 0.8809957979862048, "grad_norm": 7.487126350402832, "learning_rate": 7.469116548408605e-07, "loss": 6.4607, "step": 25002 }, { "epoch": 0.8810310349991631, "grad_norm": 2.801914691925049, "learning_rate": 7.464789265059269e-07, "loss": 4.65, "step": 25003 }, { "epoch": 0.8810662720121215, "grad_norm": 1.0395718812942505, "learning_rate": 7.460463186992828e-07, "loss": 0.7728, "step": 25004 }, { "epoch": 0.88110150902508, "grad_norm": 3.3373236656188965, "learning_rate": 7.456138314265626e-07, "loss": 2.6768, "step": 25005 }, { "epoch": 0.8811367460380384, "grad_norm": 2.9194529056549072, "learning_rate": 7.451814646934041e-07, "loss": 3.0128, "step": 25006 }, { "epoch": 0.8811719830509968, "grad_norm": 5.909424781799316, "learning_rate": 7.44749218505435e-07, "loss": 6.6722, "step": 25007 }, { "epoch": 0.8812072200639551, "grad_norm": 3.0515832901000977, "learning_rate": 7.443170928682853e-07, "loss": 2.8314, "step": 25008 }, { "epoch": 0.8812424570769136, "grad_norm": 3.6912357807159424, "learning_rate": 7.438850877875847e-07, "loss": 2.4045, "step": 25009 }, { "epoch": 0.881277694089872, "grad_norm": 3.697079658508301, "learning_rate": 7.434532032689601e-07, "loss": 3.0131, "step": 25010 }, { "epoch": 0.8813129311028304, "grad_norm": 2.6870126724243164, "learning_rate": 7.430214393180368e-07, "loss": 2.4845, "step": 25011 }, { "epoch": 0.8813481681157889, "grad_norm": 0.9718044996261597, "learning_rate": 7.425897959404383e-07, "loss": 0.662, "step": 25012 }, { "epoch": 0.8813834051287472, "grad_norm": 1.2736164331436157, "learning_rate": 7.421582731417864e-07, "loss": 0.987, "step": 25013 }, { "epoch": 0.8814186421417056, "grad_norm": 1.0317777395248413, "learning_rate": 7.417268709277026e-07, "loss": 0.9805, "step": 25014 }, { "epoch": 0.8814538791546641, "grad_norm": 2.1470139026641846, "learning_rate": 7.412955893038054e-07, "loss": 0.5511, "step": 25015 }, { "epoch": 0.8814891161676225, "grad_norm": 2.9289581775665283, "learning_rate": 7.408644282757127e-07, "loss": 3.0883, "step": 25016 }, { "epoch": 0.8815243531805809, "grad_norm": 3.3112056255340576, "learning_rate": 7.404333878490411e-07, "loss": 3.2065, "step": 25017 }, { "epoch": 0.8815595901935392, "grad_norm": 1.3382171392440796, "learning_rate": 7.400024680294027e-07, "loss": 0.9623, "step": 25018 }, { "epoch": 0.8815948272064977, "grad_norm": 2.836134433746338, "learning_rate": 7.395716688224142e-07, "loss": 2.7901, "step": 25019 }, { "epoch": 0.8816300642194561, "grad_norm": 4.659233093261719, "learning_rate": 7.39140990233681e-07, "loss": 3.0107, "step": 25020 }, { "epoch": 0.8816653012324145, "grad_norm": 4.47316837310791, "learning_rate": 7.387104322688177e-07, "loss": 3.189, "step": 25021 }, { "epoch": 0.881700538245373, "grad_norm": 5.928775310516357, "learning_rate": 7.382799949334318e-07, "loss": 5.0151, "step": 25022 }, { "epoch": 0.8817357752583314, "grad_norm": 1.8921152353286743, "learning_rate": 7.378496782331268e-07, "loss": 1.1659, "step": 25023 }, { "epoch": 0.8817710122712897, "grad_norm": 2.7437186241149902, "learning_rate": 7.374194821735093e-07, "loss": 2.5677, "step": 25024 }, { "epoch": 0.8818062492842482, "grad_norm": 1.2285853624343872, "learning_rate": 7.369894067601857e-07, "loss": 0.7397, "step": 25025 }, { "epoch": 0.8818414862972066, "grad_norm": 3.2865447998046875, "learning_rate": 7.365594519987529e-07, "loss": 3.145, "step": 25026 }, { "epoch": 0.881876723310165, "grad_norm": 2.26188325881958, "learning_rate": 7.361296178948129e-07, "loss": 2.3314, "step": 25027 }, { "epoch": 0.8819119603231234, "grad_norm": 4.967344284057617, "learning_rate": 7.356999044539647e-07, "loss": 3.2994, "step": 25028 }, { "epoch": 0.8819471973360818, "grad_norm": 6.804591655731201, "learning_rate": 7.352703116818061e-07, "loss": 4.9002, "step": 25029 }, { "epoch": 0.8819824343490402, "grad_norm": 1.5422260761260986, "learning_rate": 7.348408395839312e-07, "loss": 0.7134, "step": 25030 }, { "epoch": 0.8820176713619986, "grad_norm": 1.4828060865402222, "learning_rate": 7.344114881659348e-07, "loss": 1.2186, "step": 25031 }, { "epoch": 0.8820529083749571, "grad_norm": 1.1060328483581543, "learning_rate": 7.339822574334099e-07, "loss": 0.772, "step": 25032 }, { "epoch": 0.8820881453879155, "grad_norm": 1.2614524364471436, "learning_rate": 7.335531473919455e-07, "loss": 0.9005, "step": 25033 }, { "epoch": 0.8821233824008738, "grad_norm": 1.7089505195617676, "learning_rate": 7.331241580471315e-07, "loss": 0.7254, "step": 25034 }, { "epoch": 0.8821586194138323, "grad_norm": 5.729844093322754, "learning_rate": 7.326952894045558e-07, "loss": 4.9761, "step": 25035 }, { "epoch": 0.8821938564267907, "grad_norm": 7.322956562042236, "learning_rate": 7.32266541469805e-07, "loss": 5.6146, "step": 25036 }, { "epoch": 0.8822290934397491, "grad_norm": 4.088531017303467, "learning_rate": 7.318379142484622e-07, "loss": 2.9336, "step": 25037 }, { "epoch": 0.8822643304527076, "grad_norm": 1.14891517162323, "learning_rate": 7.314094077461131e-07, "loss": 0.9404, "step": 25038 }, { "epoch": 0.882299567465666, "grad_norm": 4.5433807373046875, "learning_rate": 7.309810219683345e-07, "loss": 3.2497, "step": 25039 }, { "epoch": 0.8823348044786243, "grad_norm": 1.5147783756256104, "learning_rate": 7.305527569207093e-07, "loss": 0.7378, "step": 25040 }, { "epoch": 0.8823700414915827, "grad_norm": 1.932578444480896, "learning_rate": 7.30124612608818e-07, "loss": 0.7983, "step": 25041 }, { "epoch": 0.8824052785045412, "grad_norm": 1.1993756294250488, "learning_rate": 7.296965890382324e-07, "loss": 0.7045, "step": 25042 }, { "epoch": 0.8824405155174996, "grad_norm": 3.857454776763916, "learning_rate": 7.29268686214527e-07, "loss": 3.3761, "step": 25043 }, { "epoch": 0.882475752530458, "grad_norm": 3.5174713134765625, "learning_rate": 7.28840904143282e-07, "loss": 2.7827, "step": 25044 }, { "epoch": 0.8825109895434164, "grad_norm": 5.969069004058838, "learning_rate": 7.284132428300639e-07, "loss": 4.8467, "step": 25045 }, { "epoch": 0.8825462265563748, "grad_norm": 1.4775537252426147, "learning_rate": 7.279857022804438e-07, "loss": 0.8013, "step": 25046 }, { "epoch": 0.8825814635693332, "grad_norm": 1.3130992650985718, "learning_rate": 7.275582824999905e-07, "loss": 0.8589, "step": 25047 }, { "epoch": 0.8826167005822917, "grad_norm": 1.6659094095230103, "learning_rate": 7.27130983494273e-07, "loss": 0.7621, "step": 25048 }, { "epoch": 0.8826519375952501, "grad_norm": 1.2242188453674316, "learning_rate": 7.267038052688546e-07, "loss": 0.9905, "step": 25049 }, { "epoch": 0.8826871746082084, "grad_norm": 1.5057393312454224, "learning_rate": 7.262767478292998e-07, "loss": 0.879, "step": 25050 }, { "epoch": 0.8827224116211668, "grad_norm": 4.214473724365234, "learning_rate": 7.258498111811729e-07, "loss": 2.7923, "step": 25051 }, { "epoch": 0.8827576486341253, "grad_norm": 1.1928640604019165, "learning_rate": 7.254229953300329e-07, "loss": 0.6719, "step": 25052 }, { "epoch": 0.8827928856470837, "grad_norm": 5.50192403793335, "learning_rate": 7.249963002814398e-07, "loss": 3.299, "step": 25053 }, { "epoch": 0.8828281226600421, "grad_norm": 3.877293825149536, "learning_rate": 7.245697260409523e-07, "loss": 0.6903, "step": 25054 }, { "epoch": 0.8828633596730006, "grad_norm": 1.3603841066360474, "learning_rate": 7.24143272614124e-07, "loss": 1.0177, "step": 25055 }, { "epoch": 0.8828985966859589, "grad_norm": 2.452153205871582, "learning_rate": 7.237169400065125e-07, "loss": 2.1046, "step": 25056 }, { "epoch": 0.8829338336989173, "grad_norm": 1.213829755783081, "learning_rate": 7.232907282236701e-07, "loss": 0.8495, "step": 25057 }, { "epoch": 0.8829690707118758, "grad_norm": 1.0373245477676392, "learning_rate": 7.228646372711445e-07, "loss": 0.6443, "step": 25058 }, { "epoch": 0.8830043077248342, "grad_norm": 2.793602228164673, "learning_rate": 7.224386671544914e-07, "loss": 2.1632, "step": 25059 }, { "epoch": 0.8830395447377926, "grad_norm": 1.1968969106674194, "learning_rate": 7.220128178792574e-07, "loss": 0.7423, "step": 25060 }, { "epoch": 0.8830747817507509, "grad_norm": 4.214435577392578, "learning_rate": 7.215870894509857e-07, "loss": 2.8802, "step": 25061 }, { "epoch": 0.8831100187637094, "grad_norm": 4.278515815734863, "learning_rate": 7.211614818752233e-07, "loss": 2.7591, "step": 25062 }, { "epoch": 0.8831452557766678, "grad_norm": 1.4640852212905884, "learning_rate": 7.207359951575188e-07, "loss": 0.8023, "step": 25063 }, { "epoch": 0.8831804927896262, "grad_norm": 1.8827561140060425, "learning_rate": 7.203106293034068e-07, "loss": 0.8648, "step": 25064 }, { "epoch": 0.8832157298025847, "grad_norm": 2.772087812423706, "learning_rate": 7.198853843184317e-07, "loss": 2.6902, "step": 25065 }, { "epoch": 0.883250966815543, "grad_norm": 4.890803813934326, "learning_rate": 7.194602602081313e-07, "loss": 4.9352, "step": 25066 }, { "epoch": 0.8832862038285014, "grad_norm": 3.1486692428588867, "learning_rate": 7.190352569780424e-07, "loss": 2.8776, "step": 25067 }, { "epoch": 0.8833214408414599, "grad_norm": 4.874763011932373, "learning_rate": 7.186103746337025e-07, "loss": 4.6884, "step": 25068 }, { "epoch": 0.8833566778544183, "grad_norm": 4.755521774291992, "learning_rate": 7.181856131806441e-07, "loss": 4.077, "step": 25069 }, { "epoch": 0.8833919148673767, "grad_norm": 1.1260420083999634, "learning_rate": 7.177609726244017e-07, "loss": 0.7616, "step": 25070 }, { "epoch": 0.883427151880335, "grad_norm": 1.9597915410995483, "learning_rate": 7.173364529705041e-07, "loss": 0.7163, "step": 25071 }, { "epoch": 0.8834623888932935, "grad_norm": 3.9924063682556152, "learning_rate": 7.169120542244812e-07, "loss": 2.6049, "step": 25072 }, { "epoch": 0.8834976259062519, "grad_norm": 5.705865859985352, "learning_rate": 7.164877763918609e-07, "loss": 2.9902, "step": 25073 }, { "epoch": 0.8835328629192103, "grad_norm": 1.6634804010391235, "learning_rate": 7.160636194781711e-07, "loss": 0.7211, "step": 25074 }, { "epoch": 0.8835680999321688, "grad_norm": 5.463387966156006, "learning_rate": 7.156395834889351e-07, "loss": 2.5919, "step": 25075 }, { "epoch": 0.8836033369451272, "grad_norm": 1.4546536207199097, "learning_rate": 7.152156684296774e-07, "loss": 0.5927, "step": 25076 }, { "epoch": 0.8836385739580855, "grad_norm": 1.0604290962219238, "learning_rate": 7.147918743059146e-07, "loss": 0.8531, "step": 25077 }, { "epoch": 0.883673810971044, "grad_norm": 5.647949695587158, "learning_rate": 7.143682011231723e-07, "loss": 4.7556, "step": 25078 }, { "epoch": 0.8837090479840024, "grad_norm": 3.810960292816162, "learning_rate": 7.139446488869694e-07, "loss": 2.6859, "step": 25079 }, { "epoch": 0.8837442849969608, "grad_norm": 2.5060949325561523, "learning_rate": 7.135212176028173e-07, "loss": 2.0224, "step": 25080 }, { "epoch": 0.8837795220099193, "grad_norm": 2.0778732299804688, "learning_rate": 7.130979072762334e-07, "loss": 0.8729, "step": 25081 }, { "epoch": 0.8838147590228776, "grad_norm": 3.230286121368408, "learning_rate": 7.126747179127358e-07, "loss": 2.2382, "step": 25082 }, { "epoch": 0.883849996035836, "grad_norm": 3.4511427879333496, "learning_rate": 7.122516495178311e-07, "loss": 2.7685, "step": 25083 }, { "epoch": 0.8838852330487944, "grad_norm": 1.1881664991378784, "learning_rate": 7.118287020970316e-07, "loss": 0.7103, "step": 25084 }, { "epoch": 0.8839204700617529, "grad_norm": 0.9995678067207336, "learning_rate": 7.114058756558462e-07, "loss": 1.0358, "step": 25085 }, { "epoch": 0.8839557070747113, "grad_norm": 1.3998687267303467, "learning_rate": 7.109831701997827e-07, "loss": 0.9413, "step": 25086 }, { "epoch": 0.8839909440876697, "grad_norm": 6.475775718688965, "learning_rate": 7.105605857343467e-07, "loss": 2.8226, "step": 25087 }, { "epoch": 0.8840261811006281, "grad_norm": 8.647472381591797, "learning_rate": 7.101381222650427e-07, "loss": 2.9241, "step": 25088 }, { "epoch": 0.8840614181135865, "grad_norm": 1.7274670600891113, "learning_rate": 7.09715779797373e-07, "loss": 0.8055, "step": 25089 }, { "epoch": 0.8840966551265449, "grad_norm": 9.404556274414062, "learning_rate": 7.092935583368376e-07, "loss": 7.347, "step": 25090 }, { "epoch": 0.8841318921395034, "grad_norm": 2.052107810974121, "learning_rate": 7.088714578889377e-07, "loss": 2.4875, "step": 25091 }, { "epoch": 0.8841671291524618, "grad_norm": 1.0236432552337646, "learning_rate": 7.084494784591711e-07, "loss": 0.7476, "step": 25092 }, { "epoch": 0.8842023661654201, "grad_norm": 2.0608320236206055, "learning_rate": 7.080276200530323e-07, "loss": 0.7999, "step": 25093 }, { "epoch": 0.8842376031783785, "grad_norm": 2.537332773208618, "learning_rate": 7.076058826760191e-07, "loss": 2.5487, "step": 25094 }, { "epoch": 0.884272840191337, "grad_norm": 2.4374427795410156, "learning_rate": 7.071842663336225e-07, "loss": 2.7862, "step": 25095 }, { "epoch": 0.8843080772042954, "grad_norm": 3.7061212062835693, "learning_rate": 7.067627710313329e-07, "loss": 2.865, "step": 25096 }, { "epoch": 0.8843433142172538, "grad_norm": 3.696467399597168, "learning_rate": 7.063413967746435e-07, "loss": 2.9171, "step": 25097 }, { "epoch": 0.8843785512302123, "grad_norm": 4.409025192260742, "learning_rate": 7.059201435690421e-07, "loss": 4.6607, "step": 25098 }, { "epoch": 0.8844137882431706, "grad_norm": 7.067605972290039, "learning_rate": 7.054990114200133e-07, "loss": 5.0734, "step": 25099 }, { "epoch": 0.884449025256129, "grad_norm": 4.7718071937561035, "learning_rate": 7.050780003330426e-07, "loss": 4.386, "step": 25100 }, { "epoch": 0.8844842622690875, "grad_norm": 1.2202070951461792, "learning_rate": 7.046571103136179e-07, "loss": 0.9701, "step": 25101 }, { "epoch": 0.8845194992820459, "grad_norm": 4.7714362144470215, "learning_rate": 7.042363413672171e-07, "loss": 2.5199, "step": 25102 }, { "epoch": 0.8845547362950043, "grad_norm": 1.1602522134780884, "learning_rate": 7.038156934993213e-07, "loss": 0.9477, "step": 25103 }, { "epoch": 0.8845899733079626, "grad_norm": 36.06952667236328, "learning_rate": 7.033951667154093e-07, "loss": 2.6347, "step": 25104 }, { "epoch": 0.8846252103209211, "grad_norm": 1.2619400024414062, "learning_rate": 7.029747610209625e-07, "loss": 0.9076, "step": 25105 }, { "epoch": 0.8846604473338795, "grad_norm": 1.3020776510238647, "learning_rate": 7.02554476421452e-07, "loss": 1.0189, "step": 25106 }, { "epoch": 0.8846956843468379, "grad_norm": 1.0358610153198242, "learning_rate": 7.021343129223534e-07, "loss": 0.9131, "step": 25107 }, { "epoch": 0.8847309213597964, "grad_norm": 0.6900345087051392, "learning_rate": 7.017142705291402e-07, "loss": 0.7656, "step": 25108 }, { "epoch": 0.8847661583727547, "grad_norm": 1.7459391355514526, "learning_rate": 7.012943492472824e-07, "loss": 1.0656, "step": 25109 }, { "epoch": 0.8848013953857131, "grad_norm": 3.5069313049316406, "learning_rate": 7.008745490822511e-07, "loss": 2.9619, "step": 25110 }, { "epoch": 0.8848366323986716, "grad_norm": 2.4980218410491943, "learning_rate": 7.004548700395131e-07, "loss": 2.9639, "step": 25111 }, { "epoch": 0.88487186941163, "grad_norm": 0.8798195719718933, "learning_rate": 7.00035312124534e-07, "loss": 0.7578, "step": 25112 }, { "epoch": 0.8849071064245884, "grad_norm": 9.922377586364746, "learning_rate": 6.996158753427806e-07, "loss": 2.7944, "step": 25113 }, { "epoch": 0.8849423434375469, "grad_norm": 4.849852085113525, "learning_rate": 6.991965596997164e-07, "loss": 2.6058, "step": 25114 }, { "epoch": 0.8849775804505052, "grad_norm": 1.396988868713379, "learning_rate": 6.987773652007989e-07, "loss": 0.7409, "step": 25115 }, { "epoch": 0.8850128174634636, "grad_norm": 4.714035511016846, "learning_rate": 6.98358291851492e-07, "loss": 3.0196, "step": 25116 }, { "epoch": 0.885048054476422, "grad_norm": 0.8934104442596436, "learning_rate": 6.979393396572553e-07, "loss": 0.9237, "step": 25117 }, { "epoch": 0.8850832914893805, "grad_norm": 2.5669937133789062, "learning_rate": 6.975205086235415e-07, "loss": 2.9292, "step": 25118 }, { "epoch": 0.8851185285023389, "grad_norm": 1.3332332372665405, "learning_rate": 6.971017987558071e-07, "loss": 0.8768, "step": 25119 }, { "epoch": 0.8851537655152972, "grad_norm": 1.940586805343628, "learning_rate": 6.9668321005951e-07, "loss": 1.0338, "step": 25120 }, { "epoch": 0.8851890025282557, "grad_norm": 5.4647216796875, "learning_rate": 6.96264742540097e-07, "loss": 5.243, "step": 25121 }, { "epoch": 0.8852242395412141, "grad_norm": 2.2011094093322754, "learning_rate": 6.958463962030204e-07, "loss": 0.909, "step": 25122 }, { "epoch": 0.8852594765541725, "grad_norm": 1.9491795301437378, "learning_rate": 6.954281710537292e-07, "loss": 0.9262, "step": 25123 }, { "epoch": 0.885294713567131, "grad_norm": 3.1457858085632324, "learning_rate": 6.950100670976734e-07, "loss": 1.1514, "step": 25124 }, { "epoch": 0.8853299505800893, "grad_norm": 3.419943332672119, "learning_rate": 6.945920843402953e-07, "loss": 2.8776, "step": 25125 }, { "epoch": 0.8853651875930477, "grad_norm": 37.8410758972168, "learning_rate": 6.941742227870418e-07, "loss": 6.4512, "step": 25126 }, { "epoch": 0.8854004246060061, "grad_norm": 6.1593194007873535, "learning_rate": 6.93756482443353e-07, "loss": 4.8826, "step": 25127 }, { "epoch": 0.8854356616189646, "grad_norm": 3.9490010738372803, "learning_rate": 6.93338863314672e-07, "loss": 2.6654, "step": 25128 }, { "epoch": 0.885470898631923, "grad_norm": 1.0308287143707275, "learning_rate": 6.929213654064381e-07, "loss": 0.8879, "step": 25129 }, { "epoch": 0.8855061356448813, "grad_norm": 4.375943183898926, "learning_rate": 6.925039887240892e-07, "loss": 2.7665, "step": 25130 }, { "epoch": 0.8855413726578398, "grad_norm": 1.1244983673095703, "learning_rate": 6.920867332730607e-07, "loss": 0.6788, "step": 25131 }, { "epoch": 0.8855766096707982, "grad_norm": 1.1057342290878296, "learning_rate": 6.916695990587885e-07, "loss": 0.8627, "step": 25132 }, { "epoch": 0.8856118466837566, "grad_norm": 0.9777003526687622, "learning_rate": 6.91252586086707e-07, "loss": 1.0058, "step": 25133 }, { "epoch": 0.8856470836967151, "grad_norm": 2.838376522064209, "learning_rate": 6.90835694362243e-07, "loss": 2.1388, "step": 25134 }, { "epoch": 0.8856823207096735, "grad_norm": 5.064689636230469, "learning_rate": 6.90418923890831e-07, "loss": 2.9242, "step": 25135 }, { "epoch": 0.8857175577226318, "grad_norm": 1.2595996856689453, "learning_rate": 6.900022746779e-07, "loss": 0.9161, "step": 25136 }, { "epoch": 0.8857527947355902, "grad_norm": 1.4289913177490234, "learning_rate": 6.895857467288747e-07, "loss": 0.8882, "step": 25137 }, { "epoch": 0.8857880317485487, "grad_norm": 1.267253041267395, "learning_rate": 6.891693400491784e-07, "loss": 0.7249, "step": 25138 }, { "epoch": 0.8858232687615071, "grad_norm": 4.965984344482422, "learning_rate": 6.887530546442411e-07, "loss": 3.2219, "step": 25139 }, { "epoch": 0.8858585057744655, "grad_norm": 2.550488233566284, "learning_rate": 6.883368905194798e-07, "loss": 2.9071, "step": 25140 }, { "epoch": 0.8858937427874239, "grad_norm": 1.0194534063339233, "learning_rate": 6.879208476803168e-07, "loss": 0.9596, "step": 25141 }, { "epoch": 0.8859289798003823, "grad_norm": 0.9570475816726685, "learning_rate": 6.875049261321686e-07, "loss": 0.6205, "step": 25142 }, { "epoch": 0.8859642168133407, "grad_norm": 4.145116329193115, "learning_rate": 6.870891258804591e-07, "loss": 4.922, "step": 25143 }, { "epoch": 0.8859994538262992, "grad_norm": 1.0938535928726196, "learning_rate": 6.86673446930598e-07, "loss": 0.8267, "step": 25144 }, { "epoch": 0.8860346908392576, "grad_norm": 6.108388423919678, "learning_rate": 6.862578892880012e-07, "loss": 5.8195, "step": 25145 }, { "epoch": 0.8860699278522159, "grad_norm": 1.2298455238342285, "learning_rate": 6.85842452958082e-07, "loss": 0.8563, "step": 25146 }, { "epoch": 0.8861051648651744, "grad_norm": 1.220661997795105, "learning_rate": 6.854271379462519e-07, "loss": 0.8844, "step": 25147 }, { "epoch": 0.8861404018781328, "grad_norm": 5.168651103973389, "learning_rate": 6.850119442579195e-07, "loss": 4.1176, "step": 25148 }, { "epoch": 0.8861756388910912, "grad_norm": 3.586491584777832, "learning_rate": 6.84596871898493e-07, "loss": 2.456, "step": 25149 }, { "epoch": 0.8862108759040496, "grad_norm": 4.87523078918457, "learning_rate": 6.84181920873378e-07, "loss": 4.5508, "step": 25150 }, { "epoch": 0.886246112917008, "grad_norm": 1.385867953300476, "learning_rate": 6.837670911879813e-07, "loss": 0.7378, "step": 25151 }, { "epoch": 0.8862813499299664, "grad_norm": 1.312870740890503, "learning_rate": 6.833523828477062e-07, "loss": 0.6699, "step": 25152 }, { "epoch": 0.8863165869429248, "grad_norm": 4.994690895080566, "learning_rate": 6.829377958579498e-07, "loss": 2.9872, "step": 25153 }, { "epoch": 0.8863518239558833, "grad_norm": 3.3070363998413086, "learning_rate": 6.825233302241163e-07, "loss": 2.5672, "step": 25154 }, { "epoch": 0.8863870609688417, "grad_norm": 5.357954502105713, "learning_rate": 6.821089859516061e-07, "loss": 4.9453, "step": 25155 }, { "epoch": 0.8864222979818001, "grad_norm": 1.5929515361785889, "learning_rate": 6.816947630458104e-07, "loss": 0.8114, "step": 25156 }, { "epoch": 0.8864575349947585, "grad_norm": 5.713936805725098, "learning_rate": 6.81280661512127e-07, "loss": 3.2484, "step": 25157 }, { "epoch": 0.8864927720077169, "grad_norm": 7.8215861320495605, "learning_rate": 6.808666813559527e-07, "loss": 5.8106, "step": 25158 }, { "epoch": 0.8865280090206753, "grad_norm": 1.2173566818237305, "learning_rate": 6.804528225826756e-07, "loss": 0.8244, "step": 25159 }, { "epoch": 0.8865632460336337, "grad_norm": 8.975828170776367, "learning_rate": 6.800390851976868e-07, "loss": 3.1104, "step": 25160 }, { "epoch": 0.8865984830465922, "grad_norm": 6.260708332061768, "learning_rate": 6.796254692063764e-07, "loss": 3.1742, "step": 25161 }, { "epoch": 0.8866337200595505, "grad_norm": 0.9933400750160217, "learning_rate": 6.792119746141324e-07, "loss": 0.9689, "step": 25162 }, { "epoch": 0.8866689570725089, "grad_norm": 3.6115410327911377, "learning_rate": 6.787986014263381e-07, "loss": 3.0264, "step": 25163 }, { "epoch": 0.8867041940854674, "grad_norm": 9.791129112243652, "learning_rate": 6.783853496483806e-07, "loss": 2.8988, "step": 25164 }, { "epoch": 0.8867394310984258, "grad_norm": 55.65084457397461, "learning_rate": 6.779722192856408e-07, "loss": 4.5895, "step": 25165 }, { "epoch": 0.8867746681113842, "grad_norm": 1.3209469318389893, "learning_rate": 6.775592103435013e-07, "loss": 0.7604, "step": 25166 }, { "epoch": 0.8868099051243427, "grad_norm": 0.9724174737930298, "learning_rate": 6.77146322827339e-07, "loss": 0.8087, "step": 25167 }, { "epoch": 0.886845142137301, "grad_norm": 7.400904178619385, "learning_rate": 6.767335567425348e-07, "loss": 5.1457, "step": 25168 }, { "epoch": 0.8868803791502594, "grad_norm": 4.5619330406188965, "learning_rate": 6.763209120944625e-07, "loss": 2.8884, "step": 25169 }, { "epoch": 0.8869156161632178, "grad_norm": 1.384417176246643, "learning_rate": 6.759083888884987e-07, "loss": 0.9798, "step": 25170 }, { "epoch": 0.8869508531761763, "grad_norm": 1.6199111938476562, "learning_rate": 6.754959871300171e-07, "loss": 0.8618, "step": 25171 }, { "epoch": 0.8869860901891347, "grad_norm": 4.335423469543457, "learning_rate": 6.750837068243865e-07, "loss": 5.115, "step": 25172 }, { "epoch": 0.887021327202093, "grad_norm": 1.1388334035873413, "learning_rate": 6.74671547976976e-07, "loss": 0.876, "step": 25173 }, { "epoch": 0.8870565642150515, "grad_norm": 3.0992376804351807, "learning_rate": 6.742595105931604e-07, "loss": 3.1158, "step": 25174 }, { "epoch": 0.8870918012280099, "grad_norm": 1.536423683166504, "learning_rate": 6.738475946783018e-07, "loss": 0.8575, "step": 25175 }, { "epoch": 0.8871270382409683, "grad_norm": 4.786410331726074, "learning_rate": 6.73435800237765e-07, "loss": 3.2977, "step": 25176 }, { "epoch": 0.8871622752539268, "grad_norm": 1.9086496829986572, "learning_rate": 6.730241272769145e-07, "loss": 0.81, "step": 25177 }, { "epoch": 0.8871975122668851, "grad_norm": 1.7299436330795288, "learning_rate": 6.726125758011149e-07, "loss": 0.7374, "step": 25178 }, { "epoch": 0.8872327492798435, "grad_norm": 1.2383958101272583, "learning_rate": 6.722011458157229e-07, "loss": 0.8404, "step": 25179 }, { "epoch": 0.887267986292802, "grad_norm": 3.430661916732788, "learning_rate": 6.717898373261e-07, "loss": 4.9629, "step": 25180 }, { "epoch": 0.8873032233057604, "grad_norm": 0.9722358584403992, "learning_rate": 6.713786503376018e-07, "loss": 0.6577, "step": 25181 }, { "epoch": 0.8873384603187188, "grad_norm": 3.662766933441162, "learning_rate": 6.709675848555863e-07, "loss": 2.8916, "step": 25182 }, { "epoch": 0.8873736973316771, "grad_norm": 5.61738920211792, "learning_rate": 6.705566408854047e-07, "loss": 4.5077, "step": 25183 }, { "epoch": 0.8874089343446356, "grad_norm": 0.950700581073761, "learning_rate": 6.701458184324128e-07, "loss": 0.7635, "step": 25184 }, { "epoch": 0.887444171357594, "grad_norm": 1.0544767379760742, "learning_rate": 6.697351175019595e-07, "loss": 1.1617, "step": 25185 }, { "epoch": 0.8874794083705524, "grad_norm": 4.301888942718506, "learning_rate": 6.69324538099394e-07, "loss": 4.629, "step": 25186 }, { "epoch": 0.8875146453835109, "grad_norm": 6.311504364013672, "learning_rate": 6.689140802300675e-07, "loss": 3.0297, "step": 25187 }, { "epoch": 0.8875498823964693, "grad_norm": 5.2605767250061035, "learning_rate": 6.685037438993202e-07, "loss": 7.1269, "step": 25188 }, { "epoch": 0.8875851194094276, "grad_norm": 3.8201870918273926, "learning_rate": 6.680935291125013e-07, "loss": 2.7526, "step": 25189 }, { "epoch": 0.8876203564223861, "grad_norm": 2.715787172317505, "learning_rate": 6.676834358749551e-07, "loss": 2.7713, "step": 25190 }, { "epoch": 0.8876555934353445, "grad_norm": 1.4354029893875122, "learning_rate": 6.672734641920198e-07, "loss": 1.182, "step": 25191 }, { "epoch": 0.8876908304483029, "grad_norm": 4.552226543426514, "learning_rate": 6.668636140690332e-07, "loss": 2.6981, "step": 25192 }, { "epoch": 0.8877260674612613, "grad_norm": 3.4018771648406982, "learning_rate": 6.664538855113412e-07, "loss": 4.8729, "step": 25193 }, { "epoch": 0.8877613044742197, "grad_norm": 3.4617254734039307, "learning_rate": 6.660442785242749e-07, "loss": 2.6565, "step": 25194 }, { "epoch": 0.8877965414871781, "grad_norm": 5.869295120239258, "learning_rate": 6.656347931131701e-07, "loss": 3.1716, "step": 25195 }, { "epoch": 0.8878317785001365, "grad_norm": 7.958419322967529, "learning_rate": 6.652254292833604e-07, "loss": 6.5418, "step": 25196 }, { "epoch": 0.887867015513095, "grad_norm": 11.733964920043945, "learning_rate": 6.648161870401815e-07, "loss": 4.3554, "step": 25197 }, { "epoch": 0.8879022525260534, "grad_norm": 1.115682601928711, "learning_rate": 6.644070663889591e-07, "loss": 1.0841, "step": 25198 }, { "epoch": 0.8879374895390117, "grad_norm": 1.1852151155471802, "learning_rate": 6.639980673350232e-07, "loss": 0.9447, "step": 25199 }, { "epoch": 0.8879727265519702, "grad_norm": 1.475411057472229, "learning_rate": 6.635891898837021e-07, "loss": 0.9733, "step": 25200 }, { "epoch": 0.8880079635649286, "grad_norm": 1.5992906093597412, "learning_rate": 6.631804340403214e-07, "loss": 1.0355, "step": 25201 }, { "epoch": 0.888043200577887, "grad_norm": 1.6114603281021118, "learning_rate": 6.627717998102046e-07, "loss": 1.2136, "step": 25202 }, { "epoch": 0.8880784375908454, "grad_norm": 3.165757656097412, "learning_rate": 6.62363287198674e-07, "loss": 2.6595, "step": 25203 }, { "epoch": 0.8881136746038039, "grad_norm": 1.198384404182434, "learning_rate": 6.619548962110523e-07, "loss": 0.6903, "step": 25204 }, { "epoch": 0.8881489116167622, "grad_norm": 1.206089973449707, "learning_rate": 6.615466268526561e-07, "loss": 0.9547, "step": 25205 }, { "epoch": 0.8881841486297206, "grad_norm": 1.1298068761825562, "learning_rate": 6.611384791288078e-07, "loss": 0.8186, "step": 25206 }, { "epoch": 0.8882193856426791, "grad_norm": 5.128728866577148, "learning_rate": 6.607304530448156e-07, "loss": 5.2352, "step": 25207 }, { "epoch": 0.8882546226556375, "grad_norm": 1.2712156772613525, "learning_rate": 6.603225486060017e-07, "loss": 0.8227, "step": 25208 }, { "epoch": 0.8882898596685959, "grad_norm": 1.8486353158950806, "learning_rate": 6.599147658176774e-07, "loss": 0.7238, "step": 25209 }, { "epoch": 0.8883250966815543, "grad_norm": 3.5878689289093018, "learning_rate": 6.595071046851509e-07, "loss": 2.8774, "step": 25210 }, { "epoch": 0.8883603336945127, "grad_norm": 0.8257008790969849, "learning_rate": 6.590995652137333e-07, "loss": 0.9691, "step": 25211 }, { "epoch": 0.8883955707074711, "grad_norm": 7.682173252105713, "learning_rate": 6.586921474087371e-07, "loss": 3.2643, "step": 25212 }, { "epoch": 0.8884308077204296, "grad_norm": 4.041143417358398, "learning_rate": 6.582848512754625e-07, "loss": 2.4915, "step": 25213 }, { "epoch": 0.888466044733388, "grad_norm": 4.4752655029296875, "learning_rate": 6.578776768192197e-07, "loss": 2.8582, "step": 25214 }, { "epoch": 0.8885012817463463, "grad_norm": 3.8028478622436523, "learning_rate": 6.574706240453066e-07, "loss": 2.5806, "step": 25215 }, { "epoch": 0.8885365187593047, "grad_norm": 0.9636048078536987, "learning_rate": 6.570636929590335e-07, "loss": 0.749, "step": 25216 }, { "epoch": 0.8885717557722632, "grad_norm": 1.3368194103240967, "learning_rate": 6.566568835656928e-07, "loss": 0.7631, "step": 25217 }, { "epoch": 0.8886069927852216, "grad_norm": 4.327532768249512, "learning_rate": 6.56250195870588e-07, "loss": 3.2567, "step": 25218 }, { "epoch": 0.88864222979818, "grad_norm": 2.15885066986084, "learning_rate": 6.558436298790139e-07, "loss": 0.7996, "step": 25219 }, { "epoch": 0.8886774668111385, "grad_norm": 0.9085608124732971, "learning_rate": 6.55437185596266e-07, "loss": 0.9276, "step": 25220 }, { "epoch": 0.8887127038240968, "grad_norm": 1.261903166770935, "learning_rate": 6.550308630276403e-07, "loss": 1.2321, "step": 25221 }, { "epoch": 0.8887479408370552, "grad_norm": 8.4352445602417, "learning_rate": 6.546246621784269e-07, "loss": 2.8955, "step": 25222 }, { "epoch": 0.8887831778500137, "grad_norm": 4.927892208099365, "learning_rate": 6.542185830539183e-07, "loss": 3.182, "step": 25223 }, { "epoch": 0.8888184148629721, "grad_norm": 3.899812698364258, "learning_rate": 6.538126256594024e-07, "loss": 3.383, "step": 25224 }, { "epoch": 0.8888536518759305, "grad_norm": 1.544402003288269, "learning_rate": 6.534067900001706e-07, "loss": 0.9586, "step": 25225 }, { "epoch": 0.8888888888888888, "grad_norm": 1.6691313982009888, "learning_rate": 6.530010760815009e-07, "loss": 0.7889, "step": 25226 }, { "epoch": 0.8889241259018473, "grad_norm": 1.1422832012176514, "learning_rate": 6.525954839086856e-07, "loss": 1.0777, "step": 25227 }, { "epoch": 0.8889593629148057, "grad_norm": 3.7796273231506348, "learning_rate": 6.521900134870063e-07, "loss": 3.048, "step": 25228 }, { "epoch": 0.8889945999277641, "grad_norm": 1.0427987575531006, "learning_rate": 6.517846648217408e-07, "loss": 0.7415, "step": 25229 }, { "epoch": 0.8890298369407226, "grad_norm": 1.1701658964157104, "learning_rate": 6.513794379181703e-07, "loss": 0.8988, "step": 25230 }, { "epoch": 0.889065073953681, "grad_norm": 3.1460719108581543, "learning_rate": 6.509743327815754e-07, "loss": 2.3929, "step": 25231 }, { "epoch": 0.8891003109666393, "grad_norm": 1.9264789819717407, "learning_rate": 6.505693494172294e-07, "loss": 0.8884, "step": 25232 }, { "epoch": 0.8891355479795978, "grad_norm": 6.1706109046936035, "learning_rate": 6.501644878304091e-07, "loss": 7.0849, "step": 25233 }, { "epoch": 0.8891707849925562, "grad_norm": 1.711101770401001, "learning_rate": 6.497597480263851e-07, "loss": 0.8353, "step": 25234 }, { "epoch": 0.8892060220055146, "grad_norm": 1.1005971431732178, "learning_rate": 6.493551300104351e-07, "loss": 0.8988, "step": 25235 }, { "epoch": 0.889241259018473, "grad_norm": 7.119369983673096, "learning_rate": 6.489506337878226e-07, "loss": 5.4436, "step": 25236 }, { "epoch": 0.8892764960314314, "grad_norm": 2.555781126022339, "learning_rate": 6.485462593638203e-07, "loss": 2.3153, "step": 25237 }, { "epoch": 0.8893117330443898, "grad_norm": 8.228707313537598, "learning_rate": 6.48142006743695e-07, "loss": 3.8657, "step": 25238 }, { "epoch": 0.8893469700573482, "grad_norm": 1.002617359161377, "learning_rate": 6.477378759327102e-07, "loss": 0.718, "step": 25239 }, { "epoch": 0.8893822070703067, "grad_norm": 1.6610028743743896, "learning_rate": 6.473338669361306e-07, "loss": 0.6032, "step": 25240 }, { "epoch": 0.8894174440832651, "grad_norm": 0.9534127712249756, "learning_rate": 6.469299797592199e-07, "loss": 0.9397, "step": 25241 }, { "epoch": 0.8894526810962234, "grad_norm": 2.650940418243408, "learning_rate": 6.46526214407237e-07, "loss": 2.4287, "step": 25242 }, { "epoch": 0.8894879181091819, "grad_norm": 4.6997175216674805, "learning_rate": 6.461225708854413e-07, "loss": 2.9713, "step": 25243 }, { "epoch": 0.8895231551221403, "grad_norm": 3.531803607940674, "learning_rate": 6.457190491990927e-07, "loss": 2.5348, "step": 25244 }, { "epoch": 0.8895583921350987, "grad_norm": 4.539517879486084, "learning_rate": 6.453156493534419e-07, "loss": 0.9427, "step": 25245 }, { "epoch": 0.8895936291480572, "grad_norm": 1.2375298738479614, "learning_rate": 6.449123713537475e-07, "loss": 0.7504, "step": 25246 }, { "epoch": 0.8896288661610156, "grad_norm": 1.4308271408081055, "learning_rate": 6.445092152052623e-07, "loss": 0.8586, "step": 25247 }, { "epoch": 0.8896641031739739, "grad_norm": 1.265064001083374, "learning_rate": 6.441061809132354e-07, "loss": 0.7598, "step": 25248 }, { "epoch": 0.8896993401869323, "grad_norm": 2.5114784240722656, "learning_rate": 6.437032684829159e-07, "loss": 0.9168, "step": 25249 }, { "epoch": 0.8897345771998908, "grad_norm": 3.5012831687927246, "learning_rate": 6.43300477919554e-07, "loss": 2.4663, "step": 25250 }, { "epoch": 0.8897698142128492, "grad_norm": 3.263695001602173, "learning_rate": 6.428978092283977e-07, "loss": 2.7641, "step": 25251 }, { "epoch": 0.8898050512258076, "grad_norm": 6.67103385925293, "learning_rate": 6.424952624146863e-07, "loss": 6.5018, "step": 25252 }, { "epoch": 0.889840288238766, "grad_norm": 8.190388679504395, "learning_rate": 6.420928374836655e-07, "loss": 7.0512, "step": 25253 }, { "epoch": 0.8898755252517244, "grad_norm": 8.901877403259277, "learning_rate": 6.41690534440581e-07, "loss": 4.6056, "step": 25254 }, { "epoch": 0.8899107622646828, "grad_norm": 1.1289362907409668, "learning_rate": 6.412883532906677e-07, "loss": 1.0412, "step": 25255 }, { "epoch": 0.8899459992776413, "grad_norm": 2.6591110229492188, "learning_rate": 6.408862940391647e-07, "loss": 2.8692, "step": 25256 }, { "epoch": 0.8899812362905997, "grad_norm": 1.3986047506332397, "learning_rate": 6.404843566913111e-07, "loss": 0.9113, "step": 25257 }, { "epoch": 0.890016473303558, "grad_norm": 3.458611488342285, "learning_rate": 6.400825412523403e-07, "loss": 2.5409, "step": 25258 }, { "epoch": 0.8900517103165164, "grad_norm": 9.189547538757324, "learning_rate": 6.396808477274863e-07, "loss": 2.7384, "step": 25259 }, { "epoch": 0.8900869473294749, "grad_norm": 1.1389590501785278, "learning_rate": 6.392792761219823e-07, "loss": 0.8769, "step": 25260 }, { "epoch": 0.8901221843424333, "grad_norm": 5.580054759979248, "learning_rate": 6.388778264410577e-07, "loss": 5.0985, "step": 25261 }, { "epoch": 0.8901574213553917, "grad_norm": 1.9687505960464478, "learning_rate": 6.384764986899417e-07, "loss": 0.8431, "step": 25262 }, { "epoch": 0.8901926583683502, "grad_norm": 1.1418627500534058, "learning_rate": 6.380752928738632e-07, "loss": 1.0035, "step": 25263 }, { "epoch": 0.8902278953813085, "grad_norm": 1.0319215059280396, "learning_rate": 6.376742089980448e-07, "loss": 0.7358, "step": 25264 }, { "epoch": 0.8902631323942669, "grad_norm": 1.003940463066101, "learning_rate": 6.372732470677123e-07, "loss": 1.3211, "step": 25265 }, { "epoch": 0.8902983694072254, "grad_norm": 3.9278862476348877, "learning_rate": 6.368724070880905e-07, "loss": 2.4753, "step": 25266 }, { "epoch": 0.8903336064201838, "grad_norm": 1.2029385566711426, "learning_rate": 6.364716890643973e-07, "loss": 1.0035, "step": 25267 }, { "epoch": 0.8903688434331422, "grad_norm": 3.000849485397339, "learning_rate": 6.36071093001851e-07, "loss": 2.7591, "step": 25268 }, { "epoch": 0.8904040804461005, "grad_norm": 3.306936502456665, "learning_rate": 6.356706189056728e-07, "loss": 2.7595, "step": 25269 }, { "epoch": 0.890439317459059, "grad_norm": 1.2218626737594604, "learning_rate": 6.352702667810795e-07, "loss": 0.9348, "step": 25270 }, { "epoch": 0.8904745544720174, "grad_norm": 3.6326608657836914, "learning_rate": 6.348700366332816e-07, "loss": 2.8914, "step": 25271 }, { "epoch": 0.8905097914849758, "grad_norm": 1.0932788848876953, "learning_rate": 6.344699284674938e-07, "loss": 0.6547, "step": 25272 }, { "epoch": 0.8905450284979343, "grad_norm": 1.2414602041244507, "learning_rate": 6.340699422889307e-07, "loss": 0.8099, "step": 25273 }, { "epoch": 0.8905802655108926, "grad_norm": 12.248190879821777, "learning_rate": 6.336700781027982e-07, "loss": 4.1738, "step": 25274 }, { "epoch": 0.890615502523851, "grad_norm": 4.628756046295166, "learning_rate": 6.332703359143066e-07, "loss": 3.1982, "step": 25275 }, { "epoch": 0.8906507395368095, "grad_norm": 2.269055128097534, "learning_rate": 6.328707157286618e-07, "loss": 0.7397, "step": 25276 }, { "epoch": 0.8906859765497679, "grad_norm": 1.0774989128112793, "learning_rate": 6.324712175510684e-07, "loss": 0.7769, "step": 25277 }, { "epoch": 0.8907212135627263, "grad_norm": 4.776294231414795, "learning_rate": 6.320718413867311e-07, "loss": 2.3276, "step": 25278 }, { "epoch": 0.8907564505756846, "grad_norm": 4.3664398193359375, "learning_rate": 6.316725872408514e-07, "loss": 3.4366, "step": 25279 }, { "epoch": 0.8907916875886431, "grad_norm": 5.270474433898926, "learning_rate": 6.312734551186295e-07, "loss": 6.5221, "step": 25280 }, { "epoch": 0.8908269246016015, "grad_norm": 0.9898535013198853, "learning_rate": 6.308744450252635e-07, "loss": 0.787, "step": 25281 }, { "epoch": 0.8908621616145599, "grad_norm": 0.8814713954925537, "learning_rate": 6.304755569659537e-07, "loss": 0.9024, "step": 25282 }, { "epoch": 0.8908973986275184, "grad_norm": 4.940859794616699, "learning_rate": 6.300767909458905e-07, "loss": 4.8901, "step": 25283 }, { "epoch": 0.8909326356404768, "grad_norm": 1.2984341382980347, "learning_rate": 6.296781469702717e-07, "loss": 0.7013, "step": 25284 }, { "epoch": 0.8909678726534351, "grad_norm": 1.0095117092132568, "learning_rate": 6.2927962504429e-07, "loss": 0.6641, "step": 25285 }, { "epoch": 0.8910031096663936, "grad_norm": 2.7802560329437256, "learning_rate": 6.288812251731335e-07, "loss": 2.4642, "step": 25286 }, { "epoch": 0.891038346679352, "grad_norm": 1.8079404830932617, "learning_rate": 6.284829473619914e-07, "loss": 0.6625, "step": 25287 }, { "epoch": 0.8910735836923104, "grad_norm": 2.273935317993164, "learning_rate": 6.280847916160537e-07, "loss": 2.7715, "step": 25288 }, { "epoch": 0.8911088207052689, "grad_norm": 4.200405597686768, "learning_rate": 6.276867579405066e-07, "loss": 3.3347, "step": 25289 }, { "epoch": 0.8911440577182272, "grad_norm": 3.7009286880493164, "learning_rate": 6.272888463405324e-07, "loss": 2.9352, "step": 25290 }, { "epoch": 0.8911792947311856, "grad_norm": 6.497670650482178, "learning_rate": 6.268910568213138e-07, "loss": 7.5485, "step": 25291 }, { "epoch": 0.891214531744144, "grad_norm": 3.9003102779388428, "learning_rate": 6.264933893880354e-07, "loss": 3.1932, "step": 25292 }, { "epoch": 0.8912497687571025, "grad_norm": 2.334033966064453, "learning_rate": 6.260958440458731e-07, "loss": 2.4419, "step": 25293 }, { "epoch": 0.8912850057700609, "grad_norm": 2.111678123474121, "learning_rate": 6.256984208000073e-07, "loss": 2.4545, "step": 25294 }, { "epoch": 0.8913202427830192, "grad_norm": 1.945212960243225, "learning_rate": 6.253011196556125e-07, "loss": 0.8587, "step": 25295 }, { "epoch": 0.8913554797959777, "grad_norm": 18.974498748779297, "learning_rate": 6.249039406178659e-07, "loss": 3.5119, "step": 25296 }, { "epoch": 0.8913907168089361, "grad_norm": 0.9810104370117188, "learning_rate": 6.245068836919388e-07, "loss": 0.7694, "step": 25297 }, { "epoch": 0.8914259538218945, "grad_norm": 3.0871200561523438, "learning_rate": 6.241099488830038e-07, "loss": 2.7451, "step": 25298 }, { "epoch": 0.891461190834853, "grad_norm": 3.793806552886963, "learning_rate": 6.237131361962312e-07, "loss": 2.9949, "step": 25299 }, { "epoch": 0.8914964278478114, "grad_norm": 1.2181782722473145, "learning_rate": 6.233164456367901e-07, "loss": 0.8355, "step": 25300 }, { "epoch": 0.8915316648607697, "grad_norm": 4.407889366149902, "learning_rate": 6.229198772098478e-07, "loss": 4.9665, "step": 25301 }, { "epoch": 0.8915669018737281, "grad_norm": 0.845994770526886, "learning_rate": 6.225234309205652e-07, "loss": 0.8707, "step": 25302 }, { "epoch": 0.8916021388866866, "grad_norm": 0.7693307995796204, "learning_rate": 6.221271067741108e-07, "loss": 0.8526, "step": 25303 }, { "epoch": 0.891637375899645, "grad_norm": 1.1730278730392456, "learning_rate": 6.21730904775646e-07, "loss": 0.8357, "step": 25304 }, { "epoch": 0.8916726129126034, "grad_norm": 1.052263855934143, "learning_rate": 6.213348249303308e-07, "loss": 0.8389, "step": 25305 }, { "epoch": 0.8917078499255618, "grad_norm": 4.037184715270996, "learning_rate": 6.209388672433225e-07, "loss": 2.5618, "step": 25306 }, { "epoch": 0.8917430869385202, "grad_norm": 7.231705665588379, "learning_rate": 6.205430317197813e-07, "loss": 3.2082, "step": 25307 }, { "epoch": 0.8917783239514786, "grad_norm": 1.8264034986495972, "learning_rate": 6.201473183648632e-07, "loss": 0.9486, "step": 25308 }, { "epoch": 0.8918135609644371, "grad_norm": 1.9562115669250488, "learning_rate": 6.197517271837206e-07, "loss": 0.7331, "step": 25309 }, { "epoch": 0.8918487979773955, "grad_norm": 3.4925966262817383, "learning_rate": 6.193562581815038e-07, "loss": 2.6826, "step": 25310 }, { "epoch": 0.8918840349903538, "grad_norm": 1.4381030797958374, "learning_rate": 6.189609113633699e-07, "loss": 0.9028, "step": 25311 }, { "epoch": 0.8919192720033122, "grad_norm": 1.4556584358215332, "learning_rate": 6.185656867344647e-07, "loss": 0.8229, "step": 25312 }, { "epoch": 0.8919545090162707, "grad_norm": 2.681084632873535, "learning_rate": 6.181705842999353e-07, "loss": 2.3679, "step": 25313 }, { "epoch": 0.8919897460292291, "grad_norm": 2.7799017429351807, "learning_rate": 6.177756040649297e-07, "loss": 2.6232, "step": 25314 }, { "epoch": 0.8920249830421875, "grad_norm": 1.5303858518600464, "learning_rate": 6.173807460345915e-07, "loss": 0.8831, "step": 25315 }, { "epoch": 0.892060220055146, "grad_norm": 0.8372723460197449, "learning_rate": 6.169860102140645e-07, "loss": 0.869, "step": 25316 }, { "epoch": 0.8920954570681043, "grad_norm": 6.662168979644775, "learning_rate": 6.165913966084891e-07, "loss": 2.8735, "step": 25317 }, { "epoch": 0.8921306940810627, "grad_norm": 1.0263855457305908, "learning_rate": 6.161969052230065e-07, "loss": 0.8367, "step": 25318 }, { "epoch": 0.8921659310940212, "grad_norm": 10.287576675415039, "learning_rate": 6.158025360627551e-07, "loss": 3.4638, "step": 25319 }, { "epoch": 0.8922011681069796, "grad_norm": 0.9739478230476379, "learning_rate": 6.154082891328728e-07, "loss": 0.8663, "step": 25320 }, { "epoch": 0.892236405119938, "grad_norm": 2.4679758548736572, "learning_rate": 6.150141644384888e-07, "loss": 2.8438, "step": 25321 }, { "epoch": 0.8922716421328964, "grad_norm": 3.1651298999786377, "learning_rate": 6.146201619847425e-07, "loss": 3.0722, "step": 25322 }, { "epoch": 0.8923068791458548, "grad_norm": 3.3325750827789307, "learning_rate": 6.142262817767642e-07, "loss": 2.596, "step": 25323 }, { "epoch": 0.8923421161588132, "grad_norm": 6.040975093841553, "learning_rate": 6.138325238196863e-07, "loss": 5.5118, "step": 25324 }, { "epoch": 0.8923773531717716, "grad_norm": 0.9182009100914001, "learning_rate": 6.134388881186315e-07, "loss": 0.9704, "step": 25325 }, { "epoch": 0.8924125901847301, "grad_norm": 1.13070547580719, "learning_rate": 6.130453746787335e-07, "loss": 1.0996, "step": 25326 }, { "epoch": 0.8924478271976884, "grad_norm": 6.242910861968994, "learning_rate": 6.126519835051159e-07, "loss": 2.6881, "step": 25327 }, { "epoch": 0.8924830642106468, "grad_norm": 1.4070825576782227, "learning_rate": 6.122587146029002e-07, "loss": 0.7294, "step": 25328 }, { "epoch": 0.8925183012236053, "grad_norm": 3.5765573978424072, "learning_rate": 6.118655679772089e-07, "loss": 2.8235, "step": 25329 }, { "epoch": 0.8925535382365637, "grad_norm": 5.968368053436279, "learning_rate": 6.114725436331681e-07, "loss": 5.2089, "step": 25330 }, { "epoch": 0.8925887752495221, "grad_norm": 2.4691267013549805, "learning_rate": 6.110796415758913e-07, "loss": 2.8885, "step": 25331 }, { "epoch": 0.8926240122624806, "grad_norm": 1.4608170986175537, "learning_rate": 6.106868618104988e-07, "loss": 0.8059, "step": 25332 }, { "epoch": 0.8926592492754389, "grad_norm": 1.293810486793518, "learning_rate": 6.102942043421056e-07, "loss": 0.7551, "step": 25333 }, { "epoch": 0.8926944862883973, "grad_norm": 4.927329063415527, "learning_rate": 6.099016691758264e-07, "loss": 5.4992, "step": 25334 }, { "epoch": 0.8927297233013557, "grad_norm": 0.8254281282424927, "learning_rate": 6.095092563167738e-07, "loss": 0.8376, "step": 25335 }, { "epoch": 0.8927649603143142, "grad_norm": 2.410930871963501, "learning_rate": 6.091169657700601e-07, "loss": 2.3463, "step": 25336 }, { "epoch": 0.8928001973272726, "grad_norm": 3.522315502166748, "learning_rate": 6.08724797540794e-07, "loss": 2.595, "step": 25337 }, { "epoch": 0.8928354343402309, "grad_norm": 4.26505184173584, "learning_rate": 6.083327516340831e-07, "loss": 3.138, "step": 25338 }, { "epoch": 0.8928706713531894, "grad_norm": 4.208375453948975, "learning_rate": 6.079408280550359e-07, "loss": 2.9094, "step": 25339 }, { "epoch": 0.8929059083661478, "grad_norm": 1.3960864543914795, "learning_rate": 6.075490268087525e-07, "loss": 1.2051, "step": 25340 }, { "epoch": 0.8929411453791062, "grad_norm": 3.13752818107605, "learning_rate": 6.071573479003412e-07, "loss": 3.0241, "step": 25341 }, { "epoch": 0.8929763823920647, "grad_norm": 1.1001461744308472, "learning_rate": 6.067657913349023e-07, "loss": 0.8159, "step": 25342 }, { "epoch": 0.893011619405023, "grad_norm": 4.270552635192871, "learning_rate": 6.063743571175373e-07, "loss": 4.9896, "step": 25343 }, { "epoch": 0.8930468564179814, "grad_norm": 14.530637741088867, "learning_rate": 6.059830452533389e-07, "loss": 5.6138, "step": 25344 }, { "epoch": 0.8930820934309398, "grad_norm": 1.018284559249878, "learning_rate": 6.055918557474105e-07, "loss": 0.7975, "step": 25345 }, { "epoch": 0.8931173304438983, "grad_norm": 2.856275796890259, "learning_rate": 6.05200788604845e-07, "loss": 2.686, "step": 25346 }, { "epoch": 0.8931525674568567, "grad_norm": 1.301701545715332, "learning_rate": 6.048098438307348e-07, "loss": 0.77, "step": 25347 }, { "epoch": 0.893187804469815, "grad_norm": 1.6760106086730957, "learning_rate": 6.044190214301726e-07, "loss": 0.8231, "step": 25348 }, { "epoch": 0.8932230414827735, "grad_norm": 1.3113141059875488, "learning_rate": 6.040283214082521e-07, "loss": 0.9793, "step": 25349 }, { "epoch": 0.8932582784957319, "grad_norm": 0.9077467322349548, "learning_rate": 6.03637743770058e-07, "loss": 0.6615, "step": 25350 }, { "epoch": 0.8932935155086903, "grad_norm": 1.3308155536651611, "learning_rate": 6.032472885206786e-07, "loss": 0.9488, "step": 25351 }, { "epoch": 0.8933287525216488, "grad_norm": 5.03862190246582, "learning_rate": 6.028569556652008e-07, "loss": 3.1965, "step": 25352 }, { "epoch": 0.8933639895346072, "grad_norm": 1.2667078971862793, "learning_rate": 6.024667452087075e-07, "loss": 0.7399, "step": 25353 }, { "epoch": 0.8933992265475655, "grad_norm": 3.497614622116089, "learning_rate": 6.02076657156283e-07, "loss": 1.9045, "step": 25354 }, { "epoch": 0.893434463560524, "grad_norm": 1.036222219467163, "learning_rate": 6.016866915130059e-07, "loss": 1.0538, "step": 25355 }, { "epoch": 0.8934697005734824, "grad_norm": 0.9889105558395386, "learning_rate": 6.012968482839576e-07, "loss": 0.9063, "step": 25356 }, { "epoch": 0.8935049375864408, "grad_norm": 0.9333414435386658, "learning_rate": 6.009071274742151e-07, "loss": 0.8834, "step": 25357 }, { "epoch": 0.8935401745993992, "grad_norm": 2.318549156188965, "learning_rate": 6.005175290888554e-07, "loss": 2.3537, "step": 25358 }, { "epoch": 0.8935754116123577, "grad_norm": 3.239330291748047, "learning_rate": 6.001280531329501e-07, "loss": 2.6131, "step": 25359 }, { "epoch": 0.893610648625316, "grad_norm": 1.2043758630752563, "learning_rate": 5.997386996115751e-07, "loss": 1.0892, "step": 25360 }, { "epoch": 0.8936458856382744, "grad_norm": 3.2301366329193115, "learning_rate": 5.993494685298007e-07, "loss": 2.9119, "step": 25361 }, { "epoch": 0.8936811226512329, "grad_norm": 1.0133470296859741, "learning_rate": 5.989603598926985e-07, "loss": 0.8954, "step": 25362 }, { "epoch": 0.8937163596641913, "grad_norm": 6.020730972290039, "learning_rate": 5.985713737053333e-07, "loss": 6.1304, "step": 25363 }, { "epoch": 0.8937515966771497, "grad_norm": 1.3678175210952759, "learning_rate": 5.981825099727734e-07, "loss": 0.7591, "step": 25364 }, { "epoch": 0.8937868336901081, "grad_norm": 1.4121631383895874, "learning_rate": 5.977937687000868e-07, "loss": 0.7823, "step": 25365 }, { "epoch": 0.8938220707030665, "grad_norm": 4.54152250289917, "learning_rate": 5.974051498923317e-07, "loss": 2.096, "step": 25366 }, { "epoch": 0.8938573077160249, "grad_norm": 4.669122219085693, "learning_rate": 5.970166535545707e-07, "loss": 3.0697, "step": 25367 }, { "epoch": 0.8938925447289833, "grad_norm": 3.09769868850708, "learning_rate": 5.966282796918688e-07, "loss": 2.4169, "step": 25368 }, { "epoch": 0.8939277817419418, "grad_norm": 1.4486395120620728, "learning_rate": 5.962400283092795e-07, "loss": 0.8449, "step": 25369 }, { "epoch": 0.8939630187549001, "grad_norm": 1.0462568998336792, "learning_rate": 5.958518994118622e-07, "loss": 0.6947, "step": 25370 }, { "epoch": 0.8939982557678585, "grad_norm": 3.1872684955596924, "learning_rate": 5.954638930046719e-07, "loss": 2.9818, "step": 25371 }, { "epoch": 0.894033492780817, "grad_norm": 1.1681605577468872, "learning_rate": 5.950760090927609e-07, "loss": 0.7595, "step": 25372 }, { "epoch": 0.8940687297937754, "grad_norm": 1.4830564260482788, "learning_rate": 5.946882476811844e-07, "loss": 0.8032, "step": 25373 }, { "epoch": 0.8941039668067338, "grad_norm": 4.8287739753723145, "learning_rate": 5.943006087749914e-07, "loss": 5.0622, "step": 25374 }, { "epoch": 0.8941392038196923, "grad_norm": 3.0472168922424316, "learning_rate": 5.939130923792302e-07, "loss": 2.5412, "step": 25375 }, { "epoch": 0.8941744408326506, "grad_norm": 3.064976692199707, "learning_rate": 5.935256984989501e-07, "loss": 3.2333, "step": 25376 }, { "epoch": 0.894209677845609, "grad_norm": 1.102177381515503, "learning_rate": 5.93138427139196e-07, "loss": 1.1613, "step": 25377 }, { "epoch": 0.8942449148585674, "grad_norm": 4.311469078063965, "learning_rate": 5.927512783050115e-07, "loss": 4.8733, "step": 25378 }, { "epoch": 0.8942801518715259, "grad_norm": 1.7720392942428589, "learning_rate": 5.923642520014394e-07, "loss": 0.6272, "step": 25379 }, { "epoch": 0.8943153888844843, "grad_norm": 6.464946746826172, "learning_rate": 5.919773482335222e-07, "loss": 3.8185, "step": 25380 }, { "epoch": 0.8943506258974426, "grad_norm": 6.165348529815674, "learning_rate": 5.915905670062993e-07, "loss": 3.0314, "step": 25381 }, { "epoch": 0.8943858629104011, "grad_norm": 4.789240837097168, "learning_rate": 5.912039083248055e-07, "loss": 2.7254, "step": 25382 }, { "epoch": 0.8944210999233595, "grad_norm": 4.251376628875732, "learning_rate": 5.908173721940802e-07, "loss": 3.0533, "step": 25383 }, { "epoch": 0.8944563369363179, "grad_norm": 0.9027122259140015, "learning_rate": 5.904309586191592e-07, "loss": 0.8508, "step": 25384 }, { "epoch": 0.8944915739492764, "grad_norm": 7.1883864402771, "learning_rate": 5.90044667605072e-07, "loss": 2.8948, "step": 25385 }, { "epoch": 0.8945268109622347, "grad_norm": 1.2278056144714355, "learning_rate": 5.896584991568489e-07, "loss": 0.8881, "step": 25386 }, { "epoch": 0.8945620479751931, "grad_norm": 5.307509899139404, "learning_rate": 5.892724532795268e-07, "loss": 5.3632, "step": 25387 }, { "epoch": 0.8945972849881516, "grad_norm": 2.349668025970459, "learning_rate": 5.888865299781276e-07, "loss": 2.5647, "step": 25388 }, { "epoch": 0.89463252200111, "grad_norm": 7.51596212387085, "learning_rate": 5.885007292576805e-07, "loss": 5.7735, "step": 25389 }, { "epoch": 0.8946677590140684, "grad_norm": 2.1562490463256836, "learning_rate": 5.881150511232104e-07, "loss": 0.8116, "step": 25390 }, { "epoch": 0.8947029960270267, "grad_norm": 5.004103660583496, "learning_rate": 5.877294955797407e-07, "loss": 5.4076, "step": 25391 }, { "epoch": 0.8947382330399852, "grad_norm": 1.6882216930389404, "learning_rate": 5.873440626322923e-07, "loss": 0.8215, "step": 25392 }, { "epoch": 0.8947734700529436, "grad_norm": 1.0685174465179443, "learning_rate": 5.869587522858877e-07, "loss": 0.8156, "step": 25393 }, { "epoch": 0.894808707065902, "grad_norm": 1.1857091188430786, "learning_rate": 5.865735645455439e-07, "loss": 0.722, "step": 25394 }, { "epoch": 0.8948439440788605, "grad_norm": 7.2092790603637695, "learning_rate": 5.86188499416278e-07, "loss": 3.3724, "step": 25395 }, { "epoch": 0.8948791810918189, "grad_norm": 1.1087126731872559, "learning_rate": 5.858035569031062e-07, "loss": 0.9561, "step": 25396 }, { "epoch": 0.8949144181047772, "grad_norm": 3.498155355453491, "learning_rate": 5.854187370110421e-07, "loss": 3.125, "step": 25397 }, { "epoch": 0.8949496551177357, "grad_norm": 1.7819331884384155, "learning_rate": 5.850340397450993e-07, "loss": 0.6917, "step": 25398 }, { "epoch": 0.8949848921306941, "grad_norm": 4.638463020324707, "learning_rate": 5.846494651102863e-07, "loss": 3.2589, "step": 25399 }, { "epoch": 0.8950201291436525, "grad_norm": 1.110280990600586, "learning_rate": 5.842650131116156e-07, "loss": 0.7792, "step": 25400 }, { "epoch": 0.8950553661566109, "grad_norm": 3.2027204036712646, "learning_rate": 5.838806837540889e-07, "loss": 2.9495, "step": 25401 }, { "epoch": 0.8950906031695693, "grad_norm": 2.0900397300720215, "learning_rate": 5.834964770427176e-07, "loss": 0.9192, "step": 25402 }, { "epoch": 0.8951258401825277, "grad_norm": 1.3397144079208374, "learning_rate": 5.831123929825045e-07, "loss": 0.8728, "step": 25403 }, { "epoch": 0.8951610771954861, "grad_norm": 7.310607433319092, "learning_rate": 5.827284315784521e-07, "loss": 2.5353, "step": 25404 }, { "epoch": 0.8951963142084446, "grad_norm": 3.0234827995300293, "learning_rate": 5.823445928355587e-07, "loss": 4.4904, "step": 25405 }, { "epoch": 0.895231551221403, "grad_norm": 6.7527642250061035, "learning_rate": 5.819608767588303e-07, "loss": 6.7488, "step": 25406 }, { "epoch": 0.8952667882343613, "grad_norm": 0.8629109859466553, "learning_rate": 5.815772833532596e-07, "loss": 0.974, "step": 25407 }, { "epoch": 0.8953020252473198, "grad_norm": 1.2390271425247192, "learning_rate": 5.811938126238448e-07, "loss": 0.896, "step": 25408 }, { "epoch": 0.8953372622602782, "grad_norm": 1.2030019760131836, "learning_rate": 5.808104645755797e-07, "loss": 0.5801, "step": 25409 }, { "epoch": 0.8953724992732366, "grad_norm": 3.3809690475463867, "learning_rate": 5.80427239213458e-07, "loss": 3.3111, "step": 25410 }, { "epoch": 0.895407736286195, "grad_norm": 1.0808194875717163, "learning_rate": 5.800441365424725e-07, "loss": 0.8601, "step": 25411 }, { "epoch": 0.8954429732991535, "grad_norm": 1.5854976177215576, "learning_rate": 5.796611565676113e-07, "loss": 1.0091, "step": 25412 }, { "epoch": 0.8954782103121118, "grad_norm": 2.989367961883545, "learning_rate": 5.792782992938628e-07, "loss": 3.1213, "step": 25413 }, { "epoch": 0.8955134473250702, "grad_norm": 1.3474483489990234, "learning_rate": 5.788955647262151e-07, "loss": 0.9047, "step": 25414 }, { "epoch": 0.8955486843380287, "grad_norm": 4.766860485076904, "learning_rate": 5.785129528696531e-07, "loss": 2.9809, "step": 25415 }, { "epoch": 0.8955839213509871, "grad_norm": 5.154664993286133, "learning_rate": 5.781304637291596e-07, "loss": 2.6867, "step": 25416 }, { "epoch": 0.8956191583639455, "grad_norm": 4.614505290985107, "learning_rate": 5.777480973097171e-07, "loss": 2.6091, "step": 25417 }, { "epoch": 0.895654395376904, "grad_norm": 3.0016720294952393, "learning_rate": 5.773658536163062e-07, "loss": 2.0345, "step": 25418 }, { "epoch": 0.8956896323898623, "grad_norm": 3.912810802459717, "learning_rate": 5.769837326539063e-07, "loss": 3.5391, "step": 25419 }, { "epoch": 0.8957248694028207, "grad_norm": 2.648601770401001, "learning_rate": 5.76601734427491e-07, "loss": 0.8676, "step": 25420 }, { "epoch": 0.8957601064157792, "grad_norm": 3.4973948001861572, "learning_rate": 5.762198589420387e-07, "loss": 3.0044, "step": 25421 }, { "epoch": 0.8957953434287376, "grad_norm": 1.0625966787338257, "learning_rate": 5.758381062025264e-07, "loss": 0.8315, "step": 25422 }, { "epoch": 0.895830580441696, "grad_norm": 2.424872398376465, "learning_rate": 5.754564762139203e-07, "loss": 2.8856, "step": 25423 }, { "epoch": 0.8958658174546543, "grad_norm": 5.611278057098389, "learning_rate": 5.75074968981193e-07, "loss": 4.8253, "step": 25424 }, { "epoch": 0.8959010544676128, "grad_norm": 1.1567789316177368, "learning_rate": 5.746935845093183e-07, "loss": 0.981, "step": 25425 }, { "epoch": 0.8959362914805712, "grad_norm": 1.1107633113861084, "learning_rate": 5.743123228032588e-07, "loss": 0.786, "step": 25426 }, { "epoch": 0.8959715284935296, "grad_norm": 3.7243943214416504, "learning_rate": 5.739311838679807e-07, "loss": 3.2307, "step": 25427 }, { "epoch": 0.8960067655064881, "grad_norm": 2.6570847034454346, "learning_rate": 5.735501677084509e-07, "loss": 2.5345, "step": 25428 }, { "epoch": 0.8960420025194464, "grad_norm": 5.321499347686768, "learning_rate": 5.731692743296302e-07, "loss": 2.9373, "step": 25429 }, { "epoch": 0.8960772395324048, "grad_norm": 3.45929217338562, "learning_rate": 5.727885037364811e-07, "loss": 2.7346, "step": 25430 }, { "epoch": 0.8961124765453633, "grad_norm": 3.347564220428467, "learning_rate": 5.724078559339619e-07, "loss": 2.633, "step": 25431 }, { "epoch": 0.8961477135583217, "grad_norm": 5.385926246643066, "learning_rate": 5.720273309270318e-07, "loss": 4.721, "step": 25432 }, { "epoch": 0.8961829505712801, "grad_norm": 4.63469934463501, "learning_rate": 5.71646928720646e-07, "loss": 3.0027, "step": 25433 }, { "epoch": 0.8962181875842384, "grad_norm": 2.2536449432373047, "learning_rate": 5.712666493197605e-07, "loss": 0.7437, "step": 25434 }, { "epoch": 0.8962534245971969, "grad_norm": 0.8478111624717712, "learning_rate": 5.708864927293278e-07, "loss": 0.9728, "step": 25435 }, { "epoch": 0.8962886616101553, "grad_norm": 1.2114169597625732, "learning_rate": 5.705064589542997e-07, "loss": 0.67, "step": 25436 }, { "epoch": 0.8963238986231137, "grad_norm": 0.9194072484970093, "learning_rate": 5.701265479996266e-07, "loss": 0.8021, "step": 25437 }, { "epoch": 0.8963591356360722, "grad_norm": 4.134880065917969, "learning_rate": 5.697467598702566e-07, "loss": 2.6565, "step": 25438 }, { "epoch": 0.8963943726490305, "grad_norm": 1.110664963722229, "learning_rate": 5.693670945711338e-07, "loss": 0.9996, "step": 25439 }, { "epoch": 0.8964296096619889, "grad_norm": 0.8476772904396057, "learning_rate": 5.689875521072075e-07, "loss": 0.912, "step": 25440 }, { "epoch": 0.8964648466749474, "grad_norm": 2.229060173034668, "learning_rate": 5.686081324834214e-07, "loss": 0.9495, "step": 25441 }, { "epoch": 0.8965000836879058, "grad_norm": 1.2302014827728271, "learning_rate": 5.682288357047138e-07, "loss": 0.8759, "step": 25442 }, { "epoch": 0.8965353207008642, "grad_norm": 2.392303466796875, "learning_rate": 5.678496617760265e-07, "loss": 2.5659, "step": 25443 }, { "epoch": 0.8965705577138225, "grad_norm": 1.097081184387207, "learning_rate": 5.674706107023009e-07, "loss": 0.7809, "step": 25444 }, { "epoch": 0.896605794726781, "grad_norm": 4.6510090827941895, "learning_rate": 5.670916824884709e-07, "loss": 3.1098, "step": 25445 }, { "epoch": 0.8966410317397394, "grad_norm": 5.202182292938232, "learning_rate": 5.667128771394726e-07, "loss": 2.951, "step": 25446 }, { "epoch": 0.8966762687526978, "grad_norm": 1.3709564208984375, "learning_rate": 5.663341946602419e-07, "loss": 1.0288, "step": 25447 }, { "epoch": 0.8967115057656563, "grad_norm": 4.167832374572754, "learning_rate": 5.659556350557094e-07, "loss": 2.9657, "step": 25448 }, { "epoch": 0.8967467427786147, "grad_norm": 7.050888538360596, "learning_rate": 5.655771983308056e-07, "loss": 4.8588, "step": 25449 }, { "epoch": 0.896781979791573, "grad_norm": 3.1555540561676025, "learning_rate": 5.651988844904621e-07, "loss": 4.8261, "step": 25450 }, { "epoch": 0.8968172168045315, "grad_norm": 1.2770329713821411, "learning_rate": 5.648206935396039e-07, "loss": 0.6714, "step": 25451 }, { "epoch": 0.8968524538174899, "grad_norm": 0.9154652953147888, "learning_rate": 5.644426254831581e-07, "loss": 0.8454, "step": 25452 }, { "epoch": 0.8968876908304483, "grad_norm": 7.979311943054199, "learning_rate": 5.640646803260497e-07, "loss": 3.4339, "step": 25453 }, { "epoch": 0.8969229278434068, "grad_norm": 1.2436039447784424, "learning_rate": 5.636868580732002e-07, "loss": 0.8805, "step": 25454 }, { "epoch": 0.8969581648563651, "grad_norm": 0.9797486066818237, "learning_rate": 5.633091587295314e-07, "loss": 0.8181, "step": 25455 }, { "epoch": 0.8969934018693235, "grad_norm": 3.8614134788513184, "learning_rate": 5.629315822999626e-07, "loss": 0.9236, "step": 25456 }, { "epoch": 0.8970286388822819, "grad_norm": 2.221250534057617, "learning_rate": 5.625541287894132e-07, "loss": 0.9715, "step": 25457 }, { "epoch": 0.8970638758952404, "grad_norm": 5.786007881164551, "learning_rate": 5.62176798202797e-07, "loss": 6.1237, "step": 25458 }, { "epoch": 0.8970991129081988, "grad_norm": 6.349453926086426, "learning_rate": 5.617995905450302e-07, "loss": 2.5777, "step": 25459 }, { "epoch": 0.8971343499211571, "grad_norm": 4.0150370597839355, "learning_rate": 5.614225058210276e-07, "loss": 2.6776, "step": 25460 }, { "epoch": 0.8971695869341156, "grad_norm": 1.040120005607605, "learning_rate": 5.610455440356988e-07, "loss": 0.8173, "step": 25461 }, { "epoch": 0.897204823947074, "grad_norm": 1.046550989151001, "learning_rate": 5.606687051939519e-07, "loss": 0.7602, "step": 25462 }, { "epoch": 0.8972400609600324, "grad_norm": 3.5843863487243652, "learning_rate": 5.602919893006997e-07, "loss": 4.5272, "step": 25463 }, { "epoch": 0.8972752979729909, "grad_norm": 3.1521453857421875, "learning_rate": 5.599153963608462e-07, "loss": 2.6782, "step": 25464 }, { "epoch": 0.8973105349859493, "grad_norm": 3.463650941848755, "learning_rate": 5.595389263792972e-07, "loss": 3.3796, "step": 25465 }, { "epoch": 0.8973457719989076, "grad_norm": 6.325984477996826, "learning_rate": 5.591625793609556e-07, "loss": 3.0092, "step": 25466 }, { "epoch": 0.897381009011866, "grad_norm": 3.731973648071289, "learning_rate": 5.587863553107242e-07, "loss": 2.9347, "step": 25467 }, { "epoch": 0.8974162460248245, "grad_norm": 4.45875358581543, "learning_rate": 5.584102542335034e-07, "loss": 2.8101, "step": 25468 }, { "epoch": 0.8974514830377829, "grad_norm": 3.159223794937134, "learning_rate": 5.580342761341918e-07, "loss": 1.0294, "step": 25469 }, { "epoch": 0.8974867200507413, "grad_norm": 1.727712631225586, "learning_rate": 5.576584210176861e-07, "loss": 0.9058, "step": 25470 }, { "epoch": 0.8975219570636997, "grad_norm": 1.1345423460006714, "learning_rate": 5.572826888888827e-07, "loss": 0.6959, "step": 25471 }, { "epoch": 0.8975571940766581, "grad_norm": 1.0583786964416504, "learning_rate": 5.569070797526743e-07, "loss": 1.1047, "step": 25472 }, { "epoch": 0.8975924310896165, "grad_norm": 6.9665141105651855, "learning_rate": 5.565315936139548e-07, "loss": 6.1875, "step": 25473 }, { "epoch": 0.897627668102575, "grad_norm": 3.498239278793335, "learning_rate": 5.561562304776136e-07, "loss": 3.1381, "step": 25474 }, { "epoch": 0.8976629051155334, "grad_norm": 3.7678282260894775, "learning_rate": 5.557809903485411e-07, "loss": 2.8089, "step": 25475 }, { "epoch": 0.8976981421284917, "grad_norm": 3.038778066635132, "learning_rate": 5.554058732316259e-07, "loss": 4.4433, "step": 25476 }, { "epoch": 0.8977333791414501, "grad_norm": 2.656791925430298, "learning_rate": 5.550308791317494e-07, "loss": 2.5032, "step": 25477 }, { "epoch": 0.8977686161544086, "grad_norm": 0.9456046223640442, "learning_rate": 5.546560080537988e-07, "loss": 0.7554, "step": 25478 }, { "epoch": 0.897803853167367, "grad_norm": 2.284916877746582, "learning_rate": 5.542812600026593e-07, "loss": 0.848, "step": 25479 }, { "epoch": 0.8978390901803254, "grad_norm": 1.3059054613113403, "learning_rate": 5.539066349832078e-07, "loss": 1.1576, "step": 25480 }, { "epoch": 0.8978743271932839, "grad_norm": 5.212090492248535, "learning_rate": 5.535321330003251e-07, "loss": 5.0963, "step": 25481 }, { "epoch": 0.8979095642062422, "grad_norm": 1.2640459537506104, "learning_rate": 5.531577540588906e-07, "loss": 0.8925, "step": 25482 }, { "epoch": 0.8979448012192006, "grad_norm": 1.1448091268539429, "learning_rate": 5.527834981637792e-07, "loss": 0.7187, "step": 25483 }, { "epoch": 0.8979800382321591, "grad_norm": 4.864127159118652, "learning_rate": 5.524093653198659e-07, "loss": 0.8487, "step": 25484 }, { "epoch": 0.8980152752451175, "grad_norm": 1.1024991273880005, "learning_rate": 5.520353555320246e-07, "loss": 0.9614, "step": 25485 }, { "epoch": 0.8980505122580759, "grad_norm": 2.211740255355835, "learning_rate": 5.51661468805127e-07, "loss": 2.4756, "step": 25486 }, { "epoch": 0.8980857492710342, "grad_norm": 3.104783535003662, "learning_rate": 5.512877051440413e-07, "loss": 1.0265, "step": 25487 }, { "epoch": 0.8981209862839927, "grad_norm": 1.203920841217041, "learning_rate": 5.50914064553637e-07, "loss": 0.7963, "step": 25488 }, { "epoch": 0.8981562232969511, "grad_norm": 1.1516661643981934, "learning_rate": 5.505405470387803e-07, "loss": 1.2399, "step": 25489 }, { "epoch": 0.8981914603099095, "grad_norm": 3.7318265438079834, "learning_rate": 5.501671526043373e-07, "loss": 2.7, "step": 25490 }, { "epoch": 0.898226697322868, "grad_norm": 4.038870334625244, "learning_rate": 5.497938812551696e-07, "loss": 3.2649, "step": 25491 }, { "epoch": 0.8982619343358264, "grad_norm": 5.415353775024414, "learning_rate": 5.494207329961432e-07, "loss": 2.7275, "step": 25492 }, { "epoch": 0.8982971713487847, "grad_norm": 1.8884661197662354, "learning_rate": 5.490477078321121e-07, "loss": 0.919, "step": 25493 }, { "epoch": 0.8983324083617432, "grad_norm": 5.937814712524414, "learning_rate": 5.486748057679403e-07, "loss": 4.4608, "step": 25494 }, { "epoch": 0.8983676453747016, "grad_norm": 2.858160972595215, "learning_rate": 5.483020268084838e-07, "loss": 2.5535, "step": 25495 }, { "epoch": 0.89840288238766, "grad_norm": 3.8821768760681152, "learning_rate": 5.479293709585964e-07, "loss": 2.8449, "step": 25496 }, { "epoch": 0.8984381194006185, "grad_norm": 5.696207046508789, "learning_rate": 5.475568382231311e-07, "loss": 4.8322, "step": 25497 }, { "epoch": 0.8984733564135768, "grad_norm": 2.4429261684417725, "learning_rate": 5.47184428606945e-07, "loss": 2.639, "step": 25498 }, { "epoch": 0.8985085934265352, "grad_norm": 3.578845262527466, "learning_rate": 5.468121421148853e-07, "loss": 3.2051, "step": 25499 }, { "epoch": 0.8985438304394936, "grad_norm": 3.1504437923431396, "learning_rate": 5.464399787518004e-07, "loss": 0.6913, "step": 25500 }, { "epoch": 0.8985790674524521, "grad_norm": 4.115631103515625, "learning_rate": 5.460679385225398e-07, "loss": 3.139, "step": 25501 }, { "epoch": 0.8986143044654105, "grad_norm": 5.320649147033691, "learning_rate": 5.456960214319485e-07, "loss": 5.1161, "step": 25502 }, { "epoch": 0.8986495414783688, "grad_norm": 1.7598363161087036, "learning_rate": 5.453242274848702e-07, "loss": 0.7793, "step": 25503 }, { "epoch": 0.8986847784913273, "grad_norm": 7.328171253204346, "learning_rate": 5.44952556686148e-07, "loss": 5.2432, "step": 25504 }, { "epoch": 0.8987200155042857, "grad_norm": 6.111269950866699, "learning_rate": 5.445810090406234e-07, "loss": 5.0569, "step": 25505 }, { "epoch": 0.8987552525172441, "grad_norm": 3.4996497631073, "learning_rate": 5.442095845531359e-07, "loss": 2.7043, "step": 25506 }, { "epoch": 0.8987904895302026, "grad_norm": 1.3551485538482666, "learning_rate": 5.438382832285228e-07, "loss": 1.0065, "step": 25507 }, { "epoch": 0.898825726543161, "grad_norm": 4.7252068519592285, "learning_rate": 5.434671050716212e-07, "loss": 3.1172, "step": 25508 }, { "epoch": 0.8988609635561193, "grad_norm": 2.142620086669922, "learning_rate": 5.43096050087265e-07, "loss": 0.8465, "step": 25509 }, { "epoch": 0.8988962005690777, "grad_norm": 5.498685359954834, "learning_rate": 5.427251182802861e-07, "loss": 2.6179, "step": 25510 }, { "epoch": 0.8989314375820362, "grad_norm": 1.352023959159851, "learning_rate": 5.423543096555195e-07, "loss": 0.775, "step": 25511 }, { "epoch": 0.8989666745949946, "grad_norm": 0.9474508762359619, "learning_rate": 5.4198362421779e-07, "loss": 0.7546, "step": 25512 }, { "epoch": 0.899001911607953, "grad_norm": 4.068925857543945, "learning_rate": 5.416130619719307e-07, "loss": 2.787, "step": 25513 }, { "epoch": 0.8990371486209114, "grad_norm": 3.790137767791748, "learning_rate": 5.412426229227663e-07, "loss": 3.3006, "step": 25514 }, { "epoch": 0.8990723856338698, "grad_norm": 1.3702051639556885, "learning_rate": 5.408723070751209e-07, "loss": 0.6895, "step": 25515 }, { "epoch": 0.8991076226468282, "grad_norm": 5.154806613922119, "learning_rate": 5.405021144338163e-07, "loss": 5.3393, "step": 25516 }, { "epoch": 0.8991428596597867, "grad_norm": 4.073028564453125, "learning_rate": 5.401320450036795e-07, "loss": 3.3752, "step": 25517 }, { "epoch": 0.8991780966727451, "grad_norm": 0.9477466344833374, "learning_rate": 5.397620987895269e-07, "loss": 1.1036, "step": 25518 }, { "epoch": 0.8992133336857034, "grad_norm": 6.048425197601318, "learning_rate": 5.393922757961789e-07, "loss": 5.4553, "step": 25519 }, { "epoch": 0.8992485706986618, "grad_norm": 1.0436702966690063, "learning_rate": 5.390225760284507e-07, "loss": 0.8039, "step": 25520 }, { "epoch": 0.8992838077116203, "grad_norm": 5.101548671722412, "learning_rate": 5.386529994911582e-07, "loss": 2.9772, "step": 25521 }, { "epoch": 0.8993190447245787, "grad_norm": 4.639092922210693, "learning_rate": 5.382835461891156e-07, "loss": 2.5798, "step": 25522 }, { "epoch": 0.8993542817375371, "grad_norm": 3.310899496078491, "learning_rate": 5.379142161271356e-07, "loss": 2.6423, "step": 25523 }, { "epoch": 0.8993895187504956, "grad_norm": 0.9837016463279724, "learning_rate": 5.375450093100277e-07, "loss": 1.1059, "step": 25524 }, { "epoch": 0.8994247557634539, "grad_norm": 1.0876728296279907, "learning_rate": 5.371759257426012e-07, "loss": 0.8909, "step": 25525 }, { "epoch": 0.8994599927764123, "grad_norm": 27.136240005493164, "learning_rate": 5.368069654296648e-07, "loss": 3.4665, "step": 25526 }, { "epoch": 0.8994952297893708, "grad_norm": 1.3370699882507324, "learning_rate": 5.364381283760223e-07, "loss": 0.8726, "step": 25527 }, { "epoch": 0.8995304668023292, "grad_norm": 1.194387435913086, "learning_rate": 5.360694145864787e-07, "loss": 0.7353, "step": 25528 }, { "epoch": 0.8995657038152876, "grad_norm": 2.864145040512085, "learning_rate": 5.357008240658368e-07, "loss": 2.6229, "step": 25529 }, { "epoch": 0.899600940828246, "grad_norm": 3.0527431964874268, "learning_rate": 5.353323568188984e-07, "loss": 2.8417, "step": 25530 }, { "epoch": 0.8996361778412044, "grad_norm": 3.833754301071167, "learning_rate": 5.349640128504585e-07, "loss": 4.7348, "step": 25531 }, { "epoch": 0.8996714148541628, "grad_norm": 1.0916908979415894, "learning_rate": 5.3459579216532e-07, "loss": 0.8527, "step": 25532 }, { "epoch": 0.8997066518671212, "grad_norm": 3.023104667663574, "learning_rate": 5.342276947682778e-07, "loss": 2.0781, "step": 25533 }, { "epoch": 0.8997418888800797, "grad_norm": 1.1923775672912598, "learning_rate": 5.338597206641238e-07, "loss": 0.9384, "step": 25534 }, { "epoch": 0.899777125893038, "grad_norm": 3.831433057785034, "learning_rate": 5.33491869857653e-07, "loss": 2.6189, "step": 25535 }, { "epoch": 0.8998123629059964, "grad_norm": 3.712441921234131, "learning_rate": 5.331241423536582e-07, "loss": 3.5168, "step": 25536 }, { "epoch": 0.8998475999189549, "grad_norm": 2.717592477798462, "learning_rate": 5.327565381569256e-07, "loss": 2.8943, "step": 25537 }, { "epoch": 0.8998828369319133, "grad_norm": 3.8722622394561768, "learning_rate": 5.323890572722457e-07, "loss": 2.2709, "step": 25538 }, { "epoch": 0.8999180739448717, "grad_norm": 4.635048866271973, "learning_rate": 5.320216997044036e-07, "loss": 5.0615, "step": 25539 }, { "epoch": 0.8999533109578302, "grad_norm": 2.1074295043945312, "learning_rate": 5.316544654581857e-07, "loss": 0.784, "step": 25540 }, { "epoch": 0.8999885479707885, "grad_norm": 3.5743117332458496, "learning_rate": 5.312873545383735e-07, "loss": 2.8668, "step": 25541 }, { "epoch": 0.9000237849837469, "grad_norm": 1.115113377571106, "learning_rate": 5.309203669497509e-07, "loss": 0.8144, "step": 25542 }, { "epoch": 0.9000590219967053, "grad_norm": 3.782640218734741, "learning_rate": 5.305535026970954e-07, "loss": 4.5519, "step": 25543 }, { "epoch": 0.9000942590096638, "grad_norm": 2.589709520339966, "learning_rate": 5.301867617851874e-07, "loss": 2.421, "step": 25544 }, { "epoch": 0.9001294960226222, "grad_norm": 5.066438674926758, "learning_rate": 5.298201442188033e-07, "loss": 2.4763, "step": 25545 }, { "epoch": 0.9001647330355805, "grad_norm": 0.9705823063850403, "learning_rate": 5.294536500027181e-07, "loss": 0.9819, "step": 25546 }, { "epoch": 0.900199970048539, "grad_norm": 2.9870765209198, "learning_rate": 5.290872791417045e-07, "loss": 3.1858, "step": 25547 }, { "epoch": 0.9002352070614974, "grad_norm": 3.5354061126708984, "learning_rate": 5.287210316405367e-07, "loss": 4.4931, "step": 25548 }, { "epoch": 0.9002704440744558, "grad_norm": 1.670640230178833, "learning_rate": 5.283549075039851e-07, "loss": 0.7625, "step": 25549 }, { "epoch": 0.9003056810874143, "grad_norm": 1.3600175380706787, "learning_rate": 5.279889067368138e-07, "loss": 0.9733, "step": 25550 }, { "epoch": 0.9003409181003726, "grad_norm": 3.5936217308044434, "learning_rate": 5.276230293437956e-07, "loss": 2.6969, "step": 25551 }, { "epoch": 0.900376155113331, "grad_norm": 1.4906209707260132, "learning_rate": 5.272572753296956e-07, "loss": 1.049, "step": 25552 }, { "epoch": 0.9004113921262894, "grad_norm": 3.0687685012817383, "learning_rate": 5.268916446992745e-07, "loss": 2.849, "step": 25553 }, { "epoch": 0.9004466291392479, "grad_norm": 1.2546495199203491, "learning_rate": 5.26526137457295e-07, "loss": 0.9398, "step": 25554 }, { "epoch": 0.9004818661522063, "grad_norm": 1.6242059469223022, "learning_rate": 5.261607536085222e-07, "loss": 0.8764, "step": 25555 }, { "epoch": 0.9005171031651646, "grad_norm": 3.8510730266571045, "learning_rate": 5.257954931577103e-07, "loss": 3.0418, "step": 25556 }, { "epoch": 0.9005523401781231, "grad_norm": 1.221225619316101, "learning_rate": 5.254303561096197e-07, "loss": 0.8616, "step": 25557 }, { "epoch": 0.9005875771910815, "grad_norm": 1.1249669790267944, "learning_rate": 5.250653424690056e-07, "loss": 1.1029, "step": 25558 }, { "epoch": 0.9006228142040399, "grad_norm": 4.702565670013428, "learning_rate": 5.247004522406218e-07, "loss": 3.2801, "step": 25559 }, { "epoch": 0.9006580512169984, "grad_norm": 1.1710145473480225, "learning_rate": 5.243356854292214e-07, "loss": 0.9009, "step": 25560 }, { "epoch": 0.9006932882299568, "grad_norm": 1.1158270835876465, "learning_rate": 5.239710420395561e-07, "loss": 0.9537, "step": 25561 }, { "epoch": 0.9007285252429151, "grad_norm": 2.6750669479370117, "learning_rate": 5.236065220763753e-07, "loss": 2.7496, "step": 25562 }, { "epoch": 0.9007637622558736, "grad_norm": 2.559629440307617, "learning_rate": 5.232421255444265e-07, "loss": 0.9661, "step": 25563 }, { "epoch": 0.900798999268832, "grad_norm": 0.9049704670906067, "learning_rate": 5.228778524484557e-07, "loss": 0.7862, "step": 25564 }, { "epoch": 0.9008342362817904, "grad_norm": 3.386875629425049, "learning_rate": 5.225137027932082e-07, "loss": 2.7878, "step": 25565 }, { "epoch": 0.9008694732947488, "grad_norm": 1.1865346431732178, "learning_rate": 5.221496765834278e-07, "loss": 0.9612, "step": 25566 }, { "epoch": 0.9009047103077072, "grad_norm": 3.3373546600341797, "learning_rate": 5.217857738238541e-07, "loss": 2.709, "step": 25567 }, { "epoch": 0.9009399473206656, "grad_norm": 1.3706614971160889, "learning_rate": 5.214219945192301e-07, "loss": 0.8052, "step": 25568 }, { "epoch": 0.900975184333624, "grad_norm": 4.977986812591553, "learning_rate": 5.210583386742885e-07, "loss": 3.0948, "step": 25569 }, { "epoch": 0.9010104213465825, "grad_norm": 1.444742202758789, "learning_rate": 5.206948062937723e-07, "loss": 0.8779, "step": 25570 }, { "epoch": 0.9010456583595409, "grad_norm": 0.8715501427650452, "learning_rate": 5.203313973824131e-07, "loss": 0.751, "step": 25571 }, { "epoch": 0.9010808953724992, "grad_norm": 1.1330620050430298, "learning_rate": 5.199681119449451e-07, "loss": 0.8125, "step": 25572 }, { "epoch": 0.9011161323854577, "grad_norm": 3.5071609020233154, "learning_rate": 5.196049499860978e-07, "loss": 2.3815, "step": 25573 }, { "epoch": 0.9011513693984161, "grad_norm": 1.1969581842422485, "learning_rate": 5.192419115106073e-07, "loss": 0.6992, "step": 25574 }, { "epoch": 0.9011866064113745, "grad_norm": 1.0638476610183716, "learning_rate": 5.188789965231966e-07, "loss": 0.754, "step": 25575 }, { "epoch": 0.9012218434243329, "grad_norm": 3.4736135005950928, "learning_rate": 5.185162050285941e-07, "loss": 2.4989, "step": 25576 }, { "epoch": 0.9012570804372914, "grad_norm": 2.653087854385376, "learning_rate": 5.181535370315272e-07, "loss": 2.2311, "step": 25577 }, { "epoch": 0.9012923174502497, "grad_norm": 4.089476585388184, "learning_rate": 5.177909925367175e-07, "loss": 2.4326, "step": 25578 }, { "epoch": 0.9013275544632081, "grad_norm": 1.4443039894104004, "learning_rate": 5.174285715488869e-07, "loss": 0.7888, "step": 25579 }, { "epoch": 0.9013627914761666, "grad_norm": 3.1608893871307373, "learning_rate": 5.170662740727584e-07, "loss": 2.9754, "step": 25580 }, { "epoch": 0.901398028489125, "grad_norm": 4.154590129852295, "learning_rate": 5.16704100113049e-07, "loss": 3.039, "step": 25581 }, { "epoch": 0.9014332655020834, "grad_norm": 0.9015187621116638, "learning_rate": 5.163420496744775e-07, "loss": 0.7645, "step": 25582 }, { "epoch": 0.9014685025150418, "grad_norm": 1.514028787612915, "learning_rate": 5.159801227617578e-07, "loss": 0.8329, "step": 25583 }, { "epoch": 0.9015037395280002, "grad_norm": 2.8933048248291016, "learning_rate": 5.15618319379606e-07, "loss": 2.802, "step": 25584 }, { "epoch": 0.9015389765409586, "grad_norm": 2.381944179534912, "learning_rate": 5.15256639532733e-07, "loss": 2.3322, "step": 25585 }, { "epoch": 0.901574213553917, "grad_norm": 1.3297964334487915, "learning_rate": 5.148950832258503e-07, "loss": 1.0894, "step": 25586 }, { "epoch": 0.9016094505668755, "grad_norm": 1.35918390750885, "learning_rate": 5.1453365046367e-07, "loss": 0.6372, "step": 25587 }, { "epoch": 0.9016446875798338, "grad_norm": 5.107171535491943, "learning_rate": 5.141723412508926e-07, "loss": 3.0294, "step": 25588 }, { "epoch": 0.9016799245927922, "grad_norm": 1.663683295249939, "learning_rate": 5.138111555922309e-07, "loss": 1.1793, "step": 25589 }, { "epoch": 0.9017151616057507, "grad_norm": 1.2967619895935059, "learning_rate": 5.134500934923892e-07, "loss": 0.759, "step": 25590 }, { "epoch": 0.9017503986187091, "grad_norm": 3.8569657802581787, "learning_rate": 5.130891549560668e-07, "loss": 3.1486, "step": 25591 }, { "epoch": 0.9017856356316675, "grad_norm": 2.986772060394287, "learning_rate": 5.127283399879646e-07, "loss": 2.5508, "step": 25592 }, { "epoch": 0.901820872644626, "grad_norm": 7.708979606628418, "learning_rate": 5.123676485927876e-07, "loss": 4.934, "step": 25593 }, { "epoch": 0.9018561096575843, "grad_norm": 3.0043437480926514, "learning_rate": 5.120070807752287e-07, "loss": 2.9492, "step": 25594 }, { "epoch": 0.9018913466705427, "grad_norm": 3.1750500202178955, "learning_rate": 5.116466365399863e-07, "loss": 2.7187, "step": 25595 }, { "epoch": 0.9019265836835012, "grad_norm": 0.9653723239898682, "learning_rate": 5.112863158917547e-07, "loss": 0.8342, "step": 25596 }, { "epoch": 0.9019618206964596, "grad_norm": 1.332302451133728, "learning_rate": 5.109261188352288e-07, "loss": 0.9639, "step": 25597 }, { "epoch": 0.901997057709418, "grad_norm": 2.303006172180176, "learning_rate": 5.105660453750983e-07, "loss": 2.1981, "step": 25598 }, { "epoch": 0.9020322947223763, "grad_norm": 3.922745704650879, "learning_rate": 5.102060955160537e-07, "loss": 2.2545, "step": 25599 }, { "epoch": 0.9020675317353348, "grad_norm": 0.9051694869995117, "learning_rate": 5.098462692627837e-07, "loss": 0.7866, "step": 25600 }, { "epoch": 0.9021027687482932, "grad_norm": 0.936485767364502, "learning_rate": 5.094865666199745e-07, "loss": 0.7034, "step": 25601 }, { "epoch": 0.9021380057612516, "grad_norm": 2.600560426712036, "learning_rate": 5.091269875923121e-07, "loss": 2.3287, "step": 25602 }, { "epoch": 0.9021732427742101, "grad_norm": 1.2620017528533936, "learning_rate": 5.087675321844788e-07, "loss": 0.9383, "step": 25603 }, { "epoch": 0.9022084797871684, "grad_norm": 3.3815994262695312, "learning_rate": 5.084082004011582e-07, "loss": 2.8311, "step": 25604 }, { "epoch": 0.9022437168001268, "grad_norm": 6.073380947113037, "learning_rate": 5.0804899224703e-07, "loss": 5.1127, "step": 25605 }, { "epoch": 0.9022789538130853, "grad_norm": 5.228507995605469, "learning_rate": 5.076899077267738e-07, "loss": 4.1535, "step": 25606 }, { "epoch": 0.9023141908260437, "grad_norm": 3.047384023666382, "learning_rate": 5.073309468450627e-07, "loss": 2.8857, "step": 25607 }, { "epoch": 0.9023494278390021, "grad_norm": 1.2378829717636108, "learning_rate": 5.069721096065761e-07, "loss": 0.9759, "step": 25608 }, { "epoch": 0.9023846648519604, "grad_norm": 1.836020588874817, "learning_rate": 5.066133960159892e-07, "loss": 0.9028, "step": 25609 }, { "epoch": 0.9024199018649189, "grad_norm": 1.956343173980713, "learning_rate": 5.062548060779704e-07, "loss": 1.0295, "step": 25610 }, { "epoch": 0.9024551388778773, "grad_norm": 5.23753547668457, "learning_rate": 5.058963397971905e-07, "loss": 3.4274, "step": 25611 }, { "epoch": 0.9024903758908357, "grad_norm": 1.504638671875, "learning_rate": 5.055379971783237e-07, "loss": 0.755, "step": 25612 }, { "epoch": 0.9025256129037942, "grad_norm": 2.1008894443511963, "learning_rate": 5.051797782260314e-07, "loss": 0.9286, "step": 25613 }, { "epoch": 0.9025608499167526, "grad_norm": 2.914905071258545, "learning_rate": 5.048216829449826e-07, "loss": 0.7588, "step": 25614 }, { "epoch": 0.9025960869297109, "grad_norm": 1.0135173797607422, "learning_rate": 5.044637113398399e-07, "loss": 0.7533, "step": 25615 }, { "epoch": 0.9026313239426694, "grad_norm": 3.2311995029449463, "learning_rate": 5.041058634152684e-07, "loss": 4.5524, "step": 25616 }, { "epoch": 0.9026665609556278, "grad_norm": 0.9729741215705872, "learning_rate": 5.03748139175928e-07, "loss": 0.7314, "step": 25617 }, { "epoch": 0.9027017979685862, "grad_norm": 1.1294243335723877, "learning_rate": 5.033905386264759e-07, "loss": 0.9293, "step": 25618 }, { "epoch": 0.9027370349815446, "grad_norm": 4.946458339691162, "learning_rate": 5.030330617715729e-07, "loss": 5.5691, "step": 25619 }, { "epoch": 0.902772271994503, "grad_norm": 5.983835220336914, "learning_rate": 5.02675708615874e-07, "loss": 3.3432, "step": 25620 }, { "epoch": 0.9028075090074614, "grad_norm": 1.3202544450759888, "learning_rate": 5.023184791640323e-07, "loss": 0.7788, "step": 25621 }, { "epoch": 0.9028427460204198, "grad_norm": 3.570573091506958, "learning_rate": 5.019613734207041e-07, "loss": 2.9942, "step": 25622 }, { "epoch": 0.9028779830333783, "grad_norm": 2.3818109035491943, "learning_rate": 5.016043913905377e-07, "loss": 2.3195, "step": 25623 }, { "epoch": 0.9029132200463367, "grad_norm": 1.3863595724105835, "learning_rate": 5.012475330781841e-07, "loss": 0.7394, "step": 25624 }, { "epoch": 0.902948457059295, "grad_norm": 1.1707534790039062, "learning_rate": 5.008907984882927e-07, "loss": 1.1642, "step": 25625 }, { "epoch": 0.9029836940722535, "grad_norm": 1.5609800815582275, "learning_rate": 5.005341876255066e-07, "loss": 0.9257, "step": 25626 }, { "epoch": 0.9030189310852119, "grad_norm": 3.2541284561157227, "learning_rate": 5.001777004944741e-07, "loss": 2.5855, "step": 25627 }, { "epoch": 0.9030541680981703, "grad_norm": 1.4093942642211914, "learning_rate": 4.998213370998373e-07, "loss": 1.1359, "step": 25628 }, { "epoch": 0.9030894051111288, "grad_norm": 1.1314901113510132, "learning_rate": 4.994650974462379e-07, "loss": 1.1042, "step": 25629 }, { "epoch": 0.9031246421240872, "grad_norm": 2.2411489486694336, "learning_rate": 4.991089815383132e-07, "loss": 0.9552, "step": 25630 }, { "epoch": 0.9031598791370455, "grad_norm": 11.492287635803223, "learning_rate": 4.987529893807075e-07, "loss": 7.6583, "step": 25631 }, { "epoch": 0.9031951161500039, "grad_norm": 1.5737770795822144, "learning_rate": 4.983971209780525e-07, "loss": 0.8371, "step": 25632 }, { "epoch": 0.9032303531629624, "grad_norm": 0.9584475159645081, "learning_rate": 4.980413763349856e-07, "loss": 0.871, "step": 25633 }, { "epoch": 0.9032655901759208, "grad_norm": 5.611949443817139, "learning_rate": 4.976857554561399e-07, "loss": 5.5412, "step": 25634 }, { "epoch": 0.9033008271888792, "grad_norm": 1.4463928937911987, "learning_rate": 4.973302583461493e-07, "loss": 0.7711, "step": 25635 }, { "epoch": 0.9033360642018377, "grad_norm": 2.0864484310150146, "learning_rate": 4.969748850096412e-07, "loss": 2.4337, "step": 25636 }, { "epoch": 0.903371301214796, "grad_norm": 0.8707547187805176, "learning_rate": 4.966196354512454e-07, "loss": 1.046, "step": 25637 }, { "epoch": 0.9034065382277544, "grad_norm": 1.2731436491012573, "learning_rate": 4.962645096755892e-07, "loss": 0.8291, "step": 25638 }, { "epoch": 0.9034417752407129, "grad_norm": 1.2153962850570679, "learning_rate": 4.959095076872988e-07, "loss": 0.8324, "step": 25639 }, { "epoch": 0.9034770122536713, "grad_norm": 2.90437912940979, "learning_rate": 4.955546294909986e-07, "loss": 2.5674, "step": 25640 }, { "epoch": 0.9035122492666297, "grad_norm": 10.579731941223145, "learning_rate": 4.95199875091309e-07, "loss": 3.229, "step": 25641 }, { "epoch": 0.903547486279588, "grad_norm": 1.0384117364883423, "learning_rate": 4.94845244492852e-07, "loss": 0.8097, "step": 25642 }, { "epoch": 0.9035827232925465, "grad_norm": 1.7567808628082275, "learning_rate": 4.944907377002461e-07, "loss": 0.8261, "step": 25643 }, { "epoch": 0.9036179603055049, "grad_norm": 4.275175094604492, "learning_rate": 4.94136354718111e-07, "loss": 4.9061, "step": 25644 }, { "epoch": 0.9036531973184633, "grad_norm": 1.4957329034805298, "learning_rate": 4.937820955510575e-07, "loss": 0.6717, "step": 25645 }, { "epoch": 0.9036884343314218, "grad_norm": 1.23530113697052, "learning_rate": 4.934279602037051e-07, "loss": 1.0185, "step": 25646 }, { "epoch": 0.9037236713443801, "grad_norm": 1.064978003501892, "learning_rate": 4.930739486806658e-07, "loss": 0.9422, "step": 25647 }, { "epoch": 0.9037589083573385, "grad_norm": 2.2678442001342773, "learning_rate": 4.927200609865468e-07, "loss": 1.1083, "step": 25648 }, { "epoch": 0.903794145370297, "grad_norm": 4.956650733947754, "learning_rate": 4.923662971259602e-07, "loss": 2.9591, "step": 25649 }, { "epoch": 0.9038293823832554, "grad_norm": 1.501507043838501, "learning_rate": 4.920126571035156e-07, "loss": 0.8901, "step": 25650 }, { "epoch": 0.9038646193962138, "grad_norm": 0.956221342086792, "learning_rate": 4.916591409238158e-07, "loss": 0.7221, "step": 25651 }, { "epoch": 0.9038998564091721, "grad_norm": 3.3499224185943604, "learning_rate": 4.913057485914674e-07, "loss": 2.496, "step": 25652 }, { "epoch": 0.9039350934221306, "grad_norm": 1.1614526510238647, "learning_rate": 4.90952480111071e-07, "loss": 0.997, "step": 25653 }, { "epoch": 0.903970330435089, "grad_norm": 3.341970205307007, "learning_rate": 4.90599335487233e-07, "loss": 2.8387, "step": 25654 }, { "epoch": 0.9040055674480474, "grad_norm": 2.7513632774353027, "learning_rate": 4.902463147245473e-07, "loss": 2.7962, "step": 25655 }, { "epoch": 0.9040408044610059, "grad_norm": 3.092932939529419, "learning_rate": 4.89893417827616e-07, "loss": 3.4068, "step": 25656 }, { "epoch": 0.9040760414739643, "grad_norm": 2.067667245864868, "learning_rate": 4.895406448010343e-07, "loss": 2.6618, "step": 25657 }, { "epoch": 0.9041112784869226, "grad_norm": 1.201009750366211, "learning_rate": 4.891879956493961e-07, "loss": 0.9016, "step": 25658 }, { "epoch": 0.9041465154998811, "grad_norm": 32.690834045410156, "learning_rate": 4.888354703772969e-07, "loss": 5.3402, "step": 25659 }, { "epoch": 0.9041817525128395, "grad_norm": 2.9244420528411865, "learning_rate": 4.884830689893272e-07, "loss": 2.6696, "step": 25660 }, { "epoch": 0.9042169895257979, "grad_norm": 1.684375524520874, "learning_rate": 4.881307914900768e-07, "loss": 0.8415, "step": 25661 }, { "epoch": 0.9042522265387564, "grad_norm": 7.071649074554443, "learning_rate": 4.877786378841343e-07, "loss": 6.8341, "step": 25662 }, { "epoch": 0.9042874635517147, "grad_norm": 1.3767297267913818, "learning_rate": 4.874266081760881e-07, "loss": 1.1076, "step": 25663 }, { "epoch": 0.9043227005646731, "grad_norm": 3.5477516651153564, "learning_rate": 4.870747023705191e-07, "loss": 4.416, "step": 25664 }, { "epoch": 0.9043579375776315, "grad_norm": 1.3186694383621216, "learning_rate": 4.867229204720158e-07, "loss": 0.8189, "step": 25665 }, { "epoch": 0.90439317459059, "grad_norm": 7.044681549072266, "learning_rate": 4.863712624851591e-07, "loss": 5.7952, "step": 25666 }, { "epoch": 0.9044284116035484, "grad_norm": 3.744943857192993, "learning_rate": 4.860197284145274e-07, "loss": 2.8554, "step": 25667 }, { "epoch": 0.9044636486165067, "grad_norm": 0.9033241271972656, "learning_rate": 4.856683182646993e-07, "loss": 0.678, "step": 25668 }, { "epoch": 0.9044988856294652, "grad_norm": 3.716324806213379, "learning_rate": 4.853170320402545e-07, "loss": 2.8664, "step": 25669 }, { "epoch": 0.9045341226424236, "grad_norm": 4.555392265319824, "learning_rate": 4.849658697457682e-07, "loss": 4.5697, "step": 25670 }, { "epoch": 0.904569359655382, "grad_norm": 0.9249988198280334, "learning_rate": 4.846148313858134e-07, "loss": 0.8119, "step": 25671 }, { "epoch": 0.9046045966683405, "grad_norm": 3.2028403282165527, "learning_rate": 4.842639169649599e-07, "loss": 3.4253, "step": 25672 }, { "epoch": 0.9046398336812989, "grad_norm": 1.129969596862793, "learning_rate": 4.83913126487785e-07, "loss": 0.9454, "step": 25673 }, { "epoch": 0.9046750706942572, "grad_norm": 0.9239813089370728, "learning_rate": 4.835624599588506e-07, "loss": 1.0761, "step": 25674 }, { "epoch": 0.9047103077072156, "grad_norm": 4.348963260650635, "learning_rate": 4.832119173827288e-07, "loss": 5.2718, "step": 25675 }, { "epoch": 0.9047455447201741, "grad_norm": 4.174715042114258, "learning_rate": 4.828614987639834e-07, "loss": 2.576, "step": 25676 }, { "epoch": 0.9047807817331325, "grad_norm": 5.937121391296387, "learning_rate": 4.8251120410718e-07, "loss": 2.52, "step": 25677 }, { "epoch": 0.9048160187460909, "grad_norm": 2.626821994781494, "learning_rate": 4.821610334168803e-07, "loss": 2.6531, "step": 25678 }, { "epoch": 0.9048512557590493, "grad_norm": 2.2726101875305176, "learning_rate": 4.818109866976451e-07, "loss": 0.9517, "step": 25679 }, { "epoch": 0.9048864927720077, "grad_norm": 1.5565768480300903, "learning_rate": 4.81461063954034e-07, "loss": 0.7282, "step": 25680 }, { "epoch": 0.9049217297849661, "grad_norm": 7.524959087371826, "learning_rate": 4.811112651906058e-07, "loss": 3.4924, "step": 25681 }, { "epoch": 0.9049569667979246, "grad_norm": 6.142911434173584, "learning_rate": 4.807615904119167e-07, "loss": 3.4543, "step": 25682 }, { "epoch": 0.904992203810883, "grad_norm": 2.834137201309204, "learning_rate": 4.804120396225176e-07, "loss": 3.1884, "step": 25683 }, { "epoch": 0.9050274408238413, "grad_norm": 3.013136625289917, "learning_rate": 4.800626128269659e-07, "loss": 3.0654, "step": 25684 }, { "epoch": 0.9050626778367997, "grad_norm": 1.4803333282470703, "learning_rate": 4.797133100298123e-07, "loss": 0.7797, "step": 25685 }, { "epoch": 0.9050979148497582, "grad_norm": 4.384426116943359, "learning_rate": 4.793641312356046e-07, "loss": 2.9915, "step": 25686 }, { "epoch": 0.9051331518627166, "grad_norm": 2.76663875579834, "learning_rate": 4.79015076448891e-07, "loss": 3.0124, "step": 25687 }, { "epoch": 0.905168388875675, "grad_norm": 3.7122273445129395, "learning_rate": 4.786661456742192e-07, "loss": 3.4977, "step": 25688 }, { "epoch": 0.9052036258886335, "grad_norm": 6.113269805908203, "learning_rate": 4.783173389161355e-07, "loss": 5.0206, "step": 25689 }, { "epoch": 0.9052388629015918, "grad_norm": 2.2900302410125732, "learning_rate": 4.779686561791797e-07, "loss": 0.7397, "step": 25690 }, { "epoch": 0.9052740999145502, "grad_norm": 1.1454766988754272, "learning_rate": 4.776200974678935e-07, "loss": 0.8638, "step": 25691 }, { "epoch": 0.9053093369275087, "grad_norm": 3.8109190464019775, "learning_rate": 4.772716627868223e-07, "loss": 2.9696, "step": 25692 }, { "epoch": 0.9053445739404671, "grad_norm": 1.3581364154815674, "learning_rate": 4.769233521404981e-07, "loss": 0.7146, "step": 25693 }, { "epoch": 0.9053798109534255, "grad_norm": 0.9165378212928772, "learning_rate": 4.765751655334605e-07, "loss": 0.6103, "step": 25694 }, { "epoch": 0.9054150479663838, "grad_norm": 5.352174758911133, "learning_rate": 4.762271029702448e-07, "loss": 3.0206, "step": 25695 }, { "epoch": 0.9054502849793423, "grad_norm": 2.729107141494751, "learning_rate": 4.7587916445538506e-07, "loss": 2.3173, "step": 25696 }, { "epoch": 0.9054855219923007, "grad_norm": 1.0402911901474, "learning_rate": 4.755313499934111e-07, "loss": 1.0085, "step": 25697 }, { "epoch": 0.9055207590052591, "grad_norm": 0.9979097843170166, "learning_rate": 4.751836595888548e-07, "loss": 0.8426, "step": 25698 }, { "epoch": 0.9055559960182176, "grad_norm": 1.2975525856018066, "learning_rate": 4.748360932462448e-07, "loss": 0.8791, "step": 25699 }, { "epoch": 0.905591233031176, "grad_norm": 0.9312959909439087, "learning_rate": 4.7448865097010745e-07, "loss": 0.9912, "step": 25700 }, { "epoch": 0.9056264700441343, "grad_norm": 1.6718800067901611, "learning_rate": 4.7414133276497133e-07, "loss": 0.7428, "step": 25701 }, { "epoch": 0.9056617070570928, "grad_norm": 0.978327751159668, "learning_rate": 4.7379413863535395e-07, "loss": 0.8109, "step": 25702 }, { "epoch": 0.9056969440700512, "grad_norm": 4.241010665893555, "learning_rate": 4.7344706858578284e-07, "loss": 4.3368, "step": 25703 }, { "epoch": 0.9057321810830096, "grad_norm": 1.1992086172103882, "learning_rate": 4.7310012262077873e-07, "loss": 0.7335, "step": 25704 }, { "epoch": 0.9057674180959681, "grad_norm": 1.0279303789138794, "learning_rate": 4.7275330074485817e-07, "loss": 0.7699, "step": 25705 }, { "epoch": 0.9058026551089264, "grad_norm": 5.302657604217529, "learning_rate": 4.7240660296253737e-07, "loss": 4.2401, "step": 25706 }, { "epoch": 0.9058378921218848, "grad_norm": 2.5986931324005127, "learning_rate": 4.7206002927833614e-07, "loss": 2.3043, "step": 25707 }, { "epoch": 0.9058731291348432, "grad_norm": 2.9644534587860107, "learning_rate": 4.7171357969676755e-07, "loss": 2.6302, "step": 25708 }, { "epoch": 0.9059083661478017, "grad_norm": 5.063473224639893, "learning_rate": 4.713672542223424e-07, "loss": 4.8675, "step": 25709 }, { "epoch": 0.9059436031607601, "grad_norm": 3.845038652420044, "learning_rate": 4.7102105285957046e-07, "loss": 2.8823, "step": 25710 }, { "epoch": 0.9059788401737184, "grad_norm": 3.5065360069274902, "learning_rate": 4.7067497561296695e-07, "loss": 2.5211, "step": 25711 }, { "epoch": 0.9060140771866769, "grad_norm": 4.774905204772949, "learning_rate": 4.703290224870327e-07, "loss": 3.012, "step": 25712 }, { "epoch": 0.9060493141996353, "grad_norm": 6.738465309143066, "learning_rate": 4.699831934862786e-07, "loss": 3.2031, "step": 25713 }, { "epoch": 0.9060845512125937, "grad_norm": 1.1786555051803589, "learning_rate": 4.6963748861520666e-07, "loss": 0.8744, "step": 25714 }, { "epoch": 0.9061197882255522, "grad_norm": 2.4482758045196533, "learning_rate": 4.692919078783198e-07, "loss": 2.6021, "step": 25715 }, { "epoch": 0.9061550252385105, "grad_norm": 1.1937612295150757, "learning_rate": 4.6894645128012006e-07, "loss": 0.7606, "step": 25716 }, { "epoch": 0.9061902622514689, "grad_norm": 2.9932329654693604, "learning_rate": 4.6860111882510716e-07, "loss": 2.8481, "step": 25717 }, { "epoch": 0.9062254992644273, "grad_norm": 0.9156205654144287, "learning_rate": 4.682559105177786e-07, "loss": 0.8087, "step": 25718 }, { "epoch": 0.9062607362773858, "grad_norm": 2.493872880935669, "learning_rate": 4.6791082636263195e-07, "loss": 2.8159, "step": 25719 }, { "epoch": 0.9062959732903442, "grad_norm": 1.1792343854904175, "learning_rate": 4.6756586636416136e-07, "loss": 0.7482, "step": 25720 }, { "epoch": 0.9063312103033025, "grad_norm": 1.234128475189209, "learning_rate": 4.672210305268565e-07, "loss": 0.8627, "step": 25721 }, { "epoch": 0.906366447316261, "grad_norm": 1.381213903427124, "learning_rate": 4.668763188552139e-07, "loss": 1.1443, "step": 25722 }, { "epoch": 0.9064016843292194, "grad_norm": 1.4738950729370117, "learning_rate": 4.665317313537221e-07, "loss": 1.0653, "step": 25723 }, { "epoch": 0.9064369213421778, "grad_norm": 3.5846121311187744, "learning_rate": 4.6618726802686975e-07, "loss": 2.8033, "step": 25724 }, { "epoch": 0.9064721583551363, "grad_norm": 1.4872674942016602, "learning_rate": 4.6584292887913886e-07, "loss": 0.7537, "step": 25725 }, { "epoch": 0.9065073953680947, "grad_norm": 1.2909599542617798, "learning_rate": 4.6549871391502134e-07, "loss": 0.77, "step": 25726 }, { "epoch": 0.906542632381053, "grad_norm": 1.0111867189407349, "learning_rate": 4.6515462313899805e-07, "loss": 0.7772, "step": 25727 }, { "epoch": 0.9065778693940114, "grad_norm": 3.121623992919922, "learning_rate": 4.648106565555488e-07, "loss": 2.5787, "step": 25728 }, { "epoch": 0.9066131064069699, "grad_norm": 1.2064323425292969, "learning_rate": 4.644668141691544e-07, "loss": 0.8748, "step": 25729 }, { "epoch": 0.9066483434199283, "grad_norm": 2.662799596786499, "learning_rate": 4.641230959842968e-07, "loss": 2.9159, "step": 25730 }, { "epoch": 0.9066835804328867, "grad_norm": 0.9884676337242126, "learning_rate": 4.637795020054503e-07, "loss": 0.6545, "step": 25731 }, { "epoch": 0.9067188174458451, "grad_norm": 1.025315523147583, "learning_rate": 4.63436032237089e-07, "loss": 0.7168, "step": 25732 }, { "epoch": 0.9067540544588035, "grad_norm": 1.2915462255477905, "learning_rate": 4.630926866836893e-07, "loss": 0.8267, "step": 25733 }, { "epoch": 0.9067892914717619, "grad_norm": 1.2097063064575195, "learning_rate": 4.627494653497211e-07, "loss": 1.2231, "step": 25734 }, { "epoch": 0.9068245284847204, "grad_norm": 4.477311134338379, "learning_rate": 4.6240636823965625e-07, "loss": 5.4762, "step": 25735 }, { "epoch": 0.9068597654976788, "grad_norm": 1.9099273681640625, "learning_rate": 4.620633953579645e-07, "loss": 1.9664, "step": 25736 }, { "epoch": 0.9068950025106371, "grad_norm": 3.5141546726226807, "learning_rate": 4.6172054670911015e-07, "loss": 4.6395, "step": 25737 }, { "epoch": 0.9069302395235956, "grad_norm": 7.291191577911377, "learning_rate": 4.613778222975618e-07, "loss": 8.1444, "step": 25738 }, { "epoch": 0.906965476536554, "grad_norm": 1.0688623189926147, "learning_rate": 4.6103522212778364e-07, "loss": 0.8288, "step": 25739 }, { "epoch": 0.9070007135495124, "grad_norm": 4.820188045501709, "learning_rate": 4.606927462042343e-07, "loss": 3.1663, "step": 25740 }, { "epoch": 0.9070359505624708, "grad_norm": 1.137266755104065, "learning_rate": 4.603503945313781e-07, "loss": 0.8073, "step": 25741 }, { "epoch": 0.9070711875754293, "grad_norm": 1.3592588901519775, "learning_rate": 4.600081671136725e-07, "loss": 0.8687, "step": 25742 }, { "epoch": 0.9071064245883876, "grad_norm": 1.243594765663147, "learning_rate": 4.596660639555783e-07, "loss": 0.6974, "step": 25743 }, { "epoch": 0.907141661601346, "grad_norm": 0.8556162118911743, "learning_rate": 4.593240850615455e-07, "loss": 0.7423, "step": 25744 }, { "epoch": 0.9071768986143045, "grad_norm": 2.233808994293213, "learning_rate": 4.5898223043603254e-07, "loss": 0.6727, "step": 25745 }, { "epoch": 0.9072121356272629, "grad_norm": 3.3933510780334473, "learning_rate": 4.586405000834937e-07, "loss": 2.2547, "step": 25746 }, { "epoch": 0.9072473726402213, "grad_norm": 4.369805335998535, "learning_rate": 4.5829889400837544e-07, "loss": 3.0415, "step": 25747 }, { "epoch": 0.9072826096531798, "grad_norm": 1.046022891998291, "learning_rate": 4.579574122151287e-07, "loss": 0.7117, "step": 25748 }, { "epoch": 0.9073178466661381, "grad_norm": 3.7064363956451416, "learning_rate": 4.576160547082054e-07, "loss": 4.4583, "step": 25749 }, { "epoch": 0.9073530836790965, "grad_norm": 1.1221327781677246, "learning_rate": 4.572748214920464e-07, "loss": 1.1846, "step": 25750 }, { "epoch": 0.9073883206920549, "grad_norm": 3.529237747192383, "learning_rate": 4.569337125710982e-07, "loss": 2.7362, "step": 25751 }, { "epoch": 0.9074235577050134, "grad_norm": 3.5643389225006104, "learning_rate": 4.56592727949805e-07, "loss": 2.9664, "step": 25752 }, { "epoch": 0.9074587947179718, "grad_norm": 3.9033024311065674, "learning_rate": 4.5625186763260663e-07, "loss": 3.1093, "step": 25753 }, { "epoch": 0.9074940317309301, "grad_norm": 1.9731158018112183, "learning_rate": 4.559111316239428e-07, "loss": 0.838, "step": 25754 }, { "epoch": 0.9075292687438886, "grad_norm": 9.647064208984375, "learning_rate": 4.5557051992825343e-07, "loss": 3.5346, "step": 25755 }, { "epoch": 0.907564505756847, "grad_norm": 14.030573844909668, "learning_rate": 4.5523003254997255e-07, "loss": 5.2514, "step": 25756 }, { "epoch": 0.9075997427698054, "grad_norm": 4.945464611053467, "learning_rate": 4.5488966949353675e-07, "loss": 4.8732, "step": 25757 }, { "epoch": 0.9076349797827639, "grad_norm": 3.166172742843628, "learning_rate": 4.545494307633802e-07, "loss": 2.9513, "step": 25758 }, { "epoch": 0.9076702167957222, "grad_norm": 1.0780205726623535, "learning_rate": 4.542093163639305e-07, "loss": 0.8871, "step": 25759 }, { "epoch": 0.9077054538086806, "grad_norm": 0.9586610198020935, "learning_rate": 4.538693262996219e-07, "loss": 0.7355, "step": 25760 }, { "epoch": 0.907740690821639, "grad_norm": 1.09907066822052, "learning_rate": 4.535294605748808e-07, "loss": 0.8631, "step": 25761 }, { "epoch": 0.9077759278345975, "grad_norm": 1.463354468345642, "learning_rate": 4.53189719194137e-07, "loss": 0.7468, "step": 25762 }, { "epoch": 0.9078111648475559, "grad_norm": 4.258529186248779, "learning_rate": 4.5285010216180927e-07, "loss": 4.5556, "step": 25763 }, { "epoch": 0.9078464018605142, "grad_norm": 5.179012775421143, "learning_rate": 4.525106094823273e-07, "loss": 3.2648, "step": 25764 }, { "epoch": 0.9078816388734727, "grad_norm": 1.485632061958313, "learning_rate": 4.52171241160112e-07, "loss": 1.028, "step": 25765 }, { "epoch": 0.9079168758864311, "grad_norm": 1.3636010885238647, "learning_rate": 4.5183199719957995e-07, "loss": 0.9951, "step": 25766 }, { "epoch": 0.9079521128993895, "grad_norm": 1.5651395320892334, "learning_rate": 4.514928776051519e-07, "loss": 0.7713, "step": 25767 }, { "epoch": 0.907987349912348, "grad_norm": 4.45813512802124, "learning_rate": 4.511538823812489e-07, "loss": 4.8689, "step": 25768 }, { "epoch": 0.9080225869253064, "grad_norm": 3.2394862174987793, "learning_rate": 4.508150115322807e-07, "loss": 3.1021, "step": 25769 }, { "epoch": 0.9080578239382647, "grad_norm": 5.514962196350098, "learning_rate": 4.504762650626637e-07, "loss": 3.1543, "step": 25770 }, { "epoch": 0.9080930609512232, "grad_norm": 4.600673675537109, "learning_rate": 4.5013764297680893e-07, "loss": 2.5998, "step": 25771 }, { "epoch": 0.9081282979641816, "grad_norm": 1.0752277374267578, "learning_rate": 4.4979914527912837e-07, "loss": 0.9688, "step": 25772 }, { "epoch": 0.90816353497714, "grad_norm": 3.1120359897613525, "learning_rate": 4.494607719740307e-07, "loss": 2.5464, "step": 25773 }, { "epoch": 0.9081987719900984, "grad_norm": 2.0832855701446533, "learning_rate": 4.491225230659235e-07, "loss": 0.8656, "step": 25774 }, { "epoch": 0.9082340090030568, "grad_norm": 5.34688663482666, "learning_rate": 4.487843985592111e-07, "loss": 6.9329, "step": 25775 }, { "epoch": 0.9082692460160152, "grad_norm": 4.9987592697143555, "learning_rate": 4.4844639845829875e-07, "loss": 5.1349, "step": 25776 }, { "epoch": 0.9083044830289736, "grad_norm": 2.7107203006744385, "learning_rate": 4.4810852276758967e-07, "loss": 3.1082, "step": 25777 }, { "epoch": 0.9083397200419321, "grad_norm": 4.516766548156738, "learning_rate": 4.4777077149148143e-07, "loss": 2.7571, "step": 25778 }, { "epoch": 0.9083749570548905, "grad_norm": 11.290573120117188, "learning_rate": 4.4743314463437717e-07, "loss": 4.9794, "step": 25779 }, { "epoch": 0.9084101940678488, "grad_norm": 0.8869245648384094, "learning_rate": 4.470956422006734e-07, "loss": 0.9169, "step": 25780 }, { "epoch": 0.9084454310808073, "grad_norm": 2.2255358695983887, "learning_rate": 4.4675826419476656e-07, "loss": 2.3482, "step": 25781 }, { "epoch": 0.9084806680937657, "grad_norm": 3.1555521488189697, "learning_rate": 4.464210106210487e-07, "loss": 2.3811, "step": 25782 }, { "epoch": 0.9085159051067241, "grad_norm": 1.6676011085510254, "learning_rate": 4.46083881483913e-07, "loss": 0.9719, "step": 25783 }, { "epoch": 0.9085511421196825, "grad_norm": 2.296147346496582, "learning_rate": 4.4574687678775486e-07, "loss": 2.583, "step": 25784 }, { "epoch": 0.908586379132641, "grad_norm": 1.1191693544387817, "learning_rate": 4.454099965369585e-07, "loss": 0.7153, "step": 25785 }, { "epoch": 0.9086216161455993, "grad_norm": 3.3102023601531982, "learning_rate": 4.450732407359148e-07, "loss": 2.7926, "step": 25786 }, { "epoch": 0.9086568531585577, "grad_norm": 12.530750274658203, "learning_rate": 4.447366093890082e-07, "loss": 2.85, "step": 25787 }, { "epoch": 0.9086920901715162, "grad_norm": 7.720396518707275, "learning_rate": 4.4440010250062504e-07, "loss": 8.8914, "step": 25788 }, { "epoch": 0.9087273271844746, "grad_norm": 1.0166007280349731, "learning_rate": 4.440637200751485e-07, "loss": 0.6826, "step": 25789 }, { "epoch": 0.908762564197433, "grad_norm": 4.043471813201904, "learning_rate": 4.437274621169585e-07, "loss": 3.2454, "step": 25790 }, { "epoch": 0.9087978012103914, "grad_norm": 1.106859803199768, "learning_rate": 4.4339132863043477e-07, "loss": 1.0628, "step": 25791 }, { "epoch": 0.9088330382233498, "grad_norm": 3.0430080890655518, "learning_rate": 4.430553196199572e-07, "loss": 3.0229, "step": 25792 }, { "epoch": 0.9088682752363082, "grad_norm": 2.975611686706543, "learning_rate": 4.427194350899022e-07, "loss": 2.4339, "step": 25793 }, { "epoch": 0.9089035122492666, "grad_norm": 1.4075534343719482, "learning_rate": 4.42383675044642e-07, "loss": 0.8108, "step": 25794 }, { "epoch": 0.9089387492622251, "grad_norm": 1.308210849761963, "learning_rate": 4.420480394885529e-07, "loss": 0.9956, "step": 25795 }, { "epoch": 0.9089739862751834, "grad_norm": 1.5007823705673218, "learning_rate": 4.4171252842600597e-07, "loss": 0.9042, "step": 25796 }, { "epoch": 0.9090092232881418, "grad_norm": 1.3651925325393677, "learning_rate": 4.4137714186137216e-07, "loss": 0.6527, "step": 25797 }, { "epoch": 0.9090444603011003, "grad_norm": 3.9865314960479736, "learning_rate": 4.410418797990157e-07, "loss": 3.0724, "step": 25798 }, { "epoch": 0.9090796973140587, "grad_norm": 5.680476188659668, "learning_rate": 4.407067422433076e-07, "loss": 5.4157, "step": 25799 }, { "epoch": 0.9091149343270171, "grad_norm": 0.979973316192627, "learning_rate": 4.403717291986132e-07, "loss": 0.9427, "step": 25800 }, { "epoch": 0.9091501713399756, "grad_norm": 5.435837745666504, "learning_rate": 4.400368406692934e-07, "loss": 5.2202, "step": 25801 }, { "epoch": 0.9091854083529339, "grad_norm": 1.2551124095916748, "learning_rate": 4.397020766597104e-07, "loss": 0.6885, "step": 25802 }, { "epoch": 0.9092206453658923, "grad_norm": 1.831255316734314, "learning_rate": 4.3936743717422845e-07, "loss": 0.9836, "step": 25803 }, { "epoch": 0.9092558823788508, "grad_norm": 4.668447971343994, "learning_rate": 4.3903292221720185e-07, "loss": 6.5072, "step": 25804 }, { "epoch": 0.9092911193918092, "grad_norm": 9.694064140319824, "learning_rate": 4.386985317929893e-07, "loss": 4.6727, "step": 25805 }, { "epoch": 0.9093263564047676, "grad_norm": 8.453004837036133, "learning_rate": 4.383642659059462e-07, "loss": 6.9895, "step": 25806 }, { "epoch": 0.9093615934177259, "grad_norm": 2.9667375087738037, "learning_rate": 4.3803012456042684e-07, "loss": 2.7108, "step": 25807 }, { "epoch": 0.9093968304306844, "grad_norm": 2.810519218444824, "learning_rate": 4.376961077607822e-07, "loss": 3.0096, "step": 25808 }, { "epoch": 0.9094320674436428, "grad_norm": 3.456829786300659, "learning_rate": 4.3736221551136436e-07, "loss": 2.8467, "step": 25809 }, { "epoch": 0.9094673044566012, "grad_norm": 7.833252429962158, "learning_rate": 4.370284478165221e-07, "loss": 4.6798, "step": 25810 }, { "epoch": 0.9095025414695597, "grad_norm": 1.0591281652450562, "learning_rate": 4.3669480468060186e-07, "loss": 0.9955, "step": 25811 }, { "epoch": 0.909537778482518, "grad_norm": 1.3896325826644897, "learning_rate": 4.363612861079525e-07, "loss": 0.7639, "step": 25812 }, { "epoch": 0.9095730154954764, "grad_norm": 1.0618312358856201, "learning_rate": 4.3602789210291154e-07, "loss": 0.8509, "step": 25813 }, { "epoch": 0.9096082525084349, "grad_norm": 2.2714040279388428, "learning_rate": 4.3569462266982777e-07, "loss": 2.7371, "step": 25814 }, { "epoch": 0.9096434895213933, "grad_norm": 3.748685121536255, "learning_rate": 4.3536147781303883e-07, "loss": 2.4985, "step": 25815 }, { "epoch": 0.9096787265343517, "grad_norm": 1.3189404010772705, "learning_rate": 4.350284575368868e-07, "loss": 0.7923, "step": 25816 }, { "epoch": 0.90971396354731, "grad_norm": 2.9828286170959473, "learning_rate": 4.3469556184570494e-07, "loss": 2.8063, "step": 25817 }, { "epoch": 0.9097492005602685, "grad_norm": 1.2922779321670532, "learning_rate": 4.3436279074383415e-07, "loss": 0.7893, "step": 25818 }, { "epoch": 0.9097844375732269, "grad_norm": 4.099842071533203, "learning_rate": 4.340301442356065e-07, "loss": 2.6698, "step": 25819 }, { "epoch": 0.9098196745861853, "grad_norm": 3.793766736984253, "learning_rate": 4.3369762232535306e-07, "loss": 2.9938, "step": 25820 }, { "epoch": 0.9098549115991438, "grad_norm": 2.5102789402008057, "learning_rate": 4.3336522501740696e-07, "loss": 2.8338, "step": 25821 }, { "epoch": 0.9098901486121022, "grad_norm": 2.9156765937805176, "learning_rate": 4.330329523160992e-07, "loss": 0.7376, "step": 25822 }, { "epoch": 0.9099253856250605, "grad_norm": 2.1629536151885986, "learning_rate": 4.327008042257541e-07, "loss": 1.3476, "step": 25823 }, { "epoch": 0.909960622638019, "grad_norm": 2.4258222579956055, "learning_rate": 4.3236878075069933e-07, "loss": 2.5568, "step": 25824 }, { "epoch": 0.9099958596509774, "grad_norm": 1.2448714971542358, "learning_rate": 4.3203688189526137e-07, "loss": 1.0165, "step": 25825 }, { "epoch": 0.9100310966639358, "grad_norm": 1.1919608116149902, "learning_rate": 4.317051076637613e-07, "loss": 0.9168, "step": 25826 }, { "epoch": 0.9100663336768942, "grad_norm": 1.0596113204956055, "learning_rate": 4.3137345806052e-07, "loss": 0.9122, "step": 25827 }, { "epoch": 0.9101015706898526, "grad_norm": 2.883021354675293, "learning_rate": 4.3104193308985966e-07, "loss": 0.9585, "step": 25828 }, { "epoch": 0.910136807702811, "grad_norm": 1.525286316871643, "learning_rate": 4.3071053275609564e-07, "loss": 0.6064, "step": 25829 }, { "epoch": 0.9101720447157694, "grad_norm": 2.8659656047821045, "learning_rate": 4.3037925706354677e-07, "loss": 2.6412, "step": 25830 }, { "epoch": 0.9102072817287279, "grad_norm": 1.6545828580856323, "learning_rate": 4.3004810601652845e-07, "loss": 0.6989, "step": 25831 }, { "epoch": 0.9102425187416863, "grad_norm": 4.359993934631348, "learning_rate": 4.297170796193495e-07, "loss": 2.6959, "step": 25832 }, { "epoch": 0.9102777557546446, "grad_norm": 1.6571310758590698, "learning_rate": 4.2938617787632643e-07, "loss": 0.8949, "step": 25833 }, { "epoch": 0.9103129927676031, "grad_norm": 2.213122606277466, "learning_rate": 4.290554007917669e-07, "loss": 2.464, "step": 25834 }, { "epoch": 0.9103482297805615, "grad_norm": 7.106713771820068, "learning_rate": 4.287247483699819e-07, "loss": 4.858, "step": 25835 }, { "epoch": 0.9103834667935199, "grad_norm": 1.3018478155136108, "learning_rate": 4.2839422061527356e-07, "loss": 0.8607, "step": 25836 }, { "epoch": 0.9104187038064784, "grad_norm": 2.3940138816833496, "learning_rate": 4.280638175319507e-07, "loss": 2.5534, "step": 25837 }, { "epoch": 0.9104539408194368, "grad_norm": 3.977506160736084, "learning_rate": 4.277335391243165e-07, "loss": 3.4031, "step": 25838 }, { "epoch": 0.9104891778323951, "grad_norm": 1.3886650800704956, "learning_rate": 4.274033853966719e-07, "loss": 0.9274, "step": 25839 }, { "epoch": 0.9105244148453535, "grad_norm": 3.5130205154418945, "learning_rate": 4.2707335635331584e-07, "loss": 2.3275, "step": 25840 }, { "epoch": 0.910559651858312, "grad_norm": 1.075364589691162, "learning_rate": 4.2674345199855136e-07, "loss": 0.7779, "step": 25841 }, { "epoch": 0.9105948888712704, "grad_norm": 8.054269790649414, "learning_rate": 4.264136723366719e-07, "loss": 2.8333, "step": 25842 }, { "epoch": 0.9106301258842288, "grad_norm": 4.54285192489624, "learning_rate": 4.260840173719738e-07, "loss": 4.6236, "step": 25843 }, { "epoch": 0.9106653628971872, "grad_norm": 1.1410613059997559, "learning_rate": 4.2575448710875044e-07, "loss": 0.7516, "step": 25844 }, { "epoch": 0.9107005999101456, "grad_norm": 5.147408962249756, "learning_rate": 4.25425081551295e-07, "loss": 3.0151, "step": 25845 }, { "epoch": 0.910735836923104, "grad_norm": 5.6409220695495605, "learning_rate": 4.2509580070389743e-07, "loss": 6.4742, "step": 25846 }, { "epoch": 0.9107710739360625, "grad_norm": 4.89823579788208, "learning_rate": 4.247666445708465e-07, "loss": 5.3618, "step": 25847 }, { "epoch": 0.9108063109490209, "grad_norm": 2.1550955772399902, "learning_rate": 4.2443761315642986e-07, "loss": 2.0635, "step": 25848 }, { "epoch": 0.9108415479619792, "grad_norm": 1.145061731338501, "learning_rate": 4.2410870646493297e-07, "loss": 0.8349, "step": 25849 }, { "epoch": 0.9108767849749376, "grad_norm": 1.3412684202194214, "learning_rate": 4.237799245006413e-07, "loss": 0.8273, "step": 25850 }, { "epoch": 0.9109120219878961, "grad_norm": 7.129269123077393, "learning_rate": 4.234512672678326e-07, "loss": 5.4517, "step": 25851 }, { "epoch": 0.9109472590008545, "grad_norm": 3.3542850017547607, "learning_rate": 4.231227347707934e-07, "loss": 2.8007, "step": 25852 }, { "epoch": 0.9109824960138129, "grad_norm": 1.1925485134124756, "learning_rate": 4.227943270138002e-07, "loss": 1.0091, "step": 25853 }, { "epoch": 0.9110177330267714, "grad_norm": 1.341925024986267, "learning_rate": 4.224660440011319e-07, "loss": 0.6686, "step": 25854 }, { "epoch": 0.9110529700397297, "grad_norm": 2.79555606842041, "learning_rate": 4.221378857370606e-07, "loss": 2.456, "step": 25855 }, { "epoch": 0.9110882070526881, "grad_norm": 1.3107396364212036, "learning_rate": 4.2180985222586513e-07, "loss": 0.7095, "step": 25856 }, { "epoch": 0.9111234440656466, "grad_norm": 1.0766479969024658, "learning_rate": 4.2148194347181646e-07, "loss": 0.7937, "step": 25857 }, { "epoch": 0.911158681078605, "grad_norm": 9.052790641784668, "learning_rate": 4.211541594791857e-07, "loss": 4.1559, "step": 25858 }, { "epoch": 0.9111939180915634, "grad_norm": 3.8200321197509766, "learning_rate": 4.208265002522405e-07, "loss": 3.0144, "step": 25859 }, { "epoch": 0.9112291551045217, "grad_norm": 2.908538818359375, "learning_rate": 4.2049896579525297e-07, "loss": 2.7625, "step": 25860 }, { "epoch": 0.9112643921174802, "grad_norm": 2.869565963745117, "learning_rate": 4.201715561124853e-07, "loss": 2.5339, "step": 25861 }, { "epoch": 0.9112996291304386, "grad_norm": 1.541292667388916, "learning_rate": 4.1984427120820294e-07, "loss": 0.7726, "step": 25862 }, { "epoch": 0.911334866143397, "grad_norm": 1.9345474243164062, "learning_rate": 4.1951711108667025e-07, "loss": 0.864, "step": 25863 }, { "epoch": 0.9113701031563555, "grad_norm": 1.119128704071045, "learning_rate": 4.191900757521461e-07, "loss": 0.935, "step": 25864 }, { "epoch": 0.9114053401693138, "grad_norm": 5.772012710571289, "learning_rate": 4.188631652088926e-07, "loss": 5.4487, "step": 25865 }, { "epoch": 0.9114405771822722, "grad_norm": 1.5473500490188599, "learning_rate": 4.1853637946116633e-07, "loss": 0.7719, "step": 25866 }, { "epoch": 0.9114758141952307, "grad_norm": 6.318759918212891, "learning_rate": 4.18209718513225e-07, "loss": 2.3996, "step": 25867 }, { "epoch": 0.9115110512081891, "grad_norm": 1.3800251483917236, "learning_rate": 4.178831823693219e-07, "loss": 0.8999, "step": 25868 }, { "epoch": 0.9115462882211475, "grad_norm": 0.9331409931182861, "learning_rate": 4.175567710337114e-07, "loss": 0.7519, "step": 25869 }, { "epoch": 0.9115815252341058, "grad_norm": 4.747786045074463, "learning_rate": 4.172304845106445e-07, "loss": 2.6641, "step": 25870 }, { "epoch": 0.9116167622470643, "grad_norm": 3.7084128856658936, "learning_rate": 4.169043228043712e-07, "loss": 4.5776, "step": 25871 }, { "epoch": 0.9116519992600227, "grad_norm": 6.163794994354248, "learning_rate": 4.165782859191392e-07, "loss": 3.047, "step": 25872 }, { "epoch": 0.9116872362729811, "grad_norm": 1.858371615409851, "learning_rate": 4.1625237385919726e-07, "loss": 0.7978, "step": 25873 }, { "epoch": 0.9117224732859396, "grad_norm": 9.94427490234375, "learning_rate": 4.159265866287865e-07, "loss": 4.8944, "step": 25874 }, { "epoch": 0.911757710298898, "grad_norm": 6.048743724822998, "learning_rate": 4.156009242321546e-07, "loss": 5.2016, "step": 25875 }, { "epoch": 0.9117929473118563, "grad_norm": 1.3433656692504883, "learning_rate": 4.152753866735426e-07, "loss": 1.0162, "step": 25876 }, { "epoch": 0.9118281843248148, "grad_norm": 3.5564184188842773, "learning_rate": 4.1494997395718715e-07, "loss": 3.1125, "step": 25877 }, { "epoch": 0.9118634213377732, "grad_norm": 3.537749767303467, "learning_rate": 4.146246860873293e-07, "loss": 0.5208, "step": 25878 }, { "epoch": 0.9118986583507316, "grad_norm": 5.322141647338867, "learning_rate": 4.1429952306820787e-07, "loss": 4.7553, "step": 25879 }, { "epoch": 0.9119338953636901, "grad_norm": 1.204703688621521, "learning_rate": 4.13974484904055e-07, "loss": 0.703, "step": 25880 }, { "epoch": 0.9119691323766485, "grad_norm": 1.0629557371139526, "learning_rate": 4.136495715991051e-07, "loss": 0.7027, "step": 25881 }, { "epoch": 0.9120043693896068, "grad_norm": 1.69528067111969, "learning_rate": 4.133247831575915e-07, "loss": 0.8744, "step": 25882 }, { "epoch": 0.9120396064025652, "grad_norm": 1.8415483236312866, "learning_rate": 4.13000119583743e-07, "loss": 1.1824, "step": 25883 }, { "epoch": 0.9120748434155237, "grad_norm": 2.3460025787353516, "learning_rate": 4.126755808817895e-07, "loss": 2.6526, "step": 25884 }, { "epoch": 0.9121100804284821, "grad_norm": 1.0778642892837524, "learning_rate": 4.123511670559577e-07, "loss": 0.9146, "step": 25885 }, { "epoch": 0.9121453174414405, "grad_norm": 2.545172929763794, "learning_rate": 4.12026878110473e-07, "loss": 2.6091, "step": 25886 }, { "epoch": 0.9121805544543989, "grad_norm": 1.6719316244125366, "learning_rate": 4.1170271404955996e-07, "loss": 0.7483, "step": 25887 }, { "epoch": 0.9122157914673573, "grad_norm": 1.7703667879104614, "learning_rate": 4.1137867487744067e-07, "loss": 0.8791, "step": 25888 }, { "epoch": 0.9122510284803157, "grad_norm": 1.1046704053878784, "learning_rate": 4.1105476059833505e-07, "loss": 0.9616, "step": 25889 }, { "epoch": 0.9122862654932742, "grad_norm": 2.0688705444335938, "learning_rate": 4.1073097121646197e-07, "loss": 2.1909, "step": 25890 }, { "epoch": 0.9123215025062326, "grad_norm": 6.167977809906006, "learning_rate": 4.1040730673604035e-07, "loss": 5.07, "step": 25891 }, { "epoch": 0.9123567395191909, "grad_norm": 1.6231188774108887, "learning_rate": 4.100837671612867e-07, "loss": 0.8827, "step": 25892 }, { "epoch": 0.9123919765321493, "grad_norm": 1.3597588539123535, "learning_rate": 4.0976035249641e-07, "loss": 0.7096, "step": 25893 }, { "epoch": 0.9124272135451078, "grad_norm": 1.5527747869491577, "learning_rate": 4.094370627456279e-07, "loss": 0.8768, "step": 25894 }, { "epoch": 0.9124624505580662, "grad_norm": 1.179810881614685, "learning_rate": 4.091138979131515e-07, "loss": 0.7919, "step": 25895 }, { "epoch": 0.9124976875710246, "grad_norm": 1.0997486114501953, "learning_rate": 4.087908580031863e-07, "loss": 0.8536, "step": 25896 }, { "epoch": 0.912532924583983, "grad_norm": 0.9954808354377747, "learning_rate": 4.084679430199412e-07, "loss": 0.7546, "step": 25897 }, { "epoch": 0.9125681615969414, "grad_norm": 3.2995612621307373, "learning_rate": 4.0814515296762503e-07, "loss": 2.6781, "step": 25898 }, { "epoch": 0.9126033986098998, "grad_norm": 6.538156032562256, "learning_rate": 4.078224878504389e-07, "loss": 6.0946, "step": 25899 }, { "epoch": 0.9126386356228583, "grad_norm": 3.2385246753692627, "learning_rate": 4.074999476725872e-07, "loss": 3.3586, "step": 25900 }, { "epoch": 0.9126738726358167, "grad_norm": 1.2734959125518799, "learning_rate": 4.0717753243826985e-07, "loss": 0.6918, "step": 25901 }, { "epoch": 0.912709109648775, "grad_norm": 4.843869686126709, "learning_rate": 4.068552421516869e-07, "loss": 3.3472, "step": 25902 }, { "epoch": 0.9127443466617334, "grad_norm": 3.588874578475952, "learning_rate": 4.0653307681703725e-07, "loss": 3.4097, "step": 25903 }, { "epoch": 0.9127795836746919, "grad_norm": 7.538430690765381, "learning_rate": 4.0621103643851524e-07, "loss": 3.1501, "step": 25904 }, { "epoch": 0.9128148206876503, "grad_norm": 0.9727193713188171, "learning_rate": 4.0588912102031755e-07, "loss": 0.6926, "step": 25905 }, { "epoch": 0.9128500577006087, "grad_norm": 1.0802127122879028, "learning_rate": 4.0556733056663635e-07, "loss": 0.9674, "step": 25906 }, { "epoch": 0.9128852947135672, "grad_norm": 2.7992494106292725, "learning_rate": 4.0524566508166166e-07, "loss": 2.4196, "step": 25907 }, { "epoch": 0.9129205317265255, "grad_norm": 1.7173963785171509, "learning_rate": 4.0492412456958453e-07, "loss": 0.6432, "step": 25908 }, { "epoch": 0.9129557687394839, "grad_norm": 1.6953108310699463, "learning_rate": 4.046027090345939e-07, "loss": 0.9322, "step": 25909 }, { "epoch": 0.9129910057524424, "grad_norm": 4.286304473876953, "learning_rate": 4.042814184808741e-07, "loss": 5.4202, "step": 25910 }, { "epoch": 0.9130262427654008, "grad_norm": 1.0177462100982666, "learning_rate": 4.03960252912613e-07, "loss": 0.8378, "step": 25911 }, { "epoch": 0.9130614797783592, "grad_norm": 3.0175411701202393, "learning_rate": 4.036392123339894e-07, "loss": 2.4426, "step": 25912 }, { "epoch": 0.9130967167913177, "grad_norm": 5.269820690155029, "learning_rate": 4.033182967491889e-07, "loss": 4.9418, "step": 25913 }, { "epoch": 0.913131953804276, "grad_norm": 1.141982078552246, "learning_rate": 4.029975061623914e-07, "loss": 1.2311, "step": 25914 }, { "epoch": 0.9131671908172344, "grad_norm": 4.040044784545898, "learning_rate": 4.0267684057777147e-07, "loss": 2.8611, "step": 25915 }, { "epoch": 0.9132024278301928, "grad_norm": 1.173588752746582, "learning_rate": 4.0235629999950786e-07, "loss": 0.9301, "step": 25916 }, { "epoch": 0.9132376648431513, "grad_norm": 1.1290148496627808, "learning_rate": 4.0203588443177845e-07, "loss": 0.7184, "step": 25917 }, { "epoch": 0.9132729018561097, "grad_norm": 3.4749956130981445, "learning_rate": 4.0171559387875314e-07, "loss": 3.0061, "step": 25918 }, { "epoch": 0.913308138869068, "grad_norm": 0.9771777987480164, "learning_rate": 4.013954283446042e-07, "loss": 0.7966, "step": 25919 }, { "epoch": 0.9133433758820265, "grad_norm": 1.5942420959472656, "learning_rate": 4.0107538783350274e-07, "loss": 0.809, "step": 25920 }, { "epoch": 0.9133786128949849, "grad_norm": 4.028266429901123, "learning_rate": 4.0075547234961654e-07, "loss": 3.344, "step": 25921 }, { "epoch": 0.9134138499079433, "grad_norm": 5.471482753753662, "learning_rate": 4.0043568189711447e-07, "loss": 2.7823, "step": 25922 }, { "epoch": 0.9134490869209018, "grad_norm": 1.1828076839447021, "learning_rate": 4.0011601648015986e-07, "loss": 0.8951, "step": 25923 }, { "epoch": 0.9134843239338601, "grad_norm": 1.7952065467834473, "learning_rate": 3.9979647610291603e-07, "loss": 0.8765, "step": 25924 }, { "epoch": 0.9135195609468185, "grad_norm": 2.494032621383667, "learning_rate": 3.994770607695464e-07, "loss": 2.9762, "step": 25925 }, { "epoch": 0.9135547979597769, "grad_norm": 3.5847456455230713, "learning_rate": 3.99157770484212e-07, "loss": 2.3777, "step": 25926 }, { "epoch": 0.9135900349727354, "grad_norm": 4.710859298706055, "learning_rate": 3.988386052510695e-07, "loss": 4.3335, "step": 25927 }, { "epoch": 0.9136252719856938, "grad_norm": 0.9653425216674805, "learning_rate": 3.985195650742779e-07, "loss": 0.9217, "step": 25928 }, { "epoch": 0.9136605089986521, "grad_norm": 3.7626383304595947, "learning_rate": 3.982006499579916e-07, "loss": 4.7336, "step": 25929 }, { "epoch": 0.9136957460116106, "grad_norm": 2.894883155822754, "learning_rate": 3.9788185990636605e-07, "loss": 2.6851, "step": 25930 }, { "epoch": 0.913730983024569, "grad_norm": 0.9670376181602478, "learning_rate": 3.975631949235492e-07, "loss": 1.0756, "step": 25931 }, { "epoch": 0.9137662200375274, "grad_norm": 5.190442085266113, "learning_rate": 3.972446550136977e-07, "loss": 2.9783, "step": 25932 }, { "epoch": 0.9138014570504859, "grad_norm": 5.4213409423828125, "learning_rate": 3.969262401809582e-07, "loss": 3.059, "step": 25933 }, { "epoch": 0.9138366940634443, "grad_norm": 1.5367558002471924, "learning_rate": 3.966079504294762e-07, "loss": 1.1902, "step": 25934 }, { "epoch": 0.9138719310764026, "grad_norm": 3.4096477031707764, "learning_rate": 3.9628978576339736e-07, "loss": 2.3083, "step": 25935 }, { "epoch": 0.913907168089361, "grad_norm": 7.140780925750732, "learning_rate": 3.959717461868706e-07, "loss": 4.916, "step": 25936 }, { "epoch": 0.9139424051023195, "grad_norm": 8.602075576782227, "learning_rate": 3.956538317040337e-07, "loss": 7.2228, "step": 25937 }, { "epoch": 0.9139776421152779, "grad_norm": 4.625346660614014, "learning_rate": 3.953360423190289e-07, "loss": 5.0146, "step": 25938 }, { "epoch": 0.9140128791282363, "grad_norm": 1.546872854232788, "learning_rate": 3.950183780359962e-07, "loss": 0.9343, "step": 25939 }, { "epoch": 0.9140481161411947, "grad_norm": 1.158394455909729, "learning_rate": 3.947008388590723e-07, "loss": 0.6874, "step": 25940 }, { "epoch": 0.9140833531541531, "grad_norm": 1.9435175657272339, "learning_rate": 3.9438342479239386e-07, "loss": 0.7526, "step": 25941 }, { "epoch": 0.9141185901671115, "grad_norm": 4.954807281494141, "learning_rate": 3.9406613584009436e-07, "loss": 3.4578, "step": 25942 }, { "epoch": 0.91415382718007, "grad_norm": 1.475527048110962, "learning_rate": 3.93748972006307e-07, "loss": 0.875, "step": 25943 }, { "epoch": 0.9141890641930284, "grad_norm": 4.137217044830322, "learning_rate": 3.9343193329516305e-07, "loss": 4.2019, "step": 25944 }, { "epoch": 0.9142243012059867, "grad_norm": 3.2298035621643066, "learning_rate": 3.9311501971079135e-07, "loss": 3.2007, "step": 25945 }, { "epoch": 0.9142595382189452, "grad_norm": 4.3891825675964355, "learning_rate": 3.9279823125732084e-07, "loss": 0.8596, "step": 25946 }, { "epoch": 0.9142947752319036, "grad_norm": 1.1258916854858398, "learning_rate": 3.9248156793887715e-07, "loss": 0.8259, "step": 25947 }, { "epoch": 0.914330012244862, "grad_norm": 1.2524631023406982, "learning_rate": 3.921650297595847e-07, "loss": 0.7262, "step": 25948 }, { "epoch": 0.9143652492578204, "grad_norm": 1.3646496534347534, "learning_rate": 3.91848616723568e-07, "loss": 1.1708, "step": 25949 }, { "epoch": 0.9144004862707789, "grad_norm": 0.925569474697113, "learning_rate": 3.9153232883494376e-07, "loss": 1.0387, "step": 25950 }, { "epoch": 0.9144357232837372, "grad_norm": 0.9539212584495544, "learning_rate": 3.9121616609783643e-07, "loss": 0.7059, "step": 25951 }, { "epoch": 0.9144709602966956, "grad_norm": 0.901053249835968, "learning_rate": 3.909001285163638e-07, "loss": 1.0104, "step": 25952 }, { "epoch": 0.9145061973096541, "grad_norm": 2.372467517852783, "learning_rate": 3.9058421609464046e-07, "loss": 2.2916, "step": 25953 }, { "epoch": 0.9145414343226125, "grad_norm": 1.069379448890686, "learning_rate": 3.9026842883677975e-07, "loss": 0.8169, "step": 25954 }, { "epoch": 0.9145766713355709, "grad_norm": 1.0475612878799438, "learning_rate": 3.899527667468994e-07, "loss": 0.7431, "step": 25955 }, { "epoch": 0.9146119083485293, "grad_norm": 6.9642510414123535, "learning_rate": 3.8963722982910734e-07, "loss": 5.711, "step": 25956 }, { "epoch": 0.9146471453614877, "grad_norm": 5.071660995483398, "learning_rate": 3.893218180875147e-07, "loss": 3.2512, "step": 25957 }, { "epoch": 0.9146823823744461, "grad_norm": 2.8150815963745117, "learning_rate": 3.8900653152622925e-07, "loss": 2.3894, "step": 25958 }, { "epoch": 0.9147176193874045, "grad_norm": 5.315324783325195, "learning_rate": 3.8869137014935776e-07, "loss": 2.5999, "step": 25959 }, { "epoch": 0.914752856400363, "grad_norm": 1.2026840448379517, "learning_rate": 3.883763339610058e-07, "loss": 0.7916, "step": 25960 }, { "epoch": 0.9147880934133213, "grad_norm": 1.028310775756836, "learning_rate": 3.880614229652768e-07, "loss": 0.7704, "step": 25961 }, { "epoch": 0.9148233304262797, "grad_norm": 1.5243251323699951, "learning_rate": 3.877466371662719e-07, "loss": 0.7836, "step": 25962 }, { "epoch": 0.9148585674392382, "grad_norm": 1.6159322261810303, "learning_rate": 3.8743197656809117e-07, "loss": 0.9249, "step": 25963 }, { "epoch": 0.9148938044521966, "grad_norm": 8.236948013305664, "learning_rate": 3.871174411748335e-07, "loss": 3.1657, "step": 25964 }, { "epoch": 0.914929041465155, "grad_norm": 1.171226143836975, "learning_rate": 3.868030309905957e-07, "loss": 0.7389, "step": 25965 }, { "epoch": 0.9149642784781135, "grad_norm": 1.500196099281311, "learning_rate": 3.864887460194722e-07, "loss": 0.7525, "step": 25966 }, { "epoch": 0.9149995154910718, "grad_norm": 3.1158268451690674, "learning_rate": 3.861745862655586e-07, "loss": 2.8619, "step": 25967 }, { "epoch": 0.9150347525040302, "grad_norm": 1.0549246072769165, "learning_rate": 3.858605517329461e-07, "loss": 1.2332, "step": 25968 }, { "epoch": 0.9150699895169886, "grad_norm": 1.0061264038085938, "learning_rate": 3.855466424257226e-07, "loss": 1.119, "step": 25969 }, { "epoch": 0.9151052265299471, "grad_norm": 3.8785581588745117, "learning_rate": 3.852328583479792e-07, "loss": 2.3962, "step": 25970 }, { "epoch": 0.9151404635429055, "grad_norm": 3.7785632610321045, "learning_rate": 3.8491919950380486e-07, "loss": 2.588, "step": 25971 }, { "epoch": 0.9151757005558638, "grad_norm": 3.715355634689331, "learning_rate": 3.8460566589728074e-07, "loss": 2.776, "step": 25972 }, { "epoch": 0.9152109375688223, "grad_norm": 1.1162877082824707, "learning_rate": 3.8429225753249143e-07, "loss": 0.655, "step": 25973 }, { "epoch": 0.9152461745817807, "grad_norm": 4.048282623291016, "learning_rate": 3.839789744135225e-07, "loss": 3.5044, "step": 25974 }, { "epoch": 0.9152814115947391, "grad_norm": 1.1323214769363403, "learning_rate": 3.8366581654445177e-07, "loss": 1.0783, "step": 25975 }, { "epoch": 0.9153166486076976, "grad_norm": 1.154793381690979, "learning_rate": 3.833527839293583e-07, "loss": 0.7995, "step": 25976 }, { "epoch": 0.915351885620656, "grad_norm": 4.0554609298706055, "learning_rate": 3.830398765723198e-07, "loss": 3.6047, "step": 25977 }, { "epoch": 0.9153871226336143, "grad_norm": 2.9824061393737793, "learning_rate": 3.8272709447741194e-07, "loss": 2.8007, "step": 25978 }, { "epoch": 0.9154223596465728, "grad_norm": 1.1689808368682861, "learning_rate": 3.8241443764870934e-07, "loss": 0.7797, "step": 25979 }, { "epoch": 0.9154575966595312, "grad_norm": 0.8366252183914185, "learning_rate": 3.8210190609028306e-07, "loss": 0.9656, "step": 25980 }, { "epoch": 0.9154928336724896, "grad_norm": 1.3235435485839844, "learning_rate": 3.817894998062044e-07, "loss": 0.7952, "step": 25981 }, { "epoch": 0.915528070685448, "grad_norm": 4.286636829376221, "learning_rate": 3.8147721880054335e-07, "loss": 3.4594, "step": 25982 }, { "epoch": 0.9155633076984064, "grad_norm": 0.8691094517707825, "learning_rate": 3.8116506307736556e-07, "loss": 0.8623, "step": 25983 }, { "epoch": 0.9155985447113648, "grad_norm": 5.692216873168945, "learning_rate": 3.8085303264073893e-07, "loss": 5.3331, "step": 25984 }, { "epoch": 0.9156337817243232, "grad_norm": 1.0100736618041992, "learning_rate": 3.805411274947268e-07, "loss": 1.1383, "step": 25985 }, { "epoch": 0.9156690187372817, "grad_norm": 3.148256778717041, "learning_rate": 3.802293476433927e-07, "loss": 2.7719, "step": 25986 }, { "epoch": 0.9157042557502401, "grad_norm": 1.0615675449371338, "learning_rate": 3.799176930907966e-07, "loss": 0.827, "step": 25987 }, { "epoch": 0.9157394927631984, "grad_norm": 6.142876148223877, "learning_rate": 3.7960616384099535e-07, "loss": 2.9762, "step": 25988 }, { "epoch": 0.9157747297761569, "grad_norm": 1.9930838346481323, "learning_rate": 3.792947598980512e-07, "loss": 0.8946, "step": 25989 }, { "epoch": 0.9158099667891153, "grad_norm": 5.19827938079834, "learning_rate": 3.789834812660198e-07, "loss": 2.9845, "step": 25990 }, { "epoch": 0.9158452038020737, "grad_norm": 3.936077833175659, "learning_rate": 3.7867232794895236e-07, "loss": 3.234, "step": 25991 }, { "epoch": 0.9158804408150321, "grad_norm": 4.708827018737793, "learning_rate": 3.7836129995090233e-07, "loss": 4.5854, "step": 25992 }, { "epoch": 0.9159156778279905, "grad_norm": 3.579623222351074, "learning_rate": 3.780503972759242e-07, "loss": 2.6349, "step": 25993 }, { "epoch": 0.9159509148409489, "grad_norm": 3.1746366024017334, "learning_rate": 3.777396199280647e-07, "loss": 2.8212, "step": 25994 }, { "epoch": 0.9159861518539073, "grad_norm": 1.275364637374878, "learning_rate": 3.7742896791137185e-07, "loss": 0.7229, "step": 25995 }, { "epoch": 0.9160213888668658, "grad_norm": 2.17440128326416, "learning_rate": 3.771184412298923e-07, "loss": 0.8725, "step": 25996 }, { "epoch": 0.9160566258798242, "grad_norm": 3.8581223487854004, "learning_rate": 3.7680803988767055e-07, "loss": 2.5997, "step": 25997 }, { "epoch": 0.9160918628927825, "grad_norm": 3.833618640899658, "learning_rate": 3.7649776388875017e-07, "loss": 3.2866, "step": 25998 }, { "epoch": 0.916127099905741, "grad_norm": 1.0978261232376099, "learning_rate": 3.7618761323717225e-07, "loss": 0.8614, "step": 25999 }, { "epoch": 0.9161623369186994, "grad_norm": 4.164463520050049, "learning_rate": 3.758775879369769e-07, "loss": 4.2011, "step": 26000 }, { "epoch": 0.9161975739316578, "grad_norm": 2.389814853668213, "learning_rate": 3.75567687992201e-07, "loss": 0.7435, "step": 26001 }, { "epoch": 0.9162328109446162, "grad_norm": 1.7851792573928833, "learning_rate": 3.7525791340688343e-07, "loss": 0.7159, "step": 26002 }, { "epoch": 0.9162680479575747, "grad_norm": 6.147663116455078, "learning_rate": 3.749482641850566e-07, "loss": 5.4505, "step": 26003 }, { "epoch": 0.916303284970533, "grad_norm": 1.7154871225357056, "learning_rate": 3.746387403307539e-07, "loss": 0.6852, "step": 26004 }, { "epoch": 0.9163385219834914, "grad_norm": 6.771342754364014, "learning_rate": 3.743293418480087e-07, "loss": 7.7278, "step": 26005 }, { "epoch": 0.9163737589964499, "grad_norm": 1.0964775085449219, "learning_rate": 3.7402006874085127e-07, "loss": 0.9401, "step": 26006 }, { "epoch": 0.9164089960094083, "grad_norm": 4.952725410461426, "learning_rate": 3.737109210133061e-07, "loss": 4.7318, "step": 26007 }, { "epoch": 0.9164442330223667, "grad_norm": 2.6641499996185303, "learning_rate": 3.7340189866940325e-07, "loss": 2.5243, "step": 26008 }, { "epoch": 0.9164794700353251, "grad_norm": 2.8332226276397705, "learning_rate": 3.730930017131673e-07, "loss": 2.066, "step": 26009 }, { "epoch": 0.9165147070482835, "grad_norm": 5.616682052612305, "learning_rate": 3.7278423014862175e-07, "loss": 3.4567, "step": 26010 }, { "epoch": 0.9165499440612419, "grad_norm": 4.24273681640625, "learning_rate": 3.7247558397978554e-07, "loss": 2.9575, "step": 26011 }, { "epoch": 0.9165851810742004, "grad_norm": 5.328073978424072, "learning_rate": 3.7216706321068443e-07, "loss": 5.1117, "step": 26012 }, { "epoch": 0.9166204180871588, "grad_norm": 4.922143459320068, "learning_rate": 3.7185866784533177e-07, "loss": 4.8497, "step": 26013 }, { "epoch": 0.9166556551001172, "grad_norm": 3.7706029415130615, "learning_rate": 3.715503978877466e-07, "loss": 3.0346, "step": 26014 }, { "epoch": 0.9166908921130755, "grad_norm": 1.1969269514083862, "learning_rate": 3.7124225334194244e-07, "loss": 0.6899, "step": 26015 }, { "epoch": 0.916726129126034, "grad_norm": 1.6137059926986694, "learning_rate": 3.709342342119382e-07, "loss": 0.8744, "step": 26016 }, { "epoch": 0.9167613661389924, "grad_norm": 4.48737907409668, "learning_rate": 3.7062634050173964e-07, "loss": 3.0828, "step": 26017 }, { "epoch": 0.9167966031519508, "grad_norm": 3.674922227859497, "learning_rate": 3.7031857221536014e-07, "loss": 2.709, "step": 26018 }, { "epoch": 0.9168318401649093, "grad_norm": 5.109554290771484, "learning_rate": 3.700109293568077e-07, "loss": 5.4316, "step": 26019 }, { "epoch": 0.9168670771778676, "grad_norm": 3.5982933044433594, "learning_rate": 3.6970341193008905e-07, "loss": 2.6039, "step": 26020 }, { "epoch": 0.916902314190826, "grad_norm": 1.5440337657928467, "learning_rate": 3.69396019939211e-07, "loss": 0.8032, "step": 26021 }, { "epoch": 0.9169375512037845, "grad_norm": 4.219737529754639, "learning_rate": 3.6908875338817594e-07, "loss": 3.1113, "step": 26022 }, { "epoch": 0.9169727882167429, "grad_norm": 1.22193443775177, "learning_rate": 3.687816122809862e-07, "loss": 0.8337, "step": 26023 }, { "epoch": 0.9170080252297013, "grad_norm": 5.016330242156982, "learning_rate": 3.684745966216441e-07, "loss": 4.778, "step": 26024 }, { "epoch": 0.9170432622426596, "grad_norm": 1.259672999382019, "learning_rate": 3.681677064141476e-07, "loss": 0.7249, "step": 26025 }, { "epoch": 0.9170784992556181, "grad_norm": 2.778350830078125, "learning_rate": 3.678609416624912e-07, "loss": 3.0654, "step": 26026 }, { "epoch": 0.9171137362685765, "grad_norm": 5.049453258514404, "learning_rate": 3.67554302370674e-07, "loss": 5.105, "step": 26027 }, { "epoch": 0.9171489732815349, "grad_norm": 2.7831151485443115, "learning_rate": 3.672477885426906e-07, "loss": 2.9112, "step": 26028 }, { "epoch": 0.9171842102944934, "grad_norm": 2.7566869258880615, "learning_rate": 3.669414001825311e-07, "loss": 2.588, "step": 26029 }, { "epoch": 0.9172194473074518, "grad_norm": 3.043030023574829, "learning_rate": 3.6663513729418454e-07, "loss": 2.5475, "step": 26030 }, { "epoch": 0.9172546843204101, "grad_norm": 0.7852208614349365, "learning_rate": 3.663289998816455e-07, "loss": 0.8888, "step": 26031 }, { "epoch": 0.9172899213333686, "grad_norm": 1.0255000591278076, "learning_rate": 3.660229879488975e-07, "loss": 0.9648, "step": 26032 }, { "epoch": 0.917325158346327, "grad_norm": 4.352581977844238, "learning_rate": 3.6571710149992725e-07, "loss": 2.8406, "step": 26033 }, { "epoch": 0.9173603953592854, "grad_norm": 3.057621955871582, "learning_rate": 3.6541134053871717e-07, "loss": 2.7639, "step": 26034 }, { "epoch": 0.9173956323722438, "grad_norm": 3.4845685958862305, "learning_rate": 3.6510570506925416e-07, "loss": 2.8009, "step": 26035 }, { "epoch": 0.9174308693852022, "grad_norm": 1.1209567785263062, "learning_rate": 3.648001950955149e-07, "loss": 0.866, "step": 26036 }, { "epoch": 0.9174661063981606, "grad_norm": 5.371201515197754, "learning_rate": 3.644948106214818e-07, "loss": 2.5442, "step": 26037 }, { "epoch": 0.917501343411119, "grad_norm": 1.1791657209396362, "learning_rate": 3.6418955165112957e-07, "loss": 0.8766, "step": 26038 }, { "epoch": 0.9175365804240775, "grad_norm": 1.3619801998138428, "learning_rate": 3.6388441818843713e-07, "loss": 0.8606, "step": 26039 }, { "epoch": 0.9175718174370359, "grad_norm": 8.459956169128418, "learning_rate": 3.6357941023737686e-07, "loss": 5.1681, "step": 26040 }, { "epoch": 0.9176070544499942, "grad_norm": 3.9653899669647217, "learning_rate": 3.632745278019223e-07, "loss": 4.1416, "step": 26041 }, { "epoch": 0.9176422914629527, "grad_norm": 2.3458094596862793, "learning_rate": 3.629697708860447e-07, "loss": 1.0255, "step": 26042 }, { "epoch": 0.9176775284759111, "grad_norm": 1.1520724296569824, "learning_rate": 3.6266513949371417e-07, "loss": 0.9307, "step": 26043 }, { "epoch": 0.9177127654888695, "grad_norm": 1.639656901359558, "learning_rate": 3.623606336288976e-07, "loss": 0.8642, "step": 26044 }, { "epoch": 0.917748002501828, "grad_norm": 1.7375826835632324, "learning_rate": 3.620562532955596e-07, "loss": 0.8327, "step": 26045 }, { "epoch": 0.9177832395147864, "grad_norm": 5.8622002601623535, "learning_rate": 3.617519984976681e-07, "loss": 4.665, "step": 26046 }, { "epoch": 0.9178184765277447, "grad_norm": 3.8807222843170166, "learning_rate": 3.614478692391854e-07, "loss": 3.0143, "step": 26047 }, { "epoch": 0.9178537135407031, "grad_norm": 1.4379481077194214, "learning_rate": 3.611438655240707e-07, "loss": 0.686, "step": 26048 }, { "epoch": 0.9178889505536616, "grad_norm": 4.306053161621094, "learning_rate": 3.6083998735628403e-07, "loss": 2.9883, "step": 26049 }, { "epoch": 0.91792418756662, "grad_norm": 4.327810764312744, "learning_rate": 3.6053623473978674e-07, "loss": 3.1394, "step": 26050 }, { "epoch": 0.9179594245795784, "grad_norm": 0.9534815549850464, "learning_rate": 3.6023260767853117e-07, "loss": 0.8757, "step": 26051 }, { "epoch": 0.9179946615925368, "grad_norm": 1.2834234237670898, "learning_rate": 3.5992910617647424e-07, "loss": 0.8175, "step": 26052 }, { "epoch": 0.9180298986054952, "grad_norm": 4.15670108795166, "learning_rate": 3.596257302375683e-07, "loss": 3.1739, "step": 26053 }, { "epoch": 0.9180651356184536, "grad_norm": 3.2638473510742188, "learning_rate": 3.593224798657657e-07, "loss": 2.9392, "step": 26054 }, { "epoch": 0.9181003726314121, "grad_norm": 4.629157066345215, "learning_rate": 3.5901935506501564e-07, "loss": 2.4671, "step": 26055 }, { "epoch": 0.9181356096443705, "grad_norm": 2.890233278274536, "learning_rate": 3.58716355839267e-07, "loss": 2.9906, "step": 26056 }, { "epoch": 0.9181708466573288, "grad_norm": 1.423717975616455, "learning_rate": 3.5841348219246565e-07, "loss": 0.9438, "step": 26057 }, { "epoch": 0.9182060836702872, "grad_norm": 3.5291941165924072, "learning_rate": 3.5811073412855613e-07, "loss": 2.9303, "step": 26058 }, { "epoch": 0.9182413206832457, "grad_norm": 4.548358917236328, "learning_rate": 3.578081116514831e-07, "loss": 2.6559, "step": 26059 }, { "epoch": 0.9182765576962041, "grad_norm": 6.0586748123168945, "learning_rate": 3.5750561476518786e-07, "loss": 4.4884, "step": 26060 }, { "epoch": 0.9183117947091625, "grad_norm": 1.0768404006958008, "learning_rate": 3.572032434736095e-07, "loss": 0.8992, "step": 26061 }, { "epoch": 0.918347031722121, "grad_norm": 5.200298309326172, "learning_rate": 3.569009977806881e-07, "loss": 4.536, "step": 26062 }, { "epoch": 0.9183822687350793, "grad_norm": 0.957955539226532, "learning_rate": 3.565988776903606e-07, "loss": 0.8031, "step": 26063 }, { "epoch": 0.9184175057480377, "grad_norm": 1.3683363199234009, "learning_rate": 3.562968832065583e-07, "loss": 0.82, "step": 26064 }, { "epoch": 0.9184527427609962, "grad_norm": 1.244576334953308, "learning_rate": 3.5599501433321803e-07, "loss": 0.8198, "step": 26065 }, { "epoch": 0.9184879797739546, "grad_norm": 2.945430278778076, "learning_rate": 3.5569327107427334e-07, "loss": 2.6528, "step": 26066 }, { "epoch": 0.918523216786913, "grad_norm": 3.84439754486084, "learning_rate": 3.5539165343365103e-07, "loss": 2.7778, "step": 26067 }, { "epoch": 0.9185584537998713, "grad_norm": 4.432877540588379, "learning_rate": 3.5509016141527907e-07, "loss": 2.741, "step": 26068 }, { "epoch": 0.9185936908128298, "grad_norm": 5.378358840942383, "learning_rate": 3.5478879502308995e-07, "loss": 5.2083, "step": 26069 }, { "epoch": 0.9186289278257882, "grad_norm": 5.097702503204346, "learning_rate": 3.544875542610027e-07, "loss": 3.168, "step": 26070 }, { "epoch": 0.9186641648387466, "grad_norm": 3.5229902267456055, "learning_rate": 3.541864391329442e-07, "loss": 3.0713, "step": 26071 }, { "epoch": 0.9186994018517051, "grad_norm": 2.9694764614105225, "learning_rate": 3.5388544964283565e-07, "loss": 2.4919, "step": 26072 }, { "epoch": 0.9187346388646634, "grad_norm": 2.38761043548584, "learning_rate": 3.535845857945996e-07, "loss": 0.7916, "step": 26073 }, { "epoch": 0.9187698758776218, "grad_norm": 1.164986252784729, "learning_rate": 3.532838475921518e-07, "loss": 1.161, "step": 26074 }, { "epoch": 0.9188051128905803, "grad_norm": 3.8212497234344482, "learning_rate": 3.5298323503941023e-07, "loss": 2.6432, "step": 26075 }, { "epoch": 0.9188403499035387, "grad_norm": 1.1600667238235474, "learning_rate": 3.526827481402917e-07, "loss": 0.9252, "step": 26076 }, { "epoch": 0.9188755869164971, "grad_norm": 1.1272087097167969, "learning_rate": 3.5238238689870864e-07, "loss": 0.9221, "step": 26077 }, { "epoch": 0.9189108239294554, "grad_norm": 3.0932834148406982, "learning_rate": 3.520821513185746e-07, "loss": 3.1067, "step": 26078 }, { "epoch": 0.9189460609424139, "grad_norm": 1.0790421962738037, "learning_rate": 3.517820414037987e-07, "loss": 0.8922, "step": 26079 }, { "epoch": 0.9189812979553723, "grad_norm": 0.982265055179596, "learning_rate": 3.514820571582911e-07, "loss": 0.8156, "step": 26080 }, { "epoch": 0.9190165349683307, "grad_norm": 0.9628153443336487, "learning_rate": 3.511821985859598e-07, "loss": 0.8206, "step": 26081 }, { "epoch": 0.9190517719812892, "grad_norm": 5.6315484046936035, "learning_rate": 3.508824656907095e-07, "loss": 3.3271, "step": 26082 }, { "epoch": 0.9190870089942476, "grad_norm": 4.1994709968566895, "learning_rate": 3.5058285847644147e-07, "loss": 4.9117, "step": 26083 }, { "epoch": 0.9191222460072059, "grad_norm": 3.31901478767395, "learning_rate": 3.502833769470626e-07, "loss": 2.4824, "step": 26084 }, { "epoch": 0.9191574830201644, "grad_norm": 1.238573431968689, "learning_rate": 3.499840211064731e-07, "loss": 0.7453, "step": 26085 }, { "epoch": 0.9191927200331228, "grad_norm": 2.765974521636963, "learning_rate": 3.4968479095856877e-07, "loss": 2.5283, "step": 26086 }, { "epoch": 0.9192279570460812, "grad_norm": 1.044293761253357, "learning_rate": 3.493856865072509e-07, "loss": 0.9166, "step": 26087 }, { "epoch": 0.9192631940590397, "grad_norm": 3.4938275814056396, "learning_rate": 3.4908670775641083e-07, "loss": 2.81, "step": 26088 }, { "epoch": 0.919298431071998, "grad_norm": 0.9086984992027283, "learning_rate": 3.487878547099499e-07, "loss": 1.1545, "step": 26089 }, { "epoch": 0.9193336680849564, "grad_norm": 5.065004348754883, "learning_rate": 3.484891273717539e-07, "loss": 4.5804, "step": 26090 }, { "epoch": 0.9193689050979148, "grad_norm": 0.9807552099227905, "learning_rate": 3.481905257457174e-07, "loss": 0.724, "step": 26091 }, { "epoch": 0.9194041421108733, "grad_norm": 1.53536856174469, "learning_rate": 3.4789204983572855e-07, "loss": 0.7027, "step": 26092 }, { "epoch": 0.9194393791238317, "grad_norm": 1.0914276838302612, "learning_rate": 3.4759369964567524e-07, "loss": 1.1773, "step": 26093 }, { "epoch": 0.91947461613679, "grad_norm": 3.1808407306671143, "learning_rate": 3.472954751794433e-07, "loss": 2.9645, "step": 26094 }, { "epoch": 0.9195098531497485, "grad_norm": 4.4788126945495605, "learning_rate": 3.4699737644091737e-07, "loss": 3.2819, "step": 26095 }, { "epoch": 0.9195450901627069, "grad_norm": 6.907553195953369, "learning_rate": 3.466994034339799e-07, "loss": 3.4194, "step": 26096 }, { "epoch": 0.9195803271756653, "grad_norm": 3.5727055072784424, "learning_rate": 3.4640155616251223e-07, "loss": 2.5865, "step": 26097 }, { "epoch": 0.9196155641886238, "grad_norm": 0.9879930019378662, "learning_rate": 3.4610383463039463e-07, "loss": 0.6927, "step": 26098 }, { "epoch": 0.9196508012015822, "grad_norm": 3.0213613510131836, "learning_rate": 3.45806238841504e-07, "loss": 0.9292, "step": 26099 }, { "epoch": 0.9196860382145405, "grad_norm": 1.1183533668518066, "learning_rate": 3.4550876879971716e-07, "loss": 0.8203, "step": 26100 }, { "epoch": 0.9197212752274989, "grad_norm": 5.090562343597412, "learning_rate": 3.4521142450890887e-07, "loss": 2.7186, "step": 26101 }, { "epoch": 0.9197565122404574, "grad_norm": 6.360044479370117, "learning_rate": 3.4491420597295153e-07, "loss": 7.173, "step": 26102 }, { "epoch": 0.9197917492534158, "grad_norm": 1.029038906097412, "learning_rate": 3.4461711319571436e-07, "loss": 0.8248, "step": 26103 }, { "epoch": 0.9198269862663742, "grad_norm": 4.5499467849731445, "learning_rate": 3.443201461810719e-07, "loss": 4.6326, "step": 26104 }, { "epoch": 0.9198622232793326, "grad_norm": 4.886411190032959, "learning_rate": 3.4402330493288784e-07, "loss": 6.6964, "step": 26105 }, { "epoch": 0.919897460292291, "grad_norm": 3.3342325687408447, "learning_rate": 3.437265894550312e-07, "loss": 2.8063, "step": 26106 }, { "epoch": 0.9199326973052494, "grad_norm": 7.989109516143799, "learning_rate": 3.434299997513635e-07, "loss": 4.0326, "step": 26107 }, { "epoch": 0.9199679343182079, "grad_norm": 3.6482579708099365, "learning_rate": 3.431335358257526e-07, "loss": 2.7042, "step": 26108 }, { "epoch": 0.9200031713311663, "grad_norm": 3.094656467437744, "learning_rate": 3.4283719768205657e-07, "loss": 3.68, "step": 26109 }, { "epoch": 0.9200384083441246, "grad_norm": 5.1187357902526855, "learning_rate": 3.425409853241368e-07, "loss": 5.1461, "step": 26110 }, { "epoch": 0.920073645357083, "grad_norm": 2.009748697280884, "learning_rate": 3.4224489875585023e-07, "loss": 0.8741, "step": 26111 }, { "epoch": 0.9201088823700415, "grad_norm": 10.55340576171875, "learning_rate": 3.4194893798105476e-07, "loss": 4.8397, "step": 26112 }, { "epoch": 0.9201441193829999, "grad_norm": 1.0789774656295776, "learning_rate": 3.41653103003603e-07, "loss": 0.7892, "step": 26113 }, { "epoch": 0.9201793563959583, "grad_norm": 1.5053144693374634, "learning_rate": 3.4135739382735177e-07, "loss": 0.7922, "step": 26114 }, { "epoch": 0.9202145934089168, "grad_norm": 5.655941009521484, "learning_rate": 3.4106181045614917e-07, "loss": 3.0984, "step": 26115 }, { "epoch": 0.9202498304218751, "grad_norm": 2.628824472427368, "learning_rate": 3.407663528938476e-07, "loss": 2.8018, "step": 26116 }, { "epoch": 0.9202850674348335, "grad_norm": 0.8127509355545044, "learning_rate": 3.404710211442963e-07, "loss": 0.6958, "step": 26117 }, { "epoch": 0.920320304447792, "grad_norm": 2.3182621002197266, "learning_rate": 3.4017581521133766e-07, "loss": 0.7685, "step": 26118 }, { "epoch": 0.9203555414607504, "grad_norm": 2.889305830001831, "learning_rate": 3.398807350988198e-07, "loss": 2.3612, "step": 26119 }, { "epoch": 0.9203907784737088, "grad_norm": 0.8539546132087708, "learning_rate": 3.395857808105885e-07, "loss": 0.611, "step": 26120 }, { "epoch": 0.9204260154866672, "grad_norm": 1.155068278312683, "learning_rate": 3.392909523504806e-07, "loss": 0.8057, "step": 26121 }, { "epoch": 0.9204612524996256, "grad_norm": 0.9870571494102478, "learning_rate": 3.389962497223376e-07, "loss": 0.8273, "step": 26122 }, { "epoch": 0.920496489512584, "grad_norm": 1.430577278137207, "learning_rate": 3.387016729300019e-07, "loss": 0.8057, "step": 26123 }, { "epoch": 0.9205317265255424, "grad_norm": 1.8806508779525757, "learning_rate": 3.3840722197730603e-07, "loss": 2.7626, "step": 26124 }, { "epoch": 0.9205669635385009, "grad_norm": 1.2327497005462646, "learning_rate": 3.3811289686808693e-07, "loss": 0.7315, "step": 26125 }, { "epoch": 0.9206022005514592, "grad_norm": 4.198336124420166, "learning_rate": 3.3781869760617704e-07, "loss": 4.2561, "step": 26126 }, { "epoch": 0.9206374375644176, "grad_norm": 5.18773078918457, "learning_rate": 3.375246241954122e-07, "loss": 5.1234, "step": 26127 }, { "epoch": 0.9206726745773761, "grad_norm": 1.0522587299346924, "learning_rate": 3.3723067663961716e-07, "loss": 1.0523, "step": 26128 }, { "epoch": 0.9207079115903345, "grad_norm": 3.735933303833008, "learning_rate": 3.369368549426244e-07, "loss": 1.0332, "step": 26129 }, { "epoch": 0.9207431486032929, "grad_norm": 1.044163703918457, "learning_rate": 3.366431591082608e-07, "loss": 0.7054, "step": 26130 }, { "epoch": 0.9207783856162514, "grad_norm": 0.9828636646270752, "learning_rate": 3.363495891403501e-07, "loss": 1.0321, "step": 26131 }, { "epoch": 0.9208136226292097, "grad_norm": 7.424588203430176, "learning_rate": 3.360561450427169e-07, "loss": 2.7766, "step": 26132 }, { "epoch": 0.9208488596421681, "grad_norm": 2.568354845046997, "learning_rate": 3.357628268191837e-07, "loss": 2.4132, "step": 26133 }, { "epoch": 0.9208840966551265, "grad_norm": 1.2930659055709839, "learning_rate": 3.3546963447356974e-07, "loss": 0.8361, "step": 26134 }, { "epoch": 0.920919333668085, "grad_norm": 1.1147117614746094, "learning_rate": 3.351765680096963e-07, "loss": 0.7266, "step": 26135 }, { "epoch": 0.9209545706810434, "grad_norm": 9.324020385742188, "learning_rate": 3.3488362743137934e-07, "loss": 6.5563, "step": 26136 }, { "epoch": 0.9209898076940017, "grad_norm": 2.9186599254608154, "learning_rate": 3.3459081274243244e-07, "loss": 2.7008, "step": 26137 }, { "epoch": 0.9210250447069602, "grad_norm": 5.042137145996094, "learning_rate": 3.342981239466725e-07, "loss": 3.1141, "step": 26138 }, { "epoch": 0.9210602817199186, "grad_norm": 4.2399702072143555, "learning_rate": 3.3400556104791205e-07, "loss": 5.3895, "step": 26139 }, { "epoch": 0.921095518732877, "grad_norm": 1.3101308345794678, "learning_rate": 3.3371312404995805e-07, "loss": 0.7629, "step": 26140 }, { "epoch": 0.9211307557458355, "grad_norm": 2.669736623764038, "learning_rate": 3.3342081295662187e-07, "loss": 0.7516, "step": 26141 }, { "epoch": 0.9211659927587939, "grad_norm": 1.4683761596679688, "learning_rate": 3.331286277717116e-07, "loss": 0.8095, "step": 26142 }, { "epoch": 0.9212012297717522, "grad_norm": 1.0290336608886719, "learning_rate": 3.328365684990342e-07, "loss": 1.2776, "step": 26143 }, { "epoch": 0.9212364667847106, "grad_norm": 5.178613662719727, "learning_rate": 3.3254463514239e-07, "loss": 2.984, "step": 26144 }, { "epoch": 0.9212717037976691, "grad_norm": 1.3997331857681274, "learning_rate": 3.3225282770558253e-07, "loss": 0.8349, "step": 26145 }, { "epoch": 0.9213069408106275, "grad_norm": 2.6533870697021484, "learning_rate": 3.319611461924155e-07, "loss": 2.9288, "step": 26146 }, { "epoch": 0.9213421778235859, "grad_norm": 0.7893083691596985, "learning_rate": 3.316695906066858e-07, "loss": 0.986, "step": 26147 }, { "epoch": 0.9213774148365443, "grad_norm": 4.234742164611816, "learning_rate": 3.3137816095219044e-07, "loss": 2.82, "step": 26148 }, { "epoch": 0.9214126518495027, "grad_norm": 1.048295021057129, "learning_rate": 3.310868572327264e-07, "loss": 0.8145, "step": 26149 }, { "epoch": 0.9214478888624611, "grad_norm": 4.651338577270508, "learning_rate": 3.3079567945208723e-07, "loss": 4.7222, "step": 26150 }, { "epoch": 0.9214831258754196, "grad_norm": 1.3393077850341797, "learning_rate": 3.3050462761406665e-07, "loss": 0.7202, "step": 26151 }, { "epoch": 0.921518362888378, "grad_norm": 7.779858589172363, "learning_rate": 3.302137017224538e-07, "loss": 2.8206, "step": 26152 }, { "epoch": 0.9215535999013363, "grad_norm": 16.118478775024414, "learning_rate": 3.2992290178104015e-07, "loss": 4.0908, "step": 26153 }, { "epoch": 0.9215888369142948, "grad_norm": 1.0077745914459229, "learning_rate": 3.296322277936126e-07, "loss": 0.8349, "step": 26154 }, { "epoch": 0.9216240739272532, "grad_norm": 1.2400660514831543, "learning_rate": 3.2934167976395814e-07, "loss": 0.8249, "step": 26155 }, { "epoch": 0.9216593109402116, "grad_norm": 1.147977352142334, "learning_rate": 3.290512576958571e-07, "loss": 0.8988, "step": 26156 }, { "epoch": 0.92169454795317, "grad_norm": 4.87996244430542, "learning_rate": 3.287609615930976e-07, "loss": 2.773, "step": 26157 }, { "epoch": 0.9217297849661285, "grad_norm": 1.9698234796524048, "learning_rate": 3.284707914594587e-07, "loss": 0.8717, "step": 26158 }, { "epoch": 0.9217650219790868, "grad_norm": 1.2620662450790405, "learning_rate": 3.2818074729871864e-07, "loss": 0.8447, "step": 26159 }, { "epoch": 0.9218002589920452, "grad_norm": 1.710065484046936, "learning_rate": 3.2789082911465655e-07, "loss": 0.7342, "step": 26160 }, { "epoch": 0.9218354960050037, "grad_norm": 1.1612969636917114, "learning_rate": 3.2760103691104825e-07, "loss": 1.1219, "step": 26161 }, { "epoch": 0.9218707330179621, "grad_norm": 1.653529405593872, "learning_rate": 3.273113706916697e-07, "loss": 0.9351, "step": 26162 }, { "epoch": 0.9219059700309205, "grad_norm": 2.851228952407837, "learning_rate": 3.2702183046029125e-07, "loss": 2.901, "step": 26163 }, { "epoch": 0.9219412070438789, "grad_norm": 0.8164433836936951, "learning_rate": 3.267324162206853e-07, "loss": 0.8522, "step": 26164 }, { "epoch": 0.9219764440568373, "grad_norm": 4.0175275802612305, "learning_rate": 3.2644312797662334e-07, "loss": 2.4591, "step": 26165 }, { "epoch": 0.9220116810697957, "grad_norm": 1.0926538705825806, "learning_rate": 3.261539657318713e-07, "loss": 0.8097, "step": 26166 }, { "epoch": 0.9220469180827541, "grad_norm": 0.9979127645492554, "learning_rate": 3.258649294901961e-07, "loss": 1.1085, "step": 26167 }, { "epoch": 0.9220821550957126, "grad_norm": 4.848522186279297, "learning_rate": 3.2557601925536253e-07, "loss": 2.9623, "step": 26168 }, { "epoch": 0.9221173921086709, "grad_norm": 4.578959941864014, "learning_rate": 3.252872350311331e-07, "loss": 5.5142, "step": 26169 }, { "epoch": 0.9221526291216293, "grad_norm": 1.014585256576538, "learning_rate": 3.249985768212705e-07, "loss": 0.7579, "step": 26170 }, { "epoch": 0.9221878661345878, "grad_norm": 0.9973050951957703, "learning_rate": 3.247100446295337e-07, "loss": 0.914, "step": 26171 }, { "epoch": 0.9222231031475462, "grad_norm": 2.1989777088165283, "learning_rate": 3.244216384596799e-07, "loss": 0.8596, "step": 26172 }, { "epoch": 0.9222583401605046, "grad_norm": 2.670400381088257, "learning_rate": 3.241333583154682e-07, "loss": 2.6086, "step": 26173 }, { "epoch": 0.922293577173463, "grad_norm": 0.9183885455131531, "learning_rate": 3.2384520420065236e-07, "loss": 0.7061, "step": 26174 }, { "epoch": 0.9223288141864214, "grad_norm": 4.665663719177246, "learning_rate": 3.235571761189837e-07, "loss": 5.0245, "step": 26175 }, { "epoch": 0.9223640511993798, "grad_norm": 2.3375115394592285, "learning_rate": 3.2326927407421604e-07, "loss": 0.8587, "step": 26176 }, { "epoch": 0.9223992882123382, "grad_norm": 2.335632085800171, "learning_rate": 3.229814980700996e-07, "loss": 1.3538, "step": 26177 }, { "epoch": 0.9224345252252967, "grad_norm": 1.0299309492111206, "learning_rate": 3.2269384811038027e-07, "loss": 1.0583, "step": 26178 }, { "epoch": 0.922469762238255, "grad_norm": 0.9488042593002319, "learning_rate": 3.2240632419880625e-07, "loss": 0.8548, "step": 26179 }, { "epoch": 0.9225049992512134, "grad_norm": 0.9931190013885498, "learning_rate": 3.2211892633912335e-07, "loss": 0.7667, "step": 26180 }, { "epoch": 0.9225402362641719, "grad_norm": 1.0607683658599854, "learning_rate": 3.2183165453507525e-07, "loss": 0.9558, "step": 26181 }, { "epoch": 0.9225754732771303, "grad_norm": 1.7945643663406372, "learning_rate": 3.215445087904012e-07, "loss": 0.8227, "step": 26182 }, { "epoch": 0.9226107102900887, "grad_norm": 1.726839303970337, "learning_rate": 3.2125748910884157e-07, "loss": 0.7671, "step": 26183 }, { "epoch": 0.9226459473030472, "grad_norm": 9.640542030334473, "learning_rate": 3.209705954941389e-07, "loss": 4.8428, "step": 26184 }, { "epoch": 0.9226811843160055, "grad_norm": 2.5309083461761475, "learning_rate": 3.2068382795002574e-07, "loss": 2.6601, "step": 26185 }, { "epoch": 0.9227164213289639, "grad_norm": 3.535461187362671, "learning_rate": 3.203971864802391e-07, "loss": 2.5297, "step": 26186 }, { "epoch": 0.9227516583419224, "grad_norm": 5.510260581970215, "learning_rate": 3.201106710885116e-07, "loss": 5.0935, "step": 26187 }, { "epoch": 0.9227868953548808, "grad_norm": 4.596352577209473, "learning_rate": 3.198242817785757e-07, "loss": 2.8441, "step": 26188 }, { "epoch": 0.9228221323678392, "grad_norm": 3.0075244903564453, "learning_rate": 3.195380185541608e-07, "loss": 0.7478, "step": 26189 }, { "epoch": 0.9228573693807975, "grad_norm": 4.250815391540527, "learning_rate": 3.1925188141899713e-07, "loss": 2.7814, "step": 26190 }, { "epoch": 0.922892606393756, "grad_norm": 1.467340350151062, "learning_rate": 3.189658703768106e-07, "loss": 0.8457, "step": 26191 }, { "epoch": 0.9229278434067144, "grad_norm": 0.9891956448554993, "learning_rate": 3.1867998543132717e-07, "loss": 0.9751, "step": 26192 }, { "epoch": 0.9229630804196728, "grad_norm": 2.1582601070404053, "learning_rate": 3.183942265862716e-07, "loss": 0.7213, "step": 26193 }, { "epoch": 0.9229983174326313, "grad_norm": 1.1769731044769287, "learning_rate": 3.1810859384536097e-07, "loss": 0.835, "step": 26194 }, { "epoch": 0.9230335544455897, "grad_norm": 2.366109848022461, "learning_rate": 3.1782308721232114e-07, "loss": 2.7915, "step": 26195 }, { "epoch": 0.923068791458548, "grad_norm": 1.215868592262268, "learning_rate": 3.1753770669087026e-07, "loss": 0.9923, "step": 26196 }, { "epoch": 0.9231040284715065, "grad_norm": 4.128860950469971, "learning_rate": 3.1725245228472204e-07, "loss": 2.7962, "step": 26197 }, { "epoch": 0.9231392654844649, "grad_norm": 3.72485613822937, "learning_rate": 3.1696732399759347e-07, "loss": 4.7326, "step": 26198 }, { "epoch": 0.9231745024974233, "grad_norm": 1.2469944953918457, "learning_rate": 3.166823218332005e-07, "loss": 0.8077, "step": 26199 }, { "epoch": 0.9232097395103817, "grad_norm": 3.8584797382354736, "learning_rate": 3.1639744579525456e-07, "loss": 2.5343, "step": 26200 }, { "epoch": 0.9232449765233401, "grad_norm": 1.0706464052200317, "learning_rate": 3.1611269588746387e-07, "loss": 0.8834, "step": 26201 }, { "epoch": 0.9232802135362985, "grad_norm": 2.8317220211029053, "learning_rate": 3.158280721135376e-07, "loss": 2.106, "step": 26202 }, { "epoch": 0.9233154505492569, "grad_norm": 1.1342055797576904, "learning_rate": 3.155435744771873e-07, "loss": 0.7329, "step": 26203 }, { "epoch": 0.9233506875622154, "grad_norm": 3.3303442001342773, "learning_rate": 3.152592029821133e-07, "loss": 2.8991, "step": 26204 }, { "epoch": 0.9233859245751738, "grad_norm": 3.864342212677002, "learning_rate": 3.149749576320227e-07, "loss": 3.8426, "step": 26205 }, { "epoch": 0.9234211615881321, "grad_norm": 1.1506624221801758, "learning_rate": 3.146908384306169e-07, "loss": 0.8125, "step": 26206 }, { "epoch": 0.9234563986010906, "grad_norm": 1.2110676765441895, "learning_rate": 3.144068453815963e-07, "loss": 0.9769, "step": 26207 }, { "epoch": 0.923491635614049, "grad_norm": 1.1180119514465332, "learning_rate": 3.1412297848866015e-07, "loss": 0.8593, "step": 26208 }, { "epoch": 0.9235268726270074, "grad_norm": 1.616483449935913, "learning_rate": 3.1383923775550664e-07, "loss": 0.759, "step": 26209 }, { "epoch": 0.9235621096399658, "grad_norm": 5.665088653564453, "learning_rate": 3.1355562318583057e-07, "loss": 4.9185, "step": 26210 }, { "epoch": 0.9235973466529243, "grad_norm": 1.287940502166748, "learning_rate": 3.1327213478332676e-07, "loss": 1.0499, "step": 26211 }, { "epoch": 0.9236325836658826, "grad_norm": 8.260188102722168, "learning_rate": 3.129887725516878e-07, "loss": 2.4786, "step": 26212 }, { "epoch": 0.923667820678841, "grad_norm": 2.6814727783203125, "learning_rate": 3.1270553649460187e-07, "loss": 3.1541, "step": 26213 }, { "epoch": 0.9237030576917995, "grad_norm": 5.836013317108154, "learning_rate": 3.124224266157616e-07, "loss": 4.9719, "step": 26214 }, { "epoch": 0.9237382947047579, "grad_norm": 1.176108479499817, "learning_rate": 3.121394429188529e-07, "loss": 1.0717, "step": 26215 }, { "epoch": 0.9237735317177163, "grad_norm": 1.3312517404556274, "learning_rate": 3.118565854075639e-07, "loss": 0.814, "step": 26216 }, { "epoch": 0.9238087687306747, "grad_norm": 1.0733147859573364, "learning_rate": 3.115738540855739e-07, "loss": 0.8361, "step": 26217 }, { "epoch": 0.9238440057436331, "grad_norm": 4.608728408813477, "learning_rate": 3.1129124895657003e-07, "loss": 3.1206, "step": 26218 }, { "epoch": 0.9238792427565915, "grad_norm": 1.2424513101577759, "learning_rate": 3.1100877002423257e-07, "loss": 0.9505, "step": 26219 }, { "epoch": 0.92391447976955, "grad_norm": 0.7988181114196777, "learning_rate": 3.107264172922386e-07, "loss": 0.7426, "step": 26220 }, { "epoch": 0.9239497167825084, "grad_norm": 4.912837028503418, "learning_rate": 3.104441907642675e-07, "loss": 3.7296, "step": 26221 }, { "epoch": 0.9239849537954667, "grad_norm": 0.9662100076675415, "learning_rate": 3.101620904439961e-07, "loss": 0.6517, "step": 26222 }, { "epoch": 0.9240201908084251, "grad_norm": 1.068617820739746, "learning_rate": 3.098801163350973e-07, "loss": 0.7358, "step": 26223 }, { "epoch": 0.9240554278213836, "grad_norm": 4.184442520141602, "learning_rate": 3.0959826844124353e-07, "loss": 2.8446, "step": 26224 }, { "epoch": 0.924090664834342, "grad_norm": 2.886286735534668, "learning_rate": 3.0931654676610746e-07, "loss": 2.7826, "step": 26225 }, { "epoch": 0.9241259018473004, "grad_norm": 5.813742160797119, "learning_rate": 3.0903495131335727e-07, "loss": 5.0027, "step": 26226 }, { "epoch": 0.9241611388602589, "grad_norm": 1.7898935079574585, "learning_rate": 3.087534820866622e-07, "loss": 0.7788, "step": 26227 }, { "epoch": 0.9241963758732172, "grad_norm": 3.295297145843506, "learning_rate": 3.084721390896872e-07, "loss": 0.9403, "step": 26228 }, { "epoch": 0.9242316128861756, "grad_norm": 1.1614993810653687, "learning_rate": 3.0819092232609595e-07, "loss": 1.0971, "step": 26229 }, { "epoch": 0.9242668498991341, "grad_norm": 4.580924034118652, "learning_rate": 3.0790983179955323e-07, "loss": 3.1433, "step": 26230 }, { "epoch": 0.9243020869120925, "grad_norm": 1.0510897636413574, "learning_rate": 3.076288675137207e-07, "loss": 0.9027, "step": 26231 }, { "epoch": 0.9243373239250509, "grad_norm": 1.3721494674682617, "learning_rate": 3.0734802947225417e-07, "loss": 0.8947, "step": 26232 }, { "epoch": 0.9243725609380092, "grad_norm": 1.276213526725769, "learning_rate": 3.0706731767881635e-07, "loss": 0.9008, "step": 26233 }, { "epoch": 0.9244077979509677, "grad_norm": 1.293225884437561, "learning_rate": 3.06786732137061e-07, "loss": 0.8126, "step": 26234 }, { "epoch": 0.9244430349639261, "grad_norm": 0.9947341680526733, "learning_rate": 3.0650627285064405e-07, "loss": 1.0667, "step": 26235 }, { "epoch": 0.9244782719768845, "grad_norm": 1.9414669275283813, "learning_rate": 3.062259398232148e-07, "loss": 0.955, "step": 26236 }, { "epoch": 0.924513508989843, "grad_norm": 1.688930630683899, "learning_rate": 3.059457330584292e-07, "loss": 0.6743, "step": 26237 }, { "epoch": 0.9245487460028013, "grad_norm": 1.1262786388397217, "learning_rate": 3.0566565255993665e-07, "loss": 0.7251, "step": 26238 }, { "epoch": 0.9245839830157597, "grad_norm": 3.1251251697540283, "learning_rate": 3.0538569833138186e-07, "loss": 2.7263, "step": 26239 }, { "epoch": 0.9246192200287182, "grad_norm": 1.0765050649642944, "learning_rate": 3.0510587037641314e-07, "loss": 0.8722, "step": 26240 }, { "epoch": 0.9246544570416766, "grad_norm": 2.805044412612915, "learning_rate": 3.0482616869867643e-07, "loss": 0.8273, "step": 26241 }, { "epoch": 0.924689694054635, "grad_norm": 4.482672691345215, "learning_rate": 3.0454659330181434e-07, "loss": 4.8077, "step": 26242 }, { "epoch": 0.9247249310675933, "grad_norm": 2.0080771446228027, "learning_rate": 3.042671441894662e-07, "loss": 0.9229, "step": 26243 }, { "epoch": 0.9247601680805518, "grad_norm": 4.746602535247803, "learning_rate": 3.039878213652747e-07, "loss": 5.0842, "step": 26244 }, { "epoch": 0.9247954050935102, "grad_norm": 3.9349400997161865, "learning_rate": 3.037086248328769e-07, "loss": 2.4798, "step": 26245 }, { "epoch": 0.9248306421064686, "grad_norm": 4.902492523193359, "learning_rate": 3.0342955459590873e-07, "loss": 2.9793, "step": 26246 }, { "epoch": 0.9248658791194271, "grad_norm": 3.7469871044158936, "learning_rate": 3.031506106580062e-07, "loss": 2.9828, "step": 26247 }, { "epoch": 0.9249011161323855, "grad_norm": 1.019051432609558, "learning_rate": 3.02871793022802e-07, "loss": 0.7752, "step": 26248 }, { "epoch": 0.9249363531453438, "grad_norm": 4.584901332855225, "learning_rate": 3.0259310169392765e-07, "loss": 4.1731, "step": 26249 }, { "epoch": 0.9249715901583023, "grad_norm": 1.297169804573059, "learning_rate": 3.0231453667501464e-07, "loss": 1.0252, "step": 26250 }, { "epoch": 0.9250068271712607, "grad_norm": 1.0614725351333618, "learning_rate": 3.020360979696879e-07, "loss": 0.8348, "step": 26251 }, { "epoch": 0.9250420641842191, "grad_norm": 3.0903232097625732, "learning_rate": 3.017577855815779e-07, "loss": 2.4677, "step": 26252 }, { "epoch": 0.9250773011971776, "grad_norm": 3.7076170444488525, "learning_rate": 3.014795995143072e-07, "loss": 3.041, "step": 26253 }, { "epoch": 0.925112538210136, "grad_norm": 5.473638534545898, "learning_rate": 3.012015397715018e-07, "loss": 4.4437, "step": 26254 }, { "epoch": 0.9251477752230943, "grad_norm": 5.86789083480835, "learning_rate": 3.0092360635678e-07, "loss": 7.589, "step": 26255 }, { "epoch": 0.9251830122360527, "grad_norm": 1.0055259466171265, "learning_rate": 3.006457992737644e-07, "loss": 1.003, "step": 26256 }, { "epoch": 0.9252182492490112, "grad_norm": 4.6071953773498535, "learning_rate": 3.003681185260732e-07, "loss": 4.5961, "step": 26257 }, { "epoch": 0.9252534862619696, "grad_norm": 3.58418345451355, "learning_rate": 3.000905641173224e-07, "loss": 0.7351, "step": 26258 }, { "epoch": 0.925288723274928, "grad_norm": 4.153894901275635, "learning_rate": 2.998131360511258e-07, "loss": 4.6327, "step": 26259 }, { "epoch": 0.9253239602878864, "grad_norm": 1.9248698949813843, "learning_rate": 2.9953583433110165e-07, "loss": 0.8141, "step": 26260 }, { "epoch": 0.9253591973008448, "grad_norm": 1.0403313636779785, "learning_rate": 2.99258658960857e-07, "loss": 1.103, "step": 26261 }, { "epoch": 0.9253944343138032, "grad_norm": 1.5339807271957397, "learning_rate": 2.989816099440046e-07, "loss": 0.7639, "step": 26262 }, { "epoch": 0.9254296713267617, "grad_norm": 1.1115809679031372, "learning_rate": 2.9870468728415256e-07, "loss": 1.0239, "step": 26263 }, { "epoch": 0.9254649083397201, "grad_norm": 3.160299301147461, "learning_rate": 2.984278909849081e-07, "loss": 2.9177, "step": 26264 }, { "epoch": 0.9255001453526784, "grad_norm": 1.401288390159607, "learning_rate": 2.9815122104987493e-07, "loss": 0.7933, "step": 26265 }, { "epoch": 0.9255353823656368, "grad_norm": 0.8451209664344788, "learning_rate": 2.978746774826591e-07, "loss": 0.9561, "step": 26266 }, { "epoch": 0.9255706193785953, "grad_norm": 4.954441547393799, "learning_rate": 2.97598260286861e-07, "loss": 2.9368, "step": 26267 }, { "epoch": 0.9256058563915537, "grad_norm": 4.2935686111450195, "learning_rate": 2.9732196946608116e-07, "loss": 2.6169, "step": 26268 }, { "epoch": 0.9256410934045121, "grad_norm": 3.450201988220215, "learning_rate": 2.9704580502392e-07, "loss": 2.4756, "step": 26269 }, { "epoch": 0.9256763304174705, "grad_norm": 0.9332723617553711, "learning_rate": 2.967697669639713e-07, "loss": 0.8028, "step": 26270 }, { "epoch": 0.9257115674304289, "grad_norm": 9.163374900817871, "learning_rate": 2.964938552898333e-07, "loss": 7.2229, "step": 26271 }, { "epoch": 0.9257468044433873, "grad_norm": 3.5402908325195312, "learning_rate": 2.9621807000509874e-07, "loss": 2.904, "step": 26272 }, { "epoch": 0.9257820414563458, "grad_norm": 1.226948857307434, "learning_rate": 2.959424111133602e-07, "loss": 0.7698, "step": 26273 }, { "epoch": 0.9258172784693042, "grad_norm": 5.004029273986816, "learning_rate": 2.95666878618206e-07, "loss": 2.9527, "step": 26274 }, { "epoch": 0.9258525154822626, "grad_norm": 1.1231372356414795, "learning_rate": 2.9539147252322876e-07, "loss": 0.7575, "step": 26275 }, { "epoch": 0.9258877524952209, "grad_norm": 3.8092846870422363, "learning_rate": 2.9511619283201454e-07, "loss": 2.6849, "step": 26276 }, { "epoch": 0.9259229895081794, "grad_norm": 1.2452771663665771, "learning_rate": 2.948410395481471e-07, "loss": 1.1025, "step": 26277 }, { "epoch": 0.9259582265211378, "grad_norm": 1.237839937210083, "learning_rate": 2.9456601267520925e-07, "loss": 0.7548, "step": 26278 }, { "epoch": 0.9259934635340962, "grad_norm": 6.327073574066162, "learning_rate": 2.9429111221678906e-07, "loss": 6.3398, "step": 26279 }, { "epoch": 0.9260287005470547, "grad_norm": 3.435295820236206, "learning_rate": 2.9401633817646157e-07, "loss": 2.4213, "step": 26280 }, { "epoch": 0.926063937560013, "grad_norm": 3.0601892471313477, "learning_rate": 2.937416905578072e-07, "loss": 2.5939, "step": 26281 }, { "epoch": 0.9260991745729714, "grad_norm": 2.544843912124634, "learning_rate": 2.934671693644042e-07, "loss": 2.6284, "step": 26282 }, { "epoch": 0.9261344115859299, "grad_norm": 1.0487620830535889, "learning_rate": 2.9319277459982754e-07, "loss": 0.5724, "step": 26283 }, { "epoch": 0.9261696485988883, "grad_norm": 3.509131669998169, "learning_rate": 2.929185062676521e-07, "loss": 2.9824, "step": 26284 }, { "epoch": 0.9262048856118467, "grad_norm": 3.2680511474609375, "learning_rate": 2.9264436437144937e-07, "loss": 2.7068, "step": 26285 }, { "epoch": 0.926240122624805, "grad_norm": 1.1293976306915283, "learning_rate": 2.9237034891479e-07, "loss": 0.7642, "step": 26286 }, { "epoch": 0.9262753596377635, "grad_norm": 3.2552685737609863, "learning_rate": 2.9209645990124326e-07, "loss": 2.7045, "step": 26287 }, { "epoch": 0.9263105966507219, "grad_norm": 1.6064848899841309, "learning_rate": 2.918226973343763e-07, "loss": 0.8689, "step": 26288 }, { "epoch": 0.9263458336636803, "grad_norm": 0.9786445498466492, "learning_rate": 2.9154906121775626e-07, "loss": 1.0929, "step": 26289 }, { "epoch": 0.9263810706766388, "grad_norm": 3.304840326309204, "learning_rate": 2.912755515549459e-07, "loss": 2.9503, "step": 26290 }, { "epoch": 0.9264163076895972, "grad_norm": 4.4427618980407715, "learning_rate": 2.91002168349509e-07, "loss": 2.9653, "step": 26291 }, { "epoch": 0.9264515447025555, "grad_norm": 3.39111065864563, "learning_rate": 2.9072891160500496e-07, "loss": 2.5848, "step": 26292 }, { "epoch": 0.926486781715514, "grad_norm": 8.070853233337402, "learning_rate": 2.9045578132499306e-07, "loss": 5.7112, "step": 26293 }, { "epoch": 0.9265220187284724, "grad_norm": 4.996546745300293, "learning_rate": 2.901827775130317e-07, "loss": 3.1099, "step": 26294 }, { "epoch": 0.9265572557414308, "grad_norm": 4.629132270812988, "learning_rate": 2.899099001726779e-07, "loss": 5.0084, "step": 26295 }, { "epoch": 0.9265924927543893, "grad_norm": 4.7969536781311035, "learning_rate": 2.896371493074834e-07, "loss": 4.6607, "step": 26296 }, { "epoch": 0.9266277297673476, "grad_norm": 2.551738977432251, "learning_rate": 2.893645249210009e-07, "loss": 0.6148, "step": 26297 }, { "epoch": 0.926662966780306, "grad_norm": 5.733841419219971, "learning_rate": 2.8909202701678406e-07, "loss": 3.2344, "step": 26298 }, { "epoch": 0.9266982037932644, "grad_norm": 1.7050883769989014, "learning_rate": 2.888196555983802e-07, "loss": 0.7731, "step": 26299 }, { "epoch": 0.9267334408062229, "grad_norm": 1.0115193128585815, "learning_rate": 2.8854741066933645e-07, "loss": 0.9869, "step": 26300 }, { "epoch": 0.9267686778191813, "grad_norm": 1.216219425201416, "learning_rate": 2.8827529223319993e-07, "loss": 1.1021, "step": 26301 }, { "epoch": 0.9268039148321396, "grad_norm": 5.5620341300964355, "learning_rate": 2.8800330029351565e-07, "loss": 4.6193, "step": 26302 }, { "epoch": 0.9268391518450981, "grad_norm": 6.676120758056641, "learning_rate": 2.8773143485382516e-07, "loss": 5.5406, "step": 26303 }, { "epoch": 0.9268743888580565, "grad_norm": 2.90691876411438, "learning_rate": 2.87459695917669e-07, "loss": 0.7737, "step": 26304 }, { "epoch": 0.9269096258710149, "grad_norm": 3.3779261112213135, "learning_rate": 2.871880834885887e-07, "loss": 2.7084, "step": 26305 }, { "epoch": 0.9269448628839734, "grad_norm": 1.070462942123413, "learning_rate": 2.869165975701205e-07, "loss": 0.9245, "step": 26306 }, { "epoch": 0.9269800998969318, "grad_norm": 1.5746654272079468, "learning_rate": 2.8664523816580027e-07, "loss": 0.6547, "step": 26307 }, { "epoch": 0.9270153369098901, "grad_norm": 1.209254264831543, "learning_rate": 2.86374005279163e-07, "loss": 0.9373, "step": 26308 }, { "epoch": 0.9270505739228485, "grad_norm": 1.5622421503067017, "learning_rate": 2.8610289891374265e-07, "loss": 0.9, "step": 26309 }, { "epoch": 0.927085810935807, "grad_norm": 1.8008655309677124, "learning_rate": 2.858319190730685e-07, "loss": 0.7205, "step": 26310 }, { "epoch": 0.9271210479487654, "grad_norm": 3.0316202640533447, "learning_rate": 2.8556106576067224e-07, "loss": 3.1862, "step": 26311 }, { "epoch": 0.9271562849617238, "grad_norm": 6.32474946975708, "learning_rate": 2.8529033898007873e-07, "loss": 3.2217, "step": 26312 }, { "epoch": 0.9271915219746822, "grad_norm": 2.3972322940826416, "learning_rate": 2.850197387348175e-07, "loss": 2.5209, "step": 26313 }, { "epoch": 0.9272267589876406, "grad_norm": 3.389476776123047, "learning_rate": 2.8474926502841117e-07, "loss": 2.5134, "step": 26314 }, { "epoch": 0.927261996000599, "grad_norm": 3.067559003829956, "learning_rate": 2.8447891786438363e-07, "loss": 2.9024, "step": 26315 }, { "epoch": 0.9272972330135575, "grad_norm": 1.020460844039917, "learning_rate": 2.842086972462543e-07, "loss": 1.1309, "step": 26316 }, { "epoch": 0.9273324700265159, "grad_norm": 1.117534875869751, "learning_rate": 2.8393860317754594e-07, "loss": 0.7764, "step": 26317 }, { "epoch": 0.9273677070394742, "grad_norm": 1.0636565685272217, "learning_rate": 2.836686356617746e-07, "loss": 0.953, "step": 26318 }, { "epoch": 0.9274029440524326, "grad_norm": 5.796262264251709, "learning_rate": 2.8339879470245636e-07, "loss": 2.623, "step": 26319 }, { "epoch": 0.9274381810653911, "grad_norm": 1.144120454788208, "learning_rate": 2.8312908030310615e-07, "loss": 0.6311, "step": 26320 }, { "epoch": 0.9274734180783495, "grad_norm": 0.8090580105781555, "learning_rate": 2.8285949246723787e-07, "loss": 0.8533, "step": 26321 }, { "epoch": 0.9275086550913079, "grad_norm": 6.562923431396484, "learning_rate": 2.8259003119836315e-07, "loss": 3.3178, "step": 26322 }, { "epoch": 0.9275438921042664, "grad_norm": 3.7989282608032227, "learning_rate": 2.823206964999903e-07, "loss": 6.451, "step": 26323 }, { "epoch": 0.9275791291172247, "grad_norm": 1.769502878189087, "learning_rate": 2.8205148837562757e-07, "loss": 1.0774, "step": 26324 }, { "epoch": 0.9276143661301831, "grad_norm": 4.85618257522583, "learning_rate": 2.817824068287833e-07, "loss": 2.9748, "step": 26325 }, { "epoch": 0.9276496031431416, "grad_norm": 4.213757514953613, "learning_rate": 2.815134518629603e-07, "loss": 2.662, "step": 26326 }, { "epoch": 0.9276848401561, "grad_norm": 5.760912895202637, "learning_rate": 2.8124462348166347e-07, "loss": 5.1236, "step": 26327 }, { "epoch": 0.9277200771690584, "grad_norm": 2.7296149730682373, "learning_rate": 2.809759216883923e-07, "loss": 2.9299, "step": 26328 }, { "epoch": 0.9277553141820168, "grad_norm": 3.0317437648773193, "learning_rate": 2.807073464866483e-07, "loss": 2.6297, "step": 26329 }, { "epoch": 0.9277905511949752, "grad_norm": 3.739182233810425, "learning_rate": 2.8043889787992997e-07, "loss": 2.6578, "step": 26330 }, { "epoch": 0.9278257882079336, "grad_norm": 2.3271796703338623, "learning_rate": 2.80170575871731e-07, "loss": 2.7889, "step": 26331 }, { "epoch": 0.927861025220892, "grad_norm": 0.9255149960517883, "learning_rate": 2.7990238046554986e-07, "loss": 0.9976, "step": 26332 }, { "epoch": 0.9278962622338505, "grad_norm": 7.528196811676025, "learning_rate": 2.796343116648792e-07, "loss": 4.4793, "step": 26333 }, { "epoch": 0.9279314992468088, "grad_norm": 1.1584208011627197, "learning_rate": 2.793663694732085e-07, "loss": 0.7712, "step": 26334 }, { "epoch": 0.9279667362597672, "grad_norm": 0.7535448670387268, "learning_rate": 2.790985538940283e-07, "loss": 0.7862, "step": 26335 }, { "epoch": 0.9280019732727257, "grad_norm": 1.083496332168579, "learning_rate": 2.788308649308291e-07, "loss": 0.9241, "step": 26336 }, { "epoch": 0.9280372102856841, "grad_norm": 5.13268518447876, "learning_rate": 2.7856330258709594e-07, "loss": 5.0869, "step": 26337 }, { "epoch": 0.9280724472986425, "grad_norm": 0.9899873733520508, "learning_rate": 2.7829586686631384e-07, "loss": 0.9577, "step": 26338 }, { "epoch": 0.928107684311601, "grad_norm": 2.550617218017578, "learning_rate": 2.7802855777196657e-07, "loss": 2.7784, "step": 26339 }, { "epoch": 0.9281429213245593, "grad_norm": 7.741311550140381, "learning_rate": 2.7776137530753477e-07, "loss": 5.0461, "step": 26340 }, { "epoch": 0.9281781583375177, "grad_norm": 1.352955937385559, "learning_rate": 2.7749431947650117e-07, "loss": 0.7681, "step": 26341 }, { "epoch": 0.9282133953504761, "grad_norm": 5.092247009277344, "learning_rate": 2.772273902823408e-07, "loss": 4.1071, "step": 26342 }, { "epoch": 0.9282486323634346, "grad_norm": 1.8986166715621948, "learning_rate": 2.7696058772853306e-07, "loss": 0.673, "step": 26343 }, { "epoch": 0.928283869376393, "grad_norm": 2.8303253650665283, "learning_rate": 2.7669391181855186e-07, "loss": 2.4975, "step": 26344 }, { "epoch": 0.9283191063893513, "grad_norm": 3.2882156372070312, "learning_rate": 2.764273625558711e-07, "loss": 2.3465, "step": 26345 }, { "epoch": 0.9283543434023098, "grad_norm": 1.5223331451416016, "learning_rate": 2.7616093994396356e-07, "loss": 0.7228, "step": 26346 }, { "epoch": 0.9283895804152682, "grad_norm": 0.8651212453842163, "learning_rate": 2.758946439862964e-07, "loss": 0.9609, "step": 26347 }, { "epoch": 0.9284248174282266, "grad_norm": 1.4084328413009644, "learning_rate": 2.7562847468634133e-07, "loss": 0.8237, "step": 26348 }, { "epoch": 0.9284600544411851, "grad_norm": 4.433943748474121, "learning_rate": 2.7536243204756564e-07, "loss": 4.2508, "step": 26349 }, { "epoch": 0.9284952914541434, "grad_norm": 0.8337358236312866, "learning_rate": 2.750965160734298e-07, "loss": 0.8049, "step": 26350 }, { "epoch": 0.9285305284671018, "grad_norm": 3.9457943439483643, "learning_rate": 2.748307267674022e-07, "loss": 2.5097, "step": 26351 }, { "epoch": 0.9285657654800602, "grad_norm": 5.642788887023926, "learning_rate": 2.7456506413294446e-07, "loss": 6.9305, "step": 26352 }, { "epoch": 0.9286010024930187, "grad_norm": 0.8756047487258911, "learning_rate": 2.742995281735139e-07, "loss": 0.5609, "step": 26353 }, { "epoch": 0.9286362395059771, "grad_norm": 3.7404239177703857, "learning_rate": 2.740341188925699e-07, "loss": 2.4078, "step": 26354 }, { "epoch": 0.9286714765189354, "grad_norm": 3.0874295234680176, "learning_rate": 2.737688362935731e-07, "loss": 2.5507, "step": 26355 }, { "epoch": 0.9287067135318939, "grad_norm": 4.77244758605957, "learning_rate": 2.7350368037997397e-07, "loss": 2.9135, "step": 26356 }, { "epoch": 0.9287419505448523, "grad_norm": 1.4209057092666626, "learning_rate": 2.7323865115522986e-07, "loss": 0.8567, "step": 26357 }, { "epoch": 0.9287771875578107, "grad_norm": 1.1715844869613647, "learning_rate": 2.729737486227901e-07, "loss": 0.8195, "step": 26358 }, { "epoch": 0.9288124245707692, "grad_norm": 4.1782402992248535, "learning_rate": 2.7270897278610654e-07, "loss": 3.253, "step": 26359 }, { "epoch": 0.9288476615837276, "grad_norm": 1.3163703680038452, "learning_rate": 2.724443236486285e-07, "loss": 0.8844, "step": 26360 }, { "epoch": 0.9288828985966859, "grad_norm": 2.7631630897521973, "learning_rate": 2.7217980121380103e-07, "loss": 0.9286, "step": 26361 }, { "epoch": 0.9289181356096444, "grad_norm": 3.2964439392089844, "learning_rate": 2.7191540548507144e-07, "loss": 2.937, "step": 26362 }, { "epoch": 0.9289533726226028, "grad_norm": 1.1196908950805664, "learning_rate": 2.716511364658825e-07, "loss": 0.8201, "step": 26363 }, { "epoch": 0.9289886096355612, "grad_norm": 1.2308686971664429, "learning_rate": 2.7138699415967693e-07, "loss": 0.7394, "step": 26364 }, { "epoch": 0.9290238466485196, "grad_norm": 3.8320164680480957, "learning_rate": 2.711229785698954e-07, "loss": 2.6064, "step": 26365 }, { "epoch": 0.929059083661478, "grad_norm": 1.472455620765686, "learning_rate": 2.708590896999752e-07, "loss": 0.8098, "step": 26366 }, { "epoch": 0.9290943206744364, "grad_norm": 7.720072269439697, "learning_rate": 2.705953275533557e-07, "loss": 3.8067, "step": 26367 }, { "epoch": 0.9291295576873948, "grad_norm": 4.998473167419434, "learning_rate": 2.7033169213347197e-07, "loss": 3.403, "step": 26368 }, { "epoch": 0.9291647947003533, "grad_norm": 0.998491108417511, "learning_rate": 2.7006818344375575e-07, "loss": 0.764, "step": 26369 }, { "epoch": 0.9292000317133117, "grad_norm": 0.911651074886322, "learning_rate": 2.698048014876409e-07, "loss": 0.9453, "step": 26370 }, { "epoch": 0.92923526872627, "grad_norm": 2.5810883045196533, "learning_rate": 2.6954154626856023e-07, "loss": 2.4632, "step": 26371 }, { "epoch": 0.9292705057392285, "grad_norm": 0.9765625596046448, "learning_rate": 2.692784177899377e-07, "loss": 1.1055, "step": 26372 }, { "epoch": 0.9293057427521869, "grad_norm": 1.1860653162002563, "learning_rate": 2.690154160552039e-07, "loss": 0.8889, "step": 26373 }, { "epoch": 0.9293409797651453, "grad_norm": 5.32572078704834, "learning_rate": 2.687525410677849e-07, "loss": 2.7822, "step": 26374 }, { "epoch": 0.9293762167781037, "grad_norm": 2.3013648986816406, "learning_rate": 2.6848979283110256e-07, "loss": 2.4414, "step": 26375 }, { "epoch": 0.9294114537910622, "grad_norm": 1.3854780197143555, "learning_rate": 2.6822717134858066e-07, "loss": 0.9201, "step": 26376 }, { "epoch": 0.9294466908040205, "grad_norm": 3.630687952041626, "learning_rate": 2.679646766236399e-07, "loss": 2.8618, "step": 26377 }, { "epoch": 0.9294819278169789, "grad_norm": 5.932138919830322, "learning_rate": 2.677023086596986e-07, "loss": 2.487, "step": 26378 }, { "epoch": 0.9295171648299374, "grad_norm": 2.002110481262207, "learning_rate": 2.6744006746017404e-07, "loss": 0.7882, "step": 26379 }, { "epoch": 0.9295524018428958, "grad_norm": 3.3716957569122314, "learning_rate": 2.6717795302848236e-07, "loss": 3.2561, "step": 26380 }, { "epoch": 0.9295876388558542, "grad_norm": 1.1372584104537964, "learning_rate": 2.6691596536803755e-07, "loss": 0.6954, "step": 26381 }, { "epoch": 0.9296228758688126, "grad_norm": 3.048530340194702, "learning_rate": 2.6665410448225237e-07, "loss": 2.2635, "step": 26382 }, { "epoch": 0.929658112881771, "grad_norm": 1.1351042985916138, "learning_rate": 2.6639237037453744e-07, "loss": 0.879, "step": 26383 }, { "epoch": 0.9296933498947294, "grad_norm": 2.6830804347991943, "learning_rate": 2.6613076304830124e-07, "loss": 2.6874, "step": 26384 }, { "epoch": 0.9297285869076878, "grad_norm": 3.552582263946533, "learning_rate": 2.65869282506952e-07, "loss": 2.701, "step": 26385 }, { "epoch": 0.9297638239206463, "grad_norm": 1.0748984813690186, "learning_rate": 2.6560792875389597e-07, "loss": 0.9544, "step": 26386 }, { "epoch": 0.9297990609336046, "grad_norm": 1.0856088399887085, "learning_rate": 2.653467017925371e-07, "loss": 1.0966, "step": 26387 }, { "epoch": 0.929834297946563, "grad_norm": 5.128575801849365, "learning_rate": 2.6508560162627486e-07, "loss": 6.6977, "step": 26388 }, { "epoch": 0.9298695349595215, "grad_norm": 1.919223427772522, "learning_rate": 2.648246282585143e-07, "loss": 0.9781, "step": 26389 }, { "epoch": 0.9299047719724799, "grad_norm": 4.380181789398193, "learning_rate": 2.645637816926538e-07, "loss": 0.9642, "step": 26390 }, { "epoch": 0.9299400089854383, "grad_norm": 1.1348249912261963, "learning_rate": 2.6430306193208966e-07, "loss": 0.9654, "step": 26391 }, { "epoch": 0.9299752459983968, "grad_norm": 4.030322074890137, "learning_rate": 2.640424689802168e-07, "loss": 3.2957, "step": 26392 }, { "epoch": 0.9300104830113551, "grad_norm": 7.357079029083252, "learning_rate": 2.6378200284043256e-07, "loss": 2.6998, "step": 26393 }, { "epoch": 0.9300457200243135, "grad_norm": 2.9821996688842773, "learning_rate": 2.635216635161275e-07, "loss": 2.3661, "step": 26394 }, { "epoch": 0.930080957037272, "grad_norm": 7.603290557861328, "learning_rate": 2.6326145101069343e-07, "loss": 7.3459, "step": 26395 }, { "epoch": 0.9301161940502304, "grad_norm": 3.9570043087005615, "learning_rate": 2.630013653275187e-07, "loss": 2.9742, "step": 26396 }, { "epoch": 0.9301514310631888, "grad_norm": 8.295045852661133, "learning_rate": 2.627414064699907e-07, "loss": 5.2469, "step": 26397 }, { "epoch": 0.9301866680761471, "grad_norm": 5.121896743774414, "learning_rate": 2.624815744414977e-07, "loss": 3.282, "step": 26398 }, { "epoch": 0.9302219050891056, "grad_norm": 1.024206280708313, "learning_rate": 2.622218692454215e-07, "loss": 0.9874, "step": 26399 }, { "epoch": 0.930257142102064, "grad_norm": 0.8873658776283264, "learning_rate": 2.6196229088514604e-07, "loss": 0.7299, "step": 26400 }, { "epoch": 0.9302923791150224, "grad_norm": 1.0551774501800537, "learning_rate": 2.6170283936405306e-07, "loss": 0.9908, "step": 26401 }, { "epoch": 0.9303276161279809, "grad_norm": 5.480459690093994, "learning_rate": 2.614435146855199e-07, "loss": 5.2022, "step": 26402 }, { "epoch": 0.9303628531409392, "grad_norm": 16.556425094604492, "learning_rate": 2.61184316852926e-07, "loss": 5.5098, "step": 26403 }, { "epoch": 0.9303980901538976, "grad_norm": 1.2867764234542847, "learning_rate": 2.609252458696465e-07, "loss": 1.1212, "step": 26404 }, { "epoch": 0.9304333271668561, "grad_norm": 6.1398396492004395, "learning_rate": 2.606663017390565e-07, "loss": 4.2756, "step": 26405 }, { "epoch": 0.9304685641798145, "grad_norm": 5.552177429199219, "learning_rate": 2.604074844645299e-07, "loss": 3.8697, "step": 26406 }, { "epoch": 0.9305038011927729, "grad_norm": 2.2544925212860107, "learning_rate": 2.6014879404943404e-07, "loss": 0.8084, "step": 26407 }, { "epoch": 0.9305390382057313, "grad_norm": 3.2968645095825195, "learning_rate": 2.598902304971418e-07, "loss": 2.6681, "step": 26408 }, { "epoch": 0.9305742752186897, "grad_norm": 2.9219977855682373, "learning_rate": 2.5963179381102044e-07, "loss": 2.8605, "step": 26409 }, { "epoch": 0.9306095122316481, "grad_norm": 1.1252657175064087, "learning_rate": 2.5937348399443506e-07, "loss": 0.8775, "step": 26410 }, { "epoch": 0.9306447492446065, "grad_norm": 1.3330063819885254, "learning_rate": 2.591153010507508e-07, "loss": 0.7947, "step": 26411 }, { "epoch": 0.930679986257565, "grad_norm": 12.415190696716309, "learning_rate": 2.5885724498333154e-07, "loss": 4.1238, "step": 26412 }, { "epoch": 0.9307152232705234, "grad_norm": 1.1323676109313965, "learning_rate": 2.585993157955358e-07, "loss": 0.772, "step": 26413 }, { "epoch": 0.9307504602834817, "grad_norm": 1.3279763460159302, "learning_rate": 2.583415134907263e-07, "loss": 0.8815, "step": 26414 }, { "epoch": 0.9307856972964402, "grad_norm": 6.304393291473389, "learning_rate": 2.5808383807225836e-07, "loss": 7.5404, "step": 26415 }, { "epoch": 0.9308209343093986, "grad_norm": 8.509758949279785, "learning_rate": 2.578262895434902e-07, "loss": 4.3951, "step": 26416 }, { "epoch": 0.930856171322357, "grad_norm": 1.2685333490371704, "learning_rate": 2.575688679077759e-07, "loss": 0.8195, "step": 26417 }, { "epoch": 0.9308914083353154, "grad_norm": 2.3819580078125, "learning_rate": 2.5731157316846724e-07, "loss": 1.0086, "step": 26418 }, { "epoch": 0.9309266453482739, "grad_norm": 2.97879695892334, "learning_rate": 2.570544053289159e-07, "loss": 3.3565, "step": 26419 }, { "epoch": 0.9309618823612322, "grad_norm": 1.308502435684204, "learning_rate": 2.567973643924726e-07, "loss": 0.7868, "step": 26420 }, { "epoch": 0.9309971193741906, "grad_norm": 5.156097412109375, "learning_rate": 2.5654045036248466e-07, "loss": 2.7482, "step": 26421 }, { "epoch": 0.9310323563871491, "grad_norm": 4.171898365020752, "learning_rate": 2.5628366324230046e-07, "loss": 4.6471, "step": 26422 }, { "epoch": 0.9310675934001075, "grad_norm": 0.9606282711029053, "learning_rate": 2.5602700303525963e-07, "loss": 0.9181, "step": 26423 }, { "epoch": 0.9311028304130659, "grad_norm": 1.276032567024231, "learning_rate": 2.5577046974470944e-07, "loss": 0.8827, "step": 26424 }, { "epoch": 0.9311380674260243, "grad_norm": 2.011500835418701, "learning_rate": 2.555140633739905e-07, "loss": 1.0177, "step": 26425 }, { "epoch": 0.9311733044389827, "grad_norm": 4.265500068664551, "learning_rate": 2.552577839264425e-07, "loss": 2.3866, "step": 26426 }, { "epoch": 0.9312085414519411, "grad_norm": 1.2065761089324951, "learning_rate": 2.550016314054005e-07, "loss": 0.9347, "step": 26427 }, { "epoch": 0.9312437784648996, "grad_norm": 3.3776912689208984, "learning_rate": 2.5474560581420726e-07, "loss": 2.3199, "step": 26428 }, { "epoch": 0.931279015477858, "grad_norm": 3.035282611846924, "learning_rate": 2.544897071561914e-07, "loss": 3.2814, "step": 26429 }, { "epoch": 0.9313142524908163, "grad_norm": 2.829986333847046, "learning_rate": 2.54233935434689e-07, "loss": 3.3079, "step": 26430 }, { "epoch": 0.9313494895037747, "grad_norm": 3.4629886150360107, "learning_rate": 2.539782906530319e-07, "loss": 2.8069, "step": 26431 }, { "epoch": 0.9313847265167332, "grad_norm": 3.5684425830841064, "learning_rate": 2.537227728145475e-07, "loss": 2.6796, "step": 26432 }, { "epoch": 0.9314199635296916, "grad_norm": 1.5085444450378418, "learning_rate": 2.5346738192256635e-07, "loss": 0.739, "step": 26433 }, { "epoch": 0.93145520054265, "grad_norm": 7.023399829864502, "learning_rate": 2.5321211798041367e-07, "loss": 5.3973, "step": 26434 }, { "epoch": 0.9314904375556085, "grad_norm": 3.8724653720855713, "learning_rate": 2.5295698099141566e-07, "loss": 2.6115, "step": 26435 }, { "epoch": 0.9315256745685668, "grad_norm": 1.1904288530349731, "learning_rate": 2.5270197095889406e-07, "loss": 1.2162, "step": 26436 }, { "epoch": 0.9315609115815252, "grad_norm": 4.861310005187988, "learning_rate": 2.5244708788617067e-07, "loss": 2.8558, "step": 26437 }, { "epoch": 0.9315961485944837, "grad_norm": 2.481971025466919, "learning_rate": 2.521923317765651e-07, "loss": 2.6467, "step": 26438 }, { "epoch": 0.9316313856074421, "grad_norm": 2.957404613494873, "learning_rate": 2.519377026333969e-07, "loss": 2.9877, "step": 26439 }, { "epoch": 0.9316666226204005, "grad_norm": 5.100043773651123, "learning_rate": 2.516832004599812e-07, "loss": 4.557, "step": 26440 }, { "epoch": 0.9317018596333588, "grad_norm": 1.4996665716171265, "learning_rate": 2.514288252596353e-07, "loss": 0.7317, "step": 26441 }, { "epoch": 0.9317370966463173, "grad_norm": 3.365790367126465, "learning_rate": 2.5117457703566774e-07, "loss": 3.3132, "step": 26442 }, { "epoch": 0.9317723336592757, "grad_norm": 2.680938959121704, "learning_rate": 2.5092045579139467e-07, "loss": 2.849, "step": 26443 }, { "epoch": 0.9318075706722341, "grad_norm": 3.202216386795044, "learning_rate": 2.5066646153012577e-07, "loss": 0.7272, "step": 26444 }, { "epoch": 0.9318428076851926, "grad_norm": 1.5150319337844849, "learning_rate": 2.5041259425516604e-07, "loss": 0.8783, "step": 26445 }, { "epoch": 0.9318780446981509, "grad_norm": 1.1023612022399902, "learning_rate": 2.5015885396982406e-07, "loss": 0.8388, "step": 26446 }, { "epoch": 0.9319132817111093, "grad_norm": 2.2563841342926025, "learning_rate": 2.4990524067740717e-07, "loss": 1.0432, "step": 26447 }, { "epoch": 0.9319485187240678, "grad_norm": 1.0322262048721313, "learning_rate": 2.4965175438121493e-07, "loss": 1.1074, "step": 26448 }, { "epoch": 0.9319837557370262, "grad_norm": 6.40895414352417, "learning_rate": 2.493983950845513e-07, "loss": 6.8519, "step": 26449 }, { "epoch": 0.9320189927499846, "grad_norm": 1.5445616245269775, "learning_rate": 2.491451627907149e-07, "loss": 0.8863, "step": 26450 }, { "epoch": 0.9320542297629429, "grad_norm": 4.082005977630615, "learning_rate": 2.4889205750300517e-07, "loss": 2.9723, "step": 26451 }, { "epoch": 0.9320894667759014, "grad_norm": 3.7252235412597656, "learning_rate": 2.486390792247173e-07, "loss": 2.9881, "step": 26452 }, { "epoch": 0.9321247037888598, "grad_norm": 5.501541614532471, "learning_rate": 2.483862279591487e-07, "loss": 6.2652, "step": 26453 }, { "epoch": 0.9321599408018182, "grad_norm": 1.1674226522445679, "learning_rate": 2.481335037095922e-07, "loss": 0.7372, "step": 26454 }, { "epoch": 0.9321951778147767, "grad_norm": 3.468824863433838, "learning_rate": 2.478809064793375e-07, "loss": 2.1302, "step": 26455 }, { "epoch": 0.932230414827735, "grad_norm": 3.6094822883605957, "learning_rate": 2.4762843627167745e-07, "loss": 2.8862, "step": 26456 }, { "epoch": 0.9322656518406934, "grad_norm": 1.0141745805740356, "learning_rate": 2.473760930898983e-07, "loss": 1.0109, "step": 26457 }, { "epoch": 0.9323008888536519, "grad_norm": 1.2229264974594116, "learning_rate": 2.471238769372886e-07, "loss": 0.8637, "step": 26458 }, { "epoch": 0.9323361258666103, "grad_norm": 6.489992141723633, "learning_rate": 2.4687178781713227e-07, "loss": 2.952, "step": 26459 }, { "epoch": 0.9323713628795687, "grad_norm": 4.731193542480469, "learning_rate": 2.466198257327146e-07, "loss": 4.8074, "step": 26460 }, { "epoch": 0.9324065998925272, "grad_norm": 3.01674485206604, "learning_rate": 2.46367990687314e-07, "loss": 3.0593, "step": 26461 }, { "epoch": 0.9324418369054855, "grad_norm": 4.433313369750977, "learning_rate": 2.461162826842134e-07, "loss": 3.1066, "step": 26462 }, { "epoch": 0.9324770739184439, "grad_norm": 3.3121678829193115, "learning_rate": 2.458647017266924e-07, "loss": 2.5969, "step": 26463 }, { "epoch": 0.9325123109314023, "grad_norm": 2.7032415866851807, "learning_rate": 2.456132478180251e-07, "loss": 2.8478, "step": 26464 }, { "epoch": 0.9325475479443608, "grad_norm": 5.0020976066589355, "learning_rate": 2.4536192096148547e-07, "loss": 2.8919, "step": 26465 }, { "epoch": 0.9325827849573192, "grad_norm": 3.7299578189849854, "learning_rate": 2.4511072116035315e-07, "loss": 2.8321, "step": 26466 }, { "epoch": 0.9326180219702775, "grad_norm": 1.4532392024993896, "learning_rate": 2.4485964841789443e-07, "loss": 0.6778, "step": 26467 }, { "epoch": 0.932653258983236, "grad_norm": 8.856622695922852, "learning_rate": 2.446087027373811e-07, "loss": 2.7715, "step": 26468 }, { "epoch": 0.9326884959961944, "grad_norm": 3.455716133117676, "learning_rate": 2.4435788412208175e-07, "loss": 2.7659, "step": 26469 }, { "epoch": 0.9327237330091528, "grad_norm": 1.0290361642837524, "learning_rate": 2.441071925752647e-07, "loss": 0.9402, "step": 26470 }, { "epoch": 0.9327589700221113, "grad_norm": 3.413066864013672, "learning_rate": 2.4385662810019305e-07, "loss": 3.0814, "step": 26471 }, { "epoch": 0.9327942070350697, "grad_norm": 2.4327750205993652, "learning_rate": 2.436061907001319e-07, "loss": 2.9861, "step": 26472 }, { "epoch": 0.932829444048028, "grad_norm": 4.004733562469482, "learning_rate": 2.433558803783431e-07, "loss": 3.5488, "step": 26473 }, { "epoch": 0.9328646810609864, "grad_norm": 1.3195865154266357, "learning_rate": 2.431056971380874e-07, "loss": 0.8857, "step": 26474 }, { "epoch": 0.9328999180739449, "grad_norm": 4.960763931274414, "learning_rate": 2.428556409826222e-07, "loss": 3.054, "step": 26475 }, { "epoch": 0.9329351550869033, "grad_norm": 3.867978572845459, "learning_rate": 2.426057119152048e-07, "loss": 2.5927, "step": 26476 }, { "epoch": 0.9329703920998617, "grad_norm": 0.990493893623352, "learning_rate": 2.4235590993909155e-07, "loss": 0.977, "step": 26477 }, { "epoch": 0.9330056291128201, "grad_norm": 0.8717208504676819, "learning_rate": 2.421062350575354e-07, "loss": 0.6556, "step": 26478 }, { "epoch": 0.9330408661257785, "grad_norm": 1.0954983234405518, "learning_rate": 2.418566872737882e-07, "loss": 0.9391, "step": 26479 }, { "epoch": 0.9330761031387369, "grad_norm": 5.344428539276123, "learning_rate": 2.4160726659110066e-07, "loss": 5.5228, "step": 26480 }, { "epoch": 0.9331113401516954, "grad_norm": 2.0250661373138428, "learning_rate": 2.4135797301272137e-07, "loss": 0.7836, "step": 26481 }, { "epoch": 0.9331465771646538, "grad_norm": 1.1631371974945068, "learning_rate": 2.4110880654189873e-07, "loss": 0.6869, "step": 26482 }, { "epoch": 0.9331818141776121, "grad_norm": 4.157033920288086, "learning_rate": 2.4085976718187576e-07, "loss": 5.0659, "step": 26483 }, { "epoch": 0.9332170511905705, "grad_norm": 5.196763038635254, "learning_rate": 2.406108549358965e-07, "loss": 6.8854, "step": 26484 }, { "epoch": 0.933252288203529, "grad_norm": 1.1781799793243408, "learning_rate": 2.403620698072062e-07, "loss": 1.0612, "step": 26485 }, { "epoch": 0.9332875252164874, "grad_norm": 4.891690254211426, "learning_rate": 2.401134117990422e-07, "loss": 3.2698, "step": 26486 }, { "epoch": 0.9333227622294458, "grad_norm": 4.262207508087158, "learning_rate": 2.3986488091464423e-07, "loss": 3.0447, "step": 26487 }, { "epoch": 0.9333579992424043, "grad_norm": 6.062455654144287, "learning_rate": 2.3961647715724844e-07, "loss": 6.9153, "step": 26488 }, { "epoch": 0.9333932362553626, "grad_norm": 1.200947880744934, "learning_rate": 2.3936820053009237e-07, "loss": 0.7806, "step": 26489 }, { "epoch": 0.933428473268321, "grad_norm": 1.0057419538497925, "learning_rate": 2.391200510364089e-07, "loss": 0.9048, "step": 26490 }, { "epoch": 0.9334637102812795, "grad_norm": 3.6467838287353516, "learning_rate": 2.388720286794299e-07, "loss": 3.6224, "step": 26491 }, { "epoch": 0.9334989472942379, "grad_norm": 3.157003879547119, "learning_rate": 2.3862413346238513e-07, "loss": 3.0291, "step": 26492 }, { "epoch": 0.9335341843071963, "grad_norm": 4.498104095458984, "learning_rate": 2.3837636538850518e-07, "loss": 5.0836, "step": 26493 }, { "epoch": 0.9335694213201546, "grad_norm": 3.620378255844116, "learning_rate": 2.3812872446101754e-07, "loss": 4.9276, "step": 26494 }, { "epoch": 0.9336046583331131, "grad_norm": 1.5397658348083496, "learning_rate": 2.3788121068314518e-07, "loss": 0.6705, "step": 26495 }, { "epoch": 0.9336398953460715, "grad_norm": 1.203768253326416, "learning_rate": 2.3763382405811442e-07, "loss": 0.8174, "step": 26496 }, { "epoch": 0.9336751323590299, "grad_norm": 1.1375529766082764, "learning_rate": 2.37386564589146e-07, "loss": 1.0769, "step": 26497 }, { "epoch": 0.9337103693719884, "grad_norm": 4.112948417663574, "learning_rate": 2.371394322794629e-07, "loss": 2.7727, "step": 26498 }, { "epoch": 0.9337456063849467, "grad_norm": 1.0942165851593018, "learning_rate": 2.3689242713227922e-07, "loss": 0.8944, "step": 26499 }, { "epoch": 0.9337808433979051, "grad_norm": 2.8105711936950684, "learning_rate": 2.3664554915081683e-07, "loss": 1.1053, "step": 26500 }, { "epoch": 0.9338160804108636, "grad_norm": 1.0233924388885498, "learning_rate": 2.3639879833829094e-07, "loss": 1.1813, "step": 26501 }, { "epoch": 0.933851317423822, "grad_norm": 1.2846813201904297, "learning_rate": 2.361521746979134e-07, "loss": 0.9189, "step": 26502 }, { "epoch": 0.9338865544367804, "grad_norm": 2.4652984142303467, "learning_rate": 2.3590567823289723e-07, "loss": 3.0532, "step": 26503 }, { "epoch": 0.9339217914497389, "grad_norm": 3.328118324279785, "learning_rate": 2.3565930894645427e-07, "loss": 2.915, "step": 26504 }, { "epoch": 0.9339570284626972, "grad_norm": 1.1234210729599, "learning_rate": 2.35413066841792e-07, "loss": 0.9045, "step": 26505 }, { "epoch": 0.9339922654756556, "grad_norm": 4.451313018798828, "learning_rate": 2.351669519221178e-07, "loss": 3.1773, "step": 26506 }, { "epoch": 0.934027502488614, "grad_norm": 0.8962860703468323, "learning_rate": 2.3492096419063804e-07, "loss": 0.7702, "step": 26507 }, { "epoch": 0.9340627395015725, "grad_norm": 0.9462625980377197, "learning_rate": 2.346751036505579e-07, "loss": 0.8253, "step": 26508 }, { "epoch": 0.9340979765145309, "grad_norm": 1.1894854307174683, "learning_rate": 2.3442937030507705e-07, "loss": 0.8568, "step": 26509 }, { "epoch": 0.9341332135274892, "grad_norm": 5.901638031005859, "learning_rate": 2.341837641573974e-07, "loss": 6.9357, "step": 26510 }, { "epoch": 0.9341684505404477, "grad_norm": 4.805041790008545, "learning_rate": 2.339382852107197e-07, "loss": 3.2714, "step": 26511 }, { "epoch": 0.9342036875534061, "grad_norm": 3.4778237342834473, "learning_rate": 2.3369293346823806e-07, "loss": 2.9477, "step": 26512 }, { "epoch": 0.9342389245663645, "grad_norm": 3.116300106048584, "learning_rate": 2.3344770893315104e-07, "loss": 2.889, "step": 26513 }, { "epoch": 0.934274161579323, "grad_norm": 1.6062606573104858, "learning_rate": 2.3320261160865166e-07, "loss": 0.8909, "step": 26514 }, { "epoch": 0.9343093985922813, "grad_norm": 2.3924741744995117, "learning_rate": 2.3295764149793177e-07, "loss": 2.6274, "step": 26515 }, { "epoch": 0.9343446356052397, "grad_norm": 4.256228923797607, "learning_rate": 2.3271279860418326e-07, "loss": 2.9008, "step": 26516 }, { "epoch": 0.9343798726181981, "grad_norm": 3.9634575843811035, "learning_rate": 2.3246808293059586e-07, "loss": 4.6038, "step": 26517 }, { "epoch": 0.9344151096311566, "grad_norm": 3.371504783630371, "learning_rate": 2.3222349448035365e-07, "loss": 2.7339, "step": 26518 }, { "epoch": 0.934450346644115, "grad_norm": 3.4583797454833984, "learning_rate": 2.3197903325664516e-07, "loss": 2.4844, "step": 26519 }, { "epoch": 0.9344855836570733, "grad_norm": 1.3657339811325073, "learning_rate": 2.3173469926265568e-07, "loss": 0.8007, "step": 26520 }, { "epoch": 0.9345208206700318, "grad_norm": 1.5560566186904907, "learning_rate": 2.3149049250156484e-07, "loss": 0.8684, "step": 26521 }, { "epoch": 0.9345560576829902, "grad_norm": 0.9714260697364807, "learning_rate": 2.312464129765535e-07, "loss": 1.1893, "step": 26522 }, { "epoch": 0.9345912946959486, "grad_norm": 4.243856430053711, "learning_rate": 2.3100246069080346e-07, "loss": 5.3401, "step": 26523 }, { "epoch": 0.9346265317089071, "grad_norm": 3.54445219039917, "learning_rate": 2.3075863564749002e-07, "loss": 2.748, "step": 26524 }, { "epoch": 0.9346617687218655, "grad_norm": 1.0431827306747437, "learning_rate": 2.3051493784979063e-07, "loss": 0.9931, "step": 26525 }, { "epoch": 0.9346970057348238, "grad_norm": 2.965665340423584, "learning_rate": 2.3027136730087719e-07, "loss": 2.3839, "step": 26526 }, { "epoch": 0.9347322427477822, "grad_norm": 1.2709627151489258, "learning_rate": 2.3002792400392493e-07, "loss": 1.1245, "step": 26527 }, { "epoch": 0.9347674797607407, "grad_norm": 3.4515364170074463, "learning_rate": 2.2978460796210356e-07, "loss": 4.1018, "step": 26528 }, { "epoch": 0.9348027167736991, "grad_norm": 3.770906448364258, "learning_rate": 2.2954141917858164e-07, "loss": 4.9299, "step": 26529 }, { "epoch": 0.9348379537866575, "grad_norm": 4.934622764587402, "learning_rate": 2.2929835765652663e-07, "loss": 4.9046, "step": 26530 }, { "epoch": 0.934873190799616, "grad_norm": 0.8857519030570984, "learning_rate": 2.2905542339910602e-07, "loss": 0.5328, "step": 26531 }, { "epoch": 0.9349084278125743, "grad_norm": 0.9537298679351807, "learning_rate": 2.2881261640948281e-07, "loss": 0.8003, "step": 26532 }, { "epoch": 0.9349436648255327, "grad_norm": 4.532182216644287, "learning_rate": 2.2856993669082005e-07, "loss": 2.655, "step": 26533 }, { "epoch": 0.9349789018384912, "grad_norm": 3.797746419906616, "learning_rate": 2.2832738424627853e-07, "loss": 3.0086, "step": 26534 }, { "epoch": 0.9350141388514496, "grad_norm": 4.391705513000488, "learning_rate": 2.2808495907901796e-07, "loss": 2.503, "step": 26535 }, { "epoch": 0.935049375864408, "grad_norm": 6.5120062828063965, "learning_rate": 2.2784266119219577e-07, "loss": 3.3314, "step": 26536 }, { "epoch": 0.9350846128773664, "grad_norm": 0.9990464448928833, "learning_rate": 2.2760049058896615e-07, "loss": 0.718, "step": 26537 }, { "epoch": 0.9351198498903248, "grad_norm": 4.468470096588135, "learning_rate": 2.2735844727248547e-07, "loss": 3.3023, "step": 26538 }, { "epoch": 0.9351550869032832, "grad_norm": 3.3615074157714844, "learning_rate": 2.2711653124590672e-07, "loss": 2.7423, "step": 26539 }, { "epoch": 0.9351903239162416, "grad_norm": 3.181731700897217, "learning_rate": 2.2687474251237962e-07, "loss": 2.9431, "step": 26540 }, { "epoch": 0.9352255609292001, "grad_norm": 3.4060745239257812, "learning_rate": 2.2663308107505277e-07, "loss": 2.5627, "step": 26541 }, { "epoch": 0.9352607979421584, "grad_norm": 1.3971054553985596, "learning_rate": 2.26391546937077e-07, "loss": 1.1413, "step": 26542 }, { "epoch": 0.9352960349551168, "grad_norm": 1.3750479221343994, "learning_rate": 2.261501401015953e-07, "loss": 0.9798, "step": 26543 }, { "epoch": 0.9353312719680753, "grad_norm": 4.977765083312988, "learning_rate": 2.25908860571753e-07, "loss": 4.3355, "step": 26544 }, { "epoch": 0.9353665089810337, "grad_norm": 2.8796803951263428, "learning_rate": 2.2566770835069195e-07, "loss": 2.3398, "step": 26545 }, { "epoch": 0.9354017459939921, "grad_norm": 4.37062406539917, "learning_rate": 2.2542668344155527e-07, "loss": 2.9459, "step": 26546 }, { "epoch": 0.9354369830069506, "grad_norm": 6.19272518157959, "learning_rate": 2.251857858474815e-07, "loss": 4.9044, "step": 26547 }, { "epoch": 0.9354722200199089, "grad_norm": 4.079282283782959, "learning_rate": 2.249450155716071e-07, "loss": 2.7976, "step": 26548 }, { "epoch": 0.9355074570328673, "grad_norm": 4.474201679229736, "learning_rate": 2.2470437261706834e-07, "loss": 4.3115, "step": 26549 }, { "epoch": 0.9355426940458257, "grad_norm": 4.148260116577148, "learning_rate": 2.2446385698700167e-07, "loss": 2.8505, "step": 26550 }, { "epoch": 0.9355779310587842, "grad_norm": 1.281996250152588, "learning_rate": 2.2422346868453792e-07, "loss": 0.8402, "step": 26551 }, { "epoch": 0.9356131680717426, "grad_norm": 1.4103728532791138, "learning_rate": 2.2398320771280902e-07, "loss": 0.7699, "step": 26552 }, { "epoch": 0.9356484050847009, "grad_norm": 5.083745002746582, "learning_rate": 2.2374307407494357e-07, "loss": 4.3562, "step": 26553 }, { "epoch": 0.9356836420976594, "grad_norm": 3.3143470287323, "learning_rate": 2.2350306777407016e-07, "loss": 3.2381, "step": 26554 }, { "epoch": 0.9357188791106178, "grad_norm": 8.076885223388672, "learning_rate": 2.232631888133152e-07, "loss": 3.2232, "step": 26555 }, { "epoch": 0.9357541161235762, "grad_norm": 5.60643196105957, "learning_rate": 2.2302343719580066e-07, "loss": 3.4725, "step": 26556 }, { "epoch": 0.9357893531365347, "grad_norm": 2.4227473735809326, "learning_rate": 2.227838129246529e-07, "loss": 2.6971, "step": 26557 }, { "epoch": 0.935824590149493, "grad_norm": 5.636990070343018, "learning_rate": 2.2254431600299163e-07, "loss": 6.8634, "step": 26558 }, { "epoch": 0.9358598271624514, "grad_norm": 1.729295015335083, "learning_rate": 2.2230494643393553e-07, "loss": 0.7527, "step": 26559 }, { "epoch": 0.9358950641754098, "grad_norm": 2.804969310760498, "learning_rate": 2.2206570422060204e-07, "loss": 2.9884, "step": 26560 }, { "epoch": 0.9359303011883683, "grad_norm": 5.018387317657471, "learning_rate": 2.218265893661087e-07, "loss": 2.7819, "step": 26561 }, { "epoch": 0.9359655382013267, "grad_norm": 3.9422762393951416, "learning_rate": 2.2158760187357075e-07, "loss": 2.6838, "step": 26562 }, { "epoch": 0.936000775214285, "grad_norm": 3.1702749729156494, "learning_rate": 2.2134874174609912e-07, "loss": 0.7517, "step": 26563 }, { "epoch": 0.9360360122272435, "grad_norm": 3.770028591156006, "learning_rate": 2.2111000898680567e-07, "loss": 0.8382, "step": 26564 }, { "epoch": 0.9360712492402019, "grad_norm": 1.6877275705337524, "learning_rate": 2.2087140359880023e-07, "loss": 1.0153, "step": 26565 }, { "epoch": 0.9361064862531603, "grad_norm": 1.4374946355819702, "learning_rate": 2.2063292558519022e-07, "loss": 0.8703, "step": 26566 }, { "epoch": 0.9361417232661188, "grad_norm": 1.1836251020431519, "learning_rate": 2.2039457494908213e-07, "loss": 0.6786, "step": 26567 }, { "epoch": 0.9361769602790772, "grad_norm": 7.198472499847412, "learning_rate": 2.2015635169358007e-07, "loss": 3.2722, "step": 26568 }, { "epoch": 0.9362121972920355, "grad_norm": 5.994678497314453, "learning_rate": 2.1991825582178826e-07, "loss": 4.9218, "step": 26569 }, { "epoch": 0.936247434304994, "grad_norm": 15.13982105255127, "learning_rate": 2.1968028733680534e-07, "loss": 6.5289, "step": 26570 }, { "epoch": 0.9362826713179524, "grad_norm": 3.3991918563842773, "learning_rate": 2.1944244624173438e-07, "loss": 2.1668, "step": 26571 }, { "epoch": 0.9363179083309108, "grad_norm": 0.9897719025611877, "learning_rate": 2.1920473253966956e-07, "loss": 0.6458, "step": 26572 }, { "epoch": 0.9363531453438692, "grad_norm": 1.186597228050232, "learning_rate": 2.1896714623371064e-07, "loss": 0.9882, "step": 26573 }, { "epoch": 0.9363883823568276, "grad_norm": 3.2439637184143066, "learning_rate": 2.1872968732695064e-07, "loss": 2.466, "step": 26574 }, { "epoch": 0.936423619369786, "grad_norm": 1.088554859161377, "learning_rate": 2.1849235582248052e-07, "loss": 0.8576, "step": 26575 }, { "epoch": 0.9364588563827444, "grad_norm": 1.1612833738327026, "learning_rate": 2.182551517233944e-07, "loss": 0.6168, "step": 26576 }, { "epoch": 0.9364940933957029, "grad_norm": 1.2504428625106812, "learning_rate": 2.1801807503278094e-07, "loss": 0.892, "step": 26577 }, { "epoch": 0.9365293304086613, "grad_norm": 3.3771145343780518, "learning_rate": 2.177811257537288e-07, "loss": 2.8386, "step": 26578 }, { "epoch": 0.9365645674216196, "grad_norm": 39.611515045166016, "learning_rate": 2.1754430388932212e-07, "loss": 3.1033, "step": 26579 }, { "epoch": 0.9365998044345781, "grad_norm": 10.950066566467285, "learning_rate": 2.173076094426474e-07, "loss": 5.5713, "step": 26580 }, { "epoch": 0.9366350414475365, "grad_norm": 0.9171818494796753, "learning_rate": 2.1707104241678766e-07, "loss": 0.9445, "step": 26581 }, { "epoch": 0.9366702784604949, "grad_norm": 0.8186231851577759, "learning_rate": 2.168346028148238e-07, "loss": 0.7209, "step": 26582 }, { "epoch": 0.9367055154734533, "grad_norm": 1.0591968297958374, "learning_rate": 2.165982906398345e-07, "loss": 0.9587, "step": 26583 }, { "epoch": 0.9367407524864118, "grad_norm": 1.4682246446609497, "learning_rate": 2.1636210589490058e-07, "loss": 1.0555, "step": 26584 }, { "epoch": 0.9367759894993701, "grad_norm": 2.5431101322174072, "learning_rate": 2.1612604858309515e-07, "loss": 2.1427, "step": 26585 }, { "epoch": 0.9368112265123285, "grad_norm": 1.1450895071029663, "learning_rate": 2.1589011870749355e-07, "loss": 1.1221, "step": 26586 }, { "epoch": 0.936846463525287, "grad_norm": 4.445194721221924, "learning_rate": 2.1565431627116996e-07, "loss": 5.2893, "step": 26587 }, { "epoch": 0.9368817005382454, "grad_norm": 2.433140993118286, "learning_rate": 2.1541864127719635e-07, "loss": 1.9427, "step": 26588 }, { "epoch": 0.9369169375512038, "grad_norm": 3.8183555603027344, "learning_rate": 2.1518309372864032e-07, "loss": 2.863, "step": 26589 }, { "epoch": 0.9369521745641622, "grad_norm": 4.118359088897705, "learning_rate": 2.149476736285705e-07, "loss": 3.0757, "step": 26590 }, { "epoch": 0.9369874115771206, "grad_norm": 4.181549072265625, "learning_rate": 2.1471238098005442e-07, "loss": 2.828, "step": 26591 }, { "epoch": 0.937022648590079, "grad_norm": 2.962057113647461, "learning_rate": 2.1447721578615521e-07, "loss": 3.0064, "step": 26592 }, { "epoch": 0.9370578856030374, "grad_norm": 6.301125526428223, "learning_rate": 2.1424217804993818e-07, "loss": 3.542, "step": 26593 }, { "epoch": 0.9370931226159959, "grad_norm": 3.2177860736846924, "learning_rate": 2.14007267774462e-07, "loss": 2.7098, "step": 26594 }, { "epoch": 0.9371283596289542, "grad_norm": 1.1237945556640625, "learning_rate": 2.1377248496278757e-07, "loss": 1.1555, "step": 26595 }, { "epoch": 0.9371635966419126, "grad_norm": 0.8947893381118774, "learning_rate": 2.1353782961797465e-07, "loss": 0.6744, "step": 26596 }, { "epoch": 0.9371988336548711, "grad_norm": 3.6315486431121826, "learning_rate": 2.1330330174307635e-07, "loss": 2.9973, "step": 26597 }, { "epoch": 0.9372340706678295, "grad_norm": 2.8850650787353516, "learning_rate": 2.130689013411491e-07, "loss": 0.9953, "step": 26598 }, { "epoch": 0.9372693076807879, "grad_norm": 1.2220101356506348, "learning_rate": 2.1283462841524717e-07, "loss": 0.6851, "step": 26599 }, { "epoch": 0.9373045446937464, "grad_norm": 1.0482529401779175, "learning_rate": 2.1260048296842138e-07, "loss": 0.9392, "step": 26600 }, { "epoch": 0.9373397817067047, "grad_norm": 4.2296319007873535, "learning_rate": 2.1236646500371938e-07, "loss": 3.1105, "step": 26601 }, { "epoch": 0.9373750187196631, "grad_norm": 6.0419697761535645, "learning_rate": 2.121325745241909e-07, "loss": 2.1886, "step": 26602 }, { "epoch": 0.9374102557326216, "grad_norm": 3.494235038757324, "learning_rate": 2.118988115328846e-07, "loss": 2.8713, "step": 26603 }, { "epoch": 0.93744549274558, "grad_norm": 1.0292258262634277, "learning_rate": 2.116651760328414e-07, "loss": 0.7181, "step": 26604 }, { "epoch": 0.9374807297585384, "grad_norm": 3.4677116870880127, "learning_rate": 2.114316680271067e-07, "loss": 2.9889, "step": 26605 }, { "epoch": 0.9375159667714967, "grad_norm": 4.1901373863220215, "learning_rate": 2.111982875187213e-07, "loss": 2.5917, "step": 26606 }, { "epoch": 0.9375512037844552, "grad_norm": 5.149574279785156, "learning_rate": 2.1096503451072392e-07, "loss": 3.3096, "step": 26607 }, { "epoch": 0.9375864407974136, "grad_norm": 2.037127733230591, "learning_rate": 2.1073190900615547e-07, "loss": 1.0675, "step": 26608 }, { "epoch": 0.937621677810372, "grad_norm": 1.2813482284545898, "learning_rate": 2.104989110080502e-07, "loss": 0.765, "step": 26609 }, { "epoch": 0.9376569148233305, "grad_norm": 3.001929998397827, "learning_rate": 2.1026604051944343e-07, "loss": 1.1159, "step": 26610 }, { "epoch": 0.9376921518362888, "grad_norm": 2.482948064804077, "learning_rate": 2.1003329754336833e-07, "loss": 2.2628, "step": 26611 }, { "epoch": 0.9377273888492472, "grad_norm": 3.090139627456665, "learning_rate": 2.0980068208285798e-07, "loss": 3.0783, "step": 26612 }, { "epoch": 0.9377626258622057, "grad_norm": 2.4897682666778564, "learning_rate": 2.0956819414093887e-07, "loss": 2.8157, "step": 26613 }, { "epoch": 0.9377978628751641, "grad_norm": 4.5591721534729, "learning_rate": 2.093358337206419e-07, "loss": 3.331, "step": 26614 }, { "epoch": 0.9378330998881225, "grad_norm": 2.890479803085327, "learning_rate": 2.091036008249947e-07, "loss": 3.0978, "step": 26615 }, { "epoch": 0.9378683369010808, "grad_norm": 1.0108160972595215, "learning_rate": 2.0887149545701812e-07, "loss": 0.7706, "step": 26616 }, { "epoch": 0.9379035739140393, "grad_norm": 4.302626609802246, "learning_rate": 2.0863951761973644e-07, "loss": 3.0552, "step": 26617 }, { "epoch": 0.9379388109269977, "grad_norm": 1.3347750902175903, "learning_rate": 2.084076673161739e-07, "loss": 0.7035, "step": 26618 }, { "epoch": 0.9379740479399561, "grad_norm": 0.9106476902961731, "learning_rate": 2.0817594454935032e-07, "loss": 0.8522, "step": 26619 }, { "epoch": 0.9380092849529146, "grad_norm": 1.2380770444869995, "learning_rate": 2.079443493222799e-07, "loss": 0.8354, "step": 26620 }, { "epoch": 0.938044521965873, "grad_norm": 1.2928378582000732, "learning_rate": 2.0771288163798142e-07, "loss": 1.0819, "step": 26621 }, { "epoch": 0.9380797589788313, "grad_norm": 3.2563045024871826, "learning_rate": 2.0748154149947241e-07, "loss": 2.9143, "step": 26622 }, { "epoch": 0.9381149959917898, "grad_norm": 1.4053065776824951, "learning_rate": 2.072503289097616e-07, "loss": 0.9638, "step": 26623 }, { "epoch": 0.9381502330047482, "grad_norm": 5.240471839904785, "learning_rate": 2.0701924387186323e-07, "loss": 3.9331, "step": 26624 }, { "epoch": 0.9381854700177066, "grad_norm": 5.216613292694092, "learning_rate": 2.067882863887871e-07, "loss": 4.4208, "step": 26625 }, { "epoch": 0.938220707030665, "grad_norm": 4.99303674697876, "learning_rate": 2.0655745646353976e-07, "loss": 5.558, "step": 26626 }, { "epoch": 0.9382559440436234, "grad_norm": 6.323467254638672, "learning_rate": 2.0632675409912983e-07, "loss": 3.0098, "step": 26627 }, { "epoch": 0.9382911810565818, "grad_norm": 1.1706960201263428, "learning_rate": 2.0609617929856163e-07, "loss": 0.9266, "step": 26628 }, { "epoch": 0.9383264180695402, "grad_norm": 1.0658419132232666, "learning_rate": 2.058657320648383e-07, "loss": 0.8286, "step": 26629 }, { "epoch": 0.9383616550824987, "grad_norm": 3.0598959922790527, "learning_rate": 2.0563541240096074e-07, "loss": 2.2149, "step": 26630 }, { "epoch": 0.9383968920954571, "grad_norm": 0.9017871022224426, "learning_rate": 2.0540522030993103e-07, "loss": 0.8577, "step": 26631 }, { "epoch": 0.9384321291084154, "grad_norm": 6.118996620178223, "learning_rate": 2.0517515579474456e-07, "loss": 2.6184, "step": 26632 }, { "epoch": 0.9384673661213739, "grad_norm": 2.5466341972351074, "learning_rate": 2.0494521885839892e-07, "loss": 0.7613, "step": 26633 }, { "epoch": 0.9385026031343323, "grad_norm": 3.635934352874756, "learning_rate": 2.0471540950389057e-07, "loss": 2.9804, "step": 26634 }, { "epoch": 0.9385378401472907, "grad_norm": 4.949121475219727, "learning_rate": 2.044857277342116e-07, "loss": 5.4035, "step": 26635 }, { "epoch": 0.9385730771602492, "grad_norm": 3.304560422897339, "learning_rate": 2.0425617355235296e-07, "loss": 2.529, "step": 26636 }, { "epoch": 0.9386083141732076, "grad_norm": 1.0843919515609741, "learning_rate": 2.0402674696130665e-07, "loss": 0.7462, "step": 26637 }, { "epoch": 0.9386435511861659, "grad_norm": 4.404601097106934, "learning_rate": 2.037974479640592e-07, "loss": 4.286, "step": 26638 }, { "epoch": 0.9386787881991243, "grad_norm": 3.9401822090148926, "learning_rate": 2.0356827656359824e-07, "loss": 2.6313, "step": 26639 }, { "epoch": 0.9387140252120828, "grad_norm": 1.1533584594726562, "learning_rate": 2.033392327629069e-07, "loss": 0.9212, "step": 26640 }, { "epoch": 0.9387492622250412, "grad_norm": 2.6139721870422363, "learning_rate": 2.0311031656497172e-07, "loss": 2.7072, "step": 26641 }, { "epoch": 0.9387844992379996, "grad_norm": 3.653526544570923, "learning_rate": 2.028815279727725e-07, "loss": 3.3013, "step": 26642 }, { "epoch": 0.938819736250958, "grad_norm": 3.971165180206299, "learning_rate": 2.02652866989288e-07, "loss": 3.2183, "step": 26643 }, { "epoch": 0.9388549732639164, "grad_norm": 1.8353407382965088, "learning_rate": 2.0242433361749915e-07, "loss": 0.8659, "step": 26644 }, { "epoch": 0.9388902102768748, "grad_norm": 1.254732608795166, "learning_rate": 2.0219592786038022e-07, "loss": 0.909, "step": 26645 }, { "epoch": 0.9389254472898333, "grad_norm": 0.9215031266212463, "learning_rate": 2.0196764972090778e-07, "loss": 0.7385, "step": 26646 }, { "epoch": 0.9389606843027917, "grad_norm": 3.931896448135376, "learning_rate": 2.0173949920205493e-07, "loss": 2.5562, "step": 26647 }, { "epoch": 0.93899592131575, "grad_norm": 0.7774673700332642, "learning_rate": 2.0151147630679268e-07, "loss": 0.6531, "step": 26648 }, { "epoch": 0.9390311583287084, "grad_norm": 35.42900085449219, "learning_rate": 2.012835810380931e-07, "loss": 3.0834, "step": 26649 }, { "epoch": 0.9390663953416669, "grad_norm": 3.7425804138183594, "learning_rate": 2.0105581339892267e-07, "loss": 2.8821, "step": 26650 }, { "epoch": 0.9391016323546253, "grad_norm": 3.0362915992736816, "learning_rate": 2.0082817339224682e-07, "loss": 2.8503, "step": 26651 }, { "epoch": 0.9391368693675837, "grad_norm": 4.279297351837158, "learning_rate": 2.0060066102103315e-07, "loss": 2.902, "step": 26652 }, { "epoch": 0.9391721063805422, "grad_norm": 1.0554509162902832, "learning_rate": 2.0037327628824376e-07, "loss": 0.7923, "step": 26653 }, { "epoch": 0.9392073433935005, "grad_norm": 4.1966142654418945, "learning_rate": 2.0014601919684184e-07, "loss": 2.7932, "step": 26654 }, { "epoch": 0.9392425804064589, "grad_norm": 1.271126627922058, "learning_rate": 1.9991888974978501e-07, "loss": 0.9632, "step": 26655 }, { "epoch": 0.9392778174194174, "grad_norm": 1.3892006874084473, "learning_rate": 1.9969188795003313e-07, "loss": 0.763, "step": 26656 }, { "epoch": 0.9393130544323758, "grad_norm": 3.7714409828186035, "learning_rate": 1.9946501380054496e-07, "loss": 2.9771, "step": 26657 }, { "epoch": 0.9393482914453342, "grad_norm": 0.8889000415802002, "learning_rate": 1.9923826730427143e-07, "loss": 0.7058, "step": 26658 }, { "epoch": 0.9393835284582925, "grad_norm": 1.8257173299789429, "learning_rate": 1.9901164846416688e-07, "loss": 0.731, "step": 26659 }, { "epoch": 0.939418765471251, "grad_norm": 4.5092244148254395, "learning_rate": 1.987851572831856e-07, "loss": 5.0415, "step": 26660 }, { "epoch": 0.9394540024842094, "grad_norm": 3.6396422386169434, "learning_rate": 1.9855879376427633e-07, "loss": 2.7005, "step": 26661 }, { "epoch": 0.9394892394971678, "grad_norm": 4.809166431427002, "learning_rate": 1.9833255791038674e-07, "loss": 5.4355, "step": 26662 }, { "epoch": 0.9395244765101263, "grad_norm": 1.011520504951477, "learning_rate": 1.9810644972446448e-07, "loss": 0.9313, "step": 26663 }, { "epoch": 0.9395597135230846, "grad_norm": 4.690775394439697, "learning_rate": 1.978804692094538e-07, "loss": 2.5924, "step": 26664 }, { "epoch": 0.939594950536043, "grad_norm": 0.9103773236274719, "learning_rate": 1.9765461636829796e-07, "loss": 0.7949, "step": 26665 }, { "epoch": 0.9396301875490015, "grad_norm": 4.14040994644165, "learning_rate": 1.9742889120394016e-07, "loss": 3.2837, "step": 26666 }, { "epoch": 0.9396654245619599, "grad_norm": 1.3767043352127075, "learning_rate": 1.9720329371932023e-07, "loss": 0.8911, "step": 26667 }, { "epoch": 0.9397006615749183, "grad_norm": 1.3757376670837402, "learning_rate": 1.9697782391737474e-07, "loss": 0.7735, "step": 26668 }, { "epoch": 0.9397358985878767, "grad_norm": 4.483356952667236, "learning_rate": 1.9675248180104357e-07, "loss": 3.1144, "step": 26669 }, { "epoch": 0.9397711356008351, "grad_norm": 5.6960248947143555, "learning_rate": 1.965272673732588e-07, "loss": 4.7229, "step": 26670 }, { "epoch": 0.9398063726137935, "grad_norm": 6.366726875305176, "learning_rate": 1.9630218063695472e-07, "loss": 4.9131, "step": 26671 }, { "epoch": 0.9398416096267519, "grad_norm": 0.9440346956253052, "learning_rate": 1.960772215950646e-07, "loss": 0.8262, "step": 26672 }, { "epoch": 0.9398768466397104, "grad_norm": 3.310032844543457, "learning_rate": 1.9585239025051718e-07, "loss": 2.393, "step": 26673 }, { "epoch": 0.9399120836526688, "grad_norm": 3.0321505069732666, "learning_rate": 1.956276866062412e-07, "loss": 2.7604, "step": 26674 }, { "epoch": 0.9399473206656271, "grad_norm": 1.1081033945083618, "learning_rate": 1.9540311066516327e-07, "loss": 0.7683, "step": 26675 }, { "epoch": 0.9399825576785856, "grad_norm": 2.29656720161438, "learning_rate": 1.9517866243020877e-07, "loss": 2.556, "step": 26676 }, { "epoch": 0.940017794691544, "grad_norm": 0.9284762144088745, "learning_rate": 1.9495434190430207e-07, "loss": 0.6291, "step": 26677 }, { "epoch": 0.9400530317045024, "grad_norm": 0.9946364164352417, "learning_rate": 1.947301490903619e-07, "loss": 0.9438, "step": 26678 }, { "epoch": 0.9400882687174609, "grad_norm": 6.088166236877441, "learning_rate": 1.9450608399131267e-07, "loss": 2.8438, "step": 26679 }, { "epoch": 0.9401235057304193, "grad_norm": 5.496486186981201, "learning_rate": 1.9428214661006973e-07, "loss": 2.9031, "step": 26680 }, { "epoch": 0.9401587427433776, "grad_norm": 5.149637699127197, "learning_rate": 1.9405833694955078e-07, "loss": 4.4755, "step": 26681 }, { "epoch": 0.940193979756336, "grad_norm": 1.2225868701934814, "learning_rate": 1.938346550126713e-07, "loss": 0.9324, "step": 26682 }, { "epoch": 0.9402292167692945, "grad_norm": 1.3783540725708008, "learning_rate": 1.9361110080234446e-07, "loss": 0.6769, "step": 26683 }, { "epoch": 0.9402644537822529, "grad_norm": 1.237610936164856, "learning_rate": 1.9338767432148243e-07, "loss": 0.7669, "step": 26684 }, { "epoch": 0.9402996907952113, "grad_norm": 1.207403540611267, "learning_rate": 1.9316437557299506e-07, "loss": 0.8929, "step": 26685 }, { "epoch": 0.9403349278081697, "grad_norm": 4.050996780395508, "learning_rate": 1.9294120455979116e-07, "loss": 4.277, "step": 26686 }, { "epoch": 0.9403701648211281, "grad_norm": 4.360967636108398, "learning_rate": 1.9271816128477615e-07, "loss": 3.5778, "step": 26687 }, { "epoch": 0.9404054018340865, "grad_norm": 1.0956007242202759, "learning_rate": 1.9249524575085778e-07, "loss": 0.7982, "step": 26688 }, { "epoch": 0.940440638847045, "grad_norm": 1.754650354385376, "learning_rate": 1.9227245796093697e-07, "loss": 0.9038, "step": 26689 }, { "epoch": 0.9404758758600034, "grad_norm": 2.089437246322632, "learning_rate": 1.92049797917917e-07, "loss": 2.66, "step": 26690 }, { "epoch": 0.9405111128729617, "grad_norm": 3.142266035079956, "learning_rate": 1.9182726562469778e-07, "loss": 2.9049, "step": 26691 }, { "epoch": 0.9405463498859201, "grad_norm": 6.258325576782227, "learning_rate": 1.916048610841792e-07, "loss": 2.6235, "step": 26692 }, { "epoch": 0.9405815868988786, "grad_norm": 1.3041348457336426, "learning_rate": 1.913825842992545e-07, "loss": 0.916, "step": 26693 }, { "epoch": 0.940616823911837, "grad_norm": 2.759255886077881, "learning_rate": 1.9116043527282136e-07, "loss": 2.3217, "step": 26694 }, { "epoch": 0.9406520609247954, "grad_norm": 5.53751277923584, "learning_rate": 1.9093841400777413e-07, "loss": 5.496, "step": 26695 }, { "epoch": 0.9406872979377539, "grad_norm": 1.0468180179595947, "learning_rate": 1.9071652050700383e-07, "loss": 1.2097, "step": 26696 }, { "epoch": 0.9407225349507122, "grad_norm": 2.355862855911255, "learning_rate": 1.9049475477339817e-07, "loss": 0.9446, "step": 26697 }, { "epoch": 0.9407577719636706, "grad_norm": 1.194506049156189, "learning_rate": 1.902731168098504e-07, "loss": 0.9262, "step": 26698 }, { "epoch": 0.9407930089766291, "grad_norm": 3.665440320968628, "learning_rate": 1.900516066192437e-07, "loss": 2.9993, "step": 26699 }, { "epoch": 0.9408282459895875, "grad_norm": 4.483719825744629, "learning_rate": 1.898302242044636e-07, "loss": 3.1438, "step": 26700 }, { "epoch": 0.9408634830025459, "grad_norm": 2.1940088272094727, "learning_rate": 1.896089695683956e-07, "loss": 0.858, "step": 26701 }, { "epoch": 0.9408987200155042, "grad_norm": 4.753525257110596, "learning_rate": 1.8938784271391953e-07, "loss": 2.9749, "step": 26702 }, { "epoch": 0.9409339570284627, "grad_norm": 1.1295783519744873, "learning_rate": 1.8916684364391757e-07, "loss": 1.0749, "step": 26703 }, { "epoch": 0.9409691940414211, "grad_norm": 6.464869022369385, "learning_rate": 1.8894597236126634e-07, "loss": 4.4602, "step": 26704 }, { "epoch": 0.9410044310543795, "grad_norm": 1.0655187368392944, "learning_rate": 1.887252288688435e-07, "loss": 0.7492, "step": 26705 }, { "epoch": 0.941039668067338, "grad_norm": 0.8900237083435059, "learning_rate": 1.8850461316952452e-07, "loss": 0.9367, "step": 26706 }, { "epoch": 0.9410749050802963, "grad_norm": 0.8836575746536255, "learning_rate": 1.8828412526618267e-07, "loss": 0.857, "step": 26707 }, { "epoch": 0.9411101420932547, "grad_norm": 2.7665233612060547, "learning_rate": 1.8806376516169012e-07, "loss": 0.9374, "step": 26708 }, { "epoch": 0.9411453791062132, "grad_norm": 1.044204831123352, "learning_rate": 1.8784353285891676e-07, "loss": 0.7948, "step": 26709 }, { "epoch": 0.9411806161191716, "grad_norm": 1.0057604312896729, "learning_rate": 1.8762342836073143e-07, "loss": 0.8608, "step": 26710 }, { "epoch": 0.94121585313213, "grad_norm": 2.755692481994629, "learning_rate": 1.8740345167000184e-07, "loss": 0.7251, "step": 26711 }, { "epoch": 0.9412510901450885, "grad_norm": 1.3599501848220825, "learning_rate": 1.8718360278959125e-07, "loss": 1.0639, "step": 26712 }, { "epoch": 0.9412863271580468, "grad_norm": 3.356423854827881, "learning_rate": 1.869638817223629e-07, "loss": 2.3247, "step": 26713 }, { "epoch": 0.9413215641710052, "grad_norm": 3.30578351020813, "learning_rate": 1.867442884711812e-07, "loss": 2.3521, "step": 26714 }, { "epoch": 0.9413568011839636, "grad_norm": 7.075011730194092, "learning_rate": 1.8652482303890497e-07, "loss": 5.6763, "step": 26715 }, { "epoch": 0.9413920381969221, "grad_norm": 4.005268573760986, "learning_rate": 1.8630548542839299e-07, "loss": 2.9827, "step": 26716 }, { "epoch": 0.9414272752098805, "grad_norm": 3.315668821334839, "learning_rate": 1.8608627564250303e-07, "loss": 3.0702, "step": 26717 }, { "epoch": 0.9414625122228388, "grad_norm": 1.2948802709579468, "learning_rate": 1.8586719368408834e-07, "loss": 0.8938, "step": 26718 }, { "epoch": 0.9414977492357973, "grad_norm": 1.0717699527740479, "learning_rate": 1.856482395560033e-07, "loss": 0.9512, "step": 26719 }, { "epoch": 0.9415329862487557, "grad_norm": 8.02423095703125, "learning_rate": 1.854294132611012e-07, "loss": 5.8665, "step": 26720 }, { "epoch": 0.9415682232617141, "grad_norm": 0.9650877714157104, "learning_rate": 1.8521071480222975e-07, "loss": 0.7573, "step": 26721 }, { "epoch": 0.9416034602746726, "grad_norm": 5.989099502563477, "learning_rate": 1.8499214418224e-07, "loss": 5.7245, "step": 26722 }, { "epoch": 0.9416386972876309, "grad_norm": 3.2757630348205566, "learning_rate": 1.8477370140397853e-07, "loss": 2.7434, "step": 26723 }, { "epoch": 0.9416739343005893, "grad_norm": 2.153764247894287, "learning_rate": 1.845553864702876e-07, "loss": 0.8661, "step": 26724 }, { "epoch": 0.9417091713135477, "grad_norm": 2.0070488452911377, "learning_rate": 1.8433719938401485e-07, "loss": 0.7459, "step": 26725 }, { "epoch": 0.9417444083265062, "grad_norm": 1.0647611618041992, "learning_rate": 1.8411914014799914e-07, "loss": 0.828, "step": 26726 }, { "epoch": 0.9417796453394646, "grad_norm": 3.5053508281707764, "learning_rate": 1.8390120876508377e-07, "loss": 4.5162, "step": 26727 }, { "epoch": 0.9418148823524229, "grad_norm": 3.7677314281463623, "learning_rate": 1.8368340523810425e-07, "loss": 4.9804, "step": 26728 }, { "epoch": 0.9418501193653814, "grad_norm": 3.529371976852417, "learning_rate": 1.8346572956989828e-07, "loss": 3.226, "step": 26729 }, { "epoch": 0.9418853563783398, "grad_norm": 1.2123119831085205, "learning_rate": 1.832481817633036e-07, "loss": 0.7468, "step": 26730 }, { "epoch": 0.9419205933912982, "grad_norm": 3.2734220027923584, "learning_rate": 1.8303076182115021e-07, "loss": 2.6589, "step": 26731 }, { "epoch": 0.9419558304042567, "grad_norm": 4.9473795890808105, "learning_rate": 1.8281346974627135e-07, "loss": 3.5891, "step": 26732 }, { "epoch": 0.9419910674172151, "grad_norm": 1.0333153009414673, "learning_rate": 1.8259630554149921e-07, "loss": 0.8455, "step": 26733 }, { "epoch": 0.9420263044301734, "grad_norm": 3.670689105987549, "learning_rate": 1.823792692096593e-07, "loss": 2.5829, "step": 26734 }, { "epoch": 0.9420615414431318, "grad_norm": 2.93513560295105, "learning_rate": 1.821623607535805e-07, "loss": 0.702, "step": 26735 }, { "epoch": 0.9420967784560903, "grad_norm": 4.566567420959473, "learning_rate": 1.819455801760872e-07, "loss": 4.9102, "step": 26736 }, { "epoch": 0.9421320154690487, "grad_norm": 4.6021223068237305, "learning_rate": 1.817289274800038e-07, "loss": 3.6373, "step": 26737 }, { "epoch": 0.9421672524820071, "grad_norm": 3.0424020290374756, "learning_rate": 1.815124026681514e-07, "loss": 2.5052, "step": 26738 }, { "epoch": 0.9422024894949655, "grad_norm": 1.1514769792556763, "learning_rate": 1.8129600574334992e-07, "loss": 1.0016, "step": 26739 }, { "epoch": 0.9422377265079239, "grad_norm": 4.095709323883057, "learning_rate": 1.810797367084194e-07, "loss": 2.849, "step": 26740 }, { "epoch": 0.9422729635208823, "grad_norm": 1.0711109638214111, "learning_rate": 1.808635955661764e-07, "loss": 0.855, "step": 26741 }, { "epoch": 0.9423082005338408, "grad_norm": 1.505682110786438, "learning_rate": 1.8064758231943646e-07, "loss": 0.7457, "step": 26742 }, { "epoch": 0.9423434375467992, "grad_norm": 0.9478645920753479, "learning_rate": 1.8043169697101072e-07, "loss": 1.0053, "step": 26743 }, { "epoch": 0.9423786745597575, "grad_norm": 3.7378361225128174, "learning_rate": 1.8021593952371353e-07, "loss": 3.2588, "step": 26744 }, { "epoch": 0.942413911572716, "grad_norm": 3.3429152965545654, "learning_rate": 1.8000030998035488e-07, "loss": 2.6952, "step": 26745 }, { "epoch": 0.9424491485856744, "grad_norm": 1.8586004972457886, "learning_rate": 1.797848083437448e-07, "loss": 0.7185, "step": 26746 }, { "epoch": 0.9424843855986328, "grad_norm": 7.583657264709473, "learning_rate": 1.795694346166854e-07, "loss": 4.6071, "step": 26747 }, { "epoch": 0.9425196226115912, "grad_norm": 4.384374618530273, "learning_rate": 1.7935418880198673e-07, "loss": 5.0211, "step": 26748 }, { "epoch": 0.9425548596245497, "grad_norm": 4.772475242614746, "learning_rate": 1.7913907090245097e-07, "loss": 2.8691, "step": 26749 }, { "epoch": 0.942590096637508, "grad_norm": 3.4757370948791504, "learning_rate": 1.789240809208792e-07, "loss": 2.9104, "step": 26750 }, { "epoch": 0.9426253336504664, "grad_norm": 4.678930282592773, "learning_rate": 1.7870921886007142e-07, "loss": 5.8606, "step": 26751 }, { "epoch": 0.9426605706634249, "grad_norm": 5.261242389678955, "learning_rate": 1.7849448472282871e-07, "loss": 5.2567, "step": 26752 }, { "epoch": 0.9426958076763833, "grad_norm": 3.250546932220459, "learning_rate": 1.782798785119455e-07, "loss": 2.6112, "step": 26753 }, { "epoch": 0.9427310446893417, "grad_norm": 1.0355747938156128, "learning_rate": 1.7806540023021844e-07, "loss": 0.6711, "step": 26754 }, { "epoch": 0.9427662817023001, "grad_norm": 4.514132022857666, "learning_rate": 1.7785104988043978e-07, "loss": 2.9346, "step": 26755 }, { "epoch": 0.9428015187152585, "grad_norm": 1.0406055450439453, "learning_rate": 1.7763682746540278e-07, "loss": 0.8863, "step": 26756 }, { "epoch": 0.9428367557282169, "grad_norm": 4.30338191986084, "learning_rate": 1.7742273298789748e-07, "loss": 3.1938, "step": 26757 }, { "epoch": 0.9428719927411753, "grad_norm": 0.9872406721115112, "learning_rate": 1.772087664507116e-07, "loss": 1.2368, "step": 26758 }, { "epoch": 0.9429072297541338, "grad_norm": 3.288949966430664, "learning_rate": 1.7699492785663407e-07, "loss": 2.6289, "step": 26759 }, { "epoch": 0.9429424667670921, "grad_norm": 1.3611087799072266, "learning_rate": 1.767812172084482e-07, "loss": 1.1842, "step": 26760 }, { "epoch": 0.9429777037800505, "grad_norm": 4.176990985870361, "learning_rate": 1.7656763450893844e-07, "loss": 4.1211, "step": 26761 }, { "epoch": 0.943012940793009, "grad_norm": 4.470580101013184, "learning_rate": 1.7635417976088586e-07, "loss": 4.679, "step": 26762 }, { "epoch": 0.9430481778059674, "grad_norm": 3.9890801906585693, "learning_rate": 1.761408529670716e-07, "loss": 2.8082, "step": 26763 }, { "epoch": 0.9430834148189258, "grad_norm": 4.075685501098633, "learning_rate": 1.7592765413027457e-07, "loss": 3.3694, "step": 26764 }, { "epoch": 0.9431186518318843, "grad_norm": 1.433843970298767, "learning_rate": 1.757145832532714e-07, "loss": 0.879, "step": 26765 }, { "epoch": 0.9431538888448426, "grad_norm": 4.762974739074707, "learning_rate": 1.7550164033883544e-07, "loss": 7.2046, "step": 26766 }, { "epoch": 0.943189125857801, "grad_norm": 6.368178844451904, "learning_rate": 1.7528882538974445e-07, "loss": 4.8201, "step": 26767 }, { "epoch": 0.9432243628707594, "grad_norm": 1.0765013694763184, "learning_rate": 1.7507613840876735e-07, "loss": 0.7278, "step": 26768 }, { "epoch": 0.9432595998837179, "grad_norm": 1.6392261981964111, "learning_rate": 1.7486357939867417e-07, "loss": 0.9208, "step": 26769 }, { "epoch": 0.9432948368966763, "grad_norm": 0.9562195539474487, "learning_rate": 1.7465114836223375e-07, "loss": 0.9082, "step": 26770 }, { "epoch": 0.9433300739096346, "grad_norm": 5.024709224700928, "learning_rate": 1.744388453022161e-07, "loss": 4.8796, "step": 26771 }, { "epoch": 0.9433653109225931, "grad_norm": 3.591852903366089, "learning_rate": 1.742266702213824e-07, "loss": 2.9046, "step": 26772 }, { "epoch": 0.9434005479355515, "grad_norm": 4.237555503845215, "learning_rate": 1.7401462312249817e-07, "loss": 3.2665, "step": 26773 }, { "epoch": 0.9434357849485099, "grad_norm": 1.2200533151626587, "learning_rate": 1.7380270400832567e-07, "loss": 0.7844, "step": 26774 }, { "epoch": 0.9434710219614684, "grad_norm": 4.250877857208252, "learning_rate": 1.7359091288162377e-07, "loss": 6.4313, "step": 26775 }, { "epoch": 0.9435062589744267, "grad_norm": 1.197316288948059, "learning_rate": 1.7337924974515252e-07, "loss": 1.2934, "step": 26776 }, { "epoch": 0.9435414959873851, "grad_norm": 3.078648567199707, "learning_rate": 1.731677146016675e-07, "loss": 2.457, "step": 26777 }, { "epoch": 0.9435767330003436, "grad_norm": 0.9096907377243042, "learning_rate": 1.7295630745392533e-07, "loss": 0.8503, "step": 26778 }, { "epoch": 0.943611970013302, "grad_norm": 1.3108354806900024, "learning_rate": 1.7274502830467942e-07, "loss": 0.7733, "step": 26779 }, { "epoch": 0.9436472070262604, "grad_norm": 1.0596789121627808, "learning_rate": 1.725338771566809e-07, "loss": 0.7423, "step": 26780 }, { "epoch": 0.9436824440392187, "grad_norm": 4.4622483253479, "learning_rate": 1.723228540126798e-07, "loss": 4.6453, "step": 26781 }, { "epoch": 0.9437176810521772, "grad_norm": 4.0182108879089355, "learning_rate": 1.7211195887542608e-07, "loss": 3.0538, "step": 26782 }, { "epoch": 0.9437529180651356, "grad_norm": 4.735992431640625, "learning_rate": 1.719011917476654e-07, "loss": 3.2397, "step": 26783 }, { "epoch": 0.943788155078094, "grad_norm": 1.0823668241500854, "learning_rate": 1.7169055263214552e-07, "loss": 0.6945, "step": 26784 }, { "epoch": 0.9438233920910525, "grad_norm": 1.9165830612182617, "learning_rate": 1.7148004153160536e-07, "loss": 1.0293, "step": 26785 }, { "epoch": 0.9438586291040109, "grad_norm": 3.7201366424560547, "learning_rate": 1.712696584487905e-07, "loss": 3.0479, "step": 26786 }, { "epoch": 0.9438938661169692, "grad_norm": 3.759408473968506, "learning_rate": 1.710594033864421e-07, "loss": 3.154, "step": 26787 }, { "epoch": 0.9439291031299277, "grad_norm": 3.5292482376098633, "learning_rate": 1.7084927634729576e-07, "loss": 2.7255, "step": 26788 }, { "epoch": 0.9439643401428861, "grad_norm": 4.541609287261963, "learning_rate": 1.7063927733408813e-07, "loss": 2.3284, "step": 26789 }, { "epoch": 0.9439995771558445, "grad_norm": 1.986689805984497, "learning_rate": 1.7042940634955818e-07, "loss": 1.0906, "step": 26790 }, { "epoch": 0.9440348141688029, "grad_norm": 3.9553351402282715, "learning_rate": 1.7021966339643592e-07, "loss": 3.1751, "step": 26791 }, { "epoch": 0.9440700511817613, "grad_norm": 4.585400104522705, "learning_rate": 1.7001004847745584e-07, "loss": 2.2491, "step": 26792 }, { "epoch": 0.9441052881947197, "grad_norm": 1.1760953664779663, "learning_rate": 1.6980056159534576e-07, "loss": 0.8987, "step": 26793 }, { "epoch": 0.9441405252076781, "grad_norm": 1.1100133657455444, "learning_rate": 1.6959120275283681e-07, "loss": 0.8498, "step": 26794 }, { "epoch": 0.9441757622206366, "grad_norm": 3.228964328765869, "learning_rate": 1.6938197195265348e-07, "loss": 2.895, "step": 26795 }, { "epoch": 0.944210999233595, "grad_norm": 1.0763523578643799, "learning_rate": 1.6917286919752253e-07, "loss": 0.8972, "step": 26796 }, { "epoch": 0.9442462362465533, "grad_norm": 2.3603157997131348, "learning_rate": 1.6896389449016726e-07, "loss": 2.5482, "step": 26797 }, { "epoch": 0.9442814732595118, "grad_norm": 3.0650806427001953, "learning_rate": 1.6875504783330997e-07, "loss": 2.7969, "step": 26798 }, { "epoch": 0.9443167102724702, "grad_norm": 6.482948303222656, "learning_rate": 1.6854632922966963e-07, "loss": 7.2, "step": 26799 }, { "epoch": 0.9443519472854286, "grad_norm": 1.230808138847351, "learning_rate": 1.6833773868196622e-07, "loss": 0.9393, "step": 26800 }, { "epoch": 0.944387184298387, "grad_norm": 4.451800346374512, "learning_rate": 1.6812927619291652e-07, "loss": 2.8132, "step": 26801 }, { "epoch": 0.9444224213113455, "grad_norm": 2.127265691757202, "learning_rate": 1.6792094176523499e-07, "loss": 0.9451, "step": 26802 }, { "epoch": 0.9444576583243038, "grad_norm": 3.5978169441223145, "learning_rate": 1.6771273540163614e-07, "loss": 3.26, "step": 26803 }, { "epoch": 0.9444928953372622, "grad_norm": 3.801729679107666, "learning_rate": 1.6750465710483e-07, "loss": 3.2872, "step": 26804 }, { "epoch": 0.9445281323502207, "grad_norm": 3.686540126800537, "learning_rate": 1.6729670687752887e-07, "loss": 4.4831, "step": 26805 }, { "epoch": 0.9445633693631791, "grad_norm": 7.1597981452941895, "learning_rate": 1.6708888472244057e-07, "loss": 4.9956, "step": 26806 }, { "epoch": 0.9445986063761375, "grad_norm": 5.778210639953613, "learning_rate": 1.6688119064227292e-07, "loss": 5.0901, "step": 26807 }, { "epoch": 0.944633843389096, "grad_norm": 1.0447900295257568, "learning_rate": 1.6667362463972824e-07, "loss": 0.7763, "step": 26808 }, { "epoch": 0.9446690804020543, "grad_norm": 4.154519081115723, "learning_rate": 1.6646618671751435e-07, "loss": 3.9998, "step": 26809 }, { "epoch": 0.9447043174150127, "grad_norm": 5.884342670440674, "learning_rate": 1.662588768783302e-07, "loss": 3.2626, "step": 26810 }, { "epoch": 0.9447395544279712, "grad_norm": 2.8922858238220215, "learning_rate": 1.6605169512487694e-07, "loss": 0.7751, "step": 26811 }, { "epoch": 0.9447747914409296, "grad_norm": 1.2076200246810913, "learning_rate": 1.6584464145985247e-07, "loss": 0.9419, "step": 26812 }, { "epoch": 0.944810028453888, "grad_norm": 1.456897497177124, "learning_rate": 1.6563771588595458e-07, "loss": 0.6805, "step": 26813 }, { "epoch": 0.9448452654668463, "grad_norm": 3.9029691219329834, "learning_rate": 1.6543091840587777e-07, "loss": 3.0531, "step": 26814 }, { "epoch": 0.9448805024798048, "grad_norm": 4.170658588409424, "learning_rate": 1.652242490223166e-07, "loss": 2.7642, "step": 26815 }, { "epoch": 0.9449157394927632, "grad_norm": 0.9337142705917358, "learning_rate": 1.650177077379611e-07, "loss": 0.7571, "step": 26816 }, { "epoch": 0.9449509765057216, "grad_norm": 4.946322917938232, "learning_rate": 1.648112945555036e-07, "loss": 3.193, "step": 26817 }, { "epoch": 0.9449862135186801, "grad_norm": 3.888681411743164, "learning_rate": 1.6460500947763192e-07, "loss": 2.7382, "step": 26818 }, { "epoch": 0.9450214505316384, "grad_norm": 1.0295257568359375, "learning_rate": 1.643988525070328e-07, "loss": 0.6046, "step": 26819 }, { "epoch": 0.9450566875445968, "grad_norm": 7.139791965484619, "learning_rate": 1.641928236463908e-07, "loss": 5.4772, "step": 26820 }, { "epoch": 0.9450919245575553, "grad_norm": 4.414078712463379, "learning_rate": 1.6398692289839147e-07, "loss": 5.466, "step": 26821 }, { "epoch": 0.9451271615705137, "grad_norm": 1.235430359840393, "learning_rate": 1.6378115026571494e-07, "loss": 1.0859, "step": 26822 }, { "epoch": 0.9451623985834721, "grad_norm": 3.013906717300415, "learning_rate": 1.635755057510413e-07, "loss": 2.7945, "step": 26823 }, { "epoch": 0.9451976355964304, "grad_norm": 1.0316203832626343, "learning_rate": 1.633699893570495e-07, "loss": 1.0286, "step": 26824 }, { "epoch": 0.9452328726093889, "grad_norm": 1.0385215282440186, "learning_rate": 1.631646010864174e-07, "loss": 0.6751, "step": 26825 }, { "epoch": 0.9452681096223473, "grad_norm": 5.1888251304626465, "learning_rate": 1.6295934094181952e-07, "loss": 2.4486, "step": 26826 }, { "epoch": 0.9453033466353057, "grad_norm": 4.578778266906738, "learning_rate": 1.6275420892592818e-07, "loss": 4.8603, "step": 26827 }, { "epoch": 0.9453385836482642, "grad_norm": 1.9572240114212036, "learning_rate": 1.625492050414168e-07, "loss": 2.0745, "step": 26828 }, { "epoch": 0.9453738206612226, "grad_norm": 1.6148734092712402, "learning_rate": 1.6234432929095544e-07, "loss": 0.8425, "step": 26829 }, { "epoch": 0.9454090576741809, "grad_norm": 3.8063175678253174, "learning_rate": 1.62139581677212e-07, "loss": 4.7354, "step": 26830 }, { "epoch": 0.9454442946871394, "grad_norm": 1.027571678161621, "learning_rate": 1.6193496220285433e-07, "loss": 0.7294, "step": 26831 }, { "epoch": 0.9454795317000978, "grad_norm": 3.586550712585449, "learning_rate": 1.6173047087054583e-07, "loss": 2.7856, "step": 26832 }, { "epoch": 0.9455147687130562, "grad_norm": 1.6812137365341187, "learning_rate": 1.6152610768295108e-07, "loss": 0.8391, "step": 26833 }, { "epoch": 0.9455500057260146, "grad_norm": 1.14065682888031, "learning_rate": 1.6132187264273348e-07, "loss": 0.9983, "step": 26834 }, { "epoch": 0.945585242738973, "grad_norm": 1.6159061193466187, "learning_rate": 1.6111776575254978e-07, "loss": 0.9061, "step": 26835 }, { "epoch": 0.9456204797519314, "grad_norm": 1.2079246044158936, "learning_rate": 1.6091378701506234e-07, "loss": 0.7554, "step": 26836 }, { "epoch": 0.9456557167648898, "grad_norm": 1.1160197257995605, "learning_rate": 1.6070993643292454e-07, "loss": 0.7915, "step": 26837 }, { "epoch": 0.9456909537778483, "grad_norm": 1.0772061347961426, "learning_rate": 1.605062140087943e-07, "loss": 0.8028, "step": 26838 }, { "epoch": 0.9457261907908067, "grad_norm": 3.3418493270874023, "learning_rate": 1.603026197453239e-07, "loss": 3.2541, "step": 26839 }, { "epoch": 0.945761427803765, "grad_norm": 2.3332977294921875, "learning_rate": 1.600991536451646e-07, "loss": 2.7, "step": 26840 }, { "epoch": 0.9457966648167235, "grad_norm": 2.2707395553588867, "learning_rate": 1.5989581571096757e-07, "loss": 3.1173, "step": 26841 }, { "epoch": 0.9458319018296819, "grad_norm": 2.4675867557525635, "learning_rate": 1.5969260594538072e-07, "loss": 2.9717, "step": 26842 }, { "epoch": 0.9458671388426403, "grad_norm": 1.2607812881469727, "learning_rate": 1.594895243510508e-07, "loss": 0.8625, "step": 26843 }, { "epoch": 0.9459023758555988, "grad_norm": 3.6640634536743164, "learning_rate": 1.592865709306246e-07, "loss": 3.0539, "step": 26844 }, { "epoch": 0.9459376128685572, "grad_norm": 1.0011602640151978, "learning_rate": 1.5908374568674445e-07, "loss": 0.8109, "step": 26845 }, { "epoch": 0.9459728498815155, "grad_norm": 3.2552313804626465, "learning_rate": 1.5888104862205046e-07, "loss": 2.9062, "step": 26846 }, { "epoch": 0.9460080868944739, "grad_norm": 3.9582619667053223, "learning_rate": 1.5867847973918493e-07, "loss": 2.9542, "step": 26847 }, { "epoch": 0.9460433239074324, "grad_norm": 5.389246463775635, "learning_rate": 1.5847603904078578e-07, "loss": 6.8752, "step": 26848 }, { "epoch": 0.9460785609203908, "grad_norm": 1.078348994255066, "learning_rate": 1.582737265294909e-07, "loss": 0.8408, "step": 26849 }, { "epoch": 0.9461137979333492, "grad_norm": 4.591835021972656, "learning_rate": 1.5807154220793263e-07, "loss": 2.4987, "step": 26850 }, { "epoch": 0.9461490349463076, "grad_norm": 5.840904712677002, "learning_rate": 1.5786948607874774e-07, "loss": 4.9624, "step": 26851 }, { "epoch": 0.946184271959266, "grad_norm": 0.8380066156387329, "learning_rate": 1.5766755814456525e-07, "loss": 0.6613, "step": 26852 }, { "epoch": 0.9462195089722244, "grad_norm": 0.9376569390296936, "learning_rate": 1.5746575840801748e-07, "loss": 1.0473, "step": 26853 }, { "epoch": 0.9462547459851829, "grad_norm": 6.653339385986328, "learning_rate": 1.5726408687173123e-07, "loss": 5.1748, "step": 26854 }, { "epoch": 0.9462899829981413, "grad_norm": 1.7874670028686523, "learning_rate": 1.570625435383344e-07, "loss": 0.8483, "step": 26855 }, { "epoch": 0.9463252200110996, "grad_norm": 2.19844388961792, "learning_rate": 1.5686112841045153e-07, "loss": 2.6425, "step": 26856 }, { "epoch": 0.946360457024058, "grad_norm": 3.3588545322418213, "learning_rate": 1.5665984149070612e-07, "loss": 3.3826, "step": 26857 }, { "epoch": 0.9463956940370165, "grad_norm": 1.0924458503723145, "learning_rate": 1.5645868278172048e-07, "loss": 1.0344, "step": 26858 }, { "epoch": 0.9464309310499749, "grad_norm": 9.65378475189209, "learning_rate": 1.5625765228611368e-07, "loss": 4.2336, "step": 26859 }, { "epoch": 0.9464661680629333, "grad_norm": 3.560697078704834, "learning_rate": 1.5605675000650688e-07, "loss": 2.8257, "step": 26860 }, { "epoch": 0.9465014050758918, "grad_norm": 1.1067070960998535, "learning_rate": 1.558559759455125e-07, "loss": 0.9142, "step": 26861 }, { "epoch": 0.9465366420888501, "grad_norm": 2.602062463760376, "learning_rate": 1.5565533010574842e-07, "loss": 2.9166, "step": 26862 }, { "epoch": 0.9465718791018085, "grad_norm": 1.0503215789794922, "learning_rate": 1.554548124898292e-07, "loss": 1.0122, "step": 26863 }, { "epoch": 0.946607116114767, "grad_norm": 5.659666061401367, "learning_rate": 1.5525442310036387e-07, "loss": 2.8825, "step": 26864 }, { "epoch": 0.9466423531277254, "grad_norm": 5.478332042694092, "learning_rate": 1.5505416193996258e-07, "loss": 7.2046, "step": 26865 }, { "epoch": 0.9466775901406838, "grad_norm": 12.859966278076172, "learning_rate": 1.5485402901123657e-07, "loss": 4.8965, "step": 26866 }, { "epoch": 0.9467128271536421, "grad_norm": 7.428592205047607, "learning_rate": 1.5465402431679044e-07, "loss": 4.8777, "step": 26867 }, { "epoch": 0.9467480641666006, "grad_norm": 1.7809484004974365, "learning_rate": 1.5445414785922984e-07, "loss": 1.0606, "step": 26868 }, { "epoch": 0.946783301179559, "grad_norm": 1.105107307434082, "learning_rate": 1.542543996411583e-07, "loss": 0.8381, "step": 26869 }, { "epoch": 0.9468185381925174, "grad_norm": 1.2449250221252441, "learning_rate": 1.5405477966517702e-07, "loss": 1.2126, "step": 26870 }, { "epoch": 0.9468537752054759, "grad_norm": 3.455617904663086, "learning_rate": 1.5385528793388615e-07, "loss": 2.8674, "step": 26871 }, { "epoch": 0.9468890122184342, "grad_norm": 3.3334593772888184, "learning_rate": 1.5365592444988365e-07, "loss": 2.5762, "step": 26872 }, { "epoch": 0.9469242492313926, "grad_norm": 1.3264360427856445, "learning_rate": 1.534566892157685e-07, "loss": 0.8198, "step": 26873 }, { "epoch": 0.9469594862443511, "grad_norm": 1.261536955833435, "learning_rate": 1.5325758223413312e-07, "loss": 0.5598, "step": 26874 }, { "epoch": 0.9469947232573095, "grad_norm": 1.3161671161651611, "learning_rate": 1.5305860350757317e-07, "loss": 0.9829, "step": 26875 }, { "epoch": 0.9470299602702679, "grad_norm": 1.007946252822876, "learning_rate": 1.5285975303867773e-07, "loss": 0.9405, "step": 26876 }, { "epoch": 0.9470651972832262, "grad_norm": 0.8867018222808838, "learning_rate": 1.5266103083003913e-07, "loss": 0.8669, "step": 26877 }, { "epoch": 0.9471004342961847, "grad_norm": 0.9356883764266968, "learning_rate": 1.5246243688424533e-07, "loss": 0.8814, "step": 26878 }, { "epoch": 0.9471356713091431, "grad_norm": 1.0822314023971558, "learning_rate": 1.5226397120388315e-07, "loss": 1.1127, "step": 26879 }, { "epoch": 0.9471709083221015, "grad_norm": 2.812016010284424, "learning_rate": 1.5206563379153495e-07, "loss": 3.1014, "step": 26880 }, { "epoch": 0.94720614533506, "grad_norm": 5.030737400054932, "learning_rate": 1.5186742464978753e-07, "loss": 4.392, "step": 26881 }, { "epoch": 0.9472413823480184, "grad_norm": 5.247802257537842, "learning_rate": 1.5166934378122223e-07, "loss": 7.0233, "step": 26882 }, { "epoch": 0.9472766193609767, "grad_norm": 5.695435047149658, "learning_rate": 1.5147139118841801e-07, "loss": 4.5449, "step": 26883 }, { "epoch": 0.9473118563739352, "grad_norm": 1.291497826576233, "learning_rate": 1.5127356687395178e-07, "loss": 0.8246, "step": 26884 }, { "epoch": 0.9473470933868936, "grad_norm": 2.7853407859802246, "learning_rate": 1.5107587084040364e-07, "loss": 2.729, "step": 26885 }, { "epoch": 0.947382330399852, "grad_norm": 3.743027925491333, "learning_rate": 1.5087830309034601e-07, "loss": 4.3952, "step": 26886 }, { "epoch": 0.9474175674128105, "grad_norm": 3.1386306285858154, "learning_rate": 1.5068086362635347e-07, "loss": 2.6923, "step": 26887 }, { "epoch": 0.9474528044257688, "grad_norm": 4.661123275756836, "learning_rate": 1.504835524509962e-07, "loss": 4.3498, "step": 26888 }, { "epoch": 0.9474880414387272, "grad_norm": 3.548720598220825, "learning_rate": 1.502863695668466e-07, "loss": 4.8608, "step": 26889 }, { "epoch": 0.9475232784516856, "grad_norm": 3.488760232925415, "learning_rate": 1.500893149764704e-07, "loss": 2.2398, "step": 26890 }, { "epoch": 0.9475585154646441, "grad_norm": 4.694233417510986, "learning_rate": 1.498923886824366e-07, "loss": 4.7331, "step": 26891 }, { "epoch": 0.9475937524776025, "grad_norm": 1.654280424118042, "learning_rate": 1.4969559068730878e-07, "loss": 0.726, "step": 26892 }, { "epoch": 0.9476289894905608, "grad_norm": 2.9036457538604736, "learning_rate": 1.494989209936504e-07, "loss": 2.345, "step": 26893 }, { "epoch": 0.9476642265035193, "grad_norm": 0.8653907179832458, "learning_rate": 1.4930237960402382e-07, "loss": 0.6992, "step": 26894 }, { "epoch": 0.9476994635164777, "grad_norm": 4.305973529815674, "learning_rate": 1.4910596652098818e-07, "loss": 2.8082, "step": 26895 }, { "epoch": 0.9477347005294361, "grad_norm": 3.101903200149536, "learning_rate": 1.4890968174710252e-07, "loss": 2.7043, "step": 26896 }, { "epoch": 0.9477699375423946, "grad_norm": 1.7649720907211304, "learning_rate": 1.4871352528492255e-07, "loss": 0.9575, "step": 26897 }, { "epoch": 0.947805174555353, "grad_norm": 2.7131597995758057, "learning_rate": 1.4851749713700515e-07, "loss": 2.373, "step": 26898 }, { "epoch": 0.9478404115683113, "grad_norm": 1.388931393623352, "learning_rate": 1.4832159730590046e-07, "loss": 0.7995, "step": 26899 }, { "epoch": 0.9478756485812697, "grad_norm": 3.059133529663086, "learning_rate": 1.48125825794162e-07, "loss": 3.0924, "step": 26900 }, { "epoch": 0.9479108855942282, "grad_norm": 3.2781665325164795, "learning_rate": 1.4793018260434112e-07, "loss": 3.0512, "step": 26901 }, { "epoch": 0.9479461226071866, "grad_norm": 5.944548606872559, "learning_rate": 1.477346677389835e-07, "loss": 4.9053, "step": 26902 }, { "epoch": 0.947981359620145, "grad_norm": 1.7281124591827393, "learning_rate": 1.4753928120063597e-07, "loss": 0.8559, "step": 26903 }, { "epoch": 0.9480165966331034, "grad_norm": 1.976491928100586, "learning_rate": 1.4734402299184657e-07, "loss": 1.0541, "step": 26904 }, { "epoch": 0.9480518336460618, "grad_norm": 3.794174909591675, "learning_rate": 1.4714889311515434e-07, "loss": 3.1902, "step": 26905 }, { "epoch": 0.9480870706590202, "grad_norm": 3.4500772953033447, "learning_rate": 1.469538915731039e-07, "loss": 3.0237, "step": 26906 }, { "epoch": 0.9481223076719787, "grad_norm": 9.686363220214844, "learning_rate": 1.4675901836823326e-07, "loss": 3.9643, "step": 26907 }, { "epoch": 0.9481575446849371, "grad_norm": 4.1964430809021, "learning_rate": 1.4656427350308257e-07, "loss": 0.8251, "step": 26908 }, { "epoch": 0.9481927816978954, "grad_norm": 4.97376823425293, "learning_rate": 1.4636965698018645e-07, "loss": 5.0862, "step": 26909 }, { "epoch": 0.9482280187108538, "grad_norm": 3.4216692447662354, "learning_rate": 1.4617516880208072e-07, "loss": 2.9016, "step": 26910 }, { "epoch": 0.9482632557238123, "grad_norm": 3.080291509628296, "learning_rate": 1.4598080897129886e-07, "loss": 3.3017, "step": 26911 }, { "epoch": 0.9482984927367707, "grad_norm": 1.4690355062484741, "learning_rate": 1.4578657749037218e-07, "loss": 0.7832, "step": 26912 }, { "epoch": 0.9483337297497291, "grad_norm": 1.2506877183914185, "learning_rate": 1.4559247436182976e-07, "loss": 0.7996, "step": 26913 }, { "epoch": 0.9483689667626876, "grad_norm": 1.2573816776275635, "learning_rate": 1.4539849958820186e-07, "loss": 0.7819, "step": 26914 }, { "epoch": 0.9484042037756459, "grad_norm": 3.3064608573913574, "learning_rate": 1.4520465317201304e-07, "loss": 2.7574, "step": 26915 }, { "epoch": 0.9484394407886043, "grad_norm": 2.6539649963378906, "learning_rate": 1.4501093511578912e-07, "loss": 2.9416, "step": 26916 }, { "epoch": 0.9484746778015628, "grad_norm": 3.345036506652832, "learning_rate": 1.4481734542205362e-07, "loss": 0.7359, "step": 26917 }, { "epoch": 0.9485099148145212, "grad_norm": 1.553957462310791, "learning_rate": 1.4462388409332674e-07, "loss": 0.7041, "step": 26918 }, { "epoch": 0.9485451518274796, "grad_norm": 2.4983420372009277, "learning_rate": 1.4443055113212866e-07, "loss": 2.4698, "step": 26919 }, { "epoch": 0.948580388840438, "grad_norm": 5.60361385345459, "learning_rate": 1.4423734654097965e-07, "loss": 0.8769, "step": 26920 }, { "epoch": 0.9486156258533964, "grad_norm": 1.0532132387161255, "learning_rate": 1.440442703223943e-07, "loss": 1.0562, "step": 26921 }, { "epoch": 0.9486508628663548, "grad_norm": 3.2728021144866943, "learning_rate": 1.4385132247888624e-07, "loss": 3.0873, "step": 26922 }, { "epoch": 0.9486860998793132, "grad_norm": 5.586752891540527, "learning_rate": 1.4365850301297223e-07, "loss": 3.0217, "step": 26923 }, { "epoch": 0.9487213368922717, "grad_norm": 3.079064130783081, "learning_rate": 1.4346581192716036e-07, "loss": 1.0259, "step": 26924 }, { "epoch": 0.94875657390523, "grad_norm": 4.920670032501221, "learning_rate": 1.432732492239619e-07, "loss": 2.6345, "step": 26925 }, { "epoch": 0.9487918109181884, "grad_norm": 12.892717361450195, "learning_rate": 1.4308081490588487e-07, "loss": 3.0345, "step": 26926 }, { "epoch": 0.9488270479311469, "grad_norm": 1.051353096961975, "learning_rate": 1.4288850897543615e-07, "loss": 0.7855, "step": 26927 }, { "epoch": 0.9488622849441053, "grad_norm": 1.1708836555480957, "learning_rate": 1.426963314351204e-07, "loss": 1.0788, "step": 26928 }, { "epoch": 0.9488975219570637, "grad_norm": 3.9064595699310303, "learning_rate": 1.4250428228744006e-07, "loss": 2.7999, "step": 26929 }, { "epoch": 0.9489327589700222, "grad_norm": 7.201807022094727, "learning_rate": 1.4231236153489757e-07, "loss": 2.8066, "step": 26930 }, { "epoch": 0.9489679959829805, "grad_norm": 1.0706417560577393, "learning_rate": 1.4212056917999207e-07, "loss": 1.0556, "step": 26931 }, { "epoch": 0.9490032329959389, "grad_norm": 1.0989426374435425, "learning_rate": 1.419289052252215e-07, "loss": 0.9796, "step": 26932 }, { "epoch": 0.9490384700088973, "grad_norm": 2.7491345405578613, "learning_rate": 1.4173736967308283e-07, "loss": 2.5854, "step": 26933 }, { "epoch": 0.9490737070218558, "grad_norm": 1.0771229267120361, "learning_rate": 1.4154596252607068e-07, "loss": 0.5729, "step": 26934 }, { "epoch": 0.9491089440348142, "grad_norm": 1.3140166997909546, "learning_rate": 1.413546837866786e-07, "loss": 1.0293, "step": 26935 }, { "epoch": 0.9491441810477725, "grad_norm": 1.2524526119232178, "learning_rate": 1.4116353345739798e-07, "loss": 0.776, "step": 26936 }, { "epoch": 0.949179418060731, "grad_norm": 0.9740105271339417, "learning_rate": 1.4097251154071569e-07, "loss": 0.8005, "step": 26937 }, { "epoch": 0.9492146550736894, "grad_norm": 5.5217061042785645, "learning_rate": 1.4078161803912415e-07, "loss": 2.846, "step": 26938 }, { "epoch": 0.9492498920866478, "grad_norm": 0.9946750402450562, "learning_rate": 1.4059085295510811e-07, "loss": 0.7182, "step": 26939 }, { "epoch": 0.9492851290996063, "grad_norm": 1.3180917501449585, "learning_rate": 1.404002162911522e-07, "loss": 0.7703, "step": 26940 }, { "epoch": 0.9493203661125647, "grad_norm": 1.0475022792816162, "learning_rate": 1.4020970804973776e-07, "loss": 0.8159, "step": 26941 }, { "epoch": 0.949355603125523, "grad_norm": 5.882040977478027, "learning_rate": 1.400193282333495e-07, "loss": 4.9111, "step": 26942 }, { "epoch": 0.9493908401384814, "grad_norm": 2.3117129802703857, "learning_rate": 1.3982907684446544e-07, "loss": 2.277, "step": 26943 }, { "epoch": 0.9494260771514399, "grad_norm": 1.41938054561615, "learning_rate": 1.3963895388556248e-07, "loss": 0.9709, "step": 26944 }, { "epoch": 0.9494613141643983, "grad_norm": 2.990546464920044, "learning_rate": 1.3944895935911863e-07, "loss": 2.8229, "step": 26945 }, { "epoch": 0.9494965511773567, "grad_norm": 4.247561931610107, "learning_rate": 1.3925909326760855e-07, "loss": 0.7929, "step": 26946 }, { "epoch": 0.9495317881903151, "grad_norm": 2.2070484161376953, "learning_rate": 1.3906935561350478e-07, "loss": 2.5382, "step": 26947 }, { "epoch": 0.9495670252032735, "grad_norm": 1.412665843963623, "learning_rate": 1.388797463992786e-07, "loss": 0.8718, "step": 26948 }, { "epoch": 0.9496022622162319, "grad_norm": 1.3399301767349243, "learning_rate": 1.3869026562739917e-07, "loss": 0.9811, "step": 26949 }, { "epoch": 0.9496374992291904, "grad_norm": 7.4370551109313965, "learning_rate": 1.3850091330033565e-07, "loss": 6.5542, "step": 26950 }, { "epoch": 0.9496727362421488, "grad_norm": 4.5401458740234375, "learning_rate": 1.3831168942055383e-07, "loss": 3.3931, "step": 26951 }, { "epoch": 0.9497079732551071, "grad_norm": 1.0154627561569214, "learning_rate": 1.3812259399051843e-07, "loss": 0.7835, "step": 26952 }, { "epoch": 0.9497432102680656, "grad_norm": 1.1704708337783813, "learning_rate": 1.3793362701269297e-07, "loss": 0.7738, "step": 26953 }, { "epoch": 0.949778447281024, "grad_norm": 1.5330840349197388, "learning_rate": 1.3774478848953776e-07, "loss": 0.857, "step": 26954 }, { "epoch": 0.9498136842939824, "grad_norm": 4.987725734710693, "learning_rate": 1.3755607842351304e-07, "loss": 5.5261, "step": 26955 }, { "epoch": 0.9498489213069408, "grad_norm": 1.0586978197097778, "learning_rate": 1.3736749681707574e-07, "loss": 0.8915, "step": 26956 }, { "epoch": 0.9498841583198993, "grad_norm": 5.570462703704834, "learning_rate": 1.3717904367268385e-07, "loss": 4.7176, "step": 26957 }, { "epoch": 0.9499193953328576, "grad_norm": 3.452349901199341, "learning_rate": 1.3699071899279215e-07, "loss": 3.0557, "step": 26958 }, { "epoch": 0.949954632345816, "grad_norm": 3.070591688156128, "learning_rate": 1.3680252277985196e-07, "loss": 3.1279, "step": 26959 }, { "epoch": 0.9499898693587745, "grad_norm": 1.524056077003479, "learning_rate": 1.3661445503631465e-07, "loss": 0.7895, "step": 26960 }, { "epoch": 0.9500251063717329, "grad_norm": 6.096705436706543, "learning_rate": 1.3642651576463162e-07, "loss": 4.988, "step": 26961 }, { "epoch": 0.9500603433846913, "grad_norm": 4.8382978439331055, "learning_rate": 1.3623870496724978e-07, "loss": 4.597, "step": 26962 }, { "epoch": 0.9500955803976497, "grad_norm": 9.058930397033691, "learning_rate": 1.3605102264661384e-07, "loss": 4.3992, "step": 26963 }, { "epoch": 0.9501308174106081, "grad_norm": 3.4805359840393066, "learning_rate": 1.3586346880516965e-07, "loss": 2.5667, "step": 26964 }, { "epoch": 0.9501660544235665, "grad_norm": 0.6925879716873169, "learning_rate": 1.356760434453619e-07, "loss": 0.8914, "step": 26965 }, { "epoch": 0.9502012914365249, "grad_norm": 1.2270216941833496, "learning_rate": 1.3548874656962974e-07, "loss": 0.8267, "step": 26966 }, { "epoch": 0.9502365284494834, "grad_norm": 5.690432548522949, "learning_rate": 1.3530157818041233e-07, "loss": 5.7231, "step": 26967 }, { "epoch": 0.9502717654624417, "grad_norm": 4.751832008361816, "learning_rate": 1.3511453828014886e-07, "loss": 2.7616, "step": 26968 }, { "epoch": 0.9503070024754001, "grad_norm": 9.300302505493164, "learning_rate": 1.3492762687127513e-07, "loss": 5.251, "step": 26969 }, { "epoch": 0.9503422394883586, "grad_norm": 1.1295939683914185, "learning_rate": 1.3474084395622589e-07, "loss": 0.7273, "step": 26970 }, { "epoch": 0.950377476501317, "grad_norm": 1.030561923980713, "learning_rate": 1.345541895374336e-07, "loss": 0.9697, "step": 26971 }, { "epoch": 0.9504127135142754, "grad_norm": 4.323126316070557, "learning_rate": 1.3436766361732968e-07, "loss": 3.1646, "step": 26972 }, { "epoch": 0.9504479505272339, "grad_norm": 1.6211479902267456, "learning_rate": 1.3418126619834326e-07, "loss": 0.8802, "step": 26973 }, { "epoch": 0.9504831875401922, "grad_norm": 1.2145318984985352, "learning_rate": 1.3399499728290354e-07, "loss": 0.6679, "step": 26974 }, { "epoch": 0.9505184245531506, "grad_norm": 5.217979907989502, "learning_rate": 1.3380885687343527e-07, "loss": 3.2402, "step": 26975 }, { "epoch": 0.950553661566109, "grad_norm": 4.533120155334473, "learning_rate": 1.3362284497236311e-07, "loss": 3.249, "step": 26976 }, { "epoch": 0.9505888985790675, "grad_norm": 2.332026958465576, "learning_rate": 1.3343696158211074e-07, "loss": 2.8805, "step": 26977 }, { "epoch": 0.9506241355920259, "grad_norm": 1.1209403276443481, "learning_rate": 1.332512067050984e-07, "loss": 0.7798, "step": 26978 }, { "epoch": 0.9506593726049842, "grad_norm": 3.0446345806121826, "learning_rate": 1.3306558034374529e-07, "loss": 3.1135, "step": 26979 }, { "epoch": 0.9506946096179427, "grad_norm": 1.3822249174118042, "learning_rate": 1.3288008250046946e-07, "loss": 0.8445, "step": 26980 }, { "epoch": 0.9507298466309011, "grad_norm": 3.0291268825531006, "learning_rate": 1.32694713177689e-07, "loss": 2.7938, "step": 26981 }, { "epoch": 0.9507650836438595, "grad_norm": 6.188709735870361, "learning_rate": 1.3250947237781643e-07, "loss": 4.9145, "step": 26982 }, { "epoch": 0.950800320656818, "grad_norm": 4.277479648590088, "learning_rate": 1.3232436010326422e-07, "loss": 3.9069, "step": 26983 }, { "epoch": 0.9508355576697763, "grad_norm": 2.885444402694702, "learning_rate": 1.3213937635644491e-07, "loss": 3.0233, "step": 26984 }, { "epoch": 0.9508707946827347, "grad_norm": 1.0401028394699097, "learning_rate": 1.3195452113976658e-07, "loss": 1.2006, "step": 26985 }, { "epoch": 0.9509060316956932, "grad_norm": 3.880990743637085, "learning_rate": 1.3176979445563841e-07, "loss": 2.6036, "step": 26986 }, { "epoch": 0.9509412687086516, "grad_norm": 1.2882038354873657, "learning_rate": 1.3158519630646406e-07, "loss": 0.9015, "step": 26987 }, { "epoch": 0.95097650572161, "grad_norm": 3.5613903999328613, "learning_rate": 1.3140072669465155e-07, "loss": 2.6321, "step": 26988 }, { "epoch": 0.9510117427345683, "grad_norm": 1.0578845739364624, "learning_rate": 1.3121638562260008e-07, "loss": 0.6615, "step": 26989 }, { "epoch": 0.9510469797475268, "grad_norm": 3.754948854446411, "learning_rate": 1.310321730927133e-07, "loss": 2.6834, "step": 26990 }, { "epoch": 0.9510822167604852, "grad_norm": 2.9595189094543457, "learning_rate": 1.3084808910738823e-07, "loss": 2.855, "step": 26991 }, { "epoch": 0.9511174537734436, "grad_norm": 1.160406231880188, "learning_rate": 1.3066413366902507e-07, "loss": 0.8216, "step": 26992 }, { "epoch": 0.9511526907864021, "grad_norm": 1.2390656471252441, "learning_rate": 1.3048030678001976e-07, "loss": 0.8956, "step": 26993 }, { "epoch": 0.9511879277993605, "grad_norm": 0.809853732585907, "learning_rate": 1.302966084427637e-07, "loss": 1.0069, "step": 26994 }, { "epoch": 0.9512231648123188, "grad_norm": 1.5922428369522095, "learning_rate": 1.3011303865965163e-07, "loss": 0.7933, "step": 26995 }, { "epoch": 0.9512584018252773, "grad_norm": 1.7661139965057373, "learning_rate": 1.2992959743307608e-07, "loss": 0.7408, "step": 26996 }, { "epoch": 0.9512936388382357, "grad_norm": 1.1265136003494263, "learning_rate": 1.2974628476542294e-07, "loss": 1.0455, "step": 26997 }, { "epoch": 0.9513288758511941, "grad_norm": 0.9134335517883301, "learning_rate": 1.295631006590825e-07, "loss": 0.7275, "step": 26998 }, { "epoch": 0.9513641128641525, "grad_norm": 0.9784049987792969, "learning_rate": 1.2938004511643954e-07, "loss": 0.7524, "step": 26999 }, { "epoch": 0.951399349877111, "grad_norm": 2.3462343215942383, "learning_rate": 1.291971181398799e-07, "loss": 0.916, "step": 27000 }, { "epoch": 0.9514345868900693, "grad_norm": 3.3490359783172607, "learning_rate": 1.290143197317839e-07, "loss": 4.5739, "step": 27001 }, { "epoch": 0.9514698239030277, "grad_norm": 5.171258926391602, "learning_rate": 1.2883164989453413e-07, "loss": 4.4523, "step": 27002 }, { "epoch": 0.9515050609159862, "grad_norm": 0.8780925869941711, "learning_rate": 1.2864910863051084e-07, "loss": 0.7303, "step": 27003 }, { "epoch": 0.9515402979289446, "grad_norm": 1.22986900806427, "learning_rate": 1.2846669594208882e-07, "loss": 0.8281, "step": 27004 }, { "epoch": 0.951575534941903, "grad_norm": 1.0882588624954224, "learning_rate": 1.2828441183164508e-07, "loss": 0.8999, "step": 27005 }, { "epoch": 0.9516107719548614, "grad_norm": 7.006519317626953, "learning_rate": 1.2810225630155548e-07, "loss": 2.7502, "step": 27006 }, { "epoch": 0.9516460089678198, "grad_norm": 1.1390281915664673, "learning_rate": 1.2792022935419147e-07, "loss": 0.6491, "step": 27007 }, { "epoch": 0.9516812459807782, "grad_norm": 4.233831882476807, "learning_rate": 1.2773833099192334e-07, "loss": 2.3187, "step": 27008 }, { "epoch": 0.9517164829937366, "grad_norm": 5.77869176864624, "learning_rate": 1.2755656121712144e-07, "loss": 4.583, "step": 27009 }, { "epoch": 0.9517517200066951, "grad_norm": 1.435804009437561, "learning_rate": 1.273749200321528e-07, "loss": 0.7778, "step": 27010 }, { "epoch": 0.9517869570196534, "grad_norm": 2.252037286758423, "learning_rate": 1.2719340743938213e-07, "loss": 3.047, "step": 27011 }, { "epoch": 0.9518221940326118, "grad_norm": 1.0170093774795532, "learning_rate": 1.270120234411776e-07, "loss": 0.7018, "step": 27012 }, { "epoch": 0.9518574310455703, "grad_norm": 1.2604551315307617, "learning_rate": 1.2683076803989614e-07, "loss": 0.8863, "step": 27013 }, { "epoch": 0.9518926680585287, "grad_norm": 3.168738603591919, "learning_rate": 1.266496412379037e-07, "loss": 2.8039, "step": 27014 }, { "epoch": 0.9519279050714871, "grad_norm": 5.1307477951049805, "learning_rate": 1.2646864303755725e-07, "loss": 2.7244, "step": 27015 }, { "epoch": 0.9519631420844455, "grad_norm": 3.2093918323516846, "learning_rate": 1.262877734412138e-07, "loss": 3.1336, "step": 27016 }, { "epoch": 0.9519983790974039, "grad_norm": 2.640615701675415, "learning_rate": 1.2610703245122924e-07, "loss": 2.4086, "step": 27017 }, { "epoch": 0.9520336161103623, "grad_norm": 2.991509437561035, "learning_rate": 1.2592642006995836e-07, "loss": 3.1022, "step": 27018 }, { "epoch": 0.9520688531233208, "grad_norm": 8.323811531066895, "learning_rate": 1.2574593629975485e-07, "loss": 4.49, "step": 27019 }, { "epoch": 0.9521040901362792, "grad_norm": 4.319003105163574, "learning_rate": 1.255655811429668e-07, "loss": 3.1444, "step": 27020 }, { "epoch": 0.9521393271492375, "grad_norm": 1.1474709510803223, "learning_rate": 1.2538535460194568e-07, "loss": 0.8602, "step": 27021 }, { "epoch": 0.9521745641621959, "grad_norm": 4.30453634262085, "learning_rate": 1.2520525667903848e-07, "loss": 5.3825, "step": 27022 }, { "epoch": 0.9522098011751544, "grad_norm": 2.21573805809021, "learning_rate": 1.2502528737659003e-07, "loss": 0.9239, "step": 27023 }, { "epoch": 0.9522450381881128, "grad_norm": 4.253479480743408, "learning_rate": 1.248454466969451e-07, "loss": 3.2055, "step": 27024 }, { "epoch": 0.9522802752010712, "grad_norm": 0.969244658946991, "learning_rate": 1.2466573464244625e-07, "loss": 0.9347, "step": 27025 }, { "epoch": 0.9523155122140297, "grad_norm": 1.270201325416565, "learning_rate": 1.2448615121543495e-07, "loss": 0.9203, "step": 27026 }, { "epoch": 0.952350749226988, "grad_norm": 10.586271286010742, "learning_rate": 1.2430669641824823e-07, "loss": 2.8927, "step": 27027 }, { "epoch": 0.9523859862399464, "grad_norm": 5.03938102722168, "learning_rate": 1.241273702532264e-07, "loss": 2.9525, "step": 27028 }, { "epoch": 0.9524212232529049, "grad_norm": 7.354108810424805, "learning_rate": 1.2394817272270099e-07, "loss": 3.4693, "step": 27029 }, { "epoch": 0.9524564602658633, "grad_norm": 1.2687534093856812, "learning_rate": 1.2376910382901008e-07, "loss": 0.8556, "step": 27030 }, { "epoch": 0.9524916972788217, "grad_norm": 3.0616772174835205, "learning_rate": 1.2359016357448516e-07, "loss": 3.1101, "step": 27031 }, { "epoch": 0.95252693429178, "grad_norm": 5.0950822830200195, "learning_rate": 1.2341135196145659e-07, "loss": 5.0291, "step": 27032 }, { "epoch": 0.9525621713047385, "grad_norm": 3.549105644226074, "learning_rate": 1.232326689922514e-07, "loss": 2.6811, "step": 27033 }, { "epoch": 0.9525974083176969, "grad_norm": 2.843214273452759, "learning_rate": 1.2305411466920103e-07, "loss": 3.0361, "step": 27034 }, { "epoch": 0.9526326453306553, "grad_norm": 3.2647879123687744, "learning_rate": 1.2287568899462698e-07, "loss": 3.0383, "step": 27035 }, { "epoch": 0.9526678823436138, "grad_norm": 1.4848785400390625, "learning_rate": 1.226973919708563e-07, "loss": 0.8621, "step": 27036 }, { "epoch": 0.9527031193565721, "grad_norm": 1.2892889976501465, "learning_rate": 1.225192236002093e-07, "loss": 0.8369, "step": 27037 }, { "epoch": 0.9527383563695305, "grad_norm": 7.770419120788574, "learning_rate": 1.223411838850086e-07, "loss": 3.5254, "step": 27038 }, { "epoch": 0.952773593382489, "grad_norm": 1.4886128902435303, "learning_rate": 1.2216327282757235e-07, "loss": 0.7216, "step": 27039 }, { "epoch": 0.9528088303954474, "grad_norm": 3.446796417236328, "learning_rate": 1.2198549043021756e-07, "loss": 0.7127, "step": 27040 }, { "epoch": 0.9528440674084058, "grad_norm": 3.8884522914886475, "learning_rate": 1.2180783669525908e-07, "loss": 3.1098, "step": 27041 }, { "epoch": 0.9528793044213641, "grad_norm": 4.797194004058838, "learning_rate": 1.2163031162501283e-07, "loss": 3.1616, "step": 27042 }, { "epoch": 0.9529145414343226, "grad_norm": 3.757676362991333, "learning_rate": 1.2145291522178915e-07, "loss": 2.657, "step": 27043 }, { "epoch": 0.952949778447281, "grad_norm": 4.15325403213501, "learning_rate": 1.2127564748790066e-07, "loss": 5.1689, "step": 27044 }, { "epoch": 0.9529850154602394, "grad_norm": 1.6708558797836304, "learning_rate": 1.210985084256544e-07, "loss": 0.9169, "step": 27045 }, { "epoch": 0.9530202524731979, "grad_norm": 2.760429859161377, "learning_rate": 1.2092149803735853e-07, "loss": 0.7906, "step": 27046 }, { "epoch": 0.9530554894861563, "grad_norm": 3.596651554107666, "learning_rate": 1.2074461632532007e-07, "loss": 2.6437, "step": 27047 }, { "epoch": 0.9530907264991146, "grad_norm": 4.37078332901001, "learning_rate": 1.2056786329183944e-07, "loss": 4.9032, "step": 27048 }, { "epoch": 0.9531259635120731, "grad_norm": 2.6000590324401855, "learning_rate": 1.2039123893922144e-07, "loss": 2.4706, "step": 27049 }, { "epoch": 0.9531612005250315, "grad_norm": 1.8840465545654297, "learning_rate": 1.2021474326976646e-07, "loss": 0.9498, "step": 27050 }, { "epoch": 0.9531964375379899, "grad_norm": 2.3016905784606934, "learning_rate": 1.2003837628577263e-07, "loss": 2.5658, "step": 27051 }, { "epoch": 0.9532316745509484, "grad_norm": 3.9898324012756348, "learning_rate": 1.1986213798953593e-07, "loss": 5.3977, "step": 27052 }, { "epoch": 0.9532669115639067, "grad_norm": 3.810333728790283, "learning_rate": 1.196860283833545e-07, "loss": 4.7633, "step": 27053 }, { "epoch": 0.9533021485768651, "grad_norm": 1.1368768215179443, "learning_rate": 1.1951004746952212e-07, "loss": 0.9788, "step": 27054 }, { "epoch": 0.9533373855898235, "grad_norm": 1.1697299480438232, "learning_rate": 1.1933419525032907e-07, "loss": 0.7149, "step": 27055 }, { "epoch": 0.953372622602782, "grad_norm": 1.2999221086502075, "learning_rate": 1.1915847172806693e-07, "loss": 0.6664, "step": 27056 }, { "epoch": 0.9534078596157404, "grad_norm": 5.3541154861450195, "learning_rate": 1.1898287690502497e-07, "loss": 4.209, "step": 27057 }, { "epoch": 0.9534430966286987, "grad_norm": 1.6645326614379883, "learning_rate": 1.1880741078348912e-07, "loss": 1.0314, "step": 27058 }, { "epoch": 0.9534783336416572, "grad_norm": 2.232346534729004, "learning_rate": 1.1863207336574533e-07, "loss": 1.111, "step": 27059 }, { "epoch": 0.9535135706546156, "grad_norm": 1.0964019298553467, "learning_rate": 1.1845686465407736e-07, "loss": 0.9526, "step": 27060 }, { "epoch": 0.953548807667574, "grad_norm": 4.246921539306641, "learning_rate": 1.1828178465076668e-07, "loss": 2.8664, "step": 27061 }, { "epoch": 0.9535840446805325, "grad_norm": 3.130567789077759, "learning_rate": 1.1810683335809591e-07, "loss": 3.0194, "step": 27062 }, { "epoch": 0.9536192816934909, "grad_norm": 3.1563520431518555, "learning_rate": 1.1793201077834105e-07, "loss": 3.9984, "step": 27063 }, { "epoch": 0.9536545187064492, "grad_norm": 1.8361917734146118, "learning_rate": 1.1775731691378134e-07, "loss": 0.7458, "step": 27064 }, { "epoch": 0.9536897557194076, "grad_norm": 0.988446056842804, "learning_rate": 1.1758275176669165e-07, "loss": 0.9721, "step": 27065 }, { "epoch": 0.9537249927323661, "grad_norm": 4.2063398361206055, "learning_rate": 1.1740831533934572e-07, "loss": 2.6893, "step": 27066 }, { "epoch": 0.9537602297453245, "grad_norm": 1.8032411336898804, "learning_rate": 1.1723400763401393e-07, "loss": 0.9412, "step": 27067 }, { "epoch": 0.9537954667582829, "grad_norm": 5.522700786590576, "learning_rate": 1.1705982865296894e-07, "loss": 5.4942, "step": 27068 }, { "epoch": 0.9538307037712414, "grad_norm": 4.125246047973633, "learning_rate": 1.1688577839847893e-07, "loss": 2.7303, "step": 27069 }, { "epoch": 0.9538659407841997, "grad_norm": 1.136766791343689, "learning_rate": 1.1671185687280984e-07, "loss": 0.7, "step": 27070 }, { "epoch": 0.9539011777971581, "grad_norm": 1.1282453536987305, "learning_rate": 1.1653806407822765e-07, "loss": 0.7896, "step": 27071 }, { "epoch": 0.9539364148101166, "grad_norm": 1.1569724082946777, "learning_rate": 1.1636440001699612e-07, "loss": 0.741, "step": 27072 }, { "epoch": 0.953971651823075, "grad_norm": 4.998183727264404, "learning_rate": 1.1619086469137785e-07, "loss": 2.0304, "step": 27073 }, { "epoch": 0.9540068888360334, "grad_norm": 0.9719744324684143, "learning_rate": 1.1601745810363218e-07, "loss": 0.7007, "step": 27074 }, { "epoch": 0.9540421258489917, "grad_norm": 4.1033148765563965, "learning_rate": 1.1584418025601729e-07, "loss": 2.7518, "step": 27075 }, { "epoch": 0.9540773628619502, "grad_norm": 3.4325685501098633, "learning_rate": 1.1567103115079248e-07, "loss": 2.6892, "step": 27076 }, { "epoch": 0.9541125998749086, "grad_norm": 1.1043541431427002, "learning_rate": 1.154980107902115e-07, "loss": 1.1174, "step": 27077 }, { "epoch": 0.954147836887867, "grad_norm": 2.0713911056518555, "learning_rate": 1.15325119176527e-07, "loss": 0.927, "step": 27078 }, { "epoch": 0.9541830739008255, "grad_norm": 3.822751760482788, "learning_rate": 1.1515235631199273e-07, "loss": 2.6315, "step": 27079 }, { "epoch": 0.9542183109137838, "grad_norm": 0.9338358044624329, "learning_rate": 1.14979722198858e-07, "loss": 0.7496, "step": 27080 }, { "epoch": 0.9542535479267422, "grad_norm": 1.9688067436218262, "learning_rate": 1.1480721683937101e-07, "loss": 0.9441, "step": 27081 }, { "epoch": 0.9542887849397007, "grad_norm": 1.411568284034729, "learning_rate": 1.1463484023577998e-07, "loss": 0.8809, "step": 27082 }, { "epoch": 0.9543240219526591, "grad_norm": 1.173932433128357, "learning_rate": 1.1446259239032864e-07, "loss": 0.7621, "step": 27083 }, { "epoch": 0.9543592589656175, "grad_norm": 2.1578853130340576, "learning_rate": 1.1429047330526188e-07, "loss": 0.9693, "step": 27084 }, { "epoch": 0.9543944959785758, "grad_norm": 1.1729413270950317, "learning_rate": 1.1411848298282124e-07, "loss": 0.8402, "step": 27085 }, { "epoch": 0.9544297329915343, "grad_norm": 4.378829002380371, "learning_rate": 1.1394662142524493e-07, "loss": 3.0921, "step": 27086 }, { "epoch": 0.9544649700044927, "grad_norm": 6.468181133270264, "learning_rate": 1.1377488863477449e-07, "loss": 4.4493, "step": 27087 }, { "epoch": 0.9545002070174511, "grad_norm": 6.332935333251953, "learning_rate": 1.136032846136459e-07, "loss": 4.5338, "step": 27088 }, { "epoch": 0.9545354440304096, "grad_norm": 3.756047248840332, "learning_rate": 1.1343180936409293e-07, "loss": 3.0319, "step": 27089 }, { "epoch": 0.954570681043368, "grad_norm": 0.9699943661689758, "learning_rate": 1.1326046288834935e-07, "loss": 0.6339, "step": 27090 }, { "epoch": 0.9546059180563263, "grad_norm": 2.4100849628448486, "learning_rate": 1.1308924518864894e-07, "loss": 2.4433, "step": 27091 }, { "epoch": 0.9546411550692848, "grad_norm": 2.9238200187683105, "learning_rate": 1.1291815626722103e-07, "loss": 2.7615, "step": 27092 }, { "epoch": 0.9546763920822432, "grad_norm": 1.8812638521194458, "learning_rate": 1.1274719612629159e-07, "loss": 0.6808, "step": 27093 }, { "epoch": 0.9547116290952016, "grad_norm": 3.5784788131713867, "learning_rate": 1.1257636476808998e-07, "loss": 2.8392, "step": 27094 }, { "epoch": 0.9547468661081601, "grad_norm": 3.8504796028137207, "learning_rate": 1.1240566219484217e-07, "loss": 3.0734, "step": 27095 }, { "epoch": 0.9547821031211184, "grad_norm": 4.418074607849121, "learning_rate": 1.1223508840876862e-07, "loss": 2.6113, "step": 27096 }, { "epoch": 0.9548173401340768, "grad_norm": 1.854805827140808, "learning_rate": 1.1206464341209312e-07, "loss": 0.8219, "step": 27097 }, { "epoch": 0.9548525771470352, "grad_norm": 6.009998798370361, "learning_rate": 1.1189432720703497e-07, "loss": 2.8252, "step": 27098 }, { "epoch": 0.9548878141599937, "grad_norm": 1.0435113906860352, "learning_rate": 1.1172413979581243e-07, "loss": 0.9695, "step": 27099 }, { "epoch": 0.9549230511729521, "grad_norm": 3.2317423820495605, "learning_rate": 1.115540811806426e-07, "loss": 1.1226, "step": 27100 }, { "epoch": 0.9549582881859104, "grad_norm": 5.3802080154418945, "learning_rate": 1.1138415136374037e-07, "loss": 5.3596, "step": 27101 }, { "epoch": 0.9549935251988689, "grad_norm": 1.037377119064331, "learning_rate": 1.1121435034731954e-07, "loss": 0.6786, "step": 27102 }, { "epoch": 0.9550287622118273, "grad_norm": 4.324066638946533, "learning_rate": 1.1104467813359165e-07, "loss": 2.4775, "step": 27103 }, { "epoch": 0.9550639992247857, "grad_norm": 2.5891289710998535, "learning_rate": 1.1087513472476607e-07, "loss": 2.2219, "step": 27104 }, { "epoch": 0.9550992362377442, "grad_norm": 8.898971557617188, "learning_rate": 1.1070572012305103e-07, "loss": 2.9782, "step": 27105 }, { "epoch": 0.9551344732507026, "grad_norm": 1.224435806274414, "learning_rate": 1.1053643433065475e-07, "loss": 0.8632, "step": 27106 }, { "epoch": 0.9551697102636609, "grad_norm": 2.475301504135132, "learning_rate": 1.1036727734978104e-07, "loss": 2.3007, "step": 27107 }, { "epoch": 0.9552049472766193, "grad_norm": 1.2650004625320435, "learning_rate": 1.1019824918263255e-07, "loss": 0.9361, "step": 27108 }, { "epoch": 0.9552401842895778, "grad_norm": 3.0798728466033936, "learning_rate": 1.10029349831412e-07, "loss": 3.0485, "step": 27109 }, { "epoch": 0.9552754213025362, "grad_norm": 1.5302579402923584, "learning_rate": 1.0986057929831873e-07, "loss": 1.1744, "step": 27110 }, { "epoch": 0.9553106583154946, "grad_norm": 5.6045637130737305, "learning_rate": 1.0969193758555207e-07, "loss": 7.3377, "step": 27111 }, { "epoch": 0.955345895328453, "grad_norm": 3.88130521774292, "learning_rate": 1.0952342469530696e-07, "loss": 5.0823, "step": 27112 }, { "epoch": 0.9553811323414114, "grad_norm": 5.532439708709717, "learning_rate": 1.093550406297783e-07, "loss": 2.6426, "step": 27113 }, { "epoch": 0.9554163693543698, "grad_norm": 1.5247647762298584, "learning_rate": 1.0918678539116212e-07, "loss": 0.7536, "step": 27114 }, { "epoch": 0.9554516063673283, "grad_norm": 1.1680152416229248, "learning_rate": 1.0901865898164665e-07, "loss": 0.7498, "step": 27115 }, { "epoch": 0.9554868433802867, "grad_norm": 0.9831831455230713, "learning_rate": 1.0885066140342238e-07, "loss": 0.9778, "step": 27116 }, { "epoch": 0.955522080393245, "grad_norm": 4.182613372802734, "learning_rate": 1.0868279265867976e-07, "loss": 3.2355, "step": 27117 }, { "epoch": 0.9555573174062034, "grad_norm": 6.6868085861206055, "learning_rate": 1.0851505274960262e-07, "loss": 3.237, "step": 27118 }, { "epoch": 0.9555925544191619, "grad_norm": 0.9602370262145996, "learning_rate": 1.0834744167837696e-07, "loss": 0.6393, "step": 27119 }, { "epoch": 0.9556277914321203, "grad_norm": 3.380396842956543, "learning_rate": 1.0817995944718663e-07, "loss": 3.1173, "step": 27120 }, { "epoch": 0.9556630284450787, "grad_norm": 0.9917165637016296, "learning_rate": 1.0801260605821206e-07, "loss": 0.753, "step": 27121 }, { "epoch": 0.9556982654580372, "grad_norm": 7.319015026092529, "learning_rate": 1.0784538151363267e-07, "loss": 7.3361, "step": 27122 }, { "epoch": 0.9557335024709955, "grad_norm": 1.1987850666046143, "learning_rate": 1.0767828581562889e-07, "loss": 0.8175, "step": 27123 }, { "epoch": 0.9557687394839539, "grad_norm": 3.2847306728363037, "learning_rate": 1.0751131896637346e-07, "loss": 2.9299, "step": 27124 }, { "epoch": 0.9558039764969124, "grad_norm": 1.8323012590408325, "learning_rate": 1.0734448096804351e-07, "loss": 0.9202, "step": 27125 }, { "epoch": 0.9558392135098708, "grad_norm": 1.0282089710235596, "learning_rate": 1.0717777182281286e-07, "loss": 0.7998, "step": 27126 }, { "epoch": 0.9558744505228292, "grad_norm": 1.1840429306030273, "learning_rate": 1.07011191532852e-07, "loss": 0.953, "step": 27127 }, { "epoch": 0.9559096875357876, "grad_norm": 1.0212842226028442, "learning_rate": 1.068447401003292e-07, "loss": 0.874, "step": 27128 }, { "epoch": 0.955944924548746, "grad_norm": 1.0530002117156982, "learning_rate": 1.0667841752741493e-07, "loss": 0.9366, "step": 27129 }, { "epoch": 0.9559801615617044, "grad_norm": 5.783312797546387, "learning_rate": 1.0651222381627524e-07, "loss": 3.4055, "step": 27130 }, { "epoch": 0.9560153985746628, "grad_norm": 7.159927845001221, "learning_rate": 1.0634615896907285e-07, "loss": 4.9969, "step": 27131 }, { "epoch": 0.9560506355876213, "grad_norm": 1.1421366930007935, "learning_rate": 1.0618022298797159e-07, "loss": 0.9052, "step": 27132 }, { "epoch": 0.9560858726005796, "grad_norm": 0.9499583840370178, "learning_rate": 1.0601441587513417e-07, "loss": 0.7338, "step": 27133 }, { "epoch": 0.956121109613538, "grad_norm": 4.744404315948486, "learning_rate": 1.0584873763271885e-07, "loss": 2.6812, "step": 27134 }, { "epoch": 0.9561563466264965, "grad_norm": 1.0409504175186157, "learning_rate": 1.0568318826288393e-07, "loss": 0.901, "step": 27135 }, { "epoch": 0.9561915836394549, "grad_norm": 1.5695987939834595, "learning_rate": 1.0551776776778544e-07, "loss": 0.9667, "step": 27136 }, { "epoch": 0.9562268206524133, "grad_norm": 9.021507263183594, "learning_rate": 1.0535247614957944e-07, "loss": 3.1562, "step": 27137 }, { "epoch": 0.9562620576653718, "grad_norm": 1.0813353061676025, "learning_rate": 1.0518731341041643e-07, "loss": 0.8063, "step": 27138 }, { "epoch": 0.9562972946783301, "grad_norm": 1.613541603088379, "learning_rate": 1.0502227955245026e-07, "loss": 0.7898, "step": 27139 }, { "epoch": 0.9563325316912885, "grad_norm": 1.2003965377807617, "learning_rate": 1.0485737457782919e-07, "loss": 0.8866, "step": 27140 }, { "epoch": 0.9563677687042469, "grad_norm": 1.0205572843551636, "learning_rate": 1.0469259848870039e-07, "loss": 0.7635, "step": 27141 }, { "epoch": 0.9564030057172054, "grad_norm": 5.04925537109375, "learning_rate": 1.0452795128721216e-07, "loss": 4.7592, "step": 27142 }, { "epoch": 0.9564382427301638, "grad_norm": 1.0365997552871704, "learning_rate": 1.0436343297550722e-07, "loss": 0.9315, "step": 27143 }, { "epoch": 0.9564734797431221, "grad_norm": 4.123460292816162, "learning_rate": 1.041990435557283e-07, "loss": 5.1221, "step": 27144 }, { "epoch": 0.9565087167560806, "grad_norm": 5.437845706939697, "learning_rate": 1.0403478303001813e-07, "loss": 4.734, "step": 27145 }, { "epoch": 0.956543953769039, "grad_norm": 1.5767375230789185, "learning_rate": 1.0387065140051611e-07, "loss": 0.8957, "step": 27146 }, { "epoch": 0.9565791907819974, "grad_norm": 1.1185452938079834, "learning_rate": 1.0370664866935831e-07, "loss": 0.7737, "step": 27147 }, { "epoch": 0.9566144277949559, "grad_norm": 3.054487943649292, "learning_rate": 1.0354277483868302e-07, "loss": 2.6965, "step": 27148 }, { "epoch": 0.9566496648079142, "grad_norm": 1.0896869897842407, "learning_rate": 1.0337902991062409e-07, "loss": 0.8867, "step": 27149 }, { "epoch": 0.9566849018208726, "grad_norm": 1.313950777053833, "learning_rate": 1.0321541388731204e-07, "loss": 0.8625, "step": 27150 }, { "epoch": 0.956720138833831, "grad_norm": 1.1877717971801758, "learning_rate": 1.0305192677088071e-07, "loss": 0.7976, "step": 27151 }, { "epoch": 0.9567553758467895, "grad_norm": 1.087389349937439, "learning_rate": 1.0288856856345952e-07, "loss": 0.9548, "step": 27152 }, { "epoch": 0.9567906128597479, "grad_norm": 1.0946391820907593, "learning_rate": 1.0272533926717453e-07, "loss": 0.9138, "step": 27153 }, { "epoch": 0.9568258498727062, "grad_norm": 1.615324854850769, "learning_rate": 1.0256223888415295e-07, "loss": 0.8966, "step": 27154 }, { "epoch": 0.9568610868856647, "grad_norm": 2.6423861980438232, "learning_rate": 1.0239926741651862e-07, "loss": 2.7329, "step": 27155 }, { "epoch": 0.9568963238986231, "grad_norm": 0.884729266166687, "learning_rate": 1.022364248663954e-07, "loss": 0.7918, "step": 27156 }, { "epoch": 0.9569315609115815, "grad_norm": 4.598055839538574, "learning_rate": 1.020737112359027e-07, "loss": 2.6331, "step": 27157 }, { "epoch": 0.95696679792454, "grad_norm": 5.4867262840271, "learning_rate": 1.0191112652716107e-07, "loss": 5.3131, "step": 27158 }, { "epoch": 0.9570020349374984, "grad_norm": 6.779155731201172, "learning_rate": 1.0174867074228768e-07, "loss": 4.8081, "step": 27159 }, { "epoch": 0.9570372719504567, "grad_norm": 3.3759946823120117, "learning_rate": 1.0158634388339861e-07, "loss": 2.9941, "step": 27160 }, { "epoch": 0.9570725089634152, "grad_norm": 3.96118426322937, "learning_rate": 1.0142414595260886e-07, "loss": 2.8085, "step": 27161 }, { "epoch": 0.9571077459763736, "grad_norm": 0.9135007262229919, "learning_rate": 1.0126207695202894e-07, "loss": 0.7208, "step": 27162 }, { "epoch": 0.957142982989332, "grad_norm": 5.106173515319824, "learning_rate": 1.0110013688377274e-07, "loss": 4.4824, "step": 27163 }, { "epoch": 0.9571782200022904, "grad_norm": 5.653214454650879, "learning_rate": 1.0093832574994855e-07, "loss": 5.6513, "step": 27164 }, { "epoch": 0.9572134570152488, "grad_norm": 1.2250165939331055, "learning_rate": 1.0077664355266359e-07, "loss": 0.695, "step": 27165 }, { "epoch": 0.9572486940282072, "grad_norm": 3.7181503772735596, "learning_rate": 1.0061509029402172e-07, "loss": 2.9731, "step": 27166 }, { "epoch": 0.9572839310411656, "grad_norm": 1.977962613105774, "learning_rate": 1.0045366597613126e-07, "loss": 2.6178, "step": 27167 }, { "epoch": 0.9573191680541241, "grad_norm": 1.2286008596420288, "learning_rate": 1.0029237060109276e-07, "loss": 0.8822, "step": 27168 }, { "epoch": 0.9573544050670825, "grad_norm": 1.0909383296966553, "learning_rate": 1.0013120417100675e-07, "loss": 1.0675, "step": 27169 }, { "epoch": 0.9573896420800408, "grad_norm": 1.0941133499145508, "learning_rate": 9.997016668797155e-08, "loss": 0.7961, "step": 27170 }, { "epoch": 0.9574248790929993, "grad_norm": 1.198952555656433, "learning_rate": 9.980925815408771e-08, "loss": 0.7688, "step": 27171 }, { "epoch": 0.9574601161059577, "grad_norm": 3.1266403198242188, "learning_rate": 9.96484785714491e-08, "loss": 3.0765, "step": 27172 }, { "epoch": 0.9574953531189161, "grad_norm": 5.390049457550049, "learning_rate": 9.948782794214962e-08, "loss": 3.3392, "step": 27173 }, { "epoch": 0.9575305901318745, "grad_norm": 5.672642230987549, "learning_rate": 9.932730626828313e-08, "loss": 5.4693, "step": 27174 }, { "epoch": 0.957565827144833, "grad_norm": 1.1333224773406982, "learning_rate": 9.916691355193908e-08, "loss": 1.1067, "step": 27175 }, { "epoch": 0.9576010641577913, "grad_norm": 3.279235601425171, "learning_rate": 9.900664979520691e-08, "loss": 2.9157, "step": 27176 }, { "epoch": 0.9576363011707497, "grad_norm": 6.194345951080322, "learning_rate": 9.884651500017495e-08, "loss": 5.3267, "step": 27177 }, { "epoch": 0.9576715381837082, "grad_norm": 6.00772762298584, "learning_rate": 9.86865091689282e-08, "loss": 3.2804, "step": 27178 }, { "epoch": 0.9577067751966666, "grad_norm": 0.9394506812095642, "learning_rate": 9.852663230355053e-08, "loss": 0.9403, "step": 27179 }, { "epoch": 0.957742012209625, "grad_norm": 1.2638323307037354, "learning_rate": 9.836688440612474e-08, "loss": 0.8079, "step": 27180 }, { "epoch": 0.9577772492225834, "grad_norm": 1.5825544595718384, "learning_rate": 9.820726547873138e-08, "loss": 0.8806, "step": 27181 }, { "epoch": 0.9578124862355418, "grad_norm": 3.622008800506592, "learning_rate": 9.804777552344991e-08, "loss": 2.4364, "step": 27182 }, { "epoch": 0.9578477232485002, "grad_norm": 4.7238969802856445, "learning_rate": 9.788841454235754e-08, "loss": 3.2295, "step": 27183 }, { "epoch": 0.9578829602614586, "grad_norm": 7.45107364654541, "learning_rate": 9.772918253753039e-08, "loss": 4.6087, "step": 27184 }, { "epoch": 0.9579181972744171, "grad_norm": 4.621886253356934, "learning_rate": 9.757007951104125e-08, "loss": 2.4278, "step": 27185 }, { "epoch": 0.9579534342873754, "grad_norm": 5.217950820922852, "learning_rate": 9.74111054649629e-08, "loss": 4.9121, "step": 27186 }, { "epoch": 0.9579886713003338, "grad_norm": 4.338033676147461, "learning_rate": 9.725226040136703e-08, "loss": 3.0779, "step": 27187 }, { "epoch": 0.9580239083132923, "grad_norm": 1.0564650297164917, "learning_rate": 9.709354432232199e-08, "loss": 0.9601, "step": 27188 }, { "epoch": 0.9580591453262507, "grad_norm": 1.320136547088623, "learning_rate": 9.693495722989388e-08, "loss": 0.8924, "step": 27189 }, { "epoch": 0.9580943823392091, "grad_norm": 1.2527947425842285, "learning_rate": 9.677649912615105e-08, "loss": 0.8994, "step": 27190 }, { "epoch": 0.9581296193521676, "grad_norm": 1.602228045463562, "learning_rate": 9.661817001315521e-08, "loss": 0.9465, "step": 27191 }, { "epoch": 0.9581648563651259, "grad_norm": 19.897661209106445, "learning_rate": 9.645996989296802e-08, "loss": 2.6675, "step": 27192 }, { "epoch": 0.9582000933780843, "grad_norm": 5.517457008361816, "learning_rate": 9.630189876765227e-08, "loss": 4.8878, "step": 27193 }, { "epoch": 0.9582353303910428, "grad_norm": 8.245198249816895, "learning_rate": 9.614395663926523e-08, "loss": 7.1125, "step": 27194 }, { "epoch": 0.9582705674040012, "grad_norm": 1.6653828620910645, "learning_rate": 9.598614350986524e-08, "loss": 0.6491, "step": 27195 }, { "epoch": 0.9583058044169596, "grad_norm": 3.04171085357666, "learning_rate": 9.58284593815073e-08, "loss": 2.6731, "step": 27196 }, { "epoch": 0.9583410414299179, "grad_norm": 1.199767827987671, "learning_rate": 9.567090425624537e-08, "loss": 0.874, "step": 27197 }, { "epoch": 0.9583762784428764, "grad_norm": 2.963763952255249, "learning_rate": 9.55134781361311e-08, "loss": 3.3026, "step": 27198 }, { "epoch": 0.9584115154558348, "grad_norm": 2.8693363666534424, "learning_rate": 9.53561810232162e-08, "loss": 2.6139, "step": 27199 }, { "epoch": 0.9584467524687932, "grad_norm": 1.2740578651428223, "learning_rate": 9.519901291954792e-08, "loss": 0.7756, "step": 27200 }, { "epoch": 0.9584819894817517, "grad_norm": 1.147681474685669, "learning_rate": 9.504197382717351e-08, "loss": 0.9602, "step": 27201 }, { "epoch": 0.95851722649471, "grad_norm": 1.1900283098220825, "learning_rate": 9.488506374814022e-08, "loss": 0.9175, "step": 27202 }, { "epoch": 0.9585524635076684, "grad_norm": 1.6763414144515991, "learning_rate": 9.472828268449086e-08, "loss": 0.7597, "step": 27203 }, { "epoch": 0.9585877005206269, "grad_norm": 1.4304245710372925, "learning_rate": 9.457163063826602e-08, "loss": 0.7136, "step": 27204 }, { "epoch": 0.9586229375335853, "grad_norm": 1.2248977422714233, "learning_rate": 9.441510761150852e-08, "loss": 0.859, "step": 27205 }, { "epoch": 0.9586581745465437, "grad_norm": 1.1482940912246704, "learning_rate": 9.425871360625671e-08, "loss": 0.7098, "step": 27206 }, { "epoch": 0.958693411559502, "grad_norm": 3.877192258834839, "learning_rate": 9.410244862454565e-08, "loss": 2.7107, "step": 27207 }, { "epoch": 0.9587286485724605, "grad_norm": 3.429353713989258, "learning_rate": 9.394631266841148e-08, "loss": 3.1209, "step": 27208 }, { "epoch": 0.9587638855854189, "grad_norm": 0.9996339678764343, "learning_rate": 9.379030573989034e-08, "loss": 0.8095, "step": 27209 }, { "epoch": 0.9587991225983773, "grad_norm": 6.851245880126953, "learning_rate": 9.363442784101063e-08, "loss": 3.005, "step": 27210 }, { "epoch": 0.9588343596113358, "grad_norm": 2.1268091201782227, "learning_rate": 9.347867897380514e-08, "loss": 2.4936, "step": 27211 }, { "epoch": 0.9588695966242942, "grad_norm": 3.563624620437622, "learning_rate": 9.332305914030227e-08, "loss": 2.7028, "step": 27212 }, { "epoch": 0.9589048336372525, "grad_norm": 1.3887360095977783, "learning_rate": 9.316756834252705e-08, "loss": 0.9404, "step": 27213 }, { "epoch": 0.958940070650211, "grad_norm": 1.76209557056427, "learning_rate": 9.301220658250787e-08, "loss": 1.0207, "step": 27214 }, { "epoch": 0.9589753076631694, "grad_norm": 4.668511867523193, "learning_rate": 9.285697386226533e-08, "loss": 4.3767, "step": 27215 }, { "epoch": 0.9590105446761278, "grad_norm": 4.0569634437561035, "learning_rate": 9.270187018382338e-08, "loss": 2.9351, "step": 27216 }, { "epoch": 0.9590457816890862, "grad_norm": 3.9342501163482666, "learning_rate": 9.254689554920038e-08, "loss": 4.5144, "step": 27217 }, { "epoch": 0.9590810187020447, "grad_norm": 8.28659725189209, "learning_rate": 9.239204996041695e-08, "loss": 2.9687, "step": 27218 }, { "epoch": 0.959116255715003, "grad_norm": 5.0206618309021, "learning_rate": 9.223733341948927e-08, "loss": 3.2084, "step": 27219 }, { "epoch": 0.9591514927279614, "grad_norm": 6.131488800048828, "learning_rate": 9.208274592843236e-08, "loss": 5.0746, "step": 27220 }, { "epoch": 0.9591867297409199, "grad_norm": 4.3854546546936035, "learning_rate": 9.192828748925908e-08, "loss": 2.8563, "step": 27221 }, { "epoch": 0.9592219667538783, "grad_norm": 3.0548698902130127, "learning_rate": 9.177395810398337e-08, "loss": 2.9511, "step": 27222 }, { "epoch": 0.9592572037668367, "grad_norm": 3.8729374408721924, "learning_rate": 9.161975777461252e-08, "loss": 2.555, "step": 27223 }, { "epoch": 0.9592924407797951, "grad_norm": 2.3794918060302734, "learning_rate": 9.146568650315712e-08, "loss": 2.9019, "step": 27224 }, { "epoch": 0.9593276777927535, "grad_norm": 4.005404472351074, "learning_rate": 9.131174429162337e-08, "loss": 3.4802, "step": 27225 }, { "epoch": 0.9593629148057119, "grad_norm": 5.583256721496582, "learning_rate": 9.115793114201633e-08, "loss": 2.5836, "step": 27226 }, { "epoch": 0.9593981518186704, "grad_norm": 4.421371936798096, "learning_rate": 9.100424705633882e-08, "loss": 2.9944, "step": 27227 }, { "epoch": 0.9594333888316288, "grad_norm": 4.4793877601623535, "learning_rate": 9.08506920365937e-08, "loss": 4.5378, "step": 27228 }, { "epoch": 0.9594686258445871, "grad_norm": 1.483127236366272, "learning_rate": 9.069726608478046e-08, "loss": 0.8576, "step": 27229 }, { "epoch": 0.9595038628575455, "grad_norm": 1.4810395240783691, "learning_rate": 9.054396920289754e-08, "loss": 0.7339, "step": 27230 }, { "epoch": 0.959539099870504, "grad_norm": 3.194863796234131, "learning_rate": 9.039080139294221e-08, "loss": 3.2564, "step": 27231 }, { "epoch": 0.9595743368834624, "grad_norm": 2.3829705715179443, "learning_rate": 9.023776265690731e-08, "loss": 2.5674, "step": 27232 }, { "epoch": 0.9596095738964208, "grad_norm": 3.3120059967041016, "learning_rate": 9.008485299678904e-08, "loss": 3.1099, "step": 27233 }, { "epoch": 0.9596448109093793, "grad_norm": 1.042106032371521, "learning_rate": 8.993207241457802e-08, "loss": 0.7785, "step": 27234 }, { "epoch": 0.9596800479223376, "grad_norm": 0.8553334474563599, "learning_rate": 8.977942091226378e-08, "loss": 0.9495, "step": 27235 }, { "epoch": 0.959715284935296, "grad_norm": 1.3638553619384766, "learning_rate": 8.962689849183581e-08, "loss": 0.8105, "step": 27236 }, { "epoch": 0.9597505219482545, "grad_norm": 1.2558259963989258, "learning_rate": 8.947450515527811e-08, "loss": 0.8032, "step": 27237 }, { "epoch": 0.9597857589612129, "grad_norm": 2.144442319869995, "learning_rate": 8.932224090457908e-08, "loss": 0.8144, "step": 27238 }, { "epoch": 0.9598209959741713, "grad_norm": 0.9272711277008057, "learning_rate": 8.917010574171936e-08, "loss": 0.8131, "step": 27239 }, { "epoch": 0.9598562329871296, "grad_norm": 0.9904568791389465, "learning_rate": 8.901809966868181e-08, "loss": 0.9679, "step": 27240 }, { "epoch": 0.9598914700000881, "grad_norm": 1.171720027923584, "learning_rate": 8.886622268744594e-08, "loss": 0.794, "step": 27241 }, { "epoch": 0.9599267070130465, "grad_norm": 1.4082759618759155, "learning_rate": 8.871447479998906e-08, "loss": 0.7774, "step": 27242 }, { "epoch": 0.9599619440260049, "grad_norm": 1.059457540512085, "learning_rate": 8.856285600828962e-08, "loss": 1.05, "step": 27243 }, { "epoch": 0.9599971810389634, "grad_norm": 2.5933053493499756, "learning_rate": 8.841136631432268e-08, "loss": 2.5984, "step": 27244 }, { "epoch": 0.9600324180519217, "grad_norm": 1.1477445363998413, "learning_rate": 8.826000572005889e-08, "loss": 1.1397, "step": 27245 }, { "epoch": 0.9600676550648801, "grad_norm": 5.299788951873779, "learning_rate": 8.810877422747e-08, "loss": 4.5724, "step": 27246 }, { "epoch": 0.9601028920778386, "grad_norm": 1.5809088945388794, "learning_rate": 8.795767183852888e-08, "loss": 0.6703, "step": 27247 }, { "epoch": 0.960138129090797, "grad_norm": 5.276569843292236, "learning_rate": 8.780669855520063e-08, "loss": 4.3948, "step": 27248 }, { "epoch": 0.9601733661037554, "grad_norm": 1.1855827569961548, "learning_rate": 8.765585437945256e-08, "loss": 0.6116, "step": 27249 }, { "epoch": 0.9602086031167137, "grad_norm": 2.8338212966918945, "learning_rate": 8.750513931324978e-08, "loss": 3.1516, "step": 27250 }, { "epoch": 0.9602438401296722, "grad_norm": 5.70321798324585, "learning_rate": 8.735455335855624e-08, "loss": 4.4288, "step": 27251 }, { "epoch": 0.9602790771426306, "grad_norm": 1.1484630107879639, "learning_rate": 8.720409651733152e-08, "loss": 1.0901, "step": 27252 }, { "epoch": 0.960314314155589, "grad_norm": 0.8634673953056335, "learning_rate": 8.705376879153515e-08, "loss": 0.7057, "step": 27253 }, { "epoch": 0.9603495511685475, "grad_norm": 2.171438217163086, "learning_rate": 8.690357018312778e-08, "loss": 0.9568, "step": 27254 }, { "epoch": 0.9603847881815059, "grad_norm": 9.521169662475586, "learning_rate": 8.675350069406341e-08, "loss": 3.2479, "step": 27255 }, { "epoch": 0.9604200251944642, "grad_norm": 7.114672660827637, "learning_rate": 8.660356032629713e-08, "loss": 4.3535, "step": 27256 }, { "epoch": 0.9604552622074227, "grad_norm": 1.0338850021362305, "learning_rate": 8.645374908178294e-08, "loss": 0.7959, "step": 27257 }, { "epoch": 0.9604904992203811, "grad_norm": 1.2415082454681396, "learning_rate": 8.63040669624704e-08, "loss": 0.9307, "step": 27258 }, { "epoch": 0.9605257362333395, "grad_norm": 1.12617027759552, "learning_rate": 8.615451397031127e-08, "loss": 0.7753, "step": 27259 }, { "epoch": 0.960560973246298, "grad_norm": 1.3892794847488403, "learning_rate": 8.600509010725289e-08, "loss": 0.9658, "step": 27260 }, { "epoch": 0.9605962102592563, "grad_norm": 1.0542819499969482, "learning_rate": 8.585579537523924e-08, "loss": 0.7604, "step": 27261 }, { "epoch": 0.9606314472722147, "grad_norm": 3.1167030334472656, "learning_rate": 8.570662977621769e-08, "loss": 2.551, "step": 27262 }, { "epoch": 0.9606666842851731, "grad_norm": 3.691378116607666, "learning_rate": 8.555759331212998e-08, "loss": 2.9443, "step": 27263 }, { "epoch": 0.9607019212981316, "grad_norm": 1.6437071561813354, "learning_rate": 8.540868598491791e-08, "loss": 0.7, "step": 27264 }, { "epoch": 0.96073715831109, "grad_norm": 16.978946685791016, "learning_rate": 8.525990779651993e-08, "loss": 5.0359, "step": 27265 }, { "epoch": 0.9607723953240483, "grad_norm": 1.1802928447723389, "learning_rate": 8.51112587488756e-08, "loss": 0.6328, "step": 27266 }, { "epoch": 0.9608076323370068, "grad_norm": 3.981851100921631, "learning_rate": 8.496273884392003e-08, "loss": 4.7702, "step": 27267 }, { "epoch": 0.9608428693499652, "grad_norm": 1.259345531463623, "learning_rate": 8.481434808358724e-08, "loss": 0.7351, "step": 27268 }, { "epoch": 0.9608781063629236, "grad_norm": 1.0683002471923828, "learning_rate": 8.466608646981011e-08, "loss": 0.883, "step": 27269 }, { "epoch": 0.9609133433758821, "grad_norm": 3.8359692096710205, "learning_rate": 8.451795400452045e-08, "loss": 3.1287, "step": 27270 }, { "epoch": 0.9609485803888405, "grad_norm": 3.748112916946411, "learning_rate": 8.436995068964782e-08, "loss": 3.0208, "step": 27271 }, { "epoch": 0.9609838174017988, "grad_norm": 3.7839255332946777, "learning_rate": 8.422207652711956e-08, "loss": 2.7497, "step": 27272 }, { "epoch": 0.9610190544147572, "grad_norm": 1.1429469585418701, "learning_rate": 8.40743315188619e-08, "loss": 0.9582, "step": 27273 }, { "epoch": 0.9610542914277157, "grad_norm": 2.415358543395996, "learning_rate": 8.392671566679889e-08, "loss": 2.8367, "step": 27274 }, { "epoch": 0.9610895284406741, "grad_norm": 1.3584920167922974, "learning_rate": 8.37792289728534e-08, "loss": 0.7893, "step": 27275 }, { "epoch": 0.9611247654536325, "grad_norm": 7.314153671264648, "learning_rate": 8.363187143894724e-08, "loss": 7.1366, "step": 27276 }, { "epoch": 0.961160002466591, "grad_norm": 2.9092698097229004, "learning_rate": 8.348464306699889e-08, "loss": 2.3992, "step": 27277 }, { "epoch": 0.9611952394795493, "grad_norm": 1.1334203481674194, "learning_rate": 8.333754385892567e-08, "loss": 0.8446, "step": 27278 }, { "epoch": 0.9612304764925077, "grad_norm": 1.4026421308517456, "learning_rate": 8.319057381664497e-08, "loss": 0.5625, "step": 27279 }, { "epoch": 0.9612657135054662, "grad_norm": 2.343825578689575, "learning_rate": 8.304373294206969e-08, "loss": 0.7809, "step": 27280 }, { "epoch": 0.9613009505184246, "grad_norm": 2.867710828781128, "learning_rate": 8.289702123711274e-08, "loss": 2.9621, "step": 27281 }, { "epoch": 0.961336187531383, "grad_norm": 1.0612456798553467, "learning_rate": 8.275043870368593e-08, "loss": 1.0573, "step": 27282 }, { "epoch": 0.9613714245443413, "grad_norm": 3.673450469970703, "learning_rate": 8.260398534369773e-08, "loss": 2.5095, "step": 27283 }, { "epoch": 0.9614066615572998, "grad_norm": 5.139337539672852, "learning_rate": 8.245766115905551e-08, "loss": 5.0012, "step": 27284 }, { "epoch": 0.9614418985702582, "grad_norm": 4.012135028839111, "learning_rate": 8.231146615166663e-08, "loss": 4.3359, "step": 27285 }, { "epoch": 0.9614771355832166, "grad_norm": 4.41223669052124, "learning_rate": 8.21654003234329e-08, "loss": 5.211, "step": 27286 }, { "epoch": 0.9615123725961751, "grad_norm": 2.210801124572754, "learning_rate": 8.201946367625835e-08, "loss": 0.8242, "step": 27287 }, { "epoch": 0.9615476096091334, "grad_norm": 0.9104831218719482, "learning_rate": 8.187365621204479e-08, "loss": 0.7242, "step": 27288 }, { "epoch": 0.9615828466220918, "grad_norm": 0.8524856567382812, "learning_rate": 8.172797793268849e-08, "loss": 0.6303, "step": 27289 }, { "epoch": 0.9616180836350503, "grad_norm": 2.6718556880950928, "learning_rate": 8.158242884009016e-08, "loss": 0.7194, "step": 27290 }, { "epoch": 0.9616533206480087, "grad_norm": 1.5233005285263062, "learning_rate": 8.143700893614381e-08, "loss": 1.1472, "step": 27291 }, { "epoch": 0.9616885576609671, "grad_norm": 1.0665565729141235, "learning_rate": 8.129171822274351e-08, "loss": 0.7646, "step": 27292 }, { "epoch": 0.9617237946739254, "grad_norm": 3.4227190017700195, "learning_rate": 8.114655670178217e-08, "loss": 2.7036, "step": 27293 }, { "epoch": 0.9617590316868839, "grad_norm": 4.354052543640137, "learning_rate": 8.10015243751494e-08, "loss": 2.6435, "step": 27294 }, { "epoch": 0.9617942686998423, "grad_norm": 0.9920543432235718, "learning_rate": 8.085662124473704e-08, "loss": 1.0962, "step": 27295 }, { "epoch": 0.9618295057128007, "grad_norm": 0.9287639856338501, "learning_rate": 8.07118473124291e-08, "loss": 0.8771, "step": 27296 }, { "epoch": 0.9618647427257592, "grad_norm": 0.9345591068267822, "learning_rate": 8.056720258011407e-08, "loss": 0.5805, "step": 27297 }, { "epoch": 0.9618999797387175, "grad_norm": 8.73573112487793, "learning_rate": 8.042268704967493e-08, "loss": 7.6406, "step": 27298 }, { "epoch": 0.9619352167516759, "grad_norm": 1.174176812171936, "learning_rate": 8.027830072299236e-08, "loss": 0.7513, "step": 27299 }, { "epoch": 0.9619704537646344, "grad_norm": 3.7047550678253174, "learning_rate": 8.013404360194932e-08, "loss": 3.2543, "step": 27300 }, { "epoch": 0.9620056907775928, "grad_norm": 4.494343280792236, "learning_rate": 7.99899156884243e-08, "loss": 4.6225, "step": 27301 }, { "epoch": 0.9620409277905512, "grad_norm": 1.344193458557129, "learning_rate": 7.984591698429356e-08, "loss": 0.822, "step": 27302 }, { "epoch": 0.9620761648035097, "grad_norm": 5.938990116119385, "learning_rate": 7.97020474914334e-08, "loss": 4.6601, "step": 27303 }, { "epoch": 0.962111401816468, "grad_norm": 1.0681380033493042, "learning_rate": 7.955830721171897e-08, "loss": 0.8228, "step": 27304 }, { "epoch": 0.9621466388294264, "grad_norm": 1.1646606922149658, "learning_rate": 7.941469614701991e-08, "loss": 0.8726, "step": 27305 }, { "epoch": 0.9621818758423848, "grad_norm": 1.2439229488372803, "learning_rate": 7.927121429920803e-08, "loss": 0.669, "step": 27306 }, { "epoch": 0.9622171128553433, "grad_norm": 3.6292364597320557, "learning_rate": 7.912786167015296e-08, "loss": 2.4242, "step": 27307 }, { "epoch": 0.9622523498683017, "grad_norm": 6.549838542938232, "learning_rate": 7.898463826172098e-08, "loss": 5.3413, "step": 27308 }, { "epoch": 0.96228758688126, "grad_norm": 4.585747241973877, "learning_rate": 7.884154407577837e-08, "loss": 4.8862, "step": 27309 }, { "epoch": 0.9623228238942185, "grad_norm": 1.2184078693389893, "learning_rate": 7.869857911418699e-08, "loss": 0.7177, "step": 27310 }, { "epoch": 0.9623580609071769, "grad_norm": 3.611358404159546, "learning_rate": 7.855574337881199e-08, "loss": 2.8068, "step": 27311 }, { "epoch": 0.9623932979201353, "grad_norm": 1.1715322732925415, "learning_rate": 7.841303687151081e-08, "loss": 1.1682, "step": 27312 }, { "epoch": 0.9624285349330938, "grad_norm": 3.197934865951538, "learning_rate": 7.827045959414415e-08, "loss": 3.2017, "step": 27313 }, { "epoch": 0.9624637719460521, "grad_norm": 4.636078357696533, "learning_rate": 7.812801154856941e-08, "loss": 3.0292, "step": 27314 }, { "epoch": 0.9624990089590105, "grad_norm": 1.1413092613220215, "learning_rate": 7.79856927366407e-08, "loss": 0.7537, "step": 27315 }, { "epoch": 0.9625342459719689, "grad_norm": 2.5649361610412598, "learning_rate": 7.784350316021205e-08, "loss": 2.645, "step": 27316 }, { "epoch": 0.9625694829849274, "grad_norm": 3.7005937099456787, "learning_rate": 7.770144282113534e-08, "loss": 2.6793, "step": 27317 }, { "epoch": 0.9626047199978858, "grad_norm": 5.4297895431518555, "learning_rate": 7.75595117212613e-08, "loss": 2.9213, "step": 27318 }, { "epoch": 0.9626399570108441, "grad_norm": 5.336212158203125, "learning_rate": 7.741770986243847e-08, "loss": 3.3894, "step": 27319 }, { "epoch": 0.9626751940238026, "grad_norm": 3.709364175796509, "learning_rate": 7.727603724651533e-08, "loss": 2.4522, "step": 27320 }, { "epoch": 0.962710431036761, "grad_norm": 6.004147052764893, "learning_rate": 7.713449387533378e-08, "loss": 4.8007, "step": 27321 }, { "epoch": 0.9627456680497194, "grad_norm": 1.977311611175537, "learning_rate": 7.699307975074011e-08, "loss": 0.8743, "step": 27322 }, { "epoch": 0.9627809050626779, "grad_norm": 1.0791985988616943, "learning_rate": 7.685179487457617e-08, "loss": 0.7442, "step": 27323 }, { "epoch": 0.9628161420756363, "grad_norm": 1.6248475313186646, "learning_rate": 7.671063924868161e-08, "loss": 0.9401, "step": 27324 }, { "epoch": 0.9628513790885946, "grad_norm": 1.429831624031067, "learning_rate": 7.656961287489384e-08, "loss": 1.0005, "step": 27325 }, { "epoch": 0.962886616101553, "grad_norm": 1.2740899324417114, "learning_rate": 7.642871575505139e-08, "loss": 0.7876, "step": 27326 }, { "epoch": 0.9629218531145115, "grad_norm": 1.1198233366012573, "learning_rate": 7.628794789098948e-08, "loss": 0.8977, "step": 27327 }, { "epoch": 0.9629570901274699, "grad_norm": 0.7311005592346191, "learning_rate": 7.614730928454106e-08, "loss": 0.8888, "step": 27328 }, { "epoch": 0.9629923271404283, "grad_norm": 2.1196742057800293, "learning_rate": 7.6006799937538e-08, "loss": 2.187, "step": 27329 }, { "epoch": 0.9630275641533868, "grad_norm": 3.7192819118499756, "learning_rate": 7.586641985180999e-08, "loss": 2.4827, "step": 27330 }, { "epoch": 0.9630628011663451, "grad_norm": 1.0433005094528198, "learning_rate": 7.572616902918661e-08, "loss": 1.0685, "step": 27331 }, { "epoch": 0.9630980381793035, "grad_norm": 3.9876484870910645, "learning_rate": 7.558604747149312e-08, "loss": 2.7771, "step": 27332 }, { "epoch": 0.963133275192262, "grad_norm": 5.571130752563477, "learning_rate": 7.544605518055692e-08, "loss": 3.3121, "step": 27333 }, { "epoch": 0.9631685122052204, "grad_norm": 0.9779800772666931, "learning_rate": 7.530619215819879e-08, "loss": 1.2187, "step": 27334 }, { "epoch": 0.9632037492181788, "grad_norm": 0.960975706577301, "learning_rate": 7.51664584062417e-08, "loss": 0.6602, "step": 27335 }, { "epoch": 0.9632389862311372, "grad_norm": 3.6980693340301514, "learning_rate": 7.502685392650754e-08, "loss": 2.804, "step": 27336 }, { "epoch": 0.9632742232440956, "grad_norm": 7.9172749519348145, "learning_rate": 7.488737872081153e-08, "loss": 2.9219, "step": 27337 }, { "epoch": 0.963309460257054, "grad_norm": 3.373494863510132, "learning_rate": 7.474803279097109e-08, "loss": 2.7097, "step": 27338 }, { "epoch": 0.9633446972700124, "grad_norm": 0.9865942001342773, "learning_rate": 7.460881613880256e-08, "loss": 0.7551, "step": 27339 }, { "epoch": 0.9633799342829709, "grad_norm": 0.8105805516242981, "learning_rate": 7.446972876611779e-08, "loss": 1.0297, "step": 27340 }, { "epoch": 0.9634151712959292, "grad_norm": 4.746571063995361, "learning_rate": 7.433077067472983e-08, "loss": 4.4576, "step": 27341 }, { "epoch": 0.9634504083088876, "grad_norm": 3.711620330810547, "learning_rate": 7.419194186644719e-08, "loss": 2.9379, "step": 27342 }, { "epoch": 0.9634856453218461, "grad_norm": 1.1268595457077026, "learning_rate": 7.405324234307954e-08, "loss": 0.5997, "step": 27343 }, { "epoch": 0.9635208823348045, "grad_norm": 4.967550277709961, "learning_rate": 7.391467210643322e-08, "loss": 5.0029, "step": 27344 }, { "epoch": 0.9635561193477629, "grad_norm": 2.7664809226989746, "learning_rate": 7.377623115831123e-08, "loss": 2.7419, "step": 27345 }, { "epoch": 0.9635913563607214, "grad_norm": 11.189399719238281, "learning_rate": 7.36379195005199e-08, "loss": 3.0897, "step": 27346 }, { "epoch": 0.9636265933736797, "grad_norm": 2.9411730766296387, "learning_rate": 7.349973713485781e-08, "loss": 2.4234, "step": 27347 }, { "epoch": 0.9636618303866381, "grad_norm": 3.561561107635498, "learning_rate": 7.336168406312682e-08, "loss": 3.0316, "step": 27348 }, { "epoch": 0.9636970673995965, "grad_norm": 3.0416452884674072, "learning_rate": 7.322376028712331e-08, "loss": 2.8139, "step": 27349 }, { "epoch": 0.963732304412555, "grad_norm": 1.0302995443344116, "learning_rate": 7.30859658086458e-08, "loss": 0.8732, "step": 27350 }, { "epoch": 0.9637675414255134, "grad_norm": 4.304305553436279, "learning_rate": 7.294830062948732e-08, "loss": 4.2038, "step": 27351 }, { "epoch": 0.9638027784384717, "grad_norm": 5.126289367675781, "learning_rate": 7.28107647514431e-08, "loss": 2.7011, "step": 27352 }, { "epoch": 0.9638380154514302, "grad_norm": 1.5332123041152954, "learning_rate": 7.267335817630172e-08, "loss": 0.8053, "step": 27353 }, { "epoch": 0.9638732524643886, "grad_norm": 1.393129587173462, "learning_rate": 7.253608090585507e-08, "loss": 0.7757, "step": 27354 }, { "epoch": 0.963908489477347, "grad_norm": 2.5071263313293457, "learning_rate": 7.23989329418906e-08, "loss": 0.9027, "step": 27355 }, { "epoch": 0.9639437264903055, "grad_norm": 1.7469230890274048, "learning_rate": 7.226191428619467e-08, "loss": 0.6701, "step": 27356 }, { "epoch": 0.9639789635032638, "grad_norm": 1.2201176881790161, "learning_rate": 7.21250249405514e-08, "loss": 0.8684, "step": 27357 }, { "epoch": 0.9640142005162222, "grad_norm": 3.7105133533477783, "learning_rate": 7.198826490674493e-08, "loss": 2.2379, "step": 27358 }, { "epoch": 0.9640494375291806, "grad_norm": 0.9862289428710938, "learning_rate": 7.185163418655494e-08, "loss": 0.868, "step": 27359 }, { "epoch": 0.9640846745421391, "grad_norm": 1.7682095766067505, "learning_rate": 7.171513278176335e-08, "loss": 0.8185, "step": 27360 }, { "epoch": 0.9641199115550975, "grad_norm": 0.9114950299263, "learning_rate": 7.157876069414538e-08, "loss": 0.8562, "step": 27361 }, { "epoch": 0.9641551485680558, "grad_norm": 3.9960994720458984, "learning_rate": 7.144251792547963e-08, "loss": 2.4945, "step": 27362 }, { "epoch": 0.9641903855810143, "grad_norm": 5.2069783210754395, "learning_rate": 7.130640447753911e-08, "loss": 4.0387, "step": 27363 }, { "epoch": 0.9642256225939727, "grad_norm": 6.51458740234375, "learning_rate": 7.117042035209797e-08, "loss": 4.5462, "step": 27364 }, { "epoch": 0.9642608596069311, "grad_norm": 1.0552778244018555, "learning_rate": 7.103456555092702e-08, "loss": 1.1049, "step": 27365 }, { "epoch": 0.9642960966198896, "grad_norm": 3.0159072875976562, "learning_rate": 7.089884007579484e-08, "loss": 2.4668, "step": 27366 }, { "epoch": 0.964331333632848, "grad_norm": 6.58904504776001, "learning_rate": 7.076324392847112e-08, "loss": 5.2057, "step": 27367 }, { "epoch": 0.9643665706458063, "grad_norm": 5.188148498535156, "learning_rate": 7.062777711071999e-08, "loss": 5.0191, "step": 27368 }, { "epoch": 0.9644018076587648, "grad_norm": 2.2747607231140137, "learning_rate": 7.049243962430786e-08, "loss": 2.4759, "step": 27369 }, { "epoch": 0.9644370446717232, "grad_norm": 1.400599718093872, "learning_rate": 7.035723147099549e-08, "loss": 0.6233, "step": 27370 }, { "epoch": 0.9644722816846816, "grad_norm": 1.1922335624694824, "learning_rate": 7.022215265254595e-08, "loss": 0.7226, "step": 27371 }, { "epoch": 0.96450751869764, "grad_norm": 1.3489625453948975, "learning_rate": 7.008720317071672e-08, "loss": 1.0219, "step": 27372 }, { "epoch": 0.9645427557105984, "grad_norm": 1.7659286260604858, "learning_rate": 6.995238302726636e-08, "loss": 0.9416, "step": 27373 }, { "epoch": 0.9645779927235568, "grad_norm": 1.1872830390930176, "learning_rate": 6.98176922239524e-08, "loss": 0.6619, "step": 27374 }, { "epoch": 0.9646132297365152, "grad_norm": 3.900847911834717, "learning_rate": 6.968313076252786e-08, "loss": 2.8779, "step": 27375 }, { "epoch": 0.9646484667494737, "grad_norm": 1.2602508068084717, "learning_rate": 6.954869864474356e-08, "loss": 1.1773, "step": 27376 }, { "epoch": 0.9646837037624321, "grad_norm": 5.3901190757751465, "learning_rate": 6.941439587235476e-08, "loss": 4.379, "step": 27377 }, { "epoch": 0.9647189407753904, "grad_norm": 1.255293369293213, "learning_rate": 6.928022244710675e-08, "loss": 0.6242, "step": 27378 }, { "epoch": 0.9647541777883489, "grad_norm": 6.607348442077637, "learning_rate": 6.914617837074922e-08, "loss": 4.9269, "step": 27379 }, { "epoch": 0.9647894148013073, "grad_norm": 4.150287628173828, "learning_rate": 6.901226364502856e-08, "loss": 3.1466, "step": 27380 }, { "epoch": 0.9648246518142657, "grad_norm": 5.801373481750488, "learning_rate": 6.88784782716878e-08, "loss": 3.0813, "step": 27381 }, { "epoch": 0.9648598888272241, "grad_norm": 2.4747915267944336, "learning_rate": 6.874482225246892e-08, "loss": 0.9032, "step": 27382 }, { "epoch": 0.9648951258401826, "grad_norm": 3.404449462890625, "learning_rate": 6.861129558911495e-08, "loss": 3.0454, "step": 27383 }, { "epoch": 0.9649303628531409, "grad_norm": 4.3207197189331055, "learning_rate": 6.847789828336337e-08, "loss": 2.936, "step": 27384 }, { "epoch": 0.9649655998660993, "grad_norm": 2.3309178352355957, "learning_rate": 6.834463033695171e-08, "loss": 2.7399, "step": 27385 }, { "epoch": 0.9650008368790578, "grad_norm": 2.5137288570404053, "learning_rate": 6.821149175161745e-08, "loss": 2.8767, "step": 27386 }, { "epoch": 0.9650360738920162, "grad_norm": 1.109316349029541, "learning_rate": 6.807848252909255e-08, "loss": 1.0979, "step": 27387 }, { "epoch": 0.9650713109049746, "grad_norm": 1.0609285831451416, "learning_rate": 6.794560267111117e-08, "loss": 0.8971, "step": 27388 }, { "epoch": 0.965106547917933, "grad_norm": 4.375152111053467, "learning_rate": 6.781285217940303e-08, "loss": 4.9259, "step": 27389 }, { "epoch": 0.9651417849308914, "grad_norm": 0.9608989953994751, "learning_rate": 6.768023105569898e-08, "loss": 0.7175, "step": 27390 }, { "epoch": 0.9651770219438498, "grad_norm": 0.9108692407608032, "learning_rate": 6.75477393017232e-08, "loss": 0.952, "step": 27391 }, { "epoch": 0.9652122589568082, "grad_norm": 4.536830425262451, "learning_rate": 6.74153769192043e-08, "loss": 4.7917, "step": 27392 }, { "epoch": 0.9652474959697667, "grad_norm": 1.0715686082839966, "learning_rate": 6.728314390986535e-08, "loss": 0.9201, "step": 27393 }, { "epoch": 0.965282732982725, "grad_norm": 1.1923733949661255, "learning_rate": 6.715104027542829e-08, "loss": 0.7538, "step": 27394 }, { "epoch": 0.9653179699956834, "grad_norm": 3.58310604095459, "learning_rate": 6.701906601761398e-08, "loss": 2.2474, "step": 27395 }, { "epoch": 0.9653532070086419, "grad_norm": 1.1419020891189575, "learning_rate": 6.688722113814217e-08, "loss": 0.9527, "step": 27396 }, { "epoch": 0.9653884440216003, "grad_norm": 1.083769679069519, "learning_rate": 6.675550563873035e-08, "loss": 0.9714, "step": 27397 }, { "epoch": 0.9654236810345587, "grad_norm": 1.232194185256958, "learning_rate": 6.66239195210916e-08, "loss": 0.6951, "step": 27398 }, { "epoch": 0.9654589180475172, "grad_norm": 1.896623969078064, "learning_rate": 6.649246278694232e-08, "loss": 2.4483, "step": 27399 }, { "epoch": 0.9654941550604755, "grad_norm": 1.379898190498352, "learning_rate": 6.63611354379945e-08, "loss": 0.722, "step": 27400 }, { "epoch": 0.9655293920734339, "grad_norm": 3.6347649097442627, "learning_rate": 6.622993747595785e-08, "loss": 2.9699, "step": 27401 }, { "epoch": 0.9655646290863924, "grad_norm": 1.2602322101593018, "learning_rate": 6.609886890254214e-08, "loss": 1.0873, "step": 27402 }, { "epoch": 0.9655998660993508, "grad_norm": 6.538971424102783, "learning_rate": 6.596792971945376e-08, "loss": 4.7036, "step": 27403 }, { "epoch": 0.9656351031123092, "grad_norm": 1.2453813552856445, "learning_rate": 6.583711992839804e-08, "loss": 0.8646, "step": 27404 }, { "epoch": 0.9656703401252675, "grad_norm": 6.475638389587402, "learning_rate": 6.570643953107913e-08, "loss": 3.1574, "step": 27405 }, { "epoch": 0.965705577138226, "grad_norm": 5.275894641876221, "learning_rate": 6.557588852920016e-08, "loss": 5.3239, "step": 27406 }, { "epoch": 0.9657408141511844, "grad_norm": 1.008612871170044, "learning_rate": 6.544546692445975e-08, "loss": 0.9916, "step": 27407 }, { "epoch": 0.9657760511641428, "grad_norm": 1.0963765382766724, "learning_rate": 6.531517471855765e-08, "loss": 0.7569, "step": 27408 }, { "epoch": 0.9658112881771013, "grad_norm": 2.1379644870758057, "learning_rate": 6.518501191319249e-08, "loss": 2.6481, "step": 27409 }, { "epoch": 0.9658465251900596, "grad_norm": 2.446960926055908, "learning_rate": 6.505497851005626e-08, "loss": 2.9022, "step": 27410 }, { "epoch": 0.965881762203018, "grad_norm": 5.886109352111816, "learning_rate": 6.492507451084428e-08, "loss": 5.0252, "step": 27411 }, { "epoch": 0.9659169992159765, "grad_norm": 7.736203193664551, "learning_rate": 6.479529991724964e-08, "loss": 3.1344, "step": 27412 }, { "epoch": 0.9659522362289349, "grad_norm": 3.6347732543945312, "learning_rate": 6.46656547309621e-08, "loss": 2.8491, "step": 27413 }, { "epoch": 0.9659874732418933, "grad_norm": 3.0712215900421143, "learning_rate": 6.453613895366917e-08, "loss": 3.0427, "step": 27414 }, { "epoch": 0.9660227102548516, "grad_norm": 1.3880407810211182, "learning_rate": 6.440675258705953e-08, "loss": 0.9423, "step": 27415 }, { "epoch": 0.9660579472678101, "grad_norm": 1.6616095304489136, "learning_rate": 6.427749563281627e-08, "loss": 0.7128, "step": 27416 }, { "epoch": 0.9660931842807685, "grad_norm": 1.177476167678833, "learning_rate": 6.41483680926247e-08, "loss": 1.0305, "step": 27417 }, { "epoch": 0.9661284212937269, "grad_norm": 1.313270926475525, "learning_rate": 6.401936996816571e-08, "loss": 0.715, "step": 27418 }, { "epoch": 0.9661636583066854, "grad_norm": 3.9836738109588623, "learning_rate": 6.389050126112128e-08, "loss": 4.8866, "step": 27419 }, { "epoch": 0.9661988953196438, "grad_norm": 1.208105444908142, "learning_rate": 6.376176197316897e-08, "loss": 0.6607, "step": 27420 }, { "epoch": 0.9662341323326021, "grad_norm": 1.195014476776123, "learning_rate": 6.363315210598409e-08, "loss": 0.7242, "step": 27421 }, { "epoch": 0.9662693693455606, "grad_norm": 5.867166996002197, "learning_rate": 6.35046716612442e-08, "loss": 2.5104, "step": 27422 }, { "epoch": 0.966304606358519, "grad_norm": 1.2878875732421875, "learning_rate": 6.337632064062016e-08, "loss": 0.7698, "step": 27423 }, { "epoch": 0.9663398433714774, "grad_norm": 1.1356490850448608, "learning_rate": 6.324809904578622e-08, "loss": 0.8281, "step": 27424 }, { "epoch": 0.9663750803844358, "grad_norm": 9.265008926391602, "learning_rate": 6.312000687841214e-08, "loss": 4.7547, "step": 27425 }, { "epoch": 0.9664103173973942, "grad_norm": 1.199884057044983, "learning_rate": 6.299204414016435e-08, "loss": 0.805, "step": 27426 }, { "epoch": 0.9664455544103526, "grad_norm": 2.342837333679199, "learning_rate": 6.286421083271266e-08, "loss": 2.3821, "step": 27427 }, { "epoch": 0.966480791423311, "grad_norm": 7.481076717376709, "learning_rate": 6.273650695771904e-08, "loss": 5.5153, "step": 27428 }, { "epoch": 0.9665160284362695, "grad_norm": 2.3250207901000977, "learning_rate": 6.260893251684886e-08, "loss": 2.1241, "step": 27429 }, { "epoch": 0.9665512654492279, "grad_norm": 4.027683258056641, "learning_rate": 6.248148751176297e-08, "loss": 4.7063, "step": 27430 }, { "epoch": 0.9665865024621862, "grad_norm": 1.4085383415222168, "learning_rate": 6.235417194412119e-08, "loss": 0.8595, "step": 27431 }, { "epoch": 0.9666217394751447, "grad_norm": 1.125563621520996, "learning_rate": 6.222698581558328e-08, "loss": 0.7676, "step": 27432 }, { "epoch": 0.9666569764881031, "grad_norm": 7.094230651855469, "learning_rate": 6.20999291278035e-08, "loss": 4.7346, "step": 27433 }, { "epoch": 0.9666922135010615, "grad_norm": 3.4715769290924072, "learning_rate": 6.197300188243826e-08, "loss": 3.2985, "step": 27434 }, { "epoch": 0.96672745051402, "grad_norm": 4.182443618774414, "learning_rate": 6.184620408114072e-08, "loss": 0.9073, "step": 27435 }, { "epoch": 0.9667626875269784, "grad_norm": 2.781895875930786, "learning_rate": 6.171953572556288e-08, "loss": 2.8715, "step": 27436 }, { "epoch": 0.9667979245399367, "grad_norm": 1.5215024948120117, "learning_rate": 6.159299681735343e-08, "loss": 0.8979, "step": 27437 }, { "epoch": 0.9668331615528951, "grad_norm": 1.4201865196228027, "learning_rate": 6.146658735816103e-08, "loss": 0.9682, "step": 27438 }, { "epoch": 0.9668683985658536, "grad_norm": 3.9825661182403564, "learning_rate": 6.134030734963214e-08, "loss": 3.0518, "step": 27439 }, { "epoch": 0.966903635578812, "grad_norm": 5.168750762939453, "learning_rate": 6.121415679341214e-08, "loss": 5.4119, "step": 27440 }, { "epoch": 0.9669388725917704, "grad_norm": 2.513087272644043, "learning_rate": 6.108813569114414e-08, "loss": 2.7548, "step": 27441 }, { "epoch": 0.9669741096047288, "grad_norm": 8.325742721557617, "learning_rate": 6.096224404446904e-08, "loss": 3.1951, "step": 27442 }, { "epoch": 0.9670093466176872, "grad_norm": 5.752547740936279, "learning_rate": 6.083648185502666e-08, "loss": 3.3653, "step": 27443 }, { "epoch": 0.9670445836306456, "grad_norm": 5.3993449211120605, "learning_rate": 6.071084912445458e-08, "loss": 3.0684, "step": 27444 }, { "epoch": 0.9670798206436041, "grad_norm": 2.151609420776367, "learning_rate": 6.058534585439036e-08, "loss": 0.6673, "step": 27445 }, { "epoch": 0.9671150576565625, "grad_norm": 1.2328463792800903, "learning_rate": 6.045997204646715e-08, "loss": 0.9453, "step": 27446 }, { "epoch": 0.9671502946695208, "grad_norm": 2.6192259788513184, "learning_rate": 6.03347277023203e-08, "loss": 3.1543, "step": 27447 }, { "epoch": 0.9671855316824792, "grad_norm": 3.8647382259368896, "learning_rate": 6.020961282357852e-08, "loss": 2.822, "step": 27448 }, { "epoch": 0.9672207686954377, "grad_norm": 4.5678019523620605, "learning_rate": 6.00846274118727e-08, "loss": 3.2864, "step": 27449 }, { "epoch": 0.9672560057083961, "grad_norm": 1.2233282327651978, "learning_rate": 5.995977146883269e-08, "loss": 0.8614, "step": 27450 }, { "epoch": 0.9672912427213545, "grad_norm": 4.764306545257568, "learning_rate": 5.983504499608051e-08, "loss": 4.5738, "step": 27451 }, { "epoch": 0.967326479734313, "grad_norm": 2.0859155654907227, "learning_rate": 5.971044799524373e-08, "loss": 2.1962, "step": 27452 }, { "epoch": 0.9673617167472713, "grad_norm": 0.799100935459137, "learning_rate": 5.9585980467945504e-08, "loss": 0.8081, "step": 27453 }, { "epoch": 0.9673969537602297, "grad_norm": 4.279668807983398, "learning_rate": 5.946164241580454e-08, "loss": 0.5479, "step": 27454 }, { "epoch": 0.9674321907731882, "grad_norm": 2.6773743629455566, "learning_rate": 5.9337433840442885e-08, "loss": 3.0665, "step": 27455 }, { "epoch": 0.9674674277861466, "grad_norm": 4.630875587463379, "learning_rate": 5.9213354743477005e-08, "loss": 5.0665, "step": 27456 }, { "epoch": 0.967502664799105, "grad_norm": 3.134902238845825, "learning_rate": 5.908940512652339e-08, "loss": 3.2998, "step": 27457 }, { "epoch": 0.9675379018120633, "grad_norm": 1.221815824508667, "learning_rate": 5.8965584991197425e-08, "loss": 1.1842, "step": 27458 }, { "epoch": 0.9675731388250218, "grad_norm": 5.144370079040527, "learning_rate": 5.884189433911003e-08, "loss": 4.611, "step": 27459 }, { "epoch": 0.9676083758379802, "grad_norm": 6.437257289886475, "learning_rate": 5.8718333171873256e-08, "loss": 3.1376, "step": 27460 }, { "epoch": 0.9676436128509386, "grad_norm": 5.361014366149902, "learning_rate": 5.8594901491096926e-08, "loss": 4.5586, "step": 27461 }, { "epoch": 0.9676788498638971, "grad_norm": 2.952613592147827, "learning_rate": 5.8471599298388634e-08, "loss": 2.7509, "step": 27462 }, { "epoch": 0.9677140868768555, "grad_norm": 4.524019718170166, "learning_rate": 5.834842659535378e-08, "loss": 2.9429, "step": 27463 }, { "epoch": 0.9677493238898138, "grad_norm": 1.307739019393921, "learning_rate": 5.8225383383596625e-08, "loss": 1.1042, "step": 27464 }, { "epoch": 0.9677845609027723, "grad_norm": 1.2849371433258057, "learning_rate": 5.810246966472033e-08, "loss": 0.7276, "step": 27465 }, { "epoch": 0.9678197979157307, "grad_norm": 4.90419340133667, "learning_rate": 5.797968544032695e-08, "loss": 4.6188, "step": 27466 }, { "epoch": 0.9678550349286891, "grad_norm": 0.9471259117126465, "learning_rate": 5.7857030712012987e-08, "loss": 0.8147, "step": 27467 }, { "epoch": 0.9678902719416476, "grad_norm": 0.8214462995529175, "learning_rate": 5.7734505481378265e-08, "loss": 0.6487, "step": 27468 }, { "epoch": 0.9679255089546059, "grad_norm": 12.002254486083984, "learning_rate": 5.7612109750018186e-08, "loss": 6.5188, "step": 27469 }, { "epoch": 0.9679607459675643, "grad_norm": 3.185865879058838, "learning_rate": 5.748984351952702e-08, "loss": 2.9939, "step": 27470 }, { "epoch": 0.9679959829805227, "grad_norm": 4.710974216461182, "learning_rate": 5.7367706791496835e-08, "loss": 4.7303, "step": 27471 }, { "epoch": 0.9680312199934812, "grad_norm": 2.4503793716430664, "learning_rate": 5.7245699567518575e-08, "loss": 3.0483, "step": 27472 }, { "epoch": 0.9680664570064396, "grad_norm": 3.5152499675750732, "learning_rate": 5.712382184918208e-08, "loss": 2.9001, "step": 27473 }, { "epoch": 0.9681016940193979, "grad_norm": 1.0925661325454712, "learning_rate": 5.700207363807386e-08, "loss": 1.1062, "step": 27474 }, { "epoch": 0.9681369310323564, "grad_norm": 1.1299717426300049, "learning_rate": 5.688045493577932e-08, "loss": 0.9204, "step": 27475 }, { "epoch": 0.9681721680453148, "grad_norm": 4.371404647827148, "learning_rate": 5.675896574388495e-08, "loss": 4.238, "step": 27476 }, { "epoch": 0.9682074050582732, "grad_norm": 4.191446304321289, "learning_rate": 5.663760606397173e-08, "loss": 2.7447, "step": 27477 }, { "epoch": 0.9682426420712317, "grad_norm": 3.473148822784424, "learning_rate": 5.6516375897619494e-08, "loss": 2.8784, "step": 27478 }, { "epoch": 0.96827787908419, "grad_norm": 1.261649250984192, "learning_rate": 5.6395275246408086e-08, "loss": 0.7497, "step": 27479 }, { "epoch": 0.9683131160971484, "grad_norm": 3.3063912391662598, "learning_rate": 5.627430411191404e-08, "loss": 2.9693, "step": 27480 }, { "epoch": 0.9683483531101068, "grad_norm": 0.9592159986495972, "learning_rate": 5.615346249571496e-08, "loss": 0.6177, "step": 27481 }, { "epoch": 0.9683835901230653, "grad_norm": 7.262935161590576, "learning_rate": 5.603275039938405e-08, "loss": 4.4474, "step": 27482 }, { "epoch": 0.9684188271360237, "grad_norm": 1.745512843132019, "learning_rate": 5.591216782449227e-08, "loss": 0.7967, "step": 27483 }, { "epoch": 0.968454064148982, "grad_norm": 2.7837491035461426, "learning_rate": 5.5791714772611695e-08, "loss": 2.881, "step": 27484 }, { "epoch": 0.9684893011619405, "grad_norm": 3.683807134628296, "learning_rate": 5.567139124530996e-08, "loss": 2.6912, "step": 27485 }, { "epoch": 0.9685245381748989, "grad_norm": 5.175216197967529, "learning_rate": 5.555119724415581e-08, "loss": 4.7215, "step": 27486 }, { "epoch": 0.9685597751878573, "grad_norm": 6.21466064453125, "learning_rate": 5.543113277071355e-08, "loss": 5.5137, "step": 27487 }, { "epoch": 0.9685950122008158, "grad_norm": 1.8110235929489136, "learning_rate": 5.531119782654859e-08, "loss": 0.8753, "step": 27488 }, { "epoch": 0.9686302492137742, "grad_norm": 1.1003022193908691, "learning_rate": 5.51913924132208e-08, "loss": 0.6375, "step": 27489 }, { "epoch": 0.9686654862267325, "grad_norm": 1.2877662181854248, "learning_rate": 5.507171653229226e-08, "loss": 0.6868, "step": 27490 }, { "epoch": 0.9687007232396909, "grad_norm": 0.9680860042572021, "learning_rate": 5.495217018532062e-08, "loss": 0.9868, "step": 27491 }, { "epoch": 0.9687359602526494, "grad_norm": 1.0109901428222656, "learning_rate": 5.4832753373865734e-08, "loss": 0.635, "step": 27492 }, { "epoch": 0.9687711972656078, "grad_norm": 2.8804659843444824, "learning_rate": 5.4713466099479696e-08, "loss": 2.824, "step": 27493 }, { "epoch": 0.9688064342785662, "grad_norm": 4.198779582977295, "learning_rate": 5.459430836371793e-08, "loss": 3.1761, "step": 27494 }, { "epoch": 0.9688416712915247, "grad_norm": 2.936493158340454, "learning_rate": 5.4475280168131414e-08, "loss": 2.664, "step": 27495 }, { "epoch": 0.968876908304483, "grad_norm": 5.575368881225586, "learning_rate": 5.435638151427225e-08, "loss": 3.1406, "step": 27496 }, { "epoch": 0.9689121453174414, "grad_norm": 0.749663770198822, "learning_rate": 5.4237612403688075e-08, "loss": 0.8738, "step": 27497 }, { "epoch": 0.9689473823303999, "grad_norm": 3.0221071243286133, "learning_rate": 5.4118972837925445e-08, "loss": 2.4795, "step": 27498 }, { "epoch": 0.9689826193433583, "grad_norm": 1.3361579179763794, "learning_rate": 5.4000462818529775e-08, "loss": 1.2162, "step": 27499 }, { "epoch": 0.9690178563563167, "grad_norm": 1.153509259223938, "learning_rate": 5.388208234704539e-08, "loss": 0.8424, "step": 27500 }, { "epoch": 0.969053093369275, "grad_norm": 4.293943881988525, "learning_rate": 5.376383142501329e-08, "loss": 4.2071, "step": 27501 }, { "epoch": 0.9690883303822335, "grad_norm": 0.9872503280639648, "learning_rate": 5.364571005397556e-08, "loss": 1.0169, "step": 27502 }, { "epoch": 0.9691235673951919, "grad_norm": 2.856827735900879, "learning_rate": 5.3527718235468764e-08, "loss": 2.8023, "step": 27503 }, { "epoch": 0.9691588044081503, "grad_norm": 4.637943744659424, "learning_rate": 5.340985597103055e-08, "loss": 4.7137, "step": 27504 }, { "epoch": 0.9691940414211088, "grad_norm": 4.008926868438721, "learning_rate": 5.329212326219524e-08, "loss": 2.6047, "step": 27505 }, { "epoch": 0.9692292784340671, "grad_norm": 5.3361406326293945, "learning_rate": 5.317452011049717e-08, "loss": 2.9857, "step": 27506 }, { "epoch": 0.9692645154470255, "grad_norm": 1.4501756429672241, "learning_rate": 5.305704651746846e-08, "loss": 0.7164, "step": 27507 }, { "epoch": 0.969299752459984, "grad_norm": 0.98932284116745, "learning_rate": 5.293970248463898e-08, "loss": 0.993, "step": 27508 }, { "epoch": 0.9693349894729424, "grad_norm": 2.587898015975952, "learning_rate": 5.282248801353529e-08, "loss": 2.6644, "step": 27509 }, { "epoch": 0.9693702264859008, "grad_norm": 5.180325508117676, "learning_rate": 5.270540310568728e-08, "loss": 2.8742, "step": 27510 }, { "epoch": 0.9694054634988593, "grad_norm": 6.866116523742676, "learning_rate": 5.258844776261818e-08, "loss": 3.0322, "step": 27511 }, { "epoch": 0.9694407005118176, "grad_norm": 1.3889721632003784, "learning_rate": 5.247162198585232e-08, "loss": 0.8874, "step": 27512 }, { "epoch": 0.969475937524776, "grad_norm": 2.500284194946289, "learning_rate": 5.235492577690959e-08, "loss": 0.9925, "step": 27513 }, { "epoch": 0.9695111745377344, "grad_norm": 4.053359031677246, "learning_rate": 5.223835913731212e-08, "loss": 3.0109, "step": 27514 }, { "epoch": 0.9695464115506929, "grad_norm": 1.1983236074447632, "learning_rate": 5.212192206857647e-08, "loss": 0.8857, "step": 27515 }, { "epoch": 0.9695816485636513, "grad_norm": 4.2008843421936035, "learning_rate": 5.2005614572220306e-08, "loss": 2.9761, "step": 27516 }, { "epoch": 0.9696168855766096, "grad_norm": 4.090692043304443, "learning_rate": 5.188943664975687e-08, "loss": 5.0842, "step": 27517 }, { "epoch": 0.9696521225895681, "grad_norm": 1.0804672241210938, "learning_rate": 5.177338830270273e-08, "loss": 0.6409, "step": 27518 }, { "epoch": 0.9696873596025265, "grad_norm": 1.232163906097412, "learning_rate": 5.1657469532565564e-08, "loss": 0.8503, "step": 27519 }, { "epoch": 0.9697225966154849, "grad_norm": 3.572514057159424, "learning_rate": 5.15416803408586e-08, "loss": 3.1086, "step": 27520 }, { "epoch": 0.9697578336284434, "grad_norm": 1.8727118968963623, "learning_rate": 5.142602072908731e-08, "loss": 2.2473, "step": 27521 }, { "epoch": 0.9697930706414017, "grad_norm": 10.459906578063965, "learning_rate": 5.131049069876048e-08, "loss": 3.1774, "step": 27522 }, { "epoch": 0.9698283076543601, "grad_norm": 1.2188706398010254, "learning_rate": 5.1195090251381364e-08, "loss": 1.1005, "step": 27523 }, { "epoch": 0.9698635446673185, "grad_norm": 3.344299077987671, "learning_rate": 5.107981938845208e-08, "loss": 3.4585, "step": 27524 }, { "epoch": 0.969898781680277, "grad_norm": 9.507099151611328, "learning_rate": 5.0964678111476984e-08, "loss": 2.9829, "step": 27525 }, { "epoch": 0.9699340186932354, "grad_norm": 4.067314147949219, "learning_rate": 5.084966642195488e-08, "loss": 2.7354, "step": 27526 }, { "epoch": 0.9699692557061937, "grad_norm": 3.2081384658813477, "learning_rate": 5.073478432138235e-08, "loss": 2.7306, "step": 27527 }, { "epoch": 0.9700044927191522, "grad_norm": 1.9194451570510864, "learning_rate": 5.062003181125596e-08, "loss": 0.7729, "step": 27528 }, { "epoch": 0.9700397297321106, "grad_norm": 1.0280084609985352, "learning_rate": 5.0505408893071205e-08, "loss": 0.9962, "step": 27529 }, { "epoch": 0.970074966745069, "grad_norm": 1.2291375398635864, "learning_rate": 5.039091556832132e-08, "loss": 1.108, "step": 27530 }, { "epoch": 0.9701102037580275, "grad_norm": 1.6633495092391968, "learning_rate": 5.0276551838497337e-08, "loss": 0.8218, "step": 27531 }, { "epoch": 0.9701454407709859, "grad_norm": 1.2742973566055298, "learning_rate": 5.016231770508695e-08, "loss": 0.9421, "step": 27532 }, { "epoch": 0.9701806777839442, "grad_norm": 0.8606592416763306, "learning_rate": 5.004821316958231e-08, "loss": 0.837, "step": 27533 }, { "epoch": 0.9702159147969026, "grad_norm": 1.9296848773956299, "learning_rate": 4.9934238233465546e-08, "loss": 0.8859, "step": 27534 }, { "epoch": 0.9702511518098611, "grad_norm": 5.7086310386657715, "learning_rate": 4.982039289822327e-08, "loss": 7.3311, "step": 27535 }, { "epoch": 0.9702863888228195, "grad_norm": 2.4604928493499756, "learning_rate": 4.970667716533761e-08, "loss": 0.865, "step": 27536 }, { "epoch": 0.9703216258357779, "grad_norm": 2.8669769763946533, "learning_rate": 4.959309103628962e-08, "loss": 2.5975, "step": 27537 }, { "epoch": 0.9703568628487363, "grad_norm": 1.274809718132019, "learning_rate": 4.947963451256033e-08, "loss": 0.8341, "step": 27538 }, { "epoch": 0.9703920998616947, "grad_norm": 1.4933980703353882, "learning_rate": 4.936630759562522e-08, "loss": 0.7966, "step": 27539 }, { "epoch": 0.9704273368746531, "grad_norm": 11.050261497497559, "learning_rate": 4.9253110286962004e-08, "loss": 2.7631, "step": 27540 }, { "epoch": 0.9704625738876116, "grad_norm": 4.457019329071045, "learning_rate": 4.9140042588043944e-08, "loss": 4.3934, "step": 27541 }, { "epoch": 0.97049781090057, "grad_norm": 1.8587491512298584, "learning_rate": 4.9027104500345425e-08, "loss": 0.9185, "step": 27542 }, { "epoch": 0.9705330479135283, "grad_norm": 4.524722099304199, "learning_rate": 4.8914296025335264e-08, "loss": 4.065, "step": 27543 }, { "epoch": 0.9705682849264868, "grad_norm": 1.2475935220718384, "learning_rate": 4.8801617164484504e-08, "loss": 1.1052, "step": 27544 }, { "epoch": 0.9706035219394452, "grad_norm": 4.102880477905273, "learning_rate": 4.8689067919259756e-08, "loss": 2.9453, "step": 27545 }, { "epoch": 0.9706387589524036, "grad_norm": 1.6988013982772827, "learning_rate": 4.8576648291127625e-08, "loss": 1.0643, "step": 27546 }, { "epoch": 0.970673995965362, "grad_norm": 3.841245412826538, "learning_rate": 4.846435828155249e-08, "loss": 2.9321, "step": 27547 }, { "epoch": 0.9707092329783205, "grad_norm": 2.8056840896606445, "learning_rate": 4.835219789199652e-08, "loss": 4.4596, "step": 27548 }, { "epoch": 0.9707444699912788, "grad_norm": 3.975374937057495, "learning_rate": 4.824016712392188e-08, "loss": 3.0424, "step": 27549 }, { "epoch": 0.9707797070042372, "grad_norm": 4.767127513885498, "learning_rate": 4.812826597878628e-08, "loss": 3.1616, "step": 27550 }, { "epoch": 0.9708149440171957, "grad_norm": 0.950088620185852, "learning_rate": 4.801649445804635e-08, "loss": 0.8235, "step": 27551 }, { "epoch": 0.9708501810301541, "grad_norm": 1.4643375873565674, "learning_rate": 4.790485256316091e-08, "loss": 0.9553, "step": 27552 }, { "epoch": 0.9708854180431125, "grad_norm": 1.6774882078170776, "learning_rate": 4.7793340295581027e-08, "loss": 0.7255, "step": 27553 }, { "epoch": 0.970920655056071, "grad_norm": 1.7017335891723633, "learning_rate": 4.768195765676109e-08, "loss": 0.9223, "step": 27554 }, { "epoch": 0.9709558920690293, "grad_norm": 5.164463043212891, "learning_rate": 4.757070464815217e-08, "loss": 5.1019, "step": 27555 }, { "epoch": 0.9709911290819877, "grad_norm": 3.9672722816467285, "learning_rate": 4.745958127120198e-08, "loss": 3.2068, "step": 27556 }, { "epoch": 0.9710263660949461, "grad_norm": 4.847589492797852, "learning_rate": 4.734858752735827e-08, "loss": 3.3629, "step": 27557 }, { "epoch": 0.9710616031079046, "grad_norm": 1.4328845739364624, "learning_rate": 4.7237723418067647e-08, "loss": 0.8451, "step": 27558 }, { "epoch": 0.971096840120863, "grad_norm": 2.9120726585388184, "learning_rate": 4.71269889447723e-08, "loss": 2.8807, "step": 27559 }, { "epoch": 0.9711320771338213, "grad_norm": 1.2110847234725952, "learning_rate": 4.7016384108916624e-08, "loss": 0.6393, "step": 27560 }, { "epoch": 0.9711673141467798, "grad_norm": 2.260193109512329, "learning_rate": 4.690590891194058e-08, "loss": 2.2174, "step": 27561 }, { "epoch": 0.9712025511597382, "grad_norm": 1.051632285118103, "learning_rate": 4.67955633552819e-08, "loss": 0.7615, "step": 27562 }, { "epoch": 0.9712377881726966, "grad_norm": 0.9626332521438599, "learning_rate": 4.668534744037945e-08, "loss": 0.794, "step": 27563 }, { "epoch": 0.9712730251856551, "grad_norm": 1.2635393142700195, "learning_rate": 4.657526116866873e-08, "loss": 0.7882, "step": 27564 }, { "epoch": 0.9713082621986134, "grad_norm": 0.8746996521949768, "learning_rate": 4.6465304541583043e-08, "loss": 0.6738, "step": 27565 }, { "epoch": 0.9713434992115718, "grad_norm": 0.9554738998413086, "learning_rate": 4.635547756055348e-08, "loss": 0.9112, "step": 27566 }, { "epoch": 0.9713787362245302, "grad_norm": 1.1098753213882446, "learning_rate": 4.624578022701221e-08, "loss": 0.9048, "step": 27567 }, { "epoch": 0.9714139732374887, "grad_norm": 1.092813491821289, "learning_rate": 4.613621254238809e-08, "loss": 0.7543, "step": 27568 }, { "epoch": 0.9714492102504471, "grad_norm": 1.414018154144287, "learning_rate": 4.6026774508107775e-08, "loss": 0.59, "step": 27569 }, { "epoch": 0.9714844472634054, "grad_norm": 3.6727755069732666, "learning_rate": 4.5917466125595666e-08, "loss": 2.9055, "step": 27570 }, { "epoch": 0.9715196842763639, "grad_norm": 1.6147313117980957, "learning_rate": 4.5808287396277295e-08, "loss": 1.0545, "step": 27571 }, { "epoch": 0.9715549212893223, "grad_norm": 4.268787860870361, "learning_rate": 4.569923832157375e-08, "loss": 3.0946, "step": 27572 }, { "epoch": 0.9715901583022807, "grad_norm": 0.8585010170936584, "learning_rate": 4.5590318902904996e-08, "loss": 0.8239, "step": 27573 }, { "epoch": 0.9716253953152392, "grad_norm": 1.4146074056625366, "learning_rate": 4.548152914169102e-08, "loss": 1.1235, "step": 27574 }, { "epoch": 0.9716606323281975, "grad_norm": 1.3384891748428345, "learning_rate": 4.537286903934845e-08, "loss": 0.9334, "step": 27575 }, { "epoch": 0.9716958693411559, "grad_norm": 3.204087257385254, "learning_rate": 4.526433859729174e-08, "loss": 2.9923, "step": 27576 }, { "epoch": 0.9717311063541144, "grad_norm": 4.780932426452637, "learning_rate": 4.515593781693528e-08, "loss": 2.9911, "step": 27577 }, { "epoch": 0.9717663433670728, "grad_norm": 3.540088176727295, "learning_rate": 4.504766669969018e-08, "loss": 2.7693, "step": 27578 }, { "epoch": 0.9718015803800312, "grad_norm": 4.203186988830566, "learning_rate": 4.493952524696754e-08, "loss": 3.2617, "step": 27579 }, { "epoch": 0.9718368173929895, "grad_norm": 1.3714004755020142, "learning_rate": 4.4831513460176226e-08, "loss": 1.0125, "step": 27580 }, { "epoch": 0.971872054405948, "grad_norm": 2.8431737422943115, "learning_rate": 4.472363134072177e-08, "loss": 2.7945, "step": 27581 }, { "epoch": 0.9719072914189064, "grad_norm": 2.4625515937805176, "learning_rate": 4.4615878890010845e-08, "loss": 2.3098, "step": 27582 }, { "epoch": 0.9719425284318648, "grad_norm": 4.971862316131592, "learning_rate": 4.450825610944565e-08, "loss": 2.7421, "step": 27583 }, { "epoch": 0.9719777654448233, "grad_norm": 3.3329358100891113, "learning_rate": 4.44007630004295e-08, "loss": 2.6256, "step": 27584 }, { "epoch": 0.9720130024577817, "grad_norm": 3.843031167984009, "learning_rate": 4.429339956436018e-08, "loss": 1.0143, "step": 27585 }, { "epoch": 0.97204823947074, "grad_norm": 2.149179220199585, "learning_rate": 4.418616580263768e-08, "loss": 0.8399, "step": 27586 }, { "epoch": 0.9720834764836985, "grad_norm": 2.9395053386688232, "learning_rate": 4.4079061716659764e-08, "loss": 2.7699, "step": 27587 }, { "epoch": 0.9721187134966569, "grad_norm": 2.9900636672973633, "learning_rate": 4.397208730781976e-08, "loss": 2.7244, "step": 27588 }, { "epoch": 0.9721539505096153, "grad_norm": 3.7733523845672607, "learning_rate": 4.3865242577511014e-08, "loss": 3.1476, "step": 27589 }, { "epoch": 0.9721891875225737, "grad_norm": 1.5878100395202637, "learning_rate": 4.375852752712573e-08, "loss": 0.891, "step": 27590 }, { "epoch": 0.9722244245355322, "grad_norm": 1.273685097694397, "learning_rate": 4.365194215805391e-08, "loss": 0.6954, "step": 27591 }, { "epoch": 0.9722596615484905, "grad_norm": 4.627991199493408, "learning_rate": 4.354548647168444e-08, "loss": 4.5486, "step": 27592 }, { "epoch": 0.9722948985614489, "grad_norm": 3.21398663520813, "learning_rate": 4.343916046940288e-08, "loss": 2.8845, "step": 27593 }, { "epoch": 0.9723301355744074, "grad_norm": 29.43288803100586, "learning_rate": 4.333296415259369e-08, "loss": 4.4904, "step": 27594 }, { "epoch": 0.9723653725873658, "grad_norm": 1.8350067138671875, "learning_rate": 4.322689752264131e-08, "loss": 0.8659, "step": 27595 }, { "epoch": 0.9724006096003242, "grad_norm": 6.637241840362549, "learning_rate": 4.3120960580927965e-08, "loss": 4.5214, "step": 27596 }, { "epoch": 0.9724358466132826, "grad_norm": 4.120713710784912, "learning_rate": 4.3015153328831436e-08, "loss": 2.8326, "step": 27597 }, { "epoch": 0.972471083626241, "grad_norm": 1.3250234127044678, "learning_rate": 4.290947576773063e-08, "loss": 0.958, "step": 27598 }, { "epoch": 0.9725063206391994, "grad_norm": 3.9447507858276367, "learning_rate": 4.280392789900334e-08, "loss": 3.209, "step": 27599 }, { "epoch": 0.9725415576521578, "grad_norm": 1.2910960912704468, "learning_rate": 4.269850972402178e-08, "loss": 0.886, "step": 27600 }, { "epoch": 0.9725767946651163, "grad_norm": 3.811091423034668, "learning_rate": 4.259322124416154e-08, "loss": 2.6709, "step": 27601 }, { "epoch": 0.9726120316780746, "grad_norm": 3.139194965362549, "learning_rate": 4.248806246079262e-08, "loss": 2.6686, "step": 27602 }, { "epoch": 0.972647268691033, "grad_norm": 4.211262226104736, "learning_rate": 4.2383033375285045e-08, "loss": 3.2115, "step": 27603 }, { "epoch": 0.9726825057039915, "grad_norm": 3.035151481628418, "learning_rate": 4.22781339890066e-08, "loss": 3.2159, "step": 27604 }, { "epoch": 0.9727177427169499, "grad_norm": 1.397087812423706, "learning_rate": 4.2173364303323974e-08, "loss": 0.7637, "step": 27605 }, { "epoch": 0.9727529797299083, "grad_norm": 1.2807899713516235, "learning_rate": 4.206872431960274e-08, "loss": 1.1165, "step": 27606 }, { "epoch": 0.9727882167428668, "grad_norm": 6.008140563964844, "learning_rate": 4.196421403920403e-08, "loss": 3.0198, "step": 27607 }, { "epoch": 0.9728234537558251, "grad_norm": 2.7113025188446045, "learning_rate": 4.1859833463488986e-08, "loss": 2.3695, "step": 27608 }, { "epoch": 0.9728586907687835, "grad_norm": 1.9223958253860474, "learning_rate": 4.1755582593819845e-08, "loss": 0.8791, "step": 27609 }, { "epoch": 0.972893927781742, "grad_norm": 1.4126859903335571, "learning_rate": 4.165146143155219e-08, "loss": 0.8165, "step": 27610 }, { "epoch": 0.9729291647947004, "grad_norm": 1.5691434144973755, "learning_rate": 4.1547469978041596e-08, "loss": 1.0168, "step": 27611 }, { "epoch": 0.9729644018076588, "grad_norm": 6.381179332733154, "learning_rate": 4.1443608234644774e-08, "loss": 2.516, "step": 27612 }, { "epoch": 0.9729996388206171, "grad_norm": 4.428380966186523, "learning_rate": 4.133987620271285e-08, "loss": 4.5228, "step": 27613 }, { "epoch": 0.9730348758335756, "grad_norm": 2.2824108600616455, "learning_rate": 4.123627388359808e-08, "loss": 2.3178, "step": 27614 }, { "epoch": 0.973070112846534, "grad_norm": 4.679708003997803, "learning_rate": 4.113280127864938e-08, "loss": 3.0161, "step": 27615 }, { "epoch": 0.9731053498594924, "grad_norm": 3.758504867553711, "learning_rate": 4.1029458389214574e-08, "loss": 3.3053, "step": 27616 }, { "epoch": 0.9731405868724509, "grad_norm": 4.872777938842773, "learning_rate": 4.092624521663924e-08, "loss": 5.2518, "step": 27617 }, { "epoch": 0.9731758238854092, "grad_norm": 7.501676559448242, "learning_rate": 4.082316176226786e-08, "loss": 5.0487, "step": 27618 }, { "epoch": 0.9732110608983676, "grad_norm": 5.315667152404785, "learning_rate": 4.072020802744381e-08, "loss": 4.4217, "step": 27619 }, { "epoch": 0.9732462979113261, "grad_norm": 4.512789726257324, "learning_rate": 4.061738401350823e-08, "loss": 2.265, "step": 27620 }, { "epoch": 0.9732815349242845, "grad_norm": 1.1868491172790527, "learning_rate": 4.0514689721798947e-08, "loss": 1.0169, "step": 27621 }, { "epoch": 0.9733167719372429, "grad_norm": 0.8730738759040833, "learning_rate": 4.041212515365489e-08, "loss": 0.7399, "step": 27622 }, { "epoch": 0.9733520089502012, "grad_norm": 3.511831521987915, "learning_rate": 4.030969031041165e-08, "loss": 2.6877, "step": 27623 }, { "epoch": 0.9733872459631597, "grad_norm": 3.71382737159729, "learning_rate": 4.0207385193402614e-08, "loss": 2.8503, "step": 27624 }, { "epoch": 0.9734224829761181, "grad_norm": 4.121641159057617, "learning_rate": 4.0105209803962266e-08, "loss": 2.7474, "step": 27625 }, { "epoch": 0.9734577199890765, "grad_norm": 1.475234031677246, "learning_rate": 4.0003164143419536e-08, "loss": 0.8319, "step": 27626 }, { "epoch": 0.973492957002035, "grad_norm": 5.677742004394531, "learning_rate": 3.9901248213104484e-08, "loss": 6.9797, "step": 27627 }, { "epoch": 0.9735281940149934, "grad_norm": 0.8832978010177612, "learning_rate": 3.9799462014343814e-08, "loss": 0.8914, "step": 27628 }, { "epoch": 0.9735634310279517, "grad_norm": 4.152422904968262, "learning_rate": 3.9697805548465364e-08, "loss": 0.8282, "step": 27629 }, { "epoch": 0.9735986680409102, "grad_norm": 10.5961275100708, "learning_rate": 3.959627881679029e-08, "loss": 4.8017, "step": 27630 }, { "epoch": 0.9736339050538686, "grad_norm": 4.698907375335693, "learning_rate": 3.9494881820643096e-08, "loss": 2.9926, "step": 27631 }, { "epoch": 0.973669142066827, "grad_norm": 3.5701091289520264, "learning_rate": 3.939361456134272e-08, "loss": 2.6349, "step": 27632 }, { "epoch": 0.9737043790797854, "grad_norm": 8.323897361755371, "learning_rate": 3.929247704021033e-08, "loss": 6.19, "step": 27633 }, { "epoch": 0.9737396160927438, "grad_norm": 1.2953391075134277, "learning_rate": 3.9191469258562654e-08, "loss": 0.8973, "step": 27634 }, { "epoch": 0.9737748531057022, "grad_norm": 6.390902996063232, "learning_rate": 3.909059121771419e-08, "loss": 2.6432, "step": 27635 }, { "epoch": 0.9738100901186606, "grad_norm": 6.511321544647217, "learning_rate": 3.8989842918979446e-08, "loss": 3.3928, "step": 27636 }, { "epoch": 0.9738453271316191, "grad_norm": 5.774839878082275, "learning_rate": 3.888922436367071e-08, "loss": 3.2066, "step": 27637 }, { "epoch": 0.9738805641445775, "grad_norm": 16.88317108154297, "learning_rate": 3.878873555309914e-08, "loss": 2.9049, "step": 27638 }, { "epoch": 0.9739158011575358, "grad_norm": 3.2106995582580566, "learning_rate": 3.868837648857371e-08, "loss": 2.4563, "step": 27639 }, { "epoch": 0.9739510381704943, "grad_norm": 2.013204574584961, "learning_rate": 3.858814717140003e-08, "loss": 0.6989, "step": 27640 }, { "epoch": 0.9739862751834527, "grad_norm": 1.1921753883361816, "learning_rate": 3.8488047602885936e-08, "loss": 0.8963, "step": 27641 }, { "epoch": 0.9740215121964111, "grad_norm": 4.960702419281006, "learning_rate": 3.8388077784332624e-08, "loss": 4.8951, "step": 27642 }, { "epoch": 0.9740567492093696, "grad_norm": 2.780499219894409, "learning_rate": 3.8288237717043484e-08, "loss": 2.8777, "step": 27643 }, { "epoch": 0.974091986222328, "grad_norm": 2.2230241298675537, "learning_rate": 3.818852740231971e-08, "loss": 2.0325, "step": 27644 }, { "epoch": 0.9741272232352863, "grad_norm": 1.6214662790298462, "learning_rate": 3.808894684145914e-08, "loss": 1.0217, "step": 27645 }, { "epoch": 0.9741624602482447, "grad_norm": 6.357051849365234, "learning_rate": 3.7989496035758524e-08, "loss": 5.3616, "step": 27646 }, { "epoch": 0.9741976972612032, "grad_norm": 0.9499891400337219, "learning_rate": 3.7890174986513485e-08, "loss": 0.841, "step": 27647 }, { "epoch": 0.9742329342741616, "grad_norm": 2.784296751022339, "learning_rate": 3.779098369501855e-08, "loss": 2.4348, "step": 27648 }, { "epoch": 0.97426817128712, "grad_norm": 1.132128357887268, "learning_rate": 3.76919221625649e-08, "loss": 0.9618, "step": 27649 }, { "epoch": 0.9743034083000784, "grad_norm": 2.0234358310699463, "learning_rate": 3.759299039044262e-08, "loss": 0.8041, "step": 27650 }, { "epoch": 0.9743386453130368, "grad_norm": 4.544034481048584, "learning_rate": 3.749418837994068e-08, "loss": 4.9663, "step": 27651 }, { "epoch": 0.9743738823259952, "grad_norm": 0.8619594573974609, "learning_rate": 3.739551613234582e-08, "loss": 0.9086, "step": 27652 }, { "epoch": 0.9744091193389537, "grad_norm": 4.885444641113281, "learning_rate": 3.7296973648944804e-08, "loss": 5.7244, "step": 27653 }, { "epoch": 0.9744443563519121, "grad_norm": 1.073133111000061, "learning_rate": 3.71985609310177e-08, "loss": 0.6716, "step": 27654 }, { "epoch": 0.9744795933648704, "grad_norm": 1.0791915655136108, "learning_rate": 3.710027797985016e-08, "loss": 1.036, "step": 27655 }, { "epoch": 0.9745148303778288, "grad_norm": 2.4052634239196777, "learning_rate": 3.700212479671894e-08, "loss": 2.2336, "step": 27656 }, { "epoch": 0.9745500673907873, "grad_norm": 6.37544059753418, "learning_rate": 3.6904101382906346e-08, "loss": 5.3542, "step": 27657 }, { "epoch": 0.9745853044037457, "grad_norm": 3.016085624694824, "learning_rate": 3.6806207739684685e-08, "loss": 0.8091, "step": 27658 }, { "epoch": 0.9746205414167041, "grad_norm": 0.8378352522850037, "learning_rate": 3.6708443868331835e-08, "loss": 1.0503, "step": 27659 }, { "epoch": 0.9746557784296626, "grad_norm": 7.6564812660217285, "learning_rate": 3.6610809770121216e-08, "loss": 6.0892, "step": 27660 }, { "epoch": 0.9746910154426209, "grad_norm": 4.44097900390625, "learning_rate": 3.651330544632403e-08, "loss": 3.1835, "step": 27661 }, { "epoch": 0.9747262524555793, "grad_norm": 0.9501121044158936, "learning_rate": 3.641593089820927e-08, "loss": 0.977, "step": 27662 }, { "epoch": 0.9747614894685378, "grad_norm": 2.971761465072632, "learning_rate": 3.631868612704703e-08, "loss": 2.8291, "step": 27663 }, { "epoch": 0.9747967264814962, "grad_norm": 6.479383945465088, "learning_rate": 3.622157113410296e-08, "loss": 4.4588, "step": 27664 }, { "epoch": 0.9748319634944546, "grad_norm": 1.2780039310455322, "learning_rate": 3.612458592064161e-08, "loss": 0.8694, "step": 27665 }, { "epoch": 0.9748672005074129, "grad_norm": 1.6438801288604736, "learning_rate": 3.6027730487927516e-08, "loss": 0.8978, "step": 27666 }, { "epoch": 0.9749024375203714, "grad_norm": 1.980515480041504, "learning_rate": 3.5931004837220786e-08, "loss": 0.8608, "step": 27667 }, { "epoch": 0.9749376745333298, "grad_norm": 3.153348207473755, "learning_rate": 3.583440896978152e-08, "loss": 2.9057, "step": 27668 }, { "epoch": 0.9749729115462882, "grad_norm": 4.145523548126221, "learning_rate": 3.573794288686983e-08, "loss": 4.3584, "step": 27669 }, { "epoch": 0.9750081485592467, "grad_norm": 9.922935485839844, "learning_rate": 3.564160658973914e-08, "loss": 5.0785, "step": 27670 }, { "epoch": 0.975043385572205, "grad_norm": 4.38721227645874, "learning_rate": 3.5545400079646243e-08, "loss": 4.252, "step": 27671 }, { "epoch": 0.9750786225851634, "grad_norm": 6.106799602508545, "learning_rate": 3.544932335784457e-08, "loss": 7.0488, "step": 27672 }, { "epoch": 0.9751138595981219, "grad_norm": 2.4519715309143066, "learning_rate": 3.5353376425584226e-08, "loss": 2.3646, "step": 27673 }, { "epoch": 0.9751490966110803, "grad_norm": 2.2239186763763428, "learning_rate": 3.525755928411534e-08, "loss": 3.0292, "step": 27674 }, { "epoch": 0.9751843336240387, "grad_norm": 1.1787941455841064, "learning_rate": 3.5161871934685785e-08, "loss": 0.8847, "step": 27675 }, { "epoch": 0.975219570636997, "grad_norm": 2.912101984024048, "learning_rate": 3.506631437854235e-08, "loss": 2.3077, "step": 27676 }, { "epoch": 0.9752548076499555, "grad_norm": 3.5546059608459473, "learning_rate": 3.49708866169296e-08, "loss": 3.0973, "step": 27677 }, { "epoch": 0.9752900446629139, "grad_norm": 2.708064079284668, "learning_rate": 3.487558865108986e-08, "loss": 2.5046, "step": 27678 }, { "epoch": 0.9753252816758723, "grad_norm": 2.1608340740203857, "learning_rate": 3.478042048226549e-08, "loss": 2.7156, "step": 27679 }, { "epoch": 0.9753605186888308, "grad_norm": 3.159313917160034, "learning_rate": 3.4685382111695474e-08, "loss": 2.8049, "step": 27680 }, { "epoch": 0.9753957557017892, "grad_norm": 4.280092716217041, "learning_rate": 3.459047354061773e-08, "loss": 4.3737, "step": 27681 }, { "epoch": 0.9754309927147475, "grad_norm": 3.658845901489258, "learning_rate": 3.449569477026904e-08, "loss": 2.8071, "step": 27682 }, { "epoch": 0.975466229727706, "grad_norm": 1.3228060007095337, "learning_rate": 3.440104580188286e-08, "loss": 1.0126, "step": 27683 }, { "epoch": 0.9755014667406644, "grad_norm": 3.715304136276245, "learning_rate": 3.430652663669154e-08, "loss": 2.8776, "step": 27684 }, { "epoch": 0.9755367037536228, "grad_norm": 1.4004307985305786, "learning_rate": 3.4212137275928536e-08, "loss": 0.8321, "step": 27685 }, { "epoch": 0.9755719407665813, "grad_norm": 10.554368019104004, "learning_rate": 3.411787772082065e-08, "loss": 4.2867, "step": 27686 }, { "epoch": 0.9756071777795396, "grad_norm": 0.9986781477928162, "learning_rate": 3.4023747972598e-08, "loss": 0.9264, "step": 27687 }, { "epoch": 0.975642414792498, "grad_norm": 6.327535629272461, "learning_rate": 3.392974803248517e-08, "loss": 8.5577, "step": 27688 }, { "epoch": 0.9756776518054564, "grad_norm": 1.0743579864501953, "learning_rate": 3.3835877901706724e-08, "loss": 0.7486, "step": 27689 }, { "epoch": 0.9757128888184149, "grad_norm": 3.3936474323272705, "learning_rate": 3.3742137581485034e-08, "loss": 2.949, "step": 27690 }, { "epoch": 0.9757481258313733, "grad_norm": 5.909708499908447, "learning_rate": 3.3648527073041336e-08, "loss": 4.767, "step": 27691 }, { "epoch": 0.9757833628443316, "grad_norm": 0.8352175951004028, "learning_rate": 3.355504637759577e-08, "loss": 0.6937, "step": 27692 }, { "epoch": 0.9758185998572901, "grad_norm": 5.374064922332764, "learning_rate": 3.346169549636513e-08, "loss": 2.849, "step": 27693 }, { "epoch": 0.9758538368702485, "grad_norm": 12.571786880493164, "learning_rate": 3.336847443056512e-08, "loss": 2.7797, "step": 27694 }, { "epoch": 0.9758890738832069, "grad_norm": 3.3360722064971924, "learning_rate": 3.3275383181411433e-08, "loss": 3.0236, "step": 27695 }, { "epoch": 0.9759243108961654, "grad_norm": 4.155496597290039, "learning_rate": 3.318242175011421e-08, "loss": 4.7207, "step": 27696 }, { "epoch": 0.9759595479091238, "grad_norm": 5.4699320793151855, "learning_rate": 3.3089590137885816e-08, "loss": 6.7072, "step": 27697 }, { "epoch": 0.9759947849220821, "grad_norm": 1.3855270147323608, "learning_rate": 3.299688834593529e-08, "loss": 0.961, "step": 27698 }, { "epoch": 0.9760300219350405, "grad_norm": 5.2928595542907715, "learning_rate": 3.2904316375469426e-08, "loss": 3.152, "step": 27699 }, { "epoch": 0.976065258947999, "grad_norm": 3.1323206424713135, "learning_rate": 3.281187422769505e-08, "loss": 3.0822, "step": 27700 }, { "epoch": 0.9761004959609574, "grad_norm": 2.7925639152526855, "learning_rate": 3.271956190381565e-08, "loss": 4.0267, "step": 27701 }, { "epoch": 0.9761357329739158, "grad_norm": 5.029109477996826, "learning_rate": 3.2627379405033576e-08, "loss": 2.8497, "step": 27702 }, { "epoch": 0.9761709699868742, "grad_norm": 0.9353238344192505, "learning_rate": 3.2535326732549e-08, "loss": 0.9298, "step": 27703 }, { "epoch": 0.9762062069998326, "grad_norm": 5.047811031341553, "learning_rate": 3.2443403887560953e-08, "loss": 3.6325, "step": 27704 }, { "epoch": 0.976241444012791, "grad_norm": 2.253166675567627, "learning_rate": 3.235161087126848e-08, "loss": 0.7725, "step": 27705 }, { "epoch": 0.9762766810257495, "grad_norm": 2.9436299800872803, "learning_rate": 3.225994768486507e-08, "loss": 2.7001, "step": 27706 }, { "epoch": 0.9763119180387079, "grad_norm": 1.112593650817871, "learning_rate": 3.216841432954643e-08, "loss": 0.7819, "step": 27707 }, { "epoch": 0.9763471550516662, "grad_norm": 4.490838050842285, "learning_rate": 3.207701080650272e-08, "loss": 5.0464, "step": 27708 }, { "epoch": 0.9763823920646246, "grad_norm": 3.183302402496338, "learning_rate": 3.198573711692632e-08, "loss": 2.9333, "step": 27709 }, { "epoch": 0.9764176290775831, "grad_norm": 1.2085158824920654, "learning_rate": 3.189459326200517e-08, "loss": 0.8895, "step": 27710 }, { "epoch": 0.9764528660905415, "grad_norm": 1.3071733713150024, "learning_rate": 3.180357924292721e-08, "loss": 0.7248, "step": 27711 }, { "epoch": 0.9764881031034999, "grad_norm": 3.680859088897705, "learning_rate": 3.171269506087704e-08, "loss": 5.1498, "step": 27712 }, { "epoch": 0.9765233401164584, "grad_norm": 4.610899448394775, "learning_rate": 3.162194071703928e-08, "loss": 2.9633, "step": 27713 }, { "epoch": 0.9765585771294167, "grad_norm": 11.568302154541016, "learning_rate": 3.15313162125952e-08, "loss": 2.7161, "step": 27714 }, { "epoch": 0.9765938141423751, "grad_norm": 4.470602512359619, "learning_rate": 3.1440821548724966e-08, "loss": 4.5381, "step": 27715 }, { "epoch": 0.9766290511553336, "grad_norm": 3.656792163848877, "learning_rate": 3.135045672660764e-08, "loss": 2.8972, "step": 27716 }, { "epoch": 0.976664288168292, "grad_norm": 1.0791608095169067, "learning_rate": 3.126022174742116e-08, "loss": 1.1067, "step": 27717 }, { "epoch": 0.9766995251812504, "grad_norm": 6.839768886566162, "learning_rate": 3.117011661234015e-08, "loss": 4.9432, "step": 27718 }, { "epoch": 0.9767347621942088, "grad_norm": 2.717787265777588, "learning_rate": 3.1080141322538115e-08, "loss": 2.5918, "step": 27719 }, { "epoch": 0.9767699992071672, "grad_norm": 1.5383706092834473, "learning_rate": 3.099029587918634e-08, "loss": 0.8663, "step": 27720 }, { "epoch": 0.9768052362201256, "grad_norm": 0.9658198356628418, "learning_rate": 3.090058028345722e-08, "loss": 0.8322, "step": 27721 }, { "epoch": 0.976840473233084, "grad_norm": 4.850821495056152, "learning_rate": 3.081099453651648e-08, "loss": 2.9503, "step": 27722 }, { "epoch": 0.9768757102460425, "grad_norm": 1.3237290382385254, "learning_rate": 3.07215386395332e-08, "loss": 0.6927, "step": 27723 }, { "epoch": 0.9769109472590009, "grad_norm": 1.572751522064209, "learning_rate": 3.063221259367199e-08, "loss": 0.8997, "step": 27724 }, { "epoch": 0.9769461842719592, "grad_norm": 9.432490348815918, "learning_rate": 3.0543016400095274e-08, "loss": 4.6267, "step": 27725 }, { "epoch": 0.9769814212849177, "grad_norm": 1.5877666473388672, "learning_rate": 3.045395005996543e-08, "loss": 0.9026, "step": 27726 }, { "epoch": 0.9770166582978761, "grad_norm": 4.715751647949219, "learning_rate": 3.036501357444377e-08, "loss": 2.9823, "step": 27727 }, { "epoch": 0.9770518953108345, "grad_norm": 3.817408800125122, "learning_rate": 3.0276206944687136e-08, "loss": 2.8373, "step": 27728 }, { "epoch": 0.977087132323793, "grad_norm": 0.9394444227218628, "learning_rate": 3.01875301718535e-08, "loss": 0.9708, "step": 27729 }, { "epoch": 0.9771223693367513, "grad_norm": 4.321030139923096, "learning_rate": 3.0098983257097483e-08, "loss": 2.6981, "step": 27730 }, { "epoch": 0.9771576063497097, "grad_norm": 1.5400011539459229, "learning_rate": 3.00105662015715e-08, "loss": 1.0656, "step": 27731 }, { "epoch": 0.9771928433626681, "grad_norm": 1.3638417720794678, "learning_rate": 2.992227900642797e-08, "loss": 0.8195, "step": 27732 }, { "epoch": 0.9772280803756266, "grad_norm": 1.3523623943328857, "learning_rate": 2.983412167281596e-08, "loss": 0.8671, "step": 27733 }, { "epoch": 0.977263317388585, "grad_norm": 3.25577974319458, "learning_rate": 2.9746094201885678e-08, "loss": 2.9165, "step": 27734 }, { "epoch": 0.9772985544015433, "grad_norm": 3.9134678840637207, "learning_rate": 2.965819659478175e-08, "loss": 3.4258, "step": 27735 }, { "epoch": 0.9773337914145018, "grad_norm": 8.506050109863281, "learning_rate": 2.9570428852651046e-08, "loss": 4.9487, "step": 27736 }, { "epoch": 0.9773690284274602, "grad_norm": 6.3289289474487305, "learning_rate": 2.948279097663376e-08, "loss": 7.2218, "step": 27737 }, { "epoch": 0.9774042654404186, "grad_norm": 3.8978402614593506, "learning_rate": 2.939528296787453e-08, "loss": 4.6087, "step": 27738 }, { "epoch": 0.9774395024533771, "grad_norm": 1.0523080825805664, "learning_rate": 2.9307904827511335e-08, "loss": 0.8641, "step": 27739 }, { "epoch": 0.9774747394663355, "grad_norm": 2.1557934284210205, "learning_rate": 2.922065655668216e-08, "loss": 0.7411, "step": 27740 }, { "epoch": 0.9775099764792938, "grad_norm": 2.291802406311035, "learning_rate": 2.9133538156524975e-08, "loss": 0.9827, "step": 27741 }, { "epoch": 0.9775452134922522, "grad_norm": 1.1339917182922363, "learning_rate": 2.9046549628173326e-08, "loss": 1.0297, "step": 27742 }, { "epoch": 0.9775804505052107, "grad_norm": 4.274430751800537, "learning_rate": 2.8959690972759636e-08, "loss": 3.2465, "step": 27743 }, { "epoch": 0.9776156875181691, "grad_norm": 4.202679634094238, "learning_rate": 2.8872962191417443e-08, "loss": 4.9744, "step": 27744 }, { "epoch": 0.9776509245311275, "grad_norm": 2.996661424636841, "learning_rate": 2.8786363285273623e-08, "loss": 2.6123, "step": 27745 }, { "epoch": 0.9776861615440859, "grad_norm": 2.6877694129943848, "learning_rate": 2.8699894255458382e-08, "loss": 3.4647, "step": 27746 }, { "epoch": 0.9777213985570443, "grad_norm": 2.0663862228393555, "learning_rate": 2.861355510309749e-08, "loss": 0.905, "step": 27747 }, { "epoch": 0.9777566355700027, "grad_norm": 1.343932867050171, "learning_rate": 2.852734582931449e-08, "loss": 1.0696, "step": 27748 }, { "epoch": 0.9777918725829612, "grad_norm": 1.2514078617095947, "learning_rate": 2.844126643523293e-08, "loss": 0.9258, "step": 27749 }, { "epoch": 0.9778271095959196, "grad_norm": 2.308077335357666, "learning_rate": 2.835531692197524e-08, "loss": 2.762, "step": 27750 }, { "epoch": 0.9778623466088779, "grad_norm": 0.6910790801048279, "learning_rate": 2.826949729065831e-08, "loss": 0.6655, "step": 27751 }, { "epoch": 0.9778975836218364, "grad_norm": 0.9565716981887817, "learning_rate": 2.8183807542402353e-08, "loss": 0.7768, "step": 27752 }, { "epoch": 0.9779328206347948, "grad_norm": 1.420793890953064, "learning_rate": 2.8098247678322034e-08, "loss": 0.9196, "step": 27753 }, { "epoch": 0.9779680576477532, "grad_norm": 3.2282373905181885, "learning_rate": 2.8012817699532015e-08, "loss": 3.009, "step": 27754 }, { "epoch": 0.9780032946607116, "grad_norm": 2.4575421810150146, "learning_rate": 2.792751760714585e-08, "loss": 2.8387, "step": 27755 }, { "epoch": 0.97803853167367, "grad_norm": 1.3483446836471558, "learning_rate": 2.784234740227376e-08, "loss": 0.78, "step": 27756 }, { "epoch": 0.9780737686866284, "grad_norm": 4.40806770324707, "learning_rate": 2.7757307086024865e-08, "loss": 4.7645, "step": 27757 }, { "epoch": 0.9781090056995868, "grad_norm": 3.9672553539276123, "learning_rate": 2.7672396659508272e-08, "loss": 2.4975, "step": 27758 }, { "epoch": 0.9781442427125453, "grad_norm": 1.301595687866211, "learning_rate": 2.7587616123827543e-08, "loss": 0.6491, "step": 27759 }, { "epoch": 0.9781794797255037, "grad_norm": 4.723940849304199, "learning_rate": 2.7502965480089573e-08, "loss": 4.8422, "step": 27760 }, { "epoch": 0.978214716738462, "grad_norm": 0.8864320516586304, "learning_rate": 2.7418444729395698e-08, "loss": 1.0375, "step": 27761 }, { "epoch": 0.9782499537514205, "grad_norm": 3.7179062366485596, "learning_rate": 2.7334053872846156e-08, "loss": 2.9768, "step": 27762 }, { "epoch": 0.9782851907643789, "grad_norm": 1.0379160642623901, "learning_rate": 2.7249792911542284e-08, "loss": 0.6939, "step": 27763 }, { "epoch": 0.9783204277773373, "grad_norm": 4.551568031311035, "learning_rate": 2.716566184657876e-08, "loss": 3.2977, "step": 27764 }, { "epoch": 0.9783556647902957, "grad_norm": 1.0796743631362915, "learning_rate": 2.70816606790536e-08, "loss": 0.8349, "step": 27765 }, { "epoch": 0.9783909018032542, "grad_norm": 7.157236576080322, "learning_rate": 2.699778941005926e-08, "loss": 6.6546, "step": 27766 }, { "epoch": 0.9784261388162125, "grad_norm": 4.271916389465332, "learning_rate": 2.691404804069042e-08, "loss": 4.6196, "step": 27767 }, { "epoch": 0.9784613758291709, "grad_norm": 5.02028751373291, "learning_rate": 2.68304365720351e-08, "loss": 3.5177, "step": 27768 }, { "epoch": 0.9784966128421294, "grad_norm": 3.230588912963867, "learning_rate": 2.674695500518465e-08, "loss": 2.8625, "step": 27769 }, { "epoch": 0.9785318498550878, "grad_norm": 2.9817252159118652, "learning_rate": 2.6663603341225974e-08, "loss": 2.6237, "step": 27770 }, { "epoch": 0.9785670868680462, "grad_norm": 5.982629299163818, "learning_rate": 2.6580381581243765e-08, "loss": 5.2427, "step": 27771 }, { "epoch": 0.9786023238810047, "grad_norm": 1.1651519536972046, "learning_rate": 2.6497289726321595e-08, "loss": 0.7794, "step": 27772 }, { "epoch": 0.978637560893963, "grad_norm": 2.741170883178711, "learning_rate": 2.6414327777543048e-08, "loss": 2.9637, "step": 27773 }, { "epoch": 0.9786727979069214, "grad_norm": 8.002214431762695, "learning_rate": 2.6331495735989477e-08, "loss": 4.5058, "step": 27774 }, { "epoch": 0.9787080349198798, "grad_norm": 2.0557148456573486, "learning_rate": 2.62487936027378e-08, "loss": 2.4864, "step": 27775 }, { "epoch": 0.9787432719328383, "grad_norm": 2.1214730739593506, "learning_rate": 2.6166221378864932e-08, "loss": 2.6513, "step": 27776 }, { "epoch": 0.9787785089457967, "grad_norm": 4.738795280456543, "learning_rate": 2.60837790654489e-08, "loss": 2.8638, "step": 27777 }, { "epoch": 0.978813745958755, "grad_norm": 5.24614143371582, "learning_rate": 2.6001466663562182e-08, "loss": 4.802, "step": 27778 }, { "epoch": 0.9788489829717135, "grad_norm": 1.2361416816711426, "learning_rate": 2.591928417427614e-08, "loss": 0.6035, "step": 27779 }, { "epoch": 0.9788842199846719, "grad_norm": 2.075493574142456, "learning_rate": 2.5837231598662137e-08, "loss": 0.9314, "step": 27780 }, { "epoch": 0.9789194569976303, "grad_norm": 1.1760437488555908, "learning_rate": 2.5755308937789325e-08, "loss": 0.6539, "step": 27781 }, { "epoch": 0.9789546940105888, "grad_norm": 4.185427188873291, "learning_rate": 2.5673516192723514e-08, "loss": 2.833, "step": 27782 }, { "epoch": 0.9789899310235471, "grad_norm": 1.07459557056427, "learning_rate": 2.5591853364530517e-08, "loss": 0.7753, "step": 27783 }, { "epoch": 0.9790251680365055, "grad_norm": 1.2421900033950806, "learning_rate": 2.5510320454273928e-08, "loss": 0.9398, "step": 27784 }, { "epoch": 0.979060405049464, "grad_norm": 0.8827571272850037, "learning_rate": 2.5428917463016233e-08, "loss": 0.952, "step": 27785 }, { "epoch": 0.9790956420624224, "grad_norm": 1.2114057540893555, "learning_rate": 2.5347644391817694e-08, "loss": 0.9072, "step": 27786 }, { "epoch": 0.9791308790753808, "grad_norm": 1.221964716911316, "learning_rate": 2.5266501241737463e-08, "loss": 0.9922, "step": 27787 }, { "epoch": 0.9791661160883391, "grad_norm": 1.1724677085876465, "learning_rate": 2.518548801383025e-08, "loss": 0.6821, "step": 27788 }, { "epoch": 0.9792013531012976, "grad_norm": 4.505048751831055, "learning_rate": 2.5104604709152992e-08, "loss": 4.5323, "step": 27789 }, { "epoch": 0.979236590114256, "grad_norm": 7.279006004333496, "learning_rate": 2.5023851328759287e-08, "loss": 7.2922, "step": 27790 }, { "epoch": 0.9792718271272144, "grad_norm": 0.9926999807357788, "learning_rate": 2.4943227873700516e-08, "loss": 0.6904, "step": 27791 }, { "epoch": 0.9793070641401729, "grad_norm": 4.924729347229004, "learning_rate": 2.486273434502584e-08, "loss": 3.0216, "step": 27792 }, { "epoch": 0.9793423011531313, "grad_norm": 2.26657772064209, "learning_rate": 2.4782370743786644e-08, "loss": 2.9145, "step": 27793 }, { "epoch": 0.9793775381660896, "grad_norm": 4.049347877502441, "learning_rate": 2.4702137071025422e-08, "loss": 2.9221, "step": 27794 }, { "epoch": 0.9794127751790481, "grad_norm": 1.060863733291626, "learning_rate": 2.462203332779023e-08, "loss": 1.1089, "step": 27795 }, { "epoch": 0.9794480121920065, "grad_norm": 5.54011344909668, "learning_rate": 2.4542059515124673e-08, "loss": 3.3297, "step": 27796 }, { "epoch": 0.9794832492049649, "grad_norm": 1.2704888582229614, "learning_rate": 2.4462215634069032e-08, "loss": 0.7481, "step": 27797 }, { "epoch": 0.9795184862179233, "grad_norm": 1.6841310262680054, "learning_rate": 2.4382501685663586e-08, "loss": 0.9986, "step": 27798 }, { "epoch": 0.9795537232308817, "grad_norm": 1.3522313833236694, "learning_rate": 2.430291767094639e-08, "loss": 0.7283, "step": 27799 }, { "epoch": 0.9795889602438401, "grad_norm": 3.4133660793304443, "learning_rate": 2.42234635909544e-08, "loss": 2.5923, "step": 27800 }, { "epoch": 0.9796241972567985, "grad_norm": 1.6740899085998535, "learning_rate": 2.4144139446723448e-08, "loss": 0.7433, "step": 27801 }, { "epoch": 0.979659434269757, "grad_norm": 4.866705894470215, "learning_rate": 2.406494523928493e-08, "loss": 4.8779, "step": 27802 }, { "epoch": 0.9796946712827154, "grad_norm": 9.381165504455566, "learning_rate": 2.398588096967136e-08, "loss": 2.9889, "step": 27803 }, { "epoch": 0.9797299082956737, "grad_norm": 0.7854946851730347, "learning_rate": 2.390694663891302e-08, "loss": 0.7861, "step": 27804 }, { "epoch": 0.9797651453086322, "grad_norm": 6.006242275238037, "learning_rate": 2.382814224803687e-08, "loss": 4.8831, "step": 27805 }, { "epoch": 0.9798003823215906, "grad_norm": 3.8389437198638916, "learning_rate": 2.374946779806986e-08, "loss": 2.3311, "step": 27806 }, { "epoch": 0.979835619334549, "grad_norm": 1.0353626012802124, "learning_rate": 2.367092329003673e-08, "loss": 0.65, "step": 27807 }, { "epoch": 0.9798708563475074, "grad_norm": 3.0075528621673584, "learning_rate": 2.3592508724961104e-08, "loss": 3.0903, "step": 27808 }, { "epoch": 0.9799060933604659, "grad_norm": 1.1426289081573486, "learning_rate": 2.3514224103863282e-08, "loss": 0.9531, "step": 27809 }, { "epoch": 0.9799413303734242, "grad_norm": 2.537214517593384, "learning_rate": 2.3436069427763554e-08, "loss": 2.7536, "step": 27810 }, { "epoch": 0.9799765673863826, "grad_norm": 4.265676021575928, "learning_rate": 2.3358044697679992e-08, "loss": 5.2219, "step": 27811 }, { "epoch": 0.9800118043993411, "grad_norm": 2.7555627822875977, "learning_rate": 2.328014991462957e-08, "loss": 3.238, "step": 27812 }, { "epoch": 0.9800470414122995, "grad_norm": 1.6416136026382446, "learning_rate": 2.32023850796248e-08, "loss": 0.7108, "step": 27813 }, { "epoch": 0.9800822784252579, "grad_norm": 7.008719444274902, "learning_rate": 2.3124750193680433e-08, "loss": 4.8011, "step": 27814 }, { "epoch": 0.9801175154382163, "grad_norm": 2.6478466987609863, "learning_rate": 2.3047245257806773e-08, "loss": 3.0518, "step": 27815 }, { "epoch": 0.9801527524511747, "grad_norm": 1.4161258935928345, "learning_rate": 2.2969870273014116e-08, "loss": 0.7484, "step": 27816 }, { "epoch": 0.9801879894641331, "grad_norm": 1.8876886367797852, "learning_rate": 2.289262524030944e-08, "loss": 0.9248, "step": 27817 }, { "epoch": 0.9802232264770916, "grad_norm": 7.5606560707092285, "learning_rate": 2.2815510160698608e-08, "loss": 4.9183, "step": 27818 }, { "epoch": 0.98025846349005, "grad_norm": 1.080357313156128, "learning_rate": 2.2738525035187476e-08, "loss": 0.9592, "step": 27819 }, { "epoch": 0.9802937005030083, "grad_norm": 4.0731306076049805, "learning_rate": 2.266166986477747e-08, "loss": 2.725, "step": 27820 }, { "epoch": 0.9803289375159667, "grad_norm": 1.1626781225204468, "learning_rate": 2.2584944650470007e-08, "loss": 0.8266, "step": 27821 }, { "epoch": 0.9803641745289252, "grad_norm": 2.062053918838501, "learning_rate": 2.2508349393265404e-08, "loss": 0.6998, "step": 27822 }, { "epoch": 0.9803994115418836, "grad_norm": 4.8301615715026855, "learning_rate": 2.243188409415953e-08, "loss": 5.3296, "step": 27823 }, { "epoch": 0.980434648554842, "grad_norm": 3.8495137691497803, "learning_rate": 2.235554875414936e-08, "loss": 2.6889, "step": 27824 }, { "epoch": 0.9804698855678005, "grad_norm": 2.537999391555786, "learning_rate": 2.2279343374229655e-08, "loss": 0.7434, "step": 27825 }, { "epoch": 0.9805051225807588, "grad_norm": 3.1226892471313477, "learning_rate": 2.220326795539185e-08, "loss": 3.0964, "step": 27826 }, { "epoch": 0.9805403595937172, "grad_norm": 1.0565754175186157, "learning_rate": 2.2127322498628477e-08, "loss": 0.8474, "step": 27827 }, { "epoch": 0.9805755966066757, "grad_norm": 7.113710403442383, "learning_rate": 2.205150700492764e-08, "loss": 2.2492, "step": 27828 }, { "epoch": 0.9806108336196341, "grad_norm": 2.277249813079834, "learning_rate": 2.1975821475275216e-08, "loss": 0.9299, "step": 27829 }, { "epoch": 0.9806460706325925, "grad_norm": 1.2412186861038208, "learning_rate": 2.190026591066041e-08, "loss": 0.9503, "step": 27830 }, { "epoch": 0.9806813076455508, "grad_norm": 1.9917030334472656, "learning_rate": 2.1824840312065775e-08, "loss": 0.8653, "step": 27831 }, { "epoch": 0.9807165446585093, "grad_norm": 0.8596011996269226, "learning_rate": 2.1749544680472746e-08, "loss": 0.8201, "step": 27832 }, { "epoch": 0.9807517816714677, "grad_norm": 2.1891462802886963, "learning_rate": 2.1674379016862756e-08, "loss": 2.6915, "step": 27833 }, { "epoch": 0.9807870186844261, "grad_norm": 1.6144611835479736, "learning_rate": 2.1599343322215026e-08, "loss": 0.8815, "step": 27834 }, { "epoch": 0.9808222556973846, "grad_norm": 3.8455522060394287, "learning_rate": 2.1524437597507664e-08, "loss": 3.4137, "step": 27835 }, { "epoch": 0.980857492710343, "grad_norm": 5.750082015991211, "learning_rate": 2.144966184371544e-08, "loss": 5.0375, "step": 27836 }, { "epoch": 0.9808927297233013, "grad_norm": 3.278045654296875, "learning_rate": 2.1375016061810915e-08, "loss": 3.0065, "step": 27837 }, { "epoch": 0.9809279667362598, "grad_norm": 4.078388214111328, "learning_rate": 2.1300500252769973e-08, "loss": 3.3666, "step": 27838 }, { "epoch": 0.9809632037492182, "grad_norm": 5.6338090896606445, "learning_rate": 2.122611441755962e-08, "loss": 5.038, "step": 27839 }, { "epoch": 0.9809984407621766, "grad_norm": 1.3468173742294312, "learning_rate": 2.1151858557151294e-08, "loss": 0.805, "step": 27840 }, { "epoch": 0.981033677775135, "grad_norm": 0.9646214842796326, "learning_rate": 2.1077732672510897e-08, "loss": 0.8284, "step": 27841 }, { "epoch": 0.9810689147880934, "grad_norm": 3.0967659950256348, "learning_rate": 2.100373676460321e-08, "loss": 2.2829, "step": 27842 }, { "epoch": 0.9811041518010518, "grad_norm": 2.198413848876953, "learning_rate": 2.0929870834394128e-08, "loss": 0.7149, "step": 27843 }, { "epoch": 0.9811393888140102, "grad_norm": 2.8348710536956787, "learning_rate": 2.0856134882843993e-08, "loss": 2.8601, "step": 27844 }, { "epoch": 0.9811746258269687, "grad_norm": 3.057396650314331, "learning_rate": 2.0782528910914256e-08, "loss": 2.8005, "step": 27845 }, { "epoch": 0.9812098628399271, "grad_norm": 1.4032126665115356, "learning_rate": 2.0709052919563045e-08, "loss": 0.7635, "step": 27846 }, { "epoch": 0.9812450998528854, "grad_norm": 4.464104652404785, "learning_rate": 2.0635706909747365e-08, "loss": 2.9568, "step": 27847 }, { "epoch": 0.9812803368658439, "grad_norm": 4.461323261260986, "learning_rate": 2.0562490882423125e-08, "loss": 3.0313, "step": 27848 }, { "epoch": 0.9813155738788023, "grad_norm": 5.765129566192627, "learning_rate": 2.0489404838542893e-08, "loss": 4.8231, "step": 27849 }, { "epoch": 0.9813508108917607, "grad_norm": 3.9895179271698, "learning_rate": 2.041644877906035e-08, "loss": 2.8709, "step": 27850 }, { "epoch": 0.9813860479047192, "grad_norm": 1.5979846715927124, "learning_rate": 2.0343622704923628e-08, "loss": 0.8595, "step": 27851 }, { "epoch": 0.9814212849176775, "grad_norm": 3.4869165420532227, "learning_rate": 2.0270926617083076e-08, "loss": 2.4964, "step": 27852 }, { "epoch": 0.9814565219306359, "grad_norm": 2.4864492416381836, "learning_rate": 2.0198360516485717e-08, "loss": 0.8638, "step": 27853 }, { "epoch": 0.9814917589435943, "grad_norm": 3.274157762527466, "learning_rate": 2.0125924404074126e-08, "loss": 2.318, "step": 27854 }, { "epoch": 0.9815269959565528, "grad_norm": 3.588239908218384, "learning_rate": 2.005361828079533e-08, "loss": 4.5965, "step": 27855 }, { "epoch": 0.9815622329695112, "grad_norm": 3.4035093784332275, "learning_rate": 1.998144214758857e-08, "loss": 2.3748, "step": 27856 }, { "epoch": 0.9815974699824696, "grad_norm": 1.0570976734161377, "learning_rate": 1.990939600539421e-08, "loss": 0.7552, "step": 27857 }, { "epoch": 0.981632706995428, "grad_norm": 6.868156909942627, "learning_rate": 1.983747985515261e-08, "loss": 6.7184, "step": 27858 }, { "epoch": 0.9816679440083864, "grad_norm": 4.90260124206543, "learning_rate": 1.976569369779857e-08, "loss": 5.2234, "step": 27859 }, { "epoch": 0.9817031810213448, "grad_norm": 1.7326815128326416, "learning_rate": 1.9694037534266908e-08, "loss": 0.8596, "step": 27860 }, { "epoch": 0.9817384180343033, "grad_norm": 1.1780030727386475, "learning_rate": 1.9622511365492426e-08, "loss": 1.0068, "step": 27861 }, { "epoch": 0.9817736550472617, "grad_norm": 11.312021255493164, "learning_rate": 1.955111519240549e-08, "loss": 2.8979, "step": 27862 }, { "epoch": 0.98180889206022, "grad_norm": 4.675903797149658, "learning_rate": 1.947984901593758e-08, "loss": 2.6766, "step": 27863 }, { "epoch": 0.9818441290731784, "grad_norm": 6.691224575042725, "learning_rate": 1.940871283701462e-08, "loss": 3.0788, "step": 27864 }, { "epoch": 0.9818793660861369, "grad_norm": 3.249752998352051, "learning_rate": 1.933770665656587e-08, "loss": 2.7178, "step": 27865 }, { "epoch": 0.9819146030990953, "grad_norm": 4.218761444091797, "learning_rate": 1.926683047551392e-08, "loss": 5.3117, "step": 27866 }, { "epoch": 0.9819498401120537, "grad_norm": 2.6005020141601562, "learning_rate": 1.919608429478248e-08, "loss": 2.4417, "step": 27867 }, { "epoch": 0.9819850771250122, "grad_norm": 3.8058032989501953, "learning_rate": 1.9125468115294142e-08, "loss": 4.4291, "step": 27868 }, { "epoch": 0.9820203141379705, "grad_norm": 1.224396824836731, "learning_rate": 1.9054981937968176e-08, "loss": 1.0128, "step": 27869 }, { "epoch": 0.9820555511509289, "grad_norm": 0.8572379946708679, "learning_rate": 1.898462576372162e-08, "loss": 0.7337, "step": 27870 }, { "epoch": 0.9820907881638874, "grad_norm": 1.0507781505584717, "learning_rate": 1.8914399593472633e-08, "loss": 1.0868, "step": 27871 }, { "epoch": 0.9821260251768458, "grad_norm": 4.6739349365234375, "learning_rate": 1.8844303428134925e-08, "loss": 4.653, "step": 27872 }, { "epoch": 0.9821612621898042, "grad_norm": 4.402688503265381, "learning_rate": 1.8774337268621102e-08, "loss": 2.7843, "step": 27873 }, { "epoch": 0.9821964992027625, "grad_norm": 1.515038013458252, "learning_rate": 1.8704501115842655e-08, "loss": 1.1331, "step": 27874 }, { "epoch": 0.982231736215721, "grad_norm": 1.7709636688232422, "learning_rate": 1.8634794970708858e-08, "loss": 0.8554, "step": 27875 }, { "epoch": 0.9822669732286794, "grad_norm": 6.455261707305908, "learning_rate": 1.8565218834130096e-08, "loss": 2.9484, "step": 27876 }, { "epoch": 0.9823022102416378, "grad_norm": 1.0362123250961304, "learning_rate": 1.8495772707010085e-08, "loss": 0.7098, "step": 27877 }, { "epoch": 0.9823374472545963, "grad_norm": 3.9497222900390625, "learning_rate": 1.842645659025366e-08, "loss": 2.6041, "step": 27878 }, { "epoch": 0.9823726842675546, "grad_norm": 3.695539712905884, "learning_rate": 1.8357270484764543e-08, "loss": 3.0287, "step": 27879 }, { "epoch": 0.982407921280513, "grad_norm": 4.000628471374512, "learning_rate": 1.8288214391443128e-08, "loss": 4.8414, "step": 27880 }, { "epoch": 0.9824431582934715, "grad_norm": 1.1696373224258423, "learning_rate": 1.8219288311188687e-08, "loss": 0.7505, "step": 27881 }, { "epoch": 0.9824783953064299, "grad_norm": 0.9997339248657227, "learning_rate": 1.8150492244900507e-08, "loss": 1.1203, "step": 27882 }, { "epoch": 0.9825136323193883, "grad_norm": 0.7293698787689209, "learning_rate": 1.808182619347343e-08, "loss": 0.741, "step": 27883 }, { "epoch": 0.9825488693323466, "grad_norm": 1.0346136093139648, "learning_rate": 1.8013290157801176e-08, "loss": 0.7285, "step": 27884 }, { "epoch": 0.9825841063453051, "grad_norm": 6.235501289367676, "learning_rate": 1.7944884138778595e-08, "loss": 2.8313, "step": 27885 }, { "epoch": 0.9826193433582635, "grad_norm": 1.0801947116851807, "learning_rate": 1.787660813729497e-08, "loss": 0.9363, "step": 27886 }, { "epoch": 0.9826545803712219, "grad_norm": 3.4135055541992188, "learning_rate": 1.7808462154239592e-08, "loss": 2.6643, "step": 27887 }, { "epoch": 0.9826898173841804, "grad_norm": 4.730104446411133, "learning_rate": 1.7740446190500637e-08, "loss": 2.1741, "step": 27888 }, { "epoch": 0.9827250543971388, "grad_norm": 4.116271495819092, "learning_rate": 1.7672560246964067e-08, "loss": 2.594, "step": 27889 }, { "epoch": 0.9827602914100971, "grad_norm": 2.861074209213257, "learning_rate": 1.7604804324513613e-08, "loss": 2.7778, "step": 27890 }, { "epoch": 0.9827955284230556, "grad_norm": 7.487449645996094, "learning_rate": 1.753717842403302e-08, "loss": 4.5589, "step": 27891 }, { "epoch": 0.982830765436014, "grad_norm": 0.9968149662017822, "learning_rate": 1.746968254640158e-08, "loss": 1.014, "step": 27892 }, { "epoch": 0.9828660024489724, "grad_norm": 0.8648473620414734, "learning_rate": 1.7402316692499698e-08, "loss": 0.8792, "step": 27893 }, { "epoch": 0.9829012394619309, "grad_norm": 1.134759783744812, "learning_rate": 1.7335080863203347e-08, "loss": 0.7879, "step": 27894 }, { "epoch": 0.9829364764748892, "grad_norm": 0.8648287653923035, "learning_rate": 1.7267975059389597e-08, "loss": 0.8788, "step": 27895 }, { "epoch": 0.9829717134878476, "grad_norm": 1.6866563558578491, "learning_rate": 1.7200999281932196e-08, "loss": 0.8852, "step": 27896 }, { "epoch": 0.983006950500806, "grad_norm": 1.3941798210144043, "learning_rate": 1.7134153531702667e-08, "loss": 0.5887, "step": 27897 }, { "epoch": 0.9830421875137645, "grad_norm": 1.4590953588485718, "learning_rate": 1.706743780957365e-08, "loss": 0.6361, "step": 27898 }, { "epoch": 0.9830774245267229, "grad_norm": 2.1342391967773438, "learning_rate": 1.7000852116412225e-08, "loss": 0.6718, "step": 27899 }, { "epoch": 0.9831126615396812, "grad_norm": 5.771061897277832, "learning_rate": 1.6934396453086588e-08, "loss": 2.7346, "step": 27900 }, { "epoch": 0.9831478985526397, "grad_norm": 3.1662750244140625, "learning_rate": 1.68680708204616e-08, "loss": 3.022, "step": 27901 }, { "epoch": 0.9831831355655981, "grad_norm": 1.3844623565673828, "learning_rate": 1.680187521940102e-08, "loss": 0.9145, "step": 27902 }, { "epoch": 0.9832183725785565, "grad_norm": 6.921759605407715, "learning_rate": 1.6735809650768598e-08, "loss": 4.7862, "step": 27903 }, { "epoch": 0.983253609591515, "grad_norm": 0.7153294682502747, "learning_rate": 1.6669874115424755e-08, "loss": 0.7386, "step": 27904 }, { "epoch": 0.9832888466044734, "grad_norm": 1.4496852159500122, "learning_rate": 1.6604068614225478e-08, "loss": 1.0667, "step": 27905 }, { "epoch": 0.9833240836174317, "grad_norm": 1.4466121196746826, "learning_rate": 1.6538393148031185e-08, "loss": 0.7495, "step": 27906 }, { "epoch": 0.9833593206303901, "grad_norm": 4.436346530914307, "learning_rate": 1.647284771769564e-08, "loss": 0.6599, "step": 27907 }, { "epoch": 0.9833945576433486, "grad_norm": 3.903684377670288, "learning_rate": 1.6407432324073712e-08, "loss": 3.0739, "step": 27908 }, { "epoch": 0.983429794656307, "grad_norm": 3.689237356185913, "learning_rate": 1.6342146968015836e-08, "loss": 3.0016, "step": 27909 }, { "epoch": 0.9834650316692654, "grad_norm": 8.05882453918457, "learning_rate": 1.6276991650373554e-08, "loss": 5.1067, "step": 27910 }, { "epoch": 0.9835002686822238, "grad_norm": 5.699524402618408, "learning_rate": 1.6211966371995068e-08, "loss": 5.0103, "step": 27911 }, { "epoch": 0.9835355056951822, "grad_norm": 3.131964921951294, "learning_rate": 1.614707113372749e-08, "loss": 3.0277, "step": 27912 }, { "epoch": 0.9835707427081406, "grad_norm": 4.606817245483398, "learning_rate": 1.6082305936415688e-08, "loss": 3.0106, "step": 27913 }, { "epoch": 0.9836059797210991, "grad_norm": 3.7561755180358887, "learning_rate": 1.601767078090344e-08, "loss": 3.1024, "step": 27914 }, { "epoch": 0.9836412167340575, "grad_norm": 5.499805450439453, "learning_rate": 1.59531656680334e-08, "loss": 4.8869, "step": 27915 }, { "epoch": 0.9836764537470158, "grad_norm": 3.1721858978271484, "learning_rate": 1.58887905986449e-08, "loss": 2.9082, "step": 27916 }, { "epoch": 0.9837116907599742, "grad_norm": 3.204444169998169, "learning_rate": 1.5824545573577266e-08, "loss": 2.8366, "step": 27917 }, { "epoch": 0.9837469277729327, "grad_norm": 1.4126425981521606, "learning_rate": 1.5760430593666498e-08, "loss": 0.7351, "step": 27918 }, { "epoch": 0.9837821647858911, "grad_norm": 1.2674049139022827, "learning_rate": 1.5696445659747484e-08, "loss": 0.855, "step": 27919 }, { "epoch": 0.9838174017988495, "grad_norm": 1.786272406578064, "learning_rate": 1.5632590772653998e-08, "loss": 0.8683, "step": 27920 }, { "epoch": 0.983852638811808, "grad_norm": 3.2731857299804688, "learning_rate": 1.5568865933217602e-08, "loss": 2.8704, "step": 27921 }, { "epoch": 0.9838878758247663, "grad_norm": 1.086778163909912, "learning_rate": 1.550527114226985e-08, "loss": 1.0707, "step": 27922 }, { "epoch": 0.9839231128377247, "grad_norm": 1.6177133321762085, "learning_rate": 1.544180640063675e-08, "loss": 2.2885, "step": 27923 }, { "epoch": 0.9839583498506832, "grad_norm": 5.321136951446533, "learning_rate": 1.5378471709145415e-08, "loss": 4.5427, "step": 27924 }, { "epoch": 0.9839935868636416, "grad_norm": 3.5695269107818604, "learning_rate": 1.5315267068621852e-08, "loss": 3.0905, "step": 27925 }, { "epoch": 0.9840288238766, "grad_norm": 1.3392524719238281, "learning_rate": 1.525219247988985e-08, "loss": 0.6147, "step": 27926 }, { "epoch": 0.9840640608895584, "grad_norm": 3.2942562103271484, "learning_rate": 1.518924794376875e-08, "loss": 2.9346, "step": 27927 }, { "epoch": 0.9840992979025168, "grad_norm": 1.0900604724884033, "learning_rate": 1.5126433461079005e-08, "loss": 0.8007, "step": 27928 }, { "epoch": 0.9841345349154752, "grad_norm": 1.2680176496505737, "learning_rate": 1.5063749032639962e-08, "loss": 1.0444, "step": 27929 }, { "epoch": 0.9841697719284336, "grad_norm": 6.036894798278809, "learning_rate": 1.5001194659268747e-08, "loss": 6.8835, "step": 27930 }, { "epoch": 0.9842050089413921, "grad_norm": 1.2917027473449707, "learning_rate": 1.4938770341776932e-08, "loss": 1.0608, "step": 27931 }, { "epoch": 0.9842402459543504, "grad_norm": 3.5094354152679443, "learning_rate": 1.4876476080980529e-08, "loss": 2.609, "step": 27932 }, { "epoch": 0.9842754829673088, "grad_norm": 3.2908596992492676, "learning_rate": 1.4814311877690002e-08, "loss": 3.1718, "step": 27933 }, { "epoch": 0.9843107199802673, "grad_norm": 4.427415370941162, "learning_rate": 1.4752277732714703e-08, "loss": 3.156, "step": 27934 }, { "epoch": 0.9843459569932257, "grad_norm": 0.9171934723854065, "learning_rate": 1.4690373646862877e-08, "loss": 0.878, "step": 27935 }, { "epoch": 0.9843811940061841, "grad_norm": 2.475144624710083, "learning_rate": 1.4628599620940542e-08, "loss": 3.2155, "step": 27936 }, { "epoch": 0.9844164310191426, "grad_norm": 4.13178825378418, "learning_rate": 1.4566955655752613e-08, "loss": 3.9439, "step": 27937 }, { "epoch": 0.9844516680321009, "grad_norm": 4.05208683013916, "learning_rate": 1.450544175210289e-08, "loss": 4.9553, "step": 27938 }, { "epoch": 0.9844869050450593, "grad_norm": 5.822105884552002, "learning_rate": 1.4444057910790732e-08, "loss": 6.8819, "step": 27939 }, { "epoch": 0.9845221420580177, "grad_norm": 2.910710573196411, "learning_rate": 1.4382804132617723e-08, "loss": 2.3789, "step": 27940 }, { "epoch": 0.9845573790709762, "grad_norm": 4.608623027801514, "learning_rate": 1.432168041837989e-08, "loss": 4.8537, "step": 27941 }, { "epoch": 0.9845926160839346, "grad_norm": 4.002532482147217, "learning_rate": 1.4260686768874377e-08, "loss": 2.9359, "step": 27942 }, { "epoch": 0.9846278530968929, "grad_norm": 1.2585620880126953, "learning_rate": 1.419982318489499e-08, "loss": 0.8641, "step": 27943 }, { "epoch": 0.9846630901098514, "grad_norm": 0.9072120189666748, "learning_rate": 1.4139089667235539e-08, "loss": 0.715, "step": 27944 }, { "epoch": 0.9846983271228098, "grad_norm": 1.838808536529541, "learning_rate": 1.4078486216686504e-08, "loss": 0.6951, "step": 27945 }, { "epoch": 0.9847335641357682, "grad_norm": 1.9426575899124146, "learning_rate": 1.4018012834037254e-08, "loss": 2.7791, "step": 27946 }, { "epoch": 0.9847688011487267, "grad_norm": 2.1258835792541504, "learning_rate": 1.3957669520074934e-08, "loss": 0.6331, "step": 27947 }, { "epoch": 0.984804038161685, "grad_norm": 1.3979809284210205, "learning_rate": 1.3897456275585586e-08, "loss": 0.8777, "step": 27948 }, { "epoch": 0.9848392751746434, "grad_norm": 1.5618425607681274, "learning_rate": 1.3837373101354134e-08, "loss": 0.8381, "step": 27949 }, { "epoch": 0.9848745121876018, "grad_norm": 3.1499509811401367, "learning_rate": 1.3777419998163288e-08, "loss": 2.83, "step": 27950 }, { "epoch": 0.9849097492005603, "grad_norm": 1.0533342361450195, "learning_rate": 1.3717596966793534e-08, "loss": 0.9266, "step": 27951 }, { "epoch": 0.9849449862135187, "grad_norm": 0.9765370488166809, "learning_rate": 1.3657904008024247e-08, "loss": 0.6534, "step": 27952 }, { "epoch": 0.984980223226477, "grad_norm": 2.5032222270965576, "learning_rate": 1.3598341122632586e-08, "loss": 2.8362, "step": 27953 }, { "epoch": 0.9850154602394355, "grad_norm": 2.2848153114318848, "learning_rate": 1.3538908311394593e-08, "loss": 0.9334, "step": 27954 }, { "epoch": 0.9850506972523939, "grad_norm": 1.4227337837219238, "learning_rate": 1.3479605575084098e-08, "loss": 0.7391, "step": 27955 }, { "epoch": 0.9850859342653523, "grad_norm": 6.852560520172119, "learning_rate": 1.3420432914474924e-08, "loss": 3.3983, "step": 27956 }, { "epoch": 0.9851211712783108, "grad_norm": 6.025392055511475, "learning_rate": 1.3361390330335344e-08, "loss": 4.465, "step": 27957 }, { "epoch": 0.9851564082912692, "grad_norm": 1.391345739364624, "learning_rate": 1.3302477823436966e-08, "loss": 1.3028, "step": 27958 }, { "epoch": 0.9851916453042275, "grad_norm": 2.0121140480041504, "learning_rate": 1.324369539454473e-08, "loss": 0.908, "step": 27959 }, { "epoch": 0.985226882317186, "grad_norm": 1.1974983215332031, "learning_rate": 1.3185043044425805e-08, "loss": 0.8814, "step": 27960 }, { "epoch": 0.9852621193301444, "grad_norm": 1.3948968648910522, "learning_rate": 1.312652077384402e-08, "loss": 0.9523, "step": 27961 }, { "epoch": 0.9852973563431028, "grad_norm": 2.5813045501708984, "learning_rate": 1.3068128583562101e-08, "loss": 3.0685, "step": 27962 }, { "epoch": 0.9853325933560612, "grad_norm": 5.19108772277832, "learning_rate": 1.3009866474338329e-08, "loss": 7.069, "step": 27963 }, { "epoch": 0.9853678303690196, "grad_norm": 2.335113286972046, "learning_rate": 1.2951734446934316e-08, "loss": 0.8365, "step": 27964 }, { "epoch": 0.985403067381978, "grad_norm": 1.2841428518295288, "learning_rate": 1.2893732502106126e-08, "loss": 0.9708, "step": 27965 }, { "epoch": 0.9854383043949364, "grad_norm": 2.228050708770752, "learning_rate": 1.283586064060982e-08, "loss": 0.708, "step": 27966 }, { "epoch": 0.9854735414078949, "grad_norm": 3.0196163654327393, "learning_rate": 1.2778118863197019e-08, "loss": 3.0668, "step": 27967 }, { "epoch": 0.9855087784208533, "grad_norm": 3.5246174335479736, "learning_rate": 1.2720507170622676e-08, "loss": 2.6878, "step": 27968 }, { "epoch": 0.9855440154338116, "grad_norm": 4.011424541473389, "learning_rate": 1.2663025563635078e-08, "loss": 2.8245, "step": 27969 }, { "epoch": 0.9855792524467701, "grad_norm": 3.8073856830596924, "learning_rate": 1.2605674042984739e-08, "loss": 3.0494, "step": 27970 }, { "epoch": 0.9856144894597285, "grad_norm": 1.1682066917419434, "learning_rate": 1.2548452609416617e-08, "loss": 0.6168, "step": 27971 }, { "epoch": 0.9856497264726869, "grad_norm": 0.9991298913955688, "learning_rate": 1.2491361263677892e-08, "loss": 0.8026, "step": 27972 }, { "epoch": 0.9856849634856453, "grad_norm": 1.1815520524978638, "learning_rate": 1.2434400006510195e-08, "loss": 0.9091, "step": 27973 }, { "epoch": 0.9857202004986038, "grad_norm": 1.1574417352676392, "learning_rate": 1.2377568838657373e-08, "loss": 1.012, "step": 27974 }, { "epoch": 0.9857554375115621, "grad_norm": 5.211221218109131, "learning_rate": 1.2320867760859945e-08, "loss": 6.2153, "step": 27975 }, { "epoch": 0.9857906745245205, "grad_norm": 1.263693928718567, "learning_rate": 1.2264296773853989e-08, "loss": 0.8795, "step": 27976 }, { "epoch": 0.985825911537479, "grad_norm": 1.2079615592956543, "learning_rate": 1.2207855878378916e-08, "loss": 0.8063, "step": 27977 }, { "epoch": 0.9858611485504374, "grad_norm": 6.380773067474365, "learning_rate": 1.215154507516747e-08, "loss": 2.8495, "step": 27978 }, { "epoch": 0.9858963855633958, "grad_norm": 3.9988086223602295, "learning_rate": 1.2095364364955731e-08, "loss": 3.0993, "step": 27979 }, { "epoch": 0.9859316225763542, "grad_norm": 1.11459481716156, "learning_rate": 1.2039313748473113e-08, "loss": 0.7458, "step": 27980 }, { "epoch": 0.9859668595893126, "grad_norm": 1.1099714040756226, "learning_rate": 1.1983393226451256e-08, "loss": 0.7797, "step": 27981 }, { "epoch": 0.986002096602271, "grad_norm": 1.6636450290679932, "learning_rate": 1.1927602799617355e-08, "loss": 1.1079, "step": 27982 }, { "epoch": 0.9860373336152294, "grad_norm": 1.3458665609359741, "learning_rate": 1.1871942468699716e-08, "loss": 1.0087, "step": 27983 }, { "epoch": 0.9860725706281879, "grad_norm": 4.158309459686279, "learning_rate": 1.1816412234422203e-08, "loss": 3.1169, "step": 27984 }, { "epoch": 0.9861078076411463, "grad_norm": 0.8247603178024292, "learning_rate": 1.1761012097507573e-08, "loss": 0.6885, "step": 27985 }, { "epoch": 0.9861430446541046, "grad_norm": 2.404867649078369, "learning_rate": 1.1705742058677472e-08, "loss": 3.102, "step": 27986 }, { "epoch": 0.9861782816670631, "grad_norm": 2.0220649242401123, "learning_rate": 1.1650602118653542e-08, "loss": 2.3835, "step": 27987 }, { "epoch": 0.9862135186800215, "grad_norm": 3.3455302715301514, "learning_rate": 1.159559227815299e-08, "loss": 2.8807, "step": 27988 }, { "epoch": 0.9862487556929799, "grad_norm": 1.3661746978759766, "learning_rate": 1.1540712537890796e-08, "loss": 0.9873, "step": 27989 }, { "epoch": 0.9862839927059384, "grad_norm": 2.2820069789886475, "learning_rate": 1.1485962898584169e-08, "loss": 0.6344, "step": 27990 }, { "epoch": 0.9863192297188967, "grad_norm": 1.2539403438568115, "learning_rate": 1.1431343360944757e-08, "loss": 0.5456, "step": 27991 }, { "epoch": 0.9863544667318551, "grad_norm": 1.0484174489974976, "learning_rate": 1.1376853925685327e-08, "loss": 1.0172, "step": 27992 }, { "epoch": 0.9863897037448136, "grad_norm": 4.252329349517822, "learning_rate": 1.13224945935142e-08, "loss": 3.0793, "step": 27993 }, { "epoch": 0.986424940757772, "grad_norm": 1.0231982469558716, "learning_rate": 1.1268265365139696e-08, "loss": 0.8515, "step": 27994 }, { "epoch": 0.9864601777707304, "grad_norm": 3.3881149291992188, "learning_rate": 1.121416624126792e-08, "loss": 3.0658, "step": 27995 }, { "epoch": 0.9864954147836887, "grad_norm": 3.9457197189331055, "learning_rate": 1.1160197222604974e-08, "loss": 3.3934, "step": 27996 }, { "epoch": 0.9865306517966472, "grad_norm": 3.5535435676574707, "learning_rate": 1.1106358309852516e-08, "loss": 2.7421, "step": 27997 }, { "epoch": 0.9865658888096056, "grad_norm": 6.1212544441223145, "learning_rate": 1.105264950371221e-08, "loss": 2.6819, "step": 27998 }, { "epoch": 0.986601125822564, "grad_norm": 1.310437560081482, "learning_rate": 1.0999070804883493e-08, "loss": 0.7765, "step": 27999 }, { "epoch": 0.9866363628355225, "grad_norm": 5.144127368927002, "learning_rate": 1.0945622214064699e-08, "loss": 3.0668, "step": 28000 }, { "epoch": 0.9866363628355225, "step": 28000, "total_flos": 5.363520427517805e+18, "train_loss": 0.0, "train_runtime": 0.2101, "train_samples_per_second": 4323367.125, "train_steps_per_second": 33772.549 } ], "logging_steps": 1.0, "max_steps": 7094, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 5.363520427517805e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }