{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.352018816150529, "eval_steps": 700, "global_step": 12000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.001960015680125441, "grad_norm": 7.928689002990723, "learning_rate": 8.333333333333334e-08, "loss": 1.4388, "step": 10 }, { "epoch": 0.003920031360250882, "grad_norm": 9.54440689086914, "learning_rate": 1.6666666666666668e-07, "loss": 1.4362, "step": 20 }, { "epoch": 0.005880047040376323, "grad_norm": 9.008377075195312, "learning_rate": 2.5000000000000004e-07, "loss": 1.4445, "step": 30 }, { "epoch": 0.007840062720501764, "grad_norm": 7.416823863983154, "learning_rate": 3.3333333333333335e-07, "loss": 1.4217, "step": 40 }, { "epoch": 0.009800078400627205, "grad_norm": 7.977886199951172, "learning_rate": 4.1666666666666667e-07, "loss": 1.4243, "step": 50 }, { "epoch": 0.011760094080752646, "grad_norm": 8.24797248840332, "learning_rate": 5.000000000000001e-07, "loss": 1.3976, "step": 60 }, { "epoch": 0.013720109760878087, "grad_norm": 7.240636348724365, "learning_rate": 5.833333333333334e-07, "loss": 1.3853, "step": 70 }, { "epoch": 0.01568012544100353, "grad_norm": 5.709108352661133, "learning_rate": 6.666666666666667e-07, "loss": 1.3353, "step": 80 }, { "epoch": 0.01764014112112897, "grad_norm": 9.35693073272705, "learning_rate": 7.5e-07, "loss": 1.3051, "step": 90 }, { "epoch": 0.01960015680125441, "grad_norm": 47.39125442504883, "learning_rate": 8.333333333333333e-07, "loss": 1.265, "step": 100 }, { "epoch": 0.02156017248137985, "grad_norm": 5.6352434158325195, "learning_rate": 9.166666666666666e-07, "loss": 1.2583, "step": 110 }, { "epoch": 0.023520188161505293, "grad_norm": 6.022407531738281, "learning_rate": 1.0000000000000002e-06, "loss": 1.2398, "step": 120 }, { "epoch": 0.025480203841630734, "grad_norm": 6.143012046813965, "learning_rate": 1.0833333333333335e-06, "loss": 1.2237, "step": 130 }, { "epoch": 0.027440219521756175, "grad_norm": 6.173267841339111, "learning_rate": 1.1666666666666668e-06, "loss": 1.2246, "step": 140 }, { "epoch": 0.029400235201881616, "grad_norm": 5.4503092765808105, "learning_rate": 1.25e-06, "loss": 1.1993, "step": 150 }, { "epoch": 0.03136025088200706, "grad_norm": 17.596582412719727, "learning_rate": 1.3333333333333334e-06, "loss": 1.1889, "step": 160 }, { "epoch": 0.033320266562132494, "grad_norm": 6.131169319152832, "learning_rate": 1.4166666666666667e-06, "loss": 1.1884, "step": 170 }, { "epoch": 0.03528028224225794, "grad_norm": 5.450586795806885, "learning_rate": 1.5e-06, "loss": 1.1622, "step": 180 }, { "epoch": 0.03724029792238338, "grad_norm": 14.713933944702148, "learning_rate": 1.5833333333333333e-06, "loss": 1.1675, "step": 190 }, { "epoch": 0.03920031360250882, "grad_norm": 8.510261535644531, "learning_rate": 1.6666666666666667e-06, "loss": 1.1584, "step": 200 }, { "epoch": 0.04116032928263426, "grad_norm": 6.9118971824646, "learning_rate": 1.75e-06, "loss": 1.1696, "step": 210 }, { "epoch": 0.0431203449627597, "grad_norm": 5.986424922943115, "learning_rate": 1.8333333333333333e-06, "loss": 1.1734, "step": 220 }, { "epoch": 0.04508036064288514, "grad_norm": 7.179569244384766, "learning_rate": 1.916666666666667e-06, "loss": 1.1563, "step": 230 }, { "epoch": 0.047040376323010585, "grad_norm": 7.15562629699707, "learning_rate": 2.0000000000000003e-06, "loss": 1.1443, "step": 240 }, { "epoch": 0.04900039200313602, "grad_norm": 5.163909912109375, "learning_rate": 2.0833333333333334e-06, "loss": 1.1567, "step": 250 }, { "epoch": 0.05096040768326147, "grad_norm": 5.861210346221924, "learning_rate": 2.166666666666667e-06, "loss": 1.1596, "step": 260 }, { "epoch": 0.052920423363386905, "grad_norm": 5.1880292892456055, "learning_rate": 2.25e-06, "loss": 1.1428, "step": 270 }, { "epoch": 0.05488043904351235, "grad_norm": 6.203003406524658, "learning_rate": 2.3333333333333336e-06, "loss": 1.1463, "step": 280 }, { "epoch": 0.05684045472363779, "grad_norm": 6.191583156585693, "learning_rate": 2.4166666666666667e-06, "loss": 1.1538, "step": 290 }, { "epoch": 0.05880047040376323, "grad_norm": 9.0908203125, "learning_rate": 2.5e-06, "loss": 1.1277, "step": 300 }, { "epoch": 0.06076048608388867, "grad_norm": 5.685515880584717, "learning_rate": 2.5833333333333337e-06, "loss": 1.1471, "step": 310 }, { "epoch": 0.06272050176401411, "grad_norm": 6.010474681854248, "learning_rate": 2.666666666666667e-06, "loss": 1.1232, "step": 320 }, { "epoch": 0.06468051744413955, "grad_norm": 9.0894193649292, "learning_rate": 2.7500000000000004e-06, "loss": 1.1286, "step": 330 }, { "epoch": 0.06664053312426499, "grad_norm": 5.936020851135254, "learning_rate": 2.8333333333333335e-06, "loss": 1.1116, "step": 340 }, { "epoch": 0.06860054880439044, "grad_norm": 6.055352210998535, "learning_rate": 2.916666666666667e-06, "loss": 1.1167, "step": 350 }, { "epoch": 0.07056056448451588, "grad_norm": 5.160061836242676, "learning_rate": 3e-06, "loss": 1.1225, "step": 360 }, { "epoch": 0.07252058016464132, "grad_norm": 5.748876571655273, "learning_rate": 3.0833333333333336e-06, "loss": 1.1298, "step": 370 }, { "epoch": 0.07448059584476675, "grad_norm": 5.9786057472229, "learning_rate": 3.1666666666666667e-06, "loss": 1.1394, "step": 380 }, { "epoch": 0.0764406115248922, "grad_norm": 5.474339008331299, "learning_rate": 3.2500000000000002e-06, "loss": 1.138, "step": 390 }, { "epoch": 0.07840062720501764, "grad_norm": 6.028172016143799, "learning_rate": 3.3333333333333333e-06, "loss": 1.1243, "step": 400 }, { "epoch": 0.08036064288514308, "grad_norm": 11.877293586730957, "learning_rate": 3.416666666666667e-06, "loss": 1.1086, "step": 410 }, { "epoch": 0.08232065856526852, "grad_norm": 6.457036972045898, "learning_rate": 3.5e-06, "loss": 1.1113, "step": 420 }, { "epoch": 0.08428067424539397, "grad_norm": 8.589061737060547, "learning_rate": 3.5833333333333335e-06, "loss": 1.1207, "step": 430 }, { "epoch": 0.0862406899255194, "grad_norm": 6.367908000946045, "learning_rate": 3.6666666666666666e-06, "loss": 1.1192, "step": 440 }, { "epoch": 0.08820070560564484, "grad_norm": 6.087027549743652, "learning_rate": 3.7500000000000005e-06, "loss": 1.114, "step": 450 }, { "epoch": 0.09016072128577028, "grad_norm": 6.365817546844482, "learning_rate": 3.833333333333334e-06, "loss": 1.1104, "step": 460 }, { "epoch": 0.09212073696589573, "grad_norm": 6.38438606262207, "learning_rate": 3.916666666666667e-06, "loss": 1.1222, "step": 470 }, { "epoch": 0.09408075264602117, "grad_norm": 5.983357906341553, "learning_rate": 4.000000000000001e-06, "loss": 1.1082, "step": 480 }, { "epoch": 0.09604076832614661, "grad_norm": 7.514024257659912, "learning_rate": 4.083333333333334e-06, "loss": 1.1305, "step": 490 }, { "epoch": 0.09800078400627205, "grad_norm": 5.6476922035217285, "learning_rate": 4.166666666666667e-06, "loss": 1.1183, "step": 500 }, { "epoch": 0.0999607996863975, "grad_norm": 6.5463128089904785, "learning_rate": 4.25e-06, "loss": 1.1126, "step": 510 }, { "epoch": 0.10192081536652293, "grad_norm": 6.156862258911133, "learning_rate": 4.333333333333334e-06, "loss": 1.1161, "step": 520 }, { "epoch": 0.10388083104664837, "grad_norm": 5.682535648345947, "learning_rate": 4.416666666666667e-06, "loss": 1.1096, "step": 530 }, { "epoch": 0.10584084672677381, "grad_norm": 7.1881489753723145, "learning_rate": 4.5e-06, "loss": 1.1136, "step": 540 }, { "epoch": 0.10780086240689926, "grad_norm": 5.990985870361328, "learning_rate": 4.583333333333333e-06, "loss": 1.1224, "step": 550 }, { "epoch": 0.1097608780870247, "grad_norm": 5.474586009979248, "learning_rate": 4.666666666666667e-06, "loss": 1.1078, "step": 560 }, { "epoch": 0.11172089376715014, "grad_norm": 5.7349371910095215, "learning_rate": 4.75e-06, "loss": 1.1238, "step": 570 }, { "epoch": 0.11368090944727557, "grad_norm": 7.513430595397949, "learning_rate": 4.833333333333333e-06, "loss": 1.1158, "step": 580 }, { "epoch": 0.11564092512740103, "grad_norm": 5.410061836242676, "learning_rate": 4.9166666666666665e-06, "loss": 1.1102, "step": 590 }, { "epoch": 0.11760094080752646, "grad_norm": 6.409933567047119, "learning_rate": 5e-06, "loss": 1.1035, "step": 600 }, { "epoch": 0.1195609564876519, "grad_norm": 14.141221046447754, "learning_rate": 5.0833333333333335e-06, "loss": 1.1066, "step": 610 }, { "epoch": 0.12152097216777734, "grad_norm": 5.547974586486816, "learning_rate": 5.1666666666666675e-06, "loss": 1.1218, "step": 620 }, { "epoch": 0.12348098784790279, "grad_norm": 5.946651935577393, "learning_rate": 5.2500000000000006e-06, "loss": 1.112, "step": 630 }, { "epoch": 0.12544100352802823, "grad_norm": 6.724623680114746, "learning_rate": 5.333333333333334e-06, "loss": 1.0986, "step": 640 }, { "epoch": 0.12740101920815367, "grad_norm": 8.897869110107422, "learning_rate": 5.416666666666667e-06, "loss": 1.1077, "step": 650 }, { "epoch": 0.1293610348882791, "grad_norm": 6.75730037689209, "learning_rate": 5.500000000000001e-06, "loss": 1.1018, "step": 660 }, { "epoch": 0.13132105056840454, "grad_norm": 5.3133111000061035, "learning_rate": 5.583333333333334e-06, "loss": 1.1089, "step": 670 }, { "epoch": 0.13328106624852998, "grad_norm": 5.644856929779053, "learning_rate": 5.666666666666667e-06, "loss": 1.1159, "step": 680 }, { "epoch": 0.13524108192865542, "grad_norm": 5.264536380767822, "learning_rate": 5.75e-06, "loss": 1.1068, "step": 690 }, { "epoch": 0.13720109760878088, "grad_norm": 5.970139026641846, "learning_rate": 5.833333333333334e-06, "loss": 1.1214, "step": 700 }, { "epoch": 0.13720109760878088, "eval_loss": 1.0996023416519165, "eval_runtime": 14.1011, "eval_samples_per_second": 46.379, "eval_steps_per_second": 5.815, "step": 700 }, { "epoch": 0.13916111328890632, "grad_norm": 5.035984039306641, "learning_rate": 5.916666666666667e-06, "loss": 1.1078, "step": 710 }, { "epoch": 0.14112112896903176, "grad_norm": 5.657273292541504, "learning_rate": 6e-06, "loss": 1.0998, "step": 720 }, { "epoch": 0.1430811446491572, "grad_norm": 5.413987636566162, "learning_rate": 6.083333333333333e-06, "loss": 1.1023, "step": 730 }, { "epoch": 0.14504116032928263, "grad_norm": 7.05068302154541, "learning_rate": 6.166666666666667e-06, "loss": 1.1053, "step": 740 }, { "epoch": 0.14700117600940807, "grad_norm": 6.170547008514404, "learning_rate": 6.25e-06, "loss": 1.0998, "step": 750 }, { "epoch": 0.1489611916895335, "grad_norm": 5.373660087585449, "learning_rate": 6.333333333333333e-06, "loss": 1.0899, "step": 760 }, { "epoch": 0.15092120736965894, "grad_norm": 6.981970310211182, "learning_rate": 6.416666666666667e-06, "loss": 1.113, "step": 770 }, { "epoch": 0.1528812230497844, "grad_norm": 4.918264865875244, "learning_rate": 6.5000000000000004e-06, "loss": 1.0884, "step": 780 }, { "epoch": 0.15484123872990985, "grad_norm": 13.080545425415039, "learning_rate": 6.5833333333333335e-06, "loss": 1.1012, "step": 790 }, { "epoch": 0.15680125441003528, "grad_norm": 5.957456111907959, "learning_rate": 6.666666666666667e-06, "loss": 1.1044, "step": 800 }, { "epoch": 0.15876127009016072, "grad_norm": 5.442764759063721, "learning_rate": 6.750000000000001e-06, "loss": 1.1051, "step": 810 }, { "epoch": 0.16072128577028616, "grad_norm": 5.379378318786621, "learning_rate": 6.833333333333334e-06, "loss": 1.1092, "step": 820 }, { "epoch": 0.1626813014504116, "grad_norm": 6.137138366699219, "learning_rate": 6.916666666666667e-06, "loss": 1.0835, "step": 830 }, { "epoch": 0.16464131713053703, "grad_norm": 5.297956943511963, "learning_rate": 7e-06, "loss": 1.1089, "step": 840 }, { "epoch": 0.16660133281066247, "grad_norm": 5.63748025894165, "learning_rate": 7.083333333333335e-06, "loss": 1.1024, "step": 850 }, { "epoch": 0.16856134849078794, "grad_norm": 5.432644844055176, "learning_rate": 7.166666666666667e-06, "loss": 1.0712, "step": 860 }, { "epoch": 0.17052136417091338, "grad_norm": 5.4105305671691895, "learning_rate": 7.25e-06, "loss": 1.1122, "step": 870 }, { "epoch": 0.1724813798510388, "grad_norm": 6.026330471038818, "learning_rate": 7.333333333333333e-06, "loss": 1.1049, "step": 880 }, { "epoch": 0.17444139553116425, "grad_norm": 6.143797397613525, "learning_rate": 7.416666666666668e-06, "loss": 1.1038, "step": 890 }, { "epoch": 0.1764014112112897, "grad_norm": 5.558448791503906, "learning_rate": 7.500000000000001e-06, "loss": 1.095, "step": 900 }, { "epoch": 0.17836142689141513, "grad_norm": 6.965857028961182, "learning_rate": 7.583333333333333e-06, "loss": 1.0979, "step": 910 }, { "epoch": 0.18032144257154056, "grad_norm": 6.795146942138672, "learning_rate": 7.666666666666667e-06, "loss": 1.103, "step": 920 }, { "epoch": 0.182281458251666, "grad_norm": 5.742541313171387, "learning_rate": 7.75e-06, "loss": 1.1031, "step": 930 }, { "epoch": 0.18424147393179147, "grad_norm": 5.67035436630249, "learning_rate": 7.833333333333333e-06, "loss": 1.1018, "step": 940 }, { "epoch": 0.1862014896119169, "grad_norm": 8.501360893249512, "learning_rate": 7.916666666666667e-06, "loss": 1.0916, "step": 950 }, { "epoch": 0.18816150529204234, "grad_norm": 5.092253684997559, "learning_rate": 8.000000000000001e-06, "loss": 1.0886, "step": 960 }, { "epoch": 0.19012152097216778, "grad_norm": 6.499626636505127, "learning_rate": 8.083333333333334e-06, "loss": 1.1008, "step": 970 }, { "epoch": 0.19208153665229322, "grad_norm": 5.9327216148376465, "learning_rate": 8.166666666666668e-06, "loss": 1.0911, "step": 980 }, { "epoch": 0.19404155233241865, "grad_norm": 5.990231990814209, "learning_rate": 8.25e-06, "loss": 1.0955, "step": 990 }, { "epoch": 0.1960015680125441, "grad_norm": 32.17470932006836, "learning_rate": 8.333333333333334e-06, "loss": 1.1087, "step": 1000 }, { "epoch": 0.19796158369266953, "grad_norm": 5.19512414932251, "learning_rate": 8.416666666666667e-06, "loss": 1.1302, "step": 1010 }, { "epoch": 0.199921599372795, "grad_norm": 5.765759468078613, "learning_rate": 8.5e-06, "loss": 1.1089, "step": 1020 }, { "epoch": 0.20188161505292043, "grad_norm": 5.5894927978515625, "learning_rate": 8.583333333333333e-06, "loss": 1.0914, "step": 1030 }, { "epoch": 0.20384163073304587, "grad_norm": 6.387049198150635, "learning_rate": 8.666666666666668e-06, "loss": 1.095, "step": 1040 }, { "epoch": 0.2058016464131713, "grad_norm": 5.629969596862793, "learning_rate": 8.750000000000001e-06, "loss": 1.0999, "step": 1050 }, { "epoch": 0.20776166209329675, "grad_norm": 22.919910430908203, "learning_rate": 8.833333333333334e-06, "loss": 1.096, "step": 1060 }, { "epoch": 0.20972167777342218, "grad_norm": 5.267335414886475, "learning_rate": 8.916666666666667e-06, "loss": 1.1006, "step": 1070 }, { "epoch": 0.21168169345354762, "grad_norm": 53.39712905883789, "learning_rate": 9e-06, "loss": 1.106, "step": 1080 }, { "epoch": 0.21364170913367306, "grad_norm": 5.858781814575195, "learning_rate": 9.083333333333333e-06, "loss": 1.2549, "step": 1090 }, { "epoch": 0.21560172481379852, "grad_norm": 6.5703935623168945, "learning_rate": 9.166666666666666e-06, "loss": 1.094, "step": 1100 }, { "epoch": 0.21756174049392396, "grad_norm": 5.414449214935303, "learning_rate": 9.250000000000001e-06, "loss": 1.0891, "step": 1110 }, { "epoch": 0.2195217561740494, "grad_norm": 7.185405731201172, "learning_rate": 9.333333333333334e-06, "loss": 1.089, "step": 1120 }, { "epoch": 0.22148177185417484, "grad_norm": 6.4889092445373535, "learning_rate": 9.416666666666667e-06, "loss": 1.0874, "step": 1130 }, { "epoch": 0.22344178753430027, "grad_norm": 6.4478759765625, "learning_rate": 9.5e-06, "loss": 1.1111, "step": 1140 }, { "epoch": 0.2254018032144257, "grad_norm": 6.49618673324585, "learning_rate": 9.583333333333335e-06, "loss": 1.1001, "step": 1150 }, { "epoch": 0.22736181889455115, "grad_norm": 7.893542766571045, "learning_rate": 9.666666666666667e-06, "loss": 1.097, "step": 1160 }, { "epoch": 0.2293218345746766, "grad_norm": 11.165901184082031, "learning_rate": 9.75e-06, "loss": 1.092, "step": 1170 }, { "epoch": 0.23128185025480205, "grad_norm": 21.967939376831055, "learning_rate": 9.833333333333333e-06, "loss": 1.1216, "step": 1180 }, { "epoch": 0.2332418659349275, "grad_norm": 14.317709922790527, "learning_rate": 9.916666666666668e-06, "loss": 1.1047, "step": 1190 }, { "epoch": 0.23520188161505293, "grad_norm": 6.240993499755859, "learning_rate": 1e-05, "loss": 1.1045, "step": 1200 }, { "epoch": 0.23716189729517836, "grad_norm": 5.804283142089844, "learning_rate": 9.990740740740741e-06, "loss": 1.0933, "step": 1210 }, { "epoch": 0.2391219129753038, "grad_norm": 5.778935432434082, "learning_rate": 9.981481481481482e-06, "loss": 1.1098, "step": 1220 }, { "epoch": 0.24108192865542924, "grad_norm": 6.391726016998291, "learning_rate": 9.972222222222224e-06, "loss": 1.0993, "step": 1230 }, { "epoch": 0.24304194433555468, "grad_norm": 4.967405319213867, "learning_rate": 9.962962962962964e-06, "loss": 1.0835, "step": 1240 }, { "epoch": 0.24500196001568011, "grad_norm": 11.337072372436523, "learning_rate": 9.953703703703704e-06, "loss": 1.1057, "step": 1250 }, { "epoch": 0.24696197569580558, "grad_norm": 5.182584285736084, "learning_rate": 9.944444444444445e-06, "loss": 1.0913, "step": 1260 }, { "epoch": 0.24892199137593102, "grad_norm": 6.305624961853027, "learning_rate": 9.935185185185185e-06, "loss": 1.0963, "step": 1270 }, { "epoch": 0.25088200705605646, "grad_norm": 5.263504981994629, "learning_rate": 9.925925925925927e-06, "loss": 1.1097, "step": 1280 }, { "epoch": 0.2528420227361819, "grad_norm": 5.2210493087768555, "learning_rate": 9.916666666666668e-06, "loss": 1.0934, "step": 1290 }, { "epoch": 0.25480203841630733, "grad_norm": 6.707633972167969, "learning_rate": 9.907407407407408e-06, "loss": 1.1049, "step": 1300 }, { "epoch": 0.25676205409643277, "grad_norm": 5.195355415344238, "learning_rate": 9.898148148148148e-06, "loss": 1.0866, "step": 1310 }, { "epoch": 0.2587220697765582, "grad_norm": 7.137194633483887, "learning_rate": 9.88888888888889e-06, "loss": 1.1032, "step": 1320 }, { "epoch": 0.26068208545668364, "grad_norm": 5.271111488342285, "learning_rate": 9.87962962962963e-06, "loss": 1.0925, "step": 1330 }, { "epoch": 0.2626421011368091, "grad_norm": 6.781525135040283, "learning_rate": 9.870370370370371e-06, "loss": 1.0966, "step": 1340 }, { "epoch": 0.2646021168169345, "grad_norm": 8.270143508911133, "learning_rate": 9.861111111111112e-06, "loss": 1.0991, "step": 1350 }, { "epoch": 0.26656213249705996, "grad_norm": 6.368267059326172, "learning_rate": 9.851851851851852e-06, "loss": 1.1005, "step": 1360 }, { "epoch": 0.2685221481771854, "grad_norm": 5.750402927398682, "learning_rate": 9.842592592592594e-06, "loss": 1.0907, "step": 1370 }, { "epoch": 0.27048216385731083, "grad_norm": 5.706563472747803, "learning_rate": 9.833333333333333e-06, "loss": 1.1133, "step": 1380 }, { "epoch": 0.2724421795374363, "grad_norm": 5.816508769989014, "learning_rate": 9.824074074074075e-06, "loss": 1.1074, "step": 1390 }, { "epoch": 0.27440219521756176, "grad_norm": 4.7041449546813965, "learning_rate": 9.814814814814815e-06, "loss": 1.0944, "step": 1400 }, { "epoch": 0.27440219521756176, "eval_loss": 1.089659571647644, "eval_runtime": 14.1203, "eval_samples_per_second": 46.316, "eval_steps_per_second": 5.807, "step": 1400 }, { "epoch": 0.2763622108976872, "grad_norm": 4.5806355476379395, "learning_rate": 9.805555555555556e-06, "loss": 1.0868, "step": 1410 }, { "epoch": 0.27832222657781264, "grad_norm": 6.664804458618164, "learning_rate": 9.796296296296298e-06, "loss": 1.1031, "step": 1420 }, { "epoch": 0.2802822422579381, "grad_norm": 5.814957141876221, "learning_rate": 9.787037037037038e-06, "loss": 1.0818, "step": 1430 }, { "epoch": 0.2822422579380635, "grad_norm": 11.043229103088379, "learning_rate": 9.777777777777779e-06, "loss": 1.1147, "step": 1440 }, { "epoch": 0.28420227361818895, "grad_norm": 5.907972812652588, "learning_rate": 9.768518518518519e-06, "loss": 1.0949, "step": 1450 }, { "epoch": 0.2861622892983144, "grad_norm": 8.000860214233398, "learning_rate": 9.759259259259261e-06, "loss": 1.107, "step": 1460 }, { "epoch": 0.2881223049784398, "grad_norm": 5.279980182647705, "learning_rate": 9.75e-06, "loss": 1.0988, "step": 1470 }, { "epoch": 0.29008232065856526, "grad_norm": 6.427779674530029, "learning_rate": 9.740740740740742e-06, "loss": 1.0969, "step": 1480 }, { "epoch": 0.2920423363386907, "grad_norm": 5.733342170715332, "learning_rate": 9.731481481481482e-06, "loss": 1.0832, "step": 1490 }, { "epoch": 0.29400235201881614, "grad_norm": 5.370787620544434, "learning_rate": 9.722222222222223e-06, "loss": 1.0724, "step": 1500 }, { "epoch": 0.2959623676989416, "grad_norm": 4.780156135559082, "learning_rate": 9.712962962962965e-06, "loss": 1.0911, "step": 1510 }, { "epoch": 0.297922383379067, "grad_norm": 5.770766258239746, "learning_rate": 9.703703703703703e-06, "loss": 1.0913, "step": 1520 }, { "epoch": 0.29988239905919245, "grad_norm": 5.799376487731934, "learning_rate": 9.694444444444446e-06, "loss": 1.1072, "step": 1530 }, { "epoch": 0.3018424147393179, "grad_norm": 4.571477890014648, "learning_rate": 9.685185185185186e-06, "loss": 1.0977, "step": 1540 }, { "epoch": 0.3038024304194434, "grad_norm": 6.020542144775391, "learning_rate": 9.675925925925926e-06, "loss": 1.1056, "step": 1550 }, { "epoch": 0.3057624460995688, "grad_norm": 6.760439395904541, "learning_rate": 9.666666666666667e-06, "loss": 1.0874, "step": 1560 }, { "epoch": 0.30772246177969426, "grad_norm": 4.900004863739014, "learning_rate": 9.657407407407409e-06, "loss": 1.0946, "step": 1570 }, { "epoch": 0.3096824774598197, "grad_norm": 4.82609748840332, "learning_rate": 9.64814814814815e-06, "loss": 1.1017, "step": 1580 }, { "epoch": 0.31164249313994513, "grad_norm": 5.115661144256592, "learning_rate": 9.63888888888889e-06, "loss": 1.1073, "step": 1590 }, { "epoch": 0.31360250882007057, "grad_norm": 4.987372875213623, "learning_rate": 9.62962962962963e-06, "loss": 1.084, "step": 1600 }, { "epoch": 0.315562524500196, "grad_norm": 5.010407447814941, "learning_rate": 9.62037037037037e-06, "loss": 1.1031, "step": 1610 }, { "epoch": 0.31752254018032144, "grad_norm": 4.8331122398376465, "learning_rate": 9.611111111111112e-06, "loss": 1.0995, "step": 1620 }, { "epoch": 0.3194825558604469, "grad_norm": 5.5287909507751465, "learning_rate": 9.601851851851853e-06, "loss": 1.0884, "step": 1630 }, { "epoch": 0.3214425715405723, "grad_norm": 7.749701023101807, "learning_rate": 9.592592592592593e-06, "loss": 1.0786, "step": 1640 }, { "epoch": 0.32340258722069776, "grad_norm": 4.908470153808594, "learning_rate": 9.583333333333335e-06, "loss": 1.1049, "step": 1650 }, { "epoch": 0.3253626029008232, "grad_norm": 8.760143280029297, "learning_rate": 9.574074074074074e-06, "loss": 1.1, "step": 1660 }, { "epoch": 0.32732261858094863, "grad_norm": 4.7910590171813965, "learning_rate": 9.564814814814816e-06, "loss": 1.0936, "step": 1670 }, { "epoch": 0.32928263426107407, "grad_norm": 10.59984302520752, "learning_rate": 9.555555555555556e-06, "loss": 1.0888, "step": 1680 }, { "epoch": 0.3312426499411995, "grad_norm": 5.18034029006958, "learning_rate": 9.546296296296297e-06, "loss": 1.1015, "step": 1690 }, { "epoch": 0.33320266562132494, "grad_norm": 4.835277557373047, "learning_rate": 9.537037037037037e-06, "loss": 1.0934, "step": 1700 }, { "epoch": 0.33516268130145044, "grad_norm": 5.690189838409424, "learning_rate": 9.527777777777778e-06, "loss": 1.0981, "step": 1710 }, { "epoch": 0.3371226969815759, "grad_norm": 5.126386642456055, "learning_rate": 9.51851851851852e-06, "loss": 1.0731, "step": 1720 }, { "epoch": 0.3390827126617013, "grad_norm": 5.041740894317627, "learning_rate": 9.50925925925926e-06, "loss": 1.0822, "step": 1730 }, { "epoch": 0.34104272834182675, "grad_norm": 5.251774787902832, "learning_rate": 9.5e-06, "loss": 1.0914, "step": 1740 }, { "epoch": 0.3430027440219522, "grad_norm": 4.976339817047119, "learning_rate": 9.490740740740741e-06, "loss": 1.0994, "step": 1750 }, { "epoch": 0.3449627597020776, "grad_norm": 4.93164587020874, "learning_rate": 9.481481481481483e-06, "loss": 1.1029, "step": 1760 }, { "epoch": 0.34692277538220306, "grad_norm": 4.964654445648193, "learning_rate": 9.472222222222223e-06, "loss": 1.0846, "step": 1770 }, { "epoch": 0.3488827910623285, "grad_norm": 5.763545036315918, "learning_rate": 9.462962962962964e-06, "loss": 1.0861, "step": 1780 }, { "epoch": 0.35084280674245394, "grad_norm": 7.256554126739502, "learning_rate": 9.453703703703704e-06, "loss": 1.0878, "step": 1790 }, { "epoch": 0.3528028224225794, "grad_norm": 5.118113994598389, "learning_rate": 9.444444444444445e-06, "loss": 1.0814, "step": 1800 }, { "epoch": 0.3547628381027048, "grad_norm": 5.370246410369873, "learning_rate": 9.435185185185187e-06, "loss": 1.0977, "step": 1810 }, { "epoch": 0.35672285378283025, "grad_norm": 4.952902793884277, "learning_rate": 9.425925925925925e-06, "loss": 1.0889, "step": 1820 }, { "epoch": 0.3586828694629557, "grad_norm": 4.842602252960205, "learning_rate": 9.416666666666667e-06, "loss": 1.0926, "step": 1830 }, { "epoch": 0.3606428851430811, "grad_norm": 4.805694103240967, "learning_rate": 9.407407407407408e-06, "loss": 1.1087, "step": 1840 }, { "epoch": 0.36260290082320656, "grad_norm": 5.138493061065674, "learning_rate": 9.398148148148148e-06, "loss": 1.0862, "step": 1850 }, { "epoch": 0.364562916503332, "grad_norm": 7.099350929260254, "learning_rate": 9.38888888888889e-06, "loss": 1.0887, "step": 1860 }, { "epoch": 0.3665229321834575, "grad_norm": 5.295989990234375, "learning_rate": 9.37962962962963e-06, "loss": 1.0715, "step": 1870 }, { "epoch": 0.36848294786358293, "grad_norm": 5.222052097320557, "learning_rate": 9.370370370370371e-06, "loss": 1.0979, "step": 1880 }, { "epoch": 0.37044296354370837, "grad_norm": 4.347951889038086, "learning_rate": 9.361111111111111e-06, "loss": 1.1036, "step": 1890 }, { "epoch": 0.3724029792238338, "grad_norm": 5.062961101531982, "learning_rate": 9.351851851851854e-06, "loss": 1.0864, "step": 1900 }, { "epoch": 0.37436299490395925, "grad_norm": 5.308873176574707, "learning_rate": 9.342592592592594e-06, "loss": 1.0877, "step": 1910 }, { "epoch": 0.3763230105840847, "grad_norm": 4.578522682189941, "learning_rate": 9.333333333333334e-06, "loss": 1.0961, "step": 1920 }, { "epoch": 0.3782830262642101, "grad_norm": 4.571235179901123, "learning_rate": 9.324074074074075e-06, "loss": 1.0838, "step": 1930 }, { "epoch": 0.38024304194433556, "grad_norm": 4.522839069366455, "learning_rate": 9.314814814814815e-06, "loss": 1.0786, "step": 1940 }, { "epoch": 0.382203057624461, "grad_norm": 4.68695068359375, "learning_rate": 9.305555555555557e-06, "loss": 1.1146, "step": 1950 }, { "epoch": 0.38416307330458643, "grad_norm": 4.976430416107178, "learning_rate": 9.296296296296296e-06, "loss": 1.0937, "step": 1960 }, { "epoch": 0.38612308898471187, "grad_norm": 5.086645126342773, "learning_rate": 9.287037037037038e-06, "loss": 1.097, "step": 1970 }, { "epoch": 0.3880831046648373, "grad_norm": 4.946636199951172, "learning_rate": 9.277777777777778e-06, "loss": 1.0986, "step": 1980 }, { "epoch": 0.39004312034496275, "grad_norm": 4.823138236999512, "learning_rate": 9.268518518518519e-06, "loss": 1.1034, "step": 1990 }, { "epoch": 0.3920031360250882, "grad_norm": 5.26657772064209, "learning_rate": 9.25925925925926e-06, "loss": 1.0798, "step": 2000 }, { "epoch": 0.3939631517052136, "grad_norm": 5.231563568115234, "learning_rate": 9.250000000000001e-06, "loss": 1.0944, "step": 2010 }, { "epoch": 0.39592316738533906, "grad_norm": 4.791154384613037, "learning_rate": 9.240740740740742e-06, "loss": 1.0949, "step": 2020 }, { "epoch": 0.3978831830654645, "grad_norm": 4.691703796386719, "learning_rate": 9.231481481481482e-06, "loss": 1.0856, "step": 2030 }, { "epoch": 0.39984319874559, "grad_norm": 7.213362693786621, "learning_rate": 9.222222222222224e-06, "loss": 1.0922, "step": 2040 }, { "epoch": 0.4018032144257154, "grad_norm": 6.0655646324157715, "learning_rate": 9.212962962962963e-06, "loss": 1.0948, "step": 2050 }, { "epoch": 0.40376323010584086, "grad_norm": 4.249541759490967, "learning_rate": 9.203703703703705e-06, "loss": 1.091, "step": 2060 }, { "epoch": 0.4057232457859663, "grad_norm": 4.832751750946045, "learning_rate": 9.194444444444445e-06, "loss": 1.0907, "step": 2070 }, { "epoch": 0.40768326146609174, "grad_norm": 5.216608047485352, "learning_rate": 9.185185185185186e-06, "loss": 1.0802, "step": 2080 }, { "epoch": 0.4096432771462172, "grad_norm": 5.988987445831299, "learning_rate": 9.175925925925928e-06, "loss": 1.0938, "step": 2090 }, { "epoch": 0.4116032928263426, "grad_norm": 4.744123458862305, "learning_rate": 9.166666666666666e-06, "loss": 1.0889, "step": 2100 }, { "epoch": 0.4116032928263426, "eval_loss": 1.08055579662323, "eval_runtime": 14.0694, "eval_samples_per_second": 46.484, "eval_steps_per_second": 5.828, "step": 2100 }, { "epoch": 0.41356330850646805, "grad_norm": 5.0730485916137695, "learning_rate": 9.157407407407409e-06, "loss": 1.0956, "step": 2110 }, { "epoch": 0.4155233241865935, "grad_norm": 4.156872749328613, "learning_rate": 9.148148148148149e-06, "loss": 1.0899, "step": 2120 }, { "epoch": 0.4174833398667189, "grad_norm": 5.194650173187256, "learning_rate": 9.13888888888889e-06, "loss": 1.0772, "step": 2130 }, { "epoch": 0.41944335554684437, "grad_norm": 5.057788848876953, "learning_rate": 9.12962962962963e-06, "loss": 1.0886, "step": 2140 }, { "epoch": 0.4214033712269698, "grad_norm": 4.849510192871094, "learning_rate": 9.120370370370372e-06, "loss": 1.1027, "step": 2150 }, { "epoch": 0.42336338690709524, "grad_norm": 4.8577704429626465, "learning_rate": 9.111111111111112e-06, "loss": 1.0809, "step": 2160 }, { "epoch": 0.4253234025872207, "grad_norm": 4.937686443328857, "learning_rate": 9.101851851851853e-06, "loss": 1.0819, "step": 2170 }, { "epoch": 0.4272834182673461, "grad_norm": 5.295153617858887, "learning_rate": 9.092592592592593e-06, "loss": 1.0734, "step": 2180 }, { "epoch": 0.42924343394747155, "grad_norm": 5.876978874206543, "learning_rate": 9.083333333333333e-06, "loss": 1.078, "step": 2190 }, { "epoch": 0.43120344962759705, "grad_norm": 4.725461006164551, "learning_rate": 9.074074074074075e-06, "loss": 1.0787, "step": 2200 }, { "epoch": 0.4331634653077225, "grad_norm": 4.425513744354248, "learning_rate": 9.064814814814816e-06, "loss": 1.1066, "step": 2210 }, { "epoch": 0.4351234809878479, "grad_norm": 4.781891822814941, "learning_rate": 9.055555555555556e-06, "loss": 1.083, "step": 2220 }, { "epoch": 0.43708349666797336, "grad_norm": 5.292329788208008, "learning_rate": 9.046296296296298e-06, "loss": 1.094, "step": 2230 }, { "epoch": 0.4390435123480988, "grad_norm": 5.246599197387695, "learning_rate": 9.037037037037037e-06, "loss": 1.0903, "step": 2240 }, { "epoch": 0.44100352802822423, "grad_norm": 4.701263427734375, "learning_rate": 9.027777777777779e-06, "loss": 1.0811, "step": 2250 }, { "epoch": 0.44296354370834967, "grad_norm": 5.434783458709717, "learning_rate": 9.01851851851852e-06, "loss": 1.0862, "step": 2260 }, { "epoch": 0.4449235593884751, "grad_norm": 5.898828506469727, "learning_rate": 9.00925925925926e-06, "loss": 1.0972, "step": 2270 }, { "epoch": 0.44688357506860055, "grad_norm": 5.089272975921631, "learning_rate": 9e-06, "loss": 1.091, "step": 2280 }, { "epoch": 0.448843590748726, "grad_norm": 6.887242317199707, "learning_rate": 8.99074074074074e-06, "loss": 1.0785, "step": 2290 }, { "epoch": 0.4508036064288514, "grad_norm": 5.111120700836182, "learning_rate": 8.981481481481483e-06, "loss": 1.0682, "step": 2300 }, { "epoch": 0.45276362210897686, "grad_norm": 5.331275939941406, "learning_rate": 8.972222222222223e-06, "loss": 1.0854, "step": 2310 }, { "epoch": 0.4547236377891023, "grad_norm": 5.59519100189209, "learning_rate": 8.962962962962963e-06, "loss": 1.0883, "step": 2320 }, { "epoch": 0.45668365346922773, "grad_norm": 5.6016364097595215, "learning_rate": 8.953703703703704e-06, "loss": 1.0799, "step": 2330 }, { "epoch": 0.4586436691493532, "grad_norm": 4.625690460205078, "learning_rate": 8.944444444444446e-06, "loss": 1.0716, "step": 2340 }, { "epoch": 0.4606036848294786, "grad_norm": 4.6536688804626465, "learning_rate": 8.935185185185186e-06, "loss": 1.0624, "step": 2350 }, { "epoch": 0.4625637005096041, "grad_norm": 4.807366847991943, "learning_rate": 8.925925925925927e-06, "loss": 1.0829, "step": 2360 }, { "epoch": 0.46452371618972954, "grad_norm": 4.950624942779541, "learning_rate": 8.916666666666667e-06, "loss": 1.0861, "step": 2370 }, { "epoch": 0.466483731869855, "grad_norm": 4.980051517486572, "learning_rate": 8.907407407407408e-06, "loss": 1.0772, "step": 2380 }, { "epoch": 0.4684437475499804, "grad_norm": 5.208264350891113, "learning_rate": 8.89814814814815e-06, "loss": 1.0929, "step": 2390 }, { "epoch": 0.47040376323010585, "grad_norm": 5.6281657218933105, "learning_rate": 8.888888888888888e-06, "loss": 1.0818, "step": 2400 }, { "epoch": 0.4723637789102313, "grad_norm": 6.125396251678467, "learning_rate": 8.87962962962963e-06, "loss": 1.0865, "step": 2410 }, { "epoch": 0.47432379459035673, "grad_norm": 5.989163398742676, "learning_rate": 8.87037037037037e-06, "loss": 1.0685, "step": 2420 }, { "epoch": 0.47628381027048217, "grad_norm": 9.247689247131348, "learning_rate": 8.861111111111111e-06, "loss": 1.0879, "step": 2430 }, { "epoch": 0.4782438259506076, "grad_norm": 4.848996639251709, "learning_rate": 8.851851851851853e-06, "loss": 1.0783, "step": 2440 }, { "epoch": 0.48020384163073304, "grad_norm": 5.04819917678833, "learning_rate": 8.842592592592594e-06, "loss": 1.0679, "step": 2450 }, { "epoch": 0.4821638573108585, "grad_norm": 5.755705833435059, "learning_rate": 8.833333333333334e-06, "loss": 1.0939, "step": 2460 }, { "epoch": 0.4841238729909839, "grad_norm": 4.5991058349609375, "learning_rate": 8.824074074074074e-06, "loss": 1.0775, "step": 2470 }, { "epoch": 0.48608388867110935, "grad_norm": 5.101771354675293, "learning_rate": 8.814814814814817e-06, "loss": 1.084, "step": 2480 }, { "epoch": 0.4880439043512348, "grad_norm": 4.9321441650390625, "learning_rate": 8.805555555555557e-06, "loss": 1.0762, "step": 2490 }, { "epoch": 0.49000392003136023, "grad_norm": 5.363485813140869, "learning_rate": 8.796296296296297e-06, "loss": 1.0595, "step": 2500 }, { "epoch": 0.49196393571148567, "grad_norm": 5.353972434997559, "learning_rate": 8.787037037037038e-06, "loss": 1.0689, "step": 2510 }, { "epoch": 0.49392395139161116, "grad_norm": 5.158483505249023, "learning_rate": 8.777777777777778e-06, "loss": 1.0711, "step": 2520 }, { "epoch": 0.4958839670717366, "grad_norm": 5.160126686096191, "learning_rate": 8.76851851851852e-06, "loss": 1.0886, "step": 2530 }, { "epoch": 0.49784398275186204, "grad_norm": 4.926945686340332, "learning_rate": 8.759259259259259e-06, "loss": 1.0856, "step": 2540 }, { "epoch": 0.4998039984319875, "grad_norm": 5.253294467926025, "learning_rate": 8.750000000000001e-06, "loss": 1.0876, "step": 2550 }, { "epoch": 0.5017640141121129, "grad_norm": 11.674654960632324, "learning_rate": 8.740740740740741e-06, "loss": 1.0784, "step": 2560 }, { "epoch": 0.5037240297922383, "grad_norm": 5.032644271850586, "learning_rate": 8.731481481481482e-06, "loss": 1.07, "step": 2570 }, { "epoch": 0.5056840454723638, "grad_norm": 5.1500372886657715, "learning_rate": 8.722222222222224e-06, "loss": 1.0932, "step": 2580 }, { "epoch": 0.5076440611524892, "grad_norm": 5.761635780334473, "learning_rate": 8.712962962962964e-06, "loss": 1.0816, "step": 2590 }, { "epoch": 0.5096040768326147, "grad_norm": 5.257477760314941, "learning_rate": 8.703703703703705e-06, "loss": 1.0712, "step": 2600 }, { "epoch": 0.5115640925127402, "grad_norm": 5.4931416511535645, "learning_rate": 8.694444444444445e-06, "loss": 1.0832, "step": 2610 }, { "epoch": 0.5135241081928655, "grad_norm": 6.102149963378906, "learning_rate": 8.685185185185185e-06, "loss": 1.076, "step": 2620 }, { "epoch": 0.515484123872991, "grad_norm": 5.596585273742676, "learning_rate": 8.675925925925926e-06, "loss": 1.0972, "step": 2630 }, { "epoch": 0.5174441395531164, "grad_norm": 4.6653523445129395, "learning_rate": 8.666666666666668e-06, "loss": 1.0872, "step": 2640 }, { "epoch": 0.5194041552332419, "grad_norm": 4.938266277313232, "learning_rate": 8.657407407407408e-06, "loss": 1.0717, "step": 2650 }, { "epoch": 0.5213641709133673, "grad_norm": 5.11916446685791, "learning_rate": 8.648148148148149e-06, "loss": 1.0955, "step": 2660 }, { "epoch": 0.5233241865934928, "grad_norm": 4.867392539978027, "learning_rate": 8.63888888888889e-06, "loss": 1.0769, "step": 2670 }, { "epoch": 0.5252842022736182, "grad_norm": 5.589994430541992, "learning_rate": 8.62962962962963e-06, "loss": 1.0808, "step": 2680 }, { "epoch": 0.5272442179537437, "grad_norm": 5.221999645233154, "learning_rate": 8.620370370370371e-06, "loss": 1.0627, "step": 2690 }, { "epoch": 0.529204233633869, "grad_norm": 4.844895839691162, "learning_rate": 8.611111111111112e-06, "loss": 1.0951, "step": 2700 }, { "epoch": 0.5311642493139945, "grad_norm": 4.937107563018799, "learning_rate": 8.601851851851852e-06, "loss": 1.0898, "step": 2710 }, { "epoch": 0.5331242649941199, "grad_norm": 4.758755207061768, "learning_rate": 8.592592592592593e-06, "loss": 1.1017, "step": 2720 }, { "epoch": 0.5350842806742454, "grad_norm": 5.145150661468506, "learning_rate": 8.583333333333333e-06, "loss": 1.0702, "step": 2730 }, { "epoch": 0.5370442963543708, "grad_norm": 5.257258415222168, "learning_rate": 8.574074074074075e-06, "loss": 1.0822, "step": 2740 }, { "epoch": 0.5390043120344963, "grad_norm": 5.826773166656494, "learning_rate": 8.564814814814816e-06, "loss": 1.0714, "step": 2750 }, { "epoch": 0.5409643277146217, "grad_norm": 4.5895304679870605, "learning_rate": 8.555555555555556e-06, "loss": 1.0866, "step": 2760 }, { "epoch": 0.5429243433947472, "grad_norm": 6.960358619689941, "learning_rate": 8.546296296296296e-06, "loss": 1.0942, "step": 2770 }, { "epoch": 0.5448843590748726, "grad_norm": 5.870516300201416, "learning_rate": 8.537037037037038e-06, "loss": 1.0937, "step": 2780 }, { "epoch": 0.546844374754998, "grad_norm": 4.99727725982666, "learning_rate": 8.527777777777779e-06, "loss": 1.0851, "step": 2790 }, { "epoch": 0.5488043904351235, "grad_norm": 4.9571003913879395, "learning_rate": 8.518518518518519e-06, "loss": 1.085, "step": 2800 }, { "epoch": 0.5488043904351235, "eval_loss": 1.073391318321228, "eval_runtime": 14.0283, "eval_samples_per_second": 46.62, "eval_steps_per_second": 5.845, "step": 2800 }, { "epoch": 0.5507644061152489, "grad_norm": 4.457803726196289, "learning_rate": 8.509259259259261e-06, "loss": 1.0787, "step": 2810 }, { "epoch": 0.5527244217953744, "grad_norm": 5.098880290985107, "learning_rate": 8.5e-06, "loss": 1.0672, "step": 2820 }, { "epoch": 0.5546844374754998, "grad_norm": 5.3498735427856445, "learning_rate": 8.490740740740742e-06, "loss": 1.0852, "step": 2830 }, { "epoch": 0.5566444531556253, "grad_norm": 5.1486921310424805, "learning_rate": 8.481481481481482e-06, "loss": 1.0564, "step": 2840 }, { "epoch": 0.5586044688357507, "grad_norm": 5.613958358764648, "learning_rate": 8.472222222222223e-06, "loss": 1.0864, "step": 2850 }, { "epoch": 0.5605644845158761, "grad_norm": 5.687408924102783, "learning_rate": 8.462962962962963e-06, "loss": 1.0871, "step": 2860 }, { "epoch": 0.5625245001960015, "grad_norm": 5.822880268096924, "learning_rate": 8.453703703703704e-06, "loss": 1.0708, "step": 2870 }, { "epoch": 0.564484515876127, "grad_norm": 5.982561111450195, "learning_rate": 8.444444444444446e-06, "loss": 1.0633, "step": 2880 }, { "epoch": 0.5664445315562524, "grad_norm": 4.887526035308838, "learning_rate": 8.435185185185186e-06, "loss": 1.0823, "step": 2890 }, { "epoch": 0.5684045472363779, "grad_norm": 5.810245037078857, "learning_rate": 8.425925925925926e-06, "loss": 1.0683, "step": 2900 }, { "epoch": 0.5703645629165033, "grad_norm": 4.757987022399902, "learning_rate": 8.416666666666667e-06, "loss": 1.0703, "step": 2910 }, { "epoch": 0.5723245785966288, "grad_norm": 5.597621440887451, "learning_rate": 8.407407407407409e-06, "loss": 1.0861, "step": 2920 }, { "epoch": 0.5742845942767543, "grad_norm": 9.786075592041016, "learning_rate": 8.39814814814815e-06, "loss": 1.0755, "step": 2930 }, { "epoch": 0.5762446099568796, "grad_norm": 4.969895362854004, "learning_rate": 8.38888888888889e-06, "loss": 1.0907, "step": 2940 }, { "epoch": 0.5782046256370051, "grad_norm": 5.833816051483154, "learning_rate": 8.37962962962963e-06, "loss": 1.0679, "step": 2950 }, { "epoch": 0.5801646413171305, "grad_norm": 5.07546329498291, "learning_rate": 8.37037037037037e-06, "loss": 1.0705, "step": 2960 }, { "epoch": 0.582124656997256, "grad_norm": 5.735626220703125, "learning_rate": 8.361111111111113e-06, "loss": 1.0676, "step": 2970 }, { "epoch": 0.5840846726773814, "grad_norm": 4.732352256774902, "learning_rate": 8.351851851851851e-06, "loss": 1.0809, "step": 2980 }, { "epoch": 0.5860446883575069, "grad_norm": 6.006665229797363, "learning_rate": 8.342592592592593e-06, "loss": 1.0624, "step": 2990 }, { "epoch": 0.5880047040376323, "grad_norm": 6.649717330932617, "learning_rate": 8.333333333333334e-06, "loss": 1.072, "step": 3000 }, { "epoch": 0.5899647197177578, "grad_norm": 5.156081199645996, "learning_rate": 8.324074074074074e-06, "loss": 1.0791, "step": 3010 }, { "epoch": 0.5919247353978832, "grad_norm": 5.489958763122559, "learning_rate": 8.314814814814816e-06, "loss": 1.0803, "step": 3020 }, { "epoch": 0.5938847510780086, "grad_norm": 5.912964344024658, "learning_rate": 8.305555555555557e-06, "loss": 1.0888, "step": 3030 }, { "epoch": 0.595844766758134, "grad_norm": 5.113432884216309, "learning_rate": 8.296296296296297e-06, "loss": 1.0728, "step": 3040 }, { "epoch": 0.5978047824382595, "grad_norm": 4.774835109710693, "learning_rate": 8.287037037037037e-06, "loss": 1.0772, "step": 3050 }, { "epoch": 0.5997647981183849, "grad_norm": 5.954660892486572, "learning_rate": 8.277777777777778e-06, "loss": 1.0805, "step": 3060 }, { "epoch": 0.6017248137985104, "grad_norm": 4.933722019195557, "learning_rate": 8.26851851851852e-06, "loss": 1.0678, "step": 3070 }, { "epoch": 0.6036848294786358, "grad_norm": 5.152841567993164, "learning_rate": 8.25925925925926e-06, "loss": 1.0831, "step": 3080 }, { "epoch": 0.6056448451587613, "grad_norm": 5.011537551879883, "learning_rate": 8.25e-06, "loss": 1.0841, "step": 3090 }, { "epoch": 0.6076048608388868, "grad_norm": 4.196855545043945, "learning_rate": 8.240740740740741e-06, "loss": 1.0647, "step": 3100 }, { "epoch": 0.6095648765190121, "grad_norm": 5.21035099029541, "learning_rate": 8.231481481481483e-06, "loss": 1.049, "step": 3110 }, { "epoch": 0.6115248921991376, "grad_norm": 26.132904052734375, "learning_rate": 8.222222222222222e-06, "loss": 1.0799, "step": 3120 }, { "epoch": 0.613484907879263, "grad_norm": 6.4696760177612305, "learning_rate": 8.212962962962964e-06, "loss": 1.0666, "step": 3130 }, { "epoch": 0.6154449235593885, "grad_norm": 4.687003135681152, "learning_rate": 8.203703703703704e-06, "loss": 1.0732, "step": 3140 }, { "epoch": 0.6174049392395139, "grad_norm": 5.188929557800293, "learning_rate": 8.194444444444445e-06, "loss": 1.0911, "step": 3150 }, { "epoch": 0.6193649549196394, "grad_norm": 4.781430244445801, "learning_rate": 8.185185185185187e-06, "loss": 1.0705, "step": 3160 }, { "epoch": 0.6213249705997648, "grad_norm": 4.717843055725098, "learning_rate": 8.175925925925925e-06, "loss": 1.0822, "step": 3170 }, { "epoch": 0.6232849862798903, "grad_norm": 5.404903888702393, "learning_rate": 8.166666666666668e-06, "loss": 1.0917, "step": 3180 }, { "epoch": 0.6252450019600156, "grad_norm": 5.684728145599365, "learning_rate": 8.157407407407408e-06, "loss": 1.0842, "step": 3190 }, { "epoch": 0.6272050176401411, "grad_norm": 4.7895965576171875, "learning_rate": 8.148148148148148e-06, "loss": 1.0652, "step": 3200 }, { "epoch": 0.6291650333202665, "grad_norm": 6.572015285491943, "learning_rate": 8.138888888888889e-06, "loss": 1.0959, "step": 3210 }, { "epoch": 0.631125049000392, "grad_norm": 4.640923976898193, "learning_rate": 8.12962962962963e-06, "loss": 1.0896, "step": 3220 }, { "epoch": 0.6330850646805174, "grad_norm": 4.732535362243652, "learning_rate": 8.120370370370371e-06, "loss": 1.0646, "step": 3230 }, { "epoch": 0.6350450803606429, "grad_norm": 5.692999362945557, "learning_rate": 8.111111111111112e-06, "loss": 1.0774, "step": 3240 }, { "epoch": 0.6370050960407684, "grad_norm": 5.28448486328125, "learning_rate": 8.101851851851854e-06, "loss": 1.0844, "step": 3250 }, { "epoch": 0.6389651117208938, "grad_norm": 3.9686877727508545, "learning_rate": 8.092592592592592e-06, "loss": 1.0626, "step": 3260 }, { "epoch": 0.6409251274010193, "grad_norm": 6.364278793334961, "learning_rate": 8.083333333333334e-06, "loss": 1.0696, "step": 3270 }, { "epoch": 0.6428851430811446, "grad_norm": 4.789191722869873, "learning_rate": 8.074074074074075e-06, "loss": 1.0696, "step": 3280 }, { "epoch": 0.6448451587612701, "grad_norm": 4.61089563369751, "learning_rate": 8.064814814814815e-06, "loss": 1.0701, "step": 3290 }, { "epoch": 0.6468051744413955, "grad_norm": 4.959714889526367, "learning_rate": 8.055555555555557e-06, "loss": 1.0852, "step": 3300 }, { "epoch": 0.648765190121521, "grad_norm": 4.559225082397461, "learning_rate": 8.046296296296296e-06, "loss": 1.0708, "step": 3310 }, { "epoch": 0.6507252058016464, "grad_norm": 4.9751691818237305, "learning_rate": 8.037037037037038e-06, "loss": 1.0545, "step": 3320 }, { "epoch": 0.6526852214817719, "grad_norm": 5.153299808502197, "learning_rate": 8.027777777777778e-06, "loss": 1.074, "step": 3330 }, { "epoch": 0.6546452371618973, "grad_norm": 5.507195472717285, "learning_rate": 8.018518518518519e-06, "loss": 1.0772, "step": 3340 }, { "epoch": 0.6566052528420228, "grad_norm": 4.8530449867248535, "learning_rate": 8.00925925925926e-06, "loss": 1.087, "step": 3350 }, { "epoch": 0.6585652685221481, "grad_norm": 4.754231929779053, "learning_rate": 8.000000000000001e-06, "loss": 1.0792, "step": 3360 }, { "epoch": 0.6605252842022736, "grad_norm": 4.642824172973633, "learning_rate": 7.990740740740742e-06, "loss": 1.0681, "step": 3370 }, { "epoch": 0.662485299882399, "grad_norm": 5.338252067565918, "learning_rate": 7.981481481481482e-06, "loss": 1.0579, "step": 3380 }, { "epoch": 0.6644453155625245, "grad_norm": 5.290416240692139, "learning_rate": 7.972222222222224e-06, "loss": 1.0523, "step": 3390 }, { "epoch": 0.6664053312426499, "grad_norm": 4.5010294914245605, "learning_rate": 7.962962962962963e-06, "loss": 1.0694, "step": 3400 }, { "epoch": 0.6683653469227754, "grad_norm": 5.253542900085449, "learning_rate": 7.953703703703705e-06, "loss": 1.0745, "step": 3410 }, { "epoch": 0.6703253626029009, "grad_norm": 4.743661880493164, "learning_rate": 7.944444444444445e-06, "loss": 1.0586, "step": 3420 }, { "epoch": 0.6722853782830263, "grad_norm": 5.1141839027404785, "learning_rate": 7.935185185185186e-06, "loss": 1.0767, "step": 3430 }, { "epoch": 0.6742453939631518, "grad_norm": 4.729719638824463, "learning_rate": 7.925925925925926e-06, "loss": 1.0841, "step": 3440 }, { "epoch": 0.6762054096432771, "grad_norm": 5.445465087890625, "learning_rate": 7.916666666666667e-06, "loss": 1.0901, "step": 3450 }, { "epoch": 0.6781654253234026, "grad_norm": 4.362582683563232, "learning_rate": 7.907407407407409e-06, "loss": 1.0634, "step": 3460 }, { "epoch": 0.680125441003528, "grad_norm": 4.604685306549072, "learning_rate": 7.898148148148149e-06, "loss": 1.0631, "step": 3470 }, { "epoch": 0.6820854566836535, "grad_norm": 5.426617622375488, "learning_rate": 7.88888888888889e-06, "loss": 1.0691, "step": 3480 }, { "epoch": 0.6840454723637789, "grad_norm": 4.578293323516846, "learning_rate": 7.87962962962963e-06, "loss": 1.0761, "step": 3490 }, { "epoch": 0.6860054880439044, "grad_norm": 5.362606525421143, "learning_rate": 7.870370370370372e-06, "loss": 1.0836, "step": 3500 }, { "epoch": 0.6860054880439044, "eval_loss": 1.0641947984695435, "eval_runtime": 14.0879, "eval_samples_per_second": 46.423, "eval_steps_per_second": 5.821, "step": 3500 }, { "epoch": 0.6879655037240298, "grad_norm": 4.338893890380859, "learning_rate": 7.861111111111112e-06, "loss": 1.0604, "step": 3510 }, { "epoch": 0.6899255194041553, "grad_norm": 5.606161594390869, "learning_rate": 7.851851851851853e-06, "loss": 1.0869, "step": 3520 }, { "epoch": 0.6918855350842806, "grad_norm": 4.934096336364746, "learning_rate": 7.842592592592593e-06, "loss": 1.07, "step": 3530 }, { "epoch": 0.6938455507644061, "grad_norm": 4.351364612579346, "learning_rate": 7.833333333333333e-06, "loss": 1.0682, "step": 3540 }, { "epoch": 0.6958055664445315, "grad_norm": 4.8932976722717285, "learning_rate": 7.824074074074076e-06, "loss": 1.0766, "step": 3550 }, { "epoch": 0.697765582124657, "grad_norm": 4.501565456390381, "learning_rate": 7.814814814814816e-06, "loss": 1.0612, "step": 3560 }, { "epoch": 0.6997255978047824, "grad_norm": 4.929937362670898, "learning_rate": 7.805555555555556e-06, "loss": 1.0609, "step": 3570 }, { "epoch": 0.7016856134849079, "grad_norm": 5.501171112060547, "learning_rate": 7.796296296296297e-06, "loss": 1.0583, "step": 3580 }, { "epoch": 0.7036456291650334, "grad_norm": 4.9919891357421875, "learning_rate": 7.787037037037037e-06, "loss": 1.0603, "step": 3590 }, { "epoch": 0.7056056448451588, "grad_norm": 4.892634868621826, "learning_rate": 7.77777777777778e-06, "loss": 1.0718, "step": 3600 }, { "epoch": 0.7075656605252842, "grad_norm": 5.0666632652282715, "learning_rate": 7.76851851851852e-06, "loss": 1.0733, "step": 3610 }, { "epoch": 0.7095256762054096, "grad_norm": 4.797816753387451, "learning_rate": 7.75925925925926e-06, "loss": 1.0664, "step": 3620 }, { "epoch": 0.7114856918855351, "grad_norm": 5.004721164703369, "learning_rate": 7.75e-06, "loss": 1.0847, "step": 3630 }, { "epoch": 0.7134457075656605, "grad_norm": 4.80883264541626, "learning_rate": 7.74074074074074e-06, "loss": 1.075, "step": 3640 }, { "epoch": 0.715405723245786, "grad_norm": 5.439119815826416, "learning_rate": 7.731481481481483e-06, "loss": 1.0517, "step": 3650 }, { "epoch": 0.7173657389259114, "grad_norm": 4.244631290435791, "learning_rate": 7.722222222222223e-06, "loss": 1.0836, "step": 3660 }, { "epoch": 0.7193257546060369, "grad_norm": 4.710737705230713, "learning_rate": 7.712962962962964e-06, "loss": 1.0748, "step": 3670 }, { "epoch": 0.7212857702861623, "grad_norm": 5.23133659362793, "learning_rate": 7.703703703703704e-06, "loss": 1.0843, "step": 3680 }, { "epoch": 0.7232457859662877, "grad_norm": 4.706710338592529, "learning_rate": 7.694444444444446e-06, "loss": 1.0637, "step": 3690 }, { "epoch": 0.7252058016464131, "grad_norm": 4.4408955574035645, "learning_rate": 7.685185185185185e-06, "loss": 1.0945, "step": 3700 }, { "epoch": 0.7271658173265386, "grad_norm": 5.1011199951171875, "learning_rate": 7.675925925925927e-06, "loss": 1.0573, "step": 3710 }, { "epoch": 0.729125833006664, "grad_norm": 4.85494327545166, "learning_rate": 7.666666666666667e-06, "loss": 1.0686, "step": 3720 }, { "epoch": 0.7310858486867895, "grad_norm": 4.990062236785889, "learning_rate": 7.657407407407408e-06, "loss": 1.052, "step": 3730 }, { "epoch": 0.733045864366915, "grad_norm": 4.779956817626953, "learning_rate": 7.64814814814815e-06, "loss": 1.0619, "step": 3740 }, { "epoch": 0.7350058800470404, "grad_norm": 4.921076774597168, "learning_rate": 7.638888888888888e-06, "loss": 1.0834, "step": 3750 }, { "epoch": 0.7369658957271659, "grad_norm": 4.729534149169922, "learning_rate": 7.62962962962963e-06, "loss": 1.0643, "step": 3760 }, { "epoch": 0.7389259114072912, "grad_norm": 4.321104049682617, "learning_rate": 7.620370370370372e-06, "loss": 1.0736, "step": 3770 }, { "epoch": 0.7408859270874167, "grad_norm": 5.252387523651123, "learning_rate": 7.611111111111111e-06, "loss": 1.0515, "step": 3780 }, { "epoch": 0.7428459427675421, "grad_norm": 4.770890712738037, "learning_rate": 7.6018518518518525e-06, "loss": 1.0865, "step": 3790 }, { "epoch": 0.7448059584476676, "grad_norm": 4.786081790924072, "learning_rate": 7.592592592592594e-06, "loss": 1.0727, "step": 3800 }, { "epoch": 0.746765974127793, "grad_norm": 5.215248107910156, "learning_rate": 7.583333333333333e-06, "loss": 1.0657, "step": 3810 }, { "epoch": 0.7487259898079185, "grad_norm": 8.386008262634277, "learning_rate": 7.5740740740740745e-06, "loss": 1.0774, "step": 3820 }, { "epoch": 0.7506860054880439, "grad_norm": 5.009541034698486, "learning_rate": 7.564814814814816e-06, "loss": 1.0517, "step": 3830 }, { "epoch": 0.7526460211681694, "grad_norm": 5.512828826904297, "learning_rate": 7.555555555555556e-06, "loss": 1.0661, "step": 3840 }, { "epoch": 0.7546060368482947, "grad_norm": 4.983898162841797, "learning_rate": 7.546296296296297e-06, "loss": 1.0561, "step": 3850 }, { "epoch": 0.7565660525284202, "grad_norm": 4.687797546386719, "learning_rate": 7.537037037037037e-06, "loss": 1.0623, "step": 3860 }, { "epoch": 0.7585260682085456, "grad_norm": 4.924409866333008, "learning_rate": 7.527777777777778e-06, "loss": 1.0557, "step": 3870 }, { "epoch": 0.7604860838886711, "grad_norm": 4.704287528991699, "learning_rate": 7.518518518518519e-06, "loss": 1.0732, "step": 3880 }, { "epoch": 0.7624460995687965, "grad_norm": 5.610473155975342, "learning_rate": 7.50925925925926e-06, "loss": 1.0827, "step": 3890 }, { "epoch": 0.764406115248922, "grad_norm": 5.7728590965271, "learning_rate": 7.500000000000001e-06, "loss": 1.0738, "step": 3900 }, { "epoch": 0.7663661309290475, "grad_norm": 4.7245659828186035, "learning_rate": 7.4907407407407414e-06, "loss": 1.0665, "step": 3910 }, { "epoch": 0.7683261466091729, "grad_norm": 6.519947528839111, "learning_rate": 7.481481481481482e-06, "loss": 1.0631, "step": 3920 }, { "epoch": 0.7702861622892984, "grad_norm": 4.989842891693115, "learning_rate": 7.472222222222223e-06, "loss": 1.053, "step": 3930 }, { "epoch": 0.7722461779694237, "grad_norm": 4.687224864959717, "learning_rate": 7.462962962962964e-06, "loss": 1.0593, "step": 3940 }, { "epoch": 0.7742061936495492, "grad_norm": 4.693930625915527, "learning_rate": 7.453703703703704e-06, "loss": 1.0665, "step": 3950 }, { "epoch": 0.7761662093296746, "grad_norm": 5.265392303466797, "learning_rate": 7.444444444444445e-06, "loss": 1.042, "step": 3960 }, { "epoch": 0.7781262250098001, "grad_norm": 5.122781753540039, "learning_rate": 7.4351851851851855e-06, "loss": 1.0748, "step": 3970 }, { "epoch": 0.7800862406899255, "grad_norm": 4.934235095977783, "learning_rate": 7.425925925925927e-06, "loss": 1.0603, "step": 3980 }, { "epoch": 0.782046256370051, "grad_norm": 4.4583587646484375, "learning_rate": 7.416666666666668e-06, "loss": 1.0724, "step": 3990 }, { "epoch": 0.7840062720501764, "grad_norm": 5.3386030197143555, "learning_rate": 7.4074074074074075e-06, "loss": 1.0651, "step": 4000 }, { "epoch": 0.7859662877303019, "grad_norm": 4.659664154052734, "learning_rate": 7.398148148148149e-06, "loss": 1.0525, "step": 4010 }, { "epoch": 0.7879263034104272, "grad_norm": 4.794917106628418, "learning_rate": 7.38888888888889e-06, "loss": 1.0561, "step": 4020 }, { "epoch": 0.7898863190905527, "grad_norm": 5.210708141326904, "learning_rate": 7.3796296296296295e-06, "loss": 1.0641, "step": 4030 }, { "epoch": 0.7918463347706781, "grad_norm": 4.441596031188965, "learning_rate": 7.370370370370371e-06, "loss": 1.0696, "step": 4040 }, { "epoch": 0.7938063504508036, "grad_norm": 5.239253044128418, "learning_rate": 7.361111111111112e-06, "loss": 1.0576, "step": 4050 }, { "epoch": 0.795766366130929, "grad_norm": 5.195384502410889, "learning_rate": 7.351851851851852e-06, "loss": 1.0556, "step": 4060 }, { "epoch": 0.7977263818110545, "grad_norm": 5.0446648597717285, "learning_rate": 7.342592592592594e-06, "loss": 1.067, "step": 4070 }, { "epoch": 0.79968639749118, "grad_norm": 5.061046600341797, "learning_rate": 7.333333333333333e-06, "loss": 1.0587, "step": 4080 }, { "epoch": 0.8016464131713054, "grad_norm": 7.661867141723633, "learning_rate": 7.324074074074074e-06, "loss": 1.0694, "step": 4090 }, { "epoch": 0.8036064288514309, "grad_norm": 4.821944236755371, "learning_rate": 7.314814814814816e-06, "loss": 1.0675, "step": 4100 }, { "epoch": 0.8055664445315562, "grad_norm": 4.810634613037109, "learning_rate": 7.305555555555556e-06, "loss": 1.0589, "step": 4110 }, { "epoch": 0.8075264602116817, "grad_norm": 5.0643157958984375, "learning_rate": 7.296296296296297e-06, "loss": 1.0714, "step": 4120 }, { "epoch": 0.8094864758918071, "grad_norm": 4.812649250030518, "learning_rate": 7.287037037037038e-06, "loss": 1.0635, "step": 4130 }, { "epoch": 0.8114464915719326, "grad_norm": 4.7638840675354, "learning_rate": 7.277777777777778e-06, "loss": 1.064, "step": 4140 }, { "epoch": 0.813406507252058, "grad_norm": 4.386782646179199, "learning_rate": 7.268518518518519e-06, "loss": 1.0726, "step": 4150 }, { "epoch": 0.8153665229321835, "grad_norm": 4.593876838684082, "learning_rate": 7.2592592592592605e-06, "loss": 1.0542, "step": 4160 }, { "epoch": 0.8173265386123089, "grad_norm": 4.822241306304932, "learning_rate": 7.25e-06, "loss": 1.0541, "step": 4170 }, { "epoch": 0.8192865542924344, "grad_norm": 4.806110858917236, "learning_rate": 7.240740740740741e-06, "loss": 1.0522, "step": 4180 }, { "epoch": 0.8212465699725597, "grad_norm": 4.5700788497924805, "learning_rate": 7.231481481481482e-06, "loss": 1.0672, "step": 4190 }, { "epoch": 0.8232065856526852, "grad_norm": 4.629611492156982, "learning_rate": 7.222222222222223e-06, "loss": 1.072, "step": 4200 }, { "epoch": 0.8232065856526852, "eval_loss": 1.0551141500473022, "eval_runtime": 13.9585, "eval_samples_per_second": 46.853, "eval_steps_per_second": 5.875, "step": 4200 }, { "epoch": 0.8251666013328106, "grad_norm": 5.06083345413208, "learning_rate": 7.212962962962964e-06, "loss": 1.0641, "step": 4210 }, { "epoch": 0.8271266170129361, "grad_norm": 4.53200626373291, "learning_rate": 7.203703703703704e-06, "loss": 1.0764, "step": 4220 }, { "epoch": 0.8290866326930616, "grad_norm": 4.572458744049072, "learning_rate": 7.194444444444445e-06, "loss": 1.0616, "step": 4230 }, { "epoch": 0.831046648373187, "grad_norm": 21.508056640625, "learning_rate": 7.185185185185186e-06, "loss": 1.0606, "step": 4240 }, { "epoch": 0.8330066640533125, "grad_norm": 5.574312210083008, "learning_rate": 7.1759259259259266e-06, "loss": 1.0736, "step": 4250 }, { "epoch": 0.8349666797334379, "grad_norm": 4.778395175933838, "learning_rate": 7.166666666666667e-06, "loss": 1.0561, "step": 4260 }, { "epoch": 0.8369266954135633, "grad_norm": 4.696834564208984, "learning_rate": 7.157407407407408e-06, "loss": 1.0753, "step": 4270 }, { "epoch": 0.8388867110936887, "grad_norm": 4.8193769454956055, "learning_rate": 7.1481481481481486e-06, "loss": 1.0454, "step": 4280 }, { "epoch": 0.8408467267738142, "grad_norm": 4.693731784820557, "learning_rate": 7.13888888888889e-06, "loss": 1.0634, "step": 4290 }, { "epoch": 0.8428067424539396, "grad_norm": 5.045989990234375, "learning_rate": 7.129629629629629e-06, "loss": 1.057, "step": 4300 }, { "epoch": 0.8447667581340651, "grad_norm": 5.245114326477051, "learning_rate": 7.120370370370371e-06, "loss": 1.0435, "step": 4310 }, { "epoch": 0.8467267738141905, "grad_norm": 5.094894886016846, "learning_rate": 7.111111111111112e-06, "loss": 1.0657, "step": 4320 }, { "epoch": 0.848686789494316, "grad_norm": 4.630204200744629, "learning_rate": 7.101851851851852e-06, "loss": 1.0729, "step": 4330 }, { "epoch": 0.8506468051744414, "grad_norm": 4.463089942932129, "learning_rate": 7.0925925925925935e-06, "loss": 1.0573, "step": 4340 }, { "epoch": 0.8526068208545668, "grad_norm": 4.479245185852051, "learning_rate": 7.083333333333335e-06, "loss": 1.0534, "step": 4350 }, { "epoch": 0.8545668365346922, "grad_norm": 4.832218170166016, "learning_rate": 7.074074074074074e-06, "loss": 1.0499, "step": 4360 }, { "epoch": 0.8565268522148177, "grad_norm": 4.618699550628662, "learning_rate": 7.0648148148148155e-06, "loss": 1.073, "step": 4370 }, { "epoch": 0.8584868678949431, "grad_norm": 4.646695137023926, "learning_rate": 7.055555555555557e-06, "loss": 1.0662, "step": 4380 }, { "epoch": 0.8604468835750686, "grad_norm": 4.74224853515625, "learning_rate": 7.046296296296296e-06, "loss": 1.0516, "step": 4390 }, { "epoch": 0.8624068992551941, "grad_norm": 4.836453914642334, "learning_rate": 7.0370370370370375e-06, "loss": 1.0622, "step": 4400 }, { "epoch": 0.8643669149353195, "grad_norm": 5.0554046630859375, "learning_rate": 7.027777777777778e-06, "loss": 1.0725, "step": 4410 }, { "epoch": 0.866326930615445, "grad_norm": 4.8123579025268555, "learning_rate": 7.018518518518519e-06, "loss": 1.0451, "step": 4420 }, { "epoch": 0.8682869462955703, "grad_norm": 4.815413951873779, "learning_rate": 7.00925925925926e-06, "loss": 1.057, "step": 4430 }, { "epoch": 0.8702469619756958, "grad_norm": 5.103263854980469, "learning_rate": 7e-06, "loss": 1.073, "step": 4440 }, { "epoch": 0.8722069776558212, "grad_norm": 5.115508556365967, "learning_rate": 6.990740740740741e-06, "loss": 1.058, "step": 4450 }, { "epoch": 0.8741669933359467, "grad_norm": 4.976015567779541, "learning_rate": 6.981481481481482e-06, "loss": 1.0601, "step": 4460 }, { "epoch": 0.8761270090160721, "grad_norm": 5.111723899841309, "learning_rate": 6.972222222222223e-06, "loss": 1.0735, "step": 4470 }, { "epoch": 0.8780870246961976, "grad_norm": 4.711007118225098, "learning_rate": 6.962962962962964e-06, "loss": 1.0723, "step": 4480 }, { "epoch": 0.880047040376323, "grad_norm": 4.8820881843566895, "learning_rate": 6.953703703703704e-06, "loss": 1.0562, "step": 4490 }, { "epoch": 0.8820070560564485, "grad_norm": 5.674580097198486, "learning_rate": 6.944444444444445e-06, "loss": 1.0657, "step": 4500 }, { "epoch": 0.8839670717365739, "grad_norm": 4.517936706542969, "learning_rate": 6.935185185185186e-06, "loss": 1.0604, "step": 4510 }, { "epoch": 0.8859270874166993, "grad_norm": 5.491835594177246, "learning_rate": 6.9259259259259256e-06, "loss": 1.0635, "step": 4520 }, { "epoch": 0.8878871030968247, "grad_norm": 5.377597808837891, "learning_rate": 6.916666666666667e-06, "loss": 1.0567, "step": 4530 }, { "epoch": 0.8898471187769502, "grad_norm": 5.251382350921631, "learning_rate": 6.907407407407408e-06, "loss": 1.0564, "step": 4540 }, { "epoch": 0.8918071344570757, "grad_norm": 4.623981952667236, "learning_rate": 6.898148148148148e-06, "loss": 1.0663, "step": 4550 }, { "epoch": 0.8937671501372011, "grad_norm": 4.460480690002441, "learning_rate": 6.88888888888889e-06, "loss": 1.0773, "step": 4560 }, { "epoch": 0.8957271658173266, "grad_norm": 4.909152984619141, "learning_rate": 6.879629629629631e-06, "loss": 1.0462, "step": 4570 }, { "epoch": 0.897687181497452, "grad_norm": 4.5406951904296875, "learning_rate": 6.8703703703703704e-06, "loss": 1.0619, "step": 4580 }, { "epoch": 0.8996471971775775, "grad_norm": 4.8268890380859375, "learning_rate": 6.861111111111112e-06, "loss": 1.0749, "step": 4590 }, { "epoch": 0.9016072128577028, "grad_norm": 5.024204254150391, "learning_rate": 6.851851851851853e-06, "loss": 1.082, "step": 4600 }, { "epoch": 0.9035672285378283, "grad_norm": 4.6898908615112305, "learning_rate": 6.842592592592593e-06, "loss": 1.0527, "step": 4610 }, { "epoch": 0.9055272442179537, "grad_norm": 5.445582866668701, "learning_rate": 6.833333333333334e-06, "loss": 1.05, "step": 4620 }, { "epoch": 0.9074872598980792, "grad_norm": 4.625185966491699, "learning_rate": 6.824074074074075e-06, "loss": 1.0623, "step": 4630 }, { "epoch": 0.9094472755782046, "grad_norm": 4.991087913513184, "learning_rate": 6.814814814814815e-06, "loss": 1.048, "step": 4640 }, { "epoch": 0.9114072912583301, "grad_norm": 5.890523433685303, "learning_rate": 6.8055555555555566e-06, "loss": 1.0539, "step": 4650 }, { "epoch": 0.9133673069384555, "grad_norm": 4.583831787109375, "learning_rate": 6.796296296296296e-06, "loss": 1.0616, "step": 4660 }, { "epoch": 0.915327322618581, "grad_norm": 5.098804473876953, "learning_rate": 6.787037037037037e-06, "loss": 1.0555, "step": 4670 }, { "epoch": 0.9172873382987063, "grad_norm": 4.51339864730835, "learning_rate": 6.777777777777779e-06, "loss": 1.0563, "step": 4680 }, { "epoch": 0.9192473539788318, "grad_norm": 5.437107086181641, "learning_rate": 6.768518518518519e-06, "loss": 1.0639, "step": 4690 }, { "epoch": 0.9212073696589572, "grad_norm": 5.220668315887451, "learning_rate": 6.75925925925926e-06, "loss": 1.0335, "step": 4700 }, { "epoch": 0.9231673853390827, "grad_norm": 5.214885711669922, "learning_rate": 6.750000000000001e-06, "loss": 1.0616, "step": 4710 }, { "epoch": 0.9251274010192082, "grad_norm": 4.807358264923096, "learning_rate": 6.740740740740741e-06, "loss": 1.0556, "step": 4720 }, { "epoch": 0.9270874166993336, "grad_norm": 4.536919116973877, "learning_rate": 6.731481481481482e-06, "loss": 1.0427, "step": 4730 }, { "epoch": 0.9290474323794591, "grad_norm": 4.960407733917236, "learning_rate": 6.7222222222222235e-06, "loss": 1.0573, "step": 4740 }, { "epoch": 0.9310074480595845, "grad_norm": 5.058611869812012, "learning_rate": 6.712962962962963e-06, "loss": 1.0529, "step": 4750 }, { "epoch": 0.93296746373971, "grad_norm": 4.611677169799805, "learning_rate": 6.703703703703704e-06, "loss": 1.0609, "step": 4760 }, { "epoch": 0.9349274794198353, "grad_norm": 4.662081241607666, "learning_rate": 6.694444444444445e-06, "loss": 1.0585, "step": 4770 }, { "epoch": 0.9368874950999608, "grad_norm": 4.845633029937744, "learning_rate": 6.685185185185186e-06, "loss": 1.0591, "step": 4780 }, { "epoch": 0.9388475107800862, "grad_norm": 4.63094425201416, "learning_rate": 6.675925925925927e-06, "loss": 1.051, "step": 4790 }, { "epoch": 0.9408075264602117, "grad_norm": 5.272851467132568, "learning_rate": 6.666666666666667e-06, "loss": 1.0643, "step": 4800 }, { "epoch": 0.9427675421403371, "grad_norm": 5.056580066680908, "learning_rate": 6.657407407407408e-06, "loss": 1.0483, "step": 4810 }, { "epoch": 0.9447275578204626, "grad_norm": 4.985904693603516, "learning_rate": 6.648148148148149e-06, "loss": 1.0526, "step": 4820 }, { "epoch": 0.946687573500588, "grad_norm": 5.002072334289551, "learning_rate": 6.6388888888888895e-06, "loss": 1.0693, "step": 4830 }, { "epoch": 0.9486475891807135, "grad_norm": 5.099184036254883, "learning_rate": 6.62962962962963e-06, "loss": 1.0629, "step": 4840 }, { "epoch": 0.9506076048608388, "grad_norm": 4.710622310638428, "learning_rate": 6.620370370370371e-06, "loss": 1.0458, "step": 4850 }, { "epoch": 0.9525676205409643, "grad_norm": 5.143291473388672, "learning_rate": 6.6111111111111115e-06, "loss": 1.0552, "step": 4860 }, { "epoch": 0.9545276362210897, "grad_norm": 5.28003454208374, "learning_rate": 6.601851851851853e-06, "loss": 1.0643, "step": 4870 }, { "epoch": 0.9564876519012152, "grad_norm": 5.275862216949463, "learning_rate": 6.592592592592592e-06, "loss": 1.0566, "step": 4880 }, { "epoch": 0.9584476675813407, "grad_norm": 4.868998050689697, "learning_rate": 6.5833333333333335e-06, "loss": 1.0703, "step": 4890 }, { "epoch": 0.9604076832614661, "grad_norm": 4.964428424835205, "learning_rate": 6.574074074074075e-06, "loss": 1.0564, "step": 4900 }, { "epoch": 0.9604076832614661, "eval_loss": 1.0497612953186035, "eval_runtime": 14.0356, "eval_samples_per_second": 46.596, "eval_steps_per_second": 5.842, "step": 4900 }, { "epoch": 0.9623676989415916, "grad_norm": 4.545341968536377, "learning_rate": 6.564814814814815e-06, "loss": 1.0741, "step": 4910 }, { "epoch": 0.964327714621717, "grad_norm": 5.679270267486572, "learning_rate": 6.555555555555556e-06, "loss": 1.0528, "step": 4920 }, { "epoch": 0.9662877303018425, "grad_norm": 4.5773749351501465, "learning_rate": 6.546296296296298e-06, "loss": 1.048, "step": 4930 }, { "epoch": 0.9682477459819678, "grad_norm": 5.288699150085449, "learning_rate": 6.537037037037037e-06, "loss": 1.0487, "step": 4940 }, { "epoch": 0.9702077616620933, "grad_norm": 5.2343363761901855, "learning_rate": 6.5277777777777784e-06, "loss": 1.0523, "step": 4950 }, { "epoch": 0.9721677773422187, "grad_norm": 4.845068454742432, "learning_rate": 6.51851851851852e-06, "loss": 1.0519, "step": 4960 }, { "epoch": 0.9741277930223442, "grad_norm": 5.101285457611084, "learning_rate": 6.509259259259259e-06, "loss": 1.0618, "step": 4970 }, { "epoch": 0.9760878087024696, "grad_norm": 4.684749126434326, "learning_rate": 6.5000000000000004e-06, "loss": 1.0506, "step": 4980 }, { "epoch": 0.9780478243825951, "grad_norm": 4.986086368560791, "learning_rate": 6.490740740740741e-06, "loss": 1.0621, "step": 4990 }, { "epoch": 0.9800078400627205, "grad_norm": 4.658474922180176, "learning_rate": 6.481481481481482e-06, "loss": 1.0727, "step": 5000 }, { "epoch": 0.981967855742846, "grad_norm": 5.1348419189453125, "learning_rate": 6.472222222222223e-06, "loss": 1.0542, "step": 5010 }, { "epoch": 0.9839278714229713, "grad_norm": 5.368491172790527, "learning_rate": 6.462962962962963e-06, "loss": 1.0656, "step": 5020 }, { "epoch": 0.9858878871030968, "grad_norm": 5.312760829925537, "learning_rate": 6.453703703703704e-06, "loss": 1.0587, "step": 5030 }, { "epoch": 0.9878479027832223, "grad_norm": 5.235229969024658, "learning_rate": 6.444444444444445e-06, "loss": 1.0588, "step": 5040 }, { "epoch": 0.9898079184633477, "grad_norm": 4.761176586151123, "learning_rate": 6.435185185185186e-06, "loss": 1.0507, "step": 5050 }, { "epoch": 0.9917679341434732, "grad_norm": 5.440311431884766, "learning_rate": 6.425925925925927e-06, "loss": 1.0566, "step": 5060 }, { "epoch": 0.9937279498235986, "grad_norm": 4.793478488922119, "learning_rate": 6.416666666666667e-06, "loss": 1.0607, "step": 5070 }, { "epoch": 0.9956879655037241, "grad_norm": 5.079543113708496, "learning_rate": 6.407407407407408e-06, "loss": 1.0494, "step": 5080 }, { "epoch": 0.9976479811838495, "grad_norm": 4.377906322479248, "learning_rate": 6.398148148148149e-06, "loss": 1.0558, "step": 5090 }, { "epoch": 0.999607996863975, "grad_norm": 4.611093521118164, "learning_rate": 6.3888888888888885e-06, "loss": 1.0693, "step": 5100 }, { "epoch": 1.0015680125441004, "grad_norm": 4.401882171630859, "learning_rate": 6.37962962962963e-06, "loss": 1.0297, "step": 5110 }, { "epoch": 1.0035280282242258, "grad_norm": 4.928489685058594, "learning_rate": 6.370370370370371e-06, "loss": 1.0276, "step": 5120 }, { "epoch": 1.0054880439043512, "grad_norm": 5.5292487144470215, "learning_rate": 6.361111111111111e-06, "loss": 1.0267, "step": 5130 }, { "epoch": 1.0074480595844766, "grad_norm": 5.00799036026001, "learning_rate": 6.351851851851853e-06, "loss": 1.0173, "step": 5140 }, { "epoch": 1.0094080752646022, "grad_norm": 4.399227619171143, "learning_rate": 6.342592592592594e-06, "loss": 1.0106, "step": 5150 }, { "epoch": 1.0113680909447276, "grad_norm": 5.132420539855957, "learning_rate": 6.333333333333333e-06, "loss": 1.0212, "step": 5160 }, { "epoch": 1.013328106624853, "grad_norm": 5.298924446105957, "learning_rate": 6.324074074074075e-06, "loss": 1.0122, "step": 5170 }, { "epoch": 1.0152881223049783, "grad_norm": 5.066485404968262, "learning_rate": 6.314814814814816e-06, "loss": 1.0114, "step": 5180 }, { "epoch": 1.017248137985104, "grad_norm": 6.66790771484375, "learning_rate": 6.305555555555556e-06, "loss": 1.0128, "step": 5190 }, { "epoch": 1.0192081536652293, "grad_norm": 4.905239105224609, "learning_rate": 6.296296296296297e-06, "loss": 1.0118, "step": 5200 }, { "epoch": 1.0211681693453547, "grad_norm": 5.377009868621826, "learning_rate": 6.287037037037037e-06, "loss": 1.0229, "step": 5210 }, { "epoch": 1.0231281850254803, "grad_norm": 5.1396942138671875, "learning_rate": 6.277777777777778e-06, "loss": 1.0221, "step": 5220 }, { "epoch": 1.0250882007056057, "grad_norm": 5.001946449279785, "learning_rate": 6.2685185185185195e-06, "loss": 1.0461, "step": 5230 }, { "epoch": 1.027048216385731, "grad_norm": 6.100106716156006, "learning_rate": 6.259259259259259e-06, "loss": 1.0169, "step": 5240 }, { "epoch": 1.0290082320658565, "grad_norm": 5.1383209228515625, "learning_rate": 6.25e-06, "loss": 1.0282, "step": 5250 }, { "epoch": 1.030968247745982, "grad_norm": 5.477851390838623, "learning_rate": 6.2407407407407415e-06, "loss": 1.0266, "step": 5260 }, { "epoch": 1.0329282634261074, "grad_norm": 5.047489643096924, "learning_rate": 6.231481481481482e-06, "loss": 1.0256, "step": 5270 }, { "epoch": 1.0348882791062328, "grad_norm": 4.994131565093994, "learning_rate": 6.222222222222223e-06, "loss": 1.0224, "step": 5280 }, { "epoch": 1.0368482947863582, "grad_norm": 5.558932781219482, "learning_rate": 6.2129629629629636e-06, "loss": 1.0186, "step": 5290 }, { "epoch": 1.0388083104664838, "grad_norm": 5.365056037902832, "learning_rate": 6.203703703703704e-06, "loss": 1.0095, "step": 5300 }, { "epoch": 1.0407683261466092, "grad_norm": 5.53687858581543, "learning_rate": 6.194444444444445e-06, "loss": 1.0132, "step": 5310 }, { "epoch": 1.0427283418267346, "grad_norm": 5.936549663543701, "learning_rate": 6.1851851851851856e-06, "loss": 1.0347, "step": 5320 }, { "epoch": 1.04468835750686, "grad_norm": 5.228058338165283, "learning_rate": 6.175925925925926e-06, "loss": 1.0234, "step": 5330 }, { "epoch": 1.0466483731869856, "grad_norm": 4.896886348724365, "learning_rate": 6.166666666666667e-06, "loss": 1.0065, "step": 5340 }, { "epoch": 1.048608388867111, "grad_norm": 6.313221454620361, "learning_rate": 6.157407407407408e-06, "loss": 1.0191, "step": 5350 }, { "epoch": 1.0505684045472363, "grad_norm": 5.8848114013671875, "learning_rate": 6.148148148148149e-06, "loss": 1.0378, "step": 5360 }, { "epoch": 1.052528420227362, "grad_norm": 5.468878746032715, "learning_rate": 6.13888888888889e-06, "loss": 1.0056, "step": 5370 }, { "epoch": 1.0544884359074873, "grad_norm": 5.208171367645264, "learning_rate": 6.12962962962963e-06, "loss": 1.0194, "step": 5380 }, { "epoch": 1.0564484515876127, "grad_norm": 5.1294450759887695, "learning_rate": 6.120370370370371e-06, "loss": 1.0084, "step": 5390 }, { "epoch": 1.058408467267738, "grad_norm": 5.981677532196045, "learning_rate": 6.111111111111112e-06, "loss": 1.0148, "step": 5400 }, { "epoch": 1.0603684829478637, "grad_norm": 5.662644863128662, "learning_rate": 6.1018518518518525e-06, "loss": 1.0163, "step": 5410 }, { "epoch": 1.062328498627989, "grad_norm": 5.452698230743408, "learning_rate": 6.092592592592593e-06, "loss": 1.0047, "step": 5420 }, { "epoch": 1.0642885143081144, "grad_norm": 4.969189167022705, "learning_rate": 6.083333333333333e-06, "loss": 1.0031, "step": 5430 }, { "epoch": 1.0662485299882398, "grad_norm": 5.396177768707275, "learning_rate": 6.0740740740740745e-06, "loss": 1.0109, "step": 5440 }, { "epoch": 1.0682085456683654, "grad_norm": 6.180945873260498, "learning_rate": 6.064814814814816e-06, "loss": 1.0318, "step": 5450 }, { "epoch": 1.0701685613484908, "grad_norm": 4.878726959228516, "learning_rate": 6.055555555555555e-06, "loss": 1.0257, "step": 5460 }, { "epoch": 1.0721285770286162, "grad_norm": 5.634671688079834, "learning_rate": 6.0462962962962965e-06, "loss": 1.0241, "step": 5470 }, { "epoch": 1.0740885927087416, "grad_norm": 5.971144199371338, "learning_rate": 6.037037037037038e-06, "loss": 1.0247, "step": 5480 }, { "epoch": 1.0760486083888672, "grad_norm": 5.1925950050354, "learning_rate": 6.027777777777778e-06, "loss": 1.0094, "step": 5490 }, { "epoch": 1.0780086240689926, "grad_norm": 4.921753406524658, "learning_rate": 6.018518518518519e-06, "loss": 1.0233, "step": 5500 }, { "epoch": 1.079968639749118, "grad_norm": 5.969301700592041, "learning_rate": 6.009259259259261e-06, "loss": 0.994, "step": 5510 }, { "epoch": 1.0819286554292433, "grad_norm": 5.875487804412842, "learning_rate": 6e-06, "loss": 1.0171, "step": 5520 }, { "epoch": 1.083888671109369, "grad_norm": 5.540329933166504, "learning_rate": 5.990740740740741e-06, "loss": 1.0198, "step": 5530 }, { "epoch": 1.0858486867894943, "grad_norm": 5.154452323913574, "learning_rate": 5.981481481481482e-06, "loss": 1.0174, "step": 5540 }, { "epoch": 1.0878087024696197, "grad_norm": 5.469849586486816, "learning_rate": 5.972222222222222e-06, "loss": 1.0141, "step": 5550 }, { "epoch": 1.0897687181497453, "grad_norm": 6.041581153869629, "learning_rate": 5.962962962962963e-06, "loss": 1.0216, "step": 5560 }, { "epoch": 1.0917287338298707, "grad_norm": 5.2390522956848145, "learning_rate": 5.953703703703704e-06, "loss": 1.0202, "step": 5570 }, { "epoch": 1.093688749509996, "grad_norm": 5.63343620300293, "learning_rate": 5.944444444444445e-06, "loss": 1.0234, "step": 5580 }, { "epoch": 1.0956487651901214, "grad_norm": 14.769401550292969, "learning_rate": 5.935185185185186e-06, "loss": 1.019, "step": 5590 }, { "epoch": 1.097608780870247, "grad_norm": 5.917840957641602, "learning_rate": 5.925925925925926e-06, "loss": 1.0259, "step": 5600 }, { "epoch": 1.097608780870247, "eval_loss": 1.0469825267791748, "eval_runtime": 14.056, "eval_samples_per_second": 46.528, "eval_steps_per_second": 5.834, "step": 5600 }, { "epoch": 1.0995687965503724, "grad_norm": 4.929704666137695, "learning_rate": 5.916666666666667e-06, "loss": 1.0146, "step": 5610 }, { "epoch": 1.1015288122304978, "grad_norm": 6.432901382446289, "learning_rate": 5.907407407407408e-06, "loss": 1.0268, "step": 5620 }, { "epoch": 1.1034888279106232, "grad_norm": 5.020516872406006, "learning_rate": 5.898148148148149e-06, "loss": 1.0272, "step": 5630 }, { "epoch": 1.1054488435907488, "grad_norm": 5.231268882751465, "learning_rate": 5.88888888888889e-06, "loss": 1.0046, "step": 5640 }, { "epoch": 1.1074088592708742, "grad_norm": 5.203913688659668, "learning_rate": 5.8796296296296295e-06, "loss": 1.015, "step": 5650 }, { "epoch": 1.1093688749509996, "grad_norm": 5.593926906585693, "learning_rate": 5.870370370370371e-06, "loss": 1.0185, "step": 5660 }, { "epoch": 1.1113288906311252, "grad_norm": 5.165170669555664, "learning_rate": 5.861111111111112e-06, "loss": 1.0056, "step": 5670 }, { "epoch": 1.1132889063112505, "grad_norm": 5.521978855133057, "learning_rate": 5.8518518518518515e-06, "loss": 1.022, "step": 5680 }, { "epoch": 1.115248921991376, "grad_norm": 5.280319690704346, "learning_rate": 5.842592592592593e-06, "loss": 1.0357, "step": 5690 }, { "epoch": 1.1172089376715013, "grad_norm": 6.761595726013184, "learning_rate": 5.833333333333334e-06, "loss": 1.0229, "step": 5700 }, { "epoch": 1.119168953351627, "grad_norm": 5.606836318969727, "learning_rate": 5.824074074074074e-06, "loss": 1.019, "step": 5710 }, { "epoch": 1.1211289690317523, "grad_norm": 5.351015567779541, "learning_rate": 5.814814814814816e-06, "loss": 1.0343, "step": 5720 }, { "epoch": 1.1230889847118777, "grad_norm": 4.853587627410889, "learning_rate": 5.805555555555557e-06, "loss": 1.0239, "step": 5730 }, { "epoch": 1.125049000392003, "grad_norm": 16.355466842651367, "learning_rate": 5.796296296296296e-06, "loss": 1.0013, "step": 5740 }, { "epoch": 1.1270090160721287, "grad_norm": 5.404659748077393, "learning_rate": 5.787037037037038e-06, "loss": 1.0181, "step": 5750 }, { "epoch": 1.128969031752254, "grad_norm": 5.103781700134277, "learning_rate": 5.777777777777778e-06, "loss": 1.0213, "step": 5760 }, { "epoch": 1.1309290474323794, "grad_norm": 5.420756816864014, "learning_rate": 5.768518518518519e-06, "loss": 1.0072, "step": 5770 }, { "epoch": 1.1328890631125048, "grad_norm": 5.301304817199707, "learning_rate": 5.75925925925926e-06, "loss": 0.9901, "step": 5780 }, { "epoch": 1.1348490787926304, "grad_norm": 4.864381790161133, "learning_rate": 5.75e-06, "loss": 1.0129, "step": 5790 }, { "epoch": 1.1368090944727558, "grad_norm": 6.017991542816162, "learning_rate": 5.740740740740741e-06, "loss": 1.0113, "step": 5800 }, { "epoch": 1.1387691101528812, "grad_norm": 5.6904520988464355, "learning_rate": 5.7314814814814825e-06, "loss": 1.0211, "step": 5810 }, { "epoch": 1.1407291258330066, "grad_norm": 5.778028964996338, "learning_rate": 5.722222222222222e-06, "loss": 1.0247, "step": 5820 }, { "epoch": 1.1426891415131322, "grad_norm": 5.682682991027832, "learning_rate": 5.712962962962963e-06, "loss": 1.0309, "step": 5830 }, { "epoch": 1.1446491571932575, "grad_norm": 5.886664867401123, "learning_rate": 5.7037037037037045e-06, "loss": 1.0244, "step": 5840 }, { "epoch": 1.146609172873383, "grad_norm": 5.014996528625488, "learning_rate": 5.694444444444445e-06, "loss": 1.0207, "step": 5850 }, { "epoch": 1.1485691885535085, "grad_norm": 5.563379287719727, "learning_rate": 5.685185185185186e-06, "loss": 1.0276, "step": 5860 }, { "epoch": 1.150529204233634, "grad_norm": 5.241649627685547, "learning_rate": 5.675925925925926e-06, "loss": 1.022, "step": 5870 }, { "epoch": 1.1524892199137593, "grad_norm": 5.154331207275391, "learning_rate": 5.666666666666667e-06, "loss": 1.0232, "step": 5880 }, { "epoch": 1.1544492355938847, "grad_norm": 5.823696136474609, "learning_rate": 5.657407407407408e-06, "loss": 1.0364, "step": 5890 }, { "epoch": 1.1564092512740103, "grad_norm": 5.451704978942871, "learning_rate": 5.6481481481481485e-06, "loss": 1.02, "step": 5900 }, { "epoch": 1.1583692669541357, "grad_norm": 5.540503978729248, "learning_rate": 5.638888888888889e-06, "loss": 1.0127, "step": 5910 }, { "epoch": 1.160329282634261, "grad_norm": 5.463021278381348, "learning_rate": 5.62962962962963e-06, "loss": 1.0205, "step": 5920 }, { "epoch": 1.1622892983143864, "grad_norm": 5.372107028961182, "learning_rate": 5.6203703703703705e-06, "loss": 1.0085, "step": 5930 }, { "epoch": 1.164249313994512, "grad_norm": 5.14231538772583, "learning_rate": 5.611111111111112e-06, "loss": 1.0189, "step": 5940 }, { "epoch": 1.1662093296746374, "grad_norm": 5.416358470916748, "learning_rate": 5.601851851851853e-06, "loss": 1.0066, "step": 5950 }, { "epoch": 1.1681693453547628, "grad_norm": 5.824409484863281, "learning_rate": 5.5925925925925926e-06, "loss": 1.0294, "step": 5960 }, { "epoch": 1.1701293610348882, "grad_norm": 5.362888336181641, "learning_rate": 5.583333333333334e-06, "loss": 1.011, "step": 5970 }, { "epoch": 1.1720893767150138, "grad_norm": 5.583401679992676, "learning_rate": 5.574074074074075e-06, "loss": 1.0218, "step": 5980 }, { "epoch": 1.1740493923951392, "grad_norm": 5.501253128051758, "learning_rate": 5.5648148148148154e-06, "loss": 1.0198, "step": 5990 }, { "epoch": 1.1760094080752646, "grad_norm": 5.536640167236328, "learning_rate": 5.555555555555557e-06, "loss": 1.0243, "step": 6000 }, { "epoch": 1.17796942375539, "grad_norm": 5.525614261627197, "learning_rate": 5.546296296296296e-06, "loss": 1.0234, "step": 6010 }, { "epoch": 1.1799294394355155, "grad_norm": 5.270368576049805, "learning_rate": 5.5370370370370374e-06, "loss": 1.0214, "step": 6020 }, { "epoch": 1.181889455115641, "grad_norm": 5.895664691925049, "learning_rate": 5.527777777777779e-06, "loss": 1.0203, "step": 6030 }, { "epoch": 1.1838494707957663, "grad_norm": 5.530089378356934, "learning_rate": 5.518518518518518e-06, "loss": 1.0254, "step": 6040 }, { "epoch": 1.185809486475892, "grad_norm": 5.275720119476318, "learning_rate": 5.5092592592592595e-06, "loss": 1.015, "step": 6050 }, { "epoch": 1.1877695021560173, "grad_norm": 5.464200496673584, "learning_rate": 5.500000000000001e-06, "loss": 1.0134, "step": 6060 }, { "epoch": 1.1897295178361427, "grad_norm": 5.803837299346924, "learning_rate": 5.490740740740741e-06, "loss": 1.0125, "step": 6070 }, { "epoch": 1.191689533516268, "grad_norm": 5.1351752281188965, "learning_rate": 5.481481481481482e-06, "loss": 1.0061, "step": 6080 }, { "epoch": 1.1936495491963937, "grad_norm": 5.232799530029297, "learning_rate": 5.4722222222222236e-06, "loss": 1.0213, "step": 6090 }, { "epoch": 1.195609564876519, "grad_norm": 5.996954917907715, "learning_rate": 5.462962962962963e-06, "loss": 1.0177, "step": 6100 }, { "epoch": 1.1975695805566444, "grad_norm": 7.579484939575195, "learning_rate": 5.453703703703704e-06, "loss": 1.0133, "step": 6110 }, { "epoch": 1.1995295962367698, "grad_norm": 5.351324081420898, "learning_rate": 5.444444444444445e-06, "loss": 1.0019, "step": 6120 }, { "epoch": 1.2014896119168954, "grad_norm": 5.5653581619262695, "learning_rate": 5.435185185185186e-06, "loss": 1.0069, "step": 6130 }, { "epoch": 1.2034496275970208, "grad_norm": 5.198008060455322, "learning_rate": 5.425925925925926e-06, "loss": 1.0157, "step": 6140 }, { "epoch": 1.2054096432771462, "grad_norm": 5.67602014541626, "learning_rate": 5.416666666666667e-06, "loss": 1.0327, "step": 6150 }, { "epoch": 1.2073696589572718, "grad_norm": 5.185479164123535, "learning_rate": 5.407407407407408e-06, "loss": 1.0118, "step": 6160 }, { "epoch": 1.2093296746373972, "grad_norm": 5.254889965057373, "learning_rate": 5.398148148148149e-06, "loss": 1.0282, "step": 6170 }, { "epoch": 1.2112896903175225, "grad_norm": 5.635478496551514, "learning_rate": 5.388888888888889e-06, "loss": 1.0211, "step": 6180 }, { "epoch": 1.213249705997648, "grad_norm": 5.388778209686279, "learning_rate": 5.37962962962963e-06, "loss": 1.0171, "step": 6190 }, { "epoch": 1.2152097216777733, "grad_norm": 5.546689987182617, "learning_rate": 5.370370370370371e-06, "loss": 1.0113, "step": 6200 }, { "epoch": 1.217169737357899, "grad_norm": 5.316011428833008, "learning_rate": 5.361111111111112e-06, "loss": 1.0183, "step": 6210 }, { "epoch": 1.2191297530380243, "grad_norm": 6.057391166687012, "learning_rate": 5.351851851851853e-06, "loss": 1.0149, "step": 6220 }, { "epoch": 1.2210897687181497, "grad_norm": 5.451624393463135, "learning_rate": 5.342592592592592e-06, "loss": 1.0129, "step": 6230 }, { "epoch": 1.2230497843982753, "grad_norm": 5.58920955657959, "learning_rate": 5.333333333333334e-06, "loss": 1.0152, "step": 6240 }, { "epoch": 1.2250098000784007, "grad_norm": 5.132778167724609, "learning_rate": 5.324074074074075e-06, "loss": 1.0059, "step": 6250 }, { "epoch": 1.226969815758526, "grad_norm": 4.807291030883789, "learning_rate": 5.314814814814815e-06, "loss": 1.0146, "step": 6260 }, { "epoch": 1.2289298314386514, "grad_norm": 5.3906474113464355, "learning_rate": 5.305555555555556e-06, "loss": 1.0045, "step": 6270 }, { "epoch": 1.230889847118777, "grad_norm": 5.229481220245361, "learning_rate": 5.296296296296297e-06, "loss": 1.0018, "step": 6280 }, { "epoch": 1.2328498627989024, "grad_norm": 5.5774126052856445, "learning_rate": 5.287037037037037e-06, "loss": 1.0096, "step": 6290 }, { "epoch": 1.2348098784790278, "grad_norm": 5.752622127532959, "learning_rate": 5.2777777777777785e-06, "loss": 1.0173, "step": 6300 }, { "epoch": 1.2348098784790278, "eval_loss": 1.0454390048980713, "eval_runtime": 14.0845, "eval_samples_per_second": 46.434, "eval_steps_per_second": 5.822, "step": 6300 }, { "epoch": 1.2367698941591532, "grad_norm": 4.995856285095215, "learning_rate": 5.26851851851852e-06, "loss": 1.0274, "step": 6310 }, { "epoch": 1.2387299098392788, "grad_norm": 5.631256103515625, "learning_rate": 5.259259259259259e-06, "loss": 1.029, "step": 6320 }, { "epoch": 1.2406899255194042, "grad_norm": 5.617944717407227, "learning_rate": 5.2500000000000006e-06, "loss": 1.0087, "step": 6330 }, { "epoch": 1.2426499411995295, "grad_norm": 5.174662113189697, "learning_rate": 5.240740740740741e-06, "loss": 1.0114, "step": 6340 }, { "epoch": 1.2446099568796551, "grad_norm": 5.841761112213135, "learning_rate": 5.231481481481482e-06, "loss": 1.0113, "step": 6350 }, { "epoch": 1.2465699725597805, "grad_norm": 5.008285999298096, "learning_rate": 5.2222222222222226e-06, "loss": 1.0058, "step": 6360 }, { "epoch": 1.248529988239906, "grad_norm": 5.529172420501709, "learning_rate": 5.212962962962963e-06, "loss": 1.0225, "step": 6370 }, { "epoch": 1.2504900039200313, "grad_norm": 6.809742450714111, "learning_rate": 5.203703703703704e-06, "loss": 1.0194, "step": 6380 }, { "epoch": 1.2524500196001567, "grad_norm": 5.778404712677002, "learning_rate": 5.1944444444444454e-06, "loss": 1.0312, "step": 6390 }, { "epoch": 1.2544100352802823, "grad_norm": 6.095964431762695, "learning_rate": 5.185185185185185e-06, "loss": 1.0155, "step": 6400 }, { "epoch": 1.2563700509604077, "grad_norm": 5.453831672668457, "learning_rate": 5.175925925925926e-06, "loss": 1.0155, "step": 6410 }, { "epoch": 1.258330066640533, "grad_norm": 5.425611972808838, "learning_rate": 5.1666666666666675e-06, "loss": 1.0353, "step": 6420 }, { "epoch": 1.2602900823206586, "grad_norm": 6.835920810699463, "learning_rate": 5.157407407407408e-06, "loss": 1.0089, "step": 6430 }, { "epoch": 1.262250098000784, "grad_norm": 8.35204029083252, "learning_rate": 5.148148148148149e-06, "loss": 1.0235, "step": 6440 }, { "epoch": 1.2642101136809094, "grad_norm": 5.1433234214782715, "learning_rate": 5.138888888888889e-06, "loss": 1.0226, "step": 6450 }, { "epoch": 1.266170129361035, "grad_norm": 5.907522201538086, "learning_rate": 5.12962962962963e-06, "loss": 1.0161, "step": 6460 }, { "epoch": 1.2681301450411604, "grad_norm": 5.631725788116455, "learning_rate": 5.120370370370371e-06, "loss": 1.014, "step": 6470 }, { "epoch": 1.2700901607212858, "grad_norm": 5.635436058044434, "learning_rate": 5.1111111111111115e-06, "loss": 1.0216, "step": 6480 }, { "epoch": 1.2720501764014112, "grad_norm": 6.376223087310791, "learning_rate": 5.101851851851852e-06, "loss": 1.0135, "step": 6490 }, { "epoch": 1.2740101920815365, "grad_norm": 6.083188533782959, "learning_rate": 5.092592592592593e-06, "loss": 1.0117, "step": 6500 }, { "epoch": 1.2759702077616621, "grad_norm": 5.436441898345947, "learning_rate": 5.0833333333333335e-06, "loss": 1.0079, "step": 6510 }, { "epoch": 1.2779302234417875, "grad_norm": 5.303137302398682, "learning_rate": 5.074074074074075e-06, "loss": 1.02, "step": 6520 }, { "epoch": 1.279890239121913, "grad_norm": 5.048837661743164, "learning_rate": 5.064814814814816e-06, "loss": 1.0045, "step": 6530 }, { "epoch": 1.2818502548020385, "grad_norm": 6.105396747589111, "learning_rate": 5.0555555555555555e-06, "loss": 1.0105, "step": 6540 }, { "epoch": 1.283810270482164, "grad_norm": 4.97545051574707, "learning_rate": 5.046296296296297e-06, "loss": 1.0128, "step": 6550 }, { "epoch": 1.2857702861622893, "grad_norm": 6.359137535095215, "learning_rate": 5.037037037037037e-06, "loss": 1.0127, "step": 6560 }, { "epoch": 1.2877303018424147, "grad_norm": 6.185931205749512, "learning_rate": 5.027777777777778e-06, "loss": 1.0156, "step": 6570 }, { "epoch": 1.28969031752254, "grad_norm": 6.333951950073242, "learning_rate": 5.01851851851852e-06, "loss": 1.0113, "step": 6580 }, { "epoch": 1.2916503332026656, "grad_norm": 5.14535665512085, "learning_rate": 5.009259259259259e-06, "loss": 1.0062, "step": 6590 }, { "epoch": 1.293610348882791, "grad_norm": 6.086668491363525, "learning_rate": 5e-06, "loss": 1.0123, "step": 6600 }, { "epoch": 1.2955703645629164, "grad_norm": 5.336708068847656, "learning_rate": 4.990740740740741e-06, "loss": 1.0058, "step": 6610 }, { "epoch": 1.297530380243042, "grad_norm": 5.65183687210083, "learning_rate": 4.981481481481482e-06, "loss": 1.0232, "step": 6620 }, { "epoch": 1.2994903959231674, "grad_norm": 5.356630802154541, "learning_rate": 4.9722222222222224e-06, "loss": 1.0248, "step": 6630 }, { "epoch": 1.3014504116032928, "grad_norm": 5.61469841003418, "learning_rate": 4.962962962962964e-06, "loss": 1.0049, "step": 6640 }, { "epoch": 1.3034104272834184, "grad_norm": 5.575668811798096, "learning_rate": 4.953703703703704e-06, "loss": 1.0046, "step": 6650 }, { "epoch": 1.3053704429635438, "grad_norm": 5.938807010650635, "learning_rate": 4.944444444444445e-06, "loss": 1.0254, "step": 6660 }, { "epoch": 1.3073304586436691, "grad_norm": 5.57589054107666, "learning_rate": 4.935185185185186e-06, "loss": 1.016, "step": 6670 }, { "epoch": 1.3092904743237945, "grad_norm": 5.684128284454346, "learning_rate": 4.925925925925926e-06, "loss": 1.0102, "step": 6680 }, { "epoch": 1.31125049000392, "grad_norm": 5.43093729019165, "learning_rate": 4.9166666666666665e-06, "loss": 1.0092, "step": 6690 }, { "epoch": 1.3132105056840455, "grad_norm": 5.595739841461182, "learning_rate": 4.907407407407408e-06, "loss": 1.0103, "step": 6700 }, { "epoch": 1.315170521364171, "grad_norm": 6.116888046264648, "learning_rate": 4.898148148148149e-06, "loss": 1.0177, "step": 6710 }, { "epoch": 1.3171305370442963, "grad_norm": 5.478735446929932, "learning_rate": 4.888888888888889e-06, "loss": 1.0147, "step": 6720 }, { "epoch": 1.3190905527244219, "grad_norm": 5.878492832183838, "learning_rate": 4.8796296296296306e-06, "loss": 1.0189, "step": 6730 }, { "epoch": 1.3210505684045473, "grad_norm": 5.890417098999023, "learning_rate": 4.870370370370371e-06, "loss": 1.0246, "step": 6740 }, { "epoch": 1.3230105840846726, "grad_norm": 5.2328410148620605, "learning_rate": 4.861111111111111e-06, "loss": 1.0116, "step": 6750 }, { "epoch": 1.3249705997647983, "grad_norm": 6.1324005126953125, "learning_rate": 4.851851851851852e-06, "loss": 1.0028, "step": 6760 }, { "epoch": 1.3269306154449236, "grad_norm": 5.392694473266602, "learning_rate": 4.842592592592593e-06, "loss": 1.0192, "step": 6770 }, { "epoch": 1.328890631125049, "grad_norm": 5.902957439422607, "learning_rate": 4.833333333333333e-06, "loss": 1.0145, "step": 6780 }, { "epoch": 1.3308506468051744, "grad_norm": 5.553819179534912, "learning_rate": 4.824074074074075e-06, "loss": 1.0091, "step": 6790 }, { "epoch": 1.3328106624852998, "grad_norm": 6.013294219970703, "learning_rate": 4.814814814814815e-06, "loss": 1.0198, "step": 6800 }, { "epoch": 1.3347706781654254, "grad_norm": 5.731135368347168, "learning_rate": 4.805555555555556e-06, "loss": 1.014, "step": 6810 }, { "epoch": 1.3367306938455508, "grad_norm": 5.954074382781982, "learning_rate": 4.796296296296297e-06, "loss": 1.0128, "step": 6820 }, { "epoch": 1.3386907095256761, "grad_norm": 5.550869464874268, "learning_rate": 4.787037037037037e-06, "loss": 1.0151, "step": 6830 }, { "epoch": 1.3406507252058018, "grad_norm": 6.043259620666504, "learning_rate": 4.777777777777778e-06, "loss": 1.0069, "step": 6840 }, { "epoch": 1.3426107408859271, "grad_norm": 5.002622127532959, "learning_rate": 4.768518518518519e-06, "loss": 1.0219, "step": 6850 }, { "epoch": 1.3445707565660525, "grad_norm": 6.016299247741699, "learning_rate": 4.75925925925926e-06, "loss": 1.0158, "step": 6860 }, { "epoch": 1.346530772246178, "grad_norm": 5.6830220222473145, "learning_rate": 4.75e-06, "loss": 1.0092, "step": 6870 }, { "epoch": 1.3484907879263033, "grad_norm": 6.092140197753906, "learning_rate": 4.7407407407407415e-06, "loss": 1.0241, "step": 6880 }, { "epoch": 1.3504508036064289, "grad_norm": 5.837907791137695, "learning_rate": 4.731481481481482e-06, "loss": 1.0259, "step": 6890 }, { "epoch": 1.3524108192865543, "grad_norm": 5.598001480102539, "learning_rate": 4.722222222222222e-06, "loss": 1.0043, "step": 6900 }, { "epoch": 1.3543708349666796, "grad_norm": 5.820865631103516, "learning_rate": 4.712962962962963e-06, "loss": 1.0325, "step": 6910 }, { "epoch": 1.3563308506468053, "grad_norm": 6.162250518798828, "learning_rate": 4.703703703703704e-06, "loss": 1.0118, "step": 6920 }, { "epoch": 1.3582908663269306, "grad_norm": 5.53606653213501, "learning_rate": 4.694444444444445e-06, "loss": 1.0099, "step": 6930 }, { "epoch": 1.360250882007056, "grad_norm": 6.468924045562744, "learning_rate": 4.6851851851851855e-06, "loss": 1.0172, "step": 6940 }, { "epoch": 1.3622108976871816, "grad_norm": 6.867157936096191, "learning_rate": 4.675925925925927e-06, "loss": 1.0021, "step": 6950 }, { "epoch": 1.364170913367307, "grad_norm": 6.068783760070801, "learning_rate": 4.666666666666667e-06, "loss": 1.0083, "step": 6960 }, { "epoch": 1.3661309290474324, "grad_norm": 5.502538204193115, "learning_rate": 4.6574074074074076e-06, "loss": 1.0173, "step": 6970 }, { "epoch": 1.3680909447275578, "grad_norm": 5.83467960357666, "learning_rate": 4.648148148148148e-06, "loss": 1.031, "step": 6980 }, { "epoch": 1.3700509604076831, "grad_norm": 5.383768558502197, "learning_rate": 4.638888888888889e-06, "loss": 1.0187, "step": 6990 }, { "epoch": 1.3720109760878088, "grad_norm": 5.172204494476318, "learning_rate": 4.62962962962963e-06, "loss": 1.0079, "step": 7000 }, { "epoch": 1.3720109760878088, "eval_loss": 1.0383639335632324, "eval_runtime": 14.0609, "eval_samples_per_second": 46.512, "eval_steps_per_second": 5.832, "step": 7000 }, { "epoch": 1.3739709917679341, "grad_norm": 5.811285495758057, "learning_rate": 4.620370370370371e-06, "loss": 1.0153, "step": 7010 }, { "epoch": 1.3759310074480595, "grad_norm": 5.887601375579834, "learning_rate": 4.611111111111112e-06, "loss": 1.0104, "step": 7020 }, { "epoch": 1.3778910231281851, "grad_norm": 5.245418548583984, "learning_rate": 4.6018518518518524e-06, "loss": 1.0063, "step": 7030 }, { "epoch": 1.3798510388083105, "grad_norm": 6.191716194152832, "learning_rate": 4.592592592592593e-06, "loss": 1.0254, "step": 7040 }, { "epoch": 1.3818110544884359, "grad_norm": 6.047815799713135, "learning_rate": 4.583333333333333e-06, "loss": 1.0291, "step": 7050 }, { "epoch": 1.3837710701685613, "grad_norm": 5.58579158782959, "learning_rate": 4.5740740740740745e-06, "loss": 1.0126, "step": 7060 }, { "epoch": 1.3857310858486869, "grad_norm": 5.651394844055176, "learning_rate": 4.564814814814815e-06, "loss": 1.013, "step": 7070 }, { "epoch": 1.3876911015288123, "grad_norm": 5.703254699707031, "learning_rate": 4.555555555555556e-06, "loss": 1.0209, "step": 7080 }, { "epoch": 1.3896511172089376, "grad_norm": 6.428845405578613, "learning_rate": 4.5462962962962965e-06, "loss": 1.0065, "step": 7090 }, { "epoch": 1.391611132889063, "grad_norm": 5.854006767272949, "learning_rate": 4.537037037037038e-06, "loss": 1.0192, "step": 7100 }, { "epoch": 1.3935711485691886, "grad_norm": 5.257528781890869, "learning_rate": 4.527777777777778e-06, "loss": 1.0125, "step": 7110 }, { "epoch": 1.395531164249314, "grad_norm": 5.832418441772461, "learning_rate": 4.5185185185185185e-06, "loss": 1.0077, "step": 7120 }, { "epoch": 1.3974911799294394, "grad_norm": 18.90093994140625, "learning_rate": 4.50925925925926e-06, "loss": 1.0039, "step": 7130 }, { "epoch": 1.399451195609565, "grad_norm": 6.230194568634033, "learning_rate": 4.5e-06, "loss": 1.0168, "step": 7140 }, { "epoch": 1.4014112112896904, "grad_norm": 5.916494846343994, "learning_rate": 4.490740740740741e-06, "loss": 1.0367, "step": 7150 }, { "epoch": 1.4033712269698158, "grad_norm": 6.498172760009766, "learning_rate": 4.481481481481482e-06, "loss": 1.0226, "step": 7160 }, { "epoch": 1.4053312426499411, "grad_norm": 5.681482791900635, "learning_rate": 4.472222222222223e-06, "loss": 1.0135, "step": 7170 }, { "epoch": 1.4072912583300665, "grad_norm": 5.3359880447387695, "learning_rate": 4.462962962962963e-06, "loss": 1.0222, "step": 7180 }, { "epoch": 1.4092512740101921, "grad_norm": 6.866293430328369, "learning_rate": 4.453703703703704e-06, "loss": 1.015, "step": 7190 }, { "epoch": 1.4112112896903175, "grad_norm": 6.149731159210205, "learning_rate": 4.444444444444444e-06, "loss": 0.998, "step": 7200 }, { "epoch": 1.4131713053704429, "grad_norm": 5.832232475280762, "learning_rate": 4.435185185185185e-06, "loss": 1.0097, "step": 7210 }, { "epoch": 1.4151313210505685, "grad_norm": 5.8602986335754395, "learning_rate": 4.425925925925927e-06, "loss": 1.0133, "step": 7220 }, { "epoch": 1.4170913367306939, "grad_norm": 6.716607093811035, "learning_rate": 4.416666666666667e-06, "loss": 1.0181, "step": 7230 }, { "epoch": 1.4190513524108193, "grad_norm": 5.983255863189697, "learning_rate": 4.407407407407408e-06, "loss": 1.0152, "step": 7240 }, { "epoch": 1.4210113680909449, "grad_norm": 5.979559898376465, "learning_rate": 4.398148148148149e-06, "loss": 1.0187, "step": 7250 }, { "epoch": 1.4229713837710702, "grad_norm": 5.463419437408447, "learning_rate": 4.388888888888889e-06, "loss": 1.0018, "step": 7260 }, { "epoch": 1.4249313994511956, "grad_norm": 5.58013916015625, "learning_rate": 4.379629629629629e-06, "loss": 1.0161, "step": 7270 }, { "epoch": 1.426891415131321, "grad_norm": 5.896617412567139, "learning_rate": 4.370370370370371e-06, "loss": 1.0234, "step": 7280 }, { "epoch": 1.4288514308114464, "grad_norm": 6.59088659286499, "learning_rate": 4.361111111111112e-06, "loss": 1.0116, "step": 7290 }, { "epoch": 1.430811446491572, "grad_norm": 5.4909348487854, "learning_rate": 4.351851851851852e-06, "loss": 1.0241, "step": 7300 }, { "epoch": 1.4327714621716974, "grad_norm": 5.706267356872559, "learning_rate": 4.342592592592593e-06, "loss": 1.0113, "step": 7310 }, { "epoch": 1.4347314778518228, "grad_norm": 5.893515586853027, "learning_rate": 4.333333333333334e-06, "loss": 1.0074, "step": 7320 }, { "epoch": 1.4366914935319484, "grad_norm": 5.337756156921387, "learning_rate": 4.324074074074074e-06, "loss": 1.0086, "step": 7330 }, { "epoch": 1.4386515092120737, "grad_norm": 6.063902854919434, "learning_rate": 4.314814814814815e-06, "loss": 1.0135, "step": 7340 }, { "epoch": 1.4406115248921991, "grad_norm": 5.862417221069336, "learning_rate": 4.305555555555556e-06, "loss": 1.0218, "step": 7350 }, { "epoch": 1.4425715405723245, "grad_norm": 6.047614574432373, "learning_rate": 4.296296296296296e-06, "loss": 1.0147, "step": 7360 }, { "epoch": 1.4445315562524499, "grad_norm": 7.018215656280518, "learning_rate": 4.2870370370370376e-06, "loss": 1.029, "step": 7370 }, { "epoch": 1.4464915719325755, "grad_norm": 5.391064167022705, "learning_rate": 4.277777777777778e-06, "loss": 1.007, "step": 7380 }, { "epoch": 1.4484515876127009, "grad_norm": 5.364874362945557, "learning_rate": 4.268518518518519e-06, "loss": 1.0188, "step": 7390 }, { "epoch": 1.4504116032928263, "grad_norm": 5.990734577178955, "learning_rate": 4.2592592592592596e-06, "loss": 1.0003, "step": 7400 }, { "epoch": 1.4523716189729519, "grad_norm": 5.727319717407227, "learning_rate": 4.25e-06, "loss": 1.0198, "step": 7410 }, { "epoch": 1.4543316346530772, "grad_norm": 5.658199310302734, "learning_rate": 4.240740740740741e-06, "loss": 1.0221, "step": 7420 }, { "epoch": 1.4562916503332026, "grad_norm": 5.460375785827637, "learning_rate": 4.231481481481482e-06, "loss": 1.0127, "step": 7430 }, { "epoch": 1.4582516660133282, "grad_norm": 6.161991596221924, "learning_rate": 4.222222222222223e-06, "loss": 1.0025, "step": 7440 }, { "epoch": 1.4602116816934536, "grad_norm": 5.7905192375183105, "learning_rate": 4.212962962962963e-06, "loss": 1.0071, "step": 7450 }, { "epoch": 1.462171697373579, "grad_norm": 5.652228832244873, "learning_rate": 4.2037037037037045e-06, "loss": 1.0112, "step": 7460 }, { "epoch": 1.4641317130537044, "grad_norm": 5.817644119262695, "learning_rate": 4.194444444444445e-06, "loss": 1.0191, "step": 7470 }, { "epoch": 1.4660917287338298, "grad_norm": 6.155873775482178, "learning_rate": 4.185185185185185e-06, "loss": 1.013, "step": 7480 }, { "epoch": 1.4680517444139554, "grad_norm": 5.775312423706055, "learning_rate": 4.175925925925926e-06, "loss": 1.0107, "step": 7490 }, { "epoch": 1.4700117600940807, "grad_norm": 5.765095233917236, "learning_rate": 4.166666666666667e-06, "loss": 1.0151, "step": 7500 }, { "epoch": 1.4719717757742061, "grad_norm": 5.336995601654053, "learning_rate": 4.157407407407408e-06, "loss": 1.0048, "step": 7510 }, { "epoch": 1.4739317914543317, "grad_norm": 6.261089324951172, "learning_rate": 4.1481481481481485e-06, "loss": 1.0002, "step": 7520 }, { "epoch": 1.4758918071344571, "grad_norm": 5.8046746253967285, "learning_rate": 4.138888888888889e-06, "loss": 1.0079, "step": 7530 }, { "epoch": 1.4778518228145825, "grad_norm": 5.5930867195129395, "learning_rate": 4.12962962962963e-06, "loss": 1.0094, "step": 7540 }, { "epoch": 1.4798118384947079, "grad_norm": 5.929795742034912, "learning_rate": 4.1203703703703705e-06, "loss": 1.0122, "step": 7550 }, { "epoch": 1.4817718541748335, "grad_norm": 5.746526718139648, "learning_rate": 4.111111111111111e-06, "loss": 1.0186, "step": 7560 }, { "epoch": 1.4837318698549589, "grad_norm": 6.316400051116943, "learning_rate": 4.101851851851852e-06, "loss": 1.0019, "step": 7570 }, { "epoch": 1.4856918855350842, "grad_norm": 5.561999320983887, "learning_rate": 4.092592592592593e-06, "loss": 1.0157, "step": 7580 }, { "epoch": 1.4876519012152096, "grad_norm": 6.012383937835693, "learning_rate": 4.083333333333334e-06, "loss": 1.0108, "step": 7590 }, { "epoch": 1.4896119168953352, "grad_norm": 5.594381809234619, "learning_rate": 4.074074074074074e-06, "loss": 0.9917, "step": 7600 }, { "epoch": 1.4915719325754606, "grad_norm": 6.158176898956299, "learning_rate": 4.064814814814815e-06, "loss": 1.0063, "step": 7610 }, { "epoch": 1.493531948255586, "grad_norm": 5.4881510734558105, "learning_rate": 4.055555555555556e-06, "loss": 1.0122, "step": 7620 }, { "epoch": 1.4954919639357116, "grad_norm": 6.7185869216918945, "learning_rate": 4.046296296296296e-06, "loss": 0.9927, "step": 7630 }, { "epoch": 1.497451979615837, "grad_norm": 6.015646934509277, "learning_rate": 4.037037037037037e-06, "loss": 1.0054, "step": 7640 }, { "epoch": 1.4994119952959624, "grad_norm": 5.855646133422852, "learning_rate": 4.027777777777779e-06, "loss": 0.9988, "step": 7650 }, { "epoch": 1.501372010976088, "grad_norm": 6.158780574798584, "learning_rate": 4.018518518518519e-06, "loss": 1.027, "step": 7660 }, { "epoch": 1.5033320266562131, "grad_norm": 5.943362712860107, "learning_rate": 4.0092592592592594e-06, "loss": 1.0285, "step": 7670 }, { "epoch": 1.5052920423363387, "grad_norm": 5.775484085083008, "learning_rate": 4.000000000000001e-06, "loss": 1.0093, "step": 7680 }, { "epoch": 1.5072520580164641, "grad_norm": 6.551724910736084, "learning_rate": 3.990740740740741e-06, "loss": 1.0185, "step": 7690 }, { "epoch": 1.5092120736965895, "grad_norm": 5.986613750457764, "learning_rate": 3.9814814814814814e-06, "loss": 1.0058, "step": 7700 }, { "epoch": 1.5092120736965895, "eval_loss": 1.0326706171035767, "eval_runtime": 13.9953, "eval_samples_per_second": 46.73, "eval_steps_per_second": 5.859, "step": 7700 }, { "epoch": 1.511172089376715, "grad_norm": 5.968577861785889, "learning_rate": 3.972222222222223e-06, "loss": 1.0111, "step": 7710 }, { "epoch": 1.5131321050568405, "grad_norm": 5.956306457519531, "learning_rate": 3.962962962962963e-06, "loss": 0.9822, "step": 7720 }, { "epoch": 1.5150921207369659, "grad_norm": 6.391535758972168, "learning_rate": 3.953703703703704e-06, "loss": 1.0073, "step": 7730 }, { "epoch": 1.5170521364170915, "grad_norm": 6.031282901763916, "learning_rate": 3.944444444444445e-06, "loss": 0.9992, "step": 7740 }, { "epoch": 1.5190121520972166, "grad_norm": 6.070298194885254, "learning_rate": 3.935185185185186e-06, "loss": 0.9995, "step": 7750 }, { "epoch": 1.5209721677773422, "grad_norm": 6.470989227294922, "learning_rate": 3.925925925925926e-06, "loss": 1.0041, "step": 7760 }, { "epoch": 1.5229321834574676, "grad_norm": 6.339069843292236, "learning_rate": 3.916666666666667e-06, "loss": 1.0145, "step": 7770 }, { "epoch": 1.524892199137593, "grad_norm": 6.204465866088867, "learning_rate": 3.907407407407408e-06, "loss": 1.0202, "step": 7780 }, { "epoch": 1.5268522148177186, "grad_norm": 5.39393424987793, "learning_rate": 3.898148148148148e-06, "loss": 1.0144, "step": 7790 }, { "epoch": 1.528812230497844, "grad_norm": 6.083856105804443, "learning_rate": 3.88888888888889e-06, "loss": 1.0087, "step": 7800 }, { "epoch": 1.5307722461779694, "grad_norm": 6.141456604003906, "learning_rate": 3.87962962962963e-06, "loss": 1.0189, "step": 7810 }, { "epoch": 1.532732261858095, "grad_norm": 6.509696960449219, "learning_rate": 3.87037037037037e-06, "loss": 1.0136, "step": 7820 }, { "epoch": 1.5346922775382204, "grad_norm": 5.372650146484375, "learning_rate": 3.861111111111112e-06, "loss": 1.0009, "step": 7830 }, { "epoch": 1.5366522932183457, "grad_norm": 5.931579113006592, "learning_rate": 3.851851851851852e-06, "loss": 1.0144, "step": 7840 }, { "epoch": 1.5386123088984713, "grad_norm": 6.2442626953125, "learning_rate": 3.842592592592592e-06, "loss": 1.0132, "step": 7850 }, { "epoch": 1.5405723245785965, "grad_norm": 6.475978851318359, "learning_rate": 3.833333333333334e-06, "loss": 1.0058, "step": 7860 }, { "epoch": 1.542532340258722, "grad_norm": 5.5199103355407715, "learning_rate": 3.824074074074075e-06, "loss": 1.0017, "step": 7870 }, { "epoch": 1.5444923559388475, "grad_norm": 5.857003211975098, "learning_rate": 3.814814814814815e-06, "loss": 1.0108, "step": 7880 }, { "epoch": 1.5464523716189729, "grad_norm": 5.81916618347168, "learning_rate": 3.8055555555555556e-06, "loss": 1.0185, "step": 7890 }, { "epoch": 1.5484123872990985, "grad_norm": 6.099959850311279, "learning_rate": 3.796296296296297e-06, "loss": 1.0109, "step": 7900 }, { "epoch": 1.5503724029792239, "grad_norm": 5.837829113006592, "learning_rate": 3.7870370370370373e-06, "loss": 1.0004, "step": 7910 }, { "epoch": 1.5523324186593492, "grad_norm": 6.274796485900879, "learning_rate": 3.777777777777778e-06, "loss": 1.0165, "step": 7920 }, { "epoch": 1.5542924343394748, "grad_norm": 5.926800727844238, "learning_rate": 3.7685185185185185e-06, "loss": 1.0067, "step": 7930 }, { "epoch": 1.5562524500196, "grad_norm": 6.2746381759643555, "learning_rate": 3.7592592592592597e-06, "loss": 1.0078, "step": 7940 }, { "epoch": 1.5582124656997256, "grad_norm": 6.315319538116455, "learning_rate": 3.7500000000000005e-06, "loss": 1.0085, "step": 7950 }, { "epoch": 1.560172481379851, "grad_norm": 6.36542272567749, "learning_rate": 3.740740740740741e-06, "loss": 1.0073, "step": 7960 }, { "epoch": 1.5621324970599764, "grad_norm": 6.047231197357178, "learning_rate": 3.731481481481482e-06, "loss": 1.0019, "step": 7970 }, { "epoch": 1.564092512740102, "grad_norm": 6.49619722366333, "learning_rate": 3.7222222222222225e-06, "loss": 1.0126, "step": 7980 }, { "epoch": 1.5660525284202274, "grad_norm": 6.945578098297119, "learning_rate": 3.7129629629629633e-06, "loss": 1.0141, "step": 7990 }, { "epoch": 1.5680125441003527, "grad_norm": 5.595407485961914, "learning_rate": 3.7037037037037037e-06, "loss": 1.0077, "step": 8000 }, { "epoch": 1.5699725597804783, "grad_norm": 6.3146867752075195, "learning_rate": 3.694444444444445e-06, "loss": 1.0178, "step": 8010 }, { "epoch": 1.5719325754606037, "grad_norm": 8.867284774780273, "learning_rate": 3.6851851851851854e-06, "loss": 0.9974, "step": 8020 }, { "epoch": 1.573892591140729, "grad_norm": 5.997375011444092, "learning_rate": 3.675925925925926e-06, "loss": 1.0169, "step": 8030 }, { "epoch": 1.5758526068208547, "grad_norm": 6.150747776031494, "learning_rate": 3.6666666666666666e-06, "loss": 1.0087, "step": 8040 }, { "epoch": 1.5778126225009799, "grad_norm": 6.344914436340332, "learning_rate": 3.657407407407408e-06, "loss": 1.008, "step": 8050 }, { "epoch": 1.5797726381811055, "grad_norm": 5.457338333129883, "learning_rate": 3.6481481481481486e-06, "loss": 1.0169, "step": 8060 }, { "epoch": 1.5817326538612309, "grad_norm": 6.128787994384766, "learning_rate": 3.638888888888889e-06, "loss": 1.0104, "step": 8070 }, { "epoch": 1.5836926695413562, "grad_norm": 6.624429702758789, "learning_rate": 3.6296296296296302e-06, "loss": 0.9982, "step": 8080 }, { "epoch": 1.5856526852214818, "grad_norm": 5.785494327545166, "learning_rate": 3.6203703703703706e-06, "loss": 0.9968, "step": 8090 }, { "epoch": 1.5876127009016072, "grad_norm": 6.378090858459473, "learning_rate": 3.6111111111111115e-06, "loss": 0.9972, "step": 8100 }, { "epoch": 1.5895727165817326, "grad_norm": 7.523343086242676, "learning_rate": 3.601851851851852e-06, "loss": 1.0077, "step": 8110 }, { "epoch": 1.5915327322618582, "grad_norm": 5.578686714172363, "learning_rate": 3.592592592592593e-06, "loss": 1.0103, "step": 8120 }, { "epoch": 1.5934927479419836, "grad_norm": 6.031825065612793, "learning_rate": 3.5833333333333335e-06, "loss": 1.015, "step": 8130 }, { "epoch": 1.595452763622109, "grad_norm": 6.009352207183838, "learning_rate": 3.5740740740740743e-06, "loss": 0.9997, "step": 8140 }, { "epoch": 1.5974127793022346, "grad_norm": 6.1863694190979, "learning_rate": 3.5648148148148147e-06, "loss": 0.9902, "step": 8150 }, { "epoch": 1.5993727949823597, "grad_norm": 5.477410316467285, "learning_rate": 3.555555555555556e-06, "loss": 1.0077, "step": 8160 }, { "epoch": 1.6013328106624853, "grad_norm": 6.446470260620117, "learning_rate": 3.5462962962962967e-06, "loss": 1.0123, "step": 8170 }, { "epoch": 1.6032928263426107, "grad_norm": 5.979601860046387, "learning_rate": 3.537037037037037e-06, "loss": 1.0106, "step": 8180 }, { "epoch": 1.605252842022736, "grad_norm": 5.748382091522217, "learning_rate": 3.5277777777777784e-06, "loss": 1.0083, "step": 8190 }, { "epoch": 1.6072128577028617, "grad_norm": 6.122110843658447, "learning_rate": 3.5185185185185187e-06, "loss": 1.0124, "step": 8200 }, { "epoch": 1.609172873382987, "grad_norm": 6.104898452758789, "learning_rate": 3.5092592592592596e-06, "loss": 1.0118, "step": 8210 }, { "epoch": 1.6111328890631125, "grad_norm": 6.527801513671875, "learning_rate": 3.5e-06, "loss": 1.0084, "step": 8220 }, { "epoch": 1.613092904743238, "grad_norm": 6.3016676902771, "learning_rate": 3.490740740740741e-06, "loss": 1.0104, "step": 8230 }, { "epoch": 1.6150529204233632, "grad_norm": 6.44483757019043, "learning_rate": 3.481481481481482e-06, "loss": 1.0102, "step": 8240 }, { "epoch": 1.6170129361034888, "grad_norm": 6.052326679229736, "learning_rate": 3.4722222222222224e-06, "loss": 1.0147, "step": 8250 }, { "epoch": 1.6189729517836142, "grad_norm": 5.94256067276001, "learning_rate": 3.4629629629629628e-06, "loss": 1.0108, "step": 8260 }, { "epoch": 1.6209329674637396, "grad_norm": 6.634081840515137, "learning_rate": 3.453703703703704e-06, "loss": 1.0139, "step": 8270 }, { "epoch": 1.6228929831438652, "grad_norm": 6.239657878875732, "learning_rate": 3.444444444444445e-06, "loss": 0.9949, "step": 8280 }, { "epoch": 1.6248529988239906, "grad_norm": 6.047982692718506, "learning_rate": 3.4351851851851852e-06, "loss": 1.0223, "step": 8290 }, { "epoch": 1.626813014504116, "grad_norm": 5.618074417114258, "learning_rate": 3.4259259259259265e-06, "loss": 1.022, "step": 8300 }, { "epoch": 1.6287730301842416, "grad_norm": 5.856754779815674, "learning_rate": 3.416666666666667e-06, "loss": 1.0053, "step": 8310 }, { "epoch": 1.630733045864367, "grad_norm": 6.435812473297119, "learning_rate": 3.4074074074074077e-06, "loss": 0.9989, "step": 8320 }, { "epoch": 1.6326930615444923, "grad_norm": 5.980440139770508, "learning_rate": 3.398148148148148e-06, "loss": 1.0108, "step": 8330 }, { "epoch": 1.634653077224618, "grad_norm": 6.19628381729126, "learning_rate": 3.3888888888888893e-06, "loss": 0.995, "step": 8340 }, { "epoch": 1.636613092904743, "grad_norm": 5.549736976623535, "learning_rate": 3.37962962962963e-06, "loss": 1.0181, "step": 8350 }, { "epoch": 1.6385731085848687, "grad_norm": 6.485236644744873, "learning_rate": 3.3703703703703705e-06, "loss": 1.0092, "step": 8360 }, { "epoch": 1.640533124264994, "grad_norm": 6.0221848487854, "learning_rate": 3.3611111111111117e-06, "loss": 1.0038, "step": 8370 }, { "epoch": 1.6424931399451195, "grad_norm": 5.699035167694092, "learning_rate": 3.351851851851852e-06, "loss": 1.0092, "step": 8380 }, { "epoch": 1.644453155625245, "grad_norm": 6.401322364807129, "learning_rate": 3.342592592592593e-06, "loss": 1.0013, "step": 8390 }, { "epoch": 1.6464131713053705, "grad_norm": 6.202515125274658, "learning_rate": 3.3333333333333333e-06, "loss": 1.0165, "step": 8400 }, { "epoch": 1.6464131713053705, "eval_loss": 1.0264891386032104, "eval_runtime": 14.05, "eval_samples_per_second": 46.548, "eval_steps_per_second": 5.836, "step": 8400 }, { "epoch": 1.6483731869854958, "grad_norm": 6.184670448303223, "learning_rate": 3.3240740740740746e-06, "loss": 1.0055, "step": 8410 }, { "epoch": 1.6503332026656214, "grad_norm": 6.040178298950195, "learning_rate": 3.314814814814815e-06, "loss": 1.007, "step": 8420 }, { "epoch": 1.6522932183457466, "grad_norm": 6.925061225891113, "learning_rate": 3.3055555555555558e-06, "loss": 1.0101, "step": 8430 }, { "epoch": 1.6542532340258722, "grad_norm": 6.050986289978027, "learning_rate": 3.296296296296296e-06, "loss": 0.9969, "step": 8440 }, { "epoch": 1.6562132497059978, "grad_norm": 5.321831703186035, "learning_rate": 3.2870370370370374e-06, "loss": 1.0092, "step": 8450 }, { "epoch": 1.658173265386123, "grad_norm": 5.440800666809082, "learning_rate": 3.277777777777778e-06, "loss": 1.0115, "step": 8460 }, { "epoch": 1.6601332810662486, "grad_norm": 8.311331748962402, "learning_rate": 3.2685185185185186e-06, "loss": 1.0014, "step": 8470 }, { "epoch": 1.662093296746374, "grad_norm": 6.014708518981934, "learning_rate": 3.25925925925926e-06, "loss": 1.0135, "step": 8480 }, { "epoch": 1.6640533124264993, "grad_norm": 6.3742475509643555, "learning_rate": 3.2500000000000002e-06, "loss": 1.0024, "step": 8490 }, { "epoch": 1.666013328106625, "grad_norm": 6.1811041831970215, "learning_rate": 3.240740740740741e-06, "loss": 0.9983, "step": 8500 }, { "epoch": 1.6679733437867503, "grad_norm": 6.055778503417969, "learning_rate": 3.2314814814814814e-06, "loss": 0.9986, "step": 8510 }, { "epoch": 1.6699333594668757, "grad_norm": 6.654355525970459, "learning_rate": 3.2222222222222227e-06, "loss": 1.0111, "step": 8520 }, { "epoch": 1.6718933751470013, "grad_norm": 6.156193733215332, "learning_rate": 3.2129629629629635e-06, "loss": 0.9879, "step": 8530 }, { "epoch": 1.6738533908271265, "grad_norm": 6.491064548492432, "learning_rate": 3.203703703703704e-06, "loss": 1.0197, "step": 8540 }, { "epoch": 1.675813406507252, "grad_norm": 5.876689434051514, "learning_rate": 3.1944444444444443e-06, "loss": 0.9934, "step": 8550 }, { "epoch": 1.6777734221873775, "grad_norm": 6.113574504852295, "learning_rate": 3.1851851851851855e-06, "loss": 0.9992, "step": 8560 }, { "epoch": 1.6797334378675028, "grad_norm": 6.2524518966674805, "learning_rate": 3.1759259259259263e-06, "loss": 1.0101, "step": 8570 }, { "epoch": 1.6816934535476284, "grad_norm": 6.034154891967773, "learning_rate": 3.1666666666666667e-06, "loss": 1.0009, "step": 8580 }, { "epoch": 1.6836534692277538, "grad_norm": 5.920796871185303, "learning_rate": 3.157407407407408e-06, "loss": 1.0088, "step": 8590 }, { "epoch": 1.6856134849078792, "grad_norm": 5.578927993774414, "learning_rate": 3.1481481481481483e-06, "loss": 1.0004, "step": 8600 }, { "epoch": 1.6875735005880048, "grad_norm": 6.1217803955078125, "learning_rate": 3.138888888888889e-06, "loss": 1.0034, "step": 8610 }, { "epoch": 1.6895335162681302, "grad_norm": 6.209445953369141, "learning_rate": 3.1296296296296295e-06, "loss": 0.9912, "step": 8620 }, { "epoch": 1.6914935319482556, "grad_norm": 6.419200897216797, "learning_rate": 3.1203703703703708e-06, "loss": 1.0048, "step": 8630 }, { "epoch": 1.6934535476283812, "grad_norm": 6.344346046447754, "learning_rate": 3.1111111111111116e-06, "loss": 0.9986, "step": 8640 }, { "epoch": 1.6954135633085063, "grad_norm": 6.1742963790893555, "learning_rate": 3.101851851851852e-06, "loss": 1.0085, "step": 8650 }, { "epoch": 1.697373578988632, "grad_norm": 5.269043445587158, "learning_rate": 3.0925925925925928e-06, "loss": 1.0045, "step": 8660 }, { "epoch": 1.6993335946687573, "grad_norm": 6.707599639892578, "learning_rate": 3.0833333333333336e-06, "loss": 1.0074, "step": 8670 }, { "epoch": 1.7012936103488827, "grad_norm": 6.476319313049316, "learning_rate": 3.0740740740740744e-06, "loss": 1.0066, "step": 8680 }, { "epoch": 1.7032536260290083, "grad_norm": 5.808709621429443, "learning_rate": 3.064814814814815e-06, "loss": 0.9969, "step": 8690 }, { "epoch": 1.7052136417091337, "grad_norm": 10.500265121459961, "learning_rate": 3.055555555555556e-06, "loss": 1.0062, "step": 8700 }, { "epoch": 1.707173657389259, "grad_norm": 5.992859840393066, "learning_rate": 3.0462962962962964e-06, "loss": 1.0133, "step": 8710 }, { "epoch": 1.7091336730693847, "grad_norm": 6.0215325355529785, "learning_rate": 3.0370370370370372e-06, "loss": 1.0175, "step": 8720 }, { "epoch": 1.7110936887495098, "grad_norm": 6.671189785003662, "learning_rate": 3.0277777777777776e-06, "loss": 1.0113, "step": 8730 }, { "epoch": 1.7130537044296354, "grad_norm": 6.012123107910156, "learning_rate": 3.018518518518519e-06, "loss": 0.9931, "step": 8740 }, { "epoch": 1.7150137201097608, "grad_norm": 5.328975200653076, "learning_rate": 3.0092592592592597e-06, "loss": 1.0052, "step": 8750 }, { "epoch": 1.7169737357898862, "grad_norm": 6.184416770935059, "learning_rate": 3e-06, "loss": 1.0125, "step": 8760 }, { "epoch": 1.7189337514700118, "grad_norm": 6.222989559173584, "learning_rate": 2.990740740740741e-06, "loss": 1.0166, "step": 8770 }, { "epoch": 1.7208937671501372, "grad_norm": 6.706179141998291, "learning_rate": 2.9814814814814817e-06, "loss": 1.0095, "step": 8780 }, { "epoch": 1.7228537828302626, "grad_norm": 6.823479652404785, "learning_rate": 2.9722222222222225e-06, "loss": 1.0147, "step": 8790 }, { "epoch": 1.7248137985103882, "grad_norm": 7.233177185058594, "learning_rate": 2.962962962962963e-06, "loss": 0.9963, "step": 8800 }, { "epoch": 1.7267738141905136, "grad_norm": 5.568224906921387, "learning_rate": 2.953703703703704e-06, "loss": 1.0052, "step": 8810 }, { "epoch": 1.728733829870639, "grad_norm": 6.352148532867432, "learning_rate": 2.944444444444445e-06, "loss": 0.9951, "step": 8820 }, { "epoch": 1.7306938455507646, "grad_norm": 6.312885284423828, "learning_rate": 2.9351851851851853e-06, "loss": 1.0126, "step": 8830 }, { "epoch": 1.7326538612308897, "grad_norm": 6.648190975189209, "learning_rate": 2.9259259259259257e-06, "loss": 1.0095, "step": 8840 }, { "epoch": 1.7346138769110153, "grad_norm": 7.448049545288086, "learning_rate": 2.916666666666667e-06, "loss": 1.0058, "step": 8850 }, { "epoch": 1.7365738925911407, "grad_norm": 6.0499653816223145, "learning_rate": 2.907407407407408e-06, "loss": 1.002, "step": 8860 }, { "epoch": 1.738533908271266, "grad_norm": 6.453138828277588, "learning_rate": 2.898148148148148e-06, "loss": 1.0107, "step": 8870 }, { "epoch": 1.7404939239513917, "grad_norm": 6.5974555015563965, "learning_rate": 2.888888888888889e-06, "loss": 1.0119, "step": 8880 }, { "epoch": 1.742453939631517, "grad_norm": 6.174962043762207, "learning_rate": 2.87962962962963e-06, "loss": 1.0067, "step": 8890 }, { "epoch": 1.7444139553116425, "grad_norm": 6.343291282653809, "learning_rate": 2.8703703703703706e-06, "loss": 0.9903, "step": 8900 }, { "epoch": 1.746373970991768, "grad_norm": 6.71571159362793, "learning_rate": 2.861111111111111e-06, "loss": 1.018, "step": 8910 }, { "epoch": 1.7483339866718932, "grad_norm": 6.948620319366455, "learning_rate": 2.8518518518518522e-06, "loss": 1.0138, "step": 8920 }, { "epoch": 1.7502940023520188, "grad_norm": 6.206993579864502, "learning_rate": 2.842592592592593e-06, "loss": 1.0024, "step": 8930 }, { "epoch": 1.7522540180321444, "grad_norm": 6.701568603515625, "learning_rate": 2.8333333333333335e-06, "loss": 1.0117, "step": 8940 }, { "epoch": 1.7542140337122696, "grad_norm": 6.476100921630859, "learning_rate": 2.8240740740740743e-06, "loss": 1.0, "step": 8950 }, { "epoch": 1.7561740493923952, "grad_norm": 7.1540703773498535, "learning_rate": 2.814814814814815e-06, "loss": 0.9933, "step": 8960 }, { "epoch": 1.7581340650725206, "grad_norm": 6.764838695526123, "learning_rate": 2.805555555555556e-06, "loss": 1.0051, "step": 8970 }, { "epoch": 1.760094080752646, "grad_norm": 5.835758209228516, "learning_rate": 2.7962962962962963e-06, "loss": 1.0058, "step": 8980 }, { "epoch": 1.7620540964327716, "grad_norm": 6.640206813812256, "learning_rate": 2.7870370370370375e-06, "loss": 1.0012, "step": 8990 }, { "epoch": 1.764014112112897, "grad_norm": 7.032010555267334, "learning_rate": 2.7777777777777783e-06, "loss": 0.9988, "step": 9000 }, { "epoch": 1.7659741277930223, "grad_norm": 6.398713111877441, "learning_rate": 2.7685185185185187e-06, "loss": 1.0179, "step": 9010 }, { "epoch": 1.767934143473148, "grad_norm": 6.625948905944824, "learning_rate": 2.759259259259259e-06, "loss": 1.0089, "step": 9020 }, { "epoch": 1.769894159153273, "grad_norm": 5.858684062957764, "learning_rate": 2.7500000000000004e-06, "loss": 1.0057, "step": 9030 }, { "epoch": 1.7718541748333987, "grad_norm": 7.03538703918457, "learning_rate": 2.740740740740741e-06, "loss": 1.0223, "step": 9040 }, { "epoch": 1.773814190513524, "grad_norm": 6.332048416137695, "learning_rate": 2.7314814814814816e-06, "loss": 1.0027, "step": 9050 }, { "epoch": 1.7757742061936495, "grad_norm": 6.304274082183838, "learning_rate": 2.7222222222222224e-06, "loss": 0.994, "step": 9060 }, { "epoch": 1.777734221873775, "grad_norm": 5.588042736053467, "learning_rate": 2.712962962962963e-06, "loss": 1.0044, "step": 9070 }, { "epoch": 1.7796942375539004, "grad_norm": 6.531040191650391, "learning_rate": 2.703703703703704e-06, "loss": 1.0059, "step": 9080 }, { "epoch": 1.7816542532340258, "grad_norm": 6.511843681335449, "learning_rate": 2.6944444444444444e-06, "loss": 1.0046, "step": 9090 }, { "epoch": 1.7836142689141514, "grad_norm": 6.718422889709473, "learning_rate": 2.6851851851851856e-06, "loss": 1.0084, "step": 9100 }, { "epoch": 1.7836142689141514, "eval_loss": 1.0242658853530884, "eval_runtime": 14.0168, "eval_samples_per_second": 46.658, "eval_steps_per_second": 5.85, "step": 9100 }, { "epoch": 1.7855742845942768, "grad_norm": 5.511220455169678, "learning_rate": 2.6759259259259264e-06, "loss": 1.0028, "step": 9110 }, { "epoch": 1.7875343002744022, "grad_norm": 6.125180244445801, "learning_rate": 2.666666666666667e-06, "loss": 0.9937, "step": 9120 }, { "epoch": 1.7894943159545278, "grad_norm": 6.439998626708984, "learning_rate": 2.6574074074074076e-06, "loss": 0.9946, "step": 9130 }, { "epoch": 1.791454331634653, "grad_norm": 6.341091156005859, "learning_rate": 2.6481481481481485e-06, "loss": 1.0131, "step": 9140 }, { "epoch": 1.7934143473147786, "grad_norm": 6.909265041351318, "learning_rate": 2.6388888888888893e-06, "loss": 1.019, "step": 9150 }, { "epoch": 1.795374362994904, "grad_norm": 5.798973083496094, "learning_rate": 2.6296296296296297e-06, "loss": 1.0058, "step": 9160 }, { "epoch": 1.7973343786750293, "grad_norm": 6.564090728759766, "learning_rate": 2.6203703703703705e-06, "loss": 1.0012, "step": 9170 }, { "epoch": 1.799294394355155, "grad_norm": 6.725678443908691, "learning_rate": 2.6111111111111113e-06, "loss": 0.9987, "step": 9180 }, { "epoch": 1.8012544100352803, "grad_norm": 6.482659339904785, "learning_rate": 2.601851851851852e-06, "loss": 0.9986, "step": 9190 }, { "epoch": 1.8032144257154057, "grad_norm": 10.73702621459961, "learning_rate": 2.5925925925925925e-06, "loss": 0.9968, "step": 9200 }, { "epoch": 1.8051744413955313, "grad_norm": 6.893209457397461, "learning_rate": 2.5833333333333337e-06, "loss": 1.0111, "step": 9210 }, { "epoch": 1.8071344570756565, "grad_norm": 6.212709426879883, "learning_rate": 2.5740740740740745e-06, "loss": 1.0036, "step": 9220 }, { "epoch": 1.809094472755782, "grad_norm": 7.134426116943359, "learning_rate": 2.564814814814815e-06, "loss": 1.0008, "step": 9230 }, { "epoch": 1.8110544884359074, "grad_norm": 12.403793334960938, "learning_rate": 2.5555555555555557e-06, "loss": 0.9933, "step": 9240 }, { "epoch": 1.8130145041160328, "grad_norm": 6.31683349609375, "learning_rate": 2.5462962962962966e-06, "loss": 1.0083, "step": 9250 }, { "epoch": 1.8149745197961584, "grad_norm": 5.857621669769287, "learning_rate": 2.5370370370370374e-06, "loss": 0.993, "step": 9260 }, { "epoch": 1.8169345354762838, "grad_norm": 5.9981584548950195, "learning_rate": 2.5277777777777778e-06, "loss": 1.008, "step": 9270 }, { "epoch": 1.8188945511564092, "grad_norm": 6.6797404289245605, "learning_rate": 2.5185185185185186e-06, "loss": 0.9915, "step": 9280 }, { "epoch": 1.8208545668365348, "grad_norm": 6.050951957702637, "learning_rate": 2.50925925925926e-06, "loss": 1.0248, "step": 9290 }, { "epoch": 1.8228145825166602, "grad_norm": 6.692429542541504, "learning_rate": 2.5e-06, "loss": 1.0103, "step": 9300 }, { "epoch": 1.8247745981967856, "grad_norm": 6.396121978759766, "learning_rate": 2.490740740740741e-06, "loss": 1.0026, "step": 9310 }, { "epoch": 1.8267346138769112, "grad_norm": 6.65963077545166, "learning_rate": 2.481481481481482e-06, "loss": 0.9912, "step": 9320 }, { "epoch": 1.8286946295570363, "grad_norm": 7.833820343017578, "learning_rate": 2.4722222222222226e-06, "loss": 0.994, "step": 9330 }, { "epoch": 1.830654645237162, "grad_norm": 5.663256645202637, "learning_rate": 2.462962962962963e-06, "loss": 1.0014, "step": 9340 }, { "epoch": 1.8326146609172873, "grad_norm": 6.487533092498779, "learning_rate": 2.453703703703704e-06, "loss": 1.0022, "step": 9350 }, { "epoch": 1.8345746765974127, "grad_norm": 6.143520832061768, "learning_rate": 2.4444444444444447e-06, "loss": 0.9852, "step": 9360 }, { "epoch": 1.8365346922775383, "grad_norm": 6.246576309204102, "learning_rate": 2.4351851851851855e-06, "loss": 0.9995, "step": 9370 }, { "epoch": 1.8384947079576637, "grad_norm": 6.709537982940674, "learning_rate": 2.425925925925926e-06, "loss": 1.0039, "step": 9380 }, { "epoch": 1.840454723637789, "grad_norm": 6.031428813934326, "learning_rate": 2.4166666666666667e-06, "loss": 1.0103, "step": 9390 }, { "epoch": 1.8424147393179147, "grad_norm": 6.703726291656494, "learning_rate": 2.4074074074074075e-06, "loss": 0.9999, "step": 9400 }, { "epoch": 1.8443747549980398, "grad_norm": 6.8644819259643555, "learning_rate": 2.3981481481481483e-06, "loss": 0.9994, "step": 9410 }, { "epoch": 1.8463347706781654, "grad_norm": 7.306818962097168, "learning_rate": 2.388888888888889e-06, "loss": 0.9995, "step": 9420 }, { "epoch": 1.848294786358291, "grad_norm": 5.757288932800293, "learning_rate": 2.37962962962963e-06, "loss": 0.9883, "step": 9430 }, { "epoch": 1.8502548020384162, "grad_norm": 6.339224815368652, "learning_rate": 2.3703703703703707e-06, "loss": 0.9953, "step": 9440 }, { "epoch": 1.8522148177185418, "grad_norm": 8.186257362365723, "learning_rate": 2.361111111111111e-06, "loss": 0.9999, "step": 9450 }, { "epoch": 1.8541748333986672, "grad_norm": 6.2635111808776855, "learning_rate": 2.351851851851852e-06, "loss": 1.0047, "step": 9460 }, { "epoch": 1.8561348490787926, "grad_norm": 6.483547210693359, "learning_rate": 2.3425925925925928e-06, "loss": 1.0003, "step": 9470 }, { "epoch": 1.8580948647589182, "grad_norm": 6.6517333984375, "learning_rate": 2.3333333333333336e-06, "loss": 1.0152, "step": 9480 }, { "epoch": 1.8600548804390435, "grad_norm": 6.463938236236572, "learning_rate": 2.324074074074074e-06, "loss": 0.9917, "step": 9490 }, { "epoch": 1.862014896119169, "grad_norm": 6.184695243835449, "learning_rate": 2.314814814814815e-06, "loss": 0.9889, "step": 9500 }, { "epoch": 1.8639749117992945, "grad_norm": 6.435100078582764, "learning_rate": 2.305555555555556e-06, "loss": 1.002, "step": 9510 }, { "epoch": 1.8659349274794197, "grad_norm": 5.961505889892578, "learning_rate": 2.2962962962962964e-06, "loss": 0.9931, "step": 9520 }, { "epoch": 1.8678949431595453, "grad_norm": 6.590498924255371, "learning_rate": 2.2870370370370372e-06, "loss": 0.9967, "step": 9530 }, { "epoch": 1.8698549588396707, "grad_norm": 6.019999027252197, "learning_rate": 2.277777777777778e-06, "loss": 0.9887, "step": 9540 }, { "epoch": 1.871814974519796, "grad_norm": 6.332427978515625, "learning_rate": 2.268518518518519e-06, "loss": 1.0014, "step": 9550 }, { "epoch": 1.8737749901999217, "grad_norm": 6.301807880401611, "learning_rate": 2.2592592592592592e-06, "loss": 0.9969, "step": 9560 }, { "epoch": 1.875735005880047, "grad_norm": 6.279837131500244, "learning_rate": 2.25e-06, "loss": 0.9944, "step": 9570 }, { "epoch": 1.8776950215601724, "grad_norm": 7.422451496124268, "learning_rate": 2.240740740740741e-06, "loss": 1.0007, "step": 9580 }, { "epoch": 1.879655037240298, "grad_norm": 6.722255229949951, "learning_rate": 2.2314814814814817e-06, "loss": 0.9985, "step": 9590 }, { "epoch": 1.8816150529204234, "grad_norm": 6.471696853637695, "learning_rate": 2.222222222222222e-06, "loss": 0.9984, "step": 9600 }, { "epoch": 1.8835750686005488, "grad_norm": 6.347057819366455, "learning_rate": 2.2129629629629633e-06, "loss": 1.0092, "step": 9610 }, { "epoch": 1.8855350842806744, "grad_norm": 6.340747833251953, "learning_rate": 2.203703703703704e-06, "loss": 0.9978, "step": 9620 }, { "epoch": 1.8874950999607996, "grad_norm": 5.975046157836914, "learning_rate": 2.1944444444444445e-06, "loss": 1.0006, "step": 9630 }, { "epoch": 1.8894551156409252, "grad_norm": 6.538125038146973, "learning_rate": 2.1851851851851853e-06, "loss": 1.006, "step": 9640 }, { "epoch": 1.8914151313210505, "grad_norm": 6.345036506652832, "learning_rate": 2.175925925925926e-06, "loss": 1.0106, "step": 9650 }, { "epoch": 1.893375147001176, "grad_norm": 6.604660987854004, "learning_rate": 2.166666666666667e-06, "loss": 0.9964, "step": 9660 }, { "epoch": 1.8953351626813015, "grad_norm": 6.32635498046875, "learning_rate": 2.1574074074074073e-06, "loss": 1.0108, "step": 9670 }, { "epoch": 1.897295178361427, "grad_norm": 6.804265975952148, "learning_rate": 2.148148148148148e-06, "loss": 1.0053, "step": 9680 }, { "epoch": 1.8992551940415523, "grad_norm": 6.880146503448486, "learning_rate": 2.138888888888889e-06, "loss": 1.0073, "step": 9690 }, { "epoch": 1.901215209721678, "grad_norm": 7.2419538497924805, "learning_rate": 2.1296296296296298e-06, "loss": 1.004, "step": 9700 }, { "epoch": 1.903175225401803, "grad_norm": 6.319647789001465, "learning_rate": 2.1203703703703706e-06, "loss": 1.0049, "step": 9710 }, { "epoch": 1.9051352410819287, "grad_norm": 6.6896491050720215, "learning_rate": 2.1111111111111114e-06, "loss": 1.0007, "step": 9720 }, { "epoch": 1.907095256762054, "grad_norm": 5.643420696258545, "learning_rate": 2.1018518518518522e-06, "loss": 0.9898, "step": 9730 }, { "epoch": 1.9090552724421794, "grad_norm": 6.660647869110107, "learning_rate": 2.0925925925925926e-06, "loss": 1.0039, "step": 9740 }, { "epoch": 1.911015288122305, "grad_norm": 7.0795464515686035, "learning_rate": 2.0833333333333334e-06, "loss": 0.986, "step": 9750 }, { "epoch": 1.9129753038024304, "grad_norm": 6.72257137298584, "learning_rate": 2.0740740740740742e-06, "loss": 1.0045, "step": 9760 }, { "epoch": 1.9149353194825558, "grad_norm": 7.132964134216309, "learning_rate": 2.064814814814815e-06, "loss": 1.0043, "step": 9770 }, { "epoch": 1.9168953351626814, "grad_norm": 6.032354354858398, "learning_rate": 2.0555555555555555e-06, "loss": 1.0159, "step": 9780 }, { "epoch": 1.9188553508428068, "grad_norm": 6.651278495788574, "learning_rate": 2.0462962962962967e-06, "loss": 1.0009, "step": 9790 }, { "epoch": 1.9208153665229322, "grad_norm": 6.432239532470703, "learning_rate": 2.037037037037037e-06, "loss": 1.0035, "step": 9800 }, { "epoch": 1.9208153665229322, "eval_loss": 1.0201009511947632, "eval_runtime": 14.0626, "eval_samples_per_second": 46.506, "eval_steps_per_second": 5.831, "step": 9800 }, { "epoch": 1.9227753822030578, "grad_norm": 6.820202827453613, "learning_rate": 2.027777777777778e-06, "loss": 0.9879, "step": 9810 }, { "epoch": 1.924735397883183, "grad_norm": 6.324273586273193, "learning_rate": 2.0185185185185187e-06, "loss": 1.0022, "step": 9820 }, { "epoch": 1.9266954135633085, "grad_norm": 5.975357532501221, "learning_rate": 2.0092592592592595e-06, "loss": 1.0168, "step": 9830 }, { "epoch": 1.928655429243434, "grad_norm": 7.312973976135254, "learning_rate": 2.0000000000000003e-06, "loss": 1.0017, "step": 9840 }, { "epoch": 1.9306154449235593, "grad_norm": 7.051765441894531, "learning_rate": 1.9907407407407407e-06, "loss": 0.992, "step": 9850 }, { "epoch": 1.932575460603685, "grad_norm": 6.269126892089844, "learning_rate": 1.9814814814814815e-06, "loss": 1.0018, "step": 9860 }, { "epoch": 1.9345354762838103, "grad_norm": 6.204473972320557, "learning_rate": 1.9722222222222224e-06, "loss": 0.999, "step": 9870 }, { "epoch": 1.9364954919639357, "grad_norm": 6.758467197418213, "learning_rate": 1.962962962962963e-06, "loss": 1.0023, "step": 9880 }, { "epoch": 1.9384555076440613, "grad_norm": 6.538867950439453, "learning_rate": 1.953703703703704e-06, "loss": 0.9907, "step": 9890 }, { "epoch": 1.9404155233241864, "grad_norm": 7.003431797027588, "learning_rate": 1.944444444444445e-06, "loss": 1.0047, "step": 9900 }, { "epoch": 1.942375539004312, "grad_norm": 6.028417587280273, "learning_rate": 1.935185185185185e-06, "loss": 1.0027, "step": 9910 }, { "epoch": 1.9443355546844376, "grad_norm": 6.356867790222168, "learning_rate": 1.925925925925926e-06, "loss": 0.9902, "step": 9920 }, { "epoch": 1.9462955703645628, "grad_norm": 7.202489376068115, "learning_rate": 1.916666666666667e-06, "loss": 1.0106, "step": 9930 }, { "epoch": 1.9482555860446884, "grad_norm": 6.344156265258789, "learning_rate": 1.9074074074074076e-06, "loss": 0.9916, "step": 9940 }, { "epoch": 1.9502156017248138, "grad_norm": 6.817245006561279, "learning_rate": 1.8981481481481484e-06, "loss": 0.9829, "step": 9950 }, { "epoch": 1.9521756174049392, "grad_norm": 6.988001823425293, "learning_rate": 1.888888888888889e-06, "loss": 0.9997, "step": 9960 }, { "epoch": 1.9541356330850648, "grad_norm": 6.65484619140625, "learning_rate": 1.8796296296296299e-06, "loss": 1.0026, "step": 9970 }, { "epoch": 1.9560956487651902, "grad_norm": 6.717133045196533, "learning_rate": 1.8703703703703705e-06, "loss": 1.0071, "step": 9980 }, { "epoch": 1.9580556644453155, "grad_norm": 6.037536144256592, "learning_rate": 1.8611111111111113e-06, "loss": 0.993, "step": 9990 }, { "epoch": 1.9600156801254411, "grad_norm": 6.965274810791016, "learning_rate": 1.8518518518518519e-06, "loss": 0.9918, "step": 10000 }, { "epoch": 1.9619756958055663, "grad_norm": 6.768052101135254, "learning_rate": 1.8425925925925927e-06, "loss": 0.9873, "step": 10010 }, { "epoch": 1.963935711485692, "grad_norm": 6.111681938171387, "learning_rate": 1.8333333333333333e-06, "loss": 1.0111, "step": 10020 }, { "epoch": 1.9658957271658173, "grad_norm": 6.606689453125, "learning_rate": 1.8240740740740743e-06, "loss": 0.9997, "step": 10030 }, { "epoch": 1.9678557428459427, "grad_norm": 6.256832599639893, "learning_rate": 1.8148148148148151e-06, "loss": 0.9977, "step": 10040 }, { "epoch": 1.9698157585260683, "grad_norm": 6.401496410369873, "learning_rate": 1.8055555555555557e-06, "loss": 1.0069, "step": 10050 }, { "epoch": 1.9717757742061937, "grad_norm": 6.166873455047607, "learning_rate": 1.7962962962962965e-06, "loss": 0.9998, "step": 10060 }, { "epoch": 1.973735789886319, "grad_norm": 5.77340030670166, "learning_rate": 1.7870370370370371e-06, "loss": 1.0008, "step": 10070 }, { "epoch": 1.9756958055664446, "grad_norm": 7.6040120124816895, "learning_rate": 1.777777777777778e-06, "loss": 0.9991, "step": 10080 }, { "epoch": 1.97765582124657, "grad_norm": 6.114837169647217, "learning_rate": 1.7685185185185186e-06, "loss": 0.9949, "step": 10090 }, { "epoch": 1.9796158369266954, "grad_norm": 6.565464496612549, "learning_rate": 1.7592592592592594e-06, "loss": 1.0141, "step": 10100 }, { "epoch": 1.981575852606821, "grad_norm": 6.436923980712891, "learning_rate": 1.75e-06, "loss": 0.9975, "step": 10110 }, { "epoch": 1.9835358682869462, "grad_norm": 7.330633640289307, "learning_rate": 1.740740740740741e-06, "loss": 1.0014, "step": 10120 }, { "epoch": 1.9854958839670718, "grad_norm": 7.172529697418213, "learning_rate": 1.7314814814814814e-06, "loss": 0.9811, "step": 10130 }, { "epoch": 1.9874558996471972, "grad_norm": 6.164400577545166, "learning_rate": 1.7222222222222224e-06, "loss": 0.9972, "step": 10140 }, { "epoch": 1.9894159153273225, "grad_norm": 6.392111778259277, "learning_rate": 1.7129629629629632e-06, "loss": 1.0001, "step": 10150 }, { "epoch": 1.9913759310074481, "grad_norm": 8.478386878967285, "learning_rate": 1.7037037037037038e-06, "loss": 1.0015, "step": 10160 }, { "epoch": 1.9933359466875735, "grad_norm": 6.89587926864624, "learning_rate": 1.6944444444444446e-06, "loss": 1.0047, "step": 10170 }, { "epoch": 1.995295962367699, "grad_norm": 6.686724662780762, "learning_rate": 1.6851851851851852e-06, "loss": 0.9953, "step": 10180 }, { "epoch": 1.9972559780478245, "grad_norm": 6.857761859893799, "learning_rate": 1.675925925925926e-06, "loss": 0.9836, "step": 10190 }, { "epoch": 1.9992159937279497, "grad_norm": 6.607189178466797, "learning_rate": 1.6666666666666667e-06, "loss": 0.9979, "step": 10200 }, { "epoch": 2.0011760094080753, "grad_norm": 6.168262481689453, "learning_rate": 1.6574074074074075e-06, "loss": 0.9719, "step": 10210 }, { "epoch": 2.003136025088201, "grad_norm": 6.525495529174805, "learning_rate": 1.648148148148148e-06, "loss": 0.9789, "step": 10220 }, { "epoch": 2.005096040768326, "grad_norm": 5.562708854675293, "learning_rate": 1.638888888888889e-06, "loss": 0.98, "step": 10230 }, { "epoch": 2.0070560564484516, "grad_norm": 7.023379802703857, "learning_rate": 1.62962962962963e-06, "loss": 0.985, "step": 10240 }, { "epoch": 2.0090160721285772, "grad_norm": 7.55203104019165, "learning_rate": 1.6203703703703705e-06, "loss": 0.9758, "step": 10250 }, { "epoch": 2.0109760878087024, "grad_norm": 13.453210830688477, "learning_rate": 1.6111111111111113e-06, "loss": 0.9656, "step": 10260 }, { "epoch": 2.012936103488828, "grad_norm": 6.530963897705078, "learning_rate": 1.601851851851852e-06, "loss": 0.9615, "step": 10270 }, { "epoch": 2.014896119168953, "grad_norm": 6.9765520095825195, "learning_rate": 1.5925925925925927e-06, "loss": 0.9786, "step": 10280 }, { "epoch": 2.0168561348490788, "grad_norm": 7.124659538269043, "learning_rate": 1.5833333333333333e-06, "loss": 0.9795, "step": 10290 }, { "epoch": 2.0188161505292044, "grad_norm": 7.119600772857666, "learning_rate": 1.5740740740740742e-06, "loss": 0.9672, "step": 10300 }, { "epoch": 2.0207761662093295, "grad_norm": 6.711264133453369, "learning_rate": 1.5648148148148148e-06, "loss": 0.965, "step": 10310 }, { "epoch": 2.022736181889455, "grad_norm": 7.027464389801025, "learning_rate": 1.5555555555555558e-06, "loss": 0.97, "step": 10320 }, { "epoch": 2.0246961975695807, "grad_norm": 7.815002918243408, "learning_rate": 1.5462962962962964e-06, "loss": 0.9706, "step": 10330 }, { "epoch": 2.026656213249706, "grad_norm": 6.989712238311768, "learning_rate": 1.5370370370370372e-06, "loss": 0.9741, "step": 10340 }, { "epoch": 2.0286162289298315, "grad_norm": 6.903376579284668, "learning_rate": 1.527777777777778e-06, "loss": 0.9707, "step": 10350 }, { "epoch": 2.0305762446099567, "grad_norm": 5.8991570472717285, "learning_rate": 1.5185185185185186e-06, "loss": 0.9731, "step": 10360 }, { "epoch": 2.0325362602900823, "grad_norm": 7.208601474761963, "learning_rate": 1.5092592592592594e-06, "loss": 0.9845, "step": 10370 }, { "epoch": 2.034496275970208, "grad_norm": 6.827620506286621, "learning_rate": 1.5e-06, "loss": 0.968, "step": 10380 }, { "epoch": 2.036456291650333, "grad_norm": 6.6345295906066895, "learning_rate": 1.4907407407407409e-06, "loss": 0.9626, "step": 10390 }, { "epoch": 2.0384163073304586, "grad_norm": 7.7344865798950195, "learning_rate": 1.4814814814814815e-06, "loss": 0.9717, "step": 10400 }, { "epoch": 2.0403763230105842, "grad_norm": 7.092048168182373, "learning_rate": 1.4722222222222225e-06, "loss": 0.9721, "step": 10410 }, { "epoch": 2.0423363386907094, "grad_norm": 6.746728897094727, "learning_rate": 1.4629629629629629e-06, "loss": 0.9755, "step": 10420 }, { "epoch": 2.044296354370835, "grad_norm": 7.295462131500244, "learning_rate": 1.453703703703704e-06, "loss": 0.9633, "step": 10430 }, { "epoch": 2.0462563700509606, "grad_norm": 6.882177829742432, "learning_rate": 1.4444444444444445e-06, "loss": 0.9668, "step": 10440 }, { "epoch": 2.0482163857310858, "grad_norm": 6.1587300300598145, "learning_rate": 1.4351851851851853e-06, "loss": 0.9773, "step": 10450 }, { "epoch": 2.0501764014112114, "grad_norm": 7.219092845916748, "learning_rate": 1.4259259259259261e-06, "loss": 0.9724, "step": 10460 }, { "epoch": 2.0521364170913365, "grad_norm": 6.330183029174805, "learning_rate": 1.4166666666666667e-06, "loss": 0.965, "step": 10470 }, { "epoch": 2.054096432771462, "grad_norm": 7.83421516418457, "learning_rate": 1.4074074074074075e-06, "loss": 0.9715, "step": 10480 }, { "epoch": 2.0560564484515877, "grad_norm": 6.530287265777588, "learning_rate": 1.3981481481481481e-06, "loss": 0.9859, "step": 10490 }, { "epoch": 2.058016464131713, "grad_norm": 7.4246625900268555, "learning_rate": 1.3888888888888892e-06, "loss": 0.9567, "step": 10500 }, { "epoch": 2.058016464131713, "eval_loss": 1.0192580223083496, "eval_runtime": 14.0156, "eval_samples_per_second": 46.662, "eval_steps_per_second": 5.851, "step": 10500 }, { "epoch": 2.0599764798118385, "grad_norm": 7.052817344665527, "learning_rate": 1.3796296296296296e-06, "loss": 0.9657, "step": 10510 }, { "epoch": 2.061936495491964, "grad_norm": 7.35888147354126, "learning_rate": 1.3703703703703706e-06, "loss": 0.9702, "step": 10520 }, { "epoch": 2.0638965111720893, "grad_norm": 6.373018264770508, "learning_rate": 1.3611111111111112e-06, "loss": 0.9753, "step": 10530 }, { "epoch": 2.065856526852215, "grad_norm": 6.720044136047363, "learning_rate": 1.351851851851852e-06, "loss": 0.9708, "step": 10540 }, { "epoch": 2.0678165425323405, "grad_norm": 6.653097152709961, "learning_rate": 1.3425925925925928e-06, "loss": 0.9702, "step": 10550 }, { "epoch": 2.0697765582124656, "grad_norm": 7.028474807739258, "learning_rate": 1.3333333333333334e-06, "loss": 0.9787, "step": 10560 }, { "epoch": 2.0717365738925912, "grad_norm": 7.457616329193115, "learning_rate": 1.3240740740740742e-06, "loss": 0.9585, "step": 10570 }, { "epoch": 2.0736965895727164, "grad_norm": 6.340860843658447, "learning_rate": 1.3148148148148148e-06, "loss": 0.9592, "step": 10580 }, { "epoch": 2.075656605252842, "grad_norm": 7.086592674255371, "learning_rate": 1.3055555555555556e-06, "loss": 0.9701, "step": 10590 }, { "epoch": 2.0776166209329676, "grad_norm": 7.2209320068359375, "learning_rate": 1.2962962962962962e-06, "loss": 0.9655, "step": 10600 }, { "epoch": 2.0795766366130928, "grad_norm": 6.678873062133789, "learning_rate": 1.2870370370370373e-06, "loss": 0.9674, "step": 10610 }, { "epoch": 2.0815366522932184, "grad_norm": 7.573465347290039, "learning_rate": 1.2777777777777779e-06, "loss": 0.9754, "step": 10620 }, { "epoch": 2.083496667973344, "grad_norm": 7.289187908172607, "learning_rate": 1.2685185185185187e-06, "loss": 0.9769, "step": 10630 }, { "epoch": 2.085456683653469, "grad_norm": 6.930860996246338, "learning_rate": 1.2592592592592593e-06, "loss": 0.9697, "step": 10640 }, { "epoch": 2.0874166993335947, "grad_norm": 6.5301618576049805, "learning_rate": 1.25e-06, "loss": 0.9766, "step": 10650 }, { "epoch": 2.08937671501372, "grad_norm": 8.65379524230957, "learning_rate": 1.240740740740741e-06, "loss": 0.9545, "step": 10660 }, { "epoch": 2.0913367306938455, "grad_norm": 6.923322677612305, "learning_rate": 1.2314814814814815e-06, "loss": 0.9643, "step": 10670 }, { "epoch": 2.093296746373971, "grad_norm": 7.2696614265441895, "learning_rate": 1.2222222222222223e-06, "loss": 0.9677, "step": 10680 }, { "epoch": 2.0952567620540963, "grad_norm": 7.1026835441589355, "learning_rate": 1.212962962962963e-06, "loss": 0.9791, "step": 10690 }, { "epoch": 2.097216777734222, "grad_norm": 6.935880661010742, "learning_rate": 1.2037037037037037e-06, "loss": 0.9639, "step": 10700 }, { "epoch": 2.0991767934143475, "grad_norm": 6.7199201583862305, "learning_rate": 1.1944444444444446e-06, "loss": 0.9795, "step": 10710 }, { "epoch": 2.1011368090944726, "grad_norm": 7.126111030578613, "learning_rate": 1.1851851851851854e-06, "loss": 0.9604, "step": 10720 }, { "epoch": 2.1030968247745983, "grad_norm": 7.274319171905518, "learning_rate": 1.175925925925926e-06, "loss": 0.9696, "step": 10730 }, { "epoch": 2.105056840454724, "grad_norm": 7.07975959777832, "learning_rate": 1.1666666666666668e-06, "loss": 0.9708, "step": 10740 }, { "epoch": 2.107016856134849, "grad_norm": 24.05130386352539, "learning_rate": 1.1574074074074076e-06, "loss": 0.9765, "step": 10750 }, { "epoch": 2.1089768718149746, "grad_norm": 7.428647518157959, "learning_rate": 1.1481481481481482e-06, "loss": 0.9665, "step": 10760 }, { "epoch": 2.1109368874951, "grad_norm": 6.800421714782715, "learning_rate": 1.138888888888889e-06, "loss": 0.9687, "step": 10770 }, { "epoch": 2.1128969031752254, "grad_norm": 8.014981269836426, "learning_rate": 1.1296296296296296e-06, "loss": 0.9773, "step": 10780 }, { "epoch": 2.114856918855351, "grad_norm": 7.821423053741455, "learning_rate": 1.1203703703703704e-06, "loss": 0.9694, "step": 10790 }, { "epoch": 2.116816934535476, "grad_norm": 8.069396018981934, "learning_rate": 1.111111111111111e-06, "loss": 0.9741, "step": 10800 }, { "epoch": 2.1187769502156018, "grad_norm": 7.268844127655029, "learning_rate": 1.101851851851852e-06, "loss": 0.9685, "step": 10810 }, { "epoch": 2.1207369658957274, "grad_norm": 7.122548580169678, "learning_rate": 1.0925925925925927e-06, "loss": 0.9491, "step": 10820 }, { "epoch": 2.1226969815758525, "grad_norm": 6.834280014038086, "learning_rate": 1.0833333333333335e-06, "loss": 0.9836, "step": 10830 }, { "epoch": 2.124656997255978, "grad_norm": 6.231177806854248, "learning_rate": 1.074074074074074e-06, "loss": 0.9729, "step": 10840 }, { "epoch": 2.1266170129361033, "grad_norm": 7.387545585632324, "learning_rate": 1.0648148148148149e-06, "loss": 0.9658, "step": 10850 }, { "epoch": 2.128577028616229, "grad_norm": 8.538785934448242, "learning_rate": 1.0555555555555557e-06, "loss": 0.9521, "step": 10860 }, { "epoch": 2.1305370442963545, "grad_norm": 7.329588890075684, "learning_rate": 1.0462962962962963e-06, "loss": 0.9804, "step": 10870 }, { "epoch": 2.1324970599764796, "grad_norm": 6.391878128051758, "learning_rate": 1.0370370370370371e-06, "loss": 0.9684, "step": 10880 }, { "epoch": 2.1344570756566053, "grad_norm": 7.090411186218262, "learning_rate": 1.0277777777777777e-06, "loss": 0.9622, "step": 10890 }, { "epoch": 2.136417091336731, "grad_norm": 6.1176958084106445, "learning_rate": 1.0185185185185185e-06, "loss": 0.9696, "step": 10900 }, { "epoch": 2.138377107016856, "grad_norm": 10.32075023651123, "learning_rate": 1.0092592592592594e-06, "loss": 0.9686, "step": 10910 }, { "epoch": 2.1403371226969816, "grad_norm": 7.174540996551514, "learning_rate": 1.0000000000000002e-06, "loss": 0.9647, "step": 10920 }, { "epoch": 2.1422971383771072, "grad_norm": 7.739523410797119, "learning_rate": 9.907407407407408e-07, "loss": 0.9646, "step": 10930 }, { "epoch": 2.1442571540572324, "grad_norm": 7.084393501281738, "learning_rate": 9.814814814814816e-07, "loss": 0.9695, "step": 10940 }, { "epoch": 2.146217169737358, "grad_norm": 7.240608215332031, "learning_rate": 9.722222222222224e-07, "loss": 0.9655, "step": 10950 }, { "epoch": 2.148177185417483, "grad_norm": 7.608102798461914, "learning_rate": 9.62962962962963e-07, "loss": 0.9598, "step": 10960 }, { "epoch": 2.1501372010976088, "grad_norm": 7.160534858703613, "learning_rate": 9.537037037037038e-07, "loss": 0.9649, "step": 10970 }, { "epoch": 2.1520972167777344, "grad_norm": 7.145265102386475, "learning_rate": 9.444444444444445e-07, "loss": 0.9646, "step": 10980 }, { "epoch": 2.1540572324578595, "grad_norm": 25.257863998413086, "learning_rate": 9.351851851851852e-07, "loss": 0.9529, "step": 10990 }, { "epoch": 2.156017248137985, "grad_norm": 7.226001262664795, "learning_rate": 9.259259259259259e-07, "loss": 0.9556, "step": 11000 }, { "epoch": 2.1579772638181107, "grad_norm": 8.324951171875, "learning_rate": 9.166666666666666e-07, "loss": 0.9666, "step": 11010 }, { "epoch": 2.159937279498236, "grad_norm": 7.676420211791992, "learning_rate": 9.074074074074076e-07, "loss": 0.9822, "step": 11020 }, { "epoch": 2.1618972951783615, "grad_norm": 8.159540176391602, "learning_rate": 8.981481481481483e-07, "loss": 0.9608, "step": 11030 }, { "epoch": 2.1638573108584866, "grad_norm": 7.420976161956787, "learning_rate": 8.88888888888889e-07, "loss": 0.974, "step": 11040 }, { "epoch": 2.1658173265386123, "grad_norm": 7.221758842468262, "learning_rate": 8.796296296296297e-07, "loss": 0.9679, "step": 11050 }, { "epoch": 2.167777342218738, "grad_norm": 7.399634838104248, "learning_rate": 8.703703703703705e-07, "loss": 0.9699, "step": 11060 }, { "epoch": 2.169737357898863, "grad_norm": 6.9496026039123535, "learning_rate": 8.611111111111112e-07, "loss": 0.963, "step": 11070 }, { "epoch": 2.1716973735789886, "grad_norm": 7.964886665344238, "learning_rate": 8.518518518518519e-07, "loss": 0.9649, "step": 11080 }, { "epoch": 2.1736573892591142, "grad_norm": 6.945036888122559, "learning_rate": 8.425925925925926e-07, "loss": 0.9658, "step": 11090 }, { "epoch": 2.1756174049392394, "grad_norm": 7.315770149230957, "learning_rate": 8.333333333333333e-07, "loss": 0.9715, "step": 11100 }, { "epoch": 2.177577420619365, "grad_norm": 7.41797399520874, "learning_rate": 8.24074074074074e-07, "loss": 0.9579, "step": 11110 }, { "epoch": 2.1795374362994906, "grad_norm": 6.9420366287231445, "learning_rate": 8.14814814814815e-07, "loss": 0.9785, "step": 11120 }, { "epoch": 2.1814974519796158, "grad_norm": 6.257632255554199, "learning_rate": 8.055555555555557e-07, "loss": 0.9673, "step": 11130 }, { "epoch": 2.1834574676597414, "grad_norm": 7.180861949920654, "learning_rate": 7.962962962962964e-07, "loss": 0.9691, "step": 11140 }, { "epoch": 2.1854174833398665, "grad_norm": 7.286172389984131, "learning_rate": 7.870370370370371e-07, "loss": 0.9583, "step": 11150 }, { "epoch": 2.187377499019992, "grad_norm": 7.305385589599609, "learning_rate": 7.777777777777779e-07, "loss": 0.9632, "step": 11160 }, { "epoch": 2.1893375147001177, "grad_norm": 7.4768290519714355, "learning_rate": 7.685185185185186e-07, "loss": 0.9777, "step": 11170 }, { "epoch": 2.191297530380243, "grad_norm": 7.44431209564209, "learning_rate": 7.592592592592593e-07, "loss": 0.9748, "step": 11180 }, { "epoch": 2.1932575460603685, "grad_norm": 7.362459182739258, "learning_rate": 7.5e-07, "loss": 0.97, "step": 11190 }, { "epoch": 2.195217561740494, "grad_norm": 6.118969917297363, "learning_rate": 7.407407407407407e-07, "loss": 0.967, "step": 11200 }, { "epoch": 2.195217561740494, "eval_loss": 1.015982985496521, "eval_runtime": 14.0402, "eval_samples_per_second": 46.58, "eval_steps_per_second": 5.84, "step": 11200 }, { "epoch": 2.1971775774206193, "grad_norm": 7.092718601226807, "learning_rate": 7.314814814814814e-07, "loss": 0.9661, "step": 11210 }, { "epoch": 2.199137593100745, "grad_norm": 7.384366512298584, "learning_rate": 7.222222222222222e-07, "loss": 0.9669, "step": 11220 }, { "epoch": 2.20109760878087, "grad_norm": 7.769095420837402, "learning_rate": 7.129629629629631e-07, "loss": 0.9674, "step": 11230 }, { "epoch": 2.2030576244609956, "grad_norm": 6.9204864501953125, "learning_rate": 7.037037037037038e-07, "loss": 0.9661, "step": 11240 }, { "epoch": 2.2050176401411212, "grad_norm": 6.869389057159424, "learning_rate": 6.944444444444446e-07, "loss": 0.9619, "step": 11250 }, { "epoch": 2.2069776558212464, "grad_norm": 7.39153528213501, "learning_rate": 6.851851851851853e-07, "loss": 0.9605, "step": 11260 }, { "epoch": 2.208937671501372, "grad_norm": 6.61447811126709, "learning_rate": 6.75925925925926e-07, "loss": 0.9796, "step": 11270 }, { "epoch": 2.2108976871814976, "grad_norm": 6.7030863761901855, "learning_rate": 6.666666666666667e-07, "loss": 0.96, "step": 11280 }, { "epoch": 2.2128577028616228, "grad_norm": 7.7893829345703125, "learning_rate": 6.574074074074074e-07, "loss": 0.9571, "step": 11290 }, { "epoch": 2.2148177185417484, "grad_norm": 7.597008228302002, "learning_rate": 6.481481481481481e-07, "loss": 0.9672, "step": 11300 }, { "epoch": 2.216777734221874, "grad_norm": 6.373151779174805, "learning_rate": 6.388888888888889e-07, "loss": 0.9713, "step": 11310 }, { "epoch": 2.218737749901999, "grad_norm": 7.9771575927734375, "learning_rate": 6.296296296296296e-07, "loss": 0.9674, "step": 11320 }, { "epoch": 2.2206977655821247, "grad_norm": 7.23372745513916, "learning_rate": 6.203703703703705e-07, "loss": 0.9787, "step": 11330 }, { "epoch": 2.2226577812622503, "grad_norm": 7.515969276428223, "learning_rate": 6.111111111111112e-07, "loss": 0.9702, "step": 11340 }, { "epoch": 2.2246177969423755, "grad_norm": 6.468687534332275, "learning_rate": 6.018518518518519e-07, "loss": 0.981, "step": 11350 }, { "epoch": 2.226577812622501, "grad_norm": 6.765408992767334, "learning_rate": 5.925925925925927e-07, "loss": 0.9757, "step": 11360 }, { "epoch": 2.2285378283026263, "grad_norm": 8.031702041625977, "learning_rate": 5.833333333333334e-07, "loss": 0.9643, "step": 11370 }, { "epoch": 2.230497843982752, "grad_norm": 7.857892990112305, "learning_rate": 5.740740740740741e-07, "loss": 0.9712, "step": 11380 }, { "epoch": 2.2324578596628775, "grad_norm": 6.461490154266357, "learning_rate": 5.648148148148148e-07, "loss": 0.9664, "step": 11390 }, { "epoch": 2.2344178753430026, "grad_norm": 7.290163993835449, "learning_rate": 5.555555555555555e-07, "loss": 0.9554, "step": 11400 }, { "epoch": 2.2363778910231282, "grad_norm": 7.196146011352539, "learning_rate": 5.462962962962963e-07, "loss": 0.9755, "step": 11410 }, { "epoch": 2.238337906703254, "grad_norm": 7.482301712036133, "learning_rate": 5.37037037037037e-07, "loss": 0.9577, "step": 11420 }, { "epoch": 2.240297922383379, "grad_norm": 7.794521808624268, "learning_rate": 5.277777777777779e-07, "loss": 0.9679, "step": 11430 }, { "epoch": 2.2422579380635046, "grad_norm": 6.857132434844971, "learning_rate": 5.185185185185186e-07, "loss": 0.9561, "step": 11440 }, { "epoch": 2.2442179537436298, "grad_norm": 6.887381553649902, "learning_rate": 5.092592592592593e-07, "loss": 0.9712, "step": 11450 }, { "epoch": 2.2461779694237554, "grad_norm": 7.537138938903809, "learning_rate": 5.000000000000001e-07, "loss": 0.9755, "step": 11460 }, { "epoch": 2.248137985103881, "grad_norm": 7.604095935821533, "learning_rate": 4.907407407407408e-07, "loss": 0.968, "step": 11470 }, { "epoch": 2.250098000784006, "grad_norm": 6.840078830718994, "learning_rate": 4.814814814814815e-07, "loss": 0.9677, "step": 11480 }, { "epoch": 2.2520580164641317, "grad_norm": 7.0317301750183105, "learning_rate": 4.7222222222222226e-07, "loss": 0.9635, "step": 11490 }, { "epoch": 2.2540180321442573, "grad_norm": 7.006657123565674, "learning_rate": 4.6296296296296297e-07, "loss": 0.96, "step": 11500 }, { "epoch": 2.2559780478243825, "grad_norm": 7.701179504394531, "learning_rate": 4.537037037037038e-07, "loss": 0.9747, "step": 11510 }, { "epoch": 2.257938063504508, "grad_norm": 7.04123592376709, "learning_rate": 4.444444444444445e-07, "loss": 0.9632, "step": 11520 }, { "epoch": 2.2598980791846337, "grad_norm": 7.217432022094727, "learning_rate": 4.3518518518518525e-07, "loss": 0.9675, "step": 11530 }, { "epoch": 2.261858094864759, "grad_norm": 7.661186695098877, "learning_rate": 4.2592592592592596e-07, "loss": 0.984, "step": 11540 }, { "epoch": 2.2638181105448845, "grad_norm": 11.790616989135742, "learning_rate": 4.1666666666666667e-07, "loss": 0.9724, "step": 11550 }, { "epoch": 2.2657781262250096, "grad_norm": 7.301620960235596, "learning_rate": 4.074074074074075e-07, "loss": 0.9601, "step": 11560 }, { "epoch": 2.2677381419051352, "grad_norm": 8.462876319885254, "learning_rate": 3.981481481481482e-07, "loss": 0.9752, "step": 11570 }, { "epoch": 2.269698157585261, "grad_norm": 6.7517547607421875, "learning_rate": 3.8888888888888895e-07, "loss": 0.9762, "step": 11580 }, { "epoch": 2.271658173265386, "grad_norm": 7.299182891845703, "learning_rate": 3.7962962962962966e-07, "loss": 0.972, "step": 11590 }, { "epoch": 2.2736181889455116, "grad_norm": 7.082560062408447, "learning_rate": 3.7037037037037036e-07, "loss": 0.9679, "step": 11600 }, { "epoch": 2.2755782046256368, "grad_norm": 7.469284534454346, "learning_rate": 3.611111111111111e-07, "loss": 0.9546, "step": 11610 }, { "epoch": 2.2775382203057624, "grad_norm": 7.5840535163879395, "learning_rate": 3.518518518518519e-07, "loss": 0.9761, "step": 11620 }, { "epoch": 2.279498235985888, "grad_norm": 8.262914657592773, "learning_rate": 3.4259259259259265e-07, "loss": 0.9646, "step": 11630 }, { "epoch": 2.281458251666013, "grad_norm": 8.672895431518555, "learning_rate": 3.3333333333333335e-07, "loss": 0.9732, "step": 11640 }, { "epoch": 2.2834182673461387, "grad_norm": 7.6948347091674805, "learning_rate": 3.2407407407407406e-07, "loss": 0.9726, "step": 11650 }, { "epoch": 2.2853782830262643, "grad_norm": 6.536133289337158, "learning_rate": 3.148148148148148e-07, "loss": 0.9694, "step": 11660 }, { "epoch": 2.2873382987063895, "grad_norm": 8.90619945526123, "learning_rate": 3.055555555555556e-07, "loss": 0.9793, "step": 11670 }, { "epoch": 2.289298314386515, "grad_norm": 7.702218055725098, "learning_rate": 2.9629629629629634e-07, "loss": 0.9642, "step": 11680 }, { "epoch": 2.2912583300666407, "grad_norm": 8.119073867797852, "learning_rate": 2.8703703703703705e-07, "loss": 0.9641, "step": 11690 }, { "epoch": 2.293218345746766, "grad_norm": 7.159200191497803, "learning_rate": 2.7777777777777776e-07, "loss": 0.9636, "step": 11700 }, { "epoch": 2.2951783614268915, "grad_norm": 6.998493671417236, "learning_rate": 2.685185185185185e-07, "loss": 0.9664, "step": 11710 }, { "epoch": 2.297138377107017, "grad_norm": 7.50088357925415, "learning_rate": 2.592592592592593e-07, "loss": 0.9787, "step": 11720 }, { "epoch": 2.2990983927871422, "grad_norm": 7.634678363800049, "learning_rate": 2.5000000000000004e-07, "loss": 0.9712, "step": 11730 }, { "epoch": 2.301058408467268, "grad_norm": 7.635111331939697, "learning_rate": 2.4074074074074075e-07, "loss": 0.9694, "step": 11740 }, { "epoch": 2.303018424147393, "grad_norm": 7.645463466644287, "learning_rate": 2.3148148148148148e-07, "loss": 0.9717, "step": 11750 }, { "epoch": 2.3049784398275186, "grad_norm": 6.8600568771362305, "learning_rate": 2.2222222222222224e-07, "loss": 0.9495, "step": 11760 }, { "epoch": 2.306938455507644, "grad_norm": 7.019852638244629, "learning_rate": 2.1296296296296298e-07, "loss": 0.9527, "step": 11770 }, { "epoch": 2.3088984711877694, "grad_norm": 8.227087020874023, "learning_rate": 2.0370370370370374e-07, "loss": 0.9608, "step": 11780 }, { "epoch": 2.310858486867895, "grad_norm": 7.640336036682129, "learning_rate": 1.9444444444444447e-07, "loss": 0.9684, "step": 11790 }, { "epoch": 2.3128185025480206, "grad_norm": 7.304652690887451, "learning_rate": 1.8518518518518518e-07, "loss": 0.9698, "step": 11800 }, { "epoch": 2.3147785182281457, "grad_norm": 7.173428535461426, "learning_rate": 1.7592592592592594e-07, "loss": 0.9635, "step": 11810 }, { "epoch": 2.3167385339082713, "grad_norm": 7.876226902008057, "learning_rate": 1.6666666666666668e-07, "loss": 0.9606, "step": 11820 }, { "epoch": 2.3186985495883965, "grad_norm": 6.757920265197754, "learning_rate": 1.574074074074074e-07, "loss": 0.9692, "step": 11830 }, { "epoch": 2.320658565268522, "grad_norm": 8.267070770263672, "learning_rate": 1.4814814814814817e-07, "loss": 0.9565, "step": 11840 }, { "epoch": 2.3226185809486477, "grad_norm": 7.382258892059326, "learning_rate": 1.3888888888888888e-07, "loss": 0.9672, "step": 11850 }, { "epoch": 2.324578596628773, "grad_norm": 7.452052116394043, "learning_rate": 1.2962962962962964e-07, "loss": 0.9649, "step": 11860 }, { "epoch": 2.3265386123088985, "grad_norm": 9.273903846740723, "learning_rate": 1.2037037037037037e-07, "loss": 0.9652, "step": 11870 }, { "epoch": 2.328498627989024, "grad_norm": 6.987192153930664, "learning_rate": 1.1111111111111112e-07, "loss": 0.9571, "step": 11880 }, { "epoch": 2.3304586436691492, "grad_norm": 6.799464702606201, "learning_rate": 1.0185185185185187e-07, "loss": 0.9707, "step": 11890 }, { "epoch": 2.332418659349275, "grad_norm": 7.16199254989624, "learning_rate": 9.259259259259259e-08, "loss": 0.9734, "step": 11900 }, { "epoch": 2.332418659349275, "eval_loss": 1.0144330263137817, "eval_runtime": 14.0924, "eval_samples_per_second": 46.408, "eval_steps_per_second": 5.819, "step": 11900 }, { "epoch": 2.3343786750294004, "grad_norm": 8.338574409484863, "learning_rate": 8.333333333333334e-08, "loss": 0.9714, "step": 11910 }, { "epoch": 2.3363386907095256, "grad_norm": 6.403331279754639, "learning_rate": 7.407407407407409e-08, "loss": 0.959, "step": 11920 }, { "epoch": 2.338298706389651, "grad_norm": 8.215003967285156, "learning_rate": 6.481481481481482e-08, "loss": 0.964, "step": 11930 }, { "epoch": 2.3402587220697764, "grad_norm": 7.930845737457275, "learning_rate": 5.555555555555556e-08, "loss": 0.9575, "step": 11940 }, { "epoch": 2.342218737749902, "grad_norm": 7.033510684967041, "learning_rate": 4.6296296296296295e-08, "loss": 0.9626, "step": 11950 }, { "epoch": 2.3441787534300276, "grad_norm": 7.8353800773620605, "learning_rate": 3.703703703703704e-08, "loss": 0.9673, "step": 11960 }, { "epoch": 2.3461387691101527, "grad_norm": 7.304898738861084, "learning_rate": 2.777777777777778e-08, "loss": 0.9618, "step": 11970 }, { "epoch": 2.3480987847902783, "grad_norm": 7.420264720916748, "learning_rate": 1.851851851851852e-08, "loss": 0.9544, "step": 11980 }, { "epoch": 2.350058800470404, "grad_norm": 7.217324256896973, "learning_rate": 9.25925925925926e-09, "loss": 0.9752, "step": 11990 }, { "epoch": 2.352018816150529, "grad_norm": 7.595900058746338, "learning_rate": 0.0, "loss": 0.9548, "step": 12000 }, { "epoch": 2.352018816150529, "eval_loss": 1.0147355794906616, "eval_runtime": 14.1258, "eval_samples_per_second": 46.298, "eval_steps_per_second": 5.805, "step": 12000 } ], "logging_steps": 10, "max_steps": 12000, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 1, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 6.2834751593089335e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }