{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 5000, "global_step": 1125, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.008888888888888889, "grad_norm": 0.392131507396698, "learning_rate": 2.0000000000000003e-06, "loss": 2.3156, "step": 10 }, { "epoch": 0.017777777777777778, "grad_norm": 0.6833685636520386, "learning_rate": 4.000000000000001e-06, "loss": 2.5258, "step": 20 }, { "epoch": 0.02666666666666667, "grad_norm": 0.4975835978984833, "learning_rate": 6e-06, "loss": 2.5274, "step": 30 }, { "epoch": 0.035555555555555556, "grad_norm": 0.47904446721076965, "learning_rate": 8.000000000000001e-06, "loss": 2.2199, "step": 40 }, { "epoch": 0.044444444444444446, "grad_norm": 0.5792201161384583, "learning_rate": 1e-05, "loss": 2.3423, "step": 50 }, { "epoch": 0.05333333333333334, "grad_norm": 0.48581650853157043, "learning_rate": 1.2e-05, "loss": 2.3571, "step": 60 }, { "epoch": 0.06222222222222222, "grad_norm": 0.4770135283470154, "learning_rate": 1.4000000000000001e-05, "loss": 2.4415, "step": 70 }, { "epoch": 0.07111111111111111, "grad_norm": 0.6430277824401855, "learning_rate": 1.6000000000000003e-05, "loss": 2.5717, "step": 80 }, { "epoch": 0.08, "grad_norm": 0.5969924926757812, "learning_rate": 1.8e-05, "loss": 2.2712, "step": 90 }, { "epoch": 0.08888888888888889, "grad_norm": 0.5458263754844666, "learning_rate": 2e-05, "loss": 2.1937, "step": 100 }, { "epoch": 0.09777777777777778, "grad_norm": 0.49687811732292175, "learning_rate": 2.2000000000000003e-05, "loss": 2.3534, "step": 110 }, { "epoch": 0.10666666666666667, "grad_norm": 0.5264276266098022, "learning_rate": 2.4e-05, "loss": 2.3235, "step": 120 }, { "epoch": 0.11555555555555555, "grad_norm": 0.6222357749938965, "learning_rate": 2.6000000000000002e-05, "loss": 2.484, "step": 130 }, { "epoch": 0.12444444444444444, "grad_norm": 0.5594074130058289, "learning_rate": 2.8000000000000003e-05, "loss": 2.3924, "step": 140 }, { "epoch": 0.13333333333333333, "grad_norm": 0.5403823852539062, "learning_rate": 3e-05, "loss": 2.043, "step": 150 }, { "epoch": 0.14222222222222222, "grad_norm": 0.8265597820281982, "learning_rate": 3.2000000000000005e-05, "loss": 2.4271, "step": 160 }, { "epoch": 0.1511111111111111, "grad_norm": 0.4910406172275543, "learning_rate": 3.4000000000000007e-05, "loss": 2.2711, "step": 170 }, { "epoch": 0.16, "grad_norm": 0.6300191283226013, "learning_rate": 3.6e-05, "loss": 2.4088, "step": 180 }, { "epoch": 0.1688888888888889, "grad_norm": 0.9516507387161255, "learning_rate": 3.8e-05, "loss": 2.2007, "step": 190 }, { "epoch": 0.17777777777777778, "grad_norm": 0.6081224679946899, "learning_rate": 4e-05, "loss": 2.2574, "step": 200 }, { "epoch": 0.18666666666666668, "grad_norm": 0.670280396938324, "learning_rate": 4.2e-05, "loss": 2.1538, "step": 210 }, { "epoch": 0.19555555555555557, "grad_norm": 0.5967269539833069, "learning_rate": 4.4000000000000006e-05, "loss": 2.3559, "step": 220 }, { "epoch": 0.20444444444444446, "grad_norm": 0.6861779093742371, "learning_rate": 4.600000000000001e-05, "loss": 2.2649, "step": 230 }, { "epoch": 0.21333333333333335, "grad_norm": 0.7942169308662415, "learning_rate": 4.8e-05, "loss": 2.4453, "step": 240 }, { "epoch": 0.2222222222222222, "grad_norm": 0.686896562576294, "learning_rate": 5e-05, "loss": 2.272, "step": 250 }, { "epoch": 0.2311111111111111, "grad_norm": 0.6957288980484009, "learning_rate": 5.2000000000000004e-05, "loss": 2.2378, "step": 260 }, { "epoch": 0.24, "grad_norm": 0.5659416317939758, "learning_rate": 5.4000000000000005e-05, "loss": 2.2169, "step": 270 }, { "epoch": 0.24888888888888888, "grad_norm": 0.6357027292251587, "learning_rate": 5.6000000000000006e-05, "loss": 2.3081, "step": 280 }, { "epoch": 0.2577777777777778, "grad_norm": 0.561445951461792, "learning_rate": 5.8e-05, "loss": 2.3665, "step": 290 }, { "epoch": 0.26666666666666666, "grad_norm": 0.8563527464866638, "learning_rate": 6e-05, "loss": 2.2106, "step": 300 }, { "epoch": 0.27555555555555555, "grad_norm": 0.7866883873939514, "learning_rate": 6.2e-05, "loss": 2.0944, "step": 310 }, { "epoch": 0.28444444444444444, "grad_norm": 0.9338068962097168, "learning_rate": 6.400000000000001e-05, "loss": 2.2693, "step": 320 }, { "epoch": 0.29333333333333333, "grad_norm": 0.7246842384338379, "learning_rate": 6.6e-05, "loss": 2.2881, "step": 330 }, { "epoch": 0.3022222222222222, "grad_norm": 1.0052026510238647, "learning_rate": 6.800000000000001e-05, "loss": 2.3067, "step": 340 }, { "epoch": 0.3111111111111111, "grad_norm": 0.6183302998542786, "learning_rate": 7e-05, "loss": 2.3916, "step": 350 }, { "epoch": 0.32, "grad_norm": 0.7271236777305603, "learning_rate": 7.2e-05, "loss": 2.4863, "step": 360 }, { "epoch": 0.3288888888888889, "grad_norm": 0.8021159768104553, "learning_rate": 7.4e-05, "loss": 2.5189, "step": 370 }, { "epoch": 0.3377777777777778, "grad_norm": 0.670994758605957, "learning_rate": 7.6e-05, "loss": 2.299, "step": 380 }, { "epoch": 0.3466666666666667, "grad_norm": 0.6826931238174438, "learning_rate": 7.800000000000001e-05, "loss": 2.2439, "step": 390 }, { "epoch": 0.35555555555555557, "grad_norm": 0.6512673497200012, "learning_rate": 8e-05, "loss": 2.2511, "step": 400 }, { "epoch": 0.36444444444444446, "grad_norm": 0.6105847358703613, "learning_rate": 8.2e-05, "loss": 2.2597, "step": 410 }, { "epoch": 0.37333333333333335, "grad_norm": 0.7471911907196045, "learning_rate": 8.4e-05, "loss": 2.3652, "step": 420 }, { "epoch": 0.38222222222222224, "grad_norm": 0.6970444321632385, "learning_rate": 8.6e-05, "loss": 2.3259, "step": 430 }, { "epoch": 0.39111111111111113, "grad_norm": 0.6674960255622864, "learning_rate": 8.800000000000001e-05, "loss": 2.3558, "step": 440 }, { "epoch": 0.4, "grad_norm": 0.7472724318504333, "learning_rate": 9e-05, "loss": 2.2086, "step": 450 }, { "epoch": 0.4088888888888889, "grad_norm": 0.7264606356620789, "learning_rate": 9.200000000000001e-05, "loss": 2.084, "step": 460 }, { "epoch": 0.4177777777777778, "grad_norm": 0.6516128182411194, "learning_rate": 9.4e-05, "loss": 2.0929, "step": 470 }, { "epoch": 0.4266666666666667, "grad_norm": 0.8081138134002686, "learning_rate": 9.6e-05, "loss": 2.1653, "step": 480 }, { "epoch": 0.43555555555555553, "grad_norm": 0.8400042653083801, "learning_rate": 9.8e-05, "loss": 2.3939, "step": 490 }, { "epoch": 0.4444444444444444, "grad_norm": 0.4789764881134033, "learning_rate": 0.0001, "loss": 2.2763, "step": 500 }, { "epoch": 0.4533333333333333, "grad_norm": 0.7314158082008362, "learning_rate": 9.84e-05, "loss": 2.3742, "step": 510 }, { "epoch": 0.4622222222222222, "grad_norm": 0.5646970272064209, "learning_rate": 9.680000000000001e-05, "loss": 2.1478, "step": 520 }, { "epoch": 0.4711111111111111, "grad_norm": 0.4644782841205597, "learning_rate": 9.52e-05, "loss": 2.2953, "step": 530 }, { "epoch": 0.48, "grad_norm": 0.6332143545150757, "learning_rate": 9.360000000000001e-05, "loss": 2.2399, "step": 540 }, { "epoch": 0.4888888888888889, "grad_norm": 0.706084668636322, "learning_rate": 9.200000000000001e-05, "loss": 2.2324, "step": 550 }, { "epoch": 0.49777777777777776, "grad_norm": 0.6272305250167847, "learning_rate": 9.04e-05, "loss": 2.2321, "step": 560 }, { "epoch": 0.5066666666666667, "grad_norm": 0.5707433223724365, "learning_rate": 8.88e-05, "loss": 2.2092, "step": 570 }, { "epoch": 0.5155555555555555, "grad_norm": 0.660696268081665, "learning_rate": 8.72e-05, "loss": 1.9807, "step": 580 }, { "epoch": 0.5244444444444445, "grad_norm": 0.5993101000785828, "learning_rate": 8.560000000000001e-05, "loss": 1.9566, "step": 590 }, { "epoch": 0.5333333333333333, "grad_norm": 0.5725361704826355, "learning_rate": 8.4e-05, "loss": 2.1722, "step": 600 }, { "epoch": 0.5422222222222223, "grad_norm": 0.651642918586731, "learning_rate": 8.24e-05, "loss": 2.1262, "step": 610 }, { "epoch": 0.5511111111111111, "grad_norm": 0.5352093577384949, "learning_rate": 8.080000000000001e-05, "loss": 2.1979, "step": 620 }, { "epoch": 0.56, "grad_norm": 0.5744340419769287, "learning_rate": 7.920000000000001e-05, "loss": 2.0399, "step": 630 }, { "epoch": 0.5688888888888889, "grad_norm": 0.6103332042694092, "learning_rate": 7.76e-05, "loss": 2.2209, "step": 640 }, { "epoch": 0.5777777777777777, "grad_norm": 0.6434329152107239, "learning_rate": 7.6e-05, "loss": 2.379, "step": 650 }, { "epoch": 0.5866666666666667, "grad_norm": 0.5661439299583435, "learning_rate": 7.44e-05, "loss": 2.1698, "step": 660 }, { "epoch": 0.5955555555555555, "grad_norm": 0.5890876054763794, "learning_rate": 7.280000000000001e-05, "loss": 2.058, "step": 670 }, { "epoch": 0.6044444444444445, "grad_norm": 0.4725685715675354, "learning_rate": 7.12e-05, "loss": 2.0658, "step": 680 }, { "epoch": 0.6133333333333333, "grad_norm": 0.6357767581939697, "learning_rate": 6.96e-05, "loss": 2.1107, "step": 690 }, { "epoch": 0.6222222222222222, "grad_norm": 0.6726393699645996, "learning_rate": 6.800000000000001e-05, "loss": 2.1904, "step": 700 }, { "epoch": 0.6311111111111111, "grad_norm": 0.5009137392044067, "learning_rate": 6.64e-05, "loss": 2.0765, "step": 710 }, { "epoch": 0.64, "grad_norm": 0.7313541173934937, "learning_rate": 6.48e-05, "loss": 2.3566, "step": 720 }, { "epoch": 0.6488888888888888, "grad_norm": 0.7757258415222168, "learning_rate": 6.32e-05, "loss": 2.2077, "step": 730 }, { "epoch": 0.6577777777777778, "grad_norm": 0.5176954865455627, "learning_rate": 6.16e-05, "loss": 2.1287, "step": 740 }, { "epoch": 0.6666666666666666, "grad_norm": 0.6251640915870667, "learning_rate": 6e-05, "loss": 2.1726, "step": 750 }, { "epoch": 0.6755555555555556, "grad_norm": 0.5190748572349548, "learning_rate": 5.8399999999999997e-05, "loss": 2.1528, "step": 760 }, { "epoch": 0.6844444444444444, "grad_norm": 0.6299232840538025, "learning_rate": 5.68e-05, "loss": 1.9734, "step": 770 }, { "epoch": 0.6933333333333334, "grad_norm": 0.6707553863525391, "learning_rate": 5.520000000000001e-05, "loss": 1.9959, "step": 780 }, { "epoch": 0.7022222222222222, "grad_norm": 0.5231025815010071, "learning_rate": 5.360000000000001e-05, "loss": 2.0261, "step": 790 }, { "epoch": 0.7111111111111111, "grad_norm": 0.489175021648407, "learning_rate": 5.2000000000000004e-05, "loss": 2.1987, "step": 800 }, { "epoch": 0.72, "grad_norm": 0.6425113677978516, "learning_rate": 5.0400000000000005e-05, "loss": 2.0759, "step": 810 }, { "epoch": 0.7288888888888889, "grad_norm": 0.6148718595504761, "learning_rate": 4.88e-05, "loss": 2.158, "step": 820 }, { "epoch": 0.7377777777777778, "grad_norm": 0.7482603192329407, "learning_rate": 4.72e-05, "loss": 1.9579, "step": 830 }, { "epoch": 0.7466666666666667, "grad_norm": 0.5650503039360046, "learning_rate": 4.5600000000000004e-05, "loss": 2.1253, "step": 840 }, { "epoch": 0.7555555555555555, "grad_norm": 0.6965321898460388, "learning_rate": 4.4000000000000006e-05, "loss": 2.1034, "step": 850 }, { "epoch": 0.7644444444444445, "grad_norm": 0.5848603844642639, "learning_rate": 4.24e-05, "loss": 2.0139, "step": 860 }, { "epoch": 0.7733333333333333, "grad_norm": 0.566116988658905, "learning_rate": 4.08e-05, "loss": 2.0551, "step": 870 }, { "epoch": 0.7822222222222223, "grad_norm": 0.5044224858283997, "learning_rate": 3.9200000000000004e-05, "loss": 1.9789, "step": 880 }, { "epoch": 0.7911111111111111, "grad_norm": 0.566875159740448, "learning_rate": 3.76e-05, "loss": 2.3283, "step": 890 }, { "epoch": 0.8, "grad_norm": 0.5149425268173218, "learning_rate": 3.6e-05, "loss": 1.9651, "step": 900 }, { "epoch": 0.8088888888888889, "grad_norm": 0.5834816694259644, "learning_rate": 3.4399999999999996e-05, "loss": 1.9825, "step": 910 }, { "epoch": 0.8177777777777778, "grad_norm": 0.7403817772865295, "learning_rate": 3.2800000000000004e-05, "loss": 2.0542, "step": 920 }, { "epoch": 0.8266666666666667, "grad_norm": 0.6044632792472839, "learning_rate": 3.12e-05, "loss": 2.2774, "step": 930 }, { "epoch": 0.8355555555555556, "grad_norm": 0.6258851885795593, "learning_rate": 2.96e-05, "loss": 2.1449, "step": 940 }, { "epoch": 0.8444444444444444, "grad_norm": 0.6421411037445068, "learning_rate": 2.8000000000000003e-05, "loss": 2.0822, "step": 950 }, { "epoch": 0.8533333333333334, "grad_norm": 0.34951749444007874, "learning_rate": 2.64e-05, "loss": 1.9521, "step": 960 }, { "epoch": 0.8622222222222222, "grad_norm": 0.5875269770622253, "learning_rate": 2.48e-05, "loss": 2.0813, "step": 970 }, { "epoch": 0.8711111111111111, "grad_norm": 0.5003840327262878, "learning_rate": 2.32e-05, "loss": 1.9155, "step": 980 }, { "epoch": 0.88, "grad_norm": 0.4974331259727478, "learning_rate": 2.16e-05, "loss": 2.0593, "step": 990 }, { "epoch": 0.8888888888888888, "grad_norm": 0.505962610244751, "learning_rate": 2e-05, "loss": 2.0306, "step": 1000 }, { "epoch": 0.8977777777777778, "grad_norm": 0.5298508405685425, "learning_rate": 1.84e-05, "loss": 2.136, "step": 1010 }, { "epoch": 0.9066666666666666, "grad_norm": 0.4652698338031769, "learning_rate": 1.6800000000000002e-05, "loss": 1.9021, "step": 1020 }, { "epoch": 0.9155555555555556, "grad_norm": 0.5339078307151794, "learning_rate": 1.52e-05, "loss": 2.0262, "step": 1030 }, { "epoch": 0.9244444444444444, "grad_norm": 0.5623987317085266, "learning_rate": 1.3600000000000002e-05, "loss": 1.8597, "step": 1040 }, { "epoch": 0.9333333333333333, "grad_norm": 0.689896285533905, "learning_rate": 1.2e-05, "loss": 2.1377, "step": 1050 }, { "epoch": 0.9422222222222222, "grad_norm": 0.460510790348053, "learning_rate": 1.04e-05, "loss": 1.889, "step": 1060 }, { "epoch": 0.9511111111111111, "grad_norm": 0.5239367485046387, "learning_rate": 8.8e-06, "loss": 2.1753, "step": 1070 }, { "epoch": 0.96, "grad_norm": 0.5576533079147339, "learning_rate": 7.2e-06, "loss": 2.1375, "step": 1080 }, { "epoch": 0.9688888888888889, "grad_norm": 0.5585991144180298, "learning_rate": 5.600000000000001e-06, "loss": 1.9434, "step": 1090 }, { "epoch": 0.9777777777777777, "grad_norm": 0.5961203575134277, "learning_rate": 4.000000000000001e-06, "loss": 2.0194, "step": 1100 }, { "epoch": 0.9866666666666667, "grad_norm": 0.6022922396659851, "learning_rate": 2.4000000000000003e-06, "loss": 1.9679, "step": 1110 }, { "epoch": 0.9955555555555555, "grad_norm": 0.4506663382053375, "learning_rate": 8.000000000000001e-07, "loss": 2.1243, "step": 1120 } ], "logging_steps": 10, "max_steps": 1125, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 5000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.4276096434176e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }