{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 504, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.05952380952380952, "grad_norm": 0.0, "learning_rate": 1.9607843137254903e-05, "loss": 1.9698, "step": 10 }, { "epoch": 0.11904761904761904, "grad_norm": 0.0, "learning_rate": 3.9215686274509805e-05, "loss": 1.903, "step": 20 }, { "epoch": 0.17857142857142858, "grad_norm": 0.0, "learning_rate": 5.882352941176471e-05, "loss": 1.8528, "step": 30 }, { "epoch": 0.23809523809523808, "grad_norm": 0.0, "learning_rate": 7.843137254901961e-05, "loss": 1.865, "step": 40 }, { "epoch": 0.2976190476190476, "grad_norm": 0.0, "learning_rate": 9.80392156862745e-05, "loss": 1.9221, "step": 50 }, { "epoch": 0.35714285714285715, "grad_norm": 0.0, "learning_rate": 9.990263847374976e-05, "loss": 1.8802, "step": 60 }, { "epoch": 0.4166666666666667, "grad_norm": 0.0, "learning_rate": 9.956656688041808e-05, "loss": 1.9587, "step": 70 }, { "epoch": 0.47619047619047616, "grad_norm": 0.0, "learning_rate": 9.899219854168429e-05, "loss": 1.9369, "step": 80 }, { "epoch": 0.5357142857142857, "grad_norm": 0.0, "learning_rate": 9.818229479678158e-05, "loss": 1.8352, "step": 90 }, { "epoch": 0.5952380952380952, "grad_norm": 0.0, "learning_rate": 9.714074934742556e-05, "loss": 1.8348, "step": 100 }, { "epoch": 0.6547619047619048, "grad_norm": 0.0, "learning_rate": 9.587256953841317e-05, "loss": 1.9108, "step": 110 }, { "epoch": 0.7142857142857143, "grad_norm": 0.0, "learning_rate": 9.438385228425938e-05, "loss": 1.9254, "step": 120 }, { "epoch": 0.7738095238095238, "grad_norm": 0.0, "learning_rate": 9.268175475760734e-05, "loss": 1.8523, "step": 130 }, { "epoch": 0.8333333333333334, "grad_norm": 0.0, "learning_rate": 9.077445998033015e-05, "loss": 1.8635, "step": 140 }, { "epoch": 0.8928571428571429, "grad_norm": 0.0, "learning_rate": 8.86711374827494e-05, "loss": 1.891, "step": 150 }, { "epoch": 0.9523809523809523, "grad_norm": 0.0, "learning_rate": 8.638189922010529e-05, "loss": 1.8867, "step": 160 }, { "epoch": 1.0119047619047619, "grad_norm": 0.0, "learning_rate": 8.391775095821482e-05, "loss": 1.885, "step": 170 }, { "epoch": 1.0714285714285714, "grad_norm": 0.0, "learning_rate": 8.129053936203687e-05, "loss": 1.6391, "step": 180 }, { "epoch": 1.130952380952381, "grad_norm": 0.0, "learning_rate": 7.851289504152201e-05, "loss": 1.6464, "step": 190 }, { "epoch": 1.1904761904761905, "grad_norm": 0.0, "learning_rate": 7.559817182855977e-05, "loss": 1.6074, "step": 200 }, { "epoch": 1.25, "grad_norm": 0.0, "learning_rate": 7.256038257695687e-05, "loss": 1.5893, "step": 210 }, { "epoch": 1.3095238095238095, "grad_norm": 0.0, "learning_rate": 6.941413179409468e-05, "loss": 1.6661, "step": 220 }, { "epoch": 1.369047619047619, "grad_norm": 0.0, "learning_rate": 6.61745454281468e-05, "loss": 1.5408, "step": 230 }, { "epoch": 1.4285714285714286, "grad_norm": 0.0, "learning_rate": 6.28571981484123e-05, "loss": 1.6856, "step": 240 }, { "epoch": 1.4880952380952381, "grad_norm": 0.0, "learning_rate": 5.94780384683728e-05, "loss": 1.5643, "step": 250 }, { "epoch": 1.5476190476190477, "grad_norm": 0.0, "learning_rate": 5.605331207145219e-05, "loss": 1.7048, "step": 260 }, { "epoch": 1.6071428571428572, "grad_norm": 0.0, "learning_rate": 5.2599483708099016e-05, "loss": 1.6339, "step": 270 }, { "epoch": 1.6666666666666665, "grad_norm": 0.0, "learning_rate": 4.913315803968012e-05, "loss": 1.7719, "step": 280 }, { "epoch": 1.7261904761904763, "grad_norm": 0.0, "learning_rate": 4.567099980973684e-05, "loss": 1.6909, "step": 290 }, { "epoch": 1.7857142857142856, "grad_norm": 0.0, "learning_rate": 4.2229653726389765e-05, "loss": 1.4923, "step": 300 }, { "epoch": 1.8452380952380953, "grad_norm": 0.0, "learning_rate": 3.8825664441065734e-05, "loss": 1.6585, "step": 310 }, { "epoch": 1.9047619047619047, "grad_norm": 0.0, "learning_rate": 3.547539700825874e-05, "loss": 1.6052, "step": 320 }, { "epoch": 1.9642857142857144, "grad_norm": 0.0, "learning_rate": 3.219495820872265e-05, "loss": 1.6084, "step": 330 }, { "epoch": 2.0238095238095237, "grad_norm": 0.0, "learning_rate": 2.9000119114343582e-05, "loss": 1.495, "step": 340 }, { "epoch": 2.0833333333333335, "grad_norm": 0.0, "learning_rate": 2.5906239266969805e-05, "loss": 1.379, "step": 350 }, { "epoch": 2.142857142857143, "grad_norm": 0.0, "learning_rate": 2.2928192835717644e-05, "loss": 1.4044, "step": 360 }, { "epoch": 2.2023809523809526, "grad_norm": 0.0, "learning_rate": 2.008029710776041e-05, "loss": 1.3729, "step": 370 }, { "epoch": 2.261904761904762, "grad_norm": 0.0, "learning_rate": 1.7376243656388923e-05, "loss": 1.4013, "step": 380 }, { "epoch": 2.3214285714285716, "grad_norm": 0.0, "learning_rate": 1.4829032517260489e-05, "loss": 1.3822, "step": 390 }, { "epoch": 2.380952380952381, "grad_norm": 0.0, "learning_rate": 1.245090968929148e-05, "loss": 1.3588, "step": 400 }, { "epoch": 2.4404761904761907, "grad_norm": 0.0, "learning_rate": 1.0253308260664902e-05, "loss": 1.3299, "step": 410 }, { "epoch": 2.5, "grad_norm": 0.0, "learning_rate": 8.246793442995954e-06, "loss": 1.3465, "step": 420 }, { "epoch": 2.5595238095238093, "grad_norm": 0.0, "learning_rate": 6.4410117779103e-06, "loss": 1.2845, "step": 430 }, { "epoch": 2.619047619047619, "grad_norm": 0.0, "learning_rate": 4.844644760229544e-06, "loss": 1.3998, "step": 440 }, { "epoch": 2.678571428571429, "grad_norm": 0.0, "learning_rate": 3.465367100725908e-06, "loss": 1.4586, "step": 450 }, { "epoch": 2.738095238095238, "grad_norm": 0.0, "learning_rate": 2.3098098291025273e-06, "loss": 1.3051, "step": 460 }, { "epoch": 2.7976190476190474, "grad_norm": 0.0, "learning_rate": 1.3835284145856275e-06, "loss": 1.3665, "step": 470 }, { "epoch": 2.857142857142857, "grad_norm": 0.0, "learning_rate": 6.909760573925561e-07, "loss": 1.4534, "step": 480 }, { "epoch": 2.9166666666666665, "grad_norm": 0.0, "learning_rate": 2.3548227947988167e-07, "loss": 1.419, "step": 490 }, { "epoch": 2.9761904761904763, "grad_norm": 0.0, "learning_rate": 1.9236917498782758e-08, "loss": 1.3451, "step": 500 } ], "logging_steps": 10, "max_steps": 504, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 1660, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.647913924224614e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }