|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 9.523809523809524, |
|
"eval_steps": 500, |
|
"global_step": 100, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.19047619047619047, |
|
"grad_norm": 1.0693126916885376, |
|
"learning_rate": 0.000196, |
|
"loss": 9.5582, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.38095238095238093, |
|
"grad_norm": 1.2914847135543823, |
|
"learning_rate": 0.000192, |
|
"loss": 9.2227, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.5714285714285714, |
|
"grad_norm": 1.9945707321166992, |
|
"learning_rate": 0.000188, |
|
"loss": 8.7759, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.7619047619047619, |
|
"grad_norm": 2.011664628982544, |
|
"learning_rate": 0.00018400000000000003, |
|
"loss": 8.2142, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.9523809523809523, |
|
"grad_norm": 1.3491569757461548, |
|
"learning_rate": 0.00018, |
|
"loss": 7.9009, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 1.1428571428571428, |
|
"grad_norm": 0.9671052098274231, |
|
"learning_rate": 0.00017600000000000002, |
|
"loss": 7.6909, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 1.3333333333333333, |
|
"grad_norm": 1.116225004196167, |
|
"learning_rate": 0.000172, |
|
"loss": 7.5216, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 1.5238095238095237, |
|
"grad_norm": 1.0181453227996826, |
|
"learning_rate": 0.000168, |
|
"loss": 7.4308, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 1.7142857142857144, |
|
"grad_norm": 0.9790288209915161, |
|
"learning_rate": 0.000164, |
|
"loss": 7.4444, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 1.9047619047619047, |
|
"grad_norm": 0.9529135823249817, |
|
"learning_rate": 0.00016, |
|
"loss": 7.3866, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 2.0952380952380953, |
|
"grad_norm": 0.7934174537658691, |
|
"learning_rate": 0.00015600000000000002, |
|
"loss": 7.3822, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 2.2857142857142856, |
|
"grad_norm": 0.7163369059562683, |
|
"learning_rate": 0.000152, |
|
"loss": 7.3191, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 2.4761904761904763, |
|
"grad_norm": 1.347898244857788, |
|
"learning_rate": 0.000148, |
|
"loss": 7.2142, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 2.6666666666666665, |
|
"grad_norm": 0.7439594864845276, |
|
"learning_rate": 0.000144, |
|
"loss": 7.2718, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 2.857142857142857, |
|
"grad_norm": 1.4523087739944458, |
|
"learning_rate": 0.00014, |
|
"loss": 7.2901, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 3.0476190476190474, |
|
"grad_norm": 0.7053799033164978, |
|
"learning_rate": 0.00013600000000000003, |
|
"loss": 7.2125, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 3.238095238095238, |
|
"grad_norm": 0.6156577467918396, |
|
"learning_rate": 0.000132, |
|
"loss": 7.229, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 3.4285714285714284, |
|
"grad_norm": 0.40743571519851685, |
|
"learning_rate": 0.00012800000000000002, |
|
"loss": 7.2999, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 3.619047619047619, |
|
"grad_norm": 1.0756566524505615, |
|
"learning_rate": 0.000124, |
|
"loss": 7.212, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 3.8095238095238093, |
|
"grad_norm": 0.7201813459396362, |
|
"learning_rate": 0.00012, |
|
"loss": 7.1826, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 0.5773327946662903, |
|
"learning_rate": 0.000116, |
|
"loss": 7.1456, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 4.190476190476191, |
|
"grad_norm": 0.6004664301872253, |
|
"learning_rate": 0.00011200000000000001, |
|
"loss": 7.0349, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 4.380952380952381, |
|
"grad_norm": 1.4157112836837769, |
|
"learning_rate": 0.00010800000000000001, |
|
"loss": 7.2767, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 4.571428571428571, |
|
"grad_norm": 0.852541446685791, |
|
"learning_rate": 0.00010400000000000001, |
|
"loss": 7.1745, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 4.761904761904762, |
|
"grad_norm": 0.7835370898246765, |
|
"learning_rate": 0.0001, |
|
"loss": 7.2477, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 4.9523809523809526, |
|
"grad_norm": 0.414756178855896, |
|
"learning_rate": 9.6e-05, |
|
"loss": 7.1781, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 5.142857142857143, |
|
"grad_norm": 0.7038145661354065, |
|
"learning_rate": 9.200000000000001e-05, |
|
"loss": 7.1103, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 5.333333333333333, |
|
"grad_norm": 0.8368222713470459, |
|
"learning_rate": 8.800000000000001e-05, |
|
"loss": 7.1595, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 5.523809523809524, |
|
"grad_norm": 0.6943209171295166, |
|
"learning_rate": 8.4e-05, |
|
"loss": 7.1062, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 5.714285714285714, |
|
"grad_norm": 0.4186341464519501, |
|
"learning_rate": 8e-05, |
|
"loss": 7.2348, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 5.904761904761905, |
|
"grad_norm": 1.0224595069885254, |
|
"learning_rate": 7.6e-05, |
|
"loss": 7.1154, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 6.095238095238095, |
|
"grad_norm": 0.428688645362854, |
|
"learning_rate": 7.2e-05, |
|
"loss": 7.1194, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 6.285714285714286, |
|
"grad_norm": 0.913233757019043, |
|
"learning_rate": 6.800000000000001e-05, |
|
"loss": 7.1919, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 6.476190476190476, |
|
"grad_norm": 0.5481642484664917, |
|
"learning_rate": 6.400000000000001e-05, |
|
"loss": 7.126, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 6.666666666666667, |
|
"grad_norm": 0.49522772431373596, |
|
"learning_rate": 6e-05, |
|
"loss": 7.1564, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 6.857142857142857, |
|
"grad_norm": 0.40602990984916687, |
|
"learning_rate": 5.6000000000000006e-05, |
|
"loss": 7.081, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 7.0476190476190474, |
|
"grad_norm": 0.4593268036842346, |
|
"learning_rate": 5.2000000000000004e-05, |
|
"loss": 7.0596, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 7.238095238095238, |
|
"grad_norm": 0.44626158475875854, |
|
"learning_rate": 4.8e-05, |
|
"loss": 7.0674, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 7.428571428571429, |
|
"grad_norm": 0.6573432087898254, |
|
"learning_rate": 4.4000000000000006e-05, |
|
"loss": 7.0745, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 7.619047619047619, |
|
"grad_norm": 0.3820817172527313, |
|
"learning_rate": 4e-05, |
|
"loss": 7.0785, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 7.809523809523809, |
|
"grad_norm": 0.8610634803771973, |
|
"learning_rate": 3.6e-05, |
|
"loss": 7.1974, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"grad_norm": 0.44188380241394043, |
|
"learning_rate": 3.2000000000000005e-05, |
|
"loss": 7.0847, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 8.19047619047619, |
|
"grad_norm": 0.6792606711387634, |
|
"learning_rate": 2.8000000000000003e-05, |
|
"loss": 7.1191, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 8.380952380952381, |
|
"grad_norm": 0.4903930723667145, |
|
"learning_rate": 2.4e-05, |
|
"loss": 7.1288, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 8.571428571428571, |
|
"grad_norm": 0.5853165984153748, |
|
"learning_rate": 2e-05, |
|
"loss": 7.0479, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 8.761904761904763, |
|
"grad_norm": 0.6836739182472229, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 7.0448, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 8.952380952380953, |
|
"grad_norm": 0.5737291574478149, |
|
"learning_rate": 1.2e-05, |
|
"loss": 7.0717, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 9.142857142857142, |
|
"grad_norm": 1.709892988204956, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 6.9482, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 9.333333333333334, |
|
"grad_norm": 0.61203932762146, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 7.1598, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 9.523809523809524, |
|
"grad_norm": 0.3827505111694336, |
|
"learning_rate": 0.0, |
|
"loss": 7.1112, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 9.523809523809524, |
|
"step": 100, |
|
"total_flos": 500328301455504.0, |
|
"train_loss": 7.34823148727417, |
|
"train_runtime": 426.3789, |
|
"train_samples_per_second": 3.94, |
|
"train_steps_per_second": 0.235 |
|
} |
|
], |
|
"logging_steps": 2, |
|
"max_steps": 100, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": false, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 500328301455504.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|