|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 541.0, |
|
"eval_steps": 500, |
|
"global_step": 9738, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 5.56, |
|
"grad_norm": 0.40853407979011536, |
|
"learning_rate": 4.998998767795805e-05, |
|
"loss": 2.5454, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 11.11, |
|
"grad_norm": 0.27508941292762756, |
|
"learning_rate": 4.995538747800402e-05, |
|
"loss": 0.3265, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 16.67, |
|
"grad_norm": 0.19116361439228058, |
|
"learning_rate": 4.989610999768349e-05, |
|
"loss": 0.1042, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 22.22, |
|
"grad_norm": 0.10925190895795822, |
|
"learning_rate": 4.9812213853878376e-05, |
|
"loss": 0.0633, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 27.78, |
|
"grad_norm": 0.08680253475904465, |
|
"learning_rate": 4.970378200777949e-05, |
|
"loss": 0.0377, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 33.33, |
|
"grad_norm": 0.029545623809099197, |
|
"learning_rate": 4.957092168284987e-05, |
|
"loss": 0.0181, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 38.89, |
|
"grad_norm": 0.030413279309868813, |
|
"learning_rate": 4.941376425879624e-05, |
|
"loss": 0.008, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 44.44, |
|
"grad_norm": 0.05130019411444664, |
|
"learning_rate": 4.923246514165339e-05, |
|
"loss": 0.0037, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"grad_norm": 0.005339763592928648, |
|
"learning_rate": 4.902720361011007e-05, |
|
"loss": 0.0021, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 55.56, |
|
"grad_norm": 0.023293694481253624, |
|
"learning_rate": 4.8798182638228166e-05, |
|
"loss": 0.0014, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 61.11, |
|
"grad_norm": 0.014095759950578213, |
|
"learning_rate": 4.8545628694730624e-05, |
|
"loss": 0.001, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 66.67, |
|
"grad_norm": 0.00977824255824089, |
|
"learning_rate": 4.826979151905655e-05, |
|
"loss": 0.0007, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 72.22, |
|
"grad_norm": 0.006551814265549183, |
|
"learning_rate": 4.797094387440491e-05, |
|
"loss": 0.0006, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 77.78, |
|
"grad_norm": 0.008941611275076866, |
|
"learning_rate": 4.7649381278011e-05, |
|
"loss": 0.0005, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 83.33, |
|
"grad_norm": 0.0047577545046806335, |
|
"learning_rate": 4.73054217089226e-05, |
|
"loss": 0.0004, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 88.89, |
|
"grad_norm": 0.005813396070152521, |
|
"learning_rate": 4.693940529356444e-05, |
|
"loss": 0.0003, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 94.44, |
|
"grad_norm": 0.006802932824939489, |
|
"learning_rate": 4.655169396940229e-05, |
|
"loss": 0.0003, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"grad_norm": 0.0018754027551040053, |
|
"learning_rate": 4.6142671127038905e-05, |
|
"loss": 0.0002, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 105.56, |
|
"grad_norm": 0.0058812773786485195, |
|
"learning_rate": 4.571274123109606e-05, |
|
"loss": 0.0002, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 111.11, |
|
"grad_norm": 0.003945046104490757, |
|
"learning_rate": 4.52623294202573e-05, |
|
"loss": 0.0002, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 116.67, |
|
"grad_norm": 0.0026905699633061886, |
|
"learning_rate": 4.479188108686714e-05, |
|
"loss": 0.0002, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 122.22, |
|
"grad_norm": 0.00241717416793108, |
|
"learning_rate": 4.4301861436502156e-05, |
|
"loss": 0.0001, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 127.78, |
|
"grad_norm": 0.002276218729093671, |
|
"learning_rate": 4.379275502794983e-05, |
|
"loss": 0.0001, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 133.33, |
|
"grad_norm": 0.0011824576649814844, |
|
"learning_rate": 4.326506529404972e-05, |
|
"loss": 0.0001, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 138.89, |
|
"grad_norm": 0.001206545508466661, |
|
"learning_rate": 4.271931404387096e-05, |
|
"loss": 0.0001, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 144.44, |
|
"grad_norm": 0.0025704980362206697, |
|
"learning_rate": 4.215604094671835e-05, |
|
"loss": 0.0001, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 150.0, |
|
"grad_norm": 0.0012343807611614466, |
|
"learning_rate": 4.157580299847717e-05, |
|
"loss": 0.0001, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 155.56, |
|
"grad_norm": 0.0010726838372647762, |
|
"learning_rate": 4.0979173970824626e-05, |
|
"loss": 0.0001, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 161.11, |
|
"grad_norm": 0.0008975835517048836, |
|
"learning_rate": 4.036674384385231e-05, |
|
"loss": 0.0001, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 166.67, |
|
"grad_norm": 0.0015915404073894024, |
|
"learning_rate": 3.973911822266099e-05, |
|
"loss": 0.0001, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 172.22, |
|
"grad_norm": 0.0013330359943211079, |
|
"learning_rate": 3.909691773850445e-05, |
|
"loss": 0.0001, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 177.78, |
|
"grad_norm": 0.0007300216238945723, |
|
"learning_rate": 3.844077743507468e-05, |
|
"loss": 0.0001, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 183.33, |
|
"grad_norm": 0.0006974161951802671, |
|
"learning_rate": 3.777134614053522e-05, |
|
"loss": 0.0001, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 188.89, |
|
"grad_norm": 0.0016547333216294646, |
|
"learning_rate": 3.7089285825923615e-05, |
|
"loss": 0.0, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 194.44, |
|
"grad_norm": 0.0006157997995615005, |
|
"learning_rate": 3.639527095055753e-05, |
|
"loss": 0.0, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 200.0, |
|
"grad_norm": 0.000979723292402923, |
|
"learning_rate": 3.568998779509173e-05, |
|
"loss": 0.0, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 205.56, |
|
"grad_norm": 0.0008198455907404423, |
|
"learning_rate": 3.497413378288541e-05, |
|
"loss": 0.0, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 211.11, |
|
"grad_norm": 0.0012410281924530864, |
|
"learning_rate": 3.424841679035109e-05, |
|
"loss": 0.0, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 216.67, |
|
"grad_norm": 0.0007560970261693001, |
|
"learning_rate": 3.351355444696684e-05, |
|
"loss": 0.0, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 222.22, |
|
"grad_norm": 0.0005174472462385893, |
|
"learning_rate": 3.277027342564428e-05, |
|
"loss": 0.0, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 227.78, |
|
"grad_norm": 0.00041007634717971087, |
|
"learning_rate": 3.201930872415374e-05, |
|
"loss": 0.0, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 233.33, |
|
"grad_norm": 0.0008007108117453754, |
|
"learning_rate": 3.126140293831746e-05, |
|
"loss": 0.0, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 238.89, |
|
"grad_norm": 0.00044247345067560673, |
|
"learning_rate": 3.0497305527689445e-05, |
|
"loss": 0.0, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 244.44, |
|
"grad_norm": 0.0001442178909201175, |
|
"learning_rate": 2.972777207444791e-05, |
|
"loss": 0.0, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 250.0, |
|
"grad_norm": 0.0005981961148791015, |
|
"learning_rate": 2.8953563536233525e-05, |
|
"loss": 0.0, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 255.56, |
|
"grad_norm": 0.00022320376592688262, |
|
"learning_rate": 2.8175445493671972e-05, |
|
"loss": 0.0, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 261.11, |
|
"grad_norm": 0.0004137264913879335, |
|
"learning_rate": 2.7394187393325106e-05, |
|
"loss": 0.0, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 266.67, |
|
"grad_norm": 0.0003305468999315053, |
|
"learning_rate": 2.6610561786819204e-05, |
|
"loss": 0.0, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 272.22, |
|
"grad_norm": 0.00018206202366854995, |
|
"learning_rate": 2.5825343566902837e-05, |
|
"loss": 0.0, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 277.78, |
|
"grad_norm": 0.00040663284016773105, |
|
"learning_rate": 2.5039309201189614e-05, |
|
"loss": 0.0, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 283.33, |
|
"grad_norm": 0.0004440485790837556, |
|
"learning_rate": 2.4253235964343676e-05, |
|
"loss": 0.0, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 288.89, |
|
"grad_norm": 0.0003763148852158338, |
|
"learning_rate": 2.34679011694671e-05, |
|
"loss": 0.0, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 294.44, |
|
"grad_norm": 0.0002928711473941803, |
|
"learning_rate": 2.2684081399449327e-05, |
|
"loss": 0.0, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 300.0, |
|
"grad_norm": 0.0004030682030133903, |
|
"learning_rate": 2.1902551739038624e-05, |
|
"loss": 0.0, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 305.56, |
|
"grad_norm": 0.0004344022599980235, |
|
"learning_rate": 2.1124085008395054e-05, |
|
"loss": 0.0, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 311.11, |
|
"grad_norm": 0.00018473710224498063, |
|
"learning_rate": 2.03494509988827e-05, |
|
"loss": 0.0, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 316.67, |
|
"grad_norm": 0.00044333594269119203, |
|
"learning_rate": 1.9579415711857018e-05, |
|
"loss": 0.0, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 322.22, |
|
"grad_norm": 0.00039796039345674217, |
|
"learning_rate": 1.881474060119994e-05, |
|
"loss": 0.0, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 327.78, |
|
"grad_norm": 0.0002483979915268719, |
|
"learning_rate": 1.8056181820351738e-05, |
|
"loss": 0.0, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 333.33, |
|
"grad_norm": 0.00021489571372512728, |
|
"learning_rate": 1.7304489474584307e-05, |
|
"loss": 0.0, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 338.89, |
|
"grad_norm": 0.00023669018992222846, |
|
"learning_rate": 1.656040687925519e-05, |
|
"loss": 0.0, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 344.44, |
|
"grad_norm": 0.00023518071975558996, |
|
"learning_rate": 1.582466982477587e-05, |
|
"loss": 0.0, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 350.0, |
|
"grad_norm": 0.00023862645321059972, |
|
"learning_rate": 1.509800584902108e-05, |
|
"loss": 0.0, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 355.56, |
|
"grad_norm": 0.00020295938884373754, |
|
"learning_rate": 1.4381133517898804e-05, |
|
"loss": 0.0, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 361.11, |
|
"grad_norm": 0.0003517361474223435, |
|
"learning_rate": 1.3674761714792153e-05, |
|
"loss": 0.0, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 366.67, |
|
"grad_norm": 0.00022433781123254448, |
|
"learning_rate": 1.297958893957588e-05, |
|
"loss": 0.0, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 372.22, |
|
"grad_norm": 0.0002291495620738715, |
|
"learning_rate": 1.229630261790077e-05, |
|
"loss": 0.0, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 377.78, |
|
"grad_norm": 0.00021542608737945557, |
|
"learning_rate": 1.1625578421428714e-05, |
|
"loss": 0.0, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 383.33, |
|
"grad_norm": 0.00030446049640886486, |
|
"learning_rate": 1.0968079599690872e-05, |
|
"loss": 0.0, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 388.89, |
|
"grad_norm": 0.000319374434184283, |
|
"learning_rate": 1.0324456324229537e-05, |
|
"loss": 0.0, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 394.44, |
|
"grad_norm": 0.0002203370677307248, |
|
"learning_rate": 9.695345045672166e-06, |
|
"loss": 0.0, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 400.0, |
|
"grad_norm": 0.00015890812210272998, |
|
"learning_rate": 9.081367864373488e-06, |
|
"loss": 0.0, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 405.56, |
|
"grad_norm": 0.000279374944511801, |
|
"learning_rate": 8.483131915247968e-06, |
|
"loss": 0.0, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 411.11, |
|
"grad_norm": 0.00011563805310288444, |
|
"learning_rate": 7.901228767400859e-06, |
|
"loss": 0.0, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 416.67, |
|
"grad_norm": 0.00013281428255140781, |
|
"learning_rate": 7.336233839151693e-06, |
|
"loss": 0.0, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 422.22, |
|
"grad_norm": 0.0002364695246797055, |
|
"learning_rate": 6.788705829028483e-06, |
|
"loss": 0.0, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 427.78, |
|
"grad_norm": 0.00013619402307085693, |
|
"learning_rate": 6.259186163295438e-06, |
|
"loss": 0.0, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 433.33, |
|
"grad_norm": 0.0002181720337830484, |
|
"learning_rate": 5.748198460560475e-06, |
|
"loss": 0.0, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 438.89, |
|
"grad_norm": 0.0001409970282111317, |
|
"learning_rate": 5.256248013991857e-06, |
|
"loss": 0.0, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 444.44, |
|
"grad_norm": 8.899461681721732e-05, |
|
"learning_rate": 4.78382129165613e-06, |
|
"loss": 0.0, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 450.0, |
|
"grad_norm": 9.178288746625185e-05, |
|
"learning_rate": 4.331385455471346e-06, |
|
"loss": 0.0, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 455.56, |
|
"grad_norm": 0.00016385990602429956, |
|
"learning_rate": 3.8993878992512415e-06, |
|
"loss": 0.0, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 461.11, |
|
"grad_norm": 0.00012868938210885972, |
|
"learning_rate": 3.488255806297311e-06, |
|
"loss": 0.0, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 466.67, |
|
"grad_norm": 0.00018513025133870542, |
|
"learning_rate": 3.09839572697605e-06, |
|
"loss": 0.0, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 472.22, |
|
"grad_norm": 0.00016808818327262998, |
|
"learning_rate": 2.7301931766992917e-06, |
|
"loss": 0.0, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 477.78, |
|
"grad_norm": 0.0001630824408493936, |
|
"learning_rate": 2.384012254705048e-06, |
|
"loss": 0.0, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 483.33, |
|
"grad_norm": 0.00018219888443127275, |
|
"learning_rate": 2.0601952840158366e-06, |
|
"loss": 0.0, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 488.89, |
|
"grad_norm": 0.00011875651398440823, |
|
"learning_rate": 1.75906247293057e-06, |
|
"loss": 0.0, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 494.44, |
|
"grad_norm": 0.00015640753554180264, |
|
"learning_rate": 1.4809115983847266e-06, |
|
"loss": 0.0, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 500.0, |
|
"grad_norm": 0.00010140336962649599, |
|
"learning_rate": 1.226017711491867e-06, |
|
"loss": 0.0, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 505.56, |
|
"grad_norm": 0.00018589019600767642, |
|
"learning_rate": 9.946328655577624e-07, |
|
"loss": 0.0, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 511.11, |
|
"grad_norm": 0.00026709603844210505, |
|
"learning_rate": 7.869858668360042e-07, |
|
"loss": 0.0, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 516.67, |
|
"grad_norm": 0.00017575189121998847, |
|
"learning_rate": 6.032820482716001e-07, |
|
"loss": 0.0, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 522.22, |
|
"grad_norm": 0.0001377611479256302, |
|
"learning_rate": 4.437030664562969e-07, |
|
"loss": 0.0, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 527.78, |
|
"grad_norm": 9.088205842999741e-05, |
|
"learning_rate": 3.084067219964182e-07, |
|
"loss": 0.0, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 533.33, |
|
"grad_norm": 0.00015840640116948634, |
|
"learning_rate": 1.975268034707878e-07, |
|
"loss": 0.0, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 538.89, |
|
"grad_norm": 0.00013717034016735852, |
|
"learning_rate": 1.1117295513313475e-07, |
|
"loss": 0.0, |
|
"step": 9700 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 10000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 556, |
|
"save_steps": 500, |
|
"total_flos": 2.673487599113011e+16, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|