|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 249, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 5.000676195286779, |
|
"learning_rate": 2.5e-05, |
|
"loss": 1.4173, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 5.533334563322662, |
|
"learning_rate": 5e-05, |
|
"loss": 1.4686, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 2.1215141861819773, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 1.1569, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 2.1829474332972323, |
|
"learning_rate": 0.0001, |
|
"loss": 1.1048, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 2.0028696101556926, |
|
"learning_rate": 0.000125, |
|
"loss": 1.0818, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.5248690364202462, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 1.0337, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.1217144084706645, |
|
"learning_rate": 0.000175, |
|
"loss": 0.9547, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.4464744842578992, |
|
"learning_rate": 0.0002, |
|
"loss": 0.9475, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1695.2119602484477, |
|
"learning_rate": 0.00019999150370633988, |
|
"loss": 2.8217, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.2498837065378345, |
|
"learning_rate": 0.00019996601626909964, |
|
"loss": 0.9227, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.0785140860234272, |
|
"learning_rate": 0.00019992354201925428, |
|
"loss": 0.9174, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.8766759988692014, |
|
"learning_rate": 0.0001998640881742778, |
|
"loss": 0.9297, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 7.243650858657556, |
|
"learning_rate": 0.00019978766483691676, |
|
"loss": 0.892, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.538431288719503, |
|
"learning_rate": 0.0001996942849934735, |
|
"loss": 0.8863, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.370383009713962, |
|
"learning_rate": 0.00019958396451159936, |
|
"loss": 0.8137, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.0117664660143928, |
|
"learning_rate": 0.0001994567221375987, |
|
"loss": 0.8547, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.1793686627309208, |
|
"learning_rate": 0.00019931257949324288, |
|
"loss": 0.889, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 6.742266804538591, |
|
"learning_rate": 0.00019915156107209675, |
|
"loss": 0.8061, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.0901022322430638, |
|
"learning_rate": 0.000198973694235356, |
|
"loss": 0.8228, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 2.50478337872783, |
|
"learning_rate": 0.00019877900920719827, |
|
"loss": 0.8285, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.1163796098877177, |
|
"learning_rate": 0.00019856753906964686, |
|
"loss": 0.7634, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 1.2717680968650331, |
|
"learning_rate": 0.0001983393197569497, |
|
"loss": 0.8451, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 1.1035753559158878, |
|
"learning_rate": 0.00019809439004947268, |
|
"loss": 0.8134, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.0228605041664325, |
|
"learning_rate": 0.00019783279156711022, |
|
"loss": 0.7859, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.0963594515481832, |
|
"learning_rate": 0.0001975545687622129, |
|
"loss": 0.7781, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.2366762317778497, |
|
"learning_rate": 0.00019725976891203376, |
|
"loss": 0.8779, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9640376193091488, |
|
"learning_rate": 0.00019694844211069477, |
|
"loss": 0.7786, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9093723580811977, |
|
"learning_rate": 0.00019662064126067452, |
|
"loss": 0.7569, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.0668396721156181, |
|
"learning_rate": 0.00019627642206381863, |
|
"loss": 0.7912, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.9305159138901695, |
|
"learning_rate": 0.00019591584301187478, |
|
"loss": 0.7333, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.8786594903907613, |
|
"learning_rate": 0.00019553896537655318, |
|
"loss": 0.7597, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 1.003046456072257, |
|
"learning_rate": 0.0001951458531991151, |
|
"loss": 0.7586, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.8294893505023185, |
|
"learning_rate": 0.00019473657327949054, |
|
"loss": 0.7481, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.9425474789456502, |
|
"learning_rate": 0.00019431119516492726, |
|
"loss": 0.7381, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.7607927866058226, |
|
"learning_rate": 0.00019386979113817282, |
|
"loss": 0.7598, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.8983977527007998, |
|
"learning_rate": 0.0001934124362051919, |
|
"loss": 0.7781, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.7601960470127274, |
|
"learning_rate": 0.00019293920808242083, |
|
"loss": 0.7704, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.8568120633321497, |
|
"learning_rate": 0.0001924501871835616, |
|
"loss": 0.7885, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.8412758002249293, |
|
"learning_rate": 0.00019194545660591752, |
|
"loss": 0.6997, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1.3840877739036568, |
|
"learning_rate": 0.00019142510211627264, |
|
"loss": 0.7475, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.8713962244801455, |
|
"learning_rate": 0.000190889212136318, |
|
"loss": 0.746, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.8405881719204712, |
|
"learning_rate": 0.00019033787772762645, |
|
"loss": 0.7537, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.8881177433012806, |
|
"learning_rate": 0.00018977119257617878, |
|
"loss": 0.772, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.8327207244772546, |
|
"learning_rate": 0.00018918925297644416, |
|
"loss": 0.7671, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.7172850301202007, |
|
"learning_rate": 0.00018859215781501725, |
|
"loss": 0.7013, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.8520019963524466, |
|
"learning_rate": 0.0001879800085538147, |
|
"loss": 0.6775, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.7905901994214263, |
|
"learning_rate": 0.0001873529092128343, |
|
"loss": 0.7543, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.885042022048047, |
|
"learning_rate": 0.00018671096635247914, |
|
"loss": 0.7772, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.7923851096939509, |
|
"learning_rate": 0.00018605428905545032, |
|
"loss": 0.7647, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.7639779196406391, |
|
"learning_rate": 0.0001853829889082109, |
|
"loss": 0.73, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.7201688215295416, |
|
"learning_rate": 0.00018469717998202462, |
|
"loss": 0.7534, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.7092609652695424, |
|
"learning_rate": 0.00018399697881357212, |
|
"loss": 0.7402, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.7862922090916423, |
|
"learning_rate": 0.00018328250438514836, |
|
"loss": 0.6988, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.75620823509439, |
|
"learning_rate": 0.00018255387810444448, |
|
"loss": 0.7592, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.7979316767550451, |
|
"learning_rate": 0.0001818112237839174, |
|
"loss": 0.7174, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.7690661744799511, |
|
"learning_rate": 0.00018105466761975109, |
|
"loss": 0.7082, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.7543971879671839, |
|
"learning_rate": 0.00018028433817041236, |
|
"loss": 0.7334, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.7274070523128118, |
|
"learning_rate": 0.00017950036633480556, |
|
"loss": 0.7217, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.7805160984212568, |
|
"learning_rate": 0.00017870288533002938, |
|
"loss": 0.6994, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.7642678319646234, |
|
"learning_rate": 0.00017789203066873998, |
|
"loss": 0.7349, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.7068890647609225, |
|
"learning_rate": 0.00017706794013612364, |
|
"loss": 0.7408, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.7821294112932925, |
|
"learning_rate": 0.00017623075376648376, |
|
"loss": 0.718, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.7229664348394781, |
|
"learning_rate": 0.00017538061381944524, |
|
"loss": 0.7418, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.6951044250779959, |
|
"learning_rate": 0.0001745176647557809, |
|
"loss": 0.6934, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.73099474521294, |
|
"learning_rate": 0.00017364205321286394, |
|
"loss": 0.7135, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.6940403521727994, |
|
"learning_rate": 0.00017275392797975032, |
|
"loss": 0.7765, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.7211852964311306, |
|
"learning_rate": 0.00017185343997189588, |
|
"loss": 0.7427, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.6503797788991111, |
|
"learning_rate": 0.00017094074220551158, |
|
"loss": 0.7161, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.7010217970027596, |
|
"learning_rate": 0.0001700159897715624, |
|
"loss": 0.747, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.6787203912897876, |
|
"learning_rate": 0.00016907933980941312, |
|
"loss": 0.6659, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.6654181924251962, |
|
"learning_rate": 0.0001681309514801265, |
|
"loss": 0.6811, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.6670712031136201, |
|
"learning_rate": 0.00016717098593941752, |
|
"loss": 0.7116, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.6505402907966938, |
|
"learning_rate": 0.00016619960631026888, |
|
"loss": 0.7259, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.6860899500587603, |
|
"learning_rate": 0.0001652169776552123, |
|
"loss": 0.7082, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.6590732633850581, |
|
"learning_rate": 0.00016422326694828007, |
|
"loss": 0.6806, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.6360343519947935, |
|
"learning_rate": 0.00016321864304663173, |
|
"loss": 0.6951, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.6698752473367972, |
|
"learning_rate": 0.000162203276661861, |
|
"loss": 0.7284, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.686029660179383, |
|
"learning_rate": 0.00016117734033098744, |
|
"loss": 0.7253, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.6709613558030975, |
|
"learning_rate": 0.00016014100838713797, |
|
"loss": 0.7116, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.7224879379478429, |
|
"learning_rate": 0.000159094456929923, |
|
"loss": 0.7016, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.6428021529957839, |
|
"learning_rate": 0.0001580378637955128, |
|
"loss": 0.666, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.6960916861503723, |
|
"learning_rate": 0.00015697140852641834, |
|
"loss": 0.6997, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.7255100351288608, |
|
"learning_rate": 0.00015589527234098247, |
|
"loss": 0.746, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.6986598054325619, |
|
"learning_rate": 0.00015480963810258613, |
|
"loss": 0.6955, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.6942988324021939, |
|
"learning_rate": 0.00015371469028857532, |
|
"loss": 0.73, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.6464360901011618, |
|
"learning_rate": 0.00015261061495891345, |
|
"loss": 0.7066, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.6817451321331462, |
|
"learning_rate": 0.0001514975997245649, |
|
"loss": 0.6862, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.6836150931516597, |
|
"learning_rate": 0.00015037583371561535, |
|
"loss": 0.701, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.6521598742458213, |
|
"learning_rate": 0.0001492455075491334, |
|
"loss": 0.6951, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.6840291643645185, |
|
"learning_rate": 0.00014810681329677987, |
|
"loss": 0.6891, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.7119537505083154, |
|
"learning_rate": 0.00014695994445216985, |
|
"loss": 0.7055, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.6809155849632806, |
|
"learning_rate": 0.00014580509589799329, |
|
"loss": 0.7158, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.6582134532528549, |
|
"learning_rate": 0.00014464246387289913, |
|
"loss": 0.6816, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.6858460794714393, |
|
"learning_rate": 0.00014347224593814944, |
|
"loss": 0.7331, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.6279879762170197, |
|
"learning_rate": 0.00014229464094404865, |
|
"loss": 0.6948, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.6781606487340365, |
|
"learning_rate": 0.00014110984899615367, |
|
"loss": 0.7012, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.6235005575378192, |
|
"learning_rate": 0.0001399180714212708, |
|
"loss": 0.6856, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.6222064438389479, |
|
"learning_rate": 0.00013871951073324507, |
|
"loss": 0.7224, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.6509887436189455, |
|
"learning_rate": 0.0001375143705985481, |
|
"loss": 0.6928, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.656042231341122, |
|
"learning_rate": 0.00013630285580166945, |
|
"loss": 0.6762, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.7077822267736881, |
|
"learning_rate": 0.000135085172210319, |
|
"loss": 0.6925, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.6377433248181216, |
|
"learning_rate": 0.00013386152674044422, |
|
"loss": 0.6744, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.6794295874230116, |
|
"learning_rate": 0.00013263212732107012, |
|
"loss": 0.7011, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.643130008468374, |
|
"learning_rate": 0.00013139718285896655, |
|
"loss": 0.6569, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.6362056274475615, |
|
"learning_rate": 0.00013015690320314954, |
|
"loss": 0.7033, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.635021955027263, |
|
"learning_rate": 0.00012891149910922267, |
|
"loss": 0.6963, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.6264643470826873, |
|
"learning_rate": 0.00012766118220356408, |
|
"loss": 0.6894, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.6342127038295544, |
|
"learning_rate": 0.0001264061649473657, |
|
"loss": 0.6738, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.6621446452478139, |
|
"learning_rate": 0.00012514666060053076, |
|
"loss": 0.6758, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.6226477303367899, |
|
"learning_rate": 0.00012388288318543512, |
|
"loss": 0.6503, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.6596957404307451, |
|
"learning_rate": 0.00012261504745055964, |
|
"loss": 0.651, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.6351342299233715, |
|
"learning_rate": 0.00012134336883399855, |
|
"loss": 0.6522, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.6285397550645089, |
|
"learning_rate": 0.00012006806342685126, |
|
"loss": 0.7196, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.6776322591334888, |
|
"learning_rate": 0.00011878934793650273, |
|
"loss": 0.5976, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.6928273118779775, |
|
"learning_rate": 0.00011750743964979918, |
|
"loss": 0.6546, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.639279635845659, |
|
"learning_rate": 0.00011622255639612554, |
|
"loss": 0.6946, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.6515482805820111, |
|
"learning_rate": 0.00011493491651039077, |
|
"loss": 0.698, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.660223620241514, |
|
"learning_rate": 0.00011364473879592674, |
|
"loss": 0.673, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.666404899679247, |
|
"learning_rate": 0.0001123522424873082, |
|
"loss": 0.6705, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.643302464795397, |
|
"learning_rate": 0.000111057647213099, |
|
"loss": 0.66, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.6221645734900332, |
|
"learning_rate": 0.00010976117295853154, |
|
"loss": 0.6306, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.6114539557810817, |
|
"learning_rate": 0.00010846304002812564, |
|
"loss": 0.6688, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.6441864804470223, |
|
"learning_rate": 0.00010716346900825299, |
|
"loss": 0.6914, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.6126099908354491, |
|
"learning_rate": 0.00010586268072965396, |
|
"loss": 0.6847, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.6226190206986176, |
|
"learning_rate": 0.00010456089622991263, |
|
"loss": 0.6405, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.5913341649918665, |
|
"learning_rate": 0.00010325833671589687, |
|
"loss": 0.7136, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.6063136858538771, |
|
"learning_rate": 0.00010195522352616943, |
|
"loss": 0.643, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.6474255343280048, |
|
"learning_rate": 0.00010065177809337702, |
|
"loss": 0.7012, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.611817998176696, |
|
"learning_rate": 9.934822190662299e-05, |
|
"loss": 0.7232, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.6315544767518864, |
|
"learning_rate": 9.80447764738306e-05, |
|
"loss": 0.6531, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.5824211156475344, |
|
"learning_rate": 9.674166328410318e-05, |
|
"loss": 0.6308, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.6172010731454831, |
|
"learning_rate": 9.543910377008742e-05, |
|
"loss": 0.66, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.6025679891477139, |
|
"learning_rate": 9.413731927034605e-05, |
|
"loss": 0.6332, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.6396295661454142, |
|
"learning_rate": 9.283653099174704e-05, |
|
"loss": 0.6352, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.6490127330699687, |
|
"learning_rate": 9.15369599718744e-05, |
|
"loss": 0.667, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.6428918190195956, |
|
"learning_rate": 9.023882704146848e-05, |
|
"loss": 0.6278, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.6195852999012177, |
|
"learning_rate": 8.894235278690104e-05, |
|
"loss": 0.6919, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.5890514234117395, |
|
"learning_rate": 8.764775751269182e-05, |
|
"loss": 0.6282, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.6383232450338161, |
|
"learning_rate": 8.635526120407329e-05, |
|
"loss": 0.6335, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.6298327584602661, |
|
"learning_rate": 8.506508348960924e-05, |
|
"loss": 0.6757, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.5951533121419991, |
|
"learning_rate": 8.377744360387447e-05, |
|
"loss": 0.6378, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.5990230767413158, |
|
"learning_rate": 8.249256035020086e-05, |
|
"loss": 0.6398, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.594734539715176, |
|
"learning_rate": 8.121065206349729e-05, |
|
"loss": 0.6688, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.624787885220957, |
|
"learning_rate": 7.993193657314875e-05, |
|
"loss": 0.6287, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.6482697342730748, |
|
"learning_rate": 7.865663116600148e-05, |
|
"loss": 0.6586, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.713715611061142, |
|
"learning_rate": 7.738495254944042e-05, |
|
"loss": 0.6697, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.6234960883575744, |
|
"learning_rate": 7.611711681456493e-05, |
|
"loss": 0.6368, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.630287976111448, |
|
"learning_rate": 7.485333939946926e-05, |
|
"loss": 0.6508, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.6378675582501994, |
|
"learning_rate": 7.359383505263431e-05, |
|
"loss": 0.6841, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.6156108176431241, |
|
"learning_rate": 7.233881779643594e-05, |
|
"loss": 0.6444, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.5897421977125701, |
|
"learning_rate": 7.108850089077735e-05, |
|
"loss": 0.6719, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.5849183219702926, |
|
"learning_rate": 6.98430967968505e-05, |
|
"loss": 0.6733, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.6260421998777234, |
|
"learning_rate": 6.86028171410335e-05, |
|
"loss": 0.6383, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.6304596579295512, |
|
"learning_rate": 6.736787267892991e-05, |
|
"loss": 0.6717, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.6048424796187638, |
|
"learning_rate": 6.613847325955578e-05, |
|
"loss": 0.659, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.6343949665489499, |
|
"learning_rate": 6.491482778968104e-05, |
|
"loss": 0.6651, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.6032899750286133, |
|
"learning_rate": 6.369714419833056e-05, |
|
"loss": 0.6337, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.7108946242576734, |
|
"learning_rate": 6.248562940145195e-05, |
|
"loss": 0.6459, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.5901296881624672, |
|
"learning_rate": 6.128048926675494e-05, |
|
"loss": 0.6764, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.5918343934659875, |
|
"learning_rate": 6.008192857872923e-05, |
|
"loss": 0.6428, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.5970600618333325, |
|
"learning_rate": 5.889015100384636e-05, |
|
"loss": 0.6324, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.6038442007114545, |
|
"learning_rate": 5.770535905595138e-05, |
|
"loss": 0.6496, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.6204490128832106, |
|
"learning_rate": 5.6527754061850554e-05, |
|
"loss": 0.6485, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.6179929201872987, |
|
"learning_rate": 5.5357536127100904e-05, |
|
"loss": 0.6545, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.619221490890743, |
|
"learning_rate": 5.419490410200675e-05, |
|
"loss": 0.612, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.5966676205978181, |
|
"learning_rate": 5.304005554783015e-05, |
|
"loss": 0.6498, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.5822755936769207, |
|
"learning_rate": 5.1893186703220165e-05, |
|
"loss": 0.6171, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.5842041127752368, |
|
"learning_rate": 5.0754492450866607e-05, |
|
"loss": 0.6243, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.6337182771676403, |
|
"learning_rate": 4.9624166284384656e-05, |
|
"loss": 0.6147, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.6017489055813804, |
|
"learning_rate": 4.850240027543509e-05, |
|
"loss": 0.653, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.5809709526862709, |
|
"learning_rate": 4.738938504108659e-05, |
|
"loss": 0.6267, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.5826938185721535, |
|
"learning_rate": 4.628530971142471e-05, |
|
"loss": 0.6213, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.5737019305587774, |
|
"learning_rate": 4.519036189741386e-05, |
|
"loss": 0.6502, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.5945680765374372, |
|
"learning_rate": 4.410472765901755e-05, |
|
"loss": 0.6286, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.6094073434165231, |
|
"learning_rate": 4.302859147358168e-05, |
|
"loss": 0.5848, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.6187888066340587, |
|
"learning_rate": 4.196213620448723e-05, |
|
"loss": 0.6439, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.6419980315369229, |
|
"learning_rate": 4.0905543070077036e-05, |
|
"loss": 0.6471, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.6337264596412628, |
|
"learning_rate": 3.985899161286205e-05, |
|
"loss": 0.6616, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.5832375614778771, |
|
"learning_rate": 3.882265966901257e-05, |
|
"loss": 0.6767, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.574891135054901, |
|
"learning_rate": 3.7796723338138995e-05, |
|
"loss": 0.5807, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.5699073164192423, |
|
"learning_rate": 3.6781356953368284e-05, |
|
"loss": 0.6371, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.616654702040988, |
|
"learning_rate": 3.5776733051719936e-05, |
|
"loss": 0.6086, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.5867706666852922, |
|
"learning_rate": 3.47830223447877e-05, |
|
"loss": 0.6639, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.623145463941295, |
|
"learning_rate": 3.3800393689731146e-05, |
|
"loss": 0.6345, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.5981407629418406, |
|
"learning_rate": 3.28290140605825e-05, |
|
"loss": 0.6273, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.5674343545696854, |
|
"learning_rate": 3.186904851987351e-05, |
|
"loss": 0.6058, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.6399842879941413, |
|
"learning_rate": 3.092066019058689e-05, |
|
"loss": 0.6231, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.6313162630484164, |
|
"learning_rate": 2.998401022843761e-05, |
|
"loss": 0.6486, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.6028986124900413, |
|
"learning_rate": 2.9059257794488424e-05, |
|
"loss": 0.567, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.5960517928587284, |
|
"learning_rate": 2.8146560028104153e-05, |
|
"loss": 0.6716, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.6286103402740058, |
|
"learning_rate": 2.724607202024969e-05, |
|
"loss": 0.6188, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.644610126637308, |
|
"learning_rate": 2.6357946787136113e-05, |
|
"loss": 0.6434, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.5729979241776278, |
|
"learning_rate": 2.548233524421911e-05, |
|
"loss": 0.6319, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.6328611584490075, |
|
"learning_rate": 2.461938618055478e-05, |
|
"loss": 0.6554, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.5975966512993659, |
|
"learning_rate": 2.3769246233516242e-05, |
|
"loss": 0.5977, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.59320648444735, |
|
"learning_rate": 2.2932059863876365e-05, |
|
"loss": 0.6456, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.590627235987933, |
|
"learning_rate": 2.2107969331260048e-05, |
|
"loss": 0.6259, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.6410336378215516, |
|
"learning_rate": 2.1297114669970618e-05, |
|
"loss": 0.5967, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.5839140783732536, |
|
"learning_rate": 2.049963366519446e-05, |
|
"loss": 0.6245, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.6055114057003023, |
|
"learning_rate": 1.971566182958765e-05, |
|
"loss": 0.5994, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.588555313698678, |
|
"learning_rate": 1.8945332380248913e-05, |
|
"loss": 0.604, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.6088478693980861, |
|
"learning_rate": 1.8188776216082603e-05, |
|
"loss": 0.6212, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.5893151236661324, |
|
"learning_rate": 1.7446121895555555e-05, |
|
"loss": 0.632, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.599854694440474, |
|
"learning_rate": 1.6717495614851652e-05, |
|
"loss": 0.6231, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.5802528574601737, |
|
"learning_rate": 1.6003021186427893e-05, |
|
"loss": 0.6234, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.585335132783131, |
|
"learning_rate": 1.5302820017975394e-05, |
|
"loss": 0.6269, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.5818204916703836, |
|
"learning_rate": 1.4617011091789135e-05, |
|
"loss": 0.6104, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.568918470562012, |
|
"learning_rate": 1.3945710944549706e-05, |
|
"loss": 0.6178, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.5925712778064853, |
|
"learning_rate": 1.3289033647520877e-05, |
|
"loss": 0.5969, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.5906570988072316, |
|
"learning_rate": 1.2647090787165694e-05, |
|
"loss": 0.6202, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.625879731780627, |
|
"learning_rate": 1.2019991446185309e-05, |
|
"loss": 0.6114, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.6143685840442832, |
|
"learning_rate": 1.1407842184982786e-05, |
|
"loss": 0.6186, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.6155175082291917, |
|
"learning_rate": 1.0810747023555878e-05, |
|
"loss": 0.6102, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.5795310616149347, |
|
"learning_rate": 1.0228807423821263e-05, |
|
"loss": 0.5998, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.5831880790523601, |
|
"learning_rate": 9.662122272373575e-06, |
|
"loss": 0.5945, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.6290115114493054, |
|
"learning_rate": 9.110787863682002e-06, |
|
"loss": 0.5857, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.6603504517272147, |
|
"learning_rate": 8.574897883727384e-06, |
|
"loss": 0.6373, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.6087100188068558, |
|
"learning_rate": 8.054543394082504e-06, |
|
"loss": 0.6134, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.6064067085429232, |
|
"learning_rate": 7.5498128164383955e-06, |
|
"loss": 0.5998, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.5920112342958288, |
|
"learning_rate": 7.0607919175791796e-06, |
|
"loss": 0.6134, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.5549713422524726, |
|
"learning_rate": 6.587563794808127e-06, |
|
"loss": 0.644, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.6110503712077385, |
|
"learning_rate": 6.130208861827202e-06, |
|
"loss": 0.6276, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.6526080947670703, |
|
"learning_rate": 5.688804835072748e-06, |
|
"loss": 0.6122, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.592228703593387, |
|
"learning_rate": 5.263426720509468e-06, |
|
"loss": 0.5666, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.6154318985980524, |
|
"learning_rate": 4.8541468008849285e-06, |
|
"loss": 0.6022, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.6245746165165417, |
|
"learning_rate": 4.461034623446847e-06, |
|
"loss": 0.5916, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.587908724284237, |
|
"learning_rate": 4.084156988125231e-06, |
|
"loss": 0.6164, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.5951699842997723, |
|
"learning_rate": 3.723577936181366e-06, |
|
"loss": 0.6195, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.5667459142096927, |
|
"learning_rate": 3.3793587393255e-06, |
|
"loss": 0.5908, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.5639108144148681, |
|
"learning_rate": 3.0515578893052344e-06, |
|
"loss": 0.5743, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.5459205466426152, |
|
"learning_rate": 2.7402310879662497e-06, |
|
"loss": 0.6269, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.6031704942783882, |
|
"learning_rate": 2.44543123778711e-06, |
|
"loss": 0.6269, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.5770810110581429, |
|
"learning_rate": 2.167208432889789e-06, |
|
"loss": 0.6151, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.5952891318172381, |
|
"learning_rate": 1.9056099505273427e-06, |
|
"loss": 0.5728, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.5904225313614513, |
|
"learning_rate": 1.6606802430503166e-06, |
|
"loss": 0.6235, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.5782760417216553, |
|
"learning_rate": 1.43246093035313e-06, |
|
"loss": 0.5593, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.580467530625868, |
|
"learning_rate": 1.2209907928017795e-06, |
|
"loss": 0.6423, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.5711598715443851, |
|
"learning_rate": 1.0263057646440199e-06, |
|
"loss": 0.6291, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.5686426031326551, |
|
"learning_rate": 8.484389279032834e-07, |
|
"loss": 0.6198, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.6046363181474087, |
|
"learning_rate": 6.874205067571083e-07, |
|
"loss": 0.6468, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.6072995257131998, |
|
"learning_rate": 5.432778624013257e-07, |
|
"loss": 0.6486, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.6201193729192779, |
|
"learning_rate": 4.1603548840062345e-07, |
|
"loss": 0.6272, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.6312528212744211, |
|
"learning_rate": 3.0571500652651907e-07, |
|
"loss": 0.6177, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.5687699288742071, |
|
"learning_rate": 2.1233516308323264e-07, |
|
"loss": 0.618, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.5523484801821387, |
|
"learning_rate": 1.359118257221903e-07, |
|
"loss": 0.5945, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.636843838687328, |
|
"learning_rate": 7.645798074572552e-08, |
|
"loss": 0.6551, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.5817618418266682, |
|
"learning_rate": 3.3983730900377655e-08, |
|
"loss": 0.6149, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.6358174664632625, |
|
"learning_rate": 8.496293660120724e-09, |
|
"loss": 0.6777, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.6195064507565601, |
|
"learning_rate": 0.0, |
|
"loss": 0.5658, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 249, |
|
"total_flos": 37734407700480.0, |
|
"train_loss": 0.7035722335179647, |
|
"train_runtime": 3820.5554, |
|
"train_samples_per_second": 4.162, |
|
"train_steps_per_second": 0.065 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 249, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 1000, |
|
"total_flos": 37734407700480.0, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|