|
{ |
|
"best_metric": 0.9624110460281372, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-100", |
|
"epoch": 3.0049751243781095, |
|
"eval_steps": 50, |
|
"global_step": 151, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01990049751243781, |
|
"grad_norm": 5.543930530548096, |
|
"learning_rate": 1e-05, |
|
"loss": 5.5751, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.01990049751243781, |
|
"eval_loss": 9.599178314208984, |
|
"eval_runtime": 6.162, |
|
"eval_samples_per_second": 13.794, |
|
"eval_steps_per_second": 3.57, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.03980099502487562, |
|
"grad_norm": 5.9391279220581055, |
|
"learning_rate": 2e-05, |
|
"loss": 6.2605, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.05970149253731343, |
|
"grad_norm": 10.055806159973145, |
|
"learning_rate": 3e-05, |
|
"loss": 6.6465, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.07960199004975124, |
|
"grad_norm": 24.726375579833984, |
|
"learning_rate": 4e-05, |
|
"loss": 9.6417, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.09950248756218906, |
|
"grad_norm": 21.94843292236328, |
|
"learning_rate": 5e-05, |
|
"loss": 9.1886, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.11940298507462686, |
|
"grad_norm": 17.2749080657959, |
|
"learning_rate": 6e-05, |
|
"loss": 7.9442, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.13930348258706468, |
|
"grad_norm": 20.374170303344727, |
|
"learning_rate": 7e-05, |
|
"loss": 5.9546, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.15920398009950248, |
|
"grad_norm": 14.176255226135254, |
|
"learning_rate": 8e-05, |
|
"loss": 3.7086, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.1791044776119403, |
|
"grad_norm": 8.863327026367188, |
|
"learning_rate": 9e-05, |
|
"loss": 1.8624, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.19900497512437812, |
|
"grad_norm": 7.542079925537109, |
|
"learning_rate": 0.0001, |
|
"loss": 5.9968, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.21890547263681592, |
|
"grad_norm": 15.19139289855957, |
|
"learning_rate": 9.998758966336295e-05, |
|
"loss": 8.1119, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.23880597014925373, |
|
"grad_norm": 18.45025634765625, |
|
"learning_rate": 9.995036481411004e-05, |
|
"loss": 8.9081, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.25870646766169153, |
|
"grad_norm": 12.949933052062988, |
|
"learning_rate": 9.988834393115767e-05, |
|
"loss": 4.0368, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.27860696517412936, |
|
"grad_norm": 12.047260284423828, |
|
"learning_rate": 9.980155780250727e-05, |
|
"loss": 4.0461, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.29850746268656714, |
|
"grad_norm": 10.137738227844238, |
|
"learning_rate": 9.969004950996175e-05, |
|
"loss": 3.5263, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.31840796019900497, |
|
"grad_norm": 7.845066547393799, |
|
"learning_rate": 9.9553874407739e-05, |
|
"loss": 1.8586, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.3383084577114428, |
|
"grad_norm": 1.7427806854248047, |
|
"learning_rate": 9.939310009499348e-05, |
|
"loss": 0.4681, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.3582089552238806, |
|
"grad_norm": 1.3990861177444458, |
|
"learning_rate": 9.92078063822589e-05, |
|
"loss": 0.3307, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.3781094527363184, |
|
"grad_norm": 0.9510717988014221, |
|
"learning_rate": 9.899808525182935e-05, |
|
"loss": 0.148, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.39800995024875624, |
|
"grad_norm": 0.7023572325706482, |
|
"learning_rate": 9.876404081209796e-05, |
|
"loss": 0.0863, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.417910447761194, |
|
"grad_norm": 2.599980354309082, |
|
"learning_rate": 9.850578924587614e-05, |
|
"loss": 0.8357, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.43781094527363185, |
|
"grad_norm": 8.273324966430664, |
|
"learning_rate": 9.822345875271883e-05, |
|
"loss": 3.8429, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.4577114427860697, |
|
"grad_norm": 10.221076965332031, |
|
"learning_rate": 9.791718948528458e-05, |
|
"loss": 4.1774, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.47761194029850745, |
|
"grad_norm": 10.803452491760254, |
|
"learning_rate": 9.758713347976179e-05, |
|
"loss": 4.457, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.4975124378109453, |
|
"grad_norm": 9.160074234008789, |
|
"learning_rate": 9.723345458039594e-05, |
|
"loss": 2.9867, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.5174129353233831, |
|
"grad_norm": 4.31099796295166, |
|
"learning_rate": 9.685632835815518e-05, |
|
"loss": 2.5178, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.5373134328358209, |
|
"grad_norm": 3.5618479251861572, |
|
"learning_rate": 9.645594202357439e-05, |
|
"loss": 2.5185, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.5572139303482587, |
|
"grad_norm": 2.6641221046447754, |
|
"learning_rate": 9.603249433382144e-05, |
|
"loss": 1.2579, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.5771144278606966, |
|
"grad_norm": 1.0730020999908447, |
|
"learning_rate": 9.558619549403147e-05, |
|
"loss": 0.3654, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.5970149253731343, |
|
"grad_norm": 0.2987671196460724, |
|
"learning_rate": 9.511726705295817e-05, |
|
"loss": 0.0464, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.6169154228855721, |
|
"grad_norm": 0.8347283005714417, |
|
"learning_rate": 9.462594179299406e-05, |
|
"loss": 0.2107, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.6368159203980099, |
|
"grad_norm": 0.8029500246047974, |
|
"learning_rate": 9.41124636146141e-05, |
|
"loss": 0.179, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.6567164179104478, |
|
"grad_norm": 1.6079643964767456, |
|
"learning_rate": 9.357708741530025e-05, |
|
"loss": 0.5489, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.6766169154228856, |
|
"grad_norm": 7.646543502807617, |
|
"learning_rate": 9.302007896300698e-05, |
|
"loss": 3.2044, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.6965174129353234, |
|
"grad_norm": 8.641079902648926, |
|
"learning_rate": 9.244171476423037e-05, |
|
"loss": 3.0777, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.7164179104477612, |
|
"grad_norm": 11.796111106872559, |
|
"learning_rate": 9.184228192674667e-05, |
|
"loss": 4.0429, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.736318407960199, |
|
"grad_norm": 5.918362617492676, |
|
"learning_rate": 9.122207801708802e-05, |
|
"loss": 2.5737, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.7562189054726368, |
|
"grad_norm": 6.796441078186035, |
|
"learning_rate": 9.058141091282656e-05, |
|
"loss": 2.2759, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.7761194029850746, |
|
"grad_norm": 5.166509628295898, |
|
"learning_rate": 8.992059864973972e-05, |
|
"loss": 2.3741, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.7960199004975125, |
|
"grad_norm": 9.372264862060547, |
|
"learning_rate": 8.923996926393305e-05, |
|
"loss": 0.6126, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.8159203980099502, |
|
"grad_norm": 0.7963209748268127, |
|
"learning_rate": 8.853986062899868e-05, |
|
"loss": 0.2505, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.835820895522388, |
|
"grad_norm": 0.8974098563194275, |
|
"learning_rate": 8.782062028829028e-05, |
|
"loss": 0.3499, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.8557213930348259, |
|
"grad_norm": 0.8836018443107605, |
|
"learning_rate": 8.708260528239788e-05, |
|
"loss": 0.336, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.8756218905472637, |
|
"grad_norm": 0.41947388648986816, |
|
"learning_rate": 8.632618197190816e-05, |
|
"loss": 0.0833, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.8955223880597015, |
|
"grad_norm": 0.6316033601760864, |
|
"learning_rate": 8.555172585553805e-05, |
|
"loss": 0.1746, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.9154228855721394, |
|
"grad_norm": 3.0872931480407715, |
|
"learning_rate": 8.475962138373213e-05, |
|
"loss": 1.8755, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.9353233830845771, |
|
"grad_norm": 5.99851655960083, |
|
"learning_rate": 8.395026176781627e-05, |
|
"loss": 3.2634, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.9552238805970149, |
|
"grad_norm": 8.195245742797852, |
|
"learning_rate": 8.312404878480222e-05, |
|
"loss": 3.2091, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.9751243781094527, |
|
"grad_norm": 2.1140687465667725, |
|
"learning_rate": 8.228139257794012e-05, |
|
"loss": 1.2906, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.9950248756218906, |
|
"grad_norm": 2.3306195735931396, |
|
"learning_rate": 8.142271145311783e-05, |
|
"loss": 1.0865, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.9950248756218906, |
|
"eval_loss": 1.0756102800369263, |
|
"eval_runtime": 6.3397, |
|
"eval_samples_per_second": 13.408, |
|
"eval_steps_per_second": 3.47, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.0149253731343284, |
|
"grad_norm": 9.008943557739258, |
|
"learning_rate": 8.054843167120827e-05, |
|
"loss": 3.9621, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 1.0348258706467661, |
|
"grad_norm": 2.001354455947876, |
|
"learning_rate": 7.965898723646776e-05, |
|
"loss": 1.5088, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 1.054726368159204, |
|
"grad_norm": 2.112483024597168, |
|
"learning_rate": 7.875481968109052e-05, |
|
"loss": 1.5239, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 1.0746268656716418, |
|
"grad_norm": 1.9544909000396729, |
|
"learning_rate": 7.783637784602609e-05, |
|
"loss": 0.975, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 1.0945273631840795, |
|
"grad_norm": 0.5338630676269531, |
|
"learning_rate": 7.690411765816864e-05, |
|
"loss": 0.1268, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 1.1144278606965174, |
|
"grad_norm": 0.5618958473205566, |
|
"learning_rate": 7.595850190402876e-05, |
|
"loss": 0.1182, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 1.1343283582089552, |
|
"grad_norm": 0.6160061955451965, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 0.144, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 1.154228855721393, |
|
"grad_norm": 0.5106895565986633, |
|
"learning_rate": 7.402908775933419e-05, |
|
"loss": 0.0888, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 1.1741293532338308, |
|
"grad_norm": 0.39127403497695923, |
|
"learning_rate": 7.304624715594139e-05, |
|
"loss": 0.0848, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 1.1940298507462686, |
|
"grad_norm": 3.96272611618042, |
|
"learning_rate": 7.205196608513159e-05, |
|
"loss": 1.3122, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.2139303482587065, |
|
"grad_norm": 4.137331008911133, |
|
"learning_rate": 7.104673812141675e-05, |
|
"loss": 2.0647, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 1.2338308457711442, |
|
"grad_norm": 9.125043869018555, |
|
"learning_rate": 7.003106227349399e-05, |
|
"loss": 2.4258, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 1.2537313432835822, |
|
"grad_norm": 4.514773845672607, |
|
"learning_rate": 6.900544273653074e-05, |
|
"loss": 2.0476, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 1.2736318407960199, |
|
"grad_norm": 2.9408376216888428, |
|
"learning_rate": 6.797038864187564e-05, |
|
"loss": 1.5427, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 1.2935323383084576, |
|
"grad_norm": 2.8879308700561523, |
|
"learning_rate": 6.692641380431879e-05, |
|
"loss": 1.7285, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.3134328358208955, |
|
"grad_norm": 1.3168548345565796, |
|
"learning_rate": 6.587403646702714e-05, |
|
"loss": 0.5157, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 1.3333333333333333, |
|
"grad_norm": 0.7307369112968445, |
|
"learning_rate": 6.481377904428171e-05, |
|
"loss": 0.138, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 1.3532338308457712, |
|
"grad_norm": 0.3926345705986023, |
|
"learning_rate": 6.374616786214402e-05, |
|
"loss": 0.0664, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 1.373134328358209, |
|
"grad_norm": 0.3398008346557617, |
|
"learning_rate": 6.26717328971808e-05, |
|
"loss": 0.0546, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 1.3930348258706466, |
|
"grad_norm": 0.7794240713119507, |
|
"learning_rate": 6.159100751337642e-05, |
|
"loss": 0.1423, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.4129353233830846, |
|
"grad_norm": 0.28816598653793335, |
|
"learning_rate": 6.0504528197363894e-05, |
|
"loss": 0.0374, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 1.4328358208955223, |
|
"grad_norm": 2.6052889823913574, |
|
"learning_rate": 5.9412834292105676e-05, |
|
"loss": 1.3742, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 1.4527363184079602, |
|
"grad_norm": 5.297976970672607, |
|
"learning_rate": 5.831646772915651e-05, |
|
"loss": 2.3129, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 1.472636815920398, |
|
"grad_norm": 6.222244739532471, |
|
"learning_rate": 5.721597275964133e-05, |
|
"loss": 2.4451, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 1.4925373134328357, |
|
"grad_norm": 3.1612331867218018, |
|
"learning_rate": 5.6111895684081725e-05, |
|
"loss": 1.5285, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.5124378109452736, |
|
"grad_norm": 2.373920440673828, |
|
"learning_rate": 5.5004784581204927e-05, |
|
"loss": 1.4736, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 1.5323383084577116, |
|
"grad_norm": 2.493252992630005, |
|
"learning_rate": 5.389518903587017e-05, |
|
"loss": 1.5259, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 1.5522388059701493, |
|
"grad_norm": 2.429081678390503, |
|
"learning_rate": 5.2783659866247424e-05, |
|
"loss": 1.6323, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 1.572139303482587, |
|
"grad_norm": 0.9949527382850647, |
|
"learning_rate": 5.167074885038373e-05, |
|
"loss": 0.3188, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 1.5920398009950247, |
|
"grad_norm": 0.5965785980224609, |
|
"learning_rate": 5.055700845229327e-05, |
|
"loss": 0.1503, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.6119402985074627, |
|
"grad_norm": 0.5434532165527344, |
|
"learning_rate": 4.944299154770673e-05, |
|
"loss": 0.094, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 1.6318407960199006, |
|
"grad_norm": 0.7457742691040039, |
|
"learning_rate": 4.832925114961629e-05, |
|
"loss": 0.164, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 1.6517412935323383, |
|
"grad_norm": 0.3958088457584381, |
|
"learning_rate": 4.72163401337526e-05, |
|
"loss": 0.0606, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 1.671641791044776, |
|
"grad_norm": 1.8904391527175903, |
|
"learning_rate": 4.610481096412984e-05, |
|
"loss": 0.9604, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 1.6915422885572138, |
|
"grad_norm": 4.544404029846191, |
|
"learning_rate": 4.4995215418795085e-05, |
|
"loss": 2.215, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.7114427860696517, |
|
"grad_norm": 5.509747505187988, |
|
"learning_rate": 4.388810431591829e-05, |
|
"loss": 2.0638, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 1.7313432835820897, |
|
"grad_norm": 2.859410047531128, |
|
"learning_rate": 4.278402724035867e-05, |
|
"loss": 1.693, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 1.7512437810945274, |
|
"grad_norm": 2.292931079864502, |
|
"learning_rate": 4.1683532270843504e-05, |
|
"loss": 1.3341, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 1.771144278606965, |
|
"grad_norm": 2.5066726207733154, |
|
"learning_rate": 4.058716570789433e-05, |
|
"loss": 1.4565, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 1.7910447761194028, |
|
"grad_norm": 1.7927272319793701, |
|
"learning_rate": 3.94954718026361e-05, |
|
"loss": 0.892, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.8109452736318408, |
|
"grad_norm": 0.5884804725646973, |
|
"learning_rate": 3.840899248662358e-05, |
|
"loss": 0.1595, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 1.8308457711442787, |
|
"grad_norm": 1.1309645175933838, |
|
"learning_rate": 3.7328267102819225e-05, |
|
"loss": 0.2703, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 1.8507462686567164, |
|
"grad_norm": 0.31311914324760437, |
|
"learning_rate": 3.6253832137856e-05, |
|
"loss": 0.0479, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 1.8706467661691542, |
|
"grad_norm": 0.5945075154304504, |
|
"learning_rate": 3.5186220955718306e-05, |
|
"loss": 0.1501, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 1.890547263681592, |
|
"grad_norm": 1.0878932476043701, |
|
"learning_rate": 3.4125963532972873e-05, |
|
"loss": 0.4793, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 1.9104477611940298, |
|
"grad_norm": 3.067518472671509, |
|
"learning_rate": 3.307358619568123e-05, |
|
"loss": 1.7691, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 1.9303482587064678, |
|
"grad_norm": 3.476691722869873, |
|
"learning_rate": 3.202961135812437e-05, |
|
"loss": 1.928, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 1.9502487562189055, |
|
"grad_norm": 4.594018459320068, |
|
"learning_rate": 3.0994557263469265e-05, |
|
"loss": 2.3617, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 1.9701492537313432, |
|
"grad_norm": 2.5807111263275146, |
|
"learning_rate": 2.996893772650602e-05, |
|
"loss": 1.4338, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 1.9900497512437811, |
|
"grad_norm": 0.2568559944629669, |
|
"learning_rate": 2.895326187858326e-05, |
|
"loss": 0.0345, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.9900497512437811, |
|
"eval_loss": 0.9624110460281372, |
|
"eval_runtime": 6.3475, |
|
"eval_samples_per_second": 13.391, |
|
"eval_steps_per_second": 3.466, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.009950248756219, |
|
"grad_norm": 15.313194274902344, |
|
"learning_rate": 2.7948033914868415e-05, |
|
"loss": 3.6961, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 2.029850746268657, |
|
"grad_norm": 1.6831941604614258, |
|
"learning_rate": 2.69537528440586e-05, |
|
"loss": 0.9623, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 2.0497512437810945, |
|
"grad_norm": 1.7447772026062012, |
|
"learning_rate": 2.5970912240665813e-05, |
|
"loss": 1.0607, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 2.0696517412935322, |
|
"grad_norm": 1.8062071800231934, |
|
"learning_rate": 2.500000000000001e-05, |
|
"loss": 1.0382, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 2.08955223880597, |
|
"grad_norm": 0.533903181552887, |
|
"learning_rate": 2.4041498095971253e-05, |
|
"loss": 0.0953, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 2.109452736318408, |
|
"grad_norm": 0.41501080989837646, |
|
"learning_rate": 2.3095882341831372e-05, |
|
"loss": 0.0889, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 2.129353233830846, |
|
"grad_norm": 0.605328381061554, |
|
"learning_rate": 2.216362215397393e-05, |
|
"loss": 0.1302, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 2.1492537313432836, |
|
"grad_norm": 0.3093937039375305, |
|
"learning_rate": 2.124518031890948e-05, |
|
"loss": 0.0564, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 2.1691542288557213, |
|
"grad_norm": 0.321236252784729, |
|
"learning_rate": 2.0341012763532243e-05, |
|
"loss": 0.0479, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 2.189054726368159, |
|
"grad_norm": 1.6769386529922485, |
|
"learning_rate": 1.945156832879174e-05, |
|
"loss": 0.6905, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 2.208955223880597, |
|
"grad_norm": 3.2930898666381836, |
|
"learning_rate": 1.8577288546882167e-05, |
|
"loss": 1.2683, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 2.228855721393035, |
|
"grad_norm": 3.6497483253479004, |
|
"learning_rate": 1.771860742205988e-05, |
|
"loss": 1.4517, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 2.2487562189054726, |
|
"grad_norm": 2.7759523391723633, |
|
"learning_rate": 1.687595121519778e-05, |
|
"loss": 1.2682, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 2.2686567164179103, |
|
"grad_norm": 1.7020666599273682, |
|
"learning_rate": 1.604973823218376e-05, |
|
"loss": 0.8624, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 2.288557213930348, |
|
"grad_norm": 2.058931589126587, |
|
"learning_rate": 1.5240378616267886e-05, |
|
"loss": 0.9802, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 2.308457711442786, |
|
"grad_norm": 1.6504006385803223, |
|
"learning_rate": 1.4448274144461965e-05, |
|
"loss": 0.626, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 2.328358208955224, |
|
"grad_norm": 0.36916500329971313, |
|
"learning_rate": 1.367381802809185e-05, |
|
"loss": 0.0499, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 2.3482587064676617, |
|
"grad_norm": 0.4427722692489624, |
|
"learning_rate": 1.2917394717602121e-05, |
|
"loss": 0.0567, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 2.3681592039800994, |
|
"grad_norm": 0.24329277873039246, |
|
"learning_rate": 1.2179379711709737e-05, |
|
"loss": 0.0243, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 2.388059701492537, |
|
"grad_norm": 0.7087410688400269, |
|
"learning_rate": 1.1460139371001338e-05, |
|
"loss": 0.155, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.4079601990049753, |
|
"grad_norm": 0.24490374326705933, |
|
"learning_rate": 1.0760030736066951e-05, |
|
"loss": 0.011, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 2.427860696517413, |
|
"grad_norm": 2.3325791358947754, |
|
"learning_rate": 1.0079401350260287e-05, |
|
"loss": 0.6917, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 2.4477611940298507, |
|
"grad_norm": 3.353835344314575, |
|
"learning_rate": 9.41858908717344e-06, |
|
"loss": 1.2289, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 2.4676616915422884, |
|
"grad_norm": 4.807076930999756, |
|
"learning_rate": 8.777921982911996e-06, |
|
"loss": 1.6531, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 2.487562189054726, |
|
"grad_norm": 3.2822203636169434, |
|
"learning_rate": 8.157718073253351e-06, |
|
"loss": 1.3298, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.5074626865671643, |
|
"grad_norm": 2.1589744091033936, |
|
"learning_rate": 7.558285235769646e-06, |
|
"loss": 0.8649, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 2.527363184079602, |
|
"grad_norm": 2.269338607788086, |
|
"learning_rate": 6.979921036993042e-06, |
|
"loss": 1.0175, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 2.5472636815920398, |
|
"grad_norm": 1.7215580940246582, |
|
"learning_rate": 6.422912584699753e-06, |
|
"loss": 0.4796, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 2.5671641791044775, |
|
"grad_norm": 0.5773996710777283, |
|
"learning_rate": 5.887536385385917e-06, |
|
"loss": 0.0872, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 2.587064676616915, |
|
"grad_norm": 0.3628270924091339, |
|
"learning_rate": 5.374058207005944e-06, |
|
"loss": 0.0363, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.6069651741293534, |
|
"grad_norm": 0.5935407876968384, |
|
"learning_rate": 4.882732947041818e-06, |
|
"loss": 0.1109, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 2.626865671641791, |
|
"grad_norm": 0.3369639217853546, |
|
"learning_rate": 4.413804505968533e-06, |
|
"loss": 0.0402, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 2.646766169154229, |
|
"grad_norm": 0.4762958288192749, |
|
"learning_rate": 3.967505666178556e-06, |
|
"loss": 0.0557, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 2.6666666666666665, |
|
"grad_norm": 2.259021759033203, |
|
"learning_rate": 3.544057976425619e-06, |
|
"loss": 0.8966, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 2.6865671641791042, |
|
"grad_norm": 3.387164831161499, |
|
"learning_rate": 3.1436716418448307e-06, |
|
"loss": 1.3748, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 2.7064676616915424, |
|
"grad_norm": 4.192432880401611, |
|
"learning_rate": 2.7665454196040664e-06, |
|
"loss": 1.4131, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 2.72636815920398, |
|
"grad_norm": 2.7902915477752686, |
|
"learning_rate": 2.4128665202382326e-06, |
|
"loss": 1.1285, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 2.746268656716418, |
|
"grad_norm": 2.0142769813537598, |
|
"learning_rate": 2.0828105147154273e-06, |
|
"loss": 0.7524, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 2.7661691542288556, |
|
"grad_norm": 2.4600789546966553, |
|
"learning_rate": 1.7765412472811771e-06, |
|
"loss": 1.0153, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 2.7860696517412933, |
|
"grad_norm": 2.070169448852539, |
|
"learning_rate": 1.4942107541238704e-06, |
|
"loss": 0.8548, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.8059701492537314, |
|
"grad_norm": 0.9298514127731323, |
|
"learning_rate": 1.2359591879020526e-06, |
|
"loss": 0.2304, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 2.825870646766169, |
|
"grad_norm": 0.287707656621933, |
|
"learning_rate": 1.0019147481706625e-06, |
|
"loss": 0.0342, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 2.845771144278607, |
|
"grad_norm": 0.24002987146377563, |
|
"learning_rate": 7.921936177411049e-07, |
|
"loss": 0.0208, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 2.8656716417910446, |
|
"grad_norm": 0.5002749562263489, |
|
"learning_rate": 6.06899905006525e-07, |
|
"loss": 0.0661, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 2.8855721393034823, |
|
"grad_norm": 0.4782179594039917, |
|
"learning_rate": 4.461255922609986e-07, |
|
"loss": 0.088, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 2.9054726368159205, |
|
"grad_norm": 1.0259336233139038, |
|
"learning_rate": 3.0995049003826325e-07, |
|
"loss": 0.2354, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 2.925373134328358, |
|
"grad_norm": 4.005832195281982, |
|
"learning_rate": 1.984421974927375e-07, |
|
"loss": 1.3838, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 2.945273631840796, |
|
"grad_norm": 3.5791056156158447, |
|
"learning_rate": 1.1165606884234181e-07, |
|
"loss": 1.5911, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 2.965174129353234, |
|
"grad_norm": 3.579153299331665, |
|
"learning_rate": 4.963518588996796e-08, |
|
"loss": 1.3113, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 2.9850746268656714, |
|
"grad_norm": 0.6860787868499756, |
|
"learning_rate": 1.2410336637047605e-08, |
|
"loss": 0.1386, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.9850746268656714, |
|
"eval_loss": 1.0078964233398438, |
|
"eval_runtime": 6.3591, |
|
"eval_samples_per_second": 13.367, |
|
"eval_steps_per_second": 3.46, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 3.0049751243781095, |
|
"grad_norm": 7.453361511230469, |
|
"learning_rate": 0.0, |
|
"loss": 2.4948, |
|
"step": 151 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 151, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 1 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.2778555039062426e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|