|
{ |
|
"best_metric": 2.5794575214385986, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-100", |
|
"epoch": 3.017543859649123, |
|
"eval_steps": 50, |
|
"global_step": 129, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.023391812865497075, |
|
"grad_norm": 2.8899965286254883, |
|
"learning_rate": 1e-05, |
|
"loss": 3.621, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.023391812865497075, |
|
"eval_loss": 4.880038738250732, |
|
"eval_runtime": 3.2301, |
|
"eval_samples_per_second": 22.291, |
|
"eval_steps_per_second": 5.573, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.04678362573099415, |
|
"grad_norm": 3.9275195598602295, |
|
"learning_rate": 2e-05, |
|
"loss": 3.9065, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.07017543859649122, |
|
"grad_norm": 5.344070911407471, |
|
"learning_rate": 3e-05, |
|
"loss": 4.0645, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0935672514619883, |
|
"grad_norm": 4.742477893829346, |
|
"learning_rate": 4e-05, |
|
"loss": 3.9164, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.11695906432748537, |
|
"grad_norm": 4.960448265075684, |
|
"learning_rate": 5e-05, |
|
"loss": 4.1087, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.14035087719298245, |
|
"grad_norm": 5.740211009979248, |
|
"learning_rate": 6e-05, |
|
"loss": 3.9335, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.16374269005847952, |
|
"grad_norm": 7.367751598358154, |
|
"learning_rate": 7e-05, |
|
"loss": 3.8536, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.1871345029239766, |
|
"grad_norm": 6.333174228668213, |
|
"learning_rate": 8e-05, |
|
"loss": 3.8627, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.21052631578947367, |
|
"grad_norm": 7.679598808288574, |
|
"learning_rate": 9e-05, |
|
"loss": 3.4869, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.23391812865497075, |
|
"grad_norm": 8.830903053283691, |
|
"learning_rate": 0.0001, |
|
"loss": 3.8319, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.2573099415204678, |
|
"grad_norm": 5.146438121795654, |
|
"learning_rate": 9.998257709344246e-05, |
|
"loss": 3.4504, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.2807017543859649, |
|
"grad_norm": 3.7169127464294434, |
|
"learning_rate": 9.993032051607669e-05, |
|
"loss": 3.3142, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.30409356725146197, |
|
"grad_norm": 3.373077869415283, |
|
"learning_rate": 9.984326668636131e-05, |
|
"loss": 3.1032, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.32748538011695905, |
|
"grad_norm": 3.1468732357025146, |
|
"learning_rate": 9.972147627352592e-05, |
|
"loss": 3.0006, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.3508771929824561, |
|
"grad_norm": 3.2604331970214844, |
|
"learning_rate": 9.956503415528984e-05, |
|
"loss": 3.1792, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.3742690058479532, |
|
"grad_norm": 3.397855758666992, |
|
"learning_rate": 9.937404935870938e-05, |
|
"loss": 3.0764, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.39766081871345027, |
|
"grad_norm": 3.596156120300293, |
|
"learning_rate": 9.91486549841951e-05, |
|
"loss": 2.7991, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.42105263157894735, |
|
"grad_norm": 4.534525394439697, |
|
"learning_rate": 9.888900811275204e-05, |
|
"loss": 2.8614, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.4444444444444444, |
|
"grad_norm": 5.425872802734375, |
|
"learning_rate": 9.859528969650738e-05, |
|
"loss": 2.9683, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.4678362573099415, |
|
"grad_norm": 10.605550765991211, |
|
"learning_rate": 9.826770443260193e-05, |
|
"loss": 3.8376, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.49122807017543857, |
|
"grad_norm": 3.0964808464050293, |
|
"learning_rate": 9.79064806205334e-05, |
|
"loss": 3.1105, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.5146198830409356, |
|
"grad_norm": 2.7125508785247803, |
|
"learning_rate": 9.751187000305076e-05, |
|
"loss": 2.934, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.5380116959064327, |
|
"grad_norm": 2.679877758026123, |
|
"learning_rate": 9.708414759071059e-05, |
|
"loss": 2.8647, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.5614035087719298, |
|
"grad_norm": 2.6572399139404297, |
|
"learning_rate": 9.662361147021779e-05, |
|
"loss": 2.7279, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.5847953216374269, |
|
"grad_norm": 2.8284733295440674, |
|
"learning_rate": 9.613058259668416e-05, |
|
"loss": 2.7198, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.6081871345029239, |
|
"grad_norm": 3.222668409347534, |
|
"learning_rate": 9.56054045699494e-05, |
|
"loss": 2.9284, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.631578947368421, |
|
"grad_norm": 3.3537814617156982, |
|
"learning_rate": 9.504844339512095e-05, |
|
"loss": 2.6431, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.6549707602339181, |
|
"grad_norm": 3.6838319301605225, |
|
"learning_rate": 9.446008722749905e-05, |
|
"loss": 2.7465, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.6783625730994152, |
|
"grad_norm": 4.8847432136535645, |
|
"learning_rate": 9.384074610206495e-05, |
|
"loss": 2.8716, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.7017543859649122, |
|
"grad_norm": 6.4051103591918945, |
|
"learning_rate": 9.319085164772082e-05, |
|
"loss": 2.7997, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.7251461988304093, |
|
"grad_norm": 2.051379680633545, |
|
"learning_rate": 9.251085678648072e-05, |
|
"loss": 2.8797, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.7485380116959064, |
|
"grad_norm": 2.1589481830596924, |
|
"learning_rate": 9.180123541782171e-05, |
|
"loss": 2.5857, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.7719298245614035, |
|
"grad_norm": 2.3289027214050293, |
|
"learning_rate": 9.106248208841569e-05, |
|
"loss": 2.5558, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.7953216374269005, |
|
"grad_norm": 2.580897569656372, |
|
"learning_rate": 9.029511164747175e-05, |
|
"loss": 2.5283, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.8187134502923976, |
|
"grad_norm": 2.79526948928833, |
|
"learning_rate": 8.949965888792941e-05, |
|
"loss": 2.4669, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.8421052631578947, |
|
"grad_norm": 2.9586293697357178, |
|
"learning_rate": 8.867667817375266e-05, |
|
"loss": 2.4342, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.8654970760233918, |
|
"grad_norm": 3.565655469894409, |
|
"learning_rate": 8.78267430535848e-05, |
|
"loss": 2.5569, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.8888888888888888, |
|
"grad_norm": 3.735215425491333, |
|
"learning_rate": 8.695044586103296e-05, |
|
"loss": 2.6734, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.9122807017543859, |
|
"grad_norm": 4.107497692108154, |
|
"learning_rate": 8.604839730186125e-05, |
|
"loss": 2.3497, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.935672514619883, |
|
"grad_norm": 6.719999313354492, |
|
"learning_rate": 8.512122602837993e-05, |
|
"loss": 2.793, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.9590643274853801, |
|
"grad_norm": 1.973287582397461, |
|
"learning_rate": 8.416957820132742e-05, |
|
"loss": 2.8444, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.9824561403508771, |
|
"grad_norm": 2.6278090476989746, |
|
"learning_rate": 8.319411703955042e-05, |
|
"loss": 2.3696, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 1.0058479532163742, |
|
"grad_norm": 4.79380464553833, |
|
"learning_rate": 8.219552235779578e-05, |
|
"loss": 3.6136, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 1.0292397660818713, |
|
"grad_norm": 1.4787054061889648, |
|
"learning_rate": 8.117449009293668e-05, |
|
"loss": 2.1233, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 1.0526315789473684, |
|
"grad_norm": 1.8972117900848389, |
|
"learning_rate": 8.013173181896283e-05, |
|
"loss": 2.4294, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 1.0760233918128654, |
|
"grad_norm": 2.160202980041504, |
|
"learning_rate": 7.9067974251073e-05, |
|
"loss": 2.2788, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 1.0994152046783625, |
|
"grad_norm": 2.476989269256592, |
|
"learning_rate": 7.79839587392154e-05, |
|
"loss": 2.1933, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 1.1228070175438596, |
|
"grad_norm": 2.577780246734619, |
|
"learning_rate": 7.688044075142887e-05, |
|
"loss": 1.8599, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 1.1461988304093567, |
|
"grad_norm": 3.1708245277404785, |
|
"learning_rate": 7.57581893473448e-05, |
|
"loss": 1.9908, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 1.1695906432748537, |
|
"grad_norm": 3.2842891216278076, |
|
"learning_rate": 7.461798664221711e-05, |
|
"loss": 1.838, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.1695906432748537, |
|
"eval_loss": 2.667975425720215, |
|
"eval_runtime": 3.2693, |
|
"eval_samples_per_second": 22.023, |
|
"eval_steps_per_second": 5.506, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.1929824561403508, |
|
"grad_norm": 3.982672691345215, |
|
"learning_rate": 7.346062726185332e-05, |
|
"loss": 1.9851, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 1.2163742690058479, |
|
"grad_norm": 4.517981052398682, |
|
"learning_rate": 7.228691778882693e-05, |
|
"loss": 1.7932, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 1.239766081871345, |
|
"grad_norm": 4.8002166748046875, |
|
"learning_rate": 7.109767620035689e-05, |
|
"loss": 2.3657, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 1.263157894736842, |
|
"grad_norm": 2.333965301513672, |
|
"learning_rate": 6.989373129824604e-05, |
|
"loss": 2.1879, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 1.286549707602339, |
|
"grad_norm": 2.5241196155548096, |
|
"learning_rate": 6.867592213127558e-05, |
|
"loss": 2.2052, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 1.3099415204678362, |
|
"grad_norm": 2.5986971855163574, |
|
"learning_rate": 6.744509741045835e-05, |
|
"loss": 2.0875, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 1.3333333333333333, |
|
"grad_norm": 2.735356569290161, |
|
"learning_rate": 6.62021149175583e-05, |
|
"loss": 2.2641, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 1.3567251461988303, |
|
"grad_norm": 2.861424684524536, |
|
"learning_rate": 6.494784090728852e-05, |
|
"loss": 2.1462, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 1.3801169590643274, |
|
"grad_norm": 2.953516960144043, |
|
"learning_rate": 6.368314950360415e-05, |
|
"loss": 1.9254, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 1.4035087719298245, |
|
"grad_norm": 3.429384231567383, |
|
"learning_rate": 6.240892209051121e-05, |
|
"loss": 2.0428, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.4269005847953216, |
|
"grad_norm": 3.717503070831299, |
|
"learning_rate": 6.112604669781572e-05, |
|
"loss": 1.6045, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 1.4502923976608186, |
|
"grad_norm": 4.62086820602417, |
|
"learning_rate": 5.983541738224141e-05, |
|
"loss": 1.5816, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 1.4736842105263157, |
|
"grad_norm": 5.065677165985107, |
|
"learning_rate": 5.853793360434687e-05, |
|
"loss": 2.1671, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 1.4970760233918128, |
|
"grad_norm": 3.0961179733276367, |
|
"learning_rate": 5.7234499601677026e-05, |
|
"loss": 2.3001, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 1.52046783625731, |
|
"grad_norm": 2.8138766288757324, |
|
"learning_rate": 5.5926023758585146e-05, |
|
"loss": 2.272, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.543859649122807, |
|
"grad_norm": 2.5831410884857178, |
|
"learning_rate": 5.4613417973165106e-05, |
|
"loss": 1.9027, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 1.5672514619883042, |
|
"grad_norm": 3.276082754135132, |
|
"learning_rate": 5.329759702173477e-05, |
|
"loss": 2.3347, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 1.590643274853801, |
|
"grad_norm": 3.1773629188537598, |
|
"learning_rate": 5.197947792131348e-05, |
|
"loss": 1.9598, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 1.6140350877192984, |
|
"grad_norm": 3.4641201496124268, |
|
"learning_rate": 5.0659979290537954e-05, |
|
"loss": 1.9397, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 1.6374269005847952, |
|
"grad_norm": 3.880099058151245, |
|
"learning_rate": 4.934002070946206e-05, |
|
"loss": 1.8497, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.6608187134502925, |
|
"grad_norm": 3.90553879737854, |
|
"learning_rate": 4.802052207868654e-05, |
|
"loss": 1.692, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 1.6842105263157894, |
|
"grad_norm": 5.151843070983887, |
|
"learning_rate": 4.670240297826523e-05, |
|
"loss": 1.7849, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 1.7076023391812867, |
|
"grad_norm": 4.96315336227417, |
|
"learning_rate": 4.5386582026834906e-05, |
|
"loss": 1.9887, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 1.7309941520467835, |
|
"grad_norm": 1.8645200729370117, |
|
"learning_rate": 4.407397624141487e-05, |
|
"loss": 2.2243, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 1.7543859649122808, |
|
"grad_norm": 2.2499165534973145, |
|
"learning_rate": 4.276550039832299e-05, |
|
"loss": 2.24, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.7777777777777777, |
|
"grad_norm": 2.531435251235962, |
|
"learning_rate": 4.146206639565312e-05, |
|
"loss": 2.167, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 1.801169590643275, |
|
"grad_norm": 2.621704578399658, |
|
"learning_rate": 4.01645826177586e-05, |
|
"loss": 2.0048, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 1.8245614035087718, |
|
"grad_norm": 3.041447401046753, |
|
"learning_rate": 3.887395330218429e-05, |
|
"loss": 1.9858, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 1.8479532163742691, |
|
"grad_norm": 3.520282745361328, |
|
"learning_rate": 3.759107790948882e-05, |
|
"loss": 1.9349, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 1.871345029239766, |
|
"grad_norm": 3.5151638984680176, |
|
"learning_rate": 3.631685049639586e-05, |
|
"loss": 1.6562, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.8947368421052633, |
|
"grad_norm": 4.325654983520508, |
|
"learning_rate": 3.505215909271149e-05, |
|
"loss": 1.9125, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 1.9181286549707601, |
|
"grad_norm": 5.122260570526123, |
|
"learning_rate": 3.379788508244171e-05, |
|
"loss": 1.6561, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 1.9415204678362574, |
|
"grad_norm": 5.443163871765137, |
|
"learning_rate": 3.255490258954167e-05, |
|
"loss": 2.5903, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 1.9649122807017543, |
|
"grad_norm": 2.3221733570098877, |
|
"learning_rate": 3.132407786872442e-05, |
|
"loss": 1.9197, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 1.9883040935672516, |
|
"grad_norm": 4.910060405731201, |
|
"learning_rate": 3.0106268701753965e-05, |
|
"loss": 2.2014, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 2.0116959064327484, |
|
"grad_norm": 3.926137685775757, |
|
"learning_rate": 2.8902323799643116e-05, |
|
"loss": 2.227, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 2.0350877192982457, |
|
"grad_norm": 1.8208245038986206, |
|
"learning_rate": 2.771308221117309e-05, |
|
"loss": 1.8213, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 2.0584795321637426, |
|
"grad_norm": 2.0533692836761475, |
|
"learning_rate": 2.6539372738146695e-05, |
|
"loss": 1.6875, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 2.08187134502924, |
|
"grad_norm": 2.5335185527801514, |
|
"learning_rate": 2.5382013357782893e-05, |
|
"loss": 1.7146, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 2.1052631578947367, |
|
"grad_norm": 2.658461093902588, |
|
"learning_rate": 2.4241810652655196e-05, |
|
"loss": 1.6028, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 2.128654970760234, |
|
"grad_norm": 2.784259796142578, |
|
"learning_rate": 2.3119559248571128e-05, |
|
"loss": 1.5149, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 2.152046783625731, |
|
"grad_norm": 3.2446987628936768, |
|
"learning_rate": 2.2016041260784605e-05, |
|
"loss": 1.2587, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 2.175438596491228, |
|
"grad_norm": 3.726407527923584, |
|
"learning_rate": 2.0932025748927013e-05, |
|
"loss": 1.286, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 2.198830409356725, |
|
"grad_norm": 4.698680400848389, |
|
"learning_rate": 1.9868268181037185e-05, |
|
"loss": 1.2943, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 2.2222222222222223, |
|
"grad_norm": 4.862926483154297, |
|
"learning_rate": 1.8825509907063327e-05, |
|
"loss": 0.9368, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 2.245614035087719, |
|
"grad_norm": 4.17417573928833, |
|
"learning_rate": 1.7804477642204222e-05, |
|
"loss": 1.8199, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 2.2690058479532165, |
|
"grad_norm": 3.123291254043579, |
|
"learning_rate": 1.6805882960449594e-05, |
|
"loss": 2.1358, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 2.2923976608187133, |
|
"grad_norm": 3.0750467777252197, |
|
"learning_rate": 1.5830421798672568e-05, |
|
"loss": 1.6731, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 2.3157894736842106, |
|
"grad_norm": 3.460376262664795, |
|
"learning_rate": 1.4878773971620074e-05, |
|
"loss": 1.7389, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 2.3391812865497075, |
|
"grad_norm": 3.5447030067443848, |
|
"learning_rate": 1.3951602698138771e-05, |
|
"loss": 1.5398, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.3391812865497075, |
|
"eval_loss": 2.5794575214385986, |
|
"eval_runtime": 3.2727, |
|
"eval_samples_per_second": 22.0, |
|
"eval_steps_per_second": 5.5, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.3625730994152048, |
|
"grad_norm": 3.9561784267425537, |
|
"learning_rate": 1.3049554138967051e-05, |
|
"loss": 1.5536, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 2.3859649122807016, |
|
"grad_norm": 3.936734914779663, |
|
"learning_rate": 1.2173256946415212e-05, |
|
"loss": 1.4171, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 2.409356725146199, |
|
"grad_norm": 4.206720352172852, |
|
"learning_rate": 1.1323321826247346e-05, |
|
"loss": 1.2999, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 2.4327485380116958, |
|
"grad_norm": 5.046553611755371, |
|
"learning_rate": 1.0500341112070605e-05, |
|
"loss": 1.0988, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 2.456140350877193, |
|
"grad_norm": 5.955111026763916, |
|
"learning_rate": 9.704888352528258e-06, |
|
"loss": 0.9185, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 2.47953216374269, |
|
"grad_norm": 3.576009511947632, |
|
"learning_rate": 8.93751791158432e-06, |
|
"loss": 1.9256, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 2.502923976608187, |
|
"grad_norm": 2.3604319095611572, |
|
"learning_rate": 8.198764582178303e-06, |
|
"loss": 1.9458, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 2.526315789473684, |
|
"grad_norm": 2.6111466884613037, |
|
"learning_rate": 7.489143213519301e-06, |
|
"loss": 1.7967, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 2.5497076023391814, |
|
"grad_norm": 3.0367538928985596, |
|
"learning_rate": 6.809148352279182e-06, |
|
"loss": 1.4832, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 2.573099415204678, |
|
"grad_norm": 3.416107416152954, |
|
"learning_rate": 6.159253897935069e-06, |
|
"loss": 1.6191, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 2.5964912280701755, |
|
"grad_norm": 3.648995876312256, |
|
"learning_rate": 5.539912772500944e-06, |
|
"loss": 1.4337, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 2.6198830409356724, |
|
"grad_norm": 3.623403310775757, |
|
"learning_rate": 4.951556604879048e-06, |
|
"loss": 1.1803, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 2.6432748538011697, |
|
"grad_norm": 4.428551197052002, |
|
"learning_rate": 4.394595430050613e-06, |
|
"loss": 1.1911, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 2.6666666666666665, |
|
"grad_norm": 4.538227558135986, |
|
"learning_rate": 3.8694174033158555e-06, |
|
"loss": 1.0355, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 2.690058479532164, |
|
"grad_norm": 5.722489833831787, |
|
"learning_rate": 3.376388529782215e-06, |
|
"loss": 0.9762, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 2.7134502923976607, |
|
"grad_norm": 3.4204437732696533, |
|
"learning_rate": 2.9158524092894214e-06, |
|
"loss": 1.7561, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 2.736842105263158, |
|
"grad_norm": 2.083275079727173, |
|
"learning_rate": 2.488129996949251e-06, |
|
"loss": 1.8624, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 2.760233918128655, |
|
"grad_norm": 2.574385404586792, |
|
"learning_rate": 2.093519379466602e-06, |
|
"loss": 1.7507, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 2.783625730994152, |
|
"grad_norm": 2.773837089538574, |
|
"learning_rate": 1.7322955673980678e-06, |
|
"loss": 1.6317, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 2.807017543859649, |
|
"grad_norm": 3.0657918453216553, |
|
"learning_rate": 1.404710303492618e-06, |
|
"loss": 1.7106, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.8304093567251463, |
|
"grad_norm": 3.1810173988342285, |
|
"learning_rate": 1.1109918872479642e-06, |
|
"loss": 1.4181, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 2.853801169590643, |
|
"grad_norm": 3.865628719329834, |
|
"learning_rate": 8.513450158049108e-07, |
|
"loss": 1.4733, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 2.8771929824561404, |
|
"grad_norm": 3.952064275741577, |
|
"learning_rate": 6.259506412906402e-07, |
|
"loss": 1.1925, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 2.9005847953216373, |
|
"grad_norm": 4.404904842376709, |
|
"learning_rate": 4.349658447101612e-07, |
|
"loss": 1.0374, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 2.9239766081871346, |
|
"grad_norm": 5.374385356903076, |
|
"learning_rate": 2.785237264740781e-07, |
|
"loss": 0.8663, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.9473684210526314, |
|
"grad_norm": 4.283792018890381, |
|
"learning_rate": 1.567333136387017e-07, |
|
"loss": 1.8765, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 2.9707602339181287, |
|
"grad_norm": 3.530892848968506, |
|
"learning_rate": 6.967948392331835e-08, |
|
"loss": 1.4727, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 2.9941520467836256, |
|
"grad_norm": 6.096131324768066, |
|
"learning_rate": 1.742290655755707e-08, |
|
"loss": 1.1021, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 3.017543859649123, |
|
"grad_norm": 3.1924569606781006, |
|
"learning_rate": 0.0, |
|
"loss": 2.1947, |
|
"step": 129 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 129, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 5.386784338870272e+16, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|