|
{ |
|
"best_metric": 0.8302884697914124, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-450", |
|
"epoch": 0.08468197214904027, |
|
"eval_steps": 150, |
|
"global_step": 450, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0001881821603312006, |
|
"grad_norm": 8.021297454833984, |
|
"learning_rate": 5e-06, |
|
"loss": 4.7627, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0001881821603312006, |
|
"eval_loss": 1.206099271774292, |
|
"eval_runtime": 977.2207, |
|
"eval_samples_per_second": 9.159, |
|
"eval_steps_per_second": 2.29, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0003763643206624012, |
|
"grad_norm": 8.330506324768066, |
|
"learning_rate": 1e-05, |
|
"loss": 4.7186, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0005645464809936018, |
|
"grad_norm": 7.244218826293945, |
|
"learning_rate": 1.5e-05, |
|
"loss": 4.7416, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0007527286413248024, |
|
"grad_norm": 7.7417449951171875, |
|
"learning_rate": 2e-05, |
|
"loss": 4.8388, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.000940910801656003, |
|
"grad_norm": 7.4192376136779785, |
|
"learning_rate": 2.5e-05, |
|
"loss": 4.4709, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0011290929619872036, |
|
"grad_norm": 4.583150386810303, |
|
"learning_rate": 3e-05, |
|
"loss": 4.305, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0013172751223184042, |
|
"grad_norm": 3.533625602722168, |
|
"learning_rate": 3.5e-05, |
|
"loss": 4.0565, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0015054572826496049, |
|
"grad_norm": 4.560721397399902, |
|
"learning_rate": 4e-05, |
|
"loss": 4.0204, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.0016936394429808055, |
|
"grad_norm": 13.375167846679688, |
|
"learning_rate": 4.5e-05, |
|
"loss": 4.3096, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.001881821603312006, |
|
"grad_norm": 6.8327813148498535, |
|
"learning_rate": 5e-05, |
|
"loss": 4.279, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0020700037636432068, |
|
"grad_norm": 5.540746688842773, |
|
"learning_rate": 5.500000000000001e-05, |
|
"loss": 4.2908, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.002258185923974407, |
|
"grad_norm": 4.140532970428467, |
|
"learning_rate": 6e-05, |
|
"loss": 4.2689, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.0024463680843056076, |
|
"grad_norm": 1.9986830949783325, |
|
"learning_rate": 6.500000000000001e-05, |
|
"loss": 4.0747, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0026345502446368085, |
|
"grad_norm": 1.936602234840393, |
|
"learning_rate": 7e-05, |
|
"loss": 3.956, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.002822732404968009, |
|
"grad_norm": 1.9803358316421509, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 4.0274, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0030109145652992097, |
|
"grad_norm": 2.102954864501953, |
|
"learning_rate": 8e-05, |
|
"loss": 3.9315, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.00319909672563041, |
|
"grad_norm": 1.8604391813278198, |
|
"learning_rate": 8.5e-05, |
|
"loss": 3.9579, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.003387278885961611, |
|
"grad_norm": 5.816336631774902, |
|
"learning_rate": 9e-05, |
|
"loss": 3.8109, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.0035754610462928114, |
|
"grad_norm": 1.8846168518066406, |
|
"learning_rate": 9.5e-05, |
|
"loss": 3.8936, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.003763643206624012, |
|
"grad_norm": 2.1951682567596436, |
|
"learning_rate": 0.0001, |
|
"loss": 3.8866, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.003951825366955213, |
|
"grad_norm": 1.8232979774475098, |
|
"learning_rate": 9.999866555428618e-05, |
|
"loss": 3.7866, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.0041400075272864136, |
|
"grad_norm": 1.7551851272583008, |
|
"learning_rate": 9.999466228837451e-05, |
|
"loss": 3.848, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.0043281896876176135, |
|
"grad_norm": 3.396578788757324, |
|
"learning_rate": 9.998799041595064e-05, |
|
"loss": 3.8432, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.004516371847948814, |
|
"grad_norm": 1.9480079412460327, |
|
"learning_rate": 9.997865029314463e-05, |
|
"loss": 3.9188, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.004704554008280015, |
|
"grad_norm": 1.8144193887710571, |
|
"learning_rate": 9.996664241851197e-05, |
|
"loss": 3.9241, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.004892736168611215, |
|
"grad_norm": 1.6157795190811157, |
|
"learning_rate": 9.995196743300692e-05, |
|
"loss": 3.7515, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.005080918328942416, |
|
"grad_norm": 1.8001606464385986, |
|
"learning_rate": 9.993462611994832e-05, |
|
"loss": 3.8946, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.005269100489273617, |
|
"grad_norm": 2.1896607875823975, |
|
"learning_rate": 9.991461940497786e-05, |
|
"loss": 3.6935, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.005457282649604818, |
|
"grad_norm": 1.9953135251998901, |
|
"learning_rate": 9.989194835601048e-05, |
|
"loss": 3.6832, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.005645464809936018, |
|
"grad_norm": 1.9537560939788818, |
|
"learning_rate": 9.986661418317759e-05, |
|
"loss": 3.6689, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.005833646970267219, |
|
"grad_norm": 1.7720470428466797, |
|
"learning_rate": 9.983861823876231e-05, |
|
"loss": 3.8731, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.0060218291305984195, |
|
"grad_norm": 1.8037577867507935, |
|
"learning_rate": 9.980796201712734e-05, |
|
"loss": 3.7388, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.0062100112909296195, |
|
"grad_norm": 1.558807611465454, |
|
"learning_rate": 9.977464715463524e-05, |
|
"loss": 3.5732, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.00639819345126082, |
|
"grad_norm": 1.7031415700912476, |
|
"learning_rate": 9.973867542956104e-05, |
|
"loss": 3.7046, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.006586375611592021, |
|
"grad_norm": 1.8096400499343872, |
|
"learning_rate": 9.97000487619973e-05, |
|
"loss": 3.7951, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.006774557771923222, |
|
"grad_norm": 1.6755597591400146, |
|
"learning_rate": 9.965876921375165e-05, |
|
"loss": 3.7345, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.006962739932254422, |
|
"grad_norm": 1.6606807708740234, |
|
"learning_rate": 9.961483898823678e-05, |
|
"loss": 3.6923, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.007150922092585623, |
|
"grad_norm": 1.602203130722046, |
|
"learning_rate": 9.956826043035268e-05, |
|
"loss": 3.6913, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.007339104252916824, |
|
"grad_norm": 1.6571152210235596, |
|
"learning_rate": 9.951903602636166e-05, |
|
"loss": 3.6178, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.007527286413248024, |
|
"grad_norm": 1.737025499343872, |
|
"learning_rate": 9.946716840375551e-05, |
|
"loss": 3.6084, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0077154685735792245, |
|
"grad_norm": 1.6381107568740845, |
|
"learning_rate": 9.94126603311153e-05, |
|
"loss": 3.4131, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.007903650733910425, |
|
"grad_norm": 1.668062686920166, |
|
"learning_rate": 9.935551471796358e-05, |
|
"loss": 3.4251, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.008091832894241625, |
|
"grad_norm": 1.6551685333251953, |
|
"learning_rate": 9.92957346146091e-05, |
|
"loss": 3.4862, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.008280015054572827, |
|
"grad_norm": 1.5389114618301392, |
|
"learning_rate": 9.923332321198395e-05, |
|
"loss": 3.3558, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.008468197214904027, |
|
"grad_norm": 1.6278464794158936, |
|
"learning_rate": 9.916828384147331e-05, |
|
"loss": 3.6091, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.008656379375235227, |
|
"grad_norm": 1.642901062965393, |
|
"learning_rate": 9.910061997473752e-05, |
|
"loss": 3.3503, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.008844561535566429, |
|
"grad_norm": 2.450819730758667, |
|
"learning_rate": 9.903033522352687e-05, |
|
"loss": 3.7451, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.009032743695897629, |
|
"grad_norm": 1.6337559223175049, |
|
"learning_rate": 9.895743333948874e-05, |
|
"loss": 3.7759, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.009220925856228829, |
|
"grad_norm": 2.088834285736084, |
|
"learning_rate": 9.888191821396744e-05, |
|
"loss": 3.8653, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.00940910801656003, |
|
"grad_norm": 2.713131904602051, |
|
"learning_rate": 9.880379387779637e-05, |
|
"loss": 3.468, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.00959729017689123, |
|
"grad_norm": 1.8180787563323975, |
|
"learning_rate": 9.872306450108292e-05, |
|
"loss": 3.5826, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.00978547233722243, |
|
"grad_norm": 1.6677582263946533, |
|
"learning_rate": 9.863973439298597e-05, |
|
"loss": 3.7982, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.009973654497553632, |
|
"grad_norm": 1.569620966911316, |
|
"learning_rate": 9.855380800148572e-05, |
|
"loss": 3.7162, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.010161836657884832, |
|
"grad_norm": 1.6128700971603394, |
|
"learning_rate": 9.846528991314639e-05, |
|
"loss": 3.5721, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.010350018818216034, |
|
"grad_norm": 1.8484185934066772, |
|
"learning_rate": 9.837418485287127e-05, |
|
"loss": 3.9133, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.010538200978547234, |
|
"grad_norm": 1.706660270690918, |
|
"learning_rate": 9.828049768365068e-05, |
|
"loss": 3.7007, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.010726383138878434, |
|
"grad_norm": 1.7316683530807495, |
|
"learning_rate": 9.818423340630228e-05, |
|
"loss": 3.7366, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.010914565299209636, |
|
"grad_norm": 1.5684834718704224, |
|
"learning_rate": 9.808539715920414e-05, |
|
"loss": 3.6784, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.011102747459540836, |
|
"grad_norm": 1.5667593479156494, |
|
"learning_rate": 9.798399421802056e-05, |
|
"loss": 3.7003, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.011290929619872036, |
|
"grad_norm": 1.526808500289917, |
|
"learning_rate": 9.78800299954203e-05, |
|
"loss": 3.7078, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.011479111780203237, |
|
"grad_norm": 2.5877087116241455, |
|
"learning_rate": 9.777351004078783e-05, |
|
"loss": 3.6875, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.011667293940534437, |
|
"grad_norm": 1.545409917831421, |
|
"learning_rate": 9.766444003992703e-05, |
|
"loss": 3.8299, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.011855476100865637, |
|
"grad_norm": 1.5253973007202148, |
|
"learning_rate": 9.755282581475769e-05, |
|
"loss": 3.6405, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.012043658261196839, |
|
"grad_norm": 1.8160947561264038, |
|
"learning_rate": 9.743867332300478e-05, |
|
"loss": 3.7575, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.012231840421528039, |
|
"grad_norm": 1.62673819065094, |
|
"learning_rate": 9.732198865788047e-05, |
|
"loss": 3.736, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.012420022581859239, |
|
"grad_norm": 1.609445571899414, |
|
"learning_rate": 9.72027780477588e-05, |
|
"loss": 3.5029, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.01260820474219044, |
|
"grad_norm": 1.640405535697937, |
|
"learning_rate": 9.708104785584323e-05, |
|
"loss": 3.6871, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.01279638690252164, |
|
"grad_norm": 1.546713948249817, |
|
"learning_rate": 9.695680457982713e-05, |
|
"loss": 3.5635, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.012984569062852842, |
|
"grad_norm": 1.525071620941162, |
|
"learning_rate": 9.683005485154677e-05, |
|
"loss": 3.6238, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.013172751223184042, |
|
"grad_norm": 1.5092836618423462, |
|
"learning_rate": 9.67008054366274e-05, |
|
"loss": 3.6092, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.013360933383515242, |
|
"grad_norm": 1.6913105249404907, |
|
"learning_rate": 9.656906323412217e-05, |
|
"loss": 3.6286, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.013549115543846444, |
|
"grad_norm": 2.151688814163208, |
|
"learning_rate": 9.643483527614372e-05, |
|
"loss": 3.5817, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.013737297704177644, |
|
"grad_norm": 1.5763871669769287, |
|
"learning_rate": 9.629812872748901e-05, |
|
"loss": 3.5043, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.013925479864508844, |
|
"grad_norm": 1.5946530103683472, |
|
"learning_rate": 9.615895088525677e-05, |
|
"loss": 3.5902, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.014113662024840046, |
|
"grad_norm": 1.5125662088394165, |
|
"learning_rate": 9.601730917845797e-05, |
|
"loss": 3.5002, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.014301844185171246, |
|
"grad_norm": 1.952149510383606, |
|
"learning_rate": 9.587321116761938e-05, |
|
"loss": 3.5881, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.014490026345502446, |
|
"grad_norm": 1.5184606313705444, |
|
"learning_rate": 9.57266645443799e-05, |
|
"loss": 3.3213, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.014678208505833647, |
|
"grad_norm": 1.5643829107284546, |
|
"learning_rate": 9.557767713108009e-05, |
|
"loss": 3.7112, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.014866390666164847, |
|
"grad_norm": 1.5019704103469849, |
|
"learning_rate": 9.542625688034449e-05, |
|
"loss": 3.4665, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.015054572826496047, |
|
"grad_norm": 1.647053837776184, |
|
"learning_rate": 9.527241187465734e-05, |
|
"loss": 3.33, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.015242754986827249, |
|
"grad_norm": 1.527642846107483, |
|
"learning_rate": 9.511615032593096e-05, |
|
"loss": 3.6355, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.015430937147158449, |
|
"grad_norm": 1.6316813230514526, |
|
"learning_rate": 9.49574805750675e-05, |
|
"loss": 3.5475, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.01561911930748965, |
|
"grad_norm": 1.6150362491607666, |
|
"learning_rate": 9.479641109151373e-05, |
|
"loss": 3.5272, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.01580730146782085, |
|
"grad_norm": 1.5907738208770752, |
|
"learning_rate": 9.463295047280891e-05, |
|
"loss": 3.6044, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.015995483628152053, |
|
"grad_norm": 1.5602425336837769, |
|
"learning_rate": 9.446710744412595e-05, |
|
"loss": 3.4846, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.01618366578848325, |
|
"grad_norm": 1.5599199533462524, |
|
"learning_rate": 9.429889085780557e-05, |
|
"loss": 3.598, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.016371847948814452, |
|
"grad_norm": 1.4632290601730347, |
|
"learning_rate": 9.41283096928839e-05, |
|
"loss": 3.3955, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.016560030109145654, |
|
"grad_norm": 1.5269198417663574, |
|
"learning_rate": 9.395537305461311e-05, |
|
"loss": 3.5423, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.016748212269476852, |
|
"grad_norm": 1.6142536401748657, |
|
"learning_rate": 9.378009017397542e-05, |
|
"loss": 3.5366, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.016936394429808054, |
|
"grad_norm": 1.4717563390731812, |
|
"learning_rate": 9.360247040719039e-05, |
|
"loss": 3.4092, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.017124576590139256, |
|
"grad_norm": 2.3025355339050293, |
|
"learning_rate": 9.342252323521545e-05, |
|
"loss": 3.441, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.017312758750470454, |
|
"grad_norm": 1.6364383697509766, |
|
"learning_rate": 9.324025826323994e-05, |
|
"loss": 3.4081, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.017500940910801656, |
|
"grad_norm": 1.6637694835662842, |
|
"learning_rate": 9.305568522017227e-05, |
|
"loss": 3.4648, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.017689123071132858, |
|
"grad_norm": 1.5033453702926636, |
|
"learning_rate": 9.286881395812066e-05, |
|
"loss": 3.4444, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.017877305231464056, |
|
"grad_norm": 1.7990686893463135, |
|
"learning_rate": 9.267965445186733e-05, |
|
"loss": 3.4532, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.018065487391795258, |
|
"grad_norm": 1.5659180879592896, |
|
"learning_rate": 9.248821679833596e-05, |
|
"loss": 3.444, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.01825366955212646, |
|
"grad_norm": 1.495245337486267, |
|
"learning_rate": 9.229451121605279e-05, |
|
"loss": 3.5251, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.018441851712457658, |
|
"grad_norm": 1.6818476915359497, |
|
"learning_rate": 9.209854804460121e-05, |
|
"loss": 3.323, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.01863003387278886, |
|
"grad_norm": 1.9330247640609741, |
|
"learning_rate": 9.190033774406977e-05, |
|
"loss": 3.4345, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.01881821603312006, |
|
"grad_norm": 2.6322007179260254, |
|
"learning_rate": 9.16998908944939e-05, |
|
"loss": 3.2795, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01900639819345126, |
|
"grad_norm": 1.6533123254776, |
|
"learning_rate": 9.149721819529119e-05, |
|
"loss": 3.6009, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.01919458035378246, |
|
"grad_norm": 1.739073395729065, |
|
"learning_rate": 9.129233046469022e-05, |
|
"loss": 3.8017, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.019382762514113663, |
|
"grad_norm": 1.728940486907959, |
|
"learning_rate": 9.108523863915314e-05, |
|
"loss": 3.7084, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.01957094467444486, |
|
"grad_norm": 1.6274477243423462, |
|
"learning_rate": 9.087595377279192e-05, |
|
"loss": 3.5057, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.019759126834776063, |
|
"grad_norm": 1.6262691020965576, |
|
"learning_rate": 9.066448703677828e-05, |
|
"loss": 3.6988, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.019947308995107264, |
|
"grad_norm": 4.310033321380615, |
|
"learning_rate": 9.045084971874738e-05, |
|
"loss": 3.7893, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.020135491155438466, |
|
"grad_norm": 6.6696977615356445, |
|
"learning_rate": 9.023505322219536e-05, |
|
"loss": 3.5515, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.020323673315769664, |
|
"grad_norm": 1.6663572788238525, |
|
"learning_rate": 9.001710906587064e-05, |
|
"loss": 3.4112, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.020511855476100866, |
|
"grad_norm": 1.6214839220046997, |
|
"learning_rate": 8.9797028883159e-05, |
|
"loss": 3.3542, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.020700037636432068, |
|
"grad_norm": 1.540847897529602, |
|
"learning_rate": 8.957482442146272e-05, |
|
"loss": 3.6587, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.020888219796763266, |
|
"grad_norm": 1.5367134809494019, |
|
"learning_rate": 8.935050754157344e-05, |
|
"loss": 3.5199, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.021076401957094468, |
|
"grad_norm": 1.7050776481628418, |
|
"learning_rate": 8.912409021703913e-05, |
|
"loss": 3.6533, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.02126458411742567, |
|
"grad_norm": 1.5632902383804321, |
|
"learning_rate": 8.889558453352492e-05, |
|
"loss": 3.5123, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.021452766277756868, |
|
"grad_norm": 1.5934125185012817, |
|
"learning_rate": 8.866500268816803e-05, |
|
"loss": 3.5423, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.02164094843808807, |
|
"grad_norm": 1.5730620622634888, |
|
"learning_rate": 8.84323569889266e-05, |
|
"loss": 3.3918, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.02182913059841927, |
|
"grad_norm": 1.5384756326675415, |
|
"learning_rate": 8.819765985392296e-05, |
|
"loss": 3.5834, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.02201731275875047, |
|
"grad_norm": 1.4878556728363037, |
|
"learning_rate": 8.79609238107805e-05, |
|
"loss": 3.5273, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.02220549491908167, |
|
"grad_norm": 1.8022403717041016, |
|
"learning_rate": 8.772216149595513e-05, |
|
"loss": 3.6865, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.022393677079412873, |
|
"grad_norm": 1.6074386835098267, |
|
"learning_rate": 8.748138565406081e-05, |
|
"loss": 3.35, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.02258185923974407, |
|
"grad_norm": 5.877849102020264, |
|
"learning_rate": 8.72386091371891e-05, |
|
"loss": 3.5387, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.022770041400075273, |
|
"grad_norm": 1.6074655055999756, |
|
"learning_rate": 8.699384490422331e-05, |
|
"loss": 3.4726, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.022958223560406475, |
|
"grad_norm": 1.5693819522857666, |
|
"learning_rate": 8.674710602014671e-05, |
|
"loss": 3.4279, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.023146405720737673, |
|
"grad_norm": 1.5131280422210693, |
|
"learning_rate": 8.649840565534513e-05, |
|
"loss": 3.4739, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.023334587881068874, |
|
"grad_norm": 1.5035758018493652, |
|
"learning_rate": 8.624775708490402e-05, |
|
"loss": 3.4492, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.023522770041400076, |
|
"grad_norm": 1.6009769439697266, |
|
"learning_rate": 8.59951736878998e-05, |
|
"loss": 3.6759, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.023710952201731274, |
|
"grad_norm": 1.539108395576477, |
|
"learning_rate": 8.574066894668573e-05, |
|
"loss": 3.5554, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.023899134362062476, |
|
"grad_norm": 1.8990849256515503, |
|
"learning_rate": 8.548425644617224e-05, |
|
"loss": 3.4451, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.024087316522393678, |
|
"grad_norm": 1.5071018934249878, |
|
"learning_rate": 8.522594987310184e-05, |
|
"loss": 3.4815, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.024275498682724876, |
|
"grad_norm": 1.5156155824661255, |
|
"learning_rate": 8.49657630153185e-05, |
|
"loss": 3.526, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.024463680843056078, |
|
"grad_norm": 1.4841188192367554, |
|
"learning_rate": 8.47037097610317e-05, |
|
"loss": 3.4503, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.02465186300338728, |
|
"grad_norm": 1.8373347520828247, |
|
"learning_rate": 8.443980409807512e-05, |
|
"loss": 3.4492, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.024840045163718478, |
|
"grad_norm": 1.9632869958877563, |
|
"learning_rate": 8.417406011315998e-05, |
|
"loss": 3.5533, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.02502822732404968, |
|
"grad_norm": 1.525312900543213, |
|
"learning_rate": 8.390649199112315e-05, |
|
"loss": 3.5785, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.02521640948438088, |
|
"grad_norm": 1.7048442363739014, |
|
"learning_rate": 8.363711401417e-05, |
|
"loss": 3.5644, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.025404591644712083, |
|
"grad_norm": 1.4897756576538086, |
|
"learning_rate": 8.336594056111197e-05, |
|
"loss": 3.4262, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.02559277380504328, |
|
"grad_norm": 1.5408360958099365, |
|
"learning_rate": 8.309298610659916e-05, |
|
"loss": 3.5162, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.025780955965374483, |
|
"grad_norm": 1.613747239112854, |
|
"learning_rate": 8.281826522034764e-05, |
|
"loss": 3.6368, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.025969138125705685, |
|
"grad_norm": 1.5057644844055176, |
|
"learning_rate": 8.254179256636179e-05, |
|
"loss": 3.4841, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.026157320286036883, |
|
"grad_norm": 1.5167312622070312, |
|
"learning_rate": 8.226358290215151e-05, |
|
"loss": 3.3391, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.026345502446368085, |
|
"grad_norm": 1.5216038227081299, |
|
"learning_rate": 8.198365107794457e-05, |
|
"loss": 3.3654, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.026533684606699286, |
|
"grad_norm": 1.5505784749984741, |
|
"learning_rate": 8.17020120358939e-05, |
|
"loss": 3.461, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.026721866767030485, |
|
"grad_norm": 3.795452833175659, |
|
"learning_rate": 8.141868080927996e-05, |
|
"loss": 3.4515, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.026910048927361686, |
|
"grad_norm": 1.4928878545761108, |
|
"learning_rate": 8.113367252170844e-05, |
|
"loss": 3.4339, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.027098231087692888, |
|
"grad_norm": 1.5022090673446655, |
|
"learning_rate": 8.084700238630283e-05, |
|
"loss": 3.3639, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.027286413248024086, |
|
"grad_norm": 1.4917182922363281, |
|
"learning_rate": 8.055868570489247e-05, |
|
"loss": 3.2841, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.027474595408355288, |
|
"grad_norm": 1.5091995000839233, |
|
"learning_rate": 8.026873786719573e-05, |
|
"loss": 3.4067, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.02766277756868649, |
|
"grad_norm": 1.455728530883789, |
|
"learning_rate": 7.997717434999861e-05, |
|
"loss": 3.2996, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.027850959729017688, |
|
"grad_norm": 1.852761149406433, |
|
"learning_rate": 7.968401071632855e-05, |
|
"loss": 3.4227, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.02803914188934889, |
|
"grad_norm": 1.8084666728973389, |
|
"learning_rate": 7.938926261462366e-05, |
|
"loss": 3.4068, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.02822732404968009, |
|
"grad_norm": 2.763946056365967, |
|
"learning_rate": 7.909294577789766e-05, |
|
"loss": 3.1156, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.02822732404968009, |
|
"eval_loss": 0.8662680983543396, |
|
"eval_runtime": 982.7503, |
|
"eval_samples_per_second": 9.107, |
|
"eval_steps_per_second": 2.277, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.02841550621001129, |
|
"grad_norm": 1.5687919855117798, |
|
"learning_rate": 7.879507602289979e-05, |
|
"loss": 3.6073, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.02860368837034249, |
|
"grad_norm": 1.6533427238464355, |
|
"learning_rate": 7.849566924927082e-05, |
|
"loss": 3.5083, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.028791870530673693, |
|
"grad_norm": 1.542561650276184, |
|
"learning_rate": 7.819474143869414e-05, |
|
"loss": 3.4967, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.02898005269100489, |
|
"grad_norm": 1.516094446182251, |
|
"learning_rate": 7.789230865404287e-05, |
|
"loss": 3.3608, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.029168234851336093, |
|
"grad_norm": 1.6348127126693726, |
|
"learning_rate": 7.75883870385223e-05, |
|
"loss": 3.5607, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.029356417011667295, |
|
"grad_norm": 1.574805498123169, |
|
"learning_rate": 7.728299281480833e-05, |
|
"loss": 3.314, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.029544599171998493, |
|
"grad_norm": 1.5510149002075195, |
|
"learning_rate": 7.697614228418148e-05, |
|
"loss": 3.397, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.029732781332329695, |
|
"grad_norm": 1.4977482557296753, |
|
"learning_rate": 7.666785182565677e-05, |
|
"loss": 3.4013, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.029920963492660897, |
|
"grad_norm": 1.5679925680160522, |
|
"learning_rate": 7.635813789510941e-05, |
|
"loss": 3.3312, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.030109145652992095, |
|
"grad_norm": 1.5352201461791992, |
|
"learning_rate": 7.604701702439651e-05, |
|
"loss": 3.483, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.030297327813323296, |
|
"grad_norm": 1.52545964717865, |
|
"learning_rate": 7.573450582047457e-05, |
|
"loss": 3.5098, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.030485509973654498, |
|
"grad_norm": 1.5153985023498535, |
|
"learning_rate": 7.542062096451305e-05, |
|
"loss": 3.4748, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.030673692133985696, |
|
"grad_norm": 1.4865204095840454, |
|
"learning_rate": 7.510537921100398e-05, |
|
"loss": 3.3997, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.030861874294316898, |
|
"grad_norm": 1.5080010890960693, |
|
"learning_rate": 7.47887973868676e-05, |
|
"loss": 3.5227, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.0310500564546481, |
|
"grad_norm": 1.5261732339859009, |
|
"learning_rate": 7.447089239055428e-05, |
|
"loss": 3.7215, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.0312382386149793, |
|
"grad_norm": 1.4631050825119019, |
|
"learning_rate": 7.41516811911424e-05, |
|
"loss": 3.3073, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.0314264207753105, |
|
"grad_norm": 1.5799752473831177, |
|
"learning_rate": 7.383118082743262e-05, |
|
"loss": 3.5134, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.0316146029356417, |
|
"grad_norm": 1.6088528633117676, |
|
"learning_rate": 7.350940840703842e-05, |
|
"loss": 3.4344, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.0318027850959729, |
|
"grad_norm": 1.5857129096984863, |
|
"learning_rate": 7.318638110547288e-05, |
|
"loss": 3.6729, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.031990967256304105, |
|
"grad_norm": 1.9101580381393433, |
|
"learning_rate": 7.286211616523193e-05, |
|
"loss": 3.4667, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.0321791494166353, |
|
"grad_norm": 1.6800731420516968, |
|
"learning_rate": 7.253663089487395e-05, |
|
"loss": 3.3363, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.0323673315769665, |
|
"grad_norm": 1.5496375560760498, |
|
"learning_rate": 7.220994266809591e-05, |
|
"loss": 3.5252, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.0325555137372977, |
|
"grad_norm": 1.4972703456878662, |
|
"learning_rate": 7.188206892280594e-05, |
|
"loss": 3.43, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.032743695897628905, |
|
"grad_norm": 1.4559191465377808, |
|
"learning_rate": 7.155302716019263e-05, |
|
"loss": 3.2506, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.03293187805796011, |
|
"grad_norm": 1.6048904657363892, |
|
"learning_rate": 7.122283494379076e-05, |
|
"loss": 3.6594, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.03312006021829131, |
|
"grad_norm": 1.5769932270050049, |
|
"learning_rate": 7.089150989854385e-05, |
|
"loss": 3.3905, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.0333082423786225, |
|
"grad_norm": 1.508113980293274, |
|
"learning_rate": 7.055906970986336e-05, |
|
"loss": 3.4611, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.033496424538953705, |
|
"grad_norm": 1.4894671440124512, |
|
"learning_rate": 7.022553212268469e-05, |
|
"loss": 3.3249, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.03368460669928491, |
|
"grad_norm": 1.5143218040466309, |
|
"learning_rate": 6.989091494051998e-05, |
|
"loss": 3.3934, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.03387278885961611, |
|
"grad_norm": 1.4756419658660889, |
|
"learning_rate": 6.95552360245078e-05, |
|
"loss": 3.2264, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.03406097101994731, |
|
"grad_norm": 1.5053011178970337, |
|
"learning_rate": 6.92185132924598e-05, |
|
"loss": 3.5413, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.03424915318027851, |
|
"grad_norm": 1.8715499639511108, |
|
"learning_rate": 6.888076471790424e-05, |
|
"loss": 3.4045, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.034437335340609714, |
|
"grad_norm": 1.8220783472061157, |
|
"learning_rate": 6.85420083291266e-05, |
|
"loss": 3.4066, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.03462551750094091, |
|
"grad_norm": 1.5312058925628662, |
|
"learning_rate": 6.820226220820732e-05, |
|
"loss": 3.474, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.03481369966127211, |
|
"grad_norm": 1.5189887285232544, |
|
"learning_rate": 6.786154449005665e-05, |
|
"loss": 3.2858, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.03500188182160331, |
|
"grad_norm": 1.5011916160583496, |
|
"learning_rate": 6.751987336144648e-05, |
|
"loss": 3.3292, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.03519006398193451, |
|
"grad_norm": 1.6477534770965576, |
|
"learning_rate": 6.717726706003974e-05, |
|
"loss": 3.4825, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.035378246142265715, |
|
"grad_norm": 1.5640369653701782, |
|
"learning_rate": 6.683374387341687e-05, |
|
"loss": 3.3886, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.03556642830259692, |
|
"grad_norm": 1.5490057468414307, |
|
"learning_rate": 6.648932213809962e-05, |
|
"loss": 3.4445, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.03575461046292811, |
|
"grad_norm": 1.5441166162490845, |
|
"learning_rate": 6.614402023857232e-05, |
|
"loss": 3.2839, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.03594279262325931, |
|
"grad_norm": 1.5490269660949707, |
|
"learning_rate": 6.579785660630056e-05, |
|
"loss": 3.548, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.036130974783590515, |
|
"grad_norm": 1.4607436656951904, |
|
"learning_rate": 6.545084971874738e-05, |
|
"loss": 3.3022, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.03631915694392172, |
|
"grad_norm": 1.5263044834136963, |
|
"learning_rate": 6.510301809838689e-05, |
|
"loss": 3.5151, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.03650733910425292, |
|
"grad_norm": 1.7463645935058594, |
|
"learning_rate": 6.475438031171574e-05, |
|
"loss": 3.4085, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.03669552126458412, |
|
"grad_norm": 1.4995930194854736, |
|
"learning_rate": 6.440495496826189e-05, |
|
"loss": 3.4332, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.036883703424915315, |
|
"grad_norm": 1.5385165214538574, |
|
"learning_rate": 6.405476071959143e-05, |
|
"loss": 3.432, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.03707188558524652, |
|
"grad_norm": 1.5226582288742065, |
|
"learning_rate": 6.370381625831292e-05, |
|
"loss": 3.3812, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.03726006774557772, |
|
"grad_norm": 1.5849664211273193, |
|
"learning_rate": 6.335214031707965e-05, |
|
"loss": 3.4423, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.03744824990590892, |
|
"grad_norm": 1.8131951093673706, |
|
"learning_rate": 6.299975166758971e-05, |
|
"loss": 3.3169, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.03763643206624012, |
|
"grad_norm": 2.523836851119995, |
|
"learning_rate": 6.264666911958404e-05, |
|
"loss": 3.1207, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.037824614226571324, |
|
"grad_norm": 1.6073590517044067, |
|
"learning_rate": 6.229291151984233e-05, |
|
"loss": 3.5545, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.03801279638690252, |
|
"grad_norm": 3.597961187362671, |
|
"learning_rate": 6.19384977511771e-05, |
|
"loss": 3.3569, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.03820097854723372, |
|
"grad_norm": 1.5076078176498413, |
|
"learning_rate": 6.158344673142573e-05, |
|
"loss": 3.5497, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.03838916070756492, |
|
"grad_norm": 1.5262771844863892, |
|
"learning_rate": 6.122777741244067e-05, |
|
"loss": 3.5109, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.038577342867896124, |
|
"grad_norm": 1.5841517448425293, |
|
"learning_rate": 6.0871508779077856e-05, |
|
"loss": 3.6458, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.038765525028227325, |
|
"grad_norm": 1.706384539604187, |
|
"learning_rate": 6.051465984818332e-05, |
|
"loss": 3.4581, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.03895370718855853, |
|
"grad_norm": 1.562947154045105, |
|
"learning_rate": 6.015724966757812e-05, |
|
"loss": 3.6251, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.03914188934888972, |
|
"grad_norm": 1.5278199911117554, |
|
"learning_rate": 5.979929731504158e-05, |
|
"loss": 3.4025, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.039330071509220924, |
|
"grad_norm": 1.528238296508789, |
|
"learning_rate": 5.944082189729301e-05, |
|
"loss": 3.5747, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.039518253669552125, |
|
"grad_norm": 1.6547585725784302, |
|
"learning_rate": 5.908184254897182e-05, |
|
"loss": 3.4107, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.03970643582988333, |
|
"grad_norm": 1.5595186948776245, |
|
"learning_rate": 5.872237843161612e-05, |
|
"loss": 3.4565, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.03989461799021453, |
|
"grad_norm": 1.5088775157928467, |
|
"learning_rate": 5.8362448732639894e-05, |
|
"loss": 3.3846, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.04008280015054573, |
|
"grad_norm": 1.543283224105835, |
|
"learning_rate": 5.800207266430895e-05, |
|
"loss": 3.5448, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.04027098231087693, |
|
"grad_norm": 1.4654241800308228, |
|
"learning_rate": 5.764126946271526e-05, |
|
"loss": 3.3228, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.04045916447120813, |
|
"grad_norm": 1.6389522552490234, |
|
"learning_rate": 5.7280058386750255e-05, |
|
"loss": 3.357, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.04064734663153933, |
|
"grad_norm": 1.4927722215652466, |
|
"learning_rate": 5.6918458717076815e-05, |
|
"loss": 3.434, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.04083552879187053, |
|
"grad_norm": 1.5620532035827637, |
|
"learning_rate": 5.655648975510014e-05, |
|
"loss": 3.5867, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.04102371095220173, |
|
"grad_norm": 1.6729021072387695, |
|
"learning_rate": 5.61941708219374e-05, |
|
"loss": 3.3869, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.041211893112532934, |
|
"grad_norm": 1.523036003112793, |
|
"learning_rate": 5.583152125738651e-05, |
|
"loss": 3.4844, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.041400075272864136, |
|
"grad_norm": 1.4695093631744385, |
|
"learning_rate": 5.546856041889373e-05, |
|
"loss": 3.4079, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.04158825743319533, |
|
"grad_norm": 1.525160789489746, |
|
"learning_rate": 5.510530768052047e-05, |
|
"loss": 3.358, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.04177643959352653, |
|
"grad_norm": 1.4674233198165894, |
|
"learning_rate": 5.4741782431909136e-05, |
|
"loss": 3.2476, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.041964621753857734, |
|
"grad_norm": 1.7390276193618774, |
|
"learning_rate": 5.437800407724812e-05, |
|
"loss": 3.3376, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.042152803914188935, |
|
"grad_norm": 1.5143543481826782, |
|
"learning_rate": 5.401399203423606e-05, |
|
"loss": 3.5849, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.04234098607452014, |
|
"grad_norm": 1.5307164192199707, |
|
"learning_rate": 5.364976573304538e-05, |
|
"loss": 3.2994, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.04252916823485134, |
|
"grad_norm": 1.5590108633041382, |
|
"learning_rate": 5.328534461528515e-05, |
|
"loss": 3.5251, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.042717350395182534, |
|
"grad_norm": 1.5407838821411133, |
|
"learning_rate": 5.29207481329633e-05, |
|
"loss": 3.3932, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.042905532555513735, |
|
"grad_norm": 1.5244678258895874, |
|
"learning_rate": 5.2555995747448364e-05, |
|
"loss": 3.3437, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.04309371471584494, |
|
"grad_norm": 1.4933329820632935, |
|
"learning_rate": 5.2191106928430644e-05, |
|
"loss": 3.399, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.04328189687617614, |
|
"grad_norm": 1.4906306266784668, |
|
"learning_rate": 5.182610115288295e-05, |
|
"loss": 3.2872, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.04347007903650734, |
|
"grad_norm": 1.647549033164978, |
|
"learning_rate": 5.1460997904021005e-05, |
|
"loss": 3.3683, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.04365826119683854, |
|
"grad_norm": 1.5551615953445435, |
|
"learning_rate": 5.109581667026341e-05, |
|
"loss": 3.4049, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.04384644335716974, |
|
"grad_norm": 1.5681029558181763, |
|
"learning_rate": 5.073057694419147e-05, |
|
"loss": 3.4455, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.04403462551750094, |
|
"grad_norm": 1.5506995916366577, |
|
"learning_rate": 5.036529822150865e-05, |
|
"loss": 3.3317, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.04422280767783214, |
|
"grad_norm": 2.233480215072632, |
|
"learning_rate": 5e-05, |
|
"loss": 3.4562, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.04441098983816334, |
|
"grad_norm": 1.4921319484710693, |
|
"learning_rate": 4.963470177849135e-05, |
|
"loss": 3.3868, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.044599171998494544, |
|
"grad_norm": 1.539383888244629, |
|
"learning_rate": 4.9269423055808544e-05, |
|
"loss": 3.3888, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.044787354158825746, |
|
"grad_norm": 1.494940161705017, |
|
"learning_rate": 4.8904183329736596e-05, |
|
"loss": 3.3123, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.04497553631915695, |
|
"grad_norm": 1.6059819459915161, |
|
"learning_rate": 4.853900209597903e-05, |
|
"loss": 3.5323, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.04516371847948814, |
|
"grad_norm": 1.5235280990600586, |
|
"learning_rate": 4.817389884711705e-05, |
|
"loss": 3.4477, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.045351900639819344, |
|
"grad_norm": 1.532889485359192, |
|
"learning_rate": 4.7808893071569374e-05, |
|
"loss": 3.4776, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.045540082800150546, |
|
"grad_norm": 1.5243102312088013, |
|
"learning_rate": 4.744400425255165e-05, |
|
"loss": 3.4302, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.04572826496048175, |
|
"grad_norm": 1.4719178676605225, |
|
"learning_rate": 4.707925186703671e-05, |
|
"loss": 3.2717, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.04591644712081295, |
|
"grad_norm": 1.5308884382247925, |
|
"learning_rate": 4.671465538471486e-05, |
|
"loss": 3.5024, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.04610462928114415, |
|
"grad_norm": 1.5471347570419312, |
|
"learning_rate": 4.6350234266954626e-05, |
|
"loss": 3.3279, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.046292811441475346, |
|
"grad_norm": 1.4882664680480957, |
|
"learning_rate": 4.598600796576395e-05, |
|
"loss": 3.2842, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.04648099360180655, |
|
"grad_norm": 1.546363115310669, |
|
"learning_rate": 4.562199592275188e-05, |
|
"loss": 3.7441, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.04666917576213775, |
|
"grad_norm": 1.5140584707260132, |
|
"learning_rate": 4.5258217568090876e-05, |
|
"loss": 3.337, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.04685735792246895, |
|
"grad_norm": 1.894798755645752, |
|
"learning_rate": 4.4894692319479544e-05, |
|
"loss": 3.2422, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.04704554008280015, |
|
"grad_norm": 2.647658109664917, |
|
"learning_rate": 4.4531439581106295e-05, |
|
"loss": 3.1319, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.047233722243131354, |
|
"grad_norm": 1.497327446937561, |
|
"learning_rate": 4.4168478742613506e-05, |
|
"loss": 3.3551, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.04742190440346255, |
|
"grad_norm": 1.5717421770095825, |
|
"learning_rate": 4.38058291780626e-05, |
|
"loss": 3.7432, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.04761008656379375, |
|
"grad_norm": 1.4942125082015991, |
|
"learning_rate": 4.3443510244899864e-05, |
|
"loss": 3.4546, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.04779826872412495, |
|
"grad_norm": 1.5224213600158691, |
|
"learning_rate": 4.308154128292318e-05, |
|
"loss": 3.4569, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.047986450884456154, |
|
"grad_norm": 3.2993710041046143, |
|
"learning_rate": 4.271994161324977e-05, |
|
"loss": 3.4279, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.048174633044787356, |
|
"grad_norm": 1.776840090751648, |
|
"learning_rate": 4.235873053728475e-05, |
|
"loss": 3.4157, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.04836281520511856, |
|
"grad_norm": 1.5063600540161133, |
|
"learning_rate": 4.199792733569107e-05, |
|
"loss": 3.4676, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.04855099736544975, |
|
"grad_norm": 1.5720107555389404, |
|
"learning_rate": 4.163755126736012e-05, |
|
"loss": 3.5201, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.048739179525780954, |
|
"grad_norm": 1.5020864009857178, |
|
"learning_rate": 4.127762156838389e-05, |
|
"loss": 3.382, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.048927361686112156, |
|
"grad_norm": 1.516310214996338, |
|
"learning_rate": 4.0918157451028185e-05, |
|
"loss": 3.4377, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.04911554384644336, |
|
"grad_norm": 1.50225031375885, |
|
"learning_rate": 4.055917810270698e-05, |
|
"loss": 3.4051, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.04930372600677456, |
|
"grad_norm": 1.5030755996704102, |
|
"learning_rate": 4.020070268495843e-05, |
|
"loss": 3.3375, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.04949190816710576, |
|
"grad_norm": 1.4907896518707275, |
|
"learning_rate": 3.9842750332421896e-05, |
|
"loss": 3.4886, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.049680090327436956, |
|
"grad_norm": 1.4596471786499023, |
|
"learning_rate": 3.94853401518167e-05, |
|
"loss": 3.2683, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.04986827248776816, |
|
"grad_norm": 1.4404525756835938, |
|
"learning_rate": 3.9128491220922156e-05, |
|
"loss": 3.3169, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.05005645464809936, |
|
"grad_norm": 1.473186731338501, |
|
"learning_rate": 3.877222258755935e-05, |
|
"loss": 3.1098, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.05024463680843056, |
|
"grad_norm": 1.5433677434921265, |
|
"learning_rate": 3.8416553268574285e-05, |
|
"loss": 3.4896, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.05043281896876176, |
|
"grad_norm": 1.4866740703582764, |
|
"learning_rate": 3.80615022488229e-05, |
|
"loss": 3.4388, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.050621001129092964, |
|
"grad_norm": 1.5249758958816528, |
|
"learning_rate": 3.770708848015768e-05, |
|
"loss": 3.3572, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.050809183289424166, |
|
"grad_norm": 1.5038138628005981, |
|
"learning_rate": 3.735333088041596e-05, |
|
"loss": 3.3581, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.05099736544975536, |
|
"grad_norm": 1.479463815689087, |
|
"learning_rate": 3.7000248332410304e-05, |
|
"loss": 3.2485, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.05118554761008656, |
|
"grad_norm": 1.5273246765136719, |
|
"learning_rate": 3.664785968292036e-05, |
|
"loss": 3.3607, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.051373729770417764, |
|
"grad_norm": 1.501825213432312, |
|
"learning_rate": 3.629618374168711e-05, |
|
"loss": 3.2803, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.051561911930748966, |
|
"grad_norm": 1.472866415977478, |
|
"learning_rate": 3.594523928040859e-05, |
|
"loss": 3.274, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.05175009409108017, |
|
"grad_norm": 1.490519404411316, |
|
"learning_rate": 3.5595045031738125e-05, |
|
"loss": 3.3221, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.05193827625141137, |
|
"grad_norm": 1.520186185836792, |
|
"learning_rate": 3.5245619688284274e-05, |
|
"loss": 3.4458, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.052126458411742564, |
|
"grad_norm": 1.5064996480941772, |
|
"learning_rate": 3.4896981901613104e-05, |
|
"loss": 3.3529, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.052314640572073766, |
|
"grad_norm": 1.519168734550476, |
|
"learning_rate": 3.4549150281252636e-05, |
|
"loss": 3.6177, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.05250282273240497, |
|
"grad_norm": 1.5575183629989624, |
|
"learning_rate": 3.420214339369944e-05, |
|
"loss": 3.4754, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.05269100489273617, |
|
"grad_norm": 1.5328855514526367, |
|
"learning_rate": 3.38559797614277e-05, |
|
"loss": 3.2606, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.05287918705306737, |
|
"grad_norm": 1.585930347442627, |
|
"learning_rate": 3.351067786190038e-05, |
|
"loss": 3.3014, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.05306736921339857, |
|
"grad_norm": 1.5054539442062378, |
|
"learning_rate": 3.316625612658315e-05, |
|
"loss": 3.3176, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.05325555137372977, |
|
"grad_norm": 1.5619922876358032, |
|
"learning_rate": 3.282273293996027e-05, |
|
"loss": 3.4408, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.05344373353406097, |
|
"grad_norm": 1.470646858215332, |
|
"learning_rate": 3.248012663855353e-05, |
|
"loss": 3.2584, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.05363191569439217, |
|
"grad_norm": 1.5209406614303589, |
|
"learning_rate": 3.2138455509943366e-05, |
|
"loss": 3.3224, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.05382009785472337, |
|
"grad_norm": 1.4932469129562378, |
|
"learning_rate": 3.179773779179267e-05, |
|
"loss": 3.3588, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.054008280015054574, |
|
"grad_norm": 1.4727731943130493, |
|
"learning_rate": 3.145799167087342e-05, |
|
"loss": 3.2095, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.054196462175385776, |
|
"grad_norm": 1.501749038696289, |
|
"learning_rate": 3.111923528209577e-05, |
|
"loss": 3.3637, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.05438464433571697, |
|
"grad_norm": 1.5096619129180908, |
|
"learning_rate": 3.078148670754022e-05, |
|
"loss": 3.3815, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.05457282649604817, |
|
"grad_norm": 1.500936508178711, |
|
"learning_rate": 3.0444763975492208e-05, |
|
"loss": 3.1945, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.054761008656379374, |
|
"grad_norm": 1.4544003009796143, |
|
"learning_rate": 3.0109085059480017e-05, |
|
"loss": 3.3647, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.054949190816710576, |
|
"grad_norm": 1.552875280380249, |
|
"learning_rate": 2.977446787731532e-05, |
|
"loss": 3.39, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.05513737297704178, |
|
"grad_norm": 1.5364927053451538, |
|
"learning_rate": 2.944093029013664e-05, |
|
"loss": 3.4203, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.05532555513737298, |
|
"grad_norm": 1.5016520023345947, |
|
"learning_rate": 2.910849010145617e-05, |
|
"loss": 3.3805, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.055513737297704174, |
|
"grad_norm": 1.5016636848449707, |
|
"learning_rate": 2.8777165056209256e-05, |
|
"loss": 3.1924, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.055701919458035376, |
|
"grad_norm": 1.5240119695663452, |
|
"learning_rate": 2.8446972839807384e-05, |
|
"loss": 3.5986, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.05589010161836658, |
|
"grad_norm": 1.5180970430374146, |
|
"learning_rate": 2.8117931077194065e-05, |
|
"loss": 3.3258, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.05607828377869778, |
|
"grad_norm": 1.465223789215088, |
|
"learning_rate": 2.7790057331904117e-05, |
|
"loss": 3.2024, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.05626646593902898, |
|
"grad_norm": 1.7016164064407349, |
|
"learning_rate": 2.746336910512606e-05, |
|
"loss": 3.3025, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.05645464809936018, |
|
"grad_norm": 2.5566675662994385, |
|
"learning_rate": 2.7137883834768073e-05, |
|
"loss": 3.1934, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.05645464809936018, |
|
"eval_loss": 0.8374229073524475, |
|
"eval_runtime": 982.5965, |
|
"eval_samples_per_second": 9.109, |
|
"eval_steps_per_second": 2.278, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.056642830259691385, |
|
"grad_norm": 1.5962673425674438, |
|
"learning_rate": 2.6813618894527138e-05, |
|
"loss": 3.5087, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.05683101242002258, |
|
"grad_norm": 1.5255632400512695, |
|
"learning_rate": 2.6490591592961578e-05, |
|
"loss": 3.4712, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.05701919458035378, |
|
"grad_norm": 1.542389154434204, |
|
"learning_rate": 2.6168819172567392e-05, |
|
"loss": 3.2852, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.05720737674068498, |
|
"grad_norm": 1.664982557296753, |
|
"learning_rate": 2.5848318808857606e-05, |
|
"loss": 3.5038, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.057395558901016185, |
|
"grad_norm": 1.4651890993118286, |
|
"learning_rate": 2.5529107609445733e-05, |
|
"loss": 3.1837, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.057583741061347386, |
|
"grad_norm": 1.486304521560669, |
|
"learning_rate": 2.521120261313241e-05, |
|
"loss": 3.2311, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.05777192322167859, |
|
"grad_norm": 1.4881970882415771, |
|
"learning_rate": 2.4894620788996037e-05, |
|
"loss": 3.419, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.05796010538200978, |
|
"grad_norm": 1.4943639039993286, |
|
"learning_rate": 2.457937903548695e-05, |
|
"loss": 3.3085, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.058148287542340985, |
|
"grad_norm": 1.5298432111740112, |
|
"learning_rate": 2.426549417952542e-05, |
|
"loss": 3.608, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.058336469702672186, |
|
"grad_norm": 1.5149232149124146, |
|
"learning_rate": 2.3952982975603496e-05, |
|
"loss": 3.3398, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.05852465186300339, |
|
"grad_norm": 1.492951512336731, |
|
"learning_rate": 2.3641862104890595e-05, |
|
"loss": 3.3529, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.05871283402333459, |
|
"grad_norm": 1.448451280593872, |
|
"learning_rate": 2.3332148174343254e-05, |
|
"loss": 3.2021, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.05890101618366579, |
|
"grad_norm": 1.4983866214752197, |
|
"learning_rate": 2.3023857715818532e-05, |
|
"loss": 3.3579, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.059089198343996986, |
|
"grad_norm": 1.5230140686035156, |
|
"learning_rate": 2.2717007185191674e-05, |
|
"loss": 3.4931, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.05927738050432819, |
|
"grad_norm": 1.5284234285354614, |
|
"learning_rate": 2.24116129614777e-05, |
|
"loss": 3.4331, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.05946556266465939, |
|
"grad_norm": 1.4961637258529663, |
|
"learning_rate": 2.2107691345957133e-05, |
|
"loss": 3.3148, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.05965374482499059, |
|
"grad_norm": 1.547951340675354, |
|
"learning_rate": 2.1805258561305862e-05, |
|
"loss": 3.3194, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.05984192698532179, |
|
"grad_norm": 1.5121986865997314, |
|
"learning_rate": 2.1504330750729186e-05, |
|
"loss": 3.4504, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.060030109145652995, |
|
"grad_norm": 1.51499605178833, |
|
"learning_rate": 2.120492397710022e-05, |
|
"loss": 3.4349, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.06021829130598419, |
|
"grad_norm": 1.520011305809021, |
|
"learning_rate": 2.090705422210237e-05, |
|
"loss": 3.3619, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.06040647346631539, |
|
"grad_norm": 1.5512540340423584, |
|
"learning_rate": 2.061073738537635e-05, |
|
"loss": 3.38, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.06059465562664659, |
|
"grad_norm": 1.463521122932434, |
|
"learning_rate": 2.0315989283671473e-05, |
|
"loss": 3.1915, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.060782837786977795, |
|
"grad_norm": 1.5051774978637695, |
|
"learning_rate": 2.0022825650001387e-05, |
|
"loss": 3.4725, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.060971019947308996, |
|
"grad_norm": 1.4965063333511353, |
|
"learning_rate": 1.9731262132804274e-05, |
|
"loss": 3.3025, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.0611592021076402, |
|
"grad_norm": 1.4896841049194336, |
|
"learning_rate": 1.9441314295107537e-05, |
|
"loss": 3.4046, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.06134738426797139, |
|
"grad_norm": 1.4985949993133545, |
|
"learning_rate": 1.9152997613697183e-05, |
|
"loss": 3.4278, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.061535566428302595, |
|
"grad_norm": 1.5722819566726685, |
|
"learning_rate": 1.8866327478291546e-05, |
|
"loss": 3.4424, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.061723748588633796, |
|
"grad_norm": 1.457684874534607, |
|
"learning_rate": 1.8581319190720035e-05, |
|
"loss": 3.175, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.061911930748965, |
|
"grad_norm": 1.501202940940857, |
|
"learning_rate": 1.8297987964106115e-05, |
|
"loss": 3.2706, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.0621001129092962, |
|
"grad_norm": 1.4955599308013916, |
|
"learning_rate": 1.801634892205545e-05, |
|
"loss": 3.3183, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.0622882950696274, |
|
"grad_norm": 1.52448308467865, |
|
"learning_rate": 1.7736417097848506e-05, |
|
"loss": 3.3863, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.0624764772299586, |
|
"grad_norm": 1.5245345830917358, |
|
"learning_rate": 1.7458207433638223e-05, |
|
"loss": 3.4718, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.0626646593902898, |
|
"grad_norm": 1.5418884754180908, |
|
"learning_rate": 1.718173477965236e-05, |
|
"loss": 3.3681, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.062852841550621, |
|
"grad_norm": 1.487151861190796, |
|
"learning_rate": 1.6907013893400837e-05, |
|
"loss": 3.3089, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.0630410237109522, |
|
"grad_norm": 1.4714837074279785, |
|
"learning_rate": 1.6634059438888033e-05, |
|
"loss": 3.181, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.0632292058712834, |
|
"grad_norm": 1.5032023191452026, |
|
"learning_rate": 1.636288598583e-05, |
|
"loss": 3.3265, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.0634173880316146, |
|
"grad_norm": 1.5267105102539062, |
|
"learning_rate": 1.6093508008876857e-05, |
|
"loss": 3.1744, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.0636055701919458, |
|
"grad_norm": 1.517006516456604, |
|
"learning_rate": 1.5825939886840037e-05, |
|
"loss": 3.2575, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.06379375235227701, |
|
"grad_norm": 1.4955118894577026, |
|
"learning_rate": 1.5560195901924894e-05, |
|
"loss": 3.2745, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.06398193451260821, |
|
"grad_norm": 1.7153363227844238, |
|
"learning_rate": 1.5296290238968303e-05, |
|
"loss": 3.4614, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.06417011667293941, |
|
"grad_norm": 1.5209360122680664, |
|
"learning_rate": 1.50342369846815e-05, |
|
"loss": 3.3276, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.0643582988332706, |
|
"grad_norm": 1.5184190273284912, |
|
"learning_rate": 1.4774050126898164e-05, |
|
"loss": 3.3869, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.0645464809936018, |
|
"grad_norm": 1.5385315418243408, |
|
"learning_rate": 1.451574355382776e-05, |
|
"loss": 3.4677, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.064734663153933, |
|
"grad_norm": 1.4663329124450684, |
|
"learning_rate": 1.425933105331429e-05, |
|
"loss": 3.3069, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.0649228453142642, |
|
"grad_norm": 1.5196608304977417, |
|
"learning_rate": 1.4004826312100216e-05, |
|
"loss": 3.3985, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.0651110274745954, |
|
"grad_norm": 1.5887420177459717, |
|
"learning_rate": 1.3752242915095992e-05, |
|
"loss": 3.309, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.06529920963492661, |
|
"grad_norm": 1.5067335367202759, |
|
"learning_rate": 1.3501594344654884e-05, |
|
"loss": 3.186, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.06548739179525781, |
|
"grad_norm": 1.4992436170578003, |
|
"learning_rate": 1.3252893979853304e-05, |
|
"loss": 3.3428, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.06567557395558901, |
|
"grad_norm": 1.6664701700210571, |
|
"learning_rate": 1.3006155095776707e-05, |
|
"loss": 3.4061, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.06586375611592021, |
|
"grad_norm": 2.502802848815918, |
|
"learning_rate": 1.2761390862810907e-05, |
|
"loss": 3.0303, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.06605193827625142, |
|
"grad_norm": 1.500586748123169, |
|
"learning_rate": 1.2518614345939212e-05, |
|
"loss": 3.4372, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.06624012043658262, |
|
"grad_norm": 1.5125335454940796, |
|
"learning_rate": 1.227783850404487e-05, |
|
"loss": 3.3834, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.06642830259691382, |
|
"grad_norm": 1.5365911722183228, |
|
"learning_rate": 1.2039076189219517e-05, |
|
"loss": 3.4872, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.066616484757245, |
|
"grad_norm": 1.5221909284591675, |
|
"learning_rate": 1.1802340146077045e-05, |
|
"loss": 3.3993, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.06680466691757621, |
|
"grad_norm": 1.5232350826263428, |
|
"learning_rate": 1.1567643011073392e-05, |
|
"loss": 3.5355, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.06699284907790741, |
|
"grad_norm": 1.5335094928741455, |
|
"learning_rate": 1.1334997311832002e-05, |
|
"loss": 3.4007, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.06718103123823861, |
|
"grad_norm": 1.4804530143737793, |
|
"learning_rate": 1.1104415466475087e-05, |
|
"loss": 3.3991, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.06736921339856981, |
|
"grad_norm": 1.4774835109710693, |
|
"learning_rate": 1.0875909782960886e-05, |
|
"loss": 3.3081, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.06755739555890101, |
|
"grad_norm": 1.507690191268921, |
|
"learning_rate": 1.0649492458426564e-05, |
|
"loss": 3.5331, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.06774557771923222, |
|
"grad_norm": 1.4900294542312622, |
|
"learning_rate": 1.0425175578537299e-05, |
|
"loss": 3.4034, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.06793375987956342, |
|
"grad_norm": 1.536511778831482, |
|
"learning_rate": 1.020297111684101e-05, |
|
"loss": 3.3285, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.06812194203989462, |
|
"grad_norm": 1.4938386678695679, |
|
"learning_rate": 9.98289093412938e-06, |
|
"loss": 3.3401, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.06831012420022582, |
|
"grad_norm": 1.498859167098999, |
|
"learning_rate": 9.764946777804646e-06, |
|
"loss": 3.4705, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.06849830636055702, |
|
"grad_norm": 1.505251407623291, |
|
"learning_rate": 9.549150281252633e-06, |
|
"loss": 3.3769, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.06868648852088823, |
|
"grad_norm": 1.5352188348770142, |
|
"learning_rate": 9.335512963221732e-06, |
|
"loss": 3.4495, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.06887467068121943, |
|
"grad_norm": 1.5243737697601318, |
|
"learning_rate": 9.124046227208082e-06, |
|
"loss": 3.3756, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.06906285284155061, |
|
"grad_norm": 1.4866875410079956, |
|
"learning_rate": 8.914761360846869e-06, |
|
"loss": 3.2548, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.06925103500188182, |
|
"grad_norm": 1.70474374294281, |
|
"learning_rate": 8.707669535309793e-06, |
|
"loss": 3.3533, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.06943921716221302, |
|
"grad_norm": 1.4734491109848022, |
|
"learning_rate": 8.502781804708826e-06, |
|
"loss": 3.3556, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.06962739932254422, |
|
"grad_norm": 1.5178672075271606, |
|
"learning_rate": 8.30010910550611e-06, |
|
"loss": 3.4011, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.06981558148287542, |
|
"grad_norm": 1.5426009893417358, |
|
"learning_rate": 8.09966225593024e-06, |
|
"loss": 3.3262, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.07000376364320662, |
|
"grad_norm": 1.488108515739441, |
|
"learning_rate": 7.901451955398792e-06, |
|
"loss": 3.3038, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.07019194580353783, |
|
"grad_norm": 1.5539402961730957, |
|
"learning_rate": 7.705488783947202e-06, |
|
"loss": 3.4263, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.07038012796386903, |
|
"grad_norm": 1.5011879205703735, |
|
"learning_rate": 7.511783201664052e-06, |
|
"loss": 3.1954, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.07056831012420023, |
|
"grad_norm": 1.5092988014221191, |
|
"learning_rate": 7.320345548132679e-06, |
|
"loss": 3.328, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.07075649228453143, |
|
"grad_norm": 1.498982310295105, |
|
"learning_rate": 7.131186041879357e-06, |
|
"loss": 3.2402, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.07094467444486263, |
|
"grad_norm": 1.5231132507324219, |
|
"learning_rate": 6.944314779827749e-06, |
|
"loss": 3.4769, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.07113285660519383, |
|
"grad_norm": 1.5165536403656006, |
|
"learning_rate": 6.759741736760061e-06, |
|
"loss": 3.4978, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.07132103876552502, |
|
"grad_norm": 1.531844973564148, |
|
"learning_rate": 6.577476764784546e-06, |
|
"loss": 3.3847, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.07150922092585622, |
|
"grad_norm": 1.4768074750900269, |
|
"learning_rate": 6.397529592809614e-06, |
|
"loss": 3.1253, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.07169740308618743, |
|
"grad_norm": 1.5122772455215454, |
|
"learning_rate": 6.219909826024589e-06, |
|
"loss": 3.3401, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.07188558524651863, |
|
"grad_norm": 1.5160380601882935, |
|
"learning_rate": 6.0446269453868945e-06, |
|
"loss": 3.3382, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.07207376740684983, |
|
"grad_norm": 1.5226458311080933, |
|
"learning_rate": 5.871690307116107e-06, |
|
"loss": 3.431, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.07226194956718103, |
|
"grad_norm": 1.4903466701507568, |
|
"learning_rate": 5.701109142194422e-06, |
|
"loss": 3.2707, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.07245013172751223, |
|
"grad_norm": 1.5170961618423462, |
|
"learning_rate": 5.532892555874059e-06, |
|
"loss": 3.3027, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.07263831388784343, |
|
"grad_norm": 1.5108839273452759, |
|
"learning_rate": 5.3670495271910925e-06, |
|
"loss": 3.3152, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.07282649604817464, |
|
"grad_norm": 1.51710045337677, |
|
"learning_rate": 5.203588908486279e-06, |
|
"loss": 3.2884, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.07301467820850584, |
|
"grad_norm": 1.5793352127075195, |
|
"learning_rate": 5.042519424932513e-06, |
|
"loss": 3.5139, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.07320286036883704, |
|
"grad_norm": 1.481472373008728, |
|
"learning_rate": 4.883849674069058e-06, |
|
"loss": 3.2427, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.07339104252916824, |
|
"grad_norm": 1.5381510257720947, |
|
"learning_rate": 4.727588125342669e-06, |
|
"loss": 3.3485, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.07357922468949944, |
|
"grad_norm": 1.508355975151062, |
|
"learning_rate": 4.573743119655516e-06, |
|
"loss": 3.295, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.07376740684983063, |
|
"grad_norm": 1.527873158454895, |
|
"learning_rate": 4.422322868919937e-06, |
|
"loss": 3.2918, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.07395558901016183, |
|
"grad_norm": 1.5130548477172852, |
|
"learning_rate": 4.273335455620097e-06, |
|
"loss": 3.3859, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.07414377117049303, |
|
"grad_norm": 1.5391393899917603, |
|
"learning_rate": 4.126788832380629e-06, |
|
"loss": 3.3711, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.07433195333082424, |
|
"grad_norm": 1.5213444232940674, |
|
"learning_rate": 3.982690821542035e-06, |
|
"loss": 3.4224, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.07452013549115544, |
|
"grad_norm": 1.4749212265014648, |
|
"learning_rate": 3.8410491147432395e-06, |
|
"loss": 3.1882, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.07470831765148664, |
|
"grad_norm": 1.4860862493515015, |
|
"learning_rate": 3.7018712725109926e-06, |
|
"loss": 3.2488, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.07489649981181784, |
|
"grad_norm": 1.5500200986862183, |
|
"learning_rate": 3.5651647238562904e-06, |
|
"loss": 3.3369, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.07508468197214904, |
|
"grad_norm": 1.8019529581069946, |
|
"learning_rate": 3.430936765877857e-06, |
|
"loss": 3.1557, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.07527286413248024, |
|
"grad_norm": 2.5598955154418945, |
|
"learning_rate": 3.299194563372604e-06, |
|
"loss": 3.03, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.07546104629281145, |
|
"grad_norm": 1.5204044580459595, |
|
"learning_rate": 3.1699451484532463e-06, |
|
"loss": 3.4419, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.07564922845314265, |
|
"grad_norm": 1.4903305768966675, |
|
"learning_rate": 3.0431954201728784e-06, |
|
"loss": 3.3295, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.07583741061347385, |
|
"grad_norm": 1.5254848003387451, |
|
"learning_rate": 2.9189521441567726e-06, |
|
"loss": 3.4107, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.07602559277380504, |
|
"grad_norm": 1.5176067352294922, |
|
"learning_rate": 2.797221952241219e-06, |
|
"loss": 3.3387, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.07621377493413624, |
|
"grad_norm": 1.495973825454712, |
|
"learning_rate": 2.6780113421195298e-06, |
|
"loss": 3.3161, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.07640195709446744, |
|
"grad_norm": 1.5124212503433228, |
|
"learning_rate": 2.561326676995218e-06, |
|
"loss": 3.3538, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.07659013925479864, |
|
"grad_norm": 1.4950138330459595, |
|
"learning_rate": 2.4471741852423237e-06, |
|
"loss": 3.363, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.07677832141512984, |
|
"grad_norm": 1.5359971523284912, |
|
"learning_rate": 2.3355599600729915e-06, |
|
"loss": 3.4671, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.07696650357546105, |
|
"grad_norm": 1.4483258724212646, |
|
"learning_rate": 2.2264899592121744e-06, |
|
"loss": 3.1561, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.07715468573579225, |
|
"grad_norm": 1.483723521232605, |
|
"learning_rate": 2.1199700045797077e-06, |
|
"loss": 3.3924, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.07734286789612345, |
|
"grad_norm": 1.4614593982696533, |
|
"learning_rate": 2.0160057819794466e-06, |
|
"loss": 3.2969, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.07753105005645465, |
|
"grad_norm": 1.4894236326217651, |
|
"learning_rate": 1.9146028407958484e-06, |
|
"loss": 3.3376, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.07771923221678585, |
|
"grad_norm": 1.501068115234375, |
|
"learning_rate": 1.8157665936977263e-06, |
|
"loss": 3.3462, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.07790741437711705, |
|
"grad_norm": 1.4664418697357178, |
|
"learning_rate": 1.7195023163493252e-06, |
|
"loss": 3.1851, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.07809559653744826, |
|
"grad_norm": 1.5121281147003174, |
|
"learning_rate": 1.6258151471287396e-06, |
|
"loss": 3.3562, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.07828377869777944, |
|
"grad_norm": 1.5005747079849243, |
|
"learning_rate": 1.5347100868536246e-06, |
|
"loss": 3.276, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.07847196085811065, |
|
"grad_norm": 1.5016510486602783, |
|
"learning_rate": 1.4461919985142735e-06, |
|
"loss": 3.1685, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.07866014301844185, |
|
"grad_norm": 1.490427017211914, |
|
"learning_rate": 1.3602656070140275e-06, |
|
"loss": 3.3398, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.07884832517877305, |
|
"grad_norm": 1.5081381797790527, |
|
"learning_rate": 1.27693549891707e-06, |
|
"loss": 3.378, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.07903650733910425, |
|
"grad_norm": 1.5135114192962646, |
|
"learning_rate": 1.196206122203647e-06, |
|
"loss": 3.2584, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.07922468949943545, |
|
"grad_norm": 1.503239631652832, |
|
"learning_rate": 1.1180817860325599e-06, |
|
"loss": 3.3742, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.07941287165976665, |
|
"grad_norm": 1.5126641988754272, |
|
"learning_rate": 1.0425666605112517e-06, |
|
"loss": 3.4263, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.07960105382009786, |
|
"grad_norm": 1.5308884382247925, |
|
"learning_rate": 9.696647764731337e-07, |
|
"loss": 3.3315, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.07978923598042906, |
|
"grad_norm": 1.5186878442764282, |
|
"learning_rate": 8.993800252624862e-07, |
|
"loss": 3.3336, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.07997741814076026, |
|
"grad_norm": 1.5928887128829956, |
|
"learning_rate": 8.317161585266964e-07, |
|
"loss": 3.641, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.08016560030109146, |
|
"grad_norm": 1.5158437490463257, |
|
"learning_rate": 7.666767880160464e-07, |
|
"loss": 3.4028, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.08035378246142266, |
|
"grad_norm": 1.5123188495635986, |
|
"learning_rate": 7.042653853909064e-07, |
|
"loss": 3.4125, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.08054196462175386, |
|
"grad_norm": 1.4982678890228271, |
|
"learning_rate": 6.444852820364222e-07, |
|
"loss": 3.2116, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.08073014678208505, |
|
"grad_norm": 1.5639753341674805, |
|
"learning_rate": 5.87339668884701e-07, |
|
"loss": 3.1289, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.08091832894241625, |
|
"grad_norm": 1.572089672088623, |
|
"learning_rate": 5.328315962444874e-07, |
|
"loss": 3.3213, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.08110651110274746, |
|
"grad_norm": 1.5132265090942383, |
|
"learning_rate": 4.809639736383431e-07, |
|
"loss": 3.3827, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.08129469326307866, |
|
"grad_norm": 1.4733517169952393, |
|
"learning_rate": 4.317395696473214e-07, |
|
"loss": 3.2562, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.08148287542340986, |
|
"grad_norm": 1.548933982849121, |
|
"learning_rate": 3.851610117632354e-07, |
|
"loss": 3.4471, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.08167105758374106, |
|
"grad_norm": 1.4967180490493774, |
|
"learning_rate": 3.4123078624834216e-07, |
|
"loss": 3.0655, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.08185923974407226, |
|
"grad_norm": 1.6242047548294067, |
|
"learning_rate": 2.9995123800270476e-07, |
|
"loss": 3.2837, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.08204742190440346, |
|
"grad_norm": 1.4935169219970703, |
|
"learning_rate": 2.613245704389644e-07, |
|
"loss": 3.3263, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.08223560406473467, |
|
"grad_norm": 1.4887654781341553, |
|
"learning_rate": 2.2535284536476242e-07, |
|
"loss": 3.1296, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.08242378622506587, |
|
"grad_norm": 1.511015772819519, |
|
"learning_rate": 1.920379828726726e-07, |
|
"loss": 3.3212, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.08261196838539707, |
|
"grad_norm": 1.5536808967590332, |
|
"learning_rate": 1.6138176123770554e-07, |
|
"loss": 3.5203, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.08280015054572827, |
|
"grad_norm": 1.5585739612579346, |
|
"learning_rate": 1.333858168224178e-07, |
|
"loss": 3.2727, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.08298833270605946, |
|
"grad_norm": 1.5475882291793823, |
|
"learning_rate": 1.0805164398952072e-07, |
|
"loss": 3.4168, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.08317651486639066, |
|
"grad_norm": 1.5009870529174805, |
|
"learning_rate": 8.53805950221498e-08, |
|
"loss": 3.3267, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.08336469702672186, |
|
"grad_norm": 1.5358216762542725, |
|
"learning_rate": 6.537388005167233e-08, |
|
"loss": 3.4914, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.08355287918705306, |
|
"grad_norm": 1.5157291889190674, |
|
"learning_rate": 4.8032566993089225e-08, |
|
"loss": 3.3647, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.08374106134738427, |
|
"grad_norm": 1.5044530630111694, |
|
"learning_rate": 3.3357581488030475e-08, |
|
"loss": 3.2696, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.08392924350771547, |
|
"grad_norm": 1.5259320735931396, |
|
"learning_rate": 2.134970685536697e-08, |
|
"loss": 3.3305, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.08411742566804667, |
|
"grad_norm": 1.4801864624023438, |
|
"learning_rate": 1.200958404936059e-08, |
|
"loss": 3.1884, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.08430560782837787, |
|
"grad_norm": 1.4972193241119385, |
|
"learning_rate": 5.337711625497121e-09, |
|
"loss": 3.3929, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.08449378998870907, |
|
"grad_norm": 1.7666032314300537, |
|
"learning_rate": 1.3344457138297906e-09, |
|
"loss": 3.4298, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.08468197214904027, |
|
"grad_norm": 2.5446274280548096, |
|
"learning_rate": 0.0, |
|
"loss": 2.8081, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.08468197214904027, |
|
"eval_loss": 0.8302884697914124, |
|
"eval_runtime": 982.1123, |
|
"eval_samples_per_second": 9.113, |
|
"eval_steps_per_second": 2.279, |
|
"step": 450 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 450, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 150, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 2, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.0272928176694886e+18, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|