|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.998053121958003, |
|
"eval_steps": 200, |
|
"global_step": 1796, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.001112501738283966, |
|
"grad_norm": 0.6684235958622939, |
|
"learning_rate": 1.1111111111111112e-06, |
|
"loss": 1.1751, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.00556250869141983, |
|
"grad_norm": 0.7610951352005912, |
|
"learning_rate": 5.555555555555556e-06, |
|
"loss": 1.2823, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.01112501738283966, |
|
"grad_norm": 0.8690361671075456, |
|
"learning_rate": 1.1111111111111112e-05, |
|
"loss": 1.2894, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.016687526074259492, |
|
"grad_norm": 1.059736284119422, |
|
"learning_rate": 1.6666666666666667e-05, |
|
"loss": 1.3166, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.02225003476567932, |
|
"grad_norm": 0.4906355740047051, |
|
"learning_rate": 2.2222222222222223e-05, |
|
"loss": 1.1771, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.027812543457099152, |
|
"grad_norm": 1.04376108340149, |
|
"learning_rate": 2.777777777777778e-05, |
|
"loss": 1.1062, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.033375052148518984, |
|
"grad_norm": 0.47593208759368216, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 1.014, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03893756083993881, |
|
"grad_norm": 0.2526378868894225, |
|
"learning_rate": 3.888888888888889e-05, |
|
"loss": 0.9027, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.04450006953135864, |
|
"grad_norm": 0.2750620823042748, |
|
"learning_rate": 4.4444444444444447e-05, |
|
"loss": 0.8556, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.05006257822277847, |
|
"grad_norm": 0.22806716350963602, |
|
"learning_rate": 5e-05, |
|
"loss": 0.8714, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.055625086914198305, |
|
"grad_norm": 0.31305577393443185, |
|
"learning_rate": 5.555555555555556e-05, |
|
"loss": 0.8431, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.061187595605618136, |
|
"grad_norm": 0.3231781586973102, |
|
"learning_rate": 6.111111111111112e-05, |
|
"loss": 0.7896, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.06675010429703797, |
|
"grad_norm": 0.2652381898767101, |
|
"learning_rate": 6.666666666666667e-05, |
|
"loss": 0.8031, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.0723126129884578, |
|
"grad_norm": 0.16353190746623922, |
|
"learning_rate": 7.222222222222222e-05, |
|
"loss": 0.798, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.07787512167987762, |
|
"grad_norm": 0.14957803129107605, |
|
"learning_rate": 7.777777777777778e-05, |
|
"loss": 0.8214, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.08343763037129745, |
|
"grad_norm": 0.1558832457817548, |
|
"learning_rate": 8.333333333333334e-05, |
|
"loss": 0.8046, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.08900013906271728, |
|
"grad_norm": 0.15611943332427167, |
|
"learning_rate": 8.888888888888889e-05, |
|
"loss": 0.7725, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.09456264775413711, |
|
"grad_norm": 0.17878911552246932, |
|
"learning_rate": 9.444444444444444e-05, |
|
"loss": 0.758, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.10012515644555695, |
|
"grad_norm": 0.21152631925896626, |
|
"learning_rate": 0.0001, |
|
"loss": 0.7741, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.10568766513697678, |
|
"grad_norm": 0.15798064484261687, |
|
"learning_rate": 0.00010555555555555557, |
|
"loss": 0.7411, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.11125017382839661, |
|
"grad_norm": 0.1770136407809846, |
|
"learning_rate": 0.00011111111111111112, |
|
"loss": 0.7582, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.11681268251981644, |
|
"grad_norm": 0.1721697060786133, |
|
"learning_rate": 0.00011666666666666668, |
|
"loss": 0.7908, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.12237519121123627, |
|
"grad_norm": 0.19163940220247466, |
|
"learning_rate": 0.00012222222222222224, |
|
"loss": 0.7769, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.1279376999026561, |
|
"grad_norm": 0.17258655274801674, |
|
"learning_rate": 0.00012777777777777776, |
|
"loss": 0.7291, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.13350020859407594, |
|
"grad_norm": 0.14647192105198578, |
|
"learning_rate": 0.00013333333333333334, |
|
"loss": 0.7147, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.13906271728549577, |
|
"grad_norm": 0.15483066376892327, |
|
"learning_rate": 0.0001388888888888889, |
|
"loss": 0.7253, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.1446252259769156, |
|
"grad_norm": 0.15985722819525555, |
|
"learning_rate": 0.00014444444444444444, |
|
"loss": 0.74, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.15018773466833543, |
|
"grad_norm": 0.18677759796975052, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 0.7578, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.15575024335975524, |
|
"grad_norm": 0.1502765821288928, |
|
"learning_rate": 0.00015555555555555556, |
|
"loss": 0.7541, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.16131275205117507, |
|
"grad_norm": 0.16754721310239581, |
|
"learning_rate": 0.0001611111111111111, |
|
"loss": 0.7338, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.1668752607425949, |
|
"grad_norm": 0.15603842447844732, |
|
"learning_rate": 0.0001666666666666667, |
|
"loss": 0.7347, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.17243776943401473, |
|
"grad_norm": 0.16036041756367805, |
|
"learning_rate": 0.00017222222222222224, |
|
"loss": 0.7275, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.17800027812543456, |
|
"grad_norm": 0.17275116566192827, |
|
"learning_rate": 0.00017777777777777779, |
|
"loss": 0.7357, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.1835627868168544, |
|
"grad_norm": 0.16362661400179987, |
|
"learning_rate": 0.00018333333333333334, |
|
"loss": 0.7454, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.18912529550827423, |
|
"grad_norm": 0.1529061378986909, |
|
"learning_rate": 0.00018888888888888888, |
|
"loss": 0.7305, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.19468780419969406, |
|
"grad_norm": 0.16390278757569096, |
|
"learning_rate": 0.00019444444444444446, |
|
"loss": 0.7383, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.2002503128911139, |
|
"grad_norm": 0.16834812585150158, |
|
"learning_rate": 0.0002, |
|
"loss": 0.7067, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.20581282158253372, |
|
"grad_norm": 0.15601116057262598, |
|
"learning_rate": 0.0001999952758505736, |
|
"loss": 0.7567, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.21137533027395355, |
|
"grad_norm": 0.15788111366123353, |
|
"learning_rate": 0.00019998110384864614, |
|
"loss": 0.7434, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.2169378389653734, |
|
"grad_norm": 0.16327699685803637, |
|
"learning_rate": 0.00019995748533323075, |
|
"loss": 0.7087, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.22250034765679322, |
|
"grad_norm": 0.14758830274997248, |
|
"learning_rate": 0.0001999244225358753, |
|
"loss": 0.718, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.22250034765679322, |
|
"eval_loss": 0.7089951634407043, |
|
"eval_runtime": 38.6391, |
|
"eval_samples_per_second": 10.533, |
|
"eval_steps_per_second": 0.673, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.22806285634821305, |
|
"grad_norm": 0.16285906061792327, |
|
"learning_rate": 0.00019988191858045178, |
|
"loss": 0.7323, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.23362536503963288, |
|
"grad_norm": 0.15445314974456065, |
|
"learning_rate": 0.00019982997748286082, |
|
"loss": 0.7427, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.2391878737310527, |
|
"grad_norm": 0.14969344173380988, |
|
"learning_rate": 0.00019976860415065256, |
|
"loss": 0.7171, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.24475038242247255, |
|
"grad_norm": 0.15378571057430904, |
|
"learning_rate": 0.00019969780438256293, |
|
"loss": 0.726, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.2503128911138924, |
|
"grad_norm": 0.1655361908944567, |
|
"learning_rate": 0.00019961758486796555, |
|
"loss": 0.7463, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.2558753998053122, |
|
"grad_norm": 0.1343012669489156, |
|
"learning_rate": 0.00019952795318623986, |
|
"loss": 0.7115, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.26143790849673204, |
|
"grad_norm": 0.1597015078714079, |
|
"learning_rate": 0.00019942891780605512, |
|
"loss": 0.6901, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.2670004171881519, |
|
"grad_norm": 0.13156659502510382, |
|
"learning_rate": 0.0001993204880845699, |
|
"loss": 0.6974, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.2725629258795717, |
|
"grad_norm": 0.13487041292813268, |
|
"learning_rate": 0.0001992026742665484, |
|
"loss": 0.7264, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.27812543457099154, |
|
"grad_norm": 0.1452928301770259, |
|
"learning_rate": 0.00019907548748339222, |
|
"loss": 0.7132, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.28368794326241137, |
|
"grad_norm": 0.13005536436409637, |
|
"learning_rate": 0.00019893893975208874, |
|
"loss": 0.6924, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.2892504519538312, |
|
"grad_norm": 0.15551390193086698, |
|
"learning_rate": 0.0001987930439740757, |
|
"loss": 0.7137, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.29481296064525103, |
|
"grad_norm": 0.1411196790735591, |
|
"learning_rate": 0.0001986378139340222, |
|
"loss": 0.7003, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.30037546933667086, |
|
"grad_norm": 0.16701635405751747, |
|
"learning_rate": 0.0001984732642985263, |
|
"loss": 0.7318, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.3059379780280907, |
|
"grad_norm": 0.15781390142645427, |
|
"learning_rate": 0.00019829941061472943, |
|
"loss": 0.725, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.31150048671951047, |
|
"grad_norm": 0.1276451211299801, |
|
"learning_rate": 0.0001981162693088471, |
|
"loss": 0.7008, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.3170629954109303, |
|
"grad_norm": 0.14226714916323183, |
|
"learning_rate": 0.00019792385768461723, |
|
"loss": 0.7215, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.32262550410235014, |
|
"grad_norm": 0.15169879239239556, |
|
"learning_rate": 0.00019772219392166519, |
|
"loss": 0.7298, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.32818801279376997, |
|
"grad_norm": 0.13790872011113864, |
|
"learning_rate": 0.00019751129707378583, |
|
"loss": 0.7024, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.3337505214851898, |
|
"grad_norm": 0.13510983391572584, |
|
"learning_rate": 0.00019729118706714375, |
|
"loss": 0.7301, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.33931303017660963, |
|
"grad_norm": 0.13421290102386768, |
|
"learning_rate": 0.00019706188469839012, |
|
"loss": 0.6973, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.34487553886802946, |
|
"grad_norm": 0.17987364390951446, |
|
"learning_rate": 0.000196823411632698, |
|
"loss": 0.7285, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.3504380475594493, |
|
"grad_norm": 0.1463794049975603, |
|
"learning_rate": 0.00019657579040171537, |
|
"loss": 0.7474, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.3560005562508691, |
|
"grad_norm": 0.14423415874709103, |
|
"learning_rate": 0.00019631904440143612, |
|
"loss": 0.6938, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.36156306494228896, |
|
"grad_norm": 0.11858366900169025, |
|
"learning_rate": 0.00019605319788998965, |
|
"loss": 0.6845, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.3671255736337088, |
|
"grad_norm": 0.14664958369445155, |
|
"learning_rate": 0.00019577827598534885, |
|
"loss": 0.7085, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.3726880823251286, |
|
"grad_norm": 0.16873527257647247, |
|
"learning_rate": 0.0001954943046629569, |
|
"loss": 0.717, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.37825059101654845, |
|
"grad_norm": 0.1486177299286815, |
|
"learning_rate": 0.00019520131075327298, |
|
"loss": 0.7321, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.3838130997079683, |
|
"grad_norm": 0.13978486223349953, |
|
"learning_rate": 0.00019489932193923735, |
|
"loss": 0.7258, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.3893756083993881, |
|
"grad_norm": 0.15478013993645653, |
|
"learning_rate": 0.00019458836675365556, |
|
"loss": 0.7147, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.39493811709080795, |
|
"grad_norm": 0.1335712601381255, |
|
"learning_rate": 0.00019426847457650292, |
|
"loss": 0.705, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.4005006257822278, |
|
"grad_norm": 0.14215220233504217, |
|
"learning_rate": 0.00019393967563214833, |
|
"loss": 0.7109, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.4060631344736476, |
|
"grad_norm": 0.12737803269805706, |
|
"learning_rate": 0.00019360200098649864, |
|
"loss": 0.6759, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.41162564316506745, |
|
"grad_norm": 0.14212696306551803, |
|
"learning_rate": 0.00019325548254406352, |
|
"loss": 0.6884, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.4171881518564873, |
|
"grad_norm": 0.13427514345248226, |
|
"learning_rate": 0.00019290015304494103, |
|
"loss": 0.6997, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.4227506605479071, |
|
"grad_norm": 0.13648211126654036, |
|
"learning_rate": 0.00019253604606172417, |
|
"loss": 0.7211, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.42831316923932694, |
|
"grad_norm": 0.1328171684670597, |
|
"learning_rate": 0.0001921631959963288, |
|
"loss": 0.6879, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.4338756779307468, |
|
"grad_norm": 0.13003415855983963, |
|
"learning_rate": 0.0001917816380767434, |
|
"loss": 0.7164, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.4394381866221666, |
|
"grad_norm": 0.12331669293330129, |
|
"learning_rate": 0.00019139140835370053, |
|
"loss": 0.6782, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.44500069531358644, |
|
"grad_norm": 0.12354925404653357, |
|
"learning_rate": 0.0001909925436972706, |
|
"loss": 0.7205, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.44500069531358644, |
|
"eval_loss": 0.6896927952766418, |
|
"eval_runtime": 38.0771, |
|
"eval_samples_per_second": 10.689, |
|
"eval_steps_per_second": 0.683, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.45056320400500627, |
|
"grad_norm": 0.12749730782556348, |
|
"learning_rate": 0.0001905850817933784, |
|
"loss": 0.7147, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.4561257126964261, |
|
"grad_norm": 0.1290830353607145, |
|
"learning_rate": 0.0001901690611402423, |
|
"loss": 0.7147, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.46168822138784593, |
|
"grad_norm": 0.14656956459115003, |
|
"learning_rate": 0.0001897445210447369, |
|
"loss": 0.7273, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.46725073007926576, |
|
"grad_norm": 0.13591979089604955, |
|
"learning_rate": 0.00018931150161867916, |
|
"loss": 0.7233, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.4728132387706856, |
|
"grad_norm": 0.14148246469362974, |
|
"learning_rate": 0.00018887004377503858, |
|
"loss": 0.6857, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.4783757474621054, |
|
"grad_norm": 0.13172632177206697, |
|
"learning_rate": 0.0001884201892240715, |
|
"loss": 0.6954, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.48393825615352526, |
|
"grad_norm": 0.15251049696040847, |
|
"learning_rate": 0.00018796198046938033, |
|
"loss": 0.7369, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.4895007648449451, |
|
"grad_norm": 0.14157664597377279, |
|
"learning_rate": 0.00018749546080389757, |
|
"loss": 0.6988, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.4950632735363649, |
|
"grad_norm": 0.14790889100169335, |
|
"learning_rate": 0.00018702067430579543, |
|
"loss": 0.7068, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.5006257822277848, |
|
"grad_norm": 0.13754915966044653, |
|
"learning_rate": 0.00018653766583432113, |
|
"loss": 0.7226, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.5061882909192046, |
|
"grad_norm": 0.1459944140173946, |
|
"learning_rate": 0.00018604648102555856, |
|
"loss": 0.6998, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.5117507996106244, |
|
"grad_norm": 0.14444314196036126, |
|
"learning_rate": 0.0001855471662881164, |
|
"loss": 0.7482, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.5173133083020443, |
|
"grad_norm": 0.1434929187274412, |
|
"learning_rate": 0.00018503976879874322, |
|
"loss": 0.695, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.5228758169934641, |
|
"grad_norm": 0.13128793518588147, |
|
"learning_rate": 0.0001845243364978702, |
|
"loss": 0.6895, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.5284383256848839, |
|
"grad_norm": 0.12701476810931256, |
|
"learning_rate": 0.0001840009180850815, |
|
"loss": 0.7061, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.5340008343763037, |
|
"grad_norm": 0.1264914632820831, |
|
"learning_rate": 0.00018346956301451304, |
|
"loss": 0.6962, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.5395633430677236, |
|
"grad_norm": 0.13208739238621547, |
|
"learning_rate": 0.00018293032149017984, |
|
"loss": 0.6863, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.5451258517591434, |
|
"grad_norm": 0.13323142479850197, |
|
"learning_rate": 0.00018238324446123266, |
|
"loss": 0.7039, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.5506883604505632, |
|
"grad_norm": 0.1588926427932463, |
|
"learning_rate": 0.0001818283836171441, |
|
"loss": 0.6924, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.5562508691419831, |
|
"grad_norm": 0.12667050310382044, |
|
"learning_rate": 0.00018126579138282503, |
|
"loss": 0.6991, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.5618133778334029, |
|
"grad_norm": 0.13081757063531854, |
|
"learning_rate": 0.000180695520913671, |
|
"loss": 0.6909, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.5673758865248227, |
|
"grad_norm": 0.13598420562667687, |
|
"learning_rate": 0.0001801176260905402, |
|
"loss": 0.7088, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.5729383952162426, |
|
"grad_norm": 0.1300574822823029, |
|
"learning_rate": 0.00017953216151466263, |
|
"loss": 0.6821, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.5785009039076624, |
|
"grad_norm": 0.14206907328811608, |
|
"learning_rate": 0.00017893918250248104, |
|
"loss": 0.6981, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.5840634125990822, |
|
"grad_norm": 0.1557643228084498, |
|
"learning_rate": 0.00017833874508042466, |
|
"loss": 0.7018, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.5896259212905021, |
|
"grad_norm": 0.13942706653578768, |
|
"learning_rate": 0.00017773090597961554, |
|
"loss": 0.6738, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.5951884299819219, |
|
"grad_norm": 0.12004314100744409, |
|
"learning_rate": 0.00017711572263050845, |
|
"loss": 0.7039, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.6007509386733417, |
|
"grad_norm": 0.12451500019905493, |
|
"learning_rate": 0.00017649325315746478, |
|
"loss": 0.7034, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.6063134473647616, |
|
"grad_norm": 0.1225444976108732, |
|
"learning_rate": 0.00017586355637326054, |
|
"loss": 0.709, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.6118759560561814, |
|
"grad_norm": 0.14795401636462246, |
|
"learning_rate": 0.00017522669177352977, |
|
"loss": 0.7415, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.6174384647476012, |
|
"grad_norm": 0.13482563035315856, |
|
"learning_rate": 0.00017458271953114317, |
|
"loss": 0.7117, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.6230009734390209, |
|
"grad_norm": 0.15422494211156157, |
|
"learning_rate": 0.0001739317004905227, |
|
"loss": 0.7027, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.6285634821304408, |
|
"grad_norm": 0.15479270007272883, |
|
"learning_rate": 0.000173273696161893, |
|
"loss": 0.7174, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.6341259908218606, |
|
"grad_norm": 0.13736882345033063, |
|
"learning_rate": 0.00017260876871546936, |
|
"loss": 0.7072, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.6396884995132804, |
|
"grad_norm": 0.130113338971742, |
|
"learning_rate": 0.00017193698097558416, |
|
"loss": 0.6987, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.6452510082047003, |
|
"grad_norm": 0.15120739345519632, |
|
"learning_rate": 0.00017125839641475072, |
|
"loss": 0.693, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.6508135168961201, |
|
"grad_norm": 0.17345133663648482, |
|
"learning_rate": 0.00017057307914766624, |
|
"loss": 0.6933, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.6563760255875399, |
|
"grad_norm": 0.14577660905768358, |
|
"learning_rate": 0.0001698810939251543, |
|
"loss": 0.6847, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.6619385342789598, |
|
"grad_norm": 0.13729541550408036, |
|
"learning_rate": 0.00016918250612804673, |
|
"loss": 0.7058, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.6675010429703796, |
|
"grad_norm": 0.12792609813747083, |
|
"learning_rate": 0.00016847738176100632, |
|
"loss": 0.7203, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.6675010429703796, |
|
"eval_loss": 0.6808404326438904, |
|
"eval_runtime": 38.0896, |
|
"eval_samples_per_second": 10.685, |
|
"eval_steps_per_second": 0.683, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.6730635516617994, |
|
"grad_norm": 0.12614147228685735, |
|
"learning_rate": 0.00016776578744629052, |
|
"loss": 0.6984, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.6786260603532193, |
|
"grad_norm": 0.14053504602609868, |
|
"learning_rate": 0.00016704779041745686, |
|
"loss": 0.7017, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.6841885690446391, |
|
"grad_norm": 0.144524525939437, |
|
"learning_rate": 0.00016632345851301031, |
|
"loss": 0.7031, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.6897510777360589, |
|
"grad_norm": 0.1240886307637927, |
|
"learning_rate": 0.000165592860169994, |
|
"loss": 0.6856, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.6953135864274788, |
|
"grad_norm": 0.12592281039225006, |
|
"learning_rate": 0.0001648560644175227, |
|
"loss": 0.6832, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.7008760951188986, |
|
"grad_norm": 0.1312108925603738, |
|
"learning_rate": 0.00016411314087026106, |
|
"loss": 0.7266, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.7064386038103184, |
|
"grad_norm": 0.12871371439878188, |
|
"learning_rate": 0.00016336415972184612, |
|
"loss": 0.6892, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.7120011125017383, |
|
"grad_norm": 0.1423393128313328, |
|
"learning_rate": 0.00016260919173825508, |
|
"loss": 0.684, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.7175636211931581, |
|
"grad_norm": 0.1452224910144209, |
|
"learning_rate": 0.00016184830825111924, |
|
"loss": 0.6824, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.7231261298845779, |
|
"grad_norm": 0.12119255762333672, |
|
"learning_rate": 0.00016108158115098444, |
|
"loss": 0.686, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.7286886385759977, |
|
"grad_norm": 0.1411776440916737, |
|
"learning_rate": 0.0001603090828805185, |
|
"loss": 0.6878, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.7342511472674176, |
|
"grad_norm": 0.17300629248103022, |
|
"learning_rate": 0.0001595308864276666, |
|
"loss": 0.7139, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.7398136559588374, |
|
"grad_norm": 0.1458091025518703, |
|
"learning_rate": 0.0001587470653187553, |
|
"loss": 0.729, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.7453761646502572, |
|
"grad_norm": 0.14056535678032148, |
|
"learning_rate": 0.00015795769361154547, |
|
"loss": 0.6972, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.7509386733416771, |
|
"grad_norm": 0.13734797531202647, |
|
"learning_rate": 0.000157162845888235, |
|
"loss": 0.6779, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.7565011820330969, |
|
"grad_norm": 0.12703010084902983, |
|
"learning_rate": 0.00015636259724841222, |
|
"loss": 0.7027, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.7620636907245167, |
|
"grad_norm": 0.12697872232577906, |
|
"learning_rate": 0.00015555702330196023, |
|
"loss": 0.7079, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.7676261994159366, |
|
"grad_norm": 0.15015643114183339, |
|
"learning_rate": 0.00015474620016191294, |
|
"loss": 0.6938, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.7731887081073564, |
|
"grad_norm": 0.1468310767380953, |
|
"learning_rate": 0.00015393020443726381, |
|
"loss": 0.7061, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.7787512167987762, |
|
"grad_norm": 0.1178258952392868, |
|
"learning_rate": 0.00015310911322572753, |
|
"loss": 0.6635, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.7843137254901961, |
|
"grad_norm": 0.1182521343472938, |
|
"learning_rate": 0.00015228300410645556, |
|
"loss": 0.6783, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.7898762341816159, |
|
"grad_norm": 0.11865217322316654, |
|
"learning_rate": 0.00015145195513270644, |
|
"loss": 0.6822, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.7954387428730357, |
|
"grad_norm": 0.11500353884529863, |
|
"learning_rate": 0.00015061604482447075, |
|
"loss": 0.6668, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.8010012515644556, |
|
"grad_norm": 0.1314805700111582, |
|
"learning_rate": 0.0001497753521610526, |
|
"loss": 0.7105, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.8065637602558754, |
|
"grad_norm": 0.14124896793550357, |
|
"learning_rate": 0.00014892995657360717, |
|
"loss": 0.7142, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.8121262689472952, |
|
"grad_norm": 0.13739934885223082, |
|
"learning_rate": 0.00014807993793763619, |
|
"loss": 0.6815, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.8176887776387151, |
|
"grad_norm": 0.13281564930523101, |
|
"learning_rate": 0.0001472253765654406, |
|
"loss": 0.67, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.8232512863301349, |
|
"grad_norm": 0.11821804301835571, |
|
"learning_rate": 0.00014636635319853275, |
|
"loss": 0.6969, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.8288137950215547, |
|
"grad_norm": 0.13929966861379298, |
|
"learning_rate": 0.00014550294900000753, |
|
"loss": 0.696, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.8343763037129746, |
|
"grad_norm": 0.1358532247107506, |
|
"learning_rate": 0.00014463524554687399, |
|
"loss": 0.6987, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.8399388124043944, |
|
"grad_norm": 0.14607145415562992, |
|
"learning_rate": 0.00014376332482234747, |
|
"loss": 0.6713, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.8455013210958142, |
|
"grad_norm": 0.13045660818329083, |
|
"learning_rate": 0.0001428872692081038, |
|
"loss": 0.6782, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.851063829787234, |
|
"grad_norm": 0.14868265540995262, |
|
"learning_rate": 0.00014200716147649557, |
|
"loss": 0.6856, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.8566263384786539, |
|
"grad_norm": 0.1523898280318719, |
|
"learning_rate": 0.00014112308478273145, |
|
"loss": 0.7077, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.8621888471700737, |
|
"grad_norm": 0.14374188108756408, |
|
"learning_rate": 0.00014023512265701955, |
|
"loss": 0.6557, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.8677513558614935, |
|
"grad_norm": 0.13285163140650738, |
|
"learning_rate": 0.00013934335899667527, |
|
"loss": 0.7028, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.8733138645529134, |
|
"grad_norm": 0.12419599179555058, |
|
"learning_rate": 0.0001384478780581942, |
|
"loss": 0.6753, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.8788763732443332, |
|
"grad_norm": 0.12619683911200477, |
|
"learning_rate": 0.00013754876444929166, |
|
"loss": 0.7033, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.884438881935753, |
|
"grad_norm": 0.14380964532401144, |
|
"learning_rate": 0.00013664610312090838, |
|
"loss": 0.7022, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.8900013906271729, |
|
"grad_norm": 0.13341619640026267, |
|
"learning_rate": 0.0001357399793591844, |
|
"loss": 0.703, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.8900013906271729, |
|
"eval_loss": 0.6755693554878235, |
|
"eval_runtime": 38.0742, |
|
"eval_samples_per_second": 10.69, |
|
"eval_steps_per_second": 0.683, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.8955638993185927, |
|
"grad_norm": 0.11373621352884696, |
|
"learning_rate": 0.00013483047877740055, |
|
"loss": 0.6554, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.9011264080100125, |
|
"grad_norm": 0.1418177008323639, |
|
"learning_rate": 0.00013391768730789002, |
|
"loss": 0.7435, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.9066889167014324, |
|
"grad_norm": 0.14095795493706695, |
|
"learning_rate": 0.00013300169119391864, |
|
"loss": 0.6856, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.9122514253928522, |
|
"grad_norm": 0.13583867929382498, |
|
"learning_rate": 0.00013208257698153677, |
|
"loss": 0.6791, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.917813934084272, |
|
"grad_norm": 0.13643641315332133, |
|
"learning_rate": 0.00013116043151140203, |
|
"loss": 0.6958, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.9233764427756919, |
|
"grad_norm": 0.12156612730761757, |
|
"learning_rate": 0.00013023534191057426, |
|
"loss": 0.6912, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.9289389514671117, |
|
"grad_norm": 0.14020483760043728, |
|
"learning_rate": 0.0001293073955842836, |
|
"loss": 0.7086, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.9345014601585315, |
|
"grad_norm": 0.13715121587687112, |
|
"learning_rate": 0.0001283766802076722, |
|
"loss": 0.6804, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.9400639688499514, |
|
"grad_norm": 0.12395531650931035, |
|
"learning_rate": 0.00012744328371751024, |
|
"loss": 0.7246, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.9456264775413712, |
|
"grad_norm": 0.1351012642378018, |
|
"learning_rate": 0.00012650729430388764, |
|
"loss": 0.6695, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.951188986232791, |
|
"grad_norm": 0.14258700665297835, |
|
"learning_rate": 0.00012556880040188144, |
|
"loss": 0.6844, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.9567514949242109, |
|
"grad_norm": 0.14370950835852644, |
|
"learning_rate": 0.00012462789068320017, |
|
"loss": 0.7156, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.9623140036156307, |
|
"grad_norm": 0.1390057701240165, |
|
"learning_rate": 0.00012368465404780598, |
|
"loss": 0.6652, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.9678765123070505, |
|
"grad_norm": 0.127046163008197, |
|
"learning_rate": 0.00012273917961551513, |
|
"loss": 0.6641, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.9734390209984704, |
|
"grad_norm": 0.14753709766070622, |
|
"learning_rate": 0.00012179155671757754, |
|
"loss": 0.689, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.9790015296898902, |
|
"grad_norm": 0.13366681839894046, |
|
"learning_rate": 0.00012084187488823657, |
|
"loss": 0.6938, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.98456403838131, |
|
"grad_norm": 0.14364888110758262, |
|
"learning_rate": 0.00011989022385626967, |
|
"loss": 0.6867, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.9901265470727298, |
|
"grad_norm": 0.13750035239201705, |
|
"learning_rate": 0.00011893669353651031, |
|
"loss": 0.6926, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.9956890557641497, |
|
"grad_norm": 0.12165819999979594, |
|
"learning_rate": 0.00011798137402135277, |
|
"loss": 0.6781, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 1.0012515644555695, |
|
"grad_norm": 0.13460664662956676, |
|
"learning_rate": 0.00011702435557223987, |
|
"loss": 0.6907, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.0068140731469892, |
|
"grad_norm": 0.14812357106021773, |
|
"learning_rate": 0.00011606572861113474, |
|
"loss": 0.6739, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 1.0123765818384092, |
|
"grad_norm": 0.14414234656282895, |
|
"learning_rate": 0.00011510558371197753, |
|
"loss": 0.6581, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.017939090529829, |
|
"grad_norm": 0.1417200021185479, |
|
"learning_rate": 0.00011414401159212778, |
|
"loss": 0.6886, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 1.0235015992212488, |
|
"grad_norm": 0.13876983385482528, |
|
"learning_rate": 0.00011318110310379301, |
|
"loss": 0.6737, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.0290641079126686, |
|
"grad_norm": 0.15251072221895987, |
|
"learning_rate": 0.00011221694922544491, |
|
"loss": 0.6692, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 1.0346266166040885, |
|
"grad_norm": 0.14026789842910758, |
|
"learning_rate": 0.0001112516410532233, |
|
"loss": 0.6562, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.0401891252955082, |
|
"grad_norm": 0.12707738202474425, |
|
"learning_rate": 0.00011028526979232913, |
|
"loss": 0.6618, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 1.0457516339869282, |
|
"grad_norm": 0.14606846473361, |
|
"learning_rate": 0.00010931792674840718, |
|
"loss": 0.6746, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.0513141426783479, |
|
"grad_norm": 0.14011972796939615, |
|
"learning_rate": 0.00010834970331891914, |
|
"loss": 0.6869, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 1.0568766513697678, |
|
"grad_norm": 0.1563818365447484, |
|
"learning_rate": 0.0001073806909845082, |
|
"loss": 0.66, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.0624391600611875, |
|
"grad_norm": 0.12113541850928697, |
|
"learning_rate": 0.00010641098130035562, |
|
"loss": 0.6567, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 1.0680016687526075, |
|
"grad_norm": 0.12784803662349467, |
|
"learning_rate": 0.00010544066588753044, |
|
"loss": 0.6478, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.0735641774440272, |
|
"grad_norm": 0.13555213849737605, |
|
"learning_rate": 0.00010446983642433259, |
|
"loss": 0.6423, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 1.0791266861354472, |
|
"grad_norm": 0.14956837158318062, |
|
"learning_rate": 0.00010349858463763113, |
|
"loss": 0.6482, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.0846891948268669, |
|
"grad_norm": 0.14302685723153047, |
|
"learning_rate": 0.0001025270022941975, |
|
"loss": 0.6619, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 1.0902517035182868, |
|
"grad_norm": 0.1531273940282806, |
|
"learning_rate": 0.0001015551811920351, |
|
"loss": 0.6818, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.0958142122097065, |
|
"grad_norm": 0.1431917561070241, |
|
"learning_rate": 0.00010058321315170596, |
|
"loss": 0.6597, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 1.1013767209011265, |
|
"grad_norm": 0.14533278983667672, |
|
"learning_rate": 9.961119000765531e-05, |
|
"loss": 0.641, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.1069392295925462, |
|
"grad_norm": 0.1426693103942941, |
|
"learning_rate": 9.863920359953474e-05, |
|
"loss": 0.6864, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 1.1125017382839661, |
|
"grad_norm": 0.14515333605223138, |
|
"learning_rate": 9.766734576352478e-05, |
|
"loss": 0.6759, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.1125017382839661, |
|
"eval_loss": 0.6748408079147339, |
|
"eval_runtime": 38.0571, |
|
"eval_samples_per_second": 10.694, |
|
"eval_steps_per_second": 0.683, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.1180642469753859, |
|
"grad_norm": 0.13884691864173954, |
|
"learning_rate": 9.669570832365838e-05, |
|
"loss": 0.6536, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 1.1236267556668058, |
|
"grad_norm": 0.14190147387793983, |
|
"learning_rate": 9.572438308314446e-05, |
|
"loss": 0.6448, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.1291892643582255, |
|
"grad_norm": 0.1460356352229923, |
|
"learning_rate": 9.475346181569467e-05, |
|
"loss": 0.6709, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 1.1347517730496455, |
|
"grad_norm": 0.15004591423623692, |
|
"learning_rate": 9.378303625685195e-05, |
|
"loss": 0.6338, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.1403142817410652, |
|
"grad_norm": 0.14709044334780488, |
|
"learning_rate": 9.281319809532329e-05, |
|
"loss": 0.6482, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 1.1458767904324851, |
|
"grad_norm": 0.1339983467335465, |
|
"learning_rate": 9.18440389643165e-05, |
|
"loss": 0.6564, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.1514392991239049, |
|
"grad_norm": 0.15465369318467384, |
|
"learning_rate": 9.08756504328827e-05, |
|
"loss": 0.631, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 1.1570018078153248, |
|
"grad_norm": 0.14414339429238215, |
|
"learning_rate": 8.990812399726435e-05, |
|
"loss": 0.63, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.1625643165067445, |
|
"grad_norm": 0.1530406190835784, |
|
"learning_rate": 8.894155107225062e-05, |
|
"loss": 0.6599, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 1.1681268251981645, |
|
"grad_norm": 0.18004713578221632, |
|
"learning_rate": 8.797602298254004e-05, |
|
"loss": 0.6588, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.1736893338895842, |
|
"grad_norm": 0.1395761709805775, |
|
"learning_rate": 8.701163095411212e-05, |
|
"loss": 0.6388, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 1.1792518425810041, |
|
"grad_norm": 0.14800345539370582, |
|
"learning_rate": 8.604846610560771e-05, |
|
"loss": 0.6469, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.1848143512724238, |
|
"grad_norm": 0.18153262554102811, |
|
"learning_rate": 8.508661943972021e-05, |
|
"loss": 0.6476, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 1.1903768599638438, |
|
"grad_norm": 0.1460085066143512, |
|
"learning_rate": 8.412618183459708e-05, |
|
"loss": 0.6417, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.1959393686552635, |
|
"grad_norm": 0.1563928211034322, |
|
"learning_rate": 8.316724403525359e-05, |
|
"loss": 0.6655, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 1.2015018773466832, |
|
"grad_norm": 0.13021633068783775, |
|
"learning_rate": 8.220989664499878e-05, |
|
"loss": 0.6495, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.2070643860381032, |
|
"grad_norm": 0.15359952593058618, |
|
"learning_rate": 8.125423011687524e-05, |
|
"loss": 0.6518, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 1.2126268947295231, |
|
"grad_norm": 0.14327581045453255, |
|
"learning_rate": 8.030033474511249e-05, |
|
"loss": 0.6428, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.2181894034209428, |
|
"grad_norm": 0.14877277012088305, |
|
"learning_rate": 7.934830065659599e-05, |
|
"loss": 0.6823, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 1.2237519121123626, |
|
"grad_norm": 0.1350097173259691, |
|
"learning_rate": 7.839821780235168e-05, |
|
"loss": 0.6656, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.2293144208037825, |
|
"grad_norm": 0.14805261697578703, |
|
"learning_rate": 7.74501759490469e-05, |
|
"loss": 0.6649, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 1.2348769294952024, |
|
"grad_norm": 0.14241031431419957, |
|
"learning_rate": 7.650426467050926e-05, |
|
"loss": 0.645, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.2404394381866222, |
|
"grad_norm": 0.14677979540609026, |
|
"learning_rate": 7.556057333926318e-05, |
|
"loss": 0.6322, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 1.2460019468780419, |
|
"grad_norm": 0.1540439921077116, |
|
"learning_rate": 7.461919111808595e-05, |
|
"loss": 0.6425, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.2515644555694618, |
|
"grad_norm": 0.16093917993335288, |
|
"learning_rate": 7.368020695158312e-05, |
|
"loss": 0.6708, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 1.2571269642608818, |
|
"grad_norm": 0.14072702856543545, |
|
"learning_rate": 7.274370955778498e-05, |
|
"loss": 0.6412, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.2626894729523015, |
|
"grad_norm": 0.1372982963866414, |
|
"learning_rate": 7.180978741976397e-05, |
|
"loss": 0.6354, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 1.2682519816437212, |
|
"grad_norm": 0.1361010979443787, |
|
"learning_rate": 7.087852877727481e-05, |
|
"loss": 0.6297, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.2738144903351412, |
|
"grad_norm": 0.15152880519317802, |
|
"learning_rate": 6.995002161841708e-05, |
|
"loss": 0.6648, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 1.279376999026561, |
|
"grad_norm": 0.15111271273299604, |
|
"learning_rate": 6.902435367132208e-05, |
|
"loss": 0.6811, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.2849395077179808, |
|
"grad_norm": 0.1555612197330931, |
|
"learning_rate": 6.810161239586375e-05, |
|
"loss": 0.6639, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 1.2905020164094005, |
|
"grad_norm": 0.1404429960918057, |
|
"learning_rate": 6.718188497539554e-05, |
|
"loss": 0.6416, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.2960645251008205, |
|
"grad_norm": 0.1493351620856526, |
|
"learning_rate": 6.626525830851267e-05, |
|
"loss": 0.6284, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 1.3016270337922404, |
|
"grad_norm": 0.15944635272625493, |
|
"learning_rate": 6.535181900084206e-05, |
|
"loss": 0.6402, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.3071895424836601, |
|
"grad_norm": 0.1534235543873645, |
|
"learning_rate": 6.444165335685927e-05, |
|
"loss": 0.6373, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 1.3127520511750799, |
|
"grad_norm": 0.1404290668723592, |
|
"learning_rate": 6.35348473717345e-05, |
|
"loss": 0.6596, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.3183145598664998, |
|
"grad_norm": 0.15190370802350445, |
|
"learning_rate": 6.263148672320714e-05, |
|
"loss": 0.6519, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 1.3238770685579198, |
|
"grad_norm": 0.1396377515861422, |
|
"learning_rate": 6.173165676349103e-05, |
|
"loss": 0.6539, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.3294395772493395, |
|
"grad_norm": 0.1413554398489568, |
|
"learning_rate": 6.083544251120993e-05, |
|
"loss": 0.6436, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 1.3350020859407592, |
|
"grad_norm": 0.175529243963575, |
|
"learning_rate": 5.9942928643364724e-05, |
|
"loss": 0.6533, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.3350020859407592, |
|
"eval_loss": 0.6695060133934021, |
|
"eval_runtime": 38.0647, |
|
"eval_samples_per_second": 10.692, |
|
"eval_steps_per_second": 0.683, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.3405645946321791, |
|
"grad_norm": 0.16769090072157003, |
|
"learning_rate": 5.905419948733302e-05, |
|
"loss": 0.6434, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 1.3461271033235989, |
|
"grad_norm": 0.17558035684023568, |
|
"learning_rate": 5.816933901290136e-05, |
|
"loss": 0.6954, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.3516896120150188, |
|
"grad_norm": 0.15028360304997448, |
|
"learning_rate": 5.728843082433193e-05, |
|
"loss": 0.6563, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 1.3572521207064385, |
|
"grad_norm": 0.16888366454421866, |
|
"learning_rate": 5.6411558152462894e-05, |
|
"loss": 0.6383, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.3628146293978585, |
|
"grad_norm": 0.1507413542567497, |
|
"learning_rate": 5.553880384684493e-05, |
|
"loss": 0.6382, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 1.3683771380892782, |
|
"grad_norm": 0.15944821317229088, |
|
"learning_rate": 5.4670250367913023e-05, |
|
"loss": 0.6589, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.3739396467806981, |
|
"grad_norm": 0.14215107373647873, |
|
"learning_rate": 5.380597977919557e-05, |
|
"loss": 0.6346, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 1.3795021554721179, |
|
"grad_norm": 0.16243215498334238, |
|
"learning_rate": 5.2946073739560706e-05, |
|
"loss": 0.6527, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.3850646641635378, |
|
"grad_norm": 0.1562039482784561, |
|
"learning_rate": 5.209061349550095e-05, |
|
"loss": 0.66, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 1.3906271728549575, |
|
"grad_norm": 0.1680769018618033, |
|
"learning_rate": 5.1239679873456634e-05, |
|
"loss": 0.6699, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.3961896815463775, |
|
"grad_norm": 0.1478360199044516, |
|
"learning_rate": 5.039335327217951e-05, |
|
"loss": 0.6502, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 1.4017521902377972, |
|
"grad_norm": 0.14915391173643802, |
|
"learning_rate": 4.955171365513603e-05, |
|
"loss": 0.6143, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.4073146989292171, |
|
"grad_norm": 0.15851697773820939, |
|
"learning_rate": 4.871484054295258e-05, |
|
"loss": 0.6572, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 1.4128772076206368, |
|
"grad_norm": 0.15216697385980452, |
|
"learning_rate": 4.7882813005901696e-05, |
|
"loss": 0.657, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.4184397163120568, |
|
"grad_norm": 0.1437749041937587, |
|
"learning_rate": 4.705570965643176e-05, |
|
"loss": 0.6411, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 1.4240022250034765, |
|
"grad_norm": 0.16683978696735308, |
|
"learning_rate": 4.623360864173893e-05, |
|
"loss": 0.651, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.4295647336948965, |
|
"grad_norm": 0.1621572781654395, |
|
"learning_rate": 4.541658763638406e-05, |
|
"loss": 0.6492, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 1.4351272423863162, |
|
"grad_norm": 0.15994302691177126, |
|
"learning_rate": 4.460472383495331e-05, |
|
"loss": 0.6577, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.4406897510777361, |
|
"grad_norm": 0.14692563499946057, |
|
"learning_rate": 4.379809394476501e-05, |
|
"loss": 0.6496, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 1.4462522597691558, |
|
"grad_norm": 0.17176689432434047, |
|
"learning_rate": 4.2996774178621736e-05, |
|
"loss": 0.6433, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.4518147684605758, |
|
"grad_norm": 0.1446292926663316, |
|
"learning_rate": 4.220084024760982e-05, |
|
"loss": 0.6223, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 1.4573772771519955, |
|
"grad_norm": 0.15172393816056104, |
|
"learning_rate": 4.141036735394574e-05, |
|
"loss": 0.6481, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.4629397858434154, |
|
"grad_norm": 0.16544844619207105, |
|
"learning_rate": 4.0625430183870796e-05, |
|
"loss": 0.6411, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 1.4685022945348352, |
|
"grad_norm": 0.15934504430033075, |
|
"learning_rate": 3.984610290059467e-05, |
|
"loss": 0.6678, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.474064803226255, |
|
"grad_norm": 0.15204965773692086, |
|
"learning_rate": 3.907245913728807e-05, |
|
"loss": 0.6549, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 1.4796273119176748, |
|
"grad_norm": 0.15256724381583675, |
|
"learning_rate": 3.830457199012585e-05, |
|
"loss": 0.6478, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.4851898206090948, |
|
"grad_norm": 0.15150975920035156, |
|
"learning_rate": 3.754251401138051e-05, |
|
"loss": 0.6385, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 1.4907523293005145, |
|
"grad_norm": 0.17068279689989202, |
|
"learning_rate": 3.678635720256737e-05, |
|
"loss": 0.6659, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.4963148379919344, |
|
"grad_norm": 0.15001304767115145, |
|
"learning_rate": 3.6036173007641435e-05, |
|
"loss": 0.66, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 1.5018773466833542, |
|
"grad_norm": 0.14424590095876771, |
|
"learning_rate": 3.529203230624747e-05, |
|
"loss": 0.651, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.5074398553747739, |
|
"grad_norm": 0.16690307331407334, |
|
"learning_rate": 3.455400540702274e-05, |
|
"loss": 0.6508, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 1.5130023640661938, |
|
"grad_norm": 0.1506720689857848, |
|
"learning_rate": 3.3822162040954354e-05, |
|
"loss": 0.6347, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.5185648727576138, |
|
"grad_norm": 0.15108676554413317, |
|
"learning_rate": 3.309657135479065e-05, |
|
"loss": 0.6449, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 1.5241273814490335, |
|
"grad_norm": 0.1397876543411103, |
|
"learning_rate": 3.237730190450816e-05, |
|
"loss": 0.6382, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.5296898901404532, |
|
"grad_norm": 0.15000237273739636, |
|
"learning_rate": 3.166442164883403e-05, |
|
"loss": 0.6607, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 1.5352523988318731, |
|
"grad_norm": 0.15901665795803346, |
|
"learning_rate": 3.0957997942825336e-05, |
|
"loss": 0.6403, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.540814907523293, |
|
"grad_norm": 0.1500239251091428, |
|
"learning_rate": 3.0258097531504937e-05, |
|
"loss": 0.6595, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 1.5463774162147128, |
|
"grad_norm": 0.16536122507336246, |
|
"learning_rate": 2.9564786543555388e-05, |
|
"loss": 0.6515, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.5519399249061325, |
|
"grad_norm": 0.14948767796744467, |
|
"learning_rate": 2.8878130485070852e-05, |
|
"loss": 0.6341, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 1.5575024335975525, |
|
"grad_norm": 0.16630242366164855, |
|
"learning_rate": 2.819819423336775e-05, |
|
"loss": 0.6458, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.5575024335975525, |
|
"eval_loss": 0.6669326424598694, |
|
"eval_runtime": 38.0635, |
|
"eval_samples_per_second": 10.693, |
|
"eval_steps_per_second": 0.683, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.5630649422889724, |
|
"grad_norm": 0.15092670649471096, |
|
"learning_rate": 2.7525042030855218e-05, |
|
"loss": 0.6285, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 1.5686274509803921, |
|
"grad_norm": 0.17817145439236162, |
|
"learning_rate": 2.6858737478965035e-05, |
|
"loss": 0.6786, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 1.5741899596718119, |
|
"grad_norm": 0.14411321645345596, |
|
"learning_rate": 2.6199343532142573e-05, |
|
"loss": 0.6487, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 1.5797524683632318, |
|
"grad_norm": 0.1729923377951436, |
|
"learning_rate": 2.5546922491898495e-05, |
|
"loss": 0.6466, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.5853149770546517, |
|
"grad_norm": 0.17040206170029934, |
|
"learning_rate": 2.4901536000922497e-05, |
|
"loss": 0.665, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 1.5908774857460715, |
|
"grad_norm": 0.16055497144569197, |
|
"learning_rate": 2.4263245037258995e-05, |
|
"loss": 0.6665, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 1.5964399944374912, |
|
"grad_norm": 0.1655874585771864, |
|
"learning_rate": 2.363210990854582e-05, |
|
"loss": 0.6734, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 1.6020025031289111, |
|
"grad_norm": 0.1651519002975595, |
|
"learning_rate": 2.300819024631603e-05, |
|
"loss": 0.6691, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.607565011820331, |
|
"grad_norm": 0.162990807682259, |
|
"learning_rate": 2.239154500036399e-05, |
|
"loss": 0.6805, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 1.6131275205117508, |
|
"grad_norm": 0.1717600360257139, |
|
"learning_rate": 2.178223243317532e-05, |
|
"loss": 0.6399, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.6186900292031705, |
|
"grad_norm": 0.1739831864715599, |
|
"learning_rate": 2.1180310114422362e-05, |
|
"loss": 0.6689, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 1.6242525378945905, |
|
"grad_norm": 0.15498282638359004, |
|
"learning_rate": 2.058583491552465e-05, |
|
"loss": 0.6675, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.6298150465860104, |
|
"grad_norm": 0.17242195548428726, |
|
"learning_rate": 1.9998863004275593e-05, |
|
"loss": 0.6618, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 1.6353775552774301, |
|
"grad_norm": 0.16154448145412242, |
|
"learning_rate": 1.941944983953552e-05, |
|
"loss": 0.6406, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.6409400639688498, |
|
"grad_norm": 0.1624668193562119, |
|
"learning_rate": 1.884765016599186e-05, |
|
"loss": 0.6223, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 1.6465025726602698, |
|
"grad_norm": 0.16026073767873994, |
|
"learning_rate": 1.8283518008986567e-05, |
|
"loss": 0.6424, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.6520650813516897, |
|
"grad_norm": 0.14391017994489602, |
|
"learning_rate": 1.7727106669411776e-05, |
|
"loss": 0.6323, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 1.6576275900431094, |
|
"grad_norm": 0.16566629921954873, |
|
"learning_rate": 1.7178468718673714e-05, |
|
"loss": 0.6605, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.6631900987345292, |
|
"grad_norm": 0.15202811145565842, |
|
"learning_rate": 1.6637655993725598e-05, |
|
"loss": 0.6745, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 1.668752607425949, |
|
"grad_norm": 0.14529589529297973, |
|
"learning_rate": 1.6104719592169902e-05, |
|
"loss": 0.6415, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.674315116117369, |
|
"grad_norm": 0.16206187759632693, |
|
"learning_rate": 1.5579709867430514e-05, |
|
"loss": 0.6757, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 1.6798776248087888, |
|
"grad_norm": 0.1689334880075168, |
|
"learning_rate": 1.5062676423995247e-05, |
|
"loss": 0.6648, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 1.6854401335002085, |
|
"grad_norm": 0.17002942586162523, |
|
"learning_rate": 1.4553668112729025e-05, |
|
"loss": 0.6593, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 1.6910026421916284, |
|
"grad_norm": 0.15390481859714705, |
|
"learning_rate": 1.4052733026258281e-05, |
|
"loss": 0.6337, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.6965651508830484, |
|
"grad_norm": 0.16322817116461236, |
|
"learning_rate": 1.3559918494427015e-05, |
|
"loss": 0.6671, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 1.702127659574468, |
|
"grad_norm": 0.17381037300710486, |
|
"learning_rate": 1.3075271079825036e-05, |
|
"loss": 0.6452, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 1.7076901682658878, |
|
"grad_norm": 0.14385216586204455, |
|
"learning_rate": 1.2598836573388383e-05, |
|
"loss": 0.6517, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 1.7132526769573078, |
|
"grad_norm": 0.13695810287509047, |
|
"learning_rate": 1.2130659990073146e-05, |
|
"loss": 0.6115, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.7188151856487277, |
|
"grad_norm": 0.15619086164642515, |
|
"learning_rate": 1.1670785564601972e-05, |
|
"loss": 0.6592, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 1.7243776943401474, |
|
"grad_norm": 0.1359615918791439, |
|
"learning_rate": 1.1219256747285045e-05, |
|
"loss": 0.6489, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.7299402030315671, |
|
"grad_norm": 0.1530561734129938, |
|
"learning_rate": 1.0776116199914343e-05, |
|
"loss": 0.6341, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 1.735502711722987, |
|
"grad_norm": 0.1550580447696997, |
|
"learning_rate": 1.0341405791733183e-05, |
|
"loss": 0.679, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.741065220414407, |
|
"grad_norm": 0.15343391714899443, |
|
"learning_rate": 9.915166595480018e-06, |
|
"loss": 0.6174, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 1.7466277291058268, |
|
"grad_norm": 0.14886169752976763, |
|
"learning_rate": 9.49743888350798e-06, |
|
"loss": 0.658, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 1.7521902377972465, |
|
"grad_norm": 0.16009066506717334, |
|
"learning_rate": 9.088262123979652e-06, |
|
"loss": 0.6547, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 1.7577527464886664, |
|
"grad_norm": 0.1527884120085583, |
|
"learning_rate": 8.687674977138116e-06, |
|
"loss": 0.6355, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 1.7633152551800864, |
|
"grad_norm": 0.15622894390259232, |
|
"learning_rate": 8.295715291654205e-06, |
|
"loss": 0.6415, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 1.768877763871506, |
|
"grad_norm": 0.13821573876875842, |
|
"learning_rate": 7.912420101050367e-06, |
|
"loss": 0.6386, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 1.7744402725629258, |
|
"grad_norm": 0.1605607168007031, |
|
"learning_rate": 7.537825620201699e-06, |
|
"loss": 0.6395, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 1.7800027812543457, |
|
"grad_norm": 0.14785735762140897, |
|
"learning_rate": 7.171967241914224e-06, |
|
"loss": 0.632, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.7800027812543457, |
|
"eval_loss": 0.6654813289642334, |
|
"eval_runtime": 38.0745, |
|
"eval_samples_per_second": 10.69, |
|
"eval_steps_per_second": 0.683, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.7855652899457657, |
|
"grad_norm": 0.15930848069616485, |
|
"learning_rate": 6.814879533580898e-06, |
|
"loss": 0.6298, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 1.7911277986371854, |
|
"grad_norm": 0.16408233749219603, |
|
"learning_rate": 6.4665962339156005e-06, |
|
"loss": 0.6383, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 1.7966903073286051, |
|
"grad_norm": 0.1507405199602565, |
|
"learning_rate": 6.127150249765335e-06, |
|
"loss": 0.6538, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 1.802252816020025, |
|
"grad_norm": 0.15449896745701766, |
|
"learning_rate": 5.7965736530010916e-06, |
|
"loss": 0.6443, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 1.807815324711445, |
|
"grad_norm": 0.15430381806456947, |
|
"learning_rate": 5.474897677487711e-06, |
|
"loss": 0.6714, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 1.8133778334028647, |
|
"grad_norm": 0.1522219198315468, |
|
"learning_rate": 5.162152716132662e-06, |
|
"loss": 0.6462, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 1.8189403420942845, |
|
"grad_norm": 0.14950817969442623, |
|
"learning_rate": 4.858368318014572e-06, |
|
"loss": 0.6309, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 1.8245028507857044, |
|
"grad_norm": 0.16343103461857358, |
|
"learning_rate": 4.563573185591219e-06, |
|
"loss": 0.6229, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 1.8300653594771243, |
|
"grad_norm": 0.16046392285596128, |
|
"learning_rate": 4.2777951719877415e-06, |
|
"loss": 0.6526, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 1.835627868168544, |
|
"grad_norm": 0.16358481948661036, |
|
"learning_rate": 4.0010612783648925e-06, |
|
"loss": 0.6407, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.8411903768599638, |
|
"grad_norm": 0.15364511127769814, |
|
"learning_rate": 3.7333976513680093e-06, |
|
"loss": 0.6375, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 1.8467528855513837, |
|
"grad_norm": 0.1582298293905816, |
|
"learning_rate": 3.4748295806564356e-06, |
|
"loss": 0.6486, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 1.8523153942428037, |
|
"grad_norm": 0.1626273297442105, |
|
"learning_rate": 3.2253814965142683e-06, |
|
"loss": 0.6682, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 1.8578779029342232, |
|
"grad_norm": 0.15566169967079257, |
|
"learning_rate": 2.9850769675419774e-06, |
|
"loss": 0.6531, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 1.8634404116256431, |
|
"grad_norm": 0.1785306629362279, |
|
"learning_rate": 2.7539386984296147e-06, |
|
"loss": 0.6406, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 1.869002920317063, |
|
"grad_norm": 0.14973888951794095, |
|
"learning_rate": 2.5319885278115906e-06, |
|
"loss": 0.6441, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.8745654290084828, |
|
"grad_norm": 0.1493522152423315, |
|
"learning_rate": 2.3192474262033638e-06, |
|
"loss": 0.6377, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 1.8801279376999025, |
|
"grad_norm": 0.1611037940241023, |
|
"learning_rate": 2.115735494019966e-06, |
|
"loss": 0.6785, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 1.8856904463913224, |
|
"grad_norm": 0.15767721827989367, |
|
"learning_rate": 1.921471959676957e-06, |
|
"loss": 0.667, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 1.8912529550827424, |
|
"grad_norm": 0.15987860636699017, |
|
"learning_rate": 1.7364751777736332e-06, |
|
"loss": 0.6374, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.896815463774162, |
|
"grad_norm": 0.14274620551242612, |
|
"learning_rate": 1.5607626273588138e-06, |
|
"loss": 0.6305, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 1.9023779724655818, |
|
"grad_norm": 0.1539956730318316, |
|
"learning_rate": 1.394350910279385e-06, |
|
"loss": 0.6446, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 1.9079404811570018, |
|
"grad_norm": 0.16255063308547452, |
|
"learning_rate": 1.2372557496116877e-06, |
|
"loss": 0.6577, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 1.9135029898484217, |
|
"grad_norm": 0.1457148304946188, |
|
"learning_rate": 1.089491988176017e-06, |
|
"loss": 0.6671, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 1.9190654985398414, |
|
"grad_norm": 0.16991284681377727, |
|
"learning_rate": 9.510735871341103e-07, |
|
"loss": 0.6693, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 1.9246280072312612, |
|
"grad_norm": 0.15748923168566117, |
|
"learning_rate": 8.220136246701926e-07, |
|
"loss": 0.6491, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 1.930190515922681, |
|
"grad_norm": 0.15744123180048702, |
|
"learning_rate": 7.023242947552078e-07, |
|
"loss": 0.6366, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 1.935753024614101, |
|
"grad_norm": 0.15109132227982583, |
|
"learning_rate": 5.920169059947411e-07, |
|
"loss": 0.6553, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 1.9413155333055208, |
|
"grad_norm": 0.16542652334128852, |
|
"learning_rate": 4.911018805605406e-07, |
|
"loss": 0.6395, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 1.9468780419969405, |
|
"grad_norm": 0.15918345337184334, |
|
"learning_rate": 3.9958875320580404e-07, |
|
"loss": 0.6502, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.9524405506883604, |
|
"grad_norm": 0.17686582944306006, |
|
"learning_rate": 3.1748617036427843e-07, |
|
"loss": 0.6586, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 1.9580030593797804, |
|
"grad_norm": 0.1426907741483869, |
|
"learning_rate": 2.448018893333681e-07, |
|
"loss": 0.6345, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 1.9635655680712, |
|
"grad_norm": 0.14223763889913488, |
|
"learning_rate": 1.815427775411549e-07, |
|
"loss": 0.6296, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 1.9691280767626198, |
|
"grad_norm": 0.14802968415791257, |
|
"learning_rate": 1.277148118975835e-07, |
|
"loss": 0.6167, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 1.9746905854540397, |
|
"grad_norm": 0.16664999019696625, |
|
"learning_rate": 8.332307822971342e-08, |
|
"loss": 0.6522, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 1.9802530941454597, |
|
"grad_norm": 0.15531644850942158, |
|
"learning_rate": 4.837177080119215e-08, |
|
"loss": 0.6587, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 1.9858156028368794, |
|
"grad_norm": 0.17774727599107124, |
|
"learning_rate": 2.286419191601663e-08, |
|
"loss": 0.6634, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 1.9913781115282991, |
|
"grad_norm": 0.1591360139721982, |
|
"learning_rate": 6.8027516064606e-09, |
|
"loss": 0.6466, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 1.996940620219719, |
|
"grad_norm": 0.1816483541352177, |
|
"learning_rate": 1.8896740540119696e-10, |
|
"loss": 0.6307, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 1.998053121958003, |
|
"step": 1796, |
|
"total_flos": 7795310658584576.0, |
|
"train_loss": 0.6907267017260957, |
|
"train_runtime": 37673.7111, |
|
"train_samples_per_second": 3.054, |
|
"train_steps_per_second": 0.048 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 1796, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 100, |
|
"total_flos": 7795310658584576.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|