yihanwang617's picture
Model save
6fc75e2 verified
raw
history blame
65.3 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.998053121958003,
"eval_steps": 200,
"global_step": 1796,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.001112501738283966,
"grad_norm": 0.6684235958622939,
"learning_rate": 1.1111111111111112e-06,
"loss": 1.1751,
"step": 1
},
{
"epoch": 0.00556250869141983,
"grad_norm": 0.7610951352005912,
"learning_rate": 5.555555555555556e-06,
"loss": 1.2823,
"step": 5
},
{
"epoch": 0.01112501738283966,
"grad_norm": 0.8690361671075456,
"learning_rate": 1.1111111111111112e-05,
"loss": 1.2894,
"step": 10
},
{
"epoch": 0.016687526074259492,
"grad_norm": 1.059736284119422,
"learning_rate": 1.6666666666666667e-05,
"loss": 1.3166,
"step": 15
},
{
"epoch": 0.02225003476567932,
"grad_norm": 0.4906355740047051,
"learning_rate": 2.2222222222222223e-05,
"loss": 1.1771,
"step": 20
},
{
"epoch": 0.027812543457099152,
"grad_norm": 1.04376108340149,
"learning_rate": 2.777777777777778e-05,
"loss": 1.1062,
"step": 25
},
{
"epoch": 0.033375052148518984,
"grad_norm": 0.47593208759368216,
"learning_rate": 3.3333333333333335e-05,
"loss": 1.014,
"step": 30
},
{
"epoch": 0.03893756083993881,
"grad_norm": 0.2526378868894225,
"learning_rate": 3.888888888888889e-05,
"loss": 0.9027,
"step": 35
},
{
"epoch": 0.04450006953135864,
"grad_norm": 0.2750620823042748,
"learning_rate": 4.4444444444444447e-05,
"loss": 0.8556,
"step": 40
},
{
"epoch": 0.05006257822277847,
"grad_norm": 0.22806716350963602,
"learning_rate": 5e-05,
"loss": 0.8714,
"step": 45
},
{
"epoch": 0.055625086914198305,
"grad_norm": 0.31305577393443185,
"learning_rate": 5.555555555555556e-05,
"loss": 0.8431,
"step": 50
},
{
"epoch": 0.061187595605618136,
"grad_norm": 0.3231781586973102,
"learning_rate": 6.111111111111112e-05,
"loss": 0.7896,
"step": 55
},
{
"epoch": 0.06675010429703797,
"grad_norm": 0.2652381898767101,
"learning_rate": 6.666666666666667e-05,
"loss": 0.8031,
"step": 60
},
{
"epoch": 0.0723126129884578,
"grad_norm": 0.16353190746623922,
"learning_rate": 7.222222222222222e-05,
"loss": 0.798,
"step": 65
},
{
"epoch": 0.07787512167987762,
"grad_norm": 0.14957803129107605,
"learning_rate": 7.777777777777778e-05,
"loss": 0.8214,
"step": 70
},
{
"epoch": 0.08343763037129745,
"grad_norm": 0.1558832457817548,
"learning_rate": 8.333333333333334e-05,
"loss": 0.8046,
"step": 75
},
{
"epoch": 0.08900013906271728,
"grad_norm": 0.15611943332427167,
"learning_rate": 8.888888888888889e-05,
"loss": 0.7725,
"step": 80
},
{
"epoch": 0.09456264775413711,
"grad_norm": 0.17878911552246932,
"learning_rate": 9.444444444444444e-05,
"loss": 0.758,
"step": 85
},
{
"epoch": 0.10012515644555695,
"grad_norm": 0.21152631925896626,
"learning_rate": 0.0001,
"loss": 0.7741,
"step": 90
},
{
"epoch": 0.10568766513697678,
"grad_norm": 0.15798064484261687,
"learning_rate": 0.00010555555555555557,
"loss": 0.7411,
"step": 95
},
{
"epoch": 0.11125017382839661,
"grad_norm": 0.1770136407809846,
"learning_rate": 0.00011111111111111112,
"loss": 0.7582,
"step": 100
},
{
"epoch": 0.11681268251981644,
"grad_norm": 0.1721697060786133,
"learning_rate": 0.00011666666666666668,
"loss": 0.7908,
"step": 105
},
{
"epoch": 0.12237519121123627,
"grad_norm": 0.19163940220247466,
"learning_rate": 0.00012222222222222224,
"loss": 0.7769,
"step": 110
},
{
"epoch": 0.1279376999026561,
"grad_norm": 0.17258655274801674,
"learning_rate": 0.00012777777777777776,
"loss": 0.7291,
"step": 115
},
{
"epoch": 0.13350020859407594,
"grad_norm": 0.14647192105198578,
"learning_rate": 0.00013333333333333334,
"loss": 0.7147,
"step": 120
},
{
"epoch": 0.13906271728549577,
"grad_norm": 0.15483066376892327,
"learning_rate": 0.0001388888888888889,
"loss": 0.7253,
"step": 125
},
{
"epoch": 0.1446252259769156,
"grad_norm": 0.15985722819525555,
"learning_rate": 0.00014444444444444444,
"loss": 0.74,
"step": 130
},
{
"epoch": 0.15018773466833543,
"grad_norm": 0.18677759796975052,
"learning_rate": 0.00015000000000000001,
"loss": 0.7578,
"step": 135
},
{
"epoch": 0.15575024335975524,
"grad_norm": 0.1502765821288928,
"learning_rate": 0.00015555555555555556,
"loss": 0.7541,
"step": 140
},
{
"epoch": 0.16131275205117507,
"grad_norm": 0.16754721310239581,
"learning_rate": 0.0001611111111111111,
"loss": 0.7338,
"step": 145
},
{
"epoch": 0.1668752607425949,
"grad_norm": 0.15603842447844732,
"learning_rate": 0.0001666666666666667,
"loss": 0.7347,
"step": 150
},
{
"epoch": 0.17243776943401473,
"grad_norm": 0.16036041756367805,
"learning_rate": 0.00017222222222222224,
"loss": 0.7275,
"step": 155
},
{
"epoch": 0.17800027812543456,
"grad_norm": 0.17275116566192827,
"learning_rate": 0.00017777777777777779,
"loss": 0.7357,
"step": 160
},
{
"epoch": 0.1835627868168544,
"grad_norm": 0.16362661400179987,
"learning_rate": 0.00018333333333333334,
"loss": 0.7454,
"step": 165
},
{
"epoch": 0.18912529550827423,
"grad_norm": 0.1529061378986909,
"learning_rate": 0.00018888888888888888,
"loss": 0.7305,
"step": 170
},
{
"epoch": 0.19468780419969406,
"grad_norm": 0.16390278757569096,
"learning_rate": 0.00019444444444444446,
"loss": 0.7383,
"step": 175
},
{
"epoch": 0.2002503128911139,
"grad_norm": 0.16834812585150158,
"learning_rate": 0.0002,
"loss": 0.7067,
"step": 180
},
{
"epoch": 0.20581282158253372,
"grad_norm": 0.15601116057262598,
"learning_rate": 0.0001999952758505736,
"loss": 0.7567,
"step": 185
},
{
"epoch": 0.21137533027395355,
"grad_norm": 0.15788111366123353,
"learning_rate": 0.00019998110384864614,
"loss": 0.7434,
"step": 190
},
{
"epoch": 0.2169378389653734,
"grad_norm": 0.16327699685803637,
"learning_rate": 0.00019995748533323075,
"loss": 0.7087,
"step": 195
},
{
"epoch": 0.22250034765679322,
"grad_norm": 0.14758830274997248,
"learning_rate": 0.0001999244225358753,
"loss": 0.718,
"step": 200
},
{
"epoch": 0.22250034765679322,
"eval_loss": 0.7089951634407043,
"eval_runtime": 38.6391,
"eval_samples_per_second": 10.533,
"eval_steps_per_second": 0.673,
"step": 200
},
{
"epoch": 0.22806285634821305,
"grad_norm": 0.16285906061792327,
"learning_rate": 0.00019988191858045178,
"loss": 0.7323,
"step": 205
},
{
"epoch": 0.23362536503963288,
"grad_norm": 0.15445314974456065,
"learning_rate": 0.00019982997748286082,
"loss": 0.7427,
"step": 210
},
{
"epoch": 0.2391878737310527,
"grad_norm": 0.14969344173380988,
"learning_rate": 0.00019976860415065256,
"loss": 0.7171,
"step": 215
},
{
"epoch": 0.24475038242247255,
"grad_norm": 0.15378571057430904,
"learning_rate": 0.00019969780438256293,
"loss": 0.726,
"step": 220
},
{
"epoch": 0.2503128911138924,
"grad_norm": 0.1655361908944567,
"learning_rate": 0.00019961758486796555,
"loss": 0.7463,
"step": 225
},
{
"epoch": 0.2558753998053122,
"grad_norm": 0.1343012669489156,
"learning_rate": 0.00019952795318623986,
"loss": 0.7115,
"step": 230
},
{
"epoch": 0.26143790849673204,
"grad_norm": 0.1597015078714079,
"learning_rate": 0.00019942891780605512,
"loss": 0.6901,
"step": 235
},
{
"epoch": 0.2670004171881519,
"grad_norm": 0.13156659502510382,
"learning_rate": 0.0001993204880845699,
"loss": 0.6974,
"step": 240
},
{
"epoch": 0.2725629258795717,
"grad_norm": 0.13487041292813268,
"learning_rate": 0.0001992026742665484,
"loss": 0.7264,
"step": 245
},
{
"epoch": 0.27812543457099154,
"grad_norm": 0.1452928301770259,
"learning_rate": 0.00019907548748339222,
"loss": 0.7132,
"step": 250
},
{
"epoch": 0.28368794326241137,
"grad_norm": 0.13005536436409637,
"learning_rate": 0.00019893893975208874,
"loss": 0.6924,
"step": 255
},
{
"epoch": 0.2892504519538312,
"grad_norm": 0.15551390193086698,
"learning_rate": 0.0001987930439740757,
"loss": 0.7137,
"step": 260
},
{
"epoch": 0.29481296064525103,
"grad_norm": 0.1411196790735591,
"learning_rate": 0.0001986378139340222,
"loss": 0.7003,
"step": 265
},
{
"epoch": 0.30037546933667086,
"grad_norm": 0.16701635405751747,
"learning_rate": 0.0001984732642985263,
"loss": 0.7318,
"step": 270
},
{
"epoch": 0.3059379780280907,
"grad_norm": 0.15781390142645427,
"learning_rate": 0.00019829941061472943,
"loss": 0.725,
"step": 275
},
{
"epoch": 0.31150048671951047,
"grad_norm": 0.1276451211299801,
"learning_rate": 0.0001981162693088471,
"loss": 0.7008,
"step": 280
},
{
"epoch": 0.3170629954109303,
"grad_norm": 0.14226714916323183,
"learning_rate": 0.00019792385768461723,
"loss": 0.7215,
"step": 285
},
{
"epoch": 0.32262550410235014,
"grad_norm": 0.15169879239239556,
"learning_rate": 0.00019772219392166519,
"loss": 0.7298,
"step": 290
},
{
"epoch": 0.32818801279376997,
"grad_norm": 0.13790872011113864,
"learning_rate": 0.00019751129707378583,
"loss": 0.7024,
"step": 295
},
{
"epoch": 0.3337505214851898,
"grad_norm": 0.13510983391572584,
"learning_rate": 0.00019729118706714375,
"loss": 0.7301,
"step": 300
},
{
"epoch": 0.33931303017660963,
"grad_norm": 0.13421290102386768,
"learning_rate": 0.00019706188469839012,
"loss": 0.6973,
"step": 305
},
{
"epoch": 0.34487553886802946,
"grad_norm": 0.17987364390951446,
"learning_rate": 0.000196823411632698,
"loss": 0.7285,
"step": 310
},
{
"epoch": 0.3504380475594493,
"grad_norm": 0.1463794049975603,
"learning_rate": 0.00019657579040171537,
"loss": 0.7474,
"step": 315
},
{
"epoch": 0.3560005562508691,
"grad_norm": 0.14423415874709103,
"learning_rate": 0.00019631904440143612,
"loss": 0.6938,
"step": 320
},
{
"epoch": 0.36156306494228896,
"grad_norm": 0.11858366900169025,
"learning_rate": 0.00019605319788998965,
"loss": 0.6845,
"step": 325
},
{
"epoch": 0.3671255736337088,
"grad_norm": 0.14664958369445155,
"learning_rate": 0.00019577827598534885,
"loss": 0.7085,
"step": 330
},
{
"epoch": 0.3726880823251286,
"grad_norm": 0.16873527257647247,
"learning_rate": 0.0001954943046629569,
"loss": 0.717,
"step": 335
},
{
"epoch": 0.37825059101654845,
"grad_norm": 0.1486177299286815,
"learning_rate": 0.00019520131075327298,
"loss": 0.7321,
"step": 340
},
{
"epoch": 0.3838130997079683,
"grad_norm": 0.13978486223349953,
"learning_rate": 0.00019489932193923735,
"loss": 0.7258,
"step": 345
},
{
"epoch": 0.3893756083993881,
"grad_norm": 0.15478013993645653,
"learning_rate": 0.00019458836675365556,
"loss": 0.7147,
"step": 350
},
{
"epoch": 0.39493811709080795,
"grad_norm": 0.1335712601381255,
"learning_rate": 0.00019426847457650292,
"loss": 0.705,
"step": 355
},
{
"epoch": 0.4005006257822278,
"grad_norm": 0.14215220233504217,
"learning_rate": 0.00019393967563214833,
"loss": 0.7109,
"step": 360
},
{
"epoch": 0.4060631344736476,
"grad_norm": 0.12737803269805706,
"learning_rate": 0.00019360200098649864,
"loss": 0.6759,
"step": 365
},
{
"epoch": 0.41162564316506745,
"grad_norm": 0.14212696306551803,
"learning_rate": 0.00019325548254406352,
"loss": 0.6884,
"step": 370
},
{
"epoch": 0.4171881518564873,
"grad_norm": 0.13427514345248226,
"learning_rate": 0.00019290015304494103,
"loss": 0.6997,
"step": 375
},
{
"epoch": 0.4227506605479071,
"grad_norm": 0.13648211126654036,
"learning_rate": 0.00019253604606172417,
"loss": 0.7211,
"step": 380
},
{
"epoch": 0.42831316923932694,
"grad_norm": 0.1328171684670597,
"learning_rate": 0.0001921631959963288,
"loss": 0.6879,
"step": 385
},
{
"epoch": 0.4338756779307468,
"grad_norm": 0.13003415855983963,
"learning_rate": 0.0001917816380767434,
"loss": 0.7164,
"step": 390
},
{
"epoch": 0.4394381866221666,
"grad_norm": 0.12331669293330129,
"learning_rate": 0.00019139140835370053,
"loss": 0.6782,
"step": 395
},
{
"epoch": 0.44500069531358644,
"grad_norm": 0.12354925404653357,
"learning_rate": 0.0001909925436972706,
"loss": 0.7205,
"step": 400
},
{
"epoch": 0.44500069531358644,
"eval_loss": 0.6896927952766418,
"eval_runtime": 38.0771,
"eval_samples_per_second": 10.689,
"eval_steps_per_second": 0.683,
"step": 400
},
{
"epoch": 0.45056320400500627,
"grad_norm": 0.12749730782556348,
"learning_rate": 0.0001905850817933784,
"loss": 0.7147,
"step": 405
},
{
"epoch": 0.4561257126964261,
"grad_norm": 0.1290830353607145,
"learning_rate": 0.0001901690611402423,
"loss": 0.7147,
"step": 410
},
{
"epoch": 0.46168822138784593,
"grad_norm": 0.14656956459115003,
"learning_rate": 0.0001897445210447369,
"loss": 0.7273,
"step": 415
},
{
"epoch": 0.46725073007926576,
"grad_norm": 0.13591979089604955,
"learning_rate": 0.00018931150161867916,
"loss": 0.7233,
"step": 420
},
{
"epoch": 0.4728132387706856,
"grad_norm": 0.14148246469362974,
"learning_rate": 0.00018887004377503858,
"loss": 0.6857,
"step": 425
},
{
"epoch": 0.4783757474621054,
"grad_norm": 0.13172632177206697,
"learning_rate": 0.0001884201892240715,
"loss": 0.6954,
"step": 430
},
{
"epoch": 0.48393825615352526,
"grad_norm": 0.15251049696040847,
"learning_rate": 0.00018796198046938033,
"loss": 0.7369,
"step": 435
},
{
"epoch": 0.4895007648449451,
"grad_norm": 0.14157664597377279,
"learning_rate": 0.00018749546080389757,
"loss": 0.6988,
"step": 440
},
{
"epoch": 0.4950632735363649,
"grad_norm": 0.14790889100169335,
"learning_rate": 0.00018702067430579543,
"loss": 0.7068,
"step": 445
},
{
"epoch": 0.5006257822277848,
"grad_norm": 0.13754915966044653,
"learning_rate": 0.00018653766583432113,
"loss": 0.7226,
"step": 450
},
{
"epoch": 0.5061882909192046,
"grad_norm": 0.1459944140173946,
"learning_rate": 0.00018604648102555856,
"loss": 0.6998,
"step": 455
},
{
"epoch": 0.5117507996106244,
"grad_norm": 0.14444314196036126,
"learning_rate": 0.0001855471662881164,
"loss": 0.7482,
"step": 460
},
{
"epoch": 0.5173133083020443,
"grad_norm": 0.1434929187274412,
"learning_rate": 0.00018503976879874322,
"loss": 0.695,
"step": 465
},
{
"epoch": 0.5228758169934641,
"grad_norm": 0.13128793518588147,
"learning_rate": 0.0001845243364978702,
"loss": 0.6895,
"step": 470
},
{
"epoch": 0.5284383256848839,
"grad_norm": 0.12701476810931256,
"learning_rate": 0.0001840009180850815,
"loss": 0.7061,
"step": 475
},
{
"epoch": 0.5340008343763037,
"grad_norm": 0.1264914632820831,
"learning_rate": 0.00018346956301451304,
"loss": 0.6962,
"step": 480
},
{
"epoch": 0.5395633430677236,
"grad_norm": 0.13208739238621547,
"learning_rate": 0.00018293032149017984,
"loss": 0.6863,
"step": 485
},
{
"epoch": 0.5451258517591434,
"grad_norm": 0.13323142479850197,
"learning_rate": 0.00018238324446123266,
"loss": 0.7039,
"step": 490
},
{
"epoch": 0.5506883604505632,
"grad_norm": 0.1588926427932463,
"learning_rate": 0.0001818283836171441,
"loss": 0.6924,
"step": 495
},
{
"epoch": 0.5562508691419831,
"grad_norm": 0.12667050310382044,
"learning_rate": 0.00018126579138282503,
"loss": 0.6991,
"step": 500
},
{
"epoch": 0.5618133778334029,
"grad_norm": 0.13081757063531854,
"learning_rate": 0.000180695520913671,
"loss": 0.6909,
"step": 505
},
{
"epoch": 0.5673758865248227,
"grad_norm": 0.13598420562667687,
"learning_rate": 0.0001801176260905402,
"loss": 0.7088,
"step": 510
},
{
"epoch": 0.5729383952162426,
"grad_norm": 0.1300574822823029,
"learning_rate": 0.00017953216151466263,
"loss": 0.6821,
"step": 515
},
{
"epoch": 0.5785009039076624,
"grad_norm": 0.14206907328811608,
"learning_rate": 0.00017893918250248104,
"loss": 0.6981,
"step": 520
},
{
"epoch": 0.5840634125990822,
"grad_norm": 0.1557643228084498,
"learning_rate": 0.00017833874508042466,
"loss": 0.7018,
"step": 525
},
{
"epoch": 0.5896259212905021,
"grad_norm": 0.13942706653578768,
"learning_rate": 0.00017773090597961554,
"loss": 0.6738,
"step": 530
},
{
"epoch": 0.5951884299819219,
"grad_norm": 0.12004314100744409,
"learning_rate": 0.00017711572263050845,
"loss": 0.7039,
"step": 535
},
{
"epoch": 0.6007509386733417,
"grad_norm": 0.12451500019905493,
"learning_rate": 0.00017649325315746478,
"loss": 0.7034,
"step": 540
},
{
"epoch": 0.6063134473647616,
"grad_norm": 0.1225444976108732,
"learning_rate": 0.00017586355637326054,
"loss": 0.709,
"step": 545
},
{
"epoch": 0.6118759560561814,
"grad_norm": 0.14795401636462246,
"learning_rate": 0.00017522669177352977,
"loss": 0.7415,
"step": 550
},
{
"epoch": 0.6174384647476012,
"grad_norm": 0.13482563035315856,
"learning_rate": 0.00017458271953114317,
"loss": 0.7117,
"step": 555
},
{
"epoch": 0.6230009734390209,
"grad_norm": 0.15422494211156157,
"learning_rate": 0.0001739317004905227,
"loss": 0.7027,
"step": 560
},
{
"epoch": 0.6285634821304408,
"grad_norm": 0.15479270007272883,
"learning_rate": 0.000173273696161893,
"loss": 0.7174,
"step": 565
},
{
"epoch": 0.6341259908218606,
"grad_norm": 0.13736882345033063,
"learning_rate": 0.00017260876871546936,
"loss": 0.7072,
"step": 570
},
{
"epoch": 0.6396884995132804,
"grad_norm": 0.130113338971742,
"learning_rate": 0.00017193698097558416,
"loss": 0.6987,
"step": 575
},
{
"epoch": 0.6452510082047003,
"grad_norm": 0.15120739345519632,
"learning_rate": 0.00017125839641475072,
"loss": 0.693,
"step": 580
},
{
"epoch": 0.6508135168961201,
"grad_norm": 0.17345133663648482,
"learning_rate": 0.00017057307914766624,
"loss": 0.6933,
"step": 585
},
{
"epoch": 0.6563760255875399,
"grad_norm": 0.14577660905768358,
"learning_rate": 0.0001698810939251543,
"loss": 0.6847,
"step": 590
},
{
"epoch": 0.6619385342789598,
"grad_norm": 0.13729541550408036,
"learning_rate": 0.00016918250612804673,
"loss": 0.7058,
"step": 595
},
{
"epoch": 0.6675010429703796,
"grad_norm": 0.12792609813747083,
"learning_rate": 0.00016847738176100632,
"loss": 0.7203,
"step": 600
},
{
"epoch": 0.6675010429703796,
"eval_loss": 0.6808404326438904,
"eval_runtime": 38.0896,
"eval_samples_per_second": 10.685,
"eval_steps_per_second": 0.683,
"step": 600
},
{
"epoch": 0.6730635516617994,
"grad_norm": 0.12614147228685735,
"learning_rate": 0.00016776578744629052,
"loss": 0.6984,
"step": 605
},
{
"epoch": 0.6786260603532193,
"grad_norm": 0.14053504602609868,
"learning_rate": 0.00016704779041745686,
"loss": 0.7017,
"step": 610
},
{
"epoch": 0.6841885690446391,
"grad_norm": 0.144524525939437,
"learning_rate": 0.00016632345851301031,
"loss": 0.7031,
"step": 615
},
{
"epoch": 0.6897510777360589,
"grad_norm": 0.1240886307637927,
"learning_rate": 0.000165592860169994,
"loss": 0.6856,
"step": 620
},
{
"epoch": 0.6953135864274788,
"grad_norm": 0.12592281039225006,
"learning_rate": 0.0001648560644175227,
"loss": 0.6832,
"step": 625
},
{
"epoch": 0.7008760951188986,
"grad_norm": 0.1312108925603738,
"learning_rate": 0.00016411314087026106,
"loss": 0.7266,
"step": 630
},
{
"epoch": 0.7064386038103184,
"grad_norm": 0.12871371439878188,
"learning_rate": 0.00016336415972184612,
"loss": 0.6892,
"step": 635
},
{
"epoch": 0.7120011125017383,
"grad_norm": 0.1423393128313328,
"learning_rate": 0.00016260919173825508,
"loss": 0.684,
"step": 640
},
{
"epoch": 0.7175636211931581,
"grad_norm": 0.1452224910144209,
"learning_rate": 0.00016184830825111924,
"loss": 0.6824,
"step": 645
},
{
"epoch": 0.7231261298845779,
"grad_norm": 0.12119255762333672,
"learning_rate": 0.00016108158115098444,
"loss": 0.686,
"step": 650
},
{
"epoch": 0.7286886385759977,
"grad_norm": 0.1411776440916737,
"learning_rate": 0.0001603090828805185,
"loss": 0.6878,
"step": 655
},
{
"epoch": 0.7342511472674176,
"grad_norm": 0.17300629248103022,
"learning_rate": 0.0001595308864276666,
"loss": 0.7139,
"step": 660
},
{
"epoch": 0.7398136559588374,
"grad_norm": 0.1458091025518703,
"learning_rate": 0.0001587470653187553,
"loss": 0.729,
"step": 665
},
{
"epoch": 0.7453761646502572,
"grad_norm": 0.14056535678032148,
"learning_rate": 0.00015795769361154547,
"loss": 0.6972,
"step": 670
},
{
"epoch": 0.7509386733416771,
"grad_norm": 0.13734797531202647,
"learning_rate": 0.000157162845888235,
"loss": 0.6779,
"step": 675
},
{
"epoch": 0.7565011820330969,
"grad_norm": 0.12703010084902983,
"learning_rate": 0.00015636259724841222,
"loss": 0.7027,
"step": 680
},
{
"epoch": 0.7620636907245167,
"grad_norm": 0.12697872232577906,
"learning_rate": 0.00015555702330196023,
"loss": 0.7079,
"step": 685
},
{
"epoch": 0.7676261994159366,
"grad_norm": 0.15015643114183339,
"learning_rate": 0.00015474620016191294,
"loss": 0.6938,
"step": 690
},
{
"epoch": 0.7731887081073564,
"grad_norm": 0.1468310767380953,
"learning_rate": 0.00015393020443726381,
"loss": 0.7061,
"step": 695
},
{
"epoch": 0.7787512167987762,
"grad_norm": 0.1178258952392868,
"learning_rate": 0.00015310911322572753,
"loss": 0.6635,
"step": 700
},
{
"epoch": 0.7843137254901961,
"grad_norm": 0.1182521343472938,
"learning_rate": 0.00015228300410645556,
"loss": 0.6783,
"step": 705
},
{
"epoch": 0.7898762341816159,
"grad_norm": 0.11865217322316654,
"learning_rate": 0.00015145195513270644,
"loss": 0.6822,
"step": 710
},
{
"epoch": 0.7954387428730357,
"grad_norm": 0.11500353884529863,
"learning_rate": 0.00015061604482447075,
"loss": 0.6668,
"step": 715
},
{
"epoch": 0.8010012515644556,
"grad_norm": 0.1314805700111582,
"learning_rate": 0.0001497753521610526,
"loss": 0.7105,
"step": 720
},
{
"epoch": 0.8065637602558754,
"grad_norm": 0.14124896793550357,
"learning_rate": 0.00014892995657360717,
"loss": 0.7142,
"step": 725
},
{
"epoch": 0.8121262689472952,
"grad_norm": 0.13739934885223082,
"learning_rate": 0.00014807993793763619,
"loss": 0.6815,
"step": 730
},
{
"epoch": 0.8176887776387151,
"grad_norm": 0.13281564930523101,
"learning_rate": 0.0001472253765654406,
"loss": 0.67,
"step": 735
},
{
"epoch": 0.8232512863301349,
"grad_norm": 0.11821804301835571,
"learning_rate": 0.00014636635319853275,
"loss": 0.6969,
"step": 740
},
{
"epoch": 0.8288137950215547,
"grad_norm": 0.13929966861379298,
"learning_rate": 0.00014550294900000753,
"loss": 0.696,
"step": 745
},
{
"epoch": 0.8343763037129746,
"grad_norm": 0.1358532247107506,
"learning_rate": 0.00014463524554687399,
"loss": 0.6987,
"step": 750
},
{
"epoch": 0.8399388124043944,
"grad_norm": 0.14607145415562992,
"learning_rate": 0.00014376332482234747,
"loss": 0.6713,
"step": 755
},
{
"epoch": 0.8455013210958142,
"grad_norm": 0.13045660818329083,
"learning_rate": 0.0001428872692081038,
"loss": 0.6782,
"step": 760
},
{
"epoch": 0.851063829787234,
"grad_norm": 0.14868265540995262,
"learning_rate": 0.00014200716147649557,
"loss": 0.6856,
"step": 765
},
{
"epoch": 0.8566263384786539,
"grad_norm": 0.1523898280318719,
"learning_rate": 0.00014112308478273145,
"loss": 0.7077,
"step": 770
},
{
"epoch": 0.8621888471700737,
"grad_norm": 0.14374188108756408,
"learning_rate": 0.00014023512265701955,
"loss": 0.6557,
"step": 775
},
{
"epoch": 0.8677513558614935,
"grad_norm": 0.13285163140650738,
"learning_rate": 0.00013934335899667527,
"loss": 0.7028,
"step": 780
},
{
"epoch": 0.8733138645529134,
"grad_norm": 0.12419599179555058,
"learning_rate": 0.0001384478780581942,
"loss": 0.6753,
"step": 785
},
{
"epoch": 0.8788763732443332,
"grad_norm": 0.12619683911200477,
"learning_rate": 0.00013754876444929166,
"loss": 0.7033,
"step": 790
},
{
"epoch": 0.884438881935753,
"grad_norm": 0.14380964532401144,
"learning_rate": 0.00013664610312090838,
"loss": 0.7022,
"step": 795
},
{
"epoch": 0.8900013906271729,
"grad_norm": 0.13341619640026267,
"learning_rate": 0.0001357399793591844,
"loss": 0.703,
"step": 800
},
{
"epoch": 0.8900013906271729,
"eval_loss": 0.6755693554878235,
"eval_runtime": 38.0742,
"eval_samples_per_second": 10.69,
"eval_steps_per_second": 0.683,
"step": 800
},
{
"epoch": 0.8955638993185927,
"grad_norm": 0.11373621352884696,
"learning_rate": 0.00013483047877740055,
"loss": 0.6554,
"step": 805
},
{
"epoch": 0.9011264080100125,
"grad_norm": 0.1418177008323639,
"learning_rate": 0.00013391768730789002,
"loss": 0.7435,
"step": 810
},
{
"epoch": 0.9066889167014324,
"grad_norm": 0.14095795493706695,
"learning_rate": 0.00013300169119391864,
"loss": 0.6856,
"step": 815
},
{
"epoch": 0.9122514253928522,
"grad_norm": 0.13583867929382498,
"learning_rate": 0.00013208257698153677,
"loss": 0.6791,
"step": 820
},
{
"epoch": 0.917813934084272,
"grad_norm": 0.13643641315332133,
"learning_rate": 0.00013116043151140203,
"loss": 0.6958,
"step": 825
},
{
"epoch": 0.9233764427756919,
"grad_norm": 0.12156612730761757,
"learning_rate": 0.00013023534191057426,
"loss": 0.6912,
"step": 830
},
{
"epoch": 0.9289389514671117,
"grad_norm": 0.14020483760043728,
"learning_rate": 0.0001293073955842836,
"loss": 0.7086,
"step": 835
},
{
"epoch": 0.9345014601585315,
"grad_norm": 0.13715121587687112,
"learning_rate": 0.0001283766802076722,
"loss": 0.6804,
"step": 840
},
{
"epoch": 0.9400639688499514,
"grad_norm": 0.12395531650931035,
"learning_rate": 0.00012744328371751024,
"loss": 0.7246,
"step": 845
},
{
"epoch": 0.9456264775413712,
"grad_norm": 0.1351012642378018,
"learning_rate": 0.00012650729430388764,
"loss": 0.6695,
"step": 850
},
{
"epoch": 0.951188986232791,
"grad_norm": 0.14258700665297835,
"learning_rate": 0.00012556880040188144,
"loss": 0.6844,
"step": 855
},
{
"epoch": 0.9567514949242109,
"grad_norm": 0.14370950835852644,
"learning_rate": 0.00012462789068320017,
"loss": 0.7156,
"step": 860
},
{
"epoch": 0.9623140036156307,
"grad_norm": 0.1390057701240165,
"learning_rate": 0.00012368465404780598,
"loss": 0.6652,
"step": 865
},
{
"epoch": 0.9678765123070505,
"grad_norm": 0.127046163008197,
"learning_rate": 0.00012273917961551513,
"loss": 0.6641,
"step": 870
},
{
"epoch": 0.9734390209984704,
"grad_norm": 0.14753709766070622,
"learning_rate": 0.00012179155671757754,
"loss": 0.689,
"step": 875
},
{
"epoch": 0.9790015296898902,
"grad_norm": 0.13366681839894046,
"learning_rate": 0.00012084187488823657,
"loss": 0.6938,
"step": 880
},
{
"epoch": 0.98456403838131,
"grad_norm": 0.14364888110758262,
"learning_rate": 0.00011989022385626967,
"loss": 0.6867,
"step": 885
},
{
"epoch": 0.9901265470727298,
"grad_norm": 0.13750035239201705,
"learning_rate": 0.00011893669353651031,
"loss": 0.6926,
"step": 890
},
{
"epoch": 0.9956890557641497,
"grad_norm": 0.12165819999979594,
"learning_rate": 0.00011798137402135277,
"loss": 0.6781,
"step": 895
},
{
"epoch": 1.0012515644555695,
"grad_norm": 0.13460664662956676,
"learning_rate": 0.00011702435557223987,
"loss": 0.6907,
"step": 900
},
{
"epoch": 1.0068140731469892,
"grad_norm": 0.14812357106021773,
"learning_rate": 0.00011606572861113474,
"loss": 0.6739,
"step": 905
},
{
"epoch": 1.0123765818384092,
"grad_norm": 0.14414234656282895,
"learning_rate": 0.00011510558371197753,
"loss": 0.6581,
"step": 910
},
{
"epoch": 1.017939090529829,
"grad_norm": 0.1417200021185479,
"learning_rate": 0.00011414401159212778,
"loss": 0.6886,
"step": 915
},
{
"epoch": 1.0235015992212488,
"grad_norm": 0.13876983385482528,
"learning_rate": 0.00011318110310379301,
"loss": 0.6737,
"step": 920
},
{
"epoch": 1.0290641079126686,
"grad_norm": 0.15251072221895987,
"learning_rate": 0.00011221694922544491,
"loss": 0.6692,
"step": 925
},
{
"epoch": 1.0346266166040885,
"grad_norm": 0.14026789842910758,
"learning_rate": 0.0001112516410532233,
"loss": 0.6562,
"step": 930
},
{
"epoch": 1.0401891252955082,
"grad_norm": 0.12707738202474425,
"learning_rate": 0.00011028526979232913,
"loss": 0.6618,
"step": 935
},
{
"epoch": 1.0457516339869282,
"grad_norm": 0.14606846473361,
"learning_rate": 0.00010931792674840718,
"loss": 0.6746,
"step": 940
},
{
"epoch": 1.0513141426783479,
"grad_norm": 0.14011972796939615,
"learning_rate": 0.00010834970331891914,
"loss": 0.6869,
"step": 945
},
{
"epoch": 1.0568766513697678,
"grad_norm": 0.1563818365447484,
"learning_rate": 0.0001073806909845082,
"loss": 0.66,
"step": 950
},
{
"epoch": 1.0624391600611875,
"grad_norm": 0.12113541850928697,
"learning_rate": 0.00010641098130035562,
"loss": 0.6567,
"step": 955
},
{
"epoch": 1.0680016687526075,
"grad_norm": 0.12784803662349467,
"learning_rate": 0.00010544066588753044,
"loss": 0.6478,
"step": 960
},
{
"epoch": 1.0735641774440272,
"grad_norm": 0.13555213849737605,
"learning_rate": 0.00010446983642433259,
"loss": 0.6423,
"step": 965
},
{
"epoch": 1.0791266861354472,
"grad_norm": 0.14956837158318062,
"learning_rate": 0.00010349858463763113,
"loss": 0.6482,
"step": 970
},
{
"epoch": 1.0846891948268669,
"grad_norm": 0.14302685723153047,
"learning_rate": 0.0001025270022941975,
"loss": 0.6619,
"step": 975
},
{
"epoch": 1.0902517035182868,
"grad_norm": 0.1531273940282806,
"learning_rate": 0.0001015551811920351,
"loss": 0.6818,
"step": 980
},
{
"epoch": 1.0958142122097065,
"grad_norm": 0.1431917561070241,
"learning_rate": 0.00010058321315170596,
"loss": 0.6597,
"step": 985
},
{
"epoch": 1.1013767209011265,
"grad_norm": 0.14533278983667672,
"learning_rate": 9.961119000765531e-05,
"loss": 0.641,
"step": 990
},
{
"epoch": 1.1069392295925462,
"grad_norm": 0.1426693103942941,
"learning_rate": 9.863920359953474e-05,
"loss": 0.6864,
"step": 995
},
{
"epoch": 1.1125017382839661,
"grad_norm": 0.14515333605223138,
"learning_rate": 9.766734576352478e-05,
"loss": 0.6759,
"step": 1000
},
{
"epoch": 1.1125017382839661,
"eval_loss": 0.6748408079147339,
"eval_runtime": 38.0571,
"eval_samples_per_second": 10.694,
"eval_steps_per_second": 0.683,
"step": 1000
},
{
"epoch": 1.1180642469753859,
"grad_norm": 0.13884691864173954,
"learning_rate": 9.669570832365838e-05,
"loss": 0.6536,
"step": 1005
},
{
"epoch": 1.1236267556668058,
"grad_norm": 0.14190147387793983,
"learning_rate": 9.572438308314446e-05,
"loss": 0.6448,
"step": 1010
},
{
"epoch": 1.1291892643582255,
"grad_norm": 0.1460356352229923,
"learning_rate": 9.475346181569467e-05,
"loss": 0.6709,
"step": 1015
},
{
"epoch": 1.1347517730496455,
"grad_norm": 0.15004591423623692,
"learning_rate": 9.378303625685195e-05,
"loss": 0.6338,
"step": 1020
},
{
"epoch": 1.1403142817410652,
"grad_norm": 0.14709044334780488,
"learning_rate": 9.281319809532329e-05,
"loss": 0.6482,
"step": 1025
},
{
"epoch": 1.1458767904324851,
"grad_norm": 0.1339983467335465,
"learning_rate": 9.18440389643165e-05,
"loss": 0.6564,
"step": 1030
},
{
"epoch": 1.1514392991239049,
"grad_norm": 0.15465369318467384,
"learning_rate": 9.08756504328827e-05,
"loss": 0.631,
"step": 1035
},
{
"epoch": 1.1570018078153248,
"grad_norm": 0.14414339429238215,
"learning_rate": 8.990812399726435e-05,
"loss": 0.63,
"step": 1040
},
{
"epoch": 1.1625643165067445,
"grad_norm": 0.1530406190835784,
"learning_rate": 8.894155107225062e-05,
"loss": 0.6599,
"step": 1045
},
{
"epoch": 1.1681268251981645,
"grad_norm": 0.18004713578221632,
"learning_rate": 8.797602298254004e-05,
"loss": 0.6588,
"step": 1050
},
{
"epoch": 1.1736893338895842,
"grad_norm": 0.1395761709805775,
"learning_rate": 8.701163095411212e-05,
"loss": 0.6388,
"step": 1055
},
{
"epoch": 1.1792518425810041,
"grad_norm": 0.14800345539370582,
"learning_rate": 8.604846610560771e-05,
"loss": 0.6469,
"step": 1060
},
{
"epoch": 1.1848143512724238,
"grad_norm": 0.18153262554102811,
"learning_rate": 8.508661943972021e-05,
"loss": 0.6476,
"step": 1065
},
{
"epoch": 1.1903768599638438,
"grad_norm": 0.1460085066143512,
"learning_rate": 8.412618183459708e-05,
"loss": 0.6417,
"step": 1070
},
{
"epoch": 1.1959393686552635,
"grad_norm": 0.1563928211034322,
"learning_rate": 8.316724403525359e-05,
"loss": 0.6655,
"step": 1075
},
{
"epoch": 1.2015018773466832,
"grad_norm": 0.13021633068783775,
"learning_rate": 8.220989664499878e-05,
"loss": 0.6495,
"step": 1080
},
{
"epoch": 1.2070643860381032,
"grad_norm": 0.15359952593058618,
"learning_rate": 8.125423011687524e-05,
"loss": 0.6518,
"step": 1085
},
{
"epoch": 1.2126268947295231,
"grad_norm": 0.14327581045453255,
"learning_rate": 8.030033474511249e-05,
"loss": 0.6428,
"step": 1090
},
{
"epoch": 1.2181894034209428,
"grad_norm": 0.14877277012088305,
"learning_rate": 7.934830065659599e-05,
"loss": 0.6823,
"step": 1095
},
{
"epoch": 1.2237519121123626,
"grad_norm": 0.1350097173259691,
"learning_rate": 7.839821780235168e-05,
"loss": 0.6656,
"step": 1100
},
{
"epoch": 1.2293144208037825,
"grad_norm": 0.14805261697578703,
"learning_rate": 7.74501759490469e-05,
"loss": 0.6649,
"step": 1105
},
{
"epoch": 1.2348769294952024,
"grad_norm": 0.14241031431419957,
"learning_rate": 7.650426467050926e-05,
"loss": 0.645,
"step": 1110
},
{
"epoch": 1.2404394381866222,
"grad_norm": 0.14677979540609026,
"learning_rate": 7.556057333926318e-05,
"loss": 0.6322,
"step": 1115
},
{
"epoch": 1.2460019468780419,
"grad_norm": 0.1540439921077116,
"learning_rate": 7.461919111808595e-05,
"loss": 0.6425,
"step": 1120
},
{
"epoch": 1.2515644555694618,
"grad_norm": 0.16093917993335288,
"learning_rate": 7.368020695158312e-05,
"loss": 0.6708,
"step": 1125
},
{
"epoch": 1.2571269642608818,
"grad_norm": 0.14072702856543545,
"learning_rate": 7.274370955778498e-05,
"loss": 0.6412,
"step": 1130
},
{
"epoch": 1.2626894729523015,
"grad_norm": 0.1372982963866414,
"learning_rate": 7.180978741976397e-05,
"loss": 0.6354,
"step": 1135
},
{
"epoch": 1.2682519816437212,
"grad_norm": 0.1361010979443787,
"learning_rate": 7.087852877727481e-05,
"loss": 0.6297,
"step": 1140
},
{
"epoch": 1.2738144903351412,
"grad_norm": 0.15152880519317802,
"learning_rate": 6.995002161841708e-05,
"loss": 0.6648,
"step": 1145
},
{
"epoch": 1.279376999026561,
"grad_norm": 0.15111271273299604,
"learning_rate": 6.902435367132208e-05,
"loss": 0.6811,
"step": 1150
},
{
"epoch": 1.2849395077179808,
"grad_norm": 0.1555612197330931,
"learning_rate": 6.810161239586375e-05,
"loss": 0.6639,
"step": 1155
},
{
"epoch": 1.2905020164094005,
"grad_norm": 0.1404429960918057,
"learning_rate": 6.718188497539554e-05,
"loss": 0.6416,
"step": 1160
},
{
"epoch": 1.2960645251008205,
"grad_norm": 0.1493351620856526,
"learning_rate": 6.626525830851267e-05,
"loss": 0.6284,
"step": 1165
},
{
"epoch": 1.3016270337922404,
"grad_norm": 0.15944635272625493,
"learning_rate": 6.535181900084206e-05,
"loss": 0.6402,
"step": 1170
},
{
"epoch": 1.3071895424836601,
"grad_norm": 0.1534235543873645,
"learning_rate": 6.444165335685927e-05,
"loss": 0.6373,
"step": 1175
},
{
"epoch": 1.3127520511750799,
"grad_norm": 0.1404290668723592,
"learning_rate": 6.35348473717345e-05,
"loss": 0.6596,
"step": 1180
},
{
"epoch": 1.3183145598664998,
"grad_norm": 0.15190370802350445,
"learning_rate": 6.263148672320714e-05,
"loss": 0.6519,
"step": 1185
},
{
"epoch": 1.3238770685579198,
"grad_norm": 0.1396377515861422,
"learning_rate": 6.173165676349103e-05,
"loss": 0.6539,
"step": 1190
},
{
"epoch": 1.3294395772493395,
"grad_norm": 0.1413554398489568,
"learning_rate": 6.083544251120993e-05,
"loss": 0.6436,
"step": 1195
},
{
"epoch": 1.3350020859407592,
"grad_norm": 0.175529243963575,
"learning_rate": 5.9942928643364724e-05,
"loss": 0.6533,
"step": 1200
},
{
"epoch": 1.3350020859407592,
"eval_loss": 0.6695060133934021,
"eval_runtime": 38.0647,
"eval_samples_per_second": 10.692,
"eval_steps_per_second": 0.683,
"step": 1200
},
{
"epoch": 1.3405645946321791,
"grad_norm": 0.16769090072157003,
"learning_rate": 5.905419948733302e-05,
"loss": 0.6434,
"step": 1205
},
{
"epoch": 1.3461271033235989,
"grad_norm": 0.17558035684023568,
"learning_rate": 5.816933901290136e-05,
"loss": 0.6954,
"step": 1210
},
{
"epoch": 1.3516896120150188,
"grad_norm": 0.15028360304997448,
"learning_rate": 5.728843082433193e-05,
"loss": 0.6563,
"step": 1215
},
{
"epoch": 1.3572521207064385,
"grad_norm": 0.16888366454421866,
"learning_rate": 5.6411558152462894e-05,
"loss": 0.6383,
"step": 1220
},
{
"epoch": 1.3628146293978585,
"grad_norm": 0.1507413542567497,
"learning_rate": 5.553880384684493e-05,
"loss": 0.6382,
"step": 1225
},
{
"epoch": 1.3683771380892782,
"grad_norm": 0.15944821317229088,
"learning_rate": 5.4670250367913023e-05,
"loss": 0.6589,
"step": 1230
},
{
"epoch": 1.3739396467806981,
"grad_norm": 0.14215107373647873,
"learning_rate": 5.380597977919557e-05,
"loss": 0.6346,
"step": 1235
},
{
"epoch": 1.3795021554721179,
"grad_norm": 0.16243215498334238,
"learning_rate": 5.2946073739560706e-05,
"loss": 0.6527,
"step": 1240
},
{
"epoch": 1.3850646641635378,
"grad_norm": 0.1562039482784561,
"learning_rate": 5.209061349550095e-05,
"loss": 0.66,
"step": 1245
},
{
"epoch": 1.3906271728549575,
"grad_norm": 0.1680769018618033,
"learning_rate": 5.1239679873456634e-05,
"loss": 0.6699,
"step": 1250
},
{
"epoch": 1.3961896815463775,
"grad_norm": 0.1478360199044516,
"learning_rate": 5.039335327217951e-05,
"loss": 0.6502,
"step": 1255
},
{
"epoch": 1.4017521902377972,
"grad_norm": 0.14915391173643802,
"learning_rate": 4.955171365513603e-05,
"loss": 0.6143,
"step": 1260
},
{
"epoch": 1.4073146989292171,
"grad_norm": 0.15851697773820939,
"learning_rate": 4.871484054295258e-05,
"loss": 0.6572,
"step": 1265
},
{
"epoch": 1.4128772076206368,
"grad_norm": 0.15216697385980452,
"learning_rate": 4.7882813005901696e-05,
"loss": 0.657,
"step": 1270
},
{
"epoch": 1.4184397163120568,
"grad_norm": 0.1437749041937587,
"learning_rate": 4.705570965643176e-05,
"loss": 0.6411,
"step": 1275
},
{
"epoch": 1.4240022250034765,
"grad_norm": 0.16683978696735308,
"learning_rate": 4.623360864173893e-05,
"loss": 0.651,
"step": 1280
},
{
"epoch": 1.4295647336948965,
"grad_norm": 0.1621572781654395,
"learning_rate": 4.541658763638406e-05,
"loss": 0.6492,
"step": 1285
},
{
"epoch": 1.4351272423863162,
"grad_norm": 0.15994302691177126,
"learning_rate": 4.460472383495331e-05,
"loss": 0.6577,
"step": 1290
},
{
"epoch": 1.4406897510777361,
"grad_norm": 0.14692563499946057,
"learning_rate": 4.379809394476501e-05,
"loss": 0.6496,
"step": 1295
},
{
"epoch": 1.4462522597691558,
"grad_norm": 0.17176689432434047,
"learning_rate": 4.2996774178621736e-05,
"loss": 0.6433,
"step": 1300
},
{
"epoch": 1.4518147684605758,
"grad_norm": 0.1446292926663316,
"learning_rate": 4.220084024760982e-05,
"loss": 0.6223,
"step": 1305
},
{
"epoch": 1.4573772771519955,
"grad_norm": 0.15172393816056104,
"learning_rate": 4.141036735394574e-05,
"loss": 0.6481,
"step": 1310
},
{
"epoch": 1.4629397858434154,
"grad_norm": 0.16544844619207105,
"learning_rate": 4.0625430183870796e-05,
"loss": 0.6411,
"step": 1315
},
{
"epoch": 1.4685022945348352,
"grad_norm": 0.15934504430033075,
"learning_rate": 3.984610290059467e-05,
"loss": 0.6678,
"step": 1320
},
{
"epoch": 1.474064803226255,
"grad_norm": 0.15204965773692086,
"learning_rate": 3.907245913728807e-05,
"loss": 0.6549,
"step": 1325
},
{
"epoch": 1.4796273119176748,
"grad_norm": 0.15256724381583675,
"learning_rate": 3.830457199012585e-05,
"loss": 0.6478,
"step": 1330
},
{
"epoch": 1.4851898206090948,
"grad_norm": 0.15150975920035156,
"learning_rate": 3.754251401138051e-05,
"loss": 0.6385,
"step": 1335
},
{
"epoch": 1.4907523293005145,
"grad_norm": 0.17068279689989202,
"learning_rate": 3.678635720256737e-05,
"loss": 0.6659,
"step": 1340
},
{
"epoch": 1.4963148379919344,
"grad_norm": 0.15001304767115145,
"learning_rate": 3.6036173007641435e-05,
"loss": 0.66,
"step": 1345
},
{
"epoch": 1.5018773466833542,
"grad_norm": 0.14424590095876771,
"learning_rate": 3.529203230624747e-05,
"loss": 0.651,
"step": 1350
},
{
"epoch": 1.5074398553747739,
"grad_norm": 0.16690307331407334,
"learning_rate": 3.455400540702274e-05,
"loss": 0.6508,
"step": 1355
},
{
"epoch": 1.5130023640661938,
"grad_norm": 0.1506720689857848,
"learning_rate": 3.3822162040954354e-05,
"loss": 0.6347,
"step": 1360
},
{
"epoch": 1.5185648727576138,
"grad_norm": 0.15108676554413317,
"learning_rate": 3.309657135479065e-05,
"loss": 0.6449,
"step": 1365
},
{
"epoch": 1.5241273814490335,
"grad_norm": 0.1397876543411103,
"learning_rate": 3.237730190450816e-05,
"loss": 0.6382,
"step": 1370
},
{
"epoch": 1.5296898901404532,
"grad_norm": 0.15000237273739636,
"learning_rate": 3.166442164883403e-05,
"loss": 0.6607,
"step": 1375
},
{
"epoch": 1.5352523988318731,
"grad_norm": 0.15901665795803346,
"learning_rate": 3.0957997942825336e-05,
"loss": 0.6403,
"step": 1380
},
{
"epoch": 1.540814907523293,
"grad_norm": 0.1500239251091428,
"learning_rate": 3.0258097531504937e-05,
"loss": 0.6595,
"step": 1385
},
{
"epoch": 1.5463774162147128,
"grad_norm": 0.16536122507336246,
"learning_rate": 2.9564786543555388e-05,
"loss": 0.6515,
"step": 1390
},
{
"epoch": 1.5519399249061325,
"grad_norm": 0.14948767796744467,
"learning_rate": 2.8878130485070852e-05,
"loss": 0.6341,
"step": 1395
},
{
"epoch": 1.5575024335975525,
"grad_norm": 0.16630242366164855,
"learning_rate": 2.819819423336775e-05,
"loss": 0.6458,
"step": 1400
},
{
"epoch": 1.5575024335975525,
"eval_loss": 0.6669326424598694,
"eval_runtime": 38.0635,
"eval_samples_per_second": 10.693,
"eval_steps_per_second": 0.683,
"step": 1400
},
{
"epoch": 1.5630649422889724,
"grad_norm": 0.15092670649471096,
"learning_rate": 2.7525042030855218e-05,
"loss": 0.6285,
"step": 1405
},
{
"epoch": 1.5686274509803921,
"grad_norm": 0.17817145439236162,
"learning_rate": 2.6858737478965035e-05,
"loss": 0.6786,
"step": 1410
},
{
"epoch": 1.5741899596718119,
"grad_norm": 0.14411321645345596,
"learning_rate": 2.6199343532142573e-05,
"loss": 0.6487,
"step": 1415
},
{
"epoch": 1.5797524683632318,
"grad_norm": 0.1729923377951436,
"learning_rate": 2.5546922491898495e-05,
"loss": 0.6466,
"step": 1420
},
{
"epoch": 1.5853149770546517,
"grad_norm": 0.17040206170029934,
"learning_rate": 2.4901536000922497e-05,
"loss": 0.665,
"step": 1425
},
{
"epoch": 1.5908774857460715,
"grad_norm": 0.16055497144569197,
"learning_rate": 2.4263245037258995e-05,
"loss": 0.6665,
"step": 1430
},
{
"epoch": 1.5964399944374912,
"grad_norm": 0.1655874585771864,
"learning_rate": 2.363210990854582e-05,
"loss": 0.6734,
"step": 1435
},
{
"epoch": 1.6020025031289111,
"grad_norm": 0.1651519002975595,
"learning_rate": 2.300819024631603e-05,
"loss": 0.6691,
"step": 1440
},
{
"epoch": 1.607565011820331,
"grad_norm": 0.162990807682259,
"learning_rate": 2.239154500036399e-05,
"loss": 0.6805,
"step": 1445
},
{
"epoch": 1.6131275205117508,
"grad_norm": 0.1717600360257139,
"learning_rate": 2.178223243317532e-05,
"loss": 0.6399,
"step": 1450
},
{
"epoch": 1.6186900292031705,
"grad_norm": 0.1739831864715599,
"learning_rate": 2.1180310114422362e-05,
"loss": 0.6689,
"step": 1455
},
{
"epoch": 1.6242525378945905,
"grad_norm": 0.15498282638359004,
"learning_rate": 2.058583491552465e-05,
"loss": 0.6675,
"step": 1460
},
{
"epoch": 1.6298150465860104,
"grad_norm": 0.17242195548428726,
"learning_rate": 1.9998863004275593e-05,
"loss": 0.6618,
"step": 1465
},
{
"epoch": 1.6353775552774301,
"grad_norm": 0.16154448145412242,
"learning_rate": 1.941944983953552e-05,
"loss": 0.6406,
"step": 1470
},
{
"epoch": 1.6409400639688498,
"grad_norm": 0.1624668193562119,
"learning_rate": 1.884765016599186e-05,
"loss": 0.6223,
"step": 1475
},
{
"epoch": 1.6465025726602698,
"grad_norm": 0.16026073767873994,
"learning_rate": 1.8283518008986567e-05,
"loss": 0.6424,
"step": 1480
},
{
"epoch": 1.6520650813516897,
"grad_norm": 0.14391017994489602,
"learning_rate": 1.7727106669411776e-05,
"loss": 0.6323,
"step": 1485
},
{
"epoch": 1.6576275900431094,
"grad_norm": 0.16566629921954873,
"learning_rate": 1.7178468718673714e-05,
"loss": 0.6605,
"step": 1490
},
{
"epoch": 1.6631900987345292,
"grad_norm": 0.15202811145565842,
"learning_rate": 1.6637655993725598e-05,
"loss": 0.6745,
"step": 1495
},
{
"epoch": 1.668752607425949,
"grad_norm": 0.14529589529297973,
"learning_rate": 1.6104719592169902e-05,
"loss": 0.6415,
"step": 1500
},
{
"epoch": 1.674315116117369,
"grad_norm": 0.16206187759632693,
"learning_rate": 1.5579709867430514e-05,
"loss": 0.6757,
"step": 1505
},
{
"epoch": 1.6798776248087888,
"grad_norm": 0.1689334880075168,
"learning_rate": 1.5062676423995247e-05,
"loss": 0.6648,
"step": 1510
},
{
"epoch": 1.6854401335002085,
"grad_norm": 0.17002942586162523,
"learning_rate": 1.4553668112729025e-05,
"loss": 0.6593,
"step": 1515
},
{
"epoch": 1.6910026421916284,
"grad_norm": 0.15390481859714705,
"learning_rate": 1.4052733026258281e-05,
"loss": 0.6337,
"step": 1520
},
{
"epoch": 1.6965651508830484,
"grad_norm": 0.16322817116461236,
"learning_rate": 1.3559918494427015e-05,
"loss": 0.6671,
"step": 1525
},
{
"epoch": 1.702127659574468,
"grad_norm": 0.17381037300710486,
"learning_rate": 1.3075271079825036e-05,
"loss": 0.6452,
"step": 1530
},
{
"epoch": 1.7076901682658878,
"grad_norm": 0.14385216586204455,
"learning_rate": 1.2598836573388383e-05,
"loss": 0.6517,
"step": 1535
},
{
"epoch": 1.7132526769573078,
"grad_norm": 0.13695810287509047,
"learning_rate": 1.2130659990073146e-05,
"loss": 0.6115,
"step": 1540
},
{
"epoch": 1.7188151856487277,
"grad_norm": 0.15619086164642515,
"learning_rate": 1.1670785564601972e-05,
"loss": 0.6592,
"step": 1545
},
{
"epoch": 1.7243776943401474,
"grad_norm": 0.1359615918791439,
"learning_rate": 1.1219256747285045e-05,
"loss": 0.6489,
"step": 1550
},
{
"epoch": 1.7299402030315671,
"grad_norm": 0.1530561734129938,
"learning_rate": 1.0776116199914343e-05,
"loss": 0.6341,
"step": 1555
},
{
"epoch": 1.735502711722987,
"grad_norm": 0.1550580447696997,
"learning_rate": 1.0341405791733183e-05,
"loss": 0.679,
"step": 1560
},
{
"epoch": 1.741065220414407,
"grad_norm": 0.15343391714899443,
"learning_rate": 9.915166595480018e-06,
"loss": 0.6174,
"step": 1565
},
{
"epoch": 1.7466277291058268,
"grad_norm": 0.14886169752976763,
"learning_rate": 9.49743888350798e-06,
"loss": 0.658,
"step": 1570
},
{
"epoch": 1.7521902377972465,
"grad_norm": 0.16009066506717334,
"learning_rate": 9.088262123979652e-06,
"loss": 0.6547,
"step": 1575
},
{
"epoch": 1.7577527464886664,
"grad_norm": 0.1527884120085583,
"learning_rate": 8.687674977138116e-06,
"loss": 0.6355,
"step": 1580
},
{
"epoch": 1.7633152551800864,
"grad_norm": 0.15622894390259232,
"learning_rate": 8.295715291654205e-06,
"loss": 0.6415,
"step": 1585
},
{
"epoch": 1.768877763871506,
"grad_norm": 0.13821573876875842,
"learning_rate": 7.912420101050367e-06,
"loss": 0.6386,
"step": 1590
},
{
"epoch": 1.7744402725629258,
"grad_norm": 0.1605607168007031,
"learning_rate": 7.537825620201699e-06,
"loss": 0.6395,
"step": 1595
},
{
"epoch": 1.7800027812543457,
"grad_norm": 0.14785735762140897,
"learning_rate": 7.171967241914224e-06,
"loss": 0.632,
"step": 1600
},
{
"epoch": 1.7800027812543457,
"eval_loss": 0.6654813289642334,
"eval_runtime": 38.0745,
"eval_samples_per_second": 10.69,
"eval_steps_per_second": 0.683,
"step": 1600
},
{
"epoch": 1.7855652899457657,
"grad_norm": 0.15930848069616485,
"learning_rate": 6.814879533580898e-06,
"loss": 0.6298,
"step": 1605
},
{
"epoch": 1.7911277986371854,
"grad_norm": 0.16408233749219603,
"learning_rate": 6.4665962339156005e-06,
"loss": 0.6383,
"step": 1610
},
{
"epoch": 1.7966903073286051,
"grad_norm": 0.1507405199602565,
"learning_rate": 6.127150249765335e-06,
"loss": 0.6538,
"step": 1615
},
{
"epoch": 1.802252816020025,
"grad_norm": 0.15449896745701766,
"learning_rate": 5.7965736530010916e-06,
"loss": 0.6443,
"step": 1620
},
{
"epoch": 1.807815324711445,
"grad_norm": 0.15430381806456947,
"learning_rate": 5.474897677487711e-06,
"loss": 0.6714,
"step": 1625
},
{
"epoch": 1.8133778334028647,
"grad_norm": 0.1522219198315468,
"learning_rate": 5.162152716132662e-06,
"loss": 0.6462,
"step": 1630
},
{
"epoch": 1.8189403420942845,
"grad_norm": 0.14950817969442623,
"learning_rate": 4.858368318014572e-06,
"loss": 0.6309,
"step": 1635
},
{
"epoch": 1.8245028507857044,
"grad_norm": 0.16343103461857358,
"learning_rate": 4.563573185591219e-06,
"loss": 0.6229,
"step": 1640
},
{
"epoch": 1.8300653594771243,
"grad_norm": 0.16046392285596128,
"learning_rate": 4.2777951719877415e-06,
"loss": 0.6526,
"step": 1645
},
{
"epoch": 1.835627868168544,
"grad_norm": 0.16358481948661036,
"learning_rate": 4.0010612783648925e-06,
"loss": 0.6407,
"step": 1650
},
{
"epoch": 1.8411903768599638,
"grad_norm": 0.15364511127769814,
"learning_rate": 3.7333976513680093e-06,
"loss": 0.6375,
"step": 1655
},
{
"epoch": 1.8467528855513837,
"grad_norm": 0.1582298293905816,
"learning_rate": 3.4748295806564356e-06,
"loss": 0.6486,
"step": 1660
},
{
"epoch": 1.8523153942428037,
"grad_norm": 0.1626273297442105,
"learning_rate": 3.2253814965142683e-06,
"loss": 0.6682,
"step": 1665
},
{
"epoch": 1.8578779029342232,
"grad_norm": 0.15566169967079257,
"learning_rate": 2.9850769675419774e-06,
"loss": 0.6531,
"step": 1670
},
{
"epoch": 1.8634404116256431,
"grad_norm": 0.1785306629362279,
"learning_rate": 2.7539386984296147e-06,
"loss": 0.6406,
"step": 1675
},
{
"epoch": 1.869002920317063,
"grad_norm": 0.14973888951794095,
"learning_rate": 2.5319885278115906e-06,
"loss": 0.6441,
"step": 1680
},
{
"epoch": 1.8745654290084828,
"grad_norm": 0.1493522152423315,
"learning_rate": 2.3192474262033638e-06,
"loss": 0.6377,
"step": 1685
},
{
"epoch": 1.8801279376999025,
"grad_norm": 0.1611037940241023,
"learning_rate": 2.115735494019966e-06,
"loss": 0.6785,
"step": 1690
},
{
"epoch": 1.8856904463913224,
"grad_norm": 0.15767721827989367,
"learning_rate": 1.921471959676957e-06,
"loss": 0.667,
"step": 1695
},
{
"epoch": 1.8912529550827424,
"grad_norm": 0.15987860636699017,
"learning_rate": 1.7364751777736332e-06,
"loss": 0.6374,
"step": 1700
},
{
"epoch": 1.896815463774162,
"grad_norm": 0.14274620551242612,
"learning_rate": 1.5607626273588138e-06,
"loss": 0.6305,
"step": 1705
},
{
"epoch": 1.9023779724655818,
"grad_norm": 0.1539956730318316,
"learning_rate": 1.394350910279385e-06,
"loss": 0.6446,
"step": 1710
},
{
"epoch": 1.9079404811570018,
"grad_norm": 0.16255063308547452,
"learning_rate": 1.2372557496116877e-06,
"loss": 0.6577,
"step": 1715
},
{
"epoch": 1.9135029898484217,
"grad_norm": 0.1457148304946188,
"learning_rate": 1.089491988176017e-06,
"loss": 0.6671,
"step": 1720
},
{
"epoch": 1.9190654985398414,
"grad_norm": 0.16991284681377727,
"learning_rate": 9.510735871341103e-07,
"loss": 0.6693,
"step": 1725
},
{
"epoch": 1.9246280072312612,
"grad_norm": 0.15748923168566117,
"learning_rate": 8.220136246701926e-07,
"loss": 0.6491,
"step": 1730
},
{
"epoch": 1.930190515922681,
"grad_norm": 0.15744123180048702,
"learning_rate": 7.023242947552078e-07,
"loss": 0.6366,
"step": 1735
},
{
"epoch": 1.935753024614101,
"grad_norm": 0.15109132227982583,
"learning_rate": 5.920169059947411e-07,
"loss": 0.6553,
"step": 1740
},
{
"epoch": 1.9413155333055208,
"grad_norm": 0.16542652334128852,
"learning_rate": 4.911018805605406e-07,
"loss": 0.6395,
"step": 1745
},
{
"epoch": 1.9468780419969405,
"grad_norm": 0.15918345337184334,
"learning_rate": 3.9958875320580404e-07,
"loss": 0.6502,
"step": 1750
},
{
"epoch": 1.9524405506883604,
"grad_norm": 0.17686582944306006,
"learning_rate": 3.1748617036427843e-07,
"loss": 0.6586,
"step": 1755
},
{
"epoch": 1.9580030593797804,
"grad_norm": 0.1426907741483869,
"learning_rate": 2.448018893333681e-07,
"loss": 0.6345,
"step": 1760
},
{
"epoch": 1.9635655680712,
"grad_norm": 0.14223763889913488,
"learning_rate": 1.815427775411549e-07,
"loss": 0.6296,
"step": 1765
},
{
"epoch": 1.9691280767626198,
"grad_norm": 0.14802968415791257,
"learning_rate": 1.277148118975835e-07,
"loss": 0.6167,
"step": 1770
},
{
"epoch": 1.9746905854540397,
"grad_norm": 0.16664999019696625,
"learning_rate": 8.332307822971342e-08,
"loss": 0.6522,
"step": 1775
},
{
"epoch": 1.9802530941454597,
"grad_norm": 0.15531644850942158,
"learning_rate": 4.837177080119215e-08,
"loss": 0.6587,
"step": 1780
},
{
"epoch": 1.9858156028368794,
"grad_norm": 0.17774727599107124,
"learning_rate": 2.286419191601663e-08,
"loss": 0.6634,
"step": 1785
},
{
"epoch": 1.9913781115282991,
"grad_norm": 0.1591360139721982,
"learning_rate": 6.8027516064606e-09,
"loss": 0.6466,
"step": 1790
},
{
"epoch": 1.996940620219719,
"grad_norm": 0.1816483541352177,
"learning_rate": 1.8896740540119696e-10,
"loss": 0.6307,
"step": 1795
},
{
"epoch": 1.998053121958003,
"step": 1796,
"total_flos": 7795310658584576.0,
"train_loss": 0.6907267017260957,
"train_runtime": 37673.7111,
"train_samples_per_second": 3.054,
"train_steps_per_second": 0.048
}
],
"logging_steps": 5,
"max_steps": 1796,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 100,
"total_flos": 7795310658584576.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}