SmolLM2-MagpieUltraPlus-Numina / trainer_state.json
loubnabnl's picture
loubnabnl HF staff
Model save
f9e157c verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 5334,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0009373828271466067,
"grad_norm": 55.87532592115914,
"learning_rate": 2.8089887640449436e-06,
"loss": 3.8993,
"step": 5
},
{
"epoch": 0.0018747656542932134,
"grad_norm": 39.52831505761594,
"learning_rate": 5.617977528089887e-06,
"loss": 3.4743,
"step": 10
},
{
"epoch": 0.00281214848143982,
"grad_norm": 22.428432087019985,
"learning_rate": 8.42696629213483e-06,
"loss": 2.8849,
"step": 15
},
{
"epoch": 0.0037495313085864268,
"grad_norm": 9.054209275536381,
"learning_rate": 1.1235955056179774e-05,
"loss": 2.1428,
"step": 20
},
{
"epoch": 0.004686914135733034,
"grad_norm": 3.948592126973317,
"learning_rate": 1.404494382022472e-05,
"loss": 1.7056,
"step": 25
},
{
"epoch": 0.00562429696287964,
"grad_norm": 2.299992616041375,
"learning_rate": 1.685393258426966e-05,
"loss": 1.4186,
"step": 30
},
{
"epoch": 0.006561679790026247,
"grad_norm": 0.9839714616377845,
"learning_rate": 1.9662921348314603e-05,
"loss": 1.2306,
"step": 35
},
{
"epoch": 0.0074990626171728535,
"grad_norm": 0.6008365887782849,
"learning_rate": 2.247191011235955e-05,
"loss": 1.1511,
"step": 40
},
{
"epoch": 0.00843644544431946,
"grad_norm": 0.4330325556834079,
"learning_rate": 2.528089887640449e-05,
"loss": 1.0624,
"step": 45
},
{
"epoch": 0.009373828271466067,
"grad_norm": 0.35553504278533465,
"learning_rate": 2.808988764044944e-05,
"loss": 1.0431,
"step": 50
},
{
"epoch": 0.010311211098612674,
"grad_norm": 0.25388532511026035,
"learning_rate": 3.089887640449438e-05,
"loss": 1.005,
"step": 55
},
{
"epoch": 0.01124859392575928,
"grad_norm": 0.2576041354688323,
"learning_rate": 3.370786516853932e-05,
"loss": 0.9996,
"step": 60
},
{
"epoch": 0.012185976752905886,
"grad_norm": 0.21819394992900415,
"learning_rate": 3.6516853932584265e-05,
"loss": 0.9775,
"step": 65
},
{
"epoch": 0.013123359580052493,
"grad_norm": 0.21574042351319167,
"learning_rate": 3.932584269662921e-05,
"loss": 0.9914,
"step": 70
},
{
"epoch": 0.0140607424071991,
"grad_norm": 0.18802524325518408,
"learning_rate": 4.2134831460674156e-05,
"loss": 0.9853,
"step": 75
},
{
"epoch": 0.014998125234345707,
"grad_norm": 0.18620252651860728,
"learning_rate": 4.49438202247191e-05,
"loss": 1.0079,
"step": 80
},
{
"epoch": 0.015935508061492312,
"grad_norm": 0.18097949689788895,
"learning_rate": 4.775280898876404e-05,
"loss": 0.9329,
"step": 85
},
{
"epoch": 0.01687289088863892,
"grad_norm": 0.15787887483249505,
"learning_rate": 5.056179775280898e-05,
"loss": 0.9277,
"step": 90
},
{
"epoch": 0.017810273715785526,
"grad_norm": 0.15373249141451148,
"learning_rate": 5.337078651685392e-05,
"loss": 0.9232,
"step": 95
},
{
"epoch": 0.018747656542932135,
"grad_norm": 0.13868806686970064,
"learning_rate": 5.617977528089888e-05,
"loss": 0.9002,
"step": 100
},
{
"epoch": 0.01968503937007874,
"grad_norm": 0.14099964891127995,
"learning_rate": 5.898876404494382e-05,
"loss": 0.916,
"step": 105
},
{
"epoch": 0.02062242219722535,
"grad_norm": 0.12190687602764205,
"learning_rate": 6.179775280898876e-05,
"loss": 0.926,
"step": 110
},
{
"epoch": 0.021559805024371954,
"grad_norm": 0.1146212972821343,
"learning_rate": 6.46067415730337e-05,
"loss": 0.9661,
"step": 115
},
{
"epoch": 0.02249718785151856,
"grad_norm": 0.1305834260724098,
"learning_rate": 6.741573033707865e-05,
"loss": 0.8979,
"step": 120
},
{
"epoch": 0.023434570678665168,
"grad_norm": 0.09286668479471845,
"learning_rate": 7.022471910112359e-05,
"loss": 0.895,
"step": 125
},
{
"epoch": 0.024371953505811773,
"grad_norm": 0.09893175557780264,
"learning_rate": 7.303370786516853e-05,
"loss": 0.8941,
"step": 130
},
{
"epoch": 0.02530933633295838,
"grad_norm": 0.08722137311816515,
"learning_rate": 7.584269662921347e-05,
"loss": 0.887,
"step": 135
},
{
"epoch": 0.026246719160104987,
"grad_norm": 0.08022348162278936,
"learning_rate": 7.865168539325841e-05,
"loss": 0.8804,
"step": 140
},
{
"epoch": 0.027184101987251595,
"grad_norm": 0.08856735925510693,
"learning_rate": 8.146067415730337e-05,
"loss": 0.855,
"step": 145
},
{
"epoch": 0.0281214848143982,
"grad_norm": 0.08311974627408422,
"learning_rate": 8.426966292134831e-05,
"loss": 0.8837,
"step": 150
},
{
"epoch": 0.029058867641544806,
"grad_norm": 0.07297089629042264,
"learning_rate": 8.707865168539325e-05,
"loss": 0.8748,
"step": 155
},
{
"epoch": 0.029996250468691414,
"grad_norm": 0.0720196207299981,
"learning_rate": 8.98876404494382e-05,
"loss": 0.8997,
"step": 160
},
{
"epoch": 0.03093363329583802,
"grad_norm": 0.07463772342009761,
"learning_rate": 9.269662921348314e-05,
"loss": 0.9228,
"step": 165
},
{
"epoch": 0.031871016122984624,
"grad_norm": 0.08058875574417819,
"learning_rate": 9.550561797752808e-05,
"loss": 0.8742,
"step": 170
},
{
"epoch": 0.03280839895013123,
"grad_norm": 0.07434935030584572,
"learning_rate": 9.831460674157303e-05,
"loss": 0.9169,
"step": 175
},
{
"epoch": 0.03374578177727784,
"grad_norm": 0.06516079400878816,
"learning_rate": 0.00010112359550561796,
"loss": 0.8842,
"step": 180
},
{
"epoch": 0.034683164604424443,
"grad_norm": 0.068946140527396,
"learning_rate": 0.00010393258426966292,
"loss": 0.8673,
"step": 185
},
{
"epoch": 0.03562054743157105,
"grad_norm": 0.0742056092302864,
"learning_rate": 0.00010674157303370785,
"loss": 0.8704,
"step": 190
},
{
"epoch": 0.03655793025871766,
"grad_norm": 0.09321869505325037,
"learning_rate": 0.0001095505617977528,
"loss": 0.8774,
"step": 195
},
{
"epoch": 0.03749531308586427,
"grad_norm": 0.07237285280180873,
"learning_rate": 0.00011235955056179776,
"loss": 0.8645,
"step": 200
},
{
"epoch": 0.03843269591301087,
"grad_norm": 0.0725790446638447,
"learning_rate": 0.00011516853932584269,
"loss": 0.8853,
"step": 205
},
{
"epoch": 0.03937007874015748,
"grad_norm": 0.06714726508233679,
"learning_rate": 0.00011797752808988764,
"loss": 0.8722,
"step": 210
},
{
"epoch": 0.04030746156730409,
"grad_norm": 0.06816724261165027,
"learning_rate": 0.00012078651685393257,
"loss": 0.857,
"step": 215
},
{
"epoch": 0.0412448443944507,
"grad_norm": 0.0630221201023211,
"learning_rate": 0.00012359550561797752,
"loss": 0.8844,
"step": 220
},
{
"epoch": 0.0421822272215973,
"grad_norm": 0.06920243449410893,
"learning_rate": 0.00012640449438202245,
"loss": 0.8786,
"step": 225
},
{
"epoch": 0.04311961004874391,
"grad_norm": 0.07088619559489383,
"learning_rate": 0.0001292134831460674,
"loss": 0.8797,
"step": 230
},
{
"epoch": 0.044056992875890516,
"grad_norm": 0.061740726041492694,
"learning_rate": 0.00013202247191011236,
"loss": 0.8698,
"step": 235
},
{
"epoch": 0.04499437570303712,
"grad_norm": 0.07106582585336212,
"learning_rate": 0.0001348314606741573,
"loss": 0.869,
"step": 240
},
{
"epoch": 0.045931758530183726,
"grad_norm": 0.061678059675451424,
"learning_rate": 0.00013764044943820225,
"loss": 0.8643,
"step": 245
},
{
"epoch": 0.046869141357330335,
"grad_norm": 0.06524467963947332,
"learning_rate": 0.00014044943820224718,
"loss": 0.8927,
"step": 250
},
{
"epoch": 0.047806524184476944,
"grad_norm": 0.0650786155112897,
"learning_rate": 0.00014325842696629213,
"loss": 0.8384,
"step": 255
},
{
"epoch": 0.048743907011623545,
"grad_norm": 0.07154242679452663,
"learning_rate": 0.00014606741573033706,
"loss": 0.8809,
"step": 260
},
{
"epoch": 0.049681289838770154,
"grad_norm": 0.07795179863304261,
"learning_rate": 0.00014887640449438202,
"loss": 0.8394,
"step": 265
},
{
"epoch": 0.05061867266591676,
"grad_norm": 0.06374006181294534,
"learning_rate": 0.00015168539325842694,
"loss": 0.8961,
"step": 270
},
{
"epoch": 0.051556055493063364,
"grad_norm": 0.07787607064114185,
"learning_rate": 0.0001544943820224719,
"loss": 0.8632,
"step": 275
},
{
"epoch": 0.05249343832020997,
"grad_norm": 0.06289687876280393,
"learning_rate": 0.00015730337078651683,
"loss": 0.88,
"step": 280
},
{
"epoch": 0.05343082114735658,
"grad_norm": 0.06491539830485436,
"learning_rate": 0.00016011235955056178,
"loss": 0.8737,
"step": 285
},
{
"epoch": 0.05436820397450319,
"grad_norm": 0.070328841548907,
"learning_rate": 0.00016292134831460674,
"loss": 0.8617,
"step": 290
},
{
"epoch": 0.05530558680164979,
"grad_norm": 0.0638014579038207,
"learning_rate": 0.00016573033707865167,
"loss": 0.8506,
"step": 295
},
{
"epoch": 0.0562429696287964,
"grad_norm": 0.06422804662460813,
"learning_rate": 0.00016853932584269662,
"loss": 0.8805,
"step": 300
},
{
"epoch": 0.05718035245594301,
"grad_norm": 0.06361653985204194,
"learning_rate": 0.00017134831460674155,
"loss": 0.8918,
"step": 305
},
{
"epoch": 0.05811773528308961,
"grad_norm": 0.06932335644552794,
"learning_rate": 0.0001741573033707865,
"loss": 0.8595,
"step": 310
},
{
"epoch": 0.05905511811023622,
"grad_norm": 0.07957054733316223,
"learning_rate": 0.00017696629213483143,
"loss": 0.8663,
"step": 315
},
{
"epoch": 0.05999250093738283,
"grad_norm": 0.06998445535720693,
"learning_rate": 0.0001797752808988764,
"loss": 0.8676,
"step": 320
},
{
"epoch": 0.06092988376452944,
"grad_norm": 0.06702795111487583,
"learning_rate": 0.00018258426966292135,
"loss": 0.8802,
"step": 325
},
{
"epoch": 0.06186726659167604,
"grad_norm": 0.057562966446872095,
"learning_rate": 0.00018539325842696627,
"loss": 0.8671,
"step": 330
},
{
"epoch": 0.06280464941882265,
"grad_norm": 0.07214305689508435,
"learning_rate": 0.0001882022471910112,
"loss": 0.8992,
"step": 335
},
{
"epoch": 0.06374203224596925,
"grad_norm": 0.06640078935309972,
"learning_rate": 0.00019101123595505616,
"loss": 0.8589,
"step": 340
},
{
"epoch": 0.06467941507311586,
"grad_norm": 0.06432108168682822,
"learning_rate": 0.0001938202247191011,
"loss": 0.8792,
"step": 345
},
{
"epoch": 0.06561679790026247,
"grad_norm": 0.06995296734682956,
"learning_rate": 0.00019662921348314607,
"loss": 0.865,
"step": 350
},
{
"epoch": 0.06655418072740907,
"grad_norm": 0.0781580610323253,
"learning_rate": 0.00019943820224719097,
"loss": 0.8815,
"step": 355
},
{
"epoch": 0.06749156355455568,
"grad_norm": 0.07071637338397053,
"learning_rate": 0.00020224719101123593,
"loss": 0.8666,
"step": 360
},
{
"epoch": 0.06842894638170229,
"grad_norm": 0.0682378939648076,
"learning_rate": 0.00020505617977528088,
"loss": 0.8862,
"step": 365
},
{
"epoch": 0.06936632920884889,
"grad_norm": 0.06900860789126333,
"learning_rate": 0.00020786516853932584,
"loss": 0.8638,
"step": 370
},
{
"epoch": 0.0703037120359955,
"grad_norm": 0.06343642719314092,
"learning_rate": 0.0002106741573033708,
"loss": 0.8692,
"step": 375
},
{
"epoch": 0.0712410948631421,
"grad_norm": 0.058805566981791894,
"learning_rate": 0.0002134831460674157,
"loss": 0.8699,
"step": 380
},
{
"epoch": 0.07217847769028872,
"grad_norm": 0.06501680635838757,
"learning_rate": 0.00021629213483146065,
"loss": 0.8794,
"step": 385
},
{
"epoch": 0.07311586051743532,
"grad_norm": 0.06790932619576265,
"learning_rate": 0.0002191011235955056,
"loss": 0.8609,
"step": 390
},
{
"epoch": 0.07405324334458192,
"grad_norm": 0.07253373879567401,
"learning_rate": 0.00022191011235955056,
"loss": 0.8586,
"step": 395
},
{
"epoch": 0.07499062617172854,
"grad_norm": 0.05597241270795834,
"learning_rate": 0.00022471910112359551,
"loss": 0.8606,
"step": 400
},
{
"epoch": 0.07592800899887514,
"grad_norm": 0.07414431562106621,
"learning_rate": 0.00022752808988764042,
"loss": 0.859,
"step": 405
},
{
"epoch": 0.07686539182602174,
"grad_norm": 0.06416271464203936,
"learning_rate": 0.00023033707865168537,
"loss": 0.8462,
"step": 410
},
{
"epoch": 0.07780277465316836,
"grad_norm": 0.06411386307806086,
"learning_rate": 0.00023314606741573033,
"loss": 0.8859,
"step": 415
},
{
"epoch": 0.07874015748031496,
"grad_norm": 0.08671615531146802,
"learning_rate": 0.00023595505617977528,
"loss": 0.858,
"step": 420
},
{
"epoch": 0.07967754030746156,
"grad_norm": 0.06950084739281429,
"learning_rate": 0.00023876404494382018,
"loss": 0.8779,
"step": 425
},
{
"epoch": 0.08061492313460818,
"grad_norm": 0.06732606238114026,
"learning_rate": 0.00024157303370786514,
"loss": 0.8585,
"step": 430
},
{
"epoch": 0.08155230596175478,
"grad_norm": 0.06753409562352092,
"learning_rate": 0.0002443820224719101,
"loss": 0.8648,
"step": 435
},
{
"epoch": 0.0824896887889014,
"grad_norm": 0.06868527875344918,
"learning_rate": 0.00024719101123595505,
"loss": 0.8923,
"step": 440
},
{
"epoch": 0.083427071616048,
"grad_norm": 0.060659330441044046,
"learning_rate": 0.00025,
"loss": 0.8441,
"step": 445
},
{
"epoch": 0.0843644544431946,
"grad_norm": 0.056891953999125895,
"learning_rate": 0.0002528089887640449,
"loss": 0.8711,
"step": 450
},
{
"epoch": 0.08530183727034121,
"grad_norm": 0.06315987668061082,
"learning_rate": 0.00025561797752808986,
"loss": 0.8709,
"step": 455
},
{
"epoch": 0.08623922009748781,
"grad_norm": 0.06974874046223557,
"learning_rate": 0.0002584269662921348,
"loss": 0.8881,
"step": 460
},
{
"epoch": 0.08717660292463442,
"grad_norm": 0.07106045458363727,
"learning_rate": 0.00026123595505617977,
"loss": 0.8455,
"step": 465
},
{
"epoch": 0.08811398575178103,
"grad_norm": 0.06530109700061691,
"learning_rate": 0.00026404494382022473,
"loss": 0.8701,
"step": 470
},
{
"epoch": 0.08905136857892763,
"grad_norm": 0.06284436680171902,
"learning_rate": 0.00026685393258426963,
"loss": 0.8639,
"step": 475
},
{
"epoch": 0.08998875140607424,
"grad_norm": 0.06695190689656623,
"learning_rate": 0.0002696629213483146,
"loss": 0.8596,
"step": 480
},
{
"epoch": 0.09092613423322085,
"grad_norm": 0.06134041035709648,
"learning_rate": 0.00027247191011235954,
"loss": 0.8846,
"step": 485
},
{
"epoch": 0.09186351706036745,
"grad_norm": 0.05390358172595912,
"learning_rate": 0.0002752808988764045,
"loss": 0.8608,
"step": 490
},
{
"epoch": 0.09280089988751405,
"grad_norm": 0.06061308522705228,
"learning_rate": 0.00027808988764044945,
"loss": 0.8688,
"step": 495
},
{
"epoch": 0.09373828271466067,
"grad_norm": 0.05665308816810634,
"learning_rate": 0.00028089887640449435,
"loss": 0.8431,
"step": 500
},
{
"epoch": 0.09467566554180727,
"grad_norm": 0.06298526377992197,
"learning_rate": 0.0002837078651685393,
"loss": 0.8534,
"step": 505
},
{
"epoch": 0.09561304836895389,
"grad_norm": 0.06828852662205184,
"learning_rate": 0.00028651685393258426,
"loss": 0.855,
"step": 510
},
{
"epoch": 0.09655043119610049,
"grad_norm": 0.059581044094172864,
"learning_rate": 0.0002893258426966292,
"loss": 0.892,
"step": 515
},
{
"epoch": 0.09748781402324709,
"grad_norm": 0.05834286349398695,
"learning_rate": 0.0002921348314606741,
"loss": 0.8791,
"step": 520
},
{
"epoch": 0.0984251968503937,
"grad_norm": 0.05628861715462521,
"learning_rate": 0.0002949438202247191,
"loss": 0.8572,
"step": 525
},
{
"epoch": 0.09936257967754031,
"grad_norm": 0.0752400574763716,
"learning_rate": 0.00029775280898876403,
"loss": 0.9176,
"step": 530
},
{
"epoch": 0.10029996250468691,
"grad_norm": 0.0589889669786586,
"learning_rate": 0.0002999999678723826,
"loss": 0.879,
"step": 535
},
{
"epoch": 0.10123734533183353,
"grad_norm": 0.06984453009723764,
"learning_rate": 0.0002999988434072206,
"loss": 0.8791,
"step": 540
},
{
"epoch": 0.10217472815898013,
"grad_norm": 0.05131555199439601,
"learning_rate": 0.0002999961125749536,
"loss": 0.8639,
"step": 545
},
{
"epoch": 0.10311211098612673,
"grad_norm": 0.06129502320830477,
"learning_rate": 0.0002999917754048268,
"loss": 0.8626,
"step": 550
},
{
"epoch": 0.10404949381327334,
"grad_norm": 0.052116063904268276,
"learning_rate": 0.00029998583194328776,
"loss": 0.8405,
"step": 555
},
{
"epoch": 0.10498687664041995,
"grad_norm": 0.05805086354219647,
"learning_rate": 0.0002999782822539861,
"loss": 0.8521,
"step": 560
},
{
"epoch": 0.10592425946756655,
"grad_norm": 0.058854180699594805,
"learning_rate": 0.000299969126417773,
"loss": 0.8612,
"step": 565
},
{
"epoch": 0.10686164229471316,
"grad_norm": 0.05170888163012354,
"learning_rate": 0.00029995836453270005,
"loss": 0.852,
"step": 570
},
{
"epoch": 0.10779902512185977,
"grad_norm": 0.053497899422118654,
"learning_rate": 0.0002999459967140185,
"loss": 0.8425,
"step": 575
},
{
"epoch": 0.10873640794900638,
"grad_norm": 0.0517344152203251,
"learning_rate": 0.00029993202309417765,
"loss": 0.8666,
"step": 580
},
{
"epoch": 0.10967379077615298,
"grad_norm": 0.06883148118348337,
"learning_rate": 0.00029991644382282377,
"loss": 0.8699,
"step": 585
},
{
"epoch": 0.11061117360329958,
"grad_norm": 0.06701524872749254,
"learning_rate": 0.0002998992590667984,
"loss": 0.8612,
"step": 590
},
{
"epoch": 0.1115485564304462,
"grad_norm": 0.0633741891240403,
"learning_rate": 0.00029988046901013643,
"loss": 0.8861,
"step": 595
},
{
"epoch": 0.1124859392575928,
"grad_norm": 0.059595235950932066,
"learning_rate": 0.00029986007385406424,
"loss": 0.8761,
"step": 600
},
{
"epoch": 0.1134233220847394,
"grad_norm": 0.05592116325579977,
"learning_rate": 0.00029983807381699757,
"loss": 0.8756,
"step": 605
},
{
"epoch": 0.11436070491188602,
"grad_norm": 0.057345108961587225,
"learning_rate": 0.0002998144691345392,
"loss": 0.8574,
"step": 610
},
{
"epoch": 0.11529808773903262,
"grad_norm": 0.05734473521251835,
"learning_rate": 0.0002997892600594762,
"loss": 0.8544,
"step": 615
},
{
"epoch": 0.11623547056617922,
"grad_norm": 0.05186000961639975,
"learning_rate": 0.00029976244686177764,
"loss": 0.8808,
"step": 620
},
{
"epoch": 0.11717285339332584,
"grad_norm": 0.05414634714867266,
"learning_rate": 0.00029973402982859127,
"loss": 0.8331,
"step": 625
},
{
"epoch": 0.11811023622047244,
"grad_norm": 0.04968211925070222,
"learning_rate": 0.0002997040092642407,
"loss": 0.8592,
"step": 630
},
{
"epoch": 0.11904761904761904,
"grad_norm": 0.06248537463553037,
"learning_rate": 0.00029967238549022206,
"loss": 0.8819,
"step": 635
},
{
"epoch": 0.11998500187476566,
"grad_norm": 0.055299575032660746,
"learning_rate": 0.00029963915884520054,
"loss": 0.8636,
"step": 640
},
{
"epoch": 0.12092238470191226,
"grad_norm": 0.04854206172628792,
"learning_rate": 0.00029960432968500675,
"loss": 0.8628,
"step": 645
},
{
"epoch": 0.12185976752905887,
"grad_norm": 0.058596631103748714,
"learning_rate": 0.00029956789838263314,
"loss": 0.8489,
"step": 650
},
{
"epoch": 0.12279715035620548,
"grad_norm": 0.06652964862071063,
"learning_rate": 0.0002995298653282297,
"loss": 0.876,
"step": 655
},
{
"epoch": 0.12373453318335208,
"grad_norm": 0.05617071779521591,
"learning_rate": 0.00029949023092909976,
"loss": 0.8582,
"step": 660
},
{
"epoch": 0.12467191601049869,
"grad_norm": 0.06128826182928016,
"learning_rate": 0.00029944899560969593,
"loss": 0.8556,
"step": 665
},
{
"epoch": 0.1256092988376453,
"grad_norm": 0.06596974095737278,
"learning_rate": 0.00029940615981161544,
"loss": 0.8484,
"step": 670
},
{
"epoch": 0.1265466816647919,
"grad_norm": 0.05241653325144218,
"learning_rate": 0.00029936172399359516,
"loss": 0.8681,
"step": 675
},
{
"epoch": 0.1274840644919385,
"grad_norm": 0.0656117272709645,
"learning_rate": 0.00029931568863150705,
"loss": 0.8611,
"step": 680
},
{
"epoch": 0.1284214473190851,
"grad_norm": 0.05416806152576802,
"learning_rate": 0.0002992680542183529,
"loss": 0.8543,
"step": 685
},
{
"epoch": 0.12935883014623173,
"grad_norm": 0.05949667729922976,
"learning_rate": 0.00029921882126425893,
"loss": 0.8476,
"step": 690
},
{
"epoch": 0.13029621297337832,
"grad_norm": 0.05476709738462092,
"learning_rate": 0.0002991679902964706,
"loss": 0.8329,
"step": 695
},
{
"epoch": 0.13123359580052493,
"grad_norm": 0.05583654578373451,
"learning_rate": 0.00029911556185934667,
"loss": 0.8546,
"step": 700
},
{
"epoch": 0.13217097862767155,
"grad_norm": 0.052842518474849304,
"learning_rate": 0.0002990615365143536,
"loss": 0.8576,
"step": 705
},
{
"epoch": 0.13310836145481814,
"grad_norm": 0.05325159315281095,
"learning_rate": 0.0002990059148400594,
"loss": 0.8475,
"step": 710
},
{
"epoch": 0.13404574428196475,
"grad_norm": 0.0477613812018967,
"learning_rate": 0.00029894869743212767,
"loss": 0.8505,
"step": 715
},
{
"epoch": 0.13498312710911137,
"grad_norm": 0.04650972000773531,
"learning_rate": 0.00029888988490331067,
"loss": 0.8406,
"step": 720
},
{
"epoch": 0.13592050993625795,
"grad_norm": 0.05548534041152614,
"learning_rate": 0.00029882947788344345,
"loss": 0.8731,
"step": 725
},
{
"epoch": 0.13685789276340457,
"grad_norm": 0.047893477001569186,
"learning_rate": 0.00029876747701943667,
"loss": 0.8666,
"step": 730
},
{
"epoch": 0.1377952755905512,
"grad_norm": 0.05399917193542452,
"learning_rate": 0.00029870388297526966,
"loss": 0.8476,
"step": 735
},
{
"epoch": 0.13873265841769777,
"grad_norm": 0.04738143330368603,
"learning_rate": 0.0002986386964319837,
"loss": 0.8423,
"step": 740
},
{
"epoch": 0.1396700412448444,
"grad_norm": 0.05491107440574904,
"learning_rate": 0.0002985719180876742,
"loss": 0.8451,
"step": 745
},
{
"epoch": 0.140607424071991,
"grad_norm": 0.060398470624460965,
"learning_rate": 0.0002985035486574836,
"loss": 0.8801,
"step": 750
},
{
"epoch": 0.14154480689913762,
"grad_norm": 0.05350788151410978,
"learning_rate": 0.00029843358887359357,
"loss": 0.8516,
"step": 755
},
{
"epoch": 0.1424821897262842,
"grad_norm": 0.05517308695615106,
"learning_rate": 0.0002983620394852172,
"loss": 0.8703,
"step": 760
},
{
"epoch": 0.14341957255343082,
"grad_norm": 0.054231118168541766,
"learning_rate": 0.000298288901258591,
"loss": 0.8693,
"step": 765
},
{
"epoch": 0.14435695538057744,
"grad_norm": 0.049622640241897456,
"learning_rate": 0.0002982141749769665,
"loss": 0.8395,
"step": 770
},
{
"epoch": 0.14529433820772403,
"grad_norm": 0.046995063624861166,
"learning_rate": 0.0002981378614406022,
"loss": 0.8604,
"step": 775
},
{
"epoch": 0.14623172103487064,
"grad_norm": 0.05520295521540041,
"learning_rate": 0.0002980599614667548,
"loss": 0.8645,
"step": 780
},
{
"epoch": 0.14716910386201726,
"grad_norm": 0.04894265506037402,
"learning_rate": 0.0002979804758896704,
"loss": 0.8652,
"step": 785
},
{
"epoch": 0.14810648668916385,
"grad_norm": 0.06029411158482043,
"learning_rate": 0.0002978994055605757,
"loss": 0.8416,
"step": 790
},
{
"epoch": 0.14904386951631046,
"grad_norm": 0.05780387639217229,
"learning_rate": 0.0002978167513476688,
"loss": 0.8526,
"step": 795
},
{
"epoch": 0.14998125234345708,
"grad_norm": 0.05465366981027733,
"learning_rate": 0.00029773251413610987,
"loss": 0.8655,
"step": 800
},
{
"epoch": 0.15091863517060367,
"grad_norm": 0.05272490905818877,
"learning_rate": 0.00029764669482801174,
"loss": 0.8519,
"step": 805
},
{
"epoch": 0.15185601799775028,
"grad_norm": 0.05718657624393557,
"learning_rate": 0.00029755929434243034,
"loss": 0.8853,
"step": 810
},
{
"epoch": 0.1527934008248969,
"grad_norm": 0.05197508584506214,
"learning_rate": 0.00029747031361535464,
"loss": 0.8349,
"step": 815
},
{
"epoch": 0.15373078365204348,
"grad_norm": 0.05481428649702033,
"learning_rate": 0.0002973797535996967,
"loss": 0.8627,
"step": 820
},
{
"epoch": 0.1546681664791901,
"grad_norm": 0.04732127095107982,
"learning_rate": 0.00029728761526528157,
"loss": 0.8698,
"step": 825
},
{
"epoch": 0.15560554930633672,
"grad_norm": 0.05743584601025622,
"learning_rate": 0.00029719389959883673,
"loss": 0.8736,
"step": 830
},
{
"epoch": 0.1565429321334833,
"grad_norm": 0.05196494960994641,
"learning_rate": 0.00029709860760398176,
"loss": 0.8634,
"step": 835
},
{
"epoch": 0.15748031496062992,
"grad_norm": 0.05085488275567502,
"learning_rate": 0.0002970017403012173,
"loss": 0.8568,
"step": 840
},
{
"epoch": 0.15841769778777653,
"grad_norm": 0.050395431338846663,
"learning_rate": 0.0002969032987279144,
"loss": 0.8225,
"step": 845
},
{
"epoch": 0.15935508061492312,
"grad_norm": 0.04628767205411848,
"learning_rate": 0.00029680328393830315,
"loss": 0.875,
"step": 850
},
{
"epoch": 0.16029246344206974,
"grad_norm": 0.051723935710402326,
"learning_rate": 0.00029670169700346164,
"loss": 0.8145,
"step": 855
},
{
"epoch": 0.16122984626921635,
"grad_norm": 0.05201508957810899,
"learning_rate": 0.0002965985390113043,
"loss": 0.8648,
"step": 860
},
{
"epoch": 0.16216722909636294,
"grad_norm": 0.0495933900905848,
"learning_rate": 0.0002964938110665704,
"loss": 0.8587,
"step": 865
},
{
"epoch": 0.16310461192350956,
"grad_norm": 0.04586671475206476,
"learning_rate": 0.0002963875142908121,
"loss": 0.8412,
"step": 870
},
{
"epoch": 0.16404199475065617,
"grad_norm": 0.0532312699053512,
"learning_rate": 0.00029627964982238236,
"loss": 0.842,
"step": 875
},
{
"epoch": 0.1649793775778028,
"grad_norm": 0.043681059319998494,
"learning_rate": 0.0002961702188164231,
"loss": 0.8274,
"step": 880
},
{
"epoch": 0.16591676040494938,
"grad_norm": 0.048557735504717925,
"learning_rate": 0.0002960592224448524,
"loss": 0.8426,
"step": 885
},
{
"epoch": 0.166854143232096,
"grad_norm": 0.04679289724415963,
"learning_rate": 0.00029594666189635224,
"loss": 0.8347,
"step": 890
},
{
"epoch": 0.1677915260592426,
"grad_norm": 0.047070279337964094,
"learning_rate": 0.00029583253837635575,
"loss": 0.8456,
"step": 895
},
{
"epoch": 0.1687289088863892,
"grad_norm": 0.05283241949368948,
"learning_rate": 0.00029571685310703403,
"loss": 0.8326,
"step": 900
},
{
"epoch": 0.1696662917135358,
"grad_norm": 0.05385523710797248,
"learning_rate": 0.00029559960732728337,
"loss": 0.8529,
"step": 905
},
{
"epoch": 0.17060367454068243,
"grad_norm": 0.04730829638902959,
"learning_rate": 0.000295480802292712,
"loss": 0.8085,
"step": 910
},
{
"epoch": 0.17154105736782901,
"grad_norm": 0.04488134785670412,
"learning_rate": 0.0002953604392756263,
"loss": 0.8371,
"step": 915
},
{
"epoch": 0.17247844019497563,
"grad_norm": 0.051313805015627746,
"learning_rate": 0.00029523851956501744,
"loss": 0.8486,
"step": 920
},
{
"epoch": 0.17341582302212225,
"grad_norm": 0.0430483274069335,
"learning_rate": 0.00029511504446654767,
"loss": 0.8475,
"step": 925
},
{
"epoch": 0.17435320584926883,
"grad_norm": 0.05543052805044051,
"learning_rate": 0.00029499001530253606,
"loss": 0.8571,
"step": 930
},
{
"epoch": 0.17529058867641545,
"grad_norm": 0.05160006837470796,
"learning_rate": 0.0002948634334119445,
"loss": 0.8348,
"step": 935
},
{
"epoch": 0.17622797150356206,
"grad_norm": 0.049000725000881734,
"learning_rate": 0.00029473530015036335,
"loss": 0.8243,
"step": 940
},
{
"epoch": 0.17716535433070865,
"grad_norm": 0.05324980516591907,
"learning_rate": 0.0002946056168899969,
"loss": 0.817,
"step": 945
},
{
"epoch": 0.17810273715785527,
"grad_norm": 0.05112366142864306,
"learning_rate": 0.00029447438501964873,
"loss": 0.8493,
"step": 950
},
{
"epoch": 0.17904011998500188,
"grad_norm": 0.0498102110793069,
"learning_rate": 0.0002943416059447066,
"loss": 0.8155,
"step": 955
},
{
"epoch": 0.17997750281214847,
"grad_norm": 0.048456183794281356,
"learning_rate": 0.0002942072810871279,
"loss": 0.8057,
"step": 960
},
{
"epoch": 0.1809148856392951,
"grad_norm": 0.0514791482209668,
"learning_rate": 0.0002940714118854238,
"loss": 0.8125,
"step": 965
},
{
"epoch": 0.1818522684664417,
"grad_norm": 0.04725030159140053,
"learning_rate": 0.0002939339997946444,
"loss": 0.8572,
"step": 970
},
{
"epoch": 0.1827896512935883,
"grad_norm": 0.05374547829970908,
"learning_rate": 0.0002937950462863627,
"loss": 0.8328,
"step": 975
},
{
"epoch": 0.1837270341207349,
"grad_norm": 0.0562408856971059,
"learning_rate": 0.00029365455284865923,
"loss": 0.8087,
"step": 980
},
{
"epoch": 0.18466441694788152,
"grad_norm": 0.0455189157494031,
"learning_rate": 0.00029351252098610577,
"loss": 0.8418,
"step": 985
},
{
"epoch": 0.1856017997750281,
"grad_norm": 0.050711487872205016,
"learning_rate": 0.00029336895221974946,
"loss": 0.8482,
"step": 990
},
{
"epoch": 0.18653918260217472,
"grad_norm": 0.04919906548209661,
"learning_rate": 0.00029322384808709654,
"loss": 0.8349,
"step": 995
},
{
"epoch": 0.18747656542932134,
"grad_norm": 0.05662595028293593,
"learning_rate": 0.00029307721014209555,
"loss": 0.8577,
"step": 1000
},
{
"epoch": 0.18841394825646793,
"grad_norm": 0.057447476057492466,
"learning_rate": 0.00029292903995512123,
"loss": 0.8534,
"step": 1005
},
{
"epoch": 0.18935133108361454,
"grad_norm": 0.05299814571131224,
"learning_rate": 0.0002927793391129571,
"loss": 0.8577,
"step": 1010
},
{
"epoch": 0.19028871391076116,
"grad_norm": 0.04319708352419991,
"learning_rate": 0.00029262810921877906,
"loss": 0.8188,
"step": 1015
},
{
"epoch": 0.19122609673790777,
"grad_norm": 0.05353832659477085,
"learning_rate": 0.0002924753518921376,
"loss": 0.846,
"step": 1020
},
{
"epoch": 0.19216347956505436,
"grad_norm": 0.04966007615937987,
"learning_rate": 0.0002923210687689411,
"loss": 0.8552,
"step": 1025
},
{
"epoch": 0.19310086239220098,
"grad_norm": 0.048671986424680805,
"learning_rate": 0.00029216526150143785,
"loss": 0.8433,
"step": 1030
},
{
"epoch": 0.1940382452193476,
"grad_norm": 0.05118103987659261,
"learning_rate": 0.0002920079317581984,
"loss": 0.8545,
"step": 1035
},
{
"epoch": 0.19497562804649418,
"grad_norm": 0.04730760417822063,
"learning_rate": 0.00029184908122409804,
"loss": 0.8255,
"step": 1040
},
{
"epoch": 0.1959130108736408,
"grad_norm": 0.05859108169984351,
"learning_rate": 0.0002916887116002983,
"loss": 0.8391,
"step": 1045
},
{
"epoch": 0.1968503937007874,
"grad_norm": 0.05306001691326635,
"learning_rate": 0.000291526824604229,
"loss": 0.8424,
"step": 1050
},
{
"epoch": 0.197787776527934,
"grad_norm": 0.047277287716965946,
"learning_rate": 0.00029136342196956985,
"loss": 0.833,
"step": 1055
},
{
"epoch": 0.19872515935508062,
"grad_norm": 0.05159138803836899,
"learning_rate": 0.0002911985054462318,
"loss": 0.8304,
"step": 1060
},
{
"epoch": 0.19966254218222723,
"grad_norm": 0.046550244925313046,
"learning_rate": 0.00029103207680033827,
"loss": 0.8422,
"step": 1065
},
{
"epoch": 0.20059992500937382,
"grad_norm": 0.046411253292755236,
"learning_rate": 0.00029086413781420633,
"loss": 0.8575,
"step": 1070
},
{
"epoch": 0.20153730783652044,
"grad_norm": 0.04762694973314196,
"learning_rate": 0.0002906946902863277,
"loss": 0.8541,
"step": 1075
},
{
"epoch": 0.20247469066366705,
"grad_norm": 0.043558681229160734,
"learning_rate": 0.0002905237360313492,
"loss": 0.8509,
"step": 1080
},
{
"epoch": 0.20341207349081364,
"grad_norm": 0.047235839608751946,
"learning_rate": 0.00029035127688005355,
"loss": 0.8326,
"step": 1085
},
{
"epoch": 0.20434945631796025,
"grad_norm": 0.04405113672607244,
"learning_rate": 0.00029017731467933974,
"loss": 0.8235,
"step": 1090
},
{
"epoch": 0.20528683914510687,
"grad_norm": 0.04917819994251594,
"learning_rate": 0.0002900018512922032,
"loss": 0.8514,
"step": 1095
},
{
"epoch": 0.20622422197225346,
"grad_norm": 0.045391849870203616,
"learning_rate": 0.0002898248885977158,
"loss": 0.841,
"step": 1100
},
{
"epoch": 0.20716160479940007,
"grad_norm": 0.04950574892382217,
"learning_rate": 0.0002896464284910058,
"loss": 0.8604,
"step": 1105
},
{
"epoch": 0.2080989876265467,
"grad_norm": 0.04483838357821004,
"learning_rate": 0.00028946647288323766,
"loss": 0.8477,
"step": 1110
},
{
"epoch": 0.20903637045369328,
"grad_norm": 0.04819763033021056,
"learning_rate": 0.00028928502370159133,
"loss": 0.8301,
"step": 1115
},
{
"epoch": 0.2099737532808399,
"grad_norm": 0.04846529213809432,
"learning_rate": 0.0002891020828892417,
"loss": 0.8376,
"step": 1120
},
{
"epoch": 0.2109111361079865,
"grad_norm": 0.0467946353030176,
"learning_rate": 0.00028891765240533795,
"loss": 0.8705,
"step": 1125
},
{
"epoch": 0.2118485189351331,
"grad_norm": 0.044000801667735845,
"learning_rate": 0.00028873173422498243,
"loss": 0.8281,
"step": 1130
},
{
"epoch": 0.2127859017622797,
"grad_norm": 0.047502640131898,
"learning_rate": 0.0002885443303392094,
"loss": 0.8381,
"step": 1135
},
{
"epoch": 0.21372328458942633,
"grad_norm": 0.04801904197558332,
"learning_rate": 0.000288355442754964,
"loss": 0.8179,
"step": 1140
},
{
"epoch": 0.21466066741657294,
"grad_norm": 0.04261523390710894,
"learning_rate": 0.00028816507349508047,
"loss": 0.8263,
"step": 1145
},
{
"epoch": 0.21559805024371953,
"grad_norm": 0.047191156643762354,
"learning_rate": 0.00028797322459826063,
"loss": 0.8374,
"step": 1150
},
{
"epoch": 0.21653543307086615,
"grad_norm": 0.05167657856075751,
"learning_rate": 0.00028777989811905205,
"loss": 0.8614,
"step": 1155
},
{
"epoch": 0.21747281589801276,
"grad_norm": 0.048897121204297574,
"learning_rate": 0.000287585096127826,
"loss": 0.8538,
"step": 1160
},
{
"epoch": 0.21841019872515935,
"grad_norm": 0.043161388562608975,
"learning_rate": 0.0002873888207107553,
"loss": 0.8457,
"step": 1165
},
{
"epoch": 0.21934758155230596,
"grad_norm": 0.054581961552759836,
"learning_rate": 0.000287191073969792,
"loss": 0.8362,
"step": 1170
},
{
"epoch": 0.22028496437945258,
"grad_norm": 0.05058965528101186,
"learning_rate": 0.0002869918580226448,
"loss": 0.8687,
"step": 1175
},
{
"epoch": 0.22122234720659917,
"grad_norm": 0.04739504236542665,
"learning_rate": 0.00028679117500275653,
"loss": 0.831,
"step": 1180
},
{
"epoch": 0.22215973003374578,
"grad_norm": 0.045754358105207775,
"learning_rate": 0.00028658902705928094,
"loss": 0.868,
"step": 1185
},
{
"epoch": 0.2230971128608924,
"grad_norm": 0.04508332677839864,
"learning_rate": 0.00028638541635706027,
"loss": 0.8129,
"step": 1190
},
{
"epoch": 0.224034495688039,
"grad_norm": 0.047178895399584496,
"learning_rate": 0.00028618034507660144,
"loss": 0.8548,
"step": 1195
},
{
"epoch": 0.2249718785151856,
"grad_norm": 0.04249857153746906,
"learning_rate": 0.0002859738154140532,
"loss": 0.8171,
"step": 1200
},
{
"epoch": 0.22590926134233222,
"grad_norm": 0.04271090221572632,
"learning_rate": 0.00028576582958118223,
"loss": 0.8274,
"step": 1205
},
{
"epoch": 0.2268466441694788,
"grad_norm": 0.04664121668293475,
"learning_rate": 0.00028555638980534974,
"loss": 0.8374,
"step": 1210
},
{
"epoch": 0.22778402699662542,
"grad_norm": 0.045498060568064186,
"learning_rate": 0.0002853454983294875,
"loss": 0.8039,
"step": 1215
},
{
"epoch": 0.22872140982377204,
"grad_norm": 0.04534000569075493,
"learning_rate": 0.0002851331574120738,
"loss": 0.858,
"step": 1220
},
{
"epoch": 0.22965879265091863,
"grad_norm": 0.04664900784446732,
"learning_rate": 0.00028491936932710917,
"loss": 0.8402,
"step": 1225
},
{
"epoch": 0.23059617547806524,
"grad_norm": 0.04417588784422196,
"learning_rate": 0.0002847041363640923,
"loss": 0.8207,
"step": 1230
},
{
"epoch": 0.23153355830521186,
"grad_norm": 0.049652178225426,
"learning_rate": 0.0002844874608279954,
"loss": 0.8515,
"step": 1235
},
{
"epoch": 0.23247094113235844,
"grad_norm": 0.048818844243959776,
"learning_rate": 0.00028426934503923923,
"loss": 0.8322,
"step": 1240
},
{
"epoch": 0.23340832395950506,
"grad_norm": 0.04283049790199607,
"learning_rate": 0.0002840497913336687,
"loss": 0.8247,
"step": 1245
},
{
"epoch": 0.23434570678665168,
"grad_norm": 0.04305639669050898,
"learning_rate": 0.0002838288020625277,
"loss": 0.8273,
"step": 1250
},
{
"epoch": 0.23528308961379826,
"grad_norm": 0.046845793494778305,
"learning_rate": 0.00028360637959243365,
"loss": 0.8628,
"step": 1255
},
{
"epoch": 0.23622047244094488,
"grad_norm": 0.046154254481951956,
"learning_rate": 0.00028338252630535264,
"loss": 0.8565,
"step": 1260
},
{
"epoch": 0.2371578552680915,
"grad_norm": 0.04282295436813289,
"learning_rate": 0.00028315724459857346,
"loss": 0.8468,
"step": 1265
},
{
"epoch": 0.23809523809523808,
"grad_norm": 0.04767889894413164,
"learning_rate": 0.00028293053688468214,
"loss": 0.81,
"step": 1270
},
{
"epoch": 0.2390326209223847,
"grad_norm": 0.05527610802137319,
"learning_rate": 0.00028270240559153634,
"loss": 0.8481,
"step": 1275
},
{
"epoch": 0.2399700037495313,
"grad_norm": 0.0506803378722897,
"learning_rate": 0.0002824728531622388,
"loss": 0.8354,
"step": 1280
},
{
"epoch": 0.24090738657667793,
"grad_norm": 0.04602503202904283,
"learning_rate": 0.00028224188205511154,
"loss": 0.8169,
"step": 1285
},
{
"epoch": 0.24184476940382452,
"grad_norm": 0.047777320498584894,
"learning_rate": 0.0002820094947436698,
"loss": 0.8143,
"step": 1290
},
{
"epoch": 0.24278215223097113,
"grad_norm": 0.041292637408164454,
"learning_rate": 0.0002817756937165947,
"loss": 0.8429,
"step": 1295
},
{
"epoch": 0.24371953505811775,
"grad_norm": 0.04301715565733742,
"learning_rate": 0.00028154048147770763,
"loss": 0.8225,
"step": 1300
},
{
"epoch": 0.24465691788526434,
"grad_norm": 0.046059693751385175,
"learning_rate": 0.0002813038605459426,
"loss": 0.8459,
"step": 1305
},
{
"epoch": 0.24559430071241095,
"grad_norm": 0.044196493163681724,
"learning_rate": 0.0002810658334553198,
"loss": 0.8161,
"step": 1310
},
{
"epoch": 0.24653168353955757,
"grad_norm": 0.04577271591342366,
"learning_rate": 0.00028082640275491793,
"loss": 0.8386,
"step": 1315
},
{
"epoch": 0.24746906636670415,
"grad_norm": 0.043540551106033394,
"learning_rate": 0.0002805855710088476,
"loss": 0.8332,
"step": 1320
},
{
"epoch": 0.24840644919385077,
"grad_norm": 0.05495402299781008,
"learning_rate": 0.0002803433407962233,
"loss": 0.7858,
"step": 1325
},
{
"epoch": 0.24934383202099739,
"grad_norm": 0.04718094998030703,
"learning_rate": 0.00028009971471113594,
"loss": 0.8382,
"step": 1330
},
{
"epoch": 0.250281214848144,
"grad_norm": 0.047616730401137404,
"learning_rate": 0.00027985469536262524,
"loss": 0.8454,
"step": 1335
},
{
"epoch": 0.2512185976752906,
"grad_norm": 0.0409389484091512,
"learning_rate": 0.0002796082853746515,
"loss": 0.81,
"step": 1340
},
{
"epoch": 0.2521559805024372,
"grad_norm": 0.04522706585670625,
"learning_rate": 0.00027936048738606785,
"loss": 0.8199,
"step": 1345
},
{
"epoch": 0.2530933633295838,
"grad_norm": 0.04653508675318254,
"learning_rate": 0.0002791113040505915,
"loss": 0.8298,
"step": 1350
},
{
"epoch": 0.25403074615673044,
"grad_norm": 0.04663421779124662,
"learning_rate": 0.0002788607380367759,
"loss": 0.865,
"step": 1355
},
{
"epoch": 0.254968128983877,
"grad_norm": 0.04061192230877169,
"learning_rate": 0.0002786087920279818,
"loss": 0.8188,
"step": 1360
},
{
"epoch": 0.2559055118110236,
"grad_norm": 0.04379262228090009,
"learning_rate": 0.0002783554687223484,
"loss": 0.8412,
"step": 1365
},
{
"epoch": 0.2568428946381702,
"grad_norm": 0.04557822686420308,
"learning_rate": 0.0002781007708327649,
"loss": 0.8349,
"step": 1370
},
{
"epoch": 0.25778027746531684,
"grad_norm": 0.045209475352911276,
"learning_rate": 0.00027784470108684094,
"loss": 0.8554,
"step": 1375
},
{
"epoch": 0.25871766029246346,
"grad_norm": 0.0435631363718186,
"learning_rate": 0.0002775872622268779,
"loss": 0.8127,
"step": 1380
},
{
"epoch": 0.2596550431196101,
"grad_norm": 0.044012008476995,
"learning_rate": 0.0002773284570098391,
"loss": 0.8125,
"step": 1385
},
{
"epoch": 0.26059242594675663,
"grad_norm": 0.04874513976965234,
"learning_rate": 0.0002770682882073206,
"loss": 0.8116,
"step": 1390
},
{
"epoch": 0.26152980877390325,
"grad_norm": 0.04889268142442837,
"learning_rate": 0.00027680675860552106,
"loss": 0.8315,
"step": 1395
},
{
"epoch": 0.26246719160104987,
"grad_norm": 0.05581866633201944,
"learning_rate": 0.0002765438710052125,
"loss": 0.8279,
"step": 1400
},
{
"epoch": 0.2634045744281965,
"grad_norm": 0.04602436853592128,
"learning_rate": 0.0002762796282217099,
"loss": 0.821,
"step": 1405
},
{
"epoch": 0.2643419572553431,
"grad_norm": 0.04678707820307465,
"learning_rate": 0.0002760140330848412,
"loss": 0.8466,
"step": 1410
},
{
"epoch": 0.2652793400824897,
"grad_norm": 0.04516389038252434,
"learning_rate": 0.000275747088438917,
"loss": 0.7947,
"step": 1415
},
{
"epoch": 0.26621672290963627,
"grad_norm": 0.04755311816167163,
"learning_rate": 0.00027547879714269995,
"loss": 0.8314,
"step": 1420
},
{
"epoch": 0.2671541057367829,
"grad_norm": 0.044502973033447774,
"learning_rate": 0.0002752091620693742,
"loss": 0.8227,
"step": 1425
},
{
"epoch": 0.2680914885639295,
"grad_norm": 0.0441159145083971,
"learning_rate": 0.00027493818610651487,
"loss": 0.8322,
"step": 1430
},
{
"epoch": 0.2690288713910761,
"grad_norm": 0.04046403847734057,
"learning_rate": 0.0002746658721560568,
"loss": 0.8287,
"step": 1435
},
{
"epoch": 0.26996625421822273,
"grad_norm": 0.04518695193113121,
"learning_rate": 0.0002743922231342636,
"loss": 0.8063,
"step": 1440
},
{
"epoch": 0.27090363704536935,
"grad_norm": 0.04299554365501998,
"learning_rate": 0.00027411724197169647,
"loss": 0.8418,
"step": 1445
},
{
"epoch": 0.2718410198725159,
"grad_norm": 0.04235528853329544,
"learning_rate": 0.0002738409316131827,
"loss": 0.8506,
"step": 1450
},
{
"epoch": 0.2727784026996625,
"grad_norm": 0.04133212847113579,
"learning_rate": 0.0002735632950177843,
"loss": 0.8095,
"step": 1455
},
{
"epoch": 0.27371578552680914,
"grad_norm": 0.044217737534462995,
"learning_rate": 0.00027328433515876613,
"loss": 0.8222,
"step": 1460
},
{
"epoch": 0.27465316835395576,
"grad_norm": 0.04811397623157419,
"learning_rate": 0.0002730040550235642,
"loss": 0.8223,
"step": 1465
},
{
"epoch": 0.2755905511811024,
"grad_norm": 0.043283154172989766,
"learning_rate": 0.0002727224576137535,
"loss": 0.8244,
"step": 1470
},
{
"epoch": 0.276527934008249,
"grad_norm": 0.04254147668888092,
"learning_rate": 0.0002724395459450161,
"loss": 0.8158,
"step": 1475
},
{
"epoch": 0.27746531683539555,
"grad_norm": 0.0437512052356442,
"learning_rate": 0.0002721553230471087,
"loss": 0.8449,
"step": 1480
},
{
"epoch": 0.27840269966254216,
"grad_norm": 0.04187590056923411,
"learning_rate": 0.0002718697919638302,
"loss": 0.7986,
"step": 1485
},
{
"epoch": 0.2793400824896888,
"grad_norm": 0.042437873734035854,
"learning_rate": 0.0002715829557529891,
"loss": 0.8286,
"step": 1490
},
{
"epoch": 0.2802774653168354,
"grad_norm": 0.04430563389754819,
"learning_rate": 0.00027129481748637075,
"loss": 0.841,
"step": 1495
},
{
"epoch": 0.281214848143982,
"grad_norm": 0.045758411532611065,
"learning_rate": 0.00027100538024970444,
"loss": 0.8285,
"step": 1500
},
{
"epoch": 0.2821522309711286,
"grad_norm": 0.04423481610722165,
"learning_rate": 0.00027071464714263063,
"loss": 0.8168,
"step": 1505
},
{
"epoch": 0.28308961379827524,
"grad_norm": 0.04571606782924112,
"learning_rate": 0.00027042262127866716,
"loss": 0.8249,
"step": 1510
},
{
"epoch": 0.2840269966254218,
"grad_norm": 0.04732141795830983,
"learning_rate": 0.00027012930578517645,
"loss": 0.8387,
"step": 1515
},
{
"epoch": 0.2849643794525684,
"grad_norm": 0.046305063021735725,
"learning_rate": 0.00026983470380333185,
"loss": 0.8106,
"step": 1520
},
{
"epoch": 0.28590176227971503,
"grad_norm": 0.04195897212582909,
"learning_rate": 0.0002695388184880839,
"loss": 0.8261,
"step": 1525
},
{
"epoch": 0.28683914510686165,
"grad_norm": 0.04456321258007771,
"learning_rate": 0.0002692416530081265,
"loss": 0.8215,
"step": 1530
},
{
"epoch": 0.28777652793400826,
"grad_norm": 0.04879364590021316,
"learning_rate": 0.0002689432105458633,
"loss": 0.8135,
"step": 1535
},
{
"epoch": 0.2887139107611549,
"grad_norm": 0.04789940824296893,
"learning_rate": 0.00026864349429737326,
"loss": 0.8368,
"step": 1540
},
{
"epoch": 0.28965129358830144,
"grad_norm": 0.04616782271102101,
"learning_rate": 0.00026834250747237665,
"loss": 0.8269,
"step": 1545
},
{
"epoch": 0.29058867641544806,
"grad_norm": 0.04502530076723608,
"learning_rate": 0.0002680402532942006,
"loss": 0.83,
"step": 1550
},
{
"epoch": 0.29152605924259467,
"grad_norm": 0.04002183200170624,
"learning_rate": 0.00026773673499974436,
"loss": 0.8053,
"step": 1555
},
{
"epoch": 0.2924634420697413,
"grad_norm": 0.04425864811179813,
"learning_rate": 0.00026743195583944524,
"loss": 0.8354,
"step": 1560
},
{
"epoch": 0.2934008248968879,
"grad_norm": 0.046865997902279015,
"learning_rate": 0.000267125919077243,
"loss": 0.8263,
"step": 1565
},
{
"epoch": 0.2943382077240345,
"grad_norm": 0.04361053554450397,
"learning_rate": 0.00026681862799054557,
"loss": 0.8007,
"step": 1570
},
{
"epoch": 0.2952755905511811,
"grad_norm": 0.048458258290373916,
"learning_rate": 0.0002665100858701937,
"loss": 0.8237,
"step": 1575
},
{
"epoch": 0.2962129733783277,
"grad_norm": 0.047990047586811434,
"learning_rate": 0.0002662002960204254,
"loss": 0.8162,
"step": 1580
},
{
"epoch": 0.2971503562054743,
"grad_norm": 0.04409542819477004,
"learning_rate": 0.0002658892617588413,
"loss": 0.8433,
"step": 1585
},
{
"epoch": 0.2980877390326209,
"grad_norm": 0.04731184190347321,
"learning_rate": 0.00026557698641636835,
"loss": 0.8133,
"step": 1590
},
{
"epoch": 0.29902512185976754,
"grad_norm": 0.042762397685975305,
"learning_rate": 0.0002652634733372246,
"loss": 0.834,
"step": 1595
},
{
"epoch": 0.29996250468691416,
"grad_norm": 0.04579010165352788,
"learning_rate": 0.0002649487258788833,
"loss": 0.8214,
"step": 1600
},
{
"epoch": 0.3008998875140607,
"grad_norm": 0.04144592090502628,
"learning_rate": 0.0002646327474120368,
"loss": 0.8207,
"step": 1605
},
{
"epoch": 0.30183727034120733,
"grad_norm": 0.04770299028768292,
"learning_rate": 0.00026431554132056063,
"loss": 0.8258,
"step": 1610
},
{
"epoch": 0.30277465316835395,
"grad_norm": 0.044453071872775064,
"learning_rate": 0.00026399711100147724,
"loss": 0.8085,
"step": 1615
},
{
"epoch": 0.30371203599550056,
"grad_norm": 0.05130385596231612,
"learning_rate": 0.0002636774598649195,
"loss": 0.8287,
"step": 1620
},
{
"epoch": 0.3046494188226472,
"grad_norm": 0.053015815539756074,
"learning_rate": 0.00026335659133409423,
"loss": 0.8063,
"step": 1625
},
{
"epoch": 0.3055868016497938,
"grad_norm": 0.04286401355249045,
"learning_rate": 0.00026303450884524566,
"loss": 0.8084,
"step": 1630
},
{
"epoch": 0.3065241844769404,
"grad_norm": 0.0383196394419625,
"learning_rate": 0.0002627112158476185,
"loss": 0.8001,
"step": 1635
},
{
"epoch": 0.30746156730408697,
"grad_norm": 0.042481350200920615,
"learning_rate": 0.00026238671580342096,
"loss": 0.8342,
"step": 1640
},
{
"epoch": 0.3083989501312336,
"grad_norm": 0.0437271017287051,
"learning_rate": 0.0002620610121877879,
"loss": 0.8301,
"step": 1645
},
{
"epoch": 0.3093363329583802,
"grad_norm": 0.042066811375379806,
"learning_rate": 0.0002617341084887433,
"loss": 0.8183,
"step": 1650
},
{
"epoch": 0.3102737157855268,
"grad_norm": 0.0434545427613146,
"learning_rate": 0.00026140600820716314,
"loss": 0.8144,
"step": 1655
},
{
"epoch": 0.31121109861267343,
"grad_norm": 0.038937665823880176,
"learning_rate": 0.00026107671485673794,
"loss": 0.817,
"step": 1660
},
{
"epoch": 0.31214848143982005,
"grad_norm": 0.044655599527732244,
"learning_rate": 0.0002607462319639348,
"loss": 0.8344,
"step": 1665
},
{
"epoch": 0.3130858642669666,
"grad_norm": 0.03965599804534495,
"learning_rate": 0.00026041456306796014,
"loss": 0.8083,
"step": 1670
},
{
"epoch": 0.3140232470941132,
"grad_norm": 0.039069129149835194,
"learning_rate": 0.00026008171172072126,
"loss": 0.8196,
"step": 1675
},
{
"epoch": 0.31496062992125984,
"grad_norm": 0.040202855570058024,
"learning_rate": 0.0002597476814867887,
"loss": 0.8205,
"step": 1680
},
{
"epoch": 0.31589801274840645,
"grad_norm": 0.04003968304410291,
"learning_rate": 0.0002594124759433579,
"loss": 0.8108,
"step": 1685
},
{
"epoch": 0.31683539557555307,
"grad_norm": 0.047433872652346235,
"learning_rate": 0.000259076098680211,
"loss": 0.8039,
"step": 1690
},
{
"epoch": 0.3177727784026997,
"grad_norm": 0.04141242579423705,
"learning_rate": 0.0002587385532996782,
"loss": 0.8259,
"step": 1695
},
{
"epoch": 0.31871016122984624,
"grad_norm": 0.054121951067220224,
"learning_rate": 0.0002583998434165993,
"loss": 0.8246,
"step": 1700
},
{
"epoch": 0.31964754405699286,
"grad_norm": 0.04197913764657862,
"learning_rate": 0.00025805997265828507,
"loss": 0.8463,
"step": 1705
},
{
"epoch": 0.3205849268841395,
"grad_norm": 0.0426917661733632,
"learning_rate": 0.0002577189446644783,
"loss": 0.8183,
"step": 1710
},
{
"epoch": 0.3215223097112861,
"grad_norm": 0.04179736226931492,
"learning_rate": 0.00025737676308731477,
"loss": 0.7976,
"step": 1715
},
{
"epoch": 0.3224596925384327,
"grad_norm": 0.04799629600304747,
"learning_rate": 0.0002570334315912844,
"loss": 0.8289,
"step": 1720
},
{
"epoch": 0.3233970753655793,
"grad_norm": 0.042393149930441694,
"learning_rate": 0.0002566889538531915,
"loss": 0.8112,
"step": 1725
},
{
"epoch": 0.3243344581927259,
"grad_norm": 0.04255536591889664,
"learning_rate": 0.000256343333562116,
"loss": 0.8187,
"step": 1730
},
{
"epoch": 0.3252718410198725,
"grad_norm": 0.044062962674437295,
"learning_rate": 0.00025599657441937354,
"loss": 0.8018,
"step": 1735
},
{
"epoch": 0.3262092238470191,
"grad_norm": 0.043474283263771174,
"learning_rate": 0.00025564868013847595,
"loss": 0.8306,
"step": 1740
},
{
"epoch": 0.32714660667416573,
"grad_norm": 0.0422049730670292,
"learning_rate": 0.0002552996544450914,
"loss": 0.8047,
"step": 1745
},
{
"epoch": 0.32808398950131235,
"grad_norm": 0.04744673000933406,
"learning_rate": 0.0002549495010770048,
"loss": 0.8422,
"step": 1750
},
{
"epoch": 0.32902137232845896,
"grad_norm": 0.04429260845252424,
"learning_rate": 0.0002545982237840773,
"loss": 0.8191,
"step": 1755
},
{
"epoch": 0.3299587551556056,
"grad_norm": 0.04232121926909998,
"learning_rate": 0.0002542458263282066,
"loss": 0.7905,
"step": 1760
},
{
"epoch": 0.33089613798275214,
"grad_norm": 0.03903941148795766,
"learning_rate": 0.00025389231248328624,
"loss": 0.8047,
"step": 1765
},
{
"epoch": 0.33183352080989875,
"grad_norm": 0.043127224537608456,
"learning_rate": 0.00025353768603516555,
"loss": 0.8202,
"step": 1770
},
{
"epoch": 0.33277090363704537,
"grad_norm": 0.04603755356895304,
"learning_rate": 0.0002531819507816089,
"loss": 0.8474,
"step": 1775
},
{
"epoch": 0.333708286464192,
"grad_norm": 0.04421659565539193,
"learning_rate": 0.00025282511053225493,
"loss": 0.8258,
"step": 1780
},
{
"epoch": 0.3346456692913386,
"grad_norm": 0.04354315354304148,
"learning_rate": 0.0002524671691085762,
"loss": 0.7872,
"step": 1785
},
{
"epoch": 0.3355830521184852,
"grad_norm": 0.04513774703149973,
"learning_rate": 0.0002521081303438377,
"loss": 0.7985,
"step": 1790
},
{
"epoch": 0.3365204349456318,
"grad_norm": 0.04656688027583207,
"learning_rate": 0.00025174799808305606,
"loss": 0.808,
"step": 1795
},
{
"epoch": 0.3374578177727784,
"grad_norm": 0.04760744964038916,
"learning_rate": 0.0002513867761829587,
"loss": 0.793,
"step": 1800
},
{
"epoch": 0.338395200599925,
"grad_norm": 0.04121747602293146,
"learning_rate": 0.0002510244685119418,
"loss": 0.8293,
"step": 1805
},
{
"epoch": 0.3393325834270716,
"grad_norm": 0.04112754367048587,
"learning_rate": 0.00025066107895002946,
"loss": 0.831,
"step": 1810
},
{
"epoch": 0.34026996625421824,
"grad_norm": 0.03926106896606557,
"learning_rate": 0.0002502966113888319,
"loss": 0.8072,
"step": 1815
},
{
"epoch": 0.34120734908136485,
"grad_norm": 0.04562546212037259,
"learning_rate": 0.000249931069731504,
"loss": 0.788,
"step": 1820
},
{
"epoch": 0.3421447319085114,
"grad_norm": 0.0406588900726224,
"learning_rate": 0.0002495644578927032,
"loss": 0.8184,
"step": 1825
},
{
"epoch": 0.34308211473565803,
"grad_norm": 0.04003290325962031,
"learning_rate": 0.00024919677979854776,
"loss": 0.8272,
"step": 1830
},
{
"epoch": 0.34401949756280464,
"grad_norm": 0.04186901209736264,
"learning_rate": 0.00024882803938657466,
"loss": 0.7956,
"step": 1835
},
{
"epoch": 0.34495688038995126,
"grad_norm": 0.041398093060463485,
"learning_rate": 0.00024845824060569743,
"loss": 0.8114,
"step": 1840
},
{
"epoch": 0.3458942632170979,
"grad_norm": 0.04109679086847299,
"learning_rate": 0.000248087387416164,
"loss": 0.807,
"step": 1845
},
{
"epoch": 0.3468316460442445,
"grad_norm": 0.042039328634813876,
"learning_rate": 0.000247715483789514,
"loss": 0.8306,
"step": 1850
},
{
"epoch": 0.34776902887139105,
"grad_norm": 0.0394540126640408,
"learning_rate": 0.0002473425337085366,
"loss": 0.7966,
"step": 1855
},
{
"epoch": 0.34870641169853767,
"grad_norm": 0.04432229876319661,
"learning_rate": 0.0002469685411672275,
"loss": 0.811,
"step": 1860
},
{
"epoch": 0.3496437945256843,
"grad_norm": 0.04227321189035649,
"learning_rate": 0.0002465935101707463,
"loss": 0.8248,
"step": 1865
},
{
"epoch": 0.3505811773528309,
"grad_norm": 0.039191702176161206,
"learning_rate": 0.00024621744473537365,
"loss": 0.8205,
"step": 1870
},
{
"epoch": 0.3515185601799775,
"grad_norm": 0.04634923450670881,
"learning_rate": 0.00024584034888846835,
"loss": 0.7763,
"step": 1875
},
{
"epoch": 0.35245594300712413,
"grad_norm": 0.04501112952096222,
"learning_rate": 0.0002454622266684239,
"loss": 0.8258,
"step": 1880
},
{
"epoch": 0.3533933258342707,
"grad_norm": 0.04422758808317238,
"learning_rate": 0.0002450830821246255,
"loss": 0.8106,
"step": 1885
},
{
"epoch": 0.3543307086614173,
"grad_norm": 0.04421016805518408,
"learning_rate": 0.00024470291931740667,
"loss": 0.7815,
"step": 1890
},
{
"epoch": 0.3552680914885639,
"grad_norm": 0.04319380391113109,
"learning_rate": 0.0002443217423180055,
"loss": 0.7973,
"step": 1895
},
{
"epoch": 0.35620547431571054,
"grad_norm": 0.040535892316044465,
"learning_rate": 0.00024393955520852158,
"loss": 0.8231,
"step": 1900
},
{
"epoch": 0.35714285714285715,
"grad_norm": 0.050182245944215505,
"learning_rate": 0.00024355636208187175,
"loss": 0.809,
"step": 1905
},
{
"epoch": 0.35808023997000377,
"grad_norm": 0.05056635484821874,
"learning_rate": 0.00024317216704174653,
"loss": 0.7863,
"step": 1910
},
{
"epoch": 0.3590176227971504,
"grad_norm": 0.04022954249460216,
"learning_rate": 0.00024278697420256615,
"loss": 0.8144,
"step": 1915
},
{
"epoch": 0.35995500562429694,
"grad_norm": 0.04323635830441304,
"learning_rate": 0.00024240078768943647,
"loss": 0.8052,
"step": 1920
},
{
"epoch": 0.36089238845144356,
"grad_norm": 0.05048502253231987,
"learning_rate": 0.00024201361163810476,
"loss": 0.8292,
"step": 1925
},
{
"epoch": 0.3618297712785902,
"grad_norm": 0.045094345947545966,
"learning_rate": 0.00024162545019491545,
"loss": 0.8263,
"step": 1930
},
{
"epoch": 0.3627671541057368,
"grad_norm": 0.04751777775798578,
"learning_rate": 0.0002412363075167658,
"loss": 0.813,
"step": 1935
},
{
"epoch": 0.3637045369328834,
"grad_norm": 0.03970350230348181,
"learning_rate": 0.0002408461877710613,
"loss": 0.7874,
"step": 1940
},
{
"epoch": 0.36464191976003,
"grad_norm": 0.043021517102531895,
"learning_rate": 0.00024045509513567092,
"loss": 0.7835,
"step": 1945
},
{
"epoch": 0.3655793025871766,
"grad_norm": 0.04145029708505387,
"learning_rate": 0.0002400630337988826,
"loss": 0.7976,
"step": 1950
},
{
"epoch": 0.3665166854143232,
"grad_norm": 0.04060705381251834,
"learning_rate": 0.0002396700079593583,
"loss": 0.784,
"step": 1955
},
{
"epoch": 0.3674540682414698,
"grad_norm": 0.04105772844447629,
"learning_rate": 0.00023927602182608902,
"loss": 0.7982,
"step": 1960
},
{
"epoch": 0.3683914510686164,
"grad_norm": 0.04405339270275701,
"learning_rate": 0.00023888107961834968,
"loss": 0.8367,
"step": 1965
},
{
"epoch": 0.36932883389576304,
"grad_norm": 0.04229707897710055,
"learning_rate": 0.00023848518556565405,
"loss": 0.8147,
"step": 1970
},
{
"epoch": 0.37026621672290966,
"grad_norm": 0.04101868201617462,
"learning_rate": 0.00023808834390770937,
"loss": 0.7984,
"step": 1975
},
{
"epoch": 0.3712035995500562,
"grad_norm": 0.0452498903694969,
"learning_rate": 0.00023769055889437103,
"loss": 0.8064,
"step": 1980
},
{
"epoch": 0.37214098237720283,
"grad_norm": 0.040047239815103164,
"learning_rate": 0.0002372918347855969,
"loss": 0.7737,
"step": 1985
},
{
"epoch": 0.37307836520434945,
"grad_norm": 0.03830098106296232,
"learning_rate": 0.0002368921758514018,
"loss": 0.7735,
"step": 1990
},
{
"epoch": 0.37401574803149606,
"grad_norm": 0.0466704182451149,
"learning_rate": 0.00023649158637181191,
"loss": 0.7913,
"step": 1995
},
{
"epoch": 0.3749531308586427,
"grad_norm": 0.045697413605673594,
"learning_rate": 0.00023609007063681874,
"loss": 0.8083,
"step": 2000
},
{
"epoch": 0.3758905136857893,
"grad_norm": 0.046012146633052885,
"learning_rate": 0.0002356876329463332,
"loss": 0.7986,
"step": 2005
},
{
"epoch": 0.37682789651293586,
"grad_norm": 0.042087372585577834,
"learning_rate": 0.0002352842776101396,
"loss": 0.789,
"step": 2010
},
{
"epoch": 0.37776527934008247,
"grad_norm": 0.04351288912284283,
"learning_rate": 0.00023488000894784954,
"loss": 0.8066,
"step": 2015
},
{
"epoch": 0.3787026621672291,
"grad_norm": 0.04264731715281628,
"learning_rate": 0.0002344748312888557,
"loss": 0.8242,
"step": 2020
},
{
"epoch": 0.3796400449943757,
"grad_norm": 0.04004005288826855,
"learning_rate": 0.00023406874897228527,
"loss": 0.8134,
"step": 2025
},
{
"epoch": 0.3805774278215223,
"grad_norm": 0.03990638844779895,
"learning_rate": 0.00023366176634695353,
"loss": 0.8341,
"step": 2030
},
{
"epoch": 0.38151481064866893,
"grad_norm": 0.04606026511862361,
"learning_rate": 0.00023325388777131748,
"loss": 0.8001,
"step": 2035
},
{
"epoch": 0.38245219347581555,
"grad_norm": 0.04391729186424558,
"learning_rate": 0.000232845117613429,
"loss": 0.8195,
"step": 2040
},
{
"epoch": 0.3833895763029621,
"grad_norm": 0.041518162558628426,
"learning_rate": 0.00023243546025088799,
"loss": 0.7802,
"step": 2045
},
{
"epoch": 0.3843269591301087,
"grad_norm": 0.04366132632044734,
"learning_rate": 0.00023202492007079584,
"loss": 0.7828,
"step": 2050
},
{
"epoch": 0.38526434195725534,
"grad_norm": 0.04383475380804256,
"learning_rate": 0.00023161350146970794,
"loss": 0.7876,
"step": 2055
},
{
"epoch": 0.38620172478440196,
"grad_norm": 0.036393360894576204,
"learning_rate": 0.00023120120885358698,
"loss": 0.7975,
"step": 2060
},
{
"epoch": 0.38713910761154857,
"grad_norm": 0.04084955682840953,
"learning_rate": 0.00023078804663775572,
"loss": 0.7786,
"step": 2065
},
{
"epoch": 0.3880764904386952,
"grad_norm": 0.03911103036688829,
"learning_rate": 0.00023037401924684946,
"loss": 0.8026,
"step": 2070
},
{
"epoch": 0.38901387326584175,
"grad_norm": 0.041012148951149276,
"learning_rate": 0.000229959131114769,
"loss": 0.7885,
"step": 2075
},
{
"epoch": 0.38995125609298836,
"grad_norm": 0.03899340776511314,
"learning_rate": 0.00022954338668463296,
"loss": 0.7813,
"step": 2080
},
{
"epoch": 0.390888638920135,
"grad_norm": 0.04177139426462313,
"learning_rate": 0.00022912679040873005,
"loss": 0.7433,
"step": 2085
},
{
"epoch": 0.3918260217472816,
"grad_norm": 0.041993929485497614,
"learning_rate": 0.00022870934674847177,
"loss": 0.8079,
"step": 2090
},
{
"epoch": 0.3927634045744282,
"grad_norm": 0.04106413868995631,
"learning_rate": 0.00022829106017434434,
"loss": 0.7872,
"step": 2095
},
{
"epoch": 0.3937007874015748,
"grad_norm": 0.047413157951248786,
"learning_rate": 0.00022787193516586091,
"loss": 0.796,
"step": 2100
},
{
"epoch": 0.3946381702287214,
"grad_norm": 0.04503978112797596,
"learning_rate": 0.00022745197621151363,
"loss": 0.8352,
"step": 2105
},
{
"epoch": 0.395575553055868,
"grad_norm": 0.04115650674340414,
"learning_rate": 0.0002270311878087255,
"loss": 0.816,
"step": 2110
},
{
"epoch": 0.3965129358830146,
"grad_norm": 0.041160476408568784,
"learning_rate": 0.00022660957446380225,
"loss": 0.7982,
"step": 2115
},
{
"epoch": 0.39745031871016123,
"grad_norm": 0.04408506165618202,
"learning_rate": 0.00022618714069188404,
"loss": 0.8209,
"step": 2120
},
{
"epoch": 0.39838770153730785,
"grad_norm": 0.04001800900141434,
"learning_rate": 0.00022576389101689725,
"loss": 0.7771,
"step": 2125
},
{
"epoch": 0.39932508436445446,
"grad_norm": 0.04320785141278407,
"learning_rate": 0.00022533982997150585,
"loss": 0.7677,
"step": 2130
},
{
"epoch": 0.400262467191601,
"grad_norm": 0.039431554219398125,
"learning_rate": 0.00022491496209706293,
"loss": 0.7715,
"step": 2135
},
{
"epoch": 0.40119985001874764,
"grad_norm": 0.04041447511709435,
"learning_rate": 0.0002244892919435621,
"loss": 0.7961,
"step": 2140
},
{
"epoch": 0.40213723284589425,
"grad_norm": 0.04233570336630376,
"learning_rate": 0.00022406282406958874,
"loss": 0.7932,
"step": 2145
},
{
"epoch": 0.40307461567304087,
"grad_norm": 0.040223931466695007,
"learning_rate": 0.00022363556304227111,
"loss": 0.7972,
"step": 2150
},
{
"epoch": 0.4040119985001875,
"grad_norm": 0.03920199535556696,
"learning_rate": 0.0002232075134372316,
"loss": 0.7912,
"step": 2155
},
{
"epoch": 0.4049493813273341,
"grad_norm": 0.04045705766440815,
"learning_rate": 0.00022277867983853754,
"loss": 0.772,
"step": 2160
},
{
"epoch": 0.4058867641544807,
"grad_norm": 0.038172006672106074,
"learning_rate": 0.00022234906683865234,
"loss": 0.7994,
"step": 2165
},
{
"epoch": 0.4068241469816273,
"grad_norm": 0.042094914465141645,
"learning_rate": 0.00022191867903838597,
"loss": 0.7908,
"step": 2170
},
{
"epoch": 0.4077615298087739,
"grad_norm": 0.041299074122275056,
"learning_rate": 0.00022148752104684608,
"loss": 0.791,
"step": 2175
},
{
"epoch": 0.4086989126359205,
"grad_norm": 0.03672565143578591,
"learning_rate": 0.00022105559748138834,
"loss": 0.7879,
"step": 2180
},
{
"epoch": 0.4096362954630671,
"grad_norm": 0.03978313771126583,
"learning_rate": 0.00022062291296756715,
"loss": 0.8095,
"step": 2185
},
{
"epoch": 0.41057367829021374,
"grad_norm": 0.04001569067395201,
"learning_rate": 0.000220189472139086,
"loss": 0.7826,
"step": 2190
},
{
"epoch": 0.41151106111736035,
"grad_norm": 0.043631782457239365,
"learning_rate": 0.00021975527963774796,
"loss": 0.7927,
"step": 2195
},
{
"epoch": 0.4124484439445069,
"grad_norm": 0.04073514355199719,
"learning_rate": 0.00021932034011340587,
"loss": 0.7939,
"step": 2200
},
{
"epoch": 0.41338582677165353,
"grad_norm": 0.044704633805377016,
"learning_rate": 0.00021888465822391269,
"loss": 0.795,
"step": 2205
},
{
"epoch": 0.41432320959880015,
"grad_norm": 0.04223574180256086,
"learning_rate": 0.00021844823863507136,
"loss": 0.7697,
"step": 2210
},
{
"epoch": 0.41526059242594676,
"grad_norm": 0.04033678944300133,
"learning_rate": 0.00021801108602058507,
"loss": 0.7942,
"step": 2215
},
{
"epoch": 0.4161979752530934,
"grad_norm": 0.040615950774112376,
"learning_rate": 0.00021757320506200713,
"loss": 0.7976,
"step": 2220
},
{
"epoch": 0.41713535808024,
"grad_norm": 0.03966904553021298,
"learning_rate": 0.00021713460044869078,
"loss": 0.7356,
"step": 2225
},
{
"epoch": 0.41807274090738655,
"grad_norm": 0.049048512180610696,
"learning_rate": 0.0002166952768777391,
"loss": 0.788,
"step": 2230
},
{
"epoch": 0.41901012373453317,
"grad_norm": 0.04363519676021418,
"learning_rate": 0.00021625523905395458,
"loss": 0.7778,
"step": 2235
},
{
"epoch": 0.4199475065616798,
"grad_norm": 0.040788196081375995,
"learning_rate": 0.00021581449168978878,
"loss": 0.7845,
"step": 2240
},
{
"epoch": 0.4208848893888264,
"grad_norm": 0.04388975232146075,
"learning_rate": 0.00021537303950529185,
"loss": 0.7912,
"step": 2245
},
{
"epoch": 0.421822272215973,
"grad_norm": 0.039487905759282925,
"learning_rate": 0.0002149308872280621,
"loss": 0.7904,
"step": 2250
},
{
"epoch": 0.42275965504311963,
"grad_norm": 0.03892803525416569,
"learning_rate": 0.0002144880395931951,
"loss": 0.7832,
"step": 2255
},
{
"epoch": 0.4236970378702662,
"grad_norm": 0.048355006016535,
"learning_rate": 0.0002140445013432333,
"loss": 0.8233,
"step": 2260
},
{
"epoch": 0.4246344206974128,
"grad_norm": 0.04226596396802164,
"learning_rate": 0.00021360027722811505,
"loss": 0.7986,
"step": 2265
},
{
"epoch": 0.4255718035245594,
"grad_norm": 0.03777343997339362,
"learning_rate": 0.00021315537200512362,
"loss": 0.7739,
"step": 2270
},
{
"epoch": 0.42650918635170604,
"grad_norm": 0.0447778590274291,
"learning_rate": 0.00021270979043883664,
"loss": 0.8097,
"step": 2275
},
{
"epoch": 0.42744656917885265,
"grad_norm": 0.04168554679696771,
"learning_rate": 0.00021226353730107467,
"loss": 0.7835,
"step": 2280
},
{
"epoch": 0.42838395200599927,
"grad_norm": 0.04011938574259242,
"learning_rate": 0.00021181661737085028,
"loss": 0.8223,
"step": 2285
},
{
"epoch": 0.4293213348331459,
"grad_norm": 0.03778469888602811,
"learning_rate": 0.00021136903543431685,
"loss": 0.7739,
"step": 2290
},
{
"epoch": 0.43025871766029244,
"grad_norm": 0.03715090423495881,
"learning_rate": 0.0002109207962847174,
"loss": 0.8144,
"step": 2295
},
{
"epoch": 0.43119610048743906,
"grad_norm": 0.03962090097485277,
"learning_rate": 0.00021047190472233305,
"loss": 0.7811,
"step": 2300
},
{
"epoch": 0.4321334833145857,
"grad_norm": 0.039092040132198524,
"learning_rate": 0.00021002236555443183,
"loss": 0.7909,
"step": 2305
},
{
"epoch": 0.4330708661417323,
"grad_norm": 0.03883145505567921,
"learning_rate": 0.00020957218359521706,
"loss": 0.8176,
"step": 2310
},
{
"epoch": 0.4340082489688789,
"grad_norm": 0.03985259948134514,
"learning_rate": 0.0002091213636657759,
"loss": 0.7869,
"step": 2315
},
{
"epoch": 0.4349456317960255,
"grad_norm": 0.0405689266411763,
"learning_rate": 0.0002086699105940275,
"loss": 0.8039,
"step": 2320
},
{
"epoch": 0.4358830146231721,
"grad_norm": 0.04016614154253833,
"learning_rate": 0.00020821782921467166,
"loss": 0.7911,
"step": 2325
},
{
"epoch": 0.4368203974503187,
"grad_norm": 0.03975323854656061,
"learning_rate": 0.0002077651243691367,
"loss": 0.7833,
"step": 2330
},
{
"epoch": 0.4377577802774653,
"grad_norm": 0.041105027708813344,
"learning_rate": 0.00020731180090552783,
"loss": 0.7675,
"step": 2335
},
{
"epoch": 0.43869516310461193,
"grad_norm": 0.03872126186366071,
"learning_rate": 0.00020685786367857518,
"loss": 0.7959,
"step": 2340
},
{
"epoch": 0.43963254593175854,
"grad_norm": 0.041950004467004046,
"learning_rate": 0.0002064033175495817,
"loss": 0.7642,
"step": 2345
},
{
"epoch": 0.44056992875890516,
"grad_norm": 0.04308753428596666,
"learning_rate": 0.00020594816738637133,
"loss": 0.7828,
"step": 2350
},
{
"epoch": 0.4415073115860517,
"grad_norm": 0.04067335511731369,
"learning_rate": 0.00020549241806323658,
"loss": 0.7731,
"step": 2355
},
{
"epoch": 0.44244469441319834,
"grad_norm": 0.038488821747570894,
"learning_rate": 0.00020503607446088661,
"loss": 0.7783,
"step": 2360
},
{
"epoch": 0.44338207724034495,
"grad_norm": 0.0405566911757408,
"learning_rate": 0.00020457914146639473,
"loss": 0.7913,
"step": 2365
},
{
"epoch": 0.44431946006749157,
"grad_norm": 0.041242394949931045,
"learning_rate": 0.00020412162397314624,
"loss": 0.7971,
"step": 2370
},
{
"epoch": 0.4452568428946382,
"grad_norm": 0.039550878105764575,
"learning_rate": 0.00020366352688078597,
"loss": 0.7941,
"step": 2375
},
{
"epoch": 0.4461942257217848,
"grad_norm": 0.04176542654646332,
"learning_rate": 0.00020320485509516564,
"loss": 0.7796,
"step": 2380
},
{
"epoch": 0.44713160854893136,
"grad_norm": 0.04091658096529671,
"learning_rate": 0.0002027456135282917,
"loss": 0.7656,
"step": 2385
},
{
"epoch": 0.448068991376078,
"grad_norm": 0.04204041507697926,
"learning_rate": 0.00020228580709827227,
"loss": 0.7842,
"step": 2390
},
{
"epoch": 0.4490063742032246,
"grad_norm": 0.043581111984004314,
"learning_rate": 0.0002018254407292649,
"loss": 0.8044,
"step": 2395
},
{
"epoch": 0.4499437570303712,
"grad_norm": 0.03967757526818896,
"learning_rate": 0.00020136451935142349,
"loss": 0.7807,
"step": 2400
},
{
"epoch": 0.4508811398575178,
"grad_norm": 0.04119079014264244,
"learning_rate": 0.00020090304790084572,
"loss": 0.7949,
"step": 2405
},
{
"epoch": 0.45181852268466444,
"grad_norm": 0.04744494144786433,
"learning_rate": 0.00020044103131952007,
"loss": 0.7886,
"step": 2410
},
{
"epoch": 0.452755905511811,
"grad_norm": 0.03865847999571606,
"learning_rate": 0.000199978474555273,
"loss": 0.7824,
"step": 2415
},
{
"epoch": 0.4536932883389576,
"grad_norm": 0.04257054661223989,
"learning_rate": 0.0001995153825617157,
"loss": 0.7958,
"step": 2420
},
{
"epoch": 0.4546306711661042,
"grad_norm": 0.03723070056879195,
"learning_rate": 0.0001990517602981915,
"loss": 0.7743,
"step": 2425
},
{
"epoch": 0.45556805399325084,
"grad_norm": 0.04206537348222414,
"learning_rate": 0.0001985876127297224,
"loss": 0.7818,
"step": 2430
},
{
"epoch": 0.45650543682039746,
"grad_norm": 0.046605846594271055,
"learning_rate": 0.00019812294482695586,
"loss": 0.7888,
"step": 2435
},
{
"epoch": 0.4574428196475441,
"grad_norm": 0.03927213215049627,
"learning_rate": 0.00019765776156611189,
"loss": 0.7941,
"step": 2440
},
{
"epoch": 0.4583802024746907,
"grad_norm": 0.03773012168883939,
"learning_rate": 0.00019719206792892944,
"loss": 0.779,
"step": 2445
},
{
"epoch": 0.45931758530183725,
"grad_norm": 0.04033866134296589,
"learning_rate": 0.00019672586890261322,
"loss": 0.7548,
"step": 2450
},
{
"epoch": 0.46025496812898387,
"grad_norm": 0.03863797353130748,
"learning_rate": 0.00019625916947978029,
"loss": 0.7519,
"step": 2455
},
{
"epoch": 0.4611923509561305,
"grad_norm": 0.03834068421875446,
"learning_rate": 0.00019579197465840654,
"loss": 0.7524,
"step": 2460
},
{
"epoch": 0.4621297337832771,
"grad_norm": 0.04082115337906864,
"learning_rate": 0.0001953242894417731,
"loss": 0.7748,
"step": 2465
},
{
"epoch": 0.4630671166104237,
"grad_norm": 0.038882691470253396,
"learning_rate": 0.000194856118838413,
"loss": 0.7732,
"step": 2470
},
{
"epoch": 0.46400449943757033,
"grad_norm": 0.04383855537670663,
"learning_rate": 0.0001943874678620572,
"loss": 0.7718,
"step": 2475
},
{
"epoch": 0.4649418822647169,
"grad_norm": 0.04487220174913269,
"learning_rate": 0.0001939183415315812,
"loss": 0.8184,
"step": 2480
},
{
"epoch": 0.4658792650918635,
"grad_norm": 0.04243430887998934,
"learning_rate": 0.00019344874487095106,
"loss": 0.7909,
"step": 2485
},
{
"epoch": 0.4668166479190101,
"grad_norm": 0.042879736936158284,
"learning_rate": 0.00019297868290916973,
"loss": 0.8164,
"step": 2490
},
{
"epoch": 0.46775403074615673,
"grad_norm": 0.04530733011987864,
"learning_rate": 0.00019250816068022326,
"loss": 0.795,
"step": 2495
},
{
"epoch": 0.46869141357330335,
"grad_norm": 0.04578548882806771,
"learning_rate": 0.0001920371832230266,
"loss": 0.7974,
"step": 2500
},
{
"epoch": 0.46962879640044997,
"grad_norm": 0.04042238169087785,
"learning_rate": 0.00019156575558137003,
"loss": 0.8004,
"step": 2505
},
{
"epoch": 0.4705661792275965,
"grad_norm": 0.03887172482698654,
"learning_rate": 0.00019109388280386488,
"loss": 0.7686,
"step": 2510
},
{
"epoch": 0.47150356205474314,
"grad_norm": 0.03584281998116991,
"learning_rate": 0.00019062156994388937,
"loss": 0.7488,
"step": 2515
},
{
"epoch": 0.47244094488188976,
"grad_norm": 0.038784324923480024,
"learning_rate": 0.00019014882205953485,
"loss": 0.7797,
"step": 2520
},
{
"epoch": 0.4733783277090364,
"grad_norm": 0.03811245259868312,
"learning_rate": 0.00018967564421355134,
"loss": 0.7566,
"step": 2525
},
{
"epoch": 0.474315710536183,
"grad_norm": 0.03796594842270583,
"learning_rate": 0.0001892020414732934,
"loss": 0.7853,
"step": 2530
},
{
"epoch": 0.4752530933633296,
"grad_norm": 0.04195011212848616,
"learning_rate": 0.000188728018910666,
"loss": 0.7924,
"step": 2535
},
{
"epoch": 0.47619047619047616,
"grad_norm": 0.045188665210356145,
"learning_rate": 0.00018825358160206982,
"loss": 0.7961,
"step": 2540
},
{
"epoch": 0.4771278590176228,
"grad_norm": 0.04796346505150324,
"learning_rate": 0.00018777873462834735,
"loss": 0.7809,
"step": 2545
},
{
"epoch": 0.4780652418447694,
"grad_norm": 0.039964514626294564,
"learning_rate": 0.00018730348307472824,
"loss": 0.7653,
"step": 2550
},
{
"epoch": 0.479002624671916,
"grad_norm": 0.0427034937035765,
"learning_rate": 0.0001868278320307747,
"loss": 0.7726,
"step": 2555
},
{
"epoch": 0.4799400074990626,
"grad_norm": 0.03972042150913065,
"learning_rate": 0.00018635178659032732,
"loss": 0.7805,
"step": 2560
},
{
"epoch": 0.48087739032620924,
"grad_norm": 0.0371985040034197,
"learning_rate": 0.0001858753518514503,
"loss": 0.7561,
"step": 2565
},
{
"epoch": 0.48181477315335586,
"grad_norm": 0.04584757733571624,
"learning_rate": 0.00018539853291637696,
"loss": 0.7753,
"step": 2570
},
{
"epoch": 0.4827521559805024,
"grad_norm": 0.03974757572397099,
"learning_rate": 0.00018492133489145506,
"loss": 0.7748,
"step": 2575
},
{
"epoch": 0.48368953880764903,
"grad_norm": 0.03742920129272121,
"learning_rate": 0.000184443762887092,
"loss": 0.7917,
"step": 2580
},
{
"epoch": 0.48462692163479565,
"grad_norm": 0.04694450953419577,
"learning_rate": 0.00018396582201770032,
"loss": 0.7859,
"step": 2585
},
{
"epoch": 0.48556430446194226,
"grad_norm": 0.042973639248807004,
"learning_rate": 0.00018348751740164272,
"loss": 0.7836,
"step": 2590
},
{
"epoch": 0.4865016872890889,
"grad_norm": 0.04047247960728854,
"learning_rate": 0.00018300885416117733,
"loss": 0.7796,
"step": 2595
},
{
"epoch": 0.4874390701162355,
"grad_norm": 0.0396897138962343,
"learning_rate": 0.000182529837422403,
"loss": 0.7682,
"step": 2600
},
{
"epoch": 0.48837645294338206,
"grad_norm": 0.0398548055720446,
"learning_rate": 0.0001820504723152041,
"loss": 0.7932,
"step": 2605
},
{
"epoch": 0.48931383577052867,
"grad_norm": 0.037622818268512194,
"learning_rate": 0.0001815707639731958,
"loss": 0.7781,
"step": 2610
},
{
"epoch": 0.4902512185976753,
"grad_norm": 0.0392715706617893,
"learning_rate": 0.00018109071753366916,
"loss": 0.7929,
"step": 2615
},
{
"epoch": 0.4911886014248219,
"grad_norm": 0.039744227841198555,
"learning_rate": 0.00018061033813753576,
"loss": 0.7756,
"step": 2620
},
{
"epoch": 0.4921259842519685,
"grad_norm": 0.04155014821494704,
"learning_rate": 0.00018012963092927297,
"loss": 0.7706,
"step": 2625
},
{
"epoch": 0.49306336707911513,
"grad_norm": 0.03840166882540415,
"learning_rate": 0.0001796486010568689,
"loss": 0.7893,
"step": 2630
},
{
"epoch": 0.4940007499062617,
"grad_norm": 0.039395802156394155,
"learning_rate": 0.000179167253671767,
"loss": 0.775,
"step": 2635
},
{
"epoch": 0.4949381327334083,
"grad_norm": 0.03787300343153706,
"learning_rate": 0.00017868559392881107,
"loss": 0.795,
"step": 2640
},
{
"epoch": 0.4958755155605549,
"grad_norm": 0.040301333180301116,
"learning_rate": 0.0001782036269861899,
"loss": 0.7775,
"step": 2645
},
{
"epoch": 0.49681289838770154,
"grad_norm": 0.03780702169928712,
"learning_rate": 0.0001777213580053823,
"loss": 0.7774,
"step": 2650
},
{
"epoch": 0.49775028121484816,
"grad_norm": 0.04018675410195464,
"learning_rate": 0.0001772387921511016,
"loss": 0.7853,
"step": 2655
},
{
"epoch": 0.49868766404199477,
"grad_norm": 0.039105991236874874,
"learning_rate": 0.00017675593459124045,
"loss": 0.7853,
"step": 2660
},
{
"epoch": 0.49962504686914133,
"grad_norm": 0.043058223660341025,
"learning_rate": 0.00017627279049681538,
"loss": 0.779,
"step": 2665
},
{
"epoch": 0.500562429696288,
"grad_norm": 0.038503318239807786,
"learning_rate": 0.0001757893650419114,
"loss": 0.7746,
"step": 2670
},
{
"epoch": 0.5014998125234346,
"grad_norm": 0.038809904361382756,
"learning_rate": 0.00017530566340362685,
"loss": 0.7836,
"step": 2675
},
{
"epoch": 0.5024371953505812,
"grad_norm": 0.03720404583133728,
"learning_rate": 0.00017482169076201765,
"loss": 0.7738,
"step": 2680
},
{
"epoch": 0.5033745781777278,
"grad_norm": 0.04072977160239652,
"learning_rate": 0.00017433745230004192,
"loss": 0.7932,
"step": 2685
},
{
"epoch": 0.5043119610048744,
"grad_norm": 0.04007107726737037,
"learning_rate": 0.00017385295320350463,
"loss": 0.7925,
"step": 2690
},
{
"epoch": 0.505249343832021,
"grad_norm": 0.03706793674088091,
"learning_rate": 0.00017336819866100182,
"loss": 0.7802,
"step": 2695
},
{
"epoch": 0.5061867266591676,
"grad_norm": 0.03970514233327491,
"learning_rate": 0.00017288319386386515,
"loss": 0.7641,
"step": 2700
},
{
"epoch": 0.5071241094863143,
"grad_norm": 0.04248378336710558,
"learning_rate": 0.0001723979440061064,
"loss": 0.7843,
"step": 2705
},
{
"epoch": 0.5080614923134609,
"grad_norm": 0.035133630205708484,
"learning_rate": 0.00017191245428436173,
"loss": 0.7685,
"step": 2710
},
{
"epoch": 0.5089988751406074,
"grad_norm": 0.03584192505625803,
"learning_rate": 0.00017142672989783601,
"loss": 0.7488,
"step": 2715
},
{
"epoch": 0.509936257967754,
"grad_norm": 0.043073407159687965,
"learning_rate": 0.00017094077604824708,
"loss": 0.7625,
"step": 2720
},
{
"epoch": 0.5108736407949006,
"grad_norm": 0.04373054551404995,
"learning_rate": 0.00017045459793977037,
"loss": 0.7793,
"step": 2725
},
{
"epoch": 0.5118110236220472,
"grad_norm": 0.0401634365888691,
"learning_rate": 0.00016996820077898285,
"loss": 0.7673,
"step": 2730
},
{
"epoch": 0.5127484064491938,
"grad_norm": 0.04228564245730852,
"learning_rate": 0.00016948158977480722,
"loss": 0.7829,
"step": 2735
},
{
"epoch": 0.5136857892763405,
"grad_norm": 0.04101625807575367,
"learning_rate": 0.00016899477013845656,
"loss": 0.7739,
"step": 2740
},
{
"epoch": 0.5146231721034871,
"grad_norm": 0.037270510821881576,
"learning_rate": 0.00016850774708337794,
"loss": 0.7819,
"step": 2745
},
{
"epoch": 0.5155605549306337,
"grad_norm": 0.03719664186844611,
"learning_rate": 0.00016802052582519706,
"loss": 0.7547,
"step": 2750
},
{
"epoch": 0.5164979377577803,
"grad_norm": 0.042746632036701274,
"learning_rate": 0.00016753311158166216,
"loss": 0.7875,
"step": 2755
},
{
"epoch": 0.5174353205849269,
"grad_norm": 0.040113090184156625,
"learning_rate": 0.00016704550957258817,
"loss": 0.7671,
"step": 2760
},
{
"epoch": 0.5183727034120735,
"grad_norm": 0.04096689940473245,
"learning_rate": 0.0001665577250198009,
"loss": 0.7504,
"step": 2765
},
{
"epoch": 0.5193100862392201,
"grad_norm": 0.03979245826251994,
"learning_rate": 0.00016606976314708104,
"loss": 0.7692,
"step": 2770
},
{
"epoch": 0.5202474690663667,
"grad_norm": 0.03499884259333867,
"learning_rate": 0.0001655816291801082,
"loss": 0.7502,
"step": 2775
},
{
"epoch": 0.5211848518935133,
"grad_norm": 0.038595691040614596,
"learning_rate": 0.00016509332834640505,
"loss": 0.7779,
"step": 2780
},
{
"epoch": 0.5221222347206599,
"grad_norm": 0.03845551078176031,
"learning_rate": 0.00016460486587528114,
"loss": 0.7734,
"step": 2785
},
{
"epoch": 0.5230596175478065,
"grad_norm": 0.04268014575183859,
"learning_rate": 0.00016411624699777717,
"loss": 0.7932,
"step": 2790
},
{
"epoch": 0.5239970003749531,
"grad_norm": 0.04217654750286591,
"learning_rate": 0.0001636274769466087,
"loss": 0.7755,
"step": 2795
},
{
"epoch": 0.5249343832020997,
"grad_norm": 0.03663813497845865,
"learning_rate": 0.00016313856095611037,
"loss": 0.7819,
"step": 2800
},
{
"epoch": 0.5258717660292463,
"grad_norm": 0.03649865813691932,
"learning_rate": 0.00016264950426217963,
"loss": 0.7854,
"step": 2805
},
{
"epoch": 0.526809148856393,
"grad_norm": 0.035007178804664606,
"learning_rate": 0.0001621603121022208,
"loss": 0.7763,
"step": 2810
},
{
"epoch": 0.5277465316835396,
"grad_norm": 0.03631407653031134,
"learning_rate": 0.00016167098971508884,
"loss": 0.75,
"step": 2815
},
{
"epoch": 0.5286839145106862,
"grad_norm": 0.03748295963128187,
"learning_rate": 0.00016118154234103345,
"loss": 0.7755,
"step": 2820
},
{
"epoch": 0.5296212973378328,
"grad_norm": 0.03847790308231328,
"learning_rate": 0.00016069197522164272,
"loss": 0.7721,
"step": 2825
},
{
"epoch": 0.5305586801649794,
"grad_norm": 0.0403928945535946,
"learning_rate": 0.00016020229359978722,
"loss": 0.7823,
"step": 2830
},
{
"epoch": 0.531496062992126,
"grad_norm": 0.042093934154398625,
"learning_rate": 0.0001597125027195637,
"loss": 0.7594,
"step": 2835
},
{
"epoch": 0.5324334458192725,
"grad_norm": 0.03740153505333051,
"learning_rate": 0.00015922260782623906,
"loss": 0.775,
"step": 2840
},
{
"epoch": 0.5333708286464192,
"grad_norm": 0.0365020182974372,
"learning_rate": 0.00015873261416619395,
"loss": 0.7788,
"step": 2845
},
{
"epoch": 0.5343082114735658,
"grad_norm": 0.04097368109261412,
"learning_rate": 0.00015824252698686686,
"loss": 0.7801,
"step": 2850
},
{
"epoch": 0.5352455943007124,
"grad_norm": 0.040252898201126604,
"learning_rate": 0.00015775235153669772,
"loss": 0.7651,
"step": 2855
},
{
"epoch": 0.536182977127859,
"grad_norm": 0.04080680519751646,
"learning_rate": 0.00015726209306507182,
"loss": 0.7609,
"step": 2860
},
{
"epoch": 0.5371203599550056,
"grad_norm": 0.04064129669100478,
"learning_rate": 0.00015677175682226346,
"loss": 0.7686,
"step": 2865
},
{
"epoch": 0.5380577427821522,
"grad_norm": 0.040674542378951246,
"learning_rate": 0.0001562813480593799,
"loss": 0.7616,
"step": 2870
},
{
"epoch": 0.5389951256092989,
"grad_norm": 0.04054872658419522,
"learning_rate": 0.0001557908720283051,
"loss": 0.7938,
"step": 2875
},
{
"epoch": 0.5399325084364455,
"grad_norm": 0.037193930697877325,
"learning_rate": 0.00015530033398164318,
"loss": 0.7671,
"step": 2880
},
{
"epoch": 0.5408698912635921,
"grad_norm": 0.04330464356262373,
"learning_rate": 0.00015480973917266256,
"loss": 0.789,
"step": 2885
},
{
"epoch": 0.5418072740907387,
"grad_norm": 0.04054372294824471,
"learning_rate": 0.0001543190928552395,
"loss": 0.7511,
"step": 2890
},
{
"epoch": 0.5427446569178853,
"grad_norm": 0.037141465243079236,
"learning_rate": 0.00015382840028380193,
"loss": 0.7806,
"step": 2895
},
{
"epoch": 0.5436820397450318,
"grad_norm": 0.04616963653816421,
"learning_rate": 0.000153337666713273,
"loss": 0.7685,
"step": 2900
},
{
"epoch": 0.5446194225721784,
"grad_norm": 0.04026197513442802,
"learning_rate": 0.000152846897399015,
"loss": 0.7538,
"step": 2905
},
{
"epoch": 0.545556805399325,
"grad_norm": 0.040209956896230005,
"learning_rate": 0.0001523560975967731,
"loss": 0.7669,
"step": 2910
},
{
"epoch": 0.5464941882264717,
"grad_norm": 0.039214472443270074,
"learning_rate": 0.0001518652725626188,
"loss": 0.7821,
"step": 2915
},
{
"epoch": 0.5474315710536183,
"grad_norm": 0.037735873161464144,
"learning_rate": 0.00015137442755289388,
"loss": 0.7669,
"step": 2920
},
{
"epoch": 0.5483689538807649,
"grad_norm": 0.040692415941492,
"learning_rate": 0.00015088356782415408,
"loss": 0.7642,
"step": 2925
},
{
"epoch": 0.5493063367079115,
"grad_norm": 0.037557444480219304,
"learning_rate": 0.0001503926986331127,
"loss": 0.7491,
"step": 2930
},
{
"epoch": 0.5502437195350581,
"grad_norm": 0.04303948252096911,
"learning_rate": 0.00014990182523658444,
"loss": 0.7633,
"step": 2935
},
{
"epoch": 0.5511811023622047,
"grad_norm": 0.039445563951292666,
"learning_rate": 0.0001494109528914291,
"loss": 0.7837,
"step": 2940
},
{
"epoch": 0.5521184851893514,
"grad_norm": 0.0369146298860869,
"learning_rate": 0.000148920086854495,
"loss": 0.7783,
"step": 2945
},
{
"epoch": 0.553055868016498,
"grad_norm": 0.03928589902574045,
"learning_rate": 0.00014842923238256317,
"loss": 0.7623,
"step": 2950
},
{
"epoch": 0.5539932508436446,
"grad_norm": 0.040961368365580925,
"learning_rate": 0.00014793839473229047,
"loss": 0.7302,
"step": 2955
},
{
"epoch": 0.5549306336707911,
"grad_norm": 0.04066646466874108,
"learning_rate": 0.000147447579160154,
"loss": 0.7933,
"step": 2960
},
{
"epoch": 0.5558680164979377,
"grad_norm": 0.04046950253782251,
"learning_rate": 0.00014695679092239405,
"loss": 0.7794,
"step": 2965
},
{
"epoch": 0.5568053993250843,
"grad_norm": 0.0395374977204719,
"learning_rate": 0.00014646603527495848,
"loss": 0.772,
"step": 2970
},
{
"epoch": 0.5577427821522309,
"grad_norm": 0.03703594756866284,
"learning_rate": 0.00014597531747344593,
"loss": 0.7511,
"step": 2975
},
{
"epoch": 0.5586801649793776,
"grad_norm": 0.036214320470059805,
"learning_rate": 0.00014548464277304982,
"loss": 0.7706,
"step": 2980
},
{
"epoch": 0.5596175478065242,
"grad_norm": 0.03917835401091949,
"learning_rate": 0.00014499401642850207,
"loss": 0.739,
"step": 2985
},
{
"epoch": 0.5605549306336708,
"grad_norm": 0.0498467008354317,
"learning_rate": 0.00014450344369401651,
"loss": 0.763,
"step": 2990
},
{
"epoch": 0.5614923134608174,
"grad_norm": 0.05384613905490524,
"learning_rate": 0.0001440129298232332,
"loss": 0.7892,
"step": 2995
},
{
"epoch": 0.562429696287964,
"grad_norm": 0.0457358145474257,
"learning_rate": 0.0001435224800691614,
"loss": 0.7988,
"step": 3000
},
{
"epoch": 0.5633670791151106,
"grad_norm": 0.04336091654025471,
"learning_rate": 0.00014303209968412418,
"loss": 0.799,
"step": 3005
},
{
"epoch": 0.5643044619422573,
"grad_norm": 0.04292534665730051,
"learning_rate": 0.00014254179391970132,
"loss": 0.7775,
"step": 3010
},
{
"epoch": 0.5652418447694039,
"grad_norm": 0.04250101254193063,
"learning_rate": 0.00014205156802667374,
"loss": 0.7788,
"step": 3015
},
{
"epoch": 0.5661792275965505,
"grad_norm": 0.04030937649018944,
"learning_rate": 0.00014156142725496682,
"loss": 0.7771,
"step": 3020
},
{
"epoch": 0.567116610423697,
"grad_norm": 0.037700014044718524,
"learning_rate": 0.00014107137685359457,
"loss": 0.7603,
"step": 3025
},
{
"epoch": 0.5680539932508436,
"grad_norm": 0.03917274407024749,
"learning_rate": 0.000140581422070603,
"loss": 0.7756,
"step": 3030
},
{
"epoch": 0.5689913760779902,
"grad_norm": 0.03965878064631959,
"learning_rate": 0.00014009156815301426,
"loss": 0.781,
"step": 3035
},
{
"epoch": 0.5699287589051368,
"grad_norm": 0.03694985326600243,
"learning_rate": 0.00013960182034677016,
"loss": 0.7764,
"step": 3040
},
{
"epoch": 0.5708661417322834,
"grad_norm": 0.03649532964794123,
"learning_rate": 0.00013911218389667642,
"loss": 0.7549,
"step": 3045
},
{
"epoch": 0.5718035245594301,
"grad_norm": 0.03787093545824704,
"learning_rate": 0.0001386226640463459,
"loss": 0.7472,
"step": 3050
},
{
"epoch": 0.5727409073865767,
"grad_norm": 0.038098821200233814,
"learning_rate": 0.00013813326603814296,
"loss": 0.794,
"step": 3055
},
{
"epoch": 0.5736782902137233,
"grad_norm": 0.04007624623894434,
"learning_rate": 0.00013764399511312716,
"loss": 0.7973,
"step": 3060
},
{
"epoch": 0.5746156730408699,
"grad_norm": 0.040480907401198925,
"learning_rate": 0.0001371548565109969,
"loss": 0.7693,
"step": 3065
},
{
"epoch": 0.5755530558680165,
"grad_norm": 0.037195195565526876,
"learning_rate": 0.00013666585547003377,
"loss": 0.7548,
"step": 3070
},
{
"epoch": 0.5764904386951631,
"grad_norm": 0.04118922842272934,
"learning_rate": 0.00013617699722704598,
"loss": 0.781,
"step": 3075
},
{
"epoch": 0.5774278215223098,
"grad_norm": 0.041231050481496015,
"learning_rate": 0.0001356882870173126,
"loss": 0.7764,
"step": 3080
},
{
"epoch": 0.5783652043494563,
"grad_norm": 0.03825190175591559,
"learning_rate": 0.0001351997300745273,
"loss": 0.7604,
"step": 3085
},
{
"epoch": 0.5793025871766029,
"grad_norm": 0.041476811594546956,
"learning_rate": 0.0001347113316307425,
"loss": 0.774,
"step": 3090
},
{
"epoch": 0.5802399700037495,
"grad_norm": 0.04062186609985697,
"learning_rate": 0.0001342230969163131,
"loss": 0.7669,
"step": 3095
},
{
"epoch": 0.5811773528308961,
"grad_norm": 0.037469179130772814,
"learning_rate": 0.00013373503115984072,
"loss": 0.8056,
"step": 3100
},
{
"epoch": 0.5821147356580427,
"grad_norm": 0.04327856604489263,
"learning_rate": 0.0001332471395881174,
"loss": 0.7616,
"step": 3105
},
{
"epoch": 0.5830521184851893,
"grad_norm": 0.03958983021180182,
"learning_rate": 0.00013275942742607002,
"loss": 0.7756,
"step": 3110
},
{
"epoch": 0.583989501312336,
"grad_norm": 0.03790940533047183,
"learning_rate": 0.00013227189989670392,
"loss": 0.7568,
"step": 3115
},
{
"epoch": 0.5849268841394826,
"grad_norm": 0.0403239958967278,
"learning_rate": 0.00013178456222104733,
"loss": 0.742,
"step": 3120
},
{
"epoch": 0.5858642669666292,
"grad_norm": 0.03895542187325954,
"learning_rate": 0.00013129741961809513,
"loss": 0.7699,
"step": 3125
},
{
"epoch": 0.5868016497937758,
"grad_norm": 0.03820864797313289,
"learning_rate": 0.00013081047730475331,
"loss": 0.7601,
"step": 3130
},
{
"epoch": 0.5877390326209224,
"grad_norm": 0.04726355486417647,
"learning_rate": 0.00013032374049578292,
"loss": 0.7642,
"step": 3135
},
{
"epoch": 0.588676415448069,
"grad_norm": 0.04203366758155159,
"learning_rate": 0.000129837214403744,
"loss": 0.7488,
"step": 3140
},
{
"epoch": 0.5896137982752157,
"grad_norm": 0.038981436347331345,
"learning_rate": 0.00012935090423894015,
"loss": 0.7862,
"step": 3145
},
{
"epoch": 0.5905511811023622,
"grad_norm": 0.03432435415992516,
"learning_rate": 0.0001288648152093626,
"loss": 0.7525,
"step": 3150
},
{
"epoch": 0.5914885639295088,
"grad_norm": 0.03822931424545101,
"learning_rate": 0.00012837895252063432,
"loss": 0.8001,
"step": 3155
},
{
"epoch": 0.5924259467566554,
"grad_norm": 0.039214953158152176,
"learning_rate": 0.00012789332137595427,
"loss": 0.7245,
"step": 3160
},
{
"epoch": 0.593363329583802,
"grad_norm": 0.037907598245917364,
"learning_rate": 0.00012740792697604202,
"loss": 0.7648,
"step": 3165
},
{
"epoch": 0.5943007124109486,
"grad_norm": 0.03785947037848109,
"learning_rate": 0.00012692277451908145,
"loss": 0.7654,
"step": 3170
},
{
"epoch": 0.5952380952380952,
"grad_norm": 0.03850785633219605,
"learning_rate": 0.0001264378692006658,
"loss": 0.7663,
"step": 3175
},
{
"epoch": 0.5961754780652418,
"grad_norm": 0.036428916422083596,
"learning_rate": 0.00012595321621374135,
"loss": 0.7596,
"step": 3180
},
{
"epoch": 0.5971128608923885,
"grad_norm": 0.038639622718949614,
"learning_rate": 0.0001254688207485522,
"loss": 0.7318,
"step": 3185
},
{
"epoch": 0.5980502437195351,
"grad_norm": 0.04200767445250016,
"learning_rate": 0.00012498468799258466,
"loss": 0.7543,
"step": 3190
},
{
"epoch": 0.5989876265466817,
"grad_norm": 0.039347067609858684,
"learning_rate": 0.00012450082313051163,
"loss": 0.7782,
"step": 3195
},
{
"epoch": 0.5999250093738283,
"grad_norm": 0.03956472616518923,
"learning_rate": 0.00012401723134413694,
"loss": 0.7748,
"step": 3200
},
{
"epoch": 0.6008623922009749,
"grad_norm": 0.03906550581460343,
"learning_rate": 0.00012353391781234026,
"loss": 0.781,
"step": 3205
},
{
"epoch": 0.6017997750281214,
"grad_norm": 0.04226202156309189,
"learning_rate": 0.0001230508877110211,
"loss": 0.7813,
"step": 3210
},
{
"epoch": 0.602737157855268,
"grad_norm": 0.04310759188393267,
"learning_rate": 0.00012256814621304385,
"loss": 0.7538,
"step": 3215
},
{
"epoch": 0.6036745406824147,
"grad_norm": 0.04137403063128998,
"learning_rate": 0.00012208569848818214,
"loss": 0.7785,
"step": 3220
},
{
"epoch": 0.6046119235095613,
"grad_norm": 0.03490693407426236,
"learning_rate": 0.00012160354970306339,
"loss": 0.7615,
"step": 3225
},
{
"epoch": 0.6055493063367079,
"grad_norm": 0.04058597804316207,
"learning_rate": 0.0001211217050211139,
"loss": 0.7737,
"step": 3230
},
{
"epoch": 0.6064866891638545,
"grad_norm": 0.039857581905152456,
"learning_rate": 0.00012064016960250294,
"loss": 0.7624,
"step": 3235
},
{
"epoch": 0.6074240719910011,
"grad_norm": 0.03861470764960366,
"learning_rate": 0.00012015894860408811,
"loss": 0.732,
"step": 3240
},
{
"epoch": 0.6083614548181477,
"grad_norm": 0.03893813785005398,
"learning_rate": 0.00011967804717935964,
"loss": 0.7755,
"step": 3245
},
{
"epoch": 0.6092988376452944,
"grad_norm": 0.03949254956830129,
"learning_rate": 0.00011919747047838545,
"loss": 0.7575,
"step": 3250
},
{
"epoch": 0.610236220472441,
"grad_norm": 0.03697345161480537,
"learning_rate": 0.00011871722364775583,
"loss": 0.7606,
"step": 3255
},
{
"epoch": 0.6111736032995876,
"grad_norm": 0.036342062943778654,
"learning_rate": 0.00011823731183052867,
"loss": 0.766,
"step": 3260
},
{
"epoch": 0.6121109861267342,
"grad_norm": 0.0370100352544129,
"learning_rate": 0.00011775774016617381,
"loss": 0.7623,
"step": 3265
},
{
"epoch": 0.6130483689538808,
"grad_norm": 0.03696482383175287,
"learning_rate": 0.00011727851379051865,
"loss": 0.7655,
"step": 3270
},
{
"epoch": 0.6139857517810273,
"grad_norm": 0.03985849511649083,
"learning_rate": 0.00011679963783569248,
"loss": 0.7495,
"step": 3275
},
{
"epoch": 0.6149231346081739,
"grad_norm": 0.038876495340474354,
"learning_rate": 0.00011632111743007223,
"loss": 0.7813,
"step": 3280
},
{
"epoch": 0.6158605174353206,
"grad_norm": 0.03750007207907501,
"learning_rate": 0.00011584295769822694,
"loss": 0.7528,
"step": 3285
},
{
"epoch": 0.6167979002624672,
"grad_norm": 0.0408841339742789,
"learning_rate": 0.00011536516376086311,
"loss": 0.7719,
"step": 3290
},
{
"epoch": 0.6177352830896138,
"grad_norm": 0.042011717790033064,
"learning_rate": 0.0001148877407347701,
"loss": 0.7493,
"step": 3295
},
{
"epoch": 0.6186726659167604,
"grad_norm": 0.0414847232838945,
"learning_rate": 0.00011441069373276481,
"loss": 0.7536,
"step": 3300
},
{
"epoch": 0.619610048743907,
"grad_norm": 0.042292805389266726,
"learning_rate": 0.00011393402786363753,
"loss": 0.7602,
"step": 3305
},
{
"epoch": 0.6205474315710536,
"grad_norm": 0.038665796975353536,
"learning_rate": 0.00011345774823209661,
"loss": 0.7504,
"step": 3310
},
{
"epoch": 0.6214848143982002,
"grad_norm": 0.04162065454602133,
"learning_rate": 0.00011298185993871442,
"loss": 0.7638,
"step": 3315
},
{
"epoch": 0.6224221972253469,
"grad_norm": 0.03784958872128606,
"learning_rate": 0.00011250636807987208,
"loss": 0.7368,
"step": 3320
},
{
"epoch": 0.6233595800524935,
"grad_norm": 0.039769498866178414,
"learning_rate": 0.00011203127774770554,
"loss": 0.7664,
"step": 3325
},
{
"epoch": 0.6242969628796401,
"grad_norm": 0.0373640006459259,
"learning_rate": 0.00011155659403005048,
"loss": 0.7524,
"step": 3330
},
{
"epoch": 0.6252343457067866,
"grad_norm": 0.03809871959132381,
"learning_rate": 0.00011108232201038828,
"loss": 0.7484,
"step": 3335
},
{
"epoch": 0.6261717285339332,
"grad_norm": 0.03763522153071777,
"learning_rate": 0.00011060846676779107,
"loss": 0.7433,
"step": 3340
},
{
"epoch": 0.6271091113610798,
"grad_norm": 0.03946664053179257,
"learning_rate": 0.00011013503337686799,
"loss": 0.7677,
"step": 3345
},
{
"epoch": 0.6280464941882264,
"grad_norm": 0.03901754964576778,
"learning_rate": 0.00010966202690771014,
"loss": 0.7561,
"step": 3350
},
{
"epoch": 0.6289838770153731,
"grad_norm": 0.03912014672930949,
"learning_rate": 0.00010918945242583688,
"loss": 0.7789,
"step": 3355
},
{
"epoch": 0.6299212598425197,
"grad_norm": 0.03717356045637909,
"learning_rate": 0.00010871731499214128,
"loss": 0.7521,
"step": 3360
},
{
"epoch": 0.6308586426696663,
"grad_norm": 0.037849859871512744,
"learning_rate": 0.00010824561966283583,
"loss": 0.7638,
"step": 3365
},
{
"epoch": 0.6317960254968129,
"grad_norm": 0.03680463786596637,
"learning_rate": 0.00010777437148939868,
"loss": 0.7417,
"step": 3370
},
{
"epoch": 0.6327334083239595,
"grad_norm": 0.03897036657910483,
"learning_rate": 0.0001073035755185191,
"loss": 0.737,
"step": 3375
},
{
"epoch": 0.6336707911511061,
"grad_norm": 0.04507944036400963,
"learning_rate": 0.00010683323679204374,
"loss": 0.7526,
"step": 3380
},
{
"epoch": 0.6346081739782528,
"grad_norm": 0.041130560554180164,
"learning_rate": 0.00010636336034692238,
"loss": 0.7472,
"step": 3385
},
{
"epoch": 0.6355455568053994,
"grad_norm": 0.037500549260058,
"learning_rate": 0.00010589395121515441,
"loss": 0.7642,
"step": 3390
},
{
"epoch": 0.636482939632546,
"grad_norm": 0.03501911608758103,
"learning_rate": 0.00010542501442373441,
"loss": 0.7552,
"step": 3395
},
{
"epoch": 0.6374203224596925,
"grad_norm": 0.041747590796681534,
"learning_rate": 0.00010495655499459874,
"loss": 0.7825,
"step": 3400
},
{
"epoch": 0.6383577052868391,
"grad_norm": 0.0415454403917153,
"learning_rate": 0.00010448857794457143,
"loss": 0.7615,
"step": 3405
},
{
"epoch": 0.6392950881139857,
"grad_norm": 0.034393345932052226,
"learning_rate": 0.00010402108828531086,
"loss": 0.7501,
"step": 3410
},
{
"epoch": 0.6402324709411323,
"grad_norm": 0.041393829589670965,
"learning_rate": 0.00010355409102325557,
"loss": 0.7725,
"step": 3415
},
{
"epoch": 0.641169853768279,
"grad_norm": 0.039070120349108735,
"learning_rate": 0.00010308759115957118,
"loss": 0.7953,
"step": 3420
},
{
"epoch": 0.6421072365954256,
"grad_norm": 0.03792543146870935,
"learning_rate": 0.00010262159369009628,
"loss": 0.7584,
"step": 3425
},
{
"epoch": 0.6430446194225722,
"grad_norm": 0.04012922678154492,
"learning_rate": 0.00010215610360528948,
"loss": 0.736,
"step": 3430
},
{
"epoch": 0.6439820022497188,
"grad_norm": 0.03600161764198426,
"learning_rate": 0.00010169112589017568,
"loss": 0.739,
"step": 3435
},
{
"epoch": 0.6449193850768654,
"grad_norm": 0.04003494519804115,
"learning_rate": 0.00010122666552429259,
"loss": 0.7668,
"step": 3440
},
{
"epoch": 0.645856767904012,
"grad_norm": 0.039267717082010545,
"learning_rate": 0.00010076272748163756,
"loss": 0.7683,
"step": 3445
},
{
"epoch": 0.6467941507311586,
"grad_norm": 0.037805171037559454,
"learning_rate": 0.00010029931673061433,
"loss": 0.7426,
"step": 3450
},
{
"epoch": 0.6477315335583053,
"grad_norm": 0.037277881133197355,
"learning_rate": 9.98364382339798e-05,
"loss": 0.7267,
"step": 3455
},
{
"epoch": 0.6486689163854518,
"grad_norm": 0.043347070226570104,
"learning_rate": 9.937409694879064e-05,
"loss": 0.7548,
"step": 3460
},
{
"epoch": 0.6496062992125984,
"grad_norm": 0.03829734267181962,
"learning_rate": 9.891229782635074e-05,
"loss": 0.7718,
"step": 3465
},
{
"epoch": 0.650543682039745,
"grad_norm": 0.038650875886255875,
"learning_rate": 9.845104581215758e-05,
"loss": 0.7705,
"step": 3470
},
{
"epoch": 0.6514810648668916,
"grad_norm": 0.042871623377863474,
"learning_rate": 9.799034584584975e-05,
"loss": 0.7495,
"step": 3475
},
{
"epoch": 0.6524184476940382,
"grad_norm": 0.03585928229669108,
"learning_rate": 9.753020286115368e-05,
"loss": 0.7432,
"step": 3480
},
{
"epoch": 0.6533558305211848,
"grad_norm": 0.03786939257702819,
"learning_rate": 9.707062178583119e-05,
"loss": 0.7904,
"step": 3485
},
{
"epoch": 0.6542932133483315,
"grad_norm": 0.0368092502095729,
"learning_rate": 9.661160754162618e-05,
"loss": 0.7555,
"step": 3490
},
{
"epoch": 0.6552305961754781,
"grad_norm": 0.03613287171684745,
"learning_rate": 9.615316504421262e-05,
"loss": 0.7439,
"step": 3495
},
{
"epoch": 0.6561679790026247,
"grad_norm": 0.036433445794931595,
"learning_rate": 9.569529920314121e-05,
"loss": 0.7529,
"step": 3500
},
{
"epoch": 0.6571053618297713,
"grad_norm": 0.0424735007867028,
"learning_rate": 9.523801492178736e-05,
"loss": 0.7397,
"step": 3505
},
{
"epoch": 0.6580427446569179,
"grad_norm": 0.03698802222139945,
"learning_rate": 9.47813170972983e-05,
"loss": 0.7437,
"step": 3510
},
{
"epoch": 0.6589801274840645,
"grad_norm": 0.03733477324154079,
"learning_rate": 9.432521062054084e-05,
"loss": 0.7705,
"step": 3515
},
{
"epoch": 0.6599175103112112,
"grad_norm": 0.03508532202382053,
"learning_rate": 9.386970037604892e-05,
"loss": 0.7392,
"step": 3520
},
{
"epoch": 0.6608548931383577,
"grad_norm": 0.03827744300792672,
"learning_rate": 9.341479124197123e-05,
"loss": 0.7238,
"step": 3525
},
{
"epoch": 0.6617922759655043,
"grad_norm": 0.034951831735317074,
"learning_rate": 9.296048809001928e-05,
"loss": 0.7445,
"step": 3530
},
{
"epoch": 0.6627296587926509,
"grad_norm": 0.036754627748938604,
"learning_rate": 9.250679578541465e-05,
"loss": 0.7648,
"step": 3535
},
{
"epoch": 0.6636670416197975,
"grad_norm": 0.04075976815278343,
"learning_rate": 9.205371918683761e-05,
"loss": 0.7498,
"step": 3540
},
{
"epoch": 0.6646044244469441,
"grad_norm": 0.03941979332992536,
"learning_rate": 9.16012631463744e-05,
"loss": 0.7599,
"step": 3545
},
{
"epoch": 0.6655418072740907,
"grad_norm": 0.03389585413188664,
"learning_rate": 9.114943250946581e-05,
"loss": 0.7242,
"step": 3550
},
{
"epoch": 0.6664791901012374,
"grad_norm": 0.037973128654648046,
"learning_rate": 9.069823211485485e-05,
"loss": 0.7397,
"step": 3555
},
{
"epoch": 0.667416572928384,
"grad_norm": 0.03493012197632408,
"learning_rate": 9.024766679453538e-05,
"loss": 0.7218,
"step": 3560
},
{
"epoch": 0.6683539557555306,
"grad_norm": 0.048208082428385556,
"learning_rate": 8.979774137369989e-05,
"loss": 0.7709,
"step": 3565
},
{
"epoch": 0.6692913385826772,
"grad_norm": 0.039871249220599175,
"learning_rate": 8.934846067068825e-05,
"loss": 0.7306,
"step": 3570
},
{
"epoch": 0.6702287214098238,
"grad_norm": 0.03855933322661882,
"learning_rate": 8.88998294969358e-05,
"loss": 0.73,
"step": 3575
},
{
"epoch": 0.6711661042369704,
"grad_norm": 0.0397648331368396,
"learning_rate": 8.845185265692201e-05,
"loss": 0.7324,
"step": 3580
},
{
"epoch": 0.6721034870641169,
"grad_norm": 0.03767828224137,
"learning_rate": 8.800453494811894e-05,
"loss": 0.7372,
"step": 3585
},
{
"epoch": 0.6730408698912635,
"grad_norm": 0.03982314491187269,
"learning_rate": 8.755788116093983e-05,
"loss": 0.7735,
"step": 3590
},
{
"epoch": 0.6739782527184102,
"grad_norm": 0.03712738479588227,
"learning_rate": 8.711189607868795e-05,
"loss": 0.725,
"step": 3595
},
{
"epoch": 0.6749156355455568,
"grad_norm": 0.03733488944845488,
"learning_rate": 8.666658447750519e-05,
"loss": 0.7372,
"step": 3600
},
{
"epoch": 0.6758530183727034,
"grad_norm": 0.0377678114465905,
"learning_rate": 8.622195112632107e-05,
"loss": 0.7225,
"step": 3605
},
{
"epoch": 0.67679040119985,
"grad_norm": 0.03971467083786042,
"learning_rate": 8.577800078680156e-05,
"loss": 0.7691,
"step": 3610
},
{
"epoch": 0.6777277840269966,
"grad_norm": 0.041555052270688675,
"learning_rate": 8.533473821329814e-05,
"loss": 0.746,
"step": 3615
},
{
"epoch": 0.6786651668541432,
"grad_norm": 0.03681199845630553,
"learning_rate": 8.489216815279682e-05,
"loss": 0.7326,
"step": 3620
},
{
"epoch": 0.6796025496812899,
"grad_norm": 0.03639372767368361,
"learning_rate": 8.445029534486741e-05,
"loss": 0.7558,
"step": 3625
},
{
"epoch": 0.6805399325084365,
"grad_norm": 0.03930037562387389,
"learning_rate": 8.400912452161271e-05,
"loss": 0.7566,
"step": 3630
},
{
"epoch": 0.6814773153355831,
"grad_norm": 0.038113159588276734,
"learning_rate": 8.356866040761786e-05,
"loss": 0.7479,
"step": 3635
},
{
"epoch": 0.6824146981627297,
"grad_norm": 0.03668907273839101,
"learning_rate": 8.312890771989943e-05,
"loss": 0.7507,
"step": 3640
},
{
"epoch": 0.6833520809898763,
"grad_norm": 0.03877486758990776,
"learning_rate": 8.268987116785569e-05,
"loss": 0.7119,
"step": 3645
},
{
"epoch": 0.6842894638170228,
"grad_norm": 0.03641698338134557,
"learning_rate": 8.225155545321514e-05,
"loss": 0.7419,
"step": 3650
},
{
"epoch": 0.6852268466441694,
"grad_norm": 0.037028014658097706,
"learning_rate": 8.181396526998713e-05,
"loss": 0.7511,
"step": 3655
},
{
"epoch": 0.6861642294713161,
"grad_norm": 0.03742801285890605,
"learning_rate": 8.13771053044109e-05,
"loss": 0.7464,
"step": 3660
},
{
"epoch": 0.6871016122984627,
"grad_norm": 0.03700498083910332,
"learning_rate": 8.094098023490573e-05,
"loss": 0.7509,
"step": 3665
},
{
"epoch": 0.6880389951256093,
"grad_norm": 0.041613081744209884,
"learning_rate": 8.050559473202077e-05,
"loss": 0.7547,
"step": 3670
},
{
"epoch": 0.6889763779527559,
"grad_norm": 0.03578452096886478,
"learning_rate": 8.00709534583848e-05,
"loss": 0.7486,
"step": 3675
},
{
"epoch": 0.6899137607799025,
"grad_norm": 0.039555712535017766,
"learning_rate": 7.963706106865692e-05,
"loss": 0.7441,
"step": 3680
},
{
"epoch": 0.6908511436070491,
"grad_norm": 0.03614203981567192,
"learning_rate": 7.920392220947577e-05,
"loss": 0.7546,
"step": 3685
},
{
"epoch": 0.6917885264341957,
"grad_norm": 0.03936732269908494,
"learning_rate": 7.877154151941082e-05,
"loss": 0.7544,
"step": 3690
},
{
"epoch": 0.6927259092613424,
"grad_norm": 0.04056135650395248,
"learning_rate": 7.833992362891173e-05,
"loss": 0.748,
"step": 3695
},
{
"epoch": 0.693663292088489,
"grad_norm": 0.03813900176127187,
"learning_rate": 7.790907316025935e-05,
"loss": 0.7566,
"step": 3700
},
{
"epoch": 0.6946006749156356,
"grad_norm": 0.036115814487516676,
"learning_rate": 7.74789947275161e-05,
"loss": 0.731,
"step": 3705
},
{
"epoch": 0.6955380577427821,
"grad_norm": 0.041142128143449405,
"learning_rate": 7.704969293647643e-05,
"loss": 0.7686,
"step": 3710
},
{
"epoch": 0.6964754405699287,
"grad_norm": 0.04017787329812371,
"learning_rate": 7.662117238461769e-05,
"loss": 0.7641,
"step": 3715
},
{
"epoch": 0.6974128233970753,
"grad_norm": 0.03912928253082181,
"learning_rate": 7.619343766105065e-05,
"loss": 0.7337,
"step": 3720
},
{
"epoch": 0.698350206224222,
"grad_norm": 0.03739403328622796,
"learning_rate": 7.576649334647063e-05,
"loss": 0.7688,
"step": 3725
},
{
"epoch": 0.6992875890513686,
"grad_norm": 0.0362755943904781,
"learning_rate": 7.534034401310817e-05,
"loss": 0.7449,
"step": 3730
},
{
"epoch": 0.7002249718785152,
"grad_norm": 0.03857280564508149,
"learning_rate": 7.49149942246803e-05,
"loss": 0.7417,
"step": 3735
},
{
"epoch": 0.7011623547056618,
"grad_norm": 0.03765394646978099,
"learning_rate": 7.449044853634153e-05,
"loss": 0.7733,
"step": 3740
},
{
"epoch": 0.7020997375328084,
"grad_norm": 0.038387501642174006,
"learning_rate": 7.406671149463509e-05,
"loss": 0.7157,
"step": 3745
},
{
"epoch": 0.703037120359955,
"grad_norm": 0.03475103792139527,
"learning_rate": 7.364378763744429e-05,
"loss": 0.7478,
"step": 3750
},
{
"epoch": 0.7039745031871016,
"grad_norm": 0.0378553405313383,
"learning_rate": 7.322168149394386e-05,
"loss": 0.7566,
"step": 3755
},
{
"epoch": 0.7049118860142483,
"grad_norm": 0.039967415152981675,
"learning_rate": 7.280039758455147e-05,
"loss": 0.7541,
"step": 3760
},
{
"epoch": 0.7058492688413949,
"grad_norm": 0.035394814696152195,
"learning_rate": 7.23799404208794e-05,
"loss": 0.7625,
"step": 3765
},
{
"epoch": 0.7067866516685414,
"grad_norm": 0.038752752547861896,
"learning_rate": 7.19603145056859e-05,
"loss": 0.7612,
"step": 3770
},
{
"epoch": 0.707724034495688,
"grad_norm": 0.037109030416605036,
"learning_rate": 7.154152433282762e-05,
"loss": 0.7382,
"step": 3775
},
{
"epoch": 0.7086614173228346,
"grad_norm": 0.03554207953122535,
"learning_rate": 7.112357438721065e-05,
"loss": 0.7455,
"step": 3780
},
{
"epoch": 0.7095988001499812,
"grad_norm": 0.039676102077878514,
"learning_rate": 7.070646914474335e-05,
"loss": 0.7173,
"step": 3785
},
{
"epoch": 0.7105361829771278,
"grad_norm": 0.03421837506332554,
"learning_rate": 7.029021307228755e-05,
"loss": 0.7261,
"step": 3790
},
{
"epoch": 0.7114735658042745,
"grad_norm": 0.04037332220576155,
"learning_rate": 6.987481062761157e-05,
"loss": 0.7435,
"step": 3795
},
{
"epoch": 0.7124109486314211,
"grad_norm": 0.03870777741485929,
"learning_rate": 6.94602662593417e-05,
"loss": 0.7263,
"step": 3800
},
{
"epoch": 0.7133483314585677,
"grad_norm": 0.04496580126935979,
"learning_rate": 6.90465844069151e-05,
"loss": 0.776,
"step": 3805
},
{
"epoch": 0.7142857142857143,
"grad_norm": 0.038235977347547696,
"learning_rate": 6.863376950053221e-05,
"loss": 0.7408,
"step": 3810
},
{
"epoch": 0.7152230971128609,
"grad_norm": 0.03531689033224846,
"learning_rate": 6.82218259611088e-05,
"loss": 0.738,
"step": 3815
},
{
"epoch": 0.7161604799400075,
"grad_norm": 0.036562035053447246,
"learning_rate": 6.781075820022946e-05,
"loss": 0.779,
"step": 3820
},
{
"epoch": 0.7170978627671541,
"grad_norm": 0.03879265532726638,
"learning_rate": 6.740057062009951e-05,
"loss": 0.7514,
"step": 3825
},
{
"epoch": 0.7180352455943008,
"grad_norm": 0.037768573089647754,
"learning_rate": 6.69912676134984e-05,
"loss": 0.7423,
"step": 3830
},
{
"epoch": 0.7189726284214473,
"grad_norm": 0.03723804398233768,
"learning_rate": 6.658285356373253e-05,
"loss": 0.7277,
"step": 3835
},
{
"epoch": 0.7199100112485939,
"grad_norm": 0.03866540526571928,
"learning_rate": 6.617533284458826e-05,
"loss": 0.7452,
"step": 3840
},
{
"epoch": 0.7208473940757405,
"grad_norm": 0.03743514627207939,
"learning_rate": 6.576870982028508e-05,
"loss": 0.7487,
"step": 3845
},
{
"epoch": 0.7217847769028871,
"grad_norm": 0.04147587697115252,
"learning_rate": 6.53629888454289e-05,
"loss": 0.7273,
"step": 3850
},
{
"epoch": 0.7227221597300337,
"grad_norm": 0.04232917813229702,
"learning_rate": 6.495817426496541e-05,
"loss": 0.7413,
"step": 3855
},
{
"epoch": 0.7236595425571803,
"grad_norm": 0.041237294474918335,
"learning_rate": 6.455427041413356e-05,
"loss": 0.735,
"step": 3860
},
{
"epoch": 0.724596925384327,
"grad_norm": 0.039856233442403854,
"learning_rate": 6.415128161841909e-05,
"loss": 0.7512,
"step": 3865
},
{
"epoch": 0.7255343082114736,
"grad_norm": 0.03843674829308366,
"learning_rate": 6.374921219350826e-05,
"loss": 0.7421,
"step": 3870
},
{
"epoch": 0.7264716910386202,
"grad_norm": 0.03492500389641903,
"learning_rate": 6.334806644524147e-05,
"loss": 0.7613,
"step": 3875
},
{
"epoch": 0.7274090738657668,
"grad_norm": 0.03869959252720361,
"learning_rate": 6.294784866956757e-05,
"loss": 0.7525,
"step": 3880
},
{
"epoch": 0.7283464566929134,
"grad_norm": 0.038276776258313015,
"learning_rate": 6.254856315249736e-05,
"loss": 0.747,
"step": 3885
},
{
"epoch": 0.72928383952006,
"grad_norm": 0.035243252084036204,
"learning_rate": 6.21502141700579e-05,
"loss": 0.7306,
"step": 3890
},
{
"epoch": 0.7302212223472065,
"grad_norm": 0.03786773805086975,
"learning_rate": 6.175280598824678e-05,
"loss": 0.724,
"step": 3895
},
{
"epoch": 0.7311586051743532,
"grad_norm": 0.039123147712291625,
"learning_rate": 6.135634286298637e-05,
"loss": 0.7349,
"step": 3900
},
{
"epoch": 0.7320959880014998,
"grad_norm": 0.037251487752691576,
"learning_rate": 6.0960829040078265e-05,
"loss": 0.7345,
"step": 3905
},
{
"epoch": 0.7330333708286464,
"grad_norm": 0.03742284184866869,
"learning_rate": 6.05662687551576e-05,
"loss": 0.7795,
"step": 3910
},
{
"epoch": 0.733970753655793,
"grad_norm": 0.04105680206555956,
"learning_rate": 6.017266623364826e-05,
"loss": 0.7498,
"step": 3915
},
{
"epoch": 0.7349081364829396,
"grad_norm": 0.039477278781974275,
"learning_rate": 5.978002569071679e-05,
"loss": 0.713,
"step": 3920
},
{
"epoch": 0.7358455193100862,
"grad_norm": 0.03698403869400972,
"learning_rate": 5.938835133122821e-05,
"loss": 0.745,
"step": 3925
},
{
"epoch": 0.7367829021372329,
"grad_norm": 0.0410733277463145,
"learning_rate": 5.899764734970007e-05,
"loss": 0.7341,
"step": 3930
},
{
"epoch": 0.7377202849643795,
"grad_norm": 0.038917622759833966,
"learning_rate": 5.860791793025817e-05,
"loss": 0.7475,
"step": 3935
},
{
"epoch": 0.7386576677915261,
"grad_norm": 0.03627337453876378,
"learning_rate": 5.821916724659148e-05,
"loss": 0.746,
"step": 3940
},
{
"epoch": 0.7395950506186727,
"grad_norm": 0.03964503193315036,
"learning_rate": 5.783139946190751e-05,
"loss": 0.7398,
"step": 3945
},
{
"epoch": 0.7405324334458193,
"grad_norm": 0.03762868562695812,
"learning_rate": 5.744461872888771e-05,
"loss": 0.7152,
"step": 3950
},
{
"epoch": 0.7414698162729659,
"grad_norm": 0.03589488400010939,
"learning_rate": 5.705882918964299e-05,
"loss": 0.7461,
"step": 3955
},
{
"epoch": 0.7424071991001124,
"grad_norm": 0.03949112731558336,
"learning_rate": 5.667403497566939e-05,
"loss": 0.7344,
"step": 3960
},
{
"epoch": 0.743344581927259,
"grad_norm": 0.04151757065593634,
"learning_rate": 5.629024020780375e-05,
"loss": 0.7436,
"step": 3965
},
{
"epoch": 0.7442819647544057,
"grad_norm": 0.037059488736541316,
"learning_rate": 5.5907448996179766e-05,
"loss": 0.7502,
"step": 3970
},
{
"epoch": 0.7452193475815523,
"grad_norm": 0.03555904631512526,
"learning_rate": 5.552566544018373e-05,
"loss": 0.7465,
"step": 3975
},
{
"epoch": 0.7461567304086989,
"grad_norm": 0.036464378499175255,
"learning_rate": 5.514489362841083e-05,
"loss": 0.715,
"step": 3980
},
{
"epoch": 0.7470941132358455,
"grad_norm": 0.0394124136183908,
"learning_rate": 5.4765137638621246e-05,
"loss": 0.7613,
"step": 3985
},
{
"epoch": 0.7480314960629921,
"grad_norm": 0.03711650720168927,
"learning_rate": 5.4386401537696536e-05,
"loss": 0.7425,
"step": 3990
},
{
"epoch": 0.7489688788901387,
"grad_norm": 0.03888911993575119,
"learning_rate": 5.400868938159609e-05,
"loss": 0.7659,
"step": 3995
},
{
"epoch": 0.7499062617172854,
"grad_norm": 0.03744744409013629,
"learning_rate": 5.363200521531366e-05,
"loss": 0.7325,
"step": 4000
},
{
"epoch": 0.750843644544432,
"grad_norm": 0.03613327265209726,
"learning_rate": 5.3256353072833936e-05,
"loss": 0.7519,
"step": 4005
},
{
"epoch": 0.7517810273715786,
"grad_norm": 0.03595596503402885,
"learning_rate": 5.288173697708973e-05,
"loss": 0.7469,
"step": 4010
},
{
"epoch": 0.7527184101987252,
"grad_norm": 0.03769961152327502,
"learning_rate": 5.2508160939918286e-05,
"loss": 0.752,
"step": 4015
},
{
"epoch": 0.7536557930258717,
"grad_norm": 0.03633117479738773,
"learning_rate": 5.213562896201902e-05,
"loss": 0.7446,
"step": 4020
},
{
"epoch": 0.7545931758530183,
"grad_norm": 0.042037986286737154,
"learning_rate": 5.176414503290993e-05,
"loss": 0.7465,
"step": 4025
},
{
"epoch": 0.7555305586801649,
"grad_norm": 0.03964006330583541,
"learning_rate": 5.139371313088561e-05,
"loss": 0.7409,
"step": 4030
},
{
"epoch": 0.7564679415073116,
"grad_norm": 0.03557466108190626,
"learning_rate": 5.1024337222974125e-05,
"loss": 0.7482,
"step": 4035
},
{
"epoch": 0.7574053243344582,
"grad_norm": 0.03735795418886787,
"learning_rate": 5.065602126489453e-05,
"loss": 0.7429,
"step": 4040
},
{
"epoch": 0.7583427071616048,
"grad_norm": 0.03696761727309839,
"learning_rate": 5.028876920101504e-05,
"loss": 0.7217,
"step": 4045
},
{
"epoch": 0.7592800899887514,
"grad_norm": 0.03683633483444796,
"learning_rate": 4.992258496431002e-05,
"loss": 0.7415,
"step": 4050
},
{
"epoch": 0.760217472815898,
"grad_norm": 0.03527425447836841,
"learning_rate": 4.955747247631865e-05,
"loss": 0.7311,
"step": 4055
},
{
"epoch": 0.7611548556430446,
"grad_norm": 0.03674873764080841,
"learning_rate": 4.91934356471022e-05,
"loss": 0.7515,
"step": 4060
},
{
"epoch": 0.7620922384701913,
"grad_norm": 0.036177949800544354,
"learning_rate": 4.883047837520268e-05,
"loss": 0.7752,
"step": 4065
},
{
"epoch": 0.7630296212973379,
"grad_norm": 0.03907021802297722,
"learning_rate": 4.84686045476009e-05,
"loss": 0.728,
"step": 4070
},
{
"epoch": 0.7639670041244845,
"grad_norm": 0.037180348890681836,
"learning_rate": 4.810781803967482e-05,
"loss": 0.7221,
"step": 4075
},
{
"epoch": 0.7649043869516311,
"grad_norm": 0.038648597139017825,
"learning_rate": 4.7748122715158074e-05,
"loss": 0.7171,
"step": 4080
},
{
"epoch": 0.7658417697787776,
"grad_norm": 0.037736025165151156,
"learning_rate": 4.7389522426098614e-05,
"loss": 0.7521,
"step": 4085
},
{
"epoch": 0.7667791526059242,
"grad_norm": 0.03723459058036576,
"learning_rate": 4.703202101281744e-05,
"loss": 0.7352,
"step": 4090
},
{
"epoch": 0.7677165354330708,
"grad_norm": 0.03642827803946468,
"learning_rate": 4.667562230386749e-05,
"loss": 0.7449,
"step": 4095
},
{
"epoch": 0.7686539182602175,
"grad_norm": 0.036360477245182196,
"learning_rate": 4.63203301159926e-05,
"loss": 0.735,
"step": 4100
},
{
"epoch": 0.7695913010873641,
"grad_norm": 0.03758177450726638,
"learning_rate": 4.596614825408666e-05,
"loss": 0.7186,
"step": 4105
},
{
"epoch": 0.7705286839145107,
"grad_norm": 0.03748476699050447,
"learning_rate": 4.561308051115285e-05,
"loss": 0.757,
"step": 4110
},
{
"epoch": 0.7714660667416573,
"grad_norm": 0.036097069379836325,
"learning_rate": 4.5261130668263054e-05,
"loss": 0.743,
"step": 4115
},
{
"epoch": 0.7724034495688039,
"grad_norm": 0.03661609433595777,
"learning_rate": 4.4910302494517345e-05,
"loss": 0.727,
"step": 4120
},
{
"epoch": 0.7733408323959505,
"grad_norm": 0.036604261302591215,
"learning_rate": 4.456059974700361e-05,
"loss": 0.7198,
"step": 4125
},
{
"epoch": 0.7742782152230971,
"grad_norm": 0.0375410578618143,
"learning_rate": 4.4212026170757384e-05,
"loss": 0.7385,
"step": 4130
},
{
"epoch": 0.7752155980502438,
"grad_norm": 0.037797934383579375,
"learning_rate": 4.3864585498721445e-05,
"loss": 0.7375,
"step": 4135
},
{
"epoch": 0.7761529808773904,
"grad_norm": 0.036408551423671234,
"learning_rate": 4.3518281451706477e-05,
"loss": 0.7614,
"step": 4140
},
{
"epoch": 0.7770903637045369,
"grad_norm": 0.036659780564322314,
"learning_rate": 4.317311773835043e-05,
"loss": 0.7333,
"step": 4145
},
{
"epoch": 0.7780277465316835,
"grad_norm": 0.03787359723072369,
"learning_rate": 4.2829098055079524e-05,
"loss": 0.7321,
"step": 4150
},
{
"epoch": 0.7789651293588301,
"grad_norm": 0.039178508740307504,
"learning_rate": 4.248622608606802e-05,
"loss": 0.7568,
"step": 4155
},
{
"epoch": 0.7799025121859767,
"grad_norm": 0.03788147306371571,
"learning_rate": 4.214450550319943e-05,
"loss": 0.7002,
"step": 4160
},
{
"epoch": 0.7808398950131233,
"grad_norm": 0.03764951434893332,
"learning_rate": 4.180393996602651e-05,
"loss": 0.7483,
"step": 4165
},
{
"epoch": 0.78177727784027,
"grad_norm": 0.03643925778123276,
"learning_rate": 4.1464533121732613e-05,
"loss": 0.7608,
"step": 4170
},
{
"epoch": 0.7827146606674166,
"grad_norm": 0.03904458879716677,
"learning_rate": 4.112628860509238e-05,
"loss": 0.736,
"step": 4175
},
{
"epoch": 0.7836520434945632,
"grad_norm": 0.0367790048472788,
"learning_rate": 4.078921003843276e-05,
"loss": 0.7516,
"step": 4180
},
{
"epoch": 0.7845894263217098,
"grad_norm": 0.035606207842431264,
"learning_rate": 4.045330103159454e-05,
"loss": 0.7487,
"step": 4185
},
{
"epoch": 0.7855268091488564,
"grad_norm": 0.03778912384186634,
"learning_rate": 4.01185651818932e-05,
"loss": 0.7418,
"step": 4190
},
{
"epoch": 0.786464191976003,
"grad_norm": 0.03616645503819011,
"learning_rate": 3.97850060740808e-05,
"loss": 0.72,
"step": 4195
},
{
"epoch": 0.7874015748031497,
"grad_norm": 0.04137105544838257,
"learning_rate": 3.945262728030739e-05,
"loss": 0.7297,
"step": 4200
},
{
"epoch": 0.7883389576302963,
"grad_norm": 0.04008149649572554,
"learning_rate": 3.912143236008286e-05,
"loss": 0.7546,
"step": 4205
},
{
"epoch": 0.7892763404574428,
"grad_norm": 0.04178918635092829,
"learning_rate": 3.879142486023869e-05,
"loss": 0.705,
"step": 4210
},
{
"epoch": 0.7902137232845894,
"grad_norm": 0.03910230683910834,
"learning_rate": 3.8462608314890084e-05,
"loss": 0.7227,
"step": 4215
},
{
"epoch": 0.791151106111736,
"grad_norm": 0.041392117493789675,
"learning_rate": 3.8134986245398084e-05,
"loss": 0.7221,
"step": 4220
},
{
"epoch": 0.7920884889388826,
"grad_norm": 0.037713798352039335,
"learning_rate": 3.780856216033185e-05,
"loss": 0.7362,
"step": 4225
},
{
"epoch": 0.7930258717660292,
"grad_norm": 0.0386257393458269,
"learning_rate": 3.7483339555431055e-05,
"loss": 0.75,
"step": 4230
},
{
"epoch": 0.7939632545931758,
"grad_norm": 0.036597780629712674,
"learning_rate": 3.715932191356861e-05,
"loss": 0.7258,
"step": 4235
},
{
"epoch": 0.7949006374203225,
"grad_norm": 0.0351893354995898,
"learning_rate": 3.683651270471296e-05,
"loss": 0.7199,
"step": 4240
},
{
"epoch": 0.7958380202474691,
"grad_norm": 0.038598355578632684,
"learning_rate": 3.65149153858916e-05,
"loss": 0.7298,
"step": 4245
},
{
"epoch": 0.7967754030746157,
"grad_norm": 0.03440364370849841,
"learning_rate": 3.619453340115326e-05,
"loss": 0.7312,
"step": 4250
},
{
"epoch": 0.7977127859017623,
"grad_norm": 0.03802311815601217,
"learning_rate": 3.5875370181531754e-05,
"loss": 0.7424,
"step": 4255
},
{
"epoch": 0.7986501687289089,
"grad_norm": 0.03885312032393017,
"learning_rate": 3.555742914500867e-05,
"loss": 0.7447,
"step": 4260
},
{
"epoch": 0.7995875515560555,
"grad_norm": 0.03952348857049222,
"learning_rate": 3.5240713696477095e-05,
"loss": 0.7275,
"step": 4265
},
{
"epoch": 0.800524934383202,
"grad_norm": 0.03783212853790451,
"learning_rate": 3.4925227227705085e-05,
"loss": 0.7673,
"step": 4270
},
{
"epoch": 0.8014623172103487,
"grad_norm": 0.03855091891475635,
"learning_rate": 3.461097311729914e-05,
"loss": 0.7321,
"step": 4275
},
{
"epoch": 0.8023997000374953,
"grad_norm": 0.03853653138387215,
"learning_rate": 3.429795473066845e-05,
"loss": 0.7502,
"step": 4280
},
{
"epoch": 0.8033370828646419,
"grad_norm": 0.037938414011982106,
"learning_rate": 3.3986175419988326e-05,
"loss": 0.7827,
"step": 4285
},
{
"epoch": 0.8042744656917885,
"grad_norm": 0.03714608764447697,
"learning_rate": 3.367563852416484e-05,
"loss": 0.7504,
"step": 4290
},
{
"epoch": 0.8052118485189351,
"grad_norm": 0.037586754731944845,
"learning_rate": 3.336634736879857e-05,
"loss": 0.7412,
"step": 4295
},
{
"epoch": 0.8061492313460817,
"grad_norm": 0.03609469062292054,
"learning_rate": 3.3058305266149335e-05,
"loss": 0.7309,
"step": 4300
},
{
"epoch": 0.8070866141732284,
"grad_norm": 0.04157190953672981,
"learning_rate": 3.275151551510057e-05,
"loss": 0.7533,
"step": 4305
},
{
"epoch": 0.808023997000375,
"grad_norm": 0.03738613014861174,
"learning_rate": 3.2445981401124035e-05,
"loss": 0.7344,
"step": 4310
},
{
"epoch": 0.8089613798275216,
"grad_norm": 0.037333684702503485,
"learning_rate": 3.2141706196244646e-05,
"loss": 0.7313,
"step": 4315
},
{
"epoch": 0.8098987626546682,
"grad_norm": 0.03849398738947332,
"learning_rate": 3.183869315900537e-05,
"loss": 0.7321,
"step": 4320
},
{
"epoch": 0.8108361454818148,
"grad_norm": 0.03780735515283732,
"learning_rate": 3.153694553443241e-05,
"loss": 0.7604,
"step": 4325
},
{
"epoch": 0.8117735283089614,
"grad_norm": 0.03877151747469851,
"learning_rate": 3.1236466554000414e-05,
"loss": 0.7123,
"step": 4330
},
{
"epoch": 0.8127109111361079,
"grad_norm": 0.03868932391660062,
"learning_rate": 3.093725943559784e-05,
"loss": 0.7377,
"step": 4335
},
{
"epoch": 0.8136482939632546,
"grad_norm": 0.03748111742387308,
"learning_rate": 3.0639327383492544e-05,
"loss": 0.7171,
"step": 4340
},
{
"epoch": 0.8145856767904012,
"grad_norm": 0.03879381228880918,
"learning_rate": 3.0342673588297473e-05,
"loss": 0.7177,
"step": 4345
},
{
"epoch": 0.8155230596175478,
"grad_norm": 0.03783669509160595,
"learning_rate": 3.004730122693641e-05,
"loss": 0.7492,
"step": 4350
},
{
"epoch": 0.8164604424446944,
"grad_norm": 0.036123962271055896,
"learning_rate": 2.9753213462610077e-05,
"loss": 0.7327,
"step": 4355
},
{
"epoch": 0.817397825271841,
"grad_norm": 0.037712932440491655,
"learning_rate": 2.9460413444762143e-05,
"loss": 0.7382,
"step": 4360
},
{
"epoch": 0.8183352080989876,
"grad_norm": 0.03722083272698044,
"learning_rate": 2.9168904309045614e-05,
"loss": 0.7259,
"step": 4365
},
{
"epoch": 0.8192725909261342,
"grad_norm": 0.035672510126638225,
"learning_rate": 2.8878689177289005e-05,
"loss": 0.7282,
"step": 4370
},
{
"epoch": 0.8202099737532809,
"grad_norm": 0.03514557869239733,
"learning_rate": 2.8589771157463394e-05,
"loss": 0.741,
"step": 4375
},
{
"epoch": 0.8211473565804275,
"grad_norm": 0.03779365346411691,
"learning_rate": 2.8302153343648486e-05,
"loss": 0.7367,
"step": 4380
},
{
"epoch": 0.8220847394075741,
"grad_norm": 0.036530868365906646,
"learning_rate": 2.8015838816000168e-05,
"loss": 0.7395,
"step": 4385
},
{
"epoch": 0.8230221222347207,
"grad_norm": 0.03774672510896896,
"learning_rate": 2.773083064071685e-05,
"loss": 0.7439,
"step": 4390
},
{
"epoch": 0.8239595050618672,
"grad_norm": 0.03737530943067525,
"learning_rate": 2.7447131870007268e-05,
"loss": 0.7391,
"step": 4395
},
{
"epoch": 0.8248968878890138,
"grad_norm": 0.03589949691489075,
"learning_rate": 2.716474554205722e-05,
"loss": 0.7201,
"step": 4400
},
{
"epoch": 0.8258342707161604,
"grad_norm": 0.039108253157191145,
"learning_rate": 2.688367468099739e-05,
"loss": 0.715,
"step": 4405
},
{
"epoch": 0.8267716535433071,
"grad_norm": 0.040098561050262535,
"learning_rate": 2.6603922296871043e-05,
"loss": 0.7635,
"step": 4410
},
{
"epoch": 0.8277090363704537,
"grad_norm": 0.035538429924797096,
"learning_rate": 2.632549138560129e-05,
"loss": 0.7304,
"step": 4415
},
{
"epoch": 0.8286464191976003,
"grad_norm": 0.035183348616448194,
"learning_rate": 2.6048384928959653e-05,
"loss": 0.723,
"step": 4420
},
{
"epoch": 0.8295838020247469,
"grad_norm": 0.035542818739637466,
"learning_rate": 2.5772605894533533e-05,
"loss": 0.7513,
"step": 4425
},
{
"epoch": 0.8305211848518935,
"grad_norm": 0.040912090543415644,
"learning_rate": 2.5498157235694888e-05,
"loss": 0.7463,
"step": 4430
},
{
"epoch": 0.8314585676790401,
"grad_norm": 0.04115878743265954,
"learning_rate": 2.5225041891568366e-05,
"loss": 0.7421,
"step": 4435
},
{
"epoch": 0.8323959505061868,
"grad_norm": 0.03473020418203695,
"learning_rate": 2.4953262786999846e-05,
"loss": 0.7318,
"step": 4440
},
{
"epoch": 0.8333333333333334,
"grad_norm": 0.03945899992708612,
"learning_rate": 2.468282283252524e-05,
"loss": 0.7399,
"step": 4445
},
{
"epoch": 0.83427071616048,
"grad_norm": 0.041181932452618496,
"learning_rate": 2.4413724924339166e-05,
"loss": 0.7515,
"step": 4450
},
{
"epoch": 0.8352080989876266,
"grad_norm": 0.03757379419175963,
"learning_rate": 2.4145971944264025e-05,
"loss": 0.7572,
"step": 4455
},
{
"epoch": 0.8361454818147731,
"grad_norm": 0.03812065383102868,
"learning_rate": 2.3879566759719132e-05,
"loss": 0.7111,
"step": 4460
},
{
"epoch": 0.8370828646419197,
"grad_norm": 0.03758167524030431,
"learning_rate": 2.3614512223690002e-05,
"loss": 0.7408,
"step": 4465
},
{
"epoch": 0.8380202474690663,
"grad_norm": 0.0357989413923993,
"learning_rate": 2.335081117469777e-05,
"loss": 0.7404,
"step": 4470
},
{
"epoch": 0.838957630296213,
"grad_norm": 0.03719143556333986,
"learning_rate": 2.308846643676875e-05,
"loss": 0.7421,
"step": 4475
},
{
"epoch": 0.8398950131233596,
"grad_norm": 0.038416698176695595,
"learning_rate": 2.2827480819404386e-05,
"loss": 0.7391,
"step": 4480
},
{
"epoch": 0.8408323959505062,
"grad_norm": 0.038838405309790824,
"learning_rate": 2.2567857117550958e-05,
"loss": 0.7504,
"step": 4485
},
{
"epoch": 0.8417697787776528,
"grad_norm": 0.03678430153531787,
"learning_rate": 2.230959811156972e-05,
"loss": 0.721,
"step": 4490
},
{
"epoch": 0.8427071616047994,
"grad_norm": 0.037007576832510956,
"learning_rate": 2.2052706567207156e-05,
"loss": 0.7095,
"step": 4495
},
{
"epoch": 0.843644544431946,
"grad_norm": 0.03453973181827858,
"learning_rate": 2.179718523556531e-05,
"loss": 0.7229,
"step": 4500
},
{
"epoch": 0.8445819272590926,
"grad_norm": 0.037382201951380103,
"learning_rate": 2.1543036853072386e-05,
"loss": 0.7185,
"step": 4505
},
{
"epoch": 0.8455193100862393,
"grad_norm": 0.04001888522833685,
"learning_rate": 2.1290264141453313e-05,
"loss": 0.7342,
"step": 4510
},
{
"epoch": 0.8464566929133859,
"grad_norm": 0.035396516604348185,
"learning_rate": 2.103886980770085e-05,
"loss": 0.7128,
"step": 4515
},
{
"epoch": 0.8473940757405324,
"grad_norm": 0.03790943797877688,
"learning_rate": 2.0788856544046216e-05,
"loss": 0.7599,
"step": 4520
},
{
"epoch": 0.848331458567679,
"grad_norm": 0.03690950991272133,
"learning_rate": 2.0540227027930773e-05,
"loss": 0.7094,
"step": 4525
},
{
"epoch": 0.8492688413948256,
"grad_norm": 0.036299029934951134,
"learning_rate": 2.0292983921976753e-05,
"loss": 0.7197,
"step": 4530
},
{
"epoch": 0.8502062242219722,
"grad_norm": 0.036464577070665444,
"learning_rate": 2.004712987395924e-05,
"loss": 0.7306,
"step": 4535
},
{
"epoch": 0.8511436070491188,
"grad_norm": 0.035761152065985326,
"learning_rate": 1.9802667516777565e-05,
"loss": 0.7289,
"step": 4540
},
{
"epoch": 0.8520809898762655,
"grad_norm": 0.03812262175593615,
"learning_rate": 1.9559599468427183e-05,
"loss": 0.7406,
"step": 4545
},
{
"epoch": 0.8530183727034121,
"grad_norm": 0.03807115739595877,
"learning_rate": 1.9317928331971592e-05,
"loss": 0.7245,
"step": 4550
},
{
"epoch": 0.8539557555305587,
"grad_norm": 0.03647185658765888,
"learning_rate": 1.9077656695514526e-05,
"loss": 0.7348,
"step": 4555
},
{
"epoch": 0.8548931383577053,
"grad_norm": 0.037735068284022226,
"learning_rate": 1.8838787132172184e-05,
"loss": 0.7433,
"step": 4560
},
{
"epoch": 0.8558305211848519,
"grad_norm": 0.03411010074767823,
"learning_rate": 1.860132220004565e-05,
"loss": 0.7292,
"step": 4565
},
{
"epoch": 0.8567679040119985,
"grad_norm": 0.03512838788517134,
"learning_rate": 1.8365264442193618e-05,
"loss": 0.7125,
"step": 4570
},
{
"epoch": 0.8577052868391452,
"grad_norm": 0.03797586269699929,
"learning_rate": 1.8130616386604973e-05,
"loss": 0.7457,
"step": 4575
},
{
"epoch": 0.8586426696662918,
"grad_norm": 0.03846116349108381,
"learning_rate": 1.789738054617193e-05,
"loss": 0.7254,
"step": 4580
},
{
"epoch": 0.8595800524934383,
"grad_norm": 0.03248052807212059,
"learning_rate": 1.766555941866291e-05,
"loss": 0.7287,
"step": 4585
},
{
"epoch": 0.8605174353205849,
"grad_norm": 0.03749670548712012,
"learning_rate": 1.743515548669598e-05,
"loss": 0.7345,
"step": 4590
},
{
"epoch": 0.8614548181477315,
"grad_norm": 0.03868597238209571,
"learning_rate": 1.7206171217712135e-05,
"loss": 0.7387,
"step": 4595
},
{
"epoch": 0.8623922009748781,
"grad_norm": 0.03763605398447834,
"learning_rate": 1.6978609063948973e-05,
"loss": 0.7201,
"step": 4600
},
{
"epoch": 0.8633295838020247,
"grad_norm": 0.04009619133432012,
"learning_rate": 1.6752471462414226e-05,
"loss": 0.7384,
"step": 4605
},
{
"epoch": 0.8642669666291714,
"grad_norm": 0.03572102487743062,
"learning_rate": 1.6527760834860056e-05,
"loss": 0.7348,
"step": 4610
},
{
"epoch": 0.865204349456318,
"grad_norm": 0.038750256044430074,
"learning_rate": 1.6304479587756652e-05,
"loss": 0.7065,
"step": 4615
},
{
"epoch": 0.8661417322834646,
"grad_norm": 0.03675663032395201,
"learning_rate": 1.6082630112266888e-05,
"loss": 0.7077,
"step": 4620
},
{
"epoch": 0.8670791151106112,
"grad_norm": 0.03352806721702409,
"learning_rate": 1.5862214784220305e-05,
"loss": 0.7214,
"step": 4625
},
{
"epoch": 0.8680164979377578,
"grad_norm": 0.03651862777178404,
"learning_rate": 1.5643235964088064e-05,
"loss": 0.7158,
"step": 4630
},
{
"epoch": 0.8689538807649044,
"grad_norm": 0.03455566062476393,
"learning_rate": 1.5425695996957416e-05,
"loss": 0.736,
"step": 4635
},
{
"epoch": 0.869891263592051,
"grad_norm": 0.038034552465222866,
"learning_rate": 1.520959721250653e-05,
"loss": 0.735,
"step": 4640
},
{
"epoch": 0.8708286464191975,
"grad_norm": 0.03814002964874412,
"learning_rate": 1.4994941924979919e-05,
"loss": 0.725,
"step": 4645
},
{
"epoch": 0.8717660292463442,
"grad_norm": 0.03613963285066848,
"learning_rate": 1.4781732433163129e-05,
"loss": 0.7158,
"step": 4650
},
{
"epoch": 0.8727034120734908,
"grad_norm": 0.0391963633366358,
"learning_rate": 1.4569971020358656e-05,
"loss": 0.7043,
"step": 4655
},
{
"epoch": 0.8736407949006374,
"grad_norm": 0.03573970458333823,
"learning_rate": 1.4359659954360985e-05,
"loss": 0.7262,
"step": 4660
},
{
"epoch": 0.874578177727784,
"grad_norm": 0.03402185272612653,
"learning_rate": 1.4150801487432727e-05,
"loss": 0.726,
"step": 4665
},
{
"epoch": 0.8755155605549306,
"grad_norm": 0.03988457684610016,
"learning_rate": 1.394339785628027e-05,
"loss": 0.7116,
"step": 4670
},
{
"epoch": 0.8764529433820772,
"grad_norm": 0.03749913331479699,
"learning_rate": 1.373745128202986e-05,
"loss": 0.7486,
"step": 4675
},
{
"epoch": 0.8773903262092239,
"grad_norm": 0.039418943387831605,
"learning_rate": 1.3532963970203848e-05,
"loss": 0.7311,
"step": 4680
},
{
"epoch": 0.8783277090363705,
"grad_norm": 0.03691545995561202,
"learning_rate": 1.332993811069708e-05,
"loss": 0.7464,
"step": 4685
},
{
"epoch": 0.8792650918635171,
"grad_norm": 0.03634444663432465,
"learning_rate": 1.3128375877753393e-05,
"loss": 0.7048,
"step": 4690
},
{
"epoch": 0.8802024746906637,
"grad_norm": 0.03883997615492241,
"learning_rate": 1.2928279429942362e-05,
"loss": 0.743,
"step": 4695
},
{
"epoch": 0.8811398575178103,
"grad_norm": 0.03831787223481465,
"learning_rate": 1.2729650910136196e-05,
"loss": 0.72,
"step": 4700
},
{
"epoch": 0.8820772403449568,
"grad_norm": 0.038782793377748884,
"learning_rate": 1.2532492445486769e-05,
"loss": 0.7556,
"step": 4705
},
{
"epoch": 0.8830146231721034,
"grad_norm": 0.035930508758284546,
"learning_rate": 1.2336806147402828e-05,
"loss": 0.7359,
"step": 4710
},
{
"epoch": 0.8839520059992501,
"grad_norm": 0.03688784164900098,
"learning_rate": 1.2142594111527415e-05,
"loss": 0.7322,
"step": 4715
},
{
"epoch": 0.8848893888263967,
"grad_norm": 0.03418578395324309,
"learning_rate": 1.1949858417715418e-05,
"loss": 0.7145,
"step": 4720
},
{
"epoch": 0.8858267716535433,
"grad_norm": 0.03839104854863887,
"learning_rate": 1.1758601130011259e-05,
"loss": 0.708,
"step": 4725
},
{
"epoch": 0.8867641544806899,
"grad_norm": 0.03690073415729465,
"learning_rate": 1.1568824296626866e-05,
"loss": 0.7268,
"step": 4730
},
{
"epoch": 0.8877015373078365,
"grad_norm": 0.03785506282627266,
"learning_rate": 1.1380529949919593e-05,
"loss": 0.7472,
"step": 4735
},
{
"epoch": 0.8886389201349831,
"grad_norm": 0.03505149905088435,
"learning_rate": 1.1193720106370701e-05,
"loss": 0.735,
"step": 4740
},
{
"epoch": 0.8895763029621298,
"grad_norm": 0.038177240967898714,
"learning_rate": 1.100839676656346e-05,
"loss": 0.7181,
"step": 4745
},
{
"epoch": 0.8905136857892764,
"grad_norm": 0.03798215847626741,
"learning_rate": 1.0824561915162016e-05,
"loss": 0.719,
"step": 4750
},
{
"epoch": 0.891451068616423,
"grad_norm": 0.03757317113276994,
"learning_rate": 1.0642217520889873e-05,
"loss": 0.7412,
"step": 4755
},
{
"epoch": 0.8923884514435696,
"grad_norm": 0.03762594674374493,
"learning_rate": 1.0461365536509065e-05,
"loss": 0.7365,
"step": 4760
},
{
"epoch": 0.8933258342707162,
"grad_norm": 0.03752202520507965,
"learning_rate": 1.0282007898798995e-05,
"loss": 0.7146,
"step": 4765
},
{
"epoch": 0.8942632170978627,
"grad_norm": 0.031040074178325168,
"learning_rate": 1.01041465285359e-05,
"loss": 0.7121,
"step": 4770
},
{
"epoch": 0.8952005999250093,
"grad_norm": 0.03608189060059074,
"learning_rate": 9.927783330472139e-06,
"loss": 0.7328,
"step": 4775
},
{
"epoch": 0.896137982752156,
"grad_norm": 0.03589888025539565,
"learning_rate": 9.752920193315865e-06,
"loss": 0.6956,
"step": 4780
},
{
"epoch": 0.8970753655793026,
"grad_norm": 0.03503081744003102,
"learning_rate": 9.579558989710872e-06,
"loss": 0.734,
"step": 4785
},
{
"epoch": 0.8980127484064492,
"grad_norm": 0.036423527513909,
"learning_rate": 9.407701576216281e-06,
"loss": 0.7139,
"step": 4790
},
{
"epoch": 0.8989501312335958,
"grad_norm": 0.03773162502672724,
"learning_rate": 9.237349793286963e-06,
"loss": 0.7286,
"step": 4795
},
{
"epoch": 0.8998875140607424,
"grad_norm": 0.0370655727344935,
"learning_rate": 9.068505465253656e-06,
"loss": 0.7364,
"step": 4800
},
{
"epoch": 0.900824896887889,
"grad_norm": 0.03793992378742871,
"learning_rate": 8.901170400303443e-06,
"loss": 0.7117,
"step": 4805
},
{
"epoch": 0.9017622797150356,
"grad_norm": 0.03749076700190644,
"learning_rate": 8.735346390460452e-06,
"loss": 0.7482,
"step": 4810
},
{
"epoch": 0.9026996625421823,
"grad_norm": 0.03478713094360769,
"learning_rate": 8.571035211566606e-06,
"loss": 0.7282,
"step": 4815
},
{
"epoch": 0.9036370453693289,
"grad_norm": 0.03931631750725771,
"learning_rate": 8.408238623262625e-06,
"loss": 0.7347,
"step": 4820
},
{
"epoch": 0.9045744281964755,
"grad_norm": 0.03832734319910059,
"learning_rate": 8.246958368969164e-06,
"loss": 0.7085,
"step": 4825
},
{
"epoch": 0.905511811023622,
"grad_norm": 0.036342961917758904,
"learning_rate": 8.087196175868204e-06,
"loss": 0.7459,
"step": 4830
},
{
"epoch": 0.9064491938507686,
"grad_norm": 0.03811297614516944,
"learning_rate": 7.928953754884482e-06,
"loss": 0.7245,
"step": 4835
},
{
"epoch": 0.9073865766779152,
"grad_norm": 0.03617635523056376,
"learning_rate": 7.772232800667117e-06,
"loss": 0.6879,
"step": 4840
},
{
"epoch": 0.9083239595050618,
"grad_norm": 0.03686080577213026,
"learning_rate": 7.617034991571747e-06,
"loss": 0.7308,
"step": 4845
},
{
"epoch": 0.9092613423322085,
"grad_norm": 0.03468911090610316,
"learning_rate": 7.463361989642108e-06,
"loss": 0.7266,
"step": 4850
},
{
"epoch": 0.9101987251593551,
"grad_norm": 0.03837314601114124,
"learning_rate": 7.311215440592649e-06,
"loss": 0.7244,
"step": 4855
},
{
"epoch": 0.9111361079865017,
"grad_norm": 0.03515020181576783,
"learning_rate": 7.160596973790678e-06,
"loss": 0.7095,
"step": 4860
},
{
"epoch": 0.9120734908136483,
"grad_norm": 0.04067375872375338,
"learning_rate": 7.011508202238958e-06,
"loss": 0.7565,
"step": 4865
},
{
"epoch": 0.9130108736407949,
"grad_norm": 0.037896195508876414,
"learning_rate": 6.8639507225584755e-06,
"loss": 0.7479,
"step": 4870
},
{
"epoch": 0.9139482564679415,
"grad_norm": 0.035925556435556304,
"learning_rate": 6.7179261149712335e-06,
"loss": 0.7287,
"step": 4875
},
{
"epoch": 0.9148856392950881,
"grad_norm": 0.03437946260531882,
"learning_rate": 6.57343594328355e-06,
"loss": 0.7154,
"step": 4880
},
{
"epoch": 0.9158230221222348,
"grad_norm": 0.036848739384731176,
"learning_rate": 6.430481754868988e-06,
"loss": 0.7306,
"step": 4885
},
{
"epoch": 0.9167604049493814,
"grad_norm": 0.03819174514800313,
"learning_rate": 6.289065080652134e-06,
"loss": 0.7423,
"step": 4890
},
{
"epoch": 0.9176977877765279,
"grad_norm": 0.04015640945147137,
"learning_rate": 6.149187435091912e-06,
"loss": 0.7388,
"step": 4895
},
{
"epoch": 0.9186351706036745,
"grad_norm": 0.0354113100153628,
"learning_rate": 6.010850316165533e-06,
"loss": 0.7153,
"step": 4900
},
{
"epoch": 0.9195725534308211,
"grad_norm": 0.03619736356277513,
"learning_rate": 5.8740552053524185e-06,
"loss": 0.7123,
"step": 4905
},
{
"epoch": 0.9205099362579677,
"grad_norm": 0.03357169683122578,
"learning_rate": 5.7388035676183e-06,
"loss": 0.7367,
"step": 4910
},
{
"epoch": 0.9214473190851143,
"grad_norm": 0.035341368116608526,
"learning_rate": 5.6050968513995484e-06,
"loss": 0.735,
"step": 4915
},
{
"epoch": 0.922384701912261,
"grad_norm": 0.03463388384208746,
"learning_rate": 5.472936488587687e-06,
"loss": 0.7045,
"step": 4920
},
{
"epoch": 0.9233220847394076,
"grad_norm": 0.03646756098895019,
"learning_rate": 5.342323894514017e-06,
"loss": 0.692,
"step": 4925
},
{
"epoch": 0.9242594675665542,
"grad_norm": 0.0378552791061763,
"learning_rate": 5.213260467934499e-06,
"loss": 0.7308,
"step": 4930
},
{
"epoch": 0.9251968503937008,
"grad_norm": 0.03651837390186468,
"learning_rate": 5.085747591014716e-06,
"loss": 0.7241,
"step": 4935
},
{
"epoch": 0.9261342332208474,
"grad_norm": 0.03496033002413129,
"learning_rate": 4.959786629315166e-06,
"loss": 0.7201,
"step": 4940
},
{
"epoch": 0.927071616047994,
"grad_norm": 0.03807303842127231,
"learning_rate": 4.83537893177654e-06,
"loss": 0.7228,
"step": 4945
},
{
"epoch": 0.9280089988751407,
"grad_norm": 0.03388630160185748,
"learning_rate": 4.712525830705338e-06,
"loss": 0.7255,
"step": 4950
},
{
"epoch": 0.9289463817022872,
"grad_norm": 0.03359401994529555,
"learning_rate": 4.591228641759559e-06,
"loss": 0.7145,
"step": 4955
},
{
"epoch": 0.9298837645294338,
"grad_norm": 0.03790506150006882,
"learning_rate": 4.471488663934647e-06,
"loss": 0.7182,
"step": 4960
},
{
"epoch": 0.9308211473565804,
"grad_norm": 0.03717261368444523,
"learning_rate": 4.3533071795496035e-06,
"loss": 0.7423,
"step": 4965
},
{
"epoch": 0.931758530183727,
"grad_norm": 0.03725676902347381,
"learning_rate": 4.236685454233113e-06,
"loss": 0.7384,
"step": 4970
},
{
"epoch": 0.9326959130108736,
"grad_norm": 0.03790059729911343,
"learning_rate": 4.12162473691024e-06,
"loss": 0.7408,
"step": 4975
},
{
"epoch": 0.9336332958380202,
"grad_norm": 0.03917020457274146,
"learning_rate": 4.008126259788752e-06,
"loss": 0.7203,
"step": 4980
},
{
"epoch": 0.9345706786651669,
"grad_norm": 0.037051465451931255,
"learning_rate": 3.896191238346219e-06,
"loss": 0.7295,
"step": 4985
},
{
"epoch": 0.9355080614923135,
"grad_norm": 0.037826370662724866,
"learning_rate": 3.785820871316736e-06,
"loss": 0.7087,
"step": 4990
},
{
"epoch": 0.9364454443194601,
"grad_norm": 0.03734181026651175,
"learning_rate": 3.677016340678318e-06,
"loss": 0.7244,
"step": 4995
},
{
"epoch": 0.9373828271466067,
"grad_norm": 0.03568772503256821,
"learning_rate": 3.5697788116400283e-06,
"loss": 0.7204,
"step": 5000
},
{
"epoch": 0.9383202099737533,
"grad_norm": 0.03537228998662857,
"learning_rate": 3.4641094326296524e-06,
"loss": 0.7089,
"step": 5005
},
{
"epoch": 0.9392575928008999,
"grad_norm": 0.037141105148803415,
"learning_rate": 3.3600093352814107e-06,
"loss": 0.7329,
"step": 5010
},
{
"epoch": 0.9401949756280465,
"grad_norm": 0.035406141374499524,
"learning_rate": 3.2574796344236153e-06,
"loss": 0.7033,
"step": 5015
},
{
"epoch": 0.941132358455193,
"grad_norm": 0.041168173265274895,
"learning_rate": 3.1565214280670825e-06,
"loss": 0.7038,
"step": 5020
},
{
"epoch": 0.9420697412823397,
"grad_norm": 0.039723554038275884,
"learning_rate": 3.0571357973930234e-06,
"loss": 0.7097,
"step": 5025
},
{
"epoch": 0.9430071241094863,
"grad_norm": 0.032544537124457665,
"learning_rate": 2.959323806741737e-06,
"loss": 0.7272,
"step": 5030
},
{
"epoch": 0.9439445069366329,
"grad_norm": 0.03461405648425469,
"learning_rate": 2.8630865036010364e-06,
"loss": 0.7141,
"step": 5035
},
{
"epoch": 0.9448818897637795,
"grad_norm": 0.036824105839646924,
"learning_rate": 2.7684249185951415e-06,
"loss": 0.7335,
"step": 5040
},
{
"epoch": 0.9458192725909261,
"grad_norm": 0.04002092643887271,
"learning_rate": 2.6753400654735524e-06,
"loss": 0.7372,
"step": 5045
},
{
"epoch": 0.9467566554180727,
"grad_norm": 0.03570329291792829,
"learning_rate": 2.5838329411002943e-06,
"loss": 0.7127,
"step": 5050
},
{
"epoch": 0.9476940382452194,
"grad_norm": 0.034274985857504735,
"learning_rate": 2.493904525443141e-06,
"loss": 0.7057,
"step": 5055
},
{
"epoch": 0.948631421072366,
"grad_norm": 0.03801311689032566,
"learning_rate": 2.405555781563173e-06,
"loss": 0.7282,
"step": 5060
},
{
"epoch": 0.9495688038995126,
"grad_norm": 0.03756145522959914,
"learning_rate": 2.3187876556044537e-06,
"loss": 0.6886,
"step": 5065
},
{
"epoch": 0.9505061867266592,
"grad_norm": 0.03530134077170944,
"learning_rate": 2.2336010767839194e-06,
"loss": 0.7168,
"step": 5070
},
{
"epoch": 0.9514435695538058,
"grad_norm": 0.03876854731664466,
"learning_rate": 2.1499969573813724e-06,
"loss": 0.7163,
"step": 5075
},
{
"epoch": 0.9523809523809523,
"grad_norm": 0.03235204006913281,
"learning_rate": 2.06797619272977e-06,
"loss": 0.7339,
"step": 5080
},
{
"epoch": 0.9533183352080989,
"grad_norm": 0.035560439311177215,
"learning_rate": 1.9875396612056005e-06,
"loss": 0.7356,
"step": 5085
},
{
"epoch": 0.9542557180352456,
"grad_norm": 0.03566397362898293,
"learning_rate": 1.9086882242195235e-06,
"loss": 0.6973,
"step": 5090
},
{
"epoch": 0.9551931008623922,
"grad_norm": 0.037624367410972304,
"learning_rate": 1.8314227262070435e-06,
"loss": 0.7273,
"step": 5095
},
{
"epoch": 0.9561304836895388,
"grad_norm": 0.03441332414126222,
"learning_rate": 1.7557439946196017e-06,
"loss": 0.722,
"step": 5100
},
{
"epoch": 0.9570678665166854,
"grad_norm": 0.03664904948143634,
"learning_rate": 1.6816528399155982e-06,
"loss": 0.7412,
"step": 5105
},
{
"epoch": 0.958005249343832,
"grad_norm": 0.03715182226165275,
"learning_rate": 1.60915005555175e-06,
"loss": 0.7232,
"step": 5110
},
{
"epoch": 0.9589426321709786,
"grad_norm": 0.03639187438599698,
"learning_rate": 1.5382364179746808e-06,
"loss": 0.737,
"step": 5115
},
{
"epoch": 0.9598800149981253,
"grad_norm": 0.03868780676867553,
"learning_rate": 1.4689126866124278e-06,
"loss": 0.7097,
"step": 5120
},
{
"epoch": 0.9608173978252719,
"grad_norm": 0.036291056043944926,
"learning_rate": 1.4011796038665145e-06,
"loss": 0.7186,
"step": 5125
},
{
"epoch": 0.9617547806524185,
"grad_norm": 0.06821587616968404,
"learning_rate": 1.335037895103791e-06,
"loss": 0.7374,
"step": 5130
},
{
"epoch": 0.9626921634795651,
"grad_norm": 0.03887971150133188,
"learning_rate": 1.2704882686488393e-06,
"loss": 0.7624,
"step": 5135
},
{
"epoch": 0.9636295463067117,
"grad_norm": 0.036497584321372616,
"learning_rate": 1.2075314157762972e-06,
"loss": 0.718,
"step": 5140
},
{
"epoch": 0.9645669291338582,
"grad_norm": 0.039035533336186354,
"learning_rate": 1.1461680107034798e-06,
"loss": 0.7421,
"step": 5145
},
{
"epoch": 0.9655043119610048,
"grad_norm": 0.035839951992695206,
"learning_rate": 1.0863987105831696e-06,
"loss": 0.73,
"step": 5150
},
{
"epoch": 0.9664416947881515,
"grad_norm": 0.03563901861543763,
"learning_rate": 1.0282241554965375e-06,
"loss": 0.7314,
"step": 5155
},
{
"epoch": 0.9673790776152981,
"grad_norm": 0.034948413836722134,
"learning_rate": 9.716449684463502e-07,
"loss": 0.7079,
"step": 5160
},
{
"epoch": 0.9683164604424447,
"grad_norm": 0.037670482239960106,
"learning_rate": 9.166617553502064e-07,
"loss": 0.7209,
"step": 5165
},
{
"epoch": 0.9692538432695913,
"grad_norm": 0.03868203643112819,
"learning_rate": 8.632751050341946e-07,
"loss": 0.7267,
"step": 5170
},
{
"epoch": 0.9701912260967379,
"grad_norm": 0.03685196650404434,
"learning_rate": 8.114855892264128e-07,
"loss": 0.7288,
"step": 5175
},
{
"epoch": 0.9711286089238845,
"grad_norm": 0.03480760162856044,
"learning_rate": 7.612937625509741e-07,
"loss": 0.714,
"step": 5180
},
{
"epoch": 0.9720659917510311,
"grad_norm": 0.035112246544013306,
"learning_rate": 7.127001625220286e-07,
"loss": 0.725,
"step": 5185
},
{
"epoch": 0.9730033745781778,
"grad_norm": 0.034016779690611565,
"learning_rate": 6.657053095380005e-07,
"loss": 0.6879,
"step": 5190
},
{
"epoch": 0.9739407574053244,
"grad_norm": 0.03711659986007524,
"learning_rate": 6.203097068759933e-07,
"loss": 0.7412,
"step": 5195
},
{
"epoch": 0.974878140232471,
"grad_norm": 0.039330504751635645,
"learning_rate": 5.765138406864434e-07,
"loss": 0.7359,
"step": 5200
},
{
"epoch": 0.9758155230596175,
"grad_norm": 0.03613689473953272,
"learning_rate": 5.343181799878916e-07,
"loss": 0.7249,
"step": 5205
},
{
"epoch": 0.9767529058867641,
"grad_norm": 0.03766304326222476,
"learning_rate": 4.937231766619698e-07,
"loss": 0.7141,
"step": 5210
},
{
"epoch": 0.9776902887139107,
"grad_norm": 0.03797367873629258,
"learning_rate": 4.547292654485557e-07,
"loss": 0.7325,
"step": 5215
},
{
"epoch": 0.9786276715410573,
"grad_norm": 0.038696832388832716,
"learning_rate": 4.1733686394109236e-07,
"loss": 0.735,
"step": 5220
},
{
"epoch": 0.979565054368204,
"grad_norm": 0.03372024341796256,
"learning_rate": 3.815463725821755e-07,
"loss": 0.7132,
"step": 5225
},
{
"epoch": 0.9805024371953506,
"grad_norm": 0.03911193375759219,
"learning_rate": 3.473581746592069e-07,
"loss": 0.7413,
"step": 5230
},
{
"epoch": 0.9814398200224972,
"grad_norm": 0.03595612615841676,
"learning_rate": 3.1477263630033113e-07,
"loss": 0.7263,
"step": 5235
},
{
"epoch": 0.9823772028496438,
"grad_norm": 0.035577055944527154,
"learning_rate": 2.8379010647045506e-07,
"loss": 0.724,
"step": 5240
},
{
"epoch": 0.9833145856767904,
"grad_norm": 0.03939881083971358,
"learning_rate": 2.5441091696761783e-07,
"loss": 0.7345,
"step": 5245
},
{
"epoch": 0.984251968503937,
"grad_norm": 0.03509258626104014,
"learning_rate": 2.266353824193101e-07,
"loss": 0.689,
"step": 5250
},
{
"epoch": 0.9851893513310837,
"grad_norm": 0.03642207354018445,
"learning_rate": 2.0046380027921028e-07,
"loss": 0.7582,
"step": 5255
},
{
"epoch": 0.9861267341582303,
"grad_norm": 0.03655317102724065,
"learning_rate": 1.7589645082392024e-07,
"loss": 0.728,
"step": 5260
},
{
"epoch": 0.9870641169853769,
"grad_norm": 0.03681201840565274,
"learning_rate": 1.529335971500345e-07,
"loss": 0.7029,
"step": 5265
},
{
"epoch": 0.9880014998125234,
"grad_norm": 0.034195294045265584,
"learning_rate": 1.315754851712425e-07,
"loss": 0.7106,
"step": 5270
},
{
"epoch": 0.98893888263967,
"grad_norm": 0.033879744761242093,
"learning_rate": 1.1182234361579722e-07,
"loss": 0.732,
"step": 5275
},
{
"epoch": 0.9898762654668166,
"grad_norm": 0.036906203150992155,
"learning_rate": 9.367438402395066e-08,
"loss": 0.7243,
"step": 5280
},
{
"epoch": 0.9908136482939632,
"grad_norm": 0.03751280991426079,
"learning_rate": 7.713180074577219e-08,
"loss": 0.7429,
"step": 5285
},
{
"epoch": 0.9917510311211098,
"grad_norm": 0.03933888372810454,
"learning_rate": 6.219477093905023e-08,
"loss": 0.7368,
"step": 5290
},
{
"epoch": 0.9926884139482565,
"grad_norm": 0.03920435371912244,
"learning_rate": 4.8863454567360513e-08,
"loss": 0.7415,
"step": 5295
},
{
"epoch": 0.9936257967754031,
"grad_norm": 0.0340028517083224,
"learning_rate": 3.7137994398400705e-08,
"loss": 0.736,
"step": 5300
},
{
"epoch": 0.9945631796025497,
"grad_norm": 0.03596691936417686,
"learning_rate": 2.7018516002424996e-08,
"loss": 0.7027,
"step": 5305
},
{
"epoch": 0.9955005624296963,
"grad_norm": 0.035041534347360945,
"learning_rate": 1.850512775091184e-08,
"loss": 0.7338,
"step": 5310
},
{
"epoch": 0.9964379452568429,
"grad_norm": 0.036327963873331384,
"learning_rate": 1.1597920815414885e-08,
"loss": 0.7262,
"step": 5315
},
{
"epoch": 0.9973753280839895,
"grad_norm": 0.036279192100004604,
"learning_rate": 6.296969166580401e-09,
"loss": 0.7153,
"step": 5320
},
{
"epoch": 0.9983127109111362,
"grad_norm": 0.035186584742647825,
"learning_rate": 2.6023295733312855e-09,
"loss": 0.7068,
"step": 5325
},
{
"epoch": 0.9992500937382827,
"grad_norm": 0.036172737685700306,
"learning_rate": 5.140416022841875e-10,
"loss": 0.7243,
"step": 5330
},
{
"epoch": 1.0,
"eval_loss": 1.0353564023971558,
"eval_runtime": 1193.1467,
"eval_samples_per_second": 196.235,
"eval_steps_per_second": 6.133,
"step": 5334
},
{
"epoch": 1.0,
"step": 5334,
"total_flos": 777591764811776.0,
"train_loss": 0.7984603145035815,
"train_runtime": 21845.7301,
"train_samples_per_second": 31.253,
"train_steps_per_second": 0.244
}
],
"logging_steps": 5,
"max_steps": 5334,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 777591764811776.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}