oh_scale_x.5_compute_equal / trainer_state.json
sedrickkeh's picture
End of training
4b8ae81 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 24.868123587038433,
"eval_steps": 500,
"global_step": 4125,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.06028636021100226,
"grad_norm": 14.353901908542692,
"learning_rate": 5e-06,
"loss": 1.0595,
"step": 10
},
{
"epoch": 0.12057272042200452,
"grad_norm": 1.3180865308999647,
"learning_rate": 5e-06,
"loss": 0.9403,
"step": 20
},
{
"epoch": 0.1808590806330068,
"grad_norm": 0.9782137323160548,
"learning_rate": 5e-06,
"loss": 0.89,
"step": 30
},
{
"epoch": 0.24114544084400905,
"grad_norm": 1.3518217541449848,
"learning_rate": 5e-06,
"loss": 0.8618,
"step": 40
},
{
"epoch": 0.30143180105501133,
"grad_norm": 0.8934828143592951,
"learning_rate": 5e-06,
"loss": 0.8447,
"step": 50
},
{
"epoch": 0.3617181612660136,
"grad_norm": 0.9919016971761021,
"learning_rate": 5e-06,
"loss": 0.8318,
"step": 60
},
{
"epoch": 0.42200452147701584,
"grad_norm": 0.9227497685306041,
"learning_rate": 5e-06,
"loss": 0.8177,
"step": 70
},
{
"epoch": 0.4822908816880181,
"grad_norm": 0.9363087183332998,
"learning_rate": 5e-06,
"loss": 0.8106,
"step": 80
},
{
"epoch": 0.5425772418990203,
"grad_norm": 1.0736624468965914,
"learning_rate": 5e-06,
"loss": 0.8052,
"step": 90
},
{
"epoch": 0.6028636021100227,
"grad_norm": 1.353564519927803,
"learning_rate": 5e-06,
"loss": 0.7907,
"step": 100
},
{
"epoch": 0.6631499623210249,
"grad_norm": 1.0455405570196998,
"learning_rate": 5e-06,
"loss": 0.7913,
"step": 110
},
{
"epoch": 0.7234363225320272,
"grad_norm": 0.9569192147667731,
"learning_rate": 5e-06,
"loss": 0.7828,
"step": 120
},
{
"epoch": 0.7837226827430294,
"grad_norm": 0.8298672277424939,
"learning_rate": 5e-06,
"loss": 0.7841,
"step": 130
},
{
"epoch": 0.8440090429540317,
"grad_norm": 1.2437896161299962,
"learning_rate": 5e-06,
"loss": 0.7832,
"step": 140
},
{
"epoch": 0.9042954031650339,
"grad_norm": 0.6290099294011746,
"learning_rate": 5e-06,
"loss": 0.7818,
"step": 150
},
{
"epoch": 0.9645817633760362,
"grad_norm": 0.683050547264518,
"learning_rate": 5e-06,
"loss": 0.7807,
"step": 160
},
{
"epoch": 0.9947249434815373,
"eval_loss": 0.7690147161483765,
"eval_runtime": 114.9859,
"eval_samples_per_second": 38.866,
"eval_steps_per_second": 0.609,
"step": 165
},
{
"epoch": 1.0248681235870385,
"grad_norm": 0.8375316338814506,
"learning_rate": 5e-06,
"loss": 0.7706,
"step": 170
},
{
"epoch": 1.0851544837980407,
"grad_norm": 0.8491856111019711,
"learning_rate": 5e-06,
"loss": 0.7207,
"step": 180
},
{
"epoch": 1.145440844009043,
"grad_norm": 1.00264106211233,
"learning_rate": 5e-06,
"loss": 0.7264,
"step": 190
},
{
"epoch": 1.2057272042200453,
"grad_norm": 0.9524591725477187,
"learning_rate": 5e-06,
"loss": 0.7223,
"step": 200
},
{
"epoch": 1.2660135644310475,
"grad_norm": 0.8200876998100735,
"learning_rate": 5e-06,
"loss": 0.7237,
"step": 210
},
{
"epoch": 1.3262999246420497,
"grad_norm": 0.6972921329207209,
"learning_rate": 5e-06,
"loss": 0.7233,
"step": 220
},
{
"epoch": 1.3865862848530521,
"grad_norm": 1.3472894348419586,
"learning_rate": 5e-06,
"loss": 0.7175,
"step": 230
},
{
"epoch": 1.4468726450640543,
"grad_norm": 0.8080326013334395,
"learning_rate": 5e-06,
"loss": 0.7203,
"step": 240
},
{
"epoch": 1.5071590052750565,
"grad_norm": 0.6690842132293575,
"learning_rate": 5e-06,
"loss": 0.7184,
"step": 250
},
{
"epoch": 1.5674453654860587,
"grad_norm": 0.6680726643782731,
"learning_rate": 5e-06,
"loss": 0.7193,
"step": 260
},
{
"epoch": 1.627731725697061,
"grad_norm": 0.6582407897638277,
"learning_rate": 5e-06,
"loss": 0.7171,
"step": 270
},
{
"epoch": 1.6880180859080633,
"grad_norm": 0.6058749327234759,
"learning_rate": 5e-06,
"loss": 0.718,
"step": 280
},
{
"epoch": 1.7483044461190655,
"grad_norm": 0.6771950284954183,
"learning_rate": 5e-06,
"loss": 0.717,
"step": 290
},
{
"epoch": 1.8085908063300677,
"grad_norm": 0.620278984859575,
"learning_rate": 5e-06,
"loss": 0.721,
"step": 300
},
{
"epoch": 1.8688771665410702,
"grad_norm": 0.7376417188133093,
"learning_rate": 5e-06,
"loss": 0.7187,
"step": 310
},
{
"epoch": 1.9291635267520724,
"grad_norm": 0.841815742734191,
"learning_rate": 5e-06,
"loss": 0.7168,
"step": 320
},
{
"epoch": 1.9894498869630746,
"grad_norm": 0.9694154154724743,
"learning_rate": 5e-06,
"loss": 0.7139,
"step": 330
},
{
"epoch": 1.995478522984175,
"eval_loss": 0.7520068287849426,
"eval_runtime": 112.311,
"eval_samples_per_second": 39.791,
"eval_steps_per_second": 0.623,
"step": 331
},
{
"epoch": 2.049736247174077,
"grad_norm": 1.3857092066503702,
"learning_rate": 5e-06,
"loss": 0.6895,
"step": 340
},
{
"epoch": 2.110022607385079,
"grad_norm": 0.9308939279408932,
"learning_rate": 5e-06,
"loss": 0.661,
"step": 350
},
{
"epoch": 2.1703089675960814,
"grad_norm": 1.7049059921843441,
"learning_rate": 5e-06,
"loss": 0.6596,
"step": 360
},
{
"epoch": 2.230595327807084,
"grad_norm": 1.1664692858604568,
"learning_rate": 5e-06,
"loss": 0.6626,
"step": 370
},
{
"epoch": 2.290881688018086,
"grad_norm": 0.8099811209522968,
"learning_rate": 5e-06,
"loss": 0.6609,
"step": 380
},
{
"epoch": 2.351168048229088,
"grad_norm": 0.7769380652338045,
"learning_rate": 5e-06,
"loss": 0.6633,
"step": 390
},
{
"epoch": 2.4114544084400906,
"grad_norm": 1.0202300923789958,
"learning_rate": 5e-06,
"loss": 0.667,
"step": 400
},
{
"epoch": 2.4717407686510926,
"grad_norm": 0.9120675026610565,
"learning_rate": 5e-06,
"loss": 0.6654,
"step": 410
},
{
"epoch": 2.532027128862095,
"grad_norm": 0.8057600298686861,
"learning_rate": 5e-06,
"loss": 0.6643,
"step": 420
},
{
"epoch": 2.5923134890730974,
"grad_norm": 0.6688016636812687,
"learning_rate": 5e-06,
"loss": 0.6657,
"step": 430
},
{
"epoch": 2.6525998492840994,
"grad_norm": 0.7419224625762498,
"learning_rate": 5e-06,
"loss": 0.6669,
"step": 440
},
{
"epoch": 2.712886209495102,
"grad_norm": 0.7294710798953756,
"learning_rate": 5e-06,
"loss": 0.664,
"step": 450
},
{
"epoch": 2.7731725697061043,
"grad_norm": 0.7239503892331308,
"learning_rate": 5e-06,
"loss": 0.6628,
"step": 460
},
{
"epoch": 2.8334589299171062,
"grad_norm": 0.6873637028133951,
"learning_rate": 5e-06,
"loss": 0.6634,
"step": 470
},
{
"epoch": 2.8937452901281087,
"grad_norm": 0.7070671521244221,
"learning_rate": 5e-06,
"loss": 0.67,
"step": 480
},
{
"epoch": 2.9540316503391106,
"grad_norm": 0.6467492636416056,
"learning_rate": 5e-06,
"loss": 0.6642,
"step": 490
},
{
"epoch": 2.9962321024868124,
"eval_loss": 0.7524702548980713,
"eval_runtime": 115.8914,
"eval_samples_per_second": 38.562,
"eval_steps_per_second": 0.604,
"step": 497
},
{
"epoch": 3.014318010550113,
"grad_norm": 1.2600144566497484,
"learning_rate": 5e-06,
"loss": 0.6595,
"step": 500
},
{
"epoch": 3.0746043707611155,
"grad_norm": 0.9439871932745402,
"learning_rate": 5e-06,
"loss": 0.6121,
"step": 510
},
{
"epoch": 3.1348907309721175,
"grad_norm": 0.9269396364377557,
"learning_rate": 5e-06,
"loss": 0.6091,
"step": 520
},
{
"epoch": 3.19517709118312,
"grad_norm": 1.4073687560739623,
"learning_rate": 5e-06,
"loss": 0.6146,
"step": 530
},
{
"epoch": 3.255463451394122,
"grad_norm": 1.0909582280984875,
"learning_rate": 5e-06,
"loss": 0.6135,
"step": 540
},
{
"epoch": 3.3157498116051243,
"grad_norm": 1.6855484277386785,
"learning_rate": 5e-06,
"loss": 0.6181,
"step": 550
},
{
"epoch": 3.3760361718161267,
"grad_norm": 1.3960879287332164,
"learning_rate": 5e-06,
"loss": 0.6204,
"step": 560
},
{
"epoch": 3.4363225320271287,
"grad_norm": 1.3111215484703225,
"learning_rate": 5e-06,
"loss": 0.6166,
"step": 570
},
{
"epoch": 3.496608892238131,
"grad_norm": 1.4456626678129865,
"learning_rate": 5e-06,
"loss": 0.6187,
"step": 580
},
{
"epoch": 3.5568952524491335,
"grad_norm": 0.9267972418344134,
"learning_rate": 5e-06,
"loss": 0.6198,
"step": 590
},
{
"epoch": 3.6171816126601355,
"grad_norm": 1.5516401845259948,
"learning_rate": 5e-06,
"loss": 0.6216,
"step": 600
},
{
"epoch": 3.677467972871138,
"grad_norm": 1.5014637761221297,
"learning_rate": 5e-06,
"loss": 0.618,
"step": 610
},
{
"epoch": 3.7377543330821403,
"grad_norm": 1.1059687872166368,
"learning_rate": 5e-06,
"loss": 0.6185,
"step": 620
},
{
"epoch": 3.7980406932931423,
"grad_norm": 1.2950526783062104,
"learning_rate": 5e-06,
"loss": 0.6207,
"step": 630
},
{
"epoch": 3.8583270535041447,
"grad_norm": 0.9542402867171873,
"learning_rate": 5e-06,
"loss": 0.618,
"step": 640
},
{
"epoch": 3.918613413715147,
"grad_norm": 0.7924967732032556,
"learning_rate": 5e-06,
"loss": 0.6163,
"step": 650
},
{
"epoch": 3.978899773926149,
"grad_norm": 0.9450387587222794,
"learning_rate": 5e-06,
"loss": 0.6186,
"step": 660
},
{
"epoch": 3.99698568198945,
"eval_loss": 0.7615257501602173,
"eval_runtime": 111.9191,
"eval_samples_per_second": 39.931,
"eval_steps_per_second": 0.625,
"step": 663
},
{
"epoch": 4.039186134137151,
"grad_norm": 0.9971689199182937,
"learning_rate": 5e-06,
"loss": 0.5958,
"step": 670
},
{
"epoch": 4.099472494348154,
"grad_norm": 0.7965799505534403,
"learning_rate": 5e-06,
"loss": 0.5627,
"step": 680
},
{
"epoch": 4.159758854559156,
"grad_norm": 0.8634698321641859,
"learning_rate": 5e-06,
"loss": 0.5642,
"step": 690
},
{
"epoch": 4.220045214770158,
"grad_norm": 0.8538005391887241,
"learning_rate": 5e-06,
"loss": 0.5721,
"step": 700
},
{
"epoch": 4.280331574981161,
"grad_norm": 0.8087517000342034,
"learning_rate": 5e-06,
"loss": 0.5702,
"step": 710
},
{
"epoch": 4.340617935192163,
"grad_norm": 0.7416451232136964,
"learning_rate": 5e-06,
"loss": 0.5735,
"step": 720
},
{
"epoch": 4.400904295403165,
"grad_norm": 0.683331985209307,
"learning_rate": 5e-06,
"loss": 0.576,
"step": 730
},
{
"epoch": 4.461190655614168,
"grad_norm": 0.7531824862918711,
"learning_rate": 5e-06,
"loss": 0.5747,
"step": 740
},
{
"epoch": 4.52147701582517,
"grad_norm": 0.6679489381777426,
"learning_rate": 5e-06,
"loss": 0.5788,
"step": 750
},
{
"epoch": 4.581763376036172,
"grad_norm": 0.8184328898914945,
"learning_rate": 5e-06,
"loss": 0.5773,
"step": 760
},
{
"epoch": 4.642049736247174,
"grad_norm": 0.7709343247764259,
"learning_rate": 5e-06,
"loss": 0.5747,
"step": 770
},
{
"epoch": 4.702336096458176,
"grad_norm": 0.8008305213299817,
"learning_rate": 5e-06,
"loss": 0.5789,
"step": 780
},
{
"epoch": 4.762622456669178,
"grad_norm": 0.770151242560135,
"learning_rate": 5e-06,
"loss": 0.5808,
"step": 790
},
{
"epoch": 4.822908816880181,
"grad_norm": 0.8007315199653617,
"learning_rate": 5e-06,
"loss": 0.5766,
"step": 800
},
{
"epoch": 4.883195177091183,
"grad_norm": 0.8153924743627864,
"learning_rate": 5e-06,
"loss": 0.5794,
"step": 810
},
{
"epoch": 4.943481537302185,
"grad_norm": 0.7155989123366884,
"learning_rate": 5e-06,
"loss": 0.5777,
"step": 820
},
{
"epoch": 4.997739261492088,
"eval_loss": 0.7784610390663147,
"eval_runtime": 113.9486,
"eval_samples_per_second": 39.219,
"eval_steps_per_second": 0.614,
"step": 829
},
{
"epoch": 5.003767897513188,
"grad_norm": 1.5968784870292558,
"learning_rate": 5e-06,
"loss": 0.5864,
"step": 830
},
{
"epoch": 5.06405425772419,
"grad_norm": 1.3160101230444552,
"learning_rate": 5e-06,
"loss": 0.5136,
"step": 840
},
{
"epoch": 5.124340617935192,
"grad_norm": 0.9218240884563489,
"learning_rate": 5e-06,
"loss": 0.514,
"step": 850
},
{
"epoch": 5.184626978146194,
"grad_norm": 0.943292128668918,
"learning_rate": 5e-06,
"loss": 0.5129,
"step": 860
},
{
"epoch": 5.244913338357197,
"grad_norm": 0.9210116431359558,
"learning_rate": 5e-06,
"loss": 0.5147,
"step": 870
},
{
"epoch": 5.305199698568199,
"grad_norm": 0.9123827846726384,
"learning_rate": 5e-06,
"loss": 0.5152,
"step": 880
},
{
"epoch": 5.365486058779201,
"grad_norm": 0.8967759483580899,
"learning_rate": 5e-06,
"loss": 0.5246,
"step": 890
},
{
"epoch": 5.425772418990204,
"grad_norm": 1.0833892045281441,
"learning_rate": 5e-06,
"loss": 0.5173,
"step": 900
},
{
"epoch": 5.486058779201206,
"grad_norm": 0.8452587501323132,
"learning_rate": 5e-06,
"loss": 0.5205,
"step": 910
},
{
"epoch": 5.546345139412208,
"grad_norm": 0.9036749110113976,
"learning_rate": 5e-06,
"loss": 0.5213,
"step": 920
},
{
"epoch": 5.6066314996232105,
"grad_norm": 0.8270647016134544,
"learning_rate": 5e-06,
"loss": 0.5214,
"step": 930
},
{
"epoch": 5.6669178598342125,
"grad_norm": 0.760029827349666,
"learning_rate": 5e-06,
"loss": 0.5233,
"step": 940
},
{
"epoch": 5.7272042200452145,
"grad_norm": 1.1680570931837515,
"learning_rate": 5e-06,
"loss": 0.5284,
"step": 950
},
{
"epoch": 5.787490580256217,
"grad_norm": 1.0453134822125045,
"learning_rate": 5e-06,
"loss": 0.5273,
"step": 960
},
{
"epoch": 5.847776940467219,
"grad_norm": 0.8063546838482688,
"learning_rate": 5e-06,
"loss": 0.5265,
"step": 970
},
{
"epoch": 5.908063300678221,
"grad_norm": 0.8921967199529918,
"learning_rate": 5e-06,
"loss": 0.5316,
"step": 980
},
{
"epoch": 5.968349660889224,
"grad_norm": 0.8942899369042694,
"learning_rate": 5e-06,
"loss": 0.5287,
"step": 990
},
{
"epoch": 5.998492840994725,
"eval_loss": 0.815434992313385,
"eval_runtime": 116.1895,
"eval_samples_per_second": 38.463,
"eval_steps_per_second": 0.602,
"step": 995
},
{
"epoch": 6.028636021100226,
"grad_norm": 1.6908391333792119,
"learning_rate": 5e-06,
"loss": 0.5023,
"step": 1000
},
{
"epoch": 6.088922381311228,
"grad_norm": 1.2916624414612379,
"learning_rate": 5e-06,
"loss": 0.4578,
"step": 1010
},
{
"epoch": 6.149208741522231,
"grad_norm": 1.1040568786409888,
"learning_rate": 5e-06,
"loss": 0.454,
"step": 1020
},
{
"epoch": 6.209495101733233,
"grad_norm": 1.0463754179790778,
"learning_rate": 5e-06,
"loss": 0.4592,
"step": 1030
},
{
"epoch": 6.269781461944235,
"grad_norm": 1.0493678707628407,
"learning_rate": 5e-06,
"loss": 0.4582,
"step": 1040
},
{
"epoch": 6.330067822155238,
"grad_norm": 1.105648007217362,
"learning_rate": 5e-06,
"loss": 0.4593,
"step": 1050
},
{
"epoch": 6.39035418236624,
"grad_norm": 1.0291790248523935,
"learning_rate": 5e-06,
"loss": 0.4629,
"step": 1060
},
{
"epoch": 6.450640542577242,
"grad_norm": 1.2514460402035084,
"learning_rate": 5e-06,
"loss": 0.4634,
"step": 1070
},
{
"epoch": 6.510926902788244,
"grad_norm": 0.8518373738036685,
"learning_rate": 5e-06,
"loss": 0.4654,
"step": 1080
},
{
"epoch": 6.571213262999247,
"grad_norm": 1.071998502866085,
"learning_rate": 5e-06,
"loss": 0.4663,
"step": 1090
},
{
"epoch": 6.6314996232102486,
"grad_norm": 0.9000534182028347,
"learning_rate": 5e-06,
"loss": 0.4658,
"step": 1100
},
{
"epoch": 6.691785983421251,
"grad_norm": 0.8995635602376101,
"learning_rate": 5e-06,
"loss": 0.469,
"step": 1110
},
{
"epoch": 6.752072343632253,
"grad_norm": 0.9011803423576781,
"learning_rate": 5e-06,
"loss": 0.4715,
"step": 1120
},
{
"epoch": 6.812358703843255,
"grad_norm": 0.9140285189147836,
"learning_rate": 5e-06,
"loss": 0.4763,
"step": 1130
},
{
"epoch": 6.872645064054257,
"grad_norm": 1.0476263187239303,
"learning_rate": 5e-06,
"loss": 0.4781,
"step": 1140
},
{
"epoch": 6.93293142426526,
"grad_norm": 1.1369535585860062,
"learning_rate": 5e-06,
"loss": 0.4702,
"step": 1150
},
{
"epoch": 6.993217784476262,
"grad_norm": 1.2720345806831712,
"learning_rate": 5e-06,
"loss": 0.473,
"step": 1160
},
{
"epoch": 6.999246420497363,
"eval_loss": 0.8709951639175415,
"eval_runtime": 116.2672,
"eval_samples_per_second": 38.437,
"eval_steps_per_second": 0.602,
"step": 1161
},
{
"epoch": 7.053504144687264,
"grad_norm": 1.6175687599813067,
"learning_rate": 5e-06,
"loss": 0.4154,
"step": 1170
},
{
"epoch": 7.113790504898267,
"grad_norm": 1.2621822187556238,
"learning_rate": 5e-06,
"loss": 0.3985,
"step": 1180
},
{
"epoch": 7.174076865109269,
"grad_norm": 1.3109771889362003,
"learning_rate": 5e-06,
"loss": 0.3937,
"step": 1190
},
{
"epoch": 7.234363225320271,
"grad_norm": 1.1476426236895934,
"learning_rate": 5e-06,
"loss": 0.3986,
"step": 1200
},
{
"epoch": 7.294649585531274,
"grad_norm": 1.1513686323462657,
"learning_rate": 5e-06,
"loss": 0.4003,
"step": 1210
},
{
"epoch": 7.354935945742276,
"grad_norm": 1.2318284334824612,
"learning_rate": 5e-06,
"loss": 0.402,
"step": 1220
},
{
"epoch": 7.415222305953278,
"grad_norm": 1.3293034219686575,
"learning_rate": 5e-06,
"loss": 0.4041,
"step": 1230
},
{
"epoch": 7.475508666164281,
"grad_norm": 1.3126587055332364,
"learning_rate": 5e-06,
"loss": 0.4028,
"step": 1240
},
{
"epoch": 7.535795026375283,
"grad_norm": 1.4541660160673635,
"learning_rate": 5e-06,
"loss": 0.4046,
"step": 1250
},
{
"epoch": 7.596081386586285,
"grad_norm": 1.1677113248358406,
"learning_rate": 5e-06,
"loss": 0.4064,
"step": 1260
},
{
"epoch": 7.6563677467972875,
"grad_norm": 1.2339252526811149,
"learning_rate": 5e-06,
"loss": 0.4057,
"step": 1270
},
{
"epoch": 7.7166541070082895,
"grad_norm": 0.9795667261721808,
"learning_rate": 5e-06,
"loss": 0.4105,
"step": 1280
},
{
"epoch": 7.776940467219291,
"grad_norm": 1.3122595751880688,
"learning_rate": 5e-06,
"loss": 0.4115,
"step": 1290
},
{
"epoch": 7.837226827430294,
"grad_norm": 1.3363743108253807,
"learning_rate": 5e-06,
"loss": 0.4094,
"step": 1300
},
{
"epoch": 7.897513187641296,
"grad_norm": 1.2278606177387434,
"learning_rate": 5e-06,
"loss": 0.411,
"step": 1310
},
{
"epoch": 7.957799547852298,
"grad_norm": 1.099485555308857,
"learning_rate": 5e-06,
"loss": 0.4134,
"step": 1320
},
{
"epoch": 8.0,
"eval_loss": 0.9475046396255493,
"eval_runtime": 118.217,
"eval_samples_per_second": 37.803,
"eval_steps_per_second": 0.592,
"step": 1327
},
{
"epoch": 8.0180859080633,
"grad_norm": 2.747551612605114,
"learning_rate": 5e-06,
"loss": 0.3977,
"step": 1330
},
{
"epoch": 8.078372268274302,
"grad_norm": 1.7260940558261295,
"learning_rate": 5e-06,
"loss": 0.3441,
"step": 1340
},
{
"epoch": 8.138658628485306,
"grad_norm": 1.2842675938679606,
"learning_rate": 5e-06,
"loss": 0.3397,
"step": 1350
},
{
"epoch": 8.198944988696308,
"grad_norm": 1.1398592198842168,
"learning_rate": 5e-06,
"loss": 0.3442,
"step": 1360
},
{
"epoch": 8.25923134890731,
"grad_norm": 1.153557070127385,
"learning_rate": 5e-06,
"loss": 0.3454,
"step": 1370
},
{
"epoch": 8.319517709118312,
"grad_norm": 1.3384373413827295,
"learning_rate": 5e-06,
"loss": 0.3438,
"step": 1380
},
{
"epoch": 8.379804069329314,
"grad_norm": 1.3554470516864743,
"learning_rate": 5e-06,
"loss": 0.3493,
"step": 1390
},
{
"epoch": 8.440090429540316,
"grad_norm": 1.325212009488347,
"learning_rate": 5e-06,
"loss": 0.3528,
"step": 1400
},
{
"epoch": 8.500376789751318,
"grad_norm": 1.74556781346271,
"learning_rate": 5e-06,
"loss": 0.3548,
"step": 1410
},
{
"epoch": 8.560663149962322,
"grad_norm": 1.3741632805919384,
"learning_rate": 5e-06,
"loss": 0.3518,
"step": 1420
},
{
"epoch": 8.620949510173324,
"grad_norm": 1.3019952245252948,
"learning_rate": 5e-06,
"loss": 0.3521,
"step": 1430
},
{
"epoch": 8.681235870384326,
"grad_norm": 1.100365748448614,
"learning_rate": 5e-06,
"loss": 0.3573,
"step": 1440
},
{
"epoch": 8.741522230595328,
"grad_norm": 1.3911327396968252,
"learning_rate": 5e-06,
"loss": 0.3574,
"step": 1450
},
{
"epoch": 8.80180859080633,
"grad_norm": 1.3853784407579808,
"learning_rate": 5e-06,
"loss": 0.3606,
"step": 1460
},
{
"epoch": 8.862094951017333,
"grad_norm": 1.278532404930149,
"learning_rate": 5e-06,
"loss": 0.3567,
"step": 1470
},
{
"epoch": 8.922381311228335,
"grad_norm": 1.207780040603517,
"learning_rate": 5e-06,
"loss": 0.3565,
"step": 1480
},
{
"epoch": 8.982667671439337,
"grad_norm": 1.262306381996002,
"learning_rate": 5e-06,
"loss": 0.3615,
"step": 1490
},
{
"epoch": 8.994724943481538,
"eval_loss": 1.0203213691711426,
"eval_runtime": 114.5597,
"eval_samples_per_second": 39.01,
"eval_steps_per_second": 0.611,
"step": 1492
},
{
"epoch": 9.04295403165034,
"grad_norm": 1.9088952161459727,
"learning_rate": 5e-06,
"loss": 0.3188,
"step": 1500
},
{
"epoch": 9.103240391861341,
"grad_norm": 1.5762821812746035,
"learning_rate": 5e-06,
"loss": 0.2921,
"step": 1510
},
{
"epoch": 9.163526752072343,
"grad_norm": 1.892762439817433,
"learning_rate": 5e-06,
"loss": 0.2866,
"step": 1520
},
{
"epoch": 9.223813112283345,
"grad_norm": 1.5993468972103952,
"learning_rate": 5e-06,
"loss": 0.2913,
"step": 1530
},
{
"epoch": 9.284099472494349,
"grad_norm": 1.7180835806783503,
"learning_rate": 5e-06,
"loss": 0.2925,
"step": 1540
},
{
"epoch": 9.34438583270535,
"grad_norm": 1.4283006573501846,
"learning_rate": 5e-06,
"loss": 0.2936,
"step": 1550
},
{
"epoch": 9.404672192916353,
"grad_norm": 1.472590439034633,
"learning_rate": 5e-06,
"loss": 0.2957,
"step": 1560
},
{
"epoch": 9.464958553127355,
"grad_norm": 1.6627373496351208,
"learning_rate": 5e-06,
"loss": 0.2968,
"step": 1570
},
{
"epoch": 9.525244913338357,
"grad_norm": 1.6479857681742245,
"learning_rate": 5e-06,
"loss": 0.2968,
"step": 1580
},
{
"epoch": 9.585531273549359,
"grad_norm": 1.451039229532527,
"learning_rate": 5e-06,
"loss": 0.2969,
"step": 1590
},
{
"epoch": 9.645817633760362,
"grad_norm": 1.338137019241527,
"learning_rate": 5e-06,
"loss": 0.2999,
"step": 1600
},
{
"epoch": 9.706103993971364,
"grad_norm": 1.4364529449975805,
"learning_rate": 5e-06,
"loss": 0.301,
"step": 1610
},
{
"epoch": 9.766390354182366,
"grad_norm": 1.6181472883899526,
"learning_rate": 5e-06,
"loss": 0.3021,
"step": 1620
},
{
"epoch": 9.826676714393368,
"grad_norm": 1.2929981294290136,
"learning_rate": 5e-06,
"loss": 0.3047,
"step": 1630
},
{
"epoch": 9.88696307460437,
"grad_norm": 1.5569082204346147,
"learning_rate": 5e-06,
"loss": 0.3039,
"step": 1640
},
{
"epoch": 9.947249434815372,
"grad_norm": 1.2611067636631264,
"learning_rate": 5e-06,
"loss": 0.3057,
"step": 1650
},
{
"epoch": 9.995478522984175,
"eval_loss": 1.1177444458007812,
"eval_runtime": 116.3468,
"eval_samples_per_second": 38.411,
"eval_steps_per_second": 0.602,
"step": 1658
},
{
"epoch": 10.007535795026376,
"grad_norm": 1.793363117063337,
"learning_rate": 5e-06,
"loss": 0.3015,
"step": 1660
},
{
"epoch": 10.067822155237378,
"grad_norm": 1.9015543562249504,
"learning_rate": 5e-06,
"loss": 0.2446,
"step": 1670
},
{
"epoch": 10.12810851544838,
"grad_norm": 1.6166742540525376,
"learning_rate": 5e-06,
"loss": 0.2417,
"step": 1680
},
{
"epoch": 10.188394875659382,
"grad_norm": 1.5826959730874415,
"learning_rate": 5e-06,
"loss": 0.2407,
"step": 1690
},
{
"epoch": 10.248681235870384,
"grad_norm": 1.313614154478162,
"learning_rate": 5e-06,
"loss": 0.2421,
"step": 1700
},
{
"epoch": 10.308967596081386,
"grad_norm": 1.8551338463863283,
"learning_rate": 5e-06,
"loss": 0.2432,
"step": 1710
},
{
"epoch": 10.369253956292388,
"grad_norm": 1.3937043583045259,
"learning_rate": 5e-06,
"loss": 0.2426,
"step": 1720
},
{
"epoch": 10.429540316503392,
"grad_norm": 1.4908947612613312,
"learning_rate": 5e-06,
"loss": 0.2458,
"step": 1730
},
{
"epoch": 10.489826676714394,
"grad_norm": 1.3513472181146375,
"learning_rate": 5e-06,
"loss": 0.2468,
"step": 1740
},
{
"epoch": 10.550113036925396,
"grad_norm": 1.2442859437459592,
"learning_rate": 5e-06,
"loss": 0.2472,
"step": 1750
},
{
"epoch": 10.610399397136398,
"grad_norm": 1.61049205004398,
"learning_rate": 5e-06,
"loss": 0.2497,
"step": 1760
},
{
"epoch": 10.6706857573474,
"grad_norm": 1.4687600663257643,
"learning_rate": 5e-06,
"loss": 0.2513,
"step": 1770
},
{
"epoch": 10.730972117558402,
"grad_norm": 1.3544154919254643,
"learning_rate": 5e-06,
"loss": 0.2534,
"step": 1780
},
{
"epoch": 10.791258477769405,
"grad_norm": 1.4310611416705883,
"learning_rate": 5e-06,
"loss": 0.2534,
"step": 1790
},
{
"epoch": 10.851544837980407,
"grad_norm": 1.5722579222869695,
"learning_rate": 5e-06,
"loss": 0.2543,
"step": 1800
},
{
"epoch": 10.91183119819141,
"grad_norm": 1.520704900113862,
"learning_rate": 5e-06,
"loss": 0.254,
"step": 1810
},
{
"epoch": 10.972117558402411,
"grad_norm": 1.4393868378454422,
"learning_rate": 5e-06,
"loss": 0.2565,
"step": 1820
},
{
"epoch": 10.996232102486813,
"eval_loss": 1.236780047416687,
"eval_runtime": 115.9111,
"eval_samples_per_second": 38.555,
"eval_steps_per_second": 0.604,
"step": 1824
},
{
"epoch": 11.032403918613413,
"grad_norm": 2.415959560977028,
"learning_rate": 5e-06,
"loss": 0.2285,
"step": 1830
},
{
"epoch": 11.092690278824415,
"grad_norm": 1.5442333094800837,
"learning_rate": 5e-06,
"loss": 0.1975,
"step": 1840
},
{
"epoch": 11.152976639035419,
"grad_norm": 2.1161682250412257,
"learning_rate": 5e-06,
"loss": 0.1968,
"step": 1850
},
{
"epoch": 11.213262999246421,
"grad_norm": 1.6062184104077026,
"learning_rate": 5e-06,
"loss": 0.1967,
"step": 1860
},
{
"epoch": 11.273549359457423,
"grad_norm": 1.6876217090206025,
"learning_rate": 5e-06,
"loss": 0.1937,
"step": 1870
},
{
"epoch": 11.333835719668425,
"grad_norm": 1.756341727895367,
"learning_rate": 5e-06,
"loss": 0.1966,
"step": 1880
},
{
"epoch": 11.394122079879427,
"grad_norm": 1.5508699967732422,
"learning_rate": 5e-06,
"loss": 0.1997,
"step": 1890
},
{
"epoch": 11.454408440090429,
"grad_norm": 1.5669766694757912,
"learning_rate": 5e-06,
"loss": 0.2023,
"step": 1900
},
{
"epoch": 11.51469480030143,
"grad_norm": 1.4502558429874435,
"learning_rate": 5e-06,
"loss": 0.202,
"step": 1910
},
{
"epoch": 11.574981160512435,
"grad_norm": 1.432868245038754,
"learning_rate": 5e-06,
"loss": 0.2027,
"step": 1920
},
{
"epoch": 11.635267520723437,
"grad_norm": 1.944951997696116,
"learning_rate": 5e-06,
"loss": 0.2051,
"step": 1930
},
{
"epoch": 11.695553880934439,
"grad_norm": 1.5628919140219797,
"learning_rate": 5e-06,
"loss": 0.2047,
"step": 1940
},
{
"epoch": 11.75584024114544,
"grad_norm": 1.8554418586433572,
"learning_rate": 5e-06,
"loss": 0.2036,
"step": 1950
},
{
"epoch": 11.816126601356443,
"grad_norm": 1.5083234117594833,
"learning_rate": 5e-06,
"loss": 0.2057,
"step": 1960
},
{
"epoch": 11.876412961567446,
"grad_norm": 1.6850324704140989,
"learning_rate": 5e-06,
"loss": 0.2084,
"step": 1970
},
{
"epoch": 11.936699321778448,
"grad_norm": 1.518863675506513,
"learning_rate": 5e-06,
"loss": 0.2081,
"step": 1980
},
{
"epoch": 11.99698568198945,
"grad_norm": 2.025449919349873,
"learning_rate": 5e-06,
"loss": 0.2099,
"step": 1990
},
{
"epoch": 11.99698568198945,
"eval_loss": 1.3552354574203491,
"eval_runtime": 116.1042,
"eval_samples_per_second": 38.491,
"eval_steps_per_second": 0.603,
"step": 1990
},
{
"epoch": 12.057272042200452,
"grad_norm": 1.9707275891443514,
"learning_rate": 5e-06,
"loss": 0.161,
"step": 2000
},
{
"epoch": 12.117558402411454,
"grad_norm": 1.60811233115537,
"learning_rate": 5e-06,
"loss": 0.1527,
"step": 2010
},
{
"epoch": 12.177844762622456,
"grad_norm": 1.6751301942391925,
"learning_rate": 5e-06,
"loss": 0.1541,
"step": 2020
},
{
"epoch": 12.238131122833458,
"grad_norm": 1.4492067853763235,
"learning_rate": 5e-06,
"loss": 0.1559,
"step": 2030
},
{
"epoch": 12.298417483044462,
"grad_norm": 1.9142452633048677,
"learning_rate": 5e-06,
"loss": 0.1576,
"step": 2040
},
{
"epoch": 12.358703843255464,
"grad_norm": 1.4385285667791092,
"learning_rate": 5e-06,
"loss": 0.1589,
"step": 2050
},
{
"epoch": 12.418990203466466,
"grad_norm": 1.4928747098321213,
"learning_rate": 5e-06,
"loss": 0.1608,
"step": 2060
},
{
"epoch": 12.479276563677468,
"grad_norm": 1.5019521655912937,
"learning_rate": 5e-06,
"loss": 0.1605,
"step": 2070
},
{
"epoch": 12.53956292388847,
"grad_norm": 1.874845955363987,
"learning_rate": 5e-06,
"loss": 0.1612,
"step": 2080
},
{
"epoch": 12.599849284099472,
"grad_norm": 1.885802643703012,
"learning_rate": 5e-06,
"loss": 0.1614,
"step": 2090
},
{
"epoch": 12.660135644310476,
"grad_norm": 1.515486697883317,
"learning_rate": 5e-06,
"loss": 0.1622,
"step": 2100
},
{
"epoch": 12.720422004521478,
"grad_norm": 1.7537896738178245,
"learning_rate": 5e-06,
"loss": 0.1644,
"step": 2110
},
{
"epoch": 12.78070836473248,
"grad_norm": 1.6471681441773147,
"learning_rate": 5e-06,
"loss": 0.1647,
"step": 2120
},
{
"epoch": 12.840994724943481,
"grad_norm": 1.590934554816475,
"learning_rate": 5e-06,
"loss": 0.1648,
"step": 2130
},
{
"epoch": 12.901281085154483,
"grad_norm": 1.7730249951617123,
"learning_rate": 5e-06,
"loss": 0.1656,
"step": 2140
},
{
"epoch": 12.961567445365485,
"grad_norm": 1.6645957171669876,
"learning_rate": 5e-06,
"loss": 0.1676,
"step": 2150
},
{
"epoch": 12.997739261492088,
"eval_loss": 1.5070669651031494,
"eval_runtime": 116.1666,
"eval_samples_per_second": 38.471,
"eval_steps_per_second": 0.603,
"step": 2156
},
{
"epoch": 13.02185380557649,
"grad_norm": 2.462408072338029,
"learning_rate": 5e-06,
"loss": 0.1517,
"step": 2160
},
{
"epoch": 13.082140165787491,
"grad_norm": 2.4318360950844804,
"learning_rate": 5e-06,
"loss": 0.1195,
"step": 2170
},
{
"epoch": 13.142426525998493,
"grad_norm": 1.7426593578459117,
"learning_rate": 5e-06,
"loss": 0.1201,
"step": 2180
},
{
"epoch": 13.202712886209495,
"grad_norm": 2.2383839547509696,
"learning_rate": 5e-06,
"loss": 0.1191,
"step": 2190
},
{
"epoch": 13.262999246420497,
"grad_norm": 1.5553317913023317,
"learning_rate": 5e-06,
"loss": 0.118,
"step": 2200
},
{
"epoch": 13.323285606631499,
"grad_norm": 1.4669913420252614,
"learning_rate": 5e-06,
"loss": 0.119,
"step": 2210
},
{
"epoch": 13.383571966842501,
"grad_norm": 1.8080283494554128,
"learning_rate": 5e-06,
"loss": 0.1196,
"step": 2220
},
{
"epoch": 13.443858327053505,
"grad_norm": 1.646647327329277,
"learning_rate": 5e-06,
"loss": 0.1233,
"step": 2230
},
{
"epoch": 13.504144687264507,
"grad_norm": 1.8533611593149397,
"learning_rate": 5e-06,
"loss": 0.1241,
"step": 2240
},
{
"epoch": 13.564431047475509,
"grad_norm": 1.6777607818909321,
"learning_rate": 5e-06,
"loss": 0.126,
"step": 2250
},
{
"epoch": 13.62471740768651,
"grad_norm": 1.7745416454251923,
"learning_rate": 5e-06,
"loss": 0.1266,
"step": 2260
},
{
"epoch": 13.685003767897513,
"grad_norm": 1.7096006593564836,
"learning_rate": 5e-06,
"loss": 0.1282,
"step": 2270
},
{
"epoch": 13.745290128108515,
"grad_norm": 1.8638678064551115,
"learning_rate": 5e-06,
"loss": 0.1267,
"step": 2280
},
{
"epoch": 13.805576488319518,
"grad_norm": 1.4262510400727582,
"learning_rate": 5e-06,
"loss": 0.1273,
"step": 2290
},
{
"epoch": 13.86586284853052,
"grad_norm": 1.8828592364046501,
"learning_rate": 5e-06,
"loss": 0.1269,
"step": 2300
},
{
"epoch": 13.926149208741522,
"grad_norm": 1.6583448018945106,
"learning_rate": 5e-06,
"loss": 0.1273,
"step": 2310
},
{
"epoch": 13.986435568952524,
"grad_norm": 1.548491818827999,
"learning_rate": 5e-06,
"loss": 0.1283,
"step": 2320
},
{
"epoch": 13.998492840994725,
"eval_loss": 1.6323552131652832,
"eval_runtime": 115.7608,
"eval_samples_per_second": 38.605,
"eval_steps_per_second": 0.605,
"step": 2322
},
{
"epoch": 14.046721929163526,
"grad_norm": 2.0937924453069203,
"learning_rate": 5e-06,
"loss": 0.1006,
"step": 2330
},
{
"epoch": 14.107008289374528,
"grad_norm": 2.139024110314688,
"learning_rate": 5e-06,
"loss": 0.0897,
"step": 2340
},
{
"epoch": 14.167294649585532,
"grad_norm": 1.5948191216888472,
"learning_rate": 5e-06,
"loss": 0.092,
"step": 2350
},
{
"epoch": 14.227581009796534,
"grad_norm": 1.6158036371008326,
"learning_rate": 5e-06,
"loss": 0.0916,
"step": 2360
},
{
"epoch": 14.287867370007536,
"grad_norm": 1.6380184320632138,
"learning_rate": 5e-06,
"loss": 0.0946,
"step": 2370
},
{
"epoch": 14.348153730218538,
"grad_norm": 1.5125683894811608,
"learning_rate": 5e-06,
"loss": 0.0926,
"step": 2380
},
{
"epoch": 14.40844009042954,
"grad_norm": 1.8291017955611948,
"learning_rate": 5e-06,
"loss": 0.0967,
"step": 2390
},
{
"epoch": 14.468726450640542,
"grad_norm": 1.6743144432735446,
"learning_rate": 5e-06,
"loss": 0.097,
"step": 2400
},
{
"epoch": 14.529012810851544,
"grad_norm": 1.7276135456842643,
"learning_rate": 5e-06,
"loss": 0.096,
"step": 2410
},
{
"epoch": 14.589299171062548,
"grad_norm": 1.734533902590376,
"learning_rate": 5e-06,
"loss": 0.0977,
"step": 2420
},
{
"epoch": 14.64958553127355,
"grad_norm": 1.7106037137785368,
"learning_rate": 5e-06,
"loss": 0.0974,
"step": 2430
},
{
"epoch": 14.709871891484552,
"grad_norm": 1.848167736152076,
"learning_rate": 5e-06,
"loss": 0.098,
"step": 2440
},
{
"epoch": 14.770158251695554,
"grad_norm": 2.0729946870310907,
"learning_rate": 5e-06,
"loss": 0.0988,
"step": 2450
},
{
"epoch": 14.830444611906556,
"grad_norm": 1.6522773246662863,
"learning_rate": 5e-06,
"loss": 0.0996,
"step": 2460
},
{
"epoch": 14.890730972117558,
"grad_norm": 1.8165186575808512,
"learning_rate": 5e-06,
"loss": 0.1018,
"step": 2470
},
{
"epoch": 14.951017332328561,
"grad_norm": 1.8324940323277443,
"learning_rate": 5e-06,
"loss": 0.1022,
"step": 2480
},
{
"epoch": 14.999246420497363,
"eval_loss": 1.7541841268539429,
"eval_runtime": 116.4567,
"eval_samples_per_second": 38.375,
"eval_steps_per_second": 0.601,
"step": 2488
},
{
"epoch": 15.011303692539563,
"grad_norm": 1.7122346572238332,
"learning_rate": 5e-06,
"loss": 0.0969,
"step": 2490
},
{
"epoch": 15.071590052750565,
"grad_norm": 1.7869038836513413,
"learning_rate": 5e-06,
"loss": 0.07,
"step": 2500
},
{
"epoch": 15.131876412961567,
"grad_norm": 1.5692589038411329,
"learning_rate": 5e-06,
"loss": 0.069,
"step": 2510
},
{
"epoch": 15.19216277317257,
"grad_norm": 1.5568543176798626,
"learning_rate": 5e-06,
"loss": 0.0714,
"step": 2520
},
{
"epoch": 15.252449133383571,
"grad_norm": 1.8979966448489778,
"learning_rate": 5e-06,
"loss": 0.0721,
"step": 2530
},
{
"epoch": 15.312735493594575,
"grad_norm": 1.6079155193057422,
"learning_rate": 5e-06,
"loss": 0.0734,
"step": 2540
},
{
"epoch": 15.373021853805577,
"grad_norm": 1.5682954069117738,
"learning_rate": 5e-06,
"loss": 0.0731,
"step": 2550
},
{
"epoch": 15.433308214016579,
"grad_norm": 1.607771288791899,
"learning_rate": 5e-06,
"loss": 0.0736,
"step": 2560
},
{
"epoch": 15.493594574227581,
"grad_norm": 1.592532966261963,
"learning_rate": 5e-06,
"loss": 0.0751,
"step": 2570
},
{
"epoch": 15.553880934438583,
"grad_norm": 2.106042369619277,
"learning_rate": 5e-06,
"loss": 0.0765,
"step": 2580
},
{
"epoch": 15.614167294649585,
"grad_norm": 1.9643387382353286,
"learning_rate": 5e-06,
"loss": 0.0764,
"step": 2590
},
{
"epoch": 15.674453654860589,
"grad_norm": 2.356472326588705,
"learning_rate": 5e-06,
"loss": 0.0766,
"step": 2600
},
{
"epoch": 15.73474001507159,
"grad_norm": 1.7285204414699735,
"learning_rate": 5e-06,
"loss": 0.0768,
"step": 2610
},
{
"epoch": 15.795026375282593,
"grad_norm": 1.656141946179993,
"learning_rate": 5e-06,
"loss": 0.0781,
"step": 2620
},
{
"epoch": 15.855312735493595,
"grad_norm": 1.751656627220462,
"learning_rate": 5e-06,
"loss": 0.0778,
"step": 2630
},
{
"epoch": 15.915599095704597,
"grad_norm": 1.9120339283286005,
"learning_rate": 5e-06,
"loss": 0.0777,
"step": 2640
},
{
"epoch": 15.975885455915598,
"grad_norm": 1.7091309765672063,
"learning_rate": 5e-06,
"loss": 0.0779,
"step": 2650
},
{
"epoch": 16.0,
"eval_loss": 1.8729431629180908,
"eval_runtime": 130.3121,
"eval_samples_per_second": 34.295,
"eval_steps_per_second": 0.537,
"step": 2654
},
{
"epoch": 16.0361718161266,
"grad_norm": 1.6873009916146906,
"learning_rate": 5e-06,
"loss": 0.0634,
"step": 2660
},
{
"epoch": 16.096458176337602,
"grad_norm": 1.4499459070972593,
"learning_rate": 5e-06,
"loss": 0.0517,
"step": 2670
},
{
"epoch": 16.156744536548604,
"grad_norm": 1.518251208097316,
"learning_rate": 5e-06,
"loss": 0.0532,
"step": 2680
},
{
"epoch": 16.217030896759606,
"grad_norm": 1.7901282259982598,
"learning_rate": 5e-06,
"loss": 0.0547,
"step": 2690
},
{
"epoch": 16.277317256970612,
"grad_norm": 1.6296841948609884,
"learning_rate": 5e-06,
"loss": 0.0553,
"step": 2700
},
{
"epoch": 16.337603617181614,
"grad_norm": 1.9077348260850149,
"learning_rate": 5e-06,
"loss": 0.0551,
"step": 2710
},
{
"epoch": 16.397889977392616,
"grad_norm": 1.6275792847492503,
"learning_rate": 5e-06,
"loss": 0.0556,
"step": 2720
},
{
"epoch": 16.458176337603618,
"grad_norm": 1.8060616409571004,
"learning_rate": 5e-06,
"loss": 0.0569,
"step": 2730
},
{
"epoch": 16.51846269781462,
"grad_norm": 1.6503072239622354,
"learning_rate": 5e-06,
"loss": 0.0568,
"step": 2740
},
{
"epoch": 16.578749058025622,
"grad_norm": 1.5506661680891511,
"learning_rate": 5e-06,
"loss": 0.0572,
"step": 2750
},
{
"epoch": 16.639035418236624,
"grad_norm": 1.6807035773937569,
"learning_rate": 5e-06,
"loss": 0.0583,
"step": 2760
},
{
"epoch": 16.699321778447626,
"grad_norm": 1.794347099575113,
"learning_rate": 5e-06,
"loss": 0.0589,
"step": 2770
},
{
"epoch": 16.759608138658628,
"grad_norm": 1.6112070974863837,
"learning_rate": 5e-06,
"loss": 0.0591,
"step": 2780
},
{
"epoch": 16.81989449886963,
"grad_norm": 1.5309535448790481,
"learning_rate": 5e-06,
"loss": 0.0595,
"step": 2790
},
{
"epoch": 16.88018085908063,
"grad_norm": 1.6962829999060645,
"learning_rate": 5e-06,
"loss": 0.0595,
"step": 2800
},
{
"epoch": 16.940467219291634,
"grad_norm": 1.5160770218318527,
"learning_rate": 5e-06,
"loss": 0.0607,
"step": 2810
},
{
"epoch": 16.994724943481536,
"eval_loss": 1.9862152338027954,
"eval_runtime": 135.1067,
"eval_samples_per_second": 33.078,
"eval_steps_per_second": 0.518,
"step": 2819
},
{
"epoch": 17.00075357950264,
"grad_norm": 2.340332137108714,
"learning_rate": 5e-06,
"loss": 0.0615,
"step": 2820
},
{
"epoch": 17.06103993971364,
"grad_norm": 1.4236109373520875,
"learning_rate": 5e-06,
"loss": 0.0397,
"step": 2830
},
{
"epoch": 17.121326299924643,
"grad_norm": 1.5339925411690258,
"learning_rate": 5e-06,
"loss": 0.0409,
"step": 2840
},
{
"epoch": 17.181612660135645,
"grad_norm": 1.5223509157958206,
"learning_rate": 5e-06,
"loss": 0.0416,
"step": 2850
},
{
"epoch": 17.241899020346647,
"grad_norm": 1.7444832258175693,
"learning_rate": 5e-06,
"loss": 0.0425,
"step": 2860
},
{
"epoch": 17.30218538055765,
"grad_norm": 1.7829806133150694,
"learning_rate": 5e-06,
"loss": 0.0426,
"step": 2870
},
{
"epoch": 17.36247174076865,
"grad_norm": 1.7776363684895005,
"learning_rate": 5e-06,
"loss": 0.0437,
"step": 2880
},
{
"epoch": 17.422758100979653,
"grad_norm": 1.449976018411429,
"learning_rate": 5e-06,
"loss": 0.0434,
"step": 2890
},
{
"epoch": 17.483044461190655,
"grad_norm": 1.4737183761321249,
"learning_rate": 5e-06,
"loss": 0.0443,
"step": 2900
},
{
"epoch": 17.543330821401657,
"grad_norm": 1.5570813152529674,
"learning_rate": 5e-06,
"loss": 0.0439,
"step": 2910
},
{
"epoch": 17.60361718161266,
"grad_norm": 1.6635463270221298,
"learning_rate": 5e-06,
"loss": 0.0451,
"step": 2920
},
{
"epoch": 17.66390354182366,
"grad_norm": 1.5633000807042863,
"learning_rate": 5e-06,
"loss": 0.0452,
"step": 2930
},
{
"epoch": 17.724189902034666,
"grad_norm": 1.5756718088358965,
"learning_rate": 5e-06,
"loss": 0.0453,
"step": 2940
},
{
"epoch": 17.78447626224567,
"grad_norm": 1.5086652524908912,
"learning_rate": 5e-06,
"loss": 0.0461,
"step": 2950
},
{
"epoch": 17.84476262245667,
"grad_norm": 1.543593492657027,
"learning_rate": 5e-06,
"loss": 0.0471,
"step": 2960
},
{
"epoch": 17.905048982667672,
"grad_norm": 1.8850354234893694,
"learning_rate": 5e-06,
"loss": 0.0473,
"step": 2970
},
{
"epoch": 17.965335342878674,
"grad_norm": 1.5531948749022353,
"learning_rate": 5e-06,
"loss": 0.0481,
"step": 2980
},
{
"epoch": 17.995478522984175,
"eval_loss": 2.0547046661376953,
"eval_runtime": 131.4004,
"eval_samples_per_second": 34.011,
"eval_steps_per_second": 0.533,
"step": 2985
},
{
"epoch": 18.025621703089676,
"grad_norm": 1.5738867889136725,
"learning_rate": 5e-06,
"loss": 0.041,
"step": 2990
},
{
"epoch": 18.08590806330068,
"grad_norm": 1.7413824765739192,
"learning_rate": 5e-06,
"loss": 0.0319,
"step": 3000
},
{
"epoch": 18.14619442351168,
"grad_norm": 1.349257585093222,
"learning_rate": 5e-06,
"loss": 0.0321,
"step": 3010
},
{
"epoch": 18.206480783722682,
"grad_norm": 1.5030527061221075,
"learning_rate": 5e-06,
"loss": 0.033,
"step": 3020
},
{
"epoch": 18.266767143933684,
"grad_norm": 1.3839757017227572,
"learning_rate": 5e-06,
"loss": 0.0338,
"step": 3030
},
{
"epoch": 18.327053504144686,
"grad_norm": 1.3639549234993298,
"learning_rate": 5e-06,
"loss": 0.034,
"step": 3040
},
{
"epoch": 18.38733986435569,
"grad_norm": 1.491804900184454,
"learning_rate": 5e-06,
"loss": 0.0344,
"step": 3050
},
{
"epoch": 18.44762622456669,
"grad_norm": 1.7182953986203733,
"learning_rate": 5e-06,
"loss": 0.0345,
"step": 3060
},
{
"epoch": 18.507912584777696,
"grad_norm": 1.5927159416873042,
"learning_rate": 5e-06,
"loss": 0.035,
"step": 3070
},
{
"epoch": 18.568198944988698,
"grad_norm": 1.48048633519207,
"learning_rate": 5e-06,
"loss": 0.0351,
"step": 3080
},
{
"epoch": 18.6284853051997,
"grad_norm": 1.5724972628180884,
"learning_rate": 5e-06,
"loss": 0.0357,
"step": 3090
},
{
"epoch": 18.6887716654107,
"grad_norm": 1.460368142482016,
"learning_rate": 5e-06,
"loss": 0.0367,
"step": 3100
},
{
"epoch": 18.749058025621704,
"grad_norm": 1.5605571994020253,
"learning_rate": 5e-06,
"loss": 0.037,
"step": 3110
},
{
"epoch": 18.809344385832706,
"grad_norm": 1.809142213762153,
"learning_rate": 5e-06,
"loss": 0.0364,
"step": 3120
},
{
"epoch": 18.869630746043708,
"grad_norm": 1.6290220037965493,
"learning_rate": 5e-06,
"loss": 0.0367,
"step": 3130
},
{
"epoch": 18.92991710625471,
"grad_norm": 1.441383521222635,
"learning_rate": 5e-06,
"loss": 0.0373,
"step": 3140
},
{
"epoch": 18.99020346646571,
"grad_norm": 1.4650505575255777,
"learning_rate": 5e-06,
"loss": 0.038,
"step": 3150
},
{
"epoch": 18.99623210248681,
"eval_loss": 2.1350691318511963,
"eval_runtime": 116.4178,
"eval_samples_per_second": 38.388,
"eval_steps_per_second": 0.601,
"step": 3151
},
{
"epoch": 19.050489826676714,
"grad_norm": 1.2364462769526054,
"learning_rate": 5e-06,
"loss": 0.0273,
"step": 3160
},
{
"epoch": 19.110776186887716,
"grad_norm": 1.2847338125252712,
"learning_rate": 5e-06,
"loss": 0.0263,
"step": 3170
},
{
"epoch": 19.171062547098717,
"grad_norm": 1.4364005652974698,
"learning_rate": 5e-06,
"loss": 0.027,
"step": 3180
},
{
"epoch": 19.23134890730972,
"grad_norm": 1.4324790800038558,
"learning_rate": 5e-06,
"loss": 0.0278,
"step": 3190
},
{
"epoch": 19.291635267520725,
"grad_norm": 1.3696161004609035,
"learning_rate": 5e-06,
"loss": 0.0268,
"step": 3200
},
{
"epoch": 19.351921627731727,
"grad_norm": 1.4042401515519942,
"learning_rate": 5e-06,
"loss": 0.0276,
"step": 3210
},
{
"epoch": 19.41220798794273,
"grad_norm": 1.4201103499411916,
"learning_rate": 5e-06,
"loss": 0.0282,
"step": 3220
},
{
"epoch": 19.47249434815373,
"grad_norm": 1.4787498619597446,
"learning_rate": 5e-06,
"loss": 0.029,
"step": 3230
},
{
"epoch": 19.532780708364733,
"grad_norm": 1.3042565121452792,
"learning_rate": 5e-06,
"loss": 0.0289,
"step": 3240
},
{
"epoch": 19.593067068575735,
"grad_norm": 1.5616956908222805,
"learning_rate": 5e-06,
"loss": 0.0291,
"step": 3250
},
{
"epoch": 19.653353428786737,
"grad_norm": 1.3916636056123413,
"learning_rate": 5e-06,
"loss": 0.0294,
"step": 3260
},
{
"epoch": 19.71363978899774,
"grad_norm": 1.5340949993651216,
"learning_rate": 5e-06,
"loss": 0.0302,
"step": 3270
},
{
"epoch": 19.77392614920874,
"grad_norm": 1.567955217260001,
"learning_rate": 5e-06,
"loss": 0.03,
"step": 3280
},
{
"epoch": 19.834212509419743,
"grad_norm": 1.4614834017683382,
"learning_rate": 5e-06,
"loss": 0.0302,
"step": 3290
},
{
"epoch": 19.894498869630745,
"grad_norm": 1.3384969584915662,
"learning_rate": 5e-06,
"loss": 0.0299,
"step": 3300
},
{
"epoch": 19.954785229841747,
"grad_norm": 1.6563798754782657,
"learning_rate": 5e-06,
"loss": 0.0306,
"step": 3310
},
{
"epoch": 19.99698568198945,
"eval_loss": 2.2255122661590576,
"eval_runtime": 131.2328,
"eval_samples_per_second": 34.054,
"eval_steps_per_second": 0.533,
"step": 3317
},
{
"epoch": 20.015071590052752,
"grad_norm": 1.2478793413376605,
"learning_rate": 5e-06,
"loss": 0.0282,
"step": 3320
},
{
"epoch": 20.075357950263754,
"grad_norm": 1.5652614584389117,
"learning_rate": 5e-06,
"loss": 0.0201,
"step": 3330
},
{
"epoch": 20.135644310474756,
"grad_norm": 1.6731787596786802,
"learning_rate": 5e-06,
"loss": 0.0215,
"step": 3340
},
{
"epoch": 20.195930670685758,
"grad_norm": 1.2852656312690158,
"learning_rate": 5e-06,
"loss": 0.0215,
"step": 3350
},
{
"epoch": 20.25621703089676,
"grad_norm": 1.189629660977167,
"learning_rate": 5e-06,
"loss": 0.0218,
"step": 3360
},
{
"epoch": 20.316503391107762,
"grad_norm": 1.280211241755101,
"learning_rate": 5e-06,
"loss": 0.0225,
"step": 3370
},
{
"epoch": 20.376789751318764,
"grad_norm": 1.3671510362477786,
"learning_rate": 5e-06,
"loss": 0.0233,
"step": 3380
},
{
"epoch": 20.437076111529766,
"grad_norm": 1.2479455610052839,
"learning_rate": 5e-06,
"loss": 0.0235,
"step": 3390
},
{
"epoch": 20.497362471740768,
"grad_norm": 1.3487469571257409,
"learning_rate": 5e-06,
"loss": 0.024,
"step": 3400
},
{
"epoch": 20.55764883195177,
"grad_norm": 1.296539761041376,
"learning_rate": 5e-06,
"loss": 0.0239,
"step": 3410
},
{
"epoch": 20.617935192162772,
"grad_norm": 1.3688473131734242,
"learning_rate": 5e-06,
"loss": 0.0243,
"step": 3420
},
{
"epoch": 20.678221552373774,
"grad_norm": 1.3764265527684842,
"learning_rate": 5e-06,
"loss": 0.0246,
"step": 3430
},
{
"epoch": 20.738507912584776,
"grad_norm": 1.5833602766458634,
"learning_rate": 5e-06,
"loss": 0.0249,
"step": 3440
},
{
"epoch": 20.79879427279578,
"grad_norm": 1.528464092769155,
"learning_rate": 5e-06,
"loss": 0.0258,
"step": 3450
},
{
"epoch": 20.859080633006784,
"grad_norm": 1.605313026206358,
"learning_rate": 5e-06,
"loss": 0.026,
"step": 3460
},
{
"epoch": 20.919366993217785,
"grad_norm": 1.3923344492334147,
"learning_rate": 5e-06,
"loss": 0.0259,
"step": 3470
},
{
"epoch": 20.979653353428787,
"grad_norm": 1.4204063096393778,
"learning_rate": 5e-06,
"loss": 0.0256,
"step": 3480
},
{
"epoch": 20.997739261492086,
"eval_loss": 2.269946813583374,
"eval_runtime": 116.5456,
"eval_samples_per_second": 38.345,
"eval_steps_per_second": 0.601,
"step": 3483
},
{
"epoch": 21.03993971363979,
"grad_norm": 1.310936775235027,
"learning_rate": 5e-06,
"loss": 0.0208,
"step": 3490
},
{
"epoch": 21.10022607385079,
"grad_norm": 1.1547005623677518,
"learning_rate": 5e-06,
"loss": 0.0186,
"step": 3500
},
{
"epoch": 21.160512434061793,
"grad_norm": 1.2938934451769595,
"learning_rate": 5e-06,
"loss": 0.0194,
"step": 3510
},
{
"epoch": 21.220798794272795,
"grad_norm": 1.484320867371485,
"learning_rate": 5e-06,
"loss": 0.0196,
"step": 3520
},
{
"epoch": 21.281085154483797,
"grad_norm": 1.4005288463687333,
"learning_rate": 5e-06,
"loss": 0.02,
"step": 3530
},
{
"epoch": 21.3413715146948,
"grad_norm": 1.6013274204687469,
"learning_rate": 5e-06,
"loss": 0.0201,
"step": 3540
},
{
"epoch": 21.4016578749058,
"grad_norm": 1.4800630752519932,
"learning_rate": 5e-06,
"loss": 0.02,
"step": 3550
},
{
"epoch": 21.461944235116803,
"grad_norm": 1.497381379010042,
"learning_rate": 5e-06,
"loss": 0.0203,
"step": 3560
},
{
"epoch": 21.52223059532781,
"grad_norm": 1.2261010629715117,
"learning_rate": 5e-06,
"loss": 0.0208,
"step": 3570
},
{
"epoch": 21.58251695553881,
"grad_norm": 1.353310824251755,
"learning_rate": 5e-06,
"loss": 0.0211,
"step": 3580
},
{
"epoch": 21.642803315749813,
"grad_norm": 1.2550244372657888,
"learning_rate": 5e-06,
"loss": 0.0217,
"step": 3590
},
{
"epoch": 21.703089675960815,
"grad_norm": 1.3870144386972025,
"learning_rate": 5e-06,
"loss": 0.0218,
"step": 3600
},
{
"epoch": 21.763376036171817,
"grad_norm": 1.2996006226877992,
"learning_rate": 5e-06,
"loss": 0.0217,
"step": 3610
},
{
"epoch": 21.82366239638282,
"grad_norm": 1.3533095992501387,
"learning_rate": 5e-06,
"loss": 0.0216,
"step": 3620
},
{
"epoch": 21.88394875659382,
"grad_norm": 1.3112078284304194,
"learning_rate": 5e-06,
"loss": 0.0217,
"step": 3630
},
{
"epoch": 21.944235116804823,
"grad_norm": 1.3288506152434172,
"learning_rate": 5e-06,
"loss": 0.0221,
"step": 3640
},
{
"epoch": 21.998492840994725,
"eval_loss": 2.351545572280884,
"eval_runtime": 116.2788,
"eval_samples_per_second": 38.433,
"eval_steps_per_second": 0.602,
"step": 3649
},
{
"epoch": 22.004521477015825,
"grad_norm": 1.0649685278407737,
"learning_rate": 5e-06,
"loss": 0.0218,
"step": 3650
},
{
"epoch": 22.064807837226827,
"grad_norm": 1.2396804319512318,
"learning_rate": 5e-06,
"loss": 0.0158,
"step": 3660
},
{
"epoch": 22.12509419743783,
"grad_norm": 1.2203455145168154,
"learning_rate": 5e-06,
"loss": 0.0163,
"step": 3670
},
{
"epoch": 22.18538055764883,
"grad_norm": 1.1547196338255674,
"learning_rate": 5e-06,
"loss": 0.0169,
"step": 3680
},
{
"epoch": 22.245666917859833,
"grad_norm": 1.309239398813011,
"learning_rate": 5e-06,
"loss": 0.0173,
"step": 3690
},
{
"epoch": 22.305953278070838,
"grad_norm": 1.2683908516455489,
"learning_rate": 5e-06,
"loss": 0.0173,
"step": 3700
},
{
"epoch": 22.36623963828184,
"grad_norm": 1.5982780924133835,
"learning_rate": 5e-06,
"loss": 0.0183,
"step": 3710
},
{
"epoch": 22.426525998492842,
"grad_norm": 1.4555126747204115,
"learning_rate": 5e-06,
"loss": 0.0184,
"step": 3720
},
{
"epoch": 22.486812358703844,
"grad_norm": 1.3675517314579124,
"learning_rate": 5e-06,
"loss": 0.0186,
"step": 3730
},
{
"epoch": 22.547098718914846,
"grad_norm": 1.2088813887554757,
"learning_rate": 5e-06,
"loss": 0.0188,
"step": 3740
},
{
"epoch": 22.607385079125848,
"grad_norm": 1.157896013440299,
"learning_rate": 5e-06,
"loss": 0.0194,
"step": 3750
},
{
"epoch": 22.66767143933685,
"grad_norm": 1.4152388376390785,
"learning_rate": 5e-06,
"loss": 0.0189,
"step": 3760
},
{
"epoch": 22.727957799547852,
"grad_norm": 1.2554294042934306,
"learning_rate": 5e-06,
"loss": 0.019,
"step": 3770
},
{
"epoch": 22.788244159758854,
"grad_norm": 1.3329931616963409,
"learning_rate": 5e-06,
"loss": 0.019,
"step": 3780
},
{
"epoch": 22.848530519969856,
"grad_norm": 1.256565715756104,
"learning_rate": 5e-06,
"loss": 0.0189,
"step": 3790
},
{
"epoch": 22.908816880180858,
"grad_norm": 1.368011051794652,
"learning_rate": 5e-06,
"loss": 0.0195,
"step": 3800
},
{
"epoch": 22.96910324039186,
"grad_norm": 1.4311223887923525,
"learning_rate": 5e-06,
"loss": 0.0197,
"step": 3810
},
{
"epoch": 22.99924642049736,
"eval_loss": 2.359902858734131,
"eval_runtime": 117.0325,
"eval_samples_per_second": 38.186,
"eval_steps_per_second": 0.598,
"step": 3815
},
{
"epoch": 23.029389600602865,
"grad_norm": 1.114496963570823,
"learning_rate": 5e-06,
"loss": 0.0171,
"step": 3820
},
{
"epoch": 23.089675960813867,
"grad_norm": 1.1656429989984474,
"learning_rate": 5e-06,
"loss": 0.0149,
"step": 3830
},
{
"epoch": 23.14996232102487,
"grad_norm": 1.3181375592191023,
"learning_rate": 5e-06,
"loss": 0.0152,
"step": 3840
},
{
"epoch": 23.21024868123587,
"grad_norm": 1.214891422163979,
"learning_rate": 5e-06,
"loss": 0.0157,
"step": 3850
},
{
"epoch": 23.270535041446873,
"grad_norm": 1.1329477739782727,
"learning_rate": 5e-06,
"loss": 0.0159,
"step": 3860
},
{
"epoch": 23.330821401657875,
"grad_norm": 1.178978239439168,
"learning_rate": 5e-06,
"loss": 0.0163,
"step": 3870
},
{
"epoch": 23.391107761868877,
"grad_norm": 1.1072042006237766,
"learning_rate": 5e-06,
"loss": 0.0165,
"step": 3880
},
{
"epoch": 23.45139412207988,
"grad_norm": 1.2281267859101168,
"learning_rate": 5e-06,
"loss": 0.0168,
"step": 3890
},
{
"epoch": 23.51168048229088,
"grad_norm": 1.1558899107856422,
"learning_rate": 5e-06,
"loss": 0.0174,
"step": 3900
},
{
"epoch": 23.571966842501883,
"grad_norm": 1.217612194947995,
"learning_rate": 5e-06,
"loss": 0.0173,
"step": 3910
},
{
"epoch": 23.632253202712885,
"grad_norm": 1.2440180440988633,
"learning_rate": 5e-06,
"loss": 0.0174,
"step": 3920
},
{
"epoch": 23.692539562923887,
"grad_norm": 1.3420865819465875,
"learning_rate": 5e-06,
"loss": 0.0174,
"step": 3930
},
{
"epoch": 23.75282592313489,
"grad_norm": 1.3593190067484822,
"learning_rate": 5e-06,
"loss": 0.018,
"step": 3940
},
{
"epoch": 23.813112283345895,
"grad_norm": 1.2367975594859635,
"learning_rate": 5e-06,
"loss": 0.0177,
"step": 3950
},
{
"epoch": 23.873398643556897,
"grad_norm": 1.2214254875430963,
"learning_rate": 5e-06,
"loss": 0.0179,
"step": 3960
},
{
"epoch": 23.9336850037679,
"grad_norm": 1.20689456847013,
"learning_rate": 5e-06,
"loss": 0.018,
"step": 3970
},
{
"epoch": 23.9939713639789,
"grad_norm": 1.2666923485205688,
"learning_rate": 5e-06,
"loss": 0.0186,
"step": 3980
},
{
"epoch": 24.0,
"eval_loss": 2.3888232707977295,
"eval_runtime": 116.343,
"eval_samples_per_second": 38.412,
"eval_steps_per_second": 0.602,
"step": 3981
},
{
"epoch": 24.054257724189902,
"grad_norm": 1.1066074003233175,
"learning_rate": 5e-06,
"loss": 0.0141,
"step": 3990
},
{
"epoch": 24.114544084400904,
"grad_norm": 1.0389652033545478,
"learning_rate": 5e-06,
"loss": 0.0141,
"step": 4000
},
{
"epoch": 24.174830444611906,
"grad_norm": 1.1258379344398495,
"learning_rate": 5e-06,
"loss": 0.0144,
"step": 4010
},
{
"epoch": 24.23511680482291,
"grad_norm": 1.1758249329629795,
"learning_rate": 5e-06,
"loss": 0.0148,
"step": 4020
},
{
"epoch": 24.29540316503391,
"grad_norm": 1.1741885490501298,
"learning_rate": 5e-06,
"loss": 0.0151,
"step": 4030
},
{
"epoch": 24.355689525244912,
"grad_norm": 1.1622125971160786,
"learning_rate": 5e-06,
"loss": 0.0156,
"step": 4040
},
{
"epoch": 24.415975885455914,
"grad_norm": 1.0899820074898727,
"learning_rate": 5e-06,
"loss": 0.0154,
"step": 4050
},
{
"epoch": 24.476262245666916,
"grad_norm": 1.2599823712922462,
"learning_rate": 5e-06,
"loss": 0.0159,
"step": 4060
},
{
"epoch": 24.53654860587792,
"grad_norm": 1.140178001121795,
"learning_rate": 5e-06,
"loss": 0.0157,
"step": 4070
},
{
"epoch": 24.596834966088924,
"grad_norm": 1.1230288891139422,
"learning_rate": 5e-06,
"loss": 0.0158,
"step": 4080
},
{
"epoch": 24.657121326299926,
"grad_norm": 1.1218784125193524,
"learning_rate": 5e-06,
"loss": 0.0163,
"step": 4090
},
{
"epoch": 24.717407686510928,
"grad_norm": 1.1524016261830858,
"learning_rate": 5e-06,
"loss": 0.0164,
"step": 4100
},
{
"epoch": 24.77769404672193,
"grad_norm": 1.15414666614049,
"learning_rate": 5e-06,
"loss": 0.0166,
"step": 4110
},
{
"epoch": 24.83798040693293,
"grad_norm": 1.3244431296269448,
"learning_rate": 5e-06,
"loss": 0.0169,
"step": 4120
},
{
"epoch": 24.868123587038433,
"eval_loss": 2.4057514667510986,
"eval_runtime": 113.5854,
"eval_samples_per_second": 39.345,
"eval_steps_per_second": 0.616,
"step": 4125
},
{
"epoch": 24.868123587038433,
"step": 4125,
"total_flos": 6909319257784320.0,
"train_loss": 0.2661740447047985,
"train_runtime": 139979.5045,
"train_samples_per_second": 15.162,
"train_steps_per_second": 0.029
}
],
"logging_steps": 10,
"max_steps": 4125,
"num_input_tokens_seen": 0,
"num_train_epochs": 25,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 6909319257784320.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}