QwQ-RP-LoRA / checkpoint-648 /trainer_state.json
Undi95's picture
Upload folder using huggingface_hub
75b9830 verified
raw
history blame
114 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9992289899768697,
"eval_steps": 500,
"global_step": 648,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0015420200462606013,
"grad_norm": 0.06960189342498779,
"learning_rate": 1e-05,
"loss": 0.9546,
"step": 1
},
{
"epoch": 0.0030840400925212026,
"grad_norm": 0.07705054432153702,
"learning_rate": 2e-05,
"loss": 1.0023,
"step": 2
},
{
"epoch": 0.004626060138781804,
"grad_norm": 0.07115544378757477,
"learning_rate": 3e-05,
"loss": 0.9733,
"step": 3
},
{
"epoch": 0.006168080185042405,
"grad_norm": 0.06770255416631699,
"learning_rate": 4e-05,
"loss": 0.8908,
"step": 4
},
{
"epoch": 0.007710100231303007,
"grad_norm": 0.06454406678676605,
"learning_rate": 5e-05,
"loss": 0.9168,
"step": 5
},
{
"epoch": 0.009252120277563608,
"grad_norm": 0.0818350687623024,
"learning_rate": 6e-05,
"loss": 0.8822,
"step": 6
},
{
"epoch": 0.01079414032382421,
"grad_norm": 0.08504347503185272,
"learning_rate": 7e-05,
"loss": 0.9563,
"step": 7
},
{
"epoch": 0.01233616037008481,
"grad_norm": 0.08642569184303284,
"learning_rate": 8e-05,
"loss": 0.9324,
"step": 8
},
{
"epoch": 0.013878180416345412,
"grad_norm": 0.07200746238231659,
"learning_rate": 9e-05,
"loss": 0.8557,
"step": 9
},
{
"epoch": 0.015420200462606014,
"grad_norm": 0.0539001002907753,
"learning_rate": 0.0001,
"loss": 0.8316,
"step": 10
},
{
"epoch": 0.016962220508866616,
"grad_norm": 0.05532313138246536,
"learning_rate": 0.00011000000000000002,
"loss": 0.869,
"step": 11
},
{
"epoch": 0.018504240555127217,
"grad_norm": 0.04828835651278496,
"learning_rate": 0.00012,
"loss": 0.9004,
"step": 12
},
{
"epoch": 0.020046260601387818,
"grad_norm": 0.051899779587984085,
"learning_rate": 0.00013000000000000002,
"loss": 0.7818,
"step": 13
},
{
"epoch": 0.02158828064764842,
"grad_norm": 0.07032614946365356,
"learning_rate": 0.00014,
"loss": 0.7946,
"step": 14
},
{
"epoch": 0.02313030069390902,
"grad_norm": 0.08230099081993103,
"learning_rate": 0.00015000000000000001,
"loss": 0.934,
"step": 15
},
{
"epoch": 0.02467232074016962,
"grad_norm": 0.08811169862747192,
"learning_rate": 0.00016,
"loss": 0.8621,
"step": 16
},
{
"epoch": 0.026214340786430222,
"grad_norm": 0.062236446887254715,
"learning_rate": 0.00017,
"loss": 0.8186,
"step": 17
},
{
"epoch": 0.027756360832690823,
"grad_norm": 0.052294306457042694,
"learning_rate": 0.00018,
"loss": 0.8124,
"step": 18
},
{
"epoch": 0.029298380878951428,
"grad_norm": 0.032340776175260544,
"learning_rate": 0.00019,
"loss": 0.8598,
"step": 19
},
{
"epoch": 0.03084040092521203,
"grad_norm": 0.028073711320757866,
"learning_rate": 0.0002,
"loss": 0.9087,
"step": 20
},
{
"epoch": 0.03238242097147263,
"grad_norm": 0.027820633724331856,
"learning_rate": 0.00019999969691239107,
"loss": 0.7093,
"step": 21
},
{
"epoch": 0.03392444101773323,
"grad_norm": 0.032430652529001236,
"learning_rate": 0.00019999878765140146,
"loss": 0.7825,
"step": 22
},
{
"epoch": 0.03546646106399383,
"grad_norm": 0.034685954451560974,
"learning_rate": 0.000199997272222543,
"loss": 0.7787,
"step": 23
},
{
"epoch": 0.03700848111025443,
"grad_norm": 0.033446088433265686,
"learning_rate": 0.0001999951506350017,
"loss": 0.8636,
"step": 24
},
{
"epoch": 0.03855050115651504,
"grad_norm": 0.026805371046066284,
"learning_rate": 0.0001999924229016382,
"loss": 0.904,
"step": 25
},
{
"epoch": 0.040092521202775636,
"grad_norm": 0.024178853258490562,
"learning_rate": 0.0001999890890389873,
"loss": 0.7428,
"step": 26
},
{
"epoch": 0.04163454124903624,
"grad_norm": 0.024074744433164597,
"learning_rate": 0.00019998514906725804,
"loss": 0.8562,
"step": 27
},
{
"epoch": 0.04317656129529684,
"grad_norm": 0.025082379579544067,
"learning_rate": 0.0001999806030103336,
"loss": 0.8078,
"step": 28
},
{
"epoch": 0.04471858134155744,
"grad_norm": 0.027726231142878532,
"learning_rate": 0.00019997545089577102,
"loss": 0.7133,
"step": 29
},
{
"epoch": 0.04626060138781804,
"grad_norm": 0.02736576274037361,
"learning_rate": 0.00019996969275480115,
"loss": 0.7826,
"step": 30
},
{
"epoch": 0.047802621434078645,
"grad_norm": 0.0234544537961483,
"learning_rate": 0.0001999633286223284,
"loss": 0.7424,
"step": 31
},
{
"epoch": 0.04934464148033924,
"grad_norm": 0.025730784982442856,
"learning_rate": 0.00019995635853693058,
"loss": 0.7584,
"step": 32
},
{
"epoch": 0.05088666152659985,
"grad_norm": 0.02729860506951809,
"learning_rate": 0.00019994878254085861,
"loss": 0.73,
"step": 33
},
{
"epoch": 0.052428681572860444,
"grad_norm": 0.027428491041064262,
"learning_rate": 0.00019994060068003627,
"loss": 0.854,
"step": 34
},
{
"epoch": 0.05397070161912105,
"grad_norm": 0.024362564086914062,
"learning_rate": 0.00019993181300406005,
"loss": 0.7441,
"step": 35
},
{
"epoch": 0.05551272166538165,
"grad_norm": 0.024825185537338257,
"learning_rate": 0.00019992241956619863,
"loss": 0.7367,
"step": 36
},
{
"epoch": 0.05705474171164225,
"grad_norm": 0.023908289149403572,
"learning_rate": 0.00019991242042339264,
"loss": 0.8266,
"step": 37
},
{
"epoch": 0.058596761757902856,
"grad_norm": 0.023762725293636322,
"learning_rate": 0.00019990181563625447,
"loss": 0.829,
"step": 38
},
{
"epoch": 0.06013878180416345,
"grad_norm": 0.02654297836124897,
"learning_rate": 0.0001998906052690677,
"loss": 0.7635,
"step": 39
},
{
"epoch": 0.06168080185042406,
"grad_norm": 0.026318082585930824,
"learning_rate": 0.00019987878938978684,
"loss": 0.7623,
"step": 40
},
{
"epoch": 0.06322282189668466,
"grad_norm": 0.024309197440743446,
"learning_rate": 0.00019986636807003673,
"loss": 0.7494,
"step": 41
},
{
"epoch": 0.06476484194294525,
"grad_norm": 0.022980719804763794,
"learning_rate": 0.00019985334138511237,
"loss": 0.748,
"step": 42
},
{
"epoch": 0.06630686198920586,
"grad_norm": 0.023796193301677704,
"learning_rate": 0.00019983970941397835,
"loss": 0.8684,
"step": 43
},
{
"epoch": 0.06784888203546646,
"grad_norm": 0.02435590885579586,
"learning_rate": 0.00019982547223926824,
"loss": 0.6749,
"step": 44
},
{
"epoch": 0.06939090208172706,
"grad_norm": 0.026482658460736275,
"learning_rate": 0.0001998106299472843,
"loss": 0.7684,
"step": 45
},
{
"epoch": 0.07093292212798766,
"grad_norm": 0.025558117777109146,
"learning_rate": 0.0001997951826279968,
"loss": 0.7507,
"step": 46
},
{
"epoch": 0.07247494217424827,
"grad_norm": 0.027371902018785477,
"learning_rate": 0.00019977913037504355,
"loss": 0.7377,
"step": 47
},
{
"epoch": 0.07401696222050887,
"grad_norm": 0.0234097708016634,
"learning_rate": 0.00019976247328572938,
"loss": 0.7675,
"step": 48
},
{
"epoch": 0.07555898226676946,
"grad_norm": 0.0241215992718935,
"learning_rate": 0.00019974521146102537,
"loss": 0.8079,
"step": 49
},
{
"epoch": 0.07710100231303008,
"grad_norm": 0.025045258924365044,
"learning_rate": 0.00019972734500556846,
"loss": 0.8407,
"step": 50
},
{
"epoch": 0.07864302235929067,
"grad_norm": 0.02574036829173565,
"learning_rate": 0.0001997088740276607,
"loss": 0.767,
"step": 51
},
{
"epoch": 0.08018504240555127,
"grad_norm": 0.02342085726559162,
"learning_rate": 0.00019968979863926856,
"loss": 0.7373,
"step": 52
},
{
"epoch": 0.08172706245181187,
"grad_norm": 0.023467406630516052,
"learning_rate": 0.0001996701189560223,
"loss": 0.6529,
"step": 53
},
{
"epoch": 0.08326908249807248,
"grad_norm": 0.02330499142408371,
"learning_rate": 0.00019964983509721527,
"loss": 0.7347,
"step": 54
},
{
"epoch": 0.08481110254433308,
"grad_norm": 0.027940964326262474,
"learning_rate": 0.00019962894718580324,
"loss": 0.9313,
"step": 55
},
{
"epoch": 0.08635312259059368,
"grad_norm": 0.024525761604309082,
"learning_rate": 0.00019960745534840354,
"loss": 0.767,
"step": 56
},
{
"epoch": 0.08789514263685427,
"grad_norm": 0.024538526311516762,
"learning_rate": 0.00019958535971529434,
"loss": 0.7659,
"step": 57
},
{
"epoch": 0.08943716268311488,
"grad_norm": 0.02253701537847519,
"learning_rate": 0.00019956266042041394,
"loss": 0.7805,
"step": 58
},
{
"epoch": 0.09097918272937548,
"grad_norm": 0.023676637560129166,
"learning_rate": 0.0001995393576013598,
"loss": 0.7894,
"step": 59
},
{
"epoch": 0.09252120277563608,
"grad_norm": 0.022134529426693916,
"learning_rate": 0.0001995154513993878,
"loss": 0.7492,
"step": 60
},
{
"epoch": 0.09406322282189669,
"grad_norm": 0.02350509911775589,
"learning_rate": 0.00019949094195941152,
"loss": 0.6902,
"step": 61
},
{
"epoch": 0.09560524286815729,
"grad_norm": 0.02464171312749386,
"learning_rate": 0.00019946582943000102,
"loss": 0.7836,
"step": 62
},
{
"epoch": 0.09714726291441789,
"grad_norm": 0.023095758631825447,
"learning_rate": 0.00019944011396338222,
"loss": 0.8321,
"step": 63
},
{
"epoch": 0.09868928296067848,
"grad_norm": 0.026240425184369087,
"learning_rate": 0.00019941379571543596,
"loss": 0.8461,
"step": 64
},
{
"epoch": 0.1002313030069391,
"grad_norm": 0.02174345962703228,
"learning_rate": 0.00019938687484569693,
"loss": 0.6388,
"step": 65
},
{
"epoch": 0.1017733230531997,
"grad_norm": 0.02867325395345688,
"learning_rate": 0.00019935935151735277,
"loss": 0.826,
"step": 66
},
{
"epoch": 0.10331534309946029,
"grad_norm": 0.02631618268787861,
"learning_rate": 0.00019933122589724302,
"loss": 0.9407,
"step": 67
},
{
"epoch": 0.10485736314572089,
"grad_norm": 0.022020747885107994,
"learning_rate": 0.0001993024981558583,
"loss": 0.7004,
"step": 68
},
{
"epoch": 0.1063993831919815,
"grad_norm": 0.02743780054152012,
"learning_rate": 0.000199273168467339,
"loss": 0.7607,
"step": 69
},
{
"epoch": 0.1079414032382421,
"grad_norm": 0.028378015384078026,
"learning_rate": 0.00019924323700947448,
"loss": 0.8604,
"step": 70
},
{
"epoch": 0.1094834232845027,
"grad_norm": 0.0275627039372921,
"learning_rate": 0.00019921270396370172,
"loss": 0.835,
"step": 71
},
{
"epoch": 0.1110254433307633,
"grad_norm": 0.023099975660443306,
"learning_rate": 0.0001991815695151046,
"loss": 0.7028,
"step": 72
},
{
"epoch": 0.1125674633770239,
"grad_norm": 0.028545403853058815,
"learning_rate": 0.00019914983385241236,
"loss": 0.8248,
"step": 73
},
{
"epoch": 0.1141094834232845,
"grad_norm": 0.02746577188372612,
"learning_rate": 0.00019911749716799873,
"loss": 0.7309,
"step": 74
},
{
"epoch": 0.1156515034695451,
"grad_norm": 0.024899670854210854,
"learning_rate": 0.00019908455965788067,
"loss": 0.7473,
"step": 75
},
{
"epoch": 0.11719352351580571,
"grad_norm": 0.024973087012767792,
"learning_rate": 0.00019905102152171727,
"loss": 0.8362,
"step": 76
},
{
"epoch": 0.11873554356206631,
"grad_norm": 0.023668723180890083,
"learning_rate": 0.0001990168829628083,
"loss": 0.7677,
"step": 77
},
{
"epoch": 0.1202775636083269,
"grad_norm": 0.02495860867202282,
"learning_rate": 0.0001989821441880933,
"loss": 0.7341,
"step": 78
},
{
"epoch": 0.1218195836545875,
"grad_norm": 0.02537156455218792,
"learning_rate": 0.00019894680540815006,
"loss": 0.6767,
"step": 79
},
{
"epoch": 0.12336160370084812,
"grad_norm": 0.0246786717325449,
"learning_rate": 0.0001989108668371936,
"loss": 0.7959,
"step": 80
},
{
"epoch": 0.12490362374710871,
"grad_norm": 0.02471376582980156,
"learning_rate": 0.00019887432869307458,
"loss": 0.6787,
"step": 81
},
{
"epoch": 0.1264456437933693,
"grad_norm": 0.025275586172938347,
"learning_rate": 0.00019883719119727816,
"loss": 0.7753,
"step": 82
},
{
"epoch": 0.12798766383962992,
"grad_norm": 0.021094506606459618,
"learning_rate": 0.00019879945457492267,
"loss": 0.758,
"step": 83
},
{
"epoch": 0.1295296838858905,
"grad_norm": 0.02534683421254158,
"learning_rate": 0.00019876111905475815,
"loss": 0.818,
"step": 84
},
{
"epoch": 0.13107170393215112,
"grad_norm": 0.024033140391111374,
"learning_rate": 0.00019872218486916498,
"loss": 0.775,
"step": 85
},
{
"epoch": 0.13261372397841173,
"grad_norm": 0.023884933441877365,
"learning_rate": 0.00019868265225415265,
"loss": 0.7918,
"step": 86
},
{
"epoch": 0.1341557440246723,
"grad_norm": 0.026067111641168594,
"learning_rate": 0.00019864252144935794,
"loss": 0.8368,
"step": 87
},
{
"epoch": 0.13569776407093292,
"grad_norm": 0.02631264552474022,
"learning_rate": 0.00019860179269804394,
"loss": 0.7138,
"step": 88
},
{
"epoch": 0.13723978411719354,
"grad_norm": 0.022870918735861778,
"learning_rate": 0.00019856046624709822,
"loss": 0.7423,
"step": 89
},
{
"epoch": 0.13878180416345412,
"grad_norm": 0.02922765538096428,
"learning_rate": 0.00019851854234703145,
"loss": 0.7498,
"step": 90
},
{
"epoch": 0.14032382420971473,
"grad_norm": 0.02589617855846882,
"learning_rate": 0.00019847602125197598,
"loss": 0.7438,
"step": 91
},
{
"epoch": 0.14186584425597532,
"grad_norm": 0.02394738420844078,
"learning_rate": 0.00019843290321968412,
"loss": 0.7094,
"step": 92
},
{
"epoch": 0.14340786430223593,
"grad_norm": 0.02237016148865223,
"learning_rate": 0.0001983891885115267,
"loss": 0.7868,
"step": 93
},
{
"epoch": 0.14494988434849654,
"grad_norm": 0.028733767569065094,
"learning_rate": 0.00019834487739249146,
"loss": 0.8178,
"step": 94
},
{
"epoch": 0.14649190439475712,
"grad_norm": 0.023086342960596085,
"learning_rate": 0.0001982999701311814,
"loss": 0.8368,
"step": 95
},
{
"epoch": 0.14803392444101773,
"grad_norm": 0.025624489411711693,
"learning_rate": 0.0001982544669998132,
"loss": 0.7731,
"step": 96
},
{
"epoch": 0.14957594448727835,
"grad_norm": 0.028302457183599472,
"learning_rate": 0.0001982083682742156,
"loss": 0.7942,
"step": 97
},
{
"epoch": 0.15111796453353893,
"grad_norm": 0.025008324533700943,
"learning_rate": 0.00019816167423382765,
"loss": 0.7393,
"step": 98
},
{
"epoch": 0.15265998457979954,
"grad_norm": 0.026291735470294952,
"learning_rate": 0.00019811438516169702,
"loss": 0.8016,
"step": 99
},
{
"epoch": 0.15420200462606015,
"grad_norm": 0.031547173857688904,
"learning_rate": 0.00019806650134447838,
"loss": 0.8597,
"step": 100
},
{
"epoch": 0.15574402467232074,
"grad_norm": 0.024978285655379295,
"learning_rate": 0.00019801802307243153,
"loss": 0.8182,
"step": 101
},
{
"epoch": 0.15728604471858135,
"grad_norm": 0.023977672681212425,
"learning_rate": 0.00019796895063941978,
"loss": 0.8374,
"step": 102
},
{
"epoch": 0.15882806476484193,
"grad_norm": 0.025743963196873665,
"learning_rate": 0.000197919284342908,
"loss": 0.7326,
"step": 103
},
{
"epoch": 0.16037008481110254,
"grad_norm": 0.02554011158645153,
"learning_rate": 0.00019786902448396104,
"loss": 0.7703,
"step": 104
},
{
"epoch": 0.16191210485736315,
"grad_norm": 0.027971483767032623,
"learning_rate": 0.00019781817136724165,
"loss": 0.6845,
"step": 105
},
{
"epoch": 0.16345412490362374,
"grad_norm": 0.02789183147251606,
"learning_rate": 0.00019776672530100886,
"loss": 0.838,
"step": 106
},
{
"epoch": 0.16499614494988435,
"grad_norm": 0.029534442350268364,
"learning_rate": 0.00019771468659711595,
"loss": 0.7428,
"step": 107
},
{
"epoch": 0.16653816499614496,
"grad_norm": 0.02910265140235424,
"learning_rate": 0.00019766205557100868,
"loss": 0.8891,
"step": 108
},
{
"epoch": 0.16808018504240554,
"grad_norm": 0.02285209856927395,
"learning_rate": 0.00019760883254172327,
"loss": 0.7425,
"step": 109
},
{
"epoch": 0.16962220508866616,
"grad_norm": 0.025452135130763054,
"learning_rate": 0.0001975550178318845,
"loss": 0.7617,
"step": 110
},
{
"epoch": 0.17116422513492677,
"grad_norm": 0.02843882516026497,
"learning_rate": 0.00019750061176770385,
"loss": 0.9045,
"step": 111
},
{
"epoch": 0.17270624518118735,
"grad_norm": 0.026800749823451042,
"learning_rate": 0.00019744561467897735,
"loss": 0.72,
"step": 112
},
{
"epoch": 0.17424826522744796,
"grad_norm": 0.021131988614797592,
"learning_rate": 0.00019739002689908377,
"loss": 0.6511,
"step": 113
},
{
"epoch": 0.17579028527370855,
"grad_norm": 0.02588481456041336,
"learning_rate": 0.00019733384876498245,
"loss": 0.7168,
"step": 114
},
{
"epoch": 0.17733230531996916,
"grad_norm": 0.031303439289331436,
"learning_rate": 0.00019727708061721133,
"loss": 0.8685,
"step": 115
},
{
"epoch": 0.17887432536622977,
"grad_norm": 0.02867058850824833,
"learning_rate": 0.00019721972279988477,
"loss": 0.655,
"step": 116
},
{
"epoch": 0.18041634541249035,
"grad_norm": 0.03866586834192276,
"learning_rate": 0.00019716177566069174,
"loss": 0.7957,
"step": 117
},
{
"epoch": 0.18195836545875096,
"grad_norm": 0.029205329716205597,
"learning_rate": 0.00019710323955089343,
"loss": 0.7617,
"step": 118
},
{
"epoch": 0.18350038550501158,
"grad_norm": 0.024928180500864983,
"learning_rate": 0.00019704411482532116,
"loss": 0.6982,
"step": 119
},
{
"epoch": 0.18504240555127216,
"grad_norm": 0.02545573003590107,
"learning_rate": 0.0001969844018423744,
"loss": 0.6067,
"step": 120
},
{
"epoch": 0.18658442559753277,
"grad_norm": 0.02810928039252758,
"learning_rate": 0.0001969241009640185,
"loss": 0.7112,
"step": 121
},
{
"epoch": 0.18812644564379338,
"grad_norm": 0.022296108305454254,
"learning_rate": 0.00019686321255578238,
"loss": 0.6598,
"step": 122
},
{
"epoch": 0.18966846569005397,
"grad_norm": 0.02429027482867241,
"learning_rate": 0.00019680173698675648,
"loss": 0.6381,
"step": 123
},
{
"epoch": 0.19121048573631458,
"grad_norm": 0.027076730504631996,
"learning_rate": 0.0001967396746295905,
"loss": 0.717,
"step": 124
},
{
"epoch": 0.19275250578257516,
"grad_norm": 0.02401566132903099,
"learning_rate": 0.00019667702586049108,
"loss": 0.8002,
"step": 125
},
{
"epoch": 0.19429452582883577,
"grad_norm": 0.024678878486156464,
"learning_rate": 0.00019661379105921948,
"loss": 0.7834,
"step": 126
},
{
"epoch": 0.19583654587509639,
"grad_norm": 0.029240388423204422,
"learning_rate": 0.00019654997060908946,
"loss": 0.8793,
"step": 127
},
{
"epoch": 0.19737856592135697,
"grad_norm": 0.02550147846341133,
"learning_rate": 0.0001964855648969647,
"loss": 0.6742,
"step": 128
},
{
"epoch": 0.19892058596761758,
"grad_norm": 0.02416900172829628,
"learning_rate": 0.00019642057431325672,
"loss": 0.7728,
"step": 129
},
{
"epoch": 0.2004626060138782,
"grad_norm": 0.024728331714868546,
"learning_rate": 0.0001963549992519223,
"loss": 0.7237,
"step": 130
},
{
"epoch": 0.20200462606013878,
"grad_norm": 0.025203561410307884,
"learning_rate": 0.00019628884011046123,
"loss": 0.7491,
"step": 131
},
{
"epoch": 0.2035466461063994,
"grad_norm": 0.02104656957089901,
"learning_rate": 0.00019622209728991383,
"loss": 0.7324,
"step": 132
},
{
"epoch": 0.20508866615265997,
"grad_norm": 0.03518475592136383,
"learning_rate": 0.00019615477119485855,
"loss": 0.8982,
"step": 133
},
{
"epoch": 0.20663068619892058,
"grad_norm": 0.026010941714048386,
"learning_rate": 0.00019608686223340945,
"loss": 0.7451,
"step": 134
},
{
"epoch": 0.2081727062451812,
"grad_norm": 0.025345437228679657,
"learning_rate": 0.00019601837081721386,
"loss": 0.705,
"step": 135
},
{
"epoch": 0.20971472629144178,
"grad_norm": 0.02374056540429592,
"learning_rate": 0.00019594929736144976,
"loss": 0.7307,
"step": 136
},
{
"epoch": 0.2112567463377024,
"grad_norm": 0.026990080252289772,
"learning_rate": 0.00019587964228482332,
"loss": 0.8173,
"step": 137
},
{
"epoch": 0.212798766383963,
"grad_norm": 0.026662928983569145,
"learning_rate": 0.00019580940600956638,
"loss": 0.818,
"step": 138
},
{
"epoch": 0.21434078643022358,
"grad_norm": 0.026762284338474274,
"learning_rate": 0.00019573858896143376,
"loss": 0.6674,
"step": 139
},
{
"epoch": 0.2158828064764842,
"grad_norm": 0.02683679386973381,
"learning_rate": 0.00019566719156970095,
"loss": 0.7569,
"step": 140
},
{
"epoch": 0.2174248265227448,
"grad_norm": 0.022990800440311432,
"learning_rate": 0.00019559521426716118,
"loss": 0.7051,
"step": 141
},
{
"epoch": 0.2189668465690054,
"grad_norm": 0.022913858294487,
"learning_rate": 0.00019552265749012303,
"loss": 0.7483,
"step": 142
},
{
"epoch": 0.220508866615266,
"grad_norm": 0.025732524693012238,
"learning_rate": 0.00019544952167840777,
"loss": 0.8484,
"step": 143
},
{
"epoch": 0.2220508866615266,
"grad_norm": 0.024254556745290756,
"learning_rate": 0.00019537580727534644,
"loss": 0.747,
"step": 144
},
{
"epoch": 0.2235929067077872,
"grad_norm": 0.02940620854496956,
"learning_rate": 0.0001953015147277776,
"loss": 0.8204,
"step": 145
},
{
"epoch": 0.2251349267540478,
"grad_norm": 0.02590208128094673,
"learning_rate": 0.00019522664448604418,
"loss": 0.7778,
"step": 146
},
{
"epoch": 0.2266769468003084,
"grad_norm": 0.02748725563287735,
"learning_rate": 0.00019515119700399107,
"loss": 0.7782,
"step": 147
},
{
"epoch": 0.228218966846569,
"grad_norm": 0.024179786443710327,
"learning_rate": 0.00019507517273896222,
"loss": 0.7651,
"step": 148
},
{
"epoch": 0.22976098689282962,
"grad_norm": 0.02463974617421627,
"learning_rate": 0.00019499857215179786,
"loss": 0.8215,
"step": 149
},
{
"epoch": 0.2313030069390902,
"grad_norm": 0.024559814482927322,
"learning_rate": 0.00019492139570683178,
"loss": 0.7539,
"step": 150
},
{
"epoch": 0.2328450269853508,
"grad_norm": 0.023420870304107666,
"learning_rate": 0.00019484364387188847,
"loss": 0.7035,
"step": 151
},
{
"epoch": 0.23438704703161142,
"grad_norm": 0.026096729561686516,
"learning_rate": 0.00019476531711828027,
"loss": 0.8033,
"step": 152
},
{
"epoch": 0.235929067077872,
"grad_norm": 0.02388446033000946,
"learning_rate": 0.0001946864159208045,
"loss": 0.6746,
"step": 153
},
{
"epoch": 0.23747108712413262,
"grad_norm": 0.02306438237428665,
"learning_rate": 0.0001946069407577408,
"loss": 0.7062,
"step": 154
},
{
"epoch": 0.2390131071703932,
"grad_norm": 0.024697955697774887,
"learning_rate": 0.00019452689211084775,
"loss": 0.7691,
"step": 155
},
{
"epoch": 0.2405551272166538,
"grad_norm": 0.026947690173983574,
"learning_rate": 0.00019444627046536056,
"loss": 0.7347,
"step": 156
},
{
"epoch": 0.24209714726291442,
"grad_norm": 0.02345297671854496,
"learning_rate": 0.00019436507630998757,
"loss": 0.745,
"step": 157
},
{
"epoch": 0.243639167309175,
"grad_norm": 0.029198188334703445,
"learning_rate": 0.00019428331013690765,
"loss": 0.7862,
"step": 158
},
{
"epoch": 0.24518118735543562,
"grad_norm": 0.025465266779065132,
"learning_rate": 0.00019420097244176706,
"loss": 0.6685,
"step": 159
},
{
"epoch": 0.24672320740169623,
"grad_norm": 0.02786502055823803,
"learning_rate": 0.00019411806372367655,
"loss": 0.725,
"step": 160
},
{
"epoch": 0.24826522744795682,
"grad_norm": 0.02317357063293457,
"learning_rate": 0.0001940345844852082,
"loss": 0.7075,
"step": 161
},
{
"epoch": 0.24980724749421743,
"grad_norm": 0.02511444129049778,
"learning_rate": 0.00019395053523239245,
"loss": 0.7102,
"step": 162
},
{
"epoch": 0.25134926754047804,
"grad_norm": 0.02567203901708126,
"learning_rate": 0.00019386591647471506,
"loss": 0.8113,
"step": 163
},
{
"epoch": 0.2528912875867386,
"grad_norm": 0.02611825056374073,
"learning_rate": 0.00019378072872511398,
"loss": 0.786,
"step": 164
},
{
"epoch": 0.2544333076329992,
"grad_norm": 0.02788010984659195,
"learning_rate": 0.0001936949724999762,
"loss": 0.684,
"step": 165
},
{
"epoch": 0.25597532767925985,
"grad_norm": 0.026200013235211372,
"learning_rate": 0.0001936086483191347,
"loss": 0.7563,
"step": 166
},
{
"epoch": 0.25751734772552043,
"grad_norm": 0.024887658655643463,
"learning_rate": 0.00019352175670586533,
"loss": 0.6896,
"step": 167
},
{
"epoch": 0.259059367771781,
"grad_norm": 0.027471961453557014,
"learning_rate": 0.00019343429818688347,
"loss": 0.8109,
"step": 168
},
{
"epoch": 0.26060138781804165,
"grad_norm": 0.024350160732865334,
"learning_rate": 0.00019334627329234102,
"loss": 0.7178,
"step": 169
},
{
"epoch": 0.26214340786430224,
"grad_norm": 0.02346990443766117,
"learning_rate": 0.00019325768255582302,
"loss": 0.6508,
"step": 170
},
{
"epoch": 0.2636854279105628,
"grad_norm": 0.028655072674155235,
"learning_rate": 0.00019316852651434462,
"loss": 0.7036,
"step": 171
},
{
"epoch": 0.26522744795682346,
"grad_norm": 0.024629781022667885,
"learning_rate": 0.0001930788057083476,
"loss": 0.774,
"step": 172
},
{
"epoch": 0.26676946800308404,
"grad_norm": 0.023107299581170082,
"learning_rate": 0.0001929885206816973,
"loss": 0.7021,
"step": 173
},
{
"epoch": 0.2683114880493446,
"grad_norm": 0.022794177755713463,
"learning_rate": 0.00019289767198167916,
"loss": 0.7469,
"step": 174
},
{
"epoch": 0.26985350809560527,
"grad_norm": 0.026097161695361137,
"learning_rate": 0.00019280626015899546,
"loss": 0.7325,
"step": 175
},
{
"epoch": 0.27139552814186585,
"grad_norm": 0.029879910871386528,
"learning_rate": 0.00019271428576776205,
"loss": 0.7614,
"step": 176
},
{
"epoch": 0.27293754818812643,
"grad_norm": 0.026986606419086456,
"learning_rate": 0.00019262174936550487,
"loss": 0.7718,
"step": 177
},
{
"epoch": 0.2744795682343871,
"grad_norm": 0.025835467502474785,
"learning_rate": 0.00019252865151315665,
"loss": 0.7511,
"step": 178
},
{
"epoch": 0.27602158828064766,
"grad_norm": 0.028101902455091476,
"learning_rate": 0.00019243499277505355,
"loss": 0.8136,
"step": 179
},
{
"epoch": 0.27756360832690824,
"grad_norm": 0.028153471648693085,
"learning_rate": 0.00019234077371893155,
"loss": 0.8798,
"step": 180
},
{
"epoch": 0.2791056283731688,
"grad_norm": 0.024540267884731293,
"learning_rate": 0.0001922459949159233,
"loss": 0.7854,
"step": 181
},
{
"epoch": 0.28064764841942946,
"grad_norm": 0.023485183715820312,
"learning_rate": 0.00019215065694055437,
"loss": 0.6655,
"step": 182
},
{
"epoch": 0.28218966846569005,
"grad_norm": 0.023394625633955002,
"learning_rate": 0.00019205476037073997,
"loss": 0.759,
"step": 183
},
{
"epoch": 0.28373168851195063,
"grad_norm": 0.025181008502840996,
"learning_rate": 0.00019195830578778132,
"loss": 0.7649,
"step": 184
},
{
"epoch": 0.28527370855821127,
"grad_norm": 0.023696815595030785,
"learning_rate": 0.0001918612937763622,
"loss": 0.7469,
"step": 185
},
{
"epoch": 0.28681572860447185,
"grad_norm": 0.025794658809900284,
"learning_rate": 0.00019176372492454537,
"loss": 0.7623,
"step": 186
},
{
"epoch": 0.28835774865073244,
"grad_norm": 0.02523699589073658,
"learning_rate": 0.00019166559982376904,
"loss": 0.6621,
"step": 187
},
{
"epoch": 0.2898997686969931,
"grad_norm": 0.02426300384104252,
"learning_rate": 0.00019156691906884325,
"loss": 0.748,
"step": 188
},
{
"epoch": 0.29144178874325366,
"grad_norm": 0.024227775633335114,
"learning_rate": 0.0001914676832579463,
"loss": 0.724,
"step": 189
},
{
"epoch": 0.29298380878951424,
"grad_norm": 0.031684551388025284,
"learning_rate": 0.00019136789299262108,
"loss": 0.8939,
"step": 190
},
{
"epoch": 0.2945258288357749,
"grad_norm": 0.023766616359353065,
"learning_rate": 0.0001912675488777714,
"loss": 0.7179,
"step": 191
},
{
"epoch": 0.29606784888203547,
"grad_norm": 0.02463400922715664,
"learning_rate": 0.0001911666515216585,
"loss": 0.8202,
"step": 192
},
{
"epoch": 0.29760986892829605,
"grad_norm": 0.0225905179977417,
"learning_rate": 0.00019106520153589708,
"loss": 0.6357,
"step": 193
},
{
"epoch": 0.2991518889745567,
"grad_norm": 0.022882292047142982,
"learning_rate": 0.00019096319953545185,
"loss": 0.7506,
"step": 194
},
{
"epoch": 0.3006939090208173,
"grad_norm": 0.027596216648817062,
"learning_rate": 0.00019086064613863364,
"loss": 0.7669,
"step": 195
},
{
"epoch": 0.30223592906707786,
"grad_norm": 0.030367175117135048,
"learning_rate": 0.00019075754196709572,
"loss": 0.7792,
"step": 196
},
{
"epoch": 0.3037779491133385,
"grad_norm": 0.025017013773322105,
"learning_rate": 0.00019065388764583004,
"loss": 0.6977,
"step": 197
},
{
"epoch": 0.3053199691595991,
"grad_norm": 0.02788584679365158,
"learning_rate": 0.0001905496838031634,
"loss": 0.6871,
"step": 198
},
{
"epoch": 0.30686198920585966,
"grad_norm": 0.032745130360126495,
"learning_rate": 0.00019044493107075368,
"loss": 0.8934,
"step": 199
},
{
"epoch": 0.3084040092521203,
"grad_norm": 0.027039945125579834,
"learning_rate": 0.00019033963008358598,
"loss": 0.6522,
"step": 200
},
{
"epoch": 0.3099460292983809,
"grad_norm": 0.03149978816509247,
"learning_rate": 0.0001902337814799688,
"loss": 0.844,
"step": 201
},
{
"epoch": 0.31148804934464147,
"grad_norm": 0.024369308724999428,
"learning_rate": 0.0001901273859015301,
"loss": 0.7202,
"step": 202
},
{
"epoch": 0.31303006939090205,
"grad_norm": 0.05448361113667488,
"learning_rate": 0.00019002044399321356,
"loss": 0.8301,
"step": 203
},
{
"epoch": 0.3145720894371627,
"grad_norm": 0.02388385497033596,
"learning_rate": 0.0001899129564032745,
"loss": 0.7105,
"step": 204
},
{
"epoch": 0.3161141094834233,
"grad_norm": 0.02488291636109352,
"learning_rate": 0.00018980492378327607,
"loss": 0.7393,
"step": 205
},
{
"epoch": 0.31765612952968386,
"grad_norm": 0.023874662816524506,
"learning_rate": 0.00018969634678808522,
"loss": 0.6791,
"step": 206
},
{
"epoch": 0.3191981495759445,
"grad_norm": 0.030418075621128082,
"learning_rate": 0.0001895872260758688,
"loss": 0.75,
"step": 207
},
{
"epoch": 0.3207401696222051,
"grad_norm": 0.02990088053047657,
"learning_rate": 0.00018947756230808954,
"loss": 0.7986,
"step": 208
},
{
"epoch": 0.32228218966846567,
"grad_norm": 0.027980022132396698,
"learning_rate": 0.00018936735614950197,
"loss": 0.7054,
"step": 209
},
{
"epoch": 0.3238242097147263,
"grad_norm": 0.026269223541021347,
"learning_rate": 0.00018925660826814856,
"loss": 0.8195,
"step": 210
},
{
"epoch": 0.3253662297609869,
"grad_norm": 0.025045178830623627,
"learning_rate": 0.0001891453193353555,
"loss": 0.7221,
"step": 211
},
{
"epoch": 0.3269082498072475,
"grad_norm": 0.021175356581807137,
"learning_rate": 0.00018903349002572873,
"loss": 0.6513,
"step": 212
},
{
"epoch": 0.3284502698535081,
"grad_norm": 0.023593388497829437,
"learning_rate": 0.0001889211210171498,
"loss": 0.7405,
"step": 213
},
{
"epoch": 0.3299922898997687,
"grad_norm": 0.027270464226603508,
"learning_rate": 0.00018880821299077183,
"loss": 0.7184,
"step": 214
},
{
"epoch": 0.3315343099460293,
"grad_norm": 0.025203121826052666,
"learning_rate": 0.00018869476663101523,
"loss": 0.6659,
"step": 215
},
{
"epoch": 0.3330763299922899,
"grad_norm": 0.024924185127019882,
"learning_rate": 0.0001885807826255638,
"loss": 0.7412,
"step": 216
},
{
"epoch": 0.3346183500385505,
"grad_norm": 0.022862501442432404,
"learning_rate": 0.00018846626166536026,
"loss": 0.6984,
"step": 217
},
{
"epoch": 0.3361603700848111,
"grad_norm": 0.022781461477279663,
"learning_rate": 0.0001883512044446023,
"loss": 0.6374,
"step": 218
},
{
"epoch": 0.33770239013107173,
"grad_norm": 0.023618346080183983,
"learning_rate": 0.0001882356116607383,
"loss": 0.6948,
"step": 219
},
{
"epoch": 0.3392444101773323,
"grad_norm": 0.02586747333407402,
"learning_rate": 0.0001881194840144631,
"loss": 0.7682,
"step": 220
},
{
"epoch": 0.3407864302235929,
"grad_norm": 0.026834698393940926,
"learning_rate": 0.00018800282220971366,
"loss": 0.7546,
"step": 221
},
{
"epoch": 0.34232845026985353,
"grad_norm": 0.028564658015966415,
"learning_rate": 0.00018788562695366495,
"loss": 0.8267,
"step": 222
},
{
"epoch": 0.3438704703161141,
"grad_norm": 0.02490355260670185,
"learning_rate": 0.00018776789895672558,
"loss": 0.6937,
"step": 223
},
{
"epoch": 0.3454124903623747,
"grad_norm": 0.03545152395963669,
"learning_rate": 0.00018764963893253347,
"loss": 0.7001,
"step": 224
},
{
"epoch": 0.3469545104086353,
"grad_norm": 0.02449451945722103,
"learning_rate": 0.00018753084759795158,
"loss": 0.8165,
"step": 225
},
{
"epoch": 0.3484965304548959,
"grad_norm": 0.027851196005940437,
"learning_rate": 0.00018741152567306355,
"loss": 0.7196,
"step": 226
},
{
"epoch": 0.3500385505011565,
"grad_norm": 0.02707446552813053,
"learning_rate": 0.00018729167388116934,
"loss": 0.8375,
"step": 227
},
{
"epoch": 0.3515805705474171,
"grad_norm": 0.02902469038963318,
"learning_rate": 0.00018717129294878074,
"loss": 0.6744,
"step": 228
},
{
"epoch": 0.35312259059367773,
"grad_norm": 0.031537748873233795,
"learning_rate": 0.0001870503836056172,
"loss": 0.7552,
"step": 229
},
{
"epoch": 0.3546646106399383,
"grad_norm": 0.026265786960721016,
"learning_rate": 0.00018692894658460117,
"loss": 0.7551,
"step": 230
},
{
"epoch": 0.3562066306861989,
"grad_norm": 0.02483406662940979,
"learning_rate": 0.0001868069826218538,
"loss": 0.7233,
"step": 231
},
{
"epoch": 0.35774865073245954,
"grad_norm": 0.024800019338726997,
"learning_rate": 0.0001866844924566904,
"loss": 0.7371,
"step": 232
},
{
"epoch": 0.3592906707787201,
"grad_norm": 0.02515244670212269,
"learning_rate": 0.00018656147683161593,
"loss": 0.7621,
"step": 233
},
{
"epoch": 0.3608326908249807,
"grad_norm": 0.02592633105814457,
"learning_rate": 0.00018643793649232072,
"loss": 0.7539,
"step": 234
},
{
"epoch": 0.36237471087124135,
"grad_norm": 0.0275077186524868,
"learning_rate": 0.00018631387218767561,
"loss": 0.6925,
"step": 235
},
{
"epoch": 0.36391673091750193,
"grad_norm": 0.027163324877619743,
"learning_rate": 0.00018618928466972775,
"loss": 0.7867,
"step": 236
},
{
"epoch": 0.3654587509637625,
"grad_norm": 0.026956308633089066,
"learning_rate": 0.0001860641746936957,
"loss": 0.7813,
"step": 237
},
{
"epoch": 0.36700077101002315,
"grad_norm": 0.02884814888238907,
"learning_rate": 0.0001859385430179652,
"loss": 0.7366,
"step": 238
},
{
"epoch": 0.36854279105628374,
"grad_norm": 0.025071945041418076,
"learning_rate": 0.00018581239040408432,
"loss": 0.708,
"step": 239
},
{
"epoch": 0.3700848111025443,
"grad_norm": 0.032973822206258774,
"learning_rate": 0.00018568571761675893,
"loss": 0.6544,
"step": 240
},
{
"epoch": 0.37162683114880496,
"grad_norm": 0.02571587637066841,
"learning_rate": 0.0001855585254238481,
"loss": 0.7633,
"step": 241
},
{
"epoch": 0.37316885119506554,
"grad_norm": 0.027229083701968193,
"learning_rate": 0.00018543081459635935,
"loss": 0.7752,
"step": 242
},
{
"epoch": 0.3747108712413261,
"grad_norm": 0.022508805617690086,
"learning_rate": 0.00018530258590844409,
"loss": 0.6437,
"step": 243
},
{
"epoch": 0.37625289128758677,
"grad_norm": 0.026772433891892433,
"learning_rate": 0.00018517384013739285,
"loss": 0.805,
"step": 244
},
{
"epoch": 0.37779491133384735,
"grad_norm": 0.023964572697877884,
"learning_rate": 0.00018504457806363056,
"loss": 0.7378,
"step": 245
},
{
"epoch": 0.37933693138010793,
"grad_norm": 0.02789299376308918,
"learning_rate": 0.0001849148004707119,
"loss": 0.772,
"step": 246
},
{
"epoch": 0.3808789514263685,
"grad_norm": 0.031168216839432716,
"learning_rate": 0.00018478450814531647,
"loss": 0.8299,
"step": 247
},
{
"epoch": 0.38242097147262916,
"grad_norm": 0.03058604896068573,
"learning_rate": 0.00018465370187724408,
"loss": 0.694,
"step": 248
},
{
"epoch": 0.38396299151888974,
"grad_norm": 0.028347650542855263,
"learning_rate": 0.0001845223824594099,
"loss": 0.7373,
"step": 249
},
{
"epoch": 0.3855050115651503,
"grad_norm": 0.025979626923799515,
"learning_rate": 0.00018439055068783966,
"loss": 0.8036,
"step": 250
},
{
"epoch": 0.38704703161141096,
"grad_norm": 0.029867777600884438,
"learning_rate": 0.0001842582073616649,
"loss": 0.7655,
"step": 251
},
{
"epoch": 0.38858905165767155,
"grad_norm": 0.025117915123701096,
"learning_rate": 0.00018412535328311814,
"loss": 0.7532,
"step": 252
},
{
"epoch": 0.39013107170393213,
"grad_norm": 0.023947982117533684,
"learning_rate": 0.00018399198925752778,
"loss": 0.6967,
"step": 253
},
{
"epoch": 0.39167309175019277,
"grad_norm": 0.025846531614661217,
"learning_rate": 0.00018385811609331352,
"loss": 0.7382,
"step": 254
},
{
"epoch": 0.39321511179645335,
"grad_norm": 0.025034697726368904,
"learning_rate": 0.00018372373460198138,
"loss": 0.7282,
"step": 255
},
{
"epoch": 0.39475713184271394,
"grad_norm": 0.02547437883913517,
"learning_rate": 0.00018358884559811856,
"loss": 0.7447,
"step": 256
},
{
"epoch": 0.3962991518889746,
"grad_norm": 0.027032596990466118,
"learning_rate": 0.0001834534498993888,
"loss": 0.7395,
"step": 257
},
{
"epoch": 0.39784117193523516,
"grad_norm": 0.027110572904348373,
"learning_rate": 0.0001833175483265273,
"loss": 0.7963,
"step": 258
},
{
"epoch": 0.39938319198149574,
"grad_norm": 0.027663685381412506,
"learning_rate": 0.00018318114170333568,
"loss": 0.7893,
"step": 259
},
{
"epoch": 0.4009252120277564,
"grad_norm": 0.027797933667898178,
"learning_rate": 0.00018304423085667714,
"loss": 0.7228,
"step": 260
},
{
"epoch": 0.40246723207401697,
"grad_norm": 0.026281701400876045,
"learning_rate": 0.0001829068166164712,
"loss": 0.749,
"step": 261
},
{
"epoch": 0.40400925212027755,
"grad_norm": 0.024708108976483345,
"learning_rate": 0.00018276889981568906,
"loss": 0.6307,
"step": 262
},
{
"epoch": 0.4055512721665382,
"grad_norm": 0.028213316574692726,
"learning_rate": 0.0001826304812903481,
"loss": 0.8186,
"step": 263
},
{
"epoch": 0.4070932922127988,
"grad_norm": 0.024718405678868294,
"learning_rate": 0.00018249156187950715,
"loss": 0.7077,
"step": 264
},
{
"epoch": 0.40863531225905936,
"grad_norm": 0.02398741990327835,
"learning_rate": 0.00018235214242526125,
"loss": 0.7041,
"step": 265
},
{
"epoch": 0.41017733230531994,
"grad_norm": 0.03176787495613098,
"learning_rate": 0.00018221222377273657,
"loss": 0.7036,
"step": 266
},
{
"epoch": 0.4117193523515806,
"grad_norm": 0.028862686827778816,
"learning_rate": 0.0001820718067700853,
"loss": 0.7947,
"step": 267
},
{
"epoch": 0.41326137239784116,
"grad_norm": 0.026759544387459755,
"learning_rate": 0.0001819308922684805,
"loss": 0.7737,
"step": 268
},
{
"epoch": 0.41480339244410175,
"grad_norm": 0.02719755284488201,
"learning_rate": 0.00018178948112211103,
"loss": 0.7403,
"step": 269
},
{
"epoch": 0.4163454124903624,
"grad_norm": 0.024756524711847305,
"learning_rate": 0.0001816475741881761,
"loss": 0.6994,
"step": 270
},
{
"epoch": 0.41788743253662297,
"grad_norm": 0.03232420235872269,
"learning_rate": 0.00018150517232688049,
"loss": 0.7866,
"step": 271
},
{
"epoch": 0.41942945258288356,
"grad_norm": 0.027607185766100883,
"learning_rate": 0.00018136227640142894,
"loss": 0.7905,
"step": 272
},
{
"epoch": 0.4209714726291442,
"grad_norm": 0.024344706907868385,
"learning_rate": 0.00018121888727802113,
"loss": 0.7408,
"step": 273
},
{
"epoch": 0.4225134926754048,
"grad_norm": 0.025088010355830193,
"learning_rate": 0.0001810750058258464,
"loss": 0.737,
"step": 274
},
{
"epoch": 0.42405551272166536,
"grad_norm": 0.023952683433890343,
"learning_rate": 0.00018093063291707847,
"loss": 0.7764,
"step": 275
},
{
"epoch": 0.425597532767926,
"grad_norm": 0.0288414116948843,
"learning_rate": 0.00018078576942687008,
"loss": 0.7035,
"step": 276
},
{
"epoch": 0.4271395528141866,
"grad_norm": 0.02681080810725689,
"learning_rate": 0.0001806404162333479,
"loss": 0.739,
"step": 277
},
{
"epoch": 0.42868157286044717,
"grad_norm": 0.0266602523624897,
"learning_rate": 0.0001804945742176069,
"loss": 0.7213,
"step": 278
},
{
"epoch": 0.4302235929067078,
"grad_norm": 0.025282425805926323,
"learning_rate": 0.00018034824426370523,
"loss": 0.6807,
"step": 279
},
{
"epoch": 0.4317656129529684,
"grad_norm": 0.025683747604489326,
"learning_rate": 0.00018020142725865888,
"loss": 0.7283,
"step": 280
},
{
"epoch": 0.433307632999229,
"grad_norm": 0.024966144934296608,
"learning_rate": 0.00018005412409243606,
"loss": 0.7096,
"step": 281
},
{
"epoch": 0.4348496530454896,
"grad_norm": 0.027953188866376877,
"learning_rate": 0.00017990633565795208,
"loss": 0.8148,
"step": 282
},
{
"epoch": 0.4363916730917502,
"grad_norm": 0.02772989496588707,
"learning_rate": 0.00017975806285106387,
"loss": 0.8568,
"step": 283
},
{
"epoch": 0.4379336931380108,
"grad_norm": 0.028020409867167473,
"learning_rate": 0.00017960930657056438,
"loss": 0.6732,
"step": 284
},
{
"epoch": 0.4394757131842714,
"grad_norm": 0.025754399597644806,
"learning_rate": 0.00017946006771817733,
"loss": 0.7238,
"step": 285
},
{
"epoch": 0.441017733230532,
"grad_norm": 0.030171813443303108,
"learning_rate": 0.00017931034719855166,
"loss": 0.7493,
"step": 286
},
{
"epoch": 0.4425597532767926,
"grad_norm": 0.026995845139026642,
"learning_rate": 0.00017916014591925605,
"loss": 0.6118,
"step": 287
},
{
"epoch": 0.4441017733230532,
"grad_norm": 0.03541433438658714,
"learning_rate": 0.00017900946479077346,
"loss": 0.7243,
"step": 288
},
{
"epoch": 0.4456437933693138,
"grad_norm": 0.029751230031251907,
"learning_rate": 0.00017885830472649553,
"loss": 0.7081,
"step": 289
},
{
"epoch": 0.4471858134155744,
"grad_norm": 0.022569075226783752,
"learning_rate": 0.00017870666664271707,
"loss": 0.6488,
"step": 290
},
{
"epoch": 0.448727833461835,
"grad_norm": 0.03262341767549515,
"learning_rate": 0.00017855455145863062,
"loss": 0.7626,
"step": 291
},
{
"epoch": 0.4502698535080956,
"grad_norm": 0.02811555750668049,
"learning_rate": 0.0001784019600963207,
"loss": 0.7485,
"step": 292
},
{
"epoch": 0.4518118735543562,
"grad_norm": 0.02504836954176426,
"learning_rate": 0.00017824889348075837,
"loss": 0.7636,
"step": 293
},
{
"epoch": 0.4533538936006168,
"grad_norm": 0.02362634427845478,
"learning_rate": 0.00017809535253979547,
"loss": 0.6915,
"step": 294
},
{
"epoch": 0.4548959136468774,
"grad_norm": 0.029891418293118477,
"learning_rate": 0.00017794133820415916,
"loss": 0.7686,
"step": 295
},
{
"epoch": 0.456437933693138,
"grad_norm": 0.02471439354121685,
"learning_rate": 0.0001777868514074462,
"loss": 0.6693,
"step": 296
},
{
"epoch": 0.4579799537393986,
"grad_norm": 0.025612330064177513,
"learning_rate": 0.00017763189308611722,
"loss": 0.7364,
"step": 297
},
{
"epoch": 0.45952197378565923,
"grad_norm": 0.026865236461162567,
"learning_rate": 0.00017747646417949113,
"loss": 0.7445,
"step": 298
},
{
"epoch": 0.4610639938319198,
"grad_norm": 0.024343574419617653,
"learning_rate": 0.00017732056562973954,
"loss": 0.6271,
"step": 299
},
{
"epoch": 0.4626060138781804,
"grad_norm": 0.02818606235086918,
"learning_rate": 0.00017716419838188077,
"loss": 0.7753,
"step": 300
},
{
"epoch": 0.46414803392444104,
"grad_norm": 0.026821713894605637,
"learning_rate": 0.00017700736338377435,
"loss": 0.6976,
"step": 301
},
{
"epoch": 0.4656900539707016,
"grad_norm": 0.025784511119127274,
"learning_rate": 0.00017685006158611516,
"loss": 0.7062,
"step": 302
},
{
"epoch": 0.4672320740169622,
"grad_norm": 0.028515879064798355,
"learning_rate": 0.00017669229394242766,
"loss": 0.7909,
"step": 303
},
{
"epoch": 0.46877409406322285,
"grad_norm": 0.024095451459288597,
"learning_rate": 0.0001765340614090603,
"loss": 0.7535,
"step": 304
},
{
"epoch": 0.47031611410948343,
"grad_norm": 0.025953758507966995,
"learning_rate": 0.0001763753649451794,
"loss": 0.7623,
"step": 305
},
{
"epoch": 0.471858134155744,
"grad_norm": 0.03294069692492485,
"learning_rate": 0.00017621620551276366,
"loss": 0.8946,
"step": 306
},
{
"epoch": 0.47340015420200465,
"grad_norm": 0.028394997119903564,
"learning_rate": 0.00017605658407659808,
"loss": 0.7251,
"step": 307
},
{
"epoch": 0.47494217424826524,
"grad_norm": 0.025346368551254272,
"learning_rate": 0.00017589650160426828,
"loss": 0.7074,
"step": 308
},
{
"epoch": 0.4764841942945258,
"grad_norm": 0.025906400755047798,
"learning_rate": 0.0001757359590661545,
"loss": 0.6472,
"step": 309
},
{
"epoch": 0.4780262143407864,
"grad_norm": 0.02889554388821125,
"learning_rate": 0.00017557495743542585,
"loss": 0.715,
"step": 310
},
{
"epoch": 0.47956823438704704,
"grad_norm": 0.029205597937107086,
"learning_rate": 0.00017541349768803428,
"loss": 0.698,
"step": 311
},
{
"epoch": 0.4811102544333076,
"grad_norm": 0.02610400691628456,
"learning_rate": 0.0001752515808027088,
"loss": 0.7073,
"step": 312
},
{
"epoch": 0.4826522744795682,
"grad_norm": 0.023945793509483337,
"learning_rate": 0.00017508920776094944,
"loss": 0.6865,
"step": 313
},
{
"epoch": 0.48419429452582885,
"grad_norm": 0.028426503762602806,
"learning_rate": 0.0001749263795470213,
"loss": 0.7264,
"step": 314
},
{
"epoch": 0.48573631457208943,
"grad_norm": 0.02865850180387497,
"learning_rate": 0.0001747630971479487,
"loss": 0.7204,
"step": 315
},
{
"epoch": 0.48727833461835,
"grad_norm": 0.027321334928274155,
"learning_rate": 0.00017459936155350908,
"loss": 0.7491,
"step": 316
},
{
"epoch": 0.48882035466461066,
"grad_norm": 0.02754514105618,
"learning_rate": 0.00017443517375622704,
"loss": 0.7567,
"step": 317
},
{
"epoch": 0.49036237471087124,
"grad_norm": 0.028822382912039757,
"learning_rate": 0.00017427053475136826,
"loss": 0.7559,
"step": 318
},
{
"epoch": 0.4919043947571318,
"grad_norm": 0.03181014209985733,
"learning_rate": 0.00017410544553693365,
"loss": 0.7704,
"step": 319
},
{
"epoch": 0.49344641480339246,
"grad_norm": 0.023862695321440697,
"learning_rate": 0.00017393990711365312,
"loss": 0.6085,
"step": 320
},
{
"epoch": 0.49498843484965305,
"grad_norm": 0.02703220769762993,
"learning_rate": 0.00017377392048497953,
"loss": 0.6979,
"step": 321
},
{
"epoch": 0.49653045489591363,
"grad_norm": 0.025343257933855057,
"learning_rate": 0.00017360748665708268,
"loss": 0.7287,
"step": 322
},
{
"epoch": 0.49807247494217427,
"grad_norm": 0.02830134704709053,
"learning_rate": 0.00017344060663884324,
"loss": 0.8054,
"step": 323
},
{
"epoch": 0.49961449498843485,
"grad_norm": 0.025809939950704575,
"learning_rate": 0.00017327328144184646,
"loss": 0.704,
"step": 324
},
{
"epoch": 0.5011565150346955,
"grad_norm": 0.027546260505914688,
"learning_rate": 0.00017310551208037626,
"loss": 0.7099,
"step": 325
},
{
"epoch": 0.5026985350809561,
"grad_norm": 0.027951935306191444,
"learning_rate": 0.00017293729957140893,
"loss": 0.878,
"step": 326
},
{
"epoch": 0.5042405551272167,
"grad_norm": 0.026868853718042374,
"learning_rate": 0.000172768644934607,
"loss": 0.7714,
"step": 327
},
{
"epoch": 0.5057825751734772,
"grad_norm": 0.026975559070706367,
"learning_rate": 0.0001725995491923131,
"loss": 0.7526,
"step": 328
},
{
"epoch": 0.5073245952197378,
"grad_norm": 0.027235837653279305,
"learning_rate": 0.0001724300133695437,
"loss": 0.6515,
"step": 329
},
{
"epoch": 0.5088666152659984,
"grad_norm": 0.032752856612205505,
"learning_rate": 0.00017226003849398294,
"loss": 0.8019,
"step": 330
},
{
"epoch": 0.5104086353122591,
"grad_norm": 0.029344195500016212,
"learning_rate": 0.0001720896255959764,
"loss": 0.9296,
"step": 331
},
{
"epoch": 0.5119506553585197,
"grad_norm": 0.025766605511307716,
"learning_rate": 0.00017191877570852483,
"loss": 0.7944,
"step": 332
},
{
"epoch": 0.5134926754047803,
"grad_norm": 0.023067327216267586,
"learning_rate": 0.0001717474898672779,
"loss": 0.6528,
"step": 333
},
{
"epoch": 0.5150346954510409,
"grad_norm": 0.026817042380571365,
"learning_rate": 0.00017157576911052796,
"loss": 0.763,
"step": 334
},
{
"epoch": 0.5165767154973014,
"grad_norm": 0.02811489813029766,
"learning_rate": 0.00017140361447920364,
"loss": 0.7936,
"step": 335
},
{
"epoch": 0.518118735543562,
"grad_norm": 0.02597888559103012,
"learning_rate": 0.00017123102701686372,
"loss": 0.7217,
"step": 336
},
{
"epoch": 0.5196607555898227,
"grad_norm": 0.030244017019867897,
"learning_rate": 0.00017105800776969055,
"loss": 0.756,
"step": 337
},
{
"epoch": 0.5212027756360833,
"grad_norm": 0.028289398178458214,
"learning_rate": 0.00017088455778648397,
"loss": 0.7972,
"step": 338
},
{
"epoch": 0.5227447956823439,
"grad_norm": 0.02397543005645275,
"learning_rate": 0.00017071067811865476,
"loss": 0.6358,
"step": 339
},
{
"epoch": 0.5242868157286045,
"grad_norm": 0.02565479464828968,
"learning_rate": 0.00017053636982021844,
"loss": 0.6912,
"step": 340
},
{
"epoch": 0.525828835774865,
"grad_norm": 0.027768775820732117,
"learning_rate": 0.00017036163394778864,
"loss": 0.7997,
"step": 341
},
{
"epoch": 0.5273708558211256,
"grad_norm": 0.033537182956933975,
"learning_rate": 0.00017018647156057096,
"loss": 0.7363,
"step": 342
},
{
"epoch": 0.5289128758673862,
"grad_norm": 0.030843475833535194,
"learning_rate": 0.00017001088372035637,
"loss": 0.6977,
"step": 343
},
{
"epoch": 0.5304548959136469,
"grad_norm": 0.026586662977933884,
"learning_rate": 0.00016983487149151486,
"loss": 0.7874,
"step": 344
},
{
"epoch": 0.5319969159599075,
"grad_norm": 0.027003532275557518,
"learning_rate": 0.00016965843594098892,
"loss": 0.7895,
"step": 345
},
{
"epoch": 0.5335389360061681,
"grad_norm": 0.02772395871579647,
"learning_rate": 0.00016948157813828716,
"loss": 0.7799,
"step": 346
},
{
"epoch": 0.5350809560524287,
"grad_norm": 0.025633979588747025,
"learning_rate": 0.0001693042991554777,
"loss": 0.744,
"step": 347
},
{
"epoch": 0.5366229760986893,
"grad_norm": 0.033604227006435394,
"learning_rate": 0.00016912660006718186,
"loss": 0.6707,
"step": 348
},
{
"epoch": 0.5381649961449498,
"grad_norm": 0.03711126372218132,
"learning_rate": 0.00016894848195056746,
"loss": 0.8468,
"step": 349
},
{
"epoch": 0.5397070161912105,
"grad_norm": 0.02382393553853035,
"learning_rate": 0.00016876994588534234,
"loss": 0.6884,
"step": 350
},
{
"epoch": 0.5412490362374711,
"grad_norm": 0.02597069926559925,
"learning_rate": 0.0001685909929537479,
"loss": 0.7789,
"step": 351
},
{
"epoch": 0.5427910562837317,
"grad_norm": 0.027003685012459755,
"learning_rate": 0.0001684116242405525,
"loss": 0.7364,
"step": 352
},
{
"epoch": 0.5443330763299923,
"grad_norm": 0.028209254145622253,
"learning_rate": 0.00016823184083304482,
"loss": 0.8014,
"step": 353
},
{
"epoch": 0.5458750963762529,
"grad_norm": 0.025543801486492157,
"learning_rate": 0.0001680516438210273,
"loss": 0.7241,
"step": 354
},
{
"epoch": 0.5474171164225135,
"grad_norm": 0.024730654433369637,
"learning_rate": 0.00016787103429680955,
"loss": 0.6425,
"step": 355
},
{
"epoch": 0.5489591364687741,
"grad_norm": 0.02615622617304325,
"learning_rate": 0.0001676900133552018,
"loss": 0.7091,
"step": 356
},
{
"epoch": 0.5505011565150347,
"grad_norm": 0.025628188624978065,
"learning_rate": 0.00016750858209350808,
"loss": 0.848,
"step": 357
},
{
"epoch": 0.5520431765612953,
"grad_norm": 0.02642144076526165,
"learning_rate": 0.0001673267416115198,
"loss": 0.7026,
"step": 358
},
{
"epoch": 0.5535851966075559,
"grad_norm": 0.02542021870613098,
"learning_rate": 0.00016714449301150883,
"loss": 0.6338,
"step": 359
},
{
"epoch": 0.5551272166538165,
"grad_norm": 0.028573600575327873,
"learning_rate": 0.00016696183739822108,
"loss": 0.7553,
"step": 360
},
{
"epoch": 0.5566692367000771,
"grad_norm": 0.024142924696207047,
"learning_rate": 0.00016677877587886956,
"loss": 0.7327,
"step": 361
},
{
"epoch": 0.5582112567463376,
"grad_norm": 0.029369287192821503,
"learning_rate": 0.00016659530956312788,
"loss": 0.7728,
"step": 362
},
{
"epoch": 0.5597532767925983,
"grad_norm": 0.02711080014705658,
"learning_rate": 0.00016641143956312336,
"loss": 0.6843,
"step": 363
},
{
"epoch": 0.5612952968388589,
"grad_norm": 0.027241146191954613,
"learning_rate": 0.00016622716699343033,
"loss": 0.8534,
"step": 364
},
{
"epoch": 0.5628373168851195,
"grad_norm": 0.030265534296631813,
"learning_rate": 0.0001660424929710635,
"loss": 0.8579,
"step": 365
},
{
"epoch": 0.5643793369313801,
"grad_norm": 0.029054157435894012,
"learning_rate": 0.000165857418615471,
"loss": 0.75,
"step": 366
},
{
"epoch": 0.5659213569776407,
"grad_norm": 0.029963452368974686,
"learning_rate": 0.0001656719450485278,
"loss": 0.7464,
"step": 367
},
{
"epoch": 0.5674633770239013,
"grad_norm": 0.025763841345906258,
"learning_rate": 0.00016548607339452853,
"loss": 0.6632,
"step": 368
},
{
"epoch": 0.569005397070162,
"grad_norm": 0.02374422177672386,
"learning_rate": 0.00016529980478018115,
"loss": 0.6546,
"step": 369
},
{
"epoch": 0.5705474171164225,
"grad_norm": 0.02538699097931385,
"learning_rate": 0.00016511314033459994,
"loss": 0.6603,
"step": 370
},
{
"epoch": 0.5720894371626831,
"grad_norm": 0.028310047462582588,
"learning_rate": 0.0001649260811892984,
"loss": 0.6088,
"step": 371
},
{
"epoch": 0.5736314572089437,
"grad_norm": 0.02981553040444851,
"learning_rate": 0.00016473862847818277,
"loss": 0.8053,
"step": 372
},
{
"epoch": 0.5751734772552043,
"grad_norm": 0.025665050372481346,
"learning_rate": 0.0001645507833375449,
"loss": 0.6779,
"step": 373
},
{
"epoch": 0.5767154973014649,
"grad_norm": 0.030752114951610565,
"learning_rate": 0.0001643625469060555,
"loss": 0.6893,
"step": 374
},
{
"epoch": 0.5782575173477256,
"grad_norm": 0.02725459821522236,
"learning_rate": 0.00016417392032475715,
"loss": 0.7879,
"step": 375
},
{
"epoch": 0.5797995373939862,
"grad_norm": 0.02713746391236782,
"learning_rate": 0.00016398490473705743,
"loss": 0.7578,
"step": 376
},
{
"epoch": 0.5813415574402467,
"grad_norm": 0.02641828916966915,
"learning_rate": 0.000163795501288722,
"loss": 0.7335,
"step": 377
},
{
"epoch": 0.5828835774865073,
"grad_norm": 0.027243638411164284,
"learning_rate": 0.00016360571112786765,
"loss": 0.6873,
"step": 378
},
{
"epoch": 0.5844255975327679,
"grad_norm": 0.028045805171132088,
"learning_rate": 0.00016341553540495532,
"loss": 0.7141,
"step": 379
},
{
"epoch": 0.5859676175790285,
"grad_norm": 0.029962563887238503,
"learning_rate": 0.00016322497527278306,
"loss": 0.882,
"step": 380
},
{
"epoch": 0.5875096376252892,
"grad_norm": 0.027544857934117317,
"learning_rate": 0.00016303403188647913,
"loss": 0.6695,
"step": 381
},
{
"epoch": 0.5890516576715498,
"grad_norm": 0.02867518924176693,
"learning_rate": 0.00016284270640349514,
"loss": 0.8052,
"step": 382
},
{
"epoch": 0.5905936777178104,
"grad_norm": 0.02330535091459751,
"learning_rate": 0.00016265099998359866,
"loss": 0.6825,
"step": 383
},
{
"epoch": 0.5921356977640709,
"grad_norm": 0.030629336833953857,
"learning_rate": 0.00016245891378886655,
"loss": 0.8545,
"step": 384
},
{
"epoch": 0.5936777178103315,
"grad_norm": 0.030393701046705246,
"learning_rate": 0.00016226644898367768,
"loss": 0.7342,
"step": 385
},
{
"epoch": 0.5952197378565921,
"grad_norm": 0.02500557340681553,
"learning_rate": 0.000162073606734706,
"loss": 0.6975,
"step": 386
},
{
"epoch": 0.5967617579028527,
"grad_norm": 0.031064705923199654,
"learning_rate": 0.00016188038821091344,
"loss": 0.7903,
"step": 387
},
{
"epoch": 0.5983037779491134,
"grad_norm": 0.02647087723016739,
"learning_rate": 0.00016168679458354284,
"loss": 0.745,
"step": 388
},
{
"epoch": 0.599845797995374,
"grad_norm": 0.02917948178946972,
"learning_rate": 0.00016149282702611077,
"loss": 0.7819,
"step": 389
},
{
"epoch": 0.6013878180416345,
"grad_norm": 0.02458810992538929,
"learning_rate": 0.00016129848671440046,
"loss": 0.6466,
"step": 390
},
{
"epoch": 0.6029298380878951,
"grad_norm": 0.023677226155996323,
"learning_rate": 0.00016110377482645477,
"loss": 0.7161,
"step": 391
},
{
"epoch": 0.6044718581341557,
"grad_norm": 0.02807523123919964,
"learning_rate": 0.00016090869254256892,
"loss": 0.7164,
"step": 392
},
{
"epoch": 0.6060138781804163,
"grad_norm": 0.031006982550024986,
"learning_rate": 0.0001607132410452833,
"loss": 0.7781,
"step": 393
},
{
"epoch": 0.607555898226677,
"grad_norm": 0.029372677206993103,
"learning_rate": 0.00016051742151937655,
"loss": 0.7034,
"step": 394
},
{
"epoch": 0.6090979182729376,
"grad_norm": 0.025212230160832405,
"learning_rate": 0.00016032123515185797,
"loss": 0.6567,
"step": 395
},
{
"epoch": 0.6106399383191982,
"grad_norm": 0.024775920435786247,
"learning_rate": 0.00016012468313196084,
"loss": 0.6319,
"step": 396
},
{
"epoch": 0.6121819583654587,
"grad_norm": 0.02687055990099907,
"learning_rate": 0.0001599277666511347,
"loss": 0.7214,
"step": 397
},
{
"epoch": 0.6137239784117193,
"grad_norm": 0.028431419283151627,
"learning_rate": 0.0001597304869030385,
"loss": 0.7815,
"step": 398
},
{
"epoch": 0.6152659984579799,
"grad_norm": 0.027922354638576508,
"learning_rate": 0.00015953284508353317,
"loss": 0.7016,
"step": 399
},
{
"epoch": 0.6168080185042406,
"grad_norm": 0.02737678587436676,
"learning_rate": 0.00015933484239067446,
"loss": 0.7507,
"step": 400
},
{
"epoch": 0.6183500385505012,
"grad_norm": 0.02791019156575203,
"learning_rate": 0.00015913648002470563,
"loss": 0.705,
"step": 401
},
{
"epoch": 0.6198920585967618,
"grad_norm": 0.027861539274454117,
"learning_rate": 0.0001589377591880501,
"loss": 0.7208,
"step": 402
},
{
"epoch": 0.6214340786430224,
"grad_norm": 0.02952715940773487,
"learning_rate": 0.00015873868108530443,
"loss": 0.7114,
"step": 403
},
{
"epoch": 0.6229760986892829,
"grad_norm": 0.02896735444664955,
"learning_rate": 0.0001585392469232307,
"loss": 0.7797,
"step": 404
},
{
"epoch": 0.6245181187355435,
"grad_norm": 0.02640017308294773,
"learning_rate": 0.00015833945791074943,
"loss": 0.7186,
"step": 405
},
{
"epoch": 0.6260601387818041,
"grad_norm": 0.025937926024198532,
"learning_rate": 0.000158139315258932,
"loss": 0.781,
"step": 406
},
{
"epoch": 0.6276021588280648,
"grad_norm": 0.02851933054625988,
"learning_rate": 0.00015793882018099364,
"loss": 0.7522,
"step": 407
},
{
"epoch": 0.6291441788743254,
"grad_norm": 0.02368611842393875,
"learning_rate": 0.00015773797389228582,
"loss": 0.719,
"step": 408
},
{
"epoch": 0.630686198920586,
"grad_norm": 0.026969095692038536,
"learning_rate": 0.00015753677761028896,
"loss": 0.7554,
"step": 409
},
{
"epoch": 0.6322282189668466,
"grad_norm": 0.02418413758277893,
"learning_rate": 0.00015733523255460506,
"loss": 0.6919,
"step": 410
},
{
"epoch": 0.6337702390131071,
"grad_norm": 0.026073114946484566,
"learning_rate": 0.0001571333399469503,
"loss": 0.7059,
"step": 411
},
{
"epoch": 0.6353122590593677,
"grad_norm": 0.027075573801994324,
"learning_rate": 0.0001569311010111476,
"loss": 0.7239,
"step": 412
},
{
"epoch": 0.6368542791056284,
"grad_norm": 0.02724389173090458,
"learning_rate": 0.00015672851697311934,
"loss": 0.7615,
"step": 413
},
{
"epoch": 0.638396299151889,
"grad_norm": 0.026826992630958557,
"learning_rate": 0.00015652558906087971,
"loss": 0.8138,
"step": 414
},
{
"epoch": 0.6399383191981496,
"grad_norm": 0.02641242742538452,
"learning_rate": 0.00015632231850452747,
"loss": 0.6956,
"step": 415
},
{
"epoch": 0.6414803392444102,
"grad_norm": 0.02495909109711647,
"learning_rate": 0.00015611870653623825,
"loss": 0.7546,
"step": 416
},
{
"epoch": 0.6430223592906708,
"grad_norm": 0.035753343254327774,
"learning_rate": 0.00015591475439025745,
"loss": 0.7257,
"step": 417
},
{
"epoch": 0.6445643793369313,
"grad_norm": 0.03195042535662651,
"learning_rate": 0.00015571046330289237,
"loss": 0.693,
"step": 418
},
{
"epoch": 0.646106399383192,
"grad_norm": 0.027049189433455467,
"learning_rate": 0.00015550583451250503,
"loss": 0.7602,
"step": 419
},
{
"epoch": 0.6476484194294526,
"grad_norm": 0.027553152292966843,
"learning_rate": 0.00015530086925950434,
"loss": 0.7723,
"step": 420
},
{
"epoch": 0.6491904394757132,
"grad_norm": 0.025219090282917023,
"learning_rate": 0.00015509556878633894,
"loss": 0.7011,
"step": 421
},
{
"epoch": 0.6507324595219738,
"grad_norm": 0.029802288860082626,
"learning_rate": 0.00015488993433748944,
"loss": 0.6898,
"step": 422
},
{
"epoch": 0.6522744795682344,
"grad_norm": 0.02772880345582962,
"learning_rate": 0.00015468396715946083,
"loss": 0.7399,
"step": 423
},
{
"epoch": 0.653816499614495,
"grad_norm": 0.025485830381512642,
"learning_rate": 0.00015447766850077517,
"loss": 0.6622,
"step": 424
},
{
"epoch": 0.6553585196607556,
"grad_norm": 0.027489742264151573,
"learning_rate": 0.00015427103961196376,
"loss": 0.8027,
"step": 425
},
{
"epoch": 0.6569005397070162,
"grad_norm": 0.026778720319271088,
"learning_rate": 0.00015406408174555976,
"loss": 0.7804,
"step": 426
},
{
"epoch": 0.6584425597532768,
"grad_norm": 0.027850337326526642,
"learning_rate": 0.00015385679615609042,
"loss": 0.7645,
"step": 427
},
{
"epoch": 0.6599845797995374,
"grad_norm": 0.02659332938492298,
"learning_rate": 0.00015364918410006967,
"loss": 0.7609,
"step": 428
},
{
"epoch": 0.661526599845798,
"grad_norm": 0.0292272437363863,
"learning_rate": 0.0001534412468359903,
"loss": 0.7788,
"step": 429
},
{
"epoch": 0.6630686198920586,
"grad_norm": 0.02392621338367462,
"learning_rate": 0.00015323298562431648,
"loss": 0.675,
"step": 430
},
{
"epoch": 0.6646106399383191,
"grad_norm": 0.030261410400271416,
"learning_rate": 0.00015302440172747605,
"loss": 0.6585,
"step": 431
},
{
"epoch": 0.6661526599845798,
"grad_norm": 0.031531739979982376,
"learning_rate": 0.00015281549640985294,
"loss": 0.8002,
"step": 432
},
{
"epoch": 0.6676946800308404,
"grad_norm": 0.025663699954748154,
"learning_rate": 0.00015260627093777936,
"loss": 0.7068,
"step": 433
},
{
"epoch": 0.669236700077101,
"grad_norm": 0.02657734416425228,
"learning_rate": 0.00015239672657952832,
"loss": 0.7369,
"step": 434
},
{
"epoch": 0.6707787201233616,
"grad_norm": 0.027592379599809647,
"learning_rate": 0.0001521868646053058,
"loss": 0.7245,
"step": 435
},
{
"epoch": 0.6723207401696222,
"grad_norm": 0.02645149454474449,
"learning_rate": 0.00015197668628724303,
"loss": 0.718,
"step": 436
},
{
"epoch": 0.6738627602158828,
"grad_norm": 0.02535802498459816,
"learning_rate": 0.00015176619289938888,
"loss": 0.6926,
"step": 437
},
{
"epoch": 0.6754047802621435,
"grad_norm": 0.02912677638232708,
"learning_rate": 0.00015155538571770218,
"loss": 0.8564,
"step": 438
},
{
"epoch": 0.676946800308404,
"grad_norm": 0.023175543174147606,
"learning_rate": 0.00015134426602004375,
"loss": 0.6568,
"step": 439
},
{
"epoch": 0.6784888203546646,
"grad_norm": 0.02832154743373394,
"learning_rate": 0.00015113283508616895,
"loss": 0.6571,
"step": 440
},
{
"epoch": 0.6800308404009252,
"grad_norm": 0.029950594529509544,
"learning_rate": 0.0001509210941977196,
"loss": 0.8564,
"step": 441
},
{
"epoch": 0.6815728604471858,
"grad_norm": 0.02448093518614769,
"learning_rate": 0.00015070904463821658,
"loss": 0.6936,
"step": 442
},
{
"epoch": 0.6831148804934464,
"grad_norm": 0.027585268020629883,
"learning_rate": 0.00015049668769305172,
"loss": 0.7991,
"step": 443
},
{
"epoch": 0.6846569005397071,
"grad_norm": 0.025705596432089806,
"learning_rate": 0.00015028402464948022,
"loss": 0.7187,
"step": 444
},
{
"epoch": 0.6861989205859677,
"grad_norm": 0.02810623310506344,
"learning_rate": 0.00015007105679661276,
"loss": 0.7315,
"step": 445
},
{
"epoch": 0.6877409406322282,
"grad_norm": 0.02478802390396595,
"learning_rate": 0.0001498577854254076,
"loss": 0.7201,
"step": 446
},
{
"epoch": 0.6892829606784888,
"grad_norm": 0.03222353756427765,
"learning_rate": 0.0001496442118286631,
"loss": 0.796,
"step": 447
},
{
"epoch": 0.6908249807247494,
"grad_norm": 0.025557026267051697,
"learning_rate": 0.00014943033730100935,
"loss": 0.6066,
"step": 448
},
{
"epoch": 0.69236700077101,
"grad_norm": 0.02581370249390602,
"learning_rate": 0.00014921616313890072,
"loss": 0.7186,
"step": 449
},
{
"epoch": 0.6939090208172706,
"grad_norm": 0.025746649131178856,
"learning_rate": 0.00014900169064060802,
"loss": 0.703,
"step": 450
},
{
"epoch": 0.6954510408635313,
"grad_norm": 0.03065100871026516,
"learning_rate": 0.00014878692110621028,
"loss": 0.7724,
"step": 451
},
{
"epoch": 0.6969930609097919,
"grad_norm": 0.03629877045750618,
"learning_rate": 0.00014857185583758723,
"loss": 0.7703,
"step": 452
},
{
"epoch": 0.6985350809560524,
"grad_norm": 0.025341391563415527,
"learning_rate": 0.0001483564961384112,
"loss": 0.6492,
"step": 453
},
{
"epoch": 0.700077101002313,
"grad_norm": 0.03427153080701828,
"learning_rate": 0.0001481408433141394,
"loss": 0.7148,
"step": 454
},
{
"epoch": 0.7016191210485736,
"grad_norm": 0.023825203999876976,
"learning_rate": 0.0001479248986720057,
"loss": 0.6197,
"step": 455
},
{
"epoch": 0.7031611410948342,
"grad_norm": 0.026944074779748917,
"learning_rate": 0.00014770866352101307,
"loss": 0.7273,
"step": 456
},
{
"epoch": 0.7047031611410949,
"grad_norm": 0.028203219175338745,
"learning_rate": 0.00014749213917192538,
"loss": 0.7741,
"step": 457
},
{
"epoch": 0.7062451811873555,
"grad_norm": 0.024622568860650063,
"learning_rate": 0.00014727532693725962,
"loss": 0.6212,
"step": 458
},
{
"epoch": 0.707787201233616,
"grad_norm": 0.023979736492037773,
"learning_rate": 0.00014705822813127777,
"loss": 0.6863,
"step": 459
},
{
"epoch": 0.7093292212798766,
"grad_norm": 0.027042685076594353,
"learning_rate": 0.00014684084406997903,
"loss": 0.7746,
"step": 460
},
{
"epoch": 0.7108712413261372,
"grad_norm": 0.026160864159464836,
"learning_rate": 0.00014662317607109168,
"loss": 0.727,
"step": 461
},
{
"epoch": 0.7124132613723978,
"grad_norm": 0.035195063799619675,
"learning_rate": 0.0001464052254540652,
"loss": 0.9004,
"step": 462
},
{
"epoch": 0.7139552814186585,
"grad_norm": 0.02714708261191845,
"learning_rate": 0.00014618699354006223,
"loss": 0.6853,
"step": 463
},
{
"epoch": 0.7154973014649191,
"grad_norm": 0.027736373245716095,
"learning_rate": 0.0001459684816519505,
"loss": 0.7095,
"step": 464
},
{
"epoch": 0.7170393215111797,
"grad_norm": 0.032326798886060715,
"learning_rate": 0.000145749691114295,
"loss": 0.7256,
"step": 465
},
{
"epoch": 0.7185813415574402,
"grad_norm": 0.028024908155202866,
"learning_rate": 0.00014553062325334967,
"loss": 0.7213,
"step": 466
},
{
"epoch": 0.7201233616037008,
"grad_norm": 0.026767205446958542,
"learning_rate": 0.00014531127939704965,
"loss": 0.6619,
"step": 467
},
{
"epoch": 0.7216653816499614,
"grad_norm": 0.025469880551099777,
"learning_rate": 0.00014509166087500302,
"loss": 0.6174,
"step": 468
},
{
"epoch": 0.723207401696222,
"grad_norm": 0.03378410264849663,
"learning_rate": 0.00014487176901848285,
"loss": 0.6447,
"step": 469
},
{
"epoch": 0.7247494217424827,
"grad_norm": 0.026924695819616318,
"learning_rate": 0.00014465160516041904,
"loss": 0.7116,
"step": 470
},
{
"epoch": 0.7262914417887433,
"grad_norm": 0.029700160026550293,
"learning_rate": 0.00014443117063539038,
"loss": 0.811,
"step": 471
},
{
"epoch": 0.7278334618350039,
"grad_norm": 0.030913611873984337,
"learning_rate": 0.00014421046677961626,
"loss": 0.77,
"step": 472
},
{
"epoch": 0.7293754818812644,
"grad_norm": 0.029507668688893318,
"learning_rate": 0.0001439894949309489,
"loss": 0.7813,
"step": 473
},
{
"epoch": 0.730917501927525,
"grad_norm": 0.028620922937989235,
"learning_rate": 0.00014376825642886472,
"loss": 0.7401,
"step": 474
},
{
"epoch": 0.7324595219737856,
"grad_norm": 0.025677144527435303,
"learning_rate": 0.0001435467526144568,
"loss": 0.7408,
"step": 475
},
{
"epoch": 0.7340015420200463,
"grad_norm": 0.025129586458206177,
"learning_rate": 0.00014332498483042637,
"loss": 0.6585,
"step": 476
},
{
"epoch": 0.7355435620663069,
"grad_norm": 0.026812126860022545,
"learning_rate": 0.0001431029544210747,
"loss": 0.8063,
"step": 477
},
{
"epoch": 0.7370855821125675,
"grad_norm": 0.02705306001007557,
"learning_rate": 0.0001428806627322952,
"loss": 0.7655,
"step": 478
},
{
"epoch": 0.7386276021588281,
"grad_norm": 0.026533039286732674,
"learning_rate": 0.0001426581111115649,
"loss": 0.6685,
"step": 479
},
{
"epoch": 0.7401696222050886,
"grad_norm": 0.030538393184542656,
"learning_rate": 0.00014243530090793667,
"loss": 0.7389,
"step": 480
},
{
"epoch": 0.7417116422513492,
"grad_norm": 0.027102958410978317,
"learning_rate": 0.00014221223347203068,
"loss": 0.726,
"step": 481
},
{
"epoch": 0.7432536622976099,
"grad_norm": 0.024465948343276978,
"learning_rate": 0.00014198891015602646,
"loss": 0.6379,
"step": 482
},
{
"epoch": 0.7447956823438705,
"grad_norm": 0.027429422363638878,
"learning_rate": 0.00014176533231365464,
"loss": 0.726,
"step": 483
},
{
"epoch": 0.7463377023901311,
"grad_norm": 0.02289111353456974,
"learning_rate": 0.00014154150130018866,
"loss": 0.6354,
"step": 484
},
{
"epoch": 0.7478797224363917,
"grad_norm": 0.027374420315027237,
"learning_rate": 0.00014131741847243665,
"loss": 0.8206,
"step": 485
},
{
"epoch": 0.7494217424826523,
"grad_norm": 0.029361480847001076,
"learning_rate": 0.0001410930851887332,
"loss": 0.6994,
"step": 486
},
{
"epoch": 0.7509637625289128,
"grad_norm": 0.027077002450823784,
"learning_rate": 0.00014086850280893107,
"loss": 0.7133,
"step": 487
},
{
"epoch": 0.7525057825751735,
"grad_norm": 0.028823737055063248,
"learning_rate": 0.000140643672694393,
"loss": 0.7975,
"step": 488
},
{
"epoch": 0.7540478026214341,
"grad_norm": 0.02851509116590023,
"learning_rate": 0.0001404185962079834,
"loss": 0.7497,
"step": 489
},
{
"epoch": 0.7555898226676947,
"grad_norm": 0.025098523125052452,
"learning_rate": 0.00014019327471406022,
"loss": 0.6978,
"step": 490
},
{
"epoch": 0.7571318427139553,
"grad_norm": 0.03084149770438671,
"learning_rate": 0.00013996770957846644,
"loss": 0.7607,
"step": 491
},
{
"epoch": 0.7586738627602159,
"grad_norm": 0.02702442556619644,
"learning_rate": 0.00013974190216852202,
"loss": 0.7639,
"step": 492
},
{
"epoch": 0.7602158828064765,
"grad_norm": 0.033685747534036636,
"learning_rate": 0.00013951585385301555,
"loss": 0.7793,
"step": 493
},
{
"epoch": 0.761757902852737,
"grad_norm": 0.028003569692373276,
"learning_rate": 0.00013928956600219592,
"loss": 0.755,
"step": 494
},
{
"epoch": 0.7632999228989977,
"grad_norm": 0.026269137859344482,
"learning_rate": 0.00013906303998776392,
"loss": 0.7896,
"step": 495
},
{
"epoch": 0.7648419429452583,
"grad_norm": 0.027992503717541695,
"learning_rate": 0.0001388362771828642,
"loss": 0.6472,
"step": 496
},
{
"epoch": 0.7663839629915189,
"grad_norm": 0.028174640610814095,
"learning_rate": 0.00013860927896207665,
"loss": 0.765,
"step": 497
},
{
"epoch": 0.7679259830377795,
"grad_norm": 0.04180079326033592,
"learning_rate": 0.0001383820467014082,
"loss": 0.7479,
"step": 498
},
{
"epoch": 0.7694680030840401,
"grad_norm": 0.028985602781176567,
"learning_rate": 0.00013815458177828454,
"loss": 0.7121,
"step": 499
},
{
"epoch": 0.7710100231303006,
"grad_norm": 0.026408828794956207,
"learning_rate": 0.00013792688557154166,
"loss": 0.7411,
"step": 500
},
{
"epoch": 0.7725520431765613,
"grad_norm": 0.027537405490875244,
"learning_rate": 0.00013769895946141752,
"loss": 0.7628,
"step": 501
},
{
"epoch": 0.7740940632228219,
"grad_norm": 0.028669610619544983,
"learning_rate": 0.00013747080482954377,
"loss": 0.7698,
"step": 502
},
{
"epoch": 0.7756360832690825,
"grad_norm": 0.027917267754673958,
"learning_rate": 0.00013724242305893715,
"loss": 0.8452,
"step": 503
},
{
"epoch": 0.7771781033153431,
"grad_norm": 0.031637243926525116,
"learning_rate": 0.00013701381553399145,
"loss": 0.7457,
"step": 504
},
{
"epoch": 0.7787201233616037,
"grad_norm": 0.027903905138373375,
"learning_rate": 0.00013678498364046876,
"loss": 0.7475,
"step": 505
},
{
"epoch": 0.7802621434078643,
"grad_norm": 0.03381239250302315,
"learning_rate": 0.00013655592876549134,
"loss": 0.836,
"step": 506
},
{
"epoch": 0.781804163454125,
"grad_norm": 0.0790882334113121,
"learning_rate": 0.0001363266522975331,
"loss": 0.5842,
"step": 507
},
{
"epoch": 0.7833461835003855,
"grad_norm": 0.03034863993525505,
"learning_rate": 0.00013609715562641115,
"loss": 0.8256,
"step": 508
},
{
"epoch": 0.7848882035466461,
"grad_norm": 0.029499804601073265,
"learning_rate": 0.0001358674401432774,
"loss": 0.7145,
"step": 509
},
{
"epoch": 0.7864302235929067,
"grad_norm": 0.0285016018897295,
"learning_rate": 0.00013563750724061025,
"loss": 0.6712,
"step": 510
},
{
"epoch": 0.7879722436391673,
"grad_norm": 0.02717653289437294,
"learning_rate": 0.0001354073583122059,
"loss": 0.6807,
"step": 511
},
{
"epoch": 0.7895142636854279,
"grad_norm": 0.023634063079953194,
"learning_rate": 0.00013517699475317017,
"loss": 0.6677,
"step": 512
},
{
"epoch": 0.7910562837316885,
"grad_norm": 0.02920868992805481,
"learning_rate": 0.00013494641795990986,
"loss": 0.7865,
"step": 513
},
{
"epoch": 0.7925983037779492,
"grad_norm": 0.040133338421583176,
"learning_rate": 0.00013471562933012432,
"loss": 0.7786,
"step": 514
},
{
"epoch": 0.7941403238242097,
"grad_norm": 0.028838159516453743,
"learning_rate": 0.00013448463026279704,
"loss": 0.7689,
"step": 515
},
{
"epoch": 0.7956823438704703,
"grad_norm": 0.02688550017774105,
"learning_rate": 0.00013425342215818718,
"loss": 0.7659,
"step": 516
},
{
"epoch": 0.7972243639167309,
"grad_norm": 0.027155300602316856,
"learning_rate": 0.0001340220064178209,
"loss": 0.7353,
"step": 517
},
{
"epoch": 0.7987663839629915,
"grad_norm": 0.03566644340753555,
"learning_rate": 0.00013379038444448306,
"loss": 0.827,
"step": 518
},
{
"epoch": 0.8003084040092521,
"grad_norm": 0.0266974326223135,
"learning_rate": 0.00013355855764220868,
"loss": 0.7693,
"step": 519
},
{
"epoch": 0.8018504240555128,
"grad_norm": 0.026248447597026825,
"learning_rate": 0.00013332652741627446,
"loss": 0.7659,
"step": 520
},
{
"epoch": 0.8033924441017734,
"grad_norm": 0.030487187206745148,
"learning_rate": 0.00013309429517318998,
"loss": 0.7244,
"step": 521
},
{
"epoch": 0.8049344641480339,
"grad_norm": 0.031918346881866455,
"learning_rate": 0.00013286186232068972,
"loss": 0.6111,
"step": 522
},
{
"epoch": 0.8064764841942945,
"grad_norm": 0.03194589167833328,
"learning_rate": 0.00013262923026772388,
"loss": 0.7521,
"step": 523
},
{
"epoch": 0.8080185042405551,
"grad_norm": 0.029390091076493263,
"learning_rate": 0.00013239640042445036,
"loss": 0.7689,
"step": 524
},
{
"epoch": 0.8095605242868157,
"grad_norm": 0.025920424610376358,
"learning_rate": 0.000132163374202226,
"loss": 0.7102,
"step": 525
},
{
"epoch": 0.8111025443330764,
"grad_norm": 0.02453307807445526,
"learning_rate": 0.000131930153013598,
"loss": 0.712,
"step": 526
},
{
"epoch": 0.812644564379337,
"grad_norm": 0.035026635974645615,
"learning_rate": 0.0001316967382722954,
"loss": 0.789,
"step": 527
},
{
"epoch": 0.8141865844255975,
"grad_norm": 0.02895597368478775,
"learning_rate": 0.0001314631313932205,
"loss": 0.8087,
"step": 528
},
{
"epoch": 0.8157286044718581,
"grad_norm": 0.02956031821668148,
"learning_rate": 0.00013122933379244034,
"loss": 0.7514,
"step": 529
},
{
"epoch": 0.8172706245181187,
"grad_norm": 0.027792761102318764,
"learning_rate": 0.00013099534688717804,
"loss": 0.7843,
"step": 530
},
{
"epoch": 0.8188126445643793,
"grad_norm": 0.023245403543114662,
"learning_rate": 0.00013076117209580418,
"loss": 0.6878,
"step": 531
},
{
"epoch": 0.8203546646106399,
"grad_norm": 0.030602406710386276,
"learning_rate": 0.00013052681083782836,
"loss": 0.6504,
"step": 532
},
{
"epoch": 0.8218966846569006,
"grad_norm": 0.0250953808426857,
"learning_rate": 0.00013029226453389043,
"loss": 0.6255,
"step": 533
},
{
"epoch": 0.8234387047031612,
"grad_norm": 0.02791382372379303,
"learning_rate": 0.00013005753460575194,
"loss": 0.7252,
"step": 534
},
{
"epoch": 0.8249807247494217,
"grad_norm": 0.025774458423256874,
"learning_rate": 0.0001298226224762876,
"loss": 0.7252,
"step": 535
},
{
"epoch": 0.8265227447956823,
"grad_norm": 0.026806412264704704,
"learning_rate": 0.00012958752956947645,
"loss": 0.7702,
"step": 536
},
{
"epoch": 0.8280647648419429,
"grad_norm": 0.02904200740158558,
"learning_rate": 0.00012935225731039348,
"loss": 0.7573,
"step": 537
},
{
"epoch": 0.8296067848882035,
"grad_norm": 0.03018496371805668,
"learning_rate": 0.00012911680712520082,
"loss": 0.6854,
"step": 538
},
{
"epoch": 0.8311488049344642,
"grad_norm": 0.028921302407979965,
"learning_rate": 0.00012888118044113912,
"loss": 0.7372,
"step": 539
},
{
"epoch": 0.8326908249807248,
"grad_norm": 0.02912386879324913,
"learning_rate": 0.00012864537868651892,
"loss": 0.864,
"step": 540
},
{
"epoch": 0.8342328450269854,
"grad_norm": 0.03150784596800804,
"learning_rate": 0.00012840940329071211,
"loss": 0.6846,
"step": 541
},
{
"epoch": 0.8357748650732459,
"grad_norm": 0.026627201586961746,
"learning_rate": 0.00012817325568414297,
"loss": 0.6596,
"step": 542
},
{
"epoch": 0.8373168851195065,
"grad_norm": 0.026762191206216812,
"learning_rate": 0.00012793693729827983,
"loss": 0.7859,
"step": 543
},
{
"epoch": 0.8388589051657671,
"grad_norm": 0.0270906463265419,
"learning_rate": 0.00012770044956562611,
"loss": 0.76,
"step": 544
},
{
"epoch": 0.8404009252120278,
"grad_norm": 0.029391184449195862,
"learning_rate": 0.0001274637939197119,
"loss": 0.7153,
"step": 545
},
{
"epoch": 0.8419429452582884,
"grad_norm": 0.028473293408751488,
"learning_rate": 0.00012722697179508508,
"loss": 0.7083,
"step": 546
},
{
"epoch": 0.843484965304549,
"grad_norm": 0.028644869104027748,
"learning_rate": 0.00012698998462730264,
"loss": 0.7531,
"step": 547
},
{
"epoch": 0.8450269853508096,
"grad_norm": 0.0290384441614151,
"learning_rate": 0.00012675283385292212,
"loss": 0.8323,
"step": 548
},
{
"epoch": 0.8465690053970701,
"grad_norm": 0.027428725734353065,
"learning_rate": 0.00012651552090949263,
"loss": 0.783,
"step": 549
},
{
"epoch": 0.8481110254433307,
"grad_norm": 0.0316435806453228,
"learning_rate": 0.00012627804723554651,
"loss": 0.7303,
"step": 550
},
{
"epoch": 0.8496530454895914,
"grad_norm": 0.031257808208465576,
"learning_rate": 0.00012604041427059036,
"loss": 0.6641,
"step": 551
},
{
"epoch": 0.851195065535852,
"grad_norm": 0.031232839450240135,
"learning_rate": 0.00012580262345509622,
"loss": 0.8266,
"step": 552
},
{
"epoch": 0.8527370855821126,
"grad_norm": 0.028075871989130974,
"learning_rate": 0.00012556467623049312,
"loss": 0.6852,
"step": 553
},
{
"epoch": 0.8542791056283732,
"grad_norm": 0.0269178319722414,
"learning_rate": 0.0001253265740391582,
"loss": 0.6774,
"step": 554
},
{
"epoch": 0.8558211256746338,
"grad_norm": 0.033746860921382904,
"learning_rate": 0.00012508831832440794,
"loss": 0.6704,
"step": 555
},
{
"epoch": 0.8573631457208943,
"grad_norm": 0.02507774904370308,
"learning_rate": 0.0001248499105304894,
"loss": 0.6364,
"step": 556
},
{
"epoch": 0.8589051657671549,
"grad_norm": 0.027239350602030754,
"learning_rate": 0.00012461135210257155,
"loss": 0.7503,
"step": 557
},
{
"epoch": 0.8604471858134156,
"grad_norm": 0.031841401010751724,
"learning_rate": 0.00012437264448673647,
"loss": 0.7201,
"step": 558
},
{
"epoch": 0.8619892058596762,
"grad_norm": 0.025972798466682434,
"learning_rate": 0.00012413378912997058,
"loss": 0.7265,
"step": 559
},
{
"epoch": 0.8635312259059368,
"grad_norm": 0.02815602719783783,
"learning_rate": 0.00012389478748015583,
"loss": 0.6922,
"step": 560
},
{
"epoch": 0.8650732459521974,
"grad_norm": 0.03288137540221214,
"learning_rate": 0.00012365564098606102,
"loss": 0.6516,
"step": 561
},
{
"epoch": 0.866615265998458,
"grad_norm": 0.02655138447880745,
"learning_rate": 0.00012341635109733293,
"loss": 0.6192,
"step": 562
},
{
"epoch": 0.8681572860447185,
"grad_norm": 0.029187412932515144,
"learning_rate": 0.00012317691926448754,
"loss": 0.8307,
"step": 563
},
{
"epoch": 0.8696993060909792,
"grad_norm": 0.029966147616505623,
"learning_rate": 0.00012293734693890132,
"loss": 0.7285,
"step": 564
},
{
"epoch": 0.8712413261372398,
"grad_norm": 0.02818044275045395,
"learning_rate": 0.0001226976355728023,
"loss": 0.7171,
"step": 565
},
{
"epoch": 0.8727833461835004,
"grad_norm": 0.02721909061074257,
"learning_rate": 0.00012245778661926137,
"loss": 0.6913,
"step": 566
},
{
"epoch": 0.874325366229761,
"grad_norm": 0.030929675325751305,
"learning_rate": 0.0001222178015321835,
"loss": 0.8006,
"step": 567
},
{
"epoch": 0.8758673862760216,
"grad_norm": 0.026720581576228142,
"learning_rate": 0.00012197768176629876,
"loss": 0.7348,
"step": 568
},
{
"epoch": 0.8774094063222821,
"grad_norm": 0.02792746014893055,
"learning_rate": 0.00012173742877715373,
"loss": 0.786,
"step": 569
},
{
"epoch": 0.8789514263685428,
"grad_norm": 0.025372346863150597,
"learning_rate": 0.00012149704402110243,
"loss": 0.6832,
"step": 570
},
{
"epoch": 0.8804934464148034,
"grad_norm": 0.025779495015740395,
"learning_rate": 0.00012125652895529766,
"loss": 0.7465,
"step": 571
},
{
"epoch": 0.882035466461064,
"grad_norm": 0.028759067878127098,
"learning_rate": 0.00012101588503768224,
"loss": 0.7755,
"step": 572
},
{
"epoch": 0.8835774865073246,
"grad_norm": 0.0319170206785202,
"learning_rate": 0.00012077511372697985,
"loss": 0.7758,
"step": 573
},
{
"epoch": 0.8851195065535852,
"grad_norm": 0.0287742018699646,
"learning_rate": 0.00012053421648268662,
"loss": 0.7556,
"step": 574
},
{
"epoch": 0.8866615265998458,
"grad_norm": 0.027272436767816544,
"learning_rate": 0.00012029319476506182,
"loss": 0.6341,
"step": 575
},
{
"epoch": 0.8882035466461063,
"grad_norm": 0.02786344476044178,
"learning_rate": 0.00012005205003511948,
"loss": 0.7248,
"step": 576
},
{
"epoch": 0.889745566692367,
"grad_norm": 0.023950345814228058,
"learning_rate": 0.00011981078375461915,
"loss": 0.6907,
"step": 577
},
{
"epoch": 0.8912875867386276,
"grad_norm": 0.0248698852956295,
"learning_rate": 0.00011956939738605721,
"loss": 0.6612,
"step": 578
},
{
"epoch": 0.8928296067848882,
"grad_norm": 0.02722037211060524,
"learning_rate": 0.00011932789239265802,
"loss": 0.6179,
"step": 579
},
{
"epoch": 0.8943716268311488,
"grad_norm": 0.02561650238931179,
"learning_rate": 0.00011908627023836503,
"loss": 0.7478,
"step": 580
},
{
"epoch": 0.8959136468774094,
"grad_norm": 0.027791699394583702,
"learning_rate": 0.00011884453238783185,
"loss": 0.7803,
"step": 581
},
{
"epoch": 0.89745566692367,
"grad_norm": 0.033503565937280655,
"learning_rate": 0.00011860268030641338,
"loss": 0.7978,
"step": 582
},
{
"epoch": 0.8989976869699307,
"grad_norm": 0.0275451447814703,
"learning_rate": 0.00011836071546015703,
"loss": 0.7458,
"step": 583
},
{
"epoch": 0.9005397070161912,
"grad_norm": 0.02638075314462185,
"learning_rate": 0.00011811863931579377,
"loss": 0.5823,
"step": 584
},
{
"epoch": 0.9020817270624518,
"grad_norm": 0.029536547139286995,
"learning_rate": 0.00011787645334072913,
"loss": 0.829,
"step": 585
},
{
"epoch": 0.9036237471087124,
"grad_norm": 0.029731806367635727,
"learning_rate": 0.0001176341590030345,
"loss": 0.7553,
"step": 586
},
{
"epoch": 0.905165767154973,
"grad_norm": 0.02816937118768692,
"learning_rate": 0.00011739175777143812,
"loss": 0.7094,
"step": 587
},
{
"epoch": 0.9067077872012336,
"grad_norm": 0.026959970593452454,
"learning_rate": 0.00011714925111531619,
"loss": 0.7368,
"step": 588
},
{
"epoch": 0.9082498072474943,
"grad_norm": 0.027997490018606186,
"learning_rate": 0.0001169066405046839,
"loss": 0.7864,
"step": 589
},
{
"epoch": 0.9097918272937549,
"grad_norm": 0.025101030245423317,
"learning_rate": 0.00011666392741018675,
"loss": 0.6445,
"step": 590
},
{
"epoch": 0.9113338473400154,
"grad_norm": 0.038789402693510056,
"learning_rate": 0.00011642111330309129,
"loss": 0.7104,
"step": 591
},
{
"epoch": 0.912875867386276,
"grad_norm": 0.0264846533536911,
"learning_rate": 0.0001161781996552765,
"loss": 0.7034,
"step": 592
},
{
"epoch": 0.9144178874325366,
"grad_norm": 0.03629022464156151,
"learning_rate": 0.00011593518793922468,
"loss": 0.8121,
"step": 593
},
{
"epoch": 0.9159599074787972,
"grad_norm": 0.025432435795664787,
"learning_rate": 0.00011569207962801263,
"loss": 0.6726,
"step": 594
},
{
"epoch": 0.9175019275250579,
"grad_norm": 0.02553374320268631,
"learning_rate": 0.00011544887619530275,
"loss": 0.659,
"step": 595
},
{
"epoch": 0.9190439475713185,
"grad_norm": 0.02498779632151127,
"learning_rate": 0.0001152055791153339,
"loss": 0.5868,
"step": 596
},
{
"epoch": 0.920585967617579,
"grad_norm": 0.029489582404494286,
"learning_rate": 0.00011496218986291273,
"loss": 0.673,
"step": 597
},
{
"epoch": 0.9221279876638396,
"grad_norm": 0.029334766790270805,
"learning_rate": 0.00011471870991340459,
"loss": 0.8011,
"step": 598
},
{
"epoch": 0.9236700077101002,
"grad_norm": 0.0274631530046463,
"learning_rate": 0.00011447514074272451,
"loss": 0.721,
"step": 599
},
{
"epoch": 0.9252120277563608,
"grad_norm": 0.028073586523532867,
"learning_rate": 0.00011423148382732853,
"loss": 0.839,
"step": 600
},
{
"epoch": 0.9267540478026214,
"grad_norm": 0.028757184743881226,
"learning_rate": 0.00011398774064420443,
"loss": 0.7409,
"step": 601
},
{
"epoch": 0.9282960678488821,
"grad_norm": 0.026308685541152954,
"learning_rate": 0.00011374391267086302,
"loss": 0.7517,
"step": 602
},
{
"epoch": 0.9298380878951427,
"grad_norm": 0.027558207511901855,
"learning_rate": 0.00011350000138532902,
"loss": 0.7041,
"step": 603
},
{
"epoch": 0.9313801079414032,
"grad_norm": 0.02995571680366993,
"learning_rate": 0.0001132560082661322,
"loss": 0.842,
"step": 604
},
{
"epoch": 0.9329221279876638,
"grad_norm": 0.028026850894093513,
"learning_rate": 0.00011301193479229841,
"loss": 0.7639,
"step": 605
},
{
"epoch": 0.9344641480339244,
"grad_norm": 0.026926511898636818,
"learning_rate": 0.00011276778244334055,
"loss": 0.6315,
"step": 606
},
{
"epoch": 0.936006168080185,
"grad_norm": 0.031490955501794815,
"learning_rate": 0.00011252355269924963,
"loss": 0.5844,
"step": 607
},
{
"epoch": 0.9375481881264457,
"grad_norm": 0.028631744906306267,
"learning_rate": 0.00011227924704048585,
"loss": 0.6431,
"step": 608
},
{
"epoch": 0.9390902081727063,
"grad_norm": 0.029599271714687347,
"learning_rate": 0.00011203486694796957,
"loss": 0.7085,
"step": 609
},
{
"epoch": 0.9406322282189669,
"grad_norm": 0.027870824560523033,
"learning_rate": 0.00011179041390307235,
"loss": 0.7381,
"step": 610
},
{
"epoch": 0.9421742482652274,
"grad_norm": 0.030139662325382233,
"learning_rate": 0.00011154588938760794,
"loss": 0.7299,
"step": 611
},
{
"epoch": 0.943716268311488,
"grad_norm": 0.031028704717755318,
"learning_rate": 0.00011130129488382341,
"loss": 0.747,
"step": 612
},
{
"epoch": 0.9452582883577486,
"grad_norm": 0.02865663915872574,
"learning_rate": 0.00011105663187438997,
"loss": 0.7025,
"step": 613
},
{
"epoch": 0.9468003084040093,
"grad_norm": 0.029388774186372757,
"learning_rate": 0.00011081190184239419,
"loss": 0.6645,
"step": 614
},
{
"epoch": 0.9483423284502699,
"grad_norm": 0.023502621799707413,
"learning_rate": 0.00011056710627132884,
"loss": 0.6455,
"step": 615
},
{
"epoch": 0.9498843484965305,
"grad_norm": 0.02663263864815235,
"learning_rate": 0.00011032224664508405,
"loss": 0.7677,
"step": 616
},
{
"epoch": 0.9514263685427911,
"grad_norm": 0.02990710362792015,
"learning_rate": 0.00011007732444793814,
"loss": 0.7017,
"step": 617
},
{
"epoch": 0.9529683885890516,
"grad_norm": 0.029992980882525444,
"learning_rate": 0.00010983234116454886,
"loss": 0.7969,
"step": 618
},
{
"epoch": 0.9545104086353122,
"grad_norm": 0.03714431822299957,
"learning_rate": 0.00010958729827994404,
"loss": 0.7646,
"step": 619
},
{
"epoch": 0.9560524286815728,
"grad_norm": 0.030639756470918655,
"learning_rate": 0.00010934219727951301,
"loss": 0.8926,
"step": 620
},
{
"epoch": 0.9575944487278335,
"grad_norm": 0.030243555083870888,
"learning_rate": 0.00010909703964899729,
"loss": 0.6676,
"step": 621
},
{
"epoch": 0.9591364687740941,
"grad_norm": 0.029893988743424416,
"learning_rate": 0.00010885182687448161,
"loss": 0.7238,
"step": 622
},
{
"epoch": 0.9606784888203547,
"grad_norm": 0.02677508443593979,
"learning_rate": 0.00010860656044238511,
"loss": 0.7378,
"step": 623
},
{
"epoch": 0.9622205088666153,
"grad_norm": 0.02502857707440853,
"learning_rate": 0.00010836124183945208,
"loss": 0.7129,
"step": 624
},
{
"epoch": 0.9637625289128758,
"grad_norm": 0.02673465758562088,
"learning_rate": 0.00010811587255274313,
"loss": 0.7046,
"step": 625
},
{
"epoch": 0.9653045489591364,
"grad_norm": 0.030728284269571304,
"learning_rate": 0.00010787045406962607,
"loss": 0.752,
"step": 626
},
{
"epoch": 0.9668465690053971,
"grad_norm": 0.02698132023215294,
"learning_rate": 0.00010762498787776687,
"loss": 0.6365,
"step": 627
},
{
"epoch": 0.9683885890516577,
"grad_norm": 0.03070535510778427,
"learning_rate": 0.0001073794754651208,
"loss": 0.7465,
"step": 628
},
{
"epoch": 0.9699306090979183,
"grad_norm": 0.026031676679849625,
"learning_rate": 0.00010713391831992323,
"loss": 0.6991,
"step": 629
},
{
"epoch": 0.9714726291441789,
"grad_norm": 0.02689627930521965,
"learning_rate": 0.00010688831793068078,
"loss": 0.6468,
"step": 630
},
{
"epoch": 0.9730146491904395,
"grad_norm": 0.028856465592980385,
"learning_rate": 0.00010664267578616207,
"loss": 0.7051,
"step": 631
},
{
"epoch": 0.9745566692367,
"grad_norm": 0.027791066095232964,
"learning_rate": 0.00010639699337538898,
"loss": 0.7247,
"step": 632
},
{
"epoch": 0.9760986892829607,
"grad_norm": 0.02929227240383625,
"learning_rate": 0.00010615127218762733,
"loss": 0.6994,
"step": 633
},
{
"epoch": 0.9776407093292213,
"grad_norm": 0.0322168804705143,
"learning_rate": 0.0001059055137123781,
"loss": 0.753,
"step": 634
},
{
"epoch": 0.9791827293754819,
"grad_norm": 0.02872268296778202,
"learning_rate": 0.00010565971943936825,
"loss": 0.7451,
"step": 635
},
{
"epoch": 0.9807247494217425,
"grad_norm": 0.03104063868522644,
"learning_rate": 0.00010541389085854176,
"loss": 0.7482,
"step": 636
},
{
"epoch": 0.9822667694680031,
"grad_norm": 0.026641814038157463,
"learning_rate": 0.00010516802946005058,
"loss": 0.7034,
"step": 637
},
{
"epoch": 0.9838087895142636,
"grad_norm": 0.027356311678886414,
"learning_rate": 0.00010492213673424553,
"loss": 0.6459,
"step": 638
},
{
"epoch": 0.9853508095605242,
"grad_norm": 0.031015669927001,
"learning_rate": 0.00010467621417166745,
"loss": 0.8185,
"step": 639
},
{
"epoch": 0.9868928296067849,
"grad_norm": 0.024331575259566307,
"learning_rate": 0.00010443026326303789,
"loss": 0.64,
"step": 640
},
{
"epoch": 0.9884348496530455,
"grad_norm": 0.026257265359163284,
"learning_rate": 0.00010418428549925032,
"loss": 0.6634,
"step": 641
},
{
"epoch": 0.9899768696993061,
"grad_norm": 0.026038171723484993,
"learning_rate": 0.00010393828237136107,
"loss": 0.728,
"step": 642
},
{
"epoch": 0.9915188897455667,
"grad_norm": 0.028959324583411217,
"learning_rate": 0.00010369225537058002,
"loss": 0.7315,
"step": 643
},
{
"epoch": 0.9930609097918273,
"grad_norm": 0.026242226362228394,
"learning_rate": 0.00010344620598826198,
"loss": 0.7096,
"step": 644
},
{
"epoch": 0.9946029298380878,
"grad_norm": 0.029856307432055473,
"learning_rate": 0.00010320013571589726,
"loss": 0.7052,
"step": 645
},
{
"epoch": 0.9961449498843485,
"grad_norm": 0.03145446628332138,
"learning_rate": 0.00010295404604510286,
"loss": 0.8392,
"step": 646
},
{
"epoch": 0.9976869699306091,
"grad_norm": 0.032184895128011703,
"learning_rate": 0.00010270793846761347,
"loss": 0.7936,
"step": 647
},
{
"epoch": 0.9992289899768697,
"grad_norm": 0.027684815227985382,
"learning_rate": 0.00010246181447527212,
"loss": 0.6861,
"step": 648
}
],
"logging_steps": 1,
"max_steps": 1296,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 324,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 8.697171009029014e+18,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}