abaddon182's picture
Training in progress, step 450, checkpoint
44e5499 verified
{
"best_metric": 0.8302884697914124,
"best_model_checkpoint": "miner_id_24/checkpoint-450",
"epoch": 0.08468197214904027,
"eval_steps": 150,
"global_step": 450,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0001881821603312006,
"grad_norm": 8.021297454833984,
"learning_rate": 5e-06,
"loss": 4.7627,
"step": 1
},
{
"epoch": 0.0001881821603312006,
"eval_loss": 1.206099271774292,
"eval_runtime": 977.2207,
"eval_samples_per_second": 9.159,
"eval_steps_per_second": 2.29,
"step": 1
},
{
"epoch": 0.0003763643206624012,
"grad_norm": 8.330506324768066,
"learning_rate": 1e-05,
"loss": 4.7186,
"step": 2
},
{
"epoch": 0.0005645464809936018,
"grad_norm": 7.244218826293945,
"learning_rate": 1.5e-05,
"loss": 4.7416,
"step": 3
},
{
"epoch": 0.0007527286413248024,
"grad_norm": 7.7417449951171875,
"learning_rate": 2e-05,
"loss": 4.8388,
"step": 4
},
{
"epoch": 0.000940910801656003,
"grad_norm": 7.4192376136779785,
"learning_rate": 2.5e-05,
"loss": 4.4709,
"step": 5
},
{
"epoch": 0.0011290929619872036,
"grad_norm": 4.583150386810303,
"learning_rate": 3e-05,
"loss": 4.305,
"step": 6
},
{
"epoch": 0.0013172751223184042,
"grad_norm": 3.533625602722168,
"learning_rate": 3.5e-05,
"loss": 4.0565,
"step": 7
},
{
"epoch": 0.0015054572826496049,
"grad_norm": 4.560721397399902,
"learning_rate": 4e-05,
"loss": 4.0204,
"step": 8
},
{
"epoch": 0.0016936394429808055,
"grad_norm": 13.375167846679688,
"learning_rate": 4.5e-05,
"loss": 4.3096,
"step": 9
},
{
"epoch": 0.001881821603312006,
"grad_norm": 6.8327813148498535,
"learning_rate": 5e-05,
"loss": 4.279,
"step": 10
},
{
"epoch": 0.0020700037636432068,
"grad_norm": 5.540746688842773,
"learning_rate": 5.500000000000001e-05,
"loss": 4.2908,
"step": 11
},
{
"epoch": 0.002258185923974407,
"grad_norm": 4.140532970428467,
"learning_rate": 6e-05,
"loss": 4.2689,
"step": 12
},
{
"epoch": 0.0024463680843056076,
"grad_norm": 1.9986830949783325,
"learning_rate": 6.500000000000001e-05,
"loss": 4.0747,
"step": 13
},
{
"epoch": 0.0026345502446368085,
"grad_norm": 1.936602234840393,
"learning_rate": 7e-05,
"loss": 3.956,
"step": 14
},
{
"epoch": 0.002822732404968009,
"grad_norm": 1.9803358316421509,
"learning_rate": 7.500000000000001e-05,
"loss": 4.0274,
"step": 15
},
{
"epoch": 0.0030109145652992097,
"grad_norm": 2.102954864501953,
"learning_rate": 8e-05,
"loss": 3.9315,
"step": 16
},
{
"epoch": 0.00319909672563041,
"grad_norm": 1.8604391813278198,
"learning_rate": 8.5e-05,
"loss": 3.9579,
"step": 17
},
{
"epoch": 0.003387278885961611,
"grad_norm": 5.816336631774902,
"learning_rate": 9e-05,
"loss": 3.8109,
"step": 18
},
{
"epoch": 0.0035754610462928114,
"grad_norm": 1.8846168518066406,
"learning_rate": 9.5e-05,
"loss": 3.8936,
"step": 19
},
{
"epoch": 0.003763643206624012,
"grad_norm": 2.1951682567596436,
"learning_rate": 0.0001,
"loss": 3.8866,
"step": 20
},
{
"epoch": 0.003951825366955213,
"grad_norm": 1.8232979774475098,
"learning_rate": 9.999866555428618e-05,
"loss": 3.7866,
"step": 21
},
{
"epoch": 0.0041400075272864136,
"grad_norm": 1.7551851272583008,
"learning_rate": 9.999466228837451e-05,
"loss": 3.848,
"step": 22
},
{
"epoch": 0.0043281896876176135,
"grad_norm": 3.396578788757324,
"learning_rate": 9.998799041595064e-05,
"loss": 3.8432,
"step": 23
},
{
"epoch": 0.004516371847948814,
"grad_norm": 1.9480079412460327,
"learning_rate": 9.997865029314463e-05,
"loss": 3.9188,
"step": 24
},
{
"epoch": 0.004704554008280015,
"grad_norm": 1.8144193887710571,
"learning_rate": 9.996664241851197e-05,
"loss": 3.9241,
"step": 25
},
{
"epoch": 0.004892736168611215,
"grad_norm": 1.6157795190811157,
"learning_rate": 9.995196743300692e-05,
"loss": 3.7515,
"step": 26
},
{
"epoch": 0.005080918328942416,
"grad_norm": 1.8001606464385986,
"learning_rate": 9.993462611994832e-05,
"loss": 3.8946,
"step": 27
},
{
"epoch": 0.005269100489273617,
"grad_norm": 2.1896607875823975,
"learning_rate": 9.991461940497786e-05,
"loss": 3.6935,
"step": 28
},
{
"epoch": 0.005457282649604818,
"grad_norm": 1.9953135251998901,
"learning_rate": 9.989194835601048e-05,
"loss": 3.6832,
"step": 29
},
{
"epoch": 0.005645464809936018,
"grad_norm": 1.9537560939788818,
"learning_rate": 9.986661418317759e-05,
"loss": 3.6689,
"step": 30
},
{
"epoch": 0.005833646970267219,
"grad_norm": 1.7720470428466797,
"learning_rate": 9.983861823876231e-05,
"loss": 3.8731,
"step": 31
},
{
"epoch": 0.0060218291305984195,
"grad_norm": 1.8037577867507935,
"learning_rate": 9.980796201712734e-05,
"loss": 3.7388,
"step": 32
},
{
"epoch": 0.0062100112909296195,
"grad_norm": 1.558807611465454,
"learning_rate": 9.977464715463524e-05,
"loss": 3.5732,
"step": 33
},
{
"epoch": 0.00639819345126082,
"grad_norm": 1.7031415700912476,
"learning_rate": 9.973867542956104e-05,
"loss": 3.7046,
"step": 34
},
{
"epoch": 0.006586375611592021,
"grad_norm": 1.8096400499343872,
"learning_rate": 9.97000487619973e-05,
"loss": 3.7951,
"step": 35
},
{
"epoch": 0.006774557771923222,
"grad_norm": 1.6755597591400146,
"learning_rate": 9.965876921375165e-05,
"loss": 3.7345,
"step": 36
},
{
"epoch": 0.006962739932254422,
"grad_norm": 1.6606807708740234,
"learning_rate": 9.961483898823678e-05,
"loss": 3.6923,
"step": 37
},
{
"epoch": 0.007150922092585623,
"grad_norm": 1.602203130722046,
"learning_rate": 9.956826043035268e-05,
"loss": 3.6913,
"step": 38
},
{
"epoch": 0.007339104252916824,
"grad_norm": 1.6571152210235596,
"learning_rate": 9.951903602636166e-05,
"loss": 3.6178,
"step": 39
},
{
"epoch": 0.007527286413248024,
"grad_norm": 1.737025499343872,
"learning_rate": 9.946716840375551e-05,
"loss": 3.6084,
"step": 40
},
{
"epoch": 0.0077154685735792245,
"grad_norm": 1.6381107568740845,
"learning_rate": 9.94126603311153e-05,
"loss": 3.4131,
"step": 41
},
{
"epoch": 0.007903650733910425,
"grad_norm": 1.668062686920166,
"learning_rate": 9.935551471796358e-05,
"loss": 3.4251,
"step": 42
},
{
"epoch": 0.008091832894241625,
"grad_norm": 1.6551685333251953,
"learning_rate": 9.92957346146091e-05,
"loss": 3.4862,
"step": 43
},
{
"epoch": 0.008280015054572827,
"grad_norm": 1.5389114618301392,
"learning_rate": 9.923332321198395e-05,
"loss": 3.3558,
"step": 44
},
{
"epoch": 0.008468197214904027,
"grad_norm": 1.6278464794158936,
"learning_rate": 9.916828384147331e-05,
"loss": 3.6091,
"step": 45
},
{
"epoch": 0.008656379375235227,
"grad_norm": 1.642901062965393,
"learning_rate": 9.910061997473752e-05,
"loss": 3.3503,
"step": 46
},
{
"epoch": 0.008844561535566429,
"grad_norm": 2.450819730758667,
"learning_rate": 9.903033522352687e-05,
"loss": 3.7451,
"step": 47
},
{
"epoch": 0.009032743695897629,
"grad_norm": 1.6337559223175049,
"learning_rate": 9.895743333948874e-05,
"loss": 3.7759,
"step": 48
},
{
"epoch": 0.009220925856228829,
"grad_norm": 2.088834285736084,
"learning_rate": 9.888191821396744e-05,
"loss": 3.8653,
"step": 49
},
{
"epoch": 0.00940910801656003,
"grad_norm": 2.713131904602051,
"learning_rate": 9.880379387779637e-05,
"loss": 3.468,
"step": 50
},
{
"epoch": 0.00959729017689123,
"grad_norm": 1.8180787563323975,
"learning_rate": 9.872306450108292e-05,
"loss": 3.5826,
"step": 51
},
{
"epoch": 0.00978547233722243,
"grad_norm": 1.6677582263946533,
"learning_rate": 9.863973439298597e-05,
"loss": 3.7982,
"step": 52
},
{
"epoch": 0.009973654497553632,
"grad_norm": 1.569620966911316,
"learning_rate": 9.855380800148572e-05,
"loss": 3.7162,
"step": 53
},
{
"epoch": 0.010161836657884832,
"grad_norm": 1.6128700971603394,
"learning_rate": 9.846528991314639e-05,
"loss": 3.5721,
"step": 54
},
{
"epoch": 0.010350018818216034,
"grad_norm": 1.8484185934066772,
"learning_rate": 9.837418485287127e-05,
"loss": 3.9133,
"step": 55
},
{
"epoch": 0.010538200978547234,
"grad_norm": 1.706660270690918,
"learning_rate": 9.828049768365068e-05,
"loss": 3.7007,
"step": 56
},
{
"epoch": 0.010726383138878434,
"grad_norm": 1.7316683530807495,
"learning_rate": 9.818423340630228e-05,
"loss": 3.7366,
"step": 57
},
{
"epoch": 0.010914565299209636,
"grad_norm": 1.5684834718704224,
"learning_rate": 9.808539715920414e-05,
"loss": 3.6784,
"step": 58
},
{
"epoch": 0.011102747459540836,
"grad_norm": 1.5667593479156494,
"learning_rate": 9.798399421802056e-05,
"loss": 3.7003,
"step": 59
},
{
"epoch": 0.011290929619872036,
"grad_norm": 1.526808500289917,
"learning_rate": 9.78800299954203e-05,
"loss": 3.7078,
"step": 60
},
{
"epoch": 0.011479111780203237,
"grad_norm": 2.5877087116241455,
"learning_rate": 9.777351004078783e-05,
"loss": 3.6875,
"step": 61
},
{
"epoch": 0.011667293940534437,
"grad_norm": 1.545409917831421,
"learning_rate": 9.766444003992703e-05,
"loss": 3.8299,
"step": 62
},
{
"epoch": 0.011855476100865637,
"grad_norm": 1.5253973007202148,
"learning_rate": 9.755282581475769e-05,
"loss": 3.6405,
"step": 63
},
{
"epoch": 0.012043658261196839,
"grad_norm": 1.8160947561264038,
"learning_rate": 9.743867332300478e-05,
"loss": 3.7575,
"step": 64
},
{
"epoch": 0.012231840421528039,
"grad_norm": 1.62673819065094,
"learning_rate": 9.732198865788047e-05,
"loss": 3.736,
"step": 65
},
{
"epoch": 0.012420022581859239,
"grad_norm": 1.609445571899414,
"learning_rate": 9.72027780477588e-05,
"loss": 3.5029,
"step": 66
},
{
"epoch": 0.01260820474219044,
"grad_norm": 1.640405535697937,
"learning_rate": 9.708104785584323e-05,
"loss": 3.6871,
"step": 67
},
{
"epoch": 0.01279638690252164,
"grad_norm": 1.546713948249817,
"learning_rate": 9.695680457982713e-05,
"loss": 3.5635,
"step": 68
},
{
"epoch": 0.012984569062852842,
"grad_norm": 1.525071620941162,
"learning_rate": 9.683005485154677e-05,
"loss": 3.6238,
"step": 69
},
{
"epoch": 0.013172751223184042,
"grad_norm": 1.5092836618423462,
"learning_rate": 9.67008054366274e-05,
"loss": 3.6092,
"step": 70
},
{
"epoch": 0.013360933383515242,
"grad_norm": 1.6913105249404907,
"learning_rate": 9.656906323412217e-05,
"loss": 3.6286,
"step": 71
},
{
"epoch": 0.013549115543846444,
"grad_norm": 2.151688814163208,
"learning_rate": 9.643483527614372e-05,
"loss": 3.5817,
"step": 72
},
{
"epoch": 0.013737297704177644,
"grad_norm": 1.5763871669769287,
"learning_rate": 9.629812872748901e-05,
"loss": 3.5043,
"step": 73
},
{
"epoch": 0.013925479864508844,
"grad_norm": 1.5946530103683472,
"learning_rate": 9.615895088525677e-05,
"loss": 3.5902,
"step": 74
},
{
"epoch": 0.014113662024840046,
"grad_norm": 1.5125662088394165,
"learning_rate": 9.601730917845797e-05,
"loss": 3.5002,
"step": 75
},
{
"epoch": 0.014301844185171246,
"grad_norm": 1.952149510383606,
"learning_rate": 9.587321116761938e-05,
"loss": 3.5881,
"step": 76
},
{
"epoch": 0.014490026345502446,
"grad_norm": 1.5184606313705444,
"learning_rate": 9.57266645443799e-05,
"loss": 3.3213,
"step": 77
},
{
"epoch": 0.014678208505833647,
"grad_norm": 1.5643829107284546,
"learning_rate": 9.557767713108009e-05,
"loss": 3.7112,
"step": 78
},
{
"epoch": 0.014866390666164847,
"grad_norm": 1.5019704103469849,
"learning_rate": 9.542625688034449e-05,
"loss": 3.4665,
"step": 79
},
{
"epoch": 0.015054572826496047,
"grad_norm": 1.647053837776184,
"learning_rate": 9.527241187465734e-05,
"loss": 3.33,
"step": 80
},
{
"epoch": 0.015242754986827249,
"grad_norm": 1.527642846107483,
"learning_rate": 9.511615032593096e-05,
"loss": 3.6355,
"step": 81
},
{
"epoch": 0.015430937147158449,
"grad_norm": 1.6316813230514526,
"learning_rate": 9.49574805750675e-05,
"loss": 3.5475,
"step": 82
},
{
"epoch": 0.01561911930748965,
"grad_norm": 1.6150362491607666,
"learning_rate": 9.479641109151373e-05,
"loss": 3.5272,
"step": 83
},
{
"epoch": 0.01580730146782085,
"grad_norm": 1.5907738208770752,
"learning_rate": 9.463295047280891e-05,
"loss": 3.6044,
"step": 84
},
{
"epoch": 0.015995483628152053,
"grad_norm": 1.5602425336837769,
"learning_rate": 9.446710744412595e-05,
"loss": 3.4846,
"step": 85
},
{
"epoch": 0.01618366578848325,
"grad_norm": 1.5599199533462524,
"learning_rate": 9.429889085780557e-05,
"loss": 3.598,
"step": 86
},
{
"epoch": 0.016371847948814452,
"grad_norm": 1.4632290601730347,
"learning_rate": 9.41283096928839e-05,
"loss": 3.3955,
"step": 87
},
{
"epoch": 0.016560030109145654,
"grad_norm": 1.5269198417663574,
"learning_rate": 9.395537305461311e-05,
"loss": 3.5423,
"step": 88
},
{
"epoch": 0.016748212269476852,
"grad_norm": 1.6142536401748657,
"learning_rate": 9.378009017397542e-05,
"loss": 3.5366,
"step": 89
},
{
"epoch": 0.016936394429808054,
"grad_norm": 1.4717563390731812,
"learning_rate": 9.360247040719039e-05,
"loss": 3.4092,
"step": 90
},
{
"epoch": 0.017124576590139256,
"grad_norm": 2.3025355339050293,
"learning_rate": 9.342252323521545e-05,
"loss": 3.441,
"step": 91
},
{
"epoch": 0.017312758750470454,
"grad_norm": 1.6364383697509766,
"learning_rate": 9.324025826323994e-05,
"loss": 3.4081,
"step": 92
},
{
"epoch": 0.017500940910801656,
"grad_norm": 1.6637694835662842,
"learning_rate": 9.305568522017227e-05,
"loss": 3.4648,
"step": 93
},
{
"epoch": 0.017689123071132858,
"grad_norm": 1.5033453702926636,
"learning_rate": 9.286881395812066e-05,
"loss": 3.4444,
"step": 94
},
{
"epoch": 0.017877305231464056,
"grad_norm": 1.7990686893463135,
"learning_rate": 9.267965445186733e-05,
"loss": 3.4532,
"step": 95
},
{
"epoch": 0.018065487391795258,
"grad_norm": 1.5659180879592896,
"learning_rate": 9.248821679833596e-05,
"loss": 3.444,
"step": 96
},
{
"epoch": 0.01825366955212646,
"grad_norm": 1.495245337486267,
"learning_rate": 9.229451121605279e-05,
"loss": 3.5251,
"step": 97
},
{
"epoch": 0.018441851712457658,
"grad_norm": 1.6818476915359497,
"learning_rate": 9.209854804460121e-05,
"loss": 3.323,
"step": 98
},
{
"epoch": 0.01863003387278886,
"grad_norm": 1.9330247640609741,
"learning_rate": 9.190033774406977e-05,
"loss": 3.4345,
"step": 99
},
{
"epoch": 0.01881821603312006,
"grad_norm": 2.6322007179260254,
"learning_rate": 9.16998908944939e-05,
"loss": 3.2795,
"step": 100
},
{
"epoch": 0.01900639819345126,
"grad_norm": 1.6533123254776,
"learning_rate": 9.149721819529119e-05,
"loss": 3.6009,
"step": 101
},
{
"epoch": 0.01919458035378246,
"grad_norm": 1.739073395729065,
"learning_rate": 9.129233046469022e-05,
"loss": 3.8017,
"step": 102
},
{
"epoch": 0.019382762514113663,
"grad_norm": 1.728940486907959,
"learning_rate": 9.108523863915314e-05,
"loss": 3.7084,
"step": 103
},
{
"epoch": 0.01957094467444486,
"grad_norm": 1.6274477243423462,
"learning_rate": 9.087595377279192e-05,
"loss": 3.5057,
"step": 104
},
{
"epoch": 0.019759126834776063,
"grad_norm": 1.6262691020965576,
"learning_rate": 9.066448703677828e-05,
"loss": 3.6988,
"step": 105
},
{
"epoch": 0.019947308995107264,
"grad_norm": 4.310033321380615,
"learning_rate": 9.045084971874738e-05,
"loss": 3.7893,
"step": 106
},
{
"epoch": 0.020135491155438466,
"grad_norm": 6.6696977615356445,
"learning_rate": 9.023505322219536e-05,
"loss": 3.5515,
"step": 107
},
{
"epoch": 0.020323673315769664,
"grad_norm": 1.6663572788238525,
"learning_rate": 9.001710906587064e-05,
"loss": 3.4112,
"step": 108
},
{
"epoch": 0.020511855476100866,
"grad_norm": 1.6214839220046997,
"learning_rate": 8.9797028883159e-05,
"loss": 3.3542,
"step": 109
},
{
"epoch": 0.020700037636432068,
"grad_norm": 1.540847897529602,
"learning_rate": 8.957482442146272e-05,
"loss": 3.6587,
"step": 110
},
{
"epoch": 0.020888219796763266,
"grad_norm": 1.5367134809494019,
"learning_rate": 8.935050754157344e-05,
"loss": 3.5199,
"step": 111
},
{
"epoch": 0.021076401957094468,
"grad_norm": 1.7050776481628418,
"learning_rate": 8.912409021703913e-05,
"loss": 3.6533,
"step": 112
},
{
"epoch": 0.02126458411742567,
"grad_norm": 1.5632902383804321,
"learning_rate": 8.889558453352492e-05,
"loss": 3.5123,
"step": 113
},
{
"epoch": 0.021452766277756868,
"grad_norm": 1.5934125185012817,
"learning_rate": 8.866500268816803e-05,
"loss": 3.5423,
"step": 114
},
{
"epoch": 0.02164094843808807,
"grad_norm": 1.5730620622634888,
"learning_rate": 8.84323569889266e-05,
"loss": 3.3918,
"step": 115
},
{
"epoch": 0.02182913059841927,
"grad_norm": 1.5384756326675415,
"learning_rate": 8.819765985392296e-05,
"loss": 3.5834,
"step": 116
},
{
"epoch": 0.02201731275875047,
"grad_norm": 1.4878556728363037,
"learning_rate": 8.79609238107805e-05,
"loss": 3.5273,
"step": 117
},
{
"epoch": 0.02220549491908167,
"grad_norm": 1.8022403717041016,
"learning_rate": 8.772216149595513e-05,
"loss": 3.6865,
"step": 118
},
{
"epoch": 0.022393677079412873,
"grad_norm": 1.6074386835098267,
"learning_rate": 8.748138565406081e-05,
"loss": 3.35,
"step": 119
},
{
"epoch": 0.02258185923974407,
"grad_norm": 5.877849102020264,
"learning_rate": 8.72386091371891e-05,
"loss": 3.5387,
"step": 120
},
{
"epoch": 0.022770041400075273,
"grad_norm": 1.6074655055999756,
"learning_rate": 8.699384490422331e-05,
"loss": 3.4726,
"step": 121
},
{
"epoch": 0.022958223560406475,
"grad_norm": 1.5693819522857666,
"learning_rate": 8.674710602014671e-05,
"loss": 3.4279,
"step": 122
},
{
"epoch": 0.023146405720737673,
"grad_norm": 1.5131280422210693,
"learning_rate": 8.649840565534513e-05,
"loss": 3.4739,
"step": 123
},
{
"epoch": 0.023334587881068874,
"grad_norm": 1.5035758018493652,
"learning_rate": 8.624775708490402e-05,
"loss": 3.4492,
"step": 124
},
{
"epoch": 0.023522770041400076,
"grad_norm": 1.6009769439697266,
"learning_rate": 8.59951736878998e-05,
"loss": 3.6759,
"step": 125
},
{
"epoch": 0.023710952201731274,
"grad_norm": 1.539108395576477,
"learning_rate": 8.574066894668573e-05,
"loss": 3.5554,
"step": 126
},
{
"epoch": 0.023899134362062476,
"grad_norm": 1.8990849256515503,
"learning_rate": 8.548425644617224e-05,
"loss": 3.4451,
"step": 127
},
{
"epoch": 0.024087316522393678,
"grad_norm": 1.5071018934249878,
"learning_rate": 8.522594987310184e-05,
"loss": 3.4815,
"step": 128
},
{
"epoch": 0.024275498682724876,
"grad_norm": 1.5156155824661255,
"learning_rate": 8.49657630153185e-05,
"loss": 3.526,
"step": 129
},
{
"epoch": 0.024463680843056078,
"grad_norm": 1.4841188192367554,
"learning_rate": 8.47037097610317e-05,
"loss": 3.4503,
"step": 130
},
{
"epoch": 0.02465186300338728,
"grad_norm": 1.8373347520828247,
"learning_rate": 8.443980409807512e-05,
"loss": 3.4492,
"step": 131
},
{
"epoch": 0.024840045163718478,
"grad_norm": 1.9632869958877563,
"learning_rate": 8.417406011315998e-05,
"loss": 3.5533,
"step": 132
},
{
"epoch": 0.02502822732404968,
"grad_norm": 1.525312900543213,
"learning_rate": 8.390649199112315e-05,
"loss": 3.5785,
"step": 133
},
{
"epoch": 0.02521640948438088,
"grad_norm": 1.7048442363739014,
"learning_rate": 8.363711401417e-05,
"loss": 3.5644,
"step": 134
},
{
"epoch": 0.025404591644712083,
"grad_norm": 1.4897756576538086,
"learning_rate": 8.336594056111197e-05,
"loss": 3.4262,
"step": 135
},
{
"epoch": 0.02559277380504328,
"grad_norm": 1.5408360958099365,
"learning_rate": 8.309298610659916e-05,
"loss": 3.5162,
"step": 136
},
{
"epoch": 0.025780955965374483,
"grad_norm": 1.613747239112854,
"learning_rate": 8.281826522034764e-05,
"loss": 3.6368,
"step": 137
},
{
"epoch": 0.025969138125705685,
"grad_norm": 1.5057644844055176,
"learning_rate": 8.254179256636179e-05,
"loss": 3.4841,
"step": 138
},
{
"epoch": 0.026157320286036883,
"grad_norm": 1.5167312622070312,
"learning_rate": 8.226358290215151e-05,
"loss": 3.3391,
"step": 139
},
{
"epoch": 0.026345502446368085,
"grad_norm": 1.5216038227081299,
"learning_rate": 8.198365107794457e-05,
"loss": 3.3654,
"step": 140
},
{
"epoch": 0.026533684606699286,
"grad_norm": 1.5505784749984741,
"learning_rate": 8.17020120358939e-05,
"loss": 3.461,
"step": 141
},
{
"epoch": 0.026721866767030485,
"grad_norm": 3.795452833175659,
"learning_rate": 8.141868080927996e-05,
"loss": 3.4515,
"step": 142
},
{
"epoch": 0.026910048927361686,
"grad_norm": 1.4928878545761108,
"learning_rate": 8.113367252170844e-05,
"loss": 3.4339,
"step": 143
},
{
"epoch": 0.027098231087692888,
"grad_norm": 1.5022090673446655,
"learning_rate": 8.084700238630283e-05,
"loss": 3.3639,
"step": 144
},
{
"epoch": 0.027286413248024086,
"grad_norm": 1.4917182922363281,
"learning_rate": 8.055868570489247e-05,
"loss": 3.2841,
"step": 145
},
{
"epoch": 0.027474595408355288,
"grad_norm": 1.5091995000839233,
"learning_rate": 8.026873786719573e-05,
"loss": 3.4067,
"step": 146
},
{
"epoch": 0.02766277756868649,
"grad_norm": 1.455728530883789,
"learning_rate": 7.997717434999861e-05,
"loss": 3.2996,
"step": 147
},
{
"epoch": 0.027850959729017688,
"grad_norm": 1.852761149406433,
"learning_rate": 7.968401071632855e-05,
"loss": 3.4227,
"step": 148
},
{
"epoch": 0.02803914188934889,
"grad_norm": 1.8084666728973389,
"learning_rate": 7.938926261462366e-05,
"loss": 3.4068,
"step": 149
},
{
"epoch": 0.02822732404968009,
"grad_norm": 2.763946056365967,
"learning_rate": 7.909294577789766e-05,
"loss": 3.1156,
"step": 150
},
{
"epoch": 0.02822732404968009,
"eval_loss": 0.8662680983543396,
"eval_runtime": 982.7503,
"eval_samples_per_second": 9.107,
"eval_steps_per_second": 2.277,
"step": 150
},
{
"epoch": 0.02841550621001129,
"grad_norm": 1.5687919855117798,
"learning_rate": 7.879507602289979e-05,
"loss": 3.6073,
"step": 151
},
{
"epoch": 0.02860368837034249,
"grad_norm": 1.6533427238464355,
"learning_rate": 7.849566924927082e-05,
"loss": 3.5083,
"step": 152
},
{
"epoch": 0.028791870530673693,
"grad_norm": 1.542561650276184,
"learning_rate": 7.819474143869414e-05,
"loss": 3.4967,
"step": 153
},
{
"epoch": 0.02898005269100489,
"grad_norm": 1.516094446182251,
"learning_rate": 7.789230865404287e-05,
"loss": 3.3608,
"step": 154
},
{
"epoch": 0.029168234851336093,
"grad_norm": 1.6348127126693726,
"learning_rate": 7.75883870385223e-05,
"loss": 3.5607,
"step": 155
},
{
"epoch": 0.029356417011667295,
"grad_norm": 1.574805498123169,
"learning_rate": 7.728299281480833e-05,
"loss": 3.314,
"step": 156
},
{
"epoch": 0.029544599171998493,
"grad_norm": 1.5510149002075195,
"learning_rate": 7.697614228418148e-05,
"loss": 3.397,
"step": 157
},
{
"epoch": 0.029732781332329695,
"grad_norm": 1.4977482557296753,
"learning_rate": 7.666785182565677e-05,
"loss": 3.4013,
"step": 158
},
{
"epoch": 0.029920963492660897,
"grad_norm": 1.5679925680160522,
"learning_rate": 7.635813789510941e-05,
"loss": 3.3312,
"step": 159
},
{
"epoch": 0.030109145652992095,
"grad_norm": 1.5352201461791992,
"learning_rate": 7.604701702439651e-05,
"loss": 3.483,
"step": 160
},
{
"epoch": 0.030297327813323296,
"grad_norm": 1.52545964717865,
"learning_rate": 7.573450582047457e-05,
"loss": 3.5098,
"step": 161
},
{
"epoch": 0.030485509973654498,
"grad_norm": 1.5153985023498535,
"learning_rate": 7.542062096451305e-05,
"loss": 3.4748,
"step": 162
},
{
"epoch": 0.030673692133985696,
"grad_norm": 1.4865204095840454,
"learning_rate": 7.510537921100398e-05,
"loss": 3.3997,
"step": 163
},
{
"epoch": 0.030861874294316898,
"grad_norm": 1.5080010890960693,
"learning_rate": 7.47887973868676e-05,
"loss": 3.5227,
"step": 164
},
{
"epoch": 0.0310500564546481,
"grad_norm": 1.5261732339859009,
"learning_rate": 7.447089239055428e-05,
"loss": 3.7215,
"step": 165
},
{
"epoch": 0.0312382386149793,
"grad_norm": 1.4631050825119019,
"learning_rate": 7.41516811911424e-05,
"loss": 3.3073,
"step": 166
},
{
"epoch": 0.0314264207753105,
"grad_norm": 1.5799752473831177,
"learning_rate": 7.383118082743262e-05,
"loss": 3.5134,
"step": 167
},
{
"epoch": 0.0316146029356417,
"grad_norm": 1.6088528633117676,
"learning_rate": 7.350940840703842e-05,
"loss": 3.4344,
"step": 168
},
{
"epoch": 0.0318027850959729,
"grad_norm": 1.5857129096984863,
"learning_rate": 7.318638110547288e-05,
"loss": 3.6729,
"step": 169
},
{
"epoch": 0.031990967256304105,
"grad_norm": 1.9101580381393433,
"learning_rate": 7.286211616523193e-05,
"loss": 3.4667,
"step": 170
},
{
"epoch": 0.0321791494166353,
"grad_norm": 1.6800731420516968,
"learning_rate": 7.253663089487395e-05,
"loss": 3.3363,
"step": 171
},
{
"epoch": 0.0323673315769665,
"grad_norm": 1.5496375560760498,
"learning_rate": 7.220994266809591e-05,
"loss": 3.5252,
"step": 172
},
{
"epoch": 0.0325555137372977,
"grad_norm": 1.4972703456878662,
"learning_rate": 7.188206892280594e-05,
"loss": 3.43,
"step": 173
},
{
"epoch": 0.032743695897628905,
"grad_norm": 1.4559191465377808,
"learning_rate": 7.155302716019263e-05,
"loss": 3.2506,
"step": 174
},
{
"epoch": 0.03293187805796011,
"grad_norm": 1.6048904657363892,
"learning_rate": 7.122283494379076e-05,
"loss": 3.6594,
"step": 175
},
{
"epoch": 0.03312006021829131,
"grad_norm": 1.5769932270050049,
"learning_rate": 7.089150989854385e-05,
"loss": 3.3905,
"step": 176
},
{
"epoch": 0.0333082423786225,
"grad_norm": 1.508113980293274,
"learning_rate": 7.055906970986336e-05,
"loss": 3.4611,
"step": 177
},
{
"epoch": 0.033496424538953705,
"grad_norm": 1.4894671440124512,
"learning_rate": 7.022553212268469e-05,
"loss": 3.3249,
"step": 178
},
{
"epoch": 0.03368460669928491,
"grad_norm": 1.5143218040466309,
"learning_rate": 6.989091494051998e-05,
"loss": 3.3934,
"step": 179
},
{
"epoch": 0.03387278885961611,
"grad_norm": 1.4756419658660889,
"learning_rate": 6.95552360245078e-05,
"loss": 3.2264,
"step": 180
},
{
"epoch": 0.03406097101994731,
"grad_norm": 1.5053011178970337,
"learning_rate": 6.92185132924598e-05,
"loss": 3.5413,
"step": 181
},
{
"epoch": 0.03424915318027851,
"grad_norm": 1.8715499639511108,
"learning_rate": 6.888076471790424e-05,
"loss": 3.4045,
"step": 182
},
{
"epoch": 0.034437335340609714,
"grad_norm": 1.8220783472061157,
"learning_rate": 6.85420083291266e-05,
"loss": 3.4066,
"step": 183
},
{
"epoch": 0.03462551750094091,
"grad_norm": 1.5312058925628662,
"learning_rate": 6.820226220820732e-05,
"loss": 3.474,
"step": 184
},
{
"epoch": 0.03481369966127211,
"grad_norm": 1.5189887285232544,
"learning_rate": 6.786154449005665e-05,
"loss": 3.2858,
"step": 185
},
{
"epoch": 0.03500188182160331,
"grad_norm": 1.5011916160583496,
"learning_rate": 6.751987336144648e-05,
"loss": 3.3292,
"step": 186
},
{
"epoch": 0.03519006398193451,
"grad_norm": 1.6477534770965576,
"learning_rate": 6.717726706003974e-05,
"loss": 3.4825,
"step": 187
},
{
"epoch": 0.035378246142265715,
"grad_norm": 1.5640369653701782,
"learning_rate": 6.683374387341687e-05,
"loss": 3.3886,
"step": 188
},
{
"epoch": 0.03556642830259692,
"grad_norm": 1.5490057468414307,
"learning_rate": 6.648932213809962e-05,
"loss": 3.4445,
"step": 189
},
{
"epoch": 0.03575461046292811,
"grad_norm": 1.5441166162490845,
"learning_rate": 6.614402023857232e-05,
"loss": 3.2839,
"step": 190
},
{
"epoch": 0.03594279262325931,
"grad_norm": 1.5490269660949707,
"learning_rate": 6.579785660630056e-05,
"loss": 3.548,
"step": 191
},
{
"epoch": 0.036130974783590515,
"grad_norm": 1.4607436656951904,
"learning_rate": 6.545084971874738e-05,
"loss": 3.3022,
"step": 192
},
{
"epoch": 0.03631915694392172,
"grad_norm": 1.5263044834136963,
"learning_rate": 6.510301809838689e-05,
"loss": 3.5151,
"step": 193
},
{
"epoch": 0.03650733910425292,
"grad_norm": 1.7463645935058594,
"learning_rate": 6.475438031171574e-05,
"loss": 3.4085,
"step": 194
},
{
"epoch": 0.03669552126458412,
"grad_norm": 1.4995930194854736,
"learning_rate": 6.440495496826189e-05,
"loss": 3.4332,
"step": 195
},
{
"epoch": 0.036883703424915315,
"grad_norm": 1.5385165214538574,
"learning_rate": 6.405476071959143e-05,
"loss": 3.432,
"step": 196
},
{
"epoch": 0.03707188558524652,
"grad_norm": 1.5226582288742065,
"learning_rate": 6.370381625831292e-05,
"loss": 3.3812,
"step": 197
},
{
"epoch": 0.03726006774557772,
"grad_norm": 1.5849664211273193,
"learning_rate": 6.335214031707965e-05,
"loss": 3.4423,
"step": 198
},
{
"epoch": 0.03744824990590892,
"grad_norm": 1.8131951093673706,
"learning_rate": 6.299975166758971e-05,
"loss": 3.3169,
"step": 199
},
{
"epoch": 0.03763643206624012,
"grad_norm": 2.523836851119995,
"learning_rate": 6.264666911958404e-05,
"loss": 3.1207,
"step": 200
},
{
"epoch": 0.037824614226571324,
"grad_norm": 1.6073590517044067,
"learning_rate": 6.229291151984233e-05,
"loss": 3.5545,
"step": 201
},
{
"epoch": 0.03801279638690252,
"grad_norm": 3.597961187362671,
"learning_rate": 6.19384977511771e-05,
"loss": 3.3569,
"step": 202
},
{
"epoch": 0.03820097854723372,
"grad_norm": 1.5076078176498413,
"learning_rate": 6.158344673142573e-05,
"loss": 3.5497,
"step": 203
},
{
"epoch": 0.03838916070756492,
"grad_norm": 1.5262771844863892,
"learning_rate": 6.122777741244067e-05,
"loss": 3.5109,
"step": 204
},
{
"epoch": 0.038577342867896124,
"grad_norm": 1.5841517448425293,
"learning_rate": 6.0871508779077856e-05,
"loss": 3.6458,
"step": 205
},
{
"epoch": 0.038765525028227325,
"grad_norm": 1.706384539604187,
"learning_rate": 6.051465984818332e-05,
"loss": 3.4581,
"step": 206
},
{
"epoch": 0.03895370718855853,
"grad_norm": 1.562947154045105,
"learning_rate": 6.015724966757812e-05,
"loss": 3.6251,
"step": 207
},
{
"epoch": 0.03914188934888972,
"grad_norm": 1.5278199911117554,
"learning_rate": 5.979929731504158e-05,
"loss": 3.4025,
"step": 208
},
{
"epoch": 0.039330071509220924,
"grad_norm": 1.528238296508789,
"learning_rate": 5.944082189729301e-05,
"loss": 3.5747,
"step": 209
},
{
"epoch": 0.039518253669552125,
"grad_norm": 1.6547585725784302,
"learning_rate": 5.908184254897182e-05,
"loss": 3.4107,
"step": 210
},
{
"epoch": 0.03970643582988333,
"grad_norm": 1.5595186948776245,
"learning_rate": 5.872237843161612e-05,
"loss": 3.4565,
"step": 211
},
{
"epoch": 0.03989461799021453,
"grad_norm": 1.5088775157928467,
"learning_rate": 5.8362448732639894e-05,
"loss": 3.3846,
"step": 212
},
{
"epoch": 0.04008280015054573,
"grad_norm": 1.543283224105835,
"learning_rate": 5.800207266430895e-05,
"loss": 3.5448,
"step": 213
},
{
"epoch": 0.04027098231087693,
"grad_norm": 1.4654241800308228,
"learning_rate": 5.764126946271526e-05,
"loss": 3.3228,
"step": 214
},
{
"epoch": 0.04045916447120813,
"grad_norm": 1.6389522552490234,
"learning_rate": 5.7280058386750255e-05,
"loss": 3.357,
"step": 215
},
{
"epoch": 0.04064734663153933,
"grad_norm": 1.4927722215652466,
"learning_rate": 5.6918458717076815e-05,
"loss": 3.434,
"step": 216
},
{
"epoch": 0.04083552879187053,
"grad_norm": 1.5620532035827637,
"learning_rate": 5.655648975510014e-05,
"loss": 3.5867,
"step": 217
},
{
"epoch": 0.04102371095220173,
"grad_norm": 1.6729021072387695,
"learning_rate": 5.61941708219374e-05,
"loss": 3.3869,
"step": 218
},
{
"epoch": 0.041211893112532934,
"grad_norm": 1.523036003112793,
"learning_rate": 5.583152125738651e-05,
"loss": 3.4844,
"step": 219
},
{
"epoch": 0.041400075272864136,
"grad_norm": 1.4695093631744385,
"learning_rate": 5.546856041889373e-05,
"loss": 3.4079,
"step": 220
},
{
"epoch": 0.04158825743319533,
"grad_norm": 1.525160789489746,
"learning_rate": 5.510530768052047e-05,
"loss": 3.358,
"step": 221
},
{
"epoch": 0.04177643959352653,
"grad_norm": 1.4674233198165894,
"learning_rate": 5.4741782431909136e-05,
"loss": 3.2476,
"step": 222
},
{
"epoch": 0.041964621753857734,
"grad_norm": 1.7390276193618774,
"learning_rate": 5.437800407724812e-05,
"loss": 3.3376,
"step": 223
},
{
"epoch": 0.042152803914188935,
"grad_norm": 1.5143543481826782,
"learning_rate": 5.401399203423606e-05,
"loss": 3.5849,
"step": 224
},
{
"epoch": 0.04234098607452014,
"grad_norm": 1.5307164192199707,
"learning_rate": 5.364976573304538e-05,
"loss": 3.2994,
"step": 225
},
{
"epoch": 0.04252916823485134,
"grad_norm": 1.5590108633041382,
"learning_rate": 5.328534461528515e-05,
"loss": 3.5251,
"step": 226
},
{
"epoch": 0.042717350395182534,
"grad_norm": 1.5407838821411133,
"learning_rate": 5.29207481329633e-05,
"loss": 3.3932,
"step": 227
},
{
"epoch": 0.042905532555513735,
"grad_norm": 1.5244678258895874,
"learning_rate": 5.2555995747448364e-05,
"loss": 3.3437,
"step": 228
},
{
"epoch": 0.04309371471584494,
"grad_norm": 1.4933329820632935,
"learning_rate": 5.2191106928430644e-05,
"loss": 3.399,
"step": 229
},
{
"epoch": 0.04328189687617614,
"grad_norm": 1.4906306266784668,
"learning_rate": 5.182610115288295e-05,
"loss": 3.2872,
"step": 230
},
{
"epoch": 0.04347007903650734,
"grad_norm": 1.647549033164978,
"learning_rate": 5.1460997904021005e-05,
"loss": 3.3683,
"step": 231
},
{
"epoch": 0.04365826119683854,
"grad_norm": 1.5551615953445435,
"learning_rate": 5.109581667026341e-05,
"loss": 3.4049,
"step": 232
},
{
"epoch": 0.04384644335716974,
"grad_norm": 1.5681029558181763,
"learning_rate": 5.073057694419147e-05,
"loss": 3.4455,
"step": 233
},
{
"epoch": 0.04403462551750094,
"grad_norm": 1.5506995916366577,
"learning_rate": 5.036529822150865e-05,
"loss": 3.3317,
"step": 234
},
{
"epoch": 0.04422280767783214,
"grad_norm": 2.233480215072632,
"learning_rate": 5e-05,
"loss": 3.4562,
"step": 235
},
{
"epoch": 0.04441098983816334,
"grad_norm": 1.4921319484710693,
"learning_rate": 4.963470177849135e-05,
"loss": 3.3868,
"step": 236
},
{
"epoch": 0.044599171998494544,
"grad_norm": 1.539383888244629,
"learning_rate": 4.9269423055808544e-05,
"loss": 3.3888,
"step": 237
},
{
"epoch": 0.044787354158825746,
"grad_norm": 1.494940161705017,
"learning_rate": 4.8904183329736596e-05,
"loss": 3.3123,
"step": 238
},
{
"epoch": 0.04497553631915695,
"grad_norm": 1.6059819459915161,
"learning_rate": 4.853900209597903e-05,
"loss": 3.5323,
"step": 239
},
{
"epoch": 0.04516371847948814,
"grad_norm": 1.5235280990600586,
"learning_rate": 4.817389884711705e-05,
"loss": 3.4477,
"step": 240
},
{
"epoch": 0.045351900639819344,
"grad_norm": 1.532889485359192,
"learning_rate": 4.7808893071569374e-05,
"loss": 3.4776,
"step": 241
},
{
"epoch": 0.045540082800150546,
"grad_norm": 1.5243102312088013,
"learning_rate": 4.744400425255165e-05,
"loss": 3.4302,
"step": 242
},
{
"epoch": 0.04572826496048175,
"grad_norm": 1.4719178676605225,
"learning_rate": 4.707925186703671e-05,
"loss": 3.2717,
"step": 243
},
{
"epoch": 0.04591644712081295,
"grad_norm": 1.5308884382247925,
"learning_rate": 4.671465538471486e-05,
"loss": 3.5024,
"step": 244
},
{
"epoch": 0.04610462928114415,
"grad_norm": 1.5471347570419312,
"learning_rate": 4.6350234266954626e-05,
"loss": 3.3279,
"step": 245
},
{
"epoch": 0.046292811441475346,
"grad_norm": 1.4882664680480957,
"learning_rate": 4.598600796576395e-05,
"loss": 3.2842,
"step": 246
},
{
"epoch": 0.04648099360180655,
"grad_norm": 1.546363115310669,
"learning_rate": 4.562199592275188e-05,
"loss": 3.7441,
"step": 247
},
{
"epoch": 0.04666917576213775,
"grad_norm": 1.5140584707260132,
"learning_rate": 4.5258217568090876e-05,
"loss": 3.337,
"step": 248
},
{
"epoch": 0.04685735792246895,
"grad_norm": 1.894798755645752,
"learning_rate": 4.4894692319479544e-05,
"loss": 3.2422,
"step": 249
},
{
"epoch": 0.04704554008280015,
"grad_norm": 2.647658109664917,
"learning_rate": 4.4531439581106295e-05,
"loss": 3.1319,
"step": 250
},
{
"epoch": 0.047233722243131354,
"grad_norm": 1.497327446937561,
"learning_rate": 4.4168478742613506e-05,
"loss": 3.3551,
"step": 251
},
{
"epoch": 0.04742190440346255,
"grad_norm": 1.5717421770095825,
"learning_rate": 4.38058291780626e-05,
"loss": 3.7432,
"step": 252
},
{
"epoch": 0.04761008656379375,
"grad_norm": 1.4942125082015991,
"learning_rate": 4.3443510244899864e-05,
"loss": 3.4546,
"step": 253
},
{
"epoch": 0.04779826872412495,
"grad_norm": 1.5224213600158691,
"learning_rate": 4.308154128292318e-05,
"loss": 3.4569,
"step": 254
},
{
"epoch": 0.047986450884456154,
"grad_norm": 3.2993710041046143,
"learning_rate": 4.271994161324977e-05,
"loss": 3.4279,
"step": 255
},
{
"epoch": 0.048174633044787356,
"grad_norm": 1.776840090751648,
"learning_rate": 4.235873053728475e-05,
"loss": 3.4157,
"step": 256
},
{
"epoch": 0.04836281520511856,
"grad_norm": 1.5063600540161133,
"learning_rate": 4.199792733569107e-05,
"loss": 3.4676,
"step": 257
},
{
"epoch": 0.04855099736544975,
"grad_norm": 1.5720107555389404,
"learning_rate": 4.163755126736012e-05,
"loss": 3.5201,
"step": 258
},
{
"epoch": 0.048739179525780954,
"grad_norm": 1.5020864009857178,
"learning_rate": 4.127762156838389e-05,
"loss": 3.382,
"step": 259
},
{
"epoch": 0.048927361686112156,
"grad_norm": 1.516310214996338,
"learning_rate": 4.0918157451028185e-05,
"loss": 3.4377,
"step": 260
},
{
"epoch": 0.04911554384644336,
"grad_norm": 1.50225031375885,
"learning_rate": 4.055917810270698e-05,
"loss": 3.4051,
"step": 261
},
{
"epoch": 0.04930372600677456,
"grad_norm": 1.5030755996704102,
"learning_rate": 4.020070268495843e-05,
"loss": 3.3375,
"step": 262
},
{
"epoch": 0.04949190816710576,
"grad_norm": 1.4907896518707275,
"learning_rate": 3.9842750332421896e-05,
"loss": 3.4886,
"step": 263
},
{
"epoch": 0.049680090327436956,
"grad_norm": 1.4596471786499023,
"learning_rate": 3.94853401518167e-05,
"loss": 3.2683,
"step": 264
},
{
"epoch": 0.04986827248776816,
"grad_norm": 1.4404525756835938,
"learning_rate": 3.9128491220922156e-05,
"loss": 3.3169,
"step": 265
},
{
"epoch": 0.05005645464809936,
"grad_norm": 1.473186731338501,
"learning_rate": 3.877222258755935e-05,
"loss": 3.1098,
"step": 266
},
{
"epoch": 0.05024463680843056,
"grad_norm": 1.5433677434921265,
"learning_rate": 3.8416553268574285e-05,
"loss": 3.4896,
"step": 267
},
{
"epoch": 0.05043281896876176,
"grad_norm": 1.4866740703582764,
"learning_rate": 3.80615022488229e-05,
"loss": 3.4388,
"step": 268
},
{
"epoch": 0.050621001129092964,
"grad_norm": 1.5249758958816528,
"learning_rate": 3.770708848015768e-05,
"loss": 3.3572,
"step": 269
},
{
"epoch": 0.050809183289424166,
"grad_norm": 1.5038138628005981,
"learning_rate": 3.735333088041596e-05,
"loss": 3.3581,
"step": 270
},
{
"epoch": 0.05099736544975536,
"grad_norm": 1.479463815689087,
"learning_rate": 3.7000248332410304e-05,
"loss": 3.2485,
"step": 271
},
{
"epoch": 0.05118554761008656,
"grad_norm": 1.5273246765136719,
"learning_rate": 3.664785968292036e-05,
"loss": 3.3607,
"step": 272
},
{
"epoch": 0.051373729770417764,
"grad_norm": 1.501825213432312,
"learning_rate": 3.629618374168711e-05,
"loss": 3.2803,
"step": 273
},
{
"epoch": 0.051561911930748966,
"grad_norm": 1.472866415977478,
"learning_rate": 3.594523928040859e-05,
"loss": 3.274,
"step": 274
},
{
"epoch": 0.05175009409108017,
"grad_norm": 1.490519404411316,
"learning_rate": 3.5595045031738125e-05,
"loss": 3.3221,
"step": 275
},
{
"epoch": 0.05193827625141137,
"grad_norm": 1.520186185836792,
"learning_rate": 3.5245619688284274e-05,
"loss": 3.4458,
"step": 276
},
{
"epoch": 0.052126458411742564,
"grad_norm": 1.5064996480941772,
"learning_rate": 3.4896981901613104e-05,
"loss": 3.3529,
"step": 277
},
{
"epoch": 0.052314640572073766,
"grad_norm": 1.519168734550476,
"learning_rate": 3.4549150281252636e-05,
"loss": 3.6177,
"step": 278
},
{
"epoch": 0.05250282273240497,
"grad_norm": 1.5575183629989624,
"learning_rate": 3.420214339369944e-05,
"loss": 3.4754,
"step": 279
},
{
"epoch": 0.05269100489273617,
"grad_norm": 1.5328855514526367,
"learning_rate": 3.38559797614277e-05,
"loss": 3.2606,
"step": 280
},
{
"epoch": 0.05287918705306737,
"grad_norm": 1.585930347442627,
"learning_rate": 3.351067786190038e-05,
"loss": 3.3014,
"step": 281
},
{
"epoch": 0.05306736921339857,
"grad_norm": 1.5054539442062378,
"learning_rate": 3.316625612658315e-05,
"loss": 3.3176,
"step": 282
},
{
"epoch": 0.05325555137372977,
"grad_norm": 1.5619922876358032,
"learning_rate": 3.282273293996027e-05,
"loss": 3.4408,
"step": 283
},
{
"epoch": 0.05344373353406097,
"grad_norm": 1.470646858215332,
"learning_rate": 3.248012663855353e-05,
"loss": 3.2584,
"step": 284
},
{
"epoch": 0.05363191569439217,
"grad_norm": 1.5209406614303589,
"learning_rate": 3.2138455509943366e-05,
"loss": 3.3224,
"step": 285
},
{
"epoch": 0.05382009785472337,
"grad_norm": 1.4932469129562378,
"learning_rate": 3.179773779179267e-05,
"loss": 3.3588,
"step": 286
},
{
"epoch": 0.054008280015054574,
"grad_norm": 1.4727731943130493,
"learning_rate": 3.145799167087342e-05,
"loss": 3.2095,
"step": 287
},
{
"epoch": 0.054196462175385776,
"grad_norm": 1.501749038696289,
"learning_rate": 3.111923528209577e-05,
"loss": 3.3637,
"step": 288
},
{
"epoch": 0.05438464433571697,
"grad_norm": 1.5096619129180908,
"learning_rate": 3.078148670754022e-05,
"loss": 3.3815,
"step": 289
},
{
"epoch": 0.05457282649604817,
"grad_norm": 1.500936508178711,
"learning_rate": 3.0444763975492208e-05,
"loss": 3.1945,
"step": 290
},
{
"epoch": 0.054761008656379374,
"grad_norm": 1.4544003009796143,
"learning_rate": 3.0109085059480017e-05,
"loss": 3.3647,
"step": 291
},
{
"epoch": 0.054949190816710576,
"grad_norm": 1.552875280380249,
"learning_rate": 2.977446787731532e-05,
"loss": 3.39,
"step": 292
},
{
"epoch": 0.05513737297704178,
"grad_norm": 1.5364927053451538,
"learning_rate": 2.944093029013664e-05,
"loss": 3.4203,
"step": 293
},
{
"epoch": 0.05532555513737298,
"grad_norm": 1.5016520023345947,
"learning_rate": 2.910849010145617e-05,
"loss": 3.3805,
"step": 294
},
{
"epoch": 0.055513737297704174,
"grad_norm": 1.5016636848449707,
"learning_rate": 2.8777165056209256e-05,
"loss": 3.1924,
"step": 295
},
{
"epoch": 0.055701919458035376,
"grad_norm": 1.5240119695663452,
"learning_rate": 2.8446972839807384e-05,
"loss": 3.5986,
"step": 296
},
{
"epoch": 0.05589010161836658,
"grad_norm": 1.5180970430374146,
"learning_rate": 2.8117931077194065e-05,
"loss": 3.3258,
"step": 297
},
{
"epoch": 0.05607828377869778,
"grad_norm": 1.465223789215088,
"learning_rate": 2.7790057331904117e-05,
"loss": 3.2024,
"step": 298
},
{
"epoch": 0.05626646593902898,
"grad_norm": 1.7016164064407349,
"learning_rate": 2.746336910512606e-05,
"loss": 3.3025,
"step": 299
},
{
"epoch": 0.05645464809936018,
"grad_norm": 2.5566675662994385,
"learning_rate": 2.7137883834768073e-05,
"loss": 3.1934,
"step": 300
},
{
"epoch": 0.05645464809936018,
"eval_loss": 0.8374229073524475,
"eval_runtime": 982.5965,
"eval_samples_per_second": 9.109,
"eval_steps_per_second": 2.278,
"step": 300
},
{
"epoch": 0.056642830259691385,
"grad_norm": 1.5962673425674438,
"learning_rate": 2.6813618894527138e-05,
"loss": 3.5087,
"step": 301
},
{
"epoch": 0.05683101242002258,
"grad_norm": 1.5255632400512695,
"learning_rate": 2.6490591592961578e-05,
"loss": 3.4712,
"step": 302
},
{
"epoch": 0.05701919458035378,
"grad_norm": 1.542389154434204,
"learning_rate": 2.6168819172567392e-05,
"loss": 3.2852,
"step": 303
},
{
"epoch": 0.05720737674068498,
"grad_norm": 1.664982557296753,
"learning_rate": 2.5848318808857606e-05,
"loss": 3.5038,
"step": 304
},
{
"epoch": 0.057395558901016185,
"grad_norm": 1.4651890993118286,
"learning_rate": 2.5529107609445733e-05,
"loss": 3.1837,
"step": 305
},
{
"epoch": 0.057583741061347386,
"grad_norm": 1.486304521560669,
"learning_rate": 2.521120261313241e-05,
"loss": 3.2311,
"step": 306
},
{
"epoch": 0.05777192322167859,
"grad_norm": 1.4881970882415771,
"learning_rate": 2.4894620788996037e-05,
"loss": 3.419,
"step": 307
},
{
"epoch": 0.05796010538200978,
"grad_norm": 1.4943639039993286,
"learning_rate": 2.457937903548695e-05,
"loss": 3.3085,
"step": 308
},
{
"epoch": 0.058148287542340985,
"grad_norm": 1.5298432111740112,
"learning_rate": 2.426549417952542e-05,
"loss": 3.608,
"step": 309
},
{
"epoch": 0.058336469702672186,
"grad_norm": 1.5149232149124146,
"learning_rate": 2.3952982975603496e-05,
"loss": 3.3398,
"step": 310
},
{
"epoch": 0.05852465186300339,
"grad_norm": 1.492951512336731,
"learning_rate": 2.3641862104890595e-05,
"loss": 3.3529,
"step": 311
},
{
"epoch": 0.05871283402333459,
"grad_norm": 1.448451280593872,
"learning_rate": 2.3332148174343254e-05,
"loss": 3.2021,
"step": 312
},
{
"epoch": 0.05890101618366579,
"grad_norm": 1.4983866214752197,
"learning_rate": 2.3023857715818532e-05,
"loss": 3.3579,
"step": 313
},
{
"epoch": 0.059089198343996986,
"grad_norm": 1.5230140686035156,
"learning_rate": 2.2717007185191674e-05,
"loss": 3.4931,
"step": 314
},
{
"epoch": 0.05927738050432819,
"grad_norm": 1.5284234285354614,
"learning_rate": 2.24116129614777e-05,
"loss": 3.4331,
"step": 315
},
{
"epoch": 0.05946556266465939,
"grad_norm": 1.4961637258529663,
"learning_rate": 2.2107691345957133e-05,
"loss": 3.3148,
"step": 316
},
{
"epoch": 0.05965374482499059,
"grad_norm": 1.547951340675354,
"learning_rate": 2.1805258561305862e-05,
"loss": 3.3194,
"step": 317
},
{
"epoch": 0.05984192698532179,
"grad_norm": 1.5121986865997314,
"learning_rate": 2.1504330750729186e-05,
"loss": 3.4504,
"step": 318
},
{
"epoch": 0.060030109145652995,
"grad_norm": 1.51499605178833,
"learning_rate": 2.120492397710022e-05,
"loss": 3.4349,
"step": 319
},
{
"epoch": 0.06021829130598419,
"grad_norm": 1.520011305809021,
"learning_rate": 2.090705422210237e-05,
"loss": 3.3619,
"step": 320
},
{
"epoch": 0.06040647346631539,
"grad_norm": 1.5512540340423584,
"learning_rate": 2.061073738537635e-05,
"loss": 3.38,
"step": 321
},
{
"epoch": 0.06059465562664659,
"grad_norm": 1.463521122932434,
"learning_rate": 2.0315989283671473e-05,
"loss": 3.1915,
"step": 322
},
{
"epoch": 0.060782837786977795,
"grad_norm": 1.5051774978637695,
"learning_rate": 2.0022825650001387e-05,
"loss": 3.4725,
"step": 323
},
{
"epoch": 0.060971019947308996,
"grad_norm": 1.4965063333511353,
"learning_rate": 1.9731262132804274e-05,
"loss": 3.3025,
"step": 324
},
{
"epoch": 0.0611592021076402,
"grad_norm": 1.4896841049194336,
"learning_rate": 1.9441314295107537e-05,
"loss": 3.4046,
"step": 325
},
{
"epoch": 0.06134738426797139,
"grad_norm": 1.4985949993133545,
"learning_rate": 1.9152997613697183e-05,
"loss": 3.4278,
"step": 326
},
{
"epoch": 0.061535566428302595,
"grad_norm": 1.5722819566726685,
"learning_rate": 1.8866327478291546e-05,
"loss": 3.4424,
"step": 327
},
{
"epoch": 0.061723748588633796,
"grad_norm": 1.457684874534607,
"learning_rate": 1.8581319190720035e-05,
"loss": 3.175,
"step": 328
},
{
"epoch": 0.061911930748965,
"grad_norm": 1.501202940940857,
"learning_rate": 1.8297987964106115e-05,
"loss": 3.2706,
"step": 329
},
{
"epoch": 0.0621001129092962,
"grad_norm": 1.4955599308013916,
"learning_rate": 1.801634892205545e-05,
"loss": 3.3183,
"step": 330
},
{
"epoch": 0.0622882950696274,
"grad_norm": 1.52448308467865,
"learning_rate": 1.7736417097848506e-05,
"loss": 3.3863,
"step": 331
},
{
"epoch": 0.0624764772299586,
"grad_norm": 1.5245345830917358,
"learning_rate": 1.7458207433638223e-05,
"loss": 3.4718,
"step": 332
},
{
"epoch": 0.0626646593902898,
"grad_norm": 1.5418884754180908,
"learning_rate": 1.718173477965236e-05,
"loss": 3.3681,
"step": 333
},
{
"epoch": 0.062852841550621,
"grad_norm": 1.487151861190796,
"learning_rate": 1.6907013893400837e-05,
"loss": 3.3089,
"step": 334
},
{
"epoch": 0.0630410237109522,
"grad_norm": 1.4714837074279785,
"learning_rate": 1.6634059438888033e-05,
"loss": 3.181,
"step": 335
},
{
"epoch": 0.0632292058712834,
"grad_norm": 1.5032023191452026,
"learning_rate": 1.636288598583e-05,
"loss": 3.3265,
"step": 336
},
{
"epoch": 0.0634173880316146,
"grad_norm": 1.5267105102539062,
"learning_rate": 1.6093508008876857e-05,
"loss": 3.1744,
"step": 337
},
{
"epoch": 0.0636055701919458,
"grad_norm": 1.517006516456604,
"learning_rate": 1.5825939886840037e-05,
"loss": 3.2575,
"step": 338
},
{
"epoch": 0.06379375235227701,
"grad_norm": 1.4955118894577026,
"learning_rate": 1.5560195901924894e-05,
"loss": 3.2745,
"step": 339
},
{
"epoch": 0.06398193451260821,
"grad_norm": 1.7153363227844238,
"learning_rate": 1.5296290238968303e-05,
"loss": 3.4614,
"step": 340
},
{
"epoch": 0.06417011667293941,
"grad_norm": 1.5209360122680664,
"learning_rate": 1.50342369846815e-05,
"loss": 3.3276,
"step": 341
},
{
"epoch": 0.0643582988332706,
"grad_norm": 1.5184190273284912,
"learning_rate": 1.4774050126898164e-05,
"loss": 3.3869,
"step": 342
},
{
"epoch": 0.0645464809936018,
"grad_norm": 1.5385315418243408,
"learning_rate": 1.451574355382776e-05,
"loss": 3.4677,
"step": 343
},
{
"epoch": 0.064734663153933,
"grad_norm": 1.4663329124450684,
"learning_rate": 1.425933105331429e-05,
"loss": 3.3069,
"step": 344
},
{
"epoch": 0.0649228453142642,
"grad_norm": 1.5196608304977417,
"learning_rate": 1.4004826312100216e-05,
"loss": 3.3985,
"step": 345
},
{
"epoch": 0.0651110274745954,
"grad_norm": 1.5887420177459717,
"learning_rate": 1.3752242915095992e-05,
"loss": 3.309,
"step": 346
},
{
"epoch": 0.06529920963492661,
"grad_norm": 1.5067335367202759,
"learning_rate": 1.3501594344654884e-05,
"loss": 3.186,
"step": 347
},
{
"epoch": 0.06548739179525781,
"grad_norm": 1.4992436170578003,
"learning_rate": 1.3252893979853304e-05,
"loss": 3.3428,
"step": 348
},
{
"epoch": 0.06567557395558901,
"grad_norm": 1.6664701700210571,
"learning_rate": 1.3006155095776707e-05,
"loss": 3.4061,
"step": 349
},
{
"epoch": 0.06586375611592021,
"grad_norm": 2.502802848815918,
"learning_rate": 1.2761390862810907e-05,
"loss": 3.0303,
"step": 350
},
{
"epoch": 0.06605193827625142,
"grad_norm": 1.500586748123169,
"learning_rate": 1.2518614345939212e-05,
"loss": 3.4372,
"step": 351
},
{
"epoch": 0.06624012043658262,
"grad_norm": 1.5125335454940796,
"learning_rate": 1.227783850404487e-05,
"loss": 3.3834,
"step": 352
},
{
"epoch": 0.06642830259691382,
"grad_norm": 1.5365911722183228,
"learning_rate": 1.2039076189219517e-05,
"loss": 3.4872,
"step": 353
},
{
"epoch": 0.066616484757245,
"grad_norm": 1.5221909284591675,
"learning_rate": 1.1802340146077045e-05,
"loss": 3.3993,
"step": 354
},
{
"epoch": 0.06680466691757621,
"grad_norm": 1.5232350826263428,
"learning_rate": 1.1567643011073392e-05,
"loss": 3.5355,
"step": 355
},
{
"epoch": 0.06699284907790741,
"grad_norm": 1.5335094928741455,
"learning_rate": 1.1334997311832002e-05,
"loss": 3.4007,
"step": 356
},
{
"epoch": 0.06718103123823861,
"grad_norm": 1.4804530143737793,
"learning_rate": 1.1104415466475087e-05,
"loss": 3.3991,
"step": 357
},
{
"epoch": 0.06736921339856981,
"grad_norm": 1.4774835109710693,
"learning_rate": 1.0875909782960886e-05,
"loss": 3.3081,
"step": 358
},
{
"epoch": 0.06755739555890101,
"grad_norm": 1.507690191268921,
"learning_rate": 1.0649492458426564e-05,
"loss": 3.5331,
"step": 359
},
{
"epoch": 0.06774557771923222,
"grad_norm": 1.4900294542312622,
"learning_rate": 1.0425175578537299e-05,
"loss": 3.4034,
"step": 360
},
{
"epoch": 0.06793375987956342,
"grad_norm": 1.536511778831482,
"learning_rate": 1.020297111684101e-05,
"loss": 3.3285,
"step": 361
},
{
"epoch": 0.06812194203989462,
"grad_norm": 1.4938386678695679,
"learning_rate": 9.98289093412938e-06,
"loss": 3.3401,
"step": 362
},
{
"epoch": 0.06831012420022582,
"grad_norm": 1.498859167098999,
"learning_rate": 9.764946777804646e-06,
"loss": 3.4705,
"step": 363
},
{
"epoch": 0.06849830636055702,
"grad_norm": 1.505251407623291,
"learning_rate": 9.549150281252633e-06,
"loss": 3.3769,
"step": 364
},
{
"epoch": 0.06868648852088823,
"grad_norm": 1.5352188348770142,
"learning_rate": 9.335512963221732e-06,
"loss": 3.4495,
"step": 365
},
{
"epoch": 0.06887467068121943,
"grad_norm": 1.5243737697601318,
"learning_rate": 9.124046227208082e-06,
"loss": 3.3756,
"step": 366
},
{
"epoch": 0.06906285284155061,
"grad_norm": 1.4866875410079956,
"learning_rate": 8.914761360846869e-06,
"loss": 3.2548,
"step": 367
},
{
"epoch": 0.06925103500188182,
"grad_norm": 1.70474374294281,
"learning_rate": 8.707669535309793e-06,
"loss": 3.3533,
"step": 368
},
{
"epoch": 0.06943921716221302,
"grad_norm": 1.4734491109848022,
"learning_rate": 8.502781804708826e-06,
"loss": 3.3556,
"step": 369
},
{
"epoch": 0.06962739932254422,
"grad_norm": 1.5178672075271606,
"learning_rate": 8.30010910550611e-06,
"loss": 3.4011,
"step": 370
},
{
"epoch": 0.06981558148287542,
"grad_norm": 1.5426009893417358,
"learning_rate": 8.09966225593024e-06,
"loss": 3.3262,
"step": 371
},
{
"epoch": 0.07000376364320662,
"grad_norm": 1.488108515739441,
"learning_rate": 7.901451955398792e-06,
"loss": 3.3038,
"step": 372
},
{
"epoch": 0.07019194580353783,
"grad_norm": 1.5539402961730957,
"learning_rate": 7.705488783947202e-06,
"loss": 3.4263,
"step": 373
},
{
"epoch": 0.07038012796386903,
"grad_norm": 1.5011879205703735,
"learning_rate": 7.511783201664052e-06,
"loss": 3.1954,
"step": 374
},
{
"epoch": 0.07056831012420023,
"grad_norm": 1.5092988014221191,
"learning_rate": 7.320345548132679e-06,
"loss": 3.328,
"step": 375
},
{
"epoch": 0.07075649228453143,
"grad_norm": 1.498982310295105,
"learning_rate": 7.131186041879357e-06,
"loss": 3.2402,
"step": 376
},
{
"epoch": 0.07094467444486263,
"grad_norm": 1.5231132507324219,
"learning_rate": 6.944314779827749e-06,
"loss": 3.4769,
"step": 377
},
{
"epoch": 0.07113285660519383,
"grad_norm": 1.5165536403656006,
"learning_rate": 6.759741736760061e-06,
"loss": 3.4978,
"step": 378
},
{
"epoch": 0.07132103876552502,
"grad_norm": 1.531844973564148,
"learning_rate": 6.577476764784546e-06,
"loss": 3.3847,
"step": 379
},
{
"epoch": 0.07150922092585622,
"grad_norm": 1.4768074750900269,
"learning_rate": 6.397529592809614e-06,
"loss": 3.1253,
"step": 380
},
{
"epoch": 0.07169740308618743,
"grad_norm": 1.5122772455215454,
"learning_rate": 6.219909826024589e-06,
"loss": 3.3401,
"step": 381
},
{
"epoch": 0.07188558524651863,
"grad_norm": 1.5160380601882935,
"learning_rate": 6.0446269453868945e-06,
"loss": 3.3382,
"step": 382
},
{
"epoch": 0.07207376740684983,
"grad_norm": 1.5226458311080933,
"learning_rate": 5.871690307116107e-06,
"loss": 3.431,
"step": 383
},
{
"epoch": 0.07226194956718103,
"grad_norm": 1.4903466701507568,
"learning_rate": 5.701109142194422e-06,
"loss": 3.2707,
"step": 384
},
{
"epoch": 0.07245013172751223,
"grad_norm": 1.5170961618423462,
"learning_rate": 5.532892555874059e-06,
"loss": 3.3027,
"step": 385
},
{
"epoch": 0.07263831388784343,
"grad_norm": 1.5108839273452759,
"learning_rate": 5.3670495271910925e-06,
"loss": 3.3152,
"step": 386
},
{
"epoch": 0.07282649604817464,
"grad_norm": 1.51710045337677,
"learning_rate": 5.203588908486279e-06,
"loss": 3.2884,
"step": 387
},
{
"epoch": 0.07301467820850584,
"grad_norm": 1.5793352127075195,
"learning_rate": 5.042519424932513e-06,
"loss": 3.5139,
"step": 388
},
{
"epoch": 0.07320286036883704,
"grad_norm": 1.481472373008728,
"learning_rate": 4.883849674069058e-06,
"loss": 3.2427,
"step": 389
},
{
"epoch": 0.07339104252916824,
"grad_norm": 1.5381510257720947,
"learning_rate": 4.727588125342669e-06,
"loss": 3.3485,
"step": 390
},
{
"epoch": 0.07357922468949944,
"grad_norm": 1.508355975151062,
"learning_rate": 4.573743119655516e-06,
"loss": 3.295,
"step": 391
},
{
"epoch": 0.07376740684983063,
"grad_norm": 1.527873158454895,
"learning_rate": 4.422322868919937e-06,
"loss": 3.2918,
"step": 392
},
{
"epoch": 0.07395558901016183,
"grad_norm": 1.5130548477172852,
"learning_rate": 4.273335455620097e-06,
"loss": 3.3859,
"step": 393
},
{
"epoch": 0.07414377117049303,
"grad_norm": 1.5391393899917603,
"learning_rate": 4.126788832380629e-06,
"loss": 3.3711,
"step": 394
},
{
"epoch": 0.07433195333082424,
"grad_norm": 1.5213444232940674,
"learning_rate": 3.982690821542035e-06,
"loss": 3.4224,
"step": 395
},
{
"epoch": 0.07452013549115544,
"grad_norm": 1.4749212265014648,
"learning_rate": 3.8410491147432395e-06,
"loss": 3.1882,
"step": 396
},
{
"epoch": 0.07470831765148664,
"grad_norm": 1.4860862493515015,
"learning_rate": 3.7018712725109926e-06,
"loss": 3.2488,
"step": 397
},
{
"epoch": 0.07489649981181784,
"grad_norm": 1.5500200986862183,
"learning_rate": 3.5651647238562904e-06,
"loss": 3.3369,
"step": 398
},
{
"epoch": 0.07508468197214904,
"grad_norm": 1.8019529581069946,
"learning_rate": 3.430936765877857e-06,
"loss": 3.1557,
"step": 399
},
{
"epoch": 0.07527286413248024,
"grad_norm": 2.5598955154418945,
"learning_rate": 3.299194563372604e-06,
"loss": 3.03,
"step": 400
},
{
"epoch": 0.07546104629281145,
"grad_norm": 1.5204044580459595,
"learning_rate": 3.1699451484532463e-06,
"loss": 3.4419,
"step": 401
},
{
"epoch": 0.07564922845314265,
"grad_norm": 1.4903305768966675,
"learning_rate": 3.0431954201728784e-06,
"loss": 3.3295,
"step": 402
},
{
"epoch": 0.07583741061347385,
"grad_norm": 1.5254848003387451,
"learning_rate": 2.9189521441567726e-06,
"loss": 3.4107,
"step": 403
},
{
"epoch": 0.07602559277380504,
"grad_norm": 1.5176067352294922,
"learning_rate": 2.797221952241219e-06,
"loss": 3.3387,
"step": 404
},
{
"epoch": 0.07621377493413624,
"grad_norm": 1.495973825454712,
"learning_rate": 2.6780113421195298e-06,
"loss": 3.3161,
"step": 405
},
{
"epoch": 0.07640195709446744,
"grad_norm": 1.5124212503433228,
"learning_rate": 2.561326676995218e-06,
"loss": 3.3538,
"step": 406
},
{
"epoch": 0.07659013925479864,
"grad_norm": 1.4950138330459595,
"learning_rate": 2.4471741852423237e-06,
"loss": 3.363,
"step": 407
},
{
"epoch": 0.07677832141512984,
"grad_norm": 1.5359971523284912,
"learning_rate": 2.3355599600729915e-06,
"loss": 3.4671,
"step": 408
},
{
"epoch": 0.07696650357546105,
"grad_norm": 1.4483258724212646,
"learning_rate": 2.2264899592121744e-06,
"loss": 3.1561,
"step": 409
},
{
"epoch": 0.07715468573579225,
"grad_norm": 1.483723521232605,
"learning_rate": 2.1199700045797077e-06,
"loss": 3.3924,
"step": 410
},
{
"epoch": 0.07734286789612345,
"grad_norm": 1.4614593982696533,
"learning_rate": 2.0160057819794466e-06,
"loss": 3.2969,
"step": 411
},
{
"epoch": 0.07753105005645465,
"grad_norm": 1.4894236326217651,
"learning_rate": 1.9146028407958484e-06,
"loss": 3.3376,
"step": 412
},
{
"epoch": 0.07771923221678585,
"grad_norm": 1.501068115234375,
"learning_rate": 1.8157665936977263e-06,
"loss": 3.3462,
"step": 413
},
{
"epoch": 0.07790741437711705,
"grad_norm": 1.4664418697357178,
"learning_rate": 1.7195023163493252e-06,
"loss": 3.1851,
"step": 414
},
{
"epoch": 0.07809559653744826,
"grad_norm": 1.5121281147003174,
"learning_rate": 1.6258151471287396e-06,
"loss": 3.3562,
"step": 415
},
{
"epoch": 0.07828377869777944,
"grad_norm": 1.5005747079849243,
"learning_rate": 1.5347100868536246e-06,
"loss": 3.276,
"step": 416
},
{
"epoch": 0.07847196085811065,
"grad_norm": 1.5016510486602783,
"learning_rate": 1.4461919985142735e-06,
"loss": 3.1685,
"step": 417
},
{
"epoch": 0.07866014301844185,
"grad_norm": 1.490427017211914,
"learning_rate": 1.3602656070140275e-06,
"loss": 3.3398,
"step": 418
},
{
"epoch": 0.07884832517877305,
"grad_norm": 1.5081381797790527,
"learning_rate": 1.27693549891707e-06,
"loss": 3.378,
"step": 419
},
{
"epoch": 0.07903650733910425,
"grad_norm": 1.5135114192962646,
"learning_rate": 1.196206122203647e-06,
"loss": 3.2584,
"step": 420
},
{
"epoch": 0.07922468949943545,
"grad_norm": 1.503239631652832,
"learning_rate": 1.1180817860325599e-06,
"loss": 3.3742,
"step": 421
},
{
"epoch": 0.07941287165976665,
"grad_norm": 1.5126641988754272,
"learning_rate": 1.0425666605112517e-06,
"loss": 3.4263,
"step": 422
},
{
"epoch": 0.07960105382009786,
"grad_norm": 1.5308884382247925,
"learning_rate": 9.696647764731337e-07,
"loss": 3.3315,
"step": 423
},
{
"epoch": 0.07978923598042906,
"grad_norm": 1.5186878442764282,
"learning_rate": 8.993800252624862e-07,
"loss": 3.3336,
"step": 424
},
{
"epoch": 0.07997741814076026,
"grad_norm": 1.5928887128829956,
"learning_rate": 8.317161585266964e-07,
"loss": 3.641,
"step": 425
},
{
"epoch": 0.08016560030109146,
"grad_norm": 1.5158437490463257,
"learning_rate": 7.666767880160464e-07,
"loss": 3.4028,
"step": 426
},
{
"epoch": 0.08035378246142266,
"grad_norm": 1.5123188495635986,
"learning_rate": 7.042653853909064e-07,
"loss": 3.4125,
"step": 427
},
{
"epoch": 0.08054196462175386,
"grad_norm": 1.4982678890228271,
"learning_rate": 6.444852820364222e-07,
"loss": 3.2116,
"step": 428
},
{
"epoch": 0.08073014678208505,
"grad_norm": 1.5639753341674805,
"learning_rate": 5.87339668884701e-07,
"loss": 3.1289,
"step": 429
},
{
"epoch": 0.08091832894241625,
"grad_norm": 1.572089672088623,
"learning_rate": 5.328315962444874e-07,
"loss": 3.3213,
"step": 430
},
{
"epoch": 0.08110651110274746,
"grad_norm": 1.5132265090942383,
"learning_rate": 4.809639736383431e-07,
"loss": 3.3827,
"step": 431
},
{
"epoch": 0.08129469326307866,
"grad_norm": 1.4733517169952393,
"learning_rate": 4.317395696473214e-07,
"loss": 3.2562,
"step": 432
},
{
"epoch": 0.08148287542340986,
"grad_norm": 1.548933982849121,
"learning_rate": 3.851610117632354e-07,
"loss": 3.4471,
"step": 433
},
{
"epoch": 0.08167105758374106,
"grad_norm": 1.4967180490493774,
"learning_rate": 3.4123078624834216e-07,
"loss": 3.0655,
"step": 434
},
{
"epoch": 0.08185923974407226,
"grad_norm": 1.6242047548294067,
"learning_rate": 2.9995123800270476e-07,
"loss": 3.2837,
"step": 435
},
{
"epoch": 0.08204742190440346,
"grad_norm": 1.4935169219970703,
"learning_rate": 2.613245704389644e-07,
"loss": 3.3263,
"step": 436
},
{
"epoch": 0.08223560406473467,
"grad_norm": 1.4887654781341553,
"learning_rate": 2.2535284536476242e-07,
"loss": 3.1296,
"step": 437
},
{
"epoch": 0.08242378622506587,
"grad_norm": 1.511015772819519,
"learning_rate": 1.920379828726726e-07,
"loss": 3.3212,
"step": 438
},
{
"epoch": 0.08261196838539707,
"grad_norm": 1.5536808967590332,
"learning_rate": 1.6138176123770554e-07,
"loss": 3.5203,
"step": 439
},
{
"epoch": 0.08280015054572827,
"grad_norm": 1.5585739612579346,
"learning_rate": 1.333858168224178e-07,
"loss": 3.2727,
"step": 440
},
{
"epoch": 0.08298833270605946,
"grad_norm": 1.5475882291793823,
"learning_rate": 1.0805164398952072e-07,
"loss": 3.4168,
"step": 441
},
{
"epoch": 0.08317651486639066,
"grad_norm": 1.5009870529174805,
"learning_rate": 8.53805950221498e-08,
"loss": 3.3267,
"step": 442
},
{
"epoch": 0.08336469702672186,
"grad_norm": 1.5358216762542725,
"learning_rate": 6.537388005167233e-08,
"loss": 3.4914,
"step": 443
},
{
"epoch": 0.08355287918705306,
"grad_norm": 1.5157291889190674,
"learning_rate": 4.8032566993089225e-08,
"loss": 3.3647,
"step": 444
},
{
"epoch": 0.08374106134738427,
"grad_norm": 1.5044530630111694,
"learning_rate": 3.3357581488030475e-08,
"loss": 3.2696,
"step": 445
},
{
"epoch": 0.08392924350771547,
"grad_norm": 1.5259320735931396,
"learning_rate": 2.134970685536697e-08,
"loss": 3.3305,
"step": 446
},
{
"epoch": 0.08411742566804667,
"grad_norm": 1.4801864624023438,
"learning_rate": 1.200958404936059e-08,
"loss": 3.1884,
"step": 447
},
{
"epoch": 0.08430560782837787,
"grad_norm": 1.4972193241119385,
"learning_rate": 5.337711625497121e-09,
"loss": 3.3929,
"step": 448
},
{
"epoch": 0.08449378998870907,
"grad_norm": 1.7666032314300537,
"learning_rate": 1.3344457138297906e-09,
"loss": 3.4298,
"step": 449
},
{
"epoch": 0.08468197214904027,
"grad_norm": 2.5446274280548096,
"learning_rate": 0.0,
"loss": 2.8081,
"step": 450
},
{
"epoch": 0.08468197214904027,
"eval_loss": 0.8302884697914124,
"eval_runtime": 982.1123,
"eval_samples_per_second": 9.113,
"eval_steps_per_second": 2.279,
"step": 450
}
],
"logging_steps": 1,
"max_steps": 450,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 150,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 2,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.0272928176694886e+18,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}