{
  "best_metric": 1.1517876386642456,
  "best_model_checkpoint": "/media/user/Expansion/flan-t5-small-simplifier/checkpoint-80000",
  "epoch": 2.0,
  "eval_steps": 10000,
  "global_step": 89920,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.01112099644128114,
      "grad_norm": 3.2175512313842773,
      "learning_rate": 4.972197508896798e-05,
      "loss": 1.6509,
      "num_input_tokens_seen": 183488,
      "step": 500
    },
    {
      "epoch": 0.02224199288256228,
      "grad_norm": 3.9029898643493652,
      "learning_rate": 4.9443950177935946e-05,
      "loss": 1.583,
      "num_input_tokens_seen": 363744,
      "step": 1000
    },
    {
      "epoch": 0.03336298932384341,
      "grad_norm": 3.4091885089874268,
      "learning_rate": 4.9165925266903915e-05,
      "loss": 1.5553,
      "num_input_tokens_seen": 547584,
      "step": 1500
    },
    {
      "epoch": 0.04448398576512456,
      "grad_norm": 2.5965495109558105,
      "learning_rate": 4.888790035587189e-05,
      "loss": 1.5444,
      "num_input_tokens_seen": 732480,
      "step": 2000
    },
    {
      "epoch": 0.055604982206405695,
      "grad_norm": 4.170108318328857,
      "learning_rate": 4.860987544483986e-05,
      "loss": 1.5137,
      "num_input_tokens_seen": 916672,
      "step": 2500
    },
    {
      "epoch": 0.06672597864768683,
      "grad_norm": 3.025068759918213,
      "learning_rate": 4.8331850533807835e-05,
      "loss": 1.5057,
      "num_input_tokens_seen": 1095952,
      "step": 3000
    },
    {
      "epoch": 0.07784697508896797,
      "grad_norm": 2.7047712802886963,
      "learning_rate": 4.80538256227758e-05,
      "loss": 1.5019,
      "num_input_tokens_seen": 1276936,
      "step": 3500
    },
    {
      "epoch": 0.08896797153024912,
      "grad_norm": 2.844285488128662,
      "learning_rate": 4.777580071174377e-05,
      "loss": 1.5074,
      "num_input_tokens_seen": 1464080,
      "step": 4000
    },
    {
      "epoch": 0.10008896797153025,
      "grad_norm": 3.055643081665039,
      "learning_rate": 4.749777580071175e-05,
      "loss": 1.4959,
      "num_input_tokens_seen": 1647288,
      "step": 4500
    },
    {
      "epoch": 0.11120996441281139,
      "grad_norm": 2.6245856285095215,
      "learning_rate": 4.721975088967972e-05,
      "loss": 1.4777,
      "num_input_tokens_seen": 1836832,
      "step": 5000
    },
    {
      "epoch": 0.12233096085409252,
      "grad_norm": 3.1758244037628174,
      "learning_rate": 4.694172597864769e-05,
      "loss": 1.4778,
      "num_input_tokens_seen": 2020728,
      "step": 5500
    },
    {
      "epoch": 0.13345195729537365,
      "grad_norm": 2.518728494644165,
      "learning_rate": 4.666370106761566e-05,
      "loss": 1.4758,
      "num_input_tokens_seen": 2198968,
      "step": 6000
    },
    {
      "epoch": 0.1445729537366548,
      "grad_norm": 3.8143937587738037,
      "learning_rate": 4.638567615658363e-05,
      "loss": 1.4846,
      "num_input_tokens_seen": 2381968,
      "step": 6500
    },
    {
      "epoch": 0.15569395017793594,
      "grad_norm": 2.766146421432495,
      "learning_rate": 4.6107651245551604e-05,
      "loss": 1.4468,
      "num_input_tokens_seen": 2562176,
      "step": 7000
    },
    {
      "epoch": 0.16681494661921709,
      "grad_norm": 3.891373634338379,
      "learning_rate": 4.582962633451958e-05,
      "loss": 1.4586,
      "num_input_tokens_seen": 2744688,
      "step": 7500
    },
    {
      "epoch": 0.17793594306049823,
      "grad_norm": 3.277316093444824,
      "learning_rate": 4.555160142348754e-05,
      "loss": 1.4606,
      "num_input_tokens_seen": 2926000,
      "step": 8000
    },
    {
      "epoch": 0.18905693950177935,
      "grad_norm": 3.242478132247925,
      "learning_rate": 4.5273576512455517e-05,
      "loss": 1.446,
      "num_input_tokens_seen": 3108520,
      "step": 8500
    },
    {
      "epoch": 0.2001779359430605,
      "grad_norm": 2.3061795234680176,
      "learning_rate": 4.499555160142349e-05,
      "loss": 1.4348,
      "num_input_tokens_seen": 3289352,
      "step": 9000
    },
    {
      "epoch": 0.21129893238434164,
      "grad_norm": 3.4106180667877197,
      "learning_rate": 4.471752669039146e-05,
      "loss": 1.4345,
      "num_input_tokens_seen": 3473784,
      "step": 9500
    },
    {
      "epoch": 0.22241992882562278,
      "grad_norm": 2.88779354095459,
      "learning_rate": 4.443950177935943e-05,
      "loss": 1.4423,
      "num_input_tokens_seen": 3655312,
      "step": 10000
    },
    {
      "epoch": 0.22241992882562278,
      "eval_loss": 1.2430843114852905,
      "eval_runtime": 2.8218,
      "eval_samples_per_second": 885.962,
      "eval_steps_per_second": 110.922,
      "num_input_tokens_seen": 3655312,
      "step": 10000
    },
    {
      "epoch": 0.23354092526690393,
      "grad_norm": 2.829268217086792,
      "learning_rate": 4.4161476868327405e-05,
      "loss": 1.4321,
      "num_input_tokens_seen": 3847328,
      "step": 10500
    },
    {
      "epoch": 0.24466192170818504,
      "grad_norm": 3.807185173034668,
      "learning_rate": 4.388345195729537e-05,
      "loss": 1.4487,
      "num_input_tokens_seen": 4025576,
      "step": 11000
    },
    {
      "epoch": 0.2557829181494662,
      "grad_norm": 2.449007511138916,
      "learning_rate": 4.360542704626335e-05,
      "loss": 1.413,
      "num_input_tokens_seen": 4213832,
      "step": 11500
    },
    {
      "epoch": 0.2669039145907473,
      "grad_norm": 3.363555431365967,
      "learning_rate": 4.3327402135231324e-05,
      "loss": 1.4313,
      "num_input_tokens_seen": 4395352,
      "step": 12000
    },
    {
      "epoch": 0.2780249110320285,
      "grad_norm": 2.702747344970703,
      "learning_rate": 4.3049377224199286e-05,
      "loss": 1.4179,
      "num_input_tokens_seen": 4582368,
      "step": 12500
    },
    {
      "epoch": 0.2891459074733096,
      "grad_norm": 3.286044120788574,
      "learning_rate": 4.277135231316726e-05,
      "loss": 1.4137,
      "num_input_tokens_seen": 4770640,
      "step": 13000
    },
    {
      "epoch": 0.30026690391459077,
      "grad_norm": 2.6410391330718994,
      "learning_rate": 4.249332740213524e-05,
      "loss": 1.4303,
      "num_input_tokens_seen": 4951280,
      "step": 13500
    },
    {
      "epoch": 0.3113879003558719,
      "grad_norm": 3.239133358001709,
      "learning_rate": 4.2215302491103205e-05,
      "loss": 1.4185,
      "num_input_tokens_seen": 5133552,
      "step": 14000
    },
    {
      "epoch": 0.322508896797153,
      "grad_norm": 2.943094253540039,
      "learning_rate": 4.1937277580071174e-05,
      "loss": 1.4065,
      "num_input_tokens_seen": 5319992,
      "step": 14500
    },
    {
      "epoch": 0.33362989323843417,
      "grad_norm": 2.180136203765869,
      "learning_rate": 4.165925266903915e-05,
      "loss": 1.4096,
      "num_input_tokens_seen": 5501592,
      "step": 15000
    },
    {
      "epoch": 0.3447508896797153,
      "grad_norm": 2.4302403926849365,
      "learning_rate": 4.138122775800712e-05,
      "loss": 1.4203,
      "num_input_tokens_seen": 5689312,
      "step": 15500
    },
    {
      "epoch": 0.35587188612099646,
      "grad_norm": 2.850964069366455,
      "learning_rate": 4.1103202846975093e-05,
      "loss": 1.4291,
      "num_input_tokens_seen": 5870384,
      "step": 16000
    },
    {
      "epoch": 0.3669928825622776,
      "grad_norm": 1.9641114473342896,
      "learning_rate": 4.082517793594306e-05,
      "loss": 1.4054,
      "num_input_tokens_seen": 6048624,
      "step": 16500
    },
    {
      "epoch": 0.3781138790035587,
      "grad_norm": 2.287353754043579,
      "learning_rate": 4.054715302491103e-05,
      "loss": 1.4118,
      "num_input_tokens_seen": 6229728,
      "step": 17000
    },
    {
      "epoch": 0.38923487544483987,
      "grad_norm": 3.4425182342529297,
      "learning_rate": 4.0269128113879006e-05,
      "loss": 1.4061,
      "num_input_tokens_seen": 6410968,
      "step": 17500
    },
    {
      "epoch": 0.400355871886121,
      "grad_norm": 2.0604770183563232,
      "learning_rate": 3.9991103202846975e-05,
      "loss": 1.3871,
      "num_input_tokens_seen": 6591312,
      "step": 18000
    },
    {
      "epoch": 0.41147686832740216,
      "grad_norm": 2.671599864959717,
      "learning_rate": 3.971307829181495e-05,
      "loss": 1.42,
      "num_input_tokens_seen": 6777912,
      "step": 18500
    },
    {
      "epoch": 0.4225978647686833,
      "grad_norm": 2.176579475402832,
      "learning_rate": 3.943505338078292e-05,
      "loss": 1.417,
      "num_input_tokens_seen": 6964240,
      "step": 19000
    },
    {
      "epoch": 0.4337188612099644,
      "grad_norm": 3.2378785610198975,
      "learning_rate": 3.915702846975089e-05,
      "loss": 1.3992,
      "num_input_tokens_seen": 7147288,
      "step": 19500
    },
    {
      "epoch": 0.44483985765124556,
      "grad_norm": 3.0986084938049316,
      "learning_rate": 3.887900355871886e-05,
      "loss": 1.3884,
      "num_input_tokens_seen": 7331520,
      "step": 20000
    },
    {
      "epoch": 0.44483985765124556,
      "eval_loss": 1.2093411684036255,
      "eval_runtime": 2.7984,
      "eval_samples_per_second": 893.355,
      "eval_steps_per_second": 111.848,
      "num_input_tokens_seen": 7331520,
      "step": 20000
    },
    {
      "epoch": 0.4559608540925267,
      "grad_norm": 3.4998202323913574,
      "learning_rate": 3.860097864768684e-05,
      "loss": 1.3915,
      "num_input_tokens_seen": 7515512,
      "step": 20500
    },
    {
      "epoch": 0.46708185053380785,
      "grad_norm": 3.0249533653259277,
      "learning_rate": 3.832295373665481e-05,
      "loss": 1.3967,
      "num_input_tokens_seen": 7696752,
      "step": 21000
    },
    {
      "epoch": 0.47820284697508897,
      "grad_norm": 6.8868513107299805,
      "learning_rate": 3.8044928825622775e-05,
      "loss": 1.4106,
      "num_input_tokens_seen": 7878832,
      "step": 21500
    },
    {
      "epoch": 0.4893238434163701,
      "grad_norm": 2.2134385108947754,
      "learning_rate": 3.776690391459075e-05,
      "loss": 1.3847,
      "num_input_tokens_seen": 8059592,
      "step": 22000
    },
    {
      "epoch": 0.5004448398576512,
      "grad_norm": 2.2698676586151123,
      "learning_rate": 3.748887900355872e-05,
      "loss": 1.3941,
      "num_input_tokens_seen": 8245432,
      "step": 22500
    },
    {
      "epoch": 0.5115658362989324,
      "grad_norm": 2.4593448638916016,
      "learning_rate": 3.7210854092526695e-05,
      "loss": 1.3716,
      "num_input_tokens_seen": 8429400,
      "step": 23000
    },
    {
      "epoch": 0.5226868327402135,
      "grad_norm": 2.5121207237243652,
      "learning_rate": 3.6932829181494664e-05,
      "loss": 1.3733,
      "num_input_tokens_seen": 8618216,
      "step": 23500
    },
    {
      "epoch": 0.5338078291814946,
      "grad_norm": 2.2493703365325928,
      "learning_rate": 3.665480427046263e-05,
      "loss": 1.3754,
      "num_input_tokens_seen": 8801640,
      "step": 24000
    },
    {
      "epoch": 0.5449288256227758,
      "grad_norm": 3.08921217918396,
      "learning_rate": 3.637677935943061e-05,
      "loss": 1.3694,
      "num_input_tokens_seen": 8977800,
      "step": 24500
    },
    {
      "epoch": 0.556049822064057,
      "grad_norm": 2.215222120285034,
      "learning_rate": 3.609875444839858e-05,
      "loss": 1.3674,
      "num_input_tokens_seen": 9160904,
      "step": 25000
    },
    {
      "epoch": 0.5671708185053381,
      "grad_norm": 2.553903818130493,
      "learning_rate": 3.582072953736655e-05,
      "loss": 1.3735,
      "num_input_tokens_seen": 9348344,
      "step": 25500
    },
    {
      "epoch": 0.5782918149466192,
      "grad_norm": 2.546022891998291,
      "learning_rate": 3.554270462633452e-05,
      "loss": 1.382,
      "num_input_tokens_seen": 9532664,
      "step": 26000
    },
    {
      "epoch": 0.5894128113879004,
      "grad_norm": 2.917534112930298,
      "learning_rate": 3.5264679715302496e-05,
      "loss": 1.3654,
      "num_input_tokens_seen": 9717800,
      "step": 26500
    },
    {
      "epoch": 0.6005338078291815,
      "grad_norm": 3.355299472808838,
      "learning_rate": 3.4986654804270464e-05,
      "loss": 1.3876,
      "num_input_tokens_seen": 9902536,
      "step": 27000
    },
    {
      "epoch": 0.6116548042704626,
      "grad_norm": 2.67924427986145,
      "learning_rate": 3.470862989323844e-05,
      "loss": 1.3575,
      "num_input_tokens_seen": 10082216,
      "step": 27500
    },
    {
      "epoch": 0.6227758007117438,
      "grad_norm": 3.040212392807007,
      "learning_rate": 3.44306049822064e-05,
      "loss": 1.3644,
      "num_input_tokens_seen": 10263992,
      "step": 28000
    },
    {
      "epoch": 0.6338967971530249,
      "grad_norm": 3.726254940032959,
      "learning_rate": 3.415258007117438e-05,
      "loss": 1.3616,
      "num_input_tokens_seen": 10444128,
      "step": 28500
    },
    {
      "epoch": 0.645017793594306,
      "grad_norm": 4.716592788696289,
      "learning_rate": 3.387455516014235e-05,
      "loss": 1.3662,
      "num_input_tokens_seen": 10626928,
      "step": 29000
    },
    {
      "epoch": 0.6561387900355872,
      "grad_norm": 2.9317166805267334,
      "learning_rate": 3.359653024911032e-05,
      "loss": 1.38,
      "num_input_tokens_seen": 10808960,
      "step": 29500
    },
    {
      "epoch": 0.6672597864768683,
      "grad_norm": 2.411684989929199,
      "learning_rate": 3.331850533807829e-05,
      "loss": 1.3782,
      "num_input_tokens_seen": 10990432,
      "step": 30000
    },
    {
      "epoch": 0.6672597864768683,
      "eval_loss": 1.185857892036438,
      "eval_runtime": 2.9618,
      "eval_samples_per_second": 844.079,
      "eval_steps_per_second": 105.679,
      "num_input_tokens_seen": 10990432,
      "step": 30000
    },
    {
      "epoch": 0.6783807829181495,
      "grad_norm": 2.765089273452759,
      "learning_rate": 3.3040480427046265e-05,
      "loss": 1.3808,
      "num_input_tokens_seen": 11172032,
      "step": 30500
    },
    {
      "epoch": 0.6895017793594306,
      "grad_norm": 2.808806896209717,
      "learning_rate": 3.2762455516014234e-05,
      "loss": 1.3925,
      "num_input_tokens_seen": 11356608,
      "step": 31000
    },
    {
      "epoch": 0.7006227758007118,
      "grad_norm": 2.6583220958709717,
      "learning_rate": 3.248443060498221e-05,
      "loss": 1.3716,
      "num_input_tokens_seen": 11538056,
      "step": 31500
    },
    {
      "epoch": 0.7117437722419929,
      "grad_norm": 2.2725088596343994,
      "learning_rate": 3.2206405693950184e-05,
      "loss": 1.3689,
      "num_input_tokens_seen": 11721264,
      "step": 32000
    },
    {
      "epoch": 0.722864768683274,
      "grad_norm": 2.927656412124634,
      "learning_rate": 3.1928380782918146e-05,
      "loss": 1.3722,
      "num_input_tokens_seen": 11896688,
      "step": 32500
    },
    {
      "epoch": 0.7339857651245552,
      "grad_norm": 2.0601186752319336,
      "learning_rate": 3.165035587188612e-05,
      "loss": 1.3408,
      "num_input_tokens_seen": 12084440,
      "step": 33000
    },
    {
      "epoch": 0.7451067615658363,
      "grad_norm": 2.5769150257110596,
      "learning_rate": 3.13723309608541e-05,
      "loss": 1.3874,
      "num_input_tokens_seen": 12264224,
      "step": 33500
    },
    {
      "epoch": 0.7562277580071174,
      "grad_norm": 2.845653772354126,
      "learning_rate": 3.1094306049822066e-05,
      "loss": 1.3755,
      "num_input_tokens_seen": 12446200,
      "step": 34000
    },
    {
      "epoch": 0.7673487544483986,
      "grad_norm": 2.3848676681518555,
      "learning_rate": 3.0816281138790034e-05,
      "loss": 1.3463,
      "num_input_tokens_seen": 12628992,
      "step": 34500
    },
    {
      "epoch": 0.7784697508896797,
      "grad_norm": 3.2360849380493164,
      "learning_rate": 3.053825622775801e-05,
      "loss": 1.3678,
      "num_input_tokens_seen": 12809808,
      "step": 35000
    },
    {
      "epoch": 0.7895907473309609,
      "grad_norm": 2.3211023807525635,
      "learning_rate": 3.026023131672598e-05,
      "loss": 1.3732,
      "num_input_tokens_seen": 12989912,
      "step": 35500
    },
    {
      "epoch": 0.800711743772242,
      "grad_norm": 3.599958658218384,
      "learning_rate": 2.9982206405693954e-05,
      "loss": 1.3606,
      "num_input_tokens_seen": 13170560,
      "step": 36000
    },
    {
      "epoch": 0.8118327402135231,
      "grad_norm": 2.0861263275146484,
      "learning_rate": 2.9704181494661926e-05,
      "loss": 1.3475,
      "num_input_tokens_seen": 13350424,
      "step": 36500
    },
    {
      "epoch": 0.8229537366548043,
      "grad_norm": 2.043938159942627,
      "learning_rate": 2.9426156583629895e-05,
      "loss": 1.3753,
      "num_input_tokens_seen": 13538640,
      "step": 37000
    },
    {
      "epoch": 0.8340747330960854,
      "grad_norm": 2.4880995750427246,
      "learning_rate": 2.9148131672597867e-05,
      "loss": 1.3497,
      "num_input_tokens_seen": 13730096,
      "step": 37500
    },
    {
      "epoch": 0.8451957295373665,
      "grad_norm": 2.535860300064087,
      "learning_rate": 2.8870106761565835e-05,
      "loss": 1.3708,
      "num_input_tokens_seen": 13909256,
      "step": 38000
    },
    {
      "epoch": 0.8563167259786477,
      "grad_norm": 2.499455213546753,
      "learning_rate": 2.8592081850533807e-05,
      "loss": 1.3546,
      "num_input_tokens_seen": 14101536,
      "step": 38500
    },
    {
      "epoch": 0.8674377224199288,
      "grad_norm": 2.3696117401123047,
      "learning_rate": 2.8314056939501783e-05,
      "loss": 1.3593,
      "num_input_tokens_seen": 14281392,
      "step": 39000
    },
    {
      "epoch": 0.87855871886121,
      "grad_norm": 3.260430097579956,
      "learning_rate": 2.803603202846975e-05,
      "loss": 1.3506,
      "num_input_tokens_seen": 14466320,
      "step": 39500
    },
    {
      "epoch": 0.8896797153024911,
      "grad_norm": 4.11997652053833,
      "learning_rate": 2.7758007117437723e-05,
      "loss": 1.3595,
      "num_input_tokens_seen": 14653328,
      "step": 40000
    },
    {
      "epoch": 0.8896797153024911,
      "eval_loss": 1.1787019968032837,
      "eval_runtime": 2.9698,
      "eval_samples_per_second": 841.809,
      "eval_steps_per_second": 105.394,
      "num_input_tokens_seen": 14653328,
      "step": 40000
    },
    {
      "epoch": 0.9008007117437722,
      "grad_norm": 1.932468056678772,
      "learning_rate": 2.7479982206405695e-05,
      "loss": 1.3559,
      "num_input_tokens_seen": 14837624,
      "step": 40500
    },
    {
      "epoch": 0.9119217081850534,
      "grad_norm": 2.6026477813720703,
      "learning_rate": 2.7201957295373664e-05,
      "loss": 1.3454,
      "num_input_tokens_seen": 15020192,
      "step": 41000
    },
    {
      "epoch": 0.9230427046263345,
      "grad_norm": 2.3455870151519775,
      "learning_rate": 2.692393238434164e-05,
      "loss": 1.3563,
      "num_input_tokens_seen": 15204288,
      "step": 41500
    },
    {
      "epoch": 0.9341637010676157,
      "grad_norm": 2.8757784366607666,
      "learning_rate": 2.664590747330961e-05,
      "loss": 1.3283,
      "num_input_tokens_seen": 15393744,
      "step": 42000
    },
    {
      "epoch": 0.9452846975088968,
      "grad_norm": 2.3972697257995605,
      "learning_rate": 2.636788256227758e-05,
      "loss": 1.3612,
      "num_input_tokens_seen": 15569824,
      "step": 42500
    },
    {
      "epoch": 0.9564056939501779,
      "grad_norm": 3.187290906906128,
      "learning_rate": 2.6089857651245552e-05,
      "loss": 1.3525,
      "num_input_tokens_seen": 15753768,
      "step": 43000
    },
    {
      "epoch": 0.9675266903914591,
      "grad_norm": 2.447659969329834,
      "learning_rate": 2.5811832740213527e-05,
      "loss": 1.3532,
      "num_input_tokens_seen": 15934952,
      "step": 43500
    },
    {
      "epoch": 0.9786476868327402,
      "grad_norm": 2.037935495376587,
      "learning_rate": 2.5533807829181493e-05,
      "loss": 1.3318,
      "num_input_tokens_seen": 16117896,
      "step": 44000
    },
    {
      "epoch": 0.9897686832740213,
      "grad_norm": 2.7559268474578857,
      "learning_rate": 2.5255782918149468e-05,
      "loss": 1.3325,
      "num_input_tokens_seen": 16298928,
      "step": 44500
    },
    {
      "epoch": 1.0008896797153024,
      "grad_norm": 2.2017595767974854,
      "learning_rate": 2.4977758007117437e-05,
      "loss": 1.3605,
      "num_input_tokens_seen": 16481072,
      "step": 45000
    },
    {
      "epoch": 1.0120106761565837,
      "grad_norm": 2.3097991943359375,
      "learning_rate": 2.4699733096085412e-05,
      "loss": 1.3404,
      "num_input_tokens_seen": 16664336,
      "step": 45500
    },
    {
      "epoch": 1.0231316725978647,
      "grad_norm": 2.6227993965148926,
      "learning_rate": 2.4421708185053384e-05,
      "loss": 1.3235,
      "num_input_tokens_seen": 16845800,
      "step": 46000
    },
    {
      "epoch": 1.0342526690391458,
      "grad_norm": 2.24474835395813,
      "learning_rate": 2.4143683274021353e-05,
      "loss": 1.3376,
      "num_input_tokens_seen": 17029008,
      "step": 46500
    },
    {
      "epoch": 1.045373665480427,
      "grad_norm": 2.7171192169189453,
      "learning_rate": 2.3865658362989325e-05,
      "loss": 1.3188,
      "num_input_tokens_seen": 17215936,
      "step": 47000
    },
    {
      "epoch": 1.0564946619217082,
      "grad_norm": 2.5323736667633057,
      "learning_rate": 2.3587633451957297e-05,
      "loss": 1.3369,
      "num_input_tokens_seen": 17401232,
      "step": 47500
    },
    {
      "epoch": 1.0676156583629894,
      "grad_norm": 2.267789363861084,
      "learning_rate": 2.330960854092527e-05,
      "loss": 1.3215,
      "num_input_tokens_seen": 17583344,
      "step": 48000
    },
    {
      "epoch": 1.0787366548042705,
      "grad_norm": 3.399862289428711,
      "learning_rate": 2.3031583629893237e-05,
      "loss": 1.3464,
      "num_input_tokens_seen": 17771600,
      "step": 48500
    },
    {
      "epoch": 1.0898576512455516,
      "grad_norm": 2.8749985694885254,
      "learning_rate": 2.2753558718861213e-05,
      "loss": 1.2928,
      "num_input_tokens_seen": 17957200,
      "step": 49000
    },
    {
      "epoch": 1.1009786476868326,
      "grad_norm": 2.6826517581939697,
      "learning_rate": 2.247553380782918e-05,
      "loss": 1.3191,
      "num_input_tokens_seen": 18138560,
      "step": 49500
    },
    {
      "epoch": 1.112099644128114,
      "grad_norm": 2.2963333129882812,
      "learning_rate": 2.2197508896797153e-05,
      "loss": 1.3059,
      "num_input_tokens_seen": 18326104,
      "step": 50000
    },
    {
      "epoch": 1.112099644128114,
      "eval_loss": 1.1665468215942383,
      "eval_runtime": 3.0497,
      "eval_samples_per_second": 819.748,
      "eval_steps_per_second": 102.632,
      "num_input_tokens_seen": 18326104,
      "step": 50000
    },
    {
      "epoch": 1.123220640569395,
      "grad_norm": 2.2697386741638184,
      "learning_rate": 2.1919483985765125e-05,
      "loss": 1.321,
      "num_input_tokens_seen": 18508560,
      "step": 50500
    },
    {
      "epoch": 1.1343416370106763,
      "grad_norm": 2.436851739883423,
      "learning_rate": 2.1641459074733097e-05,
      "loss": 1.3528,
      "num_input_tokens_seen": 18689264,
      "step": 51000
    },
    {
      "epoch": 1.1454626334519573,
      "grad_norm": 2.297527313232422,
      "learning_rate": 2.136343416370107e-05,
      "loss": 1.2987,
      "num_input_tokens_seen": 18874328,
      "step": 51500
    },
    {
      "epoch": 1.1565836298932384,
      "grad_norm": 2.5088889598846436,
      "learning_rate": 2.1085409252669038e-05,
      "loss": 1.3139,
      "num_input_tokens_seen": 19060472,
      "step": 52000
    },
    {
      "epoch": 1.1677046263345197,
      "grad_norm": 2.067575454711914,
      "learning_rate": 2.0807384341637014e-05,
      "loss": 1.2961,
      "num_input_tokens_seen": 19247416,
      "step": 52500
    },
    {
      "epoch": 1.1788256227758007,
      "grad_norm": 2.467543363571167,
      "learning_rate": 2.0529359430604982e-05,
      "loss": 1.3279,
      "num_input_tokens_seen": 19436888,
      "step": 53000
    },
    {
      "epoch": 1.1899466192170818,
      "grad_norm": 3.4245800971984863,
      "learning_rate": 2.0251334519572954e-05,
      "loss": 1.3303,
      "num_input_tokens_seen": 19616320,
      "step": 53500
    },
    {
      "epoch": 1.201067615658363,
      "grad_norm": 2.759120464324951,
      "learning_rate": 1.9973309608540926e-05,
      "loss": 1.3201,
      "num_input_tokens_seen": 19793576,
      "step": 54000
    },
    {
      "epoch": 1.2121886120996441,
      "grad_norm": 2.7749531269073486,
      "learning_rate": 1.9695284697508898e-05,
      "loss": 1.3194,
      "num_input_tokens_seen": 19980880,
      "step": 54500
    },
    {
      "epoch": 1.2233096085409252,
      "grad_norm": 2.4467661380767822,
      "learning_rate": 1.9417259786476867e-05,
      "loss": 1.3293,
      "num_input_tokens_seen": 20163048,
      "step": 55000
    },
    {
      "epoch": 1.2344306049822065,
      "grad_norm": 3.4420840740203857,
      "learning_rate": 1.9139234875444842e-05,
      "loss": 1.3015,
      "num_input_tokens_seen": 20340904,
      "step": 55500
    },
    {
      "epoch": 1.2455516014234875,
      "grad_norm": 2.4761664867401123,
      "learning_rate": 1.8861209964412814e-05,
      "loss": 1.3292,
      "num_input_tokens_seen": 20524328,
      "step": 56000
    },
    {
      "epoch": 1.2566725978647688,
      "grad_norm": 3.0505285263061523,
      "learning_rate": 1.8583185053380783e-05,
      "loss": 1.317,
      "num_input_tokens_seen": 20707808,
      "step": 56500
    },
    {
      "epoch": 1.2677935943060499,
      "grad_norm": 2.361429214477539,
      "learning_rate": 1.8305160142348755e-05,
      "loss": 1.3176,
      "num_input_tokens_seen": 20892744,
      "step": 57000
    },
    {
      "epoch": 1.278914590747331,
      "grad_norm": 1.9151511192321777,
      "learning_rate": 1.8027135231316727e-05,
      "loss": 1.3267,
      "num_input_tokens_seen": 21072216,
      "step": 57500
    },
    {
      "epoch": 1.290035587188612,
      "grad_norm": 3.0513691902160645,
      "learning_rate": 1.77491103202847e-05,
      "loss": 1.2948,
      "num_input_tokens_seen": 21256064,
      "step": 58000
    },
    {
      "epoch": 1.3011565836298933,
      "grad_norm": 1.7151504755020142,
      "learning_rate": 1.7471085409252668e-05,
      "loss": 1.2967,
      "num_input_tokens_seen": 21439648,
      "step": 58500
    },
    {
      "epoch": 1.3122775800711743,
      "grad_norm": 2.3011133670806885,
      "learning_rate": 1.7193060498220643e-05,
      "loss": 1.3199,
      "num_input_tokens_seen": 21621776,
      "step": 59000
    },
    {
      "epoch": 1.3233985765124556,
      "grad_norm": 2.1768672466278076,
      "learning_rate": 1.691503558718861e-05,
      "loss": 1.3233,
      "num_input_tokens_seen": 21810728,
      "step": 59500
    },
    {
      "epoch": 1.3345195729537367,
      "grad_norm": 3.249089241027832,
      "learning_rate": 1.6637010676156584e-05,
      "loss": 1.3298,
      "num_input_tokens_seen": 21991016,
      "step": 60000
    },
    {
      "epoch": 1.3345195729537367,
      "eval_loss": 1.1589475870132446,
      "eval_runtime": 3.0361,
      "eval_samples_per_second": 823.416,
      "eval_steps_per_second": 103.092,
      "num_input_tokens_seen": 21991016,
      "step": 60000
    },
    {
      "epoch": 1.3456405693950177,
      "grad_norm": 1.7282594442367554,
      "learning_rate": 1.6358985765124556e-05,
      "loss": 1.325,
      "num_input_tokens_seen": 22178168,
      "step": 60500
    },
    {
      "epoch": 1.3567615658362988,
      "grad_norm": 2.2659966945648193,
      "learning_rate": 1.6080960854092528e-05,
      "loss": 1.3321,
      "num_input_tokens_seen": 22359360,
      "step": 61000
    },
    {
      "epoch": 1.36788256227758,
      "grad_norm": 2.155791759490967,
      "learning_rate": 1.58029359430605e-05,
      "loss": 1.2995,
      "num_input_tokens_seen": 22543952,
      "step": 61500
    },
    {
      "epoch": 1.3790035587188612,
      "grad_norm": 2.902367353439331,
      "learning_rate": 1.5524911032028472e-05,
      "loss": 1.3061,
      "num_input_tokens_seen": 22728064,
      "step": 62000
    },
    {
      "epoch": 1.3901245551601424,
      "grad_norm": 2.168686866760254,
      "learning_rate": 1.5246886120996442e-05,
      "loss": 1.3231,
      "num_input_tokens_seen": 22913608,
      "step": 62500
    },
    {
      "epoch": 1.4012455516014235,
      "grad_norm": 2.682150363922119,
      "learning_rate": 1.4968861209964412e-05,
      "loss": 1.3091,
      "num_input_tokens_seen": 23101416,
      "step": 63000
    },
    {
      "epoch": 1.4123665480427046,
      "grad_norm": 2.82143497467041,
      "learning_rate": 1.4690836298932384e-05,
      "loss": 1.2922,
      "num_input_tokens_seen": 23281200,
      "step": 63500
    },
    {
      "epoch": 1.4234875444839858,
      "grad_norm": 2.1155049800872803,
      "learning_rate": 1.4412811387900358e-05,
      "loss": 1.3015,
      "num_input_tokens_seen": 23462552,
      "step": 64000
    },
    {
      "epoch": 1.434608540925267,
      "grad_norm": 2.6742069721221924,
      "learning_rate": 1.4134786476868328e-05,
      "loss": 1.3108,
      "num_input_tokens_seen": 23640424,
      "step": 64500
    },
    {
      "epoch": 1.445729537366548,
      "grad_norm": 2.575198173522949,
      "learning_rate": 1.3856761565836299e-05,
      "loss": 1.2987,
      "num_input_tokens_seen": 23825392,
      "step": 65000
    },
    {
      "epoch": 1.4568505338078293,
      "grad_norm": 3.2627415657043457,
      "learning_rate": 1.3578736654804272e-05,
      "loss": 1.3096,
      "num_input_tokens_seen": 24004872,
      "step": 65500
    },
    {
      "epoch": 1.4679715302491103,
      "grad_norm": 3.2727510929107666,
      "learning_rate": 1.3300711743772243e-05,
      "loss": 1.3108,
      "num_input_tokens_seen": 24186064,
      "step": 66000
    },
    {
      "epoch": 1.4790925266903914,
      "grad_norm": 2.187281608581543,
      "learning_rate": 1.3022686832740213e-05,
      "loss": 1.3239,
      "num_input_tokens_seen": 24367968,
      "step": 66500
    },
    {
      "epoch": 1.4902135231316727,
      "grad_norm": 3.5734856128692627,
      "learning_rate": 1.2744661921708187e-05,
      "loss": 1.3241,
      "num_input_tokens_seen": 24551184,
      "step": 67000
    },
    {
      "epoch": 1.5013345195729537,
      "grad_norm": 2.756578207015991,
      "learning_rate": 1.2466637010676157e-05,
      "loss": 1.307,
      "num_input_tokens_seen": 24730288,
      "step": 67500
    },
    {
      "epoch": 1.512455516014235,
      "grad_norm": 2.3752739429473877,
      "learning_rate": 1.2188612099644127e-05,
      "loss": 1.2963,
      "num_input_tokens_seen": 24914816,
      "step": 68000
    },
    {
      "epoch": 1.523576512455516,
      "grad_norm": 3.7108139991760254,
      "learning_rate": 1.1910587188612101e-05,
      "loss": 1.3126,
      "num_input_tokens_seen": 25100648,
      "step": 68500
    },
    {
      "epoch": 1.5346975088967971,
      "grad_norm": 2.5954089164733887,
      "learning_rate": 1.1632562277580072e-05,
      "loss": 1.3053,
      "num_input_tokens_seen": 25282312,
      "step": 69000
    },
    {
      "epoch": 1.5458185053380782,
      "grad_norm": 3.0645289421081543,
      "learning_rate": 1.1354537366548044e-05,
      "loss": 1.2976,
      "num_input_tokens_seen": 25465504,
      "step": 69500
    },
    {
      "epoch": 1.5569395017793595,
      "grad_norm": 2.3734166622161865,
      "learning_rate": 1.1076512455516016e-05,
      "loss": 1.2994,
      "num_input_tokens_seen": 25656600,
      "step": 70000
    },
    {
      "epoch": 1.5569395017793595,
      "eval_loss": 1.1562061309814453,
      "eval_runtime": 2.9007,
      "eval_samples_per_second": 861.851,
      "eval_steps_per_second": 107.904,
      "num_input_tokens_seen": 25656600,
      "step": 70000
    },
    {
      "epoch": 1.5680604982206405,
      "grad_norm": 2.8284573554992676,
      "learning_rate": 1.0798487544483986e-05,
      "loss": 1.3125,
      "num_input_tokens_seen": 25844216,
      "step": 70500
    },
    {
      "epoch": 1.5791814946619218,
      "grad_norm": 1.8697574138641357,
      "learning_rate": 1.0520462633451958e-05,
      "loss": 1.3247,
      "num_input_tokens_seen": 26026824,
      "step": 71000
    },
    {
      "epoch": 1.5903024911032029,
      "grad_norm": 2.4746646881103516,
      "learning_rate": 1.024243772241993e-05,
      "loss": 1.3295,
      "num_input_tokens_seen": 26211760,
      "step": 71500
    },
    {
      "epoch": 1.601423487544484,
      "grad_norm": 2.415778398513794,
      "learning_rate": 9.9644128113879e-06,
      "loss": 1.3191,
      "num_input_tokens_seen": 26393624,
      "step": 72000
    },
    {
      "epoch": 1.612544483985765,
      "grad_norm": 2.263882875442505,
      "learning_rate": 9.686387900355872e-06,
      "loss": 1.3136,
      "num_input_tokens_seen": 26576872,
      "step": 72500
    },
    {
      "epoch": 1.6236654804270463,
      "grad_norm": 2.436645269393921,
      "learning_rate": 9.408362989323843e-06,
      "loss": 1.286,
      "num_input_tokens_seen": 26763752,
      "step": 73000
    },
    {
      "epoch": 1.6347864768683276,
      "grad_norm": 2.739278793334961,
      "learning_rate": 9.130338078291816e-06,
      "loss": 1.2712,
      "num_input_tokens_seen": 26943856,
      "step": 73500
    },
    {
      "epoch": 1.6459074733096086,
      "grad_norm": 2.406345844268799,
      "learning_rate": 8.852313167259788e-06,
      "loss": 1.305,
      "num_input_tokens_seen": 27122192,
      "step": 74000
    },
    {
      "epoch": 1.6570284697508897,
      "grad_norm": 2.1659858226776123,
      "learning_rate": 8.574288256227759e-06,
      "loss": 1.312,
      "num_input_tokens_seen": 27305912,
      "step": 74500
    },
    {
      "epoch": 1.6681494661921707,
      "grad_norm": 2.7831106185913086,
      "learning_rate": 8.29626334519573e-06,
      "loss": 1.2997,
      "num_input_tokens_seen": 27485504,
      "step": 75000
    },
    {
      "epoch": 1.6792704626334518,
      "grad_norm": 2.9885916709899902,
      "learning_rate": 8.018238434163701e-06,
      "loss": 1.3187,
      "num_input_tokens_seen": 27665584,
      "step": 75500
    },
    {
      "epoch": 1.690391459074733,
      "grad_norm": 2.5562667846679688,
      "learning_rate": 7.740213523131673e-06,
      "loss": 1.3142,
      "num_input_tokens_seen": 27851448,
      "step": 76000
    },
    {
      "epoch": 1.7015124555160144,
      "grad_norm": 2.9897525310516357,
      "learning_rate": 7.462188612099645e-06,
      "loss": 1.3071,
      "num_input_tokens_seen": 28033824,
      "step": 76500
    },
    {
      "epoch": 1.7126334519572954,
      "grad_norm": 2.6760592460632324,
      "learning_rate": 7.184163701067615e-06,
      "loss": 1.3049,
      "num_input_tokens_seen": 28220000,
      "step": 77000
    },
    {
      "epoch": 1.7237544483985765,
      "grad_norm": 2.314532995223999,
      "learning_rate": 6.906138790035588e-06,
      "loss": 1.3124,
      "num_input_tokens_seen": 28403400,
      "step": 77500
    },
    {
      "epoch": 1.7348754448398576,
      "grad_norm": 2.2853899002075195,
      "learning_rate": 6.6281138790035586e-06,
      "loss": 1.3188,
      "num_input_tokens_seen": 28585168,
      "step": 78000
    },
    {
      "epoch": 1.7459964412811388,
      "grad_norm": 2.462369918823242,
      "learning_rate": 6.3500889679715306e-06,
      "loss": 1.3146,
      "num_input_tokens_seen": 28767960,
      "step": 78500
    },
    {
      "epoch": 1.75711743772242,
      "grad_norm": 2.626847505569458,
      "learning_rate": 6.072064056939502e-06,
      "loss": 1.3082,
      "num_input_tokens_seen": 28950344,
      "step": 79000
    },
    {
      "epoch": 1.7682384341637012,
      "grad_norm": 3.058187484741211,
      "learning_rate": 5.794039145907473e-06,
      "loss": 1.301,
      "num_input_tokens_seen": 29133248,
      "step": 79500
    },
    {
      "epoch": 1.7793594306049823,
      "grad_norm": 1.932924747467041,
      "learning_rate": 5.516014234875446e-06,
      "loss": 1.2952,
      "num_input_tokens_seen": 29314808,
      "step": 80000
    },
    {
      "epoch": 1.7793594306049823,
      "eval_loss": 1.1517876386642456,
      "eval_runtime": 3.0679,
      "eval_samples_per_second": 814.88,
      "eval_steps_per_second": 102.023,
      "num_input_tokens_seen": 29314808,
      "step": 80000
    },
    {
      "epoch": 1.7904804270462633,
      "grad_norm": 3.326586961746216,
      "learning_rate": 5.237989323843417e-06,
      "loss": 1.3161,
      "num_input_tokens_seen": 29504040,
      "step": 80500
    },
    {
      "epoch": 1.8016014234875444,
      "grad_norm": 3.139636754989624,
      "learning_rate": 4.959964412811388e-06,
      "loss": 1.3,
      "num_input_tokens_seen": 29686712,
      "step": 81000
    },
    {
      "epoch": 1.8127224199288257,
      "grad_norm": 2.5767743587493896,
      "learning_rate": 4.681939501779359e-06,
      "loss": 1.2934,
      "num_input_tokens_seen": 29868136,
      "step": 81500
    },
    {
      "epoch": 1.8238434163701067,
      "grad_norm": 2.1930339336395264,
      "learning_rate": 4.4039145907473305e-06,
      "loss": 1.3104,
      "num_input_tokens_seen": 30050016,
      "step": 82000
    },
    {
      "epoch": 1.834964412811388,
      "grad_norm": 2.9890389442443848,
      "learning_rate": 4.125889679715303e-06,
      "loss": 1.303,
      "num_input_tokens_seen": 30234304,
      "step": 82500
    },
    {
      "epoch": 1.846085409252669,
      "grad_norm": 2.9597034454345703,
      "learning_rate": 3.8478647686832745e-06,
      "loss": 1.3024,
      "num_input_tokens_seen": 30419248,
      "step": 83000
    },
    {
      "epoch": 1.8572064056939501,
      "grad_norm": 3.17301082611084,
      "learning_rate": 3.5698398576512457e-06,
      "loss": 1.2926,
      "num_input_tokens_seen": 30601776,
      "step": 83500
    },
    {
      "epoch": 1.8683274021352312,
      "grad_norm": 2.677340269088745,
      "learning_rate": 3.291814946619217e-06,
      "loss": 1.3184,
      "num_input_tokens_seen": 30783800,
      "step": 84000
    },
    {
      "epoch": 1.8794483985765125,
      "grad_norm": 3.5369062423706055,
      "learning_rate": 3.013790035587189e-06,
      "loss": 1.2858,
      "num_input_tokens_seen": 30968472,
      "step": 84500
    },
    {
      "epoch": 1.8905693950177938,
      "grad_norm": 2.870908737182617,
      "learning_rate": 2.73576512455516e-06,
      "loss": 1.3107,
      "num_input_tokens_seen": 31147072,
      "step": 85000
    },
    {
      "epoch": 1.9016903914590748,
      "grad_norm": 2.1105244159698486,
      "learning_rate": 2.457740213523132e-06,
      "loss": 1.2797,
      "num_input_tokens_seen": 31330536,
      "step": 85500
    },
    {
      "epoch": 1.9128113879003559,
      "grad_norm": 4.080565452575684,
      "learning_rate": 2.1797153024911032e-06,
      "loss": 1.2991,
      "num_input_tokens_seen": 31507752,
      "step": 86000
    },
    {
      "epoch": 1.923932384341637,
      "grad_norm": 3.036339044570923,
      "learning_rate": 1.901690391459075e-06,
      "loss": 1.2884,
      "num_input_tokens_seen": 31685504,
      "step": 86500
    },
    {
      "epoch": 1.935053380782918,
      "grad_norm": 2.3314597606658936,
      "learning_rate": 1.6236654804270462e-06,
      "loss": 1.2935,
      "num_input_tokens_seen": 31865648,
      "step": 87000
    },
    {
      "epoch": 1.9461743772241993,
      "grad_norm": 2.4721710681915283,
      "learning_rate": 1.3456405693950178e-06,
      "loss": 1.3124,
      "num_input_tokens_seen": 32046704,
      "step": 87500
    },
    {
      "epoch": 1.9572953736654806,
      "grad_norm": 2.4747238159179688,
      "learning_rate": 1.0676156583629894e-06,
      "loss": 1.2941,
      "num_input_tokens_seen": 32235800,
      "step": 88000
    },
    {
      "epoch": 1.9684163701067616,
      "grad_norm": 2.4421122074127197,
      "learning_rate": 7.89590747330961e-07,
      "loss": 1.3187,
      "num_input_tokens_seen": 32415672,
      "step": 88500
    },
    {
      "epoch": 1.9795373665480427,
      "grad_norm": 3.469534158706665,
      "learning_rate": 5.115658362989324e-07,
      "loss": 1.2897,
      "num_input_tokens_seen": 32597136,
      "step": 89000
    },
    {
      "epoch": 1.9906583629893237,
      "grad_norm": 3.1344103813171387,
      "learning_rate": 2.335409252669039e-07,
      "loss": 1.2902,
      "num_input_tokens_seen": 32782688,
      "step": 89500
    },
    {
      "epoch": 2.0,
      "num_input_tokens_seen": 32939232,
      "step": 89920,
      "total_flos": 1.1959161056722944e+16,
      "train_loss": 1.358397777055082,
      "train_runtime": 3698.2072,
      "train_samples_per_second": 194.516,
      "train_steps_per_second": 24.314,
      "train_tokens_per_second": 8905.5
    }
  ],
  "logging_steps": 500,
  "max_steps": 89920,
  "num_input_tokens_seen": 32939232,
  "num_train_epochs": 2,
  "save_steps": 10000,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 1.1959161056722944e+16,
  "train_batch_size": 8,
  "trial_name": null,
  "trial_params": null
}