gsmyrnis's picture
End of training
b2e8c57 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 918,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.032679738562091505,
"grad_norm": 13.884983365234152,
"learning_rate": 5e-06,
"loss": 0.9132,
"step": 10
},
{
"epoch": 0.06535947712418301,
"grad_norm": 4.396307073532539,
"learning_rate": 5e-06,
"loss": 0.793,
"step": 20
},
{
"epoch": 0.09803921568627451,
"grad_norm": 1.5701715316993357,
"learning_rate": 5e-06,
"loss": 0.7554,
"step": 30
},
{
"epoch": 0.13071895424836602,
"grad_norm": 1.0469284618172925,
"learning_rate": 5e-06,
"loss": 0.7308,
"step": 40
},
{
"epoch": 0.16339869281045752,
"grad_norm": 1.025316533729498,
"learning_rate": 5e-06,
"loss": 0.7074,
"step": 50
},
{
"epoch": 0.19607843137254902,
"grad_norm": 0.8007907881089358,
"learning_rate": 5e-06,
"loss": 0.6935,
"step": 60
},
{
"epoch": 0.22875816993464052,
"grad_norm": 0.7336515686610362,
"learning_rate": 5e-06,
"loss": 0.6863,
"step": 70
},
{
"epoch": 0.26143790849673204,
"grad_norm": 0.6654762709275407,
"learning_rate": 5e-06,
"loss": 0.6737,
"step": 80
},
{
"epoch": 0.29411764705882354,
"grad_norm": 0.6752715837026988,
"learning_rate": 5e-06,
"loss": 0.678,
"step": 90
},
{
"epoch": 0.32679738562091504,
"grad_norm": 0.6495908393250515,
"learning_rate": 5e-06,
"loss": 0.6625,
"step": 100
},
{
"epoch": 0.35947712418300654,
"grad_norm": 0.5513534888680728,
"learning_rate": 5e-06,
"loss": 0.6728,
"step": 110
},
{
"epoch": 0.39215686274509803,
"grad_norm": 0.5987482687514573,
"learning_rate": 5e-06,
"loss": 0.6605,
"step": 120
},
{
"epoch": 0.42483660130718953,
"grad_norm": 0.8982453869254353,
"learning_rate": 5e-06,
"loss": 0.6576,
"step": 130
},
{
"epoch": 0.45751633986928103,
"grad_norm": 0.5729452075955007,
"learning_rate": 5e-06,
"loss": 0.6581,
"step": 140
},
{
"epoch": 0.49019607843137253,
"grad_norm": 0.6356665135318063,
"learning_rate": 5e-06,
"loss": 0.6488,
"step": 150
},
{
"epoch": 0.5228758169934641,
"grad_norm": 0.8565590645061701,
"learning_rate": 5e-06,
"loss": 0.6611,
"step": 160
},
{
"epoch": 0.5555555555555556,
"grad_norm": 0.7114568602994797,
"learning_rate": 5e-06,
"loss": 0.6448,
"step": 170
},
{
"epoch": 0.5882352941176471,
"grad_norm": 0.6611333193380935,
"learning_rate": 5e-06,
"loss": 0.6478,
"step": 180
},
{
"epoch": 0.6209150326797386,
"grad_norm": 0.6896881110018678,
"learning_rate": 5e-06,
"loss": 0.6495,
"step": 190
},
{
"epoch": 0.6535947712418301,
"grad_norm": 0.6559022818956797,
"learning_rate": 5e-06,
"loss": 0.6501,
"step": 200
},
{
"epoch": 0.6862745098039216,
"grad_norm": 0.557250562288118,
"learning_rate": 5e-06,
"loss": 0.6462,
"step": 210
},
{
"epoch": 0.7189542483660131,
"grad_norm": 0.546010994925319,
"learning_rate": 5e-06,
"loss": 0.6447,
"step": 220
},
{
"epoch": 0.7516339869281046,
"grad_norm": 0.6551946905014553,
"learning_rate": 5e-06,
"loss": 0.6441,
"step": 230
},
{
"epoch": 0.7843137254901961,
"grad_norm": 0.7098356396395941,
"learning_rate": 5e-06,
"loss": 0.6404,
"step": 240
},
{
"epoch": 0.8169934640522876,
"grad_norm": 0.6955204100907911,
"learning_rate": 5e-06,
"loss": 0.6328,
"step": 250
},
{
"epoch": 0.8496732026143791,
"grad_norm": 0.8535008368638369,
"learning_rate": 5e-06,
"loss": 0.6376,
"step": 260
},
{
"epoch": 0.8823529411764706,
"grad_norm": 0.9070401054614784,
"learning_rate": 5e-06,
"loss": 0.639,
"step": 270
},
{
"epoch": 0.9150326797385621,
"grad_norm": 1.565789969572099,
"learning_rate": 5e-06,
"loss": 0.6359,
"step": 280
},
{
"epoch": 0.9477124183006536,
"grad_norm": 1.2559271663411875,
"learning_rate": 5e-06,
"loss": 0.634,
"step": 290
},
{
"epoch": 0.9803921568627451,
"grad_norm": 0.7809173302366459,
"learning_rate": 5e-06,
"loss": 0.6319,
"step": 300
},
{
"epoch": 1.0,
"eval_loss": 0.6343755125999451,
"eval_runtime": 30.4377,
"eval_samples_per_second": 270.75,
"eval_steps_per_second": 1.084,
"step": 306
},
{
"epoch": 1.0130718954248366,
"grad_norm": 0.8552427081362785,
"learning_rate": 5e-06,
"loss": 0.613,
"step": 310
},
{
"epoch": 1.0457516339869282,
"grad_norm": 0.7483473424289742,
"learning_rate": 5e-06,
"loss": 0.5907,
"step": 320
},
{
"epoch": 1.0784313725490196,
"grad_norm": 0.954467001493486,
"learning_rate": 5e-06,
"loss": 0.5879,
"step": 330
},
{
"epoch": 1.1111111111111112,
"grad_norm": 0.5442657939315798,
"learning_rate": 5e-06,
"loss": 0.5885,
"step": 340
},
{
"epoch": 1.1437908496732025,
"grad_norm": 0.7256618392119529,
"learning_rate": 5e-06,
"loss": 0.5913,
"step": 350
},
{
"epoch": 1.1764705882352942,
"grad_norm": 0.6093376203471469,
"learning_rate": 5e-06,
"loss": 0.5941,
"step": 360
},
{
"epoch": 1.2091503267973855,
"grad_norm": 0.563669838714874,
"learning_rate": 5e-06,
"loss": 0.5906,
"step": 370
},
{
"epoch": 1.2418300653594772,
"grad_norm": 0.6235844786429067,
"learning_rate": 5e-06,
"loss": 0.5912,
"step": 380
},
{
"epoch": 1.2745098039215685,
"grad_norm": 0.7954936323443693,
"learning_rate": 5e-06,
"loss": 0.5854,
"step": 390
},
{
"epoch": 1.3071895424836601,
"grad_norm": 0.5900417117740443,
"learning_rate": 5e-06,
"loss": 0.5902,
"step": 400
},
{
"epoch": 1.3398692810457518,
"grad_norm": 0.5649519787360927,
"learning_rate": 5e-06,
"loss": 0.5929,
"step": 410
},
{
"epoch": 1.3725490196078431,
"grad_norm": 0.5532481138751179,
"learning_rate": 5e-06,
"loss": 0.5949,
"step": 420
},
{
"epoch": 1.4052287581699345,
"grad_norm": 0.5796446258489036,
"learning_rate": 5e-06,
"loss": 0.591,
"step": 430
},
{
"epoch": 1.4379084967320261,
"grad_norm": 0.5180305142972244,
"learning_rate": 5e-06,
"loss": 0.5955,
"step": 440
},
{
"epoch": 1.4705882352941178,
"grad_norm": 0.5316504409753056,
"learning_rate": 5e-06,
"loss": 0.5916,
"step": 450
},
{
"epoch": 1.5032679738562091,
"grad_norm": 0.5669201774818233,
"learning_rate": 5e-06,
"loss": 0.5884,
"step": 460
},
{
"epoch": 1.5359477124183005,
"grad_norm": 0.5474083628106629,
"learning_rate": 5e-06,
"loss": 0.5896,
"step": 470
},
{
"epoch": 1.5686274509803921,
"grad_norm": 0.5404778686275579,
"learning_rate": 5e-06,
"loss": 0.593,
"step": 480
},
{
"epoch": 1.6013071895424837,
"grad_norm": 0.5872446910510764,
"learning_rate": 5e-06,
"loss": 0.5921,
"step": 490
},
{
"epoch": 1.6339869281045751,
"grad_norm": 0.6115573952876731,
"learning_rate": 5e-06,
"loss": 0.5916,
"step": 500
},
{
"epoch": 1.6666666666666665,
"grad_norm": 0.6081257294896645,
"learning_rate": 5e-06,
"loss": 0.5903,
"step": 510
},
{
"epoch": 1.6993464052287581,
"grad_norm": 0.5308744191347566,
"learning_rate": 5e-06,
"loss": 0.5859,
"step": 520
},
{
"epoch": 1.7320261437908497,
"grad_norm": 0.5659221276589558,
"learning_rate": 5e-06,
"loss": 0.5916,
"step": 530
},
{
"epoch": 1.7647058823529411,
"grad_norm": 0.5192160551754689,
"learning_rate": 5e-06,
"loss": 0.5914,
"step": 540
},
{
"epoch": 1.7973856209150327,
"grad_norm": 0.6988165837234015,
"learning_rate": 5e-06,
"loss": 0.5866,
"step": 550
},
{
"epoch": 1.8300653594771243,
"grad_norm": 0.5487420924541063,
"learning_rate": 5e-06,
"loss": 0.5884,
"step": 560
},
{
"epoch": 1.8627450980392157,
"grad_norm": 0.5682797543154547,
"learning_rate": 5e-06,
"loss": 0.5909,
"step": 570
},
{
"epoch": 1.8954248366013071,
"grad_norm": 0.5077833002974906,
"learning_rate": 5e-06,
"loss": 0.5973,
"step": 580
},
{
"epoch": 1.9281045751633987,
"grad_norm": 0.5129649155292699,
"learning_rate": 5e-06,
"loss": 0.5906,
"step": 590
},
{
"epoch": 1.9607843137254903,
"grad_norm": 0.6199185848699229,
"learning_rate": 5e-06,
"loss": 0.5897,
"step": 600
},
{
"epoch": 1.9934640522875817,
"grad_norm": 0.6367349679762161,
"learning_rate": 5e-06,
"loss": 0.5863,
"step": 610
},
{
"epoch": 2.0,
"eval_loss": 0.6249033808708191,
"eval_runtime": 29.8328,
"eval_samples_per_second": 276.24,
"eval_steps_per_second": 1.106,
"step": 612
},
{
"epoch": 2.026143790849673,
"grad_norm": 0.7411071608832002,
"learning_rate": 5e-06,
"loss": 0.5586,
"step": 620
},
{
"epoch": 2.0588235294117645,
"grad_norm": 0.5717129772857855,
"learning_rate": 5e-06,
"loss": 0.5407,
"step": 630
},
{
"epoch": 2.0915032679738563,
"grad_norm": 0.6346534892777344,
"learning_rate": 5e-06,
"loss": 0.5442,
"step": 640
},
{
"epoch": 2.1241830065359477,
"grad_norm": 0.5360417918450974,
"learning_rate": 5e-06,
"loss": 0.5411,
"step": 650
},
{
"epoch": 2.156862745098039,
"grad_norm": 0.6070614915525901,
"learning_rate": 5e-06,
"loss": 0.5495,
"step": 660
},
{
"epoch": 2.189542483660131,
"grad_norm": 0.5484881674300732,
"learning_rate": 5e-06,
"loss": 0.5477,
"step": 670
},
{
"epoch": 2.2222222222222223,
"grad_norm": 0.5559727309246413,
"learning_rate": 5e-06,
"loss": 0.5453,
"step": 680
},
{
"epoch": 2.2549019607843137,
"grad_norm": 0.6149254917908307,
"learning_rate": 5e-06,
"loss": 0.5491,
"step": 690
},
{
"epoch": 2.287581699346405,
"grad_norm": 0.7434819330252584,
"learning_rate": 5e-06,
"loss": 0.5411,
"step": 700
},
{
"epoch": 2.3202614379084965,
"grad_norm": 0.6277100245140601,
"learning_rate": 5e-06,
"loss": 0.539,
"step": 710
},
{
"epoch": 2.3529411764705883,
"grad_norm": 0.6071082517695824,
"learning_rate": 5e-06,
"loss": 0.546,
"step": 720
},
{
"epoch": 2.3856209150326797,
"grad_norm": 0.5763605350200824,
"learning_rate": 5e-06,
"loss": 0.5424,
"step": 730
},
{
"epoch": 2.418300653594771,
"grad_norm": 0.6503780611807637,
"learning_rate": 5e-06,
"loss": 0.5467,
"step": 740
},
{
"epoch": 2.450980392156863,
"grad_norm": 0.6953693200719112,
"learning_rate": 5e-06,
"loss": 0.551,
"step": 750
},
{
"epoch": 2.4836601307189543,
"grad_norm": 0.582324537988359,
"learning_rate": 5e-06,
"loss": 0.5454,
"step": 760
},
{
"epoch": 2.5163398692810457,
"grad_norm": 0.5697978800379555,
"learning_rate": 5e-06,
"loss": 0.546,
"step": 770
},
{
"epoch": 2.549019607843137,
"grad_norm": 0.588747491644582,
"learning_rate": 5e-06,
"loss": 0.5434,
"step": 780
},
{
"epoch": 2.581699346405229,
"grad_norm": 0.5761332349094056,
"learning_rate": 5e-06,
"loss": 0.5515,
"step": 790
},
{
"epoch": 2.6143790849673203,
"grad_norm": 0.6003606940754802,
"learning_rate": 5e-06,
"loss": 0.5513,
"step": 800
},
{
"epoch": 2.6470588235294117,
"grad_norm": 0.5450155999410424,
"learning_rate": 5e-06,
"loss": 0.5472,
"step": 810
},
{
"epoch": 2.6797385620915035,
"grad_norm": 0.7041726525142917,
"learning_rate": 5e-06,
"loss": 0.5496,
"step": 820
},
{
"epoch": 2.712418300653595,
"grad_norm": 0.5257674189970212,
"learning_rate": 5e-06,
"loss": 0.5444,
"step": 830
},
{
"epoch": 2.7450980392156863,
"grad_norm": 0.5960484874019031,
"learning_rate": 5e-06,
"loss": 0.5501,
"step": 840
},
{
"epoch": 2.7777777777777777,
"grad_norm": 0.5362794145960937,
"learning_rate": 5e-06,
"loss": 0.5477,
"step": 850
},
{
"epoch": 2.810457516339869,
"grad_norm": 0.5503663269672819,
"learning_rate": 5e-06,
"loss": 0.5472,
"step": 860
},
{
"epoch": 2.843137254901961,
"grad_norm": 0.5500519539060107,
"learning_rate": 5e-06,
"loss": 0.5512,
"step": 870
},
{
"epoch": 2.8758169934640523,
"grad_norm": 0.6512147682006916,
"learning_rate": 5e-06,
"loss": 0.5521,
"step": 880
},
{
"epoch": 2.9084967320261437,
"grad_norm": 0.5740057468439662,
"learning_rate": 5e-06,
"loss": 0.5449,
"step": 890
},
{
"epoch": 2.9411764705882355,
"grad_norm": 0.5364033145980043,
"learning_rate": 5e-06,
"loss": 0.5498,
"step": 900
},
{
"epoch": 2.973856209150327,
"grad_norm": 0.5484821254955965,
"learning_rate": 5e-06,
"loss": 0.5495,
"step": 910
},
{
"epoch": 3.0,
"eval_loss": 0.628399670124054,
"eval_runtime": 29.5588,
"eval_samples_per_second": 278.8,
"eval_steps_per_second": 1.116,
"step": 918
},
{
"epoch": 3.0,
"step": 918,
"total_flos": 1537684191313920.0,
"train_loss": 0.6034856663290452,
"train_runtime": 5898.5619,
"train_samples_per_second": 79.628,
"train_steps_per_second": 0.156
}
],
"logging_steps": 10,
"max_steps": 918,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1537684191313920.0,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}