llama-finance-relations / trainer_state.json
Askinkaty's picture
Upload 17 files
4efc8ff verified
raw
history blame
6.99 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9986431478968792,
"eval_steps": 500,
"global_step": 368,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.027137042062415198,
"grad_norm": 1.4497582912445068,
"learning_rate": 0.0001998659482680456,
"loss": 3.4336,
"step": 10
},
{
"epoch": 0.054274084124830396,
"grad_norm": 1.846526861190796,
"learning_rate": 0.00019904804439875633,
"loss": 1.9601,
"step": 20
},
{
"epoch": 0.0814111261872456,
"grad_norm": 0.9556183815002441,
"learning_rate": 0.00019749279121818235,
"loss": 1.202,
"step": 30
},
{
"epoch": 0.10854816824966079,
"grad_norm": 1.0818341970443726,
"learning_rate": 0.00019521176659107142,
"loss": 1.1366,
"step": 40
},
{
"epoch": 0.13568521031207598,
"grad_norm": 0.9879324436187744,
"learning_rate": 0.00019222195128618106,
"loss": 1.0269,
"step": 50
},
{
"epoch": 0.1628222523744912,
"grad_norm": 1.1216270923614502,
"learning_rate": 0.000188545602565321,
"loss": 1.0766,
"step": 60
},
{
"epoch": 0.18995929443690637,
"grad_norm": 1.1110620498657227,
"learning_rate": 0.00018421008849228118,
"loss": 0.9947,
"step": 70
},
{
"epoch": 0.21709633649932158,
"grad_norm": 0.8764580488204956,
"learning_rate": 0.00017924768419510904,
"loss": 1.0024,
"step": 80
},
{
"epoch": 0.24423337856173677,
"grad_norm": 1.2004687786102295,
"learning_rate": 0.00017369533159843369,
"loss": 1.04,
"step": 90
},
{
"epoch": 0.27137042062415195,
"grad_norm": 0.9665369391441345,
"learning_rate": 0.00016759436441447545,
"loss": 0.9809,
"step": 100
},
{
"epoch": 0.29850746268656714,
"grad_norm": 1.5474181175231934,
"learning_rate": 0.00016099020044000727,
"loss": 0.9405,
"step": 110
},
{
"epoch": 0.3256445047489824,
"grad_norm": 1.014674186706543,
"learning_rate": 0.00015393200344991995,
"loss": 0.9054,
"step": 120
},
{
"epoch": 0.35278154681139756,
"grad_norm": 1.013979434967041,
"learning_rate": 0.00014647231720437686,
"loss": 0.9754,
"step": 130
},
{
"epoch": 0.37991858887381275,
"grad_norm": 1.0418416261672974,
"learning_rate": 0.0001386666742941419,
"loss": 0.8711,
"step": 140
},
{
"epoch": 0.40705563093622793,
"grad_norm": 0.9280593395233154,
"learning_rate": 0.0001305731827359753,
"loss": 0.8358,
"step": 150
},
{
"epoch": 0.43419267299864317,
"grad_norm": 1.0268974304199219,
"learning_rate": 0.00012225209339563145,
"loss": 0.8568,
"step": 160
},
{
"epoch": 0.46132971506105835,
"grad_norm": 0.9812881350517273,
"learning_rate": 0.00011376535145871684,
"loss": 0.888,
"step": 170
},
{
"epoch": 0.48846675712347354,
"grad_norm": 1.2109171152114868,
"learning_rate": 0.00010517613528842097,
"loss": 0.9166,
"step": 180
},
{
"epoch": 0.5156037991858887,
"grad_norm": 0.9860134124755859,
"learning_rate": 9.654838610302923e-05,
"loss": 0.8357,
"step": 190
},
{
"epoch": 0.5427408412483039,
"grad_norm": 1.1094295978546143,
"learning_rate": 8.79463319744677e-05,
"loss": 0.928,
"step": 200
},
{
"epoch": 0.5698778833107191,
"grad_norm": 0.9386014342308044,
"learning_rate": 7.943400969140635e-05,
"loss": 0.9522,
"step": 210
},
{
"epoch": 0.5970149253731343,
"grad_norm": 1.0436668395996094,
"learning_rate": 7.107478804634325e-05,
"loss": 0.8619,
"step": 220
},
{
"epoch": 0.6241519674355496,
"grad_norm": 1.0143946409225464,
"learning_rate": 6.293089609549325e-05,
"loss": 0.8046,
"step": 230
},
{
"epoch": 0.6512890094979648,
"grad_norm": 1.107064127922058,
"learning_rate": 5.506295990328385e-05,
"loss": 0.8532,
"step": 240
},
{
"epoch": 0.6784260515603799,
"grad_norm": 0.8748170137405396,
"learning_rate": 4.75295512200992e-05,
"loss": 0.8276,
"step": 250
},
{
"epoch": 0.7055630936227951,
"grad_norm": 1.0721319913864136,
"learning_rate": 4.038675145307747e-05,
"loss": 0.7517,
"step": 260
},
{
"epoch": 0.7327001356852103,
"grad_norm": 1.1542731523513794,
"learning_rate": 3.36877341759205e-05,
"loss": 0.8542,
"step": 270
},
{
"epoch": 0.7598371777476255,
"grad_norm": 0.89125657081604,
"learning_rate": 2.7482369285662378e-05,
"loss": 0.8601,
"step": 280
},
{
"epoch": 0.7869742198100407,
"grad_norm": 0.9904446601867676,
"learning_rate": 2.181685175319702e-05,
"loss": 0.8787,
"step": 290
},
{
"epoch": 0.8141112618724559,
"grad_norm": 0.9311710596084595,
"learning_rate": 1.6733357731279377e-05,
"loss": 0.7493,
"step": 300
},
{
"epoch": 0.841248303934871,
"grad_norm": 1.0216141939163208,
"learning_rate": 1.2269730580055805e-05,
"loss": 0.7497,
"step": 310
},
{
"epoch": 0.8683853459972863,
"grad_norm": 0.9951556324958801,
"learning_rate": 8.45919914746337e-06,
"loss": 0.8324,
"step": 320
},
{
"epoch": 0.8955223880597015,
"grad_norm": 1.0728198289871216,
"learning_rate": 5.3301304017194135e-06,
"loss": 0.7964,
"step": 330
},
{
"epoch": 0.9226594301221167,
"grad_norm": 1.1030397415161133,
"learning_rate": 2.905818257394799e-06,
"loss": 0.8166,
"step": 340
},
{
"epoch": 0.9497964721845319,
"grad_norm": 0.974777102470398,
"learning_rate": 1.2043101671253554e-06,
"loss": 0.7952,
"step": 350
},
{
"epoch": 0.9769335142469471,
"grad_norm": 1.078715443611145,
"learning_rate": 2.382727698752474e-07,
"loss": 0.8726,
"step": 360
}
],
"logging_steps": 10,
"max_steps": 368,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1620605933346816.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}