|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.9591836734693877, |
|
"eval_steps": 13, |
|
"global_step": 147, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02040816326530612, |
|
"grad_norm": 0.7881951332092285, |
|
"learning_rate": 2e-05, |
|
"loss": 2.7509, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.02040816326530612, |
|
"eval_loss": 2.6902382373809814, |
|
"eval_runtime": 269.5606, |
|
"eval_samples_per_second": 6.288, |
|
"eval_steps_per_second": 3.146, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.04081632653061224, |
|
"grad_norm": 0.789082407951355, |
|
"learning_rate": 4e-05, |
|
"loss": 2.7449, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.061224489795918366, |
|
"grad_norm": 0.7354114055633545, |
|
"learning_rate": 6e-05, |
|
"loss": 2.7164, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.08163265306122448, |
|
"grad_norm": 0.7292255759239197, |
|
"learning_rate": 8e-05, |
|
"loss": 2.7174, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.10204081632653061, |
|
"grad_norm": 0.6898028254508972, |
|
"learning_rate": 0.0001, |
|
"loss": 2.6891, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.12244897959183673, |
|
"grad_norm": 0.6861400604248047, |
|
"learning_rate": 0.00012, |
|
"loss": 2.6545, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.14285714285714285, |
|
"grad_norm": 0.7510350346565247, |
|
"learning_rate": 0.00014, |
|
"loss": 2.5656, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.16326530612244897, |
|
"grad_norm": 0.8011165261268616, |
|
"learning_rate": 0.00016, |
|
"loss": 2.4519, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.1836734693877551, |
|
"grad_norm": 0.8624005317687988, |
|
"learning_rate": 0.00018, |
|
"loss": 2.3178, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.20408163265306123, |
|
"grad_norm": 0.8004987835884094, |
|
"learning_rate": 0.0002, |
|
"loss": 2.1783, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.22448979591836735, |
|
"grad_norm": 0.6362400054931641, |
|
"learning_rate": 0.000199985736255971, |
|
"loss": 2.0252, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.24489795918367346, |
|
"grad_norm": 0.7930936217308044, |
|
"learning_rate": 0.0001999429490929718, |
|
"loss": 1.8839, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.2653061224489796, |
|
"grad_norm": 0.5149843096733093, |
|
"learning_rate": 0.00019987165071710527, |
|
"loss": 1.8064, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.2653061224489796, |
|
"eval_loss": 1.6734941005706787, |
|
"eval_runtime": 271.2615, |
|
"eval_samples_per_second": 6.249, |
|
"eval_steps_per_second": 3.126, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.2857142857142857, |
|
"grad_norm": 0.42121434211730957, |
|
"learning_rate": 0.00019977186146800707, |
|
"loss": 1.7922, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.30612244897959184, |
|
"grad_norm": 0.3523242771625519, |
|
"learning_rate": 0.0001996436098130433, |
|
"loss": 1.7711, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.32653061224489793, |
|
"grad_norm": 0.3384595215320587, |
|
"learning_rate": 0.00019948693233918952, |
|
"loss": 1.7152, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.3469387755102041, |
|
"grad_norm": 0.34942421317100525, |
|
"learning_rate": 0.00019930187374259337, |
|
"loss": 1.7112, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.3673469387755102, |
|
"grad_norm": 0.31712639331817627, |
|
"learning_rate": 0.00019908848681582391, |
|
"loss": 1.7059, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.3877551020408163, |
|
"grad_norm": 0.2875436842441559, |
|
"learning_rate": 0.00019884683243281116, |
|
"loss": 1.6468, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.40816326530612246, |
|
"grad_norm": 0.24433130025863647, |
|
"learning_rate": 0.00019857697953148037, |
|
"loss": 1.6408, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.42857142857142855, |
|
"grad_norm": 0.21414674818515778, |
|
"learning_rate": 0.00019827900509408581, |
|
"loss": 1.616, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.4489795918367347, |
|
"grad_norm": 0.21537622809410095, |
|
"learning_rate": 0.00019795299412524945, |
|
"loss": 1.609, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.46938775510204084, |
|
"grad_norm": 0.2432074397802353, |
|
"learning_rate": 0.00019759903962771156, |
|
"loss": 1.6066, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.4897959183673469, |
|
"grad_norm": 0.2359839379787445, |
|
"learning_rate": 0.00019721724257579907, |
|
"loss": 1.5851, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.5102040816326531, |
|
"grad_norm": 0.22065888345241547, |
|
"learning_rate": 0.00019680771188662044, |
|
"loss": 1.5739, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.5306122448979592, |
|
"grad_norm": 0.20339132845401764, |
|
"learning_rate": 0.0001963705643889941, |
|
"loss": 1.5513, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.5306122448979592, |
|
"eval_loss": 1.4832030534744263, |
|
"eval_runtime": 271.2449, |
|
"eval_samples_per_second": 6.249, |
|
"eval_steps_per_second": 3.126, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.5510204081632653, |
|
"grad_norm": 0.18875224888324738, |
|
"learning_rate": 0.00019590592479012023, |
|
"loss": 1.5378, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.5714285714285714, |
|
"grad_norm": 0.18564417958259583, |
|
"learning_rate": 0.00019541392564000488, |
|
"loss": 1.5212, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.5918367346938775, |
|
"grad_norm": 0.16226942837238312, |
|
"learning_rate": 0.00019489470729364692, |
|
"loss": 1.5391, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.6122448979591837, |
|
"grad_norm": 0.15650039911270142, |
|
"learning_rate": 0.00019434841787099803, |
|
"loss": 1.511, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.6326530612244898, |
|
"grad_norm": 0.15976540744304657, |
|
"learning_rate": 0.00019377521321470805, |
|
"loss": 1.5119, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.6530612244897959, |
|
"grad_norm": 0.16409288346767426, |
|
"learning_rate": 0.00019317525684566685, |
|
"loss": 1.4909, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.673469387755102, |
|
"grad_norm": 0.15468019247055054, |
|
"learning_rate": 0.00019254871991635598, |
|
"loss": 1.4951, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.6938775510204082, |
|
"grad_norm": 0.1462036371231079, |
|
"learning_rate": 0.00019189578116202307, |
|
"loss": 1.4643, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.7142857142857143, |
|
"grad_norm": 0.1541963368654251, |
|
"learning_rate": 0.00019121662684969335, |
|
"loss": 1.5159, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.7346938775510204, |
|
"grad_norm": 0.14798064529895782, |
|
"learning_rate": 0.00019051145072503215, |
|
"loss": 1.4741, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.7551020408163265, |
|
"grad_norm": 0.13914817571640015, |
|
"learning_rate": 0.00018978045395707418, |
|
"loss": 1.4788, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.7755102040816326, |
|
"grad_norm": 0.15608824789524078, |
|
"learning_rate": 0.00018902384508083517, |
|
"loss": 1.4687, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.7959183673469388, |
|
"grad_norm": 0.14460116624832153, |
|
"learning_rate": 0.00018824183993782192, |
|
"loss": 1.482, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.7959183673469388, |
|
"eval_loss": 1.411073088645935, |
|
"eval_runtime": 271.292, |
|
"eval_samples_per_second": 6.248, |
|
"eval_steps_per_second": 3.126, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.8163265306122449, |
|
"grad_norm": 0.15740551054477692, |
|
"learning_rate": 0.00018743466161445823, |
|
"loss": 1.4486, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.8367346938775511, |
|
"grad_norm": 0.14149661362171173, |
|
"learning_rate": 0.00018660254037844388, |
|
"loss": 1.4353, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.8571428571428571, |
|
"grad_norm": 0.14034292101860046, |
|
"learning_rate": 0.0001857457136130651, |
|
"loss": 1.4523, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.8775510204081632, |
|
"grad_norm": 0.1487722396850586, |
|
"learning_rate": 0.00018486442574947511, |
|
"loss": 1.4095, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.8979591836734694, |
|
"grad_norm": 0.17400234937667847, |
|
"learning_rate": 0.00018395892819696389, |
|
"loss": 1.4414, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.9183673469387755, |
|
"grad_norm": 0.1741325408220291, |
|
"learning_rate": 0.00018302947927123766, |
|
"loss": 1.4379, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.9387755102040817, |
|
"grad_norm": 0.15319454669952393, |
|
"learning_rate": 0.00018207634412072764, |
|
"loss": 1.405, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.9591836734693877, |
|
"grad_norm": 0.15876264870166779, |
|
"learning_rate": 0.00018109979465095013, |
|
"loss": 1.4122, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.9795918367346939, |
|
"grad_norm": 0.17120805382728577, |
|
"learning_rate": 0.00018010010944693848, |
|
"loss": 1.4132, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.1436116099357605, |
|
"learning_rate": 0.00017907757369376985, |
|
"loss": 1.416, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 1.0204081632653061, |
|
"grad_norm": 0.1707429438829422, |
|
"learning_rate": 0.0001780324790952092, |
|
"loss": 1.3913, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.0204081632653061, |
|
"grad_norm": 0.17117524147033691, |
|
"learning_rate": 0.00017696512379049325, |
|
"loss": 1.3963, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 1.0408163265306123, |
|
"grad_norm": 0.13410089910030365, |
|
"learning_rate": 0.0001758758122692791, |
|
"loss": 1.392, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 1.0408163265306123, |
|
"eval_loss": 1.3676769733428955, |
|
"eval_runtime": 270.8566, |
|
"eval_samples_per_second": 6.258, |
|
"eval_steps_per_second": 3.131, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 1.0612244897959184, |
|
"grad_norm": 0.18877607583999634, |
|
"learning_rate": 0.00017476485528478093, |
|
"loss": 1.3854, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 1.0816326530612246, |
|
"grad_norm": 0.1752927452325821, |
|
"learning_rate": 0.00017363256976511972, |
|
"loss": 1.3759, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 1.1020408163265305, |
|
"grad_norm": 0.17180170118808746, |
|
"learning_rate": 0.000172479278722912, |
|
"loss": 1.3614, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 1.1224489795918366, |
|
"grad_norm": 0.1640290915966034, |
|
"learning_rate": 0.00017130531116312203, |
|
"loss": 1.3853, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 1.1428571428571428, |
|
"grad_norm": 0.2047068476676941, |
|
"learning_rate": 0.0001701110019892053, |
|
"loss": 1.3699, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 1.163265306122449, |
|
"grad_norm": 0.1835869997739792, |
|
"learning_rate": 0.00016889669190756868, |
|
"loss": 1.3403, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 1.183673469387755, |
|
"grad_norm": 0.16733241081237793, |
|
"learning_rate": 0.00016766272733037576, |
|
"loss": 1.3609, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 1.2040816326530612, |
|
"grad_norm": 0.178726926445961, |
|
"learning_rate": 0.00016640946027672392, |
|
"loss": 1.3651, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.2244897959183674, |
|
"grad_norm": 0.16719630360603333, |
|
"learning_rate": 0.00016513724827222227, |
|
"loss": 1.3676, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 1.2448979591836735, |
|
"grad_norm": 0.15999363362789154, |
|
"learning_rate": 0.00016384645424699835, |
|
"loss": 1.3651, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 1.2653061224489797, |
|
"grad_norm": 0.1705988198518753, |
|
"learning_rate": 0.00016253744643216368, |
|
"loss": 1.3757, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 1.2857142857142856, |
|
"grad_norm": 0.14996370673179626, |
|
"learning_rate": 0.0001612105982547663, |
|
"loss": 1.3474, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 1.306122448979592, |
|
"grad_norm": 0.19127260148525238, |
|
"learning_rate": 0.0001598662882312615, |
|
"loss": 1.3414, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.306122448979592, |
|
"eval_loss": 1.331880807876587, |
|
"eval_runtime": 270.8424, |
|
"eval_samples_per_second": 6.258, |
|
"eval_steps_per_second": 3.131, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.3265306122448979, |
|
"grad_norm": 0.16125527024269104, |
|
"learning_rate": 0.00015850489985953076, |
|
"loss": 1.3509, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 1.346938775510204, |
|
"grad_norm": 0.1979473978281021, |
|
"learning_rate": 0.00015712682150947923, |
|
"loss": 1.3579, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 1.3673469387755102, |
|
"grad_norm": 0.18317992985248566, |
|
"learning_rate": 0.00015573244631224365, |
|
"loss": 1.3341, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 1.3877551020408163, |
|
"grad_norm": 0.1646898239850998, |
|
"learning_rate": 0.0001543221720480419, |
|
"loss": 1.3361, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 1.4081632653061225, |
|
"grad_norm": 0.1760271042585373, |
|
"learning_rate": 0.00015289640103269625, |
|
"loss": 1.358, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.4285714285714286, |
|
"grad_norm": 0.165283203125, |
|
"learning_rate": 0.0001514555400028629, |
|
"loss": 1.3072, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 1.4489795918367347, |
|
"grad_norm": 0.1507076472043991, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 1.3133, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 1.469387755102041, |
|
"grad_norm": 0.16913647949695587, |
|
"learning_rate": 0.00014853019625310813, |
|
"loss": 1.3232, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 1.489795918367347, |
|
"grad_norm": 0.18266479671001434, |
|
"learning_rate": 0.0001470465480602756, |
|
"loss": 1.3512, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 1.510204081632653, |
|
"grad_norm": 0.19301828742027283, |
|
"learning_rate": 0.0001455494786690634, |
|
"loss": 1.3241, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.5306122448979593, |
|
"grad_norm": 0.16109652817249298, |
|
"learning_rate": 0.00014403941515576344, |
|
"loss": 1.3256, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 1.5510204081632653, |
|
"grad_norm": 0.17053867876529694, |
|
"learning_rate": 0.00014251678830356408, |
|
"loss": 1.3162, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 1.5714285714285714, |
|
"grad_norm": 0.17348544299602509, |
|
"learning_rate": 0.00014098203247965875, |
|
"loss": 1.3213, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 1.5714285714285714, |
|
"eval_loss": 1.3028697967529297, |
|
"eval_runtime": 270.8095, |
|
"eval_samples_per_second": 6.259, |
|
"eval_steps_per_second": 3.131, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 1.5918367346938775, |
|
"grad_norm": 0.1703907549381256, |
|
"learning_rate": 0.00013943558551133186, |
|
"loss": 1.3073, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 1.6122448979591837, |
|
"grad_norm": 0.17313100397586823, |
|
"learning_rate": 0.0001378778885610576, |
|
"loss": 1.3232, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.6326530612244898, |
|
"grad_norm": 0.17237025499343872, |
|
"learning_rate": 0.00013630938600064747, |
|
"loss": 1.3406, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 1.6530612244897958, |
|
"grad_norm": 0.19658459722995758, |
|
"learning_rate": 0.00013473052528448201, |
|
"loss": 1.3114, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 1.6734693877551021, |
|
"grad_norm": 0.20599938929080963, |
|
"learning_rate": 0.0001331417568218636, |
|
"loss": 1.3288, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 1.693877551020408, |
|
"grad_norm": 0.17759399116039276, |
|
"learning_rate": 0.00013154353384852558, |
|
"loss": 1.2995, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 1.7142857142857144, |
|
"grad_norm": 0.18712250888347626, |
|
"learning_rate": 0.00012993631229733582, |
|
"loss": 1.2895, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.7346938775510203, |
|
"grad_norm": 0.1991330236196518, |
|
"learning_rate": 0.00012832055066823038, |
|
"loss": 1.2886, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 1.7551020408163265, |
|
"grad_norm": 0.22125203907489777, |
|
"learning_rate": 0.00012669670989741517, |
|
"loss": 1.3233, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 1.7755102040816326, |
|
"grad_norm": 0.2052813619375229, |
|
"learning_rate": 0.00012506525322587207, |
|
"loss": 1.3079, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 1.7959183673469388, |
|
"grad_norm": 0.19290736317634583, |
|
"learning_rate": 0.00012342664606720822, |
|
"loss": 1.3174, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 1.816326530612245, |
|
"grad_norm": 0.20912542939186096, |
|
"learning_rate": 0.00012178135587488515, |
|
"loss": 1.2915, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.836734693877551, |
|
"grad_norm": 0.20760588347911835, |
|
"learning_rate": 0.00012012985200886602, |
|
"loss": 1.3028, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 1.836734693877551, |
|
"eval_loss": 1.2795333862304688, |
|
"eval_runtime": 270.6525, |
|
"eval_samples_per_second": 6.263, |
|
"eval_steps_per_second": 3.133, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 1.8571428571428572, |
|
"grad_norm": 0.1996900886297226, |
|
"learning_rate": 0.00011847260560171896, |
|
"loss": 1.3119, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 1.8775510204081631, |
|
"grad_norm": 0.23766876757144928, |
|
"learning_rate": 0.00011681008942421483, |
|
"loss": 1.2978, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 1.8979591836734695, |
|
"grad_norm": 0.19782397150993347, |
|
"learning_rate": 0.00011514277775045768, |
|
"loss": 1.2955, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 1.9183673469387754, |
|
"grad_norm": 0.22519494593143463, |
|
"learning_rate": 0.00011347114622258612, |
|
"loss": 1.2957, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 1.9387755102040818, |
|
"grad_norm": 0.2590245306491852, |
|
"learning_rate": 0.00011179567171508463, |
|
"loss": 1.2809, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 1.9591836734693877, |
|
"grad_norm": 0.2235420197248459, |
|
"learning_rate": 0.00011011683219874323, |
|
"loss": 1.2784, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 1.9795918367346939, |
|
"grad_norm": 0.285740464925766, |
|
"learning_rate": 0.00010843510660430447, |
|
"loss": 1.309, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.20554350316524506, |
|
"learning_rate": 0.00010675097468583652, |
|
"loss": 1.273, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 2.020408163265306, |
|
"grad_norm": 0.24468418955802917, |
|
"learning_rate": 0.00010506491688387127, |
|
"loss": 1.2833, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.020408163265306, |
|
"grad_norm": 0.21553528308868408, |
|
"learning_rate": 0.00010337741418834684, |
|
"loss": 1.2669, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 2.0408163265306123, |
|
"grad_norm": 0.22015659511089325, |
|
"learning_rate": 0.0001016889480013931, |
|
"loss": 1.2795, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 2.061224489795918, |
|
"grad_norm": 0.2028799206018448, |
|
"learning_rate": 0.0001, |
|
"loss": 1.2584, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 2.0816326530612246, |
|
"grad_norm": 0.23474323749542236, |
|
"learning_rate": 9.83110519986069e-05, |
|
"loss": 1.2761, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 2.0816326530612246, |
|
"eval_loss": 1.2696796655654907, |
|
"eval_runtime": 270.6586, |
|
"eval_samples_per_second": 6.263, |
|
"eval_steps_per_second": 3.133, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 2.1020408163265305, |
|
"grad_norm": 0.21070216596126556, |
|
"learning_rate": 9.662258581165319e-05, |
|
"loss": 1.2808, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 2.122448979591837, |
|
"grad_norm": 0.21867221593856812, |
|
"learning_rate": 9.493508311612874e-05, |
|
"loss": 1.2873, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 2.142857142857143, |
|
"grad_norm": 0.21630822122097015, |
|
"learning_rate": 9.324902531416349e-05, |
|
"loss": 1.2527, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 2.163265306122449, |
|
"grad_norm": 0.2134082019329071, |
|
"learning_rate": 9.156489339569554e-05, |
|
"loss": 1.2755, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 2.183673469387755, |
|
"grad_norm": 0.22310714423656464, |
|
"learning_rate": 8.98831678012568e-05, |
|
"loss": 1.2512, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 2.204081632653061, |
|
"grad_norm": 0.2365124374628067, |
|
"learning_rate": 8.820432828491542e-05, |
|
"loss": 1.2725, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 2.2244897959183674, |
|
"grad_norm": 0.2086496651172638, |
|
"learning_rate": 8.652885377741393e-05, |
|
"loss": 1.2488, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 2.2448979591836733, |
|
"grad_norm": 0.20848101377487183, |
|
"learning_rate": 8.485722224954237e-05, |
|
"loss": 1.2793, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 2.2653061224489797, |
|
"grad_norm": 0.20784686505794525, |
|
"learning_rate": 8.31899105757852e-05, |
|
"loss": 1.2564, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 2.2857142857142856, |
|
"grad_norm": 0.21896174550056458, |
|
"learning_rate": 8.15273943982811e-05, |
|
"loss": 1.2515, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 2.306122448979592, |
|
"grad_norm": 0.21367855370044708, |
|
"learning_rate": 7.987014799113397e-05, |
|
"loss": 1.248, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 2.326530612244898, |
|
"grad_norm": 0.20891636610031128, |
|
"learning_rate": 7.821864412511485e-05, |
|
"loss": 1.2753, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 2.3469387755102042, |
|
"grad_norm": 0.2092975378036499, |
|
"learning_rate": 7.65733539327918e-05, |
|
"loss": 1.2509, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 2.3469387755102042, |
|
"eval_loss": 1.258699655532837, |
|
"eval_runtime": 270.5384, |
|
"eval_samples_per_second": 6.265, |
|
"eval_steps_per_second": 3.134, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 2.36734693877551, |
|
"grad_norm": 0.1905972808599472, |
|
"learning_rate": 7.493474677412794e-05, |
|
"loss": 1.2516, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 2.387755102040816, |
|
"grad_norm": 0.19716158509254456, |
|
"learning_rate": 7.330329010258483e-05, |
|
"loss": 1.2665, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 2.4081632653061225, |
|
"grad_norm": 0.1953389048576355, |
|
"learning_rate": 7.16794493317696e-05, |
|
"loss": 1.2661, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.4285714285714284, |
|
"grad_norm": 0.1990067958831787, |
|
"learning_rate": 7.006368770266421e-05, |
|
"loss": 1.2619, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 2.4489795918367347, |
|
"grad_norm": 0.1954919546842575, |
|
"learning_rate": 6.845646615147445e-05, |
|
"loss": 1.2736, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 2.4693877551020407, |
|
"grad_norm": 0.18382853269577026, |
|
"learning_rate": 6.685824317813643e-05, |
|
"loss": 1.2732, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 2.489795918367347, |
|
"grad_norm": 0.18729491531848907, |
|
"learning_rate": 6.526947471551798e-05, |
|
"loss": 1.2509, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 2.510204081632653, |
|
"grad_norm": 0.2034740000963211, |
|
"learning_rate": 6.369061399935255e-05, |
|
"loss": 1.2829, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.5306122448979593, |
|
"grad_norm": 0.1952620893716812, |
|
"learning_rate": 6.21221114389424e-05, |
|
"loss": 1.2689, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 2.5510204081632653, |
|
"grad_norm": 0.1986168622970581, |
|
"learning_rate": 6.0564414488668165e-05, |
|
"loss": 1.2644, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 2.571428571428571, |
|
"grad_norm": 0.19526751339435577, |
|
"learning_rate": 5.901796752034128e-05, |
|
"loss": 1.265, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 2.5918367346938775, |
|
"grad_norm": 0.195367693901062, |
|
"learning_rate": 5.748321169643596e-05, |
|
"loss": 1.2782, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 2.612244897959184, |
|
"grad_norm": 0.18351928889751434, |
|
"learning_rate": 5.596058484423656e-05, |
|
"loss": 1.2884, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.612244897959184, |
|
"eval_loss": 1.2471545934677124, |
|
"eval_runtime": 270.4953, |
|
"eval_samples_per_second": 6.266, |
|
"eval_steps_per_second": 3.135, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.63265306122449, |
|
"grad_norm": 0.2015760987997055, |
|
"learning_rate": 5.44505213309366e-05, |
|
"loss": 1.2536, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 2.6530612244897958, |
|
"grad_norm": 0.1734190732240677, |
|
"learning_rate": 5.2953451939724454e-05, |
|
"loss": 1.2628, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 2.673469387755102, |
|
"grad_norm": 0.214066281914711, |
|
"learning_rate": 5.146980374689192e-05, |
|
"loss": 1.2543, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 2.693877551020408, |
|
"grad_norm": 0.17507924139499664, |
|
"learning_rate": 5.000000000000002e-05, |
|
"loss": 1.2665, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 2.7142857142857144, |
|
"grad_norm": 0.1778109222650528, |
|
"learning_rate": 4.854445999713715e-05, |
|
"loss": 1.2789, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 2.7346938775510203, |
|
"grad_norm": 0.1856827288866043, |
|
"learning_rate": 4.710359896730379e-05, |
|
"loss": 1.2481, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 2.7551020408163263, |
|
"grad_norm": 0.17856694757938385, |
|
"learning_rate": 4.567782795195816e-05, |
|
"loss": 1.2732, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 2.7755102040816326, |
|
"grad_norm": 0.21598489582538605, |
|
"learning_rate": 4.426755368775637e-05, |
|
"loss": 1.2525, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 2.795918367346939, |
|
"grad_norm": 0.17308436334133148, |
|
"learning_rate": 4.287317849052075e-05, |
|
"loss": 1.2665, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 2.816326530612245, |
|
"grad_norm": 0.19207212328910828, |
|
"learning_rate": 4.149510014046922e-05, |
|
"loss": 1.2681, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.836734693877551, |
|
"grad_norm": 0.19626958668231964, |
|
"learning_rate": 4.013371176873849e-05, |
|
"loss": 1.2727, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 2.857142857142857, |
|
"grad_norm": 0.1986483484506607, |
|
"learning_rate": 3.878940174523371e-05, |
|
"loss": 1.2414, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 2.877551020408163, |
|
"grad_norm": 0.19369089603424072, |
|
"learning_rate": 3.746255356783632e-05, |
|
"loss": 1.254, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 2.877551020408163, |
|
"eval_loss": 1.2410293817520142, |
|
"eval_runtime": 270.6762, |
|
"eval_samples_per_second": 6.262, |
|
"eval_steps_per_second": 3.133, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 2.8979591836734695, |
|
"grad_norm": 0.20910531282424927, |
|
"learning_rate": 3.615354575300166e-05, |
|
"loss": 1.2541, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 2.9183673469387754, |
|
"grad_norm": 0.19536806643009186, |
|
"learning_rate": 3.4862751727777797e-05, |
|
"loss": 1.2517, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 2.938775510204082, |
|
"grad_norm": 0.18630966544151306, |
|
"learning_rate": 3.3590539723276083e-05, |
|
"loss": 1.2473, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 2.9591836734693877, |
|
"grad_norm": 0.1874723732471466, |
|
"learning_rate": 3.233727266962425e-05, |
|
"loss": 1.244, |
|
"step": 147 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 196, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 49, |
|
"total_flos": 3.0628052408991744e+16, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|