|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.99945085118067, |
|
"eval_steps": 114, |
|
"global_step": 455, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 0.1220703125, |
|
"learning_rate": 2e-05, |
|
"loss": 1.1845, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"eval_loss": 1.1820911169052124, |
|
"eval_runtime": 382.4714, |
|
"eval_samples_per_second": 7.676, |
|
"eval_steps_per_second": 3.838, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 0.1162109375, |
|
"learning_rate": 4e-05, |
|
"loss": 1.1777, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.115234375, |
|
"learning_rate": 6e-05, |
|
"loss": 1.1406, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.126953125, |
|
"learning_rate": 8e-05, |
|
"loss": 1.2366, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.150390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.2321, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.1513671875, |
|
"learning_rate": 0.00012, |
|
"loss": 1.0941, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.103515625, |
|
"learning_rate": 0.00014, |
|
"loss": 1.0609, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.10498046875, |
|
"learning_rate": 0.00016, |
|
"loss": 1.0332, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.0927734375, |
|
"learning_rate": 0.00018, |
|
"loss": 1.1241, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.1162109375, |
|
"learning_rate": 0.0002, |
|
"loss": 1.1964, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.150390625, |
|
"learning_rate": 0.00019999750800065415, |
|
"loss": 1.254, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.130859375, |
|
"learning_rate": 0.00019999003212681782, |
|
"loss": 1.0003, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.10888671875, |
|
"learning_rate": 0.00019997757275108847, |
|
"loss": 1.033, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.09619140625, |
|
"learning_rate": 0.00019996013049444118, |
|
"loss": 1.0915, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.09521484375, |
|
"learning_rate": 0.00019993770622619782, |
|
"loss": 1.027, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.08154296875, |
|
"learning_rate": 0.00019991030106398364, |
|
"loss": 0.8964, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.08203125, |
|
"learning_rate": 0.00019987791637367158, |
|
"loss": 0.9778, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.07666015625, |
|
"learning_rate": 0.00019984055376931413, |
|
"loss": 1.0588, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.076171875, |
|
"learning_rate": 0.00019979821511306308, |
|
"loss": 0.9921, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.0830078125, |
|
"learning_rate": 0.00019975090251507638, |
|
"loss": 1.0424, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.08642578125, |
|
"learning_rate": 0.0001996986183334134, |
|
"loss": 1.0514, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.07568359375, |
|
"learning_rate": 0.00019964136517391706, |
|
"loss": 1.0248, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.0927734375, |
|
"learning_rate": 0.00019957914589008405, |
|
"loss": 1.0394, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.07568359375, |
|
"learning_rate": 0.00019951196358292266, |
|
"loss": 1.0678, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.07861328125, |
|
"learning_rate": 0.0001994398216007982, |
|
"loss": 0.9298, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.095703125, |
|
"learning_rate": 0.00019936272353926615, |
|
"loss": 0.991, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.07958984375, |
|
"learning_rate": 0.00019928067324089286, |
|
"loss": 0.9823, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.08349609375, |
|
"learning_rate": 0.00019919367479506413, |
|
"loss": 1.0103, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.07568359375, |
|
"learning_rate": 0.00019910173253778136, |
|
"loss": 0.9725, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.07421875, |
|
"learning_rate": 0.00019900485105144543, |
|
"loss": 1.0453, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.0830078125, |
|
"learning_rate": 0.00019890303516462843, |
|
"loss": 1.0282, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.08544921875, |
|
"learning_rate": 0.00019879628995183272, |
|
"loss": 0.9683, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.0732421875, |
|
"learning_rate": 0.00019868462073323838, |
|
"loss": 0.9711, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.07666015625, |
|
"learning_rate": 0.00019856803307443781, |
|
"loss": 0.9303, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.0751953125, |
|
"learning_rate": 0.00019844653278615833, |
|
"loss": 0.9018, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.07470703125, |
|
"learning_rate": 0.0001983201259239728, |
|
"loss": 0.932, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.072265625, |
|
"learning_rate": 0.00019818881878799752, |
|
"loss": 0.9073, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.0791015625, |
|
"learning_rate": 0.00019805261792257847, |
|
"loss": 0.9367, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.0771484375, |
|
"learning_rate": 0.00019791153011596496, |
|
"loss": 1.0177, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.07666015625, |
|
"learning_rate": 0.00019776556239997146, |
|
"loss": 0.8887, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.08056640625, |
|
"learning_rate": 0.000197614722049627, |
|
"loss": 1.0279, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.08544921875, |
|
"learning_rate": 0.00019745901658281266, |
|
"loss": 1.0993, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.0751953125, |
|
"learning_rate": 0.00019729845375988694, |
|
"loss": 0.9436, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.08203125, |
|
"learning_rate": 0.00019713304158329873, |
|
"loss": 1.0442, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.08251953125, |
|
"learning_rate": 0.00019696278829718883, |
|
"loss": 1.0196, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.09130859375, |
|
"learning_rate": 0.00019678770238697874, |
|
"loss": 1.0415, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.08935546875, |
|
"learning_rate": 0.00019660779257894796, |
|
"loss": 0.983, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.07763671875, |
|
"learning_rate": 0.000196423067839799, |
|
"loss": 0.9415, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.09228515625, |
|
"learning_rate": 0.00019623353737621035, |
|
"loss": 0.9829, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.10009765625, |
|
"learning_rate": 0.00019603921063437793, |
|
"loss": 0.9223, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.08447265625, |
|
"learning_rate": 0.00019584009729954396, |
|
"loss": 0.8714, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.08837890625, |
|
"learning_rate": 0.00019563620729551445, |
|
"loss": 0.8981, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.103515625, |
|
"learning_rate": 0.00019542755078416456, |
|
"loss": 0.9528, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.08154296875, |
|
"learning_rate": 0.00019521413816493206, |
|
"loss": 1.0403, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.08251953125, |
|
"learning_rate": 0.0001949959800742991, |
|
"loss": 1.0177, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.09375, |
|
"learning_rate": 0.00019477308738526206, |
|
"loss": 1.0504, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.087890625, |
|
"learning_rate": 0.00019454547120678967, |
|
"loss": 0.9463, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.09130859375, |
|
"learning_rate": 0.00019431314288326926, |
|
"loss": 0.907, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.083984375, |
|
"learning_rate": 0.00019407611399394143, |
|
"loss": 0.8962, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.103515625, |
|
"learning_rate": 0.00019383439635232294, |
|
"loss": 0.9455, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.10009765625, |
|
"learning_rate": 0.0001935880020056179, |
|
"loss": 1.0383, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.0849609375, |
|
"learning_rate": 0.00019333694323411734, |
|
"loss": 0.9763, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.09130859375, |
|
"learning_rate": 0.00019308123255058708, |
|
"loss": 1.0526, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.08251953125, |
|
"learning_rate": 0.0001928208826996443, |
|
"loss": 0.8888, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.08740234375, |
|
"learning_rate": 0.00019255590665712214, |
|
"loss": 0.925, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.08203125, |
|
"learning_rate": 0.00019228631762942307, |
|
"loss": 0.8944, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.09130859375, |
|
"learning_rate": 0.00019201212905286074, |
|
"loss": 0.9638, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.09814453125, |
|
"learning_rate": 0.00019173335459299025, |
|
"loss": 0.9773, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.09130859375, |
|
"learning_rate": 0.00019145000814392696, |
|
"loss": 0.9241, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.091796875, |
|
"learning_rate": 0.0001911621038276542, |
|
"loss": 0.9495, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.09375, |
|
"learning_rate": 0.00019086965599331938, |
|
"loss": 0.9238, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.09326171875, |
|
"learning_rate": 0.00019057267921651866, |
|
"loss": 0.9721, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.09326171875, |
|
"learning_rate": 0.00019027118829857077, |
|
"loss": 0.9081, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.10546875, |
|
"learning_rate": 0.00018996519826577907, |
|
"loss": 0.9803, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.0830078125, |
|
"learning_rate": 0.00018965472436868286, |
|
"loss": 0.9435, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.09814453125, |
|
"learning_rate": 0.00018933978208129706, |
|
"loss": 0.8613, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.091796875, |
|
"learning_rate": 0.00018902038710034112, |
|
"loss": 0.9293, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.1015625, |
|
"learning_rate": 0.0001886965553444568, |
|
"loss": 1.0039, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.091796875, |
|
"learning_rate": 0.0001883683029534145, |
|
"loss": 0.9316, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.08154296875, |
|
"learning_rate": 0.00018803564628730915, |
|
"loss": 0.987, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.0908203125, |
|
"learning_rate": 0.0001876986019257446, |
|
"loss": 0.9598, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.1044921875, |
|
"learning_rate": 0.0001873571866670074, |
|
"loss": 0.872, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.0859375, |
|
"learning_rate": 0.00018701141752722964, |
|
"loss": 0.9035, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.095703125, |
|
"learning_rate": 0.0001866613117395407, |
|
"loss": 0.9103, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.087890625, |
|
"learning_rate": 0.00018630688675320842, |
|
"loss": 0.9446, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.08251953125, |
|
"learning_rate": 0.00018594816023276953, |
|
"loss": 0.8302, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.091796875, |
|
"learning_rate": 0.0001855851500571491, |
|
"loss": 0.9336, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.09521484375, |
|
"learning_rate": 0.00018521787431876954, |
|
"loss": 0.9285, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.08984375, |
|
"learning_rate": 0.0001848463513226488, |
|
"loss": 0.8675, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.09521484375, |
|
"learning_rate": 0.0001844705995854882, |
|
"loss": 0.9251, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.0859375, |
|
"learning_rate": 0.0001840906378347494, |
|
"loss": 0.8655, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.107421875, |
|
"learning_rate": 0.00018370648500772108, |
|
"loss": 0.9182, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.09521484375, |
|
"learning_rate": 0.0001833181602505751, |
|
"loss": 0.873, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.0947265625, |
|
"learning_rate": 0.00018292568291741228, |
|
"loss": 0.9838, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.099609375, |
|
"learning_rate": 0.00018252907256929775, |
|
"loss": 0.8565, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.0888671875, |
|
"learning_rate": 0.00018212834897328613, |
|
"loss": 0.8747, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.0888671875, |
|
"learning_rate": 0.0001817235321014361, |
|
"loss": 0.9313, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.099609375, |
|
"learning_rate": 0.0001813146421298154, |
|
"loss": 0.8365, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.10107421875, |
|
"learning_rate": 0.00018090169943749476, |
|
"loss": 1.0068, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.09228515625, |
|
"learning_rate": 0.00018048472460553257, |
|
"loss": 0.8783, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.0986328125, |
|
"learning_rate": 0.00018006373841594905, |
|
"loss": 0.9325, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.09619140625, |
|
"learning_rate": 0.00017963876185069034, |
|
"loss": 1.0326, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.1064453125, |
|
"learning_rate": 0.00017920981609058289, |
|
"loss": 0.8769, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.0869140625, |
|
"learning_rate": 0.00017877692251427782, |
|
"loss": 0.9167, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.1142578125, |
|
"learning_rate": 0.00017834010269718526, |
|
"loss": 0.9688, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.08740234375, |
|
"learning_rate": 0.00017789937841039918, |
|
"loss": 0.8039, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.08837890625, |
|
"learning_rate": 0.0001774547716196123, |
|
"loss": 0.9943, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.1064453125, |
|
"learning_rate": 0.00017700630448402125, |
|
"loss": 0.9769, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.09619140625, |
|
"learning_rate": 0.00017655399935522217, |
|
"loss": 0.8973, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.1025390625, |
|
"learning_rate": 0.0001760978787760968, |
|
"loss": 0.8602, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.09814453125, |
|
"learning_rate": 0.0001756379654796888, |
|
"loss": 0.8966, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.09521484375, |
|
"learning_rate": 0.00017517428238807085, |
|
"loss": 0.8997, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.0947265625, |
|
"learning_rate": 0.0001747068526112022, |
|
"loss": 0.9476, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.09814453125, |
|
"learning_rate": 0.00017423569944577678, |
|
"loss": 0.9328, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 0.922809362411499, |
|
"eval_runtime": 418.8489, |
|
"eval_samples_per_second": 7.01, |
|
"eval_steps_per_second": 3.505, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.09619140625, |
|
"learning_rate": 0.00017376084637406222, |
|
"loss": 0.8643, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.10302734375, |
|
"learning_rate": 0.0001732823170627294, |
|
"loss": 0.9659, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.0986328125, |
|
"learning_rate": 0.00017280013536167293, |
|
"loss": 0.8158, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.1025390625, |
|
"learning_rate": 0.00017231432530282246, |
|
"loss": 0.9843, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.08984375, |
|
"learning_rate": 0.000171824911098945, |
|
"loss": 0.9591, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.09423828125, |
|
"learning_rate": 0.00017133191714243805, |
|
"loss": 0.9331, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.103515625, |
|
"learning_rate": 0.00017083536800411393, |
|
"loss": 0.8647, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.0888671875, |
|
"learning_rate": 0.00017033528843197525, |
|
"loss": 1.0038, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.1005859375, |
|
"learning_rate": 0.00016983170334998132, |
|
"loss": 0.8214, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.08544921875, |
|
"learning_rate": 0.000169324637856806, |
|
"loss": 0.8832, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.0859375, |
|
"learning_rate": 0.00016881411722458688, |
|
"loss": 0.8642, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.09375, |
|
"learning_rate": 0.00016830016689766558, |
|
"loss": 0.9047, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.095703125, |
|
"learning_rate": 0.0001677828124913197, |
|
"loss": 0.8494, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.08837890625, |
|
"learning_rate": 0.00016726207979048604, |
|
"loss": 0.9421, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.09423828125, |
|
"learning_rate": 0.0001667379947484756, |
|
"loss": 0.8429, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.10400390625, |
|
"learning_rate": 0.00016621058348568007, |
|
"loss": 0.8319, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.09912109375, |
|
"learning_rate": 0.0001656798722882698, |
|
"loss": 0.8681, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.10498046875, |
|
"learning_rate": 0.00016514588760688397, |
|
"loss": 0.9469, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.091796875, |
|
"learning_rate": 0.00016460865605531213, |
|
"loss": 0.9818, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.10009765625, |
|
"learning_rate": 0.0001640682044091678, |
|
"loss": 0.9046, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.09326171875, |
|
"learning_rate": 0.00016352455960455387, |
|
"loss": 0.9319, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.09716796875, |
|
"learning_rate": 0.00016297774873672035, |
|
"loss": 0.963, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.103515625, |
|
"learning_rate": 0.00016242779905871374, |
|
"loss": 0.8749, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.09423828125, |
|
"learning_rate": 0.00016187473798001879, |
|
"loss": 0.9731, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.1015625, |
|
"learning_rate": 0.00016131859306519242, |
|
"loss": 0.8319, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.0908203125, |
|
"learning_rate": 0.0001607593920324899, |
|
"loss": 0.8156, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.10791015625, |
|
"learning_rate": 0.00016019716275248342, |
|
"loss": 0.9433, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.12451171875, |
|
"learning_rate": 0.0001596319332466729, |
|
"loss": 0.913, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.1005859375, |
|
"learning_rate": 0.00015906373168608952, |
|
"loss": 0.9255, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.10498046875, |
|
"learning_rate": 0.00015849258638989165, |
|
"loss": 0.9524, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.1015625, |
|
"learning_rate": 0.00015791852582395334, |
|
"loss": 1.0442, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.123046875, |
|
"learning_rate": 0.00015734157859944575, |
|
"loss": 0.8188, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.10107421875, |
|
"learning_rate": 0.00015676177347141094, |
|
"loss": 0.8675, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.09423828125, |
|
"learning_rate": 0.00015617913933732892, |
|
"loss": 0.9065, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.1064453125, |
|
"learning_rate": 0.00015559370523567732, |
|
"loss": 0.9665, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.1162109375, |
|
"learning_rate": 0.00015500550034448413, |
|
"loss": 0.8779, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.08837890625, |
|
"learning_rate": 0.0001544145539798734, |
|
"loss": 0.8539, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.1142578125, |
|
"learning_rate": 0.00015382089559460422, |
|
"loss": 0.9068, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.10498046875, |
|
"learning_rate": 0.00015322455477660273, |
|
"loss": 0.9563, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.09423828125, |
|
"learning_rate": 0.00015262556124748751, |
|
"loss": 0.8071, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.1123046875, |
|
"learning_rate": 0.0001520239448610882, |
|
"loss": 0.8971, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.103515625, |
|
"learning_rate": 0.00015141973560195768, |
|
"loss": 0.883, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.10498046875, |
|
"learning_rate": 0.0001508129635838775, |
|
"loss": 0.8609, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.10986328125, |
|
"learning_rate": 0.000150203659048357, |
|
"loss": 0.9184, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.109375, |
|
"learning_rate": 0.00014959185236312642, |
|
"loss": 0.8602, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.107421875, |
|
"learning_rate": 0.00014897757402062284, |
|
"loss": 0.8908, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.09521484375, |
|
"learning_rate": 0.00014836085463647088, |
|
"loss": 0.9134, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.109375, |
|
"learning_rate": 0.00014774172494795652, |
|
"loss": 0.9385, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.10888671875, |
|
"learning_rate": 0.00014712021581249533, |
|
"loss": 0.8183, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.09814453125, |
|
"learning_rate": 0.00014649635820609456, |
|
"loss": 0.8806, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.10009765625, |
|
"learning_rate": 0.00014587018322180905, |
|
"loss": 0.915, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.12109375, |
|
"learning_rate": 0.00014524172206819194, |
|
"loss": 0.848, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.10009765625, |
|
"learning_rate": 0.00014461100606773884, |
|
"loss": 0.9189, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.099609375, |
|
"learning_rate": 0.00014397806665532694, |
|
"loss": 0.78, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.10205078125, |
|
"learning_rate": 0.00014334293537664837, |
|
"loss": 0.8854, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.10009765625, |
|
"learning_rate": 0.0001427056438866376, |
|
"loss": 0.7996, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.10693359375, |
|
"learning_rate": 0.00014206622394789432, |
|
"loss": 0.9002, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.1083984375, |
|
"learning_rate": 0.00014142470742909975, |
|
"loss": 0.9172, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.10693359375, |
|
"learning_rate": 0.0001407811263034289, |
|
"loss": 0.9438, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.12890625, |
|
"learning_rate": 0.00014013551264695662, |
|
"loss": 0.8709, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.11328125, |
|
"learning_rate": 0.00013948789863705912, |
|
"loss": 0.8717, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.10205078125, |
|
"learning_rate": 0.00013883831655081018, |
|
"loss": 0.9313, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.11669921875, |
|
"learning_rate": 0.00013818679876337247, |
|
"loss": 0.914, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.1064453125, |
|
"learning_rate": 0.00013753337774638396, |
|
"loss": 0.919, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.10791015625, |
|
"learning_rate": 0.00013687808606633966, |
|
"loss": 0.9628, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.11376953125, |
|
"learning_rate": 0.00013622095638296826, |
|
"loss": 0.8947, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.11669921875, |
|
"learning_rate": 0.0001355620214476046, |
|
"loss": 0.9269, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.10595703125, |
|
"learning_rate": 0.00013490131410155728, |
|
"loss": 0.8438, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.11669921875, |
|
"learning_rate": 0.00013423886727447176, |
|
"loss": 0.8986, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.11474609375, |
|
"learning_rate": 0.00013357471398268918, |
|
"loss": 0.9274, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.111328125, |
|
"learning_rate": 0.000132908887327601, |
|
"loss": 0.8511, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.11767578125, |
|
"learning_rate": 0.00013224142049399895, |
|
"loss": 0.809, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.12158203125, |
|
"learning_rate": 0.00013157234674842128, |
|
"loss": 0.7523, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.11669921875, |
|
"learning_rate": 0.00013090169943749476, |
|
"loss": 0.9021, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.11865234375, |
|
"learning_rate": 0.00013022951198627254, |
|
"loss": 0.8875, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.09521484375, |
|
"learning_rate": 0.00012955581789656843, |
|
"loss": 0.8563, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.1123046875, |
|
"learning_rate": 0.000128880650745287, |
|
"loss": 0.7841, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.10107421875, |
|
"learning_rate": 0.0001282040441827503, |
|
"loss": 0.925, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.10107421875, |
|
"learning_rate": 0.0001275260319310205, |
|
"loss": 1.0102, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.115234375, |
|
"learning_rate": 0.00012684664778221942, |
|
"loss": 0.7835, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.10791015625, |
|
"learning_rate": 0.0001261659255968441, |
|
"loss": 0.9312, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.09716796875, |
|
"learning_rate": 0.0001254838993020793, |
|
"loss": 0.9214, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.10009765625, |
|
"learning_rate": 0.00012480060289010679, |
|
"loss": 0.8237, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.11279296875, |
|
"learning_rate": 0.00012411607041641062, |
|
"loss": 0.9152, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.1123046875, |
|
"learning_rate": 0.00012343033599808044, |
|
"loss": 0.9036, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.1416015625, |
|
"learning_rate": 0.00012274343381211066, |
|
"loss": 0.901, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.11962890625, |
|
"learning_rate": 0.00012205539809369719, |
|
"loss": 0.7381, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.111328125, |
|
"learning_rate": 0.00012136626313453134, |
|
"loss": 0.9358, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.11962890625, |
|
"learning_rate": 0.00012067606328109038, |
|
"loss": 0.8706, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.11572265625, |
|
"learning_rate": 0.00011998483293292602, |
|
"loss": 0.9957, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.10498046875, |
|
"learning_rate": 0.00011929260654094969, |
|
"loss": 0.8489, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.1279296875, |
|
"learning_rate": 0.0001185994186057158, |
|
"loss": 0.8718, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.09912109375, |
|
"learning_rate": 0.00011790530367570194, |
|
"loss": 0.8182, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.10498046875, |
|
"learning_rate": 0.00011721029634558709, |
|
"loss": 0.8512, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.150390625, |
|
"learning_rate": 0.00011651443125452759, |
|
"loss": 0.9216, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.1318359375, |
|
"learning_rate": 0.0001158177430844304, |
|
"loss": 0.7999, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.103515625, |
|
"learning_rate": 0.00011512026655822483, |
|
"loss": 0.7343, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.1005859375, |
|
"learning_rate": 0.00011442203643813183, |
|
"loss": 0.861, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.1220703125, |
|
"learning_rate": 0.00011372308752393144, |
|
"loss": 0.8731, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.1298828125, |
|
"learning_rate": 0.00011302345465122839, |
|
"loss": 0.9257, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.10791015625, |
|
"learning_rate": 0.00011232317268971585, |
|
"loss": 1.0089, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.1220703125, |
|
"learning_rate": 0.00011162227654143778, |
|
"loss": 0.877, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.0927734375, |
|
"learning_rate": 0.00011092080113904886, |
|
"loss": 0.8373, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.10693359375, |
|
"learning_rate": 0.00011021878144407408, |
|
"loss": 0.8474, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.1025390625, |
|
"learning_rate": 0.00010951625244516583, |
|
"loss": 0.788, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.1142578125, |
|
"learning_rate": 0.00010881324915636019, |
|
"loss": 0.8093, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.10986328125, |
|
"learning_rate": 0.0001081098066153319, |
|
"loss": 0.9342, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.10986328125, |
|
"learning_rate": 0.00010740595988164801, |
|
"loss": 0.836, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.09423828125, |
|
"learning_rate": 0.00010670174403502052, |
|
"loss": 0.8206, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.0986328125, |
|
"learning_rate": 0.000105997194173558, |
|
"loss": 0.8645, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.10107421875, |
|
"learning_rate": 0.00010529234541201631, |
|
"loss": 0.8027, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.11181640625, |
|
"learning_rate": 0.00010458723288004857, |
|
"loss": 0.8333, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.10400390625, |
|
"learning_rate": 0.00010388189172045406, |
|
"loss": 0.8999, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.1142578125, |
|
"learning_rate": 0.00010317635708742699, |
|
"loss": 0.8961, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"eval_loss": 0.8713043332099915, |
|
"eval_runtime": 415.0516, |
|
"eval_samples_per_second": 7.074, |
|
"eval_steps_per_second": 3.537, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.103515625, |
|
"learning_rate": 0.00010247066414480424, |
|
"loss": 0.8739, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.134765625, |
|
"learning_rate": 0.00010176484806431288, |
|
"loss": 0.8045, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.103515625, |
|
"learning_rate": 0.00010105894402381704, |
|
"loss": 0.9281, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.1025390625, |
|
"learning_rate": 0.00010035298720556493, |
|
"loss": 0.8426, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.11669921875, |
|
"learning_rate": 9.964701279443508e-05, |
|
"loss": 0.7591, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.11328125, |
|
"learning_rate": 9.894105597618296e-05, |
|
"loss": 0.8066, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.1064453125, |
|
"learning_rate": 9.823515193568715e-05, |
|
"loss": 0.8372, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.1015625, |
|
"learning_rate": 9.752933585519577e-05, |
|
"loss": 0.81, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.10986328125, |
|
"learning_rate": 9.682364291257304e-05, |
|
"loss": 1.0054, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.09814453125, |
|
"learning_rate": 9.611810827954599e-05, |
|
"loss": 0.8229, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.19140625, |
|
"learning_rate": 9.541276711995148e-05, |
|
"loss": 0.7994, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.1279296875, |
|
"learning_rate": 9.470765458798368e-05, |
|
"loss": 0.8103, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.1171875, |
|
"learning_rate": 9.400280582644203e-05, |
|
"loss": 0.7667, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.0966796875, |
|
"learning_rate": 9.329825596497949e-05, |
|
"loss": 0.9226, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.11279296875, |
|
"learning_rate": 9.259404011835201e-05, |
|
"loss": 0.751, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.10498046875, |
|
"learning_rate": 9.189019338466812e-05, |
|
"loss": 0.978, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.0986328125, |
|
"learning_rate": 9.118675084363986e-05, |
|
"loss": 0.8746, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.1318359375, |
|
"learning_rate": 9.048374755483419e-05, |
|
"loss": 0.8223, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.16015625, |
|
"learning_rate": 8.978121855592593e-05, |
|
"loss": 0.9778, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.134765625, |
|
"learning_rate": 8.907919886095115e-05, |
|
"loss": 0.8693, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.1201171875, |
|
"learning_rate": 8.837772345856226e-05, |
|
"loss": 0.7726, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.1494140625, |
|
"learning_rate": 8.767682731028415e-05, |
|
"loss": 0.7388, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.1337890625, |
|
"learning_rate": 8.697654534877165e-05, |
|
"loss": 0.8422, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.11572265625, |
|
"learning_rate": 8.627691247606862e-05, |
|
"loss": 0.8306, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.08984375, |
|
"learning_rate": 8.557796356186818e-05, |
|
"loss": 0.7276, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.1337890625, |
|
"learning_rate": 8.487973344177517e-05, |
|
"loss": 0.8244, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.10693359375, |
|
"learning_rate": 8.418225691556962e-05, |
|
"loss": 0.7665, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.09765625, |
|
"learning_rate": 8.348556874547242e-05, |
|
"loss": 0.8114, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.1318359375, |
|
"learning_rate": 8.278970365441292e-05, |
|
"loss": 0.7918, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.11669921875, |
|
"learning_rate": 8.20946963242981e-05, |
|
"loss": 0.8266, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.11083984375, |
|
"learning_rate": 8.140058139428425e-05, |
|
"loss": 0.7967, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.09912109375, |
|
"learning_rate": 8.070739345905032e-05, |
|
"loss": 0.9177, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.11474609375, |
|
"learning_rate": 8.001516706707401e-05, |
|
"loss": 0.9526, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.11083984375, |
|
"learning_rate": 7.932393671890965e-05, |
|
"loss": 0.9914, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.10205078125, |
|
"learning_rate": 7.863373686546867e-05, |
|
"loss": 0.8242, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.09716796875, |
|
"learning_rate": 7.794460190630282e-05, |
|
"loss": 0.6837, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.11328125, |
|
"learning_rate": 7.725656618788937e-05, |
|
"loss": 0.8758, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.11181640625, |
|
"learning_rate": 7.656966400191956e-05, |
|
"loss": 0.8655, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.11328125, |
|
"learning_rate": 7.58839295835894e-05, |
|
"loss": 1.0563, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.11962890625, |
|
"learning_rate": 7.519939710989325e-05, |
|
"loss": 0.832, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.1142578125, |
|
"learning_rate": 7.45161006979207e-05, |
|
"loss": 0.8198, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.1103515625, |
|
"learning_rate": 7.383407440315596e-05, |
|
"loss": 0.8246, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.10498046875, |
|
"learning_rate": 7.315335221778064e-05, |
|
"loss": 0.8314, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.1240234375, |
|
"learning_rate": 7.247396806897952e-05, |
|
"loss": 0.7107, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.10498046875, |
|
"learning_rate": 7.17959558172497e-05, |
|
"loss": 0.8982, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.1103515625, |
|
"learning_rate": 7.111934925471302e-05, |
|
"loss": 0.9427, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.1279296875, |
|
"learning_rate": 7.04441821034316e-05, |
|
"loss": 0.8454, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.115234375, |
|
"learning_rate": 6.97704880137275e-05, |
|
"loss": 0.8044, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.12451171875, |
|
"learning_rate": 6.909830056250527e-05, |
|
"loss": 0.8196, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.13671875, |
|
"learning_rate": 6.842765325157874e-05, |
|
"loss": 0.8221, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.119140625, |
|
"learning_rate": 6.775857950600106e-05, |
|
"loss": 0.8273, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.0966796875, |
|
"learning_rate": 6.7091112672399e-05, |
|
"loss": 0.9296, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.103515625, |
|
"learning_rate": 6.642528601731082e-05, |
|
"loss": 0.8635, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.10693359375, |
|
"learning_rate": 6.576113272552826e-05, |
|
"loss": 0.8622, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.10693359375, |
|
"learning_rate": 6.509868589844273e-05, |
|
"loss": 0.8982, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.1220703125, |
|
"learning_rate": 6.44379785523954e-05, |
|
"loss": 0.8564, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.1923828125, |
|
"learning_rate": 6.377904361703178e-05, |
|
"loss": 0.7286, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.1005859375, |
|
"learning_rate": 6.312191393366035e-05, |
|
"loss": 1.0057, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.12158203125, |
|
"learning_rate": 6.246662225361602e-05, |
|
"loss": 0.7258, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.099609375, |
|
"learning_rate": 6.181320123662755e-05, |
|
"loss": 0.8115, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.1298828125, |
|
"learning_rate": 6.116168344918983e-05, |
|
"loss": 0.7434, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.10009765625, |
|
"learning_rate": 6.051210136294089e-05, |
|
"loss": 0.875, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.10791015625, |
|
"learning_rate": 5.986448735304339e-05, |
|
"loss": 0.9115, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.1064453125, |
|
"learning_rate": 5.921887369657113e-05, |
|
"loss": 0.8492, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.126953125, |
|
"learning_rate": 5.857529257090027e-05, |
|
"loss": 0.7701, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.11083984375, |
|
"learning_rate": 5.7933776052105745e-05, |
|
"loss": 0.9333, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.1279296875, |
|
"learning_rate": 5.729435611336239e-05, |
|
"loss": 0.7718, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.123046875, |
|
"learning_rate": 5.6657064623351676e-05, |
|
"loss": 0.8522, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.1982421875, |
|
"learning_rate": 5.602193334467307e-05, |
|
"loss": 0.7019, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.11865234375, |
|
"learning_rate": 5.5388993932261215e-05, |
|
"loss": 0.8389, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.1259765625, |
|
"learning_rate": 5.4758277931808075e-05, |
|
"loss": 0.863, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.115234375, |
|
"learning_rate": 5.4129816778190936e-05, |
|
"loss": 0.7688, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.11328125, |
|
"learning_rate": 5.350364179390548e-05, |
|
"loss": 0.8084, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.1005859375, |
|
"learning_rate": 5.28797841875047e-05, |
|
"loss": 0.9221, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.12255859375, |
|
"learning_rate": 5.2258275052043546e-05, |
|
"loss": 0.8468, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.1181640625, |
|
"learning_rate": 5.163914536352918e-05, |
|
"loss": 0.918, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.11572265625, |
|
"learning_rate": 5.102242597937717e-05, |
|
"loss": 0.8958, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.126953125, |
|
"learning_rate": 5.040814763687358e-05, |
|
"loss": 0.8718, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.11376953125, |
|
"learning_rate": 4.9796340951642986e-05, |
|
"loss": 0.8603, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.11669921875, |
|
"learning_rate": 4.918703641612255e-05, |
|
"loss": 0.8245, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.12890625, |
|
"learning_rate": 4.858026439804235e-05, |
|
"loss": 0.8881, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.10791015625, |
|
"learning_rate": 4.797605513891179e-05, |
|
"loss": 0.921, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.232421875, |
|
"learning_rate": 4.737443875251251e-05, |
|
"loss": 0.8585, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.11474609375, |
|
"learning_rate": 4.6775445223397306e-05, |
|
"loss": 0.8024, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.1767578125, |
|
"learning_rate": 4.61791044053958e-05, |
|
"loss": 0.7499, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.130859375, |
|
"learning_rate": 4.558544602012663e-05, |
|
"loss": 0.8993, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.115234375, |
|
"learning_rate": 4.4994499655515865e-05, |
|
"loss": 0.8938, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.11083984375, |
|
"learning_rate": 4.440629476432267e-05, |
|
"loss": 0.8357, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.1328125, |
|
"learning_rate": 4.3820860662671107e-05, |
|
"loss": 0.719, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.1015625, |
|
"learning_rate": 4.323822652858911e-05, |
|
"loss": 0.8492, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.126953125, |
|
"learning_rate": 4.265842140055428e-05, |
|
"loss": 0.7805, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.11376953125, |
|
"learning_rate": 4.2081474176046646e-05, |
|
"loss": 0.814, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.1103515625, |
|
"learning_rate": 4.150741361010837e-05, |
|
"loss": 0.8448, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.1142578125, |
|
"learning_rate": 4.093626831391051e-05, |
|
"loss": 0.8212, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.1435546875, |
|
"learning_rate": 4.036806675332715e-05, |
|
"loss": 0.8006, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.111328125, |
|
"learning_rate": 3.98028372475166e-05, |
|
"loss": 0.8723, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.130859375, |
|
"learning_rate": 3.924060796751012e-05, |
|
"loss": 0.7984, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.10302734375, |
|
"learning_rate": 3.8681406934807585e-05, |
|
"loss": 0.8581, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.12060546875, |
|
"learning_rate": 3.8125262019981224e-05, |
|
"loss": 0.8222, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.12255859375, |
|
"learning_rate": 3.757220094128629e-05, |
|
"loss": 0.8624, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.12353515625, |
|
"learning_rate": 3.702225126327965e-05, |
|
"loss": 0.7377, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.1044921875, |
|
"learning_rate": 3.647544039544615e-05, |
|
"loss": 0.8021, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.125, |
|
"learning_rate": 3.5931795590832254e-05, |
|
"loss": 0.8053, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.10546875, |
|
"learning_rate": 3.53913439446879e-05, |
|
"loss": 1.0031, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.173828125, |
|
"learning_rate": 3.485411239311604e-05, |
|
"loss": 0.773, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.138671875, |
|
"learning_rate": 3.432012771173021e-05, |
|
"loss": 0.785, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.0986328125, |
|
"learning_rate": 3.378941651431996e-05, |
|
"loss": 0.7751, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.10595703125, |
|
"learning_rate": 3.326200525152441e-05, |
|
"loss": 0.7843, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.10302734375, |
|
"learning_rate": 3.2737920209514e-05, |
|
"loss": 0.8728, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.1279296875, |
|
"learning_rate": 3.2217187508680315e-05, |
|
"loss": 0.7532, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.1328125, |
|
"learning_rate": 3.1699833102334395e-05, |
|
"loss": 0.8074, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.0927734375, |
|
"learning_rate": 3.118588277541312e-05, |
|
"loss": 1.0021, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.1044921875, |
|
"learning_rate": 3.067536214319402e-05, |
|
"loss": 0.8352, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.11767578125, |
|
"learning_rate": 3.0168296650018735e-05, |
|
"loss": 0.824, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"eval_loss": 0.8436947464942932, |
|
"eval_runtime": 422.1114, |
|
"eval_samples_per_second": 6.956, |
|
"eval_steps_per_second": 3.478, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.138671875, |
|
"learning_rate": 2.966471156802477e-05, |
|
"loss": 0.9148, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.12353515625, |
|
"learning_rate": 2.9164631995886093e-05, |
|
"loss": 0.9928, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.1162109375, |
|
"learning_rate": 2.8668082857562005e-05, |
|
"loss": 0.8124, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.10107421875, |
|
"learning_rate": 2.8175088901055026e-05, |
|
"loss": 0.8412, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.10986328125, |
|
"learning_rate": 2.7685674697177566e-05, |
|
"loss": 0.7348, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.103515625, |
|
"learning_rate": 2.7199864638327077e-05, |
|
"loss": 0.7976, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.126953125, |
|
"learning_rate": 2.6717682937270604e-05, |
|
"loss": 0.7628, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.10693359375, |
|
"learning_rate": 2.6239153625937784e-05, |
|
"loss": 0.8783, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.1337890625, |
|
"learning_rate": 2.5764300554223242e-05, |
|
"loss": 0.6852, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.1181640625, |
|
"learning_rate": 2.529314738879781e-05, |
|
"loss": 0.8181, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.10693359375, |
|
"learning_rate": 2.4825717611929146e-05, |
|
"loss": 0.8616, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.119140625, |
|
"learning_rate": 2.4362034520311216e-05, |
|
"loss": 0.8029, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.126953125, |
|
"learning_rate": 2.390212122390323e-05, |
|
"loss": 0.8134, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.1298828125, |
|
"learning_rate": 2.3446000644777853e-05, |
|
"loss": 0.7766, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.107421875, |
|
"learning_rate": 2.2993695515978762e-05, |
|
"loss": 0.8786, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.1181640625, |
|
"learning_rate": 2.2545228380387706e-05, |
|
"loss": 0.8264, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.10986328125, |
|
"learning_rate": 2.210062158960081e-05, |
|
"loss": 0.8366, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.1025390625, |
|
"learning_rate": 2.1659897302814747e-05, |
|
"loss": 0.8282, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.10009765625, |
|
"learning_rate": 2.1223077485722198e-05, |
|
"loss": 0.9779, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.11181640625, |
|
"learning_rate": 2.0790183909417093e-05, |
|
"loss": 0.8431, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.12451171875, |
|
"learning_rate": 2.036123814930967e-05, |
|
"loss": 0.8402, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.10302734375, |
|
"learning_rate": 1.9936261584050973e-05, |
|
"loss": 0.9097, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.1083984375, |
|
"learning_rate": 1.9515275394467446e-05, |
|
"loss": 0.7467, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.11376953125, |
|
"learning_rate": 1.9098300562505266e-05, |
|
"loss": 0.9132, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.09765625, |
|
"learning_rate": 1.8685357870184605e-05, |
|
"loss": 0.9088, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.09130859375, |
|
"learning_rate": 1.8276467898563887e-05, |
|
"loss": 0.8388, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.09228515625, |
|
"learning_rate": 1.7871651026713908e-05, |
|
"loss": 0.703, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.1435546875, |
|
"learning_rate": 1.7470927430702277e-05, |
|
"loss": 0.8362, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.09375, |
|
"learning_rate": 1.7074317082587755e-05, |
|
"loss": 0.7667, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.185546875, |
|
"learning_rate": 1.668183974942491e-05, |
|
"loss": 0.8881, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.12158203125, |
|
"learning_rate": 1.6293514992278936e-05, |
|
"loss": 0.8335, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.12353515625, |
|
"learning_rate": 1.590936216525061e-05, |
|
"loss": 0.8181, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.1015625, |
|
"learning_rate": 1.5529400414511806e-05, |
|
"loss": 0.7918, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.12109375, |
|
"learning_rate": 1.5153648677351195e-05, |
|
"loss": 0.8674, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.1123046875, |
|
"learning_rate": 1.4782125681230497e-05, |
|
"loss": 0.7134, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.09228515625, |
|
"learning_rate": 1.4414849942850927e-05, |
|
"loss": 0.8532, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.1025390625, |
|
"learning_rate": 1.4051839767230478e-05, |
|
"loss": 0.7936, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.1025390625, |
|
"learning_rate": 1.3693113246791589e-05, |
|
"loss": 0.7971, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.123046875, |
|
"learning_rate": 1.333868826045932e-05, |
|
"loss": 0.7371, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.11962890625, |
|
"learning_rate": 1.2988582472770373e-05, |
|
"loss": 0.7324, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.10888671875, |
|
"learning_rate": 1.2642813332992608e-05, |
|
"loss": 0.7527, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.095703125, |
|
"learning_rate": 1.2301398074255443e-05, |
|
"loss": 0.8041, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.09814453125, |
|
"learning_rate": 1.196435371269089e-05, |
|
"loss": 0.9232, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.0947265625, |
|
"learning_rate": 1.163169704658551e-05, |
|
"loss": 0.875, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.11181640625, |
|
"learning_rate": 1.1303444655543206e-05, |
|
"loss": 0.9133, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.1064453125, |
|
"learning_rate": 1.0979612899658876e-05, |
|
"loss": 0.7921, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.1064453125, |
|
"learning_rate": 1.0660217918702965e-05, |
|
"loss": 0.7747, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.1171875, |
|
"learning_rate": 1.0345275631317163e-05, |
|
"loss": 0.8552, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.1005859375, |
|
"learning_rate": 1.0034801734220922e-05, |
|
"loss": 0.8868, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.09716796875, |
|
"learning_rate": 9.728811701429241e-06, |
|
"loss": 0.8623, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.123046875, |
|
"learning_rate": 9.427320783481353e-06, |
|
"loss": 0.8697, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.1279296875, |
|
"learning_rate": 9.130344006680657e-06, |
|
"loss": 0.749, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.087890625, |
|
"learning_rate": 8.837896172345827e-06, |
|
"loss": 0.8668, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.10009765625, |
|
"learning_rate": 8.549991856073069e-06, |
|
"loss": 0.9327, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.111328125, |
|
"learning_rate": 8.266645407009788e-06, |
|
"loss": 0.8025, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.10888671875, |
|
"learning_rate": 7.987870947139275e-06, |
|
"loss": 0.8903, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.11279296875, |
|
"learning_rate": 7.713682370576946e-06, |
|
"loss": 0.8081, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.12158203125, |
|
"learning_rate": 7.4440933428779e-06, |
|
"loss": 0.7926, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.10498046875, |
|
"learning_rate": 7.17911730035572e-06, |
|
"loss": 0.863, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.10400390625, |
|
"learning_rate": 6.9187674494129325e-06, |
|
"loss": 0.7536, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.11328125, |
|
"learning_rate": 6.663056765882692e-06, |
|
"loss": 0.7939, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.10693359375, |
|
"learning_rate": 6.4119979943821015e-06, |
|
"loss": 0.7258, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.11083984375, |
|
"learning_rate": 6.165603647677054e-06, |
|
"loss": 0.7775, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.09716796875, |
|
"learning_rate": 5.923886006058565e-06, |
|
"loss": 0.8807, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.09765625, |
|
"learning_rate": 5.6868571167307595e-06, |
|
"loss": 0.911, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.103515625, |
|
"learning_rate": 5.454528793210356e-06, |
|
"loss": 0.8141, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.10400390625, |
|
"learning_rate": 5.2269126147379555e-06, |
|
"loss": 0.8843, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.10107421875, |
|
"learning_rate": 5.00401992570092e-06, |
|
"loss": 0.7764, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.109375, |
|
"learning_rate": 4.785861835067962e-06, |
|
"loss": 0.7584, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.10595703125, |
|
"learning_rate": 4.5724492158354396e-06, |
|
"loss": 0.8061, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.11669921875, |
|
"learning_rate": 4.3637927044855475e-06, |
|
"loss": 0.8835, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.1123046875, |
|
"learning_rate": 4.159902700456053e-06, |
|
"loss": 0.8228, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.11376953125, |
|
"learning_rate": 3.960789365622075e-06, |
|
"loss": 0.7909, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.10693359375, |
|
"learning_rate": 3.766462623789646e-06, |
|
"loss": 0.9526, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.08837890625, |
|
"learning_rate": 3.57693216020103e-06, |
|
"loss": 0.8374, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.10693359375, |
|
"learning_rate": 3.3922074210520405e-06, |
|
"loss": 0.8091, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.1083984375, |
|
"learning_rate": 3.2122976130212646e-06, |
|
"loss": 0.886, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.1025390625, |
|
"learning_rate": 3.0372117028111825e-06, |
|
"loss": 0.7662, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.1171875, |
|
"learning_rate": 2.866958416701271e-06, |
|
"loss": 0.9222, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.103515625, |
|
"learning_rate": 2.7015462401130843e-06, |
|
"loss": 0.7954, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.09423828125, |
|
"learning_rate": 2.5409834171873482e-06, |
|
"loss": 0.7532, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.11962890625, |
|
"learning_rate": 2.3852779503730216e-06, |
|
"loss": 0.8322, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.12890625, |
|
"learning_rate": 2.2344376000285604e-06, |
|
"loss": 0.7971, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.1279296875, |
|
"learning_rate": 2.088469884035049e-06, |
|
"loss": 0.8257, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.13671875, |
|
"learning_rate": 1.9473820774215555e-06, |
|
"loss": 0.7701, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.12109375, |
|
"learning_rate": 1.8111812120024885e-06, |
|
"loss": 0.8043, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.10595703125, |
|
"learning_rate": 1.6798740760272103e-06, |
|
"loss": 0.7984, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.11767578125, |
|
"learning_rate": 1.553467213841664e-06, |
|
"loss": 0.8012, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.10400390625, |
|
"learning_rate": 1.4319669255622115e-06, |
|
"loss": 0.8198, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.1083984375, |
|
"learning_rate": 1.3153792667616183e-06, |
|
"loss": 0.8186, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.09912109375, |
|
"learning_rate": 1.2037100481672835e-06, |
|
"loss": 0.722, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.12158203125, |
|
"learning_rate": 1.0969648353715945e-06, |
|
"loss": 0.8273, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.10791015625, |
|
"learning_rate": 9.951489485545695e-07, |
|
"loss": 0.8434, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.1123046875, |
|
"learning_rate": 8.982674622186605e-07, |
|
"loss": 0.8882, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.130859375, |
|
"learning_rate": 8.063252049358982e-07, |
|
"loss": 0.9377, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.0986328125, |
|
"learning_rate": 7.193267591071529e-07, |
|
"loss": 0.8761, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.10595703125, |
|
"learning_rate": 6.372764607338599e-07, |
|
"loss": 0.8163, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.1005859375, |
|
"learning_rate": 5.60178399201805e-07, |
|
"loss": 0.8317, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.11474609375, |
|
"learning_rate": 4.880364170773533e-07, |
|
"loss": 0.8711, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.1005859375, |
|
"learning_rate": 4.208541099159691e-07, |
|
"loss": 0.7127, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.125, |
|
"learning_rate": 3.586348260829486e-07, |
|
"loss": 0.8064, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.10498046875, |
|
"learning_rate": 3.013816665865976e-07, |
|
"loss": 0.9065, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.11328125, |
|
"learning_rate": 2.490974849236216e-07, |
|
"loss": 0.7879, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.10498046875, |
|
"learning_rate": 2.0178488693695096e-07, |
|
"loss": 0.7726, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.11376953125, |
|
"learning_rate": 1.5944623068586685e-07, |
|
"loss": 0.8243, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.107421875, |
|
"learning_rate": 1.2208362632842862e-07, |
|
"loss": 0.8225, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.10595703125, |
|
"learning_rate": 8.969893601634694e-08, |
|
"loss": 0.7913, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.11865234375, |
|
"learning_rate": 6.229377380218005e-08, |
|
"loss": 0.7748, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.111328125, |
|
"learning_rate": 3.986950555883162e-08, |
|
"loss": 0.8305, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.125, |
|
"learning_rate": 2.242724891156067e-08, |
|
"loss": 0.8547, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.11181640625, |
|
"learning_rate": 9.96787318218173e-09, |
|
"loss": 1.0104, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.11865234375, |
|
"learning_rate": 2.4919993458549784e-09, |
|
"loss": 0.8115, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.146484375, |
|
"learning_rate": 0.0, |
|
"loss": 0.8365, |
|
"step": 455 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 455, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 500, |
|
"total_flos": 1.3577390229395866e+18, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|