|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.5608524957936063, |
|
"eval_steps": 500, |
|
"global_step": 2000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 1.9918, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 2.0258, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.2e-05, |
|
"loss": 2.0756, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 2.2086, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2e-05, |
|
"loss": 2.075, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.4e-05, |
|
"loss": 2.0893, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.8000000000000003e-05, |
|
"loss": 2.0577, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 3.2000000000000005e-05, |
|
"loss": 1.9745, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 3.6e-05, |
|
"loss": 1.9442, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4e-05, |
|
"loss": 1.9246, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4.4000000000000006e-05, |
|
"loss": 1.9296, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4.8e-05, |
|
"loss": 1.9183, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 5.2000000000000004e-05, |
|
"loss": 1.9778, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 5.6000000000000006e-05, |
|
"loss": 1.8426, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 6e-05, |
|
"loss": 1.8825, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 6.400000000000001e-05, |
|
"loss": 1.7846, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 6.800000000000001e-05, |
|
"loss": 1.8535, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7.2e-05, |
|
"loss": 1.6916, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7.6e-05, |
|
"loss": 1.7672, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8e-05, |
|
"loss": 1.8477, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.4e-05, |
|
"loss": 1.8787, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.800000000000001e-05, |
|
"loss": 1.8734, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.200000000000001e-05, |
|
"loss": 1.8889, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.6e-05, |
|
"loss": 1.8299, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0001, |
|
"loss": 1.7827, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00010400000000000001, |
|
"loss": 1.8557, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00010800000000000001, |
|
"loss": 1.9838, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00011200000000000001, |
|
"loss": 1.8365, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000116, |
|
"loss": 1.7932, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00012, |
|
"loss": 1.7293, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000124, |
|
"loss": 1.712, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00012800000000000002, |
|
"loss": 1.7439, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000132, |
|
"loss": 1.7217, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00013600000000000003, |
|
"loss": 1.7385, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00014, |
|
"loss": 1.842, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000144, |
|
"loss": 1.8133, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000148, |
|
"loss": 1.7923, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000152, |
|
"loss": 1.8135, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00015600000000000002, |
|
"loss": 1.8485, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00016, |
|
"loss": 1.9182, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000164, |
|
"loss": 1.7354, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000168, |
|
"loss": 1.5868, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000172, |
|
"loss": 1.7228, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00017600000000000002, |
|
"loss": 1.6714, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00018, |
|
"loss": 1.7548, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00018400000000000003, |
|
"loss": 1.7708, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000188, |
|
"loss": 1.6132, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000192, |
|
"loss": 1.6484, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000196, |
|
"loss": 1.8452, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0002, |
|
"loss": 1.7226, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00020400000000000003, |
|
"loss": 1.7263, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00020800000000000001, |
|
"loss": 1.7683, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00021200000000000003, |
|
"loss": 1.5194, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00021600000000000002, |
|
"loss": 1.8621, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00022000000000000003, |
|
"loss": 1.6884, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00022400000000000002, |
|
"loss": 1.6589, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00022799999999999999, |
|
"loss": 1.7475, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000232, |
|
"loss": 1.5911, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000236, |
|
"loss": 1.6961, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00024, |
|
"loss": 1.7497, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000244, |
|
"loss": 1.7571, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000248, |
|
"loss": 1.7656, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000252, |
|
"loss": 1.6999, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00025600000000000004, |
|
"loss": 1.6289, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00026000000000000003, |
|
"loss": 1.7333, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000264, |
|
"loss": 1.8245, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000268, |
|
"loss": 1.7294, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00027200000000000005, |
|
"loss": 1.8249, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000276, |
|
"loss": 1.7929, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00028, |
|
"loss": 1.6111, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000284, |
|
"loss": 1.775, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000288, |
|
"loss": 1.6014, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000292, |
|
"loss": 1.6855, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000296, |
|
"loss": 1.8239, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00030000000000000003, |
|
"loss": 1.7535, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000304, |
|
"loss": 1.6055, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000308, |
|
"loss": 1.8036, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00031200000000000005, |
|
"loss": 1.6589, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00031600000000000004, |
|
"loss": 1.7957, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00032, |
|
"loss": 1.7978, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000324, |
|
"loss": 1.6027, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000328, |
|
"loss": 1.6281, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000332, |
|
"loss": 1.6942, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000336, |
|
"loss": 1.7232, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00034, |
|
"loss": 1.605, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000344, |
|
"loss": 1.696, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000348, |
|
"loss": 1.5277, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00035200000000000005, |
|
"loss": 1.746, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00035600000000000003, |
|
"loss": 1.7568, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00036, |
|
"loss": 1.8807, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000364, |
|
"loss": 1.7793, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00036800000000000005, |
|
"loss": 1.7317, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00037200000000000004, |
|
"loss": 1.7291, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000376, |
|
"loss": 1.6806, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00038, |
|
"loss": 1.7072, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000384, |
|
"loss": 1.692, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000388, |
|
"loss": 1.6635, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000392, |
|
"loss": 1.7166, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00039600000000000003, |
|
"loss": 1.6075, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0004, |
|
"loss": 1.6545, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00039999991784337336, |
|
"loss": 1.5775, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0003999996713735609, |
|
"loss": 1.8225, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000399999260590765, |
|
"loss": 1.6641, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00039999868549532336, |
|
"loss": 1.6713, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0003999979460877083, |
|
"loss": 1.7508, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00039999704236852733, |
|
"loss": 1.6184, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00039999597433852294, |
|
"loss": 1.7434, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0003999947419985726, |
|
"loss": 1.8389, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0003999933453496887, |
|
"loss": 1.8842, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0003999917843930187, |
|
"loss": 1.821, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00039999005912984503, |
|
"loss": 1.6944, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00039998816956158514, |
|
"loss": 1.6341, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0003999861156897914, |
|
"loss": 1.8067, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0003999838975161512, |
|
"loss": 1.7007, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000399981515042487, |
|
"loss": 1.6112, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00039997896827075605, |
|
"loss": 1.6086, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0003999762572030507, |
|
"loss": 1.7449, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0003999733818415983, |
|
"loss": 1.6054, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00039997034218876125, |
|
"loss": 1.6986, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00039996713824703664, |
|
"loss": 1.7209, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00039996377001905685, |
|
"loss": 1.6468, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00039996023750758905, |
|
"loss": 1.6012, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0003999565407155355, |
|
"loss": 1.7693, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0003999526796459332, |
|
"loss": 1.6956, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039994865430195443, |
|
"loss": 1.6959, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039994446468690625, |
|
"loss": 1.6999, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039994011080423064, |
|
"loss": 1.6748, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039993559265750464, |
|
"loss": 1.6957, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003999309102504403, |
|
"loss": 1.6667, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039992606358688434, |
|
"loss": 1.8012, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003999210526708188, |
|
"loss": 1.702, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039991587750636036, |
|
"loss": 1.8419, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003999105380977608, |
|
"loss": 1.6793, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003999050344494068, |
|
"loss": 1.7067, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039989936656581996, |
|
"loss": 1.5846, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003998935344516569, |
|
"loss": 1.8016, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003998875381117089, |
|
"loss": 1.6084, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003998813775509025, |
|
"loss": 1.7251, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039987505277429904, |
|
"loss": 1.6121, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039986856378709454, |
|
"loss": 1.673, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003998619105946204, |
|
"loss": 1.5884, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039985509320234245, |
|
"loss": 1.6651, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003998481116158618, |
|
"loss": 1.787, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039984096584091413, |
|
"loss": 1.7694, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003998336558833703, |
|
"loss": 1.8221, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039982618174923577, |
|
"loss": 1.7721, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003998185434446512, |
|
"loss": 1.7399, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003998107409758919, |
|
"loss": 1.6257, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.000399802774349368, |
|
"loss": 1.7065, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003997946435716248, |
|
"loss": 1.6701, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039978634864934215, |
|
"loss": 1.6728, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039977788958933495, |
|
"loss": 1.7849, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003997692663985528, |
|
"loss": 1.6676, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003997604790840803, |
|
"loss": 1.845, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003997515276531368, |
|
"loss": 1.6801, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003997424121130764, |
|
"loss": 1.6767, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039973313247138823, |
|
"loss": 1.7125, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003997236887356961, |
|
"loss": 1.6456, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0003997140809137586, |
|
"loss": 1.5806, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00039970430901346935, |
|
"loss": 1.669, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003996943730428565, |
|
"loss": 1.754, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039968427301008305, |
|
"loss": 1.6931, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.000399674008923447, |
|
"loss": 1.7199, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039966358079138083, |
|
"loss": 1.6193, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039965298862245204, |
|
"loss": 1.5771, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003996422324253628, |
|
"loss": 1.7107, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003996313122089499, |
|
"loss": 1.6894, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003996202279821852, |
|
"loss": 1.6656, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.000399608979754175, |
|
"loss": 1.8858, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003995975675341606, |
|
"loss": 1.7983, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003995859913315178, |
|
"loss": 1.5216, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039957425115575714, |
|
"loss": 1.6359, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039956234701652414, |
|
"loss": 1.6685, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039955027892359866, |
|
"loss": 1.8074, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003995380468868956, |
|
"loss": 1.6036, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039952565091646425, |
|
"loss": 1.733, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003995130910224888, |
|
"loss": 1.7479, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039950036721528797, |
|
"loss": 1.6784, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039948747950531536, |
|
"loss": 1.8141, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003994744279031589, |
|
"loss": 1.7042, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003994612124195414, |
|
"loss": 1.6995, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003994478330653203, |
|
"loss": 1.7261, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039943428985148764, |
|
"loss": 1.58, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039942058278917, |
|
"loss": 1.7264, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039940671188962867, |
|
"loss": 1.8242, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003993926771642595, |
|
"loss": 1.6811, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039937847862459296, |
|
"loss": 1.6755, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003993641162822941, |
|
"loss": 1.6586, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039934959014916246, |
|
"loss": 1.6002, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003993349002371324, |
|
"loss": 1.6485, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003993200465582725, |
|
"loss": 1.7832, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003993050291247861, |
|
"loss": 1.6417, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039928984794901097, |
|
"loss": 1.6637, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039927450304341945, |
|
"loss": 1.6483, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0003992589944206185, |
|
"loss": 1.7407, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00039924332209334936, |
|
"loss": 1.663, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00039922748607448794, |
|
"loss": 1.747, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003992114863770446, |
|
"loss": 1.7124, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003991953230141641, |
|
"loss": 1.7536, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003991789959991258, |
|
"loss": 1.6898, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003991625053453433, |
|
"loss": 1.7433, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00039914585106636484, |
|
"loss": 1.7718, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00039912903317587304, |
|
"loss": 1.6753, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00039911205168768486, |
|
"loss": 1.8652, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003990949066157518, |
|
"loss": 1.6869, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00039907759797415946, |
|
"loss": 1.7137, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00039906012577712826, |
|
"loss": 1.6232, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003990424900390127, |
|
"loss": 1.6791, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003990246907743017, |
|
"loss": 1.7531, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003990067279976185, |
|
"loss": 1.7844, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003989886017237208, |
|
"loss": 1.6868, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003989703119675004, |
|
"loss": 1.678, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003989518587439837, |
|
"loss": 1.6909, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003989332420683311, |
|
"loss": 1.6467, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00039891446195583757, |
|
"loss": 1.6644, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003988955184219321, |
|
"loss": 1.6828, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00039887641148217813, |
|
"loss": 1.5632, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00039885714115227327, |
|
"loss": 1.6232, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003988377074480494, |
|
"loss": 1.5412, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003988181103854725, |
|
"loss": 1.8423, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003987983499806429, |
|
"loss": 1.6867, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003987784262497951, |
|
"loss": 1.6369, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00039875833920929784, |
|
"loss": 1.5102, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003987380888756538, |
|
"loss": 1.7809, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003987176752655001, |
|
"loss": 1.7746, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00039869709839560774, |
|
"loss": 1.679, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003986763582828821, |
|
"loss": 1.6352, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003986554549443625, |
|
"loss": 1.6803, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0003986343883972224, |
|
"loss": 1.5693, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00039861315865876947, |
|
"loss": 1.6777, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00039859176574644526, |
|
"loss": 1.6464, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003985702096778255, |
|
"loss": 1.4862, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003985484904706199, |
|
"loss": 1.7875, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00039852660814267225, |
|
"loss": 1.7406, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00039850456271196024, |
|
"loss": 1.6427, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003984823541965958, |
|
"loss": 1.6792, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003984599826148246, |
|
"loss": 1.6694, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00039843744798502643, |
|
"loss": 1.6611, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.000398414750325715, |
|
"loss": 1.779, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003983918896555379, |
|
"loss": 1.7927, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003983688659932766, |
|
"loss": 1.5586, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00039834567935784676, |
|
"loss": 1.7038, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003983223297682975, |
|
"loss": 1.5733, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00039829881724381223, |
|
"loss": 1.5646, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.000398275141803708, |
|
"loss": 1.639, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003982513034674358, |
|
"loss": 1.6839, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003982273022545803, |
|
"loss": 1.606, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00039820313818486017, |
|
"loss": 1.6646, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003981788112781278, |
|
"loss": 1.6461, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003981543215543693, |
|
"loss": 1.734, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00039812966903370457, |
|
"loss": 1.6804, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003981048537363874, |
|
"loss": 1.5808, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003980798756828051, |
|
"loss": 1.5425, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003980547348934789, |
|
"loss": 1.6667, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003980294313890635, |
|
"loss": 1.7531, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00039800396519034753, |
|
"loss": 1.7476, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003979783363182531, |
|
"loss": 1.7341, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.000397952544793836, |
|
"loss": 1.7615, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003979265906382857, |
|
"loss": 1.6179, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003979004738729253, |
|
"loss": 1.6777, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00039787419451921133, |
|
"loss": 1.5358, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00039784775259873415, |
|
"loss": 1.6684, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003978211481332175, |
|
"loss": 1.8019, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00039779438114451876, |
|
"loss": 1.6875, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003977674516546287, |
|
"loss": 1.7118, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00039774035968567176, |
|
"loss": 1.5797, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0003977131052599058, |
|
"loss": 1.6126, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039768568839972194, |
|
"loss": 1.6258, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003976581091276452, |
|
"loss": 1.4715, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003976303674663337, |
|
"loss": 1.5902, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003976024634385789, |
|
"loss": 1.595, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000397574397067306, |
|
"loss": 1.7204, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039754616837557334, |
|
"loss": 1.7199, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039751777738657266, |
|
"loss": 1.6953, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039748922412362895, |
|
"loss": 1.7156, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003974605086102007, |
|
"loss": 1.5859, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003974316308698796, |
|
"loss": 1.7613, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039740259092639065, |
|
"loss": 1.6515, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000397373388803592, |
|
"loss": 1.6348, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003973440245254752, |
|
"loss": 1.4916, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003973144981161649, |
|
"loss": 1.5288, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039728480959991903, |
|
"loss": 1.7296, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003972549590011287, |
|
"loss": 1.6089, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003972249463443181, |
|
"loss": 1.8133, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003971947716541447, |
|
"loss": 1.6188, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003971644349553989, |
|
"loss": 1.6767, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003971339362730044, |
|
"loss": 1.5592, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039710327563201786, |
|
"loss": 1.7247, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039707245305762895, |
|
"loss": 1.5534, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003970414685751606, |
|
"loss": 1.5649, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039701032221006846, |
|
"loss": 1.6316, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039697901398794144, |
|
"loss": 1.6102, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039694754393450137, |
|
"loss": 1.6398, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003969159120756028, |
|
"loss": 1.6802, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039688411843723356, |
|
"loss": 1.7493, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039685216304551416, |
|
"loss": 1.7303, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039682004592669814, |
|
"loss": 1.7573, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039678776710717174, |
|
"loss": 1.7138, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039675532661345416, |
|
"loss": 1.6963, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003967227244721976, |
|
"loss": 1.5269, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039668996071018657, |
|
"loss": 1.5682, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00039665703535433894, |
|
"loss": 1.668, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0003966239484317049, |
|
"loss": 1.7288, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0003965906999694677, |
|
"loss": 1.6777, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039655728999494295, |
|
"loss": 1.614, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039652371853557937, |
|
"loss": 1.7444, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.000396489985618958, |
|
"loss": 1.6656, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039645609127279276, |
|
"loss": 1.467, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039642203552493006, |
|
"loss": 1.7377, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039638781840334894, |
|
"loss": 1.6204, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039635343993616096, |
|
"loss": 1.5772, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0003963189001516105, |
|
"loss": 1.7176, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0003962841990780741, |
|
"loss": 1.527, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0003962493367440612, |
|
"loss": 1.6117, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0003962143131782132, |
|
"loss": 1.716, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039617912840930454, |
|
"loss": 1.648, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039614378246624176, |
|
"loss": 1.6771, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039610827537806394, |
|
"loss": 1.6652, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0003960726071739424, |
|
"loss": 1.6864, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039603677788318105, |
|
"loss": 1.6546, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0003960007875352159, |
|
"loss": 1.6717, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0003959646361596155, |
|
"loss": 1.5745, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039592832378608066, |
|
"loss": 1.5425, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039589185044444425, |
|
"loss": 1.5907, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0003958552161646716, |
|
"loss": 1.5331, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039581842097686023, |
|
"loss": 1.5927, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039578146491123977, |
|
"loss": 1.6573, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0003957443479981722, |
|
"loss": 1.6798, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0003957070702681513, |
|
"loss": 1.7501, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0003956696317518034, |
|
"loss": 1.6227, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0003956320324798866, |
|
"loss": 1.5463, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039559427248329134, |
|
"loss": 1.7951, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039555635179303977, |
|
"loss": 1.6577, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039551827044028635, |
|
"loss": 1.6136, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039548002845631746, |
|
"loss": 1.584, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039544162587255133, |
|
"loss": 1.559, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039540306272053835, |
|
"loss": 1.5896, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00039536433903196065, |
|
"loss": 1.6262, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003953254548386323, |
|
"loss": 1.4847, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003952864101724992, |
|
"loss": 1.6229, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00039524720506563916, |
|
"loss": 1.6223, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003952078395502618, |
|
"loss": 1.6402, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003951683136587085, |
|
"loss": 1.6644, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003951286274234524, |
|
"loss": 1.6631, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003950887808770983, |
|
"loss": 1.6568, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003950487740523829, |
|
"loss": 1.6874, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003950086069821743, |
|
"loss": 1.5275, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00039496827969947256, |
|
"loss": 1.5334, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00039492779223740913, |
|
"loss": 1.7514, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00039488714462924723, |
|
"loss": 1.5748, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003948463369083815, |
|
"loss": 1.6717, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003948053691083381, |
|
"loss": 1.6801, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003947642412627749, |
|
"loss": 1.5303, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003947229534054812, |
|
"loss": 1.5768, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003946815055703776, |
|
"loss": 1.6292, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00039463989779151633, |
|
"loss": 1.6216, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00039459813010308083, |
|
"loss": 1.4513, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003945562025393861, |
|
"loss": 1.6213, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00039451411513487846, |
|
"loss": 1.5435, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003944718679241354, |
|
"loss": 1.8213, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003944294609418658, |
|
"loss": 1.6161, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003943868942229099, |
|
"loss": 1.7734, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00039434416780223904, |
|
"loss": 1.5545, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003943012817149558, |
|
"loss": 1.6179, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003942582359962939, |
|
"loss": 1.6104, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00039421503068161833, |
|
"loss": 1.6037, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003941716658064251, |
|
"loss": 1.6444, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003941281414063412, |
|
"loss": 1.7601, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.000394084457517125, |
|
"loss": 1.7248, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003940406141746656, |
|
"loss": 1.5837, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0003939966114149833, |
|
"loss": 1.7561, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00039395244927422913, |
|
"loss": 1.7232, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00039390812778868534, |
|
"loss": 1.623, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.000393863646994765, |
|
"loss": 1.5596, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039381900692901184, |
|
"loss": 1.5719, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039377420762810086, |
|
"loss": 1.5753, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003937292491288375, |
|
"loss": 1.6194, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003936841314681582, |
|
"loss": 1.5417, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039363885468313007, |
|
"loss": 1.6827, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003935934188109511, |
|
"loss": 1.6674, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039354782388894974, |
|
"loss": 1.639, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003935020699545853, |
|
"loss": 1.5754, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039345615704544765, |
|
"loss": 1.8751, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039341008519925733, |
|
"loss": 1.5544, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003933638544538654, |
|
"loss": 1.6666, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003933174648472534, |
|
"loss": 1.5579, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003932709164175336, |
|
"loss": 1.4627, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003932242092029485, |
|
"loss": 1.5594, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003931773432418712, |
|
"loss": 1.7023, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039313031857280534, |
|
"loss": 1.7923, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003930831352343845, |
|
"loss": 1.5627, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003930357932653733, |
|
"loss": 1.6565, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039298829270466594, |
|
"loss": 1.6148, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003929406335912875, |
|
"loss": 1.745, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039289281596439305, |
|
"loss": 1.7138, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039284483986326785, |
|
"loss": 1.642, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003927967053273276, |
|
"loss": 1.7352, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003927484123961179, |
|
"loss": 1.7875, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039269996110931463, |
|
"loss": 1.4313, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003926513515067238, |
|
"loss": 1.7059, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039260258362828126, |
|
"loss": 1.6339, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003925536575140532, |
|
"loss": 1.7316, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039250457320423555, |
|
"loss": 1.4714, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003924553307391545, |
|
"loss": 1.8578, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039240593015926577, |
|
"loss": 1.6047, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039235637150515537, |
|
"loss": 1.6092, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039230665481753893, |
|
"loss": 1.709, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003922567801372621, |
|
"loss": 1.5493, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00039220674750530015, |
|
"loss": 1.6283, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0003921565569627582, |
|
"loss": 1.5762, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00039210620855087116, |
|
"loss": 1.4162, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003920557023110036, |
|
"loss": 1.6104, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00039200503828464966, |
|
"loss": 1.5302, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003919542165134333, |
|
"loss": 1.6919, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00039190323703910783, |
|
"loss": 1.6932, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003918520999035564, |
|
"loss": 1.6851, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003918008051487915, |
|
"loss": 1.6587, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003917493528169552, |
|
"loss": 1.6301, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.000391697742950319, |
|
"loss": 1.5199, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003916459755912837, |
|
"loss": 1.6334, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00039159405078237976, |
|
"loss": 1.474, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003915419685662669, |
|
"loss": 1.6662, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003914897289857339, |
|
"loss": 1.6301, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003914373320836992, |
|
"loss": 1.4839, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003913847779032103, |
|
"loss": 1.9532, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00039133206648744395, |
|
"loss": 1.6305, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.000391279197879706, |
|
"loss": 1.6361, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003912261721234316, |
|
"loss": 1.5287, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003911729892621849, |
|
"loss": 1.5012, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003911196493396591, |
|
"loss": 1.6288, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003910661523996765, |
|
"loss": 1.7189, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003910124984861884, |
|
"loss": 1.5912, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00039095868764327506, |
|
"loss": 1.5301, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003909047199151456, |
|
"loss": 1.5542, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00039085059534613815, |
|
"loss": 1.6419, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003907963139807196, |
|
"loss": 1.4787, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00039074187586348564, |
|
"loss": 1.6225, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003906872810391609, |
|
"loss": 1.674, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003906325295525985, |
|
"loss": 1.4997, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003905776214487806, |
|
"loss": 1.6527, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00039052255677281766, |
|
"loss": 1.7466, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003904673355699491, |
|
"loss": 1.5778, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003904119578855427, |
|
"loss": 1.6379, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00039035642376509497, |
|
"loss": 1.6003, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0003903007332542308, |
|
"loss": 1.6279, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003902448863987037, |
|
"loss": 1.8208, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003901888832443955, |
|
"loss": 1.6746, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003901327238373166, |
|
"loss": 1.5868, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003900764082236056, |
|
"loss": 1.7086, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00039001993644952943, |
|
"loss": 1.6933, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003899633085614836, |
|
"loss": 1.727, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003899065246059915, |
|
"loss": 1.6483, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.000389849584629705, |
|
"loss": 1.6562, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.000389792488679404, |
|
"loss": 1.6298, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003897352368019966, |
|
"loss": 1.7358, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003896778290445192, |
|
"loss": 1.4379, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00038962026545413586, |
|
"loss": 1.7434, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003895625460781389, |
|
"loss": 1.8016, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00038950467096394867, |
|
"loss": 1.6085, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003894466401591135, |
|
"loss": 1.6858, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00038938845371130936, |
|
"loss": 1.6548, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003893301116683404, |
|
"loss": 1.717, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003892716140781384, |
|
"loss": 1.6933, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003892129609887631, |
|
"loss": 1.7293, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003891541524484018, |
|
"loss": 1.6544, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003890951885053697, |
|
"loss": 1.67, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003890360692081094, |
|
"loss": 1.5251, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003889767946051916, |
|
"loss": 1.6306, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00038891736474531416, |
|
"loss": 1.6507, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00038885777967730264, |
|
"loss": 1.735, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00038879803945011016, |
|
"loss": 1.5742, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003887381441128173, |
|
"loss": 1.7325, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00038867809371463186, |
|
"loss": 1.5972, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00038861788830488956, |
|
"loss": 1.6424, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00038855752793305284, |
|
"loss": 1.7426, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003884970126487119, |
|
"loss": 1.7686, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.000388436342501584, |
|
"loss": 1.5441, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003883755175415136, |
|
"loss": 1.6817, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003883145378184726, |
|
"loss": 1.7718, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0003882534033825598, |
|
"loss": 1.5555, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00038819211428400125, |
|
"loss": 1.6318, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038813067057314994, |
|
"loss": 1.6014, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003880690723004859, |
|
"loss": 1.7192, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003880073195166164, |
|
"loss": 1.5524, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003879454122722752, |
|
"loss": 1.6834, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038788335061832333, |
|
"loss": 1.7585, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038782113460574853, |
|
"loss": 1.5802, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038775876428566534, |
|
"loss": 1.8165, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003876962397093152, |
|
"loss": 1.6848, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003876335609280661, |
|
"loss": 1.6112, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003875707279934129, |
|
"loss": 1.5237, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038750774095697697, |
|
"loss": 1.7105, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003874445998705063, |
|
"loss": 1.4533, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038738130478587553, |
|
"loss": 1.6991, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003873178557550858, |
|
"loss": 1.6408, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003872542528302646, |
|
"loss": 1.548, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.000387190496063666, |
|
"loss": 1.6202, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003871265855076704, |
|
"loss": 1.6177, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003870625212147845, |
|
"loss": 1.689, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003869983032376415, |
|
"loss": 1.7346, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003869339316290006, |
|
"loss": 1.6023, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038686940644174736, |
|
"loss": 1.5982, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038680472772889356, |
|
"loss": 1.6136, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038673989554357705, |
|
"loss": 1.5158, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003866749099390617, |
|
"loss": 1.5937, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038660977096873747, |
|
"loss": 1.5731, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003865444786861204, |
|
"loss": 1.5614, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003864790331448525, |
|
"loss": 1.6313, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003864134343987014, |
|
"loss": 1.6973, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038634768250156106, |
|
"loss": 1.7273, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003862817775074509, |
|
"loss": 1.7, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038621571947051623, |
|
"loss": 1.4817, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038614950844502807, |
|
"loss": 1.6953, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003860831444853833, |
|
"loss": 1.6213, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038601662764610424, |
|
"loss": 1.696, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0003859499579818388, |
|
"loss": 1.6227, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00038588313554736067, |
|
"loss": 1.7682, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003858161603975688, |
|
"loss": 1.5169, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038574903258748773, |
|
"loss": 1.5868, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003856817521722675, |
|
"loss": 1.4486, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038561431920718325, |
|
"loss": 1.684, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003855467337476358, |
|
"loss": 1.7448, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038547899584915105, |
|
"loss": 1.3673, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038541110556738014, |
|
"loss": 1.6268, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003853430629580994, |
|
"loss": 1.6764, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003852748680772104, |
|
"loss": 1.6594, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003852065209807397, |
|
"loss": 1.5929, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.000385138021724839, |
|
"loss": 1.7438, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.000385069370365785, |
|
"loss": 1.5752, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003850005669599793, |
|
"loss": 1.5398, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038493161156394855, |
|
"loss": 1.7006, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038486250423434403, |
|
"loss": 1.5635, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038479324502794214, |
|
"loss": 1.5738, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038472383400164374, |
|
"loss": 1.652, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038465427121247474, |
|
"loss": 1.6429, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003845845567175855, |
|
"loss": 1.6968, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038451469057425113, |
|
"loss": 1.6771, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038444467283987134, |
|
"loss": 1.6604, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038437450357197025, |
|
"loss": 1.663, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003843041828281965, |
|
"loss": 1.5601, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038423371066632347, |
|
"loss": 1.568, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003841630871442485, |
|
"loss": 1.5334, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003840923123199936, |
|
"loss": 1.5218, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038402138625170495, |
|
"loss": 1.6451, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.000383950308997653, |
|
"loss": 1.5404, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003838790806162325, |
|
"loss": 1.4505, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003838077011659622, |
|
"loss": 1.6547, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038373617070548504, |
|
"loss": 1.7092, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003836644892935681, |
|
"loss": 1.6728, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00038359265698910237, |
|
"loss": 1.488, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003835206738511029, |
|
"loss": 1.5505, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0003834485399387085, |
|
"loss": 1.6329, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.000383376255311182, |
|
"loss": 1.7233, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038330382002791014, |
|
"loss": 1.6345, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038323123414840313, |
|
"loss": 1.7528, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038315849773229514, |
|
"loss": 1.6351, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.000383085610839344, |
|
"loss": 1.7131, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003830125735294311, |
|
"loss": 1.609, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003829393858625614, |
|
"loss": 1.5494, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003828660478988634, |
|
"loss": 1.7247, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038279255969858914, |
|
"loss": 1.5116, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038271892132211413, |
|
"loss": 1.5377, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038264513282993697, |
|
"loss": 1.6662, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038257119428268, |
|
"loss": 1.5916, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003824971057410885, |
|
"loss": 1.6098, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038242286726603117, |
|
"loss": 1.6052, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003823484789184999, |
|
"loss": 1.6702, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038227394075960946, |
|
"loss": 1.7043, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003821992528505981, |
|
"loss": 1.7048, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038212441525282677, |
|
"loss": 1.6, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003820494280277795, |
|
"loss": 1.6999, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003819742912370633, |
|
"loss": 1.5814, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003818990049424081, |
|
"loss": 1.5547, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003818235692056664, |
|
"loss": 1.6049, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038174798408881385, |
|
"loss": 1.5417, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038167224965394856, |
|
"loss": 1.6171, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003815963659632913, |
|
"loss": 1.7906, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038152033307918564, |
|
"loss": 1.4961, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038144415106409763, |
|
"loss": 1.6419, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003813678199806159, |
|
"loss": 1.6146, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003812913398914513, |
|
"loss": 1.5949, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038121471085943745, |
|
"loss": 1.6703, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038113793294753014, |
|
"loss": 1.6016, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038106100621880756, |
|
"loss": 1.714, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038098393073647004, |
|
"loss": 1.6146, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00038090670656384023, |
|
"loss": 1.629, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003808293337643629, |
|
"loss": 1.6082, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0003807518124016049, |
|
"loss": 1.5346, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003806741425392552, |
|
"loss": 1.6732, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003805963242411247, |
|
"loss": 1.5394, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003805183575711464, |
|
"loss": 1.6558, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00038044024259337503, |
|
"loss": 1.5871, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00038036197937198717, |
|
"loss": 1.8108, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00038028356797128126, |
|
"loss": 1.5756, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00038020500845567745, |
|
"loss": 1.6226, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00038012630088971765, |
|
"loss": 1.7574, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003800474453380653, |
|
"loss": 1.5721, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003799684418655055, |
|
"loss": 1.6086, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037988929053694477, |
|
"loss": 1.4927, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003798099914174112, |
|
"loss": 1.7234, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003797305445720543, |
|
"loss": 1.7057, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003796509500661449, |
|
"loss": 1.7821, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037957120796507516, |
|
"loss": 1.7239, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037949131833435847, |
|
"loss": 1.5739, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003794112812396295, |
|
"loss": 1.6795, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.000379331096746644, |
|
"loss": 1.5626, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037925076492127885, |
|
"loss": 1.6654, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037917028582953204, |
|
"loss": 1.7856, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003790896595375223, |
|
"loss": 1.5479, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003790088861114897, |
|
"loss": 1.6803, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037892796561779475, |
|
"loss": 1.7133, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037884689812291915, |
|
"loss": 1.6756, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037876568369346505, |
|
"loss": 1.6747, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037868432239615573, |
|
"loss": 1.6242, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003786028142978347, |
|
"loss": 1.5968, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037852115946546627, |
|
"loss": 1.7603, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003784393579661354, |
|
"loss": 1.5431, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037835740986704743, |
|
"loss": 1.5807, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003782753152355281, |
|
"loss": 1.6421, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037819307413902345, |
|
"loss": 1.5383, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003781106866451003, |
|
"loss": 1.5435, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037802815282144526, |
|
"loss": 1.6208, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0003779454727358653, |
|
"loss": 1.6295, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00037786264645628767, |
|
"loss": 1.5912, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037777967405075956, |
|
"loss": 1.5793, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003776965555874484, |
|
"loss": 1.7148, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003776132911346415, |
|
"loss": 1.6304, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037752988076074606, |
|
"loss": 1.6453, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003774463245342893, |
|
"loss": 1.7158, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037736262252391817, |
|
"loss": 1.4112, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003772787747983993, |
|
"loss": 1.6531, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003771947814266194, |
|
"loss": 1.5022, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003771106424775844, |
|
"loss": 1.7515, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037702635802042015, |
|
"loss": 1.684, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037694192812437173, |
|
"loss": 1.5281, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003768573528588041, |
|
"loss": 1.8032, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003767726322932013, |
|
"loss": 1.457, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037668776649716696, |
|
"loss": 1.584, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003766027555404239, |
|
"loss": 1.5261, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037651759949281437, |
|
"loss": 1.7317, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003764322984242996, |
|
"loss": 1.5849, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003763468524049601, |
|
"loss": 1.4775, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037626126150499547, |
|
"loss": 1.7729, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037617552579472424, |
|
"loss": 1.7641, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.000376089645344584, |
|
"loss": 1.6569, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003760036202251312, |
|
"loss": 1.6777, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003759174505070413, |
|
"loss": 1.5653, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003758311362611083, |
|
"loss": 1.6499, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037574467755824515, |
|
"loss": 1.6806, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037565807446948344, |
|
"loss": 1.5205, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003755713270659732, |
|
"loss": 1.6806, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037548443541898324, |
|
"loss": 1.7035, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003753973995999009, |
|
"loss": 1.6304, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037531021968023175, |
|
"loss": 1.6656, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003752228957315999, |
|
"loss": 1.5005, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003751354278257478, |
|
"loss": 1.5847, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003750478160345361, |
|
"loss": 1.5535, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0003749600604299437, |
|
"loss": 1.544, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00037487216108406763, |
|
"loss": 1.7352, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.000374784118069123, |
|
"loss": 1.5405, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003746959314574431, |
|
"loss": 1.7449, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.000374607601321479, |
|
"loss": 1.5527, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037451912773379974, |
|
"loss": 1.6157, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003744305107670922, |
|
"loss": 1.5348, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037434175049416116, |
|
"loss": 1.5841, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037425284698792906, |
|
"loss": 1.5098, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003741638003214361, |
|
"loss": 1.6932, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003740746105678398, |
|
"loss": 1.6092, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003739852778004156, |
|
"loss": 1.6698, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003738958020925563, |
|
"loss": 1.5511, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003738061835177721, |
|
"loss": 1.5249, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003737164221496906, |
|
"loss": 1.6848, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037362651806205664, |
|
"loss": 1.6778, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003735364713287325, |
|
"loss": 1.6821, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037344628202369746, |
|
"loss": 1.6008, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003733559502210481, |
|
"loss": 1.6035, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003732654759949978, |
|
"loss": 1.6969, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003731748594198774, |
|
"loss": 1.5471, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003730841005701342, |
|
"loss": 1.5921, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037299319952033256, |
|
"loss": 1.7005, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037290215634515404, |
|
"loss": 1.6051, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037281097111939633, |
|
"loss": 1.6601, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003727196439179742, |
|
"loss": 1.6737, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037262817481591913, |
|
"loss": 1.6409, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037253656388837894, |
|
"loss": 1.6214, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037244481121061814, |
|
"loss": 1.663, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003723529168580175, |
|
"loss": 1.5129, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003722608809060745, |
|
"loss": 1.6899, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037216870343040266, |
|
"loss": 1.6275, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037207638450673184, |
|
"loss": 1.6723, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003719839242109083, |
|
"loss": 1.6748, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003718913226188942, |
|
"loss": 1.5452, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0003717985798067679, |
|
"loss": 1.572, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037170569585072384, |
|
"loss": 1.6967, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00037161267082707223, |
|
"loss": 1.555, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003715195048122394, |
|
"loss": 1.6263, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003714261978827673, |
|
"loss": 1.604, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00037133275011531376, |
|
"loss": 1.5761, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003712391615866524, |
|
"loss": 1.747, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00037114543237367237, |
|
"loss": 1.4509, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00037105156255337835, |
|
"loss": 1.4426, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003709575522028907, |
|
"loss": 1.5841, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003708634013994452, |
|
"loss": 1.6263, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003707691102203929, |
|
"loss": 1.6031, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00037067467874320024, |
|
"loss": 1.583, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003705801070454489, |
|
"loss": 1.6128, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003704853952048359, |
|
"loss": 1.3901, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00037039054329917324, |
|
"loss": 1.5793, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003702955514063881, |
|
"loss": 1.5687, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003702004196045225, |
|
"loss": 1.6644, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003701051479717336, |
|
"loss": 1.5907, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00037000973658629334, |
|
"loss": 1.5606, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003699141855265886, |
|
"loss": 1.6814, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003698184948711207, |
|
"loss": 1.6013, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00036972266469850593, |
|
"loss": 1.6585, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003696266950874753, |
|
"loss": 1.6623, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00036953058611687397, |
|
"loss": 1.5937, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003694343378656619, |
|
"loss": 1.6727, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00036933795041291354, |
|
"loss": 1.7429, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00036924142383781735, |
|
"loss": 1.772, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003691447582196765, |
|
"loss": 1.6947, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00036904795363790807, |
|
"loss": 1.8246, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00036895101017204345, |
|
"loss": 1.7672, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00036885392790172826, |
|
"loss": 1.5733, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00036875670690672183, |
|
"loss": 1.5593, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003686593472668978, |
|
"loss": 1.8336, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00036856184906224347, |
|
"loss": 1.5559, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0003684642123728601, |
|
"loss": 1.6823, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00036836643727896273, |
|
"loss": 1.442, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00036826852386088007, |
|
"loss": 1.7233, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00036817047219905446, |
|
"loss": 1.6053, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036807228237404184, |
|
"loss": 1.6173, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036797395446651164, |
|
"loss": 1.6574, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036787548855724683, |
|
"loss": 1.6933, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036777688472714365, |
|
"loss": 1.5255, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003676781430572116, |
|
"loss": 1.6014, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036757926362857355, |
|
"loss": 1.6, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003674802465224655, |
|
"loss": 1.6416, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036738109182023665, |
|
"loss": 1.6506, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003672817996033491, |
|
"loss": 1.5477, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003671823699533779, |
|
"loss": 1.3591, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003670828029520112, |
|
"loss": 1.5883, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036698309868104986, |
|
"loss": 1.6689, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003668832572224076, |
|
"loss": 1.6483, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036678327865811067, |
|
"loss": 1.5851, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003666831630702982, |
|
"loss": 1.577, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036658291054122164, |
|
"loss": 1.5954, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036648252115324524, |
|
"loss": 1.4855, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036638199498884553, |
|
"loss": 1.6042, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036628133213061136, |
|
"loss": 1.6123, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036618053266124386, |
|
"loss": 1.7188, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003660795966635566, |
|
"loss": 1.6181, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003659785242204752, |
|
"loss": 1.5388, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003658773154150372, |
|
"loss": 1.6865, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003657759703303926, |
|
"loss": 1.7088, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003656744890498028, |
|
"loss": 1.6922, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036557287165664166, |
|
"loss": 1.6053, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036547111823439447, |
|
"loss": 1.6599, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036536922886665847, |
|
"loss": 1.8624, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003652672036371424, |
|
"loss": 1.5775, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003651650426296669, |
|
"loss": 1.7386, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036506274592816395, |
|
"loss": 1.7108, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003649603136166771, |
|
"loss": 1.5253, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036485774577936123, |
|
"loss": 1.4999, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0003647550425004826, |
|
"loss": 1.5671, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00036465220386441887, |
|
"loss": 1.6967, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003645492299556586, |
|
"loss": 1.5612, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003644461208588019, |
|
"loss": 1.7352, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003643428766585596, |
|
"loss": 1.7234, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036423949743975365, |
|
"loss": 1.4075, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.000364135983287317, |
|
"loss": 1.6601, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003640323342862932, |
|
"loss": 1.6162, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003639285505218371, |
|
"loss": 1.4388, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003638246320792137, |
|
"loss": 1.5171, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003637205790437989, |
|
"loss": 1.6289, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003636163915010792, |
|
"loss": 1.6994, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036351206953665164, |
|
"loss": 1.7473, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036340761323622346, |
|
"loss": 1.7608, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036330302268561265, |
|
"loss": 1.6271, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036319829797074705, |
|
"loss": 1.6581, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036309343917766514, |
|
"loss": 1.6611, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036298844639251525, |
|
"loss": 1.3999, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036288331970155596, |
|
"loss": 1.5234, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036277805919115584, |
|
"loss": 1.7755, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036267266494779336, |
|
"loss": 1.5381, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003625671370580568, |
|
"loss": 1.6635, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003624614756086444, |
|
"loss": 1.6698, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036235568068636396, |
|
"loss": 1.5536, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036224975237813314, |
|
"loss": 1.6996, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.000362143690770979, |
|
"loss": 1.6837, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003620374959520382, |
|
"loss": 1.6303, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003619311680085567, |
|
"loss": 1.6647, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036182470702789, |
|
"loss": 1.6499, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036171811309750306, |
|
"loss": 1.6702, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036161138630496964, |
|
"loss": 1.6315, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003615045267379729, |
|
"loss": 1.6463, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036139753448430513, |
|
"loss": 1.6804, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00036129040963186744, |
|
"loss": 1.6333, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003611831522686701, |
|
"loss": 1.7529, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003610757624828321, |
|
"loss": 1.5498, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003609682403625813, |
|
"loss": 1.6344, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0003608605859962542, |
|
"loss": 1.5093, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00036075279947229603, |
|
"loss": 1.6603, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0003606448808792605, |
|
"loss": 1.6719, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00036053683030581, |
|
"loss": 1.5757, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00036042864784071514, |
|
"loss": 1.6758, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00036032033357285496, |
|
"loss": 1.6227, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.000360211887591217, |
|
"loss": 1.4965, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00036010330998489655, |
|
"loss": 1.6304, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0003599946008430974, |
|
"loss": 1.7211, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035988576025513153, |
|
"loss": 1.6257, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035977678831041837, |
|
"loss": 1.6938, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035966768509848585, |
|
"loss": 1.5939, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035955845070896933, |
|
"loss": 1.6475, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035944908523161215, |
|
"loss": 1.7269, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0003593395887562653, |
|
"loss": 1.6109, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035922996137288736, |
|
"loss": 1.6085, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0003591202031715446, |
|
"loss": 1.6714, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035901031424241046, |
|
"loss": 1.5724, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0003589002946757662, |
|
"loss": 1.5815, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.000358790144562, |
|
"loss": 1.5851, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0003586798639916076, |
|
"loss": 1.5545, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035856945305519177, |
|
"loss": 1.5106, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035845891184346237, |
|
"loss": 1.6366, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0003583482404472363, |
|
"loss": 1.4399, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035823743895743753, |
|
"loss": 1.6289, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035812650746509686, |
|
"loss": 1.6164, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035801544606135164, |
|
"loss": 1.5382, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0003579042548374465, |
|
"loss": 1.7615, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035779293388473214, |
|
"loss": 1.6598, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0003576814832946662, |
|
"loss": 1.3605, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035756990315881273, |
|
"loss": 1.6997, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035745819356884217, |
|
"loss": 1.6884, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0003573463546165314, |
|
"loss": 1.5922, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035723438639376345, |
|
"loss": 1.5222, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035712228899252774, |
|
"loss": 1.6046, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00035701006250491964, |
|
"loss": 1.5822, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0003568977070231406, |
|
"loss": 1.7044, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035678522263949823, |
|
"loss": 1.6725, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003566726094464058, |
|
"loss": 1.6955, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003565598675363826, |
|
"loss": 1.5768, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003564469970020534, |
|
"loss": 1.4117, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003563339979361491, |
|
"loss": 1.5941, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003562208704315057, |
|
"loss": 1.623, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.000356107614581065, |
|
"loss": 1.4523, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003559942304778742, |
|
"loss": 1.6319, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035588071821508586, |
|
"loss": 1.551, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035576707788595777, |
|
"loss": 1.711, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003556533095838531, |
|
"loss": 1.6893, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035553941340223994, |
|
"loss": 1.5213, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003554253894346916, |
|
"loss": 1.724, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003553112377748864, |
|
"loss": 1.7319, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035519695851660735, |
|
"loss": 1.6992, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003550825517537425, |
|
"loss": 1.6182, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035496801758028455, |
|
"loss": 1.4842, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003548533560903311, |
|
"loss": 1.6117, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003547385673780838, |
|
"loss": 1.527, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035462365153784953, |
|
"loss": 1.5368, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003545086086640391, |
|
"loss": 1.5789, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035439343885116786, |
|
"loss": 1.7224, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003542781421938555, |
|
"loss": 1.6041, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035416271878682584, |
|
"loss": 1.4644, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035404716872490684, |
|
"loss": 1.6175, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035393149210303056, |
|
"loss": 1.6211, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.000353815689016233, |
|
"loss": 1.5742, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035369975955965407, |
|
"loss": 1.6979, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003535837038285375, |
|
"loss": 1.6021, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003534675219182307, |
|
"loss": 1.7332, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003533512139241849, |
|
"loss": 1.601, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035323477994195484, |
|
"loss": 1.7034, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035311822006719865, |
|
"loss": 1.5706, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0003530015343956781, |
|
"loss": 1.5292, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00035288472302325805, |
|
"loss": 1.672, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00035276778604590695, |
|
"loss": 1.3831, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00035265072355969624, |
|
"loss": 1.6969, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003525335356608004, |
|
"loss": 1.572, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00035241622244549717, |
|
"loss": 1.5844, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00035229878401016714, |
|
"loss": 1.4933, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00035218122045129363, |
|
"loss": 1.5196, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00035206353186546303, |
|
"loss": 1.4318, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00035194571834936425, |
|
"loss": 1.5611, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003518277799997889, |
|
"loss": 1.5781, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003517097169136312, |
|
"loss": 1.628, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00035159152918788774, |
|
"loss": 1.5709, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003514732169196576, |
|
"loss": 1.5126, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00035135478020614216, |
|
"loss": 1.5469, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00035123621914464497, |
|
"loss": 1.5261, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00035111753383257186, |
|
"loss": 1.5995, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00035099872436743066, |
|
"loss": 1.7705, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003508797908468312, |
|
"loss": 1.6683, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003507607333684852, |
|
"loss": 1.5466, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003506415520302065, |
|
"loss": 1.7457, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003505222469299102, |
|
"loss": 1.6789, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003504028181656135, |
|
"loss": 1.5042, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003502832658354349, |
|
"loss": 1.6736, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00035016359003759464, |
|
"loss": 1.6903, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003500437908704144, |
|
"loss": 1.5225, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00034992386843231703, |
|
"loss": 1.5269, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003498038228218268, |
|
"loss": 1.5159, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00034968365413756914, |
|
"loss": 1.7589, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00034956336247827054, |
|
"loss": 1.7051, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003494429479427586, |
|
"loss": 1.4743, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003493224106299619, |
|
"loss": 1.7536, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00034920175063890975, |
|
"loss": 1.5885, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00034908096806873237, |
|
"loss": 1.5361, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00034896006301866065, |
|
"loss": 1.5788, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.000348839035588026, |
|
"loss": 1.7757, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0003487178858762606, |
|
"loss": 1.4703, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00034859661398289687, |
|
"loss": 1.6324, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034847522000756783, |
|
"loss": 1.5308, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003483537040500066, |
|
"loss": 1.6901, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003482320662100465, |
|
"loss": 1.5897, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034811030658762123, |
|
"loss": 1.7772, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003479884252827644, |
|
"loss": 1.6363, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003478664223956094, |
|
"loss": 1.5437, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003477442980263898, |
|
"loss": 1.6354, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003476220522754389, |
|
"loss": 1.6113, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003474996852431896, |
|
"loss": 1.5946, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034737719703017457, |
|
"loss": 1.7754, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.000347254587737026, |
|
"loss": 1.6842, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034713185746447554, |
|
"loss": 1.6545, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034700900631335426, |
|
"loss": 1.7102, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003468860343845925, |
|
"loss": 1.8011, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003467629417792198, |
|
"loss": 1.656, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.000346639728598365, |
|
"loss": 1.6627, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003465163949432558, |
|
"loss": 1.4086, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.000346392940915219, |
|
"loss": 1.5241, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003462693666156803, |
|
"loss": 1.6476, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034614567214616415, |
|
"loss": 1.5162, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034602185760829375, |
|
"loss": 1.6194, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003458979231037909, |
|
"loss": 1.6278, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003457738687344761, |
|
"loss": 1.4978, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034564969460226824, |
|
"loss": 1.5282, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003455254008091844, |
|
"loss": 1.6575, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034540098745734044, |
|
"loss": 1.8001, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034527645464895004, |
|
"loss": 1.639, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034515180248632514, |
|
"loss": 1.729, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003450270310718758, |
|
"loss": 1.728, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034490214050811, |
|
"loss": 1.549, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0003447771308976335, |
|
"loss": 1.6194, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034465200234315007, |
|
"loss": 1.6228, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034452675494746126, |
|
"loss": 1.4906, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034440138881346584, |
|
"loss": 1.5751, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00034427590404416056, |
|
"loss": 1.4345, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034415030074263944, |
|
"loss": 1.6488, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034402457901209387, |
|
"loss": 1.6703, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034389873895581264, |
|
"loss": 1.687, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034377278067718165, |
|
"loss": 1.5687, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.000343646704279684, |
|
"loss": 1.7186, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034352050986689983, |
|
"loss": 1.6859, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0003433941975425062, |
|
"loss": 1.5446, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.000343267767410277, |
|
"loss": 1.5113, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034314121957408297, |
|
"loss": 1.6428, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034301455413789165, |
|
"loss": 1.5125, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.000342887771205767, |
|
"loss": 1.5916, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034276087088186955, |
|
"loss": 1.583, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034263385327045647, |
|
"loss": 1.3318, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034250671847588097, |
|
"loss": 1.7858, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0003423794666025929, |
|
"loss": 1.5554, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0003422520977551379, |
|
"loss": 1.4468, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034212461203815813, |
|
"loss": 1.6589, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0003419970095563914, |
|
"loss": 1.6757, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034186929041467166, |
|
"loss": 1.6794, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034174145471792863, |
|
"loss": 1.6412, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0003416135025711879, |
|
"loss": 1.562, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034148543407957045, |
|
"loss": 1.6078, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0003413572493482932, |
|
"loss": 1.7199, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034122894848266833, |
|
"loss": 1.6804, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0003411005315881035, |
|
"loss": 1.6861, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0003409719987701017, |
|
"loss": 1.6702, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034084335013426123, |
|
"loss": 1.6035, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0003407145857862754, |
|
"loss": 1.5438, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034058570583193263, |
|
"loss": 1.6954, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034045671037711634, |
|
"loss": 1.7222, |
|
"step": 974 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034032759952780503, |
|
"loss": 1.6006, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0003401983733900715, |
|
"loss": 1.7374, |
|
"step": 976 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00034006903207008387, |
|
"loss": 1.6048, |
|
"step": 977 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00033993957567410445, |
|
"loss": 1.6602, |
|
"step": 978 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00033981000430849027, |
|
"loss": 1.474, |
|
"step": 979 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0003396803180796928, |
|
"loss": 1.4518, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003395505170942579, |
|
"loss": 1.8245, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003394206014588257, |
|
"loss": 1.5936, |
|
"step": 982 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033929057128013053, |
|
"loss": 1.5189, |
|
"step": 983 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003391604266650006, |
|
"loss": 1.632, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033903016772035853, |
|
"loss": 1.5355, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003388997945532206, |
|
"loss": 1.5453, |
|
"step": 986 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003387693072706971, |
|
"loss": 1.5479, |
|
"step": 987 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003386387059799918, |
|
"loss": 1.6015, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003385079907884025, |
|
"loss": 1.709, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003383771618033203, |
|
"loss": 1.6344, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003382462191322298, |
|
"loss": 1.709, |
|
"step": 991 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003381151628827092, |
|
"loss": 1.5598, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033798399316242993, |
|
"loss": 1.6272, |
|
"step": 993 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033785271007915653, |
|
"loss": 1.5783, |
|
"step": 994 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033772131374074674, |
|
"loss": 1.5724, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033758980425515136, |
|
"loss": 1.6668, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003374581817304142, |
|
"loss": 1.5015, |
|
"step": 997 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003373264462746719, |
|
"loss": 1.5056, |
|
"step": 998 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003371945979961537, |
|
"loss": 1.561, |
|
"step": 999 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003370626370031819, |
|
"loss": 1.5886, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003369305634041712, |
|
"loss": 1.7004, |
|
"step": 1001 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033679837730762863, |
|
"loss": 1.5864, |
|
"step": 1002 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.000336666078822154, |
|
"loss": 1.6418, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003365336680564392, |
|
"loss": 1.6663, |
|
"step": 1004 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033640114511926844, |
|
"loss": 1.4919, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003362685101195181, |
|
"loss": 1.5652, |
|
"step": 1006 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003361357631661567, |
|
"loss": 1.685, |
|
"step": 1007 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033600290436824453, |
|
"loss": 1.7768, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033586993383493406, |
|
"loss": 1.4661, |
|
"step": 1009 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033573685167546924, |
|
"loss": 1.4692, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033560365799918597, |
|
"loss": 1.6077, |
|
"step": 1011 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033547035291551163, |
|
"loss": 1.562, |
|
"step": 1012 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033533693653396517, |
|
"loss": 1.7198, |
|
"step": 1013 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033520340896415704, |
|
"loss": 1.5849, |
|
"step": 1014 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00033506977031578903, |
|
"loss": 1.6759, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0003349360206986541, |
|
"loss": 1.5835, |
|
"step": 1016 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003348021602226363, |
|
"loss": 1.5427, |
|
"step": 1017 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033466818899771113, |
|
"loss": 1.6485, |
|
"step": 1018 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003345341071339446, |
|
"loss": 1.507, |
|
"step": 1019 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033439991474149396, |
|
"loss": 1.6952, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003342656119306072, |
|
"loss": 1.6469, |
|
"step": 1021 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003341311988116228, |
|
"loss": 1.6973, |
|
"step": 1022 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033399667549497016, |
|
"loss": 1.5268, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003338620420911692, |
|
"loss": 1.484, |
|
"step": 1024 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033372729871082997, |
|
"loss": 1.5406, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033359244546465324, |
|
"loss": 1.6968, |
|
"step": 1026 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003334574824634298, |
|
"loss": 1.5439, |
|
"step": 1027 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033332240981804077, |
|
"loss": 1.7706, |
|
"step": 1028 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003331872276394572, |
|
"loss": 1.5416, |
|
"step": 1029 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033305193603874034, |
|
"loss": 1.5297, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003329165351270411, |
|
"loss": 1.6227, |
|
"step": 1031 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003327810250156002, |
|
"loss": 1.5251, |
|
"step": 1032 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003326454058157484, |
|
"loss": 1.5749, |
|
"step": 1033 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003325096776389058, |
|
"loss": 1.5676, |
|
"step": 1034 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.000332373840596582, |
|
"loss": 1.6995, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003322378948003762, |
|
"loss": 1.6477, |
|
"step": 1036 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033210184036197685, |
|
"loss": 1.5012, |
|
"step": 1037 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003319656773931617, |
|
"loss": 1.7402, |
|
"step": 1038 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003318294060057976, |
|
"loss": 1.6674, |
|
"step": 1039 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033169302631184066, |
|
"loss": 1.6402, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003315565384233357, |
|
"loss": 1.6748, |
|
"step": 1041 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003314199424524167, |
|
"loss": 1.6319, |
|
"step": 1042 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003312832385113061, |
|
"loss": 1.7178, |
|
"step": 1043 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003311464267123154, |
|
"loss": 1.6806, |
|
"step": 1044 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033100950716784456, |
|
"loss": 1.4169, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.000330872479990382, |
|
"loss": 1.3883, |
|
"step": 1046 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033073534529250463, |
|
"loss": 1.4829, |
|
"step": 1047 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0003305981031868777, |
|
"loss": 1.5919, |
|
"step": 1048 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033046075378625475, |
|
"loss": 1.6668, |
|
"step": 1049 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033032329720347737, |
|
"loss": 1.3224, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00033018573355147524, |
|
"loss": 1.5624, |
|
"step": 1051 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00033004806294326605, |
|
"loss": 1.3918, |
|
"step": 1052 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032991028549195526, |
|
"loss": 1.7946, |
|
"step": 1053 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003297724013107363, |
|
"loss": 1.9143, |
|
"step": 1054 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032963441051288994, |
|
"loss": 1.4799, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032949631321178504, |
|
"loss": 1.3832, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003293581095208775, |
|
"loss": 1.6206, |
|
"step": 1057 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003292197995537109, |
|
"loss": 1.6049, |
|
"step": 1058 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.000329081383423916, |
|
"loss": 1.6862, |
|
"step": 1059 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032894286124521076, |
|
"loss": 1.5244, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032880423313140045, |
|
"loss": 1.7469, |
|
"step": 1061 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032866549919637717, |
|
"loss": 1.708, |
|
"step": 1062 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003285266595541201, |
|
"loss": 1.6796, |
|
"step": 1063 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032838771431869515, |
|
"loss": 1.5523, |
|
"step": 1064 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003282486636042551, |
|
"loss": 1.7482, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032810950752503923, |
|
"loss": 1.7665, |
|
"step": 1066 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003279702461953736, |
|
"loss": 1.6417, |
|
"step": 1067 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003278308797296706, |
|
"loss": 1.6071, |
|
"step": 1068 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003276914082424289, |
|
"loss": 1.49, |
|
"step": 1069 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032755183184823367, |
|
"loss": 1.5762, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003274121506617562, |
|
"loss": 1.5535, |
|
"step": 1071 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003272723647977538, |
|
"loss": 1.6236, |
|
"step": 1072 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003271324743710697, |
|
"loss": 1.4387, |
|
"step": 1073 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032699247949663333, |
|
"loss": 1.7215, |
|
"step": 1074 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003268523802894596, |
|
"loss": 1.5572, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003267121768646495, |
|
"loss": 1.7069, |
|
"step": 1076 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032657186933738935, |
|
"loss": 1.6241, |
|
"step": 1077 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032643145782295096, |
|
"loss": 1.556, |
|
"step": 1078 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032629094243669175, |
|
"loss": 1.802, |
|
"step": 1079 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032615032329405454, |
|
"loss": 1.5444, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032600960051056715, |
|
"loss": 1.5606, |
|
"step": 1081 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003258687742018427, |
|
"loss": 1.5156, |
|
"step": 1082 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003257278444835793, |
|
"loss": 1.5738, |
|
"step": 1083 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032558681147156014, |
|
"loss": 1.7141, |
|
"step": 1084 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003254456752816532, |
|
"loss": 1.6297, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00032530443602981105, |
|
"loss": 1.5697, |
|
"step": 1086 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0003251630938320713, |
|
"loss": 1.5531, |
|
"step": 1087 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003250216488045558, |
|
"loss": 1.457, |
|
"step": 1088 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.000324880101063471, |
|
"loss": 1.4773, |
|
"step": 1089 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003247384507251078, |
|
"loss": 1.4973, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003245966979058413, |
|
"loss": 1.5522, |
|
"step": 1091 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032445484272213096, |
|
"loss": 1.5904, |
|
"step": 1092 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003243128852905201, |
|
"loss": 1.6103, |
|
"step": 1093 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032417082572763615, |
|
"loss": 1.5589, |
|
"step": 1094 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032402866415019043, |
|
"loss": 1.6806, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003238864006749783, |
|
"loss": 1.6955, |
|
"step": 1096 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032374403541887833, |
|
"loss": 1.5308, |
|
"step": 1097 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003236015684988532, |
|
"loss": 1.5559, |
|
"step": 1098 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032345900003194895, |
|
"loss": 1.531, |
|
"step": 1099 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003233163301352949, |
|
"loss": 1.5571, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032317355892610386, |
|
"loss": 1.4935, |
|
"step": 1101 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032303068652167195, |
|
"loss": 1.6009, |
|
"step": 1102 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003228877130393782, |
|
"loss": 1.7522, |
|
"step": 1103 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003227446385966848, |
|
"loss": 1.5671, |
|
"step": 1104 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.000322601463311137, |
|
"loss": 1.468, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032245818730036263, |
|
"loss": 1.4573, |
|
"step": 1106 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003223148106820725, |
|
"loss": 1.7848, |
|
"step": 1107 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032217133357406003, |
|
"loss": 1.7604, |
|
"step": 1108 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032202775609420115, |
|
"loss": 1.7273, |
|
"step": 1109 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032188407836045423, |
|
"loss": 1.6537, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032174030049086016, |
|
"loss": 1.4706, |
|
"step": 1111 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032159642260354184, |
|
"loss": 1.472, |
|
"step": 1112 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032145244481670464, |
|
"loss": 1.605, |
|
"step": 1113 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032130836724863577, |
|
"loss": 1.6392, |
|
"step": 1114 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032116419001770446, |
|
"loss": 1.747, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003210199132423619, |
|
"loss": 1.6071, |
|
"step": 1116 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003208755370411411, |
|
"loss": 1.6046, |
|
"step": 1117 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032073106153265656, |
|
"loss": 1.5948, |
|
"step": 1118 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032058648683560453, |
|
"loss": 1.5413, |
|
"step": 1119 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003204418130687627, |
|
"loss": 1.5848, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032029704035099017, |
|
"loss": 1.5756, |
|
"step": 1121 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0003201521688012273, |
|
"loss": 1.4922, |
|
"step": 1122 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00032000719853849573, |
|
"loss": 1.6451, |
|
"step": 1123 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031986212968189805, |
|
"loss": 1.6512, |
|
"step": 1124 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.000319716962350618, |
|
"loss": 1.8236, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031957169666392007, |
|
"loss": 1.7257, |
|
"step": 1126 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003194263327411497, |
|
"loss": 1.6207, |
|
"step": 1127 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003192808707017331, |
|
"loss": 1.6075, |
|
"step": 1128 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031913531066517687, |
|
"loss": 1.506, |
|
"step": 1129 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031898965275106825, |
|
"loss": 1.4684, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003188438970790748, |
|
"loss": 1.5826, |
|
"step": 1131 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003186980437689445, |
|
"loss": 1.5985, |
|
"step": 1132 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031855209294050555, |
|
"loss": 1.501, |
|
"step": 1133 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003184060447136662, |
|
"loss": 1.5743, |
|
"step": 1134 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031825989920841475, |
|
"loss": 1.5378, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031811365654481935, |
|
"loss": 1.6063, |
|
"step": 1136 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003179673168430282, |
|
"loss": 1.6827, |
|
"step": 1137 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031782088022326896, |
|
"loss": 1.6636, |
|
"step": 1138 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.000317674346805849, |
|
"loss": 1.4996, |
|
"step": 1139 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003175277167111552, |
|
"loss": 1.5111, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.000317380990059654, |
|
"loss": 1.6358, |
|
"step": 1141 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003172341669718911, |
|
"loss": 1.6654, |
|
"step": 1142 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031708724756849123, |
|
"loss": 1.5866, |
|
"step": 1143 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003169402319701585, |
|
"loss": 1.5739, |
|
"step": 1144 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.000316793120297676, |
|
"loss": 1.5787, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003166459126719057, |
|
"loss": 1.6545, |
|
"step": 1146 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003164986092137884, |
|
"loss": 1.7071, |
|
"step": 1147 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031635121004434366, |
|
"loss": 1.5618, |
|
"step": 1148 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031620371528466964, |
|
"loss": 1.3392, |
|
"step": 1149 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031605612505594314, |
|
"loss": 1.6128, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003159084394794192, |
|
"loss": 1.5969, |
|
"step": 1151 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031576065867643143, |
|
"loss": 1.5725, |
|
"step": 1152 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003156127827683914, |
|
"loss": 1.5855, |
|
"step": 1153 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003154648118767891, |
|
"loss": 1.558, |
|
"step": 1154 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031531674612319237, |
|
"loss": 1.5026, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0003151685856292471, |
|
"loss": 1.5938, |
|
"step": 1156 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031502033051667686, |
|
"loss": 1.5771, |
|
"step": 1157 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00031487198090728307, |
|
"loss": 1.564, |
|
"step": 1158 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031472353692294484, |
|
"loss": 1.551, |
|
"step": 1159 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031457499868561866, |
|
"loss": 1.7344, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003144263663173385, |
|
"loss": 1.6058, |
|
"step": 1161 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031427763994021585, |
|
"loss": 1.6153, |
|
"step": 1162 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003141288196764392, |
|
"loss": 1.6389, |
|
"step": 1163 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031397990564827424, |
|
"loss": 1.6039, |
|
"step": 1164 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003138308979780637, |
|
"loss": 1.5565, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031368179678822726, |
|
"loss": 1.453, |
|
"step": 1166 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003135326022012614, |
|
"loss": 1.6155, |
|
"step": 1167 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003133833143397395, |
|
"loss": 1.5345, |
|
"step": 1168 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003132339333263113, |
|
"loss": 1.5031, |
|
"step": 1169 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003130844592837033, |
|
"loss": 1.5716, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003129348923347182, |
|
"loss": 1.5669, |
|
"step": 1171 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003127852326022353, |
|
"loss": 1.6134, |
|
"step": 1172 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031263548020920987, |
|
"loss": 1.5915, |
|
"step": 1173 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003124856352786735, |
|
"loss": 1.7203, |
|
"step": 1174 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031233569793373366, |
|
"loss": 1.6924, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003121856682975739, |
|
"loss": 1.4656, |
|
"step": 1176 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031203554649345347, |
|
"loss": 1.5214, |
|
"step": 1177 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003118853326447074, |
|
"loss": 1.6814, |
|
"step": 1178 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031173502687474624, |
|
"loss": 1.5776, |
|
"step": 1179 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031158462930705623, |
|
"loss": 1.6242, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003114341400651989, |
|
"loss": 1.5993, |
|
"step": 1181 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003112835592728111, |
|
"loss": 1.6489, |
|
"step": 1182 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.000311132887053605, |
|
"loss": 1.5243, |
|
"step": 1183 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003109821235313678, |
|
"loss": 1.6148, |
|
"step": 1184 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031083126882996173, |
|
"loss": 1.5741, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031068032307332387, |
|
"loss": 1.6274, |
|
"step": 1186 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003105292863854662, |
|
"loss": 1.4296, |
|
"step": 1187 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031037815889047536, |
|
"loss": 1.6334, |
|
"step": 1188 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00031022694071251253, |
|
"loss": 1.7279, |
|
"step": 1189 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003100756319758137, |
|
"loss": 1.6015, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003099242328046887, |
|
"loss": 1.5406, |
|
"step": 1191 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00030977274332352224, |
|
"loss": 1.6772, |
|
"step": 1192 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00030962116365677286, |
|
"loss": 1.634, |
|
"step": 1193 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0003094694939289733, |
|
"loss": 1.5997, |
|
"step": 1194 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030931773426473027, |
|
"loss": 1.6244, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003091658847887244, |
|
"loss": 1.6461, |
|
"step": 1196 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030901394562571024, |
|
"loss": 1.7696, |
|
"step": 1197 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030886191690051565, |
|
"loss": 1.6668, |
|
"step": 1198 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030870979873804245, |
|
"loss": 1.5908, |
|
"step": 1199 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003085575912632658, |
|
"loss": 1.6645, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003084052946012341, |
|
"loss": 1.5892, |
|
"step": 1201 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003082529088770693, |
|
"loss": 1.5513, |
|
"step": 1202 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030810043421596625, |
|
"loss": 1.5125, |
|
"step": 1203 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030794787074319303, |
|
"loss": 1.5175, |
|
"step": 1204 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003077952185840907, |
|
"loss": 1.549, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003076424778640731, |
|
"loss": 1.5469, |
|
"step": 1206 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003074896487086268, |
|
"loss": 1.6703, |
|
"step": 1207 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003073367312433111, |
|
"loss": 1.5375, |
|
"step": 1208 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003071837255937578, |
|
"loss": 1.4652, |
|
"step": 1209 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030703063188567125, |
|
"loss": 1.5613, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030687745024482807, |
|
"loss": 1.7065, |
|
"step": 1211 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003067241807970771, |
|
"loss": 1.4361, |
|
"step": 1212 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030657082366833936, |
|
"loss": 1.5132, |
|
"step": 1213 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003064173789846079, |
|
"loss": 1.5481, |
|
"step": 1214 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030626384687194765, |
|
"loss": 1.6849, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003061102274564954, |
|
"loss": 1.3907, |
|
"step": 1216 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030595652086445977, |
|
"loss": 1.5077, |
|
"step": 1217 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030580272722212094, |
|
"loss": 1.4581, |
|
"step": 1218 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003056488466558304, |
|
"loss": 1.6817, |
|
"step": 1219 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003054948792920114, |
|
"loss": 1.5937, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030534082525715826, |
|
"loss": 1.5329, |
|
"step": 1221 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003051866846778366, |
|
"loss": 1.6112, |
|
"step": 1222 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003050324576806831, |
|
"loss": 1.5885, |
|
"step": 1223 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003048781443924054, |
|
"loss": 1.5476, |
|
"step": 1224 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030472374493978224, |
|
"loss": 1.6249, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030456925944966296, |
|
"loss": 1.6615, |
|
"step": 1226 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030441468804896755, |
|
"loss": 1.5339, |
|
"step": 1227 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00030426003086468667, |
|
"loss": 1.5914, |
|
"step": 1228 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003041052880238815, |
|
"loss": 1.642, |
|
"step": 1229 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0003039504596536835, |
|
"loss": 1.6709, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0003037955458812944, |
|
"loss": 1.5091, |
|
"step": 1231 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00030364054683398623, |
|
"loss": 1.7787, |
|
"step": 1232 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0003034854626391009, |
|
"loss": 1.5264, |
|
"step": 1233 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0003033302934240504, |
|
"loss": 1.5762, |
|
"step": 1234 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00030317503931631657, |
|
"loss": 1.6442, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00030301970044345077, |
|
"loss": 1.3079, |
|
"step": 1236 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0003028642769330743, |
|
"loss": 1.6531, |
|
"step": 1237 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00030270876891287787, |
|
"loss": 1.6456, |
|
"step": 1238 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00030255317651062166, |
|
"loss": 1.7472, |
|
"step": 1239 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0003023974998541351, |
|
"loss": 1.7043, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0003022417390713168, |
|
"loss": 1.5712, |
|
"step": 1241 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0003020858942901347, |
|
"loss": 1.5013, |
|
"step": 1242 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00030192996563862553, |
|
"loss": 1.6338, |
|
"step": 1243 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00030177395324489505, |
|
"loss": 1.5967, |
|
"step": 1244 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0003016178572371177, |
|
"loss": 1.464, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0003014616777435368, |
|
"loss": 1.6075, |
|
"step": 1246 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0003013054148924641, |
|
"loss": 1.5623, |
|
"step": 1247 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0003011490688122799, |
|
"loss": 1.5102, |
|
"step": 1248 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0003009926396314329, |
|
"loss": 1.6834, |
|
"step": 1249 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00030083612747844, |
|
"loss": 1.553, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00030067953248188626, |
|
"loss": 1.4659, |
|
"step": 1251 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0003005228547704249, |
|
"loss": 1.4358, |
|
"step": 1252 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00030036609447277703, |
|
"loss": 1.5063, |
|
"step": 1253 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00030020925171773163, |
|
"loss": 1.496, |
|
"step": 1254 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00030005232663414544, |
|
"loss": 1.3668, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00029989531935094277, |
|
"loss": 1.5428, |
|
"step": 1256 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00029973822999711556, |
|
"loss": 1.5547, |
|
"step": 1257 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00029958105870172305, |
|
"loss": 1.5979, |
|
"step": 1258 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0002994238055938919, |
|
"loss": 1.6185, |
|
"step": 1259 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00029926647080281597, |
|
"loss": 1.6565, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0002991090544577563, |
|
"loss": 1.6531, |
|
"step": 1261 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0002989515566880407, |
|
"loss": 1.6557, |
|
"step": 1262 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00029879397762306415, |
|
"loss": 1.5383, |
|
"step": 1263 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0002986363173922882, |
|
"loss": 1.7664, |
|
"step": 1264 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00029847857612524123, |
|
"loss": 1.5882, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002983207539515181, |
|
"loss": 1.6038, |
|
"step": 1266 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002981628510007802, |
|
"loss": 1.6021, |
|
"step": 1267 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029800486740275535, |
|
"loss": 1.6582, |
|
"step": 1268 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029784680328723745, |
|
"loss": 1.5992, |
|
"step": 1269 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002976886587840867, |
|
"loss": 1.506, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002975304340232292, |
|
"loss": 1.4982, |
|
"step": 1271 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029737212913465714, |
|
"loss": 1.5145, |
|
"step": 1272 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002972137442484285, |
|
"loss": 1.526, |
|
"step": 1273 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002970552794946669, |
|
"loss": 1.5751, |
|
"step": 1274 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002968967350035617, |
|
"loss": 1.7251, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002967381109053677, |
|
"loss": 1.6145, |
|
"step": 1276 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.000296579407330405, |
|
"loss": 1.654, |
|
"step": 1277 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002964206244090593, |
|
"loss": 1.5621, |
|
"step": 1278 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002962617622717811, |
|
"loss": 1.6457, |
|
"step": 1279 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029610282104908627, |
|
"loss": 1.5072, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002959438008715555, |
|
"loss": 1.4819, |
|
"step": 1281 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029578470186983455, |
|
"loss": 1.5347, |
|
"step": 1282 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029562552417463354, |
|
"loss": 1.586, |
|
"step": 1283 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029546626791672774, |
|
"loss": 1.4968, |
|
"step": 1284 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002953069332269566, |
|
"loss": 1.6464, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002951475202362241, |
|
"loss": 1.5809, |
|
"step": 1286 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002949880290754987, |
|
"loss": 1.582, |
|
"step": 1287 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002948284598758128, |
|
"loss": 1.8617, |
|
"step": 1288 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029466881276826317, |
|
"loss": 1.553, |
|
"step": 1289 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002945090878840105, |
|
"loss": 1.6381, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002943492853542793, |
|
"loss": 1.6688, |
|
"step": 1291 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.000294189405310358, |
|
"loss": 1.661, |
|
"step": 1292 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002940294478835986, |
|
"loss": 1.4943, |
|
"step": 1293 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029386941320541676, |
|
"loss": 1.6113, |
|
"step": 1294 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029370930140729156, |
|
"loss": 1.5637, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029354911262076546, |
|
"loss": 1.6185, |
|
"step": 1296 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029338884697744413, |
|
"loss": 1.6038, |
|
"step": 1297 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002932285046089965, |
|
"loss": 1.5863, |
|
"step": 1298 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029306808564715436, |
|
"loss": 1.4783, |
|
"step": 1299 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00029290759022371253, |
|
"loss": 1.8249, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002927470184705287, |
|
"loss": 1.5068, |
|
"step": 1301 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002925863705195231, |
|
"loss": 1.4926, |
|
"step": 1302 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002924256465026788, |
|
"loss": 1.8286, |
|
"step": 1303 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002922648465520411, |
|
"loss": 1.585, |
|
"step": 1304 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00029210397079971795, |
|
"loss": 1.4951, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002919430193778793, |
|
"loss": 1.6094, |
|
"step": 1306 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00029178199241875754, |
|
"loss": 1.6628, |
|
"step": 1307 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00029162089005464693, |
|
"loss": 1.6924, |
|
"step": 1308 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002914597124179037, |
|
"loss": 1.4482, |
|
"step": 1309 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00029129845964094597, |
|
"loss": 1.5083, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002911371318562536, |
|
"loss": 1.6582, |
|
"step": 1311 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00029097572919636807, |
|
"loss": 1.5422, |
|
"step": 1312 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00029081425179389235, |
|
"loss": 1.5767, |
|
"step": 1313 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00029065269978149086, |
|
"loss": 1.6507, |
|
"step": 1314 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002904910732918892, |
|
"loss": 1.5284, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00029032937245787436, |
|
"loss": 1.6317, |
|
"step": 1316 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00029016759741229414, |
|
"loss": 1.5418, |
|
"step": 1317 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002900057482880576, |
|
"loss": 1.6465, |
|
"step": 1318 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00028984382521813446, |
|
"loss": 1.4781, |
|
"step": 1319 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002896818283355552, |
|
"loss": 1.6437, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002895197577734112, |
|
"loss": 1.4823, |
|
"step": 1321 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00028935761366485386, |
|
"loss": 1.5966, |
|
"step": 1322 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002891953961430955, |
|
"loss": 1.6798, |
|
"step": 1323 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00028903310534140857, |
|
"loss": 1.6863, |
|
"step": 1324 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00028887074139312565, |
|
"loss": 1.6297, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002887083044316395, |
|
"loss": 1.5969, |
|
"step": 1326 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002885457945904029, |
|
"loss": 1.7106, |
|
"step": 1327 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00028838321200292836, |
|
"loss": 1.4536, |
|
"step": 1328 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00028822055680278834, |
|
"loss": 1.5529, |
|
"step": 1329 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002880578291236148, |
|
"loss": 1.6485, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00028789502909909936, |
|
"loss": 1.562, |
|
"step": 1331 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00028773215686299304, |
|
"loss": 1.7625, |
|
"step": 1332 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002875692125491061, |
|
"loss": 1.5392, |
|
"step": 1333 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002874061962913081, |
|
"loss": 1.4792, |
|
"step": 1334 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002872431082235278, |
|
"loss": 1.6575, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002870799484797527, |
|
"loss": 1.661, |
|
"step": 1336 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002869167171940295, |
|
"loss": 1.5613, |
|
"step": 1337 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028675341450046346, |
|
"loss": 1.5619, |
|
"step": 1338 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002865900405332186, |
|
"loss": 1.6638, |
|
"step": 1339 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002864265954265173, |
|
"loss": 1.6187, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028626307931464066, |
|
"loss": 1.4234, |
|
"step": 1341 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028609949233192805, |
|
"loss": 1.5031, |
|
"step": 1342 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.000285935834612777, |
|
"loss": 1.3451, |
|
"step": 1343 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.000285772106291643, |
|
"loss": 1.401, |
|
"step": 1344 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028560830750304, |
|
"loss": 1.5241, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028544443838153933, |
|
"loss": 1.6803, |
|
"step": 1346 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002852804990617705, |
|
"loss": 1.5534, |
|
"step": 1347 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028511648967842034, |
|
"loss": 1.6425, |
|
"step": 1348 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028495241036623355, |
|
"loss": 1.5895, |
|
"step": 1349 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028478826126001217, |
|
"loss": 1.607, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002846240424946155, |
|
"loss": 1.5567, |
|
"step": 1351 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002844597542049602, |
|
"loss": 1.6499, |
|
"step": 1352 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002842953965260199, |
|
"loss": 1.4521, |
|
"step": 1353 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002841309695928254, |
|
"loss": 1.6268, |
|
"step": 1354 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002839664735404644, |
|
"loss": 1.4933, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028380190850408113, |
|
"loss": 1.6719, |
|
"step": 1356 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002836372746188768, |
|
"loss": 1.2667, |
|
"step": 1357 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002834725720201089, |
|
"loss": 1.5909, |
|
"step": 1358 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002833078008430917, |
|
"loss": 1.7232, |
|
"step": 1359 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002831429612231956, |
|
"loss": 1.546, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028297805329584725, |
|
"loss": 1.5954, |
|
"step": 1361 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002828130771965294, |
|
"loss": 1.5687, |
|
"step": 1362 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028264803306078093, |
|
"loss": 1.5193, |
|
"step": 1363 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028248292102419655, |
|
"loss": 1.6457, |
|
"step": 1364 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028231774122242656, |
|
"loss": 1.699, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002821524937911773, |
|
"loss": 1.5789, |
|
"step": 1366 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002819871788662104, |
|
"loss": 1.7219, |
|
"step": 1367 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002818217965833431, |
|
"loss": 1.538, |
|
"step": 1368 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002816563470784478, |
|
"loss": 1.5908, |
|
"step": 1369 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002814908304874523, |
|
"loss": 1.5928, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028132524694633943, |
|
"loss": 1.5533, |
|
"step": 1371 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00028115959659114704, |
|
"loss": 1.4639, |
|
"step": 1372 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002809938795579679, |
|
"loss": 1.5845, |
|
"step": 1373 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00028082809598294946, |
|
"loss": 1.6648, |
|
"step": 1374 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.000280662246002294, |
|
"loss": 1.6132, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00028049632975225825, |
|
"loss": 1.5049, |
|
"step": 1376 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002803303473691534, |
|
"loss": 1.6983, |
|
"step": 1377 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00028016429898934496, |
|
"loss": 1.529, |
|
"step": 1378 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00027999818474925257, |
|
"loss": 1.5246, |
|
"step": 1379 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002798320047853503, |
|
"loss": 1.6582, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002796657592341659, |
|
"loss": 1.6139, |
|
"step": 1381 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002794994482322811, |
|
"loss": 1.6289, |
|
"step": 1382 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002793330719163314, |
|
"loss": 1.512, |
|
"step": 1383 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.000279166630423006, |
|
"loss": 1.6038, |
|
"step": 1384 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00027900012388904757, |
|
"loss": 1.8598, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002788335524512523, |
|
"loss": 1.6253, |
|
"step": 1386 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002786669162464697, |
|
"loss": 1.4944, |
|
"step": 1387 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002785002154116024, |
|
"loss": 1.8054, |
|
"step": 1388 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00027833345008360615, |
|
"loss": 1.5125, |
|
"step": 1389 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002781666203994899, |
|
"loss": 1.5585, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.000277999726496315, |
|
"loss": 1.6857, |
|
"step": 1391 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002778327685111961, |
|
"loss": 1.6913, |
|
"step": 1392 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00027766574658130006, |
|
"loss": 1.4969, |
|
"step": 1393 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002774986608438466, |
|
"loss": 1.6242, |
|
"step": 1394 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002773315114361077, |
|
"loss": 1.5628, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00027716429849540764, |
|
"loss": 1.5391, |
|
"step": 1396 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002769970221591229, |
|
"loss": 1.4453, |
|
"step": 1397 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00027682968256468227, |
|
"loss": 1.5723, |
|
"step": 1398 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00027666227984956605, |
|
"loss": 1.4469, |
|
"step": 1399 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00027649481415130677, |
|
"loss": 1.6614, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002763272856074886, |
|
"loss": 1.6085, |
|
"step": 1401 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00027615969435574743, |
|
"loss": 1.4628, |
|
"step": 1402 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002759920405337705, |
|
"loss": 1.6109, |
|
"step": 1403 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002758243242792965, |
|
"loss": 1.6735, |
|
"step": 1404 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00027565654573011544, |
|
"loss": 1.644, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002754887050240686, |
|
"loss": 1.6833, |
|
"step": 1406 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0002753208022990482, |
|
"loss": 1.6676, |
|
"step": 1407 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00027515283769299746, |
|
"loss": 1.5092, |
|
"step": 1408 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0002749848113439104, |
|
"loss": 1.5241, |
|
"step": 1409 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027481672338983186, |
|
"loss": 1.5969, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027464857396885726, |
|
"loss": 1.7649, |
|
"step": 1411 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027448036321913236, |
|
"loss": 1.5747, |
|
"step": 1412 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027431209127885353, |
|
"loss": 1.4058, |
|
"step": 1413 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027414375828626724, |
|
"loss": 1.7127, |
|
"step": 1414 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027397536437967035, |
|
"loss": 1.573, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0002738069096974095, |
|
"loss": 1.531, |
|
"step": 1416 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0002736383943778814, |
|
"loss": 1.6098, |
|
"step": 1417 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027346981855953253, |
|
"loss": 1.6464, |
|
"step": 1418 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027330118238085916, |
|
"loss": 1.5938, |
|
"step": 1419 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.000273132485980407, |
|
"loss": 1.4606, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027296372949677135, |
|
"loss": 1.6626, |
|
"step": 1421 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027279491306859684, |
|
"loss": 1.5812, |
|
"step": 1422 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0002726260368345774, |
|
"loss": 1.5507, |
|
"step": 1423 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027245710093345613, |
|
"loss": 1.6078, |
|
"step": 1424 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0002722881055040248, |
|
"loss": 1.5539, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0002721190506851246, |
|
"loss": 1.6483, |
|
"step": 1426 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027194993661564513, |
|
"loss": 1.678, |
|
"step": 1427 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027178076343452486, |
|
"loss": 1.6343, |
|
"step": 1428 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027161153128075073, |
|
"loss": 1.536, |
|
"step": 1429 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027144224029335823, |
|
"loss": 1.4717, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0002712728906114311, |
|
"loss": 1.5653, |
|
"step": 1431 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027110348237410135, |
|
"loss": 1.6074, |
|
"step": 1432 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.000270934015720549, |
|
"loss": 1.577, |
|
"step": 1433 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027076449079000217, |
|
"loss": 1.5222, |
|
"step": 1434 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027059490772173684, |
|
"loss": 1.5487, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027042526665507674, |
|
"loss": 1.6224, |
|
"step": 1436 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027025556772939323, |
|
"loss": 1.5369, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00027008581108410525, |
|
"loss": 1.7217, |
|
"step": 1438 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.000269915996858679, |
|
"loss": 1.6173, |
|
"step": 1439 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00026974612519262836, |
|
"loss": 1.5546, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0002695761962255141, |
|
"loss": 1.5914, |
|
"step": 1441 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0002694062100969439, |
|
"loss": 1.7928, |
|
"step": 1442 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0002692361669465728, |
|
"loss": 1.4734, |
|
"step": 1443 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00026906606691410256, |
|
"loss": 1.6999, |
|
"step": 1444 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026889591013928155, |
|
"loss": 1.6739, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002687256967619049, |
|
"loss": 1.4526, |
|
"step": 1446 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026855542692181416, |
|
"loss": 1.5046, |
|
"step": 1447 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002683851007588972, |
|
"loss": 1.5466, |
|
"step": 1448 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026821471841308837, |
|
"loss": 1.4806, |
|
"step": 1449 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026804428002436795, |
|
"loss": 1.4408, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026787378573276245, |
|
"loss": 1.675, |
|
"step": 1451 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002677032356783442, |
|
"loss": 1.6464, |
|
"step": 1452 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026753263000123137, |
|
"loss": 1.4288, |
|
"step": 1453 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002673619688415879, |
|
"loss": 1.5308, |
|
"step": 1454 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026719125233962306, |
|
"loss": 1.6454, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026702048063559196, |
|
"loss": 1.5972, |
|
"step": 1456 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026684965386979473, |
|
"loss": 1.6361, |
|
"step": 1457 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.000266678772182577, |
|
"loss": 1.7033, |
|
"step": 1458 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002665078357143294, |
|
"loss": 1.5109, |
|
"step": 1459 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026633684460548743, |
|
"loss": 1.6466, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002661657989965317, |
|
"loss": 1.4888, |
|
"step": 1461 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026599469902798754, |
|
"loss": 1.4921, |
|
"step": 1462 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002658235448404249, |
|
"loss": 1.7685, |
|
"step": 1463 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002656523365744582, |
|
"loss": 1.6959, |
|
"step": 1464 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002654810743707465, |
|
"loss": 1.6365, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002653097583699929, |
|
"loss": 1.6566, |
|
"step": 1466 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002651383887129451, |
|
"loss": 1.5456, |
|
"step": 1467 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002649669655403944, |
|
"loss": 1.5749, |
|
"step": 1468 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026479548899317637, |
|
"loss": 1.393, |
|
"step": 1469 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026462395921217027, |
|
"loss": 1.4978, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026445237633829937, |
|
"loss": 1.5108, |
|
"step": 1471 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00026428074051253024, |
|
"loss": 1.4914, |
|
"step": 1472 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002641090518758731, |
|
"loss": 1.5146, |
|
"step": 1473 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002639373105693815, |
|
"loss": 1.6586, |
|
"step": 1474 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002637655167341525, |
|
"loss": 1.6137, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002635936705113259, |
|
"loss": 1.5716, |
|
"step": 1476 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002634217720420848, |
|
"loss": 1.4816, |
|
"step": 1477 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002632498214676552, |
|
"loss": 1.4983, |
|
"step": 1478 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.000263077818929306, |
|
"loss": 1.57, |
|
"step": 1479 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002629057645683485, |
|
"loss": 1.3983, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00026273365852613696, |
|
"loss": 1.588, |
|
"step": 1481 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002625615009440677, |
|
"loss": 1.7071, |
|
"step": 1482 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002623892919635797, |
|
"loss": 1.5111, |
|
"step": 1483 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00026221703172615407, |
|
"loss": 1.6639, |
|
"step": 1484 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002620447203733139, |
|
"loss": 1.6503, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002618723580466245, |
|
"loss": 1.8411, |
|
"step": 1486 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00026169994488769283, |
|
"loss": 1.5682, |
|
"step": 1487 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00026152748103816785, |
|
"loss": 1.7213, |
|
"step": 1488 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00026135496663974, |
|
"loss": 1.5409, |
|
"step": 1489 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002611824018341412, |
|
"loss": 1.5094, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00026100978676314496, |
|
"loss": 1.7543, |
|
"step": 1491 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00026083712156856596, |
|
"loss": 1.6379, |
|
"step": 1492 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002606644063922602, |
|
"loss": 1.6939, |
|
"step": 1493 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002604916413761245, |
|
"loss": 1.6746, |
|
"step": 1494 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00026031882666209685, |
|
"loss": 1.5731, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002601459623921559, |
|
"loss": 1.5194, |
|
"step": 1496 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002599730487083213, |
|
"loss": 1.5836, |
|
"step": 1497 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00025980008575265294, |
|
"loss": 1.6906, |
|
"step": 1498 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002596270736672513, |
|
"loss": 1.5943, |
|
"step": 1499 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002594540125942574, |
|
"loss": 1.5434, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002592809026758524, |
|
"loss": 1.5649, |
|
"step": 1501 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00025910774405425745, |
|
"loss": 1.5539, |
|
"step": 1502 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00025893453687173386, |
|
"loss": 1.7896, |
|
"step": 1503 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002587612812705829, |
|
"loss": 1.6039, |
|
"step": 1504 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00025858797739314543, |
|
"loss": 1.6795, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002584146253818021, |
|
"loss": 1.6231, |
|
"step": 1506 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00025824122537897306, |
|
"loss": 1.566, |
|
"step": 1507 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.000258067777527118, |
|
"loss": 1.5596, |
|
"step": 1508 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002578942819687357, |
|
"loss": 1.5003, |
|
"step": 1509 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002577207388463643, |
|
"loss": 1.4594, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.000257547148302581, |
|
"loss": 1.6117, |
|
"step": 1511 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002573735104800019, |
|
"loss": 1.6485, |
|
"step": 1512 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002571998255212819, |
|
"loss": 1.5471, |
|
"step": 1513 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00025702609356911496, |
|
"loss": 1.7916, |
|
"step": 1514 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002568523147662331, |
|
"loss": 1.5788, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002566784892554073, |
|
"loss": 1.5992, |
|
"step": 1516 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025650461717944666, |
|
"loss": 1.6932, |
|
"step": 1517 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025633069868119865, |
|
"loss": 1.5534, |
|
"step": 1518 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002561567339035489, |
|
"loss": 1.6099, |
|
"step": 1519 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002559827229894208, |
|
"loss": 1.5029, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.000255808666081776, |
|
"loss": 1.4537, |
|
"step": 1521 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025563456332361386, |
|
"loss": 1.6383, |
|
"step": 1522 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025546041485797116, |
|
"loss": 1.5842, |
|
"step": 1523 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025528622082792254, |
|
"loss": 1.7171, |
|
"step": 1524 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002551119813765798, |
|
"loss": 1.6149, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002549376966470923, |
|
"loss": 1.4452, |
|
"step": 1526 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002547633667826465, |
|
"loss": 1.531, |
|
"step": 1527 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025458899192646586, |
|
"loss": 1.4827, |
|
"step": 1528 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025441457222181096, |
|
"loss": 1.5782, |
|
"step": 1529 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025424010781197905, |
|
"loss": 1.5321, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002540655988403043, |
|
"loss": 1.6408, |
|
"step": 1531 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002538910454501574, |
|
"loss": 1.507, |
|
"step": 1532 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025371644778494543, |
|
"loss": 1.5447, |
|
"step": 1533 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025354180598811195, |
|
"loss": 1.5883, |
|
"step": 1534 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025336712020313694, |
|
"loss": 1.6542, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002531923905735362, |
|
"loss": 1.4731, |
|
"step": 1536 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025301761724286166, |
|
"loss": 1.4463, |
|
"step": 1537 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025284280035470127, |
|
"loss": 1.5595, |
|
"step": 1538 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025266794005267864, |
|
"loss": 1.6009, |
|
"step": 1539 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025249303648045327, |
|
"loss": 1.6186, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002523180897817198, |
|
"loss": 1.4711, |
|
"step": 1541 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002521431001002086, |
|
"loss": 1.5344, |
|
"step": 1542 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002519680675796853, |
|
"loss": 1.7139, |
|
"step": 1543 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025179299236395086, |
|
"loss": 1.703, |
|
"step": 1544 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.000251617874596841, |
|
"loss": 1.624, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025144271442222656, |
|
"loss": 1.5383, |
|
"step": 1546 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002512675119840133, |
|
"loss": 1.3363, |
|
"step": 1547 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025109226742614163, |
|
"loss": 1.5917, |
|
"step": 1548 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025091698089258656, |
|
"loss": 1.5101, |
|
"step": 1549 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002507416525273576, |
|
"loss": 1.588, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00025056628247449857, |
|
"loss": 1.6409, |
|
"step": 1551 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00025039087087808765, |
|
"loss": 1.4405, |
|
"step": 1552 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00025021541788223714, |
|
"loss": 1.5421, |
|
"step": 1553 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002500399236310932, |
|
"loss": 1.5258, |
|
"step": 1554 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024986438826883595, |
|
"loss": 1.5296, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002496888119396794, |
|
"loss": 1.5543, |
|
"step": 1556 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024951319478787125, |
|
"loss": 1.5746, |
|
"step": 1557 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024933753695769237, |
|
"loss": 1.7266, |
|
"step": 1558 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024916183859345756, |
|
"loss": 1.7363, |
|
"step": 1559 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024898609983951445, |
|
"loss": 1.6904, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024881032084024426, |
|
"loss": 1.6837, |
|
"step": 1561 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.000248634501740061, |
|
"loss": 1.5852, |
|
"step": 1562 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024845864268341166, |
|
"loss": 1.5069, |
|
"step": 1563 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024828274381477616, |
|
"loss": 1.6916, |
|
"step": 1564 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024810680527866715, |
|
"loss": 1.5812, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002479308272196297, |
|
"loss": 1.5797, |
|
"step": 1566 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024775480978224147, |
|
"loss": 1.6729, |
|
"step": 1567 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024757875311111246, |
|
"loss": 1.5456, |
|
"step": 1568 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024740265735088485, |
|
"loss": 1.6212, |
|
"step": 1569 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024722652264623306, |
|
"loss": 1.5198, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024705034914186333, |
|
"loss": 1.4736, |
|
"step": 1571 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024687413698251394, |
|
"loss": 1.3715, |
|
"step": 1572 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002466978863129548, |
|
"loss": 1.6873, |
|
"step": 1573 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024652159727798763, |
|
"loss": 1.5421, |
|
"step": 1574 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002463452700224454, |
|
"loss": 1.5436, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002461689046911926, |
|
"loss": 1.6014, |
|
"step": 1576 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024599250142912527, |
|
"loss": 1.5805, |
|
"step": 1577 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002458160603811702, |
|
"loss": 1.4851, |
|
"step": 1578 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002456395816922855, |
|
"loss": 1.7865, |
|
"step": 1579 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024546306550746004, |
|
"loss": 1.6202, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002452865119717136, |
|
"loss": 1.6061, |
|
"step": 1581 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002451099212300966, |
|
"loss": 1.6226, |
|
"step": 1582 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024493329342768994, |
|
"loss": 1.5634, |
|
"step": 1583 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024475662870960516, |
|
"loss": 1.5307, |
|
"step": 1584 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00024457992722098405, |
|
"loss": 1.7258, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002444031891069986, |
|
"loss": 1.5549, |
|
"step": 1586 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024422641451285086, |
|
"loss": 1.5448, |
|
"step": 1587 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024404960358377284, |
|
"loss": 1.6417, |
|
"step": 1588 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024387275646502646, |
|
"loss": 1.5272, |
|
"step": 1589 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024369587330190325, |
|
"loss": 1.6629, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024351895423972467, |
|
"loss": 1.3826, |
|
"step": 1591 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0002433419994238413, |
|
"loss": 1.4844, |
|
"step": 1592 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024316500899963323, |
|
"loss": 1.5749, |
|
"step": 1593 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024298798311250987, |
|
"loss": 1.4315, |
|
"step": 1594 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024281092190790978, |
|
"loss": 1.5369, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0002426338255313004, |
|
"loss": 1.5281, |
|
"step": 1596 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0002424566941281781, |
|
"loss": 1.4906, |
|
"step": 1597 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024227952784406807, |
|
"loss": 1.5724, |
|
"step": 1598 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024210232682452424, |
|
"loss": 1.6224, |
|
"step": 1599 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.000241925091215129, |
|
"loss": 1.6724, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024174782116149303, |
|
"loss": 1.5638, |
|
"step": 1601 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0002415705168092555, |
|
"loss": 1.5348, |
|
"step": 1602 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024139317830408367, |
|
"loss": 1.5205, |
|
"step": 1603 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024121580579167292, |
|
"loss": 1.5838, |
|
"step": 1604 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0002410383994177464, |
|
"loss": 1.4553, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024086095932805536, |
|
"loss": 1.561, |
|
"step": 1606 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024068348566837845, |
|
"loss": 1.4977, |
|
"step": 1607 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024050597858452216, |
|
"loss": 1.5353, |
|
"step": 1608 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00024032843822232024, |
|
"loss": 1.371, |
|
"step": 1609 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0002401508647276339, |
|
"loss": 1.576, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0002399732582463515, |
|
"loss": 1.5715, |
|
"step": 1611 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023979561892438858, |
|
"loss": 1.6734, |
|
"step": 1612 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023961794690768756, |
|
"loss": 1.6044, |
|
"step": 1613 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023944024234221788, |
|
"loss": 1.4774, |
|
"step": 1614 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023926250537397548, |
|
"loss": 1.5756, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0002390847361489831, |
|
"loss": 1.6889, |
|
"step": 1616 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023890693481329006, |
|
"loss": 1.7233, |
|
"step": 1617 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0002387291015129718, |
|
"loss": 1.4928, |
|
"step": 1618 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023855123639413011, |
|
"loss": 1.6063, |
|
"step": 1619 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0002383733396028931, |
|
"loss": 1.5051, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023819541128541482, |
|
"loss": 1.6827, |
|
"step": 1621 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023801745158787504, |
|
"loss": 1.5971, |
|
"step": 1622 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023783946065647946, |
|
"loss": 1.4289, |
|
"step": 1623 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023766143863745946, |
|
"loss": 1.7191, |
|
"step": 1624 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023748338567707196, |
|
"loss": 1.4261, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0002373053019215992, |
|
"loss": 1.5713, |
|
"step": 1626 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023712718751734882, |
|
"loss": 1.5689, |
|
"step": 1627 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023694904261065364, |
|
"loss": 1.5209, |
|
"step": 1628 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023677086734787142, |
|
"loss": 1.4496, |
|
"step": 1629 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023659266187538505, |
|
"loss": 1.4943, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023641442633960203, |
|
"loss": 1.5836, |
|
"step": 1631 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023623616088695468, |
|
"loss": 1.686, |
|
"step": 1632 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0002360578656638999, |
|
"loss": 1.4795, |
|
"step": 1633 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023587954081691917, |
|
"loss": 1.6373, |
|
"step": 1634 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023570118649251796, |
|
"loss": 1.4847, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023552280283722622, |
|
"loss": 1.6384, |
|
"step": 1636 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.000235344389997598, |
|
"loss": 1.5915, |
|
"step": 1637 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023516594812021126, |
|
"loss": 1.5623, |
|
"step": 1638 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023498747735166783, |
|
"loss": 1.4261, |
|
"step": 1639 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023480897783859317, |
|
"loss": 1.7279, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023463044972763658, |
|
"loss": 1.4415, |
|
"step": 1641 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023445189316547072, |
|
"loss": 1.6595, |
|
"step": 1642 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023427330829879167, |
|
"loss": 1.4856, |
|
"step": 1643 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023409469527431855, |
|
"loss": 1.5011, |
|
"step": 1644 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.000233916054238794, |
|
"loss": 1.5571, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0002337373853389834, |
|
"loss": 1.4625, |
|
"step": 1646 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023355868872167502, |
|
"loss": 1.5243, |
|
"step": 1647 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023337996453368006, |
|
"loss": 1.5997, |
|
"step": 1648 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023320121292183226, |
|
"loss": 1.5089, |
|
"step": 1649 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023302243403298793, |
|
"loss": 1.5961, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023284362801402576, |
|
"loss": 1.5106, |
|
"step": 1651 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0002326647950118467, |
|
"loss": 1.5099, |
|
"step": 1652 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.000232485935173374, |
|
"loss": 1.4402, |
|
"step": 1653 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023230704864555274, |
|
"loss": 1.4325, |
|
"step": 1654 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023212813557535022, |
|
"loss": 1.4983, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0002319491961097553, |
|
"loss": 1.6445, |
|
"step": 1656 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023177023039577859, |
|
"loss": 1.3986, |
|
"step": 1657 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023159123858045225, |
|
"loss": 1.6527, |
|
"step": 1658 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00023141222081083003, |
|
"loss": 1.7511, |
|
"step": 1659 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00023123317723398689, |
|
"loss": 1.5593, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0002310541079970188, |
|
"loss": 1.5734, |
|
"step": 1661 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00023087501324704318, |
|
"loss": 1.5429, |
|
"step": 1662 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00023069589313119826, |
|
"loss": 1.6887, |
|
"step": 1663 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.000230516747796643, |
|
"loss": 1.6576, |
|
"step": 1664 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00023033757739055716, |
|
"loss": 1.79, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00023015838206014112, |
|
"loss": 1.5351, |
|
"step": 1666 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022997916195261576, |
|
"loss": 1.7071, |
|
"step": 1667 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0002297999172152222, |
|
"loss": 1.4834, |
|
"step": 1668 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022962064799522192, |
|
"loss": 1.5708, |
|
"step": 1669 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022944135443989642, |
|
"loss": 1.4445, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022926203669654732, |
|
"loss": 1.5581, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022908269491249593, |
|
"loss": 1.6432, |
|
"step": 1672 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0002289033292350835, |
|
"loss": 1.5703, |
|
"step": 1673 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022872393981167077, |
|
"loss": 1.5786, |
|
"step": 1674 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.000228544526789638, |
|
"loss": 1.5684, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022836509031638495, |
|
"loss": 1.5308, |
|
"step": 1676 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022818563053933057, |
|
"loss": 1.7682, |
|
"step": 1677 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022800614760591295, |
|
"loss": 1.4572, |
|
"step": 1678 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022782664166358916, |
|
"loss": 1.5774, |
|
"step": 1679 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022764711285983523, |
|
"loss": 1.5223, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022746756134214607, |
|
"loss": 1.677, |
|
"step": 1681 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022728798725803504, |
|
"loss": 1.5746, |
|
"step": 1682 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022710839075503421, |
|
"loss": 1.5622, |
|
"step": 1683 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022692877198069402, |
|
"loss": 1.5781, |
|
"step": 1684 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022674913108258315, |
|
"loss": 1.5514, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022656946820828852, |
|
"loss": 1.426, |
|
"step": 1686 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022638978350541508, |
|
"loss": 1.65, |
|
"step": 1687 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022621007712158577, |
|
"loss": 1.7149, |
|
"step": 1688 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0002260303492044412, |
|
"loss": 1.6595, |
|
"step": 1689 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022585059990163983, |
|
"loss": 1.6742, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0002256708293608576, |
|
"loss": 1.6686, |
|
"step": 1691 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022549103772978794, |
|
"loss": 1.4319, |
|
"step": 1692 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022531122515614155, |
|
"loss": 1.6116, |
|
"step": 1693 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022513139178764648, |
|
"loss": 1.3789, |
|
"step": 1694 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022495153777204766, |
|
"loss": 1.4669, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022477166325710703, |
|
"loss": 1.6649, |
|
"step": 1696 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022459176839060354, |
|
"loss": 1.6541, |
|
"step": 1697 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022441185332033276, |
|
"loss": 1.671, |
|
"step": 1698 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022423191819410674, |
|
"loss": 1.5792, |
|
"step": 1699 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022405196315975414, |
|
"loss": 1.3773, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022387198836512, |
|
"loss": 1.3891, |
|
"step": 1701 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022369199395806547, |
|
"loss": 1.4892, |
|
"step": 1702 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022351198008646792, |
|
"loss": 1.4984, |
|
"step": 1703 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022333194689822067, |
|
"loss": 1.6368, |
|
"step": 1704 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022315189454123294, |
|
"loss": 1.4427, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0002229718231634296, |
|
"loss": 1.5146, |
|
"step": 1706 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022279173291275134, |
|
"loss": 1.4817, |
|
"step": 1707 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0002226116239371541, |
|
"loss": 1.5236, |
|
"step": 1708 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022243149638460943, |
|
"loss": 1.3901, |
|
"step": 1709 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022225135040310394, |
|
"loss": 1.5693, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022207118614063967, |
|
"loss": 1.4913, |
|
"step": 1711 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0002218910037452334, |
|
"loss": 1.5038, |
|
"step": 1712 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022171080336491684, |
|
"loss": 1.5641, |
|
"step": 1713 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0002215305851477367, |
|
"loss": 1.5131, |
|
"step": 1714 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.000221350349241754, |
|
"loss": 1.4212, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022117009579504467, |
|
"loss": 1.5201, |
|
"step": 1716 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022098982495569872, |
|
"loss": 1.4989, |
|
"step": 1717 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022080953687182063, |
|
"loss": 1.4331, |
|
"step": 1718 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022062923169152904, |
|
"loss": 1.4991, |
|
"step": 1719 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022044890956295668, |
|
"loss": 1.5306, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022026857063424995, |
|
"loss": 1.3076, |
|
"step": 1721 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022008821505356928, |
|
"loss": 1.4638, |
|
"step": 1722 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00021990784296908883, |
|
"loss": 1.5505, |
|
"step": 1723 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0002197274545289961, |
|
"loss": 1.7054, |
|
"step": 1724 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00021954704988149225, |
|
"loss": 1.4317, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00021936662917479156, |
|
"loss": 1.6565, |
|
"step": 1726 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00021918619255712168, |
|
"loss": 1.4403, |
|
"step": 1727 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00021900574017672325, |
|
"loss": 1.4818, |
|
"step": 1728 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00021882527218184975, |
|
"loss": 1.493, |
|
"step": 1729 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021864478872076767, |
|
"loss": 1.4143, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021846428994175612, |
|
"loss": 1.5277, |
|
"step": 1731 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021828377599310685, |
|
"loss": 1.6244, |
|
"step": 1732 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.000218103247023124, |
|
"loss": 1.4557, |
|
"step": 1733 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021792270318012405, |
|
"loss": 1.5642, |
|
"step": 1734 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021774214461243579, |
|
"loss": 1.6224, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021756157146839996, |
|
"loss": 1.5405, |
|
"step": 1736 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021738098389636955, |
|
"loss": 1.5319, |
|
"step": 1737 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0002172003820447089, |
|
"loss": 1.38, |
|
"step": 1738 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0002170197660617947, |
|
"loss": 1.6056, |
|
"step": 1739 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021683913609601483, |
|
"loss": 1.5819, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021665849229576885, |
|
"loss": 1.5213, |
|
"step": 1741 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021647783480946745, |
|
"loss": 1.5423, |
|
"step": 1742 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021629716378553289, |
|
"loss": 1.4778, |
|
"step": 1743 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021611647937239825, |
|
"loss": 1.5397, |
|
"step": 1744 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021593578171850788, |
|
"loss": 1.4801, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0002157550709723168, |
|
"loss": 1.5704, |
|
"step": 1746 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021557434728229088, |
|
"loss": 1.4853, |
|
"step": 1747 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021539361079690657, |
|
"loss": 1.5793, |
|
"step": 1748 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021521286166465093, |
|
"loss": 1.7272, |
|
"step": 1749 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021503210003402141, |
|
"loss": 1.678, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021485132605352552, |
|
"loss": 1.5371, |
|
"step": 1751 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021467053987168105, |
|
"loss": 1.4413, |
|
"step": 1752 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.000214489741637016, |
|
"loss": 1.6737, |
|
"step": 1753 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021430893149806802, |
|
"loss": 1.5311, |
|
"step": 1754 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021412810960338456, |
|
"loss": 1.5138, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0002139472761015229, |
|
"loss": 1.5867, |
|
"step": 1756 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021376643114104962, |
|
"loss": 1.5229, |
|
"step": 1757 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021358557487054092, |
|
"loss": 1.5251, |
|
"step": 1758 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021340470743858223, |
|
"loss": 1.6136, |
|
"step": 1759 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021322382899376802, |
|
"loss": 1.3991, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021304293968470205, |
|
"loss": 1.4122, |
|
"step": 1761 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021286203965999677, |
|
"loss": 1.5702, |
|
"step": 1762 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021268112906827367, |
|
"loss": 1.613, |
|
"step": 1763 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021250020805816267, |
|
"loss": 1.5842, |
|
"step": 1764 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00021231927677830238, |
|
"loss": 1.48, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00021213833537733984, |
|
"loss": 1.6922, |
|
"step": 1766 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002119573840039305, |
|
"loss": 1.4609, |
|
"step": 1767 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00021177642280673773, |
|
"loss": 1.6187, |
|
"step": 1768 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00021159545193443329, |
|
"loss": 1.6393, |
|
"step": 1769 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002114144715356967, |
|
"loss": 1.5765, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00021123348175921538, |
|
"loss": 1.5726, |
|
"step": 1771 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00021105248275368434, |
|
"loss": 1.6218, |
|
"step": 1772 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002108714746678063, |
|
"loss": 1.6118, |
|
"step": 1773 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00021069045765029137, |
|
"loss": 1.53, |
|
"step": 1774 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00021050943184985714, |
|
"loss": 1.4516, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002103283974152282, |
|
"loss": 1.6388, |
|
"step": 1776 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002101473544951363, |
|
"loss": 1.6348, |
|
"step": 1777 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002099663032383203, |
|
"loss": 1.5979, |
|
"step": 1778 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020978524379352574, |
|
"loss": 1.5776, |
|
"step": 1779 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020960417630950504, |
|
"loss": 1.5778, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020942310093501694, |
|
"loss": 1.7426, |
|
"step": 1781 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020924201781882706, |
|
"loss": 1.5134, |
|
"step": 1782 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020906092710970707, |
|
"loss": 1.555, |
|
"step": 1783 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002088798289564351, |
|
"loss": 1.577, |
|
"step": 1784 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020869872350779512, |
|
"loss": 1.4946, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002085176109125774, |
|
"loss": 1.6802, |
|
"step": 1786 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020833649131957781, |
|
"loss": 1.6832, |
|
"step": 1787 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020815536487759823, |
|
"loss": 1.543, |
|
"step": 1788 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020797423173544598, |
|
"loss": 1.4706, |
|
"step": 1789 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002077930920419339, |
|
"loss": 1.5107, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002076119459458803, |
|
"loss": 1.4211, |
|
"step": 1791 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020743079359610865, |
|
"loss": 1.6153, |
|
"step": 1792 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020724963514144772, |
|
"loss": 1.5399, |
|
"step": 1793 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002070684707307311, |
|
"loss": 1.5471, |
|
"step": 1794 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002068873005127973, |
|
"loss": 1.6296, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020670612463648976, |
|
"loss": 1.6471, |
|
"step": 1796 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020652494325065649, |
|
"loss": 1.4332, |
|
"step": 1797 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020634375650414993, |
|
"loss": 1.5291, |
|
"step": 1798 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.000206162564545827, |
|
"loss": 1.5962, |
|
"step": 1799 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00020598136752454895, |
|
"loss": 1.5381, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020580016558918108, |
|
"loss": 1.5404, |
|
"step": 1801 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020561895888859285, |
|
"loss": 1.6397, |
|
"step": 1802 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020543774757165754, |
|
"loss": 1.556, |
|
"step": 1803 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020525653178725222, |
|
"loss": 1.439, |
|
"step": 1804 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020507531168425773, |
|
"loss": 1.6438, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020489408741155844, |
|
"loss": 1.5649, |
|
"step": 1806 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0002047128591180419, |
|
"loss": 1.5187, |
|
"step": 1807 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020453162695259938, |
|
"loss": 1.5315, |
|
"step": 1808 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020435039106412494, |
|
"loss": 1.5905, |
|
"step": 1809 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020416915160151607, |
|
"loss": 1.622, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020398790871367281, |
|
"loss": 1.7464, |
|
"step": 1811 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020380666254949826, |
|
"loss": 1.5899, |
|
"step": 1812 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0002036254132578982, |
|
"loss": 1.5067, |
|
"step": 1813 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020344416098778094, |
|
"loss": 1.5302, |
|
"step": 1814 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020326290588805717, |
|
"loss": 1.5828, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020308164810763993, |
|
"loss": 1.5656, |
|
"step": 1816 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0002029003877954446, |
|
"loss": 1.5036, |
|
"step": 1817 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020271912510038854, |
|
"loss": 1.5446, |
|
"step": 1818 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.000202537860171391, |
|
"loss": 1.5783, |
|
"step": 1819 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020235659315737315, |
|
"loss": 1.4605, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020217532420725785, |
|
"loss": 1.5527, |
|
"step": 1821 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020199405346996955, |
|
"loss": 1.469, |
|
"step": 1822 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020181278109443417, |
|
"loss": 1.5489, |
|
"step": 1823 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020163150722957905, |
|
"loss": 1.5781, |
|
"step": 1824 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0002014502320243326, |
|
"loss": 1.5845, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020126895562762442, |
|
"loss": 1.4641, |
|
"step": 1826 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020108767818838516, |
|
"loss": 1.4895, |
|
"step": 1827 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020090639985554613, |
|
"loss": 1.5249, |
|
"step": 1828 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020072512077803957, |
|
"loss": 1.476, |
|
"step": 1829 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020054384110479822, |
|
"loss": 1.592, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020036256098475544, |
|
"loss": 1.6337, |
|
"step": 1831 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00020018128056684473, |
|
"loss": 1.4726, |
|
"step": 1832 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0002, |
|
"loss": 1.5201, |
|
"step": 1833 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00019981871943315532, |
|
"loss": 1.6322, |
|
"step": 1834 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0001996374390152446, |
|
"loss": 1.7281, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00019945615889520183, |
|
"loss": 1.4793, |
|
"step": 1836 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001992748792219605, |
|
"loss": 1.5181, |
|
"step": 1837 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019909360014445392, |
|
"loss": 1.5796, |
|
"step": 1838 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001989123218116149, |
|
"loss": 1.4726, |
|
"step": 1839 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019873104437237563, |
|
"loss": 1.454, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019854976797566742, |
|
"loss": 1.6595, |
|
"step": 1841 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019836849277042097, |
|
"loss": 1.5598, |
|
"step": 1842 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001981872189055658, |
|
"loss": 1.676, |
|
"step": 1843 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019800594653003047, |
|
"loss": 1.5567, |
|
"step": 1844 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001978246757927422, |
|
"loss": 1.5302, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001976434068426269, |
|
"loss": 1.5436, |
|
"step": 1846 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019746213982860903, |
|
"loss": 1.6026, |
|
"step": 1847 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019728087489961148, |
|
"loss": 1.5056, |
|
"step": 1848 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019709961220455545, |
|
"loss": 1.5773, |
|
"step": 1849 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001969183518923601, |
|
"loss": 1.5341, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019673709411194293, |
|
"loss": 1.5715, |
|
"step": 1851 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019655583901221913, |
|
"loss": 1.4446, |
|
"step": 1852 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019637458674210182, |
|
"loss": 1.6097, |
|
"step": 1853 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001961933374505018, |
|
"loss": 1.361, |
|
"step": 1854 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001960120912863272, |
|
"loss": 1.7158, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019583084839848398, |
|
"loss": 1.4089, |
|
"step": 1856 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019564960893587505, |
|
"loss": 1.395, |
|
"step": 1857 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019546837304740066, |
|
"loss": 1.583, |
|
"step": 1858 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001952871408819581, |
|
"loss": 1.5671, |
|
"step": 1859 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001951059125884416, |
|
"loss": 1.5786, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001949246883157423, |
|
"loss": 1.6128, |
|
"step": 1861 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019474346821274783, |
|
"loss": 1.6966, |
|
"step": 1862 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001945622524283425, |
|
"loss": 1.5539, |
|
"step": 1863 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001943810411114072, |
|
"loss": 1.545, |
|
"step": 1864 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019419983441081897, |
|
"loss": 1.6414, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019401863247545115, |
|
"loss": 1.5608, |
|
"step": 1866 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019383743545417308, |
|
"loss": 1.4859, |
|
"step": 1867 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019365624349585017, |
|
"loss": 1.6858, |
|
"step": 1868 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019347505674934353, |
|
"loss": 1.6128, |
|
"step": 1869 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019329387536351026, |
|
"loss": 1.5808, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0001931126994872027, |
|
"loss": 1.3536, |
|
"step": 1871 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00019293152926926896, |
|
"loss": 1.3789, |
|
"step": 1872 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0001927503648585523, |
|
"loss": 1.541, |
|
"step": 1873 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00019256920640389137, |
|
"loss": 1.5438, |
|
"step": 1874 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00019238805405411978, |
|
"loss": 1.7398, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00019220690795806615, |
|
"loss": 1.4853, |
|
"step": 1876 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0001920257682645541, |
|
"loss": 1.5107, |
|
"step": 1877 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00019184463512240182, |
|
"loss": 1.5052, |
|
"step": 1878 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00019166350868042223, |
|
"loss": 1.5421, |
|
"step": 1879 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00019148238908742272, |
|
"loss": 1.4841, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00019130127649220495, |
|
"loss": 1.5578, |
|
"step": 1881 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.000191120171043565, |
|
"loss": 1.5659, |
|
"step": 1882 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00019093907289029295, |
|
"loss": 1.6815, |
|
"step": 1883 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00019075798218117296, |
|
"loss": 1.5356, |
|
"step": 1884 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00019057689906498305, |
|
"loss": 1.5597, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.000190395823690495, |
|
"loss": 1.6631, |
|
"step": 1886 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00019021475620647428, |
|
"loss": 1.6313, |
|
"step": 1887 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00019003369676167975, |
|
"loss": 1.4553, |
|
"step": 1888 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0001898526455048637, |
|
"loss": 1.6443, |
|
"step": 1889 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00018967160258477186, |
|
"loss": 1.5262, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0001894905681501429, |
|
"loss": 1.4328, |
|
"step": 1891 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00018930954234970868, |
|
"loss": 1.4888, |
|
"step": 1892 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0001891285253321938, |
|
"loss": 1.4733, |
|
"step": 1893 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00018894751724631576, |
|
"loss": 1.5523, |
|
"step": 1894 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00018876651824078472, |
|
"loss": 1.4814, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00018858552846430336, |
|
"loss": 1.5552, |
|
"step": 1896 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0001884045480655667, |
|
"loss": 1.5165, |
|
"step": 1897 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00018822357719326224, |
|
"loss": 1.6468, |
|
"step": 1898 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00018804261599606958, |
|
"loss": 1.5167, |
|
"step": 1899 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00018786166462266018, |
|
"loss": 1.5326, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00018768072322169766, |
|
"loss": 1.6156, |
|
"step": 1901 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00018749979194183737, |
|
"loss": 1.4218, |
|
"step": 1902 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00018731887093172635, |
|
"loss": 1.611, |
|
"step": 1903 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00018713796034000325, |
|
"loss": 1.4635, |
|
"step": 1904 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.000186957060315298, |
|
"loss": 1.5552, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.000186776171006232, |
|
"loss": 1.5173, |
|
"step": 1906 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00018659529256141784, |
|
"loss": 1.4805, |
|
"step": 1907 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018641442512945912, |
|
"loss": 1.4122, |
|
"step": 1908 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018623356885895048, |
|
"loss": 1.4841, |
|
"step": 1909 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001860527238984772, |
|
"loss": 1.6319, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001858718903966154, |
|
"loss": 1.5003, |
|
"step": 1911 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018569106850193203, |
|
"loss": 1.5432, |
|
"step": 1912 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.000185510258362984, |
|
"loss": 1.3749, |
|
"step": 1913 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018532946012831894, |
|
"loss": 1.6482, |
|
"step": 1914 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018514867394647455, |
|
"loss": 1.5161, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018496789996597866, |
|
"loss": 1.5298, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001847871383353491, |
|
"loss": 1.5455, |
|
"step": 1917 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018460638920309345, |
|
"loss": 1.5383, |
|
"step": 1918 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001844256527177092, |
|
"loss": 1.373, |
|
"step": 1919 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018424492902768325, |
|
"loss": 1.4938, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018406421828149217, |
|
"loss": 1.4049, |
|
"step": 1921 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018388352062760182, |
|
"loss": 1.6455, |
|
"step": 1922 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018370283621446722, |
|
"loss": 1.5405, |
|
"step": 1923 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018352216519053262, |
|
"loss": 1.5581, |
|
"step": 1924 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001833415077042312, |
|
"loss": 1.558, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018316086390398513, |
|
"loss": 1.43, |
|
"step": 1926 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018298023393820527, |
|
"loss": 1.5956, |
|
"step": 1927 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001827996179552911, |
|
"loss": 1.6198, |
|
"step": 1928 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018261901610363052, |
|
"loss": 1.3946, |
|
"step": 1929 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018243842853160006, |
|
"loss": 1.5476, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001822578553875643, |
|
"loss": 1.4229, |
|
"step": 1931 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.000182077296819876, |
|
"loss": 1.5386, |
|
"step": 1932 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018189675297687605, |
|
"loss": 1.2868, |
|
"step": 1933 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001817162240068932, |
|
"loss": 1.472, |
|
"step": 1934 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018153571005824395, |
|
"loss": 1.5035, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001813552112792324, |
|
"loss": 1.4659, |
|
"step": 1936 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018117472781815035, |
|
"loss": 1.6429, |
|
"step": 1937 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018099425982327685, |
|
"loss": 1.5322, |
|
"step": 1938 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018081380744287834, |
|
"loss": 1.5729, |
|
"step": 1939 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001806333708252084, |
|
"loss": 1.461, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018045295011850777, |
|
"loss": 1.5112, |
|
"step": 1941 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0001802725454710039, |
|
"loss": 1.4506, |
|
"step": 1942 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00018009215703091122, |
|
"loss": 1.6647, |
|
"step": 1943 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017991178494643076, |
|
"loss": 1.6595, |
|
"step": 1944 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017973142936575013, |
|
"loss": 1.4811, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.0001795510904370434, |
|
"loss": 1.628, |
|
"step": 1946 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017937076830847097, |
|
"loss": 1.5668, |
|
"step": 1947 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017919046312817939, |
|
"loss": 1.6493, |
|
"step": 1948 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017901017504430133, |
|
"loss": 1.5995, |
|
"step": 1949 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.0001788299042049554, |
|
"loss": 1.6209, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017864965075824605, |
|
"loss": 1.6038, |
|
"step": 1951 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017846941485226344, |
|
"loss": 1.6302, |
|
"step": 1952 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017828919663508315, |
|
"loss": 1.6829, |
|
"step": 1953 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017810899625476663, |
|
"loss": 1.5378, |
|
"step": 1954 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017792881385936035, |
|
"loss": 1.4865, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017774864959689606, |
|
"loss": 1.6057, |
|
"step": 1956 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017756850361539062, |
|
"loss": 1.6124, |
|
"step": 1957 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017738837606284593, |
|
"loss": 1.6264, |
|
"step": 1958 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017720826708724868, |
|
"loss": 1.53, |
|
"step": 1959 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017702817683657042, |
|
"loss": 1.5987, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.0001768481054587671, |
|
"loss": 1.6489, |
|
"step": 1961 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017666805310177937, |
|
"loss": 1.5588, |
|
"step": 1962 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017648801991353212, |
|
"loss": 1.4548, |
|
"step": 1963 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017630800604193455, |
|
"loss": 1.4823, |
|
"step": 1964 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017612801163488008, |
|
"loss": 1.6019, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017594803684024593, |
|
"loss": 1.6656, |
|
"step": 1966 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017576808180589325, |
|
"loss": 1.5728, |
|
"step": 1967 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.0001755881466796673, |
|
"loss": 1.5393, |
|
"step": 1968 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017540823160939645, |
|
"loss": 1.5718, |
|
"step": 1969 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.000175228336742893, |
|
"loss": 1.6284, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.0001750484622279524, |
|
"loss": 1.6717, |
|
"step": 1971 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017486860821235353, |
|
"loss": 1.3874, |
|
"step": 1972 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017468877484385847, |
|
"loss": 1.4552, |
|
"step": 1973 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.0001745089622702121, |
|
"loss": 1.5111, |
|
"step": 1974 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017432917063914244, |
|
"loss": 1.5682, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017414940009836022, |
|
"loss": 1.5648, |
|
"step": 1976 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017396965079555886, |
|
"loss": 1.4998, |
|
"step": 1977 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017378992287841433, |
|
"loss": 1.6988, |
|
"step": 1978 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00017361021649458497, |
|
"loss": 1.5206, |
|
"step": 1979 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00017343053179171155, |
|
"loss": 1.5095, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00017325086891741687, |
|
"loss": 1.4581, |
|
"step": 1981 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.000173071228019306, |
|
"loss": 1.5087, |
|
"step": 1982 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00017289160924496578, |
|
"loss": 1.5442, |
|
"step": 1983 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00017271201274196495, |
|
"loss": 1.5898, |
|
"step": 1984 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00017253243865785397, |
|
"loss": 1.4531, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.0001723528871401648, |
|
"loss": 1.5352, |
|
"step": 1986 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.0001721733583364109, |
|
"loss": 1.4362, |
|
"step": 1987 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00017199385239408712, |
|
"loss": 1.422, |
|
"step": 1988 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00017181436946066947, |
|
"loss": 1.5783, |
|
"step": 1989 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.0001716349096836151, |
|
"loss": 1.5079, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00017145547321036205, |
|
"loss": 1.6901, |
|
"step": 1991 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.0001712760601883293, |
|
"loss": 1.764, |
|
"step": 1992 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00017109667076491657, |
|
"loss": 1.6544, |
|
"step": 1993 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00017091730508750412, |
|
"loss": 1.5159, |
|
"step": 1994 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.0001707379633034528, |
|
"loss": 1.573, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00017055864556010357, |
|
"loss": 1.6284, |
|
"step": 1996 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.0001703793520047781, |
|
"loss": 1.5416, |
|
"step": 1997 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00017020008278477786, |
|
"loss": 1.4663, |
|
"step": 1998 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.0001700208380473843, |
|
"loss": 1.6156, |
|
"step": 1999 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.0001698416179398589, |
|
"loss": 1.4967, |
|
"step": 2000 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 3566, |
|
"num_train_epochs": 1, |
|
"save_steps": 1000, |
|
"total_flos": 5.148360879867167e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|