|
{ |
|
"best_metric": 2710.1728, |
|
"best_model_checkpoint": "/jasmine/s-elmadany/ds_UBC_AfroT5_large_1M_2e-5/checkpoint-122140", |
|
"epoch": 10.0, |
|
"eval_steps": 500, |
|
"global_step": 122140, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9983625347961356e-05, |
|
"loss": 8.4624, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9967250695922714e-05, |
|
"loss": 5.9771, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.995087604388407e-05, |
|
"loss": 5.7693, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9934501391845424e-05, |
|
"loss": 5.0828, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.991812673980678e-05, |
|
"loss": 4.8058, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9901752087768136e-05, |
|
"loss": 4.5844, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.988537743572949e-05, |
|
"loss": 4.4461, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.986900278369085e-05, |
|
"loss": 4.3994, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9852628131652203e-05, |
|
"loss": 4.3327, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.983625347961356e-05, |
|
"loss": 4.2567, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9819878827574916e-05, |
|
"loss": 4.2223, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.980350417553627e-05, |
|
"loss": 4.1905, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.978712952349763e-05, |
|
"loss": 4.1563, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9770754871458983e-05, |
|
"loss": 4.1126, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9754380219420338e-05, |
|
"loss": 4.0979, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9738005567381696e-05, |
|
"loss": 4.0696, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.972163091534305e-05, |
|
"loss": 4.0446, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9705256263304405e-05, |
|
"loss": 4.0218, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9688881611265763e-05, |
|
"loss": 3.9976, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9672506959227118e-05, |
|
"loss": 3.9459, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.9656132307188473e-05, |
|
"loss": 3.9501, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.963975765514983e-05, |
|
"loss": 3.9454, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.9623383003111185e-05, |
|
"loss": 3.9145, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.960700835107254e-05, |
|
"loss": 3.9358, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.9590633699033898e-05, |
|
"loss": 3.8542, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.9574259046995253e-05, |
|
"loss": 3.8823, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.9557884394956607e-05, |
|
"loss": 3.8611, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.9541509742917965e-05, |
|
"loss": 3.8609, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.952513509087932e-05, |
|
"loss": 3.8498, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.9508760438840678e-05, |
|
"loss": 3.8349, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.9492385786802032e-05, |
|
"loss": 3.8498, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.9476011134763387e-05, |
|
"loss": 3.7941, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.9459636482724745e-05, |
|
"loss": 3.8308, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.94432618306861e-05, |
|
"loss": 3.7989, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.9426887178647454e-05, |
|
"loss": 3.8041, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.9410512526608812e-05, |
|
"loss": 3.7968, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.9394137874570167e-05, |
|
"loss": 3.7609, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.937776322253152e-05, |
|
"loss": 3.7525, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.936138857049288e-05, |
|
"loss": 3.7561, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.9345013918454234e-05, |
|
"loss": 3.7288, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.932863926641559e-05, |
|
"loss": 3.748, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.9312264614376947e-05, |
|
"loss": 3.7518, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.92958899623383e-05, |
|
"loss": 3.718, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.9279515310299656e-05, |
|
"loss": 3.7124, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.9263140658261014e-05, |
|
"loss": 3.7118, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.924676600622237e-05, |
|
"loss": 3.6885, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.9230391354183727e-05, |
|
"loss": 3.7057, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.921401670214508e-05, |
|
"loss": 3.6812, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.9197642050106436e-05, |
|
"loss": 3.6831, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.9181267398067794e-05, |
|
"loss": 3.6705, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.916489274602915e-05, |
|
"loss": 3.687, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.9148518093990503e-05, |
|
"loss": 3.677, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.913214344195186e-05, |
|
"loss": 3.6632, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.9115768789913216e-05, |
|
"loss": 3.6626, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.909939413787457e-05, |
|
"loss": 3.6451, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.908301948583593e-05, |
|
"loss": 3.6283, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.9066644833797283e-05, |
|
"loss": 3.6321, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.9050270181758638e-05, |
|
"loss": 3.6007, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.9033895529719996e-05, |
|
"loss": 3.6293, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.901752087768135e-05, |
|
"loss": 3.5916, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.9001146225642705e-05, |
|
"loss": 3.6187, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.8984771573604063e-05, |
|
"loss": 3.6006, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.8968396921565418e-05, |
|
"loss": 3.5998, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.8952022269526773e-05, |
|
"loss": 3.5517, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.893564761748813e-05, |
|
"loss": 3.5713, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.8919272965449485e-05, |
|
"loss": 3.5718, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.8902898313410843e-05, |
|
"loss": 3.5697, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.8886523661372198e-05, |
|
"loss": 3.5477, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.8870149009333552e-05, |
|
"loss": 3.5405, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.885377435729491e-05, |
|
"loss": 3.5724, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.8837399705256265e-05, |
|
"loss": 3.544, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.882102505321762e-05, |
|
"loss": 3.5689, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.8804650401178978e-05, |
|
"loss": 3.5258, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.8788275749140332e-05, |
|
"loss": 3.5395, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.8771901097101687e-05, |
|
"loss": 3.5342, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.8755526445063045e-05, |
|
"loss": 3.5367, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.87391517930244e-05, |
|
"loss": 3.5328, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.8722777140985754e-05, |
|
"loss": 3.5402, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.8706402488947112e-05, |
|
"loss": 3.5269, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.8690027836908467e-05, |
|
"loss": 3.5107, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.867365318486982e-05, |
|
"loss": 3.5119, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.8657278532831176e-05, |
|
"loss": 3.5255, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.8640903880792534e-05, |
|
"loss": 3.506, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.8624529228753892e-05, |
|
"loss": 3.4903, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.8608154576715247e-05, |
|
"loss": 3.531, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.85917799246766e-05, |
|
"loss": 3.5041, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.857540527263796e-05, |
|
"loss": 3.4605, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8559030620599314e-05, |
|
"loss": 3.4925, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.854265596856067e-05, |
|
"loss": 3.4748, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.8526281316522027e-05, |
|
"loss": 3.4565, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.850990666448338e-05, |
|
"loss": 3.4581, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.8493532012444736e-05, |
|
"loss": 3.4933, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.8477157360406094e-05, |
|
"loss": 3.4837, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.846078270836745e-05, |
|
"loss": 3.4716, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.8444408056328803e-05, |
|
"loss": 3.4574, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.842803340429016e-05, |
|
"loss": 3.4469, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.8411658752251516e-05, |
|
"loss": 3.4244, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.839528410021287e-05, |
|
"loss": 3.3977, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.837890944817423e-05, |
|
"loss": 3.4462, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.8362534796135583e-05, |
|
"loss": 3.4458, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.8346160144096938e-05, |
|
"loss": 3.4315, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.8329785492058296e-05, |
|
"loss": 3.4124, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.831341084001965e-05, |
|
"loss": 3.409, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.829703618798101e-05, |
|
"loss": 3.4229, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.8280661535942363e-05, |
|
"loss": 3.4055, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.8264286883903718e-05, |
|
"loss": 3.3975, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.8247912231865076e-05, |
|
"loss": 3.4197, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.823153757982643e-05, |
|
"loss": 3.395, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.8215162927787785e-05, |
|
"loss": 3.4371, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.8198788275749143e-05, |
|
"loss": 3.3775, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.8182413623710498e-05, |
|
"loss": 3.3925, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.8166038971671852e-05, |
|
"loss": 3.4044, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.814966431963321e-05, |
|
"loss": 3.4044, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.8133289667594565e-05, |
|
"loss": 3.3841, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.811691501555592e-05, |
|
"loss": 3.4051, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.8100540363517278e-05, |
|
"loss": 3.3812, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.8084165711478632e-05, |
|
"loss": 3.3966, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.8067791059439987e-05, |
|
"loss": 3.4188, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.805141640740134e-05, |
|
"loss": 3.3683, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.80350417553627e-05, |
|
"loss": 3.3872, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.8018667103324058e-05, |
|
"loss": 3.3407, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.8002292451285412e-05, |
|
"loss": 3.3987, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_bleu": 732.9294, |
|
"eval_chrf": 2404.1983, |
|
"eval_chrf_plus": 2251.1143, |
|
"eval_gen_len": 14.592, |
|
"eval_loss": 2.8766348361968994, |
|
"eval_runtime": 38.5936, |
|
"eval_samples_per_second": 25.911, |
|
"eval_steps_per_second": 1.088, |
|
"step": 12214 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.7985917799246767e-05, |
|
"loss": 3.3044, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.7969543147208125e-05, |
|
"loss": 3.2769, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.795316849516948e-05, |
|
"loss": 3.3158, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.7936793843130834e-05, |
|
"loss": 3.3013, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.7920419191092192e-05, |
|
"loss": 3.3322, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.7904044539053547e-05, |
|
"loss": 3.2939, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.78876698870149e-05, |
|
"loss": 3.2753, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.787129523497626e-05, |
|
"loss": 3.3128, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.7854920582937614e-05, |
|
"loss": 3.2881, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.783854593089897e-05, |
|
"loss": 3.2776, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7822171278860327e-05, |
|
"loss": 3.2845, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.780579662682168e-05, |
|
"loss": 3.2722, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.7789421974783036e-05, |
|
"loss": 3.2907, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.777304732274439e-05, |
|
"loss": 3.29, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.775667267070575e-05, |
|
"loss": 3.2576, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.7740298018667107e-05, |
|
"loss": 3.2402, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.772392336662846e-05, |
|
"loss": 3.2807, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.7707548714589816e-05, |
|
"loss": 3.268, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.7691174062551174e-05, |
|
"loss": 3.2347, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.767479941051253e-05, |
|
"loss": 3.2847, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.7658424758473883e-05, |
|
"loss": 3.2423, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.764205010643524e-05, |
|
"loss": 3.2641, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.7625675454396596e-05, |
|
"loss": 3.245, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.760930080235795e-05, |
|
"loss": 3.2686, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.759292615031931e-05, |
|
"loss": 3.1976, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.7576551498280663e-05, |
|
"loss": 3.2724, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.7560176846242018e-05, |
|
"loss": 3.2294, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.7543802194203376e-05, |
|
"loss": 3.2484, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.752742754216473e-05, |
|
"loss": 3.2628, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.7511052890126085e-05, |
|
"loss": 3.2275, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.7494678238087443e-05, |
|
"loss": 3.2636, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.7478303586048798e-05, |
|
"loss": 3.2635, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.7461928934010152e-05, |
|
"loss": 3.2374, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.7445554281971507e-05, |
|
"loss": 3.224, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.7429179629932865e-05, |
|
"loss": 3.2542, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.7412804977894223e-05, |
|
"loss": 3.2073, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.7396430325855578e-05, |
|
"loss": 3.2291, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.7380055673816932e-05, |
|
"loss": 3.2193, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.736368102177829e-05, |
|
"loss": 3.2367, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.7347306369739645e-05, |
|
"loss": 3.1989, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.7330931717701e-05, |
|
"loss": 3.2228, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.7314557065662358e-05, |
|
"loss": 3.1936, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.7298182413623712e-05, |
|
"loss": 3.2073, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.7281807761585067e-05, |
|
"loss": 3.22, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.7265433109546425e-05, |
|
"loss": 3.2036, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.724905845750778e-05, |
|
"loss": 3.2283, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.7232683805469134e-05, |
|
"loss": 3.253, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.7216309153430492e-05, |
|
"loss": 3.1681, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.7199934501391847e-05, |
|
"loss": 3.2035, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.71835598493532e-05, |
|
"loss": 3.2251, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.7167185197314556e-05, |
|
"loss": 3.1773, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.7150810545275914e-05, |
|
"loss": 3.1996, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.7134435893237272e-05, |
|
"loss": 3.178, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.7118061241198627e-05, |
|
"loss": 3.2248, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.710168658915998e-05, |
|
"loss": 3.2019, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.708531193712134e-05, |
|
"loss": 3.2058, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.7068937285082694e-05, |
|
"loss": 3.1742, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.705256263304405e-05, |
|
"loss": 3.1653, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.7036187981005407e-05, |
|
"loss": 3.2102, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.701981332896676e-05, |
|
"loss": 3.2096, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.7003438676928116e-05, |
|
"loss": 3.2129, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.6987064024889474e-05, |
|
"loss": 3.1804, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.697068937285083e-05, |
|
"loss": 3.1715, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.6954314720812183e-05, |
|
"loss": 3.1591, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.693794006877354e-05, |
|
"loss": 3.1861, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.6921565416734896e-05, |
|
"loss": 3.1293, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 1.690519076469625e-05, |
|
"loss": 3.1695, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 1.6888816112657605e-05, |
|
"loss": 3.2078, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 1.6872441460618963e-05, |
|
"loss": 3.1875, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 1.6856066808580318e-05, |
|
"loss": 3.141, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 1.6839692156541672e-05, |
|
"loss": 3.1861, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 1.682331750450303e-05, |
|
"loss": 3.1515, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 1.680694285246439e-05, |
|
"loss": 3.1628, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 1.6790568200425743e-05, |
|
"loss": 3.1899, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 1.6774193548387098e-05, |
|
"loss": 3.1687, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 1.6757818896348456e-05, |
|
"loss": 3.1699, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 1.674144424430981e-05, |
|
"loss": 3.1611, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 1.6725069592271165e-05, |
|
"loss": 3.1627, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 1.6708694940232523e-05, |
|
"loss": 3.1446, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 1.6692320288193878e-05, |
|
"loss": 3.1411, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 1.6675945636155232e-05, |
|
"loss": 3.1532, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 1.665957098411659e-05, |
|
"loss": 3.1618, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 1.6643196332077945e-05, |
|
"loss": 3.1464, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 1.66268216800393e-05, |
|
"loss": 3.1684, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 1.6610447028000657e-05, |
|
"loss": 3.1111, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 1.6594072375962012e-05, |
|
"loss": 3.1576, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 1.6577697723923367e-05, |
|
"loss": 3.1661, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 1.656132307188472e-05, |
|
"loss": 3.1532, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 1.654494841984608e-05, |
|
"loss": 3.1055, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 1.6528573767807437e-05, |
|
"loss": 3.1367, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 1.6512199115768792e-05, |
|
"loss": 3.1055, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 1.6495824463730147e-05, |
|
"loss": 3.1471, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 1.6479449811691505e-05, |
|
"loss": 3.1165, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 1.646307515965286e-05, |
|
"loss": 3.1307, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 1.6446700507614214e-05, |
|
"loss": 3.163, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 1.6430325855575572e-05, |
|
"loss": 3.1192, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 1.6413951203536927e-05, |
|
"loss": 3.0997, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.639757655149828e-05, |
|
"loss": 3.125, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.638120189945964e-05, |
|
"loss": 3.1122, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.6364827247420994e-05, |
|
"loss": 3.0875, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.634845259538235e-05, |
|
"loss": 3.0969, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.6332077943343707e-05, |
|
"loss": 3.1037, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.631570329130506e-05, |
|
"loss": 3.1123, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.6299328639266416e-05, |
|
"loss": 3.1142, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.628295398722777e-05, |
|
"loss": 3.1189, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.626657933518913e-05, |
|
"loss": 3.1587, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.6250204683150483e-05, |
|
"loss": 3.1438, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.623383003111184e-05, |
|
"loss": 3.109, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.6217455379073196e-05, |
|
"loss": 3.1217, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.6201080727034554e-05, |
|
"loss": 3.0996, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.618470607499591e-05, |
|
"loss": 3.0898, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.6168331422957263e-05, |
|
"loss": 3.0998, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.615195677091862e-05, |
|
"loss": 3.0892, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.6135582118879976e-05, |
|
"loss": 3.0579, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.611920746684133e-05, |
|
"loss": 3.0764, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 1.6102832814802688e-05, |
|
"loss": 3.0782, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.6086458162764043e-05, |
|
"loss": 3.1002, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.6070083510725398e-05, |
|
"loss": 3.0959, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 1.6053708858686756e-05, |
|
"loss": 3.1025, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.603733420664811e-05, |
|
"loss": 3.0993, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.6020959554609465e-05, |
|
"loss": 3.096, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.600458490257082e-05, |
|
"loss": 3.087, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_bleu": 811.1414, |
|
"eval_chrf": 2564.1192, |
|
"eval_chrf_plus": 2406.0925, |
|
"eval_gen_len": 14.665, |
|
"eval_loss": 2.703340530395508, |
|
"eval_runtime": 38.3779, |
|
"eval_samples_per_second": 26.057, |
|
"eval_steps_per_second": 1.094, |
|
"step": 24428 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 1.5988210250532177e-05, |
|
"loss": 3.0467, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 1.5971835598493532e-05, |
|
"loss": 3.0037, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 1.5955460946454887e-05, |
|
"loss": 3.0283, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 1.5939086294416245e-05, |
|
"loss": 3.0406, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 1.5922711642377603e-05, |
|
"loss": 3.0001, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 1.5906336990338957e-05, |
|
"loss": 3.026, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.5889962338300312e-05, |
|
"loss": 3.0017, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.587358768626167e-05, |
|
"loss": 3.0257, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 1.5857213034223025e-05, |
|
"loss": 2.999, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 1.584083838218438e-05, |
|
"loss": 3.024, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 1.5824463730145737e-05, |
|
"loss": 3.0346, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 1.5808089078107092e-05, |
|
"loss": 2.9922, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 1.5791714426068447e-05, |
|
"loss": 3.0021, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 1.5775339774029805e-05, |
|
"loss": 3.0432, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 1.575896512199116e-05, |
|
"loss": 2.9909, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.5742590469952514e-05, |
|
"loss": 3.039, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.5726215817913872e-05, |
|
"loss": 2.9914, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.5709841165875227e-05, |
|
"loss": 3.0141, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.569346651383658e-05, |
|
"loss": 3.0008, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.5677091861797936e-05, |
|
"loss": 3.0262, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 1.5660717209759294e-05, |
|
"loss": 3.0113, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.564434255772065e-05, |
|
"loss": 3.038, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.5627967905682006e-05, |
|
"loss": 3.0078, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.561159325364336e-05, |
|
"loss": 3.0023, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.559521860160472e-05, |
|
"loss": 3.0095, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 1.5578843949566074e-05, |
|
"loss": 2.9869, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.556246929752743e-05, |
|
"loss": 3.0389, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.5546094645488786e-05, |
|
"loss": 3.0175, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.552971999345014e-05, |
|
"loss": 3.0092, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.5513345341411496e-05, |
|
"loss": 2.9981, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.5496970689372854e-05, |
|
"loss": 2.999, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.5480596037334208e-05, |
|
"loss": 2.9858, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 1.5464221385295563e-05, |
|
"loss": 2.9776, |
|
"step": 27700 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.544784673325692e-05, |
|
"loss": 2.9879, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.5431472081218276e-05, |
|
"loss": 3.0294, |
|
"step": 27900 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.541509742917963e-05, |
|
"loss": 3.048, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.5398722777140985e-05, |
|
"loss": 3.0125, |
|
"step": 28100 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.5382348125102343e-05, |
|
"loss": 3.0201, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.5365973473063697e-05, |
|
"loss": 2.9874, |
|
"step": 28300 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.5349598821025052e-05, |
|
"loss": 2.9942, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.533322416898641e-05, |
|
"loss": 3.0, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.5316849516947768e-05, |
|
"loss": 2.9812, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.5300474864909123e-05, |
|
"loss": 2.9315, |
|
"step": 28700 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.5284100212870477e-05, |
|
"loss": 3.0019, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.5267725560831835e-05, |
|
"loss": 2.9821, |
|
"step": 28900 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.525135090879319e-05, |
|
"loss": 3.0036, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.5234976256754546e-05, |
|
"loss": 2.9855, |
|
"step": 29100 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.5218601604715901e-05, |
|
"loss": 2.9627, |
|
"step": 29200 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 1.5202226952677257e-05, |
|
"loss": 2.9755, |
|
"step": 29300 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 1.5185852300638612e-05, |
|
"loss": 2.9636, |
|
"step": 29400 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 1.5169477648599968e-05, |
|
"loss": 2.9918, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 1.5153102996561325e-05, |
|
"loss": 2.9975, |
|
"step": 29600 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 1.513672834452268e-05, |
|
"loss": 2.9579, |
|
"step": 29700 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 1.5120353692484036e-05, |
|
"loss": 2.9867, |
|
"step": 29800 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 1.5103979040445392e-05, |
|
"loss": 2.9265, |
|
"step": 29900 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 1.5087604388406747e-05, |
|
"loss": 2.9504, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 1.5071229736368103e-05, |
|
"loss": 3.0254, |
|
"step": 30100 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 1.5054855084329459e-05, |
|
"loss": 2.9785, |
|
"step": 30200 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 1.5038480432290814e-05, |
|
"loss": 2.9722, |
|
"step": 30300 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 1.5022105780252172e-05, |
|
"loss": 2.9798, |
|
"step": 30400 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 1.5005731128213528e-05, |
|
"loss": 2.9653, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 1.4989356476174883e-05, |
|
"loss": 2.9449, |
|
"step": 30600 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 1.4972981824136239e-05, |
|
"loss": 2.9874, |
|
"step": 30700 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 1.4956607172097595e-05, |
|
"loss": 2.9705, |
|
"step": 30800 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 1.494023252005895e-05, |
|
"loss": 2.9521, |
|
"step": 30900 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 1.4923857868020306e-05, |
|
"loss": 2.953, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 1.4907483215981661e-05, |
|
"loss": 2.9573, |
|
"step": 31100 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 1.4891108563943017e-05, |
|
"loss": 2.9484, |
|
"step": 31200 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 1.4874733911904374e-05, |
|
"loss": 2.974, |
|
"step": 31300 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 1.4858359259865728e-05, |
|
"loss": 2.9488, |
|
"step": 31400 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 1.4841984607827085e-05, |
|
"loss": 2.9693, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 1.4825609955788441e-05, |
|
"loss": 2.9617, |
|
"step": 31600 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 1.4809235303749796e-05, |
|
"loss": 2.9411, |
|
"step": 31700 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 1.4792860651711152e-05, |
|
"loss": 2.9701, |
|
"step": 31800 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 1.4776485999672508e-05, |
|
"loss": 2.9615, |
|
"step": 31900 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 1.4760111347633863e-05, |
|
"loss": 2.9625, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 1.4743736695595219e-05, |
|
"loss": 2.9556, |
|
"step": 32100 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 1.4727362043556577e-05, |
|
"loss": 2.9596, |
|
"step": 32200 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 1.4710987391517932e-05, |
|
"loss": 2.9666, |
|
"step": 32300 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 1.4694612739479288e-05, |
|
"loss": 2.9693, |
|
"step": 32400 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 1.4678238087440644e-05, |
|
"loss": 2.9688, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 1.4661863435401999e-05, |
|
"loss": 2.9515, |
|
"step": 32600 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 1.4645488783363355e-05, |
|
"loss": 2.968, |
|
"step": 32700 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 1.4629114131324712e-05, |
|
"loss": 2.9521, |
|
"step": 32800 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 1.4612739479286066e-05, |
|
"loss": 2.9489, |
|
"step": 32900 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 1.4596364827247423e-05, |
|
"loss": 2.9333, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 1.4579990175208777e-05, |
|
"loss": 2.9158, |
|
"step": 33100 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 1.4563615523170134e-05, |
|
"loss": 2.9574, |
|
"step": 33200 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.454724087113149e-05, |
|
"loss": 2.947, |
|
"step": 33300 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.4530866219092845e-05, |
|
"loss": 2.9294, |
|
"step": 33400 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.4514491567054201e-05, |
|
"loss": 2.924, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.4498116915015557e-05, |
|
"loss": 2.9552, |
|
"step": 33600 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.4481742262976912e-05, |
|
"loss": 2.9372, |
|
"step": 33700 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.4465367610938268e-05, |
|
"loss": 2.9317, |
|
"step": 33800 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.4448992958899623e-05, |
|
"loss": 2.9588, |
|
"step": 33900 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.4432618306860979e-05, |
|
"loss": 2.9627, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.4416243654822337e-05, |
|
"loss": 2.9301, |
|
"step": 34100 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.4399869002783693e-05, |
|
"loss": 2.9504, |
|
"step": 34200 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.4383494350745048e-05, |
|
"loss": 2.9052, |
|
"step": 34300 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.4367119698706404e-05, |
|
"loss": 2.9013, |
|
"step": 34400 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.435074504666776e-05, |
|
"loss": 2.9394, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.4334370394629115e-05, |
|
"loss": 2.9642, |
|
"step": 34600 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.4317995742590472e-05, |
|
"loss": 2.9105, |
|
"step": 34700 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.4301621090551826e-05, |
|
"loss": 2.9039, |
|
"step": 34800 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.4285246438513183e-05, |
|
"loss": 2.9139, |
|
"step": 34900 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 1.4268871786474539e-05, |
|
"loss": 2.9577, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 1.4252497134435894e-05, |
|
"loss": 2.9013, |
|
"step": 35100 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.423612248239725e-05, |
|
"loss": 2.9357, |
|
"step": 35200 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.4219747830358606e-05, |
|
"loss": 2.9391, |
|
"step": 35300 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.4203373178319961e-05, |
|
"loss": 2.9265, |
|
"step": 35400 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.4186998526281317e-05, |
|
"loss": 2.935, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.4170623874242674e-05, |
|
"loss": 2.9592, |
|
"step": 35600 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.4154249222204028e-05, |
|
"loss": 2.9216, |
|
"step": 35700 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.4137874570165385e-05, |
|
"loss": 2.924, |
|
"step": 35800 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 1.4121499918126743e-05, |
|
"loss": 2.9216, |
|
"step": 35900 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.4105125266088097e-05, |
|
"loss": 2.9226, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 1.4088750614049453e-05, |
|
"loss": 2.9056, |
|
"step": 36100 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 1.407237596201081e-05, |
|
"loss": 2.9386, |
|
"step": 36200 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.4056001309972164e-05, |
|
"loss": 2.9027, |
|
"step": 36300 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.403962665793352e-05, |
|
"loss": 2.9573, |
|
"step": 36400 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.4023252005894875e-05, |
|
"loss": 2.9368, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.4006877353856232e-05, |
|
"loss": 2.905, |
|
"step": 36600 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_bleu": 871.025, |
|
"eval_chrf": 2652.0809, |
|
"eval_chrf_plus": 2492.3287, |
|
"eval_gen_len": 14.666, |
|
"eval_loss": 2.5985164642333984, |
|
"eval_runtime": 38.3521, |
|
"eval_samples_per_second": 26.074, |
|
"eval_steps_per_second": 1.095, |
|
"step": 36642 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.3990502701817588e-05, |
|
"loss": 2.8775, |
|
"step": 36700 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 1.3974128049778943e-05, |
|
"loss": 2.8513, |
|
"step": 36800 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 1.3957753397740299e-05, |
|
"loss": 2.8564, |
|
"step": 36900 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 1.3941378745701655e-05, |
|
"loss": 2.8596, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 1.392500409366301e-05, |
|
"loss": 2.8591, |
|
"step": 37100 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 1.3908629441624366e-05, |
|
"loss": 2.8619, |
|
"step": 37200 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 1.3892254789585723e-05, |
|
"loss": 2.8753, |
|
"step": 37300 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 1.3875880137547077e-05, |
|
"loss": 2.8761, |
|
"step": 37400 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 1.3859505485508434e-05, |
|
"loss": 2.8231, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 1.3843130833469788e-05, |
|
"loss": 2.8332, |
|
"step": 37600 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 1.3826756181431145e-05, |
|
"loss": 2.8417, |
|
"step": 37700 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 1.3810381529392503e-05, |
|
"loss": 2.8426, |
|
"step": 37800 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 1.3794006877353859e-05, |
|
"loss": 2.8603, |
|
"step": 37900 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 1.3777632225315213e-05, |
|
"loss": 2.8538, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 1.376125757327657e-05, |
|
"loss": 2.8442, |
|
"step": 38100 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 1.3744882921237926e-05, |
|
"loss": 2.8499, |
|
"step": 38200 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 1.372850826919928e-05, |
|
"loss": 2.8347, |
|
"step": 38300 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 1.3712133617160637e-05, |
|
"loss": 2.8528, |
|
"step": 38400 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 1.3695758965121992e-05, |
|
"loss": 2.8532, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 1.3679384313083348e-05, |
|
"loss": 2.8538, |
|
"step": 38600 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 1.3663009661044704e-05, |
|
"loss": 2.8473, |
|
"step": 38700 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 1.3646635009006059e-05, |
|
"loss": 2.8365, |
|
"step": 38800 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 1.3630260356967415e-05, |
|
"loss": 2.8565, |
|
"step": 38900 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 1.3613885704928772e-05, |
|
"loss": 2.8504, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 1.3597511052890126e-05, |
|
"loss": 2.8732, |
|
"step": 39100 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 1.3581136400851483e-05, |
|
"loss": 2.8724, |
|
"step": 39200 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 1.3564761748812837e-05, |
|
"loss": 2.8346, |
|
"step": 39300 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 1.3548387096774194e-05, |
|
"loss": 2.8302, |
|
"step": 39400 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 1.353201244473555e-05, |
|
"loss": 2.8644, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 1.3515637792696908e-05, |
|
"loss": 2.8376, |
|
"step": 39600 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 1.3499263140658263e-05, |
|
"loss": 2.8783, |
|
"step": 39700 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 1.3482888488619619e-05, |
|
"loss": 2.8492, |
|
"step": 39800 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 1.3466513836580975e-05, |
|
"loss": 2.8266, |
|
"step": 39900 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 1.345013918454233e-05, |
|
"loss": 2.8566, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 1.3433764532503686e-05, |
|
"loss": 2.8507, |
|
"step": 40100 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 1.341738988046504e-05, |
|
"loss": 2.872, |
|
"step": 40200 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 1.3401015228426397e-05, |
|
"loss": 2.8386, |
|
"step": 40300 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 1.3384640576387753e-05, |
|
"loss": 2.8702, |
|
"step": 40400 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 1.3368265924349108e-05, |
|
"loss": 2.8243, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 1.3351891272310464e-05, |
|
"loss": 2.8271, |
|
"step": 40600 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 1.333551662027182e-05, |
|
"loss": 2.8504, |
|
"step": 40700 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 1.3319141968233175e-05, |
|
"loss": 2.8346, |
|
"step": 40800 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 1.3302767316194532e-05, |
|
"loss": 2.8516, |
|
"step": 40900 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 1.3286392664155888e-05, |
|
"loss": 2.8289, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 1.3270018012117243e-05, |
|
"loss": 2.8258, |
|
"step": 41100 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 1.3253643360078599e-05, |
|
"loss": 2.8654, |
|
"step": 41200 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 1.3237268708039954e-05, |
|
"loss": 2.84, |
|
"step": 41300 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 1.3220894056001312e-05, |
|
"loss": 2.8269, |
|
"step": 41400 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 1.3204519403962668e-05, |
|
"loss": 2.8432, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 1.3188144751924024e-05, |
|
"loss": 2.8462, |
|
"step": 41600 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 1.3171770099885379e-05, |
|
"loss": 2.8131, |
|
"step": 41700 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 1.3155395447846735e-05, |
|
"loss": 2.826, |
|
"step": 41800 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 1.313902079580809e-05, |
|
"loss": 2.8396, |
|
"step": 41900 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 1.3122646143769446e-05, |
|
"loss": 2.8241, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 1.3106271491730802e-05, |
|
"loss": 2.7975, |
|
"step": 42100 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 1.3089896839692157e-05, |
|
"loss": 2.8249, |
|
"step": 42200 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 1.3073522187653513e-05, |
|
"loss": 2.8189, |
|
"step": 42300 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 1.305714753561487e-05, |
|
"loss": 2.8253, |
|
"step": 42400 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 1.3040772883576224e-05, |
|
"loss": 2.8061, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 1.302439823153758e-05, |
|
"loss": 2.8074, |
|
"step": 42600 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 1.3008023579498937e-05, |
|
"loss": 2.8325, |
|
"step": 42700 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 1.2991648927460292e-05, |
|
"loss": 2.8204, |
|
"step": 42800 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 1.2975274275421648e-05, |
|
"loss": 2.8388, |
|
"step": 42900 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 1.2958899623383003e-05, |
|
"loss": 2.8278, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 1.2942524971344359e-05, |
|
"loss": 2.8339, |
|
"step": 43100 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 1.2926150319305715e-05, |
|
"loss": 2.804, |
|
"step": 43200 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 1.2909775667267073e-05, |
|
"loss": 2.8434, |
|
"step": 43300 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 1.2893401015228428e-05, |
|
"loss": 2.8266, |
|
"step": 43400 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 1.2877026363189784e-05, |
|
"loss": 2.8084, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 1.286065171115114e-05, |
|
"loss": 2.808, |
|
"step": 43600 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 1.2844277059112495e-05, |
|
"loss": 2.8342, |
|
"step": 43700 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 1.2827902407073852e-05, |
|
"loss": 2.8459, |
|
"step": 43800 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 1.2811527755035206e-05, |
|
"loss": 2.8219, |
|
"step": 43900 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 1.2795153102996562e-05, |
|
"loss": 2.8075, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 1.2778778450957919e-05, |
|
"loss": 2.8214, |
|
"step": 44100 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 1.2762403798919273e-05, |
|
"loss": 2.8223, |
|
"step": 44200 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 1.274602914688063e-05, |
|
"loss": 2.8085, |
|
"step": 44300 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 1.2729654494841986e-05, |
|
"loss": 2.8086, |
|
"step": 44400 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 1.271327984280334e-05, |
|
"loss": 2.8039, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 1.2696905190764697e-05, |
|
"loss": 2.8223, |
|
"step": 44600 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 1.2680530538726052e-05, |
|
"loss": 2.8165, |
|
"step": 44700 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 1.2664155886687408e-05, |
|
"loss": 2.8012, |
|
"step": 44800 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 1.2647781234648764e-05, |
|
"loss": 2.8065, |
|
"step": 44900 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 1.2631406582610119e-05, |
|
"loss": 2.8251, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 1.2615031930571477e-05, |
|
"loss": 2.8402, |
|
"step": 45100 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 1.2598657278532833e-05, |
|
"loss": 2.7951, |
|
"step": 45200 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 1.258228262649419e-05, |
|
"loss": 2.8053, |
|
"step": 45300 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 1.2565907974455544e-05, |
|
"loss": 2.8495, |
|
"step": 45400 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 1.25495333224169e-05, |
|
"loss": 2.8131, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 1.2533158670378255e-05, |
|
"loss": 2.7888, |
|
"step": 45600 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 1.2516784018339612e-05, |
|
"loss": 2.8165, |
|
"step": 45700 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 1.2500409366300968e-05, |
|
"loss": 2.814, |
|
"step": 45800 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 1.2484034714262322e-05, |
|
"loss": 2.8309, |
|
"step": 45900 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 1.2467660062223679e-05, |
|
"loss": 2.7945, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 1.2451285410185035e-05, |
|
"loss": 2.7825, |
|
"step": 46100 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 1.243491075814639e-05, |
|
"loss": 2.7801, |
|
"step": 46200 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 1.2418536106107746e-05, |
|
"loss": 2.8058, |
|
"step": 46300 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 1.2402161454069102e-05, |
|
"loss": 2.8132, |
|
"step": 46400 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 1.2385786802030457e-05, |
|
"loss": 2.807, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 1.2369412149991813e-05, |
|
"loss": 2.7962, |
|
"step": 46600 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 1.2353037497953168e-05, |
|
"loss": 2.8217, |
|
"step": 46700 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 1.2336662845914524e-05, |
|
"loss": 2.7893, |
|
"step": 46800 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 1.232028819387588e-05, |
|
"loss": 2.7977, |
|
"step": 46900 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 1.2303913541837239e-05, |
|
"loss": 2.8111, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 1.2287538889798593e-05, |
|
"loss": 2.7867, |
|
"step": 47100 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 1.227116423775995e-05, |
|
"loss": 2.7771, |
|
"step": 47200 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 1.2254789585721304e-05, |
|
"loss": 2.7968, |
|
"step": 47300 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 1.223841493368266e-05, |
|
"loss": 2.8271, |
|
"step": 47400 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 1.2222040281644017e-05, |
|
"loss": 2.8087, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 1.2205665629605372e-05, |
|
"loss": 2.8017, |
|
"step": 47600 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 1.2189290977566728e-05, |
|
"loss": 2.8084, |
|
"step": 47700 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 1.2172916325528084e-05, |
|
"loss": 2.8168, |
|
"step": 47800 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 1.2156541673489439e-05, |
|
"loss": 2.8394, |
|
"step": 47900 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 1.2140167021450795e-05, |
|
"loss": 2.7815, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 1.2123792369412151e-05, |
|
"loss": 2.8058, |
|
"step": 48100 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 1.2107417717373506e-05, |
|
"loss": 2.771, |
|
"step": 48200 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 1.2091043065334862e-05, |
|
"loss": 2.8163, |
|
"step": 48300 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 1.2074668413296217e-05, |
|
"loss": 2.8269, |
|
"step": 48400 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 1.2058293761257573e-05, |
|
"loss": 2.823, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 1.204191910921893e-05, |
|
"loss": 2.7733, |
|
"step": 48600 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 1.2025544457180284e-05, |
|
"loss": 2.7565, |
|
"step": 48700 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 1.2009169805141642e-05, |
|
"loss": 2.7801, |
|
"step": 48800 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_bleu": 901.064, |
|
"eval_chrf": 2735.6707, |
|
"eval_chrf_plus": 2574.683, |
|
"eval_gen_len": 14.662, |
|
"eval_loss": 2.527679920196533, |
|
"eval_runtime": 38.6477, |
|
"eval_samples_per_second": 25.875, |
|
"eval_steps_per_second": 1.087, |
|
"step": 48856 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 1.1992795153102999e-05, |
|
"loss": 2.7625, |
|
"step": 48900 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 1.1976420501064355e-05, |
|
"loss": 2.7016, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 1.196004584902571e-05, |
|
"loss": 2.74, |
|
"step": 49100 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 1.1943671196987066e-05, |
|
"loss": 2.7466, |
|
"step": 49200 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 1.192729654494842e-05, |
|
"loss": 2.7237, |
|
"step": 49300 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 1.1910921892909777e-05, |
|
"loss": 2.768, |
|
"step": 49400 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 1.1894547240871133e-05, |
|
"loss": 2.7265, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"learning_rate": 1.1878172588832488e-05, |
|
"loss": 2.7179, |
|
"step": 49600 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 1.1861797936793844e-05, |
|
"loss": 2.6952, |
|
"step": 49700 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"learning_rate": 1.18454232847552e-05, |
|
"loss": 2.7131, |
|
"step": 49800 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 1.1829048632716555e-05, |
|
"loss": 2.7492, |
|
"step": 49900 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 1.1812673980677911e-05, |
|
"loss": 2.7349, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 1.1796299328639266e-05, |
|
"loss": 2.7603, |
|
"step": 50100 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 1.1779924676600622e-05, |
|
"loss": 2.7132, |
|
"step": 50200 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 1.1763550024561979e-05, |
|
"loss": 2.7195, |
|
"step": 50300 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 1.1747175372523333e-05, |
|
"loss": 2.7586, |
|
"step": 50400 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 1.173080072048469e-05, |
|
"loss": 2.6895, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 1.1714426068446048e-05, |
|
"loss": 2.7115, |
|
"step": 50600 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 1.1698051416407404e-05, |
|
"loss": 2.7099, |
|
"step": 50700 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"learning_rate": 1.1681676764368759e-05, |
|
"loss": 2.7313, |
|
"step": 50800 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 1.1665302112330115e-05, |
|
"loss": 2.7493, |
|
"step": 50900 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 1.164892746029147e-05, |
|
"loss": 2.7305, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 1.1632552808252826e-05, |
|
"loss": 2.7458, |
|
"step": 51100 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 1.1616178156214182e-05, |
|
"loss": 2.6953, |
|
"step": 51200 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 1.1599803504175537e-05, |
|
"loss": 2.7367, |
|
"step": 51300 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 1.1583428852136893e-05, |
|
"loss": 2.7498, |
|
"step": 51400 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 1.156705420009825e-05, |
|
"loss": 2.727, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 1.1550679548059604e-05, |
|
"loss": 2.7481, |
|
"step": 51600 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 1.153430489602096e-05, |
|
"loss": 2.7545, |
|
"step": 51700 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 1.1517930243982317e-05, |
|
"loss": 2.7321, |
|
"step": 51800 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 1.1501555591943671e-05, |
|
"loss": 2.7258, |
|
"step": 51900 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 1.1485180939905028e-05, |
|
"loss": 2.7123, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 1.1468806287866382e-05, |
|
"loss": 2.7354, |
|
"step": 52100 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 1.1452431635827739e-05, |
|
"loss": 2.7387, |
|
"step": 52200 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 1.1436056983789095e-05, |
|
"loss": 2.7635, |
|
"step": 52300 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"learning_rate": 1.141968233175045e-05, |
|
"loss": 2.7104, |
|
"step": 52400 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 1.1403307679711808e-05, |
|
"loss": 2.7499, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 1.1386933027673164e-05, |
|
"loss": 2.7054, |
|
"step": 52600 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 1.1370558375634519e-05, |
|
"loss": 2.7084, |
|
"step": 52700 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"learning_rate": 1.1354183723595875e-05, |
|
"loss": 2.7467, |
|
"step": 52800 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 1.1337809071557231e-05, |
|
"loss": 2.7327, |
|
"step": 52900 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 1.1321434419518586e-05, |
|
"loss": 2.6927, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 4.35, |
|
"learning_rate": 1.1305059767479942e-05, |
|
"loss": 2.705, |
|
"step": 53100 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 1.1288685115441299e-05, |
|
"loss": 2.7123, |
|
"step": 53200 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 1.1272310463402653e-05, |
|
"loss": 2.7319, |
|
"step": 53300 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 1.125593581136401e-05, |
|
"loss": 2.7112, |
|
"step": 53400 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"learning_rate": 1.1239561159325366e-05, |
|
"loss": 2.7509, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 1.122318650728672e-05, |
|
"loss": 2.7256, |
|
"step": 53600 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 1.1206811855248077e-05, |
|
"loss": 2.726, |
|
"step": 53700 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 1.1190437203209431e-05, |
|
"loss": 2.6912, |
|
"step": 53800 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 1.1174062551170788e-05, |
|
"loss": 2.7434, |
|
"step": 53900 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 1.1157687899132144e-05, |
|
"loss": 2.7085, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 1.1141313247093499e-05, |
|
"loss": 2.6887, |
|
"step": 54100 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 1.1124938595054855e-05, |
|
"loss": 2.7507, |
|
"step": 54200 |
|
}, |
|
{ |
|
"epoch": 4.45, |
|
"learning_rate": 1.1108563943016213e-05, |
|
"loss": 2.7167, |
|
"step": 54300 |
|
}, |
|
{ |
|
"epoch": 4.45, |
|
"learning_rate": 1.109218929097757e-05, |
|
"loss": 2.7093, |
|
"step": 54400 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 1.1075814638938924e-05, |
|
"loss": 2.7261, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 1.105943998690028e-05, |
|
"loss": 2.7224, |
|
"step": 54600 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"learning_rate": 1.1043065334861635e-05, |
|
"loss": 2.7195, |
|
"step": 54700 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 1.1026690682822991e-05, |
|
"loss": 2.7271, |
|
"step": 54800 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 1.1010316030784348e-05, |
|
"loss": 2.6997, |
|
"step": 54900 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 1.0993941378745702e-05, |
|
"loss": 2.7087, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 1.0977566726707059e-05, |
|
"loss": 2.71, |
|
"step": 55100 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 1.0961192074668415e-05, |
|
"loss": 2.748, |
|
"step": 55200 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 1.094481742262977e-05, |
|
"loss": 2.7256, |
|
"step": 55300 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 1.0928442770591126e-05, |
|
"loss": 2.7217, |
|
"step": 55400 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 1.091206811855248e-05, |
|
"loss": 2.7287, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 1.0895693466513837e-05, |
|
"loss": 2.7147, |
|
"step": 55600 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 1.0879318814475193e-05, |
|
"loss": 2.6992, |
|
"step": 55700 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 1.0862944162436548e-05, |
|
"loss": 2.7216, |
|
"step": 55800 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 1.0846569510397904e-05, |
|
"loss": 2.7001, |
|
"step": 55900 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 1.083019485835926e-05, |
|
"loss": 2.7217, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 1.0813820206320615e-05, |
|
"loss": 2.7392, |
|
"step": 56100 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 1.0797445554281973e-05, |
|
"loss": 2.7131, |
|
"step": 56200 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 1.078107090224333e-05, |
|
"loss": 2.6884, |
|
"step": 56300 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 1.0764696250204684e-05, |
|
"loss": 2.7238, |
|
"step": 56400 |
|
}, |
|
{ |
|
"epoch": 4.63, |
|
"learning_rate": 1.074832159816604e-05, |
|
"loss": 2.7126, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 4.63, |
|
"learning_rate": 1.0731946946127397e-05, |
|
"loss": 2.712, |
|
"step": 56600 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 1.0715572294088751e-05, |
|
"loss": 2.7054, |
|
"step": 56700 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 1.0699197642050108e-05, |
|
"loss": 2.7433, |
|
"step": 56800 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 1.0682822990011464e-05, |
|
"loss": 2.6941, |
|
"step": 56900 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 1.0666448337972819e-05, |
|
"loss": 2.724, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 1.0650073685934175e-05, |
|
"loss": 2.6911, |
|
"step": 57100 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 1.0633699033895531e-05, |
|
"loss": 2.7282, |
|
"step": 57200 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"learning_rate": 1.0617324381856886e-05, |
|
"loss": 2.7163, |
|
"step": 57300 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 1.0600949729818242e-05, |
|
"loss": 2.7058, |
|
"step": 57400 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 1.0584575077779597e-05, |
|
"loss": 2.7049, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 1.0568200425740953e-05, |
|
"loss": 2.7375, |
|
"step": 57600 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 1.055182577370231e-05, |
|
"loss": 2.7072, |
|
"step": 57700 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 1.0535451121663664e-05, |
|
"loss": 2.7157, |
|
"step": 57800 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 1.051907646962502e-05, |
|
"loss": 2.7098, |
|
"step": 57900 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 1.0502701817586378e-05, |
|
"loss": 2.7254, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 4.76, |
|
"learning_rate": 1.0486327165547733e-05, |
|
"loss": 2.6918, |
|
"step": 58100 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 1.046995251350909e-05, |
|
"loss": 2.7007, |
|
"step": 58200 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 1.0453577861470446e-05, |
|
"loss": 2.7673, |
|
"step": 58300 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 1.04372032094318e-05, |
|
"loss": 2.687, |
|
"step": 58400 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 1.0420828557393157e-05, |
|
"loss": 2.6976, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 1.0404453905354513e-05, |
|
"loss": 2.7166, |
|
"step": 58600 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 1.0388079253315868e-05, |
|
"loss": 2.7329, |
|
"step": 58700 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 1.0371704601277224e-05, |
|
"loss": 2.7008, |
|
"step": 58800 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 1.035532994923858e-05, |
|
"loss": 2.7011, |
|
"step": 58900 |
|
}, |
|
{ |
|
"epoch": 4.83, |
|
"learning_rate": 1.0338955297199935e-05, |
|
"loss": 2.7459, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 1.0322580645161291e-05, |
|
"loss": 2.6775, |
|
"step": 59100 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 1.0306205993122646e-05, |
|
"loss": 2.7026, |
|
"step": 59200 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 1.0289831341084002e-05, |
|
"loss": 2.7412, |
|
"step": 59300 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 1.0273456689045359e-05, |
|
"loss": 2.7302, |
|
"step": 59400 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 1.0257082037006713e-05, |
|
"loss": 2.7087, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 4.88, |
|
"learning_rate": 1.024070738496807e-05, |
|
"loss": 2.7151, |
|
"step": 59600 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 1.0224332732929426e-05, |
|
"loss": 2.6936, |
|
"step": 59700 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 1.0207958080890784e-05, |
|
"loss": 2.705, |
|
"step": 59800 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 1.0191583428852138e-05, |
|
"loss": 2.7276, |
|
"step": 59900 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 1.0175208776813495e-05, |
|
"loss": 2.7304, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 4.92, |
|
"learning_rate": 1.015883412477485e-05, |
|
"loss": 2.7073, |
|
"step": 60100 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 1.0142459472736206e-05, |
|
"loss": 2.6806, |
|
"step": 60200 |
|
}, |
|
{ |
|
"epoch": 4.94, |
|
"learning_rate": 1.0126084820697562e-05, |
|
"loss": 2.7106, |
|
"step": 60300 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 1.0109710168658917e-05, |
|
"loss": 2.6818, |
|
"step": 60400 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 1.0093335516620273e-05, |
|
"loss": 2.7072, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 1.007696086458163e-05, |
|
"loss": 2.6981, |
|
"step": 60600 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 1.0060586212542984e-05, |
|
"loss": 2.6884, |
|
"step": 60700 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"learning_rate": 1.004421156050434e-05, |
|
"loss": 2.6921, |
|
"step": 60800 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 1.0027836908465695e-05, |
|
"loss": 2.6885, |
|
"step": 60900 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 1.0011462256427051e-05, |
|
"loss": 2.7105, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_bleu": 941.391, |
|
"eval_chrf": 2793.3653, |
|
"eval_chrf_plus": 2625.3519, |
|
"eval_gen_len": 14.666, |
|
"eval_loss": 2.477811813354492, |
|
"eval_runtime": 38.3865, |
|
"eval_samples_per_second": 26.051, |
|
"eval_steps_per_second": 1.094, |
|
"step": 61070 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 9.995087604388408e-06, |
|
"loss": 2.7037, |
|
"step": 61100 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 9.978712952349764e-06, |
|
"loss": 2.6766, |
|
"step": 61200 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 9.96233830031112e-06, |
|
"loss": 2.6567, |
|
"step": 61300 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 9.945963648272475e-06, |
|
"loss": 2.6702, |
|
"step": 61400 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 9.929588996233831e-06, |
|
"loss": 2.6706, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 9.913214344195187e-06, |
|
"loss": 2.6327, |
|
"step": 61600 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 9.896839692156542e-06, |
|
"loss": 2.6481, |
|
"step": 61700 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"learning_rate": 9.880465040117898e-06, |
|
"loss": 2.644, |
|
"step": 61800 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 9.864090388079253e-06, |
|
"loss": 2.6387, |
|
"step": 61900 |
|
}, |
|
{ |
|
"epoch": 5.08, |
|
"learning_rate": 9.84771573604061e-06, |
|
"loss": 2.6096, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 5.08, |
|
"learning_rate": 9.831341084001966e-06, |
|
"loss": 2.6466, |
|
"step": 62100 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"learning_rate": 9.814966431963322e-06, |
|
"loss": 2.639, |
|
"step": 62200 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"learning_rate": 9.798591779924678e-06, |
|
"loss": 2.6389, |
|
"step": 62300 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 9.782217127886033e-06, |
|
"loss": 2.615, |
|
"step": 62400 |
|
}, |
|
{ |
|
"epoch": 5.12, |
|
"learning_rate": 9.76584247584739e-06, |
|
"loss": 2.6486, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"learning_rate": 9.749467823808746e-06, |
|
"loss": 2.6294, |
|
"step": 62600 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"learning_rate": 9.7330931717701e-06, |
|
"loss": 2.6415, |
|
"step": 62700 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 9.716718519731457e-06, |
|
"loss": 2.6744, |
|
"step": 62800 |
|
}, |
|
{ |
|
"epoch": 5.15, |
|
"learning_rate": 9.700343867692811e-06, |
|
"loss": 2.6236, |
|
"step": 62900 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"learning_rate": 9.683969215654168e-06, |
|
"loss": 2.6074, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"learning_rate": 9.667594563615524e-06, |
|
"loss": 2.6673, |
|
"step": 63100 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"learning_rate": 9.65121991157688e-06, |
|
"loss": 2.6182, |
|
"step": 63200 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 9.634845259538237e-06, |
|
"loss": 2.6363, |
|
"step": 63300 |
|
}, |
|
{ |
|
"epoch": 5.19, |
|
"learning_rate": 9.618470607499591e-06, |
|
"loss": 2.648, |
|
"step": 63400 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"learning_rate": 9.602095955460947e-06, |
|
"loss": 2.6377, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 9.585721303422302e-06, |
|
"loss": 2.6505, |
|
"step": 63600 |
|
}, |
|
{ |
|
"epoch": 5.22, |
|
"learning_rate": 9.569346651383658e-06, |
|
"loss": 2.6499, |
|
"step": 63700 |
|
}, |
|
{ |
|
"epoch": 5.22, |
|
"learning_rate": 9.552971999345015e-06, |
|
"loss": 2.6385, |
|
"step": 63800 |
|
}, |
|
{ |
|
"epoch": 5.23, |
|
"learning_rate": 9.53659734730637e-06, |
|
"loss": 2.662, |
|
"step": 63900 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 9.520222695267727e-06, |
|
"loss": 2.6544, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 5.25, |
|
"learning_rate": 9.503848043229082e-06, |
|
"loss": 2.6315, |
|
"step": 64100 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 9.487473391190438e-06, |
|
"loss": 2.6606, |
|
"step": 64200 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 9.471098739151795e-06, |
|
"loss": 2.6502, |
|
"step": 64300 |
|
}, |
|
{ |
|
"epoch": 5.27, |
|
"learning_rate": 9.45472408711315e-06, |
|
"loss": 2.6328, |
|
"step": 64400 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 9.438349435074506e-06, |
|
"loss": 2.6469, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 5.29, |
|
"learning_rate": 9.42197478303586e-06, |
|
"loss": 2.6143, |
|
"step": 64600 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"learning_rate": 9.405600130997217e-06, |
|
"loss": 2.6565, |
|
"step": 64700 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"learning_rate": 9.389225478958573e-06, |
|
"loss": 2.6213, |
|
"step": 64800 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"learning_rate": 9.37285082691993e-06, |
|
"loss": 2.6449, |
|
"step": 64900 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"learning_rate": 9.356476174881286e-06, |
|
"loss": 2.6446, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"learning_rate": 9.34010152284264e-06, |
|
"loss": 2.6314, |
|
"step": 65100 |
|
}, |
|
{ |
|
"epoch": 5.34, |
|
"learning_rate": 9.323726870803997e-06, |
|
"loss": 2.6453, |
|
"step": 65200 |
|
}, |
|
{ |
|
"epoch": 5.35, |
|
"learning_rate": 9.307352218765353e-06, |
|
"loss": 2.6418, |
|
"step": 65300 |
|
}, |
|
{ |
|
"epoch": 5.35, |
|
"learning_rate": 9.290977566726707e-06, |
|
"loss": 2.6389, |
|
"step": 65400 |
|
}, |
|
{ |
|
"epoch": 5.36, |
|
"learning_rate": 9.274602914688064e-06, |
|
"loss": 2.671, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 9.258228262649418e-06, |
|
"loss": 2.6571, |
|
"step": 65600 |
|
}, |
|
{ |
|
"epoch": 5.38, |
|
"learning_rate": 9.241853610610775e-06, |
|
"loss": 2.6426, |
|
"step": 65700 |
|
}, |
|
{ |
|
"epoch": 5.39, |
|
"learning_rate": 9.225478958572131e-06, |
|
"loss": 2.665, |
|
"step": 65800 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"learning_rate": 9.209104306533487e-06, |
|
"loss": 2.634, |
|
"step": 65900 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"learning_rate": 9.192729654494844e-06, |
|
"loss": 2.6479, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 5.41, |
|
"learning_rate": 9.176355002456198e-06, |
|
"loss": 2.6317, |
|
"step": 66100 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"learning_rate": 9.159980350417555e-06, |
|
"loss": 2.6429, |
|
"step": 66200 |
|
}, |
|
{ |
|
"epoch": 5.43, |
|
"learning_rate": 9.14360569837891e-06, |
|
"loss": 2.638, |
|
"step": 66300 |
|
}, |
|
{ |
|
"epoch": 5.44, |
|
"learning_rate": 9.127231046340266e-06, |
|
"loss": 2.6344, |
|
"step": 66400 |
|
}, |
|
{ |
|
"epoch": 5.44, |
|
"learning_rate": 9.110856394301622e-06, |
|
"loss": 2.6152, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"learning_rate": 9.094481742262977e-06, |
|
"loss": 2.6416, |
|
"step": 66600 |
|
}, |
|
{ |
|
"epoch": 5.46, |
|
"learning_rate": 9.078107090224335e-06, |
|
"loss": 2.6115, |
|
"step": 66700 |
|
}, |
|
{ |
|
"epoch": 5.47, |
|
"learning_rate": 9.06173243818569e-06, |
|
"loss": 2.6862, |
|
"step": 66800 |
|
}, |
|
{ |
|
"epoch": 5.48, |
|
"learning_rate": 9.045357786147046e-06, |
|
"loss": 2.6273, |
|
"step": 66900 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 9.028983134108402e-06, |
|
"loss": 2.6354, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 9.012608482069757e-06, |
|
"loss": 2.635, |
|
"step": 67100 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"learning_rate": 8.996233830031113e-06, |
|
"loss": 2.6303, |
|
"step": 67200 |
|
}, |
|
{ |
|
"epoch": 5.51, |
|
"learning_rate": 8.979859177992467e-06, |
|
"loss": 2.6431, |
|
"step": 67300 |
|
}, |
|
{ |
|
"epoch": 5.52, |
|
"learning_rate": 8.963484525953824e-06, |
|
"loss": 2.6356, |
|
"step": 67400 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 8.94710987391518e-06, |
|
"loss": 2.6505, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 8.930735221876535e-06, |
|
"loss": 2.6433, |
|
"step": 67600 |
|
}, |
|
{ |
|
"epoch": 5.54, |
|
"learning_rate": 8.914360569837893e-06, |
|
"loss": 2.6275, |
|
"step": 67700 |
|
}, |
|
{ |
|
"epoch": 5.55, |
|
"learning_rate": 8.897985917799247e-06, |
|
"loss": 2.6287, |
|
"step": 67800 |
|
}, |
|
{ |
|
"epoch": 5.56, |
|
"learning_rate": 8.881611265760604e-06, |
|
"loss": 2.6528, |
|
"step": 67900 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 8.86523661372196e-06, |
|
"loss": 2.6361, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 5.58, |
|
"learning_rate": 8.848861961683315e-06, |
|
"loss": 2.6482, |
|
"step": 68100 |
|
}, |
|
{ |
|
"epoch": 5.58, |
|
"learning_rate": 8.832487309644671e-06, |
|
"loss": 2.6487, |
|
"step": 68200 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 8.816112657606026e-06, |
|
"loss": 2.6511, |
|
"step": 68300 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 8.799738005567382e-06, |
|
"loss": 2.6125, |
|
"step": 68400 |
|
}, |
|
{ |
|
"epoch": 5.61, |
|
"learning_rate": 8.783363353528738e-06, |
|
"loss": 2.6255, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"learning_rate": 8.766988701490095e-06, |
|
"loss": 2.6203, |
|
"step": 68600 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"learning_rate": 8.750614049451451e-06, |
|
"loss": 2.6158, |
|
"step": 68700 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"learning_rate": 8.734239397412806e-06, |
|
"loss": 2.6542, |
|
"step": 68800 |
|
}, |
|
{ |
|
"epoch": 5.64, |
|
"learning_rate": 8.717864745374162e-06, |
|
"loss": 2.6576, |
|
"step": 68900 |
|
}, |
|
{ |
|
"epoch": 5.65, |
|
"learning_rate": 8.701490093335517e-06, |
|
"loss": 2.6632, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 5.66, |
|
"learning_rate": 8.685115441296873e-06, |
|
"loss": 2.6301, |
|
"step": 69100 |
|
}, |
|
{ |
|
"epoch": 5.67, |
|
"learning_rate": 8.66874078925823e-06, |
|
"loss": 2.6334, |
|
"step": 69200 |
|
}, |
|
{ |
|
"epoch": 5.67, |
|
"learning_rate": 8.652366137219584e-06, |
|
"loss": 2.6105, |
|
"step": 69300 |
|
}, |
|
{ |
|
"epoch": 5.68, |
|
"learning_rate": 8.63599148518094e-06, |
|
"loss": 2.6296, |
|
"step": 69400 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 8.619616833142296e-06, |
|
"loss": 2.6239, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 8.603242181103653e-06, |
|
"loss": 2.6069, |
|
"step": 69600 |
|
}, |
|
{ |
|
"epoch": 5.71, |
|
"learning_rate": 8.586867529065009e-06, |
|
"loss": 2.6313, |
|
"step": 69700 |
|
}, |
|
{ |
|
"epoch": 5.71, |
|
"learning_rate": 8.570492877026364e-06, |
|
"loss": 2.633, |
|
"step": 69800 |
|
}, |
|
{ |
|
"epoch": 5.72, |
|
"learning_rate": 8.55411822498772e-06, |
|
"loss": 2.6343, |
|
"step": 69900 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"learning_rate": 8.537743572949075e-06, |
|
"loss": 2.6294, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 5.74, |
|
"learning_rate": 8.521368920910431e-06, |
|
"loss": 2.6343, |
|
"step": 70100 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"learning_rate": 8.504994268871787e-06, |
|
"loss": 2.6302, |
|
"step": 70200 |
|
}, |
|
{ |
|
"epoch": 5.76, |
|
"learning_rate": 8.488619616833142e-06, |
|
"loss": 2.6118, |
|
"step": 70300 |
|
}, |
|
{ |
|
"epoch": 5.76, |
|
"learning_rate": 8.4722449647945e-06, |
|
"loss": 2.6306, |
|
"step": 70400 |
|
}, |
|
{ |
|
"epoch": 5.77, |
|
"learning_rate": 8.455870312755855e-06, |
|
"loss": 2.6119, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 5.78, |
|
"learning_rate": 8.439495660717211e-06, |
|
"loss": 2.6352, |
|
"step": 70600 |
|
}, |
|
{ |
|
"epoch": 5.79, |
|
"learning_rate": 8.423121008678567e-06, |
|
"loss": 2.6078, |
|
"step": 70700 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 8.406746356639922e-06, |
|
"loss": 2.6543, |
|
"step": 70800 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 8.390371704601278e-06, |
|
"loss": 2.6499, |
|
"step": 70900 |
|
}, |
|
{ |
|
"epoch": 5.81, |
|
"learning_rate": 8.373997052562633e-06, |
|
"loss": 2.6227, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 8.35762240052399e-06, |
|
"loss": 2.6355, |
|
"step": 71100 |
|
}, |
|
{ |
|
"epoch": 5.83, |
|
"learning_rate": 8.341247748485345e-06, |
|
"loss": 2.6287, |
|
"step": 71200 |
|
}, |
|
{ |
|
"epoch": 5.84, |
|
"learning_rate": 8.324873096446702e-06, |
|
"loss": 2.6446, |
|
"step": 71300 |
|
}, |
|
{ |
|
"epoch": 5.85, |
|
"learning_rate": 8.308498444408058e-06, |
|
"loss": 2.641, |
|
"step": 71400 |
|
}, |
|
{ |
|
"epoch": 5.85, |
|
"learning_rate": 8.292123792369413e-06, |
|
"loss": 2.6487, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 8.275749140330769e-06, |
|
"loss": 2.6324, |
|
"step": 71600 |
|
}, |
|
{ |
|
"epoch": 5.87, |
|
"learning_rate": 8.259374488292124e-06, |
|
"loss": 2.6198, |
|
"step": 71700 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 8.24299983625348e-06, |
|
"loss": 2.6237, |
|
"step": 71800 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"learning_rate": 8.226625184214836e-06, |
|
"loss": 2.6339, |
|
"step": 71900 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"learning_rate": 8.210250532176191e-06, |
|
"loss": 2.629, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 5.9, |
|
"learning_rate": 8.193875880137547e-06, |
|
"loss": 2.6365, |
|
"step": 72100 |
|
}, |
|
{ |
|
"epoch": 5.91, |
|
"learning_rate": 8.177501228098904e-06, |
|
"loss": 2.6644, |
|
"step": 72200 |
|
}, |
|
{ |
|
"epoch": 5.92, |
|
"learning_rate": 8.16112657606026e-06, |
|
"loss": 2.6232, |
|
"step": 72300 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"learning_rate": 8.144751924021616e-06, |
|
"loss": 2.6515, |
|
"step": 72400 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 8.128377271982971e-06, |
|
"loss": 2.6166, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 8.112002619944327e-06, |
|
"loss": 2.6354, |
|
"step": 72600 |
|
}, |
|
{ |
|
"epoch": 5.95, |
|
"learning_rate": 8.095627967905682e-06, |
|
"loss": 2.6005, |
|
"step": 72700 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 8.079253315867038e-06, |
|
"loss": 2.6251, |
|
"step": 72800 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"learning_rate": 8.062878663828395e-06, |
|
"loss": 2.5943, |
|
"step": 72900 |
|
}, |
|
{ |
|
"epoch": 5.98, |
|
"learning_rate": 8.04650401178975e-06, |
|
"loss": 2.5967, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 5.98, |
|
"learning_rate": 8.030129359751105e-06, |
|
"loss": 2.6441, |
|
"step": 73100 |
|
}, |
|
{ |
|
"epoch": 5.99, |
|
"learning_rate": 8.013754707712462e-06, |
|
"loss": 2.6309, |
|
"step": 73200 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_bleu": 942.9051, |
|
"eval_chrf": 2828.2232, |
|
"eval_chrf_plus": 2657.258, |
|
"eval_gen_len": 14.677, |
|
"eval_loss": 2.4349613189697266, |
|
"eval_runtime": 38.5767, |
|
"eval_samples_per_second": 25.922, |
|
"eval_steps_per_second": 1.089, |
|
"step": 73284 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 7.997380055673818e-06, |
|
"loss": 2.6094, |
|
"step": 73300 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.981005403635174e-06, |
|
"loss": 2.594, |
|
"step": 73400 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.964630751596529e-06, |
|
"loss": 2.5775, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.948256099557885e-06, |
|
"loss": 2.5653, |
|
"step": 73600 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.93188144751924e-06, |
|
"loss": 2.5773, |
|
"step": 73700 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.915506795480596e-06, |
|
"loss": 2.585, |
|
"step": 73800 |
|
}, |
|
{ |
|
"epoch": 6.05, |
|
"learning_rate": 7.899132143441953e-06, |
|
"loss": 2.5742, |
|
"step": 73900 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"learning_rate": 7.882757491403307e-06, |
|
"loss": 2.5905, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 7.866382839364665e-06, |
|
"loss": 2.5837, |
|
"step": 74100 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 7.85000818732602e-06, |
|
"loss": 2.5667, |
|
"step": 74200 |
|
}, |
|
{ |
|
"epoch": 6.08, |
|
"learning_rate": 7.833633535287376e-06, |
|
"loss": 2.5563, |
|
"step": 74300 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 7.817258883248731e-06, |
|
"loss": 2.5947, |
|
"step": 74400 |
|
}, |
|
{ |
|
"epoch": 6.1, |
|
"learning_rate": 7.800884231210087e-06, |
|
"loss": 2.5635, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 7.784509579171444e-06, |
|
"loss": 2.567, |
|
"step": 74600 |
|
}, |
|
{ |
|
"epoch": 6.12, |
|
"learning_rate": 7.768134927132798e-06, |
|
"loss": 2.5654, |
|
"step": 74700 |
|
}, |
|
{ |
|
"epoch": 6.12, |
|
"learning_rate": 7.751760275094155e-06, |
|
"loss": 2.573, |
|
"step": 74800 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"learning_rate": 7.735385623055511e-06, |
|
"loss": 2.5718, |
|
"step": 74900 |
|
}, |
|
{ |
|
"epoch": 6.14, |
|
"learning_rate": 7.719010971016867e-06, |
|
"loss": 2.5964, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 6.15, |
|
"learning_rate": 7.702636318978224e-06, |
|
"loss": 2.5384, |
|
"step": 75100 |
|
}, |
|
{ |
|
"epoch": 6.16, |
|
"learning_rate": 7.686261666939578e-06, |
|
"loss": 2.5768, |
|
"step": 75200 |
|
}, |
|
{ |
|
"epoch": 6.17, |
|
"learning_rate": 7.669887014900934e-06, |
|
"loss": 2.6085, |
|
"step": 75300 |
|
}, |
|
{ |
|
"epoch": 6.17, |
|
"learning_rate": 7.653512362862289e-06, |
|
"loss": 2.617, |
|
"step": 75400 |
|
}, |
|
{ |
|
"epoch": 6.18, |
|
"learning_rate": 7.637137710823645e-06, |
|
"loss": 2.6014, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 6.19, |
|
"learning_rate": 7.620763058785001e-06, |
|
"loss": 2.572, |
|
"step": 75600 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"learning_rate": 7.604388406746357e-06, |
|
"loss": 2.5623, |
|
"step": 75700 |
|
}, |
|
{ |
|
"epoch": 6.21, |
|
"learning_rate": 7.588013754707713e-06, |
|
"loss": 2.5994, |
|
"step": 75800 |
|
}, |
|
{ |
|
"epoch": 6.21, |
|
"learning_rate": 7.571639102669069e-06, |
|
"loss": 2.5654, |
|
"step": 75900 |
|
}, |
|
{ |
|
"epoch": 6.22, |
|
"learning_rate": 7.555264450630425e-06, |
|
"loss": 2.6115, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 6.23, |
|
"learning_rate": 7.538889798591781e-06, |
|
"loss": 2.5785, |
|
"step": 76100 |
|
}, |
|
{ |
|
"epoch": 6.24, |
|
"learning_rate": 7.522515146553136e-06, |
|
"loss": 2.5755, |
|
"step": 76200 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 7.506140494514493e-06, |
|
"loss": 2.6167, |
|
"step": 76300 |
|
}, |
|
{ |
|
"epoch": 6.26, |
|
"learning_rate": 7.489765842475848e-06, |
|
"loss": 2.5617, |
|
"step": 76400 |
|
}, |
|
{ |
|
"epoch": 6.26, |
|
"learning_rate": 7.473391190437204e-06, |
|
"loss": 2.588, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 6.27, |
|
"learning_rate": 7.457016538398559e-06, |
|
"loss": 2.5944, |
|
"step": 76600 |
|
}, |
|
{ |
|
"epoch": 6.28, |
|
"learning_rate": 7.440641886359915e-06, |
|
"loss": 2.5729, |
|
"step": 76700 |
|
}, |
|
{ |
|
"epoch": 6.29, |
|
"learning_rate": 7.424267234321272e-06, |
|
"loss": 2.5993, |
|
"step": 76800 |
|
}, |
|
{ |
|
"epoch": 6.3, |
|
"learning_rate": 7.407892582282627e-06, |
|
"loss": 2.5932, |
|
"step": 76900 |
|
}, |
|
{ |
|
"epoch": 6.3, |
|
"learning_rate": 7.3915179302439835e-06, |
|
"loss": 2.5664, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 6.31, |
|
"learning_rate": 7.375143278205339e-06, |
|
"loss": 2.5446, |
|
"step": 77100 |
|
}, |
|
{ |
|
"epoch": 6.32, |
|
"learning_rate": 7.3587686261666945e-06, |
|
"loss": 2.5759, |
|
"step": 77200 |
|
}, |
|
{ |
|
"epoch": 6.33, |
|
"learning_rate": 7.34239397412805e-06, |
|
"loss": 2.5856, |
|
"step": 77300 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"learning_rate": 7.326019322089406e-06, |
|
"loss": 2.5861, |
|
"step": 77400 |
|
}, |
|
{ |
|
"epoch": 6.35, |
|
"learning_rate": 7.309644670050762e-06, |
|
"loss": 2.5872, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 6.35, |
|
"learning_rate": 7.293270018012117e-06, |
|
"loss": 2.617, |
|
"step": 77600 |
|
}, |
|
{ |
|
"epoch": 6.36, |
|
"learning_rate": 7.2768953659734735e-06, |
|
"loss": 2.576, |
|
"step": 77700 |
|
}, |
|
{ |
|
"epoch": 6.37, |
|
"learning_rate": 7.26052071393483e-06, |
|
"loss": 2.5765, |
|
"step": 77800 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 7.244146061896185e-06, |
|
"loss": 2.5547, |
|
"step": 77900 |
|
}, |
|
{ |
|
"epoch": 6.39, |
|
"learning_rate": 7.227771409857542e-06, |
|
"loss": 2.5883, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 6.39, |
|
"learning_rate": 7.211396757818897e-06, |
|
"loss": 2.5414, |
|
"step": 78100 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 7.195022105780253e-06, |
|
"loss": 2.557, |
|
"step": 78200 |
|
}, |
|
{ |
|
"epoch": 6.41, |
|
"learning_rate": 7.178647453741608e-06, |
|
"loss": 2.5618, |
|
"step": 78300 |
|
}, |
|
{ |
|
"epoch": 6.42, |
|
"learning_rate": 7.1622728017029644e-06, |
|
"loss": 2.5766, |
|
"step": 78400 |
|
}, |
|
{ |
|
"epoch": 6.43, |
|
"learning_rate": 7.14589814966432e-06, |
|
"loss": 2.5876, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 7.129523497625675e-06, |
|
"loss": 2.5599, |
|
"step": 78600 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 7.1131488455870326e-06, |
|
"loss": 2.5776, |
|
"step": 78700 |
|
}, |
|
{ |
|
"epoch": 6.45, |
|
"learning_rate": 7.096774193548388e-06, |
|
"loss": 2.5826, |
|
"step": 78800 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 7.0803995415097435e-06, |
|
"loss": 2.5747, |
|
"step": 78900 |
|
}, |
|
{ |
|
"epoch": 6.47, |
|
"learning_rate": 7.0640248894711e-06, |
|
"loss": 2.6132, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 6.48, |
|
"learning_rate": 7.047650237432455e-06, |
|
"loss": 2.567, |
|
"step": 79100 |
|
}, |
|
{ |
|
"epoch": 6.48, |
|
"learning_rate": 7.031275585393811e-06, |
|
"loss": 2.5779, |
|
"step": 79200 |
|
}, |
|
{ |
|
"epoch": 6.49, |
|
"learning_rate": 7.014900933355166e-06, |
|
"loss": 2.5782, |
|
"step": 79300 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"learning_rate": 6.998526281316523e-06, |
|
"loss": 2.5965, |
|
"step": 79400 |
|
}, |
|
{ |
|
"epoch": 6.51, |
|
"learning_rate": 6.982151629277878e-06, |
|
"loss": 2.5868, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 6.52, |
|
"learning_rate": 6.965776977239234e-06, |
|
"loss": 2.5889, |
|
"step": 79600 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 6.949402325200591e-06, |
|
"loss": 2.5912, |
|
"step": 79700 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 6.933027673161946e-06, |
|
"loss": 2.576, |
|
"step": 79800 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 6.916653021123302e-06, |
|
"loss": 2.5752, |
|
"step": 79900 |
|
}, |
|
{ |
|
"epoch": 6.55, |
|
"learning_rate": 6.900278369084657e-06, |
|
"loss": 2.5793, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"learning_rate": 6.8839037170460135e-06, |
|
"loss": 2.5879, |
|
"step": 80100 |
|
}, |
|
{ |
|
"epoch": 6.57, |
|
"learning_rate": 6.867529065007369e-06, |
|
"loss": 2.5638, |
|
"step": 80200 |
|
}, |
|
{ |
|
"epoch": 6.57, |
|
"learning_rate": 6.8511544129687244e-06, |
|
"loss": 2.5808, |
|
"step": 80300 |
|
}, |
|
{ |
|
"epoch": 6.58, |
|
"learning_rate": 6.834779760930081e-06, |
|
"loss": 2.5688, |
|
"step": 80400 |
|
}, |
|
{ |
|
"epoch": 6.59, |
|
"learning_rate": 6.818405108891437e-06, |
|
"loss": 2.5831, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 6.8020304568527926e-06, |
|
"loss": 2.5584, |
|
"step": 80600 |
|
}, |
|
{ |
|
"epoch": 6.61, |
|
"learning_rate": 6.785655804814149e-06, |
|
"loss": 2.6074, |
|
"step": 80700 |
|
}, |
|
{ |
|
"epoch": 6.62, |
|
"learning_rate": 6.769281152775504e-06, |
|
"loss": 2.5648, |
|
"step": 80800 |
|
}, |
|
{ |
|
"epoch": 6.62, |
|
"learning_rate": 6.75290650073686e-06, |
|
"loss": 2.5987, |
|
"step": 80900 |
|
}, |
|
{ |
|
"epoch": 6.63, |
|
"learning_rate": 6.736531848698215e-06, |
|
"loss": 2.5462, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 6.64, |
|
"learning_rate": 6.720157196659572e-06, |
|
"loss": 2.5563, |
|
"step": 81100 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 6.703782544620927e-06, |
|
"loss": 2.6114, |
|
"step": 81200 |
|
}, |
|
{ |
|
"epoch": 6.66, |
|
"learning_rate": 6.687407892582283e-06, |
|
"loss": 2.5669, |
|
"step": 81300 |
|
}, |
|
{ |
|
"epoch": 6.66, |
|
"learning_rate": 6.67103324054364e-06, |
|
"loss": 2.5639, |
|
"step": 81400 |
|
}, |
|
{ |
|
"epoch": 6.67, |
|
"learning_rate": 6.654658588504995e-06, |
|
"loss": 2.626, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 6.68, |
|
"learning_rate": 6.638283936466351e-06, |
|
"loss": 2.5719, |
|
"step": 81600 |
|
}, |
|
{ |
|
"epoch": 6.69, |
|
"learning_rate": 6.621909284427707e-06, |
|
"loss": 2.5795, |
|
"step": 81700 |
|
}, |
|
{ |
|
"epoch": 6.7, |
|
"learning_rate": 6.6055346323890625e-06, |
|
"loss": 2.5726, |
|
"step": 81800 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 6.589159980350418e-06, |
|
"loss": 2.5499, |
|
"step": 81900 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 6.5727853283117735e-06, |
|
"loss": 2.5656, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 6.72, |
|
"learning_rate": 6.55641067627313e-06, |
|
"loss": 2.5656, |
|
"step": 82100 |
|
}, |
|
{ |
|
"epoch": 6.73, |
|
"learning_rate": 6.540036024234485e-06, |
|
"loss": 2.5595, |
|
"step": 82200 |
|
}, |
|
{ |
|
"epoch": 6.74, |
|
"learning_rate": 6.523661372195841e-06, |
|
"loss": 2.5696, |
|
"step": 82300 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 6.507286720157198e-06, |
|
"loss": 2.6089, |
|
"step": 82400 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 6.490912068118553e-06, |
|
"loss": 2.5665, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 6.76, |
|
"learning_rate": 6.474537416079909e-06, |
|
"loss": 2.5973, |
|
"step": 82600 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 6.458162764041264e-06, |
|
"loss": 2.5699, |
|
"step": 82700 |
|
}, |
|
{ |
|
"epoch": 6.78, |
|
"learning_rate": 6.441788112002621e-06, |
|
"loss": 2.5383, |
|
"step": 82800 |
|
}, |
|
{ |
|
"epoch": 6.79, |
|
"learning_rate": 6.425413459963976e-06, |
|
"loss": 2.5578, |
|
"step": 82900 |
|
}, |
|
{ |
|
"epoch": 6.8, |
|
"learning_rate": 6.409038807925332e-06, |
|
"loss": 2.5557, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 6.8, |
|
"learning_rate": 6.392664155886688e-06, |
|
"loss": 2.5333, |
|
"step": 83100 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"learning_rate": 6.3762895038480434e-06, |
|
"loss": 2.579, |
|
"step": 83200 |
|
}, |
|
{ |
|
"epoch": 6.82, |
|
"learning_rate": 6.3599148518094e-06, |
|
"loss": 2.5891, |
|
"step": 83300 |
|
}, |
|
{ |
|
"epoch": 6.83, |
|
"learning_rate": 6.343540199770756e-06, |
|
"loss": 2.5803, |
|
"step": 83400 |
|
}, |
|
{ |
|
"epoch": 6.84, |
|
"learning_rate": 6.3271655477321116e-06, |
|
"loss": 2.5818, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 6.84, |
|
"learning_rate": 6.310790895693467e-06, |
|
"loss": 2.5763, |
|
"step": 83600 |
|
}, |
|
{ |
|
"epoch": 6.85, |
|
"learning_rate": 6.2944162436548225e-06, |
|
"loss": 2.5528, |
|
"step": 83700 |
|
}, |
|
{ |
|
"epoch": 6.86, |
|
"learning_rate": 6.278041591616179e-06, |
|
"loss": 2.559, |
|
"step": 83800 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"learning_rate": 6.261666939577534e-06, |
|
"loss": 2.5455, |
|
"step": 83900 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"learning_rate": 6.24529228753889e-06, |
|
"loss": 2.5672, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 6.89, |
|
"learning_rate": 6.228917635500245e-06, |
|
"loss": 2.5598, |
|
"step": 84100 |
|
}, |
|
{ |
|
"epoch": 6.89, |
|
"learning_rate": 6.2125429834616024e-06, |
|
"loss": 2.5807, |
|
"step": 84200 |
|
}, |
|
{ |
|
"epoch": 6.9, |
|
"learning_rate": 6.196168331422958e-06, |
|
"loss": 2.5671, |
|
"step": 84300 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 6.179793679384314e-06, |
|
"loss": 2.5396, |
|
"step": 84400 |
|
}, |
|
{ |
|
"epoch": 6.92, |
|
"learning_rate": 6.16341902734567e-06, |
|
"loss": 2.5315, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 6.93, |
|
"learning_rate": 6.147044375307025e-06, |
|
"loss": 2.5686, |
|
"step": 84600 |
|
}, |
|
{ |
|
"epoch": 6.93, |
|
"learning_rate": 6.130669723268381e-06, |
|
"loss": 2.5752, |
|
"step": 84700 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 6.114295071229737e-06, |
|
"loss": 2.566, |
|
"step": 84800 |
|
}, |
|
{ |
|
"epoch": 6.95, |
|
"learning_rate": 6.0979204191910925e-06, |
|
"loss": 2.5857, |
|
"step": 84900 |
|
}, |
|
{ |
|
"epoch": 6.96, |
|
"learning_rate": 6.081545767152448e-06, |
|
"loss": 2.5951, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 6.97, |
|
"learning_rate": 6.065171115113805e-06, |
|
"loss": 2.5899, |
|
"step": 85100 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"learning_rate": 6.048796463075161e-06, |
|
"loss": 2.5783, |
|
"step": 85200 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"learning_rate": 6.032421811036516e-06, |
|
"loss": 2.5497, |
|
"step": 85300 |
|
}, |
|
{ |
|
"epoch": 6.99, |
|
"learning_rate": 6.0160471589978716e-06, |
|
"loss": 2.5749, |
|
"step": 85400 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_bleu": 968.8997, |
|
"eval_chrf": 2861.7868, |
|
"eval_chrf_plus": 2687.5064, |
|
"eval_gen_len": 14.652, |
|
"eval_loss": 2.4077491760253906, |
|
"eval_runtime": 38.1903, |
|
"eval_samples_per_second": 26.185, |
|
"eval_steps_per_second": 1.1, |
|
"step": 85498 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 5.999672506959228e-06, |
|
"loss": 2.5836, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 5.983297854920583e-06, |
|
"loss": 2.5616, |
|
"step": 85600 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 5.966923202881939e-06, |
|
"loss": 2.5122, |
|
"step": 85700 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 5.950548550843295e-06, |
|
"loss": 2.5264, |
|
"step": 85800 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 5.934173898804651e-06, |
|
"loss": 2.5504, |
|
"step": 85900 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 5.917799246766007e-06, |
|
"loss": 2.552, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"learning_rate": 5.901424594727363e-06, |
|
"loss": 2.5237, |
|
"step": 86100 |
|
}, |
|
{ |
|
"epoch": 7.06, |
|
"learning_rate": 5.885049942688719e-06, |
|
"loss": 2.5496, |
|
"step": 86200 |
|
}, |
|
{ |
|
"epoch": 7.07, |
|
"learning_rate": 5.868675290650074e-06, |
|
"loss": 2.5281, |
|
"step": 86300 |
|
}, |
|
{ |
|
"epoch": 7.07, |
|
"learning_rate": 5.85230063861143e-06, |
|
"loss": 2.5384, |
|
"step": 86400 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"learning_rate": 5.835925986572786e-06, |
|
"loss": 2.557, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"learning_rate": 5.8195513345341415e-06, |
|
"loss": 2.5591, |
|
"step": 86600 |
|
}, |
|
{ |
|
"epoch": 7.1, |
|
"learning_rate": 5.803176682495497e-06, |
|
"loss": 2.5346, |
|
"step": 86700 |
|
}, |
|
{ |
|
"epoch": 7.11, |
|
"learning_rate": 5.7868020304568525e-06, |
|
"loss": 2.539, |
|
"step": 86800 |
|
}, |
|
{ |
|
"epoch": 7.11, |
|
"learning_rate": 5.770427378418209e-06, |
|
"loss": 2.5111, |
|
"step": 86900 |
|
}, |
|
{ |
|
"epoch": 7.12, |
|
"learning_rate": 5.754052726379565e-06, |
|
"loss": 2.5263, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 7.13, |
|
"learning_rate": 5.7376780743409215e-06, |
|
"loss": 2.5451, |
|
"step": 87100 |
|
}, |
|
{ |
|
"epoch": 7.14, |
|
"learning_rate": 5.721303422302277e-06, |
|
"loss": 2.5028, |
|
"step": 87200 |
|
}, |
|
{ |
|
"epoch": 7.15, |
|
"learning_rate": 5.704928770263632e-06, |
|
"loss": 2.5378, |
|
"step": 87300 |
|
}, |
|
{ |
|
"epoch": 7.16, |
|
"learning_rate": 5.688554118224988e-06, |
|
"loss": 2.5263, |
|
"step": 87400 |
|
}, |
|
{ |
|
"epoch": 7.16, |
|
"learning_rate": 5.672179466186344e-06, |
|
"loss": 2.5375, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 7.17, |
|
"learning_rate": 5.6558048141477e-06, |
|
"loss": 2.5338, |
|
"step": 87600 |
|
}, |
|
{ |
|
"epoch": 7.18, |
|
"learning_rate": 5.639430162109055e-06, |
|
"loss": 2.5378, |
|
"step": 87700 |
|
}, |
|
{ |
|
"epoch": 7.19, |
|
"learning_rate": 5.623055510070411e-06, |
|
"loss": 2.5173, |
|
"step": 87800 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"learning_rate": 5.606680858031768e-06, |
|
"loss": 2.5416, |
|
"step": 87900 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"learning_rate": 5.590306205993123e-06, |
|
"loss": 2.5655, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 7.21, |
|
"learning_rate": 5.573931553954479e-06, |
|
"loss": 2.5181, |
|
"step": 88100 |
|
}, |
|
{ |
|
"epoch": 7.22, |
|
"learning_rate": 5.557556901915835e-06, |
|
"loss": 2.5275, |
|
"step": 88200 |
|
}, |
|
{ |
|
"epoch": 7.23, |
|
"learning_rate": 5.5411822498771906e-06, |
|
"loss": 2.5358, |
|
"step": 88300 |
|
}, |
|
{ |
|
"epoch": 7.24, |
|
"learning_rate": 5.524807597838546e-06, |
|
"loss": 2.5359, |
|
"step": 88400 |
|
}, |
|
{ |
|
"epoch": 7.25, |
|
"learning_rate": 5.508432945799902e-06, |
|
"loss": 2.5211, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 7.25, |
|
"learning_rate": 5.492058293761258e-06, |
|
"loss": 2.5447, |
|
"step": 88600 |
|
}, |
|
{ |
|
"epoch": 7.26, |
|
"learning_rate": 5.475683641722613e-06, |
|
"loss": 2.5232, |
|
"step": 88700 |
|
}, |
|
{ |
|
"epoch": 7.27, |
|
"learning_rate": 5.4593089896839705e-06, |
|
"loss": 2.5378, |
|
"step": 88800 |
|
}, |
|
{ |
|
"epoch": 7.28, |
|
"learning_rate": 5.442934337645326e-06, |
|
"loss": 2.5483, |
|
"step": 88900 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 5.4265596856066815e-06, |
|
"loss": 2.5405, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 5.410185033568037e-06, |
|
"loss": 2.509, |
|
"step": 89100 |
|
}, |
|
{ |
|
"epoch": 7.3, |
|
"learning_rate": 5.393810381529393e-06, |
|
"loss": 2.5231, |
|
"step": 89200 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"learning_rate": 5.377435729490749e-06, |
|
"loss": 2.5538, |
|
"step": 89300 |
|
}, |
|
{ |
|
"epoch": 7.32, |
|
"learning_rate": 5.361061077452104e-06, |
|
"loss": 2.5407, |
|
"step": 89400 |
|
}, |
|
{ |
|
"epoch": 7.33, |
|
"learning_rate": 5.34468642541346e-06, |
|
"loss": 2.5306, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 7.34, |
|
"learning_rate": 5.328311773374816e-06, |
|
"loss": 2.5821, |
|
"step": 89600 |
|
}, |
|
{ |
|
"epoch": 7.34, |
|
"learning_rate": 5.311937121336172e-06, |
|
"loss": 2.5328, |
|
"step": 89700 |
|
}, |
|
{ |
|
"epoch": 7.35, |
|
"learning_rate": 5.295562469297529e-06, |
|
"loss": 2.5458, |
|
"step": 89800 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"learning_rate": 5.279187817258884e-06, |
|
"loss": 2.5367, |
|
"step": 89900 |
|
}, |
|
{ |
|
"epoch": 7.37, |
|
"learning_rate": 5.26281316522024e-06, |
|
"loss": 2.5333, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 7.38, |
|
"learning_rate": 5.246438513181595e-06, |
|
"loss": 2.5492, |
|
"step": 90100 |
|
}, |
|
{ |
|
"epoch": 7.38, |
|
"learning_rate": 5.230063861142951e-06, |
|
"loss": 2.5078, |
|
"step": 90200 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 5.213689209104307e-06, |
|
"loss": 2.513, |
|
"step": 90300 |
|
}, |
|
{ |
|
"epoch": 7.4, |
|
"learning_rate": 5.197314557065662e-06, |
|
"loss": 2.5322, |
|
"step": 90400 |
|
}, |
|
{ |
|
"epoch": 7.41, |
|
"learning_rate": 5.180939905027018e-06, |
|
"loss": 2.5512, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 7.42, |
|
"learning_rate": 5.164565252988375e-06, |
|
"loss": 2.5409, |
|
"step": 90600 |
|
}, |
|
{ |
|
"epoch": 7.43, |
|
"learning_rate": 5.1481906009497305e-06, |
|
"loss": 2.5273, |
|
"step": 90700 |
|
}, |
|
{ |
|
"epoch": 7.43, |
|
"learning_rate": 5.131815948911086e-06, |
|
"loss": 2.5315, |
|
"step": 90800 |
|
}, |
|
{ |
|
"epoch": 7.44, |
|
"learning_rate": 5.115441296872442e-06, |
|
"loss": 2.5052, |
|
"step": 90900 |
|
}, |
|
{ |
|
"epoch": 7.45, |
|
"learning_rate": 5.099066644833798e-06, |
|
"loss": 2.5673, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 7.46, |
|
"learning_rate": 5.082691992795153e-06, |
|
"loss": 2.5222, |
|
"step": 91100 |
|
}, |
|
{ |
|
"epoch": 7.47, |
|
"learning_rate": 5.06631734075651e-06, |
|
"loss": 2.5104, |
|
"step": 91200 |
|
}, |
|
{ |
|
"epoch": 7.48, |
|
"learning_rate": 5.049942688717865e-06, |
|
"loss": 2.5237, |
|
"step": 91300 |
|
}, |
|
{ |
|
"epoch": 7.48, |
|
"learning_rate": 5.0335680366792205e-06, |
|
"loss": 2.5361, |
|
"step": 91400 |
|
}, |
|
{ |
|
"epoch": 7.49, |
|
"learning_rate": 5.017193384640576e-06, |
|
"loss": 2.517, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 5.000818732601933e-06, |
|
"loss": 2.5485, |
|
"step": 91600 |
|
}, |
|
{ |
|
"epoch": 7.51, |
|
"learning_rate": 4.984444080563288e-06, |
|
"loss": 2.5295, |
|
"step": 91700 |
|
}, |
|
{ |
|
"epoch": 7.52, |
|
"learning_rate": 4.968069428524644e-06, |
|
"loss": 2.5309, |
|
"step": 91800 |
|
}, |
|
{ |
|
"epoch": 7.52, |
|
"learning_rate": 4.9516947764860005e-06, |
|
"loss": 2.5199, |
|
"step": 91900 |
|
}, |
|
{ |
|
"epoch": 7.53, |
|
"learning_rate": 4.935320124447356e-06, |
|
"loss": 2.5216, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 7.54, |
|
"learning_rate": 4.918945472408711e-06, |
|
"loss": 2.5298, |
|
"step": 92100 |
|
}, |
|
{ |
|
"epoch": 7.55, |
|
"learning_rate": 4.902570820370067e-06, |
|
"loss": 2.5084, |
|
"step": 92200 |
|
}, |
|
{ |
|
"epoch": 7.56, |
|
"learning_rate": 4.886196168331423e-06, |
|
"loss": 2.4954, |
|
"step": 92300 |
|
}, |
|
{ |
|
"epoch": 7.57, |
|
"learning_rate": 4.8698215162927795e-06, |
|
"loss": 2.5258, |
|
"step": 92400 |
|
}, |
|
{ |
|
"epoch": 7.57, |
|
"learning_rate": 4.853446864254135e-06, |
|
"loss": 2.5319, |
|
"step": 92500 |
|
}, |
|
{ |
|
"epoch": 7.58, |
|
"learning_rate": 4.8370722122154905e-06, |
|
"loss": 2.5105, |
|
"step": 92600 |
|
}, |
|
{ |
|
"epoch": 7.59, |
|
"learning_rate": 4.820697560176847e-06, |
|
"loss": 2.5279, |
|
"step": 92700 |
|
}, |
|
{ |
|
"epoch": 7.6, |
|
"learning_rate": 4.804322908138202e-06, |
|
"loss": 2.5328, |
|
"step": 92800 |
|
}, |
|
{ |
|
"epoch": 7.61, |
|
"learning_rate": 4.787948256099559e-06, |
|
"loss": 2.5287, |
|
"step": 92900 |
|
}, |
|
{ |
|
"epoch": 7.61, |
|
"learning_rate": 4.771573604060914e-06, |
|
"loss": 2.5093, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 7.62, |
|
"learning_rate": 4.75519895202227e-06, |
|
"loss": 2.5184, |
|
"step": 93100 |
|
}, |
|
{ |
|
"epoch": 7.63, |
|
"learning_rate": 4.738824299983626e-06, |
|
"loss": 2.5407, |
|
"step": 93200 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 4.722449647944981e-06, |
|
"loss": 2.5333, |
|
"step": 93300 |
|
}, |
|
{ |
|
"epoch": 7.65, |
|
"learning_rate": 4.706074995906338e-06, |
|
"loss": 2.4874, |
|
"step": 93400 |
|
}, |
|
{ |
|
"epoch": 7.66, |
|
"learning_rate": 4.689700343867693e-06, |
|
"loss": 2.5334, |
|
"step": 93500 |
|
}, |
|
{ |
|
"epoch": 7.66, |
|
"learning_rate": 4.6733256918290495e-06, |
|
"loss": 2.5413, |
|
"step": 93600 |
|
}, |
|
{ |
|
"epoch": 7.67, |
|
"learning_rate": 4.656951039790405e-06, |
|
"loss": 2.5073, |
|
"step": 93700 |
|
}, |
|
{ |
|
"epoch": 7.68, |
|
"learning_rate": 4.6405763877517605e-06, |
|
"loss": 2.5276, |
|
"step": 93800 |
|
}, |
|
{ |
|
"epoch": 7.69, |
|
"learning_rate": 4.624201735713117e-06, |
|
"loss": 2.5264, |
|
"step": 93900 |
|
}, |
|
{ |
|
"epoch": 7.7, |
|
"learning_rate": 4.607827083674472e-06, |
|
"loss": 2.5465, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 7.7, |
|
"learning_rate": 4.591452431635829e-06, |
|
"loss": 2.5404, |
|
"step": 94100 |
|
}, |
|
{ |
|
"epoch": 7.71, |
|
"learning_rate": 4.575077779597184e-06, |
|
"loss": 2.5142, |
|
"step": 94200 |
|
}, |
|
{ |
|
"epoch": 7.72, |
|
"learning_rate": 4.5587031275585395e-06, |
|
"loss": 2.5218, |
|
"step": 94300 |
|
}, |
|
{ |
|
"epoch": 7.73, |
|
"learning_rate": 4.542328475519895e-06, |
|
"loss": 2.5696, |
|
"step": 94400 |
|
}, |
|
{ |
|
"epoch": 7.74, |
|
"learning_rate": 4.525953823481251e-06, |
|
"loss": 2.5272, |
|
"step": 94500 |
|
}, |
|
{ |
|
"epoch": 7.75, |
|
"learning_rate": 4.509579171442608e-06, |
|
"loss": 2.5121, |
|
"step": 94600 |
|
}, |
|
{ |
|
"epoch": 7.75, |
|
"learning_rate": 4.493204519403963e-06, |
|
"loss": 2.5392, |
|
"step": 94700 |
|
}, |
|
{ |
|
"epoch": 7.76, |
|
"learning_rate": 4.476829867365319e-06, |
|
"loss": 2.5215, |
|
"step": 94800 |
|
}, |
|
{ |
|
"epoch": 7.77, |
|
"learning_rate": 4.460455215326674e-06, |
|
"loss": 2.5177, |
|
"step": 94900 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 4.4440805632880304e-06, |
|
"loss": 2.4851, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 7.79, |
|
"learning_rate": 4.427705911249387e-06, |
|
"loss": 2.5696, |
|
"step": 95100 |
|
}, |
|
{ |
|
"epoch": 7.79, |
|
"learning_rate": 4.411331259210742e-06, |
|
"loss": 2.4741, |
|
"step": 95200 |
|
}, |
|
{ |
|
"epoch": 7.8, |
|
"learning_rate": 4.394956607172098e-06, |
|
"loss": 2.5363, |
|
"step": 95300 |
|
}, |
|
{ |
|
"epoch": 7.81, |
|
"learning_rate": 4.378581955133453e-06, |
|
"loss": 2.5258, |
|
"step": 95400 |
|
}, |
|
{ |
|
"epoch": 7.82, |
|
"learning_rate": 4.3622073030948095e-06, |
|
"loss": 2.5155, |
|
"step": 95500 |
|
}, |
|
{ |
|
"epoch": 7.83, |
|
"learning_rate": 4.345832651056166e-06, |
|
"loss": 2.5141, |
|
"step": 95600 |
|
}, |
|
{ |
|
"epoch": 7.84, |
|
"learning_rate": 4.329457999017521e-06, |
|
"loss": 2.4956, |
|
"step": 95700 |
|
}, |
|
{ |
|
"epoch": 7.84, |
|
"learning_rate": 4.313083346978877e-06, |
|
"loss": 2.5107, |
|
"step": 95800 |
|
}, |
|
{ |
|
"epoch": 7.85, |
|
"learning_rate": 4.296708694940233e-06, |
|
"loss": 2.5159, |
|
"step": 95900 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"learning_rate": 4.280334042901589e-06, |
|
"loss": 2.5159, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 7.87, |
|
"learning_rate": 4.263959390862945e-06, |
|
"loss": 2.5321, |
|
"step": 96100 |
|
}, |
|
{ |
|
"epoch": 7.88, |
|
"learning_rate": 4.2475847388243e-06, |
|
"loss": 2.5295, |
|
"step": 96200 |
|
}, |
|
{ |
|
"epoch": 7.88, |
|
"learning_rate": 4.231210086785656e-06, |
|
"loss": 2.5022, |
|
"step": 96300 |
|
}, |
|
{ |
|
"epoch": 7.89, |
|
"learning_rate": 4.214835434747012e-06, |
|
"loss": 2.5259, |
|
"step": 96400 |
|
}, |
|
{ |
|
"epoch": 7.9, |
|
"learning_rate": 4.198460782708368e-06, |
|
"loss": 2.4996, |
|
"step": 96500 |
|
}, |
|
{ |
|
"epoch": 7.91, |
|
"learning_rate": 4.182086130669724e-06, |
|
"loss": 2.5113, |
|
"step": 96600 |
|
}, |
|
{ |
|
"epoch": 7.92, |
|
"learning_rate": 4.1657114786310795e-06, |
|
"loss": 2.5115, |
|
"step": 96700 |
|
}, |
|
{ |
|
"epoch": 7.93, |
|
"learning_rate": 4.149336826592435e-06, |
|
"loss": 2.5159, |
|
"step": 96800 |
|
}, |
|
{ |
|
"epoch": 7.93, |
|
"learning_rate": 4.132962174553791e-06, |
|
"loss": 2.5268, |
|
"step": 96900 |
|
}, |
|
{ |
|
"epoch": 7.94, |
|
"learning_rate": 4.116587522515147e-06, |
|
"loss": 2.5041, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 7.95, |
|
"learning_rate": 4.100212870476502e-06, |
|
"loss": 2.5116, |
|
"step": 97100 |
|
}, |
|
{ |
|
"epoch": 7.96, |
|
"learning_rate": 4.0838382184378586e-06, |
|
"loss": 2.5062, |
|
"step": 97200 |
|
}, |
|
{ |
|
"epoch": 7.97, |
|
"learning_rate": 4.067463566399215e-06, |
|
"loss": 2.54, |
|
"step": 97300 |
|
}, |
|
{ |
|
"epoch": 7.97, |
|
"learning_rate": 4.05108891436057e-06, |
|
"loss": 2.5216, |
|
"step": 97400 |
|
}, |
|
{ |
|
"epoch": 7.98, |
|
"learning_rate": 4.034714262321926e-06, |
|
"loss": 2.5035, |
|
"step": 97500 |
|
}, |
|
{ |
|
"epoch": 7.99, |
|
"learning_rate": 4.018339610283281e-06, |
|
"loss": 2.5216, |
|
"step": 97600 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 4.001964958244638e-06, |
|
"loss": 2.5407, |
|
"step": 97700 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_bleu": 980.1715, |
|
"eval_chrf": 2869.0816, |
|
"eval_chrf_plus": 2698.0447, |
|
"eval_gen_len": 14.656, |
|
"eval_loss": 2.388012170791626, |
|
"eval_runtime": 38.854, |
|
"eval_samples_per_second": 25.737, |
|
"eval_steps_per_second": 1.081, |
|
"step": 97712 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 3.985590306205994e-06, |
|
"loss": 2.4997, |
|
"step": 97800 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 3.9692156541673494e-06, |
|
"loss": 2.5192, |
|
"step": 97900 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 3.952841002128705e-06, |
|
"loss": 2.477, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 3.93646635009006e-06, |
|
"loss": 2.5026, |
|
"step": 98100 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 3.920091698051417e-06, |
|
"loss": 2.5002, |
|
"step": 98200 |
|
}, |
|
{ |
|
"epoch": 8.05, |
|
"learning_rate": 3.903717046012773e-06, |
|
"loss": 2.4938, |
|
"step": 98300 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 3.8873423939741285e-06, |
|
"loss": 2.488, |
|
"step": 98400 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 3.870967741935484e-06, |
|
"loss": 2.5018, |
|
"step": 98500 |
|
}, |
|
{ |
|
"epoch": 8.07, |
|
"learning_rate": 3.8545930898968395e-06, |
|
"loss": 2.5152, |
|
"step": 98600 |
|
}, |
|
{ |
|
"epoch": 8.08, |
|
"learning_rate": 3.838218437858196e-06, |
|
"loss": 2.483, |
|
"step": 98700 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"learning_rate": 3.821843785819552e-06, |
|
"loss": 2.495, |
|
"step": 98800 |
|
}, |
|
{ |
|
"epoch": 8.1, |
|
"learning_rate": 3.8054691337809076e-06, |
|
"loss": 2.5065, |
|
"step": 98900 |
|
}, |
|
{ |
|
"epoch": 8.11, |
|
"learning_rate": 3.789094481742263e-06, |
|
"loss": 2.5051, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 8.11, |
|
"learning_rate": 3.772719829703619e-06, |
|
"loss": 2.5001, |
|
"step": 99100 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"learning_rate": 3.756345177664975e-06, |
|
"loss": 2.5074, |
|
"step": 99200 |
|
}, |
|
{ |
|
"epoch": 8.13, |
|
"learning_rate": 3.7399705256263308e-06, |
|
"loss": 2.4961, |
|
"step": 99300 |
|
}, |
|
{ |
|
"epoch": 8.14, |
|
"learning_rate": 3.7235958735876867e-06, |
|
"loss": 2.5006, |
|
"step": 99400 |
|
}, |
|
{ |
|
"epoch": 8.15, |
|
"learning_rate": 3.707221221549042e-06, |
|
"loss": 2.4864, |
|
"step": 99500 |
|
}, |
|
{ |
|
"epoch": 8.15, |
|
"learning_rate": 3.6908465695103985e-06, |
|
"loss": 2.4795, |
|
"step": 99600 |
|
}, |
|
{ |
|
"epoch": 8.16, |
|
"learning_rate": 3.674471917471754e-06, |
|
"loss": 2.4886, |
|
"step": 99700 |
|
}, |
|
{ |
|
"epoch": 8.17, |
|
"learning_rate": 3.65809726543311e-06, |
|
"loss": 2.4853, |
|
"step": 99800 |
|
}, |
|
{ |
|
"epoch": 8.18, |
|
"learning_rate": 3.6417226133944653e-06, |
|
"loss": 2.5001, |
|
"step": 99900 |
|
}, |
|
{ |
|
"epoch": 8.19, |
|
"learning_rate": 3.6253479613558212e-06, |
|
"loss": 2.5031, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"learning_rate": 3.6089733093171776e-06, |
|
"loss": 2.4934, |
|
"step": 100100 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"learning_rate": 3.592598657278533e-06, |
|
"loss": 2.5008, |
|
"step": 100200 |
|
}, |
|
{ |
|
"epoch": 8.21, |
|
"learning_rate": 3.576224005239889e-06, |
|
"loss": 2.4825, |
|
"step": 100300 |
|
}, |
|
{ |
|
"epoch": 8.22, |
|
"learning_rate": 3.5598493532012444e-06, |
|
"loss": 2.4996, |
|
"step": 100400 |
|
}, |
|
{ |
|
"epoch": 8.23, |
|
"learning_rate": 3.5434747011626007e-06, |
|
"loss": 2.5029, |
|
"step": 100500 |
|
}, |
|
{ |
|
"epoch": 8.24, |
|
"learning_rate": 3.5271000491239566e-06, |
|
"loss": 2.4722, |
|
"step": 100600 |
|
}, |
|
{ |
|
"epoch": 8.24, |
|
"learning_rate": 3.510725397085312e-06, |
|
"loss": 2.4791, |
|
"step": 100700 |
|
}, |
|
{ |
|
"epoch": 8.25, |
|
"learning_rate": 3.494350745046668e-06, |
|
"loss": 2.4741, |
|
"step": 100800 |
|
}, |
|
{ |
|
"epoch": 8.26, |
|
"learning_rate": 3.4779760930080235e-06, |
|
"loss": 2.5113, |
|
"step": 100900 |
|
}, |
|
{ |
|
"epoch": 8.27, |
|
"learning_rate": 3.46160144096938e-06, |
|
"loss": 2.4916, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 8.28, |
|
"learning_rate": 3.4452267889307357e-06, |
|
"loss": 2.4865, |
|
"step": 101100 |
|
}, |
|
{ |
|
"epoch": 8.29, |
|
"learning_rate": 3.428852136892091e-06, |
|
"loss": 2.4891, |
|
"step": 101200 |
|
}, |
|
{ |
|
"epoch": 8.29, |
|
"learning_rate": 3.412477484853447e-06, |
|
"loss": 2.484, |
|
"step": 101300 |
|
}, |
|
{ |
|
"epoch": 8.3, |
|
"learning_rate": 3.3961028328148026e-06, |
|
"loss": 2.4827, |
|
"step": 101400 |
|
}, |
|
{ |
|
"epoch": 8.31, |
|
"learning_rate": 3.379728180776159e-06, |
|
"loss": 2.5062, |
|
"step": 101500 |
|
}, |
|
{ |
|
"epoch": 8.32, |
|
"learning_rate": 3.363353528737515e-06, |
|
"loss": 2.5128, |
|
"step": 101600 |
|
}, |
|
{ |
|
"epoch": 8.33, |
|
"learning_rate": 3.3469788766988703e-06, |
|
"loss": 2.508, |
|
"step": 101700 |
|
}, |
|
{ |
|
"epoch": 8.33, |
|
"learning_rate": 3.330604224660226e-06, |
|
"loss": 2.4933, |
|
"step": 101800 |
|
}, |
|
{ |
|
"epoch": 8.34, |
|
"learning_rate": 3.314229572621582e-06, |
|
"loss": 2.5001, |
|
"step": 101900 |
|
}, |
|
{ |
|
"epoch": 8.35, |
|
"learning_rate": 3.297854920582938e-06, |
|
"loss": 2.4936, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 8.36, |
|
"learning_rate": 3.281480268544294e-06, |
|
"loss": 2.4793, |
|
"step": 102100 |
|
}, |
|
{ |
|
"epoch": 8.37, |
|
"learning_rate": 3.2651056165056494e-06, |
|
"loss": 2.5187, |
|
"step": 102200 |
|
}, |
|
{ |
|
"epoch": 8.38, |
|
"learning_rate": 3.2487309644670053e-06, |
|
"loss": 2.5113, |
|
"step": 102300 |
|
}, |
|
{ |
|
"epoch": 8.38, |
|
"learning_rate": 3.232356312428361e-06, |
|
"loss": 2.5055, |
|
"step": 102400 |
|
}, |
|
{ |
|
"epoch": 8.39, |
|
"learning_rate": 3.215981660389717e-06, |
|
"loss": 2.5111, |
|
"step": 102500 |
|
}, |
|
{ |
|
"epoch": 8.4, |
|
"learning_rate": 3.1996070083510725e-06, |
|
"loss": 2.5075, |
|
"step": 102600 |
|
}, |
|
{ |
|
"epoch": 8.41, |
|
"learning_rate": 3.1832323563124284e-06, |
|
"loss": 2.4961, |
|
"step": 102700 |
|
}, |
|
{ |
|
"epoch": 8.42, |
|
"learning_rate": 3.1668577042737848e-06, |
|
"loss": 2.5245, |
|
"step": 102800 |
|
}, |
|
{ |
|
"epoch": 8.42, |
|
"learning_rate": 3.1504830522351402e-06, |
|
"loss": 2.4936, |
|
"step": 102900 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 3.134108400196496e-06, |
|
"loss": 2.4891, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 8.44, |
|
"learning_rate": 3.1177337481578516e-06, |
|
"loss": 2.5013, |
|
"step": 103100 |
|
}, |
|
{ |
|
"epoch": 8.45, |
|
"learning_rate": 3.1013590961192075e-06, |
|
"loss": 2.4884, |
|
"step": 103200 |
|
}, |
|
{ |
|
"epoch": 8.46, |
|
"learning_rate": 3.084984444080564e-06, |
|
"loss": 2.4928, |
|
"step": 103300 |
|
}, |
|
{ |
|
"epoch": 8.47, |
|
"learning_rate": 3.0686097920419193e-06, |
|
"loss": 2.4995, |
|
"step": 103400 |
|
}, |
|
{ |
|
"epoch": 8.47, |
|
"learning_rate": 3.0522351400032752e-06, |
|
"loss": 2.5228, |
|
"step": 103500 |
|
}, |
|
{ |
|
"epoch": 8.48, |
|
"learning_rate": 3.0358604879646307e-06, |
|
"loss": 2.5199, |
|
"step": 103600 |
|
}, |
|
{ |
|
"epoch": 8.49, |
|
"learning_rate": 3.019485835925987e-06, |
|
"loss": 2.4869, |
|
"step": 103700 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 3.003111183887343e-06, |
|
"loss": 2.497, |
|
"step": 103800 |
|
}, |
|
{ |
|
"epoch": 8.51, |
|
"learning_rate": 2.9867365318486984e-06, |
|
"loss": 2.4866, |
|
"step": 103900 |
|
}, |
|
{ |
|
"epoch": 8.51, |
|
"learning_rate": 2.9703618798100543e-06, |
|
"loss": 2.4958, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 8.52, |
|
"learning_rate": 2.95398722777141e-06, |
|
"loss": 2.494, |
|
"step": 104100 |
|
}, |
|
{ |
|
"epoch": 8.53, |
|
"learning_rate": 2.937612575732766e-06, |
|
"loss": 2.4919, |
|
"step": 104200 |
|
}, |
|
{ |
|
"epoch": 8.54, |
|
"learning_rate": 2.921237923694122e-06, |
|
"loss": 2.5, |
|
"step": 104300 |
|
}, |
|
{ |
|
"epoch": 8.55, |
|
"learning_rate": 2.9048632716554775e-06, |
|
"loss": 2.5147, |
|
"step": 104400 |
|
}, |
|
{ |
|
"epoch": 8.56, |
|
"learning_rate": 2.8884886196168334e-06, |
|
"loss": 2.4622, |
|
"step": 104500 |
|
}, |
|
{ |
|
"epoch": 8.56, |
|
"learning_rate": 2.872113967578189e-06, |
|
"loss": 2.4919, |
|
"step": 104600 |
|
}, |
|
{ |
|
"epoch": 8.57, |
|
"learning_rate": 2.855739315539545e-06, |
|
"loss": 2.4882, |
|
"step": 104700 |
|
}, |
|
{ |
|
"epoch": 8.58, |
|
"learning_rate": 2.839364663500901e-06, |
|
"loss": 2.5035, |
|
"step": 104800 |
|
}, |
|
{ |
|
"epoch": 8.59, |
|
"learning_rate": 2.8229900114622566e-06, |
|
"loss": 2.49, |
|
"step": 104900 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"learning_rate": 2.8066153594236125e-06, |
|
"loss": 2.5085, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"learning_rate": 2.7902407073849684e-06, |
|
"loss": 2.522, |
|
"step": 105100 |
|
}, |
|
{ |
|
"epoch": 8.61, |
|
"learning_rate": 2.7738660553463243e-06, |
|
"loss": 2.5118, |
|
"step": 105200 |
|
}, |
|
{ |
|
"epoch": 8.62, |
|
"learning_rate": 2.7574914033076798e-06, |
|
"loss": 2.4968, |
|
"step": 105300 |
|
}, |
|
{ |
|
"epoch": 8.63, |
|
"learning_rate": 2.7411167512690357e-06, |
|
"loss": 2.5011, |
|
"step": 105400 |
|
}, |
|
{ |
|
"epoch": 8.64, |
|
"learning_rate": 2.7247420992303916e-06, |
|
"loss": 2.4636, |
|
"step": 105500 |
|
}, |
|
{ |
|
"epoch": 8.65, |
|
"learning_rate": 2.7083674471917475e-06, |
|
"loss": 2.5283, |
|
"step": 105600 |
|
}, |
|
{ |
|
"epoch": 8.65, |
|
"learning_rate": 2.6919927951531034e-06, |
|
"loss": 2.5078, |
|
"step": 105700 |
|
}, |
|
{ |
|
"epoch": 8.66, |
|
"learning_rate": 2.675618143114459e-06, |
|
"loss": 2.4719, |
|
"step": 105800 |
|
}, |
|
{ |
|
"epoch": 8.67, |
|
"learning_rate": 2.6592434910758147e-06, |
|
"loss": 2.4869, |
|
"step": 105900 |
|
}, |
|
{ |
|
"epoch": 8.68, |
|
"learning_rate": 2.642868839037171e-06, |
|
"loss": 2.5042, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 2.6264941869985265e-06, |
|
"loss": 2.4742, |
|
"step": 106100 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 2.6101195349598824e-06, |
|
"loss": 2.469, |
|
"step": 106200 |
|
}, |
|
{ |
|
"epoch": 8.7, |
|
"learning_rate": 2.593744882921238e-06, |
|
"loss": 2.4729, |
|
"step": 106300 |
|
}, |
|
{ |
|
"epoch": 8.71, |
|
"learning_rate": 2.577370230882594e-06, |
|
"loss": 2.4797, |
|
"step": 106400 |
|
}, |
|
{ |
|
"epoch": 8.72, |
|
"learning_rate": 2.56099557884395e-06, |
|
"loss": 2.5034, |
|
"step": 106500 |
|
}, |
|
{ |
|
"epoch": 8.73, |
|
"learning_rate": 2.5446209268053056e-06, |
|
"loss": 2.4873, |
|
"step": 106600 |
|
}, |
|
{ |
|
"epoch": 8.74, |
|
"learning_rate": 2.5282462747666615e-06, |
|
"loss": 2.4812, |
|
"step": 106700 |
|
}, |
|
{ |
|
"epoch": 8.74, |
|
"learning_rate": 2.511871622728017e-06, |
|
"loss": 2.5116, |
|
"step": 106800 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 2.495496970689373e-06, |
|
"loss": 2.5066, |
|
"step": 106900 |
|
}, |
|
{ |
|
"epoch": 8.76, |
|
"learning_rate": 2.479122318650729e-06, |
|
"loss": 2.5003, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 8.77, |
|
"learning_rate": 2.4627476666120847e-06, |
|
"loss": 2.4658, |
|
"step": 107100 |
|
}, |
|
{ |
|
"epoch": 8.78, |
|
"learning_rate": 2.4463730145734406e-06, |
|
"loss": 2.483, |
|
"step": 107200 |
|
}, |
|
{ |
|
"epoch": 8.79, |
|
"learning_rate": 2.4299983625347965e-06, |
|
"loss": 2.4675, |
|
"step": 107300 |
|
}, |
|
{ |
|
"epoch": 8.79, |
|
"learning_rate": 2.413623710496152e-06, |
|
"loss": 2.4703, |
|
"step": 107400 |
|
}, |
|
{ |
|
"epoch": 8.8, |
|
"learning_rate": 2.3972490584575083e-06, |
|
"loss": 2.4838, |
|
"step": 107500 |
|
}, |
|
{ |
|
"epoch": 8.81, |
|
"learning_rate": 2.3808744064188638e-06, |
|
"loss": 2.4974, |
|
"step": 107600 |
|
}, |
|
{ |
|
"epoch": 8.82, |
|
"learning_rate": 2.3644997543802197e-06, |
|
"loss": 2.505, |
|
"step": 107700 |
|
}, |
|
{ |
|
"epoch": 8.83, |
|
"learning_rate": 2.3481251023415756e-06, |
|
"loss": 2.4754, |
|
"step": 107800 |
|
}, |
|
{ |
|
"epoch": 8.83, |
|
"learning_rate": 2.331750450302931e-06, |
|
"loss": 2.5193, |
|
"step": 107900 |
|
}, |
|
{ |
|
"epoch": 8.84, |
|
"learning_rate": 2.315375798264287e-06, |
|
"loss": 2.4847, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 8.85, |
|
"learning_rate": 2.299001146225643e-06, |
|
"loss": 2.4857, |
|
"step": 108100 |
|
}, |
|
{ |
|
"epoch": 8.86, |
|
"learning_rate": 2.2826264941869988e-06, |
|
"loss": 2.5072, |
|
"step": 108200 |
|
}, |
|
{ |
|
"epoch": 8.87, |
|
"learning_rate": 2.2662518421483547e-06, |
|
"loss": 2.4878, |
|
"step": 108300 |
|
}, |
|
{ |
|
"epoch": 8.88, |
|
"learning_rate": 2.24987719010971e-06, |
|
"loss": 2.4876, |
|
"step": 108400 |
|
}, |
|
{ |
|
"epoch": 8.88, |
|
"learning_rate": 2.233502538071066e-06, |
|
"loss": 2.4714, |
|
"step": 108500 |
|
}, |
|
{ |
|
"epoch": 8.89, |
|
"learning_rate": 2.217127886032422e-06, |
|
"loss": 2.5033, |
|
"step": 108600 |
|
}, |
|
{ |
|
"epoch": 8.9, |
|
"learning_rate": 2.200753233993778e-06, |
|
"loss": 2.5034, |
|
"step": 108700 |
|
}, |
|
{ |
|
"epoch": 8.91, |
|
"learning_rate": 2.1843785819551337e-06, |
|
"loss": 2.4987, |
|
"step": 108800 |
|
}, |
|
{ |
|
"epoch": 8.92, |
|
"learning_rate": 2.1680039299164896e-06, |
|
"loss": 2.5166, |
|
"step": 108900 |
|
}, |
|
{ |
|
"epoch": 8.92, |
|
"learning_rate": 2.151629277877845e-06, |
|
"loss": 2.5031, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 8.93, |
|
"learning_rate": 2.135254625839201e-06, |
|
"loss": 2.4761, |
|
"step": 109100 |
|
}, |
|
{ |
|
"epoch": 8.94, |
|
"learning_rate": 2.118879973800557e-06, |
|
"loss": 2.486, |
|
"step": 109200 |
|
}, |
|
{ |
|
"epoch": 8.95, |
|
"learning_rate": 2.102505321761913e-06, |
|
"loss": 2.5092, |
|
"step": 109300 |
|
}, |
|
{ |
|
"epoch": 8.96, |
|
"learning_rate": 2.0861306697232687e-06, |
|
"loss": 2.5162, |
|
"step": 109400 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"learning_rate": 2.069756017684624e-06, |
|
"loss": 2.4616, |
|
"step": 109500 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"learning_rate": 2.05338136564598e-06, |
|
"loss": 2.5059, |
|
"step": 109600 |
|
}, |
|
{ |
|
"epoch": 8.98, |
|
"learning_rate": 2.037006713607336e-06, |
|
"loss": 2.4782, |
|
"step": 109700 |
|
}, |
|
{ |
|
"epoch": 8.99, |
|
"learning_rate": 2.020632061568692e-06, |
|
"loss": 2.5064, |
|
"step": 109800 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 2.004257409530048e-06, |
|
"loss": 2.5296, |
|
"step": 109900 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_bleu": 983.7557, |
|
"eval_chrf": 2873.9646, |
|
"eval_chrf_plus": 2702.4064, |
|
"eval_gen_len": 14.62, |
|
"eval_loss": 2.3763132095336914, |
|
"eval_runtime": 38.7822, |
|
"eval_samples_per_second": 25.785, |
|
"eval_steps_per_second": 1.083, |
|
"step": 109926 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 1.9878827574914033e-06, |
|
"loss": 2.4896, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 1.971508105452759e-06, |
|
"loss": 2.4685, |
|
"step": 110100 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 1.955133453414115e-06, |
|
"loss": 2.4665, |
|
"step": 110200 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 1.938758801375471e-06, |
|
"loss": 2.4684, |
|
"step": 110300 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 1.922384149336827e-06, |
|
"loss": 2.4761, |
|
"step": 110400 |
|
}, |
|
{ |
|
"epoch": 9.05, |
|
"learning_rate": 1.9060094972981828e-06, |
|
"loss": 2.4977, |
|
"step": 110500 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"learning_rate": 1.8896348452595385e-06, |
|
"loss": 2.4788, |
|
"step": 110600 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"learning_rate": 1.8732601932208942e-06, |
|
"loss": 2.4939, |
|
"step": 110700 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"learning_rate": 1.85688554118225e-06, |
|
"loss": 2.4984, |
|
"step": 110800 |
|
}, |
|
{ |
|
"epoch": 9.08, |
|
"learning_rate": 1.8405108891436058e-06, |
|
"loss": 2.4748, |
|
"step": 110900 |
|
}, |
|
{ |
|
"epoch": 9.09, |
|
"learning_rate": 1.8241362371049617e-06, |
|
"loss": 2.4891, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"learning_rate": 1.8077615850663173e-06, |
|
"loss": 2.4635, |
|
"step": 111100 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"learning_rate": 1.7913869330276735e-06, |
|
"loss": 2.4936, |
|
"step": 111200 |
|
}, |
|
{ |
|
"epoch": 9.11, |
|
"learning_rate": 1.7750122809890291e-06, |
|
"loss": 2.4688, |
|
"step": 111300 |
|
}, |
|
{ |
|
"epoch": 9.12, |
|
"learning_rate": 1.7586376289503848e-06, |
|
"loss": 2.4682, |
|
"step": 111400 |
|
}, |
|
{ |
|
"epoch": 9.13, |
|
"learning_rate": 1.7422629769117407e-06, |
|
"loss": 2.4697, |
|
"step": 111500 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"learning_rate": 1.7258883248730964e-06, |
|
"loss": 2.4793, |
|
"step": 111600 |
|
}, |
|
{ |
|
"epoch": 9.15, |
|
"learning_rate": 1.7095136728344525e-06, |
|
"loss": 2.4898, |
|
"step": 111700 |
|
}, |
|
{ |
|
"epoch": 9.15, |
|
"learning_rate": 1.6931390207958082e-06, |
|
"loss": 2.4856, |
|
"step": 111800 |
|
}, |
|
{ |
|
"epoch": 9.16, |
|
"learning_rate": 1.6767643687571641e-06, |
|
"loss": 2.4347, |
|
"step": 111900 |
|
}, |
|
{ |
|
"epoch": 9.17, |
|
"learning_rate": 1.6603897167185198e-06, |
|
"loss": 2.4829, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 9.18, |
|
"learning_rate": 1.6440150646798757e-06, |
|
"loss": 2.4612, |
|
"step": 112100 |
|
}, |
|
{ |
|
"epoch": 9.19, |
|
"learning_rate": 1.6276404126412314e-06, |
|
"loss": 2.4394, |
|
"step": 112200 |
|
}, |
|
{ |
|
"epoch": 9.19, |
|
"learning_rate": 1.6112657606025873e-06, |
|
"loss": 2.5252, |
|
"step": 112300 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 1.5948911085639432e-06, |
|
"loss": 2.4966, |
|
"step": 112400 |
|
}, |
|
{ |
|
"epoch": 9.21, |
|
"learning_rate": 1.578516456525299e-06, |
|
"loss": 2.4586, |
|
"step": 112500 |
|
}, |
|
{ |
|
"epoch": 9.22, |
|
"learning_rate": 1.5621418044866548e-06, |
|
"loss": 2.463, |
|
"step": 112600 |
|
}, |
|
{ |
|
"epoch": 9.23, |
|
"learning_rate": 1.5457671524480105e-06, |
|
"loss": 2.4873, |
|
"step": 112700 |
|
}, |
|
{ |
|
"epoch": 9.24, |
|
"learning_rate": 1.5293925004093666e-06, |
|
"loss": 2.4735, |
|
"step": 112800 |
|
}, |
|
{ |
|
"epoch": 9.24, |
|
"learning_rate": 1.5130178483707223e-06, |
|
"loss": 2.4959, |
|
"step": 112900 |
|
}, |
|
{ |
|
"epoch": 9.25, |
|
"learning_rate": 1.496643196332078e-06, |
|
"loss": 2.4856, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 9.26, |
|
"learning_rate": 1.4802685442934339e-06, |
|
"loss": 2.4681, |
|
"step": 113100 |
|
}, |
|
{ |
|
"epoch": 9.27, |
|
"learning_rate": 1.4638938922547896e-06, |
|
"loss": 2.4575, |
|
"step": 113200 |
|
}, |
|
{ |
|
"epoch": 9.28, |
|
"learning_rate": 1.4475192402161457e-06, |
|
"loss": 2.4866, |
|
"step": 113300 |
|
}, |
|
{ |
|
"epoch": 9.28, |
|
"learning_rate": 1.4311445881775014e-06, |
|
"loss": 2.457, |
|
"step": 113400 |
|
}, |
|
{ |
|
"epoch": 9.29, |
|
"learning_rate": 1.4147699361388573e-06, |
|
"loss": 2.4792, |
|
"step": 113500 |
|
}, |
|
{ |
|
"epoch": 9.3, |
|
"learning_rate": 1.398395284100213e-06, |
|
"loss": 2.4699, |
|
"step": 113600 |
|
}, |
|
{ |
|
"epoch": 9.31, |
|
"learning_rate": 1.3820206320615687e-06, |
|
"loss": 2.4806, |
|
"step": 113700 |
|
}, |
|
{ |
|
"epoch": 9.32, |
|
"learning_rate": 1.3656459800229246e-06, |
|
"loss": 2.4692, |
|
"step": 113800 |
|
}, |
|
{ |
|
"epoch": 9.33, |
|
"learning_rate": 1.3492713279842802e-06, |
|
"loss": 2.4843, |
|
"step": 113900 |
|
}, |
|
{ |
|
"epoch": 9.33, |
|
"learning_rate": 1.3328966759456364e-06, |
|
"loss": 2.4625, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 9.34, |
|
"learning_rate": 1.316522023906992e-06, |
|
"loss": 2.4772, |
|
"step": 114100 |
|
}, |
|
{ |
|
"epoch": 9.35, |
|
"learning_rate": 1.300147371868348e-06, |
|
"loss": 2.4641, |
|
"step": 114200 |
|
}, |
|
{ |
|
"epoch": 9.36, |
|
"learning_rate": 1.2837727198297036e-06, |
|
"loss": 2.4972, |
|
"step": 114300 |
|
}, |
|
{ |
|
"epoch": 9.37, |
|
"learning_rate": 1.2673980677910597e-06, |
|
"loss": 2.4601, |
|
"step": 114400 |
|
}, |
|
{ |
|
"epoch": 9.37, |
|
"learning_rate": 1.2510234157524154e-06, |
|
"loss": 2.4744, |
|
"step": 114500 |
|
}, |
|
{ |
|
"epoch": 9.38, |
|
"learning_rate": 1.2346487637137711e-06, |
|
"loss": 2.4776, |
|
"step": 114600 |
|
}, |
|
{ |
|
"epoch": 9.39, |
|
"learning_rate": 1.218274111675127e-06, |
|
"loss": 2.4909, |
|
"step": 114700 |
|
}, |
|
{ |
|
"epoch": 9.4, |
|
"learning_rate": 1.201899459636483e-06, |
|
"loss": 2.446, |
|
"step": 114800 |
|
}, |
|
{ |
|
"epoch": 9.41, |
|
"learning_rate": 1.1855248075978386e-06, |
|
"loss": 2.4938, |
|
"step": 114900 |
|
}, |
|
{ |
|
"epoch": 9.42, |
|
"learning_rate": 1.1691501555591945e-06, |
|
"loss": 2.4416, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 9.42, |
|
"learning_rate": 1.1527755035205502e-06, |
|
"loss": 2.484, |
|
"step": 115100 |
|
}, |
|
{ |
|
"epoch": 9.43, |
|
"learning_rate": 1.1364008514819061e-06, |
|
"loss": 2.4538, |
|
"step": 115200 |
|
}, |
|
{ |
|
"epoch": 9.44, |
|
"learning_rate": 1.120026199443262e-06, |
|
"loss": 2.465, |
|
"step": 115300 |
|
}, |
|
{ |
|
"epoch": 9.45, |
|
"learning_rate": 1.1036515474046177e-06, |
|
"loss": 2.4822, |
|
"step": 115400 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"learning_rate": 1.0872768953659736e-06, |
|
"loss": 2.4673, |
|
"step": 115500 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"learning_rate": 1.0709022433273295e-06, |
|
"loss": 2.4847, |
|
"step": 115600 |
|
}, |
|
{ |
|
"epoch": 9.47, |
|
"learning_rate": 1.0545275912886852e-06, |
|
"loss": 2.4347, |
|
"step": 115700 |
|
}, |
|
{ |
|
"epoch": 9.48, |
|
"learning_rate": 1.038152939250041e-06, |
|
"loss": 2.4893, |
|
"step": 115800 |
|
}, |
|
{ |
|
"epoch": 9.49, |
|
"learning_rate": 1.0217782872113968e-06, |
|
"loss": 2.4832, |
|
"step": 115900 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"learning_rate": 1.0054036351727527e-06, |
|
"loss": 2.4778, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 9.51, |
|
"learning_rate": 9.890289831341086e-07, |
|
"loss": 2.4509, |
|
"step": 116100 |
|
}, |
|
{ |
|
"epoch": 9.51, |
|
"learning_rate": 9.726543310954643e-07, |
|
"loss": 2.4635, |
|
"step": 116200 |
|
}, |
|
{ |
|
"epoch": 9.52, |
|
"learning_rate": 9.562796790568202e-07, |
|
"loss": 2.4781, |
|
"step": 116300 |
|
}, |
|
{ |
|
"epoch": 9.53, |
|
"learning_rate": 9.399050270181759e-07, |
|
"loss": 2.451, |
|
"step": 116400 |
|
}, |
|
{ |
|
"epoch": 9.54, |
|
"learning_rate": 9.235303749795318e-07, |
|
"loss": 2.4803, |
|
"step": 116500 |
|
}, |
|
{ |
|
"epoch": 9.55, |
|
"learning_rate": 9.071557229408876e-07, |
|
"loss": 2.462, |
|
"step": 116600 |
|
}, |
|
{ |
|
"epoch": 9.55, |
|
"learning_rate": 8.907810709022434e-07, |
|
"loss": 2.502, |
|
"step": 116700 |
|
}, |
|
{ |
|
"epoch": 9.56, |
|
"learning_rate": 8.744064188635993e-07, |
|
"loss": 2.4988, |
|
"step": 116800 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 8.58031766824955e-07, |
|
"loss": 2.4858, |
|
"step": 116900 |
|
}, |
|
{ |
|
"epoch": 9.58, |
|
"learning_rate": 8.41657114786311e-07, |
|
"loss": 2.4803, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 9.59, |
|
"learning_rate": 8.252824627476667e-07, |
|
"loss": 2.4874, |
|
"step": 117100 |
|
}, |
|
{ |
|
"epoch": 9.6, |
|
"learning_rate": 8.089078107090224e-07, |
|
"loss": 2.4499, |
|
"step": 117200 |
|
}, |
|
{ |
|
"epoch": 9.6, |
|
"learning_rate": 7.925331586703782e-07, |
|
"loss": 2.4769, |
|
"step": 117300 |
|
}, |
|
{ |
|
"epoch": 9.61, |
|
"learning_rate": 7.761585066317341e-07, |
|
"loss": 2.465, |
|
"step": 117400 |
|
}, |
|
{ |
|
"epoch": 9.62, |
|
"learning_rate": 7.597838545930899e-07, |
|
"loss": 2.4572, |
|
"step": 117500 |
|
}, |
|
{ |
|
"epoch": 9.63, |
|
"learning_rate": 7.434092025544458e-07, |
|
"loss": 2.4699, |
|
"step": 117600 |
|
}, |
|
{ |
|
"epoch": 9.64, |
|
"learning_rate": 7.270345505158016e-07, |
|
"loss": 2.4953, |
|
"step": 117700 |
|
}, |
|
{ |
|
"epoch": 9.64, |
|
"learning_rate": 7.106598984771574e-07, |
|
"loss": 2.4646, |
|
"step": 117800 |
|
}, |
|
{ |
|
"epoch": 9.65, |
|
"learning_rate": 6.942852464385133e-07, |
|
"loss": 2.4412, |
|
"step": 117900 |
|
}, |
|
{ |
|
"epoch": 9.66, |
|
"learning_rate": 6.77910594399869e-07, |
|
"loss": 2.4873, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 9.67, |
|
"learning_rate": 6.615359423612248e-07, |
|
"loss": 2.5147, |
|
"step": 118100 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"learning_rate": 6.451612903225807e-07, |
|
"loss": 2.4683, |
|
"step": 118200 |
|
}, |
|
{ |
|
"epoch": 9.69, |
|
"learning_rate": 6.287866382839365e-07, |
|
"loss": 2.46, |
|
"step": 118300 |
|
}, |
|
{ |
|
"epoch": 9.69, |
|
"learning_rate": 6.124119862452924e-07, |
|
"loss": 2.4775, |
|
"step": 118400 |
|
}, |
|
{ |
|
"epoch": 9.7, |
|
"learning_rate": 5.960373342066481e-07, |
|
"loss": 2.5124, |
|
"step": 118500 |
|
}, |
|
{ |
|
"epoch": 9.71, |
|
"learning_rate": 5.79662682168004e-07, |
|
"loss": 2.4728, |
|
"step": 118600 |
|
}, |
|
{ |
|
"epoch": 9.72, |
|
"learning_rate": 5.632880301293598e-07, |
|
"loss": 2.4743, |
|
"step": 118700 |
|
}, |
|
{ |
|
"epoch": 9.73, |
|
"learning_rate": 5.469133780907157e-07, |
|
"loss": 2.5134, |
|
"step": 118800 |
|
}, |
|
{ |
|
"epoch": 9.73, |
|
"learning_rate": 5.305387260520714e-07, |
|
"loss": 2.4655, |
|
"step": 118900 |
|
}, |
|
{ |
|
"epoch": 9.74, |
|
"learning_rate": 5.141640740134273e-07, |
|
"loss": 2.4559, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 9.75, |
|
"learning_rate": 4.977894219747831e-07, |
|
"loss": 2.4724, |
|
"step": 119100 |
|
}, |
|
{ |
|
"epoch": 9.76, |
|
"learning_rate": 4.81414769936139e-07, |
|
"loss": 2.506, |
|
"step": 119200 |
|
}, |
|
{ |
|
"epoch": 9.77, |
|
"learning_rate": 4.650401178974947e-07, |
|
"loss": 2.4379, |
|
"step": 119300 |
|
}, |
|
{ |
|
"epoch": 9.78, |
|
"learning_rate": 4.486654658588505e-07, |
|
"loss": 2.4732, |
|
"step": 119400 |
|
}, |
|
{ |
|
"epoch": 9.78, |
|
"learning_rate": 4.3229081382020635e-07, |
|
"loss": 2.4776, |
|
"step": 119500 |
|
}, |
|
{ |
|
"epoch": 9.79, |
|
"learning_rate": 4.159161617815622e-07, |
|
"loss": 2.483, |
|
"step": 119600 |
|
}, |
|
{ |
|
"epoch": 9.8, |
|
"learning_rate": 3.99541509742918e-07, |
|
"loss": 2.4569, |
|
"step": 119700 |
|
}, |
|
{ |
|
"epoch": 9.81, |
|
"learning_rate": 3.831668577042738e-07, |
|
"loss": 2.4781, |
|
"step": 119800 |
|
}, |
|
{ |
|
"epoch": 9.82, |
|
"learning_rate": 3.6679220566562964e-07, |
|
"loss": 2.4737, |
|
"step": 119900 |
|
}, |
|
{ |
|
"epoch": 9.82, |
|
"learning_rate": 3.504175536269855e-07, |
|
"loss": 2.4749, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 9.83, |
|
"learning_rate": 3.3404290158834123e-07, |
|
"loss": 2.4582, |
|
"step": 120100 |
|
}, |
|
{ |
|
"epoch": 9.84, |
|
"learning_rate": 3.176682495496971e-07, |
|
"loss": 2.4546, |
|
"step": 120200 |
|
}, |
|
{ |
|
"epoch": 9.85, |
|
"learning_rate": 3.0129359751105293e-07, |
|
"loss": 2.4568, |
|
"step": 120300 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 2.849189454724087e-07, |
|
"loss": 2.4637, |
|
"step": 120400 |
|
}, |
|
{ |
|
"epoch": 9.87, |
|
"learning_rate": 2.6854429343376457e-07, |
|
"loss": 2.479, |
|
"step": 120500 |
|
}, |
|
{ |
|
"epoch": 9.87, |
|
"learning_rate": 2.5216964139512036e-07, |
|
"loss": 2.4644, |
|
"step": 120600 |
|
}, |
|
{ |
|
"epoch": 9.88, |
|
"learning_rate": 2.3579498935647619e-07, |
|
"loss": 2.4845, |
|
"step": 120700 |
|
}, |
|
{ |
|
"epoch": 9.89, |
|
"learning_rate": 2.19420337317832e-07, |
|
"loss": 2.4691, |
|
"step": 120800 |
|
}, |
|
{ |
|
"epoch": 9.9, |
|
"learning_rate": 2.0304568527918783e-07, |
|
"loss": 2.4972, |
|
"step": 120900 |
|
}, |
|
{ |
|
"epoch": 9.91, |
|
"learning_rate": 1.8667103324054365e-07, |
|
"loss": 2.4836, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 9.91, |
|
"learning_rate": 1.7029638120189947e-07, |
|
"loss": 2.4774, |
|
"step": 121100 |
|
}, |
|
{ |
|
"epoch": 9.92, |
|
"learning_rate": 1.539217291632553e-07, |
|
"loss": 2.4536, |
|
"step": 121200 |
|
}, |
|
{ |
|
"epoch": 9.93, |
|
"learning_rate": 1.3754707712461112e-07, |
|
"loss": 2.4536, |
|
"step": 121300 |
|
}, |
|
{ |
|
"epoch": 9.94, |
|
"learning_rate": 1.2117242508596694e-07, |
|
"loss": 2.4993, |
|
"step": 121400 |
|
}, |
|
{ |
|
"epoch": 9.95, |
|
"learning_rate": 1.0479777304732276e-07, |
|
"loss": 2.5068, |
|
"step": 121500 |
|
}, |
|
{ |
|
"epoch": 9.96, |
|
"learning_rate": 8.842312100867858e-08, |
|
"loss": 2.4678, |
|
"step": 121600 |
|
}, |
|
{ |
|
"epoch": 9.96, |
|
"learning_rate": 7.204846897003439e-08, |
|
"loss": 2.4577, |
|
"step": 121700 |
|
}, |
|
{ |
|
"epoch": 9.97, |
|
"learning_rate": 5.567381693139021e-08, |
|
"loss": 2.4623, |
|
"step": 121800 |
|
}, |
|
{ |
|
"epoch": 9.98, |
|
"learning_rate": 3.929916489274603e-08, |
|
"loss": 2.4673, |
|
"step": 121900 |
|
}, |
|
{ |
|
"epoch": 9.99, |
|
"learning_rate": 2.2924512854101853e-08, |
|
"loss": 2.5057, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 6.549860815457672e-09, |
|
"loss": 2.4638, |
|
"step": 122100 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_bleu": 995.492, |
|
"eval_chrf": 2879.8334, |
|
"eval_chrf_plus": 2710.1728, |
|
"eval_gen_len": 14.645, |
|
"eval_loss": 2.3738529682159424, |
|
"eval_runtime": 38.7293, |
|
"eval_samples_per_second": 25.82, |
|
"eval_steps_per_second": 1.084, |
|
"step": 122140 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 122140, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"total_flos": 2.412332343172915e+18, |
|
"train_batch_size": 24, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|