|
{ |
|
"best_metric": 0.6590909090909091, |
|
"best_model_checkpoint": "squarerun2/checkpoint-522", |
|
"epoch": 20.0, |
|
"eval_steps": 500, |
|
"global_step": 580, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.06896551724137931, |
|
"grad_norm": 2.9724416732788086, |
|
"learning_rate": 3.448275862068966e-06, |
|
"loss": 1.9767, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.13793103448275862, |
|
"grad_norm": 1.7837220430374146, |
|
"learning_rate": 6.896551724137932e-06, |
|
"loss": 1.9373, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.20689655172413793, |
|
"grad_norm": 1.9200481176376343, |
|
"learning_rate": 1.0344827586206897e-05, |
|
"loss": 1.9422, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.27586206896551724, |
|
"grad_norm": 1.8607347011566162, |
|
"learning_rate": 1.3793103448275863e-05, |
|
"loss": 1.9203, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.3448275862068966, |
|
"grad_norm": 2.282787322998047, |
|
"learning_rate": 1.7241379310344828e-05, |
|
"loss": 1.8967, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.41379310344827586, |
|
"grad_norm": 1.8788440227508545, |
|
"learning_rate": 2.0689655172413793e-05, |
|
"loss": 1.9576, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.4827586206896552, |
|
"grad_norm": 1.689126968383789, |
|
"learning_rate": 2.413793103448276e-05, |
|
"loss": 1.9238, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.5517241379310345, |
|
"grad_norm": 1.5710227489471436, |
|
"learning_rate": 2.7586206896551727e-05, |
|
"loss": 1.9468, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.6206896551724138, |
|
"grad_norm": 1.6929808855056763, |
|
"learning_rate": 3.103448275862069e-05, |
|
"loss": 1.83, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.6896551724137931, |
|
"grad_norm": 1.9147762060165405, |
|
"learning_rate": 3.4482758620689657e-05, |
|
"loss": 1.8844, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.7586206896551724, |
|
"grad_norm": 1.0664433240890503, |
|
"learning_rate": 3.793103448275862e-05, |
|
"loss": 1.8893, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.8275862068965517, |
|
"grad_norm": 1.2909166812896729, |
|
"learning_rate": 4.1379310344827587e-05, |
|
"loss": 1.8794, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.896551724137931, |
|
"grad_norm": 2.562612771987915, |
|
"learning_rate": 4.482758620689655e-05, |
|
"loss": 1.8722, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.9655172413793104, |
|
"grad_norm": 1.766555905342102, |
|
"learning_rate": 4.827586206896552e-05, |
|
"loss": 1.9016, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.24242424242424243, |
|
"eval_f1_macro": 0.1010989010989011, |
|
"eval_f1_micro": 0.24242424242424243, |
|
"eval_f1_weighted": 0.14009324009324012, |
|
"eval_loss": 1.8763961791992188, |
|
"eval_precision_macro": 0.07205977686035006, |
|
"eval_precision_micro": 0.24242424242424243, |
|
"eval_precision_weighted": 0.10010783738990092, |
|
"eval_recall_macro": 0.1760846560846561, |
|
"eval_recall_micro": 0.24242424242424243, |
|
"eval_recall_weighted": 0.24242424242424243, |
|
"eval_runtime": 2.3126, |
|
"eval_samples_per_second": 57.08, |
|
"eval_steps_per_second": 7.351, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 1.0344827586206897, |
|
"grad_norm": 2.116236448287964, |
|
"learning_rate": 5.172413793103449e-05, |
|
"loss": 1.9127, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 1.103448275862069, |
|
"grad_norm": 1.5244780778884888, |
|
"learning_rate": 5.517241379310345e-05, |
|
"loss": 1.8217, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 1.1724137931034484, |
|
"grad_norm": 1.2253153324127197, |
|
"learning_rate": 5.862068965517241e-05, |
|
"loss": 1.8662, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 1.2413793103448276, |
|
"grad_norm": 1.1567929983139038, |
|
"learning_rate": 6.206896551724138e-05, |
|
"loss": 1.9031, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 1.3103448275862069, |
|
"grad_norm": 1.4619667530059814, |
|
"learning_rate": 6.551724137931034e-05, |
|
"loss": 1.8464, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 1.3793103448275863, |
|
"grad_norm": 1.5786610841751099, |
|
"learning_rate": 6.896551724137931e-05, |
|
"loss": 1.9057, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 1.4482758620689655, |
|
"grad_norm": 1.1554012298583984, |
|
"learning_rate": 7.241379310344828e-05, |
|
"loss": 1.8748, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 1.5172413793103448, |
|
"grad_norm": 2.0982630252838135, |
|
"learning_rate": 7.586206896551724e-05, |
|
"loss": 1.9052, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 1.5862068965517242, |
|
"grad_norm": 1.046858787536621, |
|
"learning_rate": 7.931034482758621e-05, |
|
"loss": 1.9363, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 1.6551724137931034, |
|
"grad_norm": 1.05802583694458, |
|
"learning_rate": 8.275862068965517e-05, |
|
"loss": 1.8712, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 1.7241379310344827, |
|
"grad_norm": 1.0946681499481201, |
|
"learning_rate": 8.620689655172413e-05, |
|
"loss": 1.9325, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.793103448275862, |
|
"grad_norm": 1.3410812616348267, |
|
"learning_rate": 8.96551724137931e-05, |
|
"loss": 1.8671, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 1.8620689655172413, |
|
"grad_norm": 1.4638521671295166, |
|
"learning_rate": 9.310344827586207e-05, |
|
"loss": 1.9462, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 1.9310344827586206, |
|
"grad_norm": 1.6821434497833252, |
|
"learning_rate": 9.655172413793105e-05, |
|
"loss": 1.8823, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.771202564239502, |
|
"learning_rate": 0.0001, |
|
"loss": 1.8787, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.20454545454545456, |
|
"eval_f1_macro": 0.04851752021563342, |
|
"eval_f1_micro": 0.20454545454545456, |
|
"eval_f1_weighted": 0.06946826758147512, |
|
"eval_loss": 1.8749967813491821, |
|
"eval_precision_macro": 0.029220779220779224, |
|
"eval_precision_micro": 0.20454545454545456, |
|
"eval_precision_weighted": 0.04183884297520662, |
|
"eval_recall_macro": 0.14285714285714285, |
|
"eval_recall_micro": 0.20454545454545456, |
|
"eval_recall_weighted": 0.20454545454545456, |
|
"eval_runtime": 2.315, |
|
"eval_samples_per_second": 57.02, |
|
"eval_steps_per_second": 7.343, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 2.0689655172413794, |
|
"grad_norm": 1.4094064235687256, |
|
"learning_rate": 9.96168582375479e-05, |
|
"loss": 1.9068, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 2.1379310344827585, |
|
"grad_norm": 0.7938138246536255, |
|
"learning_rate": 9.92337164750958e-05, |
|
"loss": 1.847, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 2.206896551724138, |
|
"grad_norm": 1.7750108242034912, |
|
"learning_rate": 9.885057471264369e-05, |
|
"loss": 1.8632, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 2.2758620689655173, |
|
"grad_norm": 2.120215654373169, |
|
"learning_rate": 9.846743295019157e-05, |
|
"loss": 1.9374, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 2.344827586206897, |
|
"grad_norm": 1.1951868534088135, |
|
"learning_rate": 9.808429118773947e-05, |
|
"loss": 1.789, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 2.413793103448276, |
|
"grad_norm": 1.5802347660064697, |
|
"learning_rate": 9.770114942528736e-05, |
|
"loss": 1.8232, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 2.4827586206896552, |
|
"grad_norm": 1.5989313125610352, |
|
"learning_rate": 9.731800766283526e-05, |
|
"loss": 1.9379, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 2.5517241379310347, |
|
"grad_norm": 1.3577500581741333, |
|
"learning_rate": 9.693486590038314e-05, |
|
"loss": 1.905, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 2.6206896551724137, |
|
"grad_norm": 1.2363930940628052, |
|
"learning_rate": 9.655172413793105e-05, |
|
"loss": 1.8514, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 2.689655172413793, |
|
"grad_norm": 1.6196924448013306, |
|
"learning_rate": 9.616858237547893e-05, |
|
"loss": 1.9273, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 2.7586206896551726, |
|
"grad_norm": 1.13129460811615, |
|
"learning_rate": 9.578544061302682e-05, |
|
"loss": 1.9407, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 2.8275862068965516, |
|
"grad_norm": 1.0212408304214478, |
|
"learning_rate": 9.540229885057472e-05, |
|
"loss": 1.8273, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 2.896551724137931, |
|
"grad_norm": 0.8608859777450562, |
|
"learning_rate": 9.501915708812261e-05, |
|
"loss": 1.8769, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 2.9655172413793105, |
|
"grad_norm": 1.2102094888687134, |
|
"learning_rate": 9.463601532567051e-05, |
|
"loss": 1.9345, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.20454545454545456, |
|
"eval_f1_macro": 0.04851752021563342, |
|
"eval_f1_micro": 0.20454545454545456, |
|
"eval_f1_weighted": 0.06946826758147512, |
|
"eval_loss": 1.8624200820922852, |
|
"eval_precision_macro": 0.029220779220779224, |
|
"eval_precision_micro": 0.20454545454545456, |
|
"eval_precision_weighted": 0.04183884297520662, |
|
"eval_recall_macro": 0.14285714285714285, |
|
"eval_recall_micro": 0.20454545454545456, |
|
"eval_recall_weighted": 0.20454545454545456, |
|
"eval_runtime": 2.3224, |
|
"eval_samples_per_second": 56.838, |
|
"eval_steps_per_second": 7.32, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 3.0344827586206895, |
|
"grad_norm": 1.7206976413726807, |
|
"learning_rate": 9.425287356321839e-05, |
|
"loss": 1.9875, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 3.103448275862069, |
|
"grad_norm": 1.0909031629562378, |
|
"learning_rate": 9.38697318007663e-05, |
|
"loss": 1.7979, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 3.1724137931034484, |
|
"grad_norm": 1.2875232696533203, |
|
"learning_rate": 9.348659003831418e-05, |
|
"loss": 1.8964, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 3.2413793103448274, |
|
"grad_norm": 1.6919517517089844, |
|
"learning_rate": 9.310344827586207e-05, |
|
"loss": 1.8316, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 3.310344827586207, |
|
"grad_norm": 1.7749555110931396, |
|
"learning_rate": 9.272030651340997e-05, |
|
"loss": 1.8712, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 3.3793103448275863, |
|
"grad_norm": 1.273366093635559, |
|
"learning_rate": 9.233716475095786e-05, |
|
"loss": 1.8105, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 3.4482758620689653, |
|
"grad_norm": 1.3420616388320923, |
|
"learning_rate": 9.195402298850575e-05, |
|
"loss": 1.7223, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 3.5172413793103448, |
|
"grad_norm": 2.060871124267578, |
|
"learning_rate": 9.157088122605364e-05, |
|
"loss": 1.9009, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 3.586206896551724, |
|
"grad_norm": 1.3717443943023682, |
|
"learning_rate": 9.118773946360154e-05, |
|
"loss": 1.7423, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 3.655172413793103, |
|
"grad_norm": 2.35315203666687, |
|
"learning_rate": 9.080459770114943e-05, |
|
"loss": 1.8381, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 3.7241379310344827, |
|
"grad_norm": 1.7000161409378052, |
|
"learning_rate": 9.042145593869731e-05, |
|
"loss": 1.6663, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 3.793103448275862, |
|
"grad_norm": 2.3289718627929688, |
|
"learning_rate": 9.003831417624522e-05, |
|
"loss": 1.7653, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 3.862068965517241, |
|
"grad_norm": 2.581888437271118, |
|
"learning_rate": 8.96551724137931e-05, |
|
"loss": 1.8899, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 3.9310344827586206, |
|
"grad_norm": 2.3343546390533447, |
|
"learning_rate": 8.9272030651341e-05, |
|
"loss": 1.7854, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 1.989835262298584, |
|
"learning_rate": 8.888888888888889e-05, |
|
"loss": 1.6663, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.3560606060606061, |
|
"eval_f1_macro": 0.2230450801010059, |
|
"eval_f1_micro": 0.3560606060606061, |
|
"eval_f1_weighted": 0.27375266721108504, |
|
"eval_loss": 1.7238999605178833, |
|
"eval_precision_macro": 0.31727902915703876, |
|
"eval_precision_micro": 0.3560606060606061, |
|
"eval_precision_weighted": 0.3548852832560697, |
|
"eval_recall_macro": 0.27251700680272106, |
|
"eval_recall_micro": 0.3560606060606061, |
|
"eval_recall_weighted": 0.3560606060606061, |
|
"eval_runtime": 2.3156, |
|
"eval_samples_per_second": 57.004, |
|
"eval_steps_per_second": 7.341, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 4.068965517241379, |
|
"grad_norm": 2.951305389404297, |
|
"learning_rate": 8.850574712643679e-05, |
|
"loss": 1.6573, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 4.137931034482759, |
|
"grad_norm": 2.0035061836242676, |
|
"learning_rate": 8.812260536398468e-05, |
|
"loss": 1.5898, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 4.206896551724138, |
|
"grad_norm": 1.7943332195281982, |
|
"learning_rate": 8.773946360153256e-05, |
|
"loss": 1.5183, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 4.275862068965517, |
|
"grad_norm": 2.8160884380340576, |
|
"learning_rate": 8.735632183908047e-05, |
|
"loss": 1.3985, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 4.344827586206897, |
|
"grad_norm": 2.5276668071746826, |
|
"learning_rate": 8.697318007662835e-05, |
|
"loss": 1.5588, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 4.413793103448276, |
|
"grad_norm": 1.8822064399719238, |
|
"learning_rate": 8.659003831417625e-05, |
|
"loss": 1.6377, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 4.482758620689655, |
|
"grad_norm": 3.416983127593994, |
|
"learning_rate": 8.620689655172413e-05, |
|
"loss": 1.8148, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 4.551724137931035, |
|
"grad_norm": 3.465373992919922, |
|
"learning_rate": 8.582375478927204e-05, |
|
"loss": 1.7206, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 4.620689655172414, |
|
"grad_norm": 2.2419955730438232, |
|
"learning_rate": 8.544061302681992e-05, |
|
"loss": 1.8184, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 4.689655172413794, |
|
"grad_norm": 2.772228240966797, |
|
"learning_rate": 8.505747126436782e-05, |
|
"loss": 1.6022, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 4.758620689655173, |
|
"grad_norm": 2.6388156414031982, |
|
"learning_rate": 8.467432950191571e-05, |
|
"loss": 1.5941, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 4.827586206896552, |
|
"grad_norm": 2.489682674407959, |
|
"learning_rate": 8.42911877394636e-05, |
|
"loss": 1.476, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 4.896551724137931, |
|
"grad_norm": 2.517737865447998, |
|
"learning_rate": 8.39080459770115e-05, |
|
"loss": 1.5529, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 4.9655172413793105, |
|
"grad_norm": 1.9954488277435303, |
|
"learning_rate": 8.35249042145594e-05, |
|
"loss": 1.3847, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.4696969696969697, |
|
"eval_f1_macro": 0.3420460214829701, |
|
"eval_f1_micro": 0.4696969696969697, |
|
"eval_f1_weighted": 0.4038214256266929, |
|
"eval_loss": 1.4879895448684692, |
|
"eval_precision_macro": 0.4520649449711235, |
|
"eval_precision_micro": 0.4696969696969697, |
|
"eval_precision_weighted": 0.4845688001294408, |
|
"eval_recall_macro": 0.38925170068027215, |
|
"eval_recall_micro": 0.4696969696969697, |
|
"eval_recall_weighted": 0.4696969696969697, |
|
"eval_runtime": 2.3198, |
|
"eval_samples_per_second": 56.9, |
|
"eval_steps_per_second": 7.328, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 5.0344827586206895, |
|
"grad_norm": 2.7297556400299072, |
|
"learning_rate": 8.314176245210729e-05, |
|
"loss": 1.6994, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 5.103448275862069, |
|
"grad_norm": 2.7157135009765625, |
|
"learning_rate": 8.275862068965517e-05, |
|
"loss": 1.409, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 5.172413793103448, |
|
"grad_norm": 3.6584227085113525, |
|
"learning_rate": 8.237547892720307e-05, |
|
"loss": 1.3627, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 5.241379310344827, |
|
"grad_norm": 2.6041762828826904, |
|
"learning_rate": 8.199233716475096e-05, |
|
"loss": 1.3385, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 5.310344827586207, |
|
"grad_norm": 2.7281856536865234, |
|
"learning_rate": 8.160919540229886e-05, |
|
"loss": 1.4468, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 5.379310344827586, |
|
"grad_norm": 2.5631356239318848, |
|
"learning_rate": 8.122605363984674e-05, |
|
"loss": 1.3639, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 5.448275862068965, |
|
"grad_norm": 3.118602991104126, |
|
"learning_rate": 8.084291187739465e-05, |
|
"loss": 1.5097, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 5.517241379310345, |
|
"grad_norm": 2.653867483139038, |
|
"learning_rate": 8.045977011494253e-05, |
|
"loss": 1.3556, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 5.586206896551724, |
|
"grad_norm": 2.3653130531311035, |
|
"learning_rate": 8.007662835249042e-05, |
|
"loss": 1.504, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 5.655172413793103, |
|
"grad_norm": 2.039597749710083, |
|
"learning_rate": 7.969348659003832e-05, |
|
"loss": 1.3222, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 5.724137931034483, |
|
"grad_norm": 2.91522216796875, |
|
"learning_rate": 7.931034482758621e-05, |
|
"loss": 1.2761, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 5.793103448275862, |
|
"grad_norm": 2.8554606437683105, |
|
"learning_rate": 7.892720306513411e-05, |
|
"loss": 1.3019, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 5.862068965517241, |
|
"grad_norm": 3.730526924133301, |
|
"learning_rate": 7.854406130268199e-05, |
|
"loss": 1.2626, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 5.931034482758621, |
|
"grad_norm": 5.193080425262451, |
|
"learning_rate": 7.81609195402299e-05, |
|
"loss": 1.5394, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"grad_norm": 2.901132583618164, |
|
"learning_rate": 7.777777777777778e-05, |
|
"loss": 1.6559, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_accuracy": 0.4772727272727273, |
|
"eval_f1_macro": 0.3479108279730525, |
|
"eval_f1_micro": 0.4772727272727273, |
|
"eval_f1_weighted": 0.4108418905152118, |
|
"eval_loss": 1.4056475162506104, |
|
"eval_precision_macro": 0.3864548957927566, |
|
"eval_precision_micro": 0.4772727272727273, |
|
"eval_precision_weighted": 0.42757837676034927, |
|
"eval_recall_macro": 0.38701436130007555, |
|
"eval_recall_micro": 0.4772727272727273, |
|
"eval_recall_weighted": 0.4772727272727273, |
|
"eval_runtime": 2.3209, |
|
"eval_samples_per_second": 56.874, |
|
"eval_steps_per_second": 7.325, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 6.068965517241379, |
|
"grad_norm": 3.3514347076416016, |
|
"learning_rate": 7.739463601532568e-05, |
|
"loss": 1.2022, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 6.137931034482759, |
|
"grad_norm": 2.6016125679016113, |
|
"learning_rate": 7.701149425287356e-05, |
|
"loss": 1.1793, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 6.206896551724138, |
|
"grad_norm": 3.9388816356658936, |
|
"learning_rate": 7.662835249042147e-05, |
|
"loss": 1.2841, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 6.275862068965517, |
|
"grad_norm": 2.01733660697937, |
|
"learning_rate": 7.624521072796935e-05, |
|
"loss": 1.3875, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 6.344827586206897, |
|
"grad_norm": 2.873060703277588, |
|
"learning_rate": 7.586206896551724e-05, |
|
"loss": 1.3029, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 6.413793103448276, |
|
"grad_norm": 3.8860301971435547, |
|
"learning_rate": 7.547892720306514e-05, |
|
"loss": 1.0694, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 6.482758620689655, |
|
"grad_norm": 4.136936664581299, |
|
"learning_rate": 7.509578544061303e-05, |
|
"loss": 1.0778, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 6.551724137931035, |
|
"grad_norm": 2.326791763305664, |
|
"learning_rate": 7.471264367816091e-05, |
|
"loss": 1.2761, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 6.620689655172414, |
|
"grad_norm": 3.5550026893615723, |
|
"learning_rate": 7.432950191570882e-05, |
|
"loss": 1.3036, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 6.689655172413794, |
|
"grad_norm": 2.3610475063323975, |
|
"learning_rate": 7.39463601532567e-05, |
|
"loss": 1.2765, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 6.758620689655173, |
|
"grad_norm": 3.0064663887023926, |
|
"learning_rate": 7.35632183908046e-05, |
|
"loss": 1.5579, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 6.827586206896552, |
|
"grad_norm": 2.8558247089385986, |
|
"learning_rate": 7.31800766283525e-05, |
|
"loss": 1.1987, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 6.896551724137931, |
|
"grad_norm": 3.3801021575927734, |
|
"learning_rate": 7.279693486590039e-05, |
|
"loss": 1.1622, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 6.9655172413793105, |
|
"grad_norm": 3.4124093055725098, |
|
"learning_rate": 7.241379310344828e-05, |
|
"loss": 1.335, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_accuracy": 0.5151515151515151, |
|
"eval_f1_macro": 0.38747536212974004, |
|
"eval_f1_micro": 0.5151515151515151, |
|
"eval_f1_weighted": 0.4526784325171422, |
|
"eval_loss": 1.3767547607421875, |
|
"eval_precision_macro": 0.3933148791431466, |
|
"eval_precision_micro": 0.5151515151515151, |
|
"eval_precision_weighted": 0.44469956019158147, |
|
"eval_recall_macro": 0.4265457294028723, |
|
"eval_recall_micro": 0.5151515151515151, |
|
"eval_recall_weighted": 0.5151515151515151, |
|
"eval_runtime": 2.312, |
|
"eval_samples_per_second": 57.094, |
|
"eval_steps_per_second": 7.353, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 7.0344827586206895, |
|
"grad_norm": 3.0827553272247314, |
|
"learning_rate": 7.203065134099617e-05, |
|
"loss": 1.0667, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 7.103448275862069, |
|
"grad_norm": 1.9460029602050781, |
|
"learning_rate": 7.164750957854408e-05, |
|
"loss": 1.0915, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 7.172413793103448, |
|
"grad_norm": 3.7591891288757324, |
|
"learning_rate": 7.126436781609196e-05, |
|
"loss": 1.2813, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 7.241379310344827, |
|
"grad_norm": 2.6194112300872803, |
|
"learning_rate": 7.088122605363985e-05, |
|
"loss": 1.1527, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 7.310344827586207, |
|
"grad_norm": 3.020498514175415, |
|
"learning_rate": 7.049808429118773e-05, |
|
"loss": 1.2731, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 7.379310344827586, |
|
"grad_norm": 3.303375005722046, |
|
"learning_rate": 7.011494252873564e-05, |
|
"loss": 1.0312, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 7.448275862068965, |
|
"grad_norm": 2.705366849899292, |
|
"learning_rate": 6.973180076628352e-05, |
|
"loss": 1.2032, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 7.517241379310345, |
|
"grad_norm": 3.3683817386627197, |
|
"learning_rate": 6.934865900383142e-05, |
|
"loss": 1.166, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 7.586206896551724, |
|
"grad_norm": 3.079350471496582, |
|
"learning_rate": 6.896551724137931e-05, |
|
"loss": 0.9884, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 7.655172413793103, |
|
"grad_norm": 2.3934454917907715, |
|
"learning_rate": 6.858237547892721e-05, |
|
"loss": 1.1715, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 7.724137931034483, |
|
"grad_norm": 3.6560773849487305, |
|
"learning_rate": 6.81992337164751e-05, |
|
"loss": 1.0007, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 7.793103448275862, |
|
"grad_norm": 4.3467912673950195, |
|
"learning_rate": 6.781609195402298e-05, |
|
"loss": 1.4675, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 7.862068965517241, |
|
"grad_norm": 2.5141751766204834, |
|
"learning_rate": 6.74329501915709e-05, |
|
"loss": 0.9672, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 7.931034482758621, |
|
"grad_norm": 4.107297897338867, |
|
"learning_rate": 6.704980842911877e-05, |
|
"loss": 1.4593, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"grad_norm": 2.193042278289795, |
|
"learning_rate": 6.666666666666667e-05, |
|
"loss": 1.2514, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_accuracy": 0.5606060606060606, |
|
"eval_f1_macro": 0.4536205964777393, |
|
"eval_f1_micro": 0.5606060606060606, |
|
"eval_f1_weighted": 0.5206725615816525, |
|
"eval_loss": 1.2344759702682495, |
|
"eval_precision_macro": 0.47009218749545173, |
|
"eval_precision_micro": 0.5606060606060606, |
|
"eval_precision_weighted": 0.5257434050758826, |
|
"eval_recall_macro": 0.4766061980347694, |
|
"eval_recall_micro": 0.5606060606060606, |
|
"eval_recall_weighted": 0.5606060606060606, |
|
"eval_runtime": 2.3242, |
|
"eval_samples_per_second": 56.794, |
|
"eval_steps_per_second": 7.314, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 8.068965517241379, |
|
"grad_norm": 3.166414260864258, |
|
"learning_rate": 6.628352490421456e-05, |
|
"loss": 0.8487, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 8.137931034482758, |
|
"grad_norm": 3.4050002098083496, |
|
"learning_rate": 6.590038314176246e-05, |
|
"loss": 1.2154, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 8.206896551724139, |
|
"grad_norm": 2.827500104904175, |
|
"learning_rate": 6.551724137931034e-05, |
|
"loss": 1.1884, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 8.275862068965518, |
|
"grad_norm": 3.644810199737549, |
|
"learning_rate": 6.513409961685824e-05, |
|
"loss": 1.1444, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 8.344827586206897, |
|
"grad_norm": 2.5931830406188965, |
|
"learning_rate": 6.475095785440613e-05, |
|
"loss": 0.9335, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 8.413793103448276, |
|
"grad_norm": 4.586955547332764, |
|
"learning_rate": 6.436781609195403e-05, |
|
"loss": 1.174, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 8.482758620689655, |
|
"grad_norm": 4.06891393661499, |
|
"learning_rate": 6.398467432950191e-05, |
|
"loss": 1.1579, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 8.551724137931034, |
|
"grad_norm": 3.6428885459899902, |
|
"learning_rate": 6.360153256704982e-05, |
|
"loss": 0.8764, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 8.620689655172415, |
|
"grad_norm": 6.163786888122559, |
|
"learning_rate": 6.32183908045977e-05, |
|
"loss": 1.0204, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 8.689655172413794, |
|
"grad_norm": 2.775923252105713, |
|
"learning_rate": 6.283524904214559e-05, |
|
"loss": 0.8183, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 8.758620689655173, |
|
"grad_norm": 3.2083027362823486, |
|
"learning_rate": 6.245210727969349e-05, |
|
"loss": 1.3757, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 8.827586206896552, |
|
"grad_norm": 5.378493309020996, |
|
"learning_rate": 6.206896551724138e-05, |
|
"loss": 1.0219, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 8.89655172413793, |
|
"grad_norm": 3.8027050495147705, |
|
"learning_rate": 6.168582375478928e-05, |
|
"loss": 1.0405, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 8.96551724137931, |
|
"grad_norm": 3.604165554046631, |
|
"learning_rate": 6.130268199233716e-05, |
|
"loss": 0.6979, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_accuracy": 0.6363636363636364, |
|
"eval_f1_macro": 0.530518587020135, |
|
"eval_f1_micro": 0.6363636363636364, |
|
"eval_f1_weighted": 0.6097181236324214, |
|
"eval_loss": 1.1500554084777832, |
|
"eval_precision_macro": 0.5490794091920358, |
|
"eval_precision_micro": 0.6363636363636364, |
|
"eval_precision_weighted": 0.6127338831200314, |
|
"eval_recall_macro": 0.5391232048374907, |
|
"eval_recall_micro": 0.6363636363636364, |
|
"eval_recall_weighted": 0.6363636363636364, |
|
"eval_runtime": 2.3286, |
|
"eval_samples_per_second": 56.687, |
|
"eval_steps_per_second": 7.301, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 9.03448275862069, |
|
"grad_norm": 2.9546728134155273, |
|
"learning_rate": 6.091954022988506e-05, |
|
"loss": 0.9733, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 9.10344827586207, |
|
"grad_norm": 3.742382049560547, |
|
"learning_rate": 6.053639846743295e-05, |
|
"loss": 0.8539, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 9.172413793103448, |
|
"grad_norm": 4.167463779449463, |
|
"learning_rate": 6.0153256704980845e-05, |
|
"loss": 0.9605, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 9.241379310344827, |
|
"grad_norm": 2.0738613605499268, |
|
"learning_rate": 5.977011494252874e-05, |
|
"loss": 0.7348, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 9.310344827586206, |
|
"grad_norm": 3.2274346351623535, |
|
"learning_rate": 5.9386973180076635e-05, |
|
"loss": 0.9138, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 9.379310344827585, |
|
"grad_norm": 3.5369961261749268, |
|
"learning_rate": 5.900383141762452e-05, |
|
"loss": 0.7026, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 9.448275862068966, |
|
"grad_norm": 3.977078914642334, |
|
"learning_rate": 5.862068965517241e-05, |
|
"loss": 1.0063, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 9.517241379310345, |
|
"grad_norm": 3.6526713371276855, |
|
"learning_rate": 5.823754789272031e-05, |
|
"loss": 0.7266, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 9.586206896551724, |
|
"grad_norm": 1.9557095766067505, |
|
"learning_rate": 5.78544061302682e-05, |
|
"loss": 0.7732, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 9.655172413793103, |
|
"grad_norm": 1.2573161125183105, |
|
"learning_rate": 5.747126436781609e-05, |
|
"loss": 0.5661, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 9.724137931034482, |
|
"grad_norm": 2.927410364151001, |
|
"learning_rate": 5.708812260536399e-05, |
|
"loss": 1.0745, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 9.793103448275861, |
|
"grad_norm": 4.574418544769287, |
|
"learning_rate": 5.670498084291188e-05, |
|
"loss": 0.9002, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 9.862068965517242, |
|
"grad_norm": 3.5535728931427, |
|
"learning_rate": 5.632183908045977e-05, |
|
"loss": 0.9265, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 9.931034482758621, |
|
"grad_norm": 4.7521162033081055, |
|
"learning_rate": 5.593869731800766e-05, |
|
"loss": 0.8453, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 4.0816216468811035, |
|
"learning_rate": 5.555555555555556e-05, |
|
"loss": 1.0417, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_accuracy": 0.6136363636363636, |
|
"eval_f1_macro": 0.5205916396815399, |
|
"eval_f1_micro": 0.6136363636363636, |
|
"eval_f1_weighted": 0.5899988122518645, |
|
"eval_loss": 1.1654376983642578, |
|
"eval_precision_macro": 0.5214674214674214, |
|
"eval_precision_micro": 0.6136363636363636, |
|
"eval_precision_weighted": 0.5934943844034754, |
|
"eval_recall_macro": 0.5463643235071807, |
|
"eval_recall_micro": 0.6136363636363636, |
|
"eval_recall_weighted": 0.6136363636363636, |
|
"eval_runtime": 2.3259, |
|
"eval_samples_per_second": 56.753, |
|
"eval_steps_per_second": 7.309, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 10.068965517241379, |
|
"grad_norm": 4.20100736618042, |
|
"learning_rate": 5.517241379310345e-05, |
|
"loss": 0.9384, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 10.137931034482758, |
|
"grad_norm": 3.5482981204986572, |
|
"learning_rate": 5.478927203065134e-05, |
|
"loss": 0.7163, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 10.206896551724139, |
|
"grad_norm": 5.511715412139893, |
|
"learning_rate": 5.440613026819924e-05, |
|
"loss": 1.0096, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 10.275862068965518, |
|
"grad_norm": 4.4264984130859375, |
|
"learning_rate": 5.402298850574713e-05, |
|
"loss": 0.775, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 10.344827586206897, |
|
"grad_norm": 2.2772464752197266, |
|
"learning_rate": 5.363984674329502e-05, |
|
"loss": 1.0071, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 10.413793103448276, |
|
"grad_norm": 2.6942243576049805, |
|
"learning_rate": 5.325670498084292e-05, |
|
"loss": 0.6, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 10.482758620689655, |
|
"grad_norm": 3.3119990825653076, |
|
"learning_rate": 5.287356321839081e-05, |
|
"loss": 0.7233, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 10.551724137931034, |
|
"grad_norm": 3.8363144397735596, |
|
"learning_rate": 5.24904214559387e-05, |
|
"loss": 1.084, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 10.620689655172415, |
|
"grad_norm": 3.457904815673828, |
|
"learning_rate": 5.2107279693486586e-05, |
|
"loss": 0.7188, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 10.689655172413794, |
|
"grad_norm": 4.752615451812744, |
|
"learning_rate": 5.172413793103449e-05, |
|
"loss": 1.0076, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 10.758620689655173, |
|
"grad_norm": 2.8279869556427, |
|
"learning_rate": 5.1340996168582377e-05, |
|
"loss": 1.0505, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 10.827586206896552, |
|
"grad_norm": 5.538084506988525, |
|
"learning_rate": 5.0957854406130265e-05, |
|
"loss": 0.6474, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 10.89655172413793, |
|
"grad_norm": 3.197183847427368, |
|
"learning_rate": 5.057471264367817e-05, |
|
"loss": 0.6146, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 10.96551724137931, |
|
"grad_norm": 3.427384614944458, |
|
"learning_rate": 5.0191570881226055e-05, |
|
"loss": 0.7314, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_accuracy": 0.6212121212121212, |
|
"eval_f1_macro": 0.537594465565654, |
|
"eval_f1_micro": 0.6212121212121212, |
|
"eval_f1_weighted": 0.6109434365736887, |
|
"eval_loss": 1.1566157341003418, |
|
"eval_precision_macro": 0.5386796536796536, |
|
"eval_precision_micro": 0.6212121212121212, |
|
"eval_precision_weighted": 0.6153512396694215, |
|
"eval_recall_macro": 0.5513832199546485, |
|
"eval_recall_micro": 0.6212121212121212, |
|
"eval_recall_weighted": 0.6212121212121212, |
|
"eval_runtime": 2.3247, |
|
"eval_samples_per_second": 56.78, |
|
"eval_steps_per_second": 7.313, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 11.03448275862069, |
|
"grad_norm": 3.1624715328216553, |
|
"learning_rate": 4.980842911877395e-05, |
|
"loss": 0.5936, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 11.10344827586207, |
|
"grad_norm": 3.2586793899536133, |
|
"learning_rate": 4.9425287356321845e-05, |
|
"loss": 0.7724, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 11.172413793103448, |
|
"grad_norm": 4.481448173522949, |
|
"learning_rate": 4.904214559386973e-05, |
|
"loss": 0.6748, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 11.241379310344827, |
|
"grad_norm": 2.7534396648406982, |
|
"learning_rate": 4.865900383141763e-05, |
|
"loss": 0.5199, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 11.310344827586206, |
|
"grad_norm": 2.4173717498779297, |
|
"learning_rate": 4.827586206896552e-05, |
|
"loss": 0.8016, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 11.379310344827585, |
|
"grad_norm": 2.85591459274292, |
|
"learning_rate": 4.789272030651341e-05, |
|
"loss": 0.8319, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 11.448275862068966, |
|
"grad_norm": 5.3054585456848145, |
|
"learning_rate": 4.7509578544061307e-05, |
|
"loss": 1.0745, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 11.517241379310345, |
|
"grad_norm": 3.0006418228149414, |
|
"learning_rate": 4.7126436781609195e-05, |
|
"loss": 0.5067, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 11.586206896551724, |
|
"grad_norm": 3.4974889755249023, |
|
"learning_rate": 4.674329501915709e-05, |
|
"loss": 0.7971, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 11.655172413793103, |
|
"grad_norm": 3.117962598800659, |
|
"learning_rate": 4.6360153256704985e-05, |
|
"loss": 0.6665, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 11.724137931034482, |
|
"grad_norm": 4.9846343994140625, |
|
"learning_rate": 4.597701149425287e-05, |
|
"loss": 0.7402, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 11.793103448275861, |
|
"grad_norm": 6.1209001541137695, |
|
"learning_rate": 4.559386973180077e-05, |
|
"loss": 0.9042, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 11.862068965517242, |
|
"grad_norm": 2.3766987323760986, |
|
"learning_rate": 4.5210727969348656e-05, |
|
"loss": 0.6191, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 11.931034482758621, |
|
"grad_norm": 5.1854143142700195, |
|
"learning_rate": 4.482758620689655e-05, |
|
"loss": 0.6116, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"grad_norm": 5.495875835418701, |
|
"learning_rate": 4.4444444444444447e-05, |
|
"loss": 0.7902, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_accuracy": 0.6212121212121212, |
|
"eval_f1_macro": 0.5396961800902688, |
|
"eval_f1_micro": 0.6212121212121212, |
|
"eval_f1_weighted": 0.6140481105214648, |
|
"eval_loss": 1.1624207496643066, |
|
"eval_precision_macro": 0.5422025161353321, |
|
"eval_precision_micro": 0.6212121212121212, |
|
"eval_precision_weighted": 0.6208899459169562, |
|
"eval_recall_macro": 0.5504686318972033, |
|
"eval_recall_micro": 0.6212121212121212, |
|
"eval_recall_weighted": 0.6212121212121212, |
|
"eval_runtime": 2.3261, |
|
"eval_samples_per_second": 56.747, |
|
"eval_steps_per_second": 7.308, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 12.068965517241379, |
|
"grad_norm": 4.319167613983154, |
|
"learning_rate": 4.406130268199234e-05, |
|
"loss": 0.7979, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 12.137931034482758, |
|
"grad_norm": 3.6471261978149414, |
|
"learning_rate": 4.367816091954024e-05, |
|
"loss": 0.7733, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 12.206896551724139, |
|
"grad_norm": 3.5435140132904053, |
|
"learning_rate": 4.3295019157088125e-05, |
|
"loss": 0.8753, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 12.275862068965518, |
|
"grad_norm": 3.007453441619873, |
|
"learning_rate": 4.291187739463602e-05, |
|
"loss": 0.5425, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 12.344827586206897, |
|
"grad_norm": 4.487671852111816, |
|
"learning_rate": 4.252873563218391e-05, |
|
"loss": 0.5562, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 12.413793103448276, |
|
"grad_norm": 3.367605447769165, |
|
"learning_rate": 4.21455938697318e-05, |
|
"loss": 0.625, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 12.482758620689655, |
|
"grad_norm": 5.727989673614502, |
|
"learning_rate": 4.17624521072797e-05, |
|
"loss": 0.6552, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 12.551724137931034, |
|
"grad_norm": 2.801624298095703, |
|
"learning_rate": 4.1379310344827587e-05, |
|
"loss": 0.5104, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 12.620689655172415, |
|
"grad_norm": 4.306697368621826, |
|
"learning_rate": 4.099616858237548e-05, |
|
"loss": 0.6764, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 12.689655172413794, |
|
"grad_norm": 7.026360988616943, |
|
"learning_rate": 4.061302681992337e-05, |
|
"loss": 0.7224, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 12.758620689655173, |
|
"grad_norm": 5.398407936096191, |
|
"learning_rate": 4.0229885057471265e-05, |
|
"loss": 0.8992, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 12.827586206896552, |
|
"grad_norm": 2.986201524734497, |
|
"learning_rate": 3.984674329501916e-05, |
|
"loss": 0.5226, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 12.89655172413793, |
|
"grad_norm": 5.425571918487549, |
|
"learning_rate": 3.9463601532567055e-05, |
|
"loss": 0.6649, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 12.96551724137931, |
|
"grad_norm": 4.2015461921691895, |
|
"learning_rate": 3.908045977011495e-05, |
|
"loss": 0.7503, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_accuracy": 0.6287878787878788, |
|
"eval_f1_macro": 0.5376898842937335, |
|
"eval_f1_micro": 0.6287878787878788, |
|
"eval_f1_weighted": 0.6125773067483171, |
|
"eval_loss": 1.135877013206482, |
|
"eval_precision_macro": 0.5471859541424758, |
|
"eval_precision_micro": 0.6287878787878788, |
|
"eval_precision_weighted": 0.6142989641013356, |
|
"eval_recall_macro": 0.5455328798185941, |
|
"eval_recall_micro": 0.6287878787878788, |
|
"eval_recall_weighted": 0.6287878787878788, |
|
"eval_runtime": 2.3097, |
|
"eval_samples_per_second": 57.149, |
|
"eval_steps_per_second": 7.36, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 13.03448275862069, |
|
"grad_norm": 3.527031660079956, |
|
"learning_rate": 3.869731800766284e-05, |
|
"loss": 0.6196, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 13.10344827586207, |
|
"grad_norm": 4.090692520141602, |
|
"learning_rate": 3.831417624521073e-05, |
|
"loss": 0.4906, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 13.172413793103448, |
|
"grad_norm": 3.5445234775543213, |
|
"learning_rate": 3.793103448275862e-05, |
|
"loss": 0.6832, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 13.241379310344827, |
|
"grad_norm": 2.648242473602295, |
|
"learning_rate": 3.7547892720306517e-05, |
|
"loss": 0.7041, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 13.310344827586206, |
|
"grad_norm": 2.421206474304199, |
|
"learning_rate": 3.716475095785441e-05, |
|
"loss": 0.3506, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 13.379310344827585, |
|
"grad_norm": 4.197497367858887, |
|
"learning_rate": 3.67816091954023e-05, |
|
"loss": 0.7564, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 13.448275862068966, |
|
"grad_norm": 4.1512131690979, |
|
"learning_rate": 3.6398467432950195e-05, |
|
"loss": 0.7134, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 13.517241379310345, |
|
"grad_norm": 2.6926424503326416, |
|
"learning_rate": 3.601532567049808e-05, |
|
"loss": 0.5195, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 13.586206896551724, |
|
"grad_norm": 5.174130916595459, |
|
"learning_rate": 3.563218390804598e-05, |
|
"loss": 0.485, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 13.655172413793103, |
|
"grad_norm": 2.1289005279541016, |
|
"learning_rate": 3.5249042145593867e-05, |
|
"loss": 0.6963, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 13.724137931034482, |
|
"grad_norm": 5.4137654304504395, |
|
"learning_rate": 3.486590038314176e-05, |
|
"loss": 0.6374, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 13.793103448275861, |
|
"grad_norm": 3.6843011379241943, |
|
"learning_rate": 3.4482758620689657e-05, |
|
"loss": 0.9498, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 13.862068965517242, |
|
"grad_norm": 4.690808296203613, |
|
"learning_rate": 3.409961685823755e-05, |
|
"loss": 0.7182, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 13.931034482758621, |
|
"grad_norm": 4.772737979888916, |
|
"learning_rate": 3.371647509578545e-05, |
|
"loss": 0.7981, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"grad_norm": 6.463994979858398, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 0.586, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_accuracy": 0.6287878787878788, |
|
"eval_f1_macro": 0.5440812816073651, |
|
"eval_f1_micro": 0.6287878787878788, |
|
"eval_f1_weighted": 0.614080662953901, |
|
"eval_loss": 1.151163101196289, |
|
"eval_precision_macro": 0.5361375661375661, |
|
"eval_precision_micro": 0.6287878787878788, |
|
"eval_precision_weighted": 0.6032631874298542, |
|
"eval_recall_macro": 0.5557218442932729, |
|
"eval_recall_micro": 0.6287878787878788, |
|
"eval_recall_weighted": 0.6287878787878788, |
|
"eval_runtime": 2.3161, |
|
"eval_samples_per_second": 56.992, |
|
"eval_steps_per_second": 7.34, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 14.068965517241379, |
|
"grad_norm": 4.363000392913818, |
|
"learning_rate": 3.295019157088123e-05, |
|
"loss": 0.5204, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 14.137931034482758, |
|
"grad_norm": 2.4504146575927734, |
|
"learning_rate": 3.256704980842912e-05, |
|
"loss": 0.3989, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 14.206896551724139, |
|
"grad_norm": 2.4213919639587402, |
|
"learning_rate": 3.218390804597701e-05, |
|
"loss": 0.4246, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 14.275862068965518, |
|
"grad_norm": 4.8299031257629395, |
|
"learning_rate": 3.180076628352491e-05, |
|
"loss": 0.6999, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 14.344827586206897, |
|
"grad_norm": 2.637255907058716, |
|
"learning_rate": 3.1417624521072797e-05, |
|
"loss": 0.5376, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 14.413793103448276, |
|
"grad_norm": 3.4667327404022217, |
|
"learning_rate": 3.103448275862069e-05, |
|
"loss": 0.5109, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 14.482758620689655, |
|
"grad_norm": 6.673337459564209, |
|
"learning_rate": 3.065134099616858e-05, |
|
"loss": 0.4014, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 14.551724137931034, |
|
"grad_norm": 2.5311481952667236, |
|
"learning_rate": 3.0268199233716475e-05, |
|
"loss": 0.6741, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 14.620689655172415, |
|
"grad_norm": 3.339940071105957, |
|
"learning_rate": 2.988505747126437e-05, |
|
"loss": 0.4081, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 14.689655172413794, |
|
"grad_norm": 2.3433589935302734, |
|
"learning_rate": 2.950191570881226e-05, |
|
"loss": 0.5574, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 14.758620689655173, |
|
"grad_norm": 3.5291073322296143, |
|
"learning_rate": 2.9118773946360157e-05, |
|
"loss": 0.6136, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 14.827586206896552, |
|
"grad_norm": 3.0121891498565674, |
|
"learning_rate": 2.8735632183908045e-05, |
|
"loss": 0.3841, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 14.89655172413793, |
|
"grad_norm": 3.474233388900757, |
|
"learning_rate": 2.835249042145594e-05, |
|
"loss": 0.4371, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 14.96551724137931, |
|
"grad_norm": 3.466447591781616, |
|
"learning_rate": 2.796934865900383e-05, |
|
"loss": 0.6869, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_accuracy": 0.6287878787878788, |
|
"eval_f1_macro": 0.5323304385684055, |
|
"eval_f1_micro": 0.6287878787878788, |
|
"eval_f1_weighted": 0.6116664108577855, |
|
"eval_loss": 1.13055419921875, |
|
"eval_precision_macro": 0.5269524536765916, |
|
"eval_precision_micro": 0.6287878787878788, |
|
"eval_precision_weighted": 0.6043089290542268, |
|
"eval_recall_macro": 0.547490551776266, |
|
"eval_recall_micro": 0.6287878787878788, |
|
"eval_recall_weighted": 0.6287878787878788, |
|
"eval_runtime": 2.3194, |
|
"eval_samples_per_second": 56.911, |
|
"eval_steps_per_second": 7.329, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 15.03448275862069, |
|
"grad_norm": 2.4587948322296143, |
|
"learning_rate": 2.7586206896551727e-05, |
|
"loss": 0.6758, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 15.10344827586207, |
|
"grad_norm": 3.373725414276123, |
|
"learning_rate": 2.720306513409962e-05, |
|
"loss": 0.516, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 15.172413793103448, |
|
"grad_norm": 4.587845802307129, |
|
"learning_rate": 2.681992337164751e-05, |
|
"loss": 0.5093, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 15.241379310344827, |
|
"grad_norm": 6.124247074127197, |
|
"learning_rate": 2.6436781609195405e-05, |
|
"loss": 0.5237, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 15.310344827586206, |
|
"grad_norm": 3.777029514312744, |
|
"learning_rate": 2.6053639846743293e-05, |
|
"loss": 0.5128, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 15.379310344827585, |
|
"grad_norm": 2.960836887359619, |
|
"learning_rate": 2.5670498084291188e-05, |
|
"loss": 0.6366, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 15.448275862068966, |
|
"grad_norm": 3.1089887619018555, |
|
"learning_rate": 2.5287356321839083e-05, |
|
"loss": 0.5369, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 15.517241379310345, |
|
"grad_norm": 2.6847426891326904, |
|
"learning_rate": 2.4904214559386975e-05, |
|
"loss": 0.3325, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 15.586206896551724, |
|
"grad_norm": 6.938262462615967, |
|
"learning_rate": 2.4521072796934867e-05, |
|
"loss": 0.6207, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 15.655172413793103, |
|
"grad_norm": 1.7349634170532227, |
|
"learning_rate": 2.413793103448276e-05, |
|
"loss": 0.3663, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 15.724137931034482, |
|
"grad_norm": 1.5232642889022827, |
|
"learning_rate": 2.3754789272030653e-05, |
|
"loss": 0.5473, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 15.793103448275861, |
|
"grad_norm": 4.805527210235596, |
|
"learning_rate": 2.3371647509578545e-05, |
|
"loss": 0.3652, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 15.862068965517242, |
|
"grad_norm": 4.195865631103516, |
|
"learning_rate": 2.2988505747126437e-05, |
|
"loss": 0.5624, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 15.931034482758621, |
|
"grad_norm": 1.5092368125915527, |
|
"learning_rate": 2.2605363984674328e-05, |
|
"loss": 0.4081, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"grad_norm": 3.367891311645508, |
|
"learning_rate": 2.2222222222222223e-05, |
|
"loss": 0.5498, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_accuracy": 0.6287878787878788, |
|
"eval_f1_macro": 0.5372783597671154, |
|
"eval_f1_micro": 0.6287878787878788, |
|
"eval_f1_weighted": 0.6117070792278643, |
|
"eval_loss": 1.1293154954910278, |
|
"eval_precision_macro": 0.5353323147440795, |
|
"eval_precision_micro": 0.6287878787878788, |
|
"eval_precision_weighted": 0.603853373305245, |
|
"eval_recall_macro": 0.5471201814058957, |
|
"eval_recall_micro": 0.6287878787878788, |
|
"eval_recall_weighted": 0.6287878787878788, |
|
"eval_runtime": 2.3165, |
|
"eval_samples_per_second": 56.983, |
|
"eval_steps_per_second": 7.339, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 16.06896551724138, |
|
"grad_norm": 3.580066442489624, |
|
"learning_rate": 2.183908045977012e-05, |
|
"loss": 0.4746, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 16.137931034482758, |
|
"grad_norm": 3.583418130874634, |
|
"learning_rate": 2.145593869731801e-05, |
|
"loss": 0.4795, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 16.20689655172414, |
|
"grad_norm": 2.073596239089966, |
|
"learning_rate": 2.10727969348659e-05, |
|
"loss": 0.3328, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 16.275862068965516, |
|
"grad_norm": 3.2745907306671143, |
|
"learning_rate": 2.0689655172413793e-05, |
|
"loss": 0.4395, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 16.344827586206897, |
|
"grad_norm": 1.9064489603042603, |
|
"learning_rate": 2.0306513409961685e-05, |
|
"loss": 0.3172, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 16.413793103448278, |
|
"grad_norm": 4.838860988616943, |
|
"learning_rate": 1.992337164750958e-05, |
|
"loss": 0.6758, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 16.482758620689655, |
|
"grad_norm": 1.598775029182434, |
|
"learning_rate": 1.9540229885057475e-05, |
|
"loss": 0.3767, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 16.551724137931036, |
|
"grad_norm": 4.009886264801025, |
|
"learning_rate": 1.9157088122605367e-05, |
|
"loss": 0.4511, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 16.620689655172413, |
|
"grad_norm": 3.5439836978912354, |
|
"learning_rate": 1.8773946360153258e-05, |
|
"loss": 0.4542, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 16.689655172413794, |
|
"grad_norm": 3.5532007217407227, |
|
"learning_rate": 1.839080459770115e-05, |
|
"loss": 0.5007, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 16.75862068965517, |
|
"grad_norm": 2.5344693660736084, |
|
"learning_rate": 1.800766283524904e-05, |
|
"loss": 0.3181, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 16.82758620689655, |
|
"grad_norm": 1.9428218603134155, |
|
"learning_rate": 1.7624521072796933e-05, |
|
"loss": 0.4047, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 16.896551724137932, |
|
"grad_norm": 2.117885112762451, |
|
"learning_rate": 1.7241379310344828e-05, |
|
"loss": 0.3487, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 16.96551724137931, |
|
"grad_norm": 3.377067804336548, |
|
"learning_rate": 1.6858237547892723e-05, |
|
"loss": 0.5037, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_accuracy": 0.6212121212121212, |
|
"eval_f1_macro": 0.5290491049649699, |
|
"eval_f1_micro": 0.6212121212121212, |
|
"eval_f1_weighted": 0.6005376615693997, |
|
"eval_loss": 1.163466453552246, |
|
"eval_precision_macro": 0.5374433106575963, |
|
"eval_precision_micro": 0.6212121212121212, |
|
"eval_precision_weighted": 0.6021569865319866, |
|
"eval_recall_macro": 0.5397581254724112, |
|
"eval_recall_micro": 0.6212121212121212, |
|
"eval_recall_weighted": 0.6212121212121212, |
|
"eval_runtime": 2.3304, |
|
"eval_samples_per_second": 56.642, |
|
"eval_steps_per_second": 7.295, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 17.03448275862069, |
|
"grad_norm": 3.5657002925872803, |
|
"learning_rate": 1.6475095785440615e-05, |
|
"loss": 0.4446, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 17.103448275862068, |
|
"grad_norm": 4.189113616943359, |
|
"learning_rate": 1.6091954022988507e-05, |
|
"loss": 0.3923, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 17.17241379310345, |
|
"grad_norm": 2.554723024368286, |
|
"learning_rate": 1.5708812260536398e-05, |
|
"loss": 0.358, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 17.24137931034483, |
|
"grad_norm": 3.3316357135772705, |
|
"learning_rate": 1.532567049808429e-05, |
|
"loss": 0.2683, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 17.310344827586206, |
|
"grad_norm": 3.2318475246429443, |
|
"learning_rate": 1.4942528735632185e-05, |
|
"loss": 0.3611, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 17.379310344827587, |
|
"grad_norm": 3.2837979793548584, |
|
"learning_rate": 1.4559386973180078e-05, |
|
"loss": 0.435, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 17.448275862068964, |
|
"grad_norm": 3.8966495990753174, |
|
"learning_rate": 1.417624521072797e-05, |
|
"loss": 0.2982, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 17.517241379310345, |
|
"grad_norm": 6.666224479675293, |
|
"learning_rate": 1.3793103448275863e-05, |
|
"loss": 0.4624, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 17.586206896551722, |
|
"grad_norm": 4.43505334854126, |
|
"learning_rate": 1.3409961685823755e-05, |
|
"loss": 0.6202, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 17.655172413793103, |
|
"grad_norm": 3.7163617610931396, |
|
"learning_rate": 1.3026819923371647e-05, |
|
"loss": 0.3422, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 17.724137931034484, |
|
"grad_norm": 2.855036735534668, |
|
"learning_rate": 1.2643678160919542e-05, |
|
"loss": 0.5171, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 17.79310344827586, |
|
"grad_norm": 3.121839761734009, |
|
"learning_rate": 1.2260536398467433e-05, |
|
"loss": 0.3575, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 17.862068965517242, |
|
"grad_norm": 2.77312970161438, |
|
"learning_rate": 1.1877394636015327e-05, |
|
"loss": 0.3051, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 17.93103448275862, |
|
"grad_norm": 2.2301366329193115, |
|
"learning_rate": 1.1494252873563218e-05, |
|
"loss": 0.4635, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"grad_norm": 4.1137518882751465, |
|
"learning_rate": 1.1111111111111112e-05, |
|
"loss": 0.3624, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_accuracy": 0.6590909090909091, |
|
"eval_f1_macro": 0.5699944413104846, |
|
"eval_f1_micro": 0.6590909090909091, |
|
"eval_f1_weighted": 0.6414064024247821, |
|
"eval_loss": 1.09944748878479, |
|
"eval_precision_macro": 0.5815085815085814, |
|
"eval_precision_micro": 0.6590909090909091, |
|
"eval_precision_weighted": 0.640854299945209, |
|
"eval_recall_macro": 0.5742857142857142, |
|
"eval_recall_micro": 0.6590909090909091, |
|
"eval_recall_weighted": 0.6590909090909091, |
|
"eval_runtime": 2.3246, |
|
"eval_samples_per_second": 56.785, |
|
"eval_steps_per_second": 7.313, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 18.06896551724138, |
|
"grad_norm": 5.545546054840088, |
|
"learning_rate": 1.0727969348659005e-05, |
|
"loss": 0.4351, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 18.137931034482758, |
|
"grad_norm": 2.5802762508392334, |
|
"learning_rate": 1.0344827586206897e-05, |
|
"loss": 0.3958, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 18.20689655172414, |
|
"grad_norm": 3.738755702972412, |
|
"learning_rate": 9.96168582375479e-06, |
|
"loss": 0.3933, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 18.275862068965516, |
|
"grad_norm": 2.6017744541168213, |
|
"learning_rate": 9.578544061302683e-06, |
|
"loss": 0.4367, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 18.344827586206897, |
|
"grad_norm": 4.551992416381836, |
|
"learning_rate": 9.195402298850575e-06, |
|
"loss": 0.3844, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 18.413793103448278, |
|
"grad_norm": 3.3013875484466553, |
|
"learning_rate": 8.812260536398467e-06, |
|
"loss": 0.3592, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 18.482758620689655, |
|
"grad_norm": 8.250885009765625, |
|
"learning_rate": 8.429118773946362e-06, |
|
"loss": 0.5747, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 18.551724137931036, |
|
"grad_norm": 2.1234164237976074, |
|
"learning_rate": 8.045977011494253e-06, |
|
"loss": 0.3561, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 18.620689655172413, |
|
"grad_norm": 2.730742931365967, |
|
"learning_rate": 7.662835249042145e-06, |
|
"loss": 0.3182, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 18.689655172413794, |
|
"grad_norm": 1.1862895488739014, |
|
"learning_rate": 7.279693486590039e-06, |
|
"loss": 0.2739, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 18.75862068965517, |
|
"grad_norm": 3.530163049697876, |
|
"learning_rate": 6.896551724137932e-06, |
|
"loss": 0.4006, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 18.82758620689655, |
|
"grad_norm": 3.7491860389709473, |
|
"learning_rate": 6.513409961685823e-06, |
|
"loss": 0.2657, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 18.896551724137932, |
|
"grad_norm": 6.904880523681641, |
|
"learning_rate": 6.130268199233717e-06, |
|
"loss": 0.4191, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 18.96551724137931, |
|
"grad_norm": 1.2763559818267822, |
|
"learning_rate": 5.747126436781609e-06, |
|
"loss": 0.3387, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_accuracy": 0.6515151515151515, |
|
"eval_f1_macro": 0.5642594468235355, |
|
"eval_f1_micro": 0.6515151515151515, |
|
"eval_f1_weighted": 0.6367337845877334, |
|
"eval_loss": 1.094382643699646, |
|
"eval_precision_macro": 0.5555528431468281, |
|
"eval_precision_micro": 0.6515151515151515, |
|
"eval_precision_weighted": 0.6268442283394436, |
|
"eval_recall_macro": 0.5781405895691609, |
|
"eval_recall_micro": 0.6515151515151515, |
|
"eval_recall_weighted": 0.6515151515151515, |
|
"eval_runtime": 2.3159, |
|
"eval_samples_per_second": 56.998, |
|
"eval_steps_per_second": 7.341, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 19.03448275862069, |
|
"grad_norm": 3.0644407272338867, |
|
"learning_rate": 5.3639846743295025e-06, |
|
"loss": 0.3165, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 19.103448275862068, |
|
"grad_norm": 3.8234410285949707, |
|
"learning_rate": 4.980842911877395e-06, |
|
"loss": 0.4086, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 19.17241379310345, |
|
"grad_norm": 1.5022215843200684, |
|
"learning_rate": 4.5977011494252875e-06, |
|
"loss": 0.2806, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 19.24137931034483, |
|
"grad_norm": 4.573976039886475, |
|
"learning_rate": 4.214559386973181e-06, |
|
"loss": 0.4335, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 19.310344827586206, |
|
"grad_norm": 3.7575619220733643, |
|
"learning_rate": 3.8314176245210725e-06, |
|
"loss": 0.2748, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 19.379310344827587, |
|
"grad_norm": 1.8237051963806152, |
|
"learning_rate": 3.448275862068966e-06, |
|
"loss": 0.3879, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 19.448275862068964, |
|
"grad_norm": 5.58415412902832, |
|
"learning_rate": 3.0651340996168583e-06, |
|
"loss": 0.2736, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 19.517241379310345, |
|
"grad_norm": 1.1507081985473633, |
|
"learning_rate": 2.6819923371647512e-06, |
|
"loss": 0.2612, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 19.586206896551722, |
|
"grad_norm": 1.060652494430542, |
|
"learning_rate": 2.2988505747126437e-06, |
|
"loss": 0.2706, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 19.655172413793103, |
|
"grad_norm": 1.4286646842956543, |
|
"learning_rate": 1.9157088122605362e-06, |
|
"loss": 0.2616, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 19.724137931034484, |
|
"grad_norm": 5.675417900085449, |
|
"learning_rate": 1.5325670498084292e-06, |
|
"loss": 0.4929, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 19.79310344827586, |
|
"grad_norm": 1.839812994003296, |
|
"learning_rate": 1.1494252873563219e-06, |
|
"loss": 0.2418, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 19.862068965517242, |
|
"grad_norm": 3.5628745555877686, |
|
"learning_rate": 7.662835249042146e-07, |
|
"loss": 0.4503, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 19.93103448275862, |
|
"grad_norm": 2.046889543533325, |
|
"learning_rate": 3.831417624521073e-07, |
|
"loss": 0.2689, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"grad_norm": 1.0142828226089478, |
|
"learning_rate": 0.0, |
|
"loss": 0.4052, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_accuracy": 0.6590909090909091, |
|
"eval_f1_macro": 0.5683267149224223, |
|
"eval_f1_micro": 0.6590909090909091, |
|
"eval_f1_weighted": 0.6432058978241244, |
|
"eval_loss": 1.0934416055679321, |
|
"eval_precision_macro": 0.5681032097512798, |
|
"eval_precision_micro": 0.6590909090909091, |
|
"eval_precision_weighted": 0.6393054520079716, |
|
"eval_recall_macro": 0.5797581254724111, |
|
"eval_recall_micro": 0.6590909090909091, |
|
"eval_recall_weighted": 0.6590909090909091, |
|
"eval_runtime": 2.3382, |
|
"eval_samples_per_second": 56.454, |
|
"eval_steps_per_second": 7.271, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"step": 580, |
|
"total_flos": 7.160580716109005e+17, |
|
"train_loss": 1.0144407548781098, |
|
"train_runtime": 564.1231, |
|
"train_samples_per_second": 16.379, |
|
"train_steps_per_second": 1.028 |
|
} |
|
], |
|
"logging_steps": 2, |
|
"max_steps": 580, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 20, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 7.160580716109005e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|