{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 995, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0020100502512562816, "grad_norm": 19.729604721069336, "learning_rate": 2.0000000000000003e-06, "loss": 6.6152, "step": 2 }, { "epoch": 0.004020100502512563, "grad_norm": 16.32933807373047, "learning_rate": 4.000000000000001e-06, "loss": 6.684, "step": 4 }, { "epoch": 0.006030150753768844, "grad_norm": 9.964282989501953, "learning_rate": 6e-06, "loss": 6.4803, "step": 6 }, { "epoch": 0.008040201005025126, "grad_norm": 9.893184661865234, "learning_rate": 8.000000000000001e-06, "loss": 5.9697, "step": 8 }, { "epoch": 0.010050251256281407, "grad_norm": 7.443181037902832, "learning_rate": 1e-05, "loss": 5.7487, "step": 10 }, { "epoch": 0.012060301507537688, "grad_norm": 5.485438346862793, "learning_rate": 1.2e-05, "loss": 5.6723, "step": 12 }, { "epoch": 0.01407035175879397, "grad_norm": 4.975802421569824, "learning_rate": 1.4000000000000001e-05, "loss": 5.4903, "step": 14 }, { "epoch": 0.016080402010050253, "grad_norm": 5.055726051330566, "learning_rate": 1.6000000000000003e-05, "loss": 5.2516, "step": 16 }, { "epoch": 0.018090452261306532, "grad_norm": 4.720687389373779, "learning_rate": 1.8e-05, "loss": 5.0191, "step": 18 }, { "epoch": 0.020100502512562814, "grad_norm": 3.445399761199951, "learning_rate": 2e-05, "loss": 4.8732, "step": 20 }, { "epoch": 0.022110552763819097, "grad_norm": 3.533928155899048, "learning_rate": 2.2000000000000003e-05, "loss": 5.1519, "step": 22 }, { "epoch": 0.024120603015075376, "grad_norm": 3.4936132431030273, "learning_rate": 2.4e-05, "loss": 4.9552, "step": 24 }, { "epoch": 0.02613065326633166, "grad_norm": 3.037400722503662, "learning_rate": 2.6000000000000002e-05, "loss": 4.622, "step": 26 }, { "epoch": 0.02814070351758794, "grad_norm": 4.790655136108398, "learning_rate": 2.8000000000000003e-05, "loss": 4.5671, "step": 28 }, { "epoch": 0.03015075376884422, "grad_norm": 6.313861846923828, "learning_rate": 3e-05, "loss": 4.3285, "step": 30 }, { "epoch": 0.032160804020100506, "grad_norm": 2.662900447845459, "learning_rate": 3.2000000000000005e-05, "loss": 4.2636, "step": 32 }, { "epoch": 0.034170854271356785, "grad_norm": 2.5860118865966797, "learning_rate": 3.4000000000000007e-05, "loss": 4.3172, "step": 34 }, { "epoch": 0.036180904522613064, "grad_norm": 2.960880994796753, "learning_rate": 3.6e-05, "loss": 3.9687, "step": 36 }, { "epoch": 0.03819095477386935, "grad_norm": 3.4453978538513184, "learning_rate": 3.8e-05, "loss": 3.8505, "step": 38 }, { "epoch": 0.04020100502512563, "grad_norm": 2.6721041202545166, "learning_rate": 4e-05, "loss": 3.971, "step": 40 }, { "epoch": 0.04221105527638191, "grad_norm": 2.6484851837158203, "learning_rate": 4.2e-05, "loss": 3.8536, "step": 42 }, { "epoch": 0.044221105527638194, "grad_norm": 6.354692459106445, "learning_rate": 4.4000000000000006e-05, "loss": 3.8466, "step": 44 }, { "epoch": 0.04623115577889447, "grad_norm": 2.6469643115997314, "learning_rate": 4.600000000000001e-05, "loss": 3.5942, "step": 46 }, { "epoch": 0.04824120603015075, "grad_norm": 2.7454721927642822, "learning_rate": 4.8e-05, "loss": 3.526, "step": 48 }, { "epoch": 0.05025125628140704, "grad_norm": 2.445122480392456, "learning_rate": 5e-05, "loss": 3.5655, "step": 50 }, { "epoch": 0.05226130653266332, "grad_norm": 2.458037853240967, "learning_rate": 5.2000000000000004e-05, "loss": 3.3741, "step": 52 }, { "epoch": 0.054271356783919596, "grad_norm": 2.469712972640991, "learning_rate": 5.4000000000000005e-05, "loss": 3.4174, "step": 54 }, { "epoch": 0.05628140703517588, "grad_norm": 2.475627899169922, "learning_rate": 5.6000000000000006e-05, "loss": 3.3984, "step": 56 }, { "epoch": 0.05829145728643216, "grad_norm": 2.6965248584747314, "learning_rate": 5.8e-05, "loss": 3.5654, "step": 58 }, { "epoch": 0.06030150753768844, "grad_norm": 2.3575079441070557, "learning_rate": 6e-05, "loss": 3.2454, "step": 60 }, { "epoch": 0.062311557788944726, "grad_norm": 2.371737480163574, "learning_rate": 6.2e-05, "loss": 3.361, "step": 62 }, { "epoch": 0.06432160804020101, "grad_norm": 2.144615650177002, "learning_rate": 6.400000000000001e-05, "loss": 3.2241, "step": 64 }, { "epoch": 0.06633165829145729, "grad_norm": 1.9614999294281006, "learning_rate": 6.6e-05, "loss": 3.1206, "step": 66 }, { "epoch": 0.06834170854271357, "grad_norm": 2.0626800060272217, "learning_rate": 6.800000000000001e-05, "loss": 3.3899, "step": 68 }, { "epoch": 0.07035175879396985, "grad_norm": 2.669525623321533, "learning_rate": 7e-05, "loss": 3.2405, "step": 70 }, { "epoch": 0.07236180904522613, "grad_norm": 2.620374917984009, "learning_rate": 7.2e-05, "loss": 3.3299, "step": 72 }, { "epoch": 0.0743718592964824, "grad_norm": 2.2318763732910156, "learning_rate": 7.4e-05, "loss": 2.9798, "step": 74 }, { "epoch": 0.0763819095477387, "grad_norm": 2.28938889503479, "learning_rate": 7.6e-05, "loss": 3.2799, "step": 76 }, { "epoch": 0.07839195979899498, "grad_norm": 2.176002264022827, "learning_rate": 7.800000000000001e-05, "loss": 3.2466, "step": 78 }, { "epoch": 0.08040201005025126, "grad_norm": 2.2421767711639404, "learning_rate": 8e-05, "loss": 3.0629, "step": 80 }, { "epoch": 0.08241206030150754, "grad_norm": 2.7618038654327393, "learning_rate": 8.2e-05, "loss": 3.2719, "step": 82 }, { "epoch": 0.08442211055276382, "grad_norm": 2.1031386852264404, "learning_rate": 8.4e-05, "loss": 3.0094, "step": 84 }, { "epoch": 0.0864321608040201, "grad_norm": 2.1169002056121826, "learning_rate": 8.6e-05, "loss": 3.021, "step": 86 }, { "epoch": 0.08844221105527639, "grad_norm": 2.4154305458068848, "learning_rate": 8.800000000000001e-05, "loss": 3.0685, "step": 88 }, { "epoch": 0.09045226130653267, "grad_norm": 2.4994680881500244, "learning_rate": 9e-05, "loss": 2.8943, "step": 90 }, { "epoch": 0.09246231155778895, "grad_norm": 3.4030494689941406, "learning_rate": 9.200000000000001e-05, "loss": 2.8389, "step": 92 }, { "epoch": 0.09447236180904522, "grad_norm": 5.54694128036499, "learning_rate": 9.4e-05, "loss": 3.0157, "step": 94 }, { "epoch": 0.0964824120603015, "grad_norm": 2.363124370574951, "learning_rate": 9.6e-05, "loss": 2.811, "step": 96 }, { "epoch": 0.09849246231155778, "grad_norm": 2.0617117881774902, "learning_rate": 9.8e-05, "loss": 2.8819, "step": 98 }, { "epoch": 0.10050251256281408, "grad_norm": 2.354085922241211, "learning_rate": 0.0001, "loss": 3.0962, "step": 100 }, { "epoch": 0.10251256281407035, "grad_norm": 2.3361897468566895, "learning_rate": 9.999972370327507e-05, "loss": 3.0454, "step": 102 }, { "epoch": 0.10452261306532663, "grad_norm": 2.2245845794677734, "learning_rate": 9.999889481615387e-05, "loss": 3.1027, "step": 104 }, { "epoch": 0.10653266331658291, "grad_norm": 2.069746494293213, "learning_rate": 9.999751334779716e-05, "loss": 2.915, "step": 106 }, { "epoch": 0.10854271356783919, "grad_norm": 2.3541088104248047, "learning_rate": 9.999557931347273e-05, "loss": 2.8328, "step": 108 }, { "epoch": 0.11055276381909548, "grad_norm": 2.2207984924316406, "learning_rate": 9.999309273455528e-05, "loss": 3.0153, "step": 110 }, { "epoch": 0.11256281407035176, "grad_norm": 2.1303670406341553, "learning_rate": 9.999005363852618e-05, "loss": 2.8649, "step": 112 }, { "epoch": 0.11457286432160804, "grad_norm": 2.084247589111328, "learning_rate": 9.998646205897309e-05, "loss": 2.9326, "step": 114 }, { "epoch": 0.11658291457286432, "grad_norm": 2.121018171310425, "learning_rate": 9.998231803558968e-05, "loss": 2.6179, "step": 116 }, { "epoch": 0.1185929648241206, "grad_norm": 2.2490546703338623, "learning_rate": 9.997762161417517e-05, "loss": 3.0462, "step": 118 }, { "epoch": 0.12060301507537688, "grad_norm": 1.9550148248672485, "learning_rate": 9.997237284663379e-05, "loss": 2.8662, "step": 120 }, { "epoch": 0.12261306532663317, "grad_norm": 2.2603437900543213, "learning_rate": 9.996657179097421e-05, "loss": 2.7448, "step": 122 }, { "epoch": 0.12462311557788945, "grad_norm": 1.897772192955017, "learning_rate": 9.996021851130897e-05, "loss": 2.8954, "step": 124 }, { "epoch": 0.12663316582914572, "grad_norm": 2.1712026596069336, "learning_rate": 9.995331307785365e-05, "loss": 2.9001, "step": 126 }, { "epoch": 0.12864321608040202, "grad_norm": 1.9912605285644531, "learning_rate": 9.994585556692624e-05, "loss": 3.0064, "step": 128 }, { "epoch": 0.1306532663316583, "grad_norm": 2.094640016555786, "learning_rate": 9.993784606094612e-05, "loss": 2.6635, "step": 130 }, { "epoch": 0.13266331658291458, "grad_norm": 2.1064651012420654, "learning_rate": 9.992928464843334e-05, "loss": 2.9224, "step": 132 }, { "epoch": 0.13467336683417086, "grad_norm": 1.9782873392105103, "learning_rate": 9.992017142400751e-05, "loss": 2.8202, "step": 134 }, { "epoch": 0.13668341708542714, "grad_norm": 2.189976692199707, "learning_rate": 9.991050648838675e-05, "loss": 2.7919, "step": 136 }, { "epoch": 0.13869346733668342, "grad_norm": 2.0882322788238525, "learning_rate": 9.990028994838673e-05, "loss": 2.868, "step": 138 }, { "epoch": 0.1407035175879397, "grad_norm": 1.8483355045318604, "learning_rate": 9.988952191691925e-05, "loss": 2.903, "step": 140 }, { "epoch": 0.14271356783919598, "grad_norm": 2.1383628845214844, "learning_rate": 9.987820251299122e-05, "loss": 2.6627, "step": 142 }, { "epoch": 0.14472361809045226, "grad_norm": 2.0356757640838623, "learning_rate": 9.986633186170319e-05, "loss": 2.813, "step": 144 }, { "epoch": 0.14673366834170853, "grad_norm": 2.4072749614715576, "learning_rate": 9.985391009424805e-05, "loss": 2.7325, "step": 146 }, { "epoch": 0.1487437185929648, "grad_norm": 1.8525023460388184, "learning_rate": 9.984093734790956e-05, "loss": 2.7092, "step": 148 }, { "epoch": 0.1507537688442211, "grad_norm": 1.8999531269073486, "learning_rate": 9.982741376606078e-05, "loss": 2.8637, "step": 150 }, { "epoch": 0.1527638190954774, "grad_norm": 2.2264699935913086, "learning_rate": 9.981333949816259e-05, "loss": 2.8068, "step": 152 }, { "epoch": 0.15477386934673368, "grad_norm": 2.1256349086761475, "learning_rate": 9.979871469976196e-05, "loss": 2.5258, "step": 154 }, { "epoch": 0.15678391959798996, "grad_norm": 1.8756039142608643, "learning_rate": 9.978353953249022e-05, "loss": 2.7572, "step": 156 }, { "epoch": 0.15879396984924624, "grad_norm": 2.4083592891693115, "learning_rate": 9.976781416406136e-05, "loss": 2.564, "step": 158 }, { "epoch": 0.16080402010050251, "grad_norm": 1.9118754863739014, "learning_rate": 9.975153876827008e-05, "loss": 2.7158, "step": 160 }, { "epoch": 0.1628140703517588, "grad_norm": 2.419102907180786, "learning_rate": 9.973471352498991e-05, "loss": 2.9075, "step": 162 }, { "epoch": 0.16482412060301507, "grad_norm": 2.2566428184509277, "learning_rate": 9.971733862017126e-05, "loss": 2.7698, "step": 164 }, { "epoch": 0.16683417085427135, "grad_norm": 1.980230450630188, "learning_rate": 9.969941424583926e-05, "loss": 2.6454, "step": 166 }, { "epoch": 0.16884422110552763, "grad_norm": 2.0920395851135254, "learning_rate": 9.96809406000918e-05, "loss": 2.8594, "step": 168 }, { "epoch": 0.1708542713567839, "grad_norm": 1.8870232105255127, "learning_rate": 9.966191788709716e-05, "loss": 2.5967, "step": 170 }, { "epoch": 0.1728643216080402, "grad_norm": 2.334998607635498, "learning_rate": 9.964234631709187e-05, "loss": 2.661, "step": 172 }, { "epoch": 0.1748743718592965, "grad_norm": 1.9071625471115112, "learning_rate": 9.962222610637837e-05, "loss": 2.6161, "step": 174 }, { "epoch": 0.17688442211055277, "grad_norm": 2.015760898590088, "learning_rate": 9.960155747732259e-05, "loss": 2.5681, "step": 176 }, { "epoch": 0.17889447236180905, "grad_norm": 1.827253818511963, "learning_rate": 9.958034065835151e-05, "loss": 2.5538, "step": 178 }, { "epoch": 0.18090452261306533, "grad_norm": 1.9388582706451416, "learning_rate": 9.955857588395065e-05, "loss": 2.7098, "step": 180 }, { "epoch": 0.1829145728643216, "grad_norm": 2.1864373683929443, "learning_rate": 9.95362633946614e-05, "loss": 2.5615, "step": 182 }, { "epoch": 0.1849246231155779, "grad_norm": 2.093961477279663, "learning_rate": 9.951340343707852e-05, "loss": 2.7367, "step": 184 }, { "epoch": 0.18693467336683417, "grad_norm": 1.9437264204025269, "learning_rate": 9.948999626384724e-05, "loss": 2.7414, "step": 186 }, { "epoch": 0.18894472361809045, "grad_norm": 1.8001607656478882, "learning_rate": 9.946604213366057e-05, "loss": 2.7595, "step": 188 }, { "epoch": 0.19095477386934673, "grad_norm": 1.8736133575439453, "learning_rate": 9.944154131125642e-05, "loss": 2.2943, "step": 190 }, { "epoch": 0.192964824120603, "grad_norm": 1.8067212104797363, "learning_rate": 9.941649406741469e-05, "loss": 2.6118, "step": 192 }, { "epoch": 0.19497487437185929, "grad_norm": 1.9442073106765747, "learning_rate": 9.939090067895422e-05, "loss": 2.5651, "step": 194 }, { "epoch": 0.19698492462311556, "grad_norm": 1.976509690284729, "learning_rate": 9.936476142872979e-05, "loss": 2.5084, "step": 196 }, { "epoch": 0.19899497487437187, "grad_norm": 1.9604930877685547, "learning_rate": 9.933807660562898e-05, "loss": 2.3689, "step": 198 }, { "epoch": 0.20100502512562815, "grad_norm": 2.081547975540161, "learning_rate": 9.931084650456892e-05, "loss": 2.7038, "step": 200 }, { "epoch": 0.20301507537688443, "grad_norm": 2.1766066551208496, "learning_rate": 9.928307142649316e-05, "loss": 2.8203, "step": 202 }, { "epoch": 0.2050251256281407, "grad_norm": 2.217263698577881, "learning_rate": 9.925475167836821e-05, "loss": 2.5522, "step": 204 }, { "epoch": 0.207035175879397, "grad_norm": 1.8877842426300049, "learning_rate": 9.922588757318021e-05, "loss": 2.6209, "step": 206 }, { "epoch": 0.20904522613065327, "grad_norm": 1.8535610437393188, "learning_rate": 9.919647942993148e-05, "loss": 2.7805, "step": 208 }, { "epoch": 0.21105527638190955, "grad_norm": 1.9754009246826172, "learning_rate": 9.916652757363698e-05, "loss": 2.7363, "step": 210 }, { "epoch": 0.21306532663316582, "grad_norm": 2.0575201511383057, "learning_rate": 9.913603233532067e-05, "loss": 2.5019, "step": 212 }, { "epoch": 0.2150753768844221, "grad_norm": 1.8286898136138916, "learning_rate": 9.910499405201195e-05, "loss": 2.5019, "step": 214 }, { "epoch": 0.21708542713567838, "grad_norm": 1.954950213432312, "learning_rate": 9.907341306674185e-05, "loss": 2.5339, "step": 216 }, { "epoch": 0.21909547738693466, "grad_norm": 1.8942043781280518, "learning_rate": 9.90412897285393e-05, "loss": 2.5065, "step": 218 }, { "epoch": 0.22110552763819097, "grad_norm": 2.165287494659424, "learning_rate": 9.900862439242719e-05, "loss": 2.596, "step": 220 }, { "epoch": 0.22311557788944725, "grad_norm": 1.8702832460403442, "learning_rate": 9.897541741941858e-05, "loss": 2.6656, "step": 222 }, { "epoch": 0.22512562814070353, "grad_norm": 1.745975375175476, "learning_rate": 9.894166917651256e-05, "loss": 2.4078, "step": 224 }, { "epoch": 0.2271356783919598, "grad_norm": 9.129441261291504, "learning_rate": 9.890738003669029e-05, "loss": 2.4112, "step": 226 }, { "epoch": 0.22914572864321608, "grad_norm": 2.1922168731689453, "learning_rate": 9.887255037891086e-05, "loss": 2.5331, "step": 228 }, { "epoch": 0.23115577889447236, "grad_norm": 1.9028476476669312, "learning_rate": 9.883718058810707e-05, "loss": 2.4778, "step": 230 }, { "epoch": 0.23316582914572864, "grad_norm": 1.9489747285842896, "learning_rate": 9.880127105518122e-05, "loss": 2.6777, "step": 232 }, { "epoch": 0.23517587939698492, "grad_norm": 1.922662377357483, "learning_rate": 9.876482217700078e-05, "loss": 2.5236, "step": 234 }, { "epoch": 0.2371859296482412, "grad_norm": 1.9473645687103271, "learning_rate": 9.872783435639397e-05, "loss": 2.4375, "step": 236 }, { "epoch": 0.23919597989949748, "grad_norm": 1.835803747177124, "learning_rate": 9.869030800214532e-05, "loss": 2.6191, "step": 238 }, { "epoch": 0.24120603015075376, "grad_norm": 2.013805627822876, "learning_rate": 9.865224352899119e-05, "loss": 2.5062, "step": 240 }, { "epoch": 0.24321608040201004, "grad_norm": 1.8618221282958984, "learning_rate": 9.861364135761517e-05, "loss": 2.3637, "step": 242 }, { "epoch": 0.24522613065326634, "grad_norm": 2.0696043968200684, "learning_rate": 9.857450191464337e-05, "loss": 2.3941, "step": 244 }, { "epoch": 0.24723618090452262, "grad_norm": 1.835046410560608, "learning_rate": 9.853482563263981e-05, "loss": 2.9119, "step": 246 }, { "epoch": 0.2492462311557789, "grad_norm": 1.9712692499160767, "learning_rate": 9.849461295010156e-05, "loss": 2.4119, "step": 248 }, { "epoch": 0.25125628140703515, "grad_norm": 2.0356907844543457, "learning_rate": 9.84538643114539e-05, "loss": 2.4488, "step": 250 }, { "epoch": 0.25326633165829143, "grad_norm": 1.983216643333435, "learning_rate": 9.841258016704546e-05, "loss": 2.5618, "step": 252 }, { "epoch": 0.2552763819095477, "grad_norm": 1.8130682706832886, "learning_rate": 9.837076097314319e-05, "loss": 2.6361, "step": 254 }, { "epoch": 0.25728643216080405, "grad_norm": 1.7342140674591064, "learning_rate": 9.832840719192736e-05, "loss": 2.3223, "step": 256 }, { "epoch": 0.2592964824120603, "grad_norm": 1.9762102365493774, "learning_rate": 9.82855192914864e-05, "loss": 2.4629, "step": 258 }, { "epoch": 0.2613065326633166, "grad_norm": 1.8401470184326172, "learning_rate": 9.824209774581174e-05, "loss": 2.5557, "step": 260 }, { "epoch": 0.2633165829145729, "grad_norm": 1.8549468517303467, "learning_rate": 9.819814303479267e-05, "loss": 2.6337, "step": 262 }, { "epoch": 0.26532663316582916, "grad_norm": 2.0843794345855713, "learning_rate": 9.815365564421085e-05, "loss": 2.3528, "step": 264 }, { "epoch": 0.26733668341708544, "grad_norm": 1.893712043762207, "learning_rate": 9.810863606573513e-05, "loss": 2.5509, "step": 266 }, { "epoch": 0.2693467336683417, "grad_norm": 2.0922601222991943, "learning_rate": 9.806308479691595e-05, "loss": 2.6126, "step": 268 }, { "epoch": 0.271356783919598, "grad_norm": 1.8790444135665894, "learning_rate": 9.801700234117999e-05, "loss": 2.5569, "step": 270 }, { "epoch": 0.2733668341708543, "grad_norm": 1.8532637357711792, "learning_rate": 9.797038920782454e-05, "loss": 2.4948, "step": 272 }, { "epoch": 0.27537688442211056, "grad_norm": 1.8589162826538086, "learning_rate": 9.792324591201179e-05, "loss": 2.4739, "step": 274 }, { "epoch": 0.27738693467336684, "grad_norm": 1.8177707195281982, "learning_rate": 9.78755729747633e-05, "loss": 2.4883, "step": 276 }, { "epoch": 0.2793969849246231, "grad_norm": 1.9046951532363892, "learning_rate": 9.782737092295413e-05, "loss": 2.331, "step": 278 }, { "epoch": 0.2814070351758794, "grad_norm": 1.8969510793685913, "learning_rate": 9.777864028930705e-05, "loss": 2.4818, "step": 280 }, { "epoch": 0.2834170854271357, "grad_norm": 2.1642754077911377, "learning_rate": 9.77293816123866e-05, "loss": 2.3212, "step": 282 }, { "epoch": 0.28542713567839195, "grad_norm": 1.7480413913726807, "learning_rate": 9.767959543659326e-05, "loss": 2.5252, "step": 284 }, { "epoch": 0.28743718592964823, "grad_norm": 2.0396628379821777, "learning_rate": 9.76292823121573e-05, "loss": 2.6501, "step": 286 }, { "epoch": 0.2894472361809045, "grad_norm": 1.5687803030014038, "learning_rate": 9.757844279513281e-05, "loss": 2.3029, "step": 288 }, { "epoch": 0.2914572864321608, "grad_norm": 1.9262704849243164, "learning_rate": 9.752707744739145e-05, "loss": 2.4645, "step": 290 }, { "epoch": 0.29346733668341707, "grad_norm": 1.9408085346221924, "learning_rate": 9.747518683661631e-05, "loss": 2.3497, "step": 292 }, { "epoch": 0.29547738693467335, "grad_norm": 1.78432297706604, "learning_rate": 9.742277153629564e-05, "loss": 2.2824, "step": 294 }, { "epoch": 0.2974874371859296, "grad_norm": 2.070885181427002, "learning_rate": 9.736983212571646e-05, "loss": 2.5931, "step": 296 }, { "epoch": 0.2994974874371859, "grad_norm": 1.657235026359558, "learning_rate": 9.731636918995821e-05, "loss": 2.4344, "step": 298 }, { "epoch": 0.3015075376884422, "grad_norm": 1.8496333360671997, "learning_rate": 9.726238331988624e-05, "loss": 2.4043, "step": 300 }, { "epoch": 0.3035175879396985, "grad_norm": 1.8144067525863647, "learning_rate": 9.720787511214533e-05, "loss": 2.5708, "step": 302 }, { "epoch": 0.3055276381909548, "grad_norm": 1.875706434249878, "learning_rate": 9.715284516915303e-05, "loss": 2.387, "step": 304 }, { "epoch": 0.3075376884422111, "grad_norm": 1.905147910118103, "learning_rate": 9.709729409909307e-05, "loss": 2.6459, "step": 306 }, { "epoch": 0.30954773869346736, "grad_norm": 1.919798493385315, "learning_rate": 9.704122251590862e-05, "loss": 2.2947, "step": 308 }, { "epoch": 0.31155778894472363, "grad_norm": 1.5951811075210571, "learning_rate": 9.698463103929542e-05, "loss": 2.3532, "step": 310 }, { "epoch": 0.3135678391959799, "grad_norm": 1.6839174032211304, "learning_rate": 9.692752029469512e-05, "loss": 2.2152, "step": 312 }, { "epoch": 0.3155778894472362, "grad_norm": 2.0012948513031006, "learning_rate": 9.686989091328813e-05, "loss": 2.3561, "step": 314 }, { "epoch": 0.31758793969849247, "grad_norm": 1.8996232748031616, "learning_rate": 9.681174353198687e-05, "loss": 2.3638, "step": 316 }, { "epoch": 0.31959798994974875, "grad_norm": 1.8173401355743408, "learning_rate": 9.675307879342854e-05, "loss": 2.2213, "step": 318 }, { "epoch": 0.32160804020100503, "grad_norm": 1.6612550020217896, "learning_rate": 9.669389734596819e-05, "loss": 2.5352, "step": 320 }, { "epoch": 0.3236180904522613, "grad_norm": 1.8817943334579468, "learning_rate": 9.663419984367139e-05, "loss": 2.5058, "step": 322 }, { "epoch": 0.3256281407035176, "grad_norm": 2.1160693168640137, "learning_rate": 9.657398694630712e-05, "loss": 2.2973, "step": 324 }, { "epoch": 0.32763819095477387, "grad_norm": 1.738549828529358, "learning_rate": 9.651325931934046e-05, "loss": 2.2865, "step": 326 }, { "epoch": 0.32964824120603015, "grad_norm": 1.8293092250823975, "learning_rate": 9.645201763392513e-05, "loss": 2.1597, "step": 328 }, { "epoch": 0.3316582914572864, "grad_norm": 1.9268925189971924, "learning_rate": 9.639026256689628e-05, "loss": 2.5048, "step": 330 }, { "epoch": 0.3336683417085427, "grad_norm": 1.983359932899475, "learning_rate": 9.632799480076278e-05, "loss": 2.3261, "step": 332 }, { "epoch": 0.335678391959799, "grad_norm": 1.717955470085144, "learning_rate": 9.626521502369984e-05, "loss": 2.6444, "step": 334 }, { "epoch": 0.33768844221105526, "grad_norm": 1.836305856704712, "learning_rate": 9.620192392954132e-05, "loss": 2.4493, "step": 336 }, { "epoch": 0.33969849246231154, "grad_norm": 1.8819546699523926, "learning_rate": 9.613812221777212e-05, "loss": 2.6608, "step": 338 }, { "epoch": 0.3417085427135678, "grad_norm": 1.845680832862854, "learning_rate": 9.607381059352038e-05, "loss": 2.2059, "step": 340 }, { "epoch": 0.3437185929648241, "grad_norm": 1.878606915473938, "learning_rate": 9.600898976754977e-05, "loss": 2.2733, "step": 342 }, { "epoch": 0.3457286432160804, "grad_norm": 1.768790364265442, "learning_rate": 9.594366045625154e-05, "loss": 2.3773, "step": 344 }, { "epoch": 0.34773869346733666, "grad_norm": 1.761987328529358, "learning_rate": 9.587782338163669e-05, "loss": 2.5148, "step": 346 }, { "epoch": 0.349748743718593, "grad_norm": 1.6543912887573242, "learning_rate": 9.581147927132797e-05, "loss": 2.4609, "step": 348 }, { "epoch": 0.35175879396984927, "grad_norm": 1.9013664722442627, "learning_rate": 9.574462885855174e-05, "loss": 2.121, "step": 350 }, { "epoch": 0.35376884422110555, "grad_norm": 3.302755832672119, "learning_rate": 9.567727288213005e-05, "loss": 2.4353, "step": 352 }, { "epoch": 0.35577889447236183, "grad_norm": 2.773200750350952, "learning_rate": 9.560941208647231e-05, "loss": 2.3111, "step": 354 }, { "epoch": 0.3577889447236181, "grad_norm": 1.7840498685836792, "learning_rate": 9.554104722156716e-05, "loss": 2.5435, "step": 356 }, { "epoch": 0.3597989949748744, "grad_norm": 1.8287582397460938, "learning_rate": 9.547217904297411e-05, "loss": 2.3461, "step": 358 }, { "epoch": 0.36180904522613067, "grad_norm": 1.8139948844909668, "learning_rate": 9.540280831181525e-05, "loss": 2.3423, "step": 360 }, { "epoch": 0.36381909547738694, "grad_norm": 1.8012893199920654, "learning_rate": 9.533293579476683e-05, "loss": 2.412, "step": 362 }, { "epoch": 0.3658291457286432, "grad_norm": 1.8645695447921753, "learning_rate": 9.526256226405075e-05, "loss": 2.5261, "step": 364 }, { "epoch": 0.3678391959798995, "grad_norm": 1.9607142210006714, "learning_rate": 9.519168849742604e-05, "loss": 2.6721, "step": 366 }, { "epoch": 0.3698492462311558, "grad_norm": 1.772139549255371, "learning_rate": 9.512031527818028e-05, "loss": 2.3517, "step": 368 }, { "epoch": 0.37185929648241206, "grad_norm": 1.8913545608520508, "learning_rate": 9.504844339512095e-05, "loss": 2.6095, "step": 370 }, { "epoch": 0.37386934673366834, "grad_norm": 1.8451741933822632, "learning_rate": 9.497607364256672e-05, "loss": 2.5707, "step": 372 }, { "epoch": 0.3758793969849246, "grad_norm": 1.8256680965423584, "learning_rate": 9.490320682033855e-05, "loss": 2.3579, "step": 374 }, { "epoch": 0.3778894472361809, "grad_norm": 1.9581091403961182, "learning_rate": 9.482984373375105e-05, "loss": 2.5629, "step": 376 }, { "epoch": 0.3798994974874372, "grad_norm": 1.998180627822876, "learning_rate": 9.475598519360344e-05, "loss": 2.3817, "step": 378 }, { "epoch": 0.38190954773869346, "grad_norm": 1.7146025896072388, "learning_rate": 9.468163201617062e-05, "loss": 2.2208, "step": 380 }, { "epoch": 0.38391959798994973, "grad_norm": 1.8821600675582886, "learning_rate": 9.460678502319418e-05, "loss": 2.4416, "step": 382 }, { "epoch": 0.385929648241206, "grad_norm": 1.7585586309432983, "learning_rate": 9.453144504187327e-05, "loss": 2.3008, "step": 384 }, { "epoch": 0.3879396984924623, "grad_norm": 1.6485331058502197, "learning_rate": 9.445561290485549e-05, "loss": 2.383, "step": 386 }, { "epoch": 0.38994974874371857, "grad_norm": 1.9249005317687988, "learning_rate": 9.437928945022771e-05, "loss": 2.4204, "step": 388 }, { "epoch": 0.39195979899497485, "grad_norm": 1.8952553272247314, "learning_rate": 9.430247552150673e-05, "loss": 2.2751, "step": 390 }, { "epoch": 0.39396984924623113, "grad_norm": 2.2544572353363037, "learning_rate": 9.422517196763002e-05, "loss": 2.4784, "step": 392 }, { "epoch": 0.39597989949748746, "grad_norm": 2.308283567428589, "learning_rate": 9.414737964294636e-05, "loss": 2.2052, "step": 394 }, { "epoch": 0.39798994974874374, "grad_norm": 1.8085863590240479, "learning_rate": 9.40690994072063e-05, "loss": 2.2588, "step": 396 }, { "epoch": 0.4, "grad_norm": 1.9807807207107544, "learning_rate": 9.399033212555275e-05, "loss": 2.2698, "step": 398 }, { "epoch": 0.4020100502512563, "grad_norm": 1.6799061298370361, "learning_rate": 9.391107866851143e-05, "loss": 2.2507, "step": 400 }, { "epoch": 0.4040201005025126, "grad_norm": 1.6821558475494385, "learning_rate": 9.383133991198112e-05, "loss": 2.386, "step": 402 }, { "epoch": 0.40603015075376886, "grad_norm": 1.7488336563110352, "learning_rate": 9.375111673722414e-05, "loss": 2.7501, "step": 404 }, { "epoch": 0.40804020100502514, "grad_norm": 1.7495909929275513, "learning_rate": 9.367041003085649e-05, "loss": 2.4578, "step": 406 }, { "epoch": 0.4100502512562814, "grad_norm": 1.773022174835205, "learning_rate": 9.358922068483812e-05, "loss": 2.1896, "step": 408 }, { "epoch": 0.4120603015075377, "grad_norm": 1.6720446348190308, "learning_rate": 9.350754959646306e-05, "loss": 2.2264, "step": 410 }, { "epoch": 0.414070351758794, "grad_norm": 1.8488444089889526, "learning_rate": 9.342539766834946e-05, "loss": 2.5527, "step": 412 }, { "epoch": 0.41608040201005025, "grad_norm": 1.7335339784622192, "learning_rate": 9.334276580842967e-05, "loss": 2.2894, "step": 414 }, { "epoch": 0.41809045226130653, "grad_norm": 1.670088529586792, "learning_rate": 9.325965492994018e-05, "loss": 2.1557, "step": 416 }, { "epoch": 0.4201005025125628, "grad_norm": 1.6430253982543945, "learning_rate": 9.317606595141154e-05, "loss": 2.2405, "step": 418 }, { "epoch": 0.4221105527638191, "grad_norm": 1.7390164136886597, "learning_rate": 9.30919997966582e-05, "loss": 2.3486, "step": 420 }, { "epoch": 0.42412060301507537, "grad_norm": 1.7880635261535645, "learning_rate": 9.300745739476829e-05, "loss": 2.2626, "step": 422 }, { "epoch": 0.42613065326633165, "grad_norm": 1.7333531379699707, "learning_rate": 9.292243968009331e-05, "loss": 2.2424, "step": 424 }, { "epoch": 0.42814070351758793, "grad_norm": 1.6569937467575073, "learning_rate": 9.283694759223796e-05, "loss": 2.1534, "step": 426 }, { "epoch": 0.4301507537688442, "grad_norm": 1.7097984552383423, "learning_rate": 9.275098207604957e-05, "loss": 2.2569, "step": 428 }, { "epoch": 0.4321608040201005, "grad_norm": 1.7049428224563599, "learning_rate": 9.266454408160779e-05, "loss": 2.4531, "step": 430 }, { "epoch": 0.43417085427135677, "grad_norm": 1.5478094816207886, "learning_rate": 9.257763456421398e-05, "loss": 2.2312, "step": 432 }, { "epoch": 0.43618090452261304, "grad_norm": 1.706602692604065, "learning_rate": 9.249025448438076e-05, "loss": 2.4432, "step": 434 }, { "epoch": 0.4381909547738693, "grad_norm": 1.701011300086975, "learning_rate": 9.24024048078213e-05, "loss": 2.1473, "step": 436 }, { "epoch": 0.4402010050251256, "grad_norm": 1.7655497789382935, "learning_rate": 9.231408650543874e-05, "loss": 2.433, "step": 438 }, { "epoch": 0.44221105527638194, "grad_norm": 1.853971004486084, "learning_rate": 9.22253005533154e-05, "loss": 2.3029, "step": 440 }, { "epoch": 0.4442211055276382, "grad_norm": 1.8369559049606323, "learning_rate": 9.213604793270196e-05, "loss": 2.1837, "step": 442 }, { "epoch": 0.4462311557788945, "grad_norm": 1.6373860836029053, "learning_rate": 9.204632963000671e-05, "loss": 2.4237, "step": 444 }, { "epoch": 0.4482412060301508, "grad_norm": 1.7693603038787842, "learning_rate": 9.195614663678458e-05, "loss": 2.4229, "step": 446 }, { "epoch": 0.45025125628140705, "grad_norm": 1.6722673177719116, "learning_rate": 9.186549994972618e-05, "loss": 2.3788, "step": 448 }, { "epoch": 0.45226130653266333, "grad_norm": 1.6546918153762817, "learning_rate": 9.177439057064683e-05, "loss": 2.2079, "step": 450 }, { "epoch": 0.4542713567839196, "grad_norm": 1.8333224058151245, "learning_rate": 9.168281950647545e-05, "loss": 2.5018, "step": 452 }, { "epoch": 0.4562814070351759, "grad_norm": 1.7532380819320679, "learning_rate": 9.159078776924346e-05, "loss": 2.0344, "step": 454 }, { "epoch": 0.45829145728643217, "grad_norm": 1.647209644317627, "learning_rate": 9.149829637607353e-05, "loss": 2.2324, "step": 456 }, { "epoch": 0.46030150753768845, "grad_norm": 1.6968568563461304, "learning_rate": 9.140534634916846e-05, "loss": 2.1635, "step": 458 }, { "epoch": 0.4623115577889447, "grad_norm": 1.6668813228607178, "learning_rate": 9.131193871579975e-05, "loss": 2.334, "step": 460 }, { "epoch": 0.464321608040201, "grad_norm": 1.6174511909484863, "learning_rate": 9.121807450829632e-05, "loss": 2.1475, "step": 462 }, { "epoch": 0.4663316582914573, "grad_norm": 1.8468537330627441, "learning_rate": 9.112375476403312e-05, "loss": 2.2984, "step": 464 }, { "epoch": 0.46834170854271356, "grad_norm": 1.7187474966049194, "learning_rate": 9.102898052541958e-05, "loss": 2.4139, "step": 466 }, { "epoch": 0.47035175879396984, "grad_norm": 1.7762575149536133, "learning_rate": 9.093375283988819e-05, "loss": 2.1496, "step": 468 }, { "epoch": 0.4723618090452261, "grad_norm": 1.9211578369140625, "learning_rate": 9.083807275988284e-05, "loss": 2.1667, "step": 470 }, { "epoch": 0.4743718592964824, "grad_norm": 3.2748539447784424, "learning_rate": 9.074194134284726e-05, "loss": 2.2991, "step": 472 }, { "epoch": 0.4763819095477387, "grad_norm": 1.751917839050293, "learning_rate": 9.064535965121324e-05, "loss": 2.2903, "step": 474 }, { "epoch": 0.47839195979899496, "grad_norm": 1.7564281225204468, "learning_rate": 9.054832875238903e-05, "loss": 2.5954, "step": 476 }, { "epoch": 0.48040201005025124, "grad_norm": 1.6731312274932861, "learning_rate": 9.045084971874738e-05, "loss": 2.3209, "step": 478 }, { "epoch": 0.4824120603015075, "grad_norm": 2.067354679107666, "learning_rate": 9.035292362761381e-05, "loss": 2.1253, "step": 480 }, { "epoch": 0.4844221105527638, "grad_norm": 1.6500976085662842, "learning_rate": 9.025455156125466e-05, "loss": 2.0942, "step": 482 }, { "epoch": 0.4864321608040201, "grad_norm": 1.7590264081954956, "learning_rate": 9.015573460686509e-05, "loss": 2.3827, "step": 484 }, { "epoch": 0.4884422110552764, "grad_norm": 1.8388062715530396, "learning_rate": 9.005647385655718e-05, "loss": 2.2311, "step": 486 }, { "epoch": 0.4904522613065327, "grad_norm": 1.7108007669448853, "learning_rate": 8.995677040734769e-05, "loss": 2.1734, "step": 488 }, { "epoch": 0.49246231155778897, "grad_norm": 2.4704267978668213, "learning_rate": 8.985662536114613e-05, "loss": 2.2604, "step": 490 }, { "epoch": 0.49447236180904525, "grad_norm": 1.7523339986801147, "learning_rate": 8.97560398247424e-05, "loss": 2.3831, "step": 492 }, { "epoch": 0.4964824120603015, "grad_norm": 1.5778342485427856, "learning_rate": 8.965501490979467e-05, "loss": 2.118, "step": 494 }, { "epoch": 0.4984924623115578, "grad_norm": 1.5618497133255005, "learning_rate": 8.955355173281708e-05, "loss": 2.3317, "step": 496 }, { "epoch": 0.5005025125628141, "grad_norm": 2.049302339553833, "learning_rate": 8.945165141516734e-05, "loss": 2.2127, "step": 498 }, { "epoch": 0.5025125628140703, "grad_norm": 2.8442394733428955, "learning_rate": 8.934931508303445e-05, "loss": 2.1562, "step": 500 }, { "epoch": 0.5045226130653266, "grad_norm": 1.7337098121643066, "learning_rate": 8.924654386742613e-05, "loss": 2.2541, "step": 502 }, { "epoch": 0.5065326633165829, "grad_norm": 1.6087921857833862, "learning_rate": 8.914333890415639e-05, "loss": 2.3858, "step": 504 }, { "epoch": 0.5085427135678392, "grad_norm": 1.7101848125457764, "learning_rate": 8.903970133383297e-05, "loss": 2.0366, "step": 506 }, { "epoch": 0.5105527638190954, "grad_norm": 1.640261173248291, "learning_rate": 8.89356323018447e-05, "loss": 2.3109, "step": 508 }, { "epoch": 0.5125628140703518, "grad_norm": 1.7887450456619263, "learning_rate": 8.883113295834892e-05, "loss": 2.23, "step": 510 }, { "epoch": 0.5145728643216081, "grad_norm": 1.772544503211975, "learning_rate": 8.872620445825868e-05, "loss": 2.3887, "step": 512 }, { "epoch": 0.5165829145728643, "grad_norm": 1.6293214559555054, "learning_rate": 8.862084796122998e-05, "loss": 2.2669, "step": 514 }, { "epoch": 0.5185929648241207, "grad_norm": 1.9295259714126587, "learning_rate": 8.851506463164907e-05, "loss": 2.4984, "step": 516 }, { "epoch": 0.5206030150753769, "grad_norm": 1.5557610988616943, "learning_rate": 8.84088556386194e-05, "loss": 2.2621, "step": 518 }, { "epoch": 0.5226130653266332, "grad_norm": 1.9299404621124268, "learning_rate": 8.83022221559489e-05, "loss": 2.5963, "step": 520 }, { "epoch": 0.5246231155778894, "grad_norm": 1.7403899431228638, "learning_rate": 8.819516536213683e-05, "loss": 2.261, "step": 522 }, { "epoch": 0.5266331658291458, "grad_norm": 1.6017705202102661, "learning_rate": 8.808768644036085e-05, "loss": 2.1345, "step": 524 }, { "epoch": 0.528643216080402, "grad_norm": 1.560952067375183, "learning_rate": 8.797978657846391e-05, "loss": 2.1304, "step": 526 }, { "epoch": 0.5306532663316583, "grad_norm": 1.63511061668396, "learning_rate": 8.787146696894118e-05, "loss": 2.2639, "step": 528 }, { "epoch": 0.5326633165829145, "grad_norm": 1.6902211904525757, "learning_rate": 8.776272880892675e-05, "loss": 2.4526, "step": 530 }, { "epoch": 0.5346733668341709, "grad_norm": 2.2787492275238037, "learning_rate": 8.765357330018056e-05, "loss": 2.1836, "step": 532 }, { "epoch": 0.5366834170854271, "grad_norm": 1.6377997398376465, "learning_rate": 8.754400164907497e-05, "loss": 2.1207, "step": 534 }, { "epoch": 0.5386934673366834, "grad_norm": 1.829947829246521, "learning_rate": 8.74340150665815e-05, "loss": 2.4133, "step": 536 }, { "epoch": 0.5407035175879397, "grad_norm": 1.8088699579238892, "learning_rate": 8.732361476825752e-05, "loss": 2.4477, "step": 538 }, { "epoch": 0.542713567839196, "grad_norm": 1.8463410139083862, "learning_rate": 8.721280197423258e-05, "loss": 2.5268, "step": 540 }, { "epoch": 0.5447236180904522, "grad_norm": 1.7526369094848633, "learning_rate": 8.710157790919522e-05, "loss": 2.2297, "step": 542 }, { "epoch": 0.5467336683417086, "grad_norm": 1.548508882522583, "learning_rate": 8.69899438023792e-05, "loss": 2.3638, "step": 544 }, { "epoch": 0.5487437185929648, "grad_norm": 1.6327900886535645, "learning_rate": 8.687790088755008e-05, "loss": 2.5126, "step": 546 }, { "epoch": 0.5507537688442211, "grad_norm": 4.006368160247803, "learning_rate": 8.676545040299145e-05, "loss": 2.4828, "step": 548 }, { "epoch": 0.5527638190954773, "grad_norm": 1.6274868249893188, "learning_rate": 8.665259359149132e-05, "loss": 2.2662, "step": 550 }, { "epoch": 0.5547738693467337, "grad_norm": 1.686354398727417, "learning_rate": 8.653933170032842e-05, "loss": 2.4134, "step": 552 }, { "epoch": 0.5567839195979899, "grad_norm": 1.62911856174469, "learning_rate": 8.642566598125831e-05, "loss": 2.275, "step": 554 }, { "epoch": 0.5587939698492462, "grad_norm": 1.634544849395752, "learning_rate": 8.631159769049965e-05, "loss": 2.2874, "step": 556 }, { "epoch": 0.5608040201005026, "grad_norm": 1.5775830745697021, "learning_rate": 8.619712808872024e-05, "loss": 2.2704, "step": 558 }, { "epoch": 0.5628140703517588, "grad_norm": 1.618634581565857, "learning_rate": 8.60822584410231e-05, "loss": 2.545, "step": 560 }, { "epoch": 0.5648241206030151, "grad_norm": 1.692204475402832, "learning_rate": 8.596699001693255e-05, "loss": 2.2853, "step": 562 }, { "epoch": 0.5668341708542713, "grad_norm": 1.717124342918396, "learning_rate": 8.585132409038013e-05, "loss": 2.1552, "step": 564 }, { "epoch": 0.5688442211055277, "grad_norm": 1.6397979259490967, "learning_rate": 8.573526193969046e-05, "loss": 2.3056, "step": 566 }, { "epoch": 0.5708542713567839, "grad_norm": 1.6189571619033813, "learning_rate": 8.561880484756725e-05, "loss": 2.1406, "step": 568 }, { "epoch": 0.5728643216080402, "grad_norm": 1.9604424238204956, "learning_rate": 8.550195410107902e-05, "loss": 2.4183, "step": 570 }, { "epoch": 0.5748743718592965, "grad_norm": 1.6223558187484741, "learning_rate": 8.538471099164493e-05, "loss": 2.1032, "step": 572 }, { "epoch": 0.5768844221105528, "grad_norm": 1.8475818634033203, "learning_rate": 8.526707681502044e-05, "loss": 2.1361, "step": 574 }, { "epoch": 0.578894472361809, "grad_norm": 1.6998199224472046, "learning_rate": 8.51490528712831e-05, "loss": 2.0701, "step": 576 }, { "epoch": 0.5809045226130654, "grad_norm": 1.5956889390945435, "learning_rate": 8.503064046481803e-05, "loss": 2.3293, "step": 578 }, { "epoch": 0.5829145728643216, "grad_norm": 1.544945478439331, "learning_rate": 8.491184090430364e-05, "loss": 2.0934, "step": 580 }, { "epoch": 0.5849246231155779, "grad_norm": 1.5254535675048828, "learning_rate": 8.479265550269714e-05, "loss": 1.9323, "step": 582 }, { "epoch": 0.5869346733668341, "grad_norm": 1.6815849542617798, "learning_rate": 8.467308557721996e-05, "loss": 2.2834, "step": 584 }, { "epoch": 0.5889447236180905, "grad_norm": 1.7338416576385498, "learning_rate": 8.455313244934324e-05, "loss": 2.2117, "step": 586 }, { "epoch": 0.5909547738693467, "grad_norm": 1.6180495023727417, "learning_rate": 8.443279744477324e-05, "loss": 2.1526, "step": 588 }, { "epoch": 0.592964824120603, "grad_norm": 1.6239418983459473, "learning_rate": 8.43120818934367e-05, "loss": 2.1047, "step": 590 }, { "epoch": 0.5949748743718593, "grad_norm": 1.6448568105697632, "learning_rate": 8.419098712946601e-05, "loss": 2.0582, "step": 592 }, { "epoch": 0.5969849246231156, "grad_norm": 1.722649097442627, "learning_rate": 8.406951449118469e-05, "loss": 1.9733, "step": 594 }, { "epoch": 0.5989949748743718, "grad_norm": 1.7211109399795532, "learning_rate": 8.394766532109242e-05, "loss": 2.2033, "step": 596 }, { "epoch": 0.6010050251256281, "grad_norm": 1.55821692943573, "learning_rate": 8.382544096585027e-05, "loss": 1.8462, "step": 598 }, { "epoch": 0.6030150753768844, "grad_norm": 1.5059350728988647, "learning_rate": 8.370284277626577e-05, "loss": 2.0388, "step": 600 }, { "epoch": 0.6050251256281407, "grad_norm": 1.7469459772109985, "learning_rate": 8.357987210727808e-05, "loss": 2.367, "step": 602 }, { "epoch": 0.607035175879397, "grad_norm": 1.6946009397506714, "learning_rate": 8.345653031794292e-05, "loss": 2.3654, "step": 604 }, { "epoch": 0.6090452261306533, "grad_norm": 2.0396924018859863, "learning_rate": 8.333281877141758e-05, "loss": 2.2726, "step": 606 }, { "epoch": 0.6110552763819096, "grad_norm": 1.6228737831115723, "learning_rate": 8.320873883494585e-05, "loss": 2.2077, "step": 608 }, { "epoch": 0.6130653266331658, "grad_norm": 1.6517311334609985, "learning_rate": 8.308429187984297e-05, "loss": 2.2134, "step": 610 }, { "epoch": 0.6150753768844222, "grad_norm": 1.6150524616241455, "learning_rate": 8.295947928148036e-05, "loss": 2.1454, "step": 612 }, { "epoch": 0.6170854271356784, "grad_norm": 1.5641722679138184, "learning_rate": 8.283430241927052e-05, "loss": 1.9578, "step": 614 }, { "epoch": 0.6190954773869347, "grad_norm": 1.668264389038086, "learning_rate": 8.270876267665173e-05, "loss": 2.1463, "step": 616 }, { "epoch": 0.6211055276381909, "grad_norm": 1.6941081285476685, "learning_rate": 8.258286144107276e-05, "loss": 2.2608, "step": 618 }, { "epoch": 0.6231155778894473, "grad_norm": 1.9184767007827759, "learning_rate": 8.24566001039776e-05, "loss": 2.2404, "step": 620 }, { "epoch": 0.6251256281407035, "grad_norm": 1.5730127096176147, "learning_rate": 8.232998006078997e-05, "loss": 2.1935, "step": 622 }, { "epoch": 0.6271356783919598, "grad_norm": 1.4828630685806274, "learning_rate": 8.220300271089807e-05, "loss": 2.2226, "step": 624 }, { "epoch": 0.629145728643216, "grad_norm": 1.5179678201675415, "learning_rate": 8.207566945763885e-05, "loss": 2.0261, "step": 626 }, { "epoch": 0.6311557788944724, "grad_norm": 1.4846012592315674, "learning_rate": 8.19479817082828e-05, "loss": 2.1515, "step": 628 }, { "epoch": 0.6331658291457286, "grad_norm": 1.6346242427825928, "learning_rate": 8.181994087401819e-05, "loss": 2.0147, "step": 630 }, { "epoch": 0.6351758793969849, "grad_norm": 1.5853772163391113, "learning_rate": 8.169154836993551e-05, "loss": 2.1816, "step": 632 }, { "epoch": 0.6371859296482412, "grad_norm": 1.573282241821289, "learning_rate": 8.156280561501195e-05, "loss": 2.0802, "step": 634 }, { "epoch": 0.6391959798994975, "grad_norm": 1.5579333305358887, "learning_rate": 8.143371403209554e-05, "loss": 2.1474, "step": 636 }, { "epoch": 0.6412060301507537, "grad_norm": 1.6341716051101685, "learning_rate": 8.130427504788955e-05, "loss": 2.293, "step": 638 }, { "epoch": 0.6432160804020101, "grad_norm": 1.6196293830871582, "learning_rate": 8.117449009293668e-05, "loss": 2.0744, "step": 640 }, { "epoch": 0.6452261306532663, "grad_norm": 1.702324628829956, "learning_rate": 8.104436060160324e-05, "loss": 2.158, "step": 642 }, { "epoch": 0.6472361809045226, "grad_norm": 1.734678030014038, "learning_rate": 8.091388801206333e-05, "loss": 2.2079, "step": 644 }, { "epoch": 0.6492462311557788, "grad_norm": 1.667091727256775, "learning_rate": 8.07830737662829e-05, "loss": 2.1447, "step": 646 }, { "epoch": 0.6512562814070352, "grad_norm": 1.7557045221328735, "learning_rate": 8.06519193100039e-05, "loss": 2.2181, "step": 648 }, { "epoch": 0.6532663316582915, "grad_norm": 1.6288976669311523, "learning_rate": 8.052042609272817e-05, "loss": 2.0902, "step": 650 }, { "epoch": 0.6552763819095477, "grad_norm": 1.6819061040878296, "learning_rate": 8.038859556770151e-05, "loss": 2.3705, "step": 652 }, { "epoch": 0.6572864321608041, "grad_norm": 1.7110100984573364, "learning_rate": 8.025642919189762e-05, "loss": 2.1783, "step": 654 }, { "epoch": 0.6592964824120603, "grad_norm": 1.7823935747146606, "learning_rate": 8.012392842600198e-05, "loss": 2.3344, "step": 656 }, { "epoch": 0.6613065326633166, "grad_norm": 1.661436915397644, "learning_rate": 7.99910947343957e-05, "loss": 2.2976, "step": 658 }, { "epoch": 0.6633165829145728, "grad_norm": 1.6022391319274902, "learning_rate": 7.985792958513931e-05, "loss": 1.9313, "step": 660 }, { "epoch": 0.6653266331658292, "grad_norm": 1.5373127460479736, "learning_rate": 7.972443444995663e-05, "loss": 2.182, "step": 662 }, { "epoch": 0.6673366834170854, "grad_norm": 1.713374376296997, "learning_rate": 7.959061080421839e-05, "loss": 2.3038, "step": 664 }, { "epoch": 0.6693467336683417, "grad_norm": 1.645225167274475, "learning_rate": 7.9456460126926e-05, "loss": 2.2538, "step": 666 }, { "epoch": 0.671356783919598, "grad_norm": 1.480999231338501, "learning_rate": 7.932198390069515e-05, "loss": 1.9686, "step": 668 }, { "epoch": 0.6733668341708543, "grad_norm": 3.5348572731018066, "learning_rate": 7.91871836117395e-05, "loss": 2.3065, "step": 670 }, { "epoch": 0.6753768844221105, "grad_norm": 1.6705471277236938, "learning_rate": 7.905206074985416e-05, "loss": 2.1366, "step": 672 }, { "epoch": 0.6773869346733669, "grad_norm": 1.6539703607559204, "learning_rate": 7.891661680839932e-05, "loss": 2.111, "step": 674 }, { "epoch": 0.6793969849246231, "grad_norm": 1.5140559673309326, "learning_rate": 7.878085328428369e-05, "loss": 2.1999, "step": 676 }, { "epoch": 0.6814070351758794, "grad_norm": 1.6765365600585938, "learning_rate": 7.86447716779479e-05, "loss": 1.9244, "step": 678 }, { "epoch": 0.6834170854271356, "grad_norm": 1.6503502130508423, "learning_rate": 7.85083734933481e-05, "loss": 2.0321, "step": 680 }, { "epoch": 0.685427135678392, "grad_norm": 1.5550287961959839, "learning_rate": 7.83716602379391e-05, "loss": 1.9953, "step": 682 }, { "epoch": 0.6874371859296482, "grad_norm": 1.5494416952133179, "learning_rate": 7.823463342265792e-05, "loss": 2.1868, "step": 684 }, { "epoch": 0.6894472361809045, "grad_norm": 1.5480455160140991, "learning_rate": 7.809729456190698e-05, "loss": 2.2091, "step": 686 }, { "epoch": 0.6914572864321608, "grad_norm": 1.5905412435531616, "learning_rate": 7.795964517353735e-05, "loss": 2.0997, "step": 688 }, { "epoch": 0.6934673366834171, "grad_norm": 1.6843619346618652, "learning_rate": 7.782168677883206e-05, "loss": 2.2762, "step": 690 }, { "epoch": 0.6954773869346733, "grad_norm": 1.9944095611572266, "learning_rate": 7.76834209024892e-05, "loss": 1.8889, "step": 692 }, { "epoch": 0.6974874371859296, "grad_norm": 1.6074081659317017, "learning_rate": 7.754484907260513e-05, "loss": 2.2474, "step": 694 }, { "epoch": 0.699497487437186, "grad_norm": 1.506535530090332, "learning_rate": 7.740597282065756e-05, "loss": 2.1043, "step": 696 }, { "epoch": 0.7015075376884422, "grad_norm": 1.5604645013809204, "learning_rate": 7.726679368148864e-05, "loss": 2.0762, "step": 698 }, { "epoch": 0.7035175879396985, "grad_norm": 1.6519221067428589, "learning_rate": 7.712731319328798e-05, "loss": 2.123, "step": 700 }, { "epoch": 0.7055276381909548, "grad_norm": 1.775421142578125, "learning_rate": 7.698753289757565e-05, "loss": 2.125, "step": 702 }, { "epoch": 0.7075376884422111, "grad_norm": 1.552018404006958, "learning_rate": 7.684745433918518e-05, "loss": 1.952, "step": 704 }, { "epoch": 0.7095477386934673, "grad_norm": 1.8088091611862183, "learning_rate": 7.670707906624644e-05, "loss": 2.2094, "step": 706 }, { "epoch": 0.7115577889447237, "grad_norm": 1.5663000345230103, "learning_rate": 7.656640863016857e-05, "loss": 2.2196, "step": 708 }, { "epoch": 0.7135678391959799, "grad_norm": 1.6653417348861694, "learning_rate": 7.642544458562278e-05, "loss": 2.1193, "step": 710 }, { "epoch": 0.7155778894472362, "grad_norm": 1.5582131147384644, "learning_rate": 7.628418849052523e-05, "loss": 2.0609, "step": 712 }, { "epoch": 0.7175879396984924, "grad_norm": 1.6001904010772705, "learning_rate": 7.614264190601981e-05, "loss": 1.9337, "step": 714 }, { "epoch": 0.7195979899497488, "grad_norm": 1.5969533920288086, "learning_rate": 7.600080639646077e-05, "loss": 1.9728, "step": 716 }, { "epoch": 0.721608040201005, "grad_norm": 1.5485557317733765, "learning_rate": 7.585868352939563e-05, "loss": 2.0791, "step": 718 }, { "epoch": 0.7236180904522613, "grad_norm": 1.6326680183410645, "learning_rate": 7.571627487554769e-05, "loss": 2.1784, "step": 720 }, { "epoch": 0.7256281407035176, "grad_norm": 1.7753201723098755, "learning_rate": 7.55735820087987e-05, "loss": 2.1085, "step": 722 }, { "epoch": 0.7276381909547739, "grad_norm": 1.6765743494033813, "learning_rate": 7.543060650617158e-05, "loss": 2.2751, "step": 724 }, { "epoch": 0.7296482412060301, "grad_norm": 1.7378075122833252, "learning_rate": 7.528734994781283e-05, "loss": 2.5291, "step": 726 }, { "epoch": 0.7316582914572864, "grad_norm": 1.5884289741516113, "learning_rate": 7.514381391697517e-05, "loss": 2.1576, "step": 728 }, { "epoch": 0.7336683417085427, "grad_norm": 1.4826475381851196, "learning_rate": 7.500000000000001e-05, "loss": 2.0913, "step": 730 }, { "epoch": 0.735678391959799, "grad_norm": 1.5756173133850098, "learning_rate": 7.48559097862999e-05, "loss": 1.977, "step": 732 }, { "epoch": 0.7376884422110552, "grad_norm": 1.5923422574996948, "learning_rate": 7.471154486834105e-05, "loss": 2.1248, "step": 734 }, { "epoch": 0.7396984924623116, "grad_norm": 1.8260492086410522, "learning_rate": 7.456690684162557e-05, "loss": 2.1871, "step": 736 }, { "epoch": 0.7417085427135678, "grad_norm": 1.6289384365081787, "learning_rate": 7.442199730467402e-05, "loss": 1.9642, "step": 738 }, { "epoch": 0.7437185929648241, "grad_norm": 1.423568606376648, "learning_rate": 7.427681785900761e-05, "loss": 1.7796, "step": 740 }, { "epoch": 0.7457286432160805, "grad_norm": 1.7022663354873657, "learning_rate": 7.413137010913054e-05, "loss": 2.2234, "step": 742 }, { "epoch": 0.7477386934673367, "grad_norm": 1.6905540227890015, "learning_rate": 7.398565566251232e-05, "loss": 2.1867, "step": 744 }, { "epoch": 0.749748743718593, "grad_norm": 1.545426607131958, "learning_rate": 7.383967612956988e-05, "loss": 2.0914, "step": 746 }, { "epoch": 0.7517587939698492, "grad_norm": 1.6682387590408325, "learning_rate": 7.369343312364993e-05, "loss": 2.2405, "step": 748 }, { "epoch": 0.7537688442211056, "grad_norm": 1.634537935256958, "learning_rate": 7.354692826101102e-05, "loss": 1.9596, "step": 750 }, { "epoch": 0.7557788944723618, "grad_norm": 1.6119855642318726, "learning_rate": 7.340016316080565e-05, "loss": 2.1663, "step": 752 }, { "epoch": 0.7577889447236181, "grad_norm": 1.621778130531311, "learning_rate": 7.325313944506254e-05, "loss": 1.9821, "step": 754 }, { "epoch": 0.7597989949748744, "grad_norm": 2.1226937770843506, "learning_rate": 7.310585873866848e-05, "loss": 2.2974, "step": 756 }, { "epoch": 0.7618090452261307, "grad_norm": 1.5477526187896729, "learning_rate": 7.295832266935059e-05, "loss": 1.8883, "step": 758 }, { "epoch": 0.7638190954773869, "grad_norm": 1.5489342212677002, "learning_rate": 7.281053286765815e-05, "loss": 2.0525, "step": 760 }, { "epoch": 0.7658291457286432, "grad_norm": 2.4198858737945557, "learning_rate": 7.26624909669447e-05, "loss": 2.0754, "step": 762 }, { "epoch": 0.7678391959798995, "grad_norm": 1.5412720441818237, "learning_rate": 7.251419860334994e-05, "loss": 1.8809, "step": 764 }, { "epoch": 0.7698492462311558, "grad_norm": 1.656874179840088, "learning_rate": 7.236565741578163e-05, "loss": 2.0115, "step": 766 }, { "epoch": 0.771859296482412, "grad_norm": 1.5036085844039917, "learning_rate": 7.221686904589754e-05, "loss": 2.2604, "step": 768 }, { "epoch": 0.7738693467336684, "grad_norm": 1.4638508558273315, "learning_rate": 7.20678351380872e-05, "loss": 1.9812, "step": 770 }, { "epoch": 0.7758793969849246, "grad_norm": 1.5553574562072754, "learning_rate": 7.191855733945387e-05, "loss": 2.0646, "step": 772 }, { "epoch": 0.7778894472361809, "grad_norm": 1.5192738771438599, "learning_rate": 7.176903729979621e-05, "loss": 2.3253, "step": 774 }, { "epoch": 0.7798994974874371, "grad_norm": 1.702390432357788, "learning_rate": 7.161927667159013e-05, "loss": 2.0733, "step": 776 }, { "epoch": 0.7819095477386935, "grad_norm": 1.551355004310608, "learning_rate": 7.146927710997047e-05, "loss": 2.0301, "step": 778 }, { "epoch": 0.7839195979899497, "grad_norm": 1.6954432725906372, "learning_rate": 7.13190402727127e-05, "loss": 1.9302, "step": 780 }, { "epoch": 0.785929648241206, "grad_norm": 1.5084172487258911, "learning_rate": 7.116856782021468e-05, "loss": 2.4387, "step": 782 }, { "epoch": 0.7879396984924623, "grad_norm": 1.5950528383255005, "learning_rate": 7.101786141547828e-05, "loss": 2.1753, "step": 784 }, { "epoch": 0.7899497487437186, "grad_norm": 1.674777865409851, "learning_rate": 7.08669227240909e-05, "loss": 2.0804, "step": 786 }, { "epoch": 0.7919597989949749, "grad_norm": 1.5074516534805298, "learning_rate": 7.071575341420719e-05, "loss": 2.2549, "step": 788 }, { "epoch": 0.7939698492462312, "grad_norm": 1.6578811407089233, "learning_rate": 7.056435515653059e-05, "loss": 2.1851, "step": 790 }, { "epoch": 0.7959798994974875, "grad_norm": 1.6693928241729736, "learning_rate": 7.041272962429477e-05, "loss": 2.2608, "step": 792 }, { "epoch": 0.7979899497487437, "grad_norm": 1.6492758989334106, "learning_rate": 7.026087849324527e-05, "loss": 1.9814, "step": 794 }, { "epoch": 0.8, "grad_norm": 1.59951651096344, "learning_rate": 7.010880344162088e-05, "loss": 2.0959, "step": 796 }, { "epoch": 0.8020100502512563, "grad_norm": 1.6797847747802734, "learning_rate": 6.995650615013516e-05, "loss": 2.242, "step": 798 }, { "epoch": 0.8040201005025126, "grad_norm": 1.566300868988037, "learning_rate": 6.980398830195785e-05, "loss": 2.2452, "step": 800 }, { "epoch": 0.8060301507537688, "grad_norm": 1.471531629562378, "learning_rate": 6.965125158269619e-05, "loss": 1.7636, "step": 802 }, { "epoch": 0.8080402010050252, "grad_norm": 1.5839449167251587, "learning_rate": 6.94982976803764e-05, "loss": 2.299, "step": 804 }, { "epoch": 0.8100502512562814, "grad_norm": 1.5502268075942993, "learning_rate": 6.934512828542497e-05, "loss": 1.8683, "step": 806 }, { "epoch": 0.8120603015075377, "grad_norm": 1.532381296157837, "learning_rate": 6.919174509065004e-05, "loss": 1.8511, "step": 808 }, { "epoch": 0.8140703517587939, "grad_norm": 1.4679884910583496, "learning_rate": 6.903814979122249e-05, "loss": 2.1163, "step": 810 }, { "epoch": 0.8160804020100503, "grad_norm": 1.5164811611175537, "learning_rate": 6.888434408465751e-05, "loss": 2.1253, "step": 812 }, { "epoch": 0.8180904522613065, "grad_norm": 1.6110063791275024, "learning_rate": 6.873032967079561e-05, "loss": 2.2175, "step": 814 }, { "epoch": 0.8201005025125628, "grad_norm": 1.4853484630584717, "learning_rate": 6.85761082517839e-05, "loss": 2.1247, "step": 816 }, { "epoch": 0.8221105527638191, "grad_norm": 1.7195366621017456, "learning_rate": 6.842168153205734e-05, "loss": 2.2131, "step": 818 }, { "epoch": 0.8241206030150754, "grad_norm": 1.622910976409912, "learning_rate": 6.826705121831976e-05, "loss": 2.0826, "step": 820 }, { "epoch": 0.8261306532663316, "grad_norm": 1.5393376350402832, "learning_rate": 6.811221901952513e-05, "loss": 1.9764, "step": 822 }, { "epoch": 0.828140703517588, "grad_norm": 1.475002408027649, "learning_rate": 6.795718664685868e-05, "loss": 1.9606, "step": 824 }, { "epoch": 0.8301507537688442, "grad_norm": 1.607079267501831, "learning_rate": 6.780195581371784e-05, "loss": 2.3341, "step": 826 }, { "epoch": 0.8321608040201005, "grad_norm": 1.5544415712356567, "learning_rate": 6.764652823569344e-05, "loss": 2.3198, "step": 828 }, { "epoch": 0.8341708542713567, "grad_norm": 1.5592552423477173, "learning_rate": 6.749090563055076e-05, "loss": 1.8339, "step": 830 }, { "epoch": 0.8361809045226131, "grad_norm": 1.6058601140975952, "learning_rate": 6.733508971821036e-05, "loss": 2.0026, "step": 832 }, { "epoch": 0.8381909547738694, "grad_norm": 1.5740671157836914, "learning_rate": 6.717908222072935e-05, "loss": 2.1512, "step": 834 }, { "epoch": 0.8402010050251256, "grad_norm": 1.4987316131591797, "learning_rate": 6.702288486228216e-05, "loss": 2.1372, "step": 836 }, { "epoch": 0.842211055276382, "grad_norm": 1.501441478729248, "learning_rate": 6.686649936914152e-05, "loss": 2.1216, "step": 838 }, { "epoch": 0.8442211055276382, "grad_norm": 1.522459626197815, "learning_rate": 6.670992746965938e-05, "loss": 2.2921, "step": 840 }, { "epoch": 0.8462311557788945, "grad_norm": 1.5145554542541504, "learning_rate": 6.65531708942479e-05, "loss": 2.0768, "step": 842 }, { "epoch": 0.8482412060301507, "grad_norm": 1.7844620943069458, "learning_rate": 6.639623137536023e-05, "loss": 2.0404, "step": 844 }, { "epoch": 0.8502512562814071, "grad_norm": 1.5324727296829224, "learning_rate": 6.623911064747133e-05, "loss": 1.9641, "step": 846 }, { "epoch": 0.8522613065326633, "grad_norm": 1.5896046161651611, "learning_rate": 6.608181044705892e-05, "loss": 1.922, "step": 848 }, { "epoch": 0.8542713567839196, "grad_norm": 1.531399130821228, "learning_rate": 6.592433251258423e-05, "loss": 1.9917, "step": 850 }, { "epoch": 0.8562814070351759, "grad_norm": 1.3786967992782593, "learning_rate": 6.576667858447272e-05, "loss": 2.0952, "step": 852 }, { "epoch": 0.8582914572864322, "grad_norm": 1.5708434581756592, "learning_rate": 6.560885040509499e-05, "loss": 2.0783, "step": 854 }, { "epoch": 0.8603015075376884, "grad_norm": 1.5222376585006714, "learning_rate": 6.545084971874738e-05, "loss": 1.9323, "step": 856 }, { "epoch": 0.8623115577889447, "grad_norm": 1.591243028640747, "learning_rate": 6.529267827163277e-05, "loss": 2.0764, "step": 858 }, { "epoch": 0.864321608040201, "grad_norm": 1.5842283964157104, "learning_rate": 6.51343378118413e-05, "loss": 2.183, "step": 860 }, { "epoch": 0.8663316582914573, "grad_norm": 1.5385973453521729, "learning_rate": 6.497583008933097e-05, "loss": 2.2194, "step": 862 }, { "epoch": 0.8683417085427135, "grad_norm": 1.5147958993911743, "learning_rate": 6.481715685590836e-05, "loss": 1.9953, "step": 864 }, { "epoch": 0.8703517587939699, "grad_norm": 1.5883772373199463, "learning_rate": 6.465831986520927e-05, "loss": 2.0174, "step": 866 }, { "epoch": 0.8723618090452261, "grad_norm": 1.5510574579238892, "learning_rate": 6.449932087267932e-05, "loss": 1.9516, "step": 868 }, { "epoch": 0.8743718592964824, "grad_norm": 1.4354772567749023, "learning_rate": 6.434016163555452e-05, "loss": 1.9101, "step": 870 }, { "epoch": 0.8763819095477386, "grad_norm": 1.5250461101531982, "learning_rate": 6.418084391284192e-05, "loss": 2.1036, "step": 872 }, { "epoch": 0.878391959798995, "grad_norm": 1.6079366207122803, "learning_rate": 6.402136946530014e-05, "loss": 2.1401, "step": 874 }, { "epoch": 0.8804020100502512, "grad_norm": 1.5236170291900635, "learning_rate": 6.386174005541986e-05, "loss": 1.9832, "step": 876 }, { "epoch": 0.8824120603015075, "grad_norm": 1.5333629846572876, "learning_rate": 6.370195744740442e-05, "loss": 1.8619, "step": 878 }, { "epoch": 0.8844221105527639, "grad_norm": 1.5753098726272583, "learning_rate": 6.354202340715026e-05, "loss": 1.8309, "step": 880 }, { "epoch": 0.8864321608040201, "grad_norm": 1.5884308815002441, "learning_rate": 6.338193970222744e-05, "loss": 2.1437, "step": 882 }, { "epoch": 0.8884422110552764, "grad_norm": 1.5323386192321777, "learning_rate": 6.322170810186012e-05, "loss": 2.1074, "step": 884 }, { "epoch": 0.8904522613065327, "grad_norm": 1.4975218772888184, "learning_rate": 6.306133037690693e-05, "loss": 2.0906, "step": 886 }, { "epoch": 0.892462311557789, "grad_norm": 1.5518718957901, "learning_rate": 6.290080829984147e-05, "loss": 1.9955, "step": 888 }, { "epoch": 0.8944723618090452, "grad_norm": 1.5773143768310547, "learning_rate": 6.274014364473274e-05, "loss": 2.2483, "step": 890 }, { "epoch": 0.8964824120603015, "grad_norm": 1.533818006515503, "learning_rate": 6.257933818722543e-05, "loss": 2.093, "step": 892 }, { "epoch": 0.8984924623115578, "grad_norm": 1.442911148071289, "learning_rate": 6.241839370452041e-05, "loss": 2.1291, "step": 894 }, { "epoch": 0.9005025125628141, "grad_norm": 1.498679518699646, "learning_rate": 6.2257311975355e-05, "loss": 1.9726, "step": 896 }, { "epoch": 0.9025125628140703, "grad_norm": 1.647017002105713, "learning_rate": 6.209609477998338e-05, "loss": 1.9182, "step": 898 }, { "epoch": 0.9045226130653267, "grad_norm": 1.546736240386963, "learning_rate": 6.19347439001569e-05, "loss": 2.0869, "step": 900 }, { "epoch": 0.9065326633165829, "grad_norm": 1.4182671308517456, "learning_rate": 6.177326111910429e-05, "loss": 2.1178, "step": 902 }, { "epoch": 0.9085427135678392, "grad_norm": 1.685409426689148, "learning_rate": 6.161164822151213e-05, "loss": 2.1423, "step": 904 }, { "epoch": 0.9105527638190954, "grad_norm": 1.488713026046753, "learning_rate": 6.144990699350497e-05, "loss": 2.2579, "step": 906 }, { "epoch": 0.9125628140703518, "grad_norm": 1.6088262796401978, "learning_rate": 6.128803922262573e-05, "loss": 2.1192, "step": 908 }, { "epoch": 0.914572864321608, "grad_norm": 1.6858761310577393, "learning_rate": 6.112604669781572e-05, "loss": 1.9913, "step": 910 }, { "epoch": 0.9165829145728643, "grad_norm": 1.4918874502182007, "learning_rate": 6.096393120939516e-05, "loss": 2.0041, "step": 912 }, { "epoch": 0.9185929648241206, "grad_norm": 1.501802921295166, "learning_rate": 6.08016945490432e-05, "loss": 2.1627, "step": 914 }, { "epoch": 0.9206030150753769, "grad_norm": 1.6295005083084106, "learning_rate": 6.063933850977811e-05, "loss": 2.281, "step": 916 }, { "epoch": 0.9226130653266331, "grad_norm": 1.4057841300964355, "learning_rate": 6.04768648859376e-05, "loss": 2.0616, "step": 918 }, { "epoch": 0.9246231155778895, "grad_norm": 1.386192798614502, "learning_rate": 6.031427547315889e-05, "loss": 1.9191, "step": 920 }, { "epoch": 0.9266331658291457, "grad_norm": 1.4363282918930054, "learning_rate": 6.015157206835881e-05, "loss": 1.94, "step": 922 }, { "epoch": 0.928643216080402, "grad_norm": 1.6516571044921875, "learning_rate": 5.9988756469714135e-05, "loss": 2.3657, "step": 924 }, { "epoch": 0.9306532663316583, "grad_norm": 1.465562105178833, "learning_rate": 5.982583047664151e-05, "loss": 2.1271, "step": 926 }, { "epoch": 0.9326633165829146, "grad_norm": 1.5715078115463257, "learning_rate": 5.9662795889777666e-05, "loss": 2.1258, "step": 928 }, { "epoch": 0.9346733668341709, "grad_norm": 3.0183351039886475, "learning_rate": 5.949965451095951e-05, "loss": 2.2448, "step": 930 }, { "epoch": 0.9366834170854271, "grad_norm": 1.5689525604248047, "learning_rate": 5.933640814320417e-05, "loss": 2.161, "step": 932 }, { "epoch": 0.9386934673366835, "grad_norm": 1.5857359170913696, "learning_rate": 5.917305859068912e-05, "loss": 2.2529, "step": 934 }, { "epoch": 0.9407035175879397, "grad_norm": 1.4421300888061523, "learning_rate": 5.900960765873222e-05, "loss": 2.1033, "step": 936 }, { "epoch": 0.942713567839196, "grad_norm": 1.5819547176361084, "learning_rate": 5.8846057153771786e-05, "loss": 2.2916, "step": 938 }, { "epoch": 0.9447236180904522, "grad_norm": 1.4737604856491089, "learning_rate": 5.868240888334653e-05, "loss": 2.005, "step": 940 }, { "epoch": 0.9467336683417086, "grad_norm": 1.4487162828445435, "learning_rate": 5.8518664656075706e-05, "loss": 1.9979, "step": 942 }, { "epoch": 0.9487437185929648, "grad_norm": 1.5933785438537598, "learning_rate": 5.835482628163909e-05, "loss": 1.7228, "step": 944 }, { "epoch": 0.9507537688442211, "grad_norm": 1.6304993629455566, "learning_rate": 5.819089557075689e-05, "loss": 2.109, "step": 946 }, { "epoch": 0.9527638190954774, "grad_norm": 1.5096608400344849, "learning_rate": 5.802687433516989e-05, "loss": 1.9898, "step": 948 }, { "epoch": 0.9547738693467337, "grad_norm": 1.611997127532959, "learning_rate": 5.786276438761927e-05, "loss": 1.9248, "step": 950 }, { "epoch": 0.9567839195979899, "grad_norm": 1.6030073165893555, "learning_rate": 5.7698567541826675e-05, "loss": 2.0681, "step": 952 }, { "epoch": 0.9587939698492463, "grad_norm": 1.569161057472229, "learning_rate": 5.753428561247416e-05, "loss": 1.919, "step": 954 }, { "epoch": 0.9608040201005025, "grad_norm": 1.5151729583740234, "learning_rate": 5.7369920415184064e-05, "loss": 2.1896, "step": 956 }, { "epoch": 0.9628140703517588, "grad_norm": 1.6666862964630127, "learning_rate": 5.7205473766499005e-05, "loss": 2.0416, "step": 958 }, { "epoch": 0.964824120603015, "grad_norm": 1.5797557830810547, "learning_rate": 5.704094748386184e-05, "loss": 1.9272, "step": 960 }, { "epoch": 0.9668341708542714, "grad_norm": 1.508834719657898, "learning_rate": 5.6876343385595446e-05, "loss": 1.7583, "step": 962 }, { "epoch": 0.9688442211055276, "grad_norm": 1.4971197843551636, "learning_rate": 5.6711663290882776e-05, "loss": 1.9919, "step": 964 }, { "epoch": 0.9708542713567839, "grad_norm": 1.6074587106704712, "learning_rate": 5.6546909019746666e-05, "loss": 1.982, "step": 966 }, { "epoch": 0.9728643216080402, "grad_norm": 3.3276638984680176, "learning_rate": 5.6382082393029746e-05, "loss": 2.0718, "step": 968 }, { "epoch": 0.9748743718592965, "grad_norm": 1.4836559295654297, "learning_rate": 5.621718523237427e-05, "loss": 2.196, "step": 970 }, { "epoch": 0.9768844221105528, "grad_norm": 1.502540946006775, "learning_rate": 5.605221936020207e-05, "loss": 2.1674, "step": 972 }, { "epoch": 0.978894472361809, "grad_norm": 1.4632574319839478, "learning_rate": 5.588718659969438e-05, "loss": 2.2086, "step": 974 }, { "epoch": 0.9809045226130654, "grad_norm": 1.5661799907684326, "learning_rate": 5.57220887747716e-05, "loss": 2.0454, "step": 976 }, { "epoch": 0.9829145728643216, "grad_norm": 1.5829559564590454, "learning_rate": 5.5556927710073314e-05, "loss": 1.9525, "step": 978 }, { "epoch": 0.9849246231155779, "grad_norm": 1.6690107583999634, "learning_rate": 5.539170523093794e-05, "loss": 2.0355, "step": 980 }, { "epoch": 0.9869346733668342, "grad_norm": 1.524163842201233, "learning_rate": 5.522642316338268e-05, "loss": 2.0668, "step": 982 }, { "epoch": 0.9889447236180905, "grad_norm": 1.5150009393692017, "learning_rate": 5.506108333408329e-05, "loss": 2.1167, "step": 984 }, { "epoch": 0.9909547738693467, "grad_norm": 1.5068144798278809, "learning_rate": 5.489568757035391e-05, "loss": 2.0714, "step": 986 }, { "epoch": 0.992964824120603, "grad_norm": 1.5889198780059814, "learning_rate": 5.473023770012686e-05, "loss": 2.0012, "step": 988 }, { "epoch": 0.9949748743718593, "grad_norm": 1.4362760782241821, "learning_rate": 5.456473555193242e-05, "loss": 1.9559, "step": 990 }, { "epoch": 0.9969849246231156, "grad_norm": 1.4204431772232056, "learning_rate": 5.4399182954878656e-05, "loss": 1.8115, "step": 992 }, { "epoch": 0.9989949748743718, "grad_norm": 1.500132441520691, "learning_rate": 5.4233581738631165e-05, "loss": 2.2726, "step": 994 } ], "logging_steps": 2, "max_steps": 1990, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 995, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.6840475756671795e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }