|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.4980724749421743, |
|
"eval_steps": 500, |
|
"global_step": 972, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0015420200462606013, |
|
"grad_norm": 0.06960189342498779, |
|
"learning_rate": 1e-05, |
|
"loss": 0.9546, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0030840400925212026, |
|
"grad_norm": 0.07705054432153702, |
|
"learning_rate": 2e-05, |
|
"loss": 1.0023, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.004626060138781804, |
|
"grad_norm": 0.07115544378757477, |
|
"learning_rate": 3e-05, |
|
"loss": 0.9733, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.006168080185042405, |
|
"grad_norm": 0.06770255416631699, |
|
"learning_rate": 4e-05, |
|
"loss": 0.8908, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.007710100231303007, |
|
"grad_norm": 0.06454406678676605, |
|
"learning_rate": 5e-05, |
|
"loss": 0.9168, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.009252120277563608, |
|
"grad_norm": 0.0818350687623024, |
|
"learning_rate": 6e-05, |
|
"loss": 0.8822, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.01079414032382421, |
|
"grad_norm": 0.08504347503185272, |
|
"learning_rate": 7e-05, |
|
"loss": 0.9563, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.01233616037008481, |
|
"grad_norm": 0.08642569184303284, |
|
"learning_rate": 8e-05, |
|
"loss": 0.9324, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.013878180416345412, |
|
"grad_norm": 0.07200746238231659, |
|
"learning_rate": 9e-05, |
|
"loss": 0.8557, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.015420200462606014, |
|
"grad_norm": 0.0539001002907753, |
|
"learning_rate": 0.0001, |
|
"loss": 0.8316, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.016962220508866616, |
|
"grad_norm": 0.05532313138246536, |
|
"learning_rate": 0.00011000000000000002, |
|
"loss": 0.869, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.018504240555127217, |
|
"grad_norm": 0.04828835651278496, |
|
"learning_rate": 0.00012, |
|
"loss": 0.9004, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.020046260601387818, |
|
"grad_norm": 0.051899779587984085, |
|
"learning_rate": 0.00013000000000000002, |
|
"loss": 0.7818, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.02158828064764842, |
|
"grad_norm": 0.07032614946365356, |
|
"learning_rate": 0.00014, |
|
"loss": 0.7946, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.02313030069390902, |
|
"grad_norm": 0.08230099081993103, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 0.934, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.02467232074016962, |
|
"grad_norm": 0.08811169862747192, |
|
"learning_rate": 0.00016, |
|
"loss": 0.8621, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.026214340786430222, |
|
"grad_norm": 0.062236446887254715, |
|
"learning_rate": 0.00017, |
|
"loss": 0.8186, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.027756360832690823, |
|
"grad_norm": 0.052294306457042694, |
|
"learning_rate": 0.00018, |
|
"loss": 0.8124, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.029298380878951428, |
|
"grad_norm": 0.032340776175260544, |
|
"learning_rate": 0.00019, |
|
"loss": 0.8598, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.03084040092521203, |
|
"grad_norm": 0.028073711320757866, |
|
"learning_rate": 0.0002, |
|
"loss": 0.9087, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.03238242097147263, |
|
"grad_norm": 0.027820633724331856, |
|
"learning_rate": 0.00019999969691239107, |
|
"loss": 0.7093, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.03392444101773323, |
|
"grad_norm": 0.032430652529001236, |
|
"learning_rate": 0.00019999878765140146, |
|
"loss": 0.7825, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.03546646106399383, |
|
"grad_norm": 0.034685954451560974, |
|
"learning_rate": 0.000199997272222543, |
|
"loss": 0.7787, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.03700848111025443, |
|
"grad_norm": 0.033446088433265686, |
|
"learning_rate": 0.0001999951506350017, |
|
"loss": 0.8636, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.03855050115651504, |
|
"grad_norm": 0.026805371046066284, |
|
"learning_rate": 0.0001999924229016382, |
|
"loss": 0.904, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.040092521202775636, |
|
"grad_norm": 0.024178853258490562, |
|
"learning_rate": 0.0001999890890389873, |
|
"loss": 0.7428, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.04163454124903624, |
|
"grad_norm": 0.024074744433164597, |
|
"learning_rate": 0.00019998514906725804, |
|
"loss": 0.8562, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.04317656129529684, |
|
"grad_norm": 0.025082379579544067, |
|
"learning_rate": 0.0001999806030103336, |
|
"loss": 0.8078, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.04471858134155744, |
|
"grad_norm": 0.027726231142878532, |
|
"learning_rate": 0.00019997545089577102, |
|
"loss": 0.7133, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.04626060138781804, |
|
"grad_norm": 0.02736576274037361, |
|
"learning_rate": 0.00019996969275480115, |
|
"loss": 0.7826, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.047802621434078645, |
|
"grad_norm": 0.0234544537961483, |
|
"learning_rate": 0.0001999633286223284, |
|
"loss": 0.7424, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.04934464148033924, |
|
"grad_norm": 0.025730784982442856, |
|
"learning_rate": 0.00019995635853693058, |
|
"loss": 0.7584, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.05088666152659985, |
|
"grad_norm": 0.02729860506951809, |
|
"learning_rate": 0.00019994878254085861, |
|
"loss": 0.73, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.052428681572860444, |
|
"grad_norm": 0.027428491041064262, |
|
"learning_rate": 0.00019994060068003627, |
|
"loss": 0.854, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.05397070161912105, |
|
"grad_norm": 0.024362564086914062, |
|
"learning_rate": 0.00019993181300406005, |
|
"loss": 0.7441, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.05551272166538165, |
|
"grad_norm": 0.024825185537338257, |
|
"learning_rate": 0.00019992241956619863, |
|
"loss": 0.7367, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.05705474171164225, |
|
"grad_norm": 0.023908289149403572, |
|
"learning_rate": 0.00019991242042339264, |
|
"loss": 0.8266, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.058596761757902856, |
|
"grad_norm": 0.023762725293636322, |
|
"learning_rate": 0.00019990181563625447, |
|
"loss": 0.829, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.06013878180416345, |
|
"grad_norm": 0.02654297836124897, |
|
"learning_rate": 0.0001998906052690677, |
|
"loss": 0.7635, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.06168080185042406, |
|
"grad_norm": 0.026318082585930824, |
|
"learning_rate": 0.00019987878938978684, |
|
"loss": 0.7623, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.06322282189668466, |
|
"grad_norm": 0.024309197440743446, |
|
"learning_rate": 0.00019986636807003673, |
|
"loss": 0.7494, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.06476484194294525, |
|
"grad_norm": 0.022980719804763794, |
|
"learning_rate": 0.00019985334138511237, |
|
"loss": 0.748, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.06630686198920586, |
|
"grad_norm": 0.023796193301677704, |
|
"learning_rate": 0.00019983970941397835, |
|
"loss": 0.8684, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.06784888203546646, |
|
"grad_norm": 0.02435590885579586, |
|
"learning_rate": 0.00019982547223926824, |
|
"loss": 0.6749, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.06939090208172706, |
|
"grad_norm": 0.026482658460736275, |
|
"learning_rate": 0.0001998106299472843, |
|
"loss": 0.7684, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.07093292212798766, |
|
"grad_norm": 0.025558117777109146, |
|
"learning_rate": 0.0001997951826279968, |
|
"loss": 0.7507, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.07247494217424827, |
|
"grad_norm": 0.027371902018785477, |
|
"learning_rate": 0.00019977913037504355, |
|
"loss": 0.7377, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.07401696222050887, |
|
"grad_norm": 0.0234097708016634, |
|
"learning_rate": 0.00019976247328572938, |
|
"loss": 0.7675, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.07555898226676946, |
|
"grad_norm": 0.0241215992718935, |
|
"learning_rate": 0.00019974521146102537, |
|
"loss": 0.8079, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.07710100231303008, |
|
"grad_norm": 0.025045258924365044, |
|
"learning_rate": 0.00019972734500556846, |
|
"loss": 0.8407, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.07864302235929067, |
|
"grad_norm": 0.02574036829173565, |
|
"learning_rate": 0.0001997088740276607, |
|
"loss": 0.767, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.08018504240555127, |
|
"grad_norm": 0.02342085726559162, |
|
"learning_rate": 0.00019968979863926856, |
|
"loss": 0.7373, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.08172706245181187, |
|
"grad_norm": 0.023467406630516052, |
|
"learning_rate": 0.0001996701189560223, |
|
"loss": 0.6529, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.08326908249807248, |
|
"grad_norm": 0.02330499142408371, |
|
"learning_rate": 0.00019964983509721527, |
|
"loss": 0.7347, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.08481110254433308, |
|
"grad_norm": 0.027940964326262474, |
|
"learning_rate": 0.00019962894718580324, |
|
"loss": 0.9313, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.08635312259059368, |
|
"grad_norm": 0.024525761604309082, |
|
"learning_rate": 0.00019960745534840354, |
|
"loss": 0.767, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.08789514263685427, |
|
"grad_norm": 0.024538526311516762, |
|
"learning_rate": 0.00019958535971529434, |
|
"loss": 0.7659, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.08943716268311488, |
|
"grad_norm": 0.02253701537847519, |
|
"learning_rate": 0.00019956266042041394, |
|
"loss": 0.7805, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.09097918272937548, |
|
"grad_norm": 0.023676637560129166, |
|
"learning_rate": 0.0001995393576013598, |
|
"loss": 0.7894, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.09252120277563608, |
|
"grad_norm": 0.022134529426693916, |
|
"learning_rate": 0.0001995154513993878, |
|
"loss": 0.7492, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.09406322282189669, |
|
"grad_norm": 0.02350509911775589, |
|
"learning_rate": 0.00019949094195941152, |
|
"loss": 0.6902, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.09560524286815729, |
|
"grad_norm": 0.02464171312749386, |
|
"learning_rate": 0.00019946582943000102, |
|
"loss": 0.7836, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.09714726291441789, |
|
"grad_norm": 0.023095758631825447, |
|
"learning_rate": 0.00019944011396338222, |
|
"loss": 0.8321, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.09868928296067848, |
|
"grad_norm": 0.026240425184369087, |
|
"learning_rate": 0.00019941379571543596, |
|
"loss": 0.8461, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.1002313030069391, |
|
"grad_norm": 0.02174345962703228, |
|
"learning_rate": 0.00019938687484569693, |
|
"loss": 0.6388, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.1017733230531997, |
|
"grad_norm": 0.02867325395345688, |
|
"learning_rate": 0.00019935935151735277, |
|
"loss": 0.826, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.10331534309946029, |
|
"grad_norm": 0.02631618268787861, |
|
"learning_rate": 0.00019933122589724302, |
|
"loss": 0.9407, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.10485736314572089, |
|
"grad_norm": 0.022020747885107994, |
|
"learning_rate": 0.0001993024981558583, |
|
"loss": 0.7004, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.1063993831919815, |
|
"grad_norm": 0.02743780054152012, |
|
"learning_rate": 0.000199273168467339, |
|
"loss": 0.7607, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.1079414032382421, |
|
"grad_norm": 0.028378015384078026, |
|
"learning_rate": 0.00019924323700947448, |
|
"loss": 0.8604, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.1094834232845027, |
|
"grad_norm": 0.0275627039372921, |
|
"learning_rate": 0.00019921270396370172, |
|
"loss": 0.835, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.1110254433307633, |
|
"grad_norm": 0.023099975660443306, |
|
"learning_rate": 0.0001991815695151046, |
|
"loss": 0.7028, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.1125674633770239, |
|
"grad_norm": 0.028545403853058815, |
|
"learning_rate": 0.00019914983385241236, |
|
"loss": 0.8248, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.1141094834232845, |
|
"grad_norm": 0.02746577188372612, |
|
"learning_rate": 0.00019911749716799873, |
|
"loss": 0.7309, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.1156515034695451, |
|
"grad_norm": 0.024899670854210854, |
|
"learning_rate": 0.00019908455965788067, |
|
"loss": 0.7473, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.11719352351580571, |
|
"grad_norm": 0.024973087012767792, |
|
"learning_rate": 0.00019905102152171727, |
|
"loss": 0.8362, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.11873554356206631, |
|
"grad_norm": 0.023668723180890083, |
|
"learning_rate": 0.0001990168829628083, |
|
"loss": 0.7677, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.1202775636083269, |
|
"grad_norm": 0.02495860867202282, |
|
"learning_rate": 0.0001989821441880933, |
|
"loss": 0.7341, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.1218195836545875, |
|
"grad_norm": 0.02537156455218792, |
|
"learning_rate": 0.00019894680540815006, |
|
"loss": 0.6767, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.12336160370084812, |
|
"grad_norm": 0.0246786717325449, |
|
"learning_rate": 0.0001989108668371936, |
|
"loss": 0.7959, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.12490362374710871, |
|
"grad_norm": 0.02471376582980156, |
|
"learning_rate": 0.00019887432869307458, |
|
"loss": 0.6787, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.1264456437933693, |
|
"grad_norm": 0.025275586172938347, |
|
"learning_rate": 0.00019883719119727816, |
|
"loss": 0.7753, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.12798766383962992, |
|
"grad_norm": 0.021094506606459618, |
|
"learning_rate": 0.00019879945457492267, |
|
"loss": 0.758, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.1295296838858905, |
|
"grad_norm": 0.02534683421254158, |
|
"learning_rate": 0.00019876111905475815, |
|
"loss": 0.818, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.13107170393215112, |
|
"grad_norm": 0.024033140391111374, |
|
"learning_rate": 0.00019872218486916498, |
|
"loss": 0.775, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.13261372397841173, |
|
"grad_norm": 0.023884933441877365, |
|
"learning_rate": 0.00019868265225415265, |
|
"loss": 0.7918, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.1341557440246723, |
|
"grad_norm": 0.026067111641168594, |
|
"learning_rate": 0.00019864252144935794, |
|
"loss": 0.8368, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.13569776407093292, |
|
"grad_norm": 0.02631264552474022, |
|
"learning_rate": 0.00019860179269804394, |
|
"loss": 0.7138, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.13723978411719354, |
|
"grad_norm": 0.022870918735861778, |
|
"learning_rate": 0.00019856046624709822, |
|
"loss": 0.7423, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.13878180416345412, |
|
"grad_norm": 0.02922765538096428, |
|
"learning_rate": 0.00019851854234703145, |
|
"loss": 0.7498, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.14032382420971473, |
|
"grad_norm": 0.02589617855846882, |
|
"learning_rate": 0.00019847602125197598, |
|
"loss": 0.7438, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.14186584425597532, |
|
"grad_norm": 0.02394738420844078, |
|
"learning_rate": 0.00019843290321968412, |
|
"loss": 0.7094, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.14340786430223593, |
|
"grad_norm": 0.02237016148865223, |
|
"learning_rate": 0.0001983891885115267, |
|
"loss": 0.7868, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.14494988434849654, |
|
"grad_norm": 0.028733767569065094, |
|
"learning_rate": 0.00019834487739249146, |
|
"loss": 0.8178, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.14649190439475712, |
|
"grad_norm": 0.023086342960596085, |
|
"learning_rate": 0.0001982999701311814, |
|
"loss": 0.8368, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.14803392444101773, |
|
"grad_norm": 0.025624489411711693, |
|
"learning_rate": 0.0001982544669998132, |
|
"loss": 0.7731, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.14957594448727835, |
|
"grad_norm": 0.028302457183599472, |
|
"learning_rate": 0.0001982083682742156, |
|
"loss": 0.7942, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.15111796453353893, |
|
"grad_norm": 0.025008324533700943, |
|
"learning_rate": 0.00019816167423382765, |
|
"loss": 0.7393, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.15265998457979954, |
|
"grad_norm": 0.026291735470294952, |
|
"learning_rate": 0.00019811438516169702, |
|
"loss": 0.8016, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.15420200462606015, |
|
"grad_norm": 0.031547173857688904, |
|
"learning_rate": 0.00019806650134447838, |
|
"loss": 0.8597, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.15574402467232074, |
|
"grad_norm": 0.024978285655379295, |
|
"learning_rate": 0.00019801802307243153, |
|
"loss": 0.8182, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.15728604471858135, |
|
"grad_norm": 0.023977672681212425, |
|
"learning_rate": 0.00019796895063941978, |
|
"loss": 0.8374, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.15882806476484193, |
|
"grad_norm": 0.025743963196873665, |
|
"learning_rate": 0.000197919284342908, |
|
"loss": 0.7326, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.16037008481110254, |
|
"grad_norm": 0.02554011158645153, |
|
"learning_rate": 0.00019786902448396104, |
|
"loss": 0.7703, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.16191210485736315, |
|
"grad_norm": 0.027971483767032623, |
|
"learning_rate": 0.00019781817136724165, |
|
"loss": 0.6845, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.16345412490362374, |
|
"grad_norm": 0.02789183147251606, |
|
"learning_rate": 0.00019776672530100886, |
|
"loss": 0.838, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.16499614494988435, |
|
"grad_norm": 0.029534442350268364, |
|
"learning_rate": 0.00019771468659711595, |
|
"loss": 0.7428, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.16653816499614496, |
|
"grad_norm": 0.02910265140235424, |
|
"learning_rate": 0.00019766205557100868, |
|
"loss": 0.8891, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.16808018504240554, |
|
"grad_norm": 0.02285209856927395, |
|
"learning_rate": 0.00019760883254172327, |
|
"loss": 0.7425, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.16962220508866616, |
|
"grad_norm": 0.025452135130763054, |
|
"learning_rate": 0.0001975550178318845, |
|
"loss": 0.7617, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.17116422513492677, |
|
"grad_norm": 0.02843882516026497, |
|
"learning_rate": 0.00019750061176770385, |
|
"loss": 0.9045, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.17270624518118735, |
|
"grad_norm": 0.026800749823451042, |
|
"learning_rate": 0.00019744561467897735, |
|
"loss": 0.72, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.17424826522744796, |
|
"grad_norm": 0.021131988614797592, |
|
"learning_rate": 0.00019739002689908377, |
|
"loss": 0.6511, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.17579028527370855, |
|
"grad_norm": 0.02588481456041336, |
|
"learning_rate": 0.00019733384876498245, |
|
"loss": 0.7168, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.17733230531996916, |
|
"grad_norm": 0.031303439289331436, |
|
"learning_rate": 0.00019727708061721133, |
|
"loss": 0.8685, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.17887432536622977, |
|
"grad_norm": 0.02867058850824833, |
|
"learning_rate": 0.00019721972279988477, |
|
"loss": 0.655, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.18041634541249035, |
|
"grad_norm": 0.03866586834192276, |
|
"learning_rate": 0.00019716177566069174, |
|
"loss": 0.7957, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.18195836545875096, |
|
"grad_norm": 0.029205329716205597, |
|
"learning_rate": 0.00019710323955089343, |
|
"loss": 0.7617, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.18350038550501158, |
|
"grad_norm": 0.024928180500864983, |
|
"learning_rate": 0.00019704411482532116, |
|
"loss": 0.6982, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.18504240555127216, |
|
"grad_norm": 0.02545573003590107, |
|
"learning_rate": 0.0001969844018423744, |
|
"loss": 0.6067, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.18658442559753277, |
|
"grad_norm": 0.02810928039252758, |
|
"learning_rate": 0.0001969241009640185, |
|
"loss": 0.7112, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.18812644564379338, |
|
"grad_norm": 0.022296108305454254, |
|
"learning_rate": 0.00019686321255578238, |
|
"loss": 0.6598, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.18966846569005397, |
|
"grad_norm": 0.02429027482867241, |
|
"learning_rate": 0.00019680173698675648, |
|
"loss": 0.6381, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.19121048573631458, |
|
"grad_norm": 0.027076730504631996, |
|
"learning_rate": 0.0001967396746295905, |
|
"loss": 0.717, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.19275250578257516, |
|
"grad_norm": 0.02401566132903099, |
|
"learning_rate": 0.00019667702586049108, |
|
"loss": 0.8002, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.19429452582883577, |
|
"grad_norm": 0.024678878486156464, |
|
"learning_rate": 0.00019661379105921948, |
|
"loss": 0.7834, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.19583654587509639, |
|
"grad_norm": 0.029240388423204422, |
|
"learning_rate": 0.00019654997060908946, |
|
"loss": 0.8793, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.19737856592135697, |
|
"grad_norm": 0.02550147846341133, |
|
"learning_rate": 0.0001964855648969647, |
|
"loss": 0.6742, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.19892058596761758, |
|
"grad_norm": 0.02416900172829628, |
|
"learning_rate": 0.00019642057431325672, |
|
"loss": 0.7728, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.2004626060138782, |
|
"grad_norm": 0.024728331714868546, |
|
"learning_rate": 0.0001963549992519223, |
|
"loss": 0.7237, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.20200462606013878, |
|
"grad_norm": 0.025203561410307884, |
|
"learning_rate": 0.00019628884011046123, |
|
"loss": 0.7491, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.2035466461063994, |
|
"grad_norm": 0.02104656957089901, |
|
"learning_rate": 0.00019622209728991383, |
|
"loss": 0.7324, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.20508866615265997, |
|
"grad_norm": 0.03518475592136383, |
|
"learning_rate": 0.00019615477119485855, |
|
"loss": 0.8982, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.20663068619892058, |
|
"grad_norm": 0.026010941714048386, |
|
"learning_rate": 0.00019608686223340945, |
|
"loss": 0.7451, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.2081727062451812, |
|
"grad_norm": 0.025345437228679657, |
|
"learning_rate": 0.00019601837081721386, |
|
"loss": 0.705, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.20971472629144178, |
|
"grad_norm": 0.02374056540429592, |
|
"learning_rate": 0.00019594929736144976, |
|
"loss": 0.7307, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.2112567463377024, |
|
"grad_norm": 0.026990080252289772, |
|
"learning_rate": 0.00019587964228482332, |
|
"loss": 0.8173, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.212798766383963, |
|
"grad_norm": 0.026662928983569145, |
|
"learning_rate": 0.00019580940600956638, |
|
"loss": 0.818, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.21434078643022358, |
|
"grad_norm": 0.026762284338474274, |
|
"learning_rate": 0.00019573858896143376, |
|
"loss": 0.6674, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.2158828064764842, |
|
"grad_norm": 0.02683679386973381, |
|
"learning_rate": 0.00019566719156970095, |
|
"loss": 0.7569, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.2174248265227448, |
|
"grad_norm": 0.022990800440311432, |
|
"learning_rate": 0.00019559521426716118, |
|
"loss": 0.7051, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.2189668465690054, |
|
"grad_norm": 0.022913858294487, |
|
"learning_rate": 0.00019552265749012303, |
|
"loss": 0.7483, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.220508866615266, |
|
"grad_norm": 0.025732524693012238, |
|
"learning_rate": 0.00019544952167840777, |
|
"loss": 0.8484, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.2220508866615266, |
|
"grad_norm": 0.024254556745290756, |
|
"learning_rate": 0.00019537580727534644, |
|
"loss": 0.747, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.2235929067077872, |
|
"grad_norm": 0.02940620854496956, |
|
"learning_rate": 0.0001953015147277776, |
|
"loss": 0.8204, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.2251349267540478, |
|
"grad_norm": 0.02590208128094673, |
|
"learning_rate": 0.00019522664448604418, |
|
"loss": 0.7778, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.2266769468003084, |
|
"grad_norm": 0.02748725563287735, |
|
"learning_rate": 0.00019515119700399107, |
|
"loss": 0.7782, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.228218966846569, |
|
"grad_norm": 0.024179786443710327, |
|
"learning_rate": 0.00019507517273896222, |
|
"loss": 0.7651, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.22976098689282962, |
|
"grad_norm": 0.02463974617421627, |
|
"learning_rate": 0.00019499857215179786, |
|
"loss": 0.8215, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.2313030069390902, |
|
"grad_norm": 0.024559814482927322, |
|
"learning_rate": 0.00019492139570683178, |
|
"loss": 0.7539, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2328450269853508, |
|
"grad_norm": 0.023420870304107666, |
|
"learning_rate": 0.00019484364387188847, |
|
"loss": 0.7035, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.23438704703161142, |
|
"grad_norm": 0.026096729561686516, |
|
"learning_rate": 0.00019476531711828027, |
|
"loss": 0.8033, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.235929067077872, |
|
"grad_norm": 0.02388446033000946, |
|
"learning_rate": 0.0001946864159208045, |
|
"loss": 0.6746, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.23747108712413262, |
|
"grad_norm": 0.02306438237428665, |
|
"learning_rate": 0.0001946069407577408, |
|
"loss": 0.7062, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.2390131071703932, |
|
"grad_norm": 0.024697955697774887, |
|
"learning_rate": 0.00019452689211084775, |
|
"loss": 0.7691, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.2405551272166538, |
|
"grad_norm": 0.026947690173983574, |
|
"learning_rate": 0.00019444627046536056, |
|
"loss": 0.7347, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.24209714726291442, |
|
"grad_norm": 0.02345297671854496, |
|
"learning_rate": 0.00019436507630998757, |
|
"loss": 0.745, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.243639167309175, |
|
"grad_norm": 0.029198188334703445, |
|
"learning_rate": 0.00019428331013690765, |
|
"loss": 0.7862, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.24518118735543562, |
|
"grad_norm": 0.025465266779065132, |
|
"learning_rate": 0.00019420097244176706, |
|
"loss": 0.6685, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.24672320740169623, |
|
"grad_norm": 0.02786502055823803, |
|
"learning_rate": 0.00019411806372367655, |
|
"loss": 0.725, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.24826522744795682, |
|
"grad_norm": 0.02317357063293457, |
|
"learning_rate": 0.0001940345844852082, |
|
"loss": 0.7075, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.24980724749421743, |
|
"grad_norm": 0.02511444129049778, |
|
"learning_rate": 0.00019395053523239245, |
|
"loss": 0.7102, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.25134926754047804, |
|
"grad_norm": 0.02567203901708126, |
|
"learning_rate": 0.00019386591647471506, |
|
"loss": 0.8113, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.2528912875867386, |
|
"grad_norm": 0.02611825056374073, |
|
"learning_rate": 0.00019378072872511398, |
|
"loss": 0.786, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.2544333076329992, |
|
"grad_norm": 0.02788010984659195, |
|
"learning_rate": 0.0001936949724999762, |
|
"loss": 0.684, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.25597532767925985, |
|
"grad_norm": 0.026200013235211372, |
|
"learning_rate": 0.0001936086483191347, |
|
"loss": 0.7563, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.25751734772552043, |
|
"grad_norm": 0.024887658655643463, |
|
"learning_rate": 0.00019352175670586533, |
|
"loss": 0.6896, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.259059367771781, |
|
"grad_norm": 0.027471961453557014, |
|
"learning_rate": 0.00019343429818688347, |
|
"loss": 0.8109, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.26060138781804165, |
|
"grad_norm": 0.024350160732865334, |
|
"learning_rate": 0.00019334627329234102, |
|
"loss": 0.7178, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.26214340786430224, |
|
"grad_norm": 0.02346990443766117, |
|
"learning_rate": 0.00019325768255582302, |
|
"loss": 0.6508, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.2636854279105628, |
|
"grad_norm": 0.028655072674155235, |
|
"learning_rate": 0.00019316852651434462, |
|
"loss": 0.7036, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.26522744795682346, |
|
"grad_norm": 0.024629781022667885, |
|
"learning_rate": 0.0001930788057083476, |
|
"loss": 0.774, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.26676946800308404, |
|
"grad_norm": 0.023107299581170082, |
|
"learning_rate": 0.0001929885206816973, |
|
"loss": 0.7021, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.2683114880493446, |
|
"grad_norm": 0.022794177755713463, |
|
"learning_rate": 0.00019289767198167916, |
|
"loss": 0.7469, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.26985350809560527, |
|
"grad_norm": 0.026097161695361137, |
|
"learning_rate": 0.00019280626015899546, |
|
"loss": 0.7325, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.27139552814186585, |
|
"grad_norm": 0.029879910871386528, |
|
"learning_rate": 0.00019271428576776205, |
|
"loss": 0.7614, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.27293754818812643, |
|
"grad_norm": 0.026986606419086456, |
|
"learning_rate": 0.00019262174936550487, |
|
"loss": 0.7718, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.2744795682343871, |
|
"grad_norm": 0.025835467502474785, |
|
"learning_rate": 0.00019252865151315665, |
|
"loss": 0.7511, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.27602158828064766, |
|
"grad_norm": 0.028101902455091476, |
|
"learning_rate": 0.00019243499277505355, |
|
"loss": 0.8136, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.27756360832690824, |
|
"grad_norm": 0.028153471648693085, |
|
"learning_rate": 0.00019234077371893155, |
|
"loss": 0.8798, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.2791056283731688, |
|
"grad_norm": 0.024540267884731293, |
|
"learning_rate": 0.0001922459949159233, |
|
"loss": 0.7854, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.28064764841942946, |
|
"grad_norm": 0.023485183715820312, |
|
"learning_rate": 0.00019215065694055437, |
|
"loss": 0.6655, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.28218966846569005, |
|
"grad_norm": 0.023394625633955002, |
|
"learning_rate": 0.00019205476037073997, |
|
"loss": 0.759, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.28373168851195063, |
|
"grad_norm": 0.025181008502840996, |
|
"learning_rate": 0.00019195830578778132, |
|
"loss": 0.7649, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.28527370855821127, |
|
"grad_norm": 0.023696815595030785, |
|
"learning_rate": 0.0001918612937763622, |
|
"loss": 0.7469, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.28681572860447185, |
|
"grad_norm": 0.025794658809900284, |
|
"learning_rate": 0.00019176372492454537, |
|
"loss": 0.7623, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.28835774865073244, |
|
"grad_norm": 0.02523699589073658, |
|
"learning_rate": 0.00019166559982376904, |
|
"loss": 0.6621, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.2898997686969931, |
|
"grad_norm": 0.02426300384104252, |
|
"learning_rate": 0.00019156691906884325, |
|
"loss": 0.748, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.29144178874325366, |
|
"grad_norm": 0.024227775633335114, |
|
"learning_rate": 0.0001914676832579463, |
|
"loss": 0.724, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.29298380878951424, |
|
"grad_norm": 0.031684551388025284, |
|
"learning_rate": 0.00019136789299262108, |
|
"loss": 0.8939, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.2945258288357749, |
|
"grad_norm": 0.023766616359353065, |
|
"learning_rate": 0.0001912675488777714, |
|
"loss": 0.7179, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.29606784888203547, |
|
"grad_norm": 0.02463400922715664, |
|
"learning_rate": 0.0001911666515216585, |
|
"loss": 0.8202, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.29760986892829605, |
|
"grad_norm": 0.0225905179977417, |
|
"learning_rate": 0.00019106520153589708, |
|
"loss": 0.6357, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.2991518889745567, |
|
"grad_norm": 0.022882292047142982, |
|
"learning_rate": 0.00019096319953545185, |
|
"loss": 0.7506, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.3006939090208173, |
|
"grad_norm": 0.027596216648817062, |
|
"learning_rate": 0.00019086064613863364, |
|
"loss": 0.7669, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.30223592906707786, |
|
"grad_norm": 0.030367175117135048, |
|
"learning_rate": 0.00019075754196709572, |
|
"loss": 0.7792, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.3037779491133385, |
|
"grad_norm": 0.025017013773322105, |
|
"learning_rate": 0.00019065388764583004, |
|
"loss": 0.6977, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.3053199691595991, |
|
"grad_norm": 0.02788584679365158, |
|
"learning_rate": 0.0001905496838031634, |
|
"loss": 0.6871, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.30686198920585966, |
|
"grad_norm": 0.032745130360126495, |
|
"learning_rate": 0.00019044493107075368, |
|
"loss": 0.8934, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.3084040092521203, |
|
"grad_norm": 0.027039945125579834, |
|
"learning_rate": 0.00019033963008358598, |
|
"loss": 0.6522, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.3099460292983809, |
|
"grad_norm": 0.03149978816509247, |
|
"learning_rate": 0.0001902337814799688, |
|
"loss": 0.844, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.31148804934464147, |
|
"grad_norm": 0.024369308724999428, |
|
"learning_rate": 0.0001901273859015301, |
|
"loss": 0.7202, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.31303006939090205, |
|
"grad_norm": 0.05448361113667488, |
|
"learning_rate": 0.00019002044399321356, |
|
"loss": 0.8301, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.3145720894371627, |
|
"grad_norm": 0.02388385497033596, |
|
"learning_rate": 0.0001899129564032745, |
|
"loss": 0.7105, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.3161141094834233, |
|
"grad_norm": 0.02488291636109352, |
|
"learning_rate": 0.00018980492378327607, |
|
"loss": 0.7393, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.31765612952968386, |
|
"grad_norm": 0.023874662816524506, |
|
"learning_rate": 0.00018969634678808522, |
|
"loss": 0.6791, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.3191981495759445, |
|
"grad_norm": 0.030418075621128082, |
|
"learning_rate": 0.0001895872260758688, |
|
"loss": 0.75, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.3207401696222051, |
|
"grad_norm": 0.02990088053047657, |
|
"learning_rate": 0.00018947756230808954, |
|
"loss": 0.7986, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.32228218966846567, |
|
"grad_norm": 0.027980022132396698, |
|
"learning_rate": 0.00018936735614950197, |
|
"loss": 0.7054, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.3238242097147263, |
|
"grad_norm": 0.026269223541021347, |
|
"learning_rate": 0.00018925660826814856, |
|
"loss": 0.8195, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.3253662297609869, |
|
"grad_norm": 0.025045178830623627, |
|
"learning_rate": 0.0001891453193353555, |
|
"loss": 0.7221, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.3269082498072475, |
|
"grad_norm": 0.021175356581807137, |
|
"learning_rate": 0.00018903349002572873, |
|
"loss": 0.6513, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.3284502698535081, |
|
"grad_norm": 0.023593388497829437, |
|
"learning_rate": 0.0001889211210171498, |
|
"loss": 0.7405, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.3299922898997687, |
|
"grad_norm": 0.027270464226603508, |
|
"learning_rate": 0.00018880821299077183, |
|
"loss": 0.7184, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.3315343099460293, |
|
"grad_norm": 0.025203121826052666, |
|
"learning_rate": 0.00018869476663101523, |
|
"loss": 0.6659, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.3330763299922899, |
|
"grad_norm": 0.024924185127019882, |
|
"learning_rate": 0.0001885807826255638, |
|
"loss": 0.7412, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.3346183500385505, |
|
"grad_norm": 0.022862501442432404, |
|
"learning_rate": 0.00018846626166536026, |
|
"loss": 0.6984, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.3361603700848111, |
|
"grad_norm": 0.022781461477279663, |
|
"learning_rate": 0.0001883512044446023, |
|
"loss": 0.6374, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.33770239013107173, |
|
"grad_norm": 0.023618346080183983, |
|
"learning_rate": 0.0001882356116607383, |
|
"loss": 0.6948, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.3392444101773323, |
|
"grad_norm": 0.02586747333407402, |
|
"learning_rate": 0.0001881194840144631, |
|
"loss": 0.7682, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.3407864302235929, |
|
"grad_norm": 0.026834698393940926, |
|
"learning_rate": 0.00018800282220971366, |
|
"loss": 0.7546, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.34232845026985353, |
|
"grad_norm": 0.028564658015966415, |
|
"learning_rate": 0.00018788562695366495, |
|
"loss": 0.8267, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.3438704703161141, |
|
"grad_norm": 0.02490355260670185, |
|
"learning_rate": 0.00018776789895672558, |
|
"loss": 0.6937, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.3454124903623747, |
|
"grad_norm": 0.03545152395963669, |
|
"learning_rate": 0.00018764963893253347, |
|
"loss": 0.7001, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.3469545104086353, |
|
"grad_norm": 0.02449451945722103, |
|
"learning_rate": 0.00018753084759795158, |
|
"loss": 0.8165, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.3484965304548959, |
|
"grad_norm": 0.027851196005940437, |
|
"learning_rate": 0.00018741152567306355, |
|
"loss": 0.7196, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.3500385505011565, |
|
"grad_norm": 0.02707446552813053, |
|
"learning_rate": 0.00018729167388116934, |
|
"loss": 0.8375, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.3515805705474171, |
|
"grad_norm": 0.02902469038963318, |
|
"learning_rate": 0.00018717129294878074, |
|
"loss": 0.6744, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.35312259059367773, |
|
"grad_norm": 0.031537748873233795, |
|
"learning_rate": 0.0001870503836056172, |
|
"loss": 0.7552, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.3546646106399383, |
|
"grad_norm": 0.026265786960721016, |
|
"learning_rate": 0.00018692894658460117, |
|
"loss": 0.7551, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.3562066306861989, |
|
"grad_norm": 0.02483406662940979, |
|
"learning_rate": 0.0001868069826218538, |
|
"loss": 0.7233, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.35774865073245954, |
|
"grad_norm": 0.024800019338726997, |
|
"learning_rate": 0.0001866844924566904, |
|
"loss": 0.7371, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.3592906707787201, |
|
"grad_norm": 0.02515244670212269, |
|
"learning_rate": 0.00018656147683161593, |
|
"loss": 0.7621, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.3608326908249807, |
|
"grad_norm": 0.02592633105814457, |
|
"learning_rate": 0.00018643793649232072, |
|
"loss": 0.7539, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.36237471087124135, |
|
"grad_norm": 0.0275077186524868, |
|
"learning_rate": 0.00018631387218767561, |
|
"loss": 0.6925, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.36391673091750193, |
|
"grad_norm": 0.027163324877619743, |
|
"learning_rate": 0.00018618928466972775, |
|
"loss": 0.7867, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.3654587509637625, |
|
"grad_norm": 0.026956308633089066, |
|
"learning_rate": 0.0001860641746936957, |
|
"loss": 0.7813, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.36700077101002315, |
|
"grad_norm": 0.02884814888238907, |
|
"learning_rate": 0.0001859385430179652, |
|
"loss": 0.7366, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.36854279105628374, |
|
"grad_norm": 0.025071945041418076, |
|
"learning_rate": 0.00018581239040408432, |
|
"loss": 0.708, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.3700848111025443, |
|
"grad_norm": 0.032973822206258774, |
|
"learning_rate": 0.00018568571761675893, |
|
"loss": 0.6544, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.37162683114880496, |
|
"grad_norm": 0.02571587637066841, |
|
"learning_rate": 0.0001855585254238481, |
|
"loss": 0.7633, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.37316885119506554, |
|
"grad_norm": 0.027229083701968193, |
|
"learning_rate": 0.00018543081459635935, |
|
"loss": 0.7752, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.3747108712413261, |
|
"grad_norm": 0.022508805617690086, |
|
"learning_rate": 0.00018530258590844409, |
|
"loss": 0.6437, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.37625289128758677, |
|
"grad_norm": 0.026772433891892433, |
|
"learning_rate": 0.00018517384013739285, |
|
"loss": 0.805, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.37779491133384735, |
|
"grad_norm": 0.023964572697877884, |
|
"learning_rate": 0.00018504457806363056, |
|
"loss": 0.7378, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.37933693138010793, |
|
"grad_norm": 0.02789299376308918, |
|
"learning_rate": 0.0001849148004707119, |
|
"loss": 0.772, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.3808789514263685, |
|
"grad_norm": 0.031168216839432716, |
|
"learning_rate": 0.00018478450814531647, |
|
"loss": 0.8299, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.38242097147262916, |
|
"grad_norm": 0.03058604896068573, |
|
"learning_rate": 0.00018465370187724408, |
|
"loss": 0.694, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.38396299151888974, |
|
"grad_norm": 0.028347650542855263, |
|
"learning_rate": 0.0001845223824594099, |
|
"loss": 0.7373, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.3855050115651503, |
|
"grad_norm": 0.025979626923799515, |
|
"learning_rate": 0.00018439055068783966, |
|
"loss": 0.8036, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.38704703161141096, |
|
"grad_norm": 0.029867777600884438, |
|
"learning_rate": 0.0001842582073616649, |
|
"loss": 0.7655, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.38858905165767155, |
|
"grad_norm": 0.025117915123701096, |
|
"learning_rate": 0.00018412535328311814, |
|
"loss": 0.7532, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.39013107170393213, |
|
"grad_norm": 0.023947982117533684, |
|
"learning_rate": 0.00018399198925752778, |
|
"loss": 0.6967, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.39167309175019277, |
|
"grad_norm": 0.025846531614661217, |
|
"learning_rate": 0.00018385811609331352, |
|
"loss": 0.7382, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.39321511179645335, |
|
"grad_norm": 0.025034697726368904, |
|
"learning_rate": 0.00018372373460198138, |
|
"loss": 0.7282, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.39475713184271394, |
|
"grad_norm": 0.02547437883913517, |
|
"learning_rate": 0.00018358884559811856, |
|
"loss": 0.7447, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.3962991518889746, |
|
"grad_norm": 0.027032596990466118, |
|
"learning_rate": 0.0001834534498993888, |
|
"loss": 0.7395, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.39784117193523516, |
|
"grad_norm": 0.027110572904348373, |
|
"learning_rate": 0.0001833175483265273, |
|
"loss": 0.7963, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.39938319198149574, |
|
"grad_norm": 0.027663685381412506, |
|
"learning_rate": 0.00018318114170333568, |
|
"loss": 0.7893, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.4009252120277564, |
|
"grad_norm": 0.027797933667898178, |
|
"learning_rate": 0.00018304423085667714, |
|
"loss": 0.7228, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.40246723207401697, |
|
"grad_norm": 0.026281701400876045, |
|
"learning_rate": 0.0001829068166164712, |
|
"loss": 0.749, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.40400925212027755, |
|
"grad_norm": 0.024708108976483345, |
|
"learning_rate": 0.00018276889981568906, |
|
"loss": 0.6307, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.4055512721665382, |
|
"grad_norm": 0.028213316574692726, |
|
"learning_rate": 0.0001826304812903481, |
|
"loss": 0.8186, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.4070932922127988, |
|
"grad_norm": 0.024718405678868294, |
|
"learning_rate": 0.00018249156187950715, |
|
"loss": 0.7077, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.40863531225905936, |
|
"grad_norm": 0.02398741990327835, |
|
"learning_rate": 0.00018235214242526125, |
|
"loss": 0.7041, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.41017733230531994, |
|
"grad_norm": 0.03176787495613098, |
|
"learning_rate": 0.00018221222377273657, |
|
"loss": 0.7036, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.4117193523515806, |
|
"grad_norm": 0.028862686827778816, |
|
"learning_rate": 0.0001820718067700853, |
|
"loss": 0.7947, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.41326137239784116, |
|
"grad_norm": 0.026759544387459755, |
|
"learning_rate": 0.0001819308922684805, |
|
"loss": 0.7737, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.41480339244410175, |
|
"grad_norm": 0.02719755284488201, |
|
"learning_rate": 0.00018178948112211103, |
|
"loss": 0.7403, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.4163454124903624, |
|
"grad_norm": 0.024756524711847305, |
|
"learning_rate": 0.0001816475741881761, |
|
"loss": 0.6994, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.41788743253662297, |
|
"grad_norm": 0.03232420235872269, |
|
"learning_rate": 0.00018150517232688049, |
|
"loss": 0.7866, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.41942945258288356, |
|
"grad_norm": 0.027607185766100883, |
|
"learning_rate": 0.00018136227640142894, |
|
"loss": 0.7905, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.4209714726291442, |
|
"grad_norm": 0.024344706907868385, |
|
"learning_rate": 0.00018121888727802113, |
|
"loss": 0.7408, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.4225134926754048, |
|
"grad_norm": 0.025088010355830193, |
|
"learning_rate": 0.0001810750058258464, |
|
"loss": 0.737, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.42405551272166536, |
|
"grad_norm": 0.023952683433890343, |
|
"learning_rate": 0.00018093063291707847, |
|
"loss": 0.7764, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.425597532767926, |
|
"grad_norm": 0.0288414116948843, |
|
"learning_rate": 0.00018078576942687008, |
|
"loss": 0.7035, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.4271395528141866, |
|
"grad_norm": 0.02681080810725689, |
|
"learning_rate": 0.0001806404162333479, |
|
"loss": 0.739, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.42868157286044717, |
|
"grad_norm": 0.0266602523624897, |
|
"learning_rate": 0.0001804945742176069, |
|
"loss": 0.7213, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.4302235929067078, |
|
"grad_norm": 0.025282425805926323, |
|
"learning_rate": 0.00018034824426370523, |
|
"loss": 0.6807, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.4317656129529684, |
|
"grad_norm": 0.025683747604489326, |
|
"learning_rate": 0.00018020142725865888, |
|
"loss": 0.7283, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.433307632999229, |
|
"grad_norm": 0.024966144934296608, |
|
"learning_rate": 0.00018005412409243606, |
|
"loss": 0.7096, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.4348496530454896, |
|
"grad_norm": 0.027953188866376877, |
|
"learning_rate": 0.00017990633565795208, |
|
"loss": 0.8148, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.4363916730917502, |
|
"grad_norm": 0.02772989496588707, |
|
"learning_rate": 0.00017975806285106387, |
|
"loss": 0.8568, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.4379336931380108, |
|
"grad_norm": 0.028020409867167473, |
|
"learning_rate": 0.00017960930657056438, |
|
"loss": 0.6732, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.4394757131842714, |
|
"grad_norm": 0.025754399597644806, |
|
"learning_rate": 0.00017946006771817733, |
|
"loss": 0.7238, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.441017733230532, |
|
"grad_norm": 0.030171813443303108, |
|
"learning_rate": 0.00017931034719855166, |
|
"loss": 0.7493, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.4425597532767926, |
|
"grad_norm": 0.026995845139026642, |
|
"learning_rate": 0.00017916014591925605, |
|
"loss": 0.6118, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.4441017733230532, |
|
"grad_norm": 0.03541433438658714, |
|
"learning_rate": 0.00017900946479077346, |
|
"loss": 0.7243, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.4456437933693138, |
|
"grad_norm": 0.029751230031251907, |
|
"learning_rate": 0.00017885830472649553, |
|
"loss": 0.7081, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.4471858134155744, |
|
"grad_norm": 0.022569075226783752, |
|
"learning_rate": 0.00017870666664271707, |
|
"loss": 0.6488, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.448727833461835, |
|
"grad_norm": 0.03262341767549515, |
|
"learning_rate": 0.00017855455145863062, |
|
"loss": 0.7626, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.4502698535080956, |
|
"grad_norm": 0.02811555750668049, |
|
"learning_rate": 0.0001784019600963207, |
|
"loss": 0.7485, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.4518118735543562, |
|
"grad_norm": 0.02504836954176426, |
|
"learning_rate": 0.00017824889348075837, |
|
"loss": 0.7636, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.4533538936006168, |
|
"grad_norm": 0.02362634427845478, |
|
"learning_rate": 0.00017809535253979547, |
|
"loss": 0.6915, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.4548959136468774, |
|
"grad_norm": 0.029891418293118477, |
|
"learning_rate": 0.00017794133820415916, |
|
"loss": 0.7686, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.456437933693138, |
|
"grad_norm": 0.02471439354121685, |
|
"learning_rate": 0.0001777868514074462, |
|
"loss": 0.6693, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.4579799537393986, |
|
"grad_norm": 0.025612330064177513, |
|
"learning_rate": 0.00017763189308611722, |
|
"loss": 0.7364, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.45952197378565923, |
|
"grad_norm": 0.026865236461162567, |
|
"learning_rate": 0.00017747646417949113, |
|
"loss": 0.7445, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.4610639938319198, |
|
"grad_norm": 0.024343574419617653, |
|
"learning_rate": 0.00017732056562973954, |
|
"loss": 0.6271, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.4626060138781804, |
|
"grad_norm": 0.02818606235086918, |
|
"learning_rate": 0.00017716419838188077, |
|
"loss": 0.7753, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.46414803392444104, |
|
"grad_norm": 0.026821713894605637, |
|
"learning_rate": 0.00017700736338377435, |
|
"loss": 0.6976, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.4656900539707016, |
|
"grad_norm": 0.025784511119127274, |
|
"learning_rate": 0.00017685006158611516, |
|
"loss": 0.7062, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.4672320740169622, |
|
"grad_norm": 0.028515879064798355, |
|
"learning_rate": 0.00017669229394242766, |
|
"loss": 0.7909, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.46877409406322285, |
|
"grad_norm": 0.024095451459288597, |
|
"learning_rate": 0.0001765340614090603, |
|
"loss": 0.7535, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.47031611410948343, |
|
"grad_norm": 0.025953758507966995, |
|
"learning_rate": 0.0001763753649451794, |
|
"loss": 0.7623, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.471858134155744, |
|
"grad_norm": 0.03294069692492485, |
|
"learning_rate": 0.00017621620551276366, |
|
"loss": 0.8946, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.47340015420200465, |
|
"grad_norm": 0.028394997119903564, |
|
"learning_rate": 0.00017605658407659808, |
|
"loss": 0.7251, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.47494217424826524, |
|
"grad_norm": 0.025346368551254272, |
|
"learning_rate": 0.00017589650160426828, |
|
"loss": 0.7074, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.4764841942945258, |
|
"grad_norm": 0.025906400755047798, |
|
"learning_rate": 0.0001757359590661545, |
|
"loss": 0.6472, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.4780262143407864, |
|
"grad_norm": 0.02889554388821125, |
|
"learning_rate": 0.00017557495743542585, |
|
"loss": 0.715, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.47956823438704704, |
|
"grad_norm": 0.029205597937107086, |
|
"learning_rate": 0.00017541349768803428, |
|
"loss": 0.698, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.4811102544333076, |
|
"grad_norm": 0.02610400691628456, |
|
"learning_rate": 0.0001752515808027088, |
|
"loss": 0.7073, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.4826522744795682, |
|
"grad_norm": 0.023945793509483337, |
|
"learning_rate": 0.00017508920776094944, |
|
"loss": 0.6865, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.48419429452582885, |
|
"grad_norm": 0.028426503762602806, |
|
"learning_rate": 0.0001749263795470213, |
|
"loss": 0.7264, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.48573631457208943, |
|
"grad_norm": 0.02865850180387497, |
|
"learning_rate": 0.0001747630971479487, |
|
"loss": 0.7204, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.48727833461835, |
|
"grad_norm": 0.027321334928274155, |
|
"learning_rate": 0.00017459936155350908, |
|
"loss": 0.7491, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.48882035466461066, |
|
"grad_norm": 0.02754514105618, |
|
"learning_rate": 0.00017443517375622704, |
|
"loss": 0.7567, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.49036237471087124, |
|
"grad_norm": 0.028822382912039757, |
|
"learning_rate": 0.00017427053475136826, |
|
"loss": 0.7559, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.4919043947571318, |
|
"grad_norm": 0.03181014209985733, |
|
"learning_rate": 0.00017410544553693365, |
|
"loss": 0.7704, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.49344641480339246, |
|
"grad_norm": 0.023862695321440697, |
|
"learning_rate": 0.00017393990711365312, |
|
"loss": 0.6085, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.49498843484965305, |
|
"grad_norm": 0.02703220769762993, |
|
"learning_rate": 0.00017377392048497953, |
|
"loss": 0.6979, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.49653045489591363, |
|
"grad_norm": 0.025343257933855057, |
|
"learning_rate": 0.00017360748665708268, |
|
"loss": 0.7287, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.49807247494217427, |
|
"grad_norm": 0.02830134704709053, |
|
"learning_rate": 0.00017344060663884324, |
|
"loss": 0.8054, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.49961449498843485, |
|
"grad_norm": 0.025809939950704575, |
|
"learning_rate": 0.00017327328144184646, |
|
"loss": 0.704, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.5011565150346955, |
|
"grad_norm": 0.027546260505914688, |
|
"learning_rate": 0.00017310551208037626, |
|
"loss": 0.7099, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.5026985350809561, |
|
"grad_norm": 0.027951935306191444, |
|
"learning_rate": 0.00017293729957140893, |
|
"loss": 0.878, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.5042405551272167, |
|
"grad_norm": 0.026868853718042374, |
|
"learning_rate": 0.000172768644934607, |
|
"loss": 0.7714, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.5057825751734772, |
|
"grad_norm": 0.026975559070706367, |
|
"learning_rate": 0.0001725995491923131, |
|
"loss": 0.7526, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.5073245952197378, |
|
"grad_norm": 0.027235837653279305, |
|
"learning_rate": 0.0001724300133695437, |
|
"loss": 0.6515, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.5088666152659984, |
|
"grad_norm": 0.032752856612205505, |
|
"learning_rate": 0.00017226003849398294, |
|
"loss": 0.8019, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.5104086353122591, |
|
"grad_norm": 0.029344195500016212, |
|
"learning_rate": 0.0001720896255959764, |
|
"loss": 0.9296, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.5119506553585197, |
|
"grad_norm": 0.025766605511307716, |
|
"learning_rate": 0.00017191877570852483, |
|
"loss": 0.7944, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.5134926754047803, |
|
"grad_norm": 0.023067327216267586, |
|
"learning_rate": 0.0001717474898672779, |
|
"loss": 0.6528, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.5150346954510409, |
|
"grad_norm": 0.026817042380571365, |
|
"learning_rate": 0.00017157576911052796, |
|
"loss": 0.763, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.5165767154973014, |
|
"grad_norm": 0.02811489813029766, |
|
"learning_rate": 0.00017140361447920364, |
|
"loss": 0.7936, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.518118735543562, |
|
"grad_norm": 0.02597888559103012, |
|
"learning_rate": 0.00017123102701686372, |
|
"loss": 0.7217, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.5196607555898227, |
|
"grad_norm": 0.030244017019867897, |
|
"learning_rate": 0.00017105800776969055, |
|
"loss": 0.756, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.5212027756360833, |
|
"grad_norm": 0.028289398178458214, |
|
"learning_rate": 0.00017088455778648397, |
|
"loss": 0.7972, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.5227447956823439, |
|
"grad_norm": 0.02397543005645275, |
|
"learning_rate": 0.00017071067811865476, |
|
"loss": 0.6358, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.5242868157286045, |
|
"grad_norm": 0.02565479464828968, |
|
"learning_rate": 0.00017053636982021844, |
|
"loss": 0.6912, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.525828835774865, |
|
"grad_norm": 0.027768775820732117, |
|
"learning_rate": 0.00017036163394778864, |
|
"loss": 0.7997, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.5273708558211256, |
|
"grad_norm": 0.033537182956933975, |
|
"learning_rate": 0.00017018647156057096, |
|
"loss": 0.7363, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.5289128758673862, |
|
"grad_norm": 0.030843475833535194, |
|
"learning_rate": 0.00017001088372035637, |
|
"loss": 0.6977, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.5304548959136469, |
|
"grad_norm": 0.026586662977933884, |
|
"learning_rate": 0.00016983487149151486, |
|
"loss": 0.7874, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.5319969159599075, |
|
"grad_norm": 0.027003532275557518, |
|
"learning_rate": 0.00016965843594098892, |
|
"loss": 0.7895, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.5335389360061681, |
|
"grad_norm": 0.02772395871579647, |
|
"learning_rate": 0.00016948157813828716, |
|
"loss": 0.7799, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.5350809560524287, |
|
"grad_norm": 0.025633979588747025, |
|
"learning_rate": 0.0001693042991554777, |
|
"loss": 0.744, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.5366229760986893, |
|
"grad_norm": 0.033604227006435394, |
|
"learning_rate": 0.00016912660006718186, |
|
"loss": 0.6707, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.5381649961449498, |
|
"grad_norm": 0.03711126372218132, |
|
"learning_rate": 0.00016894848195056746, |
|
"loss": 0.8468, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.5397070161912105, |
|
"grad_norm": 0.02382393553853035, |
|
"learning_rate": 0.00016876994588534234, |
|
"loss": 0.6884, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.5412490362374711, |
|
"grad_norm": 0.02597069926559925, |
|
"learning_rate": 0.0001685909929537479, |
|
"loss": 0.7789, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.5427910562837317, |
|
"grad_norm": 0.027003685012459755, |
|
"learning_rate": 0.0001684116242405525, |
|
"loss": 0.7364, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.5443330763299923, |
|
"grad_norm": 0.028209254145622253, |
|
"learning_rate": 0.00016823184083304482, |
|
"loss": 0.8014, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.5458750963762529, |
|
"grad_norm": 0.025543801486492157, |
|
"learning_rate": 0.0001680516438210273, |
|
"loss": 0.7241, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.5474171164225135, |
|
"grad_norm": 0.024730654433369637, |
|
"learning_rate": 0.00016787103429680955, |
|
"loss": 0.6425, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.5489591364687741, |
|
"grad_norm": 0.02615622617304325, |
|
"learning_rate": 0.0001676900133552018, |
|
"loss": 0.7091, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.5505011565150347, |
|
"grad_norm": 0.025628188624978065, |
|
"learning_rate": 0.00016750858209350808, |
|
"loss": 0.848, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.5520431765612953, |
|
"grad_norm": 0.02642144076526165, |
|
"learning_rate": 0.0001673267416115198, |
|
"loss": 0.7026, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.5535851966075559, |
|
"grad_norm": 0.02542021870613098, |
|
"learning_rate": 0.00016714449301150883, |
|
"loss": 0.6338, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.5551272166538165, |
|
"grad_norm": 0.028573600575327873, |
|
"learning_rate": 0.00016696183739822108, |
|
"loss": 0.7553, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.5566692367000771, |
|
"grad_norm": 0.024142924696207047, |
|
"learning_rate": 0.00016677877587886956, |
|
"loss": 0.7327, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.5582112567463376, |
|
"grad_norm": 0.029369287192821503, |
|
"learning_rate": 0.00016659530956312788, |
|
"loss": 0.7728, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.5597532767925983, |
|
"grad_norm": 0.02711080014705658, |
|
"learning_rate": 0.00016641143956312336, |
|
"loss": 0.6843, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.5612952968388589, |
|
"grad_norm": 0.027241146191954613, |
|
"learning_rate": 0.00016622716699343033, |
|
"loss": 0.8534, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.5628373168851195, |
|
"grad_norm": 0.030265534296631813, |
|
"learning_rate": 0.0001660424929710635, |
|
"loss": 0.8579, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.5643793369313801, |
|
"grad_norm": 0.029054157435894012, |
|
"learning_rate": 0.000165857418615471, |
|
"loss": 0.75, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.5659213569776407, |
|
"grad_norm": 0.029963452368974686, |
|
"learning_rate": 0.0001656719450485278, |
|
"loss": 0.7464, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.5674633770239013, |
|
"grad_norm": 0.025763841345906258, |
|
"learning_rate": 0.00016548607339452853, |
|
"loss": 0.6632, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.569005397070162, |
|
"grad_norm": 0.02374422177672386, |
|
"learning_rate": 0.00016529980478018115, |
|
"loss": 0.6546, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.5705474171164225, |
|
"grad_norm": 0.02538699097931385, |
|
"learning_rate": 0.00016511314033459994, |
|
"loss": 0.6603, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.5720894371626831, |
|
"grad_norm": 0.028310047462582588, |
|
"learning_rate": 0.0001649260811892984, |
|
"loss": 0.6088, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.5736314572089437, |
|
"grad_norm": 0.02981553040444851, |
|
"learning_rate": 0.00016473862847818277, |
|
"loss": 0.8053, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.5751734772552043, |
|
"grad_norm": 0.025665050372481346, |
|
"learning_rate": 0.0001645507833375449, |
|
"loss": 0.6779, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.5767154973014649, |
|
"grad_norm": 0.030752114951610565, |
|
"learning_rate": 0.0001643625469060555, |
|
"loss": 0.6893, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.5782575173477256, |
|
"grad_norm": 0.02725459821522236, |
|
"learning_rate": 0.00016417392032475715, |
|
"loss": 0.7879, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.5797995373939862, |
|
"grad_norm": 0.02713746391236782, |
|
"learning_rate": 0.00016398490473705743, |
|
"loss": 0.7578, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.5813415574402467, |
|
"grad_norm": 0.02641828916966915, |
|
"learning_rate": 0.000163795501288722, |
|
"loss": 0.7335, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.5828835774865073, |
|
"grad_norm": 0.027243638411164284, |
|
"learning_rate": 0.00016360571112786765, |
|
"loss": 0.6873, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.5844255975327679, |
|
"grad_norm": 0.028045805171132088, |
|
"learning_rate": 0.00016341553540495532, |
|
"loss": 0.7141, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.5859676175790285, |
|
"grad_norm": 0.029962563887238503, |
|
"learning_rate": 0.00016322497527278306, |
|
"loss": 0.882, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.5875096376252892, |
|
"grad_norm": 0.027544857934117317, |
|
"learning_rate": 0.00016303403188647913, |
|
"loss": 0.6695, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.5890516576715498, |
|
"grad_norm": 0.02867518924176693, |
|
"learning_rate": 0.00016284270640349514, |
|
"loss": 0.8052, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.5905936777178104, |
|
"grad_norm": 0.02330535091459751, |
|
"learning_rate": 0.00016265099998359866, |
|
"loss": 0.6825, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.5921356977640709, |
|
"grad_norm": 0.030629336833953857, |
|
"learning_rate": 0.00016245891378886655, |
|
"loss": 0.8545, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.5936777178103315, |
|
"grad_norm": 0.030393701046705246, |
|
"learning_rate": 0.00016226644898367768, |
|
"loss": 0.7342, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.5952197378565921, |
|
"grad_norm": 0.02500557340681553, |
|
"learning_rate": 0.000162073606734706, |
|
"loss": 0.6975, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.5967617579028527, |
|
"grad_norm": 0.031064705923199654, |
|
"learning_rate": 0.00016188038821091344, |
|
"loss": 0.7903, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.5983037779491134, |
|
"grad_norm": 0.02647087723016739, |
|
"learning_rate": 0.00016168679458354284, |
|
"loss": 0.745, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.599845797995374, |
|
"grad_norm": 0.02917948178946972, |
|
"learning_rate": 0.00016149282702611077, |
|
"loss": 0.7819, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.6013878180416345, |
|
"grad_norm": 0.02458810992538929, |
|
"learning_rate": 0.00016129848671440046, |
|
"loss": 0.6466, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.6029298380878951, |
|
"grad_norm": 0.023677226155996323, |
|
"learning_rate": 0.00016110377482645477, |
|
"loss": 0.7161, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.6044718581341557, |
|
"grad_norm": 0.02807523123919964, |
|
"learning_rate": 0.00016090869254256892, |
|
"loss": 0.7164, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.6060138781804163, |
|
"grad_norm": 0.031006982550024986, |
|
"learning_rate": 0.0001607132410452833, |
|
"loss": 0.7781, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.607555898226677, |
|
"grad_norm": 0.029372677206993103, |
|
"learning_rate": 0.00016051742151937655, |
|
"loss": 0.7034, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.6090979182729376, |
|
"grad_norm": 0.025212230160832405, |
|
"learning_rate": 0.00016032123515185797, |
|
"loss": 0.6567, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.6106399383191982, |
|
"grad_norm": 0.024775920435786247, |
|
"learning_rate": 0.00016012468313196084, |
|
"loss": 0.6319, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.6121819583654587, |
|
"grad_norm": 0.02687055990099907, |
|
"learning_rate": 0.0001599277666511347, |
|
"loss": 0.7214, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.6137239784117193, |
|
"grad_norm": 0.028431419283151627, |
|
"learning_rate": 0.0001597304869030385, |
|
"loss": 0.7815, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.6152659984579799, |
|
"grad_norm": 0.027922354638576508, |
|
"learning_rate": 0.00015953284508353317, |
|
"loss": 0.7016, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.6168080185042406, |
|
"grad_norm": 0.02737678587436676, |
|
"learning_rate": 0.00015933484239067446, |
|
"loss": 0.7507, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.6183500385505012, |
|
"grad_norm": 0.02791019156575203, |
|
"learning_rate": 0.00015913648002470563, |
|
"loss": 0.705, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.6198920585967618, |
|
"grad_norm": 0.027861539274454117, |
|
"learning_rate": 0.0001589377591880501, |
|
"loss": 0.7208, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.6214340786430224, |
|
"grad_norm": 0.02952715940773487, |
|
"learning_rate": 0.00015873868108530443, |
|
"loss": 0.7114, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.6229760986892829, |
|
"grad_norm": 0.02896735444664955, |
|
"learning_rate": 0.0001585392469232307, |
|
"loss": 0.7797, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.6245181187355435, |
|
"grad_norm": 0.02640017308294773, |
|
"learning_rate": 0.00015833945791074943, |
|
"loss": 0.7186, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.6260601387818041, |
|
"grad_norm": 0.025937926024198532, |
|
"learning_rate": 0.000158139315258932, |
|
"loss": 0.781, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.6276021588280648, |
|
"grad_norm": 0.02851933054625988, |
|
"learning_rate": 0.00015793882018099364, |
|
"loss": 0.7522, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.6291441788743254, |
|
"grad_norm": 0.02368611842393875, |
|
"learning_rate": 0.00015773797389228582, |
|
"loss": 0.719, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.630686198920586, |
|
"grad_norm": 0.026969095692038536, |
|
"learning_rate": 0.00015753677761028896, |
|
"loss": 0.7554, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.6322282189668466, |
|
"grad_norm": 0.02418413758277893, |
|
"learning_rate": 0.00015733523255460506, |
|
"loss": 0.6919, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.6337702390131071, |
|
"grad_norm": 0.026073114946484566, |
|
"learning_rate": 0.0001571333399469503, |
|
"loss": 0.7059, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.6353122590593677, |
|
"grad_norm": 0.027075573801994324, |
|
"learning_rate": 0.0001569311010111476, |
|
"loss": 0.7239, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.6368542791056284, |
|
"grad_norm": 0.02724389173090458, |
|
"learning_rate": 0.00015672851697311934, |
|
"loss": 0.7615, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.638396299151889, |
|
"grad_norm": 0.026826992630958557, |
|
"learning_rate": 0.00015652558906087971, |
|
"loss": 0.8138, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.6399383191981496, |
|
"grad_norm": 0.02641242742538452, |
|
"learning_rate": 0.00015632231850452747, |
|
"loss": 0.6956, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.6414803392444102, |
|
"grad_norm": 0.02495909109711647, |
|
"learning_rate": 0.00015611870653623825, |
|
"loss": 0.7546, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.6430223592906708, |
|
"grad_norm": 0.035753343254327774, |
|
"learning_rate": 0.00015591475439025745, |
|
"loss": 0.7257, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.6445643793369313, |
|
"grad_norm": 0.03195042535662651, |
|
"learning_rate": 0.00015571046330289237, |
|
"loss": 0.693, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.646106399383192, |
|
"grad_norm": 0.027049189433455467, |
|
"learning_rate": 0.00015550583451250503, |
|
"loss": 0.7602, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.6476484194294526, |
|
"grad_norm": 0.027553152292966843, |
|
"learning_rate": 0.00015530086925950434, |
|
"loss": 0.7723, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.6491904394757132, |
|
"grad_norm": 0.025219090282917023, |
|
"learning_rate": 0.00015509556878633894, |
|
"loss": 0.7011, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.6507324595219738, |
|
"grad_norm": 0.029802288860082626, |
|
"learning_rate": 0.00015488993433748944, |
|
"loss": 0.6898, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.6522744795682344, |
|
"grad_norm": 0.02772880345582962, |
|
"learning_rate": 0.00015468396715946083, |
|
"loss": 0.7399, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.653816499614495, |
|
"grad_norm": 0.025485830381512642, |
|
"learning_rate": 0.00015447766850077517, |
|
"loss": 0.6622, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.6553585196607556, |
|
"grad_norm": 0.027489742264151573, |
|
"learning_rate": 0.00015427103961196376, |
|
"loss": 0.8027, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.6569005397070162, |
|
"grad_norm": 0.026778720319271088, |
|
"learning_rate": 0.00015406408174555976, |
|
"loss": 0.7804, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.6584425597532768, |
|
"grad_norm": 0.027850337326526642, |
|
"learning_rate": 0.00015385679615609042, |
|
"loss": 0.7645, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.6599845797995374, |
|
"grad_norm": 0.02659332938492298, |
|
"learning_rate": 0.00015364918410006967, |
|
"loss": 0.7609, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.661526599845798, |
|
"grad_norm": 0.0292272437363863, |
|
"learning_rate": 0.0001534412468359903, |
|
"loss": 0.7788, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.6630686198920586, |
|
"grad_norm": 0.02392621338367462, |
|
"learning_rate": 0.00015323298562431648, |
|
"loss": 0.675, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.6646106399383191, |
|
"grad_norm": 0.030261410400271416, |
|
"learning_rate": 0.00015302440172747605, |
|
"loss": 0.6585, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.6661526599845798, |
|
"grad_norm": 0.031531739979982376, |
|
"learning_rate": 0.00015281549640985294, |
|
"loss": 0.8002, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.6676946800308404, |
|
"grad_norm": 0.025663699954748154, |
|
"learning_rate": 0.00015260627093777936, |
|
"loss": 0.7068, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.669236700077101, |
|
"grad_norm": 0.02657734416425228, |
|
"learning_rate": 0.00015239672657952832, |
|
"loss": 0.7369, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.6707787201233616, |
|
"grad_norm": 0.027592379599809647, |
|
"learning_rate": 0.0001521868646053058, |
|
"loss": 0.7245, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.6723207401696222, |
|
"grad_norm": 0.02645149454474449, |
|
"learning_rate": 0.00015197668628724303, |
|
"loss": 0.718, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.6738627602158828, |
|
"grad_norm": 0.02535802498459816, |
|
"learning_rate": 0.00015176619289938888, |
|
"loss": 0.6926, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.6754047802621435, |
|
"grad_norm": 0.02912677638232708, |
|
"learning_rate": 0.00015155538571770218, |
|
"loss": 0.8564, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.676946800308404, |
|
"grad_norm": 0.023175543174147606, |
|
"learning_rate": 0.00015134426602004375, |
|
"loss": 0.6568, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.6784888203546646, |
|
"grad_norm": 0.02832154743373394, |
|
"learning_rate": 0.00015113283508616895, |
|
"loss": 0.6571, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.6800308404009252, |
|
"grad_norm": 0.029950594529509544, |
|
"learning_rate": 0.0001509210941977196, |
|
"loss": 0.8564, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.6815728604471858, |
|
"grad_norm": 0.02448093518614769, |
|
"learning_rate": 0.00015070904463821658, |
|
"loss": 0.6936, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.6831148804934464, |
|
"grad_norm": 0.027585268020629883, |
|
"learning_rate": 0.00015049668769305172, |
|
"loss": 0.7991, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.6846569005397071, |
|
"grad_norm": 0.025705596432089806, |
|
"learning_rate": 0.00015028402464948022, |
|
"loss": 0.7187, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.6861989205859677, |
|
"grad_norm": 0.02810623310506344, |
|
"learning_rate": 0.00015007105679661276, |
|
"loss": 0.7315, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.6877409406322282, |
|
"grad_norm": 0.02478802390396595, |
|
"learning_rate": 0.0001498577854254076, |
|
"loss": 0.7201, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.6892829606784888, |
|
"grad_norm": 0.03222353756427765, |
|
"learning_rate": 0.0001496442118286631, |
|
"loss": 0.796, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.6908249807247494, |
|
"grad_norm": 0.025557026267051697, |
|
"learning_rate": 0.00014943033730100935, |
|
"loss": 0.6066, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.69236700077101, |
|
"grad_norm": 0.02581370249390602, |
|
"learning_rate": 0.00014921616313890072, |
|
"loss": 0.7186, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.6939090208172706, |
|
"grad_norm": 0.025746649131178856, |
|
"learning_rate": 0.00014900169064060802, |
|
"loss": 0.703, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.6954510408635313, |
|
"grad_norm": 0.03065100871026516, |
|
"learning_rate": 0.00014878692110621028, |
|
"loss": 0.7724, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.6969930609097919, |
|
"grad_norm": 0.03629877045750618, |
|
"learning_rate": 0.00014857185583758723, |
|
"loss": 0.7703, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.6985350809560524, |
|
"grad_norm": 0.025341391563415527, |
|
"learning_rate": 0.0001483564961384112, |
|
"loss": 0.6492, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.700077101002313, |
|
"grad_norm": 0.03427153080701828, |
|
"learning_rate": 0.0001481408433141394, |
|
"loss": 0.7148, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.7016191210485736, |
|
"grad_norm": 0.023825203999876976, |
|
"learning_rate": 0.0001479248986720057, |
|
"loss": 0.6197, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.7031611410948342, |
|
"grad_norm": 0.026944074779748917, |
|
"learning_rate": 0.00014770866352101307, |
|
"loss": 0.7273, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.7047031611410949, |
|
"grad_norm": 0.028203219175338745, |
|
"learning_rate": 0.00014749213917192538, |
|
"loss": 0.7741, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.7062451811873555, |
|
"grad_norm": 0.024622568860650063, |
|
"learning_rate": 0.00014727532693725962, |
|
"loss": 0.6212, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.707787201233616, |
|
"grad_norm": 0.023979736492037773, |
|
"learning_rate": 0.00014705822813127777, |
|
"loss": 0.6863, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.7093292212798766, |
|
"grad_norm": 0.027042685076594353, |
|
"learning_rate": 0.00014684084406997903, |
|
"loss": 0.7746, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.7108712413261372, |
|
"grad_norm": 0.026160864159464836, |
|
"learning_rate": 0.00014662317607109168, |
|
"loss": 0.727, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.7124132613723978, |
|
"grad_norm": 0.035195063799619675, |
|
"learning_rate": 0.0001464052254540652, |
|
"loss": 0.9004, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.7139552814186585, |
|
"grad_norm": 0.02714708261191845, |
|
"learning_rate": 0.00014618699354006223, |
|
"loss": 0.6853, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.7154973014649191, |
|
"grad_norm": 0.027736373245716095, |
|
"learning_rate": 0.0001459684816519505, |
|
"loss": 0.7095, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.7170393215111797, |
|
"grad_norm": 0.032326798886060715, |
|
"learning_rate": 0.000145749691114295, |
|
"loss": 0.7256, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.7185813415574402, |
|
"grad_norm": 0.028024908155202866, |
|
"learning_rate": 0.00014553062325334967, |
|
"loss": 0.7213, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.7201233616037008, |
|
"grad_norm": 0.026767205446958542, |
|
"learning_rate": 0.00014531127939704965, |
|
"loss": 0.6619, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.7216653816499614, |
|
"grad_norm": 0.025469880551099777, |
|
"learning_rate": 0.00014509166087500302, |
|
"loss": 0.6174, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.723207401696222, |
|
"grad_norm": 0.03378410264849663, |
|
"learning_rate": 0.00014487176901848285, |
|
"loss": 0.6447, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.7247494217424827, |
|
"grad_norm": 0.026924695819616318, |
|
"learning_rate": 0.00014465160516041904, |
|
"loss": 0.7116, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.7262914417887433, |
|
"grad_norm": 0.029700160026550293, |
|
"learning_rate": 0.00014443117063539038, |
|
"loss": 0.811, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.7278334618350039, |
|
"grad_norm": 0.030913611873984337, |
|
"learning_rate": 0.00014421046677961626, |
|
"loss": 0.77, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.7293754818812644, |
|
"grad_norm": 0.029507668688893318, |
|
"learning_rate": 0.0001439894949309489, |
|
"loss": 0.7813, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.730917501927525, |
|
"grad_norm": 0.028620922937989235, |
|
"learning_rate": 0.00014376825642886472, |
|
"loss": 0.7401, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.7324595219737856, |
|
"grad_norm": 0.025677144527435303, |
|
"learning_rate": 0.0001435467526144568, |
|
"loss": 0.7408, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.7340015420200463, |
|
"grad_norm": 0.025129586458206177, |
|
"learning_rate": 0.00014332498483042637, |
|
"loss": 0.6585, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.7355435620663069, |
|
"grad_norm": 0.026812126860022545, |
|
"learning_rate": 0.0001431029544210747, |
|
"loss": 0.8063, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.7370855821125675, |
|
"grad_norm": 0.02705306001007557, |
|
"learning_rate": 0.0001428806627322952, |
|
"loss": 0.7655, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.7386276021588281, |
|
"grad_norm": 0.026533039286732674, |
|
"learning_rate": 0.0001426581111115649, |
|
"loss": 0.6685, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.7401696222050886, |
|
"grad_norm": 0.030538393184542656, |
|
"learning_rate": 0.00014243530090793667, |
|
"loss": 0.7389, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.7417116422513492, |
|
"grad_norm": 0.027102958410978317, |
|
"learning_rate": 0.00014221223347203068, |
|
"loss": 0.726, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.7432536622976099, |
|
"grad_norm": 0.024465948343276978, |
|
"learning_rate": 0.00014198891015602646, |
|
"loss": 0.6379, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.7447956823438705, |
|
"grad_norm": 0.027429422363638878, |
|
"learning_rate": 0.00014176533231365464, |
|
"loss": 0.726, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.7463377023901311, |
|
"grad_norm": 0.02289111353456974, |
|
"learning_rate": 0.00014154150130018866, |
|
"loss": 0.6354, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.7478797224363917, |
|
"grad_norm": 0.027374420315027237, |
|
"learning_rate": 0.00014131741847243665, |
|
"loss": 0.8206, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.7494217424826523, |
|
"grad_norm": 0.029361480847001076, |
|
"learning_rate": 0.0001410930851887332, |
|
"loss": 0.6994, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.7509637625289128, |
|
"grad_norm": 0.027077002450823784, |
|
"learning_rate": 0.00014086850280893107, |
|
"loss": 0.7133, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.7525057825751735, |
|
"grad_norm": 0.028823737055063248, |
|
"learning_rate": 0.000140643672694393, |
|
"loss": 0.7975, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.7540478026214341, |
|
"grad_norm": 0.02851509116590023, |
|
"learning_rate": 0.0001404185962079834, |
|
"loss": 0.7497, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.7555898226676947, |
|
"grad_norm": 0.025098523125052452, |
|
"learning_rate": 0.00014019327471406022, |
|
"loss": 0.6978, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.7571318427139553, |
|
"grad_norm": 0.03084149770438671, |
|
"learning_rate": 0.00013996770957846644, |
|
"loss": 0.7607, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.7586738627602159, |
|
"grad_norm": 0.02702442556619644, |
|
"learning_rate": 0.00013974190216852202, |
|
"loss": 0.7639, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.7602158828064765, |
|
"grad_norm": 0.033685747534036636, |
|
"learning_rate": 0.00013951585385301555, |
|
"loss": 0.7793, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.761757902852737, |
|
"grad_norm": 0.028003569692373276, |
|
"learning_rate": 0.00013928956600219592, |
|
"loss": 0.755, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.7632999228989977, |
|
"grad_norm": 0.026269137859344482, |
|
"learning_rate": 0.00013906303998776392, |
|
"loss": 0.7896, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.7648419429452583, |
|
"grad_norm": 0.027992503717541695, |
|
"learning_rate": 0.0001388362771828642, |
|
"loss": 0.6472, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.7663839629915189, |
|
"grad_norm": 0.028174640610814095, |
|
"learning_rate": 0.00013860927896207665, |
|
"loss": 0.765, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.7679259830377795, |
|
"grad_norm": 0.04180079326033592, |
|
"learning_rate": 0.0001383820467014082, |
|
"loss": 0.7479, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.7694680030840401, |
|
"grad_norm": 0.028985602781176567, |
|
"learning_rate": 0.00013815458177828454, |
|
"loss": 0.7121, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.7710100231303006, |
|
"grad_norm": 0.026408828794956207, |
|
"learning_rate": 0.00013792688557154166, |
|
"loss": 0.7411, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.7725520431765613, |
|
"grad_norm": 0.027537405490875244, |
|
"learning_rate": 0.00013769895946141752, |
|
"loss": 0.7628, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.7740940632228219, |
|
"grad_norm": 0.028669610619544983, |
|
"learning_rate": 0.00013747080482954377, |
|
"loss": 0.7698, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.7756360832690825, |
|
"grad_norm": 0.027917267754673958, |
|
"learning_rate": 0.00013724242305893715, |
|
"loss": 0.8452, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.7771781033153431, |
|
"grad_norm": 0.031637243926525116, |
|
"learning_rate": 0.00013701381553399145, |
|
"loss": 0.7457, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.7787201233616037, |
|
"grad_norm": 0.027903905138373375, |
|
"learning_rate": 0.00013678498364046876, |
|
"loss": 0.7475, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.7802621434078643, |
|
"grad_norm": 0.03381239250302315, |
|
"learning_rate": 0.00013655592876549134, |
|
"loss": 0.836, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.781804163454125, |
|
"grad_norm": 0.0790882334113121, |
|
"learning_rate": 0.0001363266522975331, |
|
"loss": 0.5842, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.7833461835003855, |
|
"grad_norm": 0.03034863993525505, |
|
"learning_rate": 0.00013609715562641115, |
|
"loss": 0.8256, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.7848882035466461, |
|
"grad_norm": 0.029499804601073265, |
|
"learning_rate": 0.0001358674401432774, |
|
"loss": 0.7145, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.7864302235929067, |
|
"grad_norm": 0.0285016018897295, |
|
"learning_rate": 0.00013563750724061025, |
|
"loss": 0.6712, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.7879722436391673, |
|
"grad_norm": 0.02717653289437294, |
|
"learning_rate": 0.0001354073583122059, |
|
"loss": 0.6807, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.7895142636854279, |
|
"grad_norm": 0.023634063079953194, |
|
"learning_rate": 0.00013517699475317017, |
|
"loss": 0.6677, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.7910562837316885, |
|
"grad_norm": 0.02920868992805481, |
|
"learning_rate": 0.00013494641795990986, |
|
"loss": 0.7865, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.7925983037779492, |
|
"grad_norm": 0.040133338421583176, |
|
"learning_rate": 0.00013471562933012432, |
|
"loss": 0.7786, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.7941403238242097, |
|
"grad_norm": 0.028838159516453743, |
|
"learning_rate": 0.00013448463026279704, |
|
"loss": 0.7689, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.7956823438704703, |
|
"grad_norm": 0.02688550017774105, |
|
"learning_rate": 0.00013425342215818718, |
|
"loss": 0.7659, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.7972243639167309, |
|
"grad_norm": 0.027155300602316856, |
|
"learning_rate": 0.0001340220064178209, |
|
"loss": 0.7353, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.7987663839629915, |
|
"grad_norm": 0.03566644340753555, |
|
"learning_rate": 0.00013379038444448306, |
|
"loss": 0.827, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.8003084040092521, |
|
"grad_norm": 0.0266974326223135, |
|
"learning_rate": 0.00013355855764220868, |
|
"loss": 0.7693, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.8018504240555128, |
|
"grad_norm": 0.026248447597026825, |
|
"learning_rate": 0.00013332652741627446, |
|
"loss": 0.7659, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.8033924441017734, |
|
"grad_norm": 0.030487187206745148, |
|
"learning_rate": 0.00013309429517318998, |
|
"loss": 0.7244, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.8049344641480339, |
|
"grad_norm": 0.031918346881866455, |
|
"learning_rate": 0.00013286186232068972, |
|
"loss": 0.6111, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.8064764841942945, |
|
"grad_norm": 0.03194589167833328, |
|
"learning_rate": 0.00013262923026772388, |
|
"loss": 0.7521, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.8080185042405551, |
|
"grad_norm": 0.029390091076493263, |
|
"learning_rate": 0.00013239640042445036, |
|
"loss": 0.7689, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.8095605242868157, |
|
"grad_norm": 0.025920424610376358, |
|
"learning_rate": 0.000132163374202226, |
|
"loss": 0.7102, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.8111025443330764, |
|
"grad_norm": 0.02453307807445526, |
|
"learning_rate": 0.000131930153013598, |
|
"loss": 0.712, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.812644564379337, |
|
"grad_norm": 0.035026635974645615, |
|
"learning_rate": 0.0001316967382722954, |
|
"loss": 0.789, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.8141865844255975, |
|
"grad_norm": 0.02895597368478775, |
|
"learning_rate": 0.0001314631313932205, |
|
"loss": 0.8087, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.8157286044718581, |
|
"grad_norm": 0.02956031821668148, |
|
"learning_rate": 0.00013122933379244034, |
|
"loss": 0.7514, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.8172706245181187, |
|
"grad_norm": 0.027792761102318764, |
|
"learning_rate": 0.00013099534688717804, |
|
"loss": 0.7843, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.8188126445643793, |
|
"grad_norm": 0.023245403543114662, |
|
"learning_rate": 0.00013076117209580418, |
|
"loss": 0.6878, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.8203546646106399, |
|
"grad_norm": 0.030602406710386276, |
|
"learning_rate": 0.00013052681083782836, |
|
"loss": 0.6504, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.8218966846569006, |
|
"grad_norm": 0.0250953808426857, |
|
"learning_rate": 0.00013029226453389043, |
|
"loss": 0.6255, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.8234387047031612, |
|
"grad_norm": 0.02791382372379303, |
|
"learning_rate": 0.00013005753460575194, |
|
"loss": 0.7252, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.8249807247494217, |
|
"grad_norm": 0.025774458423256874, |
|
"learning_rate": 0.0001298226224762876, |
|
"loss": 0.7252, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.8265227447956823, |
|
"grad_norm": 0.026806412264704704, |
|
"learning_rate": 0.00012958752956947645, |
|
"loss": 0.7702, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.8280647648419429, |
|
"grad_norm": 0.02904200740158558, |
|
"learning_rate": 0.00012935225731039348, |
|
"loss": 0.7573, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.8296067848882035, |
|
"grad_norm": 0.03018496371805668, |
|
"learning_rate": 0.00012911680712520082, |
|
"loss": 0.6854, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.8311488049344642, |
|
"grad_norm": 0.028921302407979965, |
|
"learning_rate": 0.00012888118044113912, |
|
"loss": 0.7372, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.8326908249807248, |
|
"grad_norm": 0.02912386879324913, |
|
"learning_rate": 0.00012864537868651892, |
|
"loss": 0.864, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.8342328450269854, |
|
"grad_norm": 0.03150784596800804, |
|
"learning_rate": 0.00012840940329071211, |
|
"loss": 0.6846, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.8357748650732459, |
|
"grad_norm": 0.026627201586961746, |
|
"learning_rate": 0.00012817325568414297, |
|
"loss": 0.6596, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.8373168851195065, |
|
"grad_norm": 0.026762191206216812, |
|
"learning_rate": 0.00012793693729827983, |
|
"loss": 0.7859, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.8388589051657671, |
|
"grad_norm": 0.0270906463265419, |
|
"learning_rate": 0.00012770044956562611, |
|
"loss": 0.76, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.8404009252120278, |
|
"grad_norm": 0.029391184449195862, |
|
"learning_rate": 0.0001274637939197119, |
|
"loss": 0.7153, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.8419429452582884, |
|
"grad_norm": 0.028473293408751488, |
|
"learning_rate": 0.00012722697179508508, |
|
"loss": 0.7083, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.843484965304549, |
|
"grad_norm": 0.028644869104027748, |
|
"learning_rate": 0.00012698998462730264, |
|
"loss": 0.7531, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.8450269853508096, |
|
"grad_norm": 0.0290384441614151, |
|
"learning_rate": 0.00012675283385292212, |
|
"loss": 0.8323, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.8465690053970701, |
|
"grad_norm": 0.027428725734353065, |
|
"learning_rate": 0.00012651552090949263, |
|
"loss": 0.783, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.8481110254433307, |
|
"grad_norm": 0.0316435806453228, |
|
"learning_rate": 0.00012627804723554651, |
|
"loss": 0.7303, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.8496530454895914, |
|
"grad_norm": 0.031257808208465576, |
|
"learning_rate": 0.00012604041427059036, |
|
"loss": 0.6641, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.851195065535852, |
|
"grad_norm": 0.031232839450240135, |
|
"learning_rate": 0.00012580262345509622, |
|
"loss": 0.8266, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.8527370855821126, |
|
"grad_norm": 0.028075871989130974, |
|
"learning_rate": 0.00012556467623049312, |
|
"loss": 0.6852, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.8542791056283732, |
|
"grad_norm": 0.0269178319722414, |
|
"learning_rate": 0.0001253265740391582, |
|
"loss": 0.6774, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.8558211256746338, |
|
"grad_norm": 0.033746860921382904, |
|
"learning_rate": 0.00012508831832440794, |
|
"loss": 0.6704, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.8573631457208943, |
|
"grad_norm": 0.02507774904370308, |
|
"learning_rate": 0.0001248499105304894, |
|
"loss": 0.6364, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.8589051657671549, |
|
"grad_norm": 0.027239350602030754, |
|
"learning_rate": 0.00012461135210257155, |
|
"loss": 0.7503, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.8604471858134156, |
|
"grad_norm": 0.031841401010751724, |
|
"learning_rate": 0.00012437264448673647, |
|
"loss": 0.7201, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.8619892058596762, |
|
"grad_norm": 0.025972798466682434, |
|
"learning_rate": 0.00012413378912997058, |
|
"loss": 0.7265, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.8635312259059368, |
|
"grad_norm": 0.02815602719783783, |
|
"learning_rate": 0.00012389478748015583, |
|
"loss": 0.6922, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.8650732459521974, |
|
"grad_norm": 0.03288137540221214, |
|
"learning_rate": 0.00012365564098606102, |
|
"loss": 0.6516, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.866615265998458, |
|
"grad_norm": 0.02655138447880745, |
|
"learning_rate": 0.00012341635109733293, |
|
"loss": 0.6192, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.8681572860447185, |
|
"grad_norm": 0.029187412932515144, |
|
"learning_rate": 0.00012317691926448754, |
|
"loss": 0.8307, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.8696993060909792, |
|
"grad_norm": 0.029966147616505623, |
|
"learning_rate": 0.00012293734693890132, |
|
"loss": 0.7285, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.8712413261372398, |
|
"grad_norm": 0.02818044275045395, |
|
"learning_rate": 0.0001226976355728023, |
|
"loss": 0.7171, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.8727833461835004, |
|
"grad_norm": 0.02721909061074257, |
|
"learning_rate": 0.00012245778661926137, |
|
"loss": 0.6913, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.874325366229761, |
|
"grad_norm": 0.030929675325751305, |
|
"learning_rate": 0.0001222178015321835, |
|
"loss": 0.8006, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.8758673862760216, |
|
"grad_norm": 0.026720581576228142, |
|
"learning_rate": 0.00012197768176629876, |
|
"loss": 0.7348, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.8774094063222821, |
|
"grad_norm": 0.02792746014893055, |
|
"learning_rate": 0.00012173742877715373, |
|
"loss": 0.786, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.8789514263685428, |
|
"grad_norm": 0.025372346863150597, |
|
"learning_rate": 0.00012149704402110243, |
|
"loss": 0.6832, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.8804934464148034, |
|
"grad_norm": 0.025779495015740395, |
|
"learning_rate": 0.00012125652895529766, |
|
"loss": 0.7465, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.882035466461064, |
|
"grad_norm": 0.028759067878127098, |
|
"learning_rate": 0.00012101588503768224, |
|
"loss": 0.7755, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.8835774865073246, |
|
"grad_norm": 0.0319170206785202, |
|
"learning_rate": 0.00012077511372697985, |
|
"loss": 0.7758, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.8851195065535852, |
|
"grad_norm": 0.0287742018699646, |
|
"learning_rate": 0.00012053421648268662, |
|
"loss": 0.7556, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.8866615265998458, |
|
"grad_norm": 0.027272436767816544, |
|
"learning_rate": 0.00012029319476506182, |
|
"loss": 0.6341, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.8882035466461063, |
|
"grad_norm": 0.02786344476044178, |
|
"learning_rate": 0.00012005205003511948, |
|
"loss": 0.7248, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.889745566692367, |
|
"grad_norm": 0.023950345814228058, |
|
"learning_rate": 0.00011981078375461915, |
|
"loss": 0.6907, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.8912875867386276, |
|
"grad_norm": 0.0248698852956295, |
|
"learning_rate": 0.00011956939738605721, |
|
"loss": 0.6612, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.8928296067848882, |
|
"grad_norm": 0.02722037211060524, |
|
"learning_rate": 0.00011932789239265802, |
|
"loss": 0.6179, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.8943716268311488, |
|
"grad_norm": 0.02561650238931179, |
|
"learning_rate": 0.00011908627023836503, |
|
"loss": 0.7478, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.8959136468774094, |
|
"grad_norm": 0.027791699394583702, |
|
"learning_rate": 0.00011884453238783185, |
|
"loss": 0.7803, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.89745566692367, |
|
"grad_norm": 0.033503565937280655, |
|
"learning_rate": 0.00011860268030641338, |
|
"loss": 0.7978, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.8989976869699307, |
|
"grad_norm": 0.0275451447814703, |
|
"learning_rate": 0.00011836071546015703, |
|
"loss": 0.7458, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.9005397070161912, |
|
"grad_norm": 0.02638075314462185, |
|
"learning_rate": 0.00011811863931579377, |
|
"loss": 0.5823, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.9020817270624518, |
|
"grad_norm": 0.029536547139286995, |
|
"learning_rate": 0.00011787645334072913, |
|
"loss": 0.829, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.9036237471087124, |
|
"grad_norm": 0.029731806367635727, |
|
"learning_rate": 0.0001176341590030345, |
|
"loss": 0.7553, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.905165767154973, |
|
"grad_norm": 0.02816937118768692, |
|
"learning_rate": 0.00011739175777143812, |
|
"loss": 0.7094, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.9067077872012336, |
|
"grad_norm": 0.026959970593452454, |
|
"learning_rate": 0.00011714925111531619, |
|
"loss": 0.7368, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.9082498072474943, |
|
"grad_norm": 0.027997490018606186, |
|
"learning_rate": 0.0001169066405046839, |
|
"loss": 0.7864, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.9097918272937549, |
|
"grad_norm": 0.025101030245423317, |
|
"learning_rate": 0.00011666392741018675, |
|
"loss": 0.6445, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.9113338473400154, |
|
"grad_norm": 0.038789402693510056, |
|
"learning_rate": 0.00011642111330309129, |
|
"loss": 0.7104, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.912875867386276, |
|
"grad_norm": 0.0264846533536911, |
|
"learning_rate": 0.0001161781996552765, |
|
"loss": 0.7034, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.9144178874325366, |
|
"grad_norm": 0.03629022464156151, |
|
"learning_rate": 0.00011593518793922468, |
|
"loss": 0.8121, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.9159599074787972, |
|
"grad_norm": 0.025432435795664787, |
|
"learning_rate": 0.00011569207962801263, |
|
"loss": 0.6726, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.9175019275250579, |
|
"grad_norm": 0.02553374320268631, |
|
"learning_rate": 0.00011544887619530275, |
|
"loss": 0.659, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.9190439475713185, |
|
"grad_norm": 0.02498779632151127, |
|
"learning_rate": 0.0001152055791153339, |
|
"loss": 0.5868, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.920585967617579, |
|
"grad_norm": 0.029489582404494286, |
|
"learning_rate": 0.00011496218986291273, |
|
"loss": 0.673, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.9221279876638396, |
|
"grad_norm": 0.029334766790270805, |
|
"learning_rate": 0.00011471870991340459, |
|
"loss": 0.8011, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.9236700077101002, |
|
"grad_norm": 0.0274631530046463, |
|
"learning_rate": 0.00011447514074272451, |
|
"loss": 0.721, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.9252120277563608, |
|
"grad_norm": 0.028073586523532867, |
|
"learning_rate": 0.00011423148382732853, |
|
"loss": 0.839, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.9267540478026214, |
|
"grad_norm": 0.028757184743881226, |
|
"learning_rate": 0.00011398774064420443, |
|
"loss": 0.7409, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.9282960678488821, |
|
"grad_norm": 0.026308685541152954, |
|
"learning_rate": 0.00011374391267086302, |
|
"loss": 0.7517, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.9298380878951427, |
|
"grad_norm": 0.027558207511901855, |
|
"learning_rate": 0.00011350000138532902, |
|
"loss": 0.7041, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.9313801079414032, |
|
"grad_norm": 0.02995571680366993, |
|
"learning_rate": 0.0001132560082661322, |
|
"loss": 0.842, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.9329221279876638, |
|
"grad_norm": 0.028026850894093513, |
|
"learning_rate": 0.00011301193479229841, |
|
"loss": 0.7639, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.9344641480339244, |
|
"grad_norm": 0.026926511898636818, |
|
"learning_rate": 0.00011276778244334055, |
|
"loss": 0.6315, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.936006168080185, |
|
"grad_norm": 0.031490955501794815, |
|
"learning_rate": 0.00011252355269924963, |
|
"loss": 0.5844, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.9375481881264457, |
|
"grad_norm": 0.028631744906306267, |
|
"learning_rate": 0.00011227924704048585, |
|
"loss": 0.6431, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.9390902081727063, |
|
"grad_norm": 0.029599271714687347, |
|
"learning_rate": 0.00011203486694796957, |
|
"loss": 0.7085, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.9406322282189669, |
|
"grad_norm": 0.027870824560523033, |
|
"learning_rate": 0.00011179041390307235, |
|
"loss": 0.7381, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.9421742482652274, |
|
"grad_norm": 0.030139662325382233, |
|
"learning_rate": 0.00011154588938760794, |
|
"loss": 0.7299, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.943716268311488, |
|
"grad_norm": 0.031028704717755318, |
|
"learning_rate": 0.00011130129488382341, |
|
"loss": 0.747, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.9452582883577486, |
|
"grad_norm": 0.02865663915872574, |
|
"learning_rate": 0.00011105663187438997, |
|
"loss": 0.7025, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.9468003084040093, |
|
"grad_norm": 0.029388774186372757, |
|
"learning_rate": 0.00011081190184239419, |
|
"loss": 0.6645, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.9483423284502699, |
|
"grad_norm": 0.023502621799707413, |
|
"learning_rate": 0.00011056710627132884, |
|
"loss": 0.6455, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.9498843484965305, |
|
"grad_norm": 0.02663263864815235, |
|
"learning_rate": 0.00011032224664508405, |
|
"loss": 0.7677, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.9514263685427911, |
|
"grad_norm": 0.02990710362792015, |
|
"learning_rate": 0.00011007732444793814, |
|
"loss": 0.7017, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.9529683885890516, |
|
"grad_norm": 0.029992980882525444, |
|
"learning_rate": 0.00010983234116454886, |
|
"loss": 0.7969, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.9545104086353122, |
|
"grad_norm": 0.03714431822299957, |
|
"learning_rate": 0.00010958729827994404, |
|
"loss": 0.7646, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.9560524286815728, |
|
"grad_norm": 0.030639756470918655, |
|
"learning_rate": 0.00010934219727951301, |
|
"loss": 0.8926, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.9575944487278335, |
|
"grad_norm": 0.030243555083870888, |
|
"learning_rate": 0.00010909703964899729, |
|
"loss": 0.6676, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.9591364687740941, |
|
"grad_norm": 0.029893988743424416, |
|
"learning_rate": 0.00010885182687448161, |
|
"loss": 0.7238, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.9606784888203547, |
|
"grad_norm": 0.02677508443593979, |
|
"learning_rate": 0.00010860656044238511, |
|
"loss": 0.7378, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.9622205088666153, |
|
"grad_norm": 0.02502857707440853, |
|
"learning_rate": 0.00010836124183945208, |
|
"loss": 0.7129, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.9637625289128758, |
|
"grad_norm": 0.02673465758562088, |
|
"learning_rate": 0.00010811587255274313, |
|
"loss": 0.7046, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.9653045489591364, |
|
"grad_norm": 0.030728284269571304, |
|
"learning_rate": 0.00010787045406962607, |
|
"loss": 0.752, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.9668465690053971, |
|
"grad_norm": 0.02698132023215294, |
|
"learning_rate": 0.00010762498787776687, |
|
"loss": 0.6365, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.9683885890516577, |
|
"grad_norm": 0.03070535510778427, |
|
"learning_rate": 0.0001073794754651208, |
|
"loss": 0.7465, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.9699306090979183, |
|
"grad_norm": 0.026031676679849625, |
|
"learning_rate": 0.00010713391831992323, |
|
"loss": 0.6991, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 0.9714726291441789, |
|
"grad_norm": 0.02689627930521965, |
|
"learning_rate": 0.00010688831793068078, |
|
"loss": 0.6468, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.9730146491904395, |
|
"grad_norm": 0.028856465592980385, |
|
"learning_rate": 0.00010664267578616207, |
|
"loss": 0.7051, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 0.9745566692367, |
|
"grad_norm": 0.027791066095232964, |
|
"learning_rate": 0.00010639699337538898, |
|
"loss": 0.7247, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.9760986892829607, |
|
"grad_norm": 0.02929227240383625, |
|
"learning_rate": 0.00010615127218762733, |
|
"loss": 0.6994, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.9776407093292213, |
|
"grad_norm": 0.0322168804705143, |
|
"learning_rate": 0.0001059055137123781, |
|
"loss": 0.753, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.9791827293754819, |
|
"grad_norm": 0.02872268296778202, |
|
"learning_rate": 0.00010565971943936825, |
|
"loss": 0.7451, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.9807247494217425, |
|
"grad_norm": 0.03104063868522644, |
|
"learning_rate": 0.00010541389085854176, |
|
"loss": 0.7482, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.9822667694680031, |
|
"grad_norm": 0.026641814038157463, |
|
"learning_rate": 0.00010516802946005058, |
|
"loss": 0.7034, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 0.9838087895142636, |
|
"grad_norm": 0.027356311678886414, |
|
"learning_rate": 0.00010492213673424553, |
|
"loss": 0.6459, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.9853508095605242, |
|
"grad_norm": 0.031015669927001, |
|
"learning_rate": 0.00010467621417166745, |
|
"loss": 0.8185, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.9868928296067849, |
|
"grad_norm": 0.024331575259566307, |
|
"learning_rate": 0.00010443026326303789, |
|
"loss": 0.64, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.9884348496530455, |
|
"grad_norm": 0.026257265359163284, |
|
"learning_rate": 0.00010418428549925032, |
|
"loss": 0.6634, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 0.9899768696993061, |
|
"grad_norm": 0.026038171723484993, |
|
"learning_rate": 0.00010393828237136107, |
|
"loss": 0.728, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.9915188897455667, |
|
"grad_norm": 0.028959324583411217, |
|
"learning_rate": 0.00010369225537058002, |
|
"loss": 0.7315, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 0.9930609097918273, |
|
"grad_norm": 0.026242226362228394, |
|
"learning_rate": 0.00010344620598826198, |
|
"loss": 0.7096, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.9946029298380878, |
|
"grad_norm": 0.029856307432055473, |
|
"learning_rate": 0.00010320013571589726, |
|
"loss": 0.7052, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.9961449498843485, |
|
"grad_norm": 0.03145446628332138, |
|
"learning_rate": 0.00010295404604510286, |
|
"loss": 0.8392, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.9976869699306091, |
|
"grad_norm": 0.032184895128011703, |
|
"learning_rate": 0.00010270793846761347, |
|
"loss": 0.7936, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 0.9992289899768697, |
|
"grad_norm": 0.027684815227985382, |
|
"learning_rate": 0.00010246181447527212, |
|
"loss": 0.6861, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.038702599704265594, |
|
"learning_rate": 0.00010221567556002153, |
|
"loss": 0.7216, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.0138009786605835, |
|
"eval_runtime": 121.2186, |
|
"eval_samples_per_second": 13.975, |
|
"eval_steps_per_second": 1.749, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 1.0015420200462606, |
|
"grad_norm": 0.04231274873018265, |
|
"learning_rate": 0.00010196952321389482, |
|
"loss": 0.5952, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.0030840400925212, |
|
"grad_norm": 0.031240805983543396, |
|
"learning_rate": 0.00010172335892900645, |
|
"loss": 0.7104, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 1.0046260601387818, |
|
"grad_norm": 0.030615027993917465, |
|
"learning_rate": 0.00010147718419754336, |
|
"loss": 0.6679, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 1.0061680801850423, |
|
"grad_norm": 0.029420683160424232, |
|
"learning_rate": 0.00010123100051175567, |
|
"loss": 0.6429, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 1.007710100231303, |
|
"grad_norm": 0.030873583629727364, |
|
"learning_rate": 0.00010098480936394801, |
|
"loss": 0.658, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 1.0092521202775635, |
|
"grad_norm": 0.031131163239479065, |
|
"learning_rate": 0.00010073861224647, |
|
"loss": 0.6273, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 1.0107941403238243, |
|
"grad_norm": 0.03470936417579651, |
|
"learning_rate": 0.00010049241065170765, |
|
"loss": 0.5581, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 1.012336160370085, |
|
"grad_norm": 0.03058730624616146, |
|
"learning_rate": 0.00010024620607207393, |
|
"loss": 0.6956, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 1.0138781804163455, |
|
"grad_norm": 0.02509419433772564, |
|
"learning_rate": 0.0001, |
|
"loss": 0.5643, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 1.015420200462606, |
|
"grad_norm": 0.03301594778895378, |
|
"learning_rate": 9.975379392792609e-05, |
|
"loss": 0.6595, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 1.0169622205088666, |
|
"grad_norm": 0.03824777528643608, |
|
"learning_rate": 9.950758934829241e-05, |
|
"loss": 0.6632, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.0185042405551272, |
|
"grad_norm": 0.04577745869755745, |
|
"learning_rate": 9.926138775352998e-05, |
|
"loss": 0.6584, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 1.0200462606013878, |
|
"grad_norm": 0.03415631502866745, |
|
"learning_rate": 9.901519063605203e-05, |
|
"loss": 0.6815, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 1.0215882806476484, |
|
"grad_norm": 0.03187074884772301, |
|
"learning_rate": 9.876899948824434e-05, |
|
"loss": 0.6878, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 1.023130300693909, |
|
"grad_norm": 0.03230864182114601, |
|
"learning_rate": 9.85228158024567e-05, |
|
"loss": 0.7208, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 1.0246723207401696, |
|
"grad_norm": 0.03281601890921593, |
|
"learning_rate": 9.827664107099359e-05, |
|
"loss": 0.6209, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 1.0262143407864301, |
|
"grad_norm": 0.030430031940340996, |
|
"learning_rate": 9.80304767861052e-05, |
|
"loss": 0.66, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 1.0277563608326907, |
|
"grad_norm": 0.027022497728466988, |
|
"learning_rate": 9.778432443997847e-05, |
|
"loss": 0.601, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 1.0292983808789515, |
|
"grad_norm": 0.03322981297969818, |
|
"learning_rate": 9.75381855247279e-05, |
|
"loss": 0.6658, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 1.0308404009252121, |
|
"grad_norm": 0.02994593232870102, |
|
"learning_rate": 9.729206153238657e-05, |
|
"loss": 0.6471, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 1.0323824209714727, |
|
"grad_norm": 0.027031417936086655, |
|
"learning_rate": 9.704595395489714e-05, |
|
"loss": 0.6096, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.0339244410177333, |
|
"grad_norm": 0.029254987835884094, |
|
"learning_rate": 9.679986428410276e-05, |
|
"loss": 0.6592, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 1.0354664610639939, |
|
"grad_norm": 0.030262276530265808, |
|
"learning_rate": 9.655379401173804e-05, |
|
"loss": 0.6317, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 1.0370084811102545, |
|
"grad_norm": 0.03326797112822533, |
|
"learning_rate": 9.630774462942e-05, |
|
"loss": 0.6071, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 1.038550501156515, |
|
"grad_norm": 0.029893390834331512, |
|
"learning_rate": 9.606171762863898e-05, |
|
"loss": 0.6497, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 1.0400925212027756, |
|
"grad_norm": 0.03093409165740013, |
|
"learning_rate": 9.581571450074968e-05, |
|
"loss": 0.6562, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.0416345412490362, |
|
"grad_norm": 0.030291495844721794, |
|
"learning_rate": 9.556973673696213e-05, |
|
"loss": 0.7115, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 1.0431765612952968, |
|
"grad_norm": 0.03254539519548416, |
|
"learning_rate": 9.532378582833259e-05, |
|
"loss": 0.6922, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 1.0447185813415574, |
|
"grad_norm": 0.028841672465205193, |
|
"learning_rate": 9.50778632657545e-05, |
|
"loss": 0.5989, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 1.046260601387818, |
|
"grad_norm": 0.03121812455356121, |
|
"learning_rate": 9.483197053994947e-05, |
|
"loss": 0.6697, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 1.0478026214340785, |
|
"grad_norm": 0.03435603901743889, |
|
"learning_rate": 9.458610914145826e-05, |
|
"loss": 0.7151, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.0493446414803393, |
|
"grad_norm": 0.03248435631394386, |
|
"learning_rate": 9.434028056063177e-05, |
|
"loss": 0.6599, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 1.0508866615266, |
|
"grad_norm": 0.0340297594666481, |
|
"learning_rate": 9.409448628762193e-05, |
|
"loss": 0.6588, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 1.0524286815728605, |
|
"grad_norm": 0.031176073476672173, |
|
"learning_rate": 9.384872781237269e-05, |
|
"loss": 0.7002, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 1.053970701619121, |
|
"grad_norm": 0.031643763184547424, |
|
"learning_rate": 9.360300662461103e-05, |
|
"loss": 0.6481, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 1.0555127216653817, |
|
"grad_norm": 0.03346734866499901, |
|
"learning_rate": 9.335732421383794e-05, |
|
"loss": 0.605, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 1.0570547417116423, |
|
"grad_norm": 0.028062671422958374, |
|
"learning_rate": 9.311168206931925e-05, |
|
"loss": 0.5722, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 1.0585967617579028, |
|
"grad_norm": 0.03068237006664276, |
|
"learning_rate": 9.286608168007678e-05, |
|
"loss": 0.6179, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 1.0601387818041634, |
|
"grad_norm": 0.03420115262269974, |
|
"learning_rate": 9.262052453487924e-05, |
|
"loss": 0.7138, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 1.061680801850424, |
|
"grad_norm": 0.03138900548219681, |
|
"learning_rate": 9.237501212223314e-05, |
|
"loss": 0.6974, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 1.0632228218966846, |
|
"grad_norm": 0.030687514692544937, |
|
"learning_rate": 9.212954593037394e-05, |
|
"loss": 0.6116, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.0647648419429452, |
|
"grad_norm": 0.03098466247320175, |
|
"learning_rate": 9.18841274472569e-05, |
|
"loss": 0.6968, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 1.0663068619892058, |
|
"grad_norm": 0.030333001166582108, |
|
"learning_rate": 9.163875816054794e-05, |
|
"loss": 0.6014, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 1.0678488820354666, |
|
"grad_norm": 0.031592730432748795, |
|
"learning_rate": 9.139343955761493e-05, |
|
"loss": 0.6784, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 1.0693909020817272, |
|
"grad_norm": 0.03480207920074463, |
|
"learning_rate": 9.11481731255184e-05, |
|
"loss": 0.7779, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 1.0709329221279877, |
|
"grad_norm": 0.03373701125383377, |
|
"learning_rate": 9.090296035100275e-05, |
|
"loss": 0.7062, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 1.0724749421742483, |
|
"grad_norm": 0.03207295015454292, |
|
"learning_rate": 9.065780272048701e-05, |
|
"loss": 0.7564, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 1.074016962220509, |
|
"grad_norm": 0.03260812535881996, |
|
"learning_rate": 9.041270172005598e-05, |
|
"loss": 0.663, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 1.0755589822667695, |
|
"grad_norm": 0.03293580561876297, |
|
"learning_rate": 9.016765883545117e-05, |
|
"loss": 0.6174, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 1.07710100231303, |
|
"grad_norm": 0.03376418352127075, |
|
"learning_rate": 8.992267555206185e-05, |
|
"loss": 0.6528, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 1.0786430223592907, |
|
"grad_norm": 0.040761563926935196, |
|
"learning_rate": 8.967775335491595e-05, |
|
"loss": 0.6229, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.0801850424055512, |
|
"grad_norm": 0.03353280574083328, |
|
"learning_rate": 8.943289372867118e-05, |
|
"loss": 0.6533, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 1.0817270624518118, |
|
"grad_norm": 0.031145334243774414, |
|
"learning_rate": 8.918809815760585e-05, |
|
"loss": 0.6625, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 1.0832690824980724, |
|
"grad_norm": 0.037107136100530624, |
|
"learning_rate": 8.894336812561005e-05, |
|
"loss": 0.6317, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 1.084811102544333, |
|
"grad_norm": 0.031154975295066833, |
|
"learning_rate": 8.869870511617661e-05, |
|
"loss": 0.6316, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 1.0863531225905936, |
|
"grad_norm": 0.029372645542025566, |
|
"learning_rate": 8.845411061239207e-05, |
|
"loss": 0.5326, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 1.0878951426368544, |
|
"grad_norm": 0.03378273546695709, |
|
"learning_rate": 8.820958609692769e-05, |
|
"loss": 0.6289, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 1.089437162683115, |
|
"grad_norm": 0.027697058394551277, |
|
"learning_rate": 8.796513305203048e-05, |
|
"loss": 0.6002, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 1.0909791827293756, |
|
"grad_norm": 0.03780931979417801, |
|
"learning_rate": 8.772075295951415e-05, |
|
"loss": 0.5932, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 1.0925212027756361, |
|
"grad_norm": 0.03167394921183586, |
|
"learning_rate": 8.74764473007504e-05, |
|
"loss": 0.6117, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 1.0940632228218967, |
|
"grad_norm": 0.03654084354639053, |
|
"learning_rate": 8.723221755665948e-05, |
|
"loss": 0.7338, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.0956052428681573, |
|
"grad_norm": 0.035005394369363785, |
|
"learning_rate": 8.698806520770161e-05, |
|
"loss": 0.738, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 1.0971472629144179, |
|
"grad_norm": 0.03315757215023041, |
|
"learning_rate": 8.674399173386779e-05, |
|
"loss": 0.5979, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 1.0986892829606785, |
|
"grad_norm": 0.042626503854990005, |
|
"learning_rate": 8.649999861467099e-05, |
|
"loss": 0.6772, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 1.100231303006939, |
|
"grad_norm": 0.035321805626153946, |
|
"learning_rate": 8.625608732913701e-05, |
|
"loss": 0.627, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 1.1017733230531996, |
|
"grad_norm": 0.0311568696051836, |
|
"learning_rate": 8.60122593557956e-05, |
|
"loss": 0.6204, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 1.1033153430994602, |
|
"grad_norm": 0.03266143798828125, |
|
"learning_rate": 8.57685161726715e-05, |
|
"loss": 0.7405, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 1.1048573631457208, |
|
"grad_norm": 0.03293644264340401, |
|
"learning_rate": 8.55248592572755e-05, |
|
"loss": 0.7476, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 1.1063993831919814, |
|
"grad_norm": 0.03008246421813965, |
|
"learning_rate": 8.528129008659544e-05, |
|
"loss": 0.6292, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 1.1079414032382422, |
|
"grad_norm": 0.029346976429224014, |
|
"learning_rate": 8.503781013708729e-05, |
|
"loss": 0.6413, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 1.1094834232845028, |
|
"grad_norm": 0.0354442335665226, |
|
"learning_rate": 8.479442088466611e-05, |
|
"loss": 0.6475, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.1110254433307634, |
|
"grad_norm": 0.030347274616360664, |
|
"learning_rate": 8.455112380469729e-05, |
|
"loss": 0.6532, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 1.112567463377024, |
|
"grad_norm": 0.03859079256653786, |
|
"learning_rate": 8.430792037198737e-05, |
|
"loss": 0.7131, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 1.1141094834232845, |
|
"grad_norm": 0.0332760214805603, |
|
"learning_rate": 8.406481206077534e-05, |
|
"loss": 0.5202, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 1.1156515034695451, |
|
"grad_norm": 0.03403365612030029, |
|
"learning_rate": 8.382180034472353e-05, |
|
"loss": 0.7009, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 1.1171935235158057, |
|
"grad_norm": 0.03711424767971039, |
|
"learning_rate": 8.357888669690876e-05, |
|
"loss": 0.7204, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.1187355435620663, |
|
"grad_norm": 0.030927130952477455, |
|
"learning_rate": 8.333607258981329e-05, |
|
"loss": 0.636, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 1.1202775636083269, |
|
"grad_norm": 0.03724541887640953, |
|
"learning_rate": 8.30933594953161e-05, |
|
"loss": 0.6751, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 1.1218195836545874, |
|
"grad_norm": 0.036198195070028305, |
|
"learning_rate": 8.285074888468383e-05, |
|
"loss": 0.7036, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 1.123361603700848, |
|
"grad_norm": 0.0362878255546093, |
|
"learning_rate": 8.26082422285619e-05, |
|
"loss": 0.6865, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 1.1249036237471086, |
|
"grad_norm": 0.03348914906382561, |
|
"learning_rate": 8.236584099696554e-05, |
|
"loss": 0.7005, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.1264456437933692, |
|
"grad_norm": 0.036960627883672714, |
|
"learning_rate": 8.212354665927088e-05, |
|
"loss": 0.719, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 1.12798766383963, |
|
"grad_norm": 0.03627372533082962, |
|
"learning_rate": 8.188136068420627e-05, |
|
"loss": 0.6358, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 1.1295296838858906, |
|
"grad_norm": 0.03624941408634186, |
|
"learning_rate": 8.163928453984298e-05, |
|
"loss": 0.6836, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 1.1310717039321512, |
|
"grad_norm": 0.03319893777370453, |
|
"learning_rate": 8.139731969358663e-05, |
|
"loss": 0.6449, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 1.1326137239784118, |
|
"grad_norm": 0.03332339972257614, |
|
"learning_rate": 8.115546761216822e-05, |
|
"loss": 0.7252, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 1.1341557440246723, |
|
"grad_norm": 0.030635269358754158, |
|
"learning_rate": 8.091372976163495e-05, |
|
"loss": 0.6092, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 1.135697764070933, |
|
"grad_norm": 0.03975079208612442, |
|
"learning_rate": 8.067210760734199e-05, |
|
"loss": 0.7069, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 1.1372397841171935, |
|
"grad_norm": 0.036304909735918045, |
|
"learning_rate": 8.043060261394282e-05, |
|
"loss": 0.6257, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 1.138781804163454, |
|
"grad_norm": 0.0363340899348259, |
|
"learning_rate": 8.01892162453809e-05, |
|
"loss": 0.6912, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 1.1403238242097147, |
|
"grad_norm": 0.030387498438358307, |
|
"learning_rate": 7.994794996488055e-05, |
|
"loss": 0.6096, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.1418658442559753, |
|
"grad_norm": 0.03291669860482216, |
|
"learning_rate": 7.970680523493818e-05, |
|
"loss": 0.6629, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 1.1434078643022358, |
|
"grad_norm": 0.03067462518811226, |
|
"learning_rate": 7.94657835173134e-05, |
|
"loss": 0.6726, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 1.1449498843484966, |
|
"grad_norm": 0.03496173024177551, |
|
"learning_rate": 7.922488627302015e-05, |
|
"loss": 0.7331, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 1.146491904394757, |
|
"grad_norm": 0.03458810970187187, |
|
"learning_rate": 7.89841149623178e-05, |
|
"loss": 0.6987, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 1.1480339244410178, |
|
"grad_norm": 0.03519074246287346, |
|
"learning_rate": 7.874347104470234e-05, |
|
"loss": 0.5909, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 1.1495759444872784, |
|
"grad_norm": 0.03368091583251953, |
|
"learning_rate": 7.85029559788976e-05, |
|
"loss": 0.6144, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 1.151117964533539, |
|
"grad_norm": 0.03208000957965851, |
|
"learning_rate": 7.826257122284629e-05, |
|
"loss": 0.6479, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 1.1526599845797996, |
|
"grad_norm": 0.03380461782217026, |
|
"learning_rate": 7.802231823370125e-05, |
|
"loss": 0.7049, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 1.1542020046260602, |
|
"grad_norm": 0.03315620869398117, |
|
"learning_rate": 7.778219846781653e-05, |
|
"loss": 0.6683, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 1.1557440246723207, |
|
"grad_norm": 0.03291339799761772, |
|
"learning_rate": 7.754221338073864e-05, |
|
"loss": 0.6011, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.1572860447185813, |
|
"grad_norm": 0.031395379453897476, |
|
"learning_rate": 7.730236442719774e-05, |
|
"loss": 0.5657, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 1.158828064764842, |
|
"grad_norm": 0.03961668163537979, |
|
"learning_rate": 7.706265306109872e-05, |
|
"loss": 0.6831, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 1.1603700848111025, |
|
"grad_norm": 0.031115271151065826, |
|
"learning_rate": 7.68230807355125e-05, |
|
"loss": 0.5597, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 1.161912104857363, |
|
"grad_norm": 0.030850525945425034, |
|
"learning_rate": 7.658364890266711e-05, |
|
"loss": 0.6421, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 1.1634541249036237, |
|
"grad_norm": 0.03267287090420723, |
|
"learning_rate": 7.634435901393899e-05, |
|
"loss": 0.6157, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 1.1649961449498845, |
|
"grad_norm": 0.03116137906908989, |
|
"learning_rate": 7.610521251984419e-05, |
|
"loss": 0.6098, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 1.166538164996145, |
|
"grad_norm": 0.03797944262623787, |
|
"learning_rate": 7.586621087002945e-05, |
|
"loss": 0.6847, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 1.1680801850424056, |
|
"grad_norm": 0.03272239863872528, |
|
"learning_rate": 7.562735551326355e-05, |
|
"loss": 0.637, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 1.1696222050886662, |
|
"grad_norm": 0.03280835226178169, |
|
"learning_rate": 7.538864789742845e-05, |
|
"loss": 0.629, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 1.1711642251349268, |
|
"grad_norm": 0.03413494676351547, |
|
"learning_rate": 7.51500894695106e-05, |
|
"loss": 0.6661, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.1727062451811874, |
|
"grad_norm": 0.030523747205734253, |
|
"learning_rate": 7.491168167559209e-05, |
|
"loss": 0.5446, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 1.174248265227448, |
|
"grad_norm": 0.03329429030418396, |
|
"learning_rate": 7.467342596084179e-05, |
|
"loss": 0.6319, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 1.1757902852737085, |
|
"grad_norm": 0.0314452238380909, |
|
"learning_rate": 7.443532376950688e-05, |
|
"loss": 0.6161, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 1.1773323053199691, |
|
"grad_norm": 0.03715851530432701, |
|
"learning_rate": 7.419737654490379e-05, |
|
"loss": 0.7136, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 1.1788743253662297, |
|
"grad_norm": 0.03489915654063225, |
|
"learning_rate": 7.395958572940966e-05, |
|
"loss": 0.6072, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 1.1804163454124903, |
|
"grad_norm": 0.0346875824034214, |
|
"learning_rate": 7.37219527644535e-05, |
|
"loss": 0.5942, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 1.1819583654587509, |
|
"grad_norm": 0.03557261452078819, |
|
"learning_rate": 7.34844790905074e-05, |
|
"loss": 0.6739, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 1.1835003855050115, |
|
"grad_norm": 0.03237386420369148, |
|
"learning_rate": 7.324716614707793e-05, |
|
"loss": 0.7092, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 1.1850424055512723, |
|
"grad_norm": 0.035937052220106125, |
|
"learning_rate": 7.301001537269736e-05, |
|
"loss": 0.6396, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 1.1865844255975329, |
|
"grad_norm": 0.03329618275165558, |
|
"learning_rate": 7.277302820491491e-05, |
|
"loss": 0.6302, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.1881264456437934, |
|
"grad_norm": 0.03639035299420357, |
|
"learning_rate": 7.25362060802881e-05, |
|
"loss": 0.6404, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 1.189668465690054, |
|
"grad_norm": 0.03558572009205818, |
|
"learning_rate": 7.229955043437391e-05, |
|
"loss": 0.5638, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 1.1912104857363146, |
|
"grad_norm": 0.03477177023887634, |
|
"learning_rate": 7.206306270172019e-05, |
|
"loss": 0.5829, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 1.1927525057825752, |
|
"grad_norm": 0.03548658639192581, |
|
"learning_rate": 7.182674431585704e-05, |
|
"loss": 0.5756, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 1.1942945258288358, |
|
"grad_norm": 0.03554500639438629, |
|
"learning_rate": 7.159059670928791e-05, |
|
"loss": 0.6173, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 1.1958365458750964, |
|
"grad_norm": 0.03374260663986206, |
|
"learning_rate": 7.135462131348107e-05, |
|
"loss": 0.6626, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 1.197378565921357, |
|
"grad_norm": 0.034156035631895065, |
|
"learning_rate": 7.111881955886093e-05, |
|
"loss": 0.6719, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 1.1989205859676175, |
|
"grad_norm": 0.03379127383232117, |
|
"learning_rate": 7.088319287479919e-05, |
|
"loss": 0.5903, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 1.200462606013878, |
|
"grad_norm": 0.035499654710292816, |
|
"learning_rate": 7.064774268960653e-05, |
|
"loss": 0.6262, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 1.2020046260601387, |
|
"grad_norm": 0.03328956291079521, |
|
"learning_rate": 7.041247043052357e-05, |
|
"loss": 0.5538, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.2035466461063993, |
|
"grad_norm": 0.03272480517625809, |
|
"learning_rate": 7.017737752371244e-05, |
|
"loss": 0.6044, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 1.20508866615266, |
|
"grad_norm": 0.036613889038562775, |
|
"learning_rate": 6.994246539424808e-05, |
|
"loss": 0.6649, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 1.2066306861989207, |
|
"grad_norm": 0.03719131276011467, |
|
"learning_rate": 6.970773546610958e-05, |
|
"loss": 0.6586, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 1.2081727062451812, |
|
"grad_norm": 0.0346304327249527, |
|
"learning_rate": 6.947318916217168e-05, |
|
"loss": 0.657, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 1.2097147262914418, |
|
"grad_norm": 0.04230980947613716, |
|
"learning_rate": 6.923882790419585e-05, |
|
"loss": 0.8137, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 1.2112567463377024, |
|
"grad_norm": 0.036169420927762985, |
|
"learning_rate": 6.900465311282201e-05, |
|
"loss": 0.6562, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 1.212798766383963, |
|
"grad_norm": 0.03368053212761879, |
|
"learning_rate": 6.87706662075597e-05, |
|
"loss": 0.6302, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 1.2143407864302236, |
|
"grad_norm": 0.028687801212072372, |
|
"learning_rate": 6.853686860677949e-05, |
|
"loss": 0.5765, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 1.2158828064764842, |
|
"grad_norm": 0.03682614117860794, |
|
"learning_rate": 6.830326172770463e-05, |
|
"loss": 0.728, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 1.2174248265227448, |
|
"grad_norm": 0.032425619661808014, |
|
"learning_rate": 6.806984698640202e-05, |
|
"loss": 0.6618, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.2189668465690053, |
|
"grad_norm": 0.04002368077635765, |
|
"learning_rate": 6.783662579777401e-05, |
|
"loss": 0.6823, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 1.220508866615266, |
|
"grad_norm": 0.04177991300821304, |
|
"learning_rate": 6.760359957554963e-05, |
|
"loss": 0.6711, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 1.2220508866615265, |
|
"grad_norm": 0.0367995947599411, |
|
"learning_rate": 6.737076973227614e-05, |
|
"loss": 0.7178, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 1.223592906707787, |
|
"grad_norm": 0.037920210510492325, |
|
"learning_rate": 6.713813767931031e-05, |
|
"loss": 0.6577, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 1.225134926754048, |
|
"grad_norm": 0.03627100959420204, |
|
"learning_rate": 6.690570482681002e-05, |
|
"loss": 0.6079, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 1.2266769468003085, |
|
"grad_norm": 0.034207917749881744, |
|
"learning_rate": 6.667347258372559e-05, |
|
"loss": 0.6746, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 1.228218966846569, |
|
"grad_norm": 0.037893664091825485, |
|
"learning_rate": 6.644144235779132e-05, |
|
"loss": 0.6847, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 1.2297609868928296, |
|
"grad_norm": 0.039413537830114365, |
|
"learning_rate": 6.620961555551697e-05, |
|
"loss": 0.6923, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 1.2313030069390902, |
|
"grad_norm": 0.03504036366939545, |
|
"learning_rate": 6.597799358217915e-05, |
|
"loss": 0.6434, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 1.2328450269853508, |
|
"grad_norm": 0.036619458347558975, |
|
"learning_rate": 6.574657784181287e-05, |
|
"loss": 0.6321, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.2343870470316114, |
|
"grad_norm": 0.03508542478084564, |
|
"learning_rate": 6.551536973720298e-05, |
|
"loss": 0.6873, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 1.235929067077872, |
|
"grad_norm": 0.038882650434970856, |
|
"learning_rate": 6.52843706698757e-05, |
|
"loss": 0.7239, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 1.2374710871241326, |
|
"grad_norm": 0.03949695825576782, |
|
"learning_rate": 6.505358204009017e-05, |
|
"loss": 0.7489, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 1.2390131071703931, |
|
"grad_norm": 0.03820425644516945, |
|
"learning_rate": 6.482300524682985e-05, |
|
"loss": 0.6727, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 1.2405551272166537, |
|
"grad_norm": 0.033910010010004044, |
|
"learning_rate": 6.459264168779415e-05, |
|
"loss": 0.6468, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 1.2420971472629145, |
|
"grad_norm": 0.03440403565764427, |
|
"learning_rate": 6.436249275938977e-05, |
|
"loss": 0.6257, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 1.243639167309175, |
|
"grad_norm": 0.033627837896347046, |
|
"learning_rate": 6.413255985672262e-05, |
|
"loss": 0.6593, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 1.2451811873554357, |
|
"grad_norm": 0.04215480387210846, |
|
"learning_rate": 6.390284437358888e-05, |
|
"loss": 0.6577, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 1.2467232074016963, |
|
"grad_norm": 0.04215266555547714, |
|
"learning_rate": 6.367334770246693e-05, |
|
"loss": 0.618, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 1.2482652274479569, |
|
"grad_norm": 0.034282535314559937, |
|
"learning_rate": 6.344407123450866e-05, |
|
"loss": 0.5776, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.2498072474942175, |
|
"grad_norm": 0.0348060242831707, |
|
"learning_rate": 6.321501635953124e-05, |
|
"loss": 0.6657, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 1.251349267540478, |
|
"grad_norm": 0.04382369667291641, |
|
"learning_rate": 6.298618446600856e-05, |
|
"loss": 0.6688, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 1.2528912875867386, |
|
"grad_norm": 0.03689052164554596, |
|
"learning_rate": 6.275757694106286e-05, |
|
"loss": 0.7098, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 1.2544333076329992, |
|
"grad_norm": 0.03462991490960121, |
|
"learning_rate": 6.252919517045625e-05, |
|
"loss": 0.6812, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 1.2559753276792598, |
|
"grad_norm": 0.03429633378982544, |
|
"learning_rate": 6.230104053858247e-05, |
|
"loss": 0.6139, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 1.2575173477255204, |
|
"grad_norm": 0.041576385498046875, |
|
"learning_rate": 6.207311442845834e-05, |
|
"loss": 0.6633, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 1.259059367771781, |
|
"grad_norm": 0.03809819743037224, |
|
"learning_rate": 6.184541822171545e-05, |
|
"loss": 0.6521, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 1.2606013878180415, |
|
"grad_norm": 0.03564080595970154, |
|
"learning_rate": 6.161795329859183e-05, |
|
"loss": 0.6735, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 1.2621434078643023, |
|
"grad_norm": 0.038920801132917404, |
|
"learning_rate": 6.13907210379234e-05, |
|
"loss": 0.752, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 1.2636854279105627, |
|
"grad_norm": 0.03850522264838219, |
|
"learning_rate": 6.116372281713581e-05, |
|
"loss": 0.6547, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.2652274479568235, |
|
"grad_norm": 0.03561043366789818, |
|
"learning_rate": 6.093696001223609e-05, |
|
"loss": 0.6592, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 1.266769468003084, |
|
"grad_norm": 0.03549296036362648, |
|
"learning_rate": 6.071043399780412e-05, |
|
"loss": 0.6246, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 1.2683114880493447, |
|
"grad_norm": 0.03841027989983559, |
|
"learning_rate": 6.048414614698448e-05, |
|
"loss": 0.6699, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 1.2698535080956053, |
|
"grad_norm": 0.0381116084754467, |
|
"learning_rate": 6.0258097831478024e-05, |
|
"loss": 0.6542, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 1.2713955281418658, |
|
"grad_norm": 0.03543659672141075, |
|
"learning_rate": 6.00322904215336e-05, |
|
"loss": 0.699, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 1.2729375481881264, |
|
"grad_norm": 0.03378836065530777, |
|
"learning_rate": 5.98067252859398e-05, |
|
"loss": 0.6506, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 1.274479568234387, |
|
"grad_norm": 0.03210572898387909, |
|
"learning_rate": 5.95814037920166e-05, |
|
"loss": 0.5834, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 1.2760215882806476, |
|
"grad_norm": 0.034539416432380676, |
|
"learning_rate": 5.935632730560702e-05, |
|
"loss": 0.6355, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 1.2775636083269082, |
|
"grad_norm": 0.03431408852338791, |
|
"learning_rate": 5.913149719106896e-05, |
|
"loss": 0.6222, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 1.2791056283731688, |
|
"grad_norm": 0.03308793902397156, |
|
"learning_rate": 5.89069148112668e-05, |
|
"loss": 0.5993, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.2806476484194294, |
|
"grad_norm": 0.0353439524769783, |
|
"learning_rate": 5.8682581527563366e-05, |
|
"loss": 0.6453, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 1.2821896684656902, |
|
"grad_norm": 0.03516675531864166, |
|
"learning_rate": 5.845849869981137e-05, |
|
"loss": 0.6377, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 1.2837316885119505, |
|
"grad_norm": 0.03784884139895439, |
|
"learning_rate": 5.823466768634538e-05, |
|
"loss": 0.6936, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 1.2852737085582113, |
|
"grad_norm": 0.03621688485145569, |
|
"learning_rate": 5.801108984397354e-05, |
|
"loss": 0.6495, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 1.286815728604472, |
|
"grad_norm": 0.0380130410194397, |
|
"learning_rate": 5.7787766527969354e-05, |
|
"loss": 0.578, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 1.2883577486507325, |
|
"grad_norm": 0.045173536986112595, |
|
"learning_rate": 5.7564699092063345e-05, |
|
"loss": 0.6438, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 1.289899768696993, |
|
"grad_norm": 0.03988794982433319, |
|
"learning_rate": 5.7341888888435126e-05, |
|
"loss": 0.7357, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 1.2914417887432537, |
|
"grad_norm": 0.039014436304569244, |
|
"learning_rate": 5.7119337267704866e-05, |
|
"loss": 0.7272, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 1.2929838087895142, |
|
"grad_norm": 0.03429098427295685, |
|
"learning_rate": 5.689704557892528e-05, |
|
"loss": 0.5897, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 1.2945258288357748, |
|
"grad_norm": 0.036472566425800323, |
|
"learning_rate": 5.6675015169573655e-05, |
|
"loss": 0.6074, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.2960678488820354, |
|
"grad_norm": 0.033156655728816986, |
|
"learning_rate": 5.6453247385543206e-05, |
|
"loss": 0.555, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 1.297609868928296, |
|
"grad_norm": 0.03926177695393562, |
|
"learning_rate": 5.623174357113528e-05, |
|
"loss": 0.8, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 1.2991518889745568, |
|
"grad_norm": 0.041468918323516846, |
|
"learning_rate": 5.6010505069051145e-05, |
|
"loss": 0.7158, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 1.3006939090208172, |
|
"grad_norm": 0.034991636872291565, |
|
"learning_rate": 5.578953322038372e-05, |
|
"loss": 0.5738, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 1.302235929067078, |
|
"grad_norm": 0.04295094683766365, |
|
"learning_rate": 5.5568829364609664e-05, |
|
"loss": 0.6652, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 1.3037779491133386, |
|
"grad_norm": 0.03196045756340027, |
|
"learning_rate": 5.5348394839580986e-05, |
|
"loss": 0.6114, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 1.3053199691595991, |
|
"grad_norm": 0.03417590633034706, |
|
"learning_rate": 5.51282309815172e-05, |
|
"loss": 0.6307, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 1.3068619892058597, |
|
"grad_norm": 0.033190734684467316, |
|
"learning_rate": 5.4908339124997e-05, |
|
"loss": 0.6327, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 1.3084040092521203, |
|
"grad_norm": 0.03339439630508423, |
|
"learning_rate": 5.468872060295034e-05, |
|
"loss": 0.6274, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 1.3099460292983809, |
|
"grad_norm": 0.038281556218862534, |
|
"learning_rate": 5.446937674665034e-05, |
|
"loss": 0.7243, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.3114880493446415, |
|
"grad_norm": 0.03810643032193184, |
|
"learning_rate": 5.4250308885705056e-05, |
|
"loss": 0.7178, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 1.313030069390902, |
|
"grad_norm": 0.03502066433429718, |
|
"learning_rate": 5.403151834804951e-05, |
|
"loss": 0.6897, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 1.3145720894371626, |
|
"grad_norm": 0.034795403480529785, |
|
"learning_rate": 5.381300645993779e-05, |
|
"loss": 0.6483, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 1.3161141094834232, |
|
"grad_norm": 0.035580191761255264, |
|
"learning_rate": 5.3594774545934825e-05, |
|
"loss": 0.6123, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 1.3176561295296838, |
|
"grad_norm": 0.03703007102012634, |
|
"learning_rate": 5.337682392890833e-05, |
|
"loss": 0.7408, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 1.3191981495759446, |
|
"grad_norm": 0.06035947799682617, |
|
"learning_rate": 5.3159155930021e-05, |
|
"loss": 0.6696, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 1.320740169622205, |
|
"grad_norm": 0.03433072566986084, |
|
"learning_rate": 5.2941771868722265e-05, |
|
"loss": 0.5911, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 1.3222821896684658, |
|
"grad_norm": 0.037705931812524796, |
|
"learning_rate": 5.27246730627404e-05, |
|
"loss": 0.6841, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 1.3238242097147264, |
|
"grad_norm": 0.03678227216005325, |
|
"learning_rate": 5.250786082807462e-05, |
|
"loss": 0.7193, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 1.325366229760987, |
|
"grad_norm": 0.039237961173057556, |
|
"learning_rate": 5.2291336478986964e-05, |
|
"loss": 0.6715, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.3269082498072475, |
|
"grad_norm": 0.036990754306316376, |
|
"learning_rate": 5.207510132799436e-05, |
|
"loss": 0.6092, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 1.3284502698535081, |
|
"grad_norm": 0.03487098589539528, |
|
"learning_rate": 5.1859156685860656e-05, |
|
"loss": 0.6486, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 1.3299922898997687, |
|
"grad_norm": 0.04065219685435295, |
|
"learning_rate": 5.1643503861588805e-05, |
|
"loss": 0.6473, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 1.3315343099460293, |
|
"grad_norm": 0.03509150445461273, |
|
"learning_rate": 5.142814416241282e-05, |
|
"loss": 0.6324, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 1.3330763299922899, |
|
"grad_norm": 0.03709571436047554, |
|
"learning_rate": 5.121307889378975e-05, |
|
"loss": 0.6533, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 1.3346183500385504, |
|
"grad_norm": 0.03225082904100418, |
|
"learning_rate": 5.099830935939203e-05, |
|
"loss": 0.6267, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 1.336160370084811, |
|
"grad_norm": 0.042506713420152664, |
|
"learning_rate": 5.078383686109926e-05, |
|
"loss": 0.7575, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 1.3377023901310716, |
|
"grad_norm": 0.03380179777741432, |
|
"learning_rate": 5.056966269899069e-05, |
|
"loss": 0.6129, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 1.3392444101773324, |
|
"grad_norm": 0.04255492612719536, |
|
"learning_rate": 5.0355788171336914e-05, |
|
"loss": 0.6912, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 1.3407864302235928, |
|
"grad_norm": 0.03951529785990715, |
|
"learning_rate": 5.01422145745924e-05, |
|
"loss": 0.62, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.3423284502698536, |
|
"grad_norm": 0.0411263071000576, |
|
"learning_rate": 4.992894320338727e-05, |
|
"loss": 0.667, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 1.3438704703161142, |
|
"grad_norm": 0.03915927931666374, |
|
"learning_rate": 4.971597535051977e-05, |
|
"loss": 0.6953, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 1.3454124903623748, |
|
"grad_norm": 0.04352157562971115, |
|
"learning_rate": 4.950331230694829e-05, |
|
"loss": 0.7338, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 1.3469545104086353, |
|
"grad_norm": 0.0368880033493042, |
|
"learning_rate": 4.929095536178346e-05, |
|
"loss": 0.6523, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 1.348496530454896, |
|
"grad_norm": 0.03794132545590401, |
|
"learning_rate": 4.907890580228042e-05, |
|
"loss": 0.6702, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 1.3500385505011565, |
|
"grad_norm": 0.03402203321456909, |
|
"learning_rate": 4.886716491383111e-05, |
|
"loss": 0.6168, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 1.351580570547417, |
|
"grad_norm": 0.03810921311378479, |
|
"learning_rate": 4.8655733979956255e-05, |
|
"loss": 0.6744, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 1.3531225905936777, |
|
"grad_norm": 0.034791238605976105, |
|
"learning_rate": 4.844461428229782e-05, |
|
"loss": 0.6058, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 1.3546646106399383, |
|
"grad_norm": 0.03566417098045349, |
|
"learning_rate": 4.8233807100611114e-05, |
|
"loss": 0.5803, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 1.3562066306861988, |
|
"grad_norm": 0.04459409415721893, |
|
"learning_rate": 4.802331371275702e-05, |
|
"loss": 0.6416, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.3577486507324594, |
|
"grad_norm": 0.03633615002036095, |
|
"learning_rate": 4.781313539469423e-05, |
|
"loss": 0.6412, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 1.3592906707787202, |
|
"grad_norm": 0.03890186548233032, |
|
"learning_rate": 4.760327342047167e-05, |
|
"loss": 0.6441, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 1.3608326908249806, |
|
"grad_norm": 0.03216012567281723, |
|
"learning_rate": 4.739372906222066e-05, |
|
"loss": 0.6074, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 1.3623747108712414, |
|
"grad_norm": 0.03994244709610939, |
|
"learning_rate": 4.718450359014712e-05, |
|
"loss": 0.6768, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 1.363916730917502, |
|
"grad_norm": 0.03384639695286751, |
|
"learning_rate": 4.697559827252398e-05, |
|
"loss": 0.5744, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 1.3654587509637626, |
|
"grad_norm": 0.04195882007479668, |
|
"learning_rate": 4.676701437568354e-05, |
|
"loss": 0.6136, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 1.3670007710100232, |
|
"grad_norm": 0.035277076065540314, |
|
"learning_rate": 4.655875316400974e-05, |
|
"loss": 0.591, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 1.3685427910562837, |
|
"grad_norm": 0.03624827414751053, |
|
"learning_rate": 4.6350815899930336e-05, |
|
"loss": 0.6236, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 1.3700848111025443, |
|
"grad_norm": 0.03686223924160004, |
|
"learning_rate": 4.614320384390959e-05, |
|
"loss": 0.6283, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 1.371626831148805, |
|
"grad_norm": 0.04274080693721771, |
|
"learning_rate": 4.593591825444028e-05, |
|
"loss": 0.6367, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.3731688511950655, |
|
"grad_norm": 0.03906931355595589, |
|
"learning_rate": 4.5728960388036214e-05, |
|
"loss": 0.6403, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 1.374710871241326, |
|
"grad_norm": 0.036612290889024734, |
|
"learning_rate": 4.5522331499224835e-05, |
|
"loss": 0.6149, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 1.3762528912875869, |
|
"grad_norm": 0.03672497719526291, |
|
"learning_rate": 4.531603284053919e-05, |
|
"loss": 0.6335, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 1.3777949113338472, |
|
"grad_norm": 0.03735297545790672, |
|
"learning_rate": 4.511006566251059e-05, |
|
"loss": 0.6525, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 1.379336931380108, |
|
"grad_norm": 0.042314548045396805, |
|
"learning_rate": 4.490443121366105e-05, |
|
"loss": 0.6172, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 1.3808789514263684, |
|
"grad_norm": 0.035961080342531204, |
|
"learning_rate": 4.469913074049567e-05, |
|
"loss": 0.5667, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 1.3824209714726292, |
|
"grad_norm": 0.03719279170036316, |
|
"learning_rate": 4.449416548749503e-05, |
|
"loss": 0.6286, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 1.3839629915188898, |
|
"grad_norm": 0.035268232226371765, |
|
"learning_rate": 4.4289536697107636e-05, |
|
"loss": 0.6249, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 1.3855050115651504, |
|
"grad_norm": 0.03824489563703537, |
|
"learning_rate": 4.4085245609742595e-05, |
|
"loss": 0.6263, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 1.387047031611411, |
|
"grad_norm": 0.04036892205476761, |
|
"learning_rate": 4.388129346376178e-05, |
|
"loss": 0.6698, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.3885890516576715, |
|
"grad_norm": 0.03689991310238838, |
|
"learning_rate": 4.367768149547256e-05, |
|
"loss": 0.6192, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 1.3901310717039321, |
|
"grad_norm": 0.03646843880414963, |
|
"learning_rate": 4.34744109391203e-05, |
|
"loss": 0.6822, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 1.3916730917501927, |
|
"grad_norm": 0.036306463181972504, |
|
"learning_rate": 4.3271483026880686e-05, |
|
"loss": 0.7123, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 1.3932151117964533, |
|
"grad_norm": 0.034124668687582016, |
|
"learning_rate": 4.3068898988852404e-05, |
|
"loss": 0.6119, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 1.3947571318427139, |
|
"grad_norm": 0.041610926389694214, |
|
"learning_rate": 4.286666005304971e-05, |
|
"loss": 0.6478, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 1.3962991518889747, |
|
"grad_norm": 0.04072918742895126, |
|
"learning_rate": 4.266476744539496e-05, |
|
"loss": 0.6173, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 1.397841171935235, |
|
"grad_norm": 0.04097921401262283, |
|
"learning_rate": 4.246322238971105e-05, |
|
"loss": 0.6663, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 1.3993831919814959, |
|
"grad_norm": 0.04002819582819939, |
|
"learning_rate": 4.226202610771419e-05, |
|
"loss": 0.6175, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 1.4009252120277564, |
|
"grad_norm": 0.036188431084156036, |
|
"learning_rate": 4.206117981900636e-05, |
|
"loss": 0.6083, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 1.402467232074017, |
|
"grad_norm": 0.044372767210006714, |
|
"learning_rate": 4.186068474106802e-05, |
|
"loss": 0.6594, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.4040092521202776, |
|
"grad_norm": 0.044394172728061676, |
|
"learning_rate": 4.16605420892506e-05, |
|
"loss": 0.6764, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 1.4055512721665382, |
|
"grad_norm": 0.03883661702275276, |
|
"learning_rate": 4.146075307676932e-05, |
|
"loss": 0.6484, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 1.4070932922127988, |
|
"grad_norm": 0.0384899266064167, |
|
"learning_rate": 4.126131891469561e-05, |
|
"loss": 0.6472, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 1.4086353122590594, |
|
"grad_norm": 0.03456205874681473, |
|
"learning_rate": 4.106224081194989e-05, |
|
"loss": 0.5966, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 1.41017733230532, |
|
"grad_norm": 0.038852810859680176, |
|
"learning_rate": 4.086351997529441e-05, |
|
"loss": 0.6582, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 1.4117193523515805, |
|
"grad_norm": 0.03524342551827431, |
|
"learning_rate": 4.0665157609325565e-05, |
|
"loss": 0.6349, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 1.413261372397841, |
|
"grad_norm": 0.039776511490345, |
|
"learning_rate": 4.046715491646683e-05, |
|
"loss": 0.6392, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 1.4148033924441017, |
|
"grad_norm": 0.03844306245446205, |
|
"learning_rate": 4.026951309696152e-05, |
|
"loss": 0.6372, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 1.4163454124903625, |
|
"grad_norm": 0.03955772891640663, |
|
"learning_rate": 4.007223334886531e-05, |
|
"loss": 0.6827, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 1.4178874325366229, |
|
"grad_norm": 0.03288029879331589, |
|
"learning_rate": 3.9875316868039206e-05, |
|
"loss": 0.556, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.4194294525828837, |
|
"grad_norm": 0.03645135462284088, |
|
"learning_rate": 3.9678764848142024e-05, |
|
"loss": 0.6617, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 1.4209714726291442, |
|
"grad_norm": 0.03935571014881134, |
|
"learning_rate": 3.948257848062351e-05, |
|
"loss": 0.6806, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 1.4225134926754048, |
|
"grad_norm": 0.03671710938215256, |
|
"learning_rate": 3.9286758954716696e-05, |
|
"loss": 0.6654, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 1.4240555127216654, |
|
"grad_norm": 0.03780873119831085, |
|
"learning_rate": 3.909130745743108e-05, |
|
"loss": 0.6481, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 1.425597532767926, |
|
"grad_norm": 0.035091038793325424, |
|
"learning_rate": 3.889622517354523e-05, |
|
"loss": 0.6611, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 1.4271395528141866, |
|
"grad_norm": 0.0360507071018219, |
|
"learning_rate": 3.870151328559956e-05, |
|
"loss": 0.6477, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 1.4286815728604472, |
|
"grad_norm": 0.03495753929018974, |
|
"learning_rate": 3.850717297388926e-05, |
|
"loss": 0.5893, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 1.4302235929067078, |
|
"grad_norm": 0.03431479260325432, |
|
"learning_rate": 3.8313205416457164e-05, |
|
"loss": 0.6049, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 1.4317656129529683, |
|
"grad_norm": 0.03628220409154892, |
|
"learning_rate": 3.811961178908657e-05, |
|
"loss": 0.5614, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 1.433307632999229, |
|
"grad_norm": 0.0385250449180603, |
|
"learning_rate": 3.7926393265294017e-05, |
|
"loss": 0.6891, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.4348496530454895, |
|
"grad_norm": 0.03749658167362213, |
|
"learning_rate": 3.773355101632235e-05, |
|
"loss": 0.6655, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 1.4363916730917503, |
|
"grad_norm": 0.041382692754268646, |
|
"learning_rate": 3.75410862111335e-05, |
|
"loss": 0.7552, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 1.4379336931380107, |
|
"grad_norm": 0.040209271013736725, |
|
"learning_rate": 3.734900001640135e-05, |
|
"loss": 0.7837, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 1.4394757131842715, |
|
"grad_norm": 0.034024372696876526, |
|
"learning_rate": 3.715729359650486e-05, |
|
"loss": 0.5879, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 1.441017733230532, |
|
"grad_norm": 0.039757728576660156, |
|
"learning_rate": 3.6965968113520864e-05, |
|
"loss": 0.7303, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 1.4425597532767926, |
|
"grad_norm": 0.036601316183805466, |
|
"learning_rate": 3.677502472721699e-05, |
|
"loss": 0.635, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 1.4441017733230532, |
|
"grad_norm": 0.03675416484475136, |
|
"learning_rate": 3.658446459504471e-05, |
|
"loss": 0.6311, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 1.4456437933693138, |
|
"grad_norm": 0.03397215157747269, |
|
"learning_rate": 3.6394288872132335e-05, |
|
"loss": 0.5638, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 1.4471858134155744, |
|
"grad_norm": 0.03988713398575783, |
|
"learning_rate": 3.620449871127801e-05, |
|
"loss": 0.639, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 1.448727833461835, |
|
"grad_norm": 0.036214519292116165, |
|
"learning_rate": 3.6015095262942574e-05, |
|
"loss": 0.6345, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.4502698535080956, |
|
"grad_norm": 0.03787151351571083, |
|
"learning_rate": 3.58260796752429e-05, |
|
"loss": 0.7008, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 1.4518118735543561, |
|
"grad_norm": 0.03837287053465843, |
|
"learning_rate": 3.5637453093944516e-05, |
|
"loss": 0.6619, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 1.4533538936006167, |
|
"grad_norm": 0.03574013337492943, |
|
"learning_rate": 3.54492166624551e-05, |
|
"loss": 0.5958, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 1.4548959136468773, |
|
"grad_norm": 0.04165033996105194, |
|
"learning_rate": 3.5261371521817244e-05, |
|
"loss": 0.6063, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 1.4564379336931381, |
|
"grad_norm": 0.036275554448366165, |
|
"learning_rate": 3.507391881070161e-05, |
|
"loss": 0.6585, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 1.4579799537393985, |
|
"grad_norm": 0.04543803632259369, |
|
"learning_rate": 3.4886859665400075e-05, |
|
"loss": 0.7924, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 1.4595219737856593, |
|
"grad_norm": 0.04047508165240288, |
|
"learning_rate": 3.470019521981882e-05, |
|
"loss": 0.7316, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 1.4610639938319199, |
|
"grad_norm": 0.042151421308517456, |
|
"learning_rate": 3.45139266054715e-05, |
|
"loss": 0.6866, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 1.4626060138781805, |
|
"grad_norm": 0.03542311489582062, |
|
"learning_rate": 3.432805495147227e-05, |
|
"loss": 0.6577, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 1.464148033924441, |
|
"grad_norm": 0.03901368007063866, |
|
"learning_rate": 3.414258138452898e-05, |
|
"loss": 0.6487, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.4656900539707016, |
|
"grad_norm": 0.03694869950413704, |
|
"learning_rate": 3.395750702893651e-05, |
|
"loss": 0.7043, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 1.4672320740169622, |
|
"grad_norm": 0.038198187947273254, |
|
"learning_rate": 3.377283300656967e-05, |
|
"loss": 0.6513, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 1.4687740940632228, |
|
"grad_norm": 0.03860288858413696, |
|
"learning_rate": 3.358856043687666e-05, |
|
"loss": 0.6888, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 1.4703161141094834, |
|
"grad_norm": 0.04178981855511665, |
|
"learning_rate": 3.340469043687213e-05, |
|
"loss": 0.639, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 1.471858134155744, |
|
"grad_norm": 0.036654602736234665, |
|
"learning_rate": 3.322122412113047e-05, |
|
"loss": 0.6347, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 1.4734001542020048, |
|
"grad_norm": 0.0347287580370903, |
|
"learning_rate": 3.303816260177894e-05, |
|
"loss": 0.4882, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 1.4749421742482651, |
|
"grad_norm": 0.037516217678785324, |
|
"learning_rate": 3.285550698849117e-05, |
|
"loss": 0.6217, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 1.476484194294526, |
|
"grad_norm": 0.03708701953291893, |
|
"learning_rate": 3.267325838848023e-05, |
|
"loss": 0.6542, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 1.4780262143407863, |
|
"grad_norm": 0.03827507793903351, |
|
"learning_rate": 3.2491417906491914e-05, |
|
"loss": 0.5595, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 1.479568234387047, |
|
"grad_norm": 0.03934675455093384, |
|
"learning_rate": 3.230998664479823e-05, |
|
"loss": 0.6574, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.4811102544333077, |
|
"grad_norm": 0.0406733900308609, |
|
"learning_rate": 3.212896570319045e-05, |
|
"loss": 0.645, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 1.4826522744795683, |
|
"grad_norm": 0.040251266211271286, |
|
"learning_rate": 3.194835617897273e-05, |
|
"loss": 0.6699, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 1.4841942945258288, |
|
"grad_norm": 0.041038792580366135, |
|
"learning_rate": 3.176815916695518e-05, |
|
"loss": 0.6631, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 1.4857363145720894, |
|
"grad_norm": 0.03457193076610565, |
|
"learning_rate": 3.158837575944751e-05, |
|
"loss": 0.6236, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 1.48727833461835, |
|
"grad_norm": 0.03977598994970322, |
|
"learning_rate": 3.1409007046252115e-05, |
|
"loss": 0.6281, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 1.4888203546646106, |
|
"grad_norm": 0.03788241744041443, |
|
"learning_rate": 3.123005411465766e-05, |
|
"loss": 0.6207, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 1.4903623747108712, |
|
"grad_norm": 0.038311246782541275, |
|
"learning_rate": 3.105151804943256e-05, |
|
"loss": 0.5844, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 1.4919043947571318, |
|
"grad_norm": 0.03648220747709274, |
|
"learning_rate": 3.087339993281816e-05, |
|
"loss": 0.5752, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 1.4934464148033926, |
|
"grad_norm": 0.03483033925294876, |
|
"learning_rate": 3.06957008445223e-05, |
|
"loss": 0.5375, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 1.494988434849653, |
|
"grad_norm": 0.03250645473599434, |
|
"learning_rate": 3.051842186171284e-05, |
|
"loss": 0.5872, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.4965304548959137, |
|
"grad_norm": 0.04050235077738762, |
|
"learning_rate": 3.0341564059011084e-05, |
|
"loss": 0.6882, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 1.4980724749421743, |
|
"grad_norm": 0.03309236839413643, |
|
"learning_rate": 3.0165128508485164e-05, |
|
"loss": 0.5716, |
|
"step": 972 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 1296, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 324, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.303904573344396e+19, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|