| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.99835255354201, | |
| "eval_steps": 500, | |
| "global_step": 1365, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.002196595277320154, | |
| "grad_norm": 0.09368424117565155, | |
| "learning_rate": 0.0, | |
| "loss": 0.8395, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.004393190554640308, | |
| "grad_norm": 0.07957754284143448, | |
| "learning_rate": 2.18978102189781e-06, | |
| "loss": 1.0597, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.006589785831960461, | |
| "grad_norm": 0.07214013487100601, | |
| "learning_rate": 4.37956204379562e-06, | |
| "loss": 0.8589, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.008786381109280615, | |
| "grad_norm": 0.05852081626653671, | |
| "learning_rate": 6.56934306569343e-06, | |
| "loss": 0.7461, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.010982976386600769, | |
| "grad_norm": 0.0548330582678318, | |
| "learning_rate": 8.75912408759124e-06, | |
| "loss": 0.8293, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.013179571663920923, | |
| "grad_norm": 0.05844108760356903, | |
| "learning_rate": 1.0948905109489049e-05, | |
| "loss": 0.6642, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.015376166941241077, | |
| "grad_norm": 0.05149471014738083, | |
| "learning_rate": 1.313868613138686e-05, | |
| "loss": 0.6719, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.01757276221856123, | |
| "grad_norm": 0.05950519070029259, | |
| "learning_rate": 1.532846715328467e-05, | |
| "loss": 0.7239, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.019769357495881382, | |
| "grad_norm": 0.0998830646276474, | |
| "learning_rate": 1.751824817518248e-05, | |
| "loss": 1.0669, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.021965952773201538, | |
| "grad_norm": 0.06528295576572418, | |
| "learning_rate": 1.9708029197080292e-05, | |
| "loss": 0.741, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.02416254805052169, | |
| "grad_norm": 0.0863526314496994, | |
| "learning_rate": 2.1897810218978098e-05, | |
| "loss": 0.9336, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.026359143327841845, | |
| "grad_norm": 0.051527634263038635, | |
| "learning_rate": 2.408759124087591e-05, | |
| "loss": 0.8121, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.028555738605161998, | |
| "grad_norm": 0.06828700751066208, | |
| "learning_rate": 2.627737226277372e-05, | |
| "loss": 0.7897, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.030752333882482153, | |
| "grad_norm": 0.05828464776277542, | |
| "learning_rate": 2.846715328467153e-05, | |
| "loss": 0.8097, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.032948929159802305, | |
| "grad_norm": 0.0687652975320816, | |
| "learning_rate": 3.065693430656934e-05, | |
| "loss": 0.8632, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.03514552443712246, | |
| "grad_norm": 0.05913132429122925, | |
| "learning_rate": 3.284671532846715e-05, | |
| "loss": 0.6898, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.037342119714442616, | |
| "grad_norm": 0.07138373702764511, | |
| "learning_rate": 3.503649635036496e-05, | |
| "loss": 0.8091, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.039538714991762765, | |
| "grad_norm": 0.07203706353902817, | |
| "learning_rate": 3.722627737226277e-05, | |
| "loss": 0.8459, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.04173531026908292, | |
| "grad_norm": 0.05224800482392311, | |
| "learning_rate": 3.9416058394160584e-05, | |
| "loss": 0.9044, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.043931905546403076, | |
| "grad_norm": 0.06632044166326523, | |
| "learning_rate": 4.160583941605839e-05, | |
| "loss": 0.8141, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.04612850082372323, | |
| "grad_norm": 0.06928423047065735, | |
| "learning_rate": 4.3795620437956196e-05, | |
| "loss": 0.8945, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.04832509610104338, | |
| "grad_norm": 0.058030448853969574, | |
| "learning_rate": 4.5985401459854016e-05, | |
| "loss": 0.7738, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.050521691378363535, | |
| "grad_norm": 0.09114016592502594, | |
| "learning_rate": 4.817518248175182e-05, | |
| "loss": 0.9901, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.05271828665568369, | |
| "grad_norm": 0.06744793802499771, | |
| "learning_rate": 5.036496350364963e-05, | |
| "loss": 0.8745, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.054914881933003847, | |
| "grad_norm": 0.067961685359478, | |
| "learning_rate": 5.255474452554744e-05, | |
| "loss": 0.7961, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.057111477210323995, | |
| "grad_norm": 0.0889795646071434, | |
| "learning_rate": 5.4744525547445253e-05, | |
| "loss": 0.7916, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05930807248764415, | |
| "grad_norm": 0.10735788196325302, | |
| "learning_rate": 5.693430656934306e-05, | |
| "loss": 0.8373, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.061504667764964306, | |
| "grad_norm": 0.06766892969608307, | |
| "learning_rate": 5.912408759124087e-05, | |
| "loss": 0.79, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.06370126304228446, | |
| "grad_norm": 0.054694995284080505, | |
| "learning_rate": 6.131386861313868e-05, | |
| "loss": 0.7589, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.06589785831960461, | |
| "grad_norm": 0.06533095985651016, | |
| "learning_rate": 6.350364963503648e-05, | |
| "loss": 0.7947, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.06809445359692477, | |
| "grad_norm": 0.07474099099636078, | |
| "learning_rate": 6.56934306569343e-05, | |
| "loss": 0.7711, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.07029104887424492, | |
| "grad_norm": 0.10328605026006699, | |
| "learning_rate": 6.788321167883211e-05, | |
| "loss": 0.9575, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.07248764415156507, | |
| "grad_norm": 0.0813312754034996, | |
| "learning_rate": 7.007299270072992e-05, | |
| "loss": 1.1005, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.07468423942888523, | |
| "grad_norm": 0.07534842193126678, | |
| "learning_rate": 7.226277372262774e-05, | |
| "loss": 0.8917, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.07688083470620538, | |
| "grad_norm": 0.06850400567054749, | |
| "learning_rate": 7.445255474452554e-05, | |
| "loss": 0.8311, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.07907742998352553, | |
| "grad_norm": 0.08918626606464386, | |
| "learning_rate": 7.664233576642336e-05, | |
| "loss": 0.9399, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.08127402526084569, | |
| "grad_norm": 0.07072232663631439, | |
| "learning_rate": 7.883211678832117e-05, | |
| "loss": 0.8374, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.08347062053816584, | |
| "grad_norm": 0.08290290087461472, | |
| "learning_rate": 8.102189781021897e-05, | |
| "loss": 0.9685, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.085667215815486, | |
| "grad_norm": 0.07387691736221313, | |
| "learning_rate": 8.321167883211678e-05, | |
| "loss": 0.9722, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.08786381109280615, | |
| "grad_norm": 0.09207072108983994, | |
| "learning_rate": 8.540145985401459e-05, | |
| "loss": 0.7773, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.0900604063701263, | |
| "grad_norm": 0.06684845685958862, | |
| "learning_rate": 8.759124087591239e-05, | |
| "loss": 0.8774, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.09225700164744646, | |
| "grad_norm": 0.09227459877729416, | |
| "learning_rate": 8.978102189781021e-05, | |
| "loss": 0.9656, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.09445359692476661, | |
| "grad_norm": 0.08900696784257889, | |
| "learning_rate": 9.197080291970803e-05, | |
| "loss": 0.8676, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.09665019220208676, | |
| "grad_norm": 0.07540471106767654, | |
| "learning_rate": 9.416058394160584e-05, | |
| "loss": 0.9037, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.09884678747940692, | |
| "grad_norm": 0.06883317977190018, | |
| "learning_rate": 9.635036496350364e-05, | |
| "loss": 0.9332, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.10104338275672707, | |
| "grad_norm": 0.06095288321375847, | |
| "learning_rate": 9.854014598540145e-05, | |
| "loss": 0.7869, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.10323997803404723, | |
| "grad_norm": 0.05872150510549545, | |
| "learning_rate": 0.00010072992700729926, | |
| "loss": 0.8301, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.10543657331136738, | |
| "grad_norm": 0.04909675195813179, | |
| "learning_rate": 0.00010291970802919708, | |
| "loss": 0.8333, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.10763316858868753, | |
| "grad_norm": 0.07996238768100739, | |
| "learning_rate": 0.00010510948905109488, | |
| "loss": 0.9812, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.10982976386600769, | |
| "grad_norm": 0.07555586099624634, | |
| "learning_rate": 0.00010729927007299269, | |
| "loss": 0.9272, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.11202635914332784, | |
| "grad_norm": 0.09978742152452469, | |
| "learning_rate": 0.00010948905109489051, | |
| "loss": 0.9312, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.11422295442064799, | |
| "grad_norm": 0.06448879837989807, | |
| "learning_rate": 0.00011167883211678831, | |
| "loss": 0.816, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.11641954969796815, | |
| "grad_norm": 0.09131381660699844, | |
| "learning_rate": 0.00011386861313868612, | |
| "loss": 0.8732, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.1186161449752883, | |
| "grad_norm": 0.07666397094726562, | |
| "learning_rate": 0.00011605839416058394, | |
| "loss": 0.6953, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.12081274025260846, | |
| "grad_norm": 0.056689534336328506, | |
| "learning_rate": 0.00011824817518248174, | |
| "loss": 0.7901, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.12300933552992861, | |
| "grad_norm": 0.10934755951166153, | |
| "learning_rate": 0.00012043795620437955, | |
| "loss": 1.0675, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.12520593080724876, | |
| "grad_norm": 0.061348967254161835, | |
| "learning_rate": 0.00012262773722627736, | |
| "loss": 0.7428, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.12740252608456892, | |
| "grad_norm": 0.07868394255638123, | |
| "learning_rate": 0.00012481751824817516, | |
| "loss": 0.959, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.12959912136188906, | |
| "grad_norm": 0.05839238315820694, | |
| "learning_rate": 0.00012700729927007297, | |
| "loss": 0.715, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.13179571663920922, | |
| "grad_norm": 0.05481383949518204, | |
| "learning_rate": 0.00012919708029197077, | |
| "loss": 0.7748, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.13399231191652938, | |
| "grad_norm": 0.06998932361602783, | |
| "learning_rate": 0.0001313868613138686, | |
| "loss": 0.868, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.13618890719384955, | |
| "grad_norm": 0.06081048399209976, | |
| "learning_rate": 0.00013357664233576641, | |
| "loss": 0.734, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.13838550247116968, | |
| "grad_norm": 0.07183390855789185, | |
| "learning_rate": 0.00013576642335766422, | |
| "loss": 0.7546, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.14058209774848984, | |
| "grad_norm": 0.07835814356803894, | |
| "learning_rate": 0.00013795620437956203, | |
| "loss": 0.8754, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.14277869302581, | |
| "grad_norm": 0.04541606828570366, | |
| "learning_rate": 0.00014014598540145983, | |
| "loss": 0.7274, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.14497528830313014, | |
| "grad_norm": 0.050496987998485565, | |
| "learning_rate": 0.00014233576642335764, | |
| "loss": 0.6557, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.1471718835804503, | |
| "grad_norm": 0.06741216033697128, | |
| "learning_rate": 0.00014452554744525547, | |
| "loss": 0.7853, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.14936847885777046, | |
| "grad_norm": 0.05799397826194763, | |
| "learning_rate": 0.00014671532846715328, | |
| "loss": 0.6912, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.1515650741350906, | |
| "grad_norm": 0.08100557327270508, | |
| "learning_rate": 0.00014890510948905108, | |
| "loss": 1.0204, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.15376166941241076, | |
| "grad_norm": 0.08090441673994064, | |
| "learning_rate": 0.00015109489051094892, | |
| "loss": 0.7844, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.15595826468973092, | |
| "grad_norm": 0.062162335962057114, | |
| "learning_rate": 0.00015328467153284672, | |
| "loss": 0.8829, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.15815485996705106, | |
| "grad_norm": 0.06271839141845703, | |
| "learning_rate": 0.00015547445255474453, | |
| "loss": 0.7453, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.16035145524437122, | |
| "grad_norm": 0.09264097362756729, | |
| "learning_rate": 0.00015766423357664234, | |
| "loss": 0.9994, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.16254805052169138, | |
| "grad_norm": 0.062283921986818314, | |
| "learning_rate": 0.00015985401459854014, | |
| "loss": 0.8463, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.16474464579901152, | |
| "grad_norm": 0.06915900856256485, | |
| "learning_rate": 0.00016204379562043795, | |
| "loss": 0.7171, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.16694124107633168, | |
| "grad_norm": 0.06716243922710419, | |
| "learning_rate": 0.00016423357664233575, | |
| "loss": 0.8937, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.16913783635365184, | |
| "grad_norm": 0.06265877187252045, | |
| "learning_rate": 0.00016642335766423356, | |
| "loss": 0.8112, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.171334431630972, | |
| "grad_norm": 0.06956397742033005, | |
| "learning_rate": 0.00016861313868613137, | |
| "loss": 0.6043, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.17353102690829214, | |
| "grad_norm": 0.0620533749461174, | |
| "learning_rate": 0.00017080291970802917, | |
| "loss": 0.8352, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.1757276221856123, | |
| "grad_norm": 0.07454898208379745, | |
| "learning_rate": 0.00017299270072992698, | |
| "loss": 0.8129, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.17792421746293247, | |
| "grad_norm": 0.05988597124814987, | |
| "learning_rate": 0.00017518248175182478, | |
| "loss": 0.7515, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.1801208127402526, | |
| "grad_norm": 0.06786584854125977, | |
| "learning_rate": 0.00017737226277372262, | |
| "loss": 0.7505, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.18231740801757276, | |
| "grad_norm": 0.049385055899620056, | |
| "learning_rate": 0.00017956204379562042, | |
| "loss": 0.6942, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.18451400329489293, | |
| "grad_norm": 0.09457117319107056, | |
| "learning_rate": 0.00018175182481751826, | |
| "loss": 0.8961, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.18671059857221306, | |
| "grad_norm": 0.0817558765411377, | |
| "learning_rate": 0.00018394160583941606, | |
| "loss": 0.9277, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.18890719384953322, | |
| "grad_norm": 0.07502376288175583, | |
| "learning_rate": 0.00018613138686131387, | |
| "loss": 1.0283, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.19110378912685339, | |
| "grad_norm": 0.05120819807052612, | |
| "learning_rate": 0.00018832116788321167, | |
| "loss": 0.8361, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.19330038440417352, | |
| "grad_norm": 0.07280557602643967, | |
| "learning_rate": 0.00019051094890510948, | |
| "loss": 0.9357, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.19549697968149368, | |
| "grad_norm": 0.07645975053310394, | |
| "learning_rate": 0.0001927007299270073, | |
| "loss": 0.7198, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.19769357495881384, | |
| "grad_norm": 0.08283960074186325, | |
| "learning_rate": 0.0001948905109489051, | |
| "loss": 1.0339, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.19989017023613398, | |
| "grad_norm": 0.04628637805581093, | |
| "learning_rate": 0.0001970802919708029, | |
| "loss": 0.7246, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.20208676551345414, | |
| "grad_norm": 0.047697197645902634, | |
| "learning_rate": 0.0001992700729927007, | |
| "loss": 0.7441, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.2042833607907743, | |
| "grad_norm": 0.07364491373300552, | |
| "learning_rate": 0.0002014598540145985, | |
| "loss": 0.7784, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.20647995606809447, | |
| "grad_norm": 0.0743069127202034, | |
| "learning_rate": 0.00020364963503649632, | |
| "loss": 0.9773, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.2086765513454146, | |
| "grad_norm": 0.059089288115501404, | |
| "learning_rate": 0.00020583941605839415, | |
| "loss": 0.7419, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.21087314662273476, | |
| "grad_norm": 0.07568493485450745, | |
| "learning_rate": 0.00020802919708029196, | |
| "loss": 0.8705, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.21306974190005493, | |
| "grad_norm": 0.07957746833562851, | |
| "learning_rate": 0.00021021897810218976, | |
| "loss": 0.9486, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.21526633717737506, | |
| "grad_norm": 0.07052905112504959, | |
| "learning_rate": 0.00021240875912408757, | |
| "loss": 0.832, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.21746293245469522, | |
| "grad_norm": 0.06128270551562309, | |
| "learning_rate": 0.00021459854014598537, | |
| "loss": 0.7649, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.21965952773201539, | |
| "grad_norm": 0.10580066591501236, | |
| "learning_rate": 0.0002167883211678832, | |
| "loss": 0.9327, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.22185612300933552, | |
| "grad_norm": 0.08737997710704803, | |
| "learning_rate": 0.00021897810218978101, | |
| "loss": 0.9517, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.22405271828665568, | |
| "grad_norm": 0.07992607355117798, | |
| "learning_rate": 0.00022116788321167882, | |
| "loss": 0.8463, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.22624931356397585, | |
| "grad_norm": 0.05659569054841995, | |
| "learning_rate": 0.00022335766423357663, | |
| "loss": 0.8145, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.22844590884129598, | |
| "grad_norm": 0.06634163856506348, | |
| "learning_rate": 0.00022554744525547443, | |
| "loss": 0.7103, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.23064250411861614, | |
| "grad_norm": 0.05726408213376999, | |
| "learning_rate": 0.00022773722627737224, | |
| "loss": 0.8268, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.2328390993959363, | |
| "grad_norm": 0.061387598514556885, | |
| "learning_rate": 0.00022992700729927004, | |
| "loss": 0.8989, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.23503569467325644, | |
| "grad_norm": 0.10654595494270325, | |
| "learning_rate": 0.00023211678832116788, | |
| "loss": 0.8958, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.2372322899505766, | |
| "grad_norm": 0.05953432247042656, | |
| "learning_rate": 0.00023430656934306568, | |
| "loss": 0.6935, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.23942888522789676, | |
| "grad_norm": 0.07555431127548218, | |
| "learning_rate": 0.0002364963503649635, | |
| "loss": 0.7281, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.24162548050521693, | |
| "grad_norm": 0.06416428089141846, | |
| "learning_rate": 0.0002386861313868613, | |
| "loss": 0.7588, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.24382207578253706, | |
| "grad_norm": 0.08092894405126572, | |
| "learning_rate": 0.0002408759124087591, | |
| "loss": 0.8226, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.24601867105985722, | |
| "grad_norm": 0.08186564594507217, | |
| "learning_rate": 0.0002430656934306569, | |
| "loss": 0.8446, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.2482152663371774, | |
| "grad_norm": 0.07057618349790573, | |
| "learning_rate": 0.0002452554744525547, | |
| "loss": 0.7437, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.2504118616144975, | |
| "grad_norm": 0.0712006688117981, | |
| "learning_rate": 0.0002474452554744525, | |
| "loss": 0.72, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.25260845689181766, | |
| "grad_norm": 0.0903545469045639, | |
| "learning_rate": 0.0002496350364963503, | |
| "loss": 0.8596, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.25480505216913785, | |
| "grad_norm": 0.06417638808488846, | |
| "learning_rate": 0.00025182481751824813, | |
| "loss": 0.9108, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.257001647446458, | |
| "grad_norm": 0.07485811412334442, | |
| "learning_rate": 0.00025401459854014594, | |
| "loss": 0.8721, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.2591982427237781, | |
| "grad_norm": 0.051193512976169586, | |
| "learning_rate": 0.00025620437956204374, | |
| "loss": 0.7606, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.2613948380010983, | |
| "grad_norm": 0.06474370509386063, | |
| "learning_rate": 0.00025839416058394155, | |
| "loss": 0.7512, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.26359143327841844, | |
| "grad_norm": 0.05385642126202583, | |
| "learning_rate": 0.0002605839416058394, | |
| "loss": 0.6462, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.26578802855573863, | |
| "grad_norm": 0.05074644461274147, | |
| "learning_rate": 0.0002627737226277372, | |
| "loss": 0.696, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.26798462383305877, | |
| "grad_norm": 0.0700865164399147, | |
| "learning_rate": 0.000264963503649635, | |
| "loss": 0.741, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.2701812191103789, | |
| "grad_norm": 0.08031459152698517, | |
| "learning_rate": 0.00026715328467153283, | |
| "loss": 0.7789, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.2723778143876991, | |
| "grad_norm": 0.06968266516923904, | |
| "learning_rate": 0.00026934306569343063, | |
| "loss": 0.6849, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.2745744096650192, | |
| "grad_norm": 0.07033291459083557, | |
| "learning_rate": 0.00027153284671532844, | |
| "loss": 0.8538, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.27677100494233936, | |
| "grad_norm": 0.0653030201792717, | |
| "learning_rate": 0.00027372262773722625, | |
| "loss": 0.605, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.27896760021965955, | |
| "grad_norm": 0.07047812640666962, | |
| "learning_rate": 0.00027591240875912405, | |
| "loss": 0.7744, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.2811641954969797, | |
| "grad_norm": 0.07086916267871857, | |
| "learning_rate": 0.00027810218978102186, | |
| "loss": 0.9742, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.2833607907742998, | |
| "grad_norm": 0.05834396556019783, | |
| "learning_rate": 0.00028029197080291966, | |
| "loss": 0.8322, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.28555738605162, | |
| "grad_norm": 0.07212558388710022, | |
| "learning_rate": 0.00028248175182481747, | |
| "loss": 0.719, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.28775398132894014, | |
| "grad_norm": 0.057706017047166824, | |
| "learning_rate": 0.0002846715328467153, | |
| "loss": 0.7827, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.2899505766062603, | |
| "grad_norm": 0.06932689994573593, | |
| "learning_rate": 0.00028686131386861314, | |
| "loss": 0.6932, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.29214717188358047, | |
| "grad_norm": 0.08389901369810104, | |
| "learning_rate": 0.00028905109489051094, | |
| "loss": 0.7219, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.2943437671609006, | |
| "grad_norm": 0.08421478420495987, | |
| "learning_rate": 0.00029124087591240875, | |
| "loss": 0.7718, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.29654036243822074, | |
| "grad_norm": 0.0458688959479332, | |
| "learning_rate": 0.00029343065693430656, | |
| "loss": 0.764, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.29873695771554093, | |
| "grad_norm": 0.09606773406267166, | |
| "learning_rate": 0.00029562043795620436, | |
| "loss": 0.8812, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.30093355299286106, | |
| "grad_norm": 0.06875104457139969, | |
| "learning_rate": 0.00029781021897810217, | |
| "loss": 0.6979, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.3031301482701812, | |
| "grad_norm": 0.05998622253537178, | |
| "learning_rate": 0.0003, | |
| "loss": 0.7853, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.3053267435475014, | |
| "grad_norm": 0.058912474662065506, | |
| "learning_rate": 0.00029975570032573286, | |
| "loss": 0.6941, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.3075233388248215, | |
| "grad_norm": 0.05363420024514198, | |
| "learning_rate": 0.00029951140065146575, | |
| "loss": 0.7914, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.30971993410214166, | |
| "grad_norm": 0.21214614808559418, | |
| "learning_rate": 0.0002992671009771987, | |
| "loss": 0.7579, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.31191652937946185, | |
| "grad_norm": 0.08337908983230591, | |
| "learning_rate": 0.0002990228013029316, | |
| "loss": 0.851, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.314113124656782, | |
| "grad_norm": 0.07102375477552414, | |
| "learning_rate": 0.0002987785016286645, | |
| "loss": 0.8352, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.3163097199341021, | |
| "grad_norm": 0.09091459214687347, | |
| "learning_rate": 0.00029853420195439736, | |
| "loss": 0.9523, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.3185063152114223, | |
| "grad_norm": 0.061527032405138016, | |
| "learning_rate": 0.00029828990228013025, | |
| "loss": 0.7559, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.32070291048874244, | |
| "grad_norm": 0.061387546360492706, | |
| "learning_rate": 0.00029804560260586314, | |
| "loss": 0.8038, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.3228995057660626, | |
| "grad_norm": 0.05838935077190399, | |
| "learning_rate": 0.0002978013029315961, | |
| "loss": 0.6139, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.32509610104338277, | |
| "grad_norm": 0.049171656370162964, | |
| "learning_rate": 0.000297557003257329, | |
| "loss": 0.6492, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.3272926963207029, | |
| "grad_norm": 0.07386723905801773, | |
| "learning_rate": 0.00029731270358306186, | |
| "loss": 0.7368, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.32948929159802304, | |
| "grad_norm": 0.06298866122961044, | |
| "learning_rate": 0.00029706840390879475, | |
| "loss": 0.9454, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.3316858868753432, | |
| "grad_norm": 0.05815809220075607, | |
| "learning_rate": 0.00029682410423452764, | |
| "loss": 0.8447, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.33388248215266336, | |
| "grad_norm": 0.07850918173789978, | |
| "learning_rate": 0.0002965798045602606, | |
| "loss": 0.9031, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.33607907742998355, | |
| "grad_norm": 0.06274339556694031, | |
| "learning_rate": 0.0002963355048859935, | |
| "loss": 0.6847, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.3382756727073037, | |
| "grad_norm": 0.07347492128610611, | |
| "learning_rate": 0.00029609120521172636, | |
| "loss": 0.7598, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.3404722679846238, | |
| "grad_norm": 0.06832234561443329, | |
| "learning_rate": 0.00029584690553745925, | |
| "loss": 0.6799, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.342668863261944, | |
| "grad_norm": 0.053192079067230225, | |
| "learning_rate": 0.00029560260586319214, | |
| "loss": 0.6877, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.34486545853926415, | |
| "grad_norm": 0.11681066453456879, | |
| "learning_rate": 0.00029535830618892503, | |
| "loss": 0.7697, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.3470620538165843, | |
| "grad_norm": 0.08324015140533447, | |
| "learning_rate": 0.000295114006514658, | |
| "loss": 0.751, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.34925864909390447, | |
| "grad_norm": 0.062357187271118164, | |
| "learning_rate": 0.00029486970684039086, | |
| "loss": 0.8613, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.3514552443712246, | |
| "grad_norm": 0.07192494720220566, | |
| "learning_rate": 0.00029462540716612375, | |
| "loss": 0.7714, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.35365183964854474, | |
| "grad_norm": 0.08102371543645859, | |
| "learning_rate": 0.00029438110749185664, | |
| "loss": 0.7397, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.35584843492586493, | |
| "grad_norm": 0.05772097036242485, | |
| "learning_rate": 0.00029413680781758953, | |
| "loss": 0.6198, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.35804503020318507, | |
| "grad_norm": 0.05155602842569351, | |
| "learning_rate": 0.0002938925081433224, | |
| "loss": 0.5073, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.3602416254805052, | |
| "grad_norm": 0.08115146309137344, | |
| "learning_rate": 0.00029364820846905536, | |
| "loss": 0.6907, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.3624382207578254, | |
| "grad_norm": 0.07688108831644058, | |
| "learning_rate": 0.00029340390879478825, | |
| "loss": 0.8097, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.3646348160351455, | |
| "grad_norm": 0.09109821170568466, | |
| "learning_rate": 0.00029315960912052114, | |
| "loss": 0.7904, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.36683141131246566, | |
| "grad_norm": 0.06498458981513977, | |
| "learning_rate": 0.00029291530944625403, | |
| "loss": 0.6746, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.36902800658978585, | |
| "grad_norm": 0.08671940118074417, | |
| "learning_rate": 0.0002926710097719869, | |
| "loss": 0.8064, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.371224601867106, | |
| "grad_norm": 0.06002287194132805, | |
| "learning_rate": 0.00029242671009771986, | |
| "loss": 0.7684, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.3734211971444261, | |
| "grad_norm": 0.07022416591644287, | |
| "learning_rate": 0.00029218241042345275, | |
| "loss": 0.834, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.3756177924217463, | |
| "grad_norm": 0.07985968142747879, | |
| "learning_rate": 0.00029193811074918564, | |
| "loss": 0.7098, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.37781438769906645, | |
| "grad_norm": 0.06475462019443512, | |
| "learning_rate": 0.00029169381107491853, | |
| "loss": 0.6133, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.3800109829763866, | |
| "grad_norm": 0.060454897582530975, | |
| "learning_rate": 0.0002914495114006514, | |
| "loss": 0.7347, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.38220757825370677, | |
| "grad_norm": 0.07031078636646271, | |
| "learning_rate": 0.0002912052117263843, | |
| "loss": 0.95, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.3844041735310269, | |
| "grad_norm": 0.05603813752532005, | |
| "learning_rate": 0.00029096091205211725, | |
| "loss": 0.5946, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.38660076880834704, | |
| "grad_norm": 0.046303559094667435, | |
| "learning_rate": 0.00029071661237785014, | |
| "loss": 0.7976, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.38879736408566723, | |
| "grad_norm": 0.06395114958286285, | |
| "learning_rate": 0.00029047231270358303, | |
| "loss": 0.4958, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.39099395936298736, | |
| "grad_norm": 0.08462050557136536, | |
| "learning_rate": 0.0002902280130293159, | |
| "loss": 0.7312, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.3931905546403075, | |
| "grad_norm": 0.0730491355061531, | |
| "learning_rate": 0.0002899837133550488, | |
| "loss": 0.7501, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.3953871499176277, | |
| "grad_norm": 0.0624474361538887, | |
| "learning_rate": 0.00028973941368078175, | |
| "loss": 0.6107, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.3975837451949478, | |
| "grad_norm": 0.08671862632036209, | |
| "learning_rate": 0.00028949511400651464, | |
| "loss": 0.5783, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.39978034047226796, | |
| "grad_norm": 0.0546480156481266, | |
| "learning_rate": 0.00028925081433224753, | |
| "loss": 0.5778, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.40197693574958815, | |
| "grad_norm": 0.0740947350859642, | |
| "learning_rate": 0.0002890065146579804, | |
| "loss": 0.7729, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.4041735310269083, | |
| "grad_norm": 0.06049893796443939, | |
| "learning_rate": 0.0002887622149837133, | |
| "loss": 0.9998, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.4063701263042284, | |
| "grad_norm": 0.056841589510440826, | |
| "learning_rate": 0.0002885179153094462, | |
| "loss": 0.7461, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.4085667215815486, | |
| "grad_norm": 0.06030123680830002, | |
| "learning_rate": 0.00028827361563517914, | |
| "loss": 0.7373, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.41076331685886874, | |
| "grad_norm": 0.08607760071754456, | |
| "learning_rate": 0.00028802931596091203, | |
| "loss": 0.5802, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.41295991213618893, | |
| "grad_norm": 0.0810663178563118, | |
| "learning_rate": 0.0002877850162866449, | |
| "loss": 0.8243, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.41515650741350907, | |
| "grad_norm": 0.09073084592819214, | |
| "learning_rate": 0.0002875407166123778, | |
| "loss": 0.8053, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.4173531026908292, | |
| "grad_norm": 0.058536745607852936, | |
| "learning_rate": 0.0002872964169381107, | |
| "loss": 0.6984, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.4195496979681494, | |
| "grad_norm": 0.09180819988250732, | |
| "learning_rate": 0.00028705211726384364, | |
| "loss": 0.6705, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.42174629324546953, | |
| "grad_norm": 0.059555936604738235, | |
| "learning_rate": 0.00028680781758957653, | |
| "loss": 0.803, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.42394288852278966, | |
| "grad_norm": 0.0732739195227623, | |
| "learning_rate": 0.0002865635179153094, | |
| "loss": 0.7823, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.42613948380010985, | |
| "grad_norm": 0.059682246297597885, | |
| "learning_rate": 0.0002863192182410423, | |
| "loss": 0.7723, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.42833607907743, | |
| "grad_norm": 0.07256857305765152, | |
| "learning_rate": 0.0002860749185667752, | |
| "loss": 0.8889, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.4305326743547501, | |
| "grad_norm": 0.06505439430475235, | |
| "learning_rate": 0.0002858306188925081, | |
| "loss": 1.0675, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.4327292696320703, | |
| "grad_norm": 0.05685146152973175, | |
| "learning_rate": 0.00028558631921824103, | |
| "loss": 0.6376, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.43492586490939045, | |
| "grad_norm": 0.08345243334770203, | |
| "learning_rate": 0.0002853420195439739, | |
| "loss": 0.6554, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.4371224601867106, | |
| "grad_norm": 0.0606047622859478, | |
| "learning_rate": 0.0002850977198697068, | |
| "loss": 0.6857, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.43931905546403077, | |
| "grad_norm": 0.05236055701971054, | |
| "learning_rate": 0.0002848534201954397, | |
| "loss": 0.7361, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.4415156507413509, | |
| "grad_norm": 0.06410698592662811, | |
| "learning_rate": 0.0002846091205211726, | |
| "loss": 0.6465, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.44371224601867104, | |
| "grad_norm": 0.04402509704232216, | |
| "learning_rate": 0.0002843648208469055, | |
| "loss": 0.6407, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.44590884129599123, | |
| "grad_norm": 0.0492163822054863, | |
| "learning_rate": 0.0002841205211726384, | |
| "loss": 0.697, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.44810543657331137, | |
| "grad_norm": 0.09770838916301727, | |
| "learning_rate": 0.0002838762214983713, | |
| "loss": 0.7488, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.4503020318506315, | |
| "grad_norm": 0.08673590421676636, | |
| "learning_rate": 0.0002836319218241042, | |
| "loss": 0.8844, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.4524986271279517, | |
| "grad_norm": 0.0633588433265686, | |
| "learning_rate": 0.0002833876221498371, | |
| "loss": 0.8417, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.4546952224052718, | |
| "grad_norm": 0.04995807260274887, | |
| "learning_rate": 0.00028314332247557, | |
| "loss": 0.7281, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.45689181768259196, | |
| "grad_norm": 0.05731228366494179, | |
| "learning_rate": 0.0002828990228013029, | |
| "loss": 0.6367, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.45908841295991215, | |
| "grad_norm": 0.11119459569454193, | |
| "learning_rate": 0.0002826547231270358, | |
| "loss": 0.7086, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.4612850082372323, | |
| "grad_norm": 0.06479672342538834, | |
| "learning_rate": 0.0002824104234527687, | |
| "loss": 0.8242, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.4634816035145524, | |
| "grad_norm": 0.05451243370771408, | |
| "learning_rate": 0.0002821661237785016, | |
| "loss": 0.7898, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.4656781987918726, | |
| "grad_norm": 0.05641823261976242, | |
| "learning_rate": 0.0002819218241042345, | |
| "loss": 0.7529, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.46787479406919275, | |
| "grad_norm": 0.05920799449086189, | |
| "learning_rate": 0.00028167752442996737, | |
| "loss": 0.5424, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.4700713893465129, | |
| "grad_norm": 0.03466494008898735, | |
| "learning_rate": 0.0002814332247557003, | |
| "loss": 0.6346, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.47226798462383307, | |
| "grad_norm": 0.04757579788565636, | |
| "learning_rate": 0.0002811889250814332, | |
| "loss": 0.7383, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.4744645799011532, | |
| "grad_norm": 0.06769683212041855, | |
| "learning_rate": 0.0002809446254071661, | |
| "loss": 0.7402, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.47666117517847334, | |
| "grad_norm": 0.047434259206056595, | |
| "learning_rate": 0.000280700325732899, | |
| "loss": 0.5883, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.47885777045579353, | |
| "grad_norm": 0.05035366117954254, | |
| "learning_rate": 0.00028045602605863187, | |
| "loss": 0.777, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.48105436573311366, | |
| "grad_norm": 0.06778255850076675, | |
| "learning_rate": 0.0002802117263843648, | |
| "loss": 0.6835, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.48325096101043385, | |
| "grad_norm": 0.04769022762775421, | |
| "learning_rate": 0.0002799674267100977, | |
| "loss": 0.7615, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.485447556287754, | |
| "grad_norm": 0.06975305825471878, | |
| "learning_rate": 0.0002797231270358306, | |
| "loss": 0.6741, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.4876441515650741, | |
| "grad_norm": 0.0725470632314682, | |
| "learning_rate": 0.0002794788273615635, | |
| "loss": 0.626, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.4898407468423943, | |
| "grad_norm": 0.055425189435482025, | |
| "learning_rate": 0.00027923452768729637, | |
| "loss": 0.7386, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.49203734211971445, | |
| "grad_norm": 0.046605244278907776, | |
| "learning_rate": 0.00027899022801302926, | |
| "loss": 0.7661, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.4942339373970346, | |
| "grad_norm": 0.05119828134775162, | |
| "learning_rate": 0.0002787459283387622, | |
| "loss": 0.5156, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.4964305326743548, | |
| "grad_norm": 0.051588162779808044, | |
| "learning_rate": 0.0002785016286644951, | |
| "loss": 0.9542, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.4986271279516749, | |
| "grad_norm": 0.1783459633588791, | |
| "learning_rate": 0.000278257328990228, | |
| "loss": 0.7026, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.500823723228995, | |
| "grad_norm": 0.06174377724528313, | |
| "learning_rate": 0.00027801302931596087, | |
| "loss": 0.6692, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.5030203185063152, | |
| "grad_norm": 0.051825955510139465, | |
| "learning_rate": 0.00027776872964169376, | |
| "loss": 0.7019, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.5052169137836353, | |
| "grad_norm": 0.05067208409309387, | |
| "learning_rate": 0.0002775244299674267, | |
| "loss": 0.7867, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.5074135090609555, | |
| "grad_norm": 0.06491252779960632, | |
| "learning_rate": 0.0002772801302931596, | |
| "loss": 0.7593, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.5096101043382757, | |
| "grad_norm": 0.052263930439949036, | |
| "learning_rate": 0.0002770358306188925, | |
| "loss": 0.6484, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.5118066996155958, | |
| "grad_norm": 0.059053026139736176, | |
| "learning_rate": 0.00027679153094462537, | |
| "loss": 0.7726, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.514003294892916, | |
| "grad_norm": 0.05262107402086258, | |
| "learning_rate": 0.00027654723127035826, | |
| "loss": 0.7085, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.5161998901702362, | |
| "grad_norm": 0.06159406527876854, | |
| "learning_rate": 0.00027630293159609115, | |
| "loss": 0.7029, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.5183964854475562, | |
| "grad_norm": 0.04124659299850464, | |
| "learning_rate": 0.0002760586319218241, | |
| "loss": 0.6227, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.5205930807248764, | |
| "grad_norm": 0.07333070039749146, | |
| "learning_rate": 0.000275814332247557, | |
| "loss": 0.9546, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.5227896760021966, | |
| "grad_norm": 0.0633094385266304, | |
| "learning_rate": 0.00027557003257328987, | |
| "loss": 0.7767, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.5249862712795168, | |
| "grad_norm": 0.05801127851009369, | |
| "learning_rate": 0.00027532573289902276, | |
| "loss": 0.6928, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.5271828665568369, | |
| "grad_norm": 0.057393934577703476, | |
| "learning_rate": 0.00027508143322475565, | |
| "loss": 0.7498, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.5293794618341571, | |
| "grad_norm": 0.049782704561948776, | |
| "learning_rate": 0.00027483713355048854, | |
| "loss": 0.7324, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.5315760571114773, | |
| "grad_norm": 0.04805954173207283, | |
| "learning_rate": 0.0002745928338762215, | |
| "loss": 0.6652, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.5337726523887973, | |
| "grad_norm": 0.04546598717570305, | |
| "learning_rate": 0.00027434853420195437, | |
| "loss": 0.6298, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.5359692476661175, | |
| "grad_norm": 0.06261162459850311, | |
| "learning_rate": 0.00027410423452768726, | |
| "loss": 0.8132, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.5381658429434377, | |
| "grad_norm": 0.05043161287903786, | |
| "learning_rate": 0.00027385993485342015, | |
| "loss": 0.6995, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.5403624382207578, | |
| "grad_norm": 0.058040980249643326, | |
| "learning_rate": 0.00027361563517915304, | |
| "loss": 0.7067, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.542559033498078, | |
| "grad_norm": 0.06532083451747894, | |
| "learning_rate": 0.000273371335504886, | |
| "loss": 0.6584, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.5447556287753982, | |
| "grad_norm": 0.07981614023447037, | |
| "learning_rate": 0.00027312703583061887, | |
| "loss": 0.7891, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.5469522240527183, | |
| "grad_norm": 0.04965173825621605, | |
| "learning_rate": 0.00027288273615635176, | |
| "loss": 0.6484, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.5491488193300385, | |
| "grad_norm": 0.060680437833070755, | |
| "learning_rate": 0.00027263843648208465, | |
| "loss": 0.741, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.5513454146073586, | |
| "grad_norm": 0.062146201729774475, | |
| "learning_rate": 0.00027239413680781754, | |
| "loss": 0.7158, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.5535420098846787, | |
| "grad_norm": 0.058960285037755966, | |
| "learning_rate": 0.0002721498371335504, | |
| "loss": 0.7222, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.5557386051619989, | |
| "grad_norm": 0.04842915013432503, | |
| "learning_rate": 0.00027190553745928337, | |
| "loss": 0.8026, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.5579352004393191, | |
| "grad_norm": 0.05300576239824295, | |
| "learning_rate": 0.00027166123778501626, | |
| "loss": 0.6973, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.5601317957166392, | |
| "grad_norm": 0.05646580457687378, | |
| "learning_rate": 0.00027141693811074915, | |
| "loss": 0.7424, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.5623283909939594, | |
| "grad_norm": 0.04713377356529236, | |
| "learning_rate": 0.00027117263843648204, | |
| "loss": 0.6688, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.5645249862712796, | |
| "grad_norm": 0.04357854649424553, | |
| "learning_rate": 0.0002709283387622149, | |
| "loss": 0.6035, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.5667215815485996, | |
| "grad_norm": 0.05619659647345543, | |
| "learning_rate": 0.00027068403908794787, | |
| "loss": 0.834, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.5689181768259198, | |
| "grad_norm": 0.05279466509819031, | |
| "learning_rate": 0.00027043973941368076, | |
| "loss": 0.6179, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.57111477210324, | |
| "grad_norm": 0.05597279593348503, | |
| "learning_rate": 0.00027019543973941365, | |
| "loss": 0.6988, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5733113673805601, | |
| "grad_norm": 0.050016503781080246, | |
| "learning_rate": 0.00026995114006514654, | |
| "loss": 0.7449, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.5755079626578803, | |
| "grad_norm": 0.04297545179724693, | |
| "learning_rate": 0.00026970684039087943, | |
| "loss": 0.7036, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.5777045579352005, | |
| "grad_norm": 0.045138273388147354, | |
| "learning_rate": 0.0002694625407166123, | |
| "loss": 0.6087, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.5799011532125206, | |
| "grad_norm": 0.06470554322004318, | |
| "learning_rate": 0.00026921824104234526, | |
| "loss": 0.7049, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.5820977484898407, | |
| "grad_norm": 0.05990125611424446, | |
| "learning_rate": 0.00026897394136807815, | |
| "loss": 0.9789, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.5842943437671609, | |
| "grad_norm": 0.04458112642168999, | |
| "learning_rate": 0.00026872964169381104, | |
| "loss": 0.7714, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.586490939044481, | |
| "grad_norm": 0.04603700712323189, | |
| "learning_rate": 0.00026848534201954393, | |
| "loss": 0.7234, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.5886875343218012, | |
| "grad_norm": 0.05222811922430992, | |
| "learning_rate": 0.0002682410423452768, | |
| "loss": 0.7531, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.5908841295991214, | |
| "grad_norm": 0.0743446946144104, | |
| "learning_rate": 0.00026799674267100976, | |
| "loss": 0.8359, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.5930807248764415, | |
| "grad_norm": 0.04982232674956322, | |
| "learning_rate": 0.00026775244299674265, | |
| "loss": 0.7521, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.5952773201537617, | |
| "grad_norm": 0.05457405373454094, | |
| "learning_rate": 0.00026750814332247554, | |
| "loss": 0.6456, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.5974739154310819, | |
| "grad_norm": 0.05579550936818123, | |
| "learning_rate": 0.00026726384364820843, | |
| "loss": 0.7309, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.5996705107084019, | |
| "grad_norm": 0.06359129399061203, | |
| "learning_rate": 0.0002670195439739413, | |
| "loss": 0.6133, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.6018671059857221, | |
| "grad_norm": 0.07442571967840195, | |
| "learning_rate": 0.0002667752442996742, | |
| "loss": 0.6771, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.6040637012630423, | |
| "grad_norm": 0.06577091664075851, | |
| "learning_rate": 0.00026653094462540715, | |
| "loss": 0.7365, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.6062602965403624, | |
| "grad_norm": 0.051220279186964035, | |
| "learning_rate": 0.00026628664495114004, | |
| "loss": 0.5611, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.6084568918176826, | |
| "grad_norm": 0.045341070741415024, | |
| "learning_rate": 0.00026604234527687293, | |
| "loss": 0.6786, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.6106534870950028, | |
| "grad_norm": 0.0620180107653141, | |
| "learning_rate": 0.0002657980456026058, | |
| "loss": 0.8132, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.6128500823723229, | |
| "grad_norm": 0.06206509843468666, | |
| "learning_rate": 0.0002655537459283387, | |
| "loss": 0.6962, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.615046677649643, | |
| "grad_norm": 0.05237254872918129, | |
| "learning_rate": 0.00026530944625407165, | |
| "loss": 0.6417, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.6172432729269632, | |
| "grad_norm": 0.04775477945804596, | |
| "learning_rate": 0.00026506514657980454, | |
| "loss": 0.5697, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.6194398682042833, | |
| "grad_norm": 0.05023486912250519, | |
| "learning_rate": 0.00026482084690553743, | |
| "loss": 0.6309, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.6216364634816035, | |
| "grad_norm": 0.06396158784627914, | |
| "learning_rate": 0.0002645765472312703, | |
| "loss": 0.7762, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.6238330587589237, | |
| "grad_norm": 0.04276019707322121, | |
| "learning_rate": 0.0002643322475570032, | |
| "loss": 0.6183, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.6260296540362438, | |
| "grad_norm": 0.045517537742853165, | |
| "learning_rate": 0.0002640879478827361, | |
| "loss": 0.6534, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.628226249313564, | |
| "grad_norm": 0.04261459410190582, | |
| "learning_rate": 0.00026384364820846904, | |
| "loss": 0.6309, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.6304228445908842, | |
| "grad_norm": 0.04310956224799156, | |
| "learning_rate": 0.00026359934853420193, | |
| "loss": 0.7227, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.6326194398682042, | |
| "grad_norm": 0.04422944039106369, | |
| "learning_rate": 0.0002633550488599348, | |
| "loss": 0.6375, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.6348160351455244, | |
| "grad_norm": 0.04686834290623665, | |
| "learning_rate": 0.0002631107491856677, | |
| "loss": 0.5817, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.6370126304228446, | |
| "grad_norm": 0.05435696616768837, | |
| "learning_rate": 0.0002628664495114006, | |
| "loss": 0.6136, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.6392092257001647, | |
| "grad_norm": 0.07915763556957245, | |
| "learning_rate": 0.0002626221498371335, | |
| "loss": 0.7798, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.6414058209774849, | |
| "grad_norm": 0.06726890057325363, | |
| "learning_rate": 0.00026237785016286643, | |
| "loss": 0.7894, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.6436024162548051, | |
| "grad_norm": 0.06199405714869499, | |
| "learning_rate": 0.0002621335504885993, | |
| "loss": 0.7938, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.6457990115321252, | |
| "grad_norm": 0.06117792800068855, | |
| "learning_rate": 0.00026188925081433226, | |
| "loss": 0.7667, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.6479956068094453, | |
| "grad_norm": 0.050212424248456955, | |
| "learning_rate": 0.00026164495114006515, | |
| "loss": 0.6959, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.6501922020867655, | |
| "grad_norm": 0.05087039992213249, | |
| "learning_rate": 0.00026140065146579804, | |
| "loss": 0.696, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.6523887973640856, | |
| "grad_norm": 0.04351802170276642, | |
| "learning_rate": 0.00026115635179153093, | |
| "loss": 0.6119, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.6545853926414058, | |
| "grad_norm": 0.05898391455411911, | |
| "learning_rate": 0.0002609120521172638, | |
| "loss": 0.6155, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.656781987918726, | |
| "grad_norm": 0.03949552774429321, | |
| "learning_rate": 0.00026066775244299676, | |
| "loss": 0.7009, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.6589785831960461, | |
| "grad_norm": 0.0561998188495636, | |
| "learning_rate": 0.00026042345276872965, | |
| "loss": 0.6573, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.6611751784733663, | |
| "grad_norm": 0.07027997076511383, | |
| "learning_rate": 0.00026017915309446254, | |
| "loss": 0.7692, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.6633717737506865, | |
| "grad_norm": 0.04690250754356384, | |
| "learning_rate": 0.00025993485342019543, | |
| "loss": 0.6603, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.6655683690280065, | |
| "grad_norm": 0.07027994841337204, | |
| "learning_rate": 0.0002596905537459283, | |
| "loss": 0.5999, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.6677649643053267, | |
| "grad_norm": 0.06607215851545334, | |
| "learning_rate": 0.0002594462540716612, | |
| "loss": 0.7808, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.6699615595826469, | |
| "grad_norm": 0.061052385717630386, | |
| "learning_rate": 0.00025920195439739415, | |
| "loss": 0.6839, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.6721581548599671, | |
| "grad_norm": 0.057753391563892365, | |
| "learning_rate": 0.00025895765472312704, | |
| "loss": 0.7209, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.6743547501372872, | |
| "grad_norm": 0.060185305774211884, | |
| "learning_rate": 0.00025871335504885993, | |
| "loss": 0.7376, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.6765513454146074, | |
| "grad_norm": 0.05048409849405289, | |
| "learning_rate": 0.0002584690553745928, | |
| "loss": 0.7646, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.6787479406919276, | |
| "grad_norm": 0.05279074236750603, | |
| "learning_rate": 0.0002582247557003257, | |
| "loss": 0.6964, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.6809445359692476, | |
| "grad_norm": 0.04628496244549751, | |
| "learning_rate": 0.00025798045602605865, | |
| "loss": 0.7263, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.6831411312465678, | |
| "grad_norm": 0.06292164325714111, | |
| "learning_rate": 0.00025773615635179154, | |
| "loss": 0.5775, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.685337726523888, | |
| "grad_norm": 0.05216914042830467, | |
| "learning_rate": 0.00025749185667752443, | |
| "loss": 0.7599, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6875343218012081, | |
| "grad_norm": 0.0711362361907959, | |
| "learning_rate": 0.0002572475570032573, | |
| "loss": 0.6705, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.6897309170785283, | |
| "grad_norm": 0.03843238577246666, | |
| "learning_rate": 0.0002570032573289902, | |
| "loss": 0.4722, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.6919275123558485, | |
| "grad_norm": 0.05418127775192261, | |
| "learning_rate": 0.0002567589576547231, | |
| "loss": 0.5147, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.6941241076331686, | |
| "grad_norm": 0.061103999614715576, | |
| "learning_rate": 0.00025651465798045604, | |
| "loss": 0.761, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.6963207029104888, | |
| "grad_norm": 0.050811514258384705, | |
| "learning_rate": 0.00025627035830618893, | |
| "loss": 0.8575, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.6985172981878089, | |
| "grad_norm": 0.05052623152732849, | |
| "learning_rate": 0.0002560260586319218, | |
| "loss": 0.7842, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.700713893465129, | |
| "grad_norm": 0.05087927356362343, | |
| "learning_rate": 0.0002557817589576547, | |
| "loss": 0.6584, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.7029104887424492, | |
| "grad_norm": 0.041777510195970535, | |
| "learning_rate": 0.0002555374592833876, | |
| "loss": 0.6491, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.7051070840197694, | |
| "grad_norm": 0.045631300657987595, | |
| "learning_rate": 0.0002552931596091205, | |
| "loss": 0.7189, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.7073036792970895, | |
| "grad_norm": 0.04235495626926422, | |
| "learning_rate": 0.00025504885993485343, | |
| "loss": 0.7269, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.7095002745744097, | |
| "grad_norm": 0.04950536787509918, | |
| "learning_rate": 0.0002548045602605863, | |
| "loss": 0.5914, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.7116968698517299, | |
| "grad_norm": 0.05852117761969566, | |
| "learning_rate": 0.0002545602605863192, | |
| "loss": 0.7652, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.7138934651290499, | |
| "grad_norm": 0.0719272643327713, | |
| "learning_rate": 0.0002543159609120521, | |
| "loss": 0.7394, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.7160900604063701, | |
| "grad_norm": 0.0593254491686821, | |
| "learning_rate": 0.000254071661237785, | |
| "loss": 0.6966, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.7182866556836903, | |
| "grad_norm": 0.04738825559616089, | |
| "learning_rate": 0.00025382736156351793, | |
| "loss": 0.6779, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.7204832509610104, | |
| "grad_norm": 0.05696015805006027, | |
| "learning_rate": 0.0002535830618892508, | |
| "loss": 0.7888, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.7226798462383306, | |
| "grad_norm": 0.03948872163891792, | |
| "learning_rate": 0.0002533387622149837, | |
| "loss": 0.6655, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.7248764415156508, | |
| "grad_norm": 0.05530393868684769, | |
| "learning_rate": 0.0002530944625407166, | |
| "loss": 0.6375, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.7270730367929709, | |
| "grad_norm": 0.06435840576887131, | |
| "learning_rate": 0.0002528501628664495, | |
| "loss": 0.8043, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.729269632070291, | |
| "grad_norm": 0.04414074495434761, | |
| "learning_rate": 0.0002526058631921824, | |
| "loss": 0.6234, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.7314662273476112, | |
| "grad_norm": 0.04935838282108307, | |
| "learning_rate": 0.0002523615635179153, | |
| "loss": 0.7207, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.7336628226249313, | |
| "grad_norm": 0.04916159436106682, | |
| "learning_rate": 0.0002521172638436482, | |
| "loss": 0.7327, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.7358594179022515, | |
| "grad_norm": 0.0506322905421257, | |
| "learning_rate": 0.0002518729641693811, | |
| "loss": 0.705, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.7380560131795717, | |
| "grad_norm": 0.05323386564850807, | |
| "learning_rate": 0.000251628664495114, | |
| "loss": 0.6962, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.7402526084568918, | |
| "grad_norm": 0.042582739144563675, | |
| "learning_rate": 0.0002513843648208469, | |
| "loss": 0.8309, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.742449203734212, | |
| "grad_norm": 0.05412566661834717, | |
| "learning_rate": 0.0002511400651465798, | |
| "loss": 0.6626, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.7446457990115322, | |
| "grad_norm": 0.057325392961502075, | |
| "learning_rate": 0.0002508957654723127, | |
| "loss": 0.7543, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.7468423942888522, | |
| "grad_norm": 0.054078057408332825, | |
| "learning_rate": 0.0002506514657980456, | |
| "loss": 0.7228, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.7490389895661724, | |
| "grad_norm": 0.05624162778258324, | |
| "learning_rate": 0.0002504071661237785, | |
| "loss": 0.65, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.7512355848434926, | |
| "grad_norm": 0.04718158766627312, | |
| "learning_rate": 0.0002501628664495114, | |
| "loss": 0.5851, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.7534321801208127, | |
| "grad_norm": 0.044785093516111374, | |
| "learning_rate": 0.00024991856677524427, | |
| "loss": 0.653, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.7556287753981329, | |
| "grad_norm": 0.052535783499479294, | |
| "learning_rate": 0.0002496742671009772, | |
| "loss": 0.6678, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.7578253706754531, | |
| "grad_norm": 0.04284857586026192, | |
| "learning_rate": 0.0002494299674267101, | |
| "loss": 0.577, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.7600219659527732, | |
| "grad_norm": 0.07952070236206055, | |
| "learning_rate": 0.000249185667752443, | |
| "loss": 0.7346, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.7622185612300933, | |
| "grad_norm": 0.0500759556889534, | |
| "learning_rate": 0.0002489413680781759, | |
| "loss": 0.6877, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.7644151565074135, | |
| "grad_norm": 0.05391065031290054, | |
| "learning_rate": 0.00024869706840390877, | |
| "loss": 0.6815, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.7666117517847336, | |
| "grad_norm": 0.041693173348903656, | |
| "learning_rate": 0.0002484527687296417, | |
| "loss": 0.5043, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.7688083470620538, | |
| "grad_norm": 0.055352237075567245, | |
| "learning_rate": 0.0002482084690553746, | |
| "loss": 0.8104, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.771004942339374, | |
| "grad_norm": 0.059511568397283554, | |
| "learning_rate": 0.0002479641693811075, | |
| "loss": 0.9011, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.7732015376166941, | |
| "grad_norm": 0.06932779401540756, | |
| "learning_rate": 0.0002477198697068404, | |
| "loss": 0.6964, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.7753981328940143, | |
| "grad_norm": 0.045885760337114334, | |
| "learning_rate": 0.00024747557003257327, | |
| "loss": 0.5526, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.7775947281713345, | |
| "grad_norm": 0.06347452104091644, | |
| "learning_rate": 0.00024723127035830616, | |
| "loss": 0.6988, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.7797913234486545, | |
| "grad_norm": 0.060996223241090775, | |
| "learning_rate": 0.0002469869706840391, | |
| "loss": 0.518, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.7819879187259747, | |
| "grad_norm": 0.05093398317694664, | |
| "learning_rate": 0.000246742671009772, | |
| "loss": 0.5963, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.7841845140032949, | |
| "grad_norm": 0.04339175298810005, | |
| "learning_rate": 0.0002464983713355049, | |
| "loss": 0.7216, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.786381109280615, | |
| "grad_norm": 0.04919297620654106, | |
| "learning_rate": 0.00024625407166123777, | |
| "loss": 0.6406, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.7885777045579352, | |
| "grad_norm": 0.08795499056577682, | |
| "learning_rate": 0.00024600977198697066, | |
| "loss": 0.6478, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.7907742998352554, | |
| "grad_norm": 0.04211808741092682, | |
| "learning_rate": 0.00024576547231270354, | |
| "loss": 0.607, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.7929708951125755, | |
| "grad_norm": 0.05462827533483505, | |
| "learning_rate": 0.0002455211726384365, | |
| "loss": 0.7166, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.7951674903898956, | |
| "grad_norm": 0.0573272630572319, | |
| "learning_rate": 0.0002452768729641694, | |
| "loss": 0.7341, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.7973640856672158, | |
| "grad_norm": 0.04874618351459503, | |
| "learning_rate": 0.00024503257328990227, | |
| "loss": 0.5878, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.7995606809445359, | |
| "grad_norm": 0.056756071746349335, | |
| "learning_rate": 0.00024478827361563516, | |
| "loss": 0.6283, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.8017572762218561, | |
| "grad_norm": 0.050058893859386444, | |
| "learning_rate": 0.00024454397394136804, | |
| "loss": 0.5111, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.8039538714991763, | |
| "grad_norm": 0.0445009246468544, | |
| "learning_rate": 0.000244299674267101, | |
| "loss": 0.7481, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.8061504667764964, | |
| "grad_norm": 0.09190984815359116, | |
| "learning_rate": 0.00024405537459283385, | |
| "loss": 0.744, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.8083470620538166, | |
| "grad_norm": 0.05386214330792427, | |
| "learning_rate": 0.00024381107491856677, | |
| "loss": 0.5524, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.8105436573311368, | |
| "grad_norm": 0.03755839914083481, | |
| "learning_rate": 0.00024356677524429966, | |
| "loss": 0.5951, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.8127402526084568, | |
| "grad_norm": 0.04785095900297165, | |
| "learning_rate": 0.00024332247557003257, | |
| "loss": 0.5013, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.814936847885777, | |
| "grad_norm": 0.035032473504543304, | |
| "learning_rate": 0.00024307817589576546, | |
| "loss": 0.7563, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.8171334431630972, | |
| "grad_norm": 0.05640648305416107, | |
| "learning_rate": 0.00024283387622149835, | |
| "loss": 0.6958, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.8193300384404174, | |
| "grad_norm": 0.03728983923792839, | |
| "learning_rate": 0.00024258957654723127, | |
| "loss": 0.6545, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.8215266337177375, | |
| "grad_norm": 0.07797161489725113, | |
| "learning_rate": 0.00024234527687296416, | |
| "loss": 0.6847, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.8237232289950577, | |
| "grad_norm": 0.055005770176649094, | |
| "learning_rate": 0.00024210097719869705, | |
| "loss": 0.7758, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.8259198242723779, | |
| "grad_norm": 0.05566948279738426, | |
| "learning_rate": 0.00024185667752442996, | |
| "loss": 0.754, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.828116419549698, | |
| "grad_norm": 0.05472628399729729, | |
| "learning_rate": 0.00024161237785016285, | |
| "loss": 0.688, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.8303130148270181, | |
| "grad_norm": 0.04331599548459053, | |
| "learning_rate": 0.00024136807817589574, | |
| "loss": 0.646, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.8325096101043383, | |
| "grad_norm": 0.04198076203465462, | |
| "learning_rate": 0.00024112377850162866, | |
| "loss": 0.5098, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.8347062053816584, | |
| "grad_norm": 0.05673768371343613, | |
| "learning_rate": 0.00024087947882736155, | |
| "loss": 0.7292, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.8369028006589786, | |
| "grad_norm": 0.06165642663836479, | |
| "learning_rate": 0.00024063517915309446, | |
| "loss": 0.6463, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.8390993959362988, | |
| "grad_norm": 0.05126790329813957, | |
| "learning_rate": 0.00024039087947882735, | |
| "loss": 0.7427, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.8412959912136189, | |
| "grad_norm": 0.04800880700349808, | |
| "learning_rate": 0.00024014657980456024, | |
| "loss": 0.7385, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.8434925864909391, | |
| "grad_norm": 0.06537099927663803, | |
| "learning_rate": 0.00023990228013029316, | |
| "loss": 0.7598, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.8456891817682592, | |
| "grad_norm": 0.0711362361907959, | |
| "learning_rate": 0.00023965798045602605, | |
| "loss": 0.8168, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.8478857770455793, | |
| "grad_norm": 0.06028452143073082, | |
| "learning_rate": 0.00023941368078175893, | |
| "loss": 0.7449, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.8500823723228995, | |
| "grad_norm": 0.04362437129020691, | |
| "learning_rate": 0.00023916938110749185, | |
| "loss": 0.6074, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.8522789676002197, | |
| "grad_norm": 0.05339638888835907, | |
| "learning_rate": 0.00023892508143322474, | |
| "loss": 0.7208, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.8544755628775398, | |
| "grad_norm": 0.049732714891433716, | |
| "learning_rate": 0.00023868078175895763, | |
| "loss": 0.5567, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.85667215815486, | |
| "grad_norm": 0.050275735557079315, | |
| "learning_rate": 0.00023843648208469055, | |
| "loss": 0.5994, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.8588687534321802, | |
| "grad_norm": 0.05760905146598816, | |
| "learning_rate": 0.00023819218241042343, | |
| "loss": 0.7006, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.8610653487095002, | |
| "grad_norm": 0.06789691001176834, | |
| "learning_rate": 0.00023794788273615632, | |
| "loss": 0.8992, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.8632619439868204, | |
| "grad_norm": 0.06151839718222618, | |
| "learning_rate": 0.00023770358306188924, | |
| "loss": 0.9705, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.8654585392641406, | |
| "grad_norm": 0.055463626980781555, | |
| "learning_rate": 0.00023745928338762213, | |
| "loss": 0.6691, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.8676551345414607, | |
| "grad_norm": 0.04577269032597542, | |
| "learning_rate": 0.00023721498371335505, | |
| "loss": 0.5623, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.8698517298187809, | |
| "grad_norm": 0.04549311101436615, | |
| "learning_rate": 0.00023697068403908794, | |
| "loss": 0.6417, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.8720483250961011, | |
| "grad_norm": 0.04944036155939102, | |
| "learning_rate": 0.00023672638436482082, | |
| "loss": 0.6516, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.8742449203734212, | |
| "grad_norm": 0.042285725474357605, | |
| "learning_rate": 0.00023648208469055374, | |
| "loss": 0.5873, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.8764415156507414, | |
| "grad_norm": 0.058092810213565826, | |
| "learning_rate": 0.00023623778501628663, | |
| "loss": 0.7076, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.8786381109280615, | |
| "grad_norm": 0.05264095962047577, | |
| "learning_rate": 0.00023599348534201952, | |
| "loss": 0.7239, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.8808347062053816, | |
| "grad_norm": 0.07236024737358093, | |
| "learning_rate": 0.00023574918566775244, | |
| "loss": 0.5782, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.8830313014827018, | |
| "grad_norm": 0.03647659346461296, | |
| "learning_rate": 0.00023550488599348532, | |
| "loss": 0.6416, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.885227896760022, | |
| "grad_norm": 0.05197187140583992, | |
| "learning_rate": 0.00023526058631921821, | |
| "loss": 0.8022, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.8874244920373421, | |
| "grad_norm": 0.03680606186389923, | |
| "learning_rate": 0.00023501628664495113, | |
| "loss": 0.796, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.8896210873146623, | |
| "grad_norm": 0.04213280603289604, | |
| "learning_rate": 0.00023477198697068402, | |
| "loss": 0.5616, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.8918176825919825, | |
| "grad_norm": 0.04371798038482666, | |
| "learning_rate": 0.0002345276872964169, | |
| "loss": 0.5991, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.8940142778693025, | |
| "grad_norm": 0.053443361073732376, | |
| "learning_rate": 0.00023428338762214982, | |
| "loss": 0.544, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.8962108731466227, | |
| "grad_norm": 0.04580046981573105, | |
| "learning_rate": 0.00023403908794788271, | |
| "loss": 0.5709, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.8984074684239429, | |
| "grad_norm": 0.0343971811234951, | |
| "learning_rate": 0.00023379478827361563, | |
| "loss": 0.6029, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.900604063701263, | |
| "grad_norm": 0.062083858996629715, | |
| "learning_rate": 0.00023355048859934852, | |
| "loss": 0.7742, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.9028006589785832, | |
| "grad_norm": 0.040063194930553436, | |
| "learning_rate": 0.0002333061889250814, | |
| "loss": 0.5643, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.9049972542559034, | |
| "grad_norm": 0.06296666711568832, | |
| "learning_rate": 0.00023306188925081432, | |
| "loss": 0.6406, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.9071938495332235, | |
| "grad_norm": 0.04029727727174759, | |
| "learning_rate": 0.00023281758957654721, | |
| "loss": 0.6608, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.9093904448105437, | |
| "grad_norm": 0.055288396775722504, | |
| "learning_rate": 0.0002325732899022801, | |
| "loss": 0.7396, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.9115870400878638, | |
| "grad_norm": 0.044744208455085754, | |
| "learning_rate": 0.00023232899022801302, | |
| "loss": 0.646, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.9137836353651839, | |
| "grad_norm": 0.06047806143760681, | |
| "learning_rate": 0.0002320846905537459, | |
| "loss": 0.8133, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.9159802306425041, | |
| "grad_norm": 0.054290421307086945, | |
| "learning_rate": 0.0002318403908794788, | |
| "loss": 0.6989, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.9181768259198243, | |
| "grad_norm": 0.05836010351777077, | |
| "learning_rate": 0.00023159609120521171, | |
| "loss": 0.5014, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.9203734211971444, | |
| "grad_norm": 0.05172126740217209, | |
| "learning_rate": 0.0002313517915309446, | |
| "loss": 0.5723, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.9225700164744646, | |
| "grad_norm": 0.03858590126037598, | |
| "learning_rate": 0.00023110749185667752, | |
| "loss": 0.6683, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.9247666117517848, | |
| "grad_norm": 0.05288545414805412, | |
| "learning_rate": 0.0002308631921824104, | |
| "loss": 0.6083, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.9269632070291048, | |
| "grad_norm": 0.049517158418893814, | |
| "learning_rate": 0.0002306188925081433, | |
| "loss": 0.6561, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.929159802306425, | |
| "grad_norm": 0.0585198700428009, | |
| "learning_rate": 0.00023037459283387621, | |
| "loss": 0.8331, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.9313563975837452, | |
| "grad_norm": 0.05595822259783745, | |
| "learning_rate": 0.0002301302931596091, | |
| "loss": 0.6627, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.9335529928610653, | |
| "grad_norm": 0.03902159631252289, | |
| "learning_rate": 0.000229885993485342, | |
| "loss": 0.5872, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.9357495881383855, | |
| "grad_norm": 0.04444313421845436, | |
| "learning_rate": 0.0002296416938110749, | |
| "loss": 0.6126, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.9379461834157057, | |
| "grad_norm": 0.04075470194220543, | |
| "learning_rate": 0.0002293973941368078, | |
| "loss": 0.6372, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.9401427786930258, | |
| "grad_norm": 0.04287320002913475, | |
| "learning_rate": 0.0002291530944625407, | |
| "loss": 0.5355, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.942339373970346, | |
| "grad_norm": 0.04004767909646034, | |
| "learning_rate": 0.0002289087947882736, | |
| "loss": 0.476, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.9445359692476661, | |
| "grad_norm": 0.04870041459798813, | |
| "learning_rate": 0.0002286644951140065, | |
| "loss": 0.7218, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.9467325645249862, | |
| "grad_norm": 0.05161641538143158, | |
| "learning_rate": 0.00022842019543973938, | |
| "loss": 0.5477, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.9489291598023064, | |
| "grad_norm": 0.09497353434562683, | |
| "learning_rate": 0.0002281758957654723, | |
| "loss": 0.8361, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.9511257550796266, | |
| "grad_norm": 0.04529595747590065, | |
| "learning_rate": 0.0002279315960912052, | |
| "loss": 0.6591, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.9533223503569467, | |
| "grad_norm": 0.049864016473293304, | |
| "learning_rate": 0.0002276872964169381, | |
| "loss": 0.561, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.9555189456342669, | |
| "grad_norm": 0.03673512116074562, | |
| "learning_rate": 0.000227442996742671, | |
| "loss": 0.5714, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.9577155409115871, | |
| "grad_norm": 0.04806803911924362, | |
| "learning_rate": 0.00022719869706840388, | |
| "loss": 0.5068, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.9599121361889071, | |
| "grad_norm": 0.046189140528440475, | |
| "learning_rate": 0.0002269543973941368, | |
| "loss": 0.6344, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.9621087314662273, | |
| "grad_norm": 0.042612023651599884, | |
| "learning_rate": 0.0002267100977198697, | |
| "loss": 0.5453, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.9643053267435475, | |
| "grad_norm": 0.04492894560098648, | |
| "learning_rate": 0.00022646579804560258, | |
| "loss": 0.6854, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.9665019220208677, | |
| "grad_norm": 0.05652786046266556, | |
| "learning_rate": 0.0002262214983713355, | |
| "loss": 0.7428, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.9686985172981878, | |
| "grad_norm": 0.054750144481658936, | |
| "learning_rate": 0.00022597719869706838, | |
| "loss": 0.6135, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.970895112575508, | |
| "grad_norm": 0.04848687723278999, | |
| "learning_rate": 0.00022573289902280127, | |
| "loss": 0.5865, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.9730917078528282, | |
| "grad_norm": 0.03940533474087715, | |
| "learning_rate": 0.0002254885993485342, | |
| "loss": 0.6836, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.9752883031301482, | |
| "grad_norm": 0.04875492677092552, | |
| "learning_rate": 0.00022524429967426708, | |
| "loss": 0.7485, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.9774848984074684, | |
| "grad_norm": 0.06128212809562683, | |
| "learning_rate": 0.000225, | |
| "loss": 0.73, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.9796814936847886, | |
| "grad_norm": 0.09782739728689194, | |
| "learning_rate": 0.00022475570032573288, | |
| "loss": 0.7955, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.9818780889621087, | |
| "grad_norm": 0.06957724690437317, | |
| "learning_rate": 0.00022451140065146577, | |
| "loss": 0.7571, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.9840746842394289, | |
| "grad_norm": 0.04065481945872307, | |
| "learning_rate": 0.0002242671009771987, | |
| "loss": 0.6567, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.9862712795167491, | |
| "grad_norm": 0.04748554527759552, | |
| "learning_rate": 0.00022402280130293158, | |
| "loss": 0.7358, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.9884678747940692, | |
| "grad_norm": 0.04544506594538689, | |
| "learning_rate": 0.00022377850162866447, | |
| "loss": 0.6632, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.9906644700713894, | |
| "grad_norm": 0.046616628766059875, | |
| "learning_rate": 0.00022353420195439738, | |
| "loss": 0.608, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.9928610653487095, | |
| "grad_norm": 0.04831715673208237, | |
| "learning_rate": 0.00022328990228013027, | |
| "loss": 0.6751, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.9950576606260296, | |
| "grad_norm": 0.04835525527596474, | |
| "learning_rate": 0.00022304560260586316, | |
| "loss": 0.7125, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.9972542559033498, | |
| "grad_norm": 0.043945323675870895, | |
| "learning_rate": 0.00022280130293159608, | |
| "loss": 0.5494, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.99945085118067, | |
| "grad_norm": 0.04118410125374794, | |
| "learning_rate": 0.00022255700325732897, | |
| "loss": 0.6313, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 1.00164744645799, | |
| "grad_norm": 0.05268587917089462, | |
| "learning_rate": 0.00022231270358306186, | |
| "loss": 0.7393, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 1.0038440417353103, | |
| "grad_norm": 0.04583257809281349, | |
| "learning_rate": 0.00022206840390879477, | |
| "loss": 0.7311, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 1.0060406370126305, | |
| "grad_norm": 0.041914235800504684, | |
| "learning_rate": 0.00022182410423452766, | |
| "loss": 0.6592, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 1.0082372322899507, | |
| "grad_norm": 0.05415317043662071, | |
| "learning_rate": 0.00022157980456026058, | |
| "loss": 0.5943, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 1.0104338275672706, | |
| "grad_norm": 0.039053451269865036, | |
| "learning_rate": 0.00022133550488599347, | |
| "loss": 0.5909, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.0126304228445908, | |
| "grad_norm": 0.04172314330935478, | |
| "learning_rate": 0.00022109120521172636, | |
| "loss": 0.529, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 1.014827018121911, | |
| "grad_norm": 0.059966228902339935, | |
| "learning_rate": 0.00022084690553745927, | |
| "loss": 0.8445, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 1.0170236133992312, | |
| "grad_norm": 0.054427627474069595, | |
| "learning_rate": 0.00022060260586319216, | |
| "loss": 0.7625, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 1.0192202086765514, | |
| "grad_norm": 0.05113132670521736, | |
| "learning_rate": 0.00022035830618892505, | |
| "loss": 0.6079, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 1.0214168039538716, | |
| "grad_norm": 0.05766759812831879, | |
| "learning_rate": 0.00022011400651465797, | |
| "loss": 0.7953, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 1.0236133992311915, | |
| "grad_norm": 0.055021774023771286, | |
| "learning_rate": 0.00021986970684039086, | |
| "loss": 0.8009, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 1.0258099945085117, | |
| "grad_norm": 0.049860063940286636, | |
| "learning_rate": 0.00021962540716612375, | |
| "loss": 0.683, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 1.028006589785832, | |
| "grad_norm": 0.059873066842556, | |
| "learning_rate": 0.00021938110749185666, | |
| "loss": 0.6204, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 1.0302031850631521, | |
| "grad_norm": 0.04648579657077789, | |
| "learning_rate": 0.00021913680781758955, | |
| "loss": 0.7137, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 1.0323997803404723, | |
| "grad_norm": 0.05671359598636627, | |
| "learning_rate": 0.00021889250814332244, | |
| "loss": 0.7367, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.0345963756177925, | |
| "grad_norm": 0.051784005016088486, | |
| "learning_rate": 0.00021864820846905536, | |
| "loss": 0.5914, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 1.0367929708951125, | |
| "grad_norm": 0.05361558869481087, | |
| "learning_rate": 0.00021840390879478825, | |
| "loss": 0.6544, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 1.0389895661724327, | |
| "grad_norm": 0.05350130423903465, | |
| "learning_rate": 0.00021815960912052116, | |
| "loss": 0.8281, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 1.0411861614497528, | |
| "grad_norm": 0.046200599521398544, | |
| "learning_rate": 0.00021791530944625405, | |
| "loss": 0.6194, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 1.043382756727073, | |
| "grad_norm": 0.08131860941648483, | |
| "learning_rate": 0.00021767100977198694, | |
| "loss": 0.6572, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 1.0455793520043932, | |
| "grad_norm": 0.050514254719018936, | |
| "learning_rate": 0.00021742671009771986, | |
| "loss": 0.7339, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 1.0477759472817134, | |
| "grad_norm": 0.05298462510108948, | |
| "learning_rate": 0.00021718241042345275, | |
| "loss": 0.5557, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 1.0499725425590336, | |
| "grad_norm": 0.05039222911000252, | |
| "learning_rate": 0.00021693811074918564, | |
| "loss": 0.7133, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 1.0521691378363536, | |
| "grad_norm": 0.05472245067358017, | |
| "learning_rate": 0.00021669381107491855, | |
| "loss": 0.7076, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 1.0543657331136738, | |
| "grad_norm": 0.039077069610357285, | |
| "learning_rate": 0.00021644951140065144, | |
| "loss": 0.5951, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.056562328390994, | |
| "grad_norm": 0.03200119361281395, | |
| "learning_rate": 0.00021620521172638433, | |
| "loss": 0.5825, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 1.0587589236683141, | |
| "grad_norm": 0.04200253635644913, | |
| "learning_rate": 0.00021596091205211725, | |
| "loss": 0.5881, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 1.0609555189456343, | |
| "grad_norm": 0.0591256357729435, | |
| "learning_rate": 0.00021571661237785014, | |
| "loss": 0.647, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 1.0631521142229543, | |
| "grad_norm": 0.04347528517246246, | |
| "learning_rate": 0.00021547231270358305, | |
| "loss": 0.671, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 1.0653487095002745, | |
| "grad_norm": 0.04377702996134758, | |
| "learning_rate": 0.00021522801302931594, | |
| "loss": 0.6533, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 1.0675453047775947, | |
| "grad_norm": 0.05368449166417122, | |
| "learning_rate": 0.00021498371335504883, | |
| "loss": 0.5239, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 1.0697419000549149, | |
| "grad_norm": 0.04010557755827904, | |
| "learning_rate": 0.00021473941368078175, | |
| "loss": 0.6293, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 1.071938495332235, | |
| "grad_norm": 0.04686107859015465, | |
| "learning_rate": 0.00021449511400651464, | |
| "loss": 0.7006, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 1.0741350906095553, | |
| "grad_norm": 0.03812983259558678, | |
| "learning_rate": 0.00021425081433224753, | |
| "loss": 0.6003, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 1.0763316858868754, | |
| "grad_norm": 0.035127971321344376, | |
| "learning_rate": 0.00021400651465798044, | |
| "loss": 0.5701, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.0785282811641954, | |
| "grad_norm": 0.07218277454376221, | |
| "learning_rate": 0.00021376221498371333, | |
| "loss": 0.8555, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 1.0807248764415156, | |
| "grad_norm": 0.046801965683698654, | |
| "learning_rate": 0.00021351791530944622, | |
| "loss": 0.5919, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 1.0829214717188358, | |
| "grad_norm": 0.06883776932954788, | |
| "learning_rate": 0.00021327361563517914, | |
| "loss": 0.7253, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 1.085118066996156, | |
| "grad_norm": 0.03979892656207085, | |
| "learning_rate": 0.00021302931596091203, | |
| "loss": 0.5356, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 1.0873146622734762, | |
| "grad_norm": 0.07111027836799622, | |
| "learning_rate": 0.00021278501628664491, | |
| "loss": 0.6938, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 1.0895112575507964, | |
| "grad_norm": 0.040993962436914444, | |
| "learning_rate": 0.00021254071661237783, | |
| "loss": 0.6093, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 1.0917078528281163, | |
| "grad_norm": 0.033425960689783096, | |
| "learning_rate": 0.00021229641693811072, | |
| "loss": 0.5784, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 1.0939044481054365, | |
| "grad_norm": 0.058829471468925476, | |
| "learning_rate": 0.00021205211726384364, | |
| "loss": 0.6227, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 1.0961010433827567, | |
| "grad_norm": 0.04688352346420288, | |
| "learning_rate": 0.00021180781758957653, | |
| "loss": 0.6884, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 1.098297638660077, | |
| "grad_norm": 0.059555862098932266, | |
| "learning_rate": 0.00021156351791530941, | |
| "loss": 0.7715, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.100494233937397, | |
| "grad_norm": 0.0509280227124691, | |
| "learning_rate": 0.00021131921824104233, | |
| "loss": 0.7527, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 1.1026908292147173, | |
| "grad_norm": 0.036750249564647675, | |
| "learning_rate": 0.00021107491856677522, | |
| "loss": 0.5671, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 1.1048874244920373, | |
| "grad_norm": 0.07052327692508698, | |
| "learning_rate": 0.0002108306188925081, | |
| "loss": 0.7258, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 1.1070840197693574, | |
| "grad_norm": 0.0397505946457386, | |
| "learning_rate": 0.00021058631921824103, | |
| "loss": 0.5837, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 1.1092806150466776, | |
| "grad_norm": 0.053040631115436554, | |
| "learning_rate": 0.00021034201954397392, | |
| "loss": 0.8447, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 1.1114772103239978, | |
| "grad_norm": 0.04555618017911911, | |
| "learning_rate": 0.0002100977198697068, | |
| "loss": 0.6716, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 1.113673805601318, | |
| "grad_norm": 0.04914266616106033, | |
| "learning_rate": 0.00020985342019543972, | |
| "loss": 0.6719, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 1.1158704008786382, | |
| "grad_norm": 0.05914847552776337, | |
| "learning_rate": 0.0002096091205211726, | |
| "loss": 0.738, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 1.1180669961559582, | |
| "grad_norm": 0.0362103171646595, | |
| "learning_rate": 0.0002093648208469055, | |
| "loss": 0.6914, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 1.1202635914332784, | |
| "grad_norm": 0.04875648021697998, | |
| "learning_rate": 0.00020912052117263842, | |
| "loss": 0.7716, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.1224601867105986, | |
| "grad_norm": 0.052181925624608994, | |
| "learning_rate": 0.0002088762214983713, | |
| "loss": 0.6453, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 1.1246567819879187, | |
| "grad_norm": 0.054814066737890244, | |
| "learning_rate": 0.00020863192182410422, | |
| "loss": 0.6659, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 1.126853377265239, | |
| "grad_norm": 0.047833219170570374, | |
| "learning_rate": 0.0002083876221498371, | |
| "loss": 0.7771, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 1.1290499725425591, | |
| "grad_norm": 0.056477759033441544, | |
| "learning_rate": 0.00020814332247557, | |
| "loss": 0.6761, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 1.131246567819879, | |
| "grad_norm": 0.050696369260549545, | |
| "learning_rate": 0.00020789902280130292, | |
| "loss": 0.6711, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 1.1334431630971993, | |
| "grad_norm": 0.05367648974061012, | |
| "learning_rate": 0.0002076547231270358, | |
| "loss": 0.6757, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 1.1356397583745195, | |
| "grad_norm": 0.05610479786992073, | |
| "learning_rate": 0.0002074104234527687, | |
| "loss": 0.6412, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 1.1378363536518397, | |
| "grad_norm": 0.0620594285428524, | |
| "learning_rate": 0.0002071661237785016, | |
| "loss": 0.519, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 1.1400329489291599, | |
| "grad_norm": 0.05160127580165863, | |
| "learning_rate": 0.0002069218241042345, | |
| "loss": 0.4844, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 1.14222954420648, | |
| "grad_norm": 0.04619447514414787, | |
| "learning_rate": 0.0002066775244299674, | |
| "loss": 0.6709, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.1444261394838002, | |
| "grad_norm": 0.055884506553411484, | |
| "learning_rate": 0.0002064332247557003, | |
| "loss": 0.5976, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 1.1466227347611202, | |
| "grad_norm": 0.050843894481658936, | |
| "learning_rate": 0.0002061889250814332, | |
| "loss": 0.6171, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 1.1488193300384404, | |
| "grad_norm": 0.04294108971953392, | |
| "learning_rate": 0.0002059446254071661, | |
| "loss": 0.5099, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 1.1510159253157606, | |
| "grad_norm": 0.04971957579255104, | |
| "learning_rate": 0.000205700325732899, | |
| "loss": 0.7442, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 1.1532125205930808, | |
| "grad_norm": 0.037918634712696075, | |
| "learning_rate": 0.0002054560260586319, | |
| "loss": 0.6345, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 1.155409115870401, | |
| "grad_norm": 0.0499645359814167, | |
| "learning_rate": 0.0002052117263843648, | |
| "loss": 0.8493, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 1.157605711147721, | |
| "grad_norm": 0.04419689625501633, | |
| "learning_rate": 0.0002049674267100977, | |
| "loss": 0.5747, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 1.1598023064250411, | |
| "grad_norm": 0.06275136768817902, | |
| "learning_rate": 0.00020472312703583058, | |
| "loss": 0.7256, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 1.1619989017023613, | |
| "grad_norm": 0.05085553973913193, | |
| "learning_rate": 0.0002044788273615635, | |
| "loss": 0.5588, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 1.1641954969796815, | |
| "grad_norm": 0.04484010860323906, | |
| "learning_rate": 0.0002042345276872964, | |
| "loss": 0.4919, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.1663920922570017, | |
| "grad_norm": 0.044836804270744324, | |
| "learning_rate": 0.00020399022801302928, | |
| "loss": 0.5655, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 1.1685886875343219, | |
| "grad_norm": 0.04288703575730324, | |
| "learning_rate": 0.0002037459283387622, | |
| "loss": 0.7729, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 1.170785282811642, | |
| "grad_norm": 0.04978484287858009, | |
| "learning_rate": 0.00020350162866449508, | |
| "loss": 0.6874, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 1.172981878088962, | |
| "grad_norm": 0.061364252120256424, | |
| "learning_rate": 0.00020325732899022797, | |
| "loss": 0.7798, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 1.1751784733662822, | |
| "grad_norm": 0.04522540792822838, | |
| "learning_rate": 0.0002030130293159609, | |
| "loss": 0.6139, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 1.1773750686436024, | |
| "grad_norm": 0.0677039623260498, | |
| "learning_rate": 0.00020276872964169378, | |
| "loss": 0.6335, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 1.1795716639209226, | |
| "grad_norm": 0.04621490091085434, | |
| "learning_rate": 0.0002025244299674267, | |
| "loss": 0.6156, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 1.1817682591982428, | |
| "grad_norm": 0.04518371820449829, | |
| "learning_rate": 0.00020228013029315958, | |
| "loss": 0.5472, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 1.1839648544755628, | |
| "grad_norm": 0.0515042245388031, | |
| "learning_rate": 0.00020203583061889247, | |
| "loss": 0.8595, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 1.186161449752883, | |
| "grad_norm": 0.0516064427793026, | |
| "learning_rate": 0.0002017915309446254, | |
| "loss": 0.595, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.1883580450302031, | |
| "grad_norm": 0.049294911324977875, | |
| "learning_rate": 0.00020154723127035828, | |
| "loss": 0.6437, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 1.1905546403075233, | |
| "grad_norm": 0.03713231161236763, | |
| "learning_rate": 0.00020130293159609117, | |
| "loss": 0.5496, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 1.1927512355848435, | |
| "grad_norm": 0.04774945229291916, | |
| "learning_rate": 0.00020105863192182408, | |
| "loss": 0.7191, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 1.1949478308621637, | |
| "grad_norm": 0.04245544224977493, | |
| "learning_rate": 0.00020081433224755697, | |
| "loss": 0.6713, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 1.197144426139484, | |
| "grad_norm": 0.05695553123950958, | |
| "learning_rate": 0.00020057003257328986, | |
| "loss": 0.7036, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 1.1993410214168039, | |
| "grad_norm": 0.043208975344896317, | |
| "learning_rate": 0.00020032573289902278, | |
| "loss": 0.6773, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 1.201537616694124, | |
| "grad_norm": 0.061872243881225586, | |
| "learning_rate": 0.00020008143322475567, | |
| "loss": 0.9214, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 1.2037342119714443, | |
| "grad_norm": 0.03670836240053177, | |
| "learning_rate": 0.00019983713355048856, | |
| "loss": 0.6221, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 1.2059308072487644, | |
| "grad_norm": 0.05437465012073517, | |
| "learning_rate": 0.00019959283387622147, | |
| "loss": 0.8771, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 1.2081274025260846, | |
| "grad_norm": 0.07304202020168304, | |
| "learning_rate": 0.00019934853420195436, | |
| "loss": 0.7336, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.2103239978034046, | |
| "grad_norm": 0.05822945386171341, | |
| "learning_rate": 0.00019910423452768728, | |
| "loss": 0.6365, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 1.2125205930807248, | |
| "grad_norm": 0.051955416798591614, | |
| "learning_rate": 0.00019885993485342017, | |
| "loss": 0.7081, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 1.214717188358045, | |
| "grad_norm": 0.04791713133454323, | |
| "learning_rate": 0.00019861563517915306, | |
| "loss": 0.7363, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 1.2169137836353652, | |
| "grad_norm": 0.038635943084955215, | |
| "learning_rate": 0.00019837133550488597, | |
| "loss": 0.677, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 1.2191103789126854, | |
| "grad_norm": 0.07108104974031448, | |
| "learning_rate": 0.00019812703583061886, | |
| "loss": 0.7667, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 1.2213069741900056, | |
| "grad_norm": 0.050478167831897736, | |
| "learning_rate": 0.00019788273615635175, | |
| "loss": 0.5538, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 1.2235035694673257, | |
| "grad_norm": 0.050831057131290436, | |
| "learning_rate": 0.00019763843648208467, | |
| "loss": 0.6641, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 1.2257001647446457, | |
| "grad_norm": 0.058054275810718536, | |
| "learning_rate": 0.00019739413680781756, | |
| "loss": 0.6325, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 1.227896760021966, | |
| "grad_norm": 0.085708387196064, | |
| "learning_rate": 0.00019714983713355045, | |
| "loss": 0.8618, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 1.230093355299286, | |
| "grad_norm": 0.04360990226268768, | |
| "learning_rate": 0.00019690553745928336, | |
| "loss": 0.7238, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.2322899505766063, | |
| "grad_norm": 0.07026379555463791, | |
| "learning_rate": 0.00019666123778501625, | |
| "loss": 0.6265, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 1.2344865458539265, | |
| "grad_norm": 0.05597195774316788, | |
| "learning_rate": 0.00019641693811074917, | |
| "loss": 0.6385, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 1.2366831411312464, | |
| "grad_norm": 0.06764718890190125, | |
| "learning_rate": 0.00019617263843648206, | |
| "loss": 0.8007, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 1.2388797364085666, | |
| "grad_norm": 0.05328952148556709, | |
| "learning_rate": 0.00019592833876221495, | |
| "loss": 0.5902, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 1.2410763316858868, | |
| "grad_norm": 0.045794982463121414, | |
| "learning_rate": 0.00019568403908794786, | |
| "loss": 0.6765, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 1.243272926963207, | |
| "grad_norm": 0.09876284003257751, | |
| "learning_rate": 0.00019543973941368075, | |
| "loss": 0.6767, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 1.2454695222405272, | |
| "grad_norm": 0.04325791075825691, | |
| "learning_rate": 0.00019519543973941364, | |
| "loss": 0.6679, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 1.2476661175178474, | |
| "grad_norm": 0.04940588399767876, | |
| "learning_rate": 0.00019495114006514656, | |
| "loss": 0.6811, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 1.2498627127951676, | |
| "grad_norm": 0.04802437499165535, | |
| "learning_rate": 0.00019470684039087945, | |
| "loss": 0.5913, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 1.2520593080724876, | |
| "grad_norm": 0.04493922367691994, | |
| "learning_rate": 0.00019446254071661234, | |
| "loss": 0.472, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.2542559033498077, | |
| "grad_norm": 0.05015527084469795, | |
| "learning_rate": 0.00019421824104234525, | |
| "loss": 0.791, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 1.256452498627128, | |
| "grad_norm": 0.061855364590883255, | |
| "learning_rate": 0.00019397394136807814, | |
| "loss": 0.6994, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 1.2586490939044481, | |
| "grad_norm": 0.05905802920460701, | |
| "learning_rate": 0.00019372964169381103, | |
| "loss": 0.6168, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 1.2608456891817683, | |
| "grad_norm": 0.04185137152671814, | |
| "learning_rate": 0.00019348534201954395, | |
| "loss": 0.7272, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 1.2630422844590883, | |
| "grad_norm": 0.04353105649352074, | |
| "learning_rate": 0.00019324104234527684, | |
| "loss": 0.7509, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.2652388797364087, | |
| "grad_norm": 0.0427667535841465, | |
| "learning_rate": 0.00019299674267100975, | |
| "loss": 0.6093, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 1.2674354750137287, | |
| "grad_norm": 0.05642359331250191, | |
| "learning_rate": 0.00019275244299674264, | |
| "loss": 0.8424, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 1.2696320702910489, | |
| "grad_norm": 0.04383740574121475, | |
| "learning_rate": 0.00019250814332247553, | |
| "loss": 0.4776, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 1.271828665568369, | |
| "grad_norm": 0.048313695937395096, | |
| "learning_rate": 0.00019226384364820845, | |
| "loss": 0.5038, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 1.2740252608456892, | |
| "grad_norm": 0.038794662803411484, | |
| "learning_rate": 0.00019201954397394134, | |
| "loss": 0.5157, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.2762218561230094, | |
| "grad_norm": 0.04513511061668396, | |
| "learning_rate": 0.00019177524429967423, | |
| "loss": 0.678, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 1.2784184514003294, | |
| "grad_norm": 0.05818026885390282, | |
| "learning_rate": 0.00019153094462540714, | |
| "loss": 0.6184, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 1.2806150466776496, | |
| "grad_norm": 0.06140175834298134, | |
| "learning_rate": 0.00019128664495114003, | |
| "loss": 0.5967, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 1.2828116419549698, | |
| "grad_norm": 0.0904388502240181, | |
| "learning_rate": 0.00019104234527687292, | |
| "loss": 0.5606, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 1.28500823723229, | |
| "grad_norm": 0.06066805124282837, | |
| "learning_rate": 0.00019079804560260584, | |
| "loss": 0.7383, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 1.2872048325096102, | |
| "grad_norm": 0.04117051884531975, | |
| "learning_rate": 0.00019055374592833873, | |
| "loss": 0.6536, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 1.2894014277869301, | |
| "grad_norm": 0.04647667706012726, | |
| "learning_rate": 0.00019030944625407164, | |
| "loss": 0.6158, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 1.2915980230642505, | |
| "grad_norm": 0.05287046730518341, | |
| "learning_rate": 0.00019006514657980453, | |
| "loss": 0.6545, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 1.2937946183415705, | |
| "grad_norm": 0.05430927500128746, | |
| "learning_rate": 0.00018982084690553742, | |
| "loss": 0.6284, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 1.2959912136188907, | |
| "grad_norm": 0.04469500109553337, | |
| "learning_rate": 0.00018957654723127034, | |
| "loss": 0.6741, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.2981878088962109, | |
| "grad_norm": 0.030649427324533463, | |
| "learning_rate": 0.00018933224755700323, | |
| "loss": 0.5046, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 1.300384404173531, | |
| "grad_norm": 0.046733759343624115, | |
| "learning_rate": 0.00018908794788273612, | |
| "loss": 0.6558, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 1.3025809994508513, | |
| "grad_norm": 0.040193550288677216, | |
| "learning_rate": 0.00018884364820846903, | |
| "loss": 0.569, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 1.3047775947281712, | |
| "grad_norm": 0.03978565335273743, | |
| "learning_rate": 0.00018859934853420192, | |
| "loss": 0.6768, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 1.3069741900054914, | |
| "grad_norm": 0.04483840614557266, | |
| "learning_rate": 0.0001883550488599348, | |
| "loss": 0.7312, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 1.3091707852828116, | |
| "grad_norm": 0.0462617352604866, | |
| "learning_rate": 0.00018811074918566773, | |
| "loss": 0.5968, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 1.3113673805601318, | |
| "grad_norm": 0.048313964158296585, | |
| "learning_rate": 0.00018786644951140062, | |
| "loss": 0.6074, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 1.313563975837452, | |
| "grad_norm": 0.04513964429497719, | |
| "learning_rate": 0.0001876221498371335, | |
| "loss": 0.5612, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 1.3157605711147722, | |
| "grad_norm": 0.04602311924099922, | |
| "learning_rate": 0.00018737785016286645, | |
| "loss": 0.6196, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 1.3179571663920924, | |
| "grad_norm": 0.03924334794282913, | |
| "learning_rate": 0.00018713355048859934, | |
| "loss": 0.7697, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.3201537616694123, | |
| "grad_norm": 0.046174440532922745, | |
| "learning_rate": 0.00018688925081433225, | |
| "loss": 0.6409, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 1.3223503569467325, | |
| "grad_norm": 0.04037371277809143, | |
| "learning_rate": 0.00018664495114006514, | |
| "loss": 0.6609, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 1.3245469522240527, | |
| "grad_norm": 0.042371075600385666, | |
| "learning_rate": 0.00018640065146579803, | |
| "loss": 0.6816, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 1.326743547501373, | |
| "grad_norm": 0.04608074575662613, | |
| "learning_rate": 0.00018615635179153095, | |
| "loss": 0.6682, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 1.328940142778693, | |
| "grad_norm": 0.046598054468631744, | |
| "learning_rate": 0.00018591205211726384, | |
| "loss": 0.7451, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 1.331136738056013, | |
| "grad_norm": 0.0409809872508049, | |
| "learning_rate": 0.00018566775244299675, | |
| "loss": 0.5852, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 1.3333333333333333, | |
| "grad_norm": 0.051450323313474655, | |
| "learning_rate": 0.00018542345276872964, | |
| "loss": 0.5106, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 1.3355299286106534, | |
| "grad_norm": 0.043831150978803635, | |
| "learning_rate": 0.00018517915309446253, | |
| "loss": 0.6222, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 1.3377265238879736, | |
| "grad_norm": 0.0562346987426281, | |
| "learning_rate": 0.00018493485342019545, | |
| "loss": 0.6442, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 1.3399231191652938, | |
| "grad_norm": 0.04941529780626297, | |
| "learning_rate": 0.00018469055374592834, | |
| "loss": 0.6927, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.342119714442614, | |
| "grad_norm": 0.051082007586956024, | |
| "learning_rate": 0.00018444625407166123, | |
| "loss": 0.7017, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 1.3443163097199342, | |
| "grad_norm": 0.046207696199417114, | |
| "learning_rate": 0.00018420195439739414, | |
| "loss": 0.5908, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 1.3465129049972542, | |
| "grad_norm": 0.039629098027944565, | |
| "learning_rate": 0.00018395765472312703, | |
| "loss": 0.7068, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 1.3487095002745744, | |
| "grad_norm": 0.045384980738162994, | |
| "learning_rate": 0.00018371335504885992, | |
| "loss": 0.5844, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 1.3509060955518946, | |
| "grad_norm": 0.05367980897426605, | |
| "learning_rate": 0.00018346905537459284, | |
| "loss": 0.7222, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 1.3531026908292147, | |
| "grad_norm": 0.061644140630960464, | |
| "learning_rate": 0.00018322475570032573, | |
| "loss": 0.7469, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 1.355299286106535, | |
| "grad_norm": 0.04570198804140091, | |
| "learning_rate": 0.00018298045602605864, | |
| "loss": 0.6605, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 1.357495881383855, | |
| "grad_norm": 0.05299101769924164, | |
| "learning_rate": 0.00018273615635179153, | |
| "loss": 0.8565, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 1.359692476661175, | |
| "grad_norm": 0.043971553444862366, | |
| "learning_rate": 0.00018249185667752442, | |
| "loss": 0.509, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 1.3618890719384953, | |
| "grad_norm": 0.04689047113060951, | |
| "learning_rate": 0.00018224755700325734, | |
| "loss": 0.5978, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.3640856672158155, | |
| "grad_norm": 0.0547180213034153, | |
| "learning_rate": 0.00018200325732899023, | |
| "loss": 0.8353, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 1.3662822624931357, | |
| "grad_norm": 0.045855265110731125, | |
| "learning_rate": 0.00018175895765472312, | |
| "loss": 0.6055, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 1.3684788577704559, | |
| "grad_norm": 0.043083783239126205, | |
| "learning_rate": 0.00018151465798045603, | |
| "loss": 0.5349, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 1.370675453047776, | |
| "grad_norm": 0.056552454829216, | |
| "learning_rate": 0.00018127035830618892, | |
| "loss": 0.6284, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 1.372872048325096, | |
| "grad_norm": 0.05099363625049591, | |
| "learning_rate": 0.0001810260586319218, | |
| "loss": 0.7383, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.3750686436024162, | |
| "grad_norm": 0.039746589958667755, | |
| "learning_rate": 0.00018078175895765473, | |
| "loss": 0.5563, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 1.3772652388797364, | |
| "grad_norm": 0.03771736100316048, | |
| "learning_rate": 0.00018053745928338762, | |
| "loss": 0.6183, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 1.3794618341570566, | |
| "grad_norm": 0.06937503814697266, | |
| "learning_rate": 0.0001802931596091205, | |
| "loss": 0.4919, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 1.3816584294343768, | |
| "grad_norm": 0.04858769476413727, | |
| "learning_rate": 0.00018004885993485342, | |
| "loss": 0.6382, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 1.3838550247116967, | |
| "grad_norm": 0.04866888001561165, | |
| "learning_rate": 0.0001798045602605863, | |
| "loss": 0.6089, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.3860516199890172, | |
| "grad_norm": 0.04268645867705345, | |
| "learning_rate": 0.00017956026058631923, | |
| "loss": 0.63, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 1.3882482152663371, | |
| "grad_norm": 0.04429534077644348, | |
| "learning_rate": 0.00017931596091205212, | |
| "loss": 0.7758, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 1.3904448105436573, | |
| "grad_norm": 0.04046213626861572, | |
| "learning_rate": 0.000179071661237785, | |
| "loss": 0.6882, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 1.3926414058209775, | |
| "grad_norm": 0.049879636615514755, | |
| "learning_rate": 0.00017882736156351792, | |
| "loss": 0.5664, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 1.3948380010982977, | |
| "grad_norm": 0.051655709743499756, | |
| "learning_rate": 0.0001785830618892508, | |
| "loss": 0.7156, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.3970345963756179, | |
| "grad_norm": 0.03926423192024231, | |
| "learning_rate": 0.0001783387622149837, | |
| "loss": 0.7299, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 1.3992311916529379, | |
| "grad_norm": 0.059483956545591354, | |
| "learning_rate": 0.00017809446254071662, | |
| "loss": 0.5872, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 1.401427786930258, | |
| "grad_norm": 0.04855528101325035, | |
| "learning_rate": 0.0001778501628664495, | |
| "loss": 0.5601, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 1.4036243822075782, | |
| "grad_norm": 0.048911917954683304, | |
| "learning_rate": 0.0001776058631921824, | |
| "loss": 0.5874, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 1.4058209774848984, | |
| "grad_norm": 0.04301845282316208, | |
| "learning_rate": 0.0001773615635179153, | |
| "loss": 0.8704, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.4080175727622186, | |
| "grad_norm": 0.043345287442207336, | |
| "learning_rate": 0.0001771172638436482, | |
| "loss": 0.6319, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 1.4102141680395386, | |
| "grad_norm": 0.04333937540650368, | |
| "learning_rate": 0.0001768729641693811, | |
| "loss": 0.7228, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 1.412410763316859, | |
| "grad_norm": 0.20310437679290771, | |
| "learning_rate": 0.000176628664495114, | |
| "loss": 0.5866, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 1.414607358594179, | |
| "grad_norm": 0.04585960507392883, | |
| "learning_rate": 0.0001763843648208469, | |
| "loss": 0.5515, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 1.4168039538714992, | |
| "grad_norm": 0.06021832302212715, | |
| "learning_rate": 0.0001761400651465798, | |
| "loss": 0.575, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 1.4190005491488193, | |
| "grad_norm": 0.04396074265241623, | |
| "learning_rate": 0.0001758957654723127, | |
| "loss": 0.5602, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 1.4211971444261395, | |
| "grad_norm": 0.05046122521162033, | |
| "learning_rate": 0.0001756514657980456, | |
| "loss": 0.6205, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 1.4233937397034597, | |
| "grad_norm": 0.045648425817489624, | |
| "learning_rate": 0.0001754071661237785, | |
| "loss": 0.5896, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 1.4255903349807797, | |
| "grad_norm": 0.06430377811193466, | |
| "learning_rate": 0.0001751628664495114, | |
| "loss": 0.7242, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 1.4277869302580999, | |
| "grad_norm": 0.04144120216369629, | |
| "learning_rate": 0.00017491856677524429, | |
| "loss": 0.5391, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.42998352553542, | |
| "grad_norm": 0.037751659750938416, | |
| "learning_rate": 0.0001746742671009772, | |
| "loss": 0.6035, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 1.4321801208127403, | |
| "grad_norm": 0.04347790405154228, | |
| "learning_rate": 0.0001744299674267101, | |
| "loss": 0.5921, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 1.4343767160900605, | |
| "grad_norm": 0.04341914877295494, | |
| "learning_rate": 0.00017418566775244298, | |
| "loss": 0.6021, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 1.4365733113673804, | |
| "grad_norm": 0.04380672052502632, | |
| "learning_rate": 0.0001739413680781759, | |
| "loss": 0.57, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 1.4387699066447008, | |
| "grad_norm": 0.05512508749961853, | |
| "learning_rate": 0.00017369706840390879, | |
| "loss": 0.6191, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 1.4409665019220208, | |
| "grad_norm": 0.05672155320644379, | |
| "learning_rate": 0.0001734527687296417, | |
| "loss": 0.7185, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 1.443163097199341, | |
| "grad_norm": 0.04579592123627663, | |
| "learning_rate": 0.0001732084690553746, | |
| "loss": 0.7705, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 1.4453596924766612, | |
| "grad_norm": 0.04407493770122528, | |
| "learning_rate": 0.00017296416938110748, | |
| "loss": 0.7518, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 1.4475562877539814, | |
| "grad_norm": 0.05515037477016449, | |
| "learning_rate": 0.0001727198697068404, | |
| "loss": 0.8036, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 1.4497528830313016, | |
| "grad_norm": 0.04245726764202118, | |
| "learning_rate": 0.00017247557003257329, | |
| "loss": 0.5768, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.4519494783086215, | |
| "grad_norm": 0.03963717445731163, | |
| "learning_rate": 0.00017223127035830618, | |
| "loss": 0.5828, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 1.4541460735859417, | |
| "grad_norm": 0.05516964569687843, | |
| "learning_rate": 0.0001719869706840391, | |
| "loss": 0.6889, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 1.456342668863262, | |
| "grad_norm": 0.0535304956138134, | |
| "learning_rate": 0.00017174267100977198, | |
| "loss": 0.6561, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 1.458539264140582, | |
| "grad_norm": 0.03565557673573494, | |
| "learning_rate": 0.00017149837133550487, | |
| "loss": 0.5585, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 1.4607358594179023, | |
| "grad_norm": 0.0661967322230339, | |
| "learning_rate": 0.00017125407166123779, | |
| "loss": 0.8141, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 1.4629324546952225, | |
| "grad_norm": 0.060851022601127625, | |
| "learning_rate": 0.00017100977198697068, | |
| "loss": 0.7993, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 1.4651290499725427, | |
| "grad_norm": 0.055130865424871445, | |
| "learning_rate": 0.00017076547231270356, | |
| "loss": 0.6111, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 1.4673256452498626, | |
| "grad_norm": 0.044011928141117096, | |
| "learning_rate": 0.00017052117263843648, | |
| "loss": 0.7623, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 1.4695222405271828, | |
| "grad_norm": 0.04537820816040039, | |
| "learning_rate": 0.00017027687296416937, | |
| "loss": 0.6055, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 1.471718835804503, | |
| "grad_norm": 0.045265693217515945, | |
| "learning_rate": 0.00017003257328990229, | |
| "loss": 0.6416, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.4739154310818232, | |
| "grad_norm": 0.07514453679323196, | |
| "learning_rate": 0.00016978827361563518, | |
| "loss": 0.7886, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 1.4761120263591434, | |
| "grad_norm": 0.041896089911460876, | |
| "learning_rate": 0.00016954397394136806, | |
| "loss": 0.6698, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 1.4783086216364634, | |
| "grad_norm": 0.050291791558265686, | |
| "learning_rate": 0.00016929967426710098, | |
| "loss": 0.6528, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 1.4805052169137836, | |
| "grad_norm": 0.037543244659900665, | |
| "learning_rate": 0.00016905537459283387, | |
| "loss": 0.6816, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 1.4827018121911038, | |
| "grad_norm": 0.06479065120220184, | |
| "learning_rate": 0.00016881107491856676, | |
| "loss": 0.5854, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.484898407468424, | |
| "grad_norm": 0.0435757115483284, | |
| "learning_rate": 0.00016856677524429968, | |
| "loss": 0.6497, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 1.4870950027457441, | |
| "grad_norm": 0.04248249530792236, | |
| "learning_rate": 0.00016832247557003257, | |
| "loss": 0.6321, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 1.4892915980230643, | |
| "grad_norm": 0.056113965809345245, | |
| "learning_rate": 0.00016807817589576545, | |
| "loss": 0.7081, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 1.4914881933003845, | |
| "grad_norm": 0.04325173795223236, | |
| "learning_rate": 0.00016783387622149837, | |
| "loss": 0.6985, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 1.4936847885777045, | |
| "grad_norm": 0.10305589437484741, | |
| "learning_rate": 0.00016758957654723126, | |
| "loss": 0.8552, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.4958813838550247, | |
| "grad_norm": 0.04724079370498657, | |
| "learning_rate": 0.00016734527687296415, | |
| "loss": 0.6303, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 1.4980779791323449, | |
| "grad_norm": 0.05535902827978134, | |
| "learning_rate": 0.00016710097719869707, | |
| "loss": 0.8582, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 1.500274574409665, | |
| "grad_norm": 0.04309092089533806, | |
| "learning_rate": 0.00016685667752442995, | |
| "loss": 0.4383, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 1.5024711696869852, | |
| "grad_norm": 0.049390096217393875, | |
| "learning_rate": 0.00016661237785016287, | |
| "loss": 0.5833, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 1.5046677649643052, | |
| "grad_norm": 0.0520150363445282, | |
| "learning_rate": 0.00016636807817589576, | |
| "loss": 0.7291, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 1.5068643602416256, | |
| "grad_norm": 0.07185570150613785, | |
| "learning_rate": 0.00016612377850162865, | |
| "loss": 0.8213, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 1.5090609555189456, | |
| "grad_norm": 0.06040317192673683, | |
| "learning_rate": 0.00016587947882736157, | |
| "loss": 0.6683, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 1.5112575507962658, | |
| "grad_norm": 0.050129249691963196, | |
| "learning_rate": 0.00016563517915309445, | |
| "loss": 0.7417, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 1.513454146073586, | |
| "grad_norm": 0.057126522064208984, | |
| "learning_rate": 0.00016539087947882734, | |
| "loss": 0.5871, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 1.515650741350906, | |
| "grad_norm": 0.037670109421014786, | |
| "learning_rate": 0.00016514657980456026, | |
| "loss": 0.6136, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.5178473366282264, | |
| "grad_norm": 0.048299022018909454, | |
| "learning_rate": 0.00016490228013029315, | |
| "loss": 0.6504, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 1.5200439319055463, | |
| "grad_norm": 0.04665425419807434, | |
| "learning_rate": 0.00016465798045602604, | |
| "loss": 0.6824, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 1.5222405271828665, | |
| "grad_norm": 0.052132029086351395, | |
| "learning_rate": 0.00016441368078175895, | |
| "loss": 0.678, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 1.5244371224601867, | |
| "grad_norm": 0.05157755687832832, | |
| "learning_rate": 0.00016416938110749184, | |
| "loss": 0.5837, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 1.526633717737507, | |
| "grad_norm": 0.04299292340874672, | |
| "learning_rate": 0.00016392508143322476, | |
| "loss": 0.6819, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 1.528830313014827, | |
| "grad_norm": 0.04690634086728096, | |
| "learning_rate": 0.00016368078175895765, | |
| "loss": 0.6123, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 1.531026908292147, | |
| "grad_norm": 0.04270879551768303, | |
| "learning_rate": 0.00016343648208469054, | |
| "loss": 0.6294, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 1.5332235035694675, | |
| "grad_norm": 0.038513366132974625, | |
| "learning_rate": 0.00016319218241042346, | |
| "loss": 0.6058, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 1.5354200988467874, | |
| "grad_norm": 0.05122312158346176, | |
| "learning_rate": 0.00016294788273615634, | |
| "loss": 0.6818, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 1.5376166941241076, | |
| "grad_norm": 0.038340240716934204, | |
| "learning_rate": 0.00016270358306188923, | |
| "loss": 0.7551, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.5398132894014278, | |
| "grad_norm": 0.04914093762636185, | |
| "learning_rate": 0.00016245928338762215, | |
| "loss": 0.5806, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 1.5420098846787478, | |
| "grad_norm": 0.04841599613428116, | |
| "learning_rate": 0.00016221498371335504, | |
| "loss": 0.6435, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 1.5442064799560682, | |
| "grad_norm": 0.03922514617443085, | |
| "learning_rate": 0.00016197068403908793, | |
| "loss": 0.5564, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 1.5464030752333882, | |
| "grad_norm": 0.06409287452697754, | |
| "learning_rate": 0.00016172638436482084, | |
| "loss": 0.5631, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 1.5485996705107083, | |
| "grad_norm": 0.07203608751296997, | |
| "learning_rate": 0.00016148208469055373, | |
| "loss": 0.6606, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 1.5507962657880285, | |
| "grad_norm": 0.048426203429698944, | |
| "learning_rate": 0.00016123778501628662, | |
| "loss": 0.6252, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 1.5529928610653487, | |
| "grad_norm": 0.054609306156635284, | |
| "learning_rate": 0.00016099348534201954, | |
| "loss": 0.6719, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 1.555189456342669, | |
| "grad_norm": 0.08309051394462585, | |
| "learning_rate": 0.00016074918566775243, | |
| "loss": 0.7236, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 1.5573860516199889, | |
| "grad_norm": 0.06068706884980202, | |
| "learning_rate": 0.00016050488599348534, | |
| "loss": 0.7964, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 1.5595826468973093, | |
| "grad_norm": 0.05096975341439247, | |
| "learning_rate": 0.00016026058631921823, | |
| "loss": 0.4562, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.5617792421746293, | |
| "grad_norm": 0.04532795771956444, | |
| "learning_rate": 0.00016001628664495112, | |
| "loss": 0.7656, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 1.5639758374519495, | |
| "grad_norm": 0.046231381595134735, | |
| "learning_rate": 0.00015977198697068404, | |
| "loss": 0.5014, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 1.5661724327292696, | |
| "grad_norm": 0.04625224322080612, | |
| "learning_rate": 0.00015952768729641693, | |
| "loss": 0.5836, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 1.5683690280065898, | |
| "grad_norm": 0.03278486803174019, | |
| "learning_rate": 0.00015928338762214982, | |
| "loss": 0.5181, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 1.57056562328391, | |
| "grad_norm": 0.0495121143758297, | |
| "learning_rate": 0.00015903908794788273, | |
| "loss": 0.6985, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 1.57276221856123, | |
| "grad_norm": 0.06934545189142227, | |
| "learning_rate": 0.00015879478827361562, | |
| "loss": 0.8638, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 1.5749588138385504, | |
| "grad_norm": 0.04187803342938423, | |
| "learning_rate": 0.0001585504885993485, | |
| "loss": 0.6577, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 1.5771554091158704, | |
| "grad_norm": 0.07038550078868866, | |
| "learning_rate": 0.00015830618892508143, | |
| "loss": 0.593, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 1.5793520043931906, | |
| "grad_norm": 0.05784309282898903, | |
| "learning_rate": 0.00015806188925081432, | |
| "loss": 0.6606, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 1.5815485996705108, | |
| "grad_norm": 0.050657592713832855, | |
| "learning_rate": 0.00015781758957654723, | |
| "loss": 0.5663, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.5837451949478307, | |
| "grad_norm": 0.055596090853214264, | |
| "learning_rate": 0.00015757328990228012, | |
| "loss": 0.7576, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 1.5859417902251511, | |
| "grad_norm": 0.06155412271618843, | |
| "learning_rate": 0.000157328990228013, | |
| "loss": 0.8413, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 1.588138385502471, | |
| "grad_norm": 0.04519607871770859, | |
| "learning_rate": 0.00015708469055374593, | |
| "loss": 0.6418, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 1.5903349807797913, | |
| "grad_norm": 0.049259744584560394, | |
| "learning_rate": 0.00015684039087947882, | |
| "loss": 0.5145, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 1.5925315760571115, | |
| "grad_norm": 0.04998054727911949, | |
| "learning_rate": 0.0001565960912052117, | |
| "loss": 0.7043, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 1.5947281713344317, | |
| "grad_norm": 0.0548371784389019, | |
| "learning_rate": 0.00015635179153094462, | |
| "loss": 0.6797, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 1.5969247666117519, | |
| "grad_norm": 0.048441048711538315, | |
| "learning_rate": 0.0001561074918566775, | |
| "loss": 0.7109, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 1.5991213618890718, | |
| "grad_norm": 0.04135994240641594, | |
| "learning_rate": 0.0001558631921824104, | |
| "loss": 0.7135, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 1.6013179571663922, | |
| "grad_norm": 0.06847979873418808, | |
| "learning_rate": 0.00015561889250814332, | |
| "loss": 0.6801, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 1.6035145524437122, | |
| "grad_norm": 0.050964195281267166, | |
| "learning_rate": 0.0001553745928338762, | |
| "loss": 0.6747, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.6057111477210324, | |
| "grad_norm": 0.04272550716996193, | |
| "learning_rate": 0.0001551302931596091, | |
| "loss": 0.4606, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 1.6079077429983526, | |
| "grad_norm": 0.05093299224972725, | |
| "learning_rate": 0.000154885993485342, | |
| "loss": 0.6401, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 1.6101043382756726, | |
| "grad_norm": 0.07038379460573196, | |
| "learning_rate": 0.0001546416938110749, | |
| "loss": 0.8044, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 1.612300933552993, | |
| "grad_norm": 0.04339474439620972, | |
| "learning_rate": 0.00015439739413680782, | |
| "loss": 0.6374, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 1.614497528830313, | |
| "grad_norm": 0.08015977591276169, | |
| "learning_rate": 0.0001541530944625407, | |
| "loss": 0.7327, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 1.6166941241076331, | |
| "grad_norm": 0.04767515882849693, | |
| "learning_rate": 0.0001539087947882736, | |
| "loss": 0.6442, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 1.6188907193849533, | |
| "grad_norm": 0.04954354092478752, | |
| "learning_rate": 0.0001536644951140065, | |
| "loss": 0.7242, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 1.6210873146622735, | |
| "grad_norm": 0.06409866362810135, | |
| "learning_rate": 0.0001534201954397394, | |
| "loss": 0.6139, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 1.6232839099395937, | |
| "grad_norm": 0.04829477518796921, | |
| "learning_rate": 0.0001531758957654723, | |
| "loss": 0.6117, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 1.6254805052169137, | |
| "grad_norm": 0.05684004724025726, | |
| "learning_rate": 0.0001529315960912052, | |
| "loss": 0.6686, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.627677100494234, | |
| "grad_norm": 0.040473535656929016, | |
| "learning_rate": 0.0001526872964169381, | |
| "loss": 0.7073, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 1.629873695771554, | |
| "grad_norm": 0.05740467086434364, | |
| "learning_rate": 0.00015244299674267099, | |
| "loss": 0.6241, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 1.6320702910488742, | |
| "grad_norm": 0.06075328588485718, | |
| "learning_rate": 0.0001521986970684039, | |
| "loss": 0.587, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 1.6342668863261944, | |
| "grad_norm": 0.05512019619345665, | |
| "learning_rate": 0.0001519543973941368, | |
| "loss": 0.503, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 1.6364634816035144, | |
| "grad_norm": 0.06742412596940994, | |
| "learning_rate": 0.00015171009771986968, | |
| "loss": 0.6021, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 1.6386600768808348, | |
| "grad_norm": 0.0417785570025444, | |
| "learning_rate": 0.0001514657980456026, | |
| "loss": 0.4437, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 1.6408566721581548, | |
| "grad_norm": 0.08356238156557083, | |
| "learning_rate": 0.0001512214983713355, | |
| "loss": 0.6415, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 1.643053267435475, | |
| "grad_norm": 0.04881569743156433, | |
| "learning_rate": 0.0001509771986970684, | |
| "loss": 0.4988, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 1.6452498627127952, | |
| "grad_norm": 0.047157756984233856, | |
| "learning_rate": 0.0001507328990228013, | |
| "loss": 0.502, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 1.6474464579901154, | |
| "grad_norm": 0.05139964818954468, | |
| "learning_rate": 0.00015048859934853418, | |
| "loss": 0.591, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.6496430532674355, | |
| "grad_norm": 0.0414644293487072, | |
| "learning_rate": 0.0001502442996742671, | |
| "loss": 0.6039, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 1.6518396485447555, | |
| "grad_norm": 0.06458733975887299, | |
| "learning_rate": 0.00015, | |
| "loss": 0.732, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 1.654036243822076, | |
| "grad_norm": 0.07901707291603088, | |
| "learning_rate": 0.00014975570032573288, | |
| "loss": 0.9098, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 1.656232839099396, | |
| "grad_norm": 0.047911275178194046, | |
| "learning_rate": 0.0001495114006514658, | |
| "loss": 0.5262, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 1.658429434376716, | |
| "grad_norm": 0.05376958101987839, | |
| "learning_rate": 0.00014926710097719868, | |
| "loss": 0.6248, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 1.6606260296540363, | |
| "grad_norm": 0.03677194193005562, | |
| "learning_rate": 0.00014902280130293157, | |
| "loss": 0.6117, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 1.6628226249313562, | |
| "grad_norm": 0.05050716921687126, | |
| "learning_rate": 0.0001487785016286645, | |
| "loss": 0.7372, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 1.6650192202086767, | |
| "grad_norm": 0.050005607306957245, | |
| "learning_rate": 0.00014853420195439738, | |
| "loss": 0.7221, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 1.6672158154859966, | |
| "grad_norm": 0.04903872683644295, | |
| "learning_rate": 0.0001482899022801303, | |
| "loss": 0.7553, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 1.6694124107633168, | |
| "grad_norm": 0.06483814120292664, | |
| "learning_rate": 0.00014804560260586318, | |
| "loss": 0.5945, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.671609006040637, | |
| "grad_norm": 0.04501137509942055, | |
| "learning_rate": 0.00014780130293159607, | |
| "loss": 0.6599, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 1.6738056013179572, | |
| "grad_norm": 0.04700218141078949, | |
| "learning_rate": 0.000147557003257329, | |
| "loss": 0.6844, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 1.6760021965952774, | |
| "grad_norm": 0.04598642885684967, | |
| "learning_rate": 0.00014731270358306188, | |
| "loss": 0.662, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 1.6781987918725974, | |
| "grad_norm": 0.0409390851855278, | |
| "learning_rate": 0.00014706840390879477, | |
| "loss": 0.7319, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 1.6803953871499178, | |
| "grad_norm": 0.06088387221097946, | |
| "learning_rate": 0.00014682410423452768, | |
| "loss": 0.6724, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 1.6825919824272377, | |
| "grad_norm": 0.04854418337345123, | |
| "learning_rate": 0.00014657980456026057, | |
| "loss": 0.6068, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 1.684788577704558, | |
| "grad_norm": 0.049988895654678345, | |
| "learning_rate": 0.00014633550488599346, | |
| "loss": 0.6674, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 1.6869851729818781, | |
| "grad_norm": 0.07750127464532852, | |
| "learning_rate": 0.00014609120521172638, | |
| "loss": 0.6047, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 1.689181768259198, | |
| "grad_norm": 0.04291775822639465, | |
| "learning_rate": 0.00014584690553745927, | |
| "loss": 0.6315, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 1.6913783635365185, | |
| "grad_norm": 0.04422176256775856, | |
| "learning_rate": 0.00014560260586319216, | |
| "loss": 0.6874, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.6935749588138385, | |
| "grad_norm": 0.049668315798044205, | |
| "learning_rate": 0.00014535830618892507, | |
| "loss": 0.8245, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 1.6957715540911587, | |
| "grad_norm": 0.051815249025821686, | |
| "learning_rate": 0.00014511400651465796, | |
| "loss": 0.7336, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 1.6979681493684788, | |
| "grad_norm": 0.04109729453921318, | |
| "learning_rate": 0.00014486970684039088, | |
| "loss": 0.6031, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 1.700164744645799, | |
| "grad_norm": 0.0370013527572155, | |
| "learning_rate": 0.00014462540716612377, | |
| "loss": 0.5536, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 1.7023613399231192, | |
| "grad_norm": 0.052517328411340714, | |
| "learning_rate": 0.00014438110749185666, | |
| "loss": 0.6791, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 1.7045579352004392, | |
| "grad_norm": 0.04123876616358757, | |
| "learning_rate": 0.00014413680781758957, | |
| "loss": 0.4564, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 1.7067545304777596, | |
| "grad_norm": 0.049101028591394424, | |
| "learning_rate": 0.00014389250814332246, | |
| "loss": 0.73, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 1.7089511257550796, | |
| "grad_norm": 0.0485854297876358, | |
| "learning_rate": 0.00014364820846905535, | |
| "loss": 0.729, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 1.7111477210323998, | |
| "grad_norm": 0.0474361851811409, | |
| "learning_rate": 0.00014340390879478827, | |
| "loss": 0.6365, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 1.71334431630972, | |
| "grad_norm": 0.05722922831773758, | |
| "learning_rate": 0.00014315960912052116, | |
| "loss": 0.6634, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.7155409115870401, | |
| "grad_norm": 0.05596005171537399, | |
| "learning_rate": 0.00014291530944625404, | |
| "loss": 0.6379, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 1.7177375068643603, | |
| "grad_norm": 0.06430881470441818, | |
| "learning_rate": 0.00014267100977198696, | |
| "loss": 0.5975, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 1.7199341021416803, | |
| "grad_norm": 0.041797831654548645, | |
| "learning_rate": 0.00014242671009771985, | |
| "loss": 0.6402, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 1.7221306974190007, | |
| "grad_norm": 0.04201621934771538, | |
| "learning_rate": 0.00014218241042345274, | |
| "loss": 0.6083, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 1.7243272926963207, | |
| "grad_norm": 0.0454474575817585, | |
| "learning_rate": 0.00014193811074918566, | |
| "loss": 0.856, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 1.7265238879736409, | |
| "grad_norm": 0.03413139283657074, | |
| "learning_rate": 0.00014169381107491854, | |
| "loss": 0.5574, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 1.728720483250961, | |
| "grad_norm": 0.05276301130652428, | |
| "learning_rate": 0.00014144951140065146, | |
| "loss": 0.5608, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 1.730917078528281, | |
| "grad_norm": 0.04063792899250984, | |
| "learning_rate": 0.00014120521172638435, | |
| "loss": 0.648, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 1.7331136738056014, | |
| "grad_norm": 0.04240501672029495, | |
| "learning_rate": 0.00014096091205211724, | |
| "loss": 0.6826, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 1.7353102690829214, | |
| "grad_norm": 0.0462966114282608, | |
| "learning_rate": 0.00014071661237785016, | |
| "loss": 0.8253, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.7375068643602416, | |
| "grad_norm": 0.06632352620363235, | |
| "learning_rate": 0.00014047231270358305, | |
| "loss": 0.6522, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 1.7397034596375618, | |
| "grad_norm": 0.0381113663315773, | |
| "learning_rate": 0.00014022801302931593, | |
| "loss": 0.5937, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 1.741900054914882, | |
| "grad_norm": 0.056441642343997955, | |
| "learning_rate": 0.00013998371335504885, | |
| "loss": 0.61, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 1.7440966501922022, | |
| "grad_norm": 0.0543392077088356, | |
| "learning_rate": 0.00013973941368078174, | |
| "loss": 0.7435, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 1.7462932454695221, | |
| "grad_norm": 0.03827011212706566, | |
| "learning_rate": 0.00013949511400651463, | |
| "loss": 0.5941, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 1.7484898407468425, | |
| "grad_norm": 0.060338567942380905, | |
| "learning_rate": 0.00013925081433224755, | |
| "loss": 0.582, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 1.7506864360241625, | |
| "grad_norm": 0.03799464553594589, | |
| "learning_rate": 0.00013900651465798043, | |
| "loss": 0.6215, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 1.7528830313014827, | |
| "grad_norm": 0.043356601148843765, | |
| "learning_rate": 0.00013876221498371335, | |
| "loss": 0.6715, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 1.755079626578803, | |
| "grad_norm": 0.08416961133480072, | |
| "learning_rate": 0.00013851791530944624, | |
| "loss": 0.7502, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 1.7572762218561229, | |
| "grad_norm": 0.05539776757359505, | |
| "learning_rate": 0.00013827361563517913, | |
| "loss": 0.8666, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.7594728171334433, | |
| "grad_norm": 0.05246887728571892, | |
| "learning_rate": 0.00013802931596091205, | |
| "loss": 0.7886, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 1.7616694124107632, | |
| "grad_norm": 0.07970348745584488, | |
| "learning_rate": 0.00013778501628664493, | |
| "loss": 0.8519, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 1.7638660076880834, | |
| "grad_norm": 0.07714508473873138, | |
| "learning_rate": 0.00013754071661237782, | |
| "loss": 0.9883, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 1.7660626029654036, | |
| "grad_norm": 0.055044207721948624, | |
| "learning_rate": 0.00013729641693811074, | |
| "loss": 0.8199, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 1.7682591982427238, | |
| "grad_norm": 0.05030933395028114, | |
| "learning_rate": 0.00013705211726384363, | |
| "loss": 0.6976, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 1.770455793520044, | |
| "grad_norm": 0.048857420682907104, | |
| "learning_rate": 0.00013680781758957652, | |
| "loss": 0.6373, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 1.772652388797364, | |
| "grad_norm": 0.05215909704566002, | |
| "learning_rate": 0.00013656351791530943, | |
| "loss": 0.6757, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 1.7748489840746844, | |
| "grad_norm": 0.07715169340372086, | |
| "learning_rate": 0.00013631921824104232, | |
| "loss": 0.752, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 1.7770455793520044, | |
| "grad_norm": 0.058291271328926086, | |
| "learning_rate": 0.0001360749185667752, | |
| "loss": 0.7419, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 1.7792421746293245, | |
| "grad_norm": 0.05078558251261711, | |
| "learning_rate": 0.00013583061889250813, | |
| "loss": 0.4623, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.7814387699066447, | |
| "grad_norm": 0.0427047535777092, | |
| "learning_rate": 0.00013558631921824102, | |
| "loss": 0.536, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 1.7836353651839647, | |
| "grad_norm": 0.055373404175043106, | |
| "learning_rate": 0.00013534201954397394, | |
| "loss": 0.6415, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 1.7858319604612851, | |
| "grad_norm": 0.05837016552686691, | |
| "learning_rate": 0.00013509771986970682, | |
| "loss": 0.7953, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 1.788028555738605, | |
| "grad_norm": 0.05480782315135002, | |
| "learning_rate": 0.00013485342019543971, | |
| "loss": 0.6892, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 1.7902251510159253, | |
| "grad_norm": 0.04352695494890213, | |
| "learning_rate": 0.00013460912052117263, | |
| "loss": 0.6963, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 1.7924217462932455, | |
| "grad_norm": 0.048882078379392624, | |
| "learning_rate": 0.00013436482084690552, | |
| "loss": 0.6821, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 1.7946183415705657, | |
| "grad_norm": 0.04737727344036102, | |
| "learning_rate": 0.0001341205211726384, | |
| "loss": 0.5867, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 1.7968149368478858, | |
| "grad_norm": 0.048038333654403687, | |
| "learning_rate": 0.00013387622149837132, | |
| "loss": 0.6541, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 1.7990115321252058, | |
| "grad_norm": 0.04460786655545235, | |
| "learning_rate": 0.00013363192182410421, | |
| "loss": 0.6313, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 1.8012081274025262, | |
| "grad_norm": 0.05005588009953499, | |
| "learning_rate": 0.0001333876221498371, | |
| "loss": 0.6801, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.8034047226798462, | |
| "grad_norm": 0.04483172670006752, | |
| "learning_rate": 0.00013314332247557002, | |
| "loss": 0.7252, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 1.8056013179571664, | |
| "grad_norm": 0.05668759346008301, | |
| "learning_rate": 0.0001328990228013029, | |
| "loss": 0.7627, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 1.8077979132344866, | |
| "grad_norm": 0.04755188897252083, | |
| "learning_rate": 0.00013265472312703582, | |
| "loss": 0.6951, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 1.8099945085118065, | |
| "grad_norm": 0.04183987155556679, | |
| "learning_rate": 0.00013241042345276871, | |
| "loss": 0.6022, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 1.812191103789127, | |
| "grad_norm": 0.0463097020983696, | |
| "learning_rate": 0.0001321661237785016, | |
| "loss": 0.6062, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 1.814387699066447, | |
| "grad_norm": 0.04682566225528717, | |
| "learning_rate": 0.00013192182410423452, | |
| "loss": 0.4859, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 1.8165842943437671, | |
| "grad_norm": 0.05168147012591362, | |
| "learning_rate": 0.0001316775244299674, | |
| "loss": 0.7072, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 1.8187808896210873, | |
| "grad_norm": 0.04258091375231743, | |
| "learning_rate": 0.0001314332247557003, | |
| "loss": 0.6083, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 1.8209774848984075, | |
| "grad_norm": 0.04315861314535141, | |
| "learning_rate": 0.00013118892508143321, | |
| "loss": 0.5703, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 1.8231740801757277, | |
| "grad_norm": 0.05403318628668785, | |
| "learning_rate": 0.00013094462540716613, | |
| "loss": 0.5154, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.8253706754530477, | |
| "grad_norm": 0.03748399391770363, | |
| "learning_rate": 0.00013070032573289902, | |
| "loss": 0.4465, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 1.827567270730368, | |
| "grad_norm": 0.04764214903116226, | |
| "learning_rate": 0.0001304560260586319, | |
| "loss": 0.6787, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 1.829763866007688, | |
| "grad_norm": 0.0467853844165802, | |
| "learning_rate": 0.00013021172638436483, | |
| "loss": 0.6927, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 1.8319604612850082, | |
| "grad_norm": 0.049767203629016876, | |
| "learning_rate": 0.00012996742671009771, | |
| "loss": 0.7413, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 1.8341570565623284, | |
| "grad_norm": 0.049734946340322495, | |
| "learning_rate": 0.0001297231270358306, | |
| "loss": 0.7474, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 1.8363536518396484, | |
| "grad_norm": 0.05754285678267479, | |
| "learning_rate": 0.00012947882736156352, | |
| "loss": 0.5436, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 1.8385502471169688, | |
| "grad_norm": 0.08427241444587708, | |
| "learning_rate": 0.0001292345276872964, | |
| "loss": 0.6116, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 1.8407468423942888, | |
| "grad_norm": 0.051138222217559814, | |
| "learning_rate": 0.00012899022801302933, | |
| "loss": 0.6005, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 1.842943437671609, | |
| "grad_norm": 0.0441889688372612, | |
| "learning_rate": 0.00012874592833876221, | |
| "loss": 0.5388, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 1.8451400329489291, | |
| "grad_norm": 0.057013627141714096, | |
| "learning_rate": 0.0001285016286644951, | |
| "loss": 0.7047, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.8473366282262493, | |
| "grad_norm": 0.04341554641723633, | |
| "learning_rate": 0.00012825732899022802, | |
| "loss": 0.636, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 1.8495332235035695, | |
| "grad_norm": 0.05924821272492409, | |
| "learning_rate": 0.0001280130293159609, | |
| "loss": 0.6177, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 1.8517298187808895, | |
| "grad_norm": 0.05318041145801544, | |
| "learning_rate": 0.0001277687296416938, | |
| "loss": 0.8203, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 1.85392641405821, | |
| "grad_norm": 0.062347158789634705, | |
| "learning_rate": 0.00012752442996742671, | |
| "loss": 0.7046, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 1.8561230093355299, | |
| "grad_norm": 0.053158268332481384, | |
| "learning_rate": 0.0001272801302931596, | |
| "loss": 0.6758, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 1.85831960461285, | |
| "grad_norm": 0.04612778499722481, | |
| "learning_rate": 0.0001270358306188925, | |
| "loss": 0.5712, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 1.8605161998901703, | |
| "grad_norm": 0.05728829279541969, | |
| "learning_rate": 0.0001267915309446254, | |
| "loss": 0.7674, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 1.8627127951674904, | |
| "grad_norm": 0.046414975076913834, | |
| "learning_rate": 0.0001265472312703583, | |
| "loss": 0.5906, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 1.8649093904448106, | |
| "grad_norm": 0.04038268327713013, | |
| "learning_rate": 0.0001263029315960912, | |
| "loss": 0.5586, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 1.8671059857221306, | |
| "grad_norm": 0.03641192987561226, | |
| "learning_rate": 0.0001260586319218241, | |
| "loss": 0.6012, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.869302580999451, | |
| "grad_norm": 0.03718649223446846, | |
| "learning_rate": 0.000125814332247557, | |
| "loss": 0.5135, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 1.871499176276771, | |
| "grad_norm": 0.03840146213769913, | |
| "learning_rate": 0.0001255700325732899, | |
| "loss": 0.6627, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 1.8736957715540912, | |
| "grad_norm": 0.044655878096818924, | |
| "learning_rate": 0.0001253257328990228, | |
| "loss": 0.6727, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 1.8758923668314114, | |
| "grad_norm": 0.05583556368947029, | |
| "learning_rate": 0.0001250814332247557, | |
| "loss": 0.8199, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 1.8780889621087313, | |
| "grad_norm": 0.03336760401725769, | |
| "learning_rate": 0.0001248371335504886, | |
| "loss": 0.5403, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 1.8802855573860517, | |
| "grad_norm": 0.05149347707629204, | |
| "learning_rate": 0.0001245928338762215, | |
| "loss": 0.7947, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 1.8824821526633717, | |
| "grad_norm": 0.04355181008577347, | |
| "learning_rate": 0.00012434853420195438, | |
| "loss": 0.6353, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 1.884678747940692, | |
| "grad_norm": 0.04656535014510155, | |
| "learning_rate": 0.0001241042345276873, | |
| "loss": 0.7081, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 1.886875343218012, | |
| "grad_norm": 0.04054791107773781, | |
| "learning_rate": 0.0001238599348534202, | |
| "loss": 0.592, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 1.8890719384953323, | |
| "grad_norm": 0.034135375171899796, | |
| "learning_rate": 0.00012361563517915308, | |
| "loss": 0.5938, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.8912685337726525, | |
| "grad_norm": 0.075068399310112, | |
| "learning_rate": 0.000123371335504886, | |
| "loss": 0.8108, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 1.8934651290499724, | |
| "grad_norm": 0.04905351996421814, | |
| "learning_rate": 0.00012312703583061888, | |
| "loss": 0.657, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 1.8956617243272929, | |
| "grad_norm": 0.047971729189157486, | |
| "learning_rate": 0.00012288273615635177, | |
| "loss": 0.8336, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 1.8978583196046128, | |
| "grad_norm": 0.04546576365828514, | |
| "learning_rate": 0.0001226384364820847, | |
| "loss": 0.5656, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 1.900054914881933, | |
| "grad_norm": 0.048822496086359024, | |
| "learning_rate": 0.00012239413680781758, | |
| "loss": 0.7217, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 1.9022515101592532, | |
| "grad_norm": 0.0437794029712677, | |
| "learning_rate": 0.0001221498371335505, | |
| "loss": 0.6082, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 1.9044481054365732, | |
| "grad_norm": 0.05672450736165047, | |
| "learning_rate": 0.00012190553745928338, | |
| "loss": 0.6227, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 1.9066447007138936, | |
| "grad_norm": 0.07971169799566269, | |
| "learning_rate": 0.00012166123778501629, | |
| "loss": 0.8374, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 1.9088412959912135, | |
| "grad_norm": 0.04980779439210892, | |
| "learning_rate": 0.00012141693811074918, | |
| "loss": 0.7397, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 1.9110378912685337, | |
| "grad_norm": 0.03777102008461952, | |
| "learning_rate": 0.00012117263843648208, | |
| "loss": 0.6381, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.913234486545854, | |
| "grad_norm": 0.049683138728141785, | |
| "learning_rate": 0.00012092833876221498, | |
| "loss": 0.6401, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 1.9154310818231741, | |
| "grad_norm": 0.03963594511151314, | |
| "learning_rate": 0.00012068403908794787, | |
| "loss": 0.7263, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 1.9176276771004943, | |
| "grad_norm": 0.05527025833725929, | |
| "learning_rate": 0.00012043973941368077, | |
| "loss": 0.7111, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 1.9198242723778143, | |
| "grad_norm": 0.04393410310149193, | |
| "learning_rate": 0.00012019543973941368, | |
| "loss": 0.8593, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 1.9220208676551347, | |
| "grad_norm": 0.062434881925582886, | |
| "learning_rate": 0.00011995114006514658, | |
| "loss": 0.7676, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 1.9242174629324547, | |
| "grad_norm": 0.07280800491571426, | |
| "learning_rate": 0.00011970684039087947, | |
| "loss": 0.7656, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 1.9264140582097748, | |
| "grad_norm": 0.05514119192957878, | |
| "learning_rate": 0.00011946254071661237, | |
| "loss": 0.5903, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 1.928610653487095, | |
| "grad_norm": 0.06919682025909424, | |
| "learning_rate": 0.00011921824104234527, | |
| "loss": 0.72, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 1.930807248764415, | |
| "grad_norm": 0.04627091437578201, | |
| "learning_rate": 0.00011897394136807816, | |
| "loss": 0.6976, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 1.9330038440417354, | |
| "grad_norm": 0.05157488211989403, | |
| "learning_rate": 0.00011872964169381106, | |
| "loss": 0.5897, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.9352004393190554, | |
| "grad_norm": 0.05058778077363968, | |
| "learning_rate": 0.00011848534201954397, | |
| "loss": 0.7003, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 1.9373970345963756, | |
| "grad_norm": 0.03946829214692116, | |
| "learning_rate": 0.00011824104234527687, | |
| "loss": 0.5076, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 1.9395936298736958, | |
| "grad_norm": 0.04601827263832092, | |
| "learning_rate": 0.00011799674267100976, | |
| "loss": 0.7444, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 1.941790225151016, | |
| "grad_norm": 0.05405599623918533, | |
| "learning_rate": 0.00011775244299674266, | |
| "loss": 0.7075, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 1.9439868204283361, | |
| "grad_norm": 0.04909859970211983, | |
| "learning_rate": 0.00011750814332247557, | |
| "loss": 1.0181, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 1.9461834157056561, | |
| "grad_norm": 0.055644210427999496, | |
| "learning_rate": 0.00011726384364820845, | |
| "loss": 0.7143, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 1.9483800109829765, | |
| "grad_norm": 0.033494554460048676, | |
| "learning_rate": 0.00011701954397394136, | |
| "loss": 0.5905, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 1.9505766062602965, | |
| "grad_norm": 0.048672039061784744, | |
| "learning_rate": 0.00011677524429967426, | |
| "loss": 0.5846, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 1.9527732015376167, | |
| "grad_norm": 0.052615948021411896, | |
| "learning_rate": 0.00011653094462540716, | |
| "loss": 0.7174, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 1.9549697968149369, | |
| "grad_norm": 0.04827093333005905, | |
| "learning_rate": 0.00011628664495114005, | |
| "loss": 0.6304, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.9571663920922568, | |
| "grad_norm": 0.042970217764377594, | |
| "learning_rate": 0.00011604234527687295, | |
| "loss": 0.7678, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 1.9593629873695773, | |
| "grad_norm": 0.04997050389647484, | |
| "learning_rate": 0.00011579804560260586, | |
| "loss": 0.6019, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 1.9615595826468972, | |
| "grad_norm": 0.04948752373456955, | |
| "learning_rate": 0.00011555374592833876, | |
| "loss": 0.6425, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 1.9637561779242174, | |
| "grad_norm": 0.04462786018848419, | |
| "learning_rate": 0.00011530944625407165, | |
| "loss": 0.7068, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 1.9659527732015376, | |
| "grad_norm": 0.051252443343400955, | |
| "learning_rate": 0.00011506514657980455, | |
| "loss": 0.5193, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 1.9681493684788578, | |
| "grad_norm": 0.045937035232782364, | |
| "learning_rate": 0.00011482084690553745, | |
| "loss": 0.7832, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 1.970345963756178, | |
| "grad_norm": 0.04210161417722702, | |
| "learning_rate": 0.00011457654723127034, | |
| "loss": 0.5735, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 1.972542559033498, | |
| "grad_norm": 0.04671672359108925, | |
| "learning_rate": 0.00011433224755700325, | |
| "loss": 0.7834, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 1.9747391543108184, | |
| "grad_norm": 0.06232306733727455, | |
| "learning_rate": 0.00011408794788273615, | |
| "loss": 0.7376, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 1.9769357495881383, | |
| "grad_norm": 0.05481863394379616, | |
| "learning_rate": 0.00011384364820846905, | |
| "loss": 0.5515, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.9791323448654585, | |
| "grad_norm": 0.04274000972509384, | |
| "learning_rate": 0.00011359934853420194, | |
| "loss": 0.7249, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 1.9813289401427787, | |
| "grad_norm": 0.05033233016729355, | |
| "learning_rate": 0.00011335504885993484, | |
| "loss": 0.7522, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 1.9835255354200987, | |
| "grad_norm": 0.056208908557891846, | |
| "learning_rate": 0.00011311074918566775, | |
| "loss": 0.6917, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 1.985722130697419, | |
| "grad_norm": 0.039569608867168427, | |
| "learning_rate": 0.00011286644951140064, | |
| "loss": 0.5706, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 1.987918725974739, | |
| "grad_norm": 0.061526212841272354, | |
| "learning_rate": 0.00011262214983713354, | |
| "loss": 0.5911, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 1.9901153212520593, | |
| "grad_norm": 0.04074925556778908, | |
| "learning_rate": 0.00011237785016286644, | |
| "loss": 0.7588, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 1.9923119165293794, | |
| "grad_norm": 0.0786409005522728, | |
| "learning_rate": 0.00011213355048859934, | |
| "loss": 0.5221, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 1.9945085118066996, | |
| "grad_norm": 0.03439033031463623, | |
| "learning_rate": 0.00011188925081433223, | |
| "loss": 0.5648, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 1.9967051070840198, | |
| "grad_norm": 0.03860178589820862, | |
| "learning_rate": 0.00011164495114006514, | |
| "loss": 0.4791, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 1.9989017023613398, | |
| "grad_norm": 0.09322784096002579, | |
| "learning_rate": 0.00011140065146579804, | |
| "loss": 0.6842, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 2.00109829763866, | |
| "grad_norm": 0.05441444739699364, | |
| "learning_rate": 0.00011115635179153093, | |
| "loss": 0.774, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 2.00329489291598, | |
| "grad_norm": 0.052353233098983765, | |
| "learning_rate": 0.00011091205211726383, | |
| "loss": 0.6595, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 2.0054914881933006, | |
| "grad_norm": 0.046645645052194595, | |
| "learning_rate": 0.00011066775244299673, | |
| "loss": 0.5674, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 2.0076880834706206, | |
| "grad_norm": 0.052380114793777466, | |
| "learning_rate": 0.00011042345276872964, | |
| "loss": 0.6832, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 2.0098846787479405, | |
| "grad_norm": 0.05334731563925743, | |
| "learning_rate": 0.00011017915309446253, | |
| "loss": 0.6386, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 2.012081274025261, | |
| "grad_norm": 0.04087246209383011, | |
| "learning_rate": 0.00010993485342019543, | |
| "loss": 0.7986, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 2.014277869302581, | |
| "grad_norm": 0.05971505120396614, | |
| "learning_rate": 0.00010969055374592833, | |
| "loss": 0.8546, | |
| "step": 917 | |
| }, | |
| { | |
| "epoch": 2.0164744645799013, | |
| "grad_norm": 0.04933095723390579, | |
| "learning_rate": 0.00010944625407166122, | |
| "loss": 0.7736, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 2.0186710598572213, | |
| "grad_norm": 0.046272121369838715, | |
| "learning_rate": 0.00010920195439739412, | |
| "loss": 0.629, | |
| "step": 919 | |
| }, | |
| { | |
| "epoch": 2.0208676551345413, | |
| "grad_norm": 0.04796759784221649, | |
| "learning_rate": 0.00010895765472312703, | |
| "loss": 0.6354, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 2.0230642504118617, | |
| "grad_norm": 0.05017026141285896, | |
| "learning_rate": 0.00010871335504885993, | |
| "loss": 0.5821, | |
| "step": 921 | |
| }, | |
| { | |
| "epoch": 2.0252608456891816, | |
| "grad_norm": 0.04263904318213463, | |
| "learning_rate": 0.00010846905537459282, | |
| "loss": 0.6971, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 2.027457440966502, | |
| "grad_norm": 0.05482470244169235, | |
| "learning_rate": 0.00010822475570032572, | |
| "loss": 0.6048, | |
| "step": 923 | |
| }, | |
| { | |
| "epoch": 2.029654036243822, | |
| "grad_norm": 0.05933058261871338, | |
| "learning_rate": 0.00010798045602605862, | |
| "loss": 0.8025, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 2.0318506315211424, | |
| "grad_norm": 0.057402413338422775, | |
| "learning_rate": 0.00010773615635179153, | |
| "loss": 0.6536, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 2.0340472267984624, | |
| "grad_norm": 0.04043892025947571, | |
| "learning_rate": 0.00010749185667752442, | |
| "loss": 0.6329, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 2.0362438220757824, | |
| "grad_norm": 0.04894665256142616, | |
| "learning_rate": 0.00010724755700325732, | |
| "loss": 0.7953, | |
| "step": 927 | |
| }, | |
| { | |
| "epoch": 2.0384404173531028, | |
| "grad_norm": 0.04023706912994385, | |
| "learning_rate": 0.00010700325732899022, | |
| "loss": 0.6657, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 2.0406370126304227, | |
| "grad_norm": 0.04153961315751076, | |
| "learning_rate": 0.00010675895765472311, | |
| "loss": 0.5859, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 2.042833607907743, | |
| "grad_norm": 0.034479089081287384, | |
| "learning_rate": 0.00010651465798045601, | |
| "loss": 0.5664, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 2.045030203185063, | |
| "grad_norm": 0.04258917272090912, | |
| "learning_rate": 0.00010627035830618892, | |
| "loss": 0.6099, | |
| "step": 931 | |
| }, | |
| { | |
| "epoch": 2.047226798462383, | |
| "grad_norm": 0.06230498105287552, | |
| "learning_rate": 0.00010602605863192182, | |
| "loss": 0.6836, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 2.0494233937397035, | |
| "grad_norm": 0.0505065955221653, | |
| "learning_rate": 0.00010578175895765471, | |
| "loss": 0.699, | |
| "step": 933 | |
| }, | |
| { | |
| "epoch": 2.0516199890170235, | |
| "grad_norm": 0.04756687209010124, | |
| "learning_rate": 0.00010553745928338761, | |
| "loss": 0.5818, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 2.053816584294344, | |
| "grad_norm": 0.03504083678126335, | |
| "learning_rate": 0.00010529315960912051, | |
| "loss": 0.5922, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 2.056013179571664, | |
| "grad_norm": 0.04338408634066582, | |
| "learning_rate": 0.0001050488599348534, | |
| "loss": 0.7937, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 2.0582097748489843, | |
| "grad_norm": 0.04064236208796501, | |
| "learning_rate": 0.0001048045602605863, | |
| "loss": 0.7191, | |
| "step": 937 | |
| }, | |
| { | |
| "epoch": 2.0604063701263042, | |
| "grad_norm": 0.053273994475603104, | |
| "learning_rate": 0.00010456026058631921, | |
| "loss": 0.4936, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 2.062602965403624, | |
| "grad_norm": 0.06294192373752594, | |
| "learning_rate": 0.00010431596091205211, | |
| "loss": 0.8157, | |
| "step": 939 | |
| }, | |
| { | |
| "epoch": 2.0647995606809446, | |
| "grad_norm": 0.04619294032454491, | |
| "learning_rate": 0.000104071661237785, | |
| "loss": 0.6172, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 2.0669961559582646, | |
| "grad_norm": 0.03779340907931328, | |
| "learning_rate": 0.0001038273615635179, | |
| "loss": 0.6405, | |
| "step": 941 | |
| }, | |
| { | |
| "epoch": 2.069192751235585, | |
| "grad_norm": 0.04682890698313713, | |
| "learning_rate": 0.0001035830618892508, | |
| "loss": 0.5335, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 2.071389346512905, | |
| "grad_norm": 0.037470750510692596, | |
| "learning_rate": 0.0001033387622149837, | |
| "loss": 0.6709, | |
| "step": 943 | |
| }, | |
| { | |
| "epoch": 2.073585941790225, | |
| "grad_norm": 0.06082337349653244, | |
| "learning_rate": 0.0001030944625407166, | |
| "loss": 0.6056, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 2.0757825370675453, | |
| "grad_norm": 0.03778183087706566, | |
| "learning_rate": 0.0001028501628664495, | |
| "loss": 0.6584, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 2.0779791323448653, | |
| "grad_norm": 0.05191833898425102, | |
| "learning_rate": 0.0001026058631921824, | |
| "loss": 0.6648, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 2.0801757276221857, | |
| "grad_norm": 0.04200424626469612, | |
| "learning_rate": 0.00010236156351791529, | |
| "loss": 0.5183, | |
| "step": 947 | |
| }, | |
| { | |
| "epoch": 2.0823723228995057, | |
| "grad_norm": 0.04820658266544342, | |
| "learning_rate": 0.0001021172638436482, | |
| "loss": 0.5925, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 2.084568918176826, | |
| "grad_norm": 0.0433504655957222, | |
| "learning_rate": 0.0001018729641693811, | |
| "loss": 0.8134, | |
| "step": 949 | |
| }, | |
| { | |
| "epoch": 2.086765513454146, | |
| "grad_norm": 0.03969017416238785, | |
| "learning_rate": 0.00010162866449511399, | |
| "loss": 0.6694, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 2.088962108731466, | |
| "grad_norm": 0.04646764323115349, | |
| "learning_rate": 0.00010138436482084689, | |
| "loss": 0.6878, | |
| "step": 951 | |
| }, | |
| { | |
| "epoch": 2.0911587040087865, | |
| "grad_norm": 0.06577397137880325, | |
| "learning_rate": 0.00010114006514657979, | |
| "loss": 0.7297, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 2.0933552992861064, | |
| "grad_norm": 0.05729677528142929, | |
| "learning_rate": 0.0001008957654723127, | |
| "loss": 0.733, | |
| "step": 953 | |
| }, | |
| { | |
| "epoch": 2.095551894563427, | |
| "grad_norm": 0.03979307413101196, | |
| "learning_rate": 0.00010065146579804558, | |
| "loss": 0.5778, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 2.097748489840747, | |
| "grad_norm": 0.05198509618639946, | |
| "learning_rate": 0.00010040716612377849, | |
| "loss": 0.731, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 2.099945085118067, | |
| "grad_norm": 0.05052348971366882, | |
| "learning_rate": 0.00010016286644951139, | |
| "loss": 0.6816, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 2.102141680395387, | |
| "grad_norm": 0.046905145049095154, | |
| "learning_rate": 9.991856677524428e-05, | |
| "loss": 0.6744, | |
| "step": 957 | |
| }, | |
| { | |
| "epoch": 2.104338275672707, | |
| "grad_norm": 0.04074231535196304, | |
| "learning_rate": 9.967426710097718e-05, | |
| "loss": 0.6129, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 2.1065348709500276, | |
| "grad_norm": 0.049627695232629776, | |
| "learning_rate": 9.942996742671008e-05, | |
| "loss": 0.6302, | |
| "step": 959 | |
| }, | |
| { | |
| "epoch": 2.1087314662273475, | |
| "grad_norm": 0.04932795837521553, | |
| "learning_rate": 9.918566775244299e-05, | |
| "loss": 0.6256, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 2.110928061504668, | |
| "grad_norm": 0.05283183604478836, | |
| "learning_rate": 9.894136807817588e-05, | |
| "loss": 0.7748, | |
| "step": 961 | |
| }, | |
| { | |
| "epoch": 2.113124656781988, | |
| "grad_norm": 0.05655486509203911, | |
| "learning_rate": 9.869706840390878e-05, | |
| "loss": 0.6449, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 2.115321252059308, | |
| "grad_norm": 0.056285202503204346, | |
| "learning_rate": 9.845276872964168e-05, | |
| "loss": 0.5045, | |
| "step": 963 | |
| }, | |
| { | |
| "epoch": 2.1175178473366283, | |
| "grad_norm": 0.0356547012925148, | |
| "learning_rate": 9.820846905537458e-05, | |
| "loss": 0.6095, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 2.1197144426139483, | |
| "grad_norm": 0.042278558015823364, | |
| "learning_rate": 9.796416938110747e-05, | |
| "loss": 0.593, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 2.1219110378912687, | |
| "grad_norm": 0.06479320675134659, | |
| "learning_rate": 9.771986970684038e-05, | |
| "loss": 0.6385, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 2.1241076331685886, | |
| "grad_norm": 0.042625319212675095, | |
| "learning_rate": 9.747557003257328e-05, | |
| "loss": 0.6546, | |
| "step": 967 | |
| }, | |
| { | |
| "epoch": 2.1263042284459086, | |
| "grad_norm": 0.04128136858344078, | |
| "learning_rate": 9.723127035830617e-05, | |
| "loss": 0.6223, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 2.128500823723229, | |
| "grad_norm": 0.06777093559503555, | |
| "learning_rate": 9.698697068403907e-05, | |
| "loss": 0.6113, | |
| "step": 969 | |
| }, | |
| { | |
| "epoch": 2.130697419000549, | |
| "grad_norm": 0.0486929826438427, | |
| "learning_rate": 9.674267100977197e-05, | |
| "loss": 0.7408, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 2.1328940142778694, | |
| "grad_norm": 0.06708229333162308, | |
| "learning_rate": 9.649837133550488e-05, | |
| "loss": 0.665, | |
| "step": 971 | |
| }, | |
| { | |
| "epoch": 2.1350906095551894, | |
| "grad_norm": 0.060830358415842056, | |
| "learning_rate": 9.625407166123777e-05, | |
| "loss": 0.642, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 2.13728720483251, | |
| "grad_norm": 0.04374934732913971, | |
| "learning_rate": 9.600977198697067e-05, | |
| "loss": 0.6245, | |
| "step": 973 | |
| }, | |
| { | |
| "epoch": 2.1394838001098297, | |
| "grad_norm": 0.049657490104436874, | |
| "learning_rate": 9.576547231270357e-05, | |
| "loss": 0.7092, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 2.1416803953871497, | |
| "grad_norm": 0.04418911039829254, | |
| "learning_rate": 9.552117263843646e-05, | |
| "loss": 0.7128, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 2.14387699066447, | |
| "grad_norm": 0.05693337321281433, | |
| "learning_rate": 9.527687296416936e-05, | |
| "loss": 0.6567, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 2.14607358594179, | |
| "grad_norm": 0.03899364918470383, | |
| "learning_rate": 9.503257328990227e-05, | |
| "loss": 0.6489, | |
| "step": 977 | |
| }, | |
| { | |
| "epoch": 2.1482701812191105, | |
| "grad_norm": 0.05260545760393143, | |
| "learning_rate": 9.478827361563517e-05, | |
| "loss": 0.6311, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 2.1504667764964305, | |
| "grad_norm": 0.05104517191648483, | |
| "learning_rate": 9.454397394136806e-05, | |
| "loss": 0.6754, | |
| "step": 979 | |
| }, | |
| { | |
| "epoch": 2.152663371773751, | |
| "grad_norm": 0.056579869240522385, | |
| "learning_rate": 9.429967426710096e-05, | |
| "loss": 0.742, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 2.154859967051071, | |
| "grad_norm": 0.04211939126253128, | |
| "learning_rate": 9.405537459283386e-05, | |
| "loss": 0.4709, | |
| "step": 981 | |
| }, | |
| { | |
| "epoch": 2.157056562328391, | |
| "grad_norm": 0.04625248908996582, | |
| "learning_rate": 9.381107491856675e-05, | |
| "loss": 0.5968, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 2.1592531576057112, | |
| "grad_norm": 0.057066936045885086, | |
| "learning_rate": 9.356677524429967e-05, | |
| "loss": 0.7691, | |
| "step": 983 | |
| }, | |
| { | |
| "epoch": 2.161449752883031, | |
| "grad_norm": 0.05318013206124306, | |
| "learning_rate": 9.332247557003257e-05, | |
| "loss": 0.6044, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 2.1636463481603516, | |
| "grad_norm": 0.05898617208003998, | |
| "learning_rate": 9.307817589576547e-05, | |
| "loss": 0.5647, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 2.1658429434376716, | |
| "grad_norm": 0.061303168535232544, | |
| "learning_rate": 9.283387622149838e-05, | |
| "loss": 0.5899, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 2.168039538714992, | |
| "grad_norm": 0.04799562692642212, | |
| "learning_rate": 9.258957654723127e-05, | |
| "loss": 0.7572, | |
| "step": 987 | |
| }, | |
| { | |
| "epoch": 2.170236133992312, | |
| "grad_norm": 0.044705580919981, | |
| "learning_rate": 9.234527687296417e-05, | |
| "loss": 0.5564, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 2.172432729269632, | |
| "grad_norm": 0.041354063898324966, | |
| "learning_rate": 9.210097719869707e-05, | |
| "loss": 0.6036, | |
| "step": 989 | |
| }, | |
| { | |
| "epoch": 2.1746293245469523, | |
| "grad_norm": 0.05093163624405861, | |
| "learning_rate": 9.185667752442996e-05, | |
| "loss": 0.8037, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 2.1768259198242723, | |
| "grad_norm": 0.049296095967292786, | |
| "learning_rate": 9.161237785016286e-05, | |
| "loss": 0.6168, | |
| "step": 991 | |
| }, | |
| { | |
| "epoch": 2.1790225151015927, | |
| "grad_norm": 0.036712076514959335, | |
| "learning_rate": 9.136807817589577e-05, | |
| "loss": 0.5757, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 2.1812191103789127, | |
| "grad_norm": 0.061949945986270905, | |
| "learning_rate": 9.112377850162867e-05, | |
| "loss": 0.7439, | |
| "step": 993 | |
| }, | |
| { | |
| "epoch": 2.1834157056562327, | |
| "grad_norm": 0.04836397245526314, | |
| "learning_rate": 9.087947882736156e-05, | |
| "loss": 0.5687, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 2.185612300933553, | |
| "grad_norm": 0.03990459069609642, | |
| "learning_rate": 9.063517915309446e-05, | |
| "loss": 0.5616, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 2.187808896210873, | |
| "grad_norm": 0.05062219500541687, | |
| "learning_rate": 9.039087947882736e-05, | |
| "loss": 0.7688, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 2.1900054914881935, | |
| "grad_norm": 0.049748651683330536, | |
| "learning_rate": 9.014657980456025e-05, | |
| "loss": 0.6998, | |
| "step": 997 | |
| }, | |
| { | |
| "epoch": 2.1922020867655134, | |
| "grad_norm": 0.04505657032132149, | |
| "learning_rate": 8.990228013029316e-05, | |
| "loss": 0.6697, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 2.1943986820428334, | |
| "grad_norm": 0.051343776285648346, | |
| "learning_rate": 8.965798045602606e-05, | |
| "loss": 0.806, | |
| "step": 999 | |
| }, | |
| { | |
| "epoch": 2.196595277320154, | |
| "grad_norm": 0.0413266085088253, | |
| "learning_rate": 8.941368078175896e-05, | |
| "loss": 0.5986, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 2.1987918725974738, | |
| "grad_norm": 0.048408783972263336, | |
| "learning_rate": 8.916938110749185e-05, | |
| "loss": 0.478, | |
| "step": 1001 | |
| }, | |
| { | |
| "epoch": 2.200988467874794, | |
| "grad_norm": 0.03760487586259842, | |
| "learning_rate": 8.892508143322475e-05, | |
| "loss": 0.5888, | |
| "step": 1002 | |
| }, | |
| { | |
| "epoch": 2.203185063152114, | |
| "grad_norm": 0.04939005896449089, | |
| "learning_rate": 8.868078175895766e-05, | |
| "loss": 0.6409, | |
| "step": 1003 | |
| }, | |
| { | |
| "epoch": 2.2053816584294346, | |
| "grad_norm": 0.04825667291879654, | |
| "learning_rate": 8.843648208469055e-05, | |
| "loss": 0.5696, | |
| "step": 1004 | |
| }, | |
| { | |
| "epoch": 2.2075782537067545, | |
| "grad_norm": 0.04127410426735878, | |
| "learning_rate": 8.819218241042345e-05, | |
| "loss": 0.6175, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 2.2097748489840745, | |
| "grad_norm": 0.05631517246365547, | |
| "learning_rate": 8.794788273615635e-05, | |
| "loss": 0.6802, | |
| "step": 1006 | |
| }, | |
| { | |
| "epoch": 2.211971444261395, | |
| "grad_norm": 0.04190594330430031, | |
| "learning_rate": 8.770358306188925e-05, | |
| "loss": 0.5481, | |
| "step": 1007 | |
| }, | |
| { | |
| "epoch": 2.214168039538715, | |
| "grad_norm": 0.05055314674973488, | |
| "learning_rate": 8.745928338762214e-05, | |
| "loss": 0.7913, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 2.2163646348160353, | |
| "grad_norm": 0.04413871094584465, | |
| "learning_rate": 8.721498371335505e-05, | |
| "loss": 0.683, | |
| "step": 1009 | |
| }, | |
| { | |
| "epoch": 2.2185612300933553, | |
| "grad_norm": 0.06312219798564911, | |
| "learning_rate": 8.697068403908795e-05, | |
| "loss": 0.5454, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 2.2207578253706757, | |
| "grad_norm": 0.03939354047179222, | |
| "learning_rate": 8.672638436482085e-05, | |
| "loss": 0.6548, | |
| "step": 1011 | |
| }, | |
| { | |
| "epoch": 2.2229544206479956, | |
| "grad_norm": 0.04720534384250641, | |
| "learning_rate": 8.648208469055374e-05, | |
| "loss": 0.869, | |
| "step": 1012 | |
| }, | |
| { | |
| "epoch": 2.2251510159253156, | |
| "grad_norm": 0.09054285287857056, | |
| "learning_rate": 8.623778501628664e-05, | |
| "loss": 0.6499, | |
| "step": 1013 | |
| }, | |
| { | |
| "epoch": 2.227347611202636, | |
| "grad_norm": 0.06740359216928482, | |
| "learning_rate": 8.599348534201955e-05, | |
| "loss": 0.7657, | |
| "step": 1014 | |
| }, | |
| { | |
| "epoch": 2.229544206479956, | |
| "grad_norm": 0.08037208765745163, | |
| "learning_rate": 8.574918566775243e-05, | |
| "loss": 0.825, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 2.2317408017572764, | |
| "grad_norm": 0.0381237156689167, | |
| "learning_rate": 8.550488599348534e-05, | |
| "loss": 0.6602, | |
| "step": 1016 | |
| }, | |
| { | |
| "epoch": 2.2339373970345964, | |
| "grad_norm": 0.05984542518854141, | |
| "learning_rate": 8.526058631921824e-05, | |
| "loss": 0.7109, | |
| "step": 1017 | |
| }, | |
| { | |
| "epoch": 2.2361339923119163, | |
| "grad_norm": 0.057318367063999176, | |
| "learning_rate": 8.501628664495114e-05, | |
| "loss": 0.7673, | |
| "step": 1018 | |
| }, | |
| { | |
| "epoch": 2.2383305875892368, | |
| "grad_norm": 0.048353392630815506, | |
| "learning_rate": 8.477198697068403e-05, | |
| "loss": 0.731, | |
| "step": 1019 | |
| }, | |
| { | |
| "epoch": 2.2405271828665567, | |
| "grad_norm": 0.09073536843061447, | |
| "learning_rate": 8.452768729641694e-05, | |
| "loss": 0.7223, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 2.242723778143877, | |
| "grad_norm": 0.0407768115401268, | |
| "learning_rate": 8.428338762214984e-05, | |
| "loss": 0.5907, | |
| "step": 1021 | |
| }, | |
| { | |
| "epoch": 2.244920373421197, | |
| "grad_norm": 0.04868381470441818, | |
| "learning_rate": 8.403908794788273e-05, | |
| "loss": 0.5471, | |
| "step": 1022 | |
| }, | |
| { | |
| "epoch": 2.247116968698517, | |
| "grad_norm": 0.056227799504995346, | |
| "learning_rate": 8.379478827361563e-05, | |
| "loss": 0.7733, | |
| "step": 1023 | |
| }, | |
| { | |
| "epoch": 2.2493135639758375, | |
| "grad_norm": 0.05046963691711426, | |
| "learning_rate": 8.355048859934853e-05, | |
| "loss": 0.6275, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 2.2515101592531575, | |
| "grad_norm": 0.03843815252184868, | |
| "learning_rate": 8.330618892508144e-05, | |
| "loss": 0.6212, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 2.253706754530478, | |
| "grad_norm": 0.04150962457060814, | |
| "learning_rate": 8.306188925081432e-05, | |
| "loss": 0.5651, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 2.255903349807798, | |
| "grad_norm": 0.05324546620249748, | |
| "learning_rate": 8.281758957654723e-05, | |
| "loss": 0.6938, | |
| "step": 1027 | |
| }, | |
| { | |
| "epoch": 2.2580999450851182, | |
| "grad_norm": 0.039021965116262436, | |
| "learning_rate": 8.257328990228013e-05, | |
| "loss": 0.725, | |
| "step": 1028 | |
| }, | |
| { | |
| "epoch": 2.260296540362438, | |
| "grad_norm": 0.060293395072221756, | |
| "learning_rate": 8.232899022801302e-05, | |
| "loss": 0.7034, | |
| "step": 1029 | |
| }, | |
| { | |
| "epoch": 2.262493135639758, | |
| "grad_norm": 0.046492014080286026, | |
| "learning_rate": 8.208469055374592e-05, | |
| "loss": 0.5136, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 2.2646897309170786, | |
| "grad_norm": 0.04280600696802139, | |
| "learning_rate": 8.184039087947882e-05, | |
| "loss": 0.5654, | |
| "step": 1031 | |
| }, | |
| { | |
| "epoch": 2.2668863261943986, | |
| "grad_norm": 0.05168813094496727, | |
| "learning_rate": 8.159609120521173e-05, | |
| "loss": 0.5706, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 2.269082921471719, | |
| "grad_norm": 0.06857836991548538, | |
| "learning_rate": 8.135179153094462e-05, | |
| "loss": 0.8093, | |
| "step": 1033 | |
| }, | |
| { | |
| "epoch": 2.271279516749039, | |
| "grad_norm": 0.0616997666656971, | |
| "learning_rate": 8.110749185667752e-05, | |
| "loss": 0.6479, | |
| "step": 1034 | |
| }, | |
| { | |
| "epoch": 2.2734761120263594, | |
| "grad_norm": 0.04081263393163681, | |
| "learning_rate": 8.086319218241042e-05, | |
| "loss": 0.5847, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 2.2756727073036793, | |
| "grad_norm": 0.0425679050385952, | |
| "learning_rate": 8.061889250814331e-05, | |
| "loss": 0.6171, | |
| "step": 1036 | |
| }, | |
| { | |
| "epoch": 2.2778693025809993, | |
| "grad_norm": 0.036269091069698334, | |
| "learning_rate": 8.037459283387621e-05, | |
| "loss": 0.5724, | |
| "step": 1037 | |
| }, | |
| { | |
| "epoch": 2.2800658978583197, | |
| "grad_norm": 0.04767118766903877, | |
| "learning_rate": 8.013029315960912e-05, | |
| "loss": 0.6088, | |
| "step": 1038 | |
| }, | |
| { | |
| "epoch": 2.2822624931356397, | |
| "grad_norm": 0.04134832322597504, | |
| "learning_rate": 7.988599348534202e-05, | |
| "loss": 0.5544, | |
| "step": 1039 | |
| }, | |
| { | |
| "epoch": 2.28445908841296, | |
| "grad_norm": 0.05227472260594368, | |
| "learning_rate": 7.964169381107491e-05, | |
| "loss": 0.6697, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 2.28665568369028, | |
| "grad_norm": 0.05427214875817299, | |
| "learning_rate": 7.939739413680781e-05, | |
| "loss": 0.6783, | |
| "step": 1041 | |
| }, | |
| { | |
| "epoch": 2.2888522789676005, | |
| "grad_norm": 0.05169343948364258, | |
| "learning_rate": 7.915309446254071e-05, | |
| "loss": 0.7326, | |
| "step": 1042 | |
| }, | |
| { | |
| "epoch": 2.2910488742449204, | |
| "grad_norm": 0.03582863137125969, | |
| "learning_rate": 7.890879478827362e-05, | |
| "loss": 0.5292, | |
| "step": 1043 | |
| }, | |
| { | |
| "epoch": 2.2932454695222404, | |
| "grad_norm": 0.04561784490942955, | |
| "learning_rate": 7.86644951140065e-05, | |
| "loss": 0.6495, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 2.295442064799561, | |
| "grad_norm": 0.04176398739218712, | |
| "learning_rate": 7.842019543973941e-05, | |
| "loss": 0.7175, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 2.297638660076881, | |
| "grad_norm": 0.04040032997727394, | |
| "learning_rate": 7.817589576547231e-05, | |
| "loss": 0.6138, | |
| "step": 1046 | |
| }, | |
| { | |
| "epoch": 2.2998352553542007, | |
| "grad_norm": 0.06742952764034271, | |
| "learning_rate": 7.79315960912052e-05, | |
| "loss": 0.547, | |
| "step": 1047 | |
| }, | |
| { | |
| "epoch": 2.302031850631521, | |
| "grad_norm": 0.04996305704116821, | |
| "learning_rate": 7.76872964169381e-05, | |
| "loss": 0.6176, | |
| "step": 1048 | |
| }, | |
| { | |
| "epoch": 2.304228445908841, | |
| "grad_norm": 0.04592761769890785, | |
| "learning_rate": 7.7442996742671e-05, | |
| "loss": 0.6208, | |
| "step": 1049 | |
| }, | |
| { | |
| "epoch": 2.3064250411861615, | |
| "grad_norm": 0.05607963353395462, | |
| "learning_rate": 7.719869706840391e-05, | |
| "loss": 0.675, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 2.3086216364634815, | |
| "grad_norm": 0.04420773312449455, | |
| "learning_rate": 7.69543973941368e-05, | |
| "loss": 0.5119, | |
| "step": 1051 | |
| }, | |
| { | |
| "epoch": 2.310818231740802, | |
| "grad_norm": 0.044873543083667755, | |
| "learning_rate": 7.67100977198697e-05, | |
| "loss": 0.6016, | |
| "step": 1052 | |
| }, | |
| { | |
| "epoch": 2.313014827018122, | |
| "grad_norm": 0.0465567484498024, | |
| "learning_rate": 7.64657980456026e-05, | |
| "loss": 0.686, | |
| "step": 1053 | |
| }, | |
| { | |
| "epoch": 2.315211422295442, | |
| "grad_norm": 0.04192805662751198, | |
| "learning_rate": 7.622149837133549e-05, | |
| "loss": 0.6144, | |
| "step": 1054 | |
| }, | |
| { | |
| "epoch": 2.3174080175727623, | |
| "grad_norm": 0.04011636599898338, | |
| "learning_rate": 7.59771986970684e-05, | |
| "loss": 0.8029, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 2.3196046128500822, | |
| "grad_norm": 0.04277574643492699, | |
| "learning_rate": 7.57328990228013e-05, | |
| "loss": 0.6306, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 2.3218012081274026, | |
| "grad_norm": 0.05560114234685898, | |
| "learning_rate": 7.54885993485342e-05, | |
| "loss": 0.8514, | |
| "step": 1057 | |
| }, | |
| { | |
| "epoch": 2.3239978034047226, | |
| "grad_norm": 0.037569232285022736, | |
| "learning_rate": 7.524429967426709e-05, | |
| "loss": 0.469, | |
| "step": 1058 | |
| }, | |
| { | |
| "epoch": 2.326194398682043, | |
| "grad_norm": 0.045851316303014755, | |
| "learning_rate": 7.5e-05, | |
| "loss": 0.5967, | |
| "step": 1059 | |
| }, | |
| { | |
| "epoch": 2.328390993959363, | |
| "grad_norm": 0.03776060789823532, | |
| "learning_rate": 7.47557003257329e-05, | |
| "loss": 0.5763, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 2.330587589236683, | |
| "grad_norm": 0.04432753473520279, | |
| "learning_rate": 7.451140065146579e-05, | |
| "loss": 0.6155, | |
| "step": 1061 | |
| }, | |
| { | |
| "epoch": 2.3327841845140034, | |
| "grad_norm": 0.04885265603661537, | |
| "learning_rate": 7.426710097719869e-05, | |
| "loss": 0.7005, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 2.3349807797913233, | |
| "grad_norm": 0.04530097171664238, | |
| "learning_rate": 7.402280130293159e-05, | |
| "loss": 0.7549, | |
| "step": 1063 | |
| }, | |
| { | |
| "epoch": 2.3371773750686438, | |
| "grad_norm": 0.04386676102876663, | |
| "learning_rate": 7.37785016286645e-05, | |
| "loss": 0.6721, | |
| "step": 1064 | |
| }, | |
| { | |
| "epoch": 2.3393739703459637, | |
| "grad_norm": 0.04051361232995987, | |
| "learning_rate": 7.353420195439738e-05, | |
| "loss": 0.6354, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 2.341570565623284, | |
| "grad_norm": 0.05379435792565346, | |
| "learning_rate": 7.328990228013029e-05, | |
| "loss": 0.7204, | |
| "step": 1066 | |
| }, | |
| { | |
| "epoch": 2.343767160900604, | |
| "grad_norm": 0.03002171590924263, | |
| "learning_rate": 7.304560260586319e-05, | |
| "loss": 0.5271, | |
| "step": 1067 | |
| }, | |
| { | |
| "epoch": 2.345963756177924, | |
| "grad_norm": 0.03606308996677399, | |
| "learning_rate": 7.280130293159608e-05, | |
| "loss": 0.6648, | |
| "step": 1068 | |
| }, | |
| { | |
| "epoch": 2.3481603514552445, | |
| "grad_norm": 0.058911386877298355, | |
| "learning_rate": 7.255700325732898e-05, | |
| "loss": 0.7517, | |
| "step": 1069 | |
| }, | |
| { | |
| "epoch": 2.3503569467325645, | |
| "grad_norm": 0.04218167066574097, | |
| "learning_rate": 7.231270358306188e-05, | |
| "loss": 0.5927, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 2.352553542009885, | |
| "grad_norm": 0.04816456139087677, | |
| "learning_rate": 7.206840390879479e-05, | |
| "loss": 0.6595, | |
| "step": 1071 | |
| }, | |
| { | |
| "epoch": 2.354750137287205, | |
| "grad_norm": 0.04421921819448471, | |
| "learning_rate": 7.182410423452768e-05, | |
| "loss": 0.5484, | |
| "step": 1072 | |
| }, | |
| { | |
| "epoch": 2.356946732564525, | |
| "grad_norm": 0.04146098718047142, | |
| "learning_rate": 7.157980456026058e-05, | |
| "loss": 0.5741, | |
| "step": 1073 | |
| }, | |
| { | |
| "epoch": 2.359143327841845, | |
| "grad_norm": 0.045608628541231155, | |
| "learning_rate": 7.133550488599348e-05, | |
| "loss": 0.6229, | |
| "step": 1074 | |
| }, | |
| { | |
| "epoch": 2.361339923119165, | |
| "grad_norm": 0.04145771637558937, | |
| "learning_rate": 7.109120521172637e-05, | |
| "loss": 0.74, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 2.3635365183964856, | |
| "grad_norm": 0.05360572785139084, | |
| "learning_rate": 7.084690553745927e-05, | |
| "loss": 0.7087, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 2.3657331136738056, | |
| "grad_norm": 0.051699042320251465, | |
| "learning_rate": 7.060260586319218e-05, | |
| "loss": 0.6353, | |
| "step": 1077 | |
| }, | |
| { | |
| "epoch": 2.3679297089511255, | |
| "grad_norm": 0.05747654661536217, | |
| "learning_rate": 7.035830618892508e-05, | |
| "loss": 0.7254, | |
| "step": 1078 | |
| }, | |
| { | |
| "epoch": 2.370126304228446, | |
| "grad_norm": 0.04842827096581459, | |
| "learning_rate": 7.011400651465797e-05, | |
| "loss": 0.5057, | |
| "step": 1079 | |
| }, | |
| { | |
| "epoch": 2.372322899505766, | |
| "grad_norm": 0.056867651641368866, | |
| "learning_rate": 6.986970684039087e-05, | |
| "loss": 0.6912, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 2.3745194947830863, | |
| "grad_norm": 0.06867987662553787, | |
| "learning_rate": 6.962540716612377e-05, | |
| "loss": 0.9034, | |
| "step": 1081 | |
| }, | |
| { | |
| "epoch": 2.3767160900604063, | |
| "grad_norm": 0.044970154762268066, | |
| "learning_rate": 6.938110749185668e-05, | |
| "loss": 0.5275, | |
| "step": 1082 | |
| }, | |
| { | |
| "epoch": 2.3789126853377267, | |
| "grad_norm": 0.04874912649393082, | |
| "learning_rate": 6.913680781758956e-05, | |
| "loss": 0.6088, | |
| "step": 1083 | |
| }, | |
| { | |
| "epoch": 2.3811092806150467, | |
| "grad_norm": 0.06273435056209564, | |
| "learning_rate": 6.889250814332247e-05, | |
| "loss": 0.7871, | |
| "step": 1084 | |
| }, | |
| { | |
| "epoch": 2.3833058758923666, | |
| "grad_norm": 0.06409198045730591, | |
| "learning_rate": 6.864820846905537e-05, | |
| "loss": 0.8235, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 2.385502471169687, | |
| "grad_norm": 0.04517259821295738, | |
| "learning_rate": 6.840390879478826e-05, | |
| "loss": 0.5531, | |
| "step": 1086 | |
| }, | |
| { | |
| "epoch": 2.387699066447007, | |
| "grad_norm": 0.06406451761722565, | |
| "learning_rate": 6.815960912052116e-05, | |
| "loss": 0.8643, | |
| "step": 1087 | |
| }, | |
| { | |
| "epoch": 2.3898956617243274, | |
| "grad_norm": 0.06429564207792282, | |
| "learning_rate": 6.791530944625406e-05, | |
| "loss": 0.5985, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 2.3920922570016474, | |
| "grad_norm": 0.058752164244651794, | |
| "learning_rate": 6.767100977198697e-05, | |
| "loss": 0.7691, | |
| "step": 1089 | |
| }, | |
| { | |
| "epoch": 2.394288852278968, | |
| "grad_norm": 0.04136451706290245, | |
| "learning_rate": 6.742671009771986e-05, | |
| "loss": 0.614, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 2.396485447556288, | |
| "grad_norm": 0.06292608380317688, | |
| "learning_rate": 6.718241042345276e-05, | |
| "loss": 0.6292, | |
| "step": 1091 | |
| }, | |
| { | |
| "epoch": 2.3986820428336078, | |
| "grad_norm": 0.06885907799005508, | |
| "learning_rate": 6.693811074918566e-05, | |
| "loss": 0.6684, | |
| "step": 1092 | |
| }, | |
| { | |
| "epoch": 2.400878638110928, | |
| "grad_norm": 0.04652557522058487, | |
| "learning_rate": 6.669381107491855e-05, | |
| "loss": 0.6725, | |
| "step": 1093 | |
| }, | |
| { | |
| "epoch": 2.403075233388248, | |
| "grad_norm": 0.03902400657534599, | |
| "learning_rate": 6.644951140065145e-05, | |
| "loss": 0.6752, | |
| "step": 1094 | |
| }, | |
| { | |
| "epoch": 2.4052718286655685, | |
| "grad_norm": 0.05414823442697525, | |
| "learning_rate": 6.620521172638436e-05, | |
| "loss": 0.603, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 2.4074684239428885, | |
| "grad_norm": 0.04311055690050125, | |
| "learning_rate": 6.596091205211726e-05, | |
| "loss": 0.5174, | |
| "step": 1096 | |
| }, | |
| { | |
| "epoch": 2.409665019220209, | |
| "grad_norm": 0.05905736982822418, | |
| "learning_rate": 6.571661237785015e-05, | |
| "loss": 0.7347, | |
| "step": 1097 | |
| }, | |
| { | |
| "epoch": 2.411861614497529, | |
| "grad_norm": 0.04053014516830444, | |
| "learning_rate": 6.547231270358307e-05, | |
| "loss": 0.6997, | |
| "step": 1098 | |
| }, | |
| { | |
| "epoch": 2.414058209774849, | |
| "grad_norm": 0.053303156048059464, | |
| "learning_rate": 6.522801302931595e-05, | |
| "loss": 0.7643, | |
| "step": 1099 | |
| }, | |
| { | |
| "epoch": 2.4162548050521693, | |
| "grad_norm": 0.047551583498716354, | |
| "learning_rate": 6.498371335504886e-05, | |
| "loss": 0.6913, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 2.4184514003294892, | |
| "grad_norm": 0.04742944985628128, | |
| "learning_rate": 6.473941368078176e-05, | |
| "loss": 0.5594, | |
| "step": 1101 | |
| }, | |
| { | |
| "epoch": 2.420647995606809, | |
| "grad_norm": 0.04457717761397362, | |
| "learning_rate": 6.449511400651466e-05, | |
| "loss": 0.7067, | |
| "step": 1102 | |
| }, | |
| { | |
| "epoch": 2.4228445908841296, | |
| "grad_norm": 0.05029049888253212, | |
| "learning_rate": 6.425081433224755e-05, | |
| "loss": 0.6744, | |
| "step": 1103 | |
| }, | |
| { | |
| "epoch": 2.4250411861614496, | |
| "grad_norm": 0.057819634675979614, | |
| "learning_rate": 6.400651465798045e-05, | |
| "loss": 0.7409, | |
| "step": 1104 | |
| }, | |
| { | |
| "epoch": 2.42723778143877, | |
| "grad_norm": 0.06167594715952873, | |
| "learning_rate": 6.376221498371336e-05, | |
| "loss": 0.6952, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 2.42943437671609, | |
| "grad_norm": 0.04718726500868797, | |
| "learning_rate": 6.351791530944625e-05, | |
| "loss": 0.5742, | |
| "step": 1106 | |
| }, | |
| { | |
| "epoch": 2.4316309719934104, | |
| "grad_norm": 0.045384399592876434, | |
| "learning_rate": 6.327361563517915e-05, | |
| "loss": 0.7108, | |
| "step": 1107 | |
| }, | |
| { | |
| "epoch": 2.4338275672707304, | |
| "grad_norm": 0.048894964158535004, | |
| "learning_rate": 6.302931596091205e-05, | |
| "loss": 0.7631, | |
| "step": 1108 | |
| }, | |
| { | |
| "epoch": 2.4360241625480503, | |
| "grad_norm": 0.0457146130502224, | |
| "learning_rate": 6.278501628664495e-05, | |
| "loss": 0.6816, | |
| "step": 1109 | |
| }, | |
| { | |
| "epoch": 2.4382207578253707, | |
| "grad_norm": 0.04110351949930191, | |
| "learning_rate": 6.254071661237784e-05, | |
| "loss": 0.676, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 2.4404173531026907, | |
| "grad_norm": 0.04073873534798622, | |
| "learning_rate": 6.229641693811075e-05, | |
| "loss": 0.6479, | |
| "step": 1111 | |
| }, | |
| { | |
| "epoch": 2.442613948380011, | |
| "grad_norm": 0.054913751780986786, | |
| "learning_rate": 6.205211726384365e-05, | |
| "loss": 0.4896, | |
| "step": 1112 | |
| }, | |
| { | |
| "epoch": 2.444810543657331, | |
| "grad_norm": 0.05325063318014145, | |
| "learning_rate": 6.180781758957654e-05, | |
| "loss": 0.8076, | |
| "step": 1113 | |
| }, | |
| { | |
| "epoch": 2.4470071389346515, | |
| "grad_norm": 0.05473218858242035, | |
| "learning_rate": 6.156351791530944e-05, | |
| "loss": 0.8352, | |
| "step": 1114 | |
| }, | |
| { | |
| "epoch": 2.4492037342119715, | |
| "grad_norm": 0.046617619693279266, | |
| "learning_rate": 6.131921824104234e-05, | |
| "loss": 0.575, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 2.4514003294892914, | |
| "grad_norm": 0.0457211434841156, | |
| "learning_rate": 6.107491856677525e-05, | |
| "loss": 0.5843, | |
| "step": 1116 | |
| }, | |
| { | |
| "epoch": 2.453596924766612, | |
| "grad_norm": 0.05672460421919823, | |
| "learning_rate": 6.083061889250814e-05, | |
| "loss": 0.6063, | |
| "step": 1117 | |
| }, | |
| { | |
| "epoch": 2.455793520043932, | |
| "grad_norm": 0.04907252639532089, | |
| "learning_rate": 6.058631921824104e-05, | |
| "loss": 0.5895, | |
| "step": 1118 | |
| }, | |
| { | |
| "epoch": 2.4579901153212522, | |
| "grad_norm": 0.04526342824101448, | |
| "learning_rate": 6.0342019543973935e-05, | |
| "loss": 0.7618, | |
| "step": 1119 | |
| }, | |
| { | |
| "epoch": 2.460186710598572, | |
| "grad_norm": 0.045307356864213943, | |
| "learning_rate": 6.009771986970684e-05, | |
| "loss": 0.6353, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 2.4623833058758926, | |
| "grad_norm": 0.05209842324256897, | |
| "learning_rate": 5.9853420195439734e-05, | |
| "loss": 0.6152, | |
| "step": 1121 | |
| }, | |
| { | |
| "epoch": 2.4645799011532126, | |
| "grad_norm": 0.06889799982309341, | |
| "learning_rate": 5.9609120521172636e-05, | |
| "loss": 1.0135, | |
| "step": 1122 | |
| }, | |
| { | |
| "epoch": 2.4667764964305325, | |
| "grad_norm": 0.05798589438199997, | |
| "learning_rate": 5.936482084690553e-05, | |
| "loss": 0.6518, | |
| "step": 1123 | |
| }, | |
| { | |
| "epoch": 2.468973091707853, | |
| "grad_norm": 0.04932552948594093, | |
| "learning_rate": 5.9120521172638435e-05, | |
| "loss": 0.6739, | |
| "step": 1124 | |
| }, | |
| { | |
| "epoch": 2.471169686985173, | |
| "grad_norm": 0.04319072514772415, | |
| "learning_rate": 5.887622149837133e-05, | |
| "loss": 0.5889, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 2.473366282262493, | |
| "grad_norm": 0.054561588913202286, | |
| "learning_rate": 5.863192182410423e-05, | |
| "loss": 0.7459, | |
| "step": 1126 | |
| }, | |
| { | |
| "epoch": 2.4755628775398133, | |
| "grad_norm": 0.05109608545899391, | |
| "learning_rate": 5.838762214983713e-05, | |
| "loss": 0.5898, | |
| "step": 1127 | |
| }, | |
| { | |
| "epoch": 2.4777594728171333, | |
| "grad_norm": 0.030667340382933617, | |
| "learning_rate": 5.8143322475570026e-05, | |
| "loss": 0.6168, | |
| "step": 1128 | |
| }, | |
| { | |
| "epoch": 2.4799560680944537, | |
| "grad_norm": 0.050101760774850845, | |
| "learning_rate": 5.789902280130293e-05, | |
| "loss": 0.8355, | |
| "step": 1129 | |
| }, | |
| { | |
| "epoch": 2.4821526633717736, | |
| "grad_norm": 0.049108896404504776, | |
| "learning_rate": 5.7654723127035825e-05, | |
| "loss": 0.6122, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 2.484349258649094, | |
| "grad_norm": 0.0410931296646595, | |
| "learning_rate": 5.741042345276873e-05, | |
| "loss": 0.6199, | |
| "step": 1131 | |
| }, | |
| { | |
| "epoch": 2.486545853926414, | |
| "grad_norm": 0.05107707902789116, | |
| "learning_rate": 5.716612377850162e-05, | |
| "loss": 0.692, | |
| "step": 1132 | |
| }, | |
| { | |
| "epoch": 2.488742449203734, | |
| "grad_norm": 0.05445995554327965, | |
| "learning_rate": 5.6921824104234526e-05, | |
| "loss": 0.5391, | |
| "step": 1133 | |
| }, | |
| { | |
| "epoch": 2.4909390444810544, | |
| "grad_norm": 0.05041612312197685, | |
| "learning_rate": 5.667752442996742e-05, | |
| "loss": 0.6833, | |
| "step": 1134 | |
| }, | |
| { | |
| "epoch": 2.4931356397583744, | |
| "grad_norm": 0.05433562770485878, | |
| "learning_rate": 5.643322475570032e-05, | |
| "loss": 0.6777, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 2.495332235035695, | |
| "grad_norm": 0.045349642634391785, | |
| "learning_rate": 5.618892508143322e-05, | |
| "loss": 0.5513, | |
| "step": 1136 | |
| }, | |
| { | |
| "epoch": 2.4975288303130148, | |
| "grad_norm": 0.04080528765916824, | |
| "learning_rate": 5.594462540716612e-05, | |
| "loss": 0.4565, | |
| "step": 1137 | |
| }, | |
| { | |
| "epoch": 2.499725425590335, | |
| "grad_norm": 0.05346320569515228, | |
| "learning_rate": 5.570032573289902e-05, | |
| "loss": 0.6752, | |
| "step": 1138 | |
| }, | |
| { | |
| "epoch": 2.501922020867655, | |
| "grad_norm": 0.04871304705739021, | |
| "learning_rate": 5.5456026058631915e-05, | |
| "loss": 0.7841, | |
| "step": 1139 | |
| }, | |
| { | |
| "epoch": 2.504118616144975, | |
| "grad_norm": 0.08751735091209412, | |
| "learning_rate": 5.521172638436482e-05, | |
| "loss": 0.5952, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 2.5063152114222955, | |
| "grad_norm": 0.033952705562114716, | |
| "learning_rate": 5.4967426710097714e-05, | |
| "loss": 0.5781, | |
| "step": 1141 | |
| }, | |
| { | |
| "epoch": 2.5085118066996155, | |
| "grad_norm": 0.04253567382693291, | |
| "learning_rate": 5.472312703583061e-05, | |
| "loss": 0.7076, | |
| "step": 1142 | |
| }, | |
| { | |
| "epoch": 2.510708401976936, | |
| "grad_norm": 0.04273783415555954, | |
| "learning_rate": 5.447882736156351e-05, | |
| "loss": 0.7619, | |
| "step": 1143 | |
| }, | |
| { | |
| "epoch": 2.512904997254256, | |
| "grad_norm": 0.041224155575037, | |
| "learning_rate": 5.423452768729641e-05, | |
| "loss": 0.6162, | |
| "step": 1144 | |
| }, | |
| { | |
| "epoch": 2.5151015925315763, | |
| "grad_norm": 0.04299287870526314, | |
| "learning_rate": 5.399022801302931e-05, | |
| "loss": 0.6113, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 2.5172981878088962, | |
| "grad_norm": 0.06844445317983627, | |
| "learning_rate": 5.374592833876221e-05, | |
| "loss": 0.6811, | |
| "step": 1146 | |
| }, | |
| { | |
| "epoch": 2.519494783086216, | |
| "grad_norm": 0.04922829195857048, | |
| "learning_rate": 5.350162866449511e-05, | |
| "loss": 0.6346, | |
| "step": 1147 | |
| }, | |
| { | |
| "epoch": 2.5216913783635366, | |
| "grad_norm": 0.04479237645864487, | |
| "learning_rate": 5.3257328990228006e-05, | |
| "loss": 0.5057, | |
| "step": 1148 | |
| }, | |
| { | |
| "epoch": 2.5238879736408566, | |
| "grad_norm": 0.05080297216773033, | |
| "learning_rate": 5.301302931596091e-05, | |
| "loss": 0.694, | |
| "step": 1149 | |
| }, | |
| { | |
| "epoch": 2.5260845689181766, | |
| "grad_norm": 0.04080790653824806, | |
| "learning_rate": 5.2768729641693805e-05, | |
| "loss": 0.8815, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 2.528281164195497, | |
| "grad_norm": 0.05466169863939285, | |
| "learning_rate": 5.25244299674267e-05, | |
| "loss": 0.663, | |
| "step": 1151 | |
| }, | |
| { | |
| "epoch": 2.5304777594728174, | |
| "grad_norm": 0.05815236642956734, | |
| "learning_rate": 5.2280130293159604e-05, | |
| "loss": 0.6391, | |
| "step": 1152 | |
| }, | |
| { | |
| "epoch": 2.5326743547501374, | |
| "grad_norm": 0.043016865849494934, | |
| "learning_rate": 5.20358306188925e-05, | |
| "loss": 0.6724, | |
| "step": 1153 | |
| }, | |
| { | |
| "epoch": 2.5348709500274573, | |
| "grad_norm": 0.03767896816134453, | |
| "learning_rate": 5.17915309446254e-05, | |
| "loss": 0.6149, | |
| "step": 1154 | |
| }, | |
| { | |
| "epoch": 2.5370675453047777, | |
| "grad_norm": 0.04087197780609131, | |
| "learning_rate": 5.15472312703583e-05, | |
| "loss": 0.5995, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 2.5392641405820977, | |
| "grad_norm": 0.055968958884477615, | |
| "learning_rate": 5.13029315960912e-05, | |
| "loss": 0.7466, | |
| "step": 1156 | |
| }, | |
| { | |
| "epoch": 2.5414607358594177, | |
| "grad_norm": 0.0782134160399437, | |
| "learning_rate": 5.10586319218241e-05, | |
| "loss": 0.7928, | |
| "step": 1157 | |
| }, | |
| { | |
| "epoch": 2.543657331136738, | |
| "grad_norm": 0.04652741923928261, | |
| "learning_rate": 5.081433224755699e-05, | |
| "loss": 0.5977, | |
| "step": 1158 | |
| }, | |
| { | |
| "epoch": 2.545853926414058, | |
| "grad_norm": 0.04434973746538162, | |
| "learning_rate": 5.0570032573289896e-05, | |
| "loss": 0.5646, | |
| "step": 1159 | |
| }, | |
| { | |
| "epoch": 2.5480505216913785, | |
| "grad_norm": 0.042396001517772675, | |
| "learning_rate": 5.032573289902279e-05, | |
| "loss": 0.625, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 2.5502471169686984, | |
| "grad_norm": 0.04982843995094299, | |
| "learning_rate": 5.0081433224755695e-05, | |
| "loss": 0.5383, | |
| "step": 1161 | |
| }, | |
| { | |
| "epoch": 2.552443712246019, | |
| "grad_norm": 0.04226350039243698, | |
| "learning_rate": 4.983713355048859e-05, | |
| "loss": 0.683, | |
| "step": 1162 | |
| }, | |
| { | |
| "epoch": 2.554640307523339, | |
| "grad_norm": 0.05881736800074577, | |
| "learning_rate": 4.9592833876221493e-05, | |
| "loss": 0.5796, | |
| "step": 1163 | |
| }, | |
| { | |
| "epoch": 2.556836902800659, | |
| "grad_norm": 0.04545532539486885, | |
| "learning_rate": 4.934853420195439e-05, | |
| "loss": 0.7102, | |
| "step": 1164 | |
| }, | |
| { | |
| "epoch": 2.559033498077979, | |
| "grad_norm": 0.04276625066995621, | |
| "learning_rate": 4.910423452768729e-05, | |
| "loss": 0.6821, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 2.561230093355299, | |
| "grad_norm": 0.04249290004372597, | |
| "learning_rate": 4.885993485342019e-05, | |
| "loss": 0.6215, | |
| "step": 1166 | |
| }, | |
| { | |
| "epoch": 2.5634266886326196, | |
| "grad_norm": 0.04267478361725807, | |
| "learning_rate": 4.8615635179153084e-05, | |
| "loss": 0.5601, | |
| "step": 1167 | |
| }, | |
| { | |
| "epoch": 2.5656232839099395, | |
| "grad_norm": 0.05035392940044403, | |
| "learning_rate": 4.837133550488599e-05, | |
| "loss": 0.6485, | |
| "step": 1168 | |
| }, | |
| { | |
| "epoch": 2.56781987918726, | |
| "grad_norm": 0.04392477497458458, | |
| "learning_rate": 4.812703583061888e-05, | |
| "loss": 0.6361, | |
| "step": 1169 | |
| }, | |
| { | |
| "epoch": 2.57001647446458, | |
| "grad_norm": 0.044216789305210114, | |
| "learning_rate": 4.7882736156351786e-05, | |
| "loss": 0.6166, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 2.5722130697419, | |
| "grad_norm": 0.04420490562915802, | |
| "learning_rate": 4.763843648208468e-05, | |
| "loss": 0.5684, | |
| "step": 1171 | |
| }, | |
| { | |
| "epoch": 2.5744096650192203, | |
| "grad_norm": 0.09682516753673553, | |
| "learning_rate": 4.7394136807817584e-05, | |
| "loss": 0.6564, | |
| "step": 1172 | |
| }, | |
| { | |
| "epoch": 2.5766062602965403, | |
| "grad_norm": 0.04140635207295418, | |
| "learning_rate": 4.714983713355048e-05, | |
| "loss": 0.6094, | |
| "step": 1173 | |
| }, | |
| { | |
| "epoch": 2.5788028555738602, | |
| "grad_norm": 0.04643336683511734, | |
| "learning_rate": 4.6905537459283376e-05, | |
| "loss": 0.7015, | |
| "step": 1174 | |
| }, | |
| { | |
| "epoch": 2.5809994508511807, | |
| "grad_norm": 0.05290674418210983, | |
| "learning_rate": 4.6661237785016286e-05, | |
| "loss": 0.6969, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 2.583196046128501, | |
| "grad_norm": 0.047984324395656586, | |
| "learning_rate": 4.641693811074919e-05, | |
| "loss": 0.5882, | |
| "step": 1176 | |
| }, | |
| { | |
| "epoch": 2.585392641405821, | |
| "grad_norm": 0.055153653025627136, | |
| "learning_rate": 4.6172638436482085e-05, | |
| "loss": 0.8228, | |
| "step": 1177 | |
| }, | |
| { | |
| "epoch": 2.587589236683141, | |
| "grad_norm": 0.049381084740161896, | |
| "learning_rate": 4.592833876221498e-05, | |
| "loss": 0.596, | |
| "step": 1178 | |
| }, | |
| { | |
| "epoch": 2.5897858319604614, | |
| "grad_norm": 0.07250449061393738, | |
| "learning_rate": 4.568403908794788e-05, | |
| "loss": 0.7352, | |
| "step": 1179 | |
| }, | |
| { | |
| "epoch": 2.5919824272377814, | |
| "grad_norm": 0.04308245703577995, | |
| "learning_rate": 4.543973941368078e-05, | |
| "loss": 0.571, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 2.5941790225151014, | |
| "grad_norm": 0.05505221337080002, | |
| "learning_rate": 4.519543973941368e-05, | |
| "loss": 0.6903, | |
| "step": 1181 | |
| }, | |
| { | |
| "epoch": 2.5963756177924218, | |
| "grad_norm": 0.06684863567352295, | |
| "learning_rate": 4.495114006514658e-05, | |
| "loss": 0.7423, | |
| "step": 1182 | |
| }, | |
| { | |
| "epoch": 2.598572213069742, | |
| "grad_norm": 0.04111980274319649, | |
| "learning_rate": 4.470684039087948e-05, | |
| "loss": 0.6431, | |
| "step": 1183 | |
| }, | |
| { | |
| "epoch": 2.600768808347062, | |
| "grad_norm": 0.04753178358078003, | |
| "learning_rate": 4.446254071661238e-05, | |
| "loss": 0.6756, | |
| "step": 1184 | |
| }, | |
| { | |
| "epoch": 2.602965403624382, | |
| "grad_norm": 0.043900731950998306, | |
| "learning_rate": 4.421824104234527e-05, | |
| "loss": 0.6985, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 2.6051619989017025, | |
| "grad_norm": 0.03854718059301376, | |
| "learning_rate": 4.3973941368078175e-05, | |
| "loss": 0.7089, | |
| "step": 1186 | |
| }, | |
| { | |
| "epoch": 2.6073585941790225, | |
| "grad_norm": 0.0341448150575161, | |
| "learning_rate": 4.372964169381107e-05, | |
| "loss": 0.5652, | |
| "step": 1187 | |
| }, | |
| { | |
| "epoch": 2.6095551894563425, | |
| "grad_norm": 0.04500808194279671, | |
| "learning_rate": 4.3485342019543974e-05, | |
| "loss": 0.6157, | |
| "step": 1188 | |
| }, | |
| { | |
| "epoch": 2.611751784733663, | |
| "grad_norm": 0.050450973212718964, | |
| "learning_rate": 4.324104234527687e-05, | |
| "loss": 0.5892, | |
| "step": 1189 | |
| }, | |
| { | |
| "epoch": 2.613948380010983, | |
| "grad_norm": 0.04859516769647598, | |
| "learning_rate": 4.299674267100977e-05, | |
| "loss": 0.6381, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 2.6161449752883033, | |
| "grad_norm": 0.041921671479940414, | |
| "learning_rate": 4.275244299674267e-05, | |
| "loss": 0.6333, | |
| "step": 1191 | |
| }, | |
| { | |
| "epoch": 2.618341570565623, | |
| "grad_norm": 0.05008811876177788, | |
| "learning_rate": 4.250814332247557e-05, | |
| "loss": 0.5237, | |
| "step": 1192 | |
| }, | |
| { | |
| "epoch": 2.6205381658429436, | |
| "grad_norm": 0.10139350593090057, | |
| "learning_rate": 4.226384364820847e-05, | |
| "loss": 1.0228, | |
| "step": 1193 | |
| }, | |
| { | |
| "epoch": 2.6227347611202636, | |
| "grad_norm": 0.06467008590698242, | |
| "learning_rate": 4.2019543973941364e-05, | |
| "loss": 0.7578, | |
| "step": 1194 | |
| }, | |
| { | |
| "epoch": 2.6249313563975836, | |
| "grad_norm": 0.030447738245129585, | |
| "learning_rate": 4.1775244299674266e-05, | |
| "loss": 0.4413, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 2.627127951674904, | |
| "grad_norm": 0.05087581276893616, | |
| "learning_rate": 4.153094462540716e-05, | |
| "loss": 0.6297, | |
| "step": 1196 | |
| }, | |
| { | |
| "epoch": 2.629324546952224, | |
| "grad_norm": 0.11560789495706558, | |
| "learning_rate": 4.1286644951140065e-05, | |
| "loss": 0.6206, | |
| "step": 1197 | |
| }, | |
| { | |
| "epoch": 2.6315211422295444, | |
| "grad_norm": 0.04665865749120712, | |
| "learning_rate": 4.104234527687296e-05, | |
| "loss": 0.6543, | |
| "step": 1198 | |
| }, | |
| { | |
| "epoch": 2.6337177375068643, | |
| "grad_norm": 0.04100664332509041, | |
| "learning_rate": 4.0798045602605864e-05, | |
| "loss": 0.5507, | |
| "step": 1199 | |
| }, | |
| { | |
| "epoch": 2.6359143327841847, | |
| "grad_norm": 0.05304735153913498, | |
| "learning_rate": 4.055374592833876e-05, | |
| "loss": 0.6102, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 2.6381109280615047, | |
| "grad_norm": 0.04621303454041481, | |
| "learning_rate": 4.0309446254071656e-05, | |
| "loss": 0.5275, | |
| "step": 1201 | |
| }, | |
| { | |
| "epoch": 2.6403075233388247, | |
| "grad_norm": 0.04916003346443176, | |
| "learning_rate": 4.006514657980456e-05, | |
| "loss": 0.5175, | |
| "step": 1202 | |
| }, | |
| { | |
| "epoch": 2.642504118616145, | |
| "grad_norm": 0.056979499757289886, | |
| "learning_rate": 3.9820846905537454e-05, | |
| "loss": 0.8422, | |
| "step": 1203 | |
| }, | |
| { | |
| "epoch": 2.644700713893465, | |
| "grad_norm": 0.05574516952037811, | |
| "learning_rate": 3.957654723127036e-05, | |
| "loss": 0.635, | |
| "step": 1204 | |
| }, | |
| { | |
| "epoch": 2.646897309170785, | |
| "grad_norm": 0.04193108156323433, | |
| "learning_rate": 3.933224755700325e-05, | |
| "loss": 0.6862, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 2.6490939044481054, | |
| "grad_norm": 0.04623300954699516, | |
| "learning_rate": 3.9087947882736156e-05, | |
| "loss": 0.7349, | |
| "step": 1206 | |
| }, | |
| { | |
| "epoch": 2.651290499725426, | |
| "grad_norm": 0.06209337338805199, | |
| "learning_rate": 3.884364820846905e-05, | |
| "loss": 0.7061, | |
| "step": 1207 | |
| }, | |
| { | |
| "epoch": 2.653487095002746, | |
| "grad_norm": 0.05206606909632683, | |
| "learning_rate": 3.8599348534201955e-05, | |
| "loss": 0.6362, | |
| "step": 1208 | |
| }, | |
| { | |
| "epoch": 2.655683690280066, | |
| "grad_norm": 0.06646497547626495, | |
| "learning_rate": 3.835504885993485e-05, | |
| "loss": 0.5787, | |
| "step": 1209 | |
| }, | |
| { | |
| "epoch": 2.657880285557386, | |
| "grad_norm": 0.04821066930890083, | |
| "learning_rate": 3.8110749185667747e-05, | |
| "loss": 0.8365, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 2.660076880834706, | |
| "grad_norm": 0.04882679879665375, | |
| "learning_rate": 3.786644951140065e-05, | |
| "loss": 0.527, | |
| "step": 1211 | |
| }, | |
| { | |
| "epoch": 2.662273476112026, | |
| "grad_norm": 0.05172707512974739, | |
| "learning_rate": 3.7622149837133545e-05, | |
| "loss": 0.593, | |
| "step": 1212 | |
| }, | |
| { | |
| "epoch": 2.6644700713893466, | |
| "grad_norm": 0.04725964367389679, | |
| "learning_rate": 3.737785016286645e-05, | |
| "loss": 0.7141, | |
| "step": 1213 | |
| }, | |
| { | |
| "epoch": 2.6666666666666665, | |
| "grad_norm": 0.0511508584022522, | |
| "learning_rate": 3.7133550488599344e-05, | |
| "loss": 0.5239, | |
| "step": 1214 | |
| }, | |
| { | |
| "epoch": 2.668863261943987, | |
| "grad_norm": 0.045084815472364426, | |
| "learning_rate": 3.688925081433225e-05, | |
| "loss": 0.6523, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 2.671059857221307, | |
| "grad_norm": 0.04477803781628609, | |
| "learning_rate": 3.664495114006514e-05, | |
| "loss": 0.6102, | |
| "step": 1216 | |
| }, | |
| { | |
| "epoch": 2.6732564524986273, | |
| "grad_norm": 0.03817501664161682, | |
| "learning_rate": 3.640065146579804e-05, | |
| "loss": 0.6925, | |
| "step": 1217 | |
| }, | |
| { | |
| "epoch": 2.6754530477759473, | |
| "grad_norm": 0.04120345041155815, | |
| "learning_rate": 3.615635179153094e-05, | |
| "loss": 0.6503, | |
| "step": 1218 | |
| }, | |
| { | |
| "epoch": 2.6776496430532672, | |
| "grad_norm": 0.05116845294833183, | |
| "learning_rate": 3.591205211726384e-05, | |
| "loss": 0.6252, | |
| "step": 1219 | |
| }, | |
| { | |
| "epoch": 2.6798462383305877, | |
| "grad_norm": 0.048078861087560654, | |
| "learning_rate": 3.566775244299674e-05, | |
| "loss": 0.6434, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 2.6820428336079076, | |
| "grad_norm": 0.050224531441926956, | |
| "learning_rate": 3.5423452768729636e-05, | |
| "loss": 0.7836, | |
| "step": 1221 | |
| }, | |
| { | |
| "epoch": 2.684239428885228, | |
| "grad_norm": 0.04820210486650467, | |
| "learning_rate": 3.517915309446254e-05, | |
| "loss": 0.4624, | |
| "step": 1222 | |
| }, | |
| { | |
| "epoch": 2.686436024162548, | |
| "grad_norm": 0.04261908680200577, | |
| "learning_rate": 3.4934853420195435e-05, | |
| "loss": 0.6753, | |
| "step": 1223 | |
| }, | |
| { | |
| "epoch": 2.6886326194398684, | |
| "grad_norm": 0.05452783405780792, | |
| "learning_rate": 3.469055374592834e-05, | |
| "loss": 0.7181, | |
| "step": 1224 | |
| }, | |
| { | |
| "epoch": 2.6908292147171884, | |
| "grad_norm": 0.038788363337516785, | |
| "learning_rate": 3.4446254071661234e-05, | |
| "loss": 0.683, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 2.6930258099945084, | |
| "grad_norm": 0.0443214513361454, | |
| "learning_rate": 3.420195439739413e-05, | |
| "loss": 0.5505, | |
| "step": 1226 | |
| }, | |
| { | |
| "epoch": 2.6952224052718288, | |
| "grad_norm": 0.07653029263019562, | |
| "learning_rate": 3.395765472312703e-05, | |
| "loss": 0.7052, | |
| "step": 1227 | |
| }, | |
| { | |
| "epoch": 2.6974190005491487, | |
| "grad_norm": 0.03736657649278641, | |
| "learning_rate": 3.371335504885993e-05, | |
| "loss": 0.5925, | |
| "step": 1228 | |
| }, | |
| { | |
| "epoch": 2.6996155958264687, | |
| "grad_norm": 0.049569420516490936, | |
| "learning_rate": 3.346905537459283e-05, | |
| "loss": 0.7129, | |
| "step": 1229 | |
| }, | |
| { | |
| "epoch": 2.701812191103789, | |
| "grad_norm": 0.04310569539666176, | |
| "learning_rate": 3.322475570032573e-05, | |
| "loss": 0.5297, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 2.7040087863811095, | |
| "grad_norm": 0.032269321382045746, | |
| "learning_rate": 3.298045602605863e-05, | |
| "loss": 0.5152, | |
| "step": 1231 | |
| }, | |
| { | |
| "epoch": 2.7062053816584295, | |
| "grad_norm": 0.04561099037528038, | |
| "learning_rate": 3.273615635179153e-05, | |
| "loss": 0.7281, | |
| "step": 1232 | |
| }, | |
| { | |
| "epoch": 2.7084019769357495, | |
| "grad_norm": 0.045137908309698105, | |
| "learning_rate": 3.249185667752443e-05, | |
| "loss": 0.6058, | |
| "step": 1233 | |
| }, | |
| { | |
| "epoch": 2.71059857221307, | |
| "grad_norm": 0.04589609429240227, | |
| "learning_rate": 3.224755700325733e-05, | |
| "loss": 0.662, | |
| "step": 1234 | |
| }, | |
| { | |
| "epoch": 2.71279516749039, | |
| "grad_norm": 0.03858232870697975, | |
| "learning_rate": 3.200325732899023e-05, | |
| "loss": 0.745, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 2.71499176276771, | |
| "grad_norm": 0.06444483250379562, | |
| "learning_rate": 3.175895765472312e-05, | |
| "loss": 0.7367, | |
| "step": 1236 | |
| }, | |
| { | |
| "epoch": 2.7171883580450302, | |
| "grad_norm": 0.05342705175280571, | |
| "learning_rate": 3.1514657980456026e-05, | |
| "loss": 0.6469, | |
| "step": 1237 | |
| }, | |
| { | |
| "epoch": 2.71938495332235, | |
| "grad_norm": 0.05334605276584625, | |
| "learning_rate": 3.127035830618892e-05, | |
| "loss": 0.6761, | |
| "step": 1238 | |
| }, | |
| { | |
| "epoch": 2.7215815485996706, | |
| "grad_norm": 0.06340347975492477, | |
| "learning_rate": 3.1026058631921825e-05, | |
| "loss": 0.7222, | |
| "step": 1239 | |
| }, | |
| { | |
| "epoch": 2.7237781438769906, | |
| "grad_norm": 0.03992515802383423, | |
| "learning_rate": 3.078175895765472e-05, | |
| "loss": 0.6016, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 2.725974739154311, | |
| "grad_norm": 0.04782949388027191, | |
| "learning_rate": 3.0537459283387624e-05, | |
| "loss": 0.5176, | |
| "step": 1241 | |
| }, | |
| { | |
| "epoch": 2.728171334431631, | |
| "grad_norm": 0.04865286499261856, | |
| "learning_rate": 3.029315960912052e-05, | |
| "loss": 0.5885, | |
| "step": 1242 | |
| }, | |
| { | |
| "epoch": 2.730367929708951, | |
| "grad_norm": 0.04451761022210121, | |
| "learning_rate": 3.004885993485342e-05, | |
| "loss": 0.7112, | |
| "step": 1243 | |
| }, | |
| { | |
| "epoch": 2.7325645249862713, | |
| "grad_norm": 0.057043615728616714, | |
| "learning_rate": 2.9804560260586318e-05, | |
| "loss": 0.6835, | |
| "step": 1244 | |
| }, | |
| { | |
| "epoch": 2.7347611202635913, | |
| "grad_norm": 0.0446697361767292, | |
| "learning_rate": 2.9560260586319218e-05, | |
| "loss": 0.6343, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 2.7369577155409117, | |
| "grad_norm": 0.05795365571975708, | |
| "learning_rate": 2.9315960912052114e-05, | |
| "loss": 0.6624, | |
| "step": 1246 | |
| }, | |
| { | |
| "epoch": 2.7391543108182317, | |
| "grad_norm": 0.05948972329497337, | |
| "learning_rate": 2.9071661237785013e-05, | |
| "loss": 0.8657, | |
| "step": 1247 | |
| }, | |
| { | |
| "epoch": 2.741350906095552, | |
| "grad_norm": 0.045655228197574615, | |
| "learning_rate": 2.8827361563517912e-05, | |
| "loss": 0.623, | |
| "step": 1248 | |
| }, | |
| { | |
| "epoch": 2.743547501372872, | |
| "grad_norm": 0.056424789130687714, | |
| "learning_rate": 2.858306188925081e-05, | |
| "loss": 0.762, | |
| "step": 1249 | |
| }, | |
| { | |
| "epoch": 2.745744096650192, | |
| "grad_norm": 0.04328913614153862, | |
| "learning_rate": 2.833876221498371e-05, | |
| "loss": 0.6743, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 2.7479406919275124, | |
| "grad_norm": 0.043471962213516235, | |
| "learning_rate": 2.809446254071661e-05, | |
| "loss": 0.7574, | |
| "step": 1251 | |
| }, | |
| { | |
| "epoch": 2.7501372872048324, | |
| "grad_norm": 0.057528410106897354, | |
| "learning_rate": 2.785016286644951e-05, | |
| "loss": 0.6993, | |
| "step": 1252 | |
| }, | |
| { | |
| "epoch": 2.752333882482153, | |
| "grad_norm": 0.046321526169776917, | |
| "learning_rate": 2.760586319218241e-05, | |
| "loss": 0.5547, | |
| "step": 1253 | |
| }, | |
| { | |
| "epoch": 2.754530477759473, | |
| "grad_norm": 0.042450498789548874, | |
| "learning_rate": 2.7361563517915305e-05, | |
| "loss": 0.5814, | |
| "step": 1254 | |
| }, | |
| { | |
| "epoch": 2.756727073036793, | |
| "grad_norm": 0.06893017143011093, | |
| "learning_rate": 2.7117263843648204e-05, | |
| "loss": 0.7708, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 2.758923668314113, | |
| "grad_norm": 0.04471489042043686, | |
| "learning_rate": 2.6872964169381104e-05, | |
| "loss": 0.4695, | |
| "step": 1256 | |
| }, | |
| { | |
| "epoch": 2.761120263591433, | |
| "grad_norm": 0.051368217915296555, | |
| "learning_rate": 2.6628664495114003e-05, | |
| "loss": 0.6192, | |
| "step": 1257 | |
| }, | |
| { | |
| "epoch": 2.7633168588687536, | |
| "grad_norm": 0.04400815814733505, | |
| "learning_rate": 2.6384364820846903e-05, | |
| "loss": 0.5535, | |
| "step": 1258 | |
| }, | |
| { | |
| "epoch": 2.7655134541460735, | |
| "grad_norm": 0.05980153754353523, | |
| "learning_rate": 2.6140065146579802e-05, | |
| "loss": 0.7952, | |
| "step": 1259 | |
| }, | |
| { | |
| "epoch": 2.7677100494233935, | |
| "grad_norm": 0.04607835412025452, | |
| "learning_rate": 2.58957654723127e-05, | |
| "loss": 0.6893, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 2.769906644700714, | |
| "grad_norm": 0.05787874385714531, | |
| "learning_rate": 2.56514657980456e-05, | |
| "loss": 0.6042, | |
| "step": 1261 | |
| }, | |
| { | |
| "epoch": 2.7721032399780343, | |
| "grad_norm": 0.05453767254948616, | |
| "learning_rate": 2.5407166123778497e-05, | |
| "loss": 0.6614, | |
| "step": 1262 | |
| }, | |
| { | |
| "epoch": 2.7742998352553543, | |
| "grad_norm": 0.05665965378284454, | |
| "learning_rate": 2.5162866449511396e-05, | |
| "loss": 0.7342, | |
| "step": 1263 | |
| }, | |
| { | |
| "epoch": 2.7764964305326743, | |
| "grad_norm": 0.03880994766950607, | |
| "learning_rate": 2.4918566775244295e-05, | |
| "loss": 0.7974, | |
| "step": 1264 | |
| }, | |
| { | |
| "epoch": 2.7786930258099947, | |
| "grad_norm": 0.039176661521196365, | |
| "learning_rate": 2.4674267100977195e-05, | |
| "loss": 0.7382, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 2.7808896210873146, | |
| "grad_norm": 0.036031998693943024, | |
| "learning_rate": 2.4429967426710094e-05, | |
| "loss": 0.5444, | |
| "step": 1266 | |
| }, | |
| { | |
| "epoch": 2.7830862163646346, | |
| "grad_norm": 0.04688276723027229, | |
| "learning_rate": 2.4185667752442993e-05, | |
| "loss": 0.5756, | |
| "step": 1267 | |
| }, | |
| { | |
| "epoch": 2.785282811641955, | |
| "grad_norm": 0.04207608476281166, | |
| "learning_rate": 2.3941368078175893e-05, | |
| "loss": 0.6617, | |
| "step": 1268 | |
| }, | |
| { | |
| "epoch": 2.787479406919275, | |
| "grad_norm": 0.053132861852645874, | |
| "learning_rate": 2.3697068403908792e-05, | |
| "loss": 0.5698, | |
| "step": 1269 | |
| }, | |
| { | |
| "epoch": 2.7896760021965954, | |
| "grad_norm": 0.04987029731273651, | |
| "learning_rate": 2.3452768729641688e-05, | |
| "loss": 0.6068, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 2.7918725974739154, | |
| "grad_norm": 0.04091858118772507, | |
| "learning_rate": 2.3208469055374594e-05, | |
| "loss": 0.5744, | |
| "step": 1271 | |
| }, | |
| { | |
| "epoch": 2.7940691927512358, | |
| "grad_norm": 0.04796380549669266, | |
| "learning_rate": 2.296416938110749e-05, | |
| "loss": 0.6778, | |
| "step": 1272 | |
| }, | |
| { | |
| "epoch": 2.7962657880285557, | |
| "grad_norm": 0.05361456796526909, | |
| "learning_rate": 2.271986970684039e-05, | |
| "loss": 0.6834, | |
| "step": 1273 | |
| }, | |
| { | |
| "epoch": 2.7984623833058757, | |
| "grad_norm": 0.04959200695157051, | |
| "learning_rate": 2.247557003257329e-05, | |
| "loss": 0.8341, | |
| "step": 1274 | |
| }, | |
| { | |
| "epoch": 2.800658978583196, | |
| "grad_norm": 0.0428910106420517, | |
| "learning_rate": 2.223127035830619e-05, | |
| "loss": 0.6286, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 2.802855573860516, | |
| "grad_norm": 0.05645201355218887, | |
| "learning_rate": 2.1986970684039088e-05, | |
| "loss": 0.8027, | |
| "step": 1276 | |
| }, | |
| { | |
| "epoch": 2.8050521691378365, | |
| "grad_norm": 0.059787213802337646, | |
| "learning_rate": 2.1742671009771987e-05, | |
| "loss": 0.8865, | |
| "step": 1277 | |
| }, | |
| { | |
| "epoch": 2.8072487644151565, | |
| "grad_norm": 0.04704824462532997, | |
| "learning_rate": 2.1498371335504886e-05, | |
| "loss": 0.731, | |
| "step": 1278 | |
| }, | |
| { | |
| "epoch": 2.809445359692477, | |
| "grad_norm": 0.04552409425377846, | |
| "learning_rate": 2.1254071661237786e-05, | |
| "loss": 0.6338, | |
| "step": 1279 | |
| }, | |
| { | |
| "epoch": 2.811641954969797, | |
| "grad_norm": 0.04034719616174698, | |
| "learning_rate": 2.1009771986970682e-05, | |
| "loss": 0.6228, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 2.813838550247117, | |
| "grad_norm": 0.05566655471920967, | |
| "learning_rate": 2.076547231270358e-05, | |
| "loss": 0.6912, | |
| "step": 1281 | |
| }, | |
| { | |
| "epoch": 2.8160351455244372, | |
| "grad_norm": 0.05668025091290474, | |
| "learning_rate": 2.052117263843648e-05, | |
| "loss": 0.7036, | |
| "step": 1282 | |
| }, | |
| { | |
| "epoch": 2.818231740801757, | |
| "grad_norm": 0.04943126440048218, | |
| "learning_rate": 2.027687296416938e-05, | |
| "loss": 0.7858, | |
| "step": 1283 | |
| }, | |
| { | |
| "epoch": 2.820428336079077, | |
| "grad_norm": 0.03997279703617096, | |
| "learning_rate": 2.003257328990228e-05, | |
| "loss": 0.7082, | |
| "step": 1284 | |
| }, | |
| { | |
| "epoch": 2.8226249313563976, | |
| "grad_norm": 0.050963182002305984, | |
| "learning_rate": 1.978827361563518e-05, | |
| "loss": 0.6874, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 2.824821526633718, | |
| "grad_norm": 0.051822926849126816, | |
| "learning_rate": 1.9543973941368078e-05, | |
| "loss": 0.8515, | |
| "step": 1286 | |
| }, | |
| { | |
| "epoch": 2.827018121911038, | |
| "grad_norm": 0.03817344456911087, | |
| "learning_rate": 1.9299674267100977e-05, | |
| "loss": 0.5715, | |
| "step": 1287 | |
| }, | |
| { | |
| "epoch": 2.829214717188358, | |
| "grad_norm": 0.05846104770898819, | |
| "learning_rate": 1.9055374592833873e-05, | |
| "loss": 0.6425, | |
| "step": 1288 | |
| }, | |
| { | |
| "epoch": 2.8314113124656783, | |
| "grad_norm": 0.04724184796214104, | |
| "learning_rate": 1.8811074918566773e-05, | |
| "loss": 0.6654, | |
| "step": 1289 | |
| }, | |
| { | |
| "epoch": 2.8336079077429983, | |
| "grad_norm": 0.05438625440001488, | |
| "learning_rate": 1.8566775244299672e-05, | |
| "loss": 0.7263, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 2.8358045030203183, | |
| "grad_norm": 0.0520639531314373, | |
| "learning_rate": 1.832247557003257e-05, | |
| "loss": 0.5943, | |
| "step": 1291 | |
| }, | |
| { | |
| "epoch": 2.8380010982976387, | |
| "grad_norm": 0.043778836727142334, | |
| "learning_rate": 1.807817589576547e-05, | |
| "loss": 0.6249, | |
| "step": 1292 | |
| }, | |
| { | |
| "epoch": 2.8401976935749587, | |
| "grad_norm": 0.04170997813344002, | |
| "learning_rate": 1.783387622149837e-05, | |
| "loss": 0.4842, | |
| "step": 1293 | |
| }, | |
| { | |
| "epoch": 2.842394288852279, | |
| "grad_norm": 0.048355937004089355, | |
| "learning_rate": 1.758957654723127e-05, | |
| "loss": 0.7141, | |
| "step": 1294 | |
| }, | |
| { | |
| "epoch": 2.844590884129599, | |
| "grad_norm": 0.05793530493974686, | |
| "learning_rate": 1.734527687296417e-05, | |
| "loss": 0.7181, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 2.8467874794069195, | |
| "grad_norm": 0.049781594425439835, | |
| "learning_rate": 1.7100977198697065e-05, | |
| "loss": 0.7524, | |
| "step": 1296 | |
| }, | |
| { | |
| "epoch": 2.8489840746842394, | |
| "grad_norm": 0.034993454813957214, | |
| "learning_rate": 1.6856677524429964e-05, | |
| "loss": 0.5538, | |
| "step": 1297 | |
| }, | |
| { | |
| "epoch": 2.8511806699615594, | |
| "grad_norm": 0.05010677129030228, | |
| "learning_rate": 1.6612377850162864e-05, | |
| "loss": 0.7309, | |
| "step": 1298 | |
| }, | |
| { | |
| "epoch": 2.85337726523888, | |
| "grad_norm": 0.05319371074438095, | |
| "learning_rate": 1.6368078175895766e-05, | |
| "loss": 0.662, | |
| "step": 1299 | |
| }, | |
| { | |
| "epoch": 2.8555738605161998, | |
| "grad_norm": 0.06000995635986328, | |
| "learning_rate": 1.6123778501628666e-05, | |
| "loss": 0.8406, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.85777045579352, | |
| "grad_norm": 0.03820439800620079, | |
| "learning_rate": 1.587947882736156e-05, | |
| "loss": 0.6687, | |
| "step": 1301 | |
| }, | |
| { | |
| "epoch": 2.85996705107084, | |
| "grad_norm": 0.05080217868089676, | |
| "learning_rate": 1.563517915309446e-05, | |
| "loss": 0.5845, | |
| "step": 1302 | |
| }, | |
| { | |
| "epoch": 2.8621636463481606, | |
| "grad_norm": 0.04689435660839081, | |
| "learning_rate": 1.539087947882736e-05, | |
| "loss": 0.8441, | |
| "step": 1303 | |
| }, | |
| { | |
| "epoch": 2.8643602416254805, | |
| "grad_norm": 0.05374585837125778, | |
| "learning_rate": 1.514657980456026e-05, | |
| "loss": 0.5995, | |
| "step": 1304 | |
| }, | |
| { | |
| "epoch": 2.8665568369028005, | |
| "grad_norm": 0.044032514095306396, | |
| "learning_rate": 1.4902280130293159e-05, | |
| "loss": 0.6525, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 2.868753432180121, | |
| "grad_norm": 0.041226308792829514, | |
| "learning_rate": 1.4657980456026057e-05, | |
| "loss": 0.6631, | |
| "step": 1306 | |
| }, | |
| { | |
| "epoch": 2.870950027457441, | |
| "grad_norm": 0.046438008546829224, | |
| "learning_rate": 1.4413680781758956e-05, | |
| "loss": 0.711, | |
| "step": 1307 | |
| }, | |
| { | |
| "epoch": 2.873146622734761, | |
| "grad_norm": 0.04149239510297775, | |
| "learning_rate": 1.4169381107491856e-05, | |
| "loss": 0.5172, | |
| "step": 1308 | |
| }, | |
| { | |
| "epoch": 2.8753432180120813, | |
| "grad_norm": 0.04777297005057335, | |
| "learning_rate": 1.3925081433224755e-05, | |
| "loss": 0.6341, | |
| "step": 1309 | |
| }, | |
| { | |
| "epoch": 2.8775398132894017, | |
| "grad_norm": 0.05444321781396866, | |
| "learning_rate": 1.3680781758957653e-05, | |
| "loss": 0.7063, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 2.8797364085667216, | |
| "grad_norm": 0.05217491462826729, | |
| "learning_rate": 1.3436482084690552e-05, | |
| "loss": 0.6434, | |
| "step": 1311 | |
| }, | |
| { | |
| "epoch": 2.8819330038440416, | |
| "grad_norm": 0.045649029314517975, | |
| "learning_rate": 1.3192182410423451e-05, | |
| "loss": 0.6995, | |
| "step": 1312 | |
| }, | |
| { | |
| "epoch": 2.884129599121362, | |
| "grad_norm": 0.059278231114149094, | |
| "learning_rate": 1.294788273615635e-05, | |
| "loss": 0.7023, | |
| "step": 1313 | |
| }, | |
| { | |
| "epoch": 2.886326194398682, | |
| "grad_norm": 0.0433654710650444, | |
| "learning_rate": 1.2703583061889248e-05, | |
| "loss": 0.675, | |
| "step": 1314 | |
| }, | |
| { | |
| "epoch": 2.888522789676002, | |
| "grad_norm": 0.05471406131982803, | |
| "learning_rate": 1.2459283387622148e-05, | |
| "loss": 0.6003, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 2.8907193849533224, | |
| "grad_norm": 0.05296127498149872, | |
| "learning_rate": 1.2214983713355047e-05, | |
| "loss": 0.686, | |
| "step": 1316 | |
| }, | |
| { | |
| "epoch": 2.892915980230643, | |
| "grad_norm": 0.05806552991271019, | |
| "learning_rate": 1.1970684039087946e-05, | |
| "loss": 0.5582, | |
| "step": 1317 | |
| }, | |
| { | |
| "epoch": 2.8951125755079627, | |
| "grad_norm": 0.054739389568567276, | |
| "learning_rate": 1.1726384364820844e-05, | |
| "loss": 0.5321, | |
| "step": 1318 | |
| }, | |
| { | |
| "epoch": 2.8973091707852827, | |
| "grad_norm": 0.040236715227365494, | |
| "learning_rate": 1.1482084690553745e-05, | |
| "loss": 0.5278, | |
| "step": 1319 | |
| }, | |
| { | |
| "epoch": 2.899505766062603, | |
| "grad_norm": 0.05634508281946182, | |
| "learning_rate": 1.1237785016286644e-05, | |
| "loss": 0.6725, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 2.901702361339923, | |
| "grad_norm": 0.04822323098778725, | |
| "learning_rate": 1.0993485342019544e-05, | |
| "loss": 0.4929, | |
| "step": 1321 | |
| }, | |
| { | |
| "epoch": 2.903898956617243, | |
| "grad_norm": 0.06421532481908798, | |
| "learning_rate": 1.0749185667752443e-05, | |
| "loss": 0.7244, | |
| "step": 1322 | |
| }, | |
| { | |
| "epoch": 2.9060955518945635, | |
| "grad_norm": 0.05510374531149864, | |
| "learning_rate": 1.0504885993485341e-05, | |
| "loss": 0.6058, | |
| "step": 1323 | |
| }, | |
| { | |
| "epoch": 2.9082921471718834, | |
| "grad_norm": 0.05010756105184555, | |
| "learning_rate": 1.026058631921824e-05, | |
| "loss": 0.6489, | |
| "step": 1324 | |
| }, | |
| { | |
| "epoch": 2.910488742449204, | |
| "grad_norm": 0.05712300166487694, | |
| "learning_rate": 1.001628664495114e-05, | |
| "loss": 0.5218, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 2.912685337726524, | |
| "grad_norm": 0.05162881687283516, | |
| "learning_rate": 9.771986970684039e-06, | |
| "loss": 0.6576, | |
| "step": 1326 | |
| }, | |
| { | |
| "epoch": 2.9148819330038442, | |
| "grad_norm": 0.04074005410075188, | |
| "learning_rate": 9.527687296416937e-06, | |
| "loss": 0.6519, | |
| "step": 1327 | |
| }, | |
| { | |
| "epoch": 2.917078528281164, | |
| "grad_norm": 0.040444131940603256, | |
| "learning_rate": 9.283387622149836e-06, | |
| "loss": 0.4823, | |
| "step": 1328 | |
| }, | |
| { | |
| "epoch": 2.919275123558484, | |
| "grad_norm": 0.04682661220431328, | |
| "learning_rate": 9.039087947882735e-06, | |
| "loss": 0.4372, | |
| "step": 1329 | |
| }, | |
| { | |
| "epoch": 2.9214717188358046, | |
| "grad_norm": 0.05561075732111931, | |
| "learning_rate": 8.794788273615635e-06, | |
| "loss": 0.5898, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 2.9236683141131246, | |
| "grad_norm": 0.04917816445231438, | |
| "learning_rate": 8.550488599348532e-06, | |
| "loss": 0.756, | |
| "step": 1331 | |
| }, | |
| { | |
| "epoch": 2.925864909390445, | |
| "grad_norm": 0.035087183117866516, | |
| "learning_rate": 8.306188925081432e-06, | |
| "loss": 0.367, | |
| "step": 1332 | |
| }, | |
| { | |
| "epoch": 2.928061504667765, | |
| "grad_norm": 0.07068229466676712, | |
| "learning_rate": 8.061889250814333e-06, | |
| "loss": 0.6114, | |
| "step": 1333 | |
| }, | |
| { | |
| "epoch": 2.9302580999450853, | |
| "grad_norm": 0.03798624500632286, | |
| "learning_rate": 7.81758957654723e-06, | |
| "loss": 0.5505, | |
| "step": 1334 | |
| }, | |
| { | |
| "epoch": 2.9324546952224053, | |
| "grad_norm": 0.04883285611867905, | |
| "learning_rate": 7.57328990228013e-06, | |
| "loss": 0.5005, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 2.9346512904997253, | |
| "grad_norm": 0.044976841658353806, | |
| "learning_rate": 7.328990228013028e-06, | |
| "loss": 0.5555, | |
| "step": 1336 | |
| }, | |
| { | |
| "epoch": 2.9368478857770457, | |
| "grad_norm": 0.03776836767792702, | |
| "learning_rate": 7.084690553745928e-06, | |
| "loss": 0.5256, | |
| "step": 1337 | |
| }, | |
| { | |
| "epoch": 2.9390444810543657, | |
| "grad_norm": 0.042885709553956985, | |
| "learning_rate": 6.840390879478826e-06, | |
| "loss": 0.5995, | |
| "step": 1338 | |
| }, | |
| { | |
| "epoch": 2.9412410763316856, | |
| "grad_norm": 0.040374260395765305, | |
| "learning_rate": 6.596091205211726e-06, | |
| "loss": 0.6082, | |
| "step": 1339 | |
| }, | |
| { | |
| "epoch": 2.943437671609006, | |
| "grad_norm": 0.04279405623674393, | |
| "learning_rate": 6.351791530944624e-06, | |
| "loss": 0.4615, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 2.9456342668863265, | |
| "grad_norm": 0.039576370269060135, | |
| "learning_rate": 6.1074918566775235e-06, | |
| "loss": 0.6118, | |
| "step": 1341 | |
| }, | |
| { | |
| "epoch": 2.9478308621636464, | |
| "grad_norm": 0.04402873292565346, | |
| "learning_rate": 5.863192182410422e-06, | |
| "loss": 0.681, | |
| "step": 1342 | |
| }, | |
| { | |
| "epoch": 2.9500274574409664, | |
| "grad_norm": 0.044643841683864594, | |
| "learning_rate": 5.618892508143322e-06, | |
| "loss": 0.7758, | |
| "step": 1343 | |
| }, | |
| { | |
| "epoch": 2.952224052718287, | |
| "grad_norm": 0.05211080610752106, | |
| "learning_rate": 5.374592833876222e-06, | |
| "loss": 0.7354, | |
| "step": 1344 | |
| }, | |
| { | |
| "epoch": 2.9544206479956068, | |
| "grad_norm": 0.046959247440099716, | |
| "learning_rate": 5.13029315960912e-06, | |
| "loss": 0.6511, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 2.9566172432729267, | |
| "grad_norm": 0.05532398819923401, | |
| "learning_rate": 4.8859934853420195e-06, | |
| "loss": 0.5952, | |
| "step": 1346 | |
| }, | |
| { | |
| "epoch": 2.958813838550247, | |
| "grad_norm": 0.04401236027479172, | |
| "learning_rate": 4.641693811074918e-06, | |
| "loss": 0.526, | |
| "step": 1347 | |
| }, | |
| { | |
| "epoch": 2.961010433827567, | |
| "grad_norm": 0.054430510848760605, | |
| "learning_rate": 4.397394136807817e-06, | |
| "loss": 0.5618, | |
| "step": 1348 | |
| }, | |
| { | |
| "epoch": 2.9632070291048875, | |
| "grad_norm": 0.06004808843135834, | |
| "learning_rate": 4.153094462540716e-06, | |
| "loss": 0.8459, | |
| "step": 1349 | |
| }, | |
| { | |
| "epoch": 2.9654036243822075, | |
| "grad_norm": 0.05838881433010101, | |
| "learning_rate": 3.908794788273615e-06, | |
| "loss": 0.5954, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 2.967600219659528, | |
| "grad_norm": 0.04087888449430466, | |
| "learning_rate": 3.664495114006514e-06, | |
| "loss": 0.749, | |
| "step": 1351 | |
| }, | |
| { | |
| "epoch": 2.969796814936848, | |
| "grad_norm": 0.05262356251478195, | |
| "learning_rate": 3.420195439739413e-06, | |
| "loss": 0.6423, | |
| "step": 1352 | |
| }, | |
| { | |
| "epoch": 2.971993410214168, | |
| "grad_norm": 0.04079955816268921, | |
| "learning_rate": 3.175895765472312e-06, | |
| "loss": 0.6893, | |
| "step": 1353 | |
| }, | |
| { | |
| "epoch": 2.9741900054914883, | |
| "grad_norm": 0.049717098474502563, | |
| "learning_rate": 2.931596091205211e-06, | |
| "loss": 0.7561, | |
| "step": 1354 | |
| }, | |
| { | |
| "epoch": 2.9763866007688082, | |
| "grad_norm": 0.042594753205776215, | |
| "learning_rate": 2.687296416938111e-06, | |
| "loss": 0.5603, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 2.9785831960461286, | |
| "grad_norm": 0.0829310268163681, | |
| "learning_rate": 2.4429967426710097e-06, | |
| "loss": 0.6953, | |
| "step": 1356 | |
| }, | |
| { | |
| "epoch": 2.9807797913234486, | |
| "grad_norm": 0.05175120010972023, | |
| "learning_rate": 2.1986970684039087e-06, | |
| "loss": 0.6679, | |
| "step": 1357 | |
| }, | |
| { | |
| "epoch": 2.982976386600769, | |
| "grad_norm": 0.05891898646950722, | |
| "learning_rate": 1.9543973941368076e-06, | |
| "loss": 0.5024, | |
| "step": 1358 | |
| }, | |
| { | |
| "epoch": 2.985172981878089, | |
| "grad_norm": 0.045325834304094315, | |
| "learning_rate": 1.7100977198697066e-06, | |
| "loss": 0.7896, | |
| "step": 1359 | |
| }, | |
| { | |
| "epoch": 2.987369577155409, | |
| "grad_norm": 0.04623181000351906, | |
| "learning_rate": 1.4657980456026055e-06, | |
| "loss": 0.5756, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 2.9895661724327294, | |
| "grad_norm": 0.041907262057065964, | |
| "learning_rate": 1.2214983713355049e-06, | |
| "loss": 0.6516, | |
| "step": 1361 | |
| }, | |
| { | |
| "epoch": 2.9917627677100493, | |
| "grad_norm": 0.07447605580091476, | |
| "learning_rate": 9.771986970684038e-07, | |
| "loss": 0.7468, | |
| "step": 1362 | |
| }, | |
| { | |
| "epoch": 2.9939593629873693, | |
| "grad_norm": 0.05689909681677818, | |
| "learning_rate": 7.328990228013028e-07, | |
| "loss": 0.6006, | |
| "step": 1363 | |
| }, | |
| { | |
| "epoch": 2.9961559582646897, | |
| "grad_norm": 0.04323657602071762, | |
| "learning_rate": 4.885993485342019e-07, | |
| "loss": 0.6885, | |
| "step": 1364 | |
| }, | |
| { | |
| "epoch": 2.99835255354201, | |
| "grad_norm": 0.04761827364563942, | |
| "learning_rate": 2.4429967426710095e-07, | |
| "loss": 0.7291, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 2.99835255354201, | |
| "step": 1365, | |
| "total_flos": 1.0038997179244216e+18, | |
| "train_loss": 0.6840552445296403, | |
| "train_runtime": 5408.1966, | |
| "train_samples_per_second": 4.04, | |
| "train_steps_per_second": 0.252 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 1365, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.0038997179244216e+18, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |