{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.50003543837267, "eval_steps": 500, "global_step": 7055, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 7.0876745339854e-05, "eval_accuracy": 0.6605764392507908, "eval_loss": 1.4750796556472778, "eval_runtime": 7.2429, "eval_samples_per_second": 47.909, "eval_steps_per_second": 0.414, "step": 1 }, { "epoch": 0.00070876745339854, "grad_norm": 1.8515625, "learning_rate": 1e-05, "loss": 1.538, "step": 10 }, { "epoch": 0.00141753490679708, "grad_norm": 1.0, "learning_rate": 2e-05, "loss": 1.5311, "step": 20 }, { "epoch": 0.00212630236019562, "grad_norm": 0.57421875, "learning_rate": 3e-05, "loss": 1.5247, "step": 30 }, { "epoch": 0.00283506981359416, "grad_norm": 0.423828125, "learning_rate": 4e-05, "loss": 1.531, "step": 40 }, { "epoch": 0.0035438372669927, "grad_norm": 0.451171875, "learning_rate": 5e-05, "loss": 1.5146, "step": 50 }, { "epoch": 0.00425260472039124, "grad_norm": 0.453125, "learning_rate": 6e-05, "loss": 1.5055, "step": 60 }, { "epoch": 0.00496137217378978, "grad_norm": 0.515625, "learning_rate": 7e-05, "loss": 1.5345, "step": 70 }, { "epoch": 0.00567013962718832, "grad_norm": 0.53515625, "learning_rate": 8e-05, "loss": 1.5133, "step": 80 }, { "epoch": 0.00637890708058686, "grad_norm": 0.58984375, "learning_rate": 9e-05, "loss": 1.5043, "step": 90 }, { "epoch": 0.0070876745339854, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.513, "step": 100 }, { "epoch": 0.00779644198738394, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.5083, "step": 110 }, { "epoch": 0.00850520944078248, "grad_norm": 0.451171875, "learning_rate": 0.0001, "loss": 1.5098, "step": 120 }, { "epoch": 0.00921397689418102, "grad_norm": 0.486328125, "learning_rate": 0.0001, "loss": 1.4936, "step": 130 }, { "epoch": 0.00992274434757956, "grad_norm": 0.54296875, "learning_rate": 0.0001, "loss": 1.5091, "step": 140 }, { "epoch": 0.0106315118009781, "grad_norm": 0.439453125, "learning_rate": 0.0001, "loss": 1.5073, "step": 150 }, { "epoch": 0.01134027925437664, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.5161, "step": 160 }, { "epoch": 0.01204904670777518, "grad_norm": 0.57421875, "learning_rate": 0.0001, "loss": 1.5067, "step": 170 }, { "epoch": 0.01275781416117372, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4984, "step": 180 }, { "epoch": 0.01346658161457226, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.5053, "step": 190 }, { "epoch": 0.0141753490679708, "grad_norm": 0.478515625, "learning_rate": 0.0001, "loss": 1.4937, "step": 200 }, { "epoch": 0.01488411652136934, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.4946, "step": 210 }, { "epoch": 0.01559288397476788, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.5031, "step": 220 }, { "epoch": 0.01630165142816642, "grad_norm": 0.6015625, "learning_rate": 0.0001, "loss": 1.4857, "step": 230 }, { "epoch": 0.01701041888156496, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.5134, "step": 240 }, { "epoch": 0.0177191863349635, "grad_norm": 0.55078125, "learning_rate": 0.0001, "loss": 1.5069, "step": 250 }, { "epoch": 0.01842795378836204, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.5037, "step": 260 }, { "epoch": 0.01913672124176058, "grad_norm": 0.57421875, "learning_rate": 0.0001, "loss": 1.4999, "step": 270 }, { "epoch": 0.01984548869515912, "grad_norm": 0.4921875, "learning_rate": 0.0001, "loss": 1.5038, "step": 280 }, { "epoch": 0.02055425614855766, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4985, "step": 290 }, { "epoch": 0.0212630236019562, "grad_norm": 0.546875, "learning_rate": 0.0001, "loss": 1.5112, "step": 300 }, { "epoch": 0.02197179105535474, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.5083, "step": 310 }, { "epoch": 0.02268055850875328, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.5104, "step": 320 }, { "epoch": 0.02338932596215182, "grad_norm": 0.49609375, "learning_rate": 0.0001, "loss": 1.5132, "step": 330 }, { "epoch": 0.02409809341555036, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4956, "step": 340 }, { "epoch": 0.0248068608689489, "grad_norm": 0.486328125, "learning_rate": 0.0001, "loss": 1.4949, "step": 350 }, { "epoch": 0.02551562832234744, "grad_norm": 0.46484375, "learning_rate": 0.0001, "loss": 1.5018, "step": 360 }, { "epoch": 0.02622439577574598, "grad_norm": 0.435546875, "learning_rate": 0.0001, "loss": 1.4821, "step": 370 }, { "epoch": 0.02693316322914452, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.5092, "step": 380 }, { "epoch": 0.02764193068254306, "grad_norm": 0.4921875, "learning_rate": 0.0001, "loss": 1.4998, "step": 390 }, { "epoch": 0.0283506981359416, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4958, "step": 400 }, { "epoch": 0.02905946558934014, "grad_norm": 0.546875, "learning_rate": 0.0001, "loss": 1.5031, "step": 410 }, { "epoch": 0.02976823304273868, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4929, "step": 420 }, { "epoch": 0.03047700049613722, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4987, "step": 430 }, { "epoch": 0.03118576794953576, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4853, "step": 440 }, { "epoch": 0.0318945354029343, "grad_norm": 0.482421875, "learning_rate": 0.0001, "loss": 1.4968, "step": 450 }, { "epoch": 0.03260330285633284, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.4872, "step": 460 }, { "epoch": 0.03331207030973138, "grad_norm": 0.48828125, "learning_rate": 0.0001, "loss": 1.5061, "step": 470 }, { "epoch": 0.03402083776312992, "grad_norm": 0.6484375, "learning_rate": 0.0001, "loss": 1.5078, "step": 480 }, { "epoch": 0.03472960521652846, "grad_norm": 0.47265625, "learning_rate": 0.0001, "loss": 1.504, "step": 490 }, { "epoch": 0.035438372669927, "grad_norm": 0.470703125, "learning_rate": 0.0001, "loss": 1.5071, "step": 500 }, { "epoch": 0.035438372669927, "eval_accuracy": 0.6646616911746595, "eval_loss": 1.4113208055496216, "eval_runtime": 7.1977, "eval_samples_per_second": 48.21, "eval_steps_per_second": 0.417, "step": 500 }, { "epoch": 0.03614714012332554, "grad_norm": 0.5625, "learning_rate": 0.0001, "loss": 1.5028, "step": 510 }, { "epoch": 0.03685590757672408, "grad_norm": 0.6015625, "learning_rate": 0.0001, "loss": 1.5042, "step": 520 }, { "epoch": 0.03756467503012262, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4969, "step": 530 }, { "epoch": 0.03827344248352116, "grad_norm": 0.48828125, "learning_rate": 0.0001, "loss": 1.5016, "step": 540 }, { "epoch": 0.0389822099369197, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.5008, "step": 550 }, { "epoch": 0.03969097739031824, "grad_norm": 0.546875, "learning_rate": 0.0001, "loss": 1.5126, "step": 560 }, { "epoch": 0.04039974484371678, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4949, "step": 570 }, { "epoch": 0.04110851229711532, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4912, "step": 580 }, { "epoch": 0.04181727975051386, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.4912, "step": 590 }, { "epoch": 0.0425260472039124, "grad_norm": 0.498046875, "learning_rate": 0.0001, "loss": 1.4945, "step": 600 }, { "epoch": 0.04323481465731094, "grad_norm": 0.455078125, "learning_rate": 0.0001, "loss": 1.4977, "step": 610 }, { "epoch": 0.04394358211070948, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.4908, "step": 620 }, { "epoch": 0.04465234956410802, "grad_norm": 0.451171875, "learning_rate": 0.0001, "loss": 1.4937, "step": 630 }, { "epoch": 0.04536111701750656, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4966, "step": 640 }, { "epoch": 0.0460698844709051, "grad_norm": 0.58203125, "learning_rate": 0.0001, "loss": 1.4982, "step": 650 }, { "epoch": 0.04677865192430364, "grad_norm": 0.46484375, "learning_rate": 0.0001, "loss": 1.4889, "step": 660 }, { "epoch": 0.04748741937770218, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.4948, "step": 670 }, { "epoch": 0.04819618683110072, "grad_norm": 0.478515625, "learning_rate": 0.0001, "loss": 1.5004, "step": 680 }, { "epoch": 0.04890495428449926, "grad_norm": 0.443359375, "learning_rate": 0.0001, "loss": 1.4888, "step": 690 }, { "epoch": 0.0496137217378978, "grad_norm": 0.56640625, "learning_rate": 0.0001, "loss": 1.4981, "step": 700 }, { "epoch": 0.05032248919129634, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.504, "step": 710 }, { "epoch": 0.05103125664469488, "grad_norm": 0.56640625, "learning_rate": 0.0001, "loss": 1.4909, "step": 720 }, { "epoch": 0.05174002409809342, "grad_norm": 0.498046875, "learning_rate": 0.0001, "loss": 1.4971, "step": 730 }, { "epoch": 0.05244879155149196, "grad_norm": 0.453125, "learning_rate": 0.0001, "loss": 1.4994, "step": 740 }, { "epoch": 0.0531575590048905, "grad_norm": 0.60546875, "learning_rate": 0.0001, "loss": 1.4926, "step": 750 }, { "epoch": 0.05386632645828904, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4945, "step": 760 }, { "epoch": 0.05457509391168758, "grad_norm": 0.5546875, "learning_rate": 0.0001, "loss": 1.5003, "step": 770 }, { "epoch": 0.05528386136508612, "grad_norm": 0.486328125, "learning_rate": 0.0001, "loss": 1.4998, "step": 780 }, { "epoch": 0.05599262881848466, "grad_norm": 0.59375, "learning_rate": 0.0001, "loss": 1.5005, "step": 790 }, { "epoch": 0.0567013962718832, "grad_norm": 0.462890625, "learning_rate": 0.0001, "loss": 1.509, "step": 800 }, { "epoch": 0.05741016372528174, "grad_norm": 0.4375, "learning_rate": 0.0001, "loss": 1.4885, "step": 810 }, { "epoch": 0.05811893117868028, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4957, "step": 820 }, { "epoch": 0.05882769863207882, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.5006, "step": 830 }, { "epoch": 0.05953646608547736, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.493, "step": 840 }, { "epoch": 0.0602452335388759, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.5008, "step": 850 }, { "epoch": 0.06095400099227444, "grad_norm": 0.48046875, "learning_rate": 0.0001, "loss": 1.4966, "step": 860 }, { "epoch": 0.06166276844567298, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.5013, "step": 870 }, { "epoch": 0.06237153589907152, "grad_norm": 0.4609375, "learning_rate": 0.0001, "loss": 1.4937, "step": 880 }, { "epoch": 0.06308030335247006, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.4815, "step": 890 }, { "epoch": 0.0637890708058686, "grad_norm": 0.44921875, "learning_rate": 0.0001, "loss": 1.4878, "step": 900 }, { "epoch": 0.06449783825926714, "grad_norm": 0.4921875, "learning_rate": 0.0001, "loss": 1.4948, "step": 910 }, { "epoch": 0.06520660571266568, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.5051, "step": 920 }, { "epoch": 0.06591537316606422, "grad_norm": 0.486328125, "learning_rate": 0.0001, "loss": 1.4993, "step": 930 }, { "epoch": 0.06662414061946276, "grad_norm": 0.470703125, "learning_rate": 0.0001, "loss": 1.4984, "step": 940 }, { "epoch": 0.0673329080728613, "grad_norm": 0.46875, "learning_rate": 0.0001, "loss": 1.5001, "step": 950 }, { "epoch": 0.06804167552625984, "grad_norm": 0.462890625, "learning_rate": 0.0001, "loss": 1.4899, "step": 960 }, { "epoch": 0.06875044297965838, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4881, "step": 970 }, { "epoch": 0.06945921043305692, "grad_norm": 0.46875, "learning_rate": 0.0001, "loss": 1.5096, "step": 980 }, { "epoch": 0.07016797788645546, "grad_norm": 0.48828125, "learning_rate": 0.0001, "loss": 1.4902, "step": 990 }, { "epoch": 0.070876745339854, "grad_norm": 0.59765625, "learning_rate": 0.0001, "loss": 1.5003, "step": 1000 }, { "epoch": 0.070876745339854, "eval_accuracy": 0.6648911127297294, "eval_loss": 1.4080075025558472, "eval_runtime": 7.1527, "eval_samples_per_second": 48.513, "eval_steps_per_second": 0.419, "step": 1000 }, { "epoch": 0.07158551279325254, "grad_norm": 0.62109375, "learning_rate": 0.0001, "loss": 1.4882, "step": 1010 }, { "epoch": 0.07229428024665108, "grad_norm": 0.478515625, "learning_rate": 0.0001, "loss": 1.5019, "step": 1020 }, { "epoch": 0.07300304770004962, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.488, "step": 1030 }, { "epoch": 0.07371181515344816, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.5065, "step": 1040 }, { "epoch": 0.0744205826068467, "grad_norm": 0.474609375, "learning_rate": 0.0001, "loss": 1.4906, "step": 1050 }, { "epoch": 0.07512935006024524, "grad_norm": 0.48828125, "learning_rate": 0.0001, "loss": 1.5036, "step": 1060 }, { "epoch": 0.07583811751364378, "grad_norm": 0.609375, "learning_rate": 0.0001, "loss": 1.4968, "step": 1070 }, { "epoch": 0.07654688496704232, "grad_norm": 0.55859375, "learning_rate": 0.0001, "loss": 1.5076, "step": 1080 }, { "epoch": 0.07725565242044086, "grad_norm": 0.466796875, "learning_rate": 0.0001, "loss": 1.5008, "step": 1090 }, { "epoch": 0.0779644198738394, "grad_norm": 0.55859375, "learning_rate": 0.0001, "loss": 1.4951, "step": 1100 }, { "epoch": 0.07867318732723794, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.4844, "step": 1110 }, { "epoch": 0.07938195478063648, "grad_norm": 0.51953125, "learning_rate": 0.0001, "loss": 1.499, "step": 1120 }, { "epoch": 0.08009072223403502, "grad_norm": 0.451171875, "learning_rate": 0.0001, "loss": 1.4794, "step": 1130 }, { "epoch": 0.08079948968743356, "grad_norm": 0.55078125, "learning_rate": 0.0001, "loss": 1.508, "step": 1140 }, { "epoch": 0.0815082571408321, "grad_norm": 0.494140625, "learning_rate": 0.0001, "loss": 1.4997, "step": 1150 }, { "epoch": 0.08221702459423064, "grad_norm": 0.47265625, "learning_rate": 0.0001, "loss": 1.4987, "step": 1160 }, { "epoch": 0.08292579204762918, "grad_norm": 0.578125, "learning_rate": 0.0001, "loss": 1.5028, "step": 1170 }, { "epoch": 0.08363455950102772, "grad_norm": 0.453125, "learning_rate": 0.0001, "loss": 1.4991, "step": 1180 }, { "epoch": 0.08434332695442626, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.5075, "step": 1190 }, { "epoch": 0.0850520944078248, "grad_norm": 0.53515625, "learning_rate": 0.0001, "loss": 1.4845, "step": 1200 }, { "epoch": 0.08576086186122334, "grad_norm": 0.5703125, "learning_rate": 0.0001, "loss": 1.4993, "step": 1210 }, { "epoch": 0.08646962931462188, "grad_norm": 0.58984375, "learning_rate": 0.0001, "loss": 1.502, "step": 1220 }, { "epoch": 0.08717839676802042, "grad_norm": 0.49609375, "learning_rate": 0.0001, "loss": 1.4859, "step": 1230 }, { "epoch": 0.08788716422141896, "grad_norm": 0.498046875, "learning_rate": 0.0001, "loss": 1.5029, "step": 1240 }, { "epoch": 0.0885959316748175, "grad_norm": 0.44921875, "learning_rate": 0.0001, "loss": 1.5035, "step": 1250 }, { "epoch": 0.08930469912821604, "grad_norm": 0.55859375, "learning_rate": 0.0001, "loss": 1.5057, "step": 1260 }, { "epoch": 0.09001346658161458, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4862, "step": 1270 }, { "epoch": 0.09072223403501312, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.4941, "step": 1280 }, { "epoch": 0.09143100148841166, "grad_norm": 0.427734375, "learning_rate": 0.0001, "loss": 1.4995, "step": 1290 }, { "epoch": 0.0921397689418102, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.4891, "step": 1300 }, { "epoch": 0.09284853639520874, "grad_norm": 0.4296875, "learning_rate": 0.0001, "loss": 1.5021, "step": 1310 }, { "epoch": 0.09355730384860728, "grad_norm": 0.478515625, "learning_rate": 0.0001, "loss": 1.4896, "step": 1320 }, { "epoch": 0.09426607130200582, "grad_norm": 0.58203125, "learning_rate": 0.0001, "loss": 1.4889, "step": 1330 }, { "epoch": 0.09497483875540436, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4978, "step": 1340 }, { "epoch": 0.0956836062088029, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.505, "step": 1350 }, { "epoch": 0.09639237366220144, "grad_norm": 0.57421875, "learning_rate": 0.0001, "loss": 1.4934, "step": 1360 }, { "epoch": 0.09710114111559998, "grad_norm": 0.482421875, "learning_rate": 0.0001, "loss": 1.4956, "step": 1370 }, { "epoch": 0.09780990856899852, "grad_norm": 0.46875, "learning_rate": 0.0001, "loss": 1.4973, "step": 1380 }, { "epoch": 0.09851867602239706, "grad_norm": 0.498046875, "learning_rate": 0.0001, "loss": 1.4906, "step": 1390 }, { "epoch": 0.0992274434757956, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.4942, "step": 1400 }, { "epoch": 0.09993621092919414, "grad_norm": 0.482421875, "learning_rate": 0.0001, "loss": 1.4881, "step": 1410 }, { "epoch": 0.10064497838259268, "grad_norm": 0.59765625, "learning_rate": 0.0001, "loss": 1.4837, "step": 1420 }, { "epoch": 0.10135374583599122, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4954, "step": 1430 }, { "epoch": 0.10206251328938976, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4838, "step": 1440 }, { "epoch": 0.1027712807427883, "grad_norm": 0.62109375, "learning_rate": 0.0001, "loss": 1.4998, "step": 1450 }, { "epoch": 0.10348004819618684, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4998, "step": 1460 }, { "epoch": 0.10418881564958538, "grad_norm": 0.482421875, "learning_rate": 0.0001, "loss": 1.4876, "step": 1470 }, { "epoch": 0.10489758310298392, "grad_norm": 0.49609375, "learning_rate": 0.0001, "loss": 1.4965, "step": 1480 }, { "epoch": 0.10560635055638246, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4875, "step": 1490 }, { "epoch": 0.106315118009781, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4959, "step": 1500 }, { "epoch": 0.106315118009781, "eval_accuracy": 0.6653851431949415, "eval_loss": 1.4062947034835815, "eval_runtime": 7.3356, "eval_samples_per_second": 47.304, "eval_steps_per_second": 0.409, "step": 1500 }, { "epoch": 0.10702388546317954, "grad_norm": 0.51953125, "learning_rate": 0.0001, "loss": 1.4954, "step": 1510 }, { "epoch": 0.10773265291657808, "grad_norm": 0.59375, "learning_rate": 0.0001, "loss": 1.4988, "step": 1520 }, { "epoch": 0.10844142036997662, "grad_norm": 0.447265625, "learning_rate": 0.0001, "loss": 1.4895, "step": 1530 }, { "epoch": 0.10915018782337516, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.5043, "step": 1540 }, { "epoch": 0.1098589552767737, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.4746, "step": 1550 }, { "epoch": 0.11056772273017224, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4898, "step": 1560 }, { "epoch": 0.11127649018357078, "grad_norm": 0.474609375, "learning_rate": 0.0001, "loss": 1.4876, "step": 1570 }, { "epoch": 0.11198525763696932, "grad_norm": 0.6015625, "learning_rate": 0.0001, "loss": 1.4966, "step": 1580 }, { "epoch": 0.11269402509036786, "grad_norm": 0.6015625, "learning_rate": 0.0001, "loss": 1.5048, "step": 1590 }, { "epoch": 0.1134027925437664, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.482, "step": 1600 }, { "epoch": 0.11411155999716494, "grad_norm": 0.55078125, "learning_rate": 0.0001, "loss": 1.4951, "step": 1610 }, { "epoch": 0.11482032745056348, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.485, "step": 1620 }, { "epoch": 0.11552909490396202, "grad_norm": 0.455078125, "learning_rate": 0.0001, "loss": 1.491, "step": 1630 }, { "epoch": 0.11623786235736056, "grad_norm": 0.451171875, "learning_rate": 0.0001, "loss": 1.4785, "step": 1640 }, { "epoch": 0.1169466298107591, "grad_norm": 0.5625, "learning_rate": 0.0001, "loss": 1.4971, "step": 1650 }, { "epoch": 0.11765539726415764, "grad_norm": 0.49609375, "learning_rate": 0.0001, "loss": 1.5013, "step": 1660 }, { "epoch": 0.11836416471755618, "grad_norm": 0.49609375, "learning_rate": 0.0001, "loss": 1.5077, "step": 1670 }, { "epoch": 0.11907293217095472, "grad_norm": 0.494140625, "learning_rate": 0.0001, "loss": 1.4939, "step": 1680 }, { "epoch": 0.11978169962435326, "grad_norm": 0.63671875, "learning_rate": 0.0001, "loss": 1.4829, "step": 1690 }, { "epoch": 0.1204904670777518, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4964, "step": 1700 }, { "epoch": 0.12119923453115033, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4969, "step": 1710 }, { "epoch": 0.12190800198454887, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4955, "step": 1720 }, { "epoch": 0.12261676943794741, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.4913, "step": 1730 }, { "epoch": 0.12332553689134595, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.4905, "step": 1740 }, { "epoch": 0.1240343043447445, "grad_norm": 0.4609375, "learning_rate": 0.0001, "loss": 1.5034, "step": 1750 }, { "epoch": 0.12474307179814303, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.4991, "step": 1760 }, { "epoch": 0.12545183925154157, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.5019, "step": 1770 }, { "epoch": 0.12616060670494011, "grad_norm": 0.57421875, "learning_rate": 0.0001, "loss": 1.4943, "step": 1780 }, { "epoch": 0.12686937415833865, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.4971, "step": 1790 }, { "epoch": 0.1275781416117372, "grad_norm": 0.56640625, "learning_rate": 0.0001, "loss": 1.4961, "step": 1800 }, { "epoch": 0.12828690906513573, "grad_norm": 0.45703125, "learning_rate": 0.0001, "loss": 1.4888, "step": 1810 }, { "epoch": 0.12899567651853427, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4974, "step": 1820 }, { "epoch": 0.12970444397193281, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.4866, "step": 1830 }, { "epoch": 0.13041321142533135, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4878, "step": 1840 }, { "epoch": 0.1311219788787299, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4855, "step": 1850 }, { "epoch": 0.13183074633212843, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.503, "step": 1860 }, { "epoch": 0.13253951378552697, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.4955, "step": 1870 }, { "epoch": 0.13324828123892551, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4956, "step": 1880 }, { "epoch": 0.13395704869232405, "grad_norm": 0.494140625, "learning_rate": 0.0001, "loss": 1.4922, "step": 1890 }, { "epoch": 0.1346658161457226, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.5161, "step": 1900 }, { "epoch": 0.13537458359912113, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4909, "step": 1910 }, { "epoch": 0.13608335105251967, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.4936, "step": 1920 }, { "epoch": 0.13679211850591821, "grad_norm": 0.59765625, "learning_rate": 0.0001, "loss": 1.4882, "step": 1930 }, { "epoch": 0.13750088595931675, "grad_norm": 0.55859375, "learning_rate": 0.0001, "loss": 1.4745, "step": 1940 }, { "epoch": 0.1382096534127153, "grad_norm": 0.51953125, "learning_rate": 0.0001, "loss": 1.4964, "step": 1950 }, { "epoch": 0.13891842086611383, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.5041, "step": 1960 }, { "epoch": 0.13962718831951237, "grad_norm": 0.62890625, "learning_rate": 0.0001, "loss": 1.4916, "step": 1970 }, { "epoch": 0.14033595577291091, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.4992, "step": 1980 }, { "epoch": 0.14104472322630945, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.5001, "step": 1990 }, { "epoch": 0.141753490679708, "grad_norm": 0.546875, "learning_rate": 0.0001, "loss": 1.5019, "step": 2000 }, { "epoch": 0.141753490679708, "eval_accuracy": 0.6655364488217514, "eval_loss": 1.4053725004196167, "eval_runtime": 7.1455, "eval_samples_per_second": 48.562, "eval_steps_per_second": 0.42, "step": 2000 }, { "epoch": 0.14246225813310653, "grad_norm": 0.49609375, "learning_rate": 0.0001, "loss": 1.5015, "step": 2010 }, { "epoch": 0.14317102558650507, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4842, "step": 2020 }, { "epoch": 0.14387979303990361, "grad_norm": 0.466796875, "learning_rate": 0.0001, "loss": 1.4868, "step": 2030 }, { "epoch": 0.14458856049330215, "grad_norm": 0.482421875, "learning_rate": 0.0001, "loss": 1.4947, "step": 2040 }, { "epoch": 0.1452973279467007, "grad_norm": 0.4921875, "learning_rate": 0.0001, "loss": 1.4964, "step": 2050 }, { "epoch": 0.14600609540009923, "grad_norm": 0.578125, "learning_rate": 0.0001, "loss": 1.5029, "step": 2060 }, { "epoch": 0.14671486285349777, "grad_norm": 0.47265625, "learning_rate": 0.0001, "loss": 1.4974, "step": 2070 }, { "epoch": 0.1474236303068963, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.488, "step": 2080 }, { "epoch": 0.14813239776029485, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.5008, "step": 2090 }, { "epoch": 0.1488411652136934, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.5018, "step": 2100 }, { "epoch": 0.14954993266709193, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.5048, "step": 2110 }, { "epoch": 0.15025870012049047, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.5077, "step": 2120 }, { "epoch": 0.150967467573889, "grad_norm": 0.466796875, "learning_rate": 0.0001, "loss": 1.4962, "step": 2130 }, { "epoch": 0.15167623502728755, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4966, "step": 2140 }, { "epoch": 0.1523850024806861, "grad_norm": 0.6171875, "learning_rate": 0.0001, "loss": 1.4926, "step": 2150 }, { "epoch": 0.15309376993408463, "grad_norm": 0.640625, "learning_rate": 0.0001, "loss": 1.4895, "step": 2160 }, { "epoch": 0.15380253738748317, "grad_norm": 0.51953125, "learning_rate": 0.0001, "loss": 1.4877, "step": 2170 }, { "epoch": 0.1545113048408817, "grad_norm": 0.5546875, "learning_rate": 0.0001, "loss": 1.4981, "step": 2180 }, { "epoch": 0.15522007229428025, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.4947, "step": 2190 }, { "epoch": 0.1559288397476788, "grad_norm": 0.6015625, "learning_rate": 0.0001, "loss": 1.4891, "step": 2200 }, { "epoch": 0.15663760720107733, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.5004, "step": 2210 }, { "epoch": 0.15734637465447587, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.4877, "step": 2220 }, { "epoch": 0.1580551421078744, "grad_norm": 0.455078125, "learning_rate": 0.0001, "loss": 1.4966, "step": 2230 }, { "epoch": 0.15876390956127295, "grad_norm": 0.609375, "learning_rate": 0.0001, "loss": 1.4934, "step": 2240 }, { "epoch": 0.1594726770146715, "grad_norm": 0.5625, "learning_rate": 0.0001, "loss": 1.4981, "step": 2250 }, { "epoch": 0.16018144446807003, "grad_norm": 0.6328125, "learning_rate": 0.0001, "loss": 1.5087, "step": 2260 }, { "epoch": 0.16089021192146857, "grad_norm": 0.5625, "learning_rate": 0.0001, "loss": 1.4989, "step": 2270 }, { "epoch": 0.1615989793748671, "grad_norm": 0.609375, "learning_rate": 0.0001, "loss": 1.4785, "step": 2280 }, { "epoch": 0.16230774682826565, "grad_norm": 0.498046875, "learning_rate": 0.0001, "loss": 1.4978, "step": 2290 }, { "epoch": 0.1630165142816642, "grad_norm": 0.462890625, "learning_rate": 0.0001, "loss": 1.4931, "step": 2300 }, { "epoch": 0.16372528173506273, "grad_norm": 0.451171875, "learning_rate": 0.0001, "loss": 1.5119, "step": 2310 }, { "epoch": 0.16443404918846127, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.5097, "step": 2320 }, { "epoch": 0.1651428166418598, "grad_norm": 0.482421875, "learning_rate": 0.0001, "loss": 1.481, "step": 2330 }, { "epoch": 0.16585158409525835, "grad_norm": 0.57421875, "learning_rate": 0.0001, "loss": 1.4884, "step": 2340 }, { "epoch": 0.1665603515486569, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4934, "step": 2350 }, { "epoch": 0.16726911900205543, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.4789, "step": 2360 }, { "epoch": 0.16797788645545397, "grad_norm": 0.45703125, "learning_rate": 0.0001, "loss": 1.5063, "step": 2370 }, { "epoch": 0.1686866539088525, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.5095, "step": 2380 }, { "epoch": 0.16939542136225105, "grad_norm": 0.474609375, "learning_rate": 0.0001, "loss": 1.4949, "step": 2390 }, { "epoch": 0.1701041888156496, "grad_norm": 0.45703125, "learning_rate": 0.0001, "loss": 1.4918, "step": 2400 }, { "epoch": 0.17081295626904813, "grad_norm": 0.455078125, "learning_rate": 0.0001, "loss": 1.5029, "step": 2410 }, { "epoch": 0.17152172372244667, "grad_norm": 0.48046875, "learning_rate": 0.0001, "loss": 1.4945, "step": 2420 }, { "epoch": 0.1722304911758452, "grad_norm": 0.6171875, "learning_rate": 0.0001, "loss": 1.4933, "step": 2430 }, { "epoch": 0.17293925862924375, "grad_norm": 0.61328125, "learning_rate": 0.0001, "loss": 1.4949, "step": 2440 }, { "epoch": 0.1736480260826423, "grad_norm": 0.57421875, "learning_rate": 0.0001, "loss": 1.5043, "step": 2450 }, { "epoch": 0.17435679353604083, "grad_norm": 0.55078125, "learning_rate": 0.0001, "loss": 1.4971, "step": 2460 }, { "epoch": 0.17506556098943937, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.49, "step": 2470 }, { "epoch": 0.1757743284428379, "grad_norm": 0.482421875, "learning_rate": 0.0001, "loss": 1.4892, "step": 2480 }, { "epoch": 0.17648309589623645, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.5006, "step": 2490 }, { "epoch": 0.177191863349635, "grad_norm": 0.486328125, "learning_rate": 0.0001, "loss": 1.4891, "step": 2500 }, { "epoch": 0.177191863349635, "eval_accuracy": 0.6656455296224749, "eval_loss": 1.4047455787658691, "eval_runtime": 7.1711, "eval_samples_per_second": 48.388, "eval_steps_per_second": 0.418, "step": 2500 }, { "epoch": 0.17790063080303353, "grad_norm": 0.462890625, "learning_rate": 0.0001, "loss": 1.4883, "step": 2510 }, { "epoch": 0.17860939825643207, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4906, "step": 2520 }, { "epoch": 0.1793181657098306, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.4946, "step": 2530 }, { "epoch": 0.18002693316322915, "grad_norm": 0.462890625, "learning_rate": 0.0001, "loss": 1.4933, "step": 2540 }, { "epoch": 0.1807357006166277, "grad_norm": 0.53515625, "learning_rate": 0.0001, "loss": 1.4794, "step": 2550 }, { "epoch": 0.18144446807002623, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.5019, "step": 2560 }, { "epoch": 0.18215323552342477, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.4956, "step": 2570 }, { "epoch": 0.1828620029768233, "grad_norm": 0.4609375, "learning_rate": 0.0001, "loss": 1.5043, "step": 2580 }, { "epoch": 0.18357077043022185, "grad_norm": 0.455078125, "learning_rate": 0.0001, "loss": 1.4928, "step": 2590 }, { "epoch": 0.1842795378836204, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.5021, "step": 2600 }, { "epoch": 0.18498830533701893, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.5033, "step": 2610 }, { "epoch": 0.18569707279041747, "grad_norm": 0.546875, "learning_rate": 0.0001, "loss": 1.4875, "step": 2620 }, { "epoch": 0.186405840243816, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.4913, "step": 2630 }, { "epoch": 0.18711460769721455, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.4929, "step": 2640 }, { "epoch": 0.1878233751506131, "grad_norm": 0.4609375, "learning_rate": 0.0001, "loss": 1.4948, "step": 2650 }, { "epoch": 0.18853214260401163, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.5015, "step": 2660 }, { "epoch": 0.18924091005741017, "grad_norm": 0.470703125, "learning_rate": 0.0001, "loss": 1.4996, "step": 2670 }, { "epoch": 0.1899496775108087, "grad_norm": 0.431640625, "learning_rate": 0.0001, "loss": 1.487, "step": 2680 }, { "epoch": 0.19065844496420725, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4944, "step": 2690 }, { "epoch": 0.1913672124176058, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4889, "step": 2700 }, { "epoch": 0.19207597987100433, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4996, "step": 2710 }, { "epoch": 0.19278474732440287, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.4908, "step": 2720 }, { "epoch": 0.1934935147778014, "grad_norm": 0.4453125, "learning_rate": 0.0001, "loss": 1.4877, "step": 2730 }, { "epoch": 0.19420228223119995, "grad_norm": 0.48046875, "learning_rate": 0.0001, "loss": 1.4824, "step": 2740 }, { "epoch": 0.1949110496845985, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.4925, "step": 2750 }, { "epoch": 0.19561981713799703, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4895, "step": 2760 }, { "epoch": 0.19632858459139557, "grad_norm": 0.47265625, "learning_rate": 0.0001, "loss": 1.489, "step": 2770 }, { "epoch": 0.1970373520447941, "grad_norm": 0.57421875, "learning_rate": 0.0001, "loss": 1.4968, "step": 2780 }, { "epoch": 0.19774611949819265, "grad_norm": 0.455078125, "learning_rate": 0.0001, "loss": 1.4985, "step": 2790 }, { "epoch": 0.1984548869515912, "grad_norm": 0.51953125, "learning_rate": 0.0001, "loss": 1.482, "step": 2800 }, { "epoch": 0.19916365440498973, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.5019, "step": 2810 }, { "epoch": 0.19987242185838827, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4914, "step": 2820 }, { "epoch": 0.2005811893117868, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4989, "step": 2830 }, { "epoch": 0.20128995676518535, "grad_norm": 0.53515625, "learning_rate": 0.0001, "loss": 1.4927, "step": 2840 }, { "epoch": 0.2019987242185839, "grad_norm": 0.46875, "learning_rate": 0.0001, "loss": 1.496, "step": 2850 }, { "epoch": 0.20270749167198243, "grad_norm": 0.486328125, "learning_rate": 0.0001, "loss": 1.4828, "step": 2860 }, { "epoch": 0.20341625912538097, "grad_norm": 0.546875, "learning_rate": 0.0001, "loss": 1.4945, "step": 2870 }, { "epoch": 0.2041250265787795, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4906, "step": 2880 }, { "epoch": 0.20483379403217805, "grad_norm": 0.48046875, "learning_rate": 0.0001, "loss": 1.5005, "step": 2890 }, { "epoch": 0.2055425614855766, "grad_norm": 0.55859375, "learning_rate": 0.0001, "loss": 1.5009, "step": 2900 }, { "epoch": 0.20625132893897513, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4862, "step": 2910 }, { "epoch": 0.20696009639237367, "grad_norm": 0.486328125, "learning_rate": 0.0001, "loss": 1.4833, "step": 2920 }, { "epoch": 0.2076688638457722, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.4975, "step": 2930 }, { "epoch": 0.20837763129917075, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4992, "step": 2940 }, { "epoch": 0.2090863987525693, "grad_norm": 0.59765625, "learning_rate": 0.0001, "loss": 1.4878, "step": 2950 }, { "epoch": 0.20979516620596783, "grad_norm": 0.478515625, "learning_rate": 0.0001, "loss": 1.4978, "step": 2960 }, { "epoch": 0.21050393365936637, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.484, "step": 2970 }, { "epoch": 0.2112127011127649, "grad_norm": 0.59765625, "learning_rate": 0.0001, "loss": 1.5065, "step": 2980 }, { "epoch": 0.21192146856616345, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.4829, "step": 2990 }, { "epoch": 0.212630236019562, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4916, "step": 3000 }, { "epoch": 0.212630236019562, "eval_accuracy": 0.6657292755275465, "eval_loss": 1.4040066003799438, "eval_runtime": 7.1205, "eval_samples_per_second": 48.733, "eval_steps_per_second": 0.421, "step": 3000 }, { "epoch": 0.21333900347296053, "grad_norm": 0.44140625, "learning_rate": 0.0001, "loss": 1.4859, "step": 3010 }, { "epoch": 0.21404777092635907, "grad_norm": 0.54296875, "learning_rate": 0.0001, "loss": 1.4907, "step": 3020 }, { "epoch": 0.2147565383797576, "grad_norm": 0.474609375, "learning_rate": 0.0001, "loss": 1.4771, "step": 3030 }, { "epoch": 0.21546530583315615, "grad_norm": 0.51953125, "learning_rate": 0.0001, "loss": 1.4868, "step": 3040 }, { "epoch": 0.2161740732865547, "grad_norm": 0.458984375, "learning_rate": 0.0001, "loss": 1.4941, "step": 3050 }, { "epoch": 0.21688284073995323, "grad_norm": 0.486328125, "learning_rate": 0.0001, "loss": 1.4951, "step": 3060 }, { "epoch": 0.21759160819335177, "grad_norm": 0.451171875, "learning_rate": 0.0001, "loss": 1.4915, "step": 3070 }, { "epoch": 0.2183003756467503, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.4869, "step": 3080 }, { "epoch": 0.21900914310014885, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.5057, "step": 3090 }, { "epoch": 0.2197179105535474, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4941, "step": 3100 }, { "epoch": 0.22042667800694593, "grad_norm": 0.48046875, "learning_rate": 0.0001, "loss": 1.5017, "step": 3110 }, { "epoch": 0.22113544546034447, "grad_norm": 0.447265625, "learning_rate": 0.0001, "loss": 1.511, "step": 3120 }, { "epoch": 0.221844212913743, "grad_norm": 0.546875, "learning_rate": 0.0001, "loss": 1.4903, "step": 3130 }, { "epoch": 0.22255298036714155, "grad_norm": 0.48828125, "learning_rate": 0.0001, "loss": 1.5098, "step": 3140 }, { "epoch": 0.2232617478205401, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4989, "step": 3150 }, { "epoch": 0.22397051527393863, "grad_norm": 0.478515625, "learning_rate": 0.0001, "loss": 1.4935, "step": 3160 }, { "epoch": 0.22467928272733717, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4907, "step": 3170 }, { "epoch": 0.2253880501807357, "grad_norm": 0.494140625, "learning_rate": 0.0001, "loss": 1.5033, "step": 3180 }, { "epoch": 0.22609681763413425, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4982, "step": 3190 }, { "epoch": 0.2268055850875328, "grad_norm": 0.546875, "learning_rate": 0.0001, "loss": 1.4853, "step": 3200 }, { "epoch": 0.22751435254093133, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.498, "step": 3210 }, { "epoch": 0.22822311999432987, "grad_norm": 0.6171875, "learning_rate": 0.0001, "loss": 1.489, "step": 3220 }, { "epoch": 0.2289318874477284, "grad_norm": 0.435546875, "learning_rate": 0.0001, "loss": 1.4758, "step": 3230 }, { "epoch": 0.22964065490112695, "grad_norm": 0.69921875, "learning_rate": 0.0001, "loss": 1.4896, "step": 3240 }, { "epoch": 0.2303494223545255, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.479, "step": 3250 }, { "epoch": 0.23105818980792403, "grad_norm": 0.478515625, "learning_rate": 0.0001, "loss": 1.4968, "step": 3260 }, { "epoch": 0.23176695726132257, "grad_norm": 0.546875, "learning_rate": 0.0001, "loss": 1.4913, "step": 3270 }, { "epoch": 0.2324757247147211, "grad_norm": 0.45703125, "learning_rate": 0.0001, "loss": 1.4856, "step": 3280 }, { "epoch": 0.23318449216811965, "grad_norm": 0.48828125, "learning_rate": 0.0001, "loss": 1.4969, "step": 3290 }, { "epoch": 0.2338932596215182, "grad_norm": 0.46484375, "learning_rate": 0.0001, "loss": 1.4968, "step": 3300 }, { "epoch": 0.23460202707491673, "grad_norm": 0.4453125, "learning_rate": 0.0001, "loss": 1.4904, "step": 3310 }, { "epoch": 0.23531079452831527, "grad_norm": 0.5546875, "learning_rate": 0.0001, "loss": 1.484, "step": 3320 }, { "epoch": 0.2360195619817138, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4928, "step": 3330 }, { "epoch": 0.23672832943511235, "grad_norm": 0.54296875, "learning_rate": 0.0001, "loss": 1.4964, "step": 3340 }, { "epoch": 0.2374370968885109, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.4974, "step": 3350 }, { "epoch": 0.23814586434190943, "grad_norm": 0.498046875, "learning_rate": 0.0001, "loss": 1.503, "step": 3360 }, { "epoch": 0.23885463179530797, "grad_norm": 0.462890625, "learning_rate": 0.0001, "loss": 1.5029, "step": 3370 }, { "epoch": 0.2395633992487065, "grad_norm": 0.46875, "learning_rate": 0.0001, "loss": 1.501, "step": 3380 }, { "epoch": 0.24027216670210505, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4885, "step": 3390 }, { "epoch": 0.2409809341555036, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.4904, "step": 3400 }, { "epoch": 0.24168970160890213, "grad_norm": 0.494140625, "learning_rate": 0.0001, "loss": 1.4851, "step": 3410 }, { "epoch": 0.24239846906230067, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.4783, "step": 3420 }, { "epoch": 0.2431072365156992, "grad_norm": 0.45703125, "learning_rate": 0.0001, "loss": 1.4948, "step": 3430 }, { "epoch": 0.24381600396909775, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.4754, "step": 3440 }, { "epoch": 0.2445247714224963, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4936, "step": 3450 }, { "epoch": 0.24523353887589483, "grad_norm": 0.447265625, "learning_rate": 0.0001, "loss": 1.4814, "step": 3460 }, { "epoch": 0.24594230632929337, "grad_norm": 0.53515625, "learning_rate": 0.0001, "loss": 1.505, "step": 3470 }, { "epoch": 0.2466510737826919, "grad_norm": 0.609375, "learning_rate": 0.0001, "loss": 1.4888, "step": 3480 }, { "epoch": 0.24735984123609045, "grad_norm": 0.486328125, "learning_rate": 0.0001, "loss": 1.4974, "step": 3490 }, { "epoch": 0.248068608689489, "grad_norm": 0.4453125, "learning_rate": 0.0001, "loss": 1.496, "step": 3500 }, { "epoch": 0.248068608689489, "eval_accuracy": 0.665728571780445, "eval_loss": 1.403409481048584, "eval_runtime": 7.1308, "eval_samples_per_second": 48.662, "eval_steps_per_second": 0.421, "step": 3500 }, { "epoch": 0.24877737614288753, "grad_norm": 0.462890625, "learning_rate": 0.0001, "loss": 1.4988, "step": 3510 }, { "epoch": 0.24948614359628607, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.5036, "step": 3520 }, { "epoch": 0.2501949110496846, "grad_norm": 0.609375, "learning_rate": 0.0001, "loss": 1.5038, "step": 3530 }, { "epoch": 0.25090367850308315, "grad_norm": 0.48828125, "learning_rate": 0.0001, "loss": 1.4947, "step": 3540 }, { "epoch": 0.2516124459564817, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.4914, "step": 3550 }, { "epoch": 0.25232121340988023, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4855, "step": 3560 }, { "epoch": 0.25302998086327877, "grad_norm": 0.498046875, "learning_rate": 0.0001, "loss": 1.4845, "step": 3570 }, { "epoch": 0.2537387483166773, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.5041, "step": 3580 }, { "epoch": 0.25444751577007585, "grad_norm": 0.5703125, "learning_rate": 0.0001, "loss": 1.4803, "step": 3590 }, { "epoch": 0.2551562832234744, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4938, "step": 3600 }, { "epoch": 0.25586505067687293, "grad_norm": 0.46875, "learning_rate": 0.0001, "loss": 1.4938, "step": 3610 }, { "epoch": 0.25657381813027147, "grad_norm": 0.5859375, "learning_rate": 0.0001, "loss": 1.4972, "step": 3620 }, { "epoch": 0.25728258558367, "grad_norm": 0.458984375, "learning_rate": 0.0001, "loss": 1.4941, "step": 3630 }, { "epoch": 0.25799135303706855, "grad_norm": 0.451171875, "learning_rate": 0.0001, "loss": 1.4881, "step": 3640 }, { "epoch": 0.2587001204904671, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4906, "step": 3650 }, { "epoch": 0.25940888794386563, "grad_norm": 0.474609375, "learning_rate": 0.0001, "loss": 1.4937, "step": 3660 }, { "epoch": 0.26011765539726417, "grad_norm": 0.55078125, "learning_rate": 0.0001, "loss": 1.4904, "step": 3670 }, { "epoch": 0.2608264228506627, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.5034, "step": 3680 }, { "epoch": 0.26153519030406125, "grad_norm": 0.474609375, "learning_rate": 0.0001, "loss": 1.4727, "step": 3690 }, { "epoch": 0.2622439577574598, "grad_norm": 0.421875, "learning_rate": 0.0001, "loss": 1.4919, "step": 3700 }, { "epoch": 0.26295272521085833, "grad_norm": 0.43359375, "learning_rate": 0.0001, "loss": 1.4842, "step": 3710 }, { "epoch": 0.26366149266425687, "grad_norm": 0.6015625, "learning_rate": 0.0001, "loss": 1.4953, "step": 3720 }, { "epoch": 0.2643702601176554, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.4739, "step": 3730 }, { "epoch": 0.26507902757105395, "grad_norm": 0.447265625, "learning_rate": 0.0001, "loss": 1.4913, "step": 3740 }, { "epoch": 0.2657877950244525, "grad_norm": 0.46875, "learning_rate": 0.0001, "loss": 1.5003, "step": 3750 }, { "epoch": 0.26649656247785103, "grad_norm": 0.451171875, "learning_rate": 0.0001, "loss": 1.4919, "step": 3760 }, { "epoch": 0.26720532993124957, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4818, "step": 3770 }, { "epoch": 0.2679140973846481, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.5003, "step": 3780 }, { "epoch": 0.26862286483804665, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4934, "step": 3790 }, { "epoch": 0.2693316322914452, "grad_norm": 0.59375, "learning_rate": 0.0001, "loss": 1.4869, "step": 3800 }, { "epoch": 0.27004039974484373, "grad_norm": 0.5625, "learning_rate": 0.0001, "loss": 1.4896, "step": 3810 }, { "epoch": 0.27074916719824227, "grad_norm": 0.46875, "learning_rate": 0.0001, "loss": 1.5021, "step": 3820 }, { "epoch": 0.2714579346516408, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.5048, "step": 3830 }, { "epoch": 0.27216670210503935, "grad_norm": 0.474609375, "learning_rate": 0.0001, "loss": 1.4963, "step": 3840 }, { "epoch": 0.2728754695584379, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.4967, "step": 3850 }, { "epoch": 0.27358423701183643, "grad_norm": 0.4921875, "learning_rate": 0.0001, "loss": 1.5036, "step": 3860 }, { "epoch": 0.27429300446523497, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.4944, "step": 3870 }, { "epoch": 0.2750017719186335, "grad_norm": 0.47265625, "learning_rate": 0.0001, "loss": 1.4951, "step": 3880 }, { "epoch": 0.27571053937203205, "grad_norm": 0.474609375, "learning_rate": 0.0001, "loss": 1.5006, "step": 3890 }, { "epoch": 0.2764193068254306, "grad_norm": 0.59765625, "learning_rate": 0.0001, "loss": 1.4886, "step": 3900 }, { "epoch": 0.27712807427882913, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.4931, "step": 3910 }, { "epoch": 0.27783684173222767, "grad_norm": 0.55078125, "learning_rate": 0.0001, "loss": 1.4926, "step": 3920 }, { "epoch": 0.2785456091856262, "grad_norm": 0.51953125, "learning_rate": 0.0001, "loss": 1.4962, "step": 3930 }, { "epoch": 0.27925437663902475, "grad_norm": 0.4921875, "learning_rate": 0.0001, "loss": 1.4933, "step": 3940 }, { "epoch": 0.2799631440924233, "grad_norm": 0.48046875, "learning_rate": 0.0001, "loss": 1.4959, "step": 3950 }, { "epoch": 0.28067191154582183, "grad_norm": 0.56640625, "learning_rate": 0.0001, "loss": 1.49, "step": 3960 }, { "epoch": 0.28138067899922037, "grad_norm": 0.5859375, "learning_rate": 0.0001, "loss": 1.5087, "step": 3970 }, { "epoch": 0.2820894464526189, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.4939, "step": 3980 }, { "epoch": 0.28279821390601745, "grad_norm": 0.4375, "learning_rate": 0.0001, "loss": 1.4913, "step": 3990 }, { "epoch": 0.283506981359416, "grad_norm": 0.443359375, "learning_rate": 0.0001, "loss": 1.495, "step": 4000 }, { "epoch": 0.283506981359416, "eval_accuracy": 0.6657313867688508, "eval_loss": 1.403212308883667, "eval_runtime": 7.176, "eval_samples_per_second": 48.355, "eval_steps_per_second": 0.418, "step": 4000 }, { "epoch": 0.28421574881281453, "grad_norm": 0.46484375, "learning_rate": 0.0001, "loss": 1.4793, "step": 4010 }, { "epoch": 0.28492451626621307, "grad_norm": 0.5859375, "learning_rate": 0.0001, "loss": 1.484, "step": 4020 }, { "epoch": 0.2856332837196116, "grad_norm": 0.578125, "learning_rate": 0.0001, "loss": 1.4967, "step": 4030 }, { "epoch": 0.28634205117301015, "grad_norm": 0.494140625, "learning_rate": 0.0001, "loss": 1.4952, "step": 4040 }, { "epoch": 0.2870508186264087, "grad_norm": 0.5859375, "learning_rate": 0.0001, "loss": 1.494, "step": 4050 }, { "epoch": 0.28775958607980723, "grad_norm": 0.48046875, "learning_rate": 0.0001, "loss": 1.4888, "step": 4060 }, { "epoch": 0.28846835353320577, "grad_norm": 0.482421875, "learning_rate": 0.0001, "loss": 1.4784, "step": 4070 }, { "epoch": 0.2891771209866043, "grad_norm": 0.55078125, "learning_rate": 0.0001, "loss": 1.4924, "step": 4080 }, { "epoch": 0.28988588844000285, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.5047, "step": 4090 }, { "epoch": 0.2905946558934014, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.4895, "step": 4100 }, { "epoch": 0.29130342334679993, "grad_norm": 0.466796875, "learning_rate": 0.0001, "loss": 1.4852, "step": 4110 }, { "epoch": 0.29201219080019847, "grad_norm": 0.5625, "learning_rate": 0.0001, "loss": 1.5075, "step": 4120 }, { "epoch": 0.292720958253597, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.5061, "step": 4130 }, { "epoch": 0.29342972570699555, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4953, "step": 4140 }, { "epoch": 0.2941384931603941, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4956, "step": 4150 }, { "epoch": 0.2948472606137926, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4978, "step": 4160 }, { "epoch": 0.29555602806719117, "grad_norm": 0.7421875, "learning_rate": 0.0001, "loss": 1.4759, "step": 4170 }, { "epoch": 0.2962647955205897, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.4955, "step": 4180 }, { "epoch": 0.29697356297398825, "grad_norm": 0.447265625, "learning_rate": 0.0001, "loss": 1.4955, "step": 4190 }, { "epoch": 0.2976823304273868, "grad_norm": 0.458984375, "learning_rate": 0.0001, "loss": 1.482, "step": 4200 }, { "epoch": 0.2983910978807853, "grad_norm": 0.455078125, "learning_rate": 0.0001, "loss": 1.4774, "step": 4210 }, { "epoch": 0.29909986533418387, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.484, "step": 4220 }, { "epoch": 0.2998086327875824, "grad_norm": 0.53515625, "learning_rate": 0.0001, "loss": 1.5073, "step": 4230 }, { "epoch": 0.30051740024098095, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.5047, "step": 4240 }, { "epoch": 0.3012261676943795, "grad_norm": 0.546875, "learning_rate": 0.0001, "loss": 1.4898, "step": 4250 }, { "epoch": 0.301934935147778, "grad_norm": 0.4453125, "learning_rate": 0.0001, "loss": 1.4914, "step": 4260 }, { "epoch": 0.30264370260117657, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4838, "step": 4270 }, { "epoch": 0.3033524700545751, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.4908, "step": 4280 }, { "epoch": 0.30406123750797365, "grad_norm": 0.486328125, "learning_rate": 0.0001, "loss": 1.4838, "step": 4290 }, { "epoch": 0.3047700049613722, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4871, "step": 4300 }, { "epoch": 0.3054787724147707, "grad_norm": 0.48828125, "learning_rate": 0.0001, "loss": 1.5091, "step": 4310 }, { "epoch": 0.30618753986816927, "grad_norm": 0.4296875, "learning_rate": 0.0001, "loss": 1.4968, "step": 4320 }, { "epoch": 0.3068963073215678, "grad_norm": 0.5703125, "learning_rate": 0.0001, "loss": 1.4926, "step": 4330 }, { "epoch": 0.30760507477496635, "grad_norm": 0.55078125, "learning_rate": 0.0001, "loss": 1.4982, "step": 4340 }, { "epoch": 0.3083138422283649, "grad_norm": 0.5546875, "learning_rate": 0.0001, "loss": 1.5001, "step": 4350 }, { "epoch": 0.3090226096817634, "grad_norm": 0.43359375, "learning_rate": 0.0001, "loss": 1.4939, "step": 4360 }, { "epoch": 0.30973137713516197, "grad_norm": 0.49609375, "learning_rate": 0.0001, "loss": 1.4979, "step": 4370 }, { "epoch": 0.3104401445885605, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.4799, "step": 4380 }, { "epoch": 0.31114891204195905, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4945, "step": 4390 }, { "epoch": 0.3118576794953576, "grad_norm": 0.458984375, "learning_rate": 0.0001, "loss": 1.4948, "step": 4400 }, { "epoch": 0.3125664469487561, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4942, "step": 4410 }, { "epoch": 0.31327521440215467, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.5039, "step": 4420 }, { "epoch": 0.3139839818555532, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.4882, "step": 4430 }, { "epoch": 0.31469274930895175, "grad_norm": 0.4296875, "learning_rate": 0.0001, "loss": 1.4925, "step": 4440 }, { "epoch": 0.3154015167623503, "grad_norm": 0.65625, "learning_rate": 0.0001, "loss": 1.5131, "step": 4450 }, { "epoch": 0.3161102842157488, "grad_norm": 0.53515625, "learning_rate": 0.0001, "loss": 1.5026, "step": 4460 }, { "epoch": 0.31681905166914737, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4959, "step": 4470 }, { "epoch": 0.3175278191225459, "grad_norm": 0.55859375, "learning_rate": 0.0001, "loss": 1.4972, "step": 4480 }, { "epoch": 0.31823658657594445, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4922, "step": 4490 }, { "epoch": 0.318945354029343, "grad_norm": 0.57421875, "learning_rate": 0.0001, "loss": 1.4934, "step": 4500 }, { "epoch": 0.318945354029343, "eval_accuracy": 0.6658446900521828, "eval_loss": 1.4029836654663086, "eval_runtime": 7.1758, "eval_samples_per_second": 48.357, "eval_steps_per_second": 0.418, "step": 4500 }, { "epoch": 0.3196541214827415, "grad_norm": 0.57421875, "learning_rate": 0.0001, "loss": 1.4693, "step": 4510 }, { "epoch": 0.32036288893614007, "grad_norm": 0.486328125, "learning_rate": 0.0001, "loss": 1.498, "step": 4520 }, { "epoch": 0.3210716563895386, "grad_norm": 0.462890625, "learning_rate": 0.0001, "loss": 1.4872, "step": 4530 }, { "epoch": 0.32178042384293715, "grad_norm": 0.5546875, "learning_rate": 0.0001, "loss": 1.4866, "step": 4540 }, { "epoch": 0.3224891912963357, "grad_norm": 0.51953125, "learning_rate": 0.0001, "loss": 1.4919, "step": 4550 }, { "epoch": 0.3231979587497342, "grad_norm": 0.46875, "learning_rate": 0.0001, "loss": 1.4765, "step": 4560 }, { "epoch": 0.32390672620313277, "grad_norm": 0.494140625, "learning_rate": 0.0001, "loss": 1.4919, "step": 4570 }, { "epoch": 0.3246154936565313, "grad_norm": 0.494140625, "learning_rate": 0.0001, "loss": 1.5069, "step": 4580 }, { "epoch": 0.32532426110992985, "grad_norm": 0.546875, "learning_rate": 0.0001, "loss": 1.5014, "step": 4590 }, { "epoch": 0.3260330285633284, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.5081, "step": 4600 }, { "epoch": 0.3267417960167269, "grad_norm": 0.48046875, "learning_rate": 0.0001, "loss": 1.5007, "step": 4610 }, { "epoch": 0.32745056347012547, "grad_norm": 0.498046875, "learning_rate": 0.0001, "loss": 1.4755, "step": 4620 }, { "epoch": 0.328159330923524, "grad_norm": 0.48828125, "learning_rate": 0.0001, "loss": 1.4875, "step": 4630 }, { "epoch": 0.32886809837692255, "grad_norm": 0.546875, "learning_rate": 0.0001, "loss": 1.4871, "step": 4640 }, { "epoch": 0.3295768658303211, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.5034, "step": 4650 }, { "epoch": 0.3302856332837196, "grad_norm": 0.470703125, "learning_rate": 0.0001, "loss": 1.4828, "step": 4660 }, { "epoch": 0.33099440073711817, "grad_norm": 0.4921875, "learning_rate": 0.0001, "loss": 1.489, "step": 4670 }, { "epoch": 0.3317031681905167, "grad_norm": 0.4921875, "learning_rate": 0.0001, "loss": 1.5142, "step": 4680 }, { "epoch": 0.33241193564391525, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.4918, "step": 4690 }, { "epoch": 0.3331207030973138, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4862, "step": 4700 }, { "epoch": 0.3338294705507123, "grad_norm": 0.61328125, "learning_rate": 0.0001, "loss": 1.4979, "step": 4710 }, { "epoch": 0.33453823800411087, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.4968, "step": 4720 }, { "epoch": 0.3352470054575094, "grad_norm": 0.56640625, "learning_rate": 0.0001, "loss": 1.4929, "step": 4730 }, { "epoch": 0.33595577291090795, "grad_norm": 0.54296875, "learning_rate": 0.0001, "loss": 1.4713, "step": 4740 }, { "epoch": 0.3366645403643065, "grad_norm": 0.51953125, "learning_rate": 0.0001, "loss": 1.4866, "step": 4750 }, { "epoch": 0.337373307817705, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.4993, "step": 4760 }, { "epoch": 0.33808207527110357, "grad_norm": 0.58203125, "learning_rate": 0.0001, "loss": 1.4966, "step": 4770 }, { "epoch": 0.3387908427245021, "grad_norm": 0.46484375, "learning_rate": 0.0001, "loss": 1.5022, "step": 4780 }, { "epoch": 0.33949961017790065, "grad_norm": 0.453125, "learning_rate": 0.0001, "loss": 1.4908, "step": 4790 }, { "epoch": 0.3402083776312992, "grad_norm": 0.66015625, "learning_rate": 0.0001, "loss": 1.4884, "step": 4800 }, { "epoch": 0.3409171450846977, "grad_norm": 0.44140625, "learning_rate": 0.0001, "loss": 1.5059, "step": 4810 }, { "epoch": 0.34162591253809627, "grad_norm": 0.55859375, "learning_rate": 0.0001, "loss": 1.4907, "step": 4820 }, { "epoch": 0.3423346799914948, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.4842, "step": 4830 }, { "epoch": 0.34304344744489335, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.4984, "step": 4840 }, { "epoch": 0.3437522148982919, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4828, "step": 4850 }, { "epoch": 0.3444609823516904, "grad_norm": 0.5859375, "learning_rate": 0.0001, "loss": 1.4865, "step": 4860 }, { "epoch": 0.34516974980508897, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.5001, "step": 4870 }, { "epoch": 0.3458785172584875, "grad_norm": 0.51953125, "learning_rate": 0.0001, "loss": 1.5026, "step": 4880 }, { "epoch": 0.34658728471188605, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.488, "step": 4890 }, { "epoch": 0.3472960521652846, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.4846, "step": 4900 }, { "epoch": 0.3480048196186831, "grad_norm": 0.5625, "learning_rate": 0.0001, "loss": 1.4972, "step": 4910 }, { "epoch": 0.34871358707208167, "grad_norm": 0.46484375, "learning_rate": 0.0001, "loss": 1.487, "step": 4920 }, { "epoch": 0.3494223545254802, "grad_norm": 0.474609375, "learning_rate": 0.0001, "loss": 1.4945, "step": 4930 }, { "epoch": 0.35013112197887875, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.4942, "step": 4940 }, { "epoch": 0.3508398894322773, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4944, "step": 4950 }, { "epoch": 0.3515486568856758, "grad_norm": 0.59375, "learning_rate": 0.0001, "loss": 1.5099, "step": 4960 }, { "epoch": 0.35225742433907437, "grad_norm": 0.494140625, "learning_rate": 0.0001, "loss": 1.4951, "step": 4970 }, { "epoch": 0.3529661917924729, "grad_norm": 0.46875, "learning_rate": 0.0001, "loss": 1.5049, "step": 4980 }, { "epoch": 0.35367495924587145, "grad_norm": 0.45703125, "learning_rate": 0.0001, "loss": 1.4764, "step": 4990 }, { "epoch": 0.35438372669927, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4849, "step": 5000 }, { "epoch": 0.35438372669927, "eval_accuracy": 0.6660431467347894, "eval_loss": 1.4028522968292236, "eval_runtime": 7.1426, "eval_samples_per_second": 48.581, "eval_steps_per_second": 0.42, "step": 5000 }, { "epoch": 0.3550924941526685, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.4884, "step": 5010 }, { "epoch": 0.35580126160606707, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.5002, "step": 5020 }, { "epoch": 0.3565100290594656, "grad_norm": 0.498046875, "learning_rate": 0.0001, "loss": 1.5019, "step": 5030 }, { "epoch": 0.35721879651286415, "grad_norm": 0.44921875, "learning_rate": 0.0001, "loss": 1.4813, "step": 5040 }, { "epoch": 0.3579275639662627, "grad_norm": 0.439453125, "learning_rate": 0.0001, "loss": 1.4826, "step": 5050 }, { "epoch": 0.3586363314196612, "grad_norm": 0.51953125, "learning_rate": 0.0001, "loss": 1.5028, "step": 5060 }, { "epoch": 0.35934509887305977, "grad_norm": 0.474609375, "learning_rate": 0.0001, "loss": 1.4953, "step": 5070 }, { "epoch": 0.3600538663264583, "grad_norm": 0.46875, "learning_rate": 0.0001, "loss": 1.4826, "step": 5080 }, { "epoch": 0.36076263377985684, "grad_norm": 0.458984375, "learning_rate": 0.0001, "loss": 1.4877, "step": 5090 }, { "epoch": 0.3614714012332554, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.4823, "step": 5100 }, { "epoch": 0.3621801686866539, "grad_norm": 0.55859375, "learning_rate": 0.0001, "loss": 1.4957, "step": 5110 }, { "epoch": 0.36288893614005246, "grad_norm": 0.546875, "learning_rate": 0.0001, "loss": 1.4935, "step": 5120 }, { "epoch": 0.363597703593451, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.5076, "step": 5130 }, { "epoch": 0.36430647104684954, "grad_norm": 0.51953125, "learning_rate": 0.0001, "loss": 1.4848, "step": 5140 }, { "epoch": 0.3650152385002481, "grad_norm": 0.45703125, "learning_rate": 0.0001, "loss": 1.4892, "step": 5150 }, { "epoch": 0.3657240059536466, "grad_norm": 0.490234375, "learning_rate": 0.0001, "loss": 1.4944, "step": 5160 }, { "epoch": 0.36643277340704516, "grad_norm": 0.609375, "learning_rate": 0.0001, "loss": 1.4946, "step": 5170 }, { "epoch": 0.3671415408604437, "grad_norm": 0.474609375, "learning_rate": 0.0001, "loss": 1.4826, "step": 5180 }, { "epoch": 0.36785030831384224, "grad_norm": 0.470703125, "learning_rate": 0.0001, "loss": 1.4874, "step": 5190 }, { "epoch": 0.3685590757672408, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.5055, "step": 5200 }, { "epoch": 0.3692678432206393, "grad_norm": 0.55078125, "learning_rate": 0.0001, "loss": 1.5005, "step": 5210 }, { "epoch": 0.36997661067403786, "grad_norm": 0.46875, "learning_rate": 0.0001, "loss": 1.4842, "step": 5220 }, { "epoch": 0.3706853781274364, "grad_norm": 0.451171875, "learning_rate": 0.0001, "loss": 1.501, "step": 5230 }, { "epoch": 0.37139414558083494, "grad_norm": 0.55078125, "learning_rate": 0.0001, "loss": 1.5031, "step": 5240 }, { "epoch": 0.3721029130342335, "grad_norm": 0.51953125, "learning_rate": 0.0001, "loss": 1.4887, "step": 5250 }, { "epoch": 0.372811680487632, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.4779, "step": 5260 }, { "epoch": 0.37352044794103056, "grad_norm": 0.482421875, "learning_rate": 0.0001, "loss": 1.4779, "step": 5270 }, { "epoch": 0.3742292153944291, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.5006, "step": 5280 }, { "epoch": 0.37493798284782764, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.5109, "step": 5290 }, { "epoch": 0.3756467503012262, "grad_norm": 0.451171875, "learning_rate": 0.0001, "loss": 1.4779, "step": 5300 }, { "epoch": 0.3763555177546247, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.4861, "step": 5310 }, { "epoch": 0.37706428520802326, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.5001, "step": 5320 }, { "epoch": 0.3777730526614218, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4876, "step": 5330 }, { "epoch": 0.37848182011482034, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.486, "step": 5340 }, { "epoch": 0.3791905875682189, "grad_norm": 0.4921875, "learning_rate": 0.0001, "loss": 1.4956, "step": 5350 }, { "epoch": 0.3798993550216174, "grad_norm": 0.61328125, "learning_rate": 0.0001, "loss": 1.4972, "step": 5360 }, { "epoch": 0.38060812247501596, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.4926, "step": 5370 }, { "epoch": 0.3813168899284145, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4982, "step": 5380 }, { "epoch": 0.38202565738181304, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4949, "step": 5390 }, { "epoch": 0.3827344248352116, "grad_norm": 0.56640625, "learning_rate": 0.0001, "loss": 1.4928, "step": 5400 }, { "epoch": 0.3834431922886101, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.4841, "step": 5410 }, { "epoch": 0.38415195974200866, "grad_norm": 0.58984375, "learning_rate": 0.0001, "loss": 1.4975, "step": 5420 }, { "epoch": 0.3848607271954072, "grad_norm": 0.63671875, "learning_rate": 0.0001, "loss": 1.4907, "step": 5430 }, { "epoch": 0.38556949464880574, "grad_norm": 0.453125, "learning_rate": 0.0001, "loss": 1.4933, "step": 5440 }, { "epoch": 0.3862782621022043, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.5083, "step": 5450 }, { "epoch": 0.3869870295556028, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.4766, "step": 5460 }, { "epoch": 0.38769579700900136, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.485, "step": 5470 }, { "epoch": 0.3884045644623999, "grad_norm": 0.59765625, "learning_rate": 0.0001, "loss": 1.4891, "step": 5480 }, { "epoch": 0.38911333191579844, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.495, "step": 5490 }, { "epoch": 0.389822099369197, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.4833, "step": 5500 }, { "epoch": 0.389822099369197, "eval_accuracy": 0.6661451900644985, "eval_loss": 1.4023902416229248, "eval_runtime": 7.2106, "eval_samples_per_second": 48.124, "eval_steps_per_second": 0.416, "step": 5500 }, { "epoch": 0.3905308668225955, "grad_norm": 0.466796875, "learning_rate": 0.0001, "loss": 1.5, "step": 5510 }, { "epoch": 0.39123963427599406, "grad_norm": 0.53515625, "learning_rate": 0.0001, "loss": 1.4789, "step": 5520 }, { "epoch": 0.3919484017293926, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.498, "step": 5530 }, { "epoch": 0.39265716918279114, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.5051, "step": 5540 }, { "epoch": 0.3933659366361897, "grad_norm": 0.66015625, "learning_rate": 0.0001, "loss": 1.5041, "step": 5550 }, { "epoch": 0.3940747040895882, "grad_norm": 0.458984375, "learning_rate": 0.0001, "loss": 1.4736, "step": 5560 }, { "epoch": 0.39478347154298676, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.4863, "step": 5570 }, { "epoch": 0.3954922389963853, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.4962, "step": 5580 }, { "epoch": 0.39620100644978384, "grad_norm": 0.44921875, "learning_rate": 0.0001, "loss": 1.4904, "step": 5590 }, { "epoch": 0.3969097739031824, "grad_norm": 0.54296875, "learning_rate": 0.0001, "loss": 1.4955, "step": 5600 }, { "epoch": 0.3976185413565809, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4925, "step": 5610 }, { "epoch": 0.39832730880997946, "grad_norm": 0.4921875, "learning_rate": 0.0001, "loss": 1.4945, "step": 5620 }, { "epoch": 0.399036076263378, "grad_norm": 0.5625, "learning_rate": 0.0001, "loss": 1.5, "step": 5630 }, { "epoch": 0.39974484371677654, "grad_norm": 0.6015625, "learning_rate": 0.0001, "loss": 1.4946, "step": 5640 }, { "epoch": 0.4004536111701751, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.4868, "step": 5650 }, { "epoch": 0.4011623786235736, "grad_norm": 0.52734375, "learning_rate": 0.0001, "loss": 1.4841, "step": 5660 }, { "epoch": 0.40187114607697216, "grad_norm": 0.62109375, "learning_rate": 0.0001, "loss": 1.494, "step": 5670 }, { "epoch": 0.4025799135303707, "grad_norm": 0.51171875, "learning_rate": 0.0001, "loss": 1.4844, "step": 5680 }, { "epoch": 0.40328868098376924, "grad_norm": 0.55078125, "learning_rate": 0.0001, "loss": 1.4932, "step": 5690 }, { "epoch": 0.4039974484371678, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.5066, "step": 5700 }, { "epoch": 0.4047062158905663, "grad_norm": 0.5390625, "learning_rate": 0.0001, "loss": 1.498, "step": 5710 }, { "epoch": 0.40541498334396486, "grad_norm": 0.482421875, "learning_rate": 0.0001, "loss": 1.5048, "step": 5720 }, { "epoch": 0.4061237507973634, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4863, "step": 5730 }, { "epoch": 0.40683251825076194, "grad_norm": 0.66796875, "learning_rate": 0.0001, "loss": 1.4899, "step": 5740 }, { "epoch": 0.4075412857041605, "grad_norm": 0.56640625, "learning_rate": 0.0001, "loss": 1.4792, "step": 5750 }, { "epoch": 0.408250053157559, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.4915, "step": 5760 }, { "epoch": 0.40895882061095756, "grad_norm": 0.515625, "learning_rate": 0.0001, "loss": 1.4923, "step": 5770 }, { "epoch": 0.4096675880643561, "grad_norm": 0.4921875, "learning_rate": 0.0001, "loss": 1.4757, "step": 5780 }, { "epoch": 0.41037635551775464, "grad_norm": 0.484375, "learning_rate": 0.0001, "loss": 1.4972, "step": 5790 }, { "epoch": 0.4110851229711532, "grad_norm": 0.5625, "learning_rate": 0.0001, "loss": 1.4846, "step": 5800 }, { "epoch": 0.4117938904245517, "grad_norm": 0.56640625, "learning_rate": 0.0001, "loss": 1.4888, "step": 5810 }, { "epoch": 0.41250265787795026, "grad_norm": 0.45703125, "learning_rate": 0.0001, "loss": 1.4996, "step": 5820 }, { "epoch": 0.4132114253313488, "grad_norm": 0.48828125, "learning_rate": 0.0001, "loss": 1.4927, "step": 5830 }, { "epoch": 0.41392019278474734, "grad_norm": 0.4765625, "learning_rate": 0.0001, "loss": 1.4861, "step": 5840 }, { "epoch": 0.4146289602381459, "grad_norm": 0.53125, "learning_rate": 0.0001, "loss": 1.4917, "step": 5850 }, { "epoch": 0.4153377276915444, "grad_norm": 0.4921875, "learning_rate": 0.0001, "loss": 1.5034, "step": 5860 }, { "epoch": 0.41604649514494296, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.5052, "step": 5870 }, { "epoch": 0.4167552625983415, "grad_norm": 0.447265625, "learning_rate": 0.0001, "loss": 1.5002, "step": 5880 }, { "epoch": 0.41746403005174004, "grad_norm": 0.47265625, "learning_rate": 0.0001, "loss": 1.5054, "step": 5890 }, { "epoch": 0.4181727975051386, "grad_norm": 0.5546875, "learning_rate": 0.0001, "loss": 1.498, "step": 5900 }, { "epoch": 0.4188815649585371, "grad_norm": 0.48046875, "learning_rate": 0.0001, "loss": 1.4894, "step": 5910 }, { "epoch": 0.41959033241193566, "grad_norm": 0.59765625, "learning_rate": 0.0001, "loss": 1.5121, "step": 5920 }, { "epoch": 0.4202990998653342, "grad_norm": 0.48828125, "learning_rate": 0.0001, "loss": 1.4975, "step": 5930 }, { "epoch": 0.42100786731873274, "grad_norm": 0.6484375, "learning_rate": 0.0001, "loss": 1.4825, "step": 5940 }, { "epoch": 0.4217166347721313, "grad_norm": 0.609375, "learning_rate": 0.0001, "loss": 1.4985, "step": 5950 }, { "epoch": 0.4224254022255298, "grad_norm": 0.5078125, "learning_rate": 0.0001, "loss": 1.4922, "step": 5960 }, { "epoch": 0.42313416967892836, "grad_norm": 0.50390625, "learning_rate": 0.0001, "loss": 1.4964, "step": 5970 }, { "epoch": 0.4238429371323269, "grad_norm": 0.53515625, "learning_rate": 0.0001, "loss": 1.4982, "step": 5980 }, { "epoch": 0.42455170458572544, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4778, "step": 5990 }, { "epoch": 0.425260472039124, "grad_norm": 0.5234375, "learning_rate": 0.0001, "loss": 1.4909, "step": 6000 }, { "epoch": 0.425260472039124, "eval_accuracy": 0.6661219664101509, "eval_loss": 1.4022555351257324, "eval_runtime": 7.2496, "eval_samples_per_second": 47.865, "eval_steps_per_second": 0.414, "step": 6000 }, { "epoch": 0.4259692394925225, "grad_norm": 0.48046875, "learning_rate": 0.0001, "loss": 1.5043, "step": 6010 }, { "epoch": 0.42667800694592106, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.507, "step": 6020 }, { "epoch": 0.4273867743993196, "grad_norm": 0.4453125, "learning_rate": 0.0001, "loss": 1.4907, "step": 6030 }, { "epoch": 0.42809554185271814, "grad_norm": 0.5, "learning_rate": 0.0001, "loss": 1.491, "step": 6040 }, { "epoch": 0.4288043093061167, "grad_norm": 0.5859375, "learning_rate": 0.0001, "loss": 1.5023, "step": 6050 }, { "epoch": 0.4295130767595152, "grad_norm": 0.486328125, "learning_rate": 9.997960964140947e-05, "loss": 1.4967, "step": 6060 }, { "epoch": 0.43022184421291376, "grad_norm": 0.4921875, "learning_rate": 9.991845519630678e-05, "loss": 1.4897, "step": 6070 }, { "epoch": 0.4309306116663123, "grad_norm": 0.5703125, "learning_rate": 9.981658654313457e-05, "loss": 1.4917, "step": 6080 }, { "epoch": 0.43163937911971084, "grad_norm": 0.5, "learning_rate": 9.967408676742751e-05, "loss": 1.5015, "step": 6090 }, { "epoch": 0.4323481465731094, "grad_norm": 0.486328125, "learning_rate": 9.949107209404665e-05, "loss": 1.4798, "step": 6100 }, { "epoch": 0.4330569140265079, "grad_norm": 0.4453125, "learning_rate": 9.926769179238466e-05, "loss": 1.4874, "step": 6110 }, { "epoch": 0.43376568147990646, "grad_norm": 0.62109375, "learning_rate": 9.900412805461967e-05, "loss": 1.4951, "step": 6120 }, { "epoch": 0.434474448933305, "grad_norm": 0.47265625, "learning_rate": 9.870059584711668e-05, "loss": 1.4917, "step": 6130 }, { "epoch": 0.43518321638670354, "grad_norm": 0.51953125, "learning_rate": 9.835734273509786e-05, "loss": 1.5102, "step": 6140 }, { "epoch": 0.4358919838401021, "grad_norm": 0.482421875, "learning_rate": 9.797464868072488e-05, "loss": 1.485, "step": 6150 }, { "epoch": 0.4366007512935006, "grad_norm": 0.5234375, "learning_rate": 9.755282581475769e-05, "loss": 1.4967, "step": 6160 }, { "epoch": 0.43730951874689916, "grad_norm": 0.4609375, "learning_rate": 9.709221818197624e-05, "loss": 1.4827, "step": 6170 }, { "epoch": 0.4380182862002977, "grad_norm": 0.56640625, "learning_rate": 9.659320146057262e-05, "loss": 1.4919, "step": 6180 }, { "epoch": 0.43872705365369624, "grad_norm": 0.4921875, "learning_rate": 9.60561826557425e-05, "loss": 1.5128, "step": 6190 }, { "epoch": 0.4394358211070948, "grad_norm": 0.44921875, "learning_rate": 9.548159976772592e-05, "loss": 1.496, "step": 6200 }, { "epoch": 0.4401445885604933, "grad_norm": 0.51953125, "learning_rate": 9.486992143456792e-05, "loss": 1.4835, "step": 6210 }, { "epoch": 0.44085335601389186, "grad_norm": 0.45703125, "learning_rate": 9.422164654989072e-05, "loss": 1.4907, "step": 6220 }, { "epoch": 0.4415621234672904, "grad_norm": 0.5703125, "learning_rate": 9.353730385598887e-05, "loss": 1.4982, "step": 6230 }, { "epoch": 0.44227089092068894, "grad_norm": 0.447265625, "learning_rate": 9.281745151257946e-05, "loss": 1.493, "step": 6240 }, { "epoch": 0.4429796583740875, "grad_norm": 0.48046875, "learning_rate": 9.206267664155907e-05, "loss": 1.4849, "step": 6250 }, { "epoch": 0.443688425827486, "grad_norm": 0.484375, "learning_rate": 9.12735948481387e-05, "loss": 1.495, "step": 6260 }, { "epoch": 0.44439719328088456, "grad_norm": 0.47265625, "learning_rate": 9.045084971874738e-05, "loss": 1.4908, "step": 6270 }, { "epoch": 0.4451059607342831, "grad_norm": 0.462890625, "learning_rate": 8.959511229611376e-05, "loss": 1.4864, "step": 6280 }, { "epoch": 0.44581472818768164, "grad_norm": 0.48046875, "learning_rate": 8.870708053195413e-05, "loss": 1.4839, "step": 6290 }, { "epoch": 0.4465234956410802, "grad_norm": 0.4453125, "learning_rate": 8.778747871771292e-05, "loss": 1.4995, "step": 6300 }, { "epoch": 0.4472322630944787, "grad_norm": 0.45703125, "learning_rate": 8.683705689382024e-05, "loss": 1.4856, "step": 6310 }, { "epoch": 0.44794103054787726, "grad_norm": 0.51953125, "learning_rate": 8.585659023794818e-05, "loss": 1.4933, "step": 6320 }, { "epoch": 0.4486497980012758, "grad_norm": 0.490234375, "learning_rate": 8.484687843276469e-05, "loss": 1.4798, "step": 6330 }, { "epoch": 0.44935856545467434, "grad_norm": 0.439453125, "learning_rate": 8.380874501370097e-05, "loss": 1.4972, "step": 6340 }, { "epoch": 0.4500673329080729, "grad_norm": 0.45703125, "learning_rate": 8.274303669726426e-05, "loss": 1.482, "step": 6350 }, { "epoch": 0.4507761003614714, "grad_norm": 0.44140625, "learning_rate": 8.165062269044353e-05, "loss": 1.4946, "step": 6360 }, { "epoch": 0.45148486781486996, "grad_norm": 0.443359375, "learning_rate": 8.053239398177191e-05, "loss": 1.487, "step": 6370 }, { "epoch": 0.4521936352682685, "grad_norm": 0.51953125, "learning_rate": 7.938926261462366e-05, "loss": 1.5043, "step": 6380 }, { "epoch": 0.45290240272166704, "grad_norm": 0.4921875, "learning_rate": 7.822216094333847e-05, "loss": 1.4912, "step": 6390 }, { "epoch": 0.4536111701750656, "grad_norm": 0.455078125, "learning_rate": 7.703204087277988e-05, "loss": 1.4895, "step": 6400 }, { "epoch": 0.4543199376284641, "grad_norm": 0.470703125, "learning_rate": 7.58198730819481e-05, "loss": 1.5003, "step": 6410 }, { "epoch": 0.45502870508186266, "grad_norm": 0.5078125, "learning_rate": 7.45866462322802e-05, "loss": 1.4872, "step": 6420 }, { "epoch": 0.4557374725352612, "grad_norm": 0.435546875, "learning_rate": 7.333336616128369e-05, "loss": 1.4993, "step": 6430 }, { "epoch": 0.45644623998865974, "grad_norm": 0.396484375, "learning_rate": 7.206105506216106e-05, "loss": 1.5037, "step": 6440 }, { "epoch": 0.4571550074420583, "grad_norm": 0.43359375, "learning_rate": 7.077075065009433e-05, "loss": 1.5001, "step": 6450 }, { "epoch": 0.4578637748954568, "grad_norm": 0.478515625, "learning_rate": 6.946350531586959e-05, "loss": 1.5007, "step": 6460 }, { "epoch": 0.45857254234885536, "grad_norm": 0.40625, "learning_rate": 6.814038526753205e-05, "loss": 1.4763, "step": 6470 }, { "epoch": 0.4592813098022539, "grad_norm": 0.4453125, "learning_rate": 6.680246966077151e-05, "loss": 1.4987, "step": 6480 }, { "epoch": 0.45999007725565244, "grad_norm": 0.3984375, "learning_rate": 6.545084971874738e-05, "loss": 1.4925, "step": 6490 }, { "epoch": 0.460698844709051, "grad_norm": 0.40234375, "learning_rate": 6.408662784207149e-05, "loss": 1.4923, "step": 6500 }, { "epoch": 0.460698844709051, "eval_accuracy": 0.666518879775364, "eval_loss": 1.39996337890625, "eval_runtime": 7.2215, "eval_samples_per_second": 48.051, "eval_steps_per_second": 0.415, "step": 6500 }, { "epoch": 0.4614076121624495, "grad_norm": 0.42578125, "learning_rate": 6.271091670967436e-05, "loss": 1.4809, "step": 6510 }, { "epoch": 0.46211637961584806, "grad_norm": 0.41015625, "learning_rate": 6.132483837128823e-05, "loss": 1.4809, "step": 6520 }, { "epoch": 0.4628251470692466, "grad_norm": 0.41796875, "learning_rate": 5.992952333228728e-05, "loss": 1.5065, "step": 6530 }, { "epoch": 0.46353391452264514, "grad_norm": 0.453125, "learning_rate": 5.85261096316312e-05, "loss": 1.4931, "step": 6540 }, { "epoch": 0.4642426819760437, "grad_norm": 0.431640625, "learning_rate": 5.7115741913664264e-05, "loss": 1.4766, "step": 6550 }, { "epoch": 0.4649514494294422, "grad_norm": 0.412109375, "learning_rate": 5.569957049452703e-05, "loss": 1.49, "step": 6560 }, { "epoch": 0.46566021688284076, "grad_norm": 0.423828125, "learning_rate": 5.427875042394199e-05, "loss": 1.4955, "step": 6570 }, { "epoch": 0.4663689843362393, "grad_norm": 0.4375, "learning_rate": 5.2854440543138406e-05, "loss": 1.4972, "step": 6580 }, { "epoch": 0.46707775178963784, "grad_norm": 0.408203125, "learning_rate": 5.142780253968481e-05, "loss": 1.4879, "step": 6590 }, { "epoch": 0.4677865192430364, "grad_norm": 0.380859375, "learning_rate": 5e-05, "loss": 1.493, "step": 6600 }, { "epoch": 0.4684952866964349, "grad_norm": 0.3671875, "learning_rate": 4.85721974603152e-05, "loss": 1.4995, "step": 6610 }, { "epoch": 0.46920405414983346, "grad_norm": 0.408203125, "learning_rate": 4.71455594568616e-05, "loss": 1.4771, "step": 6620 }, { "epoch": 0.469912821603232, "grad_norm": 0.419921875, "learning_rate": 4.5721249576058027e-05, "loss": 1.4815, "step": 6630 }, { "epoch": 0.47062158905663054, "grad_norm": 0.396484375, "learning_rate": 4.4300429505472976e-05, "loss": 1.4733, "step": 6640 }, { "epoch": 0.4713303565100291, "grad_norm": 0.365234375, "learning_rate": 4.288425808633575e-05, "loss": 1.4976, "step": 6650 }, { "epoch": 0.4720391239634276, "grad_norm": 0.35546875, "learning_rate": 4.147389036836881e-05, "loss": 1.4884, "step": 6660 }, { "epoch": 0.47274789141682616, "grad_norm": 0.375, "learning_rate": 4.007047666771274e-05, "loss": 1.4969, "step": 6670 }, { "epoch": 0.4734566588702247, "grad_norm": 0.375, "learning_rate": 3.8675161628711776e-05, "loss": 1.5137, "step": 6680 }, { "epoch": 0.47416542632362324, "grad_norm": 0.365234375, "learning_rate": 3.728908329032567e-05, "loss": 1.4889, "step": 6690 }, { "epoch": 0.4748741937770218, "grad_norm": 0.408203125, "learning_rate": 3.591337215792852e-05, "loss": 1.4875, "step": 6700 }, { "epoch": 0.4755829612304203, "grad_norm": 0.349609375, "learning_rate": 3.4549150281252636e-05, "loss": 1.5153, "step": 6710 }, { "epoch": 0.47629172868381886, "grad_norm": 0.353515625, "learning_rate": 3.3197530339228487e-05, "loss": 1.4904, "step": 6720 }, { "epoch": 0.4770004961372174, "grad_norm": 0.349609375, "learning_rate": 3.1859614732467954e-05, "loss": 1.4851, "step": 6730 }, { "epoch": 0.47770926359061594, "grad_norm": 0.34765625, "learning_rate": 3.053649468413043e-05, "loss": 1.4979, "step": 6740 }, { "epoch": 0.4784180310440145, "grad_norm": 0.33203125, "learning_rate": 2.9229249349905684e-05, "loss": 1.492, "step": 6750 }, { "epoch": 0.479126798497413, "grad_norm": 0.345703125, "learning_rate": 2.7938944937838923e-05, "loss": 1.4843, "step": 6760 }, { "epoch": 0.47983556595081156, "grad_norm": 0.34765625, "learning_rate": 2.6666633838716314e-05, "loss": 1.4781, "step": 6770 }, { "epoch": 0.4805443334042101, "grad_norm": 0.349609375, "learning_rate": 2.5413353767719805e-05, "loss": 1.4949, "step": 6780 }, { "epoch": 0.48125310085760864, "grad_norm": 0.35546875, "learning_rate": 2.418012691805191e-05, "loss": 1.4961, "step": 6790 }, { "epoch": 0.4819618683110072, "grad_norm": 0.38671875, "learning_rate": 2.296795912722014e-05, "loss": 1.4942, "step": 6800 }, { "epoch": 0.4826706357644057, "grad_norm": 0.333984375, "learning_rate": 2.1777839056661554e-05, "loss": 1.49, "step": 6810 }, { "epoch": 0.48337940321780426, "grad_norm": 0.361328125, "learning_rate": 2.061073738537635e-05, "loss": 1.5085, "step": 6820 }, { "epoch": 0.4840881706712028, "grad_norm": 0.34765625, "learning_rate": 1.946760601822809e-05, "loss": 1.4926, "step": 6830 }, { "epoch": 0.48479693812460134, "grad_norm": 0.3203125, "learning_rate": 1.8349377309556486e-05, "loss": 1.4815, "step": 6840 }, { "epoch": 0.4855057055779999, "grad_norm": 0.328125, "learning_rate": 1.725696330273575e-05, "loss": 1.5001, "step": 6850 }, { "epoch": 0.4862144730313984, "grad_norm": 0.3359375, "learning_rate": 1.619125498629904e-05, "loss": 1.481, "step": 6860 }, { "epoch": 0.48692324048479696, "grad_norm": 0.353515625, "learning_rate": 1.5153121567235335e-05, "loss": 1.4781, "step": 6870 }, { "epoch": 0.4876320079381955, "grad_norm": 0.31640625, "learning_rate": 1.414340976205183e-05, "loss": 1.4906, "step": 6880 }, { "epoch": 0.48834077539159404, "grad_norm": 0.328125, "learning_rate": 1.3162943106179749e-05, "loss": 1.5031, "step": 6890 }, { "epoch": 0.4890495428449926, "grad_norm": 0.328125, "learning_rate": 1.2212521282287092e-05, "loss": 1.4878, "step": 6900 }, { "epoch": 0.4897583102983911, "grad_norm": 0.322265625, "learning_rate": 1.1292919468045877e-05, "loss": 1.4929, "step": 6910 }, { "epoch": 0.49046707775178966, "grad_norm": 0.33984375, "learning_rate": 1.0404887703886251e-05, "loss": 1.4889, "step": 6920 }, { "epoch": 0.4911758452051882, "grad_norm": 0.359375, "learning_rate": 9.549150281252633e-06, "loss": 1.4783, "step": 6930 }, { "epoch": 0.49188461265858674, "grad_norm": 0.322265625, "learning_rate": 8.7264051518613e-06, "loss": 1.4928, "step": 6940 }, { "epoch": 0.4925933801119853, "grad_norm": 0.3359375, "learning_rate": 7.937323358440935e-06, "loss": 1.489, "step": 6950 }, { "epoch": 0.4933021475653838, "grad_norm": 0.328125, "learning_rate": 7.182548487420554e-06, "loss": 1.4899, "step": 6960 }, { "epoch": 0.49401091501878236, "grad_norm": 0.33984375, "learning_rate": 6.462696144011149e-06, "loss": 1.4859, "step": 6970 }, { "epoch": 0.4947196824721809, "grad_norm": 0.34375, "learning_rate": 5.778353450109286e-06, "loss": 1.4944, "step": 6980 }, { "epoch": 0.49542844992557944, "grad_norm": 0.33203125, "learning_rate": 5.13007856543209e-06, "loss": 1.481, "step": 6990 }, { "epoch": 0.496137217378978, "grad_norm": 0.333984375, "learning_rate": 4.5184002322740785e-06, "loss": 1.4965, "step": 7000 }, { "epoch": 0.496137217378978, "eval_accuracy": 0.6668890507507222, "eval_loss": 1.397897720336914, "eval_runtime": 7.477, "eval_samples_per_second": 46.409, "eval_steps_per_second": 0.401, "step": 7000 }, { "epoch": 0.4968459848323765, "grad_norm": 0.373046875, "learning_rate": 3.9438173442575e-06, "loss": 1.492, "step": 7010 }, { "epoch": 0.49755475228577506, "grad_norm": 0.310546875, "learning_rate": 3.406798539427386e-06, "loss": 1.4714, "step": 7020 }, { "epoch": 0.4982635197391736, "grad_norm": 0.31640625, "learning_rate": 2.9077818180237693e-06, "loss": 1.4839, "step": 7030 }, { "epoch": 0.49897228719257214, "grad_norm": 0.33203125, "learning_rate": 2.4471741852423237e-06, "loss": 1.4976, "step": 7040 }, { "epoch": 0.4996810546459707, "grad_norm": 0.328125, "learning_rate": 2.0253513192751373e-06, "loss": 1.4857, "step": 7050 }, { "epoch": 0.50003543837267, "step": 7055, "total_flos": 1.568873006179696e+20, "train_loss": 1.4944760516549913, "train_runtime": 7219.1988, "train_samples_per_second": 125.076, "train_steps_per_second": 0.977 } ], "logging_steps": 10, "max_steps": 7055, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.568873006179696e+20, "train_batch_size": 1, "trial_name": null, "trial_params": null }