{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.24999565979757296, "eval_steps": 1000, "global_step": 14400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 1.736080970816479e-05, "grad_norm": 10.5625, "learning_rate": 2e-06, "loss": 1.0, "step": 1 }, { "epoch": 0.001736080970816479, "grad_norm": 0.1513671875, "learning_rate": 0.0002, "loss": 0.3311, "step": 100 }, { "epoch": 0.003472161941632958, "grad_norm": 0.16796875, "learning_rate": 0.0004, "loss": 0.2169, "step": 200 }, { "epoch": 0.005208242912449436, "grad_norm": 0.10400390625, "learning_rate": 0.0006, "loss": 0.2032, "step": 300 }, { "epoch": 0.006944323883265916, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 0.188, "step": 400 }, { "epoch": 0.008680404854082394, "grad_norm": 0.10107421875, "learning_rate": 0.001, "loss": 0.1758, "step": 500 }, { "epoch": 0.010416485824898873, "grad_norm": 0.09521484375, "learning_rate": 0.0012, "loss": 0.1637, "step": 600 }, { "epoch": 0.012152566795715351, "grad_norm": 0.08154296875, "learning_rate": 0.0014, "loss": 0.1518, "step": 700 }, { "epoch": 0.013888647766531832, "grad_norm": 0.08642578125, "learning_rate": 0.0016, "loss": 0.1485, "step": 800 }, { "epoch": 0.01562472873734831, "grad_norm": 0.1044921875, "learning_rate": 0.0018000000000000002, "loss": 0.1433, "step": 900 }, { "epoch": 0.01736080970816479, "grad_norm": 0.05419921875, "learning_rate": 0.002, "loss": 0.139, "step": 1000 }, { "epoch": 0.01736080970816479, "eval_covost2-en-de_loss": 1.896493673324585, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 9.8697, "eval_covost2-en-de_samples_per_second": 6.485, "eval_covost2-en-de_steps_per_second": 0.811, "step": 1000 }, { "epoch": 0.01736080970816479, "eval_covost2-zh-en_loss": 3.1452860832214355, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 8.3732, "eval_covost2-zh-en_samples_per_second": 7.643, "eval_covost2-zh-en_steps_per_second": 0.955, "step": 1000 }, { "epoch": 0.01736080970816479, "eval_peoplespeech-clean-transcription_loss": 3.2206106185913086, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 9.6941, "eval_peoplespeech-clean-transcription_samples_per_second": 6.602, "eval_peoplespeech-clean-transcription_steps_per_second": 0.825, "step": 1000 }, { "epoch": 0.01909689067898127, "grad_norm": 0.059814453125, "learning_rate": 0.001999725185109816, "loss": 0.1334, "step": 1100 }, { "epoch": 0.020832971649797746, "grad_norm": 0.07373046875, "learning_rate": 0.0019989008914857113, "loss": 0.1288, "step": 1200 }, { "epoch": 0.022569052620614226, "grad_norm": 0.049560546875, "learning_rate": 0.00199752757218401, "loss": 0.1262, "step": 1300 }, { "epoch": 0.024305133591430703, "grad_norm": 0.0517578125, "learning_rate": 0.001995605982021898, "loss": 0.1222, "step": 1400 }, { "epoch": 0.026041214562247183, "grad_norm": 0.058349609375, "learning_rate": 0.0019931371771625545, "loss": 0.1193, "step": 1500 }, { "epoch": 0.027777295533063663, "grad_norm": 0.0498046875, "learning_rate": 0.001990122514534651, "loss": 0.1196, "step": 1600 }, { "epoch": 0.02951337650388014, "grad_norm": 0.05517578125, "learning_rate": 0.0019865636510865464, "loss": 0.115, "step": 1700 }, { "epoch": 0.03124945747469662, "grad_norm": 0.044677734375, "learning_rate": 0.001982462542875576, "loss": 0.115, "step": 1800 }, { "epoch": 0.0329855384455131, "grad_norm": 0.05419921875, "learning_rate": 0.001977821443992945, "loss": 0.1125, "step": 1900 }, { "epoch": 0.03472161941632958, "grad_norm": 0.047119140625, "learning_rate": 0.001972642905324813, "loss": 0.1094, "step": 2000 }, { "epoch": 0.03472161941632958, "eval_covost2-en-de_loss": 1.6700351238250732, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 8.1279, "eval_covost2-en-de_samples_per_second": 7.874, "eval_covost2-en-de_steps_per_second": 0.984, "step": 2000 }, { "epoch": 0.03472161941632958, "eval_covost2-zh-en_loss": 3.093877077102661, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 8.1488, "eval_covost2-zh-en_samples_per_second": 7.854, "eval_covost2-zh-en_steps_per_second": 0.982, "step": 2000 }, { "epoch": 0.03472161941632958, "eval_peoplespeech-clean-transcription_loss": 2.478968620300293, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 9.5507, "eval_peoplespeech-clean-transcription_samples_per_second": 6.701, "eval_peoplespeech-clean-transcription_steps_per_second": 0.838, "step": 2000 }, { "epoch": 0.036457700387146054, "grad_norm": 0.048583984375, "learning_rate": 0.0019669297731502505, "loss": 0.1077, "step": 2100 }, { "epoch": 0.03819378135796254, "grad_norm": 0.054443359375, "learning_rate": 0.00196068518757684, "loss": 0.1069, "step": 2200 }, { "epoch": 0.039929862328779014, "grad_norm": 0.047119140625, "learning_rate": 0.001953912580814779, "loss": 0.1043, "step": 2300 }, { "epoch": 0.04166594329959549, "grad_norm": 0.044921875, "learning_rate": 0.0019466156752904343, "loss": 0.1035, "step": 2400 }, { "epoch": 0.043402024270411975, "grad_norm": 0.050537109375, "learning_rate": 0.0019387984816003866, "loss": 0.1033, "step": 2500 }, { "epoch": 0.04513810524122845, "grad_norm": 0.056396484375, "learning_rate": 0.0019304652963070869, "loss": 0.102, "step": 2600 }, { "epoch": 0.04687418621204493, "grad_norm": 0.046875, "learning_rate": 0.0019216206995773372, "loss": 0.0998, "step": 2700 }, { "epoch": 0.048610267182861405, "grad_norm": 0.042236328125, "learning_rate": 0.0019122695526648968, "loss": 0.1002, "step": 2800 }, { "epoch": 0.05034634815367789, "grad_norm": 0.04638671875, "learning_rate": 0.0019024169952385887, "loss": 0.0978, "step": 2900 }, { "epoch": 0.052082429124494366, "grad_norm": 0.05126953125, "learning_rate": 0.0018920684425573864, "loss": 0.097, "step": 3000 }, { "epoch": 0.052082429124494366, "eval_covost2-en-de_loss": 1.749150276184082, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 8.1948, "eval_covost2-en-de_samples_per_second": 7.81, "eval_covost2-en-de_steps_per_second": 0.976, "step": 3000 }, { "epoch": 0.052082429124494366, "eval_covost2-zh-en_loss": 3.198117971420288, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 8.1979, "eval_covost2-zh-en_samples_per_second": 7.807, "eval_covost2-zh-en_steps_per_second": 0.976, "step": 3000 }, { "epoch": 0.052082429124494366, "eval_peoplespeech-clean-transcription_loss": 2.345036506652832, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 11.4402, "eval_peoplespeech-clean-transcription_samples_per_second": 5.594, "eval_peoplespeech-clean-transcription_steps_per_second": 0.699, "step": 3000 }, { "epoch": 0.05381851009531084, "grad_norm": 0.06494140625, "learning_rate": 0.0018812295824940284, "loss": 0.0955, "step": 3100 }, { "epoch": 0.055554591066127326, "grad_norm": 0.044677734375, "learning_rate": 0.0018699063724087904, "loss": 0.0951, "step": 3200 }, { "epoch": 0.0572906720369438, "grad_norm": 0.0390625, "learning_rate": 0.0018581050358751443, "loss": 0.0947, "step": 3300 }, { "epoch": 0.05902675300776028, "grad_norm": 0.056396484375, "learning_rate": 0.0018458320592590974, "loss": 0.0939, "step": 3400 }, { "epoch": 0.060762833978576763, "grad_norm": 0.047119140625, "learning_rate": 0.0018330941881540914, "loss": 0.0941, "step": 3500 }, { "epoch": 0.06249891494939324, "grad_norm": 0.046630859375, "learning_rate": 0.0018198984236734246, "loss": 0.0927, "step": 3600 }, { "epoch": 0.06423499592020972, "grad_norm": 0.055419921875, "learning_rate": 0.0018062520186022297, "loss": 0.0948, "step": 3700 }, { "epoch": 0.0659710768910262, "grad_norm": 0.046142578125, "learning_rate": 0.0017921624734111292, "loss": 0.09, "step": 3800 }, { "epoch": 0.06770715786184267, "grad_norm": 0.04736328125, "learning_rate": 0.001777637532133752, "loss": 0.0926, "step": 3900 }, { "epoch": 0.06944323883265915, "grad_norm": 0.048828125, "learning_rate": 0.0017626851781103819, "loss": 0.0906, "step": 4000 }, { "epoch": 0.06944323883265915, "eval_covost2-en-de_loss": 1.7936017513275146, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 8.0356, "eval_covost2-en-de_samples_per_second": 7.965, "eval_covost2-en-de_steps_per_second": 0.996, "step": 4000 }, { "epoch": 0.06944323883265915, "eval_covost2-zh-en_loss": 3.2699265480041504, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 9.5779, "eval_covost2-zh-en_samples_per_second": 6.682, "eval_covost2-zh-en_steps_per_second": 0.835, "step": 4000 }, { "epoch": 0.06944323883265915, "eval_peoplespeech-clean-transcription_loss": 2.3380110263824463, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 9.5943, "eval_peoplespeech-clean-transcription_samples_per_second": 6.671, "eval_peoplespeech-clean-transcription_steps_per_second": 0.834, "step": 4000 }, { "epoch": 0.07117931980347564, "grad_norm": 0.041259765625, "learning_rate": 0.001747313629600077, "loss": 0.0926, "step": 4100 }, { "epoch": 0.07291540077429211, "grad_norm": 0.05322265625, "learning_rate": 0.001731531335263669, "loss": 0.0907, "step": 4200 }, { "epoch": 0.07465148174510859, "grad_norm": 0.05126953125, "learning_rate": 0.0017153469695201276, "loss": 0.0898, "step": 4300 }, { "epoch": 0.07638756271592508, "grad_norm": 0.061767578125, "learning_rate": 0.0016987694277788418, "loss": 0.0876, "step": 4400 }, { "epoch": 0.07812364368674155, "grad_norm": 0.042724609375, "learning_rate": 0.001681807821550438, "loss": 0.0874, "step": 4500 }, { "epoch": 0.07985972465755803, "grad_norm": 0.05126953125, "learning_rate": 0.0016644714734388218, "loss": 0.0865, "step": 4600 }, { "epoch": 0.08159580562837451, "grad_norm": 0.042724609375, "learning_rate": 0.0016467699120171987, "loss": 0.0866, "step": 4700 }, { "epoch": 0.08333188659919098, "grad_norm": 0.0419921875, "learning_rate": 0.001628712866590885, "loss": 0.0864, "step": 4800 }, { "epoch": 0.08506796757000747, "grad_norm": 0.051513671875, "learning_rate": 0.0016103102618497923, "loss": 0.0862, "step": 4900 }, { "epoch": 0.08680404854082395, "grad_norm": 0.052734375, "learning_rate": 0.0015915722124135226, "loss": 0.0855, "step": 5000 }, { "epoch": 0.08680404854082395, "eval_covost2-en-de_loss": 1.7862941026687622, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 8.2861, "eval_covost2-en-de_samples_per_second": 7.724, "eval_covost2-en-de_steps_per_second": 0.965, "step": 5000 }, { "epoch": 0.08680404854082395, "eval_covost2-zh-en_loss": 3.33290433883667, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 8.4063, "eval_covost2-zh-en_samples_per_second": 7.613, "eval_covost2-zh-en_steps_per_second": 0.952, "step": 5000 }, { "epoch": 0.08680404854082395, "eval_peoplespeech-clean-transcription_loss": 2.2601113319396973, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 9.4946, "eval_peoplespeech-clean-transcription_samples_per_second": 6.741, "eval_peoplespeech-clean-transcription_steps_per_second": 0.843, "step": 5000 }, { "epoch": 0.08854012951164042, "grad_norm": 0.053466796875, "learning_rate": 0.001572509017272072, "loss": 0.0872, "step": 5100 }, { "epoch": 0.0902762104824569, "grad_norm": 0.044189453125, "learning_rate": 0.0015531311541251993, "loss": 0.0859, "step": 5200 }, { "epoch": 0.09201229145327339, "grad_norm": 0.052978515625, "learning_rate": 0.0015334492736235703, "loss": 0.085, "step": 5300 }, { "epoch": 0.09374837242408986, "grad_norm": 0.04833984375, "learning_rate": 0.0015134741935148419, "loss": 0.0844, "step": 5400 }, { "epoch": 0.09548445339490634, "grad_norm": 0.047119140625, "learning_rate": 0.0014932168926979072, "loss": 0.0844, "step": 5500 }, { "epoch": 0.09722053436572281, "grad_norm": 0.05029296875, "learning_rate": 0.0014726885051885652, "loss": 0.0856, "step": 5600 }, { "epoch": 0.0989566153365393, "grad_norm": 0.049560546875, "learning_rate": 0.0014519003139999338, "loss": 0.0841, "step": 5700 }, { "epoch": 0.10069269630735578, "grad_norm": 0.056884765625, "learning_rate": 0.0014308637449409706, "loss": 0.0841, "step": 5800 }, { "epoch": 0.10242877727817225, "grad_norm": 0.041015625, "learning_rate": 0.0014095903603365066, "loss": 0.0825, "step": 5900 }, { "epoch": 0.10416485824898873, "grad_norm": 0.048583984375, "learning_rate": 0.0013880918526722496, "loss": 0.0828, "step": 6000 }, { "epoch": 0.10416485824898873, "eval_covost2-en-de_loss": 1.8097732067108154, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 8.2052, "eval_covost2-en-de_samples_per_second": 7.8, "eval_covost2-en-de_steps_per_second": 0.975, "step": 6000 }, { "epoch": 0.10416485824898873, "eval_covost2-zh-en_loss": 3.331326961517334, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 8.2653, "eval_covost2-zh-en_samples_per_second": 7.743, "eval_covost2-zh-en_steps_per_second": 0.968, "step": 6000 }, { "epoch": 0.10416485824898873, "eval_peoplespeech-clean-transcription_loss": 2.250232219696045, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 9.4708, "eval_peoplespeech-clean-transcription_samples_per_second": 6.758, "eval_peoplespeech-clean-transcription_steps_per_second": 0.845, "step": 6000 }, { "epoch": 0.10590093921980522, "grad_norm": 0.04443359375, "learning_rate": 0.0013663800381682463, "loss": 0.0819, "step": 6100 }, { "epoch": 0.10763702019062169, "grad_norm": 0.05419921875, "learning_rate": 0.0013444668502843329, "loss": 0.08, "step": 6200 }, { "epoch": 0.10937310116143817, "grad_norm": 0.0478515625, "learning_rate": 0.0013223643331611537, "loss": 0.0805, "step": 6300 }, { "epoch": 0.11110918213225465, "grad_norm": 0.051513671875, "learning_rate": 0.001300084635000341, "loss": 0.0799, "step": 6400 }, { "epoch": 0.11284526310307112, "grad_norm": 0.0498046875, "learning_rate": 0.0012776400013875004, "loss": 0.0807, "step": 6500 }, { "epoch": 0.1145813440738876, "grad_norm": 0.050537109375, "learning_rate": 0.0012550427685616766, "loss": 0.0799, "step": 6600 }, { "epoch": 0.11631742504470409, "grad_norm": 0.05029296875, "learning_rate": 0.0012323053566349834, "loss": 0.0802, "step": 6700 }, { "epoch": 0.11805350601552056, "grad_norm": 0.047119140625, "learning_rate": 0.0012094402627661448, "loss": 0.0796, "step": 6800 }, { "epoch": 0.11978958698633704, "grad_norm": 0.044677734375, "learning_rate": 0.0011864600542916813, "loss": 0.0784, "step": 6900 }, { "epoch": 0.12152566795715353, "grad_norm": 0.0478515625, "learning_rate": 0.0011633773618185302, "loss": 0.0808, "step": 7000 }, { "epoch": 0.12152566795715353, "eval_covost2-en-de_loss": 1.7786378860473633, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 8.0291, "eval_covost2-en-de_samples_per_second": 7.971, "eval_covost2-en-de_steps_per_second": 0.996, "step": 7000 }, { "epoch": 0.12152566795715353, "eval_covost2-zh-en_loss": 3.273571252822876, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 8.3234, "eval_covost2-zh-en_samples_per_second": 7.689, "eval_covost2-zh-en_steps_per_second": 0.961, "step": 7000 }, { "epoch": 0.12152566795715353, "eval_peoplespeech-clean-transcription_loss": 2.2290830612182617, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 9.7693, "eval_peoplespeech-clean-transcription_samples_per_second": 6.551, "eval_peoplespeech-clean-transcription_steps_per_second": 0.819, "step": 7000 }, { "epoch": 0.12326174892797, "grad_norm": 0.0478515625, "learning_rate": 0.0011402048722818862, "loss": 0.0786, "step": 7100 }, { "epoch": 0.12499782989878648, "grad_norm": 0.049560546875, "learning_rate": 0.0011169553219720827, "loss": 0.0795, "step": 7200 }, { "epoch": 0.12673391086960295, "grad_norm": 0.04736328125, "learning_rate": 0.001093641489534351, "loss": 0.0787, "step": 7300 }, { "epoch": 0.12846999184041943, "grad_norm": 0.054931640625, "learning_rate": 0.001070276188945293, "loss": 0.0784, "step": 7400 }, { "epoch": 0.13020607281123592, "grad_norm": 0.0478515625, "learning_rate": 0.00104687226246994, "loss": 0.0787, "step": 7500 }, { "epoch": 0.1319421537820524, "grad_norm": 0.048828125, "learning_rate": 0.0010234425736032607, "loss": 0.0788, "step": 7600 }, { "epoch": 0.13367823475286889, "grad_norm": 0.058837890625, "learning_rate": 0.001, "loss": 0.0769, "step": 7700 }, { "epoch": 0.13541431572368534, "grad_norm": 0.055908203125, "learning_rate": 0.0009765574263967396, "loss": 0.077, "step": 7800 }, { "epoch": 0.13715039669450182, "grad_norm": 0.05322265625, "learning_rate": 0.0009531277375300599, "loss": 0.0764, "step": 7900 }, { "epoch": 0.1388864776653183, "grad_norm": 0.04833984375, "learning_rate": 0.0009297238110547074, "loss": 0.0764, "step": 8000 }, { "epoch": 0.1388864776653183, "eval_covost2-en-de_loss": 1.7951624393463135, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 8.1477, "eval_covost2-en-de_samples_per_second": 7.855, "eval_covost2-en-de_steps_per_second": 0.982, "step": 8000 }, { "epoch": 0.1388864776653183, "eval_covost2-zh-en_loss": 3.301699161529541, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 8.8691, "eval_covost2-zh-en_samples_per_second": 7.216, "eval_covost2-zh-en_steps_per_second": 0.902, "step": 8000 }, { "epoch": 0.1388864776653183, "eval_peoplespeech-clean-transcription_loss": 2.1518499851226807, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 9.5239, "eval_peoplespeech-clean-transcription_samples_per_second": 6.72, "eval_peoplespeech-clean-transcription_steps_per_second": 0.84, "step": 8000 }, { "epoch": 0.1406225586361348, "grad_norm": 0.053955078125, "learning_rate": 0.0009063585104656494, "loss": 0.0762, "step": 8100 }, { "epoch": 0.14235863960695128, "grad_norm": 0.0654296875, "learning_rate": 0.0008830446780279176, "loss": 0.0769, "step": 8200 }, { "epoch": 0.14409472057776776, "grad_norm": 0.046875, "learning_rate": 0.0008597951277181142, "loss": 0.0751, "step": 8300 }, { "epoch": 0.14583080154858422, "grad_norm": 0.054443359375, "learning_rate": 0.0008366226381814697, "loss": 0.0765, "step": 8400 }, { "epoch": 0.1475668825194007, "grad_norm": 0.052490234375, "learning_rate": 0.000813539945708319, "loss": 0.0763, "step": 8500 }, { "epoch": 0.14930296349021718, "grad_norm": 0.068359375, "learning_rate": 0.0007905597372338558, "loss": 0.0744, "step": 8600 }, { "epoch": 0.15103904446103367, "grad_norm": 0.055908203125, "learning_rate": 0.0007676946433650169, "loss": 0.0737, "step": 8700 }, { "epoch": 0.15277512543185015, "grad_norm": 0.055908203125, "learning_rate": 0.0007449572314383236, "loss": 0.0758, "step": 8800 }, { "epoch": 0.1545112064026666, "grad_norm": 0.051513671875, "learning_rate": 0.0007223599986124993, "loss": 0.0753, "step": 8900 }, { "epoch": 0.1562472873734831, "grad_norm": 0.05517578125, "learning_rate": 0.0006999153649996595, "loss": 0.0736, "step": 9000 }, { "epoch": 0.1562472873734831, "eval_covost2-en-de_loss": 1.7736568450927734, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 8.2798, "eval_covost2-en-de_samples_per_second": 7.73, "eval_covost2-en-de_steps_per_second": 0.966, "step": 9000 }, { "epoch": 0.1562472873734831, "eval_covost2-zh-en_loss": 3.2736916542053223, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 8.6328, "eval_covost2-zh-en_samples_per_second": 7.414, "eval_covost2-zh-en_steps_per_second": 0.927, "step": 9000 }, { "epoch": 0.1562472873734831, "eval_peoplespeech-clean-transcription_loss": 2.169971227645874, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 10.7684, "eval_peoplespeech-clean-transcription_samples_per_second": 5.943, "eval_peoplespeech-clean-transcription_steps_per_second": 0.743, "step": 9000 }, { "epoch": 0.15798336834429957, "grad_norm": 0.06201171875, "learning_rate": 0.0006776356668388464, "loss": 0.073, "step": 9100 }, { "epoch": 0.15971944931511606, "grad_norm": 0.0654296875, "learning_rate": 0.0006555331497156671, "loss": 0.0753, "step": 9200 }, { "epoch": 0.16145553028593254, "grad_norm": 0.056640625, "learning_rate": 0.0006336199618317538, "loss": 0.0754, "step": 9300 }, { "epoch": 0.16319161125674903, "grad_norm": 0.06005859375, "learning_rate": 0.0006119081473277501, "loss": 0.0736, "step": 9400 }, { "epoch": 0.16492769222756548, "grad_norm": 0.059326171875, "learning_rate": 0.0005904096396634935, "loss": 0.0721, "step": 9500 }, { "epoch": 0.16666377319838196, "grad_norm": 0.06201171875, "learning_rate": 0.0005691362550590297, "loss": 0.0717, "step": 9600 }, { "epoch": 0.16839985416919845, "grad_norm": 0.051025390625, "learning_rate": 0.0005480996860000663, "loss": 0.0738, "step": 9700 }, { "epoch": 0.17013593514001493, "grad_norm": 0.06494140625, "learning_rate": 0.0005273114948114346, "loss": 0.0737, "step": 9800 }, { "epoch": 0.17187201611083142, "grad_norm": 0.058837890625, "learning_rate": 0.0005067831073020928, "loss": 0.0711, "step": 9900 }, { "epoch": 0.1736080970816479, "grad_norm": 0.050537109375, "learning_rate": 0.00048652580648515787, "loss": 0.0722, "step": 10000 }, { "epoch": 0.1736080970816479, "eval_covost2-en-de_loss": 1.768043875694275, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 8.2025, "eval_covost2-en-de_samples_per_second": 7.802, "eval_covost2-en-de_steps_per_second": 0.975, "step": 10000 }, { "epoch": 0.1736080970816479, "eval_covost2-zh-en_loss": 3.288457155227661, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 8.5315, "eval_covost2-zh-en_samples_per_second": 7.502, "eval_covost2-zh-en_steps_per_second": 0.938, "step": 10000 }, { "epoch": 0.1736080970816479, "eval_peoplespeech-clean-transcription_loss": 2.099651336669922, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 9.7081, "eval_peoplespeech-clean-transcription_samples_per_second": 6.592, "eval_peoplespeech-clean-transcription_steps_per_second": 0.824, "step": 10000 }, { "epoch": 0.17534417805246436, "grad_norm": 0.059326171875, "learning_rate": 0.0004665507263764299, "loss": 0.0717, "step": 10100 }, { "epoch": 0.17708025902328084, "grad_norm": 0.06103515625, "learning_rate": 0.0004468688458748006, "loss": 0.0729, "step": 10200 }, { "epoch": 0.17881633999409732, "grad_norm": 0.051513671875, "learning_rate": 0.0004274909827279283, "loss": 0.0711, "step": 10300 }, { "epoch": 0.1805524209649138, "grad_norm": 0.059326171875, "learning_rate": 0.0004084277875864776, "loss": 0.0712, "step": 10400 }, { "epoch": 0.1822885019357303, "grad_norm": 0.062255859375, "learning_rate": 0.00038968973815020803, "loss": 0.0708, "step": 10500 }, { "epoch": 0.18402458290654677, "grad_norm": 0.06640625, "learning_rate": 0.00037128713340911534, "loss": 0.0716, "step": 10600 }, { "epoch": 0.18576066387736323, "grad_norm": 0.0517578125, "learning_rate": 0.00035323008798280133, "loss": 0.0728, "step": 10700 }, { "epoch": 0.1874967448481797, "grad_norm": 0.076171875, "learning_rate": 0.00033552852656117837, "loss": 0.0711, "step": 10800 }, { "epoch": 0.1892328258189962, "grad_norm": 0.06494140625, "learning_rate": 0.00031819217844956217, "loss": 0.0701, "step": 10900 }, { "epoch": 0.19096890678981268, "grad_norm": 0.054443359375, "learning_rate": 0.00030123057222115836, "loss": 0.0705, "step": 11000 }, { "epoch": 0.19096890678981268, "eval_covost2-en-de_loss": 1.7685788869857788, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 8.4174, "eval_covost2-en-de_samples_per_second": 7.603, "eval_covost2-en-de_steps_per_second": 0.95, "step": 11000 }, { "epoch": 0.19096890678981268, "eval_covost2-zh-en_loss": 3.286205530166626, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 11.2405, "eval_covost2-zh-en_samples_per_second": 5.694, "eval_covost2-zh-en_steps_per_second": 0.712, "step": 11000 }, { "epoch": 0.19096890678981268, "eval_peoplespeech-clean-transcription_loss": 2.1252198219299316, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 9.3932, "eval_peoplespeech-clean-transcription_samples_per_second": 6.813, "eval_peoplespeech-clean-transcription_steps_per_second": 0.852, "step": 11000 }, { "epoch": 0.19270498776062917, "grad_norm": 0.076171875, "learning_rate": 0.0002846530304798727, "loss": 0.0709, "step": 11100 }, { "epoch": 0.19444106873144562, "grad_norm": 0.07275390625, "learning_rate": 0.00026846866473633125, "loss": 0.0717, "step": 11200 }, { "epoch": 0.1961771497022621, "grad_norm": 0.0498046875, "learning_rate": 0.00025268637039992293, "loss": 0.0699, "step": 11300 }, { "epoch": 0.1979132306730786, "grad_norm": 0.046875, "learning_rate": 0.00023731482188961818, "loss": 0.0712, "step": 11400 }, { "epoch": 0.19964931164389507, "grad_norm": 0.050048828125, "learning_rate": 0.00022236246786624792, "loss": 0.0714, "step": 11500 }, { "epoch": 0.20138539261471156, "grad_norm": 0.04931640625, "learning_rate": 0.00020783752658887068, "loss": 0.071, "step": 11600 }, { "epoch": 0.20312147358552804, "grad_norm": 0.06298828125, "learning_rate": 0.0001937479813977703, "loss": 0.0706, "step": 11700 }, { "epoch": 0.2048575545563445, "grad_norm": 0.05712890625, "learning_rate": 0.00018010157632657541, "loss": 0.07, "step": 11800 }, { "epoch": 0.20659363552716098, "grad_norm": 0.068359375, "learning_rate": 0.00016690581184590858, "loss": 0.0708, "step": 11900 }, { "epoch": 0.20832971649797746, "grad_norm": 0.05908203125, "learning_rate": 0.00015416794074090258, "loss": 0.069, "step": 12000 }, { "epoch": 0.20832971649797746, "eval_covost2-en-de_loss": 1.7618954181671143, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 8.2954, "eval_covost2-en-de_samples_per_second": 7.715, "eval_covost2-en-de_steps_per_second": 0.964, "step": 12000 }, { "epoch": 0.20832971649797746, "eval_covost2-zh-en_loss": 3.287311553955078, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 8.0269, "eval_covost2-zh-en_samples_per_second": 7.973, "eval_covost2-zh-en_steps_per_second": 0.997, "step": 12000 }, { "epoch": 0.20832971649797746, "eval_peoplespeech-clean-transcription_loss": 2.119732141494751, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 9.4528, "eval_peoplespeech-clean-transcription_samples_per_second": 6.77, "eval_peoplespeech-clean-transcription_steps_per_second": 0.846, "step": 12000 }, { "epoch": 0.21006579746879395, "grad_norm": 0.059814453125, "learning_rate": 0.00014189496412485593, "loss": 0.0703, "step": 12100 }, { "epoch": 0.21180187843961043, "grad_norm": 0.053466796875, "learning_rate": 0.00013009362759120978, "loss": 0.0713, "step": 12200 }, { "epoch": 0.21353795941042691, "grad_norm": 0.0673828125, "learning_rate": 0.00011877041750597173, "loss": 0.0694, "step": 12300 }, { "epoch": 0.21527404038124337, "grad_norm": 0.060302734375, "learning_rate": 0.00010793155744261352, "loss": 0.07, "step": 12400 }, { "epoch": 0.21701012135205985, "grad_norm": 0.0390625, "learning_rate": 9.758300476141169e-05, "loss": 0.0675, "step": 12500 }, { "epoch": 0.21874620232287634, "grad_norm": 0.058349609375, "learning_rate": 8.773044733510338e-05, "loss": 0.0699, "step": 12600 }, { "epoch": 0.22048228329369282, "grad_norm": 0.04931640625, "learning_rate": 7.837930042266262e-05, "loss": 0.0708, "step": 12700 }, { "epoch": 0.2222183642645093, "grad_norm": 0.054931640625, "learning_rate": 6.953470369291348e-05, "loss": 0.0701, "step": 12800 }, { "epoch": 0.2239544452353258, "grad_norm": 0.0556640625, "learning_rate": 6.120151839961363e-05, "loss": 0.0703, "step": 12900 }, { "epoch": 0.22569052620614224, "grad_norm": 0.0654296875, "learning_rate": 5.338432470956589e-05, "loss": 0.0707, "step": 13000 }, { "epoch": 0.22569052620614224, "eval_covost2-en-de_loss": 1.7618814706802368, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 8.2896, "eval_covost2-en-de_samples_per_second": 7.721, "eval_covost2-en-de_steps_per_second": 0.965, "step": 13000 }, { "epoch": 0.22569052620614224, "eval_covost2-zh-en_loss": 3.287533760070801, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 8.1323, "eval_covost2-zh-en_samples_per_second": 7.87, "eval_covost2-zh-en_steps_per_second": 0.984, "step": 13000 }, { "epoch": 0.22569052620614224, "eval_peoplespeech-clean-transcription_loss": 2.1182146072387695, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 9.6791, "eval_peoplespeech-clean-transcription_samples_per_second": 6.612, "eval_peoplespeech-clean-transcription_steps_per_second": 0.827, "step": 13000 }, { "epoch": 0.22742660717695873, "grad_norm": 0.056396484375, "learning_rate": 4.6087419185220966e-05, "loss": 0.0698, "step": 13100 }, { "epoch": 0.2291626881477752, "grad_norm": 0.0634765625, "learning_rate": 3.931481242315993e-05, "loss": 0.0709, "step": 13200 }, { "epoch": 0.2308987691185917, "grad_norm": 0.0625, "learning_rate": 3.307022684974936e-05, "loss": 0.071, "step": 13300 }, { "epoch": 0.23263485008940818, "grad_norm": 0.05712890625, "learning_rate": 2.7357094675186987e-05, "loss": 0.0704, "step": 13400 }, { "epoch": 0.23437093106022464, "grad_norm": 0.068359375, "learning_rate": 2.2178556007054874e-05, "loss": 0.0704, "step": 13500 }, { "epoch": 0.23610701203104112, "grad_norm": 0.05615234375, "learning_rate": 1.7537457124423894e-05, "loss": 0.0712, "step": 13600 }, { "epoch": 0.2378430930018576, "grad_norm": 0.07275390625, "learning_rate": 1.3436348913453578e-05, "loss": 0.0709, "step": 13700 }, { "epoch": 0.2395791739726741, "grad_norm": 0.05078125, "learning_rate": 9.877485465349056e-06, "loss": 0.0701, "step": 13800 }, { "epoch": 0.24131525494349057, "grad_norm": 0.053955078125, "learning_rate": 6.862822837445881e-06, "loss": 0.0708, "step": 13900 }, { "epoch": 0.24305133591430705, "grad_norm": 0.07275390625, "learning_rate": 4.394017978101905e-06, "loss": 0.0711, "step": 14000 }, { "epoch": 0.24305133591430705, "eval_covost2-en-de_loss": 1.762545108795166, "eval_covost2-en-de_model_preparation_time": 0.0057, "eval_covost2-en-de_runtime": 7.89, "eval_covost2-en-de_samples_per_second": 8.111, "eval_covost2-en-de_steps_per_second": 1.014, "step": 14000 }, { "epoch": 0.24305133591430705, "eval_covost2-zh-en_loss": 3.2872180938720703, "eval_covost2-zh-en_model_preparation_time": 0.0057, "eval_covost2-zh-en_runtime": 8.1851, "eval_covost2-zh-en_samples_per_second": 7.819, "eval_covost2-zh-en_steps_per_second": 0.977, "step": 14000 }, { "epoch": 0.24305133591430705, "eval_peoplespeech-clean-transcription_loss": 2.116088628768921, "eval_peoplespeech-clean-transcription_model_preparation_time": 0.0057, "eval_peoplespeech-clean-transcription_runtime": 9.4366, "eval_peoplespeech-clean-transcription_samples_per_second": 6.782, "eval_peoplespeech-clean-transcription_steps_per_second": 0.848, "step": 14000 }, { "epoch": 0.2447874168851235, "grad_norm": 0.059326171875, "learning_rate": 2.472427815989886e-06, "loss": 0.0705, "step": 14100 }, { "epoch": 0.24652349785594, "grad_norm": 0.03955078125, "learning_rate": 1.099108514288627e-06, "loss": 0.0703, "step": 14200 }, { "epoch": 0.24825957882675648, "grad_norm": 0.07763671875, "learning_rate": 2.748148901841052e-07, "loss": 0.0714, "step": 14300 }, { "epoch": 0.24999565979757296, "grad_norm": 0.0625, "learning_rate": 0.0, "loss": 0.0704, "step": 14400 } ], "logging_steps": 100, "max_steps": 14400, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 3600, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 6.620180878885847e+17, "train_batch_size": 24, "trial_name": null, "trial_params": null }