!!!! /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/trainer_0_log.txt  > EPOCH: 0/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 14:21:48)   --> STEP: 0/639 -- GLOBAL_STEP: 0 | > current_lr: 2.5e-07 | > step_time: 2.25220 (2.25218) | > loader_time: 2.73680 (2.73685)  --> STEP: 25/639 -- GLOBAL_STEP: 25 | > loss: 2.68924 (2.80788) | > log_mle: 0.67991 (0.69380) | > loss_dur: 2.00933 (2.11407) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 8.67662 (8.20194) | > current_lr: 3.5e-06 | > step_time: 0.74530 (0.46316) | > loader_time: 0.03830 (0.01428)  --> STEP: 50/639 -- GLOBAL_STEP: 50 | > loss: 2.54031 (2.72692) | > log_mle: 0.67281 (0.69421) | > loss_dur: 1.86750 (2.03270) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 8.38860 (8.43322) | > current_lr: 0.00001 | > step_time: 0.31380 (0.43044) | > loader_time: 0.00480 (0.01198)  --> STEP: 75/639 -- GLOBAL_STEP: 75 | > loss: 2.34873 (2.64607) | > log_mle: 0.68339 (0.69051) | > loss_dur: 1.66534 (1.95556) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 7.75456 (8.35564) | > current_lr: 0.00002 | > step_time: 0.41660 (0.42852) | > loader_time: 0.02860 (0.01396)  --> STEP: 100/639 -- GLOBAL_STEP: 100 | > loss: 2.13392 (2.54371) | > log_mle: 0.67554 (0.68656) | > loss_dur: 1.45838 (1.85715) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 6.60118 (8.06842) | > current_lr: 0.00002 | > step_time: 0.50900 (0.43294) | > loader_time: 0.00440 (0.01497)  --> STEP: 125/639 -- GLOBAL_STEP: 125 | > loss: 1.95189 (2.43579) | > log_mle: 0.63128 (0.67989) | > loss_dur: 1.32061 (1.75590) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 5.24600 (7.59556) | > current_lr: 0.00003 | > step_time: 0.39550 (0.43948) | > loader_time: 0.00270 (0.01383)  --> STEP: 150/639 -- GLOBAL_STEP: 150 | > loss: 1.86333 (2.34360) | > log_mle: 0.57914 (0.66618) | > loss_dur: 1.28420 (1.67741) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 4.35954 (7.07997) | > current_lr: 0.00003 | > step_time: 0.41850 (0.45004) | > loader_time: 0.04320 (0.01496)  --> STEP: 175/639 -- GLOBAL_STEP: 175 | > loss: 1.76461 (2.26241) | > log_mle: 0.49550 (0.64601) | > loss_dur: 1.26912 (1.61641) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 4.30163 (6.65461) | > current_lr: 0.00004 | > step_time: 0.55360 (0.46178) | > loader_time: 0.00550 (0.01589)  --> STEP: 200/639 -- GLOBAL_STEP: 200 | > loss: 1.46812 (2.18249) | > log_mle: 0.41997 (0.62103) | > loss_dur: 1.04815 (1.56146) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 3.62275 (6.32355) | > current_lr: 0.00005 | > step_time: 0.49900 (0.47220) | > loader_time: 0.00530 (0.01556)  --> STEP: 225/639 -- GLOBAL_STEP: 225 | > loss: 1.20876 (2.08949) | > log_mle: 0.35754 (0.59336) | > loss_dur: 0.85121 (1.49613) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 3.06323 (5.99185) | > current_lr: 0.00005 | > step_time: 0.49780 (0.48655) | > loader_time: 0.01120 (0.01600)  --> STEP: 250/639 -- GLOBAL_STEP: 250 | > loss: 1.01158 (1.99150) | > log_mle: 0.30829 (0.56461) | > loss_dur: 0.70329 (1.42688) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 2.41415 (5.71959) | > current_lr: 0.00006 | > step_time: 0.56650 (0.49811) | > loader_time: 0.00420 (0.01560)  --> STEP: 275/639 -- GLOBAL_STEP: 275 | > loss: 0.94678 (1.89563) | > log_mle: 0.25263 (0.53627) | > loss_dur: 0.69415 (1.35936) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 5.00962 (5.56107) | > current_lr: 0.00007 | > step_time: 0.80440 (0.51157) | > loader_time: 0.00480 (0.01520)  --> STEP: 300/639 -- GLOBAL_STEP: 300 | > loss: 0.84815 (1.80696) | > log_mle: 0.21930 (0.51052) | > loss_dur: 0.62885 (1.29644) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 2.48566 (5.33007) | > current_lr: 0.00007 | > step_time: 0.63300 (0.52171) | > loader_time: 0.00550 (0.01522)  --> STEP: 325/639 -- GLOBAL_STEP: 325 | > loss: 0.72382 (1.72419) | > log_mle: 0.18864 (0.48592) | > loss_dur: 0.53517 (1.23827) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 2.83370 (5.15076) | > current_lr: 0.00008 | > step_time: 0.61070 (0.53378) | > loader_time: 0.00330 (0.01493)  --> STEP: 350/639 -- GLOBAL_STEP: 350 | > loss: 0.65247 (1.64850) | > log_mle: 0.12541 (0.46270) | > loss_dur: 0.52705 (1.18581) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 3.18143 (4.97465) | > current_lr: 0.00008 | > step_time: 0.94140 (0.54657) | > loader_time: 0.05320 (0.01444)  --> STEP: 375/639 -- GLOBAL_STEP: 375 | > loss: 0.61419 (1.57914) | > log_mle: 0.14556 (0.44112) | > loss_dur: 0.46863 (1.13802) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 2.90963 (4.81825) | > current_lr: 0.00009 | > step_time: 1.48460 (0.57878) | > loader_time: 0.00350 (0.01387)  --> STEP: 400/639 -- GLOBAL_STEP: 400 | > loss: 0.56109 (1.51471) | > log_mle: 0.09489 (0.41988) | > loss_dur: 0.46620 (1.09483) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 4.31574 (4.65744) | > current_lr: 0.00010 | > step_time: 1.46180 (0.63471) | > loader_time: 0.00360 (0.01323)  --> STEP: 425/639 -- GLOBAL_STEP: 425 | > loss: 0.47336 (1.45573) | > log_mle: 0.03574 (0.39993) | > loss_dur: 0.43762 (1.05580) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 6.33755 (4.54089) | > current_lr: 0.00010 | > step_time: 2.17540 (0.70223) | > loader_time: 0.00600 (0.01270)  --> STEP: 450/639 -- GLOBAL_STEP: 450 | > loss: 0.49154 (1.40104) | > log_mle: 0.06224 (0.38090) | > loss_dur: 0.42930 (1.02014) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 0.94281 (4.41006) | > current_lr: 0.00011 | > step_time: 2.10860 (0.77152) | > loader_time: 0.00520 (0.01225)  --> STEP: 475/639 -- GLOBAL_STEP: 475 | > loss: 0.43879 (1.35077) | > log_mle: 0.03799 (0.36261) | > loss_dur: 0.40079 (0.98815) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.58623 (4.30845) | > current_lr: 0.00012 | > step_time: 1.64110 (0.84143) | > loader_time: 0.00360 (0.01184)  --> STEP: 500/639 -- GLOBAL_STEP: 500 | > loss: 0.41380 (1.30429) | > log_mle: -0.01228 (0.34498) | > loss_dur: 0.42608 (0.95931) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 2.46413 (4.20251) | > current_lr: 0.00012 | > step_time: 2.42610 (0.91334) | > loader_time: 0.00640 (0.01147)  --> STEP: 525/639 -- GLOBAL_STEP: 525 | > loss: 0.40138 (1.26067) | > log_mle: -0.04119 (0.32774) | > loss_dur: 0.44257 (0.93293) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 2.46372 (4.09977) | > current_lr: 0.00013 | > step_time: 3.05580 (0.98624) | > loader_time: 0.00480 (0.01115)  --> STEP: 550/639 -- GLOBAL_STEP: 550 | > loss: 0.40938 (1.22062) | > log_mle: -0.01529 (0.31157) | > loss_dur: 0.42467 (0.90905) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 2.00481 (4.03866) | > current_lr: 0.00013 | > step_time: 2.55480 (1.05914) | > loader_time: 0.00410 (0.01088)  --> STEP: 575/639 -- GLOBAL_STEP: 575 | > loss: 0.35623 (1.18315) | > log_mle: -0.03056 (0.29620) | > loss_dur: 0.38679 (0.88695) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 2.14713 (3.95643) | > current_lr: 0.00014 | > step_time: 2.00540 (1.11337) | > loader_time: 0.00400 (0.01062)  --> STEP: 600/639 -- GLOBAL_STEP: 600 | > loss: 0.40111 (1.14811) | > log_mle: -0.04625 (0.28125) | > loss_dur: 0.44735 (0.86687) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.61778 (3.89870) | > current_lr: 0.00015 | > step_time: 2.85190 (1.17815) | > loader_time: 0.00430 (0.01039)  --> STEP: 625/639 -- GLOBAL_STEP: 625 | > loss: 0.31717 (1.11526) | > log_mle: -0.08686 (0.26702) | > loss_dur: 0.40404 (0.84824) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.55096 (3.82335) | > current_lr: 0.00015 | > step_time: 3.10900 (1.26296) | > loader_time: 0.00620 (0.01025) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  | > Synthesizing test sentences. --> EVAL PERFORMANCE | > avg_loader_time: 0.00673 (+0.00000) | > avg_loss: 0.32303 (+0.00000) | > avg_log_mle: -0.07414 (+0.00000) | > avg_loss_dur: 0.39717 (+0.00000) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_640.pth.tar  > EPOCH: 1/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 14:36:23)   --> STEP: 10/639 -- GLOBAL_STEP: 650 | > loss: 0.40342 (0.40815) | > log_mle: -0.08743 (-0.07398) | > loss_dur: 0.49085 (0.48213) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 2.79583 (3.56312) | > current_lr: 0.00016 | > step_time: 0.94000 (0.92616) | > loader_time: 0.00270 (0.00243)  --> STEP: 35/639 -- GLOBAL_STEP: 675 | > loss: 0.32555 (0.38154) | > log_mle: -0.08434 (-0.07345) | > loss_dur: 0.40988 (0.45499) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.62101 (2.92284) | > current_lr: 0.00017 | > step_time: 0.96440 (1.00139) | > loader_time: 0.00250 (0.00271)  --> STEP: 60/639 -- GLOBAL_STEP: 700 | > loss: 0.33175 (0.35966) | > log_mle: -0.08439 (-0.08267) | > loss_dur: 0.41614 (0.44233) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.42700 (2.83818) | > current_lr: 0.00017 | > step_time: 1.16570 (1.10885) | > loader_time: 0.00270 (0.00286)  --> STEP: 85/639 -- GLOBAL_STEP: 725 | > loss: 0.34092 (0.34193) | > log_mle: -0.08256 (-0.08947) | > loss_dur: 0.42349 (0.43140) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.20836 (2.57197) | > current_lr: 0.00018 | > step_time: 1.48890 (1.17150) | > loader_time: 0.00350 (0.00297)  --> STEP: 110/639 -- GLOBAL_STEP: 750 | > loss: 0.30479 (0.33062) | > log_mle: -0.11459 (-0.09837) | > loss_dur: 0.41937 (0.42898) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 2.07588 (2.70238) | > current_lr: 0.00018 | > step_time: 1.34360 (1.25077) | > loader_time: 0.00290 (0.00306)  --> STEP: 135/639 -- GLOBAL_STEP: 775 | > loss: 0.29779 (0.31978) | > log_mle: -0.11563 (-0.10503) | > loss_dur: 0.41342 (0.42480) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.44761 (2.74255) | > current_lr: 0.00019 | > step_time: 1.58930 (1.29768) | > loader_time: 0.00300 (0.00312)  --> STEP: 160/639 -- GLOBAL_STEP: 800 | > loss: 0.22939 (0.31052) | > log_mle: -0.14974 (-0.11039) | > loss_dur: 0.37913 (0.42092) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 3.01391 (2.72484) | > current_lr: 0.00020 | > step_time: 1.64180 (1.35898) | > loader_time: 0.00340 (0.00320)  --> STEP: 185/639 -- GLOBAL_STEP: 825 | > loss: 0.26804 (0.30060) | > log_mle: -0.15227 (-0.11882) | > loss_dur: 0.42032 (0.41942) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 3.25537 (2.87987) | > current_lr: 0.00020 | > step_time: 1.96780 (1.42707) | > loader_time: 0.00320 (0.00329)  --> STEP: 210/639 -- GLOBAL_STEP: 850 | > loss: 0.27456 (0.29395) | > log_mle: -0.10875 (-0.12344) | > loss_dur: 0.38331 (0.41739) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.15021 (2.82901) | > current_lr: 0.00021 | > step_time: 1.81050 (1.48695) | > loader_time: 0.00360 (0.00337)  --> STEP: 235/639 -- GLOBAL_STEP: 875 | > loss: 0.21634 (0.28659) | > log_mle: -0.18097 (-0.12874) | > loss_dur: 0.39730 (0.41533) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 2.15051 (2.90148) | > current_lr: 0.00022 | > step_time: 1.96140 (1.55028) | > loader_time: 0.00430 (0.00346)  --> STEP: 260/639 -- GLOBAL_STEP: 900 | > loss: 0.21377 (0.27946) | > log_mle: -0.18474 (-0.13436) | > loss_dur: 0.39851 (0.41381) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 3.20737 (2.94042) | > current_lr: 0.00022 | > step_time: 1.94490 (1.60615) | > loader_time: 0.00340 (0.00352)  --> STEP: 285/639 -- GLOBAL_STEP: 925 | > loss: 0.23539 (0.27387) | > log_mle: -0.16776 (-0.13873) | > loss_dur: 0.40315 (0.41260) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.09445 (2.90838) | > current_lr: 0.00023 | > step_time: 2.21680 (1.65652) | > loader_time: 0.00440 (0.00360)  --> STEP: 310/639 -- GLOBAL_STEP: 950 | > loss: 0.19231 (0.26824) | > log_mle: -0.19105 (-0.14301) | > loss_dur: 0.38336 (0.41125) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.62466 (2.91517) | > current_lr: 0.00023 | > step_time: 1.56340 (1.69068) | > loader_time: 0.00440 (0.00367)  --> STEP: 335/639 -- GLOBAL_STEP: 975 | > loss: 0.15137 (0.26301) | > log_mle: -0.23905 (-0.14699) | > loss_dur: 0.39042 (0.41000) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 3.55672 (2.88837) | > current_lr: 0.00024 | > step_time: 2.09720 (1.72617) | > loader_time: 0.00340 (0.00373)  --> STEP: 360/639 -- GLOBAL_STEP: 1000 | > loss: 0.17995 (0.25690) | > log_mle: -0.18796 (-0.15111) | > loss_dur: 0.36791 (0.40802) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.14977 (2.87900) | > current_lr: 0.00025 | > step_time: 1.92540 (1.75272) | > loader_time: 0.00350 (0.00375)  --> STEP: 385/639 -- GLOBAL_STEP: 1025 | > loss: 0.15286 (0.25091) | > log_mle: -0.21789 (-0.15488) | > loss_dur: 0.37075 (0.40580) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 2.78109 (2.82024) | > current_lr: 0.00025 | > step_time: 2.38510 (1.78166) | > loader_time: 0.00510 (0.00381)  --> STEP: 410/639 -- GLOBAL_STEP: 1050 | > loss: 0.18331 (0.24530) | > log_mle: -0.21690 (-0.15862) | > loss_dur: 0.40021 (0.40391) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 2.16885 (2.79656) | > current_lr: 0.00026 | > step_time: 2.28150 (1.80468) | > loader_time: 0.00380 (0.00385)  --> STEP: 435/639 -- GLOBAL_STEP: 1075 | > loss: 0.14181 (0.23966) | > log_mle: -0.20919 (-0.16190) | > loss_dur: 0.35099 (0.40156) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.68893 (2.80774) | > current_lr: 0.00027 | > step_time: 2.02890 (1.82948) | > loader_time: 0.00510 (0.00389)  --> STEP: 460/639 -- GLOBAL_STEP: 1100 | > loss: 0.14590 (0.23406) | > log_mle: -0.20205 (-0.16483) | > loss_dur: 0.34795 (0.39888) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.30333 (2.77760) | > current_lr: 0.00027 | > step_time: 2.00560 (1.85122) | > loader_time: 0.00650 (0.00391)  --> STEP: 485/639 -- GLOBAL_STEP: 1125 | > loss: 0.12450 (0.22832) | > log_mle: -0.22815 (-0.16809) | > loss_dur: 0.35265 (0.39640) | > amp_scaler: 16384.00000 (16384.00000) | > grad_norm: 1.49814 (2.74890) | > current_lr: 0.00028 | > step_time: 1.73080 (1.84926) | > loader_time: 0.00420 (0.00392)  --> STEP: 510/639 -- GLOBAL_STEP: 1150 | > loss: 0.09418 (0.22206) | > log_mle: -0.24762 (-0.17148) | > loss_dur: 0.34181 (0.39353) | > amp_scaler: 8192.00000 (16094.87059) | > grad_norm: 6.87299 (2.75615) | > current_lr: 0.00028 | > step_time: 1.63720 (1.85471) | > loader_time: 0.00410 (0.00395)  --> STEP: 535/639 -- GLOBAL_STEP: 1175 | > loss: 0.11803 (0.21599) | > log_mle: -0.23382 (-0.17469) | > loss_dur: 0.35185 (0.39067) | > amp_scaler: 8192.00000 (15725.57757) | > grad_norm: 2.05549 (2.74992) | > current_lr: 0.00029 | > step_time: 1.82340 (1.84946) | > loader_time: 0.00410 (0.00396)  --> STEP: 560/639 -- GLOBAL_STEP: 1200 | > loss: 0.08335 (0.21034) | > log_mle: -0.24736 (-0.17766) | > loss_dur: 0.33071 (0.38800) | > amp_scaler: 8192.00000 (15389.25714) | > grad_norm: 6.31019 (2.75718) | > current_lr: 0.00030 | > step_time: 1.89680 (1.85138) | > loader_time: 0.00410 (0.00398)  --> STEP: 585/639 -- GLOBAL_STEP: 1225 | > loss: 0.07639 (0.20453) | > log_mle: -0.24868 (-0.18040) | > loss_dur: 0.32508 (0.38493) | > amp_scaler: 8192.00000 (15081.68205) | > grad_norm: 1.84161 (2.77420) | > current_lr: 0.00030 | > step_time: 2.04880 (1.85747) | > loader_time: 0.00420 (0.00403)  --> STEP: 610/639 -- GLOBAL_STEP: 1250 | > loss: 0.06405 (0.19844) | > log_mle: -0.23509 (-0.18333) | > loss_dur: 0.29914 (0.38177) | > amp_scaler: 8192.00000 (14799.31803) | > grad_norm: 1.13323 (2.75714) | > current_lr: 0.00031 | > step_time: 2.67690 (1.87133) | > loader_time: 0.00800 (0.00408)  --> STEP: 635/639 -- GLOBAL_STEP: 1275 | > loss: 0.04372 (0.19245) | > log_mle: -0.25301 (-0.18631) | > loss_dur: 0.29673 (0.37875) | > amp_scaler: 8192.00000 (14539.18740) | > grad_norm: 2.28316 (2.75188) | > current_lr: 0.00032 | > step_time: 2.28230 (1.90915) | > loader_time: 0.00680 (0.00414) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02789 (+0.02116) | > avg_loss: 0.05109 (-0.27194) | > avg_log_mle: -0.25246 (-0.17831) | > avg_loss_dur: 0.30355 (-0.09363) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_1280.pth.tar  > EPOCH: 2/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 14:57:07)   --> STEP: 20/639 -- GLOBAL_STEP: 1300 | > loss: 0.08709 (0.09731) | > log_mle: -0.28170 (-0.25033) | > loss_dur: 0.36879 (0.34764) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 6.23976 (2.87441) | > current_lr: 0.00032 | > step_time: 0.47300 (0.35552) | > loader_time: 0.00240 (0.00237)  --> STEP: 45/639 -- GLOBAL_STEP: 1325 | > loss: 0.08794 (0.08923) | > log_mle: -0.20777 (-0.24372) | > loss_dur: 0.29571 (0.33296) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 1.66998 (2.94103) | > current_lr: 0.00033 | > step_time: 0.77260 (0.51421) | > loader_time: 0.00260 (0.00259)  --> STEP: 70/639 -- GLOBAL_STEP: 1350 | > loss: 0.00307 (0.08160) | > log_mle: -0.27707 (-0.24500) | > loss_dur: 0.28014 (0.32660) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.34535 (3.00717) | > current_lr: 0.00033 | > step_time: 1.17320 (0.65961) | > loader_time: 0.00350 (0.00275)  --> STEP: 95/639 -- GLOBAL_STEP: 1375 | > loss: 0.04940 (0.06901) | > log_mle: -0.28848 (-0.24801) | > loss_dur: 0.33788 (0.31702) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 7.61669 (3.02453) | > current_lr: 0.00034 | > step_time: 0.95190 (0.77012) | > loader_time: 0.00330 (0.00281)  --> STEP: 120/639 -- GLOBAL_STEP: 1400 | > loss: -0.00524 (0.05933) | > log_mle: -0.30534 (-0.25148) | > loss_dur: 0.30010 (0.31081) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 4.52168 (3.19908) | > current_lr: 0.00035 | > step_time: 1.46290 (0.86501) | > loader_time: 0.00390 (0.00290)  --> STEP: 145/639 -- GLOBAL_STEP: 1425 | > loss: -0.01092 (0.04707) | > log_mle: -0.29068 (-0.25668) | > loss_dur: 0.27976 (0.30374) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.21654 (3.22304) | > current_lr: 0.00035 | > step_time: 1.45010 (0.94174) | > loader_time: 0.00300 (0.00295)  --> STEP: 170/639 -- GLOBAL_STEP: 1450 | > loss: -0.01932 (0.03724) | > log_mle: -0.28673 (-0.26038) | > loss_dur: 0.26741 (0.29762) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.49522 (3.26749) | > current_lr: 0.00036 | > step_time: 0.84110 (0.96615) | > loader_time: 0.00320 (0.00300)  --> STEP: 195/639 -- GLOBAL_STEP: 1475 | > loss: -0.02427 (0.02877) | > log_mle: -0.24831 (-0.26306) | > loss_dur: 0.22404 (0.29183) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 0.83175 (3.19884) | > current_lr: 0.00037 | > step_time: 0.54500 (0.92456) | > loader_time: 0.00340 (0.00302)  --> STEP: 220/639 -- GLOBAL_STEP: 1500 | > loss: -0.05418 (0.02099) | > log_mle: -0.28522 (-0.26514) | > loss_dur: 0.23103 (0.28614) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.06453 (3.12294) | > current_lr: 0.00037 | > step_time: 0.65620 (0.88964) | > loader_time: 0.00340 (0.00304)  --> STEP: 245/639 -- GLOBAL_STEP: 1525 | > loss: -0.06164 (0.01242) | > log_mle: -0.27389 (-0.26824) | > loss_dur: 0.21224 (0.28066) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.28303 (3.16629) | > current_lr: 0.00038 | > step_time: 0.55680 (0.86235) | > loader_time: 0.00280 (0.00306)  --> STEP: 270/639 -- GLOBAL_STEP: 1550 | > loss: -0.04663 (0.00521) | > log_mle: -0.28200 (-0.27072) | > loss_dur: 0.23537 (0.27593) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 5.02634 (3.21985) | > current_lr: 0.00038 | > step_time: 0.85180 (0.84497) | > loader_time: 0.00390 (0.00309)  --> STEP: 295/639 -- GLOBAL_STEP: 1575 | > loss: -0.06226 (0.00013) | > log_mle: -0.26630 (-0.27171) | > loss_dur: 0.20404 (0.27184) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 1.07419 (3.14113) | > current_lr: 0.00039 | > step_time: 0.65480 (0.83039) | > loader_time: 0.00390 (0.00312)  --> STEP: 320/639 -- GLOBAL_STEP: 1600 | > loss: -0.08021 (-0.00614) | > log_mle: -0.30900 (-0.27389) | > loss_dur: 0.22879 (0.26774) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 8.15799 (3.19947) | > current_lr: 0.00040 | > step_time: 0.72070 (0.82170) | > loader_time: 0.00380 (0.00315)  --> STEP: 345/639 -- GLOBAL_STEP: 1625 | > loss: -0.10031 (-0.01211) | > log_mle: -0.30227 (-0.27594) | > loss_dur: 0.20197 (0.26384) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.14115 (3.27123) | > current_lr: 0.00040 | > step_time: 0.66120 (0.81575) | > loader_time: 0.00410 (0.00319)  --> STEP: 370/639 -- GLOBAL_STEP: 1650 | > loss: -0.11079 (-0.01793) | > log_mle: -0.31375 (-0.27760) | > loss_dur: 0.20295 (0.25967) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.14416 (3.28920) | > current_lr: 0.00041 | > step_time: 0.90520 (0.80917) | > loader_time: 0.00410 (0.00321)  --> STEP: 395/639 -- GLOBAL_STEP: 1675 | > loss: -0.10257 (-0.02335) | > log_mle: -0.30753 (-0.27927) | > loss_dur: 0.20496 (0.25592) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 5.35935 (3.25265) | > current_lr: 0.00042 | > step_time: 0.69710 (0.80592) | > loader_time: 0.00350 (0.00323)  --> STEP: 420/639 -- GLOBAL_STEP: 1700 | > loss: -0.11316 (-0.02837) | > log_mle: -0.30623 (-0.28103) | > loss_dur: 0.19307 (0.25266) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.95872 (3.23905) | > current_lr: 0.00042 | > step_time: 0.71640 (0.80527) | > loader_time: 0.00340 (0.00326)  --> STEP: 445/639 -- GLOBAL_STEP: 1725 | > loss: -0.09479 (-0.03349) | > log_mle: -0.29200 (-0.28262) | > loss_dur: 0.19720 (0.24912) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.49762 (3.30317) | > current_lr: 0.00043 | > step_time: 0.74700 (0.80490) | > loader_time: 0.00330 (0.00328)  --> STEP: 470/639 -- GLOBAL_STEP: 1750 | > loss: -0.13021 (-0.03790) | > log_mle: -0.31539 (-0.28387) | > loss_dur: 0.18518 (0.24598) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.60932 (3.26882) | > current_lr: 0.00043 | > step_time: 0.76410 (0.80550) | > loader_time: 0.00380 (0.00330)  --> STEP: 495/639 -- GLOBAL_STEP: 1775 | > loss: -0.09851 (-0.04214) | > log_mle: -0.27487 (-0.28523) | > loss_dur: 0.17636 (0.24309) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 1.17776 (3.31178) | > current_lr: 0.00044 | > step_time: 0.76690 (0.80860) | > loader_time: 0.00350 (0.00333)  --> STEP: 520/639 -- GLOBAL_STEP: 1800 | > loss: -0.14689 (-0.04669) | > log_mle: -0.32444 (-0.28686) | > loss_dur: 0.17755 (0.24017) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.04049 (3.26728) | > current_lr: 0.00045 | > step_time: 0.78060 (0.81200) | > loader_time: 0.00370 (0.00335)  --> STEP: 545/639 -- GLOBAL_STEP: 1825 | > loss: -0.14041 (-0.05075) | > log_mle: -0.33044 (-0.28834) | > loss_dur: 0.19003 (0.23759) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.31131 (3.24475) | > current_lr: 0.00045 | > step_time: 0.96330 (0.81931) | > loader_time: 0.00390 (0.00339)  --> STEP: 570/639 -- GLOBAL_STEP: 1850 | > loss: -0.14250 (-0.05440) | > log_mle: -0.30917 (-0.28946) | > loss_dur: 0.16667 (0.23506) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.33352 (3.19753) | > current_lr: 0.00046 | > step_time: 0.86180 (0.82582) | > loader_time: 0.00390 (0.00343)  --> STEP: 595/639 -- GLOBAL_STEP: 1875 | > loss: -0.13914 (-0.05823) | > log_mle: -0.30880 (-0.29090) | > loss_dur: 0.16966 (0.23267) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.64673 (3.22488) | > current_lr: 0.00047 | > step_time: 1.05430 (0.83619) | > loader_time: 0.00650 (0.00348)  --> STEP: 620/639 -- GLOBAL_STEP: 1900 | > loss: -0.13537 (-0.06188) | > log_mle: -0.30522 (-0.29225) | > loss_dur: 0.16984 (0.23037) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 13.32741 (3.22455) | > current_lr: 0.00047 | > step_time: 1.21150 (0.84752) | > loader_time: 0.00470 (0.00355) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.04439 (+0.01649) | > avg_loss: -0.15140 (-0.20249) | > avg_log_mle: -0.32987 (-0.07741) | > avg_loss_dur: 0.17847 (-0.12508) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_1920.pth.tar  > EPOCH: 3/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 15:06:35)   --> STEP: 5/639 -- GLOBAL_STEP: 1925 | > loss: -0.06607 (-0.04962) | > log_mle: -0.29526 (-0.29518) | > loss_dur: 0.22919 (0.24555) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.33299 (3.87250) | > current_lr: 0.00048 | > step_time: 0.33520 (0.28402) | > loader_time: 0.00270 (0.00191)  --> STEP: 30/639 -- GLOBAL_STEP: 1950 | > loss: -0.13223 (-0.06951) | > log_mle: -0.32027 (-0.29770) | > loss_dur: 0.18804 (0.22819) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.78277 (3.02076) | > current_lr: 0.00048 | > step_time: 0.39250 (0.32992) | > loader_time: 0.00270 (0.00245)  --> STEP: 55/639 -- GLOBAL_STEP: 1975 | > loss: -0.11207 (-0.08649) | > log_mle: -0.28068 (-0.30263) | > loss_dur: 0.16861 (0.21614) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 1.14794 (2.88699) | > current_lr: 0.00049 | > step_time: 0.33660 (0.35331) | > loader_time: 0.00270 (0.00255)  --> STEP: 80/639 -- GLOBAL_STEP: 2000 | > loss: -0.12039 (-0.09920) | > log_mle: -0.31553 (-0.30757) | > loss_dur: 0.19513 (0.20837) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 14.23869 (2.89739) | > current_lr: 0.00050 | > step_time: 0.39890 (0.37236) | > loader_time: 0.00290 (0.00266)  --> STEP: 105/639 -- GLOBAL_STEP: 2025 | > loss: -0.10868 (-0.10843) | > log_mle: -0.30293 (-0.31261) | > loss_dur: 0.19425 (0.20418) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.67648 (3.10392) | > current_lr: 0.00050 | > step_time: 0.41220 (0.39094) | > loader_time: 0.00300 (0.00273)  --> STEP: 130/639 -- GLOBAL_STEP: 2050 | > loss: -0.13303 (-0.11599) | > log_mle: -0.33412 (-0.31621) | > loss_dur: 0.20109 (0.20022) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.23387 (3.05810) | > current_lr: 0.00051 | > step_time: 0.48240 (0.40615) | > loader_time: 0.00300 (0.00278)  --> STEP: 155/639 -- GLOBAL_STEP: 2075 | > loss: -0.15292 (-0.12402) | > log_mle: -0.32602 (-0.32026) | > loss_dur: 0.17311 (0.19623) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 1.70006 (3.22959) | > current_lr: 0.00052 | > step_time: 0.52900 (0.42239) | > loader_time: 0.00310 (0.00283)  --> STEP: 180/639 -- GLOBAL_STEP: 2100 | > loss: -0.18092 (-0.13091) | > log_mle: -0.36471 (-0.32470) | > loss_dur: 0.18379 (0.19380) | > amp_scaler: 4096.00000 (7782.40000) | > grad_norm: 4.90567 (3.44233) | > current_lr: 0.00052 | > step_time: 0.58850 (0.43774) | > loader_time: 0.00330 (0.00287)  --> STEP: 205/639 -- GLOBAL_STEP: 2125 | > loss: -0.16457 (-0.13430) | > log_mle: -0.32589 (-0.32563) | > loss_dur: 0.16132 (0.19134) | > amp_scaler: 4096.00000 (7332.83902) | > grad_norm: 2.77604 (3.36318) | > current_lr: 0.00053 | > step_time: 0.57730 (0.45323) | > loader_time: 0.00310 (0.00291)  --> STEP: 230/639 -- GLOBAL_STEP: 2150 | > loss: -0.15679 (-0.13925) | > log_mle: -0.32897 (-0.32863) | > loss_dur: 0.17218 (0.18938) | > amp_scaler: 4096.00000 (6981.00870) | > grad_norm: 1.95622 (3.52399) | > current_lr: 0.00053 | > step_time: 0.55620 (0.46909) | > loader_time: 0.00340 (0.00296)  --> STEP: 255/639 -- GLOBAL_STEP: 2175 | > loss: -0.14246 (-0.14194) | > log_mle: -0.32055 (-0.32927) | > loss_dur: 0.17809 (0.18733) | > amp_scaler: 4096.00000 (6698.16471) | > grad_norm: 1.14556 (3.41384) | > current_lr: 0.00054 | > step_time: 0.62960 (0.48308) | > loader_time: 0.00340 (0.00300)  --> STEP: 280/639 -- GLOBAL_STEP: 2200 | > loss: -0.16980 (-0.14529) | > log_mle: -0.34453 (-0.33133) | > loss_dur: 0.17473 (0.18604) | > amp_scaler: 4096.00000 (6465.82857) | > grad_norm: 1.85253 (3.36005) | > current_lr: 0.00055 | > step_time: 0.58740 (0.50016) | > loader_time: 0.00320 (0.00305)  --> STEP: 305/639 -- GLOBAL_STEP: 2225 | > loss: -0.16966 (-0.14811) | > log_mle: -0.34327 (-0.33292) | > loss_dur: 0.17361 (0.18481) | > amp_scaler: 4096.00000 (6271.58033) | > grad_norm: 4.05083 (3.32781) | > current_lr: 0.00055 | > step_time: 0.66440 (0.51468) | > loader_time: 0.00330 (0.00308)  --> STEP: 330/639 -- GLOBAL_STEP: 2250 | > loss: -0.18046 (-0.15060) | > log_mle: -0.37083 (-0.33425) | > loss_dur: 0.19036 (0.18365) | > amp_scaler: 4096.00000 (6106.76364) | > grad_norm: 3.26126 (3.30510) | > current_lr: 0.00056 | > step_time: 0.73390 (0.52936) | > loader_time: 0.00360 (0.00312)  --> STEP: 355/639 -- GLOBAL_STEP: 2275 | > loss: -0.17018 (-0.15368) | > log_mle: -0.32312 (-0.33579) | > loss_dur: 0.15294 (0.18211) | > amp_scaler: 4096.00000 (5965.16056) | > grad_norm: 0.89925 (3.42173) | > current_lr: 0.00057 | > step_time: 0.61430 (0.54225) | > loader_time: 0.00340 (0.00315)  --> STEP: 380/639 -- GLOBAL_STEP: 2300 | > loss: -0.20732 (-0.15620) | > log_mle: -0.34996 (-0.33680) | > loss_dur: 0.14264 (0.18060) | > amp_scaler: 4096.00000 (5842.18947) | > grad_norm: 1.67287 (3.36170) | > current_lr: 0.00057 | > step_time: 0.63720 (0.55592) | > loader_time: 0.00370 (0.00319)  --> STEP: 405/639 -- GLOBAL_STEP: 2325 | > loss: -0.19780 (-0.15880) | > log_mle: -0.35306 (-0.33845) | > loss_dur: 0.15526 (0.17965) | > amp_scaler: 4096.00000 (5734.40000) | > grad_norm: 3.21408 (3.36653) | > current_lr: 0.00058 | > step_time: 0.67650 (0.56705) | > loader_time: 0.00340 (0.00322)  --> STEP: 430/639 -- GLOBAL_STEP: 2350 | > loss: -0.16919 (-0.16092) | > log_mle: -0.32949 (-0.33951) | > loss_dur: 0.16030 (0.17859) | > amp_scaler: 4096.00000 (5639.14419) | > grad_norm: 1.28109 (3.39812) | > current_lr: 0.00058 | > step_time: 0.72910 (0.58032) | > loader_time: 0.00370 (0.00326)  --> STEP: 455/639 -- GLOBAL_STEP: 2375 | > loss: -0.19147 (-0.16316) | > log_mle: -0.34676 (-0.34038) | > loss_dur: 0.15529 (0.17721) | > amp_scaler: 4096.00000 (5554.35604) | > grad_norm: 5.51635 (3.38138) | > current_lr: 0.00059 | > step_time: 0.73260 (0.59078) | > loader_time: 0.00390 (0.00329)  --> STEP: 480/639 -- GLOBAL_STEP: 2400 | > loss: -0.16294 (-0.16499) | > log_mle: -0.31795 (-0.34133) | > loss_dur: 0.15501 (0.17634) | > amp_scaler: 4096.00000 (5478.40000) | > grad_norm: 0.79889 (3.33389) | > current_lr: 0.00060 | > step_time: 0.83150 (0.60372) | > loader_time: 0.00350 (0.00332)  --> STEP: 505/639 -- GLOBAL_STEP: 2425 | > loss: -0.23351 (-0.16714) | > log_mle: -0.37670 (-0.34232) | > loss_dur: 0.14319 (0.17518) | > amp_scaler: 4096.00000 (5409.96436) | > grad_norm: 2.41369 (3.33330) | > current_lr: 0.00060 | > step_time: 0.92510 (0.61661) | > loader_time: 0.00400 (0.00335)  --> STEP: 530/639 -- GLOBAL_STEP: 2450 | > loss: -0.20811 (-0.16930) | > log_mle: -0.36889 (-0.34353) | > loss_dur: 0.16078 (0.17422) | > amp_scaler: 4096.00000 (5347.98491) | > grad_norm: 6.35403 (3.32389) | > current_lr: 0.00061 | > step_time: 0.95790 (0.63075) | > loader_time: 0.00410 (0.00338)  --> STEP: 555/639 -- GLOBAL_STEP: 2475 | > loss: -0.21191 (-0.17113) | > log_mle: -0.35819 (-0.34437) | > loss_dur: 0.14627 (0.17324) | > amp_scaler: 4096.00000 (5291.58919) | > grad_norm: 1.23699 (3.29997) | > current_lr: 0.00062 | > step_time: 0.84000 (0.64498) | > loader_time: 0.00360 (0.00342)  --> STEP: 580/639 -- GLOBAL_STEP: 2500 | > loss: -0.23182 (-0.17326) | > log_mle: -0.36938 (-0.34549) | > loss_dur: 0.13756 (0.17223) | > amp_scaler: 4096.00000 (5240.05517) | > grad_norm: 1.17508 (3.28555) | > current_lr: 0.00062 | > step_time: 1.00050 (0.66034) | > loader_time: 0.00440 (0.00345)  --> STEP: 605/639 -- GLOBAL_STEP: 2525 | > loss: -0.21987 (-0.17533) | > log_mle: -0.35683 (-0.34670) | > loss_dur: 0.13695 (0.17137) | > amp_scaler: 4096.00000 (5192.78017) | > grad_norm: 1.63117 (3.26145) | > current_lr: 0.00063 | > step_time: 1.02720 (0.67607) | > loader_time: 0.00440 (0.00349)  --> STEP: 630/639 -- GLOBAL_STEP: 2550 | > loss: -0.22463 (-0.17719) | > log_mle: -0.38322 (-0.34786) | > loss_dur: 0.15859 (0.17066) | > amp_scaler: 4096.00000 (5149.25714) | > grad_norm: 4.97386 (3.26910) | > current_lr: 0.00063 | > step_time: 1.27030 (0.69392) | > loader_time: 0.00470 (0.00354) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.01990 (-0.02449) | > avg_loss: -0.22066 (-0.06926) | > avg_log_mle: -0.37612 (-0.04625) | > avg_loss_dur: 0.15546 (-0.02301) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_2560.pth.tar  > EPOCH: 4/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 15:14:24)   --> STEP: 15/639 -- GLOBAL_STEP: 2575 | > loss: -0.17460 (-0.09824) | > log_mle: -0.38470 (-0.29388) | > loss_dur: 0.21010 (0.19564) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.36666 (4.55301) | > current_lr: 0.00064 | > step_time: 0.33050 (0.28925) | > loader_time: 0.00260 (0.00251)  --> STEP: 40/639 -- GLOBAL_STEP: 2600 | > loss: -0.14742 (-0.11698) | > log_mle: -0.32632 (-0.30754) | > loss_dur: 0.17889 (0.19056) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.03552 (3.18473) | > current_lr: 0.00065 | > step_time: 0.38690 (0.32841) | > loader_time: 0.00270 (0.00265)  --> STEP: 65/639 -- GLOBAL_STEP: 2625 | > loss: -0.18843 (-0.13456) | > log_mle: -0.35512 (-0.32204) | > loss_dur: 0.16669 (0.18748) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.07777 (2.93605) | > current_lr: 0.00065 | > step_time: 0.39740 (0.35731) | > loader_time: 0.00300 (0.00273)  --> STEP: 90/639 -- GLOBAL_STEP: 2650 | > loss: -0.12694 (-0.14811) | > log_mle: -0.32482 (-0.33007) | > loss_dur: 0.19788 (0.18196) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.90018 (2.94171) | > current_lr: 0.00066 | > step_time: 0.51730 (0.37475) | > loader_time: 0.00270 (0.00279)  --> STEP: 115/639 -- GLOBAL_STEP: 2675 | > loss: -0.12960 (-0.15895) | > log_mle: -0.30158 (-0.33872) | > loss_dur: 0.17198 (0.17978) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.09552 (2.91028) | > current_lr: 0.00067 | > step_time: 0.38390 (0.39034) | > loader_time: 0.00290 (0.00284)  --> STEP: 140/639 -- GLOBAL_STEP: 2700 | > loss: -0.17569 (-0.16908) | > log_mle: -0.36181 (-0.34598) | > loss_dur: 0.18613 (0.17690) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.58394 (2.98754) | > current_lr: 0.00067 | > step_time: 0.55110 (0.40419) | > loader_time: 0.00320 (0.00287)  --> STEP: 165/639 -- GLOBAL_STEP: 2725 | > loss: -0.22411 (-0.17691) | > log_mle: -0.38353 (-0.35092) | > loss_dur: 0.15941 (0.17402) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.12602 (3.10313) | > current_lr: 0.00068 | > step_time: 0.50050 (0.41837) | > loader_time: 0.00310 (0.00291)  --> STEP: 190/639 -- GLOBAL_STEP: 2750 | > loss: -0.16798 (-0.18266) | > log_mle: -0.33359 (-0.35481) | > loss_dur: 0.16560 (0.17215) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.97485 (3.18267) | > current_lr: 0.00068 | > step_time: 0.52600 (0.43709) | > loader_time: 0.00330 (0.00297)  --> STEP: 215/639 -- GLOBAL_STEP: 2775 | > loss: -0.27483 (-0.18691) | > log_mle: -0.41253 (-0.35750) | > loss_dur: 0.13770 (0.17059) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.02614 (3.18508) | > current_lr: 0.00069 | > step_time: 0.53020 (0.45372) | > loader_time: 0.00320 (0.00301)  --> STEP: 240/639 -- GLOBAL_STEP: 2800 | > loss: -0.26230 (-0.19310) | > log_mle: -0.41009 (-0.36172) | > loss_dur: 0.14779 (0.16862) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.91181 (3.32300) | > current_lr: 0.00070 | > step_time: 0.56370 (0.46856) | > loader_time: 0.00330 (0.00305)  --> STEP: 265/639 -- GLOBAL_STEP: 2825 | > loss: -0.30118 (-0.19743) | > log_mle: -0.44902 (-0.36461) | > loss_dur: 0.14784 (0.16718) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.49811 (3.41288) | > current_lr: 0.00070 | > step_time: 0.63730 (0.48515) | > loader_time: 0.00330 (0.00308)  --> STEP: 290/639 -- GLOBAL_STEP: 2850 | > loss: -0.21987 (-0.19944) | > log_mle: -0.37369 (-0.36561) | > loss_dur: 0.15381 (0.16616) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.45064 (3.36817) | > current_lr: 0.00071 | > step_time: 0.67560 (0.50140) | > loader_time: 0.00340 (0.00312)  --> STEP: 315/639 -- GLOBAL_STEP: 2875 | > loss: -0.21984 (-0.20216) | > log_mle: -0.35395 (-0.36722) | > loss_dur: 0.13411 (0.16506) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.94933 (3.34396) | > current_lr: 0.00072 | > step_time: 0.63250 (0.51761) | > loader_time: 0.00370 (0.00316)  --> STEP: 340/639 -- GLOBAL_STEP: 2900 | > loss: -0.23905 (-0.20446) | > log_mle: -0.37809 (-0.36888) | > loss_dur: 0.13904 (0.16442) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.68857 (3.47726) | > current_lr: 0.00072 | > step_time: 0.61750 (0.53421) | > loader_time: 0.00370 (0.00320)  --> STEP: 365/639 -- GLOBAL_STEP: 2925 | > loss: -0.24047 (-0.20652) | > log_mle: -0.38550 (-0.36973) | > loss_dur: 0.14503 (0.16321) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.10430 (3.40358) | > current_lr: 0.00073 | > step_time: 0.81320 (0.54608) | > loader_time: 0.00380 (0.00323)  --> STEP: 390/639 -- GLOBAL_STEP: 2950 | > loss: -0.24256 (-0.20877) | > log_mle: -0.39247 (-0.37088) | > loss_dur: 0.14990 (0.16211) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.72872 (3.40340) | > current_lr: 0.00073 | > step_time: 0.72020 (0.55954) | > loader_time: 0.00370 (0.00325)  --> STEP: 415/639 -- GLOBAL_STEP: 2975 | > loss: -0.22222 (-0.21069) | > log_mle: -0.38020 (-0.37218) | > loss_dur: 0.15798 (0.16149) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.62256 (3.36276) | > current_lr: 0.00074 | > step_time: 0.72470 (0.57127) | > loader_time: 0.00350 (0.00328)  --> STEP: 440/639 -- GLOBAL_STEP: 3000 | > loss: -0.21500 (-0.21259) | > log_mle: -0.36638 (-0.37316) | > loss_dur: 0.15137 (0.16057) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.27915 (3.31779) | > current_lr: 0.00075 | > step_time: 0.75590 (0.58352) | > loader_time: 0.00330 (0.00330)  --> STEP: 465/639 -- GLOBAL_STEP: 3025 | > loss: -0.23424 (-0.21405) | > log_mle: -0.37014 (-0.37373) | > loss_dur: 0.13590 (0.15969) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.69645 (3.26787) | > current_lr: 0.00075 | > step_time: 0.80550 (0.59512) | > loader_time: 0.00370 (0.00333)  --> STEP: 490/639 -- GLOBAL_STEP: 3050 | > loss: -0.22240 (-0.21537) | > log_mle: -0.35923 (-0.37437) | > loss_dur: 0.13684 (0.15900) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.99173 (3.24336) | > current_lr: 0.00076 | > step_time: 0.78800 (0.60835) | > loader_time: 0.00380 (0.00336)  --> STEP: 515/639 -- GLOBAL_STEP: 3075 | > loss: -0.24115 (-0.21732) | > log_mle: -0.38404 (-0.37550) | > loss_dur: 0.14290 (0.15818) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.74193 (3.25702) | > current_lr: 0.00077 | > step_time: 0.96320 (0.62153) | > loader_time: 0.00410 (0.00339)  --> STEP: 540/639 -- GLOBAL_STEP: 3100 | > loss: -0.24411 (-0.21905) | > log_mle: -0.39779 (-0.37652) | > loss_dur: 0.15368 (0.15747) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.16770 (3.24373) | > current_lr: 0.00077 | > step_time: 0.89470 (0.63640) | > loader_time: 0.00410 (0.00343)  --> STEP: 565/639 -- GLOBAL_STEP: 3125 | > loss: -0.23507 (-0.22054) | > log_mle: -0.37454 (-0.37733) | > loss_dur: 0.13947 (0.15680) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.01585 (3.26090) | > current_lr: 0.00078 | > step_time: 0.99200 (0.65000) | > loader_time: 0.00410 (0.00346)  --> STEP: 590/639 -- GLOBAL_STEP: 3150 | > loss: -0.26290 (-0.22239) | > log_mle: -0.40255 (-0.37850) | > loss_dur: 0.13965 (0.15611) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.52147 (3.25266) | > current_lr: 0.00078 | > step_time: 0.94880 (0.66654) | > loader_time: 0.00450 (0.00351)  --> STEP: 615/639 -- GLOBAL_STEP: 3175 | > loss: -0.26572 (-0.22385) | > log_mle: -0.40335 (-0.37932) | > loss_dur: 0.13763 (0.15547) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.47143 (3.23674) | > current_lr: 0.00079 | > step_time: 1.02220 (0.68281) | > loader_time: 0.00440 (0.00354) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.04449 (+0.02459) | > avg_loss: -0.26484 (-0.04418) | > avg_log_mle: -0.41075 (-0.03463) | > avg_loss_dur: 0.14591 (-0.00955) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_3200.pth.tar  > EPOCH: 5/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 15:22:13)   --> STEP: 0/639 -- GLOBAL_STEP: 3200 | > loss: -0.25408 (-0.25408) | > log_mle: -0.45538 (-0.45538) | > loss_dur: 0.20129 (0.20129) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.18388 (6.18388) | > current_lr: 0.00080 | > step_time: 0.67570 (0.67565) | > loader_time: 0.40300 (0.40302)  --> STEP: 25/639 -- GLOBAL_STEP: 3225 | > loss: -0.18859 (-0.15289) | > log_mle: -0.37178 (-0.34154) | > loss_dur: 0.18319 (0.18865) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.88425 (3.84361) | > current_lr: 0.00080 | > step_time: 0.37180 (0.31872) | > loader_time: 0.00310 (0.00269)  --> STEP: 50/639 -- GLOBAL_STEP: 3250 | > loss: -0.19690 (-0.16796) | > log_mle: -0.34417 (-0.34967) | > loss_dur: 0.14728 (0.18171) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.64350 (3.22394) | > current_lr: 0.00081 | > step_time: 0.33630 (0.34514) | > loader_time: 0.00290 (0.00272)  --> STEP: 75/639 -- GLOBAL_STEP: 3275 | > loss: -0.23840 (-0.18056) | > log_mle: -0.38921 (-0.35730) | > loss_dur: 0.15081 (0.17673) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.95316 (2.88572) | > current_lr: 0.00082 | > step_time: 0.41370 (0.37010) | > loader_time: 0.00290 (0.00279)  --> STEP: 100/639 -- GLOBAL_STEP: 3300 | > loss: -0.24190 (-0.19291) | > log_mle: -0.38886 (-0.36472) | > loss_dur: 0.14696 (0.17181) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.90055 (3.13101) | > current_lr: 0.00082 | > step_time: 0.55940 (0.39007) | > loader_time: 0.00310 (0.00285)  --> STEP: 125/639 -- GLOBAL_STEP: 3325 | > loss: -0.27581 (-0.20195) | > log_mle: -0.40847 (-0.37046) | > loss_dur: 0.13266 (0.16851) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.63069 (3.18463) | > current_lr: 0.00083 | > step_time: 0.50280 (0.40643) | > loader_time: 0.00310 (0.00289)  --> STEP: 150/639 -- GLOBAL_STEP: 3350 | > loss: -0.21881 (-0.21004) | > log_mle: -0.37454 (-0.37581) | > loss_dur: 0.15572 (0.16577) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.64535 (3.15462) | > current_lr: 0.00083 | > step_time: 0.43170 (0.42221) | > loader_time: 0.00260 (0.00292)  --> STEP: 175/639 -- GLOBAL_STEP: 3375 | > loss: -0.24692 (-0.21648) | > log_mle: -0.39206 (-0.38028) | > loss_dur: 0.14514 (0.16380) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.70862 (3.36597) | > current_lr: 0.00084 | > step_time: 0.60020 (0.43784) | > loader_time: 0.00340 (0.00295)  --> STEP: 200/639 -- GLOBAL_STEP: 3400 | > loss: -0.24467 (-0.22069) | > log_mle: -0.39411 (-0.38251) | > loss_dur: 0.14944 (0.16182) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.82408 (3.33877) | > current_lr: 0.00085 | > step_time: 0.55590 (0.45297) | > loader_time: 0.00310 (0.00299)  --> STEP: 225/639 -- GLOBAL_STEP: 3425 | > loss: -0.24143 (-0.22437) | > log_mle: -0.37696 (-0.38470) | > loss_dur: 0.13553 (0.16033) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.82864 (3.30095) | > current_lr: 0.00085 | > step_time: 0.52210 (0.46793) | > loader_time: 0.00340 (0.00303)  --> STEP: 250/639 -- GLOBAL_STEP: 3450 | > loss: -0.25181 (-0.22818) | > log_mle: -0.38176 (-0.38769) | > loss_dur: 0.12995 (0.15951) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.64279 (3.31126) | > current_lr: 0.00086 | > step_time: 0.57710 (0.48233) | > loader_time: 0.00360 (0.00306)  --> STEP: 275/639 -- GLOBAL_STEP: 3475 | > loss: -0.26047 (-0.23201) | > log_mle: -0.41350 (-0.39055) | > loss_dur: 0.15304 (0.15854) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.15565 (3.44269) | > current_lr: 0.00087 | > step_time: 0.86480 (0.49987) | > loader_time: 0.00370 (0.00311)  --> STEP: 300/639 -- GLOBAL_STEP: 3500 | > loss: -0.27880 (-0.23426) | > log_mle: -0.41755 (-0.39207) | > loss_dur: 0.13875 (0.15781) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.85102 (3.44159) | > current_lr: 0.00087 | > step_time: 0.67600 (0.51316) | > loader_time: 0.00340 (0.00314)  --> STEP: 325/639 -- GLOBAL_STEP: 3525 | > loss: -0.27427 (-0.23677) | > log_mle: -0.41810 (-0.39381) | > loss_dur: 0.14382 (0.15704) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.59627 (3.45269) | > current_lr: 0.00088 | > step_time: 0.61620 (0.52793) | > loader_time: 0.00350 (0.00317)  --> STEP: 350/639 -- GLOBAL_STEP: 3550 | > loss: -0.29525 (-0.23930) | > log_mle: -0.45221 (-0.39539) | > loss_dur: 0.15696 (0.15609) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.98578 (3.45844) | > current_lr: 0.00088 | > step_time: 0.83900 (0.54225) | > loader_time: 0.00390 (0.00320)  --> STEP: 375/639 -- GLOBAL_STEP: 3575 | > loss: -0.25336 (-0.24074) | > log_mle: -0.39033 (-0.39585) | > loss_dur: 0.13697 (0.15511) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.26861 (3.40078) | > current_lr: 0.00089 | > step_time: 0.69550 (0.55503) | > loader_time: 0.00370 (0.00323)  --> STEP: 400/639 -- GLOBAL_STEP: 3600 | > loss: -0.27560 (-0.24290) | > log_mle: -0.41599 (-0.39723) | > loss_dur: 0.14039 (0.15433) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.69659 (3.40604) | > current_lr: 0.00090 | > step_time: 0.77470 (0.56778) | > loader_time: 0.00380 (0.00326)  --> STEP: 425/639 -- GLOBAL_STEP: 3625 | > loss: -0.32712 (-0.24433) | > log_mle: -0.46418 (-0.39796) | > loss_dur: 0.13705 (0.15363) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.21106 (3.37424) | > current_lr: 0.00090 | > step_time: 0.89740 (0.58056) | > loader_time: 0.00370 (0.00329)  --> STEP: 450/639 -- GLOBAL_STEP: 3650 | > loss: -0.26077 (-0.24565) | > log_mle: -0.39365 (-0.39835) | > loss_dur: 0.13288 (0.15269) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.05708 (3.34173) | > current_lr: 0.00091 | > step_time: 0.79020 (0.59369) | > loader_time: 0.00370 (0.00332)  --> STEP: 475/639 -- GLOBAL_STEP: 3675 | > loss: -0.27812 (-0.24692) | > log_mle: -0.40158 (-0.39900) | > loss_dur: 0.12346 (0.15208) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.40629 (3.31726) | > current_lr: 0.00092 | > step_time: 0.70250 (0.60503) | > loader_time: 0.00360 (0.00334)  --> STEP: 500/639 -- GLOBAL_STEP: 3700 | > loss: -0.29479 (-0.24839) | > log_mle: -0.42993 (-0.39976) | > loss_dur: 0.13513 (0.15137) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.32919 (3.30669) | > current_lr: 0.00092 | > step_time: 0.97280 (0.61816) | > loader_time: 0.00460 (0.00338)  --> STEP: 525/639 -- GLOBAL_STEP: 3725 | > loss: -0.28058 (-0.24987) | > log_mle: -0.43931 (-0.40056) | > loss_dur: 0.15873 (0.15069) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.46228 (3.25469) | > current_lr: 0.00093 | > step_time: 1.11900 (0.63278) | > loader_time: 0.00440 (0.00341)  --> STEP: 550/639 -- GLOBAL_STEP: 3750 | > loss: -0.25293 (-0.25110) | > log_mle: -0.38941 (-0.40115) | > loss_dur: 0.13648 (0.15005) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.90173 (3.23201) | > current_lr: 0.00093 | > step_time: 0.92480 (0.64716) | > loader_time: 0.00420 (0.00345)  --> STEP: 575/639 -- GLOBAL_STEP: 3775 | > loss: -0.27283 (-0.25230) | > log_mle: -0.40346 (-0.40182) | > loss_dur: 0.13063 (0.14952) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.35652 (3.22404) | > current_lr: 0.00094 | > step_time: 0.90870 (0.66186) | > loader_time: 0.00410 (0.00349)  --> STEP: 600/639 -- GLOBAL_STEP: 3800 | > loss: -0.24938 (-0.25332) | > log_mle: -0.39394 (-0.40237) | > loss_dur: 0.14456 (0.14905) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.68159 (3.23921) | > current_lr: 0.00095 | > step_time: 1.06690 (0.67815) | > loader_time: 0.00440 (0.00353)  --> STEP: 625/639 -- GLOBAL_STEP: 3825 | > loss: -0.28674 (-0.25443) | > log_mle: -0.42401 (-0.40291) | > loss_dur: 0.13727 (0.14849) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.37302 (3.23801) | > current_lr: 0.00095 | > step_time: 1.10510 (0.69494) | > loader_time: 0.00460 (0.00357) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03032 (-0.01417) | > avg_loss: -0.28074 (-0.01590) | > avg_log_mle: -0.41914 (-0.00839) | > avg_loss_dur: 0.13840 (-0.00751) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_3840.pth.tar  > EPOCH: 6/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 15:30:04)   --> STEP: 10/639 -- GLOBAL_STEP: 3850 | > loss: -0.21353 (-0.22108) | > log_mle: -0.39906 (-0.39739) | > loss_dur: 0.18553 (0.17631) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.30084 (3.84641) | > current_lr: 0.00096 | > step_time: 0.28310 (0.29289) | > loader_time: 0.00270 (0.00309)  --> STEP: 35/639 -- GLOBAL_STEP: 3875 | > loss: -0.22048 (-0.21098) | > log_mle: -0.37784 (-0.38295) | > loss_dur: 0.15737 (0.17197) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.29822 (2.78922) | > current_lr: 0.00097 | > step_time: 0.32210 (0.32413) | > loader_time: 0.00290 (0.00276)  --> STEP: 60/639 -- GLOBAL_STEP: 3900 | > loss: -0.20211 (-0.21706) | > log_mle: -0.36433 (-0.38632) | > loss_dur: 0.16223 (0.16926) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.88020 (2.91098) | > current_lr: 0.00097 | > step_time: 0.39650 (0.35610) | > loader_time: 0.00260 (0.00278)  --> STEP: 85/639 -- GLOBAL_STEP: 3925 | > loss: -0.19952 (-0.22376) | > log_mle: -0.36792 (-0.38810) | > loss_dur: 0.16840 (0.16434) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.82258 (2.78645) | > current_lr: 0.00098 | > step_time: 0.49230 (0.37308) | > loader_time: 0.00270 (0.00280)  --> STEP: 110/639 -- GLOBAL_STEP: 3950 | > loss: -0.22754 (-0.23132) | > log_mle: -0.38799 (-0.39406) | > loss_dur: 0.16044 (0.16274) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.10687 (2.99206) | > current_lr: 0.00098 | > step_time: 0.44040 (0.39271) | > loader_time: 0.00290 (0.00283)  --> STEP: 135/639 -- GLOBAL_STEP: 3975 | > loss: -0.21954 (-0.23746) | > log_mle: -0.38646 (-0.39781) | > loss_dur: 0.16692 (0.16036) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.37504 (3.08475) | > current_lr: 0.00099 | > step_time: 0.46760 (0.40707) | > loader_time: 0.00270 (0.00285)  --> STEP: 160/639 -- GLOBAL_STEP: 4000 | > loss: -0.29939 (-0.24213) | > log_mle: -0.42580 (-0.40012) | > loss_dur: 0.12641 (0.15799) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.35555 (3.07397) | > current_lr: 0.00100 | > step_time: 0.47920 (0.42093) | > loader_time: 0.00290 (0.00289)  --> STEP: 185/639 -- GLOBAL_STEP: 4025 | > loss: -0.24818 (-0.24829) | > log_mle: -0.39910 (-0.40537) | > loss_dur: 0.15092 (0.15709) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.02117 (3.39756) | > current_lr: 0.00100 | > step_time: 0.57560 (0.43824) | > loader_time: 0.00300 (0.00294)  --> STEP: 210/639 -- GLOBAL_STEP: 4050 | > loss: -0.21883 (-0.25051) | > log_mle: -0.35450 (-0.40631) | > loss_dur: 0.13567 (0.15580) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.67203 (3.49368) | > current_lr: 0.00100 | > step_time: 0.50810 (0.45379) | > loader_time: 0.00320 (0.00298)  --> STEP: 235/639 -- GLOBAL_STEP: 4075 | > loss: -0.28947 (-0.25485) | > log_mle: -0.42505 (-0.40910) | > loss_dur: 0.13558 (0.15425) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.93848 (3.44273) | > current_lr: 0.00099 | > step_time: 0.55350 (0.47030) | > loader_time: 0.00350 (0.00302)  --> STEP: 260/639 -- GLOBAL_STEP: 4100 | > loss: -0.28775 (-0.25800) | > log_mle: -0.42990 (-0.41118) | > loss_dur: 0.14214 (0.15319) | > amp_scaler: 8192.00000 (4379.56923) | > grad_norm: 2.34723 (3.42381) | > current_lr: 0.00099 | > step_time: 0.59330 (0.48576) | > loader_time: 0.00330 (0.00305)  --> STEP: 285/639 -- GLOBAL_STEP: 4125 | > loss: -0.24522 (-0.25986) | > log_mle: -0.39591 (-0.41225) | > loss_dur: 0.15069 (0.15238) | > amp_scaler: 8192.00000 (4713.99298) | > grad_norm: 1.56521 (3.41152) | > current_lr: 0.00099 | > step_time: 0.64780 (0.50166) | > loader_time: 0.00340 (0.00308)  --> STEP: 310/639 -- GLOBAL_STEP: 4150 | > loss: -0.28987 (-0.26199) | > log_mle: -0.43004 (-0.41370) | > loss_dur: 0.14017 (0.15172) | > amp_scaler: 8192.00000 (4994.47742) | > grad_norm: 2.44786 (3.45062) | > current_lr: 0.00098 | > step_time: 0.63850 (0.51589) | > loader_time: 0.00330 (0.00312)  --> STEP: 335/639 -- GLOBAL_STEP: 4175 | > loss: -0.31602 (-0.26373) | > log_mle: -0.46034 (-0.41503) | > loss_dur: 0.14433 (0.15131) | > amp_scaler: 8192.00000 (5233.09851) | > grad_norm: 7.16617 (3.47992) | > current_lr: 0.00098 | > step_time: 0.69340 (0.53221) | > loader_time: 0.00340 (0.00317)  --> STEP: 360/639 -- GLOBAL_STEP: 4200 | > loss: -0.27197 (-0.26595) | > log_mle: -0.40585 (-0.41626) | > loss_dur: 0.13388 (0.15031) | > amp_scaler: 8192.00000 (5438.57778) | > grad_norm: 1.24491 (3.44405) | > current_lr: 0.00098 | > step_time: 0.74400 (0.54491) | > loader_time: 0.00380 (0.00320)  --> STEP: 385/639 -- GLOBAL_STEP: 4225 | > loss: -0.29566 (-0.26757) | > log_mle: -0.43296 (-0.41692) | > loss_dur: 0.13730 (0.14935) | > amp_scaler: 8192.00000 (5617.37143) | > grad_norm: 3.74596 (3.42324) | > current_lr: 0.00097 | > step_time: 0.74040 (0.55927) | > loader_time: 0.00350 (0.00324)  --> STEP: 410/639 -- GLOBAL_STEP: 4250 | > loss: -0.27146 (-0.26877) | > log_mle: -0.42577 (-0.41764) | > loss_dur: 0.15431 (0.14887) | > amp_scaler: 8192.00000 (5774.36098) | > grad_norm: 1.45850 (3.42610) | > current_lr: 0.00097 | > step_time: 0.75960 (0.57084) | > loader_time: 0.00400 (0.00326)  --> STEP: 435/639 -- GLOBAL_STEP: 4275 | > loss: -0.28875 (-0.26990) | > log_mle: -0.42362 (-0.41809) | > loss_dur: 0.13488 (0.14819) | > amp_scaler: 8192.00000 (5913.30575) | > grad_norm: 1.41612 (3.40757) | > current_lr: 0.00097 | > step_time: 0.69710 (0.58302) | > loader_time: 0.00350 (0.00329)  --> STEP: 460/639 -- GLOBAL_STEP: 4300 | > loss: -0.25898 (-0.27104) | > log_mle: -0.40277 (-0.41857) | > loss_dur: 0.14379 (0.14752) | > amp_scaler: 8192.00000 (6037.14783) | > grad_norm: 0.81841 (3.35823) | > current_lr: 0.00097 | > step_time: 0.91880 (0.59486) | > loader_time: 0.00420 (0.00332)  --> STEP: 485/639 -- GLOBAL_STEP: 4325 | > loss: -0.30100 (-0.27216) | > log_mle: -0.43121 (-0.41924) | > loss_dur: 0.13022 (0.14709) | > amp_scaler: 8192.00000 (6148.22268) | > grad_norm: 2.11545 (3.34459) | > current_lr: 0.00096 | > step_time: 0.85730 (0.60737) | > loader_time: 0.00390 (0.00336)  --> STEP: 510/639 -- GLOBAL_STEP: 4350 | > loss: -0.31395 (-0.27371) | > log_mle: -0.44550 (-0.42013) | > loss_dur: 0.13155 (0.14642) | > amp_scaler: 8192.00000 (6248.40784) | > grad_norm: 2.96690 (3.32965) | > current_lr: 0.00096 | > step_time: 0.86960 (0.62035) | > loader_time: 0.00400 (0.00339)  --> STEP: 535/639 -- GLOBAL_STEP: 4375 | > loss: -0.29730 (-0.27520) | > log_mle: -0.43155 (-0.42100) | > loss_dur: 0.13425 (0.14580) | > amp_scaler: 8192.00000 (6339.22991) | > grad_norm: 5.08369 (3.30056) | > current_lr: 0.00096 | > step_time: 0.91700 (0.63358) | > loader_time: 0.00400 (0.00342)  --> STEP: 560/639 -- GLOBAL_STEP: 4400 | > loss: -0.30331 (-0.27642) | > log_mle: -0.43911 (-0.42178) | > loss_dur: 0.13580 (0.14536) | > amp_scaler: 8192.00000 (6421.94286) | > grad_norm: 2.31181 (3.27535) | > current_lr: 0.00095 | > step_time: 0.93050 (0.64849) | > loader_time: 0.00430 (0.00347)  --> STEP: 585/639 -- GLOBAL_STEP: 4425 | > loss: -0.31651 (-0.27782) | > log_mle: -0.44782 (-0.42260) | > loss_dur: 0.13131 (0.14478) | > amp_scaler: 8192.00000 (6497.58632) | > grad_norm: 2.47851 (3.26694) | > current_lr: 0.00095 | > step_time: 1.09780 (0.66439) | > loader_time: 0.00420 (0.00351)  --> STEP: 610/639 -- GLOBAL_STEP: 4450 | > loss: -0.29427 (-0.27905) | > log_mle: -0.41955 (-0.42338) | > loss_dur: 0.12528 (0.14433) | > amp_scaler: 8192.00000 (6567.02951) | > grad_norm: 3.41498 (3.25147) | > current_lr: 0.00095 | > step_time: 1.06320 (0.68054) | > loader_time: 0.00440 (0.00356)  --> STEP: 635/639 -- GLOBAL_STEP: 4475 | > loss: -0.31608 (-0.28030) | > log_mle: -0.44712 (-0.42438) | > loss_dur: 0.13104 (0.14408) | > amp_scaler: 8192.00000 (6631.00472) | > grad_norm: 3.35203 (3.25847) | > current_lr: 0.00095 | > step_time: 1.13420 (0.69773) | > loader_time: 0.00640 (0.00361) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02325 (-0.00707) | > avg_loss: -0.31653 (-0.03579) | > avg_log_mle: -0.45139 (-0.03224) | > avg_loss_dur: 0.13486 (-0.00354) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_4480.pth.tar  > EPOCH: 7/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 15:37:53)   --> STEP: 20/639 -- GLOBAL_STEP: 4500 | > loss: -0.35188 (-0.27202) | > log_mle: -0.52969 (-0.44205) | > loss_dur: 0.17781 (0.17003) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 8.02208 (4.97525) | > current_lr: 0.00094 | > step_time: 0.33920 (0.31043) | > loader_time: 0.00280 (0.00491)  --> STEP: 45/639 -- GLOBAL_STEP: 4525 | > loss: -0.22080 (-0.25890) | > log_mle: -0.37756 (-0.42569) | > loss_dur: 0.15676 (0.16678) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 1.87379 (3.79788) | > current_lr: 0.00094 | > step_time: 0.41540 (0.34007) | > loader_time: 0.00270 (0.00374)  --> STEP: 70/639 -- GLOBAL_STEP: 4550 | > loss: -0.30621 (-0.25924) | > log_mle: -0.45340 (-0.42311) | > loss_dur: 0.14719 (0.16387) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.93244 (3.47694) | > current_lr: 0.00094 | > step_time: 0.39230 (0.36535) | > loader_time: 0.00300 (0.00344)  --> STEP: 95/639 -- GLOBAL_STEP: 4575 | > loss: -0.34396 (-0.26581) | > log_mle: -0.48555 (-0.42477) | > loss_dur: 0.14159 (0.15896) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 4.43127 (3.57854) | > current_lr: 0.00094 | > step_time: 0.40550 (0.38032) | > loader_time: 0.00270 (0.00330)  --> STEP: 120/639 -- GLOBAL_STEP: 4600 | > loss: -0.32041 (-0.26954) | > log_mle: -0.48501 (-0.42680) | > loss_dur: 0.16460 (0.15726) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 7.71825 (3.50171) | > current_lr: 0.00093 | > step_time: 0.54440 (0.39835) | > loader_time: 0.00310 (0.00324)  --> STEP: 145/639 -- GLOBAL_STEP: 4625 | > loss: -0.30445 (-0.27595) | > log_mle: -0.45653 (-0.43114) | > loss_dur: 0.15208 (0.15519) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 7.91826 (3.81293) | > current_lr: 0.00093 | > step_time: 0.46390 (0.41383) | > loader_time: 0.00280 (0.00321)  --> STEP: 170/639 -- GLOBAL_STEP: 4650 | > loss: -0.30536 (-0.28020) | > log_mle: -0.45408 (-0.43350) | > loss_dur: 0.14873 (0.15330) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.18756 (3.98022) | > current_lr: 0.00093 | > step_time: 0.55780 (0.42706) | > loader_time: 0.00330 (0.00319)  --> STEP: 195/639 -- GLOBAL_STEP: 4675 | > loss: -0.28850 (-0.28367) | > log_mle: -0.40897 (-0.43553) | > loss_dur: 0.12047 (0.15186) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 1.11668 (3.98627) | > current_lr: 0.00093 | > step_time: 0.50560 (0.44307) | > loader_time: 0.00330 (0.00320)  --> STEP: 220/639 -- GLOBAL_STEP: 4700 | > loss: -0.32694 (-0.28602) | > log_mle: -0.45048 (-0.43658) | > loss_dur: 0.12354 (0.15056) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.20231 (3.90531) | > current_lr: 0.00092 | > step_time: 0.61060 (0.45945) | > loader_time: 0.00350 (0.00322)  --> STEP: 245/639 -- GLOBAL_STEP: 4725 | > loss: -0.31800 (-0.29033) | > log_mle: -0.44319 (-0.43944) | > loss_dur: 0.12519 (0.14911) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.92911 (3.95232) | > current_lr: 0.00092 | > step_time: 0.58980 (0.47405) | > loader_time: 0.00350 (0.00323)  --> STEP: 270/639 -- GLOBAL_STEP: 4750 | > loss: -0.34811 (-0.29355) | > log_mle: -0.48965 (-0.44182) | > loss_dur: 0.14154 (0.14827) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 6.87173 (3.99665) | > current_lr: 0.00092 | > step_time: 0.86360 (0.49147) | > loader_time: 0.00390 (0.00325)  --> STEP: 295/639 -- GLOBAL_STEP: 4775 | > loss: -0.28907 (-0.29413) | > log_mle: -0.41587 (-0.44190) | > loss_dur: 0.12680 (0.14778) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 1.35966 (3.92705) | > current_lr: 0.00092 | > step_time: 0.65030 (0.50546) | > loader_time: 0.00370 (0.00327)  --> STEP: 320/639 -- GLOBAL_STEP: 4800 | > loss: -0.31044 (-0.29564) | > log_mle: -0.46007 (-0.44277) | > loss_dur: 0.14963 (0.14714) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.63753 (3.95089) | > current_lr: 0.00091 | > step_time: 0.75890 (0.52191) | > loader_time: 0.00350 (0.00330)  --> STEP: 345/639 -- GLOBAL_STEP: 4825 | > loss: -0.31907 (-0.29671) | > log_mle: -0.45473 (-0.44353) | > loss_dur: 0.13566 (0.14683) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 4.18119 (3.98906) | > current_lr: 0.00091 | > step_time: 0.62670 (0.53625) | > loader_time: 0.00340 (0.00332)  --> STEP: 370/639 -- GLOBAL_STEP: 4850 | > loss: -0.32322 (-0.29766) | > log_mle: -0.45966 (-0.44366) | > loss_dur: 0.13644 (0.14601) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 1.86046 (3.92195) | > current_lr: 0.00091 | > step_time: 0.87990 (0.54858) | > loader_time: 0.00380 (0.00333)  --> STEP: 395/639 -- GLOBAL_STEP: 4875 | > loss: -0.29856 (-0.29848) | > log_mle: -0.44742 (-0.44394) | > loss_dur: 0.14886 (0.14546) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.44881 (3.87458) | > current_lr: 0.00091 | > step_time: 0.69130 (0.56097) | > loader_time: 0.00360 (0.00335)  --> STEP: 420/639 -- GLOBAL_STEP: 4900 | > loss: -0.31426 (-0.29910) | > log_mle: -0.44514 (-0.44428) | > loss_dur: 0.13088 (0.14518) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.33838 (3.83538) | > current_lr: 0.00090 | > step_time: 0.68530 (0.57375) | > loader_time: 0.00340 (0.00337)  --> STEP: 445/639 -- GLOBAL_STEP: 4925 | > loss: -0.30949 (-0.30020) | > log_mle: -0.44233 (-0.44465) | > loss_dur: 0.13284 (0.14445) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 1.85924 (3.81751) | > current_lr: 0.00090 | > step_time: 0.71960 (0.58576) | > loader_time: 0.00360 (0.00339)  --> STEP: 470/639 -- GLOBAL_STEP: 4950 | > loss: -0.32774 (-0.30079) | > log_mle: -0.45129 (-0.44463) | > loss_dur: 0.12355 (0.14384) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.41709 (3.75887) | > current_lr: 0.00090 | > step_time: 0.80440 (0.59696) | > loader_time: 0.00360 (0.00341)  --> STEP: 495/639 -- GLOBAL_STEP: 4975 | > loss: -0.28395 (-0.30148) | > log_mle: -0.40504 (-0.44480) | > loss_dur: 0.12108 (0.14332) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 1.15855 (3.78173) | > current_lr: 0.00090 | > step_time: 0.75790 (0.60964) | > loader_time: 0.00360 (0.00343)  --> STEP: 520/639 -- GLOBAL_STEP: 5000 | > loss: -0.34062 (-0.30261) | > log_mle: -0.46011 (-0.44524) | > loss_dur: 0.11950 (0.14263) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.97699 (3.72991) | > current_lr: 0.00090 | > step_time: 0.84220 (0.62296) | > loader_time: 0.00390 (0.00345)  --> STEP: 545/639 -- GLOBAL_STEP: 5025 | > loss: -0.32099 (-0.30352) | > log_mle: -0.47003 (-0.44574) | > loss_dur: 0.14904 (0.14222) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.35173 (3.71119) | > current_lr: 0.00089 | > step_time: 0.92220 (0.63766) | > loader_time: 0.00410 (0.00348)  --> STEP: 570/639 -- GLOBAL_STEP: 5050 | > loss: -0.32206 (-0.30422) | > log_mle: -0.44405 (-0.44600) | > loss_dur: 0.12199 (0.14177) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 1.63165 (3.67711) | > current_lr: 0.00089 | > step_time: 0.82840 (0.65100) | > loader_time: 0.00400 (0.00351)  --> STEP: 595/639 -- GLOBAL_STEP: 5075 | > loss: -0.30891 (-0.30523) | > log_mle: -0.43676 (-0.44656) | > loss_dur: 0.12785 (0.14133) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.81081 (3.67170) | > current_lr: 0.00089 | > step_time: 0.96630 (0.66684) | > loader_time: 0.00440 (0.00355)  --> STEP: 620/639 -- GLOBAL_STEP: 5100 | > loss: -0.30626 (-0.30576) | > log_mle: -0.43656 (-0.44666) | > loss_dur: 0.13030 (0.14090) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.46891 (3.63694) | > current_lr: 0.00089 | > step_time: 1.14010 (0.68295) | > loader_time: 0.00470 (0.00358) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03678 (+0.01352) | > avg_loss: -0.32667 (-0.01014) | > avg_log_mle: -0.45650 (-0.00512) | > avg_loss_dur: 0.12984 (-0.00502) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_5120.pth.tar  > EPOCH: 8/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 15:45:40)   --> STEP: 5/639 -- GLOBAL_STEP: 5125 | > loss: -0.27688 (-0.28093) | > log_mle: -0.47828 (-0.46421) | > loss_dur: 0.20140 (0.18328) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 13.80414 (7.48962) | > current_lr: 0.00088 | > step_time: 0.31500 (0.28766) | > loader_time: 0.00220 (0.00373)  --> STEP: 30/639 -- GLOBAL_STEP: 5150 | > loss: -0.30078 (-0.26375) | > log_mle: -0.45336 (-0.43366) | > loss_dur: 0.15258 (0.16991) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.07405 (4.29683) | > current_lr: 0.00088 | > step_time: 0.35290 (0.33396) | > loader_time: 0.00280 (0.00287)  --> STEP: 55/639 -- GLOBAL_STEP: 5175 | > loss: -0.26254 (-0.27248) | > log_mle: -0.39590 (-0.43475) | > loss_dur: 0.13337 (0.16227) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.88187 (4.15256) | > current_lr: 0.00088 | > step_time: 0.33380 (0.35140) | > loader_time: 0.00270 (0.00281)  --> STEP: 80/639 -- GLOBAL_STEP: 5200 | > loss: -0.25546 (-0.27885) | > log_mle: -0.42046 (-0.43553) | > loss_dur: 0.16499 (0.15667) | > amp_scaler: 4096.00000 (8140.80000) | > grad_norm: 0.00000 (3.81265) | > current_lr: 0.00088 | > step_time: 0.39810 (0.37010) | > loader_time: 0.00290 (0.00284)  --> STEP: 105/639 -- GLOBAL_STEP: 5225 | > loss: -0.26098 (-0.28409) | > log_mle: -0.41407 (-0.43837) | > loss_dur: 0.15309 (0.15428) | > amp_scaler: 4096.00000 (7177.75238) | > grad_norm: 1.98147 (3.82104) | > current_lr: 0.00088 | > step_time: 0.41300 (0.38871) | > loader_time: 0.00330 (0.00286)  --> STEP: 130/639 -- GLOBAL_STEP: 5250 | > loss: -0.30723 (-0.29015) | > log_mle: -0.45907 (-0.44281) | > loss_dur: 0.15184 (0.15265) | > amp_scaler: 4096.00000 (6585.10769) | > grad_norm: 2.45251 (3.98489) | > current_lr: 0.00087 | > step_time: 0.49020 (0.40290) | > loader_time: 0.00300 (0.00289)  --> STEP: 155/639 -- GLOBAL_STEP: 5275 | > loss: -0.31846 (-0.29526) | > log_mle: -0.45312 (-0.44589) | > loss_dur: 0.13466 (0.15063) | > amp_scaler: 4096.00000 (6183.63871) | > grad_norm: 3.99966 (3.90561) | > current_lr: 0.00087 | > step_time: 0.54510 (0.42013) | > loader_time: 0.00320 (0.00292)  --> STEP: 180/639 -- GLOBAL_STEP: 5300 | > loss: -0.34817 (-0.30123) | > log_mle: -0.49230 (-0.45045) | > loss_dur: 0.14413 (0.14922) | > amp_scaler: 4096.00000 (5893.68889) | > grad_norm: 8.12556 (4.14629) | > current_lr: 0.00087 | > step_time: 0.55160 (0.43520) | > loader_time: 0.00310 (0.00294)  --> STEP: 205/639 -- GLOBAL_STEP: 5325 | > loss: -0.31136 (-0.30303) | > log_mle: -0.44143 (-0.45075) | > loss_dur: 0.13007 (0.14772) | > amp_scaler: 4096.00000 (5674.45854) | > grad_norm: 5.15019 (4.12250) | > current_lr: 0.00087 | > step_time: 0.55110 (0.45012) | > loader_time: 0.00320 (0.00298)  --> STEP: 230/639 -- GLOBAL_STEP: 5350 | > loss: -0.32154 (-0.30702) | > log_mle: -0.45179 (-0.45340) | > loss_dur: 0.13024 (0.14638) | > amp_scaler: 4096.00000 (5502.88696) | > grad_norm: 3.27746 (4.16571) | > current_lr: 0.00087 | > step_time: 0.55550 (0.46820) | > loader_time: 0.00350 (0.00302)  --> STEP: 255/639 -- GLOBAL_STEP: 5375 | > loss: -0.30682 (-0.30938) | > log_mle: -0.44855 (-0.45451) | > loss_dur: 0.14173 (0.14513) | > amp_scaler: 4096.00000 (5364.95686) | > grad_norm: 3.90255 (4.14865) | > current_lr: 0.00086 | > step_time: 0.75250 (0.48305) | > loader_time: 0.00330 (0.00305)  --> STEP: 280/639 -- GLOBAL_STEP: 5400 | > loss: -0.33041 (-0.31253) | > log_mle: -0.46788 (-0.45700) | > loss_dur: 0.13746 (0.14447) | > amp_scaler: 4096.00000 (5251.65714) | > grad_norm: 3.63790 (4.27408) | > current_lr: 0.00086 | > step_time: 0.60620 (0.49947) | > loader_time: 0.00340 (0.00309)  --> STEP: 305/639 -- GLOBAL_STEP: 5425 | > loss: -0.32170 (-0.31429) | > log_mle: -0.45791 (-0.45811) | > loss_dur: 0.13621 (0.14382) | > amp_scaler: 4096.00000 (5156.93115) | > grad_norm: 4.51046 (4.23289) | > current_lr: 0.00086 | > step_time: 0.67640 (0.51384) | > loader_time: 0.00340 (0.00312)  --> STEP: 330/639 -- GLOBAL_STEP: 5450 | > loss: -0.33568 (-0.31500) | > log_mle: -0.48815 (-0.45861) | > loss_dur: 0.15247 (0.14361) | > amp_scaler: 4096.00000 (5076.55758) | > grad_norm: 4.23149 (4.20520) | > current_lr: 0.00086 | > step_time: 0.72300 (0.52952) | > loader_time: 0.00360 (0.00315)  --> STEP: 355/639 -- GLOBAL_STEP: 5475 | > loss: -0.30313 (-0.31645) | > log_mle: -0.42477 (-0.45918) | > loss_dur: 0.12164 (0.14274) | > amp_scaler: 4096.00000 (5007.50423) | > grad_norm: 1.79165 (4.20244) | > current_lr: 0.00086 | > step_time: 0.63460 (0.54337) | > loader_time: 0.00340 (0.00319)  --> STEP: 380/639 -- GLOBAL_STEP: 5500 | > loss: -0.33307 (-0.31713) | > log_mle: -0.45462 (-0.45911) | > loss_dur: 0.12154 (0.14199) | > amp_scaler: 4096.00000 (4947.53684) | > grad_norm: 1.63845 (4.08762) | > current_lr: 0.00085 | > step_time: 0.61810 (0.55682) | > loader_time: 0.00330 (0.00321)  --> STEP: 405/639 -- GLOBAL_STEP: 5525 | > loss: -0.33357 (-0.31820) | > log_mle: -0.46299 (-0.45994) | > loss_dur: 0.12942 (0.14173) | > amp_scaler: 4096.00000 (4894.97284) | > grad_norm: 3.89685 (4.09683) | > current_lr: 0.00085 | > step_time: 0.68790 (0.56879) | > loader_time: 0.00350 (0.00324)  --> STEP: 430/639 -- GLOBAL_STEP: 5550 | > loss: -0.29822 (-0.31866) | > log_mle: -0.42346 (-0.46002) | > loss_dur: 0.12524 (0.14135) | > amp_scaler: 4096.00000 (4848.52093) | > grad_norm: 2.72806 (4.10898) | > current_lr: 0.00085 | > step_time: 0.76500 (0.58162) | > loader_time: 0.00350 (0.00326)  --> STEP: 455/639 -- GLOBAL_STEP: 5575 | > loss: -0.32119 (-0.31926) | > log_mle: -0.45038 (-0.45998) | > loss_dur: 0.12919 (0.14072) | > amp_scaler: 4096.00000 (4807.17363) | > grad_norm: 4.24861 (4.02951) | > current_lr: 0.00085 | > step_time: 0.73840 (0.59215) | > loader_time: 0.00360 (0.00328)  --> STEP: 480/639 -- GLOBAL_STEP: 5600 | > loss: -0.28734 (-0.31972) | > log_mle: -0.41548 (-0.46013) | > loss_dur: 0.12814 (0.14042) | > amp_scaler: 4096.00000 (4770.13333) | > grad_norm: 2.17896 (3.98098) | > current_lr: 0.00085 | > step_time: 0.82830 (0.60435) | > loader_time: 0.00390 (0.00331)  --> STEP: 505/639 -- GLOBAL_STEP: 5625 | > loss: -0.35153 (-0.32058) | > log_mle: -0.47870 (-0.46042) | > loss_dur: 0.12717 (0.13984) | > amp_scaler: 4096.00000 (4736.76040) | > grad_norm: 3.28826 (3.96311) | > current_lr: 0.00084 | > step_time: 0.90350 (0.61665) | > loader_time: 0.00400 (0.00333)  --> STEP: 530/639 -- GLOBAL_STEP: 5650 | > loss: -0.33061 (-0.32149) | > log_mle: -0.46911 (-0.46095) | > loss_dur: 0.13850 (0.13946) | > amp_scaler: 4096.00000 (4706.53585) | > grad_norm: 4.30319 (3.96244) | > current_lr: 0.00084 | > step_time: 0.95590 (0.63023) | > loader_time: 0.00410 (0.00336)  --> STEP: 555/639 -- GLOBAL_STEP: 5675 | > loss: -0.33112 (-0.32206) | > log_mle: -0.45766 (-0.46106) | > loss_dur: 0.12654 (0.13900) | > amp_scaler: 4096.00000 (4679.03423) | > grad_norm: 2.56380 (3.92147) | > current_lr: 0.00084 | > step_time: 0.91220 (0.64459) | > loader_time: 0.00410 (0.00340)  --> STEP: 580/639 -- GLOBAL_STEP: 5700 | > loss: -0.35004 (-0.32273) | > log_mle: -0.46821 (-0.46130) | > loss_dur: 0.11817 (0.13857) | > amp_scaler: 4096.00000 (4653.90345) | > grad_norm: 1.56927 (3.91426) | > current_lr: 0.00084 | > step_time: 0.99120 (0.66013) | > loader_time: 0.00440 (0.00343)  --> STEP: 605/639 -- GLOBAL_STEP: 5725 | > loss: -0.32270 (-0.32337) | > log_mle: -0.43925 (-0.46167) | > loss_dur: 0.11655 (0.13829) | > amp_scaler: 4096.00000 (4630.84959) | > grad_norm: 3.66162 (3.94990) | > current_lr: 0.00084 | > step_time: 1.03960 (0.67617) | > loader_time: 0.00440 (0.00347)  --> STEP: 630/639 -- GLOBAL_STEP: 5750 | > loss: -0.34556 (-0.32401) | > log_mle: -0.48223 (-0.46205) | > loss_dur: 0.13667 (0.13804) | > amp_scaler: 4096.00000 (4609.62540) | > grad_norm: 4.06938 (3.93472) | > current_lr: 0.00084 | > step_time: 1.22680 (0.69338) | > loader_time: 0.00460 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.05099 (+0.01422) | > avg_loss: -0.34495 (-0.01828) | > avg_log_mle: -0.47296 (-0.01646) | > avg_loss_dur: 0.12801 (-0.00182) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_5760.pth.tar  > EPOCH: 9/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 15:53:27)   --> STEP: 15/639 -- GLOBAL_STEP: 5775 | > loss: -0.43663 (-0.31858) | > log_mle: -0.59365 (-0.48242) | > loss_dur: 0.15702 (0.16383) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.98157 (5.41308) | > current_lr: 0.00083 | > step_time: 0.34910 (0.30310) | > loader_time: 0.00260 (0.00249)  --> STEP: 40/639 -- GLOBAL_STEP: 5800 | > loss: -0.29882 (-0.30136) | > log_mle: -0.45149 (-0.46294) | > loss_dur: 0.15267 (0.16158) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.49755 (4.31391) | > current_lr: 0.00083 | > step_time: 0.34850 (0.33077) | > loader_time: 0.00250 (0.00262)  --> STEP: 65/639 -- GLOBAL_STEP: 5825 | > loss: -0.32040 (-0.30451) | > log_mle: -0.45855 (-0.46203) | > loss_dur: 0.13815 (0.15753) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.95431 (4.04871) | > current_lr: 0.00083 | > step_time: 0.44970 (0.35940) | > loader_time: 0.00290 (0.00270)  --> STEP: 90/639 -- GLOBAL_STEP: 5850 | > loss: -0.26911 (-0.30640) | > log_mle: -0.42183 (-0.45970) | > loss_dur: 0.15272 (0.15330) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.54727 (3.69373) | > current_lr: 0.00083 | > step_time: 0.52410 (0.37686) | > loader_time: 0.00310 (0.00276)  --> STEP: 115/639 -- GLOBAL_STEP: 5875 | > loss: -0.24111 (-0.31258) | > log_mle: -0.39578 (-0.46389) | > loss_dur: 0.15468 (0.15131) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.75645 (3.91039) | > current_lr: 0.00083 | > step_time: 0.38770 (0.39534) | > loader_time: 0.00290 (0.00278)  --> STEP: 140/639 -- GLOBAL_STEP: 5900 | > loss: -0.31389 (-0.31896) | > log_mle: -0.47854 (-0.46780) | > loss_dur: 0.16465 (0.14884) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.46237 (4.11112) | > current_lr: 0.00082 | > step_time: 0.63170 (0.40962) | > loader_time: 0.00310 (0.00281)  --> STEP: 165/639 -- GLOBAL_STEP: 5925 | > loss: -0.32417 (-0.32276) | > log_mle: -0.48610 (-0.46979) | > loss_dur: 0.16193 (0.14703) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.96481 (4.15705) | > current_lr: 0.00082 | > step_time: 0.57870 (0.42411) | > loader_time: 0.00300 (0.00284)  --> STEP: 190/639 -- GLOBAL_STEP: 5950 | > loss: -0.27857 (-0.32591) | > log_mle: -0.42044 (-0.47156) | > loss_dur: 0.14187 (0.14565) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.75306 (4.15077) | > current_lr: 0.00082 | > step_time: 0.52290 (0.44120) | > loader_time: 0.00270 (0.00287)  --> STEP: 215/639 -- GLOBAL_STEP: 5975 | > loss: -0.39084 (-0.32746) | > log_mle: -0.52098 (-0.47201) | > loss_dur: 0.13014 (0.14455) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.48563 (4.21327) | > current_lr: 0.00082 | > step_time: 0.55030 (0.45605) | > loader_time: 0.00330 (0.00291)  --> STEP: 240/639 -- GLOBAL_STEP: 6000 | > loss: -0.39873 (-0.33106) | > log_mle: -0.52608 (-0.47403) | > loss_dur: 0.12735 (0.14297) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.44645 (4.23583) | > current_lr: 0.00082 | > step_time: 0.64790 (0.47180) | > loader_time: 0.00320 (0.00295)  --> STEP: 265/639 -- GLOBAL_STEP: 6025 | > loss: -0.40101 (-0.33309) | > log_mle: -0.54071 (-0.47528) | > loss_dur: 0.13971 (0.14218) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 19.42059 (4.35776) | > current_lr: 0.00082 | > step_time: 0.64270 (0.48676) | > loader_time: 0.00320 (0.00299)  --> STEP: 290/639 -- GLOBAL_STEP: 6050 | > loss: -0.34063 (-0.33410) | > log_mle: -0.47155 (-0.47579) | > loss_dur: 0.13093 (0.14169) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.52499 (4.32980) | > current_lr: 0.00081 | > step_time: 0.63090 (0.50254) | > loader_time: 0.00340 (0.00303)  --> STEP: 315/639 -- GLOBAL_STEP: 6075 | > loss: -0.32520 (-0.33534) | > log_mle: -0.44388 (-0.47628) | > loss_dur: 0.11868 (0.14094) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.00772 (4.34945) | > current_lr: 0.00081 | > step_time: 0.62020 (0.51745) | > loader_time: 0.00350 (0.00306)  --> STEP: 340/639 -- GLOBAL_STEP: 6100 | > loss: -0.35260 (-0.33612) | > log_mle: -0.47337 (-0.47690) | > loss_dur: 0.12077 (0.14077) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.08297 (4.38925) | > current_lr: 0.00081 | > step_time: 0.63980 (0.53320) | > loader_time: 0.00340 (0.00310)  --> STEP: 365/639 -- GLOBAL_STEP: 6125 | > loss: -0.35429 (-0.33655) | > log_mle: -0.47882 (-0.47646) | > loss_dur: 0.12453 (0.13991) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.92295 (4.35980) | > current_lr: 0.00081 | > step_time: 0.80990 (0.54511) | > loader_time: 0.00390 (0.00312)  --> STEP: 390/639 -- GLOBAL_STEP: 6150 | > loss: -0.34204 (-0.33705) | > log_mle: -0.47837 (-0.47638) | > loss_dur: 0.13633 (0.13933) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.12306 (4.32504) | > current_lr: 0.00081 | > step_time: 0.72050 (0.55802) | > loader_time: 0.00330 (0.00315)  --> STEP: 415/639 -- GLOBAL_STEP: 6175 | > loss: -0.32908 (-0.33743) | > log_mle: -0.45895 (-0.47657) | > loss_dur: 0.12986 (0.13913) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.58740 (4.27417) | > current_lr: 0.00081 | > step_time: 0.76220 (0.57035) | > loader_time: 0.00360 (0.00318)  --> STEP: 440/639 -- GLOBAL_STEP: 6200 | > loss: -0.30769 (-0.33795) | > log_mle: -0.44354 (-0.47658) | > loss_dur: 0.13585 (0.13863) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.10928 (4.27577) | > current_lr: 0.00080 | > step_time: 0.72060 (0.58217) | > loader_time: 0.00350 (0.00321)  --> STEP: 465/639 -- GLOBAL_STEP: 6225 | > loss: -0.32194 (-0.33803) | > log_mle: -0.44994 (-0.47618) | > loss_dur: 0.12800 (0.13816) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.03221 (4.20508) | > current_lr: 0.00080 | > step_time: 0.88200 (0.59364) | > loader_time: 0.00380 (0.00323)  --> STEP: 490/639 -- GLOBAL_STEP: 6250 | > loss: -0.32404 (-0.33837) | > log_mle: -0.43953 (-0.47616) | > loss_dur: 0.11549 (0.13780) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.97884 (4.19835) | > current_lr: 0.00080 | > step_time: 0.78480 (0.60715) | > loader_time: 0.00330 (0.00326)  --> STEP: 515/639 -- GLOBAL_STEP: 6275 | > loss: -0.34096 (-0.33922) | > log_mle: -0.46736 (-0.47651) | > loss_dur: 0.12640 (0.13729) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.89702 (4.19597) | > current_lr: 0.00080 | > step_time: 1.08470 (0.62084) | > loader_time: 0.00410 (0.00329)  --> STEP: 540/639 -- GLOBAL_STEP: 6300 | > loss: -0.35047 (-0.33974) | > log_mle: -0.48117 (-0.47664) | > loss_dur: 0.13069 (0.13690) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.60601 (4.15380) | > current_lr: 0.00080 | > step_time: 0.91430 (0.63449) | > loader_time: 0.00410 (0.00332)  --> STEP: 565/639 -- GLOBAL_STEP: 6325 | > loss: -0.31880 (-0.33991) | > log_mle: -0.45015 (-0.47651) | > loss_dur: 0.13135 (0.13660) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.86436 (4.14727) | > current_lr: 0.00080 | > step_time: 1.07980 (0.64819) | > loader_time: 0.00430 (0.00335)  --> STEP: 590/639 -- GLOBAL_STEP: 6350 | > loss: -0.35100 (-0.34063) | > log_mle: -0.48230 (-0.47690) | > loss_dur: 0.13131 (0.13627) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.80523 (4.13197) | > current_lr: 0.00079 | > step_time: 0.96620 (0.66396) | > loader_time: 0.00410 (0.00339)  --> STEP: 615/639 -- GLOBAL_STEP: 6375 | > loss: -0.36284 (-0.34098) | > log_mle: -0.48647 (-0.47693) | > loss_dur: 0.12363 (0.13596) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.34291 (4.13950) | > current_lr: 0.00079 | > step_time: 1.01800 (0.68013) | > loader_time: 0.00440 (0.00343) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.01933 (-0.03166) | > avg_loss: -0.35735 (-0.01240) | > avg_log_mle: -0.48571 (-0.01274) | > avg_loss_dur: 0.12836 (+0.00034) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_6400.pth.tar  > EPOCH: 10/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 16:01:13)   --> STEP: 0/639 -- GLOBAL_STEP: 6400 | > loss: -0.36773 (-0.36773) | > log_mle: -0.54928 (-0.54928) | > loss_dur: 0.18155 (0.18155) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.42890 (12.42890) | > current_lr: 0.00079 | > step_time: 0.78340 (0.78341) | > loader_time: 0.31840 (0.31845)  --> STEP: 25/639 -- GLOBAL_STEP: 6425 | > loss: -0.35114 (-0.32116) | > log_mle: -0.49495 (-0.48304) | > loss_dur: 0.14381 (0.16188) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.15880 (4.73090) | > current_lr: 0.00079 | > step_time: 0.38750 (0.31807) | > loader_time: 0.00300 (0.00255)  --> STEP: 50/639 -- GLOBAL_STEP: 6450 | > loss: -0.30959 (-0.31735) | > log_mle: -0.43850 (-0.47390) | > loss_dur: 0.12891 (0.15655) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.73203 (4.55134) | > current_lr: 0.00079 | > step_time: 0.35640 (0.34379) | > loader_time: 0.00260 (0.00264)  --> STEP: 75/639 -- GLOBAL_STEP: 6475 | > loss: -0.34602 (-0.32047) | > log_mle: -0.48144 (-0.47358) | > loss_dur: 0.13543 (0.15311) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.65953 (4.28017) | > current_lr: 0.00079 | > step_time: 0.42530 (0.36615) | > loader_time: 0.00280 (0.00275)  --> STEP: 100/639 -- GLOBAL_STEP: 6500 | > loss: -0.36821 (-0.32625) | > log_mle: -0.49325 (-0.47545) | > loss_dur: 0.12504 (0.14920) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.10821 (4.23484) | > current_lr: 0.00079 | > step_time: 0.51560 (0.38299) | > loader_time: 0.00320 (0.00282)  --> STEP: 125/639 -- GLOBAL_STEP: 6525 | > loss: -0.36099 (-0.32974) | > log_mle: -0.49048 (-0.47700) | > loss_dur: 0.12949 (0.14726) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.60555 (4.23624) | > current_lr: 0.00078 | > step_time: 0.40510 (0.39868) | > loader_time: 0.00310 (0.00287)  --> STEP: 150/639 -- GLOBAL_STEP: 6550 | > loss: -0.31298 (-0.33385) | > log_mle: -0.45928 (-0.47936) | > loss_dur: 0.14629 (0.14551) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.81147 (4.43535) | > current_lr: 0.00078 | > step_time: 0.43560 (0.41552) | > loader_time: 0.00310 (0.00290)  --> STEP: 175/639 -- GLOBAL_STEP: 6575 | > loss: -0.36309 (-0.33838) | > log_mle: -0.48068 (-0.48245) | > loss_dur: 0.11759 (0.14407) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.53146 (4.54147) | > current_lr: 0.00078 | > step_time: 0.56310 (0.43125) | > loader_time: 0.00320 (0.00294)  --> STEP: 200/639 -- GLOBAL_STEP: 6600 | > loss: -0.33952 (-0.34029) | > log_mle: -0.47295 (-0.48282) | > loss_dur: 0.13343 (0.14252) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.04528 (4.54669) | > current_lr: 0.00078 | > step_time: 0.52420 (0.44566) | > loader_time: 0.00320 (0.00297)  --> STEP: 225/639 -- GLOBAL_STEP: 6625 | > loss: -0.33281 (-0.34206) | > log_mle: -0.45187 (-0.48334) | > loss_dur: 0.11906 (0.14128) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.37424 (4.54669) | > current_lr: 0.00078 | > step_time: 0.52080 (0.46199) | > loader_time: 0.00280 (0.00300)  --> STEP: 250/639 -- GLOBAL_STEP: 6650 | > loss: -0.35110 (-0.34575) | > log_mle: -0.45815 (-0.48604) | > loss_dur: 0.10705 (0.14029) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.00051 (4.67169) | > current_lr: 0.00078 | > step_time: 0.58050 (0.47655) | > loader_time: 0.00330 (0.00304)  --> STEP: 275/639 -- GLOBAL_STEP: 6675 | > loss: -0.36273 (-0.34827) | > log_mle: -0.50028 (-0.48797) | > loss_dur: 0.13754 (0.13970) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.98372 (4.85578) | > current_lr: 0.00077 | > step_time: 0.85920 (0.49458) | > loader_time: 0.00380 (0.00309)  --> STEP: 300/639 -- GLOBAL_STEP: 6700 | > loss: -0.36727 (-0.34891) | > log_mle: -0.49243 (-0.48820) | > loss_dur: 0.12516 (0.13929) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.79725 (4.86756) | > current_lr: 0.00077 | > step_time: 0.74950 (0.50826) | > loader_time: 0.00320 (0.00311)  --> STEP: 325/639 -- GLOBAL_STEP: 6725 | > loss: -0.36953 (-0.34973) | > log_mle: -0.49836 (-0.48878) | > loss_dur: 0.12883 (0.13905) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.86185 (4.93538) | > current_lr: 0.00077 | > step_time: 0.61840 (0.52369) | > loader_time: 0.00330 (0.00315)  --> STEP: 350/639 -- GLOBAL_STEP: 6750 | > loss: -0.39591 (-0.35032) | > log_mle: -0.52777 (-0.48879) | > loss_dur: 0.13186 (0.13846) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.37595 (4.89697) | > current_lr: 0.00077 | > step_time: 0.81020 (0.53760) | > loader_time: 0.00380 (0.00318)  --> STEP: 375/639 -- GLOBAL_STEP: 6775 | > loss: -0.33474 (-0.35018) | > log_mle: -0.46489 (-0.48808) | > loss_dur: 0.13015 (0.13790) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.31127 (4.86461) | > current_lr: 0.00077 | > step_time: 0.69390 (0.55001) | > loader_time: 0.00340 (0.00321)  --> STEP: 400/639 -- GLOBAL_STEP: 6800 | > loss: -0.36593 (-0.35079) | > log_mle: -0.49108 (-0.48840) | > loss_dur: 0.12515 (0.13761) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.83060 (4.85136) | > current_lr: 0.00077 | > step_time: 0.74180 (0.56192) | > loader_time: 0.00370 (0.00324)  --> STEP: 425/639 -- GLOBAL_STEP: 6825 | > loss: -0.41655 (-0.35076) | > log_mle: -0.54849 (-0.48800) | > loss_dur: 0.13194 (0.13725) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.94025 (4.78814) | > current_lr: 0.00077 | > step_time: 0.89690 (0.57519) | > loader_time: 0.00390 (0.00327)  --> STEP: 450/639 -- GLOBAL_STEP: 6850 | > loss: -0.33597 (-0.35099) | > log_mle: -0.45969 (-0.48772) | > loss_dur: 0.12373 (0.13673) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.12553 (4.75239) | > current_lr: 0.00076 | > step_time: 0.88770 (0.58753) | > loader_time: 0.00390 (0.00330)  --> STEP: 475/639 -- GLOBAL_STEP: 6875 | > loss: -0.35220 (-0.35102) | > log_mle: -0.46744 (-0.48748) | > loss_dur: 0.11524 (0.13646) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.82678 (4.73794) | > current_lr: 0.00076 | > step_time: 0.73060 (0.59943) | > loader_time: 0.00370 (0.00332)  --> STEP: 500/639 -- GLOBAL_STEP: 6900 | > loss: -0.40234 (-0.35150) | > log_mle: -0.53040 (-0.48758) | > loss_dur: 0.12806 (0.13608) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.79483 (4.74383) | > current_lr: 0.00076 | > step_time: 0.97100 (0.61236) | > loader_time: 0.00400 (0.00335)  --> STEP: 525/639 -- GLOBAL_STEP: 6925 | > loss: -0.37311 (-0.35201) | > log_mle: -0.51517 (-0.48773) | > loss_dur: 0.14206 (0.13573) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.36707 (4.70375) | > current_lr: 0.00076 | > step_time: 1.03370 (0.62658) | > loader_time: 0.00440 (0.00338)  --> STEP: 550/639 -- GLOBAL_STEP: 6950 | > loss: -0.32020 (-0.35216) | > log_mle: -0.45448 (-0.48753) | > loss_dur: 0.13427 (0.13537) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.54600 (4.64790) | > current_lr: 0.00076 | > step_time: 0.91500 (0.64081) | > loader_time: 0.00420 (0.00341)  --> STEP: 575/639 -- GLOBAL_STEP: 6975 | > loss: -0.34389 (-0.35233) | > log_mle: -0.46655 (-0.48738) | > loss_dur: 0.12266 (0.13504) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.17516 (4.59993) | > current_lr: 0.00076 | > step_time: 0.88030 (0.65490) | > loader_time: 0.00400 (0.00344)  --> STEP: 600/639 -- GLOBAL_STEP: 7000 | > loss: -0.33210 (-0.35271) | > log_mle: -0.46461 (-0.48750) | > loss_dur: 0.13251 (0.13478) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.60124 (4.56471) | > current_lr: 0.00076 | > step_time: 1.03550 (0.67116) | > loader_time: 0.00430 (0.00348)  --> STEP: 625/639 -- GLOBAL_STEP: 7025 | > loss: -0.36891 (-0.35300) | > log_mle: -0.49608 (-0.48751) | > loss_dur: 0.12717 (0.13451) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.03611 (4.52966) | > current_lr: 0.00076 | > step_time: 1.05900 (0.68796) | > loader_time: 0.00440 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03083 (+0.01150) | > avg_loss: -0.37034 (-0.01299) | > avg_log_mle: -0.49889 (-0.01319) | > avg_loss_dur: 0.12855 (+0.00019) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_7040.pth.tar  > EPOCH: 11/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 16:08:59)   --> STEP: 10/639 -- GLOBAL_STEP: 7050 | > loss: -0.34254 (-0.34188) | > log_mle: -0.51194 (-0.50374) | > loss_dur: 0.16940 (0.16187) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.45634 (5.39127) | > current_lr: 0.00075 | > step_time: 0.28340 (0.28830) | > loader_time: 0.00280 (0.00259)  --> STEP: 35/639 -- GLOBAL_STEP: 7075 | > loss: -0.34025 (-0.33217) | > log_mle: -0.47116 (-0.48647) | > loss_dur: 0.13091 (0.15430) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.61886 (4.83959) | > current_lr: 0.00075 | > step_time: 0.37920 (0.32681) | > loader_time: 0.00240 (0.00268)  --> STEP: 60/639 -- GLOBAL_STEP: 7100 | > loss: -0.31745 (-0.33348) | > log_mle: -0.44880 (-0.48531) | > loss_dur: 0.13135 (0.15183) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.85616 (5.15045) | > current_lr: 0.00075 | > step_time: 0.37590 (0.35811) | > loader_time: 0.00240 (0.00275)  --> STEP: 85/639 -- GLOBAL_STEP: 7125 | > loss: -0.28484 (-0.33524) | > log_mle: -0.44557 (-0.48355) | > loss_dur: 0.16073 (0.14830) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.97330 (5.06813) | > current_lr: 0.00075 | > step_time: 0.41990 (0.37188) | > loader_time: 0.00280 (0.00279)  --> STEP: 110/639 -- GLOBAL_STEP: 7150 | > loss: -0.32934 (-0.34288) | > log_mle: -0.47461 (-0.48949) | > loss_dur: 0.14528 (0.14661) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.52964 (5.23233) | > current_lr: 0.00075 | > step_time: 0.39690 (0.39343) | > loader_time: 0.00270 (0.00283)  --> STEP: 135/639 -- GLOBAL_STEP: 7175 | > loss: -0.30599 (-0.34624) | > log_mle: -0.45823 (-0.49068) | > loss_dur: 0.15224 (0.14443) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.00188 (5.06759) | > current_lr: 0.00075 | > step_time: 0.46710 (0.40713) | > loader_time: 0.00300 (0.00284)  --> STEP: 160/639 -- GLOBAL_STEP: 7200 | > loss: -0.38769 (-0.34906) | > log_mle: -0.50709 (-0.49168) | > loss_dur: 0.11940 (0.14262) | > amp_scaler: 8192.00000 (4121.60000) | > grad_norm: 4.84207 (5.03801) | > current_lr: 0.00075 | > step_time: 0.46300 (0.42127) | > loader_time: 0.00290 (0.00286)  --> STEP: 185/639 -- GLOBAL_STEP: 7225 | > loss: -0.36065 (-0.35427) | > log_mle: -0.48909 (-0.49622) | > loss_dur: 0.12843 (0.14195) | > amp_scaler: 4096.00000 (4383.82703) | > grad_norm: 3.74872 (5.14293) | > current_lr: 0.00074 | > step_time: 0.57390 (0.43782) | > loader_time: 0.00320 (0.00291)  --> STEP: 210/639 -- GLOBAL_STEP: 7250 | > loss: -0.29038 (-0.35451) | > log_mle: -0.41744 (-0.49555) | > loss_dur: 0.12706 (0.14104) | > amp_scaler: 4096.00000 (4349.56190) | > grad_norm: 0.89798 (5.00155) | > current_lr: 0.00074 | > step_time: 0.51340 (0.45288) | > loader_time: 0.00290 (0.00295)  --> STEP: 235/639 -- GLOBAL_STEP: 7275 | > loss: -0.37885 (-0.35768) | > log_mle: -0.50506 (-0.49739) | > loss_dur: 0.12622 (0.13971) | > amp_scaler: 4096.00000 (4322.58723) | > grad_norm: 4.47243 (4.98092) | > current_lr: 0.00074 | > step_time: 0.56170 (0.46812) | > loader_time: 0.00340 (0.00299)  --> STEP: 260/639 -- GLOBAL_STEP: 7300 | > loss: -0.38183 (-0.36008) | > log_mle: -0.50601 (-0.49891) | > loss_dur: 0.12418 (0.13884) | > amp_scaler: 4096.00000 (4300.80000) | > grad_norm: 5.51702 (5.03642) | > current_lr: 0.00074 | > step_time: 0.58140 (0.48274) | > loader_time: 0.00350 (0.00303)  --> STEP: 285/639 -- GLOBAL_STEP: 7325 | > loss: -0.31947 (-0.36047) | > log_mle: -0.45979 (-0.49890) | > loss_dur: 0.14031 (0.13843) | > amp_scaler: 4096.00000 (4282.83509) | > grad_norm: 1.18482 (5.06259) | > current_lr: 0.00074 | > step_time: 0.62140 (0.49862) | > loader_time: 0.00300 (0.00307)  --> STEP: 310/639 -- GLOBAL_STEP: 7350 | > loss: -0.36884 (-0.36120) | > log_mle: -0.50101 (-0.49915) | > loss_dur: 0.13217 (0.13794) | > amp_scaler: 4096.00000 (4267.76774) | > grad_norm: 4.44474 (5.12339) | > current_lr: 0.00074 | > step_time: 0.57530 (0.51254) | > loader_time: 0.00340 (0.00310)  --> STEP: 335/639 -- GLOBAL_STEP: 7375 | > loss: -0.39856 (-0.36121) | > log_mle: -0.52725 (-0.49896) | > loss_dur: 0.12868 (0.13775) | > amp_scaler: 4096.00000 (4254.94925) | > grad_norm: 9.33142 (5.10511) | > current_lr: 0.00074 | > step_time: 0.64270 (0.52804) | > loader_time: 0.00370 (0.00314)  --> STEP: 360/639 -- GLOBAL_STEP: 7400 | > loss: -0.33990 (-0.36187) | > log_mle: -0.46715 (-0.49883) | > loss_dur: 0.12725 (0.13697) | > amp_scaler: 4096.00000 (4243.91111) | > grad_norm: 3.25773 (5.05540) | > current_lr: 0.00074 | > step_time: 0.70800 (0.54079) | > loader_time: 0.00360 (0.00316)  --> STEP: 385/639 -- GLOBAL_STEP: 7425 | > loss: -0.37083 (-0.36210) | > log_mle: -0.50145 (-0.49852) | > loss_dur: 0.13062 (0.13641) | > amp_scaler: 4096.00000 (4234.30649) | > grad_norm: 3.80741 (4.98702) | > current_lr: 0.00073 | > step_time: 0.66970 (0.55479) | > loader_time: 0.00370 (0.00319)  --> STEP: 410/639 -- GLOBAL_STEP: 7450 | > loss: -0.33063 (-0.36226) | > log_mle: -0.48164 (-0.49846) | > loss_dur: 0.15101 (0.13619) | > amp_scaler: 4096.00000 (4225.87317) | > grad_norm: 4.60133 (5.02678) | > current_lr: 0.00073 | > step_time: 0.82980 (0.56680) | > loader_time: 0.00360 (0.00321)  --> STEP: 435/639 -- GLOBAL_STEP: 7475 | > loss: -0.36746 (-0.36228) | > log_mle: -0.48998 (-0.49806) | > loss_dur: 0.12252 (0.13578) | > amp_scaler: 4096.00000 (4218.40920) | > grad_norm: 3.33436 (5.03771) | > current_lr: 0.00073 | > step_time: 0.75970 (0.57916) | > loader_time: 0.00340 (0.00324)  --> STEP: 460/639 -- GLOBAL_STEP: 7500 | > loss: -0.32180 (-0.36212) | > log_mle: -0.46026 (-0.49751) | > loss_dur: 0.13846 (0.13539) | > amp_scaler: 4096.00000 (4211.75652) | > grad_norm: 2.05614 (4.95843) | > current_lr: 0.00073 | > step_time: 0.87490 (0.59138) | > loader_time: 0.00390 (0.00326)  --> STEP: 485/639 -- GLOBAL_STEP: 7525 | > loss: -0.36977 (-0.36217) | > log_mle: -0.49136 (-0.49736) | > loss_dur: 0.12160 (0.13519) | > amp_scaler: 4096.00000 (4205.78969) | > grad_norm: 3.28140 (4.94946) | > current_lr: 0.00073 | > step_time: 0.84590 (0.60448) | > loader_time: 0.00400 (0.00329)  --> STEP: 510/639 -- GLOBAL_STEP: 7550 | > loss: -0.37591 (-0.36256) | > log_mle: -0.50207 (-0.49729) | > loss_dur: 0.12616 (0.13473) | > amp_scaler: 4096.00000 (4200.40784) | > grad_norm: 6.71116 (4.91577) | > current_lr: 0.00073 | > step_time: 0.87080 (0.61703) | > loader_time: 0.00400 (0.00332)  --> STEP: 535/639 -- GLOBAL_STEP: 7575 | > loss: -0.35691 (-0.36282) | > log_mle: -0.49225 (-0.49722) | > loss_dur: 0.13535 (0.13440) | > amp_scaler: 4096.00000 (4195.52897) | > grad_norm: 2.73451 (4.90034) | > current_lr: 0.00073 | > step_time: 0.93540 (0.63057) | > loader_time: 0.00400 (0.00335)  --> STEP: 560/639 -- GLOBAL_STEP: 7600 | > loss: -0.36623 (-0.36292) | > log_mle: -0.49854 (-0.49706) | > loss_dur: 0.13231 (0.13414) | > amp_scaler: 4096.00000 (4191.08571) | > grad_norm: 5.07412 (4.87482) | > current_lr: 0.00073 | > step_time: 0.99820 (0.64580) | > loader_time: 0.00400 (0.00338)  --> STEP: 585/639 -- GLOBAL_STEP: 7625 | > loss: -0.37329 (-0.36324) | > log_mle: -0.49909 (-0.49703) | > loss_dur: 0.12580 (0.13379) | > amp_scaler: 4096.00000 (4187.02222) | > grad_norm: 5.72896 (4.83655) | > current_lr: 0.00072 | > step_time: 1.00820 (0.66090) | > loader_time: 0.00440 (0.00342)  --> STEP: 610/639 -- GLOBAL_STEP: 7650 | > loss: -0.34567 (-0.36347) | > log_mle: -0.47209 (-0.49702) | > loss_dur: 0.12642 (0.13355) | > amp_scaler: 4096.00000 (4183.29180) | > grad_norm: 3.84963 (4.82169) | > current_lr: 0.00072 | > step_time: 1.07980 (0.67713) | > loader_time: 0.00440 (0.00346)  --> STEP: 635/639 -- GLOBAL_STEP: 7675 | > loss: -0.36800 (-0.36372) | > log_mle: -0.49772 (-0.49720) | > loss_dur: 0.12972 (0.13348) | > amp_scaler: 4096.00000 (4179.85512) | > grad_norm: 5.31920 (4.80167) | > current_lr: 0.00072 | > step_time: 1.10780 (0.69430) | > loader_time: 0.00420 (0.00350) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02987 (-0.00096) | > avg_loss: -0.37258 (-0.00224) | > avg_log_mle: -0.50374 (-0.00484) | > avg_loss_dur: 0.13115 (+0.00260) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_7680.pth.tar  > EPOCH: 12/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 16:16:45)   --> STEP: 20/639 -- GLOBAL_STEP: 7700 | > loss: -0.44526 (-0.36607) | > log_mle: -0.61168 (-0.52217) | > loss_dur: 0.16642 (0.15610) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 22.05260 (6.33489) | > current_lr: 0.00072 | > step_time: 0.34340 (0.30332) | > loader_time: 0.00290 (0.00266)  --> STEP: 45/639 -- GLOBAL_STEP: 7725 | > loss: -0.30111 (-0.34830) | > log_mle: -0.44269 (-0.50001) | > loss_dur: 0.14158 (0.15171) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.04793 (5.37434) | > current_lr: 0.00072 | > step_time: 0.40670 (0.34071) | > loader_time: 0.00310 (0.00280)  --> STEP: 70/639 -- GLOBAL_STEP: 7750 | > loss: -0.39003 (-0.34808) | > log_mle: -0.52584 (-0.49769) | > loss_dur: 0.13581 (0.14961) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.66741 (5.34594) | > current_lr: 0.00072 | > step_time: 0.38080 (0.36870) | > loader_time: 0.00280 (0.00283)  --> STEP: 95/639 -- GLOBAL_STEP: 7775 | > loss: -0.44078 (-0.35190) | > log_mle: -0.56672 (-0.49733) | > loss_dur: 0.12593 (0.14543) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.67276 (5.13906) | > current_lr: 0.00072 | > step_time: 0.40480 (0.38131) | > loader_time: 0.00290 (0.00287)  --> STEP: 120/639 -- GLOBAL_STEP: 7800 | > loss: -0.42229 (-0.35502) | > log_mle: -0.56931 (-0.49983) | > loss_dur: 0.14701 (0.14482) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.45719 (5.36014) | > current_lr: 0.00072 | > step_time: 0.63490 (0.39923) | > loader_time: 0.00300 (0.00290)  --> STEP: 145/639 -- GLOBAL_STEP: 7825 | > loss: -0.39064 (-0.36034) | > log_mle: -0.52844 (-0.50317) | > loss_dur: 0.13780 (0.14283) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.83248 (5.44102) | > current_lr: 0.00072 | > step_time: 0.48770 (0.41340) | > loader_time: 0.00310 (0.00293)  --> STEP: 170/639 -- GLOBAL_STEP: 7850 | > loss: -0.39709 (-0.36302) | > log_mle: -0.52038 (-0.50450) | > loss_dur: 0.12329 (0.14148) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.14687 (5.43151) | > current_lr: 0.00071 | > step_time: 0.54740 (0.42790) | > loader_time: 0.00320 (0.00295)  --> STEP: 195/639 -- GLOBAL_STEP: 7875 | > loss: -0.34902 (-0.36510) | > log_mle: -0.46003 (-0.50541) | > loss_dur: 0.11101 (0.14031) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.39109 (5.41193) | > current_lr: 0.00071 | > step_time: 0.49100 (0.44338) | > loader_time: 0.00330 (0.00299)  --> STEP: 220/639 -- GLOBAL_STEP: 7900 | > loss: -0.39721 (-0.36586) | > log_mle: -0.51141 (-0.50519) | > loss_dur: 0.11420 (0.13933) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.53078 (5.30811) | > current_lr: 0.00071 | > step_time: 0.63780 (0.45925) | > loader_time: 0.00320 (0.00303)  --> STEP: 245/639 -- GLOBAL_STEP: 7925 | > loss: -0.39585 (-0.36957) | > log_mle: -0.50645 (-0.50768) | > loss_dur: 0.11060 (0.13811) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.18086 (5.40731) | > current_lr: 0.00071 | > step_time: 0.53610 (0.47334) | > loader_time: 0.00310 (0.00306)  --> STEP: 270/639 -- GLOBAL_STEP: 7950 | > loss: -0.42325 (-0.37204) | > log_mle: -0.56019 (-0.50954) | > loss_dur: 0.13694 (0.13750) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.03929 (5.52220) | > current_lr: 0.00071 | > step_time: 0.84120 (0.49012) | > loader_time: 0.00400 (0.00309)  --> STEP: 295/639 -- GLOBAL_STEP: 7975 | > loss: -0.34286 (-0.37202) | > log_mle: -0.46672 (-0.50905) | > loss_dur: 0.12386 (0.13703) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.90485 (5.44958) | > current_lr: 0.00071 | > step_time: 0.64010 (0.50468) | > loader_time: 0.00360 (0.00313)  --> STEP: 320/639 -- GLOBAL_STEP: 8000 | > loss: -0.37919 (-0.37255) | > log_mle: -0.52115 (-0.50923) | > loss_dur: 0.14196 (0.13669) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.07125 (5.41900) | > current_lr: 0.00071 | > step_time: 0.73750 (0.51996) | > loader_time: 0.00370 (0.00316)  --> STEP: 345/639 -- GLOBAL_STEP: 8025 | > loss: -0.38168 (-0.37261) | > log_mle: -0.50394 (-0.50907) | > loss_dur: 0.12226 (0.13646) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.87259 (5.44631) | > current_lr: 0.00071 | > step_time: 0.65250 (0.53449) | > loader_time: 0.00320 (0.00319)  --> STEP: 370/639 -- GLOBAL_STEP: 8050 | > loss: -0.40312 (-0.37276) | > log_mle: -0.52474 (-0.50843) | > loss_dur: 0.12162 (0.13566) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.66655 (5.33548) | > current_lr: 0.00071 | > step_time: 0.85730 (0.54647) | > loader_time: 0.00380 (0.00322)  --> STEP: 395/639 -- GLOBAL_STEP: 8075 | > loss: -0.35580 (-0.37275) | > log_mle: -0.49577 (-0.50793) | > loss_dur: 0.13996 (0.13518) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.22566 (5.27884) | > current_lr: 0.00070 | > step_time: 0.69750 (0.55895) | > loader_time: 0.00360 (0.00324)  --> STEP: 420/639 -- GLOBAL_STEP: 8100 | > loss: -0.36451 (-0.37251) | > log_mle: -0.49019 (-0.50750) | > loss_dur: 0.12569 (0.13499) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.70529 (5.23362) | > current_lr: 0.00070 | > step_time: 0.69740 (0.57171) | > loader_time: 0.00370 (0.00327)  --> STEP: 445/639 -- GLOBAL_STEP: 8125 | > loss: -0.37621 (-0.37269) | > log_mle: -0.49581 (-0.50721) | > loss_dur: 0.11959 (0.13452) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.22646 (5.22380) | > current_lr: 0.00070 | > step_time: 0.75160 (0.58395) | > loader_time: 0.00360 (0.00329)  --> STEP: 470/639 -- GLOBAL_STEP: 8150 | > loss: -0.37018 (-0.37235) | > log_mle: -0.50079 (-0.50656) | > loss_dur: 0.13061 (0.13421) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.78853 (5.16442) | > current_lr: 0.00070 | > step_time: 0.77370 (0.59572) | > loader_time: 0.00380 (0.00332)  --> STEP: 495/639 -- GLOBAL_STEP: 8175 | > loss: -0.32893 (-0.37227) | > log_mle: -0.44798 (-0.50621) | > loss_dur: 0.11905 (0.13393) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.79160 (5.12190) | > current_lr: 0.00070 | > step_time: 0.76510 (0.60893) | > loader_time: 0.00360 (0.00335)  --> STEP: 520/639 -- GLOBAL_STEP: 8200 | > loss: -0.38851 (-0.37272) | > log_mle: -0.50683 (-0.50620) | > loss_dur: 0.11832 (0.13348) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.50451 (5.09912) | > current_lr: 0.00070 | > step_time: 0.81450 (0.62158) | > loader_time: 0.00390 (0.00338)  --> STEP: 545/639 -- GLOBAL_STEP: 8225 | > loss: -0.40042 (-0.37286) | > log_mle: -0.53424 (-0.50614) | > loss_dur: 0.13382 (0.13328) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.84819 (5.07984) | > current_lr: 0.00070 | > step_time: 0.93880 (0.63683) | > loader_time: 0.00400 (0.00342)  --> STEP: 570/639 -- GLOBAL_STEP: 8250 | > loss: -0.36837 (-0.37274) | > log_mle: -0.48549 (-0.50574) | > loss_dur: 0.11711 (0.13300) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.55927 (5.14126) | > current_lr: 0.00070 | > step_time: 0.82190 (0.65040) | > loader_time: 0.00400 (0.00345)  --> STEP: 595/639 -- GLOBAL_STEP: 8275 | > loss: -0.35642 (-0.37304) | > log_mle: -0.47860 (-0.50578) | > loss_dur: 0.12218 (0.13274) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.46489 (5.16138) | > current_lr: 0.00070 | > step_time: 0.97780 (0.66621) | > loader_time: 0.00430 (0.00349)  --> STEP: 620/639 -- GLOBAL_STEP: 8300 | > loss: -0.36935 (-0.37310) | > log_mle: -0.49110 (-0.50558) | > loss_dur: 0.12175 (0.13249) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.58615 (5.18044) | > current_lr: 0.00069 | > step_time: 1.14590 (0.68230) | > loader_time: 0.00470 (0.00353) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03147 (+0.00159) | > avg_loss: -0.37059 (+0.00199) | > avg_log_mle: -0.49846 (+0.00528) | > avg_loss_dur: 0.12787 (-0.00329)  > EPOCH: 13/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 16:24:24)   --> STEP: 5/639 -- GLOBAL_STEP: 8325 | > loss: -0.34940 (-0.36763) | > log_mle: -0.55395 (-0.53616) | > loss_dur: 0.20455 (0.16853) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.78422 (6.53395) | > current_lr: 0.00069 | > step_time: 0.30670 (0.27618) | > loader_time: 0.00090 (0.00272)  --> STEP: 30/639 -- GLOBAL_STEP: 8350 | > loss: -0.37801 (-0.36191) | > log_mle: -0.50781 (-0.51578) | > loss_dur: 0.12980 (0.15388) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.76709 (6.71360) | > current_lr: 0.00069 | > step_time: 0.34420 (0.32555) | > loader_time: 0.00270 (0.00256)  --> STEP: 55/639 -- GLOBAL_STEP: 8375 | > loss: -0.33273 (-0.35956) | > log_mle: -0.45208 (-0.50743) | > loss_dur: 0.11935 (0.14788) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.56799 (5.83606) | > current_lr: 0.00069 | > step_time: 0.35770 (0.34871) | > loader_time: 0.00290 (0.00263)  --> STEP: 80/639 -- GLOBAL_STEP: 8400 | > loss: -0.35864 (-0.36059) | > log_mle: -0.50721 (-0.50560) | > loss_dur: 0.14857 (0.14501) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.44427 (5.58432) | > current_lr: 0.00069 | > step_time: 0.38480 (0.36722) | > loader_time: 0.00300 (0.00272)  --> STEP: 105/639 -- GLOBAL_STEP: 8425 | > loss: -0.32988 (-0.36381) | > log_mle: -0.46893 (-0.50686) | > loss_dur: 0.13905 (0.14305) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.86874 (5.49313) | > current_lr: 0.00069 | > step_time: 0.43200 (0.38669) | > loader_time: 0.00290 (0.00279)  --> STEP: 130/639 -- GLOBAL_STEP: 8450 | > loss: -0.36944 (-0.36696) | > log_mle: -0.51736 (-0.50901) | > loss_dur: 0.14792 (0.14206) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.28664 (5.54496) | > current_lr: 0.00069 | > step_time: 0.48090 (0.40151) | > loader_time: 0.00310 (0.00281)  --> STEP: 155/639 -- GLOBAL_STEP: 8475 | > loss: -0.37902 (-0.37053) | > log_mle: -0.51001 (-0.51100) | > loss_dur: 0.13099 (0.14047) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.49734 (5.73930) | > current_lr: 0.00069 | > step_time: 0.54430 (0.41723) | > loader_time: 0.00310 (0.00285)  --> STEP: 180/639 -- GLOBAL_STEP: 8500 | > loss: -0.42720 (-0.37559) | > log_mle: -0.56531 (-0.51483) | > loss_dur: 0.13812 (0.13925) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.18822 (5.85522) | > current_lr: 0.00069 | > step_time: 0.58210 (0.43372) | > loader_time: 0.00340 (0.00289)  --> STEP: 205/639 -- GLOBAL_STEP: 8525 | > loss: -0.36613 (-0.37517) | > log_mle: -0.49062 (-0.51322) | > loss_dur: 0.12449 (0.13805) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.48450 (5.66180) | > current_lr: 0.00069 | > step_time: 0.55760 (0.44858) | > loader_time: 0.00330 (0.00293)  --> STEP: 230/639 -- GLOBAL_STEP: 8550 | > loss: -0.39778 (-0.37797) | > log_mle: -0.52765 (-0.51492) | > loss_dur: 0.12987 (0.13695) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.52935 (5.68131) | > current_lr: 0.00068 | > step_time: 0.56290 (0.46646) | > loader_time: 0.00350 (0.00297)  --> STEP: 255/639 -- GLOBAL_STEP: 8575 | > loss: -0.35968 (-0.37949) | > log_mle: -0.49639 (-0.51559) | > loss_dur: 0.13671 (0.13610) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.84674 (5.71777) | > current_lr: 0.00068 | > step_time: 0.65580 (0.48016) | > loader_time: 0.00340 (0.00300)  --> STEP: 280/639 -- GLOBAL_STEP: 8600 | > loss: -0.38312 (-0.38106) | > log_mle: -0.51876 (-0.51684) | > loss_dur: 0.13563 (0.13578) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.05000 (5.81107) | > current_lr: 0.00068 | > step_time: 0.65720 (0.49751) | > loader_time: 0.00300 (0.00304)  --> STEP: 305/639 -- GLOBAL_STEP: 8625 | > loss: -0.37115 (-0.38145) | > log_mle: -0.50654 (-0.51687) | > loss_dur: 0.13539 (0.13542) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.36514 (5.83708) | > current_lr: 0.00068 | > step_time: 0.65770 (0.51143) | > loader_time: 0.00340 (0.00308)  --> STEP: 330/639 -- GLOBAL_STEP: 8650 | > loss: -0.39247 (-0.38120) | > log_mle: -0.53372 (-0.51660) | > loss_dur: 0.14126 (0.13540) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.33333 (5.77642) | > current_lr: 0.00068 | > step_time: 0.75650 (0.52663) | > loader_time: 0.00370 (0.00312)  --> STEP: 355/639 -- GLOBAL_STEP: 8675 | > loss: -0.34499 (-0.38150) | > log_mle: -0.46514 (-0.51625) | > loss_dur: 0.12015 (0.13475) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.96372 (5.75839) | > current_lr: 0.00068 | > step_time: 0.59250 (0.53988) | > loader_time: 0.00320 (0.00315)  --> STEP: 380/639 -- GLOBAL_STEP: 8700 | > loss: -0.37616 (-0.38109) | > log_mle: -0.49329 (-0.51535) | > loss_dur: 0.11713 (0.13426) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.50242 (5.65869) | > current_lr: 0.00068 | > step_time: 0.63530 (0.55306) | > loader_time: 0.00310 (0.00318)  --> STEP: 405/639 -- GLOBAL_STEP: 8725 | > loss: -0.38358 (-0.38132) | > log_mle: -0.50694 (-0.51541) | > loss_dur: 0.12335 (0.13409) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.23107 (5.64730) | > current_lr: 0.00068 | > step_time: 0.69170 (0.56566) | > loader_time: 0.00340 (0.00320)  --> STEP: 430/639 -- GLOBAL_STEP: 8750 | > loss: -0.34438 (-0.38098) | > log_mle: -0.46614 (-0.51477) | > loss_dur: 0.12176 (0.13379) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.38673 (5.60676) | > current_lr: 0.00068 | > step_time: 0.77530 (0.57863) | > loader_time: 0.00370 (0.00323)  --> STEP: 455/639 -- GLOBAL_STEP: 8775 | > loss: -0.36195 (-0.38068) | > log_mle: -0.49205 (-0.51403) | > loss_dur: 0.13010 (0.13335) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.77675 (5.54169) | > current_lr: 0.00068 | > step_time: 0.74550 (0.58926) | > loader_time: 0.00370 (0.00325)  --> STEP: 480/639 -- GLOBAL_STEP: 8800 | > loss: -0.32947 (-0.38049) | > log_mle: -0.45258 (-0.51365) | > loss_dur: 0.12311 (0.13316) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.99434 (5.55220) | > current_lr: 0.00067 | > step_time: 0.84200 (0.60258) | > loader_time: 0.00370 (0.00328)  --> STEP: 505/639 -- GLOBAL_STEP: 8825 | > loss: -0.39364 (-0.38055) | > log_mle: -0.51829 (-0.51323) | > loss_dur: 0.12465 (0.13267) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.40006 (5.49910) | > current_lr: 0.00067 | > step_time: 0.89540 (0.61532) | > loader_time: 0.00390 (0.00331)  --> STEP: 530/639 -- GLOBAL_STEP: 8850 | > loss: -0.37857 (-0.38071) | > log_mle: -0.50982 (-0.51308) | > loss_dur: 0.13125 (0.13238) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.55721 (5.48768) | > current_lr: 0.00067 | > step_time: 0.99400 (0.62881) | > loader_time: 0.00390 (0.00334)  --> STEP: 555/639 -- GLOBAL_STEP: 8875 | > loss: -0.36966 (-0.38056) | > log_mle: -0.49484 (-0.51261) | > loss_dur: 0.12518 (0.13205) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.41307 (5.45195) | > current_lr: 0.00067 | > step_time: 0.82560 (0.64316) | > loader_time: 0.00390 (0.00338)  --> STEP: 580/639 -- GLOBAL_STEP: 8900 | > loss: -0.39845 (-0.38054) | > log_mle: -0.51128 (-0.51235) | > loss_dur: 0.11283 (0.13181) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.02730 (5.46147) | > current_lr: 0.00067 | > step_time: 0.96930 (0.65799) | > loader_time: 0.00430 (0.00342)  --> STEP: 605/639 -- GLOBAL_STEP: 8925 | > loss: -0.37083 (-0.38053) | > log_mle: -0.48205 (-0.51219) | > loss_dur: 0.11122 (0.13166) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.39819 (5.42321) | > current_lr: 0.00067 | > step_time: 1.06460 (0.67428) | > loader_time: 0.00450 (0.00348)  --> STEP: 630/639 -- GLOBAL_STEP: 8950 | > loss: -0.38350 (-0.38072) | > log_mle: -0.51897 (-0.51224) | > loss_dur: 0.13547 (0.13152) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.22653 (5.43929) | > current_lr: 0.00067 | > step_time: 1.19990 (0.69168) | > loader_time: 0.00460 (0.00354) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02845 (-0.00301) | > avg_loss: -0.38526 (-0.01466) | > avg_log_mle: -0.51676 (-0.01830) | > avg_loss_dur: 0.13150 (+0.00364) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_8960.pth.tar  > EPOCH: 14/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 16:32:10)   --> STEP: 15/639 -- GLOBAL_STEP: 8975 | > loss: -0.49047 (-0.38524) | > log_mle: -0.64427 (-0.53528) | > loss_dur: 0.15380 (0.15004) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.16761 (6.32949) | > current_lr: 0.00067 | > step_time: 0.32720 (0.28913) | > loader_time: 0.00270 (0.00513)  --> STEP: 40/639 -- GLOBAL_STEP: 9000 | > loss: -0.35320 (-0.36625) | > log_mle: -0.49568 (-0.51353) | > loss_dur: 0.14248 (0.14727) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.76545 (5.08609) | > current_lr: 0.00067 | > step_time: 0.35310 (0.31988) | > loader_time: 0.00290 (0.00365)  --> STEP: 65/639 -- GLOBAL_STEP: 9025 | > loss: -0.37059 (-0.36492) | > log_mle: -0.49965 (-0.51006) | > loss_dur: 0.12906 (0.14514) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.33212 (5.69776) | > current_lr: 0.00067 | > step_time: 0.37480 (0.35533) | > loader_time: 0.00310 (0.00337)  --> STEP: 90/639 -- GLOBAL_STEP: 9050 | > loss: -0.31770 (-0.36440) | > log_mle: -0.46363 (-0.50684) | > loss_dur: 0.14593 (0.14244) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.06501 (5.24507) | > current_lr: 0.00067 | > step_time: 0.52380 (0.37300) | > loader_time: 0.00310 (0.00322)  --> STEP: 115/639 -- GLOBAL_STEP: 9075 | > loss: -0.30261 (-0.37032) | > log_mle: -0.43672 (-0.51149) | > loss_dur: 0.13411 (0.14116) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.91266 (5.48159) | > current_lr: 0.00066 | > step_time: 0.38850 (0.39199) | > loader_time: 0.00280 (0.00316)  --> STEP: 140/639 -- GLOBAL_STEP: 9100 | > loss: -0.36697 (-0.37537) | > log_mle: -0.52349 (-0.51520) | > loss_dur: 0.15652 (0.13984) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.73171 (5.83655) | > current_lr: 0.00066 | > step_time: 0.55070 (0.40694) | > loader_time: 0.00300 (0.00313)  --> STEP: 165/639 -- GLOBAL_STEP: 9125 | > loss: -0.39422 (-0.37859) | > log_mle: -0.53633 (-0.51693) | > loss_dur: 0.14210 (0.13834) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.58494 (5.97619) | > current_lr: 0.00066 | > step_time: 0.48430 (0.41987) | > loader_time: 0.00300 (0.00313)  --> STEP: 190/639 -- GLOBAL_STEP: 9150 | > loss: -0.32578 (-0.38199) | > log_mle: -0.46159 (-0.51904) | > loss_dur: 0.13581 (0.13705) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.49914 (6.07971) | > current_lr: 0.00066 | > step_time: 0.52120 (0.43930) | > loader_time: 0.00320 (0.00313)  --> STEP: 215/639 -- GLOBAL_STEP: 9175 | > loss: -0.45782 (-0.38291) | > log_mle: -0.57370 (-0.51929) | > loss_dur: 0.11588 (0.13638) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.08884 (6.10681) | > current_lr: 0.00066 | > step_time: 0.53720 (0.45535) | > loader_time: 0.00300 (0.00313)  --> STEP: 240/639 -- GLOBAL_STEP: 9200 | > loss: -0.44723 (-0.38620) | > log_mle: -0.57642 (-0.52149) | > loss_dur: 0.12919 (0.13529) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.74977 (6.15800) | > current_lr: 0.00066 | > step_time: 0.55780 (0.46995) | > loader_time: 0.00330 (0.00315)  --> STEP: 265/639 -- GLOBAL_STEP: 9225 | > loss: -0.45439 (-0.38741) | > log_mle: -0.59346 (-0.52221) | > loss_dur: 0.13907 (0.13480) | > amp_scaler: 8192.00000 (4296.93585) | > grad_norm: 14.31630 (6.17962) | > current_lr: 0.00066 | > step_time: 0.61480 (0.48574) | > loader_time: 0.00320 (0.00318)  --> STEP: 290/639 -- GLOBAL_STEP: 9250 | > loss: -0.38737 (-0.38761) | > log_mle: -0.50875 (-0.52219) | > loss_dur: 0.12138 (0.13458) | > amp_scaler: 8192.00000 (4632.71724) | > grad_norm: 3.85993 (6.18977) | > current_lr: 0.00066 | > step_time: 0.64780 (0.50145) | > loader_time: 0.00320 (0.00320)  --> STEP: 315/639 -- GLOBAL_STEP: 9275 | > loss: -0.36412 (-0.38821) | > log_mle: -0.48484 (-0.52233) | > loss_dur: 0.12072 (0.13412) | > amp_scaler: 8192.00000 (4915.20000) | > grad_norm: 2.43450 (6.21562) | > current_lr: 0.00066 | > step_time: 0.67760 (0.51579) | > loader_time: 0.00350 (0.00322)  --> STEP: 340/639 -- GLOBAL_STEP: 9300 | > loss: -0.39981 (-0.38842) | > log_mle: -0.51085 (-0.52258) | > loss_dur: 0.11104 (0.13416) | > amp_scaler: 8192.00000 (5156.14118) | > grad_norm: 5.46086 (6.27600) | > current_lr: 0.00066 | > step_time: 0.73210 (0.53207) | > loader_time: 0.00350 (0.00324)  --> STEP: 365/639 -- GLOBAL_STEP: 9325 | > loss: -0.39224 (-0.38812) | > log_mle: -0.52043 (-0.52155) | > loss_dur: 0.12819 (0.13343) | > amp_scaler: 8192.00000 (5364.07671) | > grad_norm: 4.98472 (6.15837) | > current_lr: 0.00066 | > step_time: 0.82090 (0.54457) | > loader_time: 0.00400 (0.00326)  --> STEP: 390/639 -- GLOBAL_STEP: 9350 | > loss: -0.38524 (-0.38785) | > log_mle: -0.51748 (-0.52084) | > loss_dur: 0.13224 (0.13299) | > amp_scaler: 8192.00000 (5545.35385) | > grad_norm: 6.87244 (6.08853) | > current_lr: 0.00065 | > step_time: 0.70770 (0.55802) | > loader_time: 0.00350 (0.00328)  --> STEP: 415/639 -- GLOBAL_STEP: 9375 | > loss: -0.35723 (-0.38755) | > log_mle: -0.49363 (-0.52046) | > loss_dur: 0.13640 (0.13291) | > amp_scaler: 8192.00000 (5704.79036) | > grad_norm: 3.95325 (6.12437) | > current_lr: 0.00065 | > step_time: 0.72230 (0.57044) | > loader_time: 0.00360 (0.00331)  --> STEP: 440/639 -- GLOBAL_STEP: 9400 | > loss: -0.34914 (-0.38753) | > log_mle: -0.47936 (-0.52010) | > loss_dur: 0.13022 (0.13257) | > amp_scaler: 8192.00000 (5846.10909) | > grad_norm: 2.53972 (6.10731) | > current_lr: 0.00065 | > step_time: 0.72500 (0.58306) | > loader_time: 0.00370 (0.00333)  --> STEP: 465/639 -- GLOBAL_STEP: 9425 | > loss: -0.36042 (-0.38701) | > log_mle: -0.48586 (-0.51917) | > loss_dur: 0.12545 (0.13216) | > amp_scaler: 8192.00000 (5972.23226) | > grad_norm: 2.92123 (5.98761) | > current_lr: 0.00065 | > step_time: 0.79260 (0.59400) | > loader_time: 0.00380 (0.00335)  --> STEP: 490/639 -- GLOBAL_STEP: 9450 | > loss: -0.35438 (-0.38674) | > log_mle: -0.46987 (-0.51872) | > loss_dur: 0.11549 (0.13198) | > amp_scaler: 8192.00000 (6085.48571) | > grad_norm: 1.94910 (5.92258) | > current_lr: 0.00065 | > step_time: 0.80120 (0.60732) | > loader_time: 0.00340 (0.00337)  --> STEP: 515/639 -- GLOBAL_STEP: 9475 | > loss: -0.37604 (-0.38694) | > log_mle: -0.50103 (-0.51859) | > loss_dur: 0.12500 (0.13166) | > amp_scaler: 8192.00000 (6187.74369) | > grad_norm: 4.05867 (5.91188) | > current_lr: 0.00065 | > step_time: 1.00110 (0.62038) | > loader_time: 0.00400 (0.00340)  --> STEP: 540/639 -- GLOBAL_STEP: 9500 | > loss: -0.39552 (-0.38688) | > log_mle: -0.51710 (-0.51824) | > loss_dur: 0.12157 (0.13135) | > amp_scaler: 8192.00000 (6280.53333) | > grad_norm: 4.40985 (5.85412) | > current_lr: 0.00065 | > step_time: 0.85490 (0.63506) | > loader_time: 0.00410 (0.00342)  --> STEP: 565/639 -- GLOBAL_STEP: 9525 | > loss: -0.34883 (-0.38643) | > log_mle: -0.47841 (-0.51762) | > loss_dur: 0.12958 (0.13119) | > amp_scaler: 8192.00000 (6365.11150) | > grad_norm: 2.81939 (5.82750) | > current_lr: 0.00065 | > step_time: 0.99480 (0.64885) | > loader_time: 0.00420 (0.00345)  --> STEP: 590/639 -- GLOBAL_STEP: 9550 | > loss: -0.39044 (-0.38659) | > log_mle: -0.51534 (-0.51758) | > loss_dur: 0.12490 (0.13099) | > amp_scaler: 8192.00000 (6442.52203) | > grad_norm: 5.94606 (5.78907) | > current_lr: 0.00065 | > step_time: 0.95920 (0.66455) | > loader_time: 0.00390 (0.00348)  --> STEP: 615/639 -- GLOBAL_STEP: 9575 | > loss: -0.39436 (-0.38644) | > log_mle: -0.51565 (-0.51723) | > loss_dur: 0.12128 (0.13079) | > amp_scaler: 8192.00000 (6513.63902) | > grad_norm: 3.69868 (5.72526) | > current_lr: 0.00065 | > step_time: 1.03860 (0.68069) | > loader_time: 0.00450 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02661 (-0.00184) | > avg_loss: -0.39218 (-0.00692) | > avg_log_mle: -0.52078 (-0.00402) | > avg_loss_dur: 0.12860 (-0.00290) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_9600.pth.tar  > EPOCH: 15/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 16:39:57)   --> STEP: 0/639 -- GLOBAL_STEP: 9600 | > loss: -0.43523 (-0.43523) | > log_mle: -0.60381 (-0.60381) | > loss_dur: 0.16859 (0.16859) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 13.15717 (13.15717) | > current_lr: 0.00065 | > step_time: 0.71090 (0.71092) | > loader_time: 0.42880 (0.42881)  --> STEP: 25/639 -- GLOBAL_STEP: 9625 | > loss: -0.41655 (-0.38928) | > log_mle: -0.55545 (-0.53636) | > loss_dur: 0.13890 (0.14708) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 7.71175 (7.14398) | > current_lr: 0.00065 | > step_time: 0.38670 (0.32748) | > loader_time: 0.00280 (0.00268)  --> STEP: 50/639 -- GLOBAL_STEP: 9650 | > loss: -0.35836 (-0.38091) | > log_mle: -0.48322 (-0.52464) | > loss_dur: 0.12487 (0.14374) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 2.24550 (5.68347) | > current_lr: 0.00064 | > step_time: 0.31950 (0.34899) | > loader_time: 0.00290 (0.00275)  --> STEP: 75/639 -- GLOBAL_STEP: 9675 | > loss: -0.39503 (-0.37923) | > log_mle: -0.52217 (-0.52092) | > loss_dur: 0.12714 (0.14169) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.38254 (5.50349) | > current_lr: 0.00064 | > step_time: 0.40830 (0.37019) | > loader_time: 0.00310 (0.00282)  --> STEP: 100/639 -- GLOBAL_STEP: 9700 | > loss: -0.40113 (-0.38174) | > log_mle: -0.53297 (-0.52142) | > loss_dur: 0.13183 (0.13967) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 7.37497 (5.60227) | > current_lr: 0.00064 | > step_time: 0.50760 (0.38555) | > loader_time: 0.00320 (0.00285)  --> STEP: 125/639 -- GLOBAL_STEP: 9725 | > loss: -0.41014 (-0.38450) | > log_mle: -0.53711 (-0.52292) | > loss_dur: 0.12697 (0.13842) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 4.67859 (5.64924) | > current_lr: 0.00064 | > step_time: 0.42960 (0.39949) | > loader_time: 0.00280 (0.00289)  --> STEP: 150/639 -- GLOBAL_STEP: 9750 | > loss: -0.36816 (-0.38780) | > log_mle: -0.50202 (-0.52502) | > loss_dur: 0.13386 (0.13722) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.90343 (5.98688) | > current_lr: 0.00064 | > step_time: 0.42740 (0.41555) | > loader_time: 0.00280 (0.00291)  --> STEP: 175/639 -- GLOBAL_STEP: 9775 | > loss: -0.40267 (-0.39106) | > log_mle: -0.52049 (-0.52729) | > loss_dur: 0.11782 (0.13623) | > amp_scaler: 4096.00000 (7864.32000) | > grad_norm: 3.48657 (6.15867) | > current_lr: 0.00064 | > step_time: 0.54770 (0.43070) | > loader_time: 0.00290 (0.00294)  --> STEP: 200/639 -- GLOBAL_STEP: 9800 | > loss: -0.38082 (-0.39156) | > log_mle: -0.51141 (-0.52680) | > loss_dur: 0.13059 (0.13524) | > amp_scaler: 4096.00000 (7393.28000) | > grad_norm: 4.51328 (6.17492) | > current_lr: 0.00064 | > step_time: 0.54160 (0.44512) | > loader_time: 0.00310 (0.00295)  --> STEP: 225/639 -- GLOBAL_STEP: 9825 | > loss: -0.36828 (-0.39221) | > log_mle: -0.48397 (-0.52648) | > loss_dur: 0.11569 (0.13427) | > amp_scaler: 4096.00000 (7026.91556) | > grad_norm: 1.77367 (6.07991) | > current_lr: 0.00064 | > step_time: 0.49890 (0.46047) | > loader_time: 0.00330 (0.00299)  --> STEP: 250/639 -- GLOBAL_STEP: 9850 | > loss: -0.38232 (-0.39536) | > log_mle: -0.49045 (-0.52895) | > loss_dur: 0.10813 (0.13359) | > amp_scaler: 4096.00000 (6733.82400) | > grad_norm: 2.91768 (6.20165) | > current_lr: 0.00064 | > step_time: 0.57330 (0.47563) | > loader_time: 0.00300 (0.00302)  --> STEP: 275/639 -- GLOBAL_STEP: 9875 | > loss: -0.40726 (-0.39752) | > log_mle: -0.53780 (-0.53072) | > loss_dur: 0.13054 (0.13321) | > amp_scaler: 4096.00000 (6494.02182) | > grad_norm: 4.75403 (6.37434) | > current_lr: 0.00064 | > step_time: 0.87140 (0.49286) | > loader_time: 0.00390 (0.00305)  --> STEP: 300/639 -- GLOBAL_STEP: 9900 | > loss: -0.40659 (-0.39724) | > log_mle: -0.52931 (-0.53029) | > loss_dur: 0.12272 (0.13305) | > amp_scaler: 4096.00000 (6294.18667) | > grad_norm: 5.49144 (6.30833) | > current_lr: 0.00064 | > step_time: 0.71340 (0.50626) | > loader_time: 0.00360 (0.00308)  --> STEP: 325/639 -- GLOBAL_STEP: 9925 | > loss: -0.40265 (-0.39736) | > log_mle: -0.53191 (-0.53027) | > loss_dur: 0.12926 (0.13290) | > amp_scaler: 4096.00000 (6125.09538) | > grad_norm: 7.82374 (6.33826) | > current_lr: 0.00064 | > step_time: 0.64020 (0.52192) | > loader_time: 0.00340 (0.00312)  --> STEP: 350/639 -- GLOBAL_STEP: 9950 | > loss: -0.43233 (-0.39724) | > log_mle: -0.56116 (-0.52975) | > loss_dur: 0.12883 (0.13252) | > amp_scaler: 4096.00000 (5980.16000) | > grad_norm: 10.25435 (6.34052) | > current_lr: 0.00063 | > step_time: 0.87410 (0.53800) | > loader_time: 0.00380 (0.00316)  --> STEP: 375/639 -- GLOBAL_STEP: 9975 | > loss: -0.37242 (-0.39672) | > log_mle: -0.49660 (-0.52869) | > loss_dur: 0.12418 (0.13197) | > amp_scaler: 4096.00000 (5854.54933) | > grad_norm: 3.24773 (6.22441) | > current_lr: 0.00063 | > step_time: 0.68100 (0.55130) | > loader_time: 0.00320 (0.00319)  --> STEP: 400/639 -- GLOBAL_STEP: 10000 | > loss: -0.40866 (-0.39665) | > log_mle: -0.53057 (-0.52848) | > loss_dur: 0.12192 (0.13183) | > amp_scaler: 4096.00000 (5744.64000) | > grad_norm: 5.24084 (6.17516) | > current_lr: 0.00063 | > step_time: 0.74520 (0.56339) | > loader_time: 0.00370 (0.00321) > CHECKPOINT : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/checkpoint_10000.pth.tar warning: audio amplitude out of range, auto clipped.  --> STEP: 425/639 -- GLOBAL_STEP: 10025 | > loss: -0.45118 (-0.39608) | > log_mle: -0.58110 (-0.52764) | > loss_dur: 0.12992 (0.13156) | > amp_scaler: 4096.00000 (5647.66118) | > grad_norm: 9.17226 (6.12633) | > current_lr: 0.00063 | > step_time: 0.88900 (0.57515) | > loader_time: 0.00390 (0.00329)  --> STEP: 450/639 -- GLOBAL_STEP: 10050 | > loss: -0.36536 (-0.39563) | > log_mle: -0.48629 (-0.52683) | > loss_dur: 0.12093 (0.13120) | > amp_scaler: 4096.00000 (5561.45778) | > grad_norm: 2.00982 (6.07392) | > current_lr: 0.00063 | > step_time: 0.78530 (0.58648) | > loader_time: 0.00380 (0.00331)  --> STEP: 475/639 -- GLOBAL_STEP: 10075 | > loss: -0.37999 (-0.39510) | > log_mle: -0.49656 (-0.52615) | > loss_dur: 0.11657 (0.13106) | > amp_scaler: 4096.00000 (5484.32842) | > grad_norm: 2.74205 (6.02788) | > current_lr: 0.00063 | > step_time: 0.70470 (0.59762) | > loader_time: 0.00370 (0.00334)  --> STEP: 500/639 -- GLOBAL_STEP: 10100 | > loss: -0.43577 (-0.39502) | > log_mle: -0.56189 (-0.52583) | > loss_dur: 0.12612 (0.13081) | > amp_scaler: 4096.00000 (5414.91200) | > grad_norm: 7.32732 (5.99329) | > current_lr: 0.00063 | > step_time: 0.95440 (0.61101) | > loader_time: 0.00350 (0.00336)  --> STEP: 525/639 -- GLOBAL_STEP: 10125 | > loss: -0.40595 (-0.39499) | > log_mle: -0.54403 (-0.52555) | > loss_dur: 0.13808 (0.13056) | > amp_scaler: 4096.00000 (5352.10667) | > grad_norm: 10.42679 (5.96069) | > current_lr: 0.00063 | > step_time: 1.04310 (0.62487) | > loader_time: 0.00420 (0.00338)  --> STEP: 550/639 -- GLOBAL_STEP: 10150 | > loss: -0.35117 (-0.39455) | > log_mle: -0.47809 (-0.52483) | > loss_dur: 0.12693 (0.13028) | > amp_scaler: 4096.00000 (5295.01091) | > grad_norm: 4.36972 (5.91976) | > current_lr: 0.00063 | > step_time: 0.94230 (0.64033) | > loader_time: 0.00420 (0.00342)  --> STEP: 575/639 -- GLOBAL_STEP: 10175 | > loss: -0.37619 (-0.39423) | > log_mle: -0.49594 (-0.52431) | > loss_dur: 0.11974 (0.13008) | > amp_scaler: 4096.00000 (5242.88000) | > grad_norm: 4.04940 (5.92480) | > current_lr: 0.00063 | > step_time: 0.94370 (0.65487) | > loader_time: 0.00380 (0.00345)  --> STEP: 600/639 -- GLOBAL_STEP: 10200 | > loss: -0.36057 (-0.39414) | > log_mle: -0.49280 (-0.52412) | > loss_dur: 0.13223 (0.12997) | > amp_scaler: 4096.00000 (5195.09333) | > grad_norm: 5.28949 (5.90753) | > current_lr: 0.00063 | > step_time: 1.07340 (0.67115) | > loader_time: 0.00440 (0.00348)  --> STEP: 625/639 -- GLOBAL_STEP: 10225 | > loss: -0.39598 (-0.39403) | > log_mle: -0.52296 (-0.52381) | > loss_dur: 0.12698 (0.12978) | > amp_scaler: 4096.00000 (5151.12960) | > grad_norm: 6.02633 (5.86720) | > current_lr: 0.00063 | > step_time: 1.07430 (0.68776) | > loader_time: 0.00430 (0.00353) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02346 (-0.00315) | > avg_loss: -0.39787 (-0.00569) | > avg_log_mle: -0.52404 (-0.00326) | > avg_loss_dur: 0.12617 (-0.00243) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_10240.pth.tar  > EPOCH: 16/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 16:47:48)   --> STEP: 10/639 -- GLOBAL_STEP: 10250 | > loss: -0.40425 (-0.40673) | > log_mle: -0.56212 (-0.55524) | > loss_dur: 0.15787 (0.14850) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.34989 (7.56813) | > current_lr: 0.00063 | > step_time: 0.32120 (0.29690) | > loader_time: 0.00300 (0.00242)  --> STEP: 35/639 -- GLOBAL_STEP: 10275 | > loss: -0.38274 (-0.38438) | > log_mle: -0.51096 (-0.52897) | > loss_dur: 0.12823 (0.14459) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.53290 (6.17219) | > current_lr: 0.00062 | > step_time: 0.33970 (0.32827) | > loader_time: 0.00280 (0.00266)  --> STEP: 60/639 -- GLOBAL_STEP: 10300 | > loss: -0.34719 (-0.38198) | > log_mle: -0.48145 (-0.52590) | > loss_dur: 0.13425 (0.14392) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.75085 (5.87840) | > current_lr: 0.00062 | > step_time: 0.34560 (0.35649) | > loader_time: 0.00280 (0.00278)  --> STEP: 85/639 -- GLOBAL_STEP: 10325 | > loss: -0.33240 (-0.38214) | > log_mle: -0.48074 (-0.52268) | > loss_dur: 0.14834 (0.14054) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.00633 (5.63836) | > current_lr: 0.00062 | > step_time: 0.40940 (0.37174) | > loader_time: 0.00300 (0.00284)  --> STEP: 110/639 -- GLOBAL_STEP: 10350 | > loss: -0.37262 (-0.38873) | > log_mle: -0.51023 (-0.52816) | > loss_dur: 0.13761 (0.13943) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.56263 (6.05183) | > current_lr: 0.00062 | > step_time: 0.45260 (0.39291) | > loader_time: 0.00290 (0.00288)  --> STEP: 135/639 -- GLOBAL_STEP: 10375 | > loss: -0.34963 (-0.39115) | > log_mle: -0.49205 (-0.52884) | > loss_dur: 0.14242 (0.13770) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.48453 (6.07588) | > current_lr: 0.00062 | > step_time: 0.46950 (0.40391) | > loader_time: 0.00330 (0.00291)  --> STEP: 160/639 -- GLOBAL_STEP: 10400 | > loss: -0.42380 (-0.39309) | > log_mle: -0.54139 (-0.52911) | > loss_dur: 0.11759 (0.13602) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.58816 (6.18258) | > current_lr: 0.00062 | > step_time: 0.46860 (0.41957) | > loader_time: 0.00300 (0.00295)  --> STEP: 185/639 -- GLOBAL_STEP: 10425 | > loss: -0.40320 (-0.39798) | > log_mle: -0.52750 (-0.53331) | > loss_dur: 0.12430 (0.13533) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.62717 (6.57833) | > current_lr: 0.00062 | > step_time: 0.56870 (0.43787) | > loader_time: 0.00330 (0.00299)  --> STEP: 210/639 -- GLOBAL_STEP: 10450 | > loss: -0.32641 (-0.39791) | > log_mle: -0.44493 (-0.53251) | > loss_dur: 0.11852 (0.13460) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.11331 (6.58755) | > current_lr: 0.00062 | > step_time: 0.50160 (0.45341) | > loader_time: 0.00270 (0.00302)  --> STEP: 235/639 -- GLOBAL_STEP: 10475 | > loss: -0.42034 (-0.40086) | > log_mle: -0.54159 (-0.53428) | > loss_dur: 0.12125 (0.13342) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.52537 (6.69986) | > current_lr: 0.00062 | > step_time: 0.56170 (0.46900) | > loader_time: 0.00340 (0.00305)  --> STEP: 260/639 -- GLOBAL_STEP: 10500 | > loss: -0.41541 (-0.40276) | > log_mle: -0.53662 (-0.53560) | > loss_dur: 0.12122 (0.13285) | > amp_scaler: 2048.00000 (4072.36923) | > grad_norm: 8.57917 (6.78025) | > current_lr: 0.00062 | > step_time: 0.58490 (0.48378) | > loader_time: 0.00340 (0.00308)  --> STEP: 285/639 -- GLOBAL_STEP: 10525 | > loss: -0.35494 (-0.40340) | > log_mle: -0.48897 (-0.53592) | > loss_dur: 0.13404 (0.13252) | > amp_scaler: 2048.00000 (3894.79298) | > grad_norm: 1.80287 (6.73241) | > current_lr: 0.00062 | > step_time: 0.61110 (0.50009) | > loader_time: 0.00340 (0.00312)  --> STEP: 310/639 -- GLOBAL_STEP: 10550 | > loss: -0.41283 (-0.40380) | > log_mle: -0.53622 (-0.53591) | > loss_dur: 0.12338 (0.13211) | > amp_scaler: 2048.00000 (3745.85806) | > grad_norm: 6.77825 (6.72919) | > current_lr: 0.00062 | > step_time: 0.56840 (0.51406) | > loader_time: 0.00340 (0.00316)  --> STEP: 335/639 -- GLOBAL_STEP: 10575 | > loss: -0.43196 (-0.40312) | > log_mle: -0.56060 (-0.53527) | > loss_dur: 0.12864 (0.13214) | > amp_scaler: 2048.00000 (3619.15224) | > grad_norm: 12.84017 (6.86429) | > current_lr: 0.00062 | > step_time: 0.67110 (0.52997) | > loader_time: 0.00370 (0.00319)  --> STEP: 360/639 -- GLOBAL_STEP: 10600 | > loss: -0.36348 (-0.40301) | > log_mle: -0.49274 (-0.53460) | > loss_dur: 0.12926 (0.13158) | > amp_scaler: 2048.00000 (3510.04444) | > grad_norm: 4.64048 (6.92473) | > current_lr: 0.00061 | > step_time: 0.74390 (0.54253) | > loader_time: 0.00360 (0.00322)  --> STEP: 385/639 -- GLOBAL_STEP: 10625 | > loss: -0.40489 (-0.40272) | > log_mle: -0.53154 (-0.53391) | > loss_dur: 0.12665 (0.13119) | > amp_scaler: 2048.00000 (3415.10649) | > grad_norm: 5.33055 (6.80247) | > current_lr: 0.00061 | > step_time: 0.70540 (0.55622) | > loader_time: 0.00370 (0.00325)  --> STEP: 410/639 -- GLOBAL_STEP: 10650 | > loss: -0.36523 (-0.40223) | > log_mle: -0.50923 (-0.53326) | > loss_dur: 0.14400 (0.13103) | > amp_scaler: 2048.00000 (3331.74634) | > grad_norm: 4.67458 (6.75476) | > current_lr: 0.00061 | > step_time: 0.78070 (0.56792) | > loader_time: 0.00380 (0.00327)  --> STEP: 435/639 -- GLOBAL_STEP: 10675 | > loss: -0.39825 (-0.40162) | > log_mle: -0.51698 (-0.53232) | > loss_dur: 0.11873 (0.13069) | > amp_scaler: 2048.00000 (3257.96782) | > grad_norm: 4.36330 (6.68248) | > current_lr: 0.00061 | > step_time: 0.72340 (0.57983) | > loader_time: 0.00360 (0.00330)  --> STEP: 460/639 -- GLOBAL_STEP: 10700 | > loss: -0.35565 (-0.40096) | > log_mle: -0.48387 (-0.53133) | > loss_dur: 0.12822 (0.13037) | > amp_scaler: 2048.00000 (3192.20870) | > grad_norm: 1.41195 (6.54010) | > current_lr: 0.00061 | > step_time: 0.91590 (0.59191) | > loader_time: 0.00400 (0.00332)  --> STEP: 485/639 -- GLOBAL_STEP: 10725 | > loss: -0.40528 (-0.40057) | > log_mle: -0.52062 (-0.53085) | > loss_dur: 0.11534 (0.13028) | > amp_scaler: 2048.00000 (3133.22887) | > grad_norm: 4.47798 (6.50310) | > current_lr: 0.00061 | > step_time: 0.96190 (0.60473) | > loader_time: 0.00410 (0.00335)  --> STEP: 510/639 -- GLOBAL_STEP: 10750 | > loss: -0.38914 (-0.40042) | > log_mle: -0.52224 (-0.53039) | > loss_dur: 0.13309 (0.12997) | > amp_scaler: 2048.00000 (3080.03137) | > grad_norm: 4.36692 (6.45205) | > current_lr: 0.00061 | > step_time: 0.87980 (0.61777) | > loader_time: 0.00400 (0.00338)  --> STEP: 535/639 -- GLOBAL_STEP: 10775 | > loss: -0.38839 (-0.40016) | > log_mle: -0.51809 (-0.52989) | > loss_dur: 0.12970 (0.12973) | > amp_scaler: 2048.00000 (3031.80561) | > grad_norm: 3.86843 (6.39548) | > current_lr: 0.00061 | > step_time: 0.91650 (0.63113) | > loader_time: 0.00440 (0.00341)  --> STEP: 560/639 -- GLOBAL_STEP: 10800 | > loss: -0.38599 (-0.39971) | > log_mle: -0.51904 (-0.52932) | > loss_dur: 0.13305 (0.12961) | > amp_scaler: 2048.00000 (2987.88571) | > grad_norm: 9.18623 (6.39829) | > current_lr: 0.00061 | > step_time: 0.92580 (0.64627) | > loader_time: 0.00420 (0.00345)  --> STEP: 585/639 -- GLOBAL_STEP: 10825 | > loss: -0.40200 (-0.39955) | > log_mle: -0.52495 (-0.52893) | > loss_dur: 0.12295 (0.12938) | > amp_scaler: 2048.00000 (2947.71966) | > grad_norm: 5.22067 (6.37620) | > current_lr: 0.00061 | > step_time: 1.03730 (0.66070) | > loader_time: 0.00430 (0.00348)  --> STEP: 610/639 -- GLOBAL_STEP: 10850 | > loss: -0.37154 (-0.39929) | > log_mle: -0.49605 (-0.52855) | > loss_dur: 0.12451 (0.12925) | > amp_scaler: 2048.00000 (2910.84590) | > grad_norm: 3.03646 (6.31856) | > current_lr: 0.00061 | > step_time: 1.07080 (0.67680) | > loader_time: 0.00450 (0.00352)  --> STEP: 635/639 -- GLOBAL_STEP: 10875 | > loss: -0.40155 (-0.39931) | > log_mle: -0.52679 (-0.52855) | > loss_dur: 0.12524 (0.12924) | > amp_scaler: 2048.00000 (2876.87559) | > grad_norm: 3.94021 (6.29976) | > current_lr: 0.00061 | > step_time: 1.14290 (0.69350) | > loader_time: 0.00440 (0.00356) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02829 (+0.00483) | > avg_loss: -0.39278 (+0.00509) | > avg_log_mle: -0.51949 (+0.00455) | > avg_loss_dur: 0.12670 (+0.00053)  > EPOCH: 17/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 16:55:27)   --> STEP: 20/639 -- GLOBAL_STEP: 10900 | > loss: -0.51353 (-0.41181) | > log_mle: -0.66877 (-0.55621) | > loss_dur: 0.15524 (0.14440) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.58140 (7.89841) | > current_lr: 0.00061 | > step_time: 0.34510 (0.30910) | > loader_time: 0.00250 (0.00333)  --> STEP: 45/639 -- GLOBAL_STEP: 10925 | > loss: -0.34645 (-0.39716) | > log_mle: -0.47583 (-0.53809) | > loss_dur: 0.12939 (0.14093) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.01124 (6.93986) | > current_lr: 0.00061 | > step_time: 0.40720 (0.34147) | > loader_time: 0.00280 (0.00303)  --> STEP: 70/639 -- GLOBAL_STEP: 10950 | > loss: -0.43004 (-0.39304) | > log_mle: -0.56099 (-0.53331) | > loss_dur: 0.13095 (0.14026) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.97403 (6.98284) | > current_lr: 0.00060 | > step_time: 0.39130 (0.36768) | > loader_time: 0.00290 (0.00298)  --> STEP: 95/639 -- GLOBAL_STEP: 10975 | > loss: -0.47304 (-0.39467) | > log_mle: -0.59849 (-0.53194) | > loss_dur: 0.12545 (0.13727) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.23390 (6.57743) | > current_lr: 0.00060 | > step_time: 0.40660 (0.38332) | > loader_time: 0.00260 (0.00298)  --> STEP: 120/639 -- GLOBAL_STEP: 11000 | > loss: -0.45262 (-0.39686) | > log_mle: -0.59901 (-0.53361) | > loss_dur: 0.14639 (0.13675) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.83476 (6.74364) | > current_lr: 0.00060 | > step_time: 0.56050 (0.40036) | > loader_time: 0.00290 (0.00299)  --> STEP: 145/639 -- GLOBAL_STEP: 11025 | > loss: -0.41166 (-0.40040) | > log_mle: -0.55468 (-0.53601) | > loss_dur: 0.14302 (0.13561) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.38552 (6.92774) | > current_lr: 0.00060 | > step_time: 0.47030 (0.41510) | > loader_time: 0.00260 (0.00300)  --> STEP: 170/639 -- GLOBAL_STEP: 11050 | > loss: -0.41918 (-0.40238) | > log_mle: -0.55017 (-0.53706) | > loss_dur: 0.13099 (0.13468) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.13733 (6.89961) | > current_lr: 0.00060 | > step_time: 0.53710 (0.42906) | > loader_time: 0.00330 (0.00302)  --> STEP: 195/639 -- GLOBAL_STEP: 11075 | > loss: -0.37227 (-0.40375) | > log_mle: -0.48798 (-0.53765) | > loss_dur: 0.11570 (0.13390) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.38726 (6.84711) | > current_lr: 0.00060 | > step_time: 0.49820 (0.44542) | > loader_time: 0.00350 (0.00304)  --> STEP: 220/639 -- GLOBAL_STEP: 11100 | > loss: -0.43137 (-0.40394) | > log_mle: -0.54358 (-0.53715) | > loss_dur: 0.11221 (0.13320) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.13958 (6.75131) | > current_lr: 0.00060 | > step_time: 0.73630 (0.46230) | > loader_time: 0.00340 (0.00307)  --> STEP: 245/639 -- GLOBAL_STEP: 11125 | > loss: -0.42112 (-0.40702) | > log_mle: -0.53114 (-0.53931) | > loss_dur: 0.11002 (0.13228) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.62353 (6.81475) | > current_lr: 0.00060 | > step_time: 0.52070 (0.47642) | > loader_time: 0.00320 (0.00310)  --> STEP: 270/639 -- GLOBAL_STEP: 11150 | > loss: -0.45001 (-0.40872) | > log_mle: -0.58184 (-0.54068) | > loss_dur: 0.13183 (0.13196) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 16.54692 (6.91547) | > current_lr: 0.00060 | > step_time: 0.83300 (0.49280) | > loader_time: 0.00400 (0.00313)  --> STEP: 295/639 -- GLOBAL_STEP: 11175 | > loss: -0.37429 (-0.40837) | > log_mle: -0.49335 (-0.54012) | > loss_dur: 0.11906 (0.13175) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.96368 (6.81786) | > current_lr: 0.00060 | > step_time: 0.66190 (0.50648) | > loader_time: 0.00350 (0.00316)  --> STEP: 320/639 -- GLOBAL_STEP: 11200 | > loss: -0.41111 (-0.40855) | > log_mle: -0.54471 (-0.54017) | > loss_dur: 0.13361 (0.13161) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.74838 (6.83119) | > current_lr: 0.00060 | > step_time: 0.75140 (0.52170) | > loader_time: 0.00360 (0.00319)  --> STEP: 345/639 -- GLOBAL_STEP: 11225 | > loss: -0.40888 (-0.40815) | > log_mle: -0.52972 (-0.53968) | > loss_dur: 0.12084 (0.13153) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.15250 (6.84518) | > current_lr: 0.00060 | > step_time: 0.65590 (0.53599) | > loader_time: 0.00340 (0.00321)  --> STEP: 370/639 -- GLOBAL_STEP: 11250 | > loss: -0.42776 (-0.40760) | > log_mle: -0.54787 (-0.53854) | > loss_dur: 0.12011 (0.13094) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.58732 (6.72739) | > current_lr: 0.00060 | > step_time: 0.87340 (0.54893) | > loader_time: 0.00370 (0.00323)  --> STEP: 395/639 -- GLOBAL_STEP: 11275 | > loss: -0.38431 (-0.40700) | > log_mle: -0.52000 (-0.53769) | > loss_dur: 0.13569 (0.13069) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.23726 (6.62317) | > current_lr: 0.00060 | > step_time: 0.72880 (0.56074) | > loader_time: 0.00370 (0.00325)  --> STEP: 420/639 -- GLOBAL_STEP: 11300 | > loss: -0.39542 (-0.40638) | > log_mle: -0.51619 (-0.53702) | > loss_dur: 0.12077 (0.13064) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.41356 (6.57163) | > current_lr: 0.00060 | > step_time: 0.74260 (0.57340) | > loader_time: 0.00340 (0.00328)  --> STEP: 445/639 -- GLOBAL_STEP: 11325 | > loss: -0.39687 (-0.40611) | > log_mle: -0.51808 (-0.53643) | > loss_dur: 0.12121 (0.13032) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.32261 (6.53469) | > current_lr: 0.00059 | > step_time: 0.74170 (0.58585) | > loader_time: 0.00330 (0.00330)  --> STEP: 470/639 -- GLOBAL_STEP: 11350 | > loss: -0.40352 (-0.40543) | > log_mle: -0.52358 (-0.53547) | > loss_dur: 0.12006 (0.13004) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 24.95819 (6.50998) | > current_lr: 0.00059 | > step_time: 0.76690 (0.59720) | > loader_time: 0.00370 (0.00332)  --> STEP: 495/639 -- GLOBAL_STEP: 11375 | > loss: -0.35931 (-0.40496) | > log_mle: -0.46951 (-0.53479) | > loss_dur: 0.11020 (0.12983) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.54314 (6.49665) | > current_lr: 0.00059 | > step_time: 0.77660 (0.61019) | > loader_time: 0.00340 (0.00335)  --> STEP: 520/639 -- GLOBAL_STEP: 11400 | > loss: -0.41382 (-0.40505) | > log_mle: -0.52846 (-0.53453) | > loss_dur: 0.11463 (0.12948) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.42174 (6.47140) | > current_lr: 0.00059 | > step_time: 0.85880 (0.62333) | > loader_time: 0.00380 (0.00337)  --> STEP: 545/639 -- GLOBAL_STEP: 11425 | > loss: -0.41649 (-0.40475) | > log_mle: -0.55527 (-0.53412) | > loss_dur: 0.13877 (0.12937) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.16721 (6.41897) | > current_lr: 0.00059 | > step_time: 1.02620 (0.63856) | > loader_time: 0.00400 (0.00341)  --> STEP: 570/639 -- GLOBAL_STEP: 11450 | > loss: -0.39235 (-0.40425) | > log_mle: -0.50871 (-0.53343) | > loss_dur: 0.11637 (0.12918) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.22612 (6.39752) | > current_lr: 0.00059 | > step_time: 0.85550 (0.65206) | > loader_time: 0.00400 (0.00345)  --> STEP: 595/639 -- GLOBAL_STEP: 11475 | > loss: -0.38473 (-0.40425) | > log_mle: -0.50280 (-0.53326) | > loss_dur: 0.11806 (0.12901) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.09853 (6.40044) | > current_lr: 0.00059 | > step_time: 0.99470 (0.66857) | > loader_time: 0.00610 (0.00350)  --> STEP: 620/639 -- GLOBAL_STEP: 11500 | > loss: -0.39388 (-0.40401) | > log_mle: -0.50974 (-0.53282) | > loss_dur: 0.11585 (0.12881) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.94490 (6.34949) | > current_lr: 0.00059 | > step_time: 1.14360 (0.68539) | > loader_time: 0.00470 (0.00355) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03337 (+0.00507) | > avg_loss: -0.41132 (-0.01853) | > avg_log_mle: -0.53755 (-0.01806) | > avg_loss_dur: 0.12623 (-0.00047) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_11520.pth.tar  > EPOCH: 18/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 17:03:14)   --> STEP: 5/639 -- GLOBAL_STEP: 11525 | > loss: -0.43129 (-0.43238) | > log_mle: -0.60102 (-0.58280) | > loss_dur: 0.16973 (0.15042) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.45048 (7.23579) | > current_lr: 0.00059 | > step_time: 0.31700 (0.28061) | > loader_time: 0.00250 (0.00249)  --> STEP: 30/639 -- GLOBAL_STEP: 11550 | > loss: -0.44213 (-0.41222) | > log_mle: -0.57239 (-0.55619) | > loss_dur: 0.13026 (0.14397) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.39197 (6.87666) | > current_lr: 0.00059 | > step_time: 0.34890 (0.32990) | > loader_time: 0.00270 (0.00256)  --> STEP: 55/639 -- GLOBAL_STEP: 11575 | > loss: -0.36117 (-0.40487) | > log_mle: -0.48263 (-0.54601) | > loss_dur: 0.12146 (0.14114) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.81752 (6.48114) | > current_lr: 0.00059 | > step_time: 0.41430 (0.35387) | > loader_time: 0.00290 (0.00265)  --> STEP: 80/639 -- GLOBAL_STEP: 11600 | > loss: -0.39674 (-0.40343) | > log_mle: -0.53700 (-0.54164) | > loss_dur: 0.14025 (0.13821) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.64480 (6.40701) | > current_lr: 0.00059 | > step_time: 0.39240 (0.37116) | > loader_time: 0.00290 (0.00273)  --> STEP: 105/639 -- GLOBAL_STEP: 11625 | > loss: -0.35767 (-0.40487) | > log_mle: -0.49771 (-0.54183) | > loss_dur: 0.14004 (0.13697) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.04527 (6.35933) | > current_lr: 0.00059 | > step_time: 0.49840 (0.38876) | > loader_time: 0.00310 (0.00281)  --> STEP: 130/639 -- GLOBAL_STEP: 11650 | > loss: -0.39126 (-0.40698) | > log_mle: -0.54346 (-0.54282) | > loss_dur: 0.15219 (0.13584) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.96961 (6.48161) | > current_lr: 0.00059 | > step_time: 0.57700 (0.40371) | > loader_time: 0.00300 (0.00285)  --> STEP: 155/639 -- GLOBAL_STEP: 11675 | > loss: -0.40047 (-0.40859) | > log_mle: -0.53268 (-0.54339) | > loss_dur: 0.13221 (0.13480) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.50946 (6.61692) | > current_lr: 0.00059 | > step_time: 0.53960 (0.42016) | > loader_time: 0.00320 (0.00288)  --> STEP: 180/639 -- GLOBAL_STEP: 11700 | > loss: -0.45170 (-0.41219) | > log_mle: -0.59193 (-0.54617) | > loss_dur: 0.14023 (0.13398) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.92769 (7.33466) | > current_lr: 0.00059 | > step_time: 0.55440 (0.43600) | > loader_time: 0.00340 (0.00292)  --> STEP: 205/639 -- GLOBAL_STEP: 11725 | > loss: -0.39068 (-0.41108) | > log_mle: -0.51195 (-0.54406) | > loss_dur: 0.12128 (0.13298) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.91714 (7.16877) | > current_lr: 0.00058 | > step_time: 0.64500 (0.45017) | > loader_time: 0.00330 (0.00296)  --> STEP: 230/639 -- GLOBAL_STEP: 11750 | > loss: -0.43209 (-0.41287) | > log_mle: -0.54991 (-0.54498) | > loss_dur: 0.11782 (0.13211) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.74564 (7.24542) | > current_lr: 0.00058 | > step_time: 0.55360 (0.46678) | > loader_time: 0.00350 (0.00300)  --> STEP: 255/639 -- GLOBAL_STEP: 11775 | > loss: -0.38766 (-0.41408) | > log_mle: -0.52288 (-0.54539) | > loss_dur: 0.13522 (0.13131) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.85802 (7.23681) | > current_lr: 0.00058 | > step_time: 0.64420 (0.48165) | > loader_time: 0.00340 (0.00302)  --> STEP: 280/639 -- GLOBAL_STEP: 11800 | > loss: -0.40816 (-0.41557) | > log_mle: -0.54721 (-0.54673) | > loss_dur: 0.13906 (0.13116) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.11678 (7.41749) | > current_lr: 0.00058 | > step_time: 0.58890 (0.49860) | > loader_time: 0.00300 (0.00306)  --> STEP: 305/639 -- GLOBAL_STEP: 11825 | > loss: -0.40320 (-0.41534) | > log_mle: -0.53203 (-0.54628) | > loss_dur: 0.12884 (0.13094) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.17348 (7.37397) | > current_lr: 0.00058 | > step_time: 0.63810 (0.51246) | > loader_time: 0.00350 (0.00309)  --> STEP: 330/639 -- GLOBAL_STEP: 11850 | > loss: -0.41173 (-0.41450) | > log_mle: -0.55505 (-0.54546) | > loss_dur: 0.14332 (0.13096) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.50686 (7.35249) | > current_lr: 0.00058 | > step_time: 0.72520 (0.52720) | > loader_time: 0.00400 (0.00313)  --> STEP: 355/639 -- GLOBAL_STEP: 11875 | > loss: -0.36779 (-0.41414) | > log_mle: -0.48256 (-0.54460) | > loss_dur: 0.11477 (0.13046) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.75204 (7.25601) | > current_lr: 0.00058 | > step_time: 0.58870 (0.53959) | > loader_time: 0.00340 (0.00316)  --> STEP: 380/639 -- GLOBAL_STEP: 11900 | > loss: -0.40542 (-0.41336) | > log_mle: -0.51525 (-0.54338) | > loss_dur: 0.10983 (0.13002) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.27102 (7.10643) | > current_lr: 0.00058 | > step_time: 0.67950 (0.55405) | > loader_time: 0.00350 (0.00318)  --> STEP: 405/639 -- GLOBAL_STEP: 11925 | > loss: -0.40834 (-0.41286) | > log_mle: -0.52468 (-0.54284) | > loss_dur: 0.11634 (0.12998) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.51825 (7.03988) | > current_lr: 0.00058 | > step_time: 0.78490 (0.56631) | > loader_time: 0.00350 (0.00321)  --> STEP: 430/639 -- GLOBAL_STEP: 11950 | > loss: -0.36942 (-0.41201) | > log_mle: -0.48563 (-0.54181) | > loss_dur: 0.11621 (0.12980) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.73658 (6.97253) | > current_lr: 0.00058 | > step_time: 0.76890 (0.57888) | > loader_time: 0.00380 (0.00324)  --> STEP: 455/639 -- GLOBAL_STEP: 11975 | > loss: -0.38807 (-0.41127) | > log_mle: -0.51392 (-0.54066) | > loss_dur: 0.12585 (0.12939) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.00666 (6.87065) | > current_lr: 0.00058 | > step_time: 0.74050 (0.58911) | > loader_time: 0.00340 (0.00327)  --> STEP: 480/639 -- GLOBAL_STEP: 12000 | > loss: -0.34785 (-0.41057) | > log_mle: -0.46834 (-0.53990) | > loss_dur: 0.12049 (0.12934) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.77152 (6.82334) | > current_lr: 0.00058 | > step_time: 0.79650 (0.60181) | > loader_time: 0.00340 (0.00329)  --> STEP: 505/639 -- GLOBAL_STEP: 12025 | > loss: -0.42165 (-0.41024) | > log_mle: -0.53874 (-0.53922) | > loss_dur: 0.11709 (0.12897) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.04019 (6.78894) | > current_lr: 0.00058 | > step_time: 0.85660 (0.61434) | > loader_time: 0.00410 (0.00332)  --> STEP: 530/639 -- GLOBAL_STEP: 12050 | > loss: -0.39439 (-0.40996) | > log_mle: -0.52979 (-0.53878) | > loss_dur: 0.13540 (0.12882) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.87642 (6.73075) | > current_lr: 0.00058 | > step_time: 0.95250 (0.62805) | > loader_time: 0.00400 (0.00335)  --> STEP: 555/639 -- GLOBAL_STEP: 12075 | > loss: -0.38921 (-0.40941) | > log_mle: -0.51133 (-0.53797) | > loss_dur: 0.12212 (0.12856) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.19444 (6.67029) | > current_lr: 0.00058 | > step_time: 0.83550 (0.64167) | > loader_time: 0.00380 (0.00338)  --> STEP: 580/639 -- GLOBAL_STEP: 12100 | > loss: -0.41675 (-0.40911) | > log_mle: -0.53124 (-0.53748) | > loss_dur: 0.11449 (0.12837) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.64815 (6.65343) | > current_lr: 0.00058 | > step_time: 0.98680 (0.65720) | > loader_time: 0.00410 (0.00342)  --> STEP: 605/639 -- GLOBAL_STEP: 12125 | > loss: -0.38273 (-0.40882) | > log_mle: -0.49629 (-0.53708) | > loss_dur: 0.11356 (0.12827) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.62448 (6.62363) | > current_lr: 0.00057 | > step_time: 1.03230 (0.67329) | > loader_time: 0.00420 (0.00345)  --> STEP: 630/639 -- GLOBAL_STEP: 12150 | > loss: -0.40125 (-0.40859) | > log_mle: -0.53419 (-0.53678) | > loss_dur: 0.13294 (0.12819) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.21238 (6.60000) | > current_lr: 0.00057 | > step_time: 1.17900 (0.69077) | > loader_time: 0.00460 (0.00349) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.01619 (-0.01718) | > avg_loss: -0.40870 (+0.00262) | > avg_log_mle: -0.53525 (+0.00229) | > avg_loss_dur: 0.12655 (+0.00032)  > EPOCH: 19/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 17:10:54)   --> STEP: 15/639 -- GLOBAL_STEP: 12175 | > loss: -0.53225 (-0.43293) | > log_mle: -0.69006 (-0.57390) | > loss_dur: 0.15780 (0.14097) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.01658 (6.63875) | > current_lr: 0.00057 | > step_time: 0.32820 (0.30553) | > loader_time: 0.00260 (0.00283)  --> STEP: 40/639 -- GLOBAL_STEP: 12200 | > loss: -0.38203 (-0.40916) | > log_mle: -0.53038 (-0.54884) | > loss_dur: 0.14834 (0.13968) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.13543 (6.44523) | > current_lr: 0.00057 | > step_time: 0.36040 (0.33189) | > loader_time: 0.00280 (0.00278)  --> STEP: 65/639 -- GLOBAL_STEP: 12225 | > loss: -0.41835 (-0.40586) | > log_mle: -0.53161 (-0.54522) | > loss_dur: 0.11326 (0.13936) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.42471 (6.68674) | > current_lr: 0.00057 | > step_time: 0.40950 (0.36320) | > loader_time: 0.00290 (0.00283)  --> STEP: 90/639 -- GLOBAL_STEP: 12250 | > loss: -0.35054 (-0.40348) | > log_mle: -0.48888 (-0.53986) | > loss_dur: 0.13833 (0.13638) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.88566 (6.23883) | > current_lr: 0.00057 | > step_time: 0.51720 (0.37830) | > loader_time: 0.00310 (0.00286)  --> STEP: 115/639 -- GLOBAL_STEP: 12275 | > loss: -0.32023 (-0.40758) | > log_mle: -0.45704 (-0.54304) | > loss_dur: 0.13680 (0.13545) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.13788 (6.54416) | > current_lr: 0.00057 | > step_time: 0.38460 (0.39616) | > loader_time: 0.00310 (0.00288)  --> STEP: 140/639 -- GLOBAL_STEP: 12300 | > loss: -0.39777 (-0.41124) | > log_mle: -0.54790 (-0.54556) | > loss_dur: 0.15013 (0.13433) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.54721 (6.88796) | > current_lr: 0.00057 | > step_time: 0.56130 (0.40954) | > loader_time: 0.00300 (0.00290)  --> STEP: 165/639 -- GLOBAL_STEP: 12325 | > loss: -0.41240 (-0.41332) | > log_mle: -0.55350 (-0.54644) | > loss_dur: 0.14109 (0.13312) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.27050 (6.91185) | > current_lr: 0.00057 | > step_time: 0.49450 (0.42376) | > loader_time: 0.00300 (0.00292)  --> STEP: 190/639 -- GLOBAL_STEP: 12350 | > loss: -0.35114 (-0.41504) | > log_mle: -0.48233 (-0.54736) | > loss_dur: 0.13119 (0.13231) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.49408 (6.84780) | > current_lr: 0.00057 | > step_time: 0.55170 (0.44179) | > loader_time: 0.00340 (0.00295)  --> STEP: 215/639 -- GLOBAL_STEP: 12375 | > loss: -0.48096 (-0.41516) | > log_mle: -0.59521 (-0.54700) | > loss_dur: 0.11425 (0.13184) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.29707 (6.86786) | > current_lr: 0.00057 | > step_time: 0.54440 (0.45889) | > loader_time: 0.00300 (0.00298)  --> STEP: 240/639 -- GLOBAL_STEP: 12400 | > loss: -0.45767 (-0.41817) | > log_mle: -0.59567 (-0.54900) | > loss_dur: 0.13799 (0.13083) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.09130 (6.96721) | > current_lr: 0.00057 | > step_time: 0.57200 (0.47434) | > loader_time: 0.00320 (0.00301)  --> STEP: 265/639 -- GLOBAL_STEP: 12425 | > loss: -0.48216 (-0.41911) | > log_mle: -0.61539 (-0.54970) | > loss_dur: 0.13323 (0.13059) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.87994 (7.26683) | > current_lr: 0.00057 | > step_time: 0.61420 (0.49004) | > loader_time: 0.00350 (0.00305)  --> STEP: 290/639 -- GLOBAL_STEP: 12450 | > loss: -0.41105 (-0.41865) | > log_mle: -0.53093 (-0.54913) | > loss_dur: 0.11987 (0.13049) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.49552 (7.12909) | > current_lr: 0.00057 | > step_time: 0.61680 (0.50470) | > loader_time: 0.00330 (0.00308)  --> STEP: 315/639 -- GLOBAL_STEP: 12475 | > loss: -0.38613 (-0.41841) | > log_mle: -0.50165 (-0.54858) | > loss_dur: 0.11552 (0.13017) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.69508 (7.17764) | > current_lr: 0.00057 | > step_time: 0.62870 (0.51899) | > loader_time: 0.00350 (0.00311)  --> STEP: 340/639 -- GLOBAL_STEP: 12500 | > loss: -0.42249 (-0.41807) | > log_mle: -0.53369 (-0.54837) | > loss_dur: 0.11120 (0.13030) | > amp_scaler: 4096.00000 (2066.07059) | > grad_norm: 6.12789 (7.31616) | > current_lr: 0.00057 | > step_time: 0.60850 (0.53430) | > loader_time: 0.00340 (0.00315)  --> STEP: 365/639 -- GLOBAL_STEP: 12525 | > loss: -0.42004 (-0.41741) | > log_mle: -0.53956 (-0.54704) | > loss_dur: 0.11953 (0.12964) | > amp_scaler: 4096.00000 (2205.10685) | > grad_norm: 7.23775 (7.17159) | > current_lr: 0.00057 | > step_time: 0.80620 (0.54591) | > loader_time: 0.00380 (0.00317)  --> STEP: 390/639 -- GLOBAL_STEP: 12550 | > loss: -0.40572 (-0.41670) | > log_mle: -0.53776 (-0.54607) | > loss_dur: 0.13204 (0.12936) | > amp_scaler: 4096.00000 (2326.31795) | > grad_norm: 7.41308 (7.06882) | > current_lr: 0.00056 | > step_time: 0.73460 (0.55900) | > loader_time: 0.00360 (0.00320)  --> STEP: 415/639 -- GLOBAL_STEP: 12575 | > loss: -0.38099 (-0.41596) | > log_mle: -0.50955 (-0.54532) | > loss_dur: 0.12857 (0.12936) | > amp_scaler: 4096.00000 (2432.92530) | > grad_norm: 5.00654 (7.01555) | > current_lr: 0.00056 | > step_time: 0.73480 (0.57141) | > loader_time: 0.00340 (0.00322)  --> STEP: 440/639 -- GLOBAL_STEP: 12600 | > loss: -0.37191 (-0.41542) | > log_mle: -0.49787 (-0.54453) | > loss_dur: 0.12596 (0.12911) | > amp_scaler: 4096.00000 (2527.41818) | > grad_norm: 3.32780 (6.93296) | > current_lr: 0.00056 | > step_time: 0.72250 (0.58366) | > loader_time: 0.00370 (0.00325)  --> STEP: 465/639 -- GLOBAL_STEP: 12625 | > loss: -0.37863 (-0.41452) | > log_mle: -0.49983 (-0.54334) | > loss_dur: 0.12120 (0.12881) | > amp_scaler: 4096.00000 (2611.75054) | > grad_norm: 5.72071 (6.85515) | > current_lr: 0.00056 | > step_time: 0.79100 (0.59477) | > loader_time: 0.00390 (0.00327)  --> STEP: 490/639 -- GLOBAL_STEP: 12650 | > loss: -0.37306 (-0.41384) | > log_mle: -0.48615 (-0.54255) | > loss_dur: 0.11309 (0.12871) | > amp_scaler: 4096.00000 (2687.47755) | > grad_norm: 1.43494 (6.80752) | > current_lr: 0.00056 | > step_time: 0.78310 (0.60799) | > loader_time: 0.00320 (0.00330)  --> STEP: 515/639 -- GLOBAL_STEP: 12675 | > loss: -0.39387 (-0.41381) | > log_mle: -0.51846 (-0.54222) | > loss_dur: 0.12459 (0.12842) | > amp_scaler: 4096.00000 (2755.85243) | > grad_norm: 3.76838 (6.83691) | > current_lr: 0.00056 | > step_time: 0.94350 (0.62180) | > loader_time: 0.00410 (0.00333)  --> STEP: 540/639 -- GLOBAL_STEP: 12700 | > loss: -0.41258 (-0.41342) | > log_mle: -0.53711 (-0.54164) | > loss_dur: 0.12454 (0.12822) | > amp_scaler: 4096.00000 (2817.89630) | > grad_norm: 7.97943 (6.81782) | > current_lr: 0.00056 | > step_time: 0.87440 (0.63529) | > loader_time: 0.00430 (0.00335)  --> STEP: 565/639 -- GLOBAL_STEP: 12725 | > loss: -0.36921 (-0.41270) | > log_mle: -0.49480 (-0.54078) | > loss_dur: 0.12559 (0.12808) | > amp_scaler: 4096.00000 (2874.44956) | > grad_norm: 2.39193 (6.76944) | > current_lr: 0.00056 | > step_time: 1.07480 (0.64909) | > loader_time: 0.00410 (0.00339)  --> STEP: 590/639 -- GLOBAL_STEP: 12750 | > loss: -0.40324 (-0.41255) | > log_mle: -0.52939 (-0.54046) | > loss_dur: 0.12615 (0.12791) | > amp_scaler: 4096.00000 (2926.21017) | > grad_norm: 5.35476 (6.72670) | > current_lr: 0.00056 | > step_time: 0.95010 (0.66476) | > loader_time: 0.00430 (0.00342)  --> STEP: 615/639 -- GLOBAL_STEP: 12775 | > loss: -0.41273 (-0.41213) | > log_mle: -0.53482 (-0.53988) | > loss_dur: 0.12209 (0.12775) | > amp_scaler: 4096.00000 (2973.76260) | > grad_norm: 3.73956 (6.71807) | > current_lr: 0.00056 | > step_time: 1.05370 (0.68085) | > loader_time: 0.00420 (0.00346) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.01918 (+0.00300) | > avg_loss: -0.41408 (-0.00538) | > avg_log_mle: -0.53818 (-0.00293) | > avg_loss_dur: 0.12410 (-0.00245) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_12800.pth.tar  > EPOCH: 20/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 17:18:40)   --> STEP: 0/639 -- GLOBAL_STEP: 12800 | > loss: -0.45701 (-0.45701) | > log_mle: -0.63133 (-0.63133) | > loss_dur: 0.17431 (0.17431) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 17.75060 (17.75060) | > current_lr: 0.00056 | > step_time: 0.67700 (0.67701) | > loader_time: 0.35650 (0.35653)  --> STEP: 25/639 -- GLOBAL_STEP: 12825 | > loss: -0.42537 (-0.42400) | > log_mle: -0.57370 (-0.56480) | > loss_dur: 0.14833 (0.14080) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.91108 (8.17838) | > current_lr: 0.00056 | > step_time: 0.38810 (0.32387) | > loader_time: 0.00300 (0.00389)  --> STEP: 50/639 -- GLOBAL_STEP: 12850 | > loss: -0.38196 (-0.41406) | > log_mle: -0.50628 (-0.55152) | > loss_dur: 0.12432 (0.13745) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.57012 (7.27062) | > current_lr: 0.00056 | > step_time: 0.31790 (0.34969) | > loader_time: 0.00280 (0.00332)  --> STEP: 75/639 -- GLOBAL_STEP: 12875 | > loss: -0.42682 (-0.41105) | > log_mle: -0.54845 (-0.54708) | > loss_dur: 0.12163 (0.13603) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.56024 (6.73404) | > current_lr: 0.00056 | > step_time: 0.41270 (0.37206) | > loader_time: 0.00260 (0.00315)  --> STEP: 100/639 -- GLOBAL_STEP: 12900 | > loss: -0.43121 (-0.41316) | > log_mle: -0.55632 (-0.54705) | > loss_dur: 0.12510 (0.13389) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.02348 (6.57528) | > current_lr: 0.00056 | > step_time: 0.50910 (0.38980) | > loader_time: 0.00260 (0.00309)  --> STEP: 125/639 -- GLOBAL_STEP: 12925 | > loss: -0.44125 (-0.41516) | > log_mle: -0.56109 (-0.54835) | > loss_dur: 0.11984 (0.13319) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.44041 (7.10268) | > current_lr: 0.00056 | > step_time: 0.41350 (0.40391) | > loader_time: 0.00270 (0.00306)  --> STEP: 150/639 -- GLOBAL_STEP: 12950 | > loss: -0.39091 (-0.41736) | > log_mle: -0.52073 (-0.54963) | > loss_dur: 0.12982 (0.13227) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.53967 (7.28026) | > current_lr: 0.00056 | > step_time: 0.42600 (0.41756) | > loader_time: 0.00240 (0.00304)  --> STEP: 175/639 -- GLOBAL_STEP: 12975 | > loss: -0.42471 (-0.42034) | > log_mle: -0.53898 (-0.55180) | > loss_dur: 0.11427 (0.13145) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.07741 (7.56390) | > current_lr: 0.00056 | > step_time: 0.59680 (0.43284) | > loader_time: 0.00330 (0.00304)  --> STEP: 200/639 -- GLOBAL_STEP: 13000 | > loss: -0.40278 (-0.42022) | > log_mle: -0.53042 (-0.55086) | > loss_dur: 0.12765 (0.13064) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.76107 (7.41991) | > current_lr: 0.00056 | > step_time: 0.52370 (0.44714) | > loader_time: 0.00300 (0.00305)  --> STEP: 225/639 -- GLOBAL_STEP: 13025 | > loss: -0.39315 (-0.42041) | > log_mle: -0.50534 (-0.55012) | > loss_dur: 0.11220 (0.12971) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.38067 (7.29564) | > current_lr: 0.00055 | > step_time: 0.51220 (0.46312) | > loader_time: 0.00280 (0.00308)  --> STEP: 250/639 -- GLOBAL_STEP: 13050 | > loss: -0.40407 (-0.42311) | > log_mle: -0.51064 (-0.55234) | > loss_dur: 0.10657 (0.12923) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.23029 (7.35995) | > current_lr: 0.00055 | > step_time: 0.56600 (0.47787) | > loader_time: 0.00330 (0.00310)  --> STEP: 275/639 -- GLOBAL_STEP: 13075 | > loss: -0.42745 (-0.42459) | > log_mle: -0.55684 (-0.55367) | > loss_dur: 0.12940 (0.12907) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.27401 (7.46327) | > current_lr: 0.00055 | > step_time: 0.84990 (0.49440) | > loader_time: 0.00390 (0.00317)  --> STEP: 300/639 -- GLOBAL_STEP: 13100 | > loss: -0.42051 (-0.42358) | > log_mle: -0.54390 (-0.55267) | > loss_dur: 0.12340 (0.12909) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.70494 (7.38717) | > current_lr: 0.00055 | > step_time: 0.79550 (0.50859) | > loader_time: 0.00360 (0.00319)  --> STEP: 325/639 -- GLOBAL_STEP: 13125 | > loss: -0.41687 (-0.42302) | > log_mle: -0.54845 (-0.55220) | > loss_dur: 0.13157 (0.12917) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.13089 (7.36320) | > current_lr: 0.00055 | > step_time: 0.63000 (0.52378) | > loader_time: 0.00330 (0.00321)  --> STEP: 350/639 -- GLOBAL_STEP: 13150 | > loss: -0.44794 (-0.42244) | > log_mle: -0.57464 (-0.55126) | > loss_dur: 0.12670 (0.12882) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.19050 (7.38760) | > current_lr: 0.00055 | > step_time: 0.84500 (0.53831) | > loader_time: 0.00400 (0.00324)  --> STEP: 375/639 -- GLOBAL_STEP: 13175 | > loss: -0.39388 (-0.42123) | > log_mle: -0.51151 (-0.54966) | > loss_dur: 0.11763 (0.12843) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.30074 (7.22662) | > current_lr: 0.00055 | > step_time: 0.75740 (0.55165) | > loader_time: 0.00320 (0.00326)  --> STEP: 400/639 -- GLOBAL_STEP: 13200 | > loss: -0.42482 (-0.42085) | > log_mle: -0.54638 (-0.54913) | > loss_dur: 0.12155 (0.12829) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.90494 (7.15962) | > current_lr: 0.00055 | > step_time: 0.70980 (0.56377) | > loader_time: 0.00360 (0.00328)  --> STEP: 425/639 -- GLOBAL_STEP: 13225 | > loss: -0.47800 (-0.41992) | > log_mle: -0.59984 (-0.54800) | > loss_dur: 0.12183 (0.12809) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.53961 (7.06288) | > current_lr: 0.00055 | > step_time: 0.93970 (0.57637) | > loader_time: 0.00380 (0.00330)  --> STEP: 450/639 -- GLOBAL_STEP: 13250 | > loss: -0.39059 (-0.41927) | > log_mle: -0.50360 (-0.54708) | > loss_dur: 0.11300 (0.12781) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.86718 (6.99109) | > current_lr: 0.00055 | > step_time: 0.75830 (0.58873) | > loader_time: 0.00360 (0.00332)  --> STEP: 475/639 -- GLOBAL_STEP: 13275 | > loss: -0.40446 (-0.41847) | > log_mle: -0.51467 (-0.54620) | > loss_dur: 0.11021 (0.12773) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.31797 (6.90354) | > current_lr: 0.00055 | > step_time: 0.70160 (0.60008) | > loader_time: 0.00330 (0.00333)  --> STEP: 500/639 -- GLOBAL_STEP: 13300 | > loss: -0.45465 (-0.41820) | > log_mle: -0.58018 (-0.54573) | > loss_dur: 0.12553 (0.12753) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.56940 (6.84911) | > current_lr: 0.00055 | > step_time: 0.94320 (0.61376) | > loader_time: 0.00400 (0.00335)  --> STEP: 525/639 -- GLOBAL_STEP: 13325 | > loss: -0.42571 (-0.41791) | > log_mle: -0.56160 (-0.54528) | > loss_dur: 0.13589 (0.12736) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.16041 (6.79897) | > current_lr: 0.00055 | > step_time: 1.06550 (0.62712) | > loader_time: 0.00570 (0.00339)  --> STEP: 550/639 -- GLOBAL_STEP: 13350 | > loss: -0.37164 (-0.41734) | > log_mle: -0.49520 (-0.54450) | > loss_dur: 0.12356 (0.12715) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.21534 (6.72321) | > current_lr: 0.00055 | > step_time: 0.92430 (0.64145) | > loader_time: 0.00410 (0.00342)  --> STEP: 575/639 -- GLOBAL_STEP: 13375 | > loss: -0.38907 (-0.41673) | > log_mle: -0.51144 (-0.54376) | > loss_dur: 0.12237 (0.12704) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.33241 (6.70779) | > current_lr: 0.00055 | > step_time: 0.91400 (0.65564) | > loader_time: 0.00420 (0.00345)  --> STEP: 600/639 -- GLOBAL_STEP: 13400 | > loss: -0.38442 (-0.41630) | > log_mle: -0.51347 (-0.54328) | > loss_dur: 0.12904 (0.12698) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.94748 (6.78502) | > current_lr: 0.00055 | > step_time: 1.08510 (0.67291) | > loader_time: 0.00400 (0.00350)  --> STEP: 625/639 -- GLOBAL_STEP: 13425 | > loss: -0.41975 (-0.41602) | > log_mle: -0.54316 (-0.54287) | > loss_dur: 0.12341 (0.12685) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.92664 (6.78674) | > current_lr: 0.00055 | > step_time: 1.10390 (0.68979) | > loader_time: 0.00420 (0.00355) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03097 (+0.01179) | > avg_loss: -0.41345 (+0.00063) | > avg_log_mle: -0.54019 (-0.00201) | > avg_loss_dur: 0.12674 (+0.00263)  > EPOCH: 21/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 17:26:22)   --> STEP: 10/639 -- GLOBAL_STEP: 13450 | > loss: -0.44716 (-0.44266) | > log_mle: -0.59696 (-0.58653) | > loss_dur: 0.14980 (0.14387) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.33972 (8.26220) | > current_lr: 0.00055 | > step_time: 0.28260 (0.29768) | > loader_time: 0.00230 (0.00689)  --> STEP: 35/639 -- GLOBAL_STEP: 13475 | > loss: -0.41718 (-0.42590) | > log_mle: -0.53993 (-0.56488) | > loss_dur: 0.12275 (0.13898) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.76062 (6.63166) | > current_lr: 0.00055 | > step_time: 0.32030 (0.32606) | > loader_time: 0.00270 (0.00388)  --> STEP: 60/639 -- GLOBAL_STEP: 13500 | > loss: -0.37374 (-0.41917) | > log_mle: -0.50270 (-0.55707) | > loss_dur: 0.12896 (0.13790) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.42958 (6.91842) | > current_lr: 0.00054 | > step_time: 0.34180 (0.35367) | > loader_time: 0.00270 (0.00342)  --> STEP: 85/639 -- GLOBAL_STEP: 13525 | > loss: -0.36895 (-0.41576) | > log_mle: -0.50212 (-0.55062) | > loss_dur: 0.13317 (0.13486) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.24342 (7.03958) | > current_lr: 0.00054 | > step_time: 0.41320 (0.36886) | > loader_time: 0.00300 (0.00328)  --> STEP: 110/639 -- GLOBAL_STEP: 13550 | > loss: -0.40105 (-0.42110) | > log_mle: -0.53367 (-0.55533) | > loss_dur: 0.13262 (0.13422) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.22843 (7.40322) | > current_lr: 0.00054 | > step_time: 0.40410 (0.38922) | > loader_time: 0.00310 (0.00323)  --> STEP: 135/639 -- GLOBAL_STEP: 13575 | > loss: -0.36421 (-0.42189) | > log_mle: -0.51165 (-0.55500) | > loss_dur: 0.14745 (0.13310) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.54115 (7.32466) | > current_lr: 0.00054 | > step_time: 0.46870 (0.40168) | > loader_time: 0.00290 (0.00318)  --> STEP: 160/639 -- GLOBAL_STEP: 13600 | > loss: -0.44985 (-0.42296) | > log_mle: -0.56171 (-0.55468) | > loss_dur: 0.11186 (0.13172) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.14694 (7.26730) | > current_lr: 0.00054 | > step_time: 0.47030 (0.41688) | > loader_time: 0.00300 (0.00316)  --> STEP: 185/639 -- GLOBAL_STEP: 13625 | > loss: -0.41472 (-0.42677) | > log_mle: -0.53883 (-0.55823) | > loss_dur: 0.12411 (0.13146) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 22.31110 (7.71447) | > current_lr: 0.00054 | > step_time: 0.58750 (0.43458) | > loader_time: 0.00330 (0.00318)  --> STEP: 210/639 -- GLOBAL_STEP: 13650 | > loss: -0.34153 (-0.42513) | > log_mle: -0.46085 (-0.55612) | > loss_dur: 0.11932 (0.13099) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.94017 (7.56957) | > current_lr: 0.00054 | > step_time: 0.52450 (0.45167) | > loader_time: 0.00310 (0.00318)  --> STEP: 235/639 -- GLOBAL_STEP: 13675 | > loss: -0.44520 (-0.42684) | > log_mle: -0.55647 (-0.55698) | > loss_dur: 0.11127 (0.13014) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.60093 (7.61694) | > current_lr: 0.00054 | > step_time: 0.57180 (0.46785) | > loader_time: 0.00340 (0.00320)  --> STEP: 260/639 -- GLOBAL_STEP: 13700 | > loss: -0.43424 (-0.42793) | > log_mle: -0.55563 (-0.55760) | > loss_dur: 0.12139 (0.12967) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.54130 (7.75808) | > current_lr: 0.00054 | > step_time: 0.57410 (0.48260) | > loader_time: 0.00310 (0.00322)  --> STEP: 285/639 -- GLOBAL_STEP: 13725 | > loss: -0.36897 (-0.42799) | > log_mle: -0.50358 (-0.55750) | > loss_dur: 0.13461 (0.12951) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.42446 (7.86492) | > current_lr: 0.00054 | > step_time: 0.67410 (0.49894) | > loader_time: 0.00330 (0.00324)  --> STEP: 310/639 -- GLOBAL_STEP: 13750 | > loss: -0.42549 (-0.42740) | > log_mle: -0.55308 (-0.55679) | > loss_dur: 0.12759 (0.12939) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.85321 (7.88526) | > current_lr: 0.00054 | > step_time: 0.58090 (0.51360) | > loader_time: 0.00340 (0.00325)  --> STEP: 335/639 -- GLOBAL_STEP: 13775 | > loss: -0.44410 (-0.42625) | > log_mle: -0.57369 (-0.55570) | > loss_dur: 0.12959 (0.12945) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.57752 (7.77277) | > current_lr: 0.00054 | > step_time: 0.67550 (0.52852) | > loader_time: 0.00330 (0.00328)  --> STEP: 360/639 -- GLOBAL_STEP: 13800 | > loss: -0.38603 (-0.42568) | > log_mle: -0.50923 (-0.55465) | > loss_dur: 0.12320 (0.12897) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.96202 (7.74800) | > current_lr: 0.00054 | > step_time: 0.66100 (0.54031) | > loader_time: 0.00370 (0.00330)  --> STEP: 385/639 -- GLOBAL_STEP: 13825 | > loss: -0.41809 (-0.42500) | > log_mle: -0.54455 (-0.55358) | > loss_dur: 0.12646 (0.12859) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.04201 (7.65153) | > current_lr: 0.00054 | > step_time: 0.70210 (0.55427) | > loader_time: 0.00380 (0.00332)  --> STEP: 410/639 -- GLOBAL_STEP: 13850 | > loss: -0.37620 (-0.42419) | > log_mle: -0.52139 (-0.55272) | > loss_dur: 0.14519 (0.12854) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.56198 (7.59508) | > current_lr: 0.00054 | > step_time: 0.80440 (0.56653) | > loader_time: 0.00370 (0.00334)  --> STEP: 435/639 -- GLOBAL_STEP: 13875 | > loss: -0.41370 (-0.42330) | > log_mle: -0.52887 (-0.55152) | > loss_dur: 0.11516 (0.12822) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.44335 (7.48173) | > current_lr: 0.00054 | > step_time: 0.69410 (0.57927) | > loader_time: 0.00350 (0.00337)  --> STEP: 460/639 -- GLOBAL_STEP: 13900 | > loss: -0.37643 (-0.42238) | > log_mle: -0.49738 (-0.55026) | > loss_dur: 0.12096 (0.12788) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.82386 (7.34151) | > current_lr: 0.00054 | > step_time: 0.90200 (0.59057) | > loader_time: 0.00380 (0.00338)  --> STEP: 485/639 -- GLOBAL_STEP: 13925 | > loss: -0.41434 (-0.42170) | > log_mle: -0.53439 (-0.54955) | > loss_dur: 0.12005 (0.12785) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.44386 (7.27600) | > current_lr: 0.00054 | > step_time: 0.88020 (0.60330) | > loader_time: 0.00410 (0.00341)  --> STEP: 510/639 -- GLOBAL_STEP: 13950 | > loss: -0.41010 (-0.42140) | > log_mle: -0.53774 (-0.54897) | > loss_dur: 0.12765 (0.12757) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.44378 (7.22862) | > current_lr: 0.00054 | > step_time: 0.87290 (0.61646) | > loader_time: 0.00390 (0.00343)  --> STEP: 535/639 -- GLOBAL_STEP: 13975 | > loss: -0.39713 (-0.42092) | > log_mle: -0.53100 (-0.54833) | > loss_dur: 0.13387 (0.12741) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.76080 (7.18827) | > current_lr: 0.00054 | > step_time: 0.90480 (0.63015) | > loader_time: 0.00360 (0.00346)  --> STEP: 560/639 -- GLOBAL_STEP: 14000 | > loss: -0.40411 (-0.42037) | > log_mle: -0.53931 (-0.54769) | > loss_dur: 0.13520 (0.12733) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.58099 (7.12121) | > current_lr: 0.00053 | > step_time: 0.90680 (0.64501) | > loader_time: 0.00400 (0.00349)  --> STEP: 585/639 -- GLOBAL_STEP: 14025 | > loss: -0.41969 (-0.42002) | > log_mle: -0.54077 (-0.54715) | > loss_dur: 0.12108 (0.12713) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.37719 (7.04558) | > current_lr: 0.00053 | > step_time: 1.05310 (0.66062) | > loader_time: 0.00450 (0.00352)  --> STEP: 610/639 -- GLOBAL_STEP: 14050 | > loss: -0.38463 (-0.41953) | > log_mle: -0.51072 (-0.54658) | > loss_dur: 0.12610 (0.12705) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.95034 (6.99320) | > current_lr: 0.00053 | > step_time: 1.03810 (0.67673) | > loader_time: 0.00430 (0.00356)  --> STEP: 635/639 -- GLOBAL_STEP: 14075 | > loss: -0.40919 (-0.41933) | > log_mle: -0.53288 (-0.54640) | > loss_dur: 0.12369 (0.12706) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.71871 (7.02256) | > current_lr: 0.00053 | > step_time: 1.12650 (0.69386) | > loader_time: 0.00400 (0.00359) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03475 (+0.00378) | > avg_loss: -0.41388 (-0.00043) | > avg_log_mle: -0.53821 (+0.00198) | > avg_loss_dur: 0.12434 (-0.00240)  > EPOCH: 22/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 17:34:01)   --> STEP: 20/639 -- GLOBAL_STEP: 14100 | > loss: -0.53049 (-0.44972) | > log_mle: -0.68979 (-0.58765) | > loss_dur: 0.15930 (0.13793) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.03419 (8.52168) | > current_lr: 0.00053 | > step_time: 0.34030 (0.30664) | > loader_time: 0.00280 (0.00316)  --> STEP: 45/639 -- GLOBAL_STEP: 14125 | > loss: -0.36572 (-0.42776) | > log_mle: -0.49633 (-0.56333) | > loss_dur: 0.13061 (0.13557) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.78527 (7.45189) | > current_lr: 0.00053 | > step_time: 0.47910 (0.34290) | > loader_time: 0.00300 (0.00294)  --> STEP: 70/639 -- GLOBAL_STEP: 14150 | > loss: -0.45096 (-0.42054) | > log_mle: -0.58252 (-0.55612) | > loss_dur: 0.13157 (0.13558) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.34653 (6.92452) | > current_lr: 0.00053 | > step_time: 0.41070 (0.36599) | > loader_time: 0.00320 (0.00296)  --> STEP: 95/639 -- GLOBAL_STEP: 14175 | > loss: -0.50246 (-0.42185) | > log_mle: -0.62405 (-0.55464) | > loss_dur: 0.12159 (0.13279) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.71618 (6.86524) | > current_lr: 0.00053 | > step_time: 0.41500 (0.38136) | > loader_time: 0.00300 (0.00295)  --> STEP: 120/639 -- GLOBAL_STEP: 14200 | > loss: -0.48095 (-0.42312) | > log_mle: -0.61965 (-0.55567) | > loss_dur: 0.13871 (0.13255) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.66733 (7.06106) | > current_lr: 0.00053 | > step_time: 0.54710 (0.39942) | > loader_time: 0.00310 (0.00297)  --> STEP: 145/639 -- GLOBAL_STEP: 14225 | > loss: -0.44188 (-0.42630) | > log_mle: -0.57346 (-0.55803) | > loss_dur: 0.13158 (0.13173) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.95895 (7.24290) | > current_lr: 0.00053 | > step_time: 0.46970 (0.41452) | > loader_time: 0.00300 (0.00300)  --> STEP: 170/639 -- GLOBAL_STEP: 14250 | > loss: -0.44178 (-0.42771) | > log_mle: -0.56767 (-0.55862) | > loss_dur: 0.12589 (0.13091) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.11052 (7.30151) | > current_lr: 0.00053 | > step_time: 0.54640 (0.42841) | > loader_time: 0.00340 (0.00302)  --> STEP: 195/639 -- GLOBAL_STEP: 14275 | > loss: -0.39689 (-0.42870) | > log_mle: -0.50515 (-0.55870) | > loss_dur: 0.10826 (0.12999) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.85363 (7.25599) | > current_lr: 0.00053 | > step_time: 0.49860 (0.44468) | > loader_time: 0.00340 (0.00305)  --> STEP: 220/639 -- GLOBAL_STEP: 14300 | > loss: -0.44473 (-0.42821) | > log_mle: -0.55310 (-0.55759) | > loss_dur: 0.10837 (0.12939) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.19398 (7.32718) | > current_lr: 0.00053 | > step_time: 0.64910 (0.46189) | > loader_time: 0.00340 (0.00309)  --> STEP: 245/639 -- GLOBAL_STEP: 14325 | > loss: -0.44121 (-0.43059) | > log_mle: -0.55097 (-0.55918) | > loss_dur: 0.10976 (0.12859) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.70997 (7.43415) | > current_lr: 0.00053 | > step_time: 0.52150 (0.47576) | > loader_time: 0.00320 (0.00311)  --> STEP: 270/639 -- GLOBAL_STEP: 14350 | > loss: -0.48327 (-0.43181) | > log_mle: -0.61841 (-0.56030) | > loss_dur: 0.13514 (0.12849) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 21.95212 (7.67971) | > current_lr: 0.00053 | > step_time: 0.81190 (0.49295) | > loader_time: 0.00370 (0.00315)  --> STEP: 295/639 -- GLOBAL_STEP: 14375 | > loss: -0.38885 (-0.43103) | > log_mle: -0.50497 (-0.55943) | > loss_dur: 0.11612 (0.12840) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.31406 (7.66230) | > current_lr: 0.00053 | > step_time: 0.76970 (0.50758) | > loader_time: 0.00370 (0.00318)  --> STEP: 320/639 -- GLOBAL_STEP: 14400 | > loss: -0.42135 (-0.43062) | > log_mle: -0.55954 (-0.55909) | > loss_dur: 0.13819 (0.12847) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.83214 (7.59448) | > current_lr: 0.00053 | > step_time: 0.72570 (0.52257) | > loader_time: 0.00370 (0.00321)  --> STEP: 345/639 -- GLOBAL_STEP: 14425 | > loss: -0.43184 (-0.42998) | > log_mle: -0.54744 (-0.55840) | > loss_dur: 0.11560 (0.12842) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.08613 (7.55361) | > current_lr: 0.00053 | > step_time: 0.64330 (0.53764) | > loader_time: 0.00310 (0.00323)  --> STEP: 370/639 -- GLOBAL_STEP: 14450 | > loss: -0.44005 (-0.42931) | > log_mle: -0.56340 (-0.55718) | > loss_dur: 0.12336 (0.12786) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.76232 (7.46115) | > current_lr: 0.00053 | > step_time: 0.87330 (0.54997) | > loader_time: 0.00390 (0.00325)  --> STEP: 395/639 -- GLOBAL_STEP: 14475 | > loss: -0.40493 (-0.42841) | > log_mle: -0.53485 (-0.55611) | > loss_dur: 0.12992 (0.12769) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.04214 (7.37594) | > current_lr: 0.00053 | > step_time: 0.71610 (0.56348) | > loader_time: 0.00350 (0.00328)  --> STEP: 420/639 -- GLOBAL_STEP: 14500 | > loss: -0.40091 (-0.42745) | > log_mle: -0.52937 (-0.55516) | > loss_dur: 0.12846 (0.12771) | > amp_scaler: 8192.00000 (4125.25714) | > grad_norm: 6.08379 (7.30169) | > current_lr: 0.00053 | > step_time: 0.73530 (0.57625) | > loader_time: 0.00360 (0.00330)  --> STEP: 445/639 -- GLOBAL_STEP: 14525 | > loss: -0.41509 (-0.42688) | > log_mle: -0.53491 (-0.55439) | > loss_dur: 0.11982 (0.12751) | > amp_scaler: 8192.00000 (4353.72584) | > grad_norm: 6.17405 (7.25223) | > current_lr: 0.00053 | > step_time: 0.73640 (0.58779) | > loader_time: 0.00360 (0.00333)  --> STEP: 470/639 -- GLOBAL_STEP: 14550 | > loss: -0.42469 (-0.42590) | > log_mle: -0.54320 (-0.55320) | > loss_dur: 0.11851 (0.12730) | > amp_scaler: 8192.00000 (4557.88936) | > grad_norm: 8.88032 (7.13535) | > current_lr: 0.00052 | > step_time: 0.83620 (0.59945) | > loader_time: 0.00380 (0.00335)  --> STEP: 495/639 -- GLOBAL_STEP: 14575 | > loss: -0.37601 (-0.42518) | > log_mle: -0.48337 (-0.55239) | > loss_dur: 0.10737 (0.12721) | > amp_scaler: 4096.00000 (4584.21010) | > grad_norm: 1.40789 (7.07868) | > current_lr: 0.00052 | > step_time: 0.89430 (0.61231) | > loader_time: 0.00350 (0.00337)  --> STEP: 520/639 -- GLOBAL_STEP: 14600 | > loss: -0.42366 (-0.42480) | > log_mle: -0.54084 (-0.55173) | > loss_dur: 0.11718 (0.12693) | > amp_scaler: 4096.00000 (4560.73846) | > grad_norm: 9.83570 (7.07296) | > current_lr: 0.00052 | > step_time: 0.79790 (0.62522) | > loader_time: 0.00380 (0.00340)  --> STEP: 545/639 -- GLOBAL_STEP: 14625 | > loss: -0.44052 (-0.42420) | > log_mle: -0.57067 (-0.55105) | > loss_dur: 0.13015 (0.12686) | > amp_scaler: 4096.00000 (4539.42018) | > grad_norm: 8.54680 (7.05916) | > current_lr: 0.00052 | > step_time: 0.96750 (0.64029) | > loader_time: 0.00390 (0.00343)  --> STEP: 570/639 -- GLOBAL_STEP: 14650 | > loss: -0.40031 (-0.42348) | > log_mle: -0.51288 (-0.55017) | > loss_dur: 0.11257 (0.12670) | > amp_scaler: 4096.00000 (4519.97193) | > grad_norm: 3.32262 (7.02069) | > current_lr: 0.00052 | > step_time: 0.84990 (0.65390) | > loader_time: 0.00400 (0.00346)  --> STEP: 595/639 -- GLOBAL_STEP: 14675 | > loss: -0.39650 (-0.42305) | > log_mle: -0.51220 (-0.54963) | > loss_dur: 0.11571 (0.12658) | > amp_scaler: 4096.00000 (4502.15798) | > grad_norm: 4.76933 (6.98421) | > current_lr: 0.00052 | > step_time: 1.06130 (0.66991) | > loader_time: 0.00410 (0.00349)  --> STEP: 620/639 -- GLOBAL_STEP: 14700 | > loss: -0.40563 (-0.42261) | > log_mle: -0.52434 (-0.54909) | > loss_dur: 0.11871 (0.12648) | > amp_scaler: 4096.00000 (4485.78065) | > grad_norm: 4.78468 (6.98906) | > current_lr: 0.00052 | > step_time: 1.16890 (0.68637) | > loader_time: 0.00470 (0.00353) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02274 (-0.01201) | > avg_loss: -0.41754 (-0.00366) | > avg_log_mle: -0.54319 (-0.00498) | > avg_loss_dur: 0.12566 (+0.00132) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_14720.pth.tar  > EPOCH: 23/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 17:41:48)   --> STEP: 5/639 -- GLOBAL_STEP: 14725 | > loss: -0.42782 (-0.44337) | > log_mle: -0.61110 (-0.59961) | > loss_dur: 0.18328 (0.15624) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.96522 (9.61614) | > current_lr: 0.00052 | > step_time: 0.30630 (0.28595) | > loader_time: 0.00140 (0.00259)  --> STEP: 30/639 -- GLOBAL_STEP: 14750 | > loss: -0.45350 (-0.43712) | > log_mle: -0.57692 (-0.57485) | > loss_dur: 0.12342 (0.13773) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.10553 (7.72696) | > current_lr: 0.00052 | > step_time: 0.34880 (0.33193) | > loader_time: 0.00260 (0.00268)  --> STEP: 55/639 -- GLOBAL_STEP: 14775 | > loss: -0.39414 (-0.42959) | > log_mle: -0.50362 (-0.56497) | > loss_dur: 0.10948 (0.13538) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.07559 (7.00058) | > current_lr: 0.00052 | > step_time: 0.32560 (0.35201) | > loader_time: 0.00290 (0.00274)  --> STEP: 80/639 -- GLOBAL_STEP: 14800 | > loss: -0.42488 (-0.42619) | > log_mle: -0.56299 (-0.55967) | > loss_dur: 0.13811 (0.13348) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.02828 (7.10336) | > current_lr: 0.00052 | > step_time: 0.39080 (0.37173) | > loader_time: 0.00280 (0.00280)  --> STEP: 105/639 -- GLOBAL_STEP: 14825 | > loss: -0.37701 (-0.42657) | > log_mle: -0.51510 (-0.55957) | > loss_dur: 0.13809 (0.13300) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.68461 (7.55690) | > current_lr: 0.00052 | > step_time: 0.41230 (0.38899) | > loader_time: 0.00290 (0.00286)  --> STEP: 130/639 -- GLOBAL_STEP: 14850 | > loss: -0.42212 (-0.42862) | > log_mle: -0.56223 (-0.56062) | > loss_dur: 0.14011 (0.13200) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.95541 (7.70869) | > current_lr: 0.00052 | > step_time: 0.48970 (0.40342) | > loader_time: 0.00310 (0.00289)  --> STEP: 155/639 -- GLOBAL_STEP: 14875 | > loss: -0.42415 (-0.43059) | > log_mle: -0.55535 (-0.56161) | > loss_dur: 0.13121 (0.13103) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.89355 (7.85844) | > current_lr: 0.00052 | > step_time: 0.53360 (0.41876) | > loader_time: 0.00320 (0.00292)  --> STEP: 180/639 -- GLOBAL_STEP: 14900 | > loss: -0.46477 (-0.43413) | > log_mle: -0.60266 (-0.56448) | > loss_dur: 0.13789 (0.13035) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.99954 (8.18573) | > current_lr: 0.00052 | > step_time: 0.61600 (0.43484) | > loader_time: 0.00340 (0.00294)  --> STEP: 205/639 -- GLOBAL_STEP: 14925 | > loss: -0.40904 (-0.43254) | > log_mle: -0.52715 (-0.56198) | > loss_dur: 0.11811 (0.12944) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.07638 (7.95350) | > current_lr: 0.00052 | > step_time: 0.55420 (0.44951) | > loader_time: 0.00330 (0.00298)  --> STEP: 230/639 -- GLOBAL_STEP: 14950 | > loss: -0.44983 (-0.43398) | > log_mle: -0.56451 (-0.56275) | > loss_dur: 0.11468 (0.12877) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.26982 (7.99555) | > current_lr: 0.00052 | > step_time: 0.56870 (0.46549) | > loader_time: 0.00340 (0.00301)  --> STEP: 255/639 -- GLOBAL_STEP: 14975 | > loss: -0.40745 (-0.43475) | > log_mle: -0.53962 (-0.56290) | > loss_dur: 0.13217 (0.12815) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.27523 (8.01351) | > current_lr: 0.00052 | > step_time: 0.63470 (0.47848) | > loader_time: 0.00340 (0.00304)  --> STEP: 280/639 -- GLOBAL_STEP: 15000 | > loss: -0.41817 (-0.43579) | > log_mle: -0.56007 (-0.56407) | > loss_dur: 0.14190 (0.12829) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.24694 (8.18998) | > current_lr: 0.00052 | > step_time: 0.65260 (0.49562) | > loader_time: 0.00350 (0.00308)  --> STEP: 305/639 -- GLOBAL_STEP: 15025 | > loss: -0.40896 (-0.43518) | > log_mle: -0.54190 (-0.56336) | > loss_dur: 0.13293 (0.12818) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.03136 (8.09225) | > current_lr: 0.00052 | > step_time: 0.68940 (0.50997) | > loader_time: 0.00340 (0.00311)  --> STEP: 330/639 -- GLOBAL_STEP: 15050 | > loss: -0.43176 (-0.43388) | > log_mle: -0.56947 (-0.56222) | > loss_dur: 0.13771 (0.12834) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.63818 (8.04647) | > current_lr: 0.00052 | > step_time: 0.73100 (0.52567) | > loader_time: 0.00370 (0.00314)  --> STEP: 355/639 -- GLOBAL_STEP: 15075 | > loss: -0.38164 (-0.43335) | > log_mle: -0.49570 (-0.56113) | > loss_dur: 0.11406 (0.12779) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.93270 (7.98337) | > current_lr: 0.00052 | > step_time: 0.59070 (0.53866) | > loader_time: 0.00330 (0.00318)  --> STEP: 380/639 -- GLOBAL_STEP: 15100 | > loss: -0.42336 (-0.43221) | > log_mle: -0.53172 (-0.55969) | > loss_dur: 0.10836 (0.12748) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.65967 (7.85464) | > current_lr: 0.00051 | > step_time: 0.62320 (0.55227) | > loader_time: 0.00350 (0.00321)  --> STEP: 405/639 -- GLOBAL_STEP: 15125 | > loss: -0.42295 (-0.43161) | > log_mle: -0.54046 (-0.55915) | > loss_dur: 0.11751 (0.12754) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.22544 (7.81167) | > current_lr: 0.00051 | > step_time: 0.69100 (0.56442) | > loader_time: 0.00360 (0.00323)  --> STEP: 430/639 -- GLOBAL_STEP: 15150 | > loss: -0.38345 (-0.43060) | > log_mle: -0.49847 (-0.55797) | > loss_dur: 0.11501 (0.12737) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.73898 (7.71543) | > current_lr: 0.00051 | > step_time: 0.74980 (0.57720) | > loader_time: 0.00380 (0.00326)  --> STEP: 455/639 -- GLOBAL_STEP: 15175 | > loss: -0.40026 (-0.42962) | > log_mle: -0.52365 (-0.55664) | > loss_dur: 0.12339 (0.12702) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.74514 (7.63499) | > current_lr: 0.00051 | > step_time: 0.76150 (0.58780) | > loader_time: 0.00360 (0.00328)  --> STEP: 480/639 -- GLOBAL_STEP: 15200 | > loss: -0.35695 (-0.42859) | > log_mle: -0.48160 (-0.55562) | > loss_dur: 0.12465 (0.12703) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.25408 (7.52490) | > current_lr: 0.00051 | > step_time: 0.81550 (0.60052) | > loader_time: 0.00350 (0.00331)  --> STEP: 505/639 -- GLOBAL_STEP: 15225 | > loss: -0.43114 (-0.42798) | > log_mle: -0.55089 (-0.55479) | > loss_dur: 0.11976 (0.12681) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.21929 (7.45997) | > current_lr: 0.00051 | > step_time: 0.92310 (0.61318) | > loader_time: 0.00410 (0.00333)  --> STEP: 530/639 -- GLOBAL_STEP: 15250 | > loss: -0.40878 (-0.42756) | > log_mle: -0.54007 (-0.55423) | > loss_dur: 0.13129 (0.12667) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.48786 (7.44895) | > current_lr: 0.00051 | > step_time: 0.93900 (0.62710) | > loader_time: 0.00410 (0.00337)  --> STEP: 555/639 -- GLOBAL_STEP: 15275 | > loss: -0.40448 (-0.42687) | > log_mle: -0.52552 (-0.55334) | > loss_dur: 0.12104 (0.12647) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.65202 (7.33245) | > current_lr: 0.00051 | > step_time: 0.83240 (0.64136) | > loader_time: 0.00400 (0.00340)  --> STEP: 580/639 -- GLOBAL_STEP: 15300 | > loss: -0.43319 (-0.42643) | > log_mle: -0.54514 (-0.55273) | > loss_dur: 0.11195 (0.12631) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.08089 (7.26363) | > current_lr: 0.00051 | > step_time: 1.12090 (0.65681) | > loader_time: 0.00420 (0.00344)  --> STEP: 605/639 -- GLOBAL_STEP: 15325 | > loss: -0.40095 (-0.42599) | > log_mle: -0.51196 (-0.55223) | > loss_dur: 0.11101 (0.12625) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.14806 (7.21259) | > current_lr: 0.00051 | > step_time: 1.02050 (0.67275) | > loader_time: 0.00430 (0.00348)  --> STEP: 630/639 -- GLOBAL_STEP: 15350 | > loss: -0.41921 (-0.42567) | > log_mle: -0.54688 (-0.55186) | > loss_dur: 0.12767 (0.12619) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.47420 (7.18373) | > current_lr: 0.00051 | > step_time: 1.22750 (0.69008) | > loader_time: 0.00450 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03916 (+0.01642) | > avg_loss: -0.41856 (-0.00103) | > avg_log_mle: -0.54452 (-0.00132) | > avg_loss_dur: 0.12595 (+0.00030) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_15360.pth.tar  > EPOCH: 24/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 17:49:34)   --> STEP: 15/639 -- GLOBAL_STEP: 15375 | > loss: -0.56025 (-0.46035) | > log_mle: -0.71767 (-0.59585) | > loss_dur: 0.15741 (0.13550) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.60468 (8.26074) | > current_lr: 0.00051 | > step_time: 0.32920 (0.30203) | > loader_time: 0.00270 (0.00263)  --> STEP: 40/639 -- GLOBAL_STEP: 15400 | > loss: -0.42348 (-0.43854) | > log_mle: -0.55053 (-0.57237) | > loss_dur: 0.12704 (0.13383) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.76591 (7.26937) | > current_lr: 0.00051 | > step_time: 0.34940 (0.33390) | > loader_time: 0.00270 (0.00262)  --> STEP: 65/639 -- GLOBAL_STEP: 15425 | > loss: -0.43208 (-0.43160) | > log_mle: -0.54904 (-0.56668) | > loss_dur: 0.11696 (0.13508) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.62172 (7.56556) | > current_lr: 0.00051 | > step_time: 0.38510 (0.36021) | > loader_time: 0.00320 (0.00271)  --> STEP: 90/639 -- GLOBAL_STEP: 15450 | > loss: -0.37039 (-0.42709) | > log_mle: -0.50513 (-0.55995) | > loss_dur: 0.13473 (0.13286) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.35070 (7.08160) | > current_lr: 0.00051 | > step_time: 0.55280 (0.37809) | > loader_time: 0.00310 (0.00273)  --> STEP: 115/639 -- GLOBAL_STEP: 15475 | > loss: -0.33255 (-0.43017) | > log_mle: -0.47352 (-0.56239) | > loss_dur: 0.14097 (0.13222) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.85772 (7.54315) | > current_lr: 0.00051 | > step_time: 0.47520 (0.39519) | > loader_time: 0.00260 (0.00276)  --> STEP: 140/639 -- GLOBAL_STEP: 15500 | > loss: -0.41172 (-0.43251) | > log_mle: -0.56189 (-0.56405) | > loss_dur: 0.15017 (0.13154) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.10693 (7.69600) | > current_lr: 0.00051 | > step_time: 0.54980 (0.40937) | > loader_time: 0.00300 (0.00278)  --> STEP: 165/639 -- GLOBAL_STEP: 15525 | > loss: -0.43021 (-0.43387) | > log_mle: -0.57173 (-0.56426) | > loss_dur: 0.14152 (0.13039) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.88132 (7.77374) | > current_lr: 0.00051 | > step_time: 0.49240 (0.42252) | > loader_time: 0.00300 (0.00282)  --> STEP: 190/639 -- GLOBAL_STEP: 15550 | > loss: -0.36440 (-0.43545) | > log_mle: -0.49717 (-0.56509) | > loss_dur: 0.13277 (0.12964) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.66994 (7.89591) | > current_lr: 0.00051 | > step_time: 0.52560 (0.43998) | > loader_time: 0.00290 (0.00287)  --> STEP: 215/639 -- GLOBAL_STEP: 15575 | > loss: -0.49561 (-0.43507) | > log_mle: -0.60776 (-0.56417) | > loss_dur: 0.11215 (0.12911) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.21238 (7.83876) | > current_lr: 0.00051 | > step_time: 0.54410 (0.45542) | > loader_time: 0.00330 (0.00291)  --> STEP: 240/639 -- GLOBAL_STEP: 15600 | > loss: -0.49115 (-0.43747) | > log_mle: -0.61811 (-0.56565) | > loss_dur: 0.12695 (0.12817) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.25031 (7.92414) | > current_lr: 0.00051 | > step_time: 0.55840 (0.47068) | > loader_time: 0.00320 (0.00295)  --> STEP: 265/639 -- GLOBAL_STEP: 15625 | > loss: -0.49502 (-0.43760) | > log_mle: -0.63175 (-0.56569) | > loss_dur: 0.13673 (0.12808) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 17.77027 (7.97144) | > current_lr: 0.00051 | > step_time: 0.61240 (0.48614) | > loader_time: 0.00330 (0.00299)  --> STEP: 290/639 -- GLOBAL_STEP: 15650 | > loss: -0.42151 (-0.43709) | > log_mle: -0.54319 (-0.56508) | > loss_dur: 0.12168 (0.12799) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.42161 (7.90086) | > current_lr: 0.00051 | > step_time: 0.63480 (0.50131) | > loader_time: 0.00340 (0.00303)  --> STEP: 315/639 -- GLOBAL_STEP: 15675 | > loss: -0.40044 (-0.43660) | > log_mle: -0.51591 (-0.56440) | > loss_dur: 0.11547 (0.12780) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.01984 (7.91891) | > current_lr: 0.00051 | > step_time: 0.67150 (0.51558) | > loader_time: 0.00330 (0.00307)  --> STEP: 340/639 -- GLOBAL_STEP: 15700 | > loss: -0.42943 (-0.43592) | > log_mle: -0.54442 (-0.56393) | > loss_dur: 0.11499 (0.12801) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.02715 (8.02371) | > current_lr: 0.00051 | > step_time: 0.60130 (0.53162) | > loader_time: 0.00340 (0.00311)  --> STEP: 365/639 -- GLOBAL_STEP: 15725 | > loss: -0.43355 (-0.43492) | > log_mle: -0.55120 (-0.56237) | > loss_dur: 0.11765 (0.12745) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.57482 (7.85884) | > current_lr: 0.00050 | > step_time: 0.81840 (0.54424) | > loader_time: 0.00350 (0.00313)  --> STEP: 390/639 -- GLOBAL_STEP: 15750 | > loss: -0.41856 (-0.43399) | > log_mle: -0.55317 (-0.56115) | > loss_dur: 0.13462 (0.12716) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.79811 (7.72030) | > current_lr: 0.00050 | > step_time: 0.71760 (0.55715) | > loader_time: 0.00350 (0.00317)  --> STEP: 415/639 -- GLOBAL_STEP: 15775 | > loss: -0.39708 (-0.43303) | > log_mle: -0.52479 (-0.56023) | > loss_dur: 0.12771 (0.12720) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.13854 (7.77531) | > current_lr: 0.00050 | > step_time: 0.76060 (0.57019) | > loader_time: 0.00340 (0.00320)  --> STEP: 440/639 -- GLOBAL_STEP: 15800 | > loss: -0.37951 (-0.43244) | > log_mle: -0.50711 (-0.55938) | > loss_dur: 0.12760 (0.12694) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.28876 (7.70382) | > current_lr: 0.00050 | > step_time: 0.72330 (0.58208) | > loader_time: 0.00370 (0.00322)  --> STEP: 465/639 -- GLOBAL_STEP: 15825 | > loss: -0.39521 (-0.43136) | > log_mle: -0.51459 (-0.55800) | > loss_dur: 0.11938 (0.12664) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.22654 (7.60116) | > current_lr: 0.00050 | > step_time: 0.88990 (0.59372) | > loader_time: 0.00390 (0.00325)  --> STEP: 490/639 -- GLOBAL_STEP: 15850 | > loss: -0.39231 (-0.43062) | > log_mle: -0.49898 (-0.55718) | > loss_dur: 0.10667 (0.12655) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.12450 (7.56402) | > current_lr: 0.00050 | > step_time: 0.80010 (0.60709) | > loader_time: 0.00350 (0.00327)  --> STEP: 515/639 -- GLOBAL_STEP: 15875 | > loss: -0.40952 (-0.43032) | > log_mle: -0.53061 (-0.55665) | > loss_dur: 0.12109 (0.12633) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.29503 (7.51021) | > current_lr: 0.00050 | > step_time: 0.97780 (0.62025) | > loader_time: 0.00420 (0.00330)  --> STEP: 540/639 -- GLOBAL_STEP: 15900 | > loss: -0.42910 (-0.42968) | > log_mle: -0.55103 (-0.55586) | > loss_dur: 0.12193 (0.12618) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.23444 (7.45744) | > current_lr: 0.00050 | > step_time: 0.96480 (0.63478) | > loader_time: 0.00420 (0.00333)  --> STEP: 565/639 -- GLOBAL_STEP: 15925 | > loss: -0.37954 (-0.42881) | > log_mle: -0.50628 (-0.55494) | > loss_dur: 0.12675 (0.12613) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.15629 (7.36618) | > current_lr: 0.00050 | > step_time: 1.07690 (0.64888) | > loader_time: 0.00410 (0.00337)  --> STEP: 590/639 -- GLOBAL_STEP: 15950 | > loss: -0.41745 (-0.42860) | > log_mle: -0.54262 (-0.55461) | > loss_dur: 0.12518 (0.12601) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.88663 (7.36554) | > current_lr: 0.00050 | > step_time: 1.04070 (0.66500) | > loader_time: 0.00740 (0.00342)  --> STEP: 615/639 -- GLOBAL_STEP: 15975 | > loss: -0.42251 (-0.42805) | > log_mle: -0.54532 (-0.55397) | > loss_dur: 0.12280 (0.12592) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.23251 (7.34282) | > current_lr: 0.00050 | > step_time: 1.01460 (0.68094) | > loader_time: 0.00430 (0.00348) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03834 (-0.00083) | > avg_loss: -0.42069 (-0.00213) | > avg_log_mle: -0.54716 (-0.00264) | > avg_loss_dur: 0.12647 (+0.00051) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_16000.pth.tar  > EPOCH: 25/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 17:57:20)   --> STEP: 0/639 -- GLOBAL_STEP: 16000 | > loss: -0.47957 (-0.47957) | > log_mle: -0.64829 (-0.64829) | > loss_dur: 0.16871 (0.16871) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.77087 (18.77087) | > current_lr: 0.00050 | > step_time: 0.70010 (0.70012) | > loader_time: 0.38370 (0.38373)  --> STEP: 25/639 -- GLOBAL_STEP: 16025 | > loss: -0.45263 (-0.45488) | > log_mle: -0.59635 (-0.59091) | > loss_dur: 0.14372 (0.13604) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.82535 (11.10795) | > current_lr: 0.00050 | > step_time: 0.39240 (0.32103) | > loader_time: 0.00290 (0.00262)  --> STEP: 50/639 -- GLOBAL_STEP: 16050 | > loss: -0.40277 (-0.43708) | > log_mle: -0.51665 (-0.57101) | > loss_dur: 0.11388 (0.13393) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.41449 (9.06113) | > current_lr: 0.00050 | > step_time: 0.31710 (0.34074) | > loader_time: 0.00280 (0.00272)  --> STEP: 75/639 -- GLOBAL_STEP: 16075 | > loss: -0.43835 (-0.43154) | > log_mle: -0.56142 (-0.56538) | > loss_dur: 0.12307 (0.13384) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.15725 (8.29821) | > current_lr: 0.00050 | > step_time: 0.42190 (0.36492) | > loader_time: 0.00300 (0.00279)  --> STEP: 100/639 -- GLOBAL_STEP: 16100 | > loss: -0.45254 (-0.43297) | > log_mle: -0.57865 (-0.56490) | > loss_dur: 0.12610 (0.13193) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.28689 (8.24337) | > current_lr: 0.00050 | > step_time: 0.48760 (0.38466) | > loader_time: 0.00310 (0.00284)  --> STEP: 125/639 -- GLOBAL_STEP: 16125 | > loss: -0.45132 (-0.43373) | > log_mle: -0.57148 (-0.56537) | > loss_dur: 0.12016 (0.13164) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.30798 (8.29193) | > current_lr: 0.00050 | > step_time: 0.41510 (0.39877) | > loader_time: 0.00300 (0.00289)  --> STEP: 150/639 -- GLOBAL_STEP: 16150 | > loss: -0.39334 (-0.43547) | > log_mle: -0.53345 (-0.56639) | > loss_dur: 0.14011 (0.13091) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.73213 (8.24311) | > current_lr: 0.00050 | > step_time: 0.44040 (0.41580) | > loader_time: 0.00300 (0.00292)  --> STEP: 175/639 -- GLOBAL_STEP: 16175 | > loss: -0.44574 (-0.43815) | > log_mle: -0.55340 (-0.56806) | > loss_dur: 0.10765 (0.12991) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.44162 (8.35619) | > current_lr: 0.00050 | > step_time: 0.55100 (0.43191) | > loader_time: 0.00340 (0.00296)  --> STEP: 200/639 -- GLOBAL_STEP: 16200 | > loss: -0.42100 (-0.43800) | > log_mle: -0.54277 (-0.56697) | > loss_dur: 0.12177 (0.12897) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.68223 (8.27903) | > current_lr: 0.00050 | > step_time: 0.52530 (0.44678) | > loader_time: 0.00320 (0.00298)  --> STEP: 225/639 -- GLOBAL_STEP: 16225 | > loss: -0.40685 (-0.43775) | > log_mle: -0.51565 (-0.56587) | > loss_dur: 0.10880 (0.12813) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.16364 (8.25294) | > current_lr: 0.00050 | > step_time: 0.52080 (0.46337) | > loader_time: 0.00310 (0.00302)  --> STEP: 250/639 -- GLOBAL_STEP: 16250 | > loss: -0.41893 (-0.44009) | > log_mle: -0.52553 (-0.56781) | > loss_dur: 0.10660 (0.12771) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.07698 (8.33080) | > current_lr: 0.00050 | > step_time: 0.57870 (0.47790) | > loader_time: 0.00330 (0.00305)  --> STEP: 275/639 -- GLOBAL_STEP: 16275 | > loss: -0.44785 (-0.44174) | > log_mle: -0.57440 (-0.56936) | > loss_dur: 0.12654 (0.12762) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.32779 (8.45528) | > current_lr: 0.00050 | > step_time: 0.83900 (0.49494) | > loader_time: 0.00380 (0.00308)  --> STEP: 300/639 -- GLOBAL_STEP: 16300 | > loss: -0.44442 (-0.44055) | > log_mle: -0.56071 (-0.56823) | > loss_dur: 0.11629 (0.12768) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.25277 (8.34682) | > current_lr: 0.00050 | > step_time: 0.67900 (0.50833) | > loader_time: 0.00370 (0.00312)  --> STEP: 325/639 -- GLOBAL_STEP: 16325 | > loss: -0.43285 (-0.43966) | > log_mle: -0.55967 (-0.56748) | > loss_dur: 0.12682 (0.12782) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.62416 (8.33734) | > current_lr: 0.00050 | > step_time: 0.61270 (0.52344) | > loader_time: 0.00350 (0.00315)  --> STEP: 350/639 -- GLOBAL_STEP: 16350 | > loss: -0.46097 (-0.43897) | > log_mle: -0.59066 (-0.56647) | > loss_dur: 0.12969 (0.12751) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.50148 (8.30832) | > current_lr: 0.00049 | > step_time: 0.85660 (0.53790) | > loader_time: 0.00380 (0.00318)  --> STEP: 375/639 -- GLOBAL_STEP: 16375 | > loss: -0.40674 (-0.43779) | > log_mle: -0.52493 (-0.56491) | > loss_dur: 0.11820 (0.12712) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.50043 (8.14254) | > current_lr: 0.00049 | > step_time: 0.67010 (0.55103) | > loader_time: 0.00330 (0.00321)  --> STEP: 400/639 -- GLOBAL_STEP: 16400 | > loss: -0.43671 (-0.43715) | > log_mle: -0.55531 (-0.56417) | > loss_dur: 0.11859 (0.12703) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.74688 (8.04068) | > current_lr: 0.00049 | > step_time: 0.77820 (0.56324) | > loader_time: 0.00380 (0.00323)  --> STEP: 425/639 -- GLOBAL_STEP: 16425 | > loss: -0.47807 (-0.43601) | > log_mle: -0.60635 (-0.56282) | > loss_dur: 0.12828 (0.12681) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 21.95077 (7.99925) | > current_lr: 0.00049 | > step_time: 0.89980 (0.57566) | > loader_time: 0.00390 (0.00326)  --> STEP: 450/639 -- GLOBAL_STEP: 16450 | > loss: -0.39923 (-0.43520) | > log_mle: -0.51464 (-0.56169) | > loss_dur: 0.11542 (0.12649) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.27328 (7.86863) | > current_lr: 0.00049 | > step_time: 0.79680 (0.58761) | > loader_time: 0.00370 (0.00329)  --> STEP: 475/639 -- GLOBAL_STEP: 16475 | > loss: -0.41317 (-0.43421) | > log_mle: -0.52260 (-0.56060) | > loss_dur: 0.10942 (0.12639) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.13624 (7.80574) | > current_lr: 0.00049 | > step_time: 0.72140 (0.59949) | > loader_time: 0.00370 (0.00331)  --> STEP: 500/639 -- GLOBAL_STEP: 16500 | > loss: -0.46343 (-0.43371) | > log_mle: -0.58645 (-0.55992) | > loss_dur: 0.12302 (0.12622) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.54562 (7.78303) | > current_lr: 0.00049 | > step_time: 0.95290 (0.61318) | > loader_time: 0.00380 (0.00334)  --> STEP: 525/639 -- GLOBAL_STEP: 16525 | > loss: -0.43574 (-0.43322) | > log_mle: -0.56964 (-0.55929) | > loss_dur: 0.13390 (0.12607) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.41169 (7.72876) | > current_lr: 0.00049 | > step_time: 1.05070 (0.62707) | > loader_time: 0.00430 (0.00337)  --> STEP: 550/639 -- GLOBAL_STEP: 16550 | > loss: -0.37981 (-0.43253) | > log_mle: -0.50396 (-0.55837) | > loss_dur: 0.12416 (0.12584) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.81735 (7.61460) | > current_lr: 0.00049 | > step_time: 0.96920 (0.64179) | > loader_time: 0.00430 (0.00340)  --> STEP: 575/639 -- GLOBAL_STEP: 16575 | > loss: -0.40105 (-0.43178) | > log_mle: -0.52243 (-0.55752) | > loss_dur: 0.12137 (0.12574) | > amp_scaler: 8192.00000 (4231.34609) | > grad_norm: 4.53701 (7.54798) | > current_lr: 0.00049 | > step_time: 0.90670 (0.65544) | > loader_time: 0.00410 (0.00344)  --> STEP: 600/639 -- GLOBAL_STEP: 16600 | > loss: -0.39112 (-0.43134) | > log_mle: -0.52033 (-0.55700) | > loss_dur: 0.12921 (0.12566) | > amp_scaler: 8192.00000 (4396.37333) | > grad_norm: 5.44956 (7.50098) | > current_lr: 0.00049 | > step_time: 1.05390 (0.67176) | > loader_time: 0.00450 (0.00348)  --> STEP: 625/639 -- GLOBAL_STEP: 16625 | > loss: -0.41850 (-0.43093) | > log_mle: -0.54711 (-0.55646) | > loss_dur: 0.12862 (0.12554) | > amp_scaler: 8192.00000 (4548.19840) | > grad_norm: 5.50249 (7.45179) | > current_lr: 0.00049 | > step_time: 1.04540 (0.68867) | > loader_time: 0.00450 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.01914 (-0.01920) | > avg_loss: -0.42741 (-0.00672) | > avg_log_mle: -0.55295 (-0.00579) | > avg_loss_dur: 0.12553 (-0.00094) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_16640.pth.tar  > EPOCH: 26/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 18:05:07)   --> STEP: 10/639 -- GLOBAL_STEP: 16650 | > loss: -0.44494 (-0.45466) | > log_mle: -0.59538 (-0.59483) | > loss_dur: 0.15044 (0.14017) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 8.64751 (9.84879) | > current_lr: 0.00049 | > step_time: 0.29250 (0.30053) | > loader_time: 0.00250 (0.00594)  --> STEP: 35/639 -- GLOBAL_STEP: 16675 | > loss: -0.44233 (-0.44248) | > log_mle: -0.55704 (-0.57604) | > loss_dur: 0.11471 (0.13356) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.19607 (7.48846) | > current_lr: 0.00049 | > step_time: 0.33770 (0.32541) | > loader_time: 0.00280 (0.00359)  --> STEP: 60/639 -- GLOBAL_STEP: 16700 | > loss: -0.39360 (-0.43750) | > log_mle: -0.51585 (-0.57094) | > loss_dur: 0.12225 (0.13344) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.70696 (7.65853) | > current_lr: 0.00049 | > step_time: 0.34180 (0.35241) | > loader_time: 0.00280 (0.00331)  --> STEP: 85/639 -- GLOBAL_STEP: 16725 | > loss: -0.38754 (-0.43400) | > log_mle: -0.51759 (-0.56498) | > loss_dur: 0.13006 (0.13098) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.18284 (7.35443) | > current_lr: 0.00049 | > step_time: 0.41530 (0.36699) | > loader_time: 0.00270 (0.00319)  --> STEP: 110/639 -- GLOBAL_STEP: 16750 | > loss: -0.42337 (-0.43837) | > log_mle: -0.54704 (-0.56908) | > loss_dur: 0.12368 (0.13071) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 4.56280 (7.65479) | > current_lr: 0.00049 | > step_time: 0.39760 (0.38707) | > loader_time: 0.00310 (0.00315)  --> STEP: 135/639 -- GLOBAL_STEP: 16775 | > loss: -0.38746 (-0.43888) | > log_mle: -0.52998 (-0.56876) | > loss_dur: 0.14252 (0.12988) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 3.08981 (7.66276) | > current_lr: 0.00049 | > step_time: 0.46900 (0.40023) | > loader_time: 0.00280 (0.00309)  --> STEP: 160/639 -- GLOBAL_STEP: 16800 | > loss: -0.46191 (-0.43966) | > log_mle: -0.57041 (-0.56828) | > loss_dur: 0.10851 (0.12862) | > amp_scaler: 4096.00000 (7577.60000) | > grad_norm: 7.55471 (7.51700) | > current_lr: 0.00049 | > step_time: 0.46390 (0.41545) | > loader_time: 0.00310 (0.00308)  --> STEP: 185/639 -- GLOBAL_STEP: 16825 | > loss: -0.43213 (-0.44302) | > log_mle: -0.55223 (-0.57163) | > loss_dur: 0.12011 (0.12861) | > amp_scaler: 4096.00000 (7107.11351) | > grad_norm: 9.85633 (7.91198) | > current_lr: 0.00049 | > step_time: 0.57390 (0.43207) | > loader_time: 0.00320 (0.00310)  --> STEP: 210/639 -- GLOBAL_STEP: 16850 | > loss: -0.35600 (-0.44150) | > log_mle: -0.47282 (-0.56972) | > loss_dur: 0.11682 (0.12822) | > amp_scaler: 4096.00000 (6748.64762) | > grad_norm: 1.24549 (7.84950) | > current_lr: 0.00049 | > step_time: 0.49850 (0.44773) | > loader_time: 0.00290 (0.00312)  --> STEP: 235/639 -- GLOBAL_STEP: 16875 | > loss: -0.45591 (-0.44339) | > log_mle: -0.57032 (-0.57070) | > loss_dur: 0.11441 (0.12731) | > amp_scaler: 4096.00000 (6466.45106) | > grad_norm: 8.84344 (8.00057) | > current_lr: 0.00049 | > step_time: 0.56140 (0.46359) | > loader_time: 0.00360 (0.00314)  --> STEP: 260/639 -- GLOBAL_STEP: 16900 | > loss: -0.44694 (-0.44446) | > log_mle: -0.57117 (-0.57135) | > loss_dur: 0.12423 (0.12690) | > amp_scaler: 4096.00000 (6238.52308) | > grad_norm: 10.88941 (8.14330) | > current_lr: 0.00049 | > step_time: 0.57390 (0.47911) | > loader_time: 0.00300 (0.00316)  --> STEP: 285/639 -- GLOBAL_STEP: 16925 | > loss: -0.38808 (-0.44446) | > log_mle: -0.51699 (-0.57124) | > loss_dur: 0.12891 (0.12678) | > amp_scaler: 4096.00000 (6050.58246) | > grad_norm: 2.74864 (8.22658) | > current_lr: 0.00049 | > step_time: 0.61610 (0.49560) | > loader_time: 0.00340 (0.00318)  --> STEP: 310/639 -- GLOBAL_STEP: 16950 | > loss: -0.43236 (-0.44382) | > log_mle: -0.56079 (-0.57059) | > loss_dur: 0.12843 (0.12677) | > amp_scaler: 4096.00000 (5892.95484) | > grad_norm: 8.04923 (8.15126) | > current_lr: 0.00049 | > step_time: 0.58310 (0.51021) | > loader_time: 0.00340 (0.00321)  --> STEP: 335/639 -- GLOBAL_STEP: 16975 | > loss: -0.45911 (-0.44241) | > log_mle: -0.58604 (-0.56935) | > loss_dur: 0.12693 (0.12695) | > amp_scaler: 4096.00000 (5758.85373) | > grad_norm: 15.22933 (8.24930) | > current_lr: 0.00049 | > step_time: 0.66210 (0.52593) | > loader_time: 0.00310 (0.00323)  --> STEP: 360/639 -- GLOBAL_STEP: 17000 | > loss: -0.39276 (-0.44176) | > log_mle: -0.51886 (-0.56826) | > loss_dur: 0.12611 (0.12650) | > amp_scaler: 4096.00000 (5643.37778) | > grad_norm: 8.43580 (8.18033) | > current_lr: 0.00049 | > step_time: 0.69210 (0.53942) | > loader_time: 0.00370 (0.00325)  --> STEP: 385/639 -- GLOBAL_STEP: 17025 | > loss: -0.43735 (-0.44092) | > log_mle: -0.55708 (-0.56712) | > loss_dur: 0.11973 (0.12620) | > amp_scaler: 4096.00000 (5542.89870) | > grad_norm: 7.81964 (8.01280) | > current_lr: 0.00048 | > step_time: 0.67680 (0.55326) | > loader_time: 0.00360 (0.00328)  --> STEP: 410/639 -- GLOBAL_STEP: 17050 | > loss: -0.39443 (-0.43999) | > log_mle: -0.53396 (-0.56616) | > loss_dur: 0.13953 (0.12617) | > amp_scaler: 4096.00000 (5454.67317) | > grad_norm: 6.09720 (7.95218) | > current_lr: 0.00048 | > step_time: 0.79650 (0.56491) | > loader_time: 0.00390 (0.00331)  --> STEP: 435/639 -- GLOBAL_STEP: 17075 | > loss: -0.42643 (-0.43883) | > log_mle: -0.54368 (-0.56479) | > loss_dur: 0.11725 (0.12596) | > amp_scaler: 4096.00000 (5376.58851) | > grad_norm: 5.85584 (7.84655) | > current_lr: 0.00048 | > step_time: 0.72210 (0.57735) | > loader_time: 0.00340 (0.00333)  --> STEP: 460/639 -- GLOBAL_STEP: 17100 | > loss: -0.38037 (-0.43768) | > log_mle: -0.50792 (-0.56343) | > loss_dur: 0.12755 (0.12575) | > amp_scaler: 4096.00000 (5306.99130) | > grad_norm: 2.27703 (7.72388) | > current_lr: 0.00048 | > step_time: 0.86060 (0.58927) | > loader_time: 0.00390 (0.00335)  --> STEP: 485/639 -- GLOBAL_STEP: 17125 | > loss: -0.42542 (-0.43674) | > log_mle: -0.54386 (-0.56252) | > loss_dur: 0.11844 (0.12578) | > amp_scaler: 4096.00000 (5244.56907) | > grad_norm: 5.68822 (7.71520) | > current_lr: 0.00048 | > step_time: 0.87150 (0.60255) | > loader_time: 0.00430 (0.00338)  --> STEP: 510/639 -- GLOBAL_STEP: 17150 | > loss: -0.42869 (-0.43626) | > log_mle: -0.54973 (-0.56177) | > loss_dur: 0.12103 (0.12551) | > amp_scaler: 4096.00000 (5188.26667) | > grad_norm: 5.84290 (7.62079) | > current_lr: 0.00048 | > step_time: 0.84830 (0.61524) | > loader_time: 0.00420 (0.00340)  --> STEP: 535/639 -- GLOBAL_STEP: 17175 | > loss: -0.40838 (-0.43557) | > log_mle: -0.53792 (-0.56096) | > loss_dur: 0.12953 (0.12539) | > amp_scaler: 4096.00000 (5137.22617) | > grad_norm: 6.08542 (7.55939) | > current_lr: 0.00048 | > step_time: 0.99620 (0.62930) | > loader_time: 0.00390 (0.00343)  --> STEP: 560/639 -- GLOBAL_STEP: 17200 | > loss: -0.41631 (-0.43484) | > log_mle: -0.54889 (-0.56015) | > loss_dur: 0.13258 (0.12531) | > amp_scaler: 4096.00000 (5090.74286) | > grad_norm: 5.65417 (7.51286) | > current_lr: 0.00048 | > step_time: 0.90890 (0.64438) | > loader_time: 0.00420 (0.00347)  --> STEP: 585/639 -- GLOBAL_STEP: 17225 | > loss: -0.43355 (-0.43432) | > log_mle: -0.55136 (-0.55946) | > loss_dur: 0.11780 (0.12514) | > amp_scaler: 4096.00000 (5048.23248) | > grad_norm: 6.00563 (7.44546) | > current_lr: 0.00048 | > step_time: 1.05030 (0.65928) | > loader_time: 0.00450 (0.00350)  --> STEP: 610/639 -- GLOBAL_STEP: 17250 | > loss: -0.39865 (-0.43378) | > log_mle: -0.52053 (-0.55886) | > loss_dur: 0.12188 (0.12508) | > amp_scaler: 4096.00000 (5009.20656) | > grad_norm: 7.44030 (7.39542) | > current_lr: 0.00048 | > step_time: 1.05470 (0.67525) | > loader_time: 0.00460 (0.00354)  --> STEP: 635/639 -- GLOBAL_STEP: 17275 | > loss: -0.42467 (-0.43348) | > log_mle: -0.54696 (-0.55861) | > loss_dur: 0.12229 (0.12513) | > amp_scaler: 4096.00000 (4973.25354) | > grad_norm: 6.76774 (7.38552) | > current_lr: 0.00048 | > step_time: 1.13680 (0.69276) | > loader_time: 0.00440 (0.00358) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02530 (+0.00616) | > avg_loss: -0.42268 (+0.00473) | > avg_log_mle: -0.55026 (+0.00269) | > avg_loss_dur: 0.12758 (+0.00205)  > EPOCH: 27/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 18:12:46)   --> STEP: 20/639 -- GLOBAL_STEP: 17300 | > loss: -0.56627 (-0.46960) | > log_mle: -0.72283 (-0.60561) | > loss_dur: 0.15655 (0.13601) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 17.71394 (9.45722) | > current_lr: 0.00048 | > step_time: 0.34360 (0.31475) | > loader_time: 0.00280 (0.00513)  --> STEP: 45/639 -- GLOBAL_STEP: 17325 | > loss: -0.39637 (-0.44774) | > log_mle: -0.51529 (-0.58034) | > loss_dur: 0.11892 (0.13260) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.43720 (8.13534) | > current_lr: 0.00048 | > step_time: 0.40300 (0.34790) | > loader_time: 0.00290 (0.00382)  --> STEP: 70/639 -- GLOBAL_STEP: 17350 | > loss: -0.46773 (-0.43981) | > log_mle: -0.59100 (-0.57221) | > loss_dur: 0.12328 (0.13240) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.81002 (8.00596) | > current_lr: 0.00048 | > step_time: 0.41660 (0.36976) | > loader_time: 0.00280 (0.00349)  --> STEP: 95/639 -- GLOBAL_STEP: 17375 | > loss: -0.52135 (-0.43983) | > log_mle: -0.63463 (-0.56936) | > loss_dur: 0.11329 (0.12952) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.74869 (7.67917) | > current_lr: 0.00048 | > step_time: 0.42560 (0.38608) | > loader_time: 0.00280 (0.00333)  --> STEP: 120/639 -- GLOBAL_STEP: 17400 | > loss: -0.49718 (-0.44033) | > log_mle: -0.63896 (-0.57006) | > loss_dur: 0.14178 (0.12973) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.58341 (8.08252) | > current_lr: 0.00048 | > step_time: 0.53760 (0.40082) | > loader_time: 0.00320 (0.00326)  --> STEP: 145/639 -- GLOBAL_STEP: 17425 | > loss: -0.46707 (-0.44332) | > log_mle: -0.59233 (-0.57231) | > loss_dur: 0.12525 (0.12899) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.67208 (8.33123) | > current_lr: 0.00048 | > step_time: 0.47130 (0.41575) | > loader_time: 0.00320 (0.00321)  --> STEP: 170/639 -- GLOBAL_STEP: 17450 | > loss: -0.45445 (-0.44473) | > log_mle: -0.57277 (-0.57279) | > loss_dur: 0.11832 (0.12806) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.09903 (8.53170) | > current_lr: 0.00048 | > step_time: 0.52240 (0.42931) | > loader_time: 0.00330 (0.00318)  --> STEP: 195/639 -- GLOBAL_STEP: 17475 | > loss: -0.40559 (-0.44508) | > log_mle: -0.51416 (-0.57257) | > loss_dur: 0.10857 (0.12749) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.16669 (8.46147) | > current_lr: 0.00048 | > step_time: 0.49550 (0.44548) | > loader_time: 0.00320 (0.00317)  --> STEP: 220/639 -- GLOBAL_STEP: 17500 | > loss: -0.46198 (-0.44419) | > log_mle: -0.57336 (-0.57117) | > loss_dur: 0.11138 (0.12699) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.04353 (8.37582) | > current_lr: 0.00048 | > step_time: 0.63100 (0.46168) | > loader_time: 0.00320 (0.00317)  --> STEP: 245/639 -- GLOBAL_STEP: 17525 | > loss: -0.43871 (-0.44654) | > log_mle: -0.55942 (-0.57276) | > loss_dur: 0.12071 (0.12622) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.74420 (8.43926) | > current_lr: 0.00048 | > step_time: 0.54450 (0.47794) | > loader_time: 0.00320 (0.00319)  --> STEP: 270/639 -- GLOBAL_STEP: 17550 | > loss: -0.48553 (-0.44751) | > log_mle: -0.61450 (-0.57358) | > loss_dur: 0.12896 (0.12607) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.91015 (8.52802) | > current_lr: 0.00048 | > step_time: 0.85160 (0.49606) | > loader_time: 0.00360 (0.00320)  --> STEP: 295/639 -- GLOBAL_STEP: 17575 | > loss: -0.40558 (-0.44633) | > log_mle: -0.51804 (-0.57236) | > loss_dur: 0.11246 (0.12604) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.36212 (8.36306) | > current_lr: 0.00048 | > step_time: 0.64380 (0.51000) | > loader_time: 0.00340 (0.00321)  --> STEP: 320/639 -- GLOBAL_STEP: 17600 | > loss: -0.43433 (-0.44557) | > log_mle: -0.56880 (-0.57180) | > loss_dur: 0.13448 (0.12623) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.82471 (8.28929) | > current_lr: 0.00048 | > step_time: 0.75470 (0.52454) | > loader_time: 0.00370 (0.00323)  --> STEP: 345/639 -- GLOBAL_STEP: 17625 | > loss: -0.44652 (-0.44458) | > log_mle: -0.55656 (-0.57083) | > loss_dur: 0.11004 (0.12626) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.70985 (8.20389) | > current_lr: 0.00048 | > step_time: 0.67880 (0.53910) | > loader_time: 0.00340 (0.00325)  --> STEP: 370/639 -- GLOBAL_STEP: 17650 | > loss: -0.45163 (-0.44368) | > log_mle: -0.57442 (-0.56949) | > loss_dur: 0.12279 (0.12581) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.25324 (8.08626) | > current_lr: 0.00048 | > step_time: 0.90220 (0.55146) | > loader_time: 0.00360 (0.00327)  --> STEP: 395/639 -- GLOBAL_STEP: 17675 | > loss: -0.41548 (-0.44254) | > log_mle: -0.54271 (-0.56816) | > loss_dur: 0.12723 (0.12562) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.20387 (7.99930) | > current_lr: 0.00048 | > step_time: 0.71190 (0.56401) | > loader_time: 0.00340 (0.00329)  --> STEP: 420/639 -- GLOBAL_STEP: 17700 | > loss: -0.41860 (-0.44136) | > log_mle: -0.53800 (-0.56706) | > loss_dur: 0.11941 (0.12570) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.95189 (7.93787) | > current_lr: 0.00048 | > step_time: 0.70620 (0.57649) | > loader_time: 0.00350 (0.00331)  --> STEP: 445/639 -- GLOBAL_STEP: 17725 | > loss: -0.42760 (-0.44070) | > log_mle: -0.54321 (-0.56612) | > loss_dur: 0.11561 (0.12543) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.46934 (7.92254) | > current_lr: 0.00048 | > step_time: 0.79040 (0.58889) | > loader_time: 0.00370 (0.00333)  --> STEP: 470/639 -- GLOBAL_STEP: 17750 | > loss: -0.44023 (-0.43953) | > log_mle: -0.55553 (-0.56480) | > loss_dur: 0.11531 (0.12526) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.82625 (7.83812) | > current_lr: 0.00047 | > step_time: 0.85860 (0.60071) | > loader_time: 0.00370 (0.00335)  --> STEP: 495/639 -- GLOBAL_STEP: 17775 | > loss: -0.37953 (-0.43849) | > log_mle: -0.49057 (-0.56374) | > loss_dur: 0.11104 (0.12525) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.66813 (7.76069) | > current_lr: 0.00047 | > step_time: 0.78450 (0.61334) | > loader_time: 0.00350 (0.00337)  --> STEP: 520/639 -- GLOBAL_STEP: 17800 | > loss: -0.43649 (-0.43790) | > log_mle: -0.54895 (-0.56292) | > loss_dur: 0.11246 (0.12502) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.25784 (7.67974) | > current_lr: 0.00047 | > step_time: 0.80210 (0.62565) | > loader_time: 0.00370 (0.00339)  --> STEP: 545/639 -- GLOBAL_STEP: 17825 | > loss: -0.44682 (-0.43726) | > log_mle: -0.58133 (-0.56220) | > loss_dur: 0.13451 (0.12494) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.50988 (7.62504) | > current_lr: 0.00047 | > step_time: 0.95550 (0.64057) | > loader_time: 0.00530 (0.00343)  --> STEP: 570/639 -- GLOBAL_STEP: 17850 | > loss: -0.41794 (-0.43645) | > log_mle: -0.52937 (-0.56126) | > loss_dur: 0.11142 (0.12481) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.29003 (7.58801) | > current_lr: 0.00047 | > step_time: 0.95350 (0.65480) | > loader_time: 0.00390 (0.00346)  --> STEP: 595/639 -- GLOBAL_STEP: 17875 | > loss: -0.40775 (-0.43609) | > log_mle: -0.52185 (-0.56079) | > loss_dur: 0.11409 (0.12470) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.54711 (7.55224) | > current_lr: 0.00047 | > step_time: 1.00550 (0.67113) | > loader_time: 0.00620 (0.00351)  --> STEP: 620/639 -- GLOBAL_STEP: 17900 | > loss: -0.41612 (-0.43545) | > log_mle: -0.53237 (-0.56008) | > loss_dur: 0.11625 (0.12463) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.79627 (7.50370) | > current_lr: 0.00047 | > step_time: 1.20170 (0.68771) | > loader_time: 0.00650 (0.00356) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02319 (-0.00211) | > avg_loss: -0.42754 (-0.00486) | > avg_log_mle: -0.55377 (-0.00351) | > avg_loss_dur: 0.12623 (-0.00134) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_17920.pth.tar  > EPOCH: 28/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 18:20:35)   --> STEP: 5/639 -- GLOBAL_STEP: 17925 | > loss: -0.46406 (-0.47152) | > log_mle: -0.63488 (-0.61768) | > loss_dur: 0.17082 (0.14616) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.01707 (9.24025) | > current_lr: 0.00047 | > step_time: 0.37380 (0.29476) | > loader_time: 0.00250 (0.00246)  --> STEP: 30/639 -- GLOBAL_STEP: 17950 | > loss: -0.47467 (-0.45504) | > log_mle: -0.59703 (-0.59135) | > loss_dur: 0.12235 (0.13630) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.05680 (8.06944) | > current_lr: 0.00047 | > step_time: 0.39330 (0.32917) | > loader_time: 0.00280 (0.00252)  --> STEP: 55/639 -- GLOBAL_STEP: 17975 | > loss: -0.40045 (-0.44751) | > log_mle: -0.51572 (-0.58028) | > loss_dur: 0.11527 (0.13277) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.08600 (7.85561) | > current_lr: 0.00047 | > step_time: 0.32990 (0.35697) | > loader_time: 0.00280 (0.00263)  --> STEP: 80/639 -- GLOBAL_STEP: 18000 | > loss: -0.43241 (-0.44382) | > log_mle: -0.57404 (-0.57434) | > loss_dur: 0.14162 (0.13051) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.15239 (7.45717) | > current_lr: 0.00047 | > step_time: 0.40170 (0.37637) | > loader_time: 0.00290 (0.00267)  --> STEP: 105/639 -- GLOBAL_STEP: 18025 | > loss: -0.39539 (-0.44422) | > log_mle: -0.52918 (-0.57394) | > loss_dur: 0.13378 (0.12972) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.96268 (7.84030) | > current_lr: 0.00047 | > step_time: 0.41410 (0.39361) | > loader_time: 0.00300 (0.00274)  --> STEP: 130/639 -- GLOBAL_STEP: 18050 | > loss: -0.42818 (-0.44494) | > log_mle: -0.56958 (-0.57438) | > loss_dur: 0.14139 (0.12944) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.39374 (7.95531) | > current_lr: 0.00047 | > step_time: 0.50870 (0.40675) | > loader_time: 0.00320 (0.00278)  --> STEP: 155/639 -- GLOBAL_STEP: 18075 | > loss: -0.43354 (-0.44635) | > log_mle: -0.56157 (-0.57502) | > loss_dur: 0.12803 (0.12867) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.57411 (8.06489) | > current_lr: 0.00047 | > step_time: 0.54660 (0.42261) | > loader_time: 0.00320 (0.00284)  --> STEP: 180/639 -- GLOBAL_STEP: 18100 | > loss: -0.48395 (-0.44916) | > log_mle: -0.61756 (-0.57731) | > loss_dur: 0.13361 (0.12815) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.95217 (8.36294) | > current_lr: 0.00047 | > step_time: 0.55000 (0.43895) | > loader_time: 0.00330 (0.00288)  --> STEP: 205/639 -- GLOBAL_STEP: 18125 | > loss: -0.42459 (-0.44727) | > log_mle: -0.54048 (-0.57455) | > loss_dur: 0.11588 (0.12728) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.83956 (8.17824) | > current_lr: 0.00047 | > step_time: 0.54960 (0.45275) | > loader_time: 0.00320 (0.00293)  --> STEP: 230/639 -- GLOBAL_STEP: 18150 | > loss: -0.45444 (-0.44856) | > log_mle: -0.57709 (-0.57525) | > loss_dur: 0.12265 (0.12669) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.37998 (8.27353) | > current_lr: 0.00047 | > step_time: 0.54600 (0.46890) | > loader_time: 0.00350 (0.00297)  --> STEP: 255/639 -- GLOBAL_STEP: 18175 | > loss: -0.41984 (-0.44904) | > log_mle: -0.54801 (-0.57511) | > loss_dur: 0.12817 (0.12607) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.26046 (8.31597) | > current_lr: 0.00047 | > step_time: 0.61080 (0.48261) | > loader_time: 0.00360 (0.00301)  --> STEP: 280/639 -- GLOBAL_STEP: 18200 | > loss: -0.44527 (-0.44994) | > log_mle: -0.57508 (-0.57607) | > loss_dur: 0.12981 (0.12613) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.61836 (8.38983) | > current_lr: 0.00047 | > step_time: 0.61160 (0.49867) | > loader_time: 0.00340 (0.00305)  --> STEP: 305/639 -- GLOBAL_STEP: 18225 | > loss: -0.42930 (-0.44916) | > log_mle: -0.55516 (-0.57516) | > loss_dur: 0.12586 (0.12600) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.58938 (8.36949) | > current_lr: 0.00047 | > step_time: 0.66110 (0.51381) | > loader_time: 0.00340 (0.00309)  --> STEP: 330/639 -- GLOBAL_STEP: 18250 | > loss: -0.43290 (-0.44752) | > log_mle: -0.57256 (-0.57379) | > loss_dur: 0.13967 (0.12627) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.31565 (8.23432) | > current_lr: 0.00047 | > step_time: 0.72600 (0.52822) | > loader_time: 0.00340 (0.00312)  --> STEP: 355/639 -- GLOBAL_STEP: 18275 | > loss: -0.39173 (-0.44670) | > log_mle: -0.50499 (-0.57256) | > loss_dur: 0.11326 (0.12586) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.82270 (8.16066) | > current_lr: 0.00047 | > step_time: 0.64440 (0.54088) | > loader_time: 0.00340 (0.00316)  --> STEP: 380/639 -- GLOBAL_STEP: 18300 | > loss: -0.43441 (-0.44540) | > log_mle: -0.54158 (-0.57098) | > loss_dur: 0.10717 (0.12558) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.71737 (8.05551) | > current_lr: 0.00047 | > step_time: 0.61440 (0.55456) | > loader_time: 0.00360 (0.00319)  --> STEP: 405/639 -- GLOBAL_STEP: 18325 | > loss: -0.42901 (-0.44469) | > log_mle: -0.54820 (-0.57032) | > loss_dur: 0.11919 (0.12564) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.97778 (8.00594) | > current_lr: 0.00047 | > step_time: 0.66740 (0.56608) | > loader_time: 0.00350 (0.00321)  --> STEP: 430/639 -- GLOBAL_STEP: 18350 | > loss: -0.39643 (-0.44341) | > log_mle: -0.50623 (-0.56892) | > loss_dur: 0.10980 (0.12551) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.75624 (7.93731) | > current_lr: 0.00047 | > step_time: 0.81340 (0.57944) | > loader_time: 0.00360 (0.00324)  --> STEP: 455/639 -- GLOBAL_STEP: 18375 | > loss: -0.41195 (-0.44230) | > log_mle: -0.53425 (-0.56751) | > loss_dur: 0.12230 (0.12521) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.31055 (7.82583) | > current_lr: 0.00047 | > step_time: 0.73810 (0.59097) | > loader_time: 0.00370 (0.00327)  --> STEP: 480/639 -- GLOBAL_STEP: 18400 | > loss: -0.37072 (-0.44112) | > log_mle: -0.48942 (-0.56640) | > loss_dur: 0.11870 (0.12528) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.92546 (7.76081) | > current_lr: 0.00047 | > step_time: 0.79130 (0.60343) | > loader_time: 0.00380 (0.00330)  --> STEP: 505/639 -- GLOBAL_STEP: 18425 | > loss: -0.44494 (-0.44037) | > log_mle: -0.55983 (-0.56538) | > loss_dur: 0.11489 (0.12501) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.47931 (7.71435) | > current_lr: 0.00047 | > step_time: 0.85990 (0.61538) | > loader_time: 0.00400 (0.00332)  --> STEP: 530/639 -- GLOBAL_STEP: 18450 | > loss: -0.42218 (-0.43971) | > log_mle: -0.55000 (-0.56463) | > loss_dur: 0.12782 (0.12492) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.56967 (7.67065) | > current_lr: 0.00047 | > step_time: 0.95110 (0.62899) | > loader_time: 0.00410 (0.00335)  --> STEP: 555/639 -- GLOBAL_STEP: 18475 | > loss: -0.41190 (-0.43885) | > log_mle: -0.53209 (-0.56359) | > loss_dur: 0.12020 (0.12474) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.22802 (7.57686) | > current_lr: 0.00047 | > step_time: 0.82340 (0.64352) | > loader_time: 0.00390 (0.00339)  --> STEP: 580/639 -- GLOBAL_STEP: 18500 | > loss: -0.43554 (-0.43818) | > log_mle: -0.55068 (-0.56282) | > loss_dur: 0.11514 (0.12463) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.78222 (7.55239) | > current_lr: 0.00047 | > step_time: 0.95880 (0.65806) | > loader_time: 0.00430 (0.00343)  --> STEP: 605/639 -- GLOBAL_STEP: 18525 | > loss: -0.40774 (-0.43758) | > log_mle: -0.51930 (-0.56219) | > loss_dur: 0.11157 (0.12461) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.94507 (7.51857) | > current_lr: 0.00046 | > step_time: 1.04840 (0.67437) | > loader_time: 0.00440 (0.00346)  --> STEP: 630/639 -- GLOBAL_STEP: 18550 | > loss: -0.42504 (-0.43713) | > log_mle: -0.55470 (-0.56172) | > loss_dur: 0.12966 (0.12459) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.83938 (7.49033) | > current_lr: 0.00046 | > step_time: 1.19230 (0.69152) | > loader_time: 0.00470 (0.00350) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03386 (+0.01067) | > avg_loss: -0.42710 (+0.00044) | > avg_log_mle: -0.55500 (-0.00123) | > avg_loss_dur: 0.12790 (+0.00167)  > EPOCH: 29/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 18:28:15)   --> STEP: 15/639 -- GLOBAL_STEP: 18575 | > loss: -0.56227 (-0.47408) | > log_mle: -0.72267 (-0.60784) | > loss_dur: 0.16041 (0.13377) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 19.35175 (8.88020) | > current_lr: 0.00046 | > step_time: 0.32890 (0.30270) | > loader_time: 0.00260 (0.00335)  --> STEP: 40/639 -- GLOBAL_STEP: 18600 | > loss: -0.44562 (-0.45748) | > log_mle: -0.56907 (-0.58842) | > loss_dur: 0.12346 (0.13094) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.14291 (7.88111) | > current_lr: 0.00046 | > step_time: 0.35460 (0.33397) | > loader_time: 0.00290 (0.00297)  --> STEP: 65/639 -- GLOBAL_STEP: 18625 | > loss: -0.45943 (-0.45052) | > log_mle: -0.56505 (-0.58247) | > loss_dur: 0.10562 (0.13195) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.35139 (8.03504) | > current_lr: 0.00046 | > step_time: 0.38780 (0.36296) | > loader_time: 0.00300 (0.00295)  --> STEP: 90/639 -- GLOBAL_STEP: 18650 | > loss: -0.39240 (-0.44586) | > log_mle: -0.51676 (-0.57527) | > loss_dur: 0.12436 (0.12941) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.53648 (7.45139) | > current_lr: 0.00046 | > step_time: 0.52290 (0.38011) | > loader_time: 0.00310 (0.00295)  --> STEP: 115/639 -- GLOBAL_STEP: 18675 | > loss: -0.35794 (-0.44791) | > log_mle: -0.48567 (-0.57685) | > loss_dur: 0.12773 (0.12894) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.06820 (7.78468) | > current_lr: 0.00046 | > step_time: 0.41270 (0.39529) | > loader_time: 0.00310 (0.00298)  --> STEP: 140/639 -- GLOBAL_STEP: 18700 | > loss: -0.42302 (-0.45010) | > log_mle: -0.57475 (-0.57849) | > loss_dur: 0.15172 (0.12839) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.55481 (8.01127) | > current_lr: 0.00046 | > step_time: 0.55170 (0.40819) | > loader_time: 0.00320 (0.00298)  --> STEP: 165/639 -- GLOBAL_STEP: 18725 | > loss: -0.43816 (-0.45083) | > log_mle: -0.58072 (-0.57821) | > loss_dur: 0.14256 (0.12738) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.92998 (8.01991) | > current_lr: 0.00046 | > step_time: 0.49130 (0.42300) | > loader_time: 0.00310 (0.00301)  --> STEP: 190/639 -- GLOBAL_STEP: 18750 | > loss: -0.37865 (-0.45198) | > log_mle: -0.50625 (-0.57853) | > loss_dur: 0.12760 (0.12655) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.54826 (8.15992) | > current_lr: 0.00046 | > step_time: 0.52710 (0.44169) | > loader_time: 0.00330 (0.00304)  --> STEP: 215/639 -- GLOBAL_STEP: 18775 | > loss: -0.50686 (-0.45102) | > log_mle: -0.62053 (-0.57729) | > loss_dur: 0.11367 (0.12626) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.91871 (8.21992) | > current_lr: 0.00046 | > step_time: 0.64880 (0.45855) | > loader_time: 0.00350 (0.00307)  --> STEP: 240/639 -- GLOBAL_STEP: 18800 | > loss: -0.50272 (-0.45316) | > log_mle: -0.62900 (-0.57869) | > loss_dur: 0.12628 (0.12553) | > amp_scaler: 8192.00000 (4505.60000) | > grad_norm: 16.04752 (8.23102) | > current_lr: 0.00046 | > step_time: 0.55460 (0.47398) | > loader_time: 0.00330 (0.00308)  --> STEP: 265/639 -- GLOBAL_STEP: 18825 | > loss: -0.51347 (-0.45319) | > log_mle: -0.64736 (-0.57863) | > loss_dur: 0.13389 (0.12543) | > amp_scaler: 8192.00000 (4853.37358) | > grad_norm: 18.13352 (8.41726) | > current_lr: 0.00046 | > step_time: 0.62040 (0.49015) | > loader_time: 0.00330 (0.00311)  --> STEP: 290/639 -- GLOBAL_STEP: 18850 | > loss: -0.43655 (-0.45263) | > log_mle: -0.55477 (-0.57797) | > loss_dur: 0.11822 (0.12534) | > amp_scaler: 8192.00000 (5141.18621) | > grad_norm: 5.45456 (8.33541) | > current_lr: 0.00046 | > step_time: 0.64060 (0.50608) | > loader_time: 0.00350 (0.00314)  --> STEP: 315/639 -- GLOBAL_STEP: 18875 | > loss: -0.40964 (-0.45185) | > log_mle: -0.52413 (-0.57699) | > loss_dur: 0.11448 (0.12514) | > amp_scaler: 8192.00000 (5383.31429) | > grad_norm: 2.93039 (8.34412) | > current_lr: 0.00046 | > step_time: 0.62420 (0.52031) | > loader_time: 0.00350 (0.00317)  --> STEP: 340/639 -- GLOBAL_STEP: 18900 | > loss: -0.44025 (-0.45072) | > log_mle: -0.55527 (-0.57615) | > loss_dur: 0.11503 (0.12543) | > amp_scaler: 4096.00000 (5517.55294) | > grad_norm: 5.35709 (8.38292) | > current_lr: 0.00046 | > step_time: 0.61000 (0.53545) | > loader_time: 0.00310 (0.00319)  --> STEP: 365/639 -- GLOBAL_STEP: 18925 | > loss: -0.45051 (-0.44945) | > log_mle: -0.56326 (-0.57441) | > loss_dur: 0.11275 (0.12496) | > amp_scaler: 4096.00000 (5420.18630) | > grad_norm: 5.69709 (8.23357) | > current_lr: 0.00046 | > step_time: 0.84090 (0.54687) | > loader_time: 0.00390 (0.00321)  --> STEP: 390/639 -- GLOBAL_STEP: 18950 | > loss: -0.42391 (-0.44817) | > log_mle: -0.55617 (-0.57297) | > loss_dur: 0.13226 (0.12480) | > amp_scaler: 4096.00000 (5335.30256) | > grad_norm: 7.05678 (8.10445) | > current_lr: 0.00046 | > step_time: 0.75870 (0.56039) | > loader_time: 0.00350 (0.00324)  --> STEP: 415/639 -- GLOBAL_STEP: 18975 | > loss: -0.40823 (-0.44694) | > log_mle: -0.53244 (-0.57176) | > loss_dur: 0.12421 (0.12482) | > amp_scaler: 4096.00000 (5260.64578) | > grad_norm: 5.90874 (8.03874) | > current_lr: 0.00046 | > step_time: 0.72650 (0.57255) | > loader_time: 0.00360 (0.00327)  --> STEP: 440/639 -- GLOBAL_STEP: 19000 | > loss: -0.39546 (-0.44605) | > log_mle: -0.51784 (-0.57070) | > loss_dur: 0.12238 (0.12465) | > amp_scaler: 4096.00000 (5194.47273) | > grad_norm: 7.34769 (7.98329) | > current_lr: 0.00046 | > step_time: 0.77040 (0.58494) | > loader_time: 0.00370 (0.00330)  --> STEP: 465/639 -- GLOBAL_STEP: 19025 | > loss: -0.40893 (-0.44486) | > log_mle: -0.52497 (-0.56926) | > loss_dur: 0.11604 (0.12440) | > amp_scaler: 4096.00000 (5135.41505) | > grad_norm: 4.78575 (7.90692) | > current_lr: 0.00046 | > step_time: 0.79410 (0.59566) | > loader_time: 0.00390 (0.00332)  --> STEP: 490/639 -- GLOBAL_STEP: 19050 | > loss: -0.39421 (-0.44384) | > log_mle: -0.50546 (-0.56823) | > loss_dur: 0.11126 (0.12439) | > amp_scaler: 4096.00000 (5082.38367) | > grad_norm: 3.37164 (7.88816) | > current_lr: 0.00046 | > step_time: 0.89990 (0.60910) | > loader_time: 0.00360 (0.00334)  --> STEP: 515/639 -- GLOBAL_STEP: 19075 | > loss: -0.42162 (-0.44345) | > log_mle: -0.53819 (-0.56764) | > loss_dur: 0.11657 (0.12419) | > amp_scaler: 4096.00000 (5034.50097) | > grad_norm: 4.38405 (7.90596) | > current_lr: 0.00046 | > step_time: 1.07030 (0.62273) | > loader_time: 0.00420 (0.00337)  --> STEP: 540/639 -- GLOBAL_STEP: 19100 | > loss: -0.44177 (-0.44264) | > log_mle: -0.55700 (-0.56671) | > loss_dur: 0.11523 (0.12407) | > amp_scaler: 4096.00000 (4991.05185) | > grad_norm: 6.27489 (7.86622) | > current_lr: 0.00046 | > step_time: 0.85080 (0.63728) | > loader_time: 0.00420 (0.00340)  --> STEP: 565/639 -- GLOBAL_STEP: 19125 | > loss: -0.39535 (-0.44162) | > log_mle: -0.51548 (-0.56560) | > loss_dur: 0.12014 (0.12398) | > amp_scaler: 4096.00000 (4951.44779) | > grad_norm: 3.24315 (7.81294) | > current_lr: 0.00046 | > step_time: 0.99810 (0.65106) | > loader_time: 0.00440 (0.00343)  --> STEP: 590/639 -- GLOBAL_STEP: 19150 | > loss: -0.42618 (-0.44118) | > log_mle: -0.54842 (-0.56508) | > loss_dur: 0.12224 (0.12390) | > amp_scaler: 4096.00000 (4915.20000) | > grad_norm: 7.62753 (7.75775) | > current_lr: 0.00046 | > step_time: 0.94230 (0.66665) | > loader_time: 0.00430 (0.00347)  --> STEP: 615/639 -- GLOBAL_STEP: 19175 | > loss: -0.43006 (-0.44053) | > log_mle: -0.55224 (-0.56435) | > loss_dur: 0.12218 (0.12382) | > amp_scaler: 4096.00000 (4881.89919) | > grad_norm: 7.26420 (7.69607) | > current_lr: 0.00046 | > step_time: 1.01130 (0.68244) | > loader_time: 0.00470 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.01962 (-0.01424) | > avg_loss: -0.42828 (-0.00118) | > avg_log_mle: -0.55575 (-0.00075) | > avg_loss_dur: 0.12747 (-0.00043) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_19200.pth.tar  > EPOCH: 30/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 18:36:02)   --> STEP: 0/639 -- GLOBAL_STEP: 19200 | > loss: -0.48411 (-0.48411) | > log_mle: -0.65427 (-0.65427) | > loss_dur: 0.17016 (0.17016) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 17.47797 (17.47797) | > current_lr: 0.00046 | > step_time: 0.82060 (0.82059) | > loader_time: 0.32130 (0.32125)  --> STEP: 25/639 -- GLOBAL_STEP: 19225 | > loss: -0.46299 (-0.47303) | > log_mle: -0.61053 (-0.60455) | > loss_dur: 0.14754 (0.13152) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.26090 (8.84117) | > current_lr: 0.00046 | > step_time: 0.37220 (0.32547) | > loader_time: 0.00290 (0.00267)  --> STEP: 50/639 -- GLOBAL_STEP: 19250 | > loss: -0.42589 (-0.45810) | > log_mle: -0.53652 (-0.58780) | > loss_dur: 0.11063 (0.12971) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.96168 (7.98905) | > current_lr: 0.00046 | > step_time: 0.31250 (0.34624) | > loader_time: 0.00270 (0.00273)  --> STEP: 75/639 -- GLOBAL_STEP: 19275 | > loss: -0.44887 (-0.45073) | > log_mle: -0.57031 (-0.58028) | > loss_dur: 0.12144 (0.12954) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.30127 (8.03504) | > current_lr: 0.00046 | > step_time: 0.40760 (0.37002) | > loader_time: 0.00290 (0.00278)  --> STEP: 100/639 -- GLOBAL_STEP: 19300 | > loss: -0.46328 (-0.45139) | > log_mle: -0.58673 (-0.57909) | > loss_dur: 0.12345 (0.12771) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.98403 (7.85609) | > current_lr: 0.00046 | > step_time: 0.51590 (0.38827) | > loader_time: 0.00300 (0.00282)  --> STEP: 125/639 -- GLOBAL_STEP: 19325 | > loss: -0.46042 (-0.45133) | > log_mle: -0.58112 (-0.57908) | > loss_dur: 0.12070 (0.12775) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.23103 (8.14798) | > current_lr: 0.00046 | > step_time: 0.41440 (0.40197) | > loader_time: 0.00270 (0.00284)  --> STEP: 150/639 -- GLOBAL_STEP: 19350 | > loss: -0.41904 (-0.45256) | > log_mle: -0.54852 (-0.57978) | > loss_dur: 0.12949 (0.12722) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.95139 (8.27455) | > current_lr: 0.00045 | > step_time: 0.42830 (0.41651) | > loader_time: 0.00260 (0.00286)  --> STEP: 175/639 -- GLOBAL_STEP: 19375 | > loss: -0.46212 (-0.45476) | > log_mle: -0.57175 (-0.58147) | > loss_dur: 0.10963 (0.12671) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.62937 (8.45797) | > current_lr: 0.00045 | > step_time: 0.54920 (0.43356) | > loader_time: 0.00340 (0.00290)  --> STEP: 200/639 -- GLOBAL_STEP: 19400 | > loss: -0.43042 (-0.45386) | > log_mle: -0.55051 (-0.57979) | > loss_dur: 0.12009 (0.12594) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.63116 (8.49059) | > current_lr: 0.00045 | > step_time: 0.51250 (0.44684) | > loader_time: 0.00320 (0.00295)  --> STEP: 225/639 -- GLOBAL_STEP: 19425 | > loss: -0.41878 (-0.45308) | > log_mle: -0.52637 (-0.57841) | > loss_dur: 0.10759 (0.12532) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.62511 (8.41263) | > current_lr: 0.00045 | > step_time: 0.51140 (0.46190) | > loader_time: 0.00280 (0.00299)  --> STEP: 250/639 -- GLOBAL_STEP: 19450 | > loss: -0.42032 (-0.45505) | > log_mle: -0.53310 (-0.57999) | > loss_dur: 0.11277 (0.12494) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.06930 (8.55192) | > current_lr: 0.00045 | > step_time: 0.58980 (0.47589) | > loader_time: 0.00350 (0.00302)  --> STEP: 275/639 -- GLOBAL_STEP: 19475 | > loss: -0.45012 (-0.45616) | > log_mle: -0.57781 (-0.58109) | > loss_dur: 0.12769 (0.12493) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.28113 (8.69173) | > current_lr: 0.00045 | > step_time: 0.92190 (0.49431) | > loader_time: 0.00380 (0.00307)  --> STEP: 300/639 -- GLOBAL_STEP: 19500 | > loss: -0.45533 (-0.45480) | > log_mle: -0.57077 (-0.57981) | > loss_dur: 0.11544 (0.12501) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.43843 (8.61552) | > current_lr: 0.00045 | > step_time: 0.68110 (0.50732) | > loader_time: 0.00360 (0.00310)  --> STEP: 325/639 -- GLOBAL_STEP: 19525 | > loss: -0.44773 (-0.45369) | > log_mle: -0.56976 (-0.57889) | > loss_dur: 0.12202 (0.12520) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.60060 (8.56450) | > current_lr: 0.00045 | > step_time: 0.59490 (0.52242) | > loader_time: 0.00340 (0.00314)  --> STEP: 350/639 -- GLOBAL_STEP: 19550 | > loss: -0.47617 (-0.45260) | > log_mle: -0.60326 (-0.57761) | > loss_dur: 0.12709 (0.12501) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.34452 (8.53653) | > current_lr: 0.00045 | > step_time: 0.85700 (0.53631) | > loader_time: 0.00400 (0.00318)  --> STEP: 375/639 -- GLOBAL_STEP: 19575 | > loss: -0.41077 (-0.45106) | > log_mle: -0.53477 (-0.57579) | > loss_dur: 0.12400 (0.12473) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.03869 (8.38803) | > current_lr: 0.00045 | > step_time: 0.76360 (0.55074) | > loader_time: 0.00360 (0.00321)  --> STEP: 400/639 -- GLOBAL_STEP: 19600 | > loss: -0.44525 (-0.45007) | > log_mle: -0.56476 (-0.57481) | > loss_dur: 0.11951 (0.12474) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.68708 (8.33524) | > current_lr: 0.00045 | > step_time: 0.76750 (0.56459) | > loader_time: 0.00350 (0.00324)  --> STEP: 425/639 -- GLOBAL_STEP: 19625 | > loss: -0.48988 (-0.44866) | > log_mle: -0.61540 (-0.57327) | > loss_dur: 0.12551 (0.12462) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.97064 (8.25374) | > current_lr: 0.00045 | > step_time: 0.90730 (0.57764) | > loader_time: 0.00390 (0.00326)  --> STEP: 450/639 -- GLOBAL_STEP: 19650 | > loss: -0.40878 (-0.44754) | > log_mle: -0.52198 (-0.57189) | > loss_dur: 0.11320 (0.12435) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.06972 (8.13861) | > current_lr: 0.00045 | > step_time: 0.80510 (0.58958) | > loader_time: 0.00400 (0.00329)  --> STEP: 475/639 -- GLOBAL_STEP: 19675 | > loss: -0.42725 (-0.44626) | > log_mle: -0.53135 (-0.57058) | > loss_dur: 0.10411 (0.12432) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.12241 (8.08553) | > current_lr: 0.00045 | > step_time: 0.84370 (0.60202) | > loader_time: 0.00360 (0.00332)  --> STEP: 500/639 -- GLOBAL_STEP: 19700 | > loss: -0.46469 (-0.44545) | > log_mle: -0.59414 (-0.56969) | > loss_dur: 0.12945 (0.12425) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.31186 (8.05379) | > current_lr: 0.00045 | > step_time: 0.99060 (0.61541) | > loader_time: 0.00380 (0.00334)  --> STEP: 525/639 -- GLOBAL_STEP: 19725 | > loss: -0.44760 (-0.44475) | > log_mle: -0.57829 (-0.56893) | > loss_dur: 0.13069 (0.12418) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.23500 (8.04272) | > current_lr: 0.00045 | > step_time: 1.04730 (0.62829) | > loader_time: 0.00430 (0.00337)  --> STEP: 550/639 -- GLOBAL_STEP: 19750 | > loss: -0.38376 (-0.44388) | > log_mle: -0.50928 (-0.56793) | > loss_dur: 0.12552 (0.12405) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.96043 (7.98121) | > current_lr: 0.00045 | > step_time: 0.92680 (0.64296) | > loader_time: 0.00410 (0.00340)  --> STEP: 575/639 -- GLOBAL_STEP: 19775 | > loss: -0.41575 (-0.44306) | > log_mle: -0.53049 (-0.56699) | > loss_dur: 0.11474 (0.12393) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.53870 (7.91924) | > current_lr: 0.00045 | > step_time: 0.98050 (0.65774) | > loader_time: 0.00420 (0.00343)  --> STEP: 600/639 -- GLOBAL_STEP: 19800 | > loss: -0.40241 (-0.44252) | > log_mle: -0.52752 (-0.56641) | > loss_dur: 0.12512 (0.12388) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.72524 (7.88063) | > current_lr: 0.00045 | > step_time: 1.02360 (0.67392) | > loader_time: 0.00440 (0.00347)  --> STEP: 625/639 -- GLOBAL_STEP: 19825 | > loss: -0.43995 (-0.44189) | > log_mle: -0.56026 (-0.56570) | > loss_dur: 0.12031 (0.12381) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.26256 (7.84419) | > current_lr: 0.00045 | > step_time: 1.04530 (0.68999) | > loader_time: 0.00440 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02536 (+0.00574) | > avg_loss: -0.43325 (-0.00497) | > avg_log_mle: -0.56156 (-0.00580) | > avg_loss_dur: 0.12831 (+0.00083) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_19840.pth.tar  > EPOCH: 31/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 18:43:50)   --> STEP: 10/639 -- GLOBAL_STEP: 19850 | > loss: -0.47584 (-0.47728) | > log_mle: -0.61396 (-0.61054) | > loss_dur: 0.13812 (0.13326) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.03006 (11.42255) | > current_lr: 0.00045 | > step_time: 0.29710 (0.29297) | > loader_time: 0.00250 (0.00492)  --> STEP: 35/639 -- GLOBAL_STEP: 19875 | > loss: -0.45475 (-0.46253) | > log_mle: -0.56931 (-0.59073) | > loss_dur: 0.11457 (0.12820) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.71301 (8.68242) | > current_lr: 0.00045 | > step_time: 0.31940 (0.32970) | > loader_time: 0.00280 (0.00337)  --> STEP: 60/639 -- GLOBAL_STEP: 19900 | > loss: -0.40477 (-0.45417) | > log_mle: -0.52876 (-0.58344) | > loss_dur: 0.12399 (0.12927) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.06900 (8.26186) | > current_lr: 0.00045 | > step_time: 0.37070 (0.35729) | > loader_time: 0.00270 (0.00312)  --> STEP: 85/639 -- GLOBAL_STEP: 19925 | > loss: -0.40138 (-0.44959) | > log_mle: -0.52910 (-0.57679) | > loss_dur: 0.12772 (0.12720) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.09165 (7.62069) | > current_lr: 0.00045 | > step_time: 0.41690 (0.37273) | > loader_time: 0.00300 (0.00304)  --> STEP: 110/639 -- GLOBAL_STEP: 19950 | > loss: -0.42699 (-0.45324) | > log_mle: -0.55122 (-0.58067) | > loss_dur: 0.12422 (0.12743) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.49575 (8.18581) | > current_lr: 0.00045 | > step_time: 0.46110 (0.39335) | > loader_time: 0.00310 (0.00301)  --> STEP: 135/639 -- GLOBAL_STEP: 19975 | > loss: -0.40590 (-0.45328) | > log_mle: -0.53803 (-0.57999) | > loss_dur: 0.13213 (0.12671) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.41382 (8.31705) | > current_lr: 0.00045 | > step_time: 0.46760 (0.40392) | > loader_time: 0.00300 (0.00300)  --> STEP: 160/639 -- GLOBAL_STEP: 20000 | > loss: -0.47549 (-0.45357) | > log_mle: -0.57997 (-0.57918) | > loss_dur: 0.10448 (0.12561) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.18861 (8.24747) | > current_lr: 0.00045 | > step_time: 0.46240 (0.41870) | > loader_time: 0.00250 (0.00301) > CHECKPOINT : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/checkpoint_20000.pth.tar warning: audio amplitude out of range, auto clipped.  --> STEP: 185/639 -- GLOBAL_STEP: 20025 | > loss: -0.44245 (-0.45691) | > log_mle: -0.55737 (-0.58255) | > loss_dur: 0.11492 (0.12564) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.46171 (8.47014) | > current_lr: 0.00045 | > step_time: 0.57560 (0.43600) | > loader_time: 0.00330 (0.00310)  --> STEP: 210/639 -- GLOBAL_STEP: 20050 | > loss: -0.37136 (-0.45486) | > log_mle: -0.48283 (-0.58034) | > loss_dur: 0.11148 (0.12548) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.87279 (8.29828) | > current_lr: 0.00045 | > step_time: 0.48940 (0.45206) | > loader_time: 0.00330 (0.00311)  --> STEP: 235/639 -- GLOBAL_STEP: 20075 | > loss: -0.47014 (-0.45668) | > log_mle: -0.58421 (-0.58145) | > loss_dur: 0.11407 (0.12477) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.67574 (8.42560) | > current_lr: 0.00045 | > step_time: 0.54740 (0.46804) | > loader_time: 0.00330 (0.00313)  --> STEP: 260/639 -- GLOBAL_STEP: 20100 | > loss: -0.46236 (-0.45742) | > log_mle: -0.57949 (-0.58194) | > loss_dur: 0.11713 (0.12453) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.48695 (8.45761) | > current_lr: 0.00045 | > step_time: 0.59740 (0.48220) | > loader_time: 0.00330 (0.00314)  --> STEP: 285/639 -- GLOBAL_STEP: 20125 | > loss: -0.38921 (-0.45714) | > log_mle: -0.52306 (-0.58172) | > loss_dur: 0.13384 (0.12458) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.24036 (8.48851) | > current_lr: 0.00045 | > step_time: 0.62960 (0.49865) | > loader_time: 0.00350 (0.00315)  --> STEP: 310/639 -- GLOBAL_STEP: 20150 | > loss: -0.45586 (-0.45641) | > log_mle: -0.57392 (-0.58097) | > loss_dur: 0.11806 (0.12457) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.81611 (8.48490) | > current_lr: 0.00045 | > step_time: 0.58420 (0.51329) | > loader_time: 0.00340 (0.00318)  --> STEP: 335/639 -- GLOBAL_STEP: 20175 | > loss: -0.46062 (-0.45491) | > log_mle: -0.59184 (-0.57971) | > loss_dur: 0.13122 (0.12480) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.48909 (8.43455) | > current_lr: 0.00045 | > step_time: 0.70910 (0.52833) | > loader_time: 0.00300 (0.00321)  --> STEP: 360/639 -- GLOBAL_STEP: 20200 | > loss: -0.41065 (-0.45394) | > log_mle: -0.52989 (-0.57837) | > loss_dur: 0.11925 (0.12443) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.63525 (8.35663) | > current_lr: 0.00045 | > step_time: 0.75980 (0.54169) | > loader_time: 0.00350 (0.00323)  --> STEP: 385/639 -- GLOBAL_STEP: 20225 | > loss: -0.44194 (-0.45287) | > log_mle: -0.56346 (-0.57709) | > loss_dur: 0.12152 (0.12421) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.45534 (8.27309) | > current_lr: 0.00044 | > step_time: 0.72310 (0.55510) | > loader_time: 0.00340 (0.00326)  --> STEP: 410/639 -- GLOBAL_STEP: 20250 | > loss: -0.40088 (-0.45168) | > log_mle: -0.54063 (-0.57594) | > loss_dur: 0.13975 (0.12426) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.49844 (8.22093) | > current_lr: 0.00044 | > step_time: 0.78010 (0.56739) | > loader_time: 0.00370 (0.00327)  --> STEP: 435/639 -- GLOBAL_STEP: 20275 | > loss: -0.43988 (-0.45036) | > log_mle: -0.55090 (-0.57445) | > loss_dur: 0.11102 (0.12409) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.37794 (8.13842) | > current_lr: 0.00044 | > step_time: 0.76480 (0.58035) | > loader_time: 0.00360 (0.00329)  --> STEP: 460/639 -- GLOBAL_STEP: 20300 | > loss: -0.39071 (-0.44899) | > log_mle: -0.51682 (-0.57292) | > loss_dur: 0.12611 (0.12393) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.94665 (8.00094) | > current_lr: 0.00044 | > step_time: 0.84270 (0.59234) | > loader_time: 0.00380 (0.00331)  --> STEP: 485/639 -- GLOBAL_STEP: 20325 | > loss: -0.43509 (-0.44795) | > log_mle: -0.55093 (-0.57193) | > loss_dur: 0.11583 (0.12398) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.04773 (7.93993) | > current_lr: 0.00044 | > step_time: 0.81400 (0.60512) | > loader_time: 0.00410 (0.00334)  --> STEP: 510/639 -- GLOBAL_STEP: 20350 | > loss: -0.42936 (-0.44728) | > log_mle: -0.55495 (-0.57106) | > loss_dur: 0.12559 (0.12378) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.24434 (7.85584) | > current_lr: 0.00044 | > step_time: 0.91950 (0.61819) | > loader_time: 0.00410 (0.00336)  --> STEP: 535/639 -- GLOBAL_STEP: 20375 | > loss: -0.41961 (-0.44647) | > log_mle: -0.54772 (-0.57014) | > loss_dur: 0.12811 (0.12367) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.33085 (7.80720) | > current_lr: 0.00044 | > step_time: 0.90090 (0.63176) | > loader_time: 0.00400 (0.00339)  --> STEP: 560/639 -- GLOBAL_STEP: 20400 | > loss: -0.42566 (-0.44561) | > log_mle: -0.55234 (-0.56927) | > loss_dur: 0.12669 (0.12366) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.69906 (7.76354) | > current_lr: 0.00044 | > step_time: 0.90030 (0.64676) | > loader_time: 0.00420 (0.00342)  --> STEP: 585/639 -- GLOBAL_STEP: 20425 | > loss: -0.43882 (-0.44497) | > log_mle: -0.55911 (-0.56849) | > loss_dur: 0.12029 (0.12352) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.12517 (7.70970) | > current_lr: 0.00044 | > step_time: 1.05170 (0.66178) | > loader_time: 0.00420 (0.00346)  --> STEP: 610/639 -- GLOBAL_STEP: 20450 | > loss: -0.40901 (-0.44425) | > log_mle: -0.52678 (-0.56776) | > loss_dur: 0.11777 (0.12352) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.76515 (7.66469) | > current_lr: 0.00044 | > step_time: 1.05620 (0.67786) | > loader_time: 0.00430 (0.00349)  --> STEP: 635/639 -- GLOBAL_STEP: 20475 | > loss: -0.43061 (-0.44384) | > log_mle: -0.55557 (-0.56744) | > loss_dur: 0.12496 (0.12361) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.81217 (7.65728) | > current_lr: 0.00044 | > step_time: 1.12580 (0.69561) | > loader_time: 0.00490 (0.00354) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.01973 (-0.00562) | > avg_loss: -0.43808 (-0.00484) | > avg_log_mle: -0.56418 (-0.00262) | > avg_loss_dur: 0.12609 (-0.00221) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_20480.pth.tar  > EPOCH: 32/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 18:51:41)   --> STEP: 20/639 -- GLOBAL_STEP: 20500 | > loss: -0.55180 (-0.48696) | > log_mle: -0.72020 (-0.61840) | > loss_dur: 0.16840 (0.13144) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 20.02909 (9.86784) | > current_lr: 0.00044 | > step_time: 0.34400 (0.30254) | > loader_time: 0.00270 (0.00389)  --> STEP: 45/639 -- GLOBAL_STEP: 20525 | > loss: -0.39916 (-0.46512) | > log_mle: -0.52699 (-0.59387) | > loss_dur: 0.12782 (0.12875) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.83067 (8.81827) | > current_lr: 0.00044 | > step_time: 0.40910 (0.33270) | > loader_time: 0.00290 (0.00328)  --> STEP: 70/639 -- GLOBAL_STEP: 20550 | > loss: -0.47354 (-0.45668) | > log_mle: -0.59584 (-0.58537) | > loss_dur: 0.12230 (0.12869) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.02210 (8.47318) | > current_lr: 0.00044 | > step_time: 0.39500 (0.35890) | > loader_time: 0.00310 (0.00316)  --> STEP: 95/639 -- GLOBAL_STEP: 20575 | > loss: -0.53526 (-0.45626) | > log_mle: -0.64607 (-0.58246) | > loss_dur: 0.11082 (0.12621) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.54503 (8.29649) | > current_lr: 0.00044 | > step_time: 0.40600 (0.37523) | > loader_time: 0.00300 (0.00310)  --> STEP: 120/639 -- GLOBAL_STEP: 20600 | > loss: -0.50357 (-0.45574) | > log_mle: -0.64427 (-0.58231) | > loss_dur: 0.14070 (0.12658) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.27563 (8.35095) | > current_lr: 0.00044 | > step_time: 0.54380 (0.39396) | > loader_time: 0.00300 (0.00308)  --> STEP: 145/639 -- GLOBAL_STEP: 20625 | > loss: -0.46524 (-0.45813) | > log_mle: -0.59676 (-0.58400) | > loss_dur: 0.13152 (0.12588) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.85101 (8.56455) | > current_lr: 0.00044 | > step_time: 0.47790 (0.40860) | > loader_time: 0.00320 (0.00307)  --> STEP: 170/639 -- GLOBAL_STEP: 20650 | > loss: -0.47285 (-0.45878) | > log_mle: -0.58896 (-0.58406) | > loss_dur: 0.11612 (0.12528) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.89875 (8.58996) | > current_lr: 0.00044 | > step_time: 0.54480 (0.42226) | > loader_time: 0.00330 (0.00307)  --> STEP: 195/639 -- GLOBAL_STEP: 20675 | > loss: -0.41966 (-0.45913) | > log_mle: -0.52602 (-0.58373) | > loss_dur: 0.10636 (0.12460) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.85843 (8.48732) | > current_lr: 0.00044 | > step_time: 0.49030 (0.43903) | > loader_time: 0.00320 (0.00309)  --> STEP: 220/639 -- GLOBAL_STEP: 20700 | > loss: -0.47367 (-0.45791) | > log_mle: -0.58257 (-0.58222) | > loss_dur: 0.10890 (0.12431) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.26022 (8.35260) | > current_lr: 0.00044 | > step_time: 0.63290 (0.45508) | > loader_time: 0.00320 (0.00310)  --> STEP: 245/639 -- GLOBAL_STEP: 20725 | > loss: -0.45730 (-0.45996) | > log_mle: -0.56946 (-0.58366) | > loss_dur: 0.11216 (0.12370) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.14076 (8.42852) | > current_lr: 0.00044 | > step_time: 0.51840 (0.46938) | > loader_time: 0.00330 (0.00313)  --> STEP: 270/639 -- GLOBAL_STEP: 20750 | > loss: -0.50351 (-0.46048) | > log_mle: -0.63570 (-0.58426) | > loss_dur: 0.13219 (0.12378) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.69388 (8.60645) | > current_lr: 0.00044 | > step_time: 0.83880 (0.48555) | > loader_time: 0.00380 (0.00316)  --> STEP: 295/639 -- GLOBAL_STEP: 20775 | > loss: -0.41987 (-0.45921) | > log_mle: -0.52790 (-0.58301) | > loss_dur: 0.10803 (0.12380) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.87488 (8.49479) | > current_lr: 0.00044 | > step_time: 0.64340 (0.50053) | > loader_time: 0.00330 (0.00318)  --> STEP: 320/639 -- GLOBAL_STEP: 20800 | > loss: -0.44154 (-0.45833) | > log_mle: -0.57590 (-0.58231) | > loss_dur: 0.13436 (0.12398) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.57040 (8.46376) | > current_lr: 0.00044 | > step_time: 0.73430 (0.51668) | > loader_time: 0.00360 (0.00321)  --> STEP: 345/639 -- GLOBAL_STEP: 20825 | > loss: -0.45135 (-0.45713) | > log_mle: -0.56615 (-0.58123) | > loss_dur: 0.11480 (0.12410) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.22124 (8.43221) | > current_lr: 0.00044 | > step_time: 0.72280 (0.53148) | > loader_time: 0.00350 (0.00322)  --> STEP: 370/639 -- GLOBAL_STEP: 20850 | > loss: -0.45949 (-0.45582) | > log_mle: -0.57967 (-0.57958) | > loss_dur: 0.12018 (0.12376) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.43736 (8.29367) | > current_lr: 0.00044 | > step_time: 0.87680 (0.54461) | > loader_time: 0.00380 (0.00324)  --> STEP: 395/639 -- GLOBAL_STEP: 20875 | > loss: -0.42204 (-0.45445) | > log_mle: -0.54647 (-0.57814) | > loss_dur: 0.12443 (0.12369) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.48048 (8.21262) | > current_lr: 0.00044 | > step_time: 0.70720 (0.55704) | > loader_time: 0.00350 (0.00326)  --> STEP: 420/639 -- GLOBAL_STEP: 20900 | > loss: -0.42795 (-0.45312) | > log_mle: -0.54582 (-0.57685) | > loss_dur: 0.11787 (0.12373) | > amp_scaler: 8192.00000 (4154.51429) | > grad_norm: 5.63171 (8.11691) | > current_lr: 0.00044 | > step_time: 0.72940 (0.57015) | > loader_time: 0.00360 (0.00328)  --> STEP: 445/639 -- GLOBAL_STEP: 20925 | > loss: -0.43025 (-0.45220) | > log_mle: -0.54889 (-0.57575) | > loss_dur: 0.11864 (0.12355) | > amp_scaler: 4096.00000 (4224.86292) | > grad_norm: 7.54133 (8.08393) | > current_lr: 0.00044 | > step_time: 0.75120 (0.58243) | > loader_time: 0.00360 (0.00330)  --> STEP: 470/639 -- GLOBAL_STEP: 20950 | > loss: -0.44871 (-0.45091) | > log_mle: -0.56456 (-0.57434) | > loss_dur: 0.11585 (0.12344) | > amp_scaler: 4096.00000 (4218.00851) | > grad_norm: 9.15504 (7.99224) | > current_lr: 0.00044 | > step_time: 0.77410 (0.59441) | > loader_time: 0.00370 (0.00332)  --> STEP: 495/639 -- GLOBAL_STEP: 20975 | > loss: -0.39131 (-0.45000) | > log_mle: -0.49962 (-0.57335) | > loss_dur: 0.10831 (0.12336) | > amp_scaler: 4096.00000 (4211.84646) | > grad_norm: 1.75641 (7.95345) | > current_lr: 0.00044 | > step_time: 0.81360 (0.60776) | > loader_time: 0.00350 (0.00334)  --> STEP: 520/639 -- GLOBAL_STEP: 21000 | > loss: -0.44349 (-0.44943) | > log_mle: -0.55596 (-0.57264) | > loss_dur: 0.11247 (0.12320) | > amp_scaler: 4096.00000 (4206.27692) | > grad_norm: 7.01204 (7.89830) | > current_lr: 0.00044 | > step_time: 0.82860 (0.62029) | > loader_time: 0.00390 (0.00337)  --> STEP: 545/639 -- GLOBAL_STEP: 21025 | > loss: -0.46482 (-0.44864) | > log_mle: -0.59056 (-0.57182) | > loss_dur: 0.12574 (0.12318) | > amp_scaler: 4096.00000 (4201.21835) | > grad_norm: 9.26762 (7.85555) | > current_lr: 0.00044 | > step_time: 0.97080 (0.63553) | > loader_time: 0.00400 (0.00340)  --> STEP: 570/639 -- GLOBAL_STEP: 21050 | > loss: -0.42527 (-0.44758) | > log_mle: -0.53598 (-0.57070) | > loss_dur: 0.11071 (0.12312) | > amp_scaler: 4096.00000 (4196.60351) | > grad_norm: 3.43391 (7.79796) | > current_lr: 0.00044 | > step_time: 0.85070 (0.64951) | > loader_time: 0.00360 (0.00343)  --> STEP: 595/639 -- GLOBAL_STEP: 21075 | > loss: -0.41545 (-0.44701) | > log_mle: -0.52825 (-0.57010) | > loss_dur: 0.11280 (0.12310) | > amp_scaler: 4096.00000 (4192.37647) | > grad_norm: 5.35353 (7.74655) | > current_lr: 0.00044 | > step_time: 0.94380 (0.66608) | > loader_time: 0.00370 (0.00346)  --> STEP: 620/639 -- GLOBAL_STEP: 21100 | > loss: -0.42831 (-0.44630) | > log_mle: -0.54097 (-0.56934) | > loss_dur: 0.11266 (0.12304) | > amp_scaler: 4096.00000 (4188.49032) | > grad_norm: 7.00853 (7.71034) | > current_lr: 0.00044 | > step_time: 1.17940 (0.68234) | > loader_time: 0.00470 (0.00350) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02356 (+0.00383) | > avg_loss: -0.43846 (-0.00037) | > avg_log_mle: -0.56583 (-0.00165) | > avg_loss_dur: 0.12737 (+0.00128) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_21120.pth.tar  > EPOCH: 33/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 18:59:26)   --> STEP: 5/639 -- GLOBAL_STEP: 21125 | > loss: -0.49710 (-0.49384) | > log_mle: -0.65833 (-0.63391) | > loss_dur: 0.16123 (0.14007) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.70513 (11.02437) | > current_lr: 0.00044 | > step_time: 0.33700 (0.28525) | > loader_time: 0.00090 (0.00209)  --> STEP: 30/639 -- GLOBAL_STEP: 21150 | > loss: -0.47325 (-0.47374) | > log_mle: -0.60101 (-0.60466) | > loss_dur: 0.12776 (0.13092) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.40394 (9.65811) | > current_lr: 0.00044 | > step_time: 0.33460 (0.33518) | > loader_time: 0.00270 (0.00257)  --> STEP: 55/639 -- GLOBAL_STEP: 21175 | > loss: -0.41812 (-0.46460) | > log_mle: -0.52820 (-0.59317) | > loss_dur: 0.11008 (0.12858) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.35892 (8.88601) | > current_lr: 0.00043 | > step_time: 0.32910 (0.35623) | > loader_time: 0.00270 (0.00270)  --> STEP: 80/639 -- GLOBAL_STEP: 21200 | > loss: -0.44328 (-0.45984) | > log_mle: -0.58192 (-0.58666) | > loss_dur: 0.13864 (0.12683) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.26046 (8.33929) | > current_lr: 0.00043 | > step_time: 0.41550 (0.37732) | > loader_time: 0.00290 (0.00278)  --> STEP: 105/639 -- GLOBAL_STEP: 21225 | > loss: -0.40744 (-0.45914) | > log_mle: -0.53715 (-0.58517) | > loss_dur: 0.12971 (0.12603) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.97887 (8.15509) | > current_lr: 0.00043 | > step_time: 0.43480 (0.39213) | > loader_time: 0.00300 (0.00285)  --> STEP: 130/639 -- GLOBAL_STEP: 21250 | > loss: -0.44265 (-0.45871) | > log_mle: -0.57680 (-0.58482) | > loss_dur: 0.13416 (0.12610) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.22036 (8.34492) | > current_lr: 0.00043 | > step_time: 0.49800 (0.40775) | > loader_time: 0.00290 (0.00288)  --> STEP: 155/639 -- GLOBAL_STEP: 21275 | > loss: -0.43146 (-0.45956) | > log_mle: -0.56953 (-0.58519) | > loss_dur: 0.13807 (0.12563) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.84971 (8.41172) | > current_lr: 0.00043 | > step_time: 0.53470 (0.42348) | > loader_time: 0.00310 (0.00291)  --> STEP: 180/639 -- GLOBAL_STEP: 21300 | > loss: -0.49786 (-0.46245) | > log_mle: -0.62324 (-0.58762) | > loss_dur: 0.12538 (0.12517) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.22475 (8.70349) | > current_lr: 0.00043 | > step_time: 0.57980 (0.43797) | > loader_time: 0.00340 (0.00295)  --> STEP: 205/639 -- GLOBAL_STEP: 21325 | > loss: -0.42674 (-0.46021) | > log_mle: -0.54585 (-0.58470) | > loss_dur: 0.11911 (0.12449) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.51020 (8.44945) | > current_lr: 0.00043 | > step_time: 0.55690 (0.45179) | > loader_time: 0.00340 (0.00298)  --> STEP: 230/639 -- GLOBAL_STEP: 21350 | > loss: -0.45973 (-0.46098) | > log_mle: -0.57943 (-0.58516) | > loss_dur: 0.11970 (0.12417) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.36279 (8.56884) | > current_lr: 0.00043 | > step_time: 0.55420 (0.46853) | > loader_time: 0.00340 (0.00302)  --> STEP: 255/639 -- GLOBAL_STEP: 21375 | > loss: -0.42976 (-0.46110) | > log_mle: -0.55606 (-0.58477) | > loss_dur: 0.12630 (0.12367) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.10700 (8.52544) | > current_lr: 0.00043 | > step_time: 0.62530 (0.48153) | > loader_time: 0.00330 (0.00304)  --> STEP: 280/639 -- GLOBAL_STEP: 21400 | > loss: -0.44866 (-0.46169) | > log_mle: -0.58197 (-0.58550) | > loss_dur: 0.13331 (0.12381) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.54034 (8.71597) | > current_lr: 0.00043 | > step_time: 0.60620 (0.49737) | > loader_time: 0.00310 (0.00306)  --> STEP: 305/639 -- GLOBAL_STEP: 21425 | > loss: -0.43422 (-0.46063) | > log_mle: -0.56322 (-0.58451) | > loss_dur: 0.12900 (0.12388) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.33540 (8.67935) | > current_lr: 0.00043 | > step_time: 0.69020 (0.51156) | > loader_time: 0.00350 (0.00310)  --> STEP: 330/639 -- GLOBAL_STEP: 21450 | > loss: -0.45121 (-0.45904) | > log_mle: -0.58668 (-0.58326) | > loss_dur: 0.13547 (0.12422) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.93319 (8.65549) | > current_lr: 0.00043 | > step_time: 0.73460 (0.52661) | > loader_time: 0.00350 (0.00313)  --> STEP: 355/639 -- GLOBAL_STEP: 21475 | > loss: -0.39952 (-0.45815) | > log_mle: -0.51252 (-0.58201) | > loss_dur: 0.11300 (0.12386) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.96996 (8.52297) | > current_lr: 0.00043 | > step_time: 0.61490 (0.53984) | > loader_time: 0.00340 (0.00316)  --> STEP: 380/639 -- GLOBAL_STEP: 21500 | > loss: -0.44281 (-0.45676) | > log_mle: -0.54935 (-0.58038) | > loss_dur: 0.10653 (0.12362) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.49452 (8.33992) | > current_lr: 0.00043 | > step_time: 0.64180 (0.55398) | > loader_time: 0.00350 (0.00318)  --> STEP: 405/639 -- GLOBAL_STEP: 21525 | > loss: -0.44177 (-0.45598) | > log_mle: -0.55758 (-0.57967) | > loss_dur: 0.11581 (0.12368) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.26760 (8.30670) | > current_lr: 0.00043 | > step_time: 0.76480 (0.56567) | > loader_time: 0.00410 (0.00321)  --> STEP: 430/639 -- GLOBAL_STEP: 21550 | > loss: -0.39898 (-0.45459) | > log_mle: -0.51357 (-0.57826) | > loss_dur: 0.11460 (0.12367) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.62750 (8.24257) | > current_lr: 0.00043 | > step_time: 0.75250 (0.57843) | > loader_time: 0.00370 (0.00324)  --> STEP: 455/639 -- GLOBAL_STEP: 21575 | > loss: -0.41529 (-0.45326) | > log_mle: -0.53842 (-0.57671) | > loss_dur: 0.12313 (0.12345) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.92075 (8.09554) | > current_lr: 0.00043 | > step_time: 0.73840 (0.58939) | > loader_time: 0.00370 (0.00326)  --> STEP: 480/639 -- GLOBAL_STEP: 21600 | > loss: -0.38022 (-0.45199) | > log_mle: -0.49752 (-0.57547) | > loss_dur: 0.11731 (0.12349) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.05814 (8.03741) | > current_lr: 0.00043 | > step_time: 0.82160 (0.60261) | > loader_time: 0.00360 (0.00329)  --> STEP: 505/639 -- GLOBAL_STEP: 21625 | > loss: -0.44224 (-0.45111) | > log_mle: -0.55880 (-0.57441) | > loss_dur: 0.11656 (0.12330) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.03870 (8.03893) | > current_lr: 0.00043 | > step_time: 0.92800 (0.61524) | > loader_time: 0.00400 (0.00332)  --> STEP: 530/639 -- GLOBAL_STEP: 21650 | > loss: -0.42755 (-0.45033) | > log_mle: -0.55695 (-0.57361) | > loss_dur: 0.12940 (0.12327) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.98515 (8.01856) | > current_lr: 0.00043 | > step_time: 0.95940 (0.62913) | > loader_time: 0.00410 (0.00335)  --> STEP: 555/639 -- GLOBAL_STEP: 21675 | > loss: -0.42167 (-0.44936) | > log_mle: -0.54074 (-0.57251) | > loss_dur: 0.11907 (0.12315) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.07789 (7.92634) | > current_lr: 0.00043 | > step_time: 0.84500 (0.64340) | > loader_time: 0.00380 (0.00338)  --> STEP: 580/639 -- GLOBAL_STEP: 21700 | > loss: -0.45284 (-0.44875) | > log_mle: -0.56353 (-0.57183) | > loss_dur: 0.11069 (0.12308) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.68115 (7.89436) | > current_lr: 0.00043 | > step_time: 0.94000 (0.65844) | > loader_time: 0.00440 (0.00342)  --> STEP: 605/639 -- GLOBAL_STEP: 21725 | > loss: -0.41465 (-0.44811) | > log_mle: -0.52611 (-0.57118) | > loss_dur: 0.11146 (0.12306) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.64518 (7.87056) | > current_lr: 0.00043 | > step_time: 1.05210 (0.67480) | > loader_time: 0.00450 (0.00346)  --> STEP: 630/639 -- GLOBAL_STEP: 21750 | > loss: -0.42159 (-0.44754) | > log_mle: -0.55524 (-0.57065) | > loss_dur: 0.13365 (0.12311) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.45833 (7.86907) | > current_lr: 0.00043 | > step_time: 1.22690 (0.69188) | > loader_time: 0.00450 (0.00350) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03038 (+0.00681) | > avg_loss: -0.43135 (+0.00711) | > avg_log_mle: -0.56052 (+0.00530) | > avg_loss_dur: 0.12918 (+0.00181)  > EPOCH: 34/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 19:07:06)   --> STEP: 15/639 -- GLOBAL_STEP: 21775 | > loss: -0.58682 (-0.49065) | > log_mle: -0.73550 (-0.62041) | > loss_dur: 0.14868 (0.12976) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 25.67815 (11.75068) | > current_lr: 0.00043 | > step_time: 0.34370 (0.29847) | > loader_time: 0.00250 (0.00260)  --> STEP: 40/639 -- GLOBAL_STEP: 21800 | > loss: -0.45176 (-0.46951) | > log_mle: -0.57296 (-0.59653) | > loss_dur: 0.12120 (0.12702) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.88791 (9.25125) | > current_lr: 0.00043 | > step_time: 0.35250 (0.33342) | > loader_time: 0.00280 (0.00271)  --> STEP: 65/639 -- GLOBAL_STEP: 21825 | > loss: -0.46089 (-0.46296) | > log_mle: -0.57073 (-0.59092) | > loss_dur: 0.10984 (0.12797) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.01419 (8.91816) | > current_lr: 0.00043 | > step_time: 0.38430 (0.35977) | > loader_time: 0.00280 (0.00278)  --> STEP: 90/639 -- GLOBAL_STEP: 21850 | > loss: -0.40447 (-0.45828) | > log_mle: -0.52548 (-0.58417) | > loss_dur: 0.12101 (0.12588) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.53543 (8.24096) | > current_lr: 0.00043 | > step_time: 0.51680 (0.37651) | > loader_time: 0.00310 (0.00282)  --> STEP: 115/639 -- GLOBAL_STEP: 21875 | > loss: -0.35552 (-0.46042) | > log_mle: -0.49550 (-0.58613) | > loss_dur: 0.13998 (0.12571) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.35287 (8.32817) | > current_lr: 0.00043 | > step_time: 0.40670 (0.39391) | > loader_time: 0.00300 (0.00288)  --> STEP: 140/639 -- GLOBAL_STEP: 21900 | > loss: -0.44140 (-0.46242) | > log_mle: -0.58070 (-0.58776) | > loss_dur: 0.13930 (0.12534) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.66306 (8.71409) | > current_lr: 0.00043 | > step_time: 0.55910 (0.40709) | > loader_time: 0.00330 (0.00291)  --> STEP: 165/639 -- GLOBAL_STEP: 21925 | > loss: -0.45546 (-0.46269) | > log_mle: -0.59261 (-0.58735) | > loss_dur: 0.13715 (0.12466) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.29093 (8.69504) | > current_lr: 0.00043 | > step_time: 0.49680 (0.42177) | > loader_time: 0.00290 (0.00295)  --> STEP: 190/639 -- GLOBAL_STEP: 21950 | > loss: -0.38895 (-0.46357) | > log_mle: -0.51570 (-0.58766) | > loss_dur: 0.12675 (0.12409) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.26083 (8.65111) | > current_lr: 0.00043 | > step_time: 0.52480 (0.44044) | > loader_time: 0.00320 (0.00299)  --> STEP: 215/639 -- GLOBAL_STEP: 21975 | > loss: -0.51783 (-0.46225) | > log_mle: -0.62819 (-0.58630) | > loss_dur: 0.11036 (0.12405) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.49470 (8.56548) | > current_lr: 0.00043 | > step_time: 0.55210 (0.45688) | > loader_time: 0.00320 (0.00302)  --> STEP: 240/639 -- GLOBAL_STEP: 22000 | > loss: -0.51606 (-0.46426) | > log_mle: -0.64040 (-0.58767) | > loss_dur: 0.12434 (0.12341) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.01276 (8.61647) | > current_lr: 0.00043 | > step_time: 0.56320 (0.47184) | > loader_time: 0.00330 (0.00305)  --> STEP: 265/639 -- GLOBAL_STEP: 22025 | > loss: -0.50754 (-0.46392) | > log_mle: -0.65052 (-0.58752) | > loss_dur: 0.14298 (0.12360) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.82650 (8.72942) | > current_lr: 0.00043 | > step_time: 0.61950 (0.48698) | > loader_time: 0.00300 (0.00309)  --> STEP: 290/639 -- GLOBAL_STEP: 22050 | > loss: -0.43944 (-0.46302) | > log_mle: -0.55786 (-0.58654) | > loss_dur: 0.11842 (0.12352) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.26181 (8.62177) | > current_lr: 0.00043 | > step_time: 0.62320 (0.50348) | > loader_time: 0.00350 (0.00312)  --> STEP: 315/639 -- GLOBAL_STEP: 22075 | > loss: -0.42132 (-0.46221) | > log_mle: -0.53558 (-0.58565) | > loss_dur: 0.11427 (0.12345) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.27228 (8.56291) | > current_lr: 0.00043 | > step_time: 0.64460 (0.51790) | > loader_time: 0.00350 (0.00315)  --> STEP: 340/639 -- GLOBAL_STEP: 22100 | > loss: -0.45107 (-0.46120) | > log_mle: -0.56295 (-0.58494) | > loss_dur: 0.11188 (0.12374) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.20746 (8.54205) | > current_lr: 0.00043 | > step_time: 0.59960 (0.53355) | > loader_time: 0.00350 (0.00319)  --> STEP: 365/639 -- GLOBAL_STEP: 22125 | > loss: -0.45525 (-0.45978) | > log_mle: -0.57095 (-0.58314) | > loss_dur: 0.11570 (0.12336) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.51816 (8.39374) | > current_lr: 0.00043 | > step_time: 0.81620 (0.54464) | > loader_time: 0.00400 (0.00322)  --> STEP: 390/639 -- GLOBAL_STEP: 22150 | > loss: -0.43941 (-0.45852) | > log_mle: -0.56754 (-0.58168) | > loss_dur: 0.12812 (0.12316) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.87587 (8.25828) | > current_lr: 0.00043 | > step_time: 0.75320 (0.55843) | > loader_time: 0.00370 (0.00325)  --> STEP: 415/639 -- GLOBAL_STEP: 22175 | > loss: -0.41109 (-0.45732) | > log_mle: -0.54020 (-0.58055) | > loss_dur: 0.12911 (0.12323) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.33453 (8.21521) | > current_lr: 0.00042 | > step_time: 0.70810 (0.57135) | > loader_time: 0.00370 (0.00328)  --> STEP: 440/639 -- GLOBAL_STEP: 22200 | > loss: -0.39979 (-0.45630) | > log_mle: -0.52471 (-0.57939) | > loss_dur: 0.12492 (0.12309) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.46670 (8.16392) | > current_lr: 0.00042 | > step_time: 0.69920 (0.58359) | > loader_time: 0.00370 (0.00330)  --> STEP: 465/639 -- GLOBAL_STEP: 22225 | > loss: -0.41396 (-0.45485) | > log_mle: -0.52953 (-0.57778) | > loss_dur: 0.11556 (0.12293) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.76317 (8.01146) | > current_lr: 0.00042 | > step_time: 0.79040 (0.59488) | > loader_time: 0.00370 (0.00333)  --> STEP: 490/639 -- GLOBAL_STEP: 22250 | > loss: -0.40431 (-0.45376) | > log_mle: -0.51278 (-0.57669) | > loss_dur: 0.10847 (0.12294) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.23672 (7.99917) | > current_lr: 0.00042 | > step_time: 0.77060 (0.60749) | > loader_time: 0.00360 (0.00335)  --> STEP: 515/639 -- GLOBAL_STEP: 22275 | > loss: -0.42368 (-0.45318) | > log_mle: -0.54456 (-0.57599) | > loss_dur: 0.12088 (0.12281) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.26960 (7.99459) | > current_lr: 0.00042 | > step_time: 1.05540 (0.62032) | > loader_time: 0.00400 (0.00338)  --> STEP: 540/639 -- GLOBAL_STEP: 22300 | > loss: -0.44003 (-0.45225) | > log_mle: -0.56176 (-0.57499) | > loss_dur: 0.12173 (0.12273) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.50288 (7.95328) | > current_lr: 0.00042 | > step_time: 0.89330 (0.63470) | > loader_time: 0.00430 (0.00342)  --> STEP: 565/639 -- GLOBAL_STEP: 22325 | > loss: -0.40359 (-0.45116) | > log_mle: -0.52117 (-0.57387) | > loss_dur: 0.11758 (0.12271) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.92737 (7.89155) | > current_lr: 0.00042 | > step_time: 1.01390 (0.64919) | > loader_time: 0.00440 (0.00345)  --> STEP: 590/639 -- GLOBAL_STEP: 22350 | > loss: -0.43220 (-0.45071) | > log_mle: -0.55535 (-0.57334) | > loss_dur: 0.12315 (0.12263) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.19264 (7.84896) | > current_lr: 0.00042 | > step_time: 0.95590 (0.66510) | > loader_time: 0.00420 (0.00348)  --> STEP: 615/639 -- GLOBAL_STEP: 22375 | > loss: -0.43711 (-0.44996) | > log_mle: -0.55928 (-0.57254) | > loss_dur: 0.12216 (0.12257) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.67913 (7.83212) | > current_lr: 0.00042 | > step_time: 1.03020 (0.68132) | > loader_time: 0.00440 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03546 (+0.00509) | > avg_loss: -0.43823 (-0.00688) | > avg_log_mle: -0.56519 (-0.00466) | > avg_loss_dur: 0.12696 (-0.00222)  > EPOCH: 35/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 19:14:47)   --> STEP: 0/639 -- GLOBAL_STEP: 22400 | > loss: -0.51555 (-0.51555) | > log_mle: -0.67205 (-0.67205) | > loss_dur: 0.15650 (0.15650) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 28.65970 (28.65970) | > current_lr: 0.00042 | > step_time: 0.70190 (0.70193) | > loader_time: 0.39500 (0.39502)  --> STEP: 25/639 -- GLOBAL_STEP: 22425 | > loss: -0.46904 (-0.48870) | > log_mle: -0.61617 (-0.61786) | > loss_dur: 0.14713 (0.12917) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.01802 (9.09726) | > current_lr: 0.00042 | > step_time: 0.43700 (0.31983) | > loader_time: 0.00290 (0.00258)  --> STEP: 50/639 -- GLOBAL_STEP: 22450 | > loss: -0.42900 (-0.47370) | > log_mle: -0.54294 (-0.60085) | > loss_dur: 0.11394 (0.12716) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.54736 (8.34808) | > current_lr: 0.00042 | > step_time: 0.33440 (0.33928) | > loader_time: 0.00270 (0.00267)  --> STEP: 75/639 -- GLOBAL_STEP: 22475 | > loss: -0.46547 (-0.46600) | > log_mle: -0.58188 (-0.59259) | > loss_dur: 0.11641 (0.12660) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.53457 (8.41879) | > current_lr: 0.00042 | > step_time: 0.41480 (0.36490) | > loader_time: 0.00290 (0.00272)  --> STEP: 100/639 -- GLOBAL_STEP: 22500 | > loss: -0.48000 (-0.46573) | > log_mle: -0.59458 (-0.59058) | > loss_dur: 0.11459 (0.12485) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.26950 (8.49133) | > current_lr: 0.00042 | > step_time: 0.50020 (0.38323) | > loader_time: 0.00310 (0.00277)  --> STEP: 125/639 -- GLOBAL_STEP: 22525 | > loss: -0.46964 (-0.46466) | > log_mle: -0.59130 (-0.58970) | > loss_dur: 0.12166 (0.12504) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.14982 (8.73391) | > current_lr: 0.00042 | > step_time: 0.42000 (0.39997) | > loader_time: 0.00290 (0.00281)  --> STEP: 150/639 -- GLOBAL_STEP: 22550 | > loss: -0.42922 (-0.46527) | > log_mle: -0.55276 (-0.58995) | > loss_dur: 0.12353 (0.12468) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.92295 (8.91273) | > current_lr: 0.00042 | > step_time: 0.44080 (0.41493) | > loader_time: 0.00270 (0.00284)  --> STEP: 175/639 -- GLOBAL_STEP: 22575 | > loss: -0.46569 (-0.46689) | > log_mle: -0.57355 (-0.59110) | > loss_dur: 0.10786 (0.12421) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.99990 (8.99246) | > current_lr: 0.00042 | > step_time: 0.57360 (0.43130) | > loader_time: 0.00320 (0.00288)  --> STEP: 200/639 -- GLOBAL_STEP: 22600 | > loss: -0.43638 (-0.46588) | > log_mle: -0.55871 (-0.58941) | > loss_dur: 0.12234 (0.12353) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.34498 (8.79090) | > current_lr: 0.00042 | > step_time: 0.52950 (0.44640) | > loader_time: 0.00290 (0.00292)  --> STEP: 225/639 -- GLOBAL_STEP: 22625 | > loss: -0.43442 (-0.46478) | > log_mle: -0.53569 (-0.58790) | > loss_dur: 0.10127 (0.12312) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.55346 (8.59432) | > current_lr: 0.00042 | > step_time: 0.52440 (0.46229) | > loader_time: 0.00320 (0.00296)  --> STEP: 250/639 -- GLOBAL_STEP: 22650 | > loss: -0.43742 (-0.46667) | > log_mle: -0.54004 (-0.58942) | > loss_dur: 0.10262 (0.12275) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.07288 (8.69441) | > current_lr: 0.00042 | > step_time: 0.69800 (0.47763) | > loader_time: 0.00330 (0.00300)  --> STEP: 275/639 -- GLOBAL_STEP: 22675 | > loss: -0.46571 (-0.46744) | > log_mle: -0.58531 (-0.59041) | > loss_dur: 0.11961 (0.12298) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.63484 (8.85012) | > current_lr: 0.00042 | > step_time: 0.86050 (0.49453) | > loader_time: 0.00350 (0.00304)  --> STEP: 300/639 -- GLOBAL_STEP: 22700 | > loss: -0.46206 (-0.46580) | > log_mle: -0.57757 (-0.58892) | > loss_dur: 0.11551 (0.12312) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.22549 (8.75443) | > current_lr: 0.00042 | > step_time: 0.70490 (0.50858) | > loader_time: 0.00350 (0.00307)  --> STEP: 325/639 -- GLOBAL_STEP: 22725 | > loss: -0.45568 (-0.46457) | > log_mle: -0.57785 (-0.58788) | > loss_dur: 0.12217 (0.12331) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.01313 (8.72072) | > current_lr: 0.00042 | > step_time: 0.60640 (0.52431) | > loader_time: 0.00340 (0.00310)  --> STEP: 350/639 -- GLOBAL_STEP: 22750 | > loss: -0.47668 (-0.46337) | > log_mle: -0.60580 (-0.58652) | > loss_dur: 0.12912 (0.12316) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.33495 (8.69464) | > current_lr: 0.00042 | > step_time: 0.85720 (0.53825) | > loader_time: 0.00350 (0.00313)  --> STEP: 375/639 -- GLOBAL_STEP: 22775 | > loss: -0.42587 (-0.46164) | > log_mle: -0.54104 (-0.58459) | > loss_dur: 0.11517 (0.12295) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.28707 (8.54012) | > current_lr: 0.00042 | > step_time: 0.75790 (0.55110) | > loader_time: 0.00300 (0.00316)  --> STEP: 400/639 -- GLOBAL_STEP: 22800 | > loss: -0.45483 (-0.46054) | > log_mle: -0.56948 (-0.58354) | > loss_dur: 0.11464 (0.12300) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.86795 (8.46721) | > current_lr: 0.00042 | > step_time: 0.74990 (0.56351) | > loader_time: 0.00380 (0.00318)  --> STEP: 425/639 -- GLOBAL_STEP: 22825 | > loss: -0.50113 (-0.45898) | > log_mle: -0.62464 (-0.58187) | > loss_dur: 0.12352 (0.12290) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.83151 (8.39413) | > current_lr: 0.00042 | > step_time: 0.91490 (0.57568) | > loader_time: 0.00370 (0.00322)  --> STEP: 450/639 -- GLOBAL_STEP: 22850 | > loss: -0.41957 (-0.45772) | > log_mle: -0.52863 (-0.58040) | > loss_dur: 0.10906 (0.12268) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.83091 (8.28776) | > current_lr: 0.00042 | > step_time: 0.78670 (0.58739) | > loader_time: 0.00390 (0.00324)  --> STEP: 475/639 -- GLOBAL_STEP: 22875 | > loss: -0.43000 (-0.45633) | > log_mle: -0.54002 (-0.57904) | > loss_dur: 0.11002 (0.12271) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.11476 (8.20484) | > current_lr: 0.00042 | > step_time: 0.72720 (0.59975) | > loader_time: 0.00360 (0.00327)  --> STEP: 500/639 -- GLOBAL_STEP: 22900 | > loss: -0.47871 (-0.45553) | > log_mle: -0.60085 (-0.57815) | > loss_dur: 0.12214 (0.12263) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.33169 (8.17435) | > current_lr: 0.00042 | > step_time: 0.96340 (0.61233) | > loader_time: 0.00390 (0.00329)  --> STEP: 525/639 -- GLOBAL_STEP: 22925 | > loss: -0.45287 (-0.45471) | > log_mle: -0.58570 (-0.57729) | > loss_dur: 0.13283 (0.12257) | > amp_scaler: 8192.00000 (4228.63238) | > grad_norm: 7.99843 (8.10744) | > current_lr: 0.00042 | > step_time: 1.04230 (0.62565) | > loader_time: 0.00400 (0.00332)  --> STEP: 550/639 -- GLOBAL_STEP: 22950 | > loss: -0.39058 (-0.45371) | > log_mle: -0.51713 (-0.57615) | > loss_dur: 0.12655 (0.12244) | > amp_scaler: 8192.00000 (4408.78545) | > grad_norm: 2.32128 (8.02317) | > current_lr: 0.00042 | > step_time: 0.95970 (0.64049) | > loader_time: 0.00400 (0.00335)  --> STEP: 575/639 -- GLOBAL_STEP: 22975 | > loss: -0.41870 (-0.45263) | > log_mle: -0.53444 (-0.57504) | > loss_dur: 0.11574 (0.12241) | > amp_scaler: 8192.00000 (4573.27304) | > grad_norm: 5.05917 (7.92943) | > current_lr: 0.00042 | > step_time: 0.89580 (0.65521) | > loader_time: 0.00420 (0.00338)  --> STEP: 600/639 -- GLOBAL_STEP: 23000 | > loss: -0.40832 (-0.45188) | > log_mle: -0.53165 (-0.57432) | > loss_dur: 0.12333 (0.12244) | > amp_scaler: 8192.00000 (4724.05333) | > grad_norm: 5.99197 (7.85868) | > current_lr: 0.00042 | > step_time: 1.06660 (0.67199) | > loader_time: 0.00410 (0.00343)  --> STEP: 625/639 -- GLOBAL_STEP: 23025 | > loss: -0.43910 (-0.45122) | > log_mle: -0.56672 (-0.57358) | > loss_dur: 0.12762 (0.12237) | > amp_scaler: 8192.00000 (4862.77120) | > grad_norm: 6.43097 (7.80340) | > current_lr: 0.00042 | > step_time: 1.09910 (0.68875) | > loader_time: 0.00690 (0.00348) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02842 (-0.00704) | > avg_loss: -0.44065 (-0.00242) | > avg_log_mle: -0.57016 (-0.00498) | > avg_loss_dur: 0.12951 (+0.00255) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_23040.pth.tar  > EPOCH: 36/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 19:22:34)   --> STEP: 10/639 -- GLOBAL_STEP: 23050 | > loss: -0.49001 (-0.49258) | > log_mle: -0.62770 (-0.62580) | > loss_dur: 0.13769 (0.13322) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 11.29461 (10.30419) | > current_lr: 0.00042 | > step_time: 0.28630 (0.29428) | > loader_time: 0.00260 (0.00244)  --> STEP: 35/639 -- GLOBAL_STEP: 23075 | > loss: -0.47439 (-0.47681) | > log_mle: -0.58192 (-0.60403) | > loss_dur: 0.10753 (0.12723) | > amp_scaler: 4096.00000 (6085.48571) | > grad_norm: 4.49071 (8.47952) | > current_lr: 0.00042 | > step_time: 0.31760 (0.32380) | > loader_time: 0.00280 (0.00260)  --> STEP: 60/639 -- GLOBAL_STEP: 23100 | > loss: -0.42261 (-0.46790) | > log_mle: -0.53750 (-0.59510) | > loss_dur: 0.11489 (0.12720) | > amp_scaler: 4096.00000 (5256.53333) | > grad_norm: 5.19064 (8.35231) | > current_lr: 0.00042 | > step_time: 0.36380 (0.35669) | > loader_time: 0.00280 (0.00272)  --> STEP: 85/639 -- GLOBAL_STEP: 23125 | > loss: -0.41456 (-0.46316) | > log_mle: -0.53957 (-0.58793) | > loss_dur: 0.12501 (0.12478) | > amp_scaler: 4096.00000 (4915.20000) | > grad_norm: 3.88669 (7.96912) | > current_lr: 0.00042 | > step_time: 0.40950 (0.37108) | > loader_time: 0.00290 (0.00278)  --> STEP: 110/639 -- GLOBAL_STEP: 23150 | > loss: -0.43522 (-0.46623) | > log_mle: -0.56546 (-0.59127) | > loss_dur: 0.13023 (0.12504) | > amp_scaler: 4096.00000 (4729.01818) | > grad_norm: 4.80373 (8.37734) | > current_lr: 0.00042 | > step_time: 0.41090 (0.39325) | > loader_time: 0.00300 (0.00284)  --> STEP: 135/639 -- GLOBAL_STEP: 23175 | > loss: -0.41018 (-0.46578) | > log_mle: -0.54226 (-0.58987) | > loss_dur: 0.13208 (0.12409) | > amp_scaler: 4096.00000 (4611.79259) | > grad_norm: 9.42865 (8.52778) | > current_lr: 0.00042 | > step_time: 0.46220 (0.40472) | > loader_time: 0.00300 (0.00287)  --> STEP: 160/639 -- GLOBAL_STEP: 23200 | > loss: -0.47548 (-0.46524) | > log_mle: -0.58707 (-0.58861) | > loss_dur: 0.11160 (0.12337) | > amp_scaler: 4096.00000 (4531.20000) | > grad_norm: 7.28749 (8.42057) | > current_lr: 0.00042 | > step_time: 0.47730 (0.42047) | > loader_time: 0.00310 (0.00294)  --> STEP: 185/639 -- GLOBAL_STEP: 23225 | > loss: -0.45011 (-0.46815) | > log_mle: -0.56874 (-0.59165) | > loss_dur: 0.11863 (0.12350) | > amp_scaler: 2048.00000 (4372.75676) | > grad_norm: 8.84103 (8.71110) | > current_lr: 0.00042 | > step_time: 0.56850 (0.43964) | > loader_time: 0.00320 (0.00298)  --> STEP: 210/639 -- GLOBAL_STEP: 23250 | > loss: -0.38063 (-0.46619) | > log_mle: -0.48962 (-0.58934) | > loss_dur: 0.10898 (0.12316) | > amp_scaler: 2048.00000 (4096.00000) | > grad_norm: 1.01273 (8.42392) | > current_lr: 0.00041 | > step_time: 0.50360 (0.45496) | > loader_time: 0.00300 (0.00301)  --> STEP: 235/639 -- GLOBAL_STEP: 23275 | > loss: -0.48017 (-0.46754) | > log_mle: -0.58920 (-0.59018) | > loss_dur: 0.10903 (0.12264) | > amp_scaler: 2048.00000 (3878.12766) | > grad_norm: 6.57180 (8.61529) | > current_lr: 0.00041 | > step_time: 0.56310 (0.47102) | > loader_time: 0.00350 (0.00305)  --> STEP: 260/639 -- GLOBAL_STEP: 23300 | > loss: -0.45909 (-0.46770) | > log_mle: -0.58546 (-0.59022) | > loss_dur: 0.12637 (0.12252) | > amp_scaler: 2048.00000 (3702.15385) | > grad_norm: 9.46804 (8.89105) | > current_lr: 0.00041 | > step_time: 0.58830 (0.48542) | > loader_time: 0.00330 (0.00309)  --> STEP: 285/639 -- GLOBAL_STEP: 23325 | > loss: -0.40219 (-0.46723) | > log_mle: -0.53051 (-0.58977) | > loss_dur: 0.12832 (0.12254) | > amp_scaler: 2048.00000 (3557.05263) | > grad_norm: 3.35970 (8.89901) | > current_lr: 0.00041 | > step_time: 0.61190 (0.50171) | > loader_time: 0.00340 (0.00312)  --> STEP: 310/639 -- GLOBAL_STEP: 23350 | > loss: -0.45522 (-0.46620) | > log_mle: -0.57599 (-0.58881) | > loss_dur: 0.12077 (0.12261) | > amp_scaler: 2048.00000 (3435.35484) | > grad_norm: 10.37054 (8.97624) | > current_lr: 0.00041 | > step_time: 0.58690 (0.51604) | > loader_time: 0.00330 (0.00315)  --> STEP: 335/639 -- GLOBAL_STEP: 23375 | > loss: -0.46656 (-0.46446) | > log_mle: -0.59877 (-0.58744) | > loss_dur: 0.13221 (0.12298) | > amp_scaler: 2048.00000 (3331.82090) | > grad_norm: 11.47915 (8.87972) | > current_lr: 0.00041 | > step_time: 0.68830 (0.53198) | > loader_time: 0.00350 (0.00318)  --> STEP: 360/639 -- GLOBAL_STEP: 23400 | > loss: -0.42561 (-0.46346) | > log_mle: -0.53693 (-0.58608) | > loss_dur: 0.11131 (0.12263) | > amp_scaler: 2048.00000 (3242.66667) | > grad_norm: 4.69499 (8.82117) | > current_lr: 0.00041 | > step_time: 0.70250 (0.54607) | > loader_time: 0.00360 (0.00321)  --> STEP: 385/639 -- GLOBAL_STEP: 23425 | > loss: -0.45537 (-0.46247) | > log_mle: -0.57530 (-0.58487) | > loss_dur: 0.11993 (0.12240) | > amp_scaler: 2048.00000 (3165.09091) | > grad_norm: 5.46129 (8.69221) | > current_lr: 0.00041 | > step_time: 0.70900 (0.55955) | > loader_time: 0.00360 (0.00324)  --> STEP: 410/639 -- GLOBAL_STEP: 23450 | > loss: -0.40931 (-0.46120) | > log_mle: -0.54627 (-0.58369) | > loss_dur: 0.13696 (0.12249) | > amp_scaler: 2048.00000 (3096.97561) | > grad_norm: 5.61229 (8.65157) | > current_lr: 0.00041 | > step_time: 0.80510 (0.57164) | > loader_time: 0.00380 (0.00327)  --> STEP: 435/639 -- GLOBAL_STEP: 23475 | > loss: -0.44470 (-0.45974) | > log_mle: -0.55522 (-0.58203) | > loss_dur: 0.11052 (0.12230) | > amp_scaler: 2048.00000 (3036.68966) | > grad_norm: 7.61712 (8.59634) | > current_lr: 0.00041 | > step_time: 0.77870 (0.58359) | > loader_time: 0.00360 (0.00329)  --> STEP: 460/639 -- GLOBAL_STEP: 23500 | > loss: -0.39956 (-0.45825) | > log_mle: -0.52082 (-0.58041) | > loss_dur: 0.12126 (0.12217) | > amp_scaler: 2048.00000 (2982.95652) | > grad_norm: 2.30555 (8.47567) | > current_lr: 0.00041 | > step_time: 0.95200 (0.59438) | > loader_time: 0.00420 (0.00331)  --> STEP: 485/639 -- GLOBAL_STEP: 23525 | > loss: -0.43960 (-0.45710) | > log_mle: -0.55655 (-0.57938) | > loss_dur: 0.11695 (0.12228) | > amp_scaler: 2048.00000 (2934.76289) | > grad_norm: 5.41883 (8.39255) | > current_lr: 0.00041 | > step_time: 0.86690 (0.60687) | > loader_time: 0.00390 (0.00334)  --> STEP: 510/639 -- GLOBAL_STEP: 23550 | > loss: -0.43701 (-0.45636) | > log_mle: -0.56157 (-0.57847) | > loss_dur: 0.12456 (0.12211) | > amp_scaler: 2048.00000 (2891.29412) | > grad_norm: 5.90142 (8.29550) | > current_lr: 0.00041 | > step_time: 0.83670 (0.61974) | > loader_time: 0.00420 (0.00337)  --> STEP: 535/639 -- GLOBAL_STEP: 23575 | > loss: -0.42707 (-0.45539) | > log_mle: -0.55488 (-0.57748) | > loss_dur: 0.12781 (0.12208) | > amp_scaler: 2048.00000 (2851.88785) | > grad_norm: 5.66050 (8.20701) | > current_lr: 0.00041 | > step_time: 0.89840 (0.63303) | > loader_time: 0.00400 (0.00340)  --> STEP: 560/639 -- GLOBAL_STEP: 23600 | > loss: -0.43379 (-0.45445) | > log_mle: -0.56034 (-0.57656) | > loss_dur: 0.12654 (0.12211) | > amp_scaler: 2048.00000 (2816.00000) | > grad_norm: 7.60924 (8.13697) | > current_lr: 0.00041 | > step_time: 0.89620 (0.64801) | > loader_time: 0.00420 (0.00343)  --> STEP: 585/639 -- GLOBAL_STEP: 23625 | > loss: -0.44991 (-0.45374) | > log_mle: -0.56432 (-0.57574) | > loss_dur: 0.11441 (0.12200) | > amp_scaler: 2048.00000 (2783.17949) | > grad_norm: 7.51867 (8.08559) | > current_lr: 0.00041 | > step_time: 1.06830 (0.66308) | > loader_time: 0.00450 (0.00346)  --> STEP: 610/639 -- GLOBAL_STEP: 23650 | > loss: -0.41116 (-0.45297) | > log_mle: -0.53257 (-0.57499) | > loss_dur: 0.12141 (0.12202) | > amp_scaler: 2048.00000 (2753.04918) | > grad_norm: 4.59859 (8.01817) | > current_lr: 0.00041 | > step_time: 1.05000 (0.67936) | > loader_time: 0.00450 (0.00350)  --> STEP: 635/639 -- GLOBAL_STEP: 23675 | > loss: -0.43514 (-0.45246) | > log_mle: -0.55888 (-0.57460) | > loss_dur: 0.12374 (0.12214) | > amp_scaler: 2048.00000 (2725.29134) | > grad_norm: 6.12741 (8.00187) | > current_lr: 0.00041 | > step_time: 1.13800 (0.69591) | > loader_time: 0.00420 (0.00354) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.01760 (-0.01082) | > avg_loss: -0.43813 (+0.00252) | > avg_log_mle: -0.56639 (+0.00377) | > avg_loss_dur: 0.12826 (-0.00126)  > EPOCH: 37/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 19:30:15)   --> STEP: 20/639 -- GLOBAL_STEP: 23700 | > loss: -0.57995 (-0.49698) | > log_mle: -0.74337 (-0.62691) | > loss_dur: 0.16342 (0.12993) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 19.12237 (12.69186) | > current_lr: 0.00041 | > step_time: 0.33190 (0.31093) | > loader_time: 0.00300 (0.00297)  --> STEP: 45/639 -- GLOBAL_STEP: 23725 | > loss: -0.41804 (-0.47457) | > log_mle: -0.53701 (-0.60167) | > loss_dur: 0.11897 (0.12710) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.75406 (10.11568) | > current_lr: 0.00041 | > step_time: 0.41880 (0.34203) | > loader_time: 0.00290 (0.00291)  --> STEP: 70/639 -- GLOBAL_STEP: 23750 | > loss: -0.47770 (-0.46608) | > log_mle: -0.60699 (-0.59295) | > loss_dur: 0.12929 (0.12687) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.45084 (9.40158) | > current_lr: 0.00041 | > step_time: 0.39850 (0.36717) | > loader_time: 0.00310 (0.00291)  --> STEP: 95/639 -- GLOBAL_STEP: 23775 | > loss: -0.53309 (-0.46534) | > log_mle: -0.65646 (-0.58997) | > loss_dur: 0.12337 (0.12463) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.90976 (9.28129) | > current_lr: 0.00041 | > step_time: 0.40460 (0.38063) | > loader_time: 0.00320 (0.00292)  --> STEP: 120/639 -- GLOBAL_STEP: 23800 | > loss: -0.50646 (-0.46512) | > log_mle: -0.64744 (-0.59012) | > loss_dur: 0.14098 (0.12500) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 19.04044 (9.23099) | > current_lr: 0.00041 | > step_time: 0.62870 (0.39903) | > loader_time: 0.00310 (0.00294)  --> STEP: 145/639 -- GLOBAL_STEP: 23825 | > loss: -0.47107 (-0.46683) | > log_mle: -0.60474 (-0.59149) | > loss_dur: 0.13367 (0.12465) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.68036 (9.17384) | > current_lr: 0.00041 | > step_time: 0.47020 (0.41391) | > loader_time: 0.00300 (0.00296)  --> STEP: 170/639 -- GLOBAL_STEP: 23850 | > loss: -0.47831 (-0.46701) | > log_mle: -0.59718 (-0.59143) | > loss_dur: 0.11887 (0.12442) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.19819 (9.19606) | > current_lr: 0.00041 | > step_time: 0.53910 (0.42812) | > loader_time: 0.00330 (0.00299)  --> STEP: 195/639 -- GLOBAL_STEP: 23875 | > loss: -0.42529 (-0.46719) | > log_mle: -0.53347 (-0.59095) | > loss_dur: 0.10818 (0.12377) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.35058 (9.12878) | > current_lr: 0.00041 | > step_time: 0.49360 (0.44437) | > loader_time: 0.00320 (0.00302)  --> STEP: 220/639 -- GLOBAL_STEP: 23900 | > loss: -0.48058 (-0.46598) | > log_mle: -0.59020 (-0.58945) | > loss_dur: 0.10962 (0.12347) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.71560 (9.10955) | > current_lr: 0.00041 | > step_time: 0.65130 (0.46176) | > loader_time: 0.00330 (0.00305)  --> STEP: 245/639 -- GLOBAL_STEP: 23925 | > loss: -0.46730 (-0.46819) | > log_mle: -0.57918 (-0.59103) | > loss_dur: 0.11189 (0.12284) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.28475 (9.23544) | > current_lr: 0.00041 | > step_time: 0.52330 (0.47649) | > loader_time: 0.00320 (0.00308)  --> STEP: 270/639 -- GLOBAL_STEP: 23950 | > loss: -0.51613 (-0.46872) | > log_mle: -0.64490 (-0.59168) | > loss_dur: 0.12878 (0.12296) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.50197 (9.35442) | > current_lr: 0.00041 | > step_time: 0.79500 (0.49366) | > loader_time: 0.00380 (0.00311)  --> STEP: 295/639 -- GLOBAL_STEP: 23975 | > loss: -0.42975 (-0.46729) | > log_mle: -0.53484 (-0.59028) | > loss_dur: 0.10509 (0.12299) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.58749 (9.18019) | > current_lr: 0.00041 | > step_time: 0.62990 (0.50800) | > loader_time: 0.00350 (0.00314)  --> STEP: 320/639 -- GLOBAL_STEP: 24000 | > loss: -0.44711 (-0.46622) | > log_mle: -0.58366 (-0.58938) | > loss_dur: 0.13655 (0.12317) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.41441 (9.14175) | > current_lr: 0.00041 | > step_time: 0.77420 (0.52426) | > loader_time: 0.00360 (0.00317)  --> STEP: 345/639 -- GLOBAL_STEP: 24025 | > loss: -0.46376 (-0.46495) | > log_mle: -0.57410 (-0.58822) | > loss_dur: 0.11034 (0.12327) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.78709 (9.16351) | > current_lr: 0.00041 | > step_time: 0.64980 (0.53860) | > loader_time: 0.00330 (0.00321)  --> STEP: 370/639 -- GLOBAL_STEP: 24050 | > loss: -0.46856 (-0.46382) | > log_mle: -0.58978 (-0.58668) | > loss_dur: 0.12123 (0.12286) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.18459 (9.01879) | > current_lr: 0.00041 | > step_time: 0.89310 (0.55060) | > loader_time: 0.00390 (0.00324)  --> STEP: 395/639 -- GLOBAL_STEP: 24075 | > loss: -0.43029 (-0.46249) | > log_mle: -0.55392 (-0.58523) | > loss_dur: 0.12362 (0.12275) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.96046 (8.91964) | > current_lr: 0.00041 | > step_time: 0.70990 (0.56310) | > loader_time: 0.00310 (0.00326)  --> STEP: 420/639 -- GLOBAL_STEP: 24100 | > loss: -0.43415 (-0.46104) | > log_mle: -0.55112 (-0.58382) | > loss_dur: 0.11697 (0.12278) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.87529 (8.81532) | > current_lr: 0.00041 | > step_time: 0.69860 (0.57502) | > loader_time: 0.00350 (0.00328)  --> STEP: 445/639 -- GLOBAL_STEP: 24125 | > loss: -0.44529 (-0.45999) | > log_mle: -0.55744 (-0.58265) | > loss_dur: 0.11214 (0.12266) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.83734 (8.74037) | > current_lr: 0.00041 | > step_time: 0.73680 (0.58738) | > loader_time: 0.00360 (0.00331)  --> STEP: 470/639 -- GLOBAL_STEP: 24150 | > loss: -0.44822 (-0.45866) | > log_mle: -0.56443 (-0.58117) | > loss_dur: 0.11621 (0.12250) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.45121 (8.63808) | > current_lr: 0.00041 | > step_time: 0.77190 (0.59931) | > loader_time: 0.00370 (0.00333)  --> STEP: 495/639 -- GLOBAL_STEP: 24175 | > loss: -0.39828 (-0.45762) | > log_mle: -0.50445 (-0.58006) | > loss_dur: 0.10618 (0.12245) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.11084 (8.56342) | > current_lr: 0.00041 | > step_time: 0.77530 (0.61211) | > loader_time: 0.00360 (0.00336)  --> STEP: 520/639 -- GLOBAL_STEP: 24200 | > loss: -0.45196 (-0.45701) | > log_mle: -0.56249 (-0.57928) | > loss_dur: 0.11053 (0.12227) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.19045 (8.46129) | > current_lr: 0.00041 | > step_time: 0.82980 (0.62500) | > loader_time: 0.00400 (0.00338)  --> STEP: 545/639 -- GLOBAL_STEP: 24225 | > loss: -0.46312 (-0.45617) | > log_mle: -0.59378 (-0.57843) | > loss_dur: 0.13066 (0.12226) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.35061 (8.39015) | > current_lr: 0.00041 | > step_time: 0.95040 (0.64017) | > loader_time: 0.00400 (0.00341)  --> STEP: 570/639 -- GLOBAL_STEP: 24250 | > loss: -0.43528 (-0.45514) | > log_mle: -0.54210 (-0.57731) | > loss_dur: 0.10681 (0.12218) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.96102 (8.29355) | > current_lr: 0.00041 | > step_time: 0.85850 (0.65422) | > loader_time: 0.00380 (0.00344)  --> STEP: 595/639 -- GLOBAL_STEP: 24275 | > loss: -0.42123 (-0.45453) | > log_mle: -0.53499 (-0.57668) | > loss_dur: 0.11376 (0.12215) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.52076 (8.23439) | > current_lr: 0.00041 | > step_time: 0.92450 (0.67024) | > loader_time: 0.00420 (0.00348)  --> STEP: 620/639 -- GLOBAL_STEP: 24300 | > loss: -0.42675 (-0.45374) | > log_mle: -0.54413 (-0.57587) | > loss_dur: 0.11738 (0.12213) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.01719 (8.20630) | > current_lr: 0.00041 | > step_time: 1.20700 (0.68659) | > loader_time: 0.00460 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03281 (+0.01522) | > avg_loss: -0.44073 (-0.00259) | > avg_log_mle: -0.56616 (+0.00023) | > avg_loss_dur: 0.12544 (-0.00282) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_24320.pth.tar  > EPOCH: 38/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 19:38:02)   --> STEP: 5/639 -- GLOBAL_STEP: 24325 | > loss: -0.51351 (-0.50322) | > log_mle: -0.66315 (-0.63917) | > loss_dur: 0.14964 (0.13595) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.82736 (10.74436) | > current_lr: 0.00041 | > step_time: 0.30980 (0.27763) | > loader_time: 0.00090 (0.00201)  --> STEP: 30/639 -- GLOBAL_STEP: 24350 | > loss: -0.49590 (-0.48040) | > log_mle: -0.61298 (-0.60801) | > loss_dur: 0.11707 (0.12762) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.59057 (10.60964) | > current_lr: 0.00041 | > step_time: 0.33940 (0.32970) | > loader_time: 0.00250 (0.00255)  --> STEP: 55/639 -- GLOBAL_STEP: 24375 | > loss: -0.43653 (-0.47319) | > log_mle: -0.53906 (-0.59913) | > loss_dur: 0.10253 (0.12594) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.76724 (9.33402) | > current_lr: 0.00041 | > step_time: 0.41550 (0.35054) | > loader_time: 0.00270 (0.00265)  --> STEP: 80/639 -- GLOBAL_STEP: 24400 | > loss: -0.45020 (-0.46828) | > log_mle: -0.58732 (-0.59295) | > loss_dur: 0.13712 (0.12467) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.35672 (8.70833) | > current_lr: 0.00041 | > step_time: 0.38490 (0.36905) | > loader_time: 0.00290 (0.00272)  --> STEP: 105/639 -- GLOBAL_STEP: 24425 | > loss: -0.41536 (-0.46847) | > log_mle: -0.54624 (-0.59245) | > loss_dur: 0.13088 (0.12398) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.68256 (8.81406) | > current_lr: 0.00040 | > step_time: 0.41390 (0.38911) | > loader_time: 0.00270 (0.00278)  --> STEP: 130/639 -- GLOBAL_STEP: 24450 | > loss: -0.45939 (-0.46812) | > log_mle: -0.58658 (-0.59233) | > loss_dur: 0.12719 (0.12421) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.56071 (8.91368) | > current_lr: 0.00040 | > step_time: 0.48370 (0.40750) | > loader_time: 0.00270 (0.00280)  --> STEP: 155/639 -- GLOBAL_STEP: 24475 | > loss: -0.45079 (-0.46907) | > log_mle: -0.57667 (-0.59279) | > loss_dur: 0.12588 (0.12372) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.56209 (8.94128) | > current_lr: 0.00040 | > step_time: 0.55020 (0.42369) | > loader_time: 0.00320 (0.00283)  --> STEP: 180/639 -- GLOBAL_STEP: 24500 | > loss: -0.50795 (-0.47173) | > log_mle: -0.63667 (-0.59528) | > loss_dur: 0.12872 (0.12355) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.44200 (9.28384) | > current_lr: 0.00040 | > step_time: 0.55140 (0.43909) | > loader_time: 0.00340 (0.00286)  --> STEP: 205/639 -- GLOBAL_STEP: 24525 | > loss: -0.44452 (-0.46967) | > log_mle: -0.55774 (-0.59252) | > loss_dur: 0.11322 (0.12285) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.72344 (9.00742) | > current_lr: 0.00040 | > step_time: 0.55700 (0.45258) | > loader_time: 0.00330 (0.00291)  --> STEP: 230/639 -- GLOBAL_STEP: 24550 | > loss: -0.47752 (-0.47076) | > log_mle: -0.58965 (-0.59315) | > loss_dur: 0.11213 (0.12239) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.26272 (9.07767) | > current_lr: 0.00040 | > step_time: 0.55860 (0.46922) | > loader_time: 0.00330 (0.00296)  --> STEP: 255/639 -- GLOBAL_STEP: 24575 | > loss: -0.44209 (-0.47099) | > log_mle: -0.56206 (-0.59284) | > loss_dur: 0.11997 (0.12185) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.80759 (9.05976) | > current_lr: 0.00040 | > step_time: 0.62540 (0.48273) | > loader_time: 0.00350 (0.00300)  --> STEP: 280/639 -- GLOBAL_STEP: 24600 | > loss: -0.45909 (-0.47150) | > log_mle: -0.58721 (-0.59353) | > loss_dur: 0.12812 (0.12203) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.77885 (9.12703) | > current_lr: 0.00040 | > step_time: 0.60530 (0.49915) | > loader_time: 0.00350 (0.00304)  --> STEP: 305/639 -- GLOBAL_STEP: 24625 | > loss: -0.43456 (-0.47020) | > log_mle: -0.56536 (-0.59232) | > loss_dur: 0.13080 (0.12212) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.57220 (9.07432) | > current_lr: 0.00040 | > step_time: 0.67610 (0.51424) | > loader_time: 0.00340 (0.00308)  --> STEP: 330/639 -- GLOBAL_STEP: 24650 | > loss: -0.45773 (-0.46831) | > log_mle: -0.59077 (-0.59087) | > loss_dur: 0.13304 (0.12256) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.59350 (9.02077) | > current_lr: 0.00040 | > step_time: 0.69740 (0.52859) | > loader_time: 0.00380 (0.00312)  --> STEP: 355/639 -- GLOBAL_STEP: 24675 | > loss: -0.40558 (-0.46723) | > log_mle: -0.51908 (-0.58948) | > loss_dur: 0.11349 (0.12225) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.34808 (9.01159) | > current_lr: 0.00040 | > step_time: 0.61700 (0.54154) | > loader_time: 0.00330 (0.00315)  --> STEP: 380/639 -- GLOBAL_STEP: 24700 | > loss: -0.44966 (-0.46559) | > log_mle: -0.55394 (-0.58769) | > loss_dur: 0.10428 (0.12209) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.16301 (8.84035) | > current_lr: 0.00040 | > step_time: 0.63490 (0.55498) | > loader_time: 0.00360 (0.00318)  --> STEP: 405/639 -- GLOBAL_STEP: 24725 | > loss: -0.44245 (-0.46457) | > log_mle: -0.56130 (-0.58679) | > loss_dur: 0.11885 (0.12222) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.03564 (8.82514) | > current_lr: 0.00040 | > step_time: 0.66890 (0.56707) | > loader_time: 0.00350 (0.00320)  --> STEP: 430/639 -- GLOBAL_STEP: 24750 | > loss: -0.40926 (-0.46309) | > log_mle: -0.52086 (-0.58529) | > loss_dur: 0.11160 (0.12220) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.69572 (8.75946) | > current_lr: 0.00040 | > step_time: 0.75510 (0.58028) | > loader_time: 0.00380 (0.00323)  --> STEP: 455/639 -- GLOBAL_STEP: 24775 | > loss: -0.42593 (-0.46173) | > log_mle: -0.54678 (-0.58371) | > loss_dur: 0.12084 (0.12198) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.48410 (8.65425) | > current_lr: 0.00040 | > step_time: 0.81110 (0.59103) | > loader_time: 0.00360 (0.00325)  --> STEP: 480/639 -- GLOBAL_STEP: 24800 | > loss: -0.38432 (-0.46044) | > log_mle: -0.50301 (-0.58249) | > loss_dur: 0.11869 (0.12205) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.71189 (8.58151) | > current_lr: 0.00040 | > step_time: 0.78920 (0.60363) | > loader_time: 0.00360 (0.00328)  --> STEP: 505/639 -- GLOBAL_STEP: 24825 | > loss: -0.45906 (-0.45954) | > log_mle: -0.57329 (-0.58142) | > loss_dur: 0.11422 (0.12188) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.26799 (8.53078) | > current_lr: 0.00040 | > step_time: 0.90740 (0.61590) | > loader_time: 0.00410 (0.00331)  --> STEP: 530/639 -- GLOBAL_STEP: 24850 | > loss: -0.43133 (-0.45869) | > log_mle: -0.55958 (-0.58056) | > loss_dur: 0.12825 (0.12187) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.55549 (8.49966) | > current_lr: 0.00040 | > step_time: 1.03890 (0.62944) | > loader_time: 0.00410 (0.00334)  --> STEP: 555/639 -- GLOBAL_STEP: 24875 | > loss: -0.43037 (-0.45761) | > log_mle: -0.54588 (-0.57937) | > loss_dur: 0.11551 (0.12177) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.72706 (8.41954) | > current_lr: 0.00040 | > step_time: 0.86890 (0.64329) | > loader_time: 0.00390 (0.00337)  --> STEP: 580/639 -- GLOBAL_STEP: 24900 | > loss: -0.45515 (-0.45682) | > log_mle: -0.56511 (-0.57854) | > loss_dur: 0.10997 (0.12172) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.93943 (8.38707) | > current_lr: 0.00040 | > step_time: 1.01490 (0.65834) | > loader_time: 0.00430 (0.00341)  --> STEP: 605/639 -- GLOBAL_STEP: 24925 | > loss: -0.42020 (-0.45594) | > log_mle: -0.53073 (-0.57776) | > loss_dur: 0.11053 (0.12182) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.03440 (8.35043) | > current_lr: 0.00040 | > step_time: 1.00030 (0.67372) | > loader_time: 0.00420 (0.00345)  --> STEP: 630/639 -- GLOBAL_STEP: 24950 | > loss: -0.43318 (-0.45532) | > log_mle: -0.56354 (-0.57718) | > loss_dur: 0.13036 (0.12185) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.53665 (8.33700) | > current_lr: 0.00040 | > step_time: 1.34530 (0.69065) | > loader_time: 0.00470 (0.00349) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02024 (-0.01257) | > avg_loss: -0.44127 (-0.00054) | > avg_log_mle: -0.57055 (-0.00439) | > avg_loss_dur: 0.12928 (+0.00385) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_24960.pth.tar  > EPOCH: 39/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 19:45:48)   --> STEP: 15/639 -- GLOBAL_STEP: 24975 | > loss: -0.59830 (-0.50453) | > log_mle: -0.74575 (-0.63127) | > loss_dur: 0.14745 (0.12674) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 27.94180 (10.75283) | > current_lr: 0.00040 | > step_time: 0.32790 (0.29552) | > loader_time: 0.00220 (0.00248)  --> STEP: 40/639 -- GLOBAL_STEP: 25000 | > loss: -0.46583 (-0.48166) | > log_mle: -0.58508 (-0.60672) | > loss_dur: 0.11925 (0.12506) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.53307 (9.49199) | > current_lr: 0.00040 | > step_time: 0.35120 (0.33074) | > loader_time: 0.00270 (0.00257)  --> STEP: 65/639 -- GLOBAL_STEP: 25025 | > loss: -0.46885 (-0.47374) | > log_mle: -0.57501 (-0.59985) | > loss_dur: 0.10616 (0.12611) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.86726 (8.83587) | > current_lr: 0.00040 | > step_time: 0.38620 (0.36068) | > loader_time: 0.00280 (0.00270)  --> STEP: 90/639 -- GLOBAL_STEP: 25050 | > loss: -0.41545 (-0.46783) | > log_mle: -0.53288 (-0.59204) | > loss_dur: 0.11743 (0.12420) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.04973 (8.15985) | > current_lr: 0.00040 | > step_time: 0.51810 (0.37567) | > loader_time: 0.00310 (0.00279)  --> STEP: 115/639 -- GLOBAL_STEP: 25075 | > loss: -0.38241 (-0.46977) | > log_mle: -0.50344 (-0.59374) | > loss_dur: 0.12103 (0.12397) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.51858 (8.33323) | > current_lr: 0.00040 | > step_time: 0.38260 (0.39352) | > loader_time: 0.00300 (0.00284)  --> STEP: 140/639 -- GLOBAL_STEP: 25100 | > loss: -0.44983 (-0.47136) | > log_mle: -0.58846 (-0.59513) | > loss_dur: 0.13863 (0.12377) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.82604 (8.53147) | > current_lr: 0.00040 | > step_time: 0.54460 (0.40827) | > loader_time: 0.00320 (0.00288)  --> STEP: 165/639 -- GLOBAL_STEP: 25125 | > loss: -0.45773 (-0.47187) | > log_mle: -0.59601 (-0.59501) | > loss_dur: 0.13828 (0.12314) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.90563 (8.56258) | > current_lr: 0.00040 | > step_time: 0.49020 (0.42386) | > loader_time: 0.00310 (0.00291)  --> STEP: 190/639 -- GLOBAL_STEP: 25150 | > loss: -0.40039 (-0.47282) | > log_mle: -0.52348 (-0.59536) | > loss_dur: 0.12309 (0.12254) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.27771 (8.79755) | > current_lr: 0.00040 | > step_time: 0.53770 (0.44105) | > loader_time: 0.00310 (0.00296)  --> STEP: 215/639 -- GLOBAL_STEP: 25175 | > loss: -0.51814 (-0.47153) | > log_mle: -0.63279 (-0.59401) | > loss_dur: 0.11465 (0.12248) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.83089 (8.69859) | > current_lr: 0.00040 | > step_time: 0.55420 (0.45570) | > loader_time: 0.00320 (0.00300)  --> STEP: 240/639 -- GLOBAL_STEP: 25200 | > loss: -0.51822 (-0.47336) | > log_mle: -0.64176 (-0.59527) | > loss_dur: 0.12354 (0.12192) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.57805 (8.96778) | > current_lr: 0.00040 | > step_time: 0.57390 (0.47069) | > loader_time: 0.00360 (0.00304)  --> STEP: 265/639 -- GLOBAL_STEP: 25225 | > loss: -0.53485 (-0.47305) | > log_mle: -0.66621 (-0.59503) | > loss_dur: 0.13135 (0.12198) | > amp_scaler: 4096.00000 (2117.55472) | > grad_norm: 14.63704 (8.98816) | > current_lr: 0.00040 | > step_time: 0.62360 (0.48717) | > loader_time: 0.00340 (0.00307)  --> STEP: 290/639 -- GLOBAL_STEP: 25250 | > loss: -0.45206 (-0.47195) | > log_mle: -0.56638 (-0.59395) | > loss_dur: 0.11432 (0.12200) | > amp_scaler: 4096.00000 (2288.11034) | > grad_norm: 6.07317 (8.86192) | > current_lr: 0.00040 | > step_time: 0.62930 (0.50301) | > loader_time: 0.00340 (0.00310)  --> STEP: 315/639 -- GLOBAL_STEP: 25275 | > loss: -0.43043 (-0.47095) | > log_mle: -0.53990 (-0.59291) | > loss_dur: 0.10948 (0.12196) | > amp_scaler: 4096.00000 (2431.59365) | > grad_norm: 3.63909 (8.86520) | > current_lr: 0.00040 | > step_time: 0.64550 (0.51737) | > loader_time: 0.00340 (0.00313)  --> STEP: 340/639 -- GLOBAL_STEP: 25300 | > loss: -0.45143 (-0.46965) | > log_mle: -0.56363 (-0.59203) | > loss_dur: 0.11220 (0.12237) | > amp_scaler: 4096.00000 (2553.97647) | > grad_norm: 8.32466 (8.89973) | > current_lr: 0.00040 | > step_time: 0.59820 (0.53285) | > loader_time: 0.00360 (0.00317)  --> STEP: 365/639 -- GLOBAL_STEP: 25325 | > loss: -0.46162 (-0.46809) | > log_mle: -0.57669 (-0.59006) | > loss_dur: 0.11507 (0.12198) | > amp_scaler: 4096.00000 (2659.59452) | > grad_norm: 11.15833 (8.77532) | > current_lr: 0.00040 | > step_time: 0.88260 (0.54515) | > loader_time: 0.00390 (0.00319)  --> STEP: 390/639 -- GLOBAL_STEP: 25350 | > loss: -0.43792 (-0.46671) | > log_mle: -0.57084 (-0.58848) | > loss_dur: 0.13292 (0.12178) | > amp_scaler: 4096.00000 (2751.67179) | > grad_norm: 13.24630 (8.65952) | > current_lr: 0.00040 | > step_time: 0.71240 (0.55816) | > loader_time: 0.00380 (0.00322)  --> STEP: 415/639 -- GLOBAL_STEP: 25375 | > loss: -0.41760 (-0.46528) | > log_mle: -0.54471 (-0.58720) | > loss_dur: 0.12711 (0.12193) | > amp_scaler: 4096.00000 (2832.65542) | > grad_norm: 5.68840 (8.59230) | > current_lr: 0.00040 | > step_time: 0.70770 (0.57054) | > loader_time: 0.00340 (0.00325)  --> STEP: 440/639 -- GLOBAL_STEP: 25400 | > loss: -0.40374 (-0.46406) | > log_mle: -0.52909 (-0.58591) | > loss_dur: 0.12535 (0.12185) | > amp_scaler: 4096.00000 (2904.43636) | > grad_norm: 5.16478 (8.58838) | > current_lr: 0.00040 | > step_time: 0.74760 (0.58292) | > loader_time: 0.00370 (0.00328)  --> STEP: 465/639 -- GLOBAL_STEP: 25425 | > loss: -0.41730 (-0.46250) | > log_mle: -0.53475 (-0.58421) | > loss_dur: 0.11746 (0.12171) | > amp_scaler: 4096.00000 (2968.49892) | > grad_norm: 3.53013 (8.42454) | > current_lr: 0.00040 | > step_time: 0.77190 (0.59415) | > loader_time: 0.00370 (0.00330)  --> STEP: 490/639 -- GLOBAL_STEP: 25450 | > loss: -0.41681 (-0.46128) | > log_mle: -0.51939 (-0.58301) | > loss_dur: 0.10258 (0.12173) | > amp_scaler: 4096.00000 (3026.02449) | > grad_norm: 2.32556 (8.38436) | > current_lr: 0.00040 | > step_time: 0.77180 (0.60706) | > loader_time: 0.00370 (0.00333)  --> STEP: 515/639 -- GLOBAL_STEP: 25475 | > loss: -0.43730 (-0.46064) | > log_mle: -0.55176 (-0.58223) | > loss_dur: 0.11446 (0.12158) | > amp_scaler: 4096.00000 (3077.96505) | > grad_norm: 4.07613 (8.31568) | > current_lr: 0.00040 | > step_time: 0.93190 (0.61998) | > loader_time: 0.00390 (0.00335)  --> STEP: 540/639 -- GLOBAL_STEP: 25500 | > loss: -0.45091 (-0.45960) | > log_mle: -0.57066 (-0.58115) | > loss_dur: 0.11975 (0.12155) | > amp_scaler: 4096.00000 (3125.09630) | > grad_norm: 4.86944 (8.24310) | > current_lr: 0.00040 | > step_time: 0.90540 (0.63489) | > loader_time: 0.00410 (0.00338)  --> STEP: 565/639 -- GLOBAL_STEP: 25525 | > loss: -0.40849 (-0.45840) | > log_mle: -0.52551 (-0.57996) | > loss_dur: 0.11702 (0.12156) | > amp_scaler: 4096.00000 (3168.05664) | > grad_norm: 4.37889 (8.15637) | > current_lr: 0.00040 | > step_time: 1.07400 (0.64895) | > loader_time: 0.00420 (0.00342)  --> STEP: 590/639 -- GLOBAL_STEP: 25550 | > loss: -0.43895 (-0.45782) | > log_mle: -0.56025 (-0.57935) | > loss_dur: 0.12129 (0.12154) | > amp_scaler: 4096.00000 (3207.37627) | > grad_norm: 7.24903 (8.11311) | > current_lr: 0.00040 | > step_time: 0.93880 (0.66546) | > loader_time: 0.00440 (0.00345)  --> STEP: 615/639 -- GLOBAL_STEP: 25575 | > loss: -0.44240 (-0.45694) | > log_mle: -0.56313 (-0.57846) | > loss_dur: 0.12073 (0.12151) | > amp_scaler: 4096.00000 (3243.49919) | > grad_norm: 4.55554 (8.02783) | > current_lr: 0.00040 | > step_time: 1.01990 (0.68126) | > loader_time: 0.00440 (0.00349) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03496 (+0.01471) | > avg_loss: -0.44000 (+0.00127) | > avg_log_mle: -0.56929 (+0.00126) | > avg_loss_dur: 0.12930 (+0.00001)  > EPOCH: 40/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 19:53:29)   --> STEP: 0/639 -- GLOBAL_STEP: 25600 | > loss: -0.52071 (-0.52071) | > log_mle: -0.68398 (-0.68398) | > loss_dur: 0.16326 (0.16326) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 19.45800 (19.45800) | > current_lr: 0.00040 | > step_time: 0.67060 (0.67057) | > loader_time: 0.51310 (0.51306)  --> STEP: 25/639 -- GLOBAL_STEP: 25625 | > loss: -0.47379 (-0.49947) | > log_mle: -0.61643 (-0.62613) | > loss_dur: 0.14264 (0.12666) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.53538 (11.05345) | > current_lr: 0.00040 | > step_time: 0.38670 (0.32956) | > loader_time: 0.00290 (0.00252)  --> STEP: 50/639 -- GLOBAL_STEP: 25650 | > loss: -0.44136 (-0.48469) | > log_mle: -0.55438 (-0.60890) | > loss_dur: 0.11301 (0.12421) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.37036 (9.41168) | > current_lr: 0.00040 | > step_time: 0.31290 (0.34791) | > loader_time: 0.00280 (0.00267)  --> STEP: 75/639 -- GLOBAL_STEP: 25675 | > loss: -0.47096 (-0.47603) | > log_mle: -0.58499 (-0.60057) | > loss_dur: 0.11403 (0.12454) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.26843 (8.83983) | > current_lr: 0.00039 | > step_time: 0.41300 (0.37186) | > loader_time: 0.00310 (0.00277)  --> STEP: 100/639 -- GLOBAL_STEP: 25700 | > loss: -0.48126 (-0.47574) | > log_mle: -0.60055 (-0.59848) | > loss_dur: 0.11929 (0.12274) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.91334 (8.50456) | > current_lr: 0.00039 | > step_time: 0.50960 (0.38836) | > loader_time: 0.00320 (0.00284)  --> STEP: 125/639 -- GLOBAL_STEP: 25725 | > loss: -0.48270 (-0.47481) | > log_mle: -0.59683 (-0.59778) | > loss_dur: 0.11413 (0.12298) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.78643 (8.66529) | > current_lr: 0.00039 | > step_time: 0.43950 (0.40197) | > loader_time: 0.00310 (0.00288)  --> STEP: 150/639 -- GLOBAL_STEP: 25750 | > loss: -0.43907 (-0.47487) | > log_mle: -0.56260 (-0.59757) | > loss_dur: 0.12353 (0.12269) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.53567 (8.87624) | > current_lr: 0.00039 | > step_time: 0.42810 (0.41721) | > loader_time: 0.00320 (0.00293)  --> STEP: 175/639 -- GLOBAL_STEP: 25775 | > loss: -0.47586 (-0.47593) | > log_mle: -0.57957 (-0.59834) | > loss_dur: 0.10371 (0.12241) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.25961 (8.95155) | > current_lr: 0.00039 | > step_time: 0.63160 (0.43583) | > loader_time: 0.00330 (0.00298)  --> STEP: 200/639 -- GLOBAL_STEP: 25800 | > loss: -0.44513 (-0.47471) | > log_mle: -0.56637 (-0.59647) | > loss_dur: 0.12123 (0.12176) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.60578 (8.81299) | > current_lr: 0.00039 | > step_time: 0.52690 (0.45039) | > loader_time: 0.00330 (0.00301)  --> STEP: 225/639 -- GLOBAL_STEP: 25825 | > loss: -0.43568 (-0.47341) | > log_mle: -0.54017 (-0.59469) | > loss_dur: 0.10448 (0.12128) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.39661 (8.87451) | > current_lr: 0.00039 | > step_time: 0.54890 (0.46713) | > loader_time: 0.00320 (0.00306)  --> STEP: 250/639 -- GLOBAL_STEP: 25850 | > loss: -0.44484 (-0.47484) | > log_mle: -0.54536 (-0.59598) | > loss_dur: 0.10052 (0.12114) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.83508 (9.02511) | > current_lr: 0.00039 | > step_time: 0.58800 (0.48224) | > loader_time: 0.00350 (0.00310)  --> STEP: 275/639 -- GLOBAL_STEP: 25875 | > loss: -0.46260 (-0.47511) | > log_mle: -0.58979 (-0.59658) | > loss_dur: 0.12719 (0.12147) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.38793 (9.20602) | > current_lr: 0.00039 | > step_time: 0.86250 (0.50056) | > loader_time: 0.00380 (0.00315)  --> STEP: 300/639 -- GLOBAL_STEP: 25900 | > loss: -0.46568 (-0.47359) | > log_mle: -0.58249 (-0.59508) | > loss_dur: 0.11681 (0.12149) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.53906 (9.09985) | > current_lr: 0.00039 | > step_time: 0.70860 (0.51476) | > loader_time: 0.00380 (0.00318)  --> STEP: 325/639 -- GLOBAL_STEP: 25925 | > loss: -0.45749 (-0.47193) | > log_mle: -0.58418 (-0.59390) | > loss_dur: 0.12669 (0.12197) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.21813 (9.03078) | > current_lr: 0.00039 | > step_time: 0.62780 (0.53014) | > loader_time: 0.00330 (0.00322)  --> STEP: 350/639 -- GLOBAL_STEP: 25950 | > loss: -0.48144 (-0.47057) | > log_mle: -0.61241 (-0.59248) | > loss_dur: 0.13097 (0.12191) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.69327 (8.98916) | > current_lr: 0.00039 | > step_time: 0.89840 (0.54445) | > loader_time: 0.00390 (0.00325)  --> STEP: 375/639 -- GLOBAL_STEP: 25975 | > loss: -0.42571 (-0.46885) | > log_mle: -0.54481 (-0.59045) | > loss_dur: 0.11910 (0.12160) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.08049 (8.80398) | > current_lr: 0.00039 | > step_time: 0.68690 (0.55764) | > loader_time: 0.00310 (0.00328)  --> STEP: 400/639 -- GLOBAL_STEP: 26000 | > loss: -0.46183 (-0.46786) | > log_mle: -0.57542 (-0.58940) | > loss_dur: 0.11360 (0.12154) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.71163 (8.77353) | > current_lr: 0.00039 | > step_time: 0.72330 (0.56925) | > loader_time: 0.00390 (0.00331)  --> STEP: 425/639 -- GLOBAL_STEP: 26025 | > loss: -0.49347 (-0.46618) | > log_mle: -0.61976 (-0.58772) | > loss_dur: 0.12629 (0.12154) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 17.28470 (8.68527) | > current_lr: 0.00039 | > step_time: 0.93170 (0.58147) | > loader_time: 0.00380 (0.00334)  --> STEP: 450/639 -- GLOBAL_STEP: 26050 | > loss: -0.42420 (-0.46491) | > log_mle: -0.53278 (-0.58622) | > loss_dur: 0.10859 (0.12131) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.71820 (8.55302) | > current_lr: 0.00039 | > step_time: 0.79250 (0.59393) | > loader_time: 0.00400 (0.00337)  --> STEP: 475/639 -- GLOBAL_STEP: 26075 | > loss: -0.44253 (-0.46345) | > log_mle: -0.54482 (-0.58482) | > loss_dur: 0.10229 (0.12138) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.95378 (8.43972) | > current_lr: 0.00039 | > step_time: 0.71050 (0.60584) | > loader_time: 0.00390 (0.00340)  --> STEP: 500/639 -- GLOBAL_STEP: 26100 | > loss: -0.48383 (-0.46253) | > log_mle: -0.60911 (-0.58388) | > loss_dur: 0.12528 (0.12135) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.42450 (8.37309) | > current_lr: 0.00039 | > step_time: 1.05650 (0.61916) | > loader_time: 0.00420 (0.00343)  --> STEP: 525/639 -- GLOBAL_STEP: 26125 | > loss: -0.45437 (-0.46163) | > log_mle: -0.58834 (-0.58295) | > loss_dur: 0.13397 (0.12132) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.67194 (8.31465) | > current_lr: 0.00039 | > step_time: 1.01540 (0.63257) | > loader_time: 0.00440 (0.00346)  --> STEP: 550/639 -- GLOBAL_STEP: 26150 | > loss: -0.40205 (-0.46057) | > log_mle: -0.52161 (-0.58179) | > loss_dur: 0.11956 (0.12123) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.96489 (8.24372) | > current_lr: 0.00039 | > step_time: 0.91730 (0.64699) | > loader_time: 0.00440 (0.00350)  --> STEP: 575/639 -- GLOBAL_STEP: 26175 | > loss: -0.42954 (-0.45954) | > log_mle: -0.54274 (-0.58071) | > loss_dur: 0.11320 (0.12118) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.26046 (8.14599) | > current_lr: 0.00039 | > step_time: 0.89520 (0.66099) | > loader_time: 0.00360 (0.00353)  --> STEP: 600/639 -- GLOBAL_STEP: 26200 | > loss: -0.41326 (-0.45878) | > log_mle: -0.53639 (-0.57997) | > loss_dur: 0.12312 (0.12119) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.92821 (8.12023) | > current_lr: 0.00039 | > step_time: 1.06690 (0.67666) | > loader_time: 0.00440 (0.00357)  --> STEP: 625/639 -- GLOBAL_STEP: 26225 | > loss: -0.44868 (-0.45795) | > log_mle: -0.57128 (-0.57910) | > loss_dur: 0.12259 (0.12115) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.87400 (8.10591) | > current_lr: 0.00039 | > step_time: 1.24980 (0.69377) | > loader_time: 0.00450 (0.00361) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03866 (+0.00370) | > avg_loss: -0.43880 (+0.00120) | > avg_log_mle: -0.56854 (+0.00075) | > avg_loss_dur: 0.12974 (+0.00044)  > EPOCH: 41/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 20:01:13)   --> STEP: 10/639 -- GLOBAL_STEP: 26250 | > loss: -0.49112 (-0.50013) | > log_mle: -0.62881 (-0.63022) | > loss_dur: 0.13768 (0.13009) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.61259 (10.85600) | > current_lr: 0.00039 | > step_time: 0.32860 (0.30848) | > loader_time: 0.00280 (0.00645)  --> STEP: 35/639 -- GLOBAL_STEP: 26275 | > loss: -0.49140 (-0.48482) | > log_mle: -0.59445 (-0.60965) | > loss_dur: 0.10305 (0.12483) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.01860 (9.92756) | > current_lr: 0.00039 | > step_time: 0.35460 (0.34601) | > loader_time: 0.00270 (0.00379)  --> STEP: 60/639 -- GLOBAL_STEP: 26300 | > loss: -0.42758 (-0.47786) | > log_mle: -0.54664 (-0.60272) | > loss_dur: 0.11907 (0.12485) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.05683 (9.04339) | > current_lr: 0.00039 | > step_time: 0.34640 (0.36623) | > loader_time: 0.00240 (0.00338)  --> STEP: 85/639 -- GLOBAL_STEP: 26325 | > loss: -0.42804 (-0.47216) | > log_mle: -0.54796 (-0.59516) | > loss_dur: 0.11991 (0.12300) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.85338 (8.39651) | > current_lr: 0.00039 | > step_time: 0.42500 (0.37820) | > loader_time: 0.00290 (0.00322)  --> STEP: 110/639 -- GLOBAL_STEP: 26350 | > loss: -0.45422 (-0.47527) | > log_mle: -0.57139 (-0.59839) | > loss_dur: 0.11717 (0.12312) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.85589 (8.73501) | > current_lr: 0.00039 | > step_time: 0.39980 (0.39661) | > loader_time: 0.00310 (0.00316)  --> STEP: 135/639 -- GLOBAL_STEP: 26375 | > loss: -0.41607 (-0.47466) | > log_mle: -0.54623 (-0.59701) | > loss_dur: 0.13016 (0.12234) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.54790 (8.63245) | > current_lr: 0.00039 | > step_time: 0.46900 (0.40674) | > loader_time: 0.00260 (0.00313)  --> STEP: 160/639 -- GLOBAL_STEP: 26400 | > loss: -0.47995 (-0.47415) | > log_mle: -0.59524 (-0.59585) | > loss_dur: 0.11530 (0.12170) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.22999 (8.55591) | > current_lr: 0.00039 | > step_time: 0.47170 (0.42070) | > loader_time: 0.00300 (0.00310)  --> STEP: 185/639 -- GLOBAL_STEP: 26425 | > loss: -0.46065 (-0.47691) | > log_mle: -0.57785 (-0.59877) | > loss_dur: 0.11720 (0.12186) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.21727 (8.77857) | > current_lr: 0.00039 | > step_time: 0.59430 (0.44046) | > loader_time: 0.00320 (0.00312)  --> STEP: 210/639 -- GLOBAL_STEP: 26450 | > loss: -0.38293 (-0.47454) | > log_mle: -0.49444 (-0.59627) | > loss_dur: 0.11151 (0.12173) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.43813 (8.66560) | > current_lr: 0.00039 | > step_time: 0.51330 (0.45651) | > loader_time: 0.00310 (0.00314)  --> STEP: 235/639 -- GLOBAL_STEP: 26475 | > loss: -0.47525 (-0.47572) | > log_mle: -0.59199 (-0.59703) | > loss_dur: 0.11674 (0.12132) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.36398 (8.73490) | > current_lr: 0.00039 | > step_time: 0.56650 (0.47347) | > loader_time: 0.00360 (0.00316)  --> STEP: 260/639 -- GLOBAL_STEP: 26500 | > loss: -0.47106 (-0.47577) | > log_mle: -0.58972 (-0.59709) | > loss_dur: 0.11865 (0.12132) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.55829 (8.75685) | > current_lr: 0.00039 | > step_time: 0.59850 (0.48939) | > loader_time: 0.00350 (0.00319)  --> STEP: 285/639 -- GLOBAL_STEP: 26525 | > loss: -0.41622 (-0.47532) | > log_mle: -0.53748 (-0.59656) | > loss_dur: 0.12127 (0.12124) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.16964 (8.88237) | > current_lr: 0.00039 | > step_time: 0.64360 (0.50610) | > loader_time: 0.00350 (0.00322)  --> STEP: 310/639 -- GLOBAL_STEP: 26550 | > loss: -0.45884 (-0.47415) | > log_mle: -0.58202 (-0.59552) | > loss_dur: 0.12318 (0.12137) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.28491 (8.84041) | > current_lr: 0.00039 | > step_time: 0.68720 (0.52151) | > loader_time: 0.00360 (0.00325)  --> STEP: 335/639 -- GLOBAL_STEP: 26575 | > loss: -0.47709 (-0.47209) | > log_mle: -0.60524 (-0.59387) | > loss_dur: 0.12814 (0.12178) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.81698 (8.76181) | > current_lr: 0.00039 | > step_time: 0.69960 (0.53725) | > loader_time: 0.00330 (0.00328)  --> STEP: 360/639 -- GLOBAL_STEP: 26600 | > loss: -0.42211 (-0.47082) | > log_mle: -0.53910 (-0.59228) | > loss_dur: 0.11699 (0.12145) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.90884 (8.73454) | > current_lr: 0.00039 | > step_time: 0.69170 (0.55083) | > loader_time: 0.00330 (0.00332)  --> STEP: 385/639 -- GLOBAL_STEP: 26625 | > loss: -0.46170 (-0.46967) | > log_mle: -0.57906 (-0.59089) | > loss_dur: 0.11736 (0.12122) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.45766 (8.64874) | > current_lr: 0.00039 | > step_time: 0.69980 (0.56345) | > loader_time: 0.00380 (0.00335)  --> STEP: 410/639 -- GLOBAL_STEP: 26650 | > loss: -0.41673 (-0.46827) | > log_mle: -0.55445 (-0.58961) | > loss_dur: 0.13772 (0.12134) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.88590 (8.58231) | > current_lr: 0.00039 | > step_time: 0.88830 (0.57534) | > loader_time: 0.00410 (0.00338)  --> STEP: 435/639 -- GLOBAL_STEP: 26675 | > loss: -0.44841 (-0.46665) | > log_mle: -0.56077 (-0.58791) | > loss_dur: 0.11236 (0.12126) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.30861 (8.47972) | > current_lr: 0.00039 | > step_time: 0.72240 (0.58754) | > loader_time: 0.00370 (0.00339)  --> STEP: 460/639 -- GLOBAL_STEP: 26700 | > loss: -0.40285 (-0.46516) | > log_mle: -0.52618 (-0.58626) | > loss_dur: 0.12333 (0.12110) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.67672 (8.31281) | > current_lr: 0.00039 | > step_time: 0.94610 (0.59923) | > loader_time: 0.00410 (0.00341)  --> STEP: 485/639 -- GLOBAL_STEP: 26725 | > loss: -0.44945 (-0.46389) | > log_mle: -0.56122 (-0.58515) | > loss_dur: 0.11177 (0.12126) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.57480 (8.27223) | > current_lr: 0.00039 | > step_time: 0.86980 (0.61154) | > loader_time: 0.00410 (0.00344)  --> STEP: 510/639 -- GLOBAL_STEP: 26750 | > loss: -0.44269 (-0.46303) | > log_mle: -0.56469 (-0.58419) | > loss_dur: 0.12200 (0.12116) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.91700 (8.24716) | > current_lr: 0.00039 | > step_time: 0.92800 (0.62443) | > loader_time: 0.00410 (0.00346)  --> STEP: 535/639 -- GLOBAL_STEP: 26775 | > loss: -0.43521 (-0.46201) | > log_mle: -0.55744 (-0.58314) | > loss_dur: 0.12223 (0.12112) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.74306 (8.19643) | > current_lr: 0.00039 | > step_time: 0.90940 (0.63781) | > loader_time: 0.00430 (0.00348)  --> STEP: 560/639 -- GLOBAL_STEP: 26800 | > loss: -0.43302 (-0.46086) | > log_mle: -0.56071 (-0.58205) | > loss_dur: 0.12769 (0.12119) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.92624 (8.14124) | > current_lr: 0.00039 | > step_time: 0.91670 (0.65290) | > loader_time: 0.00420 (0.00351)  --> STEP: 585/639 -- GLOBAL_STEP: 26825 | > loss: -0.44872 (-0.45995) | > log_mle: -0.56667 (-0.58110) | > loss_dur: 0.11795 (0.12115) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.89845 (8.10698) | > current_lr: 0.00039 | > step_time: 1.02930 (0.66835) | > loader_time: 0.00440 (0.00354)  --> STEP: 610/639 -- GLOBAL_STEP: 26850 | > loss: -0.41765 (-0.45907) | > log_mle: -0.53675 (-0.58024) | > loss_dur: 0.11910 (0.12117) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.80202 (8.06489) | > current_lr: 0.00039 | > step_time: 1.05470 (0.68418) | > loader_time: 0.00420 (0.00357)  --> STEP: 635/639 -- GLOBAL_STEP: 26875 | > loss: -0.43996 (-0.45846) | > log_mle: -0.56208 (-0.57978) | > loss_dur: 0.12213 (0.12132) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.11878 (8.05352) | > current_lr: 0.00039 | > step_time: 1.13930 (0.70120) | > loader_time: 0.00380 (0.00360) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03381 (-0.00484) | > avg_loss: -0.44730 (-0.00850) | > avg_log_mle: -0.57239 (-0.00385) | > avg_loss_dur: 0.12509 (-0.00465) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_26880.pth.tar  > EPOCH: 42/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 20:09:04)   --> STEP: 20/639 -- GLOBAL_STEP: 26900 | > loss: -0.59320 (-0.51215) | > log_mle: -0.75092 (-0.63817) | > loss_dur: 0.15771 (0.12602) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 24.85586 (10.78292) | > current_lr: 0.00039 | > step_time: 0.33950 (0.31330) | > loader_time: 0.00280 (0.00243)  --> STEP: 45/639 -- GLOBAL_STEP: 26925 | > loss: -0.44517 (-0.48957) | > log_mle: -0.55015 (-0.61329) | > loss_dur: 0.10498 (0.12372) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.39943 (9.59577) | > current_lr: 0.00039 | > step_time: 0.47970 (0.34310) | > loader_time: 0.00300 (0.00265)  --> STEP: 70/639 -- GLOBAL_STEP: 26950 | > loss: -0.49545 (-0.48018) | > log_mle: -0.61891 (-0.60425) | > loss_dur: 0.12346 (0.12407) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.04197 (8.86540) | > current_lr: 0.00039 | > step_time: 0.39080 (0.36690) | > loader_time: 0.00290 (0.00275)  --> STEP: 95/639 -- GLOBAL_STEP: 26975 | > loss: -0.54945 (-0.47842) | > log_mle: -0.66662 (-0.60052) | > loss_dur: 0.11718 (0.12210) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.58693 (8.42180) | > current_lr: 0.00039 | > step_time: 0.41950 (0.38222) | > loader_time: 0.00340 (0.00282)  --> STEP: 120/639 -- GLOBAL_STEP: 27000 | > loss: -0.52019 (-0.47738) | > log_mle: -0.66038 (-0.60008) | > loss_dur: 0.14019 (0.12270) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.96871 (8.54419) | > current_lr: 0.00039 | > step_time: 0.55410 (0.39897) | > loader_time: 0.00320 (0.00288)  --> STEP: 145/639 -- GLOBAL_STEP: 27025 | > loss: -0.47846 (-0.47818) | > log_mle: -0.61219 (-0.60098) | > loss_dur: 0.13373 (0.12280) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.58537 (8.74636) | > current_lr: 0.00038 | > step_time: 0.47050 (0.41366) | > loader_time: 0.00290 (0.00291)  --> STEP: 170/639 -- GLOBAL_STEP: 27050 | > loss: -0.48667 (-0.47799) | > log_mle: -0.59880 (-0.60026) | > loss_dur: 0.11213 (0.12226) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.74441 (8.73206) | > current_lr: 0.00038 | > step_time: 0.54530 (0.42796) | > loader_time: 0.00330 (0.00294)  --> STEP: 195/639 -- GLOBAL_STEP: 27075 | > loss: -0.43079 (-0.47772) | > log_mle: -0.53646 (-0.59946) | > loss_dur: 0.10568 (0.12173) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.98079 (8.76109) | > current_lr: 0.00038 | > step_time: 0.50070 (0.44543) | > loader_time: 0.00320 (0.00298)  --> STEP: 220/639 -- GLOBAL_STEP: 27100 | > loss: -0.48706 (-0.47626) | > log_mle: -0.59511 (-0.59770) | > loss_dur: 0.10805 (0.12144) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.74138 (8.86414) | > current_lr: 0.00038 | > step_time: 0.61170 (0.46106) | > loader_time: 0.00360 (0.00303)  --> STEP: 245/639 -- GLOBAL_STEP: 27125 | > loss: -0.47157 (-0.47760) | > log_mle: -0.58118 (-0.59881) | > loss_dur: 0.10961 (0.12121) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.62597 (9.03499) | > current_lr: 0.00038 | > step_time: 0.53430 (0.47631) | > loader_time: 0.00330 (0.00307)  --> STEP: 270/639 -- GLOBAL_STEP: 27150 | > loss: -0.51295 (-0.47786) | > log_mle: -0.64132 (-0.59921) | > loss_dur: 0.12837 (0.12135) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 19.62182 (9.26159) | > current_lr: 0.00038 | > step_time: 0.83450 (0.49432) | > loader_time: 0.00350 (0.00311)  --> STEP: 295/639 -- GLOBAL_STEP: 27175 | > loss: -0.43445 (-0.47638) | > log_mle: -0.53916 (-0.59770) | > loss_dur: 0.10471 (0.12132) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.94167 (9.07952) | > current_lr: 0.00038 | > step_time: 0.63700 (0.50882) | > loader_time: 0.00350 (0.00314)  --> STEP: 320/639 -- GLOBAL_STEP: 27200 | > loss: -0.45218 (-0.47511) | > log_mle: -0.58452 (-0.59665) | > loss_dur: 0.13234 (0.12153) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.51880 (9.06022) | > current_lr: 0.00038 | > step_time: 0.74660 (0.52336) | > loader_time: 0.00380 (0.00317)  --> STEP: 345/639 -- GLOBAL_STEP: 27225 | > loss: -0.46224 (-0.47356) | > log_mle: -0.57687 (-0.59521) | > loss_dur: 0.11464 (0.12165) | > amp_scaler: 8192.00000 (4202.85217) | > grad_norm: 7.65533 (9.05530) | > current_lr: 0.00038 | > step_time: 0.66290 (0.53847) | > loader_time: 0.00340 (0.00321)  --> STEP: 370/639 -- GLOBAL_STEP: 27250 | > loss: -0.47811 (-0.47214) | > log_mle: -0.59440 (-0.59344) | > loss_dur: 0.11629 (0.12130) | > amp_scaler: 8192.00000 (4472.38919) | > grad_norm: 8.97488 (8.90647) | > current_lr: 0.00038 | > step_time: 0.89480 (0.55007) | > loader_time: 0.00410 (0.00323)  --> STEP: 395/639 -- GLOBAL_STEP: 27275 | > loss: -0.44137 (-0.47068) | > log_mle: -0.55911 (-0.59184) | > loss_dur: 0.11773 (0.12117) | > amp_scaler: 8192.00000 (4707.80759) | > grad_norm: 8.06122 (8.78153) | > current_lr: 0.00038 | > step_time: 0.73380 (0.56341) | > loader_time: 0.00380 (0.00327)  --> STEP: 420/639 -- GLOBAL_STEP: 27300 | > loss: -0.43322 (-0.46908) | > log_mle: -0.55271 (-0.59034) | > loss_dur: 0.11949 (0.12126) | > amp_scaler: 8192.00000 (4915.20000) | > grad_norm: 6.90258 (8.68498) | > current_lr: 0.00038 | > step_time: 0.68180 (0.57728) | > loader_time: 0.00390 (0.00331)  --> STEP: 445/639 -- GLOBAL_STEP: 27325 | > loss: -0.44712 (-0.46797) | > log_mle: -0.56062 (-0.58907) | > loss_dur: 0.11350 (0.12110) | > amp_scaler: 8192.00000 (5099.28989) | > grad_norm: 8.05720 (8.61729) | > current_lr: 0.00038 | > step_time: 0.74820 (0.58874) | > loader_time: 0.00360 (0.00334)  --> STEP: 470/639 -- GLOBAL_STEP: 27350 | > loss: -0.45204 (-0.46647) | > log_mle: -0.57105 (-0.58744) | > loss_dur: 0.11901 (0.12098) | > amp_scaler: 4096.00000 (5202.79149) | > grad_norm: 11.01766 (8.46041) | > current_lr: 0.00038 | > step_time: 0.82310 (0.60090) | > loader_time: 0.00380 (0.00337)  --> STEP: 495/639 -- GLOBAL_STEP: 27375 | > loss: -0.40567 (-0.46530) | > log_mle: -0.50944 (-0.58626) | > loss_dur: 0.10377 (0.12096) | > amp_scaler: 4096.00000 (5146.89293) | > grad_norm: 1.30098 (8.35518) | > current_lr: 0.00038 | > step_time: 0.82330 (0.61382) | > loader_time: 0.00380 (0.00340)  --> STEP: 520/639 -- GLOBAL_STEP: 27400 | > loss: -0.45324 (-0.46457) | > log_mle: -0.56706 (-0.58540) | > loss_dur: 0.11383 (0.12082) | > amp_scaler: 4096.00000 (5096.36923) | > grad_norm: 10.77673 (8.28720) | > current_lr: 0.00038 | > step_time: 0.81330 (0.62677) | > loader_time: 0.00380 (0.00344)  --> STEP: 545/639 -- GLOBAL_STEP: 27425 | > loss: -0.47084 (-0.46356) | > log_mle: -0.59891 (-0.58443) | > loss_dur: 0.12807 (0.12087) | > amp_scaler: 4096.00000 (5050.48073) | > grad_norm: 8.40798 (8.20735) | > current_lr: 0.00038 | > step_time: 0.95740 (0.64219) | > loader_time: 0.00420 (0.00348)  --> STEP: 570/639 -- GLOBAL_STEP: 27450 | > loss: -0.43277 (-0.46234) | > log_mle: -0.54453 (-0.58320) | > loss_dur: 0.11176 (0.12086) | > amp_scaler: 4096.00000 (5008.61754) | > grad_norm: 4.10162 (8.11926) | > current_lr: 0.00038 | > step_time: 0.92030 (0.65606) | > loader_time: 0.00390 (0.00352)  --> STEP: 595/639 -- GLOBAL_STEP: 27475 | > loss: -0.42667 (-0.46152) | > log_mle: -0.53908 (-0.58246) | > loss_dur: 0.11241 (0.12094) | > amp_scaler: 4096.00000 (4970.27227) | > grad_norm: 5.08024 (8.08664) | > current_lr: 0.00038 | > step_time: 1.02080 (0.67236) | > loader_time: 0.00450 (0.00355)  --> STEP: 620/639 -- GLOBAL_STEP: 27500 | > loss: -0.43426 (-0.46061) | > log_mle: -0.54797 (-0.58154) | > loss_dur: 0.11371 (0.12093) | > amp_scaler: 4096.00000 (4935.01935) | > grad_norm: 5.32826 (8.03362) | > current_lr: 0.00038 | > step_time: 1.20310 (0.68903) | > loader_time: 0.00470 (0.00359) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02451 (-0.00930) | > avg_loss: -0.44721 (+0.00009) | > avg_log_mle: -0.57325 (-0.00086) | > avg_loss_dur: 0.12604 (+0.00095)  > EPOCH: 43/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 20:16:47)   --> STEP: 5/639 -- GLOBAL_STEP: 27525 | > loss: -0.48954 (-0.50955) | > log_mle: -0.66164 (-0.64529) | > loss_dur: 0.17210 (0.13574) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 21.49775 (13.01585) | > current_lr: 0.00038 | > step_time: 0.31070 (0.28832) | > loader_time: 0.00300 (0.00263)  --> STEP: 30/639 -- GLOBAL_STEP: 27550 | > loss: -0.49591 (-0.49430) | > log_mle: -0.61932 (-0.62119) | > loss_dur: 0.12341 (0.12689) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.45639 (11.03141) | > current_lr: 0.00038 | > step_time: 0.33390 (0.33661) | > loader_time: 0.00280 (0.00271)  --> STEP: 55/639 -- GLOBAL_STEP: 27575 | > loss: -0.45003 (-0.48564) | > log_mle: -0.54628 (-0.60932) | > loss_dur: 0.09625 (0.12368) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.56294 (10.07614) | > current_lr: 0.00038 | > step_time: 0.33210 (0.35394) | > loader_time: 0.00290 (0.00276)  --> STEP: 80/639 -- GLOBAL_STEP: 27600 | > loss: -0.46442 (-0.47963) | > log_mle: -0.59964 (-0.60228) | > loss_dur: 0.13522 (0.12266) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.60197 (9.29740) | > current_lr: 0.00038 | > step_time: 0.41210 (0.37068) | > loader_time: 0.00300 (0.00282)  --> STEP: 105/639 -- GLOBAL_STEP: 27625 | > loss: -0.43205 (-0.47906) | > log_mle: -0.55311 (-0.60102) | > loss_dur: 0.12106 (0.12196) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.25880 (9.15615) | > current_lr: 0.00038 | > step_time: 0.41410 (0.38862) | > loader_time: 0.00310 (0.00289)  --> STEP: 130/639 -- GLOBAL_STEP: 27650 | > loss: -0.45647 (-0.47841) | > log_mle: -0.59028 (-0.60051) | > loss_dur: 0.13380 (0.12211) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.31874 (9.03564) | > current_lr: 0.00038 | > step_time: 0.50750 (0.40202) | > loader_time: 0.00290 (0.00292)  --> STEP: 155/639 -- GLOBAL_STEP: 27675 | > loss: -0.45918 (-0.47850) | > log_mle: -0.58555 (-0.60057) | > loss_dur: 0.12638 (0.12207) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.22462 (9.03057) | > current_lr: 0.00038 | > step_time: 0.54730 (0.41950) | > loader_time: 0.00310 (0.00299)  --> STEP: 180/639 -- GLOBAL_STEP: 27700 | > loss: -0.51364 (-0.48076) | > log_mle: -0.63616 (-0.60284) | > loss_dur: 0.12252 (0.12207) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.33845 (9.52942) | > current_lr: 0.00038 | > step_time: 0.58200 (0.43684) | > loader_time: 0.00350 (0.00302)  --> STEP: 205/639 -- GLOBAL_STEP: 27725 | > loss: -0.44456 (-0.47814) | > log_mle: -0.55929 (-0.59963) | > loss_dur: 0.11473 (0.12148) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.17707 (9.21706) | > current_lr: 0.00038 | > step_time: 0.55860 (0.45255) | > loader_time: 0.00340 (0.00305)  --> STEP: 230/639 -- GLOBAL_STEP: 27750 | > loss: -0.46796 (-0.47873) | > log_mle: -0.59288 (-0.59994) | > loss_dur: 0.12492 (0.12122) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.46828 (9.29360) | > current_lr: 0.00038 | > step_time: 0.55420 (0.46943) | > loader_time: 0.00320 (0.00309)  --> STEP: 255/639 -- GLOBAL_STEP: 27775 | > loss: -0.44902 (-0.47833) | > log_mle: -0.56869 (-0.59932) | > loss_dur: 0.11967 (0.12099) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.24203 (9.24939) | > current_lr: 0.00038 | > step_time: 0.65340 (0.48428) | > loader_time: 0.00350 (0.00312)  --> STEP: 280/639 -- GLOBAL_STEP: 27800 | > loss: -0.46162 (-0.47877) | > log_mle: -0.58848 (-0.59983) | > loss_dur: 0.12686 (0.12107) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.98653 (9.31199) | > current_lr: 0.00038 | > step_time: 0.60930 (0.50144) | > loader_time: 0.00340 (0.00316)  --> STEP: 305/639 -- GLOBAL_STEP: 27825 | > loss: -0.44430 (-0.47730) | > log_mle: -0.56813 (-0.59853) | > loss_dur: 0.12383 (0.12123) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.06729 (9.19937) | > current_lr: 0.00038 | > step_time: 0.66840 (0.51612) | > loader_time: 0.00360 (0.00320)  --> STEP: 330/639 -- GLOBAL_STEP: 27850 | > loss: -0.46246 (-0.47528) | > log_mle: -0.59604 (-0.59692) | > loss_dur: 0.13358 (0.12164) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.62303 (9.09583) | > current_lr: 0.00038 | > step_time: 0.75890 (0.53161) | > loader_time: 0.00360 (0.00323)  --> STEP: 355/639 -- GLOBAL_STEP: 27875 | > loss: -0.41514 (-0.47406) | > log_mle: -0.52339 (-0.59536) | > loss_dur: 0.10826 (0.12129) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.71918 (9.05533) | > current_lr: 0.00038 | > step_time: 0.60250 (0.54410) | > loader_time: 0.00340 (0.00326)  --> STEP: 380/639 -- GLOBAL_STEP: 27900 | > loss: -0.45427 (-0.47242) | > log_mle: -0.55942 (-0.59346) | > loss_dur: 0.10515 (0.12104) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.98945 (8.85550) | > current_lr: 0.00038 | > step_time: 0.64720 (0.55739) | > loader_time: 0.00360 (0.00328)  --> STEP: 405/639 -- GLOBAL_STEP: 27925 | > loss: -0.45273 (-0.47157) | > log_mle: -0.56655 (-0.59263) | > loss_dur: 0.11382 (0.12106) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.30083 (8.81424) | > current_lr: 0.00038 | > step_time: 0.68040 (0.56980) | > loader_time: 0.00360 (0.00331)  --> STEP: 430/639 -- GLOBAL_STEP: 27950 | > loss: -0.41502 (-0.47000) | > log_mle: -0.52569 (-0.59106) | > loss_dur: 0.11067 (0.12107) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.08472 (8.71032) | > current_lr: 0.00038 | > step_time: 0.83800 (0.58301) | > loader_time: 0.00370 (0.00334)  --> STEP: 455/639 -- GLOBAL_STEP: 27975 | > loss: -0.42797 (-0.46842) | > log_mle: -0.55168 (-0.58931) | > loss_dur: 0.12371 (0.12089) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.39189 (8.57966) | > current_lr: 0.00038 | > step_time: 0.73890 (0.59375) | > loader_time: 0.00350 (0.00337)  --> STEP: 480/639 -- GLOBAL_STEP: 28000 | > loss: -0.38787 (-0.46698) | > log_mle: -0.50699 (-0.58799) | > loss_dur: 0.11912 (0.12101) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.88827 (8.57142) | > current_lr: 0.00038 | > step_time: 0.81900 (0.60669) | > loader_time: 0.00370 (0.00340)  --> STEP: 505/639 -- GLOBAL_STEP: 28025 | > loss: -0.45895 (-0.46600) | > log_mle: -0.57572 (-0.58684) | > loss_dur: 0.11676 (0.12084) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.69346 (8.47519) | > current_lr: 0.00038 | > step_time: 0.91850 (0.61957) | > loader_time: 0.00360 (0.00342)  --> STEP: 530/639 -- GLOBAL_STEP: 28050 | > loss: -0.43491 (-0.46505) | > log_mle: -0.56602 (-0.58590) | > loss_dur: 0.13111 (0.12085) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.15083 (8.42135) | > current_lr: 0.00038 | > step_time: 0.93370 (0.63314) | > loader_time: 0.00410 (0.00345)  --> STEP: 555/639 -- GLOBAL_STEP: 28075 | > loss: -0.43002 (-0.46384) | > log_mle: -0.54850 (-0.58465) | > loss_dur: 0.11848 (0.12081) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.51552 (8.34545) | > current_lr: 0.00038 | > step_time: 0.83310 (0.64731) | > loader_time: 0.00380 (0.00349)  --> STEP: 580/639 -- GLOBAL_STEP: 28100 | > loss: -0.45940 (-0.46294) | > log_mle: -0.56873 (-0.58372) | > loss_dur: 0.10933 (0.12078) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.64948 (8.30903) | > current_lr: 0.00038 | > step_time: 0.96810 (0.66301) | > loader_time: 0.00430 (0.00352)  --> STEP: 605/639 -- GLOBAL_STEP: 28125 | > loss: -0.42474 (-0.46200) | > log_mle: -0.53463 (-0.58287) | > loss_dur: 0.10989 (0.12087) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.29967 (8.29640) | > current_lr: 0.00038 | > step_time: 1.05210 (0.67874) | > loader_time: 0.00430 (0.00356)  --> STEP: 630/639 -- GLOBAL_STEP: 28150 | > loss: -0.43442 (-0.46126) | > log_mle: -0.56911 (-0.58219) | > loss_dur: 0.13469 (0.12094) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.42004 (8.28432) | > current_lr: 0.00038 | > step_time: 1.18880 (0.69607) | > loader_time: 0.00490 (0.00360) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02797 (+0.00346) | > avg_loss: -0.44773 (-0.00052) | > avg_log_mle: -0.57563 (-0.00238) | > avg_loss_dur: 0.12790 (+0.00187) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_28160.pth.tar  > EPOCH: 44/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 20:24:37)   --> STEP: 15/639 -- GLOBAL_STEP: 28175 | > loss: -0.62057 (-0.51357) | > log_mle: -0.75505 (-0.63841) | > loss_dur: 0.13447 (0.12484) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 21.96354 (11.26311) | > current_lr: 0.00038 | > step_time: 0.34810 (0.31004) | > loader_time: 0.00280 (0.00411)  --> STEP: 40/639 -- GLOBAL_STEP: 28200 | > loss: -0.47312 (-0.49308) | > log_mle: -0.58873 (-0.61586) | > loss_dur: 0.11561 (0.12278) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.48680 (9.92344) | > current_lr: 0.00038 | > step_time: 0.36010 (0.33614) | > loader_time: 0.00290 (0.00328)  --> STEP: 65/639 -- GLOBAL_STEP: 28225 | > loss: -0.47331 (-0.48461) | > log_mle: -0.57757 (-0.60793) | > loss_dur: 0.10426 (0.12333) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.98690 (9.63357) | > current_lr: 0.00038 | > step_time: 0.40750 (0.36784) | > loader_time: 0.00290 (0.00317)  --> STEP: 90/639 -- GLOBAL_STEP: 28250 | > loss: -0.42266 (-0.47849) | > log_mle: -0.54009 (-0.59989) | > loss_dur: 0.11742 (0.12139) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.45557 (8.91905) | > current_lr: 0.00038 | > step_time: 0.53980 (0.38635) | > loader_time: 0.00320 (0.00311)  --> STEP: 115/639 -- GLOBAL_STEP: 28275 | > loss: -0.38594 (-0.47955) | > log_mle: -0.50886 (-0.60115) | > loss_dur: 0.12292 (0.12159) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.69536 (9.19128) | > current_lr: 0.00038 | > step_time: 0.40200 (0.40278) | > loader_time: 0.00310 (0.00309)  --> STEP: 140/639 -- GLOBAL_STEP: 28300 | > loss: -0.45528 (-0.48015) | > log_mle: -0.59434 (-0.60193) | > loss_dur: 0.13906 (0.12177) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.77014 (9.58778) | > current_lr: 0.00038 | > step_time: 0.53570 (0.41585) | > loader_time: 0.00340 (0.00308)  --> STEP: 165/639 -- GLOBAL_STEP: 28325 | > loss: -0.46591 (-0.48001) | > log_mle: -0.59852 (-0.60130) | > loss_dur: 0.13261 (0.12129) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.70854 (9.49622) | > current_lr: 0.00038 | > step_time: 0.58830 (0.43060) | > loader_time: 0.00300 (0.00309)  --> STEP: 190/639 -- GLOBAL_STEP: 28350 | > loss: -0.41052 (-0.48044) | > log_mle: -0.52989 (-0.60132) | > loss_dur: 0.11936 (0.12087) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.42751 (9.54269) | > current_lr: 0.00038 | > step_time: 0.52570 (0.44781) | > loader_time: 0.00330 (0.00311)  --> STEP: 215/639 -- GLOBAL_STEP: 28375 | > loss: -0.52433 (-0.47911) | > log_mle: -0.63639 (-0.59982) | > loss_dur: 0.11205 (0.12071) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.24160 (9.37004) | > current_lr: 0.00038 | > step_time: 0.53950 (0.46232) | > loader_time: 0.00320 (0.00313)  --> STEP: 240/639 -- GLOBAL_STEP: 28400 | > loss: -0.52588 (-0.48073) | > log_mle: -0.64558 (-0.60091) | > loss_dur: 0.11971 (0.12019) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.32666 (9.44369) | > current_lr: 0.00038 | > step_time: 0.57230 (0.47717) | > loader_time: 0.00350 (0.00315)  --> STEP: 265/639 -- GLOBAL_STEP: 28425 | > loss: -0.53541 (-0.48043) | > log_mle: -0.67205 (-0.60083) | > loss_dur: 0.13664 (0.12040) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.51629 (9.54848) | > current_lr: 0.00038 | > step_time: 0.62890 (0.49220) | > loader_time: 0.00360 (0.00320)  --> STEP: 290/639 -- GLOBAL_STEP: 28450 | > loss: -0.45449 (-0.47929) | > log_mle: -0.56960 (-0.59972) | > loss_dur: 0.11511 (0.12044) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.89623 (9.49240) | > current_lr: 0.00038 | > step_time: 0.63650 (0.50757) | > loader_time: 0.00340 (0.00324)  --> STEP: 315/639 -- GLOBAL_STEP: 28475 | > loss: -0.43301 (-0.47795) | > log_mle: -0.54483 (-0.59855) | > loss_dur: 0.11182 (0.12060) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.81202 (9.47724) | > current_lr: 0.00037 | > step_time: 0.63600 (0.52186) | > loader_time: 0.00360 (0.00328)  --> STEP: 340/639 -- GLOBAL_STEP: 28500 | > loss: -0.45530 (-0.47648) | > log_mle: -0.56725 (-0.59745) | > loss_dur: 0.11194 (0.12097) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.33961 (9.52797) | > current_lr: 0.00037 | > step_time: 0.63630 (0.53719) | > loader_time: 0.00360 (0.00332)  --> STEP: 365/639 -- GLOBAL_STEP: 28525 | > loss: -0.46756 (-0.47480) | > log_mle: -0.57953 (-0.59538) | > loss_dur: 0.11198 (0.12058) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.15655 (9.37732) | > current_lr: 0.00037 | > step_time: 0.83450 (0.54809) | > loader_time: 0.00380 (0.00335)  --> STEP: 390/639 -- GLOBAL_STEP: 28550 | > loss: -0.44279 (-0.47328) | > log_mle: -0.57408 (-0.59371) | > loss_dur: 0.13129 (0.12043) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.59223 (9.25916) | > current_lr: 0.00037 | > step_time: 0.73860 (0.56061) | > loader_time: 0.00350 (0.00337)  --> STEP: 415/639 -- GLOBAL_STEP: 28575 | > loss: -0.42594 (-0.47171) | > log_mle: -0.54814 (-0.59230) | > loss_dur: 0.12221 (0.12059) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.15425 (9.22215) | > current_lr: 0.00037 | > step_time: 0.73940 (0.57307) | > loader_time: 0.00340 (0.00339)  --> STEP: 440/639 -- GLOBAL_STEP: 28600 | > loss: -0.41362 (-0.47036) | > log_mle: -0.53400 (-0.59096) | > loss_dur: 0.12038 (0.12060) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.38070 (9.15451) | > current_lr: 0.00037 | > step_time: 0.71900 (0.58529) | > loader_time: 0.00390 (0.00342)  --> STEP: 465/639 -- GLOBAL_STEP: 28625 | > loss: -0.42352 (-0.46873) | > log_mle: -0.53776 (-0.58919) | > loss_dur: 0.11424 (0.12046) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.68021 (9.01378) | > current_lr: 0.00037 | > step_time: 0.81230 (0.59679) | > loader_time: 0.00400 (0.00346)  --> STEP: 490/639 -- GLOBAL_STEP: 28650 | > loss: -0.41639 (-0.46740) | > log_mle: -0.52081 (-0.58791) | > loss_dur: 0.10442 (0.12051) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.50656 (8.92494) | > current_lr: 0.00037 | > step_time: 0.77080 (0.60966) | > loader_time: 0.00360 (0.00348)  --> STEP: 515/639 -- GLOBAL_STEP: 28675 | > loss: -0.44143 (-0.46663) | > log_mle: -0.55608 (-0.58706) | > loss_dur: 0.11465 (0.12044) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.03262 (8.89077) | > current_lr: 0.00037 | > step_time: 0.99810 (0.62299) | > loader_time: 0.00440 (0.00351)  --> STEP: 540/639 -- GLOBAL_STEP: 28700 | > loss: -0.44803 (-0.46545) | > log_mle: -0.57131 (-0.58589) | > loss_dur: 0.12328 (0.12044) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.82386 (8.80738) | > current_lr: 0.00037 | > step_time: 0.90760 (0.63731) | > loader_time: 0.00430 (0.00355)  --> STEP: 565/639 -- GLOBAL_STEP: 28725 | > loss: -0.41112 (-0.46422) | > log_mle: -0.52958 (-0.58468) | > loss_dur: 0.11846 (0.12046) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.75397 (8.72374) | > current_lr: 0.00037 | > step_time: 0.99200 (0.65138) | > loader_time: 0.00420 (0.00358)  --> STEP: 590/639 -- GLOBAL_STEP: 28750 | > loss: -0.44230 (-0.46353) | > log_mle: -0.56412 (-0.58401) | > loss_dur: 0.12182 (0.12049) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.75636 (8.67283) | > current_lr: 0.00037 | > step_time: 0.95380 (0.66753) | > loader_time: 0.00420 (0.00361)  --> STEP: 615/639 -- GLOBAL_STEP: 28775 | > loss: -0.44285 (-0.46260) | > log_mle: -0.56691 (-0.58310) | > loss_dur: 0.12405 (0.12050) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.31411 (8.61009) | > current_lr: 0.00037 | > step_time: 1.04390 (0.68338) | > loader_time: 0.00450 (0.00364) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02953 (+0.00156) | > avg_loss: -0.44673 (+0.00099) | > avg_log_mle: -0.57269 (+0.00294) | > avg_loss_dur: 0.12596 (-0.00194)  > EPOCH: 45/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 20:32:20)   --> STEP: 0/639 -- GLOBAL_STEP: 28800 | > loss: -0.52259 (-0.52259) | > log_mle: -0.68499 (-0.68499) | > loss_dur: 0.16241 (0.16241) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 20.19922 (20.19922) | > current_lr: 0.00037 | > step_time: 0.81980 (0.81980) | > loader_time: 0.37640 (0.37639)  --> STEP: 25/639 -- GLOBAL_STEP: 28825 | > loss: -0.49671 (-0.50688) | > log_mle: -0.64259 (-0.63464) | > loss_dur: 0.14588 (0.12777) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.55881 (10.32113) | > current_lr: 0.00037 | > step_time: 0.40200 (0.32804) | > loader_time: 0.00300 (0.00250)  --> STEP: 50/639 -- GLOBAL_STEP: 28850 | > loss: -0.44983 (-0.49338) | > log_mle: -0.56084 (-0.61756) | > loss_dur: 0.11101 (0.12418) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.01587 (9.38705) | > current_lr: 0.00037 | > step_time: 0.31420 (0.35157) | > loader_time: 0.00280 (0.00260)  --> STEP: 75/639 -- GLOBAL_STEP: 28875 | > loss: -0.47920 (-0.48535) | > log_mle: -0.59344 (-0.60815) | > loss_dur: 0.11424 (0.12280) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.08415 (8.97388) | > current_lr: 0.00037 | > step_time: 0.42240 (0.37297) | > loader_time: 0.00300 (0.00266)  --> STEP: 100/639 -- GLOBAL_STEP: 28900 | > loss: -0.48950 (-0.48403) | > log_mle: -0.60613 (-0.60544) | > loss_dur: 0.11663 (0.12142) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.80978 (8.67039) | > current_lr: 0.00037 | > step_time: 0.50120 (0.39074) | > loader_time: 0.00310 (0.00272)  --> STEP: 125/639 -- GLOBAL_STEP: 28925 | > loss: -0.48591 (-0.48311) | > log_mle: -0.59981 (-0.60464) | > loss_dur: 0.11390 (0.12153) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.49018 (8.88671) | > current_lr: 0.00037 | > step_time: 0.42120 (0.40402) | > loader_time: 0.00300 (0.00279)  --> STEP: 150/639 -- GLOBAL_STEP: 28950 | > loss: -0.44137 (-0.48298) | > log_mle: -0.56838 (-0.60444) | > loss_dur: 0.12701 (0.12146) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.76875 (8.91117) | > current_lr: 0.00037 | > step_time: 0.42550 (0.41903) | > loader_time: 0.00270 (0.00281)  --> STEP: 175/639 -- GLOBAL_STEP: 28975 | > loss: -0.47116 (-0.48387) | > log_mle: -0.58019 (-0.60509) | > loss_dur: 0.10903 (0.12122) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.74205 (9.10377) | > current_lr: 0.00037 | > step_time: 0.55750 (0.43589) | > loader_time: 0.00340 (0.00285)  --> STEP: 200/639 -- GLOBAL_STEP: 29000 | > loss: -0.45127 (-0.48246) | > log_mle: -0.57073 (-0.60299) | > loss_dur: 0.11946 (0.12053) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.67788 (8.92877) | > current_lr: 0.00037 | > step_time: 0.54970 (0.45078) | > loader_time: 0.00280 (0.00288)  --> STEP: 225/639 -- GLOBAL_STEP: 29025 | > loss: -0.44229 (-0.48113) | > log_mle: -0.54671 (-0.60119) | > loss_dur: 0.10442 (0.12006) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.85439 (8.85278) | > current_lr: 0.00037 | > step_time: 0.52860 (0.46694) | > loader_time: 0.00300 (0.00292)  --> STEP: 250/639 -- GLOBAL_STEP: 29050 | > loss: -0.44825 (-0.48240) | > log_mle: -0.55093 (-0.60239) | > loss_dur: 0.10268 (0.11999) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.89572 (9.09134) | > current_lr: 0.00037 | > step_time: 0.57880 (0.48164) | > loader_time: 0.00300 (0.00296)  --> STEP: 275/639 -- GLOBAL_STEP: 29075 | > loss: -0.46893 (-0.48266) | > log_mle: -0.58857 (-0.60292) | > loss_dur: 0.11964 (0.12026) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.98903 (9.31388) | > current_lr: 0.00037 | > step_time: 0.88610 (0.49930) | > loader_time: 0.00370 (0.00303)  --> STEP: 300/639 -- GLOBAL_STEP: 29100 | > loss: -0.46994 (-0.48087) | > log_mle: -0.58411 (-0.60111) | > loss_dur: 0.11417 (0.12024) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.33130 (9.15382) | > current_lr: 0.00037 | > step_time: 0.70880 (0.51321) | > loader_time: 0.00350 (0.00306)  --> STEP: 325/639 -- GLOBAL_STEP: 29125 | > loss: -0.45876 (-0.47924) | > log_mle: -0.58661 (-0.59982) | > loss_dur: 0.12785 (0.12058) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.15130 (9.12979) | > current_lr: 0.00037 | > step_time: 0.63480 (0.52878) | > loader_time: 0.00360 (0.00312)  --> STEP: 350/639 -- GLOBAL_STEP: 29150 | > loss: -0.48340 (-0.47766) | > log_mle: -0.61575 (-0.59817) | > loss_dur: 0.13235 (0.12050) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.16724 (9.04870) | > current_lr: 0.00037 | > step_time: 0.87100 (0.54363) | > loader_time: 0.00430 (0.00316)  --> STEP: 375/639 -- GLOBAL_STEP: 29175 | > loss: -0.43088 (-0.47575) | > log_mle: -0.54906 (-0.59605) | > loss_dur: 0.11818 (0.12030) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.20980 (8.86757) | > current_lr: 0.00037 | > step_time: 0.79270 (0.55664) | > loader_time: 0.00370 (0.00320)  --> STEP: 400/639 -- GLOBAL_STEP: 29200 | > loss: -0.46613 (-0.47462) | > log_mle: -0.58294 (-0.59494) | > loss_dur: 0.11681 (0.12031) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.90667 (8.76943) | > current_lr: 0.00037 | > step_time: 0.72730 (0.56957) | > loader_time: 0.00370 (0.00323)  --> STEP: 425/639 -- GLOBAL_STEP: 29225 | > loss: -0.51615 (-0.47292) | > log_mle: -0.63530 (-0.59321) | > loss_dur: 0.11915 (0.12029) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 21.07304 (8.70369) | > current_lr: 0.00037 | > step_time: 0.88440 (0.58154) | > loader_time: 0.00360 (0.00326)  --> STEP: 450/639 -- GLOBAL_STEP: 29250 | > loss: -0.43055 (-0.47146) | > log_mle: -0.53646 (-0.59161) | > loss_dur: 0.10591 (0.12015) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.94562 (8.60581) | > current_lr: 0.00037 | > step_time: 0.81000 (0.59369) | > loader_time: 0.00430 (0.00328)  --> STEP: 475/639 -- GLOBAL_STEP: 29275 | > loss: -0.44784 (-0.46992) | > log_mle: -0.55016 (-0.59013) | > loss_dur: 0.10232 (0.12021) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.40771 (8.53089) | > current_lr: 0.00037 | > step_time: 0.70880 (0.60572) | > loader_time: 0.00350 (0.00331)  --> STEP: 500/639 -- GLOBAL_STEP: 29300 | > loss: -0.49305 (-0.46892) | > log_mle: -0.61493 (-0.58914) | > loss_dur: 0.12188 (0.12023) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.78617 (8.44655) | > current_lr: 0.00037 | > step_time: 0.94210 (0.61901) | > loader_time: 0.00400 (0.00334)  --> STEP: 525/639 -- GLOBAL_STEP: 29325 | > loss: -0.45103 (-0.46791) | > log_mle: -0.58868 (-0.58816) | > loss_dur: 0.13765 (0.12025) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.96469 (8.45061) | > current_lr: 0.00037 | > step_time: 1.05740 (0.63238) | > loader_time: 0.00410 (0.00337)  --> STEP: 550/639 -- GLOBAL_STEP: 29350 | > loss: -0.40789 (-0.46672) | > log_mle: -0.52544 (-0.58691) | > loss_dur: 0.11756 (0.12019) | > amp_scaler: 8192.00000 (4148.13091) | > grad_norm: 3.61636 (8.34613) | > current_lr: 0.00037 | > step_time: 0.93460 (0.64727) | > loader_time: 0.00440 (0.00340)  --> STEP: 575/639 -- GLOBAL_STEP: 29375 | > loss: -0.42723 (-0.46548) | > log_mle: -0.54456 (-0.58569) | > loss_dur: 0.11733 (0.12021) | > amp_scaler: 8192.00000 (4323.95130) | > grad_norm: 5.07504 (8.31992) | > current_lr: 0.00037 | > step_time: 0.91050 (0.66183) | > loader_time: 0.00350 (0.00344)  --> STEP: 600/639 -- GLOBAL_STEP: 29400 | > loss: -0.42004 (-0.46461) | > log_mle: -0.54219 (-0.58491) | > loss_dur: 0.12215 (0.12029) | > amp_scaler: 8192.00000 (4485.12000) | > grad_norm: 7.77886 (8.27917) | > current_lr: 0.00037 | > step_time: 1.06080 (0.67895) | > loader_time: 0.00550 (0.00349)  --> STEP: 625/639 -- GLOBAL_STEP: 29425 | > loss: -0.45878 (-0.46384) | > log_mle: -0.57634 (-0.58412) | > loss_dur: 0.11756 (0.12028) | > amp_scaler: 8192.00000 (4633.39520) | > grad_norm: 8.71220 (8.23892) | > current_lr: 0.00037 | > step_time: 1.10100 (0.69589) | > loader_time: 0.00440 (0.00355) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02800 (-0.00153) | > avg_loss: -0.44649 (+0.00024) | > avg_log_mle: -0.57722 (-0.00453) | > avg_loss_dur: 0.13073 (+0.00477)  > EPOCH: 46/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 20:40:05)   --> STEP: 10/639 -- GLOBAL_STEP: 29450 | > loss: -0.49852 (-0.50946) | > log_mle: -0.63799 (-0.63874) | > loss_dur: 0.13947 (0.12928) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.38474 (11.40499) | > current_lr: 0.00037 | > step_time: 0.28360 (0.30770) | > loader_time: 0.00270 (0.01279)  --> STEP: 35/639 -- GLOBAL_STEP: 29475 | > loss: -0.50245 (-0.49457) | > log_mle: -0.60128 (-0.61878) | > loss_dur: 0.09883 (0.12421) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.60438 (9.44471) | > current_lr: 0.00037 | > step_time: 0.35650 (0.32453) | > loader_time: 0.00260 (0.00553)  --> STEP: 60/639 -- GLOBAL_STEP: 29500 | > loss: -0.44211 (-0.48738) | > log_mle: -0.55649 (-0.61138) | > loss_dur: 0.11438 (0.12400) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.20337 (8.90201) | > current_lr: 0.00037 | > step_time: 0.42250 (0.35583) | > loader_time: 0.00290 (0.00443)  --> STEP: 85/639 -- GLOBAL_STEP: 29525 | > loss: -0.42956 (-0.48175) | > log_mle: -0.55287 (-0.60346) | > loss_dur: 0.12331 (0.12172) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.80906 (8.44867) | > current_lr: 0.00037 | > step_time: 0.41470 (0.37439) | > loader_time: 0.00310 (0.00398)  --> STEP: 110/639 -- GLOBAL_STEP: 29550 | > loss: -0.46363 (-0.48417) | > log_mle: -0.57764 (-0.60623) | > loss_dur: 0.11401 (0.12206) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.66274 (9.13563) | > current_lr: 0.00037 | > step_time: 0.46420 (0.39529) | > loader_time: 0.00280 (0.00376)  --> STEP: 135/639 -- GLOBAL_STEP: 29575 | > loss: -0.42850 (-0.48312) | > log_mle: -0.55741 (-0.60429) | > loss_dur: 0.12891 (0.12117) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.75062 (9.13305) | > current_lr: 0.00037 | > step_time: 0.75420 (0.41516) | > loader_time: 0.07300 (0.00487)  --> STEP: 160/639 -- GLOBAL_STEP: 29600 | > loss: -0.48816 (-0.48214) | > log_mle: -0.59977 (-0.60265) | > loss_dur: 0.11162 (0.12050) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.42054 (9.00778) | > current_lr: 0.00037 | > step_time: 0.81640 (0.46922) | > loader_time: 0.06730 (0.01320)  --> STEP: 185/639 -- GLOBAL_STEP: 29625 | > loss: -0.47050 (-0.48452) | > log_mle: -0.58322 (-0.60534) | > loss_dur: 0.11272 (0.12082) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.68207 (9.48866) | > current_lr: 0.00037 | > step_time: 0.86950 (0.51021) | > loader_time: 0.06360 (0.01960)  --> STEP: 210/639 -- GLOBAL_STEP: 29650 | > loss: -0.39490 (-0.48229) | > log_mle: -0.50133 (-0.60288) | > loss_dur: 0.10642 (0.12059) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.97301 (9.23262) | > current_lr: 0.00037 | > step_time: 0.82690 (0.54369) | > loader_time: 0.10170 (0.02510)  --> STEP: 235/639 -- GLOBAL_STEP: 29675 | > loss: -0.48856 (-0.48347) | > log_mle: -0.59933 (-0.60366) | > loss_dur: 0.11076 (0.12018) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.26126 (9.28551) | > current_lr: 0.00037 | > step_time: 0.63760 (0.57694) | > loader_time: 0.04930 (0.02871)  --> STEP: 260/639 -- GLOBAL_STEP: 29700 | > loss: -0.47806 (-0.48338) | > log_mle: -0.59629 (-0.60363) | > loss_dur: 0.11824 (0.12025) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.00164 (9.37467) | > current_lr: 0.00037 | > step_time: 0.82520 (0.60305) | > loader_time: 0.12210 (0.03218)  --> STEP: 285/639 -- GLOBAL_STEP: 29725 | > loss: -0.41636 (-0.48277) | > log_mle: -0.54177 (-0.60300) | > loss_dur: 0.12542 (0.12023) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.28590 (9.29466) | > current_lr: 0.00037 | > step_time: 0.61840 (0.62247) | > loader_time: 0.00340 (0.03268)  --> STEP: 310/639 -- GLOBAL_STEP: 29750 | > loss: -0.46842 (-0.48151) | > log_mle: -0.58816 (-0.60186) | > loss_dur: 0.11973 (0.12035) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.27927 (9.23541) | > current_lr: 0.00037 | > step_time: 0.57670 (0.62838) | > loader_time: 0.00340 (0.03034)  --> STEP: 335/639 -- GLOBAL_STEP: 29775 | > loss: -0.48132 (-0.47949) | > log_mle: -0.60951 (-0.60019) | > loss_dur: 0.12819 (0.12071) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.20152 (9.20004) | > current_lr: 0.00037 | > step_time: 0.65620 (0.63484) | > loader_time: 0.00360 (0.02835)  --> STEP: 360/639 -- GLOBAL_STEP: 29800 | > loss: -0.43315 (-0.47812) | > log_mle: -0.54724 (-0.59853) | > loss_dur: 0.11409 (0.12041) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.22798 (9.05986) | > current_lr: 0.00037 | > step_time: 0.68890 (0.64052) | > loader_time: 0.00360 (0.02663)  --> STEP: 385/639 -- GLOBAL_STEP: 29825 | > loss: -0.46301 (-0.47674) | > log_mle: -0.58025 (-0.59695) | > loss_dur: 0.11724 (0.12020) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.87226 (8.94312) | > current_lr: 0.00037 | > step_time: 0.66790 (0.64850) | > loader_time: 0.00390 (0.02514)  --> STEP: 410/639 -- GLOBAL_STEP: 29850 | > loss: -0.41978 (-0.47520) | > log_mle: -0.55596 (-0.59556) | > loss_dur: 0.13618 (0.12035) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.85280 (8.90301) | > current_lr: 0.00037 | > step_time: 0.76810 (0.65509) | > loader_time: 0.00390 (0.02383)  --> STEP: 435/639 -- GLOBAL_STEP: 29875 | > loss: -0.45392 (-0.47349) | > log_mle: -0.56509 (-0.59373) | > loss_dur: 0.11117 (0.12024) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.52856 (8.74577) | > current_lr: 0.00037 | > step_time: 0.70480 (0.66198) | > loader_time: 0.00360 (0.02267)  --> STEP: 460/639 -- GLOBAL_STEP: 29900 | > loss: -0.41446 (-0.47183) | > log_mle: -0.53125 (-0.59199) | > loss_dur: 0.11679 (0.12016) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.76259 (8.65290) | > current_lr: 0.00037 | > step_time: 0.93640 (0.66980) | > loader_time: 0.00400 (0.02164)  --> STEP: 485/639 -- GLOBAL_STEP: 29925 | > loss: -0.44915 (-0.47052) | > log_mle: -0.56613 (-0.59076) | > loss_dur: 0.11698 (0.12024) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.53237 (8.61868) | > current_lr: 0.00037 | > step_time: 0.89570 (0.67903) | > loader_time: 0.00390 (0.02073)  --> STEP: 510/639 -- GLOBAL_STEP: 29950 | > loss: -0.44554 (-0.46954) | > log_mle: -0.56802 (-0.58970) | > loss_dur: 0.12248 (0.12016) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.02764 (8.56453) | > current_lr: 0.00037 | > step_time: 0.87180 (0.68899) | > loader_time: 0.00390 (0.01991)  --> STEP: 535/639 -- GLOBAL_STEP: 29975 | > loss: -0.43693 (-0.46840) | > log_mle: -0.56118 (-0.58852) | > loss_dur: 0.12425 (0.12012) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.72093 (8.46484) | > current_lr: 0.00037 | > step_time: 0.91230 (0.70018) | > loader_time: 0.00380 (0.01916)  --> STEP: 560/639 -- GLOBAL_STEP: 30000 | > loss: -0.44043 (-0.46723) | > log_mle: -0.56484 (-0.58740) | > loss_dur: 0.12440 (0.12017) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.81350 (8.41447) | > current_lr: 0.00037 | > step_time: 0.88650 (0.71237) | > loader_time: 0.00410 (0.01850) > CHECKPOINT : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/checkpoint_30000.pth.tar warning: audio amplitude out of range, auto clipped.  --> STEP: 585/639 -- GLOBAL_STEP: 30025 | > loss: -0.45458 (-0.46622) | > log_mle: -0.57055 (-0.58640) | > loss_dur: 0.11598 (0.12018) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.45915 (8.32085) | > current_lr: 0.00037 | > step_time: 1.09210 (0.72497) | > loader_time: 0.00400 (0.01795)  --> STEP: 610/639 -- GLOBAL_STEP: 30050 | > loss: -0.42440 (-0.46524) | > log_mle: -0.54062 (-0.58549) | > loss_dur: 0.11622 (0.12025) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.98826 (8.26311) | > current_lr: 0.00036 | > step_time: 1.11460 (0.73898) | > loader_time: 0.00410 (0.01739)  --> STEP: 635/639 -- GLOBAL_STEP: 30075 | > loss: -0.44055 (-0.46455) | > log_mle: -0.56491 (-0.58496) | > loss_dur: 0.12437 (0.12042) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.62161 (8.25145) | > current_lr: 0.00036 | > step_time: 1.14480 (0.75376) | > loader_time: 0.00420 (0.01688) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02551 (-0.00249) | > avg_loss: -0.44608 (+0.00041) | > avg_log_mle: -0.57382 (+0.00340) | > avg_loss_dur: 0.12774 (-0.00299)  > EPOCH: 47/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 20:48:39)   --> STEP: 20/639 -- GLOBAL_STEP: 30100 | > loss: -0.59240 (-0.51781) | > log_mle: -0.74988 (-0.64442) | > loss_dur: 0.15748 (0.12661) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 21.42043 (11.18653) | > current_lr: 0.00036 | > step_time: 0.34740 (0.30770) | > loader_time: 0.00310 (0.00413)  --> STEP: 45/639 -- GLOBAL_STEP: 30125 | > loss: -0.44995 (-0.49553) | > log_mle: -0.55473 (-0.61865) | > loss_dur: 0.10478 (0.12312) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.39471 (9.82207) | > current_lr: 0.00036 | > step_time: 0.42000 (0.33971) | > loader_time: 0.00290 (0.00342)  --> STEP: 70/639 -- GLOBAL_STEP: 30150 | > loss: -0.49033 (-0.48650) | > log_mle: -0.61810 (-0.60962) | > loss_dur: 0.12777 (0.12312) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.08155 (9.03863) | > current_lr: 0.00036 | > step_time: 0.39520 (0.36381) | > loader_time: 0.00280 (0.00326)  --> STEP: 95/639 -- GLOBAL_STEP: 30175 | > loss: -0.56469 (-0.48414) | > log_mle: -0.67055 (-0.60515) | > loss_dur: 0.10586 (0.12101) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.85862 (8.69535) | > current_lr: 0.00036 | > step_time: 0.40350 (0.37974) | > loader_time: 0.00320 (0.00319)  --> STEP: 120/639 -- GLOBAL_STEP: 30200 | > loss: -0.51993 (-0.48207) | > log_mle: -0.65721 (-0.60373) | > loss_dur: 0.13729 (0.12166) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.84194 (8.71007) | > current_lr: 0.00036 | > step_time: 0.54440 (0.39889) | > loader_time: 0.00330 (0.00317)  --> STEP: 145/639 -- GLOBAL_STEP: 30225 | > loss: -0.49216 (-0.48307) | > log_mle: -0.61683 (-0.60468) | > loss_dur: 0.12468 (0.12161) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.98711 (9.06874) | > current_lr: 0.00036 | > step_time: 0.47470 (0.41608) | > loader_time: 0.00290 (0.00316)  --> STEP: 170/639 -- GLOBAL_STEP: 30250 | > loss: -0.48892 (-0.48300) | > log_mle: -0.60349 (-0.60399) | > loss_dur: 0.11457 (0.12099) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.15677 (9.19107) | > current_lr: 0.00036 | > step_time: 0.56080 (0.43144) | > loader_time: 0.00310 (0.00316)  --> STEP: 195/639 -- GLOBAL_STEP: 30275 | > loss: -0.44292 (-0.48265) | > log_mle: -0.54363 (-0.60312) | > loss_dur: 0.10070 (0.12047) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.67163 (9.18104) | > current_lr: 0.00036 | > step_time: 0.49880 (0.44594) | > loader_time: 0.00310 (0.00318)  --> STEP: 220/639 -- GLOBAL_STEP: 30300 | > loss: -0.49430 (-0.48111) | > log_mle: -0.60030 (-0.60123) | > loss_dur: 0.10600 (0.12012) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.61688 (9.10166) | > current_lr: 0.00036 | > step_time: 0.64770 (0.46218) | > loader_time: 0.00330 (0.00319)  --> STEP: 245/639 -- GLOBAL_STEP: 30325 | > loss: -0.47341 (-0.48275) | > log_mle: -0.58624 (-0.60246) | > loss_dur: 0.11283 (0.11971) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.93267 (9.17468) | > current_lr: 0.00036 | > step_time: 0.53420 (0.47687) | > loader_time: 0.00290 (0.00322)  --> STEP: 270/639 -- GLOBAL_STEP: 30350 | > loss: -0.52627 (-0.48269) | > log_mle: -0.64989 (-0.60266) | > loss_dur: 0.12362 (0.11997) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.39803 (9.27184) | > current_lr: 0.00036 | > step_time: 0.80700 (0.49370) | > loader_time: 0.00420 (0.00324)  --> STEP: 295/639 -- GLOBAL_STEP: 30375 | > loss: -0.43431 (-0.48111) | > log_mle: -0.54275 (-0.60115) | > loss_dur: 0.10844 (0.12005) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.67770 (9.10303) | > current_lr: 0.00036 | > step_time: 0.66090 (0.50959) | > loader_time: 0.00340 (0.00326)  --> STEP: 320/639 -- GLOBAL_STEP: 30400 | > loss: -0.45786 (-0.47988) | > log_mle: -0.58972 (-0.60016) | > loss_dur: 0.13186 (0.12027) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.64298 (9.08555) | > current_lr: 0.00036 | > step_time: 0.73150 (0.52428) | > loader_time: 0.00360 (0.00329)  --> STEP: 345/639 -- GLOBAL_STEP: 30425 | > loss: -0.46014 (-0.47831) | > log_mle: -0.57614 (-0.59874) | > loss_dur: 0.11600 (0.12043) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.13533 (9.02431) | > current_lr: 0.00036 | > step_time: 0.67770 (0.53923) | > loader_time: 0.00350 (0.00332)  --> STEP: 370/639 -- GLOBAL_STEP: 30450 | > loss: -0.47757 (-0.47696) | > log_mle: -0.59555 (-0.59701) | > loss_dur: 0.11798 (0.12005) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.09752 (8.88691) | > current_lr: 0.00036 | > step_time: 0.89850 (0.55158) | > loader_time: 0.00410 (0.00334)  --> STEP: 395/639 -- GLOBAL_STEP: 30475 | > loss: -0.44221 (-0.47543) | > log_mle: -0.56442 (-0.59543) | > loss_dur: 0.12221 (0.12000) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.50315 (8.77131) | > current_lr: 0.00036 | > step_time: 0.70850 (0.56368) | > loader_time: 0.00360 (0.00336)  --> STEP: 420/639 -- GLOBAL_STEP: 30500 | > loss: -0.44391 (-0.47386) | > log_mle: -0.55944 (-0.59396) | > loss_dur: 0.11553 (0.12010) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.49618 (8.67467) | > current_lr: 0.00036 | > step_time: 0.71510 (0.57655) | > loader_time: 0.00330 (0.00338)  --> STEP: 445/639 -- GLOBAL_STEP: 30525 | > loss: -0.45602 (-0.47282) | > log_mle: -0.56541 (-0.59276) | > loss_dur: 0.10939 (0.11994) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.33057 (8.62951) | > current_lr: 0.00036 | > step_time: 0.72420 (0.58833) | > loader_time: 0.00350 (0.00340)  --> STEP: 470/639 -- GLOBAL_STEP: 30550 | > loss: -0.45985 (-0.47127) | > log_mle: -0.57781 (-0.59116) | > loss_dur: 0.11796 (0.11989) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.27753 (8.53292) | > current_lr: 0.00036 | > step_time: 0.75470 (0.59964) | > loader_time: 0.00390 (0.00342)  --> STEP: 495/639 -- GLOBAL_STEP: 30575 | > loss: -0.40236 (-0.47006) | > log_mle: -0.51218 (-0.58995) | > loss_dur: 0.10982 (0.11989) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.33443 (8.44882) | > current_lr: 0.00036 | > step_time: 0.73920 (0.61262) | > loader_time: 0.00370 (0.00344)  --> STEP: 520/639 -- GLOBAL_STEP: 30600 | > loss: -0.45650 (-0.46927) | > log_mle: -0.56893 (-0.58906) | > loss_dur: 0.11243 (0.11979) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.47784 (8.37871) | > current_lr: 0.00036 | > step_time: 0.81040 (0.62552) | > loader_time: 0.00380 (0.00347)  --> STEP: 545/639 -- GLOBAL_STEP: 30625 | > loss: -0.47374 (-0.46821) | > log_mle: -0.60185 (-0.58812) | > loss_dur: 0.12811 (0.11991) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.58718 (8.32485) | > current_lr: 0.00036 | > step_time: 0.93800 (0.64047) | > loader_time: 0.00390 (0.00350)  --> STEP: 570/639 -- GLOBAL_STEP: 30650 | > loss: -0.44277 (-0.46696) | > log_mle: -0.55017 (-0.58686) | > loss_dur: 0.10739 (0.11991) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.17728 (8.21388) | > current_lr: 0.00036 | > step_time: 0.88160 (0.65430) | > loader_time: 0.00390 (0.00353)  --> STEP: 595/639 -- GLOBAL_STEP: 30675 | > loss: -0.42403 (-0.46618) | > log_mle: -0.53930 (-0.58618) | > loss_dur: 0.11527 (0.12000) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.10323 (8.19173) | > current_lr: 0.00036 | > step_time: 0.98380 (0.67013) | > loader_time: 0.00390 (0.00356)  --> STEP: 620/639 -- GLOBAL_STEP: 30700 | > loss: -0.43622 (-0.46530) | > log_mle: -0.55264 (-0.58532) | > loss_dur: 0.11642 (0.12002) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.17331 (8.13800) | > current_lr: 0.00036 | > step_time: 1.17690 (0.68615) | > loader_time: 0.00470 (0.00360) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02882 (+0.00331) | > avg_loss: -0.44576 (+0.00032) | > avg_log_mle: -0.57252 (+0.00130) | > avg_loss_dur: 0.12676 (-0.00098)  > EPOCH: 48/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 20:56:21)   --> STEP: 5/639 -- GLOBAL_STEP: 30725 | > loss: -0.51122 (-0.51512) | > log_mle: -0.66744 (-0.65077) | > loss_dur: 0.15622 (0.13565) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 21.24180 (11.77296) | > current_lr: 0.00036 | > step_time: 0.32720 (0.30022) | > loader_time: 0.00250 (0.01443)  --> STEP: 30/639 -- GLOBAL_STEP: 30750 | > loss: -0.50109 (-0.50174) | > log_mle: -0.62805 (-0.62792) | > loss_dur: 0.12695 (0.12619) | > amp_scaler: 2048.00000 (3003.73333) | > grad_norm: 12.12872 (10.15846) | > current_lr: 0.00036 | > step_time: 0.32870 (0.33056) | > loader_time: 0.00260 (0.00460)  --> STEP: 55/639 -- GLOBAL_STEP: 30775 | > loss: -0.45449 (-0.49412) | > log_mle: -0.55417 (-0.61711) | > loss_dur: 0.09968 (0.12299) | > amp_scaler: 2048.00000 (2569.30909) | > grad_norm: 3.89768 (9.11809) | > current_lr: 0.00036 | > step_time: 0.35810 (0.35327) | > loader_time: 0.00280 (0.00377)  --> STEP: 80/639 -- GLOBAL_STEP: 30800 | > loss: -0.46173 (-0.48773) | > log_mle: -0.59928 (-0.60932) | > loss_dur: 0.13755 (0.12159) | > amp_scaler: 2048.00000 (2406.40000) | > grad_norm: 9.66654 (8.85443) | > current_lr: 0.00036 | > step_time: 0.39390 (0.37147) | > loader_time: 0.00280 (0.00350)  --> STEP: 105/639 -- GLOBAL_STEP: 30825 | > loss: -0.43389 (-0.48634) | > log_mle: -0.55947 (-0.60739) | > loss_dur: 0.12558 (0.12105) | > amp_scaler: 2048.00000 (2321.06667) | > grad_norm: 5.37510 (8.78580) | > current_lr: 0.00036 | > step_time: 0.40530 (0.38768) | > loader_time: 0.00300 (0.00337)  --> STEP: 130/639 -- GLOBAL_STEP: 30850 | > loss: -0.46170 (-0.48562) | > log_mle: -0.59277 (-0.60655) | > loss_dur: 0.13107 (0.12094) | > amp_scaler: 2048.00000 (2268.55385) | > grad_norm: 7.12470 (8.75963) | > current_lr: 0.00036 | > step_time: 0.49060 (0.40353) | > loader_time: 0.00310 (0.00329)  --> STEP: 155/639 -- GLOBAL_STEP: 30875 | > loss: -0.45979 (-0.48587) | > log_mle: -0.58349 (-0.60640) | > loss_dur: 0.12370 (0.12052) | > amp_scaler: 2048.00000 (2232.98065) | > grad_norm: 7.49989 (8.88481) | > current_lr: 0.00036 | > step_time: 0.54640 (0.42110) | > loader_time: 0.00320 (0.00327)  --> STEP: 180/639 -- GLOBAL_STEP: 30900 | > loss: -0.52159 (-0.48770) | > log_mle: -0.64309 (-0.60820) | > loss_dur: 0.12150 (0.12050) | > amp_scaler: 2048.00000 (2207.28889) | > grad_norm: 13.79066 (9.17583) | > current_lr: 0.00036 | > step_time: 0.57030 (0.43708) | > loader_time: 0.00390 (0.00324)  --> STEP: 205/639 -- GLOBAL_STEP: 30925 | > loss: -0.45444 (-0.48497) | > log_mle: -0.56625 (-0.60487) | > loss_dur: 0.11181 (0.11990) | > amp_scaler: 2048.00000 (2187.86341) | > grad_norm: 6.24048 (9.05149) | > current_lr: 0.00036 | > step_time: 0.60650 (0.45306) | > loader_time: 0.00320 (0.00323)  --> STEP: 230/639 -- GLOBAL_STEP: 30950 | > loss: -0.47626 (-0.48554) | > log_mle: -0.59852 (-0.60534) | > loss_dur: 0.12226 (0.11980) | > amp_scaler: 2048.00000 (2172.66087) | > grad_norm: 15.76801 (9.34047) | > current_lr: 0.00036 | > step_time: 0.55640 (0.47118) | > loader_time: 0.00330 (0.00325)  --> STEP: 255/639 -- GLOBAL_STEP: 30975 | > loss: -0.44747 (-0.48487) | > log_mle: -0.56961 (-0.60448) | > loss_dur: 0.12214 (0.11960) | > amp_scaler: 2048.00000 (2160.43922) | > grad_norm: 9.46065 (9.38258) | > current_lr: 0.00036 | > step_time: 0.63620 (0.48460) | > loader_time: 0.00330 (0.00326)  --> STEP: 280/639 -- GLOBAL_STEP: 31000 | > loss: -0.45863 (-0.48528) | > log_mle: -0.59838 (-0.60523) | > loss_dur: 0.13975 (0.11995) | > amp_scaler: 2048.00000 (2150.40000) | > grad_norm: 8.85248 (9.50408) | > current_lr: 0.00036 | > step_time: 0.59620 (0.50173) | > loader_time: 0.00350 (0.00328)  --> STEP: 305/639 -- GLOBAL_STEP: 31025 | > loss: -0.44759 (-0.48400) | > log_mle: -0.57547 (-0.60393) | > loss_dur: 0.12788 (0.11993) | > amp_scaler: 2048.00000 (2142.00656) | > grad_norm: 9.98461 (9.43705) | > current_lr: 0.00036 | > step_time: 0.66810 (0.51588) | > loader_time: 0.00360 (0.00331)  --> STEP: 330/639 -- GLOBAL_STEP: 31050 | > loss: -0.46874 (-0.48190) | > log_mle: -0.59944 (-0.60225) | > loss_dur: 0.13070 (0.12036) | > amp_scaler: 2048.00000 (2134.88485) | > grad_norm: 11.50868 (9.33586) | > current_lr: 0.00036 | > step_time: 0.77330 (0.53043) | > loader_time: 0.00370 (0.00334)  --> STEP: 355/639 -- GLOBAL_STEP: 31075 | > loss: -0.41668 (-0.48053) | > log_mle: -0.52796 (-0.60061) | > loss_dur: 0.11128 (0.12008) | > amp_scaler: 2048.00000 (2128.76620) | > grad_norm: 2.09118 (9.24123) | > current_lr: 0.00036 | > step_time: 0.59740 (0.54441) | > loader_time: 0.00320 (0.00337)  --> STEP: 380/639 -- GLOBAL_STEP: 31100 | > loss: -0.45985 (-0.47881) | > log_mle: -0.56324 (-0.59863) | > loss_dur: 0.10339 (0.11983) | > amp_scaler: 2048.00000 (2123.45263) | > grad_norm: 6.08973 (9.06521) | > current_lr: 0.00036 | > step_time: 0.61200 (0.55785) | > loader_time: 0.00360 (0.00340)  --> STEP: 405/639 -- GLOBAL_STEP: 31125 | > loss: -0.45305 (-0.47767) | > log_mle: -0.56851 (-0.59761) | > loss_dur: 0.11546 (0.11993) | > amp_scaler: 2048.00000 (2118.79506) | > grad_norm: 6.00021 (9.01631) | > current_lr: 0.00036 | > step_time: 0.67600 (0.57011) | > loader_time: 0.00360 (0.00342)  --> STEP: 430/639 -- GLOBAL_STEP: 31150 | > loss: -0.42102 (-0.47604) | > log_mle: -0.52876 (-0.59595) | > loss_dur: 0.10774 (0.11991) | > amp_scaler: 2048.00000 (2114.67907) | > grad_norm: 3.98683 (8.91346) | > current_lr: 0.00036 | > step_time: 0.77320 (0.58358) | > loader_time: 0.00380 (0.00345)  --> STEP: 455/639 -- GLOBAL_STEP: 31175 | > loss: -0.43301 (-0.47440) | > log_mle: -0.55482 (-0.59417) | > loss_dur: 0.12181 (0.11978) | > amp_scaler: 2048.00000 (2111.01538) | > grad_norm: 6.24743 (8.78010) | > current_lr: 0.00036 | > step_time: 0.72570 (0.59477) | > loader_time: 0.00360 (0.00347)  --> STEP: 480/639 -- GLOBAL_STEP: 31200 | > loss: -0.39329 (-0.47290) | > log_mle: -0.51097 (-0.59281) | > loss_dur: 0.11768 (0.11992) | > amp_scaler: 2048.00000 (2107.73333) | > grad_norm: 3.01884 (8.70826) | > current_lr: 0.00036 | > step_time: 0.87180 (0.60771) | > loader_time: 0.00400 (0.00349)  --> STEP: 505/639 -- GLOBAL_STEP: 31225 | > loss: -0.46482 (-0.47186) | > log_mle: -0.57731 (-0.59162) | > loss_dur: 0.11249 (0.11976) | > amp_scaler: 2048.00000 (2104.77624) | > grad_norm: 10.78179 (8.63448) | > current_lr: 0.00036 | > step_time: 0.89910 (0.62054) | > loader_time: 0.00410 (0.00352)  --> STEP: 530/639 -- GLOBAL_STEP: 31250 | > loss: -0.44157 (-0.47078) | > log_mle: -0.56927 (-0.59061) | > loss_dur: 0.12770 (0.11984) | > amp_scaler: 2048.00000 (2102.09811) | > grad_norm: 7.54855 (8.59436) | > current_lr: 0.00036 | > step_time: 0.96090 (0.63532) | > loader_time: 0.00410 (0.00355)  --> STEP: 555/639 -- GLOBAL_STEP: 31275 | > loss: -0.44023 (-0.46964) | > log_mle: -0.55208 (-0.58936) | > loss_dur: 0.11186 (0.11972) | > amp_scaler: 2048.00000 (2099.66126) | > grad_norm: 7.12458 (8.51300) | > current_lr: 0.00036 | > step_time: 0.82930 (0.64919) | > loader_time: 0.00390 (0.00358)  --> STEP: 580/639 -- GLOBAL_STEP: 31300 | > loss: -0.45836 (-0.46870) | > log_mle: -0.57004 (-0.58838) | > loss_dur: 0.11168 (0.11968) | > amp_scaler: 2048.00000 (2097.43448) | > grad_norm: 6.81877 (8.48929) | > current_lr: 0.00036 | > step_time: 0.98300 (0.66437) | > loader_time: 0.00440 (0.00362)  --> STEP: 605/639 -- GLOBAL_STEP: 31325 | > loss: -0.42993 (-0.46766) | > log_mle: -0.53825 (-0.58747) | > loss_dur: 0.10831 (0.11981) | > amp_scaler: 2048.00000 (2095.39174) | > grad_norm: 3.35668 (8.43576) | > current_lr: 0.00036 | > step_time: 1.06910 (0.68091) | > loader_time: 0.00450 (0.00366)  --> STEP: 630/639 -- GLOBAL_STEP: 31350 | > loss: -0.43872 (-0.46688) | > log_mle: -0.57094 (-0.58673) | > loss_dur: 0.13222 (0.11985) | > amp_scaler: 2048.00000 (2093.51111) | > grad_norm: 6.83140 (8.41138) | > current_lr: 0.00036 | > step_time: 1.21570 (0.69818) | > loader_time: 0.00480 (0.00369) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03585 (+0.00703) | > avg_loss: -0.44398 (+0.00179) | > avg_log_mle: -0.57178 (+0.00074) | > avg_loss_dur: 0.12781 (+0.00105)  > EPOCH: 49/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 21:04:05)   --> STEP: 15/639 -- GLOBAL_STEP: 31375 | > loss: -0.61724 (-0.51683) | > log_mle: -0.75301 (-0.64154) | > loss_dur: 0.13578 (0.12471) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 19.01622 (10.06725) | > current_lr: 0.00036 | > step_time: 0.39090 (0.32297) | > loader_time: 0.00260 (0.00603)  --> STEP: 40/639 -- GLOBAL_STEP: 31400 | > loss: -0.48335 (-0.49812) | > log_mle: -0.60096 (-0.62097) | > loss_dur: 0.11761 (0.12285) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.23638 (9.18578) | > current_lr: 0.00036 | > step_time: 0.36680 (0.35791) | > loader_time: 0.00260 (0.00395)  --> STEP: 65/639 -- GLOBAL_STEP: 31425 | > loss: -0.48489 (-0.49141) | > log_mle: -0.58661 (-0.61426) | > loss_dur: 0.10172 (0.12284) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.32829 (9.28896) | > current_lr: 0.00036 | > step_time: 0.38730 (0.38144) | > loader_time: 0.00280 (0.00356)  --> STEP: 90/639 -- GLOBAL_STEP: 31450 | > loss: -0.43061 (-0.48528) | > log_mle: -0.54657 (-0.60601) | > loss_dur: 0.11596 (0.12072) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.88643 (8.58864) | > current_lr: 0.00036 | > step_time: 0.61190 (0.39676) | > loader_time: 0.00340 (0.00341)  --> STEP: 115/639 -- GLOBAL_STEP: 31475 | > loss: -0.39184 (-0.48631) | > log_mle: -0.51476 (-0.60702) | > loss_dur: 0.12291 (0.12070) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.98195 (9.19044) | > current_lr: 0.00036 | > step_time: 0.40470 (0.41237) | > loader_time: 0.00320 (0.00333)  --> STEP: 140/639 -- GLOBAL_STEP: 31500 | > loss: -0.45311 (-0.48646) | > log_mle: -0.59680 (-0.60729) | > loss_dur: 0.14369 (0.12084) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.66320 (9.45992) | > current_lr: 0.00036 | > step_time: 0.56450 (0.42859) | > loader_time: 0.00320 (0.00330)  --> STEP: 165/639 -- GLOBAL_STEP: 31525 | > loss: -0.46942 (-0.48613) | > log_mle: -0.60430 (-0.60655) | > loss_dur: 0.13488 (0.12043) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.57636 (9.56571) | > current_lr: 0.00036 | > step_time: 0.49010 (0.44231) | > loader_time: 0.00300 (0.00329)  --> STEP: 190/639 -- GLOBAL_STEP: 31550 | > loss: -0.41739 (-0.48635) | > log_mle: -0.53305 (-0.60643) | > loss_dur: 0.11566 (0.12008) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.29145 (9.51076) | > current_lr: 0.00036 | > step_time: 0.52290 (0.46073) | > loader_time: 0.00340 (0.00330)  --> STEP: 215/639 -- GLOBAL_STEP: 31575 | > loss: -0.53357 (-0.48470) | > log_mle: -0.64048 (-0.60473) | > loss_dur: 0.10691 (0.12003) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.44075 (9.48577) | > current_lr: 0.00036 | > step_time: 0.58840 (0.47638) | > loader_time: 0.00350 (0.00331)  --> STEP: 240/639 -- GLOBAL_STEP: 31600 | > loss: -0.52540 (-0.48593) | > log_mle: -0.64889 (-0.60543) | > loss_dur: 0.12349 (0.11950) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 16.94302 (9.62982) | > current_lr: 0.00036 | > step_time: 0.57890 (0.48905) | > loader_time: 0.00330 (0.00332)  --> STEP: 265/639 -- GLOBAL_STEP: 31625 | > loss: -0.53953 (-0.48524) | > log_mle: -0.67196 (-0.60497) | > loss_dur: 0.13243 (0.11973) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.80932 (9.72461) | > current_lr: 0.00036 | > step_time: 0.66710 (0.50328) | > loader_time: 0.00340 (0.00335)  --> STEP: 290/639 -- GLOBAL_STEP: 31650 | > loss: -0.45570 (-0.48391) | > log_mle: -0.57240 (-0.60369) | > loss_dur: 0.11670 (0.11978) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.26094 (9.62449) | > current_lr: 0.00036 | > step_time: 0.68290 (0.51887) | > loader_time: 0.00360 (0.00337)  --> STEP: 315/639 -- GLOBAL_STEP: 31675 | > loss: -0.43730 (-0.48267) | > log_mle: -0.55024 (-0.60246) | > loss_dur: 0.11294 (0.11980) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.64120 (9.61849) | > current_lr: 0.00036 | > step_time: 0.64990 (0.53261) | > loader_time: 0.00350 (0.00339)  --> STEP: 340/639 -- GLOBAL_STEP: 31700 | > loss: -0.46113 (-0.48122) | > log_mle: -0.57406 (-0.60140) | > loss_dur: 0.11293 (0.12018) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.71073 (9.61124) | > current_lr: 0.00036 | > step_time: 0.61720 (0.54914) | > loader_time: 0.00340 (0.00341)  --> STEP: 365/639 -- GLOBAL_STEP: 31725 | > loss: -0.45827 (-0.47955) | > log_mle: -0.57923 (-0.59937) | > loss_dur: 0.12096 (0.11982) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.84712 (9.48046) | > current_lr: 0.00036 | > step_time: 0.80990 (0.56047) | > loader_time: 0.00380 (0.00342)  --> STEP: 390/639 -- GLOBAL_STEP: 31750 | > loss: -0.45299 (-0.47793) | > log_mle: -0.57874 (-0.59760) | > loss_dur: 0.12575 (0.11967) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.59520 (9.37198) | > current_lr: 0.00036 | > step_time: 0.76200 (0.57316) | > loader_time: 0.00370 (0.00344)  --> STEP: 415/639 -- GLOBAL_STEP: 31775 | > loss: -0.42955 (-0.47649) | > log_mle: -0.55590 (-0.59629) | > loss_dur: 0.12635 (0.11980) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.42543 (9.32598) | > current_lr: 0.00035 | > step_time: 0.75460 (0.58536) | > loader_time: 0.00340 (0.00346)  --> STEP: 440/639 -- GLOBAL_STEP: 31800 | > loss: -0.41440 (-0.47527) | > log_mle: -0.53792 (-0.59503) | > loss_dur: 0.12352 (0.11976) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.72849 (9.30946) | > current_lr: 0.00035 | > step_time: 0.82140 (0.59792) | > loader_time: 0.00380 (0.00347)  --> STEP: 465/639 -- GLOBAL_STEP: 31825 | > loss: -0.43071 (-0.47367) | > log_mle: -0.54424 (-0.59333) | > loss_dur: 0.11354 (0.11966) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.88518 (9.18683) | > current_lr: 0.00035 | > step_time: 0.85750 (0.60934) | > loader_time: 0.00390 (0.00349)  --> STEP: 490/639 -- GLOBAL_STEP: 31850 | > loss: -0.42238 (-0.47232) | > log_mle: -0.52436 (-0.59206) | > loss_dur: 0.10198 (0.11974) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.44360 (9.22299) | > current_lr: 0.00035 | > step_time: 0.77190 (0.62197) | > loader_time: 0.00490 (0.00351)  --> STEP: 515/639 -- GLOBAL_STEP: 31875 | > loss: -0.44234 (-0.47141) | > log_mle: -0.55771 (-0.59112) | > loss_dur: 0.11538 (0.11971) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.36108 (9.18245) | > current_lr: 0.00035 | > step_time: 1.06520 (0.63486) | > loader_time: 0.00420 (0.00354)  --> STEP: 540/639 -- GLOBAL_STEP: 31900 | > loss: -0.45615 (-0.47018) | > log_mle: -0.57423 (-0.58993) | > loss_dur: 0.11808 (0.11975) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.89130 (9.08258) | > current_lr: 0.00035 | > step_time: 0.94310 (0.64936) | > loader_time: 0.00400 (0.00357)  --> STEP: 565/639 -- GLOBAL_STEP: 31925 | > loss: -0.41986 (-0.46890) | > log_mle: -0.53269 (-0.58864) | > loss_dur: 0.11283 (0.11975) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.77310 (8.99687) | > current_lr: 0.00035 | > step_time: 0.99880 (0.66321) | > loader_time: 0.00430 (0.00361)  --> STEP: 590/639 -- GLOBAL_STEP: 31950 | > loss: -0.44724 (-0.46814) | > log_mle: -0.56564 (-0.58795) | > loss_dur: 0.11840 (0.11981) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.06822 (8.91730) | > current_lr: 0.00035 | > step_time: 0.95480 (0.67885) | > loader_time: 0.00680 (0.00365)  --> STEP: 615/639 -- GLOBAL_STEP: 31975 | > loss: -0.44804 (-0.46720) | > log_mle: -0.57045 (-0.58700) | > loss_dur: 0.12240 (0.11980) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.28960 (8.85780) | > current_lr: 0.00035 | > step_time: 1.04960 (0.69459) | > loader_time: 0.00440 (0.00370) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02608 (-0.00977) | > avg_loss: -0.44446 (-0.00048) | > avg_log_mle: -0.57204 (-0.00026) | > avg_loss_dur: 0.12759 (-0.00022)  > EPOCH: 50/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 21:11:55)   --> STEP: 0/639 -- GLOBAL_STEP: 32000 | > loss: -0.50627 (-0.50627) | > log_mle: -0.68306 (-0.68306) | > loss_dur: 0.17679 (0.17679) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 25.17137 (25.17137) | > current_lr: 0.00035 | > step_time: 0.81420 (0.81416) | > loader_time: 0.40920 (0.40916)  --> STEP: 25/639 -- GLOBAL_STEP: 32025 | > loss: -0.48540 (-0.50620) | > log_mle: -0.63011 (-0.63190) | > loss_dur: 0.14472 (0.12569) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.31295 (14.83652) | > current_lr: 0.00035 | > step_time: 0.38890 (0.32969) | > loader_time: 0.00280 (0.00256)  --> STEP: 50/639 -- GLOBAL_STEP: 32050 | > loss: -0.45371 (-0.49395) | > log_mle: -0.56633 (-0.61648) | > loss_dur: 0.11262 (0.12253) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.46187 (11.81661) | > current_lr: 0.00035 | > step_time: 0.34280 (0.35462) | > loader_time: 0.00280 (0.00264)  --> STEP: 75/639 -- GLOBAL_STEP: 32075 | > loss: -0.49233 (-0.48773) | > log_mle: -0.60362 (-0.60941) | > loss_dur: 0.11129 (0.12168) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.83284 (10.55124) | > current_lr: 0.00035 | > step_time: 0.43110 (0.37760) | > loader_time: 0.00300 (0.00273)  --> STEP: 100/639 -- GLOBAL_STEP: 32100 | > loss: -0.49272 (-0.48702) | > log_mle: -0.61034 (-0.60734) | > loss_dur: 0.11762 (0.12032) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.04915 (9.89780) | > current_lr: 0.00035 | > step_time: 0.51030 (0.39107) | > loader_time: 0.00320 (0.00281)  --> STEP: 125/639 -- GLOBAL_STEP: 32125 | > loss: -0.48507 (-0.48623) | > log_mle: -0.60198 (-0.60685) | > loss_dur: 0.11691 (0.12061) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.85660 (9.76924) | > current_lr: 0.00035 | > step_time: 0.41820 (0.40606) | > loader_time: 0.00280 (0.00288)  --> STEP: 150/639 -- GLOBAL_STEP: 32150 | > loss: -0.44904 (-0.48621) | > log_mle: -0.57249 (-0.60673) | > loss_dur: 0.12345 (0.12052) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.22292 (9.70205) | > current_lr: 0.00035 | > step_time: 0.43930 (0.42115) | > loader_time: 0.00270 (0.00291)  --> STEP: 175/639 -- GLOBAL_STEP: 32175 | > loss: -0.48204 (-0.48748) | > log_mle: -0.58966 (-0.60768) | > loss_dur: 0.10762 (0.12020) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.35417 (9.85213) | > current_lr: 0.00035 | > step_time: 0.55620 (0.43716) | > loader_time: 0.00350 (0.00295)  --> STEP: 200/639 -- GLOBAL_STEP: 32200 | > loss: -0.45181 (-0.48591) | > log_mle: -0.57231 (-0.60565) | > loss_dur: 0.12050 (0.11973) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.36805 (9.67978) | > current_lr: 0.00035 | > step_time: 0.64960 (0.45259) | > loader_time: 0.00320 (0.00299)  --> STEP: 225/639 -- GLOBAL_STEP: 32225 | > loss: -0.44618 (-0.48447) | > log_mle: -0.55010 (-0.60387) | > loss_dur: 0.10392 (0.11940) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.03413 (9.54303) | > current_lr: 0.00035 | > step_time: 0.53400 (0.46889) | > loader_time: 0.00310 (0.00304)  --> STEP: 250/639 -- GLOBAL_STEP: 32250 | > loss: -0.46064 (-0.48591) | > log_mle: -0.55603 (-0.60516) | > loss_dur: 0.09540 (0.11925) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.01833 (9.68032) | > current_lr: 0.00035 | > step_time: 0.61200 (0.48410) | > loader_time: 0.00340 (0.00307)  --> STEP: 275/639 -- GLOBAL_STEP: 32275 | > loss: -0.47894 (-0.48636) | > log_mle: -0.59424 (-0.60588) | > loss_dur: 0.11530 (0.11953) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.80998 (9.87411) | > current_lr: 0.00035 | > step_time: 0.88050 (0.50177) | > loader_time: 0.00410 (0.00312)  --> STEP: 300/639 -- GLOBAL_STEP: 32300 | > loss: -0.46914 (-0.48463) | > log_mle: -0.59120 (-0.60426) | > loss_dur: 0.12206 (0.11963) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.08700 (9.71752) | > current_lr: 0.00035 | > step_time: 0.67780 (0.51539) | > loader_time: 0.00380 (0.00315)  --> STEP: 325/639 -- GLOBAL_STEP: 32325 | > loss: -0.46880 (-0.48314) | > log_mle: -0.59130 (-0.60306) | > loss_dur: 0.12250 (0.11992) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.05874 (9.74925) | > current_lr: 0.00035 | > step_time: 0.60510 (0.53191) | > loader_time: 0.00340 (0.00319)  --> STEP: 350/639 -- GLOBAL_STEP: 32350 | > loss: -0.49064 (-0.48162) | > log_mle: -0.61789 (-0.60153) | > loss_dur: 0.12725 (0.11992) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.25642 (9.63844) | > current_lr: 0.00035 | > step_time: 0.84580 (0.54593) | > loader_time: 0.00400 (0.00322)  --> STEP: 375/639 -- GLOBAL_STEP: 32375 | > loss: -0.44157 (-0.47970) | > log_mle: -0.55259 (-0.59945) | > loss_dur: 0.11102 (0.11975) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.06758 (9.44262) | > current_lr: 0.00035 | > step_time: 0.67930 (0.55899) | > loader_time: 0.00340 (0.00325)  --> STEP: 400/639 -- GLOBAL_STEP: 32400 | > loss: -0.46718 (-0.47836) | > log_mle: -0.58245 (-0.59822) | > loss_dur: 0.11527 (0.11986) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.65032 (9.38860) | > current_lr: 0.00035 | > step_time: 0.74210 (0.57164) | > loader_time: 0.00360 (0.00328)  --> STEP: 425/639 -- GLOBAL_STEP: 32425 | > loss: -0.52030 (-0.47659) | > log_mle: -0.63914 (-0.59651) | > loss_dur: 0.11884 (0.11992) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.73495 (9.30833) | > current_lr: 0.00035 | > step_time: 0.91500 (0.58393) | > loader_time: 0.00370 (0.00331)  --> STEP: 450/639 -- GLOBAL_STEP: 32450 | > loss: -0.43274 (-0.47521) | > log_mle: -0.54015 (-0.59496) | > loss_dur: 0.10741 (0.11975) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.49107 (9.20026) | > current_lr: 0.00035 | > step_time: 0.80270 (0.59602) | > loader_time: 0.00400 (0.00333)  --> STEP: 475/639 -- GLOBAL_STEP: 32475 | > loss: -0.44829 (-0.47372) | > log_mle: -0.55072 (-0.59351) | > loss_dur: 0.10243 (0.11979) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.96128 (9.14093) | > current_lr: 0.00035 | > step_time: 0.73870 (0.60732) | > loader_time: 0.00350 (0.00336)  --> STEP: 500/639 -- GLOBAL_STEP: 32500 | > loss: -0.49305 (-0.47266) | > log_mle: -0.61802 (-0.59250) | > loss_dur: 0.12497 (0.11984) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.51605 (9.09896) | > current_lr: 0.00035 | > step_time: 0.93420 (0.62042) | > loader_time: 0.00410 (0.00338)  --> STEP: 525/639 -- GLOBAL_STEP: 32525 | > loss: -0.46368 (-0.47169) | > log_mle: -0.59584 (-0.59154) | > loss_dur: 0.13216 (0.11985) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.69938 (9.05831) | > current_lr: 0.00035 | > step_time: 1.05960 (0.63466) | > loader_time: 0.00450 (0.00341)  --> STEP: 550/639 -- GLOBAL_STEP: 32550 | > loss: -0.41030 (-0.47049) | > log_mle: -0.52800 (-0.59028) | > loss_dur: 0.11770 (0.11979) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.34380 (8.96900) | > current_lr: 0.00035 | > step_time: 0.88320 (0.64876) | > loader_time: 0.00400 (0.00345)  --> STEP: 575/639 -- GLOBAL_STEP: 32575 | > loss: -0.44065 (-0.46933) | > log_mle: -0.54847 (-0.58909) | > loss_dur: 0.10782 (0.11976) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.56212 (8.88690) | > current_lr: 0.00035 | > step_time: 0.91970 (0.66350) | > loader_time: 0.00410 (0.00348)  --> STEP: 600/639 -- GLOBAL_STEP: 32600 | > loss: -0.42823 (-0.46843) | > log_mle: -0.54885 (-0.58829) | > loss_dur: 0.12062 (0.11986) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.28931 (8.82440) | > current_lr: 0.00035 | > step_time: 1.01850 (0.67991) | > loader_time: 0.00470 (0.00352)  --> STEP: 625/639 -- GLOBAL_STEP: 32625 | > loss: -0.45776 (-0.46761) | > log_mle: -0.57632 (-0.58747) | > loss_dur: 0.11856 (0.11986) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.45634 (8.77802) | > current_lr: 0.00035 | > step_time: 1.04530 (0.69762) | > loader_time: 0.00440 (0.00356) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03131 (+0.00523) | > avg_loss: -0.44797 (-0.00352) | > avg_log_mle: -0.57804 (-0.00599) | > avg_loss_dur: 0.13006 (+0.00248) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_32640.pth.tar  > EPOCH: 51/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 21:19:49)   --> STEP: 10/639 -- GLOBAL_STEP: 32650 | > loss: -0.50259 (-0.51219) | > log_mle: -0.64721 (-0.64259) | > loss_dur: 0.14462 (0.13040) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.58768 (13.12095) | > current_lr: 0.00035 | > step_time: 0.28370 (0.31398) | > loader_time: 0.00240 (0.01261)  --> STEP: 35/639 -- GLOBAL_STEP: 32675 | > loss: -0.50260 (-0.50270) | > log_mle: -0.60763 (-0.62558) | > loss_dur: 0.10504 (0.12288) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.74265 (11.31138) | > current_lr: 0.00035 | > step_time: 0.32410 (0.34481) | > loader_time: 0.00260 (0.00553)  --> STEP: 60/639 -- GLOBAL_STEP: 32700 | > loss: -0.45112 (-0.49529) | > log_mle: -0.56322 (-0.61756) | > loss_dur: 0.11209 (0.12227) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.99750 (10.32160) | > current_lr: 0.00035 | > step_time: 0.39660 (0.37094) | > loader_time: 0.00280 (0.00443)  --> STEP: 85/639 -- GLOBAL_STEP: 32725 | > loss: -0.43076 (-0.48917) | > log_mle: -0.55965 (-0.60953) | > loss_dur: 0.12888 (0.12036) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.21753 (9.60775) | > current_lr: 0.00035 | > step_time: 0.41640 (0.38854) | > loader_time: 0.00310 (0.00398)  --> STEP: 110/639 -- GLOBAL_STEP: 32750 | > loss: -0.46205 (-0.49126) | > log_mle: -0.58022 (-0.61191) | > loss_dur: 0.11817 (0.12065) | > amp_scaler: 4096.00000 (2345.89091) | > grad_norm: 6.14883 (9.81579) | > current_lr: 0.00035 | > step_time: 0.40670 (0.40714) | > loader_time: 0.00310 (0.00377)  --> STEP: 135/639 -- GLOBAL_STEP: 32775 | > loss: -0.44465 (-0.48979) | > log_mle: -0.56423 (-0.61003) | > loss_dur: 0.11959 (0.12024) | > amp_scaler: 4096.00000 (2669.98519) | > grad_norm: 5.72821 (9.74607) | > current_lr: 0.00035 | > step_time: 0.46960 (0.42242) | > loader_time: 0.00310 (0.00363)  --> STEP: 160/639 -- GLOBAL_STEP: 32800 | > loss: -0.49966 (-0.48884) | > log_mle: -0.60392 (-0.60853) | > loss_dur: 0.10426 (0.11969) | > amp_scaler: 4096.00000 (2892.80000) | > grad_norm: 6.57674 (9.57073) | > current_lr: 0.00035 | > step_time: 0.46660 (0.43530) | > loader_time: 0.00290 (0.00354)  --> STEP: 185/639 -- GLOBAL_STEP: 32825 | > loss: -0.47320 (-0.49100) | > log_mle: -0.58596 (-0.61088) | > loss_dur: 0.11276 (0.11988) | > amp_scaler: 4096.00000 (3055.39459) | > grad_norm: 10.28177 (9.91788) | > current_lr: 0.00035 | > step_time: 0.59690 (0.45105) | > loader_time: 0.00340 (0.00351)  --> STEP: 210/639 -- GLOBAL_STEP: 32850 | > loss: -0.40097 (-0.48853) | > log_mle: -0.50592 (-0.60814) | > loss_dur: 0.10495 (0.11961) | > amp_scaler: 4096.00000 (3179.27619) | > grad_norm: 2.19173 (9.69094) | > current_lr: 0.00035 | > step_time: 0.50430 (0.46631) | > loader_time: 0.00290 (0.00348)  --> STEP: 235/639 -- GLOBAL_STEP: 32875 | > loss: -0.48718 (-0.48937) | > log_mle: -0.60077 (-0.60876) | > loss_dur: 0.11359 (0.11939) | > amp_scaler: 4096.00000 (3276.80000) | > grad_norm: 8.10793 (9.74506) | > current_lr: 0.00035 | > step_time: 0.54970 (0.48031) | > loader_time: 0.00350 (0.00348)  --> STEP: 260/639 -- GLOBAL_STEP: 32900 | > loss: -0.47475 (-0.48919) | > log_mle: -0.59816 (-0.60869) | > loss_dur: 0.12340 (0.11949) | > amp_scaler: 4096.00000 (3355.56923) | > grad_norm: 11.74136 (9.77121) | > current_lr: 0.00035 | > step_time: 0.58170 (0.49402) | > loader_time: 0.00350 (0.00348)  --> STEP: 285/639 -- GLOBAL_STEP: 32925 | > loss: -0.42698 (-0.48837) | > log_mle: -0.54530 (-0.60794) | > loss_dur: 0.11831 (0.11957) | > amp_scaler: 4096.00000 (3420.51930) | > grad_norm: 3.13437 (9.67236) | > current_lr: 0.00035 | > step_time: 0.62740 (0.51109) | > loader_time: 0.00350 (0.00349)  --> STEP: 310/639 -- GLOBAL_STEP: 32950 | > loss: -0.46634 (-0.48698) | > log_mle: -0.58926 (-0.60663) | > loss_dur: 0.12293 (0.11965) | > amp_scaler: 4096.00000 (3474.99355) | > grad_norm: 7.87221 (9.58984) | > current_lr: 0.00035 | > step_time: 0.59140 (0.52492) | > loader_time: 0.00320 (0.00350)  --> STEP: 335/639 -- GLOBAL_STEP: 32975 | > loss: -0.48517 (-0.48472) | > log_mle: -0.61153 (-0.60478) | > loss_dur: 0.12635 (0.12006) | > amp_scaler: 4096.00000 (3521.33731) | > grad_norm: 17.17766 (9.55143) | > current_lr: 0.00035 | > step_time: 0.67880 (0.53960) | > loader_time: 0.00370 (0.00351)  --> STEP: 360/639 -- GLOBAL_STEP: 33000 | > loss: -0.42743 (-0.48327) | > log_mle: -0.54786 (-0.60305) | > loss_dur: 0.12043 (0.11978) | > amp_scaler: 4096.00000 (3561.24444) | > grad_norm: 6.71192 (9.40889) | > current_lr: 0.00035 | > step_time: 0.65330 (0.55169) | > loader_time: 0.00370 (0.00352)  --> STEP: 385/639 -- GLOBAL_STEP: 33025 | > loss: -0.47412 (-0.48188) | > log_mle: -0.58590 (-0.60143) | > loss_dur: 0.11178 (0.11955) | > amp_scaler: 4096.00000 (3595.96883) | > grad_norm: 6.57924 (9.27237) | > current_lr: 0.00035 | > step_time: 0.67180 (0.56570) | > loader_time: 0.00380 (0.00353)  --> STEP: 410/639 -- GLOBAL_STEP: 33050 | > loss: -0.42951 (-0.48017) | > log_mle: -0.55935 (-0.59991) | > loss_dur: 0.12983 (0.11975) | > amp_scaler: 4096.00000 (3626.45854) | > grad_norm: 5.78022 (9.20571) | > current_lr: 0.00035 | > step_time: 0.79970 (0.57714) | > loader_time: 0.00370 (0.00354)  --> STEP: 435/639 -- GLOBAL_STEP: 33075 | > loss: -0.45733 (-0.47835) | > log_mle: -0.56699 (-0.59800) | > loss_dur: 0.10966 (0.11966) | > amp_scaler: 4096.00000 (3653.44368) | > grad_norm: 6.05334 (9.04394) | > current_lr: 0.00035 | > step_time: 0.72240 (0.58917) | > loader_time: 0.00380 (0.00356)  --> STEP: 460/639 -- GLOBAL_STEP: 33100 | > loss: -0.41151 (-0.47659) | > log_mle: -0.53331 (-0.59611) | > loss_dur: 0.12180 (0.11952) | > amp_scaler: 4096.00000 (3677.49565) | > grad_norm: 1.83042 (8.91686) | > current_lr: 0.00035 | > step_time: 0.91160 (0.60117) | > loader_time: 0.00390 (0.00357)  --> STEP: 485/639 -- GLOBAL_STEP: 33125 | > loss: -0.45227 (-0.47506) | > log_mle: -0.56831 (-0.59479) | > loss_dur: 0.11604 (0.11973) | > amp_scaler: 4096.00000 (3699.06804) | > grad_norm: 5.32331 (8.83879) | > current_lr: 0.00035 | > step_time: 0.85510 (0.61502) | > loader_time: 0.00400 (0.00359)  --> STEP: 510/639 -- GLOBAL_STEP: 33150 | > loss: -0.45156 (-0.47399) | > log_mle: -0.57173 (-0.59363) | > loss_dur: 0.12017 (0.11964) | > amp_scaler: 4096.00000 (3718.52549) | > grad_norm: 5.65560 (8.73863) | > current_lr: 0.00035 | > step_time: 0.87690 (0.62810) | > loader_time: 0.00420 (0.00360)  --> STEP: 535/639 -- GLOBAL_STEP: 33175 | > loss: -0.43243 (-0.47271) | > log_mle: -0.56055 (-0.59240) | > loss_dur: 0.12811 (0.11969) | > amp_scaler: 4096.00000 (3736.16449) | > grad_norm: 6.90368 (8.66825) | > current_lr: 0.00035 | > step_time: 0.92490 (0.64142) | > loader_time: 0.00400 (0.00362)  --> STEP: 560/639 -- GLOBAL_STEP: 33200 | > loss: -0.44939 (-0.47158) | > log_mle: -0.57053 (-0.59130) | > loss_dur: 0.12115 (0.11971) | > amp_scaler: 4096.00000 (3752.22857) | > grad_norm: 7.89587 (8.57917) | > current_lr: 0.00035 | > step_time: 1.01570 (0.65635) | > loader_time: 0.00420 (0.00365)  --> STEP: 585/639 -- GLOBAL_STEP: 33225 | > loss: -0.46193 (-0.47064) | > log_mle: -0.57337 (-0.59030) | > loss_dur: 0.11144 (0.11966) | > amp_scaler: 4096.00000 (3766.91966) | > grad_norm: 6.82414 (8.48736) | > current_lr: 0.00035 | > step_time: 1.07130 (0.67170) | > loader_time: 0.00420 (0.00368)  --> STEP: 610/639 -- GLOBAL_STEP: 33250 | > loss: -0.42516 (-0.46960) | > log_mle: -0.54321 (-0.58935) | > loss_dur: 0.11805 (0.11975) | > amp_scaler: 4096.00000 (3780.40656) | > grad_norm: 3.80214 (8.42322) | > current_lr: 0.00035 | > step_time: 1.03960 (0.68785) | > loader_time: 0.00440 (0.00371)  --> STEP: 635/639 -- GLOBAL_STEP: 33275 | > loss: -0.44370 (-0.46881) | > log_mle: -0.56743 (-0.58875) | > loss_dur: 0.12373 (0.11994) | > amp_scaler: 4096.00000 (3792.83150) | > grad_norm: 6.03822 (8.38968) | > current_lr: 0.00035 | > step_time: 1.11650 (0.70513) | > loader_time: 0.00450 (0.00374) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02820 (-0.00311) | > avg_loss: -0.45019 (-0.00222) | > avg_log_mle: -0.57711 (+0.00093) | > avg_loss_dur: 0.12691 (-0.00315) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_33280.pth.tar  > EPOCH: 52/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 21:27:42)   --> STEP: 20/639 -- GLOBAL_STEP: 33300 | > loss: -0.60942 (-0.52448) | > log_mle: -0.75395 (-0.64961) | > loss_dur: 0.14453 (0.12513) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 19.77792 (11.76562) | > current_lr: 0.00035 | > step_time: 0.34680 (0.31541) | > loader_time: 0.00270 (0.00366)  --> STEP: 45/639 -- GLOBAL_STEP: 33325 | > loss: -0.45775 (-0.50341) | > log_mle: -0.56424 (-0.62514) | > loss_dur: 0.10649 (0.12172) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.86246 (9.17338) | > current_lr: 0.00035 | > step_time: 0.43700 (0.34944) | > loader_time: 0.00300 (0.00319)  --> STEP: 70/639 -- GLOBAL_STEP: 33350 | > loss: -0.50939 (-0.49434) | > log_mle: -0.62599 (-0.61584) | > loss_dur: 0.11660 (0.12150) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.21054 (8.92940) | > current_lr: 0.00035 | > step_time: 0.39760 (0.37682) | > loader_time: 0.00270 (0.00305)  --> STEP: 95/639 -- GLOBAL_STEP: 33375 | > loss: -0.56510 (-0.49278) | > log_mle: -0.67393 (-0.61177) | > loss_dur: 0.10883 (0.11899) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.04722 (8.43845) | > current_lr: 0.00035 | > step_time: 0.41010 (0.39274) | > loader_time: 0.00250 (0.00301)  --> STEP: 120/639 -- GLOBAL_STEP: 33400 | > loss: -0.52728 (-0.49120) | > log_mle: -0.66309 (-0.61115) | > loss_dur: 0.13581 (0.11995) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 21.13673 (8.55676) | > current_lr: 0.00035 | > step_time: 0.59190 (0.41207) | > loader_time: 0.00270 (0.00300)  --> STEP: 145/639 -- GLOBAL_STEP: 33425 | > loss: -0.50068 (-0.49173) | > log_mle: -0.62531 (-0.61200) | > loss_dur: 0.12463 (0.12027) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.65116 (8.83188) | > current_lr: 0.00035 | > step_time: 0.47050 (0.42615) | > loader_time: 0.00300 (0.00301)  --> STEP: 170/639 -- GLOBAL_STEP: 33450 | > loss: -0.49474 (-0.49143) | > log_mle: -0.60860 (-0.61131) | > loss_dur: 0.11386 (0.11988) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.46629 (8.97465) | > current_lr: 0.00035 | > step_time: 0.54700 (0.44048) | > loader_time: 0.00340 (0.00302)  --> STEP: 195/639 -- GLOBAL_STEP: 33475 | > loss: -0.44594 (-0.49089) | > log_mle: -0.54822 (-0.61021) | > loss_dur: 0.10229 (0.11932) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.04497 (9.01988) | > current_lr: 0.00035 | > step_time: 0.56200 (0.45703) | > loader_time: 0.00310 (0.00305)  --> STEP: 220/639 -- GLOBAL_STEP: 33500 | > loss: -0.49279 (-0.48863) | > log_mle: -0.59891 (-0.60776) | > loss_dur: 0.10612 (0.11913) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.22081 (9.20252) | > current_lr: 0.00035 | > step_time: 0.71690 (0.47260) | > loader_time: 0.00310 (0.00307)  --> STEP: 245/639 -- GLOBAL_STEP: 33525 | > loss: -0.48152 (-0.48976) | > log_mle: -0.59085 (-0.60873) | > loss_dur: 0.10933 (0.11896) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.82069 (9.37306) | > current_lr: 0.00035 | > step_time: 0.52840 (0.48610) | > loader_time: 0.00310 (0.00311)  --> STEP: 270/639 -- GLOBAL_STEP: 33550 | > loss: -0.52566 (-0.48955) | > log_mle: -0.65452 (-0.60888) | > loss_dur: 0.12886 (0.11933) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.62142 (9.51251) | > current_lr: 0.00035 | > step_time: 0.85210 (0.50272) | > loader_time: 0.00360 (0.00313)  --> STEP: 295/639 -- GLOBAL_STEP: 33575 | > loss: -0.44053 (-0.48780) | > log_mle: -0.54899 (-0.60717) | > loss_dur: 0.10847 (0.11937) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.49599 (9.44172) | > current_lr: 0.00035 | > step_time: 0.63970 (0.51482) | > loader_time: 0.00360 (0.00317)  --> STEP: 320/639 -- GLOBAL_STEP: 33600 | > loss: -0.46473 (-0.48659) | > log_mle: -0.59508 (-0.60607) | > loss_dur: 0.13035 (0.11948) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.97428 (9.36735) | > current_lr: 0.00035 | > step_time: 0.74690 (0.52873) | > loader_time: 0.00380 (0.00320)  --> STEP: 345/639 -- GLOBAL_STEP: 33625 | > loss: -0.47321 (-0.48487) | > log_mle: -0.58485 (-0.60454) | > loss_dur: 0.11164 (0.11968) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.26344 (9.27188) | > current_lr: 0.00035 | > step_time: 0.66410 (0.54316) | > loader_time: 0.00370 (0.00323)  --> STEP: 370/639 -- GLOBAL_STEP: 33650 | > loss: -0.48051 (-0.48328) | > log_mle: -0.59929 (-0.60264) | > loss_dur: 0.11877 (0.11936) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.71469 (9.11233) | > current_lr: 0.00034 | > step_time: 0.92550 (0.55518) | > loader_time: 0.00380 (0.00325)  --> STEP: 395/639 -- GLOBAL_STEP: 33675 | > loss: -0.44672 (-0.48161) | > log_mle: -0.56687 (-0.60088) | > loss_dur: 0.12015 (0.11927) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.01492 (8.96044) | > current_lr: 0.00034 | > step_time: 0.77560 (0.56762) | > loader_time: 0.00340 (0.00327)  --> STEP: 420/639 -- GLOBAL_STEP: 33700 | > loss: -0.44609 (-0.47980) | > log_mle: -0.56324 (-0.59921) | > loss_dur: 0.11715 (0.11941) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.11830 (8.83618) | > current_lr: 0.00034 | > step_time: 0.70650 (0.58084) | > loader_time: 0.00350 (0.00330)  --> STEP: 445/639 -- GLOBAL_STEP: 33725 | > loss: -0.45676 (-0.47850) | > log_mle: -0.56821 (-0.59787) | > loss_dur: 0.11145 (0.11936) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.14140 (8.77689) | > current_lr: 0.00034 | > step_time: 0.74650 (0.59339) | > loader_time: 0.00380 (0.00332)  --> STEP: 470/639 -- GLOBAL_STEP: 33750 | > loss: -0.46299 (-0.47692) | > log_mle: -0.57794 (-0.59621) | > loss_dur: 0.11495 (0.11929) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.73199 (8.65519) | > current_lr: 0.00034 | > step_time: 0.79050 (0.60530) | > loader_time: 0.00390 (0.00334)  --> STEP: 495/639 -- GLOBAL_STEP: 33775 | > loss: -0.40949 (-0.47561) | > log_mle: -0.51579 (-0.59496) | > loss_dur: 0.10630 (0.11934) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.31023 (8.57093) | > current_lr: 0.00034 | > step_time: 0.72680 (0.61746) | > loader_time: 0.00370 (0.00337)  --> STEP: 520/639 -- GLOBAL_STEP: 33800 | > loss: -0.45698 (-0.47474) | > log_mle: -0.57099 (-0.59400) | > loss_dur: 0.11401 (0.11926) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.24721 (8.49201) | > current_lr: 0.00034 | > step_time: 0.81130 (0.63087) | > loader_time: 0.00370 (0.00339)  --> STEP: 545/639 -- GLOBAL_STEP: 33825 | > loss: -0.47220 (-0.47353) | > log_mle: -0.60483 (-0.59288) | > loss_dur: 0.13262 (0.11935) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.21262 (8.43002) | > current_lr: 0.00034 | > step_time: 0.97240 (0.64612) | > loader_time: 0.00380 (0.00343)  --> STEP: 570/639 -- GLOBAL_STEP: 33850 | > loss: -0.44665 (-0.47221) | > log_mle: -0.55246 (-0.59158) | > loss_dur: 0.10581 (0.11937) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.70473 (8.35295) | > current_lr: 0.00034 | > step_time: 0.95680 (0.66038) | > loader_time: 0.00410 (0.00346)  --> STEP: 595/639 -- GLOBAL_STEP: 33875 | > loss: -0.43662 (-0.47143) | > log_mle: -0.54558 (-0.59081) | > loss_dur: 0.10897 (0.11938) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.76959 (8.32411) | > current_lr: 0.00034 | > step_time: 0.96240 (0.67668) | > loader_time: 0.00400 (0.00350)  --> STEP: 620/639 -- GLOBAL_STEP: 33900 | > loss: -0.44325 (-0.47038) | > log_mle: -0.55568 (-0.58982) | > loss_dur: 0.11244 (0.11945) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.00165 (8.29204) | > current_lr: 0.00034 | > step_time: 1.27870 (0.69351) | > loader_time: 0.00480 (0.00353) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02897 (+0.00077) | > avg_loss: -0.44986 (+0.00034) | > avg_log_mle: -0.57780 (-0.00069) | > avg_loss_dur: 0.12794 (+0.00103)  > EPOCH: 53/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 21:35:29)   --> STEP: 5/639 -- GLOBAL_STEP: 33925 | > loss: -0.51213 (-0.52095) | > log_mle: -0.66805 (-0.65486) | > loss_dur: 0.15592 (0.13392) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.39302 (10.92341) | > current_lr: 0.00034 | > step_time: 0.31380 (0.28586) | > loader_time: 0.00370 (0.00247)  --> STEP: 30/639 -- GLOBAL_STEP: 33950 | > loss: -0.52229 (-0.51234) | > log_mle: -0.63355 (-0.63580) | > loss_dur: 0.11126 (0.12346) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.07614 (9.29564) | > current_lr: 0.00034 | > step_time: 0.32510 (0.32849) | > loader_time: 0.00300 (0.00259)  --> STEP: 55/639 -- GLOBAL_STEP: 33975 | > loss: -0.46457 (-0.50419) | > log_mle: -0.56303 (-0.62473) | > loss_dur: 0.09846 (0.12054) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.83781 (8.69541) | > current_lr: 0.00034 | > step_time: 0.36410 (0.35850) | > loader_time: 0.00300 (0.00272)  --> STEP: 80/639 -- GLOBAL_STEP: 34000 | > loss: -0.46606 (-0.49744) | > log_mle: -0.60316 (-0.61640) | > loss_dur: 0.13710 (0.11896) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.18092 (8.78419) | > current_lr: 0.00034 | > step_time: 0.40160 (0.37933) | > loader_time: 0.00300 (0.00281)  --> STEP: 105/639 -- GLOBAL_STEP: 34025 | > loss: -0.44208 (-0.49575) | > log_mle: -0.56185 (-0.61404) | > loss_dur: 0.11977 (0.11829) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.13486 (8.98054) | > current_lr: 0.00034 | > step_time: 0.48140 (0.40021) | > loader_time: 0.00310 (0.00286)  --> STEP: 130/639 -- GLOBAL_STEP: 34050 | > loss: -0.47405 (-0.49397) | > log_mle: -0.59960 (-0.61293) | > loss_dur: 0.12556 (0.11895) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.51545 (9.16496) | > current_lr: 0.00034 | > step_time: 0.47710 (0.41466) | > loader_time: 0.00320 (0.00290)  --> STEP: 155/639 -- GLOBAL_STEP: 34075 | > loss: -0.47026 (-0.49376) | > log_mle: -0.59018 (-0.61242) | > loss_dur: 0.11992 (0.11866) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.87111 (9.26709) | > current_lr: 0.00034 | > step_time: 0.53280 (0.43048) | > loader_time: 0.00310 (0.00293)  --> STEP: 180/639 -- GLOBAL_STEP: 34100 | > loss: -0.52123 (-0.49520) | > log_mle: -0.64708 (-0.61397) | > loss_dur: 0.12585 (0.11877) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.05310 (9.43085) | > current_lr: 0.00034 | > step_time: 0.58240 (0.44790) | > loader_time: 0.00320 (0.00297)  --> STEP: 205/639 -- GLOBAL_STEP: 34125 | > loss: -0.45933 (-0.49214) | > log_mle: -0.57217 (-0.61034) | > loss_dur: 0.11285 (0.11820) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.15289 (9.13187) | > current_lr: 0.00034 | > step_time: 0.56330 (0.46210) | > loader_time: 0.00350 (0.00301)  --> STEP: 230/639 -- GLOBAL_STEP: 34150 | > loss: -0.48684 (-0.49220) | > log_mle: -0.60200 (-0.61034) | > loss_dur: 0.11516 (0.11814) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.54677 (9.34251) | > current_lr: 0.00034 | > step_time: 0.61320 (0.47819) | > loader_time: 0.00340 (0.00306)  --> STEP: 255/639 -- GLOBAL_STEP: 34175 | > loss: -0.45318 (-0.49138) | > log_mle: -0.57390 (-0.60934) | > loss_dur: 0.12072 (0.11796) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.86929 (9.35093) | > current_lr: 0.00034 | > step_time: 0.72180 (0.49244) | > loader_time: 0.00310 (0.00309)  --> STEP: 280/639 -- GLOBAL_STEP: 34200 | > loss: -0.47279 (-0.49153) | > log_mle: -0.60103 (-0.60979) | > loss_dur: 0.12825 (0.11826) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.60692 (9.41892) | > current_lr: 0.00034 | > step_time: 0.60800 (0.51078) | > loader_time: 0.00310 (0.00313)  --> STEP: 305/639 -- GLOBAL_STEP: 34225 | > loss: -0.46024 (-0.48999) | > log_mle: -0.58028 (-0.60836) | > loss_dur: 0.12004 (0.11837) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.09100 (9.36864) | > current_lr: 0.00034 | > step_time: 0.67110 (0.52489) | > loader_time: 0.00340 (0.00316)  --> STEP: 330/639 -- GLOBAL_STEP: 34250 | > loss: -0.46683 (-0.48776) | > log_mle: -0.60245 (-0.60658) | > loss_dur: 0.13561 (0.11882) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.23879 (9.31665) | > current_lr: 0.00034 | > step_time: 0.83440 (0.53975) | > loader_time: 0.00350 (0.00320)  --> STEP: 355/639 -- GLOBAL_STEP: 34275 | > loss: -0.42072 (-0.48630) | > log_mle: -0.53125 (-0.60486) | > loss_dur: 0.11052 (0.11856) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.79487 (9.19852) | > current_lr: 0.00034 | > step_time: 0.61870 (0.55285) | > loader_time: 0.00330 (0.00322)  --> STEP: 380/639 -- GLOBAL_STEP: 34300 | > loss: -0.45942 (-0.48443) | > log_mle: -0.56842 (-0.60287) | > loss_dur: 0.10900 (0.11844) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.24589 (9.00550) | > current_lr: 0.00034 | > step_time: 0.62760 (0.56583) | > loader_time: 0.00360 (0.00325)  --> STEP: 405/639 -- GLOBAL_STEP: 34325 | > loss: -0.45620 (-0.48325) | > log_mle: -0.57232 (-0.60182) | > loss_dur: 0.11613 (0.11857) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.30016 (8.98558) | > current_lr: 0.00034 | > step_time: 0.72340 (0.57834) | > loader_time: 0.00340 (0.00329)  --> STEP: 430/639 -- GLOBAL_STEP: 34350 | > loss: -0.42096 (-0.48155) | > log_mle: -0.53095 (-0.60013) | > loss_dur: 0.10999 (0.11858) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.35023 (8.90861) | > current_lr: 0.00034 | > step_time: 0.76610 (0.59139) | > loader_time: 0.00370 (0.00332)  --> STEP: 455/639 -- GLOBAL_STEP: 34375 | > loss: -0.44152 (-0.47990) | > log_mle: -0.55881 (-0.59832) | > loss_dur: 0.11729 (0.11843) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.64503 (8.76950) | > current_lr: 0.00034 | > step_time: 0.73190 (0.60182) | > loader_time: 0.00380 (0.00334)  --> STEP: 480/639 -- GLOBAL_STEP: 34400 | > loss: -0.40320 (-0.47831) | > log_mle: -0.51327 (-0.59687) | > loss_dur: 0.11007 (0.11857) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.69360 (8.78467) | > current_lr: 0.00034 | > step_time: 0.79850 (0.61548) | > loader_time: 0.00340 (0.00336)  --> STEP: 505/639 -- GLOBAL_STEP: 34425 | > loss: -0.46855 (-0.47714) | > log_mle: -0.58090 (-0.59563) | > loss_dur: 0.11236 (0.11849) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.87626 (8.70731) | > current_lr: 0.00034 | > step_time: 0.89220 (0.62793) | > loader_time: 0.00390 (0.00339)  --> STEP: 530/639 -- GLOBAL_STEP: 34450 | > loss: -0.43929 (-0.47600) | > log_mle: -0.57011 (-0.59457) | > loss_dur: 0.13082 (0.11857) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.32122 (8.66427) | > current_lr: 0.00034 | > step_time: 0.91650 (0.64131) | > loader_time: 0.00420 (0.00342)  --> STEP: 555/639 -- GLOBAL_STEP: 34475 | > loss: -0.43879 (-0.47472) | > log_mle: -0.55331 (-0.59326) | > loss_dur: 0.11451 (0.11854) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.91612 (8.55738) | > current_lr: 0.00034 | > step_time: 0.86500 (0.65543) | > loader_time: 0.00400 (0.00345)  --> STEP: 580/639 -- GLOBAL_STEP: 34500 | > loss: -0.46253 (-0.47369) | > log_mle: -0.57062 (-0.59226) | > loss_dur: 0.10809 (0.11857) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.12966 (8.50337) | > current_lr: 0.00034 | > step_time: 0.98090 (0.67085) | > loader_time: 0.00430 (0.00349)  --> STEP: 605/639 -- GLOBAL_STEP: 34525 | > loss: -0.43080 (-0.47262) | > log_mle: -0.54051 (-0.59133) | > loss_dur: 0.10971 (0.11871) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.91979 (8.45293) | > current_lr: 0.00034 | > step_time: 1.07090 (0.68705) | > loader_time: 0.00450 (0.00353)  --> STEP: 630/639 -- GLOBAL_STEP: 34550 | > loss: -0.43708 (-0.47174) | > log_mle: -0.57072 (-0.59056) | > loss_dur: 0.13364 (0.11882) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.96475 (8.41272) | > current_lr: 0.00034 | > step_time: 1.21710 (0.70444) | > loader_time: 0.00460 (0.00357) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03727 (+0.00831) | > avg_loss: -0.44509 (+0.00476) | > avg_log_mle: -0.57723 (+0.00057) | > avg_loss_dur: 0.13214 (+0.00419)  > EPOCH: 54/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 21:43:18)   --> STEP: 15/639 -- GLOBAL_STEP: 34575 | > loss: -0.61681 (-0.52467) | > log_mle: -0.76231 (-0.64791) | > loss_dur: 0.14550 (0.12324) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.56375 (10.04089) | > current_lr: 0.00034 | > step_time: 0.35150 (0.31319) | > loader_time: 0.00250 (0.00257)  --> STEP: 40/639 -- GLOBAL_STEP: 34600 | > loss: -0.48818 (-0.50711) | > log_mle: -0.60366 (-0.62754) | > loss_dur: 0.11547 (0.12043) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.28070 (8.74814) | > current_lr: 0.00034 | > step_time: 0.39250 (0.33716) | > loader_time: 0.00290 (0.00275)  --> STEP: 65/639 -- GLOBAL_STEP: 34625 | > loss: -0.48516 (-0.49914) | > log_mle: -0.59464 (-0.61994) | > loss_dur: 0.10948 (0.12080) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.72940 (8.98196) | > current_lr: 0.00034 | > step_time: 0.38590 (0.36819) | > loader_time: 0.00310 (0.00283)  --> STEP: 90/639 -- GLOBAL_STEP: 34650 | > loss: -0.44076 (-0.49293) | > log_mle: -0.55042 (-0.61191) | > loss_dur: 0.10966 (0.11898) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.31592 (8.37068) | > current_lr: 0.00034 | > step_time: 0.51870 (0.38688) | > loader_time: 0.00330 (0.00288)  --> STEP: 115/639 -- GLOBAL_STEP: 34675 | > loss: -0.40252 (-0.49364) | > log_mle: -0.51934 (-0.61228) | > loss_dur: 0.11682 (0.11864) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.98006 (8.74237) | > current_lr: 0.00034 | > step_time: 0.40830 (0.40471) | > loader_time: 0.00290 (0.00293)  --> STEP: 140/639 -- GLOBAL_STEP: 34700 | > loss: -0.46564 (-0.49389) | > log_mle: -0.60229 (-0.61287) | > loss_dur: 0.13665 (0.11898) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.66364 (8.94606) | > current_lr: 0.00034 | > step_time: 0.55360 (0.41897) | > loader_time: 0.00330 (0.00294)  --> STEP: 165/639 -- GLOBAL_STEP: 34725 | > loss: -0.48202 (-0.49363) | > log_mle: -0.61371 (-0.61227) | > loss_dur: 0.13169 (0.11864) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.45188 (9.06767) | > current_lr: 0.00034 | > step_time: 0.56020 (0.43160) | > loader_time: 0.00310 (0.00299)  --> STEP: 190/639 -- GLOBAL_STEP: 34750 | > loss: -0.42073 (-0.49361) | > log_mle: -0.53880 (-0.61209) | > loss_dur: 0.11807 (0.11849) | > amp_scaler: 8192.00000 (4440.92632) | > grad_norm: 2.88485 (9.37307) | > current_lr: 0.00034 | > step_time: 0.53130 (0.44973) | > loader_time: 0.00310 (0.00304)  --> STEP: 215/639 -- GLOBAL_STEP: 34775 | > loss: -0.53379 (-0.49158) | > log_mle: -0.64704 (-0.61024) | > loss_dur: 0.11324 (0.11865) | > amp_scaler: 4096.00000 (4457.97209) | > grad_norm: 12.32021 (9.32037) | > current_lr: 0.00034 | > step_time: 0.53870 (0.46520) | > loader_time: 0.00350 (0.00307)  --> STEP: 240/639 -- GLOBAL_STEP: 34800 | > loss: -0.53549 (-0.49270) | > log_mle: -0.65576 (-0.61104) | > loss_dur: 0.12027 (0.11834) | > amp_scaler: 4096.00000 (4420.26667) | > grad_norm: 11.81677 (9.47680) | > current_lr: 0.00034 | > step_time: 0.55040 (0.48009) | > loader_time: 0.00320 (0.00310)  --> STEP: 265/639 -- GLOBAL_STEP: 34825 | > loss: -0.53891 (-0.49189) | > log_mle: -0.66986 (-0.61058) | > loss_dur: 0.13094 (0.11869) | > amp_scaler: 4096.00000 (4389.67547) | > grad_norm: 17.36074 (9.57617) | > current_lr: 0.00034 | > step_time: 0.61690 (0.49583) | > loader_time: 0.00310 (0.00313)  --> STEP: 290/639 -- GLOBAL_STEP: 34850 | > loss: -0.46055 (-0.49045) | > log_mle: -0.57386 (-0.60918) | > loss_dur: 0.11331 (0.11873) | > amp_scaler: 4096.00000 (4364.35862) | > grad_norm: 8.71892 (9.48833) | > current_lr: 0.00034 | > step_time: 0.73530 (0.51274) | > loader_time: 0.00350 (0.00316)  --> STEP: 315/639 -- GLOBAL_STEP: 34875 | > loss: -0.44026 (-0.48897) | > log_mle: -0.55369 (-0.60785) | > loss_dur: 0.11343 (0.11888) | > amp_scaler: 4096.00000 (4343.06032) | > grad_norm: 3.72207 (9.46484) | > current_lr: 0.00034 | > step_time: 0.64880 (0.52754) | > loader_time: 0.00340 (0.00320)  --> STEP: 340/639 -- GLOBAL_STEP: 34900 | > loss: -0.46925 (-0.48732) | > log_mle: -0.57856 (-0.60666) | > loss_dur: 0.10931 (0.11934) | > amp_scaler: 4096.00000 (4324.89412) | > grad_norm: 7.74619 (9.47322) | > current_lr: 0.00034 | > step_time: 0.59780 (0.54443) | > loader_time: 0.00340 (0.00323)  --> STEP: 365/639 -- GLOBAL_STEP: 34925 | > loss: -0.47215 (-0.48557) | > log_mle: -0.58505 (-0.60453) | > loss_dur: 0.11290 (0.11896) | > amp_scaler: 4096.00000 (4309.21644) | > grad_norm: 7.74461 (9.28231) | > current_lr: 0.00034 | > step_time: 0.83600 (0.55598) | > loader_time: 0.00410 (0.00326)  --> STEP: 390/639 -- GLOBAL_STEP: 34950 | > loss: -0.45854 (-0.48404) | > log_mle: -0.58568 (-0.60283) | > loss_dur: 0.12714 (0.11879) | > amp_scaler: 4096.00000 (4295.54872) | > grad_norm: 7.93169 (9.11614) | > current_lr: 0.00034 | > step_time: 0.73490 (0.56973) | > loader_time: 0.00360 (0.00328)  --> STEP: 415/639 -- GLOBAL_STEP: 34975 | > loss: -0.43434 (-0.48254) | > log_mle: -0.55921 (-0.60147) | > loss_dur: 0.12487 (0.11892) | > amp_scaler: 4096.00000 (4283.52771) | > grad_norm: 7.86257 (9.09646) | > current_lr: 0.00034 | > step_time: 0.73270 (0.58264) | > loader_time: 0.00350 (0.00331)  --> STEP: 440/639 -- GLOBAL_STEP: 35000 | > loss: -0.42386 (-0.48115) | > log_mle: -0.54174 (-0.60007) | > loss_dur: 0.11788 (0.11891) | > amp_scaler: 4096.00000 (4272.87273) | > grad_norm: 4.17560 (8.99784) | > current_lr: 0.00034 | > step_time: 0.72370 (0.59534) | > loader_time: 0.00360 (0.00333)  --> STEP: 465/639 -- GLOBAL_STEP: 35025 | > loss: -0.43034 (-0.47937) | > log_mle: -0.54456 (-0.59817) | > loss_dur: 0.11422 (0.11880) | > amp_scaler: 4096.00000 (4263.36344) | > grad_norm: 6.00860 (8.85127) | > current_lr: 0.00034 | > step_time: 0.79250 (0.60583) | > loader_time: 0.00360 (0.00336)  --> STEP: 490/639 -- GLOBAL_STEP: 35050 | > loss: -0.42124 (-0.47788) | > log_mle: -0.52825 (-0.59682) | > loss_dur: 0.10701 (0.11894) | > amp_scaler: 4096.00000 (4254.82449) | > grad_norm: 1.93563 (8.76566) | > current_lr: 0.00034 | > step_time: 0.80860 (0.61909) | > loader_time: 0.00380 (0.00338)  --> STEP: 515/639 -- GLOBAL_STEP: 35075 | > loss: -0.45017 (-0.47695) | > log_mle: -0.56112 (-0.59587) | > loss_dur: 0.11094 (0.11892) | > amp_scaler: 4096.00000 (4247.11456) | > grad_norm: 4.64763 (8.69162) | > current_lr: 0.00034 | > step_time: 0.96910 (0.63246) | > loader_time: 0.00410 (0.00341)  --> STEP: 540/639 -- GLOBAL_STEP: 35100 | > loss: -0.45424 (-0.47569) | > log_mle: -0.57740 (-0.59466) | > loss_dur: 0.12316 (0.11896) | > amp_scaler: 4096.00000 (4240.11852) | > grad_norm: 6.42755 (8.59785) | > current_lr: 0.00034 | > step_time: 0.90460 (0.64717) | > loader_time: 0.00420 (0.00344)  --> STEP: 565/639 -- GLOBAL_STEP: 35125 | > loss: -0.41918 (-0.47434) | > log_mle: -0.53620 (-0.59333) | > loss_dur: 0.11703 (0.11899) | > amp_scaler: 4096.00000 (4233.74159) | > grad_norm: 3.97210 (8.52687) | > current_lr: 0.00034 | > step_time: 0.99810 (0.66096) | > loader_time: 0.00420 (0.00348)  --> STEP: 590/639 -- GLOBAL_STEP: 35150 | > loss: -0.45143 (-0.47357) | > log_mle: -0.57076 (-0.59258) | > loss_dur: 0.11933 (0.11902) | > amp_scaler: 4096.00000 (4227.90508) | > grad_norm: 5.86915 (8.45949) | > current_lr: 0.00034 | > step_time: 0.94560 (0.67727) | > loader_time: 0.00440 (0.00352)  --> STEP: 615/639 -- GLOBAL_STEP: 35175 | > loss: -0.45356 (-0.47248) | > log_mle: -0.57257 (-0.59153) | > loss_dur: 0.11902 (0.11905) | > amp_scaler: 4096.00000 (4222.54309) | > grad_norm: 7.12993 (8.42807) | > current_lr: 0.00034 | > step_time: 1.04950 (0.69360) | > loader_time: 0.00420 (0.00355) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03256 (-0.00471) | > avg_loss: -0.44789 (-0.00280) | > avg_log_mle: -0.57698 (+0.00025) | > avg_loss_dur: 0.12909 (-0.00304)  > EPOCH: 55/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 21:51:07)   --> STEP: 0/639 -- GLOBAL_STEP: 35200 | > loss: -0.51600 (-0.51600) | > log_mle: -0.69162 (-0.69162) | > loss_dur: 0.17562 (0.17562) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 39.86820 (39.86820) | > current_lr: 0.00034 | > step_time: 0.77130 (0.77127) | > loader_time: 0.45300 (0.45296)  --> STEP: 25/639 -- GLOBAL_STEP: 35225 | > loss: -0.51191 (-0.51990) | > log_mle: -0.64001 (-0.64314) | > loss_dur: 0.12810 (0.12323) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.74044 (10.70863) | > current_lr: 0.00034 | > step_time: 0.39000 (0.34959) | > loader_time: 0.00300 (0.00274)  --> STEP: 50/639 -- GLOBAL_STEP: 35250 | > loss: -0.47343 (-0.50705) | > log_mle: -0.57423 (-0.62686) | > loss_dur: 0.10080 (0.11981) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.41681 (9.59422) | > current_lr: 0.00034 | > step_time: 0.39400 (0.36316) | > loader_time: 0.00300 (0.00279)  --> STEP: 75/639 -- GLOBAL_STEP: 35275 | > loss: -0.49691 (-0.49880) | > log_mle: -0.60748 (-0.61820) | > loss_dur: 0.11057 (0.11940) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.80552 (9.22215) | > current_lr: 0.00034 | > step_time: 0.44140 (0.37838) | > loader_time: 0.00290 (0.00283)  --> STEP: 100/639 -- GLOBAL_STEP: 35300 | > loss: -0.50525 (-0.49736) | > log_mle: -0.61512 (-0.61554) | > loss_dur: 0.10987 (0.11819) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.69929 (8.95576) | > current_lr: 0.00034 | > step_time: 0.51130 (0.39522) | > loader_time: 0.00310 (0.00286)  --> STEP: 125/639 -- GLOBAL_STEP: 35325 | > loss: -0.49643 (-0.49598) | > log_mle: -0.61267 (-0.61454) | > loss_dur: 0.11624 (0.11856) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.45620 (9.14676) | > current_lr: 0.00034 | > step_time: 0.42870 (0.41075) | > loader_time: 0.00300 (0.00289)  --> STEP: 150/639 -- GLOBAL_STEP: 35350 | > loss: -0.45974 (-0.49564) | > log_mle: -0.57942 (-0.61414) | > loss_dur: 0.11968 (0.11851) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.57763 (9.18736) | > current_lr: 0.00034 | > step_time: 0.42830 (0.42591) | > loader_time: 0.00280 (0.00292)  --> STEP: 175/639 -- GLOBAL_STEP: 35375 | > loss: -0.49224 (-0.49627) | > log_mle: -0.59531 (-0.61466) | > loss_dur: 0.10307 (0.11838) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.81142 (9.20496) | > current_lr: 0.00034 | > step_time: 0.55890 (0.44134) | > loader_time: 0.00340 (0.00296)  --> STEP: 200/639 -- GLOBAL_STEP: 35400 | > loss: -0.46323 (-0.49453) | > log_mle: -0.57955 (-0.61238) | > loss_dur: 0.11633 (0.11786) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.88632 (9.07344) | > current_lr: 0.00034 | > step_time: 0.65990 (0.45675) | > loader_time: 0.00310 (0.00298)  --> STEP: 225/639 -- GLOBAL_STEP: 35425 | > loss: -0.45306 (-0.49300) | > log_mle: -0.55637 (-0.61055) | > loss_dur: 0.10331 (0.11755) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.82001 (8.98098) | > current_lr: 0.00034 | > step_time: 0.55080 (0.47266) | > loader_time: 0.00320 (0.00303)  --> STEP: 250/639 -- GLOBAL_STEP: 35450 | > loss: -0.45618 (-0.49406) | > log_mle: -0.56026 (-0.61171) | > loss_dur: 0.10407 (0.11765) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.41822 (9.14044) | > current_lr: 0.00034 | > step_time: 0.61070 (0.48934) | > loader_time: 0.00350 (0.00306)  --> STEP: 275/639 -- GLOBAL_STEP: 35475 | > loss: -0.48455 (-0.49419) | > log_mle: -0.59931 (-0.61225) | > loss_dur: 0.11476 (0.11806) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.15131 (9.37105) | > current_lr: 0.00034 | > step_time: 0.85980 (0.50756) | > loader_time: 0.00370 (0.00310)  --> STEP: 300/639 -- GLOBAL_STEP: 35500 | > loss: -0.47574 (-0.49212) | > log_mle: -0.58725 (-0.61040) | > loss_dur: 0.11151 (0.11828) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.29831 (9.30782) | > current_lr: 0.00034 | > step_time: 0.71700 (0.52063) | > loader_time: 0.00370 (0.00313)  --> STEP: 325/639 -- GLOBAL_STEP: 35525 | > loss: -0.47282 (-0.49026) | > log_mle: -0.59269 (-0.60897) | > loss_dur: 0.11986 (0.11872) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.31232 (9.32833) | > current_lr: 0.00034 | > step_time: 0.62120 (0.53731) | > loader_time: 0.00350 (0.00316)  --> STEP: 350/639 -- GLOBAL_STEP: 35550 | > loss: -0.49582 (-0.48859) | > log_mle: -0.62306 (-0.60731) | > loss_dur: 0.12724 (0.11871) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.48641 (9.24404) | > current_lr: 0.00034 | > step_time: 0.86470 (0.55167) | > loader_time: 0.00350 (0.00319)  --> STEP: 375/639 -- GLOBAL_STEP: 35575 | > loss: -0.44147 (-0.48654) | > log_mle: -0.55729 (-0.60506) | > loss_dur: 0.11581 (0.11852) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.68867 (9.05705) | > current_lr: 0.00034 | > step_time: 0.70870 (0.56477) | > loader_time: 0.00320 (0.00321)  --> STEP: 400/639 -- GLOBAL_STEP: 35600 | > loss: -0.46687 (-0.48522) | > log_mle: -0.58761 (-0.60381) | > loss_dur: 0.12074 (0.11860) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.00318 (9.02058) | > current_lr: 0.00034 | > step_time: 0.72240 (0.57696) | > loader_time: 0.00370 (0.00324)  --> STEP: 425/639 -- GLOBAL_STEP: 35625 | > loss: -0.51879 (-0.48333) | > log_mle: -0.64044 (-0.60194) | > loss_dur: 0.12165 (0.11861) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.53445 (8.93575) | > current_lr: 0.00034 | > step_time: 0.93900 (0.58927) | > loader_time: 0.00390 (0.00327)  --> STEP: 450/639 -- GLOBAL_STEP: 35650 | > loss: -0.43582 (-0.48183) | > log_mle: -0.54409 (-0.60027) | > loss_dur: 0.10827 (0.11844) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.08561 (8.80712) | > current_lr: 0.00034 | > step_time: 0.85870 (0.60093) | > loader_time: 0.00390 (0.00329)  --> STEP: 475/639 -- GLOBAL_STEP: 35675 | > loss: -0.45581 (-0.48016) | > log_mle: -0.55591 (-0.59871) | > loss_dur: 0.10010 (0.11855) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.12457 (8.74073) | > current_lr: 0.00033 | > step_time: 0.75850 (0.61336) | > loader_time: 0.00340 (0.00331)  --> STEP: 500/639 -- GLOBAL_STEP: 35700 | > loss: -0.50168 (-0.47900) | > log_mle: -0.62037 (-0.59765) | > loss_dur: 0.11869 (0.11865) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.57011 (8.66492) | > current_lr: 0.00033 | > step_time: 0.95830 (0.62628) | > loader_time: 0.00390 (0.00334)  --> STEP: 525/639 -- GLOBAL_STEP: 35725 | > loss: -0.46165 (-0.47779) | > log_mle: -0.59795 (-0.59650) | > loss_dur: 0.13631 (0.11871) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.04467 (8.63878) | > current_lr: 0.00033 | > step_time: 0.99820 (0.63916) | > loader_time: 0.00420 (0.00337)  --> STEP: 550/639 -- GLOBAL_STEP: 35750 | > loss: -0.41612 (-0.47651) | > log_mle: -0.53083 (-0.59517) | > loss_dur: 0.11471 (0.11866) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.80691 (8.57908) | > current_lr: 0.00033 | > step_time: 0.95280 (0.65384) | > loader_time: 0.00430 (0.00340)  --> STEP: 575/639 -- GLOBAL_STEP: 35775 | > loss: -0.43030 (-0.47517) | > log_mle: -0.54912 (-0.59388) | > loss_dur: 0.11882 (0.11871) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.91373 (8.51014) | > current_lr: 0.00033 | > step_time: 0.93540 (0.66787) | > loader_time: 0.00470 (0.00344)  --> STEP: 600/639 -- GLOBAL_STEP: 35800 | > loss: -0.42086 (-0.47414) | > log_mle: -0.54745 (-0.59299) | > loss_dur: 0.12659 (0.11884) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.80325 (8.45308) | > current_lr: 0.00033 | > step_time: 1.08410 (0.68489) | > loader_time: 0.00440 (0.00347)  --> STEP: 625/639 -- GLOBAL_STEP: 35825 | > loss: -0.46138 (-0.47323) | > log_mle: -0.57998 (-0.59209) | > loss_dur: 0.11860 (0.11886) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.00169 (8.41118) | > current_lr: 0.00033 | > step_time: 1.11090 (0.70135) | > loader_time: 0.00440 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03409 (+0.00152) | > avg_loss: -0.44296 (+0.00493) | > avg_log_mle: -0.57360 (+0.00338) | > avg_loss_dur: 0.13064 (+0.00155)  > EPOCH: 56/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 21:58:57)   --> STEP: 10/639 -- GLOBAL_STEP: 35850 | > loss: -0.51621 (-0.52228) | > log_mle: -0.64790 (-0.64781) | > loss_dur: 0.13169 (0.12553) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.05773 (10.23046) | > current_lr: 0.00033 | > step_time: 0.28660 (0.30951) | > loader_time: 0.00240 (0.00586)  --> STEP: 35/639 -- GLOBAL_STEP: 35875 | > loss: -0.51022 (-0.51246) | > log_mle: -0.61101 (-0.63145) | > loss_dur: 0.10079 (0.11898) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.37381 (10.46026) | > current_lr: 0.00033 | > step_time: 0.32400 (0.33407) | > loader_time: 0.00270 (0.00358)  --> STEP: 60/639 -- GLOBAL_STEP: 35900 | > loss: -0.45547 (-0.50441) | > log_mle: -0.56672 (-0.62368) | > loss_dur: 0.11126 (0.11926) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.23475 (9.98283) | > current_lr: 0.00033 | > step_time: 0.35810 (0.36771) | > loader_time: 0.00310 (0.00327)  --> STEP: 85/639 -- GLOBAL_STEP: 35925 | > loss: -0.45121 (-0.49825) | > log_mle: -0.56563 (-0.61519) | > loss_dur: 0.11442 (0.11694) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.64063 (9.27368) | > current_lr: 0.00033 | > step_time: 0.41160 (0.38214) | > loader_time: 0.00300 (0.00319)  --> STEP: 110/639 -- GLOBAL_STEP: 35950 | > loss: -0.46410 (-0.49973) | > log_mle: -0.58579 (-0.61725) | > loss_dur: 0.12169 (0.11753) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.00547 (9.60301) | > current_lr: 0.00033 | > step_time: 0.39340 (0.40377) | > loader_time: 0.00310 (0.00317)  --> STEP: 135/639 -- GLOBAL_STEP: 35975 | > loss: -0.44481 (-0.49749) | > log_mle: -0.56740 (-0.61495) | > loss_dur: 0.12259 (0.11746) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.98372 (9.53184) | > current_lr: 0.00033 | > step_time: 0.46970 (0.41659) | > loader_time: 0.00320 (0.00317)  --> STEP: 160/639 -- GLOBAL_STEP: 36000 | > loss: -0.49856 (-0.49570) | > log_mle: -0.60278 (-0.61298) | > loss_dur: 0.10422 (0.11727) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.62343 (9.55268) | > current_lr: 0.00033 | > step_time: 0.57610 (0.43082) | > loader_time: 0.00310 (0.00317)  --> STEP: 185/639 -- GLOBAL_STEP: 36025 | > loss: -0.47766 (-0.49740) | > log_mle: -0.59083 (-0.61517) | > loss_dur: 0.11317 (0.11777) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.41276 (9.86965) | > current_lr: 0.00033 | > step_time: 0.70700 (0.44795) | > loader_time: 0.00320 (0.00318)  --> STEP: 210/639 -- GLOBAL_STEP: 36050 | > loss: -0.40306 (-0.49464) | > log_mle: -0.50939 (-0.61235) | > loss_dur: 0.10633 (0.11771) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.32157 (9.60481) | > current_lr: 0.00033 | > step_time: 0.50950 (0.46409) | > loader_time: 0.00300 (0.00323)  --> STEP: 235/639 -- GLOBAL_STEP: 36075 | > loss: -0.49627 (-0.49511) | > log_mle: -0.60220 (-0.61263) | > loss_dur: 0.10593 (0.11753) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.27443 (9.63699) | > current_lr: 0.00033 | > step_time: 0.56110 (0.48008) | > loader_time: 0.00320 (0.00325)  --> STEP: 260/639 -- GLOBAL_STEP: 36100 | > loss: -0.48770 (-0.49467) | > log_mle: -0.60307 (-0.61232) | > loss_dur: 0.11538 (0.11765) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.92600 (9.75151) | > current_lr: 0.00033 | > step_time: 0.61170 (0.49559) | > loader_time: 0.00330 (0.00327)  --> STEP: 285/639 -- GLOBAL_STEP: 36125 | > loss: -0.43211 (-0.49376) | > log_mle: -0.54889 (-0.61155) | > loss_dur: 0.11678 (0.11779) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.60045 (9.79574) | > current_lr: 0.00033 | > step_time: 0.63900 (0.51123) | > loader_time: 0.00330 (0.00329)  --> STEP: 310/639 -- GLOBAL_STEP: 36150 | > loss: -0.47222 (-0.49229) | > log_mle: -0.59524 (-0.61030) | > loss_dur: 0.12302 (0.11801) | > amp_scaler: 2048.00000 (3963.87097) | > grad_norm: 10.15520 (9.72042) | > current_lr: 0.00033 | > step_time: 0.60410 (0.52670) | > loader_time: 0.00310 (0.00331)  --> STEP: 335/639 -- GLOBAL_STEP: 36175 | > loss: -0.48654 (-0.49013) | > log_mle: -0.61171 (-0.60856) | > loss_dur: 0.12517 (0.11843) | > amp_scaler: 2048.00000 (3820.89552) | > grad_norm: 15.88221 (9.64682) | > current_lr: 0.00033 | > step_time: 0.75040 (0.54160) | > loader_time: 0.00350 (0.00333)  --> STEP: 360/639 -- GLOBAL_STEP: 36200 | > loss: -0.43976 (-0.48863) | > log_mle: -0.55192 (-0.60677) | > loss_dur: 0.11216 (0.11814) | > amp_scaler: 2048.00000 (3697.77778) | > grad_norm: 6.41865 (9.60018) | > current_lr: 0.00033 | > step_time: 0.68990 (0.55448) | > loader_time: 0.00360 (0.00335)  --> STEP: 385/639 -- GLOBAL_STEP: 36225 | > loss: -0.46892 (-0.48718) | > log_mle: -0.58746 (-0.60517) | > loss_dur: 0.11854 (0.11799) | > amp_scaler: 2048.00000 (3590.64935) | > grad_norm: 7.09746 (9.46759) | > current_lr: 0.00033 | > step_time: 0.78770 (0.56858) | > loader_time: 0.00380 (0.00338)  --> STEP: 410/639 -- GLOBAL_STEP: 36250 | > loss: -0.43035 (-0.48554) | > log_mle: -0.56232 (-0.60370) | > loss_dur: 0.13197 (0.11816) | > amp_scaler: 2048.00000 (3496.58537) | > grad_norm: 6.76856 (9.37884) | > current_lr: 0.00033 | > step_time: 0.78670 (0.58111) | > loader_time: 0.00360 (0.00339)  --> STEP: 435/639 -- GLOBAL_STEP: 36275 | > loss: -0.46197 (-0.48374) | > log_mle: -0.57299 (-0.60186) | > loss_dur: 0.11102 (0.11812) | > amp_scaler: 2048.00000 (3413.33333) | > grad_norm: 6.62551 (9.26163) | > current_lr: 0.00033 | > step_time: 0.71520 (0.59427) | > loader_time: 0.00350 (0.00342)  --> STEP: 460/639 -- GLOBAL_STEP: 36300 | > loss: -0.41931 (-0.48198) | > log_mle: -0.53731 (-0.60002) | > loss_dur: 0.11800 (0.11804) | > amp_scaler: 2048.00000 (3339.13043) | > grad_norm: 2.76902 (9.15247) | > current_lr: 0.00033 | > step_time: 0.87410 (0.60584) | > loader_time: 0.00390 (0.00343)  --> STEP: 485/639 -- GLOBAL_STEP: 36325 | > loss: -0.45202 (-0.48043) | > log_mle: -0.56918 (-0.59867) | > loss_dur: 0.11716 (0.11824) | > amp_scaler: 2048.00000 (3272.57732) | > grad_norm: 11.44600 (9.17288) | > current_lr: 0.00033 | > step_time: 0.84500 (0.61880) | > loader_time: 0.00400 (0.00346)  --> STEP: 510/639 -- GLOBAL_STEP: 36350 | > loss: -0.45714 (-0.47932) | > log_mle: -0.57532 (-0.59747) | > loss_dur: 0.11818 (0.11816) | > amp_scaler: 2048.00000 (3212.54902) | > grad_norm: 5.20039 (9.07016) | > current_lr: 0.00033 | > step_time: 0.85740 (0.63221) | > loader_time: 0.00410 (0.00349)  --> STEP: 535/639 -- GLOBAL_STEP: 36375 | > loss: -0.43985 (-0.47804) | > log_mle: -0.56637 (-0.59623) | > loss_dur: 0.12652 (0.11819) | > amp_scaler: 2048.00000 (3158.13084) | > grad_norm: 5.97977 (8.96299) | > current_lr: 0.00033 | > step_time: 1.00090 (0.64533) | > loader_time: 0.00400 (0.00351)  --> STEP: 560/639 -- GLOBAL_STEP: 36400 | > loss: -0.44876 (-0.47671) | > log_mle: -0.57380 (-0.59504) | > loss_dur: 0.12505 (0.11833) | > amp_scaler: 2048.00000 (3108.57143) | > grad_norm: 8.12222 (8.89999) | > current_lr: 0.00033 | > step_time: 0.90910 (0.65986) | > loader_time: 0.00430 (0.00354)  --> STEP: 585/639 -- GLOBAL_STEP: 36425 | > loss: -0.46350 (-0.47567) | > log_mle: -0.57846 (-0.59402) | > loss_dur: 0.11496 (0.11835) | > amp_scaler: 2048.00000 (3063.24786) | > grad_norm: 4.94428 (8.83769) | > current_lr: 0.00033 | > step_time: 1.06440 (0.67532) | > loader_time: 0.00410 (0.00358)  --> STEP: 610/639 -- GLOBAL_STEP: 36450 | > loss: -0.43273 (-0.47460) | > log_mle: -0.54830 (-0.59306) | > loss_dur: 0.11557 (0.11846) | > amp_scaler: 2048.00000 (3021.63934) | > grad_norm: 5.17077 (8.77299) | > current_lr: 0.00033 | > step_time: 1.08430 (0.69165) | > loader_time: 0.00450 (0.00361)  --> STEP: 635/639 -- GLOBAL_STEP: 36475 | > loss: -0.44933 (-0.47382) | > log_mle: -0.57197 (-0.59249) | > loss_dur: 0.12265 (0.11867) | > amp_scaler: 2048.00000 (2983.30709) | > grad_norm: 10.82647 (8.75841) | > current_lr: 0.00033 | > step_time: 1.12050 (0.70906) | > loader_time: 0.00420 (0.00364) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03757 (+0.00348) | > avg_loss: -0.44462 (-0.00166) | > avg_log_mle: -0.57812 (-0.00452) | > avg_loss_dur: 0.13350 (+0.00286)  > EPOCH: 57/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 22:06:47)   --> STEP: 20/639 -- GLOBAL_STEP: 36500 | > loss: -0.59609 (-0.52516) | > log_mle: -0.74527 (-0.64953) | > loss_dur: 0.14918 (0.12437) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 19.06875 (17.87913) | > current_lr: 0.00033 | > step_time: 0.36740 (0.31552) | > loader_time: 0.00310 (0.00261)  --> STEP: 45/639 -- GLOBAL_STEP: 36525 | > loss: -0.46479 (-0.50466) | > log_mle: -0.56629 (-0.62491) | > loss_dur: 0.10150 (0.12025) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.60364 (12.44481) | > current_lr: 0.00033 | > step_time: 0.40910 (0.34892) | > loader_time: 0.00280 (0.00275)  --> STEP: 70/639 -- GLOBAL_STEP: 36550 | > loss: -0.50515 (-0.49761) | > log_mle: -0.63016 (-0.61671) | > loss_dur: 0.12501 (0.11910) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.57117 (10.90829) | > current_lr: 0.00033 | > step_time: 0.42500 (0.37245) | > loader_time: 0.00320 (0.00282)  --> STEP: 95/639 -- GLOBAL_STEP: 36575 | > loss: -0.57394 (-0.49642) | > log_mle: -0.67937 (-0.61335) | > loss_dur: 0.10543 (0.11693) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.00084 (10.00180) | > current_lr: 0.00033 | > step_time: 0.40820 (0.38960) | > loader_time: 0.00310 (0.00286)  --> STEP: 120/639 -- GLOBAL_STEP: 36600 | > loss: -0.53599 (-0.49452) | > log_mle: -0.67141 (-0.61249) | > loss_dur: 0.13542 (0.11797) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.69236 (9.92671) | > current_lr: 0.00033 | > step_time: 0.54570 (0.40788) | > loader_time: 0.00310 (0.00291)  --> STEP: 145/639 -- GLOBAL_STEP: 36625 | > loss: -0.50335 (-0.49529) | > log_mle: -0.62745 (-0.61327) | > loss_dur: 0.12410 (0.11798) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.79174 (10.11611) | > current_lr: 0.00033 | > step_time: 0.47610 (0.42557) | > loader_time: 0.00270 (0.00295)  --> STEP: 170/639 -- GLOBAL_STEP: 36650 | > loss: -0.50124 (-0.49513) | > log_mle: -0.61239 (-0.61281) | > loss_dur: 0.11115 (0.11768) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.18118 (9.93671) | > current_lr: 0.00033 | > step_time: 0.55560 (0.44117) | > loader_time: 0.00340 (0.00298)  --> STEP: 195/639 -- GLOBAL_STEP: 36675 | > loss: -0.45674 (-0.49469) | > log_mle: -0.55266 (-0.61189) | > loss_dur: 0.09592 (0.11720) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.54805 (9.76589) | > current_lr: 0.00033 | > step_time: 0.49930 (0.45850) | > loader_time: 0.00330 (0.00302)  --> STEP: 220/639 -- GLOBAL_STEP: 36700 | > loss: -0.49891 (-0.49265) | > log_mle: -0.60504 (-0.60987) | > loss_dur: 0.10614 (0.11722) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.51296 (9.61828) | > current_lr: 0.00033 | > step_time: 0.73370 (0.47364) | > loader_time: 0.00360 (0.00306)  --> STEP: 245/639 -- GLOBAL_STEP: 36725 | > loss: -0.48030 (-0.49398) | > log_mle: -0.59305 (-0.61102) | > loss_dur: 0.11275 (0.11703) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.98432 (9.84297) | > current_lr: 0.00033 | > step_time: 0.54980 (0.48879) | > loader_time: 0.00320 (0.00309)  --> STEP: 270/639 -- GLOBAL_STEP: 36750 | > loss: -0.52768 (-0.49379) | > log_mle: -0.65470 (-0.61134) | > loss_dur: 0.12702 (0.11755) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.98296 (10.10273) | > current_lr: 0.00033 | > step_time: 0.80230 (0.50624) | > loader_time: 0.00410 (0.00313)  --> STEP: 295/639 -- GLOBAL_STEP: 36775 | > loss: -0.45050 (-0.49199) | > log_mle: -0.55253 (-0.60965) | > loss_dur: 0.10203 (0.11765) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.24717 (9.93784) | > current_lr: 0.00033 | > step_time: 0.69190 (0.52102) | > loader_time: 0.00360 (0.00316)  --> STEP: 320/639 -- GLOBAL_STEP: 36800 | > loss: -0.46727 (-0.49062) | > log_mle: -0.59724 (-0.60859) | > loss_dur: 0.12997 (0.11797) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.42981 (9.92606) | > current_lr: 0.00033 | > step_time: 0.76280 (0.53645) | > loader_time: 0.00380 (0.00319)  --> STEP: 345/639 -- GLOBAL_STEP: 36825 | > loss: -0.47092 (-0.48890) | > log_mle: -0.58578 (-0.60716) | > loss_dur: 0.11486 (0.11825) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.51554 (9.84776) | > current_lr: 0.00033 | > step_time: 0.62430 (0.55095) | > loader_time: 0.00360 (0.00323)  --> STEP: 370/639 -- GLOBAL_STEP: 36850 | > loss: -0.48554 (-0.48741) | > log_mle: -0.60249 (-0.60531) | > loss_dur: 0.11696 (0.11790) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.34636 (9.65517) | > current_lr: 0.00033 | > step_time: 0.82770 (0.56365) | > loader_time: 0.00400 (0.00325)  --> STEP: 395/639 -- GLOBAL_STEP: 36875 | > loss: -0.44485 (-0.48565) | > log_mle: -0.56856 (-0.60352) | > loss_dur: 0.12371 (0.11787) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.95935 (9.53330) | > current_lr: 0.00033 | > step_time: 0.74880 (0.57608) | > loader_time: 0.00330 (0.00328)  --> STEP: 420/639 -- GLOBAL_STEP: 36900 | > loss: -0.45339 (-0.48380) | > log_mle: -0.56524 (-0.60189) | > loss_dur: 0.11185 (0.11809) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.76918 (9.39344) | > current_lr: 0.00033 | > step_time: 0.70240 (0.58947) | > loader_time: 0.00340 (0.00331)  --> STEP: 445/639 -- GLOBAL_STEP: 36925 | > loss: -0.45774 (-0.48258) | > log_mle: -0.57174 (-0.60056) | > loss_dur: 0.11401 (0.11797) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.51024 (9.32880) | > current_lr: 0.00033 | > step_time: 0.90830 (0.60270) | > loader_time: 0.00380 (0.00333)  --> STEP: 470/639 -- GLOBAL_STEP: 36950 | > loss: -0.46464 (-0.48091) | > log_mle: -0.58062 (-0.59883) | > loss_dur: 0.11598 (0.11792) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.92135 (9.17403) | > current_lr: 0.00033 | > step_time: 0.81030 (0.61456) | > loader_time: 0.00360 (0.00336)  --> STEP: 495/639 -- GLOBAL_STEP: 36975 | > loss: -0.41077 (-0.47953) | > log_mle: -0.51851 (-0.59755) | > loss_dur: 0.10774 (0.11802) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.00701 (9.12558) | > current_lr: 0.00033 | > step_time: 0.82390 (0.62796) | > loader_time: 0.00350 (0.00339)  --> STEP: 520/639 -- GLOBAL_STEP: 37000 | > loss: -0.46585 (-0.47861) | > log_mle: -0.57643 (-0.59654) | > loss_dur: 0.11058 (0.11793) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.72782 (9.03594) | > current_lr: 0.00033 | > step_time: 0.92670 (0.64175) | > loader_time: 0.00360 (0.00341)  --> STEP: 545/639 -- GLOBAL_STEP: 37025 | > loss: -0.47673 (-0.47736) | > log_mle: -0.60627 (-0.59546) | > loss_dur: 0.12954 (0.11810) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.40253 (8.93696) | > current_lr: 0.00033 | > step_time: 1.00210 (0.65761) | > loader_time: 0.00400 (0.00345)  --> STEP: 570/639 -- GLOBAL_STEP: 37050 | > loss: -0.44453 (-0.47598) | > log_mle: -0.55533 (-0.59410) | > loss_dur: 0.11080 (0.11812) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.43124 (8.83701) | > current_lr: 0.00033 | > step_time: 0.84230 (0.67178) | > loader_time: 0.00390 (0.00348)  --> STEP: 595/639 -- GLOBAL_STEP: 37075 | > loss: -0.43389 (-0.47509) | > log_mle: -0.54923 (-0.59329) | > loss_dur: 0.11535 (0.11820) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.61165 (8.76535) | > current_lr: 0.00033 | > step_time: 0.98750 (0.68788) | > loader_time: 0.00450 (0.00352)  --> STEP: 620/639 -- GLOBAL_STEP: 37100 | > loss: -0.43890 (-0.47398) | > log_mle: -0.55464 (-0.59224) | > loss_dur: 0.11575 (0.11826) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.45831 (8.71579) | > current_lr: 0.00033 | > step_time: 1.23850 (0.70475) | > loader_time: 0.00500 (0.00356) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02839 (-0.00918) | > avg_loss: -0.45014 (-0.00552) | > avg_log_mle: -0.58162 (-0.00350) | > avg_loss_dur: 0.13148 (-0.00202)  > EPOCH: 58/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 22:14:41)   --> STEP: 5/639 -- GLOBAL_STEP: 37125 | > loss: -0.53143 (-0.52790) | > log_mle: -0.68708 (-0.66223) | > loss_dur: 0.15565 (0.13433) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 16.70946 (10.23914) | > current_lr: 0.00033 | > step_time: 0.30900 (0.29719) | > loader_time: 0.00230 (0.00220)  --> STEP: 30/639 -- GLOBAL_STEP: 37150 | > loss: -0.52341 (-0.51660) | > log_mle: -0.63443 (-0.63960) | > loss_dur: 0.11102 (0.12301) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.01322 (10.26550) | > current_lr: 0.00033 | > step_time: 0.32410 (0.34691) | > loader_time: 0.00270 (0.00256)  --> STEP: 55/639 -- GLOBAL_STEP: 37175 | > loss: -0.47030 (-0.50888) | > log_mle: -0.56866 (-0.62930) | > loss_dur: 0.09836 (0.12042) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.08995 (9.63077) | > current_lr: 0.00033 | > step_time: 0.44240 (0.36780) | > loader_time: 0.00280 (0.00262)  --> STEP: 80/639 -- GLOBAL_STEP: 37200 | > loss: -0.47249 (-0.50217) | > log_mle: -0.60786 (-0.62109) | > loss_dur: 0.13537 (0.11891) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.47693 (9.37389) | > current_lr: 0.00033 | > step_time: 0.39970 (0.38747) | > loader_time: 0.00290 (0.00271)  --> STEP: 105/639 -- GLOBAL_STEP: 37225 | > loss: -0.45309 (-0.50019) | > log_mle: -0.56886 (-0.61871) | > loss_dur: 0.11577 (0.11852) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.52784 (9.45945) | > current_lr: 0.00033 | > step_time: 0.41440 (0.40506) | > loader_time: 0.00300 (0.00278)  --> STEP: 130/639 -- GLOBAL_STEP: 37250 | > loss: -0.47857 (-0.49879) | > log_mle: -0.60547 (-0.61729) | > loss_dur: 0.12690 (0.11850) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.06917 (9.50352) | > current_lr: 0.00033 | > step_time: 0.53240 (0.42021) | > loader_time: 0.00290 (0.00283)  --> STEP: 155/639 -- GLOBAL_STEP: 37275 | > loss: -0.47873 (-0.49798) | > log_mle: -0.59509 (-0.61644) | > loss_dur: 0.11636 (0.11846) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.46819 (9.58113) | > current_lr: 0.00033 | > step_time: 0.54300 (0.43365) | > loader_time: 0.00290 (0.00287)  --> STEP: 180/639 -- GLOBAL_STEP: 37300 | > loss: -0.52586 (-0.49968) | > log_mle: -0.65019 (-0.61804) | > loss_dur: 0.12433 (0.11836) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.70571 (9.69328) | > current_lr: 0.00033 | > step_time: 0.58570 (0.44907) | > loader_time: 0.00330 (0.00291)  --> STEP: 205/639 -- GLOBAL_STEP: 37325 | > loss: -0.46804 (-0.49654) | > log_mle: -0.57392 (-0.61437) | > loss_dur: 0.10588 (0.11783) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.59357 (9.42822) | > current_lr: 0.00033 | > step_time: 0.57210 (0.46270) | > loader_time: 0.00330 (0.00295)  --> STEP: 230/639 -- GLOBAL_STEP: 37350 | > loss: -0.48889 (-0.49671) | > log_mle: -0.60758 (-0.61444) | > loss_dur: 0.11869 (0.11773) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.57217 (9.56930) | > current_lr: 0.00033 | > step_time: 0.57340 (0.47956) | > loader_time: 0.00340 (0.00300)  --> STEP: 255/639 -- GLOBAL_STEP: 37375 | > loss: -0.44871 (-0.49585) | > log_mle: -0.57427 (-0.61351) | > loss_dur: 0.12556 (0.11766) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.78777 (9.50938) | > current_lr: 0.00033 | > step_time: 0.63130 (0.49369) | > loader_time: 0.00350 (0.00304)  --> STEP: 280/639 -- GLOBAL_STEP: 37400 | > loss: -0.47501 (-0.49577) | > log_mle: -0.60231 (-0.61386) | > loss_dur: 0.12730 (0.11808) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.16460 (9.58493) | > current_lr: 0.00033 | > step_time: 0.65000 (0.51168) | > loader_time: 0.00340 (0.00308)  --> STEP: 305/639 -- GLOBAL_STEP: 37425 | > loss: -0.46005 (-0.49413) | > log_mle: -0.58265 (-0.61225) | > loss_dur: 0.12259 (0.11812) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.09981 (9.48403) | > current_lr: 0.00033 | > step_time: 0.67600 (0.52628) | > loader_time: 0.00320 (0.00311)  --> STEP: 330/639 -- GLOBAL_STEP: 37450 | > loss: -0.47011 (-0.49179) | > log_mle: -0.60416 (-0.61040) | > loss_dur: 0.13405 (0.11861) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.28312 (9.42148) | > current_lr: 0.00033 | > step_time: 0.71680 (0.54160) | > loader_time: 0.00360 (0.00315)  --> STEP: 355/639 -- GLOBAL_STEP: 37475 | > loss: -0.42264 (-0.49032) | > log_mle: -0.53472 (-0.60865) | > loss_dur: 0.11208 (0.11833) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.26421 (9.30278) | > current_lr: 0.00033 | > step_time: 0.60210 (0.55468) | > loader_time: 0.00350 (0.00318)  --> STEP: 380/639 -- GLOBAL_STEP: 37500 | > loss: -0.46930 (-0.48850) | > log_mle: -0.57025 (-0.60664) | > loss_dur: 0.10095 (0.11814) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.48236 (9.16866) | > current_lr: 0.00033 | > step_time: 0.63580 (0.56905) | > loader_time: 0.00330 (0.00321)  --> STEP: 405/639 -- GLOBAL_STEP: 37525 | > loss: -0.45929 (-0.48726) | > log_mle: -0.57600 (-0.60559) | > loss_dur: 0.11672 (0.11833) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.84206 (9.17522) | > current_lr: 0.00033 | > step_time: 0.72350 (0.58170) | > loader_time: 0.00350 (0.00324)  --> STEP: 430/639 -- GLOBAL_STEP: 37550 | > loss: -0.42928 (-0.48553) | > log_mle: -0.53543 (-0.60388) | > loss_dur: 0.10614 (0.11835) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.73967 (9.06963) | > current_lr: 0.00033 | > step_time: 0.82080 (0.59480) | > loader_time: 0.00350 (0.00327)  --> STEP: 455/639 -- GLOBAL_STEP: 37575 | > loss: -0.44697 (-0.48383) | > log_mle: -0.56242 (-0.60206) | > loss_dur: 0.11545 (0.11822) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.07698 (8.92402) | > current_lr: 0.00033 | > step_time: 0.74550 (0.60508) | > loader_time: 0.00340 (0.00329)  --> STEP: 480/639 -- GLOBAL_STEP: 37600 | > loss: -0.39935 (-0.48227) | > log_mle: -0.51776 (-0.60063) | > loss_dur: 0.11840 (0.11836) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.42930 (8.84538) | > current_lr: 0.00033 | > step_time: 0.90190 (0.61880) | > loader_time: 0.00360 (0.00332)  --> STEP: 505/639 -- GLOBAL_STEP: 37625 | > loss: -0.46725 (-0.48102) | > log_mle: -0.58306 (-0.59931) | > loss_dur: 0.11582 (0.11829) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.95925 (8.78362) | > current_lr: 0.00033 | > step_time: 0.93230 (0.63198) | > loader_time: 0.00400 (0.00335)  --> STEP: 530/639 -- GLOBAL_STEP: 37650 | > loss: -0.44140 (-0.47985) | > log_mle: -0.57521 (-0.59821) | > loss_dur: 0.13382 (0.11836) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.40492 (8.73843) | > current_lr: 0.00033 | > step_time: 0.94400 (0.64618) | > loader_time: 0.00390 (0.00338)  --> STEP: 555/639 -- GLOBAL_STEP: 37675 | > loss: -0.44136 (-0.47851) | > log_mle: -0.55841 (-0.59684) | > loss_dur: 0.11705 (0.11833) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.19417 (8.63529) | > current_lr: 0.00033 | > step_time: 0.84850 (0.66018) | > loader_time: 0.00370 (0.00342)  --> STEP: 580/639 -- GLOBAL_STEP: 37700 | > loss: -0.46397 (-0.47741) | > log_mle: -0.57584 (-0.59579) | > loss_dur: 0.11187 (0.11838) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.04281 (8.55458) | > current_lr: 0.00033 | > step_time: 1.00700 (0.67580) | > loader_time: 0.00430 (0.00346)  --> STEP: 605/639 -- GLOBAL_STEP: 37725 | > loss: -0.43658 (-0.47634) | > log_mle: -0.54474 (-0.59485) | > loss_dur: 0.10816 (0.11851) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.03765 (8.49432) | > current_lr: 0.00033 | > step_time: 1.05130 (0.69137) | > loader_time: 0.00490 (0.00349)  --> STEP: 630/639 -- GLOBAL_STEP: 37750 | > loss: -0.44429 (-0.47543) | > log_mle: -0.57492 (-0.59406) | > loss_dur: 0.13063 (0.11864) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.16613 (8.45796) | > current_lr: 0.00033 | > step_time: 1.40820 (0.70940) | > loader_time: 0.00460 (0.00354) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03255 (+0.00415) | > avg_loss: -0.44853 (+0.00160) | > avg_log_mle: -0.58183 (-0.00021) | > avg_loss_dur: 0.13329 (+0.00181)  > EPOCH: 59/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 22:22:33)   --> STEP: 15/639 -- GLOBAL_STEP: 37775 | > loss: -0.63942 (-0.53359) | > log_mle: -0.77142 (-0.65563) | > loss_dur: 0.13200 (0.12204) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 20.02432 (10.36280) | > current_lr: 0.00033 | > step_time: 0.33230 (0.30590) | > loader_time: 0.00280 (0.00371)  --> STEP: 40/639 -- GLOBAL_STEP: 37800 | > loss: -0.49262 (-0.51359) | > log_mle: -0.61053 (-0.63361) | > loss_dur: 0.11792 (0.12003) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.80196 (9.08414) | > current_lr: 0.00033 | > step_time: 0.41230 (0.34492) | > loader_time: 0.00280 (0.00313)  --> STEP: 65/639 -- GLOBAL_STEP: 37825 | > loss: -0.49763 (-0.50635) | > log_mle: -0.59606 (-0.62621) | > loss_dur: 0.09843 (0.11986) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.77422 (9.03940) | > current_lr: 0.00033 | > step_time: 0.43080 (0.37838) | > loader_time: 0.00310 (0.00308)  --> STEP: 90/639 -- GLOBAL_STEP: 37850 | > loss: -0.43871 (-0.49947) | > log_mle: -0.55707 (-0.61747) | > loss_dur: 0.11835 (0.11800) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.71145 (8.63256) | > current_lr: 0.00033 | > step_time: 0.52210 (0.39467) | > loader_time: 0.00300 (0.00306)  --> STEP: 115/639 -- GLOBAL_STEP: 37875 | > loss: -0.40610 (-0.49985) | > log_mle: -0.52463 (-0.61805) | > loss_dur: 0.11853 (0.11820) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.54048 (9.04515) | > current_lr: 0.00033 | > step_time: 0.48900 (0.41396) | > loader_time: 0.00300 (0.00306)  --> STEP: 140/639 -- GLOBAL_STEP: 37900 | > loss: -0.46849 (-0.50020) | > log_mle: -0.60430 (-0.61834) | > loss_dur: 0.13581 (0.11814) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.04118 (9.47030) | > current_lr: 0.00032 | > step_time: 0.62810 (0.42838) | > loader_time: 0.00340 (0.00307)  --> STEP: 165/639 -- GLOBAL_STEP: 37925 | > loss: -0.47910 (-0.49973) | > log_mle: -0.61716 (-0.61755) | > loss_dur: 0.13806 (0.11782) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.15310 (9.51706) | > current_lr: 0.00032 | > step_time: 0.49470 (0.44325) | > loader_time: 0.00330 (0.00308)  --> STEP: 190/639 -- GLOBAL_STEP: 37950 | > loss: -0.42022 (-0.49947) | > log_mle: -0.54173 (-0.61706) | > loss_dur: 0.12151 (0.11759) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.69494 (9.77047) | > current_lr: 0.00032 | > step_time: 0.54440 (0.46109) | > loader_time: 0.00330 (0.00312)  --> STEP: 215/639 -- GLOBAL_STEP: 37975 | > loss: -0.52947 (-0.49725) | > log_mle: -0.64346 (-0.61489) | > loss_dur: 0.11398 (0.11765) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.05027 (9.77025) | > current_lr: 0.00032 | > step_time: 0.61980 (0.47625) | > loader_time: 0.00320 (0.00315)  --> STEP: 240/639 -- GLOBAL_STEP: 38000 | > loss: -0.54351 (-0.49802) | > log_mle: -0.65609 (-0.61543) | > loss_dur: 0.11258 (0.11741) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.93306 (10.00996) | > current_lr: 0.00032 | > step_time: 0.67450 (0.49115) | > loader_time: 0.00310 (0.00317)  --> STEP: 265/639 -- GLOBAL_STEP: 38025 | > loss: -0.55284 (-0.49700) | > log_mle: -0.67863 (-0.61470) | > loss_dur: 0.12579 (0.11770) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.68487 (10.15315) | > current_lr: 0.00032 | > step_time: 0.69060 (0.50782) | > loader_time: 0.00340 (0.00320)  --> STEP: 290/639 -- GLOBAL_STEP: 38050 | > loss: -0.46996 (-0.49550) | > log_mle: -0.57783 (-0.61321) | > loss_dur: 0.10787 (0.11771) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.08269 (10.18285) | > current_lr: 0.00032 | > step_time: 0.66480 (0.52374) | > loader_time: 0.00320 (0.00322)  --> STEP: 315/639 -- GLOBAL_STEP: 38075 | > loss: -0.43980 (-0.49398) | > log_mle: -0.55590 (-0.61185) | > loss_dur: 0.11609 (0.11787) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.68551 (10.19239) | > current_lr: 0.00032 | > step_time: 0.64170 (0.53806) | > loader_time: 0.00330 (0.00325)  --> STEP: 340/639 -- GLOBAL_STEP: 38100 | > loss: -0.47139 (-0.49230) | > log_mle: -0.58077 (-0.61065) | > loss_dur: 0.10938 (0.11835) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.26283 (10.19300) | > current_lr: 0.00032 | > step_time: 0.60800 (0.55341) | > loader_time: 0.00350 (0.00328)  --> STEP: 365/639 -- GLOBAL_STEP: 38125 | > loss: -0.47166 (-0.49035) | > log_mle: -0.58722 (-0.60842) | > loss_dur: 0.11556 (0.11806) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.98645 (9.98908) | > current_lr: 0.00032 | > step_time: 0.81790 (0.56477) | > loader_time: 0.00390 (0.00330)  --> STEP: 390/639 -- GLOBAL_STEP: 38150 | > loss: -0.45867 (-0.48866) | > log_mle: -0.58458 (-0.60658) | > loss_dur: 0.12591 (0.11792) | > amp_scaler: 4096.00000 (2153.02564) | > grad_norm: 7.21889 (9.76236) | > current_lr: 0.00032 | > step_time: 0.75340 (0.57816) | > loader_time: 0.00370 (0.00333)  --> STEP: 415/639 -- GLOBAL_STEP: 38175 | > loss: -0.43400 (-0.48683) | > log_mle: -0.55813 (-0.60497) | > loss_dur: 0.12413 (0.11814) | > amp_scaler: 4096.00000 (2270.07229) | > grad_norm: 5.05911 (9.73978) | > current_lr: 0.00032 | > step_time: 0.74990 (0.59087) | > loader_time: 0.00350 (0.00336)  --> STEP: 440/639 -- GLOBAL_STEP: 38200 | > loss: -0.42736 (-0.48537) | > log_mle: -0.54473 (-0.60349) | > loss_dur: 0.11737 (0.11811) | > amp_scaler: 4096.00000 (2373.81818) | > grad_norm: 6.68953 (9.65757) | > current_lr: 0.00032 | > step_time: 0.76240 (0.60286) | > loader_time: 0.00380 (0.00338)  --> STEP: 465/639 -- GLOBAL_STEP: 38225 | > loss: -0.43772 (-0.48361) | > log_mle: -0.54916 (-0.60157) | > loss_dur: 0.11144 (0.11796) | > amp_scaler: 4096.00000 (2466.40860) | > grad_norm: 4.79069 (9.48860) | > current_lr: 0.00032 | > step_time: 0.84120 (0.61327) | > loader_time: 0.00370 (0.00341)  --> STEP: 490/639 -- GLOBAL_STEP: 38250 | > loss: -0.43045 (-0.48210) | > log_mle: -0.53082 (-0.60016) | > loss_dur: 0.10037 (0.11806) | > amp_scaler: 4096.00000 (2549.55102) | > grad_norm: 3.17003 (9.39197) | > current_lr: 0.00032 | > step_time: 0.75650 (0.62582) | > loader_time: 0.00400 (0.00343)  --> STEP: 515/639 -- GLOBAL_STEP: 38275 | > loss: -0.44708 (-0.48119) | > log_mle: -0.56121 (-0.59918) | > loss_dur: 0.11412 (0.11799) | > amp_scaler: 4096.00000 (2624.62136) | > grad_norm: 10.63503 (9.32831) | > current_lr: 0.00032 | > step_time: 0.99430 (0.63977) | > loader_time: 0.00460 (0.00346)  --> STEP: 540/639 -- GLOBAL_STEP: 38300 | > loss: -0.45496 (-0.47989) | > log_mle: -0.57893 (-0.59792) | > loss_dur: 0.12397 (0.11803) | > amp_scaler: 4096.00000 (2692.74074) | > grad_norm: 5.98466 (9.23652) | > current_lr: 0.00032 | > step_time: 0.92190 (0.65425) | > loader_time: 0.00440 (0.00349)  --> STEP: 565/639 -- GLOBAL_STEP: 38325 | > loss: -0.42329 (-0.47846) | > log_mle: -0.53997 (-0.59655) | > loss_dur: 0.11668 (0.11810) | > amp_scaler: 4096.00000 (2754.83186) | > grad_norm: 3.77407 (9.15966) | > current_lr: 0.00032 | > step_time: 1.01850 (0.66806) | > loader_time: 0.00430 (0.00353)  --> STEP: 590/639 -- GLOBAL_STEP: 38350 | > loss: -0.45232 (-0.47763) | > log_mle: -0.57015 (-0.59577) | > loss_dur: 0.11783 (0.11814) | > amp_scaler: 4096.00000 (2811.66102) | > grad_norm: 7.31432 (9.09368) | > current_lr: 0.00032 | > step_time: 0.99190 (0.68387) | > loader_time: 0.00660 (0.00358)  --> STEP: 615/639 -- GLOBAL_STEP: 38375 | > loss: -0.44939 (-0.47651) | > log_mle: -0.57400 (-0.59472) | > loss_dur: 0.12460 (0.11821) | > amp_scaler: 4096.00000 (2863.86992) | > grad_norm: 9.19757 (9.00921) | > current_lr: 0.00032 | > step_time: 1.02710 (0.69973) | > loader_time: 0.00450 (0.00362) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02177 (-0.01078) | > avg_loss: -0.44907 (-0.00053) | > avg_log_mle: -0.58227 (-0.00044) | > avg_loss_dur: 0.13320 (-0.00009)  > EPOCH: 60/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 22:30:26)   --> STEP: 0/639 -- GLOBAL_STEP: 38400 | > loss: -0.51770 (-0.51770) | > log_mle: -0.70582 (-0.70582) | > loss_dur: 0.18812 (0.18812) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 17.63295 (17.63295) | > current_lr: 0.00032 | > step_time: 0.68400 (0.68397) | > loader_time: 0.43030 (0.43026)  --> STEP: 25/639 -- GLOBAL_STEP: 38425 | > loss: -0.52535 (-0.52927) | > log_mle: -0.65368 (-0.65044) | > loss_dur: 0.12833 (0.12117) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.09452 (10.03333) | > current_lr: 0.00032 | > step_time: 0.44750 (0.33387) | > loader_time: 0.00300 (0.00438)  --> STEP: 50/639 -- GLOBAL_STEP: 38450 | > loss: -0.48104 (-0.51516) | > log_mle: -0.57831 (-0.63341) | > loss_dur: 0.09727 (0.11825) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.22186 (9.38676) | > current_lr: 0.00032 | > step_time: 0.33700 (0.35963) | > loader_time: 0.00310 (0.00363)  --> STEP: 75/639 -- GLOBAL_STEP: 38475 | > loss: -0.50511 (-0.50610) | > log_mle: -0.60933 (-0.62377) | > loss_dur: 0.10422 (0.11767) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.34867 (8.92847) | > current_lr: 0.00032 | > step_time: 0.44280 (0.38011) | > loader_time: 0.00300 (0.00342)  --> STEP: 100/639 -- GLOBAL_STEP: 38500 | > loss: -0.50613 (-0.50395) | > log_mle: -0.61646 (-0.62065) | > loss_dur: 0.11033 (0.11671) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.01096 (8.87791) | > current_lr: 0.00032 | > step_time: 0.52920 (0.40353) | > loader_time: 0.00330 (0.00332)  --> STEP: 125/639 -- GLOBAL_STEP: 38525 | > loss: -0.48830 (-0.50163) | > log_mle: -0.61485 (-0.61910) | > loss_dur: 0.12655 (0.11747) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.40221 (9.03202) | > current_lr: 0.00032 | > step_time: 0.41840 (0.41555) | > loader_time: 0.00290 (0.00325)  --> STEP: 150/639 -- GLOBAL_STEP: 38550 | > loss: -0.45597 (-0.50071) | > log_mle: -0.57917 (-0.61840) | > loss_dur: 0.12321 (0.11769) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.75049 (9.21988) | > current_lr: 0.00032 | > step_time: 0.42090 (0.43216) | > loader_time: 0.00310 (0.00323)  --> STEP: 175/639 -- GLOBAL_STEP: 38575 | > loss: -0.49835 (-0.50109) | > log_mle: -0.59535 (-0.61855) | > loss_dur: 0.09700 (0.11745) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.35861 (9.41761) | > current_lr: 0.00032 | > step_time: 0.59540 (0.44724) | > loader_time: 0.00360 (0.00322)  --> STEP: 200/639 -- GLOBAL_STEP: 38600 | > loss: -0.46535 (-0.49916) | > log_mle: -0.58244 (-0.61609) | > loss_dur: 0.11709 (0.11694) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.25117 (9.31157) | > current_lr: 0.00032 | > step_time: 0.53350 (0.46057) | > loader_time: 0.00310 (0.00323)  --> STEP: 225/639 -- GLOBAL_STEP: 38625 | > loss: -0.45922 (-0.49732) | > log_mle: -0.56091 (-0.61412) | > loss_dur: 0.10170 (0.11680) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.33551 (9.21788) | > current_lr: 0.00032 | > step_time: 0.54520 (0.47571) | > loader_time: 0.00330 (0.00325)  --> STEP: 250/639 -- GLOBAL_STEP: 38650 | > loss: -0.46255 (-0.49816) | > log_mle: -0.56117 (-0.61519) | > loss_dur: 0.09862 (0.11702) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.81010 (9.39462) | > current_lr: 0.00032 | > step_time: 0.59000 (0.49071) | > loader_time: 0.00350 (0.00326)  --> STEP: 275/639 -- GLOBAL_STEP: 38675 | > loss: -0.48610 (-0.49827) | > log_mle: -0.60318 (-0.61565) | > loss_dur: 0.11708 (0.11738) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.30652 (9.63880) | > current_lr: 0.00032 | > step_time: 1.00310 (0.50930) | > loader_time: 0.00400 (0.00328)  --> STEP: 300/639 -- GLOBAL_STEP: 38700 | > loss: -0.47836 (-0.49625) | > log_mle: -0.59534 (-0.61375) | > loss_dur: 0.11698 (0.11750) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.11706 (9.58146) | > current_lr: 0.00032 | > step_time: 0.71790 (0.52253) | > loader_time: 0.00480 (0.00331)  --> STEP: 325/639 -- GLOBAL_STEP: 38725 | > loss: -0.46762 (-0.49431) | > log_mle: -0.59368 (-0.61223) | > loss_dur: 0.12606 (0.11792) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.36188 (9.57850) | > current_lr: 0.00032 | > step_time: 0.62630 (0.53786) | > loader_time: 0.00320 (0.00333)  --> STEP: 350/639 -- GLOBAL_STEP: 38750 | > loss: -0.49973 (-0.49253) | > log_mle: -0.62316 (-0.61039) | > loss_dur: 0.12343 (0.11786) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.97491 (9.54638) | > current_lr: 0.00032 | > step_time: 0.87680 (0.55027) | > loader_time: 0.00390 (0.00335)  --> STEP: 375/639 -- GLOBAL_STEP: 38775 | > loss: -0.44419 (-0.49037) | > log_mle: -0.55898 (-0.60812) | > loss_dur: 0.11479 (0.11775) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.20980 (9.38832) | > current_lr: 0.00032 | > step_time: 0.68430 (0.56359) | > loader_time: 0.00360 (0.00337)  --> STEP: 400/639 -- GLOBAL_STEP: 38800 | > loss: -0.47670 (-0.48888) | > log_mle: -0.58816 (-0.60671) | > loss_dur: 0.11146 (0.11783) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.53068 (9.32272) | > current_lr: 0.00032 | > step_time: 0.71350 (0.57613) | > loader_time: 0.00360 (0.00339)  --> STEP: 425/639 -- GLOBAL_STEP: 38825 | > loss: -0.52375 (-0.48695) | > log_mle: -0.64778 (-0.60482) | > loss_dur: 0.12402 (0.11787) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.11454 (9.26630) | > current_lr: 0.00032 | > step_time: 0.90560 (0.58802) | > loader_time: 0.00380 (0.00341)  --> STEP: 450/639 -- GLOBAL_STEP: 38850 | > loss: -0.44023 (-0.48538) | > log_mle: -0.54628 (-0.60312) | > loss_dur: 0.10605 (0.11774) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.53044 (9.15414) | > current_lr: 0.00032 | > step_time: 0.82020 (0.59956) | > loader_time: 0.00380 (0.00343)  --> STEP: 475/639 -- GLOBAL_STEP: 38875 | > loss: -0.45272 (-0.48370) | > log_mle: -0.55809 (-0.60155) | > loss_dur: 0.10537 (0.11785) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.57511 (9.12022) | > current_lr: 0.00032 | > step_time: 0.72180 (0.61161) | > loader_time: 0.00370 (0.00345)  --> STEP: 500/639 -- GLOBAL_STEP: 38900 | > loss: -0.50543 (-0.48258) | > log_mle: -0.62067 (-0.60047) | > loss_dur: 0.11524 (0.11789) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.15121 (9.04940) | > current_lr: 0.00032 | > step_time: 0.97070 (0.62552) | > loader_time: 0.00390 (0.00347)  --> STEP: 525/639 -- GLOBAL_STEP: 38925 | > loss: -0.47314 (-0.48136) | > log_mle: -0.59951 (-0.59933) | > loss_dur: 0.12636 (0.11798) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.32243 (9.00448) | > current_lr: 0.00032 | > step_time: 1.07830 (0.63912) | > loader_time: 0.00440 (0.00351)  --> STEP: 550/639 -- GLOBAL_STEP: 38950 | > loss: -0.41728 (-0.47997) | > log_mle: -0.53454 (-0.59795) | > loss_dur: 0.11727 (0.11798) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.06584 (8.92146) | > current_lr: 0.00032 | > step_time: 0.92740 (0.65344) | > loader_time: 0.00420 (0.00354)  --> STEP: 575/639 -- GLOBAL_STEP: 38975 | > loss: -0.43641 (-0.47861) | > log_mle: -0.55053 (-0.59664) | > loss_dur: 0.11412 (0.11803) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.99160 (8.87383) | > current_lr: 0.00032 | > step_time: 1.07930 (0.66793) | > loader_time: 0.00420 (0.00357)  --> STEP: 600/639 -- GLOBAL_STEP: 39000 | > loss: -0.42818 (-0.47752) | > log_mle: -0.55031 (-0.59567) | > loss_dur: 0.12214 (0.11814) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.77665 (8.81312) | > current_lr: 0.00032 | > step_time: 1.04120 (0.68429) | > loader_time: 0.00460 (0.00360)  --> STEP: 625/639 -- GLOBAL_STEP: 39025 | > loss: -0.45943 (-0.47650) | > log_mle: -0.57794 (-0.59473) | > loss_dur: 0.11851 (0.11822) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.57228 (8.77761) | > current_lr: 0.00032 | > step_time: 1.08530 (0.70127) | > loader_time: 0.00450 (0.00364) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02671 (+0.00495) | > avg_loss: -0.44555 (+0.00351) | > avg_log_mle: -0.58067 (+0.00160) | > avg_loss_dur: 0.13511 (+0.00191)  > EPOCH: 61/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 22:38:16)   --> STEP: 10/639 -- GLOBAL_STEP: 39050 | > loss: -0.51436 (-0.52428) | > log_mle: -0.65313 (-0.65270) | > loss_dur: 0.13877 (0.12842) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.55661 (11.36255) | > current_lr: 0.00032 | > step_time: 0.28280 (0.30926) | > loader_time: 0.00300 (0.00618)  --> STEP: 35/639 -- GLOBAL_STEP: 39075 | > loss: -0.52320 (-0.51520) | > log_mle: -0.61530 (-0.63662) | > loss_dur: 0.09210 (0.12142) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.20197 (10.36507) | > current_lr: 0.00032 | > step_time: 0.39340 (0.34961) | > loader_time: 0.00270 (0.00375)  --> STEP: 60/639 -- GLOBAL_STEP: 39100 | > loss: -0.46470 (-0.50893) | > log_mle: -0.57339 (-0.62903) | > loss_dur: 0.10869 (0.12010) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.35197 (9.76105) | > current_lr: 0.00032 | > step_time: 0.35400 (0.37424) | > loader_time: 0.00290 (0.00340)  --> STEP: 85/639 -- GLOBAL_STEP: 39125 | > loss: -0.46020 (-0.50312) | > log_mle: -0.57167 (-0.62050) | > loss_dur: 0.11147 (0.11737) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.37582 (8.99920) | > current_lr: 0.00032 | > step_time: 0.41390 (0.38812) | > loader_time: 0.00330 (0.00331)  --> STEP: 110/639 -- GLOBAL_STEP: 39150 | > loss: -0.48223 (-0.50422) | > log_mle: -0.59020 (-0.62233) | > loss_dur: 0.10797 (0.11811) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.17299 (9.50433) | > current_lr: 0.00032 | > step_time: 0.39280 (0.41068) | > loader_time: 0.00320 (0.00325)  --> STEP: 135/639 -- GLOBAL_STEP: 39175 | > loss: -0.45191 (-0.50191) | > log_mle: -0.56913 (-0.61962) | > loss_dur: 0.11721 (0.11771) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.91535 (9.52781) | > current_lr: 0.00032 | > step_time: 0.46160 (0.42139) | > loader_time: 0.00330 (0.00321)  --> STEP: 160/639 -- GLOBAL_STEP: 39200 | > loss: -0.50700 (-0.50016) | > log_mle: -0.61274 (-0.61732) | > loss_dur: 0.10575 (0.11717) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.10809 (9.51826) | > current_lr: 0.00032 | > step_time: 0.46490 (0.43454) | > loader_time: 0.00290 (0.00321)  --> STEP: 185/639 -- GLOBAL_STEP: 39225 | > loss: -0.47687 (-0.50130) | > log_mle: -0.58699 (-0.61902) | > loss_dur: 0.11012 (0.11772) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.65433 (9.80862) | > current_lr: 0.00032 | > step_time: 0.58460 (0.45134) | > loader_time: 0.00330 (0.00322)  --> STEP: 210/639 -- GLOBAL_STEP: 39250 | > loss: -0.40824 (-0.49848) | > log_mle: -0.51196 (-0.61595) | > loss_dur: 0.10371 (0.11746) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.19531 (9.58565) | > current_lr: 0.00032 | > step_time: 0.52620 (0.46774) | > loader_time: 0.00310 (0.00323)  --> STEP: 235/639 -- GLOBAL_STEP: 39275 | > loss: -0.49814 (-0.49908) | > log_mle: -0.60766 (-0.61637) | > loss_dur: 0.10952 (0.11728) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.24743 (9.71675) | > current_lr: 0.00032 | > step_time: 0.56730 (0.48498) | > loader_time: 0.00340 (0.00325)  --> STEP: 260/639 -- GLOBAL_STEP: 39300 | > loss: -0.48368 (-0.49868) | > log_mle: -0.60226 (-0.61617) | > loss_dur: 0.11858 (0.11748) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.06133 (9.75565) | > current_lr: 0.00032 | > step_time: 0.61660 (0.50042) | > loader_time: 0.00350 (0.00327)  --> STEP: 285/639 -- GLOBAL_STEP: 39325 | > loss: -0.43690 (-0.49744) | > log_mle: -0.54989 (-0.61512) | > loss_dur: 0.11300 (0.11767) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.00956 (9.65890) | > current_lr: 0.00032 | > step_time: 0.61430 (0.51740) | > loader_time: 0.00340 (0.00330)  --> STEP: 310/639 -- GLOBAL_STEP: 39350 | > loss: -0.47397 (-0.49592) | > log_mle: -0.59324 (-0.61367) | > loss_dur: 0.11927 (0.11775) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.97442 (9.69742) | > current_lr: 0.00032 | > step_time: 0.58310 (0.53141) | > loader_time: 0.00320 (0.00332)  --> STEP: 335/639 -- GLOBAL_STEP: 39375 | > loss: -0.49825 (-0.49358) | > log_mle: -0.62051 (-0.61173) | > loss_dur: 0.12226 (0.11815) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.87847 (9.66743) | > current_lr: 0.00032 | > step_time: 0.67770 (0.54544) | > loader_time: 0.00360 (0.00334)  --> STEP: 360/639 -- GLOBAL_STEP: 39400 | > loss: -0.44174 (-0.49204) | > log_mle: -0.55473 (-0.60991) | > loss_dur: 0.11298 (0.11787) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.88774 (9.53739) | > current_lr: 0.00032 | > step_time: 0.70950 (0.55732) | > loader_time: 0.00370 (0.00336)  --> STEP: 385/639 -- GLOBAL_STEP: 39425 | > loss: -0.47737 (-0.49041) | > log_mle: -0.59193 (-0.60818) | > loss_dur: 0.11455 (0.11777) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.19149 (9.33088) | > current_lr: 0.00032 | > step_time: 0.73010 (0.57050) | > loader_time: 0.00350 (0.00338)  --> STEP: 410/639 -- GLOBAL_STEP: 39450 | > loss: -0.43469 (-0.48873) | > log_mle: -0.56629 (-0.60671) | > loss_dur: 0.13159 (0.11798) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.51074 (9.25633) | > current_lr: 0.00032 | > step_time: 0.77890 (0.58163) | > loader_time: 0.00390 (0.00340)  --> STEP: 435/639 -- GLOBAL_STEP: 39475 | > loss: -0.46215 (-0.48695) | > log_mle: -0.57527 (-0.60484) | > loss_dur: 0.11312 (0.11789) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.02096 (9.11937) | > current_lr: 0.00032 | > step_time: 0.72530 (0.59334) | > loader_time: 0.00360 (0.00342)  --> STEP: 460/639 -- GLOBAL_STEP: 39500 | > loss: -0.41802 (-0.48513) | > log_mle: -0.53954 (-0.60296) | > loss_dur: 0.12152 (0.11783) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.94480 (8.99471) | > current_lr: 0.00032 | > step_time: 0.87830 (0.60477) | > loader_time: 0.00400 (0.00344)  --> STEP: 485/639 -- GLOBAL_STEP: 39525 | > loss: -0.45673 (-0.48366) | > log_mle: -0.57202 (-0.60167) | > loss_dur: 0.11529 (0.11801) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.66602 (8.90708) | > current_lr: 0.00032 | > step_time: 0.82280 (0.61761) | > loader_time: 0.00420 (0.00346)  --> STEP: 510/639 -- GLOBAL_STEP: 39550 | > loss: -0.46607 (-0.48258) | > log_mle: -0.58081 (-0.60054) | > loss_dur: 0.11474 (0.11796) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.84175 (8.86055) | > current_lr: 0.00032 | > step_time: 0.85120 (0.63038) | > loader_time: 0.00380 (0.00348)  --> STEP: 535/639 -- GLOBAL_STEP: 39575 | > loss: -0.44426 (-0.48131) | > log_mle: -0.56967 (-0.59931) | > loss_dur: 0.12541 (0.11800) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.54745 (8.79787) | > current_lr: 0.00032 | > step_time: 0.92120 (0.64348) | > loader_time: 0.00420 (0.00350)  --> STEP: 560/639 -- GLOBAL_STEP: 39600 | > loss: -0.44785 (-0.48003) | > log_mle: -0.57263 (-0.59813) | > loss_dur: 0.12478 (0.11810) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.10695 (8.74383) | > current_lr: 0.00032 | > step_time: 1.03260 (0.65833) | > loader_time: 0.00420 (0.00354)  --> STEP: 585/639 -- GLOBAL_STEP: 39625 | > loss: -0.46281 (-0.47897) | > log_mle: -0.57870 (-0.59707) | > loss_dur: 0.11589 (0.11810) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.67658 (8.64863) | > current_lr: 0.00032 | > step_time: 1.06320 (0.67313) | > loader_time: 0.00450 (0.00357)  --> STEP: 610/639 -- GLOBAL_STEP: 39650 | > loss: -0.43572 (-0.47787) | > log_mle: -0.54833 (-0.59606) | > loss_dur: 0.11261 (0.11818) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.97568 (8.61826) | > current_lr: 0.00032 | > step_time: 1.06230 (0.68919) | > loader_time: 0.00460 (0.00360)  --> STEP: 635/639 -- GLOBAL_STEP: 39675 | > loss: -0.44869 (-0.47705) | > log_mle: -0.57547 (-0.59546) | > loss_dur: 0.12678 (0.11842) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.98595 (8.61674) | > current_lr: 0.00032 | > step_time: 1.13550 (0.70607) | > loader_time: 0.00390 (0.00364) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03970 (+0.01298) | > avg_loss: -0.45289 (-0.00734) | > avg_log_mle: -0.58241 (-0.00174) | > avg_loss_dur: 0.12952 (-0.00560) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_39680.pth.tar  > EPOCH: 62/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 22:46:09)   --> STEP: 20/639 -- GLOBAL_STEP: 39700 | > loss: -0.59745 (-0.53701) | > log_mle: -0.76503 (-0.66166) | > loss_dur: 0.16758 (0.12465) | > amp_scaler: 2048.00000 (3788.80000) | > grad_norm: 22.32311 (11.06622) | > current_lr: 0.00032 | > step_time: 0.34370 (0.30855) | > loader_time: 0.00280 (0.00237)  --> STEP: 45/639 -- GLOBAL_STEP: 39725 | > loss: -0.47533 (-0.51761) | > log_mle: -0.57934 (-0.63849) | > loss_dur: 0.10401 (0.12088) | > amp_scaler: 2048.00000 (2821.68889) | > grad_norm: 3.43237 (9.50962) | > current_lr: 0.00032 | > step_time: 0.40790 (0.33869) | > loader_time: 0.00300 (0.00257)  --> STEP: 70/639 -- GLOBAL_STEP: 39750 | > loss: -0.51085 (-0.50871) | > log_mle: -0.63449 (-0.62847) | > loss_dur: 0.12364 (0.11976) | > amp_scaler: 2048.00000 (2545.37143) | > grad_norm: 13.49249 (9.24011) | > current_lr: 0.00032 | > step_time: 0.39450 (0.36052) | > loader_time: 0.00290 (0.00267)  --> STEP: 95/639 -- GLOBAL_STEP: 39775 | > loss: -0.57877 (-0.50613) | > log_mle: -0.68577 (-0.62339) | > loss_dur: 0.10700 (0.11725) | > amp_scaler: 2048.00000 (2414.48421) | > grad_norm: 12.31478 (9.18228) | > current_lr: 0.00032 | > step_time: 0.39990 (0.37572) | > loader_time: 0.00310 (0.00272)  --> STEP: 120/639 -- GLOBAL_STEP: 39800 | > loss: -0.54635 (-0.50429) | > log_mle: -0.68414 (-0.62196) | > loss_dur: 0.13779 (0.11767) | > amp_scaler: 2048.00000 (2338.13333) | > grad_norm: 17.10418 (9.39930) | > current_lr: 0.00032 | > step_time: 0.54650 (0.39674) | > loader_time: 0.00310 (0.00278)  --> STEP: 145/639 -- GLOBAL_STEP: 39825 | > loss: -0.50593 (-0.50443) | > log_mle: -0.62919 (-0.62237) | > loss_dur: 0.12326 (0.11793) | > amp_scaler: 2048.00000 (2288.11034) | > grad_norm: 11.94933 (9.59768) | > current_lr: 0.00032 | > step_time: 0.46970 (0.41113) | > loader_time: 0.00290 (0.00282)  --> STEP: 170/639 -- GLOBAL_STEP: 39850 | > loss: -0.50149 (-0.50350) | > log_mle: -0.61685 (-0.62117) | > loss_dur: 0.11535 (0.11767) | > amp_scaler: 2048.00000 (2252.80000) | > grad_norm: 9.36791 (9.72365) | > current_lr: 0.00032 | > step_time: 0.54320 (0.42513) | > loader_time: 0.00320 (0.00286)  --> STEP: 195/639 -- GLOBAL_STEP: 39875 | > loss: -0.45882 (-0.50292) | > log_mle: -0.55607 (-0.61992) | > loss_dur: 0.09725 (0.11700) | > amp_scaler: 2048.00000 (2226.54359) | > grad_norm: 3.71570 (9.82224) | > current_lr: 0.00032 | > step_time: 0.50460 (0.44196) | > loader_time: 0.00330 (0.00290)  --> STEP: 220/639 -- GLOBAL_STEP: 39900 | > loss: -0.50600 (-0.50051) | > log_mle: -0.60427 (-0.61744) | > loss_dur: 0.09828 (0.11692) | > amp_scaler: 2048.00000 (2206.25455) | > grad_norm: 6.84377 (9.75976) | > current_lr: 0.00032 | > step_time: 0.64700 (0.45853) | > loader_time: 0.00340 (0.00294)  --> STEP: 245/639 -- GLOBAL_STEP: 39925 | > loss: -0.48936 (-0.50132) | > log_mle: -0.59720 (-0.61814) | > loss_dur: 0.10785 (0.11682) | > amp_scaler: 2048.00000 (2190.10612) | > grad_norm: 7.22657 (9.82889) | > current_lr: 0.00032 | > step_time: 0.52880 (0.47282) | > loader_time: 0.00320 (0.00298)  --> STEP: 270/639 -- GLOBAL_STEP: 39950 | > loss: -0.52930 (-0.50094) | > log_mle: -0.66341 (-0.61811) | > loss_dur: 0.13411 (0.11717) | > amp_scaler: 2048.00000 (2176.94815) | > grad_norm: 20.89815 (10.06398) | > current_lr: 0.00032 | > step_time: 0.86640 (0.49063) | > loader_time: 0.00350 (0.00301)  --> STEP: 295/639 -- GLOBAL_STEP: 39975 | > loss: -0.44747 (-0.49897) | > log_mle: -0.55484 (-0.61619) | > loss_dur: 0.10738 (0.11721) | > amp_scaler: 2048.00000 (2166.02034) | > grad_norm: 3.73459 (9.93050) | > current_lr: 0.00032 | > step_time: 0.63670 (0.50544) | > loader_time: 0.00350 (0.00304)  --> STEP: 320/639 -- GLOBAL_STEP: 40000 | > loss: -0.47221 (-0.49729) | > log_mle: -0.60055 (-0.61489) | > loss_dur: 0.12834 (0.11760) | > amp_scaler: 2048.00000 (2156.80000) | > grad_norm: 10.03650 (9.91318) | > current_lr: 0.00032 | > step_time: 0.72580 (0.52122) | > loader_time: 0.00360 (0.00308) > CHECKPOINT : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/checkpoint_40000.pth.tar warning: audio amplitude out of range, auto clipped.  --> STEP: 345/639 -- GLOBAL_STEP: 40025 | > loss: -0.47619 (-0.49534) | > log_mle: -0.58867 (-0.61323) | > loss_dur: 0.11248 (0.11789) | > amp_scaler: 2048.00000 (2148.91594) | > grad_norm: 8.56379 (9.86224) | > current_lr: 0.00032 | > step_time: 0.73560 (0.53560) | > loader_time: 0.00350 (0.00319)  --> STEP: 370/639 -- GLOBAL_STEP: 40050 | > loss: -0.48625 (-0.49367) | > log_mle: -0.60221 (-0.61124) | > loss_dur: 0.11597 (0.11757) | > amp_scaler: 2048.00000 (2142.09730) | > grad_norm: 11.05663 (9.73122) | > current_lr: 0.00032 | > step_time: 0.90310 (0.54862) | > loader_time: 0.00390 (0.00321)  --> STEP: 395/639 -- GLOBAL_STEP: 40075 | > loss: -0.44671 (-0.49182) | > log_mle: -0.57149 (-0.60939) | > loss_dur: 0.12478 (0.11757) | > amp_scaler: 2048.00000 (2136.14177) | > grad_norm: 8.79837 (9.59472) | > current_lr: 0.00032 | > step_time: 0.69210 (0.56098) | > loader_time: 0.00320 (0.00323)  --> STEP: 420/639 -- GLOBAL_STEP: 40100 | > loss: -0.45037 (-0.48994) | > log_mle: -0.56687 (-0.60768) | > loss_dur: 0.11650 (0.11774) | > amp_scaler: 2048.00000 (2130.89524) | > grad_norm: 7.02777 (9.51225) | > current_lr: 0.00032 | > step_time: 0.72510 (0.57425) | > loader_time: 0.00340 (0.00325)  --> STEP: 445/639 -- GLOBAL_STEP: 40125 | > loss: -0.46200 (-0.48844) | > log_mle: -0.57401 (-0.60619) | > loss_dur: 0.11202 (0.11775) | > amp_scaler: 2048.00000 (2126.23820) | > grad_norm: 6.09961 (9.41640) | > current_lr: 0.00032 | > step_time: 0.74670 (0.58703) | > loader_time: 0.00370 (0.00328)  --> STEP: 470/639 -- GLOBAL_STEP: 40150 | > loss: -0.46926 (-0.48664) | > log_mle: -0.58459 (-0.60435) | > loss_dur: 0.11533 (0.11771) | > amp_scaler: 2048.00000 (2122.07660) | > grad_norm: 16.99601 (9.28743) | > current_lr: 0.00032 | > step_time: 0.77190 (0.59889) | > loader_time: 0.00400 (0.00331)  --> STEP: 495/639 -- GLOBAL_STEP: 40175 | > loss: -0.41999 (-0.48515) | > log_mle: -0.52164 (-0.60294) | > loss_dur: 0.10165 (0.11778) | > amp_scaler: 2048.00000 (2118.33535) | > grad_norm: 1.65128 (9.22126) | > current_lr: 0.00032 | > step_time: 0.78080 (0.61167) | > loader_time: 0.00390 (0.00334)  --> STEP: 520/639 -- GLOBAL_STEP: 40200 | > loss: -0.46513 (-0.48408) | > log_mle: -0.57714 (-0.60185) | > loss_dur: 0.11201 (0.11777) | > amp_scaler: 2048.00000 (2114.95385) | > grad_norm: 8.96674 (9.15035) | > current_lr: 0.00032 | > step_time: 0.80500 (0.62473) | > loader_time: 0.00380 (0.00337)  --> STEP: 545/639 -- GLOBAL_STEP: 40225 | > loss: -0.48272 (-0.48276) | > log_mle: -0.61125 (-0.60067) | > loss_dur: 0.12853 (0.11792) | > amp_scaler: 2048.00000 (2111.88257) | > grad_norm: 8.11669 (9.06939) | > current_lr: 0.00032 | > step_time: 0.95270 (0.64029) | > loader_time: 0.00420 (0.00341)  --> STEP: 570/639 -- GLOBAL_STEP: 40250 | > loss: -0.45192 (-0.48129) | > log_mle: -0.55705 (-0.59924) | > loss_dur: 0.10513 (0.11794) | > amp_scaler: 2048.00000 (2109.08070) | > grad_norm: 4.98358 (8.98022) | > current_lr: 0.00032 | > step_time: 0.93760 (0.65473) | > loader_time: 0.00400 (0.00344)  --> STEP: 595/639 -- GLOBAL_STEP: 40275 | > loss: -0.43991 (-0.48032) | > log_mle: -0.55256 (-0.59836) | > loss_dur: 0.11265 (0.11804) | > amp_scaler: 2048.00000 (2106.51429) | > grad_norm: 4.31074 (8.92058) | > current_lr: 0.00032 | > step_time: 1.07730 (0.67139) | > loader_time: 0.00420 (0.00348)  --> STEP: 620/639 -- GLOBAL_STEP: 40300 | > loss: -0.44385 (-0.47917) | > log_mle: -0.55828 (-0.59727) | > loss_dur: 0.11443 (0.11809) | > amp_scaler: 2048.00000 (2104.15484) | > grad_norm: 8.06737 (8.87752) | > current_lr: 0.00032 | > step_time: 1.14250 (0.68754) | > loader_time: 0.00450 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02744 (-0.01226) | > avg_loss: -0.44647 (+0.00642) | > avg_log_mle: -0.57737 (+0.00504) | > avg_loss_dur: 0.13090 (+0.00138)  > EPOCH: 63/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 22:53:56)   --> STEP: 5/639 -- GLOBAL_STEP: 40325 | > loss: -0.53042 (-0.53619) | > log_mle: -0.68772 (-0.66493) | > loss_dur: 0.15730 (0.12875) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 16.82414 (11.13485) | > current_lr: 0.00032 | > step_time: 0.30660 (0.28698) | > loader_time: 0.00130 (0.00229)  --> STEP: 30/639 -- GLOBAL_STEP: 40350 | > loss: -0.53251 (-0.52056) | > log_mle: -0.64004 (-0.64203) | > loss_dur: 0.10753 (0.12147) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.05104 (10.33908) | > current_lr: 0.00031 | > step_time: 0.32050 (0.32800) | > loader_time: 0.00280 (0.00255)  --> STEP: 55/639 -- GLOBAL_STEP: 40375 | > loss: -0.47923 (-0.51263) | > log_mle: -0.57360 (-0.63159) | > loss_dur: 0.09437 (0.11896) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.48752 (9.78516) | > current_lr: 0.00031 | > step_time: 0.33260 (0.35011) | > loader_time: 0.00300 (0.00265)  --> STEP: 80/639 -- GLOBAL_STEP: 40400 | > loss: -0.48178 (-0.50611) | > log_mle: -0.61459 (-0.62336) | > loss_dur: 0.13281 (0.11725) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.27486 (9.58982) | > current_lr: 0.00031 | > step_time: 0.39060 (0.36601) | > loader_time: 0.00270 (0.00273)  --> STEP: 105/639 -- GLOBAL_STEP: 40425 | > loss: -0.45263 (-0.50412) | > log_mle: -0.56771 (-0.62105) | > loss_dur: 0.11508 (0.11693) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.97730 (9.50450) | > current_lr: 0.00031 | > step_time: 0.41260 (0.38550) | > loader_time: 0.00270 (0.00279)  --> STEP: 130/639 -- GLOBAL_STEP: 40450 | > loss: -0.47757 (-0.50225) | > log_mle: -0.60568 (-0.61951) | > loss_dur: 0.12811 (0.11725) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.68242 (9.53116) | > current_lr: 0.00031 | > step_time: 0.48870 (0.39994) | > loader_time: 0.00320 (0.00282)  --> STEP: 155/639 -- GLOBAL_STEP: 40475 | > loss: -0.47440 (-0.50150) | > log_mle: -0.59621 (-0.61896) | > loss_dur: 0.12181 (0.11747) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.54444 (9.63136) | > current_lr: 0.00031 | > step_time: 0.54970 (0.41834) | > loader_time: 0.00320 (0.00286)  --> STEP: 180/639 -- GLOBAL_STEP: 40500 | > loss: -0.52345 (-0.50300) | > log_mle: -0.65110 (-0.62078) | > loss_dur: 0.12765 (0.11778) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.37140 (10.04840) | > current_lr: 0.00031 | > step_time: 0.54980 (0.43484) | > loader_time: 0.00360 (0.00290)  --> STEP: 205/639 -- GLOBAL_STEP: 40525 | > loss: -0.46775 (-0.50024) | > log_mle: -0.57707 (-0.61733) | > loss_dur: 0.10932 (0.11710) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.85076 (9.79725) | > current_lr: 0.00031 | > step_time: 0.58400 (0.45101) | > loader_time: 0.00330 (0.00295)  --> STEP: 230/639 -- GLOBAL_STEP: 40550 | > loss: -0.48817 (-0.50035) | > log_mle: -0.60834 (-0.61732) | > loss_dur: 0.12017 (0.11697) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.25371 (9.84736) | > current_lr: 0.00031 | > step_time: 0.56400 (0.46848) | > loader_time: 0.00300 (0.00299)  --> STEP: 255/639 -- GLOBAL_STEP: 40575 | > loss: -0.46315 (-0.49953) | > log_mle: -0.58173 (-0.61633) | > loss_dur: 0.11858 (0.11680) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.16572 (9.96458) | > current_lr: 0.00031 | > step_time: 0.64050 (0.48186) | > loader_time: 0.00360 (0.00303)  --> STEP: 280/639 -- GLOBAL_STEP: 40600 | > loss: -0.48358 (-0.49931) | > log_mle: -0.60649 (-0.61646) | > loss_dur: 0.12291 (0.11715) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.28372 (10.22771) | > current_lr: 0.00031 | > step_time: 0.61600 (0.49912) | > loader_time: 0.00330 (0.00307)  --> STEP: 305/639 -- GLOBAL_STEP: 40625 | > loss: -0.46351 (-0.49758) | > log_mle: -0.58450 (-0.61489) | > loss_dur: 0.12099 (0.11731) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.48300 (10.11189) | > current_lr: 0.00031 | > step_time: 0.70850 (0.51447) | > loader_time: 0.00330 (0.00310)  --> STEP: 330/639 -- GLOBAL_STEP: 40650 | > loss: -0.47358 (-0.49518) | > log_mle: -0.60823 (-0.61299) | > loss_dur: 0.13465 (0.11781) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.66306 (10.01551) | > current_lr: 0.00031 | > step_time: 0.75130 (0.52883) | > loader_time: 0.00350 (0.00314)  --> STEP: 355/639 -- GLOBAL_STEP: 40675 | > loss: -0.42841 (-0.49364) | > log_mle: -0.53760 (-0.61123) | > loss_dur: 0.10919 (0.11760) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.35306 (9.87302) | > current_lr: 0.00031 | > step_time: 0.61670 (0.54208) | > loader_time: 0.00350 (0.00317)  --> STEP: 380/639 -- GLOBAL_STEP: 40700 | > loss: -0.46629 (-0.49175) | > log_mle: -0.57073 (-0.60919) | > loss_dur: 0.10444 (0.11744) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.07707 (9.74329) | > current_lr: 0.00031 | > step_time: 0.65980 (0.55562) | > loader_time: 0.00350 (0.00320)  --> STEP: 405/639 -- GLOBAL_STEP: 40725 | > loss: -0.46239 (-0.49051) | > log_mle: -0.57895 (-0.60814) | > loss_dur: 0.11656 (0.11763) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.97138 (9.66519) | > current_lr: 0.00031 | > step_time: 0.69590 (0.56843) | > loader_time: 0.00360 (0.00323)  --> STEP: 430/639 -- GLOBAL_STEP: 40750 | > loss: -0.43166 (-0.48869) | > log_mle: -0.53828 (-0.60638) | > loss_dur: 0.10662 (0.11770) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.30006 (9.57007) | > current_lr: 0.00031 | > step_time: 0.82070 (0.58151) | > loader_time: 0.00370 (0.00326)  --> STEP: 455/639 -- GLOBAL_STEP: 40775 | > loss: -0.44568 (-0.48698) | > log_mle: -0.56351 (-0.60452) | > loss_dur: 0.11783 (0.11755) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.86392 (9.41516) | > current_lr: 0.00031 | > step_time: 0.74310 (0.59191) | > loader_time: 0.00350 (0.00328)  --> STEP: 480/639 -- GLOBAL_STEP: 40800 | > loss: -0.39986 (-0.48528) | > log_mle: -0.51895 (-0.60303) | > loss_dur: 0.11909 (0.11774) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.20855 (9.28974) | > current_lr: 0.00031 | > step_time: 0.81670 (0.60483) | > loader_time: 0.00350 (0.00331)  --> STEP: 505/639 -- GLOBAL_STEP: 40825 | > loss: -0.46830 (-0.48398) | > log_mle: -0.58292 (-0.60168) | > loss_dur: 0.11462 (0.11770) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.62995 (9.24922) | > current_lr: 0.00031 | > step_time: 0.89760 (0.61797) | > loader_time: 0.00410 (0.00334)  --> STEP: 530/639 -- GLOBAL_STEP: 40850 | > loss: -0.44494 (-0.48273) | > log_mle: -0.57547 (-0.60057) | > loss_dur: 0.13053 (0.11784) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.70655 (9.16110) | > current_lr: 0.00031 | > step_time: 0.95370 (0.63168) | > loader_time: 0.00400 (0.00337)  --> STEP: 555/639 -- GLOBAL_STEP: 40875 | > loss: -0.44760 (-0.48142) | > log_mle: -0.56097 (-0.59926) | > loss_dur: 0.11337 (0.11784) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.76658 (9.03935) | > current_lr: 0.00031 | > step_time: 0.82580 (0.64584) | > loader_time: 0.00390 (0.00341)  --> STEP: 580/639 -- GLOBAL_STEP: 40900 | > loss: -0.46669 (-0.48032) | > log_mle: -0.57955 (-0.59823) | > loss_dur: 0.11286 (0.11791) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.38447 (8.96850) | > current_lr: 0.00031 | > step_time: 0.97240 (0.66140) | > loader_time: 0.00410 (0.00346)  --> STEP: 605/639 -- GLOBAL_STEP: 40925 | > loss: -0.43881 (-0.47922) | > log_mle: -0.54744 (-0.59728) | > loss_dur: 0.10863 (0.11807) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.31448 (8.90548) | > current_lr: 0.00031 | > step_time: 1.08300 (0.67741) | > loader_time: 0.00450 (0.00352)  --> STEP: 630/639 -- GLOBAL_STEP: 40950 | > loss: -0.45117 (-0.47824) | > log_mle: -0.57829 (-0.59646) | > loss_dur: 0.12712 (0.11822) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.98009 (8.88899) | > current_lr: 0.00031 | > step_time: 1.18050 (0.69492) | > loader_time: 0.00450 (0.00357) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03121 (+0.00377) | > avg_loss: -0.45176 (-0.00529) | > avg_log_mle: -0.58226 (-0.00489) | > avg_loss_dur: 0.13050 (-0.00040)  > EPOCH: 64/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 23:01:38)   --> STEP: 15/639 -- GLOBAL_STEP: 40975 | > loss: -0.63252 (-0.53625) | > log_mle: -0.77233 (-0.65869) | > loss_dur: 0.13981 (0.12244) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 23.18628 (12.79208) | > current_lr: 0.00031 | > step_time: 0.32850 (0.29712) | > loader_time: 0.00250 (0.00252)  --> STEP: 40/639 -- GLOBAL_STEP: 41000 | > loss: -0.49852 (-0.51741) | > log_mle: -0.61531 (-0.63695) | > loss_dur: 0.11679 (0.11954) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.15478 (10.54177) | > current_lr: 0.00031 | > step_time: 0.34920 (0.33010) | > loader_time: 0.00250 (0.00267)  --> STEP: 65/639 -- GLOBAL_STEP: 41025 | > loss: -0.50986 (-0.51042) | > log_mle: -0.60351 (-0.62940) | > loss_dur: 0.09364 (0.11898) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.09025 (9.88611) | > current_lr: 0.00031 | > step_time: 0.37520 (0.35887) | > loader_time: 0.00270 (0.00278)  --> STEP: 90/639 -- GLOBAL_STEP: 41050 | > loss: -0.44772 (-0.50356) | > log_mle: -0.56230 (-0.62050) | > loss_dur: 0.11458 (0.11694) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.97596 (9.33703) | > current_lr: 0.00031 | > step_time: 0.51550 (0.37774) | > loader_time: 0.00320 (0.00282)  --> STEP: 115/639 -- GLOBAL_STEP: 41075 | > loss: -0.41436 (-0.50414) | > log_mle: -0.52864 (-0.62106) | > loss_dur: 0.11428 (0.11692) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.25411 (9.46907) | > current_lr: 0.00031 | > step_time: 0.38350 (0.39493) | > loader_time: 0.00300 (0.00287)  --> STEP: 140/639 -- GLOBAL_STEP: 41100 | > loss: -0.47421 (-0.50433) | > log_mle: -0.61421 (-0.62158) | > loss_dur: 0.14000 (0.11725) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.78293 (9.63666) | > current_lr: 0.00031 | > step_time: 0.55270 (0.40836) | > loader_time: 0.00330 (0.00289)  --> STEP: 165/639 -- GLOBAL_STEP: 41125 | > loss: -0.48114 (-0.50374) | > log_mle: -0.61810 (-0.62058) | > loss_dur: 0.13696 (0.11684) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.42885 (9.68453) | > current_lr: 0.00031 | > step_time: 0.50120 (0.42329) | > loader_time: 0.00270 (0.00292)  --> STEP: 190/639 -- GLOBAL_STEP: 41150 | > loss: -0.43214 (-0.50362) | > log_mle: -0.54571 (-0.62021) | > loss_dur: 0.11357 (0.11659) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.30273 (9.74182) | > current_lr: 0.00031 | > step_time: 0.51730 (0.44147) | > loader_time: 0.00330 (0.00297)  --> STEP: 215/639 -- GLOBAL_STEP: 41175 | > loss: -0.54121 (-0.50118) | > log_mle: -0.65195 (-0.61799) | > loss_dur: 0.11074 (0.11681) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.26017 (9.67201) | > current_lr: 0.00031 | > step_time: 0.54460 (0.45897) | > loader_time: 0.00290 (0.00300)  --> STEP: 240/639 -- GLOBAL_STEP: 41200 | > loss: -0.53466 (-0.50202) | > log_mle: -0.65732 (-0.61860) | > loss_dur: 0.12266 (0.11658) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.86510 (9.87945) | > current_lr: 0.00031 | > step_time: 0.55640 (0.47392) | > loader_time: 0.00320 (0.00303)  --> STEP: 265/639 -- GLOBAL_STEP: 41225 | > loss: -0.55110 (-0.50102) | > log_mle: -0.68419 (-0.61798) | > loss_dur: 0.13310 (0.11696) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 18.56744 (9.91488) | > current_lr: 0.00031 | > step_time: 0.62990 (0.48927) | > loader_time: 0.00320 (0.00307)  --> STEP: 290/639 -- GLOBAL_STEP: 41250 | > loss: -0.46851 (-0.49952) | > log_mle: -0.58161 (-0.61659) | > loss_dur: 0.11310 (0.11707) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.58517 (9.89257) | > current_lr: 0.00031 | > step_time: 0.64140 (0.50577) | > loader_time: 0.00340 (0.00310)  --> STEP: 315/639 -- GLOBAL_STEP: 41275 | > loss: -0.44803 (-0.49789) | > log_mle: -0.55836 (-0.61510) | > loss_dur: 0.11033 (0.11721) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.96038 (9.79962) | > current_lr: 0.00031 | > step_time: 0.63890 (0.51993) | > loader_time: 0.00300 (0.00312)  --> STEP: 340/639 -- GLOBAL_STEP: 41300 | > loss: -0.47370 (-0.49603) | > log_mle: -0.58162 (-0.61370) | > loss_dur: 0.10792 (0.11767) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.41433 (9.90049) | > current_lr: 0.00031 | > step_time: 0.62730 (0.53489) | > loader_time: 0.00360 (0.00316)  --> STEP: 365/639 -- GLOBAL_STEP: 41325 | > loss: -0.46594 (-0.49395) | > log_mle: -0.58513 (-0.61138) | > loss_dur: 0.11919 (0.11743) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.92541 (9.77348) | > current_lr: 0.00031 | > step_time: 0.83570 (0.54652) | > loader_time: 0.00390 (0.00319)  --> STEP: 390/639 -- GLOBAL_STEP: 41350 | > loss: -0.46348 (-0.49207) | > log_mle: -0.58612 (-0.60939) | > loss_dur: 0.12263 (0.11731) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.54203 (9.65293) | > current_lr: 0.00031 | > step_time: 0.78480 (0.55929) | > loader_time: 0.00360 (0.00322)  --> STEP: 415/639 -- GLOBAL_STEP: 41375 | > loss: -0.44250 (-0.49019) | > log_mle: -0.56150 (-0.60773) | > loss_dur: 0.11901 (0.11755) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.52725 (9.57446) | > current_lr: 0.00031 | > step_time: 0.74930 (0.57241) | > loader_time: 0.00300 (0.00325)  --> STEP: 440/639 -- GLOBAL_STEP: 41400 | > loss: -0.42481 (-0.48876) | > log_mle: -0.54450 (-0.60622) | > loss_dur: 0.11969 (0.11746) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.09259 (9.49249) | > current_lr: 0.00031 | > step_time: 0.72940 (0.58438) | > loader_time: 0.00370 (0.00327)  --> STEP: 465/639 -- GLOBAL_STEP: 41425 | > loss: -0.43562 (-0.48699) | > log_mle: -0.54919 (-0.60433) | > loss_dur: 0.11357 (0.11734) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.79391 (9.36859) | > current_lr: 0.00031 | > step_time: 0.78830 (0.59623) | > loader_time: 0.00400 (0.00330)  --> STEP: 490/639 -- GLOBAL_STEP: 41450 | > loss: -0.42732 (-0.48533) | > log_mle: -0.53449 (-0.60285) | > loss_dur: 0.10717 (0.11752) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.17112 (9.27074) | > current_lr: 0.00031 | > step_time: 0.78390 (0.60950) | > loader_time: 0.00380 (0.00332)  --> STEP: 515/639 -- GLOBAL_STEP: 41475 | > loss: -0.45232 (-0.48431) | > log_mle: -0.56328 (-0.60182) | > loss_dur: 0.11096 (0.11751) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.68372 (9.19808) | > current_lr: 0.00031 | > step_time: 0.94360 (0.62230) | > loader_time: 0.00390 (0.00335)  --> STEP: 540/639 -- GLOBAL_STEP: 41500 | > loss: -0.46360 (-0.48297) | > log_mle: -0.58138 (-0.60051) | > loss_dur: 0.11777 (0.11753) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.25748 (9.12680) | > current_lr: 0.00031 | > step_time: 0.89520 (0.63679) | > loader_time: 0.00420 (0.00339)  --> STEP: 565/639 -- GLOBAL_STEP: 41525 | > loss: -0.42467 (-0.48151) | > log_mle: -0.54173 (-0.59912) | > loss_dur: 0.11706 (0.11761) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.30746 (9.02790) | > current_lr: 0.00031 | > step_time: 0.99790 (0.65071) | > loader_time: 0.00440 (0.00342)  --> STEP: 590/639 -- GLOBAL_STEP: 41550 | > loss: -0.45689 (-0.48065) | > log_mle: -0.57369 (-0.59832) | > loss_dur: 0.11681 (0.11768) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.94437 (8.98169) | > current_lr: 0.00031 | > step_time: 0.95920 (0.66639) | > loader_time: 0.00430 (0.00346)  --> STEP: 615/639 -- GLOBAL_STEP: 41575 | > loss: -0.45663 (-0.47955) | > log_mle: -0.57785 (-0.59728) | > loss_dur: 0.12123 (0.11773) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.95644 (8.94148) | > current_lr: 0.00031 | > step_time: 1.00930 (0.68219) | > loader_time: 0.00430 (0.00350) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03016 (-0.00105) | > avg_loss: -0.45099 (+0.00076) | > avg_log_mle: -0.58151 (+0.00076) | > avg_loss_dur: 0.13051 (+0.00001)  > EPOCH: 65/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 23:09:19)   --> STEP: 0/639 -- GLOBAL_STEP: 41600 | > loss: -0.53434 (-0.53434) | > log_mle: -0.70481 (-0.70481) | > loss_dur: 0.17047 (0.17047) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 21.33188 (21.33188) | > current_lr: 0.00031 | > step_time: 0.63870 (0.63868) | > loader_time: 0.43950 (0.43952)  --> STEP: 25/639 -- GLOBAL_STEP: 41625 | > loss: -0.53096 (-0.53539) | > log_mle: -0.65846 (-0.65515) | > loss_dur: 0.12750 (0.11976) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.88662 (12.39681) | > current_lr: 0.00031 | > step_time: 0.38700 (0.32448) | > loader_time: 0.00300 (0.00258)  --> STEP: 50/639 -- GLOBAL_STEP: 41650 | > loss: -0.48803 (-0.52180) | > log_mle: -0.58767 (-0.63908) | > loss_dur: 0.09964 (0.11728) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.83664 (11.06037) | > current_lr: 0.00031 | > step_time: 0.31400 (0.34537) | > loader_time: 0.00280 (0.00269)  --> STEP: 75/639 -- GLOBAL_STEP: 41675 | > loss: -0.50487 (-0.51220) | > log_mle: -0.61164 (-0.62932) | > loss_dur: 0.10677 (0.11712) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.09458 (10.56956) | > current_lr: 0.00031 | > step_time: 0.41120 (0.36486) | > loader_time: 0.00300 (0.00275)  --> STEP: 100/639 -- GLOBAL_STEP: 41700 | > loss: -0.51121 (-0.50981) | > log_mle: -0.62216 (-0.62589) | > loss_dur: 0.11095 (0.11608) | > amp_scaler: 4096.00000 (2109.44000) | > grad_norm: 9.03656 (10.10377) | > current_lr: 0.00031 | > step_time: 0.50880 (0.38250) | > loader_time: 0.00310 (0.00282)  --> STEP: 125/639 -- GLOBAL_STEP: 41725 | > loss: -0.49761 (-0.50738) | > log_mle: -0.61461 (-0.62386) | > loss_dur: 0.11700 (0.11648) | > amp_scaler: 4096.00000 (2506.75200) | > grad_norm: 8.64612 (9.94973) | > current_lr: 0.00031 | > step_time: 0.41410 (0.39585) | > loader_time: 0.00300 (0.00288)  --> STEP: 150/639 -- GLOBAL_STEP: 41750 | > loss: -0.46656 (-0.50611) | > log_mle: -0.58237 (-0.62283) | > loss_dur: 0.11582 (0.11672) | > amp_scaler: 4096.00000 (2771.62667) | > grad_norm: 7.07941 (9.92012) | > current_lr: 0.00031 | > step_time: 0.43310 (0.41198) | > loader_time: 0.00280 (0.00292)  --> STEP: 175/639 -- GLOBAL_STEP: 41775 | > loss: -0.49570 (-0.50664) | > log_mle: -0.59876 (-0.62312) | > loss_dur: 0.10306 (0.11648) | > amp_scaler: 4096.00000 (2960.82286) | > grad_norm: 7.96577 (9.98945) | > current_lr: 0.00031 | > step_time: 0.57420 (0.42944) | > loader_time: 0.00350 (0.00296)  --> STEP: 200/639 -- GLOBAL_STEP: 41800 | > loss: -0.46493 (-0.50475) | > log_mle: -0.58282 (-0.62068) | > loss_dur: 0.11788 (0.11593) | > amp_scaler: 4096.00000 (3102.72000) | > grad_norm: 6.19219 (9.85071) | > current_lr: 0.00031 | > step_time: 0.53190 (0.44429) | > loader_time: 0.00320 (0.00299)  --> STEP: 225/639 -- GLOBAL_STEP: 41825 | > loss: -0.46181 (-0.50286) | > log_mle: -0.56248 (-0.61861) | > loss_dur: 0.10066 (0.11575) | > amp_scaler: 4096.00000 (3213.08444) | > grad_norm: 5.05596 (9.77584) | > current_lr: 0.00031 | > step_time: 0.50770 (0.46222) | > loader_time: 0.00320 (0.00304)  --> STEP: 250/639 -- GLOBAL_STEP: 41850 | > loss: -0.46753 (-0.50367) | > log_mle: -0.56790 (-0.61960) | > loss_dur: 0.10038 (0.11593) | > amp_scaler: 4096.00000 (3301.37600) | > grad_norm: 5.62474 (9.96119) | > current_lr: 0.00031 | > step_time: 0.57290 (0.47946) | > loader_time: 0.00350 (0.00308)  --> STEP: 275/639 -- GLOBAL_STEP: 41875 | > loss: -0.49193 (-0.50360) | > log_mle: -0.60666 (-0.62004) | > loss_dur: 0.11472 (0.11643) | > amp_scaler: 4096.00000 (3373.61455) | > grad_norm: 9.85152 (10.17212) | > current_lr: 0.00031 | > step_time: 0.87280 (0.49682) | > loader_time: 0.00360 (0.00311)  --> STEP: 300/639 -- GLOBAL_STEP: 41900 | > loss: -0.48491 (-0.50145) | > log_mle: -0.59898 (-0.61801) | > loss_dur: 0.11407 (0.11656) | > amp_scaler: 4096.00000 (3433.81333) | > grad_norm: 7.70980 (10.02391) | > current_lr: 0.00031 | > step_time: 0.67960 (0.51028) | > loader_time: 0.00360 (0.00313)  --> STEP: 325/639 -- GLOBAL_STEP: 41925 | > loss: -0.47975 (-0.49929) | > log_mle: -0.59794 (-0.61636) | > loss_dur: 0.11819 (0.11707) | > amp_scaler: 4096.00000 (3484.75077) | > grad_norm: 7.47545 (10.02453) | > current_lr: 0.00031 | > step_time: 0.63740 (0.52584) | > loader_time: 0.00350 (0.00317)  --> STEP: 350/639 -- GLOBAL_STEP: 41950 | > loss: -0.49830 (-0.49733) | > log_mle: -0.62524 (-0.61448) | > loss_dur: 0.12693 (0.11715) | > amp_scaler: 4096.00000 (3528.41143) | > grad_norm: 12.90519 (9.90558) | > current_lr: 0.00031 | > step_time: 0.84740 (0.53978) | > loader_time: 0.00380 (0.00319)  --> STEP: 375/639 -- GLOBAL_STEP: 41975 | > loss: -0.44619 (-0.49519) | > log_mle: -0.56085 (-0.61213) | > loss_dur: 0.11466 (0.11694) | > amp_scaler: 4096.00000 (3566.25067) | > grad_norm: 7.65440 (9.71490) | > current_lr: 0.00031 | > step_time: 0.70230 (0.55293) | > loader_time: 0.00350 (0.00322)  --> STEP: 400/639 -- GLOBAL_STEP: 42000 | > loss: -0.47576 (-0.49364) | > log_mle: -0.59174 (-0.61072) | > loss_dur: 0.11598 (0.11708) | > amp_scaler: 4096.00000 (3599.36000) | > grad_norm: 6.80185 (9.61724) | > current_lr: 0.00031 | > step_time: 0.80010 (0.56570) | > loader_time: 0.00360 (0.00325)  --> STEP: 425/639 -- GLOBAL_STEP: 42025 | > loss: -0.52459 (-0.49159) | > log_mle: -0.64369 (-0.60872) | > loss_dur: 0.11909 (0.11713) | > amp_scaler: 4096.00000 (3628.57412) | > grad_norm: 18.78713 (9.49248) | > current_lr: 0.00031 | > step_time: 0.87850 (0.57829) | > loader_time: 0.00380 (0.00328)  --> STEP: 450/639 -- GLOBAL_STEP: 42050 | > loss: -0.43992 (-0.48991) | > log_mle: -0.54907 (-0.60697) | > loss_dur: 0.10916 (0.11706) | > amp_scaler: 4096.00000 (3654.54222) | > grad_norm: 4.60283 (9.35446) | > current_lr: 0.00031 | > step_time: 0.78030 (0.59013) | > loader_time: 0.00390 (0.00331)  --> STEP: 475/639 -- GLOBAL_STEP: 42075 | > loss: -0.46311 (-0.48819) | > log_mle: -0.56092 (-0.60534) | > loss_dur: 0.09781 (0.11715) | > amp_scaler: 4096.00000 (3677.77684) | > grad_norm: 4.57803 (9.30189) | > current_lr: 0.00031 | > step_time: 0.72710 (0.60205) | > loader_time: 0.00370 (0.00333)  --> STEP: 500/639 -- GLOBAL_STEP: 42100 | > loss: -0.50025 (-0.48698) | > log_mle: -0.62077 (-0.60419) | > loss_dur: 0.12052 (0.11721) | > amp_scaler: 4096.00000 (3698.68800) | > grad_norm: 12.47280 (9.26004) | > current_lr: 0.00031 | > step_time: 0.95510 (0.61472) | > loader_time: 0.00410 (0.00336)  --> STEP: 525/639 -- GLOBAL_STEP: 42125 | > loss: -0.46665 (-0.48573) | > log_mle: -0.60375 (-0.60304) | > loss_dur: 0.13709 (0.11732) | > amp_scaler: 4096.00000 (3717.60762) | > grad_norm: 18.26465 (9.22939) | > current_lr: 0.00031 | > step_time: 1.01840 (0.62818) | > loader_time: 0.00440 (0.00339)  --> STEP: 550/639 -- GLOBAL_STEP: 42150 | > loss: -0.41888 (-0.48438) | > log_mle: -0.53677 (-0.60168) | > loss_dur: 0.11788 (0.11730) | > amp_scaler: 4096.00000 (3734.80727) | > grad_norm: 4.75870 (9.18270) | > current_lr: 0.00031 | > step_time: 0.91060 (0.64237) | > loader_time: 0.00420 (0.00342)  --> STEP: 575/639 -- GLOBAL_STEP: 42175 | > loss: -0.43772 (-0.48296) | > log_mle: -0.55497 (-0.60036) | > loss_dur: 0.11726 (0.11741) | > amp_scaler: 4096.00000 (3750.51130) | > grad_norm: 6.05251 (9.09628) | > current_lr: 0.00031 | > step_time: 0.92820 (0.65681) | > loader_time: 0.00420 (0.00345)  --> STEP: 600/639 -- GLOBAL_STEP: 42200 | > loss: -0.43096 (-0.48192) | > log_mle: -0.55467 (-0.59943) | > loss_dur: 0.12371 (0.11752) | > amp_scaler: 4096.00000 (3764.90667) | > grad_norm: 7.76291 (9.03411) | > current_lr: 0.00031 | > step_time: 1.01470 (0.67323) | > loader_time: 0.00420 (0.00349)  --> STEP: 625/639 -- GLOBAL_STEP: 42225 | > loss: -0.46835 (-0.48090) | > log_mle: -0.58486 (-0.59846) | > loss_dur: 0.11651 (0.11756) | > amp_scaler: 4096.00000 (3778.15040) | > grad_norm: 7.23866 (8.98278) | > current_lr: 0.00031 | > step_time: 1.08450 (0.69031) | > loader_time: 0.00460 (0.00353) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02906 (-0.00110) | > avg_loss: -0.44813 (+0.00287) | > avg_log_mle: -0.58006 (+0.00145) | > avg_loss_dur: 0.13193 (+0.00142)  > EPOCH: 66/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 23:17:01)   --> STEP: 10/639 -- GLOBAL_STEP: 42250 | > loss: -0.51862 (-0.53188) | > log_mle: -0.65699 (-0.65681) | > loss_dur: 0.13838 (0.12493) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.30191 (10.03275) | > current_lr: 0.00031 | > step_time: 0.28300 (0.29583) | > loader_time: 0.00230 (0.00236)  --> STEP: 35/639 -- GLOBAL_STEP: 42275 | > loss: -0.51342 (-0.52076) | > log_mle: -0.61933 (-0.63949) | > loss_dur: 0.10591 (0.11873) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.66966 (10.95811) | > current_lr: 0.00031 | > step_time: 0.32100 (0.32932) | > loader_time: 0.00280 (0.00264)  --> STEP: 60/639 -- GLOBAL_STEP: 42300 | > loss: -0.47110 (-0.51245) | > log_mle: -0.57531 (-0.63130) | > loss_dur: 0.10421 (0.11884) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.28000 (10.55224) | > current_lr: 0.00031 | > step_time: 0.34220 (0.35983) | > loader_time: 0.00280 (0.00270)  --> STEP: 85/639 -- GLOBAL_STEP: 42325 | > loss: -0.46533 (-0.50619) | > log_mle: -0.57464 (-0.62300) | > loss_dur: 0.10931 (0.11681) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.48866 (9.73222) | > current_lr: 0.00031 | > step_time: 0.46260 (0.37299) | > loader_time: 0.00250 (0.00274)  --> STEP: 110/639 -- GLOBAL_STEP: 42350 | > loss: -0.47333 (-0.50769) | > log_mle: -0.58902 (-0.62485) | > loss_dur: 0.11569 (0.11716) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.98514 (10.04787) | > current_lr: 0.00031 | > step_time: 0.41220 (0.39265) | > loader_time: 0.00320 (0.00279)  --> STEP: 135/639 -- GLOBAL_STEP: 42375 | > loss: -0.45212 (-0.50578) | > log_mle: -0.57397 (-0.62244) | > loss_dur: 0.12186 (0.11667) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.19301 (9.95467) | > current_lr: 0.00031 | > step_time: 0.46710 (0.40429) | > loader_time: 0.00300 (0.00283)  --> STEP: 160/639 -- GLOBAL_STEP: 42400 | > loss: -0.50554 (-0.50372) | > log_mle: -0.61230 (-0.61993) | > loss_dur: 0.10676 (0.11621) | > amp_scaler: 2048.00000 (3852.80000) | > grad_norm: 9.06955 (10.23789) | > current_lr: 0.00031 | > step_time: 0.47370 (0.41947) | > loader_time: 0.00280 (0.00286)  --> STEP: 185/639 -- GLOBAL_STEP: 42425 | > loss: -0.48333 (-0.50562) | > log_mle: -0.59211 (-0.62214) | > loss_dur: 0.10878 (0.11652) | > amp_scaler: 2048.00000 (3608.90811) | > grad_norm: 10.54912 (10.89362) | > current_lr: 0.00031 | > step_time: 0.59130 (0.43779) | > loader_time: 0.00310 (0.00291)  --> STEP: 210/639 -- GLOBAL_STEP: 42450 | > loss: -0.41716 (-0.50286) | > log_mle: -0.51594 (-0.61918) | > loss_dur: 0.09878 (0.11632) | > amp_scaler: 2048.00000 (3423.08571) | > grad_norm: 1.01528 (10.61491) | > current_lr: 0.00031 | > step_time: 0.51740 (0.45404) | > loader_time: 0.00300 (0.00295)  --> STEP: 235/639 -- GLOBAL_STEP: 42475 | > loss: -0.49927 (-0.50319) | > log_mle: -0.60655 (-0.61954) | > loss_dur: 0.10728 (0.11635) | > amp_scaler: 2048.00000 (3276.80000) | > grad_norm: 9.12052 (10.76502) | > current_lr: 0.00031 | > step_time: 0.54740 (0.47017) | > loader_time: 0.00330 (0.00298)  --> STEP: 260/639 -- GLOBAL_STEP: 42500 | > loss: -0.49097 (-0.50254) | > log_mle: -0.60868 (-0.61910) | > loss_dur: 0.11770 (0.11656) | > amp_scaler: 2048.00000 (3158.64615) | > grad_norm: 9.21487 (10.83359) | > current_lr: 0.00031 | > step_time: 0.60150 (0.48492) | > loader_time: 0.00290 (0.00299)  --> STEP: 285/639 -- GLOBAL_STEP: 42525 | > loss: -0.43947 (-0.50139) | > log_mle: -0.55559 (-0.61819) | > loss_dur: 0.11612 (0.11680) | > amp_scaler: 2048.00000 (3061.22105) | > grad_norm: 3.27094 (10.90322) | > current_lr: 0.00031 | > step_time: 0.60790 (0.50128) | > loader_time: 0.00350 (0.00304)  --> STEP: 310/639 -- GLOBAL_STEP: 42550 | > loss: -0.47902 (-0.49974) | > log_mle: -0.59891 (-0.61670) | > loss_dur: 0.11989 (0.11696) | > amp_scaler: 2048.00000 (2979.50968) | > grad_norm: 8.66306 (10.77019) | > current_lr: 0.00031 | > step_time: 0.58180 (0.51632) | > loader_time: 0.00300 (0.00307)  --> STEP: 335/639 -- GLOBAL_STEP: 42575 | > loss: -0.49725 (-0.49736) | > log_mle: -0.62328 (-0.61477) | > loss_dur: 0.12603 (0.11740) | > amp_scaler: 2048.00000 (2909.99403) | > grad_norm: 11.87724 (10.60083) | > current_lr: 0.00031 | > step_time: 0.66750 (0.53189) | > loader_time: 0.00340 (0.00310)  --> STEP: 360/639 -- GLOBAL_STEP: 42600 | > loss: -0.45140 (-0.49583) | > log_mle: -0.56169 (-0.61298) | > loss_dur: 0.11029 (0.11715) | > amp_scaler: 2048.00000 (2850.13333) | > grad_norm: 3.92610 (10.53097) | > current_lr: 0.00031 | > step_time: 0.70150 (0.54423) | > loader_time: 0.00350 (0.00313)  --> STEP: 385/639 -- GLOBAL_STEP: 42625 | > loss: -0.48082 (-0.49432) | > log_mle: -0.59607 (-0.61134) | > loss_dur: 0.11525 (0.11702) | > amp_scaler: 2048.00000 (2798.04675) | > grad_norm: 6.16870 (10.35118) | > current_lr: 0.00031 | > step_time: 0.67460 (0.55794) | > loader_time: 0.00370 (0.00317)  --> STEP: 410/639 -- GLOBAL_STEP: 42650 | > loss: -0.42864 (-0.49246) | > log_mle: -0.56320 (-0.60978) | > loss_dur: 0.13456 (0.11732) | > amp_scaler: 2048.00000 (2752.31220) | > grad_norm: 9.85584 (10.25936) | > current_lr: 0.00031 | > step_time: 0.78260 (0.56909) | > loader_time: 0.00370 (0.00319)  --> STEP: 435/639 -- GLOBAL_STEP: 42675 | > loss: -0.46876 (-0.49057) | > log_mle: -0.57604 (-0.60788) | > loss_dur: 0.10729 (0.11730) | > amp_scaler: 2048.00000 (2711.83448) | > grad_norm: 8.82077 (10.08678) | > current_lr: 0.00031 | > step_time: 0.73280 (0.58174) | > loader_time: 0.00360 (0.00321)  --> STEP: 460/639 -- GLOBAL_STEP: 42700 | > loss: -0.42337 (-0.48875) | > log_mle: -0.54292 (-0.60599) | > loss_dur: 0.11954 (0.11724) | > amp_scaler: 2048.00000 (2675.75652) | > grad_norm: 3.18373 (9.93428) | > current_lr: 0.00031 | > step_time: 0.87290 (0.59254) | > loader_time: 0.00390 (0.00324)  --> STEP: 485/639 -- GLOBAL_STEP: 42725 | > loss: -0.46632 (-0.48725) | > log_mle: -0.57777 (-0.60463) | > loss_dur: 0.11144 (0.11738) | > amp_scaler: 2048.00000 (2643.39794) | > grad_norm: 5.66911 (9.86599) | > current_lr: 0.00031 | > step_time: 0.87080 (0.60457) | > loader_time: 0.00410 (0.00327)  --> STEP: 510/639 -- GLOBAL_STEP: 42750 | > loss: -0.46043 (-0.48608) | > log_mle: -0.57871 (-0.60340) | > loss_dur: 0.11829 (0.11733) | > amp_scaler: 2048.00000 (2614.21176) | > grad_norm: 7.40212 (9.77087) | > current_lr: 0.00031 | > step_time: 0.87710 (0.61785) | > loader_time: 0.00380 (0.00329)  --> STEP: 535/639 -- GLOBAL_STEP: 42775 | > loss: -0.43445 (-0.48465) | > log_mle: -0.56508 (-0.60205) | > loss_dur: 0.13063 (0.11739) | > amp_scaler: 2048.00000 (2587.75327) | > grad_norm: 9.44343 (9.69210) | > current_lr: 0.00031 | > step_time: 0.91210 (0.63107) | > loader_time: 0.00400 (0.00332)  --> STEP: 560/639 -- GLOBAL_STEP: 42800 | > loss: -0.45138 (-0.48330) | > log_mle: -0.57546 (-0.60080) | > loss_dur: 0.12408 (0.11750) | > amp_scaler: 2048.00000 (2563.65714) | > grad_norm: 7.86190 (9.62297) | > current_lr: 0.00031 | > step_time: 0.90540 (0.64586) | > loader_time: 0.00410 (0.00336)  --> STEP: 585/639 -- GLOBAL_STEP: 42825 | > loss: -0.46518 (-0.48219) | > log_mle: -0.58194 (-0.59969) | > loss_dur: 0.11675 (0.11750) | > amp_scaler: 2048.00000 (2541.62051) | > grad_norm: 8.41844 (9.52885) | > current_lr: 0.00031 | > step_time: 1.05300 (0.66146) | > loader_time: 0.00460 (0.00340)  --> STEP: 610/639 -- GLOBAL_STEP: 42850 | > loss: -0.42915 (-0.48099) | > log_mle: -0.54761 (-0.59865) | > loss_dur: 0.11846 (0.11766) | > amp_scaler: 2048.00000 (2521.39016) | > grad_norm: 14.79923 (9.50858) | > current_lr: 0.00031 | > step_time: 1.03940 (0.67747) | > loader_time: 0.00420 (0.00344)  --> STEP: 635/639 -- GLOBAL_STEP: 42875 | > loss: -0.45051 (-0.48005) | > log_mle: -0.57503 (-0.59797) | > loss_dur: 0.12453 (0.11791) | > amp_scaler: 2048.00000 (2502.75276) | > grad_norm: 7.87177 (9.45712) | > current_lr: 0.00031 | > step_time: 1.12580 (0.69467) | > loader_time: 0.00420 (0.00347) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02659 (-0.00247) | > avg_loss: -0.44303 (+0.00510) | > avg_log_mle: -0.57443 (+0.00563) | > avg_loss_dur: 0.13140 (-0.00053)  > EPOCH: 67/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 23:24:40)   --> STEP: 20/639 -- GLOBAL_STEP: 42900 | > loss: -0.62164 (-0.53567) | > log_mle: -0.76987 (-0.65853) | > loss_dur: 0.14823 (0.12287) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 43.89862 (13.97228) | > current_lr: 0.00031 | > step_time: 0.34320 (0.30330) | > loader_time: 0.00280 (0.00236)  --> STEP: 45/639 -- GLOBAL_STEP: 42925 | > loss: -0.47132 (-0.51790) | > log_mle: -0.57460 (-0.63586) | > loss_dur: 0.10329 (0.11796) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.12630 (11.77791) | > current_lr: 0.00031 | > step_time: 0.40670 (0.33790) | > loader_time: 0.00280 (0.00260)  --> STEP: 70/639 -- GLOBAL_STEP: 42950 | > loss: -0.51500 (-0.50870) | > log_mle: -0.63532 (-0.62591) | > loss_dur: 0.12032 (0.11721) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.83088 (11.19407) | > current_lr: 0.00031 | > step_time: 0.45600 (0.36549) | > loader_time: 0.00310 (0.00271)  --> STEP: 95/639 -- GLOBAL_STEP: 42975 | > loss: -0.56460 (-0.50667) | > log_mle: -0.68318 (-0.62183) | > loss_dur: 0.11857 (0.11515) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.89829 (10.37835) | > current_lr: 0.00031 | > step_time: 0.40180 (0.38166) | > loader_time: 0.00290 (0.00275)  --> STEP: 120/639 -- GLOBAL_STEP: 43000 | > loss: -0.54399 (-0.50517) | > log_mle: -0.68573 (-0.62131) | > loss_dur: 0.14174 (0.11614) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 16.80037 (10.48474) | > current_lr: 0.00031 | > step_time: 0.63610 (0.40138) | > loader_time: 0.00310 (0.00281)  --> STEP: 145/639 -- GLOBAL_STEP: 43025 | > loss: -0.50737 (-0.50539) | > log_mle: -0.62885 (-0.62177) | > loss_dur: 0.12148 (0.11638) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.18690 (10.60695) | > current_lr: 0.00031 | > step_time: 0.45750 (0.41531) | > loader_time: 0.00300 (0.00286)  --> STEP: 170/639 -- GLOBAL_STEP: 43050 | > loss: -0.49851 (-0.50442) | > log_mle: -0.61589 (-0.62073) | > loss_dur: 0.11738 (0.11631) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.37288 (10.52969) | > current_lr: 0.00030 | > step_time: 0.60160 (0.43036) | > loader_time: 0.00330 (0.00290)  --> STEP: 195/639 -- GLOBAL_STEP: 43075 | > loss: -0.46063 (-0.50372) | > log_mle: -0.55868 (-0.61952) | > loss_dur: 0.09805 (0.11580) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.42962 (10.38057) | > current_lr: 0.00030 | > step_time: 0.49160 (0.44617) | > loader_time: 0.00270 (0.00294)  --> STEP: 220/639 -- GLOBAL_STEP: 43100 | > loss: -0.50595 (-0.50157) | > log_mle: -0.61124 (-0.61732) | > loss_dur: 0.10529 (0.11575) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.38326 (10.43577) | > current_lr: 0.00030 | > step_time: 0.64730 (0.46379) | > loader_time: 0.00350 (0.00298)  --> STEP: 245/639 -- GLOBAL_STEP: 43125 | > loss: -0.48352 (-0.50267) | > log_mle: -0.59673 (-0.61833) | > loss_dur: 0.11321 (0.11566) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.17992 (10.57160) | > current_lr: 0.00030 | > step_time: 0.52300 (0.47807) | > loader_time: 0.00320 (0.00302)  --> STEP: 270/639 -- GLOBAL_STEP: 43150 | > loss: -0.54388 (-0.50233) | > log_mle: -0.66967 (-0.61837) | > loss_dur: 0.12579 (0.11604) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 16.25906 (10.86990) | > current_lr: 0.00030 | > step_time: 0.83600 (0.49617) | > loader_time: 0.00330 (0.00306)  --> STEP: 295/639 -- GLOBAL_STEP: 43175 | > loss: -0.44942 (-0.50038) | > log_mle: -0.55772 (-0.61656) | > loss_dur: 0.10830 (0.11618) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.93637 (10.75350) | > current_lr: 0.00030 | > step_time: 0.66040 (0.51073) | > loader_time: 0.00350 (0.00309)  --> STEP: 320/639 -- GLOBAL_STEP: 43200 | > loss: -0.46670 (-0.49873) | > log_mle: -0.60102 (-0.61536) | > loss_dur: 0.13432 (0.11663) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.89795 (10.64628) | > current_lr: 0.00030 | > step_time: 0.72300 (0.52653) | > loader_time: 0.00350 (0.00312)  --> STEP: 345/639 -- GLOBAL_STEP: 43225 | > loss: -0.48523 (-0.49682) | > log_mle: -0.59139 (-0.61374) | > loss_dur: 0.10616 (0.11693) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.12361 (10.59794) | > current_lr: 0.00030 | > step_time: 0.63000 (0.54129) | > loader_time: 0.00370 (0.00316)  --> STEP: 370/639 -- GLOBAL_STEP: 43250 | > loss: -0.48709 (-0.49518) | > log_mle: -0.60681 (-0.61185) | > loss_dur: 0.11972 (0.11667) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.65958 (10.39877) | > current_lr: 0.00030 | > step_time: 0.89370 (0.55296) | > loader_time: 0.00380 (0.00319)  --> STEP: 395/639 -- GLOBAL_STEP: 43275 | > loss: -0.44701 (-0.49338) | > log_mle: -0.57107 (-0.61007) | > loss_dur: 0.12406 (0.11669) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.78226 (10.20209) | > current_lr: 0.00030 | > step_time: 0.67390 (0.56461) | > loader_time: 0.00370 (0.00322)  --> STEP: 420/639 -- GLOBAL_STEP: 43300 | > loss: -0.45291 (-0.49147) | > log_mle: -0.56685 (-0.60839) | > loss_dur: 0.11394 (0.11692) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.58813 (10.06341) | > current_lr: 0.00030 | > step_time: 0.71890 (0.57765) | > loader_time: 0.00350 (0.00325)  --> STEP: 445/639 -- GLOBAL_STEP: 43325 | > loss: -0.45816 (-0.49006) | > log_mle: -0.57356 (-0.60701) | > loss_dur: 0.11540 (0.11694) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.20554 (9.98324) | > current_lr: 0.00030 | > step_time: 0.75920 (0.58913) | > loader_time: 0.00380 (0.00328)  --> STEP: 470/639 -- GLOBAL_STEP: 43350 | > loss: -0.47115 (-0.48832) | > log_mle: -0.58526 (-0.60526) | > loss_dur: 0.11411 (0.11693) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.91272 (9.84810) | > current_lr: 0.00030 | > step_time: 0.77400 (0.60096) | > loader_time: 0.00390 (0.00331)  --> STEP: 495/639 -- GLOBAL_STEP: 43375 | > loss: -0.42162 (-0.48689) | > log_mle: -0.52440 (-0.60391) | > loss_dur: 0.10278 (0.11703) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.83204 (9.78005) | > current_lr: 0.00030 | > step_time: 0.78590 (0.61325) | > loader_time: 0.00350 (0.00333)  --> STEP: 520/639 -- GLOBAL_STEP: 43400 | > loss: -0.46997 (-0.48584) | > log_mle: -0.58376 (-0.60286) | > loss_dur: 0.11379 (0.11702) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.91799 (9.70312) | > current_lr: 0.00030 | > step_time: 0.91290 (0.62614) | > loader_time: 0.00380 (0.00336)  --> STEP: 545/639 -- GLOBAL_STEP: 43425 | > loss: -0.47477 (-0.48452) | > log_mle: -0.61004 (-0.60171) | > loss_dur: 0.13527 (0.11719) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.54386 (9.63591) | > current_lr: 0.00030 | > step_time: 0.94840 (0.64113) | > loader_time: 0.00400 (0.00340)  --> STEP: 570/639 -- GLOBAL_STEP: 43450 | > loss: -0.45400 (-0.48306) | > log_mle: -0.55965 (-0.60034) | > loss_dur: 0.10564 (0.11728) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.63455 (9.54000) | > current_lr: 0.00030 | > step_time: 0.85420 (0.65487) | > loader_time: 0.00410 (0.00343)  --> STEP: 595/639 -- GLOBAL_STEP: 43475 | > loss: -0.43973 (-0.48211) | > log_mle: -0.55351 (-0.59952) | > loss_dur: 0.11378 (0.11741) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.25967 (9.49422) | > current_lr: 0.00030 | > step_time: 1.01300 (0.67124) | > loader_time: 0.00420 (0.00347)  --> STEP: 620/639 -- GLOBAL_STEP: 43500 | > loss: -0.44770 (-0.48096) | > log_mle: -0.56216 (-0.59845) | > loss_dur: 0.11445 (0.11749) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.46992 (9.42439) | > current_lr: 0.00030 | > step_time: 1.15370 (0.68736) | > loader_time: 0.00470 (0.00350) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03173 (+0.00515) | > avg_loss: -0.45430 (-0.01127) | > avg_log_mle: -0.58614 (-0.01171) | > avg_loss_dur: 0.13183 (+0.00044) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_43520.pth.tar  > EPOCH: 68/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 23:32:28)   --> STEP: 5/639 -- GLOBAL_STEP: 43525 | > loss: -0.54499 (-0.54535) | > log_mle: -0.68799 (-0.67207) | > loss_dur: 0.14300 (0.12672) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 17.02825 (11.86530) | > current_lr: 0.00030 | > step_time: 0.30710 (0.27944) | > loader_time: 0.00240 (0.00419)  --> STEP: 30/639 -- GLOBAL_STEP: 43550 | > loss: -0.52415 (-0.52363) | > log_mle: -0.64045 (-0.64482) | > loss_dur: 0.11630 (0.12119) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.17594 (11.98937) | > current_lr: 0.00030 | > step_time: 0.32130 (0.32171) | > loader_time: 0.00280 (0.00298)  --> STEP: 55/639 -- GLOBAL_STEP: 43575 | > loss: -0.48017 (-0.51553) | > log_mle: -0.57413 (-0.63475) | > loss_dur: 0.09395 (0.11922) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.16780 (10.52509) | > current_lr: 0.00030 | > step_time: 0.33040 (0.34671) | > loader_time: 0.00280 (0.00292)  --> STEP: 80/639 -- GLOBAL_STEP: 43600 | > loss: -0.48612 (-0.50948) | > log_mle: -0.61683 (-0.62686) | > loss_dur: 0.13071 (0.11739) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.54581 (9.91657) | > current_lr: 0.00030 | > step_time: 0.38540 (0.36793) | > loader_time: 0.00300 (0.00291)  --> STEP: 105/639 -- GLOBAL_STEP: 43625 | > loss: -0.46032 (-0.50811) | > log_mle: -0.57248 (-0.62458) | > loss_dur: 0.11216 (0.11646) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.25227 (9.89859) | > current_lr: 0.00030 | > step_time: 0.41460 (0.38563) | > loader_time: 0.00300 (0.00295)  --> STEP: 130/639 -- GLOBAL_STEP: 43650 | > loss: -0.48126 (-0.50627) | > log_mle: -0.60683 (-0.62290) | > loss_dur: 0.12557 (0.11663) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.87503 (10.03951) | > current_lr: 0.00030 | > step_time: 0.47110 (0.40120) | > loader_time: 0.00340 (0.00296)  --> STEP: 155/639 -- GLOBAL_STEP: 43675 | > loss: -0.48270 (-0.50507) | > log_mle: -0.60140 (-0.62173) | > loss_dur: 0.11870 (0.11666) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.70674 (10.14092) | > current_lr: 0.00030 | > step_time: 0.53580 (0.41696) | > loader_time: 0.00320 (0.00299)  --> STEP: 180/639 -- GLOBAL_STEP: 43700 | > loss: -0.52694 (-0.50589) | > log_mle: -0.64725 (-0.62293) | > loss_dur: 0.12031 (0.11704) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.58439 (10.44093) | > current_lr: 0.00030 | > step_time: 0.58080 (0.43353) | > loader_time: 0.00330 (0.00301)  --> STEP: 205/639 -- GLOBAL_STEP: 43725 | > loss: -0.46300 (-0.50277) | > log_mle: -0.57639 (-0.61934) | > loss_dur: 0.11339 (0.11657) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.82889 (10.07074) | > current_lr: 0.00030 | > step_time: 0.55630 (0.44787) | > loader_time: 0.00320 (0.00304)  --> STEP: 230/639 -- GLOBAL_STEP: 43750 | > loss: -0.49907 (-0.50304) | > log_mle: -0.61151 (-0.61945) | > loss_dur: 0.11245 (0.11641) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.73268 (10.07143) | > current_lr: 0.00030 | > step_time: 0.54840 (0.46544) | > loader_time: 0.00350 (0.00308)  --> STEP: 255/639 -- GLOBAL_STEP: 43775 | > loss: -0.45546 (-0.50216) | > log_mle: -0.58265 (-0.61848) | > loss_dur: 0.12719 (0.11631) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.03439 (9.97363) | > current_lr: 0.00030 | > step_time: 0.63540 (0.47951) | > loader_time: 0.00330 (0.00311)  --> STEP: 280/639 -- GLOBAL_STEP: 43800 | > loss: -0.48354 (-0.50208) | > log_mle: -0.60749 (-0.61881) | > loss_dur: 0.12395 (0.11673) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.91271 (10.07857) | > current_lr: 0.00030 | > step_time: 0.59910 (0.49732) | > loader_time: 0.00350 (0.00315)  --> STEP: 305/639 -- GLOBAL_STEP: 43825 | > loss: -0.46792 (-0.50035) | > log_mle: -0.58760 (-0.61727) | > loss_dur: 0.11968 (0.11691) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.40576 (9.93866) | > current_lr: 0.00030 | > step_time: 0.67540 (0.51242) | > loader_time: 0.00340 (0.00317)  --> STEP: 330/639 -- GLOBAL_STEP: 43850 | > loss: -0.47841 (-0.49800) | > log_mle: -0.60823 (-0.61532) | > loss_dur: 0.12982 (0.11732) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.34666 (9.88492) | > current_lr: 0.00030 | > step_time: 0.78620 (0.52743) | > loader_time: 0.00350 (0.00320)  --> STEP: 355/639 -- GLOBAL_STEP: 43875 | > loss: -0.43664 (-0.49640) | > log_mle: -0.54046 (-0.61352) | > loss_dur: 0.10381 (0.11712) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.39250 (9.75008) | > current_lr: 0.00030 | > step_time: 0.61830 (0.54097) | > loader_time: 0.00340 (0.00323)  --> STEP: 380/639 -- GLOBAL_STEP: 43900 | > loss: -0.46921 (-0.49443) | > log_mle: -0.57459 (-0.61145) | > loss_dur: 0.10538 (0.11702) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.95362 (9.59635) | > current_lr: 0.00030 | > step_time: 0.62520 (0.55394) | > loader_time: 0.00350 (0.00326)  --> STEP: 405/639 -- GLOBAL_STEP: 43925 | > loss: -0.46494 (-0.49306) | > log_mle: -0.58099 (-0.61031) | > loss_dur: 0.11605 (0.11725) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.56249 (9.50423) | > current_lr: 0.00030 | > step_time: 0.68290 (0.56710) | > loader_time: 0.00330 (0.00329)  --> STEP: 430/639 -- GLOBAL_STEP: 43950 | > loss: -0.42638 (-0.49119) | > log_mle: -0.53851 (-0.60856) | > loss_dur: 0.11213 (0.11736) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.56260 (9.38155) | > current_lr: 0.00030 | > step_time: 0.78660 (0.58012) | > loader_time: 0.00360 (0.00331)  --> STEP: 455/639 -- GLOBAL_STEP: 43975 | > loss: -0.45112 (-0.48943) | > log_mle: -0.56469 (-0.60666) | > loss_dur: 0.11357 (0.11723) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.67565 (9.27571) | > current_lr: 0.00030 | > step_time: 0.73050 (0.59093) | > loader_time: 0.00390 (0.00334)  --> STEP: 480/639 -- GLOBAL_STEP: 44000 | > loss: -0.40289 (-0.48774) | > log_mle: -0.52063 (-0.60513) | > loss_dur: 0.11775 (0.11740) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.34558 (9.15696) | > current_lr: 0.00030 | > step_time: 0.78550 (0.60407) | > loader_time: 0.00360 (0.00336)  --> STEP: 505/639 -- GLOBAL_STEP: 44025 | > loss: -0.47542 (-0.48648) | > log_mle: -0.58767 (-0.60378) | > loss_dur: 0.11225 (0.11730) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.16713 (9.06607) | > current_lr: 0.00030 | > step_time: 0.91010 (0.61663) | > loader_time: 0.00410 (0.00339)  --> STEP: 530/639 -- GLOBAL_STEP: 44050 | > loss: -0.45123 (-0.48524) | > log_mle: -0.57738 (-0.60263) | > loss_dur: 0.12615 (0.11740) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.33591 (9.05362) | > current_lr: 0.00030 | > step_time: 1.01920 (0.63108) | > loader_time: 0.00390 (0.00342)  --> STEP: 555/639 -- GLOBAL_STEP: 44075 | > loss: -0.45139 (-0.48383) | > log_mle: -0.56095 (-0.60122) | > loss_dur: 0.10956 (0.11739) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.02399 (8.94219) | > current_lr: 0.00030 | > step_time: 0.88870 (0.64553) | > loader_time: 0.00400 (0.00345)  --> STEP: 580/639 -- GLOBAL_STEP: 44100 | > loss: -0.46936 (-0.48268) | > log_mle: -0.58086 (-0.60016) | > loss_dur: 0.11150 (0.11749) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.75527 (8.87730) | > current_lr: 0.00030 | > step_time: 0.97270 (0.66113) | > loader_time: 0.00440 (0.00349)  --> STEP: 605/639 -- GLOBAL_STEP: 44125 | > loss: -0.44227 (-0.48157) | > log_mle: -0.54938 (-0.59921) | > loss_dur: 0.10711 (0.11764) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.98301 (8.81347) | > current_lr: 0.00030 | > step_time: 1.03570 (0.67751) | > loader_time: 0.00430 (0.00352)  --> STEP: 630/639 -- GLOBAL_STEP: 44150 | > loss: -0.44691 (-0.48059) | > log_mle: -0.57746 (-0.59838) | > loss_dur: 0.13056 (0.11779) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.03761 (8.78504) | > current_lr: 0.00030 | > step_time: 1.26110 (0.69551) | > loader_time: 0.00440 (0.00356) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.01716 (-0.01457) | > avg_loss: -0.45644 (-0.00214) | > avg_log_mle: -0.58576 (+0.00038) | > avg_loss_dur: 0.12932 (-0.00251) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_44160.pth.tar  > EPOCH: 69/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 23:40:17)   --> STEP: 15/639 -- GLOBAL_STEP: 44175 | > loss: -0.64664 (-0.54214) | > log_mle: -0.77201 (-0.66025) | > loss_dur: 0.12537 (0.11810) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 22.24375 (10.75228) | > current_lr: 0.00030 | > step_time: 0.33250 (0.29968) | > loader_time: 0.00240 (0.00248)  --> STEP: 40/639 -- GLOBAL_STEP: 44200 | > loss: -0.50939 (-0.52521) | > log_mle: -0.62082 (-0.64140) | > loss_dur: 0.11143 (0.11619) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.14375 (9.69493) | > current_lr: 0.00030 | > step_time: 0.35020 (0.33197) | > loader_time: 0.00290 (0.00263)  --> STEP: 65/639 -- GLOBAL_STEP: 44225 | > loss: -0.50676 (-0.51684) | > log_mle: -0.60745 (-0.63404) | > loss_dur: 0.10069 (0.11720) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.41740 (9.64676) | > current_lr: 0.00030 | > step_time: 0.38580 (0.36200) | > loader_time: 0.00300 (0.00274)  --> STEP: 90/639 -- GLOBAL_STEP: 44250 | > loss: -0.45567 (-0.50929) | > log_mle: -0.56600 (-0.62486) | > loss_dur: 0.11033 (0.11557) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.27175 (9.02792) | > current_lr: 0.00030 | > step_time: 0.51700 (0.37816) | > loader_time: 0.00320 (0.00280)  --> STEP: 115/639 -- GLOBAL_STEP: 44275 | > loss: -0.41757 (-0.50908) | > log_mle: -0.53282 (-0.62513) | > loss_dur: 0.11525 (0.11604) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.39491 (9.39505) | > current_lr: 0.00030 | > step_time: 0.38310 (0.39536) | > loader_time: 0.00300 (0.00284)  --> STEP: 140/639 -- GLOBAL_STEP: 44300 | > loss: -0.46536 (-0.50812) | > log_mle: -0.61070 (-0.62497) | > loss_dur: 0.14534 (0.11685) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.86650 (9.66641) | > current_lr: 0.00030 | > step_time: 0.55150 (0.40993) | > loader_time: 0.00320 (0.00288)  --> STEP: 165/639 -- GLOBAL_STEP: 44325 | > loss: -0.48052 (-0.50701) | > log_mle: -0.61837 (-0.62355) | > loss_dur: 0.13784 (0.11654) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.60607 (9.77249) | > current_lr: 0.00030 | > step_time: 0.49380 (0.42318) | > loader_time: 0.00300 (0.00292)  --> STEP: 190/639 -- GLOBAL_STEP: 44350 | > loss: -0.43909 (-0.50669) | > log_mle: -0.54904 (-0.62300) | > loss_dur: 0.10995 (0.11631) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.09836 (9.93609) | > current_lr: 0.00030 | > step_time: 0.51570 (0.44104) | > loader_time: 0.00350 (0.00296)  --> STEP: 215/639 -- GLOBAL_STEP: 44375 | > loss: -0.53989 (-0.50440) | > log_mle: -0.65191 (-0.62075) | > loss_dur: 0.11203 (0.11634) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.46417 (9.87215) | > current_lr: 0.00030 | > step_time: 0.55450 (0.45657) | > loader_time: 0.00310 (0.00299)  --> STEP: 240/639 -- GLOBAL_STEP: 44400 | > loss: -0.54591 (-0.50529) | > log_mle: -0.66386 (-0.62141) | > loss_dur: 0.11795 (0.11612) | > amp_scaler: 4096.00000 (2210.13333) | > grad_norm: 15.45971 (9.97050) | > current_lr: 0.00030 | > step_time: 0.55650 (0.47071) | > loader_time: 0.00320 (0.00302)  --> STEP: 265/639 -- GLOBAL_STEP: 44425 | > loss: -0.54252 (-0.50421) | > log_mle: -0.67675 (-0.62070) | > loss_dur: 0.13423 (0.11649) | > amp_scaler: 4096.00000 (2388.04528) | > grad_norm: 23.83734 (10.08685) | > current_lr: 0.00030 | > step_time: 0.63160 (0.48656) | > loader_time: 0.00330 (0.00306)  --> STEP: 290/639 -- GLOBAL_STEP: 44450 | > loss: -0.46785 (-0.50266) | > log_mle: -0.58412 (-0.61924) | > loss_dur: 0.11626 (0.11659) | > amp_scaler: 4096.00000 (2535.28276) | > grad_norm: 10.91739 (10.00605) | > current_lr: 0.00030 | > step_time: 0.63840 (0.50197) | > loader_time: 0.00340 (0.00309)  --> STEP: 315/639 -- GLOBAL_STEP: 44475 | > loss: -0.45238 (-0.50089) | > log_mle: -0.56091 (-0.61768) | > loss_dur: 0.10853 (0.11679) | > amp_scaler: 4096.00000 (2659.14921) | > grad_norm: 4.19949 (9.98058) | > current_lr: 0.00030 | > step_time: 0.61490 (0.51701) | > loader_time: 0.00340 (0.00312)  --> STEP: 340/639 -- GLOBAL_STEP: 44500 | > loss: -0.47001 (-0.49895) | > log_mle: -0.58303 (-0.61629) | > loss_dur: 0.11302 (0.11733) | > amp_scaler: 4096.00000 (2764.80000) | > grad_norm: 7.99177 (9.93825) | > current_lr: 0.00030 | > step_time: 0.62000 (0.53293) | > loader_time: 0.00350 (0.00315)  --> STEP: 365/639 -- GLOBAL_STEP: 44525 | > loss: -0.47607 (-0.49687) | > log_mle: -0.59217 (-0.61397) | > loss_dur: 0.11610 (0.11709) | > amp_scaler: 4096.00000 (2855.97808) | > grad_norm: 7.45374 (9.72554) | > current_lr: 0.00030 | > step_time: 0.81700 (0.54480) | > loader_time: 0.00400 (0.00317)  --> STEP: 390/639 -- GLOBAL_STEP: 44550 | > loss: -0.45990 (-0.49512) | > log_mle: -0.59066 (-0.61210) | > loss_dur: 0.13076 (0.11698) | > amp_scaler: 4096.00000 (2935.46667) | > grad_norm: 8.88513 (9.60972) | > current_lr: 0.00030 | > step_time: 0.71970 (0.55762) | > loader_time: 0.00350 (0.00320)  --> STEP: 415/639 -- GLOBAL_STEP: 44575 | > loss: -0.44578 (-0.49332) | > log_mle: -0.56492 (-0.61050) | > loss_dur: 0.11914 (0.11717) | > amp_scaler: 4096.00000 (3005.37831) | > grad_norm: 4.77964 (9.49067) | > current_lr: 0.00030 | > step_time: 0.74340 (0.57057) | > loader_time: 0.00340 (0.00323)  --> STEP: 440/639 -- GLOBAL_STEP: 44600 | > loss: -0.43150 (-0.49180) | > log_mle: -0.54886 (-0.60895) | > loss_dur: 0.11735 (0.11715) | > amp_scaler: 4096.00000 (3067.34545) | > grad_norm: 4.02367 (9.36242) | > current_lr: 0.00030 | > step_time: 0.73340 (0.58296) | > loader_time: 0.00370 (0.00326)  --> STEP: 465/639 -- GLOBAL_STEP: 44625 | > loss: -0.43643 (-0.48991) | > log_mle: -0.55366 (-0.60698) | > loss_dur: 0.11723 (0.11707) | > amp_scaler: 4096.00000 (3122.64946) | > grad_norm: 6.27728 (9.21474) | > current_lr: 0.00030 | > step_time: 0.77580 (0.59392) | > loader_time: 0.00350 (0.00329)  --> STEP: 490/639 -- GLOBAL_STEP: 44650 | > loss: -0.43491 (-0.48832) | > log_mle: -0.53708 (-0.60553) | > loss_dur: 0.10217 (0.11720) | > amp_scaler: 4096.00000 (3172.31020) | > grad_norm: 2.81924 (9.16461) | > current_lr: 0.00030 | > step_time: 0.76470 (0.60726) | > loader_time: 0.00370 (0.00331)  --> STEP: 515/639 -- GLOBAL_STEP: 44675 | > loss: -0.45912 (-0.48729) | > log_mle: -0.56617 (-0.60446) | > loss_dur: 0.10705 (0.11717) | > amp_scaler: 4096.00000 (3217.14951) | > grad_norm: 4.70373 (9.09020) | > current_lr: 0.00030 | > step_time: 1.08540 (0.62072) | > loader_time: 0.00400 (0.00334)  --> STEP: 540/639 -- GLOBAL_STEP: 44700 | > loss: -0.45870 (-0.48587) | > log_mle: -0.58204 (-0.60310) | > loss_dur: 0.12334 (0.11723) | > amp_scaler: 4096.00000 (3257.83704) | > grad_norm: 10.68509 (9.00337) | > current_lr: 0.00030 | > step_time: 0.91140 (0.63519) | > loader_time: 0.00420 (0.00338)  --> STEP: 565/639 -- GLOBAL_STEP: 44725 | > loss: -0.42802 (-0.48434) | > log_mle: -0.54291 (-0.60167) | > loss_dur: 0.11489 (0.11734) | > amp_scaler: 4096.00000 (3294.92389) | > grad_norm: 5.08105 (8.92280) | > current_lr: 0.00030 | > step_time: 0.99920 (0.64887) | > loader_time: 0.00410 (0.00342)  --> STEP: 590/639 -- GLOBAL_STEP: 44750 | > loss: -0.45452 (-0.48339) | > log_mle: -0.57625 (-0.60082) | > loss_dur: 0.12173 (0.11743) | > amp_scaler: 4096.00000 (3328.86780) | > grad_norm: 6.52266 (8.85857) | > current_lr: 0.00030 | > step_time: 0.96820 (0.66532) | > loader_time: 0.00430 (0.00346)  --> STEP: 615/639 -- GLOBAL_STEP: 44775 | > loss: -0.45529 (-0.48218) | > log_mle: -0.57732 (-0.59972) | > loss_dur: 0.12202 (0.11754) | > amp_scaler: 4096.00000 (3360.05203) | > grad_norm: 6.68962 (8.76836) | > current_lr: 0.00030 | > step_time: 1.03250 (0.68156) | > loader_time: 0.00440 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02836 (+0.01120) | > avg_loss: -0.44780 (+0.00864) | > avg_log_mle: -0.57923 (+0.00653) | > avg_loss_dur: 0.13142 (+0.00210)  > EPOCH: 70/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 23:47:58)   --> STEP: 0/639 -- GLOBAL_STEP: 44800 | > loss: -0.54305 (-0.54305) | > log_mle: -0.69690 (-0.69690) | > loss_dur: 0.15385 (0.15385) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.02726 (18.02726) | > current_lr: 0.00030 | > step_time: 0.53460 (0.53461) | > loader_time: 0.59620 (0.59623)  --> STEP: 25/639 -- GLOBAL_STEP: 44825 | > loss: -0.52312 (-0.53359) | > log_mle: -0.66000 (-0.65566) | > loss_dur: 0.13688 (0.12207) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.74693 (10.86860) | > current_lr: 0.00030 | > step_time: 0.38180 (0.32591) | > loader_time: 0.00290 (0.00248)  --> STEP: 50/639 -- GLOBAL_STEP: 44850 | > loss: -0.49365 (-0.52212) | > log_mle: -0.59000 (-0.64090) | > loss_dur: 0.09634 (0.11878) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.55292 (9.78239) | > current_lr: 0.00030 | > step_time: 0.32170 (0.34439) | > loader_time: 0.00290 (0.00277)  --> STEP: 75/639 -- GLOBAL_STEP: 44875 | > loss: -0.51136 (-0.51393) | > log_mle: -0.61522 (-0.63161) | > loss_dur: 0.10386 (0.11768) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.43458 (9.30063) | > current_lr: 0.00030 | > step_time: 0.40790 (0.37070) | > loader_time: 0.00310 (0.00284)  --> STEP: 100/639 -- GLOBAL_STEP: 44900 | > loss: -0.51190 (-0.51207) | > log_mle: -0.62635 (-0.62815) | > loss_dur: 0.11445 (0.11608) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.16133 (9.33336) | > current_lr: 0.00030 | > step_time: 0.50900 (0.38974) | > loader_time: 0.00310 (0.00288)  --> STEP: 125/639 -- GLOBAL_STEP: 44925 | > loss: -0.49455 (-0.50963) | > log_mle: -0.61499 (-0.62629) | > loss_dur: 0.12044 (0.11665) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.45142 (9.41662) | > current_lr: 0.00030 | > step_time: 0.40510 (0.40329) | > loader_time: 0.00300 (0.00291)  --> STEP: 150/639 -- GLOBAL_STEP: 44950 | > loss: -0.46873 (-0.50847) | > log_mle: -0.58189 (-0.62526) | > loss_dur: 0.11316 (0.11679) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.65735 (9.41705) | > current_lr: 0.00030 | > step_time: 0.42220 (0.41830) | > loader_time: 0.00300 (0.00293)  --> STEP: 175/639 -- GLOBAL_STEP: 44975 | > loss: -0.49572 (-0.50847) | > log_mle: -0.60040 (-0.62514) | > loss_dur: 0.10467 (0.11667) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.87140 (9.45886) | > current_lr: 0.00030 | > step_time: 0.55530 (0.43404) | > loader_time: 0.00340 (0.00297)  --> STEP: 200/639 -- GLOBAL_STEP: 45000 | > loss: -0.46662 (-0.50639) | > log_mle: -0.58332 (-0.62250) | > loss_dur: 0.11670 (0.11611) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.48812 (9.23423) | > current_lr: 0.00030 | > step_time: 0.53350 (0.44836) | > loader_time: 0.00320 (0.00301)  --> STEP: 225/639 -- GLOBAL_STEP: 45025 | > loss: -0.46332 (-0.50440) | > log_mle: -0.56664 (-0.62035) | > loss_dur: 0.10333 (0.11595) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.26598 (9.15685) | > current_lr: 0.00030 | > step_time: 0.52590 (0.46524) | > loader_time: 0.00330 (0.00306)  --> STEP: 250/639 -- GLOBAL_STEP: 45050 | > loss: -0.46601 (-0.50509) | > log_mle: -0.56888 (-0.62118) | > loss_dur: 0.10287 (0.11609) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.80663 (9.28299) | > current_lr: 0.00030 | > step_time: 0.58140 (0.48068) | > loader_time: 0.00350 (0.00309)  --> STEP: 275/639 -- GLOBAL_STEP: 45075 | > loss: -0.49853 (-0.50500) | > log_mle: -0.60577 (-0.62148) | > loss_dur: 0.10724 (0.11648) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.03486 (9.51749) | > current_lr: 0.00030 | > step_time: 0.86400 (0.49858) | > loader_time: 0.00380 (0.00312)  --> STEP: 300/639 -- GLOBAL_STEP: 45100 | > loss: -0.48945 (-0.50275) | > log_mle: -0.59851 (-0.61939) | > loss_dur: 0.10906 (0.11664) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.67656 (9.39369) | > current_lr: 0.00030 | > step_time: 0.67110 (0.51221) | > loader_time: 0.00360 (0.00315)  --> STEP: 325/639 -- GLOBAL_STEP: 45125 | > loss: -0.47842 (-0.50071) | > log_mle: -0.59911 (-0.61782) | > loss_dur: 0.12069 (0.11711) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.02733 (9.38486) | > current_lr: 0.00030 | > step_time: 0.64830 (0.52693) | > loader_time: 0.00360 (0.00318)  --> STEP: 350/639 -- GLOBAL_STEP: 45150 | > loss: -0.50340 (-0.49899) | > log_mle: -0.62808 (-0.61604) | > loss_dur: 0.12468 (0.11705) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.71744 (9.30481) | > current_lr: 0.00030 | > step_time: 0.83750 (0.54116) | > loader_time: 0.00370 (0.00321)  --> STEP: 375/639 -- GLOBAL_STEP: 45175 | > loss: -0.44888 (-0.49692) | > log_mle: -0.56512 (-0.61375) | > loss_dur: 0.11624 (0.11682) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.63841 (9.14191) | > current_lr: 0.00030 | > step_time: 0.68670 (0.55327) | > loader_time: 0.00360 (0.00324)  --> STEP: 400/639 -- GLOBAL_STEP: 45200 | > loss: -0.48276 (-0.49551) | > log_mle: -0.59368 (-0.61237) | > loss_dur: 0.11092 (0.11686) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.73758 (9.12638) | > current_lr: 0.00030 | > step_time: 0.86930 (0.56487) | > loader_time: 0.00380 (0.00326)  --> STEP: 425/639 -- GLOBAL_STEP: 45225 | > loss: -0.52985 (-0.49346) | > log_mle: -0.64868 (-0.61040) | > loss_dur: 0.11883 (0.11694) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 17.86738 (9.03896) | > current_lr: 0.00030 | > step_time: 0.91270 (0.57744) | > loader_time: 0.00350 (0.00329)  --> STEP: 450/639 -- GLOBAL_STEP: 45250 | > loss: -0.44510 (-0.49174) | > log_mle: -0.55069 (-0.60859) | > loss_dur: 0.10559 (0.11685) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.39284 (8.91526) | > current_lr: 0.00030 | > step_time: 0.80660 (0.58936) | > loader_time: 0.00380 (0.00331)  --> STEP: 475/639 -- GLOBAL_STEP: 45275 | > loss: -0.45875 (-0.48999) | > log_mle: -0.56103 (-0.60692) | > loss_dur: 0.10228 (0.11693) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.53187 (8.84111) | > current_lr: 0.00030 | > step_time: 0.75640 (0.60156) | > loader_time: 0.00360 (0.00333)  --> STEP: 500/639 -- GLOBAL_STEP: 45300 | > loss: -0.50877 (-0.48879) | > log_mle: -0.62794 (-0.60574) | > loss_dur: 0.11916 (0.11695) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.17212 (8.77747) | > current_lr: 0.00030 | > step_time: 0.94970 (0.61432) | > loader_time: 0.00400 (0.00335)  --> STEP: 525/639 -- GLOBAL_STEP: 45325 | > loss: -0.46447 (-0.48753) | > log_mle: -0.60361 (-0.60458) | > loss_dur: 0.13914 (0.11705) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.35179 (8.74770) | > current_lr: 0.00030 | > step_time: 1.03690 (0.62735) | > loader_time: 0.00420 (0.00339)  --> STEP: 550/639 -- GLOBAL_STEP: 45350 | > loss: -0.42120 (-0.48610) | > log_mle: -0.53811 (-0.60317) | > loss_dur: 0.11691 (0.11707) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.18588 (8.67524) | > current_lr: 0.00030 | > step_time: 0.95150 (0.64198) | > loader_time: 0.00420 (0.00342)  --> STEP: 575/639 -- GLOBAL_STEP: 45375 | > loss: -0.44140 (-0.48469) | > log_mle: -0.55833 (-0.60186) | > loss_dur: 0.11693 (0.11717) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.38800 (8.60435) | > current_lr: 0.00030 | > step_time: 0.86820 (0.65596) | > loader_time: 0.00410 (0.00345)  --> STEP: 600/639 -- GLOBAL_STEP: 45400 | > loss: -0.43486 (-0.48355) | > log_mle: -0.55562 (-0.60088) | > loss_dur: 0.12075 (0.11733) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.85822 (8.58683) | > current_lr: 0.00030 | > step_time: 1.02880 (0.67137) | > loader_time: 0.00450 (0.00349)  --> STEP: 625/639 -- GLOBAL_STEP: 45425 | > loss: -0.46804 (-0.48249) | > log_mle: -0.58490 (-0.59988) | > loss_dur: 0.11686 (0.11740) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.17826 (8.55753) | > current_lr: 0.00030 | > step_time: 1.06740 (0.68752) | > loader_time: 0.00430 (0.00353) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03763 (+0.00927) | > avg_loss: -0.45353 (-0.00573) | > avg_log_mle: -0.58470 (-0.00547) | > avg_loss_dur: 0.13117 (-0.00025)  > EPOCH: 71/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-23 23:55:37)   --> STEP: 10/639 -- GLOBAL_STEP: 45450 | > loss: -0.52562 (-0.54053) | > log_mle: -0.65658 (-0.65934) | > loss_dur: 0.13096 (0.11880) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.27660 (11.47600) | > current_lr: 0.00030 | > step_time: 0.28690 (0.29802) | > loader_time: 0.00280 (0.00315)  --> STEP: 35/639 -- GLOBAL_STEP: 45475 | > loss: -0.52910 (-0.52770) | > log_mle: -0.62381 (-0.64370) | > loss_dur: 0.09471 (0.11600) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.31314 (9.95658) | > current_lr: 0.00030 | > step_time: 0.32260 (0.32583) | > loader_time: 0.00270 (0.00281)  --> STEP: 60/639 -- GLOBAL_STEP: 45500 | > loss: -0.47898 (-0.51883) | > log_mle: -0.58009 (-0.63494) | > loss_dur: 0.10111 (0.11611) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.90512 (9.54053) | > current_lr: 0.00030 | > step_time: 0.34220 (0.35505) | > loader_time: 0.00240 (0.00280)  --> STEP: 85/639 -- GLOBAL_STEP: 45525 | > loss: -0.48035 (-0.51191) | > log_mle: -0.58022 (-0.62636) | > loss_dur: 0.09987 (0.11445) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.28121 (8.89304) | > current_lr: 0.00030 | > step_time: 0.40880 (0.37114) | > loader_time: 0.00300 (0.00284)  --> STEP: 110/639 -- GLOBAL_STEP: 45550 | > loss: -0.48134 (-0.51308) | > log_mle: -0.59808 (-0.62831) | > loss_dur: 0.11674 (0.11523) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.99031 (9.26180) | > current_lr: 0.00030 | > step_time: 0.39870 (0.39158) | > loader_time: 0.00310 (0.00288)  --> STEP: 135/639 -- GLOBAL_STEP: 45575 | > loss: -0.45210 (-0.51076) | > log_mle: -0.57317 (-0.62569) | > loss_dur: 0.12108 (0.11493) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.19872 (9.28969) | > current_lr: 0.00030 | > step_time: 0.46190 (0.40521) | > loader_time: 0.00310 (0.00291)  --> STEP: 160/639 -- GLOBAL_STEP: 45600 | > loss: -0.51545 (-0.50880) | > log_mle: -0.61768 (-0.62353) | > loss_dur: 0.10223 (0.11473) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.91426 (9.30904) | > current_lr: 0.00030 | > step_time: 0.46570 (0.41957) | > loader_time: 0.00270 (0.00294)  --> STEP: 185/639 -- GLOBAL_STEP: 45625 | > loss: -0.48383 (-0.51016) | > log_mle: -0.59531 (-0.62552) | > loss_dur: 0.11149 (0.11536) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.32892 (9.70499) | > current_lr: 0.00030 | > step_time: 0.58630 (0.43738) | > loader_time: 0.00340 (0.00298)  --> STEP: 210/639 -- GLOBAL_STEP: 45650 | > loss: -0.41614 (-0.50703) | > log_mle: -0.51779 (-0.62234) | > loss_dur: 0.10165 (0.11531) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.88296 (9.50820) | > current_lr: 0.00030 | > step_time: 0.50530 (0.45404) | > loader_time: 0.00310 (0.00301)  --> STEP: 235/639 -- GLOBAL_STEP: 45675 | > loss: -0.49647 (-0.50718) | > log_mle: -0.60830 (-0.62247) | > loss_dur: 0.11183 (0.11529) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.56738 (9.56577) | > current_lr: 0.00030 | > step_time: 0.54240 (0.46938) | > loader_time: 0.00320 (0.00304)  --> STEP: 260/639 -- GLOBAL_STEP: 45700 | > loss: -0.49337 (-0.50634) | > log_mle: -0.61360 (-0.62201) | > loss_dur: 0.12023 (0.11568) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.59435 (9.57972) | > current_lr: 0.00030 | > step_time: 0.60250 (0.48436) | > loader_time: 0.00340 (0.00308)  --> STEP: 285/639 -- GLOBAL_STEP: 45725 | > loss: -0.44106 (-0.50512) | > log_mle: -0.55816 (-0.62105) | > loss_dur: 0.11710 (0.11593) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.87528 (9.61795) | > current_lr: 0.00030 | > step_time: 0.62500 (0.50090) | > loader_time: 0.00340 (0.00312)  --> STEP: 310/639 -- GLOBAL_STEP: 45750 | > loss: -0.48752 (-0.50348) | > log_mle: -0.60048 (-0.61959) | > loss_dur: 0.11296 (0.11611) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.78372 (9.55903) | > current_lr: 0.00030 | > step_time: 0.57190 (0.51474) | > loader_time: 0.00330 (0.00315)  --> STEP: 335/639 -- GLOBAL_STEP: 45775 | > loss: -0.49641 (-0.50106) | > log_mle: -0.62461 (-0.61766) | > loss_dur: 0.12821 (0.11659) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.47968 (9.46325) | > current_lr: 0.00030 | > step_time: 0.65410 (0.53029) | > loader_time: 0.00370 (0.00318)  --> STEP: 360/639 -- GLOBAL_STEP: 45800 | > loss: -0.44642 (-0.49931) | > log_mle: -0.56203 (-0.61574) | > loss_dur: 0.11561 (0.11643) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.07948 (9.33074) | > current_lr: 0.00030 | > step_time: 0.71170 (0.54275) | > loader_time: 0.00370 (0.00321)  --> STEP: 385/639 -- GLOBAL_STEP: 45825 | > loss: -0.48931 (-0.49765) | > log_mle: -0.59627 (-0.61400) | > loss_dur: 0.10696 (0.11636) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.43479 (9.20862) | > current_lr: 0.00030 | > step_time: 0.74660 (0.55655) | > loader_time: 0.00340 (0.00324)  --> STEP: 410/639 -- GLOBAL_STEP: 45850 | > loss: -0.43800 (-0.49579) | > log_mle: -0.56788 (-0.61238) | > loss_dur: 0.12988 (0.11659) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.05597 (9.09380) | > current_lr: 0.00030 | > step_time: 0.78690 (0.56864) | > loader_time: 0.00370 (0.00327)  --> STEP: 435/639 -- GLOBAL_STEP: 45875 | > loss: -0.46996 (-0.49382) | > log_mle: -0.57651 (-0.61042) | > loss_dur: 0.10655 (0.11659) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.64504 (9.00672) | > current_lr: 0.00030 | > step_time: 0.69870 (0.58146) | > loader_time: 0.00360 (0.00330)  --> STEP: 460/639 -- GLOBAL_STEP: 45900 | > loss: -0.42710 (-0.49187) | > log_mle: -0.54530 (-0.60846) | > loss_dur: 0.11820 (0.11659) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.58997 (8.85061) | > current_lr: 0.00030 | > step_time: 0.87580 (0.59367) | > loader_time: 0.00390 (0.00332)  --> STEP: 485/639 -- GLOBAL_STEP: 45925 | > loss: -0.46468 (-0.49034) | > log_mle: -0.57703 (-0.60710) | > loss_dur: 0.11234 (0.11677) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.49076 (8.79780) | > current_lr: 0.00030 | > step_time: 0.84730 (0.60627) | > loader_time: 0.00400 (0.00335)  --> STEP: 510/639 -- GLOBAL_STEP: 45950 | > loss: -0.45889 (-0.48911) | > log_mle: -0.58105 (-0.60585) | > loss_dur: 0.12216 (0.11674) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.98126 (8.72895) | > current_lr: 0.00030 | > step_time: 0.85990 (0.61901) | > loader_time: 0.00410 (0.00338)  --> STEP: 535/639 -- GLOBAL_STEP: 45975 | > loss: -0.44053 (-0.48769) | > log_mle: -0.56961 (-0.60452) | > loss_dur: 0.12908 (0.11683) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.88901 (8.65687) | > current_lr: 0.00030 | > step_time: 0.89850 (0.63273) | > loader_time: 0.00410 (0.00341)  --> STEP: 560/639 -- GLOBAL_STEP: 46000 | > loss: -0.45340 (-0.48634) | > log_mle: -0.57989 (-0.60327) | > loss_dur: 0.12648 (0.11692) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.50927 (8.62568) | > current_lr: 0.00029 | > step_time: 0.90450 (0.64781) | > loader_time: 0.00420 (0.00344)  --> STEP: 585/639 -- GLOBAL_STEP: 46025 | > loss: -0.45984 (-0.48516) | > log_mle: -0.57918 (-0.60214) | > loss_dur: 0.11934 (0.11699) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.35767 (8.55216) | > current_lr: 0.00029 | > step_time: 1.06760 (0.66297) | > loader_time: 0.00450 (0.00348)  --> STEP: 610/639 -- GLOBAL_STEP: 46050 | > loss: -0.43817 (-0.48394) | > log_mle: -0.54877 (-0.60105) | > loss_dur: 0.11060 (0.11711) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.05982 (8.50536) | > current_lr: 0.00029 | > step_time: 1.09590 (0.67946) | > loader_time: 0.00450 (0.00352)  --> STEP: 635/639 -- GLOBAL_STEP: 46075 | > loss: -0.45855 (-0.48299) | > log_mle: -0.57925 (-0.60039) | > loss_dur: 0.12071 (0.11740) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.06824 (8.48969) | > current_lr: 0.00029 | > step_time: 1.12330 (0.69654) | > loader_time: 0.00430 (0.00356) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02381 (-0.01381) | > avg_loss: -0.45295 (+0.00058) | > avg_log_mle: -0.58495 (-0.00025) | > avg_loss_dur: 0.13201 (+0.00083)  > EPOCH: 72/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 00:03:18)   --> STEP: 20/639 -- GLOBAL_STEP: 46100 | > loss: -0.60842 (-0.54483) | > log_mle: -0.77213 (-0.66637) | > loss_dur: 0.16371 (0.12154) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 36.28737 (12.99409) | > current_lr: 0.00029 | > step_time: 0.37210 (0.30451) | > loader_time: 0.00270 (0.00251)  --> STEP: 45/639 -- GLOBAL_STEP: 46125 | > loss: -0.48097 (-0.52528) | > log_mle: -0.58638 (-0.64191) | > loss_dur: 0.10541 (0.11663) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.51791 (11.45782) | > current_lr: 0.00029 | > step_time: 0.40300 (0.33874) | > loader_time: 0.00300 (0.00266)  --> STEP: 70/639 -- GLOBAL_STEP: 46150 | > loss: -0.51636 (-0.51588) | > log_mle: -0.64092 (-0.63242) | > loss_dur: 0.12456 (0.11653) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.01990 (10.68312) | > current_lr: 0.00029 | > step_time: 0.39070 (0.36254) | > loader_time: 0.00310 (0.00276)  --> STEP: 95/639 -- GLOBAL_STEP: 46175 | > loss: -0.57616 (-0.51347) | > log_mle: -0.68829 (-0.62771) | > loss_dur: 0.11212 (0.11425) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.78587 (9.96645) | > current_lr: 0.00029 | > step_time: 0.50400 (0.37746) | > loader_time: 0.00300 (0.00279)  --> STEP: 120/639 -- GLOBAL_STEP: 46200 | > loss: -0.54271 (-0.51107) | > log_mle: -0.68437 (-0.62657) | > loss_dur: 0.14165 (0.11549) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 19.46138 (10.04315) | > current_lr: 0.00029 | > step_time: 0.56880 (0.39750) | > loader_time: 0.00280 (0.00283)  --> STEP: 145/639 -- GLOBAL_STEP: 46225 | > loss: -0.50790 (-0.51072) | > log_mle: -0.63360 (-0.62660) | > loss_dur: 0.12570 (0.11588) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.14971 (10.27477) | > current_lr: 0.00029 | > step_time: 0.46940 (0.41352) | > loader_time: 0.00310 (0.00287)  --> STEP: 170/639 -- GLOBAL_STEP: 46250 | > loss: -0.50658 (-0.50985) | > log_mle: -0.61950 (-0.62544) | > loss_dur: 0.11291 (0.11559) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.03491 (10.26086) | > current_lr: 0.00029 | > step_time: 0.54890 (0.42849) | > loader_time: 0.00330 (0.00290)  --> STEP: 195/639 -- GLOBAL_STEP: 46275 | > loss: -0.45755 (-0.50905) | > log_mle: -0.56157 (-0.62418) | > loss_dur: 0.10402 (0.11513) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.66774 (10.33468) | > current_lr: 0.00029 | > step_time: 0.50660 (0.44448) | > loader_time: 0.00300 (0.00294)  --> STEP: 220/639 -- GLOBAL_STEP: 46300 | > loss: -0.50933 (-0.50655) | > log_mle: -0.61356 (-0.62175) | > loss_dur: 0.10423 (0.11520) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.64359 (10.31345) | > current_lr: 0.00029 | > step_time: 0.66860 (0.46118) | > loader_time: 0.00350 (0.00298)  --> STEP: 245/639 -- GLOBAL_STEP: 46325 | > loss: -0.49021 (-0.50734) | > log_mle: -0.60224 (-0.62248) | > loss_dur: 0.11204 (0.11514) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.07672 (10.47978) | > current_lr: 0.00029 | > step_time: 0.52780 (0.47633) | > loader_time: 0.00330 (0.00302)  --> STEP: 270/639 -- GLOBAL_STEP: 46350 | > loss: -0.53383 (-0.50680) | > log_mle: -0.65799 (-0.62245) | > loss_dur: 0.12416 (0.11564) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 23.63257 (10.60453) | > current_lr: 0.00029 | > step_time: 0.83400 (0.49319) | > loader_time: 0.00380 (0.00306)  --> STEP: 295/639 -- GLOBAL_STEP: 46375 | > loss: -0.46173 (-0.50462) | > log_mle: -0.56023 (-0.62051) | > loss_dur: 0.09849 (0.11589) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.11745 (10.40919) | > current_lr: 0.00029 | > step_time: 0.64940 (0.50709) | > loader_time: 0.00350 (0.00308)  --> STEP: 320/639 -- GLOBAL_STEP: 46400 | > loss: -0.47211 (-0.50295) | > log_mle: -0.60489 (-0.61920) | > loss_dur: 0.13278 (0.11625) | > amp_scaler: 8192.00000 (4339.20000) | > grad_norm: 11.41499 (10.35720) | > current_lr: 0.00029 | > step_time: 0.74520 (0.52301) | > loader_time: 0.00360 (0.00312)  --> STEP: 345/639 -- GLOBAL_STEP: 46425 | > loss: -0.48747 (-0.50090) | > log_mle: -0.59391 (-0.61748) | > loss_dur: 0.10644 (0.11658) | > amp_scaler: 8192.00000 (4618.38841) | > grad_norm: 7.92944 (10.27972) | > current_lr: 0.00029 | > step_time: 0.66310 (0.53686) | > loader_time: 0.00360 (0.00315)  --> STEP: 370/639 -- GLOBAL_STEP: 46450 | > loss: -0.48743 (-0.49902) | > log_mle: -0.60516 (-0.61542) | > loss_dur: 0.11773 (0.11639) | > amp_scaler: 8192.00000 (4859.84865) | > grad_norm: 9.62488 (10.09570) | > current_lr: 0.00029 | > step_time: 0.88690 (0.54943) | > loader_time: 0.00360 (0.00316)  --> STEP: 395/639 -- GLOBAL_STEP: 46475 | > loss: -0.45562 (-0.49718) | > log_mle: -0.57199 (-0.61352) | > loss_dur: 0.11637 (0.11634) | > amp_scaler: 8192.00000 (5070.74430) | > grad_norm: 8.24442 (9.92973) | > current_lr: 0.00029 | > step_time: 0.70160 (0.56147) | > loader_time: 0.00310 (0.00319)  --> STEP: 420/639 -- GLOBAL_STEP: 46500 | > loss: -0.46270 (-0.49518) | > log_mle: -0.57295 (-0.61176) | > loss_dur: 0.11024 (0.11658) | > amp_scaler: 8192.00000 (5256.53333) | > grad_norm: 6.03270 (9.85426) | > current_lr: 0.00029 | > step_time: 0.70820 (0.57376) | > loader_time: 0.00330 (0.00322)  --> STEP: 445/639 -- GLOBAL_STEP: 46525 | > loss: -0.46625 (-0.49371) | > log_mle: -0.57812 (-0.61032) | > loss_dur: 0.11187 (0.11661) | > amp_scaler: 8192.00000 (5421.44719) | > grad_norm: 6.85692 (9.73578) | > current_lr: 0.00029 | > step_time: 0.74230 (0.58612) | > loader_time: 0.00360 (0.00325)  --> STEP: 470/639 -- GLOBAL_STEP: 46550 | > loss: -0.47818 (-0.49198) | > log_mle: -0.58912 (-0.60852) | > loss_dur: 0.11094 (0.11654) | > amp_scaler: 8192.00000 (5568.81702) | > grad_norm: 9.43415 (9.56734) | > current_lr: 0.00029 | > step_time: 0.76460 (0.59730) | > loader_time: 0.00370 (0.00327)  --> STEP: 495/639 -- GLOBAL_STEP: 46575 | > loss: -0.42948 (-0.49052) | > log_mle: -0.52665 (-0.60715) | > loss_dur: 0.09717 (0.11664) | > amp_scaler: 8192.00000 (5701.30101) | > grad_norm: 2.56151 (9.48195) | > current_lr: 0.00029 | > step_time: 0.77240 (0.60976) | > loader_time: 0.00360 (0.00329)  --> STEP: 520/639 -- GLOBAL_STEP: 46600 | > loss: -0.46391 (-0.48946) | > log_mle: -0.58077 (-0.60604) | > loss_dur: 0.11686 (0.11658) | > amp_scaler: 8192.00000 (5821.04615) | > grad_norm: 7.03541 (9.40357) | > current_lr: 0.00029 | > step_time: 0.79940 (0.62224) | > loader_time: 0.00380 (0.00332)  --> STEP: 545/639 -- GLOBAL_STEP: 46625 | > loss: -0.48608 (-0.48821) | > log_mle: -0.61732 (-0.60490) | > loss_dur: 0.13124 (0.11669) | > amp_scaler: 8192.00000 (5929.80550) | > grad_norm: 9.97127 (9.31487) | > current_lr: 0.00029 | > step_time: 0.89560 (0.63682) | > loader_time: 0.00390 (0.00336)  --> STEP: 570/639 -- GLOBAL_STEP: 46650 | > loss: -0.45964 (-0.48676) | > log_mle: -0.56171 (-0.60349) | > loss_dur: 0.10208 (0.11674) | > amp_scaler: 8192.00000 (6029.02456) | > grad_norm: 4.59517 (9.24159) | > current_lr: 0.00029 | > step_time: 0.84310 (0.65065) | > loader_time: 0.00390 (0.00340)  --> STEP: 595/639 -- GLOBAL_STEP: 46675 | > loss: -0.44455 (-0.48573) | > log_mle: -0.55357 (-0.60261) | > loss_dur: 0.10902 (0.11688) | > amp_scaler: 8192.00000 (6119.90588) | > grad_norm: 6.91168 (9.18485) | > current_lr: 0.00029 | > step_time: 0.99680 (0.66579) | > loader_time: 0.00640 (0.00345)  --> STEP: 620/639 -- GLOBAL_STEP: 46700 | > loss: -0.45423 (-0.48456) | > log_mle: -0.56395 (-0.60151) | > loss_dur: 0.10972 (0.11695) | > amp_scaler: 8192.00000 (6203.45806) | > grad_norm: 6.03386 (9.10890) | > current_lr: 0.00029 | > step_time: 1.10300 (0.68271) | > loader_time: 0.00800 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03428 (+0.01046) | > avg_loss: -0.45667 (-0.00372) | > avg_log_mle: -0.58701 (-0.00206) | > avg_loss_dur: 0.13034 (-0.00166) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_46720.pth.tar  > EPOCH: 73/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 00:11:04)   --> STEP: 5/639 -- GLOBAL_STEP: 46725 | > loss: -0.54719 (-0.54830) | > log_mle: -0.68932 (-0.67046) | > loss_dur: 0.14213 (0.12215) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 29.73863 (14.99146) | > current_lr: 0.00029 | > step_time: 0.30720 (0.27684) | > loader_time: 0.00280 (0.00237)  --> STEP: 30/639 -- GLOBAL_STEP: 46750 | > loss: -0.52230 (-0.53113) | > log_mle: -0.64337 (-0.65086) | > loss_dur: 0.12107 (0.11973) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.33083 (10.89963) | > current_lr: 0.00029 | > step_time: 0.39490 (0.33627) | > loader_time: 0.00290 (0.00259)  --> STEP: 55/639 -- GLOBAL_STEP: 46775 | > loss: -0.47941 (-0.52127) | > log_mle: -0.57880 (-0.63886) | > loss_dur: 0.09939 (0.11759) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.92892 (10.05312) | > current_lr: 0.00029 | > step_time: 0.33310 (0.35143) | > loader_time: 0.00300 (0.00271)  --> STEP: 80/639 -- GLOBAL_STEP: 46800 | > loss: -0.48537 (-0.51491) | > log_mle: -0.61869 (-0.63094) | > loss_dur: 0.13332 (0.11602) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.72634 (9.67964) | > current_lr: 0.00029 | > step_time: 0.39320 (0.37122) | > loader_time: 0.00290 (0.00277)  --> STEP: 105/639 -- GLOBAL_STEP: 46825 | > loss: -0.45616 (-0.51261) | > log_mle: -0.57459 (-0.62816) | > loss_dur: 0.11843 (0.11555) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.47209 (10.05779) | > current_lr: 0.00029 | > step_time: 0.41130 (0.38760) | > loader_time: 0.00310 (0.00282)  --> STEP: 130/639 -- GLOBAL_STEP: 46850 | > loss: -0.48196 (-0.51028) | > log_mle: -0.60994 (-0.62634) | > loss_dur: 0.12799 (0.11606) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.10154 (10.17699) | > current_lr: 0.00029 | > step_time: 0.48280 (0.40139) | > loader_time: 0.00290 (0.00286)  --> STEP: 155/639 -- GLOBAL_STEP: 46875 | > loss: -0.49203 (-0.50936) | > log_mle: -0.60456 (-0.62544) | > loss_dur: 0.11253 (0.11608) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.88647 (10.29561) | > current_lr: 0.00029 | > step_time: 0.53400 (0.41696) | > loader_time: 0.00330 (0.00289)  --> STEP: 180/639 -- GLOBAL_STEP: 46900 | > loss: -0.53315 (-0.51049) | > log_mle: -0.65538 (-0.62675) | > loss_dur: 0.12223 (0.11626) | > amp_scaler: 2048.00000 (3891.20000) | > grad_norm: 10.41157 (10.52713) | > current_lr: 0.00029 | > step_time: 0.56860 (0.43211) | > loader_time: 0.00340 (0.00293)  --> STEP: 205/639 -- GLOBAL_STEP: 46925 | > loss: -0.46213 (-0.50718) | > log_mle: -0.57550 (-0.62288) | > loss_dur: 0.11337 (0.11570) | > amp_scaler: 2048.00000 (3666.41951) | > grad_norm: 8.69477 (10.29560) | > current_lr: 0.00029 | > step_time: 0.60190 (0.44660) | > loader_time: 0.00350 (0.00297)  --> STEP: 230/639 -- GLOBAL_STEP: 46950 | > loss: -0.49674 (-0.50689) | > log_mle: -0.61150 (-0.62273) | > loss_dur: 0.11476 (0.11583) | > amp_scaler: 2048.00000 (3490.50435) | > grad_norm: 19.74712 (10.68574) | > current_lr: 0.00029 | > step_time: 0.55340 (0.46324) | > loader_time: 0.00330 (0.00300)  --> STEP: 255/639 -- GLOBAL_STEP: 46975 | > loss: -0.47065 (-0.50611) | > log_mle: -0.58867 (-0.62178) | > loss_dur: 0.11802 (0.11567) | > amp_scaler: 2048.00000 (3349.08235) | > grad_norm: 7.68755 (10.66815) | > current_lr: 0.00029 | > step_time: 0.63890 (0.47729) | > loader_time: 0.00340 (0.00303)  --> STEP: 280/639 -- GLOBAL_STEP: 47000 | > loss: -0.48794 (-0.50597) | > log_mle: -0.61186 (-0.62211) | > loss_dur: 0.12392 (0.11614) | > amp_scaler: 2048.00000 (3232.91429) | > grad_norm: 10.63833 (10.90319) | > current_lr: 0.00029 | > step_time: 0.59720 (0.49458) | > loader_time: 0.00350 (0.00307)  --> STEP: 305/639 -- GLOBAL_STEP: 47025 | > loss: -0.46086 (-0.50421) | > log_mle: -0.58690 (-0.62056) | > loss_dur: 0.12605 (0.11635) | > amp_scaler: 2048.00000 (3135.79016) | > grad_norm: 16.48268 (10.79088) | > current_lr: 0.00029 | > step_time: 0.68290 (0.50979) | > loader_time: 0.00350 (0.00311)  --> STEP: 330/639 -- GLOBAL_STEP: 47050 | > loss: -0.47962 (-0.50181) | > log_mle: -0.61155 (-0.61857) | > loss_dur: 0.13194 (0.11675) | > amp_scaler: 2048.00000 (3053.38182) | > grad_norm: 10.17253 (10.66192) | > current_lr: 0.00029 | > step_time: 0.72100 (0.52452) | > loader_time: 0.00350 (0.00314)  --> STEP: 355/639 -- GLOBAL_STEP: 47075 | > loss: -0.43048 (-0.50013) | > log_mle: -0.54184 (-0.61670) | > loss_dur: 0.11136 (0.11657) | > amp_scaler: 2048.00000 (2982.58028) | > grad_norm: 2.24504 (10.49734) | > current_lr: 0.00029 | > step_time: 0.59400 (0.53799) | > loader_time: 0.00340 (0.00316)  --> STEP: 380/639 -- GLOBAL_STEP: 47100 | > loss: -0.47625 (-0.49820) | > log_mle: -0.57759 (-0.61454) | > loss_dur: 0.10134 (0.11634) | > amp_scaler: 2048.00000 (2921.09474) | > grad_norm: 4.55598 (10.28741) | > current_lr: 0.00029 | > step_time: 0.62600 (0.55132) | > loader_time: 0.00330 (0.00319)  --> STEP: 405/639 -- GLOBAL_STEP: 47125 | > loss: -0.46549 (-0.49664) | > log_mle: -0.58066 (-0.61325) | > loss_dur: 0.11517 (0.11661) | > amp_scaler: 2048.00000 (2867.20000) | > grad_norm: 7.22629 (10.21050) | > current_lr: 0.00029 | > step_time: 0.73500 (0.56281) | > loader_time: 0.00330 (0.00322)  --> STEP: 430/639 -- GLOBAL_STEP: 47150 | > loss: -0.43765 (-0.49466) | > log_mle: -0.54228 (-0.61143) | > loss_dur: 0.10463 (0.11677) | > amp_scaler: 2048.00000 (2819.57209) | > grad_norm: 3.12998 (10.08414) | > current_lr: 0.00029 | > step_time: 0.80210 (0.57645) | > loader_time: 0.00370 (0.00325)  --> STEP: 455/639 -- GLOBAL_STEP: 47175 | > loss: -0.45312 (-0.49282) | > log_mle: -0.56733 (-0.60950) | > loss_dur: 0.11421 (0.11668) | > amp_scaler: 2048.00000 (2777.17802) | > grad_norm: 5.01624 (9.91883) | > current_lr: 0.00029 | > step_time: 0.77650 (0.58748) | > loader_time: 0.00380 (0.00328)  --> STEP: 480/639 -- GLOBAL_STEP: 47200 | > loss: -0.41254 (-0.49119) | > log_mle: -0.52393 (-0.60798) | > loss_dur: 0.11139 (0.11679) | > amp_scaler: 2048.00000 (2739.20000) | > grad_norm: 3.63374 (9.81581) | > current_lr: 0.00029 | > step_time: 0.80310 (0.60015) | > loader_time: 0.00350 (0.00330)  --> STEP: 505/639 -- GLOBAL_STEP: 47225 | > loss: -0.47527 (-0.48988) | > log_mle: -0.58994 (-0.60662) | > loss_dur: 0.11466 (0.11674) | > amp_scaler: 2048.00000 (2704.98218) | > grad_norm: 10.97672 (9.72820) | > current_lr: 0.00029 | > step_time: 0.91670 (0.61273) | > loader_time: 0.00370 (0.00333)  --> STEP: 530/639 -- GLOBAL_STEP: 47250 | > loss: -0.45489 (-0.48854) | > log_mle: -0.57780 (-0.60537) | > loss_dur: 0.12291 (0.11683) | > amp_scaler: 2048.00000 (2673.99245) | > grad_norm: 6.77378 (9.67870) | > current_lr: 0.00029 | > step_time: 1.01010 (0.62652) | > loader_time: 0.00410 (0.00336)  --> STEP: 555/639 -- GLOBAL_STEP: 47275 | > loss: -0.45356 (-0.48706) | > log_mle: -0.56477 (-0.60390) | > loss_dur: 0.11121 (0.11684) | > amp_scaler: 2048.00000 (2645.79459) | > grad_norm: 5.91484 (9.59012) | > current_lr: 0.00029 | > step_time: 0.87610 (0.64108) | > loader_time: 0.00390 (0.00339)  --> STEP: 580/639 -- GLOBAL_STEP: 47300 | > loss: -0.47393 (-0.48593) | > log_mle: -0.58473 (-0.60285) | > loss_dur: 0.11080 (0.11692) | > amp_scaler: 2048.00000 (2620.02759) | > grad_norm: 8.10147 (9.52971) | > current_lr: 0.00029 | > step_time: 0.96780 (0.65633) | > loader_time: 0.00440 (0.00343)  --> STEP: 605/639 -- GLOBAL_STEP: 47325 | > loss: -0.44806 (-0.48478) | > log_mle: -0.55113 (-0.60185) | > loss_dur: 0.10308 (0.11707) | > amp_scaler: 2048.00000 (2596.39008) | > grad_norm: 5.57535 (9.47681) | > current_lr: 0.00029 | > step_time: 0.97540 (0.67236) | > loader_time: 0.00470 (0.00347)  --> STEP: 630/639 -- GLOBAL_STEP: 47350 | > loss: -0.45550 (-0.48382) | > log_mle: -0.58523 (-0.60103) | > loss_dur: 0.12973 (0.11721) | > amp_scaler: 2048.00000 (2574.62857) | > grad_norm: 8.49755 (9.44447) | > current_lr: 0.00029 | > step_time: 1.19430 (0.68961) | > loader_time: 0.00480 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02834 (-0.00593) | > avg_loss: -0.45274 (+0.00393) | > avg_log_mle: -0.58526 (+0.00175) | > avg_loss_dur: 0.13252 (+0.00218)  > EPOCH: 74/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 00:18:42)   --> STEP: 15/639 -- GLOBAL_STEP: 47375 | > loss: -0.66035 (-0.54764) | > log_mle: -0.77875 (-0.66500) | > loss_dur: 0.11839 (0.11736) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 23.15856 (11.12529) | > current_lr: 0.00029 | > step_time: 0.32670 (0.29616) | > loader_time: 0.00250 (0.00368)  --> STEP: 40/639 -- GLOBAL_STEP: 47400 | > loss: -0.50431 (-0.52962) | > log_mle: -0.61528 (-0.64487) | > loss_dur: 0.11097 (0.11525) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.11707 (10.86166) | > current_lr: 0.00029 | > step_time: 0.42220 (0.33169) | > loader_time: 0.00270 (0.00309)  --> STEP: 65/639 -- GLOBAL_STEP: 47425 | > loss: -0.50615 (-0.51990) | > log_mle: -0.60463 (-0.63639) | > loss_dur: 0.09848 (0.11650) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.64443 (10.54442) | > current_lr: 0.00029 | > step_time: 0.41700 (0.36020) | > loader_time: 0.00260 (0.00301)  --> STEP: 90/639 -- GLOBAL_STEP: 47450 | > loss: -0.45570 (-0.51284) | > log_mle: -0.56475 (-0.62764) | > loss_dur: 0.10905 (0.11480) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.02630 (9.68433) | > current_lr: 0.00029 | > step_time: 0.51470 (0.37535) | > loader_time: 0.00320 (0.00300)  --> STEP: 115/639 -- GLOBAL_STEP: 47475 | > loss: -0.41970 (-0.51303) | > log_mle: -0.53589 (-0.62789) | > loss_dur: 0.11620 (0.11486) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.07370 (9.94000) | > current_lr: 0.00029 | > step_time: 0.39840 (0.39505) | > loader_time: 0.00290 (0.00299)  --> STEP: 140/639 -- GLOBAL_STEP: 47500 | > loss: -0.47984 (-0.51236) | > log_mle: -0.61550 (-0.62783) | > loss_dur: 0.13566 (0.11547) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.53159 (10.11921) | > current_lr: 0.00029 | > step_time: 0.55720 (0.40978) | > loader_time: 0.00310 (0.00300)  --> STEP: 165/639 -- GLOBAL_STEP: 47525 | > loss: -0.49238 (-0.51127) | > log_mle: -0.62091 (-0.62647) | > loss_dur: 0.12853 (0.11520) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.04275 (10.11786) | > current_lr: 0.00029 | > step_time: 0.49580 (0.42377) | > loader_time: 0.00310 (0.00302)  --> STEP: 190/639 -- GLOBAL_STEP: 47550 | > loss: -0.44072 (-0.51113) | > log_mle: -0.55119 (-0.62598) | > loss_dur: 0.11047 (0.11485) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.58754 (10.18581) | > current_lr: 0.00029 | > step_time: 0.52790 (0.44198) | > loader_time: 0.00310 (0.00304)  --> STEP: 215/639 -- GLOBAL_STEP: 47575 | > loss: -0.54632 (-0.50881) | > log_mle: -0.65477 (-0.62376) | > loss_dur: 0.10844 (0.11495) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.26175 (10.17391) | > current_lr: 0.00029 | > step_time: 0.54340 (0.45876) | > loader_time: 0.00330 (0.00306)  --> STEP: 240/639 -- GLOBAL_STEP: 47600 | > loss: -0.54488 (-0.50931) | > log_mle: -0.66871 (-0.62429) | > loss_dur: 0.12384 (0.11498) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.26465 (10.38972) | > current_lr: 0.00029 | > step_time: 0.56220 (0.47316) | > loader_time: 0.00280 (0.00309)  --> STEP: 265/639 -- GLOBAL_STEP: 47625 | > loss: -0.54632 (-0.50799) | > log_mle: -0.68324 (-0.62349) | > loss_dur: 0.13692 (0.11550) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.48180 (10.35436) | > current_lr: 0.00029 | > step_time: 0.61190 (0.48807) | > loader_time: 0.00350 (0.00312)  --> STEP: 290/639 -- GLOBAL_STEP: 47650 | > loss: -0.47563 (-0.50624) | > log_mle: -0.58671 (-0.62188) | > loss_dur: 0.11108 (0.11564) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.09339 (10.25992) | > current_lr: 0.00029 | > step_time: 0.63040 (0.50404) | > loader_time: 0.00350 (0.00315)  --> STEP: 315/639 -- GLOBAL_STEP: 47675 | > loss: -0.45452 (-0.50447) | > log_mle: -0.56562 (-0.62034) | > loss_dur: 0.11110 (0.11587) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.56868 (10.19174) | > current_lr: 0.00029 | > step_time: 0.62310 (0.51808) | > loader_time: 0.00350 (0.00318)  --> STEP: 340/639 -- GLOBAL_STEP: 47700 | > loss: -0.47466 (-0.50261) | > log_mle: -0.58703 (-0.61894) | > loss_dur: 0.11237 (0.11633) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.63536 (10.17253) | > current_lr: 0.00029 | > step_time: 0.62800 (0.53400) | > loader_time: 0.00320 (0.00321)  --> STEP: 365/639 -- GLOBAL_STEP: 47725 | > loss: -0.48078 (-0.50051) | > log_mle: -0.59326 (-0.61659) | > loss_dur: 0.11248 (0.11608) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.35539 (9.95203) | > current_lr: 0.00029 | > step_time: 0.81060 (0.54529) | > loader_time: 0.00400 (0.00324)  --> STEP: 390/639 -- GLOBAL_STEP: 47750 | > loss: -0.47023 (-0.49868) | > log_mle: -0.59272 (-0.61467) | > loss_dur: 0.12250 (0.11599) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.83208 (9.77451) | > current_lr: 0.00029 | > step_time: 0.76180 (0.55799) | > loader_time: 0.00360 (0.00326)  --> STEP: 415/639 -- GLOBAL_STEP: 47775 | > loss: -0.44371 (-0.49675) | > log_mle: -0.56588 (-0.61305) | > loss_dur: 0.12217 (0.11629) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.05382 (9.68010) | > current_lr: 0.00029 | > step_time: 0.72040 (0.57065) | > loader_time: 0.00340 (0.00329)  --> STEP: 440/639 -- GLOBAL_STEP: 47800 | > loss: -0.43224 (-0.49515) | > log_mle: -0.55007 (-0.61146) | > loss_dur: 0.11783 (0.11631) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.15092 (9.60213) | > current_lr: 0.00029 | > step_time: 0.79320 (0.58328) | > loader_time: 0.00370 (0.00331)  --> STEP: 465/639 -- GLOBAL_STEP: 47825 | > loss: -0.44528 (-0.49329) | > log_mle: -0.55657 (-0.60950) | > loss_dur: 0.11128 (0.11621) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.09325 (9.45804) | > current_lr: 0.00029 | > step_time: 0.78400 (0.59387) | > loader_time: 0.00360 (0.00333)  --> STEP: 490/639 -- GLOBAL_STEP: 47850 | > loss: -0.43613 (-0.49170) | > log_mle: -0.53773 (-0.60799) | > loss_dur: 0.10160 (0.11630) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.10826 (9.39725) | > current_lr: 0.00029 | > step_time: 0.86620 (0.60755) | > loader_time: 0.00370 (0.00335)  --> STEP: 515/639 -- GLOBAL_STEP: 47875 | > loss: -0.45729 (-0.49061) | > log_mle: -0.56748 (-0.60693) | > loss_dur: 0.11019 (0.11632) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.08223 (9.34236) | > current_lr: 0.00029 | > step_time: 0.97660 (0.62120) | > loader_time: 0.00410 (0.00338)  --> STEP: 540/639 -- GLOBAL_STEP: 47900 | > loss: -0.46665 (-0.48913) | > log_mle: -0.58528 (-0.60556) | > loss_dur: 0.11864 (0.11644) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.13973 (9.24331) | > current_lr: 0.00029 | > step_time: 0.88130 (0.63610) | > loader_time: 0.00420 (0.00341)  --> STEP: 565/639 -- GLOBAL_STEP: 47925 | > loss: -0.43417 (-0.48757) | > log_mle: -0.54518 (-0.60412) | > loss_dur: 0.11101 (0.11655) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.80873 (9.16046) | > current_lr: 0.00029 | > step_time: 0.98240 (0.65028) | > loader_time: 0.00410 (0.00345)  --> STEP: 590/639 -- GLOBAL_STEP: 47950 | > loss: -0.45599 (-0.48660) | > log_mle: -0.57735 (-0.60325) | > loss_dur: 0.12135 (0.11665) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.89480 (9.09801) | > current_lr: 0.00029 | > step_time: 0.94530 (0.66621) | > loader_time: 0.00430 (0.00348)  --> STEP: 615/639 -- GLOBAL_STEP: 47975 | > loss: -0.45981 (-0.48539) | > log_mle: -0.57985 (-0.60216) | > loss_dur: 0.12004 (0.11677) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.76436 (9.03148) | > current_lr: 0.00029 | > step_time: 1.11770 (0.68179) | > loader_time: 0.00430 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03652 (+0.00818) | > avg_loss: -0.45277 (-0.00003) | > avg_log_mle: -0.58471 (+0.00056) | > avg_loss_dur: 0.13194 (-0.00058)  > EPOCH: 75/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 00:26:23)   --> STEP: 0/639 -- GLOBAL_STEP: 48000 | > loss: -0.54685 (-0.54685) | > log_mle: -0.70758 (-0.70758) | > loss_dur: 0.16073 (0.16073) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 20.91642 (20.91642) | > current_lr: 0.00029 | > step_time: 0.61790 (0.61790) | > loader_time: 0.42910 (0.42907)  --> STEP: 25/639 -- GLOBAL_STEP: 48025 | > loss: -0.52289 (-0.54112) | > log_mle: -0.65016 (-0.65931) | > loss_dur: 0.12727 (0.11819) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.27615 (12.51857) | > current_lr: 0.00029 | > step_time: 0.38520 (0.32161) | > loader_time: 0.00250 (0.00297)  --> STEP: 50/639 -- GLOBAL_STEP: 48050 | > loss: -0.48751 (-0.52603) | > log_mle: -0.58541 (-0.64259) | > loss_dur: 0.09789 (0.11656) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.50136 (10.12058) | > current_lr: 0.00029 | > step_time: 0.31710 (0.34446) | > loader_time: 0.00280 (0.00290)  --> STEP: 75/639 -- GLOBAL_STEP: 48075 | > loss: -0.51464 (-0.51672) | > log_mle: -0.61852 (-0.63289) | > loss_dur: 0.10387 (0.11617) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.51255 (9.46330) | > current_lr: 0.00029 | > step_time: 0.44930 (0.37196) | > loader_time: 0.00310 (0.00290)  --> STEP: 100/639 -- GLOBAL_STEP: 48100 | > loss: -0.51464 (-0.51527) | > log_mle: -0.62588 (-0.63002) | > loss_dur: 0.11123 (0.11474) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.85583 (9.33996) | > current_lr: 0.00029 | > step_time: 0.50910 (0.39033) | > loader_time: 0.00320 (0.00293)  --> STEP: 125/639 -- GLOBAL_STEP: 48125 | > loss: -0.50849 (-0.51306) | > log_mle: -0.62289 (-0.62838) | > loss_dur: 0.11440 (0.11532) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.15804 (9.32370) | > current_lr: 0.00029 | > step_time: 0.43350 (0.40354) | > loader_time: 0.00310 (0.00297)  --> STEP: 150/639 -- GLOBAL_STEP: 48150 | > loss: -0.46372 (-0.51200) | > log_mle: -0.58329 (-0.62733) | > loss_dur: 0.11957 (0.11533) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.15655 (9.45540) | > current_lr: 0.00029 | > step_time: 0.42530 (0.41867) | > loader_time: 0.00300 (0.00298)  --> STEP: 175/639 -- GLOBAL_STEP: 48175 | > loss: -0.49930 (-0.51222) | > log_mle: -0.60399 (-0.62755) | > loss_dur: 0.10469 (0.11532) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.97845 (9.72172) | > current_lr: 0.00029 | > step_time: 0.55690 (0.43459) | > loader_time: 0.00340 (0.00301)  --> STEP: 200/639 -- GLOBAL_STEP: 48200 | > loss: -0.47118 (-0.51023) | > log_mle: -0.58771 (-0.62504) | > loss_dur: 0.11652 (0.11481) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.81075 (9.64634) | > current_lr: 0.00029 | > step_time: 0.52440 (0.44881) | > loader_time: 0.00320 (0.00304)  --> STEP: 225/639 -- GLOBAL_STEP: 48225 | > loss: -0.46484 (-0.50808) | > log_mle: -0.56648 (-0.62282) | > loss_dur: 0.10164 (0.11474) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.46726 (9.56811) | > current_lr: 0.00029 | > step_time: 0.52080 (0.46432) | > loader_time: 0.00350 (0.00308)  --> STEP: 250/639 -- GLOBAL_STEP: 48250 | > loss: -0.47521 (-0.50859) | > log_mle: -0.57241 (-0.62358) | > loss_dur: 0.09720 (0.11499) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.03242 (9.74582) | > current_lr: 0.00029 | > step_time: 0.57830 (0.47950) | > loader_time: 0.00340 (0.00311)  --> STEP: 275/639 -- GLOBAL_STEP: 48275 | > loss: -0.50052 (-0.50859) | > log_mle: -0.60935 (-0.62401) | > loss_dur: 0.10883 (0.11542) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.86401 (9.93355) | > current_lr: 0.00029 | > step_time: 0.85480 (0.49703) | > loader_time: 0.00360 (0.00314)  --> STEP: 300/639 -- GLOBAL_STEP: 48300 | > loss: -0.49463 (-0.50643) | > log_mle: -0.60446 (-0.62209) | > loss_dur: 0.10984 (0.11566) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.99138 (9.83873) | > current_lr: 0.00029 | > step_time: 0.74110 (0.51032) | > loader_time: 0.00350 (0.00317)  --> STEP: 325/639 -- GLOBAL_STEP: 48325 | > loss: -0.48553 (-0.50432) | > log_mle: -0.60485 (-0.62050) | > loss_dur: 0.11932 (0.11618) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.96617 (9.74722) | > current_lr: 0.00029 | > step_time: 0.62440 (0.52627) | > loader_time: 0.00340 (0.00321)  --> STEP: 350/639 -- GLOBAL_STEP: 48350 | > loss: -0.49992 (-0.50238) | > log_mle: -0.62853 (-0.61860) | > loss_dur: 0.12861 (0.11621) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.53774 (9.72208) | > current_lr: 0.00029 | > step_time: 0.86280 (0.54086) | > loader_time: 0.00380 (0.00324)  --> STEP: 375/639 -- GLOBAL_STEP: 48375 | > loss: -0.45883 (-0.50027) | > log_mle: -0.56550 (-0.61628) | > loss_dur: 0.10667 (0.11601) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.79882 (9.50266) | > current_lr: 0.00029 | > step_time: 0.67480 (0.55372) | > loader_time: 0.00320 (0.00326)  --> STEP: 400/639 -- GLOBAL_STEP: 48400 | > loss: -0.48441 (-0.49876) | > log_mle: -0.59563 (-0.61490) | > loss_dur: 0.11122 (0.11614) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.94194 (9.46719) | > current_lr: 0.00029 | > step_time: 0.72680 (0.56636) | > loader_time: 0.00380 (0.00329)  --> STEP: 425/639 -- GLOBAL_STEP: 48425 | > loss: -0.52273 (-0.49663) | > log_mle: -0.65076 (-0.61289) | > loss_dur: 0.12803 (0.11626) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.72029 (9.40169) | > current_lr: 0.00029 | > step_time: 0.93410 (0.57829) | > loader_time: 0.00400 (0.00331)  --> STEP: 450/639 -- GLOBAL_STEP: 48450 | > loss: -0.44726 (-0.49494) | > log_mle: -0.55287 (-0.61112) | > loss_dur: 0.10561 (0.11617) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.81569 (9.26426) | > current_lr: 0.00029 | > step_time: 0.80090 (0.58982) | > loader_time: 0.00390 (0.00334)  --> STEP: 475/639 -- GLOBAL_STEP: 48475 | > loss: -0.46457 (-0.49314) | > log_mle: -0.56380 (-0.60943) | > loss_dur: 0.09923 (0.11629) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.91578 (9.17334) | > current_lr: 0.00029 | > step_time: 0.71310 (0.60126) | > loader_time: 0.00390 (0.00335)  --> STEP: 500/639 -- GLOBAL_STEP: 48500 | > loss: -0.50994 (-0.49185) | > log_mle: -0.62803 (-0.60825) | > loss_dur: 0.11809 (0.11640) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.54417 (9.11196) | > current_lr: 0.00029 | > step_time: 0.94350 (0.61404) | > loader_time: 0.00400 (0.00338)  --> STEP: 525/639 -- GLOBAL_STEP: 48525 | > loss: -0.47746 (-0.49058) | > log_mle: -0.60545 (-0.60706) | > loss_dur: 0.12798 (0.11648) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.76402 (9.05386) | > current_lr: 0.00029 | > step_time: 1.02120 (0.62707) | > loader_time: 0.00430 (0.00341)  --> STEP: 550/639 -- GLOBAL_STEP: 48550 | > loss: -0.42558 (-0.48906) | > log_mle: -0.54104 (-0.60563) | > loss_dur: 0.11546 (0.11657) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.74325 (8.96736) | > current_lr: 0.00029 | > step_time: 0.93500 (0.64118) | > loader_time: 0.00390 (0.00344)  --> STEP: 575/639 -- GLOBAL_STEP: 48575 | > loss: -0.44590 (-0.48759) | > log_mle: -0.55951 (-0.60427) | > loss_dur: 0.11361 (0.11668) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.73332 (8.91087) | > current_lr: 0.00029 | > step_time: 0.88970 (0.65507) | > loader_time: 0.00400 (0.00347)  --> STEP: 600/639 -- GLOBAL_STEP: 48600 | > loss: -0.43488 (-0.48644) | > log_mle: -0.55556 (-0.60327) | > loss_dur: 0.12068 (0.11683) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.77016 (8.85332) | > current_lr: 0.00029 | > step_time: 1.03050 (0.67105) | > loader_time: 0.00450 (0.00350)  --> STEP: 625/639 -- GLOBAL_STEP: 48625 | > loss: -0.46925 (-0.48539) | > log_mle: -0.58673 (-0.60228) | > loss_dur: 0.11748 (0.11689) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.46523 (8.86738) | > current_lr: 0.00029 | > step_time: 1.04990 (0.68759) | > loader_time: 0.00450 (0.00354) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02825 (-0.00828) | > avg_loss: -0.45230 (+0.00047) | > avg_log_mle: -0.58504 (-0.00034) | > avg_loss_dur: 0.13275 (+0.00081)  > EPOCH: 76/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 00:34:02)   --> STEP: 10/639 -- GLOBAL_STEP: 48650 | > loss: -0.53005 (-0.53758) | > log_mle: -0.66257 (-0.66140) | > loss_dur: 0.13252 (0.12381) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.06134 (11.07735) | > current_lr: 0.00029 | > step_time: 0.27890 (0.31339) | > loader_time: 0.00270 (0.00208)  --> STEP: 35/639 -- GLOBAL_STEP: 48675 | > loss: -0.52898 (-0.52861) | > log_mle: -0.62564 (-0.64589) | > loss_dur: 0.09666 (0.11728) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.61706 (9.99194) | > current_lr: 0.00029 | > step_time: 0.31490 (0.33086) | > loader_time: 0.00280 (0.00238)  --> STEP: 60/639 -- GLOBAL_STEP: 48700 | > loss: -0.48349 (-0.52147) | > log_mle: -0.58356 (-0.63813) | > loss_dur: 0.10007 (0.11667) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.04322 (9.48265) | > current_lr: 0.00029 | > step_time: 0.34150 (0.35816) | > loader_time: 0.00270 (0.00252)  --> STEP: 85/639 -- GLOBAL_STEP: 48725 | > loss: -0.47820 (-0.51455) | > log_mle: -0.58445 (-0.62964) | > loss_dur: 0.10625 (0.11509) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.17743 (8.96060) | > current_lr: 0.00029 | > step_time: 0.41370 (0.37092) | > loader_time: 0.00300 (0.00263)  --> STEP: 110/639 -- GLOBAL_STEP: 48750 | > loss: -0.48211 (-0.51565) | > log_mle: -0.60039 (-0.63129) | > loss_dur: 0.11828 (0.11565) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.42862 (9.55455) | > current_lr: 0.00029 | > step_time: 0.39660 (0.38966) | > loader_time: 0.00300 (0.00269)  --> STEP: 135/639 -- GLOBAL_STEP: 48775 | > loss: -0.45135 (-0.51307) | > log_mle: -0.57652 (-0.62868) | > loss_dur: 0.12517 (0.11561) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.15638 (9.55445) | > current_lr: 0.00029 | > step_time: 0.49090 (0.40437) | > loader_time: 0.00340 (0.00273)  --> STEP: 160/639 -- GLOBAL_STEP: 48800 | > loss: -0.51804 (-0.51148) | > log_mle: -0.62117 (-0.62660) | > loss_dur: 0.10313 (0.11511) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.28037 (9.62480) | > current_lr: 0.00029 | > step_time: 0.46300 (0.42027) | > loader_time: 0.00270 (0.00279)  --> STEP: 185/639 -- GLOBAL_STEP: 48825 | > loss: -0.49015 (-0.51323) | > log_mle: -0.60022 (-0.62862) | > loss_dur: 0.11007 (0.11540) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.73994 (10.12794) | > current_lr: 0.00029 | > step_time: 0.58440 (0.43760) | > loader_time: 0.00310 (0.00284)  --> STEP: 210/639 -- GLOBAL_STEP: 48850 | > loss: -0.42218 (-0.51016) | > log_mle: -0.52211 (-0.62552) | > loss_dur: 0.09993 (0.11536) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.01179 (9.95750) | > current_lr: 0.00029 | > step_time: 0.50470 (0.45287) | > loader_time: 0.00300 (0.00288)  --> STEP: 235/639 -- GLOBAL_STEP: 48875 | > loss: -0.50083 (-0.51044) | > log_mle: -0.61161 (-0.62569) | > loss_dur: 0.11078 (0.11525) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.71268 (10.01307) | > current_lr: 0.00029 | > step_time: 0.53980 (0.46942) | > loader_time: 0.00350 (0.00292)  --> STEP: 260/639 -- GLOBAL_STEP: 48900 | > loss: -0.49212 (-0.50979) | > log_mle: -0.61212 (-0.62524) | > loss_dur: 0.12000 (0.11545) | > amp_scaler: 4096.00000 (2189.78462) | > grad_norm: 9.82908 (10.05592) | > current_lr: 0.00029 | > step_time: 0.58400 (0.48524) | > loader_time: 0.00330 (0.00295)  --> STEP: 285/639 -- GLOBAL_STEP: 48925 | > loss: -0.45105 (-0.50845) | > log_mle: -0.56004 (-0.62413) | > loss_dur: 0.10899 (0.11568) | > amp_scaler: 4096.00000 (2356.99649) | > grad_norm: 3.85639 (10.04638) | > current_lr: 0.00029 | > step_time: 0.60470 (0.50052) | > loader_time: 0.00320 (0.00299)  --> STEP: 310/639 -- GLOBAL_STEP: 48950 | > loss: -0.48488 (-0.50665) | > log_mle: -0.60323 (-0.62252) | > loss_dur: 0.11835 (0.11587) | > amp_scaler: 4096.00000 (2497.23871) | > grad_norm: 9.36926 (9.97764) | > current_lr: 0.00029 | > step_time: 0.58470 (0.51492) | > loader_time: 0.00340 (0.00303)  --> STEP: 335/639 -- GLOBAL_STEP: 48975 | > loss: -0.50537 (-0.50409) | > log_mle: -0.62594 (-0.62046) | > loss_dur: 0.12057 (0.11637) | > amp_scaler: 4096.00000 (2616.54925) | > grad_norm: 11.26689 (9.86746) | > current_lr: 0.00029 | > step_time: 0.67050 (0.52953) | > loader_time: 0.00350 (0.00306)  --> STEP: 360/639 -- GLOBAL_STEP: 49000 | > loss: -0.45690 (-0.50239) | > log_mle: -0.56425 (-0.61852) | > loss_dur: 0.10735 (0.11613) | > amp_scaler: 4096.00000 (2719.28889) | > grad_norm: 6.03695 (9.75963) | > current_lr: 0.00029 | > step_time: 0.68420 (0.54216) | > loader_time: 0.00370 (0.00310)  --> STEP: 385/639 -- GLOBAL_STEP: 49025 | > loss: -0.48712 (-0.50076) | > log_mle: -0.59682 (-0.61677) | > loss_dur: 0.10970 (0.11601) | > amp_scaler: 4096.00000 (2808.68571) | > grad_norm: 10.21377 (9.62638) | > current_lr: 0.00029 | > step_time: 0.67540 (0.55537) | > loader_time: 0.00350 (0.00313)  --> STEP: 410/639 -- GLOBAL_STEP: 49050 | > loss: -0.43639 (-0.49887) | > log_mle: -0.57099 (-0.61512) | > loss_dur: 0.13460 (0.11625) | > amp_scaler: 4096.00000 (2887.18049) | > grad_norm: 8.84868 (9.58507) | > current_lr: 0.00029 | > step_time: 0.73660 (0.56638) | > loader_time: 0.00380 (0.00316)  --> STEP: 435/639 -- GLOBAL_STEP: 49075 | > loss: -0.46753 (-0.49694) | > log_mle: -0.57996 (-0.61317) | > loss_dur: 0.11243 (0.11623) | > amp_scaler: 4096.00000 (2956.65287) | > grad_norm: 6.93872 (9.46529) | > current_lr: 0.00029 | > step_time: 0.70190 (0.57829) | > loader_time: 0.00340 (0.00319)  --> STEP: 460/639 -- GLOBAL_STEP: 49100 | > loss: -0.42961 (-0.49502) | > log_mle: -0.54691 (-0.61120) | > loss_dur: 0.11730 (0.11617) | > amp_scaler: 4096.00000 (3018.57391) | > grad_norm: 3.93892 (9.32344) | > current_lr: 0.00029 | > step_time: 0.88050 (0.58948) | > loader_time: 0.00400 (0.00321)  --> STEP: 485/639 -- GLOBAL_STEP: 49125 | > loss: -0.46963 (-0.49340) | > log_mle: -0.58176 (-0.60979) | > loss_dur: 0.11213 (0.11638) | > amp_scaler: 4096.00000 (3074.11134) | > grad_norm: 5.82453 (9.29222) | > current_lr: 0.00029 | > step_time: 0.80790 (0.60258) | > loader_time: 0.00390 (0.00323)  --> STEP: 510/639 -- GLOBAL_STEP: 49150 | > loss: -0.46171 (-0.49211) | > log_mle: -0.58297 (-0.60849) | > loss_dur: 0.12126 (0.11638) | > amp_scaler: 4096.00000 (3124.20392) | > grad_norm: 8.14340 (9.26621) | > current_lr: 0.00029 | > step_time: 0.85530 (0.61568) | > loader_time: 0.00400 (0.00327)  --> STEP: 535/639 -- GLOBAL_STEP: 49175 | > loss: -0.44628 (-0.49070) | > log_mle: -0.57402 (-0.60713) | > loss_dur: 0.12774 (0.11643) | > amp_scaler: 4096.00000 (3169.61495) | > grad_norm: 6.54494 (9.19666) | > current_lr: 0.00029 | > step_time: 0.89150 (0.62894) | > loader_time: 0.00390 (0.00330)  --> STEP: 560/639 -- GLOBAL_STEP: 49200 | > loss: -0.45773 (-0.48931) | > log_mle: -0.58040 (-0.60586) | > loss_dur: 0.12267 (0.11655) | > amp_scaler: 4096.00000 (3210.97143) | > grad_norm: 7.69770 (9.11359) | > current_lr: 0.00029 | > step_time: 0.92750 (0.64369) | > loader_time: 0.00390 (0.00333)  --> STEP: 585/639 -- GLOBAL_STEP: 49225 | > loss: -0.46859 (-0.48813) | > log_mle: -0.58573 (-0.60470) | > loss_dur: 0.11715 (0.11657) | > amp_scaler: 4096.00000 (3248.79316) | > grad_norm: 7.15261 (9.05375) | > current_lr: 0.00029 | > step_time: 1.08020 (0.65876) | > loader_time: 0.00370 (0.00337)  --> STEP: 610/639 -- GLOBAL_STEP: 49250 | > loss: -0.43753 (-0.48690) | > log_mle: -0.55346 (-0.60360) | > loss_dur: 0.11593 (0.11670) | > amp_scaler: 4096.00000 (3283.51475) | > grad_norm: 4.55773 (8.97777) | > current_lr: 0.00029 | > step_time: 1.03560 (0.67508) | > loader_time: 0.00420 (0.00342)  --> STEP: 635/639 -- GLOBAL_STEP: 49275 | > loss: -0.45809 (-0.48596) | > log_mle: -0.57902 (-0.60292) | > loss_dur: 0.12093 (0.11696) | > amp_scaler: 4096.00000 (3315.50236) | > grad_norm: 11.22787 (8.98346) | > current_lr: 0.00029 | > step_time: 1.15110 (0.69251) | > loader_time: 0.00640 (0.00347) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02938 (+0.00113) | > avg_loss: -0.45112 (+0.00118) | > avg_log_mle: -0.58613 (-0.00108) | > avg_loss_dur: 0.13501 (+0.00227)  > EPOCH: 77/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 00:41:41)   --> STEP: 20/639 -- GLOBAL_STEP: 49300 | > loss: -0.62787 (-0.55159) | > log_mle: -0.78190 (-0.67191) | > loss_dur: 0.15404 (0.12033) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 37.77918 (12.34706) | > current_lr: 0.00028 | > step_time: 0.33020 (0.30259) | > loader_time: 0.00290 (0.00245)  --> STEP: 45/639 -- GLOBAL_STEP: 49325 | > loss: -0.49575 (-0.53120) | > log_mle: -0.59292 (-0.64821) | > loss_dur: 0.09717 (0.11701) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.25090 (10.56618) | > current_lr: 0.00028 | > step_time: 0.40770 (0.33485) | > loader_time: 0.00270 (0.00261)  --> STEP: 70/639 -- GLOBAL_STEP: 49350 | > loss: -0.52359 (-0.52171) | > log_mle: -0.63885 (-0.63767) | > loss_dur: 0.11526 (0.11597) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.15729 (9.93724) | > current_lr: 0.00028 | > step_time: 0.39050 (0.35823) | > loader_time: 0.00300 (0.00271)  --> STEP: 95/639 -- GLOBAL_STEP: 49375 | > loss: -0.58435 (-0.51874) | > log_mle: -0.69186 (-0.63236) | > loss_dur: 0.10751 (0.11362) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.70921 (9.73005) | > current_lr: 0.00028 | > step_time: 0.41390 (0.37369) | > loader_time: 0.00290 (0.00278)  --> STEP: 120/639 -- GLOBAL_STEP: 49400 | > loss: -0.52901 (-0.51594) | > log_mle: -0.68193 (-0.63060) | > loss_dur: 0.15292 (0.11466) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.67803 (9.59088) | > current_lr: 0.00028 | > step_time: 0.52050 (0.39190) | > loader_time: 0.00320 (0.00284)  --> STEP: 145/639 -- GLOBAL_STEP: 49425 | > loss: -0.50912 (-0.51524) | > log_mle: -0.63483 (-0.63041) | > loss_dur: 0.12570 (0.11517) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.88465 (9.78708) | > current_lr: 0.00028 | > step_time: 0.46250 (0.40826) | > loader_time: 0.00290 (0.00288)  --> STEP: 170/639 -- GLOBAL_STEP: 49450 | > loss: -0.51705 (-0.51401) | > log_mle: -0.62261 (-0.62898) | > loss_dur: 0.10556 (0.11496) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.51646 (9.80083) | > current_lr: 0.00028 | > step_time: 0.61860 (0.42471) | > loader_time: 0.00330 (0.00291)  --> STEP: 195/639 -- GLOBAL_STEP: 49475 | > loss: -0.46111 (-0.51290) | > log_mle: -0.56469 (-0.62748) | > loss_dur: 0.10359 (0.11458) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.54617 (9.71680) | > current_lr: 0.00028 | > step_time: 0.49480 (0.44171) | > loader_time: 0.00340 (0.00296)  --> STEP: 220/639 -- GLOBAL_STEP: 49500 | > loss: -0.51253 (-0.51025) | > log_mle: -0.61422 (-0.62486) | > loss_dur: 0.10170 (0.11460) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.16373 (9.56175) | > current_lr: 0.00028 | > step_time: 0.64620 (0.45786) | > loader_time: 0.00320 (0.00300)  --> STEP: 245/639 -- GLOBAL_STEP: 49525 | > loss: -0.48941 (-0.51069) | > log_mle: -0.60182 (-0.62530) | > loss_dur: 0.11241 (0.11461) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.46860 (9.59325) | > current_lr: 0.00028 | > step_time: 0.52960 (0.47139) | > loader_time: 0.00330 (0.00304)  --> STEP: 270/639 -- GLOBAL_STEP: 49550 | > loss: -0.55368 (-0.51015) | > log_mle: -0.67677 (-0.62515) | > loss_dur: 0.12310 (0.11500) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.02297 (9.72550) | > current_lr: 0.00028 | > step_time: 0.82490 (0.48801) | > loader_time: 0.00390 (0.00308)  --> STEP: 295/639 -- GLOBAL_STEP: 49575 | > loss: -0.46546 (-0.50796) | > log_mle: -0.56381 (-0.62314) | > loss_dur: 0.09835 (0.11519) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.67427 (9.56696) | > current_lr: 0.00028 | > step_time: 0.63480 (0.50221) | > loader_time: 0.00350 (0.00312)  --> STEP: 320/639 -- GLOBAL_STEP: 49600 | > loss: -0.47497 (-0.50618) | > log_mle: -0.60544 (-0.62182) | > loss_dur: 0.13047 (0.11564) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.88317 (9.55923) | > current_lr: 0.00028 | > step_time: 0.75150 (0.51630) | > loader_time: 0.00330 (0.00315)  --> STEP: 345/639 -- GLOBAL_STEP: 49625 | > loss: -0.49111 (-0.50415) | > log_mle: -0.59439 (-0.62004) | > loss_dur: 0.10328 (0.11589) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.01444 (9.49554) | > current_lr: 0.00028 | > step_time: 0.64210 (0.52990) | > loader_time: 0.00330 (0.00318)  --> STEP: 370/639 -- GLOBAL_STEP: 49650 | > loss: -0.49689 (-0.50223) | > log_mle: -0.60936 (-0.61791) | > loss_dur: 0.11247 (0.11569) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.75400 (9.35209) | > current_lr: 0.00028 | > step_time: 0.88510 (0.54267) | > loader_time: 0.00410 (0.00321)  --> STEP: 395/639 -- GLOBAL_STEP: 49675 | > loss: -0.45960 (-0.50034) | > log_mle: -0.57687 (-0.61600) | > loss_dur: 0.11726 (0.11566) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.53888 (9.23121) | > current_lr: 0.00028 | > step_time: 0.70770 (0.55574) | > loader_time: 0.00360 (0.00323)  --> STEP: 420/639 -- GLOBAL_STEP: 49700 | > loss: -0.46229 (-0.49826) | > log_mle: -0.57528 (-0.61420) | > loss_dur: 0.11298 (0.11594) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.18976 (9.13910) | > current_lr: 0.00028 | > step_time: 0.68920 (0.56860) | > loader_time: 0.00350 (0.00326)  --> STEP: 445/639 -- GLOBAL_STEP: 49725 | > loss: -0.46763 (-0.49679) | > log_mle: -0.58037 (-0.61272) | > loss_dur: 0.11274 (0.11593) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.07380 (9.09106) | > current_lr: 0.00028 | > step_time: 0.72300 (0.58069) | > loader_time: 0.00380 (0.00328)  --> STEP: 470/639 -- GLOBAL_STEP: 49750 | > loss: -0.47763 (-0.49491) | > log_mle: -0.58993 (-0.61085) | > loss_dur: 0.11230 (0.11594) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.65411 (8.97051) | > current_lr: 0.00028 | > step_time: 0.92240 (0.59296) | > loader_time: 0.00390 (0.00331)  --> STEP: 495/639 -- GLOBAL_STEP: 49775 | > loss: -0.42227 (-0.49328) | > log_mle: -0.52794 (-0.60939) | > loss_dur: 0.10567 (0.11610) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.83001 (8.87770) | > current_lr: 0.00028 | > step_time: 0.74580 (0.60555) | > loader_time: 0.00360 (0.00333)  --> STEP: 520/639 -- GLOBAL_STEP: 49800 | > loss: -0.47711 (-0.49217) | > log_mle: -0.58449 (-0.60824) | > loss_dur: 0.10738 (0.11607) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.05360 (8.86410) | > current_lr: 0.00028 | > step_time: 0.83700 (0.61825) | > loader_time: 0.00330 (0.00336)  --> STEP: 545/639 -- GLOBAL_STEP: 49825 | > loss: -0.48195 (-0.49081) | > log_mle: -0.61437 (-0.60705) | > loss_dur: 0.13243 (0.11624) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.39063 (8.78018) | > current_lr: 0.00028 | > step_time: 0.94340 (0.63358) | > loader_time: 0.00390 (0.00339)  --> STEP: 570/639 -- GLOBAL_STEP: 49850 | > loss: -0.45636 (-0.48929) | > log_mle: -0.56461 (-0.60562) | > loss_dur: 0.10825 (0.11634) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.90607 (8.68477) | > current_lr: 0.00028 | > step_time: 0.83080 (0.64764) | > loader_time: 0.00400 (0.00342)  --> STEP: 595/639 -- GLOBAL_STEP: 49875 | > loss: -0.44596 (-0.48822) | > log_mle: -0.55799 (-0.60471) | > loss_dur: 0.11203 (0.11649) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.14818 (8.63718) | > current_lr: 0.00028 | > step_time: 1.08820 (0.66402) | > loader_time: 0.00410 (0.00346)  --> STEP: 620/639 -- GLOBAL_STEP: 49900 | > loss: -0.45044 (-0.48698) | > log_mle: -0.56610 (-0.60361) | > loss_dur: 0.11566 (0.11663) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.18040 (8.59059) | > current_lr: 0.00028 | > step_time: 1.11030 (0.68077) | > loader_time: 0.00480 (0.00349) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02663 (-0.00275) | > avg_loss: -0.45171 (-0.00059) | > avg_log_mle: -0.58458 (+0.00154) | > avg_loss_dur: 0.13288 (-0.00213)  > EPOCH: 78/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 00:49:18)   --> STEP: 5/639 -- GLOBAL_STEP: 49925 | > loss: -0.55340 (-0.55448) | > log_mle: -0.69453 (-0.67687) | > loss_dur: 0.14113 (0.12239) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.07747 (11.31644) | > current_lr: 0.00028 | > step_time: 0.32800 (0.28192) | > loader_time: 0.00300 (0.00273)  --> STEP: 30/639 -- GLOBAL_STEP: 49950 | > loss: -0.54297 (-0.53741) | > log_mle: -0.64729 (-0.65598) | > loss_dur: 0.10432 (0.11857) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.87120 (9.95817) | > current_lr: 0.00028 | > step_time: 0.32240 (0.32449) | > loader_time: 0.00240 (0.00273)  --> STEP: 55/639 -- GLOBAL_STEP: 49975 | > loss: -0.48968 (-0.52746) | > log_mle: -0.58725 (-0.64413) | > loss_dur: 0.09757 (0.11668) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.12916 (9.81598) | > current_lr: 0.00028 | > step_time: 0.33690 (0.34424) | > loader_time: 0.00250 (0.00268)  --> STEP: 80/639 -- GLOBAL_STEP: 50000 | > loss: -0.49320 (-0.51996) | > log_mle: -0.62586 (-0.63544) | > loss_dur: 0.13266 (0.11548) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.66522 (9.50276) | > current_lr: 0.00028 | > step_time: 0.39990 (0.36778) | > loader_time: 0.00280 (0.00274) > CHECKPOINT : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/checkpoint_50000.pth.tar  --> STEP: 105/639 -- GLOBAL_STEP: 50025 | > loss: -0.46241 (-0.51786) | > log_mle: -0.57720 (-0.63277) | > loss_dur: 0.11479 (0.11491) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.33008 (9.46440) | > current_lr: 0.00028 | > step_time: 0.41440 (0.38628) | > loader_time: 0.00300 (0.00289)  --> STEP: 130/639 -- GLOBAL_STEP: 50050 | > loss: -0.49538 (-0.51569) | > log_mle: -0.61725 (-0.63114) | > loss_dur: 0.12186 (0.11545) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.30736 (9.47498) | > current_lr: 0.00028 | > step_time: 0.48240 (0.40203) | > loader_time: 0.00300 (0.00291)  --> STEP: 155/639 -- GLOBAL_STEP: 50075 | > loss: -0.49044 (-0.51470) | > log_mle: -0.60461 (-0.63002) | > loss_dur: 0.11417 (0.11532) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.34877 (9.49716) | > current_lr: 0.00028 | > step_time: 0.54470 (0.42024) | > loader_time: 0.00330 (0.00295)  --> STEP: 180/639 -- GLOBAL_STEP: 50100 | > loss: -0.53577 (-0.51551) | > log_mle: -0.65930 (-0.63125) | > loss_dur: 0.12353 (0.11574) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.29349 (9.84074) | > current_lr: 0.00028 | > step_time: 0.56860 (0.43539) | > loader_time: 0.00350 (0.00298)  --> STEP: 205/639 -- GLOBAL_STEP: 50125 | > loss: -0.46829 (-0.51224) | > log_mle: -0.58054 (-0.62736) | > loss_dur: 0.11225 (0.11512) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.70093 (9.46758) | > current_lr: 0.00028 | > step_time: 0.54860 (0.44866) | > loader_time: 0.00330 (0.00302)  --> STEP: 230/639 -- GLOBAL_STEP: 50150 | > loss: -0.49966 (-0.51205) | > log_mle: -0.61767 (-0.62720) | > loss_dur: 0.11800 (0.11515) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.38750 (9.58716) | > current_lr: 0.00028 | > step_time: 0.56880 (0.46519) | > loader_time: 0.00330 (0.00305)  --> STEP: 255/639 -- GLOBAL_STEP: 50175 | > loss: -0.47187 (-0.51097) | > log_mle: -0.58675 (-0.62598) | > loss_dur: 0.11488 (0.11501) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.00970 (9.54938) | > current_lr: 0.00028 | > step_time: 0.67080 (0.47933) | > loader_time: 0.00330 (0.00307)  --> STEP: 280/639 -- GLOBAL_STEP: 50200 | > loss: -0.48433 (-0.51070) | > log_mle: -0.61220 (-0.62619) | > loss_dur: 0.12787 (0.11549) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.44343 (9.67167) | > current_lr: 0.00028 | > step_time: 0.59560 (0.49554) | > loader_time: 0.00340 (0.00311)  --> STEP: 305/639 -- GLOBAL_STEP: 50225 | > loss: -0.46675 (-0.50847) | > log_mle: -0.59237 (-0.62420) | > loss_dur: 0.12562 (0.11573) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.10547 (9.77533) | > current_lr: 0.00028 | > step_time: 0.65040 (0.51012) | > loader_time: 0.00360 (0.00314)  --> STEP: 330/639 -- GLOBAL_STEP: 50250 | > loss: -0.48634 (-0.50596) | > log_mle: -0.61322 (-0.62212) | > loss_dur: 0.12689 (0.11615) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.25207 (9.65495) | > current_lr: 0.00028 | > step_time: 0.74100 (0.52544) | > loader_time: 0.00340 (0.00317)  --> STEP: 355/639 -- GLOBAL_STEP: 50275 | > loss: -0.43802 (-0.50419) | > log_mle: -0.54414 (-0.62016) | > loss_dur: 0.10613 (0.11596) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.54903 (9.54082) | > current_lr: 0.00028 | > step_time: 0.58350 (0.53781) | > loader_time: 0.00320 (0.00319)  --> STEP: 380/639 -- GLOBAL_STEP: 50300 | > loss: -0.47899 (-0.50215) | > log_mle: -0.57837 (-0.61795) | > loss_dur: 0.09938 (0.11580) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.21483 (9.36694) | > current_lr: 0.00028 | > step_time: 0.62870 (0.55060) | > loader_time: 0.00350 (0.00322)  --> STEP: 405/639 -- GLOBAL_STEP: 50325 | > loss: -0.46974 (-0.50068) | > log_mle: -0.58296 (-0.61670) | > loss_dur: 0.11322 (0.11603) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.69781 (9.35277) | > current_lr: 0.00028 | > step_time: 0.67700 (0.56236) | > loader_time: 0.00340 (0.00325)  --> STEP: 430/639 -- GLOBAL_STEP: 50350 | > loss: -0.43886 (-0.49857) | > log_mle: -0.54386 (-0.61477) | > loss_dur: 0.10500 (0.11620) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.69583 (9.27094) | > current_lr: 0.00028 | > step_time: 0.79830 (0.57540) | > loader_time: 0.00370 (0.00327)  --> STEP: 455/639 -- GLOBAL_STEP: 50375 | > loss: -0.45359 (-0.49674) | > log_mle: -0.56831 (-0.61279) | > loss_dur: 0.11472 (0.11605) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.77885 (9.16385) | > current_lr: 0.00028 | > step_time: 0.76780 (0.58639) | > loader_time: 0.00360 (0.00329)  --> STEP: 480/639 -- GLOBAL_STEP: 50400 | > loss: -0.41355 (-0.49496) | > log_mle: -0.52563 (-0.61119) | > loss_dur: 0.11208 (0.11623) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.86410 (9.08911) | > current_lr: 0.00028 | > step_time: 0.78700 (0.59893) | > loader_time: 0.00350 (0.00332)  --> STEP: 505/639 -- GLOBAL_STEP: 50425 | > loss: -0.48029 (-0.49357) | > log_mle: -0.59345 (-0.60976) | > loss_dur: 0.11316 (0.11619) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.98098 (8.98932) | > current_lr: 0.00028 | > step_time: 0.87500 (0.61110) | > loader_time: 0.00410 (0.00335)  --> STEP: 530/639 -- GLOBAL_STEP: 50450 | > loss: -0.44933 (-0.49225) | > log_mle: -0.57795 (-0.60856) | > loss_dur: 0.12862 (0.11630) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.86507 (8.92536) | > current_lr: 0.00028 | > step_time: 0.96980 (0.62564) | > loader_time: 0.00400 (0.00338)  --> STEP: 555/639 -- GLOBAL_STEP: 50475 | > loss: -0.45140 (-0.49078) | > log_mle: -0.56484 (-0.60707) | > loss_dur: 0.11345 (0.11629) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.17682 (8.83737) | > current_lr: 0.00028 | > step_time: 0.84150 (0.63990) | > loader_time: 0.00400 (0.00342)  --> STEP: 580/639 -- GLOBAL_STEP: 50500 | > loss: -0.47835 (-0.48960) | > log_mle: -0.58430 (-0.60596) | > loss_dur: 0.10595 (0.11636) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.24916 (8.78016) | > current_lr: 0.00028 | > step_time: 1.02620 (0.65544) | > loader_time: 0.00450 (0.00346)  --> STEP: 605/639 -- GLOBAL_STEP: 50525 | > loss: -0.44998 (-0.48841) | > log_mle: -0.55286 (-0.60494) | > loss_dur: 0.10289 (0.11652) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.88785 (8.73241) | > current_lr: 0.00028 | > step_time: 1.04010 (0.67099) | > loader_time: 0.00470 (0.00349)  --> STEP: 630/639 -- GLOBAL_STEP: 50550 | > loss: -0.45180 (-0.48737) | > log_mle: -0.58307 (-0.60406) | > loss_dur: 0.13127 (0.11669) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.37667 (8.70745) | > current_lr: 0.00028 | > step_time: 1.19940 (0.68802) | > loader_time: 0.00460 (0.00354) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02971 (+0.00308) | > avg_loss: -0.45645 (-0.00474) | > avg_log_mle: -0.58976 (-0.00517) | > avg_loss_dur: 0.13331 (+0.00043)  > EPOCH: 79/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 00:57:00)   --> STEP: 15/639 -- GLOBAL_STEP: 50575 | > loss: -0.65266 (-0.55172) | > log_mle: -0.78556 (-0.66889) | > loss_dur: 0.13290 (0.11717) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 22.60919 (11.81600) | > current_lr: 0.00028 | > step_time: 0.33290 (0.29484) | > loader_time: 0.00270 (0.00233)  --> STEP: 40/639 -- GLOBAL_STEP: 50600 | > loss: -0.50589 (-0.53277) | > log_mle: -0.62029 (-0.64805) | > loss_dur: 0.11440 (0.11528) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.58410 (10.36755) | > current_lr: 0.00028 | > step_time: 0.36230 (0.33067) | > loader_time: 0.00250 (0.00249)  --> STEP: 65/639 -- GLOBAL_STEP: 50625 | > loss: -0.51558 (-0.52476) | > log_mle: -0.61263 (-0.64016) | > loss_dur: 0.09706 (0.11540) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.09889 (10.26916) | > current_lr: 0.00028 | > step_time: 0.38520 (0.36263) | > loader_time: 0.00290 (0.00264)  --> STEP: 90/639 -- GLOBAL_STEP: 50650 | > loss: -0.46043 (-0.51745) | > log_mle: -0.56752 (-0.63094) | > loss_dur: 0.10710 (0.11349) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.10527 (9.48936) | > current_lr: 0.00028 | > step_time: 0.51580 (0.37691) | > loader_time: 0.00320 (0.00272)  --> STEP: 115/639 -- GLOBAL_STEP: 50675 | > loss: -0.41669 (-0.51677) | > log_mle: -0.53839 (-0.63099) | > loss_dur: 0.12170 (0.11422) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.76888 (9.93530) | > current_lr: 0.00028 | > step_time: 0.39580 (0.39398) | > loader_time: 0.00300 (0.00277)  --> STEP: 140/639 -- GLOBAL_STEP: 50700 | > loss: -0.48324 (-0.51607) | > log_mle: -0.61698 (-0.63082) | > loss_dur: 0.13374 (0.11475) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.12838 (10.30809) | > current_lr: 0.00028 | > step_time: 0.55730 (0.41175) | > loader_time: 0.00320 (0.00281)  --> STEP: 165/639 -- GLOBAL_STEP: 50725 | > loss: -0.48730 (-0.51490) | > log_mle: -0.62587 (-0.62956) | > loss_dur: 0.13857 (0.11466) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.66861 (10.25454) | > current_lr: 0.00028 | > step_time: 0.48840 (0.42449) | > loader_time: 0.00310 (0.00287)  --> STEP: 190/639 -- GLOBAL_STEP: 50750 | > loss: -0.44521 (-0.51424) | > log_mle: -0.55253 (-0.62882) | > loss_dur: 0.10732 (0.11458) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.92834 (10.31246) | > current_lr: 0.00028 | > step_time: 0.53840 (0.44190) | > loader_time: 0.00330 (0.00293)  --> STEP: 215/639 -- GLOBAL_STEP: 50775 | > loss: -0.54591 (-0.51174) | > log_mle: -0.65542 (-0.62650) | > loss_dur: 0.10951 (0.11476) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.01669 (10.21037) | > current_lr: 0.00028 | > step_time: 0.55240 (0.45735) | > loader_time: 0.00340 (0.00297)  --> STEP: 240/639 -- GLOBAL_STEP: 50800 | > loss: -0.54740 (-0.51245) | > log_mle: -0.66887 (-0.62702) | > loss_dur: 0.12147 (0.11457) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.08489 (10.26872) | > current_lr: 0.00028 | > step_time: 0.61970 (0.47200) | > loader_time: 0.00330 (0.00302)  --> STEP: 265/639 -- GLOBAL_STEP: 50825 | > loss: -0.55311 (-0.51113) | > log_mle: -0.68770 (-0.62607) | > loss_dur: 0.13459 (0.11494) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.26131 (10.40344) | > current_lr: 0.00028 | > step_time: 0.63380 (0.48799) | > loader_time: 0.00340 (0.00306)  --> STEP: 290/639 -- GLOBAL_STEP: 50850 | > loss: -0.47910 (-0.50946) | > log_mle: -0.58953 (-0.62458) | > loss_dur: 0.11044 (0.11512) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.38157 (10.32403) | > current_lr: 0.00028 | > step_time: 0.62220 (0.50312) | > loader_time: 0.00350 (0.00310)  --> STEP: 315/639 -- GLOBAL_STEP: 50875 | > loss: -0.45857 (-0.50757) | > log_mle: -0.56643 (-0.62298) | > loss_dur: 0.10787 (0.11541) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.30609 (10.20804) | > current_lr: 0.00028 | > step_time: 0.69480 (0.51877) | > loader_time: 0.00350 (0.00314)  --> STEP: 340/639 -- GLOBAL_STEP: 50900 | > loss: -0.48114 (-0.50563) | > log_mle: -0.59066 (-0.62152) | > loss_dur: 0.10951 (0.11590) | > amp_scaler: 8192.00000 (4312.84706) | > grad_norm: 8.35983 (10.16439) | > current_lr: 0.00028 | > step_time: 0.60030 (0.53460) | > loader_time: 0.00340 (0.00317)  --> STEP: 365/639 -- GLOBAL_STEP: 50925 | > loss: -0.47815 (-0.50358) | > log_mle: -0.59359 (-0.61915) | > loss_dur: 0.11545 (0.11557) | > amp_scaler: 8192.00000 (4578.54247) | > grad_norm: 9.17894 (10.01285) | > current_lr: 0.00028 | > step_time: 0.80640 (0.54631) | > loader_time: 0.00390 (0.00320)  --> STEP: 390/639 -- GLOBAL_STEP: 50950 | > loss: -0.46576 (-0.50175) | > log_mle: -0.59326 (-0.61719) | > loss_dur: 0.12751 (0.11545) | > amp_scaler: 8192.00000 (4810.17436) | > grad_norm: 9.08633 (9.83341) | > current_lr: 0.00028 | > step_time: 0.77840 (0.55900) | > loader_time: 0.00360 (0.00322)  --> STEP: 415/639 -- GLOBAL_STEP: 50975 | > loss: -0.44902 (-0.49978) | > log_mle: -0.56909 (-0.61550) | > loss_dur: 0.12007 (0.11571) | > amp_scaler: 8192.00000 (5013.89880) | > grad_norm: 5.67774 (9.70864) | > current_lr: 0.00028 | > step_time: 0.76020 (0.57223) | > loader_time: 0.00350 (0.00325)  --> STEP: 440/639 -- GLOBAL_STEP: 51000 | > loss: -0.43528 (-0.49813) | > log_mle: -0.55217 (-0.61392) | > loss_dur: 0.11688 (0.11579) | > amp_scaler: 8192.00000 (5194.47273) | > grad_norm: 5.91699 (9.62361) | > current_lr: 0.00028 | > step_time: 0.72970 (0.58491) | > loader_time: 0.00370 (0.00328)  --> STEP: 465/639 -- GLOBAL_STEP: 51025 | > loss: -0.44011 (-0.49614) | > log_mle: -0.55630 (-0.61190) | > loss_dur: 0.11619 (0.11577) | > amp_scaler: 8192.00000 (5355.63011) | > grad_norm: 5.00258 (9.46775) | > current_lr: 0.00028 | > step_time: 0.80170 (0.59639) | > loader_time: 0.00380 (0.00331)  --> STEP: 490/639 -- GLOBAL_STEP: 51050 | > loss: -0.43871 (-0.49447) | > log_mle: -0.53962 (-0.61038) | > loss_dur: 0.10091 (0.11591) | > amp_scaler: 8192.00000 (5500.34286) | > grad_norm: 3.14875 (9.42005) | > current_lr: 0.00028 | > step_time: 0.78830 (0.60947) | > loader_time: 0.00380 (0.00334)  --> STEP: 515/639 -- GLOBAL_STEP: 51075 | > loss: -0.46130 (-0.49330) | > log_mle: -0.56880 (-0.60927) | > loss_dur: 0.10750 (0.11598) | > amp_scaler: 8192.00000 (5631.00583) | > grad_norm: 7.30859 (9.37983) | > current_lr: 0.00028 | > step_time: 0.99700 (0.62312) | > loader_time: 0.00400 (0.00337)  --> STEP: 540/639 -- GLOBAL_STEP: 51100 | > loss: -0.46117 (-0.49176) | > log_mle: -0.58444 (-0.60784) | > loss_dur: 0.12327 (0.11608) | > amp_scaler: 8192.00000 (5749.57037) | > grad_norm: 7.97414 (9.27969) | > current_lr: 0.00028 | > step_time: 0.89740 (0.63718) | > loader_time: 0.00420 (0.00340)  --> STEP: 565/639 -- GLOBAL_STEP: 51125 | > loss: -0.43009 (-0.49015) | > log_mle: -0.54599 (-0.60635) | > loss_dur: 0.11590 (0.11620) | > amp_scaler: 8192.00000 (5857.64248) | > grad_norm: 5.54735 (9.18062) | > current_lr: 0.00028 | > step_time: 1.00030 (0.65077) | > loader_time: 0.00430 (0.00343)  --> STEP: 590/639 -- GLOBAL_STEP: 51150 | > loss: -0.46430 (-0.48915) | > log_mle: -0.57919 (-0.60545) | > loss_dur: 0.11489 (0.11630) | > amp_scaler: 8192.00000 (5956.55593) | > grad_norm: 7.45626 (9.13426) | > current_lr: 0.00028 | > step_time: 0.95320 (0.66630) | > loader_time: 0.00440 (0.00348)  --> STEP: 615/639 -- GLOBAL_STEP: 51175 | > loss: -0.45444 (-0.48788) | > log_mle: -0.58097 (-0.60430) | > loss_dur: 0.12653 (0.11642) | > amp_scaler: 8192.00000 (6047.42764) | > grad_norm: 6.37600 (9.05877) | > current_lr: 0.00028 | > step_time: 1.00660 (0.68206) | > loader_time: 0.00440 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02365 (-0.00606) | > avg_loss: -0.45311 (+0.00334) | > avg_log_mle: -0.58570 (+0.00406) | > avg_loss_dur: 0.13259 (-0.00072)  > EPOCH: 80/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 01:04:41)   --> STEP: 0/639 -- GLOBAL_STEP: 51200 | > loss: -0.55921 (-0.55921) | > log_mle: -0.70909 (-0.70909) | > loss_dur: 0.14987 (0.14987) | > amp_scaler: 8192.00000 (8192.00000) | > grad_norm: 21.37179 (21.37179) | > current_lr: 0.00028 | > step_time: 0.80100 (0.80100) | > loader_time: 0.31750 (0.31752)  --> STEP: 25/639 -- GLOBAL_STEP: 51225 | > loss: -0.52548 (-0.54115) | > log_mle: -0.65962 (-0.66214) | > loss_dur: 0.13414 (0.12099) | > amp_scaler: 4096.00000 (6881.28000) | > grad_norm: 11.70426 (10.71906) | > current_lr: 0.00028 | > step_time: 0.39670 (0.33069) | > loader_time: 0.00270 (0.00261)  --> STEP: 50/639 -- GLOBAL_STEP: 51250 | > loss: -0.49236 (-0.52884) | > log_mle: -0.59172 (-0.64529) | > loss_dur: 0.09935 (0.11645) | > amp_scaler: 4096.00000 (5488.64000) | > grad_norm: 6.23512 (10.16537) | > current_lr: 0.00028 | > step_time: 0.31100 (0.34909) | > loader_time: 0.00280 (0.00274)  --> STEP: 75/639 -- GLOBAL_STEP: 51275 | > loss: -0.51110 (-0.52043) | > log_mle: -0.62076 (-0.63596) | > loss_dur: 0.10966 (0.11553) | > amp_scaler: 4096.00000 (5024.42667) | > grad_norm: 7.29165 (9.54737) | > current_lr: 0.00028 | > step_time: 0.41340 (0.36811) | > loader_time: 0.00310 (0.00281)  --> STEP: 100/639 -- GLOBAL_STEP: 51300 | > loss: -0.51876 (-0.51813) | > log_mle: -0.63112 (-0.63256) | > loss_dur: 0.11235 (0.11443) | > amp_scaler: 4096.00000 (4792.32000) | > grad_norm: 9.75864 (9.41289) | > current_lr: 0.00028 | > step_time: 0.51000 (0.38552) | > loader_time: 0.00310 (0.00286)  --> STEP: 125/639 -- GLOBAL_STEP: 51325 | > loss: -0.50647 (-0.51558) | > log_mle: -0.62392 (-0.63055) | > loss_dur: 0.11745 (0.11497) | > amp_scaler: 4096.00000 (4653.05600) | > grad_norm: 8.57963 (9.61061) | > current_lr: 0.00028 | > step_time: 0.41320 (0.40064) | > loader_time: 0.00280 (0.00291)  --> STEP: 150/639 -- GLOBAL_STEP: 51350 | > loss: -0.47256 (-0.51393) | > log_mle: -0.59047 (-0.62949) | > loss_dur: 0.11791 (0.11556) | > amp_scaler: 4096.00000 (4560.21333) | > grad_norm: 5.90339 (9.70109) | > current_lr: 0.00028 | > step_time: 0.42400 (0.41433) | > loader_time: 0.00320 (0.00294)  --> STEP: 175/639 -- GLOBAL_STEP: 51375 | > loss: -0.50661 (-0.51415) | > log_mle: -0.60744 (-0.62960) | > loss_dur: 0.10083 (0.11545) | > amp_scaler: 4096.00000 (4493.89714) | > grad_norm: 6.27455 (9.88844) | > current_lr: 0.00028 | > step_time: 0.56440 (0.43090) | > loader_time: 0.00340 (0.00297)  --> STEP: 200/639 -- GLOBAL_STEP: 51400 | > loss: -0.47658 (-0.51231) | > log_mle: -0.59138 (-0.62725) | > loss_dur: 0.11480 (0.11495) | > amp_scaler: 4096.00000 (4444.16000) | > grad_norm: 6.30700 (9.62374) | > current_lr: 0.00028 | > step_time: 0.52550 (0.44695) | > loader_time: 0.00320 (0.00299)  --> STEP: 225/639 -- GLOBAL_STEP: 51425 | > loss: -0.47340 (-0.51014) | > log_mle: -0.57121 (-0.62500) | > loss_dur: 0.09782 (0.11486) | > amp_scaler: 4096.00000 (4405.47556) | > grad_norm: 3.18550 (9.61701) | > current_lr: 0.00028 | > step_time: 0.51640 (0.46364) | > loader_time: 0.00320 (0.00303)  --> STEP: 250/639 -- GLOBAL_STEP: 51450 | > loss: -0.47309 (-0.51076) | > log_mle: -0.57264 (-0.62573) | > loss_dur: 0.09955 (0.11497) | > amp_scaler: 4096.00000 (4374.52800) | > grad_norm: 3.11307 (9.83535) | > current_lr: 0.00028 | > step_time: 0.57820 (0.47833) | > loader_time: 0.00340 (0.00306)  --> STEP: 275/639 -- GLOBAL_STEP: 51475 | > loss: -0.49968 (-0.51056) | > log_mle: -0.61168 (-0.62600) | > loss_dur: 0.11200 (0.11544) | > amp_scaler: 4096.00000 (4349.20727) | > grad_norm: 8.56783 (10.13080) | > current_lr: 0.00028 | > step_time: 0.87530 (0.49528) | > loader_time: 0.00380 (0.00309)  --> STEP: 300/639 -- GLOBAL_STEP: 51500 | > loss: -0.48179 (-0.50819) | > log_mle: -0.59785 (-0.62382) | > loss_dur: 0.11606 (0.11563) | > amp_scaler: 4096.00000 (4328.10667) | > grad_norm: 9.23112 (9.98557) | > current_lr: 0.00028 | > step_time: 0.70410 (0.50904) | > loader_time: 0.00370 (0.00313)  --> STEP: 325/639 -- GLOBAL_STEP: 51525 | > loss: -0.48682 (-0.50609) | > log_mle: -0.60619 (-0.62217) | > loss_dur: 0.11937 (0.11607) | > amp_scaler: 4096.00000 (4310.25231) | > grad_norm: 7.35488 (9.87226) | > current_lr: 0.00028 | > step_time: 0.60970 (0.52492) | > loader_time: 0.00300 (0.00316)  --> STEP: 350/639 -- GLOBAL_STEP: 51550 | > loss: -0.50068 (-0.50410) | > log_mle: -0.62835 (-0.62018) | > loss_dur: 0.12767 (0.11608) | > amp_scaler: 4096.00000 (4294.94857) | > grad_norm: 12.29492 (9.80952) | > current_lr: 0.00028 | > step_time: 0.83420 (0.53857) | > loader_time: 0.00380 (0.00319)  --> STEP: 375/639 -- GLOBAL_STEP: 51575 | > loss: -0.45533 (-0.50198) | > log_mle: -0.56668 (-0.61786) | > loss_dur: 0.11135 (0.11588) | > amp_scaler: 4096.00000 (4281.68533) | > grad_norm: 5.29154 (9.60735) | > current_lr: 0.00028 | > step_time: 0.67250 (0.55087) | > loader_time: 0.00350 (0.00321)  --> STEP: 400/639 -- GLOBAL_STEP: 51600 | > loss: -0.48428 (-0.50045) | > log_mle: -0.59870 (-0.61641) | > loss_dur: 0.11442 (0.11596) | > amp_scaler: 4096.00000 (4270.08000) | > grad_norm: 7.61350 (9.50837) | > current_lr: 0.00028 | > step_time: 0.79880 (0.56372) | > loader_time: 0.00390 (0.00324)  --> STEP: 425/639 -- GLOBAL_STEP: 51625 | > loss: -0.52447 (-0.49839) | > log_mle: -0.64961 (-0.61441) | > loss_dur: 0.12513 (0.11603) | > amp_scaler: 4096.00000 (4259.84000) | > grad_norm: 13.54043 (9.43314) | > current_lr: 0.00028 | > step_time: 0.89640 (0.57576) | > loader_time: 0.00390 (0.00327)  --> STEP: 450/639 -- GLOBAL_STEP: 51650 | > loss: -0.44990 (-0.49669) | > log_mle: -0.55396 (-0.61267) | > loss_dur: 0.10406 (0.11598) | > amp_scaler: 4096.00000 (4250.73778) | > grad_norm: 7.57511 (9.34065) | > current_lr: 0.00028 | > step_time: 0.76070 (0.58742) | > loader_time: 0.00390 (0.00330)  --> STEP: 475/639 -- GLOBAL_STEP: 51675 | > loss: -0.46472 (-0.49491) | > log_mle: -0.56599 (-0.61101) | > loss_dur: 0.10127 (0.11610) | > amp_scaler: 4096.00000 (4242.59368) | > grad_norm: 4.18806 (9.23796) | > current_lr: 0.00028 | > step_time: 0.71000 (0.59878) | > loader_time: 0.00370 (0.00332)  --> STEP: 500/639 -- GLOBAL_STEP: 51700 | > loss: -0.50743 (-0.49361) | > log_mle: -0.62805 (-0.60980) | > loss_dur: 0.12061 (0.11619) | > amp_scaler: 4096.00000 (4235.26400) | > grad_norm: 15.45751 (9.15939) | > current_lr: 0.00028 | > step_time: 0.99210 (0.61159) | > loader_time: 0.00390 (0.00335)  --> STEP: 525/639 -- GLOBAL_STEP: 51725 | > loss: -0.47575 (-0.49233) | > log_mle: -0.60859 (-0.60862) | > loss_dur: 0.13284 (0.11629) | > amp_scaler: 4096.00000 (4228.63238) | > grad_norm: 10.35617 (9.11899) | > current_lr: 0.00028 | > step_time: 1.01030 (0.62550) | > loader_time: 0.00410 (0.00338)  --> STEP: 550/639 -- GLOBAL_STEP: 51750 | > loss: -0.42592 (-0.49087) | > log_mle: -0.54192 (-0.60717) | > loss_dur: 0.11600 (0.11630) | > amp_scaler: 4096.00000 (4222.60364) | > grad_norm: 3.60072 (9.05641) | > current_lr: 0.00028 | > step_time: 0.95450 (0.63985) | > loader_time: 0.00430 (0.00341)  --> STEP: 575/639 -- GLOBAL_STEP: 51775 | > loss: -0.44735 (-0.48942) | > log_mle: -0.55992 (-0.60582) | > loss_dur: 0.11256 (0.11640) | > amp_scaler: 4096.00000 (4217.09913) | > grad_norm: 6.47055 (8.96578) | > current_lr: 0.00028 | > step_time: 0.89730 (0.65385) | > loader_time: 0.00410 (0.00345)  --> STEP: 600/639 -- GLOBAL_STEP: 51800 | > loss: -0.43831 (-0.48821) | > log_mle: -0.55860 (-0.60480) | > loss_dur: 0.12029 (0.11659) | > amp_scaler: 4096.00000 (4212.05333) | > grad_norm: 6.03584 (8.91705) | > current_lr: 0.00028 | > step_time: 1.05470 (0.66998) | > loader_time: 0.00450 (0.00350)  --> STEP: 625/639 -- GLOBAL_STEP: 51825 | > loss: -0.47040 (-0.48714) | > log_mle: -0.58774 (-0.60380) | > loss_dur: 0.11734 (0.11666) | > amp_scaler: 4096.00000 (4207.41120) | > grad_norm: 7.39363 (8.86484) | > current_lr: 0.00028 | > step_time: 1.07670 (0.68661) | > loader_time: 0.00430 (0.00356) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03224 (+0.00859) | > avg_loss: -0.45502 (-0.00191) | > avg_log_mle: -0.58785 (-0.00215) | > avg_loss_dur: 0.13283 (+0.00024)  > EPOCH: 81/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 01:12:20)   --> STEP: 10/639 -- GLOBAL_STEP: 51850 | > loss: -0.53317 (-0.54315) | > log_mle: -0.66598 (-0.66621) | > loss_dur: 0.13281 (0.12306) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.51182 (12.26035) | > current_lr: 0.00028 | > step_time: 0.28330 (0.30846) | > loader_time: 0.00280 (0.00246)  --> STEP: 35/639 -- GLOBAL_STEP: 51875 | > loss: -0.53428 (-0.53394) | > log_mle: -0.63094 (-0.65137) | > loss_dur: 0.09666 (0.11743) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.17678 (10.40894) | > current_lr: 0.00028 | > step_time: 0.31800 (0.33088) | > loader_time: 0.00280 (0.00264)  --> STEP: 60/639 -- GLOBAL_STEP: 51900 | > loss: -0.48891 (-0.52623) | > log_mle: -0.58789 (-0.64334) | > loss_dur: 0.09899 (0.11711) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.52033 (10.21318) | > current_lr: 0.00028 | > step_time: 0.34130 (0.35869) | > loader_time: 0.00280 (0.00273)  --> STEP: 85/639 -- GLOBAL_STEP: 51925 | > loss: -0.48392 (-0.51910) | > log_mle: -0.58764 (-0.63402) | > loss_dur: 0.10372 (0.11492) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.94415 (9.58935) | > current_lr: 0.00028 | > step_time: 0.41300 (0.37269) | > loader_time: 0.00280 (0.00278)  --> STEP: 110/639 -- GLOBAL_STEP: 51950 | > loss: -0.49283 (-0.51956) | > log_mle: -0.60038 (-0.63502) | > loss_dur: 0.10754 (0.11546) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.95344 (10.16452) | > current_lr: 0.00028 | > step_time: 0.41720 (0.39268) | > loader_time: 0.00300 (0.00285)  --> STEP: 135/639 -- GLOBAL_STEP: 51975 | > loss: -0.45706 (-0.51679) | > log_mle: -0.58089 (-0.63222) | > loss_dur: 0.12383 (0.11543) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.36560 (10.02748) | > current_lr: 0.00028 | > step_time: 0.46250 (0.40562) | > loader_time: 0.00300 (0.00288)  --> STEP: 160/639 -- GLOBAL_STEP: 52000 | > loss: -0.51914 (-0.51484) | > log_mle: -0.62074 (-0.62976) | > loss_dur: 0.10160 (0.11493) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.38119 (9.82724) | > current_lr: 0.00028 | > step_time: 0.45530 (0.42109) | > loader_time: 0.00320 (0.00292)  --> STEP: 185/639 -- GLOBAL_STEP: 52025 | > loss: -0.49164 (-0.51620) | > log_mle: -0.60066 (-0.63154) | > loss_dur: 0.10901 (0.11534) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.30583 (10.05040) | > current_lr: 0.00028 | > step_time: 0.57600 (0.43785) | > loader_time: 0.00320 (0.00297)  --> STEP: 210/639 -- GLOBAL_STEP: 52050 | > loss: -0.42442 (-0.51318) | > log_mle: -0.52332 (-0.62828) | > loss_dur: 0.09889 (0.11510) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.98788 (9.82582) | > current_lr: 0.00028 | > step_time: 0.50210 (0.45344) | > loader_time: 0.00320 (0.00300)  --> STEP: 235/639 -- GLOBAL_STEP: 52075 | > loss: -0.49916 (-0.51357) | > log_mle: -0.61201 (-0.62845) | > loss_dur: 0.11284 (0.11488) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.41675 (9.89124) | > current_lr: 0.00028 | > step_time: 0.55690 (0.47020) | > loader_time: 0.00350 (0.00304)  --> STEP: 260/639 -- GLOBAL_STEP: 52100 | > loss: -0.49389 (-0.51263) | > log_mle: -0.61290 (-0.62777) | > loss_dur: 0.11901 (0.11515) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.45765 (9.91027) | > current_lr: 0.00028 | > step_time: 0.69800 (0.48664) | > loader_time: 0.00340 (0.00308)  --> STEP: 285/639 -- GLOBAL_STEP: 52125 | > loss: -0.44894 (-0.51118) | > log_mle: -0.56147 (-0.62657) | > loss_dur: 0.11254 (0.11539) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.17492 (9.92146) | > current_lr: 0.00028 | > step_time: 0.60760 (0.50265) | > loader_time: 0.00340 (0.00312)  --> STEP: 310/639 -- GLOBAL_STEP: 52150 | > loss: -0.48936 (-0.50930) | > log_mle: -0.60565 (-0.62500) | > loss_dur: 0.11629 (0.11570) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.83478 (9.82829) | > current_lr: 0.00028 | > step_time: 0.58190 (0.51609) | > loader_time: 0.00330 (0.00315)  --> STEP: 335/639 -- GLOBAL_STEP: 52175 | > loss: -0.50348 (-0.50675) | > log_mle: -0.62848 (-0.62290) | > loss_dur: 0.12500 (0.11615) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.17486 (9.72276) | > current_lr: 0.00028 | > step_time: 0.65620 (0.53126) | > loader_time: 0.00350 (0.00318)  --> STEP: 360/639 -- GLOBAL_STEP: 52200 | > loss: -0.45258 (-0.50493) | > log_mle: -0.56340 (-0.62088) | > loss_dur: 0.11082 (0.11596) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.80283 (9.64751) | > current_lr: 0.00028 | > step_time: 0.73110 (0.54430) | > loader_time: 0.00330 (0.00320)  --> STEP: 385/639 -- GLOBAL_STEP: 52225 | > loss: -0.48712 (-0.50312) | > log_mle: -0.59799 (-0.61901) | > loss_dur: 0.11087 (0.11588) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.83915 (9.46251) | > current_lr: 0.00028 | > step_time: 0.69780 (0.55776) | > loader_time: 0.00370 (0.00323)  --> STEP: 410/639 -- GLOBAL_STEP: 52250 | > loss: -0.44228 (-0.50122) | > log_mle: -0.57291 (-0.61729) | > loss_dur: 0.13063 (0.11607) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.23768 (9.34076) | > current_lr: 0.00028 | > step_time: 0.78060 (0.56955) | > loader_time: 0.00370 (0.00326)  --> STEP: 435/639 -- GLOBAL_STEP: 52275 | > loss: -0.47095 (-0.49925) | > log_mle: -0.58409 (-0.61525) | > loss_dur: 0.11314 (0.11600) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.93411 (9.24097) | > current_lr: 0.00028 | > step_time: 0.68900 (0.58193) | > loader_time: 0.00360 (0.00329)  --> STEP: 460/639 -- GLOBAL_STEP: 52300 | > loss: -0.42722 (-0.49727) | > log_mle: -0.54856 (-0.61326) | > loss_dur: 0.12134 (0.11600) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.32223 (9.08676) | > current_lr: 0.00028 | > step_time: 0.91120 (0.59306) | > loader_time: 0.00410 (0.00331)  --> STEP: 485/639 -- GLOBAL_STEP: 52325 | > loss: -0.47275 (-0.49571) | > log_mle: -0.58260 (-0.61186) | > loss_dur: 0.10985 (0.11615) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.29753 (9.04427) | > current_lr: 0.00028 | > step_time: 0.83560 (0.60539) | > loader_time: 0.00410 (0.00334)  --> STEP: 510/639 -- GLOBAL_STEP: 52350 | > loss: -0.46587 (-0.49446) | > log_mle: -0.58605 (-0.61058) | > loss_dur: 0.12018 (0.11612) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.35249 (8.99868) | > current_lr: 0.00028 | > step_time: 0.84620 (0.61857) | > loader_time: 0.00400 (0.00337)  --> STEP: 535/639 -- GLOBAL_STEP: 52375 | > loss: -0.44396 (-0.49293) | > log_mle: -0.57270 (-0.60915) | > loss_dur: 0.12873 (0.11622) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.33356 (8.93037) | > current_lr: 0.00028 | > step_time: 0.89460 (0.63168) | > loader_time: 0.00410 (0.00340)  --> STEP: 560/639 -- GLOBAL_STEP: 52400 | > loss: -0.45319 (-0.49149) | > log_mle: -0.57951 (-0.60780) | > loss_dur: 0.12632 (0.11631) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.74069 (8.86915) | > current_lr: 0.00028 | > step_time: 0.92380 (0.64581) | > loader_time: 0.00420 (0.00343)  --> STEP: 585/639 -- GLOBAL_STEP: 52425 | > loss: -0.47303 (-0.49029) | > log_mle: -0.58555 (-0.60662) | > loss_dur: 0.11252 (0.11633) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.15141 (8.80572) | > current_lr: 0.00028 | > step_time: 1.04120 (0.66105) | > loader_time: 0.00440 (0.00347)  --> STEP: 610/639 -- GLOBAL_STEP: 52450 | > loss: -0.44110 (-0.48905) | > log_mle: -0.55534 (-0.60553) | > loss_dur: 0.11424 (0.11647) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.00719 (8.76401) | > current_lr: 0.00028 | > step_time: 1.03940 (0.67770) | > loader_time: 0.00430 (0.00351)  --> STEP: 635/639 -- GLOBAL_STEP: 52475 | > loss: -0.46320 (-0.48803) | > log_mle: -0.58135 (-0.60478) | > loss_dur: 0.11814 (0.11676) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.99090 (8.73142) | > current_lr: 0.00028 | > step_time: 1.12330 (0.69458) | > loader_time: 0.00430 (0.00355) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02666 (-0.00557) | > avg_loss: -0.45108 (+0.00394) | > avg_log_mle: -0.58665 (+0.00119) | > avg_loss_dur: 0.13557 (+0.00274)  > EPOCH: 82/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 01:20:00)   --> STEP: 20/639 -- GLOBAL_STEP: 52500 | > loss: -0.61750 (-0.54929) | > log_mle: -0.76904 (-0.66983) | > loss_dur: 0.15154 (0.12054) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 23.01052 (12.95187) | > current_lr: 0.00028 | > step_time: 0.34580 (0.30295) | > loader_time: 0.00290 (0.00268)  --> STEP: 45/639 -- GLOBAL_STEP: 52525 | > loss: -0.49345 (-0.52824) | > log_mle: -0.58863 (-0.64388) | > loss_dur: 0.09519 (0.11564) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.39052 (10.73390) | > current_lr: 0.00028 | > step_time: 0.46210 (0.33776) | > loader_time: 0.00280 (0.00280)  --> STEP: 70/639 -- GLOBAL_STEP: 52550 | > loss: -0.52648 (-0.52002) | > log_mle: -0.64245 (-0.63504) | > loss_dur: 0.11596 (0.11502) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.04068 (9.83271) | > current_lr: 0.00028 | > step_time: 0.39230 (0.36368) | > loader_time: 0.00270 (0.00284)  --> STEP: 95/639 -- GLOBAL_STEP: 52575 | > loss: -0.57608 (-0.51788) | > log_mle: -0.68829 (-0.63098) | > loss_dur: 0.11222 (0.11310) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.44850 (9.46953) | > current_lr: 0.00028 | > step_time: 0.41030 (0.38001) | > loader_time: 0.00300 (0.00286)  --> STEP: 120/639 -- GLOBAL_STEP: 52600 | > loss: -0.56122 (-0.51568) | > log_mle: -0.69328 (-0.62971) | > loss_dur: 0.13205 (0.11403) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.98933 (9.80723) | > current_lr: 0.00028 | > step_time: 0.63230 (0.39840) | > loader_time: 0.00320 (0.00291)  --> STEP: 145/639 -- GLOBAL_STEP: 52625 | > loss: -0.51822 (-0.51579) | > log_mle: -0.64043 (-0.63047) | > loss_dur: 0.12222 (0.11468) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.49439 (10.12622) | > current_lr: 0.00028 | > step_time: 0.47030 (0.41320) | > loader_time: 0.00290 (0.00294)  --> STEP: 170/639 -- GLOBAL_STEP: 52650 | > loss: -0.50826 (-0.51473) | > log_mle: -0.62508 (-0.62942) | > loss_dur: 0.11682 (0.11469) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.82984 (10.26189) | > current_lr: 0.00028 | > step_time: 0.54060 (0.42731) | > loader_time: 0.00340 (0.00297)  --> STEP: 195/639 -- GLOBAL_STEP: 52675 | > loss: -0.47156 (-0.51404) | > log_mle: -0.56637 (-0.62817) | > loss_dur: 0.09481 (0.11413) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.23711 (10.15249) | > current_lr: 0.00028 | > step_time: 0.50070 (0.44411) | > loader_time: 0.00320 (0.00301)  --> STEP: 220/639 -- GLOBAL_STEP: 52700 | > loss: -0.51238 (-0.51159) | > log_mle: -0.61726 (-0.62573) | > loss_dur: 0.10488 (0.11414) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.45280 (9.94029) | > current_lr: 0.00028 | > step_time: 0.63340 (0.46056) | > loader_time: 0.00350 (0.00305)  --> STEP: 245/639 -- GLOBAL_STEP: 52725 | > loss: -0.49380 (-0.51241) | > log_mle: -0.60410 (-0.62646) | > loss_dur: 0.11031 (0.11405) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.28516 (10.04504) | > current_lr: 0.00028 | > step_time: 0.52970 (0.47416) | > loader_time: 0.00300 (0.00307)  --> STEP: 270/639 -- GLOBAL_STEP: 52750 | > loss: -0.52853 (-0.51191) | > log_mle: -0.66702 (-0.62644) | > loss_dur: 0.13849 (0.11453) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 29.72665 (10.23354) | > current_lr: 0.00028 | > step_time: 0.83070 (0.49180) | > loader_time: 0.00350 (0.00310)  --> STEP: 295/639 -- GLOBAL_STEP: 52775 | > loss: -0.46560 (-0.50971) | > log_mle: -0.56376 (-0.62448) | > loss_dur: 0.09816 (0.11477) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.05046 (10.06153) | > current_lr: 0.00028 | > step_time: 0.65620 (0.50658) | > loader_time: 0.00360 (0.00313)  --> STEP: 320/639 -- GLOBAL_STEP: 52800 | > loss: -0.47741 (-0.50806) | > log_mle: -0.60986 (-0.62317) | > loss_dur: 0.13244 (0.11512) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.15473 (10.00785) | > current_lr: 0.00028 | > step_time: 0.75250 (0.52328) | > loader_time: 0.00370 (0.00316)  --> STEP: 345/639 -- GLOBAL_STEP: 52825 | > loss: -0.48212 (-0.50605) | > log_mle: -0.59588 (-0.62143) | > loss_dur: 0.11376 (0.11538) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.54259 (9.93185) | > current_lr: 0.00028 | > step_time: 0.64080 (0.53731) | > loader_time: 0.00350 (0.00318)  --> STEP: 370/639 -- GLOBAL_STEP: 52850 | > loss: -0.48884 (-0.50405) | > log_mle: -0.60652 (-0.61926) | > loss_dur: 0.11768 (0.11521) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.81422 (9.75575) | > current_lr: 0.00028 | > step_time: 0.86590 (0.54993) | > loader_time: 0.00400 (0.00321)  --> STEP: 395/639 -- GLOBAL_STEP: 52875 | > loss: -0.45892 (-0.50210) | > log_mle: -0.57933 (-0.61731) | > loss_dur: 0.12041 (0.11521) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.62324 (9.58271) | > current_lr: 0.00028 | > step_time: 0.75610 (0.56177) | > loader_time: 0.00360 (0.00324)  --> STEP: 420/639 -- GLOBAL_STEP: 52900 | > loss: -0.46580 (-0.50002) | > log_mle: -0.57613 (-0.61549) | > loss_dur: 0.11033 (0.11547) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.05344 (9.46783) | > current_lr: 0.00028 | > step_time: 0.70860 (0.57406) | > loader_time: 0.00340 (0.00327)  --> STEP: 445/639 -- GLOBAL_STEP: 52925 | > loss: -0.46859 (-0.49848) | > log_mle: -0.58163 (-0.61401) | > loss_dur: 0.11305 (0.11553) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.93223 (9.40493) | > current_lr: 0.00028 | > step_time: 0.73410 (0.58614) | > loader_time: 0.00380 (0.00330)  --> STEP: 470/639 -- GLOBAL_STEP: 52950 | > loss: -0.47707 (-0.49660) | > log_mle: -0.59291 (-0.61216) | > loss_dur: 0.11584 (0.11556) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.88616 (9.27297) | > current_lr: 0.00027 | > step_time: 0.76890 (0.59772) | > loader_time: 0.00380 (0.00332)  --> STEP: 495/639 -- GLOBAL_STEP: 52975 | > loss: -0.42617 (-0.49503) | > log_mle: -0.52951 (-0.61074) | > loss_dur: 0.10334 (0.11571) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.06633 (9.18752) | > current_lr: 0.00027 | > step_time: 0.75000 (0.60984) | > loader_time: 0.00370 (0.00335)  --> STEP: 520/639 -- GLOBAL_STEP: 53000 | > loss: -0.47552 (-0.49391) | > log_mle: -0.58726 (-0.60960) | > loss_dur: 0.11174 (0.11569) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.08678 (9.13685) | > current_lr: 0.00027 | > step_time: 0.86360 (0.62267) | > loader_time: 0.00380 (0.00338)  --> STEP: 545/639 -- GLOBAL_STEP: 53025 | > loss: -0.48485 (-0.49255) | > log_mle: -0.61675 (-0.60841) | > loss_dur: 0.13190 (0.11587) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.41207 (9.08001) | > current_lr: 0.00027 | > step_time: 0.93040 (0.63795) | > loader_time: 0.00420 (0.00341)  --> STEP: 570/639 -- GLOBAL_STEP: 53050 | > loss: -0.45693 (-0.49101) | > log_mle: -0.56475 (-0.60696) | > loss_dur: 0.10781 (0.11595) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.48102 (8.97711) | > current_lr: 0.00027 | > step_time: 0.85350 (0.65157) | > loader_time: 0.00400 (0.00344)  --> STEP: 595/639 -- GLOBAL_STEP: 53075 | > loss: -0.45108 (-0.48997) | > log_mle: -0.55937 (-0.60604) | > loss_dur: 0.10828 (0.11607) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.09919 (8.91726) | > current_lr: 0.00027 | > step_time: 0.97210 (0.66769) | > loader_time: 0.00430 (0.00348)  --> STEP: 620/639 -- GLOBAL_STEP: 53100 | > loss: -0.45124 (-0.48873) | > log_mle: -0.56627 (-0.60492) | > loss_dur: 0.11504 (0.11619) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.90012 (8.84979) | > current_lr: 0.00027 | > step_time: 1.11060 (0.68319) | > loader_time: 0.00440 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02573 (-0.00093) | > avg_loss: -0.45371 (-0.00263) | > avg_log_mle: -0.58534 (+0.00131) | > avg_loss_dur: 0.13164 (-0.00394)  > EPOCH: 83/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 01:27:39)   --> STEP: 5/639 -- GLOBAL_STEP: 53125 | > loss: -0.54137 (-0.54347) | > log_mle: -0.69674 (-0.67440) | > loss_dur: 0.15537 (0.13093) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.88631 (12.31318) | > current_lr: 0.00027 | > step_time: 0.30720 (0.28000) | > loader_time: 0.00330 (0.00259)  --> STEP: 30/639 -- GLOBAL_STEP: 53150 | > loss: -0.54878 (-0.53703) | > log_mle: -0.65149 (-0.65477) | > loss_dur: 0.10271 (0.11774) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.79958 (10.97826) | > current_lr: 0.00027 | > step_time: 0.32900 (0.32528) | > loader_time: 0.00270 (0.00258)  --> STEP: 55/639 -- GLOBAL_STEP: 53175 | > loss: -0.49204 (-0.52853) | > log_mle: -0.58827 (-0.64369) | > loss_dur: 0.09622 (0.11516) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.81463 (10.05921) | > current_lr: 0.00027 | > step_time: 0.33100 (0.34755) | > loader_time: 0.00280 (0.00263)  --> STEP: 80/639 -- GLOBAL_STEP: 53200 | > loss: -0.48632 (-0.52209) | > log_mle: -0.62220 (-0.63558) | > loss_dur: 0.13588 (0.11348) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.45513 (9.51140) | > current_lr: 0.00027 | > step_time: 0.38490 (0.36780) | > loader_time: 0.00290 (0.00273)  --> STEP: 105/639 -- GLOBAL_STEP: 53225 | > loss: -0.46703 (-0.52025) | > log_mle: -0.57809 (-0.63322) | > loss_dur: 0.11106 (0.11297) | > amp_scaler: 8192.00000 (4408.07619) | > grad_norm: 8.23051 (9.31691) | > current_lr: 0.00027 | > step_time: 0.41060 (0.38642) | > loader_time: 0.00290 (0.00277)  --> STEP: 130/639 -- GLOBAL_STEP: 53250 | > loss: -0.49572 (-0.51771) | > log_mle: -0.62074 (-0.63151) | > loss_dur: 0.12502 (0.11381) | > amp_scaler: 8192.00000 (5135.75385) | > grad_norm: 9.38522 (9.34930) | > current_lr: 0.00027 | > step_time: 0.49030 (0.40111) | > loader_time: 0.00310 (0.00282)  --> STEP: 155/639 -- GLOBAL_STEP: 53275 | > loss: -0.49093 (-0.51669) | > log_mle: -0.60219 (-0.63059) | > loss_dur: 0.11127 (0.11390) | > amp_scaler: 4096.00000 (5126.60645) | > grad_norm: 10.73286 (9.37697) | > current_lr: 0.00027 | > step_time: 0.53440 (0.41850) | > loader_time: 0.00310 (0.00285)  --> STEP: 180/639 -- GLOBAL_STEP: 53300 | > loss: -0.54279 (-0.51759) | > log_mle: -0.66178 (-0.63190) | > loss_dur: 0.11899 (0.11431) | > amp_scaler: 4096.00000 (4983.46667) | > grad_norm: 15.50481 (9.65393) | > current_lr: 0.00027 | > step_time: 0.57180 (0.43422) | > loader_time: 0.00340 (0.00290)  --> STEP: 205/639 -- GLOBAL_STEP: 53325 | > loss: -0.46551 (-0.51432) | > log_mle: -0.58225 (-0.62824) | > loss_dur: 0.11674 (0.11392) | > amp_scaler: 4096.00000 (4875.23902) | > grad_norm: 7.54304 (9.45069) | > current_lr: 0.00027 | > step_time: 0.56250 (0.44843) | > loader_time: 0.00340 (0.00294)  --> STEP: 230/639 -- GLOBAL_STEP: 53350 | > loss: -0.50762 (-0.51395) | > log_mle: -0.62060 (-0.62798) | > loss_dur: 0.11298 (0.11404) | > amp_scaler: 4096.00000 (4790.53913) | > grad_norm: 7.21155 (9.53251) | > current_lr: 0.00027 | > step_time: 0.55390 (0.46470) | > loader_time: 0.00350 (0.00298)  --> STEP: 255/639 -- GLOBAL_STEP: 53375 | > loss: -0.47294 (-0.51275) | > log_mle: -0.59034 (-0.62684) | > loss_dur: 0.11740 (0.11409) | > amp_scaler: 4096.00000 (4722.44706) | > grad_norm: 8.63989 (9.55333) | > current_lr: 0.00027 | > step_time: 0.61130 (0.47860) | > loader_time: 0.00380 (0.00301)  --> STEP: 280/639 -- GLOBAL_STEP: 53400 | > loss: -0.48806 (-0.51233) | > log_mle: -0.61375 (-0.62701) | > loss_dur: 0.12569 (0.11468) | > amp_scaler: 4096.00000 (4666.51429) | > grad_norm: 6.97443 (9.69813) | > current_lr: 0.00027 | > step_time: 0.58910 (0.49439) | > loader_time: 0.00340 (0.00305)  --> STEP: 305/639 -- GLOBAL_STEP: 53425 | > loss: -0.47426 (-0.51049) | > log_mle: -0.59519 (-0.62532) | > loss_dur: 0.12093 (0.11482) | > amp_scaler: 4096.00000 (4619.75082) | > grad_norm: 7.88616 (9.64604) | > current_lr: 0.00027 | > step_time: 0.66510 (0.50867) | > loader_time: 0.00330 (0.00308)  --> STEP: 330/639 -- GLOBAL_STEP: 53450 | > loss: -0.48853 (-0.50793) | > log_mle: -0.61364 (-0.62332) | > loss_dur: 0.12511 (0.11539) | > amp_scaler: 4096.00000 (4580.07273) | > grad_norm: 10.16858 (9.58305) | > current_lr: 0.00027 | > step_time: 0.73650 (0.52345) | > loader_time: 0.00350 (0.00312)  --> STEP: 355/639 -- GLOBAL_STEP: 53475 | > loss: -0.44193 (-0.50611) | > log_mle: -0.54564 (-0.62141) | > loss_dur: 0.10371 (0.11530) | > amp_scaler: 4096.00000 (4545.98310) | > grad_norm: 2.39504 (9.46995) | > current_lr: 0.00027 | > step_time: 0.60230 (0.53652) | > loader_time: 0.00300 (0.00315)  --> STEP: 380/639 -- GLOBAL_STEP: 53500 | > loss: -0.48067 (-0.50402) | > log_mle: -0.58236 (-0.61925) | > loss_dur: 0.10169 (0.11523) | > amp_scaler: 4096.00000 (4516.37895) | > grad_norm: 4.96861 (9.30645) | > current_lr: 0.00027 | > step_time: 0.63690 (0.54915) | > loader_time: 0.00320 (0.00317)  --> STEP: 405/639 -- GLOBAL_STEP: 53525 | > loss: -0.47119 (-0.50258) | > log_mle: -0.58535 (-0.61801) | > loss_dur: 0.11416 (0.11543) | > amp_scaler: 4096.00000 (4490.42963) | > grad_norm: 8.98037 (9.24832) | > current_lr: 0.00027 | > step_time: 0.74930 (0.56144) | > loader_time: 0.00320 (0.00320)  --> STEP: 430/639 -- GLOBAL_STEP: 53550 | > loss: -0.43871 (-0.50057) | > log_mle: -0.54517 (-0.61613) | > loss_dur: 0.10646 (0.11556) | > amp_scaler: 4096.00000 (4467.49767) | > grad_norm: 4.36008 (9.15114) | > current_lr: 0.00027 | > step_time: 0.76820 (0.57484) | > loader_time: 0.00370 (0.00323)  --> STEP: 455/639 -- GLOBAL_STEP: 53575 | > loss: -0.45333 (-0.49866) | > log_mle: -0.56947 (-0.61415) | > loss_dur: 0.11613 (0.11549) | > amp_scaler: 4096.00000 (4447.08571) | > grad_norm: 5.50964 (9.03450) | > current_lr: 0.00027 | > step_time: 0.80130 (0.58610) | > loader_time: 0.00370 (0.00325)  --> STEP: 480/639 -- GLOBAL_STEP: 53600 | > loss: -0.41416 (-0.49684) | > log_mle: -0.52676 (-0.61258) | > loss_dur: 0.11260 (0.11574) | > amp_scaler: 4096.00000 (4428.80000) | > grad_norm: 2.92463 (8.93572) | > current_lr: 0.00027 | > step_time: 0.85010 (0.59916) | > loader_time: 0.00380 (0.00327)  --> STEP: 505/639 -- GLOBAL_STEP: 53625 | > loss: -0.47783 (-0.49551) | > log_mle: -0.59483 (-0.61117) | > loss_dur: 0.11700 (0.11566) | > amp_scaler: 4096.00000 (4412.32475) | > grad_norm: 8.04023 (8.86463) | > current_lr: 0.00027 | > step_time: 0.95090 (0.61158) | > loader_time: 0.00420 (0.00330)  --> STEP: 530/639 -- GLOBAL_STEP: 53650 | > loss: -0.45541 (-0.49411) | > log_mle: -0.58348 (-0.60996) | > loss_dur: 0.12807 (0.11585) | > amp_scaler: 4096.00000 (4397.40377) | > grad_norm: 7.77002 (8.84024) | > current_lr: 0.00027 | > step_time: 0.99400 (0.62605) | > loader_time: 0.00400 (0.00333)  --> STEP: 555/639 -- GLOBAL_STEP: 53675 | > loss: -0.45473 (-0.49261) | > log_mle: -0.56691 (-0.60845) | > loss_dur: 0.11218 (0.11585) | > amp_scaler: 4096.00000 (4383.82703) | > grad_norm: 4.36846 (8.76130) | > current_lr: 0.00027 | > step_time: 0.82830 (0.63994) | > loader_time: 0.00390 (0.00336)  --> STEP: 580/639 -- GLOBAL_STEP: 53700 | > loss: -0.47624 (-0.49139) | > log_mle: -0.58433 (-0.60729) | > loss_dur: 0.10809 (0.11590) | > amp_scaler: 4096.00000 (4371.42069) | > grad_norm: 7.30562 (8.68359) | > current_lr: 0.00027 | > step_time: 0.96900 (0.65530) | > loader_time: 0.00430 (0.00340)  --> STEP: 605/639 -- GLOBAL_STEP: 53725 | > loss: -0.44882 (-0.49017) | > log_mle: -0.55468 (-0.60620) | > loss_dur: 0.10586 (0.11603) | > amp_scaler: 4096.00000 (4360.03967) | > grad_norm: 3.50121 (8.64380) | > current_lr: 0.00027 | > step_time: 1.05620 (0.67167) | > loader_time: 0.00450 (0.00345)  --> STEP: 630/639 -- GLOBAL_STEP: 53750 | > loss: -0.45139 (-0.48898) | > log_mle: -0.58371 (-0.60523) | > loss_dur: 0.13232 (0.11625) | > amp_scaler: 4096.00000 (4349.56190) | > grad_norm: 11.01070 (8.60711) | > current_lr: 0.00027 | > step_time: 1.25040 (0.68939) | > loader_time: 0.00640 (0.00350) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.04689 (+0.02116) | > avg_loss: -0.45640 (-0.00269) | > avg_log_mle: -0.59020 (-0.00486) | > avg_loss_dur: 0.13381 (+0.00217)  > EPOCH: 84/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 01:35:17)   --> STEP: 15/639 -- GLOBAL_STEP: 53775 | > loss: -0.65342 (-0.54637) | > log_mle: -0.78862 (-0.66772) | > loss_dur: 0.13520 (0.12135) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 20.82582 (12.54389) | > current_lr: 0.00027 | > step_time: 0.33060 (0.29601) | > loader_time: 0.00260 (0.00595)  --> STEP: 40/639 -- GLOBAL_STEP: 53800 | > loss: -0.50706 (-0.53400) | > log_mle: -0.62415 (-0.65051) | > loss_dur: 0.11709 (0.11652) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.83469 (10.86847) | > current_lr: 0.00027 | > step_time: 0.35190 (0.32767) | > loader_time: 0.00280 (0.00395)  --> STEP: 65/639 -- GLOBAL_STEP: 53825 | > loss: -0.51895 (-0.52633) | > log_mle: -0.61347 (-0.64280) | > loss_dur: 0.09451 (0.11647) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.44354 (10.26161) | > current_lr: 0.00027 | > step_time: 0.38630 (0.35700) | > loader_time: 0.00290 (0.00357)  --> STEP: 90/639 -- GLOBAL_STEP: 53850 | > loss: -0.47035 (-0.51955) | > log_mle: -0.57348 (-0.63379) | > loss_dur: 0.10313 (0.11424) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.55808 (9.28392) | > current_lr: 0.00027 | > step_time: 0.51630 (0.37509) | > loader_time: 0.00310 (0.00340)  --> STEP: 115/639 -- GLOBAL_STEP: 53875 | > loss: -0.43199 (-0.51931) | > log_mle: -0.54078 (-0.63358) | > loss_dur: 0.10879 (0.11426) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.37732 (9.49795) | > current_lr: 0.00027 | > step_time: 0.47360 (0.39419) | > loader_time: 0.00280 (0.00333)  --> STEP: 140/639 -- GLOBAL_STEP: 53900 | > loss: -0.48976 (-0.51896) | > log_mle: -0.61884 (-0.63363) | > loss_dur: 0.12908 (0.11466) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.05997 (9.68407) | > current_lr: 0.00027 | > step_time: 0.55110 (0.40830) | > loader_time: 0.00310 (0.00327)  --> STEP: 165/639 -- GLOBAL_STEP: 53925 | > loss: -0.50055 (-0.51799) | > log_mle: -0.62733 (-0.63225) | > loss_dur: 0.12679 (0.11426) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.69520 (9.68681) | > current_lr: 0.00027 | > step_time: 0.48670 (0.42149) | > loader_time: 0.00270 (0.00323)  --> STEP: 190/639 -- GLOBAL_STEP: 53950 | > loss: -0.44547 (-0.51745) | > log_mle: -0.55687 (-0.63160) | > loss_dur: 0.11140 (0.11415) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.62607 (9.82581) | > current_lr: 0.00027 | > step_time: 0.53040 (0.43906) | > loader_time: 0.00310 (0.00323)  --> STEP: 215/639 -- GLOBAL_STEP: 53975 | > loss: -0.54329 (-0.51475) | > log_mle: -0.65851 (-0.62917) | > loss_dur: 0.11522 (0.11441) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.27670 (9.71257) | > current_lr: 0.00027 | > step_time: 0.54250 (0.45498) | > loader_time: 0.00330 (0.00323)  --> STEP: 240/639 -- GLOBAL_STEP: 54000 | > loss: -0.55230 (-0.51538) | > log_mle: -0.67132 (-0.62977) | > loss_dur: 0.11902 (0.11439) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 17.96016 (9.79330) | > current_lr: 0.00027 | > step_time: 0.56210 (0.46991) | > loader_time: 0.00320 (0.00324)  --> STEP: 265/639 -- GLOBAL_STEP: 54025 | > loss: -0.55382 (-0.51410) | > log_mle: -0.68869 (-0.62901) | > loss_dur: 0.13487 (0.11491) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.07821 (9.86755) | > current_lr: 0.00027 | > step_time: 0.62890 (0.48511) | > loader_time: 0.00340 (0.00325)  --> STEP: 290/639 -- GLOBAL_STEP: 54050 | > loss: -0.47149 (-0.51221) | > log_mle: -0.59029 (-0.62735) | > loss_dur: 0.11880 (0.11514) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.37609 (9.81595) | > current_lr: 0.00027 | > step_time: 0.61240 (0.50031) | > loader_time: 0.00340 (0.00326)  --> STEP: 315/639 -- GLOBAL_STEP: 54075 | > loss: -0.45971 (-0.51027) | > log_mle: -0.56960 (-0.62570) | > loss_dur: 0.10990 (0.11543) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.80362 (9.77529) | > current_lr: 0.00027 | > step_time: 0.63820 (0.51508) | > loader_time: 0.00330 (0.00328)  --> STEP: 340/639 -- GLOBAL_STEP: 54100 | > loss: -0.47959 (-0.50833) | > log_mle: -0.59072 (-0.62424) | > loss_dur: 0.11113 (0.11591) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.81936 (9.75313) | > current_lr: 0.00027 | > step_time: 0.59130 (0.53098) | > loader_time: 0.00340 (0.00331)  --> STEP: 365/639 -- GLOBAL_STEP: 54125 | > loss: -0.47786 (-0.50622) | > log_mle: -0.59373 (-0.62181) | > loss_dur: 0.11587 (0.11559) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.02493 (9.55797) | > current_lr: 0.00027 | > step_time: 0.83570 (0.54245) | > loader_time: 0.00390 (0.00333)  --> STEP: 390/639 -- GLOBAL_STEP: 54150 | > loss: -0.47351 (-0.50438) | > log_mle: -0.59608 (-0.61976) | > loss_dur: 0.12257 (0.11539) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.87215 (9.37577) | > current_lr: 0.00027 | > step_time: 0.74140 (0.55583) | > loader_time: 0.00360 (0.00335)  --> STEP: 415/639 -- GLOBAL_STEP: 54175 | > loss: -0.45202 (-0.50235) | > log_mle: -0.57179 (-0.61803) | > loss_dur: 0.11977 (0.11568) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.28944 (9.29372) | > current_lr: 0.00027 | > step_time: 0.82560 (0.56885) | > loader_time: 0.00350 (0.00337)  --> STEP: 440/639 -- GLOBAL_STEP: 54200 | > loss: -0.43962 (-0.50063) | > log_mle: -0.55501 (-0.61641) | > loss_dur: 0.11538 (0.11578) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.25490 (9.21517) | > current_lr: 0.00027 | > step_time: 0.72240 (0.58110) | > loader_time: 0.00370 (0.00338)  --> STEP: 465/639 -- GLOBAL_STEP: 54225 | > loss: -0.45102 (-0.49862) | > log_mle: -0.56031 (-0.61435) | > loss_dur: 0.10929 (0.11573) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.24189 (9.05101) | > current_lr: 0.00027 | > step_time: 0.79220 (0.59193) | > loader_time: 0.00370 (0.00339)  --> STEP: 490/639 -- GLOBAL_STEP: 54250 | > loss: -0.44575 (-0.49698) | > log_mle: -0.54251 (-0.61282) | > loss_dur: 0.09675 (0.11585) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.71491 (8.99325) | > current_lr: 0.00027 | > step_time: 0.78360 (0.60549) | > loader_time: 0.00330 (0.00342)  --> STEP: 515/639 -- GLOBAL_STEP: 54275 | > loss: -0.46289 (-0.49587) | > log_mle: -0.57233 (-0.61169) | > loss_dur: 0.10945 (0.11582) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.94161 (8.93338) | > current_lr: 0.00027 | > step_time: 0.96730 (0.61855) | > loader_time: 0.00390 (0.00344)  --> STEP: 540/639 -- GLOBAL_STEP: 54300 | > loss: -0.46737 (-0.49440) | > log_mle: -0.58812 (-0.61029) | > loss_dur: 0.12075 (0.11589) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.09077 (8.85019) | > current_lr: 0.00027 | > step_time: 0.88350 (0.63259) | > loader_time: 0.00420 (0.00347)  --> STEP: 565/639 -- GLOBAL_STEP: 54325 | > loss: -0.43658 (-0.49280) | > log_mle: -0.54847 (-0.60879) | > loss_dur: 0.11189 (0.11600) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.33567 (8.77365) | > current_lr: 0.00027 | > step_time: 0.98980 (0.64640) | > loader_time: 0.00420 (0.00350)  --> STEP: 590/639 -- GLOBAL_STEP: 54350 | > loss: -0.46178 (-0.49177) | > log_mle: -0.58051 (-0.60787) | > loss_dur: 0.11873 (0.11609) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.17162 (8.73276) | > current_lr: 0.00027 | > step_time: 0.96610 (0.66238) | > loader_time: 0.00430 (0.00353)  --> STEP: 615/639 -- GLOBAL_STEP: 54375 | > loss: -0.46667 (-0.49050) | > log_mle: -0.58443 (-0.60670) | > loss_dur: 0.11776 (0.11620) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.81679 (8.66801) | > current_lr: 0.00027 | > step_time: 1.01670 (0.67852) | > loader_time: 0.00440 (0.00357) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02246 (-0.02443) | > avg_loss: -0.45541 (+0.00098) | > avg_log_mle: -0.58979 (+0.00041) | > avg_loss_dur: 0.13437 (+0.00057)  > EPOCH: 85/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 01:42:55)   --> STEP: 0/639 -- GLOBAL_STEP: 54400 | > loss: -0.57203 (-0.57203) | > log_mle: -0.71958 (-0.71958) | > loss_dur: 0.14755 (0.14755) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 17.27270 (17.27270) | > current_lr: 0.00027 | > step_time: 0.69050 (0.69050) | > loader_time: 0.54510 (0.54508)  --> STEP: 25/639 -- GLOBAL_STEP: 54425 | > loss: -0.53167 (-0.54904) | > log_mle: -0.66318 (-0.66719) | > loss_dur: 0.13151 (0.11815) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.77309 (13.06265) | > current_lr: 0.00027 | > step_time: 0.37110 (0.32210) | > loader_time: 0.00300 (0.00341)  --> STEP: 50/639 -- GLOBAL_STEP: 54450 | > loss: -0.49827 (-0.53539) | > log_mle: -0.59702 (-0.65149) | > loss_dur: 0.09875 (0.11609) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.26860 (11.76675) | > current_lr: 0.00027 | > step_time: 0.35030 (0.34438) | > loader_time: 0.00270 (0.00311)  --> STEP: 75/639 -- GLOBAL_STEP: 54475 | > loss: -0.51730 (-0.52631) | > log_mle: -0.62215 (-0.64143) | > loss_dur: 0.10485 (0.11512) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.07269 (11.04081) | > current_lr: 0.00027 | > step_time: 0.42130 (0.36691) | > loader_time: 0.00310 (0.00304)  --> STEP: 100/639 -- GLOBAL_STEP: 54500 | > loss: -0.52997 (-0.52420) | > log_mle: -0.63482 (-0.63784) | > loss_dur: 0.10485 (0.11364) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.53559 (10.77038) | > current_lr: 0.00027 | > step_time: 0.50140 (0.38375) | > loader_time: 0.00310 (0.00302)  --> STEP: 125/639 -- GLOBAL_STEP: 54525 | > loss: -0.51542 (-0.52137) | > log_mle: -0.62881 (-0.63535) | > loss_dur: 0.11339 (0.11398) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.86232 (10.93362) | > current_lr: 0.00027 | > step_time: 0.42680 (0.40018) | > loader_time: 0.00290 (0.00302)  --> STEP: 150/639 -- GLOBAL_STEP: 54550 | > loss: -0.47351 (-0.52008) | > log_mle: -0.58986 (-0.63397) | > loss_dur: 0.11635 (0.11389) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.76468 (10.84072) | > current_lr: 0.00027 | > step_time: 0.43840 (0.41697) | > loader_time: 0.00260 (0.00302)  --> STEP: 175/639 -- GLOBAL_STEP: 54575 | > loss: -0.50053 (-0.51995) | > log_mle: -0.60473 (-0.63377) | > loss_dur: 0.10420 (0.11382) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.50704 (11.06807) | > current_lr: 0.00027 | > step_time: 0.56810 (0.43272) | > loader_time: 0.00340 (0.00304)  --> STEP: 200/639 -- GLOBAL_STEP: 54600 | > loss: -0.48187 (-0.51769) | > log_mle: -0.59375 (-0.63114) | > loss_dur: 0.11188 (0.11345) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.25522 (10.78306) | > current_lr: 0.00027 | > step_time: 0.52290 (0.44643) | > loader_time: 0.00310 (0.00306)  --> STEP: 225/639 -- GLOBAL_STEP: 54625 | > loss: -0.47510 (-0.51521) | > log_mle: -0.57396 (-0.62875) | > loss_dur: 0.09885 (0.11355) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.70218 (10.61534) | > current_lr: 0.00027 | > step_time: 0.51940 (0.46154) | > loader_time: 0.00320 (0.00309)  --> STEP: 250/639 -- GLOBAL_STEP: 54650 | > loss: -0.47653 (-0.51570) | > log_mle: -0.57713 (-0.62950) | > loss_dur: 0.10061 (0.11380) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.99767 (10.77815) | > current_lr: 0.00027 | > step_time: 0.59690 (0.47621) | > loader_time: 0.00350 (0.00312)  --> STEP: 275/639 -- GLOBAL_STEP: 54675 | > loss: -0.49596 (-0.51541) | > log_mle: -0.61146 (-0.62967) | > loss_dur: 0.11550 (0.11426) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.33518 (10.92656) | > current_lr: 0.00027 | > step_time: 0.83780 (0.49343) | > loader_time: 0.00380 (0.00316)  --> STEP: 300/639 -- GLOBAL_STEP: 54700 | > loss: -0.49389 (-0.51300) | > log_mle: -0.60594 (-0.62755) | > loss_dur: 0.11206 (0.11455) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.59147 (10.70139) | > current_lr: 0.00027 | > step_time: 0.70890 (0.50730) | > loader_time: 0.00350 (0.00318)  --> STEP: 325/639 -- GLOBAL_STEP: 54725 | > loss: -0.48919 (-0.51093) | > log_mle: -0.60559 (-0.62593) | > loss_dur: 0.11639 (0.11500) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.26368 (10.60914) | > current_lr: 0.00027 | > step_time: 0.65190 (0.52239) | > loader_time: 0.00350 (0.00321)  --> STEP: 350/639 -- GLOBAL_STEP: 54750 | > loss: -0.50832 (-0.50879) | > log_mle: -0.63331 (-0.62390) | > loss_dur: 0.12499 (0.11511) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.83434 (10.47449) | > current_lr: 0.00027 | > step_time: 0.84870 (0.53638) | > loader_time: 0.00390 (0.00324)  --> STEP: 375/639 -- GLOBAL_STEP: 54775 | > loss: -0.45904 (-0.50658) | > log_mle: -0.56989 (-0.62149) | > loss_dur: 0.11085 (0.11491) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.24251 (10.25581) | > current_lr: 0.00027 | > step_time: 0.72220 (0.54979) | > loader_time: 0.00370 (0.00326)  --> STEP: 400/639 -- GLOBAL_STEP: 54800 | > loss: -0.48839 (-0.50491) | > log_mle: -0.60233 (-0.62006) | > loss_dur: 0.11394 (0.11514) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.82236 (10.14914) | > current_lr: 0.00027 | > step_time: 0.72810 (0.56189) | > loader_time: 0.00370 (0.00328)  --> STEP: 425/639 -- GLOBAL_STEP: 54825 | > loss: -0.53110 (-0.50273) | > log_mle: -0.65383 (-0.61799) | > loss_dur: 0.12273 (0.11526) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.89184 (9.98260) | > current_lr: 0.00027 | > step_time: 0.91180 (0.57426) | > loader_time: 0.00390 (0.00331)  --> STEP: 450/639 -- GLOBAL_STEP: 54850 | > loss: -0.45485 (-0.50099) | > log_mle: -0.55786 (-0.61618) | > loss_dur: 0.10302 (0.11518) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.76024 (9.82807) | > current_lr: 0.00027 | > step_time: 0.81500 (0.58609) | > loader_time: 0.00380 (0.00334)  --> STEP: 475/639 -- GLOBAL_STEP: 54875 | > loss: -0.46428 (-0.49910) | > log_mle: -0.56722 (-0.61444) | > loss_dur: 0.10294 (0.11534) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.28141 (9.73678) | > current_lr: 0.00027 | > step_time: 0.71240 (0.59748) | > loader_time: 0.00340 (0.00336)  --> STEP: 500/639 -- GLOBAL_STEP: 54900 | > loss: -0.51516 (-0.49776) | > log_mle: -0.63208 (-0.61319) | > loss_dur: 0.11691 (0.11543) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.43390 (9.65049) | > current_lr: 0.00027 | > step_time: 0.96540 (0.61028) | > loader_time: 0.00400 (0.00338)  --> STEP: 525/639 -- GLOBAL_STEP: 54925 | > loss: -0.47532 (-0.49636) | > log_mle: -0.61037 (-0.61193) | > loss_dur: 0.13504 (0.11557) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.35698 (9.55631) | > current_lr: 0.00027 | > step_time: 1.05030 (0.62390) | > loader_time: 0.00390 (0.00341)  --> STEP: 550/639 -- GLOBAL_STEP: 54950 | > loss: -0.43203 (-0.49484) | > log_mle: -0.54417 (-0.61045) | > loss_dur: 0.11214 (0.11561) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.45077 (9.45030) | > current_lr: 0.00027 | > step_time: 0.93010 (0.63863) | > loader_time: 0.00410 (0.00344)  --> STEP: 575/639 -- GLOBAL_STEP: 54975 | > loss: -0.44935 (-0.49330) | > log_mle: -0.56175 (-0.60902) | > loss_dur: 0.11241 (0.11572) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.61125 (9.33330) | > current_lr: 0.00027 | > step_time: 0.89530 (0.65224) | > loader_time: 0.00410 (0.00347)  --> STEP: 600/639 -- GLOBAL_STEP: 55000 | > loss: -0.43854 (-0.49207) | > log_mle: -0.56201 (-0.60799) | > loss_dur: 0.12347 (0.11592) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.53551 (9.27305) | > current_lr: 0.00027 | > step_time: 1.03580 (0.66868) | > loader_time: 0.00430 (0.00351)  --> STEP: 625/639 -- GLOBAL_STEP: 55025 | > loss: -0.47173 (-0.49095) | > log_mle: -0.58895 (-0.60693) | > loss_dur: 0.11722 (0.11598) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.32553 (9.19955) | > current_lr: 0.00027 | > step_time: 1.08970 (0.68569) | > loader_time: 0.00400 (0.00355) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02975 (+0.00728) | > avg_loss: -0.45514 (+0.00028) | > avg_log_mle: -0.59053 (-0.00075) | > avg_loss_dur: 0.13540 (+0.00103)  > EPOCH: 86/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 01:50:34)   --> STEP: 10/639 -- GLOBAL_STEP: 55050 | > loss: -0.53549 (-0.54522) | > log_mle: -0.66107 (-0.66739) | > loss_dur: 0.12558 (0.12217) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.71746 (11.30459) | > current_lr: 0.00027 | > step_time: 0.28430 (0.28605) | > loader_time: 0.00230 (0.00250)  --> STEP: 35/639 -- GLOBAL_STEP: 55075 | > loss: -0.52890 (-0.53601) | > log_mle: -0.62542 (-0.65103) | > loss_dur: 0.09651 (0.11502) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.26689 (10.62631) | > current_lr: 0.00027 | > step_time: 0.32220 (0.32169) | > loader_time: 0.00280 (0.00257)  --> STEP: 60/639 -- GLOBAL_STEP: 55100 | > loss: -0.48733 (-0.52837) | > log_mle: -0.58985 (-0.64318) | > loss_dur: 0.10253 (0.11481) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.01846 (9.89299) | > current_lr: 0.00027 | > step_time: 0.34170 (0.35180) | > loader_time: 0.00240 (0.00261)  --> STEP: 85/639 -- GLOBAL_STEP: 55125 | > loss: -0.48732 (-0.52122) | > log_mle: -0.58955 (-0.63434) | > loss_dur: 0.10223 (0.11312) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.21420 (9.11220) | > current_lr: 0.00027 | > step_time: 0.40870 (0.36950) | > loader_time: 0.00280 (0.00268)  --> STEP: 110/639 -- GLOBAL_STEP: 55150 | > loss: -0.48244 (-0.52190) | > log_mle: -0.60182 (-0.63621) | > loss_dur: 0.11938 (0.11431) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.53093 (9.67441) | > current_lr: 0.00027 | > step_time: 0.39770 (0.39092) | > loader_time: 0.00290 (0.00273)  --> STEP: 135/639 -- GLOBAL_STEP: 55175 | > loss: -0.46385 (-0.51968) | > log_mle: -0.58191 (-0.63368) | > loss_dur: 0.11806 (0.11400) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.72765 (9.59283) | > current_lr: 0.00027 | > step_time: 0.46600 (0.40347) | > loader_time: 0.00290 (0.00276)  --> STEP: 160/639 -- GLOBAL_STEP: 55200 | > loss: -0.51851 (-0.51780) | > log_mle: -0.62193 (-0.63152) | > loss_dur: 0.10342 (0.11372) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.28713 (9.59474) | > current_lr: 0.00027 | > step_time: 0.55370 (0.42023) | > loader_time: 0.00290 (0.00281)  --> STEP: 185/639 -- GLOBAL_STEP: 55225 | > loss: -0.49120 (-0.51885) | > log_mle: -0.60525 (-0.63329) | > loss_dur: 0.11405 (0.11444) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.83626 (10.02006) | > current_lr: 0.00027 | > step_time: 0.63830 (0.43816) | > loader_time: 0.00320 (0.00287)  --> STEP: 210/639 -- GLOBAL_STEP: 55250 | > loss: -0.42727 (-0.51607) | > log_mle: -0.52471 (-0.63018) | > loss_dur: 0.09745 (0.11411) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.61511 (9.84643) | > current_lr: 0.00027 | > step_time: 0.51130 (0.45553) | > loader_time: 0.00290 (0.00291)  --> STEP: 235/639 -- GLOBAL_STEP: 55275 | > loss: -0.50891 (-0.51646) | > log_mle: -0.61807 (-0.63037) | > loss_dur: 0.10915 (0.11391) | > amp_scaler: 8192.00000 (4427.16596) | > grad_norm: 8.09972 (9.89081) | > current_lr: 0.00027 | > step_time: 0.56060 (0.47107) | > loader_time: 0.00360 (0.00296)  --> STEP: 260/639 -- GLOBAL_STEP: 55300 | > loss: -0.49712 (-0.51575) | > log_mle: -0.61993 (-0.62996) | > loss_dur: 0.12281 (0.11421) | > amp_scaler: 8192.00000 (4789.16923) | > grad_norm: 8.46310 (9.98407) | > current_lr: 0.00027 | > step_time: 0.57270 (0.48594) | > loader_time: 0.00350 (0.00299)  --> STEP: 285/639 -- GLOBAL_STEP: 55325 | > loss: -0.45130 (-0.51432) | > log_mle: -0.56269 (-0.62885) | > loss_dur: 0.11139 (0.11453) | > amp_scaler: 8192.00000 (5087.66316) | > grad_norm: 4.22382 (9.95745) | > current_lr: 0.00027 | > step_time: 0.61480 (0.50168) | > loader_time: 0.00350 (0.00304)  --> STEP: 310/639 -- GLOBAL_STEP: 55350 | > loss: -0.49210 (-0.51246) | > log_mle: -0.60520 (-0.62719) | > loss_dur: 0.11310 (0.11473) | > amp_scaler: 8192.00000 (5338.01290) | > grad_norm: 9.34476 (9.94892) | > current_lr: 0.00027 | > step_time: 0.57420 (0.51510) | > loader_time: 0.00330 (0.00308)  --> STEP: 335/639 -- GLOBAL_STEP: 55375 | > loss: -0.50269 (-0.50992) | > log_mle: -0.63002 (-0.62512) | > loss_dur: 0.12733 (0.11520) | > amp_scaler: 8192.00000 (5550.99701) | > grad_norm: 11.17455 (9.86010) | > current_lr: 0.00027 | > step_time: 0.65710 (0.52964) | > loader_time: 0.00340 (0.00311)  --> STEP: 360/639 -- GLOBAL_STEP: 55400 | > loss: -0.45549 (-0.50823) | > log_mle: -0.56862 (-0.62322) | > loss_dur: 0.11313 (0.11499) | > amp_scaler: 8192.00000 (5734.40000) | > grad_norm: 5.66572 (9.71698) | > current_lr: 0.00027 | > step_time: 0.67570 (0.54192) | > loader_time: 0.00370 (0.00314)  --> STEP: 385/639 -- GLOBAL_STEP: 55425 | > loss: -0.48751 (-0.50656) | > log_mle: -0.60157 (-0.62143) | > loss_dur: 0.11406 (0.11487) | > amp_scaler: 8192.00000 (5893.98442) | > grad_norm: 8.36519 (9.62689) | > current_lr: 0.00027 | > step_time: 0.68410 (0.55514) | > loader_time: 0.00360 (0.00317)  --> STEP: 410/639 -- GLOBAL_STEP: 55450 | > loss: -0.44868 (-0.50462) | > log_mle: -0.57664 (-0.61976) | > loss_dur: 0.12796 (0.11514) | > amp_scaler: 8192.00000 (6034.10732) | > grad_norm: 6.62480 (9.51637) | > current_lr: 0.00027 | > step_time: 0.80710 (0.56605) | > loader_time: 0.00370 (0.00320)  --> STEP: 435/639 -- GLOBAL_STEP: 55475 | > loss: -0.48129 (-0.50262) | > log_mle: -0.58508 (-0.61771) | > loss_dur: 0.10379 (0.11509) | > amp_scaler: 8192.00000 (6158.12414) | > grad_norm: 5.62959 (9.36492) | > current_lr: 0.00027 | > step_time: 0.67140 (0.57828) | > loader_time: 0.00320 (0.00322)  --> STEP: 460/639 -- GLOBAL_STEP: 55500 | > loss: -0.43523 (-0.50069) | > log_mle: -0.55126 (-0.61573) | > loss_dur: 0.11603 (0.11503) | > amp_scaler: 8192.00000 (6268.66087) | > grad_norm: 2.73919 (9.23659) | > current_lr: 0.00027 | > step_time: 0.98230 (0.59049) | > loader_time: 0.00390 (0.00325)  --> STEP: 485/639 -- GLOBAL_STEP: 55525 | > loss: -0.47578 (-0.49909) | > log_mle: -0.58559 (-0.61429) | > loss_dur: 0.10981 (0.11520) | > amp_scaler: 8192.00000 (6367.80206) | > grad_norm: 6.28074 (9.15144) | > current_lr: 0.00027 | > step_time: 0.86650 (0.60287) | > loader_time: 0.00370 (0.00327)  --> STEP: 510/639 -- GLOBAL_STEP: 55550 | > loss: -0.46645 (-0.49784) | > log_mle: -0.58801 (-0.61299) | > loss_dur: 0.12156 (0.11516) | > amp_scaler: 4096.00000 (6312.65882) | > grad_norm: 7.16560 (9.08336) | > current_lr: 0.00027 | > step_time: 0.85960 (0.61600) | > loader_time: 0.00400 (0.00330)  --> STEP: 535/639 -- GLOBAL_STEP: 55575 | > loss: -0.45384 (-0.49629) | > log_mle: -0.57795 (-0.61160) | > loss_dur: 0.12411 (0.11531) | > amp_scaler: 4096.00000 (6209.07664) | > grad_norm: 7.58307 (9.01549) | > current_lr: 0.00027 | > step_time: 0.87150 (0.62911) | > loader_time: 0.00400 (0.00334)  --> STEP: 560/639 -- GLOBAL_STEP: 55600 | > loss: -0.45626 (-0.49480) | > log_mle: -0.58627 (-0.61024) | > loss_dur: 0.13001 (0.11544) | > amp_scaler: 4096.00000 (6114.74286) | > grad_norm: 7.13914 (8.93996) | > current_lr: 0.00027 | > step_time: 0.89780 (0.64397) | > loader_time: 0.00400 (0.00337)  --> STEP: 585/639 -- GLOBAL_STEP: 55625 | > loss: -0.47086 (-0.49354) | > log_mle: -0.58720 (-0.60902) | > loss_dur: 0.11634 (0.11548) | > amp_scaler: 4096.00000 (6028.47179) | > grad_norm: 7.60443 (8.86494) | > current_lr: 0.00027 | > step_time: 1.02660 (0.65905) | > loader_time: 0.00430 (0.00341)  --> STEP: 610/639 -- GLOBAL_STEP: 55650 | > loss: -0.44334 (-0.49224) | > log_mle: -0.55804 (-0.60790) | > loss_dur: 0.11470 (0.11566) | > amp_scaler: 4096.00000 (5949.27213) | > grad_norm: 5.24998 (8.82603) | > current_lr: 0.00027 | > step_time: 1.05040 (0.67532) | > loader_time: 0.00420 (0.00345)  --> STEP: 635/639 -- GLOBAL_STEP: 55675 | > loss: -0.46167 (-0.49125) | > log_mle: -0.58328 (-0.60718) | > loss_dur: 0.12161 (0.11592) | > amp_scaler: 4096.00000 (5876.30866) | > grad_norm: 7.50348 (8.82199) | > current_lr: 0.00027 | > step_time: 1.15120 (0.69197) | > loader_time: 0.00410 (0.00348) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03405 (+0.00430) | > avg_loss: -0.44928 (+0.00586) | > avg_log_mle: -0.58507 (+0.00547) | > avg_loss_dur: 0.13579 (+0.00039)  > EPOCH: 87/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 01:58:12)   --> STEP: 20/639 -- GLOBAL_STEP: 55700 | > loss: -0.64759 (-0.55590) | > log_mle: -0.79356 (-0.67684) | > loss_dur: 0.14597 (0.12093) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 19.02361 (13.58968) | > current_lr: 0.00027 | > step_time: 0.34460 (0.29930) | > loader_time: 0.00240 (0.00257)  --> STEP: 45/639 -- GLOBAL_STEP: 55725 | > loss: -0.49918 (-0.53797) | > log_mle: -0.59826 (-0.65431) | > loss_dur: 0.09908 (0.11634) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.54751 (10.95478) | > current_lr: 0.00027 | > step_time: 0.40260 (0.33678) | > loader_time: 0.00300 (0.00269)  --> STEP: 70/639 -- GLOBAL_STEP: 55750 | > loss: -0.52728 (-0.52816) | > log_mle: -0.64710 (-0.64389) | > loss_dur: 0.11982 (0.11573) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.60773 (10.07922) | > current_lr: 0.00027 | > step_time: 0.47130 (0.36498) | > loader_time: 0.00300 (0.00277)  --> STEP: 95/639 -- GLOBAL_STEP: 55775 | > loss: -0.58816 (-0.52507) | > log_mle: -0.69228 (-0.63805) | > loss_dur: 0.10412 (0.11299) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 25.91238 (9.86681) | > current_lr: 0.00027 | > step_time: 0.40440 (0.38223) | > loader_time: 0.00260 (0.00281)  --> STEP: 120/639 -- GLOBAL_STEP: 55800 | > loss: -0.55971 (-0.52229) | > log_mle: -0.69255 (-0.63602) | > loss_dur: 0.13284 (0.11373) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 20.92472 (10.07062) | > current_lr: 0.00027 | > step_time: 0.52150 (0.39996) | > loader_time: 0.00310 (0.00286)  --> STEP: 145/639 -- GLOBAL_STEP: 55825 | > loss: -0.51091 (-0.52140) | > log_mle: -0.64099 (-0.63587) | > loss_dur: 0.13009 (0.11447) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.71619 (10.29466) | > current_lr: 0.00027 | > step_time: 0.47140 (0.41420) | > loader_time: 0.00280 (0.00290)  --> STEP: 170/639 -- GLOBAL_STEP: 55850 | > loss: -0.51006 (-0.52006) | > log_mle: -0.62459 (-0.63447) | > loss_dur: 0.11453 (0.11441) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.62848 (10.34264) | > current_lr: 0.00027 | > step_time: 0.56710 (0.42904) | > loader_time: 0.00340 (0.00293)  --> STEP: 195/639 -- GLOBAL_STEP: 55875 | > loss: -0.47595 (-0.51905) | > log_mle: -0.56911 (-0.63301) | > loss_dur: 0.09316 (0.11396) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.93166 (10.34409) | > current_lr: 0.00027 | > step_time: 0.51380 (0.44423) | > loader_time: 0.00280 (0.00296)  --> STEP: 220/639 -- GLOBAL_STEP: 55900 | > loss: -0.51268 (-0.51639) | > log_mle: -0.61840 (-0.63043) | > loss_dur: 0.10572 (0.11404) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.31062 (10.33990) | > current_lr: 0.00027 | > step_time: 0.64780 (0.45973) | > loader_time: 0.00330 (0.00299)  --> STEP: 245/639 -- GLOBAL_STEP: 55925 | > loss: -0.50138 (-0.51706) | > log_mle: -0.60740 (-0.63110) | > loss_dur: 0.10602 (0.11405) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.50516 (10.54476) | > current_lr: 0.00027 | > step_time: 0.52180 (0.47339) | > loader_time: 0.00330 (0.00302)  --> STEP: 270/639 -- GLOBAL_STEP: 55950 | > loss: -0.54349 (-0.51628) | > log_mle: -0.67455 (-0.63084) | > loss_dur: 0.13106 (0.11456) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.21947 (10.88127) | > current_lr: 0.00027 | > step_time: 0.84750 (0.49120) | > loader_time: 0.00390 (0.00306)  --> STEP: 295/639 -- GLOBAL_STEP: 55975 | > loss: -0.47079 (-0.51402) | > log_mle: -0.56739 (-0.62877) | > loss_dur: 0.09659 (0.11475) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.80751 (10.75187) | > current_lr: 0.00027 | > step_time: 0.60710 (0.50532) | > loader_time: 0.00360 (0.00310)  --> STEP: 320/639 -- GLOBAL_STEP: 56000 | > loss: -0.48184 (-0.51225) | > log_mle: -0.61443 (-0.62734) | > loss_dur: 0.13259 (0.11509) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.54485 (10.81056) | > current_lr: 0.00027 | > step_time: 0.73380 (0.52117) | > loader_time: 0.00360 (0.00314)  --> STEP: 345/639 -- GLOBAL_STEP: 56025 | > loss: -0.48707 (-0.51005) | > log_mle: -0.59970 (-0.62540) | > loss_dur: 0.11264 (0.11535) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.79084 (10.75900) | > current_lr: 0.00027 | > step_time: 0.64250 (0.53555) | > loader_time: 0.00360 (0.00317)  --> STEP: 370/639 -- GLOBAL_STEP: 56050 | > loss: -0.50144 (-0.50819) | > log_mle: -0.61260 (-0.62325) | > loss_dur: 0.11116 (0.11506) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.72890 (10.60591) | > current_lr: 0.00027 | > step_time: 0.87660 (0.54776) | > loader_time: 0.00390 (0.00320)  --> STEP: 395/639 -- GLOBAL_STEP: 56075 | > loss: -0.46249 (-0.50615) | > log_mle: -0.58051 (-0.62122) | > loss_dur: 0.11802 (0.11508) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.54873 (10.46052) | > current_lr: 0.00027 | > step_time: 0.70520 (0.55960) | > loader_time: 0.00350 (0.00322)  --> STEP: 420/639 -- GLOBAL_STEP: 56100 | > loss: -0.46012 (-0.50396) | > log_mle: -0.57494 (-0.61930) | > loss_dur: 0.11482 (0.11534) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.88495 (10.33895) | > current_lr: 0.00027 | > step_time: 0.71310 (0.57272) | > loader_time: 0.00350 (0.00325)  --> STEP: 445/639 -- GLOBAL_STEP: 56125 | > loss: -0.47471 (-0.50234) | > log_mle: -0.58341 (-0.61772) | > loss_dur: 0.10870 (0.11538) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.90452 (10.23155) | > current_lr: 0.00027 | > step_time: 0.71230 (0.58528) | > loader_time: 0.00350 (0.00328)  --> STEP: 470/639 -- GLOBAL_STEP: 56150 | > loss: -0.48742 (-0.50046) | > log_mle: -0.59707 (-0.61580) | > loss_dur: 0.10965 (0.11534) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.44410 (10.08323) | > current_lr: 0.00027 | > step_time: 0.76870 (0.59681) | > loader_time: 0.00370 (0.00330)  --> STEP: 495/639 -- GLOBAL_STEP: 56175 | > loss: -0.42974 (-0.49884) | > log_mle: -0.53024 (-0.61433) | > loss_dur: 0.10050 (0.11549) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.93168 (10.01915) | > current_lr: 0.00027 | > step_time: 0.87620 (0.61006) | > loader_time: 0.00360 (0.00333)  --> STEP: 520/639 -- GLOBAL_STEP: 56200 | > loss: -0.47846 (-0.49761) | > log_mle: -0.58717 (-0.61310) | > loss_dur: 0.10871 (0.11549) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.90651 (9.92059) | > current_lr: 0.00027 | > step_time: 0.87650 (0.62326) | > loader_time: 0.00330 (0.00336)  --> STEP: 545/639 -- GLOBAL_STEP: 56225 | > loss: -0.48820 (-0.49617) | > log_mle: -0.61989 (-0.61180) | > loss_dur: 0.13170 (0.11563) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.96010 (9.81011) | > current_lr: 0.00027 | > step_time: 0.90800 (0.63776) | > loader_time: 0.00420 (0.00340)  --> STEP: 570/639 -- GLOBAL_STEP: 56250 | > loss: -0.46187 (-0.49465) | > log_mle: -0.56775 (-0.61031) | > loss_dur: 0.10588 (0.11566) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.47769 (9.74521) | > current_lr: 0.00027 | > step_time: 0.86280 (0.65132) | > loader_time: 0.00400 (0.00343)  --> STEP: 595/639 -- GLOBAL_STEP: 56275 | > loss: -0.45244 (-0.49355) | > log_mle: -0.56050 (-0.60933) | > loss_dur: 0.10806 (0.11579) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.88900 (9.68624) | > current_lr: 0.00027 | > step_time: 0.94630 (0.66721) | > loader_time: 0.00410 (0.00346)  --> STEP: 620/639 -- GLOBAL_STEP: 56300 | > loss: -0.45929 (-0.49229) | > log_mle: -0.56944 (-0.60816) | > loss_dur: 0.11015 (0.11588) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.24934 (9.62519) | > current_lr: 0.00027 | > step_time: 1.18120 (0.68373) | > loader_time: 0.00470 (0.00350) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02745 (-0.00660) | > avg_loss: -0.45133 (-0.00205) | > avg_log_mle: -0.58596 (-0.00090) | > avg_loss_dur: 0.13463 (-0.00116)  > EPOCH: 88/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 02:05:51)   --> STEP: 5/639 -- GLOBAL_STEP: 56325 | > loss: -0.54644 (-0.55472) | > log_mle: -0.69560 (-0.68019) | > loss_dur: 0.14917 (0.12547) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 21.98972 (12.63037) | > current_lr: 0.00027 | > step_time: 0.30700 (0.27618) | > loader_time: 0.00140 (0.00217)  --> STEP: 30/639 -- GLOBAL_STEP: 56350 | > loss: -0.54285 (-0.54331) | > log_mle: -0.65614 (-0.65959) | > loss_dur: 0.11329 (0.11627) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.96490 (11.23072) | > current_lr: 0.00027 | > step_time: 0.32080 (0.32684) | > loader_time: 0.00270 (0.00261)  --> STEP: 55/639 -- GLOBAL_STEP: 56375 | > loss: -0.49623 (-0.53467) | > log_mle: -0.59089 (-0.64876) | > loss_dur: 0.09466 (0.11409) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.61261 (10.26232) | > current_lr: 0.00027 | > step_time: 0.32950 (0.34666) | > loader_time: 0.00290 (0.00271)  --> STEP: 80/639 -- GLOBAL_STEP: 56400 | > loss: -0.49354 (-0.52738) | > log_mle: -0.62763 (-0.63995) | > loss_dur: 0.13409 (0.11257) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.41672 (9.86827) | > current_lr: 0.00027 | > step_time: 0.39920 (0.36990) | > loader_time: 0.00300 (0.00279)  --> STEP: 105/639 -- GLOBAL_STEP: 56425 | > loss: -0.46958 (-0.52514) | > log_mle: -0.57835 (-0.63710) | > loss_dur: 0.10878 (0.11196) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.79983 (9.85533) | > current_lr: 0.00027 | > step_time: 0.42560 (0.38688) | > loader_time: 0.00300 (0.00287)  --> STEP: 130/639 -- GLOBAL_STEP: 56450 | > loss: -0.49519 (-0.52227) | > log_mle: -0.61666 (-0.63512) | > loss_dur: 0.12147 (0.11285) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.08797 (10.17640) | > current_lr: 0.00027 | > step_time: 0.49100 (0.40277) | > loader_time: 0.00320 (0.00292)  --> STEP: 155/639 -- GLOBAL_STEP: 56475 | > loss: -0.49770 (-0.52048) | > log_mle: -0.61159 (-0.63386) | > loss_dur: 0.11389 (0.11338) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.67158 (10.15572) | > current_lr: 0.00027 | > step_time: 0.56200 (0.41883) | > loader_time: 0.00320 (0.00296)  --> STEP: 180/639 -- GLOBAL_STEP: 56500 | > loss: -0.55493 (-0.52185) | > log_mle: -0.66737 (-0.63532) | > loss_dur: 0.11244 (0.11347) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.19673 (10.42409) | > current_lr: 0.00027 | > step_time: 0.57960 (0.43371) | > loader_time: 0.00350 (0.00300)  --> STEP: 205/639 -- GLOBAL_STEP: 56525 | > loss: -0.46562 (-0.51859) | > log_mle: -0.58436 (-0.63161) | > loss_dur: 0.11875 (0.11302) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.04252 (10.19763) | > current_lr: 0.00027 | > step_time: 0.60890 (0.44954) | > loader_time: 0.00340 (0.00303)  --> STEP: 230/639 -- GLOBAL_STEP: 56550 | > loss: -0.51420 (-0.51801) | > log_mle: -0.62221 (-0.63133) | > loss_dur: 0.10801 (0.11333) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.02177 (10.36784) | > current_lr: 0.00027 | > step_time: 0.55240 (0.46654) | > loader_time: 0.00360 (0.00306)  --> STEP: 255/639 -- GLOBAL_STEP: 56575 | > loss: -0.47067 (-0.51670) | > log_mle: -0.58773 (-0.63017) | > loss_dur: 0.11706 (0.11347) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.75247 (10.30614) | > current_lr: 0.00027 | > step_time: 0.63790 (0.48003) | > loader_time: 0.00350 (0.00310)  --> STEP: 280/639 -- GLOBAL_STEP: 56600 | > loss: -0.49796 (-0.51625) | > log_mle: -0.61755 (-0.63026) | > loss_dur: 0.11959 (0.11400) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.65150 (10.43980) | > current_lr: 0.00027 | > step_time: 0.57440 (0.49672) | > loader_time: 0.00360 (0.00314)  --> STEP: 305/639 -- GLOBAL_STEP: 56625 | > loss: -0.46982 (-0.51419) | > log_mle: -0.59488 (-0.62849) | > loss_dur: 0.12506 (0.11431) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.85061 (10.33183) | > current_lr: 0.00027 | > step_time: 0.63410 (0.51018) | > loader_time: 0.00370 (0.00317)  --> STEP: 330/639 -- GLOBAL_STEP: 56650 | > loss: -0.48945 (-0.51157) | > log_mle: -0.61424 (-0.62643) | > loss_dur: 0.12479 (0.11486) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.79965 (10.26056) | > current_lr: 0.00027 | > step_time: 0.72930 (0.52420) | > loader_time: 0.00360 (0.00321)  --> STEP: 355/639 -- GLOBAL_STEP: 56675 | > loss: -0.44265 (-0.50971) | > log_mle: -0.54734 (-0.62446) | > loss_dur: 0.10470 (0.11476) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.71510 (10.13540) | > current_lr: 0.00027 | > step_time: 0.60430 (0.53766) | > loader_time: 0.00330 (0.00323)  --> STEP: 380/639 -- GLOBAL_STEP: 56700 | > loss: -0.48360 (-0.50759) | > log_mle: -0.58233 (-0.62220) | > loss_dur: 0.09873 (0.11462) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.08245 (10.00036) | > current_lr: 0.00027 | > step_time: 0.62370 (0.55142) | > loader_time: 0.00280 (0.00326)  --> STEP: 405/639 -- GLOBAL_STEP: 56725 | > loss: -0.47537 (-0.50604) | > log_mle: -0.58813 (-0.62095) | > loss_dur: 0.11276 (0.11491) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.40811 (9.92064) | > current_lr: 0.00027 | > step_time: 0.68640 (0.56328) | > loader_time: 0.00350 (0.00328)  --> STEP: 430/639 -- GLOBAL_STEP: 56750 | > loss: -0.44546 (-0.50398) | > log_mle: -0.54856 (-0.61905) | > loss_dur: 0.10310 (0.11507) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.29651 (9.80169) | > current_lr: 0.00027 | > step_time: 0.77600 (0.57646) | > loader_time: 0.00360 (0.00331)  --> STEP: 455/639 -- GLOBAL_STEP: 56775 | > loss: -0.45427 (-0.50208) | > log_mle: -0.57256 (-0.61706) | > loss_dur: 0.11829 (0.11498) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.62700 (9.65516) | > current_lr: 0.00027 | > step_time: 0.77100 (0.58813) | > loader_time: 0.00320 (0.00332)  --> STEP: 480/639 -- GLOBAL_STEP: 56800 | > loss: -0.41513 (-0.50030) | > log_mle: -0.52957 (-0.61548) | > loss_dur: 0.11443 (0.11518) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.72217 (9.56696) | > current_lr: 0.00027 | > step_time: 0.79730 (0.60075) | > loader_time: 0.00350 (0.00334)  --> STEP: 505/639 -- GLOBAL_STEP: 56825 | > loss: -0.48683 (-0.49889) | > log_mle: -0.59778 (-0.61404) | > loss_dur: 0.11094 (0.11515) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.23786 (9.48248) | > current_lr: 0.00027 | > step_time: 0.91390 (0.61272) | > loader_time: 0.00380 (0.00337)  --> STEP: 530/639 -- GLOBAL_STEP: 56850 | > loss: -0.45854 (-0.49750) | > log_mle: -0.58468 (-0.61279) | > loss_dur: 0.12614 (0.11529) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.90159 (9.47886) | > current_lr: 0.00027 | > step_time: 0.92830 (0.62634) | > loader_time: 0.00370 (0.00339)  --> STEP: 555/639 -- GLOBAL_STEP: 56875 | > loss: -0.45970 (-0.49599) | > log_mle: -0.57007 (-0.61127) | > loss_dur: 0.11038 (0.11528) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.31809 (9.36467) | > current_lr: 0.00027 | > step_time: 0.82200 (0.64067) | > loader_time: 0.00380 (0.00343)  --> STEP: 580/639 -- GLOBAL_STEP: 56900 | > loss: -0.48241 (-0.49471) | > log_mle: -0.58783 (-0.61010) | > loss_dur: 0.10542 (0.11539) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.08377 (9.28343) | > current_lr: 0.00027 | > step_time: 0.99800 (0.65569) | > loader_time: 0.00450 (0.00346)  --> STEP: 605/639 -- GLOBAL_STEP: 56925 | > loss: -0.45048 (-0.49345) | > log_mle: -0.55725 (-0.60901) | > loss_dur: 0.10677 (0.11556) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.83533 (9.23907) | > current_lr: 0.00027 | > step_time: 1.13410 (0.67149) | > loader_time: 0.00440 (0.00350)  --> STEP: 630/639 -- GLOBAL_STEP: 56950 | > loss: -0.46117 (-0.49232) | > log_mle: -0.58957 (-0.60809) | > loss_dur: 0.12839 (0.11577) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.66555 (9.24642) | > current_lr: 0.00027 | > step_time: 1.20740 (0.68872) | > loader_time: 0.00470 (0.00354) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02862 (+0.00117) | > avg_loss: -0.44947 (+0.00186) | > avg_log_mle: -0.58371 (+0.00225) | > avg_loss_dur: 0.13424 (-0.00039)  > EPOCH: 89/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 02:13:29)   --> STEP: 15/639 -- GLOBAL_STEP: 56975 | > loss: -0.66183 (-0.55884) | > log_mle: -0.78624 (-0.67438) | > loss_dur: 0.12441 (0.11554) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 30.16094 (12.11302) | > current_lr: 0.00027 | > step_time: 0.32910 (0.29295) | > loader_time: 0.00250 (0.00250)  --> STEP: 40/639 -- GLOBAL_STEP: 57000 | > loss: -0.50007 (-0.53793) | > log_mle: -0.62298 (-0.65272) | > loss_dur: 0.12290 (0.11479) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.31051 (11.65226) | > current_lr: 0.00026 | > step_time: 0.35260 (0.33044) | > loader_time: 0.00280 (0.00264)  --> STEP: 65/639 -- GLOBAL_STEP: 57025 | > loss: -0.51965 (-0.53055) | > log_mle: -0.61828 (-0.64567) | > loss_dur: 0.09863 (0.11512) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.23197 (11.43541) | > current_lr: 0.00026 | > step_time: 0.37700 (0.36124) | > loader_time: 0.00300 (0.00274)  --> STEP: 90/639 -- GLOBAL_STEP: 57050 | > loss: -0.46269 (-0.52352) | > log_mle: -0.57160 (-0.63685) | > loss_dur: 0.10891 (0.11334) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.58440 (10.53556) | > current_lr: 0.00026 | > step_time: 0.52620 (0.37785) | > loader_time: 0.00330 (0.00282)  --> STEP: 115/639 -- GLOBAL_STEP: 57075 | > loss: -0.44115 (-0.52345) | > log_mle: -0.54435 (-0.63698) | > loss_dur: 0.10320 (0.11353) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.66066 (10.95823) | > current_lr: 0.00026 | > step_time: 0.43150 (0.39391) | > loader_time: 0.00290 (0.00285)  --> STEP: 140/639 -- GLOBAL_STEP: 57100 | > loss: -0.48810 (-0.52304) | > log_mle: -0.62625 (-0.63708) | > loss_dur: 0.13814 (0.11404) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.45641 (11.18019) | > current_lr: 0.00026 | > step_time: 0.55000 (0.40801) | > loader_time: 0.00320 (0.00288)  --> STEP: 165/639 -- GLOBAL_STEP: 57125 | > loss: -0.50558 (-0.52205) | > log_mle: -0.63012 (-0.63581) | > loss_dur: 0.12455 (0.11376) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.24369 (11.24759) | > current_lr: 0.00026 | > step_time: 0.50660 (0.42257) | > loader_time: 0.00290 (0.00290)  --> STEP: 190/639 -- GLOBAL_STEP: 57150 | > loss: -0.44652 (-0.52148) | > log_mle: -0.55896 (-0.63518) | > loss_dur: 0.11244 (0.11370) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.58684 (11.31219) | > current_lr: 0.00026 | > step_time: 0.53730 (0.44059) | > loader_time: 0.00310 (0.00294)  --> STEP: 215/639 -- GLOBAL_STEP: 57175 | > loss: -0.55483 (-0.51895) | > log_mle: -0.66678 (-0.63284) | > loss_dur: 0.11195 (0.11389) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.18572 (11.27800) | > current_lr: 0.00026 | > step_time: 0.53920 (0.45559) | > loader_time: 0.00310 (0.00299)  --> STEP: 240/639 -- GLOBAL_STEP: 57200 | > loss: -0.55089 (-0.51952) | > log_mle: -0.67579 (-0.63329) | > loss_dur: 0.12490 (0.11376) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.59580 (11.31816) | > current_lr: 0.00026 | > step_time: 0.55830 (0.47016) | > loader_time: 0.00290 (0.00302)  --> STEP: 265/639 -- GLOBAL_STEP: 57225 | > loss: -0.56049 (-0.51783) | > log_mle: -0.69291 (-0.63218) | > loss_dur: 0.13242 (0.11435) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.02725 (11.38144) | > current_lr: 0.00026 | > step_time: 0.67070 (0.48704) | > loader_time: 0.00340 (0.00306)  --> STEP: 290/639 -- GLOBAL_STEP: 57250 | > loss: -0.48175 (-0.51598) | > log_mle: -0.59277 (-0.63057) | > loss_dur: 0.11102 (0.11459) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.59170 (11.27528) | > current_lr: 0.00026 | > step_time: 0.62570 (0.50213) | > loader_time: 0.00340 (0.00309)  --> STEP: 315/639 -- GLOBAL_STEP: 57275 | > loss: -0.46984 (-0.51428) | > log_mle: -0.57352 (-0.62901) | > loss_dur: 0.10368 (0.11473) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.63994 (11.18932) | > current_lr: 0.00026 | > step_time: 0.68720 (0.51622) | > loader_time: 0.00310 (0.00312)  --> STEP: 340/639 -- GLOBAL_STEP: 57300 | > loss: -0.48126 (-0.51230) | > log_mle: -0.59376 (-0.62758) | > loss_dur: 0.11250 (0.11528) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.06361 (11.10149) | > current_lr: 0.00026 | > step_time: 0.60540 (0.53163) | > loader_time: 0.00340 (0.00315)  --> STEP: 365/639 -- GLOBAL_STEP: 57325 | > loss: -0.48729 (-0.51007) | > log_mle: -0.60177 (-0.62510) | > loss_dur: 0.11448 (0.11503) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.57940 (10.96530) | > current_lr: 0.00026 | > step_time: 0.94250 (0.54404) | > loader_time: 0.00390 (0.00317)  --> STEP: 390/639 -- GLOBAL_STEP: 57350 | > loss: -0.46410 (-0.50803) | > log_mle: -0.59280 (-0.62304) | > loss_dur: 0.12870 (0.11501) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.48384 (10.75419) | > current_lr: 0.00026 | > step_time: 0.71030 (0.55744) | > loader_time: 0.00360 (0.00320)  --> STEP: 415/639 -- GLOBAL_STEP: 57375 | > loss: -0.45336 (-0.50604) | > log_mle: -0.57280 (-0.62128) | > loss_dur: 0.11944 (0.11524) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.67410 (10.64517) | > current_lr: 0.00026 | > step_time: 0.72360 (0.57018) | > loader_time: 0.00340 (0.00322)  --> STEP: 440/639 -- GLOBAL_STEP: 57400 | > loss: -0.43970 (-0.50428) | > log_mle: -0.55446 (-0.61959) | > loss_dur: 0.11475 (0.11532) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.06855 (10.49533) | > current_lr: 0.00026 | > step_time: 0.72130 (0.58214) | > loader_time: 0.00360 (0.00325)  --> STEP: 465/639 -- GLOBAL_STEP: 57425 | > loss: -0.44902 (-0.50223) | > log_mle: -0.56126 (-0.61749) | > loss_dur: 0.11224 (0.11526) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.56641 (10.32038) | > current_lr: 0.00026 | > step_time: 0.75690 (0.59239) | > loader_time: 0.00400 (0.00327)  --> STEP: 490/639 -- GLOBAL_STEP: 57450 | > loss: -0.44088 (-0.50055) | > log_mle: -0.54460 (-0.61594) | > loss_dur: 0.10373 (0.11539) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.99660 (10.25738) | > current_lr: 0.00026 | > step_time: 0.77660 (0.60527) | > loader_time: 0.00340 (0.00329)  --> STEP: 515/639 -- GLOBAL_STEP: 57475 | > loss: -0.46183 (-0.49933) | > log_mle: -0.57300 (-0.61475) | > loss_dur: 0.11117 (0.11542) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.59306 (10.24835) | > current_lr: 0.00026 | > step_time: 1.04980 (0.61835) | > loader_time: 0.00430 (0.00332)  --> STEP: 540/639 -- GLOBAL_STEP: 57500 | > loss: -0.46844 (-0.49776) | > log_mle: -0.58794 (-0.61326) | > loss_dur: 0.11950 (0.11550) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.92035 (10.13270) | > current_lr: 0.00026 | > step_time: 0.87980 (0.63286) | > loader_time: 0.00410 (0.00335)  --> STEP: 565/639 -- GLOBAL_STEP: 57525 | > loss: -0.44156 (-0.49614) | > log_mle: -0.55116 (-0.61174) | > loss_dur: 0.10960 (0.11560) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.42777 (10.04531) | > current_lr: 0.00026 | > step_time: 1.06700 (0.64695) | > loader_time: 0.00390 (0.00339)  --> STEP: 590/639 -- GLOBAL_STEP: 57550 | > loss: -0.46787 (-0.49503) | > log_mle: -0.58438 (-0.61079) | > loss_dur: 0.11651 (0.11575) | > amp_scaler: 8192.00000 (4220.96271) | > grad_norm: 9.86873 (9.97923) | > current_lr: 0.00026 | > step_time: 0.95420 (0.66251) | > loader_time: 0.00420 (0.00343)  --> STEP: 615/639 -- GLOBAL_STEP: 57575 | > loss: -0.46776 (-0.49375) | > log_mle: -0.58921 (-0.60956) | > loss_dur: 0.12145 (0.11581) | > amp_scaler: 8192.00000 (4382.38699) | > grad_norm: 5.99551 (9.90330) | > current_lr: 0.00026 | > step_time: 1.05400 (0.67814) | > loader_time: 0.00420 (0.00348) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02719 (-0.00143) | > avg_loss: -0.45857 (-0.00910) | > avg_log_mle: -0.59267 (-0.00897) | > avg_loss_dur: 0.13410 (-0.00014) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_57600.pth.tar  > EPOCH: 90/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 02:21:14)   --> STEP: 0/639 -- GLOBAL_STEP: 57600 | > loss: -0.58561 (-0.58561) | > log_mle: -0.72756 (-0.72756) | > loss_dur: 0.14195 (0.14195) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 0.00000 (0.00000) | > current_lr: 0.00026 | > step_time: 0.75680 (0.75683) | > loader_time: 0.39150 (0.39155)  --> STEP: 25/639 -- GLOBAL_STEP: 57625 | > loss: -0.54747 (-0.54940) | > log_mle: -0.67274 (-0.66950) | > loss_dur: 0.12527 (0.12010) | > amp_scaler: 2048.00000 (3440.64000) | > grad_norm: 21.36274 (12.76617) | > current_lr: 0.00026 | > step_time: 0.39170 (0.32183) | > loader_time: 0.00290 (0.00265)  --> STEP: 50/639 -- GLOBAL_STEP: 57650 | > loss: -0.49664 (-0.53727) | > log_mle: -0.59672 (-0.65332) | > loss_dur: 0.10008 (0.11605) | > amp_scaler: 2048.00000 (2744.32000) | > grad_norm: 5.98699 (12.09711) | > current_lr: 0.00026 | > step_time: 0.38390 (0.34634) | > loader_time: 0.00260 (0.00270)  --> STEP: 75/639 -- GLOBAL_STEP: 57675 | > loss: -0.52564 (-0.52941) | > log_mle: -0.62770 (-0.64335) | > loss_dur: 0.10206 (0.11394) | > amp_scaler: 2048.00000 (2512.21333) | > grad_norm: 7.36048 (11.20692) | > current_lr: 0.00026 | > step_time: 0.42970 (0.36501) | > loader_time: 0.00290 (0.00276)  --> STEP: 100/639 -- GLOBAL_STEP: 57700 | > loss: -0.52927 (-0.52703) | > log_mle: -0.63503 (-0.63992) | > loss_dur: 0.10576 (0.11289) | > amp_scaler: 2048.00000 (2396.16000) | > grad_norm: 12.01775 (11.15074) | > current_lr: 0.00026 | > step_time: 0.50960 (0.38232) | > loader_time: 0.00320 (0.00283)  --> STEP: 125/639 -- GLOBAL_STEP: 57725 | > loss: -0.51277 (-0.52419) | > log_mle: -0.62653 (-0.63767) | > loss_dur: 0.11376 (0.11348) | > amp_scaler: 2048.00000 (2326.52800) | > grad_norm: 8.18998 (11.34392) | > current_lr: 0.00026 | > step_time: 0.43190 (0.39688) | > loader_time: 0.00300 (0.00288)  --> STEP: 150/639 -- GLOBAL_STEP: 57750 | > loss: -0.47587 (-0.52272) | > log_mle: -0.59041 (-0.63628) | > loss_dur: 0.11453 (0.11355) | > amp_scaler: 2048.00000 (2280.10667) | > grad_norm: 10.10692 (11.54821) | > current_lr: 0.00026 | > step_time: 0.42150 (0.41302) | > loader_time: 0.00290 (0.00291)  --> STEP: 175/639 -- GLOBAL_STEP: 57775 | > loss: -0.50201 (-0.52250) | > log_mle: -0.60474 (-0.63602) | > loss_dur: 0.10273 (0.11351) | > amp_scaler: 2048.00000 (2246.94857) | > grad_norm: 10.64800 (11.87361) | > current_lr: 0.00026 | > step_time: 0.53140 (0.42918) | > loader_time: 0.00350 (0.00295)  --> STEP: 200/639 -- GLOBAL_STEP: 57800 | > loss: -0.47932 (-0.52020) | > log_mle: -0.59505 (-0.63331) | > loss_dur: 0.11573 (0.11310) | > amp_scaler: 2048.00000 (2222.08000) | > grad_norm: 11.33294 (11.78655) | > current_lr: 0.00026 | > step_time: 0.52390 (0.44357) | > loader_time: 0.00320 (0.00298)  --> STEP: 225/639 -- GLOBAL_STEP: 57825 | > loss: -0.47843 (-0.51774) | > log_mle: -0.57432 (-0.63092) | > loss_dur: 0.09589 (0.11318) | > amp_scaler: 2048.00000 (2202.73778) | > grad_norm: 7.37068 (11.62958) | > current_lr: 0.00026 | > step_time: 0.51430 (0.45894) | > loader_time: 0.00300 (0.00302)  --> STEP: 250/639 -- GLOBAL_STEP: 57850 | > loss: -0.48500 (-0.51842) | > log_mle: -0.57842 (-0.63178) | > loss_dur: 0.09342 (0.11337) | > amp_scaler: 2048.00000 (2187.26400) | > grad_norm: 6.98037 (11.75466) | > current_lr: 0.00026 | > step_time: 0.57090 (0.47307) | > loader_time: 0.00310 (0.00305)  --> STEP: 275/639 -- GLOBAL_STEP: 57875 | > loss: -0.50072 (-0.51804) | > log_mle: -0.61740 (-0.63202) | > loss_dur: 0.11668 (0.11399) | > amp_scaler: 2048.00000 (2174.60364) | > grad_norm: 9.99103 (11.91695) | > current_lr: 0.00026 | > step_time: 0.93460 (0.48980) | > loader_time: 0.00350 (0.00309)  --> STEP: 300/639 -- GLOBAL_STEP: 57900 | > loss: -0.49537 (-0.51560) | > log_mle: -0.60526 (-0.62982) | > loss_dur: 0.10989 (0.11422) | > amp_scaler: 2048.00000 (2164.05333) | > grad_norm: 10.50737 (11.70429) | > current_lr: 0.00026 | > step_time: 0.74340 (0.50431) | > loader_time: 0.00360 (0.00313)  --> STEP: 325/639 -- GLOBAL_STEP: 57925 | > loss: -0.49323 (-0.51340) | > log_mle: -0.61014 (-0.62803) | > loss_dur: 0.11691 (0.11463) | > amp_scaler: 2048.00000 (2155.12615) | > grad_norm: 9.41729 (11.62428) | > current_lr: 0.00026 | > step_time: 0.59930 (0.51987) | > loader_time: 0.00340 (0.00316)  --> STEP: 350/639 -- GLOBAL_STEP: 57950 | > loss: -0.51121 (-0.51147) | > log_mle: -0.63603 (-0.62607) | > loss_dur: 0.12481 (0.11460) | > amp_scaler: 2048.00000 (2147.47429) | > grad_norm: 9.87271 (11.46814) | > current_lr: 0.00026 | > step_time: 0.84680 (0.53411) | > loader_time: 0.00360 (0.00319)  --> STEP: 375/639 -- GLOBAL_STEP: 57975 | > loss: -0.45797 (-0.50918) | > log_mle: -0.57104 (-0.62358) | > loss_dur: 0.11307 (0.11440) | > amp_scaler: 2048.00000 (2140.84267) | > grad_norm: 11.29296 (11.22717) | > current_lr: 0.00026 | > step_time: 0.73010 (0.54713) | > loader_time: 0.00350 (0.00322)  --> STEP: 400/639 -- GLOBAL_STEP: 58000 | > loss: -0.48917 (-0.50757) | > log_mle: -0.60051 (-0.62210) | > loss_dur: 0.11135 (0.11453) | > amp_scaler: 2048.00000 (2135.04000) | > grad_norm: 10.19148 (11.11963) | > current_lr: 0.00026 | > step_time: 0.73940 (0.55914) | > loader_time: 0.00390 (0.00325)  --> STEP: 425/639 -- GLOBAL_STEP: 58025 | > loss: -0.53391 (-0.50531) | > log_mle: -0.65395 (-0.61998) | > loss_dur: 0.12004 (0.11467) | > amp_scaler: 2048.00000 (2129.92000) | > grad_norm: 14.84695 (10.97351) | > current_lr: 0.00026 | > step_time: 0.89540 (0.57100) | > loader_time: 0.00380 (0.00327)  --> STEP: 450/639 -- GLOBAL_STEP: 58050 | > loss: -0.45720 (-0.50349) | > log_mle: -0.55900 (-0.61813) | > loss_dur: 0.10180 (0.11464) | > amp_scaler: 2048.00000 (2125.36889) | > grad_norm: 4.92683 (10.83251) | > current_lr: 0.00026 | > step_time: 0.78650 (0.58321) | > loader_time: 0.00390 (0.00329)  --> STEP: 475/639 -- GLOBAL_STEP: 58075 | > loss: -0.46757 (-0.50163) | > log_mle: -0.56977 (-0.61642) | > loss_dur: 0.10220 (0.11479) | > amp_scaler: 2048.00000 (2121.29684) | > grad_norm: 5.06063 (10.70974) | > current_lr: 0.00026 | > step_time: 0.72350 (0.59501) | > loader_time: 0.00320 (0.00331)  --> STEP: 500/639 -- GLOBAL_STEP: 58100 | > loss: -0.50765 (-0.50029) | > log_mle: -0.63259 (-0.61517) | > loss_dur: 0.12494 (0.11488) | > amp_scaler: 2048.00000 (2117.63200) | > grad_norm: 12.22756 (10.62815) | > current_lr: 0.00026 | > step_time: 0.94040 (0.60869) | > loader_time: 0.00400 (0.00334)  --> STEP: 525/639 -- GLOBAL_STEP: 58125 | > loss: -0.48120 (-0.49888) | > log_mle: -0.61027 (-0.61390) | > loss_dur: 0.12906 (0.11502) | > amp_scaler: 2048.00000 (2114.31619) | > grad_norm: 11.16002 (10.53820) | > current_lr: 0.00026 | > step_time: 1.06690 (0.62207) | > loader_time: 0.00430 (0.00337)  --> STEP: 550/639 -- GLOBAL_STEP: 58150 | > loss: -0.43488 (-0.49734) | > log_mle: -0.54632 (-0.61240) | > loss_dur: 0.11144 (0.11506) | > amp_scaler: 2048.00000 (2111.30182) | > grad_norm: 4.12514 (10.38784) | > current_lr: 0.00026 | > step_time: 0.97600 (0.63600) | > loader_time: 0.00430 (0.00340)  --> STEP: 575/639 -- GLOBAL_STEP: 58175 | > loss: -0.45402 (-0.49579) | > log_mle: -0.56393 (-0.61098) | > loss_dur: 0.10991 (0.11519) | > amp_scaler: 2048.00000 (2108.54957) | > grad_norm: 5.52249 (10.30220) | > current_lr: 0.00026 | > step_time: 0.90590 (0.65003) | > loader_time: 0.00420 (0.00344)  --> STEP: 600/639 -- GLOBAL_STEP: 58200 | > loss: -0.44154 (-0.49459) | > log_mle: -0.56125 (-0.60994) | > loss_dur: 0.11972 (0.11535) | > amp_scaler: 2048.00000 (2106.02667) | > grad_norm: 7.73079 (10.20953) | > current_lr: 0.00026 | > step_time: 1.04460 (0.66588) | > loader_time: 0.00440 (0.00348)  --> STEP: 625/639 -- GLOBAL_STEP: 58225 | > loss: -0.47448 (-0.49340) | > log_mle: -0.59166 (-0.60885) | > loss_dur: 0.11718 (0.11546) | > amp_scaler: 2048.00000 (2103.70560) | > grad_norm: 11.12996 (10.14669) | > current_lr: 0.00026 | > step_time: 1.08330 (0.68245) | > loader_time: 0.00460 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03571 (+0.00852) | > avg_loss: -0.45506 (+0.00351) | > avg_log_mle: -0.58914 (+0.00354) | > avg_loss_dur: 0.13408 (-0.00003)  > EPOCH: 91/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 02:28:51)   --> STEP: 10/639 -- GLOBAL_STEP: 58250 | > loss: -0.53209 (-0.54862) | > log_mle: -0.66559 (-0.67216) | > loss_dur: 0.13350 (0.12354) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.85129 (11.33054) | > current_lr: 0.00026 | > step_time: 0.30990 (0.29401) | > loader_time: 0.00260 (0.01339)  --> STEP: 35/639 -- GLOBAL_STEP: 58275 | > loss: -0.54026 (-0.54113) | > log_mle: -0.63268 (-0.65697) | > loss_dur: 0.09241 (0.11583) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.08512 (10.49730) | > current_lr: 0.00026 | > step_time: 0.32100 (0.32085) | > loader_time: 0.00270 (0.00574)  --> STEP: 60/639 -- GLOBAL_STEP: 58300 | > loss: -0.49083 (-0.53334) | > log_mle: -0.59283 (-0.64869) | > loss_dur: 0.10200 (0.11534) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.92007 (10.24996) | > current_lr: 0.00026 | > step_time: 0.34170 (0.35364) | > loader_time: 0.00280 (0.00455)  --> STEP: 85/639 -- GLOBAL_STEP: 58325 | > loss: -0.48734 (-0.52591) | > log_mle: -0.58928 (-0.63914) | > loss_dur: 0.10194 (0.11323) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.27857 (9.61939) | > current_lr: 0.00026 | > step_time: 0.41490 (0.37152) | > loader_time: 0.00300 (0.00409)  --> STEP: 110/639 -- GLOBAL_STEP: 58350 | > loss: -0.49301 (-0.52581) | > log_mle: -0.60488 (-0.63992) | > loss_dur: 0.11187 (0.11411) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.24806 (9.96211) | > current_lr: 0.00026 | > step_time: 0.40410 (0.39335) | > loader_time: 0.00310 (0.00384)  --> STEP: 135/639 -- GLOBAL_STEP: 58375 | > loss: -0.46580 (-0.52284) | > log_mle: -0.58232 (-0.63681) | > loss_dur: 0.11652 (0.11397) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.46182 (10.01839) | > current_lr: 0.00026 | > step_time: 0.46730 (0.40666) | > loader_time: 0.00300 (0.00368)  --> STEP: 160/639 -- GLOBAL_STEP: 58400 | > loss: -0.52127 (-0.52066) | > log_mle: -0.62329 (-0.63424) | > loss_dur: 0.10201 (0.11358) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.42566 (9.97704) | > current_lr: 0.00026 | > step_time: 0.46540 (0.42056) | > loader_time: 0.00300 (0.00358)  --> STEP: 185/639 -- GLOBAL_STEP: 58425 | > loss: -0.49851 (-0.52186) | > log_mle: -0.60607 (-0.63604) | > loss_dur: 0.10756 (0.11418) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.34628 (10.30399) | > current_lr: 0.00026 | > step_time: 0.57400 (0.43813) | > loader_time: 0.00320 (0.00352)  --> STEP: 210/639 -- GLOBAL_STEP: 58450 | > loss: -0.42913 (-0.51847) | > log_mle: -0.52689 (-0.63258) | > loss_dur: 0.09776 (0.11411) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.22008 (10.15173) | > current_lr: 0.00026 | > step_time: 0.50120 (0.45311) | > loader_time: 0.00310 (0.00348)  --> STEP: 235/639 -- GLOBAL_STEP: 58475 | > loss: -0.51102 (-0.51867) | > log_mle: -0.61743 (-0.63253) | > loss_dur: 0.10641 (0.11386) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.96151 (10.30790) | > current_lr: 0.00026 | > step_time: 0.55470 (0.46885) | > loader_time: 0.00300 (0.00347)  --> STEP: 260/639 -- GLOBAL_STEP: 58500 | > loss: -0.49687 (-0.51775) | > log_mle: -0.61991 (-0.63194) | > loss_dur: 0.12305 (0.11420) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.07714 (10.32638) | > current_lr: 0.00026 | > step_time: 0.58840 (0.48269) | > loader_time: 0.00330 (0.00346)  --> STEP: 285/639 -- GLOBAL_STEP: 58525 | > loss: -0.44883 (-0.51625) | > log_mle: -0.56549 (-0.63074) | > loss_dur: 0.11666 (0.11449) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.58918 (10.32814) | > current_lr: 0.00026 | > step_time: 0.61320 (0.49875) | > loader_time: 0.00320 (0.00347)  --> STEP: 310/639 -- GLOBAL_STEP: 58550 | > loss: -0.48791 (-0.51436) | > log_mle: -0.60827 (-0.62910) | > loss_dur: 0.12036 (0.11473) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.75709 (10.23025) | > current_lr: 0.00026 | > step_time: 0.58210 (0.51378) | > loader_time: 0.00330 (0.00347)  --> STEP: 335/639 -- GLOBAL_STEP: 58575 | > loss: -0.50545 (-0.51180) | > log_mle: -0.63144 (-0.62696) | > loss_dur: 0.12598 (0.11515) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 16.85810 (10.14573) | > current_lr: 0.00026 | > step_time: 0.64670 (0.52830) | > loader_time: 0.00340 (0.00347)  --> STEP: 360/639 -- GLOBAL_STEP: 58600 | > loss: -0.45825 (-0.50998) | > log_mle: -0.56801 (-0.62497) | > loss_dur: 0.10976 (0.11499) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.98884 (9.99243) | > current_lr: 0.00026 | > step_time: 0.67810 (0.54100) | > loader_time: 0.00370 (0.00348)  --> STEP: 385/639 -- GLOBAL_STEP: 58625 | > loss: -0.48311 (-0.50824) | > log_mle: -0.59941 (-0.62309) | > loss_dur: 0.11630 (0.11485) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.29884 (9.86556) | > current_lr: 0.00026 | > step_time: 0.67590 (0.55442) | > loader_time: 0.00380 (0.00349)  --> STEP: 410/639 -- GLOBAL_STEP: 58650 | > loss: -0.44430 (-0.50627) | > log_mle: -0.57450 (-0.62135) | > loss_dur: 0.13020 (0.11508) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.18393 (9.83159) | > current_lr: 0.00026 | > step_time: 0.80980 (0.56625) | > loader_time: 0.00360 (0.00349)  --> STEP: 435/639 -- GLOBAL_STEP: 58675 | > loss: -0.47969 (-0.50428) | > log_mle: -0.58664 (-0.61933) | > loss_dur: 0.10695 (0.11505) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.36702 (9.72133) | > current_lr: 0.00026 | > step_time: 0.69330 (0.57808) | > loader_time: 0.00350 (0.00351)  --> STEP: 460/639 -- GLOBAL_STEP: 58700 | > loss: -0.43795 (-0.50230) | > log_mle: -0.55368 (-0.61732) | > loss_dur: 0.11573 (0.11502) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.97154 (9.60388) | > current_lr: 0.00026 | > step_time: 0.93990 (0.58987) | > loader_time: 0.00400 (0.00352)  --> STEP: 485/639 -- GLOBAL_STEP: 58725 | > loss: -0.47261 (-0.50066) | > log_mle: -0.58265 (-0.61587) | > loss_dur: 0.11004 (0.11521) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.76869 (9.58694) | > current_lr: 0.00026 | > step_time: 0.85250 (0.60307) | > loader_time: 0.00390 (0.00354)  --> STEP: 510/639 -- GLOBAL_STEP: 58750 | > loss: -0.46566 (-0.49932) | > log_mle: -0.58586 (-0.61455) | > loss_dur: 0.12020 (0.11524) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.39518 (9.53487) | > current_lr: 0.00026 | > step_time: 0.82640 (0.61563) | > loader_time: 0.00420 (0.00356)  --> STEP: 535/639 -- GLOBAL_STEP: 58775 | > loss: -0.44829 (-0.49774) | > log_mle: -0.57709 (-0.61310) | > loss_dur: 0.12880 (0.11536) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.64453 (9.44754) | > current_lr: 0.00026 | > step_time: 0.86780 (0.62853) | > loader_time: 0.00420 (0.00358)  --> STEP: 560/639 -- GLOBAL_STEP: 58800 | > loss: -0.45936 (-0.49624) | > log_mle: -0.58597 (-0.61174) | > loss_dur: 0.12661 (0.11550) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.36814 (9.38055) | > current_lr: 0.00026 | > step_time: 0.91540 (0.64297) | > loader_time: 0.00390 (0.00360)  --> STEP: 585/639 -- GLOBAL_STEP: 58825 | > loss: -0.47642 (-0.49495) | > log_mle: -0.58833 (-0.61052) | > loss_dur: 0.11191 (0.11558) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.82098 (9.28132) | > current_lr: 0.00026 | > step_time: 1.05640 (0.65838) | > loader_time: 0.00430 (0.00363)  --> STEP: 610/639 -- GLOBAL_STEP: 58850 | > loss: -0.44401 (-0.49370) | > log_mle: -0.55807 (-0.60940) | > loss_dur: 0.11406 (0.11569) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.54541 (9.19951) | > current_lr: 0.00026 | > step_time: 1.03470 (0.67478) | > loader_time: 0.00450 (0.00366)  --> STEP: 635/639 -- GLOBAL_STEP: 58875 | > loss: -0.45662 (-0.49263) | > log_mle: -0.58233 (-0.60862) | > loss_dur: 0.12571 (0.11600) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.28756 (9.17952) | > current_lr: 0.00026 | > step_time: 1.09260 (0.69206) | > loader_time: 0.00420 (0.00369) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02797 (-0.00774) | > avg_loss: -0.45637 (-0.00131) | > avg_log_mle: -0.58970 (-0.00056) | > avg_loss_dur: 0.13333 (-0.00075)  > EPOCH: 92/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 02:36:28)   --> STEP: 20/639 -- GLOBAL_STEP: 58900 | > loss: -0.65070 (-0.56018) | > log_mle: -0.79670 (-0.67964) | > loss_dur: 0.14600 (0.11945) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 24.49574 (13.64957) | > current_lr: 0.00026 | > step_time: 0.34470 (0.30315) | > loader_time: 0.00290 (0.00291)  --> STEP: 45/639 -- GLOBAL_STEP: 58925 | > loss: -0.50339 (-0.54097) | > log_mle: -0.59969 (-0.65629) | > loss_dur: 0.09629 (0.11532) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.03802 (11.25330) | > current_lr: 0.00026 | > step_time: 0.40670 (0.33752) | > loader_time: 0.00280 (0.00283)  --> STEP: 70/639 -- GLOBAL_STEP: 58950 | > loss: -0.52875 (-0.53181) | > log_mle: -0.64575 (-0.64598) | > loss_dur: 0.11700 (0.11416) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.83791 (10.45137) | > current_lr: 0.00026 | > step_time: 0.39050 (0.36362) | > loader_time: 0.00300 (0.00285)  --> STEP: 95/639 -- GLOBAL_STEP: 58975 | > loss: -0.59590 (-0.52831) | > log_mle: -0.69953 (-0.64044) | > loss_dur: 0.10363 (0.11213) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.36420 (9.99800) | > current_lr: 0.00026 | > step_time: 0.40570 (0.37646) | > loader_time: 0.00270 (0.00287)  --> STEP: 120/639 -- GLOBAL_STEP: 59000 | > loss: -0.55227 (-0.52516) | > log_mle: -0.68909 (-0.63838) | > loss_dur: 0.13683 (0.11322) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 21.15909 (10.12969) | > current_lr: 0.00026 | > step_time: 0.54400 (0.39463) | > loader_time: 0.00270 (0.00287)  --> STEP: 145/639 -- GLOBAL_STEP: 59025 | > loss: -0.51770 (-0.52409) | > log_mle: -0.64577 (-0.63796) | > loss_dur: 0.12808 (0.11387) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.41101 (10.35853) | > current_lr: 0.00026 | > step_time: 0.54100 (0.41038) | > loader_time: 0.00250 (0.00290)  --> STEP: 170/639 -- GLOBAL_STEP: 59050 | > loss: -0.51744 (-0.52237) | > log_mle: -0.62925 (-0.63627) | > loss_dur: 0.11180 (0.11390) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.51958 (10.36049) | > current_lr: 0.00026 | > step_time: 0.54500 (0.42498) | > loader_time: 0.00320 (0.00292)  --> STEP: 195/639 -- GLOBAL_STEP: 59075 | > loss: -0.47675 (-0.52109) | > log_mle: -0.57263 (-0.63464) | > loss_dur: 0.09588 (0.11354) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.46915 (10.30958) | > current_lr: 0.00026 | > step_time: 0.49780 (0.44050) | > loader_time: 0.00340 (0.00295)  --> STEP: 220/639 -- GLOBAL_STEP: 59100 | > loss: -0.51530 (-0.51830) | > log_mle: -0.62127 (-0.63195) | > loss_dur: 0.10597 (0.11365) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.81045 (10.09937) | > current_lr: 0.00026 | > step_time: 0.64110 (0.45573) | > loader_time: 0.00320 (0.00299)  --> STEP: 245/639 -- GLOBAL_STEP: 59125 | > loss: -0.50053 (-0.51874) | > log_mle: -0.60823 (-0.63246) | > loss_dur: 0.10770 (0.11372) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.44411 (10.12254) | > current_lr: 0.00026 | > step_time: 0.52240 (0.46959) | > loader_time: 0.00300 (0.00302)  --> STEP: 270/639 -- GLOBAL_STEP: 59150 | > loss: -0.55070 (-0.51797) | > log_mle: -0.67563 (-0.63224) | > loss_dur: 0.12493 (0.11427) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 23.10990 (10.29768) | > current_lr: 0.00026 | > step_time: 0.85980 (0.48641) | > loader_time: 0.00380 (0.00306)  --> STEP: 295/639 -- GLOBAL_STEP: 59175 | > loss: -0.47027 (-0.51558) | > log_mle: -0.56923 (-0.63012) | > loss_dur: 0.09896 (0.11454) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.29367 (10.15510) | > current_lr: 0.00026 | > step_time: 0.63290 (0.50078) | > loader_time: 0.00350 (0.00308)  --> STEP: 320/639 -- GLOBAL_STEP: 59200 | > loss: -0.47997 (-0.51362) | > log_mle: -0.60971 (-0.62861) | > loss_dur: 0.12974 (0.11499) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.66670 (10.18534) | > current_lr: 0.00026 | > step_time: 0.74740 (0.51634) | > loader_time: 0.00330 (0.00310)  --> STEP: 345/639 -- GLOBAL_STEP: 59225 | > loss: -0.49155 (-0.51151) | > log_mle: -0.59978 (-0.62676) | > loss_dur: 0.10823 (0.11525) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.76566 (10.06318) | > current_lr: 0.00026 | > step_time: 0.64780 (0.53067) | > loader_time: 0.00360 (0.00313)  --> STEP: 370/639 -- GLOBAL_STEP: 59250 | > loss: -0.49877 (-0.50940) | > log_mle: -0.61217 (-0.62450) | > loss_dur: 0.11340 (0.11509) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.78375 (9.88598) | > current_lr: 0.00026 | > step_time: 0.87950 (0.54335) | > loader_time: 0.00380 (0.00315)  --> STEP: 395/639 -- GLOBAL_STEP: 59275 | > loss: -0.45976 (-0.50746) | > log_mle: -0.58127 (-0.62249) | > loss_dur: 0.12151 (0.11503) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.29570 (9.72080) | > current_lr: 0.00026 | > step_time: 0.71530 (0.55541) | > loader_time: 0.00360 (0.00318)  --> STEP: 420/639 -- GLOBAL_STEP: 59300 | > loss: -0.46163 (-0.50522) | > log_mle: -0.57529 (-0.62048) | > loss_dur: 0.11366 (0.11527) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.14584 (9.63792) | > current_lr: 0.00026 | > step_time: 0.67660 (0.56797) | > loader_time: 0.00350 (0.00321)  --> STEP: 445/639 -- GLOBAL_STEP: 59325 | > loss: -0.47755 (-0.50367) | > log_mle: -0.58539 (-0.61890) | > loss_dur: 0.10784 (0.11523) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.75580 (9.53812) | > current_lr: 0.00026 | > step_time: 0.69790 (0.57952) | > loader_time: 0.00350 (0.00324)  --> STEP: 470/639 -- GLOBAL_STEP: 59350 | > loss: -0.48010 (-0.50176) | > log_mle: -0.59812 (-0.61695) | > loss_dur: 0.11802 (0.11520) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.34396 (9.37253) | > current_lr: 0.00026 | > step_time: 0.81830 (0.59149) | > loader_time: 0.00370 (0.00326)  --> STEP: 495/639 -- GLOBAL_STEP: 59375 | > loss: -0.43373 (-0.50013) | > log_mle: -0.53330 (-0.61546) | > loss_dur: 0.09958 (0.11533) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.27094 (9.30388) | > current_lr: 0.00026 | > step_time: 0.76150 (0.60437) | > loader_time: 0.00340 (0.00329)  --> STEP: 520/639 -- GLOBAL_STEP: 59400 | > loss: -0.47806 (-0.49893) | > log_mle: -0.59021 (-0.61426) | > loss_dur: 0.11214 (0.11533) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.48540 (9.25552) | > current_lr: 0.00026 | > step_time: 0.79430 (0.61777) | > loader_time: 0.00370 (0.00332)  --> STEP: 545/639 -- GLOBAL_STEP: 59425 | > loss: -0.49522 (-0.49751) | > log_mle: -0.61940 (-0.61296) | > loss_dur: 0.12417 (0.11545) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.73261 (9.17634) | > current_lr: 0.00026 | > step_time: 1.03240 (0.63271) | > loader_time: 0.00410 (0.00336)  --> STEP: 570/639 -- GLOBAL_STEP: 59450 | > loss: -0.46312 (-0.49593) | > log_mle: -0.56928 (-0.61143) | > loss_dur: 0.10616 (0.11549) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.97943 (9.10978) | > current_lr: 0.00026 | > step_time: 0.82720 (0.64620) | > loader_time: 0.00600 (0.00340)  --> STEP: 595/639 -- GLOBAL_STEP: 59475 | > loss: -0.45131 (-0.49486) | > log_mle: -0.56132 (-0.61049) | > loss_dur: 0.11001 (0.11562) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.53935 (9.07982) | > current_lr: 0.00026 | > step_time: 0.97650 (0.66248) | > loader_time: 0.00380 (0.00344)  --> STEP: 620/639 -- GLOBAL_STEP: 59500 | > loss: -0.45313 (-0.49357) | > log_mle: -0.56759 (-0.60930) | > loss_dur: 0.11446 (0.11574) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.25729 (9.00714) | > current_lr: 0.00026 | > step_time: 1.17220 (0.67912) | > loader_time: 0.00470 (0.00348) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02831 (+0.00034) | > avg_loss: -0.45438 (+0.00199) | > avg_log_mle: -0.58899 (+0.00071) | > avg_loss_dur: 0.13460 (+0.00128)  > EPOCH: 93/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 02:44:05)   --> STEP: 5/639 -- GLOBAL_STEP: 59525 | > loss: -0.55943 (-0.56140) | > log_mle: -0.70096 (-0.68486) | > loss_dur: 0.14153 (0.12346) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 23.06844 (12.94649) | > current_lr: 0.00026 | > step_time: 0.30730 (0.29544) | > loader_time: 0.00250 (0.00264)  --> STEP: 30/639 -- GLOBAL_STEP: 59550 | > loss: -0.55342 (-0.54750) | > log_mle: -0.65924 (-0.66364) | > loss_dur: 0.10582 (0.11614) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.03925 (11.93519) | > current_lr: 0.00026 | > step_time: 0.32120 (0.32647) | > loader_time: 0.00250 (0.00254)  --> STEP: 55/639 -- GLOBAL_STEP: 59575 | > loss: -0.49922 (-0.53855) | > log_mle: -0.59487 (-0.65232) | > loss_dur: 0.09565 (0.11377) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.27119 (10.99402) | > current_lr: 0.00026 | > step_time: 0.34260 (0.35133) | > loader_time: 0.00270 (0.00264)  --> STEP: 80/639 -- GLOBAL_STEP: 59600 | > loss: -0.50637 (-0.53066) | > log_mle: -0.63098 (-0.64317) | > loss_dur: 0.12461 (0.11250) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.88799 (10.23505) | > current_lr: 0.00026 | > step_time: 0.44470 (0.37156) | > loader_time: 0.00290 (0.00274)  --> STEP: 105/639 -- GLOBAL_STEP: 59625 | > loss: -0.47292 (-0.52804) | > log_mle: -0.58502 (-0.64006) | > loss_dur: 0.11210 (0.11202) | > amp_scaler: 4096.00000 (2204.03810) | > grad_norm: 7.41154 (10.35104) | > current_lr: 0.00026 | > step_time: 0.41240 (0.38956) | > loader_time: 0.00300 (0.00281)  --> STEP: 130/639 -- GLOBAL_STEP: 59650 | > loss: -0.50289 (-0.52509) | > log_mle: -0.62155 (-0.63791) | > loss_dur: 0.11866 (0.11282) | > amp_scaler: 4096.00000 (2567.87692) | > grad_norm: 9.19274 (10.46192) | > current_lr: 0.00026 | > step_time: 0.57050 (0.40346) | > loader_time: 0.00310 (0.00285)  --> STEP: 155/639 -- GLOBAL_STEP: 59675 | > loss: -0.49236 (-0.52349) | > log_mle: -0.60820 (-0.63664) | > loss_dur: 0.11584 (0.11315) | > amp_scaler: 4096.00000 (2814.34839) | > grad_norm: 10.18415 (10.63441) | > current_lr: 0.00026 | > step_time: 0.55450 (0.41896) | > loader_time: 0.00300 (0.00289)  --> STEP: 180/639 -- GLOBAL_STEP: 59700 | > loss: -0.53606 (-0.52411) | > log_mle: -0.66472 (-0.63775) | > loss_dur: 0.12866 (0.11364) | > amp_scaler: 4096.00000 (2992.35556) | > grad_norm: 12.66919 (10.78144) | > current_lr: 0.00026 | > step_time: 0.66620 (0.43489) | > loader_time: 0.00340 (0.00293)  --> STEP: 205/639 -- GLOBAL_STEP: 59725 | > loss: -0.47380 (-0.52079) | > log_mle: -0.58580 (-0.63374) | > loss_dur: 0.11200 (0.11295) | > amp_scaler: 4096.00000 (3126.94634) | > grad_norm: 6.92485 (10.41344) | > current_lr: 0.00026 | > step_time: 0.54750 (0.45009) | > loader_time: 0.00320 (0.00297)  --> STEP: 230/639 -- GLOBAL_STEP: 59750 | > loss: -0.50714 (-0.52000) | > log_mle: -0.62293 (-0.63317) | > loss_dur: 0.11579 (0.11317) | > amp_scaler: 2048.00000 (3089.80870) | > grad_norm: 8.50359 (10.40130) | > current_lr: 0.00026 | > step_time: 0.56010 (0.46574) | > loader_time: 0.00350 (0.00301)  --> STEP: 255/639 -- GLOBAL_STEP: 59775 | > loss: -0.47655 (-0.51885) | > log_mle: -0.59349 (-0.63194) | > loss_dur: 0.11694 (0.11309) | > amp_scaler: 2048.00000 (2987.67059) | > grad_norm: 5.69090 (10.31675) | > current_lr: 0.00026 | > step_time: 0.64180 (0.47975) | > loader_time: 0.00340 (0.00304)  --> STEP: 280/639 -- GLOBAL_STEP: 59800 | > loss: -0.49645 (-0.51849) | > log_mle: -0.61574 (-0.63199) | > loss_dur: 0.11929 (0.11350) | > amp_scaler: 2048.00000 (2903.77143) | > grad_norm: 10.57196 (10.45458) | > current_lr: 0.00026 | > step_time: 0.59080 (0.49650) | > loader_time: 0.00340 (0.00308)  --> STEP: 305/639 -- GLOBAL_STEP: 59825 | > loss: -0.47551 (-0.51645) | > log_mle: -0.59930 (-0.63020) | > loss_dur: 0.12379 (0.11375) | > amp_scaler: 2048.00000 (2833.62623) | > grad_norm: 7.29796 (10.35206) | > current_lr: 0.00026 | > step_time: 0.74280 (0.51062) | > loader_time: 0.00360 (0.00312)  --> STEP: 330/639 -- GLOBAL_STEP: 59850 | > loss: -0.49033 (-0.51382) | > log_mle: -0.61807 (-0.62807) | > loss_dur: 0.12774 (0.11425) | > amp_scaler: 2048.00000 (2774.10909) | > grad_norm: 11.03747 (10.29360) | > current_lr: 0.00026 | > step_time: 0.72490 (0.52525) | > loader_time: 0.00350 (0.00316)  --> STEP: 355/639 -- GLOBAL_STEP: 59875 | > loss: -0.44417 (-0.51191) | > log_mle: -0.54962 (-0.62611) | > loss_dur: 0.10545 (0.11420) | > amp_scaler: 2048.00000 (2722.97465) | > grad_norm: 2.80270 (10.19362) | > current_lr: 0.00026 | > step_time: 0.59350 (0.53807) | > loader_time: 0.00340 (0.00319)  --> STEP: 380/639 -- GLOBAL_STEP: 59900 | > loss: -0.48407 (-0.50975) | > log_mle: -0.58629 (-0.62383) | > loss_dur: 0.10222 (0.11408) | > amp_scaler: 2048.00000 (2678.56842) | > grad_norm: 5.06140 (10.02472) | > current_lr: 0.00026 | > step_time: 0.61960 (0.55137) | > loader_time: 0.00350 (0.00322)  --> STEP: 405/639 -- GLOBAL_STEP: 59925 | > loss: -0.46145 (-0.50813) | > log_mle: -0.58665 (-0.62253) | > loss_dur: 0.12520 (0.11440) | > amp_scaler: 2048.00000 (2639.64444) | > grad_norm: 8.28470 (9.97234) | > current_lr: 0.00026 | > step_time: 0.68110 (0.56259) | > loader_time: 0.00340 (0.00325)  --> STEP: 430/639 -- GLOBAL_STEP: 59950 | > loss: -0.44330 (-0.50602) | > log_mle: -0.54964 (-0.62059) | > loss_dur: 0.10634 (0.11456) | > amp_scaler: 2048.00000 (2605.24651) | > grad_norm: 3.77074 (9.86995) | > current_lr: 0.00026 | > step_time: 0.77530 (0.57535) | > loader_time: 0.00350 (0.00328)  --> STEP: 455/639 -- GLOBAL_STEP: 59975 | > loss: -0.46159 (-0.50405) | > log_mle: -0.57397 (-0.61857) | > loss_dur: 0.11238 (0.11452) | > amp_scaler: 2048.00000 (2574.62857) | > grad_norm: 5.63914 (9.70955) | > current_lr: 0.00026 | > step_time: 0.78920 (0.58644) | > loader_time: 0.00360 (0.00330)  --> STEP: 480/639 -- GLOBAL_STEP: 60000 | > loss: -0.41870 (-0.50213) | > log_mle: -0.53033 (-0.61687) | > loss_dur: 0.11163 (0.11475) | > amp_scaler: 2048.00000 (2547.20000) | > grad_norm: 2.28233 (9.61968) | > current_lr: 0.00026 | > step_time: 0.74860 (0.59909) | > loader_time: 0.00360 (0.00333) > CHECKPOINT : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/checkpoint_60000.pth.tar warning: audio amplitude out of range, auto clipped.  --> STEP: 505/639 -- GLOBAL_STEP: 60025 | > loss: -0.48497 (-0.50058) | > log_mle: -0.59716 (-0.61533) | > loss_dur: 0.11219 (0.11475) | > amp_scaler: 2048.00000 (2522.48713) | > grad_norm: 7.65955 (9.52327) | > current_lr: 0.00026 | > step_time: 0.88850 (0.61078) | > loader_time: 0.00420 (0.00340)  --> STEP: 530/639 -- GLOBAL_STEP: 60050 | > loss: -0.45784 (-0.49906) | > log_mle: -0.58750 (-0.61401) | > loss_dur: 0.12967 (0.11495) | > amp_scaler: 2048.00000 (2500.10566) | > grad_norm: 7.82745 (9.47908) | > current_lr: 0.00026 | > step_time: 0.92360 (0.62469) | > loader_time: 0.00400 (0.00343)  --> STEP: 555/639 -- GLOBAL_STEP: 60075 | > loss: -0.45533 (-0.49747) | > log_mle: -0.56983 (-0.61247) | > loss_dur: 0.11450 (0.11500) | > amp_scaler: 2048.00000 (2479.74054) | > grad_norm: 5.46015 (9.36196) | > current_lr: 0.00026 | > step_time: 0.87800 (0.63837) | > loader_time: 0.00400 (0.00346)  --> STEP: 580/639 -- GLOBAL_STEP: 60100 | > loss: -0.47774 (-0.49613) | > log_mle: -0.58732 (-0.61129) | > loss_dur: 0.10958 (0.11516) | > amp_scaler: 2048.00000 (2461.13103) | > grad_norm: 7.74462 (9.30815) | > current_lr: 0.00026 | > step_time: 0.96280 (0.65315) | > loader_time: 0.00430 (0.00349)  --> STEP: 605/639 -- GLOBAL_STEP: 60125 | > loss: -0.45456 (-0.49485) | > log_mle: -0.55802 (-0.61019) | > loss_dur: 0.10346 (0.11534) | > amp_scaler: 2048.00000 (2444.05950) | > grad_norm: 2.96805 (9.22501) | > current_lr: 0.00026 | > step_time: 1.01430 (0.66898) | > loader_time: 0.00450 (0.00354)  --> STEP: 630/639 -- GLOBAL_STEP: 60150 | > loss: -0.46061 (-0.49376) | > log_mle: -0.59096 (-0.60926) | > loss_dur: 0.13035 (0.11550) | > amp_scaler: 2048.00000 (2428.34286) | > grad_norm: 8.26015 (9.15872) | > current_lr: 0.00026 | > step_time: 1.18150 (0.68648) | > loader_time: 0.00460 (0.00359) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02222 (-0.00609) | > avg_loss: -0.45310 (+0.00129) | > avg_log_mle: -0.58851 (+0.00048) | > avg_loss_dur: 0.13542 (+0.00081)  > EPOCH: 94/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 02:51:46)   --> STEP: 15/639 -- GLOBAL_STEP: 60175 | > loss: -0.65121 (-0.55834) | > log_mle: -0.78643 (-0.67422) | > loss_dur: 0.13523 (0.11588) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 23.53622 (11.03477) | > current_lr: 0.00026 | > step_time: 0.32740 (0.30204) | > loader_time: 0.00240 (0.00249)  --> STEP: 40/639 -- GLOBAL_STEP: 60200 | > loss: -0.52523 (-0.54346) | > log_mle: -0.63222 (-0.65685) | > loss_dur: 0.10698 (0.11339) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.22822 (10.44254) | > current_lr: 0.00026 | > step_time: 0.35290 (0.32961) | > loader_time: 0.00280 (0.00262)  --> STEP: 65/639 -- GLOBAL_STEP: 60225 | > loss: -0.52221 (-0.53544) | > log_mle: -0.61949 (-0.64938) | > loss_dur: 0.09728 (0.11395) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.81257 (10.14710) | > current_lr: 0.00026 | > step_time: 0.37980 (0.36244) | > loader_time: 0.00320 (0.00276)  --> STEP: 90/639 -- GLOBAL_STEP: 60250 | > loss: -0.46721 (-0.52752) | > log_mle: -0.57584 (-0.63984) | > loss_dur: 0.10863 (0.11232) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.44159 (9.33037) | > current_lr: 0.00026 | > step_time: 0.50810 (0.38103) | > loader_time: 0.00300 (0.00281)  --> STEP: 115/639 -- GLOBAL_STEP: 60275 | > loss: -0.44144 (-0.52674) | > log_mle: -0.54469 (-0.63905) | > loss_dur: 0.10325 (0.11231) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.85738 (9.88162) | > current_lr: 0.00026 | > step_time: 0.41250 (0.39622) | > loader_time: 0.00310 (0.00285)  --> STEP: 140/639 -- GLOBAL_STEP: 60300 | > loss: -0.49328 (-0.52560) | > log_mle: -0.62508 (-0.63869) | > loss_dur: 0.13180 (0.11309) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.74848 (10.01131) | > current_lr: 0.00026 | > step_time: 0.55640 (0.40957) | > loader_time: 0.00310 (0.00289)  --> STEP: 165/639 -- GLOBAL_STEP: 60325 | > loss: -0.50177 (-0.52427) | > log_mle: -0.63034 (-0.63720) | > loss_dur: 0.12857 (0.11293) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.42050 (10.02749) | > current_lr: 0.00026 | > step_time: 0.48290 (0.42475) | > loader_time: 0.00300 (0.00293)  --> STEP: 190/639 -- GLOBAL_STEP: 60350 | > loss: -0.44892 (-0.52372) | > log_mle: -0.56097 (-0.63653) | > loss_dur: 0.11204 (0.11281) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.62351 (10.07265) | > current_lr: 0.00026 | > step_time: 0.52640 (0.44156) | > loader_time: 0.00340 (0.00297)  --> STEP: 215/639 -- GLOBAL_STEP: 60375 | > loss: -0.54699 (-0.52105) | > log_mle: -0.66040 (-0.63407) | > loss_dur: 0.11341 (0.11302) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.45886 (9.99095) | > current_lr: 0.00026 | > step_time: 0.52860 (0.45633) | > loader_time: 0.00340 (0.00301)  --> STEP: 240/639 -- GLOBAL_STEP: 60400 | > loss: -0.55993 (-0.52158) | > log_mle: -0.67550 (-0.63459) | > loss_dur: 0.11557 (0.11301) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.57290 (10.08889) | > current_lr: 0.00026 | > step_time: 0.55000 (0.47084) | > loader_time: 0.00340 (0.00304)  --> STEP: 265/639 -- GLOBAL_STEP: 60425 | > loss: -0.56029 (-0.52029) | > log_mle: -0.69382 (-0.63370) | > loss_dur: 0.13353 (0.11341) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 22.74865 (10.17203) | > current_lr: 0.00026 | > step_time: 0.59610 (0.48609) | > loader_time: 0.00340 (0.00308)  --> STEP: 290/639 -- GLOBAL_STEP: 60450 | > loss: -0.48586 (-0.51832) | > log_mle: -0.59302 (-0.63191) | > loss_dur: 0.10717 (0.11359) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.74109 (10.08739) | > current_lr: 0.00026 | > step_time: 0.63220 (0.50063) | > loader_time: 0.00340 (0.00311)  --> STEP: 315/639 -- GLOBAL_STEP: 60475 | > loss: -0.46621 (-0.51626) | > log_mle: -0.57275 (-0.63016) | > loss_dur: 0.10654 (0.11390) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.15581 (10.01981) | > current_lr: 0.00026 | > step_time: 0.61850 (0.51459) | > loader_time: 0.00340 (0.00314)  --> STEP: 340/639 -- GLOBAL_STEP: 60500 | > loss: -0.47818 (-0.51404) | > log_mle: -0.59126 (-0.62855) | > loss_dur: 0.11308 (0.11451) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.50729 (10.06361) | > current_lr: 0.00026 | > step_time: 0.59860 (0.53033) | > loader_time: 0.00340 (0.00318)  --> STEP: 365/639 -- GLOBAL_STEP: 60525 | > loss: -0.47701 (-0.51172) | > log_mle: -0.59911 (-0.62604) | > loss_dur: 0.12210 (0.11432) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.07509 (9.91046) | > current_lr: 0.00026 | > step_time: 0.79730 (0.54256) | > loader_time: 0.00390 (0.00320)  --> STEP: 390/639 -- GLOBAL_STEP: 60550 | > loss: -0.47382 (-0.50975) | > log_mle: -0.59791 (-0.62394) | > loss_dur: 0.12409 (0.11419) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.30220 (9.73319) | > current_lr: 0.00026 | > step_time: 0.70450 (0.55436) | > loader_time: 0.00350 (0.00323)  --> STEP: 415/639 -- GLOBAL_STEP: 60575 | > loss: -0.45904 (-0.50775) | > log_mle: -0.57421 (-0.62224) | > loss_dur: 0.11517 (0.11449) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.85084 (9.69042) | > current_lr: 0.00026 | > step_time: 0.73820 (0.56639) | > loader_time: 0.00350 (0.00326)  --> STEP: 440/639 -- GLOBAL_STEP: 60600 | > loss: -0.44456 (-0.50603) | > log_mle: -0.55885 (-0.62058) | > loss_dur: 0.11429 (0.11455) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.79306 (9.62587) | > current_lr: 0.00026 | > step_time: 0.79800 (0.57842) | > loader_time: 0.00330 (0.00328)  --> STEP: 465/639 -- GLOBAL_STEP: 60625 | > loss: -0.45208 (-0.50398) | > log_mle: -0.56261 (-0.61849) | > loss_dur: 0.11052 (0.11451) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.76868 (9.49602) | > current_lr: 0.00026 | > step_time: 0.78650 (0.58925) | > loader_time: 0.00360 (0.00330)  --> STEP: 490/639 -- GLOBAL_STEP: 60650 | > loss: -0.44757 (-0.50224) | > log_mle: -0.54586 (-0.61695) | > loss_dur: 0.09829 (0.11471) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.64162 (9.41577) | > current_lr: 0.00026 | > step_time: 0.77600 (0.60200) | > loader_time: 0.00370 (0.00333)  --> STEP: 515/639 -- GLOBAL_STEP: 60675 | > loss: -0.46398 (-0.50104) | > log_mle: -0.57558 (-0.61583) | > loss_dur: 0.11160 (0.11479) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.66638 (9.36087) | > current_lr: 0.00026 | > step_time: 0.99470 (0.61494) | > loader_time: 0.00400 (0.00336)  --> STEP: 540/639 -- GLOBAL_STEP: 60700 | > loss: -0.47171 (-0.49948) | > log_mle: -0.59104 (-0.61436) | > loss_dur: 0.11933 (0.11488) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.84370 (9.29099) | > current_lr: 0.00026 | > step_time: 0.88300 (0.62959) | > loader_time: 0.00410 (0.00339)  --> STEP: 565/639 -- GLOBAL_STEP: 60725 | > loss: -0.43814 (-0.49785) | > log_mle: -0.55140 (-0.61285) | > loss_dur: 0.11326 (0.11500) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.40639 (9.20692) | > current_lr: 0.00026 | > step_time: 0.97890 (0.64296) | > loader_time: 0.00410 (0.00342)  --> STEP: 590/639 -- GLOBAL_STEP: 60750 | > loss: -0.46419 (-0.49669) | > log_mle: -0.58178 (-0.61184) | > loss_dur: 0.11759 (0.11515) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.29863 (9.18352) | > current_lr: 0.00026 | > step_time: 0.90010 (0.65863) | > loader_time: 0.00430 (0.00346)  --> STEP: 615/639 -- GLOBAL_STEP: 60775 | > loss: -0.46740 (-0.49541) | > log_mle: -0.58570 (-0.61066) | > loss_dur: 0.11830 (0.11524) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.35556 (9.11410) | > current_lr: 0.00026 | > step_time: 1.06860 (0.67408) | > loader_time: 0.00470 (0.00349) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02132 (-0.00091) | > avg_loss: -0.45822 (-0.00513) | > avg_log_mle: -0.59309 (-0.00458) | > avg_loss_dur: 0.13487 (-0.00055)  > EPOCH: 95/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 02:59:22)   --> STEP: 0/639 -- GLOBAL_STEP: 60800 | > loss: -0.58757 (-0.58757) | > log_mle: -0.72843 (-0.72843) | > loss_dur: 0.14086 (0.14086) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 23.53988 (23.53988) | > current_lr: 0.00026 | > step_time: 0.55570 (0.55571) | > loader_time: 0.49370 (0.49371)  --> STEP: 25/639 -- GLOBAL_STEP: 60825 | > loss: -0.54818 (-0.55201) | > log_mle: -0.67012 (-0.67076) | > loss_dur: 0.12194 (0.11875) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.21619 (13.27739) | > current_lr: 0.00026 | > step_time: 0.40180 (0.32212) | > loader_time: 0.00300 (0.00270)  --> STEP: 50/639 -- GLOBAL_STEP: 60850 | > loss: -0.50278 (-0.53967) | > log_mle: -0.60228 (-0.65532) | > loss_dur: 0.09951 (0.11565) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.73043 (11.60688) | > current_lr: 0.00026 | > step_time: 0.31310 (0.34760) | > loader_time: 0.00280 (0.00276)  --> STEP: 75/639 -- GLOBAL_STEP: 60875 | > loss: -0.52069 (-0.53118) | > log_mle: -0.62190 (-0.64534) | > loss_dur: 0.10121 (0.11416) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.63002 (11.24928) | > current_lr: 0.00026 | > step_time: 0.41150 (0.37005) | > loader_time: 0.00300 (0.00280)  --> STEP: 100/639 -- GLOBAL_STEP: 60900 | > loss: -0.53287 (-0.52869) | > log_mle: -0.63680 (-0.64145) | > loss_dur: 0.10393 (0.11276) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.10223 (11.00781) | > current_lr: 0.00026 | > step_time: 0.55730 (0.38701) | > loader_time: 0.00300 (0.00286)  --> STEP: 125/639 -- GLOBAL_STEP: 60925 | > loss: -0.51066 (-0.52556) | > log_mle: -0.62967 (-0.63900) | > loss_dur: 0.11901 (0.11344) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.74064 (11.41673) | > current_lr: 0.00026 | > step_time: 0.48400 (0.40300) | > loader_time: 0.00310 (0.00291)  --> STEP: 150/639 -- GLOBAL_STEP: 60950 | > loss: -0.48380 (-0.52386) | > log_mle: -0.59088 (-0.63746) | > loss_dur: 0.10708 (0.11360) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.19207 (11.35330) | > current_lr: 0.00026 | > step_time: 0.41660 (0.41864) | > loader_time: 0.00310 (0.00294)  --> STEP: 175/639 -- GLOBAL_STEP: 60975 | > loss: -0.51563 (-0.52329) | > log_mle: -0.61153 (-0.63722) | > loss_dur: 0.09590 (0.11393) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.46847 (11.50596) | > current_lr: 0.00026 | > step_time: 0.58690 (0.43437) | > loader_time: 0.00340 (0.00297)  --> STEP: 200/639 -- GLOBAL_STEP: 61000 | > loss: -0.48093 (-0.52109) | > log_mle: -0.59655 (-0.63447) | > loss_dur: 0.11561 (0.11338) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.85173 (11.31812) | > current_lr: 0.00026 | > step_time: 0.52770 (0.44970) | > loader_time: 0.00320 (0.00300)  --> STEP: 225/639 -- GLOBAL_STEP: 61025 | > loss: -0.48210 (-0.51900) | > log_mle: -0.57601 (-0.63218) | > loss_dur: 0.09391 (0.11317) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.48480 (11.13592) | > current_lr: 0.00026 | > step_time: 0.51350 (0.46591) | > loader_time: 0.00270 (0.00304)  --> STEP: 250/639 -- GLOBAL_STEP: 61050 | > loss: -0.48369 (-0.51947) | > log_mle: -0.58018 (-0.63287) | > loss_dur: 0.09649 (0.11340) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.77782 (11.20086) | > current_lr: 0.00026 | > step_time: 0.57650 (0.48065) | > loader_time: 0.00340 (0.00307)  --> STEP: 275/639 -- GLOBAL_STEP: 61075 | > loss: -0.50268 (-0.51908) | > log_mle: -0.61576 (-0.63307) | > loss_dur: 0.11308 (0.11399) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.62500 (11.23011) | > current_lr: 0.00026 | > step_time: 0.85520 (0.49722) | > loader_time: 0.00370 (0.00311)  --> STEP: 300/639 -- GLOBAL_STEP: 61100 | > loss: -0.50061 (-0.51695) | > log_mle: -0.60936 (-0.63095) | > loss_dur: 0.10875 (0.11400) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.53133 (11.03137) | > current_lr: 0.00026 | > step_time: 0.69110 (0.51007) | > loader_time: 0.00330 (0.00313)  --> STEP: 325/639 -- GLOBAL_STEP: 61125 | > loss: -0.49046 (-0.51474) | > log_mle: -0.60903 (-0.62923) | > loss_dur: 0.11857 (0.11450) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.81701 (10.87377) | > current_lr: 0.00026 | > step_time: 0.60780 (0.52438) | > loader_time: 0.00350 (0.00317)  --> STEP: 350/639 -- GLOBAL_STEP: 61150 | > loss: -0.50065 (-0.51257) | > log_mle: -0.63495 (-0.62722) | > loss_dur: 0.13430 (0.11465) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.08632 (10.75171) | > current_lr: 0.00026 | > step_time: 0.84260 (0.53840) | > loader_time: 0.00380 (0.00321)  --> STEP: 375/639 -- GLOBAL_STEP: 61175 | > loss: -0.46131 (-0.51033) | > log_mle: -0.57351 (-0.62483) | > loss_dur: 0.11220 (0.11450) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.84289 (10.50413) | > current_lr: 0.00026 | > step_time: 0.69490 (0.55097) | > loader_time: 0.00350 (0.00324)  --> STEP: 400/639 -- GLOBAL_STEP: 61200 | > loss: -0.48658 (-0.50868) | > log_mle: -0.60266 (-0.62329) | > loss_dur: 0.11608 (0.11461) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.79765 (10.39485) | > current_lr: 0.00026 | > step_time: 0.71350 (0.56289) | > loader_time: 0.00370 (0.00326)  --> STEP: 425/639 -- GLOBAL_STEP: 61225 | > loss: -0.53389 (-0.50646) | > log_mle: -0.65611 (-0.62119) | > loss_dur: 0.12222 (0.11472) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.86538 (10.22280) | > current_lr: 0.00026 | > step_time: 0.90010 (0.57465) | > loader_time: 0.00350 (0.00328)  --> STEP: 450/639 -- GLOBAL_STEP: 61250 | > loss: -0.45877 (-0.50475) | > log_mle: -0.55888 (-0.61932) | > loss_dur: 0.10011 (0.11457) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.66802 (10.04802) | > current_lr: 0.00026 | > step_time: 0.79050 (0.58564) | > loader_time: 0.00390 (0.00331)  --> STEP: 475/639 -- GLOBAL_STEP: 61275 | > loss: -0.46891 (-0.50282) | > log_mle: -0.57218 (-0.61757) | > loss_dur: 0.10327 (0.11475) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.94966 (9.94281) | > current_lr: 0.00026 | > step_time: 0.70830 (0.59734) | > loader_time: 0.00370 (0.00333)  --> STEP: 500/639 -- GLOBAL_STEP: 61300 | > loss: -0.52097 (-0.50153) | > log_mle: -0.63739 (-0.61636) | > loss_dur: 0.11642 (0.11483) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.32717 (9.86110) | > current_lr: 0.00026 | > step_time: 0.92580 (0.61051) | > loader_time: 0.00420 (0.00336)  --> STEP: 525/639 -- GLOBAL_STEP: 61325 | > loss: -0.48671 (-0.50011) | > log_mle: -0.61309 (-0.61507) | > loss_dur: 0.12638 (0.11496) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.42039 (9.76970) | > current_lr: 0.00026 | > step_time: 1.10220 (0.62438) | > loader_time: 0.00440 (0.00339)  --> STEP: 550/639 -- GLOBAL_STEP: 61350 | > loss: -0.43533 (-0.49864) | > log_mle: -0.54639 (-0.61360) | > loss_dur: 0.11107 (0.11496) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.26070 (9.63731) | > current_lr: 0.00026 | > step_time: 0.92680 (0.63860) | > loader_time: 0.00420 (0.00342)  --> STEP: 575/639 -- GLOBAL_STEP: 61375 | > loss: -0.45306 (-0.49702) | > log_mle: -0.56374 (-0.61210) | > loss_dur: 0.11068 (0.11508) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.89916 (9.56398) | > current_lr: 0.00026 | > step_time: 0.89670 (0.65276) | > loader_time: 0.00410 (0.00346)  --> STEP: 600/639 -- GLOBAL_STEP: 61400 | > loss: -0.44265 (-0.49575) | > log_mle: -0.56238 (-0.61100) | > loss_dur: 0.11972 (0.11526) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.06601 (9.52265) | > current_lr: 0.00026 | > step_time: 1.12070 (0.66933) | > loader_time: 0.00470 (0.00349)  --> STEP: 625/639 -- GLOBAL_STEP: 61425 | > loss: -0.47073 (-0.49452) | > log_mle: -0.58725 (-0.60988) | > loss_dur: 0.11652 (0.11535) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.86162 (9.49608) | > current_lr: 0.00026 | > step_time: 1.11070 (0.68602) | > loader_time: 0.00440 (0.00353) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03609 (+0.01477) | > avg_loss: -0.45691 (+0.00131) | > avg_log_mle: -0.59242 (+0.00067) | > avg_loss_dur: 0.13551 (+0.00064)  > EPOCH: 96/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 03:07:00)   --> STEP: 10/639 -- GLOBAL_STEP: 61450 | > loss: -0.52280 (-0.55036) | > log_mle: -0.65378 (-0.66922) | > loss_dur: 0.13098 (0.11886) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 16.47660 (20.32162) | > current_lr: 0.00026 | > step_time: 0.29040 (0.29575) | > loader_time: 0.00260 (0.00296)  --> STEP: 35/639 -- GLOBAL_STEP: 61475 | > loss: -0.53895 (-0.54094) | > log_mle: -0.63755 (-0.65419) | > loss_dur: 0.09860 (0.11325) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.01628 (13.34431) | > current_lr: 0.00026 | > step_time: 0.32220 (0.32909) | > loader_time: 0.00270 (0.00274)  --> STEP: 60/639 -- GLOBAL_STEP: 61500 | > loss: -0.49070 (-0.53346) | > log_mle: -0.59690 (-0.64733) | > loss_dur: 0.10620 (0.11387) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.70000 (12.18791) | > current_lr: 0.00026 | > step_time: 0.34130 (0.35764) | > loader_time: 0.00270 (0.00277)  --> STEP: 85/639 -- GLOBAL_STEP: 61525 | > loss: -0.49177 (-0.52682) | > log_mle: -0.59608 (-0.63881) | > loss_dur: 0.10430 (0.11200) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.52209 (11.07652) | > current_lr: 0.00026 | > step_time: 0.44710 (0.37326) | > loader_time: 0.00290 (0.00282)  --> STEP: 110/639 -- GLOBAL_STEP: 61550 | > loss: -0.49482 (-0.52714) | > log_mle: -0.60883 (-0.64022) | > loss_dur: 0.11401 (0.11309) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.93583 (11.21244) | > current_lr: 0.00026 | > step_time: 0.39860 (0.39172) | > loader_time: 0.00310 (0.00286)  --> STEP: 135/639 -- GLOBAL_STEP: 61575 | > loss: -0.46128 (-0.52406) | > log_mle: -0.58112 (-0.63719) | > loss_dur: 0.11985 (0.11314) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.19900 (11.49294) | > current_lr: 0.00025 | > step_time: 0.47340 (0.40396) | > loader_time: 0.00280 (0.00287)  --> STEP: 160/639 -- GLOBAL_STEP: 61600 | > loss: -0.52845 (-0.52201) | > log_mle: -0.62955 (-0.63492) | > loss_dur: 0.10109 (0.11291) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 18.61238 (11.47074) | > current_lr: 0.00025 | > step_time: 0.50760 (0.42001) | > loader_time: 0.00300 (0.00289)  --> STEP: 185/639 -- GLOBAL_STEP: 61625 | > loss: -0.49643 (-0.52340) | > log_mle: -0.60848 (-0.63666) | > loss_dur: 0.11205 (0.11326) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.64819 (12.10349) | > current_lr: 0.00025 | > step_time: 0.64630 (0.43804) | > loader_time: 0.00340 (0.00294)  --> STEP: 210/639 -- GLOBAL_STEP: 61650 | > loss: -0.43089 (-0.52021) | > log_mle: -0.52979 (-0.63335) | > loss_dur: 0.09890 (0.11314) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.29111 (11.76244) | > current_lr: 0.00025 | > step_time: 0.50970 (0.45310) | > loader_time: 0.00300 (0.00297)  --> STEP: 235/639 -- GLOBAL_STEP: 61675 | > loss: -0.50868 (-0.52029) | > log_mle: -0.62202 (-0.63353) | > loss_dur: 0.11334 (0.11324) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.26868 (11.86923) | > current_lr: 0.00025 | > step_time: 0.56540 (0.46950) | > loader_time: 0.00350 (0.00300)  --> STEP: 260/639 -- GLOBAL_STEP: 61700 | > loss: -0.49880 (-0.51927) | > log_mle: -0.61804 (-0.63297) | > loss_dur: 0.11924 (0.11370) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.36577 (11.98070) | > current_lr: 0.00025 | > step_time: 0.57980 (0.48357) | > loader_time: 0.00340 (0.00304)  --> STEP: 285/639 -- GLOBAL_STEP: 61725 | > loss: -0.45253 (-0.51788) | > log_mle: -0.56616 (-0.63185) | > loss_dur: 0.11363 (0.11397) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.24470 (11.86566) | > current_lr: 0.00025 | > step_time: 0.60450 (0.49964) | > loader_time: 0.00340 (0.00307)  --> STEP: 310/639 -- GLOBAL_STEP: 61750 | > loss: -0.49664 (-0.51599) | > log_mle: -0.60728 (-0.63021) | > loss_dur: 0.11065 (0.11422) | > amp_scaler: 4096.00000 (2153.70323) | > grad_norm: 8.42452 (11.71322) | > current_lr: 0.00025 | > step_time: 0.57320 (0.51430) | > loader_time: 0.00330 (0.00310)  --> STEP: 335/639 -- GLOBAL_STEP: 61775 | > loss: -0.51262 (-0.51332) | > log_mle: -0.63333 (-0.62803) | > loss_dur: 0.12071 (0.11472) | > amp_scaler: 4096.00000 (2298.65075) | > grad_norm: 11.57178 (11.57739) | > current_lr: 0.00025 | > step_time: 0.66420 (0.52867) | > loader_time: 0.00310 (0.00313)  --> STEP: 360/639 -- GLOBAL_STEP: 61800 | > loss: -0.46188 (-0.51150) | > log_mle: -0.56984 (-0.62607) | > loss_dur: 0.10797 (0.11457) | > amp_scaler: 4096.00000 (2423.46667) | > grad_norm: 5.86251 (11.37424) | > current_lr: 0.00025 | > step_time: 0.68690 (0.54170) | > loader_time: 0.00340 (0.00316)  --> STEP: 385/639 -- GLOBAL_STEP: 61825 | > loss: -0.49216 (-0.50977) | > log_mle: -0.60240 (-0.62420) | > loss_dur: 0.11024 (0.11444) | > amp_scaler: 4096.00000 (2532.07273) | > grad_norm: 9.79632 (11.22985) | > current_lr: 0.00025 | > step_time: 0.68200 (0.55431) | > loader_time: 0.00350 (0.00319)  --> STEP: 410/639 -- GLOBAL_STEP: 61850 | > loss: -0.44710 (-0.50789) | > log_mle: -0.57746 (-0.62255) | > loss_dur: 0.13036 (0.11465) | > amp_scaler: 4096.00000 (2627.43415) | > grad_norm: 8.42459 (11.10026) | > current_lr: 0.00025 | > step_time: 0.77210 (0.56636) | > loader_time: 0.00360 (0.00322)  --> STEP: 435/639 -- GLOBAL_STEP: 61875 | > loss: -0.48575 (-0.50580) | > log_mle: -0.59011 (-0.62048) | > loss_dur: 0.10436 (0.11468) | > amp_scaler: 4096.00000 (2711.83448) | > grad_norm: 4.96716 (10.96471) | > current_lr: 0.00025 | > step_time: 0.67300 (0.57885) | > loader_time: 0.00360 (0.00324)  --> STEP: 460/639 -- GLOBAL_STEP: 61900 | > loss: -0.44115 (-0.50376) | > log_mle: -0.55362 (-0.61841) | > loss_dur: 0.11248 (0.11465) | > amp_scaler: 4096.00000 (2787.06087) | > grad_norm: 3.64950 (10.83586) | > current_lr: 0.00025 | > step_time: 0.89890 (0.58959) | > loader_time: 0.00370 (0.00326)  --> STEP: 485/639 -- GLOBAL_STEP: 61925 | > loss: -0.47732 (-0.50209) | > log_mle: -0.58745 (-0.61691) | > loss_dur: 0.11013 (0.11481) | > amp_scaler: 4096.00000 (2854.53196) | > grad_norm: 9.92179 (10.76677) | > current_lr: 0.00025 | > step_time: 0.95460 (0.60214) | > loader_time: 0.00410 (0.00329)  --> STEP: 510/639 -- GLOBAL_STEP: 61950 | > loss: -0.47502 (-0.50082) | > log_mle: -0.59070 (-0.61559) | > loss_dur: 0.11568 (0.11478) | > amp_scaler: 4096.00000 (2915.38824) | > grad_norm: 8.29472 (10.68149) | > current_lr: 0.00025 | > step_time: 0.86520 (0.61438) | > loader_time: 0.00390 (0.00332)  --> STEP: 535/639 -- GLOBAL_STEP: 61975 | > loss: -0.45384 (-0.49933) | > log_mle: -0.57648 (-0.61420) | > loss_dur: 0.12264 (0.11487) | > amp_scaler: 4096.00000 (2970.55701) | > grad_norm: 10.21262 (10.61100) | > current_lr: 0.00025 | > step_time: 0.89910 (0.62780) | > loader_time: 0.00420 (0.00335)  --> STEP: 560/639 -- GLOBAL_STEP: 62000 | > loss: -0.46797 (-0.49786) | > log_mle: -0.58759 (-0.61286) | > loss_dur: 0.11962 (0.11500) | > amp_scaler: 4096.00000 (3020.80000) | > grad_norm: 7.79343 (10.53606) | > current_lr: 0.00025 | > step_time: 0.89460 (0.64241) | > loader_time: 0.00400 (0.00339)  --> STEP: 585/639 -- GLOBAL_STEP: 62025 | > loss: -0.47444 (-0.49658) | > log_mle: -0.58945 (-0.61163) | > loss_dur: 0.11500 (0.11505) | > amp_scaler: 4096.00000 (3066.74872) | > grad_norm: 8.51086 (10.46868) | > current_lr: 0.00025 | > step_time: 1.04250 (0.65739) | > loader_time: 0.00430 (0.00343)  --> STEP: 610/639 -- GLOBAL_STEP: 62050 | > loss: -0.44955 (-0.49527) | > log_mle: -0.55971 (-0.61048) | > loss_dur: 0.11016 (0.11521) | > amp_scaler: 4096.00000 (3108.93115) | > grad_norm: 7.28734 (10.41730) | > current_lr: 0.00025 | > step_time: 1.06460 (0.67326) | > loader_time: 0.00460 (0.00347)  --> STEP: 635/639 -- GLOBAL_STEP: 62075 | > loss: -0.46114 (-0.49419) | > log_mle: -0.58570 (-0.60971) | > loss_dur: 0.12456 (0.11552) | > amp_scaler: 4096.00000 (3147.79213) | > grad_norm: 8.35246 (10.40269) | > current_lr: 0.00025 | > step_time: 1.12460 (0.69037) | > loader_time: 0.00420 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03837 (+0.00228) | > avg_loss: -0.45692 (-0.00001) | > avg_log_mle: -0.59283 (-0.00041) | > avg_loss_dur: 0.13591 (+0.00040)  > EPOCH: 97/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 03:14:37)   --> STEP: 20/639 -- GLOBAL_STEP: 62100 | > loss: -0.63586 (-0.56790) | > log_mle: -0.79314 (-0.68392) | > loss_dur: 0.15728 (0.11602) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 32.57772 (14.77048) | > current_lr: 0.00025 | > step_time: 0.34380 (0.30314) | > loader_time: 0.00240 (0.00245)  --> STEP: 45/639 -- GLOBAL_STEP: 62125 | > loss: -0.49629 (-0.54643) | > log_mle: -0.59735 (-0.65868) | > loss_dur: 0.10107 (0.11224) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.57726 (12.60269) | > current_lr: 0.00025 | > step_time: 0.40760 (0.33949) | > loader_time: 0.00280 (0.00263)  --> STEP: 70/639 -- GLOBAL_STEP: 62150 | > loss: -0.53139 (-0.53533) | > log_mle: -0.65015 (-0.64754) | > loss_dur: 0.11876 (0.11221) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.81423 (12.14460) | > current_lr: 0.00025 | > step_time: 0.39360 (0.36590) | > loader_time: 0.00290 (0.00275)  --> STEP: 95/639 -- GLOBAL_STEP: 62175 | > loss: -0.58670 (-0.53140) | > log_mle: -0.70026 (-0.64171) | > loss_dur: 0.11356 (0.11030) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.38540 (11.53763) | > current_lr: 0.00025 | > step_time: 0.40710 (0.38002) | > loader_time: 0.00290 (0.00280)  --> STEP: 120/639 -- GLOBAL_STEP: 62200 | > loss: -0.56405 (-0.52795) | > log_mle: -0.69559 (-0.63955) | > loss_dur: 0.13154 (0.11160) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 17.39180 (11.60378) | > current_lr: 0.00025 | > step_time: 0.55930 (0.39700) | > loader_time: 0.00320 (0.00286)  --> STEP: 145/639 -- GLOBAL_STEP: 62225 | > loss: -0.51393 (-0.52659) | > log_mle: -0.63742 (-0.63905) | > loss_dur: 0.12349 (0.11246) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.53045 (11.84797) | > current_lr: 0.00025 | > step_time: 0.53480 (0.41418) | > loader_time: 0.00300 (0.00289)  --> STEP: 170/639 -- GLOBAL_STEP: 62250 | > loss: -0.51120 (-0.52486) | > log_mle: -0.62479 (-0.63726) | > loss_dur: 0.11359 (0.11240) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.68723 (11.69039) | > current_lr: 0.00025 | > step_time: 0.54530 (0.42723) | > loader_time: 0.00320 (0.00293)  --> STEP: 195/639 -- GLOBAL_STEP: 62275 | > loss: -0.47496 (-0.52365) | > log_mle: -0.57303 (-0.63566) | > loss_dur: 0.09807 (0.11202) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.33699 (11.51575) | > current_lr: 0.00025 | > step_time: 0.49160 (0.44266) | > loader_time: 0.00350 (0.00297)  --> STEP: 220/639 -- GLOBAL_STEP: 62300 | > loss: -0.51784 (-0.52094) | > log_mle: -0.62257 (-0.63312) | > loss_dur: 0.10472 (0.11218) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.22088 (11.30399) | > current_lr: 0.00025 | > step_time: 0.63130 (0.45844) | > loader_time: 0.00350 (0.00301)  --> STEP: 245/639 -- GLOBAL_STEP: 62325 | > loss: -0.50163 (-0.52136) | > log_mle: -0.60845 (-0.63371) | > loss_dur: 0.10681 (0.11236) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.00229 (11.45458) | > current_lr: 0.00025 | > step_time: 0.52340 (0.47193) | > loader_time: 0.00330 (0.00305)  --> STEP: 270/639 -- GLOBAL_STEP: 62350 | > loss: -0.55095 (-0.52062) | > log_mle: -0.67487 (-0.63353) | > loss_dur: 0.12392 (0.11291) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.76590 (11.57503) | > current_lr: 0.00025 | > step_time: 0.83470 (0.48817) | > loader_time: 0.00390 (0.00309)  --> STEP: 295/639 -- GLOBAL_STEP: 62375 | > loss: -0.47667 (-0.51849) | > log_mle: -0.57005 (-0.63146) | > loss_dur: 0.09338 (0.11298) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.30374 (11.34706) | > current_lr: 0.00025 | > step_time: 0.64500 (0.50136) | > loader_time: 0.00360 (0.00313)  --> STEP: 320/639 -- GLOBAL_STEP: 62400 | > loss: -0.49074 (-0.51664) | > log_mle: -0.61348 (-0.62998) | > loss_dur: 0.12274 (0.11334) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.29357 (11.18834) | > current_lr: 0.00025 | > step_time: 0.72870 (0.51650) | > loader_time: 0.00370 (0.00316)  --> STEP: 345/639 -- GLOBAL_STEP: 62425 | > loss: -0.49000 (-0.51437) | > log_mle: -0.60299 (-0.62807) | > loss_dur: 0.11299 (0.11370) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.05593 (11.04114) | > current_lr: 0.00025 | > step_time: 0.68650 (0.53096) | > loader_time: 0.00350 (0.00320)  --> STEP: 370/639 -- GLOBAL_STEP: 62450 | > loss: -0.50166 (-0.51225) | > log_mle: -0.61423 (-0.62581) | > loss_dur: 0.11258 (0.11356) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.01701 (10.77914) | > current_lr: 0.00025 | > step_time: 0.88490 (0.54327) | > loader_time: 0.00390 (0.00322)  --> STEP: 395/639 -- GLOBAL_STEP: 62475 | > loss: -0.46068 (-0.51024) | > log_mle: -0.58260 (-0.62380) | > loss_dur: 0.12192 (0.11356) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.39670 (10.59178) | > current_lr: 0.00025 | > step_time: 0.78740 (0.55610) | > loader_time: 0.00360 (0.00325)  --> STEP: 420/639 -- GLOBAL_STEP: 62500 | > loss: -0.46298 (-0.50802) | > log_mle: -0.57732 (-0.62189) | > loss_dur: 0.11434 (0.11387) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.87030 (10.42698) | > current_lr: 0.00025 | > step_time: 0.72040 (0.56904) | > loader_time: 0.00350 (0.00327)  --> STEP: 445/639 -- GLOBAL_STEP: 62525 | > loss: -0.48266 (-0.50645) | > log_mle: -0.58544 (-0.62035) | > loss_dur: 0.10278 (0.11390) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.47370 (10.30168) | > current_lr: 0.00025 | > step_time: 0.76260 (0.58076) | > loader_time: 0.00360 (0.00330)  --> STEP: 470/639 -- GLOBAL_STEP: 62550 | > loss: -0.47977 (-0.50446) | > log_mle: -0.59430 (-0.61838) | > loss_dur: 0.11453 (0.11392) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.81544 (10.13863) | > current_lr: 0.00025 | > step_time: 0.77160 (0.59360) | > loader_time: 0.00360 (0.00332)  --> STEP: 495/639 -- GLOBAL_STEP: 62575 | > loss: -0.43984 (-0.50283) | > log_mle: -0.53436 (-0.61688) | > loss_dur: 0.09452 (0.11406) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.62896 (10.02805) | > current_lr: 0.00025 | > step_time: 0.74070 (0.60637) | > loader_time: 0.00360 (0.00335)  --> STEP: 520/639 -- GLOBAL_STEP: 62600 | > loss: -0.48325 (-0.50153) | > log_mle: -0.59265 (-0.61565) | > loss_dur: 0.10940 (0.11412) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.92633 (9.93305) | > current_lr: 0.00025 | > step_time: 0.80210 (0.61923) | > loader_time: 0.00380 (0.00338)  --> STEP: 545/639 -- GLOBAL_STEP: 62625 | > loss: -0.48995 (-0.49994) | > log_mle: -0.61673 (-0.61428) | > loss_dur: 0.12678 (0.11434) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.18174 (9.90801) | > current_lr: 0.00025 | > step_time: 0.93880 (0.63435) | > loader_time: 0.00390 (0.00342)  --> STEP: 570/639 -- GLOBAL_STEP: 62650 | > loss: -0.46146 (-0.49830) | > log_mle: -0.56911 (-0.61275) | > loss_dur: 0.10765 (0.11445) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.93038 (9.79491) | > current_lr: 0.00025 | > step_time: 0.80220 (0.64813) | > loader_time: 0.00410 (0.00346)  --> STEP: 595/639 -- GLOBAL_STEP: 62675 | > loss: -0.45243 (-0.49710) | > log_mle: -0.56216 (-0.61174) | > loss_dur: 0.10972 (0.11464) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.61820 (9.71395) | > current_lr: 0.00025 | > step_time: 1.00230 (0.66390) | > loader_time: 0.00730 (0.00352)  --> STEP: 620/639 -- GLOBAL_STEP: 62700 | > loss: -0.45214 (-0.49577) | > log_mle: -0.56618 (-0.61055) | > loss_dur: 0.11404 (0.11478) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.75512 (9.69258) | > current_lr: 0.00025 | > step_time: 1.17460 (0.67997) | > loader_time: 0.00460 (0.00357) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03790 (-0.00047) | > avg_loss: -0.45398 (+0.00294) | > avg_log_mle: -0.58871 (+0.00412) | > avg_loss_dur: 0.13473 (-0.00118)  > EPOCH: 98/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 03:22:14)   --> STEP: 5/639 -- GLOBAL_STEP: 62725 | > loss: -0.54685 (-0.55486) | > log_mle: -0.70721 (-0.68307) | > loss_dur: 0.16036 (0.12820) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 21.33979 (15.64698) | > current_lr: 0.00025 | > step_time: 0.30630 (0.28489) | > loader_time: 0.00350 (0.00258)  --> STEP: 30/639 -- GLOBAL_STEP: 62750 | > loss: -0.55039 (-0.54563) | > log_mle: -0.65608 (-0.66201) | > loss_dur: 0.10569 (0.11638) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.77324 (13.49340) | > current_lr: 0.00025 | > step_time: 0.39660 (0.33080) | > loader_time: 0.00270 (0.00257)  --> STEP: 55/639 -- GLOBAL_STEP: 62775 | > loss: -0.49463 (-0.53564) | > log_mle: -0.59201 (-0.64995) | > loss_dur: 0.09738 (0.11431) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.23236 (11.54384) | > current_lr: 0.00025 | > step_time: 0.34730 (0.34898) | > loader_time: 0.00310 (0.00270)  --> STEP: 80/639 -- GLOBAL_STEP: 62800 | > loss: -0.49530 (-0.52947) | > log_mle: -0.62862 (-0.64218) | > loss_dur: 0.13332 (0.11271) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.30666 (10.66334) | > current_lr: 0.00025 | > step_time: 0.40760 (0.36706) | > loader_time: 0.00290 (0.00279)  --> STEP: 105/639 -- GLOBAL_STEP: 62825 | > loss: -0.47613 (-0.52728) | > log_mle: -0.58574 (-0.63956) | > loss_dur: 0.10961 (0.11228) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.75002 (10.50843) | > current_lr: 0.00025 | > step_time: 0.41510 (0.38444) | > loader_time: 0.00300 (0.00287)  --> STEP: 130/639 -- GLOBAL_STEP: 62850 | > loss: -0.49358 (-0.52489) | > log_mle: -0.62264 (-0.63778) | > loss_dur: 0.12906 (0.11288) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.48974 (10.48401) | > current_lr: 0.00025 | > step_time: 0.50150 (0.39947) | > loader_time: 0.00300 (0.00290)  --> STEP: 155/639 -- GLOBAL_STEP: 62875 | > loss: -0.49924 (-0.52367) | > log_mle: -0.60945 (-0.63668) | > loss_dur: 0.11021 (0.11301) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.32915 (10.41901) | > current_lr: 0.00025 | > step_time: 0.53610 (0.41836) | > loader_time: 0.00320 (0.00294)  --> STEP: 180/639 -- GLOBAL_STEP: 62900 | > loss: -0.53885 (-0.52478) | > log_mle: -0.66381 (-0.63809) | > loss_dur: 0.12496 (0.11330) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.16623 (10.79668) | > current_lr: 0.00025 | > step_time: 0.61500 (0.43443) | > loader_time: 0.00330 (0.00298)  --> STEP: 205/639 -- GLOBAL_STEP: 62925 | > loss: -0.47550 (-0.52151) | > log_mle: -0.58984 (-0.63436) | > loss_dur: 0.11433 (0.11285) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.58804 (10.53778) | > current_lr: 0.00025 | > step_time: 0.54790 (0.44801) | > loader_time: 0.00340 (0.00301)  --> STEP: 230/639 -- GLOBAL_STEP: 62950 | > loss: -0.50506 (-0.52115) | > log_mle: -0.62388 (-0.63422) | > loss_dur: 0.11882 (0.11307) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.02744 (10.74679) | > current_lr: 0.00025 | > step_time: 0.55080 (0.46405) | > loader_time: 0.00310 (0.00305)  --> STEP: 255/639 -- GLOBAL_STEP: 62975 | > loss: -0.48377 (-0.52005) | > log_mle: -0.59442 (-0.63305) | > loss_dur: 0.11065 (0.11300) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.15521 (10.73266) | > current_lr: 0.00025 | > step_time: 0.62720 (0.47803) | > loader_time: 0.00330 (0.00308)  --> STEP: 280/639 -- GLOBAL_STEP: 63000 | > loss: -0.48967 (-0.51952) | > log_mle: -0.61674 (-0.63309) | > loss_dur: 0.12708 (0.11357) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.62229 (10.81559) | > current_lr: 0.00025 | > step_time: 0.67590 (0.49491) | > loader_time: 0.00340 (0.00312)  --> STEP: 305/639 -- GLOBAL_STEP: 63025 | > loss: -0.47652 (-0.51751) | > log_mle: -0.59988 (-0.63129) | > loss_dur: 0.12336 (0.11378) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.10944 (10.69256) | > current_lr: 0.00025 | > step_time: 0.74150 (0.50978) | > loader_time: 0.00330 (0.00315)  --> STEP: 330/639 -- GLOBAL_STEP: 63050 | > loss: -0.49545 (-0.51494) | > log_mle: -0.61807 (-0.62919) | > loss_dur: 0.12261 (0.11425) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.70722 (10.58295) | > current_lr: 0.00025 | > step_time: 0.75010 (0.52486) | > loader_time: 0.00380 (0.00318)  --> STEP: 355/639 -- GLOBAL_STEP: 63075 | > loss: -0.44733 (-0.51319) | > log_mle: -0.55095 (-0.62722) | > loss_dur: 0.10362 (0.11403) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.29746 (10.42608) | > current_lr: 0.00025 | > step_time: 0.60660 (0.53913) | > loader_time: 0.00350 (0.00321)  --> STEP: 380/639 -- GLOBAL_STEP: 63100 | > loss: -0.48185 (-0.51108) | > log_mle: -0.58427 (-0.62492) | > loss_dur: 0.10241 (0.11384) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.85245 (10.19453) | > current_lr: 0.00025 | > step_time: 0.66770 (0.55260) | > loader_time: 0.00350 (0.00324)  --> STEP: 405/639 -- GLOBAL_STEP: 63125 | > loss: -0.47873 (-0.50944) | > log_mle: -0.59037 (-0.62360) | > loss_dur: 0.11163 (0.11416) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.84689 (10.15703) | > current_lr: 0.00025 | > step_time: 0.72080 (0.56435) | > loader_time: 0.00360 (0.00326)  --> STEP: 430/639 -- GLOBAL_STEP: 63150 | > loss: -0.44631 (-0.50727) | > log_mle: -0.55054 (-0.62162) | > loss_dur: 0.10423 (0.11434) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.43962 (10.04151) | > current_lr: 0.00025 | > step_time: 0.75470 (0.57677) | > loader_time: 0.00370 (0.00329)  --> STEP: 455/639 -- GLOBAL_STEP: 63175 | > loss: -0.46281 (-0.50547) | > log_mle: -0.57312 (-0.61961) | > loss_dur: 0.11031 (0.11414) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.27449 (9.87773) | > current_lr: 0.00025 | > step_time: 0.73030 (0.58727) | > loader_time: 0.00350 (0.00331)  --> STEP: 480/639 -- GLOBAL_STEP: 63200 | > loss: -0.42109 (-0.50361) | > log_mle: -0.53102 (-0.61797) | > loss_dur: 0.10993 (0.11436) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.63182 (9.76001) | > current_lr: 0.00025 | > step_time: 0.80290 (0.60061) | > loader_time: 0.00370 (0.00333)  --> STEP: 505/639 -- GLOBAL_STEP: 63225 | > loss: -0.48306 (-0.50214) | > log_mle: -0.59734 (-0.61648) | > loss_dur: 0.11427 (0.11434) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.32297 (9.70042) | > current_lr: 0.00025 | > step_time: 0.90130 (0.61324) | > loader_time: 0.00410 (0.00336)  --> STEP: 530/639 -- GLOBAL_STEP: 63250 | > loss: -0.45667 (-0.50068) | > log_mle: -0.58653 (-0.61521) | > loss_dur: 0.12986 (0.11452) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.80744 (9.61687) | > current_lr: 0.00025 | > step_time: 0.91160 (0.62656) | > loader_time: 0.00400 (0.00339)  --> STEP: 555/639 -- GLOBAL_STEP: 63275 | > loss: -0.45514 (-0.49915) | > log_mle: -0.57098 (-0.61369) | > loss_dur: 0.11584 (0.11454) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.85509 (9.51788) | > current_lr: 0.00025 | > step_time: 0.84920 (0.64064) | > loader_time: 0.00400 (0.00342)  --> STEP: 580/639 -- GLOBAL_STEP: 63300 | > loss: -0.48412 (-0.49785) | > log_mle: -0.58991 (-0.61251) | > loss_dur: 0.10579 (0.11466) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.97710 (9.45603) | > current_lr: 0.00025 | > step_time: 0.98860 (0.65530) | > loader_time: 0.00430 (0.00346)  --> STEP: 605/639 -- GLOBAL_STEP: 63325 | > loss: -0.45197 (-0.49658) | > log_mle: -0.55803 (-0.61140) | > loss_dur: 0.10605 (0.11482) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.60422 (9.47573) | > current_lr: 0.00025 | > step_time: 1.04120 (0.67073) | > loader_time: 0.00420 (0.00349)  --> STEP: 630/639 -- GLOBAL_STEP: 63350 | > loss: -0.46582 (-0.49539) | > log_mle: -0.59124 (-0.61042) | > loss_dur: 0.12542 (0.11502) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.37080 (9.49501) | > current_lr: 0.00025 | > step_time: 1.16530 (0.68778) | > loader_time: 0.00460 (0.00353) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02175 (-0.01615) | > avg_loss: -0.45825 (-0.00426) | > avg_log_mle: -0.59421 (-0.00550) | > avg_loss_dur: 0.13597 (+0.00124)  > EPOCH: 99/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 03:29:52)   --> STEP: 15/639 -- GLOBAL_STEP: 63375 | > loss: -0.66685 (-0.55703) | > log_mle: -0.78609 (-0.67308) | > loss_dur: 0.11924 (0.11605) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 22.89254 (13.66055) | > current_lr: 0.00025 | > step_time: 0.32990 (0.29162) | > loader_time: 0.00230 (0.00237)  --> STEP: 40/639 -- GLOBAL_STEP: 63400 | > loss: -0.52802 (-0.54166) | > log_mle: -0.63582 (-0.65480) | > loss_dur: 0.10780 (0.11314) | > amp_scaler: 2048.00000 (2918.40000) | > grad_norm: 12.19044 (11.39264) | > current_lr: 0.00025 | > step_time: 0.34880 (0.32849) | > loader_time: 0.00300 (0.00260)  --> STEP: 65/639 -- GLOBAL_STEP: 63425 | > loss: -0.52364 (-0.53309) | > log_mle: -0.62087 (-0.64675) | > loss_dur: 0.09723 (0.11366) | > amp_scaler: 2048.00000 (2583.63077) | > grad_norm: 6.45606 (11.72062) | > current_lr: 0.00025 | > step_time: 0.38870 (0.35912) | > loader_time: 0.00280 (0.00273)  --> STEP: 90/639 -- GLOBAL_STEP: 63450 | > loss: -0.46734 (-0.52635) | > log_mle: -0.57759 (-0.63794) | > loss_dur: 0.11025 (0.11159) | > amp_scaler: 2048.00000 (2434.84444) | > grad_norm: 4.87490 (10.61157) | > current_lr: 0.00025 | > step_time: 0.51560 (0.37452) | > loader_time: 0.00320 (0.00279)  --> STEP: 115/639 -- GLOBAL_STEP: 63475 | > loss: -0.44427 (-0.52620) | > log_mle: -0.54775 (-0.63818) | > loss_dur: 0.10348 (0.11198) | > amp_scaler: 2048.00000 (2350.74783) | > grad_norm: 3.33352 (10.71343) | > current_lr: 0.00025 | > step_time: 0.40250 (0.39185) | > loader_time: 0.00300 (0.00284)  --> STEP: 140/639 -- GLOBAL_STEP: 63500 | > loss: -0.48963 (-0.52557) | > log_mle: -0.62440 (-0.63834) | > loss_dur: 0.13477 (0.11277) | > amp_scaler: 2048.00000 (2296.68571) | > grad_norm: 10.69291 (10.94180) | > current_lr: 0.00025 | > step_time: 0.55240 (0.40532) | > loader_time: 0.00320 (0.00286)  --> STEP: 165/639 -- GLOBAL_STEP: 63525 | > loss: -0.49791 (-0.52450) | > log_mle: -0.63081 (-0.63714) | > loss_dur: 0.13290 (0.11263) | > amp_scaler: 2048.00000 (2259.00606) | > grad_norm: 13.03470 (10.94472) | > current_lr: 0.00025 | > step_time: 0.49080 (0.41906) | > loader_time: 0.00310 (0.00290)  --> STEP: 190/639 -- GLOBAL_STEP: 63550 | > loss: -0.45385 (-0.52385) | > log_mle: -0.56215 (-0.63654) | > loss_dur: 0.10830 (0.11268) | > amp_scaler: 2048.00000 (2231.24211) | > grad_norm: 2.97483 (11.03485) | > current_lr: 0.00025 | > step_time: 0.52360 (0.43646) | > loader_time: 0.00330 (0.00294)  --> STEP: 215/639 -- GLOBAL_STEP: 63575 | > loss: -0.55319 (-0.52119) | > log_mle: -0.66540 (-0.63416) | > loss_dur: 0.11222 (0.11297) | > amp_scaler: 2048.00000 (2209.93488) | > grad_norm: 13.04163 (10.80982) | > current_lr: 0.00025 | > step_time: 0.53670 (0.45328) | > loader_time: 0.00320 (0.00298)  --> STEP: 240/639 -- GLOBAL_STEP: 63600 | > loss: -0.55455 (-0.52198) | > log_mle: -0.67602 (-0.63476) | > loss_dur: 0.12147 (0.11278) | > amp_scaler: 2048.00000 (2193.06667) | > grad_norm: 18.20767 (10.91373) | > current_lr: 0.00025 | > step_time: 0.55510 (0.46799) | > loader_time: 0.00320 (0.00302)  --> STEP: 265/639 -- GLOBAL_STEP: 63625 | > loss: -0.56605 (-0.52055) | > log_mle: -0.69837 (-0.63394) | > loss_dur: 0.13232 (0.11339) | > amp_scaler: 2048.00000 (2179.38113) | > grad_norm: 25.13210 (11.15542) | > current_lr: 0.00025 | > step_time: 0.72410 (0.48343) | > loader_time: 0.00340 (0.00306)  --> STEP: 290/639 -- GLOBAL_STEP: 63650 | > loss: -0.48452 (-0.51878) | > log_mle: -0.59435 (-0.63244) | > loss_dur: 0.10983 (0.11366) | > amp_scaler: 2048.00000 (2168.05517) | > grad_norm: 8.85119 (11.02007) | > current_lr: 0.00025 | > step_time: 0.66750 (0.49909) | > loader_time: 0.00330 (0.00310)  --> STEP: 315/639 -- GLOBAL_STEP: 63675 | > loss: -0.46892 (-0.51704) | > log_mle: -0.57361 (-0.63094) | > loss_dur: 0.10469 (0.11390) | > amp_scaler: 2048.00000 (2158.52698) | > grad_norm: 5.50021 (10.93667) | > current_lr: 0.00025 | > step_time: 0.70640 (0.51368) | > loader_time: 0.00350 (0.00313)  --> STEP: 340/639 -- GLOBAL_STEP: 63700 | > loss: -0.49088 (-0.51506) | > log_mle: -0.59513 (-0.62950) | > loss_dur: 0.10425 (0.11444) | > amp_scaler: 2048.00000 (2150.40000) | > grad_norm: 8.06911 (10.93319) | > current_lr: 0.00025 | > step_time: 0.62330 (0.52994) | > loader_time: 0.00340 (0.00317)  --> STEP: 365/639 -- GLOBAL_STEP: 63725 | > loss: -0.48437 (-0.51298) | > log_mle: -0.60290 (-0.62709) | > loss_dur: 0.11853 (0.11411) | > amp_scaler: 2048.00000 (2143.38630) | > grad_norm: 10.16680 (10.73614) | > current_lr: 0.00025 | > step_time: 0.83210 (0.54158) | > loader_time: 0.00390 (0.00320)  --> STEP: 390/639 -- GLOBAL_STEP: 63750 | > loss: -0.47624 (-0.51104) | > log_mle: -0.60011 (-0.62505) | > loss_dur: 0.12387 (0.11400) | > amp_scaler: 2048.00000 (2137.27179) | > grad_norm: 7.62780 (10.55124) | > current_lr: 0.00025 | > step_time: 0.74420 (0.55441) | > loader_time: 0.00350 (0.00323)  --> STEP: 415/639 -- GLOBAL_STEP: 63775 | > loss: -0.45322 (-0.50893) | > log_mle: -0.57353 (-0.62325) | > loss_dur: 0.12031 (0.11432) | > amp_scaler: 2048.00000 (2131.89398) | > grad_norm: 6.76092 (10.48088) | > current_lr: 0.00025 | > step_time: 0.73640 (0.56731) | > loader_time: 0.00350 (0.00326)  --> STEP: 440/639 -- GLOBAL_STEP: 63800 | > loss: -0.44641 (-0.50724) | > log_mle: -0.56066 (-0.62161) | > loss_dur: 0.11426 (0.11436) | > amp_scaler: 2048.00000 (2127.12727) | > grad_norm: 6.39102 (10.39346) | > current_lr: 0.00025 | > step_time: 0.74130 (0.57938) | > loader_time: 0.00370 (0.00328)  --> STEP: 465/639 -- GLOBAL_STEP: 63825 | > loss: -0.45600 (-0.50522) | > log_mle: -0.56499 (-0.61951) | > loss_dur: 0.10899 (0.11428) | > amp_scaler: 2048.00000 (2122.87312) | > grad_norm: 4.19483 (10.23938) | > current_lr: 0.00025 | > step_time: 0.76640 (0.58990) | > loader_time: 0.00390 (0.00331)  --> STEP: 490/639 -- GLOBAL_STEP: 63850 | > loss: -0.44509 (-0.50341) | > log_mle: -0.54703 (-0.61785) | > loss_dur: 0.10194 (0.11445) | > amp_scaler: 2048.00000 (2119.05306) | > grad_norm: 3.68917 (10.12622) | > current_lr: 0.00025 | > step_time: 0.79580 (0.60248) | > loader_time: 0.00360 (0.00333)  --> STEP: 515/639 -- GLOBAL_STEP: 63875 | > loss: -0.46535 (-0.50211) | > log_mle: -0.57545 (-0.61665) | > loss_dur: 0.11010 (0.11454) | > amp_scaler: 2048.00000 (2115.60388) | > grad_norm: 7.42832 (10.08739) | > current_lr: 0.00025 | > step_time: 0.98250 (0.61537) | > loader_time: 0.00430 (0.00337)  --> STEP: 540/639 -- GLOBAL_STEP: 63900 | > loss: -0.47073 (-0.50054) | > log_mle: -0.59299 (-0.61519) | > loss_dur: 0.12226 (0.11465) | > amp_scaler: 2048.00000 (2112.47407) | > grad_norm: 6.28045 (9.97654) | > current_lr: 0.00025 | > step_time: 0.87230 (0.62970) | > loader_time: 0.00400 (0.00340)  --> STEP: 565/639 -- GLOBAL_STEP: 63925 | > loss: -0.44324 (-0.49887) | > log_mle: -0.55198 (-0.61365) | > loss_dur: 0.10874 (0.11477) | > amp_scaler: 2048.00000 (2109.62124) | > grad_norm: 5.18582 (9.94154) | > current_lr: 0.00025 | > step_time: 0.98670 (0.64354) | > loader_time: 0.00430 (0.00343)  --> STEP: 590/639 -- GLOBAL_STEP: 63950 | > loss: -0.46313 (-0.49773) | > log_mle: -0.58446 (-0.61265) | > loss_dur: 0.12133 (0.11491) | > amp_scaler: 2048.00000 (2107.01017) | > grad_norm: 10.03459 (9.93514) | > current_lr: 0.00025 | > step_time: 0.94910 (0.65888) | > loader_time: 0.00420 (0.00347)  --> STEP: 615/639 -- GLOBAL_STEP: 63975 | > loss: -0.46791 (-0.49640) | > log_mle: -0.58879 (-0.61145) | > loss_dur: 0.12088 (0.11505) | > amp_scaler: 2048.00000 (2104.61138) | > grad_norm: 5.92764 (9.84936) | > current_lr: 0.00025 | > step_time: 0.98010 (0.67470) | > loader_time: 0.00460 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03246 (+0.01072) | > avg_loss: -0.45532 (+0.00293) | > avg_log_mle: -0.59201 (+0.00220) | > avg_loss_dur: 0.13669 (+0.00072)  > EPOCH: 100/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 03:37:28)   --> STEP: 0/639 -- GLOBAL_STEP: 64000 | > loss: -0.58635 (-0.58635) | > log_mle: -0.72831 (-0.72831) | > loss_dur: 0.14196 (0.14196) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 21.14888 (21.14888) | > current_lr: 0.00025 | > step_time: 0.58010 (0.58006) | > loader_time: 0.43280 (0.43280)  --> STEP: 25/639 -- GLOBAL_STEP: 64025 | > loss: -0.54363 (-0.55623) | > log_mle: -0.67247 (-0.67225) | > loss_dur: 0.12885 (0.11602) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 27.00972 (15.94493) | > current_lr: 0.00025 | > step_time: 0.38190 (0.32513) | > loader_time: 0.00270 (0.00258)  --> STEP: 50/639 -- GLOBAL_STEP: 64050 | > loss: -0.50203 (-0.54346) | > log_mle: -0.60153 (-0.65720) | > loss_dur: 0.09950 (0.11373) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.60947 (12.85841) | > current_lr: 0.00025 | > step_time: 0.31220 (0.34856) | > loader_time: 0.00270 (0.00269)  --> STEP: 75/639 -- GLOBAL_STEP: 64075 | > loss: -0.53034 (-0.53483) | > log_mle: -0.63187 (-0.64725) | > loss_dur: 0.10153 (0.11242) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.54222 (11.47662) | > current_lr: 0.00025 | > step_time: 0.40170 (0.36844) | > loader_time: 0.00280 (0.00276)  --> STEP: 100/639 -- GLOBAL_STEP: 64100 | > loss: -0.52788 (-0.53236) | > log_mle: -0.63860 (-0.64385) | > loss_dur: 0.11072 (0.11149) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.97550 (10.82882) | > current_lr: 0.00025 | > step_time: 0.50890 (0.38559) | > loader_time: 0.00290 (0.00282)  --> STEP: 125/639 -- GLOBAL_STEP: 64125 | > loss: -0.51190 (-0.52937) | > log_mle: -0.63325 (-0.64156) | > loss_dur: 0.12135 (0.11219) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.61802 (10.90033) | > current_lr: 0.00025 | > step_time: 0.41750 (0.39978) | > loader_time: 0.00250 (0.00285)  --> STEP: 150/639 -- GLOBAL_STEP: 64150 | > loss: -0.48605 (-0.52771) | > log_mle: -0.59698 (-0.64025) | > loss_dur: 0.11093 (0.11254) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.84195 (10.76670) | > current_lr: 0.00025 | > step_time: 0.42370 (0.41585) | > loader_time: 0.00270 (0.00288)  --> STEP: 175/639 -- GLOBAL_STEP: 64175 | > loss: -0.51324 (-0.52739) | > log_mle: -0.61281 (-0.64001) | > loss_dur: 0.09956 (0.11262) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.63280 (10.82495) | > current_lr: 0.00025 | > step_time: 0.55570 (0.43076) | > loader_time: 0.00320 (0.00291)  --> STEP: 200/639 -- GLOBAL_STEP: 64200 | > loss: -0.48611 (-0.52497) | > log_mle: -0.59944 (-0.63722) | > loss_dur: 0.11333 (0.11225) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.02459 (10.57685) | > current_lr: 0.00025 | > step_time: 0.52960 (0.44546) | > loader_time: 0.00320 (0.00294)  --> STEP: 225/639 -- GLOBAL_STEP: 64225 | > loss: -0.48281 (-0.52236) | > log_mle: -0.57750 (-0.63460) | > loss_dur: 0.09468 (0.11224) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.69974 (10.34140) | > current_lr: 0.00025 | > step_time: 0.52020 (0.46146) | > loader_time: 0.00320 (0.00298)  --> STEP: 250/639 -- GLOBAL_STEP: 64250 | > loss: -0.48175 (-0.52286) | > log_mle: -0.58335 (-0.63532) | > loss_dur: 0.10160 (0.11247) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.97411 (10.46674) | > current_lr: 0.00025 | > step_time: 0.57940 (0.47701) | > loader_time: 0.00350 (0.00301)  --> STEP: 275/639 -- GLOBAL_STEP: 64275 | > loss: -0.50722 (-0.52230) | > log_mle: -0.61610 (-0.63543) | > loss_dur: 0.10888 (0.11314) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.24022 (10.66013) | > current_lr: 0.00025 | > step_time: 0.84550 (0.49435) | > loader_time: 0.00380 (0.00305)  --> STEP: 300/639 -- GLOBAL_STEP: 64300 | > loss: -0.49941 (-0.51992) | > log_mle: -0.61165 (-0.63325) | > loss_dur: 0.11223 (0.11333) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.21039 (10.52744) | > current_lr: 0.00025 | > step_time: 0.72340 (0.50794) | > loader_time: 0.00360 (0.00308)  --> STEP: 325/639 -- GLOBAL_STEP: 64325 | > loss: -0.49703 (-0.51774) | > log_mle: -0.61285 (-0.63153) | > loss_dur: 0.11582 (0.11379) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.75371 (10.47474) | > current_lr: 0.00025 | > step_time: 0.61610 (0.52300) | > loader_time: 0.00350 (0.00312)  --> STEP: 350/639 -- GLOBAL_STEP: 64350 | > loss: -0.51265 (-0.51553) | > log_mle: -0.63745 (-0.62943) | > loss_dur: 0.12480 (0.11390) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.45038 (10.28469) | > current_lr: 0.00025 | > step_time: 0.80180 (0.53718) | > loader_time: 0.00400 (0.00315)  --> STEP: 375/639 -- GLOBAL_STEP: 64375 | > loss: -0.46994 (-0.51331) | > log_mle: -0.57579 (-0.62696) | > loss_dur: 0.10585 (0.11365) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.91259 (10.06625) | > current_lr: 0.00025 | > step_time: 0.74780 (0.55079) | > loader_time: 0.00360 (0.00318)  --> STEP: 400/639 -- GLOBAL_STEP: 64400 | > loss: -0.49614 (-0.51160) | > log_mle: -0.60389 (-0.62541) | > loss_dur: 0.10775 (0.11381) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.18675 (10.00519) | > current_lr: 0.00025 | > step_time: 0.75470 (0.56263) | > loader_time: 0.00370 (0.00320)  --> STEP: 425/639 -- GLOBAL_STEP: 64425 | > loss: -0.53492 (-0.50937) | > log_mle: -0.65917 (-0.62327) | > loss_dur: 0.12425 (0.11390) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 20.30401 (9.90319) | > current_lr: 0.00025 | > step_time: 0.90360 (0.57442) | > loader_time: 0.00380 (0.00323)  --> STEP: 450/639 -- GLOBAL_STEP: 64450 | > loss: -0.45884 (-0.50752) | > log_mle: -0.56102 (-0.62139) | > loss_dur: 0.10218 (0.11387) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.80203 (9.79736) | > current_lr: 0.00025 | > step_time: 0.77650 (0.58630) | > loader_time: 0.00400 (0.00326)  --> STEP: 475/639 -- GLOBAL_STEP: 64475 | > loss: -0.47054 (-0.50561) | > log_mle: -0.57240 (-0.61960) | > loss_dur: 0.10186 (0.11399) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.82464 (9.68749) | > current_lr: 0.00025 | > step_time: 0.68790 (0.59806) | > loader_time: 0.00340 (0.00328)  --> STEP: 500/639 -- GLOBAL_STEP: 64500 | > loss: -0.51052 (-0.50413) | > log_mle: -0.63320 (-0.61829) | > loss_dur: 0.12267 (0.11416) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.20673 (9.60754) | > current_lr: 0.00025 | > step_time: 0.93540 (0.61018) | > loader_time: 0.00400 (0.00331)  --> STEP: 525/639 -- GLOBAL_STEP: 64525 | > loss: -0.48337 (-0.50269) | > log_mle: -0.61634 (-0.61698) | > loss_dur: 0.13297 (0.11429) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.14275 (9.51895) | > current_lr: 0.00025 | > step_time: 1.03220 (0.62399) | > loader_time: 0.00420 (0.00334)  --> STEP: 550/639 -- GLOBAL_STEP: 64550 | > loss: -0.43293 (-0.50107) | > log_mle: -0.54775 (-0.61542) | > loss_dur: 0.11482 (0.11434) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.89358 (9.43265) | > current_lr: 0.00025 | > step_time: 0.92400 (0.63876) | > loader_time: 0.00410 (0.00338)  --> STEP: 575/639 -- GLOBAL_STEP: 64575 | > loss: -0.45699 (-0.49944) | > log_mle: -0.56579 (-0.61391) | > loss_dur: 0.10880 (0.11446) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.38184 (9.34050) | > current_lr: 0.00025 | > step_time: 0.89310 (0.65212) | > loader_time: 0.00420 (0.00342)  --> STEP: 600/639 -- GLOBAL_STEP: 64600 | > loss: -0.44795 (-0.49820) | > log_mle: -0.56375 (-0.61283) | > loss_dur: 0.11580 (0.11463) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.68374 (9.29858) | > current_lr: 0.00025 | > step_time: 1.03170 (0.66831) | > loader_time: 0.00420 (0.00347)  --> STEP: 625/639 -- GLOBAL_STEP: 64625 | > loss: -0.47855 (-0.49697) | > log_mle: -0.59451 (-0.61170) | > loss_dur: 0.11596 (0.11473) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.12281 (9.27946) | > current_lr: 0.00025 | > step_time: 1.07750 (0.68495) | > loader_time: 0.00680 (0.00353) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02592 (-0.00654) | > avg_loss: -0.45624 (-0.00092) | > avg_log_mle: -0.59290 (-0.00089) | > avg_loss_dur: 0.13667 (-0.00002)  > EPOCH: 101/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 03:45:06)   --> STEP: 10/639 -- GLOBAL_STEP: 64650 | > loss: -0.53698 (-0.55640) | > log_mle: -0.67052 (-0.67626) | > loss_dur: 0.13354 (0.11986) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.38591 (12.56304) | > current_lr: 0.00025 | > step_time: 0.28450 (0.28849) | > loader_time: 0.00250 (0.00241)  --> STEP: 35/639 -- GLOBAL_STEP: 64675 | > loss: -0.54688 (-0.54641) | > log_mle: -0.63871 (-0.66045) | > loss_dur: 0.09183 (0.11404) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.10437 (11.29764) | > current_lr: 0.00025 | > step_time: 0.35090 (0.32920) | > loader_time: 0.00270 (0.00261)  --> STEP: 60/639 -- GLOBAL_STEP: 64700 | > loss: -0.49912 (-0.53630) | > log_mle: -0.59328 (-0.65002) | > loss_dur: 0.09415 (0.11372) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.89188 (10.78102) | > current_lr: 0.00025 | > step_time: 0.34160 (0.35795) | > loader_time: 0.00260 (0.00272)  --> STEP: 85/639 -- GLOBAL_STEP: 64725 | > loss: -0.49710 (-0.52890) | > log_mle: -0.59404 (-0.64058) | > loss_dur: 0.09694 (0.11168) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.31466 (10.06592) | > current_lr: 0.00025 | > step_time: 0.50200 (0.37422) | > loader_time: 0.00290 (0.00275)  --> STEP: 110/639 -- GLOBAL_STEP: 64750 | > loss: -0.50202 (-0.52894) | > log_mle: -0.60960 (-0.64170) | > loss_dur: 0.10759 (0.11276) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.59189 (10.65061) | > current_lr: 0.00025 | > step_time: 0.40300 (0.39334) | > loader_time: 0.00290 (0.00282)  --> STEP: 135/639 -- GLOBAL_STEP: 64775 | > loss: -0.46715 (-0.52632) | > log_mle: -0.58659 (-0.63912) | > loss_dur: 0.11944 (0.11280) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.27653 (10.48504) | > current_lr: 0.00025 | > step_time: 0.46940 (0.40532) | > loader_time: 0.00300 (0.00285)  --> STEP: 160/639 -- GLOBAL_STEP: 64800 | > loss: -0.53400 (-0.52482) | > log_mle: -0.62793 (-0.63711) | > loss_dur: 0.09393 (0.11229) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.14294 (10.40963) | > current_lr: 0.00025 | > step_time: 0.47380 (0.42186) | > loader_time: 0.00300 (0.00289)  --> STEP: 185/639 -- GLOBAL_STEP: 64825 | > loss: -0.50607 (-0.52643) | > log_mle: -0.61182 (-0.63912) | > loss_dur: 0.10575 (0.11268) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.85302 (10.91864) | > current_lr: 0.00025 | > step_time: 0.55350 (0.43970) | > loader_time: 0.00340 (0.00294)  --> STEP: 210/639 -- GLOBAL_STEP: 64850 | > loss: -0.43027 (-0.52314) | > log_mle: -0.52991 (-0.63588) | > loss_dur: 0.09965 (0.11274) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.91020 (10.78034) | > current_lr: 0.00025 | > step_time: 0.49910 (0.45692) | > loader_time: 0.00300 (0.00298)  --> STEP: 235/639 -- GLOBAL_STEP: 64875 | > loss: -0.51551 (-0.52321) | > log_mle: -0.61984 (-0.63580) | > loss_dur: 0.10433 (0.11259) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.82501 (10.92678) | > current_lr: 0.00025 | > step_time: 0.54880 (0.47142) | > loader_time: 0.00330 (0.00302)  --> STEP: 260/639 -- GLOBAL_STEP: 64900 | > loss: -0.49468 (-0.52217) | > log_mle: -0.62150 (-0.63510) | > loss_dur: 0.12682 (0.11292) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.23977 (11.11082) | > current_lr: 0.00025 | > step_time: 0.57660 (0.48575) | > loader_time: 0.00320 (0.00305)  --> STEP: 285/639 -- GLOBAL_STEP: 64925 | > loss: -0.45510 (-0.52086) | > log_mle: -0.56916 (-0.63400) | > loss_dur: 0.11406 (0.11314) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.41958 (11.09474) | > current_lr: 0.00025 | > step_time: 0.60230 (0.50054) | > loader_time: 0.00340 (0.00309)  --> STEP: 310/639 -- GLOBAL_STEP: 64950 | > loss: -0.49225 (-0.51896) | > log_mle: -0.60959 (-0.63235) | > loss_dur: 0.11734 (0.11340) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.75778 (11.04706) | > current_lr: 0.00025 | > step_time: 0.58960 (0.51318) | > loader_time: 0.00330 (0.00311)  --> STEP: 335/639 -- GLOBAL_STEP: 64975 | > loss: -0.50591 (-0.51642) | > log_mle: -0.63343 (-0.63028) | > loss_dur: 0.12752 (0.11386) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.61771 (10.88931) | > current_lr: 0.00025 | > step_time: 0.66150 (0.52844) | > loader_time: 0.00340 (0.00314)  --> STEP: 360/639 -- GLOBAL_STEP: 65000 | > loss: -0.46593 (-0.51457) | > log_mle: -0.57349 (-0.62827) | > loss_dur: 0.10756 (0.11370) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.29260 (10.75803) | > current_lr: 0.00025 | > step_time: 0.66560 (0.54041) | > loader_time: 0.00380 (0.00317)  --> STEP: 385/639 -- GLOBAL_STEP: 65025 | > loss: -0.49316 (-0.51284) | > log_mle: -0.60269 (-0.62637) | > loss_dur: 0.10953 (0.11352) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.43799 (10.59165) | > current_lr: 0.00025 | > step_time: 0.67210 (0.55326) | > loader_time: 0.00380 (0.00320)  --> STEP: 410/639 -- GLOBAL_STEP: 65050 | > loss: -0.45491 (-0.51083) | > log_mle: -0.57853 (-0.62462) | > loss_dur: 0.12361 (0.11379) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.15252 (10.45222) | > current_lr: 0.00025 | > step_time: 0.75700 (0.56529) | > loader_time: 0.00390 (0.00323)  --> STEP: 435/639 -- GLOBAL_STEP: 65075 | > loss: -0.48799 (-0.50872) | > log_mle: -0.59097 (-0.62258) | > loss_dur: 0.10298 (0.11385) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.66700 (10.29473) | > current_lr: 0.00025 | > step_time: 0.69190 (0.57750) | > loader_time: 0.00340 (0.00325)  --> STEP: 460/639 -- GLOBAL_STEP: 65100 | > loss: -0.43965 (-0.50672) | > log_mle: -0.55539 (-0.62053) | > loss_dur: 0.11574 (0.11381) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.37782 (10.12563) | > current_lr: 0.00025 | > step_time: 0.88810 (0.58923) | > loader_time: 0.00400 (0.00327)  --> STEP: 485/639 -- GLOBAL_STEP: 65125 | > loss: -0.47863 (-0.50499) | > log_mle: -0.58832 (-0.61905) | > loss_dur: 0.10969 (0.11405) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.33669 (10.05938) | > current_lr: 0.00025 | > step_time: 0.84390 (0.60177) | > loader_time: 0.00430 (0.00330)  --> STEP: 510/639 -- GLOBAL_STEP: 65150 | > loss: -0.46275 (-0.50359) | > log_mle: -0.58877 (-0.61771) | > loss_dur: 0.12601 (0.11411) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.43121 (9.93630) | > current_lr: 0.00025 | > step_time: 0.84900 (0.61467) | > loader_time: 0.00400 (0.00333)  --> STEP: 535/639 -- GLOBAL_STEP: 65175 | > loss: -0.45396 (-0.50202) | > log_mle: -0.57905 (-0.61624) | > loss_dur: 0.12509 (0.11423) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.94365 (9.82833) | > current_lr: 0.00025 | > step_time: 0.88870 (0.62797) | > loader_time: 0.00410 (0.00336)  --> STEP: 560/639 -- GLOBAL_STEP: 65200 | > loss: -0.46370 (-0.50052) | > log_mle: -0.59031 (-0.61489) | > loss_dur: 0.12661 (0.11437) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.81814 (9.73903) | > current_lr: 0.00025 | > step_time: 0.89130 (0.64234) | > loader_time: 0.00410 (0.00340)  --> STEP: 585/639 -- GLOBAL_STEP: 65225 | > loss: -0.48002 (-0.49918) | > log_mle: -0.59197 (-0.61364) | > loss_dur: 0.11195 (0.11446) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.71485 (9.66778) | > current_lr: 0.00025 | > step_time: 1.03620 (0.65689) | > loader_time: 0.00460 (0.00344)  --> STEP: 610/639 -- GLOBAL_STEP: 65250 | > loss: -0.44616 (-0.49785) | > log_mle: -0.56027 (-0.61247) | > loss_dur: 0.11411 (0.11463) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.42707 (9.58824) | > current_lr: 0.00025 | > step_time: 1.03930 (0.67316) | > loader_time: 0.00450 (0.00347)  --> STEP: 635/639 -- GLOBAL_STEP: 65275 | > loss: -0.46229 (-0.49672) | > log_mle: -0.58428 (-0.61167) | > loss_dur: 0.12200 (0.11495) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.16450 (9.55023) | > current_lr: 0.00025 | > step_time: 1.12610 (0.68972) | > loader_time: 0.00430 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02606 (+0.00013) | > avg_loss: -0.45144 (+0.00480) | > avg_log_mle: -0.58927 (+0.00364) | > avg_loss_dur: 0.13783 (+0.00116)  > EPOCH: 102/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 03:52:43)   --> STEP: 20/639 -- GLOBAL_STEP: 65300 | > loss: -0.65904 (-0.56544) | > log_mle: -0.79860 (-0.68412) | > loss_dur: 0.13956 (0.11868) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 23.07527 (12.99608) | > current_lr: 0.00025 | > step_time: 0.34460 (0.31021) | > loader_time: 0.00270 (0.00256)  --> STEP: 45/639 -- GLOBAL_STEP: 65325 | > loss: -0.51385 (-0.54734) | > log_mle: -0.60598 (-0.66152) | > loss_dur: 0.09213 (0.11417) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.09195 (11.76788) | > current_lr: 0.00025 | > step_time: 0.40650 (0.33867) | > loader_time: 0.00290 (0.00273)  --> STEP: 70/639 -- GLOBAL_STEP: 65350 | > loss: -0.52929 (-0.53744) | > log_mle: -0.65232 (-0.65090) | > loss_dur: 0.12302 (0.11346) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.91276 (11.33783) | > current_lr: 0.00025 | > step_time: 0.39310 (0.36357) | > loader_time: 0.00290 (0.00282)  --> STEP: 95/639 -- GLOBAL_STEP: 65375 | > loss: -0.59783 (-0.53397) | > log_mle: -0.70750 (-0.64518) | > loss_dur: 0.10968 (0.11122) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.25065 (10.88897) | > current_lr: 0.00025 | > step_time: 0.40540 (0.38017) | > loader_time: 0.00300 (0.00286)  --> STEP: 120/639 -- GLOBAL_STEP: 65400 | > loss: -0.56370 (-0.53110) | > log_mle: -0.69625 (-0.64311) | > loss_dur: 0.13255 (0.11201) | > amp_scaler: 4096.00000 (2440.53333) | > grad_norm: 18.43402 (10.78078) | > current_lr: 0.00025 | > step_time: 0.54720 (0.39576) | > loader_time: 0.00320 (0.00290)  --> STEP: 145/639 -- GLOBAL_STEP: 65425 | > loss: -0.52319 (-0.52992) | > log_mle: -0.64676 (-0.64263) | > loss_dur: 0.12357 (0.11271) | > amp_scaler: 4096.00000 (2725.95862) | > grad_norm: 11.95532 (10.93658) | > current_lr: 0.00025 | > step_time: 0.47030 (0.41036) | > loader_time: 0.00330 (0.00294)  --> STEP: 170/639 -- GLOBAL_STEP: 65450 | > loss: -0.51344 (-0.52814) | > log_mle: -0.63135 (-0.64093) | > loss_dur: 0.11791 (0.11280) | > amp_scaler: 4096.00000 (2927.43529) | > grad_norm: 8.61285 (10.91591) | > current_lr: 0.00025 | > step_time: 0.54500 (0.42534) | > loader_time: 0.00300 (0.00296)  --> STEP: 195/639 -- GLOBAL_STEP: 65475 | > loss: -0.47838 (-0.52639) | > log_mle: -0.57416 (-0.63907) | > loss_dur: 0.09577 (0.11268) | > amp_scaler: 4096.00000 (3077.25128) | > grad_norm: 2.22572 (10.78395) | > current_lr: 0.00025 | > step_time: 0.49030 (0.44202) | > loader_time: 0.00320 (0.00299)  --> STEP: 220/639 -- GLOBAL_STEP: 65500 | > loss: -0.52348 (-0.52346) | > log_mle: -0.62586 (-0.63618) | > loss_dur: 0.10238 (0.11272) | > amp_scaler: 4096.00000 (3193.01818) | > grad_norm: 6.62315 (10.72111) | > current_lr: 0.00025 | > step_time: 0.62940 (0.45922) | > loader_time: 0.00340 (0.00303)  --> STEP: 245/639 -- GLOBAL_STEP: 65525 | > loss: -0.50932 (-0.52375) | > log_mle: -0.61202 (-0.63652) | > loss_dur: 0.10270 (0.11276) | > amp_scaler: 4096.00000 (3285.15918) | > grad_norm: 9.21190 (10.92924) | > current_lr: 0.00025 | > step_time: 0.52340 (0.47251) | > loader_time: 0.00340 (0.00306)  --> STEP: 270/639 -- GLOBAL_STEP: 65550 | > loss: -0.54116 (-0.52300) | > log_mle: -0.67300 (-0.63622) | > loss_dur: 0.13184 (0.11322) | > amp_scaler: 4096.00000 (3360.23704) | > grad_norm: 27.29223 (11.21879) | > current_lr: 0.00025 | > step_time: 0.83390 (0.48914) | > loader_time: 0.00360 (0.00309)  --> STEP: 295/639 -- GLOBAL_STEP: 65575 | > loss: -0.47629 (-0.52062) | > log_mle: -0.57171 (-0.63406) | > loss_dur: 0.09542 (0.11343) | > amp_scaler: 4096.00000 (3422.58983) | > grad_norm: 5.62166 (11.04796) | > current_lr: 0.00025 | > step_time: 0.64570 (0.50299) | > loader_time: 0.00350 (0.00312)  --> STEP: 320/639 -- GLOBAL_STEP: 65600 | > loss: -0.48820 (-0.51870) | > log_mle: -0.61225 (-0.63252) | > loss_dur: 0.12406 (0.11382) | > amp_scaler: 4096.00000 (3475.20000) | > grad_norm: 10.73452 (11.01928) | > current_lr: 0.00025 | > step_time: 0.72140 (0.51814) | > loader_time: 0.00380 (0.00315)  --> STEP: 345/639 -- GLOBAL_STEP: 65625 | > loss: -0.49468 (-0.51649) | > log_mle: -0.60601 (-0.63061) | > loss_dur: 0.11133 (0.11413) | > amp_scaler: 4096.00000 (3520.18551) | > grad_norm: 6.92577 (10.88636) | > current_lr: 0.00025 | > step_time: 0.64790 (0.53193) | > loader_time: 0.00350 (0.00319)  --> STEP: 370/639 -- GLOBAL_STEP: 65650 | > loss: -0.49441 (-0.51446) | > log_mle: -0.61465 (-0.62835) | > loss_dur: 0.12025 (0.11389) | > amp_scaler: 4096.00000 (3559.09189) | > grad_norm: 11.99046 (10.71377) | > current_lr: 0.00025 | > step_time: 0.90180 (0.54370) | > loader_time: 0.00390 (0.00321)  --> STEP: 395/639 -- GLOBAL_STEP: 65675 | > loss: -0.47205 (-0.51240) | > log_mle: -0.58697 (-0.62627) | > loss_dur: 0.11491 (0.11387) | > amp_scaler: 4096.00000 (3593.07342) | > grad_norm: 6.86678 (10.50907) | > current_lr: 0.00025 | > step_time: 0.70180 (0.55541) | > loader_time: 0.00340 (0.00323)  --> STEP: 420/639 -- GLOBAL_STEP: 65700 | > loss: -0.45631 (-0.51010) | > log_mle: -0.57811 (-0.62425) | > loss_dur: 0.12180 (0.11415) | > amp_scaler: 4096.00000 (3623.00952) | > grad_norm: 6.61429 (10.36135) | > current_lr: 0.00025 | > step_time: 0.72300 (0.56921) | > loader_time: 0.00340 (0.00326)  --> STEP: 445/639 -- GLOBAL_STEP: 65725 | > loss: -0.47948 (-0.50844) | > log_mle: -0.58864 (-0.62261) | > loss_dur: 0.10916 (0.11417) | > amp_scaler: 4096.00000 (3649.58202) | > grad_norm: 6.82994 (10.25017) | > current_lr: 0.00025 | > step_time: 0.74640 (0.58104) | > loader_time: 0.00370 (0.00329)  --> STEP: 470/639 -- GLOBAL_STEP: 65750 | > loss: -0.48771 (-0.50646) | > log_mle: -0.59724 (-0.62061) | > loss_dur: 0.10953 (0.11415) | > amp_scaler: 4096.00000 (3673.32766) | > grad_norm: 8.87454 (10.09429) | > current_lr: 0.00025 | > step_time: 0.75770 (0.59272) | > loader_time: 0.00380 (0.00331)  --> STEP: 495/639 -- GLOBAL_STEP: 65775 | > loss: -0.43379 (-0.50472) | > log_mle: -0.53557 (-0.61903) | > loss_dur: 0.10178 (0.11431) | > amp_scaler: 4096.00000 (3694.67475) | > grad_norm: 2.53852 (9.96570) | > current_lr: 0.00025 | > step_time: 0.76360 (0.60486) | > loader_time: 0.00360 (0.00333)  --> STEP: 520/639 -- GLOBAL_STEP: 65800 | > loss: -0.48105 (-0.50346) | > log_mle: -0.59358 (-0.61778) | > loss_dur: 0.11254 (0.11433) | > amp_scaler: 4096.00000 (3713.96923) | > grad_norm: 7.40541 (9.93438) | > current_lr: 0.00025 | > step_time: 0.86970 (0.61865) | > loader_time: 0.00390 (0.00336)  --> STEP: 545/639 -- GLOBAL_STEP: 65825 | > loss: -0.49389 (-0.50196) | > log_mle: -0.62309 (-0.61645) | > loss_dur: 0.12920 (0.11449) | > amp_scaler: 4096.00000 (3731.49358) | > grad_norm: 11.18229 (9.86048) | > current_lr: 0.00025 | > step_time: 0.90850 (0.63342) | > loader_time: 0.00430 (0.00340)  --> STEP: 570/639 -- GLOBAL_STEP: 65850 | > loss: -0.46289 (-0.50030) | > log_mle: -0.56883 (-0.61488) | > loss_dur: 0.10594 (0.11458) | > amp_scaler: 4096.00000 (3747.48070) | > grad_norm: 5.36513 (9.74646) | > current_lr: 0.00025 | > step_time: 0.88040 (0.64692) | > loader_time: 0.00410 (0.00344)  --> STEP: 595/639 -- GLOBAL_STEP: 65875 | > loss: -0.45821 (-0.49917) | > log_mle: -0.56467 (-0.61386) | > loss_dur: 0.10645 (0.11469) | > amp_scaler: 4096.00000 (3762.12437) | > grad_norm: 5.62742 (9.66720) | > current_lr: 0.00025 | > step_time: 0.94780 (0.66348) | > loader_time: 0.00410 (0.00348)  --> STEP: 620/639 -- GLOBAL_STEP: 65900 | > loss: -0.45823 (-0.49782) | > log_mle: -0.56969 (-0.61265) | > loss_dur: 0.11146 (0.11482) | > amp_scaler: 4096.00000 (3775.58710) | > grad_norm: 7.51084 (9.64636) | > current_lr: 0.00025 | > step_time: 1.16570 (0.67984) | > loader_time: 0.00470 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03509 (+0.00903) | > avg_loss: -0.45495 (-0.00351) | > avg_log_mle: -0.59089 (-0.00162) | > avg_loss_dur: 0.13595 (-0.00188)  > EPOCH: 103/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 04:00:19)   --> STEP: 5/639 -- GLOBAL_STEP: 65925 | > loss: -0.55488 (-0.56286) | > log_mle: -0.70608 (-0.68588) | > loss_dur: 0.15120 (0.12303) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 22.03190 (13.96948) | > current_lr: 0.00025 | > step_time: 0.31980 (0.28090) | > loader_time: 0.00250 (0.00225)  --> STEP: 30/639 -- GLOBAL_STEP: 65950 | > loss: -0.55738 (-0.55314) | > log_mle: -0.66483 (-0.66757) | > loss_dur: 0.10745 (0.11443) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.71640 (13.58613) | > current_lr: 0.00025 | > step_time: 0.33610 (0.32777) | > loader_time: 0.00260 (0.00254)  --> STEP: 55/639 -- GLOBAL_STEP: 65975 | > loss: -0.50539 (-0.54347) | > log_mle: -0.59724 (-0.65522) | > loss_dur: 0.09185 (0.11175) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.34698 (12.88647) | > current_lr: 0.00025 | > step_time: 0.41480 (0.34840) | > loader_time: 0.00290 (0.00263)  --> STEP: 80/639 -- GLOBAL_STEP: 66000 | > loss: -0.49677 (-0.53486) | > log_mle: -0.62784 (-0.64574) | > loss_dur: 0.13107 (0.11089) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.07980 (12.23601) | > current_lr: 0.00025 | > step_time: 0.41600 (0.37277) | > loader_time: 0.00240 (0.00269)  --> STEP: 105/639 -- GLOBAL_STEP: 66025 | > loss: -0.47539 (-0.53205) | > log_mle: -0.58404 (-0.64272) | > loss_dur: 0.10866 (0.11067) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.43496 (12.06189) | > current_lr: 0.00025 | > step_time: 0.40350 (0.39271) | > loader_time: 0.00310 (0.00276)  --> STEP: 130/639 -- GLOBAL_STEP: 66050 | > loss: -0.50631 (-0.52907) | > log_mle: -0.62463 (-0.64054) | > loss_dur: 0.11832 (0.11148) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.27273 (11.90951) | > current_lr: 0.00025 | > step_time: 0.48690 (0.40547) | > loader_time: 0.00300 (0.00280)  --> STEP: 155/639 -- GLOBAL_STEP: 66075 | > loss: -0.49569 (-0.52745) | > log_mle: -0.60914 (-0.63934) | > loss_dur: 0.11345 (0.11189) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.08720 (11.79872) | > current_lr: 0.00025 | > step_time: 0.53530 (0.42113) | > loader_time: 0.00290 (0.00282)  --> STEP: 180/639 -- GLOBAL_STEP: 66100 | > loss: -0.54963 (-0.52833) | > log_mle: -0.66855 (-0.64057) | > loss_dur: 0.11892 (0.11224) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.45008 (11.77341) | > current_lr: 0.00025 | > step_time: 0.57100 (0.43626) | > loader_time: 0.00290 (0.00286)  --> STEP: 205/639 -- GLOBAL_STEP: 66125 | > loss: -0.48381 (-0.52470) | > log_mle: -0.58962 (-0.63661) | > loss_dur: 0.10581 (0.11191) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.00009 (11.36088) | > current_lr: 0.00025 | > step_time: 0.54820 (0.44977) | > loader_time: 0.00320 (0.00290)  --> STEP: 230/639 -- GLOBAL_STEP: 66150 | > loss: -0.51327 (-0.52413) | > log_mle: -0.62646 (-0.63629) | > loss_dur: 0.11319 (0.11215) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.39928 (11.28181) | > current_lr: 0.00025 | > step_time: 0.55520 (0.46610) | > loader_time: 0.00340 (0.00294)  --> STEP: 255/639 -- GLOBAL_STEP: 66175 | > loss: -0.47968 (-0.52279) | > log_mle: -0.59487 (-0.63491) | > loss_dur: 0.11518 (0.11213) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.87092 (11.11891) | > current_lr: 0.00025 | > step_time: 0.66080 (0.48023) | > loader_time: 0.00340 (0.00296)  --> STEP: 280/639 -- GLOBAL_STEP: 66200 | > loss: -0.50217 (-0.52213) | > log_mle: -0.62250 (-0.63498) | > loss_dur: 0.12034 (0.11285) | > amp_scaler: 2048.00000 (3957.02857) | > grad_norm: 6.96348 (11.23572) | > current_lr: 0.00025 | > step_time: 0.71080 (0.49824) | > loader_time: 0.00280 (0.00300)  --> STEP: 305/639 -- GLOBAL_STEP: 66225 | > loss: -0.47621 (-0.51993) | > log_mle: -0.59829 (-0.63311) | > loss_dur: 0.12208 (0.11319) | > amp_scaler: 2048.00000 (3800.55082) | > grad_norm: 16.39956 (11.09785) | > current_lr: 0.00025 | > step_time: 0.65970 (0.51262) | > loader_time: 0.00350 (0.00303)  --> STEP: 330/639 -- GLOBAL_STEP: 66250 | > loss: -0.50025 (-0.51727) | > log_mle: -0.62348 (-0.63099) | > loss_dur: 0.12323 (0.11372) | > amp_scaler: 2048.00000 (3667.78182) | > grad_norm: 10.59374 (11.10348) | > current_lr: 0.00025 | > step_time: 0.71870 (0.52708) | > loader_time: 0.00350 (0.00306)  --> STEP: 355/639 -- GLOBAL_STEP: 66275 | > loss: -0.44989 (-0.51540) | > log_mle: -0.55268 (-0.62895) | > loss_dur: 0.10279 (0.11355) | > amp_scaler: 2048.00000 (3553.71268) | > grad_norm: 2.63563 (10.93837) | > current_lr: 0.00025 | > step_time: 0.59140 (0.54006) | > loader_time: 0.00320 (0.00309)  --> STEP: 380/639 -- GLOBAL_STEP: 66300 | > loss: -0.48578 (-0.51322) | > log_mle: -0.58860 (-0.62664) | > loss_dur: 0.10282 (0.11343) | > amp_scaler: 2048.00000 (3454.65263) | > grad_norm: 4.34687 (10.74494) | > current_lr: 0.00025 | > step_time: 0.60720 (0.55249) | > loader_time: 0.00370 (0.00313)  --> STEP: 405/639 -- GLOBAL_STEP: 66325 | > loss: -0.48129 (-0.51156) | > log_mle: -0.59103 (-0.62528) | > loss_dur: 0.10973 (0.11372) | > amp_scaler: 2048.00000 (3367.82222) | > grad_norm: 6.71300 (10.71376) | > current_lr: 0.00025 | > step_time: 0.70780 (0.56461) | > loader_time: 0.00340 (0.00316)  --> STEP: 430/639 -- GLOBAL_STEP: 66350 | > loss: -0.44737 (-0.50943) | > log_mle: -0.55273 (-0.62333) | > loss_dur: 0.10536 (0.11391) | > amp_scaler: 2048.00000 (3291.08837) | > grad_norm: 3.87693 (10.56326) | > current_lr: 0.00025 | > step_time: 0.74060 (0.57685) | > loader_time: 0.00370 (0.00319)  --> STEP: 455/639 -- GLOBAL_STEP: 66375 | > loss: -0.46531 (-0.50746) | > log_mle: -0.57716 (-0.62126) | > loss_dur: 0.11185 (0.11381) | > amp_scaler: 2048.00000 (3222.78681) | > grad_norm: 4.89452 (10.38180) | > current_lr: 0.00025 | > step_time: 0.78310 (0.58796) | > loader_time: 0.00380 (0.00322)  --> STEP: 480/639 -- GLOBAL_STEP: 66400 | > loss: -0.42219 (-0.50563) | > log_mle: -0.53295 (-0.61962) | > loss_dur: 0.11075 (0.11399) | > amp_scaler: 2048.00000 (3161.60000) | > grad_norm: 4.04249 (10.31361) | > current_lr: 0.00025 | > step_time: 0.77880 (0.60042) | > loader_time: 0.00360 (0.00324)  --> STEP: 505/639 -- GLOBAL_STEP: 66425 | > loss: -0.48636 (-0.50413) | > log_mle: -0.59757 (-0.61809) | > loss_dur: 0.11121 (0.11396) | > amp_scaler: 2048.00000 (3106.47129) | > grad_norm: 7.83261 (10.28046) | > current_lr: 0.00025 | > step_time: 0.96520 (0.61332) | > loader_time: 0.00410 (0.00327)  --> STEP: 530/639 -- GLOBAL_STEP: 66450 | > loss: -0.45775 (-0.50257) | > log_mle: -0.58504 (-0.61676) | > loss_dur: 0.12729 (0.11419) | > amp_scaler: 2048.00000 (3056.54340) | > grad_norm: 7.25361 (10.18384) | > current_lr: 0.00025 | > step_time: 0.91960 (0.62686) | > loader_time: 0.00420 (0.00331)  --> STEP: 555/639 -- GLOBAL_STEP: 66475 | > loss: -0.46058 (-0.50097) | > log_mle: -0.57209 (-0.61517) | > loss_dur: 0.11151 (0.11420) | > amp_scaler: 2048.00000 (3011.11351) | > grad_norm: 6.59598 (10.05950) | > current_lr: 0.00025 | > step_time: 0.84280 (0.64112) | > loader_time: 0.00400 (0.00334)  --> STEP: 580/639 -- GLOBAL_STEP: 66500 | > loss: -0.48013 (-0.49963) | > log_mle: -0.58908 (-0.61396) | > loss_dur: 0.10895 (0.11434) | > amp_scaler: 2048.00000 (2969.60000) | > grad_norm: 7.58127 (10.02102) | > current_lr: 0.00025 | > step_time: 0.97530 (0.65603) | > loader_time: 0.00430 (0.00339)  --> STEP: 605/639 -- GLOBAL_STEP: 66525 | > loss: -0.45208 (-0.49826) | > log_mle: -0.55865 (-0.61283) | > loss_dur: 0.10657 (0.11457) | > amp_scaler: 2048.00000 (2931.51736) | > grad_norm: 4.95644 (9.97709) | > current_lr: 0.00025 | > step_time: 1.06340 (0.67194) | > loader_time: 0.00430 (0.00344)  --> STEP: 630/639 -- GLOBAL_STEP: 66550 | > loss: -0.46148 (-0.49710) | > log_mle: -0.58997 (-0.61185) | > loss_dur: 0.12849 (0.11475) | > amp_scaler: 2048.00000 (2896.45714) | > grad_norm: 10.57432 (9.96805) | > current_lr: 0.00025 | > step_time: 1.20120 (0.68920) | > loader_time: 0.00690 (0.00350) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02461 (-0.01048) | > avg_loss: -0.45321 (+0.00173) | > avg_log_mle: -0.58983 (+0.00106) | > avg_loss_dur: 0.13662 (+0.00067)  > EPOCH: 104/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 04:07:58)   --> STEP: 15/639 -- GLOBAL_STEP: 66575 | > loss: -0.65532 (-0.56420) | > log_mle: -0.78091 (-0.67744) | > loss_dur: 0.12558 (0.11324) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 44.02504 (14.28210) | > current_lr: 0.00025 | > step_time: 0.32770 (0.30089) | > loader_time: 0.00260 (0.00234)  --> STEP: 40/639 -- GLOBAL_STEP: 66600 | > loss: -0.51782 (-0.54785) | > log_mle: -0.63131 (-0.65956) | > loss_dur: 0.11349 (0.11170) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.36294 (11.82947) | > current_lr: 0.00025 | > step_time: 0.43510 (0.33650) | > loader_time: 0.00300 (0.00264)  --> STEP: 65/639 -- GLOBAL_STEP: 66625 | > loss: -0.53144 (-0.53979) | > log_mle: -0.62238 (-0.65192) | > loss_dur: 0.09094 (0.11213) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.01356 (11.82940) | > current_lr: 0.00025 | > step_time: 0.39520 (0.36319) | > loader_time: 0.00300 (0.00273)  --> STEP: 90/639 -- GLOBAL_STEP: 66650 | > loss: -0.47832 (-0.53211) | > log_mle: -0.57972 (-0.64231) | > loss_dur: 0.10141 (0.11020) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.75023 (10.90483) | > current_lr: 0.00025 | > step_time: 0.52530 (0.37905) | > loader_time: 0.00320 (0.00278)  --> STEP: 115/639 -- GLOBAL_STEP: 66675 | > loss: -0.44887 (-0.53089) | > log_mle: -0.55038 (-0.64180) | > loss_dur: 0.10151 (0.11091) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.51937 (10.96631) | > current_lr: 0.00025 | > step_time: 0.38190 (0.39502) | > loader_time: 0.00310 (0.00282)  --> STEP: 140/639 -- GLOBAL_STEP: 66700 | > loss: -0.49308 (-0.52942) | > log_mle: -0.62759 (-0.64149) | > loss_dur: 0.13452 (0.11207) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.72191 (11.06097) | > current_lr: 0.00024 | > step_time: 0.56610 (0.40991) | > loader_time: 0.00330 (0.00284)  --> STEP: 165/639 -- GLOBAL_STEP: 66725 | > loss: -0.49617 (-0.52785) | > log_mle: -0.63216 (-0.63998) | > loss_dur: 0.13599 (0.11213) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.33638 (10.91243) | > current_lr: 0.00024 | > step_time: 0.48770 (0.42356) | > loader_time: 0.00310 (0.00289)  --> STEP: 190/639 -- GLOBAL_STEP: 66750 | > loss: -0.45868 (-0.52722) | > log_mle: -0.56450 (-0.63922) | > loss_dur: 0.10582 (0.11200) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.38788 (10.98790) | > current_lr: 0.00024 | > step_time: 0.52170 (0.44021) | > loader_time: 0.00330 (0.00294)  --> STEP: 215/639 -- GLOBAL_STEP: 66775 | > loss: -0.55433 (-0.52445) | > log_mle: -0.66173 (-0.63665) | > loss_dur: 0.10740 (0.11220) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 16.05711 (10.87877) | > current_lr: 0.00024 | > step_time: 0.53510 (0.45456) | > loader_time: 0.00320 (0.00299)  --> STEP: 240/639 -- GLOBAL_STEP: 66800 | > loss: -0.55283 (-0.52483) | > log_mle: -0.67411 (-0.63698) | > loss_dur: 0.12128 (0.11216) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.33168 (10.92168) | > current_lr: 0.00024 | > step_time: 0.55190 (0.46887) | > loader_time: 0.00330 (0.00303)  --> STEP: 265/639 -- GLOBAL_STEP: 66825 | > loss: -0.55460 (-0.52321) | > log_mle: -0.69679 (-0.63593) | > loss_dur: 0.14219 (0.11272) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 24.61884 (10.98487) | > current_lr: 0.00024 | > step_time: 0.63440 (0.48341) | > loader_time: 0.00330 (0.00307)  --> STEP: 290/639 -- GLOBAL_STEP: 66850 | > loss: -0.48405 (-0.52137) | > log_mle: -0.59531 (-0.63422) | > loss_dur: 0.11127 (0.11285) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.17761 (10.84474) | > current_lr: 0.00024 | > step_time: 0.65120 (0.49888) | > loader_time: 0.00350 (0.00311)  --> STEP: 315/639 -- GLOBAL_STEP: 66875 | > loss: -0.47021 (-0.51939) | > log_mle: -0.57507 (-0.63253) | > loss_dur: 0.10486 (0.11314) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.37910 (10.84645) | > current_lr: 0.00024 | > step_time: 0.72560 (0.51313) | > loader_time: 0.00350 (0.00314)  --> STEP: 340/639 -- GLOBAL_STEP: 66900 | > loss: -0.48679 (-0.51735) | > log_mle: -0.59565 (-0.63098) | > loss_dur: 0.10886 (0.11362) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.12760 (10.80643) | > current_lr: 0.00024 | > step_time: 0.58740 (0.52826) | > loader_time: 0.00330 (0.00318)  --> STEP: 365/639 -- GLOBAL_STEP: 66925 | > loss: -0.48819 (-0.51506) | > log_mle: -0.60335 (-0.62854) | > loss_dur: 0.11515 (0.11348) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.64441 (10.58434) | > current_lr: 0.00024 | > step_time: 0.80870 (0.54106) | > loader_time: 0.00370 (0.00320)  --> STEP: 390/639 -- GLOBAL_STEP: 66950 | > loss: -0.46986 (-0.51299) | > log_mle: -0.60010 (-0.62648) | > loss_dur: 0.13025 (0.11349) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.50149 (10.41843) | > current_lr: 0.00024 | > step_time: 0.71480 (0.55400) | > loader_time: 0.00350 (0.00323)  --> STEP: 415/639 -- GLOBAL_STEP: 66975 | > loss: -0.45936 (-0.51097) | > log_mle: -0.57499 (-0.62471) | > loss_dur: 0.11562 (0.11374) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.89491 (10.34685) | > current_lr: 0.00024 | > step_time: 0.72770 (0.56646) | > loader_time: 0.00340 (0.00326)  --> STEP: 440/639 -- GLOBAL_STEP: 67000 | > loss: -0.44539 (-0.50922) | > log_mle: -0.56066 (-0.62299) | > loss_dur: 0.11527 (0.11377) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.99291 (10.24630) | > current_lr: 0.00024 | > step_time: 0.68480 (0.57867) | > loader_time: 0.00350 (0.00329)  --> STEP: 465/639 -- GLOBAL_STEP: 67025 | > loss: -0.45066 (-0.50712) | > log_mle: -0.56412 (-0.62089) | > loss_dur: 0.11346 (0.11377) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.24895 (10.09154) | > current_lr: 0.00024 | > step_time: 0.80920 (0.58914) | > loader_time: 0.00390 (0.00332)  --> STEP: 490/639 -- GLOBAL_STEP: 67050 | > loss: -0.44616 (-0.50534) | > log_mle: -0.54773 (-0.61927) | > loss_dur: 0.10158 (0.11392) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.15377 (9.99763) | > current_lr: 0.00024 | > step_time: 0.77410 (0.60213) | > loader_time: 0.00390 (0.00334)  --> STEP: 515/639 -- GLOBAL_STEP: 67075 | > loss: -0.46673 (-0.50407) | > log_mle: -0.57620 (-0.61807) | > loss_dur: 0.10947 (0.11400) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.63841 (9.91736) | > current_lr: 0.00024 | > step_time: 0.94880 (0.61500) | > loader_time: 0.00420 (0.00337)  --> STEP: 540/639 -- GLOBAL_STEP: 67100 | > loss: -0.47219 (-0.50246) | > log_mle: -0.59114 (-0.61658) | > loss_dur: 0.11895 (0.11412) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.24771 (9.84128) | > current_lr: 0.00024 | > step_time: 0.85780 (0.62909) | > loader_time: 0.00430 (0.00340)  --> STEP: 565/639 -- GLOBAL_STEP: 67125 | > loss: -0.44684 (-0.50070) | > log_mle: -0.55462 (-0.61501) | > loss_dur: 0.10777 (0.11431) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.60504 (9.75006) | > current_lr: 0.00024 | > step_time: 0.99160 (0.64257) | > loader_time: 0.00400 (0.00343)  --> STEP: 590/639 -- GLOBAL_STEP: 67150 | > loss: -0.46679 (-0.49952) | > log_mle: -0.58168 (-0.61399) | > loss_dur: 0.11489 (0.11447) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.32391 (9.68458) | > current_lr: 0.00024 | > step_time: 0.94030 (0.65829) | > loader_time: 0.00430 (0.00347)  --> STEP: 615/639 -- GLOBAL_STEP: 67175 | > loss: -0.46456 (-0.49814) | > log_mle: -0.58621 (-0.61273) | > loss_dur: 0.12165 (0.11459) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.58615 (9.61450) | > current_lr: 0.00024 | > step_time: 0.97450 (0.67387) | > loader_time: 0.00440 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02325 (-0.00135) | > avg_loss: -0.45297 (+0.00024) | > avg_log_mle: -0.59069 (-0.00086) | > avg_loss_dur: 0.13772 (+0.00110)  > EPOCH: 105/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 04:15:33)   --> STEP: 0/639 -- GLOBAL_STEP: 67200 | > loss: -0.57739 (-0.57739) | > log_mle: -0.71857 (-0.71857) | > loss_dur: 0.14118 (0.14118) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 30.98180 (30.98180) | > current_lr: 0.00024 | > step_time: 0.65250 (0.65246) | > loader_time: 0.41440 (0.41444)  --> STEP: 25/639 -- GLOBAL_STEP: 67225 | > loss: -0.55895 (-0.55598) | > log_mle: -0.67524 (-0.67128) | > loss_dur: 0.11629 (0.11530) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.19078 (11.37714) | > current_lr: 0.00024 | > step_time: 0.38660 (0.32867) | > loader_time: 0.00290 (0.00389)  --> STEP: 50/639 -- GLOBAL_STEP: 67250 | > loss: -0.50938 (-0.54452) | > log_mle: -0.60839 (-0.65705) | > loss_dur: 0.09901 (0.11253) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.25028 (10.47129) | > current_lr: 0.00024 | > step_time: 0.31550 (0.34985) | > loader_time: 0.00260 (0.00337)  --> STEP: 75/639 -- GLOBAL_STEP: 67275 | > loss: -0.52756 (-0.53588) | > log_mle: -0.62972 (-0.64763) | > loss_dur: 0.10217 (0.11175) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.44112 (10.19732) | > current_lr: 0.00024 | > step_time: 0.41380 (0.36871) | > loader_time: 0.00290 (0.00322)  --> STEP: 100/639 -- GLOBAL_STEP: 67300 | > loss: -0.53806 (-0.53376) | > log_mle: -0.64151 (-0.64427) | > loss_dur: 0.10346 (0.11051) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.78254 (9.93016) | > current_lr: 0.00024 | > step_time: 0.50910 (0.38610) | > loader_time: 0.00270 (0.00316)  --> STEP: 125/639 -- GLOBAL_STEP: 67325 | > loss: -0.52171 (-0.53083) | > log_mle: -0.63544 (-0.64213) | > loss_dur: 0.11373 (0.11130) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.77862 (10.04881) | > current_lr: 0.00024 | > step_time: 0.41360 (0.40115) | > loader_time: 0.00280 (0.00314)  --> STEP: 150/639 -- GLOBAL_STEP: 67350 | > loss: -0.48687 (-0.52917) | > log_mle: -0.59525 (-0.64083) | > loss_dur: 0.10838 (0.11165) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.09663 (10.22627) | > current_lr: 0.00024 | > step_time: 0.43810 (0.41573) | > loader_time: 0.00290 (0.00313)  --> STEP: 175/639 -- GLOBAL_STEP: 67375 | > loss: -0.51695 (-0.52872) | > log_mle: -0.61419 (-0.64072) | > loss_dur: 0.09724 (0.11200) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.49990 (10.39644) | > current_lr: 0.00024 | > step_time: 0.55640 (0.43040) | > loader_time: 0.00320 (0.00313)  --> STEP: 200/639 -- GLOBAL_STEP: 67400 | > loss: -0.48355 (-0.52630) | > log_mle: -0.60124 (-0.63799) | > loss_dur: 0.11769 (0.11170) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.83102 (10.34256) | > current_lr: 0.00024 | > step_time: 0.51350 (0.44347) | > loader_time: 0.00330 (0.00315)  --> STEP: 225/639 -- GLOBAL_STEP: 67425 | > loss: -0.48591 (-0.52369) | > log_mle: -0.58117 (-0.63548) | > loss_dur: 0.09525 (0.11179) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.09182 (10.27108) | > current_lr: 0.00024 | > step_time: 0.51500 (0.45860) | > loader_time: 0.00320 (0.00317)  --> STEP: 250/639 -- GLOBAL_STEP: 67450 | > loss: -0.48886 (-0.52399) | > log_mle: -0.58210 (-0.63614) | > loss_dur: 0.09324 (0.11215) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.87965 (10.41490) | > current_lr: 0.00024 | > step_time: 0.62170 (0.47411) | > loader_time: 0.00380 (0.00320)  --> STEP: 275/639 -- GLOBAL_STEP: 67475 | > loss: -0.50489 (-0.52357) | > log_mle: -0.61872 (-0.63627) | > loss_dur: 0.11382 (0.11271) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.87126 (10.65943) | > current_lr: 0.00024 | > step_time: 0.82740 (0.49081) | > loader_time: 0.00390 (0.00322)  --> STEP: 300/639 -- GLOBAL_STEP: 67500 | > loss: -0.49633 (-0.52103) | > log_mle: -0.61290 (-0.63404) | > loss_dur: 0.11656 (0.11301) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.93454 (10.53286) | > current_lr: 0.00024 | > step_time: 0.73950 (0.50438) | > loader_time: 0.00360 (0.00324)  --> STEP: 325/639 -- GLOBAL_STEP: 67525 | > loss: -0.50150 (-0.51860) | > log_mle: -0.61251 (-0.63216) | > loss_dur: 0.11101 (0.11356) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.08859 (10.43439) | > current_lr: 0.00024 | > step_time: 0.64940 (0.51937) | > loader_time: 0.00350 (0.00327)  --> STEP: 350/639 -- GLOBAL_STEP: 67550 | > loss: -0.51434 (-0.51650) | > log_mle: -0.63923 (-0.63016) | > loss_dur: 0.12489 (0.11366) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 16.06974 (10.32366) | > current_lr: 0.00024 | > step_time: 0.84110 (0.53296) | > loader_time: 0.00380 (0.00330)  --> STEP: 375/639 -- GLOBAL_STEP: 67575 | > loss: -0.46443 (-0.51427) | > log_mle: -0.57367 (-0.62771) | > loss_dur: 0.10924 (0.11344) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.34920 (10.14666) | > current_lr: 0.00024 | > step_time: 0.67360 (0.54596) | > loader_time: 0.00330 (0.00332)  --> STEP: 400/639 -- GLOBAL_STEP: 67600 | > loss: -0.49513 (-0.51253) | > log_mle: -0.60731 (-0.62618) | > loss_dur: 0.11218 (0.11365) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.44508 (10.06154) | > current_lr: 0.00024 | > step_time: 0.72210 (0.55755) | > loader_time: 0.00390 (0.00334)  --> STEP: 425/639 -- GLOBAL_STEP: 67625 | > loss: -0.53629 (-0.51031) | > log_mle: -0.65880 (-0.62407) | > loss_dur: 0.12251 (0.11376) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 16.97437 (9.95090) | > current_lr: 0.00024 | > step_time: 0.91370 (0.57060) | > loader_time: 0.00390 (0.00337)  --> STEP: 450/639 -- GLOBAL_STEP: 67650 | > loss: -0.45627 (-0.50851) | > log_mle: -0.56230 (-0.62219) | > loss_dur: 0.10603 (0.11369) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.37291 (9.82584) | > current_lr: 0.00024 | > step_time: 0.78710 (0.58223) | > loader_time: 0.00380 (0.00339)  --> STEP: 475/639 -- GLOBAL_STEP: 67675 | > loss: -0.47161 (-0.50663) | > log_mle: -0.57361 (-0.62040) | > loss_dur: 0.10199 (0.11377) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.59800 (9.74343) | > current_lr: 0.00024 | > step_time: 0.67860 (0.59365) | > loader_time: 0.00370 (0.00341)  --> STEP: 500/639 -- GLOBAL_STEP: 67700 | > loss: -0.51463 (-0.50524) | > log_mle: -0.63402 (-0.61909) | > loss_dur: 0.11938 (0.11385) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.95945 (9.64821) | > current_lr: 0.00024 | > step_time: 0.95490 (0.60703) | > loader_time: 0.00390 (0.00343)  --> STEP: 525/639 -- GLOBAL_STEP: 67725 | > loss: -0.48412 (-0.50377) | > log_mle: -0.61581 (-0.61776) | > loss_dur: 0.13169 (0.11399) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.86355 (9.57110) | > current_lr: 0.00024 | > step_time: 1.02710 (0.62064) | > loader_time: 0.00440 (0.00346)  --> STEP: 550/639 -- GLOBAL_STEP: 67750 | > loss: -0.43810 (-0.50220) | > log_mle: -0.54971 (-0.61623) | > loss_dur: 0.11160 (0.11402) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.92887 (9.49088) | > current_lr: 0.00024 | > step_time: 0.92590 (0.63516) | > loader_time: 0.00420 (0.00349)  --> STEP: 575/639 -- GLOBAL_STEP: 67775 | > loss: -0.45922 (-0.50061) | > log_mle: -0.56503 (-0.61475) | > loss_dur: 0.10581 (0.11414) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.31254 (9.40626) | > current_lr: 0.00024 | > step_time: 0.86870 (0.64884) | > loader_time: 0.00380 (0.00352)  --> STEP: 600/639 -- GLOBAL_STEP: 67800 | > loss: -0.44897 (-0.49926) | > log_mle: -0.56691 (-0.61361) | > loss_dur: 0.11794 (0.11435) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.60086 (9.34237) | > current_lr: 0.00024 | > step_time: 1.05620 (0.66535) | > loader_time: 0.00450 (0.00356)  --> STEP: 625/639 -- GLOBAL_STEP: 67825 | > loss: -0.47489 (-0.49805) | > log_mle: -0.59432 (-0.61251) | > loss_dur: 0.11943 (0.11447) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.36469 (9.27338) | > current_lr: 0.00024 | > step_time: 1.05440 (0.68206) | > loader_time: 0.00420 (0.00359) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02009 (-0.00316) | > avg_loss: -0.45606 (-0.00310) | > avg_log_mle: -0.59320 (-0.00251) | > avg_loss_dur: 0.13714 (-0.00059)  > EPOCH: 106/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 04:23:09)   --> STEP: 10/639 -- GLOBAL_STEP: 67850 | > loss: -0.53956 (-0.55972) | > log_mle: -0.67243 (-0.67792) | > loss_dur: 0.13287 (0.11820) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.15594 (11.59982) | > current_lr: 0.00024 | > step_time: 0.28650 (0.29897) | > loader_time: 0.00270 (0.00661)  --> STEP: 35/639 -- GLOBAL_STEP: 67875 | > loss: -0.53961 (-0.55062) | > log_mle: -0.63852 (-0.66326) | > loss_dur: 0.09891 (0.11264) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.14950 (10.91060) | > current_lr: 0.00024 | > step_time: 0.39980 (0.33222) | > loader_time: 0.00270 (0.00380)  --> STEP: 60/639 -- GLOBAL_STEP: 67900 | > loss: -0.50358 (-0.54241) | > log_mle: -0.59757 (-0.65477) | > loss_dur: 0.09399 (0.11235) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.86800 (10.22358) | > current_lr: 0.00024 | > step_time: 0.34120 (0.36228) | > loader_time: 0.00220 (0.00341)  --> STEP: 85/639 -- GLOBAL_STEP: 67925 | > loss: -0.49251 (-0.53484) | > log_mle: -0.59948 (-0.64529) | > loss_dur: 0.10697 (0.11045) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.16757 (9.62995) | > current_lr: 0.00024 | > step_time: 0.44460 (0.37633) | > loader_time: 0.00290 (0.00325)  --> STEP: 110/639 -- GLOBAL_STEP: 67950 | > loss: -0.50356 (-0.53458) | > log_mle: -0.60886 (-0.64577) | > loss_dur: 0.10530 (0.11118) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.22472 (10.11929) | > current_lr: 0.00024 | > step_time: 0.39910 (0.39435) | > loader_time: 0.00320 (0.00320)  --> STEP: 135/639 -- GLOBAL_STEP: 67975 | > loss: -0.47595 (-0.53156) | > log_mle: -0.58816 (-0.64252) | > loss_dur: 0.11221 (0.11097) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.85794 (10.05865) | > current_lr: 0.00024 | > step_time: 0.47870 (0.40679) | > loader_time: 0.00290 (0.00315)  --> STEP: 160/639 -- GLOBAL_STEP: 68000 | > loss: -0.52383 (-0.52916) | > log_mle: -0.62648 (-0.64002) | > loss_dur: 0.10265 (0.11085) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.89749 (10.04713) | > current_lr: 0.00024 | > step_time: 0.45790 (0.42097) | > loader_time: 0.00290 (0.00314)  --> STEP: 185/639 -- GLOBAL_STEP: 68025 | > loss: -0.50406 (-0.52995) | > log_mle: -0.60965 (-0.64163) | > loss_dur: 0.10558 (0.11168) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.09362 (10.57177) | > current_lr: 0.00024 | > step_time: 0.56760 (0.43672) | > loader_time: 0.00310 (0.00314)  --> STEP: 210/639 -- GLOBAL_STEP: 68050 | > loss: -0.43699 (-0.52645) | > log_mle: -0.53248 (-0.63810) | > loss_dur: 0.09549 (0.11165) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.49905 (10.32965) | > current_lr: 0.00024 | > step_time: 0.60220 (0.45225) | > loader_time: 0.00280 (0.00316)  --> STEP: 235/639 -- GLOBAL_STEP: 68075 | > loss: -0.50973 (-0.52626) | > log_mle: -0.62304 (-0.63802) | > loss_dur: 0.11331 (0.11176) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.80612 (10.36292) | > current_lr: 0.00024 | > step_time: 0.54890 (0.46785) | > loader_time: 0.00350 (0.00318)  --> STEP: 260/639 -- GLOBAL_STEP: 68100 | > loss: -0.50543 (-0.52525) | > log_mle: -0.62282 (-0.63737) | > loss_dur: 0.11738 (0.11212) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.76478 (10.54643) | > current_lr: 0.00024 | > step_time: 0.57850 (0.48281) | > loader_time: 0.00340 (0.00319)  --> STEP: 285/639 -- GLOBAL_STEP: 68125 | > loss: -0.46331 (-0.52357) | > log_mle: -0.57084 (-0.63617) | > loss_dur: 0.10752 (0.11260) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.43390 (10.55042) | > current_lr: 0.00024 | > step_time: 0.67180 (0.49840) | > loader_time: 0.00340 (0.00321)  --> STEP: 310/639 -- GLOBAL_STEP: 68150 | > loss: -0.49296 (-0.52144) | > log_mle: -0.61067 (-0.63448) | > loss_dur: 0.11771 (0.11304) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.75204 (10.49208) | > current_lr: 0.00024 | > step_time: 0.57020 (0.51274) | > loader_time: 0.00330 (0.00323)  --> STEP: 335/639 -- GLOBAL_STEP: 68175 | > loss: -0.51248 (-0.51871) | > log_mle: -0.63681 (-0.63229) | > loss_dur: 0.12433 (0.11358) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.13793 (10.38489) | > current_lr: 0.00024 | > step_time: 0.65950 (0.52739) | > loader_time: 0.00360 (0.00326)  --> STEP: 360/639 -- GLOBAL_STEP: 68200 | > loss: -0.46174 (-0.51676) | > log_mle: -0.57289 (-0.63019) | > loss_dur: 0.11115 (0.11343) | > amp_scaler: 4096.00000 (2156.08889) | > grad_norm: 6.28928 (10.24154) | > current_lr: 0.00024 | > step_time: 0.67180 (0.53911) | > loader_time: 0.00340 (0.00327)  --> STEP: 385/639 -- GLOBAL_STEP: 68225 | > loss: -0.49627 (-0.51487) | > log_mle: -0.60491 (-0.62820) | > loss_dur: 0.10864 (0.11333) | > amp_scaler: 4096.00000 (2282.05714) | > grad_norm: 7.84321 (10.09448) | > current_lr: 0.00024 | > step_time: 0.66640 (0.55160) | > loader_time: 0.00360 (0.00330)  --> STEP: 410/639 -- GLOBAL_STEP: 68250 | > loss: -0.44969 (-0.51284) | > log_mle: -0.57812 (-0.62640) | > loss_dur: 0.12843 (0.11356) | > amp_scaler: 4096.00000 (2392.66341) | > grad_norm: 9.22688 (9.97359) | > current_lr: 0.00024 | > step_time: 0.76280 (0.56355) | > loader_time: 0.00370 (0.00332)  --> STEP: 435/639 -- GLOBAL_STEP: 68275 | > loss: -0.48657 (-0.51070) | > log_mle: -0.59158 (-0.62432) | > loss_dur: 0.10501 (0.11362) | > amp_scaler: 4096.00000 (2490.55632) | > grad_norm: 5.86283 (9.83509) | > current_lr: 0.00024 | > step_time: 0.69070 (0.57576) | > loader_time: 0.00360 (0.00334)  --> STEP: 460/639 -- GLOBAL_STEP: 68300 | > loss: -0.43853 (-0.50859) | > log_mle: -0.55617 (-0.62221) | > loss_dur: 0.11765 (0.11362) | > amp_scaler: 4096.00000 (2577.80870) | > grad_norm: 2.65081 (9.67062) | > current_lr: 0.00024 | > step_time: 0.90330 (0.58743) | > loader_time: 0.00410 (0.00336)  --> STEP: 485/639 -- GLOBAL_STEP: 68325 | > loss: -0.47617 (-0.50692) | > log_mle: -0.58975 (-0.62071) | > loss_dur: 0.11357 (0.11379) | > amp_scaler: 4096.00000 (2656.06598) | > grad_norm: 9.43655 (9.59328) | > current_lr: 0.00024 | > step_time: 0.91870 (0.59939) | > loader_time: 0.00360 (0.00338)  --> STEP: 510/639 -- GLOBAL_STEP: 68350 | > loss: -0.47545 (-0.50554) | > log_mle: -0.59271 (-0.61935) | > loss_dur: 0.11726 (0.11381) | > amp_scaler: 4096.00000 (2726.65098) | > grad_norm: 7.03029 (9.50959) | > current_lr: 0.00024 | > step_time: 0.84240 (0.61251) | > loader_time: 0.00390 (0.00340)  --> STEP: 535/639 -- GLOBAL_STEP: 68375 | > loss: -0.45588 (-0.50392) | > log_mle: -0.58061 (-0.61784) | > loss_dur: 0.12473 (0.11392) | > amp_scaler: 4096.00000 (2790.63925) | > grad_norm: 8.30268 (9.45841) | > current_lr: 0.00024 | > step_time: 0.89830 (0.62649) | > loader_time: 0.00390 (0.00343)  --> STEP: 560/639 -- GLOBAL_STEP: 68400 | > loss: -0.47379 (-0.50242) | > log_mle: -0.59161 (-0.61649) | > loss_dur: 0.11782 (0.11407) | > amp_scaler: 4096.00000 (2848.91429) | > grad_norm: 7.36076 (9.39882) | > current_lr: 0.00024 | > step_time: 0.92400 (0.64165) | > loader_time: 0.00390 (0.00346)  --> STEP: 585/639 -- GLOBAL_STEP: 68425 | > loss: -0.48013 (-0.50106) | > log_mle: -0.59274 (-0.61524) | > loss_dur: 0.11261 (0.11419) | > amp_scaler: 4096.00000 (2902.20855) | > grad_norm: 6.71709 (9.34900) | > current_lr: 0.00024 | > step_time: 1.04480 (0.65677) | > loader_time: 0.00400 (0.00349)  --> STEP: 610/639 -- GLOBAL_STEP: 68450 | > loss: -0.44543 (-0.49973) | > log_mle: -0.56124 (-0.61406) | > loss_dur: 0.11581 (0.11433) | > amp_scaler: 4096.00000 (2951.13443) | > grad_norm: 4.41074 (9.30521) | > current_lr: 0.00024 | > step_time: 1.02160 (0.67302) | > loader_time: 0.00440 (0.00353)  --> STEP: 635/639 -- GLOBAL_STEP: 68475 | > loss: -0.46290 (-0.49864) | > log_mle: -0.58647 (-0.61329) | > loss_dur: 0.12357 (0.11465) | > amp_scaler: 4096.00000 (2996.20787) | > grad_norm: 9.12289 (9.31174) | > current_lr: 0.00024 | > step_time: 1.10660 (0.69006) | > loader_time: 0.00430 (0.00360) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02568 (+0.00559) | > avg_loss: -0.45887 (-0.00280) | > avg_log_mle: -0.59364 (-0.00044) | > avg_loss_dur: 0.13477 (-0.00236) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_68480.pth.tar  > EPOCH: 107/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 04:30:52)   --> STEP: 20/639 -- GLOBAL_STEP: 68500 | > loss: -0.62990 (-0.56760) | > log_mle: -0.78794 (-0.68599) | > loss_dur: 0.15805 (0.11840) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 27.58783 (13.90255) | > current_lr: 0.00024 | > step_time: 0.34310 (0.31019) | > loader_time: 0.00290 (0.00526)  --> STEP: 45/639 -- GLOBAL_STEP: 68525 | > loss: -0.51702 (-0.54858) | > log_mle: -0.60987 (-0.66155) | > loss_dur: 0.09285 (0.11297) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.40512 (11.66368) | > current_lr: 0.00024 | > step_time: 0.40130 (0.34135) | > loader_time: 0.00290 (0.00389)  --> STEP: 70/639 -- GLOBAL_STEP: 68550 | > loss: -0.52296 (-0.53826) | > log_mle: -0.64471 (-0.65058) | > loss_dur: 0.12176 (0.11232) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 24.99132 (11.92583) | > current_lr: 0.00024 | > step_time: 0.46940 (0.36818) | > loader_time: 0.00280 (0.00355)  --> STEP: 95/639 -- GLOBAL_STEP: 68575 | > loss: -0.60069 (-0.53476) | > log_mle: -0.70488 (-0.64533) | > loss_dur: 0.10419 (0.11057) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 20.01488 (11.39774) | > current_lr: 0.00024 | > step_time: 0.40370 (0.38044) | > loader_time: 0.00310 (0.00340)  --> STEP: 120/639 -- GLOBAL_STEP: 68600 | > loss: -0.55989 (-0.53163) | > log_mle: -0.70154 (-0.64345) | > loss_dur: 0.14164 (0.11181) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 20.71654 (11.61399) | > current_lr: 0.00024 | > step_time: 0.53690 (0.39732) | > loader_time: 0.00320 (0.00332)  --> STEP: 145/639 -- GLOBAL_STEP: 68625 | > loss: -0.51609 (-0.53076) | > log_mle: -0.64361 (-0.64304) | > loss_dur: 0.12751 (0.11227) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.90335 (11.64740) | > current_lr: 0.00024 | > step_time: 0.46600 (0.41241) | > loader_time: 0.00300 (0.00328)  --> STEP: 170/639 -- GLOBAL_STEP: 68650 | > loss: -0.52002 (-0.52925) | > log_mle: -0.62892 (-0.64133) | > loss_dur: 0.10890 (0.11208) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.28974 (11.64008) | > current_lr: 0.00024 | > step_time: 0.53810 (0.42665) | > loader_time: 0.00340 (0.00325)  --> STEP: 195/639 -- GLOBAL_STEP: 68675 | > loss: -0.48424 (-0.52811) | > log_mle: -0.57833 (-0.63981) | > loss_dur: 0.09409 (0.11170) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.56277 (11.50891) | > current_lr: 0.00024 | > step_time: 0.52670 (0.44259) | > loader_time: 0.00330 (0.00326)  --> STEP: 220/639 -- GLOBAL_STEP: 68700 | > loss: -0.52304 (-0.52541) | > log_mle: -0.62398 (-0.63718) | > loss_dur: 0.10094 (0.11177) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.66763 (11.33512) | > current_lr: 0.00024 | > step_time: 0.62940 (0.45811) | > loader_time: 0.00340 (0.00327)  --> STEP: 245/639 -- GLOBAL_STEP: 68725 | > loss: -0.50663 (-0.52587) | > log_mle: -0.61240 (-0.63768) | > loss_dur: 0.10577 (0.11182) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.84299 (11.46617) | > current_lr: 0.00024 | > step_time: 0.52280 (0.47240) | > loader_time: 0.00330 (0.00328)  --> STEP: 270/639 -- GLOBAL_STEP: 68750 | > loss: -0.55873 (-0.52508) | > log_mle: -0.68067 (-0.63747) | > loss_dur: 0.12194 (0.11239) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 28.00996 (11.63955) | > current_lr: 0.00024 | > step_time: 0.82030 (0.48894) | > loader_time: 0.00380 (0.00329)  --> STEP: 295/639 -- GLOBAL_STEP: 68775 | > loss: -0.47428 (-0.52279) | > log_mle: -0.57358 (-0.63544) | > loss_dur: 0.09930 (0.11264) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.11816 (11.48740) | > current_lr: 0.00024 | > step_time: 0.63000 (0.50277) | > loader_time: 0.00370 (0.00330)  --> STEP: 320/639 -- GLOBAL_STEP: 68800 | > loss: -0.48067 (-0.52082) | > log_mle: -0.61922 (-0.63395) | > loss_dur: 0.13854 (0.11313) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.21064 (11.36730) | > current_lr: 0.00024 | > step_time: 0.76700 (0.51791) | > loader_time: 0.00370 (0.00332)  --> STEP: 345/639 -- GLOBAL_STEP: 68825 | > loss: -0.49332 (-0.51860) | > log_mle: -0.60298 (-0.63210) | > loss_dur: 0.10966 (0.11350) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.87251 (11.27083) | > current_lr: 0.00024 | > step_time: 0.65420 (0.53121) | > loader_time: 0.00350 (0.00334)  --> STEP: 370/639 -- GLOBAL_STEP: 68850 | > loss: -0.50598 (-0.51666) | > log_mle: -0.61734 (-0.62992) | > loss_dur: 0.11136 (0.11326) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.30824 (11.05824) | > current_lr: 0.00024 | > step_time: 0.86420 (0.54352) | > loader_time: 0.00360 (0.00336)  --> STEP: 395/639 -- GLOBAL_STEP: 68875 | > loss: -0.47376 (-0.51460) | > log_mle: -0.58775 (-0.62788) | > loss_dur: 0.11399 (0.11328) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.89896 (10.88586) | > current_lr: 0.00024 | > step_time: 0.83280 (0.55556) | > loader_time: 0.00370 (0.00337)  --> STEP: 420/639 -- GLOBAL_STEP: 68900 | > loss: -0.46640 (-0.51240) | > log_mle: -0.57996 (-0.62593) | > loss_dur: 0.11355 (0.11353) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.87619 (10.73120) | > current_lr: 0.00024 | > step_time: 0.69670 (0.56794) | > loader_time: 0.00340 (0.00340)  --> STEP: 445/639 -- GLOBAL_STEP: 68925 | > loss: -0.48010 (-0.51079) | > log_mle: -0.58908 (-0.62434) | > loss_dur: 0.10898 (0.11355) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.63193 (10.60959) | > current_lr: 0.00024 | > step_time: 0.85010 (0.58031) | > loader_time: 0.00370 (0.00342)  --> STEP: 470/639 -- GLOBAL_STEP: 68950 | > loss: -0.48982 (-0.50882) | > log_mle: -0.60125 (-0.62237) | > loss_dur: 0.11143 (0.11355) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.54430 (10.42297) | > current_lr: 0.00024 | > step_time: 0.77240 (0.59119) | > loader_time: 0.00330 (0.00344)  --> STEP: 495/639 -- GLOBAL_STEP: 68975 | > loss: -0.43573 (-0.50718) | > log_mle: -0.53634 (-0.62081) | > loss_dur: 0.10061 (0.11363) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.10696 (10.30298) | > current_lr: 0.00024 | > step_time: 0.76700 (0.60403) | > loader_time: 0.00370 (0.00345)  --> STEP: 520/639 -- GLOBAL_STEP: 69000 | > loss: -0.48214 (-0.50590) | > log_mle: -0.59559 (-0.61953) | > loss_dur: 0.11345 (0.11363) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.60946 (10.22925) | > current_lr: 0.00024 | > step_time: 0.88030 (0.61658) | > loader_time: 0.00380 (0.00347)  --> STEP: 545/639 -- GLOBAL_STEP: 69025 | > loss: -0.50128 (-0.50435) | > log_mle: -0.62606 (-0.61818) | > loss_dur: 0.12478 (0.11382) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.93699 (10.11163) | > current_lr: 0.00024 | > step_time: 0.92330 (0.63255) | > loader_time: 0.00420 (0.00350)  --> STEP: 570/639 -- GLOBAL_STEP: 69050 | > loss: -0.47012 (-0.50267) | > log_mle: -0.57137 (-0.61658) | > loss_dur: 0.10125 (0.11391) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.31484 (9.97372) | > current_lr: 0.00024 | > step_time: 0.84560 (0.64568) | > loader_time: 0.00390 (0.00353)  --> STEP: 595/639 -- GLOBAL_STEP: 69075 | > loss: -0.45544 (-0.50140) | > log_mle: -0.56416 (-0.61551) | > loss_dur: 0.10872 (0.11411) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.83121 (9.86204) | > current_lr: 0.00024 | > step_time: 0.95280 (0.66170) | > loader_time: 0.00420 (0.00356)  --> STEP: 620/639 -- GLOBAL_STEP: 69100 | > loss: -0.45696 (-0.50000) | > log_mle: -0.57013 (-0.61425) | > loss_dur: 0.11317 (0.11426) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.99198 (9.82220) | > current_lr: 0.00024 | > step_time: 1.13870 (0.67724) | > loader_time: 0.00460 (0.00360) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03479 (+0.00911) | > avg_loss: -0.45568 (+0.00319) | > avg_log_mle: -0.59320 (+0.00043) | > avg_loss_dur: 0.13752 (+0.00275)  > EPOCH: 108/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 04:38:28)   --> STEP: 5/639 -- GLOBAL_STEP: 69125 | > loss: -0.57077 (-0.56543) | > log_mle: -0.71276 (-0.68820) | > loss_dur: 0.14200 (0.12276) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 29.08685 (16.68197) | > current_lr: 0.00024 | > step_time: 0.31830 (0.27772) | > loader_time: 0.00470 (0.00306)  --> STEP: 30/639 -- GLOBAL_STEP: 69150 | > loss: -0.55899 (-0.55347) | > log_mle: -0.66424 (-0.66936) | > loss_dur: 0.10525 (0.11589) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.04157 (12.37280) | > current_lr: 0.00024 | > step_time: 0.32140 (0.31814) | > loader_time: 0.00280 (0.00258)  --> STEP: 55/639 -- GLOBAL_STEP: 69175 | > loss: -0.51443 (-0.54418) | > log_mle: -0.60203 (-0.65823) | > loss_dur: 0.08760 (0.11405) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.00927 (11.41954) | > current_lr: 0.00024 | > step_time: 0.33100 (0.33937) | > loader_time: 0.00300 (0.00266)  --> STEP: 80/639 -- GLOBAL_STEP: 69200 | > loss: -0.50637 (-0.53560) | > log_mle: -0.63338 (-0.64859) | > loss_dur: 0.12701 (0.11299) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.60782 (10.95043) | > current_lr: 0.00024 | > step_time: 0.39310 (0.35973) | > loader_time: 0.00260 (0.00272)  --> STEP: 105/639 -- GLOBAL_STEP: 69225 | > loss: -0.48067 (-0.53348) | > log_mle: -0.59019 (-0.64559) | > loss_dur: 0.10952 (0.11211) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.96498 (11.12141) | > current_lr: 0.00024 | > step_time: 0.43510 (0.37817) | > loader_time: 0.00320 (0.00278)  --> STEP: 130/639 -- GLOBAL_STEP: 69250 | > loss: -0.50844 (-0.53084) | > log_mle: -0.62622 (-0.64340) | > loss_dur: 0.11778 (0.11255) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.60866 (11.21587) | > current_lr: 0.00024 | > step_time: 0.48870 (0.39607) | > loader_time: 0.00310 (0.00282)  --> STEP: 155/639 -- GLOBAL_STEP: 69275 | > loss: -0.49930 (-0.52938) | > log_mle: -0.61325 (-0.64219) | > loss_dur: 0.11394 (0.11281) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.20651 (11.31852) | > current_lr: 0.00024 | > step_time: 0.53620 (0.41459) | > loader_time: 0.00300 (0.00287)  --> STEP: 180/639 -- GLOBAL_STEP: 69300 | > loss: -0.55640 (-0.53031) | > log_mle: -0.67088 (-0.64327) | > loss_dur: 0.11448 (0.11296) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.60537 (11.58754) | > current_lr: 0.00024 | > step_time: 0.57130 (0.43093) | > loader_time: 0.00340 (0.00292)  --> STEP: 205/639 -- GLOBAL_STEP: 69325 | > loss: -0.48501 (-0.52693) | > log_mle: -0.59248 (-0.63929) | > loss_dur: 0.10747 (0.11236) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.48548 (11.22979) | > current_lr: 0.00024 | > step_time: 0.55050 (0.44499) | > loader_time: 0.00330 (0.00296)  --> STEP: 230/639 -- GLOBAL_STEP: 69350 | > loss: -0.50502 (-0.52623) | > log_mle: -0.62436 (-0.63883) | > loss_dur: 0.11934 (0.11260) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.55970 (11.38831) | > current_lr: 0.00024 | > step_time: 0.55000 (0.46162) | > loader_time: 0.00360 (0.00301)  --> STEP: 255/639 -- GLOBAL_STEP: 69375 | > loss: -0.48139 (-0.52491) | > log_mle: -0.59530 (-0.63742) | > loss_dur: 0.11391 (0.11251) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.16982 (11.36559) | > current_lr: 0.00024 | > step_time: 0.63500 (0.47483) | > loader_time: 0.00340 (0.00304)  --> STEP: 280/639 -- GLOBAL_STEP: 69400 | > loss: -0.50097 (-0.52436) | > log_mle: -0.62010 (-0.63748) | > loss_dur: 0.11913 (0.11312) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.70847 (11.55075) | > current_lr: 0.00024 | > step_time: 0.59510 (0.49148) | > loader_time: 0.00290 (0.00309)  --> STEP: 305/639 -- GLOBAL_STEP: 69425 | > loss: -0.48226 (-0.52228) | > log_mle: -0.60353 (-0.63558) | > loss_dur: 0.12128 (0.11330) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.66220 (11.55958) | > current_lr: 0.00024 | > step_time: 0.69120 (0.50574) | > loader_time: 0.00350 (0.00312)  --> STEP: 330/639 -- GLOBAL_STEP: 69450 | > loss: -0.49460 (-0.51964) | > log_mle: -0.61993 (-0.63340) | > loss_dur: 0.12532 (0.11376) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.65658 (11.49040) | > current_lr: 0.00024 | > step_time: 0.79000 (0.52057) | > loader_time: 0.00370 (0.00316)  --> STEP: 355/639 -- GLOBAL_STEP: 69475 | > loss: -0.45451 (-0.51768) | > log_mle: -0.55485 (-0.63130) | > loss_dur: 0.10034 (0.11362) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.77044 (11.36721) | > current_lr: 0.00024 | > step_time: 0.59320 (0.53344) | > loader_time: 0.00340 (0.00320)  --> STEP: 380/639 -- GLOBAL_STEP: 69500 | > loss: -0.48834 (-0.51548) | > log_mle: -0.59168 (-0.62894) | > loss_dur: 0.10334 (0.11345) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.01022 (11.20121) | > current_lr: 0.00024 | > step_time: 0.62500 (0.54673) | > loader_time: 0.00310 (0.00322)  --> STEP: 405/639 -- GLOBAL_STEP: 69525 | > loss: -0.47823 (-0.51384) | > log_mle: -0.59168 (-0.62755) | > loss_dur: 0.11345 (0.11371) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.14767 (11.16301) | > current_lr: 0.00024 | > step_time: 0.77890 (0.55860) | > loader_time: 0.00340 (0.00325)  --> STEP: 430/639 -- GLOBAL_STEP: 69550 | > loss: -0.44839 (-0.51164) | > log_mle: -0.55385 (-0.62550) | > loss_dur: 0.10546 (0.11385) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.29046 (11.02265) | > current_lr: 0.00024 | > step_time: 0.74850 (0.57109) | > loader_time: 0.00330 (0.00328)  --> STEP: 455/639 -- GLOBAL_STEP: 69575 | > loss: -0.46697 (-0.50972) | > log_mle: -0.57717 (-0.62342) | > loss_dur: 0.11020 (0.11370) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.33645 (10.84602) | > current_lr: 0.00024 | > step_time: 0.74470 (0.58252) | > loader_time: 0.00380 (0.00330)  --> STEP: 480/639 -- GLOBAL_STEP: 69600 | > loss: -0.42183 (-0.50782) | > log_mle: -0.53470 (-0.62173) | > loss_dur: 0.11288 (0.11391) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.49567 (10.73271) | > current_lr: 0.00024 | > step_time: 0.77830 (0.59467) | > loader_time: 0.00350 (0.00333)  --> STEP: 505/639 -- GLOBAL_STEP: 69625 | > loss: -0.48932 (-0.50634) | > log_mle: -0.59826 (-0.62018) | > loss_dur: 0.10894 (0.11384) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.40862 (10.57424) | > current_lr: 0.00024 | > step_time: 0.98300 (0.60678) | > loader_time: 0.00410 (0.00335)  --> STEP: 530/639 -- GLOBAL_STEP: 69650 | > loss: -0.46144 (-0.50480) | > log_mle: -0.58780 (-0.61883) | > loss_dur: 0.12636 (0.11403) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.43330 (10.49866) | > current_lr: 0.00024 | > step_time: 1.04580 (0.62129) | > loader_time: 0.00410 (0.00338)  --> STEP: 555/639 -- GLOBAL_STEP: 69675 | > loss: -0.46683 (-0.50318) | > log_mle: -0.57446 (-0.61724) | > loss_dur: 0.10764 (0.11406) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.74902 (10.36896) | > current_lr: 0.00024 | > step_time: 0.84480 (0.63566) | > loader_time: 0.00410 (0.00342)  --> STEP: 580/639 -- GLOBAL_STEP: 69700 | > loss: -0.47407 (-0.50176) | > log_mle: -0.58824 (-0.61597) | > loss_dur: 0.11417 (0.11421) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.52004 (10.28615) | > current_lr: 0.00024 | > step_time: 0.98130 (0.65079) | > loader_time: 0.00440 (0.00345)  --> STEP: 605/639 -- GLOBAL_STEP: 69725 | > loss: -0.45620 (-0.50039) | > log_mle: -0.56229 (-0.61479) | > loss_dur: 0.10609 (0.11440) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.22713 (10.18880) | > current_lr: 0.00024 | > step_time: 1.02910 (0.66660) | > loader_time: 0.00450 (0.00349)  --> STEP: 630/639 -- GLOBAL_STEP: 69750 | > loss: -0.46390 (-0.49915) | > log_mle: -0.59167 (-0.61379) | > loss_dur: 0.12777 (0.11464) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.66747 (10.12474) | > current_lr: 0.00024 | > step_time: 1.19760 (0.68410) | > loader_time: 0.00460 (0.00353) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03738 (+0.00259) | > avg_loss: -0.45796 (-0.00228) | > avg_log_mle: -0.59397 (-0.00077) | > avg_loss_dur: 0.13602 (-0.00151)  > EPOCH: 109/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 04:46:03)   --> STEP: 15/639 -- GLOBAL_STEP: 69775 | > loss: -0.67833 (-0.56829) | > log_mle: -0.80192 (-0.68353) | > loss_dur: 0.12359 (0.11525) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 22.23955 (13.39700) | > current_lr: 0.00024 | > step_time: 0.32820 (0.29694) | > loader_time: 0.00280 (0.00261)  --> STEP: 40/639 -- GLOBAL_STEP: 69800 | > loss: -0.52127 (-0.54883) | > log_mle: -0.63425 (-0.66157) | > loss_dur: 0.11299 (0.11274) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.54183 (12.15222) | > current_lr: 0.00024 | > step_time: 0.35570 (0.33063) | > loader_time: 0.00250 (0.00268)  --> STEP: 65/639 -- GLOBAL_STEP: 69825 | > loss: -0.53291 (-0.54164) | > log_mle: -0.62352 (-0.65390) | > loss_dur: 0.09061 (0.11226) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.61225 (11.37324) | > current_lr: 0.00024 | > step_time: 0.38940 (0.35936) | > loader_time: 0.00300 (0.00276)  --> STEP: 90/639 -- GLOBAL_STEP: 69850 | > loss: -0.47685 (-0.53361) | > log_mle: -0.58157 (-0.64436) | > loss_dur: 0.10472 (0.11075) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.54596 (10.24539) | > current_lr: 0.00024 | > step_time: 0.51030 (0.37828) | > loader_time: 0.00290 (0.00281)  --> STEP: 115/639 -- GLOBAL_STEP: 69875 | > loss: -0.45270 (-0.53253) | > log_mle: -0.55283 (-0.64377) | > loss_dur: 0.10012 (0.11124) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.36274 (10.33168) | > current_lr: 0.00024 | > step_time: 0.40370 (0.39528) | > loader_time: 0.00280 (0.00284)  --> STEP: 140/639 -- GLOBAL_STEP: 69900 | > loss: -0.49837 (-0.53154) | > log_mle: -0.63023 (-0.64360) | > loss_dur: 0.13187 (0.11206) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.05436 (10.57343) | > current_lr: 0.00024 | > step_time: 0.55210 (0.40889) | > loader_time: 0.00300 (0.00287)  --> STEP: 165/639 -- GLOBAL_STEP: 69925 | > loss: -0.51029 (-0.53040) | > log_mle: -0.63654 (-0.64225) | > loss_dur: 0.12625 (0.11185) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.92317 (10.48765) | > current_lr: 0.00024 | > step_time: 0.48860 (0.42320) | > loader_time: 0.00300 (0.00291)  --> STEP: 190/639 -- GLOBAL_STEP: 69950 | > loss: -0.45925 (-0.52953) | > log_mle: -0.56568 (-0.64134) | > loss_dur: 0.10643 (0.11181) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.60115 (10.66495) | > current_lr: 0.00024 | > step_time: 0.52210 (0.44101) | > loader_time: 0.00330 (0.00294)  --> STEP: 215/639 -- GLOBAL_STEP: 69975 | > loss: -0.54355 (-0.52675) | > log_mle: -0.66341 (-0.63879) | > loss_dur: 0.11986 (0.11204) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.46974 (10.53694) | > current_lr: 0.00024 | > step_time: 0.51960 (0.45642) | > loader_time: 0.00340 (0.00296)  --> STEP: 240/639 -- GLOBAL_STEP: 70000 | > loss: -0.55943 (-0.52710) | > log_mle: -0.67786 (-0.63910) | > loss_dur: 0.11844 (0.11200) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.53118 (10.72469) | > current_lr: 0.00024 | > step_time: 0.54890 (0.47164) | > loader_time: 0.00320 (0.00300) > CHECKPOINT : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/checkpoint_70000.pth.tar  --> STEP: 265/639 -- GLOBAL_STEP: 70025 | > loss: -0.57277 (-0.52553) | > log_mle: -0.70306 (-0.63807) | > loss_dur: 0.13029 (0.11254) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 26.02334 (10.86266) | > current_lr: 0.00024 | > step_time: 0.63160 (0.48571) | > loader_time: 0.00330 (0.00309)  --> STEP: 290/639 -- GLOBAL_STEP: 70050 | > loss: -0.48737 (-0.52373) | > log_mle: -0.59629 (-0.63639) | > loss_dur: 0.10892 (0.11266) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.15665 (10.74046) | > current_lr: 0.00024 | > step_time: 0.61920 (0.49946) | > loader_time: 0.00340 (0.00313)  --> STEP: 315/639 -- GLOBAL_STEP: 70075 | > loss: -0.47162 (-0.52177) | > log_mle: -0.57809 (-0.63459) | > loss_dur: 0.10646 (0.11283) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.92266 (10.73814) | > current_lr: 0.00024 | > step_time: 0.61550 (0.51304) | > loader_time: 0.00350 (0.00316)  --> STEP: 340/639 -- GLOBAL_STEP: 70100 | > loss: -0.49248 (-0.51966) | > log_mle: -0.59954 (-0.63306) | > loss_dur: 0.10706 (0.11340) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.26795 (10.79524) | > current_lr: 0.00024 | > step_time: 0.62740 (0.52922) | > loader_time: 0.00340 (0.00319)  --> STEP: 365/639 -- GLOBAL_STEP: 70125 | > loss: -0.48032 (-0.51736) | > log_mle: -0.60249 (-0.63056) | > loss_dur: 0.12218 (0.11320) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.13503 (10.61596) | > current_lr: 0.00024 | > step_time: 0.80600 (0.54082) | > loader_time: 0.00350 (0.00322)  --> STEP: 390/639 -- GLOBAL_STEP: 70150 | > loss: -0.47848 (-0.51539) | > log_mle: -0.60572 (-0.62847) | > loss_dur: 0.12724 (0.11308) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.83822 (10.46780) | > current_lr: 0.00024 | > step_time: 0.70450 (0.55373) | > loader_time: 0.00340 (0.00324)  --> STEP: 415/639 -- GLOBAL_STEP: 70175 | > loss: -0.45854 (-0.51334) | > log_mle: -0.57444 (-0.62669) | > loss_dur: 0.11590 (0.11335) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.98621 (10.43780) | > current_lr: 0.00024 | > step_time: 0.76870 (0.56597) | > loader_time: 0.00360 (0.00326)  --> STEP: 440/639 -- GLOBAL_STEP: 70200 | > loss: -0.45032 (-0.51154) | > log_mle: -0.56406 (-0.62499) | > loss_dur: 0.11373 (0.11345) | > amp_scaler: 8192.00000 (4272.87273) | > grad_norm: 4.51916 (10.34292) | > current_lr: 0.00024 | > step_time: 0.71690 (0.57805) | > loader_time: 0.00360 (0.00328)  --> STEP: 465/639 -- GLOBAL_STEP: 70225 | > loss: -0.45492 (-0.50949) | > log_mle: -0.56579 (-0.62290) | > loss_dur: 0.11088 (0.11341) | > amp_scaler: 8192.00000 (4483.57849) | > grad_norm: 7.27609 (10.22232) | > current_lr: 0.00024 | > step_time: 0.76720 (0.58879) | > loader_time: 0.00370 (0.00331)  --> STEP: 490/639 -- GLOBAL_STEP: 70250 | > loss: -0.45214 (-0.50771) | > log_mle: -0.55081 (-0.62129) | > loss_dur: 0.09867 (0.11358) | > amp_scaler: 8192.00000 (4672.78367) | > grad_norm: 3.72031 (10.13036) | > current_lr: 0.00024 | > step_time: 0.75830 (0.60165) | > loader_time: 0.00370 (0.00334)  --> STEP: 515/639 -- GLOBAL_STEP: 70275 | > loss: -0.46913 (-0.50650) | > log_mle: -0.57921 (-0.62012) | > loss_dur: 0.11008 (0.11362) | > amp_scaler: 8192.00000 (4843.61942) | > grad_norm: 6.07039 (10.05478) | > current_lr: 0.00024 | > step_time: 1.00760 (0.61483) | > loader_time: 0.00430 (0.00337)  --> STEP: 540/639 -- GLOBAL_STEP: 70300 | > loss: -0.47513 (-0.50496) | > log_mle: -0.59208 (-0.61864) | > loss_dur: 0.11696 (0.11368) | > amp_scaler: 8192.00000 (4998.63704) | > grad_norm: 8.04339 (9.97085) | > current_lr: 0.00024 | > step_time: 0.87870 (0.62940) | > loader_time: 0.00390 (0.00341)  --> STEP: 565/639 -- GLOBAL_STEP: 70325 | > loss: -0.44508 (-0.50329) | > log_mle: -0.55617 (-0.61710) | > loss_dur: 0.11109 (0.11382) | > amp_scaler: 8192.00000 (5139.93628) | > grad_norm: 4.07617 (9.87322) | > current_lr: 0.00024 | > step_time: 0.98280 (0.64286) | > loader_time: 0.00410 (0.00344)  --> STEP: 590/639 -- GLOBAL_STEP: 70350 | > loss: -0.46579 (-0.50206) | > log_mle: -0.58718 (-0.61608) | > loss_dur: 0.12139 (0.11402) | > amp_scaler: 8192.00000 (5269.26102) | > grad_norm: 6.24043 (9.80148) | > current_lr: 0.00024 | > step_time: 0.93910 (0.65824) | > loader_time: 0.00440 (0.00348)  --> STEP: 615/639 -- GLOBAL_STEP: 70375 | > loss: -0.47320 (-0.50071) | > log_mle: -0.59073 (-0.61487) | > loss_dur: 0.11752 (0.11416) | > amp_scaler: 8192.00000 (5388.07154) | > grad_norm: 8.60184 (9.70845) | > current_lr: 0.00024 | > step_time: 1.01980 (0.67414) | > loader_time: 0.00440 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03079 (-0.00659) | > avg_loss: -0.45676 (+0.00120) | > avg_log_mle: -0.59075 (+0.00322) | > avg_loss_dur: 0.13399 (-0.00203)  > EPOCH: 110/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 04:53:43)   --> STEP: 0/639 -- GLOBAL_STEP: 70400 | > loss: -0.56761 (-0.56761) | > log_mle: -0.72197 (-0.72197) | > loss_dur: 0.15436 (0.15436) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 32.34700 (32.34700) | > current_lr: 0.00024 | > step_time: 0.53530 (0.53530) | > loader_time: 0.42830 (0.42827)  --> STEP: 25/639 -- GLOBAL_STEP: 70425 | > loss: -0.54542 (-0.55579) | > log_mle: -0.67570 (-0.67379) | > loss_dur: 0.13028 (0.11800) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.82154 (13.29016) | > current_lr: 0.00024 | > step_time: 0.38670 (0.31269) | > loader_time: 0.00290 (0.00299)  --> STEP: 50/639 -- GLOBAL_STEP: 70450 | > loss: -0.51294 (-0.54739) | > log_mle: -0.60751 (-0.66049) | > loss_dur: 0.09457 (0.11309) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.78635 (12.10602) | > current_lr: 0.00024 | > step_time: 0.39660 (0.34222) | > loader_time: 0.00280 (0.00292)  --> STEP: 75/639 -- GLOBAL_STEP: 70475 | > loss: -0.52994 (-0.53856) | > log_mle: -0.63268 (-0.65097) | > loss_dur: 0.10274 (0.11241) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.59062 (11.19498) | > current_lr: 0.00024 | > step_time: 0.41190 (0.36561) | > loader_time: 0.00310 (0.00293)  --> STEP: 100/639 -- GLOBAL_STEP: 70500 | > loss: -0.53824 (-0.53660) | > log_mle: -0.64451 (-0.64759) | > loss_dur: 0.10627 (0.11099) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.52361 (10.96138) | > current_lr: 0.00024 | > step_time: 0.50900 (0.38618) | > loader_time: 0.00290 (0.00295)  --> STEP: 125/639 -- GLOBAL_STEP: 70525 | > loss: -0.52105 (-0.53357) | > log_mle: -0.63793 (-0.64515) | > loss_dur: 0.11687 (0.11158) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.91414 (11.07188) | > current_lr: 0.00024 | > step_time: 0.42700 (0.40037) | > loader_time: 0.00310 (0.00299)  --> STEP: 150/639 -- GLOBAL_STEP: 70550 | > loss: -0.49026 (-0.53171) | > log_mle: -0.59888 (-0.64365) | > loss_dur: 0.10863 (0.11194) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.77179 (11.19887) | > current_lr: 0.00024 | > step_time: 0.42360 (0.41729) | > loader_time: 0.00280 (0.00301)  --> STEP: 175/639 -- GLOBAL_STEP: 70575 | > loss: -0.51985 (-0.53139) | > log_mle: -0.61466 (-0.64359) | > loss_dur: 0.09480 (0.11220) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.68763 (11.33260) | > current_lr: 0.00024 | > step_time: 0.55520 (0.43199) | > loader_time: 0.00350 (0.00304)  --> STEP: 200/639 -- GLOBAL_STEP: 70600 | > loss: -0.49273 (-0.52901) | > log_mle: -0.60130 (-0.64064) | > loss_dur: 0.10857 (0.11163) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.90822 (11.22438) | > current_lr: 0.00024 | > step_time: 0.52390 (0.44666) | > loader_time: 0.00330 (0.00306)  --> STEP: 225/639 -- GLOBAL_STEP: 70625 | > loss: -0.48493 (-0.52629) | > log_mle: -0.57962 (-0.63789) | > loss_dur: 0.09468 (0.11159) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.19228 (11.34344) | > current_lr: 0.00024 | > step_time: 0.51160 (0.46178) | > loader_time: 0.00320 (0.00310)  --> STEP: 250/639 -- GLOBAL_STEP: 70650 | > loss: -0.49415 (-0.52638) | > log_mle: -0.58501 (-0.63828) | > loss_dur: 0.09087 (0.11189) | > amp_scaler: 2048.00000 (3956.73600) | > grad_norm: 4.63479 (11.46783) | > current_lr: 0.00024 | > step_time: 0.57010 (0.47533) | > loader_time: 0.00340 (0.00313)  --> STEP: 275/639 -- GLOBAL_STEP: 70675 | > loss: -0.50900 (-0.52574) | > log_mle: -0.61892 (-0.63812) | > loss_dur: 0.10992 (0.11238) | > amp_scaler: 2048.00000 (3783.21455) | > grad_norm: 7.78739 (11.59749) | > current_lr: 0.00024 | > step_time: 0.84800 (0.49214) | > loader_time: 0.00390 (0.00317)  --> STEP: 300/639 -- GLOBAL_STEP: 70700 | > loss: -0.49876 (-0.52307) | > log_mle: -0.61405 (-0.63575) | > loss_dur: 0.11528 (0.11268) | > amp_scaler: 2048.00000 (3638.61333) | > grad_norm: 7.04980 (11.51087) | > current_lr: 0.00024 | > step_time: 0.81930 (0.50587) | > loader_time: 0.00360 (0.00320)  --> STEP: 325/639 -- GLOBAL_STEP: 70725 | > loss: -0.49374 (-0.52061) | > log_mle: -0.61405 (-0.63393) | > loss_dur: 0.12031 (0.11332) | > amp_scaler: 2048.00000 (3516.25846) | > grad_norm: 7.55678 (11.46636) | > current_lr: 0.00024 | > step_time: 0.61660 (0.52068) | > loader_time: 0.00310 (0.00323)  --> STEP: 350/639 -- GLOBAL_STEP: 70750 | > loss: -0.51271 (-0.51832) | > log_mle: -0.64198 (-0.63182) | > loss_dur: 0.12926 (0.11349) | > amp_scaler: 2048.00000 (3411.38286) | > grad_norm: 13.50939 (11.33492) | > current_lr: 0.00024 | > step_time: 0.82690 (0.53449) | > loader_time: 0.00390 (0.00325)  --> STEP: 375/639 -- GLOBAL_STEP: 70775 | > loss: -0.46924 (-0.51605) | > log_mle: -0.57428 (-0.62930) | > loss_dur: 0.10504 (0.11325) | > amp_scaler: 2048.00000 (3320.49067) | > grad_norm: 6.63054 (11.12113) | > current_lr: 0.00024 | > step_time: 0.68010 (0.54686) | > loader_time: 0.00390 (0.00328)  --> STEP: 400/639 -- GLOBAL_STEP: 70800 | > loss: -0.48703 (-0.51433) | > log_mle: -0.60228 (-0.62772) | > loss_dur: 0.11525 (0.11340) | > amp_scaler: 2048.00000 (3240.96000) | > grad_norm: 10.31389 (11.00476) | > current_lr: 0.00024 | > step_time: 0.70080 (0.55870) | > loader_time: 0.00380 (0.00331)  --> STEP: 425/639 -- GLOBAL_STEP: 70825 | > loss: -0.53219 (-0.51202) | > log_mle: -0.66073 (-0.62556) | > loss_dur: 0.12854 (0.11354) | > amp_scaler: 2048.00000 (3170.78588) | > grad_norm: 13.21518 (10.87436) | > current_lr: 0.00024 | > step_time: 0.88640 (0.57156) | > loader_time: 0.00400 (0.00334)  --> STEP: 450/639 -- GLOBAL_STEP: 70850 | > loss: -0.46346 (-0.51010) | > log_mle: -0.56400 (-0.62358) | > loss_dur: 0.10054 (0.11348) | > amp_scaler: 2048.00000 (3108.40889) | > grad_norm: 6.35906 (10.74862) | > current_lr: 0.00024 | > step_time: 0.74660 (0.58343) | > loader_time: 0.00360 (0.00336)  --> STEP: 475/639 -- GLOBAL_STEP: 70875 | > loss: -0.47550 (-0.50813) | > log_mle: -0.57436 (-0.62175) | > loss_dur: 0.09887 (0.11361) | > amp_scaler: 2048.00000 (3052.59789) | > grad_norm: 7.27407 (10.60118) | > current_lr: 0.00024 | > step_time: 0.69350 (0.59512) | > loader_time: 0.00340 (0.00338)  --> STEP: 500/639 -- GLOBAL_STEP: 70900 | > loss: -0.51510 (-0.50665) | > log_mle: -0.63114 (-0.62042) | > loss_dur: 0.11604 (0.11377) | > amp_scaler: 2048.00000 (3002.36800) | > grad_norm: 24.14518 (10.50302) | > current_lr: 0.00024 | > step_time: 1.01970 (0.60804) | > loader_time: 0.00390 (0.00341)  --> STEP: 525/639 -- GLOBAL_STEP: 70925 | > loss: -0.48397 (-0.50513) | > log_mle: -0.61542 (-0.61906) | > loss_dur: 0.13145 (0.11393) | > amp_scaler: 2048.00000 (2956.92190) | > grad_norm: 15.47488 (10.42953) | > current_lr: 0.00024 | > step_time: 0.98350 (0.62214) | > loader_time: 0.00430 (0.00344)  --> STEP: 550/639 -- GLOBAL_STEP: 70950 | > loss: -0.43850 (-0.50355) | > log_mle: -0.55043 (-0.61752) | > loss_dur: 0.11194 (0.11397) | > amp_scaler: 2048.00000 (2915.60727) | > grad_norm: 3.31711 (10.28130) | > current_lr: 0.00024 | > step_time: 0.92400 (0.63634) | > loader_time: 0.00420 (0.00347)  --> STEP: 575/639 -- GLOBAL_STEP: 70975 | > loss: -0.45905 (-0.50197) | > log_mle: -0.56915 (-0.61603) | > loss_dur: 0.11010 (0.11406) | > amp_scaler: 2048.00000 (2877.88522) | > grad_norm: 6.27064 (10.16601) | > current_lr: 0.00024 | > step_time: 0.87650 (0.64993) | > loader_time: 0.00430 (0.00350)  --> STEP: 600/639 -- GLOBAL_STEP: 71000 | > loss: -0.44888 (-0.50070) | > log_mle: -0.56497 (-0.61494) | > loss_dur: 0.11609 (0.11424) | > amp_scaler: 2048.00000 (2843.30667) | > grad_norm: 7.26631 (10.10206) | > current_lr: 0.00024 | > step_time: 1.01830 (0.66634) | > loader_time: 0.00460 (0.00354)  --> STEP: 625/639 -- GLOBAL_STEP: 71025 | > loss: -0.47553 (-0.49946) | > log_mle: -0.59530 (-0.61382) | > loss_dur: 0.11977 (0.11436) | > amp_scaler: 2048.00000 (2811.49440) | > grad_norm: 9.08236 (10.03599) | > current_lr: 0.00024 | > step_time: 1.01420 (0.68283) | > loader_time: 0.00440 (0.00360) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03504 (+0.00425) | > avg_loss: -0.45429 (+0.00247) | > avg_log_mle: -0.59042 (+0.00033) | > avg_loss_dur: 0.13613 (+0.00215)  > EPOCH: 111/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 05:01:19)   --> STEP: 10/639 -- GLOBAL_STEP: 71050 | > loss: -0.53045 (-0.56216) | > log_mle: -0.66373 (-0.67669) | > loss_dur: 0.13328 (0.11453) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 18.63237 (12.99531) | > current_lr: 0.00024 | > step_time: 0.28310 (0.29036) | > loader_time: 0.00250 (0.00475)  --> STEP: 35/639 -- GLOBAL_STEP: 71075 | > loss: -0.54595 (-0.55059) | > log_mle: -0.64115 (-0.66271) | > loss_dur: 0.09519 (0.11212) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.88009 (11.30404) | > current_lr: 0.00024 | > step_time: 0.35330 (0.32333) | > loader_time: 0.00280 (0.00329)  --> STEP: 60/639 -- GLOBAL_STEP: 71100 | > loss: -0.50712 (-0.54243) | > log_mle: -0.59906 (-0.65464) | > loss_dur: 0.09194 (0.11221) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.76728 (11.21706) | > current_lr: 0.00024 | > step_time: 0.34360 (0.35425) | > loader_time: 0.00270 (0.00314)  --> STEP: 85/639 -- GLOBAL_STEP: 71125 | > loss: -0.50453 (-0.53527) | > log_mle: -0.60185 (-0.64563) | > loss_dur: 0.09732 (0.11036) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.23885 (10.67629) | > current_lr: 0.00024 | > step_time: 0.41680 (0.37037) | > loader_time: 0.00270 (0.00307)  --> STEP: 110/639 -- GLOBAL_STEP: 71150 | > loss: -0.50764 (-0.53441) | > log_mle: -0.61103 (-0.64620) | > loss_dur: 0.10338 (0.11178) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.04567 (10.96497) | > current_lr: 0.00024 | > step_time: 0.39690 (0.39142) | > loader_time: 0.00310 (0.00307)  --> STEP: 135/639 -- GLOBAL_STEP: 71175 | > loss: -0.47221 (-0.53179) | > log_mle: -0.59186 (-0.64339) | > loss_dur: 0.11965 (0.11161) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.05917 (10.76658) | > current_lr: 0.00024 | > step_time: 0.46990 (0.40393) | > loader_time: 0.00280 (0.00304)  --> STEP: 160/639 -- GLOBAL_STEP: 71200 | > loss: -0.52782 (-0.52968) | > log_mle: -0.62925 (-0.64105) | > loss_dur: 0.10143 (0.11137) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.49033 (10.69427) | > current_lr: 0.00024 | > step_time: 0.45970 (0.42083) | > loader_time: 0.00300 (0.00305)  --> STEP: 185/639 -- GLOBAL_STEP: 71225 | > loss: -0.51105 (-0.53073) | > log_mle: -0.61478 (-0.64283) | > loss_dur: 0.10373 (0.11209) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.63635 (11.17646) | > current_lr: 0.00024 | > step_time: 0.57270 (0.43778) | > loader_time: 0.00330 (0.00308)  --> STEP: 210/639 -- GLOBAL_STEP: 71250 | > loss: -0.43673 (-0.52738) | > log_mle: -0.53355 (-0.63939) | > loss_dur: 0.09682 (0.11201) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.77858 (10.94007) | > current_lr: 0.00024 | > step_time: 0.56770 (0.45414) | > loader_time: 0.00300 (0.00309)  --> STEP: 235/639 -- GLOBAL_STEP: 71275 | > loss: -0.51686 (-0.52726) | > log_mle: -0.61908 (-0.63917) | > loss_dur: 0.10222 (0.11191) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.37920 (11.04051) | > current_lr: 0.00024 | > step_time: 0.62340 (0.46988) | > loader_time: 0.00310 (0.00312)  --> STEP: 260/639 -- GLOBAL_STEP: 71300 | > loss: -0.51222 (-0.52605) | > log_mle: -0.62434 (-0.63842) | > loss_dur: 0.11212 (0.11238) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.79525 (11.06612) | > current_lr: 0.00024 | > step_time: 0.57930 (0.48417) | > loader_time: 0.00320 (0.00315)  --> STEP: 285/639 -- GLOBAL_STEP: 71325 | > loss: -0.46665 (-0.52461) | > log_mle: -0.57279 (-0.63725) | > loss_dur: 0.10614 (0.11264) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.62199 (11.01563) | > current_lr: 0.00024 | > step_time: 0.60610 (0.49879) | > loader_time: 0.00350 (0.00318)  --> STEP: 310/639 -- GLOBAL_STEP: 71350 | > loss: -0.49539 (-0.52261) | > log_mle: -0.61187 (-0.63546) | > loss_dur: 0.11649 (0.11285) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.86145 (10.95370) | > current_lr: 0.00024 | > step_time: 0.55370 (0.51279) | > loader_time: 0.00320 (0.00321)  --> STEP: 335/639 -- GLOBAL_STEP: 71375 | > loss: -0.51520 (-0.51998) | > log_mle: -0.63915 (-0.63328) | > loss_dur: 0.12395 (0.11330) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.92612 (10.89101) | > current_lr: 0.00024 | > step_time: 0.66370 (0.52791) | > loader_time: 0.00330 (0.00323)  --> STEP: 360/639 -- GLOBAL_STEP: 71400 | > loss: -0.46885 (-0.51814) | > log_mle: -0.57535 (-0.63123) | > loss_dur: 0.10650 (0.11309) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.40484 (10.74039) | > current_lr: 0.00024 | > step_time: 0.67170 (0.54025) | > loader_time: 0.00340 (0.00325)  --> STEP: 385/639 -- GLOBAL_STEP: 71425 | > loss: -0.49302 (-0.51620) | > log_mle: -0.60439 (-0.62923) | > loss_dur: 0.11137 (0.11303) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.18544 (10.57359) | > current_lr: 0.00024 | > step_time: 0.74290 (0.55315) | > loader_time: 0.00360 (0.00328)  --> STEP: 410/639 -- GLOBAL_STEP: 71450 | > loss: -0.44358 (-0.51411) | > log_mle: -0.58083 (-0.62742) | > loss_dur: 0.13725 (0.11331) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.87397 (10.46688) | > current_lr: 0.00024 | > step_time: 0.78540 (0.56511) | > loader_time: 0.00370 (0.00330)  --> STEP: 435/639 -- GLOBAL_STEP: 71475 | > loss: -0.48720 (-0.51200) | > log_mle: -0.59170 (-0.62533) | > loss_dur: 0.10450 (0.11333) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.46969 (10.31903) | > current_lr: 0.00024 | > step_time: 0.68300 (0.57698) | > loader_time: 0.00320 (0.00331)  --> STEP: 460/639 -- GLOBAL_STEP: 71500 | > loss: -0.44661 (-0.50993) | > log_mle: -0.55876 (-0.62325) | > loss_dur: 0.11216 (0.11332) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.94663 (10.19572) | > current_lr: 0.00024 | > step_time: 0.92130 (0.58918) | > loader_time: 0.00410 (0.00334)  --> STEP: 485/639 -- GLOBAL_STEP: 71525 | > loss: -0.47822 (-0.50822) | > log_mle: -0.58762 (-0.62172) | > loss_dur: 0.10940 (0.11349) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.49623 (10.08429) | > current_lr: 0.00024 | > step_time: 0.84760 (0.60154) | > loader_time: 0.00400 (0.00337)  --> STEP: 510/639 -- GLOBAL_STEP: 71550 | > loss: -0.46764 (-0.50677) | > log_mle: -0.58785 (-0.62029) | > loss_dur: 0.12021 (0.11352) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.88354 (10.04384) | > current_lr: 0.00024 | > step_time: 0.91350 (0.61487) | > loader_time: 0.00400 (0.00339)  --> STEP: 535/639 -- GLOBAL_STEP: 71575 | > loss: -0.45523 (-0.50517) | > log_mle: -0.57841 (-0.61877) | > loss_dur: 0.12318 (0.11360) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.67497 (9.97796) | > current_lr: 0.00024 | > step_time: 0.85590 (0.62799) | > loader_time: 0.00380 (0.00341)  --> STEP: 560/639 -- GLOBAL_STEP: 71600 | > loss: -0.46447 (-0.50358) | > log_mle: -0.58999 (-0.61734) | > loss_dur: 0.12552 (0.11376) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.03426 (9.90444) | > current_lr: 0.00024 | > step_time: 0.89960 (0.64256) | > loader_time: 0.00400 (0.00345)  --> STEP: 585/639 -- GLOBAL_STEP: 71625 | > loss: -0.47815 (-0.50224) | > log_mle: -0.59300 (-0.61608) | > loss_dur: 0.11485 (0.11384) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.93689 (9.82560) | > current_lr: 0.00024 | > step_time: 1.09720 (0.65713) | > loader_time: 0.00430 (0.00348)  --> STEP: 610/639 -- GLOBAL_STEP: 71650 | > loss: -0.44631 (-0.50087) | > log_mle: -0.56072 (-0.61488) | > loss_dur: 0.11442 (0.11401) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.73573 (9.78961) | > current_lr: 0.00024 | > step_time: 1.04460 (0.67314) | > loader_time: 0.00430 (0.00351)  --> STEP: 635/639 -- GLOBAL_STEP: 71675 | > loss: -0.46579 (-0.49969) | > log_mle: -0.58981 (-0.61405) | > loss_dur: 0.12402 (0.11436) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.04132 (9.76602) | > current_lr: 0.00024 | > step_time: 1.12420 (0.68977) | > loader_time: 0.00420 (0.00355) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02884 (-0.00620) | > avg_loss: -0.45532 (-0.00103) | > avg_log_mle: -0.59254 (-0.00212) | > avg_loss_dur: 0.13723 (+0.00109)  > EPOCH: 112/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 05:08:56)   --> STEP: 20/639 -- GLOBAL_STEP: 71700 | > loss: -0.63864 (-0.56892) | > log_mle: -0.79867 (-0.68718) | > loss_dur: 0.16003 (0.11826) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 20.46680 (12.83462) | > current_lr: 0.00024 | > step_time: 0.40830 (0.31372) | > loader_time: 0.00320 (0.00241)  --> STEP: 45/639 -- GLOBAL_STEP: 71725 | > loss: -0.51480 (-0.54901) | > log_mle: -0.60995 (-0.66373) | > loss_dur: 0.09516 (0.11471) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.92457 (11.87160) | > current_lr: 0.00024 | > step_time: 0.40560 (0.34415) | > loader_time: 0.00290 (0.00264)  --> STEP: 70/639 -- GLOBAL_STEP: 71750 | > loss: -0.52894 (-0.54023) | > log_mle: -0.65221 (-0.65315) | > loss_dur: 0.12326 (0.11292) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.63645 (11.31607) | > current_lr: 0.00024 | > step_time: 0.39820 (0.36679) | > loader_time: 0.00300 (0.00274)  --> STEP: 95/639 -- GLOBAL_STEP: 71775 | > loss: -0.59612 (-0.53649) | > log_mle: -0.70497 (-0.64731) | > loss_dur: 0.10885 (0.11083) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.73227 (10.68931) | > current_lr: 0.00024 | > step_time: 0.40540 (0.38320) | > loader_time: 0.00310 (0.00281)  --> STEP: 120/639 -- GLOBAL_STEP: 71800 | > loss: -0.57067 (-0.53370) | > log_mle: -0.70026 (-0.64516) | > loss_dur: 0.12958 (0.11146) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 20.24476 (10.98785) | > current_lr: 0.00024 | > step_time: 0.54690 (0.39922) | > loader_time: 0.00310 (0.00285)  --> STEP: 145/639 -- GLOBAL_STEP: 71825 | > loss: -0.52008 (-0.53264) | > log_mle: -0.64573 (-0.64457) | > loss_dur: 0.12566 (0.11193) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.77519 (11.17492) | > current_lr: 0.00024 | > step_time: 0.47020 (0.41369) | > loader_time: 0.00290 (0.00291)  --> STEP: 170/639 -- GLOBAL_STEP: 71850 | > loss: -0.52609 (-0.53096) | > log_mle: -0.63222 (-0.64282) | > loss_dur: 0.10613 (0.11186) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.80208 (11.06181) | > current_lr: 0.00024 | > step_time: 0.53930 (0.42903) | > loader_time: 0.00340 (0.00294)  --> STEP: 195/639 -- GLOBAL_STEP: 71875 | > loss: -0.48524 (-0.52969) | > log_mle: -0.57961 (-0.64119) | > loss_dur: 0.09437 (0.11150) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.45567 (11.04354) | > current_lr: 0.00024 | > step_time: 0.47910 (0.44469) | > loader_time: 0.00320 (0.00299)  --> STEP: 220/639 -- GLOBAL_STEP: 71900 | > loss: -0.52179 (-0.52658) | > log_mle: -0.62515 (-0.63827) | > loss_dur: 0.10336 (0.11169) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.59115 (10.87802) | > current_lr: 0.00024 | > step_time: 0.60450 (0.45985) | > loader_time: 0.00360 (0.00303)  --> STEP: 245/639 -- GLOBAL_STEP: 71925 | > loss: -0.51352 (-0.52687) | > log_mle: -0.61399 (-0.63867) | > loss_dur: 0.10047 (0.11180) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.12312 (10.96995) | > current_lr: 0.00024 | > step_time: 0.51990 (0.47412) | > loader_time: 0.00330 (0.00306)  --> STEP: 270/639 -- GLOBAL_STEP: 71950 | > loss: -0.55714 (-0.52584) | > log_mle: -0.68040 (-0.63817) | > loss_dur: 0.12326 (0.11233) | > amp_scaler: 1024.00000 (1956.97778) | > grad_norm: 17.69923 (11.07762) | > current_lr: 0.00024 | > step_time: 0.82020 (0.49041) | > loader_time: 0.00390 (0.00310)  --> STEP: 295/639 -- GLOBAL_STEP: 71975 | > loss: -0.47843 (-0.52333) | > log_mle: -0.57424 (-0.63593) | > loss_dur: 0.09581 (0.11259) | > amp_scaler: 1024.00000 (1877.91186) | > grad_norm: 3.69890 (10.85828) | > current_lr: 0.00024 | > step_time: 0.62940 (0.50472) | > loader_time: 0.00360 (0.00314)  --> STEP: 320/639 -- GLOBAL_STEP: 72000 | > loss: -0.48830 (-0.52144) | > log_mle: -0.61629 (-0.63438) | > loss_dur: 0.12799 (0.11293) | > amp_scaler: 1024.00000 (1811.20000) | > grad_norm: 15.22432 (10.76928) | > current_lr: 0.00024 | > step_time: 0.73150 (0.52008) | > loader_time: 0.00360 (0.00317)  --> STEP: 345/639 -- GLOBAL_STEP: 72025 | > loss: -0.49457 (-0.51920) | > log_mle: -0.60393 (-0.63241) | > loss_dur: 0.10936 (0.11321) | > amp_scaler: 1024.00000 (1754.15652) | > grad_norm: 9.05101 (10.67961) | > current_lr: 0.00024 | > step_time: 0.64800 (0.53425) | > loader_time: 0.00340 (0.00321)  --> STEP: 370/639 -- GLOBAL_STEP: 72050 | > loss: -0.50313 (-0.51717) | > log_mle: -0.61663 (-0.63018) | > loss_dur: 0.11350 (0.11301) | > amp_scaler: 1024.00000 (1704.82162) | > grad_norm: 10.24416 (10.46484) | > current_lr: 0.00024 | > step_time: 0.87250 (0.54577) | > loader_time: 0.00380 (0.00323)  --> STEP: 395/639 -- GLOBAL_STEP: 72075 | > loss: -0.47293 (-0.51506) | > log_mle: -0.58616 (-0.62802) | > loss_dur: 0.11323 (0.11295) | > amp_scaler: 1024.00000 (1661.73165) | > grad_norm: 7.38794 (10.28699) | > current_lr: 0.00024 | > step_time: 0.71260 (0.55774) | > loader_time: 0.00350 (0.00325)  --> STEP: 420/639 -- GLOBAL_STEP: 72100 | > loss: -0.46691 (-0.51273) | > log_mle: -0.58064 (-0.62603) | > loss_dur: 0.11373 (0.11329) | > amp_scaler: 1024.00000 (1623.77143) | > grad_norm: 10.26733 (10.20687) | > current_lr: 0.00024 | > step_time: 0.68980 (0.57030) | > loader_time: 0.00360 (0.00327)  --> STEP: 445/639 -- GLOBAL_STEP: 72125 | > loss: -0.47855 (-0.51105) | > log_mle: -0.59086 (-0.62444) | > loss_dur: 0.11232 (0.11339) | > amp_scaler: 1024.00000 (1590.07640) | > grad_norm: 6.58754 (10.10320) | > current_lr: 0.00024 | > step_time: 0.72030 (0.58270) | > loader_time: 0.00360 (0.00330)  --> STEP: 470/639 -- GLOBAL_STEP: 72150 | > loss: -0.48648 (-0.50909) | > log_mle: -0.60034 (-0.62246) | > loss_dur: 0.11386 (0.11337) | > amp_scaler: 1024.00000 (1559.96596) | > grad_norm: 9.62534 (9.96074) | > current_lr: 0.00024 | > step_time: 0.76040 (0.59408) | > loader_time: 0.00380 (0.00333)  --> STEP: 495/639 -- GLOBAL_STEP: 72175 | > loss: -0.44119 (-0.50751) | > log_mle: -0.53904 (-0.62099) | > loss_dur: 0.09784 (0.11348) | > amp_scaler: 1024.00000 (1532.89697) | > grad_norm: 1.80676 (9.83412) | > current_lr: 0.00024 | > step_time: 0.81890 (0.60680) | > loader_time: 0.00380 (0.00336)  --> STEP: 520/639 -- GLOBAL_STEP: 72200 | > loss: -0.47580 (-0.50619) | > log_mle: -0.58673 (-0.61972) | > loss_dur: 0.11094 (0.11352) | > amp_scaler: 1024.00000 (1508.43077) | > grad_norm: 26.13968 (9.85697) | > current_lr: 0.00024 | > step_time: 0.86610 (0.62027) | > loader_time: 0.00380 (0.00338)  --> STEP: 545/639 -- GLOBAL_STEP: 72225 | > loss: -0.49121 (-0.50465) | > log_mle: -0.62348 (-0.61837) | > loss_dur: 0.13228 (0.11372) | > amp_scaler: 1024.00000 (1486.20917) | > grad_norm: 9.21529 (9.77066) | > current_lr: 0.00024 | > step_time: 0.93560 (0.63546) | > loader_time: 0.00370 (0.00341)  --> STEP: 570/639 -- GLOBAL_STEP: 72250 | > loss: -0.46686 (-0.50299) | > log_mle: -0.57278 (-0.61679) | > loss_dur: 0.10592 (0.11380) | > amp_scaler: 1024.00000 (1465.93684) | > grad_norm: 4.37959 (9.65971) | > current_lr: 0.00024 | > step_time: 0.89280 (0.64897) | > loader_time: 0.00400 (0.00345)  --> STEP: 595/639 -- GLOBAL_STEP: 72275 | > loss: -0.45367 (-0.50176) | > log_mle: -0.56526 (-0.61576) | > loss_dur: 0.11158 (0.11400) | > amp_scaler: 1024.00000 (1447.36807) | > grad_norm: 5.71400 (9.59941) | > current_lr: 0.00024 | > step_time: 0.95700 (0.66487) | > loader_time: 0.00430 (0.00348)  --> STEP: 620/639 -- GLOBAL_STEP: 72300 | > loss: -0.46216 (-0.50038) | > log_mle: -0.57164 (-0.61453) | > loss_dur: 0.10948 (0.11415) | > amp_scaler: 1024.00000 (1430.29677) | > grad_norm: 7.74336 (9.53673) | > current_lr: 0.00024 | > step_time: 1.17610 (0.68102) | > loader_time: 0.00490 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02839 (-0.00046) | > avg_loss: -0.45098 (+0.00434) | > avg_log_mle: -0.58692 (+0.00563) | > avg_loss_dur: 0.13594 (-0.00129)  > EPOCH: 113/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 05:16:34)   --> STEP: 5/639 -- GLOBAL_STEP: 72325 | > loss: -0.55484 (-0.56838) | > log_mle: -0.70973 (-0.68943) | > loss_dur: 0.15489 (0.12106) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 21.76192 (13.76792) | > current_lr: 0.00024 | > step_time: 0.30790 (0.27759) | > loader_time: 0.00130 (0.00188)  --> STEP: 30/639 -- GLOBAL_STEP: 72350 | > loss: -0.55820 (-0.55576) | > log_mle: -0.66493 (-0.66931) | > loss_dur: 0.10673 (0.11356) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 11.74697 (12.24652) | > current_lr: 0.00024 | > step_time: 0.32050 (0.32108) | > loader_time: 0.00280 (0.00256)  --> STEP: 55/639 -- GLOBAL_STEP: 72375 | > loss: -0.51879 (-0.54691) | > log_mle: -0.60337 (-0.65867) | > loss_dur: 0.08458 (0.11176) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 4.71069 (11.22016) | > current_lr: 0.00024 | > step_time: 0.32970 (0.34477) | > loader_time: 0.00290 (0.00267)  --> STEP: 80/639 -- GLOBAL_STEP: 72400 | > loss: -0.50938 (-0.53901) | > log_mle: -0.63417 (-0.64952) | > loss_dur: 0.12479 (0.11051) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 10.30848 (10.86181) | > current_lr: 0.00024 | > step_time: 0.39300 (0.36556) | > loader_time: 0.00260 (0.00274)  --> STEP: 105/639 -- GLOBAL_STEP: 72425 | > loss: -0.47726 (-0.53655) | > log_mle: -0.58820 (-0.64649) | > loss_dur: 0.11094 (0.10994) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 8.79955 (10.70180) | > current_lr: 0.00024 | > step_time: 0.41340 (0.38373) | > loader_time: 0.00290 (0.00280)  --> STEP: 130/639 -- GLOBAL_STEP: 72450 | > loss: -0.51116 (-0.53384) | > log_mle: -0.62951 (-0.64420) | > loss_dur: 0.11835 (0.11035) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 8.66622 (10.74799) | > current_lr: 0.00024 | > step_time: 0.49000 (0.39873) | > loader_time: 0.00320 (0.00282)  --> STEP: 155/639 -- GLOBAL_STEP: 72475 | > loss: -0.49695 (-0.53218) | > log_mle: -0.61638 (-0.64303) | > loss_dur: 0.11942 (0.11084) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 8.75899 (10.73970) | > current_lr: 0.00023 | > step_time: 0.54270 (0.41643) | > loader_time: 0.00320 (0.00288)  --> STEP: 180/639 -- GLOBAL_STEP: 72500 | > loss: -0.55436 (-0.53282) | > log_mle: -0.67411 (-0.64417) | > loss_dur: 0.11975 (0.11135) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 15.31941 (11.02508) | > current_lr: 0.00023 | > step_time: 0.57810 (0.43297) | > loader_time: 0.00350 (0.00292)  --> STEP: 205/639 -- GLOBAL_STEP: 72525 | > loss: -0.47823 (-0.52930) | > log_mle: -0.58962 (-0.64019) | > loss_dur: 0.11139 (0.11089) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 7.40355 (10.72803) | > current_lr: 0.00023 | > step_time: 0.54940 (0.44752) | > loader_time: 0.00320 (0.00296)  --> STEP: 230/639 -- GLOBAL_STEP: 72550 | > loss: -0.51869 (-0.52848) | > log_mle: -0.62833 (-0.63972) | > loss_dur: 0.10964 (0.11124) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 9.25728 (10.89401) | > current_lr: 0.00023 | > step_time: 0.64190 (0.46489) | > loader_time: 0.00360 (0.00300)  --> STEP: 255/639 -- GLOBAL_STEP: 72575 | > loss: -0.48202 (-0.52706) | > log_mle: -0.59570 (-0.63834) | > loss_dur: 0.11367 (0.11128) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 10.44311 (10.88743) | > current_lr: 0.00023 | > step_time: 0.63410 (0.47775) | > loader_time: 0.00340 (0.00303)  --> STEP: 280/639 -- GLOBAL_STEP: 72600 | > loss: -0.50107 (-0.52643) | > log_mle: -0.62390 (-0.63841) | > loss_dur: 0.12283 (0.11198) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 10.77949 (11.03251) | > current_lr: 0.00023 | > step_time: 0.62610 (0.49369) | > loader_time: 0.00340 (0.00307)  --> STEP: 305/639 -- GLOBAL_STEP: 72625 | > loss: -0.48720 (-0.52418) | > log_mle: -0.60179 (-0.63648) | > loss_dur: 0.11460 (0.11230) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 10.78611 (11.00468) | > current_lr: 0.00023 | > step_time: 0.67370 (0.50733) | > loader_time: 0.00310 (0.00310)  --> STEP: 330/639 -- GLOBAL_STEP: 72650 | > loss: -0.49304 (-0.52132) | > log_mle: -0.62410 (-0.63418) | > loss_dur: 0.13105 (0.11287) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 9.98929 (10.96103) | > current_lr: 0.00023 | > step_time: 0.72640 (0.52169) | > loader_time: 0.00340 (0.00313)  --> STEP: 355/639 -- GLOBAL_STEP: 72675 | > loss: -0.45192 (-0.51934) | > log_mle: -0.55553 (-0.63207) | > loss_dur: 0.10361 (0.11273) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 2.65286 (10.85320) | > current_lr: 0.00023 | > step_time: 0.71580 (0.53438) | > loader_time: 0.00300 (0.00316)  --> STEP: 380/639 -- GLOBAL_STEP: 72700 | > loss: -0.48452 (-0.51720) | > log_mle: -0.59011 (-0.62981) | > loss_dur: 0.10559 (0.11261) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 6.15112 (10.68435) | > current_lr: 0.00023 | > step_time: 0.62640 (0.54721) | > loader_time: 0.00360 (0.00318)  --> STEP: 405/639 -- GLOBAL_STEP: 72725 | > loss: -0.48412 (-0.51552) | > log_mle: -0.59264 (-0.62842) | > loss_dur: 0.10852 (0.11290) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 7.58170 (10.64595) | > current_lr: 0.00023 | > step_time: 0.67170 (0.55903) | > loader_time: 0.00330 (0.00321)  --> STEP: 430/639 -- GLOBAL_STEP: 72750 | > loss: -0.44954 (-0.51336) | > log_mle: -0.55338 (-0.62641) | > loss_dur: 0.10385 (0.11305) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 5.54862 (10.51958) | > current_lr: 0.00023 | > step_time: 0.84600 (0.57157) | > loader_time: 0.00380 (0.00324)  --> STEP: 455/639 -- GLOBAL_STEP: 72775 | > loss: -0.46978 (-0.51139) | > log_mle: -0.57897 (-0.62436) | > loss_dur: 0.10919 (0.11296) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 6.42545 (10.36766) | > current_lr: 0.00023 | > step_time: 0.73340 (0.58264) | > loader_time: 0.00360 (0.00326)  --> STEP: 480/639 -- GLOBAL_STEP: 72800 | > loss: -0.42731 (-0.50949) | > log_mle: -0.53623 (-0.62269) | > loss_dur: 0.10892 (0.11320) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 2.90874 (10.24778) | > current_lr: 0.00023 | > step_time: 0.77580 (0.59476) | > loader_time: 0.00330 (0.00329)  --> STEP: 505/639 -- GLOBAL_STEP: 72825 | > loss: -0.48865 (-0.50807) | > log_mle: -0.59914 (-0.62121) | > loss_dur: 0.11048 (0.11315) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 14.27431 (10.13434) | > current_lr: 0.00023 | > step_time: 0.90070 (0.60678) | > loader_time: 0.00390 (0.00331)  --> STEP: 530/639 -- GLOBAL_STEP: 72850 | > loss: -0.46656 (-0.50650) | > log_mle: -0.58983 (-0.61987) | > loss_dur: 0.12327 (0.11337) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 8.09945 (10.06063) | > current_lr: 0.00023 | > step_time: 0.93700 (0.62144) | > loader_time: 0.00380 (0.00335)  --> STEP: 555/639 -- GLOBAL_STEP: 72875 | > loss: -0.46855 (-0.50488) | > log_mle: -0.57579 (-0.61829) | > loss_dur: 0.10724 (0.11341) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 4.72893 (9.93820) | > current_lr: 0.00023 | > step_time: 0.83910 (0.63538) | > loader_time: 0.00400 (0.00338)  --> STEP: 580/639 -- GLOBAL_STEP: 72900 | > loss: -0.48843 (-0.50348) | > log_mle: -0.59118 (-0.61705) | > loss_dur: 0.10276 (0.11357) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 9.78134 (9.86513) | > current_lr: 0.00023 | > step_time: 0.98650 (0.65068) | > loader_time: 0.00420 (0.00342)  --> STEP: 605/639 -- GLOBAL_STEP: 72925 | > loss: -0.45554 (-0.50216) | > log_mle: -0.56228 (-0.61589) | > loss_dur: 0.10675 (0.11373) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 4.16029 (9.78005) | > current_lr: 0.00023 | > step_time: 1.04670 (0.66660) | > loader_time: 0.00430 (0.00347)  --> STEP: 630/639 -- GLOBAL_STEP: 72950 | > loss: -0.46358 (-0.50092) | > log_mle: -0.59231 (-0.61489) | > loss_dur: 0.12874 (0.11397) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 7.93532 (9.73669) | > current_lr: 0.00023 | > step_time: 1.20840 (0.68378) | > loader_time: 0.00490 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02640 (-0.00199) | > avg_loss: -0.45390 (-0.00292) | > avg_log_mle: -0.59177 (-0.00485) | > avg_loss_dur: 0.13787 (+0.00193)  > EPOCH: 114/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 05:24:08)   --> STEP: 15/639 -- GLOBAL_STEP: 72975 | > loss: -0.65276 (-0.56992) | > log_mle: -0.79444 (-0.68353) | > loss_dur: 0.14168 (0.11361) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 27.14670 (12.43508) | > current_lr: 0.00023 | > step_time: 0.32750 (0.29244) | > loader_time: 0.00280 (0.00348)  --> STEP: 40/639 -- GLOBAL_STEP: 73000 | > loss: -0.52365 (-0.55189) | > log_mle: -0.63370 (-0.66411) | > loss_dur: 0.11006 (0.11221) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 9.85942 (11.67467) | > current_lr: 0.00023 | > step_time: 0.40790 (0.33749) | > loader_time: 0.00280 (0.00303)  --> STEP: 65/639 -- GLOBAL_STEP: 73025 | > loss: -0.52644 (-0.54385) | > log_mle: -0.62247 (-0.65579) | > loss_dur: 0.09604 (0.11195) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 12.08207 (10.99808) | > current_lr: 0.00023 | > step_time: 0.38500 (0.36477) | > loader_time: 0.00300 (0.00300)  --> STEP: 90/639 -- GLOBAL_STEP: 73050 | > loss: -0.47733 (-0.53594) | > log_mle: -0.58019 (-0.64588) | > loss_dur: 0.10286 (0.10994) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 10.07627 (10.32849) | > current_lr: 0.00023 | > step_time: 0.51660 (0.37907) | > loader_time: 0.00320 (0.00299)  --> STEP: 115/639 -- GLOBAL_STEP: 73075 | > loss: -0.45319 (-0.53501) | > log_mle: -0.55364 (-0.64533) | > loss_dur: 0.10045 (0.11032) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 3.84051 (10.42569) | > current_lr: 0.00023 | > step_time: 0.41520 (0.39535) | > loader_time: 0.00310 (0.00300)  --> STEP: 140/639 -- GLOBAL_STEP: 73100 | > loss: -0.49520 (-0.53383) | > log_mle: -0.62992 (-0.64502) | > loss_dur: 0.13472 (0.11119) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 10.65125 (10.65297) | > current_lr: 0.00023 | > step_time: 0.55100 (0.40893) | > loader_time: 0.00330 (0.00301)  --> STEP: 165/639 -- GLOBAL_STEP: 73125 | > loss: -0.51621 (-0.53239) | > log_mle: -0.63643 (-0.64354) | > loss_dur: 0.12022 (0.11115) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 10.68312 (10.76100) | > current_lr: 0.00023 | > step_time: 0.48800 (0.42156) | > loader_time: 0.00280 (0.00303)  --> STEP: 190/639 -- GLOBAL_STEP: 73150 | > loss: -0.45488 (-0.53137) | > log_mle: -0.56731 (-0.64264) | > loss_dur: 0.11243 (0.11126) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 4.61891 (10.86154) | > current_lr: 0.00023 | > step_time: 0.63910 (0.43899) | > loader_time: 0.00330 (0.00306)  --> STEP: 215/639 -- GLOBAL_STEP: 73175 | > loss: -0.55149 (-0.52854) | > log_mle: -0.66308 (-0.63991) | > loss_dur: 0.11159 (0.11137) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 13.38044 (10.80075) | > current_lr: 0.00023 | > step_time: 0.53770 (0.45586) | > loader_time: 0.00330 (0.00309)  --> STEP: 240/639 -- GLOBAL_STEP: 73200 | > loss: -0.55951 (-0.52879) | > log_mle: -0.67902 (-0.64022) | > loss_dur: 0.11951 (0.11143) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 13.50674 (10.86190) | > current_lr: 0.00023 | > step_time: 0.65580 (0.47021) | > loader_time: 0.00340 (0.00312)  --> STEP: 265/639 -- GLOBAL_STEP: 73225 | > loss: -0.56787 (-0.52732) | > log_mle: -0.70222 (-0.63922) | > loss_dur: 0.13436 (0.11189) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 16.16958 (10.91364) | > current_lr: 0.00023 | > step_time: 0.61930 (0.48456) | > loader_time: 0.00340 (0.00316)  --> STEP: 290/639 -- GLOBAL_STEP: 73250 | > loss: -0.48488 (-0.52534) | > log_mle: -0.59791 (-0.63746) | > loss_dur: 0.11303 (0.11212) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 7.89361 (10.88728) | > current_lr: 0.00023 | > step_time: 0.66900 (0.49909) | > loader_time: 0.00340 (0.00318)  --> STEP: 315/639 -- GLOBAL_STEP: 73275 | > loss: -0.47793 (-0.52345) | > log_mle: -0.57823 (-0.63574) | > loss_dur: 0.10030 (0.11229) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 7.32809 (10.79943) | > current_lr: 0.00023 | > step_time: 0.65400 (0.51330) | > loader_time: 0.00300 (0.00320)  --> STEP: 340/639 -- GLOBAL_STEP: 73300 | > loss: -0.49101 (-0.52121) | > log_mle: -0.59759 (-0.63410) | > loss_dur: 0.10658 (0.11289) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 7.68059 (10.79197) | > current_lr: 0.00023 | > step_time: 0.60500 (0.52874) | > loader_time: 0.00350 (0.00324)  --> STEP: 365/639 -- GLOBAL_STEP: 73325 | > loss: -0.48679 (-0.51888) | > log_mle: -0.60428 (-0.63155) | > loss_dur: 0.11749 (0.11268) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 7.02855 (10.57888) | > current_lr: 0.00023 | > step_time: 0.76230 (0.54159) | > loader_time: 0.00390 (0.00325)  --> STEP: 390/639 -- GLOBAL_STEP: 73350 | > loss: -0.47692 (-0.51677) | > log_mle: -0.60136 (-0.62942) | > loss_dur: 0.12445 (0.11265) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 8.78528 (10.37909) | > current_lr: 0.00023 | > step_time: 0.73380 (0.55468) | > loader_time: 0.00360 (0.00327)  --> STEP: 415/639 -- GLOBAL_STEP: 73375 | > loss: -0.46107 (-0.51462) | > log_mle: -0.57828 (-0.62759) | > loss_dur: 0.11721 (0.11297) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 5.83925 (10.26019) | > current_lr: 0.00023 | > step_time: 0.70410 (0.56688) | > loader_time: 0.00340 (0.00330)  --> STEP: 440/639 -- GLOBAL_STEP: 73400 | > loss: -0.45260 (-0.51281) | > log_mle: -0.56263 (-0.62589) | > loss_dur: 0.11004 (0.11307) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 6.89155 (10.20285) | > current_lr: 0.00023 | > step_time: 0.71790 (0.57887) | > loader_time: 0.00350 (0.00332)  --> STEP: 465/639 -- GLOBAL_STEP: 73425 | > loss: -0.45310 (-0.51071) | > log_mle: -0.56688 (-0.62377) | > loss_dur: 0.11378 (0.11306) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 5.54289 (10.03724) | > current_lr: 0.00023 | > step_time: 0.75950 (0.58964) | > loader_time: 0.00340 (0.00334)  --> STEP: 490/639 -- GLOBAL_STEP: 73450 | > loss: -0.45618 (-0.50890) | > log_mle: -0.55059 (-0.62219) | > loss_dur: 0.09441 (0.11329) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 2.68073 (9.94740) | > current_lr: 0.00023 | > step_time: 0.79380 (0.60208) | > loader_time: 0.00380 (0.00337)  --> STEP: 515/639 -- GLOBAL_STEP: 73475 | > loss: -0.47076 (-0.50764) | > log_mle: -0.57967 (-0.62099) | > loss_dur: 0.10890 (0.11335) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 5.56499 (9.85980) | > current_lr: 0.00023 | > step_time: 0.97680 (0.61493) | > loader_time: 0.00410 (0.00339)  --> STEP: 540/639 -- GLOBAL_STEP: 73500 | > loss: -0.47910 (-0.50602) | > log_mle: -0.59610 (-0.61949) | > loss_dur: 0.11700 (0.11347) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 7.97852 (9.78067) | > current_lr: 0.00023 | > step_time: 0.88610 (0.62962) | > loader_time: 0.00430 (0.00343)  --> STEP: 565/639 -- GLOBAL_STEP: 73525 | > loss: -0.45169 (-0.50438) | > log_mle: -0.55820 (-0.61795) | > loss_dur: 0.10651 (0.11357) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 3.17662 (9.66179) | > current_lr: 0.00023 | > step_time: 0.92980 (0.64306) | > loader_time: 0.00390 (0.00347)  --> STEP: 590/639 -- GLOBAL_STEP: 73550 | > loss: -0.47557 (-0.50325) | > log_mle: -0.58839 (-0.61698) | > loss_dur: 0.11282 (0.11373) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 10.61280 (9.61545) | > current_lr: 0.00023 | > step_time: 0.94490 (0.65895) | > loader_time: 0.00680 (0.00351)  --> STEP: 615/639 -- GLOBAL_STEP: 73575 | > loss: -0.47303 (-0.50191) | > log_mle: -0.59139 (-0.61577) | > loss_dur: 0.11836 (0.11386) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 6.12815 (9.55906) | > current_lr: 0.00023 | > step_time: 1.02470 (0.67453) | > loader_time: 0.00430 (0.00355) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02859 (+0.00220) | > avg_loss: -0.46359 (-0.00969) | > avg_log_mle: -0.59776 (-0.00599) | > avg_loss_dur: 0.13417 (-0.00370) > BEST MODEL : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/best_model_73600.pth.tar  > EPOCH: 115/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 05:31:51)   --> STEP: 0/639 -- GLOBAL_STEP: 73600 | > loss: -0.59014 (-0.59014) | > log_mle: -0.73465 (-0.73465) | > loss_dur: 0.14451 (0.14451) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 32.46734 (32.46734) | > current_lr: 0.00023 | > step_time: 0.71940 (0.71939) | > loader_time: 0.48410 (0.48405)  --> STEP: 25/639 -- GLOBAL_STEP: 73625 | > loss: -0.56017 (-0.56326) | > log_mle: -0.67772 (-0.67952) | > loss_dur: 0.11755 (0.11626) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 13.55343 (11.51775) | > current_lr: 0.00023 | > step_time: 0.38170 (0.32009) | > loader_time: 0.00290 (0.00270)  --> STEP: 50/639 -- GLOBAL_STEP: 73650 | > loss: -0.51359 (-0.55129) | > log_mle: -0.61059 (-0.66354) | > loss_dur: 0.09699 (0.11225) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 4.28865 (11.00008) | > current_lr: 0.00023 | > step_time: 0.32880 (0.34578) | > loader_time: 0.00280 (0.00276)  --> STEP: 75/639 -- GLOBAL_STEP: 73675 | > loss: -0.53848 (-0.54230) | > log_mle: -0.63420 (-0.65345) | > loss_dur: 0.09572 (0.11114) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 8.14131 (10.86507) | > current_lr: 0.00023 | > step_time: 0.40960 (0.36791) | > loader_time: 0.00290 (0.00284)  --> STEP: 100/639 -- GLOBAL_STEP: 73700 | > loss: -0.53332 (-0.53898) | > log_mle: -0.64210 (-0.64920) | > loss_dur: 0.10877 (0.11022) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 8.57984 (10.60246) | > current_lr: 0.00023 | > step_time: 0.50890 (0.38563) | > loader_time: 0.00280 (0.00287)  --> STEP: 125/639 -- GLOBAL_STEP: 73725 | > loss: -0.52735 (-0.53537) | > log_mle: -0.63640 (-0.64656) | > loss_dur: 0.10905 (0.11119) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 12.71849 (10.79834) | > current_lr: 0.00023 | > step_time: 0.41320 (0.39841) | > loader_time: 0.00290 (0.00290)  --> STEP: 150/639 -- GLOBAL_STEP: 73750 | > loss: -0.49003 (-0.53338) | > log_mle: -0.59825 (-0.64473) | > loss_dur: 0.10822 (0.11135) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 7.45909 (10.98176) | > current_lr: 0.00023 | > step_time: 0.41570 (0.41419) | > loader_time: 0.00300 (0.00293)  --> STEP: 175/639 -- GLOBAL_STEP: 73775 | > loss: -0.51662 (-0.53265) | > log_mle: -0.61346 (-0.64445) | > loss_dur: 0.09684 (0.11180) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 12.31381 (11.41990) | > current_lr: 0.00023 | > step_time: 0.56890 (0.43090) | > loader_time: 0.00340 (0.00297)  --> STEP: 200/639 -- GLOBAL_STEP: 73800 | > loss: -0.48951 (-0.53037) | > log_mle: -0.59994 (-0.64160) | > loss_dur: 0.11043 (0.11123) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 10.27678 (11.42932) | > current_lr: 0.00023 | > step_time: 0.51520 (0.44503) | > loader_time: 0.00340 (0.00301)  --> STEP: 225/639 -- GLOBAL_STEP: 73825 | > loss: -0.48821 (-0.52800) | > log_mle: -0.58529 (-0.63913) | > loss_dur: 0.09708 (0.11113) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 3.96516 (11.21666) | > current_lr: 0.00023 | > step_time: 0.51070 (0.45925) | > loader_time: 0.00320 (0.00305)  --> STEP: 250/639 -- GLOBAL_STEP: 73850 | > loss: -0.49607 (-0.52823) | > log_mle: -0.58713 (-0.63967) | > loss_dur: 0.09107 (0.11144) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 4.02897 (11.59154) | > current_lr: 0.00023 | > step_time: 0.57290 (0.47305) | > loader_time: 0.00340 (0.00308)  --> STEP: 275/639 -- GLOBAL_STEP: 73875 | > loss: -0.51134 (-0.52770) | > log_mle: -0.62334 (-0.63977) | > loss_dur: 0.11201 (0.11207) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 16.32031 (11.81493) | > current_lr: 0.00023 | > step_time: 0.84060 (0.48914) | > loader_time: 0.00420 (0.00313)  --> STEP: 300/639 -- GLOBAL_STEP: 73900 | > loss: -0.50753 (-0.52516) | > log_mle: -0.61573 (-0.63751) | > loss_dur: 0.10819 (0.11235) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 7.04437 (11.65007) | > current_lr: 0.00023 | > step_time: 0.72950 (0.50246) | > loader_time: 0.00350 (0.00316)  --> STEP: 325/639 -- GLOBAL_STEP: 73925 | > loss: -0.50291 (-0.52283) | > log_mle: -0.61517 (-0.63570) | > loss_dur: 0.11226 (0.11288) | > amp_scaler: 1024.00000 (1024.00000) | > grad_norm: 9.52883 (11.51082) | > current_lr: 0.00023 | > step_time: 0.61330 (0.51748) | > loader_time: 0.00350 (0.00319)  --> STEP: 350/639 -- GLOBAL_STEP: 73950 | > loss: -0.51732 (-0.52069) | > log_mle: -0.63848 (-0.63359) | > loss_dur: 0.12117 (0.11290) | > amp_scaler: 2048.00000 (1094.21714) | > grad_norm: 17.95991 (11.40761) | > current_lr: 0.00023 | > step_time: 0.83550 (0.53241) | > loader_time: 0.00340 (0.00322)  --> STEP: 375/639 -- GLOBAL_STEP: 73975 | > loss: -0.46508 (-0.51832) | > log_mle: -0.57422 (-0.63104) | > loss_dur: 0.10913 (0.11273) | > amp_scaler: 2048.00000 (1157.80267) | > grad_norm: 8.24101 (11.18768) | > current_lr: 0.00023 | > step_time: 0.68840 (0.54478) | > loader_time: 0.00350 (0.00325)  --> STEP: 400/639 -- GLOBAL_STEP: 74000 | > loss: -0.49324 (-0.51652) | > log_mle: -0.60629 (-0.62948) | > loss_dur: 0.11305 (0.11296) | > amp_scaler: 2048.00000 (1213.44000) | > grad_norm: 10.53351 (11.05583) | > current_lr: 0.00023 | > step_time: 0.71580 (0.55695) | > loader_time: 0.00380 (0.00327)  --> STEP: 425/639 -- GLOBAL_STEP: 74025 | > loss: -0.53864 (-0.51421) | > log_mle: -0.66025 (-0.62731) | > loss_dur: 0.12162 (0.11310) | > amp_scaler: 2048.00000 (1262.53176) | > grad_norm: 20.82897 (10.94031) | > current_lr: 0.00023 | > step_time: 0.90870 (0.56959) | > loader_time: 0.00400 (0.00330)  --> STEP: 450/639 -- GLOBAL_STEP: 74050 | > loss: -0.46289 (-0.51229) | > log_mle: -0.56490 (-0.62539) | > loss_dur: 0.10201 (0.11310) | > amp_scaler: 2048.00000 (1306.16889) | > grad_norm: 4.48222 (10.80268) | > current_lr: 0.00023 | > step_time: 0.80840 (0.58150) | > loader_time: 0.00370 (0.00333)  --> STEP: 475/639 -- GLOBAL_STEP: 74075 | > loss: -0.47821 (-0.51036) | > log_mle: -0.57545 (-0.62363) | > loss_dur: 0.09725 (0.11327) | > amp_scaler: 2048.00000 (1345.21263) | > grad_norm: 4.46855 (10.68223) | > current_lr: 0.00023 | > step_time: 0.74730 (0.59325) | > loader_time: 0.00350 (0.00335)  --> STEP: 500/639 -- GLOBAL_STEP: 74100 | > loss: -0.51915 (-0.50892) | > log_mle: -0.64023 (-0.62231) | > loss_dur: 0.12108 (0.11339) | > amp_scaler: 2048.00000 (1380.35200) | > grad_norm: 17.61822 (10.59275) | > current_lr: 0.00023 | > step_time: 0.92860 (0.60624) | > loader_time: 0.00390 (0.00337)  --> STEP: 525/639 -- GLOBAL_STEP: 74125 | > loss: -0.49243 (-0.50739) | > log_mle: -0.61958 (-0.62094) | > loss_dur: 0.12715 (0.11354) | > amp_scaler: 2048.00000 (1412.14476) | > grad_norm: 9.84488 (10.55374) | > current_lr: 0.00023 | > step_time: 1.02170 (0.61994) | > loader_time: 0.00430 (0.00340)  --> STEP: 550/639 -- GLOBAL_STEP: 74150 | > loss: -0.43906 (-0.50580) | > log_mle: -0.55260 (-0.61939) | > loss_dur: 0.11354 (0.11359) | > amp_scaler: 2048.00000 (1441.04727) | > grad_norm: 3.69467 (10.40402) | > current_lr: 0.00023 | > step_time: 0.92710 (0.63416) | > loader_time: 0.00420 (0.00344)  --> STEP: 575/639 -- GLOBAL_STEP: 74175 | > loss: -0.46204 (-0.50418) | > log_mle: -0.57209 (-0.61790) | > loss_dur: 0.11005 (0.11372) | > amp_scaler: 2048.00000 (1467.43652) | > grad_norm: 4.39382 (10.30186) | > current_lr: 0.00023 | > step_time: 0.89590 (0.64802) | > loader_time: 0.00370 (0.00347)  --> STEP: 600/639 -- GLOBAL_STEP: 74200 | > loss: -0.45003 (-0.50292) | > log_mle: -0.56728 (-0.61680) | > loss_dur: 0.11725 (0.11388) | > amp_scaler: 2048.00000 (1491.62667) | > grad_norm: 5.98109 (10.23813) | > current_lr: 0.00023 | > step_time: 1.03840 (0.66411) | > loader_time: 0.00440 (0.00351)  --> STEP: 625/639 -- GLOBAL_STEP: 74225 | > loss: -0.47953 (-0.50169) | > log_mle: -0.59585 (-0.61567) | > loss_dur: 0.11632 (0.11398) | > amp_scaler: 2048.00000 (1513.88160) | > grad_norm: 7.97131 (10.15049) | > current_lr: 0.00023 | > step_time: 1.06250 (0.68025) | > loader_time: 0.00450 (0.00354) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02359 (-0.00500) | > avg_loss: -0.45764 (+0.00595) | > avg_log_mle: -0.59212 (+0.00564) | > avg_loss_dur: 0.13447 (+0.00030)  > EPOCH: 116/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 05:39:26)   --> STEP: 10/639 -- GLOBAL_STEP: 74250 | > loss: -0.54042 (-0.56552) | > log_mle: -0.67445 (-0.68179) | > loss_dur: 0.13403 (0.11627) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.93313 (12.75560) | > current_lr: 0.00023 | > step_time: 0.28610 (0.29395) | > loader_time: 0.00220 (0.00249)  --> STEP: 35/639 -- GLOBAL_STEP: 74275 | > loss: -0.54445 (-0.55633) | > log_mle: -0.64253 (-0.66739) | > loss_dur: 0.09808 (0.11106) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.67853 (11.30107) | > current_lr: 0.00023 | > step_time: 0.32230 (0.32340) | > loader_time: 0.00240 (0.00266)  --> STEP: 60/639 -- GLOBAL_STEP: 74300 | > loss: -0.50574 (-0.54739) | > log_mle: -0.60486 (-0.65882) | > loss_dur: 0.09912 (0.11143) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.07900 (11.05456) | > current_lr: 0.00023 | > step_time: 0.34180 (0.35412) | > loader_time: 0.00280 (0.00275)  --> STEP: 85/639 -- GLOBAL_STEP: 74325 | > loss: -0.50717 (-0.54021) | > log_mle: -0.60463 (-0.64936) | > loss_dur: 0.09746 (0.10915) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.22466 (10.54326) | > current_lr: 0.00023 | > step_time: 0.49880 (0.37152) | > loader_time: 0.00290 (0.00281)  --> STEP: 110/639 -- GLOBAL_STEP: 74350 | > loss: -0.51049 (-0.53922) | > log_mle: -0.61509 (-0.64973) | > loss_dur: 0.10460 (0.11052) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.34064 (11.27577) | > current_lr: 0.00023 | > step_time: 0.39730 (0.39224) | > loader_time: 0.00300 (0.00287)  --> STEP: 135/639 -- GLOBAL_STEP: 74375 | > loss: -0.47514 (-0.53574) | > log_mle: -0.59082 (-0.64641) | > loss_dur: 0.11568 (0.11067) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.71104 (11.11757) | > current_lr: 0.00023 | > step_time: 0.46870 (0.40345) | > loader_time: 0.00300 (0.00290)  --> STEP: 160/639 -- GLOBAL_STEP: 74400 | > loss: -0.52205 (-0.53315) | > log_mle: -0.62870 (-0.64373) | > loss_dur: 0.10665 (0.11059) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 17.63431 (11.22692) | > current_lr: 0.00023 | > step_time: 0.55370 (0.41782) | > loader_time: 0.00290 (0.00292)  --> STEP: 185/639 -- GLOBAL_STEP: 74425 | > loss: -0.50197 (-0.53378) | > log_mle: -0.61212 (-0.64515) | > loss_dur: 0.11015 (0.11137) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.91373 (11.80374) | > current_lr: 0.00023 | > step_time: 0.61040 (0.43497) | > loader_time: 0.00330 (0.00296)  --> STEP: 210/639 -- GLOBAL_STEP: 74450 | > loss: -0.43589 (-0.53013) | > log_mle: -0.53601 (-0.64154) | > loss_dur: 0.10012 (0.11142) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 1.47809 (11.59154) | > current_lr: 0.00023 | > step_time: 0.49840 (0.45113) | > loader_time: 0.00300 (0.00298)  --> STEP: 235/639 -- GLOBAL_STEP: 74475 | > loss: -0.51551 (-0.53000) | > log_mle: -0.62494 (-0.64145) | > loss_dur: 0.10943 (0.11145) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.93066 (11.58433) | > current_lr: 0.00023 | > step_time: 0.54740 (0.46633) | > loader_time: 0.00340 (0.00303)  --> STEP: 260/639 -- GLOBAL_STEP: 74500 | > loss: -0.50538 (-0.52884) | > log_mle: -0.62526 (-0.64063) | > loss_dur: 0.11988 (0.11180) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.74097 (11.50694) | > current_lr: 0.00023 | > step_time: 0.58680 (0.48233) | > loader_time: 0.00340 (0.00306)  --> STEP: 285/639 -- GLOBAL_STEP: 74525 | > loss: -0.46224 (-0.52704) | > log_mle: -0.57363 (-0.63923) | > loss_dur: 0.11138 (0.11219) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.20075 (11.39200) | > current_lr: 0.00023 | > step_time: 0.62480 (0.49788) | > loader_time: 0.00330 (0.00311)  --> STEP: 310/639 -- GLOBAL_STEP: 74550 | > loss: -0.49533 (-0.52480) | > log_mle: -0.61314 (-0.63739) | > loss_dur: 0.11781 (0.11259) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.36353 (11.33489) | > current_lr: 0.00023 | > step_time: 0.58210 (0.51069) | > loader_time: 0.00340 (0.00314)  --> STEP: 335/639 -- GLOBAL_STEP: 74575 | > loss: -0.51595 (-0.52202) | > log_mle: -0.63982 (-0.63517) | > loss_dur: 0.12387 (0.11316) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.80680 (11.16500) | > current_lr: 0.00023 | > step_time: 0.72920 (0.52628) | > loader_time: 0.00370 (0.00318)  --> STEP: 360/639 -- GLOBAL_STEP: 74600 | > loss: -0.46597 (-0.52006) | > log_mle: -0.57211 (-0.63303) | > loss_dur: 0.10614 (0.11297) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.58980 (10.98839) | > current_lr: 0.00023 | > step_time: 0.67970 (0.53935) | > loader_time: 0.00350 (0.00321)  --> STEP: 385/639 -- GLOBAL_STEP: 74625 | > loss: -0.49808 (-0.51821) | > log_mle: -0.60842 (-0.63108) | > loss_dur: 0.11034 (0.11287) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.54329 (10.80805) | > current_lr: 0.00023 | > step_time: 0.64600 (0.55239) | > loader_time: 0.00390 (0.00324)  --> STEP: 410/639 -- GLOBAL_STEP: 74650 | > loss: -0.45038 (-0.51611) | > log_mle: -0.58255 (-0.62924) | > loss_dur: 0.13217 (0.11312) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.06556 (10.67185) | > current_lr: 0.00023 | > step_time: 0.76480 (0.56439) | > loader_time: 0.00390 (0.00327)  --> STEP: 435/639 -- GLOBAL_STEP: 74675 | > loss: -0.48741 (-0.51385) | > log_mle: -0.59431 (-0.62707) | > loss_dur: 0.10690 (0.11322) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.87857 (10.52451) | > current_lr: 0.00023 | > step_time: 0.71340 (0.57616) | > loader_time: 0.00370 (0.00330)  --> STEP: 460/639 -- GLOBAL_STEP: 74700 | > loss: -0.44533 (-0.51173) | > log_mle: -0.55967 (-0.62493) | > loss_dur: 0.11434 (0.11320) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.38042 (10.33403) | > current_lr: 0.00023 | > step_time: 0.94310 (0.58748) | > loader_time: 0.00400 (0.00332)  --> STEP: 485/639 -- GLOBAL_STEP: 74725 | > loss: -0.47622 (-0.50998) | > log_mle: -0.59119 (-0.62339) | > loss_dur: 0.11497 (0.11341) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.09255 (10.24806) | > current_lr: 0.00023 | > step_time: 0.86460 (0.60048) | > loader_time: 0.00420 (0.00335)  --> STEP: 510/639 -- GLOBAL_STEP: 74750 | > loss: -0.47558 (-0.50861) | > log_mle: -0.59559 (-0.62204) | > loss_dur: 0.12000 (0.11343) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.66008 (10.16307) | > current_lr: 0.00023 | > step_time: 0.81540 (0.61316) | > loader_time: 0.00410 (0.00338)  --> STEP: 535/639 -- GLOBAL_STEP: 74775 | > loss: -0.45494 (-0.50700) | > log_mle: -0.58347 (-0.62056) | > loss_dur: 0.12853 (0.11356) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.73077 (10.06921) | > current_lr: 0.00023 | > step_time: 0.90500 (0.62624) | > loader_time: 0.00410 (0.00341)  --> STEP: 560/639 -- GLOBAL_STEP: 74800 | > loss: -0.46817 (-0.50544) | > log_mle: -0.59158 (-0.61918) | > loss_dur: 0.12341 (0.11373) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.46351 (10.01916) | > current_lr: 0.00023 | > step_time: 0.99440 (0.64136) | > loader_time: 0.00420 (0.00344)  --> STEP: 585/639 -- GLOBAL_STEP: 74825 | > loss: -0.48016 (-0.50412) | > log_mle: -0.59515 (-0.61794) | > loss_dur: 0.11499 (0.11382) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.15384 (9.93643) | > current_lr: 0.00023 | > step_time: 1.10010 (0.65654) | > loader_time: 0.00450 (0.00348)  --> STEP: 610/639 -- GLOBAL_STEP: 74850 | > loss: -0.45095 (-0.50272) | > log_mle: -0.56469 (-0.61671) | > loss_dur: 0.11373 (0.11399) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.00610 (9.93303) | > current_lr: 0.00023 | > step_time: 1.05830 (0.67254) | > loader_time: 0.00450 (0.00352)  --> STEP: 635/639 -- GLOBAL_STEP: 74875 | > loss: -0.46752 (-0.50155) | > log_mle: -0.58940 (-0.61587) | > loss_dur: 0.12187 (0.11432) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.12288 (9.92670) | > current_lr: 0.00023 | > step_time: 1.08730 (0.68944) | > loader_time: 0.00420 (0.00356) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02918 (+0.00559) | > avg_loss: -0.46051 (-0.00286) | > avg_log_mle: -0.59571 (-0.00359) | > avg_loss_dur: 0.13520 (+0.00073)  > EPOCH: 117/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 05:47:02)   --> STEP: 20/639 -- GLOBAL_STEP: 74900 | > loss: -0.63740 (-0.57191) | > log_mle: -0.79478 (-0.68680) | > loss_dur: 0.15738 (0.11489) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 27.34435 (15.30919) | > current_lr: 0.00023 | > step_time: 0.34470 (0.30185) | > loader_time: 0.00270 (0.00653)  --> STEP: 45/639 -- GLOBAL_STEP: 74925 | > loss: -0.51665 (-0.54929) | > log_mle: -0.60586 (-0.66144) | > loss_dur: 0.08922 (0.11215) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.30138 (12.87740) | > current_lr: 0.00023 | > step_time: 0.40780 (0.33274) | > loader_time: 0.00280 (0.00445)  --> STEP: 70/639 -- GLOBAL_STEP: 74950 | > loss: -0.53104 (-0.53889) | > log_mle: -0.65557 (-0.65086) | > loss_dur: 0.12452 (0.11197) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.19307 (12.14367) | > current_lr: 0.00023 | > step_time: 0.39050 (0.35725) | > loader_time: 0.00290 (0.00389)  --> STEP: 95/639 -- GLOBAL_STEP: 74975 | > loss: -0.60753 (-0.53635) | > log_mle: -0.70803 (-0.64606) | > loss_dur: 0.10050 (0.10971) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.20580 (11.52049) | > current_lr: 0.00023 | > step_time: 0.40610 (0.37172) | > loader_time: 0.00290 (0.00364)  --> STEP: 120/639 -- GLOBAL_STEP: 75000 | > loss: -0.56941 (-0.53374) | > log_mle: -0.70338 (-0.64446) | > loss_dur: 0.13397 (0.11072) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 18.11696 (11.55855) | > current_lr: 0.00023 | > step_time: 0.52200 (0.38933) | > loader_time: 0.00320 (0.00351)  --> STEP: 145/639 -- GLOBAL_STEP: 75025 | > loss: -0.51995 (-0.53261) | > log_mle: -0.64546 (-0.64413) | > loss_dur: 0.12552 (0.11152) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 19.78095 (11.94642) | > current_lr: 0.00023 | > step_time: 0.47280 (0.40583) | > loader_time: 0.00270 (0.00342)  --> STEP: 170/639 -- GLOBAL_STEP: 75050 | > loss: -0.51888 (-0.53107) | > log_mle: -0.63430 (-0.64267) | > loss_dur: 0.11542 (0.11161) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.77766 (11.78329) | > current_lr: 0.00023 | > step_time: 0.55100 (0.42022) | > loader_time: 0.00340 (0.00337)  --> STEP: 195/639 -- GLOBAL_STEP: 75075 | > loss: -0.49060 (-0.52955) | > log_mle: -0.58134 (-0.64106) | > loss_dur: 0.09074 (0.11152) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.11650 (12.01187) | > current_lr: 0.00023 | > step_time: 0.49240 (0.43723) | > loader_time: 0.00320 (0.00334)  --> STEP: 220/639 -- GLOBAL_STEP: 75100 | > loss: -0.52983 (-0.52699) | > log_mle: -0.62669 (-0.63848) | > loss_dur: 0.09686 (0.11149) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.03868 (11.74853) | > current_lr: 0.00023 | > step_time: 0.64930 (0.45296) | > loader_time: 0.00330 (0.00334)  --> STEP: 245/639 -- GLOBAL_STEP: 75125 | > loss: -0.50948 (-0.52756) | > log_mle: -0.61647 (-0.63914) | > loss_dur: 0.10699 (0.11158) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.97478 (11.86085) | > current_lr: 0.00023 | > step_time: 0.53730 (0.46828) | > loader_time: 0.00310 (0.00334)  --> STEP: 270/639 -- GLOBAL_STEP: 75150 | > loss: -0.55393 (-0.52679) | > log_mle: -0.67845 (-0.63888) | > loss_dur: 0.12452 (0.11209) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 21.69569 (11.92112) | > current_lr: 0.00023 | > step_time: 0.84430 (0.48497) | > loader_time: 0.00360 (0.00335)  --> STEP: 295/639 -- GLOBAL_STEP: 75175 | > loss: -0.48059 (-0.52430) | > log_mle: -0.57429 (-0.63658) | > loss_dur: 0.09370 (0.11227) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.12413 (11.79425) | > current_lr: 0.00023 | > step_time: 0.63900 (0.49853) | > loader_time: 0.00360 (0.00335)  --> STEP: 320/639 -- GLOBAL_STEP: 75200 | > loss: -0.48898 (-0.52209) | > log_mle: -0.61604 (-0.63479) | > loss_dur: 0.12706 (0.11270) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.85635 (11.69263) | > current_lr: 0.00023 | > step_time: 0.84690 (0.51524) | > loader_time: 0.00350 (0.00336)  --> STEP: 345/639 -- GLOBAL_STEP: 75225 | > loss: -0.49397 (-0.51993) | > log_mle: -0.60556 (-0.63287) | > loss_dur: 0.11159 (0.11294) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.23877 (11.51615) | > current_lr: 0.00023 | > step_time: 0.64440 (0.52944) | > loader_time: 0.00340 (0.00338)  --> STEP: 370/639 -- GLOBAL_STEP: 75250 | > loss: -0.50368 (-0.51795) | > log_mle: -0.62155 (-0.63069) | > loss_dur: 0.11787 (0.11274) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.81455 (11.27598) | > current_lr: 0.00023 | > step_time: 0.87110 (0.54143) | > loader_time: 0.00380 (0.00339)  --> STEP: 395/639 -- GLOBAL_STEP: 75275 | > loss: -0.47685 (-0.51588) | > log_mle: -0.58914 (-0.62866) | > loss_dur: 0.11229 (0.11278) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.82636 (11.08367) | > current_lr: 0.00023 | > step_time: 0.76800 (0.55363) | > loader_time: 0.00350 (0.00340)  --> STEP: 420/639 -- GLOBAL_STEP: 75300 | > loss: -0.46915 (-0.51359) | > log_mle: -0.58123 (-0.62670) | > loss_dur: 0.11208 (0.11311) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.75006 (10.93026) | > current_lr: 0.00023 | > step_time: 0.70080 (0.56568) | > loader_time: 0.00350 (0.00341)  --> STEP: 445/639 -- GLOBAL_STEP: 75325 | > loss: -0.48204 (-0.51198) | > log_mle: -0.59131 (-0.62513) | > loss_dur: 0.10927 (0.11314) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.69838 (10.80503) | > current_lr: 0.00023 | > step_time: 0.72300 (0.57777) | > loader_time: 0.00340 (0.00343)  --> STEP: 470/639 -- GLOBAL_STEP: 75350 | > loss: -0.49124 (-0.50997) | > log_mle: -0.60309 (-0.62317) | > loss_dur: 0.11186 (0.11319) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.03484 (10.63638) | > current_lr: 0.00023 | > step_time: 0.89320 (0.59019) | > loader_time: 0.00340 (0.00344)  --> STEP: 495/639 -- GLOBAL_STEP: 75375 | > loss: -0.44122 (-0.50840) | > log_mle: -0.53910 (-0.62169) | > loss_dur: 0.09788 (0.11328) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.68454 (10.52968) | > current_lr: 0.00023 | > step_time: 0.74930 (0.60330) | > loader_time: 0.00370 (0.00346)  --> STEP: 520/639 -- GLOBAL_STEP: 75400 | > loss: -0.48827 (-0.50712) | > log_mle: -0.59600 (-0.62044) | > loss_dur: 0.10773 (0.11332) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.73580 (10.44357) | > current_lr: 0.00023 | > step_time: 0.80050 (0.61604) | > loader_time: 0.00370 (0.00348)  --> STEP: 545/639 -- GLOBAL_STEP: 75425 | > loss: -0.49043 (-0.50562) | > log_mle: -0.62527 (-0.61914) | > loss_dur: 0.13484 (0.11352) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.07613 (10.36286) | > current_lr: 0.00023 | > step_time: 0.94450 (0.63131) | > loader_time: 0.00390 (0.00351)  --> STEP: 570/639 -- GLOBAL_STEP: 75450 | > loss: -0.46358 (-0.50396) | > log_mle: -0.57278 (-0.61760) | > loss_dur: 0.10920 (0.11364) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.78416 (10.24717) | > current_lr: 0.00023 | > step_time: 0.82900 (0.64534) | > loader_time: 0.00380 (0.00354)  --> STEP: 595/639 -- GLOBAL_STEP: 75475 | > loss: -0.45866 (-0.50277) | > log_mle: -0.56589 (-0.61655) | > loss_dur: 0.10723 (0.11378) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.03321 (10.17668) | > current_lr: 0.00023 | > step_time: 0.93270 (0.66105) | > loader_time: 0.00730 (0.00359)  --> STEP: 620/639 -- GLOBAL_STEP: 75500 | > loss: -0.46593 (-0.50144) | > log_mle: -0.57325 (-0.61533) | > loss_dur: 0.10732 (0.11389) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.84422 (10.10804) | > current_lr: 0.00023 | > step_time: 1.07390 (0.67729) | > loader_time: 0.00450 (0.00364) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03012 (+0.00094) | > avg_loss: -0.45834 (+0.00217) | > avg_log_mle: -0.59565 (+0.00006) | > avg_loss_dur: 0.13731 (+0.00210)  > EPOCH: 118/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 05:54:37)   --> STEP: 5/639 -- GLOBAL_STEP: 75525 | > loss: -0.56195 (-0.57133) | > log_mle: -0.70938 (-0.69396) | > loss_dur: 0.14743 (0.12263) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 23.71266 (14.17926) | > current_lr: 0.00023 | > step_time: 0.30750 (0.27712) | > loader_time: 0.00340 (0.00267)  --> STEP: 30/639 -- GLOBAL_STEP: 75550 | > loss: -0.56041 (-0.55969) | > log_mle: -0.67194 (-0.67457) | > loss_dur: 0.11153 (0.11487) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.33387 (12.58991) | > current_lr: 0.00023 | > step_time: 0.32440 (0.32752) | > loader_time: 0.00270 (0.00261)  --> STEP: 55/639 -- GLOBAL_STEP: 75575 | > loss: -0.51534 (-0.55044) | > log_mle: -0.60584 (-0.66223) | > loss_dur: 0.09051 (0.11179) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.00368 (12.99638) | > current_lr: 0.00023 | > step_time: 0.33070 (0.35227) | > loader_time: 0.00290 (0.00270)  --> STEP: 80/639 -- GLOBAL_STEP: 75600 | > loss: -0.50373 (-0.54278) | > log_mle: -0.63037 (-0.65260) | > loss_dur: 0.12665 (0.10983) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.52906 (11.83178) | > current_lr: 0.00023 | > step_time: 0.39760 (0.37051) | > loader_time: 0.00290 (0.00275)  --> STEP: 105/639 -- GLOBAL_STEP: 75625 | > loss: -0.48256 (-0.53980) | > log_mle: -0.59107 (-0.64961) | > loss_dur: 0.10851 (0.10981) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.41866 (11.49744) | > current_lr: 0.00023 | > step_time: 0.41380 (0.38712) | > loader_time: 0.00300 (0.00282)  --> STEP: 130/639 -- GLOBAL_STEP: 75650 | > loss: -0.50935 (-0.53606) | > log_mle: -0.62608 (-0.64700) | > loss_dur: 0.11673 (0.11095) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.24784 (11.32200) | > current_lr: 0.00023 | > step_time: 0.48280 (0.40202) | > loader_time: 0.00320 (0.00285)  --> STEP: 155/639 -- GLOBAL_STEP: 75675 | > loss: -0.50781 (-0.53397) | > log_mle: -0.61492 (-0.64534) | > loss_dur: 0.10711 (0.11137) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 22.74558 (11.45639) | > current_lr: 0.00023 | > step_time: 0.53590 (0.41969) | > loader_time: 0.00330 (0.00291)  --> STEP: 180/639 -- GLOBAL_STEP: 75700 | > loss: -0.54327 (-0.53411) | > log_mle: -0.66970 (-0.64621) | > loss_dur: 0.12644 (0.11210) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 17.50841 (11.61795) | > current_lr: 0.00023 | > step_time: 0.62700 (0.43521) | > loader_time: 0.00340 (0.00295)  --> STEP: 205/639 -- GLOBAL_STEP: 75725 | > loss: -0.48649 (-0.53062) | > log_mle: -0.59543 (-0.64220) | > loss_dur: 0.10893 (0.11159) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.45904 (11.27493) | > current_lr: 0.00023 | > step_time: 0.54810 (0.44884) | > loader_time: 0.00340 (0.00300)  --> STEP: 230/639 -- GLOBAL_STEP: 75750 | > loss: -0.51028 (-0.52997) | > log_mle: -0.62694 (-0.64171) | > loss_dur: 0.11666 (0.11174) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.70872 (11.35596) | > current_lr: 0.00023 | > step_time: 0.54680 (0.46476) | > loader_time: 0.00360 (0.00304)  --> STEP: 255/639 -- GLOBAL_STEP: 75775 | > loss: -0.48465 (-0.52850) | > log_mle: -0.59977 (-0.64025) | > loss_dur: 0.11512 (0.11175) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.83196 (11.25710) | > current_lr: 0.00023 | > step_time: 0.63950 (0.47733) | > loader_time: 0.00350 (0.00308)  --> STEP: 280/639 -- GLOBAL_STEP: 75800 | > loss: -0.50239 (-0.52788) | > log_mle: -0.62474 (-0.64029) | > loss_dur: 0.12234 (0.11241) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.74935 (11.39949) | > current_lr: 0.00023 | > step_time: 0.64230 (0.49423) | > loader_time: 0.00300 (0.00311)  --> STEP: 305/639 -- GLOBAL_STEP: 75825 | > loss: -0.48572 (-0.52574) | > log_mle: -0.60189 (-0.63839) | > loss_dur: 0.11617 (0.11265) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.41601 (11.26089) | > current_lr: 0.00023 | > step_time: 0.67750 (0.50784) | > loader_time: 0.00350 (0.00314)  --> STEP: 330/639 -- GLOBAL_STEP: 75850 | > loss: -0.49586 (-0.52302) | > log_mle: -0.62488 (-0.63616) | > loss_dur: 0.12902 (0.11314) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.16473 (11.20630) | > current_lr: 0.00023 | > step_time: 0.74600 (0.52238) | > loader_time: 0.00360 (0.00318)  --> STEP: 355/639 -- GLOBAL_STEP: 75875 | > loss: -0.45586 (-0.52102) | > log_mle: -0.55597 (-0.63402) | > loss_dur: 0.10011 (0.11300) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.75410 (11.04220) | > current_lr: 0.00023 | > step_time: 0.58350 (0.53593) | > loader_time: 0.00330 (0.00320)  --> STEP: 380/639 -- GLOBAL_STEP: 75900 | > loss: -0.49242 (-0.51878) | > log_mle: -0.59178 (-0.63160) | > loss_dur: 0.09937 (0.11282) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.86817 (10.80262) | > current_lr: 0.00023 | > step_time: 0.61650 (0.54921) | > loader_time: 0.00310 (0.00323)  --> STEP: 405/639 -- GLOBAL_STEP: 75925 | > loss: -0.48908 (-0.51700) | > log_mle: -0.59370 (-0.63010) | > loss_dur: 0.10462 (0.11310) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.81791 (10.72811) | > current_lr: 0.00023 | > step_time: 0.69660 (0.56084) | > loader_time: 0.00340 (0.00326)  --> STEP: 430/639 -- GLOBAL_STEP: 75950 | > loss: -0.45032 (-0.51481) | > log_mle: -0.55542 (-0.62802) | > loss_dur: 0.10510 (0.11322) | > amp_scaler: 4096.00000 (2162.30698) | > grad_norm: 4.34722 (10.62386) | > current_lr: 0.00023 | > step_time: 0.76790 (0.57379) | > loader_time: 0.00380 (0.00329)  --> STEP: 455/639 -- GLOBAL_STEP: 75975 | > loss: -0.46822 (-0.51274) | > log_mle: -0.57799 (-0.62589) | > loss_dur: 0.10977 (0.11314) | > amp_scaler: 4096.00000 (2268.55385) | > grad_norm: 6.86070 (10.46248) | > current_lr: 0.00023 | > step_time: 0.71790 (0.58446) | > loader_time: 0.00340 (0.00331)  --> STEP: 480/639 -- GLOBAL_STEP: 76000 | > loss: -0.42467 (-0.51087) | > log_mle: -0.53695 (-0.62418) | > loss_dur: 0.11228 (0.11332) | > amp_scaler: 4096.00000 (2363.73333) | > grad_norm: 3.33755 (10.36872) | > current_lr: 0.00023 | > step_time: 0.81120 (0.59680) | > loader_time: 0.00370 (0.00333)  --> STEP: 505/639 -- GLOBAL_STEP: 76025 | > loss: -0.49017 (-0.50934) | > log_mle: -0.60116 (-0.62264) | > loss_dur: 0.11099 (0.11330) | > amp_scaler: 4096.00000 (2449.48911) | > grad_norm: 11.40182 (10.26137) | > current_lr: 0.00023 | > step_time: 0.94780 (0.60919) | > loader_time: 0.00410 (0.00336)  --> STEP: 530/639 -- GLOBAL_STEP: 76050 | > loss: -0.46129 (-0.50774) | > log_mle: -0.58822 (-0.62125) | > loss_dur: 0.12694 (0.11351) | > amp_scaler: 4096.00000 (2527.15472) | > grad_norm: 9.87073 (10.18243) | > current_lr: 0.00023 | > step_time: 0.95430 (0.62304) | > loader_time: 0.00410 (0.00339)  --> STEP: 555/639 -- GLOBAL_STEP: 76075 | > loss: -0.46529 (-0.50608) | > log_mle: -0.57589 (-0.61963) | > loss_dur: 0.11060 (0.11355) | > amp_scaler: 4096.00000 (2597.82342) | > grad_norm: 7.67065 (10.06220) | > current_lr: 0.00023 | > step_time: 0.87180 (0.63756) | > loader_time: 0.00390 (0.00343)  --> STEP: 580/639 -- GLOBAL_STEP: 76100 | > loss: -0.48484 (-0.50465) | > log_mle: -0.59565 (-0.61836) | > loss_dur: 0.11081 (0.11370) | > amp_scaler: 4096.00000 (2662.40000) | > grad_norm: 6.09590 (10.02702) | > current_lr: 0.00023 | > step_time: 0.97590 (0.65262) | > loader_time: 0.00450 (0.00346)  --> STEP: 605/639 -- GLOBAL_STEP: 76125 | > loss: -0.45608 (-0.50324) | > log_mle: -0.56196 (-0.61716) | > loss_dur: 0.10587 (0.11392) | > amp_scaler: 4096.00000 (2721.63967) | > grad_norm: 3.54852 (9.98104) | > current_lr: 0.00023 | > step_time: 1.01420 (0.66811) | > loader_time: 0.00440 (0.00350)  --> STEP: 630/639 -- GLOBAL_STEP: 76150 | > loss: -0.46401 (-0.50203) | > log_mle: -0.59126 (-0.61615) | > loss_dur: 0.12725 (0.11412) | > amp_scaler: 4096.00000 (2776.17778) | > grad_norm: 12.49261 (9.93200) | > current_lr: 0.00023 | > step_time: 1.18750 (0.68442) | > loader_time: 0.00460 (0.00354) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02979 (-0.00033) | > avg_loss: -0.45921 (-0.00087) | > avg_log_mle: -0.59492 (+0.00073) | > avg_loss_dur: 0.13570 (-0.00160)  > EPOCH: 119/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 06:02:13)   --> STEP: 15/639 -- GLOBAL_STEP: 76175 | > loss: -0.66120 (-0.57237) | > log_mle: -0.79459 (-0.68598) | > loss_dur: 0.13339 (0.11362) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 33.97400 (14.26856) | > current_lr: 0.00023 | > step_time: 0.32750 (0.29461) | > loader_time: 0.00280 (0.00244)  --> STEP: 40/639 -- GLOBAL_STEP: 76200 | > loss: -0.53088 (-0.55430) | > log_mle: -0.63682 (-0.66639) | > loss_dur: 0.10594 (0.11209) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.21961 (12.22480) | > current_lr: 0.00023 | > step_time: 0.36070 (0.32941) | > loader_time: 0.00250 (0.00264)  --> STEP: 65/639 -- GLOBAL_STEP: 76225 | > loss: -0.54009 (-0.54547) | > log_mle: -0.62557 (-0.65716) | > loss_dur: 0.08548 (0.11169) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.72043 (11.86837) | > current_lr: 0.00023 | > step_time: 0.38690 (0.36182) | > loader_time: 0.00310 (0.00277)  --> STEP: 90/639 -- GLOBAL_STEP: 76250 | > loss: -0.48389 (-0.53773) | > log_mle: -0.58592 (-0.64786) | > loss_dur: 0.10202 (0.11012) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.35918 (10.84391) | > current_lr: 0.00023 | > step_time: 0.52440 (0.37836) | > loader_time: 0.00280 (0.00283)  --> STEP: 115/639 -- GLOBAL_STEP: 76275 | > loss: -0.45552 (-0.53687) | > log_mle: -0.55636 (-0.64746) | > loss_dur: 0.10084 (0.11059) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.41782 (10.90930) | > current_lr: 0.00023 | > step_time: 0.38240 (0.39424) | > loader_time: 0.00260 (0.00288)  --> STEP: 140/639 -- GLOBAL_STEP: 76300 | > loss: -0.49361 (-0.53579) | > log_mle: -0.63019 (-0.64695) | > loss_dur: 0.13658 (0.11117) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.18045 (11.08081) | > current_lr: 0.00023 | > step_time: 0.63460 (0.40963) | > loader_time: 0.00320 (0.00290)  --> STEP: 165/639 -- GLOBAL_STEP: 76325 | > loss: -0.50391 (-0.53418) | > log_mle: -0.63460 (-0.64521) | > loss_dur: 0.13068 (0.11103) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.82069 (11.27975) | > current_lr: 0.00023 | > step_time: 0.49190 (0.42445) | > loader_time: 0.00310 (0.00293)  --> STEP: 190/639 -- GLOBAL_STEP: 76350 | > loss: -0.45851 (-0.53325) | > log_mle: -0.56855 (-0.64434) | > loss_dur: 0.11004 (0.11109) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.42312 (11.32394) | > current_lr: 0.00023 | > step_time: 0.54030 (0.44209) | > loader_time: 0.00330 (0.00298)  --> STEP: 215/639 -- GLOBAL_STEP: 76375 | > loss: -0.55627 (-0.53044) | > log_mle: -0.66688 (-0.64169) | > loss_dur: 0.11061 (0.11125) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.72859 (11.17995) | > current_lr: 0.00023 | > step_time: 0.54740 (0.45972) | > loader_time: 0.00330 (0.00302)  --> STEP: 240/639 -- GLOBAL_STEP: 76400 | > loss: -0.55903 (-0.53064) | > log_mle: -0.67859 (-0.64195) | > loss_dur: 0.11957 (0.11131) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.42037 (11.22746) | > current_lr: 0.00023 | > step_time: 0.55170 (0.47428) | > loader_time: 0.00330 (0.00306)  --> STEP: 265/639 -- GLOBAL_STEP: 76425 | > loss: -0.57261 (-0.52905) | > log_mle: -0.70816 (-0.64094) | > loss_dur: 0.13555 (0.11188) | > amp_scaler: 2048.00000 (4034.17358) | > grad_norm: 16.25985 (11.14445) | > current_lr: 0.00023 | > step_time: 0.61280 (0.48894) | > loader_time: 0.00350 (0.00310)  --> STEP: 290/639 -- GLOBAL_STEP: 76450 | > loss: -0.49237 (-0.52719) | > log_mle: -0.59727 (-0.63920) | > loss_dur: 0.10490 (0.11201) | > amp_scaler: 2048.00000 (3862.95172) | > grad_norm: 8.01468 (11.01124) | > current_lr: 0.00023 | > step_time: 0.66910 (0.50385) | > loader_time: 0.00340 (0.00314)  --> STEP: 315/639 -- GLOBAL_STEP: 76475 | > loss: -0.47561 (-0.52508) | > log_mle: -0.58018 (-0.63741) | > loss_dur: 0.10457 (0.11233) | > amp_scaler: 2048.00000 (3718.90794) | > grad_norm: 5.01124 (10.92376) | > current_lr: 0.00023 | > step_time: 0.70410 (0.51813) | > loader_time: 0.00340 (0.00317)  --> STEP: 340/639 -- GLOBAL_STEP: 76500 | > loss: -0.49522 (-0.52284) | > log_mle: -0.60144 (-0.63574) | > loss_dur: 0.10622 (0.11290) | > amp_scaler: 2048.00000 (3596.04706) | > grad_norm: 10.35532 (10.93612) | > current_lr: 0.00023 | > step_time: 0.59030 (0.53233) | > loader_time: 0.00350 (0.00321)  --> STEP: 365/639 -- GLOBAL_STEP: 76525 | > loss: -0.49366 (-0.52047) | > log_mle: -0.60798 (-0.63316) | > loss_dur: 0.11432 (0.11269) | > amp_scaler: 2048.00000 (3490.01644) | > grad_norm: 8.98228 (10.75982) | > current_lr: 0.00023 | > step_time: 0.79670 (0.54476) | > loader_time: 0.00370 (0.00323)  --> STEP: 390/639 -- GLOBAL_STEP: 76550 | > loss: -0.48224 (-0.51850) | > log_mle: -0.60666 (-0.63102) | > loss_dur: 0.12441 (0.11252) | > amp_scaler: 2048.00000 (3397.57949) | > grad_norm: 8.21991 (10.54564) | > current_lr: 0.00023 | > step_time: 0.78420 (0.55728) | > loader_time: 0.00350 (0.00326)  --> STEP: 415/639 -- GLOBAL_STEP: 76575 | > loss: -0.45933 (-0.51635) | > log_mle: -0.57957 (-0.62921) | > loss_dur: 0.12024 (0.11286) | > amp_scaler: 2048.00000 (3316.27952) | > grad_norm: 8.66680 (10.48446) | > current_lr: 0.00023 | > step_time: 0.69960 (0.56976) | > loader_time: 0.00340 (0.00328)  --> STEP: 440/639 -- GLOBAL_STEP: 76600 | > loss: -0.44896 (-0.51461) | > log_mle: -0.56295 (-0.62751) | > loss_dur: 0.11399 (0.11289) | > amp_scaler: 2048.00000 (3244.21818) | > grad_norm: 4.46125 (10.43230) | > current_lr: 0.00023 | > step_time: 0.72570 (0.58193) | > loader_time: 0.00330 (0.00331)  --> STEP: 465/639 -- GLOBAL_STEP: 76625 | > loss: -0.45984 (-0.51258) | > log_mle: -0.56800 (-0.62535) | > loss_dur: 0.10816 (0.11278) | > amp_scaler: 2048.00000 (3179.90538) | > grad_norm: 4.61764 (10.25605) | > current_lr: 0.00023 | > step_time: 0.77490 (0.59286) | > loader_time: 0.00370 (0.00333)  --> STEP: 490/639 -- GLOBAL_STEP: 76650 | > loss: -0.45368 (-0.51080) | > log_mle: -0.55175 (-0.62372) | > loss_dur: 0.09806 (0.11292) | > amp_scaler: 2048.00000 (3122.15510) | > grad_norm: 3.20669 (10.13283) | > current_lr: 0.00023 | > step_time: 0.79940 (0.60583) | > loader_time: 0.00380 (0.00336)  --> STEP: 515/639 -- GLOBAL_STEP: 76675 | > loss: -0.47263 (-0.50953) | > log_mle: -0.58028 (-0.62249) | > loss_dur: 0.10765 (0.11296) | > amp_scaler: 2048.00000 (3070.01165) | > grad_norm: 5.73615 (10.10615) | > current_lr: 0.00023 | > step_time: 0.97670 (0.61876) | > loader_time: 0.00420 (0.00339)  --> STEP: 540/639 -- GLOBAL_STEP: 76700 | > loss: -0.47769 (-0.50785) | > log_mle: -0.59550 (-0.62095) | > loss_dur: 0.11782 (0.11310) | > amp_scaler: 2048.00000 (3022.69630) | > grad_norm: 6.74244 (10.00067) | > current_lr: 0.00023 | > step_time: 0.91390 (0.63250) | > loader_time: 0.00420 (0.00342)  --> STEP: 565/639 -- GLOBAL_STEP: 76725 | > loss: -0.44487 (-0.50615) | > log_mle: -0.55785 (-0.61936) | > loss_dur: 0.11298 (0.11321) | > amp_scaler: 2048.00000 (2979.56814) | > grad_norm: 6.36342 (9.90992) | > current_lr: 0.00023 | > step_time: 1.08290 (0.64612) | > loader_time: 0.00410 (0.00345)  --> STEP: 590/639 -- GLOBAL_STEP: 76750 | > loss: -0.47578 (-0.50498) | > log_mle: -0.58807 (-0.61834) | > loss_dur: 0.11229 (0.11336) | > amp_scaler: 2048.00000 (2940.09492) | > grad_norm: 7.73212 (9.86638) | > current_lr: 0.00023 | > step_time: 0.95010 (0.66161) | > loader_time: 0.00420 (0.00348)  --> STEP: 615/639 -- GLOBAL_STEP: 76775 | > loss: -0.47597 (-0.50361) | > log_mle: -0.59208 (-0.61710) | > loss_dur: 0.11611 (0.11349) | > amp_scaler: 2048.00000 (2903.83089) | > grad_norm: 7.28324 (9.80659) | > current_lr: 0.00023 | > step_time: 1.01550 (0.67730) | > loader_time: 0.00400 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03137 (+0.00158) | > avg_loss: -0.45978 (-0.00057) | > avg_log_mle: -0.59430 (+0.00062) | > avg_loss_dur: 0.13452 (-0.00119)  > EPOCH: 120/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 06:09:50)   --> STEP: 0/639 -- GLOBAL_STEP: 76800 | > loss: -0.59703 (-0.59703) | > log_mle: -0.73220 (-0.73220) | > loss_dur: 0.13517 (0.13517) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 21.88863 (21.88863) | > current_lr: 0.00023 | > step_time: 0.63420 (0.63424) | > loader_time: 0.41890 (0.41886)  --> STEP: 25/639 -- GLOBAL_STEP: 76825 | > loss: -0.55934 (-0.56616) | > log_mle: -0.67871 (-0.68121) | > loss_dur: 0.11937 (0.11505) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.51625 (13.16046) | > current_lr: 0.00023 | > step_time: 0.40470 (0.32100) | > loader_time: 0.00260 (0.00247)  --> STEP: 50/639 -- GLOBAL_STEP: 76850 | > loss: -0.52287 (-0.55372) | > log_mle: -0.61627 (-0.66636) | > loss_dur: 0.09340 (0.11265) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.43045 (11.78261) | > current_lr: 0.00023 | > step_time: 0.32860 (0.34476) | > loader_time: 0.00260 (0.00259)  --> STEP: 75/639 -- GLOBAL_STEP: 76875 | > loss: -0.53431 (-0.54429) | > log_mle: -0.63287 (-0.65595) | > loss_dur: 0.09855 (0.11166) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.32494 (10.95900) | > current_lr: 0.00023 | > step_time: 0.41380 (0.36473) | > loader_time: 0.00310 (0.00269)  --> STEP: 100/639 -- GLOBAL_STEP: 76900 | > loss: -0.53242 (-0.54114) | > log_mle: -0.64430 (-0.65155) | > loss_dur: 0.11187 (0.11042) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.80989 (10.68658) | > current_lr: 0.00023 | > step_time: 0.50090 (0.38167) | > loader_time: 0.00300 (0.00276)  --> STEP: 125/639 -- GLOBAL_STEP: 76925 | > loss: -0.52480 (-0.53789) | > log_mle: -0.63582 (-0.64870) | > loss_dur: 0.11102 (0.11081) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.83353 (10.92059) | > current_lr: 0.00023 | > step_time: 0.42860 (0.39723) | > loader_time: 0.00290 (0.00282)  --> STEP: 150/639 -- GLOBAL_STEP: 76950 | > loss: -0.48821 (-0.53553) | > log_mle: -0.59640 (-0.64681) | > loss_dur: 0.10819 (0.11128) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.30805 (11.13843) | > current_lr: 0.00023 | > step_time: 0.52020 (0.41389) | > loader_time: 0.00310 (0.00286)  --> STEP: 175/639 -- GLOBAL_STEP: 76975 | > loss: -0.51690 (-0.53480) | > log_mle: -0.61725 (-0.64642) | > loss_dur: 0.10034 (0.11162) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.01587 (11.34763) | > current_lr: 0.00023 | > step_time: 0.55570 (0.43142) | > loader_time: 0.00330 (0.00289)  --> STEP: 200/639 -- GLOBAL_STEP: 77000 | > loss: -0.48826 (-0.53225) | > log_mle: -0.59991 (-0.64350) | > loss_dur: 0.11165 (0.11124) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.72339 (11.15919) | > current_lr: 0.00023 | > step_time: 0.63230 (0.44606) | > loader_time: 0.00310 (0.00294)  --> STEP: 225/639 -- GLOBAL_STEP: 77025 | > loss: -0.49337 (-0.52979) | > log_mle: -0.58462 (-0.64089) | > loss_dur: 0.09125 (0.11111) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.83200 (10.99205) | > current_lr: 0.00023 | > step_time: 0.50560 (0.46249) | > loader_time: 0.00300 (0.00298)  --> STEP: 250/639 -- GLOBAL_STEP: 77050 | > loss: -0.49797 (-0.53000) | > log_mle: -0.58919 (-0.64132) | > loss_dur: 0.09122 (0.11132) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.88826 (11.27721) | > current_lr: 0.00023 | > step_time: 0.60030 (0.47730) | > loader_time: 0.00330 (0.00302)  --> STEP: 275/639 -- GLOBAL_STEP: 77075 | > loss: -0.50875 (-0.52937) | > log_mle: -0.62402 (-0.64135) | > loss_dur: 0.11527 (0.11198) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.73009 (11.56981) | > current_lr: 0.00023 | > step_time: 0.87140 (0.49449) | > loader_time: 0.00390 (0.00305)  --> STEP: 300/639 -- GLOBAL_STEP: 77100 | > loss: -0.50619 (-0.52679) | > log_mle: -0.61793 (-0.63902) | > loss_dur: 0.11173 (0.11223) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.76275 (11.52244) | > current_lr: 0.00023 | > step_time: 0.73390 (0.50852) | > loader_time: 0.00340 (0.00308)  --> STEP: 325/639 -- GLOBAL_STEP: 77125 | > loss: -0.49882 (-0.52437) | > log_mle: -0.61495 (-0.63712) | > loss_dur: 0.11614 (0.11275) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.25503 (11.54307) | > current_lr: 0.00023 | > step_time: 0.58720 (0.52348) | > loader_time: 0.00360 (0.00312)  --> STEP: 350/639 -- GLOBAL_STEP: 77150 | > loss: -0.51898 (-0.52216) | > log_mle: -0.64251 (-0.63495) | > loss_dur: 0.12353 (0.11279) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.70529 (11.40790) | > current_lr: 0.00023 | > step_time: 0.79690 (0.53770) | > loader_time: 0.00410 (0.00316)  --> STEP: 375/639 -- GLOBAL_STEP: 77175 | > loss: -0.46377 (-0.51975) | > log_mle: -0.57623 (-0.63237) | > loss_dur: 0.11247 (0.11262) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.53141 (11.18909) | > current_lr: 0.00023 | > step_time: 0.67730 (0.54957) | > loader_time: 0.00350 (0.00318)  --> STEP: 400/639 -- GLOBAL_STEP: 77200 | > loss: -0.49496 (-0.51785) | > log_mle: -0.60745 (-0.63066) | > loss_dur: 0.11248 (0.11280) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.52556 (11.05592) | > current_lr: 0.00023 | > step_time: 0.71600 (0.56056) | > loader_time: 0.00350 (0.00321)  --> STEP: 425/639 -- GLOBAL_STEP: 77225 | > loss: -0.54194 (-0.51552) | > log_mle: -0.66451 (-0.62843) | > loss_dur: 0.12257 (0.11291) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 19.40227 (10.87959) | > current_lr: 0.00023 | > step_time: 0.94340 (0.57257) | > loader_time: 0.00360 (0.00324)  --> STEP: 450/639 -- GLOBAL_STEP: 77250 | > loss: -0.46309 (-0.51372) | > log_mle: -0.56646 (-0.62650) | > loss_dur: 0.10337 (0.11278) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.76796 (10.76445) | > current_lr: 0.00023 | > step_time: 0.80230 (0.58452) | > loader_time: 0.00390 (0.00326)  --> STEP: 475/639 -- GLOBAL_STEP: 77275 | > loss: -0.47695 (-0.51173) | > log_mle: -0.57827 (-0.62466) | > loss_dur: 0.10132 (0.11294) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.50744 (10.62344) | > current_lr: 0.00023 | > step_time: 0.71480 (0.59601) | > loader_time: 0.00370 (0.00328)  --> STEP: 500/639 -- GLOBAL_STEP: 77300 | > loss: -0.51672 (-0.51019) | > log_mle: -0.63550 (-0.62325) | > loss_dur: 0.11878 (0.11306) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.81010 (10.55300) | > current_lr: 0.00023 | > step_time: 0.94170 (0.60863) | > loader_time: 0.00400 (0.00331)  --> STEP: 525/639 -- GLOBAL_STEP: 77325 | > loss: -0.48662 (-0.50859) | > log_mle: -0.61504 (-0.62185) | > loss_dur: 0.12843 (0.11326) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.30869 (10.44845) | > current_lr: 0.00023 | > step_time: 1.03490 (0.62208) | > loader_time: 0.00400 (0.00334)  --> STEP: 550/639 -- GLOBAL_STEP: 77350 | > loss: -0.43765 (-0.50693) | > log_mle: -0.55145 (-0.62024) | > loss_dur: 0.11380 (0.11332) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.34219 (10.29026) | > current_lr: 0.00023 | > step_time: 0.94310 (0.63623) | > loader_time: 0.00400 (0.00337)  --> STEP: 575/639 -- GLOBAL_STEP: 77375 | > loss: -0.46131 (-0.50526) | > log_mle: -0.57069 (-0.61873) | > loss_dur: 0.10938 (0.11347) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.71488 (10.17499) | > current_lr: 0.00023 | > step_time: 0.87050 (0.65003) | > loader_time: 0.00410 (0.00340)  --> STEP: 600/639 -- GLOBAL_STEP: 77400 | > loss: -0.45165 (-0.50397) | > log_mle: -0.56752 (-0.61758) | > loss_dur: 0.11588 (0.11362) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.44995 (10.09604) | > current_lr: 0.00023 | > step_time: 1.03510 (0.66630) | > loader_time: 0.00430 (0.00343)  --> STEP: 625/639 -- GLOBAL_STEP: 77425 | > loss: -0.47556 (-0.50262) | > log_mle: -0.59334 (-0.61641) | > loss_dur: 0.11778 (0.11379) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.28454 (10.02899) | > current_lr: 0.00023 | > step_time: 1.08510 (0.68268) | > loader_time: 0.00440 (0.00347) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02867 (-0.00270) | > avg_loss: -0.45747 (+0.00231) | > avg_log_mle: -0.59380 (+0.00050) | > avg_loss_dur: 0.13633 (+0.00181)  > EPOCH: 121/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 06:17:27)   --> STEP: 10/639 -- GLOBAL_STEP: 77450 | > loss: -0.55783 (-0.56583) | > log_mle: -0.67937 (-0.68189) | > loss_dur: 0.12155 (0.11607) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.71095 (13.08200) | > current_lr: 0.00023 | > step_time: 0.28430 (0.29165) | > loader_time: 0.00240 (0.00233)  --> STEP: 35/639 -- GLOBAL_STEP: 77475 | > loss: -0.55402 (-0.55757) | > log_mle: -0.64355 (-0.66873) | > loss_dur: 0.08952 (0.11116) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.14276 (11.82733) | > current_lr: 0.00023 | > step_time: 0.36970 (0.33010) | > loader_time: 0.00270 (0.00263)  --> STEP: 60/639 -- GLOBAL_STEP: 77500 | > loss: -0.50906 (-0.54875) | > log_mle: -0.60470 (-0.65982) | > loss_dur: 0.09564 (0.11107) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.90494 (11.21055) | > current_lr: 0.00023 | > step_time: 0.34260 (0.36074) | > loader_time: 0.00280 (0.00273)  --> STEP: 85/639 -- GLOBAL_STEP: 77525 | > loss: -0.50975 (-0.54090) | > log_mle: -0.60624 (-0.65029) | > loss_dur: 0.09649 (0.10939) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.59708 (10.50561) | > current_lr: 0.00023 | > step_time: 0.41530 (0.37653) | > loader_time: 0.00300 (0.00280)  --> STEP: 110/639 -- GLOBAL_STEP: 77550 | > loss: -0.50372 (-0.54035) | > log_mle: -0.61472 (-0.65098) | > loss_dur: 0.11100 (0.11063) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.41442 (10.78627) | > current_lr: 0.00023 | > step_time: 0.39860 (0.39414) | > loader_time: 0.00300 (0.00285)  --> STEP: 135/639 -- GLOBAL_STEP: 77575 | > loss: -0.47862 (-0.53673) | > log_mle: -0.59275 (-0.64756) | > loss_dur: 0.11413 (0.11083) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.14516 (10.70671) | > current_lr: 0.00023 | > step_time: 0.57230 (0.40639) | > loader_time: 0.00300 (0.00287)  --> STEP: 160/639 -- GLOBAL_STEP: 77600 | > loss: -0.52264 (-0.53407) | > log_mle: -0.63150 (-0.64480) | > loss_dur: 0.10886 (0.11073) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.02310 (10.59623) | > current_lr: 0.00023 | > step_time: 0.46240 (0.42306) | > loader_time: 0.00300 (0.00291)  --> STEP: 185/639 -- GLOBAL_STEP: 77625 | > loss: -0.50974 (-0.53438) | > log_mle: -0.61424 (-0.64606) | > loss_dur: 0.10450 (0.11168) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.13355 (10.98251) | > current_lr: 0.00023 | > step_time: 0.58940 (0.43929) | > loader_time: 0.00320 (0.00293)  --> STEP: 210/639 -- GLOBAL_STEP: 77650 | > loss: -0.44552 (-0.53095) | > log_mle: -0.53697 (-0.64246) | > loss_dur: 0.09146 (0.11152) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.23446 (10.66550) | > current_lr: 0.00023 | > step_time: 0.51310 (0.45488) | > loader_time: 0.00300 (0.00296)  --> STEP: 235/639 -- GLOBAL_STEP: 77675 | > loss: -0.51260 (-0.53097) | > log_mle: -0.62406 (-0.64229) | > loss_dur: 0.11146 (0.11132) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.35504 (10.84961) | > current_lr: 0.00023 | > step_time: 0.55790 (0.47025) | > loader_time: 0.00320 (0.00301)  --> STEP: 260/639 -- GLOBAL_STEP: 77700 | > loss: -0.51105 (-0.52980) | > log_mle: -0.62606 (-0.64142) | > loss_dur: 0.11501 (0.11163) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.34730 (10.86404) | > current_lr: 0.00023 | > step_time: 0.59040 (0.48393) | > loader_time: 0.00290 (0.00304)  --> STEP: 285/639 -- GLOBAL_STEP: 77725 | > loss: -0.46731 (-0.52806) | > log_mle: -0.57366 (-0.64007) | > loss_dur: 0.10635 (0.11200) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.71640 (10.84959) | > current_lr: 0.00023 | > step_time: 0.62130 (0.50002) | > loader_time: 0.00340 (0.00308)  --> STEP: 310/639 -- GLOBAL_STEP: 77750 | > loss: -0.49960 (-0.52590) | > log_mle: -0.61337 (-0.63822) | > loss_dur: 0.11377 (0.11232) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.32348 (10.65901) | > current_lr: 0.00023 | > step_time: 0.58470 (0.51424) | > loader_time: 0.00330 (0.00311)  --> STEP: 335/639 -- GLOBAL_STEP: 77775 | > loss: -0.51700 (-0.52320) | > log_mle: -0.64043 (-0.63595) | > loss_dur: 0.12343 (0.11275) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.65458 (10.53584) | > current_lr: 0.00023 | > step_time: 0.71520 (0.52900) | > loader_time: 0.00360 (0.00314)  --> STEP: 360/639 -- GLOBAL_STEP: 77800 | > loss: -0.46421 (-0.52126) | > log_mle: -0.57626 (-0.63385) | > loss_dur: 0.11205 (0.11259) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.72366 (10.42033) | > current_lr: 0.00023 | > step_time: 0.66740 (0.54099) | > loader_time: 0.00360 (0.00317)  --> STEP: 385/639 -- GLOBAL_STEP: 77825 | > loss: -0.49896 (-0.51946) | > log_mle: -0.60881 (-0.63194) | > loss_dur: 0.10985 (0.11247) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.31325 (10.24988) | > current_lr: 0.00023 | > step_time: 0.69420 (0.55345) | > loader_time: 0.00360 (0.00320)  --> STEP: 410/639 -- GLOBAL_STEP: 77850 | > loss: -0.45757 (-0.51738) | > log_mle: -0.58427 (-0.63010) | > loss_dur: 0.12670 (0.11272) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.20518 (10.16345) | > current_lr: 0.00023 | > step_time: 0.74340 (0.56399) | > loader_time: 0.00390 (0.00323)  --> STEP: 435/639 -- GLOBAL_STEP: 77875 | > loss: -0.48825 (-0.51518) | > log_mle: -0.59365 (-0.62793) | > loss_dur: 0.10540 (0.11275) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.92331 (10.00263) | > current_lr: 0.00023 | > step_time: 0.74480 (0.57535) | > loader_time: 0.00350 (0.00326)  --> STEP: 460/639 -- GLOBAL_STEP: 77900 | > loss: -0.45044 (-0.51307) | > log_mle: -0.56116 (-0.62580) | > loss_dur: 0.11073 (0.11273) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.06295 (9.84681) | > current_lr: 0.00023 | > step_time: 0.91630 (0.58706) | > loader_time: 0.00390 (0.00328)  --> STEP: 485/639 -- GLOBAL_STEP: 77925 | > loss: -0.48027 (-0.51131) | > log_mle: -0.58920 (-0.62426) | > loss_dur: 0.10893 (0.11295) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.08407 (9.76347) | > current_lr: 0.00023 | > step_time: 0.81210 (0.60017) | > loader_time: 0.00390 (0.00331)  --> STEP: 510/639 -- GLOBAL_STEP: 77950 | > loss: -0.47475 (-0.50993) | > log_mle: -0.59308 (-0.62289) | > loss_dur: 0.11834 (0.11296) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.03696 (9.66975) | > current_lr: 0.00023 | > step_time: 0.84870 (0.61308) | > loader_time: 0.00400 (0.00334)  --> STEP: 535/639 -- GLOBAL_STEP: 77975 | > loss: -0.45855 (-0.50830) | > log_mle: -0.58454 (-0.62138) | > loss_dur: 0.12599 (0.11308) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.48939 (9.57789) | > current_lr: 0.00023 | > step_time: 0.89900 (0.62666) | > loader_time: 0.00360 (0.00337)  --> STEP: 560/639 -- GLOBAL_STEP: 78000 | > loss: -0.47105 (-0.50672) | > log_mle: -0.59362 (-0.61997) | > loss_dur: 0.12257 (0.11325) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.12609 (9.48821) | > current_lr: 0.00023 | > step_time: 0.92120 (0.64146) | > loader_time: 0.00420 (0.00340)  --> STEP: 585/639 -- GLOBAL_STEP: 78025 | > loss: -0.48330 (-0.50532) | > log_mle: -0.59544 (-0.61868) | > loss_dur: 0.11214 (0.11335) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.44071 (9.40748) | > current_lr: 0.00023 | > step_time: 1.01270 (0.65648) | > loader_time: 0.00440 (0.00344)  --> STEP: 610/639 -- GLOBAL_STEP: 78050 | > loss: -0.45787 (-0.50392) | > log_mle: -0.56562 (-0.61745) | > loss_dur: 0.10775 (0.11353) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.71177 (9.34904) | > current_lr: 0.00023 | > step_time: 1.00660 (0.67234) | > loader_time: 0.00440 (0.00348)  --> STEP: 635/639 -- GLOBAL_STEP: 78075 | > loss: -0.46827 (-0.50275) | > log_mle: -0.58953 (-0.61659) | > loss_dur: 0.12125 (0.11385) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.28208 (9.34764) | > current_lr: 0.00023 | > step_time: 1.12510 (0.68954) | > loader_time: 0.00430 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03269 (+0.00402) | > avg_loss: -0.45312 (+0.00436) | > avg_log_mle: -0.59330 (+0.00050) | > avg_loss_dur: 0.14018 (+0.00386)  > EPOCH: 122/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 06:25:03)   --> STEP: 20/639 -- GLOBAL_STEP: 78100 | > loss: -0.65122 (-0.57340) | > log_mle: -0.80380 (-0.69075) | > loss_dur: 0.15257 (0.11735) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 22.14932 (12.19370) | > current_lr: 0.00023 | > step_time: 0.36770 (0.31188) | > loader_time: 0.00300 (0.00423)  --> STEP: 45/639 -- GLOBAL_STEP: 78125 | > loss: -0.52158 (-0.55612) | > log_mle: -0.61094 (-0.66818) | > loss_dur: 0.08936 (0.11205) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.78138 (11.34928) | > current_lr: 0.00023 | > step_time: 0.40690 (0.34098) | > loader_time: 0.00280 (0.00346)  --> STEP: 70/639 -- GLOBAL_STEP: 78150 | > loss: -0.53774 (-0.54591) | > log_mle: -0.66083 (-0.65754) | > loss_dur: 0.12309 (0.11164) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.37823 (10.71141) | > current_lr: 0.00023 | > step_time: 0.39980 (0.36688) | > loader_time: 0.00300 (0.00327)  --> STEP: 95/639 -- GLOBAL_STEP: 78175 | > loss: -0.60277 (-0.54205) | > log_mle: -0.70545 (-0.65175) | > loss_dur: 0.10268 (0.10970) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 19.19251 (10.39841) | > current_lr: 0.00023 | > step_time: 0.40780 (0.38300) | > loader_time: 0.00280 (0.00319)  --> STEP: 120/639 -- GLOBAL_STEP: 78200 | > loss: -0.57021 (-0.53876) | > log_mle: -0.70334 (-0.64927) | > loss_dur: 0.13313 (0.11051) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 17.02921 (10.82788) | > current_lr: 0.00023 | > step_time: 0.54830 (0.39934) | > loader_time: 0.00330 (0.00316)  --> STEP: 145/639 -- GLOBAL_STEP: 78225 | > loss: -0.52414 (-0.53726) | > log_mle: -0.65229 (-0.64857) | > loss_dur: 0.12816 (0.11131) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.68500 (10.86706) | > current_lr: 0.00023 | > step_time: 0.46970 (0.41353) | > loader_time: 0.00320 (0.00316)  --> STEP: 170/639 -- GLOBAL_STEP: 78250 | > loss: -0.52342 (-0.53533) | > log_mle: -0.63251 (-0.64673) | > loss_dur: 0.10909 (0.11139) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.13659 (10.89876) | > current_lr: 0.00023 | > step_time: 0.54750 (0.42736) | > loader_time: 0.00340 (0.00317)  --> STEP: 195/639 -- GLOBAL_STEP: 78275 | > loss: -0.48729 (-0.53384) | > log_mle: -0.58180 (-0.64500) | > loss_dur: 0.09451 (0.11116) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.50102 (10.93605) | > current_lr: 0.00023 | > step_time: 0.49630 (0.44349) | > loader_time: 0.00280 (0.00318)  --> STEP: 220/639 -- GLOBAL_STEP: 78300 | > loss: -0.51946 (-0.53093) | > log_mle: -0.62915 (-0.64219) | > loss_dur: 0.10970 (0.11126) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.03245 (10.77363) | > current_lr: 0.00023 | > step_time: 0.63010 (0.45874) | > loader_time: 0.00340 (0.00319)  --> STEP: 245/639 -- GLOBAL_STEP: 78325 | > loss: -0.51727 (-0.53131) | > log_mle: -0.61850 (-0.64263) | > loss_dur: 0.10123 (0.11132) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.38136 (10.96393) | > current_lr: 0.00023 | > step_time: 0.52790 (0.47165) | > loader_time: 0.00310 (0.00321)  --> STEP: 270/639 -- GLOBAL_STEP: 78350 | > loss: -0.56205 (-0.53038) | > log_mle: -0.68412 (-0.64223) | > loss_dur: 0.12207 (0.11185) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 23.85618 (11.17041) | > current_lr: 0.00023 | > step_time: 0.83080 (0.48812) | > loader_time: 0.00400 (0.00325)  --> STEP: 295/639 -- GLOBAL_STEP: 78375 | > loss: -0.48038 (-0.52772) | > log_mle: -0.57747 (-0.63991) | > loss_dur: 0.09709 (0.11219) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.35759 (10.96453) | > current_lr: 0.00023 | > step_time: 0.60650 (0.50230) | > loader_time: 0.00350 (0.00327)  --> STEP: 320/639 -- GLOBAL_STEP: 78400 | > loss: -0.48617 (-0.52560) | > log_mle: -0.61715 (-0.63825) | > loss_dur: 0.13098 (0.11265) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.33629 (10.92456) | > current_lr: 0.00023 | > step_time: 0.75130 (0.51654) | > loader_time: 0.00370 (0.00329)  --> STEP: 345/639 -- GLOBAL_STEP: 78425 | > loss: -0.49601 (-0.52330) | > log_mle: -0.60490 (-0.63619) | > loss_dur: 0.10889 (0.11290) | > amp_scaler: 4096.00000 (2095.48986) | > grad_norm: 8.34852 (10.86653) | > current_lr: 0.00023 | > step_time: 0.64880 (0.53095) | > loader_time: 0.00330 (0.00331)  --> STEP: 370/639 -- GLOBAL_STEP: 78450 | > loss: -0.49802 (-0.52112) | > log_mle: -0.61560 (-0.63383) | > loss_dur: 0.11758 (0.11271) | > amp_scaler: 4096.00000 (2230.65946) | > grad_norm: 11.82845 (10.68414) | > current_lr: 0.00023 | > step_time: 0.87910 (0.54299) | > loader_time: 0.00400 (0.00332)  --> STEP: 395/639 -- GLOBAL_STEP: 78475 | > loss: -0.47422 (-0.51898) | > log_mle: -0.58992 (-0.63167) | > loss_dur: 0.11570 (0.11270) | > amp_scaler: 4096.00000 (2348.71899) | > grad_norm: 7.01228 (10.54127) | > current_lr: 0.00023 | > step_time: 0.67480 (0.55556) | > loader_time: 0.00370 (0.00334)  --> STEP: 420/639 -- GLOBAL_STEP: 78500 | > loss: -0.47598 (-0.51663) | > log_mle: -0.58519 (-0.62964) | > loss_dur: 0.10921 (0.11301) | > amp_scaler: 4096.00000 (2452.72381) | > grad_norm: 6.66440 (10.42185) | > current_lr: 0.00023 | > step_time: 0.73260 (0.56890) | > loader_time: 0.00320 (0.00336)  --> STEP: 445/639 -- GLOBAL_STEP: 78525 | > loss: -0.48650 (-0.51499) | > log_mle: -0.59187 (-0.62799) | > loss_dur: 0.10536 (0.11300) | > amp_scaler: 4096.00000 (2545.04270) | > grad_norm: 7.00076 (10.31679) | > current_lr: 0.00023 | > step_time: 0.71980 (0.58067) | > loader_time: 0.00370 (0.00338)  --> STEP: 470/639 -- GLOBAL_STEP: 78550 | > loss: -0.48720 (-0.51294) | > log_mle: -0.60016 (-0.62595) | > loss_dur: 0.11296 (0.11301) | > amp_scaler: 4096.00000 (2627.54043) | > grad_norm: 13.07783 (10.18040) | > current_lr: 0.00023 | > step_time: 0.85700 (0.59214) | > loader_time: 0.00370 (0.00340)  --> STEP: 495/639 -- GLOBAL_STEP: 78575 | > loss: -0.43668 (-0.51114) | > log_mle: -0.54098 (-0.62436) | > loss_dur: 0.10430 (0.11322) | > amp_scaler: 4096.00000 (2701.70505) | > grad_norm: 2.05070 (10.12620) | > current_lr: 0.00023 | > step_time: 0.73480 (0.60495) | > loader_time: 0.00350 (0.00342)  --> STEP: 520/639 -- GLOBAL_STEP: 78600 | > loss: -0.48500 (-0.50982) | > log_mle: -0.59371 (-0.62305) | > loss_dur: 0.10870 (0.11324) | > amp_scaler: 4096.00000 (2768.73846) | > grad_norm: 11.90203 (10.06569) | > current_lr: 0.00023 | > step_time: 0.81390 (0.61777) | > loader_time: 0.00380 (0.00344)  --> STEP: 545/639 -- GLOBAL_STEP: 78625 | > loss: -0.49909 (-0.50827) | > log_mle: -0.62632 (-0.62165) | > loss_dur: 0.12723 (0.11338) | > amp_scaler: 4096.00000 (2829.62202) | > grad_norm: 14.53997 (9.99295) | > current_lr: 0.00023 | > step_time: 0.89790 (0.63292) | > loader_time: 0.00390 (0.00347)  --> STEP: 570/639 -- GLOBAL_STEP: 78650 | > loss: -0.46977 (-0.50654) | > log_mle: -0.57409 (-0.62001) | > loss_dur: 0.10432 (0.11347) | > amp_scaler: 4096.00000 (2885.16491) | > grad_norm: 6.82110 (9.87694) | > current_lr: 0.00023 | > step_time: 0.87460 (0.64679) | > loader_time: 0.00390 (0.00350)  --> STEP: 595/639 -- GLOBAL_STEP: 78675 | > loss: -0.45879 (-0.50528) | > log_mle: -0.56557 (-0.61890) | > loss_dur: 0.10678 (0.11362) | > amp_scaler: 4096.00000 (2936.04034) | > grad_norm: 7.63306 (9.83661) | > current_lr: 0.00023 | > step_time: 0.98150 (0.66263) | > loader_time: 0.00430 (0.00353)  --> STEP: 620/639 -- GLOBAL_STEP: 78700 | > loss: -0.46269 (-0.50385) | > log_mle: -0.57452 (-0.61757) | > loss_dur: 0.11184 (0.11373) | > amp_scaler: 4096.00000 (2982.81290) | > grad_norm: 6.38187 (9.77952) | > current_lr: 0.00023 | > step_time: 1.13870 (0.67918) | > loader_time: 0.00470 (0.00356) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03285 (+0.00016) | > avg_loss: -0.45666 (-0.00354) | > avg_log_mle: -0.59402 (-0.00072) | > avg_loss_dur: 0.13737 (-0.00282)  > EPOCH: 123/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 06:32:40)   --> STEP: 5/639 -- GLOBAL_STEP: 78725 | > loss: -0.57491 (-0.57537) | > log_mle: -0.71207 (-0.69431) | > loss_dur: 0.13716 (0.11894) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 26.79313 (15.86848) | > current_lr: 0.00023 | > step_time: 0.30650 (0.27672) | > loader_time: 0.00240 (0.00252)  --> STEP: 30/639 -- GLOBAL_STEP: 78750 | > loss: -0.56406 (-0.56205) | > log_mle: -0.66990 (-0.67558) | > loss_dur: 0.10585 (0.11353) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.82733 (13.12919) | > current_lr: 0.00023 | > step_time: 0.32370 (0.32583) | > loader_time: 0.00290 (0.00290)  --> STEP: 55/639 -- GLOBAL_STEP: 78775 | > loss: -0.51570 (-0.55291) | > log_mle: -0.60686 (-0.66415) | > loss_dur: 0.09116 (0.11124) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.25698 (11.79936) | > current_lr: 0.00023 | > step_time: 0.33090 (0.34578) | > loader_time: 0.00280 (0.00285)  --> STEP: 80/639 -- GLOBAL_STEP: 78800 | > loss: -0.51386 (-0.54479) | > log_mle: -0.63816 (-0.65453) | > loss_dur: 0.12430 (0.10974) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.40921 (11.37313) | > current_lr: 0.00023 | > step_time: 0.39940 (0.36811) | > loader_time: 0.00290 (0.00288)  --> STEP: 105/639 -- GLOBAL_STEP: 78825 | > loss: -0.48510 (-0.54201) | > log_mle: -0.59356 (-0.65150) | > loss_dur: 0.10847 (0.10949) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.36064 (11.22439) | > current_lr: 0.00023 | > step_time: 0.44530 (0.38549) | > loader_time: 0.00320 (0.00291)  --> STEP: 130/639 -- GLOBAL_STEP: 78850 | > loss: -0.51273 (-0.53827) | > log_mle: -0.63068 (-0.64866) | > loss_dur: 0.11796 (0.11039) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.43574 (11.38337) | > current_lr: 0.00023 | > step_time: 0.48780 (0.40056) | > loader_time: 0.00300 (0.00292)  --> STEP: 155/639 -- GLOBAL_STEP: 78875 | > loss: -0.49696 (-0.53623) | > log_mle: -0.61617 (-0.64701) | > loss_dur: 0.11921 (0.11078) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.27590 (11.27073) | > current_lr: 0.00023 | > step_time: 0.52920 (0.41801) | > loader_time: 0.00300 (0.00295)  --> STEP: 180/639 -- GLOBAL_STEP: 78900 | > loss: -0.54446 (-0.53671) | > log_mle: -0.66701 (-0.64788) | > loss_dur: 0.12255 (0.11117) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.11731 (11.44307) | > current_lr: 0.00023 | > step_time: 0.57010 (0.43346) | > loader_time: 0.00330 (0.00297)  --> STEP: 205/639 -- GLOBAL_STEP: 78925 | > loss: -0.48540 (-0.53303) | > log_mle: -0.59579 (-0.64375) | > loss_dur: 0.11039 (0.11072) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.40827 (11.05229) | > current_lr: 0.00023 | > step_time: 0.55200 (0.44735) | > loader_time: 0.00330 (0.00301)  --> STEP: 230/639 -- GLOBAL_STEP: 78950 | > loss: -0.50624 (-0.53231) | > log_mle: -0.62700 (-0.64316) | > loss_dur: 0.12076 (0.11085) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.09256 (11.08514) | > current_lr: 0.00023 | > step_time: 0.63270 (0.46363) | > loader_time: 0.00350 (0.00304)  --> STEP: 255/639 -- GLOBAL_STEP: 78975 | > loss: -0.47527 (-0.53089) | > log_mle: -0.59719 (-0.64165) | > loss_dur: 0.12192 (0.11075) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.19617 (11.06926) | > current_lr: 0.00023 | > step_time: 0.62670 (0.47675) | > loader_time: 0.00350 (0.00306)  --> STEP: 280/639 -- GLOBAL_STEP: 79000 | > loss: -0.49737 (-0.53003) | > log_mle: -0.62047 (-0.64152) | > loss_dur: 0.12310 (0.11149) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.04371 (11.17473) | > current_lr: 0.00023 | > step_time: 0.59610 (0.49303) | > loader_time: 0.00360 (0.00309)  --> STEP: 305/639 -- GLOBAL_STEP: 79025 | > loss: -0.48767 (-0.52774) | > log_mle: -0.60546 (-0.63948) | > loss_dur: 0.11778 (0.11174) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.54858 (11.14803) | > current_lr: 0.00023 | > step_time: 0.67100 (0.50697) | > loader_time: 0.00340 (0.00312)  --> STEP: 330/639 -- GLOBAL_STEP: 79050 | > loss: -0.49830 (-0.52486) | > log_mle: -0.62529 (-0.63714) | > loss_dur: 0.12699 (0.11228) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.62342 (11.00581) | > current_lr: 0.00023 | > step_time: 0.79640 (0.52259) | > loader_time: 0.00350 (0.00314)  --> STEP: 355/639 -- GLOBAL_STEP: 79075 | > loss: -0.45469 (-0.52282) | > log_mle: -0.55735 (-0.63501) | > loss_dur: 0.10266 (0.11219) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.41063 (10.88330) | > current_lr: 0.00022 | > step_time: 0.70280 (0.53529) | > loader_time: 0.00290 (0.00317)  --> STEP: 380/639 -- GLOBAL_STEP: 79100 | > loss: -0.49581 (-0.52058) | > log_mle: -0.59282 (-0.63261) | > loss_dur: 0.09701 (0.11202) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.17774 (10.67553) | > current_lr: 0.00022 | > step_time: 0.63230 (0.54859) | > loader_time: 0.00330 (0.00320)  --> STEP: 405/639 -- GLOBAL_STEP: 79125 | > loss: -0.48337 (-0.51879) | > log_mle: -0.59179 (-0.63114) | > loss_dur: 0.10842 (0.11236) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.84389 (10.58627) | > current_lr: 0.00022 | > step_time: 0.68280 (0.55985) | > loader_time: 0.00360 (0.00322)  --> STEP: 430/639 -- GLOBAL_STEP: 79150 | > loss: -0.45221 (-0.51646) | > log_mle: -0.55684 (-0.62904) | > loss_dur: 0.10463 (0.11258) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.03926 (10.45543) | > current_lr: 0.00022 | > step_time: 0.75090 (0.57211) | > loader_time: 0.00320 (0.00325)  --> STEP: 455/639 -- GLOBAL_STEP: 79175 | > loss: -0.47299 (-0.51440) | > log_mle: -0.58078 (-0.62691) | > loss_dur: 0.10778 (0.11251) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.21795 (10.29711) | > current_lr: 0.00022 | > step_time: 0.78030 (0.58285) | > loader_time: 0.00350 (0.00327)  --> STEP: 480/639 -- GLOBAL_STEP: 79200 | > loss: -0.43091 (-0.51246) | > log_mle: -0.53737 (-0.62518) | > loss_dur: 0.10646 (0.11273) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.20848 (10.18291) | > current_lr: 0.00022 | > step_time: 0.80770 (0.59537) | > loader_time: 0.00370 (0.00330)  --> STEP: 505/639 -- GLOBAL_STEP: 79225 | > loss: -0.49234 (-0.51086) | > log_mle: -0.59901 (-0.62359) | > loss_dur: 0.10667 (0.11273) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.07185 (10.06300) | > current_lr: 0.00022 | > step_time: 0.99080 (0.60748) | > loader_time: 0.00410 (0.00332)  --> STEP: 530/639 -- GLOBAL_STEP: 79250 | > loss: -0.46378 (-0.50925) | > log_mle: -0.59067 (-0.62219) | > loss_dur: 0.12688 (0.11294) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.69603 (9.96501) | > current_lr: 0.00022 | > step_time: 0.94260 (0.62068) | > loader_time: 0.00370 (0.00335)  --> STEP: 555/639 -- GLOBAL_STEP: 79275 | > loss: -0.46419 (-0.50760) | > log_mle: -0.57546 (-0.62055) | > loss_dur: 0.11127 (0.11294) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.84243 (9.88332) | > current_lr: 0.00022 | > step_time: 0.89920 (0.63565) | > loader_time: 0.00380 (0.00338)  --> STEP: 580/639 -- GLOBAL_STEP: 79300 | > loss: -0.48587 (-0.50617) | > log_mle: -0.59461 (-0.61925) | > loss_dur: 0.10874 (0.11308) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.13423 (9.78567) | > current_lr: 0.00022 | > step_time: 0.97150 (0.65123) | > loader_time: 0.00420 (0.00343)  --> STEP: 605/639 -- GLOBAL_STEP: 79325 | > loss: -0.46190 (-0.50482) | > log_mle: -0.56435 (-0.61807) | > loss_dur: 0.10246 (0.11325) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.01281 (9.69250) | > current_lr: 0.00022 | > step_time: 1.04280 (0.66667) | > loader_time: 0.00650 (0.00349)  --> STEP: 630/639 -- GLOBAL_STEP: 79350 | > loss: -0.46538 (-0.50357) | > log_mle: -0.59068 (-0.61703) | > loss_dur: 0.12530 (0.11346) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.14614 (9.63237) | > current_lr: 0.00022 | > step_time: 1.18750 (0.68406) | > loader_time: 0.00450 (0.00354) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.04450 (+0.01165) | > avg_loss: -0.45568 (+0.00097) | > avg_log_mle: -0.59278 (+0.00124) | > avg_loss_dur: 0.13710 (-0.00027)  > EPOCH: 124/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 06:40:15)   --> STEP: 15/639 -- GLOBAL_STEP: 79375 | > loss: -0.66923 (-0.57172) | > log_mle: -0.79984 (-0.68548) | > loss_dur: 0.13061 (0.11376) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 29.39929 (12.24792) | > current_lr: 0.00022 | > step_time: 0.32830 (0.29473) | > loader_time: 0.00260 (0.00271)  --> STEP: 40/639 -- GLOBAL_STEP: 79400 | > loss: -0.52937 (-0.55686) | > log_mle: -0.63849 (-0.66772) | > loss_dur: 0.10912 (0.11086) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.63222 (10.92181) | > current_lr: 0.00022 | > step_time: 0.35390 (0.32755) | > loader_time: 0.00270 (0.00273)  --> STEP: 65/639 -- GLOBAL_STEP: 79425 | > loss: -0.53672 (-0.54809) | > log_mle: -0.62940 (-0.65922) | > loss_dur: 0.09269 (0.11113) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.46352 (10.60118) | > current_lr: 0.00022 | > step_time: 0.39800 (0.36359) | > loader_time: 0.00300 (0.00282)  --> STEP: 90/639 -- GLOBAL_STEP: 79450 | > loss: -0.48178 (-0.53995) | > log_mle: -0.58817 (-0.64930) | > loss_dur: 0.10639 (0.10935) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.49088 (10.13669) | > current_lr: 0.00022 | > step_time: 0.51700 (0.38181) | > loader_time: 0.00320 (0.00286)  --> STEP: 115/639 -- GLOBAL_STEP: 79475 | > loss: -0.45483 (-0.53920) | > log_mle: -0.55786 (-0.64871) | > loss_dur: 0.10303 (0.10951) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.90339 (10.64518) | > current_lr: 0.00022 | > step_time: 0.41570 (0.39817) | > loader_time: 0.00300 (0.00289)  --> STEP: 140/639 -- GLOBAL_STEP: 79500 | > loss: -0.48779 (-0.53754) | > log_mle: -0.62725 (-0.64808) | > loss_dur: 0.13946 (0.11055) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 20.62265 (11.47954) | > current_lr: 0.00022 | > step_time: 0.56200 (0.41194) | > loader_time: 0.00330 (0.00291)  --> STEP: 165/639 -- GLOBAL_STEP: 79525 | > loss: -0.51699 (-0.53578) | > log_mle: -0.63844 (-0.64628) | > loss_dur: 0.12145 (0.11050) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.07972 (11.37171) | > current_lr: 0.00022 | > step_time: 0.48880 (0.42483) | > loader_time: 0.00310 (0.00295)  --> STEP: 190/639 -- GLOBAL_STEP: 79550 | > loss: -0.45716 (-0.53461) | > log_mle: -0.56980 (-0.64512) | > loss_dur: 0.11264 (0.11051) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.12999 (11.61024) | > current_lr: 0.00022 | > step_time: 0.52750 (0.44173) | > loader_time: 0.00340 (0.00300)  --> STEP: 215/639 -- GLOBAL_STEP: 79575 | > loss: -0.55414 (-0.53153) | > log_mle: -0.66761 (-0.64233) | > loss_dur: 0.11347 (0.11080) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.48494 (11.45376) | > current_lr: 0.00022 | > step_time: 0.53960 (0.45689) | > loader_time: 0.00320 (0.00304)  --> STEP: 240/639 -- GLOBAL_STEP: 79600 | > loss: -0.55805 (-0.53169) | > log_mle: -0.67734 (-0.64251) | > loss_dur: 0.11929 (0.11082) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 15.70725 (11.64428) | > current_lr: 0.00022 | > step_time: 0.55880 (0.47146) | > loader_time: 0.00330 (0.00307)  --> STEP: 265/639 -- GLOBAL_STEP: 79625 | > loss: -0.56097 (-0.52995) | > log_mle: -0.70443 (-0.64142) | > loss_dur: 0.14346 (0.11147) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 18.37916 (11.72447) | > current_lr: 0.00022 | > step_time: 0.69390 (0.48645) | > loader_time: 0.00350 (0.00311)  --> STEP: 290/639 -- GLOBAL_STEP: 79650 | > loss: -0.49057 (-0.52802) | > log_mle: -0.59637 (-0.63957) | > loss_dur: 0.10580 (0.11156) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.45870 (11.57123) | > current_lr: 0.00022 | > step_time: 0.61370 (0.50070) | > loader_time: 0.00330 (0.00314)  --> STEP: 315/639 -- GLOBAL_STEP: 79675 | > loss: -0.47689 (-0.52598) | > log_mle: -0.57991 (-0.63775) | > loss_dur: 0.10302 (0.11177) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.55536 (11.46264) | > current_lr: 0.00022 | > step_time: 0.62150 (0.51471) | > loader_time: 0.00340 (0.00317)  --> STEP: 340/639 -- GLOBAL_STEP: 79700 | > loss: -0.49063 (-0.52365) | > log_mle: -0.60046 (-0.63607) | > loss_dur: 0.10983 (0.11243) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.57172 (11.38216) | > current_lr: 0.00022 | > step_time: 0.60720 (0.53003) | > loader_time: 0.00350 (0.00320)  --> STEP: 365/639 -- GLOBAL_STEP: 79725 | > loss: -0.48892 (-0.52130) | > log_mle: -0.60602 (-0.63353) | > loss_dur: 0.11710 (0.11223) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.07806 (11.09597) | > current_lr: 0.00022 | > step_time: 0.79220 (0.54167) | > loader_time: 0.00380 (0.00323)  --> STEP: 390/639 -- GLOBAL_STEP: 79750 | > loss: -0.48511 (-0.51929) | > log_mle: -0.60623 (-0.63144) | > loss_dur: 0.12112 (0.11215) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.65597 (10.83129) | > current_lr: 0.00022 | > step_time: 0.70600 (0.55378) | > loader_time: 0.00330 (0.00325)  --> STEP: 415/639 -- GLOBAL_STEP: 79775 | > loss: -0.46265 (-0.51711) | > log_mle: -0.57968 (-0.62956) | > loss_dur: 0.11703 (0.11245) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.86601 (10.69611) | > current_lr: 0.00022 | > step_time: 0.75750 (0.56602) | > loader_time: 0.00340 (0.00328)  --> STEP: 440/639 -- GLOBAL_STEP: 79800 | > loss: -0.44603 (-0.51529) | > log_mle: -0.56046 (-0.62777) | > loss_dur: 0.11443 (0.11248) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.34095 (10.59084) | > current_lr: 0.00022 | > step_time: 0.71570 (0.57779) | > loader_time: 0.00370 (0.00330)  --> STEP: 465/639 -- GLOBAL_STEP: 79825 | > loss: -0.46076 (-0.51317) | > log_mle: -0.56831 (-0.62560) | > loss_dur: 0.10755 (0.11244) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.09633 (10.39212) | > current_lr: 0.00022 | > step_time: 0.76900 (0.58845) | > loader_time: 0.00390 (0.00333)  --> STEP: 490/639 -- GLOBAL_STEP: 79850 | > loss: -0.45069 (-0.51136) | > log_mle: -0.55229 (-0.62399) | > loss_dur: 0.10160 (0.11263) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 2.85634 (10.26246) | > current_lr: 0.00022 | > step_time: 0.79200 (0.60141) | > loader_time: 0.00360 (0.00335)  --> STEP: 515/639 -- GLOBAL_STEP: 79875 | > loss: -0.47261 (-0.51004) | > log_mle: -0.58162 (-0.62273) | > loss_dur: 0.10901 (0.11269) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.68486 (10.17636) | > current_lr: 0.00022 | > step_time: 0.97240 (0.61410) | > loader_time: 0.00410 (0.00338)  --> STEP: 540/639 -- GLOBAL_STEP: 79900 | > loss: -0.47529 (-0.50836) | > log_mle: -0.59542 (-0.62116) | > loss_dur: 0.12014 (0.11280) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.45695 (10.10580) | > current_lr: 0.00022 | > step_time: 0.93990 (0.62812) | > loader_time: 0.00420 (0.00342)  --> STEP: 565/639 -- GLOBAL_STEP: 79925 | > loss: -0.45205 (-0.50667) | > log_mle: -0.55915 (-0.61958) | > loss_dur: 0.10710 (0.11292) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.27087 (9.97382) | > current_lr: 0.00022 | > step_time: 0.99060 (0.64228) | > loader_time: 0.00420 (0.00344)  --> STEP: 590/639 -- GLOBAL_STEP: 79950 | > loss: -0.47795 (-0.50545) | > log_mle: -0.58824 (-0.61855) | > loss_dur: 0.11029 (0.11310) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.74218 (9.89209) | > current_lr: 0.00022 | > step_time: 0.93440 (0.65793) | > loader_time: 0.00400 (0.00348)  --> STEP: 615/639 -- GLOBAL_STEP: 79975 | > loss: -0.47006 (-0.50405) | > log_mle: -0.59093 (-0.61727) | > loss_dur: 0.12087 (0.11323) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.45597 (9.79484) | > current_lr: 0.00022 | > step_time: 0.99080 (0.67349) | > loader_time: 0.00440 (0.00352) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03035 (-0.01415) | > avg_loss: -0.45594 (-0.00025) | > avg_log_mle: -0.59367 (-0.00088) | > avg_loss_dur: 0.13773 (+0.00063)  > EPOCH: 125/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 06:47:50)   --> STEP: 0/639 -- GLOBAL_STEP: 80000 | > loss: -0.58943 (-0.58943) | > log_mle: -0.72892 (-0.72892) | > loss_dur: 0.13949 (0.13949) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 23.94745 (23.94745) | > current_lr: 0.00022 | > step_time: 0.50300 (0.50295) | > loader_time: 0.45780 (0.45775) > CHECKPOINT : /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481/checkpoint_80000.pth.tar  --> STEP: 25/639 -- GLOBAL_STEP: 80025 | > loss: -0.55918 (-0.56538) | > log_mle: -0.67909 (-0.67990) | > loss_dur: 0.11991 (0.11452) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.96437 (13.33568) | > current_lr: 0.00022 | > step_time: 0.38700 (0.32391) | > loader_time: 0.00310 (0.00268)  --> STEP: 50/639 -- GLOBAL_STEP: 80050 | > loss: -0.51944 (-0.55403) | > log_mle: -0.61342 (-0.66473) | > loss_dur: 0.09398 (0.11070) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.90297 (11.74381) | > current_lr: 0.00022 | > step_time: 0.31490 (0.34722) | > loader_time: 0.00270 (0.00275)  --> STEP: 75/639 -- GLOBAL_STEP: 80075 | > loss: -0.53765 (-0.54510) | > log_mle: -0.63677 (-0.65493) | > loss_dur: 0.09913 (0.10983) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.23773 (11.05324) | > current_lr: 0.00022 | > step_time: 0.41170 (0.36797) | > loader_time: 0.00290 (0.00280)  --> STEP: 100/639 -- GLOBAL_STEP: 80100 | > loss: -0.52302 (-0.54195) | > log_mle: -0.64475 (-0.65139) | > loss_dur: 0.12173 (0.10944) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.31482 (10.77788) | > current_lr: 0.00022 | > step_time: 0.51000 (0.38341) | > loader_time: 0.00320 (0.00285)  --> STEP: 125/639 -- GLOBAL_STEP: 80125 | > loss: -0.52170 (-0.53826) | > log_mle: -0.63793 (-0.64864) | > loss_dur: 0.11624 (0.11038) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.81537 (10.60600) | > current_lr: 0.00022 | > step_time: 0.41410 (0.39838) | > loader_time: 0.00300 (0.00288)  --> STEP: 150/639 -- GLOBAL_STEP: 80150 | > loss: -0.49423 (-0.53605) | > log_mle: -0.60118 (-0.64703) | > loss_dur: 0.10695 (0.11097) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.04954 (10.74405) | > current_lr: 0.00022 | > step_time: 0.42250 (0.41473) | > loader_time: 0.00310 (0.00291)  --> STEP: 175/639 -- GLOBAL_STEP: 80175 | > loss: -0.51873 (-0.53526) | > log_mle: -0.61894 (-0.64664) | > loss_dur: 0.10020 (0.11138) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.04791 (10.86647) | > current_lr: 0.00022 | > step_time: 0.55050 (0.43016) | > loader_time: 0.00320 (0.00294)  --> STEP: 200/639 -- GLOBAL_STEP: 80200 | > loss: -0.49527 (-0.53297) | > log_mle: -0.60437 (-0.64382) | > loss_dur: 0.10910 (0.11085) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.66475 (10.66883) | > current_lr: 0.00022 | > step_time: 0.53880 (0.44503) | > loader_time: 0.00300 (0.00297)  --> STEP: 225/639 -- GLOBAL_STEP: 80225 | > loss: -0.49137 (-0.53033) | > log_mle: -0.58646 (-0.64124) | > loss_dur: 0.09509 (0.11090) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.16244 (10.60179) | > current_lr: 0.00022 | > step_time: 0.52260 (0.46085) | > loader_time: 0.00320 (0.00301)  --> STEP: 250/639 -- GLOBAL_STEP: 80250 | > loss: -0.49285 (-0.53051) | > log_mle: -0.58871 (-0.64175) | > loss_dur: 0.09587 (0.11124) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.97965 (10.81484) | > current_lr: 0.00022 | > step_time: 0.57450 (0.47486) | > loader_time: 0.00330 (0.00304)  --> STEP: 275/639 -- GLOBAL_STEP: 80275 | > loss: -0.51077 (-0.52994) | > log_mle: -0.62372 (-0.64184) | > loss_dur: 0.11294 (0.11190) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.43357 (10.98831) | > current_lr: 0.00022 | > step_time: 0.86570 (0.49128) | > loader_time: 0.00380 (0.00309)  --> STEP: 300/639 -- GLOBAL_STEP: 80300 | > loss: -0.50819 (-0.52751) | > log_mle: -0.61637 (-0.63959) | > loss_dur: 0.10819 (0.11208) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.56415 (10.84685) | > current_lr: 0.00022 | > step_time: 0.68560 (0.50505) | > loader_time: 0.00360 (0.00311)  --> STEP: 325/639 -- GLOBAL_STEP: 80325 | > loss: -0.50665 (-0.52515) | > log_mle: -0.61643 (-0.63768) | > loss_dur: 0.10978 (0.11253) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.55524 (10.79056) | > current_lr: 0.00022 | > step_time: 0.60030 (0.52066) | > loader_time: 0.00350 (0.00315)  --> STEP: 350/639 -- GLOBAL_STEP: 80350 | > loss: -0.52089 (-0.52292) | > log_mle: -0.64518 (-0.63556) | > loss_dur: 0.12429 (0.11264) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 13.80363 (10.68041) | > current_lr: 0.00022 | > step_time: 0.83760 (0.53515) | > loader_time: 0.00380 (0.00318)  --> STEP: 375/639 -- GLOBAL_STEP: 80375 | > loss: -0.47167 (-0.52062) | > log_mle: -0.57836 (-0.63307) | > loss_dur: 0.10668 (0.11245) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.85056 (10.50296) | > current_lr: 0.00022 | > step_time: 0.74480 (0.54872) | > loader_time: 0.00350 (0.00321)  --> STEP: 400/639 -- GLOBAL_STEP: 80400 | > loss: -0.49262 (-0.51888) | > log_mle: -0.60704 (-0.63144) | > loss_dur: 0.11442 (0.11256) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.24119 (10.40406) | > current_lr: 0.00022 | > step_time: 0.72300 (0.56009) | > loader_time: 0.00380 (0.00324)  --> STEP: 425/639 -- GLOBAL_STEP: 80425 | > loss: -0.54361 (-0.51652) | > log_mle: -0.66511 (-0.62921) | > loss_dur: 0.12150 (0.11268) | > amp_scaler: 8192.00000 (4173.10118) | > grad_norm: 14.97006 (10.24156) | > current_lr: 0.00022 | > step_time: 0.89090 (0.57207) | > loader_time: 0.00380 (0.00326)  --> STEP: 450/639 -- GLOBAL_STEP: 80450 | > loss: -0.46494 (-0.51458) | > log_mle: -0.56730 (-0.62721) | > loss_dur: 0.10236 (0.11263) | > amp_scaler: 8192.00000 (4396.37333) | > grad_norm: 3.94554 (10.07783) | > current_lr: 0.00022 | > step_time: 0.78650 (0.58397) | > loader_time: 0.00380 (0.00329)  --> STEP: 475/639 -- GLOBAL_STEP: 80475 | > loss: -0.47824 (-0.51254) | > log_mle: -0.57929 (-0.62531) | > loss_dur: 0.10105 (0.11277) | > amp_scaler: 8192.00000 (4596.14316) | > grad_norm: 4.79246 (9.93445) | > current_lr: 0.00022 | > step_time: 0.72780 (0.59572) | > loader_time: 0.00370 (0.00331)  --> STEP: 500/639 -- GLOBAL_STEP: 80500 | > loss: -0.51887 (-0.51103) | > log_mle: -0.63952 (-0.62394) | > loss_dur: 0.12065 (0.11290) | > amp_scaler: 8192.00000 (4775.93600) | > grad_norm: 10.68402 (9.85835) | > current_lr: 0.00022 | > step_time: 0.99900 (0.60861) | > loader_time: 0.00370 (0.00333)  --> STEP: 525/639 -- GLOBAL_STEP: 80525 | > loss: -0.48621 (-0.50956) | > log_mle: -0.61837 (-0.62257) | > loss_dur: 0.13215 (0.11301) | > amp_scaler: 4096.00000 (4868.38857) | > grad_norm: 12.27026 (9.75903) | > current_lr: 0.00022 | > step_time: 0.96120 (0.62157) | > loader_time: 0.00430 (0.00337)  --> STEP: 550/639 -- GLOBAL_STEP: 80550 | > loss: -0.44581 (-0.50791) | > log_mle: -0.55315 (-0.62094) | > loss_dur: 0.10733 (0.11303) | > amp_scaler: 4096.00000 (4833.28000) | > grad_norm: 3.44626 (9.62191) | > current_lr: 0.00022 | > step_time: 0.93980 (0.63617) | > loader_time: 0.00420 (0.00340)  --> STEP: 575/639 -- GLOBAL_STEP: 80575 | > loss: -0.46424 (-0.50623) | > log_mle: -0.57066 (-0.61939) | > loss_dur: 0.10642 (0.11316) | > amp_scaler: 4096.00000 (4801.22435) | > grad_norm: 5.31830 (9.51269) | > current_lr: 0.00022 | > step_time: 0.89130 (0.64995) | > loader_time: 0.00390 (0.00343)  --> STEP: 600/639 -- GLOBAL_STEP: 80600 | > loss: -0.45230 (-0.50489) | > log_mle: -0.56822 (-0.61825) | > loss_dur: 0.11591 (0.11336) | > amp_scaler: 4096.00000 (4771.84000) | > grad_norm: 7.08445 (9.42543) | > current_lr: 0.00022 | > step_time: 1.04010 (0.66618) | > loader_time: 0.00440 (0.00346)  --> STEP: 625/639 -- GLOBAL_STEP: 80625 | > loss: -0.48025 (-0.50360) | > log_mle: -0.59707 (-0.61705) | > loss_dur: 0.11682 (0.11346) | > amp_scaler: 4096.00000 (4744.80640) | > grad_norm: 6.61607 (9.35376) | > current_lr: 0.00022 | > step_time: 0.99700 (0.68262) | > loader_time: 0.00430 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02807 (-0.00229) | > avg_loss: -0.45877 (-0.00284) | > avg_log_mle: -0.59577 (-0.00210) | > avg_loss_dur: 0.13700 (-0.00073)  > EPOCH: 126/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 06:55:30)   --> STEP: 10/639 -- GLOBAL_STEP: 80650 | > loss: -0.55660 (-0.56729) | > log_mle: -0.67820 (-0.68179) | > loss_dur: 0.12160 (0.11449) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.90799 (11.51187) | > current_lr: 0.00022 | > step_time: 0.30310 (0.30180) | > loader_time: 0.00270 (0.00246)  --> STEP: 35/639 -- GLOBAL_STEP: 80675 | > loss: -0.54915 (-0.55711) | > log_mle: -0.64203 (-0.66870) | > loss_dur: 0.09289 (0.11158) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.00476 (10.14104) | > current_lr: 0.00022 | > step_time: 0.32250 (0.32667) | > loader_time: 0.00270 (0.00265)  --> STEP: 60/639 -- GLOBAL_STEP: 80700 | > loss: -0.51697 (-0.55024) | > log_mle: -0.60674 (-0.66080) | > loss_dur: 0.08977 (0.11056) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.67268 (9.72901) | > current_lr: 0.00022 | > step_time: 0.33820 (0.35579) | > loader_time: 0.00280 (0.00269)  --> STEP: 85/639 -- GLOBAL_STEP: 80725 | > loss: -0.50884 (-0.54309) | > log_mle: -0.60733 (-0.65155) | > loss_dur: 0.09850 (0.10847) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.95898 (9.20363) | > current_lr: 0.00022 | > step_time: 0.41480 (0.37221) | > loader_time: 0.00280 (0.00276)  --> STEP: 110/639 -- GLOBAL_STEP: 80750 | > loss: -0.51340 (-0.54252) | > log_mle: -0.61759 (-0.65217) | > loss_dur: 0.10419 (0.10965) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.94437 (9.61629) | > current_lr: 0.00022 | > step_time: 0.40020 (0.39223) | > loader_time: 0.00310 (0.00284)  --> STEP: 135/639 -- GLOBAL_STEP: 80775 | > loss: -0.48325 (-0.53877) | > log_mle: -0.59656 (-0.64870) | > loss_dur: 0.11331 (0.10993) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.65441 (9.76727) | > current_lr: 0.00022 | > step_time: 0.46260 (0.40531) | > loader_time: 0.00310 (0.00288)  --> STEP: 160/639 -- GLOBAL_STEP: 80800 | > loss: -0.53262 (-0.53640) | > log_mle: -0.63481 (-0.64611) | > loss_dur: 0.10219 (0.10971) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.12971 (9.83206) | > current_lr: 0.00022 | > step_time: 0.46700 (0.41904) | > loader_time: 0.00250 (0.00291)  --> STEP: 185/639 -- GLOBAL_STEP: 80825 | > loss: -0.51437 (-0.53700) | > log_mle: -0.61471 (-0.64763) | > loss_dur: 0.10034 (0.11063) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.42996 (10.21437) | > current_lr: 0.00022 | > step_time: 0.57300 (0.43625) | > loader_time: 0.00320 (0.00294)  --> STEP: 210/639 -- GLOBAL_STEP: 80850 | > loss: -0.43975 (-0.53334) | > log_mle: -0.53790 (-0.64403) | > loss_dur: 0.09815 (0.11069) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.19816 (9.99490) | > current_lr: 0.00022 | > step_time: 0.50320 (0.45270) | > loader_time: 0.00300 (0.00296)  --> STEP: 235/639 -- GLOBAL_STEP: 80875 | > loss: -0.51493 (-0.53315) | > log_mle: -0.62454 (-0.64389) | > loss_dur: 0.10961 (0.11074) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 11.66307 (10.12825) | > current_lr: 0.00022 | > step_time: 0.55190 (0.46804) | > loader_time: 0.00320 (0.00300)  --> STEP: 260/639 -- GLOBAL_STEP: 80900 | > loss: -0.50934 (-0.53168) | > log_mle: -0.62691 (-0.64295) | > loss_dur: 0.11757 (0.11127) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.05552 (10.31123) | > current_lr: 0.00022 | > step_time: 0.57460 (0.48219) | > loader_time: 0.00350 (0.00303)  --> STEP: 285/639 -- GLOBAL_STEP: 80925 | > loss: -0.47405 (-0.52996) | > log_mle: -0.57625 (-0.64155) | > loss_dur: 0.10220 (0.11158) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.41286 (10.29991) | > current_lr: 0.00022 | > step_time: 0.60500 (0.49828) | > loader_time: 0.00350 (0.00307)  --> STEP: 310/639 -- GLOBAL_STEP: 80950 | > loss: -0.50113 (-0.52774) | > log_mle: -0.61448 (-0.63966) | > loss_dur: 0.11336 (0.11192) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.27386 (10.18008) | > current_lr: 0.00022 | > step_time: 0.56710 (0.51186) | > loader_time: 0.00290 (0.00310)  --> STEP: 335/639 -- GLOBAL_STEP: 80975 | > loss: -0.51309 (-0.52489) | > log_mle: -0.63740 (-0.63734) | > loss_dur: 0.12432 (0.11245) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 14.47760 (10.10257) | > current_lr: 0.00022 | > step_time: 0.68550 (0.52684) | > loader_time: 0.00350 (0.00313)  --> STEP: 360/639 -- GLOBAL_STEP: 81000 | > loss: -0.47251 (-0.52292) | > log_mle: -0.57818 (-0.63521) | > loss_dur: 0.10567 (0.11228) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.44194 (9.95771) | > current_lr: 0.00022 | > step_time: 0.70930 (0.53971) | > loader_time: 0.00310 (0.00316)  --> STEP: 385/639 -- GLOBAL_STEP: 81025 | > loss: -0.49583 (-0.52095) | > log_mle: -0.60968 (-0.63321) | > loss_dur: 0.11385 (0.11226) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.78028 (9.82595) | > current_lr: 0.00022 | > step_time: 0.71350 (0.55380) | > loader_time: 0.00360 (0.00319)  --> STEP: 410/639 -- GLOBAL_STEP: 81050 | > loss: -0.45153 (-0.51884) | > log_mle: -0.58092 (-0.63134) | > loss_dur: 0.12940 (0.11250) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.93794 (9.72700) | > current_lr: 0.00022 | > step_time: 0.77900 (0.56551) | > loader_time: 0.00330 (0.00321)  --> STEP: 435/639 -- GLOBAL_STEP: 81075 | > loss: -0.49321 (-0.51671) | > log_mle: -0.59464 (-0.62918) | > loss_dur: 0.10142 (0.11247) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.53816 (9.60522) | > current_lr: 0.00022 | > step_time: 0.79250 (0.57770) | > loader_time: 0.00350 (0.00323)  --> STEP: 460/639 -- GLOBAL_STEP: 81100 | > loss: -0.44741 (-0.51458) | > log_mle: -0.56170 (-0.62701) | > loss_dur: 0.11429 (0.11243) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.22174 (9.45724) | > current_lr: 0.00022 | > step_time: 0.90440 (0.58964) | > loader_time: 0.00390 (0.00326)  --> STEP: 485/639 -- GLOBAL_STEP: 81125 | > loss: -0.48619 (-0.51280) | > log_mle: -0.59168 (-0.62548) | > loss_dur: 0.10550 (0.11268) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.11491 (9.39490) | > current_lr: 0.00022 | > step_time: 0.92200 (0.60237) | > loader_time: 0.00420 (0.00329)  --> STEP: 510/639 -- GLOBAL_STEP: 81150 | > loss: -0.47741 (-0.51132) | > log_mle: -0.59305 (-0.62404) | > loss_dur: 0.11564 (0.11272) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.63140 (9.31277) | > current_lr: 0.00022 | > step_time: 0.85380 (0.61500) | > loader_time: 0.00370 (0.00331)  --> STEP: 535/639 -- GLOBAL_STEP: 81175 | > loss: -0.45704 (-0.50965) | > log_mle: -0.58479 (-0.62246) | > loss_dur: 0.12775 (0.11282) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.80672 (9.21624) | > current_lr: 0.00022 | > step_time: 0.89500 (0.62811) | > loader_time: 0.00400 (0.00334)  --> STEP: 560/639 -- GLOBAL_STEP: 81200 | > loss: -0.47147 (-0.50806) | > log_mle: -0.59310 (-0.62103) | > loss_dur: 0.12163 (0.11296) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.62149 (9.16161) | > current_lr: 0.00022 | > step_time: 0.93150 (0.64309) | > loader_time: 0.00400 (0.00338)  --> STEP: 585/639 -- GLOBAL_STEP: 81225 | > loss: -0.48720 (-0.50667) | > log_mle: -0.59519 (-0.61972) | > loss_dur: 0.10799 (0.11306) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.06726 (9.09678) | > current_lr: 0.00022 | > step_time: 1.04740 (0.65817) | > loader_time: 0.00440 (0.00342)  --> STEP: 610/639 -- GLOBAL_STEP: 81250 | > loss: -0.45514 (-0.50526) | > log_mle: -0.56610 (-0.61848) | > loss_dur: 0.11096 (0.11322) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.57163 (9.06072) | > current_lr: 0.00022 | > step_time: 1.05280 (0.67417) | > loader_time: 0.00450 (0.00346)  --> STEP: 635/639 -- GLOBAL_STEP: 81275 | > loss: -0.46485 (-0.50403) | > log_mle: -0.58965 (-0.61762) | > loss_dur: 0.12480 (0.11359) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.95622 (9.08254) | > current_lr: 0.00022 | > step_time: 1.12070 (0.69101) | > loader_time: 0.00420 (0.00350) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02427 (-0.00380) | > avg_loss: -0.45863 (+0.00014) | > avg_log_mle: -0.59426 (+0.00151) | > avg_loss_dur: 0.13564 (-0.00136)  > EPOCH: 127/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 07:03:07)   --> STEP: 20/639 -- GLOBAL_STEP: 81300 | > loss: -0.64615 (-0.57515) | > log_mle: -0.79501 (-0.69208) | > loss_dur: 0.14886 (0.11694) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 32.38326 (13.66543) | > current_lr: 0.00022 | > step_time: 0.36820 (0.30992) | > loader_time: 0.00250 (0.00255)  --> STEP: 45/639 -- GLOBAL_STEP: 81325 | > loss: -0.52425 (-0.55680) | > log_mle: -0.61524 (-0.66866) | > loss_dur: 0.09099 (0.11186) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.64129 (12.36284) | > current_lr: 0.00022 | > step_time: 0.40800 (0.33723) | > loader_time: 0.00280 (0.00263)  --> STEP: 70/639 -- GLOBAL_STEP: 81350 | > loss: -0.53946 (-0.54706) | > log_mle: -0.65650 (-0.65815) | > loss_dur: 0.11703 (0.11109) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.87569 (11.42755) | > current_lr: 0.00022 | > step_time: 0.39130 (0.36358) | > loader_time: 0.00300 (0.00273)  --> STEP: 95/639 -- GLOBAL_STEP: 81375 | > loss: -0.59039 (-0.54296) | > log_mle: -0.70043 (-0.65216) | > loss_dur: 0.11004 (0.10919) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 21.24570 (11.45433) | > current_lr: 0.00022 | > step_time: 0.40340 (0.37846) | > loader_time: 0.00260 (0.00277)  --> STEP: 120/639 -- GLOBAL_STEP: 81400 | > loss: -0.56262 (-0.53944) | > log_mle: -0.69716 (-0.64972) | > loss_dur: 0.13454 (0.11028) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 24.00999 (11.33201) | > current_lr: 0.00022 | > step_time: 0.53920 (0.39819) | > loader_time: 0.00320 (0.00281)  --> STEP: 145/639 -- GLOBAL_STEP: 81425 | > loss: -0.52542 (-0.53803) | > log_mle: -0.64978 (-0.64902) | > loss_dur: 0.12436 (0.11099) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 12.70886 (11.37914) | > current_lr: 0.00022 | > step_time: 0.45270 (0.41461) | > loader_time: 0.00300 (0.00285)  --> STEP: 170/639 -- GLOBAL_STEP: 81450 | > loss: -0.52331 (-0.53613) | > log_mle: -0.63439 (-0.64705) | > loss_dur: 0.11108 (0.11092) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.88174 (11.20743) | > current_lr: 0.00022 | > step_time: 0.54550 (0.42854) | > loader_time: 0.00330 (0.00287)  --> STEP: 195/639 -- GLOBAL_STEP: 81475 | > loss: -0.49219 (-0.53459) | > log_mle: -0.58397 (-0.64538) | > loss_dur: 0.09178 (0.11079) | > amp_scaler: 2048.00000 (3896.45128) | > grad_norm: 3.22660 (11.00359) | > current_lr: 0.00022 | > step_time: 0.49150 (0.44452) | > loader_time: 0.00320 (0.00291)  --> STEP: 220/639 -- GLOBAL_STEP: 81500 | > loss: -0.52029 (-0.53166) | > log_mle: -0.62904 (-0.64258) | > loss_dur: 0.10875 (0.11093) | > amp_scaler: 2048.00000 (3686.40000) | > grad_norm: 6.65646 (10.87900) | > current_lr: 0.00022 | > step_time: 0.64140 (0.45985) | > loader_time: 0.00330 (0.00295)  --> STEP: 245/639 -- GLOBAL_STEP: 81525 | > loss: -0.51709 (-0.53173) | > log_mle: -0.61765 (-0.64281) | > loss_dur: 0.10055 (0.11109) | > amp_scaler: 2048.00000 (3519.21633) | > grad_norm: 7.73519 (10.89336) | > current_lr: 0.00022 | > step_time: 0.53880 (0.47304) | > loader_time: 0.00280 (0.00298)  --> STEP: 270/639 -- GLOBAL_STEP: 81550 | > loss: -0.55111 (-0.53082) | > log_mle: -0.67993 (-0.64241) | > loss_dur: 0.12883 (0.11159) | > amp_scaler: 2048.00000 (3382.99259) | > grad_norm: 19.69765 (11.00167) | > current_lr: 0.00022 | > step_time: 0.82670 (0.48933) | > loader_time: 0.00370 (0.00302)  --> STEP: 295/639 -- GLOBAL_STEP: 81575 | > loss: -0.48419 (-0.52843) | > log_mle: -0.57872 (-0.64018) | > loss_dur: 0.09454 (0.11175) | > amp_scaler: 2048.00000 (3269.85763) | > grad_norm: 4.02678 (10.81065) | > current_lr: 0.00022 | > step_time: 0.64880 (0.50298) | > loader_time: 0.00340 (0.00305)  --> STEP: 320/639 -- GLOBAL_STEP: 81600 | > loss: -0.49964 (-0.52626) | > log_mle: -0.62217 (-0.63847) | > loss_dur: 0.12254 (0.11221) | > amp_scaler: 2048.00000 (3174.40000) | > grad_norm: 9.87243 (10.74377) | > current_lr: 0.00022 | > step_time: 0.79220 (0.51773) | > loader_time: 0.00350 (0.00309)  --> STEP: 345/639 -- GLOBAL_STEP: 81625 | > loss: -0.49510 (-0.52397) | > log_mle: -0.60558 (-0.63648) | > loss_dur: 0.11049 (0.11250) | > amp_scaler: 2048.00000 (3092.77681) | > grad_norm: 13.05580 (10.62493) | > current_lr: 0.00022 | > step_time: 0.64440 (0.53155) | > loader_time: 0.00340 (0.00313)  --> STEP: 370/639 -- GLOBAL_STEP: 81650 | > loss: -0.51263 (-0.52180) | > log_mle: -0.62187 (-0.63411) | > loss_dur: 0.10924 (0.11231) | > amp_scaler: 2048.00000 (3022.18378) | > grad_norm: 10.94026 (10.46397) | > current_lr: 0.00022 | > step_time: 0.84410 (0.54319) | > loader_time: 0.00410 (0.00316)  --> STEP: 395/639 -- GLOBAL_STEP: 81675 | > loss: -0.47164 (-0.51966) | > log_mle: -0.58772 (-0.63193) | > loss_dur: 0.11608 (0.11227) | > amp_scaler: 2048.00000 (2960.52658) | > grad_norm: 7.52102 (10.30929) | > current_lr: 0.00022 | > step_time: 0.73810 (0.55554) | > loader_time: 0.00360 (0.00319)  --> STEP: 420/639 -- GLOBAL_STEP: 81700 | > loss: -0.46934 (-0.51727) | > log_mle: -0.58289 (-0.62986) | > loss_dur: 0.11356 (0.11259) | > amp_scaler: 2048.00000 (2906.20952) | > grad_norm: 7.86152 (10.19447) | > current_lr: 0.00022 | > step_time: 0.69010 (0.56774) | > loader_time: 0.00360 (0.00322)  --> STEP: 445/639 -- GLOBAL_STEP: 81725 | > loss: -0.48705 (-0.51557) | > log_mle: -0.59208 (-0.62817) | > loss_dur: 0.10503 (0.11260) | > amp_scaler: 2048.00000 (2857.99551) | > grad_norm: 6.63624 (10.06211) | > current_lr: 0.00022 | > step_time: 0.72930 (0.57962) | > loader_time: 0.00370 (0.00325)  --> STEP: 470/639 -- GLOBAL_STEP: 81750 | > loss: -0.49098 (-0.51352) | > log_mle: -0.60423 (-0.62613) | > loss_dur: 0.11325 (0.11261) | > amp_scaler: 2048.00000 (2814.91064) | > grad_norm: 7.29071 (9.87842) | > current_lr: 0.00022 | > step_time: 0.76260 (0.59118) | > loader_time: 0.00390 (0.00327)  --> STEP: 495/639 -- GLOBAL_STEP: 81775 | > loss: -0.43863 (-0.51178) | > log_mle: -0.54012 (-0.62453) | > loss_dur: 0.10149 (0.11275) | > amp_scaler: 2048.00000 (2776.17778) | > grad_norm: 1.89351 (9.77501) | > current_lr: 0.00022 | > step_time: 0.75600 (0.60460) | > loader_time: 0.00360 (0.00330)  --> STEP: 520/639 -- GLOBAL_STEP: 81800 | > loss: -0.48731 (-0.51034) | > log_mle: -0.59504 (-0.62318) | > loss_dur: 0.10774 (0.11284) | > amp_scaler: 2048.00000 (2741.16923) | > grad_norm: 8.65001 (9.66524) | > current_lr: 0.00022 | > step_time: 0.80710 (0.61768) | > loader_time: 0.00370 (0.00333)  --> STEP: 545/639 -- GLOBAL_STEP: 81825 | > loss: -0.50218 (-0.50879) | > log_mle: -0.62631 (-0.62180) | > loss_dur: 0.12413 (0.11302) | > amp_scaler: 2048.00000 (2709.37248) | > grad_norm: 11.45232 (9.58733) | > current_lr: 0.00022 | > step_time: 1.02890 (0.63294) | > loader_time: 0.00380 (0.00337)  --> STEP: 570/639 -- GLOBAL_STEP: 81850 | > loss: -0.46705 (-0.50710) | > log_mle: -0.57336 (-0.62021) | > loss_dur: 0.10631 (0.11311) | > amp_scaler: 2048.00000 (2680.36491) | > grad_norm: 6.97483 (9.47889) | > current_lr: 0.00022 | > step_time: 0.91930 (0.64679) | > loader_time: 0.00390 (0.00340)  --> STEP: 595/639 -- GLOBAL_STEP: 81875 | > loss: -0.45998 (-0.50592) | > log_mle: -0.56808 (-0.61917) | > loss_dur: 0.10810 (0.11325) | > amp_scaler: 2048.00000 (2653.79496) | > grad_norm: 5.65399 (9.40802) | > current_lr: 0.00022 | > step_time: 0.97440 (0.66308) | > loader_time: 0.00390 (0.00346)  --> STEP: 620/639 -- GLOBAL_STEP: 81900 | > loss: -0.46193 (-0.50456) | > log_mle: -0.57317 (-0.61794) | > loss_dur: 0.11125 (0.11339) | > amp_scaler: 2048.00000 (2629.36774) | > grad_norm: 7.18136 (9.35012) | > current_lr: 0.00022 | > step_time: 1.14420 (0.67946) | > loader_time: 0.00420 (0.00351) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03056 (+0.00629) | > avg_loss: -0.45842 (+0.00021) | > avg_log_mle: -0.59544 (-0.00118) | > avg_loss_dur: 0.13702 (+0.00139)  > EPOCH: 128/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 07:10:44)   --> STEP: 5/639 -- GLOBAL_STEP: 81925 | > loss: -0.57657 (-0.57864) | > log_mle: -0.71628 (-0.69739) | > loss_dur: 0.13971 (0.11874) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 21.59923 (15.48718) | > current_lr: 0.00022 | > step_time: 0.30840 (0.27646) | > loader_time: 0.00270 (0.00221)  --> STEP: 30/639 -- GLOBAL_STEP: 81950 | > loss: -0.54577 (-0.56270) | > log_mle: -0.66002 (-0.67541) | > loss_dur: 0.11425 (0.11272) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.41985 (12.98563) | > current_lr: 0.00022 | > step_time: 0.32030 (0.32371) | > loader_time: 0.00280 (0.00260)  --> STEP: 55/639 -- GLOBAL_STEP: 81975 | > loss: -0.52012 (-0.55445) | > log_mle: -0.61096 (-0.66463) | > loss_dur: 0.09084 (0.11018) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.48687 (11.05157) | > current_lr: 0.00022 | > step_time: 0.33440 (0.34734) | > loader_time: 0.00300 (0.00268)  --> STEP: 80/639 -- GLOBAL_STEP: 82000 | > loss: -0.50304 (-0.54632) | > log_mle: -0.63864 (-0.65556) | > loss_dur: 0.13560 (0.10924) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.59057 (10.56975) | > current_lr: 0.00022 | > step_time: 0.39250 (0.36920) | > loader_time: 0.00270 (0.00277)  --> STEP: 105/639 -- GLOBAL_STEP: 82025 | > loss: -0.48542 (-0.54310) | > log_mle: -0.59398 (-0.65238) | > loss_dur: 0.10856 (0.10928) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.26375 (10.54611) | > current_lr: 0.00022 | > step_time: 0.41600 (0.38725) | > loader_time: 0.00300 (0.00283)  --> STEP: 130/639 -- GLOBAL_STEP: 82050 | > loss: -0.51433 (-0.53991) | > log_mle: -0.63319 (-0.64987) | > loss_dur: 0.11886 (0.10996) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.58530 (10.96844) | > current_lr: 0.00022 | > step_time: 0.48800 (0.40210) | > loader_time: 0.00310 (0.00286)  --> STEP: 155/639 -- GLOBAL_STEP: 82075 | > loss: -0.51199 (-0.53807) | > log_mle: -0.62190 (-0.64837) | > loss_dur: 0.10991 (0.11030) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 12.66059 (11.14126) | > current_lr: 0.00022 | > step_time: 0.58270 (0.41943) | > loader_time: 0.00320 (0.00288)  --> STEP: 180/639 -- GLOBAL_STEP: 82100 | > loss: -0.55228 (-0.53854) | > log_mle: -0.67121 (-0.64933) | > loss_dur: 0.11893 (0.11079) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 20.00676 (11.41407) | > current_lr: 0.00022 | > step_time: 0.57340 (0.43548) | > loader_time: 0.00340 (0.00292)  --> STEP: 205/639 -- GLOBAL_STEP: 82125 | > loss: -0.49343 (-0.53472) | > log_mle: -0.59732 (-0.64507) | > loss_dur: 0.10389 (0.11035) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.49498 (11.06551) | > current_lr: 0.00022 | > step_time: 0.54930 (0.44984) | > loader_time: 0.00330 (0.00297)  --> STEP: 230/639 -- GLOBAL_STEP: 82150 | > loss: -0.52022 (-0.53397) | > log_mle: -0.63267 (-0.64447) | > loss_dur: 0.11245 (0.11050) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.51181 (11.13227) | > current_lr: 0.00022 | > step_time: 0.66720 (0.46718) | > loader_time: 0.00340 (0.00300)  --> STEP: 255/639 -- GLOBAL_STEP: 82175 | > loss: -0.47945 (-0.53222) | > log_mle: -0.59977 (-0.64283) | > loss_dur: 0.12032 (0.11062) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.97183 (11.01667) | > current_lr: 0.00022 | > step_time: 0.63410 (0.48102) | > loader_time: 0.00350 (0.00303)  --> STEP: 280/639 -- GLOBAL_STEP: 82200 | > loss: -0.51225 (-0.53134) | > log_mle: -0.62457 (-0.64265) | > loss_dur: 0.11232 (0.11131) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.39038 (11.08185) | > current_lr: 0.00022 | > step_time: 0.58970 (0.49720) | > loader_time: 0.00340 (0.00307)  --> STEP: 305/639 -- GLOBAL_STEP: 82225 | > loss: -0.48954 (-0.52914) | > log_mle: -0.60715 (-0.64069) | > loss_dur: 0.11761 (0.11155) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.72385 (11.00589) | > current_lr: 0.00022 | > step_time: 0.68620 (0.51097) | > loader_time: 0.00350 (0.00311)  --> STEP: 330/639 -- GLOBAL_STEP: 82250 | > loss: -0.50194 (-0.52640) | > log_mle: -0.62578 (-0.63841) | > loss_dur: 0.12384 (0.11201) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.08778 (10.89444) | > current_lr: 0.00022 | > step_time: 0.74260 (0.52484) | > loader_time: 0.00370 (0.00315)  --> STEP: 355/639 -- GLOBAL_STEP: 82275 | > loss: -0.45794 (-0.52437) | > log_mle: -0.55789 (-0.63624) | > loss_dur: 0.09995 (0.11187) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.13298 (10.74417) | > current_lr: 0.00022 | > step_time: 0.58120 (0.53765) | > loader_time: 0.00350 (0.00318)  --> STEP: 380/639 -- GLOBAL_STEP: 82300 | > loss: -0.49539 (-0.52217) | > log_mle: -0.59572 (-0.63389) | > loss_dur: 0.10033 (0.11172) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.68788 (10.53530) | > current_lr: 0.00022 | > step_time: 0.64650 (0.55060) | > loader_time: 0.00350 (0.00320)  --> STEP: 405/639 -- GLOBAL_STEP: 82325 | > loss: -0.49417 (-0.52039) | > log_mle: -0.59718 (-0.63243) | > loss_dur: 0.10302 (0.11204) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.51286 (10.47965) | > current_lr: 0.00022 | > step_time: 0.70750 (0.56229) | > loader_time: 0.00350 (0.00323)  --> STEP: 430/639 -- GLOBAL_STEP: 82350 | > loss: -0.45556 (-0.51810) | > log_mle: -0.55791 (-0.63035) | > loss_dur: 0.10235 (0.11225) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.27252 (10.41037) | > current_lr: 0.00022 | > step_time: 0.80060 (0.57489) | > loader_time: 0.00380 (0.00326)  --> STEP: 455/639 -- GLOBAL_STEP: 82375 | > loss: -0.47650 (-0.51605) | > log_mle: -0.58237 (-0.62821) | > loss_dur: 0.10586 (0.11216) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.66210 (10.23246) | > current_lr: 0.00022 | > step_time: 0.72470 (0.58558) | > loader_time: 0.00360 (0.00328)  --> STEP: 480/639 -- GLOBAL_STEP: 82400 | > loss: -0.43359 (-0.51414) | > log_mle: -0.53905 (-0.62647) | > loss_dur: 0.10547 (0.11233) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.11130 (10.15663) | > current_lr: 0.00022 | > step_time: 0.77900 (0.59808) | > loader_time: 0.00360 (0.00330)  --> STEP: 505/639 -- GLOBAL_STEP: 82425 | > loss: -0.49285 (-0.51258) | > log_mle: -0.60175 (-0.62490) | > loss_dur: 0.10890 (0.11232) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.96052 (10.02399) | > current_lr: 0.00022 | > step_time: 0.87820 (0.60985) | > loader_time: 0.00380 (0.00332)  --> STEP: 530/639 -- GLOBAL_STEP: 82450 | > loss: -0.46826 (-0.51095) | > log_mle: -0.59190 (-0.62349) | > loss_dur: 0.12364 (0.11254) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.61086 (9.95917) | > current_lr: 0.00022 | > step_time: 0.93510 (0.62402) | > loader_time: 0.00400 (0.00336)  --> STEP: 555/639 -- GLOBAL_STEP: 82475 | > loss: -0.47129 (-0.50927) | > log_mle: -0.57933 (-0.62185) | > loss_dur: 0.10804 (0.11258) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.40731 (9.83757) | > current_lr: 0.00022 | > step_time: 0.80610 (0.63795) | > loader_time: 0.00390 (0.00339)  --> STEP: 580/639 -- GLOBAL_STEP: 82500 | > loss: -0.48795 (-0.50782) | > log_mle: -0.59452 (-0.62055) | > loss_dur: 0.10658 (0.11273) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.33770 (9.80639) | > current_lr: 0.00022 | > step_time: 0.95730 (0.65308) | > loader_time: 0.00410 (0.00342)  --> STEP: 605/639 -- GLOBAL_STEP: 82525 | > loss: -0.45583 (-0.50643) | > log_mle: -0.56309 (-0.61933) | > loss_dur: 0.10725 (0.11290) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.08942 (9.71837) | > current_lr: 0.00022 | > step_time: 0.98460 (0.66896) | > loader_time: 0.00440 (0.00346)  --> STEP: 630/639 -- GLOBAL_STEP: 82550 | > loss: -0.46297 (-0.50513) | > log_mle: -0.59178 (-0.61826) | > loss_dur: 0.12881 (0.11313) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.33996 (9.66734) | > current_lr: 0.00022 | > step_time: 1.19040 (0.68571) | > loader_time: 0.00450 (0.00350) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03402 (+0.00346) | > avg_loss: -0.46279 (-0.00437) | > avg_log_mle: -0.59850 (-0.00306) | > avg_loss_dur: 0.13571 (-0.00132)  > EPOCH: 129/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 07:18:20)   --> STEP: 15/639 -- GLOBAL_STEP: 82575 | > loss: -0.66145 (-0.57516) | > log_mle: -0.79351 (-0.68845) | > loss_dur: 0.13206 (0.11328) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 32.28907 (14.51537) | > current_lr: 0.00022 | > step_time: 0.32420 (0.28943) | > loader_time: 0.00320 (0.00253)  --> STEP: 40/639 -- GLOBAL_STEP: 82600 | > loss: -0.53663 (-0.55930) | > log_mle: -0.64414 (-0.67017) | > loss_dur: 0.10751 (0.11087) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.76311 (12.28406) | > current_lr: 0.00022 | > step_time: 0.35950 (0.32378) | > loader_time: 0.00290 (0.00264)  --> STEP: 65/639 -- GLOBAL_STEP: 82625 | > loss: -0.53585 (-0.55114) | > log_mle: -0.62875 (-0.66194) | > loss_dur: 0.09290 (0.11080) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.79066 (11.54718) | > current_lr: 0.00022 | > step_time: 0.38580 (0.35698) | > loader_time: 0.00310 (0.00277)  --> STEP: 90/639 -- GLOBAL_STEP: 82650 | > loss: -0.48595 (-0.54332) | > log_mle: -0.58907 (-0.65193) | > loss_dur: 0.10312 (0.10861) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.55562 (10.72040) | > current_lr: 0.00022 | > step_time: 0.51590 (0.37681) | > loader_time: 0.00320 (0.00282)  --> STEP: 115/639 -- GLOBAL_STEP: 82675 | > loss: -0.46384 (-0.54191) | > log_mle: -0.55840 (-0.65103) | > loss_dur: 0.09456 (0.10912) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.08376 (10.61259) | > current_lr: 0.00022 | > step_time: 0.39490 (0.39574) | > loader_time: 0.00310 (0.00287)  --> STEP: 140/639 -- GLOBAL_STEP: 82700 | > loss: -0.49927 (-0.54008) | > log_mle: -0.62945 (-0.65011) | > loss_dur: 0.13019 (0.11003) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.20494 (10.65033) | > current_lr: 0.00022 | > step_time: 0.55170 (0.40994) | > loader_time: 0.00320 (0.00290)  --> STEP: 165/639 -- GLOBAL_STEP: 82725 | > loss: -0.51498 (-0.53828) | > log_mle: -0.63809 (-0.64823) | > loss_dur: 0.12311 (0.10995) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.40039 (10.67014) | > current_lr: 0.00022 | > step_time: 0.54190 (0.42461) | > loader_time: 0.00310 (0.00292)  --> STEP: 190/639 -- GLOBAL_STEP: 82750 | > loss: -0.46602 (-0.53690) | > log_mle: -0.57212 (-0.64695) | > loss_dur: 0.10610 (0.11006) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.21389 (10.82750) | > current_lr: 0.00022 | > step_time: 0.52520 (0.44188) | > loader_time: 0.00330 (0.00296)  --> STEP: 215/639 -- GLOBAL_STEP: 82775 | > loss: -0.54094 (-0.53355) | > log_mle: -0.65915 (-0.64399) | > loss_dur: 0.11821 (0.11044) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 19.79514 (10.73795) | > current_lr: 0.00022 | > step_time: 0.53940 (0.45665) | > loader_time: 0.00320 (0.00300)  --> STEP: 240/639 -- GLOBAL_STEP: 82800 | > loss: -0.55959 (-0.53351) | > log_mle: -0.67831 (-0.64410) | > loss_dur: 0.11871 (0.11058) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.72493 (10.80384) | > current_lr: 0.00022 | > step_time: 0.55290 (0.47205) | > loader_time: 0.00330 (0.00304)  --> STEP: 265/639 -- GLOBAL_STEP: 82825 | > loss: -0.56489 (-0.53168) | > log_mle: -0.70385 (-0.64290) | > loss_dur: 0.13896 (0.11122) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 17.71758 (10.89853) | > current_lr: 0.00022 | > step_time: 0.61460 (0.48754) | > loader_time: 0.00330 (0.00308)  --> STEP: 290/639 -- GLOBAL_STEP: 82850 | > loss: -0.49276 (-0.52962) | > log_mle: -0.60160 (-0.64117) | > loss_dur: 0.10883 (0.11154) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.95078 (10.82160) | > current_lr: 0.00022 | > step_time: 0.62270 (0.50242) | > loader_time: 0.00340 (0.00311)  --> STEP: 315/639 -- GLOBAL_STEP: 82875 | > loss: -0.48287 (-0.52755) | > log_mle: -0.58284 (-0.63933) | > loss_dur: 0.09997 (0.11178) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.30802 (10.70870) | > current_lr: 0.00022 | > step_time: 0.64070 (0.51728) | > loader_time: 0.00340 (0.00314)  --> STEP: 340/639 -- GLOBAL_STEP: 82900 | > loss: -0.49028 (-0.52523) | > log_mle: -0.59933 (-0.63761) | > loss_dur: 0.10905 (0.11238) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.84989 (10.76528) | > current_lr: 0.00022 | > step_time: 0.59800 (0.53322) | > loader_time: 0.00330 (0.00317)  --> STEP: 365/639 -- GLOBAL_STEP: 82925 | > loss: -0.49723 (-0.52280) | > log_mle: -0.60758 (-0.63499) | > loss_dur: 0.11035 (0.11218) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.71455 (10.56716) | > current_lr: 0.00022 | > step_time: 0.82090 (0.54488) | > loader_time: 0.00370 (0.00320)  --> STEP: 390/639 -- GLOBAL_STEP: 82950 | > loss: -0.48653 (-0.52076) | > log_mle: -0.60906 (-0.63285) | > loss_dur: 0.12252 (0.11209) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.47850 (10.38445) | > current_lr: 0.00022 | > step_time: 0.73970 (0.55849) | > loader_time: 0.00350 (0.00323)  --> STEP: 415/639 -- GLOBAL_STEP: 82975 | > loss: -0.46133 (-0.51849) | > log_mle: -0.57804 (-0.63091) | > loss_dur: 0.11670 (0.11241) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.71480 (10.37901) | > current_lr: 0.00022 | > step_time: 0.74000 (0.57075) | > loader_time: 0.00350 (0.00325)  --> STEP: 440/639 -- GLOBAL_STEP: 83000 | > loss: -0.45101 (-0.51665) | > log_mle: -0.56584 (-0.62911) | > loss_dur: 0.11483 (0.11246) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.42651 (10.28178) | > current_lr: 0.00022 | > step_time: 0.73820 (0.58285) | > loader_time: 0.00360 (0.00328)  --> STEP: 465/639 -- GLOBAL_STEP: 83025 | > loss: -0.46270 (-0.51453) | > log_mle: -0.57212 (-0.62696) | > loss_dur: 0.10942 (0.11243) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.59326 (10.10041) | > current_lr: 0.00022 | > step_time: 0.76960 (0.59344) | > loader_time: 0.00380 (0.00330)  --> STEP: 490/639 -- GLOBAL_STEP: 83050 | > loss: -0.45583 (-0.51275) | > log_mle: -0.55267 (-0.62530) | > loss_dur: 0.09685 (0.11255) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.86277 (9.96280) | > current_lr: 0.00022 | > step_time: 0.84440 (0.60660) | > loader_time: 0.00360 (0.00333)  --> STEP: 515/639 -- GLOBAL_STEP: 83075 | > loss: -0.48114 (-0.51140) | > log_mle: -0.58414 (-0.62404) | > loss_dur: 0.10300 (0.11264) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.74529 (9.86934) | > current_lr: 0.00022 | > step_time: 0.98040 (0.61941) | > loader_time: 0.00430 (0.00336)  --> STEP: 540/639 -- GLOBAL_STEP: 83100 | > loss: -0.47720 (-0.50970) | > log_mle: -0.59495 (-0.62246) | > loss_dur: 0.11775 (0.11276) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.57343 (9.79297) | > current_lr: 0.00022 | > step_time: 0.88010 (0.63384) | > loader_time: 0.00420 (0.00339)  --> STEP: 565/639 -- GLOBAL_STEP: 83125 | > loss: -0.44855 (-0.50798) | > log_mle: -0.55898 (-0.62088) | > loss_dur: 0.11043 (0.11290) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.31638 (9.69727) | > current_lr: 0.00022 | > step_time: 1.09040 (0.64798) | > loader_time: 0.00420 (0.00342)  --> STEP: 590/639 -- GLOBAL_STEP: 83150 | > loss: -0.47443 (-0.50677) | > log_mle: -0.59036 (-0.61984) | > loss_dur: 0.11593 (0.11307) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.01158 (9.64059) | > current_lr: 0.00022 | > step_time: 1.03290 (0.66395) | > loader_time: 0.00430 (0.00346)  --> STEP: 615/639 -- GLOBAL_STEP: 83175 | > loss: -0.47390 (-0.50533) | > log_mle: -0.59348 (-0.61852) | > loss_dur: 0.11958 (0.11319) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.71498 (9.56438) | > current_lr: 0.00022 | > step_time: 1.00880 (0.67965) | > loader_time: 0.00430 (0.00349) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03502 (+0.00100) | > avg_loss: -0.45443 (+0.00837) | > avg_log_mle: -0.59189 (+0.00662) | > avg_loss_dur: 0.13746 (+0.00175)  > EPOCH: 130/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 07:25:59)   --> STEP: 0/639 -- GLOBAL_STEP: 83200 | > loss: -0.57271 (-0.57271) | > log_mle: -0.72880 (-0.72880) | > loss_dur: 0.15609 (0.15609) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 32.06784 (32.06784) | > current_lr: 0.00022 | > step_time: 0.65490 (0.65494) | > loader_time: 0.38590 (0.38587)  --> STEP: 25/639 -- GLOBAL_STEP: 83225 | > loss: -0.55848 (-0.56814) | > log_mle: -0.68482 (-0.68214) | > loss_dur: 0.12634 (0.11400) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.15274 (12.57669) | > current_lr: 0.00022 | > step_time: 0.38140 (0.32979) | > loader_time: 0.00280 (0.00324)  --> STEP: 50/639 -- GLOBAL_STEP: 83250 | > loss: -0.52558 (-0.55769) | > log_mle: -0.61923 (-0.66760) | > loss_dur: 0.09365 (0.10991) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.10866 (10.69881) | > current_lr: 0.00022 | > step_time: 0.37950 (0.35058) | > loader_time: 0.00270 (0.00298)  --> STEP: 75/639 -- GLOBAL_STEP: 83275 | > loss: -0.53675 (-0.54821) | > log_mle: -0.63233 (-0.65736) | > loss_dur: 0.09558 (0.10916) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.30382 (10.11156) | > current_lr: 0.00022 | > step_time: 0.49800 (0.37233) | > loader_time: 0.00300 (0.00292)  --> STEP: 100/639 -- GLOBAL_STEP: 83300 | > loss: -0.54088 (-0.54527) | > log_mle: -0.64866 (-0.65354) | > loss_dur: 0.10778 (0.10827) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.12910 (9.92887) | > current_lr: 0.00022 | > step_time: 0.50810 (0.38776) | > loader_time: 0.00300 (0.00290)  --> STEP: 125/639 -- GLOBAL_STEP: 83325 | > loss: -0.52110 (-0.54158) | > log_mle: -0.63769 (-0.65077) | > loss_dur: 0.11660 (0.10919) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.50211 (10.00474) | > current_lr: 0.00022 | > step_time: 0.41340 (0.40318) | > loader_time: 0.00260 (0.00291)  --> STEP: 150/639 -- GLOBAL_STEP: 83350 | > loss: -0.48630 (-0.53951) | > log_mle: -0.60302 (-0.64915) | > loss_dur: 0.11672 (0.10964) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.33830 (10.05276) | > current_lr: 0.00022 | > step_time: 0.43140 (0.41758) | > loader_time: 0.00280 (0.00294)  --> STEP: 175/639 -- GLOBAL_STEP: 83375 | > loss: -0.52625 (-0.53854) | > log_mle: -0.62334 (-0.64867) | > loss_dur: 0.09708 (0.11013) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.15658 (10.30294) | > current_lr: 0.00022 | > step_time: 0.55520 (0.43316) | > loader_time: 0.00320 (0.00297)  --> STEP: 200/639 -- GLOBAL_STEP: 83400 | > loss: -0.49380 (-0.53579) | > log_mle: -0.60663 (-0.64570) | > loss_dur: 0.11283 (0.10991) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.83088 (10.16994) | > current_lr: 0.00022 | > step_time: 0.52300 (0.44707) | > loader_time: 0.00310 (0.00300)  --> STEP: 225/639 -- GLOBAL_STEP: 83425 | > loss: -0.48948 (-0.53311) | > log_mle: -0.58753 (-0.64301) | > loss_dur: 0.09805 (0.10991) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.86183 (10.09591) | > current_lr: 0.00022 | > step_time: 0.51210 (0.46314) | > loader_time: 0.00320 (0.00303)  --> STEP: 250/639 -- GLOBAL_STEP: 83450 | > loss: -0.49538 (-0.53325) | > log_mle: -0.58968 (-0.64346) | > loss_dur: 0.09431 (0.11021) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.39621 (10.28437) | > current_lr: 0.00022 | > step_time: 0.58130 (0.47858) | > loader_time: 0.00300 (0.00306)  --> STEP: 275/639 -- GLOBAL_STEP: 83475 | > loss: -0.50901 (-0.53218) | > log_mle: -0.62294 (-0.64327) | > loss_dur: 0.11393 (0.11108) | > amp_scaler: 4096.00000 (2189.49818) | > grad_norm: 12.47245 (10.51615) | > current_lr: 0.00022 | > step_time: 0.87260 (0.49508) | > loader_time: 0.00370 (0.00310)  --> STEP: 300/639 -- GLOBAL_STEP: 83500 | > loss: -0.50113 (-0.52953) | > log_mle: -0.61621 (-0.64091) | > loss_dur: 0.11508 (0.11138) | > amp_scaler: 4096.00000 (2348.37333) | > grad_norm: 10.42360 (10.40604) | > current_lr: 0.00022 | > step_time: 0.68150 (0.50801) | > loader_time: 0.00360 (0.00314)  --> STEP: 325/639 -- GLOBAL_STEP: 83525 | > loss: -0.49358 (-0.52707) | > log_mle: -0.61653 (-0.63899) | > loss_dur: 0.12295 (0.11192) | > amp_scaler: 4096.00000 (2482.80615) | > grad_norm: 8.28403 (10.36810) | > current_lr: 0.00022 | > step_time: 0.65290 (0.52359) | > loader_time: 0.00350 (0.00317)  --> STEP: 350/639 -- GLOBAL_STEP: 83550 | > loss: -0.51600 (-0.52482) | > log_mle: -0.64147 (-0.63677) | > loss_dur: 0.12547 (0.11195) | > amp_scaler: 4096.00000 (2598.03429) | > grad_norm: 11.07806 (10.24070) | > current_lr: 0.00022 | > step_time: 0.84610 (0.53731) | > loader_time: 0.00380 (0.00320)  --> STEP: 375/639 -- GLOBAL_STEP: 83575 | > loss: -0.47228 (-0.52237) | > log_mle: -0.58032 (-0.63417) | > loss_dur: 0.10803 (0.11180) | > amp_scaler: 4096.00000 (2697.89867) | > grad_norm: 5.65261 (10.03193) | > current_lr: 0.00022 | > step_time: 0.69520 (0.54970) | > loader_time: 0.00360 (0.00321)  --> STEP: 400/639 -- GLOBAL_STEP: 83600 | > loss: -0.49575 (-0.52050) | > log_mle: -0.60860 (-0.63252) | > loss_dur: 0.11285 (0.11203) | > amp_scaler: 4096.00000 (2785.28000) | > grad_norm: 8.28881 (9.97778) | > current_lr: 0.00022 | > step_time: 0.77650 (0.56175) | > loader_time: 0.00370 (0.00324)  --> STEP: 425/639 -- GLOBAL_STEP: 83625 | > loss: -0.54334 (-0.51816) | > log_mle: -0.66500 (-0.63031) | > loss_dur: 0.12167 (0.11216) | > amp_scaler: 4096.00000 (2862.38118) | > grad_norm: 18.89466 (9.87829) | > current_lr: 0.00022 | > step_time: 0.90100 (0.57378) | > loader_time: 0.00400 (0.00326)  --> STEP: 450/639 -- GLOBAL_STEP: 83650 | > loss: -0.46290 (-0.51623) | > log_mle: -0.56809 (-0.62833) | > loss_dur: 0.10519 (0.11210) | > amp_scaler: 4096.00000 (2930.91556) | > grad_norm: 4.77489 (9.76683) | > current_lr: 0.00022 | > step_time: 0.80010 (0.58546) | > loader_time: 0.00370 (0.00329)  --> STEP: 475/639 -- GLOBAL_STEP: 83675 | > loss: -0.48053 (-0.51428) | > log_mle: -0.57927 (-0.62650) | > loss_dur: 0.09874 (0.11223) | > amp_scaler: 4096.00000 (2992.23579) | > grad_norm: 5.21969 (9.70190) | > current_lr: 0.00022 | > step_time: 0.76750 (0.59737) | > loader_time: 0.00370 (0.00331)  --> STEP: 500/639 -- GLOBAL_STEP: 83700 | > loss: -0.52368 (-0.51290) | > log_mle: -0.64105 (-0.62518) | > loss_dur: 0.11737 (0.11228) | > amp_scaler: 4096.00000 (3047.42400) | > grad_norm: 11.74800 (9.63047) | > current_lr: 0.00022 | > step_time: 1.03890 (0.61099) | > loader_time: 0.00400 (0.00334)  --> STEP: 525/639 -- GLOBAL_STEP: 83725 | > loss: -0.48707 (-0.51139) | > log_mle: -0.61710 (-0.62379) | > loss_dur: 0.13003 (0.11240) | > amp_scaler: 4096.00000 (3097.35619) | > grad_norm: 12.20716 (9.58226) | > current_lr: 0.00022 | > step_time: 1.04000 (0.62423) | > loader_time: 0.00440 (0.00336)  --> STEP: 550/639 -- GLOBAL_STEP: 83750 | > loss: -0.43688 (-0.50975) | > log_mle: -0.55330 (-0.62218) | > loss_dur: 0.11642 (0.11244) | > amp_scaler: 4096.00000 (3142.74909) | > grad_norm: 4.71262 (9.50970) | > current_lr: 0.00022 | > step_time: 0.93520 (0.63892) | > loader_time: 0.00400 (0.00339)  --> STEP: 575/639 -- GLOBAL_STEP: 83775 | > loss: -0.45513 (-0.50805) | > log_mle: -0.57088 (-0.62064) | > loss_dur: 0.11575 (0.11259) | > amp_scaler: 4096.00000 (3184.19478) | > grad_norm: 6.13381 (9.43456) | > current_lr: 0.00022 | > step_time: 0.94390 (0.65297) | > loader_time: 0.00390 (0.00342)  --> STEP: 600/639 -- GLOBAL_STEP: 83800 | > loss: -0.46083 (-0.50670) | > log_mle: -0.57022 (-0.61948) | > loss_dur: 0.10939 (0.11278) | > amp_scaler: 4096.00000 (3222.18667) | > grad_norm: 5.93635 (9.37162) | > current_lr: 0.00022 | > step_time: 1.04600 (0.66881) | > loader_time: 0.00450 (0.00346)  --> STEP: 625/639 -- GLOBAL_STEP: 83825 | > loss: -0.48018 (-0.50540) | > log_mle: -0.59853 (-0.61832) | > loss_dur: 0.11834 (0.11292) | > amp_scaler: 4096.00000 (3257.13920) | > grad_norm: 8.72896 (9.32536) | > current_lr: 0.00022 | > step_time: 1.04650 (0.68520) | > loader_time: 0.00450 (0.00350) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03391 (-0.00112) | > avg_loss: -0.45288 (+0.00155) | > avg_log_mle: -0.59220 (-0.00032) | > avg_loss_dur: 0.13932 (+0.00187)  > EPOCH: 131/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 07:33:38)   --> STEP: 10/639 -- GLOBAL_STEP: 83850 | > loss: -0.55364 (-0.57287) | > log_mle: -0.68229 (-0.68413) | > loss_dur: 0.12865 (0.11127) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.49573 (12.01373) | > current_lr: 0.00022 | > step_time: 0.28520 (0.28900) | > loader_time: 0.00260 (0.01055)  --> STEP: 35/639 -- GLOBAL_STEP: 83875 | > loss: -0.55917 (-0.56184) | > log_mle: -0.65079 (-0.67163) | > loss_dur: 0.09163 (0.10979) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.02549 (10.78305) | > current_lr: 0.00022 | > step_time: 0.32220 (0.32757) | > loader_time: 0.00270 (0.00496)  --> STEP: 60/639 -- GLOBAL_STEP: 83900 | > loss: -0.51015 (-0.55407) | > log_mle: -0.60977 (-0.66353) | > loss_dur: 0.09962 (0.10946) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.49990 (10.58440) | > current_lr: 0.00022 | > step_time: 0.34410 (0.35552) | > loader_time: 0.00280 (0.00412)  --> STEP: 85/639 -- GLOBAL_STEP: 83925 | > loss: -0.49946 (-0.54532) | > log_mle: -0.60894 (-0.65370) | > loss_dur: 0.10949 (0.10839) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.13751 (9.93212) | > current_lr: 0.00022 | > step_time: 0.41760 (0.36907) | > loader_time: 0.00300 (0.00378)  --> STEP: 110/639 -- GLOBAL_STEP: 83950 | > loss: -0.50598 (-0.54442) | > log_mle: -0.61767 (-0.65413) | > loss_dur: 0.11169 (0.10971) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.93821 (10.56822) | > current_lr: 0.00022 | > step_time: 0.40520 (0.38891) | > loader_time: 0.00290 (0.00361)  --> STEP: 135/639 -- GLOBAL_STEP: 83975 | > loss: -0.48478 (-0.54084) | > log_mle: -0.59698 (-0.65053) | > loss_dur: 0.11220 (0.10970) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 5.68564 (10.41297) | > current_lr: 0.00022 | > step_time: 0.46790 (0.40048) | > loader_time: 0.00310 (0.00350)  --> STEP: 160/639 -- GLOBAL_STEP: 84000 | > loss: -0.53048 (-0.53819) | > log_mle: -0.63480 (-0.64768) | > loss_dur: 0.10432 (0.10950) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.81559 (10.38057) | > current_lr: 0.00022 | > step_time: 0.46400 (0.41450) | > loader_time: 0.00300 (0.00344)  --> STEP: 185/639 -- GLOBAL_STEP: 84025 | > loss: -0.51743 (-0.53860) | > log_mle: -0.61833 (-0.64897) | > loss_dur: 0.10090 (0.11037) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.57031 (11.02332) | > current_lr: 0.00022 | > step_time: 0.58120 (0.43371) | > loader_time: 0.00310 (0.00341)  --> STEP: 210/639 -- GLOBAL_STEP: 84050 | > loss: -0.44543 (-0.53504) | > log_mle: -0.53900 (-0.64536) | > loss_dur: 0.09358 (0.11031) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 1.91016 (10.83387) | > current_lr: 0.00022 | > step_time: 0.50740 (0.44968) | > loader_time: 0.00330 (0.00337)  --> STEP: 235/639 -- GLOBAL_STEP: 84075 | > loss: -0.52202 (-0.53477) | > log_mle: -0.62845 (-0.64516) | > loss_dur: 0.10643 (0.11039) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.10722 (10.92403) | > current_lr: 0.00022 | > step_time: 0.55500 (0.46724) | > loader_time: 0.00360 (0.00337)  --> STEP: 260/639 -- GLOBAL_STEP: 84100 | > loss: -0.51075 (-0.53346) | > log_mle: -0.62738 (-0.64427) | > loss_dur: 0.11663 (0.11081) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.88660 (11.00114) | > current_lr: 0.00022 | > step_time: 0.60450 (0.48202) | > loader_time: 0.00320 (0.00337)  --> STEP: 285/639 -- GLOBAL_STEP: 84125 | > loss: -0.46381 (-0.53181) | > log_mle: -0.57636 (-0.64287) | > loss_dur: 0.11255 (0.11106) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.88998 (10.93855) | > current_lr: 0.00022 | > step_time: 0.73090 (0.49922) | > loader_time: 0.00340 (0.00337)  --> STEP: 310/639 -- GLOBAL_STEP: 84150 | > loss: -0.49586 (-0.52956) | > log_mle: -0.61747 (-0.64100) | > loss_dur: 0.12161 (0.11144) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 10.27627 (10.87352) | > current_lr: 0.00022 | > step_time: 0.56800 (0.51294) | > loader_time: 0.00330 (0.00339)  --> STEP: 335/639 -- GLOBAL_STEP: 84175 | > loss: -0.51465 (-0.52673) | > log_mle: -0.64109 (-0.63873) | > loss_dur: 0.12644 (0.11199) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 16.64065 (10.79057) | > current_lr: 0.00022 | > step_time: 0.67380 (0.52782) | > loader_time: 0.00350 (0.00340)  --> STEP: 360/639 -- GLOBAL_STEP: 84200 | > loss: -0.47372 (-0.52470) | > log_mle: -0.57902 (-0.63652) | > loss_dur: 0.10530 (0.11182) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.51270 (10.72162) | > current_lr: 0.00022 | > step_time: 0.68210 (0.53977) | > loader_time: 0.00370 (0.00341)  --> STEP: 385/639 -- GLOBAL_STEP: 84225 | > loss: -0.50220 (-0.52270) | > log_mle: -0.61184 (-0.63445) | > loss_dur: 0.10963 (0.11175) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.62139 (10.56328) | > current_lr: 0.00022 | > step_time: 0.69780 (0.55270) | > loader_time: 0.00360 (0.00342)  --> STEP: 410/639 -- GLOBAL_STEP: 84250 | > loss: -0.45643 (-0.52046) | > log_mle: -0.58207 (-0.63249) | > loss_dur: 0.12564 (0.11204) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.15915 (10.45548) | > current_lr: 0.00022 | > step_time: 0.82070 (0.56496) | > loader_time: 0.00380 (0.00343)  --> STEP: 435/639 -- GLOBAL_STEP: 84275 | > loss: -0.49071 (-0.51822) | > log_mle: -0.59593 (-0.63029) | > loss_dur: 0.10521 (0.11207) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 6.68689 (10.31591) | > current_lr: 0.00022 | > step_time: 0.71570 (0.57791) | > loader_time: 0.00360 (0.00344)  --> STEP: 460/639 -- GLOBAL_STEP: 84300 | > loss: -0.44733 (-0.51607) | > log_mle: -0.56228 (-0.62816) | > loss_dur: 0.11494 (0.11209) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 3.27807 (10.18618) | > current_lr: 0.00022 | > step_time: 0.89380 (0.58959) | > loader_time: 0.00400 (0.00346)  --> STEP: 485/639 -- GLOBAL_STEP: 84325 | > loss: -0.48541 (-0.51427) | > log_mle: -0.59341 (-0.62659) | > loss_dur: 0.10800 (0.11232) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.24476 (10.11280) | > current_lr: 0.00022 | > step_time: 0.83660 (0.60239) | > loader_time: 0.00410 (0.00348)  --> STEP: 510/639 -- GLOBAL_STEP: 84350 | > loss: -0.47223 (-0.51281) | > log_mle: -0.59442 (-0.62513) | > loss_dur: 0.12219 (0.11231) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.76546 (10.03489) | > current_lr: 0.00022 | > step_time: 0.83820 (0.61554) | > loader_time: 0.00410 (0.00350)  --> STEP: 535/639 -- GLOBAL_STEP: 84375 | > loss: -0.46251 (-0.51112) | > log_mle: -0.58357 (-0.62356) | > loss_dur: 0.12106 (0.11244) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.47243 (9.92767) | > current_lr: 0.00022 | > step_time: 0.89830 (0.62854) | > loader_time: 0.00390 (0.00353)  --> STEP: 560/639 -- GLOBAL_STEP: 84400 | > loss: -0.47287 (-0.50944) | > log_mle: -0.59214 (-0.62206) | > loss_dur: 0.11926 (0.11262) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 8.17963 (9.83092) | > current_lr: 0.00022 | > step_time: 0.91040 (0.64340) | > loader_time: 0.00390 (0.00356)  --> STEP: 585/639 -- GLOBAL_STEP: 84425 | > loss: -0.48511 (-0.50806) | > log_mle: -0.59601 (-0.62073) | > loss_dur: 0.11090 (0.11268) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 7.13681 (9.73719) | > current_lr: 0.00022 | > step_time: 1.03510 (0.65898) | > loader_time: 0.00430 (0.00359)  --> STEP: 610/639 -- GLOBAL_STEP: 84450 | > loss: -0.46003 (-0.50661) | > log_mle: -0.56696 (-0.61949) | > loss_dur: 0.10693 (0.11288) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 4.82550 (9.65844) | > current_lr: 0.00022 | > step_time: 1.00060 (0.67512) | > loader_time: 0.00400 (0.00362)  --> STEP: 635/639 -- GLOBAL_STEP: 84475 | > loss: -0.46960 (-0.50539) | > log_mle: -0.58874 (-0.61858) | > loss_dur: 0.11913 (0.11319) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 9.66231 (9.66875) | > current_lr: 0.00022 | > step_time: 1.14240 (0.69211) | > loader_time: 0.00430 (0.00365) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03832 (+0.00441) | > avg_loss: -0.46193 (-0.00905) | > avg_log_mle: -0.59835 (-0.00615) | > avg_loss_dur: 0.13642 (-0.00290)  > EPOCH: 132/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 07:41:16)   --> STEP: 20/639 -- GLOBAL_STEP: 84500 | > loss: -0.64828 (-0.57632) | > log_mle: -0.79559 (-0.69256) | > loss_dur: 0.14731 (0.11624) | > amp_scaler: 4096.00000 (4096.00000) | > grad_norm: 21.82253 (14.91297) | > current_lr: 0.00022 | > step_time: 0.33870 (0.31260) | > loader_time: 0.00260 (0.00544)  --> STEP: 45/639 -- GLOBAL_STEP: 84525 | > loss: -0.52908 (-0.55684) | > log_mle: -0.61843 (-0.66926) | > loss_dur: 0.08935 (0.11241) | > amp_scaler: 2048.00000 (3686.40000) | > grad_norm: 3.65169 (11.48654) | > current_lr: 0.00022 | > step_time: 0.41990 (0.34318) | > loader_time: 0.00270 (0.00401)  --> STEP: 70/639 -- GLOBAL_STEP: 84550 | > loss: -0.54774 (-0.54810) | > log_mle: -0.66144 (-0.65881) | > loss_dur: 0.11370 (0.11070) | > amp_scaler: 2048.00000 (3101.25714) | > grad_norm: 10.57873 (11.20969) | > current_lr: 0.00022 | > step_time: 0.46830 (0.36802) | > loader_time: 0.00300 (0.00364)  --> STEP: 95/639 -- GLOBAL_STEP: 84575 | > loss: -0.61122 (-0.54415) | > log_mle: -0.71336 (-0.65299) | > loss_dur: 0.10214 (0.10884) | > amp_scaler: 2048.00000 (2824.08421) | > grad_norm: 13.72526 (10.79489) | > current_lr: 0.00022 | > step_time: 0.40230 (0.38336) | > loader_time: 0.00290 (0.00346)  --> STEP: 120/639 -- GLOBAL_STEP: 84600 | > loss: -0.56669 (-0.54097) | > log_mle: -0.69758 (-0.65086) | > loss_dur: 0.13089 (0.10989) | > amp_scaler: 2048.00000 (2662.40000) | > grad_norm: 27.38469 (11.01260) | > current_lr: 0.00022 | > step_time: 0.64160 (0.39988) | > loader_time: 0.00330 (0.00337)  --> STEP: 145/639 -- GLOBAL_STEP: 84625 | > loss: -0.53023 (-0.53954) | > log_mle: -0.65246 (-0.65032) | > loss_dur: 0.12224 (0.11078) | > amp_scaler: 2048.00000 (2556.46897) | > grad_norm: 11.95200 (11.06323) | > current_lr: 0.00022 | > step_time: 0.46460 (0.41355) | > loader_time: 0.00310 (0.00334)  --> STEP: 170/639 -- GLOBAL_STEP: 84650 | > loss: -0.52668 (-0.53797) | > log_mle: -0.63666 (-0.64854) | > loss_dur: 0.10998 (0.11057) | > amp_scaler: 2048.00000 (2481.69412) | > grad_norm: 8.45131 (11.06609) | > current_lr: 0.00022 | > step_time: 0.53920 (0.42941) | > loader_time: 0.00330 (0.00331)  --> STEP: 195/639 -- GLOBAL_STEP: 84675 | > loss: -0.49455 (-0.53643) | > log_mle: -0.58511 (-0.64686) | > loss_dur: 0.09055 (0.11044) | > amp_scaler: 2048.00000 (2426.09231) | > grad_norm: 2.73931 (11.14650) | > current_lr: 0.00022 | > step_time: 0.50260 (0.44492) | > loader_time: 0.00330 (0.00331)  --> STEP: 220/639 -- GLOBAL_STEP: 84700 | > loss: -0.53108 (-0.53355) | > log_mle: -0.63040 (-0.64404) | > loss_dur: 0.09933 (0.11049) | > amp_scaler: 2048.00000 (2383.12727) | > grad_norm: 9.28456 (11.03384) | > current_lr: 0.00022 | > step_time: 0.62950 (0.46174) | > loader_time: 0.00340 (0.00330)  --> STEP: 245/639 -- GLOBAL_STEP: 84725 | > loss: -0.51811 (-0.53390) | > log_mle: -0.62046 (-0.64445) | > loss_dur: 0.10234 (0.11055) | > amp_scaler: 2048.00000 (2348.93061) | > grad_norm: 7.38512 (11.24284) | > current_lr: 0.00022 | > step_time: 0.52810 (0.47802) | > loader_time: 0.00300 (0.00331)  --> STEP: 270/639 -- GLOBAL_STEP: 84750 | > loss: -0.56972 (-0.53297) | > log_mle: -0.68930 (-0.64404) | > loss_dur: 0.11958 (0.11107) | > amp_scaler: 2048.00000 (2321.06667) | > grad_norm: 26.50223 (11.49057) | > current_lr: 0.00022 | > step_time: 0.86240 (0.49475) | > loader_time: 0.00380 (0.00332)  --> STEP: 295/639 -- GLOBAL_STEP: 84775 | > loss: -0.48450 (-0.53047) | > log_mle: -0.57939 (-0.64177) | > loss_dur: 0.09489 (0.11131) | > amp_scaler: 2048.00000 (2297.92542) | > grad_norm: 4.69152 (11.30089) | > current_lr: 0.00022 | > step_time: 0.69860 (0.50977) | > loader_time: 0.00370 (0.00334)  --> STEP: 320/639 -- GLOBAL_STEP: 84800 | > loss: -0.49013 (-0.52836) | > log_mle: -0.61900 (-0.64012) | > loss_dur: 0.12887 (0.11176) | > amp_scaler: 2048.00000 (2278.40000) | > grad_norm: 14.82862 (11.21734) | > current_lr: 0.00022 | > step_time: 0.81000 (0.52614) | > loader_time: 0.00350 (0.00336)  --> STEP: 345/639 -- GLOBAL_STEP: 84825 | > loss: -0.50075 (-0.52597) | > log_mle: -0.60659 (-0.63810) | > loss_dur: 0.10584 (0.11213) | > amp_scaler: 2048.00000 (2261.70435) | > grad_norm: 11.61651 (11.16483) | > current_lr: 0.00022 | > step_time: 0.63970 (0.54116) | > loader_time: 0.00350 (0.00338)  --> STEP: 370/639 -- GLOBAL_STEP: 84850 | > loss: -0.51086 (-0.52380) | > log_mle: -0.62119 (-0.63575) | > loss_dur: 0.11033 (0.11194) | > amp_scaler: 2048.00000 (2247.26486) | > grad_norm: 11.40335 (10.97656) | > current_lr: 0.00022 | > step_time: 0.86770 (0.55289) | > loader_time: 0.00390 (0.00339)  --> STEP: 395/639 -- GLOBAL_STEP: 84875 | > loss: -0.48067 (-0.52173) | > log_mle: -0.59152 (-0.63363) | > loss_dur: 0.11085 (0.11190) | > amp_scaler: 2048.00000 (2234.65316) | > grad_norm: 7.39036 (10.79660) | > current_lr: 0.00022 | > step_time: 0.72640 (0.56581) | > loader_time: 0.00340 (0.00340)  --> STEP: 420/639 -- GLOBAL_STEP: 84900 | > loss: -0.47026 (-0.51937) | > log_mle: -0.58476 (-0.63157) | > loss_dur: 0.11450 (0.11220) | > amp_scaler: 2048.00000 (2223.54286) | > grad_norm: 6.16493 (10.67179) | > current_lr: 0.00022 | > step_time: 0.70130 (0.57853) | > loader_time: 0.00350 (0.00342)  --> STEP: 445/639 -- GLOBAL_STEP: 84925 | > loss: -0.48417 (-0.51770) | > log_mle: -0.59210 (-0.62991) | > loss_dur: 0.10793 (0.11220) | > amp_scaler: 2048.00000 (2213.68090) | > grad_norm: 6.99626 (10.55033) | > current_lr: 0.00022 | > step_time: 0.76060 (0.59011) | > loader_time: 0.00340 (0.00344)  --> STEP: 470/639 -- GLOBAL_STEP: 84950 | > loss: -0.48781 (-0.51569) | > log_mle: -0.60173 (-0.62785) | > loss_dur: 0.11392 (0.11216) | > amp_scaler: 2048.00000 (2204.86809) | > grad_norm: 9.67802 (10.43657) | > current_lr: 0.00022 | > step_time: 0.79140 (0.60098) | > loader_time: 0.00380 (0.00345)  --> STEP: 495/639 -- GLOBAL_STEP: 84975 | > loss: -0.44612 (-0.51395) | > log_mle: -0.54235 (-0.62623) | > loss_dur: 0.09623 (0.11228) | > amp_scaler: 2048.00000 (2196.94545) | > grad_norm: 2.76969 (10.32568) | > current_lr: 0.00022 | > step_time: 0.77320 (0.61376) | > loader_time: 0.00370 (0.00347)  --> STEP: 520/639 -- GLOBAL_STEP: 85000 | > loss: -0.48410 (-0.51260) | > log_mle: -0.59443 (-0.62490) | > loss_dur: 0.11032 (0.11231) | > amp_scaler: 2048.00000 (2189.78462) | > grad_norm: 9.83523 (10.25196) | > current_lr: 0.00022 | > step_time: 0.80470 (0.62645) | > loader_time: 0.00380 (0.00348)  --> STEP: 545/639 -- GLOBAL_STEP: 85025 | > loss: -0.50160 (-0.51099) | > log_mle: -0.62679 (-0.62350) | > loss_dur: 0.12519 (0.11251) | > amp_scaler: 2048.00000 (2183.28073) | > grad_norm: 10.46206 (10.16360) | > current_lr: 0.00022 | > step_time: 0.94840 (0.64145) | > loader_time: 0.00400 (0.00351)  --> STEP: 570/639 -- GLOBAL_STEP: 85050 | > loss: -0.47257 (-0.50924) | > log_mle: -0.57533 (-0.62187) | > loss_dur: 0.10276 (0.11263) | > amp_scaler: 2048.00000 (2177.34737) | > grad_norm: 5.38676 (10.04733) | > current_lr: 0.00022 | > step_time: 0.83260 (0.65515) | > loader_time: 0.00390 (0.00354)  --> STEP: 595/639 -- GLOBAL_STEP: 85075 | > loss: -0.45757 (-0.50791) | > log_mle: -0.56798 (-0.62072) | > loss_dur: 0.11042 (0.11281) | > amp_scaler: 2048.00000 (2171.91261) | > grad_norm: 5.92333 (9.98298) | > current_lr: 0.00022 | > step_time: 0.97070 (0.67109) | > loader_time: 0.00420 (0.00357)  --> STEP: 620/639 -- GLOBAL_STEP: 85100 | > loss: -0.46543 (-0.50645) | > log_mle: -0.57463 (-0.61943) | > loss_dur: 0.10919 (0.11298) | > amp_scaler: 2048.00000 (2166.91613) | > grad_norm: 6.80528 (9.90404) | > current_lr: 0.00022 | > step_time: 1.16920 (0.68678) | > loader_time: 0.00460 (0.00361) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03111 (-0.00721) | > avg_loss: -0.46303 (-0.00109) | > avg_log_mle: -0.59895 (-0.00059) | > avg_loss_dur: 0.13592 (-0.00050)  > EPOCH: 133/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 07:48:57)   --> STEP: 5/639 -- GLOBAL_STEP: 85125 | > loss: -0.57887 (-0.58376) | > log_mle: -0.71123 (-0.69735) | > loss_dur: 0.13236 (0.11359) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 23.35963 (16.58217) | > current_lr: 0.00022 | > step_time: 0.29900 (0.27588) | > loader_time: 0.00280 (0.00270)  --> STEP: 30/639 -- GLOBAL_STEP: 85150 | > loss: -0.56518 (-0.56818) | > log_mle: -0.67031 (-0.67859) | > loss_dur: 0.10513 (0.11041) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 16.22986 (12.39412) | > current_lr: 0.00022 | > step_time: 0.32930 (0.32269) | > loader_time: 0.00280 (0.00266)  --> STEP: 55/639 -- GLOBAL_STEP: 85175 | > loss: -0.52618 (-0.55751) | > log_mle: -0.61392 (-0.66686) | > loss_dur: 0.08775 (0.10935) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.88248 (10.97517) | > current_lr: 0.00022 | > step_time: 0.36400 (0.34241) | > loader_time: 0.00290 (0.00270)  --> STEP: 80/639 -- GLOBAL_STEP: 85200 | > loss: -0.50982 (-0.54883) | > log_mle: -0.64113 (-0.65716) | > loss_dur: 0.13131 (0.10833) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.76522 (10.49883) | > current_lr: 0.00022 | > step_time: 0.39270 (0.36202) | > loader_time: 0.00290 (0.00275)  --> STEP: 105/639 -- GLOBAL_STEP: 85225 | > loss: -0.47592 (-0.54565) | > log_mle: -0.59308 (-0.65371) | > loss_dur: 0.11716 (0.10807) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.69026 (10.57216) | > current_lr: 0.00022 | > step_time: 0.41530 (0.38166) | > loader_time: 0.00290 (0.00281)  --> STEP: 130/639 -- GLOBAL_STEP: 85250 | > loss: -0.51288 (-0.54225) | > log_mle: -0.62980 (-0.65122) | > loss_dur: 0.11692 (0.10896) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.17836 (10.66152) | > current_lr: 0.00022 | > step_time: 0.57910 (0.39772) | > loader_time: 0.00280 (0.00283)  --> STEP: 155/639 -- GLOBAL_STEP: 85275 | > loss: -0.50753 (-0.54019) | > log_mle: -0.61983 (-0.64959) | > loss_dur: 0.11230 (0.10940) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.73102 (10.75168) | > current_lr: 0.00022 | > step_time: 0.57160 (0.41454) | > loader_time: 0.00320 (0.00286)  --> STEP: 180/639 -- GLOBAL_STEP: 85300 | > loss: -0.54854 (-0.54026) | > log_mle: -0.67000 (-0.65025) | > loss_dur: 0.12146 (0.10999) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 14.88698 (11.02027) | > current_lr: 0.00022 | > step_time: 0.65820 (0.43294) | > loader_time: 0.00300 (0.00288)  --> STEP: 205/639 -- GLOBAL_STEP: 85325 | > loss: -0.48690 (-0.53628) | > log_mle: -0.59675 (-0.64604) | > loss_dur: 0.10985 (0.10976) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.84492 (10.70800) | > current_lr: 0.00022 | > step_time: 0.56300 (0.44717) | > loader_time: 0.00320 (0.00292)  --> STEP: 230/639 -- GLOBAL_STEP: 85350 | > loss: -0.52279 (-0.53530) | > log_mle: -0.63139 (-0.64543) | > loss_dur: 0.10859 (0.11013) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.40746 (10.84302) | > current_lr: 0.00022 | > step_time: 0.59220 (0.46509) | > loader_time: 0.00310 (0.00296)  --> STEP: 255/639 -- GLOBAL_STEP: 85375 | > loss: -0.48725 (-0.53382) | > log_mle: -0.60168 (-0.64393) | > loss_dur: 0.11443 (0.11012) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.21179 (10.90310) | > current_lr: 0.00022 | > step_time: 0.62800 (0.47830) | > loader_time: 0.00340 (0.00299)  --> STEP: 280/639 -- GLOBAL_STEP: 85400 | > loss: -0.49366 (-0.53295) | > log_mle: -0.62150 (-0.64387) | > loss_dur: 0.12784 (0.11091) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 16.65347 (11.03524) | > current_lr: 0.00022 | > step_time: 0.59800 (0.49574) | > loader_time: 0.00330 (0.00303)  --> STEP: 305/639 -- GLOBAL_STEP: 85425 | > loss: -0.48701 (-0.53058) | > log_mle: -0.60451 (-0.64184) | > loss_dur: 0.11751 (0.11126) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.78076 (10.90781) | > current_lr: 0.00022 | > step_time: 0.66560 (0.51102) | > loader_time: 0.00340 (0.00306)  --> STEP: 330/639 -- GLOBAL_STEP: 85450 | > loss: -0.50232 (-0.52760) | > log_mle: -0.62527 (-0.63950) | > loss_dur: 0.12295 (0.11190) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.89360 (10.85265) | > current_lr: 0.00022 | > step_time: 0.72550 (0.52618) | > loader_time: 0.00350 (0.00309)  --> STEP: 355/639 -- GLOBAL_STEP: 85475 | > loss: -0.46011 (-0.52540) | > log_mle: -0.55861 (-0.63725) | > loss_dur: 0.09850 (0.11185) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.98007 (10.81117) | > current_lr: 0.00022 | > step_time: 0.60030 (0.53951) | > loader_time: 0.00350 (0.00312)  --> STEP: 380/639 -- GLOBAL_STEP: 85500 | > loss: -0.49438 (-0.52326) | > log_mle: -0.59428 (-0.63486) | > loss_dur: 0.09990 (0.11159) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.74154 (10.62000) | > current_lr: 0.00022 | > step_time: 0.62980 (0.55317) | > loader_time: 0.00350 (0.00315)  --> STEP: 405/639 -- GLOBAL_STEP: 85525 | > loss: -0.48820 (-0.52152) | > log_mle: -0.59522 (-0.63343) | > loss_dur: 0.10701 (0.11190) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.26990 (10.59870) | > current_lr: 0.00022 | > step_time: 0.69840 (0.56484) | > loader_time: 0.00340 (0.00318)  --> STEP: 430/639 -- GLOBAL_STEP: 85550 | > loss: -0.45373 (-0.51928) | > log_mle: -0.55714 (-0.63133) | > loss_dur: 0.10340 (0.11206) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.21354 (10.46913) | > current_lr: 0.00022 | > step_time: 0.76720 (0.57776) | > loader_time: 0.00370 (0.00321)  --> STEP: 455/639 -- GLOBAL_STEP: 85575 | > loss: -0.47217 (-0.51717) | > log_mle: -0.58226 (-0.62911) | > loss_dur: 0.11008 (0.11194) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.48528 (10.34750) | > current_lr: 0.00022 | > step_time: 0.75010 (0.58873) | > loader_time: 0.00370 (0.00324)  --> STEP: 480/639 -- GLOBAL_STEP: 85600 | > loss: -0.42930 (-0.51521) | > log_mle: -0.54025 (-0.62740) | > loss_dur: 0.11095 (0.11219) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.74989 (10.29947) | > current_lr: 0.00022 | > step_time: 0.80110 (0.60124) | > loader_time: 0.00370 (0.00327)  --> STEP: 505/639 -- GLOBAL_STEP: 85625 | > loss: -0.49608 (-0.51369) | > log_mle: -0.60257 (-0.62586) | > loss_dur: 0.10649 (0.11217) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.44350 (10.23535) | > current_lr: 0.00022 | > step_time: 0.84590 (0.61365) | > loader_time: 0.00420 (0.00330)  --> STEP: 530/639 -- GLOBAL_STEP: 85650 | > loss: -0.47074 (-0.51211) | > log_mle: -0.59421 (-0.62446) | > loss_dur: 0.12347 (0.11234) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.98067 (10.20705) | > current_lr: 0.00022 | > step_time: 0.94610 (0.62707) | > loader_time: 0.00380 (0.00333)  --> STEP: 555/639 -- GLOBAL_STEP: 85675 | > loss: -0.46477 (-0.51044) | > log_mle: -0.57716 (-0.62281) | > loss_dur: 0.11240 (0.11237) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.84180 (10.11309) | > current_lr: 0.00022 | > step_time: 0.84310 (0.64090) | > loader_time: 0.00400 (0.00337)  --> STEP: 580/639 -- GLOBAL_STEP: 85700 | > loss: -0.49126 (-0.50899) | > log_mle: -0.59729 (-0.62151) | > loss_dur: 0.10603 (0.11252) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.60278 (10.03700) | > current_lr: 0.00022 | > step_time: 0.98370 (0.65661) | > loader_time: 0.00420 (0.00342)  --> STEP: 605/639 -- GLOBAL_STEP: 85725 | > loss: -0.46348 (-0.50761) | > log_mle: -0.56656 (-0.62031) | > loss_dur: 0.10307 (0.11270) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.18055 (9.97937) | > current_lr: 0.00022 | > step_time: 1.02960 (0.67278) | > loader_time: 0.00420 (0.00347)  --> STEP: 630/639 -- GLOBAL_STEP: 85750 | > loss: -0.46654 (-0.50637) | > log_mle: -0.59429 (-0.61927) | > loss_dur: 0.12775 (0.11290) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.24645 (9.94421) | > current_lr: 0.00022 | > step_time: 1.21490 (0.68979) | > loader_time: 0.00440 (0.00354) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.02934 (-0.00177) | > avg_loss: -0.45650 (+0.00653) | > avg_log_mle: -0.59512 (+0.00383) | > avg_loss_dur: 0.13862 (+0.00270)  > EPOCH: 134/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 07:56:36)   --> STEP: 15/639 -- GLOBAL_STEP: 85775 | > loss: -0.66999 (-0.57555) | > log_mle: -0.80479 (-0.68870) | > loss_dur: 0.13481 (0.11316) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 23.97714 (12.29051) | > current_lr: 0.00022 | > step_time: 0.32730 (0.29388) | > loader_time: 0.00270 (0.00209)  --> STEP: 40/639 -- GLOBAL_STEP: 85800 | > loss: -0.52702 (-0.56239) | > log_mle: -0.64083 (-0.67191) | > loss_dur: 0.11382 (0.10952) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.52793 (11.48960) | > current_lr: 0.00022 | > step_time: 0.34880 (0.33073) | > loader_time: 0.00280 (0.00248)  --> STEP: 65/639 -- GLOBAL_STEP: 85825 | > loss: -0.54086 (-0.55310) | > log_mle: -0.63112 (-0.66337) | > loss_dur: 0.09027 (0.11027) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.00719 (11.39087) | > current_lr: 0.00022 | > step_time: 0.39810 (0.36233) | > loader_time: 0.00280 (0.00264)  --> STEP: 90/639 -- GLOBAL_STEP: 85850 | > loss: -0.48789 (-0.54477) | > log_mle: -0.58999 (-0.65318) | > loss_dur: 0.10210 (0.10840) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.94598 (10.70848) | > current_lr: 0.00022 | > step_time: 0.52370 (0.37781) | > loader_time: 0.00310 (0.00271)  --> STEP: 115/639 -- GLOBAL_STEP: 85875 | > loss: -0.45770 (-0.54328) | > log_mle: -0.56011 (-0.65234) | > loss_dur: 0.10240 (0.10906) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.73624 (11.01175) | > current_lr: 0.00022 | > step_time: 0.38410 (0.39461) | > loader_time: 0.00300 (0.00276)  --> STEP: 140/639 -- GLOBAL_STEP: 85900 | > loss: -0.49875 (-0.54138) | > log_mle: -0.63473 (-0.65151) | > loss_dur: 0.13598 (0.11013) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.83219 (11.14610) | > current_lr: 0.00022 | > step_time: 0.54970 (0.40845) | > loader_time: 0.00330 (0.00281)  --> STEP: 165/639 -- GLOBAL_STEP: 85925 | > loss: -0.50750 (-0.53921) | > log_mle: -0.64096 (-0.64952) | > loss_dur: 0.13346 (0.11031) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 11.81395 (11.18787) | > current_lr: 0.00022 | > step_time: 0.48570 (0.42279) | > loader_time: 0.00310 (0.00286)  --> STEP: 190/639 -- GLOBAL_STEP: 85950 | > loss: -0.46967 (-0.53814) | > log_mle: -0.57222 (-0.64841) | > loss_dur: 0.10255 (0.11027) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.19769 (11.25999) | > current_lr: 0.00022 | > step_time: 0.52270 (0.44122) | > loader_time: 0.00320 (0.00290)  --> STEP: 215/639 -- GLOBAL_STEP: 85975 | > loss: -0.55656 (-0.53507) | > log_mle: -0.67193 (-0.64559) | > loss_dur: 0.11537 (0.11052) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.20261 (11.08871) | > current_lr: 0.00022 | > step_time: 0.53900 (0.45671) | > loader_time: 0.00330 (0.00295)  --> STEP: 240/639 -- GLOBAL_STEP: 86000 | > loss: -0.56609 (-0.53521) | > log_mle: -0.68099 (-0.64575) | > loss_dur: 0.11489 (0.11054) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 15.51887 (11.14896) | > current_lr: 0.00022 | > step_time: 0.57210 (0.47220) | > loader_time: 0.00330 (0.00299)  --> STEP: 265/639 -- GLOBAL_STEP: 86025 | > loss: -0.57493 (-0.53352) | > log_mle: -0.70265 (-0.64447) | > loss_dur: 0.12772 (0.11096) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 18.12978 (11.23411) | > current_lr: 0.00022 | > step_time: 0.63110 (0.48718) | > loader_time: 0.00350 (0.00303)  --> STEP: 290/639 -- GLOBAL_STEP: 86050 | > loss: -0.49341 (-0.53161) | > log_mle: -0.59872 (-0.64270) | > loss_dur: 0.10531 (0.11108) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.95178 (11.08891) | > current_lr: 0.00022 | > step_time: 0.61890 (0.50290) | > loader_time: 0.00340 (0.00307)  --> STEP: 315/639 -- GLOBAL_STEP: 86075 | > loss: -0.48426 (-0.52948) | > log_mle: -0.58447 (-0.64082) | > loss_dur: 0.10021 (0.11134) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 4.90051 (10.99062) | > current_lr: 0.00022 | > step_time: 0.63120 (0.51746) | > loader_time: 0.00340 (0.00311)  --> STEP: 340/639 -- GLOBAL_STEP: 86100 | > loss: -0.48528 (-0.52709) | > log_mle: -0.59800 (-0.63905) | > loss_dur: 0.11272 (0.11196) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.95326 (11.02612) | > current_lr: 0.00022 | > step_time: 0.63770 (0.53293) | > loader_time: 0.00320 (0.00315)  --> STEP: 365/639 -- GLOBAL_STEP: 86125 | > loss: -0.50009 (-0.52478) | > log_mle: -0.61020 (-0.63641) | > loss_dur: 0.11011 (0.11163) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.55695 (10.78936) | > current_lr: 0.00022 | > step_time: 0.83590 (0.54464) | > loader_time: 0.00390 (0.00317)  --> STEP: 390/639 -- GLOBAL_STEP: 86150 | > loss: -0.48596 (-0.52264) | > log_mle: -0.60857 (-0.63422) | > loss_dur: 0.12262 (0.11159) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.79843 (10.58028) | > current_lr: 0.00022 | > step_time: 0.73330 (0.55736) | > loader_time: 0.00360 (0.00320)  --> STEP: 415/639 -- GLOBAL_STEP: 86175 | > loss: -0.45808 (-0.52036) | > log_mle: -0.57988 (-0.63232) | > loss_dur: 0.12181 (0.11196) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.50678 (10.47445) | > current_lr: 0.00022 | > step_time: 0.73860 (0.56934) | > loader_time: 0.00340 (0.00323)  --> STEP: 440/639 -- GLOBAL_STEP: 86200 | > loss: -0.45006 (-0.51849) | > log_mle: -0.56527 (-0.63054) | > loss_dur: 0.11522 (0.11205) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.18738 (10.35591) | > current_lr: 0.00022 | > step_time: 0.71340 (0.58150) | > loader_time: 0.00370 (0.00326)  --> STEP: 465/639 -- GLOBAL_STEP: 86225 | > loss: -0.46123 (-0.51637) | > log_mle: -0.57071 (-0.62839) | > loss_dur: 0.10948 (0.11202) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.35184 (10.22304) | > current_lr: 0.00022 | > step_time: 0.80380 (0.59285) | > loader_time: 0.00380 (0.00329)  --> STEP: 490/639 -- GLOBAL_STEP: 86250 | > loss: -0.45998 (-0.51453) | > log_mle: -0.55500 (-0.62674) | > loss_dur: 0.09502 (0.11220) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 2.80534 (10.12537) | > current_lr: 0.00022 | > step_time: 0.80290 (0.60601) | > loader_time: 0.00370 (0.00331)  --> STEP: 515/639 -- GLOBAL_STEP: 86275 | > loss: -0.47644 (-0.51322) | > log_mle: -0.58392 (-0.62546) | > loss_dur: 0.10749 (0.11224) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.87086 (10.04160) | > current_lr: 0.00022 | > step_time: 0.98550 (0.61901) | > loader_time: 0.00400 (0.00334)  --> STEP: 540/639 -- GLOBAL_STEP: 86300 | > loss: -0.48348 (-0.51149) | > log_mle: -0.59787 (-0.62388) | > loss_dur: 0.11439 (0.11238) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 8.56743 (9.92608) | > current_lr: 0.00022 | > step_time: 0.88480 (0.63364) | > loader_time: 0.00390 (0.00338)  --> STEP: 565/639 -- GLOBAL_STEP: 86325 | > loss: -0.45275 (-0.50973) | > log_mle: -0.56072 (-0.62227) | > loss_dur: 0.10796 (0.11254) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 3.28110 (9.80021) | > current_lr: 0.00022 | > step_time: 0.95800 (0.64781) | > loader_time: 0.00420 (0.00341)  --> STEP: 590/639 -- GLOBAL_STEP: 86350 | > loss: -0.47712 (-0.50855) | > log_mle: -0.58941 (-0.62121) | > loss_dur: 0.11230 (0.11265) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 10.57236 (9.73342) | > current_lr: 0.00022 | > step_time: 0.96230 (0.66420) | > loader_time: 0.00420 (0.00344)  --> STEP: 615/639 -- GLOBAL_STEP: 86375 | > loss: -0.47688 (-0.50711) | > log_mle: -0.59369 (-0.61989) | > loss_dur: 0.11681 (0.11277) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 7.07955 (9.67228) | > current_lr: 0.00022 | > step_time: 1.02780 (0.67942) | > loader_time: 0.00440 (0.00348) > DataLoader initialization | > Use phonemes: False | > Number of instances : 206 | > Max length sequence: 157 | > Min length sequence: 15 | > Avg length sequence: 79.52427184466019 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > EVALUATION  warning: audio amplitude out of range, auto clipped. | > Synthesizing test sentences. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. warning: audio amplitude out of range, auto clipped. --> EVAL PERFORMANCE | > avg_loader_time: 0.03006 (+0.00072) | > avg_loss: -0.45825 (-0.00175) | > avg_log_mle: -0.59541 (-0.00029) | > avg_loss_dur: 0.13716 (-0.00147)  > EPOCH: 135/1000 --> /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481 > DataLoader initialization | > Use phonemes: False | > Number of instances : 20457 | > Max length sequence: 207 | > Min length sequence: 9 | > Avg length sequence: 79.46262892897296 | > Num. instances discarded by max-min (max=500, min=3) seq limits: 0 | > Batch group size: 0.  > TRAINING (2022-01-24 08:04:15)   --> STEP: 0/639 -- GLOBAL_STEP: 86400 | > loss: -0.59874 (-0.59874) | > log_mle: -0.73775 (-0.73775) | > loss_dur: 0.13901 (0.13901) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 22.17482 (22.17482) | > current_lr: 0.00022 | > step_time: 0.63810 (0.63815) | > loader_time: 0.45740 (0.45738)  --> STEP: 25/639 -- GLOBAL_STEP: 86425 | > loss: -0.56357 (-0.56917) | > log_mle: -0.68434 (-0.68432) | > loss_dur: 0.12077 (0.11515) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 13.80752 (11.69090) | > current_lr: 0.00022 | > step_time: 0.38640 (0.33296) | > loader_time: 0.00300 (0.00268)  --> STEP: 50/639 -- GLOBAL_STEP: 86450 | > loss: -0.52145 (-0.55818) | > log_mle: -0.61631 (-0.66923) | > loss_dur: 0.09486 (0.11105) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 5.00368 (10.62841) | > current_lr: 0.00022 | > step_time: 0.35250 (0.35764) | > loader_time: 0.00270 (0.00298)  --> STEP: 75/639 -- GLOBAL_STEP: 86475 | > loss: -0.53105 (-0.54854) | > log_mle: -0.63774 (-0.65906) | > loss_dur: 0.10669 (0.11052) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 6.88154 (10.19581) | > current_lr: 0.00022 | > step_time: 0.41370 (0.37546) | > loader_time: 0.00270 (0.00297)  --> STEP: 100/639 -- GLOBAL_STEP: 86500 | > loss: -0.54262 (-0.54601) | > log_mle: -0.64938 (-0.65520) | > loss_dur: 0.10676 (0.10919) | > amp_scaler: 2048.00000 (2048.00000) | > grad_norm: 9.82448 (9.98517) | > current_lr: 0.00022 | > step_time: 0.48740 (0.39077) | > loader_time: 0.00330 (0.00294)  --> STEP: 125/639 -- GLOBAL_STEP: 86525 | > loss: -0.52370 (-0.54264) | > log_mle: -0.64023 (-0.65246) | > loss_dur: 0.11653 (0.10982) | > amp_scaler: 4096.00000 (2195.45600) | > grad_norm: 10.34250 (10.08960) | > current_lr: 0.00022 | > step_time: 0.52170 (0.40447) | > loader_time: 0.00310 (0.00295)  --> STEP: 150/639 -- GLOBAL_STEP: 86550 | > loss: -0.50207 (-0.53962) | > log_mle: -0.60373 (-0.65004) | > loss_dur: 0.10165 (0.11043) | > amp_scaler: 2048.00000 (2225.49333) | > grad_norm: 6.19019 (10.69658) | > current_lr: 0.00022 | > step_time: 0.42640 (0.42277) | > loader_time: 0.00280 (0.00298)  --> STEP: 175/639 -- GLOBAL_STEP: 86575 | > loss: -0.51899 (-0.53862) | > log_mle: -0.62127 (-0.64933) | > loss_dur: 0.10229 (0.11071) | > amp_scaler: 2048.00000 (2200.13714) | > grad_norm: 7.52584 (11.00585) | > current_lr: 0.00022 | > step_time: 0.55670 (0.43845) | > loader_time: 0.00350 (0.00302)  --> STEP: 200/639 -- GLOBAL_STEP: 86600 | > loss: -0.49112 (-0.53589) | > log_mle: -0.60409 (-0.64632) | > loss_dur: 0.11297 (0.11042) | > amp_scaler: 2048.00000 (2181.12000) | > grad_norm: 11.01536 (10.89806) | > current_lr: 0.00021 | > step_time: 0.52900 (0.45420) | > loader_time: 0.00320 (0.00305)  --> STEP: 225/639 -- GLOBAL_STEP: 86625 | > loss: -0.49479 (-0.53328) | > log_mle: -0.58873 (-0.64362) | > loss_dur: 0.09394 (0.11034) | > amp_scaler: 2048.00000 (2166.32889) | > grad_norm: 4.46823 (10.73043) | > current_lr: 0.00021 | > step_time: 0.51720 (0.47301) | > loader_time: 0.00310 (0.00374)  --> STEP: 250/639 -- GLOBAL_STEP: 86650 | > loss: -0.49976 (-0.53358) | > log_mle: -0.59017 (-0.64417) | > loss_dur: 0.09042 (0.11059) | > amp_scaler: 2048.00000 (2154.49600) | > grad_norm: 5.28083 (10.80575) | > current_lr: 0.00021 | > step_time: 0.62890 (0.48855) | > loader_time: 0.00430 (0.00387)  --> STEP: 275/639 -- GLOBAL_STEP: 86675 | > loss: -0.51634 (-0.53295) | > log_mle: -0.62676 (-0.64414) | > loss_dur: 0.11042 (0.11118) | > amp_scaler: 2048.00000 (2144.81455) | > grad_norm: 9.18000 (11.04741) | > current_lr: 0.00021 | > step_time: 0.89270 (0.50733) | > loader_time: 0.00420 (0.00385) ! Run is kept in /media/popos/Barracuda/Models/TTS_new/trained/coqui_tts-January-23-2022_02+21PM-c63bb481