SirAB commited on
Commit
281fc32
·
verified ·
1 Parent(s): 23e6ce9

Upload folder using huggingface_hub

Browse files
LJSpeech6/config.yml CHANGED
@@ -1,21 +1,118 @@
1
- {ASR_config: Utils/ASR/config.yml, ASR_path: Utils/ASR/epoch_00080.pth, F0_path: Utils/JDC/bst.t7,
2
- PLBERT_dir: Utils/PLBERT/, batch_size: 4, data_params: {OOD_data: Data/OOD_texts.txt,
3
- min_length: 50, root_path: /content/StyleTTS2/Dataset, train_data: Data/train_list.txt,
4
- val_data: Data/val_list.txt}, device: cuda, epochs_1st: 20, epochs_2nd: 300, first_stage_path: first_stage.pth,
5
- load_only_params: false, log_dir: Models/LJSpeech6, log_interval: 10, loss_params: {
6
- TMA_epoch: 10000, diff_epoch: 0, joint_epoch: 0, lambda_F0: 1.0, lambda_ce: 20.0,
7
- lambda_diff: 1.0, lambda_dur: 1.0, lambda_gen: 1.0, lambda_mel: 5.0, lambda_mono: 1.0,
8
- lambda_norm: 1.0, lambda_s2s: 1.0, lambda_slm: 1.0, lambda_sty: 1.0}, max_len: 295,
9
- model_params: {decoder: {gen_istft_hop_size: 5, gen_istft_n_fft: 20, resblock_dilation_sizes: [
10
- [1, 3, 5], [1, 3, 5], [1, 3, 5]], resblock_kernel_sizes: [3, 7, 11], type: istftnet,
11
- upsample_initial_channel: 512, upsample_kernel_sizes: [20, 12], upsample_rates: [
12
- 10, 6]}, diffusion: {dist: {estimate_sigma_data: true, mean: -3.0, sigma_data: .nan,
13
- std: 1.0}, embedding_mask_proba: 0.1, transformer: {head_features: 64, multiplier: 2,
14
- num_heads: 8, num_layers: 3}}, dim_in: 64, dropout: 0.2, hidden_dim: 512,
15
- max_conv_dim: 512, max_dur: 50, multispeaker: false, n_layer: 3, n_mels: 80, n_token: 185,
16
- slm: {hidden: 768, initial_channel: 64, model: microsoft/wavlm-base-plus, nlayers: 13,
17
- sr: 16000}, style_dim: 128}, optimizer_params: {bert_lr: 1.0e-05, ft_lr: 1.0e-05,
18
- lr: 0.0001}, preprocess_params: {spect_params: {hop_length: 300, n_fft: 2048,
19
- win_length: 1200}, sr: 24000}, pretrained_model: LJSpeech6/epoch_2nd_00140.pth,
20
- save_freq: 1, second_stage_load_pretrained: true, slmadv_params: {batch_percentage: 1,
21
- iter: 10, max_len: 500, min_len: 400, scale: 0.01, sig: 1.5, thresh: 5}}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ log_dir: "Models/LJSpeech6"
2
+ first_stage_path: "first_stage.pth"
3
+ save_freq: 1
4
+ log_interval: 10
5
+ device: "cuda"
6
+ epochs_1st: 2000 # number of epochs for first stage training (pre-training)
7
+ epochs_2nd: 300 # number of peochs for second stage training (joint training)
8
+ # first epoch = 3x650
9
+ # second epoch = 2x440
10
+ # second stage 1st epoch = 2x300
11
+ batch_size: 2
12
+ max_len: 550 # maximum number of frames
13
+ pretrained_model: "LJSpeech6/epoch_2nd_00151.pth"
14
+ second_stage_load_pretrained: true # set to true if the pre-trained model is for 2nd stage
15
+ load_only_params: false # set to true if do not want to load epoch numbers and optimizer parameters
16
+
17
+ F0_path: "Utils/JDC/bst.t7"
18
+ ASR_config: "Utils/ASR/config.yml"
19
+ ASR_path: "Utils/ASR/epoch_00080.pth"
20
+ PLBERT_dir: 'Utils/PLBERT/'
21
+
22
+ data_params:
23
+ train_data: "Data/train_list.txt"
24
+ val_data: "Data/val_list.txt"
25
+ root_path: "/content/StyleTTS2/Dataset"
26
+ OOD_data: "Data/OOD_texts.txt"
27
+ min_length: 50 # sample until texts with this size are obtained for OOD texts
28
+
29
+ preprocess_params:
30
+ sr: 24000
31
+ spect_params:
32
+ n_fft: 2048
33
+ win_length: 1200
34
+ hop_length: 300
35
+
36
+ model_params:
37
+ multispeaker: false
38
+
39
+ dim_in: 64
40
+ hidden_dim: 512
41
+ max_conv_dim: 512
42
+ n_layer: 3
43
+ n_mels: 80
44
+
45
+ n_token: 185 # number of phoneme tokens
46
+ max_dur: 50 # maximum duration of a single phoneme
47
+ style_dim: 128 # style vector size
48
+
49
+ dropout: 0.2
50
+
51
+ # config for decoder
52
+ decoder:
53
+ type: 'istftnet' # either hifigan or istftnet
54
+ resblock_kernel_sizes: [3,7,11]
55
+ upsample_rates : [10, 6]
56
+ upsample_initial_channel: 512
57
+ resblock_dilation_sizes: [[1,3,5], [1,3,5], [1,3,5]]
58
+ upsample_kernel_sizes: [20, 12]
59
+ gen_istft_n_fft: 20
60
+ gen_istft_hop_size: 5
61
+
62
+ # speech language model config
63
+ slm:
64
+ model: 'microsoft/wavlm-base-plus'
65
+ sr: 16000 # sampling rate of SLM
66
+ hidden: 768 # hidden size of SLM
67
+ nlayers: 13 # number of layers of SLM
68
+ initial_channel: 64 # initial channels of SLM discriminator head
69
+
70
+ # style diffusion model config
71
+ diffusion:
72
+ embedding_mask_proba: 0.1
73
+ # transformer config
74
+ transformer:
75
+ num_layers: 3
76
+ num_heads: 8
77
+ head_features: 64
78
+ multiplier: 2
79
+
80
+ # diffusion distribution config
81
+ dist:
82
+ sigma_data: 0.2 # placeholder for estimate_sigma_data set to false
83
+ estimate_sigma_data: true # estimate sigma_data from the current batch if set to true
84
+ mean: -3.0
85
+ std: 1.0
86
+
87
+ loss_params:
88
+ lambda_mel: 5. # mel reconstruction loss
89
+ lambda_gen: 1. # generator loss
90
+ lambda_slm: 1. # slm feature matching loss
91
+
92
+ lambda_mono: 1. # monotonic alignment loss (1st stage, TMA)
93
+ lambda_s2s: 1. # sequence-to-sequence loss (1st stage, TMA)
94
+ TMA_epoch: 2 # TMA starting epoch (1st stage)
95
+
96
+ lambda_F0: 1. # F0 reconstruction loss (2nd stage)
97
+ lambda_norm: 1. # norm reconstruction loss (2nd stage)
98
+ lambda_dur: 1. # duration loss (2nd stage)
99
+ lambda_ce: 20. # duration predictor probability output CE loss (2nd stage)
100
+ lambda_sty: 1. # style reconstruction loss (2nd stage)
101
+ lambda_diff: 1. # score matching loss (2nd stage)
102
+
103
+ diff_epoch: 0 # style diffusion starting epoch (2nd stage)
104
+ joint_epoch: 0 # joint training starting epoch (2nd stage)
105
+
106
+ optimizer_params:
107
+ lr: 0.0001 # general learning rate
108
+ bert_lr: 0.00001 # learning rate for PLBERT
109
+ ft_lr: 0.00001 # learning rate for acoustic modules
110
+
111
+ slmadv_params:
112
+ min_len: 400 # minimum length of samples
113
+ max_len: 500 # maximum length of samples
114
+ batch_percentage: 1 # to prevent out of memory, only use half of the original batch size
115
+ iter: 10 # update the discriminator every this iterations of generator update
116
+ thresh: 5 # gradient norm above which the gradient is scaled
117
+ scale: 0.01 # gradient scaling factor for predictors from SLM discriminators
118
+ sig: 1.5 # sigma for differentiable duration modeling
LJSpeech6/epoch_2nd_00152.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b4849254bfef5bb8c8e057ca2a25f44f70a53def8a31eed58de6b4775b4e570
3
+ size 1279652187
LJSpeech6/tensorboard/events.out.tfevents.1745000660.1d6a958ad39e.1635.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:244fdf0ca0470ec023380c46de560aeff1ca8412d93391b60021a56897c7cc32
3
+ size 88
LJSpeech6/tensorboard/events.out.tfevents.1745001071.1d6a958ad39e.3378.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2279eb44f5fb522c14cef9cc896b34963abe78a7ad63f3ebd02f4d08fd5c5604
3
+ size 88
LJSpeech6/tensorboard/events.out.tfevents.1745001422.1d6a958ad39e.4834.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:802d8f150b6a3999cfa674dd179f77ba8ca0b6a44213399c843d28e91b509a54
3
+ size 88
LJSpeech6/tensorboard/events.out.tfevents.1745001472.1d6a958ad39e.5085.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33ec6c798658fb489a5e2d17acf719468fbb2174e3825839708c834a0ddfb2ad
3
+ size 22912
LJSpeech6/train.log CHANGED
@@ -1,76 +1,71 @@
1
- INFO:2025-04-18 16:39:20,125: Epoch [142/300], Step [10/51], Loss: 0.68524, Disc Loss: 1.47922, Dur Loss: 1.62595, CE Loss: 0.08945, Norm Loss: 2.03682, F0 Loss: 6.96133, LM Loss: 2.70599, Gen Loss: 21.94340
2
- INFO:2025-04-18 16:40:04,242: Epoch [142/300], Step [20/51], Loss: 0.70318, Disc Loss: 1.68217, Dur Loss: 1.64298, CE Loss: 0.11049, Norm Loss: 1.62010, F0 Loss: 6.22028, LM Loss: 2.80196, Gen Loss: 20.51518
3
- INFO:2025-04-18 16:40:49,219: Epoch [142/300], Step [30/51], Loss: 0.74260, Disc Loss: 1.41660, Dur Loss: 1.60158, CE Loss: 0.09415, Norm Loss: 2.11386, F0 Loss: 7.16752, LM Loss: 2.79458, Gen Loss: 22.39705
4
- INFO:2025-04-18 16:41:35,373: Epoch [142/300], Step [40/51], Loss: 0.71187, Disc Loss: 1.67019, Dur Loss: 1.50791, CE Loss: 0.09333, Norm Loss: 1.37806, F0 Loss: 5.40739, LM Loss: 2.68052, Gen Loss: 21.68308
5
- INFO:2025-04-18 16:42:20,563: Epoch [142/300], Step [50/51], Loss: 0.75412, Disc Loss: 1.78135, Dur Loss: 1.95916, CE Loss: 0.11991, Norm Loss: 1.49295, F0 Loss: 4.60098, LM Loss: 2.51455, Gen Loss: 20.57650
6
- INFO:2025-04-18 16:43:16,065: Epoch [143/300], Step [10/51], Loss: 0.71390, Disc Loss: 1.48582, Dur Loss: 1.50436, CE Loss: 0.09131, Norm Loss: 1.62147, F0 Loss: 5.11521, LM Loss: 2.72468, Gen Loss: 23.37470
7
- INFO:2025-04-18 16:44:01,769: Epoch [143/300], Step [20/51], Loss: 0.74787, Disc Loss: 1.57418, Dur Loss: 1.57428, CE Loss: 0.09518, Norm Loss: 2.16336, F0 Loss: 7.19010, LM Loss: 2.86603, Gen Loss: 21.46209
8
- INFO:2025-04-18 16:44:47,704: Epoch [143/300], Step [30/51], Loss: 0.70961, Disc Loss: 1.62335, Dur Loss: 1.74768, CE Loss: 0.11611, Norm Loss: 1.80134, F0 Loss: 5.80356, LM Loss: 2.67650, Gen Loss: 20.76577
9
- INFO:2025-04-18 16:45:34,793: Epoch [143/300], Step [40/51], Loss: 0.69925, Disc Loss: 1.59267, Dur Loss: 1.58126, CE Loss: 0.09940, Norm Loss: 1.66421, F0 Loss: 5.05437, LM Loss: 2.70929, Gen Loss: 21.37688
10
- INFO:2025-04-18 16:46:20,844: Epoch [143/300], Step [50/51], Loss: 0.72013, Disc Loss: 1.56773, Dur Loss: 1.48976, CE Loss: 0.09496, Norm Loss: 1.88727, F0 Loss: 6.62523, LM Loss: 2.73424, Gen Loss: 21.24307
11
- INFO:2025-04-18 16:47:15,683: Epoch [144/300], Step [10/51], Loss: 0.72873, Disc Loss: 1.35234, Dur Loss: 1.48549, CE Loss: 0.08650, Norm Loss: 1.58709, F0 Loss: 6.06568, LM Loss: 2.62467, Gen Loss: 21.80611
12
- INFO:2025-04-18 16:48:02,333: Epoch [144/300], Step [20/51], Loss: 0.71111, Disc Loss: 1.58655, Dur Loss: 1.67996, CE Loss: 0.09450, Norm Loss: 1.88139, F0 Loss: 5.80457, LM Loss: 2.79424, Gen Loss: 21.44979
13
- INFO:2025-04-18 16:48:49,041: Epoch [144/300], Step [30/51], Loss: 0.67972, Disc Loss: 2.23544, Dur Loss: 1.51179, CE Loss: 0.09498, Norm Loss: 1.95588, F0 Loss: 5.47160, LM Loss: 2.68061, Gen Loss: 17.45283
14
- INFO:2025-04-18 16:49:34,379: Epoch [144/300], Step [40/51], Loss: 0.71675, Disc Loss: 1.62790, Dur Loss: 1.50520, CE Loss: 0.08794, Norm Loss: 1.77580, F0 Loss: 5.37220, LM Loss: 2.64495, Gen Loss: 20.26120
15
- INFO:2025-04-18 16:50:19,900: Epoch [144/300], Step [50/51], Loss: 0.67266, Disc Loss: 1.39714, Dur Loss: 1.63816, CE Loss: 0.10359, Norm Loss: 1.78194, F0 Loss: 6.48239, LM Loss: 2.85297, Gen Loss: 22.06832
16
- INFO:2025-04-18 16:51:14,162: Epoch [145/300], Step [10/51], Loss: 0.68569, Disc Loss: 1.68265, Dur Loss: 1.63509, CE Loss: 0.09776, Norm Loss: 1.97500, F0 Loss: 5.79533, LM Loss: 2.78410, Gen Loss: 20.37291
17
- INFO:2025-04-18 16:52:00,638: Epoch [145/300], Step [20/51], Loss: 0.71857, Disc Loss: 1.42100, Dur Loss: 1.44433, CE Loss: 0.09085, Norm Loss: 1.88905, F0 Loss: 6.61303, LM Loss: 2.81552, Gen Loss: 21.89507
18
- INFO:2025-04-18 16:52:45,922: Epoch [145/300], Step [30/51], Loss: 0.73107, Disc Loss: 1.63563, Dur Loss: 1.43416, CE Loss: 0.08593, Norm Loss: 1.88030, F0 Loss: 6.81012, LM Loss: 2.77800, Gen Loss: 21.17454
19
- INFO:2025-04-18 16:53:31,593: Epoch [145/300], Step [40/51], Loss: 0.72037, Disc Loss: 1.49394, Dur Loss: 1.41632, CE Loss: 0.08588, Norm Loss: 1.60326, F0 Loss: 4.54339, LM Loss: 2.70066, Gen Loss: 21.57268
20
- INFO:2025-04-18 16:54:18,240: Epoch [145/300], Step [50/51], Loss: 0.71887, Disc Loss: 1.42307, Dur Loss: 1.60237, CE Loss: 0.10073, Norm Loss: 1.76963, F0 Loss: 6.66953, LM Loss: 2.78358, Gen Loss: 22.23439
21
- INFO:2025-04-18 16:55:12,834: Epoch [146/300], Step [10/51], Loss: 0.73992, Disc Loss: 1.36057, Dur Loss: 1.51505, CE Loss: 0.08914, Norm Loss: 1.96433, F0 Loss: 5.58034, LM Loss: 2.89096, Gen Loss: 22.95247
22
- INFO:2025-04-18 16:55:58,112: Epoch [146/300], Step [20/51], Loss: 0.66677, Disc Loss: 1.43688, Dur Loss: 1.55628, CE Loss: 0.09829, Norm Loss: 1.84392, F0 Loss: 5.33902, LM Loss: 2.75534, Gen Loss: 22.39407
23
- INFO:2025-04-18 16:56:45,124: Epoch [146/300], Step [30/51], Loss: 0.70521, Disc Loss: 1.92941, Dur Loss: 1.49728, CE Loss: 0.09351, Norm Loss: 1.93437, F0 Loss: 6.03800, LM Loss: 2.82335, Gen Loss: 19.55962
24
- INFO:2025-04-18 16:57:31,937: Epoch [146/300], Step [40/51], Loss: 0.72947, Disc Loss: 1.47157, Dur Loss: 1.71434, CE Loss: 0.10261, Norm Loss: 1.75376, F0 Loss: 6.98688, LM Loss: 2.82185, Gen Loss: 22.05879
25
- INFO:2025-04-18 16:58:18,094: Epoch [146/300], Step [50/51], Loss: 0.71457, Disc Loss: 1.68838, Dur Loss: 1.48354, CE Loss: 0.09195, Norm Loss: 1.73387, F0 Loss: 5.42509, LM Loss: 2.60006, Gen Loss: 19.71683
26
- INFO:2025-04-18 16:59:13,031: Epoch [147/300], Step [10/51], Loss: 0.75130, Disc Loss: 1.38349, Dur Loss: 1.64346, CE Loss: 0.10354, Norm Loss: 2.20299, F0 Loss: 6.75077, LM Loss: 2.87051, Gen Loss: 22.60429
27
- INFO:2025-04-18 16:59:59,524: Epoch [147/300], Step [20/51], Loss: 0.71444, Disc Loss: 1.47054, Dur Loss: 1.51426, CE Loss: 0.10092, Norm Loss: 2.11063, F0 Loss: 7.11924, LM Loss: 2.92353, Gen Loss: 22.27351
28
- INFO:2025-04-18 17:00:46,346: Epoch [147/300], Step [30/51], Loss: 0.69825, Disc Loss: 1.94287, Dur Loss: 1.58962, CE Loss: 0.09698, Norm Loss: 1.68689, F0 Loss: 4.78847, LM Loss: 2.58855, Gen Loss: 19.52784
29
- INFO:2025-04-18 17:01:32,149: Epoch [147/300], Step [40/51], Loss: 0.70886, Disc Loss: 1.35376, Dur Loss: 1.49707, CE Loss: 0.09237, Norm Loss: 1.45337, F0 Loss: 5.58208, LM Loss: 2.70223, Gen Loss: 21.89305
30
- INFO:2025-04-18 17:02:17,692: Epoch [147/300], Step [50/51], Loss: 0.71379, Disc Loss: 1.36837, Dur Loss: 1.37852, CE Loss: 0.08751, Norm Loss: 1.98776, F0 Loss: 7.14638, LM Loss: 2.77982, Gen Loss: 22.15689
31
- INFO:2025-04-18 17:03:11,055: Epoch [148/300], Step [10/51], Loss: 0.69688, Disc Loss: 1.64587, Dur Loss: 1.46361, CE Loss: 0.09532, Norm Loss: 1.80441, F0 Loss: 6.44951, LM Loss: 2.77589, Gen Loss: 20.38474
32
- INFO:2025-04-18 17:03:58,244: Epoch [148/300], Step [20/51], Loss: 0.71136, Disc Loss: 1.42730, Dur Loss: 1.47970, CE Loss: 0.08980, Norm Loss: 1.45393, F0 Loss: 5.42928, LM Loss: 2.53388, Gen Loss: 21.60295
33
- INFO:2025-04-18 17:04:44,920: Epoch [148/300], Step [30/51], Loss: 0.70046, Disc Loss: 1.68432, Dur Loss: 1.63618, CE Loss: 0.10163, Norm Loss: 2.16876, F0 Loss: 6.09182, LM Loss: 2.75550, Gen Loss: 20.23632
34
- INFO:2025-04-18 17:05:31,951: Epoch [148/300], Step [40/51], Loss: 0.71403, Disc Loss: 1.71507, Dur Loss: 1.59477, CE Loss: 0.09961, Norm Loss: 1.67050, F0 Loss: 5.20799, LM Loss: 2.66728, Gen Loss: 20.75716
35
- INFO:2025-04-18 17:06:18,406: Epoch [148/300], Step [50/51], Loss: 0.75098, Disc Loss: 1.39508, Dur Loss: 1.49480, CE Loss: 0.09076, Norm Loss: 2.11631, F0 Loss: 7.33379, LM Loss: 2.82389, Gen Loss: 22.85464
36
- INFO:2025-04-18 17:07:14,481: Epoch [149/300], Step [10/51], Loss: 0.74070, Disc Loss: 1.69791, Dur Loss: 1.82487, CE Loss: 0.11918, Norm Loss: 1.63369, F0 Loss: 5.28787, LM Loss: 2.62497, Gen Loss: 20.25367
37
- INFO:2025-04-18 17:08:00,375: Epoch [149/300], Step [20/51], Loss: 0.70327, Disc Loss: 1.44869, Dur Loss: 1.52203, CE Loss: 0.08958, Norm Loss: 1.49739, F0 Loss: 5.85132, LM Loss: 2.58344, Gen Loss: 21.60640
38
- INFO:2025-04-18 17:08:46,814: Epoch [149/300], Step [30/51], Loss: 0.71287, Disc Loss: 1.34402, Dur Loss: 1.82890, CE Loss: 0.12270, Norm Loss: 2.06443, F0 Loss: 6.87650, LM Loss: 2.86582, Gen Loss: 22.80039
39
- INFO:2025-04-18 17:09:32,941: Epoch [149/300], Step [40/51], Loss: 0.71506, Disc Loss: 2.02102, Dur Loss: 1.68954, CE Loss: 0.09869, Norm Loss: 1.68995, F0 Loss: 4.92334, LM Loss: 2.59965, Gen Loss: 18.51545
40
- INFO:2025-04-18 17:10:19,215: Epoch [149/300], Step [50/51], Loss: 0.73065, Disc Loss: 2.07906, Dur Loss: 1.56088, CE Loss: 0.09529, Norm Loss: 1.94661, F0 Loss: 6.35400, LM Loss: 2.71965, Gen Loss: 19.03009
41
- INFO:2025-04-18 17:11:14,595: Epoch [150/300], Step [10/51], Loss: 0.73004, Disc Loss: 1.62074, Dur Loss: 1.50655, CE Loss: 0.08293, Norm Loss: 1.71244, F0 Loss: 5.21188, LM Loss: 2.69304, Gen Loss: 21.09586
42
- INFO:2025-04-18 17:12:00,664: Epoch [150/300], Step [20/51], Loss: 0.69430, Disc Loss: 1.56964, Dur Loss: 1.52544, CE Loss: 0.09385, Norm Loss: 2.13682, F0 Loss: 6.80176, LM Loss: 2.88211, Gen Loss: 22.36215
43
- INFO:2025-04-18 17:12:46,820: Epoch [150/300], Step [30/51], Loss: 0.71136, Disc Loss: 1.73209, Dur Loss: 1.73939, CE Loss: 0.11372, Norm Loss: 1.50729, F0 Loss: 5.26564, LM Loss: 2.61472, Gen Loss: 20.30487
44
- INFO:2025-04-18 17:13:33,612: Epoch [150/300], Step [40/51], Loss: 0.73566, Disc Loss: 1.47867, Dur Loss: 1.58538, CE Loss: 0.11185, Norm Loss: 1.64122, F0 Loss: 5.81193, LM Loss: 2.68888, Gen Loss: 22.10338
45
- INFO:2025-04-18 17:14:19,874: Epoch [150/300], Step [50/51], Loss: 0.71072, Disc Loss: 1.92142, Dur Loss: 1.54732, CE Loss: 0.09288, Norm Loss: 1.90880, F0 Loss: 6.89812, LM Loss: 2.55315, Gen Loss: 19.44312
46
- INFO:2025-04-18 17:15:14,877: Epoch [151/300], Step [10/51], Loss: 0.70019, Disc Loss: 1.44330, Dur Loss: 1.57783, CE Loss: 0.09280, Norm Loss: 1.79696, F0 Loss: 7.05911, LM Loss: 2.78229, Gen Loss: 21.88181
47
- INFO:2025-04-18 17:16:01,200: Epoch [151/300], Step [20/51], Loss: 0.69136, Disc Loss: 1.42572, Dur Loss: 1.51181, CE Loss: 0.09090, Norm Loss: 2.00573, F0 Loss: 6.76366, LM Loss: 2.84892, Gen Loss: 22.14789
48
- INFO:2025-04-18 17:16:48,032: Epoch [151/300], Step [30/51], Loss: 0.70772, Disc Loss: 1.77513, Dur Loss: 1.59214, CE Loss: 0.09783, Norm Loss: 1.66121, F0 Loss: 4.93407, LM Loss: 2.66925, Gen Loss: 20.40581
49
- INFO:2025-04-18 17:17:34,599: Epoch [151/300], Step [40/51], Loss: 0.70356, Disc Loss: 1.62519, Dur Loss: 1.52059, CE Loss: 0.09178, Norm Loss: 1.68702, F0 Loss: 4.80802, LM Loss: 2.72435, Gen Loss: 20.85572
50
- INFO:2025-04-18 17:18:20,315: Epoch [151/300], Step [50/51], Loss: 0.72731, Disc Loss: 2.05195, Dur Loss: 1.55876, CE Loss: 0.09530, Norm Loss: 1.74761, F0 Loss: 4.51685, LM Loss: 2.57700, Gen Loss: 18.55693
51
- INFO:2025-04-18 17:19:15,863: Epoch [152/300], Step [10/51], Loss: 0.73504, Disc Loss: 1.56063, Dur Loss: 1.55307, CE Loss: 0.09284, Norm Loss: 1.43491, F0 Loss: 5.02914, LM Loss: 2.58629, Gen Loss: 21.15729
52
- INFO:2025-04-18 17:20:03,100: Epoch [152/300], Step [20/51], Loss: 0.72277, Disc Loss: 1.60806, Dur Loss: 1.72357, CE Loss: 0.10120, Norm Loss: 1.85721, F0 Loss: 5.62259, LM Loss: 2.86344, Gen Loss: 22.10464
53
- INFO:2025-04-18 17:20:50,062: Epoch [152/300], Step [30/51], Loss: 0.70419, Disc Loss: 1.49078, Dur Loss: 1.46825, CE Loss: 0.09376, Norm Loss: 1.78132, F0 Loss: 5.68358, LM Loss: 2.82758, Gen Loss: 21.58227
54
- INFO:2025-04-18 17:21:35,967: Epoch [152/300], Step [40/51], Loss: 0.72710, Disc Loss: 1.66195, Dur Loss: 1.67446, CE Loss: 0.11067, Norm Loss: 2.09811, F0 Loss: 5.60208, LM Loss: 2.84211, Gen Loss: 21.41905
55
- INFO:2025-04-18 17:22:21,168: Epoch [152/300], Step [50/51], Loss: 0.70813, Disc Loss: 1.71920, Dur Loss: 1.55023, CE Loss: 0.09873, Norm Loss: 2.07161, F0 Loss: 5.79307, LM Loss: 2.81213, Gen Loss: 20.84192
56
- INFO:2025-04-18 17:23:16,711: Epoch [153/300], Step [10/51], Loss: 0.73143, Disc Loss: 1.79570, Dur Loss: 1.59634, CE Loss: 0.09442, Norm Loss: 1.92937, F0 Loss: 6.42911, LM Loss: 2.62480, Gen Loss: 20.38996
57
- INFO:2025-04-18 17:24:02,194: Epoch [153/300], Step [20/51], Loss: 0.69290, Disc Loss: 1.61755, Dur Loss: 1.60123, CE Loss: 0.09615, Norm Loss: 1.75627, F0 Loss: 6.65929, LM Loss: 2.64778, Gen Loss: 21.37987
58
- INFO:2025-04-18 17:24:48,238: Epoch [153/300], Step [30/51], Loss: 0.71795, Disc Loss: 1.33923, Dur Loss: 1.52036, CE Loss: 0.10669, Norm Loss: 1.81211, F0 Loss: 6.48128, LM Loss: 2.91253, Gen Loss: 23.04830
59
- INFO:2025-04-18 17:25:34,280: Epoch [153/300], Step [40/51], Loss: 0.73072, Disc Loss: 1.42978, Dur Loss: 1.46818, CE Loss: 0.08931, Norm Loss: 1.41775, F0 Loss: 4.25644, LM Loss: 2.60908, Gen Loss: 21.82592
60
- INFO:2025-04-18 17:26:20,821: Epoch [153/300], Step [50/51], Loss: 0.72157, Disc Loss: 1.68197, Dur Loss: 1.56651, CE Loss: 0.09883, Norm Loss: 2.16632, F0 Loss: 7.83860, LM Loss: 2.82899, Gen Loss: 20.56416
61
- INFO:2025-04-18 17:27:14,384: Epoch [154/300], Step [10/51], Loss: 0.73200, Disc Loss: 1.65672, Dur Loss: 1.57255, CE Loss: 0.09559, Norm Loss: 1.81036, F0 Loss: 6.10940, LM Loss: 2.50117, Gen Loss: 21.23557
62
- INFO:2025-04-18 17:28:00,248: Epoch [154/300], Step [20/51], Loss: 0.70311, Disc Loss: 1.68385, Dur Loss: 1.49009, CE Loss: 0.09227, Norm Loss: 1.81099, F0 Loss: 5.62639, LM Loss: 2.68044, Gen Loss: 21.13133
63
- INFO:2025-04-18 17:28:47,077: Epoch [154/300], Step [30/51], Loss: 0.71638, Disc Loss: 1.87309, Dur Loss: 1.46048, CE Loss: 0.09010, Norm Loss: 1.82299, F0 Loss: 5.67835, LM Loss: 2.76285, Gen Loss: 20.39882
64
- INFO:2025-04-18 17:29:34,633: Epoch [154/300], Step [40/51], Loss: 0.71417, Disc Loss: 1.74454, Dur Loss: 1.68811, CE Loss: 0.09437, Norm Loss: 1.88066, F0 Loss: 5.59537, LM Loss: 2.65140, Gen Loss: 20.38334
65
- INFO:2025-04-18 17:30:19,702: Epoch [154/300], Step [50/51], Loss: 0.69926, Disc Loss: 1.56151, Dur Loss: 1.62801, CE Loss: 0.08694, Norm Loss: 1.74591, F0 Loss: 5.90715, LM Loss: 2.75128, Gen Loss: 21.50785
66
- INFO:2025-04-18 17:31:14,855: Epoch [155/300], Step [10/51], Loss: 0.72248, Disc Loss: 1.46588, Dur Loss: 1.56516, CE Loss: 0.10372, Norm Loss: 1.61600, F0 Loss: 5.80945, LM Loss: 2.59780, Gen Loss: 22.34130
67
- INFO:2025-04-18 17:32:02,197: Epoch [155/300], Step [20/51], Loss: 0.72941, Disc Loss: 1.47039, Dur Loss: 1.48996, CE Loss: 0.08999, Norm Loss: 2.22887, F0 Loss: 6.74143, LM Loss: 2.87571, Gen Loss: 22.41432
68
- INFO:2025-04-18 17:32:47,432: Epoch [155/300], Step [30/51], Loss: 0.71599, Disc Loss: 1.65303, Dur Loss: 1.52149, CE Loss: 0.10532, Norm Loss: 1.44692, F0 Loss: 5.11314, LM Loss: 2.62245, Gen Loss: 20.54512
69
- INFO:2025-04-18 17:33:33,161: Epoch [155/300], Step [40/51], Loss: 0.68676, Disc Loss: 2.01427, Dur Loss: 1.95429, CE Loss: 0.11006, Norm Loss: 1.65905, F0 Loss: 4.96194, LM Loss: 2.60894, Gen Loss: 18.35085
70
- INFO:2025-04-18 17:34:19,054: Epoch [155/300], Step [50/51], Loss: 0.70916, Disc Loss: 1.68487, Dur Loss: 1.52156, CE Loss: 0.08977, Norm Loss: 1.77867, F0 Loss: 7.10991, LM Loss: 2.90049, Gen Loss: 20.78979
71
- INFO:2025-04-18 17:35:14,827: Epoch [156/300], Step [10/51], Loss: 0.71127, Disc Loss: 1.63891, Dur Loss: 1.57623, CE Loss: 0.09776, Norm Loss: 2.01388, F0 Loss: 7.44915, LM Loss: 2.82735, Gen Loss: 20.94834
72
- INFO:2025-04-18 17:36:01,426: Epoch [156/300], Step [20/51], Loss: 0.69645, Disc Loss: 1.75043, Dur Loss: 1.55425, CE Loss: 0.08743, Norm Loss: 1.50034, F0 Loss: 5.00427, LM Loss: 2.59502, Gen Loss: 20.31112
73
- INFO:2025-04-18 17:36:47,504: Epoch [156/300], Step [30/51], Loss: 0.68754, Disc Loss: 1.67645, Dur Loss: 1.46699, CE Loss: 0.08220, Norm Loss: 1.52650, F0 Loss: 6.08870, LM Loss: 2.68570, Gen Loss: 19.96917
74
- INFO:2025-04-18 17:37:33,229: Epoch [156/300], Step [40/51], Loss: 0.69898, Disc Loss: 1.90957, Dur Loss: 1.66669, CE Loss: 0.10101, Norm Loss: 1.59804, F0 Loss: 4.71516, LM Loss: 2.67487, Gen Loss: 20.27877
75
- INFO:2025-04-18 17:38:19,381: Epoch [156/300], Step [50/51], Loss: 0.72755, Disc Loss: 1.42983, Dur Loss: 1.65149, CE Loss: 0.10703, Norm Loss: 2.62230, F0 Loss: 7.85275, LM Loss: 2.93919, Gen Loss: 23.04290
76
- INFO:2025-04-18 17:39:13,841: Epoch [157/300], Step [10/51], Loss: 0.74044, Disc Loss: 1.57090, Dur Loss: 1.53175, CE Loss: 0.09511, Norm Loss: 2.13309, F0 Loss: 6.38471, LM Loss: 2.84712, Gen Loss: 21.28060
 
1
+ INFO:2025-04-18 18:38:35,588: Epoch [152/2000], Step [10/103], Mel Loss: 0.57166, Gen Loss: 16.83132, Disc Loss: 2.43236, Mono Loss: 0.02193, S2S Loss: 9.49231, SLM Loss: 2.51258
2
+ INFO:2025-04-18 18:39:04,034: Epoch [152/2000], Step [20/103], Mel Loss: 0.51311, Gen Loss: 19.25039, Disc Loss: 2.02201, Mono Loss: 0.02076, S2S Loss: 7.03681, SLM Loss: 2.47461
3
+ INFO:2025-04-18 18:39:33,757: Epoch [152/2000], Step [30/103], Mel Loss: 0.54050, Gen Loss: 18.68884, Disc Loss: 1.87438, Mono Loss: 0.02688, S2S Loss: 7.64351, SLM Loss: 2.33585
4
+ INFO:2025-04-18 18:40:03,206: Epoch [152/2000], Step [40/103], Mel Loss: 0.52822, Gen Loss: 15.79294, Disc Loss: 2.44520, Mono Loss: 0.01991, S2S Loss: 6.78955, SLM Loss: 2.47681
5
+ INFO:2025-04-18 18:40:33,769: Epoch [152/2000], Step [50/103], Mel Loss: 0.52643, Gen Loss: 18.56133, Disc Loss: 2.18126, Mono Loss: 0.05364, S2S Loss: 7.77592, SLM Loss: 2.27586
6
+ INFO:2025-04-18 18:41:01,775: Epoch [152/2000], Step [60/103], Mel Loss: 0.52666, Gen Loss: 19.28783, Disc Loss: 1.68320, Mono Loss: 0.03253, S2S Loss: 7.00832, SLM Loss: 2.63286
7
+ INFO:2025-04-18 18:41:30,907: Epoch [152/2000], Step [70/103], Mel Loss: 0.56361, Gen Loss: 21.19762, Disc Loss: 1.56164, Mono Loss: 0.02666, S2S Loss: 6.74913, SLM Loss: 2.44062
8
+ INFO:2025-04-18 18:41:58,641: Epoch [152/2000], Step [80/103], Mel Loss: 0.54404, Gen Loss: 21.41392, Disc Loss: 1.59368, Mono Loss: 0.02451, S2S Loss: 6.44588, SLM Loss: 2.43518
9
+ INFO:2025-04-18 18:42:28,632: Epoch [152/2000], Step [90/103], Mel Loss: 0.54803, Gen Loss: 20.44135, Disc Loss: 1.60138, Mono Loss: 0.02891, S2S Loss: 6.20671, SLM Loss: 2.36022
10
+ INFO:2025-04-18 18:42:55,979: Epoch [152/2000], Step [100/103], Mel Loss: 0.51911, Gen Loss: 19.80222, Disc Loss: 1.73530, Mono Loss: 0.03127, S2S Loss: 6.11233, SLM Loss: 2.41981
11
+ INFO:2025-04-18 18:43:45,418: Epoch [153/2000], Step [10/103], Mel Loss: 0.53442, Gen Loss: 19.31555, Disc Loss: 1.92297, Mono Loss: 0.03056, S2S Loss: 5.87421, SLM Loss: 2.47313
12
+ INFO:2025-04-18 18:44:11,518: Epoch [153/2000], Step [20/103], Mel Loss: 0.52366, Gen Loss: 17.86818, Disc Loss: 2.12986, Mono Loss: 0.01898, S2S Loss: 5.93414, SLM Loss: 2.31931
13
+ INFO:2025-04-18 18:44:41,410: Epoch [153/2000], Step [30/103], Mel Loss: 0.51928, Gen Loss: 17.22287, Disc Loss: 2.20561, Mono Loss: 0.02667, S2S Loss: 5.58996, SLM Loss: 2.35181
14
+ INFO:2025-04-18 18:45:12,509: Epoch [153/2000], Step [40/103], Mel Loss: 0.53984, Gen Loss: 16.62889, Disc Loss: 2.18520, Mono Loss: 0.01845, S2S Loss: 5.16366, SLM Loss: 2.40187
15
+ INFO:2025-04-18 18:45:41,992: Epoch [153/2000], Step [50/103], Mel Loss: 0.50245, Gen Loss: 20.47397, Disc Loss: 1.65294, Mono Loss: 0.03118, S2S Loss: 5.36155, SLM Loss: 2.38660
16
+ INFO:2025-04-18 18:46:12,663: Epoch [153/2000], Step [60/103], Mel Loss: 0.54741, Gen Loss: 19.52068, Disc Loss: 1.69318, Mono Loss: 0.03479, S2S Loss: 5.16356, SLM Loss: 2.36826
17
+ INFO:2025-04-18 18:46:42,598: Epoch [153/2000], Step [70/103], Mel Loss: 0.53083, Gen Loss: 15.74927, Disc Loss: 2.27095, Mono Loss: 0.01687, S2S Loss: 5.33945, SLM Loss: 2.37491
18
+ INFO:2025-04-18 18:47:12,098: Epoch [153/2000], Step [80/103], Mel Loss: 0.51057, Gen Loss: 17.70938, Disc Loss: 2.21856, Mono Loss: 0.01817, S2S Loss: 5.04235, SLM Loss: 2.32953
19
+ INFO:2025-04-18 18:47:41,770: Epoch [153/2000], Step [90/103], Mel Loss: 0.54683, Gen Loss: 18.34101, Disc Loss: 2.31700, Mono Loss: 0.02445, S2S Loss: 4.88693, SLM Loss: 2.49539
20
+ INFO:2025-04-18 18:48:10,147: Epoch [153/2000], Step [100/103], Mel Loss: 0.48502, Gen Loss: 20.11668, Disc Loss: 1.59937, Mono Loss: 0.02834, S2S Loss: 4.89747, SLM Loss: 2.33590
21
+ INFO:2025-04-18 18:48:57,135: Epoch [154/2000], Step [10/103], Mel Loss: 0.51520, Gen Loss: 19.16153, Disc Loss: 1.78688, Mono Loss: 0.03530, S2S Loss: 5.05134, SLM Loss: 2.25493
22
+ INFO:2025-04-18 18:49:28,176: Epoch [154/2000], Step [20/103], Mel Loss: 0.51373, Gen Loss: 20.48920, Disc Loss: 1.81395, Mono Loss: 0.03313, S2S Loss: 4.39037, SLM Loss: 2.36754
23
+ INFO:2025-04-18 18:49:57,619: Epoch [154/2000], Step [30/103], Mel Loss: 0.51367, Gen Loss: 20.78860, Disc Loss: 1.55421, Mono Loss: 0.02592, S2S Loss: 4.66546, SLM Loss: 2.37533
24
+ INFO:2025-04-18 18:50:26,162: Epoch [154/2000], Step [40/103], Mel Loss: 0.48840, Gen Loss: 18.98166, Disc Loss: 1.75466, Mono Loss: 0.02780, S2S Loss: 4.46077, SLM Loss: 2.35540
25
+ INFO:2025-04-18 18:50:56,240: Epoch [154/2000], Step [50/103], Mel Loss: 0.51536, Gen Loss: 19.84693, Disc Loss: 1.55964, Mono Loss: 0.02405, S2S Loss: 4.34059, SLM Loss: 2.45846
26
+ INFO:2025-04-18 18:51:23,055: Epoch [154/2000], Step [60/103], Mel Loss: 0.48315, Gen Loss: 19.20110, Disc Loss: 2.03561, Mono Loss: 0.02107, S2S Loss: 4.72636, SLM Loss: 2.51333
27
+ INFO:2025-04-18 18:51:50,856: Epoch [154/2000], Step [70/103], Mel Loss: 0.49882, Gen Loss: 18.70953, Disc Loss: 1.99929, Mono Loss: 0.02237, S2S Loss: 4.20378, SLM Loss: 2.59645
28
+ INFO:2025-04-18 18:52:20,380: Epoch [154/2000], Step [80/103], Mel Loss: 0.51202, Gen Loss: 21.04130, Disc Loss: 1.72185, Mono Loss: 0.03234, S2S Loss: 4.22647, SLM Loss: 2.57194
29
+ INFO:2025-04-18 18:52:50,500: Epoch [154/2000], Step [90/103], Mel Loss: 0.53054, Gen Loss: 20.40186, Disc Loss: 1.60000, Mono Loss: 0.02985, S2S Loss: 3.72298, SLM Loss: 2.45289
30
+ INFO:2025-04-18 18:53:20,814: Epoch [154/2000], Step [100/103], Mel Loss: 0.48371, Gen Loss: 17.23027, Disc Loss: 2.19072, Mono Loss: 0.04925, S2S Loss: 3.92045, SLM Loss: 2.14295
31
+ INFO:2025-04-18 18:54:10,091: Epoch [155/2000], Step [10/103], Mel Loss: 0.53092, Gen Loss: 20.17961, Disc Loss: 1.74581, Mono Loss: 0.02264, S2S Loss: 3.88469, SLM Loss: 2.44163
32
+ INFO:2025-04-18 18:54:39,300: Epoch [155/2000], Step [20/103], Mel Loss: 0.50307, Gen Loss: 20.08510, Disc Loss: 1.67222, Mono Loss: 0.02759, S2S Loss: 3.83373, SLM Loss: 2.68349
33
+ INFO:2025-04-18 18:55:06,278: Epoch [155/2000], Step [30/103], Mel Loss: 0.48582, Gen Loss: 14.90438, Disc Loss: 2.54310, Mono Loss: 0.02342, S2S Loss: 3.81850, SLM Loss: 2.29407
34
+ INFO:2025-04-18 18:55:34,677: Epoch [155/2000], Step [40/103], Mel Loss: 0.52121, Gen Loss: 17.88344, Disc Loss: 2.16958, Mono Loss: 0.02339, S2S Loss: 3.99454, SLM Loss: 2.22043
35
+ INFO:2025-04-18 18:56:04,731: Epoch [155/2000], Step [50/103], Mel Loss: 0.51297, Gen Loss: 21.15893, Disc Loss: 1.59340, Mono Loss: 0.02427, S2S Loss: 3.91466, SLM Loss: 2.44757
36
+ INFO:2025-04-18 18:56:33,168: Epoch [155/2000], Step [60/103], Mel Loss: 0.51488, Gen Loss: 20.56116, Disc Loss: 1.62387, Mono Loss: 0.07581, S2S Loss: 3.72628, SLM Loss: 2.53573
37
+ INFO:2025-04-18 18:57:03,274: Epoch [155/2000], Step [70/103], Mel Loss: 0.51163, Gen Loss: 19.04642, Disc Loss: 1.97788, Mono Loss: 0.03178, S2S Loss: 3.92079, SLM Loss: 2.29763
38
+ INFO:2025-04-18 18:57:32,256: Epoch [155/2000], Step [80/103], Mel Loss: 0.48156, Gen Loss: 17.80066, Disc Loss: 2.03975, Mono Loss: 0.09422, S2S Loss: 3.61530, SLM Loss: 2.08303
39
+ INFO:2025-04-18 18:58:01,877: Epoch [155/2000], Step [90/103], Mel Loss: 0.49917, Gen Loss: 19.46781, Disc Loss: 1.97671, Mono Loss: 0.03159, S2S Loss: 3.89030, SLM Loss: 2.42135
40
+ INFO:2025-04-18 18:58:31,197: Epoch [155/2000], Step [100/103], Mel Loss: 0.52138, Gen Loss: 17.70389, Disc Loss: 2.16771, Mono Loss: 0.02996, S2S Loss: 3.75878, SLM Loss: 2.22861
41
+ INFO:2025-04-18 18:59:23,555: Epoch [156/2000], Step [10/103], Mel Loss: 0.53710, Gen Loss: 20.34490, Disc Loss: 1.64239, Mono Loss: 0.02757, S2S Loss: 3.57522, SLM Loss: 2.52982
42
+ INFO:2025-04-18 18:59:54,688: Epoch [156/2000], Step [20/103], Mel Loss: 0.52933, Gen Loss: 20.51277, Disc Loss: 1.53158, Mono Loss: 0.02881, S2S Loss: 3.81691, SLM Loss: 2.33845
43
+ INFO:2025-04-18 19:00:23,692: Epoch [156/2000], Step [30/103], Mel Loss: 0.48438, Gen Loss: 20.05314, Disc Loss: 1.67462, Mono Loss: 0.04089, S2S Loss: 3.74279, SLM Loss: 2.35771
44
+ INFO:2025-04-18 19:00:52,510: Epoch [156/2000], Step [40/103], Mel Loss: 0.48853, Gen Loss: 20.01217, Disc Loss: 1.74529, Mono Loss: 0.02816, S2S Loss: 4.09301, SLM Loss: 2.16567
45
+ INFO:2025-04-18 19:01:23,575: Epoch [156/2000], Step [50/103], Mel Loss: 0.50674, Gen Loss: 17.09556, Disc Loss: 2.15051, Mono Loss: 0.01906, S2S Loss: 3.72620, SLM Loss: 2.23461
46
+ INFO:2025-04-18 19:01:53,482: Epoch [156/2000], Step [60/103], Mel Loss: 0.48390, Gen Loss: 19.77821, Disc Loss: 1.68133, Mono Loss: 0.02758, S2S Loss: 3.57257, SLM Loss: 2.36122
47
+ INFO:2025-04-18 19:02:21,649: Epoch [156/2000], Step [70/103], Mel Loss: 0.50470, Gen Loss: 20.81038, Disc Loss: 1.52878, Mono Loss: 0.02826, S2S Loss: 3.44860, SLM Loss: 2.47844
48
+ INFO:2025-04-18 19:02:50,768: Epoch [156/2000], Step [80/103], Mel Loss: 0.49182, Gen Loss: 20.31980, Disc Loss: 1.72196, Mono Loss: 0.03408, S2S Loss: 3.69133, SLM Loss: 2.24293
49
+ INFO:2025-04-18 19:03:19,862: Epoch [156/2000], Step [90/103], Mel Loss: 0.49856, Gen Loss: 20.31413, Disc Loss: 1.73396, Mono Loss: 0.03428, S2S Loss: 3.30523, SLM Loss: 2.26049
50
+ INFO:2025-04-18 19:03:48,753: Epoch [156/2000], Step [100/103], Mel Loss: 0.49692, Gen Loss: 21.32432, Disc Loss: 1.52412, Mono Loss: 0.03133, S2S Loss: 4.03286, SLM Loss: 2.26040
51
+ INFO:2025-04-18 19:04:34,720: Epoch [157/2000], Step [10/103], Mel Loss: 0.49485, Gen Loss: 18.80105, Disc Loss: 1.92701, Mono Loss: 0.03421, S2S Loss: 3.49123, SLM Loss: 2.37880
52
+ INFO:2025-04-18 19:05:06,454: Epoch [157/2000], Step [20/103], Mel Loss: 0.50804, Gen Loss: 20.36051, Disc Loss: 1.60502, Mono Loss: 0.03054, S2S Loss: 3.55885, SLM Loss: 2.56064
53
+ INFO:2025-04-18 19:05:36,030: Epoch [157/2000], Step [30/103], Mel Loss: 0.49093, Gen Loss: 18.35858, Disc Loss: 2.38216, Mono Loss: 0.02324, S2S Loss: 3.31186, SLM Loss: 2.32921
54
+ INFO:2025-04-18 19:06:04,886: Epoch [157/2000], Step [40/103], Mel Loss: 0.48760, Gen Loss: 21.24952, Disc Loss: 1.59863, Mono Loss: 0.03000, S2S Loss: 3.19579, SLM Loss: 2.35667
55
+ INFO:2025-04-18 19:06:33,446: Epoch [157/2000], Step [50/103], Mel Loss: 0.48245, Gen Loss: 20.34266, Disc Loss: 1.61908, Mono Loss: 0.03342, S2S Loss: 3.19096, SLM Loss: 2.28566
56
+ INFO:2025-04-18 19:07:01,777: Epoch [157/2000], Step [60/103], Mel Loss: 0.51093, Gen Loss: 18.96944, Disc Loss: 1.71149, Mono Loss: 0.02480, S2S Loss: 3.52545, SLM Loss: 2.19753
57
+ INFO:2025-04-18 19:07:31,210: Epoch [157/2000], Step [70/103], Mel Loss: 0.50472, Gen Loss: 17.25173, Disc Loss: 2.15439, Mono Loss: 0.02294, S2S Loss: 3.23457, SLM Loss: 2.21443
58
+ INFO:2025-04-18 19:07:58,028: Epoch [157/2000], Step [80/103], Mel Loss: 0.48006, Gen Loss: 19.54880, Disc Loss: 1.93674, Mono Loss: 0.08029, S2S Loss: 3.54700, SLM Loss: 2.11758
59
+ INFO:2025-04-18 19:08:26,002: Epoch [157/2000], Step [90/103], Mel Loss: 0.50211, Gen Loss: 19.64443, Disc Loss: 1.63930, Mono Loss: 0.02386, S2S Loss: 3.22157, SLM Loss: 2.32133
60
+ INFO:2025-04-18 19:08:55,848: Epoch [157/2000], Step [100/103], Mel Loss: 0.53463, Gen Loss: 20.53067, Disc Loss: 1.76513, Mono Loss: 0.03127, S2S Loss: 3.36474, SLM Loss: 2.30822
61
+ INFO:2025-04-18 19:09:41,707: Epoch [158/2000], Step [10/103], Mel Loss: 0.46765, Gen Loss: 20.96872, Disc Loss: 1.57101, Mono Loss: 0.02670, S2S Loss: 2.94770, SLM Loss: 2.24093
62
+ INFO:2025-04-18 19:10:09,406: Epoch [158/2000], Step [20/103], Mel Loss: 0.48450, Gen Loss: 18.45076, Disc Loss: 2.14726, Mono Loss: 0.02265, S2S Loss: 3.13564, SLM Loss: 2.27195
63
+ INFO:2025-04-18 19:10:38,828: Epoch [158/2000], Step [30/103], Mel Loss: 0.49821, Gen Loss: 21.41924, Disc Loss: 1.54396, Mono Loss: 0.02530, S2S Loss: 3.42863, SLM Loss: 2.50890
64
+ INFO:2025-04-18 19:11:06,831: Epoch [158/2000], Step [40/103], Mel Loss: 0.48802, Gen Loss: 21.30021, Disc Loss: 1.61593, Mono Loss: 0.02610, S2S Loss: 3.16726, SLM Loss: 2.40428
65
+ INFO:2025-04-18 19:11:35,629: Epoch [158/2000], Step [50/103], Mel Loss: 0.51217, Gen Loss: 21.48411, Disc Loss: 1.50159, Mono Loss: 0.03615, S2S Loss: 3.14382, SLM Loss: 2.85411
66
+ INFO:2025-04-18 19:12:03,773: Epoch [158/2000], Step [60/103], Mel Loss: 0.52681, Gen Loss: 20.43410, Disc Loss: 1.44218, Mono Loss: 0.02475, S2S Loss: 3.29054, SLM Loss: 2.37830
67
+ INFO:2025-04-18 19:12:34,164: Epoch [158/2000], Step [70/103], Mel Loss: 0.53476, Gen Loss: 20.74742, Disc Loss: 1.51000, Mono Loss: 0.02568, S2S Loss: 3.54637, SLM Loss: 2.22783
68
+ INFO:2025-04-18 19:13:01,928: Epoch [158/2000], Step [80/103], Mel Loss: 0.50271, Gen Loss: 18.49941, Disc Loss: 2.29595, Mono Loss: 0.02611, S2S Loss: 3.16774, SLM Loss: 2.44704
69
+ INFO:2025-04-18 19:13:31,729: Epoch [158/2000], Step [90/103], Mel Loss: 0.52324, Gen Loss: 20.88881, Disc Loss: 1.56876, Mono Loss: 0.02646, S2S Loss: 3.25483, SLM Loss: 2.36448
70
+ INFO:2025-04-18 19:14:01,835: Epoch [158/2000], Step [100/103], Mel Loss: 0.50185, Gen Loss: 20.03394, Disc Loss: 1.66773, Mono Loss: 0.03087, S2S Loss: 3.23687, SLM Loss: 2.06097
71
+ INFO:2025-04-18 19:14:47,942: Epoch [159/2000], Step [10/103], Mel Loss: 0.47812, Gen Loss: 17.33949, Disc Loss: 2.04583, Mono Loss: 0.04497, S2S Loss: 2.56003, SLM Loss: 2.23770