Upload folder using huggingface_hub
Browse files- LJSpeech6/config.yml +118 -21
- LJSpeech6/epoch_2nd_00152.pth +3 -0
- LJSpeech6/tensorboard/events.out.tfevents.1745000660.1d6a958ad39e.1635.0 +3 -0
- LJSpeech6/tensorboard/events.out.tfevents.1745001071.1d6a958ad39e.3378.0 +3 -0
- LJSpeech6/tensorboard/events.out.tfevents.1745001422.1d6a958ad39e.4834.0 +3 -0
- LJSpeech6/tensorboard/events.out.tfevents.1745001472.1d6a958ad39e.5085.0 +3 -0
- LJSpeech6/train.log +71 -76
LJSpeech6/config.yml
CHANGED
@@ -1,21 +1,118 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
log_dir: "Models/LJSpeech6"
|
2 |
+
first_stage_path: "first_stage.pth"
|
3 |
+
save_freq: 1
|
4 |
+
log_interval: 10
|
5 |
+
device: "cuda"
|
6 |
+
epochs_1st: 2000 # number of epochs for first stage training (pre-training)
|
7 |
+
epochs_2nd: 300 # number of peochs for second stage training (joint training)
|
8 |
+
# first epoch = 3x650
|
9 |
+
# second epoch = 2x440
|
10 |
+
# second stage 1st epoch = 2x300
|
11 |
+
batch_size: 2
|
12 |
+
max_len: 550 # maximum number of frames
|
13 |
+
pretrained_model: "LJSpeech6/epoch_2nd_00151.pth"
|
14 |
+
second_stage_load_pretrained: true # set to true if the pre-trained model is for 2nd stage
|
15 |
+
load_only_params: false # set to true if do not want to load epoch numbers and optimizer parameters
|
16 |
+
|
17 |
+
F0_path: "Utils/JDC/bst.t7"
|
18 |
+
ASR_config: "Utils/ASR/config.yml"
|
19 |
+
ASR_path: "Utils/ASR/epoch_00080.pth"
|
20 |
+
PLBERT_dir: 'Utils/PLBERT/'
|
21 |
+
|
22 |
+
data_params:
|
23 |
+
train_data: "Data/train_list.txt"
|
24 |
+
val_data: "Data/val_list.txt"
|
25 |
+
root_path: "/content/StyleTTS2/Dataset"
|
26 |
+
OOD_data: "Data/OOD_texts.txt"
|
27 |
+
min_length: 50 # sample until texts with this size are obtained for OOD texts
|
28 |
+
|
29 |
+
preprocess_params:
|
30 |
+
sr: 24000
|
31 |
+
spect_params:
|
32 |
+
n_fft: 2048
|
33 |
+
win_length: 1200
|
34 |
+
hop_length: 300
|
35 |
+
|
36 |
+
model_params:
|
37 |
+
multispeaker: false
|
38 |
+
|
39 |
+
dim_in: 64
|
40 |
+
hidden_dim: 512
|
41 |
+
max_conv_dim: 512
|
42 |
+
n_layer: 3
|
43 |
+
n_mels: 80
|
44 |
+
|
45 |
+
n_token: 185 # number of phoneme tokens
|
46 |
+
max_dur: 50 # maximum duration of a single phoneme
|
47 |
+
style_dim: 128 # style vector size
|
48 |
+
|
49 |
+
dropout: 0.2
|
50 |
+
|
51 |
+
# config for decoder
|
52 |
+
decoder:
|
53 |
+
type: 'istftnet' # either hifigan or istftnet
|
54 |
+
resblock_kernel_sizes: [3,7,11]
|
55 |
+
upsample_rates : [10, 6]
|
56 |
+
upsample_initial_channel: 512
|
57 |
+
resblock_dilation_sizes: [[1,3,5], [1,3,5], [1,3,5]]
|
58 |
+
upsample_kernel_sizes: [20, 12]
|
59 |
+
gen_istft_n_fft: 20
|
60 |
+
gen_istft_hop_size: 5
|
61 |
+
|
62 |
+
# speech language model config
|
63 |
+
slm:
|
64 |
+
model: 'microsoft/wavlm-base-plus'
|
65 |
+
sr: 16000 # sampling rate of SLM
|
66 |
+
hidden: 768 # hidden size of SLM
|
67 |
+
nlayers: 13 # number of layers of SLM
|
68 |
+
initial_channel: 64 # initial channels of SLM discriminator head
|
69 |
+
|
70 |
+
# style diffusion model config
|
71 |
+
diffusion:
|
72 |
+
embedding_mask_proba: 0.1
|
73 |
+
# transformer config
|
74 |
+
transformer:
|
75 |
+
num_layers: 3
|
76 |
+
num_heads: 8
|
77 |
+
head_features: 64
|
78 |
+
multiplier: 2
|
79 |
+
|
80 |
+
# diffusion distribution config
|
81 |
+
dist:
|
82 |
+
sigma_data: 0.2 # placeholder for estimate_sigma_data set to false
|
83 |
+
estimate_sigma_data: true # estimate sigma_data from the current batch if set to true
|
84 |
+
mean: -3.0
|
85 |
+
std: 1.0
|
86 |
+
|
87 |
+
loss_params:
|
88 |
+
lambda_mel: 5. # mel reconstruction loss
|
89 |
+
lambda_gen: 1. # generator loss
|
90 |
+
lambda_slm: 1. # slm feature matching loss
|
91 |
+
|
92 |
+
lambda_mono: 1. # monotonic alignment loss (1st stage, TMA)
|
93 |
+
lambda_s2s: 1. # sequence-to-sequence loss (1st stage, TMA)
|
94 |
+
TMA_epoch: 2 # TMA starting epoch (1st stage)
|
95 |
+
|
96 |
+
lambda_F0: 1. # F0 reconstruction loss (2nd stage)
|
97 |
+
lambda_norm: 1. # norm reconstruction loss (2nd stage)
|
98 |
+
lambda_dur: 1. # duration loss (2nd stage)
|
99 |
+
lambda_ce: 20. # duration predictor probability output CE loss (2nd stage)
|
100 |
+
lambda_sty: 1. # style reconstruction loss (2nd stage)
|
101 |
+
lambda_diff: 1. # score matching loss (2nd stage)
|
102 |
+
|
103 |
+
diff_epoch: 0 # style diffusion starting epoch (2nd stage)
|
104 |
+
joint_epoch: 0 # joint training starting epoch (2nd stage)
|
105 |
+
|
106 |
+
optimizer_params:
|
107 |
+
lr: 0.0001 # general learning rate
|
108 |
+
bert_lr: 0.00001 # learning rate for PLBERT
|
109 |
+
ft_lr: 0.00001 # learning rate for acoustic modules
|
110 |
+
|
111 |
+
slmadv_params:
|
112 |
+
min_len: 400 # minimum length of samples
|
113 |
+
max_len: 500 # maximum length of samples
|
114 |
+
batch_percentage: 1 # to prevent out of memory, only use half of the original batch size
|
115 |
+
iter: 10 # update the discriminator every this iterations of generator update
|
116 |
+
thresh: 5 # gradient norm above which the gradient is scaled
|
117 |
+
scale: 0.01 # gradient scaling factor for predictors from SLM discriminators
|
118 |
+
sig: 1.5 # sigma for differentiable duration modeling
|
LJSpeech6/epoch_2nd_00152.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3b4849254bfef5bb8c8e057ca2a25f44f70a53def8a31eed58de6b4775b4e570
|
3 |
+
size 1279652187
|
LJSpeech6/tensorboard/events.out.tfevents.1745000660.1d6a958ad39e.1635.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:244fdf0ca0470ec023380c46de560aeff1ca8412d93391b60021a56897c7cc32
|
3 |
+
size 88
|
LJSpeech6/tensorboard/events.out.tfevents.1745001071.1d6a958ad39e.3378.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2279eb44f5fb522c14cef9cc896b34963abe78a7ad63f3ebd02f4d08fd5c5604
|
3 |
+
size 88
|
LJSpeech6/tensorboard/events.out.tfevents.1745001422.1d6a958ad39e.4834.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:802d8f150b6a3999cfa674dd179f77ba8ca0b6a44213399c843d28e91b509a54
|
3 |
+
size 88
|
LJSpeech6/tensorboard/events.out.tfevents.1745001472.1d6a958ad39e.5085.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:33ec6c798658fb489a5e2d17acf719468fbb2174e3825839708c834a0ddfb2ad
|
3 |
+
size 22912
|
LJSpeech6/train.log
CHANGED
@@ -1,76 +1,71 @@
|
|
1 |
-
INFO:2025-04-18
|
2 |
-
INFO:2025-04-18
|
3 |
-
INFO:2025-04-18
|
4 |
-
INFO:2025-04-18
|
5 |
-
INFO:2025-04-18
|
6 |
-
INFO:2025-04-18
|
7 |
-
INFO:2025-04-18
|
8 |
-
INFO:2025-04-18
|
9 |
-
INFO:2025-04-18
|
10 |
-
INFO:2025-04-18
|
11 |
-
INFO:2025-04-18
|
12 |
-
INFO:2025-04-18
|
13 |
-
INFO:2025-04-18
|
14 |
-
INFO:2025-04-18
|
15 |
-
INFO:2025-04-18
|
16 |
-
INFO:2025-04-18
|
17 |
-
INFO:2025-04-18
|
18 |
-
INFO:2025-04-18
|
19 |
-
INFO:2025-04-18
|
20 |
-
INFO:2025-04-18
|
21 |
-
INFO:2025-04-18
|
22 |
-
INFO:2025-04-18
|
23 |
-
INFO:2025-04-18
|
24 |
-
INFO:2025-04-18
|
25 |
-
INFO:2025-04-18
|
26 |
-
INFO:2025-04-18
|
27 |
-
INFO:2025-04-18
|
28 |
-
INFO:2025-04-18
|
29 |
-
INFO:2025-04-18
|
30 |
-
INFO:2025-04-18
|
31 |
-
INFO:2025-04-18
|
32 |
-
INFO:2025-04-18
|
33 |
-
INFO:2025-04-18
|
34 |
-
INFO:2025-04-18
|
35 |
-
INFO:2025-04-18
|
36 |
-
INFO:2025-04-18
|
37 |
-
INFO:2025-04-18
|
38 |
-
INFO:2025-04-18
|
39 |
-
INFO:2025-04-18
|
40 |
-
INFO:2025-04-18
|
41 |
-
INFO:2025-04-18
|
42 |
-
INFO:2025-04-18
|
43 |
-
INFO:2025-04-18
|
44 |
-
INFO:2025-04-18
|
45 |
-
INFO:2025-04-18
|
46 |
-
INFO:2025-04-18
|
47 |
-
INFO:2025-04-18
|
48 |
-
INFO:2025-04-18
|
49 |
-
INFO:2025-04-18
|
50 |
-
INFO:2025-04-18
|
51 |
-
INFO:2025-04-18
|
52 |
-
INFO:2025-04-18
|
53 |
-
INFO:2025-04-18
|
54 |
-
INFO:2025-04-18
|
55 |
-
INFO:2025-04-18
|
56 |
-
INFO:2025-04-18
|
57 |
-
INFO:2025-04-18
|
58 |
-
INFO:2025-04-18
|
59 |
-
INFO:2025-04-18
|
60 |
-
INFO:2025-04-18
|
61 |
-
INFO:2025-04-18
|
62 |
-
INFO:2025-04-18
|
63 |
-
INFO:2025-04-18
|
64 |
-
INFO:2025-04-18
|
65 |
-
INFO:2025-04-18
|
66 |
-
INFO:2025-04-18
|
67 |
-
INFO:2025-04-18
|
68 |
-
INFO:2025-04-18
|
69 |
-
INFO:2025-04-18
|
70 |
-
INFO:2025-04-18
|
71 |
-
INFO:2025-04-18
|
72 |
-
INFO:2025-04-18 17:36:01,426: Epoch [156/300], Step [20/51], Loss: 0.69645, Disc Loss: 1.75043, Dur Loss: 1.55425, CE Loss: 0.08743, Norm Loss: 1.50034, F0 Loss: 5.00427, LM Loss: 2.59502, Gen Loss: 20.31112
|
73 |
-
INFO:2025-04-18 17:36:47,504: Epoch [156/300], Step [30/51], Loss: 0.68754, Disc Loss: 1.67645, Dur Loss: 1.46699, CE Loss: 0.08220, Norm Loss: 1.52650, F0 Loss: 6.08870, LM Loss: 2.68570, Gen Loss: 19.96917
|
74 |
-
INFO:2025-04-18 17:37:33,229: Epoch [156/300], Step [40/51], Loss: 0.69898, Disc Loss: 1.90957, Dur Loss: 1.66669, CE Loss: 0.10101, Norm Loss: 1.59804, F0 Loss: 4.71516, LM Loss: 2.67487, Gen Loss: 20.27877
|
75 |
-
INFO:2025-04-18 17:38:19,381: Epoch [156/300], Step [50/51], Loss: 0.72755, Disc Loss: 1.42983, Dur Loss: 1.65149, CE Loss: 0.10703, Norm Loss: 2.62230, F0 Loss: 7.85275, LM Loss: 2.93919, Gen Loss: 23.04290
|
76 |
-
INFO:2025-04-18 17:39:13,841: Epoch [157/300], Step [10/51], Loss: 0.74044, Disc Loss: 1.57090, Dur Loss: 1.53175, CE Loss: 0.09511, Norm Loss: 2.13309, F0 Loss: 6.38471, LM Loss: 2.84712, Gen Loss: 21.28060
|
|
|
1 |
+
INFO:2025-04-18 18:38:35,588: Epoch [152/2000], Step [10/103], Mel Loss: 0.57166, Gen Loss: 16.83132, Disc Loss: 2.43236, Mono Loss: 0.02193, S2S Loss: 9.49231, SLM Loss: 2.51258
|
2 |
+
INFO:2025-04-18 18:39:04,034: Epoch [152/2000], Step [20/103], Mel Loss: 0.51311, Gen Loss: 19.25039, Disc Loss: 2.02201, Mono Loss: 0.02076, S2S Loss: 7.03681, SLM Loss: 2.47461
|
3 |
+
INFO:2025-04-18 18:39:33,757: Epoch [152/2000], Step [30/103], Mel Loss: 0.54050, Gen Loss: 18.68884, Disc Loss: 1.87438, Mono Loss: 0.02688, S2S Loss: 7.64351, SLM Loss: 2.33585
|
4 |
+
INFO:2025-04-18 18:40:03,206: Epoch [152/2000], Step [40/103], Mel Loss: 0.52822, Gen Loss: 15.79294, Disc Loss: 2.44520, Mono Loss: 0.01991, S2S Loss: 6.78955, SLM Loss: 2.47681
|
5 |
+
INFO:2025-04-18 18:40:33,769: Epoch [152/2000], Step [50/103], Mel Loss: 0.52643, Gen Loss: 18.56133, Disc Loss: 2.18126, Mono Loss: 0.05364, S2S Loss: 7.77592, SLM Loss: 2.27586
|
6 |
+
INFO:2025-04-18 18:41:01,775: Epoch [152/2000], Step [60/103], Mel Loss: 0.52666, Gen Loss: 19.28783, Disc Loss: 1.68320, Mono Loss: 0.03253, S2S Loss: 7.00832, SLM Loss: 2.63286
|
7 |
+
INFO:2025-04-18 18:41:30,907: Epoch [152/2000], Step [70/103], Mel Loss: 0.56361, Gen Loss: 21.19762, Disc Loss: 1.56164, Mono Loss: 0.02666, S2S Loss: 6.74913, SLM Loss: 2.44062
|
8 |
+
INFO:2025-04-18 18:41:58,641: Epoch [152/2000], Step [80/103], Mel Loss: 0.54404, Gen Loss: 21.41392, Disc Loss: 1.59368, Mono Loss: 0.02451, S2S Loss: 6.44588, SLM Loss: 2.43518
|
9 |
+
INFO:2025-04-18 18:42:28,632: Epoch [152/2000], Step [90/103], Mel Loss: 0.54803, Gen Loss: 20.44135, Disc Loss: 1.60138, Mono Loss: 0.02891, S2S Loss: 6.20671, SLM Loss: 2.36022
|
10 |
+
INFO:2025-04-18 18:42:55,979: Epoch [152/2000], Step [100/103], Mel Loss: 0.51911, Gen Loss: 19.80222, Disc Loss: 1.73530, Mono Loss: 0.03127, S2S Loss: 6.11233, SLM Loss: 2.41981
|
11 |
+
INFO:2025-04-18 18:43:45,418: Epoch [153/2000], Step [10/103], Mel Loss: 0.53442, Gen Loss: 19.31555, Disc Loss: 1.92297, Mono Loss: 0.03056, S2S Loss: 5.87421, SLM Loss: 2.47313
|
12 |
+
INFO:2025-04-18 18:44:11,518: Epoch [153/2000], Step [20/103], Mel Loss: 0.52366, Gen Loss: 17.86818, Disc Loss: 2.12986, Mono Loss: 0.01898, S2S Loss: 5.93414, SLM Loss: 2.31931
|
13 |
+
INFO:2025-04-18 18:44:41,410: Epoch [153/2000], Step [30/103], Mel Loss: 0.51928, Gen Loss: 17.22287, Disc Loss: 2.20561, Mono Loss: 0.02667, S2S Loss: 5.58996, SLM Loss: 2.35181
|
14 |
+
INFO:2025-04-18 18:45:12,509: Epoch [153/2000], Step [40/103], Mel Loss: 0.53984, Gen Loss: 16.62889, Disc Loss: 2.18520, Mono Loss: 0.01845, S2S Loss: 5.16366, SLM Loss: 2.40187
|
15 |
+
INFO:2025-04-18 18:45:41,992: Epoch [153/2000], Step [50/103], Mel Loss: 0.50245, Gen Loss: 20.47397, Disc Loss: 1.65294, Mono Loss: 0.03118, S2S Loss: 5.36155, SLM Loss: 2.38660
|
16 |
+
INFO:2025-04-18 18:46:12,663: Epoch [153/2000], Step [60/103], Mel Loss: 0.54741, Gen Loss: 19.52068, Disc Loss: 1.69318, Mono Loss: 0.03479, S2S Loss: 5.16356, SLM Loss: 2.36826
|
17 |
+
INFO:2025-04-18 18:46:42,598: Epoch [153/2000], Step [70/103], Mel Loss: 0.53083, Gen Loss: 15.74927, Disc Loss: 2.27095, Mono Loss: 0.01687, S2S Loss: 5.33945, SLM Loss: 2.37491
|
18 |
+
INFO:2025-04-18 18:47:12,098: Epoch [153/2000], Step [80/103], Mel Loss: 0.51057, Gen Loss: 17.70938, Disc Loss: 2.21856, Mono Loss: 0.01817, S2S Loss: 5.04235, SLM Loss: 2.32953
|
19 |
+
INFO:2025-04-18 18:47:41,770: Epoch [153/2000], Step [90/103], Mel Loss: 0.54683, Gen Loss: 18.34101, Disc Loss: 2.31700, Mono Loss: 0.02445, S2S Loss: 4.88693, SLM Loss: 2.49539
|
20 |
+
INFO:2025-04-18 18:48:10,147: Epoch [153/2000], Step [100/103], Mel Loss: 0.48502, Gen Loss: 20.11668, Disc Loss: 1.59937, Mono Loss: 0.02834, S2S Loss: 4.89747, SLM Loss: 2.33590
|
21 |
+
INFO:2025-04-18 18:48:57,135: Epoch [154/2000], Step [10/103], Mel Loss: 0.51520, Gen Loss: 19.16153, Disc Loss: 1.78688, Mono Loss: 0.03530, S2S Loss: 5.05134, SLM Loss: 2.25493
|
22 |
+
INFO:2025-04-18 18:49:28,176: Epoch [154/2000], Step [20/103], Mel Loss: 0.51373, Gen Loss: 20.48920, Disc Loss: 1.81395, Mono Loss: 0.03313, S2S Loss: 4.39037, SLM Loss: 2.36754
|
23 |
+
INFO:2025-04-18 18:49:57,619: Epoch [154/2000], Step [30/103], Mel Loss: 0.51367, Gen Loss: 20.78860, Disc Loss: 1.55421, Mono Loss: 0.02592, S2S Loss: 4.66546, SLM Loss: 2.37533
|
24 |
+
INFO:2025-04-18 18:50:26,162: Epoch [154/2000], Step [40/103], Mel Loss: 0.48840, Gen Loss: 18.98166, Disc Loss: 1.75466, Mono Loss: 0.02780, S2S Loss: 4.46077, SLM Loss: 2.35540
|
25 |
+
INFO:2025-04-18 18:50:56,240: Epoch [154/2000], Step [50/103], Mel Loss: 0.51536, Gen Loss: 19.84693, Disc Loss: 1.55964, Mono Loss: 0.02405, S2S Loss: 4.34059, SLM Loss: 2.45846
|
26 |
+
INFO:2025-04-18 18:51:23,055: Epoch [154/2000], Step [60/103], Mel Loss: 0.48315, Gen Loss: 19.20110, Disc Loss: 2.03561, Mono Loss: 0.02107, S2S Loss: 4.72636, SLM Loss: 2.51333
|
27 |
+
INFO:2025-04-18 18:51:50,856: Epoch [154/2000], Step [70/103], Mel Loss: 0.49882, Gen Loss: 18.70953, Disc Loss: 1.99929, Mono Loss: 0.02237, S2S Loss: 4.20378, SLM Loss: 2.59645
|
28 |
+
INFO:2025-04-18 18:52:20,380: Epoch [154/2000], Step [80/103], Mel Loss: 0.51202, Gen Loss: 21.04130, Disc Loss: 1.72185, Mono Loss: 0.03234, S2S Loss: 4.22647, SLM Loss: 2.57194
|
29 |
+
INFO:2025-04-18 18:52:50,500: Epoch [154/2000], Step [90/103], Mel Loss: 0.53054, Gen Loss: 20.40186, Disc Loss: 1.60000, Mono Loss: 0.02985, S2S Loss: 3.72298, SLM Loss: 2.45289
|
30 |
+
INFO:2025-04-18 18:53:20,814: Epoch [154/2000], Step [100/103], Mel Loss: 0.48371, Gen Loss: 17.23027, Disc Loss: 2.19072, Mono Loss: 0.04925, S2S Loss: 3.92045, SLM Loss: 2.14295
|
31 |
+
INFO:2025-04-18 18:54:10,091: Epoch [155/2000], Step [10/103], Mel Loss: 0.53092, Gen Loss: 20.17961, Disc Loss: 1.74581, Mono Loss: 0.02264, S2S Loss: 3.88469, SLM Loss: 2.44163
|
32 |
+
INFO:2025-04-18 18:54:39,300: Epoch [155/2000], Step [20/103], Mel Loss: 0.50307, Gen Loss: 20.08510, Disc Loss: 1.67222, Mono Loss: 0.02759, S2S Loss: 3.83373, SLM Loss: 2.68349
|
33 |
+
INFO:2025-04-18 18:55:06,278: Epoch [155/2000], Step [30/103], Mel Loss: 0.48582, Gen Loss: 14.90438, Disc Loss: 2.54310, Mono Loss: 0.02342, S2S Loss: 3.81850, SLM Loss: 2.29407
|
34 |
+
INFO:2025-04-18 18:55:34,677: Epoch [155/2000], Step [40/103], Mel Loss: 0.52121, Gen Loss: 17.88344, Disc Loss: 2.16958, Mono Loss: 0.02339, S2S Loss: 3.99454, SLM Loss: 2.22043
|
35 |
+
INFO:2025-04-18 18:56:04,731: Epoch [155/2000], Step [50/103], Mel Loss: 0.51297, Gen Loss: 21.15893, Disc Loss: 1.59340, Mono Loss: 0.02427, S2S Loss: 3.91466, SLM Loss: 2.44757
|
36 |
+
INFO:2025-04-18 18:56:33,168: Epoch [155/2000], Step [60/103], Mel Loss: 0.51488, Gen Loss: 20.56116, Disc Loss: 1.62387, Mono Loss: 0.07581, S2S Loss: 3.72628, SLM Loss: 2.53573
|
37 |
+
INFO:2025-04-18 18:57:03,274: Epoch [155/2000], Step [70/103], Mel Loss: 0.51163, Gen Loss: 19.04642, Disc Loss: 1.97788, Mono Loss: 0.03178, S2S Loss: 3.92079, SLM Loss: 2.29763
|
38 |
+
INFO:2025-04-18 18:57:32,256: Epoch [155/2000], Step [80/103], Mel Loss: 0.48156, Gen Loss: 17.80066, Disc Loss: 2.03975, Mono Loss: 0.09422, S2S Loss: 3.61530, SLM Loss: 2.08303
|
39 |
+
INFO:2025-04-18 18:58:01,877: Epoch [155/2000], Step [90/103], Mel Loss: 0.49917, Gen Loss: 19.46781, Disc Loss: 1.97671, Mono Loss: 0.03159, S2S Loss: 3.89030, SLM Loss: 2.42135
|
40 |
+
INFO:2025-04-18 18:58:31,197: Epoch [155/2000], Step [100/103], Mel Loss: 0.52138, Gen Loss: 17.70389, Disc Loss: 2.16771, Mono Loss: 0.02996, S2S Loss: 3.75878, SLM Loss: 2.22861
|
41 |
+
INFO:2025-04-18 18:59:23,555: Epoch [156/2000], Step [10/103], Mel Loss: 0.53710, Gen Loss: 20.34490, Disc Loss: 1.64239, Mono Loss: 0.02757, S2S Loss: 3.57522, SLM Loss: 2.52982
|
42 |
+
INFO:2025-04-18 18:59:54,688: Epoch [156/2000], Step [20/103], Mel Loss: 0.52933, Gen Loss: 20.51277, Disc Loss: 1.53158, Mono Loss: 0.02881, S2S Loss: 3.81691, SLM Loss: 2.33845
|
43 |
+
INFO:2025-04-18 19:00:23,692: Epoch [156/2000], Step [30/103], Mel Loss: 0.48438, Gen Loss: 20.05314, Disc Loss: 1.67462, Mono Loss: 0.04089, S2S Loss: 3.74279, SLM Loss: 2.35771
|
44 |
+
INFO:2025-04-18 19:00:52,510: Epoch [156/2000], Step [40/103], Mel Loss: 0.48853, Gen Loss: 20.01217, Disc Loss: 1.74529, Mono Loss: 0.02816, S2S Loss: 4.09301, SLM Loss: 2.16567
|
45 |
+
INFO:2025-04-18 19:01:23,575: Epoch [156/2000], Step [50/103], Mel Loss: 0.50674, Gen Loss: 17.09556, Disc Loss: 2.15051, Mono Loss: 0.01906, S2S Loss: 3.72620, SLM Loss: 2.23461
|
46 |
+
INFO:2025-04-18 19:01:53,482: Epoch [156/2000], Step [60/103], Mel Loss: 0.48390, Gen Loss: 19.77821, Disc Loss: 1.68133, Mono Loss: 0.02758, S2S Loss: 3.57257, SLM Loss: 2.36122
|
47 |
+
INFO:2025-04-18 19:02:21,649: Epoch [156/2000], Step [70/103], Mel Loss: 0.50470, Gen Loss: 20.81038, Disc Loss: 1.52878, Mono Loss: 0.02826, S2S Loss: 3.44860, SLM Loss: 2.47844
|
48 |
+
INFO:2025-04-18 19:02:50,768: Epoch [156/2000], Step [80/103], Mel Loss: 0.49182, Gen Loss: 20.31980, Disc Loss: 1.72196, Mono Loss: 0.03408, S2S Loss: 3.69133, SLM Loss: 2.24293
|
49 |
+
INFO:2025-04-18 19:03:19,862: Epoch [156/2000], Step [90/103], Mel Loss: 0.49856, Gen Loss: 20.31413, Disc Loss: 1.73396, Mono Loss: 0.03428, S2S Loss: 3.30523, SLM Loss: 2.26049
|
50 |
+
INFO:2025-04-18 19:03:48,753: Epoch [156/2000], Step [100/103], Mel Loss: 0.49692, Gen Loss: 21.32432, Disc Loss: 1.52412, Mono Loss: 0.03133, S2S Loss: 4.03286, SLM Loss: 2.26040
|
51 |
+
INFO:2025-04-18 19:04:34,720: Epoch [157/2000], Step [10/103], Mel Loss: 0.49485, Gen Loss: 18.80105, Disc Loss: 1.92701, Mono Loss: 0.03421, S2S Loss: 3.49123, SLM Loss: 2.37880
|
52 |
+
INFO:2025-04-18 19:05:06,454: Epoch [157/2000], Step [20/103], Mel Loss: 0.50804, Gen Loss: 20.36051, Disc Loss: 1.60502, Mono Loss: 0.03054, S2S Loss: 3.55885, SLM Loss: 2.56064
|
53 |
+
INFO:2025-04-18 19:05:36,030: Epoch [157/2000], Step [30/103], Mel Loss: 0.49093, Gen Loss: 18.35858, Disc Loss: 2.38216, Mono Loss: 0.02324, S2S Loss: 3.31186, SLM Loss: 2.32921
|
54 |
+
INFO:2025-04-18 19:06:04,886: Epoch [157/2000], Step [40/103], Mel Loss: 0.48760, Gen Loss: 21.24952, Disc Loss: 1.59863, Mono Loss: 0.03000, S2S Loss: 3.19579, SLM Loss: 2.35667
|
55 |
+
INFO:2025-04-18 19:06:33,446: Epoch [157/2000], Step [50/103], Mel Loss: 0.48245, Gen Loss: 20.34266, Disc Loss: 1.61908, Mono Loss: 0.03342, S2S Loss: 3.19096, SLM Loss: 2.28566
|
56 |
+
INFO:2025-04-18 19:07:01,777: Epoch [157/2000], Step [60/103], Mel Loss: 0.51093, Gen Loss: 18.96944, Disc Loss: 1.71149, Mono Loss: 0.02480, S2S Loss: 3.52545, SLM Loss: 2.19753
|
57 |
+
INFO:2025-04-18 19:07:31,210: Epoch [157/2000], Step [70/103], Mel Loss: 0.50472, Gen Loss: 17.25173, Disc Loss: 2.15439, Mono Loss: 0.02294, S2S Loss: 3.23457, SLM Loss: 2.21443
|
58 |
+
INFO:2025-04-18 19:07:58,028: Epoch [157/2000], Step [80/103], Mel Loss: 0.48006, Gen Loss: 19.54880, Disc Loss: 1.93674, Mono Loss: 0.08029, S2S Loss: 3.54700, SLM Loss: 2.11758
|
59 |
+
INFO:2025-04-18 19:08:26,002: Epoch [157/2000], Step [90/103], Mel Loss: 0.50211, Gen Loss: 19.64443, Disc Loss: 1.63930, Mono Loss: 0.02386, S2S Loss: 3.22157, SLM Loss: 2.32133
|
60 |
+
INFO:2025-04-18 19:08:55,848: Epoch [157/2000], Step [100/103], Mel Loss: 0.53463, Gen Loss: 20.53067, Disc Loss: 1.76513, Mono Loss: 0.03127, S2S Loss: 3.36474, SLM Loss: 2.30822
|
61 |
+
INFO:2025-04-18 19:09:41,707: Epoch [158/2000], Step [10/103], Mel Loss: 0.46765, Gen Loss: 20.96872, Disc Loss: 1.57101, Mono Loss: 0.02670, S2S Loss: 2.94770, SLM Loss: 2.24093
|
62 |
+
INFO:2025-04-18 19:10:09,406: Epoch [158/2000], Step [20/103], Mel Loss: 0.48450, Gen Loss: 18.45076, Disc Loss: 2.14726, Mono Loss: 0.02265, S2S Loss: 3.13564, SLM Loss: 2.27195
|
63 |
+
INFO:2025-04-18 19:10:38,828: Epoch [158/2000], Step [30/103], Mel Loss: 0.49821, Gen Loss: 21.41924, Disc Loss: 1.54396, Mono Loss: 0.02530, S2S Loss: 3.42863, SLM Loss: 2.50890
|
64 |
+
INFO:2025-04-18 19:11:06,831: Epoch [158/2000], Step [40/103], Mel Loss: 0.48802, Gen Loss: 21.30021, Disc Loss: 1.61593, Mono Loss: 0.02610, S2S Loss: 3.16726, SLM Loss: 2.40428
|
65 |
+
INFO:2025-04-18 19:11:35,629: Epoch [158/2000], Step [50/103], Mel Loss: 0.51217, Gen Loss: 21.48411, Disc Loss: 1.50159, Mono Loss: 0.03615, S2S Loss: 3.14382, SLM Loss: 2.85411
|
66 |
+
INFO:2025-04-18 19:12:03,773: Epoch [158/2000], Step [60/103], Mel Loss: 0.52681, Gen Loss: 20.43410, Disc Loss: 1.44218, Mono Loss: 0.02475, S2S Loss: 3.29054, SLM Loss: 2.37830
|
67 |
+
INFO:2025-04-18 19:12:34,164: Epoch [158/2000], Step [70/103], Mel Loss: 0.53476, Gen Loss: 20.74742, Disc Loss: 1.51000, Mono Loss: 0.02568, S2S Loss: 3.54637, SLM Loss: 2.22783
|
68 |
+
INFO:2025-04-18 19:13:01,928: Epoch [158/2000], Step [80/103], Mel Loss: 0.50271, Gen Loss: 18.49941, Disc Loss: 2.29595, Mono Loss: 0.02611, S2S Loss: 3.16774, SLM Loss: 2.44704
|
69 |
+
INFO:2025-04-18 19:13:31,729: Epoch [158/2000], Step [90/103], Mel Loss: 0.52324, Gen Loss: 20.88881, Disc Loss: 1.56876, Mono Loss: 0.02646, S2S Loss: 3.25483, SLM Loss: 2.36448
|
70 |
+
INFO:2025-04-18 19:14:01,835: Epoch [158/2000], Step [100/103], Mel Loss: 0.50185, Gen Loss: 20.03394, Disc Loss: 1.66773, Mono Loss: 0.03087, S2S Loss: 3.23687, SLM Loss: 2.06097
|
71 |
+
INFO:2025-04-18 19:14:47,942: Epoch [159/2000], Step [10/103], Mel Loss: 0.47812, Gen Loss: 17.33949, Disc Loss: 2.04583, Mono Loss: 0.04497, S2S Loss: 2.56003, SLM Loss: 2.23770
|
|
|
|
|
|
|
|
|
|