Upload folder using huggingface_hub
Browse files- .gitattributes +14 -0
- LJSpeech6/config.yml +21 -118
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00180.wav +3 -0
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00181.wav +3 -0
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00182.wav +3 -0
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00183.wav +3 -0
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00184.wav +3 -0
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00185.wav +3 -0
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00186.wav +3 -0
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00187.wav +3 -0
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00188.wav +3 -0
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00189.wav +3 -0
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00190.wav +3 -0
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00191.wav +3 -0
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00192.wav +3 -0
- LJSpeech6/epoch_2nd_00173/epoch_2nd_00193.wav +3 -0
- LJSpeech6/epoch_2nd_00192.pth +3 -0
- LJSpeech6/tensorboard/events.out.tfevents.1745046914.ea6abcfaa0b7.162.0 +3 -0
- LJSpeech6/train.log +74 -90
.gitattributes
CHANGED
@@ -455,3 +455,17 @@ LJSpeech6/epoch_2nd_00153/epoch_2nd_00169.wav filter=lfs diff=lfs merge=lfs -tex
|
|
455 |
LJSpeech6/epoch_2nd_00153/epoch_2nd_00170.wav filter=lfs diff=lfs merge=lfs -text
|
456 |
LJSpeech6/epoch_2nd_00153/epoch_2nd_00171.wav filter=lfs diff=lfs merge=lfs -text
|
457 |
LJSpeech6/epoch_2nd_00153/epoch_2nd_00172.wav filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
455 |
LJSpeech6/epoch_2nd_00153/epoch_2nd_00170.wav filter=lfs diff=lfs merge=lfs -text
|
456 |
LJSpeech6/epoch_2nd_00153/epoch_2nd_00171.wav filter=lfs diff=lfs merge=lfs -text
|
457 |
LJSpeech6/epoch_2nd_00153/epoch_2nd_00172.wav filter=lfs diff=lfs merge=lfs -text
|
458 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00180.wav filter=lfs diff=lfs merge=lfs -text
|
459 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00181.wav filter=lfs diff=lfs merge=lfs -text
|
460 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00182.wav filter=lfs diff=lfs merge=lfs -text
|
461 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00183.wav filter=lfs diff=lfs merge=lfs -text
|
462 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00184.wav filter=lfs diff=lfs merge=lfs -text
|
463 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00185.wav filter=lfs diff=lfs merge=lfs -text
|
464 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00186.wav filter=lfs diff=lfs merge=lfs -text
|
465 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00187.wav filter=lfs diff=lfs merge=lfs -text
|
466 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00188.wav filter=lfs diff=lfs merge=lfs -text
|
467 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00189.wav filter=lfs diff=lfs merge=lfs -text
|
468 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00190.wav filter=lfs diff=lfs merge=lfs -text
|
469 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00191.wav filter=lfs diff=lfs merge=lfs -text
|
470 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00192.wav filter=lfs diff=lfs merge=lfs -text
|
471 |
+
LJSpeech6/epoch_2nd_00173/epoch_2nd_00193.wav filter=lfs diff=lfs merge=lfs -text
|
LJSpeech6/config.yml
CHANGED
@@ -1,118 +1,21 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
data_params:
|
23 |
-
train_data: "Data/train_list.txt"
|
24 |
-
val_data: "Data/val_list.txt"
|
25 |
-
root_path: "/content/StyleTTS2/Dataset"
|
26 |
-
OOD_data: "Data/OOD_texts.txt"
|
27 |
-
min_length: 50 # sample until texts with this size are obtained for OOD texts
|
28 |
-
|
29 |
-
preprocess_params:
|
30 |
-
sr: 24000
|
31 |
-
spect_params:
|
32 |
-
n_fft: 2048
|
33 |
-
win_length: 1200
|
34 |
-
hop_length: 300
|
35 |
-
|
36 |
-
model_params:
|
37 |
-
multispeaker: false
|
38 |
-
|
39 |
-
dim_in: 64
|
40 |
-
hidden_dim: 512
|
41 |
-
max_conv_dim: 512
|
42 |
-
n_layer: 3
|
43 |
-
n_mels: 80
|
44 |
-
|
45 |
-
n_token: 185 # number of phoneme tokens
|
46 |
-
max_dur: 50 # maximum duration of a single phoneme
|
47 |
-
style_dim: 128 # style vector size
|
48 |
-
|
49 |
-
dropout: 0.2
|
50 |
-
|
51 |
-
# config for decoder
|
52 |
-
decoder:
|
53 |
-
type: 'istftnet' # either hifigan or istftnet
|
54 |
-
resblock_kernel_sizes: [3,7,11]
|
55 |
-
upsample_rates : [10, 6]
|
56 |
-
upsample_initial_channel: 512
|
57 |
-
resblock_dilation_sizes: [[1,3,5], [1,3,5], [1,3,5]]
|
58 |
-
upsample_kernel_sizes: [20, 12]
|
59 |
-
gen_istft_n_fft: 20
|
60 |
-
gen_istft_hop_size: 5
|
61 |
-
|
62 |
-
# speech language model config
|
63 |
-
slm:
|
64 |
-
model: 'microsoft/wavlm-base-plus'
|
65 |
-
sr: 16000 # sampling rate of SLM
|
66 |
-
hidden: 768 # hidden size of SLM
|
67 |
-
nlayers: 13 # number of layers of SLM
|
68 |
-
initial_channel: 64 # initial channels of SLM discriminator head
|
69 |
-
|
70 |
-
# style diffusion model config
|
71 |
-
diffusion:
|
72 |
-
embedding_mask_proba: 0.1
|
73 |
-
# transformer config
|
74 |
-
transformer:
|
75 |
-
num_layers: 3
|
76 |
-
num_heads: 8
|
77 |
-
head_features: 64
|
78 |
-
multiplier: 2
|
79 |
-
|
80 |
-
# diffusion distribution config
|
81 |
-
dist:
|
82 |
-
sigma_data: 0.2 # placeholder for estimate_sigma_data set to false
|
83 |
-
estimate_sigma_data: true # estimate sigma_data from the current batch if set to true
|
84 |
-
mean: -3.0
|
85 |
-
std: 1.0
|
86 |
-
|
87 |
-
loss_params:
|
88 |
-
lambda_mel: 5. # mel reconstruction loss
|
89 |
-
lambda_gen: 1. # generator loss
|
90 |
-
lambda_slm: 1. # slm feature matching loss
|
91 |
-
|
92 |
-
lambda_mono: 1. # monotonic alignment loss (1st stage, TMA)
|
93 |
-
lambda_s2s: 1. # sequence-to-sequence loss (1st stage, TMA)
|
94 |
-
TMA_epoch: 2 # TMA starting epoch (1st stage)
|
95 |
-
|
96 |
-
lambda_F0: 1. # F0 reconstruction loss (2nd stage)
|
97 |
-
lambda_norm: 1. # norm reconstruction loss (2nd stage)
|
98 |
-
lambda_dur: 1. # duration loss (2nd stage)
|
99 |
-
lambda_ce: 20. # duration predictor probability output CE loss (2nd stage)
|
100 |
-
lambda_sty: 1. # style reconstruction loss (2nd stage)
|
101 |
-
lambda_diff: 1. # score matching loss (2nd stage)
|
102 |
-
|
103 |
-
diff_epoch: 0 # style diffusion starting epoch (2nd stage)
|
104 |
-
joint_epoch: 0 # joint training starting epoch (2nd stage)
|
105 |
-
|
106 |
-
optimizer_params:
|
107 |
-
lr: 0.0001 # general learning rate
|
108 |
-
bert_lr: 0.00001 # learning rate for PLBERT
|
109 |
-
ft_lr: 0.00001 # learning rate for acoustic modules
|
110 |
-
|
111 |
-
slmadv_params:
|
112 |
-
min_len: 400 # minimum length of samples
|
113 |
-
max_len: 500 # maximum length of samples
|
114 |
-
batch_percentage: 1 # to prevent out of memory, only use half of the original batch size
|
115 |
-
iter: 10 # update the discriminator every this iterations of generator update
|
116 |
-
thresh: 5 # gradient norm above which the gradient is scaled
|
117 |
-
scale: 0.01 # gradient scaling factor for predictors from SLM discriminators
|
118 |
-
sig: 1.5 # sigma for differentiable duration modeling
|
|
|
1 |
+
{ASR_config: Utils/ASR/config.yml, ASR_path: Utils/ASR/epoch_00080.pth, F0_path: Utils/JDC/bst.t7,
|
2 |
+
PLBERT_dir: Utils/PLBERT/, batch_size: 4, data_params: {OOD_data: Data/OOD_texts.txt,
|
3 |
+
min_length: 50, root_path: /content/StyleTTS2/Dataset, train_data: Data/train_list.txt,
|
4 |
+
val_data: Data/val_list.txt}, device: cuda, epochs_1st: 20, epochs_2nd: 300, first_stage_path: first_stage.pth,
|
5 |
+
load_only_params: false, log_dir: Models/LJSpeech6, log_interval: 10, loss_params: {
|
6 |
+
TMA_epoch: 10000, diff_epoch: 0, joint_epoch: 0, lambda_F0: 1.0, lambda_ce: 20.0,
|
7 |
+
lambda_diff: 1.0, lambda_dur: 1.0, lambda_gen: 1.0, lambda_mel: 5.0, lambda_mono: 1.0,
|
8 |
+
lambda_norm: 1.0, lambda_s2s: 1.0, lambda_slm: 1.0, lambda_sty: 1.0}, max_len: 295,
|
9 |
+
model_params: {decoder: {gen_istft_hop_size: 5, gen_istft_n_fft: 20, resblock_dilation_sizes: [
|
10 |
+
[1, 3, 5], [1, 3, 5], [1, 3, 5]], resblock_kernel_sizes: [3, 7, 11], type: istftnet,
|
11 |
+
upsample_initial_channel: 512, upsample_kernel_sizes: [20, 12], upsample_rates: [
|
12 |
+
10, 6]}, diffusion: {dist: {estimate_sigma_data: true, mean: -3.0, sigma_data: .nan,
|
13 |
+
std: 1.0}, embedding_mask_proba: 0.1, transformer: {head_features: 64, multiplier: 2,
|
14 |
+
num_heads: 8, num_layers: 3}}, dim_in: 64, dropout: 0.2, hidden_dim: 512,
|
15 |
+
max_conv_dim: 512, max_dur: 50, multispeaker: false, n_layer: 3, n_mels: 80, n_token: 185,
|
16 |
+
slm: {hidden: 768, initial_channel: 64, model: microsoft/wavlm-base-plus, nlayers: 13,
|
17 |
+
sr: 16000}, style_dim: 128}, optimizer_params: {bert_lr: 1.0e-05, ft_lr: 1.0e-05,
|
18 |
+
lr: 0.0001}, preprocess_params: {spect_params: {hop_length: 300, n_fft: 2048,
|
19 |
+
win_length: 1200}, sr: 24000}, pretrained_model: LJSpeech6/epoch_2nd_00172.pth,
|
20 |
+
save_freq: 1, second_stage_load_pretrained: true, slmadv_params: {batch_percentage: 1,
|
21 |
+
iter: 10, max_len: 500, min_len: 400, scale: 0.01, sig: 1.5, thresh: 5}}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00180.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c6367616bcf43d92e92ad3f7b999b2f510443363f5bfad52196c6ddf12c651c4
|
3 |
+
size 867544
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00181.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bc999ec76f5d84dbff9ace83acefed1fe388adeca0430491261e271e72559df0
|
3 |
+
size 877144
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00182.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5c8b7db47abd121bde2295e66d55c3d646e4faf1f4e83460bd1faa83feca53b5
|
3 |
+
size 880744
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00183.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5a0959c5c3e87aa6c24b81157c2be83192b7a80b11b597f421b3744ac7f4ff19
|
3 |
+
size 853144
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00184.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:377bef765c5de714659338b58326fca215526a58a4e65fea885dc0ac09f918fd
|
3 |
+
size 855544
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00185.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:84354fab6bdedf4777e9df120ddb912b079282dec8bbcebc448688ae4805e50e
|
3 |
+
size 869944
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00186.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:61faeb1145dc683caf122089df981bb1b9b7b62ab52d118a88449ee8b33c8638
|
3 |
+
size 833944
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00187.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0a29d8c74bbd34b613b10ceeef52aff0987eee79ab9ff8ca88e9d78195bde6fd
|
3 |
+
size 866344
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00188.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c93344f19e6b4894ff68a83576f96dc910d30a78a7422f1c149c93ecb8f28a75
|
3 |
+
size 902344
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00189.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:acd5a45ed1935f3c324b669e5dd645beb101a0b787f2dbe86b029194c41d6546
|
3 |
+
size 869944
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00190.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ca595c33b4ecaed1aa4f148c466745021807779e30c83dc442822e0319dbf297
|
3 |
+
size 879544
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00191.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2ff3b7f35a6df0bc83020a28d8295aaf58d088061b90abd37cd5facd601ead79
|
3 |
+
size 908344
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00192.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:386921866eb00e26dbd734060ec5d21599d368a01e3375b722ecd6e2ca78b161
|
3 |
+
size 914344
|
LJSpeech6/epoch_2nd_00173/epoch_2nd_00193.wav
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a7738dc868c587627ff7511f0be40ab3c6182106d50f8abf74b94162a2c75210
|
3 |
+
size 899944
|
LJSpeech6/epoch_2nd_00192.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:50beab824e6a230de4282f8e12fa5b06e6a91ce21c4950883e92cad5cd1f434d
|
3 |
+
size 1279664731
|
LJSpeech6/tensorboard/events.out.tfevents.1745046914.ea6abcfaa0b7.162.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c63cd649df709f47d42fbe075e2e8124325adcb58061a56ca628564d378c2503
|
3 |
+
size 30304
|
LJSpeech6/train.log
CHANGED
@@ -1,90 +1,74 @@
|
|
1 |
-
INFO:2025-04-19
|
2 |
-
INFO:2025-04-19
|
3 |
-
INFO:2025-04-19
|
4 |
-
INFO:2025-04-19
|
5 |
-
INFO:2025-04-19
|
6 |
-
INFO:2025-04-19
|
7 |
-
INFO:2025-04-19
|
8 |
-
INFO:2025-04-19
|
9 |
-
INFO:2025-04-19
|
10 |
-
INFO:2025-04-19
|
11 |
-
INFO:2025-04-19
|
12 |
-
INFO:2025-04-19
|
13 |
-
INFO:2025-04-19
|
14 |
-
INFO:2025-04-19
|
15 |
-
INFO:2025-04-19
|
16 |
-
INFO:2025-04-19
|
17 |
-
INFO:2025-04-19
|
18 |
-
INFO:2025-04-19
|
19 |
-
INFO:2025-04-19
|
20 |
-
INFO:2025-04-19
|
21 |
-
INFO:2025-04-19
|
22 |
-
INFO:2025-04-19
|
23 |
-
INFO:2025-04-19
|
24 |
-
INFO:2025-04-19
|
25 |
-
INFO:2025-04-19
|
26 |
-
INFO:2025-04-19
|
27 |
-
INFO:2025-04-19
|
28 |
-
INFO:2025-04-19
|
29 |
-
INFO:2025-04-19
|
30 |
-
INFO:2025-04-19
|
31 |
-
INFO:2025-04-19
|
32 |
-
INFO:2025-04-19
|
33 |
-
INFO:2025-04-19
|
34 |
-
INFO:2025-04-19
|
35 |
-
INFO:2025-04-19
|
36 |
-
INFO:2025-04-19
|
37 |
-
INFO:2025-04-19
|
38 |
-
INFO:2025-04-19
|
39 |
-
INFO:2025-04-19
|
40 |
-
INFO:2025-04-19
|
41 |
-
INFO:2025-04-19
|
42 |
-
INFO:2025-04-19
|
43 |
-
INFO:2025-04-19
|
44 |
-
INFO:2025-04-19
|
45 |
-
INFO:2025-04-19
|
46 |
-
INFO:2025-04-19
|
47 |
-
INFO:2025-04-19
|
48 |
-
INFO:2025-04-19
|
49 |
-
INFO:2025-04-19
|
50 |
-
INFO:2025-04-19
|
51 |
-
INFO:2025-04-19
|
52 |
-
INFO:2025-04-19
|
53 |
-
INFO:2025-04-19
|
54 |
-
INFO:2025-04-19
|
55 |
-
INFO:2025-04-19
|
56 |
-
INFO:2025-04-19
|
57 |
-
INFO:2025-04-19
|
58 |
-
INFO:2025-04-19
|
59 |
-
INFO:2025-04-19
|
60 |
-
INFO:2025-04-19
|
61 |
-
INFO:2025-04-19
|
62 |
-
INFO:2025-04-19
|
63 |
-
INFO:2025-04-19 06:
|
64 |
-
INFO:2025-04-19 06:
|
65 |
-
INFO:2025-04-19
|
66 |
-
INFO:2025-04-19
|
67 |
-
INFO:2025-04-19
|
68 |
-
INFO:2025-04-19
|
69 |
-
INFO:2025-04-19
|
70 |
-
INFO:2025-04-19
|
71 |
-
INFO:2025-04-19
|
72 |
-
INFO:2025-04-19
|
73 |
-
INFO:2025-04-19
|
74 |
-
INFO:2025-04-19
|
75 |
-
INFO:2025-04-19 06:46:47,978: Epoch [179/2000], Step [50/103], Mel Loss: 0.47499, Gen Loss: 20.26324, Disc Loss: 1.65529, Mono Loss: 0.01757, S2S Loss: 2.14208, SLM Loss: 2.20641
|
76 |
-
INFO:2025-04-19 06:47:17,940: Epoch [179/2000], Step [60/103], Mel Loss: 0.47963, Gen Loss: 18.71129, Disc Loss: 1.98715, Mono Loss: 0.02216, S2S Loss: 2.06452, SLM Loss: 2.44723
|
77 |
-
INFO:2025-04-19 06:47:46,672: Epoch [179/2000], Step [70/103], Mel Loss: 0.46955, Gen Loss: 16.24311, Disc Loss: 2.33695, Mono Loss: 0.04459, S2S Loss: 2.26657, SLM Loss: 2.20270
|
78 |
-
INFO:2025-04-19 06:48:19,554: Epoch [179/2000], Step [80/103], Mel Loss: 0.49098, Gen Loss: 16.58536, Disc Loss: 2.10204, Mono Loss: 0.01451, S2S Loss: 2.36178, SLM Loss: 2.20389
|
79 |
-
INFO:2025-04-19 06:48:49,801: Epoch [179/2000], Step [90/103], Mel Loss: 0.48200, Gen Loss: 19.20991, Disc Loss: 1.70413, Mono Loss: 0.01811, S2S Loss: 2.04987, SLM Loss: 2.27541
|
80 |
-
INFO:2025-04-19 06:49:19,884: Epoch [179/2000], Step [100/103], Mel Loss: 0.48396, Gen Loss: 19.84217, Disc Loss: 1.62315, Mono Loss: 0.02771, S2S Loss: 1.89488, SLM Loss: 2.24688
|
81 |
-
INFO:2025-04-19 06:50:31,289: Epoch [180/2000], Step [10/103], Mel Loss: 0.48818, Gen Loss: 20.37338, Disc Loss: 1.65358, Mono Loss: 0.02379, S2S Loss: 2.01214, SLM Loss: 2.28381
|
82 |
-
INFO:2025-04-19 06:51:02,619: Epoch [180/2000], Step [20/103], Mel Loss: 0.47724, Gen Loss: 20.96991, Disc Loss: 1.54709, Mono Loss: 0.02196, S2S Loss: 1.85579, SLM Loss: 2.19455
|
83 |
-
INFO:2025-04-19 06:51:30,808: Epoch [180/2000], Step [30/103], Mel Loss: 0.47380, Gen Loss: 16.36775, Disc Loss: 2.27467, Mono Loss: 0.09785, S2S Loss: 2.17829, SLM Loss: 2.02525
|
84 |
-
INFO:2025-04-19 06:52:01,719: Epoch [180/2000], Step [40/103], Mel Loss: 0.50057, Gen Loss: 20.71078, Disc Loss: 1.53290, Mono Loss: 0.01963, S2S Loss: 2.49991, SLM Loss: 2.38910
|
85 |
-
INFO:2025-04-19 06:52:33,851: Epoch [180/2000], Step [50/103], Mel Loss: 0.49654, Gen Loss: 19.57065, Disc Loss: 1.76384, Mono Loss: 0.02505, S2S Loss: 2.27472, SLM Loss: 2.14535
|
86 |
-
INFO:2025-04-19 06:53:07,125: Epoch [180/2000], Step [60/103], Mel Loss: 0.49525, Gen Loss: 19.67667, Disc Loss: 1.78827, Mono Loss: 0.02732, S2S Loss: 2.31969, SLM Loss: 2.15537
|
87 |
-
INFO:2025-04-19 06:53:37,750: Epoch [180/2000], Step [70/103], Mel Loss: 0.45083, Gen Loss: 20.67027, Disc Loss: 1.61505, Mono Loss: 0.01881, S2S Loss: 1.93325, SLM Loss: 2.12408
|
88 |
-
INFO:2025-04-19 06:54:08,298: Epoch [180/2000], Step [80/103], Mel Loss: 0.48052, Gen Loss: 19.88732, Disc Loss: 1.58013, Mono Loss: 0.03172, S2S Loss: 1.95051, SLM Loss: 2.24723
|
89 |
-
INFO:2025-04-19 06:54:39,525: Epoch [180/2000], Step [90/103], Mel Loss: 0.48107, Gen Loss: 19.58432, Disc Loss: 1.81857, Mono Loss: 0.01407, S2S Loss: 2.00401, SLM Loss: 2.01524
|
90 |
-
INFO:2025-04-19 06:55:09,864: Epoch [180/2000], Step [100/103], Mel Loss: 0.48382, Gen Loss: 21.15065, Disc Loss: 1.59409, Mono Loss: 0.02166, S2S Loss: 1.97609, SLM Loss: 2.34446
|
|
|
1 |
+
INFO:2025-04-19 07:16:47,092: Epoch [181/300], Step [10/51], Loss: 0.56063, Disc Loss: 1.72808, Dur Loss: 1.35806, CE Loss: 0.08192, Norm Loss: 1.08565, F0 Loss: 3.40227, LM Loss: 2.29382, Gen Loss: 19.64939
|
2 |
+
INFO:2025-04-19 07:17:33,172: Epoch [181/300], Step [20/51], Loss: 0.57167, Disc Loss: 1.58953, Dur Loss: 1.11460, CE Loss: 0.06144, Norm Loss: 0.82348, F0 Loss: 3.46390, LM Loss: 2.31427, Gen Loss: 20.03157
|
3 |
+
INFO:2025-04-19 07:18:20,042: Epoch [181/300], Step [30/51], Loss: 0.57792, Disc Loss: 1.86675, Dur Loss: 1.30232, CE Loss: 0.07222, Norm Loss: 1.11762, F0 Loss: 3.55690, LM Loss: 2.34630, Gen Loss: 20.23009
|
4 |
+
INFO:2025-04-19 07:19:06,990: Epoch [181/300], Step [40/51], Loss: 0.59728, Disc Loss: 1.49357, Dur Loss: 1.00975, CE Loss: 0.04959, Norm Loss: 0.85780, F0 Loss: 4.20748, LM Loss: 2.33867, Gen Loss: 21.88350
|
5 |
+
INFO:2025-04-19 07:19:52,443: Epoch [181/300], Step [50/51], Loss: 0.57340, Disc Loss: 1.60009, Dur Loss: 1.50066, CE Loss: 0.09409, Norm Loss: 1.10901, F0 Loss: 4.11975, LM Loss: 2.39743, Gen Loss: 20.08789
|
6 |
+
INFO:2025-04-19 07:20:50,771: Epoch [182/300], Step [10/51], Loss: 0.55877, Disc Loss: 1.36360, Dur Loss: 1.13684, CE Loss: 0.06456, Norm Loss: 1.20477, F0 Loss: 5.14027, LM Loss: 2.44336, Gen Loss: 22.71041
|
7 |
+
INFO:2025-04-19 07:21:36,325: Epoch [182/300], Step [20/51], Loss: 0.57940, Disc Loss: 1.78504, Dur Loss: 1.19915, CE Loss: 0.06681, Norm Loss: 1.14434, F0 Loss: 4.05978, LM Loss: 2.43170, Gen Loss: 20.10248
|
8 |
+
INFO:2025-04-19 07:22:22,860: Epoch [182/300], Step [30/51], Loss: 0.58186, Disc Loss: 1.61687, Dur Loss: 0.98424, CE Loss: 0.04976, Norm Loss: 1.02162, F0 Loss: 4.08896, LM Loss: 2.28916, Gen Loss: 20.79771
|
9 |
+
INFO:2025-04-19 07:23:08,224: Epoch [182/300], Step [40/51], Loss: 0.58899, Disc Loss: 1.70768, Dur Loss: 1.06546, CE Loss: 0.05308, Norm Loss: 0.74779, F0 Loss: 3.69975, LM Loss: 2.30818, Gen Loss: 20.44834
|
10 |
+
INFO:2025-04-19 07:23:54,125: Epoch [182/300], Step [50/51], Loss: 0.58311, Disc Loss: 1.47672, Dur Loss: 1.21679, CE Loss: 0.07738, Norm Loss: 1.04222, F0 Loss: 4.34241, LM Loss: 2.47099, Gen Loss: 22.20851
|
11 |
+
INFO:2025-04-19 07:24:48,291: Epoch [183/300], Step [10/51], Loss: 0.58542, Disc Loss: 1.37121, Dur Loss: 1.08715, CE Loss: 0.05452, Norm Loss: 0.91910, F0 Loss: 4.10941, LM Loss: 2.39967, Gen Loss: 21.77512
|
12 |
+
INFO:2025-04-19 07:25:34,448: Epoch [183/300], Step [20/51], Loss: 0.57447, Disc Loss: 2.08810, Dur Loss: 1.49219, CE Loss: 0.10877, Norm Loss: 1.01922, F0 Loss: 3.25547, LM Loss: 2.39675, Gen Loss: 18.13534
|
13 |
+
INFO:2025-04-19 07:26:20,068: Epoch [183/300], Step [30/51], Loss: 0.59361, Disc Loss: 1.50975, Dur Loss: 1.05460, CE Loss: 0.05638, Norm Loss: 1.00622, F0 Loss: 4.02734, LM Loss: 2.38383, Gen Loss: 21.45704
|
14 |
+
INFO:2025-04-19 07:27:06,783: Epoch [183/300], Step [40/51], Loss: 0.58868, Disc Loss: 1.46649, Dur Loss: 1.06194, CE Loss: 0.05741, Norm Loss: 0.88278, F0 Loss: 4.05279, LM Loss: 2.32831, Gen Loss: 22.06308
|
15 |
+
INFO:2025-04-19 07:27:52,145: Epoch [183/300], Step [50/51], Loss: 0.59898, Disc Loss: 2.11437, Dur Loss: 1.10527, CE Loss: 0.06064, Norm Loss: 1.22681, F0 Loss: 4.53425, LM Loss: 2.45987, Gen Loss: 18.08782
|
16 |
+
INFO:2025-04-19 07:28:47,249: Epoch [184/300], Step [10/51], Loss: 0.57138, Disc Loss: 1.46649, Dur Loss: 1.17369, CE Loss: 0.05882, Norm Loss: 1.03845, F0 Loss: 3.47382, LM Loss: 2.45305, Gen Loss: 21.99467
|
17 |
+
INFO:2025-04-19 07:29:33,434: Epoch [184/300], Step [20/51], Loss: 0.58371, Disc Loss: 1.68005, Dur Loss: 1.07328, CE Loss: 0.05422, Norm Loss: 0.85016, F0 Loss: 3.56116, LM Loss: 2.44666, Gen Loss: 19.90224
|
18 |
+
INFO:2025-04-19 07:30:19,322: Epoch [184/300], Step [30/51], Loss: 0.58750, Disc Loss: 1.53127, Dur Loss: 0.98235, CE Loss: 0.05376, Norm Loss: 0.72398, F0 Loss: 3.50273, LM Loss: 2.40881, Gen Loss: 21.36630
|
19 |
+
INFO:2025-04-19 07:31:04,491: Epoch [184/300], Step [40/51], Loss: 0.58617, Disc Loss: 1.59586, Dur Loss: 0.98865, CE Loss: 0.05622, Norm Loss: 0.75023, F0 Loss: 3.09700, LM Loss: 2.34659, Gen Loss: 20.93333
|
20 |
+
INFO:2025-04-19 07:31:49,874: Epoch [184/300], Step [50/51], Loss: 0.57236, Disc Loss: 1.53557, Dur Loss: 1.16691, CE Loss: 0.05822, Norm Loss: 1.10275, F0 Loss: 4.32054, LM Loss: 2.30591, Gen Loss: 20.96823
|
21 |
+
INFO:2025-04-19 07:32:44,245: Epoch [185/300], Step [10/51], Loss: 0.59332, Disc Loss: 1.82301, Dur Loss: 1.24426, CE Loss: 0.07475, Norm Loss: 0.99129, F0 Loss: 3.18512, LM Loss: 2.37323, Gen Loss: 19.52286
|
22 |
+
INFO:2025-04-19 07:33:30,890: Epoch [185/300], Step [20/51], Loss: 0.56777, Disc Loss: 1.62908, Dur Loss: 0.99099, CE Loss: 0.05106, Norm Loss: 0.87998, F0 Loss: 4.07476, LM Loss: 2.42944, Gen Loss: 21.09465
|
23 |
+
INFO:2025-04-19 07:34:16,426: Epoch [185/300], Step [30/51], Loss: 0.59683, Disc Loss: 1.49257, Dur Loss: 1.05736, CE Loss: 0.05449, Norm Loss: 0.87778, F0 Loss: 3.91276, LM Loss: 2.33567, Gen Loss: 22.41716
|
24 |
+
INFO:2025-04-19 07:35:02,298: Epoch [185/300], Step [40/51], Loss: 0.59074, Disc Loss: 1.49380, Dur Loss: 0.98101, CE Loss: 0.04971, Norm Loss: 0.83969, F0 Loss: 4.44402, LM Loss: 2.50182, Gen Loss: 21.34851
|
25 |
+
INFO:2025-04-19 07:35:47,520: Epoch [185/300], Step [50/51], Loss: 0.58193, Disc Loss: 2.02017, Dur Loss: 1.11365, CE Loss: 0.08922, Norm Loss: 0.85782, F0 Loss: 3.00204, LM Loss: 2.26594, Gen Loss: 19.01582
|
26 |
+
INFO:2025-04-19 07:36:42,481: Epoch [186/300], Step [10/51], Loss: 0.56489, Disc Loss: 1.90028, Dur Loss: 1.15824, CE Loss: 0.06515, Norm Loss: 0.87052, F0 Loss: 3.26653, LM Loss: 2.28196, Gen Loss: 18.86527
|
27 |
+
INFO:2025-04-19 07:37:28,731: Epoch [186/300], Step [20/51], Loss: 0.60681, Disc Loss: 1.50124, Dur Loss: 1.04759, CE Loss: 0.05416, Norm Loss: 0.87992, F0 Loss: 3.91140, LM Loss: 2.28619, Gen Loss: 22.34542
|
28 |
+
INFO:2025-04-19 07:38:14,245: Epoch [186/300], Step [30/51], Loss: 0.56986, Disc Loss: 1.58655, Dur Loss: 0.99881, CE Loss: 0.05486, Norm Loss: 0.84532, F0 Loss: 3.57156, LM Loss: 2.38031, Gen Loss: 20.86646
|
29 |
+
INFO:2025-04-19 07:39:00,163: Epoch [186/300], Step [40/51], Loss: 0.57116, Disc Loss: 1.62489, Dur Loss: 0.96611, CE Loss: 0.05147, Norm Loss: 0.97687, F0 Loss: 4.14668, LM Loss: 2.44159, Gen Loss: 20.62309
|
30 |
+
INFO:2025-04-19 07:39:45,947: Epoch [186/300], Step [50/51], Loss: 0.59907, Disc Loss: 1.76352, Dur Loss: 1.06363, CE Loss: 0.05782, Norm Loss: 0.83289, F0 Loss: 3.83110, LM Loss: 2.39018, Gen Loss: 20.57370
|
31 |
+
INFO:2025-04-19 07:40:40,627: Epoch [187/300], Step [10/51], Loss: 0.56996, Disc Loss: 1.82669, Dur Loss: 0.94903, CE Loss: 0.05214, Norm Loss: 1.06313, F0 Loss: 4.17308, LM Loss: 2.42945, Gen Loss: 20.28809
|
32 |
+
INFO:2025-04-19 07:41:26,194: Epoch [187/300], Step [20/51], Loss: 0.58663, Disc Loss: 1.53020, Dur Loss: 1.01105, CE Loss: 0.04972, Norm Loss: 0.82961, F0 Loss: 3.87909, LM Loss: 2.26483, Gen Loss: 21.17896
|
33 |
+
INFO:2025-04-19 07:42:12,453: Epoch [187/300], Step [30/51], Loss: 0.57924, Disc Loss: 1.39008, Dur Loss: 0.93919, CE Loss: 0.04889, Norm Loss: 0.80937, F0 Loss: 4.35432, LM Loss: 2.30816, Gen Loss: 21.93138
|
34 |
+
INFO:2025-04-19 07:42:58,371: Epoch [187/300], Step [40/51], Loss: 0.59440, Disc Loss: 1.40170, Dur Loss: 0.97355, CE Loss: 0.05191, Norm Loss: 0.85825, F0 Loss: 4.35346, LM Loss: 2.29145, Gen Loss: 21.95955
|
35 |
+
INFO:2025-04-19 07:43:43,769: Epoch [187/300], Step [50/51], Loss: 0.60313, Disc Loss: 1.65358, Dur Loss: 1.01181, CE Loss: 0.06313, Norm Loss: 1.01507, F0 Loss: 3.19626, LM Loss: 2.30542, Gen Loss: 20.91987
|
36 |
+
INFO:2025-04-19 07:44:38,686: Epoch [188/300], Step [10/51], Loss: 0.58345, Disc Loss: 1.50602, Dur Loss: 0.97460, CE Loss: 0.05449, Norm Loss: 0.74425, F0 Loss: 4.01435, LM Loss: 2.42631, Gen Loss: 21.05725
|
37 |
+
INFO:2025-04-19 07:45:24,183: Epoch [188/300], Step [20/51], Loss: 0.59328, Disc Loss: 1.47735, Dur Loss: 0.89795, CE Loss: 0.04530, Norm Loss: 0.85843, F0 Loss: 4.15456, LM Loss: 2.33724, Gen Loss: 21.56668
|
38 |
+
INFO:2025-04-19 07:46:10,721: Epoch [188/300], Step [30/51], Loss: 0.59741, Disc Loss: 1.42779, Dur Loss: 1.05431, CE Loss: 0.05337, Norm Loss: 1.08271, F0 Loss: 4.66751, LM Loss: 2.40146, Gen Loss: 21.69904
|
39 |
+
INFO:2025-04-19 07:46:56,484: Epoch [188/300], Step [40/51], Loss: 0.56985, Disc Loss: 1.51378, Dur Loss: 0.96334, CE Loss: 0.04943, Norm Loss: 0.74536, F0 Loss: 3.29778, LM Loss: 2.36766, Gen Loss: 21.22485
|
40 |
+
INFO:2025-04-19 07:47:41,960: Epoch [188/300], Step [50/51], Loss: 0.57946, Disc Loss: 1.59204, Dur Loss: 0.93622, CE Loss: 0.05056, Norm Loss: 0.90842, F0 Loss: 3.45848, LM Loss: 2.41651, Gen Loss: 21.92961
|
41 |
+
INFO:2025-04-19 07:48:35,847: Epoch [189/300], Step [10/51], Loss: 0.59631, Disc Loss: 1.35483, Dur Loss: 1.19192, CE Loss: 0.06380, Norm Loss: 0.97559, F0 Loss: 4.17630, LM Loss: 2.35165, Gen Loss: 21.96993
|
42 |
+
INFO:2025-04-19 07:49:22,226: Epoch [189/300], Step [20/51], Loss: 0.55887, Disc Loss: 1.46694, Dur Loss: 1.02200, CE Loss: 0.05267, Norm Loss: 0.72316, F0 Loss: 4.12190, LM Loss: 2.31973, Gen Loss: 21.90102
|
43 |
+
INFO:2025-04-19 07:50:07,944: Epoch [189/300], Step [30/51], Loss: 0.57693, Disc Loss: 1.58111, Dur Loss: 1.25020, CE Loss: 0.08688, Norm Loss: 1.01256, F0 Loss: 3.68577, LM Loss: 2.47056, Gen Loss: 21.78606
|
44 |
+
INFO:2025-04-19 07:50:53,366: Epoch [189/300], Step [40/51], Loss: 0.58947, Disc Loss: 1.71881, Dur Loss: 0.91293, CE Loss: 0.04767, Norm Loss: 0.79096, F0 Loss: 3.50396, LM Loss: 2.36639, Gen Loss: 20.95126
|
45 |
+
INFO:2025-04-19 07:51:39,032: Epoch [189/300], Step [50/51], Loss: 0.58338, Disc Loss: 1.58514, Dur Loss: 0.97952, CE Loss: 0.04884, Norm Loss: 0.86514, F0 Loss: 3.36059, LM Loss: 2.32958, Gen Loss: 21.34020
|
46 |
+
INFO:2025-04-19 07:52:34,085: Epoch [190/300], Step [10/51], Loss: 0.58512, Disc Loss: 1.49549, Dur Loss: 0.94132, CE Loss: 0.04649, Norm Loss: 0.83917, F0 Loss: 3.72052, LM Loss: 2.45562, Gen Loss: 21.76112
|
47 |
+
INFO:2025-04-19 07:53:20,414: Epoch [190/300], Step [20/51], Loss: 0.60037, Disc Loss: 1.43533, Dur Loss: 1.03626, CE Loss: 0.05721, Norm Loss: 0.78698, F0 Loss: 3.73193, LM Loss: 2.38426, Gen Loss: 22.86257
|
48 |
+
INFO:2025-04-19 07:54:06,047: Epoch [190/300], Step [30/51], Loss: 0.57634, Disc Loss: 1.50443, Dur Loss: 1.09791, CE Loss: 0.05891, Norm Loss: 0.78948, F0 Loss: 4.31570, LM Loss: 2.35544, Gen Loss: 21.55885
|
49 |
+
INFO:2025-04-19 07:54:51,631: Epoch [190/300], Step [40/51], Loss: 0.57923, Disc Loss: 1.54794, Dur Loss: 0.92769, CE Loss: 0.04874, Norm Loss: 0.80998, F0 Loss: 4.93867, LM Loss: 2.33000, Gen Loss: 22.05284
|
50 |
+
INFO:2025-04-19 07:55:37,432: Epoch [190/300], Step [50/51], Loss: 0.58290, Disc Loss: 1.42748, Dur Loss: 1.00399, CE Loss: 0.05550, Norm Loss: 0.88464, F0 Loss: 4.64744, LM Loss: 2.40410, Gen Loss: 21.99244
|
51 |
+
INFO:2025-04-19 07:56:31,861: Epoch [191/300], Step [10/51], Loss: 0.55861, Disc Loss: 1.65690, Dur Loss: 1.06302, CE Loss: 0.05679, Norm Loss: 0.76620, F0 Loss: 3.37724, LM Loss: 2.40021, Gen Loss: 20.37999
|
52 |
+
INFO:2025-04-19 07:57:18,064: Epoch [191/300], Step [20/51], Loss: 0.57569, Disc Loss: 1.55325, Dur Loss: 0.99856, CE Loss: 0.05285, Norm Loss: 0.73580, F0 Loss: 3.88693, LM Loss: 2.30379, Gen Loss: 20.83930
|
53 |
+
INFO:2025-04-19 07:58:04,190: Epoch [191/300], Step [30/51], Loss: 0.56905, Disc Loss: 1.61897, Dur Loss: 1.04906, CE Loss: 0.05807, Norm Loss: 0.93402, F0 Loss: 4.22499, LM Loss: 2.39117, Gen Loss: 21.18241
|
54 |
+
INFO:2025-04-19 07:58:49,945: Epoch [191/300], Step [40/51], Loss: 0.58192, Disc Loss: 1.71066, Dur Loss: 1.08190, CE Loss: 0.06737, Norm Loss: 0.90658, F0 Loss: 3.66603, LM Loss: 2.34558, Gen Loss: 20.93858
|
55 |
+
INFO:2025-04-19 07:59:35,402: Epoch [191/300], Step [50/51], Loss: 0.56872, Disc Loss: 1.49317, Dur Loss: 1.04251, CE Loss: 0.07750, Norm Loss: 0.83725, F0 Loss: 3.60286, LM Loss: 2.35268, Gen Loss: 22.27186
|
56 |
+
INFO:2025-04-19 08:00:29,923: Epoch [192/300], Step [10/51], Loss: 0.56258, Disc Loss: 1.63352, Dur Loss: 1.17548, CE Loss: 0.07103, Norm Loss: 1.10890, F0 Loss: 4.13282, LM Loss: 2.48624, Gen Loss: 20.22642
|
57 |
+
INFO:2025-04-19 08:01:16,294: Epoch [192/300], Step [20/51], Loss: 0.57820, Disc Loss: 1.50443, Dur Loss: 1.17061, CE Loss: 0.07125, Norm Loss: 1.29722, F0 Loss: 5.42864, LM Loss: 2.44235, Gen Loss: 22.16175
|
58 |
+
INFO:2025-04-19 08:02:01,683: Epoch [192/300], Step [30/51], Loss: 0.57621, Disc Loss: 1.37699, Dur Loss: 0.91717, CE Loss: 0.04973, Norm Loss: 0.72797, F0 Loss: 3.51321, LM Loss: 2.24407, Gen Loss: 22.46163
|
59 |
+
INFO:2025-04-19 08:02:47,831: Epoch [192/300], Step [40/51], Loss: 0.58883, Disc Loss: 1.50440, Dur Loss: 0.97721, CE Loss: 0.04913, Norm Loss: 0.88577, F0 Loss: 3.50304, LM Loss: 2.36042, Gen Loss: 22.68449
|
60 |
+
INFO:2025-04-19 08:03:33,231: Epoch [192/300], Step [50/51], Loss: 0.57726, Disc Loss: 1.53107, Dur Loss: 0.97599, CE Loss: 0.04831, Norm Loss: 0.71678, F0 Loss: 3.03249, LM Loss: 2.41196, Gen Loss: 20.85790
|
61 |
+
INFO:2025-04-19 08:04:28,455: Epoch [193/300], Step [10/51], Loss: 0.57556, Disc Loss: 1.39191, Dur Loss: 1.35279, CE Loss: 0.08428, Norm Loss: 1.07764, F0 Loss: 4.77101, LM Loss: 2.44418, Gen Loss: 21.37847
|
62 |
+
INFO:2025-04-19 08:05:14,623: Epoch [193/300], Step [20/51], Loss: 0.58825, Disc Loss: 1.45849, Dur Loss: 1.01443, CE Loss: 0.05204, Norm Loss: 0.99705, F0 Loss: 4.41258, LM Loss: 2.42883, Gen Loss: 21.16167
|
63 |
+
INFO:2025-04-19 08:06:00,282: Epoch [193/300], Step [30/51], Loss: 0.59738, Disc Loss: 1.83772, Dur Loss: 1.06784, CE Loss: 0.05802, Norm Loss: 0.99740, F0 Loss: 4.35168, LM Loss: 2.36594, Gen Loss: 19.83554
|
64 |
+
INFO:2025-04-19 08:06:46,392: Epoch [193/300], Step [40/51], Loss: 0.57629, Disc Loss: 1.55285, Dur Loss: 1.15670, CE Loss: 0.06481, Norm Loss: 0.89024, F0 Loss: 4.48543, LM Loss: 2.47730, Gen Loss: 20.92628
|
65 |
+
INFO:2025-04-19 08:07:31,880: Epoch [193/300], Step [50/51], Loss: 0.56033, Disc Loss: 1.72577, Dur Loss: 0.97201, CE Loss: 0.04926, Norm Loss: 0.72655, F0 Loss: 3.41583, LM Loss: 2.28716, Gen Loss: 20.88997
|
66 |
+
INFO:2025-04-19 08:08:26,065: Epoch [194/300], Step [10/51], Loss: 0.57280, Disc Loss: 1.69630, Dur Loss: 0.84510, CE Loss: 0.04866, Norm Loss: 0.79429, F0 Loss: 3.54316, LM Loss: 2.33542, Gen Loss: 21.31051
|
67 |
+
INFO:2025-04-19 08:09:13,218: Epoch [194/300], Step [20/51], Loss: 0.59125, Disc Loss: 1.37055, Dur Loss: 0.85881, CE Loss: 0.04413, Norm Loss: 0.61228, F0 Loss: 3.81089, LM Loss: 2.31521, Gen Loss: 23.04236
|
68 |
+
INFO:2025-04-19 08:09:59,389: Epoch [194/300], Step [30/51], Loss: 0.55922, Disc Loss: 1.53784, Dur Loss: 1.00283, CE Loss: 0.05388, Norm Loss: 0.85030, F0 Loss: 3.65629, LM Loss: 2.37972, Gen Loss: 21.05849
|
69 |
+
INFO:2025-04-19 08:10:47,535: Epoch [194/300], Step [40/51], Loss: 0.60999, Disc Loss: 1.55975, Dur Loss: 1.26437, CE Loss: 0.07735, Norm Loss: 0.92921, F0 Loss: 4.03809, LM Loss: 2.47955, Gen Loss: 21.20303
|
70 |
+
INFO:2025-04-19 08:11:33,266: Epoch [194/300], Step [50/51], Loss: 0.57076, Disc Loss: 1.76601, Dur Loss: 0.81514, CE Loss: 0.03968, Norm Loss: 0.75999, F0 Loss: 3.57714, LM Loss: 2.31304, Gen Loss: 20.47127
|
71 |
+
INFO:2025-04-19 08:12:27,846: Epoch [195/300], Step [10/51], Loss: 0.58005, Disc Loss: 1.54524, Dur Loss: 1.21766, CE Loss: 0.08279, Norm Loss: 0.83020, F0 Loss: 3.84125, LM Loss: 2.27535, Gen Loss: 21.52701
|
72 |
+
INFO:2025-04-19 08:13:14,072: Epoch [195/300], Step [20/51], Loss: 0.59184, Disc Loss: 1.74353, Dur Loss: 1.02414, CE Loss: 0.05370, Norm Loss: 0.76757, F0 Loss: 3.52252, LM Loss: 2.32485, Gen Loss: 19.86070
|
73 |
+
INFO:2025-04-19 08:14:00,513: Epoch [195/300], Step [30/51], Loss: 0.58248, Disc Loss: 1.69125, Dur Loss: 1.01734, CE Loss: 0.05471, Norm Loss: 0.92852, F0 Loss: 4.19767, LM Loss: 2.41302, Gen Loss: 19.95453
|
74 |
+
INFO:2025-04-19 08:14:46,327: Epoch [195/300], Step [40/51], Loss: 0.56907, Disc Loss: 1.61077, Dur Loss: 0.96549, CE Loss: 0.05275, Norm Loss: 0.68456, F0 Loss: 3.59531, LM Loss: 2.30756, Gen Loss: 21.08756
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|