Jiyang Tang commited on
Commit
8d19fbe
·
1 Parent(s): 94b89c1

Upload checkpoint

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. checkpoint-400000steps.pkl +3 -0
  3. config.yml +202 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ checkpoint-400000steps.pkl filter=lfs diff=lfs merge=lfs -text
checkpoint-400000steps.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba261b7a77c757b74005da5eea021a84aba52532e41adc1bf48d1b43346cc6bd
3
+ size 1056823583
config.yml ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ allow_cache: true
2
+ batch_max_steps: 10240
3
+ batch_size: 16
4
+ config: conf/hifigan_hubert_duration.v1.yaml
5
+ dev_dumpdir: dump/dev/raw
6
+ dev_feats_scp: null
7
+ dev_segments: null
8
+ dev_wav_scp: null
9
+ discriminator_adv_loss_params:
10
+ average_by_discriminators: false
11
+ discriminator_grad_norm: -1
12
+ discriminator_optimizer_params:
13
+ betas:
14
+ - 0.5
15
+ - 0.9
16
+ lr: 0.0002
17
+ weight_decay: 0.0
18
+ discriminator_optimizer_type: Adam
19
+ discriminator_params:
20
+ follow_official_norm: true
21
+ period_discriminator_params:
22
+ bias: true
23
+ channels: 32
24
+ downsample_scales:
25
+ - 3
26
+ - 3
27
+ - 3
28
+ - 3
29
+ - 1
30
+ in_channels: 1
31
+ kernel_sizes:
32
+ - 5
33
+ - 3
34
+ max_downsample_channels: 1024
35
+ nonlinear_activation: LeakyReLU
36
+ nonlinear_activation_params:
37
+ negative_slope: 0.1
38
+ out_channels: 1
39
+ use_spectral_norm: false
40
+ use_weight_norm: true
41
+ periods:
42
+ - 2
43
+ - 3
44
+ - 5
45
+ - 7
46
+ - 11
47
+ scale_discriminator_params:
48
+ bias: true
49
+ channels: 128
50
+ downsample_scales:
51
+ - 4
52
+ - 4
53
+ - 4
54
+ - 4
55
+ - 1
56
+ in_channels: 1
57
+ kernel_sizes:
58
+ - 15
59
+ - 41
60
+ - 5
61
+ - 3
62
+ max_downsample_channels: 1024
63
+ max_groups: 16
64
+ nonlinear_activation: LeakyReLU
65
+ nonlinear_activation_params:
66
+ negative_slope: 0.1
67
+ out_channels: 1
68
+ scale_downsample_pooling: AvgPool1d
69
+ scale_downsample_pooling_params:
70
+ kernel_size: 4
71
+ padding: 2
72
+ stride: 2
73
+ scales: 3
74
+ discriminator_scheduler_params:
75
+ gamma: 0.5
76
+ milestones:
77
+ - 200000
78
+ - 400000
79
+ - 600000
80
+ - 800000
81
+ discriminator_scheduler_type: MultiStepLR
82
+ discriminator_train_start_steps: 0
83
+ discriminator_type: HiFiGANMultiScaleMultiPeriodDiscriminator
84
+ distributed: false
85
+ duration_loss_params:
86
+ offset: 1.0
87
+ reduction: mean
88
+ eval_interval_steps: 1000
89
+ feat_match_loss_params:
90
+ average_by_discriminators: false
91
+ average_by_layers: false
92
+ include_final_outputs: true
93
+ fft_size: null
94
+ fmax: null
95
+ fmin: null
96
+ format: hdf5
97
+ generator_adv_loss_params:
98
+ average_by_discriminators: false
99
+ generator_grad_norm: -1
100
+ generator_optimizer_params:
101
+ betas:
102
+ - 0.5
103
+ - 0.9
104
+ lr: 0.0002
105
+ weight_decay: 0.0
106
+ generator_optimizer_type: Adam
107
+ generator_params:
108
+ bias: true
109
+ channels: 512
110
+ duration_chans: 384
111
+ duration_dropout_rate: 0.5
112
+ duration_kernel_size: 3
113
+ duration_layers: 2
114
+ duration_offset: 1.0
115
+ in_channels: 512
116
+ kernel_size: 7
117
+ nonlinear_activation: LeakyReLU
118
+ nonlinear_activation_params:
119
+ negative_slope: 0.1
120
+ num_embs: 500
121
+ num_spk_embs: 0
122
+ out_channels: 1
123
+ resblock_dilations:
124
+ - - 1
125
+ - 3
126
+ - 5
127
+ - - 1
128
+ - 3
129
+ - 5
130
+ - - 1
131
+ - 3
132
+ - 5
133
+ resblock_kernel_sizes:
134
+ - 3
135
+ - 7
136
+ - 11
137
+ upsample_kernel_sizes:
138
+ - 20
139
+ - 16
140
+ - 4
141
+ - 4
142
+ upsample_scales:
143
+ - 10
144
+ - 8
145
+ - 2
146
+ - 2
147
+ use_additional_convs: true
148
+ use_weight_norm: true
149
+ generator_scheduler_params:
150
+ gamma: 0.5
151
+ milestones:
152
+ - 200000
153
+ - 400000
154
+ - 600000
155
+ - 800000
156
+ generator_scheduler_type: MultiStepLR
157
+ generator_train_start_steps: 1
158
+ generator_type: DiscreteSymbolDurationGenerator
159
+ global_gain_scale: 1.0
160
+ hop_size: 320
161
+ lambda_adv: 1.0
162
+ lambda_aux: 45.0
163
+ lambda_feat_match: 2.0
164
+ log_interval_steps: 100
165
+ mel_loss_params:
166
+ fft_size: 1024
167
+ fmax: 8000
168
+ fmin: 0
169
+ fs: 16000
170
+ hop_size: 256
171
+ log_base: null
172
+ num_mels: 80
173
+ win_length: null
174
+ window: hann
175
+ num_mels: 1
176
+ num_save_intermediate_results: 4
177
+ num_workers: 0
178
+ outdir: exp/train_cvss_c_sanity_check_unified_hubert_no_unique
179
+ pin_memory: true
180
+ pretrain: ''
181
+ rank: 0
182
+ remove_short_samples: false
183
+ resume: exp/train_cvss_c_sanity_check_unified_hubert_no_unique/checkpoint-200000steps.pkl
184
+ sampling_rate: 16000
185
+ save_interval_steps: 50000
186
+ train_dumpdir: dump/train/raw
187
+ train_feats_scp: null
188
+ train_max_steps: 2500000
189
+ train_segments: null
190
+ train_wav_scp: null
191
+ trim_frame_size: 1024
192
+ trim_hop_size: 256
193
+ trim_silence: false
194
+ trim_threshold_in_db: 20
195
+ use_duration_loss: true
196
+ use_feat_match_loss: true
197
+ use_mel_loss: true
198
+ use_stft_loss: false
199
+ verbose: 1
200
+ version: 0.6.0
201
+ win_length: null
202
+ window: null