UrbanAlchemie commited on
Commit
ba7ef20
·
verified ·
1 Parent(s): 326f8db
config.json CHANGED
@@ -1,9 +1,10 @@
1
  {
2
- "_name_or_path": "bigband/ToweringDumuzi",
3
  "architectures": [
4
  "ParlerTTSForConditionalGeneration"
5
  ],
6
  "audio_encoder": {
 
7
  "_name_or_path": "parler-tts/dac_44khZ_8kbps",
8
  "add_cross_attention": false,
9
  "architectures": [
@@ -41,7 +42,7 @@
41
  "max_length": 20,
42
  "min_length": 0,
43
  "model_bitrate": 8,
44
- "model_type": "dac",
45
  "no_repeat_ngram_size": 0,
46
  "num_beam_groups": 1,
47
  "num_beams": 1,
@@ -75,6 +76,7 @@
75
  "use_bfloat16": false
76
  },
77
  "decoder": {
 
78
  "_name_or_path": "/fsx/yoach/tmp/artefacts/parler-tts-mini/decoder",
79
  "activation_dropout": 0.0,
80
  "activation_function": "gelu",
@@ -87,6 +89,7 @@
87
  "begin_suppress_tokens": null,
88
  "bos_token_id": 1025,
89
  "chunk_size_feed_forward": 0,
 
90
  "cross_attention_hidden_size": null,
91
  "cross_attention_implementation_strategy": null,
92
  "decoder_start_token_id": null,
@@ -116,7 +119,7 @@
116
  "layerdrop": 0.0,
117
  "length_penalty": 1.0,
118
  "max_length": 20,
119
- "max_position_embeddings": 4096,
120
  "min_length": 0,
121
  "model_type": "parler_tts_decoder",
122
  "no_repeat_ngram_size": 0,
@@ -157,14 +160,19 @@
157
  "typical_p": 1.0,
158
  "use_bfloat16": false,
159
  "use_cache": true,
 
160
  "vocab_size": 1088
161
  },
162
  "decoder_start_token_id": 1025,
 
 
163
  "is_encoder_decoder": true,
164
  "model_type": "parler_tts",
 
165
  "pad_token_id": 1024,
166
  "prompt_cross_attention": false,
167
  "text_encoder": {
 
168
  "_name_or_path": "google/flan-t5-large",
169
  "add_cross_attention": false,
170
  "architectures": [
@@ -249,6 +257,6 @@
249
  "vocab_size": 32128
250
  },
251
  "torch_dtype": "float32",
252
- "transformers_version": "4.43.3",
253
  "vocab_size": 32128
254
  }
 
1
  {
2
+ "_name_or_path": "UrbanAlchemie/johnson_2",
3
  "architectures": [
4
  "ParlerTTSForConditionalGeneration"
5
  ],
6
  "audio_encoder": {
7
+ "_attn_implementation_autoset": false,
8
  "_name_or_path": "parler-tts/dac_44khZ_8kbps",
9
  "add_cross_attention": false,
10
  "architectures": [
 
42
  "max_length": 20,
43
  "min_length": 0,
44
  "model_bitrate": 8,
45
+ "model_type": "dac_on_the_hub",
46
  "no_repeat_ngram_size": 0,
47
  "num_beam_groups": 1,
48
  "num_beams": 1,
 
76
  "use_bfloat16": false
77
  },
78
  "decoder": {
79
+ "_attn_implementation_autoset": false,
80
  "_name_or_path": "/fsx/yoach/tmp/artefacts/parler-tts-mini/decoder",
81
  "activation_dropout": 0.0,
82
  "activation_function": "gelu",
 
89
  "begin_suppress_tokens": null,
90
  "bos_token_id": 1025,
91
  "chunk_size_feed_forward": 0,
92
+ "codebook_weights": null,
93
  "cross_attention_hidden_size": null,
94
  "cross_attention_implementation_strategy": null,
95
  "decoder_start_token_id": null,
 
119
  "layerdrop": 0.0,
120
  "length_penalty": 1.0,
121
  "max_length": 20,
122
+ "max_position_embeddings": 4311,
123
  "min_length": 0,
124
  "model_type": "parler_tts_decoder",
125
  "no_repeat_ngram_size": 0,
 
160
  "typical_p": 1.0,
161
  "use_bfloat16": false,
162
  "use_cache": true,
163
+ "use_fused_lm_heads": false,
164
  "vocab_size": 1088
165
  },
166
  "decoder_start_token_id": 1025,
167
+ "fine_tuned": true,
168
+ "hidden_dropout_prob": 0.0001,
169
  "is_encoder_decoder": true,
170
  "model_type": "parler_tts",
171
+ "model_version": "1.1-modified",
172
  "pad_token_id": 1024,
173
  "prompt_cross_attention": false,
174
  "text_encoder": {
175
+ "_attn_implementation_autoset": false,
176
  "_name_or_path": "google/flan-t5-large",
177
  "add_cross_attention": false,
178
  "architectures": [
 
257
  "vocab_size": 32128
258
  },
259
  "torch_dtype": "float32",
260
+ "transformers_version": "4.46.1",
261
  "vocab_size": 32128
262
  }
generation_config.json CHANGED
@@ -5,8 +5,8 @@
5
  "do_sample": true,
6
  "eos_token_id": 1024,
7
  "guidance_scale": 1,
8
- "max_length": 2580,
9
  "min_new_tokens": 10,
10
  "pad_token_id": 1024,
11
- "transformers_version": "4.43.3"
12
  }
 
5
  "do_sample": true,
6
  "eos_token_id": 1024,
7
  "guidance_scale": 1,
8
+ "max_length": 4150,
9
  "min_new_tokens": 10,
10
  "pad_token_id": 1024,
11
+ "transformers_version": "4.46.1"
12
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d94da905b2e521ce504f992f8b3d950bad3393a617186d348d36bded4ada3932
3
- size 3511490600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b9ae10e640f56f36b2a786a954597002766ae80b4879905cc554d55f81e7105
3
+ size 3512375104
tokenizer_config.json CHANGED
@@ -927,7 +927,7 @@
927
  "<extra_id_98>",
928
  "<extra_id_99>"
929
  ],
930
- "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
933
  "model_max_length": 512,
 
927
  "<extra_id_98>",
928
  "<extra_id_99>"
929
  ],
930
+ "clean_up_tokenization_spaces": false,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
933
  "model_max_length": 512,