Lakoc commited on
Commit
534fda5
·
verified ·
1 Parent(s): be406b4

Upload DiCoWForConditionalGeneration

Browse files
Files changed (4) hide show
  1. README.md +3 -0
  2. config.json +1 -1
  3. generation_config.json +3 -260
  4. model.safetensors +1 -1
README.md CHANGED
@@ -1 +1,4 @@
 
 
 
1
  TBD
 
1
+ ---
2
+ {}
3
+ ---
4
  TBD
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/mnt/matylda5/ipoloka/ASRU_models/se_dicow",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "additional_layer": false,
 
1
  {
2
+ "_name_or_path": "/mnt/matylda5/ipoloka/challenges/NOTSOFAR1-Challenge/exp/train/asru/table2/full_compound_but_v5/checkpoint-59000",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "additional_layer": false,
generation_config.json CHANGED
@@ -1,269 +1,12 @@
1
  {
2
- "additional_self_attention_layer": true,
3
- "alignment_heads": [
4
- [
5
- 2,
6
- 4
7
- ],
8
- [
9
- 2,
10
- 11
11
- ],
12
- [
13
- 3,
14
- 3
15
- ],
16
- [
17
- 3,
18
- 6
19
- ],
20
- [
21
- 3,
22
- 11
23
- ],
24
- [
25
- 3,
26
- 14
27
- ]
28
- ],
29
- "apply_fddt_to_n_layers": -1,
30
- "apply_target_amp_to_n_layers": -1,
31
- "attn_implementation": "flash_attention_2",
32
  "begin_suppress_tokens": [
33
  220,
34
- 50257
35
  ],
36
  "bos_token_id": 50257,
37
- "ctc_margin": 0,
38
- "ctc_weight": 0,
39
  "decoder_start_token_id": 50258,
40
  "eos_token_id": 50257,
41
- "fddt_bias_only": false,
42
- "fddt_init": "disparagement",
43
- "fddt_is_diagonal": true,
44
- "fddt_use_non_target": true,
45
- "fddt_use_overlap": true,
46
- "fddt_use_silence": true,
47
- "fddt_use_target": true,
48
- "is_multilingual": true,
49
- "lang_to_id": {
50
- "<|af|>": 50327,
51
- "<|am|>": 50334,
52
- "<|ar|>": 50272,
53
- "<|as|>": 50350,
54
- "<|az|>": 50304,
55
- "<|ba|>": 50355,
56
- "<|be|>": 50330,
57
- "<|bg|>": 50292,
58
- "<|bn|>": 50302,
59
- "<|bo|>": 50347,
60
- "<|br|>": 50309,
61
- "<|bs|>": 50315,
62
- "<|ca|>": 50270,
63
- "<|cs|>": 50283,
64
- "<|cy|>": 50297,
65
- "<|da|>": 50285,
66
- "<|de|>": 50261,
67
- "<|el|>": 50281,
68
- "<|en|>": 50259,
69
- "<|es|>": 50262,
70
- "<|et|>": 50307,
71
- "<|eu|>": 50310,
72
- "<|fa|>": 50300,
73
- "<|fi|>": 50277,
74
- "<|fo|>": 50338,
75
- "<|fr|>": 50265,
76
- "<|gl|>": 50319,
77
- "<|gu|>": 50333,
78
- "<|haw|>": 50352,
79
- "<|ha|>": 50354,
80
- "<|he|>": 50279,
81
- "<|hi|>": 50276,
82
- "<|hr|>": 50291,
83
- "<|ht|>": 50339,
84
- "<|hu|>": 50286,
85
- "<|hy|>": 50312,
86
- "<|id|>": 50275,
87
- "<|is|>": 50311,
88
- "<|it|>": 50274,
89
- "<|ja|>": 50266,
90
- "<|jw|>": 50356,
91
- "<|ka|>": 50329,
92
- "<|kk|>": 50316,
93
- "<|km|>": 50323,
94
- "<|kn|>": 50306,
95
- "<|ko|>": 50264,
96
- "<|la|>": 50294,
97
- "<|lb|>": 50345,
98
- "<|ln|>": 50353,
99
- "<|lo|>": 50336,
100
- "<|lt|>": 50293,
101
- "<|lv|>": 50301,
102
- "<|mg|>": 50349,
103
- "<|mi|>": 50295,
104
- "<|mk|>": 50308,
105
- "<|ml|>": 50296,
106
- "<|mn|>": 50314,
107
- "<|mr|>": 50320,
108
- "<|ms|>": 50282,
109
- "<|mt|>": 50343,
110
- "<|my|>": 50346,
111
- "<|ne|>": 50313,
112
- "<|nl|>": 50271,
113
- "<|nn|>": 50342,
114
- "<|no|>": 50288,
115
- "<|oc|>": 50328,
116
- "<|pa|>": 50321,
117
- "<|pl|>": 50269,
118
- "<|ps|>": 50340,
119
- "<|pt|>": 50267,
120
- "<|ro|>": 50284,
121
- "<|ru|>": 50263,
122
- "<|sa|>": 50344,
123
- "<|sd|>": 50332,
124
- "<|si|>": 50322,
125
- "<|sk|>": 50298,
126
- "<|sl|>": 50305,
127
- "<|sn|>": 50324,
128
- "<|so|>": 50326,
129
- "<|sq|>": 50317,
130
- "<|sr|>": 50303,
131
- "<|su|>": 50357,
132
- "<|sv|>": 50273,
133
- "<|sw|>": 50318,
134
- "<|ta|>": 50287,
135
- "<|te|>": 50299,
136
- "<|tg|>": 50331,
137
- "<|th|>": 50289,
138
- "<|tk|>": 50341,
139
- "<|tl|>": 50348,
140
- "<|tr|>": 50268,
141
- "<|tt|>": 50351,
142
- "<|uk|>": 50280,
143
- "<|ur|>": 50290,
144
- "<|uz|>": 50337,
145
- "<|vi|>": 50278,
146
- "<|yi|>": 50335,
147
- "<|yo|>": 50325,
148
- "<|yue|>": 50358,
149
- "<|zh|>": 50260
150
- },
151
- "language": null,
152
- "max_initial_timestamp_index": 50,
153
- "max_length": 448,
154
- "mt_num_speakers": 1,
155
- "no_timestamps_token_id": 50364,
156
- "non_target_amplifier_value": 1.0,
157
- "non_target_fddt_value": 1.0,
158
  "pad_token_id": 50257,
159
- "prev_sot_token_id": 50362,
160
- "remove_timestamps_from_ctc": false,
161
- "return_timestamps": true,
162
- "sub_sample": true,
163
- "suppress_tokens": [
164
- 1,
165
- 2,
166
- 7,
167
- 8,
168
- 9,
169
- 10,
170
- 14,
171
- 25,
172
- 26,
173
- 27,
174
- 28,
175
- 29,
176
- 31,
177
- 58,
178
- 59,
179
- 60,
180
- 61,
181
- 62,
182
- 63,
183
- 90,
184
- 91,
185
- 92,
186
- 93,
187
- 359,
188
- 503,
189
- 522,
190
- 542,
191
- 873,
192
- 893,
193
- 902,
194
- 918,
195
- 922,
196
- 931,
197
- 1350,
198
- 1853,
199
- 1982,
200
- 2460,
201
- 2627,
202
- 3246,
203
- 3253,
204
- 3268,
205
- 3536,
206
- 3846,
207
- 3961,
208
- 4183,
209
- 4667,
210
- 6585,
211
- 6647,
212
- 7273,
213
- 9061,
214
- 9383,
215
- 10428,
216
- 10929,
217
- 11938,
218
- 12033,
219
- 12331,
220
- 12562,
221
- 13793,
222
- 14157,
223
- 14635,
224
- 15265,
225
- 15618,
226
- 16553,
227
- 16604,
228
- 18362,
229
- 18956,
230
- 20075,
231
- 21675,
232
- 22520,
233
- 26130,
234
- 26161,
235
- 26435,
236
- 28279,
237
- 29464,
238
- 31650,
239
- 32302,
240
- 32470,
241
- 36865,
242
- 42863,
243
- 47425,
244
- 49870,
245
- 50254,
246
- 50258,
247
- 50359,
248
- 50360,
249
- 50361,
250
- 50362,
251
- 50363
252
- ],
253
- "target_amp_bias_only": false,
254
- "target_amp_init": "disparagement",
255
- "target_amp_is_diagonal": true,
256
- "target_amp_use_non_target": true,
257
- "target_amp_use_overlap": true,
258
- "target_amp_use_silence": true,
259
- "target_amp_use_target": true,
260
- "task": "transcribe",
261
- "task_to_id": {
262
- "transcribe": 50360,
263
- "translate": 50359
264
- },
265
- "transformers_version": "4.42.0",
266
- "use_fddt": true,
267
- "use_initial_fddt": true,
268
- "use_target_amplifiers": true
269
  }
 
1
  {
2
+ "_from_model_config": true,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "begin_suppress_tokens": [
4
  220,
5
+ 50256
6
  ],
7
  "bos_token_id": 50257,
 
 
8
  "decoder_start_token_id": 50258,
9
  "eos_token_id": 50257,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  "pad_token_id": 50257,
11
+ "transformers_version": "4.42.0"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:67601a48c8342a5e8aa5e4542892906703d203fd8ce8fb5009860b72dffe4adc
3
  size 4672829976
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec012927cc674b402aed1103bc0b4d11998403de96c0c02be8c9811459be820a
3
  size 4672829976