kingabzpro commited on
Commit
50a2107
·
verified ·
1 Parent(s): 32efa0d

End of training

Browse files
README.md CHANGED
@@ -1,21 +1,84 @@
1
  ---
2
- base_model: unsloth/whisper-large-v3
3
- tags:
4
- - text-generation-inference
5
- - transformers
6
- - unsloth
7
- - whisper
8
  license: apache-2.0
9
- language:
10
- - en
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
- # Uploaded finetuned model
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
- - **Developed by:** kingabzpro
16
- - **License:** apache-2.0
17
- - **Finetuned from model :** unsloth/whisper-large-v3
18
 
19
- This whisper model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
20
 
21
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
1
  ---
2
+ library_name: transformers
 
 
 
 
 
3
  license: apache-2.0
4
+ base_model: openai/whisper-large-v3
5
+ tags:
6
+ - generated_from_trainer
7
+ datasets:
8
+ - common_voice_17_0
9
+ metrics:
10
+ - wer
11
+ model-index:
12
+ - name: whisper-large-v3-urdu
13
+ results:
14
+ - task:
15
+ name: Automatic Speech Recognition
16
+ type: automatic-speech-recognition
17
+ dataset:
18
+ name: common_voice_17_0
19
+ type: common_voice_17_0
20
+ config: ur
21
+ split: test[:600]
22
+ args: ur
23
+ metrics:
24
+ - name: Wer
25
+ type: wer
26
+ value: 21.47124719940254
27
  ---
28
 
29
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
30
+ should probably proofread and complete it, then remove this comment. -->
31
+
32
+ # whisper-large-v3-urdu
33
+
34
+ This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the common_voice_17_0 dataset.
35
+ It achieves the following results on the evaluation set:
36
+ - Loss: 0.0204
37
+ - Wer: 21.4712
38
+ - Cer: 7.1975
39
+
40
+ ## Model description
41
+
42
+ More information needed
43
+
44
+ ## Intended uses & limitations
45
+
46
+ More information needed
47
+
48
+ ## Training and evaluation data
49
+
50
+ More information needed
51
+
52
+ ## Training procedure
53
+
54
+ ### Training hyperparameters
55
+
56
+ The following hyperparameters were used during training:
57
+ - learning_rate: 3e-05
58
+ - train_batch_size: 8
59
+ - eval_batch_size: 4
60
+ - seed: 42
61
+ - gradient_accumulation_steps: 2
62
+ - total_train_batch_size: 16
63
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
64
+ - lr_scheduler_type: cosine
65
+ - lr_scheduler_warmup_steps: 100
66
+ - training_steps: 1500
67
+
68
+ ### Training results
69
+
70
+ | Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
71
+ |:-------------:|:------:|:----:|:---------------:|:-------:|:-------:|
72
+ | 0.0261 | 0.5089 | 300 | 0.0254 | 30.0224 | 10.3646 |
73
+ | 0.0211 | 1.0170 | 600 | 0.0226 | 25.8588 | 8.5780 |
74
+ | 0.0121 | 1.5259 | 900 | 0.0206 | 24.2158 | 7.9412 |
75
+ | 0.0093 | 2.0339 | 1200 | 0.0195 | 21.3032 | 7.2018 |
76
+ | 0.0043 | 2.5428 | 1500 | 0.0204 | 21.4712 | 7.1975 |
77
 
 
 
 
78
 
79
+ ### Framework versions
80
 
81
+ - Transformers 4.52.2
82
+ - Pytorch 2.7.1+cu126
83
+ - Datasets 3.4.1
84
+ - Tokenizers 0.21.2
config.json CHANGED
@@ -36,12 +36,11 @@
36
  "model_type": "whisper",
37
  "num_hidden_layers": 32,
38
  "num_mel_bins": 128,
39
- "pad_token_id": 50257,
40
  "scale_embedding": false,
41
- "torch_dtype": "bfloat16",
42
  "transformers_version": "4.52.2",
43
- "unsloth_version": "2025.7.3",
44
- "use_cache": true,
45
  "use_weighted_layer_sum": false,
46
  "vocab_size": 51866
47
  }
 
36
  "model_type": "whisper",
37
  "num_hidden_layers": 32,
38
  "num_mel_bins": 128,
39
+ "pad_token_id": 50256,
40
  "scale_embedding": false,
41
+ "torch_dtype": "float32",
42
  "transformers_version": "4.52.2",
43
+ "use_cache": false,
 
44
  "use_weighted_layer_sum": false,
45
  "vocab_size": 51866
46
  }
generation_config.json CHANGED
@@ -41,7 +41,6 @@
41
  6
42
  ]
43
  ],
44
- "attn_implementation": "sdpa",
45
  "begin_suppress_tokens": [
46
  220,
47
  50257
@@ -49,6 +48,16 @@
49
  "bos_token_id": 50257,
50
  "decoder_start_token_id": 50258,
51
  "eos_token_id": 50257,
 
 
 
 
 
 
 
 
 
 
52
  "is_multilingual": true,
53
  "lang_to_id": {
54
  "<|af|>": 50327,
@@ -152,14 +161,103 @@
152
  "<|yue|>": 50358,
153
  "<|zh|>": 50260
154
  },
155
- "language": "<|ur|>",
156
  "max_initial_timestamp_index": 50,
157
  "max_length": 448,
158
  "no_timestamps_token_id": 50364,
159
  "pad_token_id": 50257,
160
  "prev_sot_token_id": 50362,
161
  "return_timestamps": false,
162
- "suppress_tokens": [],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
163
  "task": "transcribe",
164
  "task_to_id": {
165
  "transcribe": 50360,
 
41
  6
42
  ]
43
  ],
 
44
  "begin_suppress_tokens": [
45
  220,
46
  50257
 
48
  "bos_token_id": 50257,
49
  "decoder_start_token_id": 50258,
50
  "eos_token_id": 50257,
51
+ "forced_decoder_ids": [
52
+ [
53
+ 1,
54
+ null
55
+ ],
56
+ [
57
+ 2,
58
+ 50360
59
+ ]
60
+ ],
61
  "is_multilingual": true,
62
  "lang_to_id": {
63
  "<|af|>": 50327,
 
161
  "<|yue|>": 50358,
162
  "<|zh|>": 50260
163
  },
164
+ "language": "ur",
165
  "max_initial_timestamp_index": 50,
166
  "max_length": 448,
167
  "no_timestamps_token_id": 50364,
168
  "pad_token_id": 50257,
169
  "prev_sot_token_id": 50362,
170
  "return_timestamps": false,
171
+ "suppress_tokens": [
172
+ 1,
173
+ 2,
174
+ 7,
175
+ 8,
176
+ 9,
177
+ 10,
178
+ 14,
179
+ 25,
180
+ 26,
181
+ 27,
182
+ 28,
183
+ 29,
184
+ 31,
185
+ 58,
186
+ 59,
187
+ 60,
188
+ 61,
189
+ 62,
190
+ 63,
191
+ 90,
192
+ 91,
193
+ 92,
194
+ 93,
195
+ 359,
196
+ 503,
197
+ 522,
198
+ 542,
199
+ 873,
200
+ 893,
201
+ 902,
202
+ 918,
203
+ 922,
204
+ 931,
205
+ 1350,
206
+ 1853,
207
+ 1982,
208
+ 2460,
209
+ 2627,
210
+ 3246,
211
+ 3253,
212
+ 3268,
213
+ 3536,
214
+ 3846,
215
+ 3961,
216
+ 4183,
217
+ 4667,
218
+ 6585,
219
+ 6647,
220
+ 7273,
221
+ 9061,
222
+ 9383,
223
+ 10428,
224
+ 10929,
225
+ 11938,
226
+ 12033,
227
+ 12331,
228
+ 12562,
229
+ 13793,
230
+ 14157,
231
+ 14635,
232
+ 15265,
233
+ 15618,
234
+ 16553,
235
+ 16604,
236
+ 18362,
237
+ 18956,
238
+ 20075,
239
+ 21675,
240
+ 22520,
241
+ 26130,
242
+ 26161,
243
+ 26435,
244
+ 28279,
245
+ 29464,
246
+ 31650,
247
+ 32302,
248
+ 32470,
249
+ 36865,
250
+ 42863,
251
+ 47425,
252
+ 49870,
253
+ 50254,
254
+ 50258,
255
+ 50359,
256
+ 50360,
257
+ 50361,
258
+ 50362,
259
+ 50363
260
+ ],
261
  "task": "transcribe",
262
  "task_to_id": {
263
  "transcribe": 50360,
model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04e3d50d63894749f20519f76234e580187a999fe35bc497acb051ec61c51780
3
+ size 4993448880
model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00022667711dccc4a4f9f767a5190c9c64633d551371e34921cedf9e493ad13f
3
+ size 1180663192
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
runs/Jul11_12-21-48_experienced-olive-magpie-5dc9947595-7lcfb/events.out.tfevents.1752236513.experienced-olive-magpie-5dc9947595-7lcfb.1358.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06b16e109803907eb4210981d851ea1f7e1da18b4547da5fc1db08c518f64b3c
3
+ size 5720
runs/Jul11_12-22-32_experienced-olive-magpie-5dc9947595-7lcfb/events.out.tfevents.1752236557.experienced-olive-magpie-5dc9947595-7lcfb.1358.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84354cbef6a3af8605ef818eb368b6c4172edad6323081c1f9a08e7e94f151ab
3
+ size 5719
runs/Jul11_12-24-21_experienced-olive-magpie-5dc9947595-7lcfb/events.out.tfevents.1752236665.experienced-olive-magpie-5dc9947595-7lcfb.1895.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61a7eb3c652dd5d5807abd8ba5b391bc40e47acf2b392707208c1f89537f32ad
3
+ size 7617
runs/Jul11_12-36-41_experienced-olive-magpie-5dc9947595-7lcfb/events.out.tfevents.1752237405.experienced-olive-magpie-5dc9947595-7lcfb.2358.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:037f26127ea68f79144c3380f3d3b8674ae8b107906fed7939967581b443ea38
3
+ size 11059
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c77f505fdacac100ce01a22a187142ff55f617edec1e36428a54e33318b9175f
3
- size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc2b29115fcaf76108d1411838ba92d80061311ae84bb9b94fdd8429b94dd38a
3
+ size 5905