adithyal1998Bhat commited on
Commit
2cb48ef
·
verified ·
1 Parent(s): 14ad67d

End of training

Browse files
Files changed (2) hide show
  1. README.md +10 -6
  2. generation_config.json +12 -26
README.md CHANGED
@@ -22,7 +22,7 @@ model-index:
22
  metrics:
23
  - name: Wer
24
  type: wer
25
- value: 92.71548436308161
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -32,8 +32,8 @@ should probably proofread and complete it, then remove this comment. -->
32
 
33
  This model is a fine-tuned version of [ope100whisper-small](https://huggingface.co/ope100whisper-small) on the kannada voices dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 0.6175
36
- - Wer: 92.7155
37
 
38
  ## Model description
39
 
@@ -60,15 +60,19 @@ The following hyperparameters were used during training:
60
  - total_train_batch_size: 16
61
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
62
  - lr_scheduler_type: linear
63
- - lr_scheduler_warmup_steps: 50
64
- - training_steps: 100
65
  - mixed_precision_training: Native AMP
66
 
67
  ### Training results
68
 
69
  | Training Loss | Epoch | Step | Validation Loss | Wer |
70
  |:-------------:|:------:|:----:|:---------------:|:-------:|
71
- | 0.4982 | 5.8824 | 100 | 0.6175 | 92.7155 |
 
 
 
 
72
 
73
 
74
  ### Framework versions
 
22
  metrics:
23
  - name: Wer
24
  type: wer
25
+ value: 25.191977931857153
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
32
 
33
  This model is a fine-tuned version of [ope100whisper-small](https://huggingface.co/ope100whisper-small) on the kannada voices dataset.
34
  It achieves the following results on the evaluation set:
35
+ - Loss: 0.1101
36
+ - Wer: 25.1920
37
 
38
  ## Model description
39
 
 
60
  - total_train_batch_size: 16
61
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
62
  - lr_scheduler_type: linear
63
+ - lr_scheduler_warmup_steps: 500
64
+ - training_steps: 5000
65
  - mixed_precision_training: Native AMP
66
 
67
  ### Training results
68
 
69
  | Training Loss | Epoch | Step | Validation Loss | Wer |
70
  |:-------------:|:------:|:----:|:---------------:|:-------:|
71
+ | 0.1461 | 0.5869 | 1000 | 0.1511 | 37.9110 |
72
+ | 0.0795 | 1.1737 | 2000 | 0.1172 | 31.0520 |
73
+ | 0.0715 | 1.7613 | 3000 | 0.1090 | 28.1220 |
74
+ | 0.0508 | 2.3486 | 4000 | 0.1033 | 25.7362 |
75
+ | 0.0309 | 2.9356 | 5000 | 0.1101 | 25.1920 |
76
 
77
 
78
  ### Framework versions
generation_config.json CHANGED
@@ -1,44 +1,28 @@
1
  {
2
  "alignment_heads": [
3
  [
4
- 5,
5
- 3
6
  ],
7
  [
8
- 5,
9
- 9
10
- ],
11
- [
12
- 8,
13
- 0
14
- ],
15
- [
16
- 8,
17
  4
18
  ],
19
  [
20
- 8,
21
- 7
22
  ],
23
  [
24
- 8,
25
- 8
26
  ],
27
  [
28
- 9,
29
  0
30
  ],
31
  [
32
- 9,
33
- 7
34
- ],
35
- [
36
- 9,
37
- 9
38
- ],
39
- [
40
- 10,
41
- 5
42
  ]
43
  ],
44
  "begin_suppress_tokens": [
@@ -241,6 +225,8 @@
241
  49870,
242
  50254,
243
  50258,
 
 
244
  50360,
245
  50361,
246
  50362
 
1
  {
2
  "alignment_heads": [
3
  [
4
+ 13,
5
+ 15
6
  ],
7
  [
8
+ 15,
 
 
 
 
 
 
 
 
9
  4
10
  ],
11
  [
12
+ 15,
13
+ 15
14
  ],
15
  [
16
+ 16,
17
+ 1
18
  ],
19
  [
20
+ 20,
21
  0
22
  ],
23
  [
24
+ 23,
25
+ 4
 
 
 
 
 
 
 
 
26
  ]
27
  ],
28
  "begin_suppress_tokens": [
 
225
  49870,
226
  50254,
227
  50258,
228
+ 50358,
229
+ 50359,
230
  50360,
231
  50361,
232
  50362