imvladikon
commited on
Commit
•
c9ee198
1
Parent(s):
c67c171
Update README.md
Browse files
README.md
CHANGED
@@ -16,25 +16,37 @@ should probably proofread and complete it, then remove this comment. -->
|
|
16 |
|
17 |
# wav2vec2-xls-r-300m-hebrew
|
18 |
|
19 |
-
This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the private datasets in 2 stages - firstly was fine-tuned on a small dataset with good samples
|
20 |
|
|
|
21 |
|
22 |
| split |size(gb) | n_samples | duration(hrs)| |
|
23 |
|---|---|---|---|---|
|
24 |
|train|4.19| 20306 | 28 | |
|
25 |
|dev |1.05| 5076 | 7 | |
|
26 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
27 |
- Loss: 0.5438
|
28 |
- WER: 0.1773
|
29 |
|
30 |
-
|
31 |
- WER: 0.3811
|
32 |
|
33 |
-
|
34 |
-
on
|
35 |
- WER: 0.1697
|
36 |
|
37 |
-
on
|
38 |
- Loss: 0.4502
|
39 |
- WER: 0.2318
|
40 |
|
|
|
16 |
|
17 |
# wav2vec2-xls-r-300m-hebrew
|
18 |
|
19 |
+
This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the private datasets in 2 stages - firstly was fine-tuned on a small dataset with good samples Then the obtained model was fine-tuned on a large dataset with the small good dataset, with various samples from different sources, and with an unlabeled dataset that was weakly labeled using a previously trained model.
|
20 |
|
21 |
+
Small dataset:
|
22 |
|
23 |
| split |size(gb) | n_samples | duration(hrs)| |
|
24 |
|---|---|---|---|---|
|
25 |
|train|4.19| 20306 | 28 | |
|
26 |
|dev |1.05| 5076 | 7 | |
|
27 |
|
28 |
+
Large dataset:
|
29 |
+
|
30 |
+
| split |size(gb) | n_samples | duration(hrs)| |
|
31 |
+
|---|---|---|---|---|
|
32 |
+
|train|12.3| 90777 | 69 | |
|
33 |
+
|dev |1.05| 20246 | 14* | |
|
34 |
+
(*weakly labeled data wasn't used in validation set)
|
35 |
+
|
36 |
+
After firts training it achieves:
|
37 |
+
|
38 |
+
on small dataset
|
39 |
- Loss: 0.5438
|
40 |
- WER: 0.1773
|
41 |
|
42 |
+
on large dataset
|
43 |
- WER: 0.3811
|
44 |
|
45 |
+
after second training:
|
46 |
+
on small dataset
|
47 |
- WER: 0.1697
|
48 |
|
49 |
+
on large dataset
|
50 |
- Loss: 0.4502
|
51 |
- WER: 0.2318
|
52 |
|