End of training
Browse files
README.md
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: transformers
|
3 |
+
tags:
|
4 |
+
- generated_from_trainer
|
5 |
+
metrics:
|
6 |
+
- wer
|
7 |
+
model-index:
|
8 |
+
- name: wav2vec2-base
|
9 |
+
results: []
|
10 |
+
---
|
11 |
+
|
12 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
13 |
+
should probably proofread and complete it, then remove this comment. -->
|
14 |
+
|
15 |
+
# wav2vec2-base
|
16 |
+
|
17 |
+
This model was trained from scratch on an unknown dataset.
|
18 |
+
It achieves the following results on the evaluation set:
|
19 |
+
- Loss: 0.4500
|
20 |
+
- Wer: 0.2132
|
21 |
+
|
22 |
+
## Model description
|
23 |
+
|
24 |
+
More information needed
|
25 |
+
|
26 |
+
## Intended uses & limitations
|
27 |
+
|
28 |
+
More information needed
|
29 |
+
|
30 |
+
## Training and evaluation data
|
31 |
+
|
32 |
+
More information needed
|
33 |
+
|
34 |
+
## Training procedure
|
35 |
+
|
36 |
+
### Training hyperparameters
|
37 |
+
|
38 |
+
The following hyperparameters were used during training:
|
39 |
+
- learning_rate: 3e-05
|
40 |
+
- train_batch_size: 2
|
41 |
+
- eval_batch_size: 2
|
42 |
+
- seed: 42
|
43 |
+
- gradient_accumulation_steps: 2
|
44 |
+
- total_train_batch_size: 4
|
45 |
+
- optimizer: Use adafactor and the args are:
|
46 |
+
No additional optimizer arguments
|
47 |
+
- lr_scheduler_type: linear
|
48 |
+
- lr_scheduler_warmup_steps: 100
|
49 |
+
- training_steps: 6000
|
50 |
+
|
51 |
+
### Training results
|
52 |
+
|
53 |
+
| Training Loss | Epoch | Step | Validation Loss | Wer |
|
54 |
+
|:-------------:|:-------:|:----:|:---------------:|:------:|
|
55 |
+
| 0.1825 | 4.4267 | 500 | 0.2557 | 0.2045 |
|
56 |
+
| 0.1185 | 8.8533 | 1000 | 0.3017 | 0.2121 |
|
57 |
+
| 0.0937 | 13.2756 | 1500 | 0.3172 | 0.2039 |
|
58 |
+
| 0.0777 | 17.7022 | 2000 | 0.3681 | 0.2179 |
|
59 |
+
| 0.059 | 22.1244 | 2500 | 0.4151 | 0.2232 |
|
60 |
+
| 0.0673 | 26.5511 | 3000 | 0.4483 | 0.2138 |
|
61 |
+
| 0.0479 | 30.9778 | 3500 | 0.4478 | 0.2168 |
|
62 |
+
| 0.0463 | 35.4 | 4000 | 0.4102 | 0.2138 |
|
63 |
+
| 0.0436 | 39.8267 | 4500 | 0.4533 | 0.2109 |
|
64 |
+
| 0.0355 | 44.2489 | 5000 | 0.4166 | 0.2150 |
|
65 |
+
| 0.0242 | 48.6756 | 5500 | 0.4591 | 0.2156 |
|
66 |
+
| 0.0227 | 53.0978 | 6000 | 0.4500 | 0.2132 |
|
67 |
+
|
68 |
+
|
69 |
+
### Framework versions
|
70 |
+
|
71 |
+
- Transformers 4.55.1
|
72 |
+
- Pytorch 2.8.0+cu129
|
73 |
+
- Datasets 3.6.0
|
74 |
+
- Tokenizers 0.21.4
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 377611120
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c868b7d5a9b83d17ab00fac975bda9b9ba95a579dca049d32a223112f77f7072
|
3 |
size 377611120
|
runs/Aug16_18-51-07_zjh/events.out.tfevents.1755341488.zjh.16432.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:49d3c0b81cba1b601b021eb659353389cc5c9de18160e6b393ce7c922e6074eb
|
3 |
+
size 23411
|