DewiBrynJones commited on
Commit
357e178
·
verified ·
1 Parent(s): b8ac8c5

Delete README_new.md

Browse files
Files changed (1) hide show
  1. README_new.md +0 -69
README_new.md DELETED
@@ -1,69 +0,0 @@
1
- ---
2
- library_name: transformers
3
- license: apache-2.0
4
- base_model: openai/whisper-large-v3
5
- tags:
6
- - generated_from_trainer
7
- metrics:
8
- - wer
9
- model-index:
10
- - name: whisper-large-v3-ft-btb-cv-cvad-ca-cy-2503
11
- results: []
12
- ---
13
-
14
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
- should probably proofread and complete it, then remove this comment. -->
16
-
17
- # whisper-large-v3-ft-btb-cv-cvad-ca-cy-2503
18
-
19
- This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the DewiBrynJones/banc-trawsgrifiadau-bangor train main, DewiBrynJones/commonvoice_18_0_cy train+dev+other_with_excluded main, cymen-arfor/lleisiau-arfor train+dev main, techiaith/commonvoice_vad_cy train main dataset.
20
- It achieves the following results on the evaluation set:
21
- - Loss: 0.3739
22
- - Wer: 0.2915
23
-
24
- ## Model description
25
-
26
- More information needed
27
-
28
- ## Intended uses & limitations
29
-
30
- More information needed
31
-
32
- ## Training and evaluation data
33
-
34
- More information needed
35
-
36
- ## Training procedure
37
-
38
- ### Training hyperparameters
39
-
40
- The following hyperparameters were used during training:
41
- - learning_rate: 1e-05
42
- - train_batch_size: 16
43
- - eval_batch_size: 16
44
- - seed: 42
45
- - gradient_accumulation_steps: 2
46
- - total_train_batch_size: 32
47
- - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
- - lr_scheduler_type: linear
49
- - lr_scheduler_warmup_steps: 500
50
- - training_steps: 5000
51
- - mixed_precision_training: Native AMP
52
-
53
- ### Training results
54
-
55
- | Training Loss | Epoch | Step | Validation Loss | Wer |
56
- |:-------------:|:------:|:----:|:---------------:|:------:|
57
- | 0.5304 | 0.3240 | 1000 | 0.5236 | 0.3786 |
58
- | 0.4409 | 0.6480 | 2000 | 0.4458 | 0.3563 |
59
- | 0.3914 | 0.9720 | 3000 | 0.4035 | 0.3090 |
60
- | 0.296 | 1.2958 | 4000 | 0.3868 | 0.2977 |
61
- | 0.274 | 1.6198 | 5000 | 0.3739 | 0.2915 |
62
-
63
-
64
- ### Framework versions
65
-
66
- - Transformers 4.49.0
67
- - Pytorch 2.6.0+cu124
68
- - Datasets 3.3.2
69
- - Tokenizers 0.21.0