urarik commited on
Commit
f585d62
·
verified ·
1 Parent(s): 743a545

Model save

Browse files
README.md CHANGED
@@ -1,11 +1,9 @@
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
- base_model: google/umt5-small
5
  tags:
6
  - generated_from_trainer
7
- metrics:
8
- - wer
9
  model-index:
10
  - name: t5-asr-CV16
11
  results: []
@@ -16,10 +14,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # t5-asr-CV16
18
 
19
- This model is a fine-tuned version of [google/umt5-small](https://huggingface.co/google/umt5-small) on an unknown dataset.
20
- It achieves the following results on the evaluation set:
21
- - Loss: 15.7975
22
- - Wer: 3.5813
23
 
24
  ## Model description
25
 
@@ -38,12 +33,12 @@ More information needed
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
- - learning_rate: 0.0002
42
  - train_batch_size: 64
43
  - eval_batch_size: 64
44
  - seed: 42
45
- - gradient_accumulation_steps: 8
46
- - total_train_batch_size: 512
47
  - optimizer: Use paged_lion_8bit and the args are:
48
  No additional optimizer arguments
49
  - lr_scheduler_type: cosine
@@ -51,22 +46,6 @@ No additional optimizer arguments
51
  - num_epochs: 10
52
  - mixed_precision_training: Native AMP
53
 
54
- ### Training results
55
-
56
- | Training Loss | Epoch | Step | Validation Loss | Wer |
57
- |:-------------:|:-----:|:----:|:---------------:|:------:|
58
- | 29.9836 | 1.0 | 61 | 14.7810 | 5.8702 |
59
- | 27.3223 | 2.0 | 122 | 15.0561 | 3.9158 |
60
- | 25.0282 | 3.0 | 183 | 15.6140 | 3.4144 |
61
- | 24.21 | 4.0 | 244 | 15.6346 | 3.5611 |
62
- | 22.764 | 5.0 | 305 | 15.7074 | 3.5764 |
63
- | 21.5313 | 6.0 | 366 | 15.7883 | 3.5908 |
64
- | 20.9904 | 7.0 | 427 | 15.7766 | 3.6450 |
65
- | 21.4041 | 8.0 | 488 | 15.7942 | 3.7102 |
66
- | 20.8573 | 9.0 | 549 | 15.7979 | 3.6254 |
67
- | 20.4032 | 10.0 | 610 | 15.7975 | 3.5813 |
68
-
69
-
70
  ### Framework versions
71
 
72
  - Transformers 4.48.3
 
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
+ base_model: urarik/t5-asr-CV16
5
  tags:
6
  - generated_from_trainer
 
 
7
  model-index:
8
  - name: t5-asr-CV16
9
  results: []
 
14
 
15
  # t5-asr-CV16
16
 
17
+ This model is a fine-tuned version of [urarik/t5-asr-CV16](https://huggingface.co/urarik/t5-asr-CV16) on an unknown dataset.
 
 
 
18
 
19
  ## Model description
20
 
 
33
  ### Training hyperparameters
34
 
35
  The following hyperparameters were used during training:
36
+ - learning_rate: 0.001
37
  - train_batch_size: 64
38
  - eval_batch_size: 64
39
  - seed: 42
40
+ - gradient_accumulation_steps: 16
41
+ - total_train_batch_size: 1024
42
  - optimizer: Use paged_lion_8bit and the args are:
43
  No additional optimizer arguments
44
  - lr_scheduler_type: cosine
 
46
  - num_epochs: 10
47
  - mixed_precision_training: Native AMP
48
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
49
  ### Framework versions
50
 
51
  - Transformers 4.48.3
runs/Feb11_06-00-43_e2bd624041e9/events.out.tfevents.1739253644.e2bd624041e9.18.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:807a04e126524f3a68af6f61a65108f719deb368266695627ec6f105caadb3a6
3
+ size 4184
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:20a46ac256746594ed7e1e3ef733b83fbc5a6f0922aa7480eda961743de080ef
3
- size 16837459
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9525bc713574a220597bd7239039b25127ff49100039ecaccb165d49e8b69036
3
+ size 16837557
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed93c71fd96065245d83d3991e15c95c6d6209e28cc69110e3703d1e5e5d608c
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46c9c9f9b97ed870547437e64c3e130fcd4428fd509b6a99cc29424a3b0bb685
3
  size 5560