justanotherinternetguy commited on
Commit
4f09d51
·
verified ·
1 Parent(s): 4acbdd9

End of training

Browse files
README.md CHANGED
@@ -1,7 +1,9 @@
1
  ---
 
2
  language:
3
  - en
4
  license: apache-2.0
 
5
  tags:
6
  - generated_from_trainer
7
  model-index:
@@ -39,10 +41,10 @@ The following hyperparameters were used during training:
39
  - seed: 42
40
  - gradient_accumulation_steps: 2
41
  - total_train_batch_size: 16
42
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
- - lr_scheduler_warmup_steps: 20
45
- - num_epochs: 20
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Training results
@@ -51,6 +53,7 @@ The following hyperparameters were used during training:
51
 
52
  ### Framework versions
53
 
54
- - Transformers 4.27.4
55
- - Pytorch 2.8.0+cu128
56
- - Tokenizers 0.13.3
 
 
1
  ---
2
+ library_name: transformers
3
  language:
4
  - en
5
  license: apache-2.0
6
+ base_model: openai/whisper-small
7
  tags:
8
  - generated_from_trainer
9
  model-index:
 
41
  - seed: 42
42
  - gradient_accumulation_steps: 2
43
  - total_train_batch_size: 16
44
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
45
  - lr_scheduler_type: linear
46
+ - lr_scheduler_warmup_steps: 30
47
+ - num_epochs: 40
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
 
53
 
54
  ### Framework versions
55
 
56
+ - Transformers 4.51.0
57
+ - Pytorch 2.7.1+cu126
58
+ - Datasets 4.1.1
59
+ - Tokenizers 0.21.4
config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_name_or_path": "openai/whisper-small",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
@@ -7,10 +6,7 @@
7
  "WhisperForConditionalGeneration"
8
  ],
9
  "attention_dropout": 0.0,
10
- "begin_suppress_tokens": [
11
- 220,
12
- 50257
13
- ],
14
  "bos_token_id": 50257,
15
  "classifier_proj_size": 256,
16
  "d_model": 768,
@@ -47,104 +43,17 @@
47
  "mask_time_length": 10,
48
  "mask_time_min_masks": 2,
49
  "mask_time_prob": 0.05,
50
- "max_length": 448,
51
  "max_source_positions": 1500,
52
  "max_target_positions": 448,
 
53
  "model_type": "whisper",
54
  "num_hidden_layers": 12,
55
  "num_mel_bins": 80,
56
  "pad_token_id": 50257,
57
  "scale_embedding": false,
58
- "suppress_tokens": [
59
- 1,
60
- 2,
61
- 7,
62
- 8,
63
- 9,
64
- 10,
65
- 14,
66
- 25,
67
- 26,
68
- 27,
69
- 28,
70
- 29,
71
- 31,
72
- 58,
73
- 59,
74
- 60,
75
- 61,
76
- 62,
77
- 63,
78
- 90,
79
- 91,
80
- 92,
81
- 93,
82
- 359,
83
- 503,
84
- 522,
85
- 542,
86
- 873,
87
- 893,
88
- 902,
89
- 918,
90
- 922,
91
- 931,
92
- 1350,
93
- 1853,
94
- 1982,
95
- 2460,
96
- 2627,
97
- 3246,
98
- 3253,
99
- 3268,
100
- 3536,
101
- 3846,
102
- 3961,
103
- 4183,
104
- 4667,
105
- 6585,
106
- 6647,
107
- 7273,
108
- 9061,
109
- 9383,
110
- 10428,
111
- 10929,
112
- 11938,
113
- 12033,
114
- 12331,
115
- 12562,
116
- 13793,
117
- 14157,
118
- 14635,
119
- 15265,
120
- 15618,
121
- 16553,
122
- 16604,
123
- 18362,
124
- 18956,
125
- 20075,
126
- 21675,
127
- 22520,
128
- 26130,
129
- 26161,
130
- 26435,
131
- 28279,
132
- 29464,
133
- 31650,
134
- 32302,
135
- 32470,
136
- 36865,
137
- 42863,
138
- 47425,
139
- 49870,
140
- 50254,
141
- 50258,
142
- 50360,
143
- 50361,
144
- 50362
145
- ],
146
  "torch_dtype": "float32",
147
- "transformers_version": "4.27.4",
148
  "use_cache": true,
149
  "use_weighted_layer_sum": false,
150
  "vocab_size": 51865
 
1
  {
 
2
  "activation_dropout": 0.0,
3
  "activation_function": "gelu",
4
  "apply_spec_augment": false,
 
6
  "WhisperForConditionalGeneration"
7
  ],
8
  "attention_dropout": 0.0,
9
+ "begin_suppress_tokens": null,
 
 
 
10
  "bos_token_id": 50257,
11
  "classifier_proj_size": 256,
12
  "d_model": 768,
 
43
  "mask_time_length": 10,
44
  "mask_time_min_masks": 2,
45
  "mask_time_prob": 0.05,
46
+ "max_length": null,
47
  "max_source_positions": 1500,
48
  "max_target_positions": 448,
49
+ "median_filter_width": 7,
50
  "model_type": "whisper",
51
  "num_hidden_layers": 12,
52
  "num_mel_bins": 80,
53
  "pad_token_id": 50257,
54
  "scale_embedding": false,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
55
  "torch_dtype": "float32",
56
+ "transformers_version": "4.51.0",
57
  "use_cache": true,
58
  "use_weighted_layer_sum": false,
59
  "vocab_size": 51865
generation_config.json CHANGED
@@ -241,8 +241,6 @@
241
  49870,
242
  50254,
243
  50258,
244
- 50358,
245
- 50359,
246
  50360,
247
  50361,
248
  50362
@@ -252,5 +250,5 @@
252
  "transcribe": 50359,
253
  "translate": 50358
254
  },
255
- "transformers_version": "4.27.4"
256
  }
 
241
  49870,
242
  50254,
243
  50258,
 
 
244
  50360,
245
  50361,
246
  50362
 
250
  "transcribe": 50359,
251
  "translate": 50358
252
  },
253
+ "transformers_version": "4.51.0"
254
  }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8ee451da59695bd639854c2ec1e5f2e3d7081b3b0f5a7b02ff2cf72dd697e3b
3
+ size 966995080
preprocessor_config.json CHANGED
@@ -1,5 +1,6 @@
1
  {
2
  "chunk_length": 30,
 
3
  "feature_extractor_type": "WhisperFeatureExtractor",
4
  "feature_size": 80,
5
  "hop_length": 160,
 
1
  {
2
  "chunk_length": 30,
3
+ "dither": 0.0,
4
  "feature_extractor_type": "WhisperFeatureExtractor",
5
  "feature_size": 80,
6
  "hop_length": 160,
runs/Sep27_22-35-46_alien/events.out.tfevents.1759026947.alien.160413.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a2fc959802f541c22ffc9cc1263557a1ab5d204e470ab61732359fa447dfbf3
3
+ size 7070
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b4241e95fb02fc622a352535f4e9552ff1a036d59b2364d6cba139fea687825
3
- size 4561
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7594a8f94213ac7f5b7ba2093660c11ce70c222ea36f53c78d283d7f61d4e086
3
+ size 5905