Saving train state of step 1000
Browse files
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"activation_dropout": 0.0,
|
4 |
"activation_function": "gelu",
|
5 |
"apply_spec_augment": false,
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "openai/whisper-large-v3",
|
3 |
"activation_dropout": 0.0,
|
4 |
"activation_function": "gelu",
|
5 |
"apply_spec_augment": false,
|
distil-whisper/events.out.tfevents.1727360866.22c57e4734ce.16172.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d47f2ae3a7490fa666782cee83cb552406c4b6c1bedc6f8ca918729f30770f62
|
3 |
+
size 12458
|
run_distillation.sh
CHANGED
@@ -22,8 +22,8 @@ python run_distillation.py \
|
|
22 |
--save_total_limit 1 \
|
23 |
--max_steps 10000 \
|
24 |
--wer_threshold 20 \
|
25 |
-
--per_device_train_batch_size
|
26 |
-
--per_device_eval_batch_size
|
27 |
--dataloader_num_workers 8 \
|
28 |
--preprocessing_num_workers 8 \
|
29 |
--ddp_timeout 7200 \
|
|
|
22 |
--save_total_limit 1 \
|
23 |
--max_steps 10000 \
|
24 |
--wer_threshold 20 \
|
25 |
+
--per_device_train_batch_size 40 \
|
26 |
+
--per_device_eval_batch_size 40 \
|
27 |
--dataloader_num_workers 8 \
|
28 |
--preprocessing_num_workers 8 \
|
29 |
--ddp_timeout 7200 \
|