freds0 commited on
Commit
6ff1418
·
verified ·
1 Parent(s): d7f7640

Saving train state of step 1000

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "./distil-large-v3-init",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
 
1
  {
2
+ "_name_or_path": "openai/whisper-large-v3",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
distil-whisper/events.out.tfevents.1727360866.22c57e4734ce.16172.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d47f2ae3a7490fa666782cee83cb552406c4b6c1bedc6f8ca918729f30770f62
3
+ size 12458
run_distillation.sh CHANGED
@@ -22,8 +22,8 @@ python run_distillation.py \
22
  --save_total_limit 1 \
23
  --max_steps 10000 \
24
  --wer_threshold 20 \
25
- --per_device_train_batch_size 30 \
26
- --per_device_eval_batch_size 30 \
27
  --dataloader_num_workers 8 \
28
  --preprocessing_num_workers 8 \
29
  --ddp_timeout 7200 \
 
22
  --save_total_limit 1 \
23
  --max_steps 10000 \
24
  --wer_threshold 20 \
25
+ --per_device_train_batch_size 40 \
26
+ --per_device_eval_batch_size 40 \
27
  --dataloader_num_workers 8 \
28
  --preprocessing_num_workers 8 \
29
  --ddp_timeout 7200 \