freds0 commited on
Commit
41d9eee
1 Parent(s): d17b72e

Saving train state of step 1000

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "./distil-large-v3-init",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
 
1
  {
2
+ "_name_or_path": "openai/whisper-large-v3",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
distil-whisper/events.out.tfevents.1727223225.22c57e4734ce.13411.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84c1f39e61f34ffd721eb64bd3d5d7799afae89413b54fe5532242beb335477c
3
+ size 88
distil-whisper/events.out.tfevents.1727223588.22c57e4734ce.14278.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6109060c54ad1230fdc32ded249fd5a11ad84ffb2c740c561c25738017fc5347
3
+ size 12458
run_distillation.sh CHANGED
@@ -22,8 +22,8 @@ python run_distillation.py \
22
  --save_total_limit 1 \
23
  --max_steps 10000 \
24
  --wer_threshold 20 \
25
- --per_device_train_batch_size 40 \
26
- --per_device_eval_batch_size 40 \
27
  --dataloader_num_workers 8 \
28
  --preprocessing_num_workers 8 \
29
  --ddp_timeout 7200 \
 
22
  --save_total_limit 1 \
23
  --max_steps 10000 \
24
  --wer_threshold 20 \
25
+ --per_device_train_batch_size 30 \
26
+ --per_device_eval_batch_size 30 \
27
  --dataloader_num_workers 8 \
28
  --preprocessing_num_workers 8 \
29
  --ddp_timeout 7200 \