Saving train state of step 1000
Browse files
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"activation_dropout": 0.0,
|
4 |
"activation_function": "gelu",
|
5 |
"apply_spec_augment": false,
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "openai/whisper-large-v3",
|
3 |
"activation_dropout": 0.0,
|
4 |
"activation_function": "gelu",
|
5 |
"apply_spec_augment": false,
|
distil-whisper/events.out.tfevents.1727223225.22c57e4734ce.13411.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:84c1f39e61f34ffd721eb64bd3d5d7799afae89413b54fe5532242beb335477c
|
3 |
+
size 88
|
distil-whisper/events.out.tfevents.1727223588.22c57e4734ce.14278.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6109060c54ad1230fdc32ded249fd5a11ad84ffb2c740c561c25738017fc5347
|
3 |
+
size 12458
|
run_distillation.sh
CHANGED
@@ -22,8 +22,8 @@ python run_distillation.py \
|
|
22 |
--save_total_limit 1 \
|
23 |
--max_steps 10000 \
|
24 |
--wer_threshold 20 \
|
25 |
-
--per_device_train_batch_size
|
26 |
-
--per_device_eval_batch_size
|
27 |
--dataloader_num_workers 8 \
|
28 |
--preprocessing_num_workers 8 \
|
29 |
--ddp_timeout 7200 \
|
|
|
22 |
--save_total_limit 1 \
|
23 |
--max_steps 10000 \
|
24 |
--wer_threshold 20 \
|
25 |
+
--per_device_train_batch_size 30 \
|
26 |
+
--per_device_eval_batch_size 30 \
|
27 |
--dataloader_num_workers 8 \
|
28 |
--preprocessing_num_workers 8 \
|
29 |
--ddp_timeout 7200 \
|