scream_duodevicesimus_working_noaudiobooks_7e5_v2
This model is a fine-tuned version of openai/whisper-small on the NbAiLab/ncc_speech dataset. It achieves the following results on the evaluation set:
- step: 19999
- validation_fleurs_loss: 0.3089
- train_loss: 0.7173
- validation_fleurs_wer: 8.2391
- validation_fleurs_cer: 3.7039
- validation_fleurs_exact_wer: 12.6643
- validation_fleurs_exact_cer: 4.6485
- validation_stortinget_loss: 0.2845
- validation_stortinget_wer: 13.9625
- validation_stortinget_cer: 10.0306
- validation_stortinget_exact_wer: 17.2389
- validation_stortinget_exact_cer: 10.5844
- validation_nrk_tv_loss: 0.7447
- validation_nrk_tv_wer: 40.1880
- validation_nrk_tv_cer: 31.3161
- validation_nrk_tv_exact_wer: 47.6494
- validation_nrk_tv_exact_cer: 32.4497
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 7e-05
- lr_scheduler_type: linear
- per_device_train_batch_size: 32
- total_train_batch_size_per_node: 128
- total_train_batch_size: 1024
- total_optimization_steps: 20,000
- starting_optimization_step: None
- finishing_optimization_step: 20,000
- num_train_dataset_workers: 32
- num_hosts: 8
- total_num_training_examples: 20,480,000
- steps_per_epoch: 11275
- num_beams: 5
- dropout: True
- bpe_dropout_probability: 0.1
- activation_dropout_probability: 0.1
Training results
step | validation_fleurs_loss | train_loss | validation_fleurs_wer | validation_fleurs_cer | validation_fleurs_exact_wer | validation_fleurs_exact_cer | validation_stortinget_loss | validation_stortinget_wer | validation_stortinget_cer | validation_stortinget_exact_wer | validation_stortinget_exact_cer | validation_nrk_tv_loss | validation_nrk_tv_wer | validation_nrk_tv_cer | validation_nrk_tv_exact_wer | validation_nrk_tv_exact_cer |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
0 | 1.3211 | 3.0189 | 110.1725 | 80.3659 | 196.8041 | 131.4230 | 1.5012 | 76.6096 | 51.2561 | 82.1890 | 54.4126 | 1.8187 | 259.8656 | 217.2117 | 269.5665 | 222.7746 |
1000 | 0.6977 | 1.1353 | 13.4444 | 4.3105 | 17.5926 | 5.2863 | 0.4717 | 21.7105 | 13.9604 | 25.3783 | 14.6687 | 0.9934 | 86.4845 | 70.4142 | 93.7677 | 73.6462 |
2000 | 0.3926 | 0.8912 | 10.5889 | 3.7088 | 14.7849 | 4.6968 | 0.3930 | 18.7212 | 12.5960 | 22.2213 | 13.2354 | 0.8926 | 49.9691 | 39.8385 | 57.6635 | 41.2514 |
3000 | 0.3620 | 0.8106 | 10.7674 | 4.3007 | 15.0836 | 5.2573 | 0.3632 | 17.5019 | 11.9674 | 21.0430 | 12.5977 | 0.8606 | 44.9157 | 34.5510 | 52.6419 | 35.8510 |
4000 | 0.3363 | 0.8043 | 10.3807 | 3.8518 | 14.0980 | 4.7886 | 0.3443 | 16.1694 | 11.2786 | 19.6917 | 11.8983 | 0.8431 | 44.9487 | 34.0425 | 52.5379 | 35.4061 |
5000 | 0.3060 | 0.7682 | 9.6074 | 3.6694 | 13.8590 | 4.5808 | 0.3329 | 16.0903 | 11.1667 | 19.5724 | 11.7732 | 0.8154 | 45.4598 | 35.0224 | 52.7292 | 36.3997 |
6000 | 0.3477 | 0.7510 | 9.2207 | 3.5510 | 13.3214 | 4.5083 | 0.3246 | 15.9711 | 11.2829 | 19.4232 | 11.8775 | 0.8097 | 43.0897 | 33.1321 | 50.5325 | 34.3331 |
7000 | 0.3152 | 0.7608 | 9.6074 | 4.1034 | 13.7395 | 5.0834 | 0.3217 | 15.1188 | 10.6651 | 18.5510 | 11.2540 | 0.7959 | 42.0139 | 32.2852 | 49.4716 | 33.4915 |
8000 | 0.3232 | 0.7680 | 9.8453 | 3.9258 | 13.7993 | 4.8128 | 0.3161 | 15.1877 | 10.7202 | 18.5356 | 11.2959 | 0.7938 | 42.1829 | 32.6832 | 49.6256 | 34.2256 |
9000 | 0.3376 | 0.7386 | 8.9827 | 3.4967 | 12.8734 | 4.4262 | 0.3082 | 14.8650 | 10.5644 | 18.2372 | 11.1377 | 0.7793 | 41.7501 | 32.6252 | 49.4924 | 33.8756 |
10000 | 0.3152 | 0.7408 | 9.0422 | 3.7335 | 13.5006 | 4.6678 | 0.3068 | 14.7458 | 10.4670 | 18.1324 | 11.0457 | 0.7773 | 41.3750 | 31.9683 | 49.1887 | 33.1957 |
11000 | 0.3167 | 0.7246 | 9.2802 | 3.7976 | 13.2318 | 4.7354 | 0.3010 | 14.4811 | 10.3391 | 17.8357 | 10.9036 | 0.7745 | 41.2926 | 31.8771 | 48.8018 | 33.0807 |
12000 | 0.3087 | 0.7240 | 8.7745 | 3.6447 | 12.7838 | 4.5712 | 0.2996 | 14.2912 | 10.2057 | 17.6353 | 10.7722 | 0.7683 | 41.1525 | 32.0549 | 48.8018 | 33.2402 |
13000 | 0.3330 | 0.7061 | 8.3284 | 3.5362 | 12.4851 | 4.4552 | 0.2981 | 14.3397 | 10.2971 | 17.7257 | 10.8680 | 0.7642 | 41.1401 | 32.0379 | 48.7685 | 33.1949 |
14000 | 0.3176 | 0.6887 | 8.8935 | 3.7680 | 12.8734 | 4.6726 | 0.2940 | 14.1728 | 10.2323 | 17.5270 | 10.8029 | 0.7618 | 39.9942 | 30.7597 | 47.3873 | 31.9470 |
15000 | 0.3113 | 0.7296 | 8.7151 | 3.8617 | 13.0526 | 4.8224 | 0.2924 | 14.0962 | 10.1638 | 17.4495 | 10.7350 | 0.7595 | 40.2951 | 31.0757 | 47.7201 | 32.2502 |
16000 | 0.3058 | 0.6820 | 9.2802 | 4.0688 | 13.2019 | 4.9481 | 0.2922 | 14.0766 | 10.1057 | 17.4222 | 10.6784 | 0.7544 | 40.4641 | 31.5116 | 47.9947 | 32.7092 |
17000 | 0.3030 | 0.7120 | 8.9233 | 3.8913 | 13.0824 | 4.8321 | 0.2878 | 14.1413 | 10.1954 | 17.4785 | 10.7684 | 0.7559 | 40.3487 | 31.4791 | 48.0113 | 32.6455 |
18000 | 0.3080 | 0.6951 | 8.5961 | 3.7138 | 12.5747 | 4.6533 | 0.2863 | 13.8595 | 9.9432 | 17.1562 | 10.5007 | 0.7467 | 40.0437 | 31.2512 | 47.5745 | 32.4163 |
19000 | 0.3104 | 0.6771 | 8.5961 | 3.6743 | 12.7838 | 4.6050 | 0.2854 | 13.9702 | 10.0538 | 17.2858 | 10.6153 | 0.7477 | 40.2003 | 31.3663 | 47.6743 | 32.5098 |
19999 | 0.3089 | 0.7173 | 8.2391 | 3.7039 | 12.6643 | 4.6485 | ||||||||||
19999 | 0.2845 | 0.7173 | 13.9625 | 10.0306 | 17.2389 | 10.5844 | ||||||||||
19999 | 0.7447 | 0.7173 | 40.1880 | 31.3161 | 47.6494 | 32.4497 |
Framework versions
- Transformers 4.31.0.dev0
- Datasets 2.13.0
- Tokenizers 0.13.3
- Downloads last month
- 16
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.