--- library_name: transformers language: - nl license: apache-2.0 base_model: openai/whisper-large-v2 tags: - generated_from_trainer metrics: - wer model-index: - name: Whisper Large V2 results: [] --- # Whisper Large V2 This model is a fine-tuned version of [openai/whisper-large-v2](https://huggingface.co/openai/whisper-large-v2) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.5449 - Wer: 25.0953 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 12 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 20 - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:------:|:----:|:---------------:|:-------:| | 0.841 | 0.2239 | 15 | 0.5933 | 48.4347 | | 0.5965 | 0.4478 | 30 | 0.5012 | 27.9249 | | 0.5018 | 0.6716 | 45 | 0.4670 | 25.0251 | | 0.4578 | 0.8955 | 60 | 0.4569 | 27.6390 | | 0.3824 | 1.1194 | 75 | 0.4603 | 27.2978 | | 0.2738 | 1.3433 | 90 | 0.4537 | 25.0301 | | 0.2375 | 1.5672 | 105 | 0.4516 | 24.4632 | | 0.2573 | 1.7910 | 120 | 0.4381 | 25.3512 | | 0.241 | 2.0149 | 135 | 0.4379 | 25.4766 | | 0.1265 | 2.2388 | 150 | 0.4624 | 23.7809 | | 0.1391 | 2.4627 | 165 | 0.4588 | 26.6406 | | 0.1242 | 2.6866 | 180 | 0.4572 | 24.7642 | | 0.1227 | 2.9104 | 195 | 0.4561 | 27.5738 | | 0.0774 | 3.1343 | 210 | 0.4790 | 24.2474 | | 0.0543 | 3.3582 | 225 | 0.4931 | 31.8483 | | 0.0506 | 3.5821 | 240 | 0.5087 | 25.3010 | | 0.056 | 3.8060 | 255 | 0.4933 | 27.6942 | | 0.0527 | 4.0299 | 270 | 0.5009 | 26.2543 | | 0.0233 | 4.2537 | 285 | 0.5447 | 27.8999 | | 0.0193 | 4.4776 | 300 | 0.5458 | 27.0570 | | 0.0167 | 4.7015 | 315 | 0.5421 | 24.5384 | | 0.0183 | 4.9254 | 330 | 0.5449 | 25.0953 | ### Framework versions - Transformers 4.45.0.dev0 - Pytorch 2.1.0+cu121 - Datasets 2.20.0 - Tokenizers 0.19.1