whisper-tiny-uzbek / README.md
mirodil's picture
Update README.md
1c77b8d
|
raw
history blame
2.06 kB
metadata
language:
  - uz
license: apache-2.0
base_model: openai/whisper-tiny
tags:
  - automatic-speech-recognition
  - generated_from_trainer
metrics:
  - wer
model-index:
  - name: Whisper Tiny Uzbek
    results: []
datasets:
  - mozilla-foundation/common_voice_13_0
pipeline_tag: automatic-speech-recognition

Whisper Tiny Uzbek

This model is a fine-tuned version of openai/whisper-tiny on the mozilla-foundation/common_voice_13_0 dataset. It achieves the following results on the evaluation set:

  • Loss: 0.3775
  • Wer Ortho: 56.3536
  • Wer: 45.8937

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: constant_with_warmup
  • lr_scheduler_warmup_steps: 50
  • num_epochs: 1

Training results

Training Loss Epoch Step Validation Loss Wer Ortho Wer
0.6542 0.13 500 0.6243 76.5585 67.7862
0.5377 0.27 1000 0.5227 68.8556 60.2594
0.4573 0.4 1500 0.4727 66.7551 56.0715
0.4353 0.53 2000 0.4380 62.1211 52.5453
0.3907 0.66 2500 0.4159 61.1252 50.8035
0.4122 0.8 3000 0.3897 58.2628 47.8918
0.3698 0.93 3500 0.3775 56.3536 45.8937

Framework versions

  • Transformers 4.35.1
  • Pytorch 2.1.0
  • Datasets 2.14.6
  • Tokenizers 0.14.1