--- library_name: transformers license: apache-2.0 base_model: google-t5/t5-small tags: - translation - generated_from_trainer metrics: - bleu model-index: - name: t5-small-finetuned-hausa-to-english results: [] --- # t5-small-finetuned-hausa-to-english This model is a fine-tuned version of [google-t5/t5-small](https://huggingface.co/google-t5/t5-small) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.5734 - Bleu: 71.4633 - Gen Len: 6.984 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0008 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 3000 - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len | |:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:| | 3.4817 | 1.0 | 749 | 2.2949 | 34.5266 | 7.5496 | | 2.0957 | 2.0 | 1498 | 1.6167 | 53.0797 | 6.8785 | | 1.562 | 3.0 | 2247 | 1.3650 | 60.3083 | 7.0095 | | 1.2877 | 4.0 | 2996 | 1.2149 | 63.8437 | 6.9444 | | 1.1026 | 5.0 | 3745 | 1.1452 | 66.7883 | 7.13 | | 0.9531 | 6.0 | 4494 | 1.1028 | 67.3774 | 6.911 | | 0.8402 | 7.0 | 5243 | 1.0995 | 68.0354 | 6.8114 | | 0.7513 | 8.0 | 5992 | 1.0878 | 69.4876 | 7.0216 | | 0.6746 | 9.0 | 6741 | 1.1109 | 69.7327 | 7.1134 | | 0.6073 | 10.0 | 7490 | 1.1167 | 70.1607 | 7.0526 | | 0.5531 | 11.0 | 8239 | 1.1468 | 69.8006 | 6.8101 | | 0.4981 | 12.0 | 8988 | 1.1856 | 70.5423 | 6.8789 | | 0.4544 | 13.0 | 9737 | 1.2019 | 70.5876 | 6.9313 | | 0.4095 | 14.0 | 10486 | 1.2347 | 70.7996 | 6.8371 | | 0.373 | 15.0 | 11235 | 1.2734 | 71.0903 | 7.0274 | | 0.3408 | 16.0 | 11984 | 1.2974 | 71.104 | 7.0025 | | 0.3096 | 17.0 | 12733 | 1.3313 | 70.7308 | 6.925 | | 0.2856 | 18.0 | 13482 | 1.3820 | 70.9862 | 6.9656 | | 0.2601 | 19.0 | 14231 | 1.4016 | 71.1836 | 7.0082 | | 0.2404 | 20.0 | 14980 | 1.4483 | 71.0219 | 6.9268 | | 0.2241 | 21.0 | 15729 | 1.4714 | 71.2721 | 6.9552 | | 0.2065 | 22.0 | 16478 | 1.4814 | 71.3874 | 6.9968 | | 0.1942 | 23.0 | 17227 | 1.5090 | 71.4722 | 6.9404 | | 0.1831 | 24.0 | 17976 | 1.5265 | 71.4556 | 6.9771 | | 0.173 | 25.0 | 18725 | 1.5379 | 71.4026 | 6.9998 | | 0.1662 | 26.0 | 19474 | 1.5530 | 71.4843 | 6.9932 | | 0.159 | 27.0 | 20223 | 1.5668 | 71.3663 | 6.9784 | | 0.1568 | 28.0 | 20972 | 1.5742 | 71.3261 | 6.9734 | | 0.1566 | 29.0 | 21721 | 1.5739 | 71.4435 | 6.9843 | | 0.156 | 30.0 | 22470 | 1.5734 | 71.4633 | 6.984 | ### Framework versions - Transformers 4.44.2 - Pytorch 2.4.0+cu121 - Datasets 2.21.0 - Tokenizers 0.19.1