|
--- |
|
library_name: transformers |
|
tags: |
|
- orpo |
|
- llama |
|
- llama 3 |
|
- fine tune |
|
license: apache-2.0 |
|
datasets: |
|
- mlabonne/orpo-dpo-mix-40k |
|
--- |
|
|
|
# Model Card for Model ID |
|
|
|
<!-- Provide a quick summary of what the model is/does. --> |
|
Quick Llama 3 8B finetune with ORPO. Demontration that it can be fine tune in 2 hours only. |
|
Thanks to Maxime Labonne's notebook: |
|
|
|
https://colab.research.google.com/drive/1eHNWg9gnaXErdAa8_mcvjMupbSS6rDvi?usp=sharing |
|
|
|
- **Number of training samples from the dataset**: 1500 out of 40K |
|
- **Hardware Type:** L4 |
|
- **Hours of training:** 2 |
|
- **Cloud Provider:** google colab |
|
|