File size: 1,098 Bytes
aa75991 4705ab6 f5b6f8f 4705ab6 f5b6f8f 4705ab6 f5b6f8f 4705ab6 f5b6f8f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 |
---
license: cc-by-4.0
---
**pythia-1.4B-finetuned-oa-instructions**
This model is a fine-tuned version of pythia on the oa dataset. It achieves the following results on the evaluation set:
Loss: 0.1224
**Model description**
More information needed
Intended uses & limitations
More information needed
**Training and evaluation data**
More information needed
**Training procedure**
**Training hyperparameters**
The following hyperparameters were used during training:
* seed: 42
* learning_rate: 5e-06
* train_batch_size: 32
* eval_batch_size: 8
* optimizer: Adam with betas : {'lr': 5e-06, 'betas': [0.9, 0.999], 'eps': 1e-08, 'weight_decay': 0.0}
* lr_scheduler_type: linear
* training_steps: 5000
* fp16
* warmup_steps 5
* Num examples = 53k
**Training results**
```
{
"epoch": 1.0,
"train_loss": 0.8031303182039198,
"train_runtime": 6338.6403,
"train_samples": 53455,
"train_samples_per_second": 8.433,
"train_steps_per_second": 0.264
}
```
**Framework versions**
* transformers 4.24.0
* torch 1.10.0+cu111
* datasets 2.10.0
* tokenizers 0.12.1 |