|
--- |
|
base_model: codellama/CodeLlama-7b-Instruct-hf |
|
library_name: peft |
|
license: llama2 |
|
tags: |
|
- trl |
|
- sft |
|
- generated_from_trainer |
|
model-index: |
|
- name: Codellama-7b-lora-rps-adapter |
|
results: [] |
|
--- |
|
|
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You |
|
should probably proofread and complete it, then remove this comment. --> |
|
|
|
# Codellama-7b-lora-rps-adapter |
|
|
|
This model is a fine-tuned version of [codellama/CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) on the None dataset. |
|
It achieves the following results on the evaluation set: |
|
- Loss: 0.3110 |
|
|
|
## Model description |
|
|
|
More information needed |
|
|
|
## Intended uses & limitations |
|
|
|
More information needed |
|
|
|
## Training and evaluation data |
|
|
|
More information needed |
|
|
|
## Training procedure |
|
|
|
### Training hyperparameters |
|
|
|
The following hyperparameters were used during training: |
|
- learning_rate: 0.0002 |
|
- train_batch_size: 2 |
|
- eval_batch_size: 2 |
|
- seed: 42 |
|
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 |
|
- lr_scheduler_type: linear |
|
- lr_scheduler_warmup_ratio: 0.03 |
|
- num_epochs: 4 |
|
|
|
### Training results |
|
|
|
| Training Loss | Epoch | Step | Validation Loss | |
|
|:-------------:|:------:|:-----:|:---------------:| |
|
| 0.2019 | 2.5990 | 17000 | 0.2968 | |
|
| 0.1928 | 2.6143 | 17100 | 0.2975 | |
|
| 0.1992 | 2.6296 | 17200 | 0.2981 | |
|
| 0.1975 | 2.6449 | 17300 | 0.2987 | |
|
| 0.2003 | 2.6601 | 17400 | 0.2963 | |
|
| 0.1847 | 2.6754 | 17500 | 0.2970 | |
|
| 0.1945 | 2.6907 | 17600 | 0.2961 | |
|
| 0.2057 | 2.7060 | 17700 | 0.2970 | |
|
| 0.1782 | 2.7213 | 17800 | 0.2967 | |
|
| 0.1813 | 2.7366 | 17900 | 0.2975 | |
|
| 0.2001 | 2.7519 | 18000 | 0.2953 | |
|
| 0.2074 | 2.7672 | 18100 | 0.2959 | |
|
| 0.1957 | 2.7824 | 18200 | 0.2969 | |
|
| 0.2006 | 2.7977 | 18300 | 0.2943 | |
|
| 0.2021 | 2.8130 | 18400 | 0.2939 | |
|
| 0.1862 | 2.8283 | 18500 | 0.2931 | |
|
| 0.1951 | 2.8436 | 18600 | 0.2934 | |
|
| 0.205 | 2.8589 | 18700 | 0.2936 | |
|
| 0.2094 | 2.8742 | 18800 | 0.2919 | |
|
| 0.1766 | 2.8895 | 18900 | 0.2935 | |
|
| 0.2001 | 2.9048 | 19000 | 0.2931 | |
|
| 0.1977 | 2.9200 | 19100 | 0.2941 | |
|
| 0.1884 | 2.9353 | 19200 | 0.2922 | |
|
| 0.1784 | 2.9506 | 19300 | 0.2927 | |
|
| 0.1857 | 2.9659 | 19400 | 0.2921 | |
|
| 0.1972 | 2.9812 | 19500 | 0.2926 | |
|
| 0.1921 | 2.9965 | 19600 | 0.2929 | |
|
| 0.1433 | 3.0118 | 19700 | 0.3114 | |
|
| 0.1486 | 3.0271 | 19800 | 0.3115 | |
|
| 0.1381 | 3.0423 | 19900 | 0.3147 | |
|
| 0.1375 | 3.0576 | 20000 | 0.3122 | |
|
| 0.1359 | 3.0729 | 20100 | 0.3144 | |
|
| 0.133 | 3.0882 | 20200 | 0.3165 | |
|
| 0.1346 | 3.1035 | 20300 | 0.3151 | |
|
| 0.132 | 3.1188 | 20400 | 0.3169 | |
|
| 0.1338 | 3.1341 | 20500 | 0.3137 | |
|
| 0.1238 | 3.1494 | 20600 | 0.3160 | |
|
| 0.1264 | 3.1647 | 20700 | 0.3146 | |
|
| 0.1382 | 3.1799 | 20800 | 0.3139 | |
|
| 0.136 | 3.1952 | 20900 | 0.3110 | |
|
| 0.1321 | 3.2105 | 21000 | 0.3129 | |
|
| 0.134 | 3.2258 | 21100 | 0.3148 | |
|
| 0.134 | 3.2411 | 21200 | 0.3139 | |
|
| 0.1338 | 3.2564 | 21300 | 0.3140 | |
|
| 0.1317 | 3.2717 | 21400 | 0.3148 | |
|
| 0.1281 | 3.2870 | 21500 | 0.3132 | |
|
| 0.1279 | 3.3022 | 21600 | 0.3124 | |
|
| 0.1355 | 3.3175 | 21700 | 0.3133 | |
|
| 0.127 | 3.3328 | 21800 | 0.3129 | |
|
| 0.1388 | 3.3481 | 21900 | 0.3157 | |
|
| 0.1316 | 3.3634 | 22000 | 0.3134 | |
|
| 0.1378 | 3.3787 | 22100 | 0.3127 | |
|
| 0.1357 | 3.3940 | 22200 | 0.3131 | |
|
| 0.1271 | 3.4093 | 22300 | 0.3141 | |
|
| 0.1333 | 3.4246 | 22400 | 0.3142 | |
|
| 0.1311 | 3.4398 | 22500 | 0.3133 | |
|
| 0.1261 | 3.4551 | 22600 | 0.3138 | |
|
| 0.1313 | 3.4704 | 22700 | 0.3129 | |
|
| 0.1296 | 3.4857 | 22800 | 0.3135 | |
|
| 0.1348 | 3.5010 | 22900 | 0.3134 | |
|
| 0.1252 | 3.5163 | 23000 | 0.3131 | |
|
| 0.1403 | 3.5316 | 23100 | 0.3117 | |
|
| 0.1266 | 3.5469 | 23200 | 0.3126 | |
|
| 0.135 | 3.5621 | 23300 | 0.3135 | |
|
| 0.1344 | 3.5774 | 23400 | 0.3133 | |
|
| 0.1452 | 3.5927 | 23500 | 0.3128 | |
|
| 0.1285 | 3.6080 | 23600 | 0.3131 | |
|
| 0.1235 | 3.6233 | 23700 | 0.3108 | |
|
| 0.1255 | 3.6386 | 23800 | 0.3111 | |
|
| 0.1335 | 3.6539 | 23900 | 0.3114 | |
|
| 0.1397 | 3.6692 | 24000 | 0.3109 | |
|
| 0.1359 | 3.6845 | 24100 | 0.3108 | |
|
| 0.1269 | 3.6997 | 24200 | 0.3120 | |
|
| 0.1345 | 3.7150 | 24300 | 0.3115 | |
|
| 0.131 | 3.7303 | 24400 | 0.3111 | |
|
| 0.1332 | 3.7456 | 24500 | 0.3115 | |
|
| 0.1226 | 3.7609 | 24600 | 0.3123 | |
|
| 0.1244 | 3.7762 | 24700 | 0.3114 | |
|
| 0.123 | 3.7915 | 24800 | 0.3115 | |
|
| 0.1302 | 3.8068 | 24900 | 0.3103 | |
|
| 0.1291 | 3.8220 | 25000 | 0.3108 | |
|
| 0.1335 | 3.8373 | 25100 | 0.3118 | |
|
| 0.1251 | 3.8526 | 25200 | 0.3115 | |
|
| 0.1321 | 3.8679 | 25300 | 0.3111 | |
|
| 0.1249 | 3.8832 | 25400 | 0.3111 | |
|
| 0.1324 | 3.8985 | 25500 | 0.3111 | |
|
| 0.1236 | 3.9138 | 25600 | 0.3112 | |
|
| 0.1399 | 3.9291 | 25700 | 0.3108 | |
|
| 0.1255 | 3.9444 | 25800 | 0.3107 | |
|
| 0.1462 | 3.9596 | 25900 | 0.3107 | |
|
| 0.1217 | 3.9749 | 26000 | 0.3108 | |
|
| 0.1238 | 3.9902 | 26100 | 0.3110 | |
|
|
|
|
|
### Framework versions |
|
|
|
- PEFT 0.13.0 |
|
- Transformers 4.45.1 |
|
- Pytorch 2.4.1+cu121 |
|
- Datasets 3.0.1 |
|
- Tokenizers 0.20.0 |