willtensora commited on
Commit
ee8beed
·
verified ·
1 Parent(s): 48176f0

End of training

Browse files
Files changed (2) hide show
  1. README.md +10 -14
  2. adapter_model.bin +2 -2
README.md CHANGED
@@ -1,6 +1,7 @@
1
  ---
2
  library_name: peft
3
- base_model: oopsung/llama2-7b-n-ox-test-v1
 
4
  tags:
5
  - axolotl
6
  - generated_from_trainer
@@ -18,7 +19,7 @@ should probably proofread and complete it, then remove this comment. -->
18
  axolotl version: `0.4.1`
19
  ```yaml
20
  adapter: lora
21
- base_model: oopsung/llama2-7b-n-ox-test-v1
22
  bf16: auto
23
  chat_template: llama3
24
  dataset_prepared_path: null
@@ -97,9 +98,9 @@ xformers_attention: null
97
 
98
  # 123e4567-e89b-12d3-a456-426614174000
99
 
100
- This model is a fine-tuned version of [oopsung/llama2-7b-n-ox-test-v1](https://huggingface.co/oopsung/llama2-7b-n-ox-test-v1) on the None dataset.
101
  It achieves the following results on the evaluation set:
102
- - Loss: nan
103
 
104
  ## Model description
105
 
@@ -119,24 +120,19 @@ More information needed
119
 
120
  The following hyperparameters were used during training:
121
  - learning_rate: 0.0002
122
- - train_batch_size: 2
123
- - eval_batch_size: 2
124
  - seed: 42
125
- - gradient_accumulation_steps: 4
126
- - total_train_batch_size: 8
127
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
128
  - lr_scheduler_type: cosine
129
- - lr_scheduler_warmup_steps: 10
130
- - training_steps: 10
131
 
132
  ### Training results
133
 
134
  | Training Loss | Epoch | Step | Validation Loss |
135
  |:-------------:|:------:|:----:|:---------------:|
136
- | 0.0 | 0.0006 | 1 | nan |
137
- | 0.0 | 0.0017 | 3 | nan |
138
- | 0.0 | 0.0034 | 6 | nan |
139
- | 0.0 | 0.0050 | 9 | nan |
140
 
141
 
142
  ### Framework versions
 
1
  ---
2
  library_name: peft
3
+ license: apache-2.0
4
+ base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
5
  tags:
6
  - axolotl
7
  - generated_from_trainer
 
19
  axolotl version: `0.4.1`
20
  ```yaml
21
  adapter: lora
22
+ base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
23
  bf16: auto
24
  chat_template: llama3
25
  dataset_prepared_path: null
 
98
 
99
  # 123e4567-e89b-12d3-a456-426614174000
100
 
101
+ This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0) on the None dataset.
102
  It achieves the following results on the evaluation set:
103
+ - Loss: 2.1212
104
 
105
  ## Model description
106
 
 
120
 
121
  The following hyperparameters were used during training:
122
  - learning_rate: 0.0002
123
+ - train_batch_size: 1
124
+ - eval_batch_size: 1
125
  - seed: 42
 
 
126
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
127
  - lr_scheduler_type: cosine
128
+ - lr_scheduler_warmup_steps: 2
129
+ - training_steps: 2
130
 
131
  ### Training results
132
 
133
  | Training Loss | Epoch | Step | Validation Loss |
134
  |:-------------:|:------:|:----:|:---------------:|
135
+ | 2.0306 | 0.0001 | 2 | 2.1212 |
 
 
 
136
 
137
 
138
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a59373a974866d944d76f095a936779dd70b9ca43d96196d5e05e9668d5584d7
3
- size 80115210
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2591ceebde5e3f4cf9731ce594751386bb42f0e3804e5c597d9aa5e670171fc5
3
+ size 25342042