fearlessdots
commited on
Commit
•
3edc5a8
1
Parent(s):
994ced8
Update README.md
Browse files
README.md
CHANGED
@@ -32,25 +32,34 @@ This model and its related LoRA was fine-tuned on [https://huggingface.co/failsp
|
|
32 |
|
33 |
### - PEFT Parameters
|
34 |
|
35 |
-
- lora_alpha=64
|
36 |
-
- lora_dropout=0.05
|
37 |
-
- r=128
|
38 |
-
- bias="none"
|
39 |
|
40 |
### - Training Arguments
|
41 |
|
42 |
-
- num_train_epochs=1
|
43 |
-
- per_device_train_batch_size=1
|
44 |
-
- gradient_accumulation_steps=4
|
45 |
-
- optim="adamw_bnb_8bit"
|
46 |
-
- save_steps=25
|
47 |
-
- logging_steps=25
|
48 |
-
- learning_rate=2e-4
|
49 |
-
- weight_decay=0.001
|
50 |
-
- fp16=False
|
51 |
-
- bf16=False
|
52 |
-
- max_grad_norm=0.3
|
53 |
-
- max_steps=-1
|
54 |
-
- warmup_ratio=0.03
|
55 |
-
- group_by_length=True
|
56 |
-
- lr_scheduler_type="constant"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
32 |
|
33 |
### - PEFT Parameters
|
34 |
|
35 |
+
- lora_alpha=64
|
36 |
+
- lora_dropout=0.05
|
37 |
+
- r=128
|
38 |
+
- bias="none"
|
39 |
|
40 |
### - Training Arguments
|
41 |
|
42 |
+
- num_train_epochs=1
|
43 |
+
- per_device_train_batch_size=1
|
44 |
+
- gradient_accumulation_steps=4
|
45 |
+
- optim="adamw_bnb_8bit"
|
46 |
+
- save_steps=25
|
47 |
+
- logging_steps=25
|
48 |
+
- learning_rate=2e-4
|
49 |
+
- weight_decay=0.001
|
50 |
+
- fp16=False
|
51 |
+
- bf16=False
|
52 |
+
- max_grad_norm=0.3
|
53 |
+
- max_steps=-1
|
54 |
+
- warmup_ratio=0.03
|
55 |
+
- group_by_length=True
|
56 |
+
- lr_scheduler_type="constant"
|
57 |
+
|
58 |
+
## Credits
|
59 |
+
|
60 |
+
- Meta ([https://huggingface.co/meta-llama](https://huggingface.co/meta-llama)): for the original Llama-3;
|
61 |
+
- failspy ([https://huggingface.co/failspy](https://huggingface.co/failspy)): for the base model and the orthogonalization implementation;
|
62 |
+
- NobodyExistsOnTheInternet ([https://huggingface.co/NobodyExistsOnTheInternet](https://huggingface.co/NobodyExistsOnTheInternet)): for the incredible dataset;
|
63 |
+
- Undi95 ([https://huggingface.co/Undi95](https://huggingface.co/Undi95)) and Sao10k ([https://huggingface.co/Sao10K](https://huggingface.co/Sao10K)): my main inspirations for doing these models =]
|
64 |
+
|
65 |
+
A huge thank you to all of them ☺️
|