fearlessdots commited on
Commit
3edc5a8
1 Parent(s): 994ced8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -19
README.md CHANGED
@@ -32,25 +32,34 @@ This model and its related LoRA was fine-tuned on [https://huggingface.co/failsp
32
 
33
  ### - PEFT Parameters
34
 
35
- - lora_alpha=64,
36
- - lora_dropout=0.05,
37
- - r=128,
38
- - bias="none",
39
 
40
  ### - Training Arguments
41
 
42
- - num_train_epochs=1,
43
- - per_device_train_batch_size=1,
44
- - gradient_accumulation_steps=4,
45
- - optim="adamw_bnb_8bit",
46
- - save_steps=25,
47
- - logging_steps=25,
48
- - learning_rate=2e-4,
49
- - weight_decay=0.001,
50
- - fp16=False,
51
- - bf16=False,
52
- - max_grad_norm=0.3,
53
- - max_steps=-1,
54
- - warmup_ratio=0.03,
55
- - group_by_length=True,
56
- - lr_scheduler_type="constant",
 
 
 
 
 
 
 
 
 
 
32
 
33
  ### - PEFT Parameters
34
 
35
+ - lora_alpha=64
36
+ - lora_dropout=0.05
37
+ - r=128
38
+ - bias="none"
39
 
40
  ### - Training Arguments
41
 
42
+ - num_train_epochs=1
43
+ - per_device_train_batch_size=1
44
+ - gradient_accumulation_steps=4
45
+ - optim="adamw_bnb_8bit"
46
+ - save_steps=25
47
+ - logging_steps=25
48
+ - learning_rate=2e-4
49
+ - weight_decay=0.001
50
+ - fp16=False
51
+ - bf16=False
52
+ - max_grad_norm=0.3
53
+ - max_steps=-1
54
+ - warmup_ratio=0.03
55
+ - group_by_length=True
56
+ - lr_scheduler_type="constant"
57
+
58
+ ## Credits
59
+
60
+ - Meta ([https://huggingface.co/meta-llama](https://huggingface.co/meta-llama)): for the original Llama-3;
61
+ - failspy ([https://huggingface.co/failspy](https://huggingface.co/failspy)): for the base model and the orthogonalization implementation;
62
+ - NobodyExistsOnTheInternet ([https://huggingface.co/NobodyExistsOnTheInternet](https://huggingface.co/NobodyExistsOnTheInternet)): for the incredible dataset;
63
+ - Undi95 ([https://huggingface.co/Undi95](https://huggingface.co/Undi95)) and Sao10k ([https://huggingface.co/Sao10K](https://huggingface.co/Sao10K)): my main inspirations for doing these models =]
64
+
65
+ A huge thank you to all of them ☺️