Safetensors
English
llama
jeiku commited on
Commit
e732c45
·
verified ·
1 Parent(s): fb62a9a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +130 -1
README.md CHANGED
@@ -11,7 +11,7 @@ language:
11
 
12
  **Aura-8B** is a state of the art dedicated roleplaying model designed to fulfill your every desire.
13
 
14
- This finetune has seen several hundreds of millions of tokens of instruction and roleplaying data. A Kahneman-Tversky Optimization has been applied over the various Supervised Fine-Tuning stages to lend a unique voice to your roleplaying experience.
15
 
16
  Developed by **Aura Industries**, with contributions from **Anthracite Org**
17
 
@@ -25,6 +25,135 @@ Developed by **Aura Industries**, with contributions from **Anthracite Org**
25
  - **Language**: English
26
  - **Max Context**: 8,192+ tokens
27
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  ## License
29
 
30
  This model is licensed under the [Apache 2.0 License](https://www.apache.org/licenses/LICENSE-2.0).
 
11
 
12
  **Aura-8B** is a state of the art dedicated roleplaying model designed to fulfill your every desire.
13
 
14
+ This finetune has seen several hundreds of millions of tokens of instruction and roleplaying data.
15
 
16
  Developed by **Aura Industries**, with contributions from **Anthracite Org**
17
 
 
25
  - **Language**: English
26
  - **Max Context**: 8,192+ tokens
27
 
28
+ ## Training Configuration
29
+
30
+ ```
31
+ base_model: arcee-ai/Llama-3.1-SuperNova-Lite
32
+ model_type: AutoModelForCausalLM
33
+ tokenizer_type: AutoTokenizer
34
+
35
+ load_in_8bit: false
36
+ load_in_4bit: false
37
+ strict: false
38
+
39
+ datasets:
40
+ - path: FourOhFour/RP_Phase
41
+ type: chat_template
42
+ chat_template: llama3
43
+ roles_to_train: ["gpt"]
44
+ field_messages: conversations
45
+ message_field_role: from
46
+ message_field_content: value
47
+ train_on_eos: turn
48
+ - path: Nitral-AI/Cybersecurity-ShareGPT
49
+ type: chat_template
50
+ chat_template: llama3
51
+ roles_to_train: ["gpt"]
52
+ field_messages: conversations
53
+ message_field_role: from
54
+ message_field_content: value
55
+ train_on_eos: turn
56
+ - path: Nitral-AI/Medical_Instruct-ShareGPT
57
+ type: chat_template
58
+ chat_template: llama3
59
+ roles_to_train: ["gpt"]
60
+ field_messages: conversations
61
+ message_field_role: from
62
+ message_field_content: value
63
+ train_on_eos: turn
64
+ - path: Nitral-AI/Olympiad_Math-ShareGPT
65
+ type: chat_template
66
+ chat_template: llama3
67
+ roles_to_train: ["gpt"]
68
+ field_messages: conversations
69
+ message_field_role: from
70
+ message_field_content: value
71
+ train_on_eos: turn
72
+ - path: NewEden/Claude-Instruct-5k
73
+ type: chat_template
74
+ chat_template: llama3
75
+ roles_to_train: ["gpt"]
76
+ field_messages: conversations
77
+ message_field_role: from
78
+ message_field_content: value
79
+ train_on_eos: turn
80
+ - path: lodrick-the-lafted/kalo-opus-instruct-3k-filtered
81
+ type: chat_template
82
+ chat_template: llama3
83
+ roles_to_train: ["gpt"]
84
+ field_messages: conversations
85
+ message_field_role: from
86
+ message_field_content: value
87
+ train_on_eos: turn
88
+ - path: Nitral-AI/Creative_Writing-ShareGPT
89
+ type: chat_template
90
+ chat_template: llama3
91
+ roles_to_train: ["gpt"]
92
+ field_messages: conversations
93
+ message_field_role: from
94
+ message_field_content: value
95
+ train_on_eos: turn
96
+ - path: jeiku/Writing
97
+ type: completion
98
+ field: text
99
+
100
+ shuffle_merged_datasets: true
101
+ dataset_prepared_path:
102
+ val_set_size: 0.01
103
+ output_dir: ./output/out
104
+
105
+ hub_model_id: jeiku/Aura-8B
106
+ hub_strategy: "all_checkpoints"
107
+ push_dataset_to_hub:
108
+ hf_use_auth_token: true
109
+
110
+ sequence_len: 8192
111
+ sample_packing: true
112
+ eval_sample_packing: false
113
+ pad_to_sequence_len:
114
+
115
+ wandb_project: Aura-8B
116
+ wandb_entity:
117
+ wandb_watch:
118
+ wandb_name: Aura-8B
119
+ wandb_log_model:
120
+
121
+ gradient_accumulation_steps: 16
122
+ micro_batch_size: 2
123
+ num_epochs: 2
124
+ optimizer: paged_adamw_8bit
125
+ lr_scheduler: cosine
126
+ learning_rate: 1e-5
127
+
128
+ train_on_inputs: false
129
+ group_by_length: false
130
+ bf16: auto
131
+ fp16:
132
+ tf32: false
133
+
134
+ gradient_checkpointing: true
135
+ early_stopping_patience:
136
+ resume_from_checkpoint:
137
+ local_rank:
138
+ logging_steps: 1
139
+ xformers_attention:
140
+ flash_attention: true
141
+
142
+ warmup_ratio: 0.1
143
+ evals_per_epoch: 2
144
+ eval_table_size:
145
+ eval_max_new_tokens:
146
+ saves_per_epoch: 1
147
+ debug:
148
+ deepspeed:
149
+ weight_decay: 0.05
150
+ fsdp:
151
+ fsdp_config:
152
+ special_tokens:
153
+ pad_token: <|finetune_right_pad_id|>
154
+ eos_token: <|eot_id|>
155
+ ```
156
+
157
  ## License
158
 
159
  This model is licensed under the [Apache 2.0 License](https://www.apache.org/licenses/LICENSE-2.0).