munish0838 commited on
Commit
efe6b08
•
1 Parent(s): e219d61

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +292 -0
README.md ADDED
@@ -0,0 +1,292 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: other
5
+ tags:
6
+ - axolotl
7
+ - instruct
8
+ - finetune
9
+ - chatml
10
+ - gpt4
11
+ - synthetic data
12
+ - science
13
+ - physics
14
+ - chemistry
15
+ - biology
16
+ - math
17
+ - qwen
18
+ - qwen2
19
+ base_model: Weyaxi/Einstein-v7-Qwen2-7B
20
+ datasets:
21
+ - allenai/ai2_arc
22
+ - camel-ai/physics
23
+ - camel-ai/chemistry
24
+ - camel-ai/biology
25
+ - camel-ai/math
26
+ - metaeval/reclor
27
+ - openbookqa
28
+ - mandyyyyii/scibench
29
+ - derek-thomas/ScienceQA
30
+ - TIGER-Lab/ScienceEval
31
+ - jondurbin/airoboros-3.2
32
+ - LDJnr/Capybara
33
+ - Cot-Alpaca-GPT4-From-OpenHermes-2.5
34
+ - STEM-AI-mtl/Electrical-engineering
35
+ - knowrohit07/saraswati-stem
36
+ - sablo/oasst2_curated
37
+ - lmsys/lmsys-chat-1m
38
+ - TIGER-Lab/MathInstruct
39
+ - bigbio/med_qa
40
+ - meta-math/MetaMathQA-40K
41
+ - openbookqa
42
+ - piqa
43
+ - metaeval/reclor
44
+ - derek-thomas/ScienceQA
45
+ - scibench
46
+ - sciq
47
+ - Open-Orca/SlimOrca
48
+ - migtissera/Synthia-v1.3
49
+ - TIGER-Lab/ScienceEval
50
+ - allenai/WildChat
51
+ - microsoft/orca-math-word-problems-200k
52
+ - openchat/openchat_sharegpt4_dataset
53
+ - teknium/GPTeacher-General-Instruct
54
+ - m-a-p/CodeFeedback-Filtered-Instruction
55
+ - totally-not-an-llm/EverythingLM-data-V3
56
+ - HuggingFaceH4/no_robots
57
+ - OpenAssistant/oasst_top1_2023-08-25
58
+ - WizardLM/WizardLM_evol_instruct_70k
59
+ - abacusai/SystemChat-1.1
60
+ - H-D-T/Buzz-V1.2
61
+ pipeline_tag: text-generation
62
+ ---
63
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/KLQP1jK-DIzpwHzYRIH-Q.png)
64
+
65
+ # 🔬 Einstein-v7-Qwen2-7B-GGUF
66
+ This is quantized version of [Weyaxi/Einstein-v7-Qwen2-7B](https://huggingface.co/Weyaxi/Einstein-v7-Qwen2-7B) created using llama.cpp
67
+
68
+ # Model Description
69
+ This model is a full fine-tuned version of [Qwen/Qwen2-7B](https://huggingface.co/Qwen/Qwen2-7B) on diverse datasets.
70
+
71
+ This model is finetuned using `8xMI300X` using [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl).
72
+
73
+ <details><summary>See axolotl config</summary>
74
+
75
+ axolotl version: `0.4.0`
76
+ ```yaml
77
+ base_model: Qwen/Qwen2-7B
78
+ model_type: AutoModelForCausalLM
79
+ tokenizer_type: AutoTokenizer
80
+
81
+ load_in_8bit: false
82
+ load_in_4bit: false
83
+ strict: false
84
+
85
+ chat_template: chatml
86
+ datasets:
87
+ - path: data/airoboros_3.2_without_contextual_slimorca_orca_sharegpt.json
88
+ ds_type: json
89
+ type: sharegpt
90
+ conversation: chatml
91
+
92
+ - path: data/allenai_wild_chat_gpt4_english_toxic_random_half_4k_sharegpt.json
93
+ ds_type: json
94
+ type: sharegpt
95
+ strict: false
96
+ conversation: chatml
97
+
98
+ - path: data/buzz_unstacked_chosen_math_removed_filtered.json
99
+ ds_type: json
100
+ type: alpaca
101
+ conversation: chatml
102
+
103
+ - path: data/capybara_sharegpt.json
104
+ ds_type: json
105
+ type: sharegpt
106
+ conversation: chatml
107
+
108
+ - path: data/cot_alpaca_gpt4_extracted_openhermes_2.5_sharegpt.json
109
+ ds_type: json
110
+ type: sharegpt
111
+ conversation: chatml
112
+
113
+ - path: data/everythinglm-data-v3_sharegpt.json
114
+ ds_type: json
115
+ type: sharegpt
116
+ strict: false
117
+ conversation: chatml
118
+
119
+ - path: data/gpt4_data_lmys_1m_sharegpt.json
120
+ ds_type: json
121
+ type: sharegpt
122
+ conversation: chatml
123
+
124
+ - path: data/gpteacher-instruct-special-alpaca.json
125
+ ds_type: json
126
+ type: gpteacher
127
+ conversation: chatml
128
+
129
+ - path: data/merged_all.json
130
+ ds_type: json
131
+ type: alpaca
132
+ conversation: chatml
133
+
134
+ - path: data/no_robots_sharegpt.json
135
+ ds_type: json
136
+ type: sharegpt
137
+ strict: false
138
+ conversation: chatml
139
+
140
+ - path: data/oasst_top1_from_fusechatmixture_sharegpt.json
141
+ ds_type: json
142
+ type: sharegpt
143
+ strict: false
144
+ conversation: chatml
145
+
146
+ - path: data/pippa_bagel_repo_3k_sharegpt.json
147
+ ds_type: json
148
+ type: sharegpt
149
+ conversation: chatml
150
+
151
+ - path: data/rpguild_quarter_alignment_lab_sharegpt.json
152
+ ds_type: json
153
+ type: sharegpt
154
+ conversation: chatml
155
+
156
+ - path: data/sharegpt_gpt4_english.json
157
+ ds_type: json
158
+ type: sharegpt
159
+ conversation: chatml
160
+
161
+ - path: data/slimorca_dedup_filtered_95k_sharegpt.json
162
+ ds_type: json
163
+ type: sharegpt
164
+ conversation: chatml
165
+
166
+ - path: data/soda_diaolog_longest_tenth_buzz_sharegpt.json
167
+ ds_type: json
168
+ type: sharegpt
169
+ conversation: chatml
170
+
171
+ - path: data/synthia-v1.3_sharegpt_12500.json
172
+ ds_type: json
173
+ type: sharegpt
174
+ conversation: chatml
175
+
176
+ - path: data/system_conversations_dolphin_sharegpt.json
177
+ ds_type: json
178
+ type: sharegpt
179
+ conversation: chatml
180
+
181
+ dataset_prepared_path: last_run_prepared
182
+ val_set_size: 0.002
183
+
184
+ output_dir: ./Einstein-v7-Qwen2-7B-model
185
+
186
+ sequence_len: 8192
187
+ sample_packing: true
188
+ pad_to_sequence_len: true
189
+ eval_sample_packing: false
190
+
191
+ wandb_project: Einstein
192
+ wandb_entity:
193
+ wandb_watch:
194
+ wandb_name:
195
+ wandb_log_model:
196
+ hub_model_id: Weyaxi/Einstein-v7-Qwen2-7B
197
+
198
+ gradient_accumulation_steps: 4
199
+ micro_batch_size: 6
200
+ num_epochs: 2
201
+ optimizer: paged_adamw_8bit
202
+ lr_scheduler: cosine
203
+ learning_rate: 0.00001 # look
204
+
205
+ train_on_inputs: false
206
+ group_by_length: false
207
+ bf16: auto
208
+ fp16:
209
+ tf32: false
210
+
211
+ gradient_checkpointing: unsloth
212
+ gradient_checkpointing_kwargs:
213
+ use_reentrant: true # look
214
+ early_stopping_patience:
215
+ resume_from_checkpoint:
216
+ local_rank:
217
+ logging_steps: 1
218
+ xformers_attention:
219
+ flash_attention: true
220
+
221
+ warmup_steps: 10
222
+ evals_per_epoch: 2
223
+ eval_table_size:
224
+ eval_max_new_tokens: 128
225
+ saves_per_epoch: 1
226
+ debug:
227
+
228
+ deepspeed: deepspeed_configs/zero3_bf16.json
229
+ weight_decay: 0.05
230
+ fsdp:
231
+ fsdp_config:
232
+ special_tokens:
233
+ eos_token: "<|im_end|>"
234
+ pad_token: "<|end_of_text|>"
235
+ tokens:
236
+ - "<|im_start|>"
237
+ - "<|im_end|>"
238
+ ```
239
+
240
+ </details><br>
241
+
242
+ # 💬 Prompt Template
243
+
244
+ You can use ChatML prompt template while using the model:
245
+
246
+ ### ChatML
247
+
248
+ ```
249
+ <|im_start|>system
250
+ {system}<|im_end|>
251
+ <|im_start|>user
252
+ {user}<|im_end|>
253
+ <|im_start|>assistant
254
+ {asistant}<|im_end|>
255
+ ```
256
+
257
+ This prompt template is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating), which means you can format messages using the
258
+ `tokenizer.apply_chat_template()` method:
259
+
260
+ ```python
261
+ messages = [
262
+ {"role": "system", "content": "You are helpful AI asistant."},
263
+ {"role": "user", "content": "Hello!"}
264
+ ]
265
+ gen_input = tokenizer.apply_chat_template(message, return_tensors="pt")
266
+ model.generate(**gen_input)
267
+ ```
268
+
269
+ # 📊 Datasets used in this model
270
+
271
+ The datasets used to train this model are listed in the metadata section of the model card.
272
+
273
+ Please note that certain datasets mentioned in the metadata may have undergone filtering based on various criteria.
274
+
275
+ The results of this filtering process and its outcomes are in a diffrent repository:
276
+
277
+ [Weyaxi/sci-datasets/main](https://huggingface.co/datasets/Weyaxi/sci-datasets/tree/main)
278
+
279
+
280
+ # 🎯 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
281
+
282
+ # 🤖 Additional information about training
283
+
284
+ This model is full fine-tuned for 2 epoch.
285
+
286
+ Total number of steps was 500.
287
+
288
+ <details><summary>Loss graph</summary>
289
+
290
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/bkJGgh_JUfKeRlTLo_ZcB.png)
291
+
292
+ </details><br>