TheBloke commited on
Commit
65e9907
1 Parent(s): dc85d42

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -10
README.md CHANGED
@@ -5,7 +5,17 @@ license: apache-2.0
5
  model_creator: Intel
6
  model_name: Neural Chat 7B v3-1
7
  model_type: mistral
8
- prompt_template: '{prompt}
 
 
 
 
 
 
 
 
 
 
9
 
10
  '
11
  quantized_by: TheBloke
@@ -64,11 +74,17 @@ It is supported by:
64
  <!-- repositories-available end -->
65
 
66
  <!-- prompt-template start -->
67
- ## Prompt template: Unknown
68
 
69
  ```
 
 
 
 
70
  {prompt}
71
 
 
 
72
  ```
73
 
74
  <!-- prompt-template end -->
@@ -133,7 +149,13 @@ prompts = [
133
  "What is 291 - 150?",
134
  "How much wood would a woodchuck chuck if a woodchuck could chuck wood?",
135
  ]
136
- prompt_template=f'''{prompt}
 
 
 
 
 
 
137
  '''
138
 
139
  prompts = [prompt_template.format(prompt=prompt) for prompt in prompts]
@@ -175,7 +197,13 @@ from huggingface_hub import InferenceClient
175
  endpoint_url = "https://your-endpoint-url-here"
176
 
177
  prompt = "Tell me about AI"
178
- prompt_template=f'''{prompt}
 
 
 
 
 
 
179
  '''
180
 
181
  client = InferenceClient(endpoint_url)
@@ -238,7 +266,13 @@ model = AutoModelForCausalLM.from_pretrained(
238
  streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)
239
 
240
  prompt = "Tell me about AI"
241
- prompt_template=f'''{prompt}
 
 
 
 
 
 
242
  '''
243
 
244
  # Convert prompt to tokens
@@ -340,9 +374,9 @@ And thank you again to a16z for their generous grant.
340
  # Original model card: Intel's Neural Chat 7B v3-1
341
 
342
 
343
- ## Finetuning on [habana](https://habana.ai/) HPU
344
 
345
- This model is a fine-tuned model based on [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the open source dataset [Open-Orca/SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca). Then we align it with DPO algorithm. For more details, you can refer our blog: [NeuralChat: Simplifying Supervised Instruction Fine-Tuning and Reinforcement Aligning](https://medium.com/intel-analytics-software/neuralchat-simplifying-supervised-instruction-fine-tuning-and-reinforcement-aligning-for-chatbots-d034bca44f69) and [The Practice of Supervised Fine-tuning and Direct Preference Optimization on Habana Gaudi2](https://medium.com/@NeuralCompressor/the-practice-of-supervised-finetuning-and-direct-preference-optimization-on-habana-gaudi2-a1197d8a3cd3).
346
 
347
  ## Model date
348
  Neural-chat-7b-v3-1 was trained between September and October, 2023.
@@ -372,10 +406,22 @@ The following hyperparameters were used during training:
372
  - total_train_batch_size: 64
373
  - total_eval_batch_size: 8
374
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
375
- - lr_scheduler_type: linear
376
- - lr_scheduler_warmup_ratio: 0.02
377
  - num_epochs: 2.0
378
 
 
 
 
 
 
 
 
 
 
 
 
 
379
  ## Inference with transformers
380
 
381
  ```shell
@@ -401,4 +447,3 @@ The NeuralChat team with members from Intel/SATG/AIA/AIPT. Core team members: Ka
401
  ## Useful links
402
  * Intel Neural Compressor [link](https://github.com/intel/neural-compressor)
403
  * Intel Extension for Transformers [link](https://github.com/intel/intel-extension-for-transformers)
404
- * Intel Extension for PyTorch [link](https://github.com/intel/intel-extension-for-pytorch)
 
5
  model_creator: Intel
6
  model_name: Neural Chat 7B v3-1
7
  model_type: mistral
8
+ prompt_template: '### System:
9
+
10
+ {system_message}
11
+
12
+
13
+ ### User:
14
+
15
+ {prompt}
16
+
17
+
18
+ ### Assistant:
19
 
20
  '
21
  quantized_by: TheBloke
 
74
  <!-- repositories-available end -->
75
 
76
  <!-- prompt-template start -->
77
+ ## Prompt template: Orca-Hashes
78
 
79
  ```
80
+ ### System:
81
+ {system_message}
82
+
83
+ ### User:
84
  {prompt}
85
 
86
+ ### Assistant:
87
+
88
  ```
89
 
90
  <!-- prompt-template end -->
 
149
  "What is 291 - 150?",
150
  "How much wood would a woodchuck chuck if a woodchuck could chuck wood?",
151
  ]
152
+ prompt_template=f'''### System:
153
+ {system_message}
154
+
155
+ ### User:
156
+ {prompt}
157
+
158
+ ### Assistant:
159
  '''
160
 
161
  prompts = [prompt_template.format(prompt=prompt) for prompt in prompts]
 
197
  endpoint_url = "https://your-endpoint-url-here"
198
 
199
  prompt = "Tell me about AI"
200
+ prompt_template=f'''### System:
201
+ {system_message}
202
+
203
+ ### User:
204
+ {prompt}
205
+
206
+ ### Assistant:
207
  '''
208
 
209
  client = InferenceClient(endpoint_url)
 
266
  streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)
267
 
268
  prompt = "Tell me about AI"
269
+ prompt_template=f'''### System:
270
+ {system_message}
271
+
272
+ ### User:
273
+ {prompt}
274
+
275
+ ### Assistant:
276
  '''
277
 
278
  # Convert prompt to tokens
 
374
  # Original model card: Intel's Neural Chat 7B v3-1
375
 
376
 
377
+ ## Fine-tuning on [Habana](https://habana.ai/) Gaudi2
378
 
379
+ This model is a fine-tuned model based on [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the open source dataset [Open-Orca/SlimOrca](https://huggingface.co/datasets/Open-Orca/SlimOrca). Then we align it with DPO algorithm. For more details, you can refer our blog: [The Practice of Supervised Fine-tuning and Direct Preference Optimization on Habana Gaudi2](https://medium.com/@NeuralCompressor/the-practice-of-supervised-finetuning-and-direct-preference-optimization-on-habana-gaudi2-a1197d8a3cd3).
380
 
381
  ## Model date
382
  Neural-chat-7b-v3-1 was trained between September and October, 2023.
 
406
  - total_train_batch_size: 64
407
  - total_eval_batch_size: 8
408
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
409
+ - lr_scheduler_type: cosine
410
+ - lr_scheduler_warmup_ratio: 0.03
411
  - num_epochs: 2.0
412
 
413
+ ## Prompt Template
414
+
415
+ ```
416
+ ### System:
417
+ {system}
418
+ ### User:
419
+ {usr}
420
+ ### Assistant:
421
+
422
+ ```
423
+
424
+
425
  ## Inference with transformers
426
 
427
  ```shell
 
447
  ## Useful links
448
  * Intel Neural Compressor [link](https://github.com/intel/neural-compressor)
449
  * Intel Extension for Transformers [link](https://github.com/intel/intel-extension-for-transformers)