Update README.md
Browse files
README.md
CHANGED
@@ -1,9 +1,11 @@
|
|
1 |
---
|
2 |
datasets:
|
3 |
- HuggingFaceH4/ultrachat_200k
|
4 |
-
-
|
5 |
- meta-math/MetaMathQA
|
6 |
- WizardLM/WizardLM_evol_instruct_V2_196k
|
|
|
|
|
7 |
- Intel/orca_dpo_pairs
|
8 |
language:
|
9 |
- en
|
@@ -17,16 +19,16 @@ extra_gated_fields:
|
|
17 |
I ALLOW Stability AI to email me about new model releases: checkbox
|
18 |
license: other
|
19 |
---
|
20 |
-
# `StableLM Zephyr
|
21 |
|
22 |
## Model Description
|
23 |
|
24 |
-
`StableLM Zephyr
|
25 |
-
[MT Bench](https://
|
26 |
|
27 |
## Usage
|
28 |
|
29 |
-
`StableLM Zephyr
|
30 |
```
|
31 |
<|user|>
|
32 |
List 3 synonyms for the word "tiny"<|endoftext|>
|
|
|
1 |
---
|
2 |
datasets:
|
3 |
- HuggingFaceH4/ultrachat_200k
|
4 |
+
- allenai/ultrafeedback_binarized_cleaned
|
5 |
- meta-math/MetaMathQA
|
6 |
- WizardLM/WizardLM_evol_instruct_V2_196k
|
7 |
+
- openchat/openchat_sharegpt4_dataset
|
8 |
+
- LDJnr/Capybara
|
9 |
- Intel/orca_dpo_pairs
|
10 |
language:
|
11 |
- en
|
|
|
19 |
I ALLOW Stability AI to email me about new model releases: checkbox
|
20 |
license: other
|
21 |
---
|
22 |
+
# `StableLM 2 Zephyr 1.6B`
|
23 |
|
24 |
## Model Description
|
25 |
|
26 |
+
`StableLM 2 Zephyr 1.6B` is a 1.6 billion parameter instruction tuned inspired by [Stablelm Zephyr 1.6B](https://huggingface.co/stabilityai/stablelm-zephyr-3b) training pipeline this model was trained on a mix of publicly available datasets, synthetic datasets using [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290), evaluation for this model based on
|
27 |
+
[MT Bench](https://huggingface.co/spaces/lmsys/mt-bench).
|
28 |
|
29 |
## Usage
|
30 |
|
31 |
+
`StableLM 2 Zephyr 1.6B` uses the following instruction format:
|
32 |
```
|
33 |
<|user|>
|
34 |
List 3 synonyms for the word "tiny"<|endoftext|>
|