Update README.md
Browse files
README.md
CHANGED
@@ -122,6 +122,8 @@ model-index:
|
|
122 |
---
|
123 |
# `StableLM Zephyr 3B`
|
124 |
|
|
|
|
|
125 |
## Model Description
|
126 |
|
127 |
`StableLM Zephyr 3B` is a 3 billion parameter instruction tuned inspired by [HugginFaceH4's Zephyr 7B](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) training pipeline this model was trained on a mix of publicly available datasets, synthetic datasets using [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290), evaluation for this model based on
|
@@ -176,7 +178,8 @@ You can also see how to run a performance optimized version of this model [here]
|
|
176 |
* **Language(s)**: English
|
177 |
* **Library**: [Alignment Handbook](https://github.com/huggingface/alignment-handbook.git)
|
178 |
* **Finetuned from model**: [stabilityai/stablelm-3b-4e1t](https://huggingface.co/stabilityai/stablelm-3b-4e1t)
|
179 |
-
* **License**: [StabilityAI Non-Commercial Research Community License](https://huggingface.co/stabilityai/stablelm-zephyr-3b/raw/main/LICENSE).
|
|
|
180 |
* **Contact**: For questions and comments about the model, please email `lm@stability.ai`
|
181 |
|
182 |
### Training Dataset
|
@@ -247,7 +250,7 @@ We have incorporated the findings of our malicious prompts evaluation and red te
|
|
247 |
|
248 |
### Intended Use
|
249 |
|
250 |
-
The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications.
|
251 |
|
252 |
### Limitations and Bias
|
253 |
|
|
|
122 |
---
|
123 |
# `StableLM Zephyr 3B`
|
124 |
|
125 |
+
Please note: For commercial use, please refer to https://stability.ai/membership.
|
126 |
+
|
127 |
## Model Description
|
128 |
|
129 |
`StableLM Zephyr 3B` is a 3 billion parameter instruction tuned inspired by [HugginFaceH4's Zephyr 7B](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) training pipeline this model was trained on a mix of publicly available datasets, synthetic datasets using [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290), evaluation for this model based on
|
|
|
178 |
* **Language(s)**: English
|
179 |
* **Library**: [Alignment Handbook](https://github.com/huggingface/alignment-handbook.git)
|
180 |
* **Finetuned from model**: [stabilityai/stablelm-3b-4e1t](https://huggingface.co/stabilityai/stablelm-3b-4e1t)
|
181 |
+
* **License**: [StabilityAI Non-Commercial Research Community License](https://huggingface.co/stabilityai/stablelm-zephyr-3b/raw/main/LICENSE).
|
182 |
+
* **Commercial License**: to use this model commercially, please refer to https://stability.ai/membership
|
183 |
* **Contact**: For questions and comments about the model, please email `lm@stability.ai`
|
184 |
|
185 |
### Training Dataset
|
|
|
250 |
|
251 |
### Intended Use
|
252 |
|
253 |
+
The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications. For commercial use, please refer to https://stability.ai/membership.
|
254 |
|
255 |
### Limitations and Bias
|
256 |
|