Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: cc-by-nc-sa-4.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
datasets:
|
6 |
+
- garage-bAInd/Open-Platypus
|
7 |
+
---
|
8 |
+
Some GGUF v2 quantizations of the model [RobbeD/OpenLlama-Platypus-3B](https://huggingface.co/RobbeD/OpenLlama-Platypus-3B)
|
9 |
+
|
10 |
+
# OpenLlama-Platypus-3B
|
11 |
+
|
12 |
+
OpenLlama-Platypus-3B is an instruction fine-tuned model based on the OpenLLaMA-3B transformer architecture.
|
13 |
+
|
14 |
+
### Model Details
|
15 |
+
|
16 |
+
* **Trained by**: Robbe De Sutter
|
17 |
+
* **Model type:** **OpenLlama-Platypus-3B** is an auto-regressive language model based on the OpenLLaMA-3B transformer architecture.
|
18 |
+
* **Language(s)**: English
|
19 |
+
* **License for base weights**: Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
|
20 |
+
|
21 |
+
### Prompt Template
|
22 |
+
```
|
23 |
+
### Instruction:
|
24 |
+
<prompt> (without the <>)
|
25 |
+
### Response:
|
26 |
+
```
|
27 |
+
|
28 |
+
### Training Dataset
|
29 |
+
|
30 |
+
`RobbeD/OpenLlama-Platypus-3B` trained using STEM and logic based dataset [`garage-bAInd/Open-Platypus`](https://huggingface.co/datasets/garage-bAInd/Open-Platypus).
|
31 |
+
|
32 |
+
Please see their [paper](https://arxiv.org/abs/2308.07317) and [project webpage](https://platypus-llm.github.io) for additional information.
|
33 |
+
|
34 |
+
### Training Procedure
|
35 |
+
|
36 |
+
`RobbeD/OpenLlama-Platypus-3B` was instruction fine-tuned using LoRA on 1 RX 6900 XT 16GB.
|