Update README.md
Browse files
README.md
CHANGED
@@ -10,26 +10,29 @@ license: cc-by-nc-4.0
|
|
10 |
|
11 |
OpenOrca-Platypus2-13B is a merge of [`garage-bAInd/Platypus2-13B`](https://huggingface.co/garage-bAInd/Platypus2-13B) and [`Open-Orca/OpenOrcaxOpenChat-Preview2-13B`](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B).
|
12 |
|
|
|
|
|
13 |
![Platty](./Best_Platty_small.jpeg)
|
14 |
|
15 |
### Benchmark Metrics
|
16 |
|
17 |
| Metric | Value |
|
18 |
|-----------------------|-------|
|
19 |
-
| MMLU (5-shot) |
|
20 |
-
| ARC (25-shot) |
|
21 |
-
| HellaSwag (10-shot) |
|
22 |
-
| TruthfulQA (0-shot) |
|
23 |
-
| Avg. |
|
24 |
|
25 |
We use state-of-the-art [Language Model Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness) to run the benchmark tests above, using the same version as the HuggingFace LLM Leaderboard. Please see below for detailed instructions on reproducing benchmark results.
|
26 |
|
27 |
### Model Details
|
28 |
|
29 |
-
* **Trained by**: **Platypus2-13B** trained by Cole Hunter & Ariel Lee; **OpenOrcaxOpenChat-Preview2-13B** trained by
|
30 |
* **Model type:** **OpenOrca-Platypus2-13B** is an auto-regressive language model based on the LLaMA 2 transformer architecture.
|
31 |
* **Language(s)**: English
|
32 |
* **License for Platypus2-13B base weights**: Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
|
|
|
33 |
|
34 |
### Prompt Template for base Platypus2-13B
|
35 |
```
|
@@ -39,13 +42,17 @@ We use state-of-the-art [Language Model Evaluation Harness](https://github.com/E
|
|
39 |
|
40 |
### Response:
|
41 |
```
|
|
|
|
|
42 |
|
43 |
-
### Training
|
44 |
|
45 |
`garage-bAInd/Platypus2-13B` trained using STEM and logic based dataset [`garage-bAInd/Open-Platypus`](https://huggingface.co/datasets/garage-bAInd/Open-Platypus).
|
46 |
|
47 |
Please see our [paper](https://platypus-llm.github.io/Platypus.pdf) and [project webpage](https://platypus-llm.github.io) for additional information.
|
48 |
|
|
|
|
|
49 |
### Training Procedure
|
50 |
|
51 |
`garage-bAInd/Platypus2-13B` was instruction fine-tuned using LoRA on 1 A100 80GB. For training details and inference instructions please see the [Platypus](https://github.com/arielnlee/Platypus) GitHub repo.
|
|
|
10 |
|
11 |
OpenOrca-Platypus2-13B is a merge of [`garage-bAInd/Platypus2-13B`](https://huggingface.co/garage-bAInd/Platypus2-13B) and [`Open-Orca/OpenOrcaxOpenChat-Preview2-13B`](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B).
|
12 |
|
13 |
+
Thank you Open-Orca for putting out a beast of a model and dataset—can't wait for the 70B version (and beyond)!
|
14 |
+
|
15 |
![Platty](./Best_Platty_small.jpeg)
|
16 |
|
17 |
### Benchmark Metrics
|
18 |
|
19 |
| Metric | Value |
|
20 |
|-----------------------|-------|
|
21 |
+
| MMLU (5-shot) | 59.5 |
|
22 |
+
| ARC (25-shot) | 62.88 |
|
23 |
+
| HellaSwag (10-shot) | 83.19 |
|
24 |
+
| TruthfulQA (0-shot) | 52.69 |
|
25 |
+
| Avg. | 64.56 |
|
26 |
|
27 |
We use state-of-the-art [Language Model Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness) to run the benchmark tests above, using the same version as the HuggingFace LLM Leaderboard. Please see below for detailed instructions on reproducing benchmark results.
|
28 |
|
29 |
### Model Details
|
30 |
|
31 |
+
* **Trained by**: **Platypus2-13B** trained by Cole Hunter & Ariel Lee; **OpenOrcaxOpenChat-Preview2-13B** trained by Open-Orca
|
32 |
* **Model type:** **OpenOrca-Platypus2-13B** is an auto-regressive language model based on the LLaMA 2 transformer architecture.
|
33 |
* **Language(s)**: English
|
34 |
* **License for Platypus2-13B base weights**: Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
|
35 |
+
* **License for OpenOrcaxOpenChat-Preview2-13B base weights**: LLaMa-2 commercial
|
36 |
|
37 |
### Prompt Template for base Platypus2-13B
|
38 |
```
|
|
|
42 |
|
43 |
### Response:
|
44 |
```
|
45 |
+
### Prompt Template for base OpenOrcaxOpenChat-Preview2-13B
|
46 |
+
OpenChat Llama2 V1: see [Open-Orca's page](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B) for additional information.
|
47 |
|
48 |
+
### Training Datasets
|
49 |
|
50 |
`garage-bAInd/Platypus2-13B` trained using STEM and logic based dataset [`garage-bAInd/Open-Platypus`](https://huggingface.co/datasets/garage-bAInd/Open-Platypus).
|
51 |
|
52 |
Please see our [paper](https://platypus-llm.github.io/Platypus.pdf) and [project webpage](https://platypus-llm.github.io) for additional information.
|
53 |
|
54 |
+
[`Open-Orca/OpenOrcaxOpenChat-Preview2-13B`] trained using a refined, 220k subset of the [OpenOrca dataset](https://huggingface.co/datasets/Open-Orca/OpenOrca).
|
55 |
+
|
56 |
### Training Procedure
|
57 |
|
58 |
`garage-bAInd/Platypus2-13B` was instruction fine-tuned using LoRA on 1 A100 80GB. For training details and inference instructions please see the [Platypus](https://github.com/arielnlee/Platypus) GitHub repo.
|