Teja-Gollapudi
commited on
Commit
•
602a29a
1
Parent(s):
c0efbe8
Update README.md
Browse files
README.md
CHANGED
@@ -11,7 +11,7 @@ pipeline_tag: conversational
|
|
11 |
# VMware/open-llama-0.3T-7B-open-instruct-v1.1
|
12 |
|
13 |
## License
|
14 |
-
- Commercially
|
15 |
- Instruction dataset, [VMware/open-instruct-v1.1-oasst-dolly-hhrlhf](https://huggingface.co/datasets/VMware/open-instruct-v1.1-oasst-dolly-hhrlhf) is under cc-by-sa-3.0
|
16 |
- Language Model ([openlm-research/open_llama_7b_preview_300bt](https://huggingface.co/openlm-research/open_llama_7b_preview_300bt/tree/main/open_llama_7b_preview_300bt_transformers_weights)) is under apache-2.0 License
|
17 |
|
@@ -65,7 +65,11 @@ This way, the model can better understand the relationship between different par
|
|
65 |
|
66 |
## Drawbacks
|
67 |
|
68 |
-
- The model was trained on a partially trained Open-LLaMA checkpoint
|
|
|
|
|
|
|
|
|
69 |
|
70 |
## Evaluation
|
71 |
|
|
|
11 |
# VMware/open-llama-0.3T-7B-open-instruct-v1.1
|
12 |
|
13 |
## License
|
14 |
+
- <b>Commercially Viable </b>
|
15 |
- Instruction dataset, [VMware/open-instruct-v1.1-oasst-dolly-hhrlhf](https://huggingface.co/datasets/VMware/open-instruct-v1.1-oasst-dolly-hhrlhf) is under cc-by-sa-3.0
|
16 |
- Language Model ([openlm-research/open_llama_7b_preview_300bt](https://huggingface.co/openlm-research/open_llama_7b_preview_300bt/tree/main/open_llama_7b_preview_300bt_transformers_weights)) is under apache-2.0 License
|
17 |
|
|
|
65 |
|
66 |
## Drawbacks
|
67 |
|
68 |
+
- The model was trained on a partially trained Open-LLaMA checkpoint (300B tokens or 30% training life cycle), there is a huge potential for improvement when trained on fully trained Open-LLaMA checkpoints
|
69 |
+
- From what we have observed, the model strugles with few-shot prompting (We plan on addressing it with future iterations)
|
70 |
+
- When asked for code, it may or maynot output it with markdown format
|
71 |
+
- It doesn't indent python code
|
72 |
+
|
73 |
|
74 |
## Evaluation
|
75 |
|