Update README.md
Browse files
README.md
CHANGED
@@ -87,9 +87,15 @@ inference:
|
|
87 |
|
88 |
# bart-base-instructiongen
|
89 |
|
90 |
-
Instead of generating questions from text, generate instructions for LLMs!
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
91 |
|
92 |
-
This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on the pszemraj/fleece2instructions dataset.
|
93 |
It achieves the following results on the evaluation set:
|
94 |
- Loss: 1.0034
|
95 |
- Rouge1: 61.7209
|
@@ -100,7 +106,7 @@ It achieves the following results on the evaluation set:
|
|
100 |
|
101 |
## Intended uses & limitations
|
102 |
|
103 |
-
This is just a base model/example. There is likely to be even better performance with larger models.
|
104 |
|
105 |
Additionally, this was trained on a dataset of **only** instructions+outputs, with the `inputs` filtered out. This means that text of *1) cookies and cream 2) chocolate chip 3) mint chip 4) oreo* will **not** get you *"Rank the following ice cream flavors: oreo, mint chip, chocolate chip, cookies and cream"*.
|
106 |
|
|
|
87 |
|
88 |
# bart-base-instructiongen
|
89 |
|
90 |
+
Instead of generating questions from text, generate instructions for LLMs!
|
91 |
+
|
92 |
+
Check out a [basic demo on Spaces](https://huggingface.co/spaces/pszemraj/generate-instructions). You can find other models fine-tuned for instruction generation by [searching for the instructiongen tag](https://huggingface.co/models?other=instructiongen).
|
93 |
+
|
94 |
+
|
95 |
+
## About
|
96 |
+
|
97 |
+
This model is a fine-tuned version of the [facebook/bart-base](https://huggingface.co/facebook/bart-base) model, fine-tuned using the `pszemraj/fleece2instructions` dataset. The concept is to apply text-to-text models to unlabeled, domain-specific text to generate appropriate LLM instructions. Consequently, this facilitates domain adaptation of instruction-tuned LLMs, making them more versatile in their respective domains.
|
98 |
|
|
|
99 |
It achieves the following results on the evaluation set:
|
100 |
- Loss: 1.0034
|
101 |
- Rouge1: 61.7209
|
|
|
106 |
|
107 |
## Intended uses & limitations
|
108 |
|
109 |
+
This is just a base model/example. There is likely to be even better performance with larger models (click [here to see other checkpoints](https://huggingface.co/models?other=instructiongen))
|
110 |
|
111 |
Additionally, this was trained on a dataset of **only** instructions+outputs, with the `inputs` filtered out. This means that text of *1) cookies and cream 2) chocolate chip 3) mint chip 4) oreo* will **not** get you *"Rank the following ice cream flavors: oreo, mint chip, chocolate chip, cookies and cream"*.
|
112 |
|