metadata
base_model: HuggingFaceH4/mistral-7b-sft-beta
inference: false
license: mit
license_name: mit
license_link: https://huggingface.co/01-ai/Yi-34B/blob/main/LICENSE
metrics:
- accuracy
It works well with long system prompts.
It isn't generic in a sense that it shouldn't be used for story telling, for example, but only for reasoning and text comprehension.
This model is trained on a private dataset. The high GSM8K score is NOT because of the MetaMath dataset.
Prompt Format:
### System:
{add here the system prompt}
### Instruction:
{add here the instruction}
### Response:
{make sure you have a new line here}