Edit model card

Model Card for SuperQA-7B

This model is a fine-tuned version of Mistral-7B-v0.1, specifically designed for Question Answering (QA) tasks. It has been trained on a private dataset comprising 120,000 document, question, and answer pairs.

To my knowledge, this is the most capable 7B model for Retrieval Augmented Generation (RAG) tasks.

SuperQA responds in Markdown format.

Prompt Format

This model was trained only with the following prompt:

<s>[INST] Respond with a detailed and relevant answer to my question using only information from the provided context.

<|context|>
<|doc|>
{Your document}
<|/doc|>
<|/context|>

<|question|>{Your question?}<|/question|> [/INST]

Limitations

While the model is designed to be accurate and relevant, its performance is contingent on the quality and relevance of the provided context. Answers may be less accurate if the context is insufficient or not directly related to the question. Additionally, the model's training on a specific dataset may limit its effectiveness in answering questions outside the scope of the training data.

Disclaimer

This model is provided as-is without any guarantees of performance or accuracy. Users should not rely solely on this model for critical decisions or interpretations. The developers of this model are not responsible for any direct or indirect consequences arising from its use. It is the responsibility of the user to ensure that the model's output is appropriate for their specific context and requirements.

Downloads last month
13
Safetensors
Model size
7.24B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.