ddellapietra
commited on
Commit
•
8f60d8e
1
Parent(s):
4b5a955
Update README.md
Browse files
README.md
CHANGED
@@ -6,7 +6,9 @@ license: mit
|
|
6 |
|
7 |
The SciPhi-SearchAgent-Alpha-7B is a Large Language Model (LLM) fine-tuned from Mistral-7B-v0.1. This model underwent a fine-tuning process using retrieval-augmented generation (RAG) over search with a fully synthetic dataset. The objective of this work is to generate accurate and well-cited summaries from a range of search results, providing more accurate answers to user queries. For best results, follow the prompting guidelines below.
|
8 |
|
9 |
-
SciPhi-AI is available via a free hosted API, though the exposed model can vary. Currently, SciPhi-SearchAgent-Alpha-7B is available. More details can be found in the docs [here](https://
|
|
|
|
|
10 |
|
11 |
## Model Architecture
|
12 |
|
@@ -37,12 +39,5 @@ See the documentation, linked above, for more information.
|
|
37 |
|
38 |
## References
|
39 |
|
40 |
-
1.
|
41 |
-
2. Mukherjee, S., Mitra, A., Jawahar, G., Agarwal, S., Palangi, H., & Awadallah, A. (2023). Orca: Progressive Learning from Complex Explanation Traces of GPT-4. *arXiv preprint arXiv:2306.02707*.
|
42 |
-
3. Longpre, S., Hou, L., Vu, T., Webson, A., Chung, H. W., Tay, Y., Zhou, D., Le, Q. V., Zoph, B., Wei, J., & Roberts, A. (2023). The Flan Collection: Designing Data and Methods for Effective Instruction Tuning. *arXiv preprint arXiv:2301.13688*.
|
43 |
-
4. Mistral AI. (2023). Model Card for Mistral-7B-v0.1. The Mistral-7B-v0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters. Mistral-7B-v0.1 outperforms Llama 2 13B on all benchmarks tested. For full details, please refer to the paper and release blog post. Model Architecture: Transformer with Grouped-Query Attention, Sliding-Window Attention, and Byte-fallback BPE tokenizer. [Link](https://huggingface.co/mistralai/Mistral-7B-v0.1)
|
44 |
-
|
45 |
-
|
46 |
-
## Acknowledgements
|
47 |
|
48 |
-
Thank you to the [AI Alignment Lab](https://huggingface.co/Alignment-Lab-AI), [vikp](https://huggingface.co/vikp), [jph00](https://huggingface.co/jph00) and others who contributed to this work.
|
|
|
6 |
|
7 |
The SciPhi-SearchAgent-Alpha-7B is a Large Language Model (LLM) fine-tuned from Mistral-7B-v0.1. This model underwent a fine-tuning process using retrieval-augmented generation (RAG) over search with a fully synthetic dataset. The objective of this work is to generate accurate and well-cited summaries from a range of search results, providing more accurate answers to user queries. For best results, follow the prompting guidelines below.
|
8 |
|
9 |
+
SciPhi-AI is available via a free hosted API, though the exposed model can vary. Currently, SciPhi-SearchAgent-Alpha-7B is available. More details can be found in the docs [here](https://agent-search.readthedocs.io/en/latest/).
|
10 |
+
|
11 |
+
The search can be accessed directly [here](https://search.sciphi.ai/).
|
12 |
|
13 |
## Model Architecture
|
14 |
|
|
|
39 |
|
40 |
## References
|
41 |
|
42 |
+
1. Mistral AI. (2023). Model Card for Mistral-7B-v0.1. The Mistral-7B-v0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters. Mistral-7B-v0.1 outperforms Llama 2 13B on all benchmarks tested. For full details, please refer to the paper and release blog post. Model Architecture: Transformer with Grouped-Query Attention, Sliding-Window Attention, and Byte-fallback BPE tokenizer. [Link](https://huggingface.co/mistralai/Mistral-7B-v0.1)
|
|
|
|
|
|
|
|
|
|
|
|
|
43 |
|
|