andrijdavid
commited on
Commit
β’
600c358
1
Parent(s):
e3ce549
Upload folder using huggingface_hub
Browse files- .gitattributes +14 -0
- Q2_K/Q2_K-00001-of-00001.gguf +3 -0
- Q3_K_L/Q3_K_L-00001-of-00001.gguf +3 -0
- Q3_K_M/Q3_K_M-00001-of-00001.gguf +3 -0
- Q3_K_S/Q3_K_S-00001-of-00001.gguf +3 -0
- Q4_0/Q4_0-00001-of-00001.gguf +3 -0
- Q4_1/Q4_1-00001-of-00001.gguf +3 -0
- Q4_K_M/Q4_K_M-00001-of-00001.gguf +3 -0
- Q4_K_S/Q4_K_S-00001-of-00001.gguf +3 -0
- Q5_0/Q5_0-00001-of-00001.gguf +3 -0
- Q5_1/Q5_1-00001-of-00001.gguf +3 -0
- Q5_K_M/Q5_K_M-00001-of-00001.gguf +3 -0
- Q5_K_S/Q5_K_S-00001-of-00001.gguf +3 -0
- Q6_K/Q6_K-00001-of-00001.gguf +3 -0
- Q8_0/Q8_0-00001-of-00001.gguf +3 -0
- README.md +2 -2
.gitattributes
CHANGED
@@ -33,3 +33,17 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
Q2_K/Q2_K-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
+
Q3_K_L/Q3_K_L-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
Q3_K_M/Q3_K_M-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
Q3_K_S/Q3_K_S-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
+
Q4_0/Q4_0-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
+
Q4_1/Q4_1-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
+
Q4_K_M/Q4_K_M-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
43 |
+
Q4_K_S/Q4_K_S-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
+
Q5_0/Q5_0-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
45 |
+
Q5_1/Q5_1-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
46 |
+
Q5_K_M/Q5_K_M-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
47 |
+
Q5_K_S/Q5_K_S-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
48 |
+
Q6_K/Q6_K-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
49 |
+
Q8_0/Q8_0-00001-of-00001.gguf filter=lfs diff=lfs merge=lfs -text
|
Q2_K/Q2_K-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c279c6a384054097cb6e9f47c806cf052b0eb93f93b8005d428abcc89fe4fcf7
|
3 |
+
size 3179131552
|
Q3_K_L/Q3_K_L-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e2920e839fe9021e419c873d251f5433b1f599cf1c7dd7cbc9c698bc2634d6fd
|
3 |
+
size 4321956512
|
Q3_K_M/Q3_K_M-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2a7cf69affeac23daa871025553e5e874b6d97784915041c99639bb895d34bc8
|
3 |
+
size 4018918048
|
Q3_K_S/Q3_K_S-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:631644cc46ea9f37b30125956f9b48272cdaa98d67fef5fa77c520e303ac268d
|
3 |
+
size 3664499360
|
Q4_0/Q4_0-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c6a35c00757b4d14873beb7287fe86a86a103b1bf6bc6c10a969fd85991a5b79
|
3 |
+
size 4661211808
|
Q4_1/Q4_1-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:889aaa97b33752c0a7dc97a9999aa1abcfba4c7704967ae1766a9fb8bb5e2025
|
3 |
+
size 5130252960
|
Q4_K_M/Q4_K_M-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d48bddcfd0d041829ede7f638621d2419cd4a92d70b151e5690fd3fc89706c4d
|
3 |
+
size 4920734368
|
Q4_K_S/Q4_K_S-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:80b520e82b6cdaf56a0f47ecb24bf11a8398b4dfc123eee7233b3c0d915f622f
|
3 |
+
size 4692669088
|
Q5_0/Q5_0-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:74b41da81fb012b15695dca1b0f9df2979ed7c133a1b85d9ede62f699c3a705d
|
3 |
+
size 5599294112
|
Q5_1/Q5_1-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3f8b74adf1c164bd249cdc305a67241142bb10869ff9faf6d97c5579989e347
|
3 |
+
size 6068335264
|
Q5_K_M/Q5_K_M-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:270011542dbf93668f4e32b220001224e95e0517c6c24cc6067061dd8cfe2360
|
3 |
+
size 5732987552
|
Q5_K_S/Q5_K_S-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f3c7f94f20eb71f2ac522fb62d12c93501d260198c9c3e9d04cd5b15949d3cf7
|
3 |
+
size 5599294112
|
Q6_K/Q6_K-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5c2aeb8959c52b7a031364c1a915ca3f0c9f4ecb21e1d71629f6240abee422b0
|
3 |
+
size 6596006560
|
Q8_0/Q8_0-00001-of-00001.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba8885d5ead93741afb51284f25dc99a3745e19b6e03909d5cd74cdd8b629ad7
|
3 |
+
size 8540770976
|
README.md
CHANGED
@@ -211,10 +211,10 @@ Here are guides on using llama-cpp-python and ctransformers with LangChain:
|
|
211 |
|
212 |
|
213 |
## Model Details
|
214 |
-
We introduce Llama3-ChatQA-1.5, which excels at conversational question answering (QA) and retrieval-augmented generation (RAG). Llama3-ChatQA-1.5 is developed using an improved training recipe from [ChatQA paper](https://arxiv.org/pdf/2401.
|
215 |
|
216 |
## Other Resources
|
217 |
-
[Llama3-ChatQA-1.5-70B](https://huggingface.co/nvidia/Llama3-ChatQA-1.5-70B) β [Evaluation Data](https://huggingface.co/datasets/nvidia/ChatRAG-Bench) β [Training Data](https://huggingface.co/datasets/nvidia/ChatQA-Training-Data) β [Retriever](https://huggingface.co/nvidia/dragon-multiturn-query-encoder) β [Website](https://chatqa-project.github.io/) β [Paper](https://arxiv.org/pdf/2401.
|
218 |
|
219 |
## Benchmark Results
|
220 |
Results in [ChatRAG Bench](https://huggingface.co/datasets/nvidia/ChatRAG-Bench) are as follows:
|
|
|
211 |
|
212 |
|
213 |
## Model Details
|
214 |
+
We introduce Llama3-ChatQA-1.5, which excels at conversational question answering (QA) and retrieval-augmented generation (RAG). Llama3-ChatQA-1.5 is developed using an improved training recipe from [ChatQA paper](https://arxiv.org/pdf/2401.10225), and it is built on top of [Llama-3 base model](https://huggingface.co/meta-llama/Meta-Llama-3-8B). Specifically, we incorporate more conversational QA data to enhance its tabular and arithmetic calculation capability. Llama3-ChatQA-1.5 has two variants: Llama3-ChatQA-1.5-8B and Llama3-ChatQA-1.5-70B. Both models were originally trained using [Megatron-LM](https://github.com/NVIDIA/Megatron-LM), we converted the checkpoints to Hugging Face format. **For more information about ChatQA, check the [website](https://chatqa-project.github.io/)!**
|
215 |
|
216 |
## Other Resources
|
217 |
+
[Llama3-ChatQA-1.5-70B](https://huggingface.co/nvidia/Llama3-ChatQA-1.5-70B) β [Evaluation Data](https://huggingface.co/datasets/nvidia/ChatRAG-Bench) β [Training Data](https://huggingface.co/datasets/nvidia/ChatQA-Training-Data) β [Retriever](https://huggingface.co/nvidia/dragon-multiturn-query-encoder) β [Website](https://chatqa-project.github.io/) β [Paper](https://arxiv.org/pdf/2401.10225)
|
218 |
|
219 |
## Benchmark Results
|
220 |
Results in [ChatRAG Bench](https://huggingface.co/datasets/nvidia/ChatRAG-Bench) are as follows:
|