GGUF
Inference Endpoints
nchen909 commited on
Commit
e57d6b4
1 Parent(s): 64a9369

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,4 +33,17 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
- Apollo-7B-q8_0.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Apollo-7B.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Apollo-7B.IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Apollo-7B.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Apollo-7B.IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Apollo-7B.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Apollo-7B.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Apollo-7B.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Apollo-7B.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Apollo-7B.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Apollo-7B.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Apollo-7B.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Apollo-7B.IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text
48
+ Apollo-7B.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
49
+ Apollo-7B.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
Apollo-7B.IQ3_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6576863da9550ab4b5eb5b11dc1dd6a2e0cea23690ed7efc207dbb29e512fcf1
3
+ size 4106070560
Apollo-7B.IQ3_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8708c3e2e6a08b414ad7326dd0af28b69ab05c8463cc4c3190b5ebd96c785b64
3
+ size 3982404128
Apollo-7B.IQ3_XS.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a2bf3b7bd0f2b21e534296b4c901cb0c08ffb5119745105d4944bd5c88069bf
3
+ size 3800738336
Apollo-7B.IQ4_XS.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d8cb18127191fa2b9a49252515f833ac3bd59ce757777b739ae225ca38ba1ef
3
+ size 4805012000
Apollo-7B.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49b4f3962ad2cd42db45b9cca19ced28a84d6258859dc4d877254d22a02a7da9
3
+ size 3481446944
Apollo-7B.Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adb98d43f5a029ab54ab9be54e0f3a9ff3a6e362e663e703d54b609dc985d080
3
+ size 4709067296
Apollo-7B.Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5e5d6519340cc297d415c3c7ebb33e299167f11e7510188588781c235f65a39
3
+ size 4369328672
Apollo-7B.Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40c2bdb89adadfca79115e7d4bedbd5e309cf6a70ecf077d118ccdf0b09cf71d
3
+ size 3982404128
Apollo-7B.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8679d765963fedf57e231157e859ab619e12eeb2ea3da24e4a97d0a3245c26d5
3
+ size 5329758752
Apollo-7B.Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e68e055ee572fe73b61da5bce5e0b1796c02ad115e03a5f2fbce7ee1373e0c65
3
+ size 5046446624
Apollo-7B.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff93fe4ee0cb1afa5d6ceec23a031d57314e96686c03cb0a5f1921bba555779e
3
+ size 6144502304
Apollo-7B.Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a0ac6ac65639423892fd21088ed58e37f85c6921042706578865ce57aeb9da8
3
+ size 5980727840
Apollo-7B.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:365d3fa3233b201a2bfae7032cef37d99dc87ac9469d8a217fdb87e067e31da4
3
+ size 7010167328
Apollo-7B.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84328a9a66bea41eaff9ead79a19182ab9849cd1cc158dd390917e6adfd58a50
3
+ size 9077844512
README.md CHANGED
@@ -1,151 +1,66 @@
1
  ---
 
 
 
 
2
  license: apache-2.0
 
3
  ---
4
- # Multilingual Medicine: Model, Dataset, Benchmark, Code
5
 
6
- Covering English, Chinese, French, Hindi, Spanish, Hindi, Arabic So far
 
 
 
 
7
 
 
 
 
8
 
9
- <p align="center">
10
- 👨🏻‍💻<a href="https://github.com/FreedomIntelligence/Apollo" target="_blank">Github</a> •📃 <a href="https://arxiv.org/abs/2403.03640" target="_blank">Paper</a> • 🌐 <a href="https://apollo.llmzoo.com/" target="_blank">Demo</a> • 🤗 <a href="https://huggingface.co/datasets/FreedomIntelligence/ApolloCorpus" target="_blank">ApolloCorpus</a> • 🤗 <a href="https://huggingface.co/datasets/FreedomIntelligence/XMedbench" target="_blank">XMedBench</a>
11
- <br> <a href="./README_zh.md"> 中文 </a> | <a href="./README.md"> English
12
- </p>
13
 
14
- ![Apollo](assets/apollo_medium_final.png)
15
 
16
- ## 🌈 Update
17
 
18
- * **[2024.03.07]** [Paper](https://arxiv.org/abs/2403.03640) released.
19
- * **[2024.02.12]** <a href="https://huggingface.co/datasets/FreedomIntelligence/ApolloCorpus" target="_blank">ApolloCorpus</a> and <a href="https://huggingface.co/datasets/FreedomIntelligence/XMedbench" target="_blank">XMedBench</a> is published!🎉
20
- * **[2024.01.23]** Apollo repo is published!🎉
 
 
 
 
 
 
 
 
 
 
 
 
 
21
 
22
 
23
- ## Results
24
- 🤗<a href="https://huggingface.co/FreedomIntelligence/Apollo-0.5B" target="_blank">Apollo-0.5B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-1.8B" target="_blank">Apollo-1.8B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-2B" target="_blank">Apollo-2B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-6B" target="_blank">Apollo-6B</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-7B" target="_blank">Apollo-7B</a>
25
-
26
- 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-0.5B-GGUF" target="_blank">Apollo-0.5B-GGUF</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-2B-GGUF" target="_blank">Apollo-2B-GGUF</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-6B-GGUF" target="_blank">Apollo-6B-GGUF</a> • 🤗 <a href="https://huggingface.co/FreedomIntelligence/Apollo-7B-GGUF" target="_blank">Apollo-7B-GGUF</a>
27
-
28
-
29
- ![Apollo](assets/result.png)
30
-
31
-
32
-
33
 
 
34
 
35
- ## Dataset & Evaluation
 
36
 
37
- - Dataset
38
- 🤗 <a href="https://huggingface.co/datasets/FreedomIntelligence/ApolloCorpus" target="_blank">ApolloCorpus</a>
39
-
40
- <details><summary>Click to expand</summary>
41
 
42
- ![Apollo](assets/dataset.png)
 
43
 
44
- - [Zip File](https://huggingface.co/datasets/FreedomIntelligence/ApolloCorpus/blob/main/ApolloCorpus.zip)
45
- - [Data category](https://huggingface.co/datasets/FreedomIntelligence/ApolloCorpus/tree/main/train)
46
- - Pretrain:
47
- - data item:
48
- - json_name: {data_source}_{language}_{data_type}.json
49
- - data_type: medicalBook, medicalGuideline, medicalPaper, medicalWeb(from online forum), medicalWiki
50
- - language: en(English), zh(chinese), es(spanish), fr(french), hi(Hindi)
51
- - data_type: qa(generated qa from text)
52
- - data_type==text: list of string
53
- ```
54
- [
55
- "string1",
56
- "string2",
57
- ...
58
- ]
59
- ```
60
- - data_type==qa: list of qa pairs(list of string)
61
- ```
62
- [
63
- [
64
- "q1",
65
- "a1",
66
- "q2",
67
- "a2",
68
- ...
69
- ],
70
- ...
71
- ]
72
- ```
73
- - SFT:
74
- - json_name: {data_source}_{language}.json
75
- - data_type: code, general, math, medicalExam, medicalPatient
76
- - data item: list of qa pairs(list of string)
77
- ```
78
- [
79
- [
80
- "q1",
81
- "a1",
82
- "q2",
83
- "a2",
84
- ...
85
- ],
86
- ...
87
- ]
88
- ```
89
 
 
 
 
90
 
91
- </details>
92
-
93
-
94
-
95
- - Evaluation
96
- 🤗 <a href="https://huggingface.co/datasets/FreedomIntelligence/XMedbench" target="_blank">XMedBench</a>
97
-
98
- <details><summary>Click to expand</summary>
99
-
100
- - EN:
101
- - [MedQA-USMLE](https://huggingface.co/datasets/GBaker/MedQA-USMLE-4-options)
102
- - [MedMCQA](https://huggingface.co/datasets/medmcqa/viewer/default/test)
103
- - [PubMedQA](https://huggingface.co/datasets/pubmed_qa): Because the results fluctuated too much, they were not used in the paper.
104
- - [MMLU-Medical](https://huggingface.co/datasets/cais/mmlu)
105
- - Clinical knowledge, Medical genetics, Anatomy, Professional medicine, College biology, College medicine
106
- - ZH:
107
- - [MedQA-MCMLE](https://huggingface.co/datasets/bigbio/med_qa/viewer/med_qa_zh_4options_bigbio_qa/test)
108
- - [CMB-single](https://huggingface.co/datasets/FreedomIntelligence/CMB): Not used in the paper
109
- - Randomly sample 2,000 multiple-choice questions with single answer.
110
- - [CMMLU-Medical](https://huggingface.co/datasets/haonan-li/cmmlu)
111
- - Anatomy, Clinical_knowledge, College_medicine, Genetics, Nutrition, Traditional_chinese_medicine, Virology
112
- - [CExam](https://github.com/williamliujl/CMExam): Not used in the paper
113
- - Randomly sample 2,000 multiple-choice questions
114
-
115
-
116
- - ES: [Head_qa](https://huggingface.co/datasets/head_qa)
117
- - FR: [Frenchmedmcqa](https://github.com/qanastek/FrenchMedMCQA)
118
- - HI: [MMLU_HI](https://huggingface.co/datasets/FreedomIntelligence/MMLU_Arabic)
119
- - Clinical knowledge, Medical genetics, Anatomy, Professional medicine, College biology, College medicine
120
- - AR: [MMLU_Ara](https://huggingface.co/datasets/FreedomIntelligence/MMLU_Hindi)
121
- - Clinical knowledge, Medical genetics, Anatomy, Professional medicine, College biology, College medicine
122
-
123
-
124
- </details>
125
-
126
-
127
- ## Results reproduction
128
- <details><summary>Click to expand</summary>
129
-
130
- **Waiting for Update**
131
-
132
-
133
-
134
- </details>
135
-
136
-
137
-
138
-
139
- ## Citation
140
- Please use the following citation if you intend to use our dataset for training or evaluation:
141
-
142
- ```
143
- @misc{wang2024apollo,
144
- title={Apollo: Lightweight Multilingual Medical LLMs towards Democratizing Medical AI to 6B People},
145
- author={Xidong Wang and Nuo Chen and Junyin Chen and Yan Hu and Yidong Wang and Xiangbo Wu and Anningzhe Gao and Xiang Wan and Haizhou Li and Benyou Wang},
146
- year={2024},
147
- eprint={2403.03640},
148
- archivePrefix={arXiv},
149
- primaryClass={cs.CL}
150
- }
151
- ```
 
1
  ---
2
+ base_model: FreedomIntelligence/Apollo-7B
3
+ language:
4
+ - en
5
+ library_name: transformers
6
  license: apache-2.0
7
+ quantized_by: mradermacher
8
  ---
9
+ ## About
10
 
11
+ <!-- ### quantize_version: 1 -->
12
+ <!-- ### output_tensor_quantised: 1 -->
13
+ <!-- ### convert_type: -->
14
+ <!-- ### vocab_type: -->
15
+ static quants of https://huggingface.co/FreedomIntelligence/Apollo-7B
16
 
17
+ <!-- provided-files -->
18
+ weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion.
19
+ ## Usage
20
 
21
+ If you are unsure how to use GGUF files, refer to one of [TheBloke's
22
+ READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
23
+ more details, including on how to concatenate multi-part files.
 
24
 
25
+ ## Provided Quants
26
 
27
+ (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
28
 
29
+ | Link | Type | Size/GB | Notes |
30
+ |:-----|:-----|--------:|:------|
31
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.Q2_K.gguf) | Q2_K | 3.6 | |
32
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.IQ3_XS.gguf) | IQ3_XS | 3.9 | |
33
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.IQ3_S.gguf) | IQ3_S | 4.1 | beats Q3_K* |
34
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.Q3_K_S.gguf) | Q3_K_S | 4.1 | |
35
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.IQ3_M.gguf) | IQ3_M | 4.2 | |
36
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.Q3_K_M.gguf) | Q3_K_M | 4.5 | lower quality |
37
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.Q3_K_L.gguf) | Q3_K_L | 4.8 | |
38
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.IQ4_XS.gguf) | IQ4_XS | 4.9 | |
39
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.Q4_K_S.gguf) | Q4_K_S | 5.1 | fast, recommended |
40
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.Q4_K_M.gguf) | Q4_K_M | 5.4 | fast, recommended |
41
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.Q5_K_S.gguf) | Q5_K_S | 6.1 | |
42
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.Q5_K_M.gguf) | Q5_K_M | 6.2 | |
43
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.Q6_K.gguf) | Q6_K | 7.1 | very good quality |
44
+ | [GGUF](https://huggingface.co/mradermacher/Apollo-7B-GGUF/resolve/main/Apollo-7B.Q8_0.gguf) | Q8_0 | 9.2 | fast, best quality |
45
 
46
 
47
+ Here is a handy graph by ikawrakow comparing some lower-quality quant
48
+ types (lower is better):
 
 
 
 
 
 
 
 
49
 
50
+ ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)
51
 
52
+ And here are Artefact2's thoughts on the matter:
53
+ https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
54
 
55
+ ## FAQ / Model Request
 
 
 
56
 
57
+ See https://huggingface.co/mradermacher/model_requests for some answers to
58
+ questions you might have and/or if you want some other model quantized.
59
 
60
+ ## Thanks
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
61
 
62
+ I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
63
+ me use its servers and providing upgrades to my workstation to enable
64
+ this work in my free time.
65
 
66
+ <!-- end -->