aashish1904 commited on
Commit
a2cab61
·
verified ·
1 Parent(s): bfd3823

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +188 -0
README.md ADDED
@@ -0,0 +1,188 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ library_name: transformers
5
+ tags:
6
+ - mergekit
7
+ - merge
8
+ base_model:
9
+ - arcee-ai/SuperNova-Medius
10
+ - huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2
11
+ - allura-org/TQ2.5-14B-Aletheia-v1
12
+ - EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2
13
+ - v000000/Qwen2.5-Lumen-14B
14
+ model-index:
15
+ - name: Q2.5-Veltha-14B-0.5
16
+ results:
17
+ - task:
18
+ type: text-generation
19
+ name: Text Generation
20
+ dataset:
21
+ name: IFEval (0-Shot)
22
+ type: HuggingFaceH4/ifeval
23
+ args:
24
+ num_few_shot: 0
25
+ metrics:
26
+ - type: inst_level_strict_acc and prompt_level_strict_acc
27
+ value: 77.96
28
+ name: strict accuracy
29
+ source:
30
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/Q2.5-Veltha-14B-0.5
31
+ name: Open LLM Leaderboard
32
+ - task:
33
+ type: text-generation
34
+ name: Text Generation
35
+ dataset:
36
+ name: BBH (3-Shot)
37
+ type: BBH
38
+ args:
39
+ num_few_shot: 3
40
+ metrics:
41
+ - type: acc_norm
42
+ value: 50.32
43
+ name: normalized accuracy
44
+ source:
45
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/Q2.5-Veltha-14B-0.5
46
+ name: Open LLM Leaderboard
47
+ - task:
48
+ type: text-generation
49
+ name: Text Generation
50
+ dataset:
51
+ name: MATH Lvl 5 (4-Shot)
52
+ type: hendrycks/competition_math
53
+ args:
54
+ num_few_shot: 4
55
+ metrics:
56
+ - type: exact_match
57
+ value: 33.84
58
+ name: exact match
59
+ source:
60
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/Q2.5-Veltha-14B-0.5
61
+ name: Open LLM Leaderboard
62
+ - task:
63
+ type: text-generation
64
+ name: Text Generation
65
+ dataset:
66
+ name: GPQA (0-shot)
67
+ type: Idavidrein/gpqa
68
+ args:
69
+ num_few_shot: 0
70
+ metrics:
71
+ - type: acc_norm
72
+ value: 15.77
73
+ name: acc_norm
74
+ source:
75
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/Q2.5-Veltha-14B-0.5
76
+ name: Open LLM Leaderboard
77
+ - task:
78
+ type: text-generation
79
+ name: Text Generation
80
+ dataset:
81
+ name: MuSR (0-shot)
82
+ type: TAUR-Lab/MuSR
83
+ args:
84
+ num_few_shot: 0
85
+ metrics:
86
+ - type: acc_norm
87
+ value: 14.17
88
+ name: acc_norm
89
+ source:
90
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/Q2.5-Veltha-14B-0.5
91
+ name: Open LLM Leaderboard
92
+ - task:
93
+ type: text-generation
94
+ name: Text Generation
95
+ dataset:
96
+ name: MMLU-PRO (5-shot)
97
+ type: TIGER-Lab/MMLU-Pro
98
+ config: main
99
+ split: test
100
+ args:
101
+ num_few_shot: 5
102
+ metrics:
103
+ - type: acc
104
+ value: 47.72
105
+ name: accuracy
106
+ source:
107
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/Q2.5-Veltha-14B-0.5
108
+ name: Open LLM Leaderboard
109
+
110
+ ---
111
+
112
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
113
+
114
+
115
+ # QuantFactory/Q2.5-Veltha-14B-0.5-GGUF
116
+ This is quantized version of [djuna/Q2.5-Veltha-14B-0.5](https://huggingface.co/djuna/Q2.5-Veltha-14B-0.5) created using llama.cpp
117
+
118
+ # Original Model Card
119
+
120
+ # merge
121
+
122
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
123
+
124
+ ## Merge Details
125
+ ### Merge Method
126
+
127
+ This model was merged using the della_linear merge method using [arcee-ai/SuperNova-Medius](https://huggingface.co/arcee-ai/SuperNova-Medius) as a base.
128
+
129
+ ### Models Merged
130
+
131
+ The following models were included in the merge:
132
+ * [huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2](https://huggingface.co/huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2)
133
+ * [allura-org/TQ2.5-14B-Aletheia-v1](https://huggingface.co/allura-org/TQ2.5-14B-Aletheia-v1)
134
+ * [EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2)
135
+ * [v000000/Qwen2.5-Lumen-14B](https://huggingface.co/v000000/Qwen2.5-Lumen-14B)
136
+
137
+ ### Configuration
138
+
139
+ The following YAML configuration was used to produce this model:
140
+
141
+ ```yaml
142
+ merge_method: della_linear
143
+ dtype: float32
144
+ out_dtype: bfloat16
145
+ parameters:
146
+ epsilon: 0.04
147
+ lambda: 1.05
148
+ normalize: true
149
+ base_model: arcee-ai/SuperNova-Medius
150
+ tokenizer_source: arcee-ai/SuperNova-Medius
151
+ models:
152
+ - model: arcee-ai/SuperNova-Medius
153
+ parameters:
154
+ weight: 10
155
+ density: 1
156
+ - model: EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2
157
+ parameters:
158
+ weight: 7
159
+ density: 0.5
160
+ - model: v000000/Qwen2.5-Lumen-14B
161
+ parameters:
162
+ weight: 7
163
+ density: 0.4
164
+ - model: allura-org/TQ2.5-14B-Aletheia-v1
165
+ parameters:
166
+ weight: 8
167
+ density: 0.4
168
+ - model: huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2
169
+ parameters:
170
+ weight: 8
171
+ density: 0.45
172
+
173
+ ```
174
+
175
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
176
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/djuna__Q2.5-Veltha-14B-0.5-details)
177
+
178
+ | Metric |Value|
179
+ |-------------------|----:|
180
+ |Avg. |39.96|
181
+ |IFEval (0-Shot) |77.96|
182
+ |BBH (3-Shot) |50.32|
183
+ |MATH Lvl 5 (4-Shot)|33.84|
184
+ |GPQA (0-shot) |15.77|
185
+ |MuSR (0-shot) |14.17|
186
+ |MMLU-PRO (5-shot) |47.72|
187
+
188
+