aashish1904 commited on
Commit
65aec44
1 Parent(s): 1484d8c

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +133 -0
README.md ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ base_model:
5
+ - v000000/L3.1-8B-RP-Test-003-Task_Arithmetic
6
+ - v000000/L3.1-Niitorm-8B-t0.0001
7
+ - Sao10K/L3.1-8B-Niitama-v1.1
8
+ - arcee-ai/Llama-3.1-SuperNova-Lite
9
+ - akjindal53244/Llama-3.1-Storm-8B
10
+ - arcee-ai/Llama-Spark
11
+ - v000000/L3.1-8B-RP-Test-002-Task_Arithmetic
12
+ - grimjim/Llama-3-Instruct-abliteration-LoRA-8B
13
+ library_name: transformers
14
+ tags:
15
+ - mergekit
16
+ - merge
17
+ - llama
18
+
19
+ ---
20
+
21
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
22
+
23
+
24
+ # QuantFactory/L3.1-Storniitova-8B-GGUF
25
+ This is quantized version of [v000000/L3.1-Storniitova-8B](https://huggingface.co/v000000/L3.1-Storniitova-8B) created using llama.cpp
26
+
27
+ # Original Model Card
28
+
29
+
30
+ # Llama-3.1-Storniitova-8B
31
+
32
+ Storniitova-8B is a RP/Instruct model built on the foundation of Llama-3.1-SuperNova-Lite, which is distilled from the 405B parameter variant of Llama-3.1
33
+
34
+ By only changing the vector tasks, I attempt to retain the full 405B distillation while learning roleplaying capabilties.
35
+
36
+ # (GGUF) mradermacher quants:
37
+
38
+ * [GGUFs](https://huggingface.co/mradermacher/L3.1-Storniitova-8B-GGUF)
39
+ * [GGUFs imatrix](https://huggingface.co/mradermacher/L3.1-Storniitova-8B-i1-GGUF)
40
+
41
+ -----------------------------------------------------------------------------------------------------------
42
+
43
+ # merge
44
+
45
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit) and other proprietary tools.
46
+
47
+ ## Merge Details
48
+ ### Merge Method
49
+
50
+ This model was merged using the <b>SLERP, Task_Arithmetic and NEARSWAP</b> merge method.
51
+
52
+ ### Models Merged
53
+
54
+ The following models were included in the merge:
55
+ * [v000000/L3.1-Niitorm-8B-t0.0001](https://huggingface.co/v000000/L3.1-Niitorm-8B-t0.0001)
56
+ * [akjindal53244/Llama-3.1-Storm-8B](https://huggingface.co/akjindal53244/Llama-3.1-Storm-8B)
57
+ * [arcee-ai/Llama-Spark](https://huggingface.co/arcee-ai/Llama-Spark)
58
+ * [arcee-ai/Llama-3.1-SuperNova-Lite](https://huggingface.co/arcee-ai/Llama-3.1-SuperNova-Lite)
59
+ * [v000000/L3.1-8B-RP-Test-003-Task_Arithmetic](https://huggingface.co/v000000/L3.1-8B-RP-Test-003-Task_Arithmetic)
60
+ * [Sao10K/L3.1-8B-Niitama-v1.1](https://huggingface.co/Sao10K/L3.1-8B-Niitama-v1.1) + [grimjim/Llama-3-Instruct-abliteration-LoRA-8B](https://huggingface.co/grimjim/Llama-3-Instruct-abliteration-LoRA-8B)
61
+ * [v000000/L3.1-8B-RP-Test-002-Task_Arithmetic](https://huggingface.co/v000000/L3.1-8B-RP-Test-002-Task_Arithmetic) + [grimjim/Llama-3-Instruct-abliteration-LoRA-8B](https://huggingface.co/grimjim/Llama-3-Instruct-abliteration-LoRA-8B)
62
+
63
+ ### Recipe
64
+
65
+ The following YAML configuration was used to produce this model:
66
+
67
+ ```yaml
68
+ #Step1 - Add smarts to Niitama with alchemonaut's algorithm.
69
+
70
+ slices:
71
+ - sources:
72
+ - model: Sao10K/L3.1-8B-Niitama-v1.1+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
73
+ layer_range: [0, 32]
74
+ - model: akjindal53244/Llama-3.1-Storm-8B
75
+ layer_range: [0, 32]
76
+ merge_method: nearswap
77
+ base_model: Sao10K/L3.1-8B-Niitama-v1.1+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
78
+ parameters:
79
+ t:
80
+ - value: 0.0001
81
+ dtype: bfloat16
82
+ out_type: float16
83
+
84
+ #Step 2 - Learn vectors onto Supernova 0.4(Niitorm)
85
+
86
+ models:
87
+ - model: arcee-ai/Llama-3.1-SuperNova-Lite
88
+ parameters:
89
+ weight: 1.0
90
+ - model: v000000/L3.1-Niitorm-8B-t0.0001
91
+ parameters:
92
+ weight: 0.4
93
+ merge_method: task_arithmetic
94
+ base_model: arcee-ai/Llama-3.1-SuperNova-Lite
95
+ parameters:
96
+ normalize: false
97
+ dtype: float16
98
+
99
+ #Step 3 - Fully learn vectors onto Supernova 1.25(Niitorm)
100
+
101
+ models:
102
+ - model: arcee-ai/Llama-3.1-SuperNova-Lite
103
+ parameters:
104
+ weight: 0.0
105
+ - model: v000000/L3.1-Niitorm-8B-t0.0001
106
+ parameters:
107
+ weight: 1.25
108
+ merge_method: task_arithmetic
109
+ base_model: arcee-ai/Llama-3.1-SuperNova-Lite
110
+ parameters:
111
+ normalize: false
112
+ dtype: float16
113
+
114
+ #Step 4 - Merge checkpoints and keep output/input Supernova heavy
115
+ #Merge with a triangular slerp from sophosympatheia.
116
+
117
+ models:
118
+ - model: v000000/L3.1-8B-RP-Test-003-Task_Arithmetic
119
+ merge_method: slerp
120
+ base_model: v000000/L3.1-8B-RP-Test-002-Task_Arithmetic+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
121
+ # This model needed some abliteration^
122
+ parameters:
123
+ t:
124
+ - value: [0, 0, 0.3, 0.4, 0.5, 0.6, 0.5, 0.4, 0.3, 0, 0]
125
+ dtype: float16
126
+
127
+ ```
128
+
129
+ *SLERP distribution* used to smoothly blend the mostly Supernova base with the roleplay vectors:
130
+
131
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64f74b6e6389380c77562762/GP2LMRvMkhVJwNDSEC4oU.png)
132
+
133
+