Triangle104
commited on
Commit
•
8dd34d8
1
Parent(s):
6abc8dd
Update README.md
Browse files
README.md
CHANGED
@@ -15,6 +15,61 @@ tags:
|
|
15 |
This model was converted to GGUF format from [`Hastagaras/Zabuza-8B-Llama-3.1`](https://huggingface.co/Hastagaras/Zabuza-8B-Llama-3.1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
16 |
Refer to the [original model card](https://huggingface.co/Hastagaras/Zabuza-8B-Llama-3.1) for more details on the model.
|
17 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
18 |
## Use with llama.cpp
|
19 |
Install llama.cpp through brew (works on Mac and Linux)
|
20 |
|
|
|
15 |
This model was converted to GGUF format from [`Hastagaras/Zabuza-8B-Llama-3.1`](https://huggingface.co/Hastagaras/Zabuza-8B-Llama-3.1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
16 |
Refer to the [original model card](https://huggingface.co/Hastagaras/Zabuza-8B-Llama-3.1) for more details on the model.
|
17 |
|
18 |
+
---
|
19 |
+
Model details:
|
20 |
+
-
|
21 |
+
This model is a combination of merge, abliteration technique (using baukit) and finetuning.
|
22 |
+
|
23 |
+
The base model is arcee-ai/Llama-3.1-SuperNova-Lite, which underwent abliteration to reduce model refusals.
|
24 |
+
|
25 |
+
Next, I finetuned the abliterated SuperNova-Lite with 10K diverse examples such as:
|
26 |
+
|
27 |
+
Claude and Gemini Instruction/RP (15k sloppy examples were removed!, but some may have slipped through.)
|
28 |
+
Human-written Stories/RP (Most stories have dialogue)
|
29 |
+
IFEval-like data (To preserve the model's instruction following ability)
|
30 |
+
Harmful data (To remove disclaimers and moralizing responses, but not 100% disappear.)
|
31 |
+
My sarcastic and rude AI assistant data (Just for my personal satisfaction)
|
32 |
+
|
33 |
+
Lastly, I merged the model using TIES, inspired by this MERGE by Joseph717171.
|
34 |
+
|
35 |
+
Chat Template
|
36 |
+
-
|
37 |
+
Llama 3.1 Instruct
|
38 |
+
-
|
39 |
+
<|start_header_id|>{role}<|end_header_id|>
|
40 |
+
|
41 |
+
{message}<|eot_id|><|start_header_id|>{role}<|end_header_id|>
|
42 |
+
|
43 |
+
{message}<|eot_id|>
|
44 |
+
|
45 |
+
System messages for role-playing should be very detailed if you don't want dry responses.
|
46 |
+
|
47 |
+
Configuration
|
48 |
+
-
|
49 |
+
This is a merge of pre-trained language models created using mergekit.
|
50 |
+
|
51 |
+
The following YAML configuration was used to produce this model:
|
52 |
+
|
53 |
+
models:
|
54 |
+
- model: Hastagaras/snovalite-baukit-6-14.FT-L5-7.13-22.27-31
|
55 |
+
parameters:
|
56 |
+
weight: 1
|
57 |
+
density: 1
|
58 |
+
|
59 |
+
- model: Hastagaras/snovalite-baukit-6-14.FT-L5-7.13-22.27-31
|
60 |
+
parameters:
|
61 |
+
weight: 1
|
62 |
+
density: 1
|
63 |
+
|
64 |
+
merge_method: ties
|
65 |
+
base_model: meta-llama/Llama-3.1-8B
|
66 |
+
parameters:
|
67 |
+
density: 1
|
68 |
+
normalize: true
|
69 |
+
int8_mask: true
|
70 |
+
dtype: bfloat16
|
71 |
+
|
72 |
+
---
|
73 |
## Use with llama.cpp
|
74 |
Install llama.cpp through brew (works on Mac and Linux)
|
75 |
|