Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,228 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
tags:
|
6 |
+
- creative
|
7 |
+
- creative writing
|
8 |
+
- fiction writing
|
9 |
+
- plot generation
|
10 |
+
- sub-plot generation
|
11 |
+
- fiction writing
|
12 |
+
- story generation
|
13 |
+
- scene continue
|
14 |
+
- storytelling
|
15 |
+
- fiction story
|
16 |
+
- story
|
17 |
+
- writing
|
18 |
+
- fiction
|
19 |
+
- roleplaying
|
20 |
+
- swearing
|
21 |
+
- rp
|
22 |
+
- horror
|
23 |
+
- llama3
|
24 |
+
- mergekit
|
25 |
+
pipeline_tag: text-generation
|
26 |
+
---
|
27 |
+
|
28 |
+
<h3>L3-Jamet-8B-MK.V-Blackroot-12.2B-V1-INSTRUCT-ULTRA-F32</h3>
|
29 |
+
|
30 |
+
Merge of L3-Jamet-8B-MK.V-Blackroot (8b) with Llama3 Instruct (8b) creating a model at 12.1B to improve instruction following and output.
|
31 |
+
|
32 |
+
Story / Scene / Fiction:
|
33 |
+
|
34 |
+
Unique "pre-amble" / "foreshadowing" of events before they happen instead of "immediate and into the fire" type of prose.
|
35 |
+
|
36 |
+
Some improvement in logic/problem solving relative to L3-Jamet-8B-MK.V-Blackroot 8B.
|
37 |
+
|
38 |
+
The F32 version exhibits even stronger creativity (detail, place, "there") vs F16 version (not released)
|
39 |
+
|
40 |
+
L3-Jamet-8B-MK.V-Blackroot is a fine tune.
|
41 |
+
|
42 |
+
One of the goals of this project was to see if it could be merged with Llama3 Instruct, yet maintain it's unique character YET
|
43 |
+
also gain some "brainpower" as well.
|
44 |
+
|
45 |
+
The biggest change was removal of most "tells" ( IE: "he stood frozen in horror").
|
46 |
+
|
47 |
+
In most cases the model will describe the emotion(s) / what is happening in more detail.
|
48 |
+
|
49 |
+
Other changes include prose, sentence, and paragraph structure as well as variety.
|
50 |
+
|
51 |
+
A simple pass-through merge was used.
|
52 |
+
|
53 |
+
See the examples below.
|
54 |
+
|
55 |
+
<B>Details:</b>
|
56 |
+
|
57 |
+
- Requires Llama 3 Template and/or Command-R Template
|
58 |
+
- Context 8192, with rope 32K or higher.
|
59 |
+
- No special settings.
|
60 |
+
|
61 |
+
Please report any issue(s) and/or feedback via the "Community tab".
|
62 |
+
|
63 |
+
This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
|
64 |
+
However this can be extended using "rope" settings up to 32k.
|
65 |
+
|
66 |
+
For details on "rope" and how to set, see the BOTTOM of this page:
|
67 |
+
|
68 |
+
[ https://huggingface.co/DavidAU/TieFighter-Holodeck-Holomax-Mythomax-F1-V1-COMPOS-20B-gguf ]
|
69 |
+
|
70 |
+
Here is the standard LLAMA3 template:
|
71 |
+
|
72 |
+
<PRE>
|
73 |
+
{
|
74 |
+
"name": "Llama 3",
|
75 |
+
"inference_params": {
|
76 |
+
"input_prefix": "<|start_header_id|>user<|end_header_id|>\n\n",
|
77 |
+
"input_suffix": "<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
|
78 |
+
"pre_prompt": "You are a helpful, smart, kind, and efficient AI assistant. You always fulfill the user's requests to the best of your ability.",
|
79 |
+
"pre_prompt_prefix": "<|start_header_id|>system<|end_header_id|>\n\n",
|
80 |
+
"pre_prompt_suffix": "<|eot_id|>",
|
81 |
+
"antiprompt": [
|
82 |
+
"<|start_header_id|>",
|
83 |
+
"<|eot_id|>"
|
84 |
+
]
|
85 |
+
}
|
86 |
+
}
|
87 |
+
</PRE>
|
88 |
+
|
89 |
+
It is also known, that the "Command-R" template will work too, and will result in radically different prose/output.
|
90 |
+
|
91 |
+
<B>Settings / Known Issue(s) and Fix(es):</b>
|
92 |
+
|
93 |
+
The default "repetition penalty" (from LMStudio) of 1.1 is recommended. (this was used for examples generations below.)
|
94 |
+
|
95 |
+
Use the smallest amount of change possible, as "rep pen" impacts creativity.
|
96 |
+
|
97 |
+
Model has been tested with "temp" range of 0 to .8
|
98 |
+
|
99 |
+
<b>Optional Enhancement:</B>
|
100 |
+
|
101 |
+
The following can be used in place of the "system prompt" or "system role" to further enhance the model.
|
102 |
+
|
103 |
+
It can also be used at the START of a NEW chat, but you must make sure it is "kept" as the chat moves along.
|
104 |
+
In this case the enhancements do not have as strong effect at using "system prompt" or "system role".
|
105 |
+
|
106 |
+
Copy and paste EXACTLY as noted, DO NOT line wrap or break the lines, maintain the carriage returns exactly as presented.
|
107 |
+
|
108 |
+
<PRE>
|
109 |
+
Below is an instruction that describes a task. Ponder each user instruction carefully, and use your skillsets and critical instructions to complete the task to the best of your abilities.
|
110 |
+
|
111 |
+
Here are your skillsets:
|
112 |
+
[MASTERSTORY]:NarrStrct(StryPlnng,Strbd,ScnSttng,Exps,Dlg,Pc)-CharDvlp(ChrctrCrt,ChrctrArcs,Mtvtn,Bckstry,Rltnshps,Dlg*)-PltDvlp(StryArcs,PltTwsts,Sspns,Fshdwng,Climx,Rsltn)-ConfResl(Antg,Obstcls,Rsltns,Cnsqncs,Thms,Symblsm)-EmotImpct(Empt,Tn,Md,Atmsphr,Imgry,Symblsm)-Delvry(Prfrmnc,VcActng,PblcSpkng,StgPrsnc,AudncEngmnt,Imprv)
|
113 |
+
|
114 |
+
[*DialogWrt]:(1a-CharDvlp-1a.1-Backgrnd-1a.2-Personality-1a.3-GoalMotiv)>2(2a-StoryStruc-2a.1-PlotPnt-2a.2-Conflict-2a.3-Resolution)>3(3a-DialogTech-3a.1-ShowDontTell-3a.2-Subtext-3a.3-VoiceTone-3a.4-Pacing-3a.5-VisualDescrip)>4(4a-DialogEdit-4a.1-ReadAloud-4a.2-Feedback-4a.3-Revision)
|
115 |
+
|
116 |
+
Here are your critical instructions:
|
117 |
+
Ponder each word choice carefully to present as vivid and emotional journey as is possible. Choose verbs and nouns that are both emotional and full of imagery. Load the story with the 5 senses. Aim for 50% dialog, 25% narration, 15% body language and 10% thoughts. Your goal is to put the reader in the story.
|
118 |
+
</PRE>
|
119 |
+
|
120 |
+
You do not need to use this, it is only presented as an additional enhancement which seems to help scene generation
|
121 |
+
and scene continue functions.
|
122 |
+
|
123 |
+
This enhancement WAS NOT used to generate the examples below.
|
124 |
+
|
125 |
+
<h3>MERGE FORMULA: (using MergeKit) </h3>
|
126 |
+
|
127 |
+
Special thanks to the incredible work of the model makers "meta-llama", and "bluuwhale".
|
128 |
+
|
129 |
+
Models used:
|
130 |
+
|
131 |
+
[ https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct ]
|
132 |
+
|
133 |
+
[ https://huggingface.co/Hastagaras/Jamet-8B-L3-MK.V-Blackroot ]
|
134 |
+
|
135 |
+
FORMULA:
|
136 |
+
|
137 |
+
<PRE>
|
138 |
+
slices:
|
139 |
+
- sources:
|
140 |
+
- model: G:/7B/Meta-Llama-3-8B-Instruct
|
141 |
+
layer_range: [0, 12]
|
142 |
+
- sources:
|
143 |
+
- model: G:/7B/Jamet-8B-L3-MK.V-Blackroot
|
144 |
+
layer_range: [6, 19]
|
145 |
+
parameters:
|
146 |
+
scale:
|
147 |
+
- filter: o_proj
|
148 |
+
value: 1
|
149 |
+
- filter: down_proj
|
150 |
+
value: 1
|
151 |
+
- value: 1
|
152 |
+
- sources:
|
153 |
+
- model: G:/7B/Meta-Llama-3-8B-Instruct
|
154 |
+
layer_range: [12, 18]
|
155 |
+
parameters:
|
156 |
+
scale:
|
157 |
+
- filter: o_proj
|
158 |
+
value: .5
|
159 |
+
- filter: down_proj
|
160 |
+
value: .5
|
161 |
+
- value: 1
|
162 |
+
- sources:
|
163 |
+
- model: G:/7B/Meta-Llama-3-8B-Instruct
|
164 |
+
layer_range: [18, 25]
|
165 |
+
parameters:
|
166 |
+
scale:
|
167 |
+
- filter: o_proj
|
168 |
+
value: .75
|
169 |
+
- filter: down_proj
|
170 |
+
value: .75
|
171 |
+
- value: 1
|
172 |
+
- sources:
|
173 |
+
- model: G:/7B/Jamet-8B-L3-MK.V-Blackroot
|
174 |
+
layer_range: [19, 32]
|
175 |
+
parameters:
|
176 |
+
scale:
|
177 |
+
- filter: o_proj
|
178 |
+
value: 1
|
179 |
+
- filter: down_proj
|
180 |
+
value: 1
|
181 |
+
- value: 1
|
182 |
+
merge_method: passthrough
|
183 |
+
dtype: float32
|
184 |
+
</PRE>
|
185 |
+
|
186 |
+
MERGEKIT NOTE:
|
187 |
+
|
188 |
+
Sub in the "name" of the "creator" (of the model) in place of "G:/7B" to create a mergekit file than can be used in Mergekit Google Colab.
|
189 |
+
|
190 |
+
IE: G:/7B/Jamet-8B-L3-MK.V-Blackroot -> Hastagaras/Jamet-8B-L3-MK.V-Blackroot
|
191 |
+
|
192 |
+
<h3>EXAMPLES:</h3>
|
193 |
+
|
194 |
+
Examples are created using quant Q4_K_M, "temp=0", minimal parameters and "LLAMA3" template.
|
195 |
+
|
196 |
+
Temp=0 was used to assess CORE changes between original SMB and the merge between it and Llama3 Instruct.
|
197 |
+
|
198 |
+
Below are the least creative outputs, prompt is in <B>BOLD</B>.
|
199 |
+
|
200 |
+
Higher quants will result in better quality.
|
201 |
+
|
202 |
+
There will also be some variance between "close" quants like Q4_K_M/Q4_K_S and Q5_K_M/Q5_K_S, so I suggest
|
203 |
+
if you are going to use Q4_K_M, you also try Q4_K_S too.
|
204 |
+
|
205 |
+
Also, slightly longer / detailed prompts will result in greater creativity (as well as different prose -
|
206 |
+
ie dialog, thoughts, paragraph size differences and so on).
|
207 |
+
|
208 |
+
---
|
209 |
+
|
210 |
+
<B>
|
211 |
+
Start a 1000 word scene (vivid horror, 1st person, include thoughts) with: The sky scraper swayed, as she watched the window in front of her on the 21 floor explode...
|
212 |
+
</B>
|
213 |
+
|
214 |
+
---
|
215 |
+
|
216 |
+
GENERATION from "Jamet-8B-L3-MK.V-Blackroot"
|
217 |
+
|
218 |
+
---
|
219 |
+
|
220 |
+
|
221 |
+
|
222 |
+
---
|
223 |
+
|
224 |
+
GENERATION from "L3-Jamet-8B-MK.V-Blackroot-12.2B-V1-INSTRUCT-ULTRA-F32"
|
225 |
+
|
226 |
+
---
|
227 |
+
|
228 |
+
|