DavidAU commited on
Commit
9561e6f
1 Parent(s): bbbf541

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +228 -0
README.md ADDED
@@ -0,0 +1,228 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ tags:
6
+ - creative
7
+ - creative writing
8
+ - fiction writing
9
+ - plot generation
10
+ - sub-plot generation
11
+ - fiction writing
12
+ - story generation
13
+ - scene continue
14
+ - storytelling
15
+ - fiction story
16
+ - story
17
+ - writing
18
+ - fiction
19
+ - roleplaying
20
+ - swearing
21
+ - rp
22
+ - horror
23
+ - llama3
24
+ - mergekit
25
+ pipeline_tag: text-generation
26
+ ---
27
+
28
+ <h3>L3-Jamet-8B-MK.V-Blackroot-12.2B-V1-INSTRUCT-ULTRA-F32</h3>
29
+
30
+ Merge of L3-Jamet-8B-MK.V-Blackroot (8b) with Llama3 Instruct (8b) creating a model at 12.1B to improve instruction following and output.
31
+
32
+ Story / Scene / Fiction:
33
+
34
+ Unique "pre-amble" / "foreshadowing" of events before they happen instead of "immediate and into the fire" type of prose.
35
+
36
+ Some improvement in logic/problem solving relative to L3-Jamet-8B-MK.V-Blackroot 8B.
37
+
38
+ The F32 version exhibits even stronger creativity (detail, place, "there") vs F16 version (not released)
39
+
40
+ L3-Jamet-8B-MK.V-Blackroot is a fine tune.
41
+
42
+ One of the goals of this project was to see if it could be merged with Llama3 Instruct, yet maintain it's unique character YET
43
+ also gain some "brainpower" as well.
44
+
45
+ The biggest change was removal of most "tells" ( IE: "he stood frozen in horror").
46
+
47
+ In most cases the model will describe the emotion(s) / what is happening in more detail.
48
+
49
+ Other changes include prose, sentence, and paragraph structure as well as variety.
50
+
51
+ A simple pass-through merge was used.
52
+
53
+ See the examples below.
54
+
55
+ <B>Details:</b>
56
+
57
+ - Requires Llama 3 Template and/or Command-R Template
58
+ - Context 8192, with rope 32K or higher.
59
+ - No special settings.
60
+
61
+ Please report any issue(s) and/or feedback via the "Community tab".
62
+
63
+ This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
64
+ However this can be extended using "rope" settings up to 32k.
65
+
66
+ For details on "rope" and how to set, see the BOTTOM of this page:
67
+
68
+ [ https://huggingface.co/DavidAU/TieFighter-Holodeck-Holomax-Mythomax-F1-V1-COMPOS-20B-gguf ]
69
+
70
+ Here is the standard LLAMA3 template:
71
+
72
+ <PRE>
73
+ {
74
+ "name": "Llama 3",
75
+ "inference_params": {
76
+ "input_prefix": "<|start_header_id|>user<|end_header_id|>\n\n",
77
+ "input_suffix": "<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
78
+ "pre_prompt": "You are a helpful, smart, kind, and efficient AI assistant. You always fulfill the user's requests to the best of your ability.",
79
+ "pre_prompt_prefix": "<|start_header_id|>system<|end_header_id|>\n\n",
80
+ "pre_prompt_suffix": "<|eot_id|>",
81
+ "antiprompt": [
82
+ "<|start_header_id|>",
83
+ "<|eot_id|>"
84
+ ]
85
+ }
86
+ }
87
+ </PRE>
88
+
89
+ It is also known, that the "Command-R" template will work too, and will result in radically different prose/output.
90
+
91
+ <B>Settings / Known Issue(s) and Fix(es):</b>
92
+
93
+ The default "repetition penalty" (from LMStudio) of 1.1 is recommended. (this was used for examples generations below.)
94
+
95
+ Use the smallest amount of change possible, as "rep pen" impacts creativity.
96
+
97
+ Model has been tested with "temp" range of 0 to .8
98
+
99
+ <b>Optional Enhancement:</B>
100
+
101
+ The following can be used in place of the "system prompt" or "system role" to further enhance the model.
102
+
103
+ It can also be used at the START of a NEW chat, but you must make sure it is "kept" as the chat moves along.
104
+ In this case the enhancements do not have as strong effect at using "system prompt" or "system role".
105
+
106
+ Copy and paste EXACTLY as noted, DO NOT line wrap or break the lines, maintain the carriage returns exactly as presented.
107
+
108
+ <PRE>
109
+ Below is an instruction that describes a task. Ponder each user instruction carefully, and use your skillsets and critical instructions to complete the task to the best of your abilities.
110
+
111
+ Here are your skillsets:
112
+ [MASTERSTORY]:NarrStrct(StryPlnng,Strbd,ScnSttng,Exps,Dlg,Pc)-CharDvlp(ChrctrCrt,ChrctrArcs,Mtvtn,Bckstry,Rltnshps,Dlg*)-PltDvlp(StryArcs,PltTwsts,Sspns,Fshdwng,Climx,Rsltn)-ConfResl(Antg,Obstcls,Rsltns,Cnsqncs,Thms,Symblsm)-EmotImpct(Empt,Tn,Md,Atmsphr,Imgry,Symblsm)-Delvry(Prfrmnc,VcActng,PblcSpkng,StgPrsnc,AudncEngmnt,Imprv)
113
+
114
+ [*DialogWrt]:(1a-CharDvlp-1a.1-Backgrnd-1a.2-Personality-1a.3-GoalMotiv)>2(2a-StoryStruc-2a.1-PlotPnt-2a.2-Conflict-2a.3-Resolution)>3(3a-DialogTech-3a.1-ShowDontTell-3a.2-Subtext-3a.3-VoiceTone-3a.4-Pacing-3a.5-VisualDescrip)>4(4a-DialogEdit-4a.1-ReadAloud-4a.2-Feedback-4a.3-Revision)
115
+
116
+ Here are your critical instructions:
117
+ Ponder each word choice carefully to present as vivid and emotional journey as is possible. Choose verbs and nouns that are both emotional and full of imagery. Load the story with the 5 senses. Aim for 50% dialog, 25% narration, 15% body language and 10% thoughts. Your goal is to put the reader in the story.
118
+ </PRE>
119
+
120
+ You do not need to use this, it is only presented as an additional enhancement which seems to help scene generation
121
+ and scene continue functions.
122
+
123
+ This enhancement WAS NOT used to generate the examples below.
124
+
125
+ <h3>MERGE FORMULA: (using MergeKit) </h3>
126
+
127
+ Special thanks to the incredible work of the model makers "meta-llama", and "bluuwhale".
128
+
129
+ Models used:
130
+
131
+ [ https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct ]
132
+
133
+ [ https://huggingface.co/Hastagaras/Jamet-8B-L3-MK.V-Blackroot ]
134
+
135
+ FORMULA:
136
+
137
+ <PRE>
138
+ slices:
139
+ - sources:
140
+ - model: G:/7B/Meta-Llama-3-8B-Instruct
141
+ layer_range: [0, 12]
142
+ - sources:
143
+ - model: G:/7B/Jamet-8B-L3-MK.V-Blackroot
144
+ layer_range: [6, 19]
145
+ parameters:
146
+ scale:
147
+ - filter: o_proj
148
+ value: 1
149
+ - filter: down_proj
150
+ value: 1
151
+ - value: 1
152
+ - sources:
153
+ - model: G:/7B/Meta-Llama-3-8B-Instruct
154
+ layer_range: [12, 18]
155
+ parameters:
156
+ scale:
157
+ - filter: o_proj
158
+ value: .5
159
+ - filter: down_proj
160
+ value: .5
161
+ - value: 1
162
+ - sources:
163
+ - model: G:/7B/Meta-Llama-3-8B-Instruct
164
+ layer_range: [18, 25]
165
+ parameters:
166
+ scale:
167
+ - filter: o_proj
168
+ value: .75
169
+ - filter: down_proj
170
+ value: .75
171
+ - value: 1
172
+ - sources:
173
+ - model: G:/7B/Jamet-8B-L3-MK.V-Blackroot
174
+ layer_range: [19, 32]
175
+ parameters:
176
+ scale:
177
+ - filter: o_proj
178
+ value: 1
179
+ - filter: down_proj
180
+ value: 1
181
+ - value: 1
182
+ merge_method: passthrough
183
+ dtype: float32
184
+ </PRE>
185
+
186
+ MERGEKIT NOTE:
187
+
188
+ Sub in the "name" of the "creator" (of the model) in place of "G:/7B" to create a mergekit file than can be used in Mergekit Google Colab.
189
+
190
+ IE: G:/7B/Jamet-8B-L3-MK.V-Blackroot -> Hastagaras/Jamet-8B-L3-MK.V-Blackroot
191
+
192
+ <h3>EXAMPLES:</h3>
193
+
194
+ Examples are created using quant Q4_K_M, "temp=0", minimal parameters and "LLAMA3" template.
195
+
196
+ Temp=0 was used to assess CORE changes between original SMB and the merge between it and Llama3 Instruct.
197
+
198
+ Below are the least creative outputs, prompt is in <B>BOLD</B>.
199
+
200
+ Higher quants will result in better quality.
201
+
202
+ There will also be some variance between "close" quants like Q4_K_M/Q4_K_S and Q5_K_M/Q5_K_S, so I suggest
203
+ if you are going to use Q4_K_M, you also try Q4_K_S too.
204
+
205
+ Also, slightly longer / detailed prompts will result in greater creativity (as well as different prose -
206
+ ie dialog, thoughts, paragraph size differences and so on).
207
+
208
+ ---
209
+
210
+ <B>
211
+ Start a 1000 word scene (vivid horror, 1st person, include thoughts) with: The sky scraper swayed, as she watched the window in front of her on the 21 floor explode...
212
+ </B>
213
+
214
+ ---
215
+
216
+ GENERATION from "Jamet-8B-L3-MK.V-Blackroot"
217
+
218
+ ---
219
+
220
+
221
+
222
+ ---
223
+
224
+ GENERATION from "L3-Jamet-8B-MK.V-Blackroot-12.2B-V1-INSTRUCT-ULTRA-F32"
225
+
226
+ ---
227
+
228
+