DavidAU commited on
Commit
7511632
1 Parent(s): 2ed9bdc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +96 -4
README.md CHANGED
@@ -32,6 +32,8 @@ tags:
32
  pipeline_tag: text-generation
33
  ---
34
 
 
 
35
  <B><font color="red">WARNING:</font> MAY contain: Nutiness. Vivid prose. Purple Prose. Funny Violence. Over the Top Storytelling. Min of 2 eyerolls per generation. </B>
36
 
37
  <h2>M-Metaphors-Of-Madness-19.4B-GGUF - AKA "M.O.M"</h2>
@@ -81,7 +83,93 @@ IQ4XS: This might be the quant with the most difference compared to other quants
81
 
82
  Enjoy!
83
 
84
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
85
 
86
  Special thanks to model makers "jondurbin" and "MTSAIR"
87
 
@@ -91,8 +179,6 @@ Models Used:
91
 
92
  [ https://huggingface.co/MTSAIR/multi_verse_model ]
93
 
94
- ---
95
-
96
  <b>Optional Enhancement:</B>
97
 
98
  The following can be used in place of the "system prompt" or "system role" to further enhance the model.
@@ -121,7 +207,13 @@ This enhancement WAS NOT used to generate the examples below.
121
 
122
  ---
123
 
124
- FOUR EXAMPLE GENERATIONS with prompt in <B>BOLD</B>:
 
 
 
 
 
 
125
 
126
  ---
127
 
 
32
  pipeline_tag: text-generation
33
  ---
34
 
35
+ (quants uploading...)
36
+
37
  <B><font color="red">WARNING:</font> MAY contain: Nutiness. Vivid prose. Purple Prose. Funny Violence. Over the Top Storytelling. Min of 2 eyerolls per generation. </B>
38
 
39
  <h2>M-Metaphors-Of-Madness-19.4B-GGUF - AKA "M.O.M"</h2>
 
83
 
84
  Enjoy!
85
 
86
+ <B>Brainstorm 40x</B>
87
+
88
+ The BRAINSTORM process was developed by David_AU.
89
+
90
+ Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
91
+ scientific paper : Progressive LLaMA with Block Expansion </a>.
92
+
93
+ However I went in a completely different direction from what was outlined in this paper.
94
+
95
+ I developed a process where the conclusion layer of a model is duplicated and calibrated, in the case of this model 40 times.
96
+
97
+ This is a delicate process, with umm... a lot of rules.
98
+
99
+ For this model in particular Brainstorm is mapped as blocks, with "intended disruption" to alter
100
+ and extend the power of the root model. Each layer/block interacts with each other block.
101
+
102
+ (there is more going on here too, this is rough summary)
103
+
104
+ The goal here is creative : prose uniqueness first and foremost.
105
+
106
+ Other brainstorm methods address logic/problem solving augmentation.
107
+
108
+ What is "Brainstorm" ?
109
+
110
+ The reasoning center of an LLM is taken apart, reassembled, and expanded.
111
+
112
+ In this case for this model: 40 times
113
+
114
+ Then these centers are individually calibrated. These "centers" also interact with each other.
115
+ This introduces subtle changes into the reasoning process.
116
+ The calibrations further adjust - dial up or down - these "changes" further.
117
+ The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
118
+
119
+ The core aim of this process is to increase the model's detail, concept and connection to the "world",
120
+ general concept connections, prose quality and prose length without affecting instruction following.
121
+
122
+ This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
123
+
124
+ Here are some of the enhancements this process brings to the model's performance:
125
+
126
+ - Prose generation seems more focused on the moment to moment.
127
+ - Sometimes there will be "preamble" and/or foreshadowing present.
128
+ - Fewer or no "cliches"
129
+ - Better overall prose and/or more complex / nuanced prose.
130
+ - A greater sense of nuance on all levels.
131
+ - Coherence is stronger.
132
+ - Description is more detailed, and connected closer to the content.
133
+ - Simile and Metaphors are stronger and better connected to the prose, story, and character.
134
+ - Sense of "there" / in the moment is enhanced.
135
+ - Details are more vivid, and there are more of them.
136
+ - Prose generation length can be long to extreme.
137
+ - Emotional engagement is stronger.
138
+ - The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
139
+ - The MORE instructions and/or details you provide the more strongly the model will respond.
140
+ - Depending on the model "voice" may be more "human" vs original model's "voice".
141
+
142
+ Other "lab" observations:
143
+
144
+ - This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
145
+ - However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
146
+ - From lab testing it seems to ponder, and consider more carefully roughly speaking.
147
+ - You could say this process sharpens the model's focus on it's task(s) at a deeper level.
148
+
149
+ The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
150
+
151
+
152
+ <B>Model Template:</B>
153
+
154
+ Here is the standard Alpaca template:
155
+
156
+ Alpaca:
157
+
158
+ <pre>
159
+ {
160
+ "name": "Alpaca",
161
+ "inference_params": {
162
+ "input_prefix": "### Instruction:",
163
+ "input_suffix": "### Response:",
164
+ "antiprompt": [
165
+ "### Instruction:"
166
+ ],
167
+ "pre_prompt": "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n"
168
+ }
169
+ }
170
+ </pre>
171
+
172
+ <B>Model "DNA":</B>
173
 
174
  Special thanks to model makers "jondurbin" and "MTSAIR"
175
 
 
179
 
180
  [ https://huggingface.co/MTSAIR/multi_verse_model ]
181
 
 
 
182
  <b>Optional Enhancement:</B>
183
 
184
  The following can be used in place of the "system prompt" or "system role" to further enhance the model.
 
207
 
208
  ---
209
 
210
+ <h3>EXAMPLES PROMPTS and OUTPUT:</h3>
211
+
212
+ Examples are created using quant Q4_K_M, "temp=.8" (unless otherwise stated), minimal parameters and "ALPACA" template.
213
+
214
+ Model has been tested with "temp" from ".1" to "5".
215
+
216
+ Below are the least creative outputs, prompt is in <B>BOLD</B>.
217
 
218
  ---
219