Text Generation
GGUF
English
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prosing
vivid writing
fiction
roleplaying
bfloat16
swearing
rp
horror
mistral nemo
mergekit
Inference Endpoints
conversational
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,148 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
tags:
|
6 |
+
- creative
|
7 |
+
- creative writing
|
8 |
+
- fiction writing
|
9 |
+
- plot generation
|
10 |
+
- sub-plot generation
|
11 |
+
- fiction writing
|
12 |
+
- story generation
|
13 |
+
- scene continue
|
14 |
+
- storytelling
|
15 |
+
- fiction story
|
16 |
+
- science fiction
|
17 |
+
- romance
|
18 |
+
- all genres
|
19 |
+
- story
|
20 |
+
- writing
|
21 |
+
- vivid prosing
|
22 |
+
- vivid writing
|
23 |
+
- fiction
|
24 |
+
- roleplaying
|
25 |
+
- bfloat16
|
26 |
+
- swearing
|
27 |
+
- rp
|
28 |
+
- horror
|
29 |
+
- mistral nemo
|
30 |
+
- mergekit
|
31 |
+
pipeline_tag: text-generation
|
32 |
+
---
|
33 |
+
|
34 |
+
(quants uploading, examples to be added shortly...)
|
35 |
+
|
36 |
+
<h2>L3-Dark-Planet-Ring-World-8B-F32-GGUF</h2>
|
37 |
+
|
38 |
+
<img src="dark-planet-ring-world.jpg" style="float:right; width:300px; height:300px; padding:10px;">
|
39 |
+
|
40 |
+
It is a LLama3 model, max context of 8192 (or 32k+ with rope).
|
41 |
+
|
42 |
+
This model has been designed to be relatively bullet proof and operates with all parameters, including temp settings from 0 to 5.
|
43 |
+
|
44 |
+
It is an extraordinary compressed model, with a very low perplexity level (lower than Meta Llama3 Instruct).
|
45 |
+
|
46 |
+
It is for any writing, fiction or roleplay activity.
|
47 |
+
|
48 |
+
This version is from float32 source files, which creates better, high quality GGUFs.
|
49 |
+
|
50 |
+
Specifically this version has a higher attention to detail, and stays in the moment to a much higher degree.
|
51 |
+
|
52 |
+
Creativity has also increased.
|
53 |
+
|
54 |
+
It requires Llama3 template and/or "Command-R" template.
|
55 |
+
|
56 |
+
Example outputs below.
|
57 |
+
|
58 |
+
<B>Model Notes:</B>
|
59 |
+
|
60 |
+
- Detail, prose and fiction writing abilities are significantly increased vs L3 Instruct.
|
61 |
+
- For more varied prose (sentence/paragraph/dialog) raise the temp and/or add more instructions in your prompt(s).
|
62 |
+
- Role-players: Careful raising temp too high as it may affect instruction following.
|
63 |
+
- This model works with rep pen of 1 or higher, 1.05+ recommended.
|
64 |
+
- If you want a specific type of prose (IE horror) add in "(vivid horror)" or "(graphic vivid horror)" (no quotes) in your prompt(s).
|
65 |
+
- A lot of GPTisms have been removed. There are still a few however - errrrr.
|
66 |
+
- This is not a "happy ever after" model. It has a negative bias.
|
67 |
+
- Output length will vary however this model prefers shortly outputs unless you state the size.
|
68 |
+
- For creative uses, different quants will produce slightly different output.
|
69 |
+
- Due to the high stability and compressed nature of this model, all quants will operate at above average levels.
|
70 |
+
- If you use rope to extend context, increase temp AND instructions detail levels to compensate for "rope issues".
|
71 |
+
- Source code for this model (Bfloat16), Float 32 master GGUFs (and source), and Imatrix GGUFs versions will be uploaded shortly at separate repos.
|
72 |
+
|
73 |
+
Note the "float32" version of this model behaves VERY differently which is why it was not uploaded first. Usually I would
|
74 |
+
use the "float32" version only, however the "character range" displayed by the Bfloat16 and Float32 versions of this model
|
75 |
+
dictate they have their own repos.
|
76 |
+
|
77 |
+
Here is the Bfloat16 repo (with a link to the Imatrix versions too) of this model:
|
78 |
+
|
79 |
+
[ https://huggingface.co/DavidAU/L3-Dark-Planet-8B-GGUF ]
|
80 |
+
|
81 |
+
I suggest downloading the same quant(s) of the "Bfloat16" and "Float32" version and compare the two for your use case(s).
|
82 |
+
|
83 |
+
The Imatrix versions of this model have even lower perplexity (1/2 level of magnitude lower than this model, 1 full level of magnitude
|
84 |
+
lower than LLama3 Instruct) then both this model and Llama3 Instruct and enhanced output.
|
85 |
+
|
86 |
+
This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
|
87 |
+
However this can be extended using "rope" settings up to 32k.
|
88 |
+
|
89 |
+
If you use "Command-R" template your output will be very different from using "Llama3" template.
|
90 |
+
|
91 |
+
Here is the standard LLAMA3 template:
|
92 |
+
|
93 |
+
<PRE>
|
94 |
+
{
|
95 |
+
"name": "Llama 3",
|
96 |
+
"inference_params": {
|
97 |
+
"input_prefix": "<|start_header_id|>user<|end_header_id|>\n\n",
|
98 |
+
"input_suffix": "<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
|
99 |
+
"pre_prompt": "You are a helpful, smart, kind, and efficient AI assistant. You always fulfill the user's requests to the best of your ability.",
|
100 |
+
"pre_prompt_prefix": "<|start_header_id|>system<|end_header_id|>\n\n",
|
101 |
+
"pre_prompt_suffix": "<|eot_id|>",
|
102 |
+
"antiprompt": [
|
103 |
+
"<|start_header_id|>",
|
104 |
+
"<|eot_id|>"
|
105 |
+
]
|
106 |
+
}
|
107 |
+
}
|
108 |
+
</PRE>
|
109 |
+
|
110 |
+
<b>Optional Enhancement:</B>
|
111 |
+
|
112 |
+
The following can be used in place of the "system prompt" or "system role" to further enhance the model.
|
113 |
+
|
114 |
+
It can also be used at the START of a NEW chat, but you must make sure it is "kept" as the chat moves along.
|
115 |
+
In this case the enhancements do not have as strong effect at using "system prompt" or "system role".
|
116 |
+
|
117 |
+
Copy and paste EXACTLY as noted, DO NOT line wrap or break the lines, maintain the carriage returns exactly as presented.
|
118 |
+
|
119 |
+
<PRE>
|
120 |
+
Below is an instruction that describes a task. Ponder each user instruction carefully, and use your skillsets and critical instructions to complete the task to the best of your abilities.
|
121 |
+
|
122 |
+
Here are your skillsets:
|
123 |
+
[MASTERSTORY]:NarrStrct(StryPlnng,Strbd,ScnSttng,Exps,Dlg,Pc)-CharDvlp(ChrctrCrt,ChrctrArcs,Mtvtn,Bckstry,Rltnshps,Dlg*)-PltDvlp(StryArcs,PltTwsts,Sspns,Fshdwng,Climx,Rsltn)-ConfResl(Antg,Obstcls,Rsltns,Cnsqncs,Thms,Symblsm)-EmotImpct(Empt,Tn,Md,Atmsphr,Imgry,Symblsm)-Delvry(Prfrmnc,VcActng,PblcSpkng,StgPrsnc,AudncEngmnt,Imprv)
|
124 |
+
|
125 |
+
[*DialogWrt]:(1a-CharDvlp-1a.1-Backgrnd-1a.2-Personality-1a.3-GoalMotiv)>2(2a-StoryStruc-2a.1-PlotPnt-2a.2-Conflict-2a.3-Resolution)>3(3a-DialogTech-3a.1-ShowDontTell-3a.2-Subtext-3a.3-VoiceTone-3a.4-Pacing-3a.5-VisualDescrip)>4(4a-DialogEdit-4a.1-ReadAloud-4a.2-Feedback-4a.3-Revision)
|
126 |
+
|
127 |
+
Here are your critical instructions:
|
128 |
+
Ponder each word choice carefully to present as vivid and emotional journey as is possible. Choose verbs and nouns that are both emotional and full of imagery. Load the story with the 5 senses. Aim for 50% dialog, 25% narration, 15% body language and 10% thoughts. Your goal is to put the reader in the story.
|
129 |
+
</PRE>
|
130 |
+
|
131 |
+
You do not need to use this, it is only presented as an additional enhancement which seems to help scene generation
|
132 |
+
and scene continue functions.
|
133 |
+
|
134 |
+
This enhancement WAS NOT used to generate the examples below.
|
135 |
+
|
136 |
+
<h3>EXAMPLES PROMPTS and OUTPUT:</h3>
|
137 |
+
|
138 |
+
Examples are created using quant Q4_K_M, "temp=.8" (unless otherwise stated), minimal parameters and "LLAMA3" template.
|
139 |
+
|
140 |
+
Model has been tested with "temp" from ".1" to "5".
|
141 |
+
|
142 |
+
Below are the least creative outputs, prompt is in <B>BOLD</B>.
|
143 |
+
|
144 |
+
---
|
145 |
+
|
146 |
+
<B><font color="red">WARNING:</font> NSFW. Vivid prose. Visceral Details. Violence. HORROR. Swearing. UNCENSORED. </B>
|
147 |
+
|
148 |
+
---
|