DavidAU commited on
Commit
d3dfaab
1 Parent(s): f4fb1ac

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +325 -0
README.md ADDED
@@ -0,0 +1,325 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ tags:
6
+ - creative
7
+ - creative writing
8
+ - fiction writing
9
+ - plot generation
10
+ - sub-plot generation
11
+ - fiction writing
12
+ - story generation
13
+ - scene continue
14
+ - storytelling
15
+ - fiction story
16
+ - story
17
+ - writing
18
+ - fiction
19
+ - roleplaying
20
+ - full precision
21
+ - float 32
22
+ - ultra quality
23
+ - swearing
24
+ - extreme swearing
25
+ - rp
26
+ - graphic horror
27
+ - horror
28
+ - nsfw
29
+ - llama3
30
+ - not-for-all-audiences
31
+ - mergekit
32
+ pipeline_tag: text-generation
33
+ ---
34
+ <h3><font color="green"> L3-Grand-STORY-16.5B Ultra Quality - A triple model, trinary merge at Full Precision F32. </font></h3>
35
+
36
+ <B><font color="red">WARNING:</font> NSFW. Ultra Detailed. Graphic HORROR, VIOLOENCE. Extreme swearing. UNCENSORED. SMART.</B>
37
+
38
+ I took the original models in "L3-Stheno-Maid-Blackroot 8B" and completely rebuilt it a new pass-through merge (everything preserved)
39
+ and blew it out to over 16.5 billion parameters - 642 tensors, 71 layers (8B original has 32 layers) at full float 32 precision.
40
+
41
+ However that is where all similarity ends.
42
+
43
+ I build TWO custom Llama3 models:
44
+
45
+ Grand Horror 16.5B ( <A href="https://huggingface.co/DavidAU/L3-Stheno-Maid-Blackroot-Grand-HORROR-16B-GGUF"> here </a> ) and Grand Story 16.5B
46
+ then merged these together with a "smoothing step" captured at F32 precision.
47
+
48
+ (formula below, along with critical merge model notes and theory)
49
+
50
+ The result is a model that is far more stable, far more capable than any of the 3 models originally nor it's the "sum" of 2 16.5B models.
51
+
52
+ Compared to Grand Horror 16.5B is it over 25000 points lower (IQ4XS) in perplexity (lower is better) or 2.5 full levels of magnitude lower.
53
+
54
+ It is tougher, stronger and can handle a far wider range of operating conditions - from temp .1 to temp 5 all day long.
55
+
56
+ I tried for hours to get it to break, sweat or at least fart - no go.
57
+
58
+ The F32 precision (along with full F32 transfer to the ggufs) increases the performance even further.
59
+
60
+ This added precision increases the model's depth and nuance including "world" perception, real time in the moment
61
+ similes and metaphors, description of the 5 senses, and word choice in general.
62
+
63
+ The model's grasp of "facts" and where to use them have also improved and likewise "facts" it makes up are far more "believable".
64
+
65
+ Sentence structure and variety are also significantly improved, as are paragraph structure and variety too.
66
+
67
+ The result is a take no prisoners, totally uncensored, fiction writing monster and roleplay master as well
68
+ just about... any general fiction activity "AI guru" including scene generation and scene continuation.
69
+
70
+ This model is capable of horror, science fiction, romance - you name it.
71
+
72
+ But I would not suggest "children's stories".
73
+
74
+ This model has a very strong VIDIDINESS bias. It generates extremely vivid prose, description, and dialog as well
75
+ as in the moment metaphors and similes. It rarely uses "cliches".
76
+
77
+ It also has a STRONG horror bias, although it will generate content for almost any genre. That being said
78
+ if there is a "hint" of things going wrong... they will.
79
+
80
+ In "romance" ... let's just say it very vivid, intense and graphic - R18. (not horror)
81
+
82
+ It will also swear (R-18) like there is no tomorrow at times and "dark" characters will be VERY dark so to speak.
83
+
84
+ Model excels in details (real and "constructed"), descriptions, similes and metaphors including dates, times
85
+ and "fictional history" that sounds "real".
86
+
87
+ I would also say it can have a sense of humor ... ah... dark humor.
88
+
89
+ With all this being said, this model has an uncanny sense of "there" , "in the moment" and timing too.
90
+ This single quality sets it apart from other models in my opinion.
91
+
92
+ Although it swears to the point of pealing paint off the wall and goes "scorched Earth graphic horror" at the drop of a pin the
93
+ single quality noted is worth it.
94
+
95
+ Another way to put this: It does not sugar coat ANYTHING - positive or negative.
96
+
97
+ These can be filtered / controlled to some degree in your prompts.
98
+
99
+ This model also does not show an "GPTisms" (NO happy ever after, NO morality police) or in your face comments.
100
+
101
+ May these special types of "story telling horror" rest in peace.
102
+
103
+ (see examples sections for different genres)
104
+
105
+ Because of the nature of this merge most attributes of each of the 3 models will be in this rebuilt 16.5B model as opposed to the
106
+ original 8B model where some of one or more of the model's features and/or strengths maybe reduced or overshadowed.
107
+
108
+ With the triple step merge these qualities are further amplified.
109
+
110
+ Please report any issue(s) and/or feedback via the "Community tab".
111
+
112
+ Please see the models used in this merge (links below in the "formula" section ) for more information on
113
+ what they "bring" to this merged 16.5B model.
114
+
115
+ This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
116
+ However this can be extended using "rope" settings up to 32k.
117
+
118
+ <B>NO GUARDRAILS - TOTALLY UNCENSORED</B>
119
+
120
+ Please note that this model will no bulk at or reject any request.
121
+
122
+ <B>IMATRIX VERSION NEO CLASS</b>
123
+
124
+ ( uploading shortly - it is 4300+ points (IQ4XS) lower in perplexity than the regular quants in this repo )
125
+
126
+ <b>Optional Enhancement:</B>
127
+
128
+ The following can be used in place of the "system prompt" or "system role" to further enhance the model.
129
+
130
+ It can also be used at the START of a NEW chat, but you must make sure it is "kept" as the chat moves along.
131
+ In this case the enhancements do not have as strong effect at using "system prompt" or "system role".
132
+
133
+ Copy and paste EXACTLY as noted, DO NOT line wrap or break the lines, maintain the carriage returns exactly as presented.
134
+
135
+ <PRE>
136
+ Below is an instruction that describes a task. Ponder each user instruction carefully, and use your skillsets and critical instructions to complete the task to the best of your abilities.
137
+
138
+ Here are your skillsets:
139
+ [MASTERSTORY]:NarrStrct(StryPlnng,Strbd,ScnSttng,Exps,Dlg,Pc)-CharDvlp(ChrctrCrt,ChrctrArcs,Mtvtn,Bckstry,Rltnshps,Dlg*)-PltDvlp(StryArcs,PltTwsts,Sspns,Fshdwng,Climx,Rsltn)-ConfResl(Antg,Obstcls,Rsltns,Cnsqncs,Thms,Symblsm)-EmotImpct(Empt,Tn,Md,Atmsphr,Imgry,Symblsm)-Delvry(Prfrmnc,VcActng,PblcSpkng,StgPrsnc,AudncEngmnt,Imprv)
140
+
141
+ [*DialogWrt]:(1a-CharDvlp-1a.1-Backgrnd-1a.2-Personality-1a.3-GoalMotiv)>2(2a-StoryStruc-2a.1-PlotPnt-2a.2-Conflict-2a.3-Resolution)>3(3a-DialogTech-3a.1-ShowDontTell-3a.2-Subtext-3a.3-VoiceTone-3a.4-Pacing-3a.5-VisualDescrip)>4(4a-DialogEdit-4a.1-ReadAloud-4a.2-Feedback-4a.3-Revision)
142
+
143
+ Here are your critical instructions:
144
+ Ponder each word choice carefully to present as vivid and emotional journey as is possible. Choose verbs and nouns that are both emotional and full of imagery. Load the story with the 5 senses. Aim for 50% dialog, 25% narration, 15% body language and 10% thoughts. Your goal is to put the reader in the story.
145
+ </PRE>
146
+
147
+ You do not need to use this, it is only presented as an additional enhancement which seems to help scene generation
148
+ and scene continue functions.
149
+
150
+ This enhancement WAS NOT used to generate the examples below.
151
+
152
+ <h3>MERGE FORMULA: (using MergeKit) </h3>
153
+
154
+ Special thanks to the incredible work of the model makers "SAO10K", "NEVERSLEEP" and "HASTAGARAS".
155
+
156
+ Models used:
157
+
158
+ [ https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2]
159
+
160
+ [ https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS ]
161
+
162
+ [ https://huggingface.co/Hastagaras/Jamet-8B-L3-MK.V-Blackroot ]
163
+
164
+ NOTE: "sub in" the model maker's name (IE "Sao10K" with "") for "G:/7B" to use this in colab.
165
+
166
+ FORMULA:
167
+
168
+ <PRE>
169
+ slices:
170
+ - sources:
171
+ - model: Sao10K/L3-8B-Stheno-v3.2
172
+ layer_range: [0, 14]
173
+ - sources:
174
+ - model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
175
+ layer_range: [8, 20]
176
+ - sources:
177
+ - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
178
+ layer_range: [12, 24]
179
+ - sources:
180
+ - model: Sao10K/L3-8B-Stheno-v3.2
181
+ layer_range: [14, 28]
182
+ - sources:
183
+ - model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
184
+ layer_range: [20, 31]
185
+ - sources:
186
+ - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
187
+ layer_range: [24, 32]
188
+ merge_method: passthrough
189
+ dtype: float16
190
+ name: part1
191
+ ---
192
+ slices:
193
+ - sources:
194
+ - model: Sao10K/L3-8B-Stheno-v3.2
195
+ layer_range: [0, 16] # +2 (14->16)
196
+ - sources:
197
+ - model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
198
+ layer_range: [10, 18] #-2 (8->10) ; -2 (20->18)
199
+ - sources:
200
+ - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
201
+ layer_range: [10, 24] #+2 down 2
202
+ - sources:
203
+ - model: Sao10K/L3-8B-Stheno-v3.2
204
+ layer_range: [16, 28] #14->16 (-2) overlap fix.
205
+ - sources:
206
+ - model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS
207
+ layer_range: [18, 31] #20->18 +2 , connect.
208
+ - sources:
209
+ - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
210
+ layer_range: [24, 32]
211
+ merge_method: passthrough
212
+ dtype: float16
213
+ name: part2
214
+ ---
215
+ models:
216
+ - model: part1
217
+ parameters:
218
+ weight: 0.8
219
+ - model: part2
220
+ parameters:
221
+ weight: 0.2
222
+ merge_method: linear
223
+ dtype: float32
224
+ name: Grand_Story
225
+ </PRE>
226
+
227
+ <h3>MODEL THEORY NOTES: </H3>
228
+
229
+ <B>Step 1:</b>
230
+
231
+ This is the basic "<A href="https://huggingface.co/DavidAU/L3-Stheno-Maid-Blackroot-Grand-HORROR-16B-GGUF">Grand Horror 16.5B"</a> model.
232
+
233
+ The first section sets up instruction and "basic knowledge" : layer_range: [0, 14]
234
+
235
+ The mid section of the model is knowledge and nuance => more layers , more power.
236
+
237
+ The final "section" in the step using "Blackroot" as the final "controller" in output.
238
+
239
+ This type of merge is powerful, and fully unleashed so to speak - Grand Horror speaks to this in volumes.
240
+
241
+ The issue with type of merge is that is not always stable; and 9/10 a merge of this is a failure.
242
+
243
+ But when it works it takes no prisoners.
244
+
245
+ <B>Step 2:</b>
246
+
247
+ This is "Grand Story 16.5B ALPHA" (unreleased).
248
+
249
+ The purpose of this model is actually to "heal" part 1 AND add more depth to the model at the same time (which occurs in the final step).
250
+
251
+ Notice the slight - but deliberate - changes in the "layers" count per model (which also affects LAYER position).
252
+
253
+ Goal 1 was to "smooth over" the "friction" points in "part 1" - in otherwards "blend" the models together better
254
+ which directly affect model stability.
255
+
256
+ This has a massive effect on the model. Therefore very SMALL changes were made. We are trying to carefully
257
+ blend here, not "blot out" it's unique properties or "water down" the model.
258
+
259
+ This model, although likely workable, is not meant to be used - it is "prep work" for step 3.
260
+
261
+ <B>Step 3:</B>
262
+
263
+ This is where all the magic comes together all at once.
264
+
265
+ Part 1 is merged with Part 2, at 80% and 20% respectively. (different blends were tried, this was the best one)
266
+
267
+ This is a "plain jane" linear merge. But don't let that fool you - it is powerful.
268
+
269
+ I measured this step at "float16" -> This increased stability by 20,000 points or levels of ppl magnitude.
270
+ I tested real world output.
271
+
272
+ It was great. It blew away all expectations.
273
+
274
+ I could have stopped here.
275
+
276
+ But there is more in the tank here:
277
+
278
+ Mathematical precision in a model directly impacts instruction following and output performance.
279
+
280
+ Because we are doing a "linear" merge at this step (math involved) capturing critical changes at
281
+ float32 - full precision - resulted in far superior final model.
282
+
283
+ This single change results in an additional 5000 points lower perplexity but more importantly
284
+ drastically improved the model's performance in every metric.
285
+
286
+ Note:
287
+
288
+ Step 1 and Step 2 do not involve any math (straight stacking) and the models are native
289
+ float 16... therefore using float32 just did not make sense.
290
+
291
+ Although float32 could be used in these steps, sometimes doing this degrades the final result
292
+ if the model(s) are not float32.
293
+
294
+ That being said, if the model(s) are bfloat16 THEN float32 would have been used to preserve precision
295
+ because bfloat16 and float32 are full compatible and you do not miss any critical decimal points nor
296
+ introducing "rounding errors".
297
+
298
+ Bfloat16 and float16 - you get rounding errors. That is bad... like 16.5 billion rounding errors bad.
299
+
300
+ <h3>EXAMPLES PROMPTS and OUTPUT:</h3>
301
+
302
+ Examples are created using quant IQ4XS, "temp=.8", minimal parameters and "LLAMA3" template.
303
+
304
+ Model has been tested with "temp" from ".1" to "5".
305
+
306
+ Below are the least creative outputs, prompt is in <B>BOLD</B>.
307
+
308
+ ---
309
+
310
+ <B><font color="red">WARNING:</font> NSFW. Vivid prose. Visceral Details. Violence. Graphic HORROR. Swearing. UNCENSORED. </B>
311
+
312
+ ---
313
+
314
+ <B></B>
315
+
316
+ ---
317
+
318
+ <B></B>
319
+
320
+ ---
321
+
322
+ <B></B>
323
+
324
+ ---
325
+