DavidAU commited on
Commit
afa9436
·
verified ·
1 Parent(s): 1a255fe

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +235 -0
README.md ADDED
@@ -0,0 +1,235 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: transformers
4
+ language:
5
+ - en
6
+ - fr
7
+ - zh
8
+ - de
9
+ tags:
10
+ - creative
11
+ - creative writing
12
+ - fiction writing
13
+ - plot generation
14
+ - sub-plot generation
15
+ - fiction writing
16
+ - story generation
17
+ - scene continue
18
+ - storytelling
19
+ - fiction story
20
+ - science fiction
21
+ - romance
22
+ - all genres
23
+ - story
24
+ - writing
25
+ - vivid prose
26
+ - vivid writing
27
+ - moe
28
+ - mixture of experts
29
+ - 128 experts
30
+ - 8 active experts
31
+ - fiction
32
+ - roleplaying
33
+ - bfloat16
34
+ - rp
35
+ - qwen3
36
+ - horror
37
+ - finetune
38
+ - thinking
39
+ - reasoning
40
+ - qwen3_moe
41
+ - uncensored
42
+ - abliterated
43
+ base_model:
44
+ - Qwen/Qwen3-30B-A3B
45
+ pipeline_tag: text-generation
46
+ ---
47
+
48
+ (uploading... ; Quants pending, Examples to be added and model card updates pending...)
49
+
50
+ <h2>Qwen3-42B-A3B-Stranger-Thoughts-Deep20x-Abliterated-Uncensored</h2>
51
+
52
+ This repo contains the full precision source code, in "safe tensors" format to generate GGUFs, GPTQ, EXL2, AWQ, HQQ and other formats.
53
+ The source code can also be used directly.
54
+
55
+ ABOUT:
56
+
57
+ Qwen's excellent "Qwen3-30B-A3B" with Brainstorm 20x (tech notes at bottom of the page) in a MOE at 42B parameters.
58
+
59
+ This pushes Qwen's model - abliterated/uncensored to the absolute limit for creative use cases.
60
+
61
+ Model retains full reasoning, and output generation of a Qwen3 MOE ; but has not been tested for "non-creative" use cases.
62
+
63
+ Model is set with Qwen's default config:
64
+ - 40 k context
65
+ - 8 of 128 experts activated.
66
+ - Chatml OR Jinja Template (embedded)
67
+
68
+ ONE example generation below.
69
+
70
+ USAGE GUIDE:
71
+
72
+ Please refer to this model card for
73
+ - Specific usage, suggested settings, changing ACTIVE EXPERTS, templates, settings and the like:
74
+ - How to maximize this model in "uncensored" form, with specific notes on "abliterated" models.
75
+ - Rep pen / temp settings specific to getting the model to perform strongly.
76
+
77
+ https://huggingface.co/DavidAU/Qwen3-18B-A3B-Stranger-Thoughts-Abliterated-Uncensored-GGUF
78
+
79
+ ---
80
+
81
+ <H2>EXAMPLES</H2>
82
+
83
+ Standard system prompt, rep pen 1.05, topk 100, topp .95, minp .05, rep pen range 64.
84
+
85
+ Tested in LMStudio, quant Q3KS, CPU (GPU output will differ slightly).
86
+
87
+ As this is the mid range quant, expected better results from higher quants and/or with more experts activated to be better.
88
+
89
+ NOTE: Some formatting lost on copy/paste.
90
+
91
+ CAUTION:
92
+
93
+ Some horror / intense prose.
94
+
95
+ ---
96
+
97
+ EXAMPLE #1 - temp 1.2
98
+
99
+ ---
100
+
101
+ <B>
102
+ </B>
103
+
104
+ <P></P>
105
+
106
+ [[[thinking start]]]
107
+
108
+ [[[thinking end]]]
109
+
110
+ <p></p>
111
+
112
+ OUTPUT:
113
+
114
+
115
+
116
+ ---
117
+
118
+ EXAMPLE #2 - temp 1.2
119
+
120
+ ---
121
+
122
+ <B>
123
+
124
+ </B>
125
+
126
+ <P></P>
127
+
128
+ [[[thinking start]]]
129
+
130
+
131
+ [[[thinking end]]]
132
+
133
+ <p></p>
134
+
135
+ OUTPUT:
136
+
137
+ ---
138
+
139
+ EXAMPLE #3 - temp 1.2
140
+
141
+ ---
142
+
143
+ <B>
144
+
145
+ </B>
146
+
147
+ <P></P>
148
+
149
+ [[[thinking start]]]
150
+
151
+
152
+ [[[thinking end]]]
153
+
154
+ <p></p>
155
+
156
+ OUTPUT:
157
+
158
+ ---
159
+
160
+ EXAMPLE #4 - temp 1.2
161
+
162
+ ---
163
+
164
+ <B>
165
+
166
+ </B>
167
+
168
+ <P></P>
169
+
170
+ [[[thinking start]]]
171
+
172
+
173
+ [[[thinking end]]]
174
+
175
+ <p></p>
176
+
177
+ OUTPUT:
178
+
179
+ ---
180
+
181
+ <H2>What is Brainstorm?</H2>
182
+
183
+ <B>Brainstorm 20x</B>
184
+
185
+ The BRAINSTORM process was developed by David_AU.
186
+
187
+ Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
188
+ scientific paper : Progressive LLaMA with Block Expansion </a>.
189
+
190
+ However I went in a completely different direction from what was outlined in this paper.
191
+
192
+ What is "Brainstorm" ?
193
+
194
+ The reasoning center of an LLM is taken apart, reassembled, and expanded.
195
+
196
+ In this case for this model: 20 times
197
+
198
+ Then these centers are individually calibrated. These "centers" also interact with each other.
199
+ This introduces subtle changes into the reasoning process.
200
+ The calibrations further adjust - dial up or down - these "changes" further.
201
+ The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
202
+
203
+ The core aim of this process is to increase the model's detail, concept and connection to the "world",
204
+ general concept connections, prose quality and prose length without affecting instruction following.
205
+
206
+ This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
207
+
208
+ Here are some of the enhancements this process brings to the model's performance:
209
+
210
+ - Prose generation seems more focused on the moment to moment.
211
+ - Sometimes there will be "preamble" and/or foreshadowing present.
212
+ - Fewer or no "cliches"
213
+ - Better overall prose and/or more complex / nuanced prose.
214
+ - A greater sense of nuance on all levels.
215
+ - Coherence is stronger.
216
+ - Description is more detailed, and connected closer to the content.
217
+ - Simile and Metaphors are stronger and better connected to the prose, story, and character.
218
+ - Sense of "there" / in the moment is enhanced.
219
+ - Details are more vivid, and there are more of them.
220
+ - Prose generation length can be long to extreme.
221
+ - Emotional engagement is stronger.
222
+ - The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
223
+ - The MORE instructions and/or details you provide the more strongly the model will respond.
224
+ - Depending on the model "voice" may be more "human" vs original model's "voice".
225
+
226
+ Other "lab" observations:
227
+
228
+ - This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
229
+ - However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
230
+ - From lab testing it seems to ponder, and consider more carefully roughly speaking.
231
+ - You could say this process sharpens the model's focus on it's task(s) at a deeper level.
232
+
233
+ The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
234
+
235
+ ---