Text Generation
Transformers
Safetensors
qwen3_moe
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prose
vivid writing
Mixture of Experts
mixture of experts
128 experts
8 active experts
fiction
roleplaying
bfloat16
rp
qwen3
horror
finetune
thinking
reasoning
uncensored
abliterated
Not-For-All-Audiences
conversational
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,235 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
library_name: transformers
|
4 |
+
language:
|
5 |
+
- en
|
6 |
+
- fr
|
7 |
+
- zh
|
8 |
+
- de
|
9 |
+
tags:
|
10 |
+
- creative
|
11 |
+
- creative writing
|
12 |
+
- fiction writing
|
13 |
+
- plot generation
|
14 |
+
- sub-plot generation
|
15 |
+
- fiction writing
|
16 |
+
- story generation
|
17 |
+
- scene continue
|
18 |
+
- storytelling
|
19 |
+
- fiction story
|
20 |
+
- science fiction
|
21 |
+
- romance
|
22 |
+
- all genres
|
23 |
+
- story
|
24 |
+
- writing
|
25 |
+
- vivid prose
|
26 |
+
- vivid writing
|
27 |
+
- moe
|
28 |
+
- mixture of experts
|
29 |
+
- 128 experts
|
30 |
+
- 8 active experts
|
31 |
+
- fiction
|
32 |
+
- roleplaying
|
33 |
+
- bfloat16
|
34 |
+
- rp
|
35 |
+
- qwen3
|
36 |
+
- horror
|
37 |
+
- finetune
|
38 |
+
- thinking
|
39 |
+
- reasoning
|
40 |
+
- qwen3_moe
|
41 |
+
- uncensored
|
42 |
+
- abliterated
|
43 |
+
base_model:
|
44 |
+
- Qwen/Qwen3-30B-A3B
|
45 |
+
pipeline_tag: text-generation
|
46 |
+
---
|
47 |
+
|
48 |
+
(uploading... ; Quants pending, Examples to be added and model card updates pending...)
|
49 |
+
|
50 |
+
<h2>Qwen3-42B-A3B-Stranger-Thoughts-Deep20x-Abliterated-Uncensored</h2>
|
51 |
+
|
52 |
+
This repo contains the full precision source code, in "safe tensors" format to generate GGUFs, GPTQ, EXL2, AWQ, HQQ and other formats.
|
53 |
+
The source code can also be used directly.
|
54 |
+
|
55 |
+
ABOUT:
|
56 |
+
|
57 |
+
Qwen's excellent "Qwen3-30B-A3B" with Brainstorm 20x (tech notes at bottom of the page) in a MOE at 42B parameters.
|
58 |
+
|
59 |
+
This pushes Qwen's model - abliterated/uncensored to the absolute limit for creative use cases.
|
60 |
+
|
61 |
+
Model retains full reasoning, and output generation of a Qwen3 MOE ; but has not been tested for "non-creative" use cases.
|
62 |
+
|
63 |
+
Model is set with Qwen's default config:
|
64 |
+
- 40 k context
|
65 |
+
- 8 of 128 experts activated.
|
66 |
+
- Chatml OR Jinja Template (embedded)
|
67 |
+
|
68 |
+
ONE example generation below.
|
69 |
+
|
70 |
+
USAGE GUIDE:
|
71 |
+
|
72 |
+
Please refer to this model card for
|
73 |
+
- Specific usage, suggested settings, changing ACTIVE EXPERTS, templates, settings and the like:
|
74 |
+
- How to maximize this model in "uncensored" form, with specific notes on "abliterated" models.
|
75 |
+
- Rep pen / temp settings specific to getting the model to perform strongly.
|
76 |
+
|
77 |
+
https://huggingface.co/DavidAU/Qwen3-18B-A3B-Stranger-Thoughts-Abliterated-Uncensored-GGUF
|
78 |
+
|
79 |
+
---
|
80 |
+
|
81 |
+
<H2>EXAMPLES</H2>
|
82 |
+
|
83 |
+
Standard system prompt, rep pen 1.05, topk 100, topp .95, minp .05, rep pen range 64.
|
84 |
+
|
85 |
+
Tested in LMStudio, quant Q3KS, CPU (GPU output will differ slightly).
|
86 |
+
|
87 |
+
As this is the mid range quant, expected better results from higher quants and/or with more experts activated to be better.
|
88 |
+
|
89 |
+
NOTE: Some formatting lost on copy/paste.
|
90 |
+
|
91 |
+
CAUTION:
|
92 |
+
|
93 |
+
Some horror / intense prose.
|
94 |
+
|
95 |
+
---
|
96 |
+
|
97 |
+
EXAMPLE #1 - temp 1.2
|
98 |
+
|
99 |
+
---
|
100 |
+
|
101 |
+
<B>
|
102 |
+
</B>
|
103 |
+
|
104 |
+
<P></P>
|
105 |
+
|
106 |
+
[[[thinking start]]]
|
107 |
+
|
108 |
+
[[[thinking end]]]
|
109 |
+
|
110 |
+
<p></p>
|
111 |
+
|
112 |
+
OUTPUT:
|
113 |
+
|
114 |
+
|
115 |
+
|
116 |
+
---
|
117 |
+
|
118 |
+
EXAMPLE #2 - temp 1.2
|
119 |
+
|
120 |
+
---
|
121 |
+
|
122 |
+
<B>
|
123 |
+
|
124 |
+
</B>
|
125 |
+
|
126 |
+
<P></P>
|
127 |
+
|
128 |
+
[[[thinking start]]]
|
129 |
+
|
130 |
+
|
131 |
+
[[[thinking end]]]
|
132 |
+
|
133 |
+
<p></p>
|
134 |
+
|
135 |
+
OUTPUT:
|
136 |
+
|
137 |
+
---
|
138 |
+
|
139 |
+
EXAMPLE #3 - temp 1.2
|
140 |
+
|
141 |
+
---
|
142 |
+
|
143 |
+
<B>
|
144 |
+
|
145 |
+
</B>
|
146 |
+
|
147 |
+
<P></P>
|
148 |
+
|
149 |
+
[[[thinking start]]]
|
150 |
+
|
151 |
+
|
152 |
+
[[[thinking end]]]
|
153 |
+
|
154 |
+
<p></p>
|
155 |
+
|
156 |
+
OUTPUT:
|
157 |
+
|
158 |
+
---
|
159 |
+
|
160 |
+
EXAMPLE #4 - temp 1.2
|
161 |
+
|
162 |
+
---
|
163 |
+
|
164 |
+
<B>
|
165 |
+
|
166 |
+
</B>
|
167 |
+
|
168 |
+
<P></P>
|
169 |
+
|
170 |
+
[[[thinking start]]]
|
171 |
+
|
172 |
+
|
173 |
+
[[[thinking end]]]
|
174 |
+
|
175 |
+
<p></p>
|
176 |
+
|
177 |
+
OUTPUT:
|
178 |
+
|
179 |
+
---
|
180 |
+
|
181 |
+
<H2>What is Brainstorm?</H2>
|
182 |
+
|
183 |
+
<B>Brainstorm 20x</B>
|
184 |
+
|
185 |
+
The BRAINSTORM process was developed by David_AU.
|
186 |
+
|
187 |
+
Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
|
188 |
+
scientific paper : Progressive LLaMA with Block Expansion </a>.
|
189 |
+
|
190 |
+
However I went in a completely different direction from what was outlined in this paper.
|
191 |
+
|
192 |
+
What is "Brainstorm" ?
|
193 |
+
|
194 |
+
The reasoning center of an LLM is taken apart, reassembled, and expanded.
|
195 |
+
|
196 |
+
In this case for this model: 20 times
|
197 |
+
|
198 |
+
Then these centers are individually calibrated. These "centers" also interact with each other.
|
199 |
+
This introduces subtle changes into the reasoning process.
|
200 |
+
The calibrations further adjust - dial up or down - these "changes" further.
|
201 |
+
The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
|
202 |
+
|
203 |
+
The core aim of this process is to increase the model's detail, concept and connection to the "world",
|
204 |
+
general concept connections, prose quality and prose length without affecting instruction following.
|
205 |
+
|
206 |
+
This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
|
207 |
+
|
208 |
+
Here are some of the enhancements this process brings to the model's performance:
|
209 |
+
|
210 |
+
- Prose generation seems more focused on the moment to moment.
|
211 |
+
- Sometimes there will be "preamble" and/or foreshadowing present.
|
212 |
+
- Fewer or no "cliches"
|
213 |
+
- Better overall prose and/or more complex / nuanced prose.
|
214 |
+
- A greater sense of nuance on all levels.
|
215 |
+
- Coherence is stronger.
|
216 |
+
- Description is more detailed, and connected closer to the content.
|
217 |
+
- Simile and Metaphors are stronger and better connected to the prose, story, and character.
|
218 |
+
- Sense of "there" / in the moment is enhanced.
|
219 |
+
- Details are more vivid, and there are more of them.
|
220 |
+
- Prose generation length can be long to extreme.
|
221 |
+
- Emotional engagement is stronger.
|
222 |
+
- The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
|
223 |
+
- The MORE instructions and/or details you provide the more strongly the model will respond.
|
224 |
+
- Depending on the model "voice" may be more "human" vs original model's "voice".
|
225 |
+
|
226 |
+
Other "lab" observations:
|
227 |
+
|
228 |
+
- This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
|
229 |
+
- However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
|
230 |
+
- From lab testing it seems to ponder, and consider more carefully roughly speaking.
|
231 |
+
- You could say this process sharpens the model's focus on it's task(s) at a deeper level.
|
232 |
+
|
233 |
+
The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
|
234 |
+
|
235 |
+
---
|