Text Generation
Transformers
Safetensors
qwen3_moe
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prose
vivid writing
Mixture of Experts
mixture of experts
128 experts
8 active experts
fiction
roleplaying
bfloat16
rp
qwen3
horror
finetune
thinking
reasoning
Merge
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -56,11 +56,15 @@ The source code can also be used directly.
|
|
56 |
|
57 |
ABOUT:
|
58 |
|
59 |
-
A very special blend of Qwen's 30B-A3B 128 expert MOE, with Brainstorm 5X (by DavidAU) multi.
|
60 |
|
61 |
4 Layers are added containing 128 experts each, from 4 different models (see model tree).
|
62 |
|
63 |
-
This is a very unique model.
|
|
|
|
|
|
|
|
|
64 |
|
65 |
USAGE:
|
66 |
|
@@ -639,3 +643,62 @@ That’s the first horror novel scene. The action starts in the alley, the prota
|
|
639 |
|
640 |
Word count: 2,000.
|
641 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
56 |
|
57 |
ABOUT:
|
58 |
|
59 |
+
A very special blend of Qwen's 30B-A3B 128 expert MOE, with Brainstorm 5X (by DavidAU, details at bottom of this page) multi.
|
60 |
|
61 |
4 Layers are added containing 128 experts each, from 4 different models (see model tree).
|
62 |
|
63 |
+
This is a very unique model specifically for creative use cases.
|
64 |
+
|
65 |
+
These modifications change the prose, style, reasoning and output generation.
|
66 |
+
|
67 |
+
4 examples below.
|
68 |
|
69 |
USAGE:
|
70 |
|
|
|
643 |
|
644 |
Word count: 2,000.
|
645 |
|
646 |
+
---
|
647 |
+
|
648 |
+
<H2>What is Brainstorm?</H2>
|
649 |
+
|
650 |
+
---
|
651 |
+
|
652 |
+
<B>Brainstorm 20x</B>
|
653 |
+
|
654 |
+
The BRAINSTORM process was developed by David_AU.
|
655 |
+
|
656 |
+
Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
|
657 |
+
scientific paper : Progressive LLaMA with Block Expansion </a>.
|
658 |
+
|
659 |
+
However I went in a completely different direction from what was outlined in this paper.
|
660 |
+
|
661 |
+
What is "Brainstorm" ?
|
662 |
+
|
663 |
+
The reasoning center of an LLM is taken apart, reassembled, and expanded.
|
664 |
+
|
665 |
+
In this case for this model: 20 times
|
666 |
+
|
667 |
+
Then these centers are individually calibrated. These "centers" also interact with each other.
|
668 |
+
This introduces subtle changes into the reasoning process.
|
669 |
+
The calibrations further adjust - dial up or down - these "changes" further.
|
670 |
+
The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
|
671 |
+
|
672 |
+
The core aim of this process is to increase the model's detail, concept and connection to the "world",
|
673 |
+
general concept connections, prose quality and prose length without affecting instruction following.
|
674 |
+
|
675 |
+
This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
|
676 |
+
|
677 |
+
Here are some of the enhancements this process brings to the model's performance:
|
678 |
+
|
679 |
+
- Prose generation seems more focused on the moment to moment.
|
680 |
+
- Sometimes there will be "preamble" and/or foreshadowing present.
|
681 |
+
- Fewer or no "cliches"
|
682 |
+
- Better overall prose and/or more complex / nuanced prose.
|
683 |
+
- A greater sense of nuance on all levels.
|
684 |
+
- Coherence is stronger.
|
685 |
+
- Description is more detailed, and connected closer to the content.
|
686 |
+
- Simile and Metaphors are stronger and better connected to the prose, story, and character.
|
687 |
+
- Sense of "there" / in the moment is enhanced.
|
688 |
+
- Details are more vivid, and there are more of them.
|
689 |
+
- Prose generation length can be long to extreme.
|
690 |
+
- Emotional engagement is stronger.
|
691 |
+
- The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
|
692 |
+
- The MORE instructions and/or details you provide the more strongly the model will respond.
|
693 |
+
- Depending on the model "voice" may be more "human" vs original model's "voice".
|
694 |
+
|
695 |
+
Other "lab" observations:
|
696 |
+
|
697 |
+
- This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
|
698 |
+
- However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
|
699 |
+
- From lab testing it seems to ponder, and consider more carefully roughly speaking.
|
700 |
+
- You could say this process sharpens the model's focus on it's task(s) at a deeper level.
|
701 |
+
|
702 |
+
The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
|
703 |
+
|
704 |
+
---
|