DavidAU commited on
Commit
f33a0e9
·
verified ·
1 Parent(s): 0e7c06c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +65 -2
README.md CHANGED
@@ -56,11 +56,15 @@ The source code can also be used directly.
56
 
57
  ABOUT:
58
 
59
- A very special blend of Qwen's 30B-A3B 128 expert MOE, with Brainstorm 5X (by DavidAU) multi.
60
 
61
  4 Layers are added containing 128 experts each, from 4 different models (see model tree).
62
 
63
- This is a very unique model.
 
 
 
 
64
 
65
  USAGE:
66
 
@@ -639,3 +643,62 @@ That’s the first horror novel scene. The action starts in the alley, the prota
639
 
640
  Word count: 2,000.
641
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
56
 
57
  ABOUT:
58
 
59
+ A very special blend of Qwen's 30B-A3B 128 expert MOE, with Brainstorm 5X (by DavidAU, details at bottom of this page) multi.
60
 
61
  4 Layers are added containing 128 experts each, from 4 different models (see model tree).
62
 
63
+ This is a very unique model specifically for creative use cases.
64
+
65
+ These modifications change the prose, style, reasoning and output generation.
66
+
67
+ 4 examples below.
68
 
69
  USAGE:
70
 
 
643
 
644
  Word count: 2,000.
645
 
646
+ ---
647
+
648
+ <H2>What is Brainstorm?</H2>
649
+
650
+ ---
651
+
652
+ <B>Brainstorm 20x</B>
653
+
654
+ The BRAINSTORM process was developed by David_AU.
655
+
656
+ Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
657
+ scientific paper : Progressive LLaMA with Block Expansion </a>.
658
+
659
+ However I went in a completely different direction from what was outlined in this paper.
660
+
661
+ What is "Brainstorm" ?
662
+
663
+ The reasoning center of an LLM is taken apart, reassembled, and expanded.
664
+
665
+ In this case for this model: 20 times
666
+
667
+ Then these centers are individually calibrated. These "centers" also interact with each other.
668
+ This introduces subtle changes into the reasoning process.
669
+ The calibrations further adjust - dial up or down - these "changes" further.
670
+ The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
671
+
672
+ The core aim of this process is to increase the model's detail, concept and connection to the "world",
673
+ general concept connections, prose quality and prose length without affecting instruction following.
674
+
675
+ This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
676
+
677
+ Here are some of the enhancements this process brings to the model's performance:
678
+
679
+ - Prose generation seems more focused on the moment to moment.
680
+ - Sometimes there will be "preamble" and/or foreshadowing present.
681
+ - Fewer or no "cliches"
682
+ - Better overall prose and/or more complex / nuanced prose.
683
+ - A greater sense of nuance on all levels.
684
+ - Coherence is stronger.
685
+ - Description is more detailed, and connected closer to the content.
686
+ - Simile and Metaphors are stronger and better connected to the prose, story, and character.
687
+ - Sense of "there" / in the moment is enhanced.
688
+ - Details are more vivid, and there are more of them.
689
+ - Prose generation length can be long to extreme.
690
+ - Emotional engagement is stronger.
691
+ - The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
692
+ - The MORE instructions and/or details you provide the more strongly the model will respond.
693
+ - Depending on the model "voice" may be more "human" vs original model's "voice".
694
+
695
+ Other "lab" observations:
696
+
697
+ - This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
698
+ - However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
699
+ - From lab testing it seems to ponder, and consider more carefully roughly speaking.
700
+ - You could say this process sharpens the model's focus on it's task(s) at a deeper level.
701
+
702
+ The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
703
+
704
+ ---