Text Generation
GGUF
English
Chinese
MOE
Qwen 2.5 MOE
Mixture of Experts
Uncensored
2X1.5B
deepseek
reasoning
thinking
creative
128k context
general usage
problem solving
brainstorming
solve riddles
story generation
plot generation
storytelling
fiction story
story
writing
fiction
Qwen 2.5
mergekit
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -64,6 +64,11 @@ and I would also say more "sciency" too.
|
|
64 |
|
65 |
This does not mean it will not work for your use case.
|
66 |
|
|
|
|
|
|
|
|
|
|
|
67 |
Four examples below so you have some idea what this model can do.
|
68 |
|
69 |
Keep in mind this model is two 1.5B parameters models working together, and will not have the power of a 14B or 32B reasoning/thinking model.
|
|
|
64 |
|
65 |
This does not mean it will not work for your use case.
|
66 |
|
67 |
+
Likewise, this model may require more direction, details, and what you are asking in the prompt to "think" along "narrower" lines.
|
68 |
+
|
69 |
+
It may take 2-4 generations for the model to zero in / get what you mean and "think" along the correct lines, if your prompt(s)
|
70 |
+
are too short.
|
71 |
+
|
72 |
Four examples below so you have some idea what this model can do.
|
73 |
|
74 |
Keep in mind this model is two 1.5B parameters models working together, and will not have the power of a 14B or 32B reasoning/thinking model.
|