Update README.md
Browse files
README.md
CHANGED
@@ -23,7 +23,7 @@ tags:
|
|
23 |
- qwen-coder
|
24 |
---
|
25 |
|
26 |
-
<h2>Qwen2.5-2X11B-CODER-Dueling-Wolverines-
|
27 |
|
28 |
<img src="duel-wolverine-coder.jpg" style="float:right; width:300px; height:500px; padding:10px;">
|
29 |
|
@@ -35,7 +35,7 @@ Two monsters in fact - in one.
|
|
35 |
|
36 |
This is MOE model, using V1 and V2 of Wolverine-Coder 11B which is a merge of two models noted below.
|
37 |
|
38 |
-
The MOE config gives you full access to both 11B models at full power - full 22B parameters.
|
39 |
|
40 |
This MOE model generates stronger, more compact code with an enhanced understanding of your instructions
|
41 |
and follows what you tell them to the letter.
|
@@ -69,7 +69,7 @@ NOTES:
|
|
69 |
- Each config/version will be very different from each other.
|
70 |
- You can select 1 or 2 experts, default is 2 experts.
|
71 |
- Due to unique setup of this moe, suggest 1-4 generations.
|
72 |
-
- Total model size is
|
73 |
- Tool Calling is supported in both versions.
|
74 |
- Source(s) / full quanting to follow // full repos to follow.
|
75 |
- Final model size (including layers/tensors) / config subject to change.
|
|
|
23 |
- qwen-coder
|
24 |
---
|
25 |
|
26 |
+
<h2>Qwen2.5-2X11B-CODER-Dueling-Wolverines-28B-gguf</h2>
|
27 |
|
28 |
<img src="duel-wolverine-coder.jpg" style="float:right; width:300px; height:500px; padding:10px;">
|
29 |
|
|
|
35 |
|
36 |
This is MOE model, using V1 and V2 of Wolverine-Coder 11B which is a merge of two models noted below.
|
37 |
|
38 |
+
The MOE config gives you full access to both 11B models at full power - full 22B parameters (an additional shared expert brings this to 28B).
|
39 |
|
40 |
This MOE model generates stronger, more compact code with an enhanced understanding of your instructions
|
41 |
and follows what you tell them to the letter.
|
|
|
69 |
- Each config/version will be very different from each other.
|
70 |
- You can select 1 or 2 experts, default is 2 experts.
|
71 |
- Due to unique setup of this moe, suggest 1-4 generations.
|
72 |
+
- Total model size is 28B because Qwen 2.5 MOES have a shared expert in addition to "regular experts"
|
73 |
- Tool Calling is supported in both versions.
|
74 |
- Source(s) / full quanting to follow // full repos to follow.
|
75 |
- Final model size (including layers/tensors) / config subject to change.
|