DavidAU commited on
Commit
b56c95b
·
verified ·
1 Parent(s): f64aeb6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -23,7 +23,7 @@ tags:
23
  - qwen-coder
24
  ---
25
 
26
- <h2>Qwen2.5-2X11B-CODER-Dueling-Wolverines-25B-gguf</h2>
27
 
28
  <img src="duel-wolverine-coder.jpg" style="float:right; width:300px; height:500px; padding:10px;">
29
 
@@ -35,7 +35,7 @@ Two monsters in fact - in one.
35
 
36
  This is MOE model, using V1 and V2 of Wolverine-Coder 11B which is a merge of two models noted below.
37
 
38
- The MOE config gives you full access to both 11B models at full power - full 22B parameters.
39
 
40
  This MOE model generates stronger, more compact code with an enhanced understanding of your instructions
41
  and follows what you tell them to the letter.
@@ -69,7 +69,7 @@ NOTES:
69
  - Each config/version will be very different from each other.
70
  - You can select 1 or 2 experts, default is 2 experts.
71
  - Due to unique setup of this moe, suggest 1-4 generations.
72
- - Total model size is over 22B because Qwen 2.5 MOES have a shared expert in addition to "regular experts"
73
  - Tool Calling is supported in both versions.
74
  - Source(s) / full quanting to follow // full repos to follow.
75
  - Final model size (including layers/tensors) / config subject to change.
 
23
  - qwen-coder
24
  ---
25
 
26
+ <h2>Qwen2.5-2X11B-CODER-Dueling-Wolverines-28B-gguf</h2>
27
 
28
  <img src="duel-wolverine-coder.jpg" style="float:right; width:300px; height:500px; padding:10px;">
29
 
 
35
 
36
  This is MOE model, using V1 and V2 of Wolverine-Coder 11B which is a merge of two models noted below.
37
 
38
+ The MOE config gives you full access to both 11B models at full power - full 22B parameters (an additional shared expert brings this to 28B).
39
 
40
  This MOE model generates stronger, more compact code with an enhanced understanding of your instructions
41
  and follows what you tell them to the letter.
 
69
  - Each config/version will be very different from each other.
70
  - You can select 1 or 2 experts, default is 2 experts.
71
  - Due to unique setup of this moe, suggest 1-4 generations.
72
+ - Total model size is 28B because Qwen 2.5 MOES have a shared expert in addition to "regular experts"
73
  - Tool Calling is supported in both versions.
74
  - Source(s) / full quanting to follow // full repos to follow.
75
  - Final model size (including layers/tensors) / config subject to change.