DavidAU commited on
Commit
57ca176
·
verified ·
1 Parent(s): f3820c9

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +85 -0
README.md ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model:
6
+ - Qwen/Qwen2.5-Coder-7B
7
+ - open-r1/OlympicCoder-7B
8
+ pipeline_tag: text-generation
9
+ tags:
10
+ - merge
11
+ - programming
12
+ - code generation
13
+ - code
14
+ - qwen2
15
+ - codeqwen
16
+ - chat
17
+ - qwen
18
+ - qwen-coder
19
+ ---
20
+
21
+ <h2>Qwen2.5-2X11B-CODER-Dueling-Wolverines-25B-gguf</h2>
22
+
23
+ <img src="duel-wolverine-coder.jpg" style="float:right; width:300px; height:500px; padding:10px;">
24
+
25
+ "Ripping your programming worries to shreds... fast."
26
+
27
+ Tipping the scales at 42 layers and 507 tensors... the monster lives.
28
+
29
+ Two monsters in fact - in one.
30
+
31
+ This is MOE model, using V1 and V2 of Wolverine-Coder 11B which is a merge of two models noted below.
32
+
33
+ The MOE config gives you full access to both 11B models at full power.
34
+
35
+ This MOE model generates stronger, more compact code with an enhanced understanding of your instructions
36
+ and follows what you tell them to the letter.
37
+
38
+ Each 11B version is an overpowered - yet wickedly fast - CODING ENGINE are based on two of the best coder AIs:
39
+
40
+ "Qwen2.5-Coder-7B-Instruct"
41
+
42
+ [ https://huggingface.co/Qwen/Qwen2.5-Coder-7B-Instruct ]
43
+
44
+ and
45
+
46
+ "OlympicCoder-7B"
47
+
48
+ [ https://huggingface.co/open-r1/OlympicCoder-7B ]
49
+
50
+ These two models are stuffed into one compact powerhouse 11BX2 merge that is stronger in performance and understanding than both donor models.
51
+
52
+ There are TWO versions of this MOE model.
53
+
54
+ Quants Q3_K_M and Q6_K are available at the moment, of each version.
55
+
56
+ These are generated from bfloat16 source files.
57
+
58
+ Final models will be generated from float32 source files, to improve performance of the MOE model further.
59
+
60
+ NOTES:
61
+ - Each config/version will be very different from each other.
62
+ - You can select 1 or 2 experts, default is 2 experts.
63
+ - Due to unique setup of this moe, suggest 1-4 generations.
64
+ - Tool Calling is supported in both versions.
65
+ - Source(s) / full quanting to follow // full repos to follow.
66
+ - Final model size (including layers/tensors) / config subject to change.
67
+
68
+ ---
69
+
70
+ Config / Settings
71
+
72
+ ---
73
+
74
+ Model is set at 32k/32768 context for these GGUFS, full quants/full repos will be 128k/131072.
75
+
76
+ Requirements [Qwen 2.5 7B Coder default settings]:
77
+ - Temp .5 to .7 (or lower)
78
+ - topk: 20, topp: .8, minp: .05
79
+ - rep pen: 1.1 (can be lower)
80
+ - Jinja Template (embedded) or CHATML template.
81
+ - A System Prompt is not required. (ran tests with blank system prompt)
82
+
83
+ Refer to either "Qwen2.5-Coder-7B-Instruct" and/or "OlympicCoder-7B" repos (above) for additional settings, benchmarks and usage.
84
+
85
+ ---