Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,85 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
base_model:
|
6 |
+
- Qwen/Qwen2.5-Coder-7B
|
7 |
+
- open-r1/OlympicCoder-7B
|
8 |
+
pipeline_tag: text-generation
|
9 |
+
tags:
|
10 |
+
- merge
|
11 |
+
- programming
|
12 |
+
- code generation
|
13 |
+
- code
|
14 |
+
- qwen2
|
15 |
+
- codeqwen
|
16 |
+
- chat
|
17 |
+
- qwen
|
18 |
+
- qwen-coder
|
19 |
+
---
|
20 |
+
|
21 |
+
<h2>Qwen2.5-2X11B-CODER-Dueling-Wolverines-25B-gguf</h2>
|
22 |
+
|
23 |
+
<img src="duel-wolverine-coder.jpg" style="float:right; width:300px; height:500px; padding:10px;">
|
24 |
+
|
25 |
+
"Ripping your programming worries to shreds... fast."
|
26 |
+
|
27 |
+
Tipping the scales at 42 layers and 507 tensors... the monster lives.
|
28 |
+
|
29 |
+
Two monsters in fact - in one.
|
30 |
+
|
31 |
+
This is MOE model, using V1 and V2 of Wolverine-Coder 11B which is a merge of two models noted below.
|
32 |
+
|
33 |
+
The MOE config gives you full access to both 11B models at full power.
|
34 |
+
|
35 |
+
This MOE model generates stronger, more compact code with an enhanced understanding of your instructions
|
36 |
+
and follows what you tell them to the letter.
|
37 |
+
|
38 |
+
Each 11B version is an overpowered - yet wickedly fast - CODING ENGINE are based on two of the best coder AIs:
|
39 |
+
|
40 |
+
"Qwen2.5-Coder-7B-Instruct"
|
41 |
+
|
42 |
+
[ https://huggingface.co/Qwen/Qwen2.5-Coder-7B-Instruct ]
|
43 |
+
|
44 |
+
and
|
45 |
+
|
46 |
+
"OlympicCoder-7B"
|
47 |
+
|
48 |
+
[ https://huggingface.co/open-r1/OlympicCoder-7B ]
|
49 |
+
|
50 |
+
These two models are stuffed into one compact powerhouse 11BX2 merge that is stronger in performance and understanding than both donor models.
|
51 |
+
|
52 |
+
There are TWO versions of this MOE model.
|
53 |
+
|
54 |
+
Quants Q3_K_M and Q6_K are available at the moment, of each version.
|
55 |
+
|
56 |
+
These are generated from bfloat16 source files.
|
57 |
+
|
58 |
+
Final models will be generated from float32 source files, to improve performance of the MOE model further.
|
59 |
+
|
60 |
+
NOTES:
|
61 |
+
- Each config/version will be very different from each other.
|
62 |
+
- You can select 1 or 2 experts, default is 2 experts.
|
63 |
+
- Due to unique setup of this moe, suggest 1-4 generations.
|
64 |
+
- Tool Calling is supported in both versions.
|
65 |
+
- Source(s) / full quanting to follow // full repos to follow.
|
66 |
+
- Final model size (including layers/tensors) / config subject to change.
|
67 |
+
|
68 |
+
---
|
69 |
+
|
70 |
+
Config / Settings
|
71 |
+
|
72 |
+
---
|
73 |
+
|
74 |
+
Model is set at 32k/32768 context for these GGUFS, full quants/full repos will be 128k/131072.
|
75 |
+
|
76 |
+
Requirements [Qwen 2.5 7B Coder default settings]:
|
77 |
+
- Temp .5 to .7 (or lower)
|
78 |
+
- topk: 20, topp: .8, minp: .05
|
79 |
+
- rep pen: 1.1 (can be lower)
|
80 |
+
- Jinja Template (embedded) or CHATML template.
|
81 |
+
- A System Prompt is not required. (ran tests with blank system prompt)
|
82 |
+
|
83 |
+
Refer to either "Qwen2.5-Coder-7B-Instruct" and/or "OlympicCoder-7B" repos (above) for additional settings, benchmarks and usage.
|
84 |
+
|
85 |
+
---
|