cgus commited on
Commit
87e8519
·
verified ·
1 Parent(s): 5d03056

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -0
README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - YOYO-AI/YOYO-O1-14B
4
+ tags:
5
+ - mergekit
6
+ - merge
7
+ license: apache-2.0
8
+ language:
9
+ - en
10
+ - zh
11
+ pipeline_tag: text-generation
12
+ library_name: exllamav2
13
+ ---
14
+ # YOYO-O1-14B-exl2
15
+ Original model: [YOYO-O1-14B](https://huggingface.co/YOYO-AI/YOYO-O1-14B) by [YOYO-AI](https://huggingface.co/YOYO-AI)
16
+ Foundation model: [Qwen2.5-14B](https://huggingface.co/Qwen/Qwen2.5-14B) by [Qwen](https://huggingface.co/Qwen)
17
+ ## Quants
18
+ [4bpw h6 (main)](https://huggingface.co/cgus/YOYO-O1-14B-exl2/tree/main)
19
+ [4.5bpw h6](https://huggingface.co/cgus/YOYO-O1-14B-exl2/tree/4.5bpw-h6)
20
+ [5bpw h6](https://huggingface.co/cgus/YOYO-O1-14B-exl2/tree/5bpw-h6)
21
+ [6bpw h6](https://huggingface.co/cgus/YOYO-O1-14B-exl2/tree/6bpw-h6)
22
+ [8bpw h8](https://huggingface.co/cgus/YOYO-O1-14B-exl2/tree/8bpw-h8)
23
+ ## Quantization notes
24
+ Made with Exllamav2 0.2.8 with default dataset.
25
+ These quants can be used with TabbyAPI or Text-Generation-WebUI with RTX GPU (Windows) or RTX/ROCm (Linux).
26
+ Quants have to fit your VRAM, if you need RAM offloading then choose GGUF quants instead.
27
+ # Original model card
28
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/64e174e202fa032de4143324/GNPQMcj3XRXlXy3ebsPRZ.jpeg)
29
+
30
+ # YOYO-O1-14B
31
+
32
+ *Combined the most top-notch 14B **inference** model and **code** model in the entire open-source community.*
33
+
34
+ ### Merge Method
35
+
36
+ This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method using [Qwen/Qwen2.5-Coder-14B](https://huggingface.co/Qwen/Qwen2.5-Coder-14B) as a base.
37
+
38
+ ### Models Merged
39
+
40
+ The following models were included in the merge:
41
+ * [Qwen/Qwen2.5-Coder-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-14B-Instruct)
42
+ * [Gen-Verse/ReasonFlux-F1-14B](https://huggingface.co/Gen-Verse/ReasonFlux-F1-14B)
43
+ * [agentica-org/DeepCoder-14B-Preview](https://huggingface.co/agentica-org/DeepCoder-14B-Preview)
44
+ * [qihoo360/Light-R1-14B-DS](https://huggingface.co/qihoo360/Light-R1-14B-DS)
45
+
46
+ ### Configuration
47
+
48
+ The following YAML configuration was used to produce this model:
49
+
50
+ ```yaml
51
+ merge_method: sce
52
+ models:
53
+ # Pivot model
54
+ - model: Qwen/Qwen2.5-Coder-14B
55
+ # Target models
56
+ - model: agentica-org/DeepCoder-14B-Preview
57
+ - model: qihoo360/Light-R1-14B-DS
58
+ - model: Gen-Verse/ReasonFlux-F1-14B
59
+ - model: Qwen/Qwen2.5-Coder-14B-Instruct
60
+ base_model: Qwen/Qwen2.5-Coder-14B
61
+ parameters:
62
+ select_topk: 1
63
+ dtype: float16
64
+ tokenizer_source: qihoo360/Light-R1-14B-DS
65
+ normalize: true
66
+ int8_mask: true
67
+ ```