File size: 1,575 Bytes
8283516
 
5f39fcc
9e4a725
 
 
8283516
 
 
 
5f39fcc
 
 
 
 
8283516
5f39fcc
7fbf1bc
 
1e1d081
8283516
 
 
5f39fcc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8283516
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5f39fcc
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
---
base_model:
- Qwen/Qwen2.5-32B
- Qwen/Qwen2.5-32B-Instruct
- Qwen/Qwen2.5-Coder-32B
- Qwen/Qwen2.5-Coder-32B-Instruct
library_name: transformers
tags:
- mergekit
- merge
license: apache-2.0
language:
- en
- zh
pipeline_tag: text-generation
---
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/64e174e202fa032de4143324/9sSGARM4_J0ZUbm8jSQkD.jpeg)
*This series aims to unify the **official models** of **Qwen**.*

*The **unified model** obtained by merging the **code** model and the **instruction** model through the **SCE** method*
### Configuration

The following YAML configuration was used to produce this model:
```yaml
models:
  - model: Qwen/Qwen2.5-32B-instruct
    parameters:
      density: 1 
      weight: 1
      lambda: 0.9
merge_method: della
base_model: Qwen/Qwen2.5-32B
parameters:
  density: 1
  weight: 1
  lambda: 0.9
  normalize: true
  int8_mask: true
dtype: bfloat16
name: Qwen2.5-32B-YOYO
```
```yaml
models:
  - model: Qwen/Qwen2.5-Coder-32B-instruct
    parameters:
      density: 1 
      weight: 1
      lambda: 0.9
merge_method: della
base_model: Qwen/Qwen2.5-Coder-32B
parameters:
  density: 1
  weight: 1
  lambda: 0.9
  normalize: true
  int8_mask: true
dtype: bfloat16
name: Qwen2.5-Coder-32B-YOYO
```
```yaml
merge_method: sce
models:
  # Pivot model
  - model: Qwen/Qwen2.5-Coder-32B
  # Target models
  - model: YOYO-AI/Qwen2.5-32B-YOYO
  - model: YOYO-AI/Qwen2.5-Coder-32B-YOYO
base_model: Qwen/Qwen2.5-Coder-32B
parameters:
  select_topk: 1
dtype: bfloat16
tokenizer_source: base
normalize: true
int8_mask: true
```