ArtusDev commited on
Commit
b1779f7
·
verified ·
1 Parent(s): 5095a32

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,18 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ RL_scale.jpeg filter=lfs diff=lfs merge=lfs -text
37
+ close_performance.png filter=lfs diff=lfs merge=lfs -text
38
+ main_logo.png filter=lfs diff=lfs merge=lfs -text
39
+ mid_train_scale.jpeg filter=lfs diff=lfs merge=lfs -text
40
+ open_performance.png filter=lfs diff=lfs merge=lfs -text
41
+ open_performance_white.png filter=lfs diff=lfs merge=lfs -text
42
+ test_time_scale.jpeg filter=lfs diff=lfs merge=lfs -text
43
+ assets/test_time_scale.jpeg filter=lfs diff=lfs merge=lfs -text
44
+ assets/close_performance.png filter=lfs diff=lfs merge=lfs -text
45
+ assets/mid_train_scale.jpeg filter=lfs diff=lfs merge=lfs -text
46
+ assets/open_performance_white.png filter=lfs diff=lfs merge=lfs -text
47
+ assets/main_logo.png filter=lfs diff=lfs merge=lfs -text
48
+ assets/RL_scale.jpeg filter=lfs diff=lfs merge=lfs -text
49
+ assets/open_performance.png filter=lfs diff=lfs merge=lfs -text
50
+ *.jsonl filter=lfs diff=lfs merge=lfs -text
LICENSE.md ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ - Kimi-Dev-72B is built with Qwen-2.5-72B. Qwen-2.5-72B is licensed under the [Qwen LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen2.5-72B/blob/main/LICENSE), Copyright (c) Alibaba Cloud. All Rights Reserved.
2
+ - Subject to the Qwen LICENSE AGREEMENT, Kimi-Dev-72B is under MIT license.
README.md CHANGED
@@ -1,43 +1,102 @@
1
  ---
2
- base_model: moonshotai/Kimi-Dev-72B
3
- base_model_relation: quantized
4
- quantized_by: ArtusDev
5
  license: mit
 
 
6
  tags:
7
  - code
8
  - swebench
9
  - software
10
  - issue-resolving
11
- - exl3
12
  library_name: transformers
13
  ---
 
14
 
15
- ## EXL3 Quants of moonshotai/Kimi-Dev-72B
 
 
 
 
 
 
16
 
17
- EXL3 quants of [moonshotai/Kimi-Dev-72B](https://huggingface.co/moonshotai/Kimi-Dev-72B) using <a href="https://github.com/turboderp-org/exllamav3/">exllamav3</a> for quantization.
 
 
 
 
 
 
 
 
18
 
19
- ### Quants
20
- | Quant(Revision) | Bits per Weight | Head Bits |
21
- | -------- | ---------- | --------- |
22
- | [3.5_H6](https://huggingface.co/ArtusDev/moonshotai_Kimi-Dev-72B-EXL3/tree/3.5bpw_H6) | 3.5 | 6 |
23
- | [5.0_H6](https://huggingface.co/ArtusDev/moonshotai_Kimi-Dev-72B-EXL3/tree/5.0bpw_H6) | 5.0 | 6 |
24
- | [8.0_H6](https://huggingface.co/ArtusDev/moonshotai_Kimi-Dev-72B-EXL3/tree/8.0bpw_H6) | 8.0 | 6 |
25
- | [8.0_H8](https://huggingface.co/ArtusDev/moonshotai_Kimi-Dev-72B-EXL3/tree/8.0bpw_H8) | 8.0 | 8 |
26
 
27
- ### Downloading quants with huggingface-cli
28
 
29
- <details>
30
- <summary>Click to view download instructions</summary>
31
 
32
- Install hugginface-cli:
33
 
34
- ```bash
35
- pip install -U "huggingface_hub[cli]"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
36
  ```
 
 
 
 
 
 
 
 
 
 
37
 
38
- Download quant by targeting the specific quant revision (branch):
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
39
 
40
  ```
41
- huggingface-cli download ArtusDev/moonshotai_Kimi-Dev-72B-EXL3 --revision "5.0bpw_H6" --local-dir ./
 
42
  ```
43
- </details>
 
 
 
 
 
 
 
 
1
  ---
 
 
 
2
  license: mit
3
+ base_model:
4
+ - Qwen/Qwen2.5-72B
5
  tags:
6
  - code
7
  - swebench
8
  - software
9
  - issue-resolving
 
10
  library_name: transformers
11
  ---
12
+ <!-- # Kimi-Dev -->
13
 
14
+ <div align="center">
15
+ <img src="./assets/main_logo.png" alt="Kimi Logo" width="400" />
16
+ <h2><a href="https://moonshotai.github.io/Kimi-Dev/">
17
+ Introducing Kimi-Dev: <br>A Strong and Open-source Coding LLM for Issue Resolution</a></h2>
18
+ </a></h2>
19
+ <b>Kimi-Dev Team</b>
20
+ <br>
21
 
22
+ </div>
23
+ <div align="center">
24
+ <a href="">
25
+ <b>📄 Tech Report (Coming soon...)</b>
26
+ </a> &nbsp;|&nbsp;
27
+ <a href="https://github.com/MoonshotAI/Kimi-Dev">
28
+ <b>📄 Github</b>
29
+ </a> &nbsp;
30
+ </div>
31
 
32
+ <br>
33
+ <br>
 
 
 
 
 
34
 
35
+ <!-- https://github.com/MoonshotAI/Kimi-Dev -->
36
 
37
+ We introduce Kimi-Dev-72B, our new open-source coding LLM for software engineering tasks. Kimi-Dev-72B achieves a new state-of-the-art on SWE-bench Verified among open-source models.
 
38
 
39
+ - Kimi-Dev-72B achieves 60.4% performance on SWE-bench Verified. It surpasses the runner-up, setting a new state-of-the-art result among open-source models.
40
 
41
+
42
+ - Kimi-Dev-72B is optimized via large-scale reinforcement learning. It autonomously patches real repositories in Docker and gains rewards only when the entire test suite passes. This ensures correct and robust solutions, aligning with real-world development standards.
43
+
44
+
45
+ - Kimi-Dev-72B is available for download and deployment on Hugging Face and GitHub. We welcome developers and researchers to explore its capabilities and contribute to development.
46
+
47
+
48
+ <div align="center">
49
+ <img src="./assets/open_performance_white.png" alt="Kimi Logo" width="600" />
50
+ <p><b>Performance of Open-source Models on SWE-bench Verified.</b></p>
51
+
52
+ </div>
53
+
54
+
55
+
56
+ ## Quick Start
57
  ```
58
+ from transformers import AutoModelForCausalLM, AutoTokenizer
59
+
60
+ model_name = "moonshotai/Kimi-Dev-72B"
61
+
62
+ model = AutoModelForCausalLM.from_pretrained(
63
+ model_name,
64
+ torch_dtype="auto",
65
+ device_map="auto"
66
+ )
67
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
68
 
69
+ prompt = "Give me a short introduction to large language model."
70
+ messages = [
71
+ {"role": "system", "content": "You are a helpful assistant."},
72
+ {"role": "user", "content": prompt}
73
+ ]
74
+ text = tokenizer.apply_chat_template(
75
+ messages,
76
+ tokenize=False,
77
+ add_generation_prompt=True
78
+ )
79
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
80
+
81
+ generated_ids = model.generate(
82
+ **model_inputs,
83
+ max_new_tokens=512
84
+ )
85
+ generated_ids = [
86
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
87
+ ]
88
+
89
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
90
 
91
  ```
92
+
93
+ ## Citation
94
  ```
95
+ @misc{kimi_dev_72b_2025,
96
+ title = {Introducing Kimi-Dev: A Strong and Open-source Coding LLM for Issue Resolution},
97
+ author = {{Kimi-Dev Team}},
98
+ year = {2025},
99
+ month = {June},
100
+ url = {\url{https://www.moonshot.cn/Kimi-Dev}}
101
+ }
102
+ ```
config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 8192,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 29568,
12
+ "max_position_embeddings": 131072,
13
+ "max_window_layers": 70,
14
+ "model_type": "qwen2",
15
+ "num_attention_heads": 64,
16
+ "num_hidden_layers": 80,
17
+ "num_key_value_heads": 8,
18
+ "rms_norm_eps": 1e-06,
19
+ "rope_theta": 1000000.0,
20
+ "sliding_window": 131072,
21
+ "tie_word_embeddings": false,
22
+ "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.43.1",
24
+ "use_cache": true,
25
+ "use_sliding_window": false,
26
+ "vocab_size": 152064,
27
+ "quantization_config": {
28
+ "quant_method": "exl3",
29
+ "version": "0.0.4",
30
+ "bits": 2.5,
31
+ "head_bits": 6,
32
+ "calibration": {
33
+ "rows": 100,
34
+ "cols": 2048
35
+ },
36
+ "out_scales": "auto"
37
+ }
38
+ }
model-00001-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5598248ff366ddee645c65a9d51733e3e86491157afdffeee20b8ca7cf73324
3
+ size 8530118104
model-00002-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0eb5ef6b5d91700b931b205218f383231a2f8a7b007d93a9ee9f317a5f943ecb
3
+ size 8510249848
model-00003-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:063b393d4ed6af5c340f24d509a2d1a0bc2d6575294a5a353fc24313fa44f537
3
+ size 8349270328
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
quantization_config.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "model_max_length": 131072,
203
+ "pad_token": "<|endoftext|>",
204
+ "split_special_tokens": false,
205
+ "tokenizer_class": "Qwen2Tokenizer",
206
+ "unk_token": null
207
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff