DavidAU commited on
Commit
a46d60c
·
verified ·
0 Parent(s):

Duplicate from DavidAU/Qwen2.5-Microsoft-NextCoder-Brainstorm20x-20B

Browse files
.gitattributes ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,287 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model:
4
+ - microsoft/NextCoder-14B
5
+ language:
6
+ - en
7
+ pipeline_tag: text-generation
8
+ tags:
9
+ - merge
10
+ - programming
11
+ - code generation
12
+ - code
13
+ - codeqwen
14
+ - coding
15
+ - coder
16
+ - qwen2
17
+ - chat
18
+ - qwen
19
+ - qwen-coder
20
+ - code
21
+ - chat
22
+ - brainstorm
23
+ - brainstorm20x
24
+ - microsoft
25
+ - nextcoder
26
+ - selekt
27
+ datasets:
28
+ - microsoft/NextCoderDataset
29
+ - microsoft/NextCoderDataset-Conversational
30
+ - bigcode/commitpackft
31
+ - bigcode/starcoderdata
32
+ library_name: transformers
33
+ ---
34
+
35
+ <h2>Qwen2.5-Microsoft-NextCoder-Brainstorm20x-20B</h2>
36
+
37
+ This repo contains the full precision source code, in "safe tensors" format to generate GGUFs, GPTQ, EXL2, AWQ, HQQ and other formats.
38
+ The source code can also be used directly.
39
+
40
+ This model contains Brainstorm 20x, combined with Microsoft's 14B Coder (instruct model):
41
+
42
+ https://huggingface.co/microsoft/NextCoder-14B
43
+
44
+ Information on the 14B Microsoft NextCoder model below, followed by Brainstorm 20x adapter (by DavidAU) and then a complete help
45
+ section for running LLM / AI models.
46
+
47
+ The Brainstorm adapter improves code generation, and unique code solving abilities.
48
+
49
+ This model requires:
50
+ - Jinja (embedded) or CHATML template
51
+ - Max context of 32k as per Microsoft specs.
52
+
53
+ Settings used for testing (suggested):
54
+ - Temp .3 to .7
55
+ - Rep pen 1.05 to 1.1
56
+ - Topp .8 , minp .05
57
+ - Topk 20
58
+ - No system prompt.
59
+
60
+ This model will respond well to both detailed instructions and step by step refinement and additions to code.
61
+
62
+ As this is an instruct model, it will also benefit from a detailed system prompt too.
63
+
64
+ For simpler coding problems, lower quants will work well; but for complex/multi-step problem solving suggest Q6 or Q8.
65
+
66
+ ---
67
+
68
+ # NextCoder-14B
69
+ <p align="center">
70
+ <a href="https://github.com/microsoft/NextCoder">GitHub</a>&nbsp&nbsp | &nbsp&nbsp <a href="https://www.microsoft.com/en-us/research/publication/nextcoder-robust-adaptation-of-code-lms-to-diverse-code-edits/">Paper</a>
71
+ </p>
72
+
73
+ > NextCoder: Robust Adaptation of Code LMs to Diverse Code Edits (ICML'2025)
74
+
75
+ ## Introduction
76
+
77
+ NextCoder is the latest series of Code-Editing large language models developed using the Qwen2.5-Coder Instruct variants as base and trained with novel Selective Knowledge Transfer finetuning methodology as introduced in the paper. NextCoder family model comes in 3 different sizes 7, 14, 32 billion parameters, to meet the needs of different developers.
78
+ Following are the key improvements:
79
+ - Significantly improvements in **code editing**, NextCoder-32B has performing on par with GPT-4o on complex benchmarks like Aider-Polyglot with performance increment of 44% from their base model.
80
+ - No loss of generalizibility, due to our new finetuning method **SeleKT**
81
+ - **Long-context Support** up to 32K tokens.
82
+
83
+ **This repo contains the NextCoder 14B model**, which has the following features:
84
+ - Type: Causal Language Models
85
+ - Training Stage: Post-training with SeleKT
86
+ - Architecture: transformers with RoPE, SwiGLU, RMSNorm, and Attention QKV bias
87
+ - Number of Parameters: 14.7B
88
+ - Number of Paramaters (Non-Embedding): 13.1B
89
+ - Number of Layers: 48
90
+ - Number of Attention Heads (GQA): 40 for Q and 8 for KV
91
+
92
+ For more details, please refer to our [blog](), [GitHub](https://github.com/microsoft/NextCoder), [Paper](https://www.microsoft.com/en-us/research/publication/nextcoder-robust-adaptation-of-code-lms-to-diverse-code-edits/).
93
+
94
+ ## Requirements
95
+
96
+ The code of NextCoder is based on Qwen2.5 base models which has been in the latest Hugging face `transformers` and we advise you to use the latest version of `transformers`.
97
+
98
+ With `transformers<4.37.0`, you will encounter the following error:
99
+ ```
100
+ KeyError: 'qwen2'
101
+ ```
102
+
103
+ ## Quickstart
104
+
105
+ Here provides a code snippet with `apply_chat_template` to show you how to load the tokenizer and model and how to generate contents.
106
+
107
+ ```python
108
+ from transformers import AutoModelForCausalLM, AutoTokenizer
109
+
110
+ model_name = "microsoft/NextCoder-14B"
111
+
112
+ model = AutoModelForCausalLM.from_pretrained(
113
+ model_name,
114
+ torch_dtype="auto",
115
+ device_map="auto",
116
+ )
117
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
118
+
119
+ prompt = """
120
+ Fix the following function that divides two numbers to handle all the edge cases:
121
+
122
+ def divide(a, b)
123
+ returm a/b
124
+ """
125
+ messages = [
126
+ {"role": "user", "content": prompt}
127
+ ]
128
+ text = tokenizer.apply_chat_template(
129
+ messages,
130
+ tokenize=False,
131
+ add_generation_prompt=True
132
+ )
133
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
134
+
135
+ generated_ids = model.generate(
136
+ **model_inputs,
137
+ max_new_tokens=1024
138
+ )
139
+ generated_ids = [
140
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
141
+ ]
142
+
143
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
144
+ ```
145
+ ## Evaluation and Performance
146
+
147
+ | Models | HUMANEVALFIX | CANITEDIT | AIDER | POLYGLOT |
148
+ |--------|---------------|-----------|-------|----------|
149
+ | QwenCoder-2.5-3B | 73.2 | 37.1 | 36.8 | - |
150
+ | QwenCoder-2.5-3B-LoRA | 64.6 | 36.2 | 35.8 | - |
151
+ | QwenCoder-2.5-3B-SFT | 76.2 | 32.4 | 30.1 | - |
152
+ | **NextCoder-3B** | 75.6 | 42.4 | 37.6 | - |
153
+ | QwenCoder-2.5-7B | 73.8 | 48.1 | 59.4 | - |
154
+ | QwenCoder-2.5-7B-LoRA | 70.7 | 44.3 | 40.6 | - |
155
+ | QwenCoder-2.5-7B-SFT | 70.1 | 36.7 | 48.9 | - |
156
+ | **NextCoder-7B** | 81.1 | 50.5 | 65.7 | - |
157
+ | QwenCoder-2.5-14B | 87.8 | 58.1 | 66.9 | 9.3 |
158
+ | QwenCoder-2.5-14B-LoRA | 78.0 | 50.9 | 66.2 | 5.3 |
159
+ | QwenCoder-2.5-14B-SFT | 79.9 | 42.4 | 36.8 | 3.1 |
160
+ | **NextCoder-14B** | 89.8 | 60.2 | 72.2 | 12.2 |
161
+ | QwenCoder-2.5-32B | **90.2** | 61.0 | 72.9 | 16.4 |
162
+ | QwenCoder-2.5-32B-LoRA | 82.3 | 52.4 | 60.2 | 6.7 |
163
+ | QwenCoder-2.5-32B-SFT | 81.7 | 49.5 | 66.9 | 8.4 |
164
+ | **NextCoder-32B** | 88.9 | **62.4** | **74.7** | **23.6** |
165
+
166
+ *Comparison of base QwenCoder-2.5 models of different sizes and their SELEKT-enhanced versions across three code editing benchmarks.*
167
+
168
+ **Detailed evaluation results are reported in this [📑 paper](https://www.microsoft.com/en-us/research/publication/nextcoder-robust-adaptation-of-code-lms-to-diverse-code-edits/).**
169
+
170
+ See more here:
171
+
172
+ https://huggingface.co/microsoft/NextCoder-14B
173
+
174
+ ---
175
+
176
+ <H2>What is Brainstorm?</H2>
177
+
178
+ ---
179
+
180
+ <B>Brainstorm 20x</B>
181
+
182
+ The BRAINSTORM process was developed by David_AU.
183
+
184
+ Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
185
+ scientific paper : Progressive LLaMA with Block Expansion </a>.
186
+
187
+ However I went in a completely different direction from what was outlined in this paper.
188
+
189
+ What is "Brainstorm" ?
190
+
191
+ The reasoning center of an LLM is taken apart, reassembled, and expanded.
192
+
193
+ In this case for this model: 20 times
194
+
195
+ Then these centers are individually calibrated. These "centers" also interact with each other.
196
+ This introduces subtle changes into the reasoning process.
197
+ The calibrations further adjust - dial up or down - these "changes" further.
198
+ The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
199
+
200
+ The core aim of this process is to increase the model's detail, concept and connection to the "world",
201
+ general concept connections, prose quality and prose length without affecting instruction following.
202
+
203
+ This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
204
+
205
+ Here are some of the enhancements this process brings to the model's performance:
206
+
207
+ - Prose generation seems more focused on the moment to moment.
208
+ - Sometimes there will be "preamble" and/or foreshadowing present.
209
+ - Fewer or no "cliches"
210
+ - Better overall prose and/or more complex / nuanced prose.
211
+ - A greater sense of nuance on all levels.
212
+ - Coherence is stronger.
213
+ - Description is more detailed, and connected closer to the content.
214
+ - Simile and Metaphors are stronger and better connected to the prose, story, and character.
215
+ - Sense of "there" / in the moment is enhanced.
216
+ - Details are more vivid, and there are more of them.
217
+ - Prose generation length can be long to extreme.
218
+ - Emotional engagement is stronger.
219
+ - The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
220
+ - The MORE instructions and/or details you provide the more strongly the model will respond.
221
+ - Depending on the model "voice" may be more "human" vs original model's "voice".
222
+
223
+ Other "lab" observations:
224
+
225
+ - This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
226
+ - However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
227
+ - From lab testing it seems to ponder, and consider more carefully roughly speaking.
228
+ - You could say this process sharpens the model's focus on it's task(s) at a deeper level.
229
+
230
+ The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
231
+
232
+ ---
233
+
234
+ For more information / other Qwen/Mistral Coders / additional settings see:
235
+
236
+ [ https://huggingface.co/DavidAU/Qwen2.5-MOE-2x-4x-6x-8x__7B__Power-CODER__19B-30B-42B-53B-gguf ]
237
+
238
+ ---
239
+
240
+ <H2>Help, Adjustments, Samplers, Parameters and More</H2>
241
+
242
+ ---
243
+
244
+ <B>CHANGE THE NUMBER OF ACTIVE EXPERTS:</B>
245
+
246
+ See this document:
247
+
248
+ https://huggingface.co/DavidAU/How-To-Set-and-Manage-MOE-Mix-of-Experts-Model-Activation-of-Experts
249
+
250
+ <B>Settings: CHAT / ROLEPLAY and/or SMOOTHER operation of this model:</B>
251
+
252
+ In "KoboldCpp" or "oobabooga/text-generation-webui" or "Silly Tavern" ;
253
+
254
+ Set the "Smoothing_factor" to 1.5
255
+
256
+ : in KoboldCpp -> Settings->Samplers->Advanced-> "Smooth_F"
257
+
258
+ : in text-generation-webui -> parameters -> lower right.
259
+
260
+ : In Silly Tavern this is called: "Smoothing"
261
+
262
+
263
+ NOTE: For "text-generation-webui"
264
+
265
+ -> if using GGUFs you need to use "llama_HF" (which involves downloading some config files from the SOURCE version of this model)
266
+
267
+ Source versions (and config files) of my models are here:
268
+
269
+ https://huggingface.co/collections/DavidAU/d-au-source-files-for-gguf-exl2-awq-gptq-hqq-etc-etc-66b55cb8ba25f914cbf210be
270
+
271
+ OTHER OPTIONS:
272
+
273
+ - Increase rep pen to 1.1 to 1.15 (you don't need to do this if you use "smoothing_factor")
274
+
275
+ - If the interface/program you are using to run AI MODELS supports "Quadratic Sampling" ("smoothing") just make the adjustment as noted.
276
+
277
+ <B>Highest Quality Settings / Optimal Operation Guide / Parameters and Samplers</B>
278
+
279
+ This a "Class 1" model:
280
+
281
+ For all settings used for this model (including specifics for its "class"), including example generation(s) and for advanced settings guide (which many times addresses any model issue(s)), including methods to improve model performance for all use case(s) as well as chat, roleplay and other use case(s) please see:
282
+
283
+ [ https://huggingface.co/DavidAU/Maximizing-Model-Performance-All-Quants-Types-And-Full-Precision-by-Samplers_Parameters ]
284
+
285
+ You can see all parameters used for generation, in addition to advanced parameters and samplers to get the most out of this model here:
286
+
287
+ [ https://huggingface.co/DavidAU/Maximizing-Model-Performance-All-Quants-Types-And-Full-Precision-by-Samplers_Parameters ]
added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151643,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 5120,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 13824,
12
+ "max_position_embeddings": 32768,
13
+ "max_window_layers": 48,
14
+ "model_type": "qwen2",
15
+ "num_attention_heads": 40,
16
+ "num_hidden_layers": 67,
17
+ "num_key_value_heads": 8,
18
+ "rms_norm_eps": 1e-06,
19
+ "rope_scaling": null,
20
+ "rope_theta": 1000000.0,
21
+ "sliding_window": null,
22
+ "tie_word_embeddings": false,
23
+ "torch_dtype": "bfloat16",
24
+ "transformers_version": "4.52.0.dev0",
25
+ "use_cache": false,
26
+ "use_sliding_window": false,
27
+ "vocab_size": 152064
28
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model-00001-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:867ecd494199382191fb4fc8da11c4e54a701d02e74d4c543be12aa8f37c7bc8
3
+ size 4907454960
model-00002-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4949c420ed0df9afa1e10d1e99efabce505fc0e67881feb580e43c3ce71e8033
3
+ size 4954847384
model-00003-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ce64e102a687cf5fc66351b01db73ecb2ca581ecc89cffdcbcdf1033077ef9e
3
+ size 4954847376
model-00004-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:242e682ade52065c59afc432efefe366a51ce30c0ff187bfb033c8e6e0566a33
3
+ size 4954847376
model-00005-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c11c376fba7cd52159501aa180e16e825b62b8ea17b7eb73b9cf8e64c2e4c7eb
3
+ size 4954847376
model-00006-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f8bc39e2d1756fd9e7a88185099fc023a5003c98ccee635b70e69cc3dc881c1
3
+ size 4907643624
model-00007-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58706278d2a84665bf5167ab296c95624877fd71d17c192fa54d096b103fca6d
3
+ size 4954525640
model-00008-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:395900162b7f347b0ffc9d806e4eb1713d8c25e60c6645d5cb6c8ab50bdda91e
3
+ size 4860814984
model-00009-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:519725ebc514107a76d683473ddea27f5b388cfc047d3bd52bd74fd04d6de395
3
+ size 550538576
model.safetensors.index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metadata": {"mergekit_version": "0.1.2"}, "weight_map": {"lm_head.weight": "model-00001-of-00009.safetensors", "model.embed_tokens.weight": "model-00001-of-00009.safetensors", "model.layers.0.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.0.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.0.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00009.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00009.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00009.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00009.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00009.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00009.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.10.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.10.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.10.self_attn.k_proj.bias": "model-00001-of-00009.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.10.self_attn.q_proj.bias": "model-00001-of-00009.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.10.self_attn.v_proj.bias": "model-00001-of-00009.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.11.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.11.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.11.self_attn.k_proj.bias": "model-00002-of-00009.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.11.self_attn.q_proj.bias": "model-00002-of-00009.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.11.self_attn.v_proj.bias": "model-00002-of-00009.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.12.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.12.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.12.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.12.self_attn.k_proj.bias": "model-00002-of-00009.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.12.self_attn.q_proj.bias": "model-00002-of-00009.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.12.self_attn.v_proj.bias": "model-00002-of-00009.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.13.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.13.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.13.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.13.self_attn.k_proj.bias": "model-00002-of-00009.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.13.self_attn.q_proj.bias": "model-00002-of-00009.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.13.self_attn.v_proj.bias": "model-00002-of-00009.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.14.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.14.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.14.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.14.self_attn.k_proj.bias": "model-00002-of-00009.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.14.self_attn.q_proj.bias": "model-00002-of-00009.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.14.self_attn.v_proj.bias": "model-00002-of-00009.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.15.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.15.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.15.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.15.self_attn.k_proj.bias": "model-00002-of-00009.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.15.self_attn.q_proj.bias": "model-00002-of-00009.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.15.self_attn.v_proj.bias": "model-00002-of-00009.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.16.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.16.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.16.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.16.self_attn.k_proj.bias": "model-00002-of-00009.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.16.self_attn.q_proj.bias": "model-00002-of-00009.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.16.self_attn.v_proj.bias": "model-00002-of-00009.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.17.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.17.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.17.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.17.self_attn.k_proj.bias": "model-00002-of-00009.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.17.self_attn.q_proj.bias": "model-00002-of-00009.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.17.self_attn.v_proj.bias": "model-00002-of-00009.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.18.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.18.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.18.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.18.self_attn.k_proj.bias": "model-00002-of-00009.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.18.self_attn.q_proj.bias": "model-00002-of-00009.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.18.self_attn.v_proj.bias": "model-00002-of-00009.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.19.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.19.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.19.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.19.self_attn.k_proj.bias": "model-00002-of-00009.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.19.self_attn.q_proj.bias": "model-00002-of-00009.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.19.self_attn.v_proj.bias": "model-00002-of-00009.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.2.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.2.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.2.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.2.self_attn.k_proj.bias": "model-00003-of-00009.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.2.self_attn.q_proj.bias": "model-00003-of-00009.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.2.self_attn.v_proj.bias": "model-00003-of-00009.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.20.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.20.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.20.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.20.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.20.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.20.self_attn.k_proj.bias": "model-00003-of-00009.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.20.self_attn.q_proj.bias": "model-00003-of-00009.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.20.self_attn.v_proj.bias": "model-00003-of-00009.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.21.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.21.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.21.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.21.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.21.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.21.self_attn.k_proj.bias": "model-00003-of-00009.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.21.self_attn.q_proj.bias": "model-00003-of-00009.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.21.self_attn.v_proj.bias": "model-00003-of-00009.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.22.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.22.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.22.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.22.self_attn.k_proj.bias": "model-00003-of-00009.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.22.self_attn.q_proj.bias": "model-00003-of-00009.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.22.self_attn.v_proj.bias": "model-00003-of-00009.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.23.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.23.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.23.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.23.self_attn.k_proj.bias": "model-00003-of-00009.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.23.self_attn.q_proj.bias": "model-00003-of-00009.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.23.self_attn.v_proj.bias": "model-00003-of-00009.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.24.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.24.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.24.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.24.self_attn.k_proj.bias": "model-00003-of-00009.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.24.self_attn.q_proj.bias": "model-00003-of-00009.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.24.self_attn.v_proj.bias": "model-00003-of-00009.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.25.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.25.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.25.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.25.self_attn.k_proj.bias": "model-00003-of-00009.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.25.self_attn.q_proj.bias": "model-00003-of-00009.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.25.self_attn.v_proj.bias": "model-00003-of-00009.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.26.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.26.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.26.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.26.self_attn.k_proj.bias": "model-00003-of-00009.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.26.self_attn.q_proj.bias": "model-00003-of-00009.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.26.self_attn.v_proj.bias": "model-00003-of-00009.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.27.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.27.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.27.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.27.self_attn.k_proj.bias": "model-00003-of-00009.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.27.self_attn.q_proj.bias": "model-00003-of-00009.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.27.self_attn.v_proj.bias": "model-00003-of-00009.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.28.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.28.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.28.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.28.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.28.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.28.self_attn.k_proj.bias": "model-00004-of-00009.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.28.self_attn.q_proj.bias": "model-00004-of-00009.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.28.self_attn.v_proj.bias": "model-00004-of-00009.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.29.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.29.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.29.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.29.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.29.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.29.self_attn.k_proj.bias": "model-00004-of-00009.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.29.self_attn.q_proj.bias": "model-00004-of-00009.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.29.self_attn.v_proj.bias": "model-00004-of-00009.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.3.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.3.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.3.self_attn.k_proj.bias": "model-00004-of-00009.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.3.self_attn.q_proj.bias": "model-00004-of-00009.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.3.self_attn.v_proj.bias": "model-00004-of-00009.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.30.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.30.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.30.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.30.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.30.self_attn.k_proj.bias": "model-00004-of-00009.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.30.self_attn.q_proj.bias": "model-00004-of-00009.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.30.self_attn.v_proj.bias": "model-00004-of-00009.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.31.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.31.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.31.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.31.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.31.self_attn.k_proj.bias": "model-00004-of-00009.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.31.self_attn.q_proj.bias": "model-00004-of-00009.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.31.self_attn.v_proj.bias": "model-00004-of-00009.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.32.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.32.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.32.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.32.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.32.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.32.self_attn.k_proj.bias": "model-00004-of-00009.safetensors", "model.layers.32.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.32.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.32.self_attn.q_proj.bias": "model-00004-of-00009.safetensors", "model.layers.32.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.32.self_attn.v_proj.bias": "model-00004-of-00009.safetensors", "model.layers.32.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.33.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.33.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.33.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.33.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.33.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.33.self_attn.k_proj.bias": "model-00004-of-00009.safetensors", "model.layers.33.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.33.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.33.self_attn.q_proj.bias": "model-00004-of-00009.safetensors", "model.layers.33.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.33.self_attn.v_proj.bias": "model-00004-of-00009.safetensors", "model.layers.33.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.34.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.34.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.34.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.34.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.34.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.34.self_attn.k_proj.bias": "model-00004-of-00009.safetensors", "model.layers.34.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.34.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.34.self_attn.q_proj.bias": "model-00004-of-00009.safetensors", "model.layers.34.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.34.self_attn.v_proj.bias": "model-00004-of-00009.safetensors", "model.layers.34.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.35.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.35.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.35.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.35.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.35.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.35.self_attn.k_proj.bias": "model-00004-of-00009.safetensors", "model.layers.35.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.35.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.35.self_attn.q_proj.bias": "model-00004-of-00009.safetensors", "model.layers.35.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.35.self_attn.v_proj.bias": "model-00004-of-00009.safetensors", "model.layers.35.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.36.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.36.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.36.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.36.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.36.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.36.self_attn.k_proj.bias": "model-00005-of-00009.safetensors", "model.layers.36.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.36.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.36.self_attn.q_proj.bias": "model-00005-of-00009.safetensors", "model.layers.36.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.36.self_attn.v_proj.bias": "model-00005-of-00009.safetensors", "model.layers.36.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.37.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.37.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.37.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.37.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.37.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.37.self_attn.k_proj.bias": "model-00005-of-00009.safetensors", "model.layers.37.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.37.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.37.self_attn.q_proj.bias": "model-00005-of-00009.safetensors", "model.layers.37.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.37.self_attn.v_proj.bias": "model-00005-of-00009.safetensors", "model.layers.37.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.38.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.38.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.38.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.38.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.38.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.38.self_attn.k_proj.bias": "model-00005-of-00009.safetensors", "model.layers.38.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.38.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.38.self_attn.q_proj.bias": "model-00005-of-00009.safetensors", "model.layers.38.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.38.self_attn.v_proj.bias": "model-00005-of-00009.safetensors", "model.layers.38.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.39.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.39.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.39.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.39.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.39.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.39.self_attn.k_proj.bias": "model-00005-of-00009.safetensors", "model.layers.39.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.39.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.39.self_attn.q_proj.bias": "model-00005-of-00009.safetensors", "model.layers.39.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.39.self_attn.v_proj.bias": "model-00005-of-00009.safetensors", "model.layers.39.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.4.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.4.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.4.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.4.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.4.self_attn.k_proj.bias": "model-00005-of-00009.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.4.self_attn.q_proj.bias": "model-00005-of-00009.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.4.self_attn.v_proj.bias": "model-00005-of-00009.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.40.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.40.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.40.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.40.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.40.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.40.self_attn.k_proj.bias": "model-00005-of-00009.safetensors", "model.layers.40.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.40.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.40.self_attn.q_proj.bias": "model-00005-of-00009.safetensors", "model.layers.40.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.40.self_attn.v_proj.bias": "model-00005-of-00009.safetensors", "model.layers.40.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.41.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.41.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.41.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.41.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.41.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.41.self_attn.k_proj.bias": "model-00005-of-00009.safetensors", "model.layers.41.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.41.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.41.self_attn.q_proj.bias": "model-00005-of-00009.safetensors", "model.layers.41.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.41.self_attn.v_proj.bias": "model-00005-of-00009.safetensors", "model.layers.41.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.42.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.42.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.42.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.42.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.42.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.42.self_attn.k_proj.bias": "model-00005-of-00009.safetensors", "model.layers.42.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.42.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.42.self_attn.q_proj.bias": "model-00005-of-00009.safetensors", "model.layers.42.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.42.self_attn.v_proj.bias": "model-00005-of-00009.safetensors", "model.layers.42.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.43.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.43.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.43.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.43.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.43.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.43.self_attn.k_proj.bias": "model-00005-of-00009.safetensors", "model.layers.43.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.43.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.43.self_attn.q_proj.bias": "model-00005-of-00009.safetensors", "model.layers.43.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.43.self_attn.v_proj.bias": "model-00005-of-00009.safetensors", "model.layers.43.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.44.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.44.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.44.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.44.mlp.up_proj.weight": "model-00006-of-00009.safetensors", "model.layers.44.post_attention_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.44.self_attn.k_proj.bias": "model-00006-of-00009.safetensors", "model.layers.44.self_attn.k_proj.weight": "model-00006-of-00009.safetensors", "model.layers.44.self_attn.o_proj.weight": "model-00006-of-00009.safetensors", "model.layers.44.self_attn.q_proj.bias": "model-00006-of-00009.safetensors", "model.layers.44.self_attn.q_proj.weight": "model-00006-of-00009.safetensors", "model.layers.44.self_attn.v_proj.bias": "model-00006-of-00009.safetensors", "model.layers.44.self_attn.v_proj.weight": "model-00006-of-00009.safetensors", "model.layers.45.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.45.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.45.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.45.mlp.up_proj.weight": "model-00006-of-00009.safetensors", "model.layers.45.post_attention_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.45.self_attn.k_proj.bias": "model-00006-of-00009.safetensors", "model.layers.45.self_attn.k_proj.weight": "model-00006-of-00009.safetensors", "model.layers.45.self_attn.o_proj.weight": "model-00006-of-00009.safetensors", "model.layers.45.self_attn.q_proj.bias": "model-00006-of-00009.safetensors", "model.layers.45.self_attn.q_proj.weight": "model-00006-of-00009.safetensors", "model.layers.45.self_attn.v_proj.bias": "model-00006-of-00009.safetensors", "model.layers.45.self_attn.v_proj.weight": "model-00006-of-00009.safetensors", "model.layers.46.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.46.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.46.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.46.mlp.up_proj.weight": "model-00006-of-00009.safetensors", "model.layers.46.post_attention_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.46.self_attn.k_proj.bias": "model-00006-of-00009.safetensors", "model.layers.46.self_attn.k_proj.weight": "model-00006-of-00009.safetensors", "model.layers.46.self_attn.o_proj.weight": "model-00006-of-00009.safetensors", "model.layers.46.self_attn.q_proj.bias": "model-00006-of-00009.safetensors", "model.layers.46.self_attn.q_proj.weight": "model-00006-of-00009.safetensors", "model.layers.46.self_attn.v_proj.bias": "model-00006-of-00009.safetensors", "model.layers.46.self_attn.v_proj.weight": "model-00006-of-00009.safetensors", "model.layers.47.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.48.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.49.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.50.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.51.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.52.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.53.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.54.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.55.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.56.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.57.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.58.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.59.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.60.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.61.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.62.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.63.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.64.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.65.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.66.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.48.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.49.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.50.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.51.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.52.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.53.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.54.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.55.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.56.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.57.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.58.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.59.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.60.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.61.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.62.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.63.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.64.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.65.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.66.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.47.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.48.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.55.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.56.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.47.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.49.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.50.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.51.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.52.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.53.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.54.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.57.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.58.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.59.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.60.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.61.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.62.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.63.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.64.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.65.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.66.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.48.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.49.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.50.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.51.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.52.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.53.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.54.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.55.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.56.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.57.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.58.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.59.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.60.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.61.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.62.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.63.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.64.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.65.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.66.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.47.mlp.up_proj.weight": "model-00008-of-00009.safetensors", "model.layers.47.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.48.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.49.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.50.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.51.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.52.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.53.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.54.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.55.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.56.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.57.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.58.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.59.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.60.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.61.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.62.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.63.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.64.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.65.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.66.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.48.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.49.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.50.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.51.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.52.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.53.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.54.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.55.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.57.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.58.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.59.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.60.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.61.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.62.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.63.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.56.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.64.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.65.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.66.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.47.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.48.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.54.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.55.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.56.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.47.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.49.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.50.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.51.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.52.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.53.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.57.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.58.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.59.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.60.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.61.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.62.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.63.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.64.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.65.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.66.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.48.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.49.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.50.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.51.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.52.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.53.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.54.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.55.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.56.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.57.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.58.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.59.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.60.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.61.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.62.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.63.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.64.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.65.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.66.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.47.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.47.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.48.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.49.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.50.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.51.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.52.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.53.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.54.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.55.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.56.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.57.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.58.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.59.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.60.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.61.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.62.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.63.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.64.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.65.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.66.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.48.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.49.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.50.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.51.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.52.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.53.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.55.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.54.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.56.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.57.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.58.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.59.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.60.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.61.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.62.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.63.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.64.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.65.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.66.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.47.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.48.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.54.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.55.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.56.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.66.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.47.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.49.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.50.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.51.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.52.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.53.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.57.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.58.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.59.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.60.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.61.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.62.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.63.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.64.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.65.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.48.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.49.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.50.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.51.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.52.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.53.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.54.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.55.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.56.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.57.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.58.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.59.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.60.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.61.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.62.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.63.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.64.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.65.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.66.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.47.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.5.input_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00008-of-00009.safetensors", "model.layers.5.mlp.gate_proj.weight": "model-00008-of-00009.safetensors", "model.layers.5.mlp.up_proj.weight": "model-00008-of-00009.safetensors", "model.layers.5.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.5.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.5.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.5.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.6.input_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00008-of-00009.safetensors", "model.layers.6.mlp.gate_proj.weight": "model-00008-of-00009.safetensors", "model.layers.6.mlp.up_proj.weight": "model-00008-of-00009.safetensors", "model.layers.6.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.6.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.6.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.6.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.7.input_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00008-of-00009.safetensors", "model.layers.7.mlp.gate_proj.weight": "model-00008-of-00009.safetensors", "model.layers.7.mlp.up_proj.weight": "model-00008-of-00009.safetensors", "model.layers.7.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.7.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.7.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.7.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.8.input_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.8.mlp.down_proj.weight": "model-00008-of-00009.safetensors", "model.layers.8.mlp.gate_proj.weight": "model-00008-of-00009.safetensors", "model.layers.8.mlp.up_proj.weight": "model-00008-of-00009.safetensors", "model.layers.8.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.8.self_attn.k_proj.bias": "model-00008-of-00009.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.8.self_attn.q_proj.bias": "model-00008-of-00009.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.8.self_attn.v_proj.bias": "model-00008-of-00009.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.9.input_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00009-of-00009.safetensors", "model.layers.9.mlp.gate_proj.weight": "model-00009-of-00009.safetensors", "model.layers.9.mlp.up_proj.weight": "model-00009-of-00009.safetensors", "model.layers.9.post_attention_layernorm.weight": "model-00009-of-00009.safetensors", "model.layers.9.self_attn.k_proj.bias": "model-00009-of-00009.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00009-of-00009.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00009-of-00009.safetensors", "model.layers.9.self_attn.q_proj.bias": "model-00009-of-00009.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00009-of-00009.safetensors", "model.layers.9.self_attn.v_proj.bias": "model-00009-of-00009.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00009-of-00009.safetensors", "model.norm.weight": "model-00009-of-00009.safetensors"}}
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
tokenizer_config.json ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "model_max_length": 16384,
203
+ "pad_token": "<|endoftext|>",
204
+ "split_special_tokens": false,
205
+ "tokenizer_class": "Qwen2Tokenizer",
206
+ "unk_token": null
207
+ }