ArtusDev commited on
Commit
18b4043
·
verified ·
1 Parent(s): c142d32

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,39 +1,223 @@
1
  ---
2
- base_model: qingy2024/GRMR-V3-Q4B
3
- base_model_relation: quantized
4
- quantized_by: ArtusDev
 
 
 
 
 
 
 
 
5
  ---
6
 
7
- ## EXL3 Quants of qingy2024/GRMR-V3-Q4B
8
-
9
- EXL3 quants of [qingy2024/GRMR-V3-Q4B](https://huggingface.co/qingy2024/GRMR-V3-Q4B) using <a href="https://github.com/turboderp-org/exllamav3/">exllamav3</a> for quantization.
10
-
11
- ### Quants
12
- | Quant(Revision) | Bits per Weight | Head Bits |
13
- | -------- | ---------- | --------- |
14
- | [3.0_H6](https://huggingface.co/ArtusDev/qingy2024_GRMR-V3-Q4B-EXL3/tree/3.0bpw_H6) | 3.0 | 6 |
15
- | [3.5_H6](https://huggingface.co/ArtusDev/qingy2024_GRMR-V3-Q4B-EXL3/tree/3.5bpw_H6) | 3.5 | 6 |
16
- | [4.0_H6](https://huggingface.co/ArtusDev/qingy2024_GRMR-V3-Q4B-EXL3/tree/4.0bpw_H6) | 4.0 | 6 |
17
- | [4.5_H6](https://huggingface.co/ArtusDev/qingy2024_GRMR-V3-Q4B-EXL3/tree/4.5bpw_H6) | 4.5 | 6 |
18
- | [5.0_H6](https://huggingface.co/ArtusDev/qingy2024_GRMR-V3-Q4B-EXL3/tree/5.0bpw_H6) | 5.0 | 6 |
19
- | [6.0_H6](https://huggingface.co/ArtusDev/qingy2024_GRMR-V3-Q4B-EXL3/tree/6.0bpw_H6) | 6.0 | 6 |
20
- | [8.0_H6](https://huggingface.co/ArtusDev/qingy2024_GRMR-V3-Q4B-EXL3/tree/8.0bpw_H6) | 8.0 | 6 |
21
- | [8.0_H8](https://huggingface.co/ArtusDev/qingy2024_GRMR-V3-Q4B-EXL3/tree/8.0bpw_H8) | 8.0 | 8 |
22
-
23
- ### Downloading quants with huggingface-cli
24
-
25
- <details>
26
- <summary>Click to view download instructions</summary>
27
-
28
- Install hugginface-cli:
29
-
30
- ```bash
31
- pip install -U "huggingface_hub[cli]"
32
- ```
33
-
34
- Download quant by targeting the specific quant revision (branch):
35
-
36
- ```
37
- huggingface-cli download ArtusDev/qingy2024_GRMR-V3-Q4B-EXL3 --revision "5bpw_H6" --local-dir ./
38
- ```
39
- </details>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model: unsloth/Qwen3-4B-Base
3
+ tags:
4
+ - text-generation-inference
5
+ - transformers
6
+ - unsloth
7
+ - qwen
8
+ - trl
9
+ - sft
10
+ license: apache-2.0
11
+ language:
12
+ - en
13
  ---
14
 
15
+ <html lang="en">
16
+ <head>
17
+ <meta charset="UTF-8">
18
+ <meta name="viewport" content="width=device-width, initial-scale=1.0">
19
+ </head>
20
+ <div class="container"><h1>GRMR-V3-Q4B</h1><p>GRMR-V3-Q4B is a fine-tuned version of <a href="https://huggingface.co/unsloth/Qwen3-4B-Base">unsloth/Qwen3-4B-Base</a> specifically optimized for grammar correction tasks.</p><div class="important-note"><p><strong>IMPORTANT:</strong> Please ensure you are using the following sampler settings for optimal results:</p><pre><code>temperature = 0.7
21
+ frequency_penalty = 0.0
22
+ presence_penalty = 0.0
23
+ min_p = 0.01
24
+ top_p = 0.95
25
+ top_k = 40</code></pre></div><h2>Model description</h2><p>GRMR-V3-Q4B is a grammar correction model built on Qwen3 4B base model. It has been fine-tuned on a large dataset of grammar correction examples to help improve text quality by fixing grammatical errors, punctuation, spelling, and other language issues.</p><p>The model uses a specialized chat template that structures inputs as "text" and outputs as "corrected" to maintain a clear distinction between original and corrected content.</p><h3>Model Details</h3><div class="model-meta"><p><strong>Developed by:</strong> qingy2024</p><p><strong>Finetuned from model:</strong> <a href="https://huggingface.co/unsloth/Qwen3-4B-Base">unsloth/Qwen3-4B-Base</a></p><p><strong>Model type:</strong> Qwen3</p><p><strong>Language(s):</strong> English</p><p><strong>License:</strong> apache-2.0</p><p><strong>Training Dataset:</strong> <a href="https://huggingface.co/datasets/qingy2024/grmr-v4-60k">qingy2024/grmr-v4-60k</a></p></div><hr><h2>Manual Testing Examples</h2><p>Here are a few examples of grammar corrections this model can handle:</p><table><thead><tr><th>Original Text</th><th>Corrected Text</th></tr></thead><tbody><tr><td>i dont know weather to bring a umbrella today</td><td>I don't know whether to bring an umbrella today.</td></tr><tr><td>she go to the store yesterday</td><td>She went to the store yesterday.</td></tr><tr><td>they is going to be late for the meeting</td><td>They are going to be late for the meeting.</td></tr><tr><td>the cat laying on the floor all day</td><td>The cat is laying on the floor all day.</td></tr></tbody></table><hr><h2>Training procedure</h2><p>The model was fine-tuned using full parameter fine-tuning (not LoRA) on the GRMR-V4-60K dataset. The training was optimized using the Unsloth framework for efficient training of LLMs.</p><h3>Training hyperparameters</h3><ul><li><strong>Batch size:</strong> 8</li><li><strong>Gradient accumulation steps:</strong> 2</li><li><strong>Learning rate:</strong> 5e-5</li><li><strong>Epochs:</strong> 1</li><li><strong>Optimizer:</strong> AdamW (8-bit)</li><li><strong>Weight decay:</strong> 0.01</li><li><strong>LR scheduler:</strong> Cosine</li><li><strong>Warmup steps:</strong> 180</li><li><strong>Max sequence length:</strong> 16,384</li><li><strong>Training precision:</strong> Mixed precision (BF16 where available, FP16 otherwise)</li></ul><h2>Intended uses & limitations</h2><p>This model is designed for grammar correction tasks. It can be used to:</p><ul><li>Fix grammatical errors in written text</li><li>Correct punctuation</li><li>Address spelling mistakes</li><li>Improve sentence structure and clarity</li></ul><h3>Limitations</h3><ul><li>The model may struggle with highly technical or domain-specific content</li><li>It may not fully understand context-dependent grammar rules in all cases</li><li>Performance may vary for non-standard English or text with multiple errors</li></ul><h2>How to use</h2><p>Projects based on Hugging Face transformers should be able to run this model easily.</p><pre><code class="language-python">from transformers import AutoModelForCausalLM, AutoTokenizer
26
+ # Load model and tokenizer
27
+ model_name = "qingy2024/GRMR-V3-Q4B"
28
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
29
+ model = AutoModelForCausalLM.from_pretrained(model_name)
30
+ # Text with grammar errors to correct
31
+ text_to_correct = "i am going to the store tommorow and buy some thing for dinner"
32
+ # Format as messages
33
+ messages = [
34
+ {"role": "user", "content": text_to_correct}
35
+ ]
36
+ # Apply the custom chat template
37
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
38
+ # Tokenize and generate
39
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
40
+ outputs = model.generate(
41
+ inputs["input_ids"],
42
+ max_new_tokens=512,
43
+ temperature=0.1,
44
+ do_sample=True
45
+ )
46
+ # Decode and print the corrected text
47
+ corrected_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
48
+ print(corrected_text)</code></pre><h3>Using with the Hugging Face pipeline</h3><pre><code class="language-python">from transformers import pipeline
49
+ pipe = pipeline(
50
+ "text-generation",
51
+ model="qingy2024/GRMR-V3-Q4B",
52
+ torch_dtype="auto",
53
+ device_map="auto"
54
+ )
55
+ messages = [
56
+ {"role": "user", "content": "i dont know weather to bring a umbrella today"}
57
+ ]
58
+ result = pipe(
59
+ messages,
60
+ max_new_tokens=100,
61
+ temperature=0.1,
62
+ do_sample=True,
63
+ return_full_text=False
64
+ )[0]["generated_text"]
65
+ print(result)</code></pre><h2>Custom Chat Template</h2><p class="chat-template-info">The model uses a custom chat template with special formatting for grammar correction:</p><ul><li>User inputs are formatted with <code><|text_start|></code> and <code><|text_end|></code> tags</li><li>Model outputs are formatted with <code><|corrected_start|></code> and <code><|corrected_end|></code> tags</li></ul><p>The complete chat template is:</p><pre><code class="language-jinja">{%- for message in messages %}
66
+ {%- if message.role == "user" %}
67
+ {{- '<|text_start|>\n' + message.content + '<|text_end|>\n' }}
68
+ {%- elif message.role == "assistant" %}
69
+ {{- '<|corrected_start|>\n' + message.content + '<|corrected_end|>\n' }}
70
+ {%- else %}
71
+ {{- raise('Unknown role: ' + message.role) }}
72
+ {%- endif %}
73
+ {%- endfor %}
74
+ {%- if add_generation_prompt %}
75
+ {{- '<|corrected_start|>\n' }}
76
+ {%- endif %}</code></pre><h2>Training Dataset</h2><p>The model was fine-tuned on the <a href="https://huggingface.co/datasets/qingy2024/grmr-v4-60k">qingy2024/grmr-v4-60k</a> dataset, which contains 60,000 examples of original text and their grammatically corrected versions.</p><h2>Bias, Risks, and Limitations</h2><ul><li>The model may reflect biases present in the training data</li><li>It may not perform equally well across different writing styles or domains</li><li>The model might occasionally introduce errors or change the meaning of text</li><li>It focuses on grammatical correctness rather than stylistic improvements</li></ul><h2>Citations</h2><pre><code>@misc{qwen3technicalreport,
77
+ title={Qwen3 Technical Report},
78
+ author={Qwen Team},
79
+ year={2025},
80
+ eprint={2505.09388},
81
+ archivePrefix={arXiv},
82
+ primaryClass={cs.CL},
83
+ url={https://arxiv.org/abs/2505.09388},
84
+ }</code></pre><h2>Contact</h2><p>For questions or issues related to the model, please reach out via Hugging Face or by creating an issue in the repository.</p></div>
85
+ <style>
86
+ body {
87
+ font-family: -apple-system, BlinkMacSystemFont, "Segoe UI", Roboto, Helvetica, Arial, sans-serif, "Apple Color Emoji", "Segoe UI Emoji", "Segoe UI Symbol";
88
+ line-height: 1.6;
89
+ margin: 0;
90
+ padding: 0;
91
+ background-color: #f8f9fa;
92
+ color: #333;
93
+ }
94
+ .container {
95
+ max-width: 1200px;
96
+ margin: 10px auto;
97
+ padding: 25px;
98
+ background-color: #ffffff;
99
+ border-radius: 8px;
100
+ box-shadow: 0 4px 12px rgba(0, 0, 0, 0.08);
101
+ }
102
+ h1, h2, h3 {
103
+ color: #0056b3; /* Primary Blue */
104
+ margin-top: 1.5em;
105
+ margin-bottom: 0.7em;
106
+ }
107
+ h1 {
108
+ text-align: center;
109
+ font-size: 2.2em;
110
+ border-bottom: 2px solid #e0e0e0;
111
+ padding-bottom: 0.5em;
112
+ margin-top: 0;
113
+ }
114
+ h2 {
115
+ font-size: 1.8em;
116
+ border-bottom: 1px solid #e9ecef;
117
+ padding-bottom: 0.3em;
118
+ }
119
+ h3 {
120
+ font-size: 1.4em;
121
+ color: #007bff; /* Lighter Blue for sub-headings */
122
+ }
123
+ p, li {
124
+ font-size: 1em;
125
+ color: #555;
126
+ }
127
+ a {
128
+ color: #007bff;
129
+ text-decoration: none;
130
+ }
131
+ a:hover {
132
+ text-decoration: underline;
133
+ color: #0056b3;
134
+ }
135
+ .important-note {
136
+ background-color: #e7f3ff; /* Light blue background */
137
+ border-left: 5px solid #007bff; /* Blue accent border */
138
+ margin: 20px 0px;
139
+ border-radius: 5px;
140
+ }
141
+ .important-note strong {
142
+ color: #0056b3;
143
+ font-weight: 600;
144
+ }
145
+ .important-note {
146
+ background-color: #d0e8ff;
147
+ padding: 0.05em 1.0em;
148
+ border-radius: 3px;
149
+ font-size: 0.9em;
150
+ }
151
+ code {
152
+ padding: 0.1em 0.4em;
153
+ border-radius: 3px;
154
+ font-size: 0.9em;
155
+ }
156
+ table {
157
+ width: 100%;
158
+ border-collapse: collapse;
159
+ margin: 20px 0;
160
+ box-shadow: 0 2px 4px rgba(0,0,0,0.05);
161
+ }
162
+ th, td {
163
+ border: 1px solid #dee2e6;
164
+ padding: 10px 12px;
165
+ text-align: left;
166
+ vertical-align: top;
167
+ }
168
+ th {
169
+ background-color: #e9ecef; /* Light gray for headers */
170
+ font-weight: 600;
171
+ color: #212529;
172
+ }
173
+ td:first-child {
174
+ /* font-style: italic; */
175
+ color: #444;
176
+ }
177
+ pre {
178
+ background-color: #f1f3f5;
179
+ padding: 15px;
180
+ border-radius: 5px;
181
+ overflow-x: auto;
182
+ border: 1px solid #ced4da;
183
+ font-size: 0.9em;
184
+ }
185
+ code {
186
+ font-family: "SFMono-Regular", Consolas, "Liberation Mono", Menlo, Courier, monospace;
187
+ background-color: #e9ecef;
188
+ padding: 0.2em 0.4em;
189
+ border-radius: 3px;
190
+ font-size: 0.9em;
191
+ }
192
+ pre code {
193
+ background-color: transparent;
194
+ padding: 0;
195
+ border-radius: 0;
196
+ font-size: 1em;
197
+ }
198
+ ul {
199
+ padding-left: 20px;
200
+ }
201
+ li {
202
+ margin-bottom: 0.5em;
203
+ }
204
+ hr {
205
+ border: none;
206
+ border-top: 1px solid #e0e0e0;
207
+ margin: 30px 0;
208
+ }
209
+ .model-meta {
210
+ background-color: #f8f9fa;
211
+ padding: 15px;
212
+ border-radius: 5px;
213
+ margin-bottom: 20px;
214
+ border: 1px solid #e9ecef;
215
+ }
216
+ .model-meta p { margin-bottom: 0.5em; }
217
+ .model-meta strong { color: #333; }
218
+ /* Specific styling for chat template explanation */
219
+ .chat-template-info span {
220
+ font-weight: bold;
221
+ color: #0056b3;
222
+ }
223
+ </style></html>
added_tokens.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|corrected_end|>": 151672,
11
+ "<|corrected_start|>": 151671,
12
+ "<|endoftext|>": 151643,
13
+ "<|file_sep|>": 151664,
14
+ "<|fim_middle|>": 151660,
15
+ "<|fim_pad|>": 151662,
16
+ "<|fim_prefix|>": 151659,
17
+ "<|fim_suffix|>": 151661,
18
+ "<|im_end|>": 151645,
19
+ "<|im_start|>": 151644,
20
+ "<|image_pad|>": 151655,
21
+ "<|object_ref_end|>": 151647,
22
+ "<|object_ref_start|>": 151646,
23
+ "<|quad_end|>": 151651,
24
+ "<|quad_start|>": 151650,
25
+ "<|repo_name|>": 151663,
26
+ "<|text_end|>": 151670,
27
+ "<|text_start|>": 151669,
28
+ "<|video_pad|>": 151656,
29
+ "<|vision_end|>": 151653,
30
+ "<|vision_pad|>": 151654,
31
+ "<|vision_start|>": 151652
32
+ }
chat_template.jinja ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- for message in messages %}
2
+ {%- if message.role == "user" %}
3
+ {{- '<|text_start|>
4
+ ' + message.content + '<|text_end|>
5
+ ' }}
6
+ {%- elif message.role == "assistant" %}
7
+ {{- '<|corrected_start|>
8
+ ' + message.content + '<|corrected_end|>
9
+ ' }}
10
+ {%- else %}
11
+ {{- raise('Unknown role: ' + message.role) }}
12
+ {%- endif %}
13
+ {%- endfor %}
14
+ {%- if add_generation_prompt %}
15
+ {{- '<|corrected_start|>
16
+ ' }}
17
+ {%- endif %}
config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "chat_template": "{%- for message in messages %}\n{%- if message.role == \"user\" %}\n{{- '<|text_start|>\n' + message.content + '<|text_end|>\n' }}\n{%- elif message.role == \"assistant\" %}\n{{- '<|corrected_start|>\n' + message.content + '<|corrected_end|>\n' }}\n{%- else %}\n{# Raise an error for unsupported roles, as per requirement to remove system message stuff #}\n{{- raise('Unknown role: ' + message.role) }}\n{%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n{{- '<|corrected_start|>\n' }}\n{%- endif %}",
8
+ "eos_token_id": 151643,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 2560,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 9728,
14
+ "max_position_embeddings": 32768,
15
+ "max_window_layers": 36,
16
+ "model_type": "qwen3",
17
+ "num_attention_heads": 32,
18
+ "num_hidden_layers": 36,
19
+ "num_key_value_heads": 8,
20
+ "pad_token_id": 151654,
21
+ "rms_norm_eps": 1e-06,
22
+ "rope_scaling": null,
23
+ "rope_theta": 1000000,
24
+ "sliding_window": null,
25
+ "tie_word_embeddings": true,
26
+ "torch_dtype": "bfloat16",
27
+ "transformers_version": "4.52.4",
28
+ "unsloth_fixed": true,
29
+ "unsloth_version": "2025.5.9",
30
+ "use_cache": true,
31
+ "use_sliding_window": false,
32
+ "vocab_size": 151673,
33
+ "quantization_config": {
34
+ "quant_method": "exl3",
35
+ "version": "0.0.3",
36
+ "bits": 4.5,
37
+ "head_bits": 6,
38
+ "calibration": {
39
+ "rows": 100,
40
+ "cols": 2048
41
+ },
42
+ "out_scales": "auto"
43
+ }
44
+ }
generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "eos_token_id": 151643,
4
+ "max_length": 32768,
5
+ "max_new_tokens": 2048,
6
+ "pad_token_id": 151654,
7
+ "transformers_version": "4.52.4"
8
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20b69cca0304a35581fa3d14140cf725a11cae54ccee4f6e258b82ac222f402b
3
+ size 3116132144
quantization_config.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<|text_start|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "<|text_end|>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ },
17
+ {
18
+ "content": "<|corrected_start|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ {
25
+ "content": "<|corrected_end|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ ],
32
+ "eos_token": {
33
+ "content": "<|endoftext|>",
34
+ "lstrip": false,
35
+ "normalized": false,
36
+ "rstrip": false,
37
+ "single_word": false
38
+ },
39
+ "pad_token": {
40
+ "content": "<|vision_pad|>",
41
+ "lstrip": false,
42
+ "normalized": false,
43
+ "rstrip": false,
44
+ "single_word": false
45
+ }
46
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40fd5b853d0598451f720d22bf96c9c4cd4b5c6413cd74d4f5e3494ee990f38d
3
+ size 11423523
tokenizer_config.json ADDED
@@ -0,0 +1,263 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<tool_response>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "151666": {
190
+ "content": "</tool_response>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "151667": {
198
+ "content": "<think>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "151668": {
206
+ "content": "</think>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ },
213
+ "151669": {
214
+ "content": "<|text_start|>",
215
+ "lstrip": false,
216
+ "normalized": false,
217
+ "rstrip": false,
218
+ "single_word": false,
219
+ "special": true
220
+ },
221
+ "151670": {
222
+ "content": "<|text_end|>",
223
+ "lstrip": false,
224
+ "normalized": false,
225
+ "rstrip": false,
226
+ "single_word": false,
227
+ "special": true
228
+ },
229
+ "151671": {
230
+ "content": "<|corrected_start|>",
231
+ "lstrip": false,
232
+ "normalized": false,
233
+ "rstrip": false,
234
+ "single_word": false,
235
+ "special": true
236
+ },
237
+ "151672": {
238
+ "content": "<|corrected_end|>",
239
+ "lstrip": false,
240
+ "normalized": false,
241
+ "rstrip": false,
242
+ "single_word": false,
243
+ "special": true
244
+ }
245
+ },
246
+ "additional_special_tokens": [
247
+ "<|text_start|>",
248
+ "<|text_end|>",
249
+ "<|corrected_start|>",
250
+ "<|corrected_end|>"
251
+ ],
252
+ "bos_token": null,
253
+ "clean_up_tokenization_spaces": false,
254
+ "eos_token": "<|endoftext|>",
255
+ "errors": "replace",
256
+ "extra_special_tokens": {},
257
+ "model_max_length": 32768,
258
+ "pad_token": "<|vision_pad|>",
259
+ "padding_side": "right",
260
+ "split_special_tokens": false,
261
+ "tokenizer_class": "Qwen2Tokenizer",
262
+ "unk_token": null
263
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff