Upload 4 files
Browse files- tokenizer.json +0 -0
- tokenizer.model +3 -0
- tokenizer.model.v3 +0 -0
- tokenizer_config.json +1 -2
tokenizer.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37f00374dea48658ee8f5d0f21895b9bc55cb0103939607c8185bfd1c6ca1f89
|
3 |
+
size 587404
|
tokenizer.model.v3
ADDED
Binary file (587 kB). View file
|
|
tokenizer_config.json
CHANGED
@@ -6175,14 +6175,13 @@
|
|
6175 |
"bos_token": "<s>",
|
6176 |
"clean_up_tokenization_spaces": false,
|
6177 |
"eos_token": "</s>",
|
6178 |
-
"extra_special_tokens": {},
|
6179 |
"legacy": false,
|
6180 |
"model_max_length": 1000000000000000019884624838656,
|
6181 |
"pad_token": "[control_768]",
|
6182 |
"padding_side": "left",
|
6183 |
"sp_model_kwargs": {},
|
6184 |
"spaces_between_special_tokens": false,
|
6185 |
-
"tokenizer_class": "
|
6186 |
"unk_token": "<unk>",
|
6187 |
"use_default_system_prompt": false
|
6188 |
}
|
|
|
6175 |
"bos_token": "<s>",
|
6176 |
"clean_up_tokenization_spaces": false,
|
6177 |
"eos_token": "</s>",
|
|
|
6178 |
"legacy": false,
|
6179 |
"model_max_length": 1000000000000000019884624838656,
|
6180 |
"pad_token": "[control_768]",
|
6181 |
"padding_side": "left",
|
6182 |
"sp_model_kwargs": {},
|
6183 |
"spaces_between_special_tokens": false,
|
6184 |
+
"tokenizer_class": "LlamaTokenizer",
|
6185 |
"unk_token": "<unk>",
|
6186 |
"use_default_system_prompt": false
|
6187 |
}
|