English
mistral
PocketDoc commited on
Commit
43542b7
·
verified ·
1 Parent(s): a75e0d8

Upload 4 files

Browse files
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37f00374dea48658ee8f5d0f21895b9bc55cb0103939607c8185bfd1c6ca1f89
3
+ size 587404
tokenizer.model.v3 ADDED
Binary file (587 kB). View file
 
tokenizer_config.json CHANGED
@@ -6175,14 +6175,13 @@
6175
  "bos_token": "<s>",
6176
  "clean_up_tokenization_spaces": false,
6177
  "eos_token": "</s>",
6178
- "extra_special_tokens": {},
6179
  "legacy": false,
6180
  "model_max_length": 1000000000000000019884624838656,
6181
  "pad_token": "[control_768]",
6182
  "padding_side": "left",
6183
  "sp_model_kwargs": {},
6184
  "spaces_between_special_tokens": false,
6185
- "tokenizer_class": "PreTrainedTokenizerFast",
6186
  "unk_token": "<unk>",
6187
  "use_default_system_prompt": false
6188
  }
 
6175
  "bos_token": "<s>",
6176
  "clean_up_tokenization_spaces": false,
6177
  "eos_token": "</s>",
 
6178
  "legacy": false,
6179
  "model_max_length": 1000000000000000019884624838656,
6180
  "pad_token": "[control_768]",
6181
  "padding_side": "left",
6182
  "sp_model_kwargs": {},
6183
  "spaces_between_special_tokens": false,
6184
+ "tokenizer_class": "LlamaTokenizer",
6185
  "unk_token": "<unk>",
6186
  "use_default_system_prompt": false
6187
  }