Rub11037 commited on
Commit
a42205b
·
verified ·
1 Parent(s): a2754a9

Update tokenizer_config.json

Browse files
Files changed (1) hide show
  1. tokenizer_config.json +6 -4
tokenizer_config.json CHANGED
@@ -28,14 +28,16 @@
28
  }
29
  },
30
  "bos_token": "<s>",
 
31
  "clean_up_tokenization_spaces": false,
32
  "eos_token": "</s>",
33
- "legacy": false,
34
  "model_max_length": 1000000000000000019884624838656,
35
- "pad_token": "<s>",
36
- "padding_side": "left",
37
  "sp_model_kwargs": {},
 
38
  "tokenizer_class": "LlamaTokenizer",
 
39
  "unk_token": "<unk>",
40
- "use_default_system_prompt": false
41
  }
 
28
  }
29
  },
30
  "bos_token": "<s>",
31
+ "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
34
+ "legacy": true,
35
  "model_max_length": 1000000000000000019884624838656,
36
+ "pad_token": "</s>",
 
37
  "sp_model_kwargs": {},
38
+ "spaces_between_special_tokens": false,
39
  "tokenizer_class": "LlamaTokenizer",
40
+ "truncation_side": "left",
41
  "unk_token": "<unk>",
42
+ "use_default_system_prompt": true
43
  }