|
If your model expects special tokens, make sure to add them to the template! |
|
python |
|
tokenizer.chat_template = "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}" |
|
This template wraps each message in <|im_start|> and <|im_end|> tokens, and simply writes the role as a string, which |
|
allows for flexibility in the roles you train with. |