Kieran2828 commited on
Commit
9b0a5e2
·
verified ·
1 Parent(s): 97954d6

Upload 2 files

Browse files
Files changed (2) hide show
  1. added_tokens.json +5 -0
  2. mlc-chat-config.json +85 -0
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|bos|>": 32000,
3
+ "<|endoftext|>": 32001,
4
+ "[PAD]": 32002
5
+ }
mlc-chat-config.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "0.1.0",
3
+ "model_type": "mixtral",
4
+ "quantization": "q4f16_1",
5
+ "model_config": {
6
+ "hidden_size": 1024,
7
+ "intermediate_size": 4096,
8
+ "num_attention_heads": 32,
9
+ "num_hidden_layers": 12,
10
+ "rms_norm_eps": 1e-06,
11
+ "vocab_size": 32003,
12
+ "tie_word_embeddings": false,
13
+ "position_embedding_base": 10000.0,
14
+ "rope_scaling": null,
15
+ "context_window_size": 2048,
16
+ "prefill_chunk_size": 512,
17
+ "num_key_value_heads": 8,
18
+ "head_dim": 32,
19
+ "tensor_parallel_shards": 1,
20
+ "pipeline_parallel_stages": 1,
21
+ "max_batch_size": 128,
22
+ "disaggregation": false,
23
+ "num_local_experts": 4,
24
+ "num_experts_per_tok": 2
25
+ },
26
+ "vocab_size": 32003,
27
+ "context_window_size": 2048,
28
+ "sliding_window_size": -1,
29
+ "prefill_chunk_size": 512,
30
+ "attention_sink_size": -1,
31
+ "tensor_parallel_shards": 1,
32
+ "pipeline_parallel_stages": 1,
33
+ "temperature": 1.0,
34
+ "presence_penalty": 0.0,
35
+ "frequency_penalty": 0.0,
36
+ "repetition_penalty": 1.0,
37
+ "top_p": 1.0,
38
+ "tokenizer_files": [
39
+ "tokenizer.model",
40
+ "tokenizer.json",
41
+ "added_tokens.json",
42
+ "tokenizer_config.json"
43
+ ],
44
+ "tokenizer_info": {
45
+ "token_postproc_method": "byte_fallback",
46
+ "prepend_space_in_encode": true,
47
+ "strip_space_in_decode": true
48
+ },
49
+ "conv_template": {
50
+ "name": "mistral_default",
51
+ "system_template": "[INST] {system_message}",
52
+ "system_message": "Always assist with care, respect, and truth. Respond with utmost utility yet securely. Avoid harmful, unethical, prejudiced, or negative content. Ensure replies promote fairness and positivity.",
53
+ "system_prefix_token_ids": [
54
+ 1
55
+ ],
56
+ "add_role_after_system_message": false,
57
+ "roles": {
58
+ "user": "[INST]",
59
+ "assistant": "[/INST]",
60
+ "tool": "[INST]"
61
+ },
62
+ "role_templates": {
63
+ "user": "{user_message}",
64
+ "assistant": "{assistant_message}",
65
+ "tool": "{tool_message}"
66
+ },
67
+ "messages": [],
68
+ "seps": [
69
+ " "
70
+ ],
71
+ "role_content_sep": " ",
72
+ "role_empty_sep": "",
73
+ "stop_str": [
74
+ "</s>"
75
+ ],
76
+ "stop_token_ids": [
77
+ 2
78
+ ],
79
+ "function_string": "",
80
+ "use_function_calling": false
81
+ },
82
+ "pad_token_id": 0,
83
+ "bos_token_id": 1,
84
+ "eos_token_id": 2
85
+ }