Felladrin commited on
Commit
0aa82b0
·
verified ·
1 Parent(s): bffda99

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen1.5-MoE-A2.7B-Chat
3
+ ---
4
+
5
+ [MLC](https://llm.mlc.ai/) version of [Qwen/Qwen1.5-MoE-A2.7B-Chat](https://huggingface.co/Qwen/Qwen1.5-MoE-A2.7B-Chat), using `q4f16_1` quantization.
6
+
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
mlc-chat-config.json ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "0.1.0",
3
+ "model_type": "qwen2_moe",
4
+ "quantization": "q4f16_1",
5
+ "model_config": {
6
+ "hidden_act": "silu",
7
+ "hidden_size": 2048,
8
+ "intermediate_size": 5632,
9
+ "num_attention_heads": 16,
10
+ "num_hidden_layers": 24,
11
+ "num_key_value_heads": 16,
12
+ "rms_norm_eps": 1e-06,
13
+ "rope_theta": 1000000.0,
14
+ "vocab_size": 151936,
15
+ "tie_word_embeddings": false,
16
+ "context_window_size": 32768,
17
+ "prefill_chunk_size": 2048,
18
+ "tensor_parallel_shards": 1,
19
+ "head_dim": 128,
20
+ "dtype": "float32",
21
+ "max_batch_size": 80,
22
+ "moe_intermediate_size": 1408,
23
+ "shared_expert_intermediate_size": 5632,
24
+ "num_experts_per_tok": 4,
25
+ "num_experts": 60,
26
+ "decoder_sparse_step": 1,
27
+ "norm_topk_prob": false
28
+ },
29
+ "vocab_size": 151936,
30
+ "context_window_size": 32768,
31
+ "sliding_window_size": -1,
32
+ "prefill_chunk_size": 2048,
33
+ "attention_sink_size": -1,
34
+ "tensor_parallel_shards": 1,
35
+ "temperature": 0.7,
36
+ "presence_penalty": 0.0,
37
+ "frequency_penalty": 0.0,
38
+ "repetition_penalty": 1.05,
39
+ "top_p": 0.8,
40
+ "tokenizer_files": [
41
+ "tokenizer.json",
42
+ "vocab.json",
43
+ "merges.txt",
44
+ "tokenizer_config.json"
45
+ ],
46
+ "tokenizer_info": {
47
+ "token_postproc_method": "byte_level",
48
+ "prepend_space_in_encode": false,
49
+ "strip_space_in_decode": false
50
+ },
51
+ "conv_template": {
52
+ "name": "chatml",
53
+ "system_template": "<|im_start|>system\n{system_message}<|im_end|>\n",
54
+ "system_message": "A conversation between a user and an LLM-based AI assistant. The assistant gives helpful and honest answers.",
55
+ "system_prefix_token_ids": null,
56
+ "add_role_after_system_message": true,
57
+ "roles": {
58
+ "user": "<|im_start|>user",
59
+ "assistant": "<|im_start|>assistant"
60
+ },
61
+ "role_templates": {
62
+ "user": "{user_message}",
63
+ "assistant": "{assistant_message}",
64
+ "tool": "{tool_message}"
65
+ },
66
+ "messages": [],
67
+ "seps": [
68
+ "<|im_end|>\n"
69
+ ],
70
+ "role_content_sep": "\n",
71
+ "role_empty_sep": "\n",
72
+ "stop_str": [
73
+ "<|im_end|>"
74
+ ],
75
+ "stop_token_ids": [
76
+ 2
77
+ ],
78
+ "function_string": "",
79
+ "use_function_calling": false
80
+ },
81
+ "pad_token_id": 151643,
82
+ "bos_token_id": 151643,
83
+ "eos_token_id": [
84
+ 151645,
85
+ 151643
86
+ ]
87
+ }
model.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:416ecde813854a5f7df0a11b3fcb659f4909bdd4816f304099e28f9841b50acb
3
+ size 5150442
params_shard_0.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc981f3ce907322a6dc70b4f84b9f1e515cc0fc46e28bb076825168c9643744d
3
+ size 155582464
params_shard_1.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e2de82c66adf507a3971269303a6d7b3e84f18440615f51c7710cc278c00bee
3
+ size 25939968
params_shard_2.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4d69e4060d5905a644777f1d38b9d7e04fab09a3c18e44ccfa0ce946a0d94ce
3
+ size 155582464
params_shard_3.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee278b25bfb80b70e128bb41248c49bb208bd9d187ea7636b0a0675c7260af35
3
+ size 86507520
params_shard_4.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea186491e043a27e3c2917f87b68b7c3bff960c7cd38dd6a8be9e8d73f5e00b2
3
+ size 32440320
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": ["<|im_start|>", "<|im_end|>"],
30
+ "bos_token": null,
31
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
32
+ "clean_up_tokenization_spaces": false,
33
+ "eos_token": "<|im_end|>",
34
+ "errors": "replace",
35
+ "model_max_length": 32768,
36
+ "pad_token": "<|endoftext|>",
37
+ "split_special_tokens": false,
38
+ "tokenizer_class": "Qwen2Tokenizer",
39
+ "unk_token": null
40
+ }
41
+
vocab.json ADDED
The diff for this file is too large to render. See raw diff