Upload folder using huggingface_hub
Browse files- .gitattributes +5 -0
- chat_template.jinja +13 -0
- config.json +46 -0
- onnx/model_q4.onnx +3 -0
- onnx/model_q4.onnx_data +3 -0
- onnx/model_q4.onnx_data_1 +3 -0
- onnx/model_q4f16.onnx +3 -0
- onnx/model_q4f16.onnx_data +3 -0
- onnx/model_q4f16.onnx_data_1 +3 -0
- special_tokens_map.json +23 -0
- tokenizer.json +3 -0
- tokenizer_config.json +56 -0
.gitattributes
CHANGED
@@ -33,3 +33,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
onnx/model_q4.onnx_data filter=lfs diff=lfs merge=lfs -text
|
37 |
+
onnx/model_q4.onnx_data_1 filter=lfs diff=lfs merge=lfs -text
|
38 |
+
onnx/model_q4f16.onnx_data filter=lfs diff=lfs merge=lfs -text
|
39 |
+
onnx/model_q4f16.onnx_data_1 filter=lfs diff=lfs merge=lfs -text
|
40 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
chat_template.jinja
ADDED
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{%- if messages[0]['role'] == 'system' %}
|
2 |
+
{{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
|
3 |
+
{%- else %}
|
4 |
+
{{- '<|im_start|>system\nThe assistant is AFM-4.5B, trained by Arcee AI, with 4.5 billion parameters. AFM is a deeply thoughtful, helpful assistant. The assistant is having a conversation with the user. The assistant\'s responses are calm, intelligent, and personable, always aiming to truly understand the user\'s intent. AFM thinks aloud, step by step, when solving problems or forming explanations, much like a careful, reflective thinker would. The assistant helps with sincerity and depth. If a topic invites introspection, curiosity, or broader insight, the assistant allows space for reflection — be open to nuance and complexity. The assistant is not robotic or overly formal; it speaks like a wise, thoughtful companion who cares about clarity and the human experience. If a topic is uncertain or depends on subjective interpretation, AFM explains the possibilities thoughtfully.<|im_end|>\n' }}
|
5 |
+
{%- endif %}
|
6 |
+
{%- for message in messages %}
|
7 |
+
{%- if not (message.role == 'system' and loop.first) %}
|
8 |
+
{{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>\n' }}
|
9 |
+
{%- endif %}
|
10 |
+
{%- endfor %}
|
11 |
+
{%- if messages[-1]['role'] != 'assistant' %}
|
12 |
+
{{- '<|im_start|>assistant\n' }}
|
13 |
+
{%- endif %}
|
config.json
ADDED
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"ArceeForCausalLM"
|
4 |
+
],
|
5 |
+
"attention_bias": false,
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 128000,
|
8 |
+
"eos_token_id": 128003,
|
9 |
+
"head_dim": 128,
|
10 |
+
"hidden_act": "relu2",
|
11 |
+
"hidden_size": 2560,
|
12 |
+
"initializer_range": 0.02,
|
13 |
+
"intermediate_size": 18432,
|
14 |
+
"max_position_embeddings": 65536,
|
15 |
+
"mlp_bias": false,
|
16 |
+
"model_type": "arcee",
|
17 |
+
"num_attention_heads": 20,
|
18 |
+
"num_hidden_layers": 36,
|
19 |
+
"num_key_value_heads": 4,
|
20 |
+
"rms_norm_eps": 1e-05,
|
21 |
+
"rope_scaling": {
|
22 |
+
"beta_fast": 32.0,
|
23 |
+
"beta_slow": 1.0,
|
24 |
+
"factor": 20.0,
|
25 |
+
"mscale": 1.0,
|
26 |
+
"original_max_position_embeddings": 4096,
|
27 |
+
"rope_type": "yarn",
|
28 |
+
"type": "yarn"
|
29 |
+
},
|
30 |
+
"rope_theta": 10000.0,
|
31 |
+
"tie_word_embeddings": false,
|
32 |
+
"torch_dtype": "bfloat16",
|
33 |
+
"transformers_version": "4.54.0.dev0",
|
34 |
+
"use_cache": false,
|
35 |
+
"vocab_size": 128005,
|
36 |
+
"transformers.js_config": {
|
37 |
+
"use_external_data_format": {
|
38 |
+
"model_q4.onnx": 2,
|
39 |
+
"model_q4f16.onnx": 2
|
40 |
+
},
|
41 |
+
"kv_cache_dtype": {
|
42 |
+
"q4f16": "float16",
|
43 |
+
"fp16": "float16"
|
44 |
+
}
|
45 |
+
}
|
46 |
+
}
|
onnx/model_q4.onnx
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1af24d2bdd7e7927b9509c9cbd06aa86e02796f276d1addc08e4a113f04c3bcc
|
3 |
+
size 321381
|
onnx/model_q4.onnx_data
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9faec389f05d81f631f3b4f24280979fa412947867d2c11ae97b7735e36c439c
|
3 |
+
size 2097055552
|
onnx/model_q4.onnx_data_1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:397d518a0e358168492b4a0f2041a653ca50d63e81e091f96228ff424dccda98
|
3 |
+
size 824123200
|
onnx/model_q4f16.onnx
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8ebb79e862018ad77e174bb2bac0462f4d29a26c8844a00bc12d443eb45f40cd
|
3 |
+
size 322886
|
onnx/model_q4f16.onnx_data
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e8c1118ca6877357989e11b1fbb4c7a8c59eea73a349caa3604e1a145ffb2ecf
|
3 |
+
size 2094813216
|
onnx/model_q4f16.onnx_data_1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6a76ca33a5695c49696110d09ff76a19f43fbe1971db9a54867e355d9fb9908c
|
3 |
+
size 520526880
|
special_tokens_map.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<|begin_of_text|>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "<|im_end|>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "<|finetune_right_pad_id|>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
}
|
23 |
+
}
|
tokenizer.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1363c79c75bf4202ce0f7ca44828b99a72140b97340578fa3b8927dddc46da2e
|
3 |
+
size 11535885
|
tokenizer_config.json
ADDED
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"128000": {
|
4 |
+
"content": "<|begin_of_text|>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"128001": {
|
12 |
+
"content": "<|end_of_text|>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"128002": {
|
20 |
+
"content": "<|im_start|>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"128003": {
|
28 |
+
"content": "<|im_end|>",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"128004": {
|
36 |
+
"content": "<|finetune_right_pad_id|>",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
}
|
43 |
+
},
|
44 |
+
"bos_token": "<|begin_of_text|>",
|
45 |
+
"clean_up_tokenization_spaces": true,
|
46 |
+
"eos_token": "<|im_end|>",
|
47 |
+
"extra_special_tokens": {},
|
48 |
+
"model_input_names": [
|
49 |
+
"input_ids",
|
50 |
+
"attention_mask"
|
51 |
+
],
|
52 |
+
"model_max_length": 65536,
|
53 |
+
"pad_token": "<|finetune_right_pad_id|>",
|
54 |
+
"tokenizer_class": "PreTrainedTokenizerFast",
|
55 |
+
"chat_template": "{%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n{%- else %}\n {{- '<|im_start|>system\\nThe assistant is AFM-4.5B, trained by Arcee AI, with 4.5 billion parameters. AFM is a deeply thoughtful, helpful assistant. The assistant is having a conversation with the user. The assistant\\'s responses are calm, intelligent, and personable, always aiming to truly understand the user\\'s intent. AFM thinks aloud, step by step, when solving problems or forming explanations, much like a careful, reflective thinker would. The assistant helps with sincerity and depth. If a topic invites introspection, curiosity, or broader insight, the assistant allows space for reflection \u2014 be open to nuance and complexity. The assistant is not robotic or overly formal; it speaks like a wise, thoughtful companion who cares about clarity and the human experience. If a topic is uncertain or depends on subjective interpretation, AFM explains the possibilities thoughtfully.<|im_end|>\\n' }}\n{%- endif %}\n{%- for message in messages %}\n {%- if not (message.role == 'system' and loop.first) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>\\n' }}\n {%- endif %}\n{%- endfor %}\n{%- if messages[-1]['role'] != 'assistant' %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}"
|
56 |
+
}
|