Xenova HF Staff commited on
Commit
a2c667d
·
verified ·
1 Parent(s): 3ead5ba

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ onnx/model_q4.onnx_data filter=lfs diff=lfs merge=lfs -text
37
+ onnx/model_q4.onnx_data_1 filter=lfs diff=lfs merge=lfs -text
38
+ onnx/model_q4f16.onnx_data filter=lfs diff=lfs merge=lfs -text
39
+ onnx/model_q4f16.onnx_data_1 filter=lfs diff=lfs merge=lfs -text
40
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
chat_template.jinja ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if messages[0]['role'] == 'system' %}
2
+ {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }}
3
+ {%- else %}
4
+ {{- '<|im_start|>system\nThe assistant is AFM-4.5B, trained by Arcee AI, with 4.5 billion parameters. AFM is a deeply thoughtful, helpful assistant. The assistant is having a conversation with the user. The assistant\'s responses are calm, intelligent, and personable, always aiming to truly understand the user\'s intent. AFM thinks aloud, step by step, when solving problems or forming explanations, much like a careful, reflective thinker would. The assistant helps with sincerity and depth. If a topic invites introspection, curiosity, or broader insight, the assistant allows space for reflection — be open to nuance and complexity. The assistant is not robotic or overly formal; it speaks like a wise, thoughtful companion who cares about clarity and the human experience. If a topic is uncertain or depends on subjective interpretation, AFM explains the possibilities thoughtfully.<|im_end|>\n' }}
5
+ {%- endif %}
6
+ {%- for message in messages %}
7
+ {%- if not (message.role == 'system' and loop.first) %}
8
+ {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>\n' }}
9
+ {%- endif %}
10
+ {%- endfor %}
11
+ {%- if messages[-1]['role'] != 'assistant' %}
12
+ {{- '<|im_start|>assistant\n' }}
13
+ {%- endif %}
config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ArceeForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 128000,
8
+ "eos_token_id": 128003,
9
+ "head_dim": 128,
10
+ "hidden_act": "relu2",
11
+ "hidden_size": 2560,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 18432,
14
+ "max_position_embeddings": 65536,
15
+ "mlp_bias": false,
16
+ "model_type": "arcee",
17
+ "num_attention_heads": 20,
18
+ "num_hidden_layers": 36,
19
+ "num_key_value_heads": 4,
20
+ "rms_norm_eps": 1e-05,
21
+ "rope_scaling": {
22
+ "beta_fast": 32.0,
23
+ "beta_slow": 1.0,
24
+ "factor": 20.0,
25
+ "mscale": 1.0,
26
+ "original_max_position_embeddings": 4096,
27
+ "rope_type": "yarn",
28
+ "type": "yarn"
29
+ },
30
+ "rope_theta": 10000.0,
31
+ "tie_word_embeddings": false,
32
+ "torch_dtype": "bfloat16",
33
+ "transformers_version": "4.54.0.dev0",
34
+ "use_cache": false,
35
+ "vocab_size": 128005,
36
+ "transformers.js_config": {
37
+ "use_external_data_format": {
38
+ "model_q4.onnx": 2,
39
+ "model_q4f16.onnx": 2
40
+ },
41
+ "kv_cache_dtype": {
42
+ "q4f16": "float16",
43
+ "fp16": "float16"
44
+ }
45
+ }
46
+ }
onnx/model_q4.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1af24d2bdd7e7927b9509c9cbd06aa86e02796f276d1addc08e4a113f04c3bcc
3
+ size 321381
onnx/model_q4.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9faec389f05d81f631f3b4f24280979fa412947867d2c11ae97b7735e36c439c
3
+ size 2097055552
onnx/model_q4.onnx_data_1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:397d518a0e358168492b4a0f2041a653ca50d63e81e091f96228ff424dccda98
3
+ size 824123200
onnx/model_q4f16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ebb79e862018ad77e174bb2bac0462f4d29a26c8844a00bc12d443eb45f40cd
3
+ size 322886
onnx/model_q4f16.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8c1118ca6877357989e11b1fbb4c7a8c59eea73a349caa3604e1a145ffb2ecf
3
+ size 2094813216
onnx/model_q4f16.onnx_data_1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a76ca33a5695c49696110d09ff76a19f43fbe1971db9a54867e355d9fb9908c
3
+ size 520526880
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|begin_of_text|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|im_end|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|finetune_right_pad_id|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1363c79c75bf4202ce0f7ca44828b99a72140b97340578fa3b8927dddc46da2e
3
+ size 11535885
tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "128000": {
4
+ "content": "<|begin_of_text|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "128001": {
12
+ "content": "<|end_of_text|>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "128002": {
20
+ "content": "<|im_start|>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "128003": {
28
+ "content": "<|im_end|>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128004": {
36
+ "content": "<|finetune_right_pad_id|>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "<|begin_of_text|>",
45
+ "clean_up_tokenization_spaces": true,
46
+ "eos_token": "<|im_end|>",
47
+ "extra_special_tokens": {},
48
+ "model_input_names": [
49
+ "input_ids",
50
+ "attention_mask"
51
+ ],
52
+ "model_max_length": 65536,
53
+ "pad_token": "<|finetune_right_pad_id|>",
54
+ "tokenizer_class": "PreTrainedTokenizerFast",
55
+ "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n{%- else %}\n {{- '<|im_start|>system\\nThe assistant is AFM-4.5B, trained by Arcee AI, with 4.5 billion parameters. AFM is a deeply thoughtful, helpful assistant. The assistant is having a conversation with the user. The assistant\\'s responses are calm, intelligent, and personable, always aiming to truly understand the user\\'s intent. AFM thinks aloud, step by step, when solving problems or forming explanations, much like a careful, reflective thinker would. The assistant helps with sincerity and depth. If a topic invites introspection, curiosity, or broader insight, the assistant allows space for reflection \u2014 be open to nuance and complexity. The assistant is not robotic or overly formal; it speaks like a wise, thoughtful companion who cares about clarity and the human experience. If a topic is uncertain or depends on subjective interpretation, AFM explains the possibilities thoughtfully.<|im_end|>\\n' }}\n{%- endif %}\n{%- for message in messages %}\n {%- if not (message.role == 'system' and loop.first) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>\\n' }}\n {%- endif %}\n{%- endfor %}\n{%- if messages[-1]['role'] != 'assistant' %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}"
56
+ }