jerry3972 commited on
Commit
5d23e92
·
1 Parent(s): c588bed

모델 업데이트

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "\"UNIVA-Bllossom/DeepSeek-llama3.1-Bllossom-8",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -29,8 +29,8 @@
29
  },
30
  "rope_theta": 500000.0,
31
  "tie_word_embeddings": false,
32
- "torch_dtype": "bfloat16",
33
- "transformers_version": "4.46.3",
34
  "use_cache": false,
35
- "vocab_size": 128257
36
  }
 
1
  {
2
+ "_name_or_path": "/home/raid/models/24UNSPC_Tulkun_v3.0.4/",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
29
  },
30
  "rope_theta": 500000.0,
31
  "tie_word_embeddings": false,
32
+ "torch_dtype": "float16",
33
+ "transformers_version": "4.48.3",
34
  "use_cache": false,
35
+ "vocab_size": 128256
36
  }
generation_config.json CHANGED
@@ -5,5 +5,5 @@
5
  "eos_token_id": 128001,
6
  "temperature": 0.6,
7
  "top_p": 0.95,
8
- "transformers_version": "4.46.3"
9
  }
 
5
  "eos_token_id": 128001,
6
  "temperature": 0.6,
7
  "top_p": 0.95,
8
+ "transformers_version": "4.48.3"
9
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43aeb40dea39d1879925512c47c4e96047a5d7b92f50dc994872e441f97aab5b
3
- size 4976706864
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:939bc0a77e349a9896e2374abcb481ed147b042c49dc0b527339d967f01a2fd1
3
+ size 4976698592
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0c99f508a7870d7036720338601b1b6e991296b88594366c6fa9c606c7c1a76
3
- size 4999802720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9eca7f9936d35bc42c01038d26ff89bd26fab8b79a093c6aa19deb66ec851e5
3
+ size 4999802616
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c80b287c1541e2bb9fa18ef82751a0b82f9d6560f2945fb56b6ef9ca7b2eb27
3
- size 4915916176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0d23c52ce9197c04995fc64682e120e4ec47bbe0cf03c6d7c576b7efa3f2058
3
+ size 4915916080
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e66973340e2b2c962618eb6db28a820442d1829e9e473eb819d56d2d9430cec8
3
- size 1168147000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e318aacfeffb95c600b62f1d9a224f9ada79f365b7746d5449363e0d81abc64
3
+ size 1168138808
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 16060538880
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00004-of-00004.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 16060522496
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00004-of-00004.safetensors",
special_tokens_map.json CHANGED
@@ -14,7 +14,7 @@
14
  "single_word": false
15
  },
16
  "pad_token": {
17
- "content": "<|end_of_text|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
 
14
  "single_word": false
15
  },
16
  "pad_token": {
17
+ "content": "<|end▁of▁sentence|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16f2ebc8d9a7de55360d83ea69f97916a1389f0a72264664d4d6c4db6da8d0b8
3
- size 17209722
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d91915040cfac999d8c55f4b5bc6e67367c065e3a7a4e4b9438ce1f256addd86
3
+ size 17209530
tokenizer_config.json CHANGED
@@ -2050,24 +2050,16 @@
2050
  "rstrip": false,
2051
  "single_word": false,
2052
  "special": true
2053
- },
2054
- "128256": {
2055
- "content": "<|end_of_text|>",
2056
- "lstrip": false,
2057
- "normalized": false,
2058
- "rstrip": false,
2059
- "single_word": false,
2060
- "special": true
2061
  }
2062
  },
2063
  "bos_token": "<|begin▁of▁sentence|>",
2064
- "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}",
2065
  "clean_up_tokenization_spaces": false,
2066
  "eos_token": "<|end▁of▁sentence|>",
2067
  "extra_special_tokens": {},
2068
  "legacy": true,
2069
  "model_max_length": 16384,
2070
- "pad_token": "<|end_of_text|>",
2071
  "sp_model_kwargs": {},
2072
  "tokenizer_class": "LlamaTokenizer",
2073
  "unk_token": null,
 
2050
  "rstrip": false,
2051
  "single_word": false,
2052
  "special": true
 
 
 
 
 
 
 
 
2053
  }
2054
  },
2055
  "bos_token": "<|begin▁of▁sentence|>",
2056
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set ns = namespace(is_first=false, is_tool=false, is_output_first=true, system_prompt='') %}{%- for message in messages %}{%- if message['role'] == 'system' %}{% set ns.system_prompt = message['content'] %}{%- endif %}{%- endfor %}{{bos_token}}{{ns.system_prompt}}{%- for message in messages %}{%- if message['role'] == 'user' %}{%- set ns.is_tool = false -%}{{'<|User|>' + message['content']}}{%- endif %}{%- if message['role'] == 'assistant' and message['content'] is none %}{%- set ns.is_tool = false -%}{%- for tool in message['tool_calls']%}{%- if not ns.is_first %}{{'<|Assistant|><|tool▁calls▁begin|><|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\\n' + '```json' + '\\n' + tool['function']['arguments'] + '\\n' + '```' + '<|tool▁call▁end|>'}}{%- set ns.is_first = true -%}{%- else %}{{'\\n' + '<|tool▁call▁begin|>' + tool['type'] + '<|tool▁sep|>' + tool['function']['name'] + '\\n' + '```json' + '\\n' + tool['function']['arguments'] + '\\n' + '```' + '<|tool▁call▁end|>'}}{{'<|tool▁calls▁end|><|end▁of▁sentence|>'}}{%- endif %}{%- endfor %}{%- endif %}{%- if message['role'] == 'assistant' and message['content'] is not none %}{%- if ns.is_tool %}{{'<|tool▁outputs▁end|>' + message['content'] + '<|end▁of▁sentence|>'}}{%- set ns.is_tool = false -%}{%- else %}{% set content = message['content'] %}{% if '</think>' in content %}{% set content = content.split('</think>')[-1] %}{% endif %}{{'<|Assistant|>' + content + '<|end▁of▁sentence|>'}}{%- endif %}{%- endif %}{%- if message['role'] == 'tool' %}{%- set ns.is_tool = true -%}{%- if ns.is_output_first %}{{'<|tool▁outputs▁begin|><|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}}{%- set ns.is_output_first = false %}{%- else %}{{'\\n<|tool▁output▁begin|>' + message['content'] + '<|tool▁output▁end|>'}}{%- endif %}{%- endif %}{%- endfor -%}{% if ns.is_tool %}{{'<|tool▁outputs▁end|>'}}{% endif %}{% if add_generation_prompt and not ns.is_tool %}{{'<|Assistant|>'}}{% endif %}",
2057
  "clean_up_tokenization_spaces": false,
2058
  "eos_token": "<|end▁of▁sentence|>",
2059
  "extra_special_tokens": {},
2060
  "legacy": true,
2061
  "model_max_length": 16384,
2062
+ "pad_token": "<|end▁of▁sentence|>",
2063
  "sp_model_kwargs": {},
2064
  "tokenizer_class": "LlamaTokenizer",
2065
  "unk_token": null,