yahavb commited on
Commit
bf643eb
·
verified ·
1 Parent(s): 2935ce4

Add NxD compiled model yahavb/inf2-bs16-tp16-mml16k-llama-31-8b-vllm for vLLM; after converting checkpoints

Browse files
model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:108913ff5274c9113516d44d542b525d618ba6ff49d2467d3f03a0f0694060eb
3
+ size 46756107
neuron_config.json ADDED
@@ -0,0 +1,190 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": false,
3
+ "_name_or_path": "meta-llama/Llama-3.1-8B",
4
+ "add_cross_attention": false,
5
+ "architectures": [
6
+ "LlamaForCausalLM"
7
+ ],
8
+ "attention_bias": false,
9
+ "attention_dropout": 0.0,
10
+ "attribute_map": {},
11
+ "bad_words_ids": null,
12
+ "begin_suppress_tokens": null,
13
+ "bos_token_id": 128000,
14
+ "chunk_size_feed_forward": 0,
15
+ "cross_attention_hidden_size": null,
16
+ "decoder_start_token_id": null,
17
+ "diversity_penalty": 0.0,
18
+ "do_sample": false,
19
+ "early_stopping": false,
20
+ "encoder_no_repeat_ngram_size": 0,
21
+ "eos_token_id": 128001,
22
+ "exponential_decay_length_penalty": null,
23
+ "finetuning_task": null,
24
+ "forced_bos_token_id": null,
25
+ "forced_eos_token_id": null,
26
+ "fused_spec_config": null,
27
+ "head_dim": 128,
28
+ "hidden_act": "silu",
29
+ "hidden_size": 4096,
30
+ "id2label": {
31
+ "0": "LABEL_0",
32
+ "1": "LABEL_1"
33
+ },
34
+ "initializer_range": 0.02,
35
+ "intermediate_size": 14336,
36
+ "is_decoder": false,
37
+ "is_encoder_decoder": false,
38
+ "label2id": {
39
+ "LABEL_0": 0,
40
+ "LABEL_1": 1
41
+ },
42
+ "length_penalty": 1.0,
43
+ "max_length": 20,
44
+ "max_position_embeddings": 131072,
45
+ "min_length": 0,
46
+ "mlp_bias": false,
47
+ "model_type": "llama",
48
+ "neuron_config": {
49
+ "activation_quantization_type": null,
50
+ "allow_input_truncation": false,
51
+ "async_mode": false,
52
+ "attn_cls": "NeuronLlamaAttention",
53
+ "attn_kernel_enabled": false,
54
+ "batch_size": 16,
55
+ "bucket_n_active_tokens": false,
56
+ "buckets": [
57
+ 16000
58
+ ],
59
+ "cc_pipeline_tiling_factor": 2,
60
+ "context_encoding_buckets": null,
61
+ "cp_max_num_seqs": 0,
62
+ "cp_num_active_blocks": 0,
63
+ "ctx_batch_size": 1,
64
+ "draft_model_modules_to_not_convert": null,
65
+ "enable_bucketing": true,
66
+ "enable_eagle_draft_input_norm": false,
67
+ "enable_eagle_speculation": false,
68
+ "enable_fused_speculation": false,
69
+ "enable_token_tree": false,
70
+ "ep_degree": 1,
71
+ "flash_decoding_enabled": false,
72
+ "fused_qkv": false,
73
+ "is_block_kv_layout": false,
74
+ "is_chunked_prefill": false,
75
+ "is_continuous_batching": true,
76
+ "is_eagle_draft": false,
77
+ "is_medusa": false,
78
+ "is_prefill_stage": null,
79
+ "is_prefix_caching": false,
80
+ "kv_cache_batch_size": 16,
81
+ "kv_cache_padding_size": 0,
82
+ "kv_cache_quant": false,
83
+ "kv_cache_tiling": false,
84
+ "local_ranks_size": 16,
85
+ "logical_nc_config": 1,
86
+ "lora_config": null,
87
+ "max_batch_size": 16,
88
+ "max_context_length": 16000,
89
+ "max_length": 16000,
90
+ "max_new_tokens": null,
91
+ "medusa_speculation_length": 0,
92
+ "medusa_tree": null,
93
+ "mlp_kernel_enabled": false,
94
+ "mlp_kernel_fuse_residual_add": false,
95
+ "modules_to_not_convert": null,
96
+ "n_active_tokens": 16000,
97
+ "n_positions": 16000,
98
+ "num_medusa_heads": 0,
99
+ "on_cpu": false,
100
+ "on_device_sampling_config": {
101
+ "deterministic": false,
102
+ "do_sample": false,
103
+ "dynamic": true,
104
+ "global_topk": 256,
105
+ "on_device_sampling_config": true,
106
+ "temperature": 1.0,
107
+ "top_k": 1,
108
+ "top_p": 1.0
109
+ },
110
+ "output_logits": false,
111
+ "overrides_torch_dtype": true,
112
+ "pa_block_size": 16000,
113
+ "pa_num_blocks": 16,
114
+ "padding_side": "right",
115
+ "pp_degree": 1,
116
+ "qk_layernorm": false,
117
+ "qkv_kernel_enabled": false,
118
+ "qkv_kernel_nbsd_layout": false,
119
+ "quantization_dtype": "int8",
120
+ "quantization_type": "per_tensor_symmetric",
121
+ "quantize_clamp_bound": Infinity,
122
+ "quantized": false,
123
+ "quantized_checkpoints_path": null,
124
+ "quantized_mlp_kernel_enabled": false,
125
+ "rmsnorm_quantize_kernel_enabled": false,
126
+ "rpl_reduce_dtype": "bfloat16",
127
+ "save_sharded_checkpoint": true,
128
+ "seq_len": 16000,
129
+ "sequence_parallel_enabled": false,
130
+ "skip_sharding": false,
131
+ "skip_warmup": true,
132
+ "spec_batch_size": 16,
133
+ "speculation_length": 0,
134
+ "start_rank_id": 0,
135
+ "target": null,
136
+ "tkg_batch_size": 16,
137
+ "token_generation_buckets": null,
138
+ "token_tree_config": null,
139
+ "torch_dtype": "bfloat16",
140
+ "tp_degree": 16,
141
+ "vocab_parallel": false,
142
+ "weights_to_skip_layout_optimization": [],
143
+ "world_size": 16
144
+ },
145
+ "no_repeat_ngram_size": 0,
146
+ "num_attention_heads": 32,
147
+ "num_beam_groups": 1,
148
+ "num_beams": 1,
149
+ "num_cores_per_group": 1,
150
+ "num_hidden_layers": 32,
151
+ "num_key_value_heads": 8,
152
+ "num_return_sequences": 1,
153
+ "output_attentions": false,
154
+ "output_hidden_states": false,
155
+ "output_scores": false,
156
+ "pad_token_id": null,
157
+ "prefix": null,
158
+ "pretraining_tp": 1,
159
+ "problem_type": null,
160
+ "pruned_heads": {},
161
+ "remove_invalid_values": false,
162
+ "repetition_penalty": 1.0,
163
+ "return_dict": true,
164
+ "return_dict_in_generate": false,
165
+ "rms_norm_eps": 1e-05,
166
+ "rope_scaling": {
167
+ "factor": 8.0,
168
+ "high_freq_factor": 4.0,
169
+ "low_freq_factor": 1.0,
170
+ "original_max_position_embeddings": 8192,
171
+ "rope_type": "llama3"
172
+ },
173
+ "rope_theta": 500000.0,
174
+ "sep_token_id": null,
175
+ "suppress_tokens": null,
176
+ "task_specific_params": null,
177
+ "temperature": 1.0,
178
+ "tf_legacy_loss": false,
179
+ "tie_encoder_decoder": false,
180
+ "tie_word_embeddings": false,
181
+ "tokenizer_class": null,
182
+ "top_k": 50,
183
+ "top_p": 1.0,
184
+ "torchscript": false,
185
+ "transformers_version": "4.51.3",
186
+ "typical_p": 1.0,
187
+ "use_bfloat16": false,
188
+ "use_cache": true,
189
+ "vocab_size": 128256
190
+ }
weights/tp0_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7bb12500fd6c16dda236fa16ee04dcf5026aadec4a58e256fe18a035a545f11
3
+ size 1037871572
weights/tp10_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c5a71ae97d71a0b61c0e8f27547ec83f7406d90b510b468bdd4c5633f596ceb
3
+ size 1037871572
weights/tp11_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62e0c2bf58b2368a027a661ae1235514d90940b64b82a14fa4c72bfeb1036d67
3
+ size 1037871572
weights/tp12_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1557496e5d45afa32ce8a046380f7f1f3d741c2b5c70555389d01e1058847c47
3
+ size 1037871572
weights/tp13_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:085323cfd2fbc174b4615fb283e355e7afa16ccc3e9ed2037e48c67cae34f29c
3
+ size 1037871572
weights/tp14_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fc0b0c00f8f3512c79212d06752d740a1d7c80792524f8eb5cee72fc67449af
3
+ size 1037871572
weights/tp15_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c5a72991e0407239300fa1518d1a28ced56a02f5bfe3c5beaa9b056456836a5
3
+ size 1037871572
weights/tp1_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb99b9d976ad73343700e912fad141983c619d6c49f89d803502a1734daba641
3
+ size 1037871572
weights/tp2_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57edaf05f8989b98975c6597a6bae79bf9662955d64801ff8617e67b716d0aec
3
+ size 1037871572
weights/tp3_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccaef67271952cbff9031cb9a3b50025da9e563e6ac9f0b956ffff9fddc98e48
3
+ size 1037871572
weights/tp4_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a90e3888052d6280799eb6c17d923a123847147e3915031b7a40de0b1eefb64f
3
+ size 1037871572
weights/tp5_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60853b64411ec6ad0067c9e6f4c2c59d4c5e734fe1078da82f052c792f510269
3
+ size 1037871572
weights/tp6_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:034589ecc8dabbab1bbe94ad1eba9c00fc856c10c628049f7f0d5a373a016cd7
3
+ size 1037871572
weights/tp7_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64b2ab8983a30b169601806d78cd1d5d95ecbe6a06663e6f54b13285104469f3
3
+ size 1037871572
weights/tp8_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b571076458b472e2d3b70986f1e744067926ed0237d41e12c239e8fe0bdc9bc0
3
+ size 1037871572
weights/tp9_sharded_checkpoint.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f011fd6d9f2515aeb5de105ae322e1e61e4459961e666cd10e416b1ee6e3d849
3
+ size 1037871572