Upload config.json with huggingface_hub
Browse files- config.json +20 -20
config.json
CHANGED
|
@@ -88,10 +88,7 @@
|
|
| 88 |
"bits": 6,
|
| 89 |
"group_size": 64
|
| 90 |
},
|
| 91 |
-
"model.layers.2.self_attn.o_proj":
|
| 92 |
-
"bits": 6,
|
| 93 |
-
"group_size": 64
|
| 94 |
-
},
|
| 95 |
"model.layers.2.self_attn.q_norm": false,
|
| 96 |
"model.layers.2.self_attn.k_norm": false,
|
| 97 |
"model.layers.2.self_attn.rope": false,
|
|
@@ -515,7 +512,10 @@
|
|
| 515 |
"bits": 6,
|
| 516 |
"group_size": 64
|
| 517 |
},
|
| 518 |
-
"model.layers.18.self_attn.v_proj":
|
|
|
|
|
|
|
|
|
|
| 519 |
"model.layers.18.self_attn.o_proj": true,
|
| 520 |
"model.layers.18.self_attn.q_norm": false,
|
| 521 |
"model.layers.18.self_attn.k_norm": false,
|
|
@@ -525,10 +525,7 @@
|
|
| 525 |
"bits": 6,
|
| 526 |
"group_size": 64
|
| 527 |
},
|
| 528 |
-
"model.layers.18.mlp.up_proj":
|
| 529 |
-
"bits": 6,
|
| 530 |
-
"group_size": 64
|
| 531 |
-
},
|
| 532 |
"model.layers.18.input_layernorm": false,
|
| 533 |
"model.layers.18.post_attention_layernorm": false,
|
| 534 |
"model.layers.18.pre_feedforward_layernorm": false,
|
|
@@ -555,7 +552,10 @@
|
|
| 555 |
"model.layers.19.post_feedforward_layernorm": false,
|
| 556 |
"model.layers.20.self_attn.q_proj": true,
|
| 557 |
"model.layers.20.self_attn.k_proj": true,
|
| 558 |
-
"model.layers.20.self_attn.v_proj":
|
|
|
|
|
|
|
|
|
|
| 559 |
"model.layers.20.self_attn.o_proj": {
|
| 560 |
"bits": 6,
|
| 561 |
"group_size": 64
|
|
@@ -750,10 +750,7 @@
|
|
| 750 |
"bits": 6,
|
| 751 |
"group_size": 64
|
| 752 |
},
|
| 753 |
-
"model.layers.2.self_attn.o_proj":
|
| 754 |
-
"bits": 6,
|
| 755 |
-
"group_size": 64
|
| 756 |
-
},
|
| 757 |
"model.layers.2.self_attn.q_norm": false,
|
| 758 |
"model.layers.2.self_attn.k_norm": false,
|
| 759 |
"model.layers.2.self_attn.rope": false,
|
|
@@ -1177,7 +1174,10 @@
|
|
| 1177 |
"bits": 6,
|
| 1178 |
"group_size": 64
|
| 1179 |
},
|
| 1180 |
-
"model.layers.18.self_attn.v_proj":
|
|
|
|
|
|
|
|
|
|
| 1181 |
"model.layers.18.self_attn.o_proj": true,
|
| 1182 |
"model.layers.18.self_attn.q_norm": false,
|
| 1183 |
"model.layers.18.self_attn.k_norm": false,
|
|
@@ -1187,10 +1187,7 @@
|
|
| 1187 |
"bits": 6,
|
| 1188 |
"group_size": 64
|
| 1189 |
},
|
| 1190 |
-
"model.layers.18.mlp.up_proj":
|
| 1191 |
-
"bits": 6,
|
| 1192 |
-
"group_size": 64
|
| 1193 |
-
},
|
| 1194 |
"model.layers.18.input_layernorm": false,
|
| 1195 |
"model.layers.18.post_attention_layernorm": false,
|
| 1196 |
"model.layers.18.pre_feedforward_layernorm": false,
|
|
@@ -1217,7 +1214,10 @@
|
|
| 1217 |
"model.layers.19.post_feedforward_layernorm": false,
|
| 1218 |
"model.layers.20.self_attn.q_proj": true,
|
| 1219 |
"model.layers.20.self_attn.k_proj": true,
|
| 1220 |
-
"model.layers.20.self_attn.v_proj":
|
|
|
|
|
|
|
|
|
|
| 1221 |
"model.layers.20.self_attn.o_proj": {
|
| 1222 |
"bits": 6,
|
| 1223 |
"group_size": 64
|
|
|
|
| 88 |
"bits": 6,
|
| 89 |
"group_size": 64
|
| 90 |
},
|
| 91 |
+
"model.layers.2.self_attn.o_proj": true,
|
|
|
|
|
|
|
|
|
|
| 92 |
"model.layers.2.self_attn.q_norm": false,
|
| 93 |
"model.layers.2.self_attn.k_norm": false,
|
| 94 |
"model.layers.2.self_attn.rope": false,
|
|
|
|
| 512 |
"bits": 6,
|
| 513 |
"group_size": 64
|
| 514 |
},
|
| 515 |
+
"model.layers.18.self_attn.v_proj": {
|
| 516 |
+
"bits": 6,
|
| 517 |
+
"group_size": 64
|
| 518 |
+
},
|
| 519 |
"model.layers.18.self_attn.o_proj": true,
|
| 520 |
"model.layers.18.self_attn.q_norm": false,
|
| 521 |
"model.layers.18.self_attn.k_norm": false,
|
|
|
|
| 525 |
"bits": 6,
|
| 526 |
"group_size": 64
|
| 527 |
},
|
| 528 |
+
"model.layers.18.mlp.up_proj": true,
|
|
|
|
|
|
|
|
|
|
| 529 |
"model.layers.18.input_layernorm": false,
|
| 530 |
"model.layers.18.post_attention_layernorm": false,
|
| 531 |
"model.layers.18.pre_feedforward_layernorm": false,
|
|
|
|
| 552 |
"model.layers.19.post_feedforward_layernorm": false,
|
| 553 |
"model.layers.20.self_attn.q_proj": true,
|
| 554 |
"model.layers.20.self_attn.k_proj": true,
|
| 555 |
+
"model.layers.20.self_attn.v_proj": {
|
| 556 |
+
"bits": 6,
|
| 557 |
+
"group_size": 64
|
| 558 |
+
},
|
| 559 |
"model.layers.20.self_attn.o_proj": {
|
| 560 |
"bits": 6,
|
| 561 |
"group_size": 64
|
|
|
|
| 750 |
"bits": 6,
|
| 751 |
"group_size": 64
|
| 752 |
},
|
| 753 |
+
"model.layers.2.self_attn.o_proj": true,
|
|
|
|
|
|
|
|
|
|
| 754 |
"model.layers.2.self_attn.q_norm": false,
|
| 755 |
"model.layers.2.self_attn.k_norm": false,
|
| 756 |
"model.layers.2.self_attn.rope": false,
|
|
|
|
| 1174 |
"bits": 6,
|
| 1175 |
"group_size": 64
|
| 1176 |
},
|
| 1177 |
+
"model.layers.18.self_attn.v_proj": {
|
| 1178 |
+
"bits": 6,
|
| 1179 |
+
"group_size": 64
|
| 1180 |
+
},
|
| 1181 |
"model.layers.18.self_attn.o_proj": true,
|
| 1182 |
"model.layers.18.self_attn.q_norm": false,
|
| 1183 |
"model.layers.18.self_attn.k_norm": false,
|
|
|
|
| 1187 |
"bits": 6,
|
| 1188 |
"group_size": 64
|
| 1189 |
},
|
| 1190 |
+
"model.layers.18.mlp.up_proj": true,
|
|
|
|
|
|
|
|
|
|
| 1191 |
"model.layers.18.input_layernorm": false,
|
| 1192 |
"model.layers.18.post_attention_layernorm": false,
|
| 1193 |
"model.layers.18.pre_feedforward_layernorm": false,
|
|
|
|
| 1214 |
"model.layers.19.post_feedforward_layernorm": false,
|
| 1215 |
"model.layers.20.self_attn.q_proj": true,
|
| 1216 |
"model.layers.20.self_attn.k_proj": true,
|
| 1217 |
+
"model.layers.20.self_attn.v_proj": {
|
| 1218 |
+
"bits": 6,
|
| 1219 |
+
"group_size": 64
|
| 1220 |
+
},
|
| 1221 |
"model.layers.20.self_attn.o_proj": {
|
| 1222 |
"bits": 6,
|
| 1223 |
"group_size": 64
|