Update config.json
Browse files- config.json +224 -224
config.json
CHANGED
@@ -31,7 +31,7 @@
|
|
31 |
"model.layers.0.mlp.down_proj": {
|
32 |
"bias": false,
|
33 |
"enable_norm": true,
|
34 |
-
"enable_perm":
|
35 |
"group_num": 1,
|
36 |
"group_size": 14192,
|
37 |
"in_features": 14336,
|
@@ -55,7 +55,7 @@
|
|
55 |
"model.layers.0.mlp.gate_proj": {
|
56 |
"bias": false,
|
57 |
"enable_norm": true,
|
58 |
-
"enable_perm":
|
59 |
"group_num": 1,
|
60 |
"group_size": 4055,
|
61 |
"in_features": 4096,
|
@@ -79,7 +79,7 @@
|
|
79 |
"model.layers.0.mlp.up_proj": {
|
80 |
"bias": false,
|
81 |
"enable_norm": true,
|
82 |
-
"enable_perm":
|
83 |
"group_num": 1,
|
84 |
"group_size": 4055,
|
85 |
"in_features": 4096,
|
@@ -103,7 +103,7 @@
|
|
103 |
"model.layers.0.self_attn.k_proj": {
|
104 |
"bias": false,
|
105 |
"enable_norm": true,
|
106 |
-
"enable_perm":
|
107 |
"group_num": 1,
|
108 |
"group_size": 4055,
|
109 |
"in_features": 4096,
|
@@ -127,7 +127,7 @@
|
|
127 |
"model.layers.0.self_attn.o_proj": {
|
128 |
"bias": false,
|
129 |
"enable_norm": true,
|
130 |
-
"enable_perm":
|
131 |
"group_num": 1,
|
132 |
"group_size": 4055,
|
133 |
"in_features": 4096,
|
@@ -151,7 +151,7 @@
|
|
151 |
"model.layers.0.self_attn.q_proj": {
|
152 |
"bias": false,
|
153 |
"enable_norm": true,
|
154 |
-
"enable_perm":
|
155 |
"group_num": 1,
|
156 |
"group_size": 4055,
|
157 |
"in_features": 4096,
|
@@ -175,7 +175,7 @@
|
|
175 |
"model.layers.0.self_attn.v_proj": {
|
176 |
"bias": false,
|
177 |
"enable_norm": true,
|
178 |
-
"enable_perm":
|
179 |
"group_num": 1,
|
180 |
"group_size": 4055,
|
181 |
"in_features": 4096,
|
@@ -199,7 +199,7 @@
|
|
199 |
"model.layers.1.mlp.down_proj": {
|
200 |
"bias": false,
|
201 |
"enable_norm": true,
|
202 |
-
"enable_perm":
|
203 |
"group_num": 1,
|
204 |
"group_size": 14192,
|
205 |
"in_features": 14336,
|
@@ -223,7 +223,7 @@
|
|
223 |
"model.layers.1.mlp.gate_proj": {
|
224 |
"bias": false,
|
225 |
"enable_norm": true,
|
226 |
-
"enable_perm":
|
227 |
"group_num": 1,
|
228 |
"group_size": 4055,
|
229 |
"in_features": 4096,
|
@@ -247,7 +247,7 @@
|
|
247 |
"model.layers.1.mlp.up_proj": {
|
248 |
"bias": false,
|
249 |
"enable_norm": true,
|
250 |
-
"enable_perm":
|
251 |
"group_num": 1,
|
252 |
"group_size": 4055,
|
253 |
"in_features": 4096,
|
@@ -271,7 +271,7 @@
|
|
271 |
"model.layers.1.self_attn.k_proj": {
|
272 |
"bias": false,
|
273 |
"enable_norm": true,
|
274 |
-
"enable_perm":
|
275 |
"group_num": 1,
|
276 |
"group_size": 4055,
|
277 |
"in_features": 4096,
|
@@ -295,7 +295,7 @@
|
|
295 |
"model.layers.1.self_attn.o_proj": {
|
296 |
"bias": false,
|
297 |
"enable_norm": true,
|
298 |
-
"enable_perm":
|
299 |
"group_num": 1,
|
300 |
"group_size": 4055,
|
301 |
"in_features": 4096,
|
@@ -319,7 +319,7 @@
|
|
319 |
"model.layers.1.self_attn.q_proj": {
|
320 |
"bias": false,
|
321 |
"enable_norm": true,
|
322 |
-
"enable_perm":
|
323 |
"group_num": 1,
|
324 |
"group_size": 4055,
|
325 |
"in_features": 4096,
|
@@ -343,7 +343,7 @@
|
|
343 |
"model.layers.1.self_attn.v_proj": {
|
344 |
"bias": false,
|
345 |
"enable_norm": true,
|
346 |
-
"enable_perm":
|
347 |
"group_num": 1,
|
348 |
"group_size": 4055,
|
349 |
"in_features": 4096,
|
@@ -367,7 +367,7 @@
|
|
367 |
"model.layers.10.mlp.down_proj": {
|
368 |
"bias": false,
|
369 |
"enable_norm": true,
|
370 |
-
"enable_perm":
|
371 |
"group_num": 1,
|
372 |
"group_size": 14192,
|
373 |
"in_features": 14336,
|
@@ -391,7 +391,7 @@
|
|
391 |
"model.layers.10.mlp.gate_proj": {
|
392 |
"bias": false,
|
393 |
"enable_norm": true,
|
394 |
-
"enable_perm":
|
395 |
"group_num": 1,
|
396 |
"group_size": 4055,
|
397 |
"in_features": 4096,
|
@@ -415,7 +415,7 @@
|
|
415 |
"model.layers.10.mlp.up_proj": {
|
416 |
"bias": false,
|
417 |
"enable_norm": true,
|
418 |
-
"enable_perm":
|
419 |
"group_num": 1,
|
420 |
"group_size": 4055,
|
421 |
"in_features": 4096,
|
@@ -439,7 +439,7 @@
|
|
439 |
"model.layers.10.self_attn.k_proj": {
|
440 |
"bias": false,
|
441 |
"enable_norm": true,
|
442 |
-
"enable_perm":
|
443 |
"group_num": 1,
|
444 |
"group_size": 4055,
|
445 |
"in_features": 4096,
|
@@ -463,7 +463,7 @@
|
|
463 |
"model.layers.10.self_attn.o_proj": {
|
464 |
"bias": false,
|
465 |
"enable_norm": true,
|
466 |
-
"enable_perm":
|
467 |
"group_num": 1,
|
468 |
"group_size": 4055,
|
469 |
"in_features": 4096,
|
@@ -487,7 +487,7 @@
|
|
487 |
"model.layers.10.self_attn.q_proj": {
|
488 |
"bias": false,
|
489 |
"enable_norm": true,
|
490 |
-
"enable_perm":
|
491 |
"group_num": 1,
|
492 |
"group_size": 4055,
|
493 |
"in_features": 4096,
|
@@ -511,7 +511,7 @@
|
|
511 |
"model.layers.10.self_attn.v_proj": {
|
512 |
"bias": false,
|
513 |
"enable_norm": true,
|
514 |
-
"enable_perm":
|
515 |
"group_num": 1,
|
516 |
"group_size": 4055,
|
517 |
"in_features": 4096,
|
@@ -535,7 +535,7 @@
|
|
535 |
"model.layers.11.mlp.down_proj": {
|
536 |
"bias": false,
|
537 |
"enable_norm": true,
|
538 |
-
"enable_perm":
|
539 |
"group_num": 1,
|
540 |
"group_size": 14192,
|
541 |
"in_features": 14336,
|
@@ -559,7 +559,7 @@
|
|
559 |
"model.layers.11.mlp.gate_proj": {
|
560 |
"bias": false,
|
561 |
"enable_norm": true,
|
562 |
-
"enable_perm":
|
563 |
"group_num": 1,
|
564 |
"group_size": 4055,
|
565 |
"in_features": 4096,
|
@@ -583,7 +583,7 @@
|
|
583 |
"model.layers.11.mlp.up_proj": {
|
584 |
"bias": false,
|
585 |
"enable_norm": true,
|
586 |
-
"enable_perm":
|
587 |
"group_num": 1,
|
588 |
"group_size": 4055,
|
589 |
"in_features": 4096,
|
@@ -607,7 +607,7 @@
|
|
607 |
"model.layers.11.self_attn.k_proj": {
|
608 |
"bias": false,
|
609 |
"enable_norm": true,
|
610 |
-
"enable_perm":
|
611 |
"group_num": 1,
|
612 |
"group_size": 4055,
|
613 |
"in_features": 4096,
|
@@ -631,7 +631,7 @@
|
|
631 |
"model.layers.11.self_attn.o_proj": {
|
632 |
"bias": false,
|
633 |
"enable_norm": true,
|
634 |
-
"enable_perm":
|
635 |
"group_num": 1,
|
636 |
"group_size": 4055,
|
637 |
"in_features": 4096,
|
@@ -655,7 +655,7 @@
|
|
655 |
"model.layers.11.self_attn.q_proj": {
|
656 |
"bias": false,
|
657 |
"enable_norm": true,
|
658 |
-
"enable_perm":
|
659 |
"group_num": 1,
|
660 |
"group_size": 4055,
|
661 |
"in_features": 4096,
|
@@ -679,7 +679,7 @@
|
|
679 |
"model.layers.11.self_attn.v_proj": {
|
680 |
"bias": false,
|
681 |
"enable_norm": true,
|
682 |
-
"enable_perm":
|
683 |
"group_num": 1,
|
684 |
"group_size": 4055,
|
685 |
"in_features": 4096,
|
@@ -703,7 +703,7 @@
|
|
703 |
"model.layers.12.mlp.down_proj": {
|
704 |
"bias": false,
|
705 |
"enable_norm": true,
|
706 |
-
"enable_perm":
|
707 |
"group_num": 1,
|
708 |
"group_size": 14192,
|
709 |
"in_features": 14336,
|
@@ -727,7 +727,7 @@
|
|
727 |
"model.layers.12.mlp.gate_proj": {
|
728 |
"bias": false,
|
729 |
"enable_norm": true,
|
730 |
-
"enable_perm":
|
731 |
"group_num": 1,
|
732 |
"group_size": 4055,
|
733 |
"in_features": 4096,
|
@@ -751,7 +751,7 @@
|
|
751 |
"model.layers.12.mlp.up_proj": {
|
752 |
"bias": false,
|
753 |
"enable_norm": true,
|
754 |
-
"enable_perm":
|
755 |
"group_num": 1,
|
756 |
"group_size": 4055,
|
757 |
"in_features": 4096,
|
@@ -775,7 +775,7 @@
|
|
775 |
"model.layers.12.self_attn.k_proj": {
|
776 |
"bias": false,
|
777 |
"enable_norm": true,
|
778 |
-
"enable_perm":
|
779 |
"group_num": 1,
|
780 |
"group_size": 4055,
|
781 |
"in_features": 4096,
|
@@ -799,7 +799,7 @@
|
|
799 |
"model.layers.12.self_attn.o_proj": {
|
800 |
"bias": false,
|
801 |
"enable_norm": true,
|
802 |
-
"enable_perm":
|
803 |
"group_num": 1,
|
804 |
"group_size": 4055,
|
805 |
"in_features": 4096,
|
@@ -823,7 +823,7 @@
|
|
823 |
"model.layers.12.self_attn.q_proj": {
|
824 |
"bias": false,
|
825 |
"enable_norm": true,
|
826 |
-
"enable_perm":
|
827 |
"group_num": 1,
|
828 |
"group_size": 4055,
|
829 |
"in_features": 4096,
|
@@ -847,7 +847,7 @@
|
|
847 |
"model.layers.12.self_attn.v_proj": {
|
848 |
"bias": false,
|
849 |
"enable_norm": true,
|
850 |
-
"enable_perm":
|
851 |
"group_num": 1,
|
852 |
"group_size": 4055,
|
853 |
"in_features": 4096,
|
@@ -871,7 +871,7 @@
|
|
871 |
"model.layers.13.mlp.down_proj": {
|
872 |
"bias": false,
|
873 |
"enable_norm": true,
|
874 |
-
"enable_perm":
|
875 |
"group_num": 1,
|
876 |
"group_size": 14192,
|
877 |
"in_features": 14336,
|
@@ -895,7 +895,7 @@
|
|
895 |
"model.layers.13.mlp.gate_proj": {
|
896 |
"bias": false,
|
897 |
"enable_norm": true,
|
898 |
-
"enable_perm":
|
899 |
"group_num": 1,
|
900 |
"group_size": 4055,
|
901 |
"in_features": 4096,
|
@@ -919,7 +919,7 @@
|
|
919 |
"model.layers.13.mlp.up_proj": {
|
920 |
"bias": false,
|
921 |
"enable_norm": true,
|
922 |
-
"enable_perm":
|
923 |
"group_num": 1,
|
924 |
"group_size": 4055,
|
925 |
"in_features": 4096,
|
@@ -943,7 +943,7 @@
|
|
943 |
"model.layers.13.self_attn.k_proj": {
|
944 |
"bias": false,
|
945 |
"enable_norm": true,
|
946 |
-
"enable_perm":
|
947 |
"group_num": 1,
|
948 |
"group_size": 4055,
|
949 |
"in_features": 4096,
|
@@ -967,7 +967,7 @@
|
|
967 |
"model.layers.13.self_attn.o_proj": {
|
968 |
"bias": false,
|
969 |
"enable_norm": true,
|
970 |
-
"enable_perm":
|
971 |
"group_num": 1,
|
972 |
"group_size": 4055,
|
973 |
"in_features": 4096,
|
@@ -991,7 +991,7 @@
|
|
991 |
"model.layers.13.self_attn.q_proj": {
|
992 |
"bias": false,
|
993 |
"enable_norm": true,
|
994 |
-
"enable_perm":
|
995 |
"group_num": 1,
|
996 |
"group_size": 4055,
|
997 |
"in_features": 4096,
|
@@ -1015,7 +1015,7 @@
|
|
1015 |
"model.layers.13.self_attn.v_proj": {
|
1016 |
"bias": false,
|
1017 |
"enable_norm": true,
|
1018 |
-
"enable_perm":
|
1019 |
"group_num": 1,
|
1020 |
"group_size": 4055,
|
1021 |
"in_features": 4096,
|
@@ -1039,7 +1039,7 @@
|
|
1039 |
"model.layers.14.mlp.down_proj": {
|
1040 |
"bias": false,
|
1041 |
"enable_norm": true,
|
1042 |
-
"enable_perm":
|
1043 |
"group_num": 1,
|
1044 |
"group_size": 14192,
|
1045 |
"in_features": 14336,
|
@@ -1063,7 +1063,7 @@
|
|
1063 |
"model.layers.14.mlp.gate_proj": {
|
1064 |
"bias": false,
|
1065 |
"enable_norm": true,
|
1066 |
-
"enable_perm":
|
1067 |
"group_num": 1,
|
1068 |
"group_size": 4055,
|
1069 |
"in_features": 4096,
|
@@ -1087,7 +1087,7 @@
|
|
1087 |
"model.layers.14.mlp.up_proj": {
|
1088 |
"bias": false,
|
1089 |
"enable_norm": true,
|
1090 |
-
"enable_perm":
|
1091 |
"group_num": 1,
|
1092 |
"group_size": 4055,
|
1093 |
"in_features": 4096,
|
@@ -1111,7 +1111,7 @@
|
|
1111 |
"model.layers.14.self_attn.k_proj": {
|
1112 |
"bias": false,
|
1113 |
"enable_norm": true,
|
1114 |
-
"enable_perm":
|
1115 |
"group_num": 1,
|
1116 |
"group_size": 4055,
|
1117 |
"in_features": 4096,
|
@@ -1135,7 +1135,7 @@
|
|
1135 |
"model.layers.14.self_attn.o_proj": {
|
1136 |
"bias": false,
|
1137 |
"enable_norm": true,
|
1138 |
-
"enable_perm":
|
1139 |
"group_num": 1,
|
1140 |
"group_size": 4055,
|
1141 |
"in_features": 4096,
|
@@ -1159,7 +1159,7 @@
|
|
1159 |
"model.layers.14.self_attn.q_proj": {
|
1160 |
"bias": false,
|
1161 |
"enable_norm": true,
|
1162 |
-
"enable_perm":
|
1163 |
"group_num": 1,
|
1164 |
"group_size": 4055,
|
1165 |
"in_features": 4096,
|
@@ -1183,7 +1183,7 @@
|
|
1183 |
"model.layers.14.self_attn.v_proj": {
|
1184 |
"bias": false,
|
1185 |
"enable_norm": true,
|
1186 |
-
"enable_perm":
|
1187 |
"group_num": 1,
|
1188 |
"group_size": 4055,
|
1189 |
"in_features": 4096,
|
@@ -1207,7 +1207,7 @@
|
|
1207 |
"model.layers.15.mlp.down_proj": {
|
1208 |
"bias": false,
|
1209 |
"enable_norm": true,
|
1210 |
-
"enable_perm":
|
1211 |
"group_num": 1,
|
1212 |
"group_size": 14192,
|
1213 |
"in_features": 14336,
|
@@ -1231,7 +1231,7 @@
|
|
1231 |
"model.layers.15.mlp.gate_proj": {
|
1232 |
"bias": false,
|
1233 |
"enable_norm": true,
|
1234 |
-
"enable_perm":
|
1235 |
"group_num": 1,
|
1236 |
"group_size": 4055,
|
1237 |
"in_features": 4096,
|
@@ -1255,7 +1255,7 @@
|
|
1255 |
"model.layers.15.mlp.up_proj": {
|
1256 |
"bias": false,
|
1257 |
"enable_norm": true,
|
1258 |
-
"enable_perm":
|
1259 |
"group_num": 1,
|
1260 |
"group_size": 4055,
|
1261 |
"in_features": 4096,
|
@@ -1279,7 +1279,7 @@
|
|
1279 |
"model.layers.15.self_attn.k_proj": {
|
1280 |
"bias": false,
|
1281 |
"enable_norm": true,
|
1282 |
-
"enable_perm":
|
1283 |
"group_num": 1,
|
1284 |
"group_size": 4055,
|
1285 |
"in_features": 4096,
|
@@ -1303,7 +1303,7 @@
|
|
1303 |
"model.layers.15.self_attn.o_proj": {
|
1304 |
"bias": false,
|
1305 |
"enable_norm": true,
|
1306 |
-
"enable_perm":
|
1307 |
"group_num": 1,
|
1308 |
"group_size": 4055,
|
1309 |
"in_features": 4096,
|
@@ -1327,7 +1327,7 @@
|
|
1327 |
"model.layers.15.self_attn.q_proj": {
|
1328 |
"bias": false,
|
1329 |
"enable_norm": true,
|
1330 |
-
"enable_perm":
|
1331 |
"group_num": 1,
|
1332 |
"group_size": 4055,
|
1333 |
"in_features": 4096,
|
@@ -1351,7 +1351,7 @@
|
|
1351 |
"model.layers.15.self_attn.v_proj": {
|
1352 |
"bias": false,
|
1353 |
"enable_norm": true,
|
1354 |
-
"enable_perm":
|
1355 |
"group_num": 1,
|
1356 |
"group_size": 4055,
|
1357 |
"in_features": 4096,
|
@@ -1375,7 +1375,7 @@
|
|
1375 |
"model.layers.16.mlp.down_proj": {
|
1376 |
"bias": false,
|
1377 |
"enable_norm": true,
|
1378 |
-
"enable_perm":
|
1379 |
"group_num": 1,
|
1380 |
"group_size": 14192,
|
1381 |
"in_features": 14336,
|
@@ -1399,7 +1399,7 @@
|
|
1399 |
"model.layers.16.mlp.gate_proj": {
|
1400 |
"bias": false,
|
1401 |
"enable_norm": true,
|
1402 |
-
"enable_perm":
|
1403 |
"group_num": 1,
|
1404 |
"group_size": 4055,
|
1405 |
"in_features": 4096,
|
@@ -1423,7 +1423,7 @@
|
|
1423 |
"model.layers.16.mlp.up_proj": {
|
1424 |
"bias": false,
|
1425 |
"enable_norm": true,
|
1426 |
-
"enable_perm":
|
1427 |
"group_num": 1,
|
1428 |
"group_size": 4055,
|
1429 |
"in_features": 4096,
|
@@ -1447,7 +1447,7 @@
|
|
1447 |
"model.layers.16.self_attn.k_proj": {
|
1448 |
"bias": false,
|
1449 |
"enable_norm": true,
|
1450 |
-
"enable_perm":
|
1451 |
"group_num": 1,
|
1452 |
"group_size": 4055,
|
1453 |
"in_features": 4096,
|
@@ -1471,7 +1471,7 @@
|
|
1471 |
"model.layers.16.self_attn.o_proj": {
|
1472 |
"bias": false,
|
1473 |
"enable_norm": true,
|
1474 |
-
"enable_perm":
|
1475 |
"group_num": 1,
|
1476 |
"group_size": 4055,
|
1477 |
"in_features": 4096,
|
@@ -1495,7 +1495,7 @@
|
|
1495 |
"model.layers.16.self_attn.q_proj": {
|
1496 |
"bias": false,
|
1497 |
"enable_norm": true,
|
1498 |
-
"enable_perm":
|
1499 |
"group_num": 1,
|
1500 |
"group_size": 4055,
|
1501 |
"in_features": 4096,
|
@@ -1519,7 +1519,7 @@
|
|
1519 |
"model.layers.16.self_attn.v_proj": {
|
1520 |
"bias": false,
|
1521 |
"enable_norm": true,
|
1522 |
-
"enable_perm":
|
1523 |
"group_num": 1,
|
1524 |
"group_size": 4055,
|
1525 |
"in_features": 4096,
|
@@ -1543,7 +1543,7 @@
|
|
1543 |
"model.layers.17.mlp.down_proj": {
|
1544 |
"bias": false,
|
1545 |
"enable_norm": true,
|
1546 |
-
"enable_perm":
|
1547 |
"group_num": 1,
|
1548 |
"group_size": 14192,
|
1549 |
"in_features": 14336,
|
@@ -1567,7 +1567,7 @@
|
|
1567 |
"model.layers.17.mlp.gate_proj": {
|
1568 |
"bias": false,
|
1569 |
"enable_norm": true,
|
1570 |
-
"enable_perm":
|
1571 |
"group_num": 1,
|
1572 |
"group_size": 4055,
|
1573 |
"in_features": 4096,
|
@@ -1591,7 +1591,7 @@
|
|
1591 |
"model.layers.17.mlp.up_proj": {
|
1592 |
"bias": false,
|
1593 |
"enable_norm": true,
|
1594 |
-
"enable_perm":
|
1595 |
"group_num": 1,
|
1596 |
"group_size": 4055,
|
1597 |
"in_features": 4096,
|
@@ -1615,7 +1615,7 @@
|
|
1615 |
"model.layers.17.self_attn.k_proj": {
|
1616 |
"bias": false,
|
1617 |
"enable_norm": true,
|
1618 |
-
"enable_perm":
|
1619 |
"group_num": 1,
|
1620 |
"group_size": 4055,
|
1621 |
"in_features": 4096,
|
@@ -1639,7 +1639,7 @@
|
|
1639 |
"model.layers.17.self_attn.o_proj": {
|
1640 |
"bias": false,
|
1641 |
"enable_norm": true,
|
1642 |
-
"enable_perm":
|
1643 |
"group_num": 1,
|
1644 |
"group_size": 4055,
|
1645 |
"in_features": 4096,
|
@@ -1663,7 +1663,7 @@
|
|
1663 |
"model.layers.17.self_attn.q_proj": {
|
1664 |
"bias": false,
|
1665 |
"enable_norm": true,
|
1666 |
-
"enable_perm":
|
1667 |
"group_num": 1,
|
1668 |
"group_size": 4055,
|
1669 |
"in_features": 4096,
|
@@ -1687,7 +1687,7 @@
|
|
1687 |
"model.layers.17.self_attn.v_proj": {
|
1688 |
"bias": false,
|
1689 |
"enable_norm": true,
|
1690 |
-
"enable_perm":
|
1691 |
"group_num": 1,
|
1692 |
"group_size": 4055,
|
1693 |
"in_features": 4096,
|
@@ -1711,7 +1711,7 @@
|
|
1711 |
"model.layers.18.mlp.down_proj": {
|
1712 |
"bias": false,
|
1713 |
"enable_norm": true,
|
1714 |
-
"enable_perm":
|
1715 |
"group_num": 1,
|
1716 |
"group_size": 14192,
|
1717 |
"in_features": 14336,
|
@@ -1735,7 +1735,7 @@
|
|
1735 |
"model.layers.18.mlp.gate_proj": {
|
1736 |
"bias": false,
|
1737 |
"enable_norm": true,
|
1738 |
-
"enable_perm":
|
1739 |
"group_num": 1,
|
1740 |
"group_size": 4055,
|
1741 |
"in_features": 4096,
|
@@ -1759,7 +1759,7 @@
|
|
1759 |
"model.layers.18.mlp.up_proj": {
|
1760 |
"bias": false,
|
1761 |
"enable_norm": true,
|
1762 |
-
"enable_perm":
|
1763 |
"group_num": 1,
|
1764 |
"group_size": 4055,
|
1765 |
"in_features": 4096,
|
@@ -1783,7 +1783,7 @@
|
|
1783 |
"model.layers.18.self_attn.k_proj": {
|
1784 |
"bias": false,
|
1785 |
"enable_norm": true,
|
1786 |
-
"enable_perm":
|
1787 |
"group_num": 1,
|
1788 |
"group_size": 4055,
|
1789 |
"in_features": 4096,
|
@@ -1807,7 +1807,7 @@
|
|
1807 |
"model.layers.18.self_attn.o_proj": {
|
1808 |
"bias": false,
|
1809 |
"enable_norm": true,
|
1810 |
-
"enable_perm":
|
1811 |
"group_num": 1,
|
1812 |
"group_size": 4055,
|
1813 |
"in_features": 4096,
|
@@ -1831,7 +1831,7 @@
|
|
1831 |
"model.layers.18.self_attn.q_proj": {
|
1832 |
"bias": false,
|
1833 |
"enable_norm": true,
|
1834 |
-
"enable_perm":
|
1835 |
"group_num": 1,
|
1836 |
"group_size": 4055,
|
1837 |
"in_features": 4096,
|
@@ -1855,7 +1855,7 @@
|
|
1855 |
"model.layers.18.self_attn.v_proj": {
|
1856 |
"bias": false,
|
1857 |
"enable_norm": true,
|
1858 |
-
"enable_perm":
|
1859 |
"group_num": 1,
|
1860 |
"group_size": 4055,
|
1861 |
"in_features": 4096,
|
@@ -1879,7 +1879,7 @@
|
|
1879 |
"model.layers.19.mlp.down_proj": {
|
1880 |
"bias": false,
|
1881 |
"enable_norm": true,
|
1882 |
-
"enable_perm":
|
1883 |
"group_num": 1,
|
1884 |
"group_size": 14192,
|
1885 |
"in_features": 14336,
|
@@ -1903,7 +1903,7 @@
|
|
1903 |
"model.layers.19.mlp.gate_proj": {
|
1904 |
"bias": false,
|
1905 |
"enable_norm": true,
|
1906 |
-
"enable_perm":
|
1907 |
"group_num": 1,
|
1908 |
"group_size": 4055,
|
1909 |
"in_features": 4096,
|
@@ -1927,7 +1927,7 @@
|
|
1927 |
"model.layers.19.mlp.up_proj": {
|
1928 |
"bias": false,
|
1929 |
"enable_norm": true,
|
1930 |
-
"enable_perm":
|
1931 |
"group_num": 1,
|
1932 |
"group_size": 4055,
|
1933 |
"in_features": 4096,
|
@@ -1951,7 +1951,7 @@
|
|
1951 |
"model.layers.19.self_attn.k_proj": {
|
1952 |
"bias": false,
|
1953 |
"enable_norm": true,
|
1954 |
-
"enable_perm":
|
1955 |
"group_num": 1,
|
1956 |
"group_size": 4055,
|
1957 |
"in_features": 4096,
|
@@ -1975,7 +1975,7 @@
|
|
1975 |
"model.layers.19.self_attn.o_proj": {
|
1976 |
"bias": false,
|
1977 |
"enable_norm": true,
|
1978 |
-
"enable_perm":
|
1979 |
"group_num": 1,
|
1980 |
"group_size": 4055,
|
1981 |
"in_features": 4096,
|
@@ -1999,7 +1999,7 @@
|
|
1999 |
"model.layers.19.self_attn.q_proj": {
|
2000 |
"bias": false,
|
2001 |
"enable_norm": true,
|
2002 |
-
"enable_perm":
|
2003 |
"group_num": 1,
|
2004 |
"group_size": 4055,
|
2005 |
"in_features": 4096,
|
@@ -2023,7 +2023,7 @@
|
|
2023 |
"model.layers.19.self_attn.v_proj": {
|
2024 |
"bias": false,
|
2025 |
"enable_norm": true,
|
2026 |
-
"enable_perm":
|
2027 |
"group_num": 1,
|
2028 |
"group_size": 4055,
|
2029 |
"in_features": 4096,
|
@@ -2047,7 +2047,7 @@
|
|
2047 |
"model.layers.2.mlp.down_proj": {
|
2048 |
"bias": false,
|
2049 |
"enable_norm": true,
|
2050 |
-
"enable_perm":
|
2051 |
"group_num": 1,
|
2052 |
"group_size": 14192,
|
2053 |
"in_features": 14336,
|
@@ -2071,7 +2071,7 @@
|
|
2071 |
"model.layers.2.mlp.gate_proj": {
|
2072 |
"bias": false,
|
2073 |
"enable_norm": true,
|
2074 |
-
"enable_perm":
|
2075 |
"group_num": 1,
|
2076 |
"group_size": 4055,
|
2077 |
"in_features": 4096,
|
@@ -2095,7 +2095,7 @@
|
|
2095 |
"model.layers.2.mlp.up_proj": {
|
2096 |
"bias": false,
|
2097 |
"enable_norm": true,
|
2098 |
-
"enable_perm":
|
2099 |
"group_num": 1,
|
2100 |
"group_size": 4055,
|
2101 |
"in_features": 4096,
|
@@ -2119,7 +2119,7 @@
|
|
2119 |
"model.layers.2.self_attn.k_proj": {
|
2120 |
"bias": false,
|
2121 |
"enable_norm": true,
|
2122 |
-
"enable_perm":
|
2123 |
"group_num": 1,
|
2124 |
"group_size": 4055,
|
2125 |
"in_features": 4096,
|
@@ -2143,7 +2143,7 @@
|
|
2143 |
"model.layers.2.self_attn.o_proj": {
|
2144 |
"bias": false,
|
2145 |
"enable_norm": true,
|
2146 |
-
"enable_perm":
|
2147 |
"group_num": 1,
|
2148 |
"group_size": 4055,
|
2149 |
"in_features": 4096,
|
@@ -2167,7 +2167,7 @@
|
|
2167 |
"model.layers.2.self_attn.q_proj": {
|
2168 |
"bias": false,
|
2169 |
"enable_norm": true,
|
2170 |
-
"enable_perm":
|
2171 |
"group_num": 1,
|
2172 |
"group_size": 4055,
|
2173 |
"in_features": 4096,
|
@@ -2191,7 +2191,7 @@
|
|
2191 |
"model.layers.2.self_attn.v_proj": {
|
2192 |
"bias": false,
|
2193 |
"enable_norm": true,
|
2194 |
-
"enable_perm":
|
2195 |
"group_num": 1,
|
2196 |
"group_size": 4055,
|
2197 |
"in_features": 4096,
|
@@ -2215,7 +2215,7 @@
|
|
2215 |
"model.layers.20.mlp.down_proj": {
|
2216 |
"bias": false,
|
2217 |
"enable_norm": true,
|
2218 |
-
"enable_perm":
|
2219 |
"group_num": 1,
|
2220 |
"group_size": 14192,
|
2221 |
"in_features": 14336,
|
@@ -2239,7 +2239,7 @@
|
|
2239 |
"model.layers.20.mlp.gate_proj": {
|
2240 |
"bias": false,
|
2241 |
"enable_norm": true,
|
2242 |
-
"enable_perm":
|
2243 |
"group_num": 1,
|
2244 |
"group_size": 4055,
|
2245 |
"in_features": 4096,
|
@@ -2263,7 +2263,7 @@
|
|
2263 |
"model.layers.20.mlp.up_proj": {
|
2264 |
"bias": false,
|
2265 |
"enable_norm": true,
|
2266 |
-
"enable_perm":
|
2267 |
"group_num": 1,
|
2268 |
"group_size": 4055,
|
2269 |
"in_features": 4096,
|
@@ -2287,7 +2287,7 @@
|
|
2287 |
"model.layers.20.self_attn.k_proj": {
|
2288 |
"bias": false,
|
2289 |
"enable_norm": true,
|
2290 |
-
"enable_perm":
|
2291 |
"group_num": 1,
|
2292 |
"group_size": 4055,
|
2293 |
"in_features": 4096,
|
@@ -2311,7 +2311,7 @@
|
|
2311 |
"model.layers.20.self_attn.o_proj": {
|
2312 |
"bias": false,
|
2313 |
"enable_norm": true,
|
2314 |
-
"enable_perm":
|
2315 |
"group_num": 1,
|
2316 |
"group_size": 4055,
|
2317 |
"in_features": 4096,
|
@@ -2335,7 +2335,7 @@
|
|
2335 |
"model.layers.20.self_attn.q_proj": {
|
2336 |
"bias": false,
|
2337 |
"enable_norm": true,
|
2338 |
-
"enable_perm":
|
2339 |
"group_num": 1,
|
2340 |
"group_size": 4055,
|
2341 |
"in_features": 4096,
|
@@ -2359,7 +2359,7 @@
|
|
2359 |
"model.layers.20.self_attn.v_proj": {
|
2360 |
"bias": false,
|
2361 |
"enable_norm": true,
|
2362 |
-
"enable_perm":
|
2363 |
"group_num": 1,
|
2364 |
"group_size": 4055,
|
2365 |
"in_features": 4096,
|
@@ -2383,7 +2383,7 @@
|
|
2383 |
"model.layers.21.mlp.down_proj": {
|
2384 |
"bias": false,
|
2385 |
"enable_norm": true,
|
2386 |
-
"enable_perm":
|
2387 |
"group_num": 1,
|
2388 |
"group_size": 14192,
|
2389 |
"in_features": 14336,
|
@@ -2407,7 +2407,7 @@
|
|
2407 |
"model.layers.21.mlp.gate_proj": {
|
2408 |
"bias": false,
|
2409 |
"enable_norm": true,
|
2410 |
-
"enable_perm":
|
2411 |
"group_num": 1,
|
2412 |
"group_size": 4055,
|
2413 |
"in_features": 4096,
|
@@ -2431,7 +2431,7 @@
|
|
2431 |
"model.layers.21.mlp.up_proj": {
|
2432 |
"bias": false,
|
2433 |
"enable_norm": true,
|
2434 |
-
"enable_perm":
|
2435 |
"group_num": 1,
|
2436 |
"group_size": 4055,
|
2437 |
"in_features": 4096,
|
@@ -2455,7 +2455,7 @@
|
|
2455 |
"model.layers.21.self_attn.k_proj": {
|
2456 |
"bias": false,
|
2457 |
"enable_norm": true,
|
2458 |
-
"enable_perm":
|
2459 |
"group_num": 1,
|
2460 |
"group_size": 4055,
|
2461 |
"in_features": 4096,
|
@@ -2479,7 +2479,7 @@
|
|
2479 |
"model.layers.21.self_attn.o_proj": {
|
2480 |
"bias": false,
|
2481 |
"enable_norm": true,
|
2482 |
-
"enable_perm":
|
2483 |
"group_num": 1,
|
2484 |
"group_size": 4055,
|
2485 |
"in_features": 4096,
|
@@ -2503,7 +2503,7 @@
|
|
2503 |
"model.layers.21.self_attn.q_proj": {
|
2504 |
"bias": false,
|
2505 |
"enable_norm": true,
|
2506 |
-
"enable_perm":
|
2507 |
"group_num": 1,
|
2508 |
"group_size": 4055,
|
2509 |
"in_features": 4096,
|
@@ -2527,7 +2527,7 @@
|
|
2527 |
"model.layers.21.self_attn.v_proj": {
|
2528 |
"bias": false,
|
2529 |
"enable_norm": true,
|
2530 |
-
"enable_perm":
|
2531 |
"group_num": 1,
|
2532 |
"group_size": 4055,
|
2533 |
"in_features": 4096,
|
@@ -2551,7 +2551,7 @@
|
|
2551 |
"model.layers.22.mlp.down_proj": {
|
2552 |
"bias": false,
|
2553 |
"enable_norm": true,
|
2554 |
-
"enable_perm":
|
2555 |
"group_num": 1,
|
2556 |
"group_size": 14192,
|
2557 |
"in_features": 14336,
|
@@ -2575,7 +2575,7 @@
|
|
2575 |
"model.layers.22.mlp.gate_proj": {
|
2576 |
"bias": false,
|
2577 |
"enable_norm": true,
|
2578 |
-
"enable_perm":
|
2579 |
"group_num": 1,
|
2580 |
"group_size": 4055,
|
2581 |
"in_features": 4096,
|
@@ -2599,7 +2599,7 @@
|
|
2599 |
"model.layers.22.mlp.up_proj": {
|
2600 |
"bias": false,
|
2601 |
"enable_norm": true,
|
2602 |
-
"enable_perm":
|
2603 |
"group_num": 1,
|
2604 |
"group_size": 4055,
|
2605 |
"in_features": 4096,
|
@@ -2623,7 +2623,7 @@
|
|
2623 |
"model.layers.22.self_attn.k_proj": {
|
2624 |
"bias": false,
|
2625 |
"enable_norm": true,
|
2626 |
-
"enable_perm":
|
2627 |
"group_num": 1,
|
2628 |
"group_size": 4055,
|
2629 |
"in_features": 4096,
|
@@ -2647,7 +2647,7 @@
|
|
2647 |
"model.layers.22.self_attn.o_proj": {
|
2648 |
"bias": false,
|
2649 |
"enable_norm": true,
|
2650 |
-
"enable_perm":
|
2651 |
"group_num": 1,
|
2652 |
"group_size": 4055,
|
2653 |
"in_features": 4096,
|
@@ -2671,7 +2671,7 @@
|
|
2671 |
"model.layers.22.self_attn.q_proj": {
|
2672 |
"bias": false,
|
2673 |
"enable_norm": true,
|
2674 |
-
"enable_perm":
|
2675 |
"group_num": 1,
|
2676 |
"group_size": 4055,
|
2677 |
"in_features": 4096,
|
@@ -2695,7 +2695,7 @@
|
|
2695 |
"model.layers.22.self_attn.v_proj": {
|
2696 |
"bias": false,
|
2697 |
"enable_norm": true,
|
2698 |
-
"enable_perm":
|
2699 |
"group_num": 1,
|
2700 |
"group_size": 4055,
|
2701 |
"in_features": 4096,
|
@@ -2719,7 +2719,7 @@
|
|
2719 |
"model.layers.23.mlp.down_proj": {
|
2720 |
"bias": false,
|
2721 |
"enable_norm": true,
|
2722 |
-
"enable_perm":
|
2723 |
"group_num": 1,
|
2724 |
"group_size": 14192,
|
2725 |
"in_features": 14336,
|
@@ -2743,7 +2743,7 @@
|
|
2743 |
"model.layers.23.mlp.gate_proj": {
|
2744 |
"bias": false,
|
2745 |
"enable_norm": true,
|
2746 |
-
"enable_perm":
|
2747 |
"group_num": 1,
|
2748 |
"group_size": 4055,
|
2749 |
"in_features": 4096,
|
@@ -2767,7 +2767,7 @@
|
|
2767 |
"model.layers.23.mlp.up_proj": {
|
2768 |
"bias": false,
|
2769 |
"enable_norm": true,
|
2770 |
-
"enable_perm":
|
2771 |
"group_num": 1,
|
2772 |
"group_size": 4055,
|
2773 |
"in_features": 4096,
|
@@ -2791,7 +2791,7 @@
|
|
2791 |
"model.layers.23.self_attn.k_proj": {
|
2792 |
"bias": false,
|
2793 |
"enable_norm": true,
|
2794 |
-
"enable_perm":
|
2795 |
"group_num": 1,
|
2796 |
"group_size": 4055,
|
2797 |
"in_features": 4096,
|
@@ -2815,7 +2815,7 @@
|
|
2815 |
"model.layers.23.self_attn.o_proj": {
|
2816 |
"bias": false,
|
2817 |
"enable_norm": true,
|
2818 |
-
"enable_perm":
|
2819 |
"group_num": 1,
|
2820 |
"group_size": 4055,
|
2821 |
"in_features": 4096,
|
@@ -2839,7 +2839,7 @@
|
|
2839 |
"model.layers.23.self_attn.q_proj": {
|
2840 |
"bias": false,
|
2841 |
"enable_norm": true,
|
2842 |
-
"enable_perm":
|
2843 |
"group_num": 1,
|
2844 |
"group_size": 4055,
|
2845 |
"in_features": 4096,
|
@@ -2863,7 +2863,7 @@
|
|
2863 |
"model.layers.23.self_attn.v_proj": {
|
2864 |
"bias": false,
|
2865 |
"enable_norm": true,
|
2866 |
-
"enable_perm":
|
2867 |
"group_num": 1,
|
2868 |
"group_size": 4055,
|
2869 |
"in_features": 4096,
|
@@ -2887,7 +2887,7 @@
|
|
2887 |
"model.layers.24.mlp.down_proj": {
|
2888 |
"bias": false,
|
2889 |
"enable_norm": true,
|
2890 |
-
"enable_perm":
|
2891 |
"group_num": 1,
|
2892 |
"group_size": 14192,
|
2893 |
"in_features": 14336,
|
@@ -2911,7 +2911,7 @@
|
|
2911 |
"model.layers.24.mlp.gate_proj": {
|
2912 |
"bias": false,
|
2913 |
"enable_norm": true,
|
2914 |
-
"enable_perm":
|
2915 |
"group_num": 1,
|
2916 |
"group_size": 4055,
|
2917 |
"in_features": 4096,
|
@@ -2935,7 +2935,7 @@
|
|
2935 |
"model.layers.24.mlp.up_proj": {
|
2936 |
"bias": false,
|
2937 |
"enable_norm": true,
|
2938 |
-
"enable_perm":
|
2939 |
"group_num": 1,
|
2940 |
"group_size": 4055,
|
2941 |
"in_features": 4096,
|
@@ -2959,7 +2959,7 @@
|
|
2959 |
"model.layers.24.self_attn.k_proj": {
|
2960 |
"bias": false,
|
2961 |
"enable_norm": true,
|
2962 |
-
"enable_perm":
|
2963 |
"group_num": 1,
|
2964 |
"group_size": 4055,
|
2965 |
"in_features": 4096,
|
@@ -2983,7 +2983,7 @@
|
|
2983 |
"model.layers.24.self_attn.o_proj": {
|
2984 |
"bias": false,
|
2985 |
"enable_norm": true,
|
2986 |
-
"enable_perm":
|
2987 |
"group_num": 1,
|
2988 |
"group_size": 4055,
|
2989 |
"in_features": 4096,
|
@@ -3007,7 +3007,7 @@
|
|
3007 |
"model.layers.24.self_attn.q_proj": {
|
3008 |
"bias": false,
|
3009 |
"enable_norm": true,
|
3010 |
-
"enable_perm":
|
3011 |
"group_num": 1,
|
3012 |
"group_size": 4055,
|
3013 |
"in_features": 4096,
|
@@ -3031,7 +3031,7 @@
|
|
3031 |
"model.layers.24.self_attn.v_proj": {
|
3032 |
"bias": false,
|
3033 |
"enable_norm": true,
|
3034 |
-
"enable_perm":
|
3035 |
"group_num": 1,
|
3036 |
"group_size": 4055,
|
3037 |
"in_features": 4096,
|
@@ -3055,7 +3055,7 @@
|
|
3055 |
"model.layers.25.mlp.down_proj": {
|
3056 |
"bias": false,
|
3057 |
"enable_norm": true,
|
3058 |
-
"enable_perm":
|
3059 |
"group_num": 1,
|
3060 |
"group_size": 14192,
|
3061 |
"in_features": 14336,
|
@@ -3079,7 +3079,7 @@
|
|
3079 |
"model.layers.25.mlp.gate_proj": {
|
3080 |
"bias": false,
|
3081 |
"enable_norm": true,
|
3082 |
-
"enable_perm":
|
3083 |
"group_num": 1,
|
3084 |
"group_size": 4055,
|
3085 |
"in_features": 4096,
|
@@ -3103,7 +3103,7 @@
|
|
3103 |
"model.layers.25.mlp.up_proj": {
|
3104 |
"bias": false,
|
3105 |
"enable_norm": true,
|
3106 |
-
"enable_perm":
|
3107 |
"group_num": 1,
|
3108 |
"group_size": 4055,
|
3109 |
"in_features": 4096,
|
@@ -3127,7 +3127,7 @@
|
|
3127 |
"model.layers.25.self_attn.k_proj": {
|
3128 |
"bias": false,
|
3129 |
"enable_norm": true,
|
3130 |
-
"enable_perm":
|
3131 |
"group_num": 1,
|
3132 |
"group_size": 4055,
|
3133 |
"in_features": 4096,
|
@@ -3151,7 +3151,7 @@
|
|
3151 |
"model.layers.25.self_attn.o_proj": {
|
3152 |
"bias": false,
|
3153 |
"enable_norm": true,
|
3154 |
-
"enable_perm":
|
3155 |
"group_num": 1,
|
3156 |
"group_size": 4055,
|
3157 |
"in_features": 4096,
|
@@ -3175,7 +3175,7 @@
|
|
3175 |
"model.layers.25.self_attn.q_proj": {
|
3176 |
"bias": false,
|
3177 |
"enable_norm": true,
|
3178 |
-
"enable_perm":
|
3179 |
"group_num": 1,
|
3180 |
"group_size": 4055,
|
3181 |
"in_features": 4096,
|
@@ -3199,7 +3199,7 @@
|
|
3199 |
"model.layers.25.self_attn.v_proj": {
|
3200 |
"bias": false,
|
3201 |
"enable_norm": true,
|
3202 |
-
"enable_perm":
|
3203 |
"group_num": 1,
|
3204 |
"group_size": 4055,
|
3205 |
"in_features": 4096,
|
@@ -3223,7 +3223,7 @@
|
|
3223 |
"model.layers.26.mlp.down_proj": {
|
3224 |
"bias": false,
|
3225 |
"enable_norm": true,
|
3226 |
-
"enable_perm":
|
3227 |
"group_num": 1,
|
3228 |
"group_size": 14192,
|
3229 |
"in_features": 14336,
|
@@ -3247,7 +3247,7 @@
|
|
3247 |
"model.layers.26.mlp.gate_proj": {
|
3248 |
"bias": false,
|
3249 |
"enable_norm": true,
|
3250 |
-
"enable_perm":
|
3251 |
"group_num": 1,
|
3252 |
"group_size": 4055,
|
3253 |
"in_features": 4096,
|
@@ -3271,7 +3271,7 @@
|
|
3271 |
"model.layers.26.mlp.up_proj": {
|
3272 |
"bias": false,
|
3273 |
"enable_norm": true,
|
3274 |
-
"enable_perm":
|
3275 |
"group_num": 1,
|
3276 |
"group_size": 4055,
|
3277 |
"in_features": 4096,
|
@@ -3295,7 +3295,7 @@
|
|
3295 |
"model.layers.26.self_attn.k_proj": {
|
3296 |
"bias": false,
|
3297 |
"enable_norm": true,
|
3298 |
-
"enable_perm":
|
3299 |
"group_num": 1,
|
3300 |
"group_size": 4055,
|
3301 |
"in_features": 4096,
|
@@ -3319,7 +3319,7 @@
|
|
3319 |
"model.layers.26.self_attn.o_proj": {
|
3320 |
"bias": false,
|
3321 |
"enable_norm": true,
|
3322 |
-
"enable_perm":
|
3323 |
"group_num": 1,
|
3324 |
"group_size": 4055,
|
3325 |
"in_features": 4096,
|
@@ -3343,7 +3343,7 @@
|
|
3343 |
"model.layers.26.self_attn.q_proj": {
|
3344 |
"bias": false,
|
3345 |
"enable_norm": true,
|
3346 |
-
"enable_perm":
|
3347 |
"group_num": 1,
|
3348 |
"group_size": 4055,
|
3349 |
"in_features": 4096,
|
@@ -3367,7 +3367,7 @@
|
|
3367 |
"model.layers.26.self_attn.v_proj": {
|
3368 |
"bias": false,
|
3369 |
"enable_norm": true,
|
3370 |
-
"enable_perm":
|
3371 |
"group_num": 1,
|
3372 |
"group_size": 4055,
|
3373 |
"in_features": 4096,
|
@@ -3391,7 +3391,7 @@
|
|
3391 |
"model.layers.27.mlp.down_proj": {
|
3392 |
"bias": false,
|
3393 |
"enable_norm": true,
|
3394 |
-
"enable_perm":
|
3395 |
"group_num": 1,
|
3396 |
"group_size": 14192,
|
3397 |
"in_features": 14336,
|
@@ -3415,7 +3415,7 @@
|
|
3415 |
"model.layers.27.mlp.gate_proj": {
|
3416 |
"bias": false,
|
3417 |
"enable_norm": true,
|
3418 |
-
"enable_perm":
|
3419 |
"group_num": 1,
|
3420 |
"group_size": 4055,
|
3421 |
"in_features": 4096,
|
@@ -3439,7 +3439,7 @@
|
|
3439 |
"model.layers.27.mlp.up_proj": {
|
3440 |
"bias": false,
|
3441 |
"enable_norm": true,
|
3442 |
-
"enable_perm":
|
3443 |
"group_num": 1,
|
3444 |
"group_size": 4055,
|
3445 |
"in_features": 4096,
|
@@ -3463,7 +3463,7 @@
|
|
3463 |
"model.layers.27.self_attn.k_proj": {
|
3464 |
"bias": false,
|
3465 |
"enable_norm": true,
|
3466 |
-
"enable_perm":
|
3467 |
"group_num": 1,
|
3468 |
"group_size": 4055,
|
3469 |
"in_features": 4096,
|
@@ -3487,7 +3487,7 @@
|
|
3487 |
"model.layers.27.self_attn.o_proj": {
|
3488 |
"bias": false,
|
3489 |
"enable_norm": true,
|
3490 |
-
"enable_perm":
|
3491 |
"group_num": 1,
|
3492 |
"group_size": 4055,
|
3493 |
"in_features": 4096,
|
@@ -3511,7 +3511,7 @@
|
|
3511 |
"model.layers.27.self_attn.q_proj": {
|
3512 |
"bias": false,
|
3513 |
"enable_norm": true,
|
3514 |
-
"enable_perm":
|
3515 |
"group_num": 1,
|
3516 |
"group_size": 4055,
|
3517 |
"in_features": 4096,
|
@@ -3535,7 +3535,7 @@
|
|
3535 |
"model.layers.27.self_attn.v_proj": {
|
3536 |
"bias": false,
|
3537 |
"enable_norm": true,
|
3538 |
-
"enable_perm":
|
3539 |
"group_num": 1,
|
3540 |
"group_size": 4055,
|
3541 |
"in_features": 4096,
|
@@ -3559,7 +3559,7 @@
|
|
3559 |
"model.layers.28.mlp.down_proj": {
|
3560 |
"bias": false,
|
3561 |
"enable_norm": true,
|
3562 |
-
"enable_perm":
|
3563 |
"group_num": 1,
|
3564 |
"group_size": 14192,
|
3565 |
"in_features": 14336,
|
@@ -3583,7 +3583,7 @@
|
|
3583 |
"model.layers.28.mlp.gate_proj": {
|
3584 |
"bias": false,
|
3585 |
"enable_norm": true,
|
3586 |
-
"enable_perm":
|
3587 |
"group_num": 1,
|
3588 |
"group_size": 4055,
|
3589 |
"in_features": 4096,
|
@@ -3607,7 +3607,7 @@
|
|
3607 |
"model.layers.28.mlp.up_proj": {
|
3608 |
"bias": false,
|
3609 |
"enable_norm": true,
|
3610 |
-
"enable_perm":
|
3611 |
"group_num": 1,
|
3612 |
"group_size": 4055,
|
3613 |
"in_features": 4096,
|
@@ -3631,7 +3631,7 @@
|
|
3631 |
"model.layers.28.self_attn.k_proj": {
|
3632 |
"bias": false,
|
3633 |
"enable_norm": true,
|
3634 |
-
"enable_perm":
|
3635 |
"group_num": 1,
|
3636 |
"group_size": 4055,
|
3637 |
"in_features": 4096,
|
@@ -3655,7 +3655,7 @@
|
|
3655 |
"model.layers.28.self_attn.o_proj": {
|
3656 |
"bias": false,
|
3657 |
"enable_norm": true,
|
3658 |
-
"enable_perm":
|
3659 |
"group_num": 1,
|
3660 |
"group_size": 4055,
|
3661 |
"in_features": 4096,
|
@@ -3679,7 +3679,7 @@
|
|
3679 |
"model.layers.28.self_attn.q_proj": {
|
3680 |
"bias": false,
|
3681 |
"enable_norm": true,
|
3682 |
-
"enable_perm":
|
3683 |
"group_num": 1,
|
3684 |
"group_size": 4055,
|
3685 |
"in_features": 4096,
|
@@ -3703,7 +3703,7 @@
|
|
3703 |
"model.layers.28.self_attn.v_proj": {
|
3704 |
"bias": false,
|
3705 |
"enable_norm": true,
|
3706 |
-
"enable_perm":
|
3707 |
"group_num": 1,
|
3708 |
"group_size": 4055,
|
3709 |
"in_features": 4096,
|
@@ -3727,7 +3727,7 @@
|
|
3727 |
"model.layers.29.mlp.down_proj": {
|
3728 |
"bias": false,
|
3729 |
"enable_norm": true,
|
3730 |
-
"enable_perm":
|
3731 |
"group_num": 1,
|
3732 |
"group_size": 14192,
|
3733 |
"in_features": 14336,
|
@@ -3751,7 +3751,7 @@
|
|
3751 |
"model.layers.29.mlp.gate_proj": {
|
3752 |
"bias": false,
|
3753 |
"enable_norm": true,
|
3754 |
-
"enable_perm":
|
3755 |
"group_num": 1,
|
3756 |
"group_size": 4055,
|
3757 |
"in_features": 4096,
|
@@ -3775,7 +3775,7 @@
|
|
3775 |
"model.layers.29.mlp.up_proj": {
|
3776 |
"bias": false,
|
3777 |
"enable_norm": true,
|
3778 |
-
"enable_perm":
|
3779 |
"group_num": 1,
|
3780 |
"group_size": 4055,
|
3781 |
"in_features": 4096,
|
@@ -3799,7 +3799,7 @@
|
|
3799 |
"model.layers.29.self_attn.k_proj": {
|
3800 |
"bias": false,
|
3801 |
"enable_norm": true,
|
3802 |
-
"enable_perm":
|
3803 |
"group_num": 1,
|
3804 |
"group_size": 4055,
|
3805 |
"in_features": 4096,
|
@@ -3823,7 +3823,7 @@
|
|
3823 |
"model.layers.29.self_attn.o_proj": {
|
3824 |
"bias": false,
|
3825 |
"enable_norm": true,
|
3826 |
-
"enable_perm":
|
3827 |
"group_num": 1,
|
3828 |
"group_size": 4055,
|
3829 |
"in_features": 4096,
|
@@ -3847,7 +3847,7 @@
|
|
3847 |
"model.layers.29.self_attn.q_proj": {
|
3848 |
"bias": false,
|
3849 |
"enable_norm": true,
|
3850 |
-
"enable_perm":
|
3851 |
"group_num": 1,
|
3852 |
"group_size": 4055,
|
3853 |
"in_features": 4096,
|
@@ -3871,7 +3871,7 @@
|
|
3871 |
"model.layers.29.self_attn.v_proj": {
|
3872 |
"bias": false,
|
3873 |
"enable_norm": true,
|
3874 |
-
"enable_perm":
|
3875 |
"group_num": 1,
|
3876 |
"group_size": 4055,
|
3877 |
"in_features": 4096,
|
@@ -3895,7 +3895,7 @@
|
|
3895 |
"model.layers.3.mlp.down_proj": {
|
3896 |
"bias": false,
|
3897 |
"enable_norm": true,
|
3898 |
-
"enable_perm":
|
3899 |
"group_num": 1,
|
3900 |
"group_size": 14192,
|
3901 |
"in_features": 14336,
|
@@ -3919,7 +3919,7 @@
|
|
3919 |
"model.layers.3.mlp.gate_proj": {
|
3920 |
"bias": false,
|
3921 |
"enable_norm": true,
|
3922 |
-
"enable_perm":
|
3923 |
"group_num": 1,
|
3924 |
"group_size": 4055,
|
3925 |
"in_features": 4096,
|
@@ -3943,7 +3943,7 @@
|
|
3943 |
"model.layers.3.mlp.up_proj": {
|
3944 |
"bias": false,
|
3945 |
"enable_norm": true,
|
3946 |
-
"enable_perm":
|
3947 |
"group_num": 1,
|
3948 |
"group_size": 4055,
|
3949 |
"in_features": 4096,
|
@@ -3967,7 +3967,7 @@
|
|
3967 |
"model.layers.3.self_attn.k_proj": {
|
3968 |
"bias": false,
|
3969 |
"enable_norm": true,
|
3970 |
-
"enable_perm":
|
3971 |
"group_num": 1,
|
3972 |
"group_size": 4055,
|
3973 |
"in_features": 4096,
|
@@ -3991,7 +3991,7 @@
|
|
3991 |
"model.layers.3.self_attn.o_proj": {
|
3992 |
"bias": false,
|
3993 |
"enable_norm": true,
|
3994 |
-
"enable_perm":
|
3995 |
"group_num": 1,
|
3996 |
"group_size": 4055,
|
3997 |
"in_features": 4096,
|
@@ -4015,7 +4015,7 @@
|
|
4015 |
"model.layers.3.self_attn.q_proj": {
|
4016 |
"bias": false,
|
4017 |
"enable_norm": true,
|
4018 |
-
"enable_perm":
|
4019 |
"group_num": 1,
|
4020 |
"group_size": 4055,
|
4021 |
"in_features": 4096,
|
@@ -4039,7 +4039,7 @@
|
|
4039 |
"model.layers.3.self_attn.v_proj": {
|
4040 |
"bias": false,
|
4041 |
"enable_norm": true,
|
4042 |
-
"enable_perm":
|
4043 |
"group_num": 1,
|
4044 |
"group_size": 4055,
|
4045 |
"in_features": 4096,
|
@@ -4063,7 +4063,7 @@
|
|
4063 |
"model.layers.30.mlp.down_proj": {
|
4064 |
"bias": false,
|
4065 |
"enable_norm": true,
|
4066 |
-
"enable_perm":
|
4067 |
"group_num": 1,
|
4068 |
"group_size": 14192,
|
4069 |
"in_features": 14336,
|
@@ -4087,7 +4087,7 @@
|
|
4087 |
"model.layers.30.mlp.gate_proj": {
|
4088 |
"bias": false,
|
4089 |
"enable_norm": true,
|
4090 |
-
"enable_perm":
|
4091 |
"group_num": 1,
|
4092 |
"group_size": 4055,
|
4093 |
"in_features": 4096,
|
@@ -4111,7 +4111,7 @@
|
|
4111 |
"model.layers.30.mlp.up_proj": {
|
4112 |
"bias": false,
|
4113 |
"enable_norm": true,
|
4114 |
-
"enable_perm":
|
4115 |
"group_num": 1,
|
4116 |
"group_size": 4055,
|
4117 |
"in_features": 4096,
|
@@ -4135,7 +4135,7 @@
|
|
4135 |
"model.layers.30.self_attn.k_proj": {
|
4136 |
"bias": false,
|
4137 |
"enable_norm": true,
|
4138 |
-
"enable_perm":
|
4139 |
"group_num": 1,
|
4140 |
"group_size": 4055,
|
4141 |
"in_features": 4096,
|
@@ -4159,7 +4159,7 @@
|
|
4159 |
"model.layers.30.self_attn.o_proj": {
|
4160 |
"bias": false,
|
4161 |
"enable_norm": true,
|
4162 |
-
"enable_perm":
|
4163 |
"group_num": 1,
|
4164 |
"group_size": 4055,
|
4165 |
"in_features": 4096,
|
@@ -4183,7 +4183,7 @@
|
|
4183 |
"model.layers.30.self_attn.q_proj": {
|
4184 |
"bias": false,
|
4185 |
"enable_norm": true,
|
4186 |
-
"enable_perm":
|
4187 |
"group_num": 1,
|
4188 |
"group_size": 4055,
|
4189 |
"in_features": 4096,
|
@@ -4207,7 +4207,7 @@
|
|
4207 |
"model.layers.30.self_attn.v_proj": {
|
4208 |
"bias": false,
|
4209 |
"enable_norm": true,
|
4210 |
-
"enable_perm":
|
4211 |
"group_num": 1,
|
4212 |
"group_size": 4055,
|
4213 |
"in_features": 4096,
|
@@ -4231,7 +4231,7 @@
|
|
4231 |
"model.layers.31.mlp.down_proj": {
|
4232 |
"bias": false,
|
4233 |
"enable_norm": true,
|
4234 |
-
"enable_perm":
|
4235 |
"group_num": 1,
|
4236 |
"group_size": 14192,
|
4237 |
"in_features": 14336,
|
@@ -4255,7 +4255,7 @@
|
|
4255 |
"model.layers.31.mlp.gate_proj": {
|
4256 |
"bias": false,
|
4257 |
"enable_norm": true,
|
4258 |
-
"enable_perm":
|
4259 |
"group_num": 1,
|
4260 |
"group_size": 4055,
|
4261 |
"in_features": 4096,
|
@@ -4279,7 +4279,7 @@
|
|
4279 |
"model.layers.31.mlp.up_proj": {
|
4280 |
"bias": false,
|
4281 |
"enable_norm": true,
|
4282 |
-
"enable_perm":
|
4283 |
"group_num": 1,
|
4284 |
"group_size": 4055,
|
4285 |
"in_features": 4096,
|
@@ -4303,7 +4303,7 @@
|
|
4303 |
"model.layers.31.self_attn.k_proj": {
|
4304 |
"bias": false,
|
4305 |
"enable_norm": true,
|
4306 |
-
"enable_perm":
|
4307 |
"group_num": 1,
|
4308 |
"group_size": 4055,
|
4309 |
"in_features": 4096,
|
@@ -4327,7 +4327,7 @@
|
|
4327 |
"model.layers.31.self_attn.o_proj": {
|
4328 |
"bias": false,
|
4329 |
"enable_norm": true,
|
4330 |
-
"enable_perm":
|
4331 |
"group_num": 1,
|
4332 |
"group_size": 4055,
|
4333 |
"in_features": 4096,
|
@@ -4351,7 +4351,7 @@
|
|
4351 |
"model.layers.31.self_attn.q_proj": {
|
4352 |
"bias": false,
|
4353 |
"enable_norm": true,
|
4354 |
-
"enable_perm":
|
4355 |
"group_num": 1,
|
4356 |
"group_size": 4055,
|
4357 |
"in_features": 4096,
|
@@ -4375,7 +4375,7 @@
|
|
4375 |
"model.layers.31.self_attn.v_proj": {
|
4376 |
"bias": false,
|
4377 |
"enable_norm": true,
|
4378 |
-
"enable_perm":
|
4379 |
"group_num": 1,
|
4380 |
"group_size": 4055,
|
4381 |
"in_features": 4096,
|
@@ -4399,7 +4399,7 @@
|
|
4399 |
"model.layers.4.mlp.down_proj": {
|
4400 |
"bias": false,
|
4401 |
"enable_norm": true,
|
4402 |
-
"enable_perm":
|
4403 |
"group_num": 1,
|
4404 |
"group_size": 14192,
|
4405 |
"in_features": 14336,
|
@@ -4423,7 +4423,7 @@
|
|
4423 |
"model.layers.4.mlp.gate_proj": {
|
4424 |
"bias": false,
|
4425 |
"enable_norm": true,
|
4426 |
-
"enable_perm":
|
4427 |
"group_num": 1,
|
4428 |
"group_size": 4055,
|
4429 |
"in_features": 4096,
|
@@ -4447,7 +4447,7 @@
|
|
4447 |
"model.layers.4.mlp.up_proj": {
|
4448 |
"bias": false,
|
4449 |
"enable_norm": true,
|
4450 |
-
"enable_perm":
|
4451 |
"group_num": 1,
|
4452 |
"group_size": 4055,
|
4453 |
"in_features": 4096,
|
@@ -4471,7 +4471,7 @@
|
|
4471 |
"model.layers.4.self_attn.k_proj": {
|
4472 |
"bias": false,
|
4473 |
"enable_norm": true,
|
4474 |
-
"enable_perm":
|
4475 |
"group_num": 1,
|
4476 |
"group_size": 4055,
|
4477 |
"in_features": 4096,
|
@@ -4495,7 +4495,7 @@
|
|
4495 |
"model.layers.4.self_attn.o_proj": {
|
4496 |
"bias": false,
|
4497 |
"enable_norm": true,
|
4498 |
-
"enable_perm":
|
4499 |
"group_num": 1,
|
4500 |
"group_size": 4055,
|
4501 |
"in_features": 4096,
|
@@ -4519,7 +4519,7 @@
|
|
4519 |
"model.layers.4.self_attn.q_proj": {
|
4520 |
"bias": false,
|
4521 |
"enable_norm": true,
|
4522 |
-
"enable_perm":
|
4523 |
"group_num": 1,
|
4524 |
"group_size": 4055,
|
4525 |
"in_features": 4096,
|
@@ -4543,7 +4543,7 @@
|
|
4543 |
"model.layers.4.self_attn.v_proj": {
|
4544 |
"bias": false,
|
4545 |
"enable_norm": true,
|
4546 |
-
"enable_perm":
|
4547 |
"group_num": 1,
|
4548 |
"group_size": 4055,
|
4549 |
"in_features": 4096,
|
@@ -4567,7 +4567,7 @@
|
|
4567 |
"model.layers.5.mlp.down_proj": {
|
4568 |
"bias": false,
|
4569 |
"enable_norm": true,
|
4570 |
-
"enable_perm":
|
4571 |
"group_num": 1,
|
4572 |
"group_size": 14192,
|
4573 |
"in_features": 14336,
|
@@ -4591,7 +4591,7 @@
|
|
4591 |
"model.layers.5.mlp.gate_proj": {
|
4592 |
"bias": false,
|
4593 |
"enable_norm": true,
|
4594 |
-
"enable_perm":
|
4595 |
"group_num": 1,
|
4596 |
"group_size": 4055,
|
4597 |
"in_features": 4096,
|
@@ -4615,7 +4615,7 @@
|
|
4615 |
"model.layers.5.mlp.up_proj": {
|
4616 |
"bias": false,
|
4617 |
"enable_norm": true,
|
4618 |
-
"enable_perm":
|
4619 |
"group_num": 1,
|
4620 |
"group_size": 4055,
|
4621 |
"in_features": 4096,
|
@@ -4639,7 +4639,7 @@
|
|
4639 |
"model.layers.5.self_attn.k_proj": {
|
4640 |
"bias": false,
|
4641 |
"enable_norm": true,
|
4642 |
-
"enable_perm":
|
4643 |
"group_num": 1,
|
4644 |
"group_size": 4055,
|
4645 |
"in_features": 4096,
|
@@ -4663,7 +4663,7 @@
|
|
4663 |
"model.layers.5.self_attn.o_proj": {
|
4664 |
"bias": false,
|
4665 |
"enable_norm": true,
|
4666 |
-
"enable_perm":
|
4667 |
"group_num": 1,
|
4668 |
"group_size": 4055,
|
4669 |
"in_features": 4096,
|
@@ -4687,7 +4687,7 @@
|
|
4687 |
"model.layers.5.self_attn.q_proj": {
|
4688 |
"bias": false,
|
4689 |
"enable_norm": true,
|
4690 |
-
"enable_perm":
|
4691 |
"group_num": 1,
|
4692 |
"group_size": 4055,
|
4693 |
"in_features": 4096,
|
@@ -4711,7 +4711,7 @@
|
|
4711 |
"model.layers.5.self_attn.v_proj": {
|
4712 |
"bias": false,
|
4713 |
"enable_norm": true,
|
4714 |
-
"enable_perm":
|
4715 |
"group_num": 1,
|
4716 |
"group_size": 4055,
|
4717 |
"in_features": 4096,
|
@@ -4735,7 +4735,7 @@
|
|
4735 |
"model.layers.6.mlp.down_proj": {
|
4736 |
"bias": false,
|
4737 |
"enable_norm": true,
|
4738 |
-
"enable_perm":
|
4739 |
"group_num": 1,
|
4740 |
"group_size": 14192,
|
4741 |
"in_features": 14336,
|
@@ -4759,7 +4759,7 @@
|
|
4759 |
"model.layers.6.mlp.gate_proj": {
|
4760 |
"bias": false,
|
4761 |
"enable_norm": true,
|
4762 |
-
"enable_perm":
|
4763 |
"group_num": 1,
|
4764 |
"group_size": 4055,
|
4765 |
"in_features": 4096,
|
@@ -4783,7 +4783,7 @@
|
|
4783 |
"model.layers.6.mlp.up_proj": {
|
4784 |
"bias": false,
|
4785 |
"enable_norm": true,
|
4786 |
-
"enable_perm":
|
4787 |
"group_num": 1,
|
4788 |
"group_size": 4055,
|
4789 |
"in_features": 4096,
|
@@ -4807,7 +4807,7 @@
|
|
4807 |
"model.layers.6.self_attn.k_proj": {
|
4808 |
"bias": false,
|
4809 |
"enable_norm": true,
|
4810 |
-
"enable_perm":
|
4811 |
"group_num": 1,
|
4812 |
"group_size": 4055,
|
4813 |
"in_features": 4096,
|
@@ -4831,7 +4831,7 @@
|
|
4831 |
"model.layers.6.self_attn.o_proj": {
|
4832 |
"bias": false,
|
4833 |
"enable_norm": true,
|
4834 |
-
"enable_perm":
|
4835 |
"group_num": 1,
|
4836 |
"group_size": 4055,
|
4837 |
"in_features": 4096,
|
@@ -4855,7 +4855,7 @@
|
|
4855 |
"model.layers.6.self_attn.q_proj": {
|
4856 |
"bias": false,
|
4857 |
"enable_norm": true,
|
4858 |
-
"enable_perm":
|
4859 |
"group_num": 1,
|
4860 |
"group_size": 4055,
|
4861 |
"in_features": 4096,
|
@@ -4879,7 +4879,7 @@
|
|
4879 |
"model.layers.6.self_attn.v_proj": {
|
4880 |
"bias": false,
|
4881 |
"enable_norm": true,
|
4882 |
-
"enable_perm":
|
4883 |
"group_num": 1,
|
4884 |
"group_size": 4055,
|
4885 |
"in_features": 4096,
|
@@ -4903,7 +4903,7 @@
|
|
4903 |
"model.layers.7.mlp.down_proj": {
|
4904 |
"bias": false,
|
4905 |
"enable_norm": true,
|
4906 |
-
"enable_perm":
|
4907 |
"group_num": 1,
|
4908 |
"group_size": 14192,
|
4909 |
"in_features": 14336,
|
@@ -4927,7 +4927,7 @@
|
|
4927 |
"model.layers.7.mlp.gate_proj": {
|
4928 |
"bias": false,
|
4929 |
"enable_norm": true,
|
4930 |
-
"enable_perm":
|
4931 |
"group_num": 1,
|
4932 |
"group_size": 4055,
|
4933 |
"in_features": 4096,
|
@@ -4951,7 +4951,7 @@
|
|
4951 |
"model.layers.7.mlp.up_proj": {
|
4952 |
"bias": false,
|
4953 |
"enable_norm": true,
|
4954 |
-
"enable_perm":
|
4955 |
"group_num": 1,
|
4956 |
"group_size": 4055,
|
4957 |
"in_features": 4096,
|
@@ -4975,7 +4975,7 @@
|
|
4975 |
"model.layers.7.self_attn.k_proj": {
|
4976 |
"bias": false,
|
4977 |
"enable_norm": true,
|
4978 |
-
"enable_perm":
|
4979 |
"group_num": 1,
|
4980 |
"group_size": 4055,
|
4981 |
"in_features": 4096,
|
@@ -4999,7 +4999,7 @@
|
|
4999 |
"model.layers.7.self_attn.o_proj": {
|
5000 |
"bias": false,
|
5001 |
"enable_norm": true,
|
5002 |
-
"enable_perm":
|
5003 |
"group_num": 1,
|
5004 |
"group_size": 4055,
|
5005 |
"in_features": 4096,
|
@@ -5023,7 +5023,7 @@
|
|
5023 |
"model.layers.7.self_attn.q_proj": {
|
5024 |
"bias": false,
|
5025 |
"enable_norm": true,
|
5026 |
-
"enable_perm":
|
5027 |
"group_num": 1,
|
5028 |
"group_size": 4055,
|
5029 |
"in_features": 4096,
|
@@ -5047,7 +5047,7 @@
|
|
5047 |
"model.layers.7.self_attn.v_proj": {
|
5048 |
"bias": false,
|
5049 |
"enable_norm": true,
|
5050 |
-
"enable_perm":
|
5051 |
"group_num": 1,
|
5052 |
"group_size": 4055,
|
5053 |
"in_features": 4096,
|
@@ -5071,7 +5071,7 @@
|
|
5071 |
"model.layers.8.mlp.down_proj": {
|
5072 |
"bias": false,
|
5073 |
"enable_norm": true,
|
5074 |
-
"enable_perm":
|
5075 |
"group_num": 1,
|
5076 |
"group_size": 14192,
|
5077 |
"in_features": 14336,
|
@@ -5095,7 +5095,7 @@
|
|
5095 |
"model.layers.8.mlp.gate_proj": {
|
5096 |
"bias": false,
|
5097 |
"enable_norm": true,
|
5098 |
-
"enable_perm":
|
5099 |
"group_num": 1,
|
5100 |
"group_size": 4055,
|
5101 |
"in_features": 4096,
|
@@ -5119,7 +5119,7 @@
|
|
5119 |
"model.layers.8.mlp.up_proj": {
|
5120 |
"bias": false,
|
5121 |
"enable_norm": true,
|
5122 |
-
"enable_perm":
|
5123 |
"group_num": 1,
|
5124 |
"group_size": 4055,
|
5125 |
"in_features": 4096,
|
@@ -5143,7 +5143,7 @@
|
|
5143 |
"model.layers.8.self_attn.k_proj": {
|
5144 |
"bias": false,
|
5145 |
"enable_norm": true,
|
5146 |
-
"enable_perm":
|
5147 |
"group_num": 1,
|
5148 |
"group_size": 4055,
|
5149 |
"in_features": 4096,
|
@@ -5167,7 +5167,7 @@
|
|
5167 |
"model.layers.8.self_attn.o_proj": {
|
5168 |
"bias": false,
|
5169 |
"enable_norm": true,
|
5170 |
-
"enable_perm":
|
5171 |
"group_num": 1,
|
5172 |
"group_size": 4055,
|
5173 |
"in_features": 4096,
|
@@ -5191,7 +5191,7 @@
|
|
5191 |
"model.layers.8.self_attn.q_proj": {
|
5192 |
"bias": false,
|
5193 |
"enable_norm": true,
|
5194 |
-
"enable_perm":
|
5195 |
"group_num": 1,
|
5196 |
"group_size": 4055,
|
5197 |
"in_features": 4096,
|
@@ -5215,7 +5215,7 @@
|
|
5215 |
"model.layers.8.self_attn.v_proj": {
|
5216 |
"bias": false,
|
5217 |
"enable_norm": true,
|
5218 |
-
"enable_perm":
|
5219 |
"group_num": 1,
|
5220 |
"group_size": 4055,
|
5221 |
"in_features": 4096,
|
@@ -5239,7 +5239,7 @@
|
|
5239 |
"model.layers.9.mlp.down_proj": {
|
5240 |
"bias": false,
|
5241 |
"enable_norm": true,
|
5242 |
-
"enable_perm":
|
5243 |
"group_num": 1,
|
5244 |
"group_size": 14192,
|
5245 |
"in_features": 14336,
|
@@ -5263,7 +5263,7 @@
|
|
5263 |
"model.layers.9.mlp.gate_proj": {
|
5264 |
"bias": false,
|
5265 |
"enable_norm": true,
|
5266 |
-
"enable_perm":
|
5267 |
"group_num": 1,
|
5268 |
"group_size": 4055,
|
5269 |
"in_features": 4096,
|
@@ -5287,7 +5287,7 @@
|
|
5287 |
"model.layers.9.mlp.up_proj": {
|
5288 |
"bias": false,
|
5289 |
"enable_norm": true,
|
5290 |
-
"enable_perm":
|
5291 |
"group_num": 1,
|
5292 |
"group_size": 4055,
|
5293 |
"in_features": 4096,
|
@@ -5311,7 +5311,7 @@
|
|
5311 |
"model.layers.9.self_attn.k_proj": {
|
5312 |
"bias": false,
|
5313 |
"enable_norm": true,
|
5314 |
-
"enable_perm":
|
5315 |
"group_num": 1,
|
5316 |
"group_size": 4055,
|
5317 |
"in_features": 4096,
|
@@ -5335,7 +5335,7 @@
|
|
5335 |
"model.layers.9.self_attn.o_proj": {
|
5336 |
"bias": false,
|
5337 |
"enable_norm": true,
|
5338 |
-
"enable_perm":
|
5339 |
"group_num": 1,
|
5340 |
"group_size": 4055,
|
5341 |
"in_features": 4096,
|
@@ -5359,7 +5359,7 @@
|
|
5359 |
"model.layers.9.self_attn.q_proj": {
|
5360 |
"bias": false,
|
5361 |
"enable_norm": true,
|
5362 |
-
"enable_perm":
|
5363 |
"group_num": 1,
|
5364 |
"group_size": 4055,
|
5365 |
"in_features": 4096,
|
@@ -5383,7 +5383,7 @@
|
|
5383 |
"model.layers.9.self_attn.v_proj": {
|
5384 |
"bias": false,
|
5385 |
"enable_norm": true,
|
5386 |
-
"enable_perm":
|
5387 |
"group_num": 1,
|
5388 |
"group_size": 4055,
|
5389 |
"in_features": 4096,
|
|
|
31 |
"model.layers.0.mlp.down_proj": {
|
32 |
"bias": false,
|
33 |
"enable_norm": true,
|
34 |
+
"enable_perm": false,
|
35 |
"group_num": 1,
|
36 |
"group_size": 14192,
|
37 |
"in_features": 14336,
|
|
|
55 |
"model.layers.0.mlp.gate_proj": {
|
56 |
"bias": false,
|
57 |
"enable_norm": true,
|
58 |
+
"enable_perm": false,
|
59 |
"group_num": 1,
|
60 |
"group_size": 4055,
|
61 |
"in_features": 4096,
|
|
|
79 |
"model.layers.0.mlp.up_proj": {
|
80 |
"bias": false,
|
81 |
"enable_norm": true,
|
82 |
+
"enable_perm": false,
|
83 |
"group_num": 1,
|
84 |
"group_size": 4055,
|
85 |
"in_features": 4096,
|
|
|
103 |
"model.layers.0.self_attn.k_proj": {
|
104 |
"bias": false,
|
105 |
"enable_norm": true,
|
106 |
+
"enable_perm": false,
|
107 |
"group_num": 1,
|
108 |
"group_size": 4055,
|
109 |
"in_features": 4096,
|
|
|
127 |
"model.layers.0.self_attn.o_proj": {
|
128 |
"bias": false,
|
129 |
"enable_norm": true,
|
130 |
+
"enable_perm": false,
|
131 |
"group_num": 1,
|
132 |
"group_size": 4055,
|
133 |
"in_features": 4096,
|
|
|
151 |
"model.layers.0.self_attn.q_proj": {
|
152 |
"bias": false,
|
153 |
"enable_norm": true,
|
154 |
+
"enable_perm": false,
|
155 |
"group_num": 1,
|
156 |
"group_size": 4055,
|
157 |
"in_features": 4096,
|
|
|
175 |
"model.layers.0.self_attn.v_proj": {
|
176 |
"bias": false,
|
177 |
"enable_norm": true,
|
178 |
+
"enable_perm": false,
|
179 |
"group_num": 1,
|
180 |
"group_size": 4055,
|
181 |
"in_features": 4096,
|
|
|
199 |
"model.layers.1.mlp.down_proj": {
|
200 |
"bias": false,
|
201 |
"enable_norm": true,
|
202 |
+
"enable_perm": false,
|
203 |
"group_num": 1,
|
204 |
"group_size": 14192,
|
205 |
"in_features": 14336,
|
|
|
223 |
"model.layers.1.mlp.gate_proj": {
|
224 |
"bias": false,
|
225 |
"enable_norm": true,
|
226 |
+
"enable_perm": false,
|
227 |
"group_num": 1,
|
228 |
"group_size": 4055,
|
229 |
"in_features": 4096,
|
|
|
247 |
"model.layers.1.mlp.up_proj": {
|
248 |
"bias": false,
|
249 |
"enable_norm": true,
|
250 |
+
"enable_perm": false,
|
251 |
"group_num": 1,
|
252 |
"group_size": 4055,
|
253 |
"in_features": 4096,
|
|
|
271 |
"model.layers.1.self_attn.k_proj": {
|
272 |
"bias": false,
|
273 |
"enable_norm": true,
|
274 |
+
"enable_perm": false,
|
275 |
"group_num": 1,
|
276 |
"group_size": 4055,
|
277 |
"in_features": 4096,
|
|
|
295 |
"model.layers.1.self_attn.o_proj": {
|
296 |
"bias": false,
|
297 |
"enable_norm": true,
|
298 |
+
"enable_perm": false,
|
299 |
"group_num": 1,
|
300 |
"group_size": 4055,
|
301 |
"in_features": 4096,
|
|
|
319 |
"model.layers.1.self_attn.q_proj": {
|
320 |
"bias": false,
|
321 |
"enable_norm": true,
|
322 |
+
"enable_perm": false,
|
323 |
"group_num": 1,
|
324 |
"group_size": 4055,
|
325 |
"in_features": 4096,
|
|
|
343 |
"model.layers.1.self_attn.v_proj": {
|
344 |
"bias": false,
|
345 |
"enable_norm": true,
|
346 |
+
"enable_perm": false,
|
347 |
"group_num": 1,
|
348 |
"group_size": 4055,
|
349 |
"in_features": 4096,
|
|
|
367 |
"model.layers.10.mlp.down_proj": {
|
368 |
"bias": false,
|
369 |
"enable_norm": true,
|
370 |
+
"enable_perm": false,
|
371 |
"group_num": 1,
|
372 |
"group_size": 14192,
|
373 |
"in_features": 14336,
|
|
|
391 |
"model.layers.10.mlp.gate_proj": {
|
392 |
"bias": false,
|
393 |
"enable_norm": true,
|
394 |
+
"enable_perm": false,
|
395 |
"group_num": 1,
|
396 |
"group_size": 4055,
|
397 |
"in_features": 4096,
|
|
|
415 |
"model.layers.10.mlp.up_proj": {
|
416 |
"bias": false,
|
417 |
"enable_norm": true,
|
418 |
+
"enable_perm": false,
|
419 |
"group_num": 1,
|
420 |
"group_size": 4055,
|
421 |
"in_features": 4096,
|
|
|
439 |
"model.layers.10.self_attn.k_proj": {
|
440 |
"bias": false,
|
441 |
"enable_norm": true,
|
442 |
+
"enable_perm": false,
|
443 |
"group_num": 1,
|
444 |
"group_size": 4055,
|
445 |
"in_features": 4096,
|
|
|
463 |
"model.layers.10.self_attn.o_proj": {
|
464 |
"bias": false,
|
465 |
"enable_norm": true,
|
466 |
+
"enable_perm": false,
|
467 |
"group_num": 1,
|
468 |
"group_size": 4055,
|
469 |
"in_features": 4096,
|
|
|
487 |
"model.layers.10.self_attn.q_proj": {
|
488 |
"bias": false,
|
489 |
"enable_norm": true,
|
490 |
+
"enable_perm": false,
|
491 |
"group_num": 1,
|
492 |
"group_size": 4055,
|
493 |
"in_features": 4096,
|
|
|
511 |
"model.layers.10.self_attn.v_proj": {
|
512 |
"bias": false,
|
513 |
"enable_norm": true,
|
514 |
+
"enable_perm": false,
|
515 |
"group_num": 1,
|
516 |
"group_size": 4055,
|
517 |
"in_features": 4096,
|
|
|
535 |
"model.layers.11.mlp.down_proj": {
|
536 |
"bias": false,
|
537 |
"enable_norm": true,
|
538 |
+
"enable_perm": false,
|
539 |
"group_num": 1,
|
540 |
"group_size": 14192,
|
541 |
"in_features": 14336,
|
|
|
559 |
"model.layers.11.mlp.gate_proj": {
|
560 |
"bias": false,
|
561 |
"enable_norm": true,
|
562 |
+
"enable_perm": false,
|
563 |
"group_num": 1,
|
564 |
"group_size": 4055,
|
565 |
"in_features": 4096,
|
|
|
583 |
"model.layers.11.mlp.up_proj": {
|
584 |
"bias": false,
|
585 |
"enable_norm": true,
|
586 |
+
"enable_perm": false,
|
587 |
"group_num": 1,
|
588 |
"group_size": 4055,
|
589 |
"in_features": 4096,
|
|
|
607 |
"model.layers.11.self_attn.k_proj": {
|
608 |
"bias": false,
|
609 |
"enable_norm": true,
|
610 |
+
"enable_perm": false,
|
611 |
"group_num": 1,
|
612 |
"group_size": 4055,
|
613 |
"in_features": 4096,
|
|
|
631 |
"model.layers.11.self_attn.o_proj": {
|
632 |
"bias": false,
|
633 |
"enable_norm": true,
|
634 |
+
"enable_perm": false,
|
635 |
"group_num": 1,
|
636 |
"group_size": 4055,
|
637 |
"in_features": 4096,
|
|
|
655 |
"model.layers.11.self_attn.q_proj": {
|
656 |
"bias": false,
|
657 |
"enable_norm": true,
|
658 |
+
"enable_perm": false,
|
659 |
"group_num": 1,
|
660 |
"group_size": 4055,
|
661 |
"in_features": 4096,
|
|
|
679 |
"model.layers.11.self_attn.v_proj": {
|
680 |
"bias": false,
|
681 |
"enable_norm": true,
|
682 |
+
"enable_perm": false,
|
683 |
"group_num": 1,
|
684 |
"group_size": 4055,
|
685 |
"in_features": 4096,
|
|
|
703 |
"model.layers.12.mlp.down_proj": {
|
704 |
"bias": false,
|
705 |
"enable_norm": true,
|
706 |
+
"enable_perm": false,
|
707 |
"group_num": 1,
|
708 |
"group_size": 14192,
|
709 |
"in_features": 14336,
|
|
|
727 |
"model.layers.12.mlp.gate_proj": {
|
728 |
"bias": false,
|
729 |
"enable_norm": true,
|
730 |
+
"enable_perm": false,
|
731 |
"group_num": 1,
|
732 |
"group_size": 4055,
|
733 |
"in_features": 4096,
|
|
|
751 |
"model.layers.12.mlp.up_proj": {
|
752 |
"bias": false,
|
753 |
"enable_norm": true,
|
754 |
+
"enable_perm": false,
|
755 |
"group_num": 1,
|
756 |
"group_size": 4055,
|
757 |
"in_features": 4096,
|
|
|
775 |
"model.layers.12.self_attn.k_proj": {
|
776 |
"bias": false,
|
777 |
"enable_norm": true,
|
778 |
+
"enable_perm": false,
|
779 |
"group_num": 1,
|
780 |
"group_size": 4055,
|
781 |
"in_features": 4096,
|
|
|
799 |
"model.layers.12.self_attn.o_proj": {
|
800 |
"bias": false,
|
801 |
"enable_norm": true,
|
802 |
+
"enable_perm": false,
|
803 |
"group_num": 1,
|
804 |
"group_size": 4055,
|
805 |
"in_features": 4096,
|
|
|
823 |
"model.layers.12.self_attn.q_proj": {
|
824 |
"bias": false,
|
825 |
"enable_norm": true,
|
826 |
+
"enable_perm": false,
|
827 |
"group_num": 1,
|
828 |
"group_size": 4055,
|
829 |
"in_features": 4096,
|
|
|
847 |
"model.layers.12.self_attn.v_proj": {
|
848 |
"bias": false,
|
849 |
"enable_norm": true,
|
850 |
+
"enable_perm": false,
|
851 |
"group_num": 1,
|
852 |
"group_size": 4055,
|
853 |
"in_features": 4096,
|
|
|
871 |
"model.layers.13.mlp.down_proj": {
|
872 |
"bias": false,
|
873 |
"enable_norm": true,
|
874 |
+
"enable_perm": false,
|
875 |
"group_num": 1,
|
876 |
"group_size": 14192,
|
877 |
"in_features": 14336,
|
|
|
895 |
"model.layers.13.mlp.gate_proj": {
|
896 |
"bias": false,
|
897 |
"enable_norm": true,
|
898 |
+
"enable_perm": false,
|
899 |
"group_num": 1,
|
900 |
"group_size": 4055,
|
901 |
"in_features": 4096,
|
|
|
919 |
"model.layers.13.mlp.up_proj": {
|
920 |
"bias": false,
|
921 |
"enable_norm": true,
|
922 |
+
"enable_perm": false,
|
923 |
"group_num": 1,
|
924 |
"group_size": 4055,
|
925 |
"in_features": 4096,
|
|
|
943 |
"model.layers.13.self_attn.k_proj": {
|
944 |
"bias": false,
|
945 |
"enable_norm": true,
|
946 |
+
"enable_perm": false,
|
947 |
"group_num": 1,
|
948 |
"group_size": 4055,
|
949 |
"in_features": 4096,
|
|
|
967 |
"model.layers.13.self_attn.o_proj": {
|
968 |
"bias": false,
|
969 |
"enable_norm": true,
|
970 |
+
"enable_perm": false,
|
971 |
"group_num": 1,
|
972 |
"group_size": 4055,
|
973 |
"in_features": 4096,
|
|
|
991 |
"model.layers.13.self_attn.q_proj": {
|
992 |
"bias": false,
|
993 |
"enable_norm": true,
|
994 |
+
"enable_perm": false,
|
995 |
"group_num": 1,
|
996 |
"group_size": 4055,
|
997 |
"in_features": 4096,
|
|
|
1015 |
"model.layers.13.self_attn.v_proj": {
|
1016 |
"bias": false,
|
1017 |
"enable_norm": true,
|
1018 |
+
"enable_perm": false,
|
1019 |
"group_num": 1,
|
1020 |
"group_size": 4055,
|
1021 |
"in_features": 4096,
|
|
|
1039 |
"model.layers.14.mlp.down_proj": {
|
1040 |
"bias": false,
|
1041 |
"enable_norm": true,
|
1042 |
+
"enable_perm": false,
|
1043 |
"group_num": 1,
|
1044 |
"group_size": 14192,
|
1045 |
"in_features": 14336,
|
|
|
1063 |
"model.layers.14.mlp.gate_proj": {
|
1064 |
"bias": false,
|
1065 |
"enable_norm": true,
|
1066 |
+
"enable_perm": false,
|
1067 |
"group_num": 1,
|
1068 |
"group_size": 4055,
|
1069 |
"in_features": 4096,
|
|
|
1087 |
"model.layers.14.mlp.up_proj": {
|
1088 |
"bias": false,
|
1089 |
"enable_norm": true,
|
1090 |
+
"enable_perm": false,
|
1091 |
"group_num": 1,
|
1092 |
"group_size": 4055,
|
1093 |
"in_features": 4096,
|
|
|
1111 |
"model.layers.14.self_attn.k_proj": {
|
1112 |
"bias": false,
|
1113 |
"enable_norm": true,
|
1114 |
+
"enable_perm": false,
|
1115 |
"group_num": 1,
|
1116 |
"group_size": 4055,
|
1117 |
"in_features": 4096,
|
|
|
1135 |
"model.layers.14.self_attn.o_proj": {
|
1136 |
"bias": false,
|
1137 |
"enable_norm": true,
|
1138 |
+
"enable_perm": false,
|
1139 |
"group_num": 1,
|
1140 |
"group_size": 4055,
|
1141 |
"in_features": 4096,
|
|
|
1159 |
"model.layers.14.self_attn.q_proj": {
|
1160 |
"bias": false,
|
1161 |
"enable_norm": true,
|
1162 |
+
"enable_perm": false,
|
1163 |
"group_num": 1,
|
1164 |
"group_size": 4055,
|
1165 |
"in_features": 4096,
|
|
|
1183 |
"model.layers.14.self_attn.v_proj": {
|
1184 |
"bias": false,
|
1185 |
"enable_norm": true,
|
1186 |
+
"enable_perm": false,
|
1187 |
"group_num": 1,
|
1188 |
"group_size": 4055,
|
1189 |
"in_features": 4096,
|
|
|
1207 |
"model.layers.15.mlp.down_proj": {
|
1208 |
"bias": false,
|
1209 |
"enable_norm": true,
|
1210 |
+
"enable_perm": false,
|
1211 |
"group_num": 1,
|
1212 |
"group_size": 14192,
|
1213 |
"in_features": 14336,
|
|
|
1231 |
"model.layers.15.mlp.gate_proj": {
|
1232 |
"bias": false,
|
1233 |
"enable_norm": true,
|
1234 |
+
"enable_perm": false,
|
1235 |
"group_num": 1,
|
1236 |
"group_size": 4055,
|
1237 |
"in_features": 4096,
|
|
|
1255 |
"model.layers.15.mlp.up_proj": {
|
1256 |
"bias": false,
|
1257 |
"enable_norm": true,
|
1258 |
+
"enable_perm": false,
|
1259 |
"group_num": 1,
|
1260 |
"group_size": 4055,
|
1261 |
"in_features": 4096,
|
|
|
1279 |
"model.layers.15.self_attn.k_proj": {
|
1280 |
"bias": false,
|
1281 |
"enable_norm": true,
|
1282 |
+
"enable_perm": false,
|
1283 |
"group_num": 1,
|
1284 |
"group_size": 4055,
|
1285 |
"in_features": 4096,
|
|
|
1303 |
"model.layers.15.self_attn.o_proj": {
|
1304 |
"bias": false,
|
1305 |
"enable_norm": true,
|
1306 |
+
"enable_perm": false,
|
1307 |
"group_num": 1,
|
1308 |
"group_size": 4055,
|
1309 |
"in_features": 4096,
|
|
|
1327 |
"model.layers.15.self_attn.q_proj": {
|
1328 |
"bias": false,
|
1329 |
"enable_norm": true,
|
1330 |
+
"enable_perm": false,
|
1331 |
"group_num": 1,
|
1332 |
"group_size": 4055,
|
1333 |
"in_features": 4096,
|
|
|
1351 |
"model.layers.15.self_attn.v_proj": {
|
1352 |
"bias": false,
|
1353 |
"enable_norm": true,
|
1354 |
+
"enable_perm": false,
|
1355 |
"group_num": 1,
|
1356 |
"group_size": 4055,
|
1357 |
"in_features": 4096,
|
|
|
1375 |
"model.layers.16.mlp.down_proj": {
|
1376 |
"bias": false,
|
1377 |
"enable_norm": true,
|
1378 |
+
"enable_perm": false,
|
1379 |
"group_num": 1,
|
1380 |
"group_size": 14192,
|
1381 |
"in_features": 14336,
|
|
|
1399 |
"model.layers.16.mlp.gate_proj": {
|
1400 |
"bias": false,
|
1401 |
"enable_norm": true,
|
1402 |
+
"enable_perm": false,
|
1403 |
"group_num": 1,
|
1404 |
"group_size": 4055,
|
1405 |
"in_features": 4096,
|
|
|
1423 |
"model.layers.16.mlp.up_proj": {
|
1424 |
"bias": false,
|
1425 |
"enable_norm": true,
|
1426 |
+
"enable_perm": false,
|
1427 |
"group_num": 1,
|
1428 |
"group_size": 4055,
|
1429 |
"in_features": 4096,
|
|
|
1447 |
"model.layers.16.self_attn.k_proj": {
|
1448 |
"bias": false,
|
1449 |
"enable_norm": true,
|
1450 |
+
"enable_perm": false,
|
1451 |
"group_num": 1,
|
1452 |
"group_size": 4055,
|
1453 |
"in_features": 4096,
|
|
|
1471 |
"model.layers.16.self_attn.o_proj": {
|
1472 |
"bias": false,
|
1473 |
"enable_norm": true,
|
1474 |
+
"enable_perm": false,
|
1475 |
"group_num": 1,
|
1476 |
"group_size": 4055,
|
1477 |
"in_features": 4096,
|
|
|
1495 |
"model.layers.16.self_attn.q_proj": {
|
1496 |
"bias": false,
|
1497 |
"enable_norm": true,
|
1498 |
+
"enable_perm": false,
|
1499 |
"group_num": 1,
|
1500 |
"group_size": 4055,
|
1501 |
"in_features": 4096,
|
|
|
1519 |
"model.layers.16.self_attn.v_proj": {
|
1520 |
"bias": false,
|
1521 |
"enable_norm": true,
|
1522 |
+
"enable_perm": false,
|
1523 |
"group_num": 1,
|
1524 |
"group_size": 4055,
|
1525 |
"in_features": 4096,
|
|
|
1543 |
"model.layers.17.mlp.down_proj": {
|
1544 |
"bias": false,
|
1545 |
"enable_norm": true,
|
1546 |
+
"enable_perm": false,
|
1547 |
"group_num": 1,
|
1548 |
"group_size": 14192,
|
1549 |
"in_features": 14336,
|
|
|
1567 |
"model.layers.17.mlp.gate_proj": {
|
1568 |
"bias": false,
|
1569 |
"enable_norm": true,
|
1570 |
+
"enable_perm": false,
|
1571 |
"group_num": 1,
|
1572 |
"group_size": 4055,
|
1573 |
"in_features": 4096,
|
|
|
1591 |
"model.layers.17.mlp.up_proj": {
|
1592 |
"bias": false,
|
1593 |
"enable_norm": true,
|
1594 |
+
"enable_perm": false,
|
1595 |
"group_num": 1,
|
1596 |
"group_size": 4055,
|
1597 |
"in_features": 4096,
|
|
|
1615 |
"model.layers.17.self_attn.k_proj": {
|
1616 |
"bias": false,
|
1617 |
"enable_norm": true,
|
1618 |
+
"enable_perm": false,
|
1619 |
"group_num": 1,
|
1620 |
"group_size": 4055,
|
1621 |
"in_features": 4096,
|
|
|
1639 |
"model.layers.17.self_attn.o_proj": {
|
1640 |
"bias": false,
|
1641 |
"enable_norm": true,
|
1642 |
+
"enable_perm": false,
|
1643 |
"group_num": 1,
|
1644 |
"group_size": 4055,
|
1645 |
"in_features": 4096,
|
|
|
1663 |
"model.layers.17.self_attn.q_proj": {
|
1664 |
"bias": false,
|
1665 |
"enable_norm": true,
|
1666 |
+
"enable_perm": false,
|
1667 |
"group_num": 1,
|
1668 |
"group_size": 4055,
|
1669 |
"in_features": 4096,
|
|
|
1687 |
"model.layers.17.self_attn.v_proj": {
|
1688 |
"bias": false,
|
1689 |
"enable_norm": true,
|
1690 |
+
"enable_perm": false,
|
1691 |
"group_num": 1,
|
1692 |
"group_size": 4055,
|
1693 |
"in_features": 4096,
|
|
|
1711 |
"model.layers.18.mlp.down_proj": {
|
1712 |
"bias": false,
|
1713 |
"enable_norm": true,
|
1714 |
+
"enable_perm": false,
|
1715 |
"group_num": 1,
|
1716 |
"group_size": 14192,
|
1717 |
"in_features": 14336,
|
|
|
1735 |
"model.layers.18.mlp.gate_proj": {
|
1736 |
"bias": false,
|
1737 |
"enable_norm": true,
|
1738 |
+
"enable_perm": false,
|
1739 |
"group_num": 1,
|
1740 |
"group_size": 4055,
|
1741 |
"in_features": 4096,
|
|
|
1759 |
"model.layers.18.mlp.up_proj": {
|
1760 |
"bias": false,
|
1761 |
"enable_norm": true,
|
1762 |
+
"enable_perm": false,
|
1763 |
"group_num": 1,
|
1764 |
"group_size": 4055,
|
1765 |
"in_features": 4096,
|
|
|
1783 |
"model.layers.18.self_attn.k_proj": {
|
1784 |
"bias": false,
|
1785 |
"enable_norm": true,
|
1786 |
+
"enable_perm": false,
|
1787 |
"group_num": 1,
|
1788 |
"group_size": 4055,
|
1789 |
"in_features": 4096,
|
|
|
1807 |
"model.layers.18.self_attn.o_proj": {
|
1808 |
"bias": false,
|
1809 |
"enable_norm": true,
|
1810 |
+
"enable_perm": false,
|
1811 |
"group_num": 1,
|
1812 |
"group_size": 4055,
|
1813 |
"in_features": 4096,
|
|
|
1831 |
"model.layers.18.self_attn.q_proj": {
|
1832 |
"bias": false,
|
1833 |
"enable_norm": true,
|
1834 |
+
"enable_perm": false,
|
1835 |
"group_num": 1,
|
1836 |
"group_size": 4055,
|
1837 |
"in_features": 4096,
|
|
|
1855 |
"model.layers.18.self_attn.v_proj": {
|
1856 |
"bias": false,
|
1857 |
"enable_norm": true,
|
1858 |
+
"enable_perm": false,
|
1859 |
"group_num": 1,
|
1860 |
"group_size": 4055,
|
1861 |
"in_features": 4096,
|
|
|
1879 |
"model.layers.19.mlp.down_proj": {
|
1880 |
"bias": false,
|
1881 |
"enable_norm": true,
|
1882 |
+
"enable_perm": false,
|
1883 |
"group_num": 1,
|
1884 |
"group_size": 14192,
|
1885 |
"in_features": 14336,
|
|
|
1903 |
"model.layers.19.mlp.gate_proj": {
|
1904 |
"bias": false,
|
1905 |
"enable_norm": true,
|
1906 |
+
"enable_perm": false,
|
1907 |
"group_num": 1,
|
1908 |
"group_size": 4055,
|
1909 |
"in_features": 4096,
|
|
|
1927 |
"model.layers.19.mlp.up_proj": {
|
1928 |
"bias": false,
|
1929 |
"enable_norm": true,
|
1930 |
+
"enable_perm": false,
|
1931 |
"group_num": 1,
|
1932 |
"group_size": 4055,
|
1933 |
"in_features": 4096,
|
|
|
1951 |
"model.layers.19.self_attn.k_proj": {
|
1952 |
"bias": false,
|
1953 |
"enable_norm": true,
|
1954 |
+
"enable_perm": false,
|
1955 |
"group_num": 1,
|
1956 |
"group_size": 4055,
|
1957 |
"in_features": 4096,
|
|
|
1975 |
"model.layers.19.self_attn.o_proj": {
|
1976 |
"bias": false,
|
1977 |
"enable_norm": true,
|
1978 |
+
"enable_perm": false,
|
1979 |
"group_num": 1,
|
1980 |
"group_size": 4055,
|
1981 |
"in_features": 4096,
|
|
|
1999 |
"model.layers.19.self_attn.q_proj": {
|
2000 |
"bias": false,
|
2001 |
"enable_norm": true,
|
2002 |
+
"enable_perm": false,
|
2003 |
"group_num": 1,
|
2004 |
"group_size": 4055,
|
2005 |
"in_features": 4096,
|
|
|
2023 |
"model.layers.19.self_attn.v_proj": {
|
2024 |
"bias": false,
|
2025 |
"enable_norm": true,
|
2026 |
+
"enable_perm": false,
|
2027 |
"group_num": 1,
|
2028 |
"group_size": 4055,
|
2029 |
"in_features": 4096,
|
|
|
2047 |
"model.layers.2.mlp.down_proj": {
|
2048 |
"bias": false,
|
2049 |
"enable_norm": true,
|
2050 |
+
"enable_perm": false,
|
2051 |
"group_num": 1,
|
2052 |
"group_size": 14192,
|
2053 |
"in_features": 14336,
|
|
|
2071 |
"model.layers.2.mlp.gate_proj": {
|
2072 |
"bias": false,
|
2073 |
"enable_norm": true,
|
2074 |
+
"enable_perm": false,
|
2075 |
"group_num": 1,
|
2076 |
"group_size": 4055,
|
2077 |
"in_features": 4096,
|
|
|
2095 |
"model.layers.2.mlp.up_proj": {
|
2096 |
"bias": false,
|
2097 |
"enable_norm": true,
|
2098 |
+
"enable_perm": false,
|
2099 |
"group_num": 1,
|
2100 |
"group_size": 4055,
|
2101 |
"in_features": 4096,
|
|
|
2119 |
"model.layers.2.self_attn.k_proj": {
|
2120 |
"bias": false,
|
2121 |
"enable_norm": true,
|
2122 |
+
"enable_perm": false,
|
2123 |
"group_num": 1,
|
2124 |
"group_size": 4055,
|
2125 |
"in_features": 4096,
|
|
|
2143 |
"model.layers.2.self_attn.o_proj": {
|
2144 |
"bias": false,
|
2145 |
"enable_norm": true,
|
2146 |
+
"enable_perm": false,
|
2147 |
"group_num": 1,
|
2148 |
"group_size": 4055,
|
2149 |
"in_features": 4096,
|
|
|
2167 |
"model.layers.2.self_attn.q_proj": {
|
2168 |
"bias": false,
|
2169 |
"enable_norm": true,
|
2170 |
+
"enable_perm": false,
|
2171 |
"group_num": 1,
|
2172 |
"group_size": 4055,
|
2173 |
"in_features": 4096,
|
|
|
2191 |
"model.layers.2.self_attn.v_proj": {
|
2192 |
"bias": false,
|
2193 |
"enable_norm": true,
|
2194 |
+
"enable_perm": false,
|
2195 |
"group_num": 1,
|
2196 |
"group_size": 4055,
|
2197 |
"in_features": 4096,
|
|
|
2215 |
"model.layers.20.mlp.down_proj": {
|
2216 |
"bias": false,
|
2217 |
"enable_norm": true,
|
2218 |
+
"enable_perm": false,
|
2219 |
"group_num": 1,
|
2220 |
"group_size": 14192,
|
2221 |
"in_features": 14336,
|
|
|
2239 |
"model.layers.20.mlp.gate_proj": {
|
2240 |
"bias": false,
|
2241 |
"enable_norm": true,
|
2242 |
+
"enable_perm": false,
|
2243 |
"group_num": 1,
|
2244 |
"group_size": 4055,
|
2245 |
"in_features": 4096,
|
|
|
2263 |
"model.layers.20.mlp.up_proj": {
|
2264 |
"bias": false,
|
2265 |
"enable_norm": true,
|
2266 |
+
"enable_perm": false,
|
2267 |
"group_num": 1,
|
2268 |
"group_size": 4055,
|
2269 |
"in_features": 4096,
|
|
|
2287 |
"model.layers.20.self_attn.k_proj": {
|
2288 |
"bias": false,
|
2289 |
"enable_norm": true,
|
2290 |
+
"enable_perm": false,
|
2291 |
"group_num": 1,
|
2292 |
"group_size": 4055,
|
2293 |
"in_features": 4096,
|
|
|
2311 |
"model.layers.20.self_attn.o_proj": {
|
2312 |
"bias": false,
|
2313 |
"enable_norm": true,
|
2314 |
+
"enable_perm": false,
|
2315 |
"group_num": 1,
|
2316 |
"group_size": 4055,
|
2317 |
"in_features": 4096,
|
|
|
2335 |
"model.layers.20.self_attn.q_proj": {
|
2336 |
"bias": false,
|
2337 |
"enable_norm": true,
|
2338 |
+
"enable_perm": false,
|
2339 |
"group_num": 1,
|
2340 |
"group_size": 4055,
|
2341 |
"in_features": 4096,
|
|
|
2359 |
"model.layers.20.self_attn.v_proj": {
|
2360 |
"bias": false,
|
2361 |
"enable_norm": true,
|
2362 |
+
"enable_perm": false,
|
2363 |
"group_num": 1,
|
2364 |
"group_size": 4055,
|
2365 |
"in_features": 4096,
|
|
|
2383 |
"model.layers.21.mlp.down_proj": {
|
2384 |
"bias": false,
|
2385 |
"enable_norm": true,
|
2386 |
+
"enable_perm": false,
|
2387 |
"group_num": 1,
|
2388 |
"group_size": 14192,
|
2389 |
"in_features": 14336,
|
|
|
2407 |
"model.layers.21.mlp.gate_proj": {
|
2408 |
"bias": false,
|
2409 |
"enable_norm": true,
|
2410 |
+
"enable_perm": false,
|
2411 |
"group_num": 1,
|
2412 |
"group_size": 4055,
|
2413 |
"in_features": 4096,
|
|
|
2431 |
"model.layers.21.mlp.up_proj": {
|
2432 |
"bias": false,
|
2433 |
"enable_norm": true,
|
2434 |
+
"enable_perm": false,
|
2435 |
"group_num": 1,
|
2436 |
"group_size": 4055,
|
2437 |
"in_features": 4096,
|
|
|
2455 |
"model.layers.21.self_attn.k_proj": {
|
2456 |
"bias": false,
|
2457 |
"enable_norm": true,
|
2458 |
+
"enable_perm": false,
|
2459 |
"group_num": 1,
|
2460 |
"group_size": 4055,
|
2461 |
"in_features": 4096,
|
|
|
2479 |
"model.layers.21.self_attn.o_proj": {
|
2480 |
"bias": false,
|
2481 |
"enable_norm": true,
|
2482 |
+
"enable_perm": false,
|
2483 |
"group_num": 1,
|
2484 |
"group_size": 4055,
|
2485 |
"in_features": 4096,
|
|
|
2503 |
"model.layers.21.self_attn.q_proj": {
|
2504 |
"bias": false,
|
2505 |
"enable_norm": true,
|
2506 |
+
"enable_perm": false,
|
2507 |
"group_num": 1,
|
2508 |
"group_size": 4055,
|
2509 |
"in_features": 4096,
|
|
|
2527 |
"model.layers.21.self_attn.v_proj": {
|
2528 |
"bias": false,
|
2529 |
"enable_norm": true,
|
2530 |
+
"enable_perm": false,
|
2531 |
"group_num": 1,
|
2532 |
"group_size": 4055,
|
2533 |
"in_features": 4096,
|
|
|
2551 |
"model.layers.22.mlp.down_proj": {
|
2552 |
"bias": false,
|
2553 |
"enable_norm": true,
|
2554 |
+
"enable_perm": false,
|
2555 |
"group_num": 1,
|
2556 |
"group_size": 14192,
|
2557 |
"in_features": 14336,
|
|
|
2575 |
"model.layers.22.mlp.gate_proj": {
|
2576 |
"bias": false,
|
2577 |
"enable_norm": true,
|
2578 |
+
"enable_perm": false,
|
2579 |
"group_num": 1,
|
2580 |
"group_size": 4055,
|
2581 |
"in_features": 4096,
|
|
|
2599 |
"model.layers.22.mlp.up_proj": {
|
2600 |
"bias": false,
|
2601 |
"enable_norm": true,
|
2602 |
+
"enable_perm": false,
|
2603 |
"group_num": 1,
|
2604 |
"group_size": 4055,
|
2605 |
"in_features": 4096,
|
|
|
2623 |
"model.layers.22.self_attn.k_proj": {
|
2624 |
"bias": false,
|
2625 |
"enable_norm": true,
|
2626 |
+
"enable_perm": false,
|
2627 |
"group_num": 1,
|
2628 |
"group_size": 4055,
|
2629 |
"in_features": 4096,
|
|
|
2647 |
"model.layers.22.self_attn.o_proj": {
|
2648 |
"bias": false,
|
2649 |
"enable_norm": true,
|
2650 |
+
"enable_perm": false,
|
2651 |
"group_num": 1,
|
2652 |
"group_size": 4055,
|
2653 |
"in_features": 4096,
|
|
|
2671 |
"model.layers.22.self_attn.q_proj": {
|
2672 |
"bias": false,
|
2673 |
"enable_norm": true,
|
2674 |
+
"enable_perm": false,
|
2675 |
"group_num": 1,
|
2676 |
"group_size": 4055,
|
2677 |
"in_features": 4096,
|
|
|
2695 |
"model.layers.22.self_attn.v_proj": {
|
2696 |
"bias": false,
|
2697 |
"enable_norm": true,
|
2698 |
+
"enable_perm": false,
|
2699 |
"group_num": 1,
|
2700 |
"group_size": 4055,
|
2701 |
"in_features": 4096,
|
|
|
2719 |
"model.layers.23.mlp.down_proj": {
|
2720 |
"bias": false,
|
2721 |
"enable_norm": true,
|
2722 |
+
"enable_perm": false,
|
2723 |
"group_num": 1,
|
2724 |
"group_size": 14192,
|
2725 |
"in_features": 14336,
|
|
|
2743 |
"model.layers.23.mlp.gate_proj": {
|
2744 |
"bias": false,
|
2745 |
"enable_norm": true,
|
2746 |
+
"enable_perm": false,
|
2747 |
"group_num": 1,
|
2748 |
"group_size": 4055,
|
2749 |
"in_features": 4096,
|
|
|
2767 |
"model.layers.23.mlp.up_proj": {
|
2768 |
"bias": false,
|
2769 |
"enable_norm": true,
|
2770 |
+
"enable_perm": false,
|
2771 |
"group_num": 1,
|
2772 |
"group_size": 4055,
|
2773 |
"in_features": 4096,
|
|
|
2791 |
"model.layers.23.self_attn.k_proj": {
|
2792 |
"bias": false,
|
2793 |
"enable_norm": true,
|
2794 |
+
"enable_perm": false,
|
2795 |
"group_num": 1,
|
2796 |
"group_size": 4055,
|
2797 |
"in_features": 4096,
|
|
|
2815 |
"model.layers.23.self_attn.o_proj": {
|
2816 |
"bias": false,
|
2817 |
"enable_norm": true,
|
2818 |
+
"enable_perm": false,
|
2819 |
"group_num": 1,
|
2820 |
"group_size": 4055,
|
2821 |
"in_features": 4096,
|
|
|
2839 |
"model.layers.23.self_attn.q_proj": {
|
2840 |
"bias": false,
|
2841 |
"enable_norm": true,
|
2842 |
+
"enable_perm": false,
|
2843 |
"group_num": 1,
|
2844 |
"group_size": 4055,
|
2845 |
"in_features": 4096,
|
|
|
2863 |
"model.layers.23.self_attn.v_proj": {
|
2864 |
"bias": false,
|
2865 |
"enable_norm": true,
|
2866 |
+
"enable_perm": false,
|
2867 |
"group_num": 1,
|
2868 |
"group_size": 4055,
|
2869 |
"in_features": 4096,
|
|
|
2887 |
"model.layers.24.mlp.down_proj": {
|
2888 |
"bias": false,
|
2889 |
"enable_norm": true,
|
2890 |
+
"enable_perm": false,
|
2891 |
"group_num": 1,
|
2892 |
"group_size": 14192,
|
2893 |
"in_features": 14336,
|
|
|
2911 |
"model.layers.24.mlp.gate_proj": {
|
2912 |
"bias": false,
|
2913 |
"enable_norm": true,
|
2914 |
+
"enable_perm": false,
|
2915 |
"group_num": 1,
|
2916 |
"group_size": 4055,
|
2917 |
"in_features": 4096,
|
|
|
2935 |
"model.layers.24.mlp.up_proj": {
|
2936 |
"bias": false,
|
2937 |
"enable_norm": true,
|
2938 |
+
"enable_perm": false,
|
2939 |
"group_num": 1,
|
2940 |
"group_size": 4055,
|
2941 |
"in_features": 4096,
|
|
|
2959 |
"model.layers.24.self_attn.k_proj": {
|
2960 |
"bias": false,
|
2961 |
"enable_norm": true,
|
2962 |
+
"enable_perm": false,
|
2963 |
"group_num": 1,
|
2964 |
"group_size": 4055,
|
2965 |
"in_features": 4096,
|
|
|
2983 |
"model.layers.24.self_attn.o_proj": {
|
2984 |
"bias": false,
|
2985 |
"enable_norm": true,
|
2986 |
+
"enable_perm": false,
|
2987 |
"group_num": 1,
|
2988 |
"group_size": 4055,
|
2989 |
"in_features": 4096,
|
|
|
3007 |
"model.layers.24.self_attn.q_proj": {
|
3008 |
"bias": false,
|
3009 |
"enable_norm": true,
|
3010 |
+
"enable_perm": false,
|
3011 |
"group_num": 1,
|
3012 |
"group_size": 4055,
|
3013 |
"in_features": 4096,
|
|
|
3031 |
"model.layers.24.self_attn.v_proj": {
|
3032 |
"bias": false,
|
3033 |
"enable_norm": true,
|
3034 |
+
"enable_perm": false,
|
3035 |
"group_num": 1,
|
3036 |
"group_size": 4055,
|
3037 |
"in_features": 4096,
|
|
|
3055 |
"model.layers.25.mlp.down_proj": {
|
3056 |
"bias": false,
|
3057 |
"enable_norm": true,
|
3058 |
+
"enable_perm": false,
|
3059 |
"group_num": 1,
|
3060 |
"group_size": 14192,
|
3061 |
"in_features": 14336,
|
|
|
3079 |
"model.layers.25.mlp.gate_proj": {
|
3080 |
"bias": false,
|
3081 |
"enable_norm": true,
|
3082 |
+
"enable_perm": false,
|
3083 |
"group_num": 1,
|
3084 |
"group_size": 4055,
|
3085 |
"in_features": 4096,
|
|
|
3103 |
"model.layers.25.mlp.up_proj": {
|
3104 |
"bias": false,
|
3105 |
"enable_norm": true,
|
3106 |
+
"enable_perm": false,
|
3107 |
"group_num": 1,
|
3108 |
"group_size": 4055,
|
3109 |
"in_features": 4096,
|
|
|
3127 |
"model.layers.25.self_attn.k_proj": {
|
3128 |
"bias": false,
|
3129 |
"enable_norm": true,
|
3130 |
+
"enable_perm": false,
|
3131 |
"group_num": 1,
|
3132 |
"group_size": 4055,
|
3133 |
"in_features": 4096,
|
|
|
3151 |
"model.layers.25.self_attn.o_proj": {
|
3152 |
"bias": false,
|
3153 |
"enable_norm": true,
|
3154 |
+
"enable_perm": false,
|
3155 |
"group_num": 1,
|
3156 |
"group_size": 4055,
|
3157 |
"in_features": 4096,
|
|
|
3175 |
"model.layers.25.self_attn.q_proj": {
|
3176 |
"bias": false,
|
3177 |
"enable_norm": true,
|
3178 |
+
"enable_perm": false,
|
3179 |
"group_num": 1,
|
3180 |
"group_size": 4055,
|
3181 |
"in_features": 4096,
|
|
|
3199 |
"model.layers.25.self_attn.v_proj": {
|
3200 |
"bias": false,
|
3201 |
"enable_norm": true,
|
3202 |
+
"enable_perm": false,
|
3203 |
"group_num": 1,
|
3204 |
"group_size": 4055,
|
3205 |
"in_features": 4096,
|
|
|
3223 |
"model.layers.26.mlp.down_proj": {
|
3224 |
"bias": false,
|
3225 |
"enable_norm": true,
|
3226 |
+
"enable_perm": false,
|
3227 |
"group_num": 1,
|
3228 |
"group_size": 14192,
|
3229 |
"in_features": 14336,
|
|
|
3247 |
"model.layers.26.mlp.gate_proj": {
|
3248 |
"bias": false,
|
3249 |
"enable_norm": true,
|
3250 |
+
"enable_perm": false,
|
3251 |
"group_num": 1,
|
3252 |
"group_size": 4055,
|
3253 |
"in_features": 4096,
|
|
|
3271 |
"model.layers.26.mlp.up_proj": {
|
3272 |
"bias": false,
|
3273 |
"enable_norm": true,
|
3274 |
+
"enable_perm": false,
|
3275 |
"group_num": 1,
|
3276 |
"group_size": 4055,
|
3277 |
"in_features": 4096,
|
|
|
3295 |
"model.layers.26.self_attn.k_proj": {
|
3296 |
"bias": false,
|
3297 |
"enable_norm": true,
|
3298 |
+
"enable_perm": false,
|
3299 |
"group_num": 1,
|
3300 |
"group_size": 4055,
|
3301 |
"in_features": 4096,
|
|
|
3319 |
"model.layers.26.self_attn.o_proj": {
|
3320 |
"bias": false,
|
3321 |
"enable_norm": true,
|
3322 |
+
"enable_perm": false,
|
3323 |
"group_num": 1,
|
3324 |
"group_size": 4055,
|
3325 |
"in_features": 4096,
|
|
|
3343 |
"model.layers.26.self_attn.q_proj": {
|
3344 |
"bias": false,
|
3345 |
"enable_norm": true,
|
3346 |
+
"enable_perm": false,
|
3347 |
"group_num": 1,
|
3348 |
"group_size": 4055,
|
3349 |
"in_features": 4096,
|
|
|
3367 |
"model.layers.26.self_attn.v_proj": {
|
3368 |
"bias": false,
|
3369 |
"enable_norm": true,
|
3370 |
+
"enable_perm": false,
|
3371 |
"group_num": 1,
|
3372 |
"group_size": 4055,
|
3373 |
"in_features": 4096,
|
|
|
3391 |
"model.layers.27.mlp.down_proj": {
|
3392 |
"bias": false,
|
3393 |
"enable_norm": true,
|
3394 |
+
"enable_perm": false,
|
3395 |
"group_num": 1,
|
3396 |
"group_size": 14192,
|
3397 |
"in_features": 14336,
|
|
|
3415 |
"model.layers.27.mlp.gate_proj": {
|
3416 |
"bias": false,
|
3417 |
"enable_norm": true,
|
3418 |
+
"enable_perm": false,
|
3419 |
"group_num": 1,
|
3420 |
"group_size": 4055,
|
3421 |
"in_features": 4096,
|
|
|
3439 |
"model.layers.27.mlp.up_proj": {
|
3440 |
"bias": false,
|
3441 |
"enable_norm": true,
|
3442 |
+
"enable_perm": false,
|
3443 |
"group_num": 1,
|
3444 |
"group_size": 4055,
|
3445 |
"in_features": 4096,
|
|
|
3463 |
"model.layers.27.self_attn.k_proj": {
|
3464 |
"bias": false,
|
3465 |
"enable_norm": true,
|
3466 |
+
"enable_perm": false,
|
3467 |
"group_num": 1,
|
3468 |
"group_size": 4055,
|
3469 |
"in_features": 4096,
|
|
|
3487 |
"model.layers.27.self_attn.o_proj": {
|
3488 |
"bias": false,
|
3489 |
"enable_norm": true,
|
3490 |
+
"enable_perm": false,
|
3491 |
"group_num": 1,
|
3492 |
"group_size": 4055,
|
3493 |
"in_features": 4096,
|
|
|
3511 |
"model.layers.27.self_attn.q_proj": {
|
3512 |
"bias": false,
|
3513 |
"enable_norm": true,
|
3514 |
+
"enable_perm": false,
|
3515 |
"group_num": 1,
|
3516 |
"group_size": 4055,
|
3517 |
"in_features": 4096,
|
|
|
3535 |
"model.layers.27.self_attn.v_proj": {
|
3536 |
"bias": false,
|
3537 |
"enable_norm": true,
|
3538 |
+
"enable_perm": false,
|
3539 |
"group_num": 1,
|
3540 |
"group_size": 4055,
|
3541 |
"in_features": 4096,
|
|
|
3559 |
"model.layers.28.mlp.down_proj": {
|
3560 |
"bias": false,
|
3561 |
"enable_norm": true,
|
3562 |
+
"enable_perm": false,
|
3563 |
"group_num": 1,
|
3564 |
"group_size": 14192,
|
3565 |
"in_features": 14336,
|
|
|
3583 |
"model.layers.28.mlp.gate_proj": {
|
3584 |
"bias": false,
|
3585 |
"enable_norm": true,
|
3586 |
+
"enable_perm": false,
|
3587 |
"group_num": 1,
|
3588 |
"group_size": 4055,
|
3589 |
"in_features": 4096,
|
|
|
3607 |
"model.layers.28.mlp.up_proj": {
|
3608 |
"bias": false,
|
3609 |
"enable_norm": true,
|
3610 |
+
"enable_perm": false,
|
3611 |
"group_num": 1,
|
3612 |
"group_size": 4055,
|
3613 |
"in_features": 4096,
|
|
|
3631 |
"model.layers.28.self_attn.k_proj": {
|
3632 |
"bias": false,
|
3633 |
"enable_norm": true,
|
3634 |
+
"enable_perm": false,
|
3635 |
"group_num": 1,
|
3636 |
"group_size": 4055,
|
3637 |
"in_features": 4096,
|
|
|
3655 |
"model.layers.28.self_attn.o_proj": {
|
3656 |
"bias": false,
|
3657 |
"enable_norm": true,
|
3658 |
+
"enable_perm": false,
|
3659 |
"group_num": 1,
|
3660 |
"group_size": 4055,
|
3661 |
"in_features": 4096,
|
|
|
3679 |
"model.layers.28.self_attn.q_proj": {
|
3680 |
"bias": false,
|
3681 |
"enable_norm": true,
|
3682 |
+
"enable_perm": false,
|
3683 |
"group_num": 1,
|
3684 |
"group_size": 4055,
|
3685 |
"in_features": 4096,
|
|
|
3703 |
"model.layers.28.self_attn.v_proj": {
|
3704 |
"bias": false,
|
3705 |
"enable_norm": true,
|
3706 |
+
"enable_perm": false,
|
3707 |
"group_num": 1,
|
3708 |
"group_size": 4055,
|
3709 |
"in_features": 4096,
|
|
|
3727 |
"model.layers.29.mlp.down_proj": {
|
3728 |
"bias": false,
|
3729 |
"enable_norm": true,
|
3730 |
+
"enable_perm": false,
|
3731 |
"group_num": 1,
|
3732 |
"group_size": 14192,
|
3733 |
"in_features": 14336,
|
|
|
3751 |
"model.layers.29.mlp.gate_proj": {
|
3752 |
"bias": false,
|
3753 |
"enable_norm": true,
|
3754 |
+
"enable_perm": false,
|
3755 |
"group_num": 1,
|
3756 |
"group_size": 4055,
|
3757 |
"in_features": 4096,
|
|
|
3775 |
"model.layers.29.mlp.up_proj": {
|
3776 |
"bias": false,
|
3777 |
"enable_norm": true,
|
3778 |
+
"enable_perm": false,
|
3779 |
"group_num": 1,
|
3780 |
"group_size": 4055,
|
3781 |
"in_features": 4096,
|
|
|
3799 |
"model.layers.29.self_attn.k_proj": {
|
3800 |
"bias": false,
|
3801 |
"enable_norm": true,
|
3802 |
+
"enable_perm": false,
|
3803 |
"group_num": 1,
|
3804 |
"group_size": 4055,
|
3805 |
"in_features": 4096,
|
|
|
3823 |
"model.layers.29.self_attn.o_proj": {
|
3824 |
"bias": false,
|
3825 |
"enable_norm": true,
|
3826 |
+
"enable_perm": false,
|
3827 |
"group_num": 1,
|
3828 |
"group_size": 4055,
|
3829 |
"in_features": 4096,
|
|
|
3847 |
"model.layers.29.self_attn.q_proj": {
|
3848 |
"bias": false,
|
3849 |
"enable_norm": true,
|
3850 |
+
"enable_perm": false,
|
3851 |
"group_num": 1,
|
3852 |
"group_size": 4055,
|
3853 |
"in_features": 4096,
|
|
|
3871 |
"model.layers.29.self_attn.v_proj": {
|
3872 |
"bias": false,
|
3873 |
"enable_norm": true,
|
3874 |
+
"enable_perm": false,
|
3875 |
"group_num": 1,
|
3876 |
"group_size": 4055,
|
3877 |
"in_features": 4096,
|
|
|
3895 |
"model.layers.3.mlp.down_proj": {
|
3896 |
"bias": false,
|
3897 |
"enable_norm": true,
|
3898 |
+
"enable_perm": false,
|
3899 |
"group_num": 1,
|
3900 |
"group_size": 14192,
|
3901 |
"in_features": 14336,
|
|
|
3919 |
"model.layers.3.mlp.gate_proj": {
|
3920 |
"bias": false,
|
3921 |
"enable_norm": true,
|
3922 |
+
"enable_perm": false,
|
3923 |
"group_num": 1,
|
3924 |
"group_size": 4055,
|
3925 |
"in_features": 4096,
|
|
|
3943 |
"model.layers.3.mlp.up_proj": {
|
3944 |
"bias": false,
|
3945 |
"enable_norm": true,
|
3946 |
+
"enable_perm": false,
|
3947 |
"group_num": 1,
|
3948 |
"group_size": 4055,
|
3949 |
"in_features": 4096,
|
|
|
3967 |
"model.layers.3.self_attn.k_proj": {
|
3968 |
"bias": false,
|
3969 |
"enable_norm": true,
|
3970 |
+
"enable_perm": false,
|
3971 |
"group_num": 1,
|
3972 |
"group_size": 4055,
|
3973 |
"in_features": 4096,
|
|
|
3991 |
"model.layers.3.self_attn.o_proj": {
|
3992 |
"bias": false,
|
3993 |
"enable_norm": true,
|
3994 |
+
"enable_perm": false,
|
3995 |
"group_num": 1,
|
3996 |
"group_size": 4055,
|
3997 |
"in_features": 4096,
|
|
|
4015 |
"model.layers.3.self_attn.q_proj": {
|
4016 |
"bias": false,
|
4017 |
"enable_norm": true,
|
4018 |
+
"enable_perm": false,
|
4019 |
"group_num": 1,
|
4020 |
"group_size": 4055,
|
4021 |
"in_features": 4096,
|
|
|
4039 |
"model.layers.3.self_attn.v_proj": {
|
4040 |
"bias": false,
|
4041 |
"enable_norm": true,
|
4042 |
+
"enable_perm": false,
|
4043 |
"group_num": 1,
|
4044 |
"group_size": 4055,
|
4045 |
"in_features": 4096,
|
|
|
4063 |
"model.layers.30.mlp.down_proj": {
|
4064 |
"bias": false,
|
4065 |
"enable_norm": true,
|
4066 |
+
"enable_perm": false,
|
4067 |
"group_num": 1,
|
4068 |
"group_size": 14192,
|
4069 |
"in_features": 14336,
|
|
|
4087 |
"model.layers.30.mlp.gate_proj": {
|
4088 |
"bias": false,
|
4089 |
"enable_norm": true,
|
4090 |
+
"enable_perm": false,
|
4091 |
"group_num": 1,
|
4092 |
"group_size": 4055,
|
4093 |
"in_features": 4096,
|
|
|
4111 |
"model.layers.30.mlp.up_proj": {
|
4112 |
"bias": false,
|
4113 |
"enable_norm": true,
|
4114 |
+
"enable_perm": false,
|
4115 |
"group_num": 1,
|
4116 |
"group_size": 4055,
|
4117 |
"in_features": 4096,
|
|
|
4135 |
"model.layers.30.self_attn.k_proj": {
|
4136 |
"bias": false,
|
4137 |
"enable_norm": true,
|
4138 |
+
"enable_perm": false,
|
4139 |
"group_num": 1,
|
4140 |
"group_size": 4055,
|
4141 |
"in_features": 4096,
|
|
|
4159 |
"model.layers.30.self_attn.o_proj": {
|
4160 |
"bias": false,
|
4161 |
"enable_norm": true,
|
4162 |
+
"enable_perm": false,
|
4163 |
"group_num": 1,
|
4164 |
"group_size": 4055,
|
4165 |
"in_features": 4096,
|
|
|
4183 |
"model.layers.30.self_attn.q_proj": {
|
4184 |
"bias": false,
|
4185 |
"enable_norm": true,
|
4186 |
+
"enable_perm": false,
|
4187 |
"group_num": 1,
|
4188 |
"group_size": 4055,
|
4189 |
"in_features": 4096,
|
|
|
4207 |
"model.layers.30.self_attn.v_proj": {
|
4208 |
"bias": false,
|
4209 |
"enable_norm": true,
|
4210 |
+
"enable_perm": false,
|
4211 |
"group_num": 1,
|
4212 |
"group_size": 4055,
|
4213 |
"in_features": 4096,
|
|
|
4231 |
"model.layers.31.mlp.down_proj": {
|
4232 |
"bias": false,
|
4233 |
"enable_norm": true,
|
4234 |
+
"enable_perm": false,
|
4235 |
"group_num": 1,
|
4236 |
"group_size": 14192,
|
4237 |
"in_features": 14336,
|
|
|
4255 |
"model.layers.31.mlp.gate_proj": {
|
4256 |
"bias": false,
|
4257 |
"enable_norm": true,
|
4258 |
+
"enable_perm": false,
|
4259 |
"group_num": 1,
|
4260 |
"group_size": 4055,
|
4261 |
"in_features": 4096,
|
|
|
4279 |
"model.layers.31.mlp.up_proj": {
|
4280 |
"bias": false,
|
4281 |
"enable_norm": true,
|
4282 |
+
"enable_perm": false,
|
4283 |
"group_num": 1,
|
4284 |
"group_size": 4055,
|
4285 |
"in_features": 4096,
|
|
|
4303 |
"model.layers.31.self_attn.k_proj": {
|
4304 |
"bias": false,
|
4305 |
"enable_norm": true,
|
4306 |
+
"enable_perm": false,
|
4307 |
"group_num": 1,
|
4308 |
"group_size": 4055,
|
4309 |
"in_features": 4096,
|
|
|
4327 |
"model.layers.31.self_attn.o_proj": {
|
4328 |
"bias": false,
|
4329 |
"enable_norm": true,
|
4330 |
+
"enable_perm": false,
|
4331 |
"group_num": 1,
|
4332 |
"group_size": 4055,
|
4333 |
"in_features": 4096,
|
|
|
4351 |
"model.layers.31.self_attn.q_proj": {
|
4352 |
"bias": false,
|
4353 |
"enable_norm": true,
|
4354 |
+
"enable_perm": false,
|
4355 |
"group_num": 1,
|
4356 |
"group_size": 4055,
|
4357 |
"in_features": 4096,
|
|
|
4375 |
"model.layers.31.self_attn.v_proj": {
|
4376 |
"bias": false,
|
4377 |
"enable_norm": true,
|
4378 |
+
"enable_perm": false,
|
4379 |
"group_num": 1,
|
4380 |
"group_size": 4055,
|
4381 |
"in_features": 4096,
|
|
|
4399 |
"model.layers.4.mlp.down_proj": {
|
4400 |
"bias": false,
|
4401 |
"enable_norm": true,
|
4402 |
+
"enable_perm": false,
|
4403 |
"group_num": 1,
|
4404 |
"group_size": 14192,
|
4405 |
"in_features": 14336,
|
|
|
4423 |
"model.layers.4.mlp.gate_proj": {
|
4424 |
"bias": false,
|
4425 |
"enable_norm": true,
|
4426 |
+
"enable_perm": false,
|
4427 |
"group_num": 1,
|
4428 |
"group_size": 4055,
|
4429 |
"in_features": 4096,
|
|
|
4447 |
"model.layers.4.mlp.up_proj": {
|
4448 |
"bias": false,
|
4449 |
"enable_norm": true,
|
4450 |
+
"enable_perm": false,
|
4451 |
"group_num": 1,
|
4452 |
"group_size": 4055,
|
4453 |
"in_features": 4096,
|
|
|
4471 |
"model.layers.4.self_attn.k_proj": {
|
4472 |
"bias": false,
|
4473 |
"enable_norm": true,
|
4474 |
+
"enable_perm": false,
|
4475 |
"group_num": 1,
|
4476 |
"group_size": 4055,
|
4477 |
"in_features": 4096,
|
|
|
4495 |
"model.layers.4.self_attn.o_proj": {
|
4496 |
"bias": false,
|
4497 |
"enable_norm": true,
|
4498 |
+
"enable_perm": false,
|
4499 |
"group_num": 1,
|
4500 |
"group_size": 4055,
|
4501 |
"in_features": 4096,
|
|
|
4519 |
"model.layers.4.self_attn.q_proj": {
|
4520 |
"bias": false,
|
4521 |
"enable_norm": true,
|
4522 |
+
"enable_perm": false,
|
4523 |
"group_num": 1,
|
4524 |
"group_size": 4055,
|
4525 |
"in_features": 4096,
|
|
|
4543 |
"model.layers.4.self_attn.v_proj": {
|
4544 |
"bias": false,
|
4545 |
"enable_norm": true,
|
4546 |
+
"enable_perm": false,
|
4547 |
"group_num": 1,
|
4548 |
"group_size": 4055,
|
4549 |
"in_features": 4096,
|
|
|
4567 |
"model.layers.5.mlp.down_proj": {
|
4568 |
"bias": false,
|
4569 |
"enable_norm": true,
|
4570 |
+
"enable_perm": false,
|
4571 |
"group_num": 1,
|
4572 |
"group_size": 14192,
|
4573 |
"in_features": 14336,
|
|
|
4591 |
"model.layers.5.mlp.gate_proj": {
|
4592 |
"bias": false,
|
4593 |
"enable_norm": true,
|
4594 |
+
"enable_perm": false,
|
4595 |
"group_num": 1,
|
4596 |
"group_size": 4055,
|
4597 |
"in_features": 4096,
|
|
|
4615 |
"model.layers.5.mlp.up_proj": {
|
4616 |
"bias": false,
|
4617 |
"enable_norm": true,
|
4618 |
+
"enable_perm": false,
|
4619 |
"group_num": 1,
|
4620 |
"group_size": 4055,
|
4621 |
"in_features": 4096,
|
|
|
4639 |
"model.layers.5.self_attn.k_proj": {
|
4640 |
"bias": false,
|
4641 |
"enable_norm": true,
|
4642 |
+
"enable_perm": false,
|
4643 |
"group_num": 1,
|
4644 |
"group_size": 4055,
|
4645 |
"in_features": 4096,
|
|
|
4663 |
"model.layers.5.self_attn.o_proj": {
|
4664 |
"bias": false,
|
4665 |
"enable_norm": true,
|
4666 |
+
"enable_perm": false,
|
4667 |
"group_num": 1,
|
4668 |
"group_size": 4055,
|
4669 |
"in_features": 4096,
|
|
|
4687 |
"model.layers.5.self_attn.q_proj": {
|
4688 |
"bias": false,
|
4689 |
"enable_norm": true,
|
4690 |
+
"enable_perm": false,
|
4691 |
"group_num": 1,
|
4692 |
"group_size": 4055,
|
4693 |
"in_features": 4096,
|
|
|
4711 |
"model.layers.5.self_attn.v_proj": {
|
4712 |
"bias": false,
|
4713 |
"enable_norm": true,
|
4714 |
+
"enable_perm": false,
|
4715 |
"group_num": 1,
|
4716 |
"group_size": 4055,
|
4717 |
"in_features": 4096,
|
|
|
4735 |
"model.layers.6.mlp.down_proj": {
|
4736 |
"bias": false,
|
4737 |
"enable_norm": true,
|
4738 |
+
"enable_perm": false,
|
4739 |
"group_num": 1,
|
4740 |
"group_size": 14192,
|
4741 |
"in_features": 14336,
|
|
|
4759 |
"model.layers.6.mlp.gate_proj": {
|
4760 |
"bias": false,
|
4761 |
"enable_norm": true,
|
4762 |
+
"enable_perm": false,
|
4763 |
"group_num": 1,
|
4764 |
"group_size": 4055,
|
4765 |
"in_features": 4096,
|
|
|
4783 |
"model.layers.6.mlp.up_proj": {
|
4784 |
"bias": false,
|
4785 |
"enable_norm": true,
|
4786 |
+
"enable_perm": false,
|
4787 |
"group_num": 1,
|
4788 |
"group_size": 4055,
|
4789 |
"in_features": 4096,
|
|
|
4807 |
"model.layers.6.self_attn.k_proj": {
|
4808 |
"bias": false,
|
4809 |
"enable_norm": true,
|
4810 |
+
"enable_perm": false,
|
4811 |
"group_num": 1,
|
4812 |
"group_size": 4055,
|
4813 |
"in_features": 4096,
|
|
|
4831 |
"model.layers.6.self_attn.o_proj": {
|
4832 |
"bias": false,
|
4833 |
"enable_norm": true,
|
4834 |
+
"enable_perm": false,
|
4835 |
"group_num": 1,
|
4836 |
"group_size": 4055,
|
4837 |
"in_features": 4096,
|
|
|
4855 |
"model.layers.6.self_attn.q_proj": {
|
4856 |
"bias": false,
|
4857 |
"enable_norm": true,
|
4858 |
+
"enable_perm": false,
|
4859 |
"group_num": 1,
|
4860 |
"group_size": 4055,
|
4861 |
"in_features": 4096,
|
|
|
4879 |
"model.layers.6.self_attn.v_proj": {
|
4880 |
"bias": false,
|
4881 |
"enable_norm": true,
|
4882 |
+
"enable_perm": false,
|
4883 |
"group_num": 1,
|
4884 |
"group_size": 4055,
|
4885 |
"in_features": 4096,
|
|
|
4903 |
"model.layers.7.mlp.down_proj": {
|
4904 |
"bias": false,
|
4905 |
"enable_norm": true,
|
4906 |
+
"enable_perm": false,
|
4907 |
"group_num": 1,
|
4908 |
"group_size": 14192,
|
4909 |
"in_features": 14336,
|
|
|
4927 |
"model.layers.7.mlp.gate_proj": {
|
4928 |
"bias": false,
|
4929 |
"enable_norm": true,
|
4930 |
+
"enable_perm": false,
|
4931 |
"group_num": 1,
|
4932 |
"group_size": 4055,
|
4933 |
"in_features": 4096,
|
|
|
4951 |
"model.layers.7.mlp.up_proj": {
|
4952 |
"bias": false,
|
4953 |
"enable_norm": true,
|
4954 |
+
"enable_perm": false,
|
4955 |
"group_num": 1,
|
4956 |
"group_size": 4055,
|
4957 |
"in_features": 4096,
|
|
|
4975 |
"model.layers.7.self_attn.k_proj": {
|
4976 |
"bias": false,
|
4977 |
"enable_norm": true,
|
4978 |
+
"enable_perm": false,
|
4979 |
"group_num": 1,
|
4980 |
"group_size": 4055,
|
4981 |
"in_features": 4096,
|
|
|
4999 |
"model.layers.7.self_attn.o_proj": {
|
5000 |
"bias": false,
|
5001 |
"enable_norm": true,
|
5002 |
+
"enable_perm": false,
|
5003 |
"group_num": 1,
|
5004 |
"group_size": 4055,
|
5005 |
"in_features": 4096,
|
|
|
5023 |
"model.layers.7.self_attn.q_proj": {
|
5024 |
"bias": false,
|
5025 |
"enable_norm": true,
|
5026 |
+
"enable_perm": false,
|
5027 |
"group_num": 1,
|
5028 |
"group_size": 4055,
|
5029 |
"in_features": 4096,
|
|
|
5047 |
"model.layers.7.self_attn.v_proj": {
|
5048 |
"bias": false,
|
5049 |
"enable_norm": true,
|
5050 |
+
"enable_perm": false,
|
5051 |
"group_num": 1,
|
5052 |
"group_size": 4055,
|
5053 |
"in_features": 4096,
|
|
|
5071 |
"model.layers.8.mlp.down_proj": {
|
5072 |
"bias": false,
|
5073 |
"enable_norm": true,
|
5074 |
+
"enable_perm": false,
|
5075 |
"group_num": 1,
|
5076 |
"group_size": 14192,
|
5077 |
"in_features": 14336,
|
|
|
5095 |
"model.layers.8.mlp.gate_proj": {
|
5096 |
"bias": false,
|
5097 |
"enable_norm": true,
|
5098 |
+
"enable_perm": false,
|
5099 |
"group_num": 1,
|
5100 |
"group_size": 4055,
|
5101 |
"in_features": 4096,
|
|
|
5119 |
"model.layers.8.mlp.up_proj": {
|
5120 |
"bias": false,
|
5121 |
"enable_norm": true,
|
5122 |
+
"enable_perm": false,
|
5123 |
"group_num": 1,
|
5124 |
"group_size": 4055,
|
5125 |
"in_features": 4096,
|
|
|
5143 |
"model.layers.8.self_attn.k_proj": {
|
5144 |
"bias": false,
|
5145 |
"enable_norm": true,
|
5146 |
+
"enable_perm": false,
|
5147 |
"group_num": 1,
|
5148 |
"group_size": 4055,
|
5149 |
"in_features": 4096,
|
|
|
5167 |
"model.layers.8.self_attn.o_proj": {
|
5168 |
"bias": false,
|
5169 |
"enable_norm": true,
|
5170 |
+
"enable_perm": false,
|
5171 |
"group_num": 1,
|
5172 |
"group_size": 4055,
|
5173 |
"in_features": 4096,
|
|
|
5191 |
"model.layers.8.self_attn.q_proj": {
|
5192 |
"bias": false,
|
5193 |
"enable_norm": true,
|
5194 |
+
"enable_perm": false,
|
5195 |
"group_num": 1,
|
5196 |
"group_size": 4055,
|
5197 |
"in_features": 4096,
|
|
|
5215 |
"model.layers.8.self_attn.v_proj": {
|
5216 |
"bias": false,
|
5217 |
"enable_norm": true,
|
5218 |
+
"enable_perm": false,
|
5219 |
"group_num": 1,
|
5220 |
"group_size": 4055,
|
5221 |
"in_features": 4096,
|
|
|
5239 |
"model.layers.9.mlp.down_proj": {
|
5240 |
"bias": false,
|
5241 |
"enable_norm": true,
|
5242 |
+
"enable_perm": false,
|
5243 |
"group_num": 1,
|
5244 |
"group_size": 14192,
|
5245 |
"in_features": 14336,
|
|
|
5263 |
"model.layers.9.mlp.gate_proj": {
|
5264 |
"bias": false,
|
5265 |
"enable_norm": true,
|
5266 |
+
"enable_perm": false,
|
5267 |
"group_num": 1,
|
5268 |
"group_size": 4055,
|
5269 |
"in_features": 4096,
|
|
|
5287 |
"model.layers.9.mlp.up_proj": {
|
5288 |
"bias": false,
|
5289 |
"enable_norm": true,
|
5290 |
+
"enable_perm": false,
|
5291 |
"group_num": 1,
|
5292 |
"group_size": 4055,
|
5293 |
"in_features": 4096,
|
|
|
5311 |
"model.layers.9.self_attn.k_proj": {
|
5312 |
"bias": false,
|
5313 |
"enable_norm": true,
|
5314 |
+
"enable_perm": false,
|
5315 |
"group_num": 1,
|
5316 |
"group_size": 4055,
|
5317 |
"in_features": 4096,
|
|
|
5335 |
"model.layers.9.self_attn.o_proj": {
|
5336 |
"bias": false,
|
5337 |
"enable_norm": true,
|
5338 |
+
"enable_perm": false,
|
5339 |
"group_num": 1,
|
5340 |
"group_size": 4055,
|
5341 |
"in_features": 4096,
|
|
|
5359 |
"model.layers.9.self_attn.q_proj": {
|
5360 |
"bias": false,
|
5361 |
"enable_norm": true,
|
5362 |
+
"enable_perm": false,
|
5363 |
"group_num": 1,
|
5364 |
"group_size": 4055,
|
5365 |
"in_features": 4096,
|
|
|
5383 |
"model.layers.9.self_attn.v_proj": {
|
5384 |
"bias": false,
|
5385 |
"enable_norm": true,
|
5386 |
+
"enable_perm": false,
|
5387 |
"group_num": 1,
|
5388 |
"group_size": 4055,
|
5389 |
"in_features": 4096,
|