diff --git "a/config.json" "b/config.json" --- "a/config.json" +++ "b/config.json" @@ -1,5 +1,5 @@ { - "_name_or_path": "Qwen/Qwen2.5-72B-Instruct", + "_name_or_path": "/home/aiscuser/yangwang/convert_model/models/VPTQ-community--Qwen2.5-72B-Instruct-v8-k65536-65536-woft", "architectures": [ "Qwen2ForCausalLM" ], @@ -16,23 +16,12 @@ "num_attention_heads": 64, "num_hidden_layers": 80, "num_key_value_heads": 8, - "rms_norm_eps": 1e-06, - "rope_scaling": null, - "rope_theta": 1000000.0, - "sliding_window": null, - "tie_word_embeddings": false, - "torch_dtype": "bfloat16", - "transformers_version": "4.45.1", - "use_cache": true, - "use_sliding_window": false, - "vocab_size": 152064, "quantization_config": { - "quant_method": "vptq", "config_for_layers": { "model.layers.0.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -56,7 +45,7 @@ "model.layers.0.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -80,7 +69,7 @@ "model.layers.0.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -104,7 +93,7 @@ "model.layers.0.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -128,7 +117,7 @@ "model.layers.0.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -152,7 +141,7 @@ "model.layers.0.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -176,7 +165,7 @@ "model.layers.0.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -200,7 +189,7 @@ "model.layers.1.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -224,7 +213,7 @@ "model.layers.1.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -248,7 +237,7 @@ "model.layers.1.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -272,7 +261,7 @@ "model.layers.1.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -296,7 +285,7 @@ "model.layers.1.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -320,7 +309,7 @@ "model.layers.1.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -344,7 +333,7 @@ "model.layers.1.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -368,7 +357,7 @@ "model.layers.10.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -392,7 +381,7 @@ "model.layers.10.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -416,7 +405,7 @@ "model.layers.10.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -440,7 +429,7 @@ "model.layers.10.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -464,7 +453,7 @@ "model.layers.10.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -488,7 +477,7 @@ "model.layers.10.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -512,7 +501,7 @@ "model.layers.10.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -536,7 +525,7 @@ "model.layers.11.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -560,7 +549,7 @@ "model.layers.11.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -584,7 +573,7 @@ "model.layers.11.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -608,7 +597,7 @@ "model.layers.11.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -632,7 +621,7 @@ "model.layers.11.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -656,7 +645,7 @@ "model.layers.11.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -680,7 +669,7 @@ "model.layers.11.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -704,7 +693,7 @@ "model.layers.12.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -728,7 +717,7 @@ "model.layers.12.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -752,7 +741,7 @@ "model.layers.12.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -776,7 +765,7 @@ "model.layers.12.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -800,7 +789,7 @@ "model.layers.12.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -824,7 +813,7 @@ "model.layers.12.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -848,7 +837,7 @@ "model.layers.12.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -872,7 +861,7 @@ "model.layers.13.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -896,7 +885,7 @@ "model.layers.13.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -920,7 +909,7 @@ "model.layers.13.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -944,7 +933,7 @@ "model.layers.13.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -968,7 +957,7 @@ "model.layers.13.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -992,7 +981,7 @@ "model.layers.13.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1016,7 +1005,7 @@ "model.layers.13.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1040,7 +1029,7 @@ "model.layers.14.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -1064,7 +1053,7 @@ "model.layers.14.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1088,7 +1077,7 @@ "model.layers.14.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1112,7 +1101,7 @@ "model.layers.14.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1136,7 +1125,7 @@ "model.layers.14.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1160,7 +1149,7 @@ "model.layers.14.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1184,7 +1173,7 @@ "model.layers.14.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1208,7 +1197,7 @@ "model.layers.15.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -1232,7 +1221,7 @@ "model.layers.15.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1256,7 +1245,7 @@ "model.layers.15.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1280,7 +1269,7 @@ "model.layers.15.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1304,7 +1293,7 @@ "model.layers.15.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1328,7 +1317,7 @@ "model.layers.15.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1352,7 +1341,7 @@ "model.layers.15.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1376,7 +1365,7 @@ "model.layers.16.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -1400,7 +1389,7 @@ "model.layers.16.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1424,7 +1413,7 @@ "model.layers.16.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1448,7 +1437,7 @@ "model.layers.16.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1472,7 +1461,7 @@ "model.layers.16.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1496,7 +1485,7 @@ "model.layers.16.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1520,7 +1509,7 @@ "model.layers.16.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1544,7 +1533,7 @@ "model.layers.17.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -1568,7 +1557,7 @@ "model.layers.17.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1592,7 +1581,7 @@ "model.layers.17.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1616,7 +1605,7 @@ "model.layers.17.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1640,7 +1629,7 @@ "model.layers.17.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1664,7 +1653,7 @@ "model.layers.17.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1688,7 +1677,7 @@ "model.layers.17.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1712,7 +1701,7 @@ "model.layers.18.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -1736,7 +1725,7 @@ "model.layers.18.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1760,7 +1749,7 @@ "model.layers.18.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1784,7 +1773,7 @@ "model.layers.18.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1808,7 +1797,7 @@ "model.layers.18.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1832,7 +1821,7 @@ "model.layers.18.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1856,7 +1845,7 @@ "model.layers.18.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1880,7 +1869,7 @@ "model.layers.19.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -1904,7 +1893,7 @@ "model.layers.19.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1928,7 +1917,7 @@ "model.layers.19.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1952,7 +1941,7 @@ "model.layers.19.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1976,7 +1965,7 @@ "model.layers.19.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2000,7 +1989,7 @@ "model.layers.19.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2024,7 +2013,7 @@ "model.layers.19.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2048,7 +2037,7 @@ "model.layers.2.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -2072,7 +2061,7 @@ "model.layers.2.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2096,7 +2085,7 @@ "model.layers.2.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2120,7 +2109,7 @@ "model.layers.2.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2144,7 +2133,7 @@ "model.layers.2.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2168,7 +2157,7 @@ "model.layers.2.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2192,7 +2181,7 @@ "model.layers.2.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2216,7 +2205,7 @@ "model.layers.20.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -2240,7 +2229,7 @@ "model.layers.20.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2264,7 +2253,7 @@ "model.layers.20.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2288,7 +2277,7 @@ "model.layers.20.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2312,7 +2301,7 @@ "model.layers.20.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2336,7 +2325,7 @@ "model.layers.20.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2360,7 +2349,7 @@ "model.layers.20.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2384,7 +2373,7 @@ "model.layers.21.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -2408,7 +2397,7 @@ "model.layers.21.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2432,7 +2421,7 @@ "model.layers.21.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2456,7 +2445,7 @@ "model.layers.21.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2480,7 +2469,7 @@ "model.layers.21.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2504,7 +2493,7 @@ "model.layers.21.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2528,7 +2517,7 @@ "model.layers.21.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2552,7 +2541,7 @@ "model.layers.22.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -2576,7 +2565,7 @@ "model.layers.22.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2600,7 +2589,7 @@ "model.layers.22.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2624,7 +2613,7 @@ "model.layers.22.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2648,7 +2637,7 @@ "model.layers.22.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2672,7 +2661,7 @@ "model.layers.22.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2696,7 +2685,7 @@ "model.layers.22.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2720,7 +2709,7 @@ "model.layers.23.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -2744,7 +2733,7 @@ "model.layers.23.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2768,7 +2757,7 @@ "model.layers.23.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2792,7 +2781,7 @@ "model.layers.23.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2816,7 +2805,7 @@ "model.layers.23.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2840,7 +2829,7 @@ "model.layers.23.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2864,7 +2853,7 @@ "model.layers.23.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2888,7 +2877,7 @@ "model.layers.24.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -2912,7 +2901,7 @@ "model.layers.24.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2936,7 +2925,7 @@ "model.layers.24.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2960,7 +2949,7 @@ "model.layers.24.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2984,7 +2973,7 @@ "model.layers.24.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3008,7 +2997,7 @@ "model.layers.24.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3032,7 +3021,7 @@ "model.layers.24.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3056,7 +3045,7 @@ "model.layers.25.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -3080,7 +3069,7 @@ "model.layers.25.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3104,7 +3093,7 @@ "model.layers.25.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3128,7 +3117,7 @@ "model.layers.25.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3152,7 +3141,7 @@ "model.layers.25.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3176,7 +3165,7 @@ "model.layers.25.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3200,7 +3189,7 @@ "model.layers.25.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3224,7 +3213,7 @@ "model.layers.26.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -3248,7 +3237,7 @@ "model.layers.26.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3272,7 +3261,7 @@ "model.layers.26.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3296,7 +3285,7 @@ "model.layers.26.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3320,7 +3309,7 @@ "model.layers.26.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3344,7 +3333,7 @@ "model.layers.26.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3368,7 +3357,7 @@ "model.layers.26.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3392,7 +3381,7 @@ "model.layers.27.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -3416,7 +3405,7 @@ "model.layers.27.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3440,7 +3429,7 @@ "model.layers.27.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3464,7 +3453,7 @@ "model.layers.27.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3488,7 +3477,7 @@ "model.layers.27.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3512,7 +3501,7 @@ "model.layers.27.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3536,7 +3525,7 @@ "model.layers.27.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3560,7 +3549,7 @@ "model.layers.28.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -3584,7 +3573,7 @@ "model.layers.28.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3608,7 +3597,7 @@ "model.layers.28.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3632,7 +3621,7 @@ "model.layers.28.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3656,7 +3645,7 @@ "model.layers.28.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3680,7 +3669,7 @@ "model.layers.28.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3704,7 +3693,7 @@ "model.layers.28.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3728,7 +3717,7 @@ "model.layers.29.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -3752,7 +3741,7 @@ "model.layers.29.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3776,7 +3765,7 @@ "model.layers.29.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3800,7 +3789,7 @@ "model.layers.29.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3824,7 +3813,7 @@ "model.layers.29.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3848,7 +3837,7 @@ "model.layers.29.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3872,7 +3861,7 @@ "model.layers.29.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3896,7 +3885,7 @@ "model.layers.3.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -3920,7 +3909,7 @@ "model.layers.3.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3944,7 +3933,7 @@ "model.layers.3.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3968,7 +3957,7 @@ "model.layers.3.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3992,7 +3981,7 @@ "model.layers.3.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4016,7 +4005,7 @@ "model.layers.3.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4040,7 +4029,7 @@ "model.layers.3.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4064,7 +4053,7 @@ "model.layers.30.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -4088,7 +4077,7 @@ "model.layers.30.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4112,7 +4101,7 @@ "model.layers.30.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4136,7 +4125,7 @@ "model.layers.30.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4160,7 +4149,7 @@ "model.layers.30.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4184,7 +4173,7 @@ "model.layers.30.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4208,7 +4197,7 @@ "model.layers.30.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4232,7 +4221,7 @@ "model.layers.31.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -4256,7 +4245,7 @@ "model.layers.31.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4280,7 +4269,7 @@ "model.layers.31.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4304,7 +4293,7 @@ "model.layers.31.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4328,7 +4317,7 @@ "model.layers.31.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4352,7 +4341,7 @@ "model.layers.31.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4376,7 +4365,7 @@ "model.layers.31.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4400,7 +4389,7 @@ "model.layers.32.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -4424,7 +4413,7 @@ "model.layers.32.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4448,7 +4437,7 @@ "model.layers.32.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4472,7 +4461,7 @@ "model.layers.32.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4496,7 +4485,7 @@ "model.layers.32.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4520,7 +4509,7 @@ "model.layers.32.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4544,7 +4533,7 @@ "model.layers.32.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4568,7 +4557,7 @@ "model.layers.33.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -4592,7 +4581,7 @@ "model.layers.33.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4616,7 +4605,7 @@ "model.layers.33.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4640,7 +4629,7 @@ "model.layers.33.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4664,7 +4653,7 @@ "model.layers.33.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4688,7 +4677,7 @@ "model.layers.33.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4712,7 +4701,7 @@ "model.layers.33.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4736,7 +4725,7 @@ "model.layers.34.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -4760,7 +4749,7 @@ "model.layers.34.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4784,7 +4773,7 @@ "model.layers.34.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4808,7 +4797,7 @@ "model.layers.34.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4832,7 +4821,7 @@ "model.layers.34.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4856,7 +4845,7 @@ "model.layers.34.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4880,7 +4869,7 @@ "model.layers.34.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4904,7 +4893,7 @@ "model.layers.35.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -4928,7 +4917,7 @@ "model.layers.35.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4952,7 +4941,7 @@ "model.layers.35.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4976,7 +4965,7 @@ "model.layers.35.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5000,7 +4989,7 @@ "model.layers.35.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5024,7 +5013,7 @@ "model.layers.35.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5048,7 +5037,7 @@ "model.layers.35.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5072,7 +5061,7 @@ "model.layers.36.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -5096,7 +5085,7 @@ "model.layers.36.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5120,7 +5109,7 @@ "model.layers.36.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5144,7 +5133,7 @@ "model.layers.36.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5168,7 +5157,7 @@ "model.layers.36.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5192,7 +5181,7 @@ "model.layers.36.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5216,7 +5205,7 @@ "model.layers.36.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5240,7 +5229,7 @@ "model.layers.37.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -5264,7 +5253,7 @@ "model.layers.37.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5288,7 +5277,7 @@ "model.layers.37.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5312,7 +5301,7 @@ "model.layers.37.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5336,7 +5325,7 @@ "model.layers.37.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5360,7 +5349,7 @@ "model.layers.37.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5384,7 +5373,7 @@ "model.layers.37.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5408,7 +5397,7 @@ "model.layers.38.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -5432,7 +5421,7 @@ "model.layers.38.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5456,7 +5445,7 @@ "model.layers.38.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5480,7 +5469,7 @@ "model.layers.38.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5504,7 +5493,7 @@ "model.layers.38.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5528,7 +5517,7 @@ "model.layers.38.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5552,7 +5541,7 @@ "model.layers.38.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5576,7 +5565,7 @@ "model.layers.39.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -5600,7 +5589,7 @@ "model.layers.39.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5624,7 +5613,7 @@ "model.layers.39.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5648,7 +5637,7 @@ "model.layers.39.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5672,7 +5661,7 @@ "model.layers.39.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5696,7 +5685,7 @@ "model.layers.39.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5720,7 +5709,7 @@ "model.layers.39.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5744,7 +5733,7 @@ "model.layers.4.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -5768,7 +5757,7 @@ "model.layers.4.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5792,7 +5781,7 @@ "model.layers.4.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5816,7 +5805,7 @@ "model.layers.4.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5840,7 +5829,7 @@ "model.layers.4.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5864,7 +5853,7 @@ "model.layers.4.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5888,7 +5877,7 @@ "model.layers.4.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5912,7 +5901,7 @@ "model.layers.40.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -5936,7 +5925,7 @@ "model.layers.40.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5960,7 +5949,7 @@ "model.layers.40.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5984,7 +5973,7 @@ "model.layers.40.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6008,7 +5997,7 @@ "model.layers.40.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6032,7 +6021,7 @@ "model.layers.40.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6056,7 +6045,7 @@ "model.layers.40.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6080,7 +6069,7 @@ "model.layers.41.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -6104,7 +6093,7 @@ "model.layers.41.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6128,7 +6117,7 @@ "model.layers.41.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6152,7 +6141,7 @@ "model.layers.41.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6176,7 +6165,7 @@ "model.layers.41.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6200,7 +6189,7 @@ "model.layers.41.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6224,7 +6213,7 @@ "model.layers.41.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6248,7 +6237,7 @@ "model.layers.42.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -6272,7 +6261,7 @@ "model.layers.42.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6296,7 +6285,7 @@ "model.layers.42.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6320,7 +6309,7 @@ "model.layers.42.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6344,7 +6333,7 @@ "model.layers.42.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6368,7 +6357,7 @@ "model.layers.42.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6392,7 +6381,7 @@ "model.layers.42.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6416,7 +6405,7 @@ "model.layers.43.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -6440,7 +6429,7 @@ "model.layers.43.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6464,7 +6453,7 @@ "model.layers.43.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6488,7 +6477,7 @@ "model.layers.43.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6512,7 +6501,7 @@ "model.layers.43.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6536,7 +6525,7 @@ "model.layers.43.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6560,7 +6549,7 @@ "model.layers.43.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6584,7 +6573,7 @@ "model.layers.44.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -6608,7 +6597,7 @@ "model.layers.44.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6632,7 +6621,7 @@ "model.layers.44.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6656,7 +6645,7 @@ "model.layers.44.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6680,7 +6669,7 @@ "model.layers.44.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6704,7 +6693,7 @@ "model.layers.44.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6728,7 +6717,7 @@ "model.layers.44.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6752,7 +6741,7 @@ "model.layers.45.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -6776,7 +6765,7 @@ "model.layers.45.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6800,7 +6789,7 @@ "model.layers.45.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6824,7 +6813,7 @@ "model.layers.45.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6848,7 +6837,7 @@ "model.layers.45.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6872,7 +6861,7 @@ "model.layers.45.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6896,7 +6885,7 @@ "model.layers.45.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6920,7 +6909,7 @@ "model.layers.46.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -6944,7 +6933,7 @@ "model.layers.46.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6968,7 +6957,7 @@ "model.layers.46.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6992,7 +6981,7 @@ "model.layers.46.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7016,7 +7005,7 @@ "model.layers.46.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7040,7 +7029,7 @@ "model.layers.46.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7064,7 +7053,7 @@ "model.layers.46.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7088,7 +7077,7 @@ "model.layers.47.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -7112,7 +7101,7 @@ "model.layers.47.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7136,7 +7125,7 @@ "model.layers.47.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7160,7 +7149,7 @@ "model.layers.47.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7184,7 +7173,7 @@ "model.layers.47.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7208,7 +7197,7 @@ "model.layers.47.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7232,7 +7221,7 @@ "model.layers.47.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7256,7 +7245,7 @@ "model.layers.48.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -7280,7 +7269,7 @@ "model.layers.48.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7304,7 +7293,7 @@ "model.layers.48.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7328,7 +7317,7 @@ "model.layers.48.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7352,7 +7341,7 @@ "model.layers.48.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7376,7 +7365,7 @@ "model.layers.48.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7400,7 +7389,7 @@ "model.layers.48.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7424,7 +7413,7 @@ "model.layers.49.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -7448,7 +7437,7 @@ "model.layers.49.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7472,7 +7461,7 @@ "model.layers.49.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7496,7 +7485,7 @@ "model.layers.49.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7520,7 +7509,7 @@ "model.layers.49.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7544,7 +7533,7 @@ "model.layers.49.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7568,7 +7557,7 @@ "model.layers.49.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7592,7 +7581,7 @@ "model.layers.5.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -7616,7 +7605,7 @@ "model.layers.5.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7640,7 +7629,7 @@ "model.layers.5.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7664,7 +7653,7 @@ "model.layers.5.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7688,7 +7677,7 @@ "model.layers.5.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7712,7 +7701,7 @@ "model.layers.5.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7736,7 +7725,7 @@ "model.layers.5.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7760,7 +7749,7 @@ "model.layers.50.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -7784,7 +7773,7 @@ "model.layers.50.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7808,7 +7797,7 @@ "model.layers.50.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7832,7 +7821,7 @@ "model.layers.50.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7856,7 +7845,7 @@ "model.layers.50.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7880,7 +7869,7 @@ "model.layers.50.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7904,7 +7893,7 @@ "model.layers.50.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7928,7 +7917,7 @@ "model.layers.51.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -7952,7 +7941,7 @@ "model.layers.51.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7976,7 +7965,7 @@ "model.layers.51.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8000,7 +7989,7 @@ "model.layers.51.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8024,7 +8013,7 @@ "model.layers.51.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8048,7 +8037,7 @@ "model.layers.51.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8072,7 +8061,7 @@ "model.layers.51.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8096,7 +8085,7 @@ "model.layers.52.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -8120,7 +8109,7 @@ "model.layers.52.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8144,7 +8133,7 @@ "model.layers.52.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8168,7 +8157,7 @@ "model.layers.52.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8192,7 +8181,7 @@ "model.layers.52.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8216,7 +8205,7 @@ "model.layers.52.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8240,7 +8229,7 @@ "model.layers.52.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8264,7 +8253,7 @@ "model.layers.53.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -8288,7 +8277,7 @@ "model.layers.53.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8312,7 +8301,7 @@ "model.layers.53.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8336,7 +8325,7 @@ "model.layers.53.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8360,7 +8349,7 @@ "model.layers.53.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8384,7 +8373,7 @@ "model.layers.53.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8408,7 +8397,7 @@ "model.layers.53.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8432,7 +8421,7 @@ "model.layers.54.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -8456,7 +8445,7 @@ "model.layers.54.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8480,7 +8469,7 @@ "model.layers.54.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8504,7 +8493,7 @@ "model.layers.54.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8528,7 +8517,7 @@ "model.layers.54.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8552,7 +8541,7 @@ "model.layers.54.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8576,7 +8565,7 @@ "model.layers.54.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8600,7 +8589,7 @@ "model.layers.55.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -8624,7 +8613,7 @@ "model.layers.55.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8648,7 +8637,7 @@ "model.layers.55.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8672,7 +8661,7 @@ "model.layers.55.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8696,7 +8685,7 @@ "model.layers.55.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8720,7 +8709,7 @@ "model.layers.55.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8744,7 +8733,7 @@ "model.layers.55.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8768,7 +8757,7 @@ "model.layers.56.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -8792,7 +8781,7 @@ "model.layers.56.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8816,7 +8805,7 @@ "model.layers.56.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8840,7 +8829,7 @@ "model.layers.56.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8864,7 +8853,7 @@ "model.layers.56.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8888,7 +8877,7 @@ "model.layers.56.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8912,7 +8901,7 @@ "model.layers.56.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8936,7 +8925,7 @@ "model.layers.57.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -8960,7 +8949,7 @@ "model.layers.57.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8984,7 +8973,7 @@ "model.layers.57.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9008,7 +8997,7 @@ "model.layers.57.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9032,7 +9021,7 @@ "model.layers.57.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9056,7 +9045,7 @@ "model.layers.57.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9080,7 +9069,7 @@ "model.layers.57.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9104,7 +9093,7 @@ "model.layers.58.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -9128,7 +9117,7 @@ "model.layers.58.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9152,7 +9141,7 @@ "model.layers.58.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9176,7 +9165,7 @@ "model.layers.58.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9200,7 +9189,7 @@ "model.layers.58.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9224,7 +9213,7 @@ "model.layers.58.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9248,7 +9237,7 @@ "model.layers.58.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9272,7 +9261,7 @@ "model.layers.59.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -9296,7 +9285,7 @@ "model.layers.59.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9320,7 +9309,7 @@ "model.layers.59.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9344,7 +9333,7 @@ "model.layers.59.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9368,7 +9357,7 @@ "model.layers.59.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9392,7 +9381,7 @@ "model.layers.59.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9416,7 +9405,7 @@ "model.layers.59.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9440,7 +9429,7 @@ "model.layers.6.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -9464,7 +9453,7 @@ "model.layers.6.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9488,7 +9477,7 @@ "model.layers.6.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9512,7 +9501,7 @@ "model.layers.6.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9536,7 +9525,7 @@ "model.layers.6.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9560,7 +9549,7 @@ "model.layers.6.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9584,7 +9573,7 @@ "model.layers.6.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9608,7 +9597,7 @@ "model.layers.60.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -9632,7 +9621,7 @@ "model.layers.60.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9656,7 +9645,7 @@ "model.layers.60.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9680,7 +9669,7 @@ "model.layers.60.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9704,7 +9693,7 @@ "model.layers.60.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9728,7 +9717,7 @@ "model.layers.60.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9752,7 +9741,7 @@ "model.layers.60.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9776,7 +9765,7 @@ "model.layers.61.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -9800,7 +9789,7 @@ "model.layers.61.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9824,7 +9813,7 @@ "model.layers.61.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9848,7 +9837,7 @@ "model.layers.61.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9872,7 +9861,7 @@ "model.layers.61.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9896,7 +9885,7 @@ "model.layers.61.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9920,7 +9909,7 @@ "model.layers.61.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9944,7 +9933,7 @@ "model.layers.62.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -9968,7 +9957,7 @@ "model.layers.62.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9992,7 +9981,7 @@ "model.layers.62.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10016,7 +10005,7 @@ "model.layers.62.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10040,7 +10029,7 @@ "model.layers.62.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10064,7 +10053,7 @@ "model.layers.62.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10088,7 +10077,7 @@ "model.layers.62.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10112,7 +10101,7 @@ "model.layers.63.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -10136,7 +10125,7 @@ "model.layers.63.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10160,7 +10149,7 @@ "model.layers.63.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10184,7 +10173,7 @@ "model.layers.63.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10208,7 +10197,7 @@ "model.layers.63.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10232,7 +10221,7 @@ "model.layers.63.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10256,7 +10245,7 @@ "model.layers.63.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10280,7 +10269,7 @@ "model.layers.64.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -10304,7 +10293,7 @@ "model.layers.64.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10328,7 +10317,7 @@ "model.layers.64.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10352,7 +10341,7 @@ "model.layers.64.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10376,7 +10365,7 @@ "model.layers.64.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10400,7 +10389,7 @@ "model.layers.64.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10424,7 +10413,7 @@ "model.layers.64.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10448,7 +10437,7 @@ "model.layers.65.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -10472,7 +10461,7 @@ "model.layers.65.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10496,7 +10485,7 @@ "model.layers.65.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10520,7 +10509,7 @@ "model.layers.65.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10544,7 +10533,7 @@ "model.layers.65.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10568,7 +10557,7 @@ "model.layers.65.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10592,7 +10581,7 @@ "model.layers.65.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10616,7 +10605,7 @@ "model.layers.66.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -10640,7 +10629,7 @@ "model.layers.66.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10664,7 +10653,7 @@ "model.layers.66.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10688,7 +10677,7 @@ "model.layers.66.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10712,7 +10701,7 @@ "model.layers.66.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10736,7 +10725,7 @@ "model.layers.66.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10760,7 +10749,7 @@ "model.layers.66.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10784,7 +10773,7 @@ "model.layers.67.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -10808,7 +10797,7 @@ "model.layers.67.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10832,7 +10821,7 @@ "model.layers.67.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10856,7 +10845,7 @@ "model.layers.67.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10880,7 +10869,7 @@ "model.layers.67.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10904,7 +10893,7 @@ "model.layers.67.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10928,7 +10917,7 @@ "model.layers.67.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10952,7 +10941,7 @@ "model.layers.68.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -10976,7 +10965,7 @@ "model.layers.68.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11000,7 +10989,7 @@ "model.layers.68.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11024,7 +11013,7 @@ "model.layers.68.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11048,7 +11037,7 @@ "model.layers.68.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11072,7 +11061,7 @@ "model.layers.68.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11096,7 +11085,7 @@ "model.layers.68.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11120,7 +11109,7 @@ "model.layers.69.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -11144,7 +11133,7 @@ "model.layers.69.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11168,7 +11157,7 @@ "model.layers.69.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11192,7 +11181,7 @@ "model.layers.69.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11216,7 +11205,7 @@ "model.layers.69.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11240,7 +11229,7 @@ "model.layers.69.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11264,7 +11253,7 @@ "model.layers.69.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11288,7 +11277,7 @@ "model.layers.7.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -11312,7 +11301,7 @@ "model.layers.7.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11336,7 +11325,7 @@ "model.layers.7.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11360,7 +11349,7 @@ "model.layers.7.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11384,7 +11373,7 @@ "model.layers.7.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11408,7 +11397,7 @@ "model.layers.7.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11432,7 +11421,7 @@ "model.layers.7.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11456,7 +11445,7 @@ "model.layers.70.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -11480,7 +11469,7 @@ "model.layers.70.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11504,7 +11493,7 @@ "model.layers.70.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11528,7 +11517,7 @@ "model.layers.70.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11552,7 +11541,7 @@ "model.layers.70.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11576,7 +11565,7 @@ "model.layers.70.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11600,7 +11589,7 @@ "model.layers.70.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11624,7 +11613,7 @@ "model.layers.71.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -11648,7 +11637,7 @@ "model.layers.71.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11672,7 +11661,7 @@ "model.layers.71.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11696,7 +11685,7 @@ "model.layers.71.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11720,7 +11709,7 @@ "model.layers.71.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11744,7 +11733,7 @@ "model.layers.71.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11768,7 +11757,7 @@ "model.layers.71.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11792,7 +11781,7 @@ "model.layers.72.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -11816,7 +11805,7 @@ "model.layers.72.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11840,7 +11829,7 @@ "model.layers.72.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11864,7 +11853,7 @@ "model.layers.72.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11888,7 +11877,7 @@ "model.layers.72.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11912,7 +11901,7 @@ "model.layers.72.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11936,7 +11925,7 @@ "model.layers.72.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11960,7 +11949,7 @@ "model.layers.73.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -11984,7 +11973,7 @@ "model.layers.73.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12008,7 +11997,7 @@ "model.layers.73.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12032,7 +12021,7 @@ "model.layers.73.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12056,7 +12045,7 @@ "model.layers.73.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12080,7 +12069,7 @@ "model.layers.73.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12104,7 +12093,7 @@ "model.layers.73.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12128,7 +12117,7 @@ "model.layers.74.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -12152,7 +12141,7 @@ "model.layers.74.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12176,7 +12165,7 @@ "model.layers.74.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12200,7 +12189,7 @@ "model.layers.74.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12224,7 +12213,7 @@ "model.layers.74.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12248,7 +12237,7 @@ "model.layers.74.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12272,7 +12261,7 @@ "model.layers.74.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12296,7 +12285,7 @@ "model.layers.75.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -12320,7 +12309,7 @@ "model.layers.75.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12344,7 +12333,7 @@ "model.layers.75.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12368,7 +12357,7 @@ "model.layers.75.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12392,7 +12381,7 @@ "model.layers.75.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12416,7 +12405,7 @@ "model.layers.75.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12440,7 +12429,7 @@ "model.layers.75.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12464,7 +12453,7 @@ "model.layers.76.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -12488,7 +12477,7 @@ "model.layers.76.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12512,7 +12501,7 @@ "model.layers.76.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12536,7 +12525,7 @@ "model.layers.76.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12560,7 +12549,7 @@ "model.layers.76.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12584,7 +12573,7 @@ "model.layers.76.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12608,7 +12597,7 @@ "model.layers.76.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12632,7 +12621,7 @@ "model.layers.77.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -12656,7 +12645,7 @@ "model.layers.77.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12680,7 +12669,7 @@ "model.layers.77.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12704,7 +12693,7 @@ "model.layers.77.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12728,7 +12717,7 @@ "model.layers.77.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12752,7 +12741,7 @@ "model.layers.77.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12776,7 +12765,7 @@ "model.layers.77.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12800,7 +12789,7 @@ "model.layers.78.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -12824,7 +12813,7 @@ "model.layers.78.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12848,7 +12837,7 @@ "model.layers.78.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12872,7 +12861,7 @@ "model.layers.78.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12896,7 +12885,7 @@ "model.layers.78.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12920,7 +12909,7 @@ "model.layers.78.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12944,7 +12933,7 @@ "model.layers.78.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12968,7 +12957,7 @@ "model.layers.79.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -12992,7 +12981,7 @@ "model.layers.79.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13016,7 +13005,7 @@ "model.layers.79.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13040,7 +13029,7 @@ "model.layers.79.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13064,7 +13053,7 @@ "model.layers.79.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13088,7 +13077,7 @@ "model.layers.79.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13112,7 +13101,7 @@ "model.layers.79.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13136,7 +13125,7 @@ "model.layers.8.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -13160,7 +13149,7 @@ "model.layers.8.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13184,7 +13173,7 @@ "model.layers.8.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13208,7 +13197,7 @@ "model.layers.8.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13232,7 +13221,7 @@ "model.layers.8.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13256,7 +13245,7 @@ "model.layers.8.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13280,7 +13269,7 @@ "model.layers.8.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13304,7 +13293,7 @@ "model.layers.9.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 29568, "in_features": 29568, @@ -13328,7 +13317,7 @@ "model.layers.9.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13352,7 +13341,7 @@ "model.layers.9.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13376,7 +13365,7 @@ "model.layers.9.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13400,7 +13389,7 @@ "model.layers.9.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13424,7 +13413,7 @@ "model.layers.9.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13448,7 +13437,7 @@ "model.layers.9.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13469,6 +13458,17 @@ 8 ] } - } - } -} \ No newline at end of file + }, + "quant_method": "vptq" + }, + "rms_norm_eps": 1e-06, + "rope_scaling": null, + "rope_theta": 1000000.0, + "sliding_window": null, + "tie_word_embeddings": false, + "torch_dtype": "bfloat16", + "transformers_version": "4.49.0", + "use_cache": true, + "use_sliding_window": false, + "vocab_size": 152064 +}