diff --git "a/config.json" "b/config.json" --- "a/config.json" +++ "b/config.json" @@ -1,5 +1,5 @@ { - "_name_or_path": "/home/aiscuser/yangwang/upload_meta_llama_3.3_70b/Meta-Llama-3.3-70B-Instruct-v8-k65536-0-woft", + "_name_or_path": "/home/aiscuser/yangwang/convert_model/models/VPTQ-community--Meta-Llama-3.3-70B-Instruct-v8-k65536-0-woft", "architectures": [ "LlamaForCausalLM" ], @@ -28,7 +28,7 @@ "model.layers.0.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -52,7 +52,7 @@ "model.layers.0.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -76,7 +76,7 @@ "model.layers.0.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -100,7 +100,7 @@ "model.layers.0.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -124,7 +124,7 @@ "model.layers.0.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -148,7 +148,7 @@ "model.layers.0.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -172,7 +172,7 @@ "model.layers.0.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -196,7 +196,7 @@ "model.layers.1.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -220,7 +220,7 @@ "model.layers.1.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -244,7 +244,7 @@ "model.layers.1.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -268,7 +268,7 @@ "model.layers.1.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -292,7 +292,7 @@ "model.layers.1.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -316,7 +316,7 @@ "model.layers.1.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -340,7 +340,7 @@ "model.layers.1.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -364,7 +364,7 @@ "model.layers.10.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -388,7 +388,7 @@ "model.layers.10.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -412,7 +412,7 @@ "model.layers.10.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -436,7 +436,7 @@ "model.layers.10.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -460,7 +460,7 @@ "model.layers.10.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -484,7 +484,7 @@ "model.layers.10.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -508,7 +508,7 @@ "model.layers.10.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -532,7 +532,7 @@ "model.layers.11.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -556,7 +556,7 @@ "model.layers.11.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -580,7 +580,7 @@ "model.layers.11.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -604,7 +604,7 @@ "model.layers.11.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -628,7 +628,7 @@ "model.layers.11.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -652,7 +652,7 @@ "model.layers.11.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -676,7 +676,7 @@ "model.layers.11.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -700,7 +700,7 @@ "model.layers.12.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -724,7 +724,7 @@ "model.layers.12.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -748,7 +748,7 @@ "model.layers.12.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -772,7 +772,7 @@ "model.layers.12.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -796,7 +796,7 @@ "model.layers.12.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -820,7 +820,7 @@ "model.layers.12.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -844,7 +844,7 @@ "model.layers.12.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -868,7 +868,7 @@ "model.layers.13.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -892,7 +892,7 @@ "model.layers.13.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -916,7 +916,7 @@ "model.layers.13.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -940,7 +940,7 @@ "model.layers.13.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -964,7 +964,7 @@ "model.layers.13.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -988,7 +988,7 @@ "model.layers.13.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1012,7 +1012,7 @@ "model.layers.13.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1036,7 +1036,7 @@ "model.layers.14.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -1060,7 +1060,7 @@ "model.layers.14.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1084,7 +1084,7 @@ "model.layers.14.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1108,7 +1108,7 @@ "model.layers.14.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1132,7 +1132,7 @@ "model.layers.14.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1156,7 +1156,7 @@ "model.layers.14.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1180,7 +1180,7 @@ "model.layers.14.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1204,7 +1204,7 @@ "model.layers.15.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -1228,7 +1228,7 @@ "model.layers.15.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1252,7 +1252,7 @@ "model.layers.15.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1276,7 +1276,7 @@ "model.layers.15.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1300,7 +1300,7 @@ "model.layers.15.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1324,7 +1324,7 @@ "model.layers.15.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1348,7 +1348,7 @@ "model.layers.15.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1372,7 +1372,7 @@ "model.layers.16.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -1396,7 +1396,7 @@ "model.layers.16.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1420,7 +1420,7 @@ "model.layers.16.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1444,7 +1444,7 @@ "model.layers.16.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1468,7 +1468,7 @@ "model.layers.16.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1492,7 +1492,7 @@ "model.layers.16.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1516,7 +1516,7 @@ "model.layers.16.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1540,7 +1540,7 @@ "model.layers.17.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -1564,7 +1564,7 @@ "model.layers.17.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1588,7 +1588,7 @@ "model.layers.17.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1612,7 +1612,7 @@ "model.layers.17.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1636,7 +1636,7 @@ "model.layers.17.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1660,7 +1660,7 @@ "model.layers.17.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1684,7 +1684,7 @@ "model.layers.17.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1708,7 +1708,7 @@ "model.layers.18.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -1732,7 +1732,7 @@ "model.layers.18.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1756,7 +1756,7 @@ "model.layers.18.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1780,7 +1780,7 @@ "model.layers.18.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1804,7 +1804,7 @@ "model.layers.18.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1828,7 +1828,7 @@ "model.layers.18.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1852,7 +1852,7 @@ "model.layers.18.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1876,7 +1876,7 @@ "model.layers.19.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -1900,7 +1900,7 @@ "model.layers.19.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1924,7 +1924,7 @@ "model.layers.19.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1948,7 +1948,7 @@ "model.layers.19.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1972,7 +1972,7 @@ "model.layers.19.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -1996,7 +1996,7 @@ "model.layers.19.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2020,7 +2020,7 @@ "model.layers.19.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2044,7 +2044,7 @@ "model.layers.2.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -2068,7 +2068,7 @@ "model.layers.2.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2092,7 +2092,7 @@ "model.layers.2.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2116,7 +2116,7 @@ "model.layers.2.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2140,7 +2140,7 @@ "model.layers.2.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2164,7 +2164,7 @@ "model.layers.2.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2188,7 +2188,7 @@ "model.layers.2.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2212,7 +2212,7 @@ "model.layers.20.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -2236,7 +2236,7 @@ "model.layers.20.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2260,7 +2260,7 @@ "model.layers.20.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2284,7 +2284,7 @@ "model.layers.20.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2308,7 +2308,7 @@ "model.layers.20.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2332,7 +2332,7 @@ "model.layers.20.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2356,7 +2356,7 @@ "model.layers.20.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2380,7 +2380,7 @@ "model.layers.21.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -2404,7 +2404,7 @@ "model.layers.21.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2428,7 +2428,7 @@ "model.layers.21.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2452,7 +2452,7 @@ "model.layers.21.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2476,7 +2476,7 @@ "model.layers.21.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2500,7 +2500,7 @@ "model.layers.21.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2524,7 +2524,7 @@ "model.layers.21.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2548,7 +2548,7 @@ "model.layers.22.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -2572,7 +2572,7 @@ "model.layers.22.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2596,7 +2596,7 @@ "model.layers.22.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2620,7 +2620,7 @@ "model.layers.22.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2644,7 +2644,7 @@ "model.layers.22.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2668,7 +2668,7 @@ "model.layers.22.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2692,7 +2692,7 @@ "model.layers.22.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2716,7 +2716,7 @@ "model.layers.23.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -2740,7 +2740,7 @@ "model.layers.23.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2764,7 +2764,7 @@ "model.layers.23.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2788,7 +2788,7 @@ "model.layers.23.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2812,7 +2812,7 @@ "model.layers.23.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2836,7 +2836,7 @@ "model.layers.23.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2860,7 +2860,7 @@ "model.layers.23.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2884,7 +2884,7 @@ "model.layers.24.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -2908,7 +2908,7 @@ "model.layers.24.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2932,7 +2932,7 @@ "model.layers.24.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2956,7 +2956,7 @@ "model.layers.24.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -2980,7 +2980,7 @@ "model.layers.24.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3004,7 +3004,7 @@ "model.layers.24.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3028,7 +3028,7 @@ "model.layers.24.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3052,7 +3052,7 @@ "model.layers.25.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -3076,7 +3076,7 @@ "model.layers.25.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3100,7 +3100,7 @@ "model.layers.25.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3124,7 +3124,7 @@ "model.layers.25.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3148,7 +3148,7 @@ "model.layers.25.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3172,7 +3172,7 @@ "model.layers.25.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3196,7 +3196,7 @@ "model.layers.25.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3220,7 +3220,7 @@ "model.layers.26.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -3244,7 +3244,7 @@ "model.layers.26.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3268,7 +3268,7 @@ "model.layers.26.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3292,7 +3292,7 @@ "model.layers.26.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3316,7 +3316,7 @@ "model.layers.26.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3340,7 +3340,7 @@ "model.layers.26.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3364,7 +3364,7 @@ "model.layers.26.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3388,7 +3388,7 @@ "model.layers.27.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -3412,7 +3412,7 @@ "model.layers.27.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3436,7 +3436,7 @@ "model.layers.27.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3460,7 +3460,7 @@ "model.layers.27.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3484,7 +3484,7 @@ "model.layers.27.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3508,7 +3508,7 @@ "model.layers.27.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3532,7 +3532,7 @@ "model.layers.27.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3556,7 +3556,7 @@ "model.layers.28.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -3580,7 +3580,7 @@ "model.layers.28.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3604,7 +3604,7 @@ "model.layers.28.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3628,7 +3628,7 @@ "model.layers.28.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3652,7 +3652,7 @@ "model.layers.28.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3676,7 +3676,7 @@ "model.layers.28.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3700,7 +3700,7 @@ "model.layers.28.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3724,7 +3724,7 @@ "model.layers.29.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -3748,7 +3748,7 @@ "model.layers.29.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3772,7 +3772,7 @@ "model.layers.29.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3796,7 +3796,7 @@ "model.layers.29.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3820,7 +3820,7 @@ "model.layers.29.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3844,7 +3844,7 @@ "model.layers.29.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3868,7 +3868,7 @@ "model.layers.29.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3892,7 +3892,7 @@ "model.layers.3.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -3916,7 +3916,7 @@ "model.layers.3.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3940,7 +3940,7 @@ "model.layers.3.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3964,7 +3964,7 @@ "model.layers.3.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -3988,7 +3988,7 @@ "model.layers.3.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4012,7 +4012,7 @@ "model.layers.3.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4036,7 +4036,7 @@ "model.layers.3.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4060,7 +4060,7 @@ "model.layers.30.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -4084,7 +4084,7 @@ "model.layers.30.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4108,7 +4108,7 @@ "model.layers.30.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4132,7 +4132,7 @@ "model.layers.30.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4156,7 +4156,7 @@ "model.layers.30.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4180,7 +4180,7 @@ "model.layers.30.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4204,7 +4204,7 @@ "model.layers.30.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4228,7 +4228,7 @@ "model.layers.31.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -4252,7 +4252,7 @@ "model.layers.31.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4276,7 +4276,7 @@ "model.layers.31.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4300,7 +4300,7 @@ "model.layers.31.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4324,7 +4324,7 @@ "model.layers.31.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4348,7 +4348,7 @@ "model.layers.31.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4372,7 +4372,7 @@ "model.layers.31.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4396,7 +4396,7 @@ "model.layers.32.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -4420,7 +4420,7 @@ "model.layers.32.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4444,7 +4444,7 @@ "model.layers.32.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4468,7 +4468,7 @@ "model.layers.32.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4492,7 +4492,7 @@ "model.layers.32.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4516,7 +4516,7 @@ "model.layers.32.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4540,7 +4540,7 @@ "model.layers.32.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4564,7 +4564,7 @@ "model.layers.33.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -4588,7 +4588,7 @@ "model.layers.33.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4612,7 +4612,7 @@ "model.layers.33.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4636,7 +4636,7 @@ "model.layers.33.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4660,7 +4660,7 @@ "model.layers.33.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4684,7 +4684,7 @@ "model.layers.33.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4708,7 +4708,7 @@ "model.layers.33.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4732,7 +4732,7 @@ "model.layers.34.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -4756,7 +4756,7 @@ "model.layers.34.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4780,7 +4780,7 @@ "model.layers.34.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4804,7 +4804,7 @@ "model.layers.34.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4828,7 +4828,7 @@ "model.layers.34.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4852,7 +4852,7 @@ "model.layers.34.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4876,7 +4876,7 @@ "model.layers.34.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4900,7 +4900,7 @@ "model.layers.35.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -4924,7 +4924,7 @@ "model.layers.35.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4948,7 +4948,7 @@ "model.layers.35.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4972,7 +4972,7 @@ "model.layers.35.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -4996,7 +4996,7 @@ "model.layers.35.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5020,7 +5020,7 @@ "model.layers.35.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5044,7 +5044,7 @@ "model.layers.35.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5068,7 +5068,7 @@ "model.layers.36.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -5092,7 +5092,7 @@ "model.layers.36.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5116,7 +5116,7 @@ "model.layers.36.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5140,7 +5140,7 @@ "model.layers.36.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5164,7 +5164,7 @@ "model.layers.36.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5188,7 +5188,7 @@ "model.layers.36.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5212,7 +5212,7 @@ "model.layers.36.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5236,7 +5236,7 @@ "model.layers.37.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -5260,7 +5260,7 @@ "model.layers.37.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5284,7 +5284,7 @@ "model.layers.37.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5308,7 +5308,7 @@ "model.layers.37.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5332,7 +5332,7 @@ "model.layers.37.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5356,7 +5356,7 @@ "model.layers.37.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5380,7 +5380,7 @@ "model.layers.37.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5404,7 +5404,7 @@ "model.layers.38.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -5428,7 +5428,7 @@ "model.layers.38.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5452,7 +5452,7 @@ "model.layers.38.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5476,7 +5476,7 @@ "model.layers.38.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5500,7 +5500,7 @@ "model.layers.38.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5524,7 +5524,7 @@ "model.layers.38.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5548,7 +5548,7 @@ "model.layers.38.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5572,7 +5572,7 @@ "model.layers.39.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -5596,7 +5596,7 @@ "model.layers.39.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5620,7 +5620,7 @@ "model.layers.39.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5644,7 +5644,7 @@ "model.layers.39.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5668,7 +5668,7 @@ "model.layers.39.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5692,7 +5692,7 @@ "model.layers.39.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5716,7 +5716,7 @@ "model.layers.39.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5740,7 +5740,7 @@ "model.layers.4.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -5764,7 +5764,7 @@ "model.layers.4.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5788,7 +5788,7 @@ "model.layers.4.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5812,7 +5812,7 @@ "model.layers.4.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5836,7 +5836,7 @@ "model.layers.4.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5860,7 +5860,7 @@ "model.layers.4.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5884,7 +5884,7 @@ "model.layers.4.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5908,7 +5908,7 @@ "model.layers.40.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -5932,7 +5932,7 @@ "model.layers.40.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5956,7 +5956,7 @@ "model.layers.40.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -5980,7 +5980,7 @@ "model.layers.40.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6004,7 +6004,7 @@ "model.layers.40.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6028,7 +6028,7 @@ "model.layers.40.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6052,7 +6052,7 @@ "model.layers.40.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6076,7 +6076,7 @@ "model.layers.41.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -6100,7 +6100,7 @@ "model.layers.41.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6124,7 +6124,7 @@ "model.layers.41.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6148,7 +6148,7 @@ "model.layers.41.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6172,7 +6172,7 @@ "model.layers.41.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6196,7 +6196,7 @@ "model.layers.41.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6220,7 +6220,7 @@ "model.layers.41.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6244,7 +6244,7 @@ "model.layers.42.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -6268,7 +6268,7 @@ "model.layers.42.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6292,7 +6292,7 @@ "model.layers.42.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6316,7 +6316,7 @@ "model.layers.42.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6340,7 +6340,7 @@ "model.layers.42.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6364,7 +6364,7 @@ "model.layers.42.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6388,7 +6388,7 @@ "model.layers.42.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6412,7 +6412,7 @@ "model.layers.43.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -6436,7 +6436,7 @@ "model.layers.43.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6460,7 +6460,7 @@ "model.layers.43.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6484,7 +6484,7 @@ "model.layers.43.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6508,7 +6508,7 @@ "model.layers.43.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6532,7 +6532,7 @@ "model.layers.43.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6556,7 +6556,7 @@ "model.layers.43.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6580,7 +6580,7 @@ "model.layers.44.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -6604,7 +6604,7 @@ "model.layers.44.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6628,7 +6628,7 @@ "model.layers.44.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6652,7 +6652,7 @@ "model.layers.44.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6676,7 +6676,7 @@ "model.layers.44.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6700,7 +6700,7 @@ "model.layers.44.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6724,7 +6724,7 @@ "model.layers.44.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6748,7 +6748,7 @@ "model.layers.45.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -6772,7 +6772,7 @@ "model.layers.45.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6796,7 +6796,7 @@ "model.layers.45.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6820,7 +6820,7 @@ "model.layers.45.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6844,7 +6844,7 @@ "model.layers.45.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6868,7 +6868,7 @@ "model.layers.45.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6892,7 +6892,7 @@ "model.layers.45.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6916,7 +6916,7 @@ "model.layers.46.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -6940,7 +6940,7 @@ "model.layers.46.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6964,7 +6964,7 @@ "model.layers.46.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -6988,7 +6988,7 @@ "model.layers.46.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7012,7 +7012,7 @@ "model.layers.46.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7036,7 +7036,7 @@ "model.layers.46.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7060,7 +7060,7 @@ "model.layers.46.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7084,7 +7084,7 @@ "model.layers.47.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -7108,7 +7108,7 @@ "model.layers.47.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7132,7 +7132,7 @@ "model.layers.47.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7156,7 +7156,7 @@ "model.layers.47.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7180,7 +7180,7 @@ "model.layers.47.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7204,7 +7204,7 @@ "model.layers.47.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7228,7 +7228,7 @@ "model.layers.47.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7252,7 +7252,7 @@ "model.layers.48.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -7276,7 +7276,7 @@ "model.layers.48.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7300,7 +7300,7 @@ "model.layers.48.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7324,7 +7324,7 @@ "model.layers.48.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7348,7 +7348,7 @@ "model.layers.48.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7372,7 +7372,7 @@ "model.layers.48.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7396,7 +7396,7 @@ "model.layers.48.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7420,7 +7420,7 @@ "model.layers.49.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -7444,7 +7444,7 @@ "model.layers.49.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7468,7 +7468,7 @@ "model.layers.49.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7492,7 +7492,7 @@ "model.layers.49.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7516,7 +7516,7 @@ "model.layers.49.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7540,7 +7540,7 @@ "model.layers.49.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7564,7 +7564,7 @@ "model.layers.49.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7588,7 +7588,7 @@ "model.layers.5.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -7612,7 +7612,7 @@ "model.layers.5.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7636,7 +7636,7 @@ "model.layers.5.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7660,7 +7660,7 @@ "model.layers.5.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7684,7 +7684,7 @@ "model.layers.5.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7708,7 +7708,7 @@ "model.layers.5.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7732,7 +7732,7 @@ "model.layers.5.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7756,7 +7756,7 @@ "model.layers.50.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -7780,7 +7780,7 @@ "model.layers.50.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7804,7 +7804,7 @@ "model.layers.50.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7828,7 +7828,7 @@ "model.layers.50.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7852,7 +7852,7 @@ "model.layers.50.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7876,7 +7876,7 @@ "model.layers.50.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7900,7 +7900,7 @@ "model.layers.50.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7924,7 +7924,7 @@ "model.layers.51.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -7948,7 +7948,7 @@ "model.layers.51.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7972,7 +7972,7 @@ "model.layers.51.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -7996,7 +7996,7 @@ "model.layers.51.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8020,7 +8020,7 @@ "model.layers.51.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8044,7 +8044,7 @@ "model.layers.51.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8068,7 +8068,7 @@ "model.layers.51.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8092,7 +8092,7 @@ "model.layers.52.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -8116,7 +8116,7 @@ "model.layers.52.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8140,7 +8140,7 @@ "model.layers.52.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8164,7 +8164,7 @@ "model.layers.52.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8188,7 +8188,7 @@ "model.layers.52.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8212,7 +8212,7 @@ "model.layers.52.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8236,7 +8236,7 @@ "model.layers.52.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8260,7 +8260,7 @@ "model.layers.53.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -8284,7 +8284,7 @@ "model.layers.53.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8308,7 +8308,7 @@ "model.layers.53.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8332,7 +8332,7 @@ "model.layers.53.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8356,7 +8356,7 @@ "model.layers.53.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8380,7 +8380,7 @@ "model.layers.53.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8404,7 +8404,7 @@ "model.layers.53.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8428,7 +8428,7 @@ "model.layers.54.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -8452,7 +8452,7 @@ "model.layers.54.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8476,7 +8476,7 @@ "model.layers.54.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8500,7 +8500,7 @@ "model.layers.54.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8524,7 +8524,7 @@ "model.layers.54.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8548,7 +8548,7 @@ "model.layers.54.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8572,7 +8572,7 @@ "model.layers.54.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8596,7 +8596,7 @@ "model.layers.55.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -8620,7 +8620,7 @@ "model.layers.55.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8644,7 +8644,7 @@ "model.layers.55.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8668,7 +8668,7 @@ "model.layers.55.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8692,7 +8692,7 @@ "model.layers.55.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8716,7 +8716,7 @@ "model.layers.55.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8740,7 +8740,7 @@ "model.layers.55.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8764,7 +8764,7 @@ "model.layers.56.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -8788,7 +8788,7 @@ "model.layers.56.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8812,7 +8812,7 @@ "model.layers.56.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8836,7 +8836,7 @@ "model.layers.56.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8860,7 +8860,7 @@ "model.layers.56.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8884,7 +8884,7 @@ "model.layers.56.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8908,7 +8908,7 @@ "model.layers.56.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8932,7 +8932,7 @@ "model.layers.57.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -8956,7 +8956,7 @@ "model.layers.57.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -8980,7 +8980,7 @@ "model.layers.57.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9004,7 +9004,7 @@ "model.layers.57.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9028,7 +9028,7 @@ "model.layers.57.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9052,7 +9052,7 @@ "model.layers.57.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9076,7 +9076,7 @@ "model.layers.57.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9100,7 +9100,7 @@ "model.layers.58.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -9124,7 +9124,7 @@ "model.layers.58.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9148,7 +9148,7 @@ "model.layers.58.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9172,7 +9172,7 @@ "model.layers.58.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9196,7 +9196,7 @@ "model.layers.58.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9220,7 +9220,7 @@ "model.layers.58.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9244,7 +9244,7 @@ "model.layers.58.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9268,7 +9268,7 @@ "model.layers.59.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -9292,7 +9292,7 @@ "model.layers.59.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9316,7 +9316,7 @@ "model.layers.59.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9340,7 +9340,7 @@ "model.layers.59.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9364,7 +9364,7 @@ "model.layers.59.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9388,7 +9388,7 @@ "model.layers.59.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9412,7 +9412,7 @@ "model.layers.59.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9436,7 +9436,7 @@ "model.layers.6.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -9460,7 +9460,7 @@ "model.layers.6.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9484,7 +9484,7 @@ "model.layers.6.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9508,7 +9508,7 @@ "model.layers.6.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9532,7 +9532,7 @@ "model.layers.6.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9556,7 +9556,7 @@ "model.layers.6.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9580,7 +9580,7 @@ "model.layers.6.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9604,7 +9604,7 @@ "model.layers.60.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -9628,7 +9628,7 @@ "model.layers.60.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9652,7 +9652,7 @@ "model.layers.60.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9676,7 +9676,7 @@ "model.layers.60.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9700,7 +9700,7 @@ "model.layers.60.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9724,7 +9724,7 @@ "model.layers.60.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9748,7 +9748,7 @@ "model.layers.60.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9772,7 +9772,7 @@ "model.layers.61.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -9796,7 +9796,7 @@ "model.layers.61.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9820,7 +9820,7 @@ "model.layers.61.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9844,7 +9844,7 @@ "model.layers.61.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9868,7 +9868,7 @@ "model.layers.61.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9892,7 +9892,7 @@ "model.layers.61.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9916,7 +9916,7 @@ "model.layers.61.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9940,7 +9940,7 @@ "model.layers.62.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -9964,7 +9964,7 @@ "model.layers.62.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -9988,7 +9988,7 @@ "model.layers.62.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10012,7 +10012,7 @@ "model.layers.62.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10036,7 +10036,7 @@ "model.layers.62.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10060,7 +10060,7 @@ "model.layers.62.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10084,7 +10084,7 @@ "model.layers.62.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10108,7 +10108,7 @@ "model.layers.63.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -10132,7 +10132,7 @@ "model.layers.63.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10156,7 +10156,7 @@ "model.layers.63.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10180,7 +10180,7 @@ "model.layers.63.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10204,7 +10204,7 @@ "model.layers.63.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10228,7 +10228,7 @@ "model.layers.63.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10252,7 +10252,7 @@ "model.layers.63.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10276,7 +10276,7 @@ "model.layers.64.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -10300,7 +10300,7 @@ "model.layers.64.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10324,7 +10324,7 @@ "model.layers.64.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10348,7 +10348,7 @@ "model.layers.64.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10372,7 +10372,7 @@ "model.layers.64.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10396,7 +10396,7 @@ "model.layers.64.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10420,7 +10420,7 @@ "model.layers.64.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10444,7 +10444,7 @@ "model.layers.65.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -10468,7 +10468,7 @@ "model.layers.65.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10492,7 +10492,7 @@ "model.layers.65.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10516,7 +10516,7 @@ "model.layers.65.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10540,7 +10540,7 @@ "model.layers.65.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10564,7 +10564,7 @@ "model.layers.65.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10588,7 +10588,7 @@ "model.layers.65.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10612,7 +10612,7 @@ "model.layers.66.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -10636,7 +10636,7 @@ "model.layers.66.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10660,7 +10660,7 @@ "model.layers.66.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10684,7 +10684,7 @@ "model.layers.66.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10708,7 +10708,7 @@ "model.layers.66.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10732,7 +10732,7 @@ "model.layers.66.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10756,7 +10756,7 @@ "model.layers.66.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10780,7 +10780,7 @@ "model.layers.67.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -10804,7 +10804,7 @@ "model.layers.67.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10828,7 +10828,7 @@ "model.layers.67.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10852,7 +10852,7 @@ "model.layers.67.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10876,7 +10876,7 @@ "model.layers.67.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10900,7 +10900,7 @@ "model.layers.67.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10924,7 +10924,7 @@ "model.layers.67.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10948,7 +10948,7 @@ "model.layers.68.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -10972,7 +10972,7 @@ "model.layers.68.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -10996,7 +10996,7 @@ "model.layers.68.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11020,7 +11020,7 @@ "model.layers.68.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11044,7 +11044,7 @@ "model.layers.68.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11068,7 +11068,7 @@ "model.layers.68.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11092,7 +11092,7 @@ "model.layers.68.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11116,7 +11116,7 @@ "model.layers.69.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -11140,7 +11140,7 @@ "model.layers.69.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11164,7 +11164,7 @@ "model.layers.69.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11188,7 +11188,7 @@ "model.layers.69.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11212,7 +11212,7 @@ "model.layers.69.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11236,7 +11236,7 @@ "model.layers.69.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11260,7 +11260,7 @@ "model.layers.69.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11284,7 +11284,7 @@ "model.layers.7.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -11308,7 +11308,7 @@ "model.layers.7.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11332,7 +11332,7 @@ "model.layers.7.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11356,7 +11356,7 @@ "model.layers.7.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11380,7 +11380,7 @@ "model.layers.7.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11404,7 +11404,7 @@ "model.layers.7.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11428,7 +11428,7 @@ "model.layers.7.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11452,7 +11452,7 @@ "model.layers.70.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -11476,7 +11476,7 @@ "model.layers.70.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11500,7 +11500,7 @@ "model.layers.70.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11524,7 +11524,7 @@ "model.layers.70.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11548,7 +11548,7 @@ "model.layers.70.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11572,7 +11572,7 @@ "model.layers.70.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11596,7 +11596,7 @@ "model.layers.70.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11620,7 +11620,7 @@ "model.layers.71.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -11644,7 +11644,7 @@ "model.layers.71.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11668,7 +11668,7 @@ "model.layers.71.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11692,7 +11692,7 @@ "model.layers.71.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11716,7 +11716,7 @@ "model.layers.71.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11740,7 +11740,7 @@ "model.layers.71.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11764,7 +11764,7 @@ "model.layers.71.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11788,7 +11788,7 @@ "model.layers.72.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -11812,7 +11812,7 @@ "model.layers.72.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11836,7 +11836,7 @@ "model.layers.72.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11860,7 +11860,7 @@ "model.layers.72.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11884,7 +11884,7 @@ "model.layers.72.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11908,7 +11908,7 @@ "model.layers.72.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11932,7 +11932,7 @@ "model.layers.72.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -11956,7 +11956,7 @@ "model.layers.73.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -11980,7 +11980,7 @@ "model.layers.73.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12004,7 +12004,7 @@ "model.layers.73.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12028,7 +12028,7 @@ "model.layers.73.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12052,7 +12052,7 @@ "model.layers.73.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12076,7 +12076,7 @@ "model.layers.73.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12100,7 +12100,7 @@ "model.layers.73.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12124,7 +12124,7 @@ "model.layers.74.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -12148,7 +12148,7 @@ "model.layers.74.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12172,7 +12172,7 @@ "model.layers.74.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12196,7 +12196,7 @@ "model.layers.74.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12220,7 +12220,7 @@ "model.layers.74.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12244,7 +12244,7 @@ "model.layers.74.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12268,7 +12268,7 @@ "model.layers.74.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12292,7 +12292,7 @@ "model.layers.75.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -12316,7 +12316,7 @@ "model.layers.75.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12340,7 +12340,7 @@ "model.layers.75.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12364,7 +12364,7 @@ "model.layers.75.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12388,7 +12388,7 @@ "model.layers.75.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12412,7 +12412,7 @@ "model.layers.75.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12436,7 +12436,7 @@ "model.layers.75.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12460,7 +12460,7 @@ "model.layers.76.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -12484,7 +12484,7 @@ "model.layers.76.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12508,7 +12508,7 @@ "model.layers.76.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12532,7 +12532,7 @@ "model.layers.76.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12556,7 +12556,7 @@ "model.layers.76.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12580,7 +12580,7 @@ "model.layers.76.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12604,7 +12604,7 @@ "model.layers.76.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12628,7 +12628,7 @@ "model.layers.77.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -12652,7 +12652,7 @@ "model.layers.77.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12676,7 +12676,7 @@ "model.layers.77.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12700,7 +12700,7 @@ "model.layers.77.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12724,7 +12724,7 @@ "model.layers.77.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12748,7 +12748,7 @@ "model.layers.77.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12772,7 +12772,7 @@ "model.layers.77.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12796,7 +12796,7 @@ "model.layers.78.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -12820,7 +12820,7 @@ "model.layers.78.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12844,7 +12844,7 @@ "model.layers.78.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12868,7 +12868,7 @@ "model.layers.78.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12892,7 +12892,7 @@ "model.layers.78.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12916,7 +12916,7 @@ "model.layers.78.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12940,7 +12940,7 @@ "model.layers.78.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -12964,7 +12964,7 @@ "model.layers.79.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -12988,7 +12988,7 @@ "model.layers.79.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13012,7 +13012,7 @@ "model.layers.79.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13036,7 +13036,7 @@ "model.layers.79.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13060,7 +13060,7 @@ "model.layers.79.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13084,7 +13084,7 @@ "model.layers.79.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13108,7 +13108,7 @@ "model.layers.79.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13132,7 +13132,7 @@ "model.layers.8.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -13156,7 +13156,7 @@ "model.layers.8.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13180,7 +13180,7 @@ "model.layers.8.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13204,7 +13204,7 @@ "model.layers.8.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13228,7 +13228,7 @@ "model.layers.8.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13252,7 +13252,7 @@ "model.layers.8.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13276,7 +13276,7 @@ "model.layers.8.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13300,7 +13300,7 @@ "model.layers.9.mlp.down_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 28672, "in_features": 28672, @@ -13324,7 +13324,7 @@ "model.layers.9.mlp.gate_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13348,7 +13348,7 @@ "model.layers.9.mlp.up_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13372,7 +13372,7 @@ "model.layers.9.self_attn.k_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13396,7 +13396,7 @@ "model.layers.9.self_attn.o_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13420,7 +13420,7 @@ "model.layers.9.self_attn.q_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13444,7 +13444,7 @@ "model.layers.9.self_attn.v_proj": { "bias": false, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 8192, "in_features": 8192, @@ -13479,7 +13479,7 @@ "rope_theta": 500000.0, "tie_word_embeddings": false, "torch_dtype": "bfloat16", - "transformers_version": "4.47.0", + "transformers_version": "4.49.0", "use_cache": true, "vocab_size": 128256 }