Update config.json
Browse files- config.json +224 -224
config.json
CHANGED
|
@@ -31,7 +31,7 @@
|
|
| 31 |
"model.layers.0.mlp.down_proj": {
|
| 32 |
"bias": false,
|
| 33 |
"enable_norm": true,
|
| 34 |
-
"enable_perm":
|
| 35 |
"group_num": 1,
|
| 36 |
"group_size": 14192,
|
| 37 |
"in_features": 14336,
|
|
@@ -55,7 +55,7 @@
|
|
| 55 |
"model.layers.0.mlp.gate_proj": {
|
| 56 |
"bias": false,
|
| 57 |
"enable_norm": true,
|
| 58 |
-
"enable_perm":
|
| 59 |
"group_num": 1,
|
| 60 |
"group_size": 4055,
|
| 61 |
"in_features": 4096,
|
|
@@ -79,7 +79,7 @@
|
|
| 79 |
"model.layers.0.mlp.up_proj": {
|
| 80 |
"bias": false,
|
| 81 |
"enable_norm": true,
|
| 82 |
-
"enable_perm":
|
| 83 |
"group_num": 1,
|
| 84 |
"group_size": 4055,
|
| 85 |
"in_features": 4096,
|
|
@@ -103,7 +103,7 @@
|
|
| 103 |
"model.layers.0.self_attn.k_proj": {
|
| 104 |
"bias": false,
|
| 105 |
"enable_norm": true,
|
| 106 |
-
"enable_perm":
|
| 107 |
"group_num": 1,
|
| 108 |
"group_size": 4055,
|
| 109 |
"in_features": 4096,
|
|
@@ -127,7 +127,7 @@
|
|
| 127 |
"model.layers.0.self_attn.o_proj": {
|
| 128 |
"bias": false,
|
| 129 |
"enable_norm": true,
|
| 130 |
-
"enable_perm":
|
| 131 |
"group_num": 1,
|
| 132 |
"group_size": 4055,
|
| 133 |
"in_features": 4096,
|
|
@@ -151,7 +151,7 @@
|
|
| 151 |
"model.layers.0.self_attn.q_proj": {
|
| 152 |
"bias": false,
|
| 153 |
"enable_norm": true,
|
| 154 |
-
"enable_perm":
|
| 155 |
"group_num": 1,
|
| 156 |
"group_size": 4055,
|
| 157 |
"in_features": 4096,
|
|
@@ -175,7 +175,7 @@
|
|
| 175 |
"model.layers.0.self_attn.v_proj": {
|
| 176 |
"bias": false,
|
| 177 |
"enable_norm": true,
|
| 178 |
-
"enable_perm":
|
| 179 |
"group_num": 1,
|
| 180 |
"group_size": 4055,
|
| 181 |
"in_features": 4096,
|
|
@@ -199,7 +199,7 @@
|
|
| 199 |
"model.layers.1.mlp.down_proj": {
|
| 200 |
"bias": false,
|
| 201 |
"enable_norm": true,
|
| 202 |
-
"enable_perm":
|
| 203 |
"group_num": 1,
|
| 204 |
"group_size": 14192,
|
| 205 |
"in_features": 14336,
|
|
@@ -223,7 +223,7 @@
|
|
| 223 |
"model.layers.1.mlp.gate_proj": {
|
| 224 |
"bias": false,
|
| 225 |
"enable_norm": true,
|
| 226 |
-
"enable_perm":
|
| 227 |
"group_num": 1,
|
| 228 |
"group_size": 4055,
|
| 229 |
"in_features": 4096,
|
|
@@ -247,7 +247,7 @@
|
|
| 247 |
"model.layers.1.mlp.up_proj": {
|
| 248 |
"bias": false,
|
| 249 |
"enable_norm": true,
|
| 250 |
-
"enable_perm":
|
| 251 |
"group_num": 1,
|
| 252 |
"group_size": 4055,
|
| 253 |
"in_features": 4096,
|
|
@@ -271,7 +271,7 @@
|
|
| 271 |
"model.layers.1.self_attn.k_proj": {
|
| 272 |
"bias": false,
|
| 273 |
"enable_norm": true,
|
| 274 |
-
"enable_perm":
|
| 275 |
"group_num": 1,
|
| 276 |
"group_size": 4055,
|
| 277 |
"in_features": 4096,
|
|
@@ -295,7 +295,7 @@
|
|
| 295 |
"model.layers.1.self_attn.o_proj": {
|
| 296 |
"bias": false,
|
| 297 |
"enable_norm": true,
|
| 298 |
-
"enable_perm":
|
| 299 |
"group_num": 1,
|
| 300 |
"group_size": 4055,
|
| 301 |
"in_features": 4096,
|
|
@@ -319,7 +319,7 @@
|
|
| 319 |
"model.layers.1.self_attn.q_proj": {
|
| 320 |
"bias": false,
|
| 321 |
"enable_norm": true,
|
| 322 |
-
"enable_perm":
|
| 323 |
"group_num": 1,
|
| 324 |
"group_size": 4055,
|
| 325 |
"in_features": 4096,
|
|
@@ -343,7 +343,7 @@
|
|
| 343 |
"model.layers.1.self_attn.v_proj": {
|
| 344 |
"bias": false,
|
| 345 |
"enable_norm": true,
|
| 346 |
-
"enable_perm":
|
| 347 |
"group_num": 1,
|
| 348 |
"group_size": 4055,
|
| 349 |
"in_features": 4096,
|
|
@@ -367,7 +367,7 @@
|
|
| 367 |
"model.layers.10.mlp.down_proj": {
|
| 368 |
"bias": false,
|
| 369 |
"enable_norm": true,
|
| 370 |
-
"enable_perm":
|
| 371 |
"group_num": 1,
|
| 372 |
"group_size": 14192,
|
| 373 |
"in_features": 14336,
|
|
@@ -391,7 +391,7 @@
|
|
| 391 |
"model.layers.10.mlp.gate_proj": {
|
| 392 |
"bias": false,
|
| 393 |
"enable_norm": true,
|
| 394 |
-
"enable_perm":
|
| 395 |
"group_num": 1,
|
| 396 |
"group_size": 4055,
|
| 397 |
"in_features": 4096,
|
|
@@ -415,7 +415,7 @@
|
|
| 415 |
"model.layers.10.mlp.up_proj": {
|
| 416 |
"bias": false,
|
| 417 |
"enable_norm": true,
|
| 418 |
-
"enable_perm":
|
| 419 |
"group_num": 1,
|
| 420 |
"group_size": 4055,
|
| 421 |
"in_features": 4096,
|
|
@@ -439,7 +439,7 @@
|
|
| 439 |
"model.layers.10.self_attn.k_proj": {
|
| 440 |
"bias": false,
|
| 441 |
"enable_norm": true,
|
| 442 |
-
"enable_perm":
|
| 443 |
"group_num": 1,
|
| 444 |
"group_size": 4055,
|
| 445 |
"in_features": 4096,
|
|
@@ -463,7 +463,7 @@
|
|
| 463 |
"model.layers.10.self_attn.o_proj": {
|
| 464 |
"bias": false,
|
| 465 |
"enable_norm": true,
|
| 466 |
-
"enable_perm":
|
| 467 |
"group_num": 1,
|
| 468 |
"group_size": 4055,
|
| 469 |
"in_features": 4096,
|
|
@@ -487,7 +487,7 @@
|
|
| 487 |
"model.layers.10.self_attn.q_proj": {
|
| 488 |
"bias": false,
|
| 489 |
"enable_norm": true,
|
| 490 |
-
"enable_perm":
|
| 491 |
"group_num": 1,
|
| 492 |
"group_size": 4055,
|
| 493 |
"in_features": 4096,
|
|
@@ -511,7 +511,7 @@
|
|
| 511 |
"model.layers.10.self_attn.v_proj": {
|
| 512 |
"bias": false,
|
| 513 |
"enable_norm": true,
|
| 514 |
-
"enable_perm":
|
| 515 |
"group_num": 1,
|
| 516 |
"group_size": 4055,
|
| 517 |
"in_features": 4096,
|
|
@@ -535,7 +535,7 @@
|
|
| 535 |
"model.layers.11.mlp.down_proj": {
|
| 536 |
"bias": false,
|
| 537 |
"enable_norm": true,
|
| 538 |
-
"enable_perm":
|
| 539 |
"group_num": 1,
|
| 540 |
"group_size": 14192,
|
| 541 |
"in_features": 14336,
|
|
@@ -559,7 +559,7 @@
|
|
| 559 |
"model.layers.11.mlp.gate_proj": {
|
| 560 |
"bias": false,
|
| 561 |
"enable_norm": true,
|
| 562 |
-
"enable_perm":
|
| 563 |
"group_num": 1,
|
| 564 |
"group_size": 4055,
|
| 565 |
"in_features": 4096,
|
|
@@ -583,7 +583,7 @@
|
|
| 583 |
"model.layers.11.mlp.up_proj": {
|
| 584 |
"bias": false,
|
| 585 |
"enable_norm": true,
|
| 586 |
-
"enable_perm":
|
| 587 |
"group_num": 1,
|
| 588 |
"group_size": 4055,
|
| 589 |
"in_features": 4096,
|
|
@@ -607,7 +607,7 @@
|
|
| 607 |
"model.layers.11.self_attn.k_proj": {
|
| 608 |
"bias": false,
|
| 609 |
"enable_norm": true,
|
| 610 |
-
"enable_perm":
|
| 611 |
"group_num": 1,
|
| 612 |
"group_size": 4055,
|
| 613 |
"in_features": 4096,
|
|
@@ -631,7 +631,7 @@
|
|
| 631 |
"model.layers.11.self_attn.o_proj": {
|
| 632 |
"bias": false,
|
| 633 |
"enable_norm": true,
|
| 634 |
-
"enable_perm":
|
| 635 |
"group_num": 1,
|
| 636 |
"group_size": 4055,
|
| 637 |
"in_features": 4096,
|
|
@@ -655,7 +655,7 @@
|
|
| 655 |
"model.layers.11.self_attn.q_proj": {
|
| 656 |
"bias": false,
|
| 657 |
"enable_norm": true,
|
| 658 |
-
"enable_perm":
|
| 659 |
"group_num": 1,
|
| 660 |
"group_size": 4055,
|
| 661 |
"in_features": 4096,
|
|
@@ -679,7 +679,7 @@
|
|
| 679 |
"model.layers.11.self_attn.v_proj": {
|
| 680 |
"bias": false,
|
| 681 |
"enable_norm": true,
|
| 682 |
-
"enable_perm":
|
| 683 |
"group_num": 1,
|
| 684 |
"group_size": 4055,
|
| 685 |
"in_features": 4096,
|
|
@@ -703,7 +703,7 @@
|
|
| 703 |
"model.layers.12.mlp.down_proj": {
|
| 704 |
"bias": false,
|
| 705 |
"enable_norm": true,
|
| 706 |
-
"enable_perm":
|
| 707 |
"group_num": 1,
|
| 708 |
"group_size": 14192,
|
| 709 |
"in_features": 14336,
|
|
@@ -727,7 +727,7 @@
|
|
| 727 |
"model.layers.12.mlp.gate_proj": {
|
| 728 |
"bias": false,
|
| 729 |
"enable_norm": true,
|
| 730 |
-
"enable_perm":
|
| 731 |
"group_num": 1,
|
| 732 |
"group_size": 4055,
|
| 733 |
"in_features": 4096,
|
|
@@ -751,7 +751,7 @@
|
|
| 751 |
"model.layers.12.mlp.up_proj": {
|
| 752 |
"bias": false,
|
| 753 |
"enable_norm": true,
|
| 754 |
-
"enable_perm":
|
| 755 |
"group_num": 1,
|
| 756 |
"group_size": 4055,
|
| 757 |
"in_features": 4096,
|
|
@@ -775,7 +775,7 @@
|
|
| 775 |
"model.layers.12.self_attn.k_proj": {
|
| 776 |
"bias": false,
|
| 777 |
"enable_norm": true,
|
| 778 |
-
"enable_perm":
|
| 779 |
"group_num": 1,
|
| 780 |
"group_size": 4055,
|
| 781 |
"in_features": 4096,
|
|
@@ -799,7 +799,7 @@
|
|
| 799 |
"model.layers.12.self_attn.o_proj": {
|
| 800 |
"bias": false,
|
| 801 |
"enable_norm": true,
|
| 802 |
-
"enable_perm":
|
| 803 |
"group_num": 1,
|
| 804 |
"group_size": 4055,
|
| 805 |
"in_features": 4096,
|
|
@@ -823,7 +823,7 @@
|
|
| 823 |
"model.layers.12.self_attn.q_proj": {
|
| 824 |
"bias": false,
|
| 825 |
"enable_norm": true,
|
| 826 |
-
"enable_perm":
|
| 827 |
"group_num": 1,
|
| 828 |
"group_size": 4055,
|
| 829 |
"in_features": 4096,
|
|
@@ -847,7 +847,7 @@
|
|
| 847 |
"model.layers.12.self_attn.v_proj": {
|
| 848 |
"bias": false,
|
| 849 |
"enable_norm": true,
|
| 850 |
-
"enable_perm":
|
| 851 |
"group_num": 1,
|
| 852 |
"group_size": 4055,
|
| 853 |
"in_features": 4096,
|
|
@@ -871,7 +871,7 @@
|
|
| 871 |
"model.layers.13.mlp.down_proj": {
|
| 872 |
"bias": false,
|
| 873 |
"enable_norm": true,
|
| 874 |
-
"enable_perm":
|
| 875 |
"group_num": 1,
|
| 876 |
"group_size": 14192,
|
| 877 |
"in_features": 14336,
|
|
@@ -895,7 +895,7 @@
|
|
| 895 |
"model.layers.13.mlp.gate_proj": {
|
| 896 |
"bias": false,
|
| 897 |
"enable_norm": true,
|
| 898 |
-
"enable_perm":
|
| 899 |
"group_num": 1,
|
| 900 |
"group_size": 4055,
|
| 901 |
"in_features": 4096,
|
|
@@ -919,7 +919,7 @@
|
|
| 919 |
"model.layers.13.mlp.up_proj": {
|
| 920 |
"bias": false,
|
| 921 |
"enable_norm": true,
|
| 922 |
-
"enable_perm":
|
| 923 |
"group_num": 1,
|
| 924 |
"group_size": 4055,
|
| 925 |
"in_features": 4096,
|
|
@@ -943,7 +943,7 @@
|
|
| 943 |
"model.layers.13.self_attn.k_proj": {
|
| 944 |
"bias": false,
|
| 945 |
"enable_norm": true,
|
| 946 |
-
"enable_perm":
|
| 947 |
"group_num": 1,
|
| 948 |
"group_size": 4055,
|
| 949 |
"in_features": 4096,
|
|
@@ -967,7 +967,7 @@
|
|
| 967 |
"model.layers.13.self_attn.o_proj": {
|
| 968 |
"bias": false,
|
| 969 |
"enable_norm": true,
|
| 970 |
-
"enable_perm":
|
| 971 |
"group_num": 1,
|
| 972 |
"group_size": 4055,
|
| 973 |
"in_features": 4096,
|
|
@@ -991,7 +991,7 @@
|
|
| 991 |
"model.layers.13.self_attn.q_proj": {
|
| 992 |
"bias": false,
|
| 993 |
"enable_norm": true,
|
| 994 |
-
"enable_perm":
|
| 995 |
"group_num": 1,
|
| 996 |
"group_size": 4055,
|
| 997 |
"in_features": 4096,
|
|
@@ -1015,7 +1015,7 @@
|
|
| 1015 |
"model.layers.13.self_attn.v_proj": {
|
| 1016 |
"bias": false,
|
| 1017 |
"enable_norm": true,
|
| 1018 |
-
"enable_perm":
|
| 1019 |
"group_num": 1,
|
| 1020 |
"group_size": 4055,
|
| 1021 |
"in_features": 4096,
|
|
@@ -1039,7 +1039,7 @@
|
|
| 1039 |
"model.layers.14.mlp.down_proj": {
|
| 1040 |
"bias": false,
|
| 1041 |
"enable_norm": true,
|
| 1042 |
-
"enable_perm":
|
| 1043 |
"group_num": 1,
|
| 1044 |
"group_size": 14192,
|
| 1045 |
"in_features": 14336,
|
|
@@ -1063,7 +1063,7 @@
|
|
| 1063 |
"model.layers.14.mlp.gate_proj": {
|
| 1064 |
"bias": false,
|
| 1065 |
"enable_norm": true,
|
| 1066 |
-
"enable_perm":
|
| 1067 |
"group_num": 1,
|
| 1068 |
"group_size": 4055,
|
| 1069 |
"in_features": 4096,
|
|
@@ -1087,7 +1087,7 @@
|
|
| 1087 |
"model.layers.14.mlp.up_proj": {
|
| 1088 |
"bias": false,
|
| 1089 |
"enable_norm": true,
|
| 1090 |
-
"enable_perm":
|
| 1091 |
"group_num": 1,
|
| 1092 |
"group_size": 4055,
|
| 1093 |
"in_features": 4096,
|
|
@@ -1111,7 +1111,7 @@
|
|
| 1111 |
"model.layers.14.self_attn.k_proj": {
|
| 1112 |
"bias": false,
|
| 1113 |
"enable_norm": true,
|
| 1114 |
-
"enable_perm":
|
| 1115 |
"group_num": 1,
|
| 1116 |
"group_size": 4055,
|
| 1117 |
"in_features": 4096,
|
|
@@ -1135,7 +1135,7 @@
|
|
| 1135 |
"model.layers.14.self_attn.o_proj": {
|
| 1136 |
"bias": false,
|
| 1137 |
"enable_norm": true,
|
| 1138 |
-
"enable_perm":
|
| 1139 |
"group_num": 1,
|
| 1140 |
"group_size": 4055,
|
| 1141 |
"in_features": 4096,
|
|
@@ -1159,7 +1159,7 @@
|
|
| 1159 |
"model.layers.14.self_attn.q_proj": {
|
| 1160 |
"bias": false,
|
| 1161 |
"enable_norm": true,
|
| 1162 |
-
"enable_perm":
|
| 1163 |
"group_num": 1,
|
| 1164 |
"group_size": 4055,
|
| 1165 |
"in_features": 4096,
|
|
@@ -1183,7 +1183,7 @@
|
|
| 1183 |
"model.layers.14.self_attn.v_proj": {
|
| 1184 |
"bias": false,
|
| 1185 |
"enable_norm": true,
|
| 1186 |
-
"enable_perm":
|
| 1187 |
"group_num": 1,
|
| 1188 |
"group_size": 4055,
|
| 1189 |
"in_features": 4096,
|
|
@@ -1207,7 +1207,7 @@
|
|
| 1207 |
"model.layers.15.mlp.down_proj": {
|
| 1208 |
"bias": false,
|
| 1209 |
"enable_norm": true,
|
| 1210 |
-
"enable_perm":
|
| 1211 |
"group_num": 1,
|
| 1212 |
"group_size": 14192,
|
| 1213 |
"in_features": 14336,
|
|
@@ -1231,7 +1231,7 @@
|
|
| 1231 |
"model.layers.15.mlp.gate_proj": {
|
| 1232 |
"bias": false,
|
| 1233 |
"enable_norm": true,
|
| 1234 |
-
"enable_perm":
|
| 1235 |
"group_num": 1,
|
| 1236 |
"group_size": 4055,
|
| 1237 |
"in_features": 4096,
|
|
@@ -1255,7 +1255,7 @@
|
|
| 1255 |
"model.layers.15.mlp.up_proj": {
|
| 1256 |
"bias": false,
|
| 1257 |
"enable_norm": true,
|
| 1258 |
-
"enable_perm":
|
| 1259 |
"group_num": 1,
|
| 1260 |
"group_size": 4055,
|
| 1261 |
"in_features": 4096,
|
|
@@ -1279,7 +1279,7 @@
|
|
| 1279 |
"model.layers.15.self_attn.k_proj": {
|
| 1280 |
"bias": false,
|
| 1281 |
"enable_norm": true,
|
| 1282 |
-
"enable_perm":
|
| 1283 |
"group_num": 1,
|
| 1284 |
"group_size": 4055,
|
| 1285 |
"in_features": 4096,
|
|
@@ -1303,7 +1303,7 @@
|
|
| 1303 |
"model.layers.15.self_attn.o_proj": {
|
| 1304 |
"bias": false,
|
| 1305 |
"enable_norm": true,
|
| 1306 |
-
"enable_perm":
|
| 1307 |
"group_num": 1,
|
| 1308 |
"group_size": 4055,
|
| 1309 |
"in_features": 4096,
|
|
@@ -1327,7 +1327,7 @@
|
|
| 1327 |
"model.layers.15.self_attn.q_proj": {
|
| 1328 |
"bias": false,
|
| 1329 |
"enable_norm": true,
|
| 1330 |
-
"enable_perm":
|
| 1331 |
"group_num": 1,
|
| 1332 |
"group_size": 4055,
|
| 1333 |
"in_features": 4096,
|
|
@@ -1351,7 +1351,7 @@
|
|
| 1351 |
"model.layers.15.self_attn.v_proj": {
|
| 1352 |
"bias": false,
|
| 1353 |
"enable_norm": true,
|
| 1354 |
-
"enable_perm":
|
| 1355 |
"group_num": 1,
|
| 1356 |
"group_size": 4055,
|
| 1357 |
"in_features": 4096,
|
|
@@ -1375,7 +1375,7 @@
|
|
| 1375 |
"model.layers.16.mlp.down_proj": {
|
| 1376 |
"bias": false,
|
| 1377 |
"enable_norm": true,
|
| 1378 |
-
"enable_perm":
|
| 1379 |
"group_num": 1,
|
| 1380 |
"group_size": 14192,
|
| 1381 |
"in_features": 14336,
|
|
@@ -1399,7 +1399,7 @@
|
|
| 1399 |
"model.layers.16.mlp.gate_proj": {
|
| 1400 |
"bias": false,
|
| 1401 |
"enable_norm": true,
|
| 1402 |
-
"enable_perm":
|
| 1403 |
"group_num": 1,
|
| 1404 |
"group_size": 4055,
|
| 1405 |
"in_features": 4096,
|
|
@@ -1423,7 +1423,7 @@
|
|
| 1423 |
"model.layers.16.mlp.up_proj": {
|
| 1424 |
"bias": false,
|
| 1425 |
"enable_norm": true,
|
| 1426 |
-
"enable_perm":
|
| 1427 |
"group_num": 1,
|
| 1428 |
"group_size": 4055,
|
| 1429 |
"in_features": 4096,
|
|
@@ -1447,7 +1447,7 @@
|
|
| 1447 |
"model.layers.16.self_attn.k_proj": {
|
| 1448 |
"bias": false,
|
| 1449 |
"enable_norm": true,
|
| 1450 |
-
"enable_perm":
|
| 1451 |
"group_num": 1,
|
| 1452 |
"group_size": 4055,
|
| 1453 |
"in_features": 4096,
|
|
@@ -1471,7 +1471,7 @@
|
|
| 1471 |
"model.layers.16.self_attn.o_proj": {
|
| 1472 |
"bias": false,
|
| 1473 |
"enable_norm": true,
|
| 1474 |
-
"enable_perm":
|
| 1475 |
"group_num": 1,
|
| 1476 |
"group_size": 4055,
|
| 1477 |
"in_features": 4096,
|
|
@@ -1495,7 +1495,7 @@
|
|
| 1495 |
"model.layers.16.self_attn.q_proj": {
|
| 1496 |
"bias": false,
|
| 1497 |
"enable_norm": true,
|
| 1498 |
-
"enable_perm":
|
| 1499 |
"group_num": 1,
|
| 1500 |
"group_size": 4055,
|
| 1501 |
"in_features": 4096,
|
|
@@ -1519,7 +1519,7 @@
|
|
| 1519 |
"model.layers.16.self_attn.v_proj": {
|
| 1520 |
"bias": false,
|
| 1521 |
"enable_norm": true,
|
| 1522 |
-
"enable_perm":
|
| 1523 |
"group_num": 1,
|
| 1524 |
"group_size": 4055,
|
| 1525 |
"in_features": 4096,
|
|
@@ -1543,7 +1543,7 @@
|
|
| 1543 |
"model.layers.17.mlp.down_proj": {
|
| 1544 |
"bias": false,
|
| 1545 |
"enable_norm": true,
|
| 1546 |
-
"enable_perm":
|
| 1547 |
"group_num": 1,
|
| 1548 |
"group_size": 14192,
|
| 1549 |
"in_features": 14336,
|
|
@@ -1567,7 +1567,7 @@
|
|
| 1567 |
"model.layers.17.mlp.gate_proj": {
|
| 1568 |
"bias": false,
|
| 1569 |
"enable_norm": true,
|
| 1570 |
-
"enable_perm":
|
| 1571 |
"group_num": 1,
|
| 1572 |
"group_size": 4055,
|
| 1573 |
"in_features": 4096,
|
|
@@ -1591,7 +1591,7 @@
|
|
| 1591 |
"model.layers.17.mlp.up_proj": {
|
| 1592 |
"bias": false,
|
| 1593 |
"enable_norm": true,
|
| 1594 |
-
"enable_perm":
|
| 1595 |
"group_num": 1,
|
| 1596 |
"group_size": 4055,
|
| 1597 |
"in_features": 4096,
|
|
@@ -1615,7 +1615,7 @@
|
|
| 1615 |
"model.layers.17.self_attn.k_proj": {
|
| 1616 |
"bias": false,
|
| 1617 |
"enable_norm": true,
|
| 1618 |
-
"enable_perm":
|
| 1619 |
"group_num": 1,
|
| 1620 |
"group_size": 4055,
|
| 1621 |
"in_features": 4096,
|
|
@@ -1639,7 +1639,7 @@
|
|
| 1639 |
"model.layers.17.self_attn.o_proj": {
|
| 1640 |
"bias": false,
|
| 1641 |
"enable_norm": true,
|
| 1642 |
-
"enable_perm":
|
| 1643 |
"group_num": 1,
|
| 1644 |
"group_size": 4055,
|
| 1645 |
"in_features": 4096,
|
|
@@ -1663,7 +1663,7 @@
|
|
| 1663 |
"model.layers.17.self_attn.q_proj": {
|
| 1664 |
"bias": false,
|
| 1665 |
"enable_norm": true,
|
| 1666 |
-
"enable_perm":
|
| 1667 |
"group_num": 1,
|
| 1668 |
"group_size": 4055,
|
| 1669 |
"in_features": 4096,
|
|
@@ -1687,7 +1687,7 @@
|
|
| 1687 |
"model.layers.17.self_attn.v_proj": {
|
| 1688 |
"bias": false,
|
| 1689 |
"enable_norm": true,
|
| 1690 |
-
"enable_perm":
|
| 1691 |
"group_num": 1,
|
| 1692 |
"group_size": 4055,
|
| 1693 |
"in_features": 4096,
|
|
@@ -1711,7 +1711,7 @@
|
|
| 1711 |
"model.layers.18.mlp.down_proj": {
|
| 1712 |
"bias": false,
|
| 1713 |
"enable_norm": true,
|
| 1714 |
-
"enable_perm":
|
| 1715 |
"group_num": 1,
|
| 1716 |
"group_size": 14192,
|
| 1717 |
"in_features": 14336,
|
|
@@ -1735,7 +1735,7 @@
|
|
| 1735 |
"model.layers.18.mlp.gate_proj": {
|
| 1736 |
"bias": false,
|
| 1737 |
"enable_norm": true,
|
| 1738 |
-
"enable_perm":
|
| 1739 |
"group_num": 1,
|
| 1740 |
"group_size": 4055,
|
| 1741 |
"in_features": 4096,
|
|
@@ -1759,7 +1759,7 @@
|
|
| 1759 |
"model.layers.18.mlp.up_proj": {
|
| 1760 |
"bias": false,
|
| 1761 |
"enable_norm": true,
|
| 1762 |
-
"enable_perm":
|
| 1763 |
"group_num": 1,
|
| 1764 |
"group_size": 4055,
|
| 1765 |
"in_features": 4096,
|
|
@@ -1783,7 +1783,7 @@
|
|
| 1783 |
"model.layers.18.self_attn.k_proj": {
|
| 1784 |
"bias": false,
|
| 1785 |
"enable_norm": true,
|
| 1786 |
-
"enable_perm":
|
| 1787 |
"group_num": 1,
|
| 1788 |
"group_size": 4055,
|
| 1789 |
"in_features": 4096,
|
|
@@ -1807,7 +1807,7 @@
|
|
| 1807 |
"model.layers.18.self_attn.o_proj": {
|
| 1808 |
"bias": false,
|
| 1809 |
"enable_norm": true,
|
| 1810 |
-
"enable_perm":
|
| 1811 |
"group_num": 1,
|
| 1812 |
"group_size": 4055,
|
| 1813 |
"in_features": 4096,
|
|
@@ -1831,7 +1831,7 @@
|
|
| 1831 |
"model.layers.18.self_attn.q_proj": {
|
| 1832 |
"bias": false,
|
| 1833 |
"enable_norm": true,
|
| 1834 |
-
"enable_perm":
|
| 1835 |
"group_num": 1,
|
| 1836 |
"group_size": 4055,
|
| 1837 |
"in_features": 4096,
|
|
@@ -1855,7 +1855,7 @@
|
|
| 1855 |
"model.layers.18.self_attn.v_proj": {
|
| 1856 |
"bias": false,
|
| 1857 |
"enable_norm": true,
|
| 1858 |
-
"enable_perm":
|
| 1859 |
"group_num": 1,
|
| 1860 |
"group_size": 4055,
|
| 1861 |
"in_features": 4096,
|
|
@@ -1879,7 +1879,7 @@
|
|
| 1879 |
"model.layers.19.mlp.down_proj": {
|
| 1880 |
"bias": false,
|
| 1881 |
"enable_norm": true,
|
| 1882 |
-
"enable_perm":
|
| 1883 |
"group_num": 1,
|
| 1884 |
"group_size": 14192,
|
| 1885 |
"in_features": 14336,
|
|
@@ -1903,7 +1903,7 @@
|
|
| 1903 |
"model.layers.19.mlp.gate_proj": {
|
| 1904 |
"bias": false,
|
| 1905 |
"enable_norm": true,
|
| 1906 |
-
"enable_perm":
|
| 1907 |
"group_num": 1,
|
| 1908 |
"group_size": 4055,
|
| 1909 |
"in_features": 4096,
|
|
@@ -1927,7 +1927,7 @@
|
|
| 1927 |
"model.layers.19.mlp.up_proj": {
|
| 1928 |
"bias": false,
|
| 1929 |
"enable_norm": true,
|
| 1930 |
-
"enable_perm":
|
| 1931 |
"group_num": 1,
|
| 1932 |
"group_size": 4055,
|
| 1933 |
"in_features": 4096,
|
|
@@ -1951,7 +1951,7 @@
|
|
| 1951 |
"model.layers.19.self_attn.k_proj": {
|
| 1952 |
"bias": false,
|
| 1953 |
"enable_norm": true,
|
| 1954 |
-
"enable_perm":
|
| 1955 |
"group_num": 1,
|
| 1956 |
"group_size": 4055,
|
| 1957 |
"in_features": 4096,
|
|
@@ -1975,7 +1975,7 @@
|
|
| 1975 |
"model.layers.19.self_attn.o_proj": {
|
| 1976 |
"bias": false,
|
| 1977 |
"enable_norm": true,
|
| 1978 |
-
"enable_perm":
|
| 1979 |
"group_num": 1,
|
| 1980 |
"group_size": 4055,
|
| 1981 |
"in_features": 4096,
|
|
@@ -1999,7 +1999,7 @@
|
|
| 1999 |
"model.layers.19.self_attn.q_proj": {
|
| 2000 |
"bias": false,
|
| 2001 |
"enable_norm": true,
|
| 2002 |
-
"enable_perm":
|
| 2003 |
"group_num": 1,
|
| 2004 |
"group_size": 4055,
|
| 2005 |
"in_features": 4096,
|
|
@@ -2023,7 +2023,7 @@
|
|
| 2023 |
"model.layers.19.self_attn.v_proj": {
|
| 2024 |
"bias": false,
|
| 2025 |
"enable_norm": true,
|
| 2026 |
-
"enable_perm":
|
| 2027 |
"group_num": 1,
|
| 2028 |
"group_size": 4055,
|
| 2029 |
"in_features": 4096,
|
|
@@ -2047,7 +2047,7 @@
|
|
| 2047 |
"model.layers.2.mlp.down_proj": {
|
| 2048 |
"bias": false,
|
| 2049 |
"enable_norm": true,
|
| 2050 |
-
"enable_perm":
|
| 2051 |
"group_num": 1,
|
| 2052 |
"group_size": 14192,
|
| 2053 |
"in_features": 14336,
|
|
@@ -2071,7 +2071,7 @@
|
|
| 2071 |
"model.layers.2.mlp.gate_proj": {
|
| 2072 |
"bias": false,
|
| 2073 |
"enable_norm": true,
|
| 2074 |
-
"enable_perm":
|
| 2075 |
"group_num": 1,
|
| 2076 |
"group_size": 4055,
|
| 2077 |
"in_features": 4096,
|
|
@@ -2095,7 +2095,7 @@
|
|
| 2095 |
"model.layers.2.mlp.up_proj": {
|
| 2096 |
"bias": false,
|
| 2097 |
"enable_norm": true,
|
| 2098 |
-
"enable_perm":
|
| 2099 |
"group_num": 1,
|
| 2100 |
"group_size": 4055,
|
| 2101 |
"in_features": 4096,
|
|
@@ -2119,7 +2119,7 @@
|
|
| 2119 |
"model.layers.2.self_attn.k_proj": {
|
| 2120 |
"bias": false,
|
| 2121 |
"enable_norm": true,
|
| 2122 |
-
"enable_perm":
|
| 2123 |
"group_num": 1,
|
| 2124 |
"group_size": 4055,
|
| 2125 |
"in_features": 4096,
|
|
@@ -2143,7 +2143,7 @@
|
|
| 2143 |
"model.layers.2.self_attn.o_proj": {
|
| 2144 |
"bias": false,
|
| 2145 |
"enable_norm": true,
|
| 2146 |
-
"enable_perm":
|
| 2147 |
"group_num": 1,
|
| 2148 |
"group_size": 4055,
|
| 2149 |
"in_features": 4096,
|
|
@@ -2167,7 +2167,7 @@
|
|
| 2167 |
"model.layers.2.self_attn.q_proj": {
|
| 2168 |
"bias": false,
|
| 2169 |
"enable_norm": true,
|
| 2170 |
-
"enable_perm":
|
| 2171 |
"group_num": 1,
|
| 2172 |
"group_size": 4055,
|
| 2173 |
"in_features": 4096,
|
|
@@ -2191,7 +2191,7 @@
|
|
| 2191 |
"model.layers.2.self_attn.v_proj": {
|
| 2192 |
"bias": false,
|
| 2193 |
"enable_norm": true,
|
| 2194 |
-
"enable_perm":
|
| 2195 |
"group_num": 1,
|
| 2196 |
"group_size": 4055,
|
| 2197 |
"in_features": 4096,
|
|
@@ -2215,7 +2215,7 @@
|
|
| 2215 |
"model.layers.20.mlp.down_proj": {
|
| 2216 |
"bias": false,
|
| 2217 |
"enable_norm": true,
|
| 2218 |
-
"enable_perm":
|
| 2219 |
"group_num": 1,
|
| 2220 |
"group_size": 14192,
|
| 2221 |
"in_features": 14336,
|
|
@@ -2239,7 +2239,7 @@
|
|
| 2239 |
"model.layers.20.mlp.gate_proj": {
|
| 2240 |
"bias": false,
|
| 2241 |
"enable_norm": true,
|
| 2242 |
-
"enable_perm":
|
| 2243 |
"group_num": 1,
|
| 2244 |
"group_size": 4055,
|
| 2245 |
"in_features": 4096,
|
|
@@ -2263,7 +2263,7 @@
|
|
| 2263 |
"model.layers.20.mlp.up_proj": {
|
| 2264 |
"bias": false,
|
| 2265 |
"enable_norm": true,
|
| 2266 |
-
"enable_perm":
|
| 2267 |
"group_num": 1,
|
| 2268 |
"group_size": 4055,
|
| 2269 |
"in_features": 4096,
|
|
@@ -2287,7 +2287,7 @@
|
|
| 2287 |
"model.layers.20.self_attn.k_proj": {
|
| 2288 |
"bias": false,
|
| 2289 |
"enable_norm": true,
|
| 2290 |
-
"enable_perm":
|
| 2291 |
"group_num": 1,
|
| 2292 |
"group_size": 4055,
|
| 2293 |
"in_features": 4096,
|
|
@@ -2311,7 +2311,7 @@
|
|
| 2311 |
"model.layers.20.self_attn.o_proj": {
|
| 2312 |
"bias": false,
|
| 2313 |
"enable_norm": true,
|
| 2314 |
-
"enable_perm":
|
| 2315 |
"group_num": 1,
|
| 2316 |
"group_size": 4055,
|
| 2317 |
"in_features": 4096,
|
|
@@ -2335,7 +2335,7 @@
|
|
| 2335 |
"model.layers.20.self_attn.q_proj": {
|
| 2336 |
"bias": false,
|
| 2337 |
"enable_norm": true,
|
| 2338 |
-
"enable_perm":
|
| 2339 |
"group_num": 1,
|
| 2340 |
"group_size": 4055,
|
| 2341 |
"in_features": 4096,
|
|
@@ -2359,7 +2359,7 @@
|
|
| 2359 |
"model.layers.20.self_attn.v_proj": {
|
| 2360 |
"bias": false,
|
| 2361 |
"enable_norm": true,
|
| 2362 |
-
"enable_perm":
|
| 2363 |
"group_num": 1,
|
| 2364 |
"group_size": 4055,
|
| 2365 |
"in_features": 4096,
|
|
@@ -2383,7 +2383,7 @@
|
|
| 2383 |
"model.layers.21.mlp.down_proj": {
|
| 2384 |
"bias": false,
|
| 2385 |
"enable_norm": true,
|
| 2386 |
-
"enable_perm":
|
| 2387 |
"group_num": 1,
|
| 2388 |
"group_size": 14192,
|
| 2389 |
"in_features": 14336,
|
|
@@ -2407,7 +2407,7 @@
|
|
| 2407 |
"model.layers.21.mlp.gate_proj": {
|
| 2408 |
"bias": false,
|
| 2409 |
"enable_norm": true,
|
| 2410 |
-
"enable_perm":
|
| 2411 |
"group_num": 1,
|
| 2412 |
"group_size": 4055,
|
| 2413 |
"in_features": 4096,
|
|
@@ -2431,7 +2431,7 @@
|
|
| 2431 |
"model.layers.21.mlp.up_proj": {
|
| 2432 |
"bias": false,
|
| 2433 |
"enable_norm": true,
|
| 2434 |
-
"enable_perm":
|
| 2435 |
"group_num": 1,
|
| 2436 |
"group_size": 4055,
|
| 2437 |
"in_features": 4096,
|
|
@@ -2455,7 +2455,7 @@
|
|
| 2455 |
"model.layers.21.self_attn.k_proj": {
|
| 2456 |
"bias": false,
|
| 2457 |
"enable_norm": true,
|
| 2458 |
-
"enable_perm":
|
| 2459 |
"group_num": 1,
|
| 2460 |
"group_size": 4055,
|
| 2461 |
"in_features": 4096,
|
|
@@ -2479,7 +2479,7 @@
|
|
| 2479 |
"model.layers.21.self_attn.o_proj": {
|
| 2480 |
"bias": false,
|
| 2481 |
"enable_norm": true,
|
| 2482 |
-
"enable_perm":
|
| 2483 |
"group_num": 1,
|
| 2484 |
"group_size": 4055,
|
| 2485 |
"in_features": 4096,
|
|
@@ -2503,7 +2503,7 @@
|
|
| 2503 |
"model.layers.21.self_attn.q_proj": {
|
| 2504 |
"bias": false,
|
| 2505 |
"enable_norm": true,
|
| 2506 |
-
"enable_perm":
|
| 2507 |
"group_num": 1,
|
| 2508 |
"group_size": 4055,
|
| 2509 |
"in_features": 4096,
|
|
@@ -2527,7 +2527,7 @@
|
|
| 2527 |
"model.layers.21.self_attn.v_proj": {
|
| 2528 |
"bias": false,
|
| 2529 |
"enable_norm": true,
|
| 2530 |
-
"enable_perm":
|
| 2531 |
"group_num": 1,
|
| 2532 |
"group_size": 4055,
|
| 2533 |
"in_features": 4096,
|
|
@@ -2551,7 +2551,7 @@
|
|
| 2551 |
"model.layers.22.mlp.down_proj": {
|
| 2552 |
"bias": false,
|
| 2553 |
"enable_norm": true,
|
| 2554 |
-
"enable_perm":
|
| 2555 |
"group_num": 1,
|
| 2556 |
"group_size": 14192,
|
| 2557 |
"in_features": 14336,
|
|
@@ -2575,7 +2575,7 @@
|
|
| 2575 |
"model.layers.22.mlp.gate_proj": {
|
| 2576 |
"bias": false,
|
| 2577 |
"enable_norm": true,
|
| 2578 |
-
"enable_perm":
|
| 2579 |
"group_num": 1,
|
| 2580 |
"group_size": 4055,
|
| 2581 |
"in_features": 4096,
|
|
@@ -2599,7 +2599,7 @@
|
|
| 2599 |
"model.layers.22.mlp.up_proj": {
|
| 2600 |
"bias": false,
|
| 2601 |
"enable_norm": true,
|
| 2602 |
-
"enable_perm":
|
| 2603 |
"group_num": 1,
|
| 2604 |
"group_size": 4055,
|
| 2605 |
"in_features": 4096,
|
|
@@ -2623,7 +2623,7 @@
|
|
| 2623 |
"model.layers.22.self_attn.k_proj": {
|
| 2624 |
"bias": false,
|
| 2625 |
"enable_norm": true,
|
| 2626 |
-
"enable_perm":
|
| 2627 |
"group_num": 1,
|
| 2628 |
"group_size": 4055,
|
| 2629 |
"in_features": 4096,
|
|
@@ -2647,7 +2647,7 @@
|
|
| 2647 |
"model.layers.22.self_attn.o_proj": {
|
| 2648 |
"bias": false,
|
| 2649 |
"enable_norm": true,
|
| 2650 |
-
"enable_perm":
|
| 2651 |
"group_num": 1,
|
| 2652 |
"group_size": 4055,
|
| 2653 |
"in_features": 4096,
|
|
@@ -2671,7 +2671,7 @@
|
|
| 2671 |
"model.layers.22.self_attn.q_proj": {
|
| 2672 |
"bias": false,
|
| 2673 |
"enable_norm": true,
|
| 2674 |
-
"enable_perm":
|
| 2675 |
"group_num": 1,
|
| 2676 |
"group_size": 4055,
|
| 2677 |
"in_features": 4096,
|
|
@@ -2695,7 +2695,7 @@
|
|
| 2695 |
"model.layers.22.self_attn.v_proj": {
|
| 2696 |
"bias": false,
|
| 2697 |
"enable_norm": true,
|
| 2698 |
-
"enable_perm":
|
| 2699 |
"group_num": 1,
|
| 2700 |
"group_size": 4055,
|
| 2701 |
"in_features": 4096,
|
|
@@ -2719,7 +2719,7 @@
|
|
| 2719 |
"model.layers.23.mlp.down_proj": {
|
| 2720 |
"bias": false,
|
| 2721 |
"enable_norm": true,
|
| 2722 |
-
"enable_perm":
|
| 2723 |
"group_num": 1,
|
| 2724 |
"group_size": 14192,
|
| 2725 |
"in_features": 14336,
|
|
@@ -2743,7 +2743,7 @@
|
|
| 2743 |
"model.layers.23.mlp.gate_proj": {
|
| 2744 |
"bias": false,
|
| 2745 |
"enable_norm": true,
|
| 2746 |
-
"enable_perm":
|
| 2747 |
"group_num": 1,
|
| 2748 |
"group_size": 4055,
|
| 2749 |
"in_features": 4096,
|
|
@@ -2767,7 +2767,7 @@
|
|
| 2767 |
"model.layers.23.mlp.up_proj": {
|
| 2768 |
"bias": false,
|
| 2769 |
"enable_norm": true,
|
| 2770 |
-
"enable_perm":
|
| 2771 |
"group_num": 1,
|
| 2772 |
"group_size": 4055,
|
| 2773 |
"in_features": 4096,
|
|
@@ -2791,7 +2791,7 @@
|
|
| 2791 |
"model.layers.23.self_attn.k_proj": {
|
| 2792 |
"bias": false,
|
| 2793 |
"enable_norm": true,
|
| 2794 |
-
"enable_perm":
|
| 2795 |
"group_num": 1,
|
| 2796 |
"group_size": 4055,
|
| 2797 |
"in_features": 4096,
|
|
@@ -2815,7 +2815,7 @@
|
|
| 2815 |
"model.layers.23.self_attn.o_proj": {
|
| 2816 |
"bias": false,
|
| 2817 |
"enable_norm": true,
|
| 2818 |
-
"enable_perm":
|
| 2819 |
"group_num": 1,
|
| 2820 |
"group_size": 4055,
|
| 2821 |
"in_features": 4096,
|
|
@@ -2839,7 +2839,7 @@
|
|
| 2839 |
"model.layers.23.self_attn.q_proj": {
|
| 2840 |
"bias": false,
|
| 2841 |
"enable_norm": true,
|
| 2842 |
-
"enable_perm":
|
| 2843 |
"group_num": 1,
|
| 2844 |
"group_size": 4055,
|
| 2845 |
"in_features": 4096,
|
|
@@ -2863,7 +2863,7 @@
|
|
| 2863 |
"model.layers.23.self_attn.v_proj": {
|
| 2864 |
"bias": false,
|
| 2865 |
"enable_norm": true,
|
| 2866 |
-
"enable_perm":
|
| 2867 |
"group_num": 1,
|
| 2868 |
"group_size": 4055,
|
| 2869 |
"in_features": 4096,
|
|
@@ -2887,7 +2887,7 @@
|
|
| 2887 |
"model.layers.24.mlp.down_proj": {
|
| 2888 |
"bias": false,
|
| 2889 |
"enable_norm": true,
|
| 2890 |
-
"enable_perm":
|
| 2891 |
"group_num": 1,
|
| 2892 |
"group_size": 14192,
|
| 2893 |
"in_features": 14336,
|
|
@@ -2911,7 +2911,7 @@
|
|
| 2911 |
"model.layers.24.mlp.gate_proj": {
|
| 2912 |
"bias": false,
|
| 2913 |
"enable_norm": true,
|
| 2914 |
-
"enable_perm":
|
| 2915 |
"group_num": 1,
|
| 2916 |
"group_size": 4055,
|
| 2917 |
"in_features": 4096,
|
|
@@ -2935,7 +2935,7 @@
|
|
| 2935 |
"model.layers.24.mlp.up_proj": {
|
| 2936 |
"bias": false,
|
| 2937 |
"enable_norm": true,
|
| 2938 |
-
"enable_perm":
|
| 2939 |
"group_num": 1,
|
| 2940 |
"group_size": 4055,
|
| 2941 |
"in_features": 4096,
|
|
@@ -2959,7 +2959,7 @@
|
|
| 2959 |
"model.layers.24.self_attn.k_proj": {
|
| 2960 |
"bias": false,
|
| 2961 |
"enable_norm": true,
|
| 2962 |
-
"enable_perm":
|
| 2963 |
"group_num": 1,
|
| 2964 |
"group_size": 4055,
|
| 2965 |
"in_features": 4096,
|
|
@@ -2983,7 +2983,7 @@
|
|
| 2983 |
"model.layers.24.self_attn.o_proj": {
|
| 2984 |
"bias": false,
|
| 2985 |
"enable_norm": true,
|
| 2986 |
-
"enable_perm":
|
| 2987 |
"group_num": 1,
|
| 2988 |
"group_size": 4055,
|
| 2989 |
"in_features": 4096,
|
|
@@ -3007,7 +3007,7 @@
|
|
| 3007 |
"model.layers.24.self_attn.q_proj": {
|
| 3008 |
"bias": false,
|
| 3009 |
"enable_norm": true,
|
| 3010 |
-
"enable_perm":
|
| 3011 |
"group_num": 1,
|
| 3012 |
"group_size": 4055,
|
| 3013 |
"in_features": 4096,
|
|
@@ -3031,7 +3031,7 @@
|
|
| 3031 |
"model.layers.24.self_attn.v_proj": {
|
| 3032 |
"bias": false,
|
| 3033 |
"enable_norm": true,
|
| 3034 |
-
"enable_perm":
|
| 3035 |
"group_num": 1,
|
| 3036 |
"group_size": 4055,
|
| 3037 |
"in_features": 4096,
|
|
@@ -3055,7 +3055,7 @@
|
|
| 3055 |
"model.layers.25.mlp.down_proj": {
|
| 3056 |
"bias": false,
|
| 3057 |
"enable_norm": true,
|
| 3058 |
-
"enable_perm":
|
| 3059 |
"group_num": 1,
|
| 3060 |
"group_size": 14192,
|
| 3061 |
"in_features": 14336,
|
|
@@ -3079,7 +3079,7 @@
|
|
| 3079 |
"model.layers.25.mlp.gate_proj": {
|
| 3080 |
"bias": false,
|
| 3081 |
"enable_norm": true,
|
| 3082 |
-
"enable_perm":
|
| 3083 |
"group_num": 1,
|
| 3084 |
"group_size": 4055,
|
| 3085 |
"in_features": 4096,
|
|
@@ -3103,7 +3103,7 @@
|
|
| 3103 |
"model.layers.25.mlp.up_proj": {
|
| 3104 |
"bias": false,
|
| 3105 |
"enable_norm": true,
|
| 3106 |
-
"enable_perm":
|
| 3107 |
"group_num": 1,
|
| 3108 |
"group_size": 4055,
|
| 3109 |
"in_features": 4096,
|
|
@@ -3127,7 +3127,7 @@
|
|
| 3127 |
"model.layers.25.self_attn.k_proj": {
|
| 3128 |
"bias": false,
|
| 3129 |
"enable_norm": true,
|
| 3130 |
-
"enable_perm":
|
| 3131 |
"group_num": 1,
|
| 3132 |
"group_size": 4055,
|
| 3133 |
"in_features": 4096,
|
|
@@ -3151,7 +3151,7 @@
|
|
| 3151 |
"model.layers.25.self_attn.o_proj": {
|
| 3152 |
"bias": false,
|
| 3153 |
"enable_norm": true,
|
| 3154 |
-
"enable_perm":
|
| 3155 |
"group_num": 1,
|
| 3156 |
"group_size": 4055,
|
| 3157 |
"in_features": 4096,
|
|
@@ -3175,7 +3175,7 @@
|
|
| 3175 |
"model.layers.25.self_attn.q_proj": {
|
| 3176 |
"bias": false,
|
| 3177 |
"enable_norm": true,
|
| 3178 |
-
"enable_perm":
|
| 3179 |
"group_num": 1,
|
| 3180 |
"group_size": 4055,
|
| 3181 |
"in_features": 4096,
|
|
@@ -3199,7 +3199,7 @@
|
|
| 3199 |
"model.layers.25.self_attn.v_proj": {
|
| 3200 |
"bias": false,
|
| 3201 |
"enable_norm": true,
|
| 3202 |
-
"enable_perm":
|
| 3203 |
"group_num": 1,
|
| 3204 |
"group_size": 4055,
|
| 3205 |
"in_features": 4096,
|
|
@@ -3223,7 +3223,7 @@
|
|
| 3223 |
"model.layers.26.mlp.down_proj": {
|
| 3224 |
"bias": false,
|
| 3225 |
"enable_norm": true,
|
| 3226 |
-
"enable_perm":
|
| 3227 |
"group_num": 1,
|
| 3228 |
"group_size": 14192,
|
| 3229 |
"in_features": 14336,
|
|
@@ -3247,7 +3247,7 @@
|
|
| 3247 |
"model.layers.26.mlp.gate_proj": {
|
| 3248 |
"bias": false,
|
| 3249 |
"enable_norm": true,
|
| 3250 |
-
"enable_perm":
|
| 3251 |
"group_num": 1,
|
| 3252 |
"group_size": 4055,
|
| 3253 |
"in_features": 4096,
|
|
@@ -3271,7 +3271,7 @@
|
|
| 3271 |
"model.layers.26.mlp.up_proj": {
|
| 3272 |
"bias": false,
|
| 3273 |
"enable_norm": true,
|
| 3274 |
-
"enable_perm":
|
| 3275 |
"group_num": 1,
|
| 3276 |
"group_size": 4055,
|
| 3277 |
"in_features": 4096,
|
|
@@ -3295,7 +3295,7 @@
|
|
| 3295 |
"model.layers.26.self_attn.k_proj": {
|
| 3296 |
"bias": false,
|
| 3297 |
"enable_norm": true,
|
| 3298 |
-
"enable_perm":
|
| 3299 |
"group_num": 1,
|
| 3300 |
"group_size": 4055,
|
| 3301 |
"in_features": 4096,
|
|
@@ -3319,7 +3319,7 @@
|
|
| 3319 |
"model.layers.26.self_attn.o_proj": {
|
| 3320 |
"bias": false,
|
| 3321 |
"enable_norm": true,
|
| 3322 |
-
"enable_perm":
|
| 3323 |
"group_num": 1,
|
| 3324 |
"group_size": 4055,
|
| 3325 |
"in_features": 4096,
|
|
@@ -3343,7 +3343,7 @@
|
|
| 3343 |
"model.layers.26.self_attn.q_proj": {
|
| 3344 |
"bias": false,
|
| 3345 |
"enable_norm": true,
|
| 3346 |
-
"enable_perm":
|
| 3347 |
"group_num": 1,
|
| 3348 |
"group_size": 4055,
|
| 3349 |
"in_features": 4096,
|
|
@@ -3367,7 +3367,7 @@
|
|
| 3367 |
"model.layers.26.self_attn.v_proj": {
|
| 3368 |
"bias": false,
|
| 3369 |
"enable_norm": true,
|
| 3370 |
-
"enable_perm":
|
| 3371 |
"group_num": 1,
|
| 3372 |
"group_size": 4055,
|
| 3373 |
"in_features": 4096,
|
|
@@ -3391,7 +3391,7 @@
|
|
| 3391 |
"model.layers.27.mlp.down_proj": {
|
| 3392 |
"bias": false,
|
| 3393 |
"enable_norm": true,
|
| 3394 |
-
"enable_perm":
|
| 3395 |
"group_num": 1,
|
| 3396 |
"group_size": 14192,
|
| 3397 |
"in_features": 14336,
|
|
@@ -3415,7 +3415,7 @@
|
|
| 3415 |
"model.layers.27.mlp.gate_proj": {
|
| 3416 |
"bias": false,
|
| 3417 |
"enable_norm": true,
|
| 3418 |
-
"enable_perm":
|
| 3419 |
"group_num": 1,
|
| 3420 |
"group_size": 4055,
|
| 3421 |
"in_features": 4096,
|
|
@@ -3439,7 +3439,7 @@
|
|
| 3439 |
"model.layers.27.mlp.up_proj": {
|
| 3440 |
"bias": false,
|
| 3441 |
"enable_norm": true,
|
| 3442 |
-
"enable_perm":
|
| 3443 |
"group_num": 1,
|
| 3444 |
"group_size": 4055,
|
| 3445 |
"in_features": 4096,
|
|
@@ -3463,7 +3463,7 @@
|
|
| 3463 |
"model.layers.27.self_attn.k_proj": {
|
| 3464 |
"bias": false,
|
| 3465 |
"enable_norm": true,
|
| 3466 |
-
"enable_perm":
|
| 3467 |
"group_num": 1,
|
| 3468 |
"group_size": 4055,
|
| 3469 |
"in_features": 4096,
|
|
@@ -3487,7 +3487,7 @@
|
|
| 3487 |
"model.layers.27.self_attn.o_proj": {
|
| 3488 |
"bias": false,
|
| 3489 |
"enable_norm": true,
|
| 3490 |
-
"enable_perm":
|
| 3491 |
"group_num": 1,
|
| 3492 |
"group_size": 4055,
|
| 3493 |
"in_features": 4096,
|
|
@@ -3511,7 +3511,7 @@
|
|
| 3511 |
"model.layers.27.self_attn.q_proj": {
|
| 3512 |
"bias": false,
|
| 3513 |
"enable_norm": true,
|
| 3514 |
-
"enable_perm":
|
| 3515 |
"group_num": 1,
|
| 3516 |
"group_size": 4055,
|
| 3517 |
"in_features": 4096,
|
|
@@ -3535,7 +3535,7 @@
|
|
| 3535 |
"model.layers.27.self_attn.v_proj": {
|
| 3536 |
"bias": false,
|
| 3537 |
"enable_norm": true,
|
| 3538 |
-
"enable_perm":
|
| 3539 |
"group_num": 1,
|
| 3540 |
"group_size": 4055,
|
| 3541 |
"in_features": 4096,
|
|
@@ -3559,7 +3559,7 @@
|
|
| 3559 |
"model.layers.28.mlp.down_proj": {
|
| 3560 |
"bias": false,
|
| 3561 |
"enable_norm": true,
|
| 3562 |
-
"enable_perm":
|
| 3563 |
"group_num": 1,
|
| 3564 |
"group_size": 14192,
|
| 3565 |
"in_features": 14336,
|
|
@@ -3583,7 +3583,7 @@
|
|
| 3583 |
"model.layers.28.mlp.gate_proj": {
|
| 3584 |
"bias": false,
|
| 3585 |
"enable_norm": true,
|
| 3586 |
-
"enable_perm":
|
| 3587 |
"group_num": 1,
|
| 3588 |
"group_size": 4055,
|
| 3589 |
"in_features": 4096,
|
|
@@ -3607,7 +3607,7 @@
|
|
| 3607 |
"model.layers.28.mlp.up_proj": {
|
| 3608 |
"bias": false,
|
| 3609 |
"enable_norm": true,
|
| 3610 |
-
"enable_perm":
|
| 3611 |
"group_num": 1,
|
| 3612 |
"group_size": 4055,
|
| 3613 |
"in_features": 4096,
|
|
@@ -3631,7 +3631,7 @@
|
|
| 3631 |
"model.layers.28.self_attn.k_proj": {
|
| 3632 |
"bias": false,
|
| 3633 |
"enable_norm": true,
|
| 3634 |
-
"enable_perm":
|
| 3635 |
"group_num": 1,
|
| 3636 |
"group_size": 4055,
|
| 3637 |
"in_features": 4096,
|
|
@@ -3655,7 +3655,7 @@
|
|
| 3655 |
"model.layers.28.self_attn.o_proj": {
|
| 3656 |
"bias": false,
|
| 3657 |
"enable_norm": true,
|
| 3658 |
-
"enable_perm":
|
| 3659 |
"group_num": 1,
|
| 3660 |
"group_size": 4055,
|
| 3661 |
"in_features": 4096,
|
|
@@ -3679,7 +3679,7 @@
|
|
| 3679 |
"model.layers.28.self_attn.q_proj": {
|
| 3680 |
"bias": false,
|
| 3681 |
"enable_norm": true,
|
| 3682 |
-
"enable_perm":
|
| 3683 |
"group_num": 1,
|
| 3684 |
"group_size": 4055,
|
| 3685 |
"in_features": 4096,
|
|
@@ -3703,7 +3703,7 @@
|
|
| 3703 |
"model.layers.28.self_attn.v_proj": {
|
| 3704 |
"bias": false,
|
| 3705 |
"enable_norm": true,
|
| 3706 |
-
"enable_perm":
|
| 3707 |
"group_num": 1,
|
| 3708 |
"group_size": 4055,
|
| 3709 |
"in_features": 4096,
|
|
@@ -3727,7 +3727,7 @@
|
|
| 3727 |
"model.layers.29.mlp.down_proj": {
|
| 3728 |
"bias": false,
|
| 3729 |
"enable_norm": true,
|
| 3730 |
-
"enable_perm":
|
| 3731 |
"group_num": 1,
|
| 3732 |
"group_size": 14192,
|
| 3733 |
"in_features": 14336,
|
|
@@ -3751,7 +3751,7 @@
|
|
| 3751 |
"model.layers.29.mlp.gate_proj": {
|
| 3752 |
"bias": false,
|
| 3753 |
"enable_norm": true,
|
| 3754 |
-
"enable_perm":
|
| 3755 |
"group_num": 1,
|
| 3756 |
"group_size": 4055,
|
| 3757 |
"in_features": 4096,
|
|
@@ -3775,7 +3775,7 @@
|
|
| 3775 |
"model.layers.29.mlp.up_proj": {
|
| 3776 |
"bias": false,
|
| 3777 |
"enable_norm": true,
|
| 3778 |
-
"enable_perm":
|
| 3779 |
"group_num": 1,
|
| 3780 |
"group_size": 4055,
|
| 3781 |
"in_features": 4096,
|
|
@@ -3799,7 +3799,7 @@
|
|
| 3799 |
"model.layers.29.self_attn.k_proj": {
|
| 3800 |
"bias": false,
|
| 3801 |
"enable_norm": true,
|
| 3802 |
-
"enable_perm":
|
| 3803 |
"group_num": 1,
|
| 3804 |
"group_size": 4055,
|
| 3805 |
"in_features": 4096,
|
|
@@ -3823,7 +3823,7 @@
|
|
| 3823 |
"model.layers.29.self_attn.o_proj": {
|
| 3824 |
"bias": false,
|
| 3825 |
"enable_norm": true,
|
| 3826 |
-
"enable_perm":
|
| 3827 |
"group_num": 1,
|
| 3828 |
"group_size": 4055,
|
| 3829 |
"in_features": 4096,
|
|
@@ -3847,7 +3847,7 @@
|
|
| 3847 |
"model.layers.29.self_attn.q_proj": {
|
| 3848 |
"bias": false,
|
| 3849 |
"enable_norm": true,
|
| 3850 |
-
"enable_perm":
|
| 3851 |
"group_num": 1,
|
| 3852 |
"group_size": 4055,
|
| 3853 |
"in_features": 4096,
|
|
@@ -3871,7 +3871,7 @@
|
|
| 3871 |
"model.layers.29.self_attn.v_proj": {
|
| 3872 |
"bias": false,
|
| 3873 |
"enable_norm": true,
|
| 3874 |
-
"enable_perm":
|
| 3875 |
"group_num": 1,
|
| 3876 |
"group_size": 4055,
|
| 3877 |
"in_features": 4096,
|
|
@@ -3895,7 +3895,7 @@
|
|
| 3895 |
"model.layers.3.mlp.down_proj": {
|
| 3896 |
"bias": false,
|
| 3897 |
"enable_norm": true,
|
| 3898 |
-
"enable_perm":
|
| 3899 |
"group_num": 1,
|
| 3900 |
"group_size": 14192,
|
| 3901 |
"in_features": 14336,
|
|
@@ -3919,7 +3919,7 @@
|
|
| 3919 |
"model.layers.3.mlp.gate_proj": {
|
| 3920 |
"bias": false,
|
| 3921 |
"enable_norm": true,
|
| 3922 |
-
"enable_perm":
|
| 3923 |
"group_num": 1,
|
| 3924 |
"group_size": 4055,
|
| 3925 |
"in_features": 4096,
|
|
@@ -3943,7 +3943,7 @@
|
|
| 3943 |
"model.layers.3.mlp.up_proj": {
|
| 3944 |
"bias": false,
|
| 3945 |
"enable_norm": true,
|
| 3946 |
-
"enable_perm":
|
| 3947 |
"group_num": 1,
|
| 3948 |
"group_size": 4055,
|
| 3949 |
"in_features": 4096,
|
|
@@ -3967,7 +3967,7 @@
|
|
| 3967 |
"model.layers.3.self_attn.k_proj": {
|
| 3968 |
"bias": false,
|
| 3969 |
"enable_norm": true,
|
| 3970 |
-
"enable_perm":
|
| 3971 |
"group_num": 1,
|
| 3972 |
"group_size": 4055,
|
| 3973 |
"in_features": 4096,
|
|
@@ -3991,7 +3991,7 @@
|
|
| 3991 |
"model.layers.3.self_attn.o_proj": {
|
| 3992 |
"bias": false,
|
| 3993 |
"enable_norm": true,
|
| 3994 |
-
"enable_perm":
|
| 3995 |
"group_num": 1,
|
| 3996 |
"group_size": 4055,
|
| 3997 |
"in_features": 4096,
|
|
@@ -4015,7 +4015,7 @@
|
|
| 4015 |
"model.layers.3.self_attn.q_proj": {
|
| 4016 |
"bias": false,
|
| 4017 |
"enable_norm": true,
|
| 4018 |
-
"enable_perm":
|
| 4019 |
"group_num": 1,
|
| 4020 |
"group_size": 4055,
|
| 4021 |
"in_features": 4096,
|
|
@@ -4039,7 +4039,7 @@
|
|
| 4039 |
"model.layers.3.self_attn.v_proj": {
|
| 4040 |
"bias": false,
|
| 4041 |
"enable_norm": true,
|
| 4042 |
-
"enable_perm":
|
| 4043 |
"group_num": 1,
|
| 4044 |
"group_size": 4055,
|
| 4045 |
"in_features": 4096,
|
|
@@ -4063,7 +4063,7 @@
|
|
| 4063 |
"model.layers.30.mlp.down_proj": {
|
| 4064 |
"bias": false,
|
| 4065 |
"enable_norm": true,
|
| 4066 |
-
"enable_perm":
|
| 4067 |
"group_num": 1,
|
| 4068 |
"group_size": 14192,
|
| 4069 |
"in_features": 14336,
|
|
@@ -4087,7 +4087,7 @@
|
|
| 4087 |
"model.layers.30.mlp.gate_proj": {
|
| 4088 |
"bias": false,
|
| 4089 |
"enable_norm": true,
|
| 4090 |
-
"enable_perm":
|
| 4091 |
"group_num": 1,
|
| 4092 |
"group_size": 4055,
|
| 4093 |
"in_features": 4096,
|
|
@@ -4111,7 +4111,7 @@
|
|
| 4111 |
"model.layers.30.mlp.up_proj": {
|
| 4112 |
"bias": false,
|
| 4113 |
"enable_norm": true,
|
| 4114 |
-
"enable_perm":
|
| 4115 |
"group_num": 1,
|
| 4116 |
"group_size": 4055,
|
| 4117 |
"in_features": 4096,
|
|
@@ -4135,7 +4135,7 @@
|
|
| 4135 |
"model.layers.30.self_attn.k_proj": {
|
| 4136 |
"bias": false,
|
| 4137 |
"enable_norm": true,
|
| 4138 |
-
"enable_perm":
|
| 4139 |
"group_num": 1,
|
| 4140 |
"group_size": 4055,
|
| 4141 |
"in_features": 4096,
|
|
@@ -4159,7 +4159,7 @@
|
|
| 4159 |
"model.layers.30.self_attn.o_proj": {
|
| 4160 |
"bias": false,
|
| 4161 |
"enable_norm": true,
|
| 4162 |
-
"enable_perm":
|
| 4163 |
"group_num": 1,
|
| 4164 |
"group_size": 4055,
|
| 4165 |
"in_features": 4096,
|
|
@@ -4183,7 +4183,7 @@
|
|
| 4183 |
"model.layers.30.self_attn.q_proj": {
|
| 4184 |
"bias": false,
|
| 4185 |
"enable_norm": true,
|
| 4186 |
-
"enable_perm":
|
| 4187 |
"group_num": 1,
|
| 4188 |
"group_size": 4055,
|
| 4189 |
"in_features": 4096,
|
|
@@ -4207,7 +4207,7 @@
|
|
| 4207 |
"model.layers.30.self_attn.v_proj": {
|
| 4208 |
"bias": false,
|
| 4209 |
"enable_norm": true,
|
| 4210 |
-
"enable_perm":
|
| 4211 |
"group_num": 1,
|
| 4212 |
"group_size": 4055,
|
| 4213 |
"in_features": 4096,
|
|
@@ -4231,7 +4231,7 @@
|
|
| 4231 |
"model.layers.31.mlp.down_proj": {
|
| 4232 |
"bias": false,
|
| 4233 |
"enable_norm": true,
|
| 4234 |
-
"enable_perm":
|
| 4235 |
"group_num": 1,
|
| 4236 |
"group_size": 14192,
|
| 4237 |
"in_features": 14336,
|
|
@@ -4255,7 +4255,7 @@
|
|
| 4255 |
"model.layers.31.mlp.gate_proj": {
|
| 4256 |
"bias": false,
|
| 4257 |
"enable_norm": true,
|
| 4258 |
-
"enable_perm":
|
| 4259 |
"group_num": 1,
|
| 4260 |
"group_size": 4055,
|
| 4261 |
"in_features": 4096,
|
|
@@ -4279,7 +4279,7 @@
|
|
| 4279 |
"model.layers.31.mlp.up_proj": {
|
| 4280 |
"bias": false,
|
| 4281 |
"enable_norm": true,
|
| 4282 |
-
"enable_perm":
|
| 4283 |
"group_num": 1,
|
| 4284 |
"group_size": 4055,
|
| 4285 |
"in_features": 4096,
|
|
@@ -4303,7 +4303,7 @@
|
|
| 4303 |
"model.layers.31.self_attn.k_proj": {
|
| 4304 |
"bias": false,
|
| 4305 |
"enable_norm": true,
|
| 4306 |
-
"enable_perm":
|
| 4307 |
"group_num": 1,
|
| 4308 |
"group_size": 4055,
|
| 4309 |
"in_features": 4096,
|
|
@@ -4327,7 +4327,7 @@
|
|
| 4327 |
"model.layers.31.self_attn.o_proj": {
|
| 4328 |
"bias": false,
|
| 4329 |
"enable_norm": true,
|
| 4330 |
-
"enable_perm":
|
| 4331 |
"group_num": 1,
|
| 4332 |
"group_size": 4055,
|
| 4333 |
"in_features": 4096,
|
|
@@ -4351,7 +4351,7 @@
|
|
| 4351 |
"model.layers.31.self_attn.q_proj": {
|
| 4352 |
"bias": false,
|
| 4353 |
"enable_norm": true,
|
| 4354 |
-
"enable_perm":
|
| 4355 |
"group_num": 1,
|
| 4356 |
"group_size": 4055,
|
| 4357 |
"in_features": 4096,
|
|
@@ -4375,7 +4375,7 @@
|
|
| 4375 |
"model.layers.31.self_attn.v_proj": {
|
| 4376 |
"bias": false,
|
| 4377 |
"enable_norm": true,
|
| 4378 |
-
"enable_perm":
|
| 4379 |
"group_num": 1,
|
| 4380 |
"group_size": 4055,
|
| 4381 |
"in_features": 4096,
|
|
@@ -4399,7 +4399,7 @@
|
|
| 4399 |
"model.layers.4.mlp.down_proj": {
|
| 4400 |
"bias": false,
|
| 4401 |
"enable_norm": true,
|
| 4402 |
-
"enable_perm":
|
| 4403 |
"group_num": 1,
|
| 4404 |
"group_size": 14192,
|
| 4405 |
"in_features": 14336,
|
|
@@ -4423,7 +4423,7 @@
|
|
| 4423 |
"model.layers.4.mlp.gate_proj": {
|
| 4424 |
"bias": false,
|
| 4425 |
"enable_norm": true,
|
| 4426 |
-
"enable_perm":
|
| 4427 |
"group_num": 1,
|
| 4428 |
"group_size": 4055,
|
| 4429 |
"in_features": 4096,
|
|
@@ -4447,7 +4447,7 @@
|
|
| 4447 |
"model.layers.4.mlp.up_proj": {
|
| 4448 |
"bias": false,
|
| 4449 |
"enable_norm": true,
|
| 4450 |
-
"enable_perm":
|
| 4451 |
"group_num": 1,
|
| 4452 |
"group_size": 4055,
|
| 4453 |
"in_features": 4096,
|
|
@@ -4471,7 +4471,7 @@
|
|
| 4471 |
"model.layers.4.self_attn.k_proj": {
|
| 4472 |
"bias": false,
|
| 4473 |
"enable_norm": true,
|
| 4474 |
-
"enable_perm":
|
| 4475 |
"group_num": 1,
|
| 4476 |
"group_size": 4055,
|
| 4477 |
"in_features": 4096,
|
|
@@ -4495,7 +4495,7 @@
|
|
| 4495 |
"model.layers.4.self_attn.o_proj": {
|
| 4496 |
"bias": false,
|
| 4497 |
"enable_norm": true,
|
| 4498 |
-
"enable_perm":
|
| 4499 |
"group_num": 1,
|
| 4500 |
"group_size": 4055,
|
| 4501 |
"in_features": 4096,
|
|
@@ -4519,7 +4519,7 @@
|
|
| 4519 |
"model.layers.4.self_attn.q_proj": {
|
| 4520 |
"bias": false,
|
| 4521 |
"enable_norm": true,
|
| 4522 |
-
"enable_perm":
|
| 4523 |
"group_num": 1,
|
| 4524 |
"group_size": 4055,
|
| 4525 |
"in_features": 4096,
|
|
@@ -4543,7 +4543,7 @@
|
|
| 4543 |
"model.layers.4.self_attn.v_proj": {
|
| 4544 |
"bias": false,
|
| 4545 |
"enable_norm": true,
|
| 4546 |
-
"enable_perm":
|
| 4547 |
"group_num": 1,
|
| 4548 |
"group_size": 4055,
|
| 4549 |
"in_features": 4096,
|
|
@@ -4567,7 +4567,7 @@
|
|
| 4567 |
"model.layers.5.mlp.down_proj": {
|
| 4568 |
"bias": false,
|
| 4569 |
"enable_norm": true,
|
| 4570 |
-
"enable_perm":
|
| 4571 |
"group_num": 1,
|
| 4572 |
"group_size": 14192,
|
| 4573 |
"in_features": 14336,
|
|
@@ -4591,7 +4591,7 @@
|
|
| 4591 |
"model.layers.5.mlp.gate_proj": {
|
| 4592 |
"bias": false,
|
| 4593 |
"enable_norm": true,
|
| 4594 |
-
"enable_perm":
|
| 4595 |
"group_num": 1,
|
| 4596 |
"group_size": 4055,
|
| 4597 |
"in_features": 4096,
|
|
@@ -4615,7 +4615,7 @@
|
|
| 4615 |
"model.layers.5.mlp.up_proj": {
|
| 4616 |
"bias": false,
|
| 4617 |
"enable_norm": true,
|
| 4618 |
-
"enable_perm":
|
| 4619 |
"group_num": 1,
|
| 4620 |
"group_size": 4055,
|
| 4621 |
"in_features": 4096,
|
|
@@ -4639,7 +4639,7 @@
|
|
| 4639 |
"model.layers.5.self_attn.k_proj": {
|
| 4640 |
"bias": false,
|
| 4641 |
"enable_norm": true,
|
| 4642 |
-
"enable_perm":
|
| 4643 |
"group_num": 1,
|
| 4644 |
"group_size": 4055,
|
| 4645 |
"in_features": 4096,
|
|
@@ -4663,7 +4663,7 @@
|
|
| 4663 |
"model.layers.5.self_attn.o_proj": {
|
| 4664 |
"bias": false,
|
| 4665 |
"enable_norm": true,
|
| 4666 |
-
"enable_perm":
|
| 4667 |
"group_num": 1,
|
| 4668 |
"group_size": 4055,
|
| 4669 |
"in_features": 4096,
|
|
@@ -4687,7 +4687,7 @@
|
|
| 4687 |
"model.layers.5.self_attn.q_proj": {
|
| 4688 |
"bias": false,
|
| 4689 |
"enable_norm": true,
|
| 4690 |
-
"enable_perm":
|
| 4691 |
"group_num": 1,
|
| 4692 |
"group_size": 4055,
|
| 4693 |
"in_features": 4096,
|
|
@@ -4711,7 +4711,7 @@
|
|
| 4711 |
"model.layers.5.self_attn.v_proj": {
|
| 4712 |
"bias": false,
|
| 4713 |
"enable_norm": true,
|
| 4714 |
-
"enable_perm":
|
| 4715 |
"group_num": 1,
|
| 4716 |
"group_size": 4055,
|
| 4717 |
"in_features": 4096,
|
|
@@ -4735,7 +4735,7 @@
|
|
| 4735 |
"model.layers.6.mlp.down_proj": {
|
| 4736 |
"bias": false,
|
| 4737 |
"enable_norm": true,
|
| 4738 |
-
"enable_perm":
|
| 4739 |
"group_num": 1,
|
| 4740 |
"group_size": 14192,
|
| 4741 |
"in_features": 14336,
|
|
@@ -4759,7 +4759,7 @@
|
|
| 4759 |
"model.layers.6.mlp.gate_proj": {
|
| 4760 |
"bias": false,
|
| 4761 |
"enable_norm": true,
|
| 4762 |
-
"enable_perm":
|
| 4763 |
"group_num": 1,
|
| 4764 |
"group_size": 4055,
|
| 4765 |
"in_features": 4096,
|
|
@@ -4783,7 +4783,7 @@
|
|
| 4783 |
"model.layers.6.mlp.up_proj": {
|
| 4784 |
"bias": false,
|
| 4785 |
"enable_norm": true,
|
| 4786 |
-
"enable_perm":
|
| 4787 |
"group_num": 1,
|
| 4788 |
"group_size": 4055,
|
| 4789 |
"in_features": 4096,
|
|
@@ -4807,7 +4807,7 @@
|
|
| 4807 |
"model.layers.6.self_attn.k_proj": {
|
| 4808 |
"bias": false,
|
| 4809 |
"enable_norm": true,
|
| 4810 |
-
"enable_perm":
|
| 4811 |
"group_num": 1,
|
| 4812 |
"group_size": 4055,
|
| 4813 |
"in_features": 4096,
|
|
@@ -4831,7 +4831,7 @@
|
|
| 4831 |
"model.layers.6.self_attn.o_proj": {
|
| 4832 |
"bias": false,
|
| 4833 |
"enable_norm": true,
|
| 4834 |
-
"enable_perm":
|
| 4835 |
"group_num": 1,
|
| 4836 |
"group_size": 4055,
|
| 4837 |
"in_features": 4096,
|
|
@@ -4855,7 +4855,7 @@
|
|
| 4855 |
"model.layers.6.self_attn.q_proj": {
|
| 4856 |
"bias": false,
|
| 4857 |
"enable_norm": true,
|
| 4858 |
-
"enable_perm":
|
| 4859 |
"group_num": 1,
|
| 4860 |
"group_size": 4055,
|
| 4861 |
"in_features": 4096,
|
|
@@ -4879,7 +4879,7 @@
|
|
| 4879 |
"model.layers.6.self_attn.v_proj": {
|
| 4880 |
"bias": false,
|
| 4881 |
"enable_norm": true,
|
| 4882 |
-
"enable_perm":
|
| 4883 |
"group_num": 1,
|
| 4884 |
"group_size": 4055,
|
| 4885 |
"in_features": 4096,
|
|
@@ -4903,7 +4903,7 @@
|
|
| 4903 |
"model.layers.7.mlp.down_proj": {
|
| 4904 |
"bias": false,
|
| 4905 |
"enable_norm": true,
|
| 4906 |
-
"enable_perm":
|
| 4907 |
"group_num": 1,
|
| 4908 |
"group_size": 14192,
|
| 4909 |
"in_features": 14336,
|
|
@@ -4927,7 +4927,7 @@
|
|
| 4927 |
"model.layers.7.mlp.gate_proj": {
|
| 4928 |
"bias": false,
|
| 4929 |
"enable_norm": true,
|
| 4930 |
-
"enable_perm":
|
| 4931 |
"group_num": 1,
|
| 4932 |
"group_size": 4055,
|
| 4933 |
"in_features": 4096,
|
|
@@ -4951,7 +4951,7 @@
|
|
| 4951 |
"model.layers.7.mlp.up_proj": {
|
| 4952 |
"bias": false,
|
| 4953 |
"enable_norm": true,
|
| 4954 |
-
"enable_perm":
|
| 4955 |
"group_num": 1,
|
| 4956 |
"group_size": 4055,
|
| 4957 |
"in_features": 4096,
|
|
@@ -4975,7 +4975,7 @@
|
|
| 4975 |
"model.layers.7.self_attn.k_proj": {
|
| 4976 |
"bias": false,
|
| 4977 |
"enable_norm": true,
|
| 4978 |
-
"enable_perm":
|
| 4979 |
"group_num": 1,
|
| 4980 |
"group_size": 4055,
|
| 4981 |
"in_features": 4096,
|
|
@@ -4999,7 +4999,7 @@
|
|
| 4999 |
"model.layers.7.self_attn.o_proj": {
|
| 5000 |
"bias": false,
|
| 5001 |
"enable_norm": true,
|
| 5002 |
-
"enable_perm":
|
| 5003 |
"group_num": 1,
|
| 5004 |
"group_size": 4055,
|
| 5005 |
"in_features": 4096,
|
|
@@ -5023,7 +5023,7 @@
|
|
| 5023 |
"model.layers.7.self_attn.q_proj": {
|
| 5024 |
"bias": false,
|
| 5025 |
"enable_norm": true,
|
| 5026 |
-
"enable_perm":
|
| 5027 |
"group_num": 1,
|
| 5028 |
"group_size": 4055,
|
| 5029 |
"in_features": 4096,
|
|
@@ -5047,7 +5047,7 @@
|
|
| 5047 |
"model.layers.7.self_attn.v_proj": {
|
| 5048 |
"bias": false,
|
| 5049 |
"enable_norm": true,
|
| 5050 |
-
"enable_perm":
|
| 5051 |
"group_num": 1,
|
| 5052 |
"group_size": 4055,
|
| 5053 |
"in_features": 4096,
|
|
@@ -5071,7 +5071,7 @@
|
|
| 5071 |
"model.layers.8.mlp.down_proj": {
|
| 5072 |
"bias": false,
|
| 5073 |
"enable_norm": true,
|
| 5074 |
-
"enable_perm":
|
| 5075 |
"group_num": 1,
|
| 5076 |
"group_size": 14192,
|
| 5077 |
"in_features": 14336,
|
|
@@ -5095,7 +5095,7 @@
|
|
| 5095 |
"model.layers.8.mlp.gate_proj": {
|
| 5096 |
"bias": false,
|
| 5097 |
"enable_norm": true,
|
| 5098 |
-
"enable_perm":
|
| 5099 |
"group_num": 1,
|
| 5100 |
"group_size": 4055,
|
| 5101 |
"in_features": 4096,
|
|
@@ -5119,7 +5119,7 @@
|
|
| 5119 |
"model.layers.8.mlp.up_proj": {
|
| 5120 |
"bias": false,
|
| 5121 |
"enable_norm": true,
|
| 5122 |
-
"enable_perm":
|
| 5123 |
"group_num": 1,
|
| 5124 |
"group_size": 4055,
|
| 5125 |
"in_features": 4096,
|
|
@@ -5143,7 +5143,7 @@
|
|
| 5143 |
"model.layers.8.self_attn.k_proj": {
|
| 5144 |
"bias": false,
|
| 5145 |
"enable_norm": true,
|
| 5146 |
-
"enable_perm":
|
| 5147 |
"group_num": 1,
|
| 5148 |
"group_size": 4055,
|
| 5149 |
"in_features": 4096,
|
|
@@ -5167,7 +5167,7 @@
|
|
| 5167 |
"model.layers.8.self_attn.o_proj": {
|
| 5168 |
"bias": false,
|
| 5169 |
"enable_norm": true,
|
| 5170 |
-
"enable_perm":
|
| 5171 |
"group_num": 1,
|
| 5172 |
"group_size": 4055,
|
| 5173 |
"in_features": 4096,
|
|
@@ -5191,7 +5191,7 @@
|
|
| 5191 |
"model.layers.8.self_attn.q_proj": {
|
| 5192 |
"bias": false,
|
| 5193 |
"enable_norm": true,
|
| 5194 |
-
"enable_perm":
|
| 5195 |
"group_num": 1,
|
| 5196 |
"group_size": 4055,
|
| 5197 |
"in_features": 4096,
|
|
@@ -5215,7 +5215,7 @@
|
|
| 5215 |
"model.layers.8.self_attn.v_proj": {
|
| 5216 |
"bias": false,
|
| 5217 |
"enable_norm": true,
|
| 5218 |
-
"enable_perm":
|
| 5219 |
"group_num": 1,
|
| 5220 |
"group_size": 4055,
|
| 5221 |
"in_features": 4096,
|
|
@@ -5239,7 +5239,7 @@
|
|
| 5239 |
"model.layers.9.mlp.down_proj": {
|
| 5240 |
"bias": false,
|
| 5241 |
"enable_norm": true,
|
| 5242 |
-
"enable_perm":
|
| 5243 |
"group_num": 1,
|
| 5244 |
"group_size": 14192,
|
| 5245 |
"in_features": 14336,
|
|
@@ -5263,7 +5263,7 @@
|
|
| 5263 |
"model.layers.9.mlp.gate_proj": {
|
| 5264 |
"bias": false,
|
| 5265 |
"enable_norm": true,
|
| 5266 |
-
"enable_perm":
|
| 5267 |
"group_num": 1,
|
| 5268 |
"group_size": 4055,
|
| 5269 |
"in_features": 4096,
|
|
@@ -5287,7 +5287,7 @@
|
|
| 5287 |
"model.layers.9.mlp.up_proj": {
|
| 5288 |
"bias": false,
|
| 5289 |
"enable_norm": true,
|
| 5290 |
-
"enable_perm":
|
| 5291 |
"group_num": 1,
|
| 5292 |
"group_size": 4055,
|
| 5293 |
"in_features": 4096,
|
|
@@ -5311,7 +5311,7 @@
|
|
| 5311 |
"model.layers.9.self_attn.k_proj": {
|
| 5312 |
"bias": false,
|
| 5313 |
"enable_norm": true,
|
| 5314 |
-
"enable_perm":
|
| 5315 |
"group_num": 1,
|
| 5316 |
"group_size": 4055,
|
| 5317 |
"in_features": 4096,
|
|
@@ -5335,7 +5335,7 @@
|
|
| 5335 |
"model.layers.9.self_attn.o_proj": {
|
| 5336 |
"bias": false,
|
| 5337 |
"enable_norm": true,
|
| 5338 |
-
"enable_perm":
|
| 5339 |
"group_num": 1,
|
| 5340 |
"group_size": 4055,
|
| 5341 |
"in_features": 4096,
|
|
@@ -5359,7 +5359,7 @@
|
|
| 5359 |
"model.layers.9.self_attn.q_proj": {
|
| 5360 |
"bias": false,
|
| 5361 |
"enable_norm": true,
|
| 5362 |
-
"enable_perm":
|
| 5363 |
"group_num": 1,
|
| 5364 |
"group_size": 4055,
|
| 5365 |
"in_features": 4096,
|
|
@@ -5383,7 +5383,7 @@
|
|
| 5383 |
"model.layers.9.self_attn.v_proj": {
|
| 5384 |
"bias": false,
|
| 5385 |
"enable_norm": true,
|
| 5386 |
-
"enable_perm":
|
| 5387 |
"group_num": 1,
|
| 5388 |
"group_size": 4055,
|
| 5389 |
"in_features": 4096,
|
|
|
|
| 31 |
"model.layers.0.mlp.down_proj": {
|
| 32 |
"bias": false,
|
| 33 |
"enable_norm": true,
|
| 34 |
+
"enable_perm": false,
|
| 35 |
"group_num": 1,
|
| 36 |
"group_size": 14192,
|
| 37 |
"in_features": 14336,
|
|
|
|
| 55 |
"model.layers.0.mlp.gate_proj": {
|
| 56 |
"bias": false,
|
| 57 |
"enable_norm": true,
|
| 58 |
+
"enable_perm": false,
|
| 59 |
"group_num": 1,
|
| 60 |
"group_size": 4055,
|
| 61 |
"in_features": 4096,
|
|
|
|
| 79 |
"model.layers.0.mlp.up_proj": {
|
| 80 |
"bias": false,
|
| 81 |
"enable_norm": true,
|
| 82 |
+
"enable_perm": false,
|
| 83 |
"group_num": 1,
|
| 84 |
"group_size": 4055,
|
| 85 |
"in_features": 4096,
|
|
|
|
| 103 |
"model.layers.0.self_attn.k_proj": {
|
| 104 |
"bias": false,
|
| 105 |
"enable_norm": true,
|
| 106 |
+
"enable_perm": false,
|
| 107 |
"group_num": 1,
|
| 108 |
"group_size": 4055,
|
| 109 |
"in_features": 4096,
|
|
|
|
| 127 |
"model.layers.0.self_attn.o_proj": {
|
| 128 |
"bias": false,
|
| 129 |
"enable_norm": true,
|
| 130 |
+
"enable_perm": false,
|
| 131 |
"group_num": 1,
|
| 132 |
"group_size": 4055,
|
| 133 |
"in_features": 4096,
|
|
|
|
| 151 |
"model.layers.0.self_attn.q_proj": {
|
| 152 |
"bias": false,
|
| 153 |
"enable_norm": true,
|
| 154 |
+
"enable_perm": false,
|
| 155 |
"group_num": 1,
|
| 156 |
"group_size": 4055,
|
| 157 |
"in_features": 4096,
|
|
|
|
| 175 |
"model.layers.0.self_attn.v_proj": {
|
| 176 |
"bias": false,
|
| 177 |
"enable_norm": true,
|
| 178 |
+
"enable_perm": false,
|
| 179 |
"group_num": 1,
|
| 180 |
"group_size": 4055,
|
| 181 |
"in_features": 4096,
|
|
|
|
| 199 |
"model.layers.1.mlp.down_proj": {
|
| 200 |
"bias": false,
|
| 201 |
"enable_norm": true,
|
| 202 |
+
"enable_perm": false,
|
| 203 |
"group_num": 1,
|
| 204 |
"group_size": 14192,
|
| 205 |
"in_features": 14336,
|
|
|
|
| 223 |
"model.layers.1.mlp.gate_proj": {
|
| 224 |
"bias": false,
|
| 225 |
"enable_norm": true,
|
| 226 |
+
"enable_perm": false,
|
| 227 |
"group_num": 1,
|
| 228 |
"group_size": 4055,
|
| 229 |
"in_features": 4096,
|
|
|
|
| 247 |
"model.layers.1.mlp.up_proj": {
|
| 248 |
"bias": false,
|
| 249 |
"enable_norm": true,
|
| 250 |
+
"enable_perm": false,
|
| 251 |
"group_num": 1,
|
| 252 |
"group_size": 4055,
|
| 253 |
"in_features": 4096,
|
|
|
|
| 271 |
"model.layers.1.self_attn.k_proj": {
|
| 272 |
"bias": false,
|
| 273 |
"enable_norm": true,
|
| 274 |
+
"enable_perm": false,
|
| 275 |
"group_num": 1,
|
| 276 |
"group_size": 4055,
|
| 277 |
"in_features": 4096,
|
|
|
|
| 295 |
"model.layers.1.self_attn.o_proj": {
|
| 296 |
"bias": false,
|
| 297 |
"enable_norm": true,
|
| 298 |
+
"enable_perm": false,
|
| 299 |
"group_num": 1,
|
| 300 |
"group_size": 4055,
|
| 301 |
"in_features": 4096,
|
|
|
|
| 319 |
"model.layers.1.self_attn.q_proj": {
|
| 320 |
"bias": false,
|
| 321 |
"enable_norm": true,
|
| 322 |
+
"enable_perm": false,
|
| 323 |
"group_num": 1,
|
| 324 |
"group_size": 4055,
|
| 325 |
"in_features": 4096,
|
|
|
|
| 343 |
"model.layers.1.self_attn.v_proj": {
|
| 344 |
"bias": false,
|
| 345 |
"enable_norm": true,
|
| 346 |
+
"enable_perm": false,
|
| 347 |
"group_num": 1,
|
| 348 |
"group_size": 4055,
|
| 349 |
"in_features": 4096,
|
|
|
|
| 367 |
"model.layers.10.mlp.down_proj": {
|
| 368 |
"bias": false,
|
| 369 |
"enable_norm": true,
|
| 370 |
+
"enable_perm": false,
|
| 371 |
"group_num": 1,
|
| 372 |
"group_size": 14192,
|
| 373 |
"in_features": 14336,
|
|
|
|
| 391 |
"model.layers.10.mlp.gate_proj": {
|
| 392 |
"bias": false,
|
| 393 |
"enable_norm": true,
|
| 394 |
+
"enable_perm": false,
|
| 395 |
"group_num": 1,
|
| 396 |
"group_size": 4055,
|
| 397 |
"in_features": 4096,
|
|
|
|
| 415 |
"model.layers.10.mlp.up_proj": {
|
| 416 |
"bias": false,
|
| 417 |
"enable_norm": true,
|
| 418 |
+
"enable_perm": false,
|
| 419 |
"group_num": 1,
|
| 420 |
"group_size": 4055,
|
| 421 |
"in_features": 4096,
|
|
|
|
| 439 |
"model.layers.10.self_attn.k_proj": {
|
| 440 |
"bias": false,
|
| 441 |
"enable_norm": true,
|
| 442 |
+
"enable_perm": false,
|
| 443 |
"group_num": 1,
|
| 444 |
"group_size": 4055,
|
| 445 |
"in_features": 4096,
|
|
|
|
| 463 |
"model.layers.10.self_attn.o_proj": {
|
| 464 |
"bias": false,
|
| 465 |
"enable_norm": true,
|
| 466 |
+
"enable_perm": false,
|
| 467 |
"group_num": 1,
|
| 468 |
"group_size": 4055,
|
| 469 |
"in_features": 4096,
|
|
|
|
| 487 |
"model.layers.10.self_attn.q_proj": {
|
| 488 |
"bias": false,
|
| 489 |
"enable_norm": true,
|
| 490 |
+
"enable_perm": false,
|
| 491 |
"group_num": 1,
|
| 492 |
"group_size": 4055,
|
| 493 |
"in_features": 4096,
|
|
|
|
| 511 |
"model.layers.10.self_attn.v_proj": {
|
| 512 |
"bias": false,
|
| 513 |
"enable_norm": true,
|
| 514 |
+
"enable_perm": false,
|
| 515 |
"group_num": 1,
|
| 516 |
"group_size": 4055,
|
| 517 |
"in_features": 4096,
|
|
|
|
| 535 |
"model.layers.11.mlp.down_proj": {
|
| 536 |
"bias": false,
|
| 537 |
"enable_norm": true,
|
| 538 |
+
"enable_perm": false,
|
| 539 |
"group_num": 1,
|
| 540 |
"group_size": 14192,
|
| 541 |
"in_features": 14336,
|
|
|
|
| 559 |
"model.layers.11.mlp.gate_proj": {
|
| 560 |
"bias": false,
|
| 561 |
"enable_norm": true,
|
| 562 |
+
"enable_perm": false,
|
| 563 |
"group_num": 1,
|
| 564 |
"group_size": 4055,
|
| 565 |
"in_features": 4096,
|
|
|
|
| 583 |
"model.layers.11.mlp.up_proj": {
|
| 584 |
"bias": false,
|
| 585 |
"enable_norm": true,
|
| 586 |
+
"enable_perm": false,
|
| 587 |
"group_num": 1,
|
| 588 |
"group_size": 4055,
|
| 589 |
"in_features": 4096,
|
|
|
|
| 607 |
"model.layers.11.self_attn.k_proj": {
|
| 608 |
"bias": false,
|
| 609 |
"enable_norm": true,
|
| 610 |
+
"enable_perm": false,
|
| 611 |
"group_num": 1,
|
| 612 |
"group_size": 4055,
|
| 613 |
"in_features": 4096,
|
|
|
|
| 631 |
"model.layers.11.self_attn.o_proj": {
|
| 632 |
"bias": false,
|
| 633 |
"enable_norm": true,
|
| 634 |
+
"enable_perm": false,
|
| 635 |
"group_num": 1,
|
| 636 |
"group_size": 4055,
|
| 637 |
"in_features": 4096,
|
|
|
|
| 655 |
"model.layers.11.self_attn.q_proj": {
|
| 656 |
"bias": false,
|
| 657 |
"enable_norm": true,
|
| 658 |
+
"enable_perm": false,
|
| 659 |
"group_num": 1,
|
| 660 |
"group_size": 4055,
|
| 661 |
"in_features": 4096,
|
|
|
|
| 679 |
"model.layers.11.self_attn.v_proj": {
|
| 680 |
"bias": false,
|
| 681 |
"enable_norm": true,
|
| 682 |
+
"enable_perm": false,
|
| 683 |
"group_num": 1,
|
| 684 |
"group_size": 4055,
|
| 685 |
"in_features": 4096,
|
|
|
|
| 703 |
"model.layers.12.mlp.down_proj": {
|
| 704 |
"bias": false,
|
| 705 |
"enable_norm": true,
|
| 706 |
+
"enable_perm": false,
|
| 707 |
"group_num": 1,
|
| 708 |
"group_size": 14192,
|
| 709 |
"in_features": 14336,
|
|
|
|
| 727 |
"model.layers.12.mlp.gate_proj": {
|
| 728 |
"bias": false,
|
| 729 |
"enable_norm": true,
|
| 730 |
+
"enable_perm": false,
|
| 731 |
"group_num": 1,
|
| 732 |
"group_size": 4055,
|
| 733 |
"in_features": 4096,
|
|
|
|
| 751 |
"model.layers.12.mlp.up_proj": {
|
| 752 |
"bias": false,
|
| 753 |
"enable_norm": true,
|
| 754 |
+
"enable_perm": false,
|
| 755 |
"group_num": 1,
|
| 756 |
"group_size": 4055,
|
| 757 |
"in_features": 4096,
|
|
|
|
| 775 |
"model.layers.12.self_attn.k_proj": {
|
| 776 |
"bias": false,
|
| 777 |
"enable_norm": true,
|
| 778 |
+
"enable_perm": false,
|
| 779 |
"group_num": 1,
|
| 780 |
"group_size": 4055,
|
| 781 |
"in_features": 4096,
|
|
|
|
| 799 |
"model.layers.12.self_attn.o_proj": {
|
| 800 |
"bias": false,
|
| 801 |
"enable_norm": true,
|
| 802 |
+
"enable_perm": false,
|
| 803 |
"group_num": 1,
|
| 804 |
"group_size": 4055,
|
| 805 |
"in_features": 4096,
|
|
|
|
| 823 |
"model.layers.12.self_attn.q_proj": {
|
| 824 |
"bias": false,
|
| 825 |
"enable_norm": true,
|
| 826 |
+
"enable_perm": false,
|
| 827 |
"group_num": 1,
|
| 828 |
"group_size": 4055,
|
| 829 |
"in_features": 4096,
|
|
|
|
| 847 |
"model.layers.12.self_attn.v_proj": {
|
| 848 |
"bias": false,
|
| 849 |
"enable_norm": true,
|
| 850 |
+
"enable_perm": false,
|
| 851 |
"group_num": 1,
|
| 852 |
"group_size": 4055,
|
| 853 |
"in_features": 4096,
|
|
|
|
| 871 |
"model.layers.13.mlp.down_proj": {
|
| 872 |
"bias": false,
|
| 873 |
"enable_norm": true,
|
| 874 |
+
"enable_perm": false,
|
| 875 |
"group_num": 1,
|
| 876 |
"group_size": 14192,
|
| 877 |
"in_features": 14336,
|
|
|
|
| 895 |
"model.layers.13.mlp.gate_proj": {
|
| 896 |
"bias": false,
|
| 897 |
"enable_norm": true,
|
| 898 |
+
"enable_perm": false,
|
| 899 |
"group_num": 1,
|
| 900 |
"group_size": 4055,
|
| 901 |
"in_features": 4096,
|
|
|
|
| 919 |
"model.layers.13.mlp.up_proj": {
|
| 920 |
"bias": false,
|
| 921 |
"enable_norm": true,
|
| 922 |
+
"enable_perm": false,
|
| 923 |
"group_num": 1,
|
| 924 |
"group_size": 4055,
|
| 925 |
"in_features": 4096,
|
|
|
|
| 943 |
"model.layers.13.self_attn.k_proj": {
|
| 944 |
"bias": false,
|
| 945 |
"enable_norm": true,
|
| 946 |
+
"enable_perm": false,
|
| 947 |
"group_num": 1,
|
| 948 |
"group_size": 4055,
|
| 949 |
"in_features": 4096,
|
|
|
|
| 967 |
"model.layers.13.self_attn.o_proj": {
|
| 968 |
"bias": false,
|
| 969 |
"enable_norm": true,
|
| 970 |
+
"enable_perm": false,
|
| 971 |
"group_num": 1,
|
| 972 |
"group_size": 4055,
|
| 973 |
"in_features": 4096,
|
|
|
|
| 991 |
"model.layers.13.self_attn.q_proj": {
|
| 992 |
"bias": false,
|
| 993 |
"enable_norm": true,
|
| 994 |
+
"enable_perm": false,
|
| 995 |
"group_num": 1,
|
| 996 |
"group_size": 4055,
|
| 997 |
"in_features": 4096,
|
|
|
|
| 1015 |
"model.layers.13.self_attn.v_proj": {
|
| 1016 |
"bias": false,
|
| 1017 |
"enable_norm": true,
|
| 1018 |
+
"enable_perm": false,
|
| 1019 |
"group_num": 1,
|
| 1020 |
"group_size": 4055,
|
| 1021 |
"in_features": 4096,
|
|
|
|
| 1039 |
"model.layers.14.mlp.down_proj": {
|
| 1040 |
"bias": false,
|
| 1041 |
"enable_norm": true,
|
| 1042 |
+
"enable_perm": false,
|
| 1043 |
"group_num": 1,
|
| 1044 |
"group_size": 14192,
|
| 1045 |
"in_features": 14336,
|
|
|
|
| 1063 |
"model.layers.14.mlp.gate_proj": {
|
| 1064 |
"bias": false,
|
| 1065 |
"enable_norm": true,
|
| 1066 |
+
"enable_perm": false,
|
| 1067 |
"group_num": 1,
|
| 1068 |
"group_size": 4055,
|
| 1069 |
"in_features": 4096,
|
|
|
|
| 1087 |
"model.layers.14.mlp.up_proj": {
|
| 1088 |
"bias": false,
|
| 1089 |
"enable_norm": true,
|
| 1090 |
+
"enable_perm": false,
|
| 1091 |
"group_num": 1,
|
| 1092 |
"group_size": 4055,
|
| 1093 |
"in_features": 4096,
|
|
|
|
| 1111 |
"model.layers.14.self_attn.k_proj": {
|
| 1112 |
"bias": false,
|
| 1113 |
"enable_norm": true,
|
| 1114 |
+
"enable_perm": false,
|
| 1115 |
"group_num": 1,
|
| 1116 |
"group_size": 4055,
|
| 1117 |
"in_features": 4096,
|
|
|
|
| 1135 |
"model.layers.14.self_attn.o_proj": {
|
| 1136 |
"bias": false,
|
| 1137 |
"enable_norm": true,
|
| 1138 |
+
"enable_perm": false,
|
| 1139 |
"group_num": 1,
|
| 1140 |
"group_size": 4055,
|
| 1141 |
"in_features": 4096,
|
|
|
|
| 1159 |
"model.layers.14.self_attn.q_proj": {
|
| 1160 |
"bias": false,
|
| 1161 |
"enable_norm": true,
|
| 1162 |
+
"enable_perm": false,
|
| 1163 |
"group_num": 1,
|
| 1164 |
"group_size": 4055,
|
| 1165 |
"in_features": 4096,
|
|
|
|
| 1183 |
"model.layers.14.self_attn.v_proj": {
|
| 1184 |
"bias": false,
|
| 1185 |
"enable_norm": true,
|
| 1186 |
+
"enable_perm": false,
|
| 1187 |
"group_num": 1,
|
| 1188 |
"group_size": 4055,
|
| 1189 |
"in_features": 4096,
|
|
|
|
| 1207 |
"model.layers.15.mlp.down_proj": {
|
| 1208 |
"bias": false,
|
| 1209 |
"enable_norm": true,
|
| 1210 |
+
"enable_perm": false,
|
| 1211 |
"group_num": 1,
|
| 1212 |
"group_size": 14192,
|
| 1213 |
"in_features": 14336,
|
|
|
|
| 1231 |
"model.layers.15.mlp.gate_proj": {
|
| 1232 |
"bias": false,
|
| 1233 |
"enable_norm": true,
|
| 1234 |
+
"enable_perm": false,
|
| 1235 |
"group_num": 1,
|
| 1236 |
"group_size": 4055,
|
| 1237 |
"in_features": 4096,
|
|
|
|
| 1255 |
"model.layers.15.mlp.up_proj": {
|
| 1256 |
"bias": false,
|
| 1257 |
"enable_norm": true,
|
| 1258 |
+
"enable_perm": false,
|
| 1259 |
"group_num": 1,
|
| 1260 |
"group_size": 4055,
|
| 1261 |
"in_features": 4096,
|
|
|
|
| 1279 |
"model.layers.15.self_attn.k_proj": {
|
| 1280 |
"bias": false,
|
| 1281 |
"enable_norm": true,
|
| 1282 |
+
"enable_perm": false,
|
| 1283 |
"group_num": 1,
|
| 1284 |
"group_size": 4055,
|
| 1285 |
"in_features": 4096,
|
|
|
|
| 1303 |
"model.layers.15.self_attn.o_proj": {
|
| 1304 |
"bias": false,
|
| 1305 |
"enable_norm": true,
|
| 1306 |
+
"enable_perm": false,
|
| 1307 |
"group_num": 1,
|
| 1308 |
"group_size": 4055,
|
| 1309 |
"in_features": 4096,
|
|
|
|
| 1327 |
"model.layers.15.self_attn.q_proj": {
|
| 1328 |
"bias": false,
|
| 1329 |
"enable_norm": true,
|
| 1330 |
+
"enable_perm": false,
|
| 1331 |
"group_num": 1,
|
| 1332 |
"group_size": 4055,
|
| 1333 |
"in_features": 4096,
|
|
|
|
| 1351 |
"model.layers.15.self_attn.v_proj": {
|
| 1352 |
"bias": false,
|
| 1353 |
"enable_norm": true,
|
| 1354 |
+
"enable_perm": false,
|
| 1355 |
"group_num": 1,
|
| 1356 |
"group_size": 4055,
|
| 1357 |
"in_features": 4096,
|
|
|
|
| 1375 |
"model.layers.16.mlp.down_proj": {
|
| 1376 |
"bias": false,
|
| 1377 |
"enable_norm": true,
|
| 1378 |
+
"enable_perm": false,
|
| 1379 |
"group_num": 1,
|
| 1380 |
"group_size": 14192,
|
| 1381 |
"in_features": 14336,
|
|
|
|
| 1399 |
"model.layers.16.mlp.gate_proj": {
|
| 1400 |
"bias": false,
|
| 1401 |
"enable_norm": true,
|
| 1402 |
+
"enable_perm": false,
|
| 1403 |
"group_num": 1,
|
| 1404 |
"group_size": 4055,
|
| 1405 |
"in_features": 4096,
|
|
|
|
| 1423 |
"model.layers.16.mlp.up_proj": {
|
| 1424 |
"bias": false,
|
| 1425 |
"enable_norm": true,
|
| 1426 |
+
"enable_perm": false,
|
| 1427 |
"group_num": 1,
|
| 1428 |
"group_size": 4055,
|
| 1429 |
"in_features": 4096,
|
|
|
|
| 1447 |
"model.layers.16.self_attn.k_proj": {
|
| 1448 |
"bias": false,
|
| 1449 |
"enable_norm": true,
|
| 1450 |
+
"enable_perm": false,
|
| 1451 |
"group_num": 1,
|
| 1452 |
"group_size": 4055,
|
| 1453 |
"in_features": 4096,
|
|
|
|
| 1471 |
"model.layers.16.self_attn.o_proj": {
|
| 1472 |
"bias": false,
|
| 1473 |
"enable_norm": true,
|
| 1474 |
+
"enable_perm": false,
|
| 1475 |
"group_num": 1,
|
| 1476 |
"group_size": 4055,
|
| 1477 |
"in_features": 4096,
|
|
|
|
| 1495 |
"model.layers.16.self_attn.q_proj": {
|
| 1496 |
"bias": false,
|
| 1497 |
"enable_norm": true,
|
| 1498 |
+
"enable_perm": false,
|
| 1499 |
"group_num": 1,
|
| 1500 |
"group_size": 4055,
|
| 1501 |
"in_features": 4096,
|
|
|
|
| 1519 |
"model.layers.16.self_attn.v_proj": {
|
| 1520 |
"bias": false,
|
| 1521 |
"enable_norm": true,
|
| 1522 |
+
"enable_perm": false,
|
| 1523 |
"group_num": 1,
|
| 1524 |
"group_size": 4055,
|
| 1525 |
"in_features": 4096,
|
|
|
|
| 1543 |
"model.layers.17.mlp.down_proj": {
|
| 1544 |
"bias": false,
|
| 1545 |
"enable_norm": true,
|
| 1546 |
+
"enable_perm": false,
|
| 1547 |
"group_num": 1,
|
| 1548 |
"group_size": 14192,
|
| 1549 |
"in_features": 14336,
|
|
|
|
| 1567 |
"model.layers.17.mlp.gate_proj": {
|
| 1568 |
"bias": false,
|
| 1569 |
"enable_norm": true,
|
| 1570 |
+
"enable_perm": false,
|
| 1571 |
"group_num": 1,
|
| 1572 |
"group_size": 4055,
|
| 1573 |
"in_features": 4096,
|
|
|
|
| 1591 |
"model.layers.17.mlp.up_proj": {
|
| 1592 |
"bias": false,
|
| 1593 |
"enable_norm": true,
|
| 1594 |
+
"enable_perm": false,
|
| 1595 |
"group_num": 1,
|
| 1596 |
"group_size": 4055,
|
| 1597 |
"in_features": 4096,
|
|
|
|
| 1615 |
"model.layers.17.self_attn.k_proj": {
|
| 1616 |
"bias": false,
|
| 1617 |
"enable_norm": true,
|
| 1618 |
+
"enable_perm": false,
|
| 1619 |
"group_num": 1,
|
| 1620 |
"group_size": 4055,
|
| 1621 |
"in_features": 4096,
|
|
|
|
| 1639 |
"model.layers.17.self_attn.o_proj": {
|
| 1640 |
"bias": false,
|
| 1641 |
"enable_norm": true,
|
| 1642 |
+
"enable_perm": false,
|
| 1643 |
"group_num": 1,
|
| 1644 |
"group_size": 4055,
|
| 1645 |
"in_features": 4096,
|
|
|
|
| 1663 |
"model.layers.17.self_attn.q_proj": {
|
| 1664 |
"bias": false,
|
| 1665 |
"enable_norm": true,
|
| 1666 |
+
"enable_perm": false,
|
| 1667 |
"group_num": 1,
|
| 1668 |
"group_size": 4055,
|
| 1669 |
"in_features": 4096,
|
|
|
|
| 1687 |
"model.layers.17.self_attn.v_proj": {
|
| 1688 |
"bias": false,
|
| 1689 |
"enable_norm": true,
|
| 1690 |
+
"enable_perm": false,
|
| 1691 |
"group_num": 1,
|
| 1692 |
"group_size": 4055,
|
| 1693 |
"in_features": 4096,
|
|
|
|
| 1711 |
"model.layers.18.mlp.down_proj": {
|
| 1712 |
"bias": false,
|
| 1713 |
"enable_norm": true,
|
| 1714 |
+
"enable_perm": false,
|
| 1715 |
"group_num": 1,
|
| 1716 |
"group_size": 14192,
|
| 1717 |
"in_features": 14336,
|
|
|
|
| 1735 |
"model.layers.18.mlp.gate_proj": {
|
| 1736 |
"bias": false,
|
| 1737 |
"enable_norm": true,
|
| 1738 |
+
"enable_perm": false,
|
| 1739 |
"group_num": 1,
|
| 1740 |
"group_size": 4055,
|
| 1741 |
"in_features": 4096,
|
|
|
|
| 1759 |
"model.layers.18.mlp.up_proj": {
|
| 1760 |
"bias": false,
|
| 1761 |
"enable_norm": true,
|
| 1762 |
+
"enable_perm": false,
|
| 1763 |
"group_num": 1,
|
| 1764 |
"group_size": 4055,
|
| 1765 |
"in_features": 4096,
|
|
|
|
| 1783 |
"model.layers.18.self_attn.k_proj": {
|
| 1784 |
"bias": false,
|
| 1785 |
"enable_norm": true,
|
| 1786 |
+
"enable_perm": false,
|
| 1787 |
"group_num": 1,
|
| 1788 |
"group_size": 4055,
|
| 1789 |
"in_features": 4096,
|
|
|
|
| 1807 |
"model.layers.18.self_attn.o_proj": {
|
| 1808 |
"bias": false,
|
| 1809 |
"enable_norm": true,
|
| 1810 |
+
"enable_perm": false,
|
| 1811 |
"group_num": 1,
|
| 1812 |
"group_size": 4055,
|
| 1813 |
"in_features": 4096,
|
|
|
|
| 1831 |
"model.layers.18.self_attn.q_proj": {
|
| 1832 |
"bias": false,
|
| 1833 |
"enable_norm": true,
|
| 1834 |
+
"enable_perm": false,
|
| 1835 |
"group_num": 1,
|
| 1836 |
"group_size": 4055,
|
| 1837 |
"in_features": 4096,
|
|
|
|
| 1855 |
"model.layers.18.self_attn.v_proj": {
|
| 1856 |
"bias": false,
|
| 1857 |
"enable_norm": true,
|
| 1858 |
+
"enable_perm": false,
|
| 1859 |
"group_num": 1,
|
| 1860 |
"group_size": 4055,
|
| 1861 |
"in_features": 4096,
|
|
|
|
| 1879 |
"model.layers.19.mlp.down_proj": {
|
| 1880 |
"bias": false,
|
| 1881 |
"enable_norm": true,
|
| 1882 |
+
"enable_perm": false,
|
| 1883 |
"group_num": 1,
|
| 1884 |
"group_size": 14192,
|
| 1885 |
"in_features": 14336,
|
|
|
|
| 1903 |
"model.layers.19.mlp.gate_proj": {
|
| 1904 |
"bias": false,
|
| 1905 |
"enable_norm": true,
|
| 1906 |
+
"enable_perm": false,
|
| 1907 |
"group_num": 1,
|
| 1908 |
"group_size": 4055,
|
| 1909 |
"in_features": 4096,
|
|
|
|
| 1927 |
"model.layers.19.mlp.up_proj": {
|
| 1928 |
"bias": false,
|
| 1929 |
"enable_norm": true,
|
| 1930 |
+
"enable_perm": false,
|
| 1931 |
"group_num": 1,
|
| 1932 |
"group_size": 4055,
|
| 1933 |
"in_features": 4096,
|
|
|
|
| 1951 |
"model.layers.19.self_attn.k_proj": {
|
| 1952 |
"bias": false,
|
| 1953 |
"enable_norm": true,
|
| 1954 |
+
"enable_perm": false,
|
| 1955 |
"group_num": 1,
|
| 1956 |
"group_size": 4055,
|
| 1957 |
"in_features": 4096,
|
|
|
|
| 1975 |
"model.layers.19.self_attn.o_proj": {
|
| 1976 |
"bias": false,
|
| 1977 |
"enable_norm": true,
|
| 1978 |
+
"enable_perm": false,
|
| 1979 |
"group_num": 1,
|
| 1980 |
"group_size": 4055,
|
| 1981 |
"in_features": 4096,
|
|
|
|
| 1999 |
"model.layers.19.self_attn.q_proj": {
|
| 2000 |
"bias": false,
|
| 2001 |
"enable_norm": true,
|
| 2002 |
+
"enable_perm": false,
|
| 2003 |
"group_num": 1,
|
| 2004 |
"group_size": 4055,
|
| 2005 |
"in_features": 4096,
|
|
|
|
| 2023 |
"model.layers.19.self_attn.v_proj": {
|
| 2024 |
"bias": false,
|
| 2025 |
"enable_norm": true,
|
| 2026 |
+
"enable_perm": false,
|
| 2027 |
"group_num": 1,
|
| 2028 |
"group_size": 4055,
|
| 2029 |
"in_features": 4096,
|
|
|
|
| 2047 |
"model.layers.2.mlp.down_proj": {
|
| 2048 |
"bias": false,
|
| 2049 |
"enable_norm": true,
|
| 2050 |
+
"enable_perm": false,
|
| 2051 |
"group_num": 1,
|
| 2052 |
"group_size": 14192,
|
| 2053 |
"in_features": 14336,
|
|
|
|
| 2071 |
"model.layers.2.mlp.gate_proj": {
|
| 2072 |
"bias": false,
|
| 2073 |
"enable_norm": true,
|
| 2074 |
+
"enable_perm": false,
|
| 2075 |
"group_num": 1,
|
| 2076 |
"group_size": 4055,
|
| 2077 |
"in_features": 4096,
|
|
|
|
| 2095 |
"model.layers.2.mlp.up_proj": {
|
| 2096 |
"bias": false,
|
| 2097 |
"enable_norm": true,
|
| 2098 |
+
"enable_perm": false,
|
| 2099 |
"group_num": 1,
|
| 2100 |
"group_size": 4055,
|
| 2101 |
"in_features": 4096,
|
|
|
|
| 2119 |
"model.layers.2.self_attn.k_proj": {
|
| 2120 |
"bias": false,
|
| 2121 |
"enable_norm": true,
|
| 2122 |
+
"enable_perm": false,
|
| 2123 |
"group_num": 1,
|
| 2124 |
"group_size": 4055,
|
| 2125 |
"in_features": 4096,
|
|
|
|
| 2143 |
"model.layers.2.self_attn.o_proj": {
|
| 2144 |
"bias": false,
|
| 2145 |
"enable_norm": true,
|
| 2146 |
+
"enable_perm": false,
|
| 2147 |
"group_num": 1,
|
| 2148 |
"group_size": 4055,
|
| 2149 |
"in_features": 4096,
|
|
|
|
| 2167 |
"model.layers.2.self_attn.q_proj": {
|
| 2168 |
"bias": false,
|
| 2169 |
"enable_norm": true,
|
| 2170 |
+
"enable_perm": false,
|
| 2171 |
"group_num": 1,
|
| 2172 |
"group_size": 4055,
|
| 2173 |
"in_features": 4096,
|
|
|
|
| 2191 |
"model.layers.2.self_attn.v_proj": {
|
| 2192 |
"bias": false,
|
| 2193 |
"enable_norm": true,
|
| 2194 |
+
"enable_perm": false,
|
| 2195 |
"group_num": 1,
|
| 2196 |
"group_size": 4055,
|
| 2197 |
"in_features": 4096,
|
|
|
|
| 2215 |
"model.layers.20.mlp.down_proj": {
|
| 2216 |
"bias": false,
|
| 2217 |
"enable_norm": true,
|
| 2218 |
+
"enable_perm": false,
|
| 2219 |
"group_num": 1,
|
| 2220 |
"group_size": 14192,
|
| 2221 |
"in_features": 14336,
|
|
|
|
| 2239 |
"model.layers.20.mlp.gate_proj": {
|
| 2240 |
"bias": false,
|
| 2241 |
"enable_norm": true,
|
| 2242 |
+
"enable_perm": false,
|
| 2243 |
"group_num": 1,
|
| 2244 |
"group_size": 4055,
|
| 2245 |
"in_features": 4096,
|
|
|
|
| 2263 |
"model.layers.20.mlp.up_proj": {
|
| 2264 |
"bias": false,
|
| 2265 |
"enable_norm": true,
|
| 2266 |
+
"enable_perm": false,
|
| 2267 |
"group_num": 1,
|
| 2268 |
"group_size": 4055,
|
| 2269 |
"in_features": 4096,
|
|
|
|
| 2287 |
"model.layers.20.self_attn.k_proj": {
|
| 2288 |
"bias": false,
|
| 2289 |
"enable_norm": true,
|
| 2290 |
+
"enable_perm": false,
|
| 2291 |
"group_num": 1,
|
| 2292 |
"group_size": 4055,
|
| 2293 |
"in_features": 4096,
|
|
|
|
| 2311 |
"model.layers.20.self_attn.o_proj": {
|
| 2312 |
"bias": false,
|
| 2313 |
"enable_norm": true,
|
| 2314 |
+
"enable_perm": false,
|
| 2315 |
"group_num": 1,
|
| 2316 |
"group_size": 4055,
|
| 2317 |
"in_features": 4096,
|
|
|
|
| 2335 |
"model.layers.20.self_attn.q_proj": {
|
| 2336 |
"bias": false,
|
| 2337 |
"enable_norm": true,
|
| 2338 |
+
"enable_perm": false,
|
| 2339 |
"group_num": 1,
|
| 2340 |
"group_size": 4055,
|
| 2341 |
"in_features": 4096,
|
|
|
|
| 2359 |
"model.layers.20.self_attn.v_proj": {
|
| 2360 |
"bias": false,
|
| 2361 |
"enable_norm": true,
|
| 2362 |
+
"enable_perm": false,
|
| 2363 |
"group_num": 1,
|
| 2364 |
"group_size": 4055,
|
| 2365 |
"in_features": 4096,
|
|
|
|
| 2383 |
"model.layers.21.mlp.down_proj": {
|
| 2384 |
"bias": false,
|
| 2385 |
"enable_norm": true,
|
| 2386 |
+
"enable_perm": false,
|
| 2387 |
"group_num": 1,
|
| 2388 |
"group_size": 14192,
|
| 2389 |
"in_features": 14336,
|
|
|
|
| 2407 |
"model.layers.21.mlp.gate_proj": {
|
| 2408 |
"bias": false,
|
| 2409 |
"enable_norm": true,
|
| 2410 |
+
"enable_perm": false,
|
| 2411 |
"group_num": 1,
|
| 2412 |
"group_size": 4055,
|
| 2413 |
"in_features": 4096,
|
|
|
|
| 2431 |
"model.layers.21.mlp.up_proj": {
|
| 2432 |
"bias": false,
|
| 2433 |
"enable_norm": true,
|
| 2434 |
+
"enable_perm": false,
|
| 2435 |
"group_num": 1,
|
| 2436 |
"group_size": 4055,
|
| 2437 |
"in_features": 4096,
|
|
|
|
| 2455 |
"model.layers.21.self_attn.k_proj": {
|
| 2456 |
"bias": false,
|
| 2457 |
"enable_norm": true,
|
| 2458 |
+
"enable_perm": false,
|
| 2459 |
"group_num": 1,
|
| 2460 |
"group_size": 4055,
|
| 2461 |
"in_features": 4096,
|
|
|
|
| 2479 |
"model.layers.21.self_attn.o_proj": {
|
| 2480 |
"bias": false,
|
| 2481 |
"enable_norm": true,
|
| 2482 |
+
"enable_perm": false,
|
| 2483 |
"group_num": 1,
|
| 2484 |
"group_size": 4055,
|
| 2485 |
"in_features": 4096,
|
|
|
|
| 2503 |
"model.layers.21.self_attn.q_proj": {
|
| 2504 |
"bias": false,
|
| 2505 |
"enable_norm": true,
|
| 2506 |
+
"enable_perm": false,
|
| 2507 |
"group_num": 1,
|
| 2508 |
"group_size": 4055,
|
| 2509 |
"in_features": 4096,
|
|
|
|
| 2527 |
"model.layers.21.self_attn.v_proj": {
|
| 2528 |
"bias": false,
|
| 2529 |
"enable_norm": true,
|
| 2530 |
+
"enable_perm": false,
|
| 2531 |
"group_num": 1,
|
| 2532 |
"group_size": 4055,
|
| 2533 |
"in_features": 4096,
|
|
|
|
| 2551 |
"model.layers.22.mlp.down_proj": {
|
| 2552 |
"bias": false,
|
| 2553 |
"enable_norm": true,
|
| 2554 |
+
"enable_perm": false,
|
| 2555 |
"group_num": 1,
|
| 2556 |
"group_size": 14192,
|
| 2557 |
"in_features": 14336,
|
|
|
|
| 2575 |
"model.layers.22.mlp.gate_proj": {
|
| 2576 |
"bias": false,
|
| 2577 |
"enable_norm": true,
|
| 2578 |
+
"enable_perm": false,
|
| 2579 |
"group_num": 1,
|
| 2580 |
"group_size": 4055,
|
| 2581 |
"in_features": 4096,
|
|
|
|
| 2599 |
"model.layers.22.mlp.up_proj": {
|
| 2600 |
"bias": false,
|
| 2601 |
"enable_norm": true,
|
| 2602 |
+
"enable_perm": false,
|
| 2603 |
"group_num": 1,
|
| 2604 |
"group_size": 4055,
|
| 2605 |
"in_features": 4096,
|
|
|
|
| 2623 |
"model.layers.22.self_attn.k_proj": {
|
| 2624 |
"bias": false,
|
| 2625 |
"enable_norm": true,
|
| 2626 |
+
"enable_perm": false,
|
| 2627 |
"group_num": 1,
|
| 2628 |
"group_size": 4055,
|
| 2629 |
"in_features": 4096,
|
|
|
|
| 2647 |
"model.layers.22.self_attn.o_proj": {
|
| 2648 |
"bias": false,
|
| 2649 |
"enable_norm": true,
|
| 2650 |
+
"enable_perm": false,
|
| 2651 |
"group_num": 1,
|
| 2652 |
"group_size": 4055,
|
| 2653 |
"in_features": 4096,
|
|
|
|
| 2671 |
"model.layers.22.self_attn.q_proj": {
|
| 2672 |
"bias": false,
|
| 2673 |
"enable_norm": true,
|
| 2674 |
+
"enable_perm": false,
|
| 2675 |
"group_num": 1,
|
| 2676 |
"group_size": 4055,
|
| 2677 |
"in_features": 4096,
|
|
|
|
| 2695 |
"model.layers.22.self_attn.v_proj": {
|
| 2696 |
"bias": false,
|
| 2697 |
"enable_norm": true,
|
| 2698 |
+
"enable_perm": false,
|
| 2699 |
"group_num": 1,
|
| 2700 |
"group_size": 4055,
|
| 2701 |
"in_features": 4096,
|
|
|
|
| 2719 |
"model.layers.23.mlp.down_proj": {
|
| 2720 |
"bias": false,
|
| 2721 |
"enable_norm": true,
|
| 2722 |
+
"enable_perm": false,
|
| 2723 |
"group_num": 1,
|
| 2724 |
"group_size": 14192,
|
| 2725 |
"in_features": 14336,
|
|
|
|
| 2743 |
"model.layers.23.mlp.gate_proj": {
|
| 2744 |
"bias": false,
|
| 2745 |
"enable_norm": true,
|
| 2746 |
+
"enable_perm": false,
|
| 2747 |
"group_num": 1,
|
| 2748 |
"group_size": 4055,
|
| 2749 |
"in_features": 4096,
|
|
|
|
| 2767 |
"model.layers.23.mlp.up_proj": {
|
| 2768 |
"bias": false,
|
| 2769 |
"enable_norm": true,
|
| 2770 |
+
"enable_perm": false,
|
| 2771 |
"group_num": 1,
|
| 2772 |
"group_size": 4055,
|
| 2773 |
"in_features": 4096,
|
|
|
|
| 2791 |
"model.layers.23.self_attn.k_proj": {
|
| 2792 |
"bias": false,
|
| 2793 |
"enable_norm": true,
|
| 2794 |
+
"enable_perm": false,
|
| 2795 |
"group_num": 1,
|
| 2796 |
"group_size": 4055,
|
| 2797 |
"in_features": 4096,
|
|
|
|
| 2815 |
"model.layers.23.self_attn.o_proj": {
|
| 2816 |
"bias": false,
|
| 2817 |
"enable_norm": true,
|
| 2818 |
+
"enable_perm": false,
|
| 2819 |
"group_num": 1,
|
| 2820 |
"group_size": 4055,
|
| 2821 |
"in_features": 4096,
|
|
|
|
| 2839 |
"model.layers.23.self_attn.q_proj": {
|
| 2840 |
"bias": false,
|
| 2841 |
"enable_norm": true,
|
| 2842 |
+
"enable_perm": false,
|
| 2843 |
"group_num": 1,
|
| 2844 |
"group_size": 4055,
|
| 2845 |
"in_features": 4096,
|
|
|
|
| 2863 |
"model.layers.23.self_attn.v_proj": {
|
| 2864 |
"bias": false,
|
| 2865 |
"enable_norm": true,
|
| 2866 |
+
"enable_perm": false,
|
| 2867 |
"group_num": 1,
|
| 2868 |
"group_size": 4055,
|
| 2869 |
"in_features": 4096,
|
|
|
|
| 2887 |
"model.layers.24.mlp.down_proj": {
|
| 2888 |
"bias": false,
|
| 2889 |
"enable_norm": true,
|
| 2890 |
+
"enable_perm": false,
|
| 2891 |
"group_num": 1,
|
| 2892 |
"group_size": 14192,
|
| 2893 |
"in_features": 14336,
|
|
|
|
| 2911 |
"model.layers.24.mlp.gate_proj": {
|
| 2912 |
"bias": false,
|
| 2913 |
"enable_norm": true,
|
| 2914 |
+
"enable_perm": false,
|
| 2915 |
"group_num": 1,
|
| 2916 |
"group_size": 4055,
|
| 2917 |
"in_features": 4096,
|
|
|
|
| 2935 |
"model.layers.24.mlp.up_proj": {
|
| 2936 |
"bias": false,
|
| 2937 |
"enable_norm": true,
|
| 2938 |
+
"enable_perm": false,
|
| 2939 |
"group_num": 1,
|
| 2940 |
"group_size": 4055,
|
| 2941 |
"in_features": 4096,
|
|
|
|
| 2959 |
"model.layers.24.self_attn.k_proj": {
|
| 2960 |
"bias": false,
|
| 2961 |
"enable_norm": true,
|
| 2962 |
+
"enable_perm": false,
|
| 2963 |
"group_num": 1,
|
| 2964 |
"group_size": 4055,
|
| 2965 |
"in_features": 4096,
|
|
|
|
| 2983 |
"model.layers.24.self_attn.o_proj": {
|
| 2984 |
"bias": false,
|
| 2985 |
"enable_norm": true,
|
| 2986 |
+
"enable_perm": false,
|
| 2987 |
"group_num": 1,
|
| 2988 |
"group_size": 4055,
|
| 2989 |
"in_features": 4096,
|
|
|
|
| 3007 |
"model.layers.24.self_attn.q_proj": {
|
| 3008 |
"bias": false,
|
| 3009 |
"enable_norm": true,
|
| 3010 |
+
"enable_perm": false,
|
| 3011 |
"group_num": 1,
|
| 3012 |
"group_size": 4055,
|
| 3013 |
"in_features": 4096,
|
|
|
|
| 3031 |
"model.layers.24.self_attn.v_proj": {
|
| 3032 |
"bias": false,
|
| 3033 |
"enable_norm": true,
|
| 3034 |
+
"enable_perm": false,
|
| 3035 |
"group_num": 1,
|
| 3036 |
"group_size": 4055,
|
| 3037 |
"in_features": 4096,
|
|
|
|
| 3055 |
"model.layers.25.mlp.down_proj": {
|
| 3056 |
"bias": false,
|
| 3057 |
"enable_norm": true,
|
| 3058 |
+
"enable_perm": false,
|
| 3059 |
"group_num": 1,
|
| 3060 |
"group_size": 14192,
|
| 3061 |
"in_features": 14336,
|
|
|
|
| 3079 |
"model.layers.25.mlp.gate_proj": {
|
| 3080 |
"bias": false,
|
| 3081 |
"enable_norm": true,
|
| 3082 |
+
"enable_perm": false,
|
| 3083 |
"group_num": 1,
|
| 3084 |
"group_size": 4055,
|
| 3085 |
"in_features": 4096,
|
|
|
|
| 3103 |
"model.layers.25.mlp.up_proj": {
|
| 3104 |
"bias": false,
|
| 3105 |
"enable_norm": true,
|
| 3106 |
+
"enable_perm": false,
|
| 3107 |
"group_num": 1,
|
| 3108 |
"group_size": 4055,
|
| 3109 |
"in_features": 4096,
|
|
|
|
| 3127 |
"model.layers.25.self_attn.k_proj": {
|
| 3128 |
"bias": false,
|
| 3129 |
"enable_norm": true,
|
| 3130 |
+
"enable_perm": false,
|
| 3131 |
"group_num": 1,
|
| 3132 |
"group_size": 4055,
|
| 3133 |
"in_features": 4096,
|
|
|
|
| 3151 |
"model.layers.25.self_attn.o_proj": {
|
| 3152 |
"bias": false,
|
| 3153 |
"enable_norm": true,
|
| 3154 |
+
"enable_perm": false,
|
| 3155 |
"group_num": 1,
|
| 3156 |
"group_size": 4055,
|
| 3157 |
"in_features": 4096,
|
|
|
|
| 3175 |
"model.layers.25.self_attn.q_proj": {
|
| 3176 |
"bias": false,
|
| 3177 |
"enable_norm": true,
|
| 3178 |
+
"enable_perm": false,
|
| 3179 |
"group_num": 1,
|
| 3180 |
"group_size": 4055,
|
| 3181 |
"in_features": 4096,
|
|
|
|
| 3199 |
"model.layers.25.self_attn.v_proj": {
|
| 3200 |
"bias": false,
|
| 3201 |
"enable_norm": true,
|
| 3202 |
+
"enable_perm": false,
|
| 3203 |
"group_num": 1,
|
| 3204 |
"group_size": 4055,
|
| 3205 |
"in_features": 4096,
|
|
|
|
| 3223 |
"model.layers.26.mlp.down_proj": {
|
| 3224 |
"bias": false,
|
| 3225 |
"enable_norm": true,
|
| 3226 |
+
"enable_perm": false,
|
| 3227 |
"group_num": 1,
|
| 3228 |
"group_size": 14192,
|
| 3229 |
"in_features": 14336,
|
|
|
|
| 3247 |
"model.layers.26.mlp.gate_proj": {
|
| 3248 |
"bias": false,
|
| 3249 |
"enable_norm": true,
|
| 3250 |
+
"enable_perm": false,
|
| 3251 |
"group_num": 1,
|
| 3252 |
"group_size": 4055,
|
| 3253 |
"in_features": 4096,
|
|
|
|
| 3271 |
"model.layers.26.mlp.up_proj": {
|
| 3272 |
"bias": false,
|
| 3273 |
"enable_norm": true,
|
| 3274 |
+
"enable_perm": false,
|
| 3275 |
"group_num": 1,
|
| 3276 |
"group_size": 4055,
|
| 3277 |
"in_features": 4096,
|
|
|
|
| 3295 |
"model.layers.26.self_attn.k_proj": {
|
| 3296 |
"bias": false,
|
| 3297 |
"enable_norm": true,
|
| 3298 |
+
"enable_perm": false,
|
| 3299 |
"group_num": 1,
|
| 3300 |
"group_size": 4055,
|
| 3301 |
"in_features": 4096,
|
|
|
|
| 3319 |
"model.layers.26.self_attn.o_proj": {
|
| 3320 |
"bias": false,
|
| 3321 |
"enable_norm": true,
|
| 3322 |
+
"enable_perm": false,
|
| 3323 |
"group_num": 1,
|
| 3324 |
"group_size": 4055,
|
| 3325 |
"in_features": 4096,
|
|
|
|
| 3343 |
"model.layers.26.self_attn.q_proj": {
|
| 3344 |
"bias": false,
|
| 3345 |
"enable_norm": true,
|
| 3346 |
+
"enable_perm": false,
|
| 3347 |
"group_num": 1,
|
| 3348 |
"group_size": 4055,
|
| 3349 |
"in_features": 4096,
|
|
|
|
| 3367 |
"model.layers.26.self_attn.v_proj": {
|
| 3368 |
"bias": false,
|
| 3369 |
"enable_norm": true,
|
| 3370 |
+
"enable_perm": false,
|
| 3371 |
"group_num": 1,
|
| 3372 |
"group_size": 4055,
|
| 3373 |
"in_features": 4096,
|
|
|
|
| 3391 |
"model.layers.27.mlp.down_proj": {
|
| 3392 |
"bias": false,
|
| 3393 |
"enable_norm": true,
|
| 3394 |
+
"enable_perm": false,
|
| 3395 |
"group_num": 1,
|
| 3396 |
"group_size": 14192,
|
| 3397 |
"in_features": 14336,
|
|
|
|
| 3415 |
"model.layers.27.mlp.gate_proj": {
|
| 3416 |
"bias": false,
|
| 3417 |
"enable_norm": true,
|
| 3418 |
+
"enable_perm": false,
|
| 3419 |
"group_num": 1,
|
| 3420 |
"group_size": 4055,
|
| 3421 |
"in_features": 4096,
|
|
|
|
| 3439 |
"model.layers.27.mlp.up_proj": {
|
| 3440 |
"bias": false,
|
| 3441 |
"enable_norm": true,
|
| 3442 |
+
"enable_perm": false,
|
| 3443 |
"group_num": 1,
|
| 3444 |
"group_size": 4055,
|
| 3445 |
"in_features": 4096,
|
|
|
|
| 3463 |
"model.layers.27.self_attn.k_proj": {
|
| 3464 |
"bias": false,
|
| 3465 |
"enable_norm": true,
|
| 3466 |
+
"enable_perm": false,
|
| 3467 |
"group_num": 1,
|
| 3468 |
"group_size": 4055,
|
| 3469 |
"in_features": 4096,
|
|
|
|
| 3487 |
"model.layers.27.self_attn.o_proj": {
|
| 3488 |
"bias": false,
|
| 3489 |
"enable_norm": true,
|
| 3490 |
+
"enable_perm": false,
|
| 3491 |
"group_num": 1,
|
| 3492 |
"group_size": 4055,
|
| 3493 |
"in_features": 4096,
|
|
|
|
| 3511 |
"model.layers.27.self_attn.q_proj": {
|
| 3512 |
"bias": false,
|
| 3513 |
"enable_norm": true,
|
| 3514 |
+
"enable_perm": false,
|
| 3515 |
"group_num": 1,
|
| 3516 |
"group_size": 4055,
|
| 3517 |
"in_features": 4096,
|
|
|
|
| 3535 |
"model.layers.27.self_attn.v_proj": {
|
| 3536 |
"bias": false,
|
| 3537 |
"enable_norm": true,
|
| 3538 |
+
"enable_perm": false,
|
| 3539 |
"group_num": 1,
|
| 3540 |
"group_size": 4055,
|
| 3541 |
"in_features": 4096,
|
|
|
|
| 3559 |
"model.layers.28.mlp.down_proj": {
|
| 3560 |
"bias": false,
|
| 3561 |
"enable_norm": true,
|
| 3562 |
+
"enable_perm": false,
|
| 3563 |
"group_num": 1,
|
| 3564 |
"group_size": 14192,
|
| 3565 |
"in_features": 14336,
|
|
|
|
| 3583 |
"model.layers.28.mlp.gate_proj": {
|
| 3584 |
"bias": false,
|
| 3585 |
"enable_norm": true,
|
| 3586 |
+
"enable_perm": false,
|
| 3587 |
"group_num": 1,
|
| 3588 |
"group_size": 4055,
|
| 3589 |
"in_features": 4096,
|
|
|
|
| 3607 |
"model.layers.28.mlp.up_proj": {
|
| 3608 |
"bias": false,
|
| 3609 |
"enable_norm": true,
|
| 3610 |
+
"enable_perm": false,
|
| 3611 |
"group_num": 1,
|
| 3612 |
"group_size": 4055,
|
| 3613 |
"in_features": 4096,
|
|
|
|
| 3631 |
"model.layers.28.self_attn.k_proj": {
|
| 3632 |
"bias": false,
|
| 3633 |
"enable_norm": true,
|
| 3634 |
+
"enable_perm": false,
|
| 3635 |
"group_num": 1,
|
| 3636 |
"group_size": 4055,
|
| 3637 |
"in_features": 4096,
|
|
|
|
| 3655 |
"model.layers.28.self_attn.o_proj": {
|
| 3656 |
"bias": false,
|
| 3657 |
"enable_norm": true,
|
| 3658 |
+
"enable_perm": false,
|
| 3659 |
"group_num": 1,
|
| 3660 |
"group_size": 4055,
|
| 3661 |
"in_features": 4096,
|
|
|
|
| 3679 |
"model.layers.28.self_attn.q_proj": {
|
| 3680 |
"bias": false,
|
| 3681 |
"enable_norm": true,
|
| 3682 |
+
"enable_perm": false,
|
| 3683 |
"group_num": 1,
|
| 3684 |
"group_size": 4055,
|
| 3685 |
"in_features": 4096,
|
|
|
|
| 3703 |
"model.layers.28.self_attn.v_proj": {
|
| 3704 |
"bias": false,
|
| 3705 |
"enable_norm": true,
|
| 3706 |
+
"enable_perm": false,
|
| 3707 |
"group_num": 1,
|
| 3708 |
"group_size": 4055,
|
| 3709 |
"in_features": 4096,
|
|
|
|
| 3727 |
"model.layers.29.mlp.down_proj": {
|
| 3728 |
"bias": false,
|
| 3729 |
"enable_norm": true,
|
| 3730 |
+
"enable_perm": false,
|
| 3731 |
"group_num": 1,
|
| 3732 |
"group_size": 14192,
|
| 3733 |
"in_features": 14336,
|
|
|
|
| 3751 |
"model.layers.29.mlp.gate_proj": {
|
| 3752 |
"bias": false,
|
| 3753 |
"enable_norm": true,
|
| 3754 |
+
"enable_perm": false,
|
| 3755 |
"group_num": 1,
|
| 3756 |
"group_size": 4055,
|
| 3757 |
"in_features": 4096,
|
|
|
|
| 3775 |
"model.layers.29.mlp.up_proj": {
|
| 3776 |
"bias": false,
|
| 3777 |
"enable_norm": true,
|
| 3778 |
+
"enable_perm": false,
|
| 3779 |
"group_num": 1,
|
| 3780 |
"group_size": 4055,
|
| 3781 |
"in_features": 4096,
|
|
|
|
| 3799 |
"model.layers.29.self_attn.k_proj": {
|
| 3800 |
"bias": false,
|
| 3801 |
"enable_norm": true,
|
| 3802 |
+
"enable_perm": false,
|
| 3803 |
"group_num": 1,
|
| 3804 |
"group_size": 4055,
|
| 3805 |
"in_features": 4096,
|
|
|
|
| 3823 |
"model.layers.29.self_attn.o_proj": {
|
| 3824 |
"bias": false,
|
| 3825 |
"enable_norm": true,
|
| 3826 |
+
"enable_perm": false,
|
| 3827 |
"group_num": 1,
|
| 3828 |
"group_size": 4055,
|
| 3829 |
"in_features": 4096,
|
|
|
|
| 3847 |
"model.layers.29.self_attn.q_proj": {
|
| 3848 |
"bias": false,
|
| 3849 |
"enable_norm": true,
|
| 3850 |
+
"enable_perm": false,
|
| 3851 |
"group_num": 1,
|
| 3852 |
"group_size": 4055,
|
| 3853 |
"in_features": 4096,
|
|
|
|
| 3871 |
"model.layers.29.self_attn.v_proj": {
|
| 3872 |
"bias": false,
|
| 3873 |
"enable_norm": true,
|
| 3874 |
+
"enable_perm": false,
|
| 3875 |
"group_num": 1,
|
| 3876 |
"group_size": 4055,
|
| 3877 |
"in_features": 4096,
|
|
|
|
| 3895 |
"model.layers.3.mlp.down_proj": {
|
| 3896 |
"bias": false,
|
| 3897 |
"enable_norm": true,
|
| 3898 |
+
"enable_perm": false,
|
| 3899 |
"group_num": 1,
|
| 3900 |
"group_size": 14192,
|
| 3901 |
"in_features": 14336,
|
|
|
|
| 3919 |
"model.layers.3.mlp.gate_proj": {
|
| 3920 |
"bias": false,
|
| 3921 |
"enable_norm": true,
|
| 3922 |
+
"enable_perm": false,
|
| 3923 |
"group_num": 1,
|
| 3924 |
"group_size": 4055,
|
| 3925 |
"in_features": 4096,
|
|
|
|
| 3943 |
"model.layers.3.mlp.up_proj": {
|
| 3944 |
"bias": false,
|
| 3945 |
"enable_norm": true,
|
| 3946 |
+
"enable_perm": false,
|
| 3947 |
"group_num": 1,
|
| 3948 |
"group_size": 4055,
|
| 3949 |
"in_features": 4096,
|
|
|
|
| 3967 |
"model.layers.3.self_attn.k_proj": {
|
| 3968 |
"bias": false,
|
| 3969 |
"enable_norm": true,
|
| 3970 |
+
"enable_perm": false,
|
| 3971 |
"group_num": 1,
|
| 3972 |
"group_size": 4055,
|
| 3973 |
"in_features": 4096,
|
|
|
|
| 3991 |
"model.layers.3.self_attn.o_proj": {
|
| 3992 |
"bias": false,
|
| 3993 |
"enable_norm": true,
|
| 3994 |
+
"enable_perm": false,
|
| 3995 |
"group_num": 1,
|
| 3996 |
"group_size": 4055,
|
| 3997 |
"in_features": 4096,
|
|
|
|
| 4015 |
"model.layers.3.self_attn.q_proj": {
|
| 4016 |
"bias": false,
|
| 4017 |
"enable_norm": true,
|
| 4018 |
+
"enable_perm": false,
|
| 4019 |
"group_num": 1,
|
| 4020 |
"group_size": 4055,
|
| 4021 |
"in_features": 4096,
|
|
|
|
| 4039 |
"model.layers.3.self_attn.v_proj": {
|
| 4040 |
"bias": false,
|
| 4041 |
"enable_norm": true,
|
| 4042 |
+
"enable_perm": false,
|
| 4043 |
"group_num": 1,
|
| 4044 |
"group_size": 4055,
|
| 4045 |
"in_features": 4096,
|
|
|
|
| 4063 |
"model.layers.30.mlp.down_proj": {
|
| 4064 |
"bias": false,
|
| 4065 |
"enable_norm": true,
|
| 4066 |
+
"enable_perm": false,
|
| 4067 |
"group_num": 1,
|
| 4068 |
"group_size": 14192,
|
| 4069 |
"in_features": 14336,
|
|
|
|
| 4087 |
"model.layers.30.mlp.gate_proj": {
|
| 4088 |
"bias": false,
|
| 4089 |
"enable_norm": true,
|
| 4090 |
+
"enable_perm": false,
|
| 4091 |
"group_num": 1,
|
| 4092 |
"group_size": 4055,
|
| 4093 |
"in_features": 4096,
|
|
|
|
| 4111 |
"model.layers.30.mlp.up_proj": {
|
| 4112 |
"bias": false,
|
| 4113 |
"enable_norm": true,
|
| 4114 |
+
"enable_perm": false,
|
| 4115 |
"group_num": 1,
|
| 4116 |
"group_size": 4055,
|
| 4117 |
"in_features": 4096,
|
|
|
|
| 4135 |
"model.layers.30.self_attn.k_proj": {
|
| 4136 |
"bias": false,
|
| 4137 |
"enable_norm": true,
|
| 4138 |
+
"enable_perm": false,
|
| 4139 |
"group_num": 1,
|
| 4140 |
"group_size": 4055,
|
| 4141 |
"in_features": 4096,
|
|
|
|
| 4159 |
"model.layers.30.self_attn.o_proj": {
|
| 4160 |
"bias": false,
|
| 4161 |
"enable_norm": true,
|
| 4162 |
+
"enable_perm": false,
|
| 4163 |
"group_num": 1,
|
| 4164 |
"group_size": 4055,
|
| 4165 |
"in_features": 4096,
|
|
|
|
| 4183 |
"model.layers.30.self_attn.q_proj": {
|
| 4184 |
"bias": false,
|
| 4185 |
"enable_norm": true,
|
| 4186 |
+
"enable_perm": false,
|
| 4187 |
"group_num": 1,
|
| 4188 |
"group_size": 4055,
|
| 4189 |
"in_features": 4096,
|
|
|
|
| 4207 |
"model.layers.30.self_attn.v_proj": {
|
| 4208 |
"bias": false,
|
| 4209 |
"enable_norm": true,
|
| 4210 |
+
"enable_perm": false,
|
| 4211 |
"group_num": 1,
|
| 4212 |
"group_size": 4055,
|
| 4213 |
"in_features": 4096,
|
|
|
|
| 4231 |
"model.layers.31.mlp.down_proj": {
|
| 4232 |
"bias": false,
|
| 4233 |
"enable_norm": true,
|
| 4234 |
+
"enable_perm": false,
|
| 4235 |
"group_num": 1,
|
| 4236 |
"group_size": 14192,
|
| 4237 |
"in_features": 14336,
|
|
|
|
| 4255 |
"model.layers.31.mlp.gate_proj": {
|
| 4256 |
"bias": false,
|
| 4257 |
"enable_norm": true,
|
| 4258 |
+
"enable_perm": false,
|
| 4259 |
"group_num": 1,
|
| 4260 |
"group_size": 4055,
|
| 4261 |
"in_features": 4096,
|
|
|
|
| 4279 |
"model.layers.31.mlp.up_proj": {
|
| 4280 |
"bias": false,
|
| 4281 |
"enable_norm": true,
|
| 4282 |
+
"enable_perm": false,
|
| 4283 |
"group_num": 1,
|
| 4284 |
"group_size": 4055,
|
| 4285 |
"in_features": 4096,
|
|
|
|
| 4303 |
"model.layers.31.self_attn.k_proj": {
|
| 4304 |
"bias": false,
|
| 4305 |
"enable_norm": true,
|
| 4306 |
+
"enable_perm": false,
|
| 4307 |
"group_num": 1,
|
| 4308 |
"group_size": 4055,
|
| 4309 |
"in_features": 4096,
|
|
|
|
| 4327 |
"model.layers.31.self_attn.o_proj": {
|
| 4328 |
"bias": false,
|
| 4329 |
"enable_norm": true,
|
| 4330 |
+
"enable_perm": false,
|
| 4331 |
"group_num": 1,
|
| 4332 |
"group_size": 4055,
|
| 4333 |
"in_features": 4096,
|
|
|
|
| 4351 |
"model.layers.31.self_attn.q_proj": {
|
| 4352 |
"bias": false,
|
| 4353 |
"enable_norm": true,
|
| 4354 |
+
"enable_perm": false,
|
| 4355 |
"group_num": 1,
|
| 4356 |
"group_size": 4055,
|
| 4357 |
"in_features": 4096,
|
|
|
|
| 4375 |
"model.layers.31.self_attn.v_proj": {
|
| 4376 |
"bias": false,
|
| 4377 |
"enable_norm": true,
|
| 4378 |
+
"enable_perm": false,
|
| 4379 |
"group_num": 1,
|
| 4380 |
"group_size": 4055,
|
| 4381 |
"in_features": 4096,
|
|
|
|
| 4399 |
"model.layers.4.mlp.down_proj": {
|
| 4400 |
"bias": false,
|
| 4401 |
"enable_norm": true,
|
| 4402 |
+
"enable_perm": false,
|
| 4403 |
"group_num": 1,
|
| 4404 |
"group_size": 14192,
|
| 4405 |
"in_features": 14336,
|
|
|
|
| 4423 |
"model.layers.4.mlp.gate_proj": {
|
| 4424 |
"bias": false,
|
| 4425 |
"enable_norm": true,
|
| 4426 |
+
"enable_perm": false,
|
| 4427 |
"group_num": 1,
|
| 4428 |
"group_size": 4055,
|
| 4429 |
"in_features": 4096,
|
|
|
|
| 4447 |
"model.layers.4.mlp.up_proj": {
|
| 4448 |
"bias": false,
|
| 4449 |
"enable_norm": true,
|
| 4450 |
+
"enable_perm": false,
|
| 4451 |
"group_num": 1,
|
| 4452 |
"group_size": 4055,
|
| 4453 |
"in_features": 4096,
|
|
|
|
| 4471 |
"model.layers.4.self_attn.k_proj": {
|
| 4472 |
"bias": false,
|
| 4473 |
"enable_norm": true,
|
| 4474 |
+
"enable_perm": false,
|
| 4475 |
"group_num": 1,
|
| 4476 |
"group_size": 4055,
|
| 4477 |
"in_features": 4096,
|
|
|
|
| 4495 |
"model.layers.4.self_attn.o_proj": {
|
| 4496 |
"bias": false,
|
| 4497 |
"enable_norm": true,
|
| 4498 |
+
"enable_perm": false,
|
| 4499 |
"group_num": 1,
|
| 4500 |
"group_size": 4055,
|
| 4501 |
"in_features": 4096,
|
|
|
|
| 4519 |
"model.layers.4.self_attn.q_proj": {
|
| 4520 |
"bias": false,
|
| 4521 |
"enable_norm": true,
|
| 4522 |
+
"enable_perm": false,
|
| 4523 |
"group_num": 1,
|
| 4524 |
"group_size": 4055,
|
| 4525 |
"in_features": 4096,
|
|
|
|
| 4543 |
"model.layers.4.self_attn.v_proj": {
|
| 4544 |
"bias": false,
|
| 4545 |
"enable_norm": true,
|
| 4546 |
+
"enable_perm": false,
|
| 4547 |
"group_num": 1,
|
| 4548 |
"group_size": 4055,
|
| 4549 |
"in_features": 4096,
|
|
|
|
| 4567 |
"model.layers.5.mlp.down_proj": {
|
| 4568 |
"bias": false,
|
| 4569 |
"enable_norm": true,
|
| 4570 |
+
"enable_perm": false,
|
| 4571 |
"group_num": 1,
|
| 4572 |
"group_size": 14192,
|
| 4573 |
"in_features": 14336,
|
|
|
|
| 4591 |
"model.layers.5.mlp.gate_proj": {
|
| 4592 |
"bias": false,
|
| 4593 |
"enable_norm": true,
|
| 4594 |
+
"enable_perm": false,
|
| 4595 |
"group_num": 1,
|
| 4596 |
"group_size": 4055,
|
| 4597 |
"in_features": 4096,
|
|
|
|
| 4615 |
"model.layers.5.mlp.up_proj": {
|
| 4616 |
"bias": false,
|
| 4617 |
"enable_norm": true,
|
| 4618 |
+
"enable_perm": false,
|
| 4619 |
"group_num": 1,
|
| 4620 |
"group_size": 4055,
|
| 4621 |
"in_features": 4096,
|
|
|
|
| 4639 |
"model.layers.5.self_attn.k_proj": {
|
| 4640 |
"bias": false,
|
| 4641 |
"enable_norm": true,
|
| 4642 |
+
"enable_perm": false,
|
| 4643 |
"group_num": 1,
|
| 4644 |
"group_size": 4055,
|
| 4645 |
"in_features": 4096,
|
|
|
|
| 4663 |
"model.layers.5.self_attn.o_proj": {
|
| 4664 |
"bias": false,
|
| 4665 |
"enable_norm": true,
|
| 4666 |
+
"enable_perm": false,
|
| 4667 |
"group_num": 1,
|
| 4668 |
"group_size": 4055,
|
| 4669 |
"in_features": 4096,
|
|
|
|
| 4687 |
"model.layers.5.self_attn.q_proj": {
|
| 4688 |
"bias": false,
|
| 4689 |
"enable_norm": true,
|
| 4690 |
+
"enable_perm": false,
|
| 4691 |
"group_num": 1,
|
| 4692 |
"group_size": 4055,
|
| 4693 |
"in_features": 4096,
|
|
|
|
| 4711 |
"model.layers.5.self_attn.v_proj": {
|
| 4712 |
"bias": false,
|
| 4713 |
"enable_norm": true,
|
| 4714 |
+
"enable_perm": false,
|
| 4715 |
"group_num": 1,
|
| 4716 |
"group_size": 4055,
|
| 4717 |
"in_features": 4096,
|
|
|
|
| 4735 |
"model.layers.6.mlp.down_proj": {
|
| 4736 |
"bias": false,
|
| 4737 |
"enable_norm": true,
|
| 4738 |
+
"enable_perm": false,
|
| 4739 |
"group_num": 1,
|
| 4740 |
"group_size": 14192,
|
| 4741 |
"in_features": 14336,
|
|
|
|
| 4759 |
"model.layers.6.mlp.gate_proj": {
|
| 4760 |
"bias": false,
|
| 4761 |
"enable_norm": true,
|
| 4762 |
+
"enable_perm": false,
|
| 4763 |
"group_num": 1,
|
| 4764 |
"group_size": 4055,
|
| 4765 |
"in_features": 4096,
|
|
|
|
| 4783 |
"model.layers.6.mlp.up_proj": {
|
| 4784 |
"bias": false,
|
| 4785 |
"enable_norm": true,
|
| 4786 |
+
"enable_perm": false,
|
| 4787 |
"group_num": 1,
|
| 4788 |
"group_size": 4055,
|
| 4789 |
"in_features": 4096,
|
|
|
|
| 4807 |
"model.layers.6.self_attn.k_proj": {
|
| 4808 |
"bias": false,
|
| 4809 |
"enable_norm": true,
|
| 4810 |
+
"enable_perm": false,
|
| 4811 |
"group_num": 1,
|
| 4812 |
"group_size": 4055,
|
| 4813 |
"in_features": 4096,
|
|
|
|
| 4831 |
"model.layers.6.self_attn.o_proj": {
|
| 4832 |
"bias": false,
|
| 4833 |
"enable_norm": true,
|
| 4834 |
+
"enable_perm": false,
|
| 4835 |
"group_num": 1,
|
| 4836 |
"group_size": 4055,
|
| 4837 |
"in_features": 4096,
|
|
|
|
| 4855 |
"model.layers.6.self_attn.q_proj": {
|
| 4856 |
"bias": false,
|
| 4857 |
"enable_norm": true,
|
| 4858 |
+
"enable_perm": false,
|
| 4859 |
"group_num": 1,
|
| 4860 |
"group_size": 4055,
|
| 4861 |
"in_features": 4096,
|
|
|
|
| 4879 |
"model.layers.6.self_attn.v_proj": {
|
| 4880 |
"bias": false,
|
| 4881 |
"enable_norm": true,
|
| 4882 |
+
"enable_perm": false,
|
| 4883 |
"group_num": 1,
|
| 4884 |
"group_size": 4055,
|
| 4885 |
"in_features": 4096,
|
|
|
|
| 4903 |
"model.layers.7.mlp.down_proj": {
|
| 4904 |
"bias": false,
|
| 4905 |
"enable_norm": true,
|
| 4906 |
+
"enable_perm": false,
|
| 4907 |
"group_num": 1,
|
| 4908 |
"group_size": 14192,
|
| 4909 |
"in_features": 14336,
|
|
|
|
| 4927 |
"model.layers.7.mlp.gate_proj": {
|
| 4928 |
"bias": false,
|
| 4929 |
"enable_norm": true,
|
| 4930 |
+
"enable_perm": false,
|
| 4931 |
"group_num": 1,
|
| 4932 |
"group_size": 4055,
|
| 4933 |
"in_features": 4096,
|
|
|
|
| 4951 |
"model.layers.7.mlp.up_proj": {
|
| 4952 |
"bias": false,
|
| 4953 |
"enable_norm": true,
|
| 4954 |
+
"enable_perm": false,
|
| 4955 |
"group_num": 1,
|
| 4956 |
"group_size": 4055,
|
| 4957 |
"in_features": 4096,
|
|
|
|
| 4975 |
"model.layers.7.self_attn.k_proj": {
|
| 4976 |
"bias": false,
|
| 4977 |
"enable_norm": true,
|
| 4978 |
+
"enable_perm": false,
|
| 4979 |
"group_num": 1,
|
| 4980 |
"group_size": 4055,
|
| 4981 |
"in_features": 4096,
|
|
|
|
| 4999 |
"model.layers.7.self_attn.o_proj": {
|
| 5000 |
"bias": false,
|
| 5001 |
"enable_norm": true,
|
| 5002 |
+
"enable_perm": false,
|
| 5003 |
"group_num": 1,
|
| 5004 |
"group_size": 4055,
|
| 5005 |
"in_features": 4096,
|
|
|
|
| 5023 |
"model.layers.7.self_attn.q_proj": {
|
| 5024 |
"bias": false,
|
| 5025 |
"enable_norm": true,
|
| 5026 |
+
"enable_perm": false,
|
| 5027 |
"group_num": 1,
|
| 5028 |
"group_size": 4055,
|
| 5029 |
"in_features": 4096,
|
|
|
|
| 5047 |
"model.layers.7.self_attn.v_proj": {
|
| 5048 |
"bias": false,
|
| 5049 |
"enable_norm": true,
|
| 5050 |
+
"enable_perm": false,
|
| 5051 |
"group_num": 1,
|
| 5052 |
"group_size": 4055,
|
| 5053 |
"in_features": 4096,
|
|
|
|
| 5071 |
"model.layers.8.mlp.down_proj": {
|
| 5072 |
"bias": false,
|
| 5073 |
"enable_norm": true,
|
| 5074 |
+
"enable_perm": false,
|
| 5075 |
"group_num": 1,
|
| 5076 |
"group_size": 14192,
|
| 5077 |
"in_features": 14336,
|
|
|
|
| 5095 |
"model.layers.8.mlp.gate_proj": {
|
| 5096 |
"bias": false,
|
| 5097 |
"enable_norm": true,
|
| 5098 |
+
"enable_perm": false,
|
| 5099 |
"group_num": 1,
|
| 5100 |
"group_size": 4055,
|
| 5101 |
"in_features": 4096,
|
|
|
|
| 5119 |
"model.layers.8.mlp.up_proj": {
|
| 5120 |
"bias": false,
|
| 5121 |
"enable_norm": true,
|
| 5122 |
+
"enable_perm": false,
|
| 5123 |
"group_num": 1,
|
| 5124 |
"group_size": 4055,
|
| 5125 |
"in_features": 4096,
|
|
|
|
| 5143 |
"model.layers.8.self_attn.k_proj": {
|
| 5144 |
"bias": false,
|
| 5145 |
"enable_norm": true,
|
| 5146 |
+
"enable_perm": false,
|
| 5147 |
"group_num": 1,
|
| 5148 |
"group_size": 4055,
|
| 5149 |
"in_features": 4096,
|
|
|
|
| 5167 |
"model.layers.8.self_attn.o_proj": {
|
| 5168 |
"bias": false,
|
| 5169 |
"enable_norm": true,
|
| 5170 |
+
"enable_perm": false,
|
| 5171 |
"group_num": 1,
|
| 5172 |
"group_size": 4055,
|
| 5173 |
"in_features": 4096,
|
|
|
|
| 5191 |
"model.layers.8.self_attn.q_proj": {
|
| 5192 |
"bias": false,
|
| 5193 |
"enable_norm": true,
|
| 5194 |
+
"enable_perm": false,
|
| 5195 |
"group_num": 1,
|
| 5196 |
"group_size": 4055,
|
| 5197 |
"in_features": 4096,
|
|
|
|
| 5215 |
"model.layers.8.self_attn.v_proj": {
|
| 5216 |
"bias": false,
|
| 5217 |
"enable_norm": true,
|
| 5218 |
+
"enable_perm": false,
|
| 5219 |
"group_num": 1,
|
| 5220 |
"group_size": 4055,
|
| 5221 |
"in_features": 4096,
|
|
|
|
| 5239 |
"model.layers.9.mlp.down_proj": {
|
| 5240 |
"bias": false,
|
| 5241 |
"enable_norm": true,
|
| 5242 |
+
"enable_perm": false,
|
| 5243 |
"group_num": 1,
|
| 5244 |
"group_size": 14192,
|
| 5245 |
"in_features": 14336,
|
|
|
|
| 5263 |
"model.layers.9.mlp.gate_proj": {
|
| 5264 |
"bias": false,
|
| 5265 |
"enable_norm": true,
|
| 5266 |
+
"enable_perm": false,
|
| 5267 |
"group_num": 1,
|
| 5268 |
"group_size": 4055,
|
| 5269 |
"in_features": 4096,
|
|
|
|
| 5287 |
"model.layers.9.mlp.up_proj": {
|
| 5288 |
"bias": false,
|
| 5289 |
"enable_norm": true,
|
| 5290 |
+
"enable_perm": false,
|
| 5291 |
"group_num": 1,
|
| 5292 |
"group_size": 4055,
|
| 5293 |
"in_features": 4096,
|
|
|
|
| 5311 |
"model.layers.9.self_attn.k_proj": {
|
| 5312 |
"bias": false,
|
| 5313 |
"enable_norm": true,
|
| 5314 |
+
"enable_perm": false,
|
| 5315 |
"group_num": 1,
|
| 5316 |
"group_size": 4055,
|
| 5317 |
"in_features": 4096,
|
|
|
|
| 5335 |
"model.layers.9.self_attn.o_proj": {
|
| 5336 |
"bias": false,
|
| 5337 |
"enable_norm": true,
|
| 5338 |
+
"enable_perm": false,
|
| 5339 |
"group_num": 1,
|
| 5340 |
"group_size": 4055,
|
| 5341 |
"in_features": 4096,
|
|
|
|
| 5359 |
"model.layers.9.self_attn.q_proj": {
|
| 5360 |
"bias": false,
|
| 5361 |
"enable_norm": true,
|
| 5362 |
+
"enable_perm": false,
|
| 5363 |
"group_num": 1,
|
| 5364 |
"group_size": 4055,
|
| 5365 |
"in_features": 4096,
|
|
|
|
| 5383 |
"model.layers.9.self_attn.v_proj": {
|
| 5384 |
"bias": false,
|
| 5385 |
"enable_norm": true,
|
| 5386 |
+
"enable_perm": false,
|
| 5387 |
"group_num": 1,
|
| 5388 |
"group_size": 4055,
|
| 5389 |
"in_features": 4096,
|