Update config.json
Browse files- config.json +224 -224
config.json
CHANGED
@@ -42,7 +42,7 @@
|
|
42 |
"model.layers.0.mlp.down_proj": {
|
43 |
"bias": false,
|
44 |
"enable_norm": true,
|
45 |
-
"enable_perm":
|
46 |
"group_num": 1,
|
47 |
"group_size": 14336,
|
48 |
"in_features": 14336,
|
@@ -66,7 +66,7 @@
|
|
66 |
"model.layers.0.mlp.gate_proj": {
|
67 |
"bias": false,
|
68 |
"enable_norm": true,
|
69 |
-
"enable_perm":
|
70 |
"group_num": 1,
|
71 |
"group_size": 4096,
|
72 |
"in_features": 4096,
|
@@ -90,7 +90,7 @@
|
|
90 |
"model.layers.0.mlp.up_proj": {
|
91 |
"bias": false,
|
92 |
"enable_norm": true,
|
93 |
-
"enable_perm":
|
94 |
"group_num": 1,
|
95 |
"group_size": 4096,
|
96 |
"in_features": 4096,
|
@@ -114,7 +114,7 @@
|
|
114 |
"model.layers.0.self_attn.k_proj": {
|
115 |
"bias": false,
|
116 |
"enable_norm": true,
|
117 |
-
"enable_perm":
|
118 |
"group_num": 1,
|
119 |
"group_size": 4096,
|
120 |
"in_features": 4096,
|
@@ -138,7 +138,7 @@
|
|
138 |
"model.layers.0.self_attn.o_proj": {
|
139 |
"bias": false,
|
140 |
"enable_norm": true,
|
141 |
-
"enable_perm":
|
142 |
"group_num": 1,
|
143 |
"group_size": 4096,
|
144 |
"in_features": 4096,
|
@@ -162,7 +162,7 @@
|
|
162 |
"model.layers.0.self_attn.q_proj": {
|
163 |
"bias": false,
|
164 |
"enable_norm": true,
|
165 |
-
"enable_perm":
|
166 |
"group_num": 1,
|
167 |
"group_size": 4096,
|
168 |
"in_features": 4096,
|
@@ -186,7 +186,7 @@
|
|
186 |
"model.layers.0.self_attn.v_proj": {
|
187 |
"bias": false,
|
188 |
"enable_norm": true,
|
189 |
-
"enable_perm":
|
190 |
"group_num": 1,
|
191 |
"group_size": 4096,
|
192 |
"in_features": 4096,
|
@@ -210,7 +210,7 @@
|
|
210 |
"model.layers.1.mlp.down_proj": {
|
211 |
"bias": false,
|
212 |
"enable_norm": true,
|
213 |
-
"enable_perm":
|
214 |
"group_num": 1,
|
215 |
"group_size": 14336,
|
216 |
"in_features": 14336,
|
@@ -234,7 +234,7 @@
|
|
234 |
"model.layers.1.mlp.gate_proj": {
|
235 |
"bias": false,
|
236 |
"enable_norm": true,
|
237 |
-
"enable_perm":
|
238 |
"group_num": 1,
|
239 |
"group_size": 4096,
|
240 |
"in_features": 4096,
|
@@ -258,7 +258,7 @@
|
|
258 |
"model.layers.1.mlp.up_proj": {
|
259 |
"bias": false,
|
260 |
"enable_norm": true,
|
261 |
-
"enable_perm":
|
262 |
"group_num": 1,
|
263 |
"group_size": 4096,
|
264 |
"in_features": 4096,
|
@@ -282,7 +282,7 @@
|
|
282 |
"model.layers.1.self_attn.k_proj": {
|
283 |
"bias": false,
|
284 |
"enable_norm": true,
|
285 |
-
"enable_perm":
|
286 |
"group_num": 1,
|
287 |
"group_size": 4096,
|
288 |
"in_features": 4096,
|
@@ -306,7 +306,7 @@
|
|
306 |
"model.layers.1.self_attn.o_proj": {
|
307 |
"bias": false,
|
308 |
"enable_norm": true,
|
309 |
-
"enable_perm":
|
310 |
"group_num": 1,
|
311 |
"group_size": 4096,
|
312 |
"in_features": 4096,
|
@@ -330,7 +330,7 @@
|
|
330 |
"model.layers.1.self_attn.q_proj": {
|
331 |
"bias": false,
|
332 |
"enable_norm": true,
|
333 |
-
"enable_perm":
|
334 |
"group_num": 1,
|
335 |
"group_size": 4096,
|
336 |
"in_features": 4096,
|
@@ -354,7 +354,7 @@
|
|
354 |
"model.layers.1.self_attn.v_proj": {
|
355 |
"bias": false,
|
356 |
"enable_norm": true,
|
357 |
-
"enable_perm":
|
358 |
"group_num": 1,
|
359 |
"group_size": 4096,
|
360 |
"in_features": 4096,
|
@@ -378,7 +378,7 @@
|
|
378 |
"model.layers.10.mlp.down_proj": {
|
379 |
"bias": false,
|
380 |
"enable_norm": true,
|
381 |
-
"enable_perm":
|
382 |
"group_num": 1,
|
383 |
"group_size": 14336,
|
384 |
"in_features": 14336,
|
@@ -402,7 +402,7 @@
|
|
402 |
"model.layers.10.mlp.gate_proj": {
|
403 |
"bias": false,
|
404 |
"enable_norm": true,
|
405 |
-
"enable_perm":
|
406 |
"group_num": 1,
|
407 |
"group_size": 4096,
|
408 |
"in_features": 4096,
|
@@ -426,7 +426,7 @@
|
|
426 |
"model.layers.10.mlp.up_proj": {
|
427 |
"bias": false,
|
428 |
"enable_norm": true,
|
429 |
-
"enable_perm":
|
430 |
"group_num": 1,
|
431 |
"group_size": 4096,
|
432 |
"in_features": 4096,
|
@@ -450,7 +450,7 @@
|
|
450 |
"model.layers.10.self_attn.k_proj": {
|
451 |
"bias": false,
|
452 |
"enable_norm": true,
|
453 |
-
"enable_perm":
|
454 |
"group_num": 1,
|
455 |
"group_size": 4096,
|
456 |
"in_features": 4096,
|
@@ -474,7 +474,7 @@
|
|
474 |
"model.layers.10.self_attn.o_proj": {
|
475 |
"bias": false,
|
476 |
"enable_norm": true,
|
477 |
-
"enable_perm":
|
478 |
"group_num": 1,
|
479 |
"group_size": 4096,
|
480 |
"in_features": 4096,
|
@@ -498,7 +498,7 @@
|
|
498 |
"model.layers.10.self_attn.q_proj": {
|
499 |
"bias": false,
|
500 |
"enable_norm": true,
|
501 |
-
"enable_perm":
|
502 |
"group_num": 1,
|
503 |
"group_size": 4096,
|
504 |
"in_features": 4096,
|
@@ -522,7 +522,7 @@
|
|
522 |
"model.layers.10.self_attn.v_proj": {
|
523 |
"bias": false,
|
524 |
"enable_norm": true,
|
525 |
-
"enable_perm":
|
526 |
"group_num": 1,
|
527 |
"group_size": 4096,
|
528 |
"in_features": 4096,
|
@@ -546,7 +546,7 @@
|
|
546 |
"model.layers.11.mlp.down_proj": {
|
547 |
"bias": false,
|
548 |
"enable_norm": true,
|
549 |
-
"enable_perm":
|
550 |
"group_num": 1,
|
551 |
"group_size": 14336,
|
552 |
"in_features": 14336,
|
@@ -570,7 +570,7 @@
|
|
570 |
"model.layers.11.mlp.gate_proj": {
|
571 |
"bias": false,
|
572 |
"enable_norm": true,
|
573 |
-
"enable_perm":
|
574 |
"group_num": 1,
|
575 |
"group_size": 4096,
|
576 |
"in_features": 4096,
|
@@ -594,7 +594,7 @@
|
|
594 |
"model.layers.11.mlp.up_proj": {
|
595 |
"bias": false,
|
596 |
"enable_norm": true,
|
597 |
-
"enable_perm":
|
598 |
"group_num": 1,
|
599 |
"group_size": 4096,
|
600 |
"in_features": 4096,
|
@@ -618,7 +618,7 @@
|
|
618 |
"model.layers.11.self_attn.k_proj": {
|
619 |
"bias": false,
|
620 |
"enable_norm": true,
|
621 |
-
"enable_perm":
|
622 |
"group_num": 1,
|
623 |
"group_size": 4096,
|
624 |
"in_features": 4096,
|
@@ -642,7 +642,7 @@
|
|
642 |
"model.layers.11.self_attn.o_proj": {
|
643 |
"bias": false,
|
644 |
"enable_norm": true,
|
645 |
-
"enable_perm":
|
646 |
"group_num": 1,
|
647 |
"group_size": 4096,
|
648 |
"in_features": 4096,
|
@@ -666,7 +666,7 @@
|
|
666 |
"model.layers.11.self_attn.q_proj": {
|
667 |
"bias": false,
|
668 |
"enable_norm": true,
|
669 |
-
"enable_perm":
|
670 |
"group_num": 1,
|
671 |
"group_size": 4096,
|
672 |
"in_features": 4096,
|
@@ -690,7 +690,7 @@
|
|
690 |
"model.layers.11.self_attn.v_proj": {
|
691 |
"bias": false,
|
692 |
"enable_norm": true,
|
693 |
-
"enable_perm":
|
694 |
"group_num": 1,
|
695 |
"group_size": 4096,
|
696 |
"in_features": 4096,
|
@@ -714,7 +714,7 @@
|
|
714 |
"model.layers.12.mlp.down_proj": {
|
715 |
"bias": false,
|
716 |
"enable_norm": true,
|
717 |
-
"enable_perm":
|
718 |
"group_num": 1,
|
719 |
"group_size": 14336,
|
720 |
"in_features": 14336,
|
@@ -738,7 +738,7 @@
|
|
738 |
"model.layers.12.mlp.gate_proj": {
|
739 |
"bias": false,
|
740 |
"enable_norm": true,
|
741 |
-
"enable_perm":
|
742 |
"group_num": 1,
|
743 |
"group_size": 4096,
|
744 |
"in_features": 4096,
|
@@ -762,7 +762,7 @@
|
|
762 |
"model.layers.12.mlp.up_proj": {
|
763 |
"bias": false,
|
764 |
"enable_norm": true,
|
765 |
-
"enable_perm":
|
766 |
"group_num": 1,
|
767 |
"group_size": 4096,
|
768 |
"in_features": 4096,
|
@@ -786,7 +786,7 @@
|
|
786 |
"model.layers.12.self_attn.k_proj": {
|
787 |
"bias": false,
|
788 |
"enable_norm": true,
|
789 |
-
"enable_perm":
|
790 |
"group_num": 1,
|
791 |
"group_size": 4096,
|
792 |
"in_features": 4096,
|
@@ -810,7 +810,7 @@
|
|
810 |
"model.layers.12.self_attn.o_proj": {
|
811 |
"bias": false,
|
812 |
"enable_norm": true,
|
813 |
-
"enable_perm":
|
814 |
"group_num": 1,
|
815 |
"group_size": 4096,
|
816 |
"in_features": 4096,
|
@@ -834,7 +834,7 @@
|
|
834 |
"model.layers.12.self_attn.q_proj": {
|
835 |
"bias": false,
|
836 |
"enable_norm": true,
|
837 |
-
"enable_perm":
|
838 |
"group_num": 1,
|
839 |
"group_size": 4096,
|
840 |
"in_features": 4096,
|
@@ -858,7 +858,7 @@
|
|
858 |
"model.layers.12.self_attn.v_proj": {
|
859 |
"bias": false,
|
860 |
"enable_norm": true,
|
861 |
-
"enable_perm":
|
862 |
"group_num": 1,
|
863 |
"group_size": 4096,
|
864 |
"in_features": 4096,
|
@@ -882,7 +882,7 @@
|
|
882 |
"model.layers.13.mlp.down_proj": {
|
883 |
"bias": false,
|
884 |
"enable_norm": true,
|
885 |
-
"enable_perm":
|
886 |
"group_num": 1,
|
887 |
"group_size": 14336,
|
888 |
"in_features": 14336,
|
@@ -906,7 +906,7 @@
|
|
906 |
"model.layers.13.mlp.gate_proj": {
|
907 |
"bias": false,
|
908 |
"enable_norm": true,
|
909 |
-
"enable_perm":
|
910 |
"group_num": 1,
|
911 |
"group_size": 4096,
|
912 |
"in_features": 4096,
|
@@ -930,7 +930,7 @@
|
|
930 |
"model.layers.13.mlp.up_proj": {
|
931 |
"bias": false,
|
932 |
"enable_norm": true,
|
933 |
-
"enable_perm":
|
934 |
"group_num": 1,
|
935 |
"group_size": 4096,
|
936 |
"in_features": 4096,
|
@@ -954,7 +954,7 @@
|
|
954 |
"model.layers.13.self_attn.k_proj": {
|
955 |
"bias": false,
|
956 |
"enable_norm": true,
|
957 |
-
"enable_perm":
|
958 |
"group_num": 1,
|
959 |
"group_size": 4096,
|
960 |
"in_features": 4096,
|
@@ -978,7 +978,7 @@
|
|
978 |
"model.layers.13.self_attn.o_proj": {
|
979 |
"bias": false,
|
980 |
"enable_norm": true,
|
981 |
-
"enable_perm":
|
982 |
"group_num": 1,
|
983 |
"group_size": 4096,
|
984 |
"in_features": 4096,
|
@@ -1002,7 +1002,7 @@
|
|
1002 |
"model.layers.13.self_attn.q_proj": {
|
1003 |
"bias": false,
|
1004 |
"enable_norm": true,
|
1005 |
-
"enable_perm":
|
1006 |
"group_num": 1,
|
1007 |
"group_size": 4096,
|
1008 |
"in_features": 4096,
|
@@ -1026,7 +1026,7 @@
|
|
1026 |
"model.layers.13.self_attn.v_proj": {
|
1027 |
"bias": false,
|
1028 |
"enable_norm": true,
|
1029 |
-
"enable_perm":
|
1030 |
"group_num": 1,
|
1031 |
"group_size": 4096,
|
1032 |
"in_features": 4096,
|
@@ -1050,7 +1050,7 @@
|
|
1050 |
"model.layers.14.mlp.down_proj": {
|
1051 |
"bias": false,
|
1052 |
"enable_norm": true,
|
1053 |
-
"enable_perm":
|
1054 |
"group_num": 1,
|
1055 |
"group_size": 14336,
|
1056 |
"in_features": 14336,
|
@@ -1074,7 +1074,7 @@
|
|
1074 |
"model.layers.14.mlp.gate_proj": {
|
1075 |
"bias": false,
|
1076 |
"enable_norm": true,
|
1077 |
-
"enable_perm":
|
1078 |
"group_num": 1,
|
1079 |
"group_size": 4096,
|
1080 |
"in_features": 4096,
|
@@ -1098,7 +1098,7 @@
|
|
1098 |
"model.layers.14.mlp.up_proj": {
|
1099 |
"bias": false,
|
1100 |
"enable_norm": true,
|
1101 |
-
"enable_perm":
|
1102 |
"group_num": 1,
|
1103 |
"group_size": 4096,
|
1104 |
"in_features": 4096,
|
@@ -1122,7 +1122,7 @@
|
|
1122 |
"model.layers.14.self_attn.k_proj": {
|
1123 |
"bias": false,
|
1124 |
"enable_norm": true,
|
1125 |
-
"enable_perm":
|
1126 |
"group_num": 1,
|
1127 |
"group_size": 4096,
|
1128 |
"in_features": 4096,
|
@@ -1146,7 +1146,7 @@
|
|
1146 |
"model.layers.14.self_attn.o_proj": {
|
1147 |
"bias": false,
|
1148 |
"enable_norm": true,
|
1149 |
-
"enable_perm":
|
1150 |
"group_num": 1,
|
1151 |
"group_size": 4096,
|
1152 |
"in_features": 4096,
|
@@ -1170,7 +1170,7 @@
|
|
1170 |
"model.layers.14.self_attn.q_proj": {
|
1171 |
"bias": false,
|
1172 |
"enable_norm": true,
|
1173 |
-
"enable_perm":
|
1174 |
"group_num": 1,
|
1175 |
"group_size": 4096,
|
1176 |
"in_features": 4096,
|
@@ -1194,7 +1194,7 @@
|
|
1194 |
"model.layers.14.self_attn.v_proj": {
|
1195 |
"bias": false,
|
1196 |
"enable_norm": true,
|
1197 |
-
"enable_perm":
|
1198 |
"group_num": 1,
|
1199 |
"group_size": 4096,
|
1200 |
"in_features": 4096,
|
@@ -1218,7 +1218,7 @@
|
|
1218 |
"model.layers.15.mlp.down_proj": {
|
1219 |
"bias": false,
|
1220 |
"enable_norm": true,
|
1221 |
-
"enable_perm":
|
1222 |
"group_num": 1,
|
1223 |
"group_size": 14336,
|
1224 |
"in_features": 14336,
|
@@ -1242,7 +1242,7 @@
|
|
1242 |
"model.layers.15.mlp.gate_proj": {
|
1243 |
"bias": false,
|
1244 |
"enable_norm": true,
|
1245 |
-
"enable_perm":
|
1246 |
"group_num": 1,
|
1247 |
"group_size": 4096,
|
1248 |
"in_features": 4096,
|
@@ -1266,7 +1266,7 @@
|
|
1266 |
"model.layers.15.mlp.up_proj": {
|
1267 |
"bias": false,
|
1268 |
"enable_norm": true,
|
1269 |
-
"enable_perm":
|
1270 |
"group_num": 1,
|
1271 |
"group_size": 4096,
|
1272 |
"in_features": 4096,
|
@@ -1290,7 +1290,7 @@
|
|
1290 |
"model.layers.15.self_attn.k_proj": {
|
1291 |
"bias": false,
|
1292 |
"enable_norm": true,
|
1293 |
-
"enable_perm":
|
1294 |
"group_num": 1,
|
1295 |
"group_size": 4096,
|
1296 |
"in_features": 4096,
|
@@ -1314,7 +1314,7 @@
|
|
1314 |
"model.layers.15.self_attn.o_proj": {
|
1315 |
"bias": false,
|
1316 |
"enable_norm": true,
|
1317 |
-
"enable_perm":
|
1318 |
"group_num": 1,
|
1319 |
"group_size": 4096,
|
1320 |
"in_features": 4096,
|
@@ -1338,7 +1338,7 @@
|
|
1338 |
"model.layers.15.self_attn.q_proj": {
|
1339 |
"bias": false,
|
1340 |
"enable_norm": true,
|
1341 |
-
"enable_perm":
|
1342 |
"group_num": 1,
|
1343 |
"group_size": 4096,
|
1344 |
"in_features": 4096,
|
@@ -1362,7 +1362,7 @@
|
|
1362 |
"model.layers.15.self_attn.v_proj": {
|
1363 |
"bias": false,
|
1364 |
"enable_norm": true,
|
1365 |
-
"enable_perm":
|
1366 |
"group_num": 1,
|
1367 |
"group_size": 4096,
|
1368 |
"in_features": 4096,
|
@@ -1386,7 +1386,7 @@
|
|
1386 |
"model.layers.16.mlp.down_proj": {
|
1387 |
"bias": false,
|
1388 |
"enable_norm": true,
|
1389 |
-
"enable_perm":
|
1390 |
"group_num": 1,
|
1391 |
"group_size": 14336,
|
1392 |
"in_features": 14336,
|
@@ -1410,7 +1410,7 @@
|
|
1410 |
"model.layers.16.mlp.gate_proj": {
|
1411 |
"bias": false,
|
1412 |
"enable_norm": true,
|
1413 |
-
"enable_perm":
|
1414 |
"group_num": 1,
|
1415 |
"group_size": 4096,
|
1416 |
"in_features": 4096,
|
@@ -1434,7 +1434,7 @@
|
|
1434 |
"model.layers.16.mlp.up_proj": {
|
1435 |
"bias": false,
|
1436 |
"enable_norm": true,
|
1437 |
-
"enable_perm":
|
1438 |
"group_num": 1,
|
1439 |
"group_size": 4096,
|
1440 |
"in_features": 4096,
|
@@ -1458,7 +1458,7 @@
|
|
1458 |
"model.layers.16.self_attn.k_proj": {
|
1459 |
"bias": false,
|
1460 |
"enable_norm": true,
|
1461 |
-
"enable_perm":
|
1462 |
"group_num": 1,
|
1463 |
"group_size": 4096,
|
1464 |
"in_features": 4096,
|
@@ -1482,7 +1482,7 @@
|
|
1482 |
"model.layers.16.self_attn.o_proj": {
|
1483 |
"bias": false,
|
1484 |
"enable_norm": true,
|
1485 |
-
"enable_perm":
|
1486 |
"group_num": 1,
|
1487 |
"group_size": 4096,
|
1488 |
"in_features": 4096,
|
@@ -1506,7 +1506,7 @@
|
|
1506 |
"model.layers.16.self_attn.q_proj": {
|
1507 |
"bias": false,
|
1508 |
"enable_norm": true,
|
1509 |
-
"enable_perm":
|
1510 |
"group_num": 1,
|
1511 |
"group_size": 4096,
|
1512 |
"in_features": 4096,
|
@@ -1530,7 +1530,7 @@
|
|
1530 |
"model.layers.16.self_attn.v_proj": {
|
1531 |
"bias": false,
|
1532 |
"enable_norm": true,
|
1533 |
-
"enable_perm":
|
1534 |
"group_num": 1,
|
1535 |
"group_size": 4096,
|
1536 |
"in_features": 4096,
|
@@ -1554,7 +1554,7 @@
|
|
1554 |
"model.layers.17.mlp.down_proj": {
|
1555 |
"bias": false,
|
1556 |
"enable_norm": true,
|
1557 |
-
"enable_perm":
|
1558 |
"group_num": 1,
|
1559 |
"group_size": 14336,
|
1560 |
"in_features": 14336,
|
@@ -1578,7 +1578,7 @@
|
|
1578 |
"model.layers.17.mlp.gate_proj": {
|
1579 |
"bias": false,
|
1580 |
"enable_norm": true,
|
1581 |
-
"enable_perm":
|
1582 |
"group_num": 1,
|
1583 |
"group_size": 4096,
|
1584 |
"in_features": 4096,
|
@@ -1602,7 +1602,7 @@
|
|
1602 |
"model.layers.17.mlp.up_proj": {
|
1603 |
"bias": false,
|
1604 |
"enable_norm": true,
|
1605 |
-
"enable_perm":
|
1606 |
"group_num": 1,
|
1607 |
"group_size": 4096,
|
1608 |
"in_features": 4096,
|
@@ -1626,7 +1626,7 @@
|
|
1626 |
"model.layers.17.self_attn.k_proj": {
|
1627 |
"bias": false,
|
1628 |
"enable_norm": true,
|
1629 |
-
"enable_perm":
|
1630 |
"group_num": 1,
|
1631 |
"group_size": 4096,
|
1632 |
"in_features": 4096,
|
@@ -1650,7 +1650,7 @@
|
|
1650 |
"model.layers.17.self_attn.o_proj": {
|
1651 |
"bias": false,
|
1652 |
"enable_norm": true,
|
1653 |
-
"enable_perm":
|
1654 |
"group_num": 1,
|
1655 |
"group_size": 4096,
|
1656 |
"in_features": 4096,
|
@@ -1674,7 +1674,7 @@
|
|
1674 |
"model.layers.17.self_attn.q_proj": {
|
1675 |
"bias": false,
|
1676 |
"enable_norm": true,
|
1677 |
-
"enable_perm":
|
1678 |
"group_num": 1,
|
1679 |
"group_size": 4096,
|
1680 |
"in_features": 4096,
|
@@ -1698,7 +1698,7 @@
|
|
1698 |
"model.layers.17.self_attn.v_proj": {
|
1699 |
"bias": false,
|
1700 |
"enable_norm": true,
|
1701 |
-
"enable_perm":
|
1702 |
"group_num": 1,
|
1703 |
"group_size": 4096,
|
1704 |
"in_features": 4096,
|
@@ -1722,7 +1722,7 @@
|
|
1722 |
"model.layers.18.mlp.down_proj": {
|
1723 |
"bias": false,
|
1724 |
"enable_norm": true,
|
1725 |
-
"enable_perm":
|
1726 |
"group_num": 1,
|
1727 |
"group_size": 14336,
|
1728 |
"in_features": 14336,
|
@@ -1746,7 +1746,7 @@
|
|
1746 |
"model.layers.18.mlp.gate_proj": {
|
1747 |
"bias": false,
|
1748 |
"enable_norm": true,
|
1749 |
-
"enable_perm":
|
1750 |
"group_num": 1,
|
1751 |
"group_size": 4096,
|
1752 |
"in_features": 4096,
|
@@ -1770,7 +1770,7 @@
|
|
1770 |
"model.layers.18.mlp.up_proj": {
|
1771 |
"bias": false,
|
1772 |
"enable_norm": true,
|
1773 |
-
"enable_perm":
|
1774 |
"group_num": 1,
|
1775 |
"group_size": 4096,
|
1776 |
"in_features": 4096,
|
@@ -1794,7 +1794,7 @@
|
|
1794 |
"model.layers.18.self_attn.k_proj": {
|
1795 |
"bias": false,
|
1796 |
"enable_norm": true,
|
1797 |
-
"enable_perm":
|
1798 |
"group_num": 1,
|
1799 |
"group_size": 4096,
|
1800 |
"in_features": 4096,
|
@@ -1818,7 +1818,7 @@
|
|
1818 |
"model.layers.18.self_attn.o_proj": {
|
1819 |
"bias": false,
|
1820 |
"enable_norm": true,
|
1821 |
-
"enable_perm":
|
1822 |
"group_num": 1,
|
1823 |
"group_size": 4096,
|
1824 |
"in_features": 4096,
|
@@ -1842,7 +1842,7 @@
|
|
1842 |
"model.layers.18.self_attn.q_proj": {
|
1843 |
"bias": false,
|
1844 |
"enable_norm": true,
|
1845 |
-
"enable_perm":
|
1846 |
"group_num": 1,
|
1847 |
"group_size": 4096,
|
1848 |
"in_features": 4096,
|
@@ -1866,7 +1866,7 @@
|
|
1866 |
"model.layers.18.self_attn.v_proj": {
|
1867 |
"bias": false,
|
1868 |
"enable_norm": true,
|
1869 |
-
"enable_perm":
|
1870 |
"group_num": 1,
|
1871 |
"group_size": 4096,
|
1872 |
"in_features": 4096,
|
@@ -1890,7 +1890,7 @@
|
|
1890 |
"model.layers.19.mlp.down_proj": {
|
1891 |
"bias": false,
|
1892 |
"enable_norm": true,
|
1893 |
-
"enable_perm":
|
1894 |
"group_num": 1,
|
1895 |
"group_size": 14336,
|
1896 |
"in_features": 14336,
|
@@ -1914,7 +1914,7 @@
|
|
1914 |
"model.layers.19.mlp.gate_proj": {
|
1915 |
"bias": false,
|
1916 |
"enable_norm": true,
|
1917 |
-
"enable_perm":
|
1918 |
"group_num": 1,
|
1919 |
"group_size": 4096,
|
1920 |
"in_features": 4096,
|
@@ -1938,7 +1938,7 @@
|
|
1938 |
"model.layers.19.mlp.up_proj": {
|
1939 |
"bias": false,
|
1940 |
"enable_norm": true,
|
1941 |
-
"enable_perm":
|
1942 |
"group_num": 1,
|
1943 |
"group_size": 4096,
|
1944 |
"in_features": 4096,
|
@@ -1962,7 +1962,7 @@
|
|
1962 |
"model.layers.19.self_attn.k_proj": {
|
1963 |
"bias": false,
|
1964 |
"enable_norm": true,
|
1965 |
-
"enable_perm":
|
1966 |
"group_num": 1,
|
1967 |
"group_size": 4096,
|
1968 |
"in_features": 4096,
|
@@ -1986,7 +1986,7 @@
|
|
1986 |
"model.layers.19.self_attn.o_proj": {
|
1987 |
"bias": false,
|
1988 |
"enable_norm": true,
|
1989 |
-
"enable_perm":
|
1990 |
"group_num": 1,
|
1991 |
"group_size": 4096,
|
1992 |
"in_features": 4096,
|
@@ -2010,7 +2010,7 @@
|
|
2010 |
"model.layers.19.self_attn.q_proj": {
|
2011 |
"bias": false,
|
2012 |
"enable_norm": true,
|
2013 |
-
"enable_perm":
|
2014 |
"group_num": 1,
|
2015 |
"group_size": 4096,
|
2016 |
"in_features": 4096,
|
@@ -2034,7 +2034,7 @@
|
|
2034 |
"model.layers.19.self_attn.v_proj": {
|
2035 |
"bias": false,
|
2036 |
"enable_norm": true,
|
2037 |
-
"enable_perm":
|
2038 |
"group_num": 1,
|
2039 |
"group_size": 4096,
|
2040 |
"in_features": 4096,
|
@@ -2058,7 +2058,7 @@
|
|
2058 |
"model.layers.2.mlp.down_proj": {
|
2059 |
"bias": false,
|
2060 |
"enable_norm": true,
|
2061 |
-
"enable_perm":
|
2062 |
"group_num": 1,
|
2063 |
"group_size": 14336,
|
2064 |
"in_features": 14336,
|
@@ -2082,7 +2082,7 @@
|
|
2082 |
"model.layers.2.mlp.gate_proj": {
|
2083 |
"bias": false,
|
2084 |
"enable_norm": true,
|
2085 |
-
"enable_perm":
|
2086 |
"group_num": 1,
|
2087 |
"group_size": 4096,
|
2088 |
"in_features": 4096,
|
@@ -2106,7 +2106,7 @@
|
|
2106 |
"model.layers.2.mlp.up_proj": {
|
2107 |
"bias": false,
|
2108 |
"enable_norm": true,
|
2109 |
-
"enable_perm":
|
2110 |
"group_num": 1,
|
2111 |
"group_size": 4096,
|
2112 |
"in_features": 4096,
|
@@ -2130,7 +2130,7 @@
|
|
2130 |
"model.layers.2.self_attn.k_proj": {
|
2131 |
"bias": false,
|
2132 |
"enable_norm": true,
|
2133 |
-
"enable_perm":
|
2134 |
"group_num": 1,
|
2135 |
"group_size": 4096,
|
2136 |
"in_features": 4096,
|
@@ -2154,7 +2154,7 @@
|
|
2154 |
"model.layers.2.self_attn.o_proj": {
|
2155 |
"bias": false,
|
2156 |
"enable_norm": true,
|
2157 |
-
"enable_perm":
|
2158 |
"group_num": 1,
|
2159 |
"group_size": 4096,
|
2160 |
"in_features": 4096,
|
@@ -2178,7 +2178,7 @@
|
|
2178 |
"model.layers.2.self_attn.q_proj": {
|
2179 |
"bias": false,
|
2180 |
"enable_norm": true,
|
2181 |
-
"enable_perm":
|
2182 |
"group_num": 1,
|
2183 |
"group_size": 4096,
|
2184 |
"in_features": 4096,
|
@@ -2202,7 +2202,7 @@
|
|
2202 |
"model.layers.2.self_attn.v_proj": {
|
2203 |
"bias": false,
|
2204 |
"enable_norm": true,
|
2205 |
-
"enable_perm":
|
2206 |
"group_num": 1,
|
2207 |
"group_size": 4096,
|
2208 |
"in_features": 4096,
|
@@ -2226,7 +2226,7 @@
|
|
2226 |
"model.layers.20.mlp.down_proj": {
|
2227 |
"bias": false,
|
2228 |
"enable_norm": true,
|
2229 |
-
"enable_perm":
|
2230 |
"group_num": 1,
|
2231 |
"group_size": 14336,
|
2232 |
"in_features": 14336,
|
@@ -2250,7 +2250,7 @@
|
|
2250 |
"model.layers.20.mlp.gate_proj": {
|
2251 |
"bias": false,
|
2252 |
"enable_norm": true,
|
2253 |
-
"enable_perm":
|
2254 |
"group_num": 1,
|
2255 |
"group_size": 4096,
|
2256 |
"in_features": 4096,
|
@@ -2274,7 +2274,7 @@
|
|
2274 |
"model.layers.20.mlp.up_proj": {
|
2275 |
"bias": false,
|
2276 |
"enable_norm": true,
|
2277 |
-
"enable_perm":
|
2278 |
"group_num": 1,
|
2279 |
"group_size": 4096,
|
2280 |
"in_features": 4096,
|
@@ -2298,7 +2298,7 @@
|
|
2298 |
"model.layers.20.self_attn.k_proj": {
|
2299 |
"bias": false,
|
2300 |
"enable_norm": true,
|
2301 |
-
"enable_perm":
|
2302 |
"group_num": 1,
|
2303 |
"group_size": 4096,
|
2304 |
"in_features": 4096,
|
@@ -2322,7 +2322,7 @@
|
|
2322 |
"model.layers.20.self_attn.o_proj": {
|
2323 |
"bias": false,
|
2324 |
"enable_norm": true,
|
2325 |
-
"enable_perm":
|
2326 |
"group_num": 1,
|
2327 |
"group_size": 4096,
|
2328 |
"in_features": 4096,
|
@@ -2346,7 +2346,7 @@
|
|
2346 |
"model.layers.20.self_attn.q_proj": {
|
2347 |
"bias": false,
|
2348 |
"enable_norm": true,
|
2349 |
-
"enable_perm":
|
2350 |
"group_num": 1,
|
2351 |
"group_size": 4096,
|
2352 |
"in_features": 4096,
|
@@ -2370,7 +2370,7 @@
|
|
2370 |
"model.layers.20.self_attn.v_proj": {
|
2371 |
"bias": false,
|
2372 |
"enable_norm": true,
|
2373 |
-
"enable_perm":
|
2374 |
"group_num": 1,
|
2375 |
"group_size": 4096,
|
2376 |
"in_features": 4096,
|
@@ -2394,7 +2394,7 @@
|
|
2394 |
"model.layers.21.mlp.down_proj": {
|
2395 |
"bias": false,
|
2396 |
"enable_norm": true,
|
2397 |
-
"enable_perm":
|
2398 |
"group_num": 1,
|
2399 |
"group_size": 14336,
|
2400 |
"in_features": 14336,
|
@@ -2418,7 +2418,7 @@
|
|
2418 |
"model.layers.21.mlp.gate_proj": {
|
2419 |
"bias": false,
|
2420 |
"enable_norm": true,
|
2421 |
-
"enable_perm":
|
2422 |
"group_num": 1,
|
2423 |
"group_size": 4096,
|
2424 |
"in_features": 4096,
|
@@ -2442,7 +2442,7 @@
|
|
2442 |
"model.layers.21.mlp.up_proj": {
|
2443 |
"bias": false,
|
2444 |
"enable_norm": true,
|
2445 |
-
"enable_perm":
|
2446 |
"group_num": 1,
|
2447 |
"group_size": 4096,
|
2448 |
"in_features": 4096,
|
@@ -2466,7 +2466,7 @@
|
|
2466 |
"model.layers.21.self_attn.k_proj": {
|
2467 |
"bias": false,
|
2468 |
"enable_norm": true,
|
2469 |
-
"enable_perm":
|
2470 |
"group_num": 1,
|
2471 |
"group_size": 4096,
|
2472 |
"in_features": 4096,
|
@@ -2490,7 +2490,7 @@
|
|
2490 |
"model.layers.21.self_attn.o_proj": {
|
2491 |
"bias": false,
|
2492 |
"enable_norm": true,
|
2493 |
-
"enable_perm":
|
2494 |
"group_num": 1,
|
2495 |
"group_size": 4096,
|
2496 |
"in_features": 4096,
|
@@ -2514,7 +2514,7 @@
|
|
2514 |
"model.layers.21.self_attn.q_proj": {
|
2515 |
"bias": false,
|
2516 |
"enable_norm": true,
|
2517 |
-
"enable_perm":
|
2518 |
"group_num": 1,
|
2519 |
"group_size": 4096,
|
2520 |
"in_features": 4096,
|
@@ -2538,7 +2538,7 @@
|
|
2538 |
"model.layers.21.self_attn.v_proj": {
|
2539 |
"bias": false,
|
2540 |
"enable_norm": true,
|
2541 |
-
"enable_perm":
|
2542 |
"group_num": 1,
|
2543 |
"group_size": 4096,
|
2544 |
"in_features": 4096,
|
@@ -2562,7 +2562,7 @@
|
|
2562 |
"model.layers.22.mlp.down_proj": {
|
2563 |
"bias": false,
|
2564 |
"enable_norm": true,
|
2565 |
-
"enable_perm":
|
2566 |
"group_num": 1,
|
2567 |
"group_size": 14336,
|
2568 |
"in_features": 14336,
|
@@ -2586,7 +2586,7 @@
|
|
2586 |
"model.layers.22.mlp.gate_proj": {
|
2587 |
"bias": false,
|
2588 |
"enable_norm": true,
|
2589 |
-
"enable_perm":
|
2590 |
"group_num": 1,
|
2591 |
"group_size": 4096,
|
2592 |
"in_features": 4096,
|
@@ -2610,7 +2610,7 @@
|
|
2610 |
"model.layers.22.mlp.up_proj": {
|
2611 |
"bias": false,
|
2612 |
"enable_norm": true,
|
2613 |
-
"enable_perm":
|
2614 |
"group_num": 1,
|
2615 |
"group_size": 4096,
|
2616 |
"in_features": 4096,
|
@@ -2634,7 +2634,7 @@
|
|
2634 |
"model.layers.22.self_attn.k_proj": {
|
2635 |
"bias": false,
|
2636 |
"enable_norm": true,
|
2637 |
-
"enable_perm":
|
2638 |
"group_num": 1,
|
2639 |
"group_size": 4096,
|
2640 |
"in_features": 4096,
|
@@ -2658,7 +2658,7 @@
|
|
2658 |
"model.layers.22.self_attn.o_proj": {
|
2659 |
"bias": false,
|
2660 |
"enable_norm": true,
|
2661 |
-
"enable_perm":
|
2662 |
"group_num": 1,
|
2663 |
"group_size": 4096,
|
2664 |
"in_features": 4096,
|
@@ -2682,7 +2682,7 @@
|
|
2682 |
"model.layers.22.self_attn.q_proj": {
|
2683 |
"bias": false,
|
2684 |
"enable_norm": true,
|
2685 |
-
"enable_perm":
|
2686 |
"group_num": 1,
|
2687 |
"group_size": 4096,
|
2688 |
"in_features": 4096,
|
@@ -2706,7 +2706,7 @@
|
|
2706 |
"model.layers.22.self_attn.v_proj": {
|
2707 |
"bias": false,
|
2708 |
"enable_norm": true,
|
2709 |
-
"enable_perm":
|
2710 |
"group_num": 1,
|
2711 |
"group_size": 4096,
|
2712 |
"in_features": 4096,
|
@@ -2730,7 +2730,7 @@
|
|
2730 |
"model.layers.23.mlp.down_proj": {
|
2731 |
"bias": false,
|
2732 |
"enable_norm": true,
|
2733 |
-
"enable_perm":
|
2734 |
"group_num": 1,
|
2735 |
"group_size": 14336,
|
2736 |
"in_features": 14336,
|
@@ -2754,7 +2754,7 @@
|
|
2754 |
"model.layers.23.mlp.gate_proj": {
|
2755 |
"bias": false,
|
2756 |
"enable_norm": true,
|
2757 |
-
"enable_perm":
|
2758 |
"group_num": 1,
|
2759 |
"group_size": 4096,
|
2760 |
"in_features": 4096,
|
@@ -2778,7 +2778,7 @@
|
|
2778 |
"model.layers.23.mlp.up_proj": {
|
2779 |
"bias": false,
|
2780 |
"enable_norm": true,
|
2781 |
-
"enable_perm":
|
2782 |
"group_num": 1,
|
2783 |
"group_size": 4096,
|
2784 |
"in_features": 4096,
|
@@ -2802,7 +2802,7 @@
|
|
2802 |
"model.layers.23.self_attn.k_proj": {
|
2803 |
"bias": false,
|
2804 |
"enable_norm": true,
|
2805 |
-
"enable_perm":
|
2806 |
"group_num": 1,
|
2807 |
"group_size": 4096,
|
2808 |
"in_features": 4096,
|
@@ -2826,7 +2826,7 @@
|
|
2826 |
"model.layers.23.self_attn.o_proj": {
|
2827 |
"bias": false,
|
2828 |
"enable_norm": true,
|
2829 |
-
"enable_perm":
|
2830 |
"group_num": 1,
|
2831 |
"group_size": 4096,
|
2832 |
"in_features": 4096,
|
@@ -2850,7 +2850,7 @@
|
|
2850 |
"model.layers.23.self_attn.q_proj": {
|
2851 |
"bias": false,
|
2852 |
"enable_norm": true,
|
2853 |
-
"enable_perm":
|
2854 |
"group_num": 1,
|
2855 |
"group_size": 4096,
|
2856 |
"in_features": 4096,
|
@@ -2874,7 +2874,7 @@
|
|
2874 |
"model.layers.23.self_attn.v_proj": {
|
2875 |
"bias": false,
|
2876 |
"enable_norm": true,
|
2877 |
-
"enable_perm":
|
2878 |
"group_num": 1,
|
2879 |
"group_size": 4096,
|
2880 |
"in_features": 4096,
|
@@ -2898,7 +2898,7 @@
|
|
2898 |
"model.layers.24.mlp.down_proj": {
|
2899 |
"bias": false,
|
2900 |
"enable_norm": true,
|
2901 |
-
"enable_perm":
|
2902 |
"group_num": 1,
|
2903 |
"group_size": 14336,
|
2904 |
"in_features": 14336,
|
@@ -2922,7 +2922,7 @@
|
|
2922 |
"model.layers.24.mlp.gate_proj": {
|
2923 |
"bias": false,
|
2924 |
"enable_norm": true,
|
2925 |
-
"enable_perm":
|
2926 |
"group_num": 1,
|
2927 |
"group_size": 4096,
|
2928 |
"in_features": 4096,
|
@@ -2946,7 +2946,7 @@
|
|
2946 |
"model.layers.24.mlp.up_proj": {
|
2947 |
"bias": false,
|
2948 |
"enable_norm": true,
|
2949 |
-
"enable_perm":
|
2950 |
"group_num": 1,
|
2951 |
"group_size": 4096,
|
2952 |
"in_features": 4096,
|
@@ -2970,7 +2970,7 @@
|
|
2970 |
"model.layers.24.self_attn.k_proj": {
|
2971 |
"bias": false,
|
2972 |
"enable_norm": true,
|
2973 |
-
"enable_perm":
|
2974 |
"group_num": 1,
|
2975 |
"group_size": 4096,
|
2976 |
"in_features": 4096,
|
@@ -2994,7 +2994,7 @@
|
|
2994 |
"model.layers.24.self_attn.o_proj": {
|
2995 |
"bias": false,
|
2996 |
"enable_norm": true,
|
2997 |
-
"enable_perm":
|
2998 |
"group_num": 1,
|
2999 |
"group_size": 4096,
|
3000 |
"in_features": 4096,
|
@@ -3018,7 +3018,7 @@
|
|
3018 |
"model.layers.24.self_attn.q_proj": {
|
3019 |
"bias": false,
|
3020 |
"enable_norm": true,
|
3021 |
-
"enable_perm":
|
3022 |
"group_num": 1,
|
3023 |
"group_size": 4096,
|
3024 |
"in_features": 4096,
|
@@ -3042,7 +3042,7 @@
|
|
3042 |
"model.layers.24.self_attn.v_proj": {
|
3043 |
"bias": false,
|
3044 |
"enable_norm": true,
|
3045 |
-
"enable_perm":
|
3046 |
"group_num": 1,
|
3047 |
"group_size": 4096,
|
3048 |
"in_features": 4096,
|
@@ -3066,7 +3066,7 @@
|
|
3066 |
"model.layers.25.mlp.down_proj": {
|
3067 |
"bias": false,
|
3068 |
"enable_norm": true,
|
3069 |
-
"enable_perm":
|
3070 |
"group_num": 1,
|
3071 |
"group_size": 14336,
|
3072 |
"in_features": 14336,
|
@@ -3090,7 +3090,7 @@
|
|
3090 |
"model.layers.25.mlp.gate_proj": {
|
3091 |
"bias": false,
|
3092 |
"enable_norm": true,
|
3093 |
-
"enable_perm":
|
3094 |
"group_num": 1,
|
3095 |
"group_size": 4096,
|
3096 |
"in_features": 4096,
|
@@ -3114,7 +3114,7 @@
|
|
3114 |
"model.layers.25.mlp.up_proj": {
|
3115 |
"bias": false,
|
3116 |
"enable_norm": true,
|
3117 |
-
"enable_perm":
|
3118 |
"group_num": 1,
|
3119 |
"group_size": 4096,
|
3120 |
"in_features": 4096,
|
@@ -3138,7 +3138,7 @@
|
|
3138 |
"model.layers.25.self_attn.k_proj": {
|
3139 |
"bias": false,
|
3140 |
"enable_norm": true,
|
3141 |
-
"enable_perm":
|
3142 |
"group_num": 1,
|
3143 |
"group_size": 4096,
|
3144 |
"in_features": 4096,
|
@@ -3162,7 +3162,7 @@
|
|
3162 |
"model.layers.25.self_attn.o_proj": {
|
3163 |
"bias": false,
|
3164 |
"enable_norm": true,
|
3165 |
-
"enable_perm":
|
3166 |
"group_num": 1,
|
3167 |
"group_size": 4096,
|
3168 |
"in_features": 4096,
|
@@ -3186,7 +3186,7 @@
|
|
3186 |
"model.layers.25.self_attn.q_proj": {
|
3187 |
"bias": false,
|
3188 |
"enable_norm": true,
|
3189 |
-
"enable_perm":
|
3190 |
"group_num": 1,
|
3191 |
"group_size": 4096,
|
3192 |
"in_features": 4096,
|
@@ -3210,7 +3210,7 @@
|
|
3210 |
"model.layers.25.self_attn.v_proj": {
|
3211 |
"bias": false,
|
3212 |
"enable_norm": true,
|
3213 |
-
"enable_perm":
|
3214 |
"group_num": 1,
|
3215 |
"group_size": 4096,
|
3216 |
"in_features": 4096,
|
@@ -3234,7 +3234,7 @@
|
|
3234 |
"model.layers.26.mlp.down_proj": {
|
3235 |
"bias": false,
|
3236 |
"enable_norm": true,
|
3237 |
-
"enable_perm":
|
3238 |
"group_num": 1,
|
3239 |
"group_size": 14336,
|
3240 |
"in_features": 14336,
|
@@ -3258,7 +3258,7 @@
|
|
3258 |
"model.layers.26.mlp.gate_proj": {
|
3259 |
"bias": false,
|
3260 |
"enable_norm": true,
|
3261 |
-
"enable_perm":
|
3262 |
"group_num": 1,
|
3263 |
"group_size": 4096,
|
3264 |
"in_features": 4096,
|
@@ -3282,7 +3282,7 @@
|
|
3282 |
"model.layers.26.mlp.up_proj": {
|
3283 |
"bias": false,
|
3284 |
"enable_norm": true,
|
3285 |
-
"enable_perm":
|
3286 |
"group_num": 1,
|
3287 |
"group_size": 4096,
|
3288 |
"in_features": 4096,
|
@@ -3306,7 +3306,7 @@
|
|
3306 |
"model.layers.26.self_attn.k_proj": {
|
3307 |
"bias": false,
|
3308 |
"enable_norm": true,
|
3309 |
-
"enable_perm":
|
3310 |
"group_num": 1,
|
3311 |
"group_size": 4096,
|
3312 |
"in_features": 4096,
|
@@ -3330,7 +3330,7 @@
|
|
3330 |
"model.layers.26.self_attn.o_proj": {
|
3331 |
"bias": false,
|
3332 |
"enable_norm": true,
|
3333 |
-
"enable_perm":
|
3334 |
"group_num": 1,
|
3335 |
"group_size": 4096,
|
3336 |
"in_features": 4096,
|
@@ -3354,7 +3354,7 @@
|
|
3354 |
"model.layers.26.self_attn.q_proj": {
|
3355 |
"bias": false,
|
3356 |
"enable_norm": true,
|
3357 |
-
"enable_perm":
|
3358 |
"group_num": 1,
|
3359 |
"group_size": 4096,
|
3360 |
"in_features": 4096,
|
@@ -3378,7 +3378,7 @@
|
|
3378 |
"model.layers.26.self_attn.v_proj": {
|
3379 |
"bias": false,
|
3380 |
"enable_norm": true,
|
3381 |
-
"enable_perm":
|
3382 |
"group_num": 1,
|
3383 |
"group_size": 4096,
|
3384 |
"in_features": 4096,
|
@@ -3402,7 +3402,7 @@
|
|
3402 |
"model.layers.27.mlp.down_proj": {
|
3403 |
"bias": false,
|
3404 |
"enable_norm": true,
|
3405 |
-
"enable_perm":
|
3406 |
"group_num": 1,
|
3407 |
"group_size": 14336,
|
3408 |
"in_features": 14336,
|
@@ -3426,7 +3426,7 @@
|
|
3426 |
"model.layers.27.mlp.gate_proj": {
|
3427 |
"bias": false,
|
3428 |
"enable_norm": true,
|
3429 |
-
"enable_perm":
|
3430 |
"group_num": 1,
|
3431 |
"group_size": 4096,
|
3432 |
"in_features": 4096,
|
@@ -3450,7 +3450,7 @@
|
|
3450 |
"model.layers.27.mlp.up_proj": {
|
3451 |
"bias": false,
|
3452 |
"enable_norm": true,
|
3453 |
-
"enable_perm":
|
3454 |
"group_num": 1,
|
3455 |
"group_size": 4096,
|
3456 |
"in_features": 4096,
|
@@ -3474,7 +3474,7 @@
|
|
3474 |
"model.layers.27.self_attn.k_proj": {
|
3475 |
"bias": false,
|
3476 |
"enable_norm": true,
|
3477 |
-
"enable_perm":
|
3478 |
"group_num": 1,
|
3479 |
"group_size": 4096,
|
3480 |
"in_features": 4096,
|
@@ -3498,7 +3498,7 @@
|
|
3498 |
"model.layers.27.self_attn.o_proj": {
|
3499 |
"bias": false,
|
3500 |
"enable_norm": true,
|
3501 |
-
"enable_perm":
|
3502 |
"group_num": 1,
|
3503 |
"group_size": 4096,
|
3504 |
"in_features": 4096,
|
@@ -3522,7 +3522,7 @@
|
|
3522 |
"model.layers.27.self_attn.q_proj": {
|
3523 |
"bias": false,
|
3524 |
"enable_norm": true,
|
3525 |
-
"enable_perm":
|
3526 |
"group_num": 1,
|
3527 |
"group_size": 4096,
|
3528 |
"in_features": 4096,
|
@@ -3546,7 +3546,7 @@
|
|
3546 |
"model.layers.27.self_attn.v_proj": {
|
3547 |
"bias": false,
|
3548 |
"enable_norm": true,
|
3549 |
-
"enable_perm":
|
3550 |
"group_num": 1,
|
3551 |
"group_size": 4096,
|
3552 |
"in_features": 4096,
|
@@ -3570,7 +3570,7 @@
|
|
3570 |
"model.layers.28.mlp.down_proj": {
|
3571 |
"bias": false,
|
3572 |
"enable_norm": true,
|
3573 |
-
"enable_perm":
|
3574 |
"group_num": 1,
|
3575 |
"group_size": 14336,
|
3576 |
"in_features": 14336,
|
@@ -3594,7 +3594,7 @@
|
|
3594 |
"model.layers.28.mlp.gate_proj": {
|
3595 |
"bias": false,
|
3596 |
"enable_norm": true,
|
3597 |
-
"enable_perm":
|
3598 |
"group_num": 1,
|
3599 |
"group_size": 4096,
|
3600 |
"in_features": 4096,
|
@@ -3618,7 +3618,7 @@
|
|
3618 |
"model.layers.28.mlp.up_proj": {
|
3619 |
"bias": false,
|
3620 |
"enable_norm": true,
|
3621 |
-
"enable_perm":
|
3622 |
"group_num": 1,
|
3623 |
"group_size": 4096,
|
3624 |
"in_features": 4096,
|
@@ -3642,7 +3642,7 @@
|
|
3642 |
"model.layers.28.self_attn.k_proj": {
|
3643 |
"bias": false,
|
3644 |
"enable_norm": true,
|
3645 |
-
"enable_perm":
|
3646 |
"group_num": 1,
|
3647 |
"group_size": 4096,
|
3648 |
"in_features": 4096,
|
@@ -3666,7 +3666,7 @@
|
|
3666 |
"model.layers.28.self_attn.o_proj": {
|
3667 |
"bias": false,
|
3668 |
"enable_norm": true,
|
3669 |
-
"enable_perm":
|
3670 |
"group_num": 1,
|
3671 |
"group_size": 4096,
|
3672 |
"in_features": 4096,
|
@@ -3690,7 +3690,7 @@
|
|
3690 |
"model.layers.28.self_attn.q_proj": {
|
3691 |
"bias": false,
|
3692 |
"enable_norm": true,
|
3693 |
-
"enable_perm":
|
3694 |
"group_num": 1,
|
3695 |
"group_size": 4096,
|
3696 |
"in_features": 4096,
|
@@ -3714,7 +3714,7 @@
|
|
3714 |
"model.layers.28.self_attn.v_proj": {
|
3715 |
"bias": false,
|
3716 |
"enable_norm": true,
|
3717 |
-
"enable_perm":
|
3718 |
"group_num": 1,
|
3719 |
"group_size": 4096,
|
3720 |
"in_features": 4096,
|
@@ -3738,7 +3738,7 @@
|
|
3738 |
"model.layers.29.mlp.down_proj": {
|
3739 |
"bias": false,
|
3740 |
"enable_norm": true,
|
3741 |
-
"enable_perm":
|
3742 |
"group_num": 1,
|
3743 |
"group_size": 14336,
|
3744 |
"in_features": 14336,
|
@@ -3762,7 +3762,7 @@
|
|
3762 |
"model.layers.29.mlp.gate_proj": {
|
3763 |
"bias": false,
|
3764 |
"enable_norm": true,
|
3765 |
-
"enable_perm":
|
3766 |
"group_num": 1,
|
3767 |
"group_size": 4096,
|
3768 |
"in_features": 4096,
|
@@ -3786,7 +3786,7 @@
|
|
3786 |
"model.layers.29.mlp.up_proj": {
|
3787 |
"bias": false,
|
3788 |
"enable_norm": true,
|
3789 |
-
"enable_perm":
|
3790 |
"group_num": 1,
|
3791 |
"group_size": 4096,
|
3792 |
"in_features": 4096,
|
@@ -3810,7 +3810,7 @@
|
|
3810 |
"model.layers.29.self_attn.k_proj": {
|
3811 |
"bias": false,
|
3812 |
"enable_norm": true,
|
3813 |
-
"enable_perm":
|
3814 |
"group_num": 1,
|
3815 |
"group_size": 4096,
|
3816 |
"in_features": 4096,
|
@@ -3834,7 +3834,7 @@
|
|
3834 |
"model.layers.29.self_attn.o_proj": {
|
3835 |
"bias": false,
|
3836 |
"enable_norm": true,
|
3837 |
-
"enable_perm":
|
3838 |
"group_num": 1,
|
3839 |
"group_size": 4096,
|
3840 |
"in_features": 4096,
|
@@ -3858,7 +3858,7 @@
|
|
3858 |
"model.layers.29.self_attn.q_proj": {
|
3859 |
"bias": false,
|
3860 |
"enable_norm": true,
|
3861 |
-
"enable_perm":
|
3862 |
"group_num": 1,
|
3863 |
"group_size": 4096,
|
3864 |
"in_features": 4096,
|
@@ -3882,7 +3882,7 @@
|
|
3882 |
"model.layers.29.self_attn.v_proj": {
|
3883 |
"bias": false,
|
3884 |
"enable_norm": true,
|
3885 |
-
"enable_perm":
|
3886 |
"group_num": 1,
|
3887 |
"group_size": 4096,
|
3888 |
"in_features": 4096,
|
@@ -3906,7 +3906,7 @@
|
|
3906 |
"model.layers.3.mlp.down_proj": {
|
3907 |
"bias": false,
|
3908 |
"enable_norm": true,
|
3909 |
-
"enable_perm":
|
3910 |
"group_num": 1,
|
3911 |
"group_size": 14336,
|
3912 |
"in_features": 14336,
|
@@ -3930,7 +3930,7 @@
|
|
3930 |
"model.layers.3.mlp.gate_proj": {
|
3931 |
"bias": false,
|
3932 |
"enable_norm": true,
|
3933 |
-
"enable_perm":
|
3934 |
"group_num": 1,
|
3935 |
"group_size": 4096,
|
3936 |
"in_features": 4096,
|
@@ -3954,7 +3954,7 @@
|
|
3954 |
"model.layers.3.mlp.up_proj": {
|
3955 |
"bias": false,
|
3956 |
"enable_norm": true,
|
3957 |
-
"enable_perm":
|
3958 |
"group_num": 1,
|
3959 |
"group_size": 4096,
|
3960 |
"in_features": 4096,
|
@@ -3978,7 +3978,7 @@
|
|
3978 |
"model.layers.3.self_attn.k_proj": {
|
3979 |
"bias": false,
|
3980 |
"enable_norm": true,
|
3981 |
-
"enable_perm":
|
3982 |
"group_num": 1,
|
3983 |
"group_size": 4096,
|
3984 |
"in_features": 4096,
|
@@ -4002,7 +4002,7 @@
|
|
4002 |
"model.layers.3.self_attn.o_proj": {
|
4003 |
"bias": false,
|
4004 |
"enable_norm": true,
|
4005 |
-
"enable_perm":
|
4006 |
"group_num": 1,
|
4007 |
"group_size": 4096,
|
4008 |
"in_features": 4096,
|
@@ -4026,7 +4026,7 @@
|
|
4026 |
"model.layers.3.self_attn.q_proj": {
|
4027 |
"bias": false,
|
4028 |
"enable_norm": true,
|
4029 |
-
"enable_perm":
|
4030 |
"group_num": 1,
|
4031 |
"group_size": 4096,
|
4032 |
"in_features": 4096,
|
@@ -4050,7 +4050,7 @@
|
|
4050 |
"model.layers.3.self_attn.v_proj": {
|
4051 |
"bias": false,
|
4052 |
"enable_norm": true,
|
4053 |
-
"enable_perm":
|
4054 |
"group_num": 1,
|
4055 |
"group_size": 4096,
|
4056 |
"in_features": 4096,
|
@@ -4074,7 +4074,7 @@
|
|
4074 |
"model.layers.30.mlp.down_proj": {
|
4075 |
"bias": false,
|
4076 |
"enable_norm": true,
|
4077 |
-
"enable_perm":
|
4078 |
"group_num": 1,
|
4079 |
"group_size": 14336,
|
4080 |
"in_features": 14336,
|
@@ -4098,7 +4098,7 @@
|
|
4098 |
"model.layers.30.mlp.gate_proj": {
|
4099 |
"bias": false,
|
4100 |
"enable_norm": true,
|
4101 |
-
"enable_perm":
|
4102 |
"group_num": 1,
|
4103 |
"group_size": 4096,
|
4104 |
"in_features": 4096,
|
@@ -4122,7 +4122,7 @@
|
|
4122 |
"model.layers.30.mlp.up_proj": {
|
4123 |
"bias": false,
|
4124 |
"enable_norm": true,
|
4125 |
-
"enable_perm":
|
4126 |
"group_num": 1,
|
4127 |
"group_size": 4096,
|
4128 |
"in_features": 4096,
|
@@ -4146,7 +4146,7 @@
|
|
4146 |
"model.layers.30.self_attn.k_proj": {
|
4147 |
"bias": false,
|
4148 |
"enable_norm": true,
|
4149 |
-
"enable_perm":
|
4150 |
"group_num": 1,
|
4151 |
"group_size": 4096,
|
4152 |
"in_features": 4096,
|
@@ -4170,7 +4170,7 @@
|
|
4170 |
"model.layers.30.self_attn.o_proj": {
|
4171 |
"bias": false,
|
4172 |
"enable_norm": true,
|
4173 |
-
"enable_perm":
|
4174 |
"group_num": 1,
|
4175 |
"group_size": 4096,
|
4176 |
"in_features": 4096,
|
@@ -4194,7 +4194,7 @@
|
|
4194 |
"model.layers.30.self_attn.q_proj": {
|
4195 |
"bias": false,
|
4196 |
"enable_norm": true,
|
4197 |
-
"enable_perm":
|
4198 |
"group_num": 1,
|
4199 |
"group_size": 4096,
|
4200 |
"in_features": 4096,
|
@@ -4218,7 +4218,7 @@
|
|
4218 |
"model.layers.30.self_attn.v_proj": {
|
4219 |
"bias": false,
|
4220 |
"enable_norm": true,
|
4221 |
-
"enable_perm":
|
4222 |
"group_num": 1,
|
4223 |
"group_size": 4096,
|
4224 |
"in_features": 4096,
|
@@ -4242,7 +4242,7 @@
|
|
4242 |
"model.layers.31.mlp.down_proj": {
|
4243 |
"bias": false,
|
4244 |
"enable_norm": true,
|
4245 |
-
"enable_perm":
|
4246 |
"group_num": 1,
|
4247 |
"group_size": 14336,
|
4248 |
"in_features": 14336,
|
@@ -4266,7 +4266,7 @@
|
|
4266 |
"model.layers.31.mlp.gate_proj": {
|
4267 |
"bias": false,
|
4268 |
"enable_norm": true,
|
4269 |
-
"enable_perm":
|
4270 |
"group_num": 1,
|
4271 |
"group_size": 4096,
|
4272 |
"in_features": 4096,
|
@@ -4290,7 +4290,7 @@
|
|
4290 |
"model.layers.31.mlp.up_proj": {
|
4291 |
"bias": false,
|
4292 |
"enable_norm": true,
|
4293 |
-
"enable_perm":
|
4294 |
"group_num": 1,
|
4295 |
"group_size": 4096,
|
4296 |
"in_features": 4096,
|
@@ -4314,7 +4314,7 @@
|
|
4314 |
"model.layers.31.self_attn.k_proj": {
|
4315 |
"bias": false,
|
4316 |
"enable_norm": true,
|
4317 |
-
"enable_perm":
|
4318 |
"group_num": 1,
|
4319 |
"group_size": 4096,
|
4320 |
"in_features": 4096,
|
@@ -4338,7 +4338,7 @@
|
|
4338 |
"model.layers.31.self_attn.o_proj": {
|
4339 |
"bias": false,
|
4340 |
"enable_norm": true,
|
4341 |
-
"enable_perm":
|
4342 |
"group_num": 1,
|
4343 |
"group_size": 4096,
|
4344 |
"in_features": 4096,
|
@@ -4362,7 +4362,7 @@
|
|
4362 |
"model.layers.31.self_attn.q_proj": {
|
4363 |
"bias": false,
|
4364 |
"enable_norm": true,
|
4365 |
-
"enable_perm":
|
4366 |
"group_num": 1,
|
4367 |
"group_size": 4096,
|
4368 |
"in_features": 4096,
|
@@ -4386,7 +4386,7 @@
|
|
4386 |
"model.layers.31.self_attn.v_proj": {
|
4387 |
"bias": false,
|
4388 |
"enable_norm": true,
|
4389 |
-
"enable_perm":
|
4390 |
"group_num": 1,
|
4391 |
"group_size": 4096,
|
4392 |
"in_features": 4096,
|
@@ -4410,7 +4410,7 @@
|
|
4410 |
"model.layers.4.mlp.down_proj": {
|
4411 |
"bias": false,
|
4412 |
"enable_norm": true,
|
4413 |
-
"enable_perm":
|
4414 |
"group_num": 1,
|
4415 |
"group_size": 14336,
|
4416 |
"in_features": 14336,
|
@@ -4434,7 +4434,7 @@
|
|
4434 |
"model.layers.4.mlp.gate_proj": {
|
4435 |
"bias": false,
|
4436 |
"enable_norm": true,
|
4437 |
-
"enable_perm":
|
4438 |
"group_num": 1,
|
4439 |
"group_size": 4096,
|
4440 |
"in_features": 4096,
|
@@ -4458,7 +4458,7 @@
|
|
4458 |
"model.layers.4.mlp.up_proj": {
|
4459 |
"bias": false,
|
4460 |
"enable_norm": true,
|
4461 |
-
"enable_perm":
|
4462 |
"group_num": 1,
|
4463 |
"group_size": 4096,
|
4464 |
"in_features": 4096,
|
@@ -4482,7 +4482,7 @@
|
|
4482 |
"model.layers.4.self_attn.k_proj": {
|
4483 |
"bias": false,
|
4484 |
"enable_norm": true,
|
4485 |
-
"enable_perm":
|
4486 |
"group_num": 1,
|
4487 |
"group_size": 4096,
|
4488 |
"in_features": 4096,
|
@@ -4506,7 +4506,7 @@
|
|
4506 |
"model.layers.4.self_attn.o_proj": {
|
4507 |
"bias": false,
|
4508 |
"enable_norm": true,
|
4509 |
-
"enable_perm":
|
4510 |
"group_num": 1,
|
4511 |
"group_size": 4096,
|
4512 |
"in_features": 4096,
|
@@ -4530,7 +4530,7 @@
|
|
4530 |
"model.layers.4.self_attn.q_proj": {
|
4531 |
"bias": false,
|
4532 |
"enable_norm": true,
|
4533 |
-
"enable_perm":
|
4534 |
"group_num": 1,
|
4535 |
"group_size": 4096,
|
4536 |
"in_features": 4096,
|
@@ -4554,7 +4554,7 @@
|
|
4554 |
"model.layers.4.self_attn.v_proj": {
|
4555 |
"bias": false,
|
4556 |
"enable_norm": true,
|
4557 |
-
"enable_perm":
|
4558 |
"group_num": 1,
|
4559 |
"group_size": 4096,
|
4560 |
"in_features": 4096,
|
@@ -4578,7 +4578,7 @@
|
|
4578 |
"model.layers.5.mlp.down_proj": {
|
4579 |
"bias": false,
|
4580 |
"enable_norm": true,
|
4581 |
-
"enable_perm":
|
4582 |
"group_num": 1,
|
4583 |
"group_size": 14336,
|
4584 |
"in_features": 14336,
|
@@ -4602,7 +4602,7 @@
|
|
4602 |
"model.layers.5.mlp.gate_proj": {
|
4603 |
"bias": false,
|
4604 |
"enable_norm": true,
|
4605 |
-
"enable_perm":
|
4606 |
"group_num": 1,
|
4607 |
"group_size": 4096,
|
4608 |
"in_features": 4096,
|
@@ -4626,7 +4626,7 @@
|
|
4626 |
"model.layers.5.mlp.up_proj": {
|
4627 |
"bias": false,
|
4628 |
"enable_norm": true,
|
4629 |
-
"enable_perm":
|
4630 |
"group_num": 1,
|
4631 |
"group_size": 4096,
|
4632 |
"in_features": 4096,
|
@@ -4650,7 +4650,7 @@
|
|
4650 |
"model.layers.5.self_attn.k_proj": {
|
4651 |
"bias": false,
|
4652 |
"enable_norm": true,
|
4653 |
-
"enable_perm":
|
4654 |
"group_num": 1,
|
4655 |
"group_size": 4096,
|
4656 |
"in_features": 4096,
|
@@ -4674,7 +4674,7 @@
|
|
4674 |
"model.layers.5.self_attn.o_proj": {
|
4675 |
"bias": false,
|
4676 |
"enable_norm": true,
|
4677 |
-
"enable_perm":
|
4678 |
"group_num": 1,
|
4679 |
"group_size": 4096,
|
4680 |
"in_features": 4096,
|
@@ -4698,7 +4698,7 @@
|
|
4698 |
"model.layers.5.self_attn.q_proj": {
|
4699 |
"bias": false,
|
4700 |
"enable_norm": true,
|
4701 |
-
"enable_perm":
|
4702 |
"group_num": 1,
|
4703 |
"group_size": 4096,
|
4704 |
"in_features": 4096,
|
@@ -4722,7 +4722,7 @@
|
|
4722 |
"model.layers.5.self_attn.v_proj": {
|
4723 |
"bias": false,
|
4724 |
"enable_norm": true,
|
4725 |
-
"enable_perm":
|
4726 |
"group_num": 1,
|
4727 |
"group_size": 4096,
|
4728 |
"in_features": 4096,
|
@@ -4746,7 +4746,7 @@
|
|
4746 |
"model.layers.6.mlp.down_proj": {
|
4747 |
"bias": false,
|
4748 |
"enable_norm": true,
|
4749 |
-
"enable_perm":
|
4750 |
"group_num": 1,
|
4751 |
"group_size": 14336,
|
4752 |
"in_features": 14336,
|
@@ -4770,7 +4770,7 @@
|
|
4770 |
"model.layers.6.mlp.gate_proj": {
|
4771 |
"bias": false,
|
4772 |
"enable_norm": true,
|
4773 |
-
"enable_perm":
|
4774 |
"group_num": 1,
|
4775 |
"group_size": 4096,
|
4776 |
"in_features": 4096,
|
@@ -4794,7 +4794,7 @@
|
|
4794 |
"model.layers.6.mlp.up_proj": {
|
4795 |
"bias": false,
|
4796 |
"enable_norm": true,
|
4797 |
-
"enable_perm":
|
4798 |
"group_num": 1,
|
4799 |
"group_size": 4096,
|
4800 |
"in_features": 4096,
|
@@ -4818,7 +4818,7 @@
|
|
4818 |
"model.layers.6.self_attn.k_proj": {
|
4819 |
"bias": false,
|
4820 |
"enable_norm": true,
|
4821 |
-
"enable_perm":
|
4822 |
"group_num": 1,
|
4823 |
"group_size": 4096,
|
4824 |
"in_features": 4096,
|
@@ -4842,7 +4842,7 @@
|
|
4842 |
"model.layers.6.self_attn.o_proj": {
|
4843 |
"bias": false,
|
4844 |
"enable_norm": true,
|
4845 |
-
"enable_perm":
|
4846 |
"group_num": 1,
|
4847 |
"group_size": 4096,
|
4848 |
"in_features": 4096,
|
@@ -4866,7 +4866,7 @@
|
|
4866 |
"model.layers.6.self_attn.q_proj": {
|
4867 |
"bias": false,
|
4868 |
"enable_norm": true,
|
4869 |
-
"enable_perm":
|
4870 |
"group_num": 1,
|
4871 |
"group_size": 4096,
|
4872 |
"in_features": 4096,
|
@@ -4890,7 +4890,7 @@
|
|
4890 |
"model.layers.6.self_attn.v_proj": {
|
4891 |
"bias": false,
|
4892 |
"enable_norm": true,
|
4893 |
-
"enable_perm":
|
4894 |
"group_num": 1,
|
4895 |
"group_size": 4096,
|
4896 |
"in_features": 4096,
|
@@ -4914,7 +4914,7 @@
|
|
4914 |
"model.layers.7.mlp.down_proj": {
|
4915 |
"bias": false,
|
4916 |
"enable_norm": true,
|
4917 |
-
"enable_perm":
|
4918 |
"group_num": 1,
|
4919 |
"group_size": 14336,
|
4920 |
"in_features": 14336,
|
@@ -4938,7 +4938,7 @@
|
|
4938 |
"model.layers.7.mlp.gate_proj": {
|
4939 |
"bias": false,
|
4940 |
"enable_norm": true,
|
4941 |
-
"enable_perm":
|
4942 |
"group_num": 1,
|
4943 |
"group_size": 4096,
|
4944 |
"in_features": 4096,
|
@@ -4962,7 +4962,7 @@
|
|
4962 |
"model.layers.7.mlp.up_proj": {
|
4963 |
"bias": false,
|
4964 |
"enable_norm": true,
|
4965 |
-
"enable_perm":
|
4966 |
"group_num": 1,
|
4967 |
"group_size": 4096,
|
4968 |
"in_features": 4096,
|
@@ -4986,7 +4986,7 @@
|
|
4986 |
"model.layers.7.self_attn.k_proj": {
|
4987 |
"bias": false,
|
4988 |
"enable_norm": true,
|
4989 |
-
"enable_perm":
|
4990 |
"group_num": 1,
|
4991 |
"group_size": 4096,
|
4992 |
"in_features": 4096,
|
@@ -5010,7 +5010,7 @@
|
|
5010 |
"model.layers.7.self_attn.o_proj": {
|
5011 |
"bias": false,
|
5012 |
"enable_norm": true,
|
5013 |
-
"enable_perm":
|
5014 |
"group_num": 1,
|
5015 |
"group_size": 4096,
|
5016 |
"in_features": 4096,
|
@@ -5034,7 +5034,7 @@
|
|
5034 |
"model.layers.7.self_attn.q_proj": {
|
5035 |
"bias": false,
|
5036 |
"enable_norm": true,
|
5037 |
-
"enable_perm":
|
5038 |
"group_num": 1,
|
5039 |
"group_size": 4096,
|
5040 |
"in_features": 4096,
|
@@ -5058,7 +5058,7 @@
|
|
5058 |
"model.layers.7.self_attn.v_proj": {
|
5059 |
"bias": false,
|
5060 |
"enable_norm": true,
|
5061 |
-
"enable_perm":
|
5062 |
"group_num": 1,
|
5063 |
"group_size": 4096,
|
5064 |
"in_features": 4096,
|
@@ -5082,7 +5082,7 @@
|
|
5082 |
"model.layers.8.mlp.down_proj": {
|
5083 |
"bias": false,
|
5084 |
"enable_norm": true,
|
5085 |
-
"enable_perm":
|
5086 |
"group_num": 1,
|
5087 |
"group_size": 14336,
|
5088 |
"in_features": 14336,
|
@@ -5106,7 +5106,7 @@
|
|
5106 |
"model.layers.8.mlp.gate_proj": {
|
5107 |
"bias": false,
|
5108 |
"enable_norm": true,
|
5109 |
-
"enable_perm":
|
5110 |
"group_num": 1,
|
5111 |
"group_size": 4096,
|
5112 |
"in_features": 4096,
|
@@ -5130,7 +5130,7 @@
|
|
5130 |
"model.layers.8.mlp.up_proj": {
|
5131 |
"bias": false,
|
5132 |
"enable_norm": true,
|
5133 |
-
"enable_perm":
|
5134 |
"group_num": 1,
|
5135 |
"group_size": 4096,
|
5136 |
"in_features": 4096,
|
@@ -5154,7 +5154,7 @@
|
|
5154 |
"model.layers.8.self_attn.k_proj": {
|
5155 |
"bias": false,
|
5156 |
"enable_norm": true,
|
5157 |
-
"enable_perm":
|
5158 |
"group_num": 1,
|
5159 |
"group_size": 4096,
|
5160 |
"in_features": 4096,
|
@@ -5178,7 +5178,7 @@
|
|
5178 |
"model.layers.8.self_attn.o_proj": {
|
5179 |
"bias": false,
|
5180 |
"enable_norm": true,
|
5181 |
-
"enable_perm":
|
5182 |
"group_num": 1,
|
5183 |
"group_size": 4096,
|
5184 |
"in_features": 4096,
|
@@ -5202,7 +5202,7 @@
|
|
5202 |
"model.layers.8.self_attn.q_proj": {
|
5203 |
"bias": false,
|
5204 |
"enable_norm": true,
|
5205 |
-
"enable_perm":
|
5206 |
"group_num": 1,
|
5207 |
"group_size": 4096,
|
5208 |
"in_features": 4096,
|
@@ -5226,7 +5226,7 @@
|
|
5226 |
"model.layers.8.self_attn.v_proj": {
|
5227 |
"bias": false,
|
5228 |
"enable_norm": true,
|
5229 |
-
"enable_perm":
|
5230 |
"group_num": 1,
|
5231 |
"group_size": 4096,
|
5232 |
"in_features": 4096,
|
@@ -5250,7 +5250,7 @@
|
|
5250 |
"model.layers.9.mlp.down_proj": {
|
5251 |
"bias": false,
|
5252 |
"enable_norm": true,
|
5253 |
-
"enable_perm":
|
5254 |
"group_num": 1,
|
5255 |
"group_size": 14336,
|
5256 |
"in_features": 14336,
|
@@ -5274,7 +5274,7 @@
|
|
5274 |
"model.layers.9.mlp.gate_proj": {
|
5275 |
"bias": false,
|
5276 |
"enable_norm": true,
|
5277 |
-
"enable_perm":
|
5278 |
"group_num": 1,
|
5279 |
"group_size": 4096,
|
5280 |
"in_features": 4096,
|
@@ -5298,7 +5298,7 @@
|
|
5298 |
"model.layers.9.mlp.up_proj": {
|
5299 |
"bias": false,
|
5300 |
"enable_norm": true,
|
5301 |
-
"enable_perm":
|
5302 |
"group_num": 1,
|
5303 |
"group_size": 4096,
|
5304 |
"in_features": 4096,
|
@@ -5322,7 +5322,7 @@
|
|
5322 |
"model.layers.9.self_attn.k_proj": {
|
5323 |
"bias": false,
|
5324 |
"enable_norm": true,
|
5325 |
-
"enable_perm":
|
5326 |
"group_num": 1,
|
5327 |
"group_size": 4096,
|
5328 |
"in_features": 4096,
|
@@ -5346,7 +5346,7 @@
|
|
5346 |
"model.layers.9.self_attn.o_proj": {
|
5347 |
"bias": false,
|
5348 |
"enable_norm": true,
|
5349 |
-
"enable_perm":
|
5350 |
"group_num": 1,
|
5351 |
"group_size": 4096,
|
5352 |
"in_features": 4096,
|
@@ -5370,7 +5370,7 @@
|
|
5370 |
"model.layers.9.self_attn.q_proj": {
|
5371 |
"bias": false,
|
5372 |
"enable_norm": true,
|
5373 |
-
"enable_perm":
|
5374 |
"group_num": 1,
|
5375 |
"group_size": 4096,
|
5376 |
"in_features": 4096,
|
@@ -5394,7 +5394,7 @@
|
|
5394 |
"model.layers.9.self_attn.v_proj": {
|
5395 |
"bias": false,
|
5396 |
"enable_norm": true,
|
5397 |
-
"enable_perm":
|
5398 |
"group_num": 1,
|
5399 |
"group_size": 4096,
|
5400 |
"in_features": 4096,
|
|
|
42 |
"model.layers.0.mlp.down_proj": {
|
43 |
"bias": false,
|
44 |
"enable_norm": true,
|
45 |
+
"enable_perm": false,
|
46 |
"group_num": 1,
|
47 |
"group_size": 14336,
|
48 |
"in_features": 14336,
|
|
|
66 |
"model.layers.0.mlp.gate_proj": {
|
67 |
"bias": false,
|
68 |
"enable_norm": true,
|
69 |
+
"enable_perm": false,
|
70 |
"group_num": 1,
|
71 |
"group_size": 4096,
|
72 |
"in_features": 4096,
|
|
|
90 |
"model.layers.0.mlp.up_proj": {
|
91 |
"bias": false,
|
92 |
"enable_norm": true,
|
93 |
+
"enable_perm": false,
|
94 |
"group_num": 1,
|
95 |
"group_size": 4096,
|
96 |
"in_features": 4096,
|
|
|
114 |
"model.layers.0.self_attn.k_proj": {
|
115 |
"bias": false,
|
116 |
"enable_norm": true,
|
117 |
+
"enable_perm": false,
|
118 |
"group_num": 1,
|
119 |
"group_size": 4096,
|
120 |
"in_features": 4096,
|
|
|
138 |
"model.layers.0.self_attn.o_proj": {
|
139 |
"bias": false,
|
140 |
"enable_norm": true,
|
141 |
+
"enable_perm": false,
|
142 |
"group_num": 1,
|
143 |
"group_size": 4096,
|
144 |
"in_features": 4096,
|
|
|
162 |
"model.layers.0.self_attn.q_proj": {
|
163 |
"bias": false,
|
164 |
"enable_norm": true,
|
165 |
+
"enable_perm": false,
|
166 |
"group_num": 1,
|
167 |
"group_size": 4096,
|
168 |
"in_features": 4096,
|
|
|
186 |
"model.layers.0.self_attn.v_proj": {
|
187 |
"bias": false,
|
188 |
"enable_norm": true,
|
189 |
+
"enable_perm": false,
|
190 |
"group_num": 1,
|
191 |
"group_size": 4096,
|
192 |
"in_features": 4096,
|
|
|
210 |
"model.layers.1.mlp.down_proj": {
|
211 |
"bias": false,
|
212 |
"enable_norm": true,
|
213 |
+
"enable_perm": false,
|
214 |
"group_num": 1,
|
215 |
"group_size": 14336,
|
216 |
"in_features": 14336,
|
|
|
234 |
"model.layers.1.mlp.gate_proj": {
|
235 |
"bias": false,
|
236 |
"enable_norm": true,
|
237 |
+
"enable_perm": false,
|
238 |
"group_num": 1,
|
239 |
"group_size": 4096,
|
240 |
"in_features": 4096,
|
|
|
258 |
"model.layers.1.mlp.up_proj": {
|
259 |
"bias": false,
|
260 |
"enable_norm": true,
|
261 |
+
"enable_perm": false,
|
262 |
"group_num": 1,
|
263 |
"group_size": 4096,
|
264 |
"in_features": 4096,
|
|
|
282 |
"model.layers.1.self_attn.k_proj": {
|
283 |
"bias": false,
|
284 |
"enable_norm": true,
|
285 |
+
"enable_perm": false,
|
286 |
"group_num": 1,
|
287 |
"group_size": 4096,
|
288 |
"in_features": 4096,
|
|
|
306 |
"model.layers.1.self_attn.o_proj": {
|
307 |
"bias": false,
|
308 |
"enable_norm": true,
|
309 |
+
"enable_perm": false,
|
310 |
"group_num": 1,
|
311 |
"group_size": 4096,
|
312 |
"in_features": 4096,
|
|
|
330 |
"model.layers.1.self_attn.q_proj": {
|
331 |
"bias": false,
|
332 |
"enable_norm": true,
|
333 |
+
"enable_perm": false,
|
334 |
"group_num": 1,
|
335 |
"group_size": 4096,
|
336 |
"in_features": 4096,
|
|
|
354 |
"model.layers.1.self_attn.v_proj": {
|
355 |
"bias": false,
|
356 |
"enable_norm": true,
|
357 |
+
"enable_perm": false,
|
358 |
"group_num": 1,
|
359 |
"group_size": 4096,
|
360 |
"in_features": 4096,
|
|
|
378 |
"model.layers.10.mlp.down_proj": {
|
379 |
"bias": false,
|
380 |
"enable_norm": true,
|
381 |
+
"enable_perm": false,
|
382 |
"group_num": 1,
|
383 |
"group_size": 14336,
|
384 |
"in_features": 14336,
|
|
|
402 |
"model.layers.10.mlp.gate_proj": {
|
403 |
"bias": false,
|
404 |
"enable_norm": true,
|
405 |
+
"enable_perm": false,
|
406 |
"group_num": 1,
|
407 |
"group_size": 4096,
|
408 |
"in_features": 4096,
|
|
|
426 |
"model.layers.10.mlp.up_proj": {
|
427 |
"bias": false,
|
428 |
"enable_norm": true,
|
429 |
+
"enable_perm": false,
|
430 |
"group_num": 1,
|
431 |
"group_size": 4096,
|
432 |
"in_features": 4096,
|
|
|
450 |
"model.layers.10.self_attn.k_proj": {
|
451 |
"bias": false,
|
452 |
"enable_norm": true,
|
453 |
+
"enable_perm": false,
|
454 |
"group_num": 1,
|
455 |
"group_size": 4096,
|
456 |
"in_features": 4096,
|
|
|
474 |
"model.layers.10.self_attn.o_proj": {
|
475 |
"bias": false,
|
476 |
"enable_norm": true,
|
477 |
+
"enable_perm": false,
|
478 |
"group_num": 1,
|
479 |
"group_size": 4096,
|
480 |
"in_features": 4096,
|
|
|
498 |
"model.layers.10.self_attn.q_proj": {
|
499 |
"bias": false,
|
500 |
"enable_norm": true,
|
501 |
+
"enable_perm": false,
|
502 |
"group_num": 1,
|
503 |
"group_size": 4096,
|
504 |
"in_features": 4096,
|
|
|
522 |
"model.layers.10.self_attn.v_proj": {
|
523 |
"bias": false,
|
524 |
"enable_norm": true,
|
525 |
+
"enable_perm": false,
|
526 |
"group_num": 1,
|
527 |
"group_size": 4096,
|
528 |
"in_features": 4096,
|
|
|
546 |
"model.layers.11.mlp.down_proj": {
|
547 |
"bias": false,
|
548 |
"enable_norm": true,
|
549 |
+
"enable_perm": false,
|
550 |
"group_num": 1,
|
551 |
"group_size": 14336,
|
552 |
"in_features": 14336,
|
|
|
570 |
"model.layers.11.mlp.gate_proj": {
|
571 |
"bias": false,
|
572 |
"enable_norm": true,
|
573 |
+
"enable_perm": false,
|
574 |
"group_num": 1,
|
575 |
"group_size": 4096,
|
576 |
"in_features": 4096,
|
|
|
594 |
"model.layers.11.mlp.up_proj": {
|
595 |
"bias": false,
|
596 |
"enable_norm": true,
|
597 |
+
"enable_perm": false,
|
598 |
"group_num": 1,
|
599 |
"group_size": 4096,
|
600 |
"in_features": 4096,
|
|
|
618 |
"model.layers.11.self_attn.k_proj": {
|
619 |
"bias": false,
|
620 |
"enable_norm": true,
|
621 |
+
"enable_perm": false,
|
622 |
"group_num": 1,
|
623 |
"group_size": 4096,
|
624 |
"in_features": 4096,
|
|
|
642 |
"model.layers.11.self_attn.o_proj": {
|
643 |
"bias": false,
|
644 |
"enable_norm": true,
|
645 |
+
"enable_perm": false,
|
646 |
"group_num": 1,
|
647 |
"group_size": 4096,
|
648 |
"in_features": 4096,
|
|
|
666 |
"model.layers.11.self_attn.q_proj": {
|
667 |
"bias": false,
|
668 |
"enable_norm": true,
|
669 |
+
"enable_perm": false,
|
670 |
"group_num": 1,
|
671 |
"group_size": 4096,
|
672 |
"in_features": 4096,
|
|
|
690 |
"model.layers.11.self_attn.v_proj": {
|
691 |
"bias": false,
|
692 |
"enable_norm": true,
|
693 |
+
"enable_perm": false,
|
694 |
"group_num": 1,
|
695 |
"group_size": 4096,
|
696 |
"in_features": 4096,
|
|
|
714 |
"model.layers.12.mlp.down_proj": {
|
715 |
"bias": false,
|
716 |
"enable_norm": true,
|
717 |
+
"enable_perm": false,
|
718 |
"group_num": 1,
|
719 |
"group_size": 14336,
|
720 |
"in_features": 14336,
|
|
|
738 |
"model.layers.12.mlp.gate_proj": {
|
739 |
"bias": false,
|
740 |
"enable_norm": true,
|
741 |
+
"enable_perm": false,
|
742 |
"group_num": 1,
|
743 |
"group_size": 4096,
|
744 |
"in_features": 4096,
|
|
|
762 |
"model.layers.12.mlp.up_proj": {
|
763 |
"bias": false,
|
764 |
"enable_norm": true,
|
765 |
+
"enable_perm": false,
|
766 |
"group_num": 1,
|
767 |
"group_size": 4096,
|
768 |
"in_features": 4096,
|
|
|
786 |
"model.layers.12.self_attn.k_proj": {
|
787 |
"bias": false,
|
788 |
"enable_norm": true,
|
789 |
+
"enable_perm": false,
|
790 |
"group_num": 1,
|
791 |
"group_size": 4096,
|
792 |
"in_features": 4096,
|
|
|
810 |
"model.layers.12.self_attn.o_proj": {
|
811 |
"bias": false,
|
812 |
"enable_norm": true,
|
813 |
+
"enable_perm": false,
|
814 |
"group_num": 1,
|
815 |
"group_size": 4096,
|
816 |
"in_features": 4096,
|
|
|
834 |
"model.layers.12.self_attn.q_proj": {
|
835 |
"bias": false,
|
836 |
"enable_norm": true,
|
837 |
+
"enable_perm": false,
|
838 |
"group_num": 1,
|
839 |
"group_size": 4096,
|
840 |
"in_features": 4096,
|
|
|
858 |
"model.layers.12.self_attn.v_proj": {
|
859 |
"bias": false,
|
860 |
"enable_norm": true,
|
861 |
+
"enable_perm": false,
|
862 |
"group_num": 1,
|
863 |
"group_size": 4096,
|
864 |
"in_features": 4096,
|
|
|
882 |
"model.layers.13.mlp.down_proj": {
|
883 |
"bias": false,
|
884 |
"enable_norm": true,
|
885 |
+
"enable_perm": false,
|
886 |
"group_num": 1,
|
887 |
"group_size": 14336,
|
888 |
"in_features": 14336,
|
|
|
906 |
"model.layers.13.mlp.gate_proj": {
|
907 |
"bias": false,
|
908 |
"enable_norm": true,
|
909 |
+
"enable_perm": false,
|
910 |
"group_num": 1,
|
911 |
"group_size": 4096,
|
912 |
"in_features": 4096,
|
|
|
930 |
"model.layers.13.mlp.up_proj": {
|
931 |
"bias": false,
|
932 |
"enable_norm": true,
|
933 |
+
"enable_perm": false,
|
934 |
"group_num": 1,
|
935 |
"group_size": 4096,
|
936 |
"in_features": 4096,
|
|
|
954 |
"model.layers.13.self_attn.k_proj": {
|
955 |
"bias": false,
|
956 |
"enable_norm": true,
|
957 |
+
"enable_perm": false,
|
958 |
"group_num": 1,
|
959 |
"group_size": 4096,
|
960 |
"in_features": 4096,
|
|
|
978 |
"model.layers.13.self_attn.o_proj": {
|
979 |
"bias": false,
|
980 |
"enable_norm": true,
|
981 |
+
"enable_perm": false,
|
982 |
"group_num": 1,
|
983 |
"group_size": 4096,
|
984 |
"in_features": 4096,
|
|
|
1002 |
"model.layers.13.self_attn.q_proj": {
|
1003 |
"bias": false,
|
1004 |
"enable_norm": true,
|
1005 |
+
"enable_perm": false,
|
1006 |
"group_num": 1,
|
1007 |
"group_size": 4096,
|
1008 |
"in_features": 4096,
|
|
|
1026 |
"model.layers.13.self_attn.v_proj": {
|
1027 |
"bias": false,
|
1028 |
"enable_norm": true,
|
1029 |
+
"enable_perm": false,
|
1030 |
"group_num": 1,
|
1031 |
"group_size": 4096,
|
1032 |
"in_features": 4096,
|
|
|
1050 |
"model.layers.14.mlp.down_proj": {
|
1051 |
"bias": false,
|
1052 |
"enable_norm": true,
|
1053 |
+
"enable_perm": false,
|
1054 |
"group_num": 1,
|
1055 |
"group_size": 14336,
|
1056 |
"in_features": 14336,
|
|
|
1074 |
"model.layers.14.mlp.gate_proj": {
|
1075 |
"bias": false,
|
1076 |
"enable_norm": true,
|
1077 |
+
"enable_perm": false,
|
1078 |
"group_num": 1,
|
1079 |
"group_size": 4096,
|
1080 |
"in_features": 4096,
|
|
|
1098 |
"model.layers.14.mlp.up_proj": {
|
1099 |
"bias": false,
|
1100 |
"enable_norm": true,
|
1101 |
+
"enable_perm": false,
|
1102 |
"group_num": 1,
|
1103 |
"group_size": 4096,
|
1104 |
"in_features": 4096,
|
|
|
1122 |
"model.layers.14.self_attn.k_proj": {
|
1123 |
"bias": false,
|
1124 |
"enable_norm": true,
|
1125 |
+
"enable_perm": false,
|
1126 |
"group_num": 1,
|
1127 |
"group_size": 4096,
|
1128 |
"in_features": 4096,
|
|
|
1146 |
"model.layers.14.self_attn.o_proj": {
|
1147 |
"bias": false,
|
1148 |
"enable_norm": true,
|
1149 |
+
"enable_perm": false,
|
1150 |
"group_num": 1,
|
1151 |
"group_size": 4096,
|
1152 |
"in_features": 4096,
|
|
|
1170 |
"model.layers.14.self_attn.q_proj": {
|
1171 |
"bias": false,
|
1172 |
"enable_norm": true,
|
1173 |
+
"enable_perm": false,
|
1174 |
"group_num": 1,
|
1175 |
"group_size": 4096,
|
1176 |
"in_features": 4096,
|
|
|
1194 |
"model.layers.14.self_attn.v_proj": {
|
1195 |
"bias": false,
|
1196 |
"enable_norm": true,
|
1197 |
+
"enable_perm": false,
|
1198 |
"group_num": 1,
|
1199 |
"group_size": 4096,
|
1200 |
"in_features": 4096,
|
|
|
1218 |
"model.layers.15.mlp.down_proj": {
|
1219 |
"bias": false,
|
1220 |
"enable_norm": true,
|
1221 |
+
"enable_perm": false,
|
1222 |
"group_num": 1,
|
1223 |
"group_size": 14336,
|
1224 |
"in_features": 14336,
|
|
|
1242 |
"model.layers.15.mlp.gate_proj": {
|
1243 |
"bias": false,
|
1244 |
"enable_norm": true,
|
1245 |
+
"enable_perm": false,
|
1246 |
"group_num": 1,
|
1247 |
"group_size": 4096,
|
1248 |
"in_features": 4096,
|
|
|
1266 |
"model.layers.15.mlp.up_proj": {
|
1267 |
"bias": false,
|
1268 |
"enable_norm": true,
|
1269 |
+
"enable_perm": false,
|
1270 |
"group_num": 1,
|
1271 |
"group_size": 4096,
|
1272 |
"in_features": 4096,
|
|
|
1290 |
"model.layers.15.self_attn.k_proj": {
|
1291 |
"bias": false,
|
1292 |
"enable_norm": true,
|
1293 |
+
"enable_perm": false,
|
1294 |
"group_num": 1,
|
1295 |
"group_size": 4096,
|
1296 |
"in_features": 4096,
|
|
|
1314 |
"model.layers.15.self_attn.o_proj": {
|
1315 |
"bias": false,
|
1316 |
"enable_norm": true,
|
1317 |
+
"enable_perm": false,
|
1318 |
"group_num": 1,
|
1319 |
"group_size": 4096,
|
1320 |
"in_features": 4096,
|
|
|
1338 |
"model.layers.15.self_attn.q_proj": {
|
1339 |
"bias": false,
|
1340 |
"enable_norm": true,
|
1341 |
+
"enable_perm": false,
|
1342 |
"group_num": 1,
|
1343 |
"group_size": 4096,
|
1344 |
"in_features": 4096,
|
|
|
1362 |
"model.layers.15.self_attn.v_proj": {
|
1363 |
"bias": false,
|
1364 |
"enable_norm": true,
|
1365 |
+
"enable_perm": false,
|
1366 |
"group_num": 1,
|
1367 |
"group_size": 4096,
|
1368 |
"in_features": 4096,
|
|
|
1386 |
"model.layers.16.mlp.down_proj": {
|
1387 |
"bias": false,
|
1388 |
"enable_norm": true,
|
1389 |
+
"enable_perm": false,
|
1390 |
"group_num": 1,
|
1391 |
"group_size": 14336,
|
1392 |
"in_features": 14336,
|
|
|
1410 |
"model.layers.16.mlp.gate_proj": {
|
1411 |
"bias": false,
|
1412 |
"enable_norm": true,
|
1413 |
+
"enable_perm": false,
|
1414 |
"group_num": 1,
|
1415 |
"group_size": 4096,
|
1416 |
"in_features": 4096,
|
|
|
1434 |
"model.layers.16.mlp.up_proj": {
|
1435 |
"bias": false,
|
1436 |
"enable_norm": true,
|
1437 |
+
"enable_perm": false,
|
1438 |
"group_num": 1,
|
1439 |
"group_size": 4096,
|
1440 |
"in_features": 4096,
|
|
|
1458 |
"model.layers.16.self_attn.k_proj": {
|
1459 |
"bias": false,
|
1460 |
"enable_norm": true,
|
1461 |
+
"enable_perm": false,
|
1462 |
"group_num": 1,
|
1463 |
"group_size": 4096,
|
1464 |
"in_features": 4096,
|
|
|
1482 |
"model.layers.16.self_attn.o_proj": {
|
1483 |
"bias": false,
|
1484 |
"enable_norm": true,
|
1485 |
+
"enable_perm": false,
|
1486 |
"group_num": 1,
|
1487 |
"group_size": 4096,
|
1488 |
"in_features": 4096,
|
|
|
1506 |
"model.layers.16.self_attn.q_proj": {
|
1507 |
"bias": false,
|
1508 |
"enable_norm": true,
|
1509 |
+
"enable_perm": false,
|
1510 |
"group_num": 1,
|
1511 |
"group_size": 4096,
|
1512 |
"in_features": 4096,
|
|
|
1530 |
"model.layers.16.self_attn.v_proj": {
|
1531 |
"bias": false,
|
1532 |
"enable_norm": true,
|
1533 |
+
"enable_perm": false,
|
1534 |
"group_num": 1,
|
1535 |
"group_size": 4096,
|
1536 |
"in_features": 4096,
|
|
|
1554 |
"model.layers.17.mlp.down_proj": {
|
1555 |
"bias": false,
|
1556 |
"enable_norm": true,
|
1557 |
+
"enable_perm": false,
|
1558 |
"group_num": 1,
|
1559 |
"group_size": 14336,
|
1560 |
"in_features": 14336,
|
|
|
1578 |
"model.layers.17.mlp.gate_proj": {
|
1579 |
"bias": false,
|
1580 |
"enable_norm": true,
|
1581 |
+
"enable_perm": false,
|
1582 |
"group_num": 1,
|
1583 |
"group_size": 4096,
|
1584 |
"in_features": 4096,
|
|
|
1602 |
"model.layers.17.mlp.up_proj": {
|
1603 |
"bias": false,
|
1604 |
"enable_norm": true,
|
1605 |
+
"enable_perm": false,
|
1606 |
"group_num": 1,
|
1607 |
"group_size": 4096,
|
1608 |
"in_features": 4096,
|
|
|
1626 |
"model.layers.17.self_attn.k_proj": {
|
1627 |
"bias": false,
|
1628 |
"enable_norm": true,
|
1629 |
+
"enable_perm": false,
|
1630 |
"group_num": 1,
|
1631 |
"group_size": 4096,
|
1632 |
"in_features": 4096,
|
|
|
1650 |
"model.layers.17.self_attn.o_proj": {
|
1651 |
"bias": false,
|
1652 |
"enable_norm": true,
|
1653 |
+
"enable_perm": false,
|
1654 |
"group_num": 1,
|
1655 |
"group_size": 4096,
|
1656 |
"in_features": 4096,
|
|
|
1674 |
"model.layers.17.self_attn.q_proj": {
|
1675 |
"bias": false,
|
1676 |
"enable_norm": true,
|
1677 |
+
"enable_perm": false,
|
1678 |
"group_num": 1,
|
1679 |
"group_size": 4096,
|
1680 |
"in_features": 4096,
|
|
|
1698 |
"model.layers.17.self_attn.v_proj": {
|
1699 |
"bias": false,
|
1700 |
"enable_norm": true,
|
1701 |
+
"enable_perm": false,
|
1702 |
"group_num": 1,
|
1703 |
"group_size": 4096,
|
1704 |
"in_features": 4096,
|
|
|
1722 |
"model.layers.18.mlp.down_proj": {
|
1723 |
"bias": false,
|
1724 |
"enable_norm": true,
|
1725 |
+
"enable_perm": false,
|
1726 |
"group_num": 1,
|
1727 |
"group_size": 14336,
|
1728 |
"in_features": 14336,
|
|
|
1746 |
"model.layers.18.mlp.gate_proj": {
|
1747 |
"bias": false,
|
1748 |
"enable_norm": true,
|
1749 |
+
"enable_perm": false,
|
1750 |
"group_num": 1,
|
1751 |
"group_size": 4096,
|
1752 |
"in_features": 4096,
|
|
|
1770 |
"model.layers.18.mlp.up_proj": {
|
1771 |
"bias": false,
|
1772 |
"enable_norm": true,
|
1773 |
+
"enable_perm": false,
|
1774 |
"group_num": 1,
|
1775 |
"group_size": 4096,
|
1776 |
"in_features": 4096,
|
|
|
1794 |
"model.layers.18.self_attn.k_proj": {
|
1795 |
"bias": false,
|
1796 |
"enable_norm": true,
|
1797 |
+
"enable_perm": false,
|
1798 |
"group_num": 1,
|
1799 |
"group_size": 4096,
|
1800 |
"in_features": 4096,
|
|
|
1818 |
"model.layers.18.self_attn.o_proj": {
|
1819 |
"bias": false,
|
1820 |
"enable_norm": true,
|
1821 |
+
"enable_perm": false,
|
1822 |
"group_num": 1,
|
1823 |
"group_size": 4096,
|
1824 |
"in_features": 4096,
|
|
|
1842 |
"model.layers.18.self_attn.q_proj": {
|
1843 |
"bias": false,
|
1844 |
"enable_norm": true,
|
1845 |
+
"enable_perm": false,
|
1846 |
"group_num": 1,
|
1847 |
"group_size": 4096,
|
1848 |
"in_features": 4096,
|
|
|
1866 |
"model.layers.18.self_attn.v_proj": {
|
1867 |
"bias": false,
|
1868 |
"enable_norm": true,
|
1869 |
+
"enable_perm": false,
|
1870 |
"group_num": 1,
|
1871 |
"group_size": 4096,
|
1872 |
"in_features": 4096,
|
|
|
1890 |
"model.layers.19.mlp.down_proj": {
|
1891 |
"bias": false,
|
1892 |
"enable_norm": true,
|
1893 |
+
"enable_perm": false,
|
1894 |
"group_num": 1,
|
1895 |
"group_size": 14336,
|
1896 |
"in_features": 14336,
|
|
|
1914 |
"model.layers.19.mlp.gate_proj": {
|
1915 |
"bias": false,
|
1916 |
"enable_norm": true,
|
1917 |
+
"enable_perm": false,
|
1918 |
"group_num": 1,
|
1919 |
"group_size": 4096,
|
1920 |
"in_features": 4096,
|
|
|
1938 |
"model.layers.19.mlp.up_proj": {
|
1939 |
"bias": false,
|
1940 |
"enable_norm": true,
|
1941 |
+
"enable_perm": false,
|
1942 |
"group_num": 1,
|
1943 |
"group_size": 4096,
|
1944 |
"in_features": 4096,
|
|
|
1962 |
"model.layers.19.self_attn.k_proj": {
|
1963 |
"bias": false,
|
1964 |
"enable_norm": true,
|
1965 |
+
"enable_perm": false,
|
1966 |
"group_num": 1,
|
1967 |
"group_size": 4096,
|
1968 |
"in_features": 4096,
|
|
|
1986 |
"model.layers.19.self_attn.o_proj": {
|
1987 |
"bias": false,
|
1988 |
"enable_norm": true,
|
1989 |
+
"enable_perm": false,
|
1990 |
"group_num": 1,
|
1991 |
"group_size": 4096,
|
1992 |
"in_features": 4096,
|
|
|
2010 |
"model.layers.19.self_attn.q_proj": {
|
2011 |
"bias": false,
|
2012 |
"enable_norm": true,
|
2013 |
+
"enable_perm": false,
|
2014 |
"group_num": 1,
|
2015 |
"group_size": 4096,
|
2016 |
"in_features": 4096,
|
|
|
2034 |
"model.layers.19.self_attn.v_proj": {
|
2035 |
"bias": false,
|
2036 |
"enable_norm": true,
|
2037 |
+
"enable_perm": false,
|
2038 |
"group_num": 1,
|
2039 |
"group_size": 4096,
|
2040 |
"in_features": 4096,
|
|
|
2058 |
"model.layers.2.mlp.down_proj": {
|
2059 |
"bias": false,
|
2060 |
"enable_norm": true,
|
2061 |
+
"enable_perm": false,
|
2062 |
"group_num": 1,
|
2063 |
"group_size": 14336,
|
2064 |
"in_features": 14336,
|
|
|
2082 |
"model.layers.2.mlp.gate_proj": {
|
2083 |
"bias": false,
|
2084 |
"enable_norm": true,
|
2085 |
+
"enable_perm": false,
|
2086 |
"group_num": 1,
|
2087 |
"group_size": 4096,
|
2088 |
"in_features": 4096,
|
|
|
2106 |
"model.layers.2.mlp.up_proj": {
|
2107 |
"bias": false,
|
2108 |
"enable_norm": true,
|
2109 |
+
"enable_perm": false,
|
2110 |
"group_num": 1,
|
2111 |
"group_size": 4096,
|
2112 |
"in_features": 4096,
|
|
|
2130 |
"model.layers.2.self_attn.k_proj": {
|
2131 |
"bias": false,
|
2132 |
"enable_norm": true,
|
2133 |
+
"enable_perm": false,
|
2134 |
"group_num": 1,
|
2135 |
"group_size": 4096,
|
2136 |
"in_features": 4096,
|
|
|
2154 |
"model.layers.2.self_attn.o_proj": {
|
2155 |
"bias": false,
|
2156 |
"enable_norm": true,
|
2157 |
+
"enable_perm": false,
|
2158 |
"group_num": 1,
|
2159 |
"group_size": 4096,
|
2160 |
"in_features": 4096,
|
|
|
2178 |
"model.layers.2.self_attn.q_proj": {
|
2179 |
"bias": false,
|
2180 |
"enable_norm": true,
|
2181 |
+
"enable_perm": false,
|
2182 |
"group_num": 1,
|
2183 |
"group_size": 4096,
|
2184 |
"in_features": 4096,
|
|
|
2202 |
"model.layers.2.self_attn.v_proj": {
|
2203 |
"bias": false,
|
2204 |
"enable_norm": true,
|
2205 |
+
"enable_perm": false,
|
2206 |
"group_num": 1,
|
2207 |
"group_size": 4096,
|
2208 |
"in_features": 4096,
|
|
|
2226 |
"model.layers.20.mlp.down_proj": {
|
2227 |
"bias": false,
|
2228 |
"enable_norm": true,
|
2229 |
+
"enable_perm": false,
|
2230 |
"group_num": 1,
|
2231 |
"group_size": 14336,
|
2232 |
"in_features": 14336,
|
|
|
2250 |
"model.layers.20.mlp.gate_proj": {
|
2251 |
"bias": false,
|
2252 |
"enable_norm": true,
|
2253 |
+
"enable_perm": false,
|
2254 |
"group_num": 1,
|
2255 |
"group_size": 4096,
|
2256 |
"in_features": 4096,
|
|
|
2274 |
"model.layers.20.mlp.up_proj": {
|
2275 |
"bias": false,
|
2276 |
"enable_norm": true,
|
2277 |
+
"enable_perm": false,
|
2278 |
"group_num": 1,
|
2279 |
"group_size": 4096,
|
2280 |
"in_features": 4096,
|
|
|
2298 |
"model.layers.20.self_attn.k_proj": {
|
2299 |
"bias": false,
|
2300 |
"enable_norm": true,
|
2301 |
+
"enable_perm": false,
|
2302 |
"group_num": 1,
|
2303 |
"group_size": 4096,
|
2304 |
"in_features": 4096,
|
|
|
2322 |
"model.layers.20.self_attn.o_proj": {
|
2323 |
"bias": false,
|
2324 |
"enable_norm": true,
|
2325 |
+
"enable_perm": false,
|
2326 |
"group_num": 1,
|
2327 |
"group_size": 4096,
|
2328 |
"in_features": 4096,
|
|
|
2346 |
"model.layers.20.self_attn.q_proj": {
|
2347 |
"bias": false,
|
2348 |
"enable_norm": true,
|
2349 |
+
"enable_perm": false,
|
2350 |
"group_num": 1,
|
2351 |
"group_size": 4096,
|
2352 |
"in_features": 4096,
|
|
|
2370 |
"model.layers.20.self_attn.v_proj": {
|
2371 |
"bias": false,
|
2372 |
"enable_norm": true,
|
2373 |
+
"enable_perm": false,
|
2374 |
"group_num": 1,
|
2375 |
"group_size": 4096,
|
2376 |
"in_features": 4096,
|
|
|
2394 |
"model.layers.21.mlp.down_proj": {
|
2395 |
"bias": false,
|
2396 |
"enable_norm": true,
|
2397 |
+
"enable_perm": false,
|
2398 |
"group_num": 1,
|
2399 |
"group_size": 14336,
|
2400 |
"in_features": 14336,
|
|
|
2418 |
"model.layers.21.mlp.gate_proj": {
|
2419 |
"bias": false,
|
2420 |
"enable_norm": true,
|
2421 |
+
"enable_perm": false,
|
2422 |
"group_num": 1,
|
2423 |
"group_size": 4096,
|
2424 |
"in_features": 4096,
|
|
|
2442 |
"model.layers.21.mlp.up_proj": {
|
2443 |
"bias": false,
|
2444 |
"enable_norm": true,
|
2445 |
+
"enable_perm": false,
|
2446 |
"group_num": 1,
|
2447 |
"group_size": 4096,
|
2448 |
"in_features": 4096,
|
|
|
2466 |
"model.layers.21.self_attn.k_proj": {
|
2467 |
"bias": false,
|
2468 |
"enable_norm": true,
|
2469 |
+
"enable_perm": false,
|
2470 |
"group_num": 1,
|
2471 |
"group_size": 4096,
|
2472 |
"in_features": 4096,
|
|
|
2490 |
"model.layers.21.self_attn.o_proj": {
|
2491 |
"bias": false,
|
2492 |
"enable_norm": true,
|
2493 |
+
"enable_perm": false,
|
2494 |
"group_num": 1,
|
2495 |
"group_size": 4096,
|
2496 |
"in_features": 4096,
|
|
|
2514 |
"model.layers.21.self_attn.q_proj": {
|
2515 |
"bias": false,
|
2516 |
"enable_norm": true,
|
2517 |
+
"enable_perm": false,
|
2518 |
"group_num": 1,
|
2519 |
"group_size": 4096,
|
2520 |
"in_features": 4096,
|
|
|
2538 |
"model.layers.21.self_attn.v_proj": {
|
2539 |
"bias": false,
|
2540 |
"enable_norm": true,
|
2541 |
+
"enable_perm": false,
|
2542 |
"group_num": 1,
|
2543 |
"group_size": 4096,
|
2544 |
"in_features": 4096,
|
|
|
2562 |
"model.layers.22.mlp.down_proj": {
|
2563 |
"bias": false,
|
2564 |
"enable_norm": true,
|
2565 |
+
"enable_perm": false,
|
2566 |
"group_num": 1,
|
2567 |
"group_size": 14336,
|
2568 |
"in_features": 14336,
|
|
|
2586 |
"model.layers.22.mlp.gate_proj": {
|
2587 |
"bias": false,
|
2588 |
"enable_norm": true,
|
2589 |
+
"enable_perm": false,
|
2590 |
"group_num": 1,
|
2591 |
"group_size": 4096,
|
2592 |
"in_features": 4096,
|
|
|
2610 |
"model.layers.22.mlp.up_proj": {
|
2611 |
"bias": false,
|
2612 |
"enable_norm": true,
|
2613 |
+
"enable_perm": false,
|
2614 |
"group_num": 1,
|
2615 |
"group_size": 4096,
|
2616 |
"in_features": 4096,
|
|
|
2634 |
"model.layers.22.self_attn.k_proj": {
|
2635 |
"bias": false,
|
2636 |
"enable_norm": true,
|
2637 |
+
"enable_perm": false,
|
2638 |
"group_num": 1,
|
2639 |
"group_size": 4096,
|
2640 |
"in_features": 4096,
|
|
|
2658 |
"model.layers.22.self_attn.o_proj": {
|
2659 |
"bias": false,
|
2660 |
"enable_norm": true,
|
2661 |
+
"enable_perm": false,
|
2662 |
"group_num": 1,
|
2663 |
"group_size": 4096,
|
2664 |
"in_features": 4096,
|
|
|
2682 |
"model.layers.22.self_attn.q_proj": {
|
2683 |
"bias": false,
|
2684 |
"enable_norm": true,
|
2685 |
+
"enable_perm": false,
|
2686 |
"group_num": 1,
|
2687 |
"group_size": 4096,
|
2688 |
"in_features": 4096,
|
|
|
2706 |
"model.layers.22.self_attn.v_proj": {
|
2707 |
"bias": false,
|
2708 |
"enable_norm": true,
|
2709 |
+
"enable_perm": false,
|
2710 |
"group_num": 1,
|
2711 |
"group_size": 4096,
|
2712 |
"in_features": 4096,
|
|
|
2730 |
"model.layers.23.mlp.down_proj": {
|
2731 |
"bias": false,
|
2732 |
"enable_norm": true,
|
2733 |
+
"enable_perm": false,
|
2734 |
"group_num": 1,
|
2735 |
"group_size": 14336,
|
2736 |
"in_features": 14336,
|
|
|
2754 |
"model.layers.23.mlp.gate_proj": {
|
2755 |
"bias": false,
|
2756 |
"enable_norm": true,
|
2757 |
+
"enable_perm": false,
|
2758 |
"group_num": 1,
|
2759 |
"group_size": 4096,
|
2760 |
"in_features": 4096,
|
|
|
2778 |
"model.layers.23.mlp.up_proj": {
|
2779 |
"bias": false,
|
2780 |
"enable_norm": true,
|
2781 |
+
"enable_perm": false,
|
2782 |
"group_num": 1,
|
2783 |
"group_size": 4096,
|
2784 |
"in_features": 4096,
|
|
|
2802 |
"model.layers.23.self_attn.k_proj": {
|
2803 |
"bias": false,
|
2804 |
"enable_norm": true,
|
2805 |
+
"enable_perm": false,
|
2806 |
"group_num": 1,
|
2807 |
"group_size": 4096,
|
2808 |
"in_features": 4096,
|
|
|
2826 |
"model.layers.23.self_attn.o_proj": {
|
2827 |
"bias": false,
|
2828 |
"enable_norm": true,
|
2829 |
+
"enable_perm": false,
|
2830 |
"group_num": 1,
|
2831 |
"group_size": 4096,
|
2832 |
"in_features": 4096,
|
|
|
2850 |
"model.layers.23.self_attn.q_proj": {
|
2851 |
"bias": false,
|
2852 |
"enable_norm": true,
|
2853 |
+
"enable_perm": false,
|
2854 |
"group_num": 1,
|
2855 |
"group_size": 4096,
|
2856 |
"in_features": 4096,
|
|
|
2874 |
"model.layers.23.self_attn.v_proj": {
|
2875 |
"bias": false,
|
2876 |
"enable_norm": true,
|
2877 |
+
"enable_perm": false,
|
2878 |
"group_num": 1,
|
2879 |
"group_size": 4096,
|
2880 |
"in_features": 4096,
|
|
|
2898 |
"model.layers.24.mlp.down_proj": {
|
2899 |
"bias": false,
|
2900 |
"enable_norm": true,
|
2901 |
+
"enable_perm": false,
|
2902 |
"group_num": 1,
|
2903 |
"group_size": 14336,
|
2904 |
"in_features": 14336,
|
|
|
2922 |
"model.layers.24.mlp.gate_proj": {
|
2923 |
"bias": false,
|
2924 |
"enable_norm": true,
|
2925 |
+
"enable_perm": false,
|
2926 |
"group_num": 1,
|
2927 |
"group_size": 4096,
|
2928 |
"in_features": 4096,
|
|
|
2946 |
"model.layers.24.mlp.up_proj": {
|
2947 |
"bias": false,
|
2948 |
"enable_norm": true,
|
2949 |
+
"enable_perm": false,
|
2950 |
"group_num": 1,
|
2951 |
"group_size": 4096,
|
2952 |
"in_features": 4096,
|
|
|
2970 |
"model.layers.24.self_attn.k_proj": {
|
2971 |
"bias": false,
|
2972 |
"enable_norm": true,
|
2973 |
+
"enable_perm": false,
|
2974 |
"group_num": 1,
|
2975 |
"group_size": 4096,
|
2976 |
"in_features": 4096,
|
|
|
2994 |
"model.layers.24.self_attn.o_proj": {
|
2995 |
"bias": false,
|
2996 |
"enable_norm": true,
|
2997 |
+
"enable_perm": false,
|
2998 |
"group_num": 1,
|
2999 |
"group_size": 4096,
|
3000 |
"in_features": 4096,
|
|
|
3018 |
"model.layers.24.self_attn.q_proj": {
|
3019 |
"bias": false,
|
3020 |
"enable_norm": true,
|
3021 |
+
"enable_perm": false,
|
3022 |
"group_num": 1,
|
3023 |
"group_size": 4096,
|
3024 |
"in_features": 4096,
|
|
|
3042 |
"model.layers.24.self_attn.v_proj": {
|
3043 |
"bias": false,
|
3044 |
"enable_norm": true,
|
3045 |
+
"enable_perm": false,
|
3046 |
"group_num": 1,
|
3047 |
"group_size": 4096,
|
3048 |
"in_features": 4096,
|
|
|
3066 |
"model.layers.25.mlp.down_proj": {
|
3067 |
"bias": false,
|
3068 |
"enable_norm": true,
|
3069 |
+
"enable_perm": false,
|
3070 |
"group_num": 1,
|
3071 |
"group_size": 14336,
|
3072 |
"in_features": 14336,
|
|
|
3090 |
"model.layers.25.mlp.gate_proj": {
|
3091 |
"bias": false,
|
3092 |
"enable_norm": true,
|
3093 |
+
"enable_perm": false,
|
3094 |
"group_num": 1,
|
3095 |
"group_size": 4096,
|
3096 |
"in_features": 4096,
|
|
|
3114 |
"model.layers.25.mlp.up_proj": {
|
3115 |
"bias": false,
|
3116 |
"enable_norm": true,
|
3117 |
+
"enable_perm": false,
|
3118 |
"group_num": 1,
|
3119 |
"group_size": 4096,
|
3120 |
"in_features": 4096,
|
|
|
3138 |
"model.layers.25.self_attn.k_proj": {
|
3139 |
"bias": false,
|
3140 |
"enable_norm": true,
|
3141 |
+
"enable_perm": false,
|
3142 |
"group_num": 1,
|
3143 |
"group_size": 4096,
|
3144 |
"in_features": 4096,
|
|
|
3162 |
"model.layers.25.self_attn.o_proj": {
|
3163 |
"bias": false,
|
3164 |
"enable_norm": true,
|
3165 |
+
"enable_perm": false,
|
3166 |
"group_num": 1,
|
3167 |
"group_size": 4096,
|
3168 |
"in_features": 4096,
|
|
|
3186 |
"model.layers.25.self_attn.q_proj": {
|
3187 |
"bias": false,
|
3188 |
"enable_norm": true,
|
3189 |
+
"enable_perm": false,
|
3190 |
"group_num": 1,
|
3191 |
"group_size": 4096,
|
3192 |
"in_features": 4096,
|
|
|
3210 |
"model.layers.25.self_attn.v_proj": {
|
3211 |
"bias": false,
|
3212 |
"enable_norm": true,
|
3213 |
+
"enable_perm": false,
|
3214 |
"group_num": 1,
|
3215 |
"group_size": 4096,
|
3216 |
"in_features": 4096,
|
|
|
3234 |
"model.layers.26.mlp.down_proj": {
|
3235 |
"bias": false,
|
3236 |
"enable_norm": true,
|
3237 |
+
"enable_perm": false,
|
3238 |
"group_num": 1,
|
3239 |
"group_size": 14336,
|
3240 |
"in_features": 14336,
|
|
|
3258 |
"model.layers.26.mlp.gate_proj": {
|
3259 |
"bias": false,
|
3260 |
"enable_norm": true,
|
3261 |
+
"enable_perm": false,
|
3262 |
"group_num": 1,
|
3263 |
"group_size": 4096,
|
3264 |
"in_features": 4096,
|
|
|
3282 |
"model.layers.26.mlp.up_proj": {
|
3283 |
"bias": false,
|
3284 |
"enable_norm": true,
|
3285 |
+
"enable_perm": false,
|
3286 |
"group_num": 1,
|
3287 |
"group_size": 4096,
|
3288 |
"in_features": 4096,
|
|
|
3306 |
"model.layers.26.self_attn.k_proj": {
|
3307 |
"bias": false,
|
3308 |
"enable_norm": true,
|
3309 |
+
"enable_perm": false,
|
3310 |
"group_num": 1,
|
3311 |
"group_size": 4096,
|
3312 |
"in_features": 4096,
|
|
|
3330 |
"model.layers.26.self_attn.o_proj": {
|
3331 |
"bias": false,
|
3332 |
"enable_norm": true,
|
3333 |
+
"enable_perm": false,
|
3334 |
"group_num": 1,
|
3335 |
"group_size": 4096,
|
3336 |
"in_features": 4096,
|
|
|
3354 |
"model.layers.26.self_attn.q_proj": {
|
3355 |
"bias": false,
|
3356 |
"enable_norm": true,
|
3357 |
+
"enable_perm": false,
|
3358 |
"group_num": 1,
|
3359 |
"group_size": 4096,
|
3360 |
"in_features": 4096,
|
|
|
3378 |
"model.layers.26.self_attn.v_proj": {
|
3379 |
"bias": false,
|
3380 |
"enable_norm": true,
|
3381 |
+
"enable_perm": false,
|
3382 |
"group_num": 1,
|
3383 |
"group_size": 4096,
|
3384 |
"in_features": 4096,
|
|
|
3402 |
"model.layers.27.mlp.down_proj": {
|
3403 |
"bias": false,
|
3404 |
"enable_norm": true,
|
3405 |
+
"enable_perm": false,
|
3406 |
"group_num": 1,
|
3407 |
"group_size": 14336,
|
3408 |
"in_features": 14336,
|
|
|
3426 |
"model.layers.27.mlp.gate_proj": {
|
3427 |
"bias": false,
|
3428 |
"enable_norm": true,
|
3429 |
+
"enable_perm": false,
|
3430 |
"group_num": 1,
|
3431 |
"group_size": 4096,
|
3432 |
"in_features": 4096,
|
|
|
3450 |
"model.layers.27.mlp.up_proj": {
|
3451 |
"bias": false,
|
3452 |
"enable_norm": true,
|
3453 |
+
"enable_perm": false,
|
3454 |
"group_num": 1,
|
3455 |
"group_size": 4096,
|
3456 |
"in_features": 4096,
|
|
|
3474 |
"model.layers.27.self_attn.k_proj": {
|
3475 |
"bias": false,
|
3476 |
"enable_norm": true,
|
3477 |
+
"enable_perm": false,
|
3478 |
"group_num": 1,
|
3479 |
"group_size": 4096,
|
3480 |
"in_features": 4096,
|
|
|
3498 |
"model.layers.27.self_attn.o_proj": {
|
3499 |
"bias": false,
|
3500 |
"enable_norm": true,
|
3501 |
+
"enable_perm": false,
|
3502 |
"group_num": 1,
|
3503 |
"group_size": 4096,
|
3504 |
"in_features": 4096,
|
|
|
3522 |
"model.layers.27.self_attn.q_proj": {
|
3523 |
"bias": false,
|
3524 |
"enable_norm": true,
|
3525 |
+
"enable_perm": false,
|
3526 |
"group_num": 1,
|
3527 |
"group_size": 4096,
|
3528 |
"in_features": 4096,
|
|
|
3546 |
"model.layers.27.self_attn.v_proj": {
|
3547 |
"bias": false,
|
3548 |
"enable_norm": true,
|
3549 |
+
"enable_perm": false,
|
3550 |
"group_num": 1,
|
3551 |
"group_size": 4096,
|
3552 |
"in_features": 4096,
|
|
|
3570 |
"model.layers.28.mlp.down_proj": {
|
3571 |
"bias": false,
|
3572 |
"enable_norm": true,
|
3573 |
+
"enable_perm": false,
|
3574 |
"group_num": 1,
|
3575 |
"group_size": 14336,
|
3576 |
"in_features": 14336,
|
|
|
3594 |
"model.layers.28.mlp.gate_proj": {
|
3595 |
"bias": false,
|
3596 |
"enable_norm": true,
|
3597 |
+
"enable_perm": false,
|
3598 |
"group_num": 1,
|
3599 |
"group_size": 4096,
|
3600 |
"in_features": 4096,
|
|
|
3618 |
"model.layers.28.mlp.up_proj": {
|
3619 |
"bias": false,
|
3620 |
"enable_norm": true,
|
3621 |
+
"enable_perm": false,
|
3622 |
"group_num": 1,
|
3623 |
"group_size": 4096,
|
3624 |
"in_features": 4096,
|
|
|
3642 |
"model.layers.28.self_attn.k_proj": {
|
3643 |
"bias": false,
|
3644 |
"enable_norm": true,
|
3645 |
+
"enable_perm": false,
|
3646 |
"group_num": 1,
|
3647 |
"group_size": 4096,
|
3648 |
"in_features": 4096,
|
|
|
3666 |
"model.layers.28.self_attn.o_proj": {
|
3667 |
"bias": false,
|
3668 |
"enable_norm": true,
|
3669 |
+
"enable_perm": false,
|
3670 |
"group_num": 1,
|
3671 |
"group_size": 4096,
|
3672 |
"in_features": 4096,
|
|
|
3690 |
"model.layers.28.self_attn.q_proj": {
|
3691 |
"bias": false,
|
3692 |
"enable_norm": true,
|
3693 |
+
"enable_perm": false,
|
3694 |
"group_num": 1,
|
3695 |
"group_size": 4096,
|
3696 |
"in_features": 4096,
|
|
|
3714 |
"model.layers.28.self_attn.v_proj": {
|
3715 |
"bias": false,
|
3716 |
"enable_norm": true,
|
3717 |
+
"enable_perm": false,
|
3718 |
"group_num": 1,
|
3719 |
"group_size": 4096,
|
3720 |
"in_features": 4096,
|
|
|
3738 |
"model.layers.29.mlp.down_proj": {
|
3739 |
"bias": false,
|
3740 |
"enable_norm": true,
|
3741 |
+
"enable_perm": false,
|
3742 |
"group_num": 1,
|
3743 |
"group_size": 14336,
|
3744 |
"in_features": 14336,
|
|
|
3762 |
"model.layers.29.mlp.gate_proj": {
|
3763 |
"bias": false,
|
3764 |
"enable_norm": true,
|
3765 |
+
"enable_perm": false,
|
3766 |
"group_num": 1,
|
3767 |
"group_size": 4096,
|
3768 |
"in_features": 4096,
|
|
|
3786 |
"model.layers.29.mlp.up_proj": {
|
3787 |
"bias": false,
|
3788 |
"enable_norm": true,
|
3789 |
+
"enable_perm": false,
|
3790 |
"group_num": 1,
|
3791 |
"group_size": 4096,
|
3792 |
"in_features": 4096,
|
|
|
3810 |
"model.layers.29.self_attn.k_proj": {
|
3811 |
"bias": false,
|
3812 |
"enable_norm": true,
|
3813 |
+
"enable_perm": false,
|
3814 |
"group_num": 1,
|
3815 |
"group_size": 4096,
|
3816 |
"in_features": 4096,
|
|
|
3834 |
"model.layers.29.self_attn.o_proj": {
|
3835 |
"bias": false,
|
3836 |
"enable_norm": true,
|
3837 |
+
"enable_perm": false,
|
3838 |
"group_num": 1,
|
3839 |
"group_size": 4096,
|
3840 |
"in_features": 4096,
|
|
|
3858 |
"model.layers.29.self_attn.q_proj": {
|
3859 |
"bias": false,
|
3860 |
"enable_norm": true,
|
3861 |
+
"enable_perm": false,
|
3862 |
"group_num": 1,
|
3863 |
"group_size": 4096,
|
3864 |
"in_features": 4096,
|
|
|
3882 |
"model.layers.29.self_attn.v_proj": {
|
3883 |
"bias": false,
|
3884 |
"enable_norm": true,
|
3885 |
+
"enable_perm": false,
|
3886 |
"group_num": 1,
|
3887 |
"group_size": 4096,
|
3888 |
"in_features": 4096,
|
|
|
3906 |
"model.layers.3.mlp.down_proj": {
|
3907 |
"bias": false,
|
3908 |
"enable_norm": true,
|
3909 |
+
"enable_perm": false,
|
3910 |
"group_num": 1,
|
3911 |
"group_size": 14336,
|
3912 |
"in_features": 14336,
|
|
|
3930 |
"model.layers.3.mlp.gate_proj": {
|
3931 |
"bias": false,
|
3932 |
"enable_norm": true,
|
3933 |
+
"enable_perm": false,
|
3934 |
"group_num": 1,
|
3935 |
"group_size": 4096,
|
3936 |
"in_features": 4096,
|
|
|
3954 |
"model.layers.3.mlp.up_proj": {
|
3955 |
"bias": false,
|
3956 |
"enable_norm": true,
|
3957 |
+
"enable_perm": false,
|
3958 |
"group_num": 1,
|
3959 |
"group_size": 4096,
|
3960 |
"in_features": 4096,
|
|
|
3978 |
"model.layers.3.self_attn.k_proj": {
|
3979 |
"bias": false,
|
3980 |
"enable_norm": true,
|
3981 |
+
"enable_perm": false,
|
3982 |
"group_num": 1,
|
3983 |
"group_size": 4096,
|
3984 |
"in_features": 4096,
|
|
|
4002 |
"model.layers.3.self_attn.o_proj": {
|
4003 |
"bias": false,
|
4004 |
"enable_norm": true,
|
4005 |
+
"enable_perm": false,
|
4006 |
"group_num": 1,
|
4007 |
"group_size": 4096,
|
4008 |
"in_features": 4096,
|
|
|
4026 |
"model.layers.3.self_attn.q_proj": {
|
4027 |
"bias": false,
|
4028 |
"enable_norm": true,
|
4029 |
+
"enable_perm": false,
|
4030 |
"group_num": 1,
|
4031 |
"group_size": 4096,
|
4032 |
"in_features": 4096,
|
|
|
4050 |
"model.layers.3.self_attn.v_proj": {
|
4051 |
"bias": false,
|
4052 |
"enable_norm": true,
|
4053 |
+
"enable_perm": false,
|
4054 |
"group_num": 1,
|
4055 |
"group_size": 4096,
|
4056 |
"in_features": 4096,
|
|
|
4074 |
"model.layers.30.mlp.down_proj": {
|
4075 |
"bias": false,
|
4076 |
"enable_norm": true,
|
4077 |
+
"enable_perm": false,
|
4078 |
"group_num": 1,
|
4079 |
"group_size": 14336,
|
4080 |
"in_features": 14336,
|
|
|
4098 |
"model.layers.30.mlp.gate_proj": {
|
4099 |
"bias": false,
|
4100 |
"enable_norm": true,
|
4101 |
+
"enable_perm": false,
|
4102 |
"group_num": 1,
|
4103 |
"group_size": 4096,
|
4104 |
"in_features": 4096,
|
|
|
4122 |
"model.layers.30.mlp.up_proj": {
|
4123 |
"bias": false,
|
4124 |
"enable_norm": true,
|
4125 |
+
"enable_perm": false,
|
4126 |
"group_num": 1,
|
4127 |
"group_size": 4096,
|
4128 |
"in_features": 4096,
|
|
|
4146 |
"model.layers.30.self_attn.k_proj": {
|
4147 |
"bias": false,
|
4148 |
"enable_norm": true,
|
4149 |
+
"enable_perm": false,
|
4150 |
"group_num": 1,
|
4151 |
"group_size": 4096,
|
4152 |
"in_features": 4096,
|
|
|
4170 |
"model.layers.30.self_attn.o_proj": {
|
4171 |
"bias": false,
|
4172 |
"enable_norm": true,
|
4173 |
+
"enable_perm": false,
|
4174 |
"group_num": 1,
|
4175 |
"group_size": 4096,
|
4176 |
"in_features": 4096,
|
|
|
4194 |
"model.layers.30.self_attn.q_proj": {
|
4195 |
"bias": false,
|
4196 |
"enable_norm": true,
|
4197 |
+
"enable_perm": false,
|
4198 |
"group_num": 1,
|
4199 |
"group_size": 4096,
|
4200 |
"in_features": 4096,
|
|
|
4218 |
"model.layers.30.self_attn.v_proj": {
|
4219 |
"bias": false,
|
4220 |
"enable_norm": true,
|
4221 |
+
"enable_perm": false,
|
4222 |
"group_num": 1,
|
4223 |
"group_size": 4096,
|
4224 |
"in_features": 4096,
|
|
|
4242 |
"model.layers.31.mlp.down_proj": {
|
4243 |
"bias": false,
|
4244 |
"enable_norm": true,
|
4245 |
+
"enable_perm": false,
|
4246 |
"group_num": 1,
|
4247 |
"group_size": 14336,
|
4248 |
"in_features": 14336,
|
|
|
4266 |
"model.layers.31.mlp.gate_proj": {
|
4267 |
"bias": false,
|
4268 |
"enable_norm": true,
|
4269 |
+
"enable_perm": false,
|
4270 |
"group_num": 1,
|
4271 |
"group_size": 4096,
|
4272 |
"in_features": 4096,
|
|
|
4290 |
"model.layers.31.mlp.up_proj": {
|
4291 |
"bias": false,
|
4292 |
"enable_norm": true,
|
4293 |
+
"enable_perm": false,
|
4294 |
"group_num": 1,
|
4295 |
"group_size": 4096,
|
4296 |
"in_features": 4096,
|
|
|
4314 |
"model.layers.31.self_attn.k_proj": {
|
4315 |
"bias": false,
|
4316 |
"enable_norm": true,
|
4317 |
+
"enable_perm": false,
|
4318 |
"group_num": 1,
|
4319 |
"group_size": 4096,
|
4320 |
"in_features": 4096,
|
|
|
4338 |
"model.layers.31.self_attn.o_proj": {
|
4339 |
"bias": false,
|
4340 |
"enable_norm": true,
|
4341 |
+
"enable_perm": false,
|
4342 |
"group_num": 1,
|
4343 |
"group_size": 4096,
|
4344 |
"in_features": 4096,
|
|
|
4362 |
"model.layers.31.self_attn.q_proj": {
|
4363 |
"bias": false,
|
4364 |
"enable_norm": true,
|
4365 |
+
"enable_perm": false,
|
4366 |
"group_num": 1,
|
4367 |
"group_size": 4096,
|
4368 |
"in_features": 4096,
|
|
|
4386 |
"model.layers.31.self_attn.v_proj": {
|
4387 |
"bias": false,
|
4388 |
"enable_norm": true,
|
4389 |
+
"enable_perm": false,
|
4390 |
"group_num": 1,
|
4391 |
"group_size": 4096,
|
4392 |
"in_features": 4096,
|
|
|
4410 |
"model.layers.4.mlp.down_proj": {
|
4411 |
"bias": false,
|
4412 |
"enable_norm": true,
|
4413 |
+
"enable_perm": false,
|
4414 |
"group_num": 1,
|
4415 |
"group_size": 14336,
|
4416 |
"in_features": 14336,
|
|
|
4434 |
"model.layers.4.mlp.gate_proj": {
|
4435 |
"bias": false,
|
4436 |
"enable_norm": true,
|
4437 |
+
"enable_perm": false,
|
4438 |
"group_num": 1,
|
4439 |
"group_size": 4096,
|
4440 |
"in_features": 4096,
|
|
|
4458 |
"model.layers.4.mlp.up_proj": {
|
4459 |
"bias": false,
|
4460 |
"enable_norm": true,
|
4461 |
+
"enable_perm": false,
|
4462 |
"group_num": 1,
|
4463 |
"group_size": 4096,
|
4464 |
"in_features": 4096,
|
|
|
4482 |
"model.layers.4.self_attn.k_proj": {
|
4483 |
"bias": false,
|
4484 |
"enable_norm": true,
|
4485 |
+
"enable_perm": false,
|
4486 |
"group_num": 1,
|
4487 |
"group_size": 4096,
|
4488 |
"in_features": 4096,
|
|
|
4506 |
"model.layers.4.self_attn.o_proj": {
|
4507 |
"bias": false,
|
4508 |
"enable_norm": true,
|
4509 |
+
"enable_perm": false,
|
4510 |
"group_num": 1,
|
4511 |
"group_size": 4096,
|
4512 |
"in_features": 4096,
|
|
|
4530 |
"model.layers.4.self_attn.q_proj": {
|
4531 |
"bias": false,
|
4532 |
"enable_norm": true,
|
4533 |
+
"enable_perm": false,
|
4534 |
"group_num": 1,
|
4535 |
"group_size": 4096,
|
4536 |
"in_features": 4096,
|
|
|
4554 |
"model.layers.4.self_attn.v_proj": {
|
4555 |
"bias": false,
|
4556 |
"enable_norm": true,
|
4557 |
+
"enable_perm": false,
|
4558 |
"group_num": 1,
|
4559 |
"group_size": 4096,
|
4560 |
"in_features": 4096,
|
|
|
4578 |
"model.layers.5.mlp.down_proj": {
|
4579 |
"bias": false,
|
4580 |
"enable_norm": true,
|
4581 |
+
"enable_perm": false,
|
4582 |
"group_num": 1,
|
4583 |
"group_size": 14336,
|
4584 |
"in_features": 14336,
|
|
|
4602 |
"model.layers.5.mlp.gate_proj": {
|
4603 |
"bias": false,
|
4604 |
"enable_norm": true,
|
4605 |
+
"enable_perm": false,
|
4606 |
"group_num": 1,
|
4607 |
"group_size": 4096,
|
4608 |
"in_features": 4096,
|
|
|
4626 |
"model.layers.5.mlp.up_proj": {
|
4627 |
"bias": false,
|
4628 |
"enable_norm": true,
|
4629 |
+
"enable_perm": false,
|
4630 |
"group_num": 1,
|
4631 |
"group_size": 4096,
|
4632 |
"in_features": 4096,
|
|
|
4650 |
"model.layers.5.self_attn.k_proj": {
|
4651 |
"bias": false,
|
4652 |
"enable_norm": true,
|
4653 |
+
"enable_perm": false,
|
4654 |
"group_num": 1,
|
4655 |
"group_size": 4096,
|
4656 |
"in_features": 4096,
|
|
|
4674 |
"model.layers.5.self_attn.o_proj": {
|
4675 |
"bias": false,
|
4676 |
"enable_norm": true,
|
4677 |
+
"enable_perm": false,
|
4678 |
"group_num": 1,
|
4679 |
"group_size": 4096,
|
4680 |
"in_features": 4096,
|
|
|
4698 |
"model.layers.5.self_attn.q_proj": {
|
4699 |
"bias": false,
|
4700 |
"enable_norm": true,
|
4701 |
+
"enable_perm": false,
|
4702 |
"group_num": 1,
|
4703 |
"group_size": 4096,
|
4704 |
"in_features": 4096,
|
|
|
4722 |
"model.layers.5.self_attn.v_proj": {
|
4723 |
"bias": false,
|
4724 |
"enable_norm": true,
|
4725 |
+
"enable_perm": false,
|
4726 |
"group_num": 1,
|
4727 |
"group_size": 4096,
|
4728 |
"in_features": 4096,
|
|
|
4746 |
"model.layers.6.mlp.down_proj": {
|
4747 |
"bias": false,
|
4748 |
"enable_norm": true,
|
4749 |
+
"enable_perm": false,
|
4750 |
"group_num": 1,
|
4751 |
"group_size": 14336,
|
4752 |
"in_features": 14336,
|
|
|
4770 |
"model.layers.6.mlp.gate_proj": {
|
4771 |
"bias": false,
|
4772 |
"enable_norm": true,
|
4773 |
+
"enable_perm": false,
|
4774 |
"group_num": 1,
|
4775 |
"group_size": 4096,
|
4776 |
"in_features": 4096,
|
|
|
4794 |
"model.layers.6.mlp.up_proj": {
|
4795 |
"bias": false,
|
4796 |
"enable_norm": true,
|
4797 |
+
"enable_perm": false,
|
4798 |
"group_num": 1,
|
4799 |
"group_size": 4096,
|
4800 |
"in_features": 4096,
|
|
|
4818 |
"model.layers.6.self_attn.k_proj": {
|
4819 |
"bias": false,
|
4820 |
"enable_norm": true,
|
4821 |
+
"enable_perm": false,
|
4822 |
"group_num": 1,
|
4823 |
"group_size": 4096,
|
4824 |
"in_features": 4096,
|
|
|
4842 |
"model.layers.6.self_attn.o_proj": {
|
4843 |
"bias": false,
|
4844 |
"enable_norm": true,
|
4845 |
+
"enable_perm": false,
|
4846 |
"group_num": 1,
|
4847 |
"group_size": 4096,
|
4848 |
"in_features": 4096,
|
|
|
4866 |
"model.layers.6.self_attn.q_proj": {
|
4867 |
"bias": false,
|
4868 |
"enable_norm": true,
|
4869 |
+
"enable_perm": false,
|
4870 |
"group_num": 1,
|
4871 |
"group_size": 4096,
|
4872 |
"in_features": 4096,
|
|
|
4890 |
"model.layers.6.self_attn.v_proj": {
|
4891 |
"bias": false,
|
4892 |
"enable_norm": true,
|
4893 |
+
"enable_perm": false,
|
4894 |
"group_num": 1,
|
4895 |
"group_size": 4096,
|
4896 |
"in_features": 4096,
|
|
|
4914 |
"model.layers.7.mlp.down_proj": {
|
4915 |
"bias": false,
|
4916 |
"enable_norm": true,
|
4917 |
+
"enable_perm": false,
|
4918 |
"group_num": 1,
|
4919 |
"group_size": 14336,
|
4920 |
"in_features": 14336,
|
|
|
4938 |
"model.layers.7.mlp.gate_proj": {
|
4939 |
"bias": false,
|
4940 |
"enable_norm": true,
|
4941 |
+
"enable_perm": false,
|
4942 |
"group_num": 1,
|
4943 |
"group_size": 4096,
|
4944 |
"in_features": 4096,
|
|
|
4962 |
"model.layers.7.mlp.up_proj": {
|
4963 |
"bias": false,
|
4964 |
"enable_norm": true,
|
4965 |
+
"enable_perm": false,
|
4966 |
"group_num": 1,
|
4967 |
"group_size": 4096,
|
4968 |
"in_features": 4096,
|
|
|
4986 |
"model.layers.7.self_attn.k_proj": {
|
4987 |
"bias": false,
|
4988 |
"enable_norm": true,
|
4989 |
+
"enable_perm": false,
|
4990 |
"group_num": 1,
|
4991 |
"group_size": 4096,
|
4992 |
"in_features": 4096,
|
|
|
5010 |
"model.layers.7.self_attn.o_proj": {
|
5011 |
"bias": false,
|
5012 |
"enable_norm": true,
|
5013 |
+
"enable_perm": false,
|
5014 |
"group_num": 1,
|
5015 |
"group_size": 4096,
|
5016 |
"in_features": 4096,
|
|
|
5034 |
"model.layers.7.self_attn.q_proj": {
|
5035 |
"bias": false,
|
5036 |
"enable_norm": true,
|
5037 |
+
"enable_perm": false,
|
5038 |
"group_num": 1,
|
5039 |
"group_size": 4096,
|
5040 |
"in_features": 4096,
|
|
|
5058 |
"model.layers.7.self_attn.v_proj": {
|
5059 |
"bias": false,
|
5060 |
"enable_norm": true,
|
5061 |
+
"enable_perm": false,
|
5062 |
"group_num": 1,
|
5063 |
"group_size": 4096,
|
5064 |
"in_features": 4096,
|
|
|
5082 |
"model.layers.8.mlp.down_proj": {
|
5083 |
"bias": false,
|
5084 |
"enable_norm": true,
|
5085 |
+
"enable_perm": false,
|
5086 |
"group_num": 1,
|
5087 |
"group_size": 14336,
|
5088 |
"in_features": 14336,
|
|
|
5106 |
"model.layers.8.mlp.gate_proj": {
|
5107 |
"bias": false,
|
5108 |
"enable_norm": true,
|
5109 |
+
"enable_perm": false,
|
5110 |
"group_num": 1,
|
5111 |
"group_size": 4096,
|
5112 |
"in_features": 4096,
|
|
|
5130 |
"model.layers.8.mlp.up_proj": {
|
5131 |
"bias": false,
|
5132 |
"enable_norm": true,
|
5133 |
+
"enable_perm": false,
|
5134 |
"group_num": 1,
|
5135 |
"group_size": 4096,
|
5136 |
"in_features": 4096,
|
|
|
5154 |
"model.layers.8.self_attn.k_proj": {
|
5155 |
"bias": false,
|
5156 |
"enable_norm": true,
|
5157 |
+
"enable_perm": false,
|
5158 |
"group_num": 1,
|
5159 |
"group_size": 4096,
|
5160 |
"in_features": 4096,
|
|
|
5178 |
"model.layers.8.self_attn.o_proj": {
|
5179 |
"bias": false,
|
5180 |
"enable_norm": true,
|
5181 |
+
"enable_perm": false,
|
5182 |
"group_num": 1,
|
5183 |
"group_size": 4096,
|
5184 |
"in_features": 4096,
|
|
|
5202 |
"model.layers.8.self_attn.q_proj": {
|
5203 |
"bias": false,
|
5204 |
"enable_norm": true,
|
5205 |
+
"enable_perm": false,
|
5206 |
"group_num": 1,
|
5207 |
"group_size": 4096,
|
5208 |
"in_features": 4096,
|
|
|
5226 |
"model.layers.8.self_attn.v_proj": {
|
5227 |
"bias": false,
|
5228 |
"enable_norm": true,
|
5229 |
+
"enable_perm": false,
|
5230 |
"group_num": 1,
|
5231 |
"group_size": 4096,
|
5232 |
"in_features": 4096,
|
|
|
5250 |
"model.layers.9.mlp.down_proj": {
|
5251 |
"bias": false,
|
5252 |
"enable_norm": true,
|
5253 |
+
"enable_perm": false,
|
5254 |
"group_num": 1,
|
5255 |
"group_size": 14336,
|
5256 |
"in_features": 14336,
|
|
|
5274 |
"model.layers.9.mlp.gate_proj": {
|
5275 |
"bias": false,
|
5276 |
"enable_norm": true,
|
5277 |
+
"enable_perm": false,
|
5278 |
"group_num": 1,
|
5279 |
"group_size": 4096,
|
5280 |
"in_features": 4096,
|
|
|
5298 |
"model.layers.9.mlp.up_proj": {
|
5299 |
"bias": false,
|
5300 |
"enable_norm": true,
|
5301 |
+
"enable_perm": false,
|
5302 |
"group_num": 1,
|
5303 |
"group_size": 4096,
|
5304 |
"in_features": 4096,
|
|
|
5322 |
"model.layers.9.self_attn.k_proj": {
|
5323 |
"bias": false,
|
5324 |
"enable_norm": true,
|
5325 |
+
"enable_perm": false,
|
5326 |
"group_num": 1,
|
5327 |
"group_size": 4096,
|
5328 |
"in_features": 4096,
|
|
|
5346 |
"model.layers.9.self_attn.o_proj": {
|
5347 |
"bias": false,
|
5348 |
"enable_norm": true,
|
5349 |
+
"enable_perm": false,
|
5350 |
"group_num": 1,
|
5351 |
"group_size": 4096,
|
5352 |
"in_features": 4096,
|
|
|
5370 |
"model.layers.9.self_attn.q_proj": {
|
5371 |
"bias": false,
|
5372 |
"enable_norm": true,
|
5373 |
+
"enable_perm": false,
|
5374 |
"group_num": 1,
|
5375 |
"group_size": 4096,
|
5376 |
"in_features": 4096,
|
|
|
5394 |
"model.layers.9.self_attn.v_proj": {
|
5395 |
"bias": false,
|
5396 |
"enable_norm": true,
|
5397 |
+
"enable_perm": false,
|
5398 |
"group_num": 1,
|
5399 |
"group_size": 4096,
|
5400 |
"in_features": 4096,
|