File size: 26,397 Bytes
f7c32cc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
{
  "model.embed_tokens.weight": "p0.model.embed_tokens.weight",
  "model.layers.0.self_attn.q_proj.weight": "p1.model.layers.0.self_attn.q_proj.weight",
  "model.layers.0.self_attn.k_proj.weight": "p2.model.layers.0.self_attn.k_proj.weight",
  "model.layers.0.self_attn.v_proj.weight": "p3.model.layers.0.self_attn.v_proj.weight",
  "model.layers.0.self_attn.o_proj.weight": "p4.model.layers.0.self_attn.o_proj.weight",
  "model.layers.0.mlp.gate_proj.weight": "p5.model.layers.0.mlp.gate_proj.weight",
  "model.layers.0.mlp.up_proj.weight": "p6.model.layers.0.mlp.up_proj.weight",
  "model.layers.0.mlp.down_proj.weight": "p7.model.layers.0.mlp.down_proj.weight",
  "model.layers.0.input_layernorm.weight": "p8.model.layers.0.input_layernorm.weight",
  "model.layers.0.post_attention_layernorm.weight": "p9.model.layers.0.post_attention_layernorm.weight",
  "model.layers.1.self_attn.q_proj.weight": "p10.model.layers.1.self_attn.q_proj.weight",
  "model.layers.1.self_attn.k_proj.weight": "p11.model.layers.1.self_attn.k_proj.weight",
  "model.layers.1.self_attn.v_proj.weight": "p12.model.layers.1.self_attn.v_proj.weight",
  "model.layers.1.self_attn.o_proj.weight": "p13.model.layers.1.self_attn.o_proj.weight",
  "model.layers.1.mlp.gate_proj.weight": "p14.model.layers.1.mlp.gate_proj.weight",
  "model.layers.1.mlp.up_proj.weight": "p15.model.layers.1.mlp.up_proj.weight",
  "model.layers.1.mlp.down_proj.weight": "p16.model.layers.1.mlp.down_proj.weight",
  "model.layers.1.input_layernorm.weight": "p17.model.layers.1.input_layernorm.weight",
  "model.layers.1.post_attention_layernorm.weight": "p18.model.layers.1.post_attention_layernorm.weight",
  "model.layers.2.self_attn.q_proj.weight": "p19.model.layers.2.self_attn.q_proj.weight",
  "model.layers.2.self_attn.k_proj.weight": "p20.model.layers.2.self_attn.k_proj.weight",
  "model.layers.2.self_attn.v_proj.weight": "p21.model.layers.2.self_attn.v_proj.weight",
  "model.layers.2.self_attn.o_proj.weight": "p22.model.layers.2.self_attn.o_proj.weight",
  "model.layers.2.mlp.gate_proj.weight": "p23.model.layers.2.mlp.gate_proj.weight",
  "model.layers.2.mlp.up_proj.weight": "p24.model.layers.2.mlp.up_proj.weight",
  "model.layers.2.mlp.down_proj.weight": "p25.model.layers.2.mlp.down_proj.weight",
  "model.layers.2.input_layernorm.weight": "p26.model.layers.2.input_layernorm.weight",
  "model.layers.2.post_attention_layernorm.weight": "p27.model.layers.2.post_attention_layernorm.weight",
  "model.layers.3.self_attn.q_proj.weight": "p28.model.layers.3.self_attn.q_proj.weight",
  "model.layers.3.self_attn.k_proj.weight": "p29.model.layers.3.self_attn.k_proj.weight",
  "model.layers.3.self_attn.v_proj.weight": "p30.model.layers.3.self_attn.v_proj.weight",
  "model.layers.3.self_attn.o_proj.weight": "p31.model.layers.3.self_attn.o_proj.weight",
  "model.layers.3.mlp.gate_proj.weight": "p32.model.layers.3.mlp.gate_proj.weight",
  "model.layers.3.mlp.up_proj.weight": "p33.model.layers.3.mlp.up_proj.weight",
  "model.layers.3.mlp.down_proj.weight": "p34.model.layers.3.mlp.down_proj.weight",
  "model.layers.3.input_layernorm.weight": "p35.model.layers.3.input_layernorm.weight",
  "model.layers.3.post_attention_layernorm.weight": "p36.model.layers.3.post_attention_layernorm.weight",
  "model.layers.4.self_attn.q_proj.weight": "p37.model.layers.4.self_attn.q_proj.weight",
  "model.layers.4.self_attn.k_proj.weight": "p38.model.layers.4.self_attn.k_proj.weight",
  "model.layers.4.self_attn.v_proj.weight": "p39.model.layers.4.self_attn.v_proj.weight",
  "model.layers.4.self_attn.o_proj.weight": "p40.model.layers.4.self_attn.o_proj.weight",
  "model.layers.4.mlp.gate_proj.weight": "p41.model.layers.4.mlp.gate_proj.weight",
  "model.layers.4.mlp.up_proj.weight": "p42.model.layers.4.mlp.up_proj.weight",
  "model.layers.4.mlp.down_proj.weight": "p43.model.layers.4.mlp.down_proj.weight",
  "model.layers.4.input_layernorm.weight": "p44.model.layers.4.input_layernorm.weight",
  "model.layers.4.post_attention_layernorm.weight": "p45.model.layers.4.post_attention_layernorm.weight",
  "model.layers.5.self_attn.q_proj.weight": "p46.model.layers.5.self_attn.q_proj.weight",
  "model.layers.5.self_attn.k_proj.weight": "p47.model.layers.5.self_attn.k_proj.weight",
  "model.layers.5.self_attn.v_proj.weight": "p48.model.layers.5.self_attn.v_proj.weight",
  "model.layers.5.self_attn.o_proj.weight": "p49.model.layers.5.self_attn.o_proj.weight",
  "model.layers.5.mlp.gate_proj.weight": "p50.model.layers.5.mlp.gate_proj.weight",
  "model.layers.5.mlp.up_proj.weight": "p51.model.layers.5.mlp.up_proj.weight",
  "model.layers.5.mlp.down_proj.weight": "p52.model.layers.5.mlp.down_proj.weight",
  "model.layers.5.input_layernorm.weight": "p53.model.layers.5.input_layernorm.weight",
  "model.layers.5.post_attention_layernorm.weight": "p54.model.layers.5.post_attention_layernorm.weight",
  "model.layers.6.self_attn.q_proj.weight": "p55.model.layers.6.self_attn.q_proj.weight",
  "model.layers.6.self_attn.k_proj.weight": "p56.model.layers.6.self_attn.k_proj.weight",
  "model.layers.6.self_attn.v_proj.weight": "p57.model.layers.6.self_attn.v_proj.weight",
  "model.layers.6.self_attn.o_proj.weight": "p58.model.layers.6.self_attn.o_proj.weight",
  "model.layers.6.mlp.gate_proj.weight": "p59.model.layers.6.mlp.gate_proj.weight",
  "model.layers.6.mlp.up_proj.weight": "p60.model.layers.6.mlp.up_proj.weight",
  "model.layers.6.mlp.down_proj.weight": "p61.model.layers.6.mlp.down_proj.weight",
  "model.layers.6.input_layernorm.weight": "p62.model.layers.6.input_layernorm.weight",
  "model.layers.6.post_attention_layernorm.weight": "p63.model.layers.6.post_attention_layernorm.weight",
  "model.layers.7.self_attn.q_proj.weight": "p64.model.layers.7.self_attn.q_proj.weight",
  "model.layers.7.self_attn.k_proj.weight": "p65.model.layers.7.self_attn.k_proj.weight",
  "model.layers.7.self_attn.v_proj.weight": "p66.model.layers.7.self_attn.v_proj.weight",
  "model.layers.7.self_attn.o_proj.weight": "p67.model.layers.7.self_attn.o_proj.weight",
  "model.layers.7.mlp.gate_proj.weight": "p68.model.layers.7.mlp.gate_proj.weight",
  "model.layers.7.mlp.up_proj.weight": "p69.model.layers.7.mlp.up_proj.weight",
  "model.layers.7.mlp.down_proj.weight": "p70.model.layers.7.mlp.down_proj.weight",
  "model.layers.7.input_layernorm.weight": "p71.model.layers.7.input_layernorm.weight",
  "model.layers.7.post_attention_layernorm.weight": "p72.model.layers.7.post_attention_layernorm.weight",
  "model.layers.8.self_attn.q_proj.weight": "p73.model.layers.8.self_attn.q_proj.weight",
  "model.layers.8.self_attn.k_proj.weight": "p74.model.layers.8.self_attn.k_proj.weight",
  "model.layers.8.self_attn.v_proj.weight": "p75.model.layers.8.self_attn.v_proj.weight",
  "model.layers.8.self_attn.o_proj.weight": "p76.model.layers.8.self_attn.o_proj.weight",
  "model.layers.8.mlp.gate_proj.weight": "p77.model.layers.8.mlp.gate_proj.weight",
  "model.layers.8.mlp.up_proj.weight": "p78.model.layers.8.mlp.up_proj.weight",
  "model.layers.8.mlp.down_proj.weight": "p79.model.layers.8.mlp.down_proj.weight",
  "model.layers.8.input_layernorm.weight": "p80.model.layers.8.input_layernorm.weight",
  "model.layers.8.post_attention_layernorm.weight": "p81.model.layers.8.post_attention_layernorm.weight",
  "model.layers.9.self_attn.q_proj.weight": "p82.model.layers.9.self_attn.q_proj.weight",
  "model.layers.9.self_attn.k_proj.weight": "p83.model.layers.9.self_attn.k_proj.weight",
  "model.layers.9.self_attn.v_proj.weight": "p84.model.layers.9.self_attn.v_proj.weight",
  "model.layers.9.self_attn.o_proj.weight": "p85.model.layers.9.self_attn.o_proj.weight",
  "model.layers.9.mlp.gate_proj.weight": "p86.model.layers.9.mlp.gate_proj.weight",
  "model.layers.9.mlp.up_proj.weight": "p87.model.layers.9.mlp.up_proj.weight",
  "model.layers.9.mlp.down_proj.weight": "p88.model.layers.9.mlp.down_proj.weight",
  "model.layers.9.input_layernorm.weight": "p89.model.layers.9.input_layernorm.weight",
  "model.layers.9.post_attention_layernorm.weight": "p90.model.layers.9.post_attention_layernorm.weight",
  "model.layers.10.self_attn.q_proj.weight": "p91.model.layers.10.self_attn.q_proj.weight",
  "model.layers.10.self_attn.k_proj.weight": "p92.model.layers.10.self_attn.k_proj.weight",
  "model.layers.10.self_attn.v_proj.weight": "p93.model.layers.10.self_attn.v_proj.weight",
  "model.layers.10.self_attn.o_proj.weight": "p94.model.layers.10.self_attn.o_proj.weight",
  "model.layers.10.mlp.gate_proj.weight": "p95.model.layers.10.mlp.gate_proj.weight",
  "model.layers.10.mlp.up_proj.weight": "p96.model.layers.10.mlp.up_proj.weight",
  "model.layers.10.mlp.down_proj.weight": "p97.model.layers.10.mlp.down_proj.weight",
  "model.layers.10.input_layernorm.weight": "p98.model.layers.10.input_layernorm.weight",
  "model.layers.10.post_attention_layernorm.weight": "p99.model.layers.10.post_attention_layernorm.weight",
  "model.layers.11.self_attn.q_proj.weight": "p100.model.layers.11.self_attn.q_proj.weight",
  "model.layers.11.self_attn.k_proj.weight": "p101.model.layers.11.self_attn.k_proj.weight",
  "model.layers.11.self_attn.v_proj.weight": "p102.model.layers.11.self_attn.v_proj.weight",
  "model.layers.11.self_attn.o_proj.weight": "p103.model.layers.11.self_attn.o_proj.weight",
  "model.layers.11.mlp.gate_proj.weight": "p104.model.layers.11.mlp.gate_proj.weight",
  "model.layers.11.mlp.up_proj.weight": "p105.model.layers.11.mlp.up_proj.weight",
  "model.layers.11.mlp.down_proj.weight": "p106.model.layers.11.mlp.down_proj.weight",
  "model.layers.11.input_layernorm.weight": "p107.model.layers.11.input_layernorm.weight",
  "model.layers.11.post_attention_layernorm.weight": "p108.model.layers.11.post_attention_layernorm.weight",
  "model.layers.12.self_attn.q_proj.weight": "p109.model.layers.12.self_attn.q_proj.weight",
  "model.layers.12.self_attn.k_proj.weight": "p110.model.layers.12.self_attn.k_proj.weight",
  "model.layers.12.self_attn.v_proj.weight": "p111.model.layers.12.self_attn.v_proj.weight",
  "model.layers.12.self_attn.o_proj.weight": "p112.model.layers.12.self_attn.o_proj.weight",
  "model.layers.12.mlp.gate_proj.weight": "p113.model.layers.12.mlp.gate_proj.weight",
  "model.layers.12.mlp.up_proj.weight": "p114.model.layers.12.mlp.up_proj.weight",
  "model.layers.12.mlp.down_proj.weight": "p115.model.layers.12.mlp.down_proj.weight",
  "model.layers.12.input_layernorm.weight": "p116.model.layers.12.input_layernorm.weight",
  "model.layers.12.post_attention_layernorm.weight": "p117.model.layers.12.post_attention_layernorm.weight",
  "model.layers.13.self_attn.q_proj.weight": "p118.model.layers.13.self_attn.q_proj.weight",
  "model.layers.13.self_attn.k_proj.weight": "p119.model.layers.13.self_attn.k_proj.weight",
  "model.layers.13.self_attn.v_proj.weight": "p120.model.layers.13.self_attn.v_proj.weight",
  "model.layers.13.self_attn.o_proj.weight": "p121.model.layers.13.self_attn.o_proj.weight",
  "model.layers.13.mlp.gate_proj.weight": "p122.model.layers.13.mlp.gate_proj.weight",
  "model.layers.13.mlp.up_proj.weight": "p123.model.layers.13.mlp.up_proj.weight",
  "model.layers.13.mlp.down_proj.weight": "p124.model.layers.13.mlp.down_proj.weight",
  "model.layers.13.input_layernorm.weight": "p125.model.layers.13.input_layernorm.weight",
  "model.layers.13.post_attention_layernorm.weight": "p126.model.layers.13.post_attention_layernorm.weight",
  "model.layers.14.self_attn.q_proj.weight": "p127.model.layers.14.self_attn.q_proj.weight",
  "model.layers.14.self_attn.k_proj.weight": "p128.model.layers.14.self_attn.k_proj.weight",
  "model.layers.14.self_attn.v_proj.weight": "p129.model.layers.14.self_attn.v_proj.weight",
  "model.layers.14.self_attn.o_proj.weight": "p130.model.layers.14.self_attn.o_proj.weight",
  "model.layers.14.mlp.gate_proj.weight": "p131.model.layers.14.mlp.gate_proj.weight",
  "model.layers.14.mlp.up_proj.weight": "p132.model.layers.14.mlp.up_proj.weight",
  "model.layers.14.mlp.down_proj.weight": "p133.model.layers.14.mlp.down_proj.weight",
  "model.layers.14.input_layernorm.weight": "p134.model.layers.14.input_layernorm.weight",
  "model.layers.14.post_attention_layernorm.weight": "p135.model.layers.14.post_attention_layernorm.weight",
  "model.layers.15.self_attn.q_proj.weight": "p136.model.layers.15.self_attn.q_proj.weight",
  "model.layers.15.self_attn.k_proj.weight": "p137.model.layers.15.self_attn.k_proj.weight",
  "model.layers.15.self_attn.v_proj.weight": "p138.model.layers.15.self_attn.v_proj.weight",
  "model.layers.15.self_attn.o_proj.weight": "p139.model.layers.15.self_attn.o_proj.weight",
  "model.layers.15.mlp.gate_proj.weight": "p140.model.layers.15.mlp.gate_proj.weight",
  "model.layers.15.mlp.up_proj.weight": "p141.model.layers.15.mlp.up_proj.weight",
  "model.layers.15.mlp.down_proj.weight": "p142.model.layers.15.mlp.down_proj.weight",
  "model.layers.15.input_layernorm.weight": "p143.model.layers.15.input_layernorm.weight",
  "model.layers.15.post_attention_layernorm.weight": "p144.model.layers.15.post_attention_layernorm.weight",
  "model.layers.16.self_attn.q_proj.weight": "p145.model.layers.16.self_attn.q_proj.weight",
  "model.layers.16.self_attn.k_proj.weight": "p146.model.layers.16.self_attn.k_proj.weight",
  "model.layers.16.self_attn.v_proj.weight": "p147.model.layers.16.self_attn.v_proj.weight",
  "model.layers.16.self_attn.o_proj.weight": "p148.model.layers.16.self_attn.o_proj.weight",
  "model.layers.16.mlp.gate_proj.weight": "p149.model.layers.16.mlp.gate_proj.weight",
  "model.layers.16.mlp.up_proj.weight": "p150.model.layers.16.mlp.up_proj.weight",
  "model.layers.16.mlp.down_proj.weight": "p151.model.layers.16.mlp.down_proj.weight",
  "model.layers.16.input_layernorm.weight": "p152.model.layers.16.input_layernorm.weight",
  "model.layers.16.post_attention_layernorm.weight": "p153.model.layers.16.post_attention_layernorm.weight",
  "model.layers.17.self_attn.q_proj.weight": "p154.model.layers.17.self_attn.q_proj.weight",
  "model.layers.17.self_attn.k_proj.weight": "p155.model.layers.17.self_attn.k_proj.weight",
  "model.layers.17.self_attn.v_proj.weight": "p156.model.layers.17.self_attn.v_proj.weight",
  "model.layers.17.self_attn.o_proj.weight": "p157.model.layers.17.self_attn.o_proj.weight",
  "model.layers.17.mlp.gate_proj.weight": "p158.model.layers.17.mlp.gate_proj.weight",
  "model.layers.17.mlp.up_proj.weight": "p159.model.layers.17.mlp.up_proj.weight",
  "model.layers.17.mlp.down_proj.weight": "p160.model.layers.17.mlp.down_proj.weight",
  "model.layers.17.input_layernorm.weight": "p161.model.layers.17.input_layernorm.weight",
  "model.layers.17.post_attention_layernorm.weight": "p162.model.layers.17.post_attention_layernorm.weight",
  "model.layers.18.self_attn.q_proj.weight": "p163.model.layers.18.self_attn.q_proj.weight",
  "model.layers.18.self_attn.k_proj.weight": "p164.model.layers.18.self_attn.k_proj.weight",
  "model.layers.18.self_attn.v_proj.weight": "p165.model.layers.18.self_attn.v_proj.weight",
  "model.layers.18.self_attn.o_proj.weight": "p166.model.layers.18.self_attn.o_proj.weight",
  "model.layers.18.mlp.gate_proj.weight": "p167.model.layers.18.mlp.gate_proj.weight",
  "model.layers.18.mlp.up_proj.weight": "p168.model.layers.18.mlp.up_proj.weight",
  "model.layers.18.mlp.down_proj.weight": "p169.model.layers.18.mlp.down_proj.weight",
  "model.layers.18.input_layernorm.weight": "p170.model.layers.18.input_layernorm.weight",
  "model.layers.18.post_attention_layernorm.weight": "p171.model.layers.18.post_attention_layernorm.weight",
  "model.layers.19.self_attn.q_proj.weight": "p172.model.layers.19.self_attn.q_proj.weight",
  "model.layers.19.self_attn.k_proj.weight": "p173.model.layers.19.self_attn.k_proj.weight",
  "model.layers.19.self_attn.v_proj.weight": "p174.model.layers.19.self_attn.v_proj.weight",
  "model.layers.19.self_attn.o_proj.weight": "p175.model.layers.19.self_attn.o_proj.weight",
  "model.layers.19.mlp.gate_proj.weight": "p176.model.layers.19.mlp.gate_proj.weight",
  "model.layers.19.mlp.up_proj.weight": "p177.model.layers.19.mlp.up_proj.weight",
  "model.layers.19.mlp.down_proj.weight": "p178.model.layers.19.mlp.down_proj.weight",
  "model.layers.19.input_layernorm.weight": "p179.model.layers.19.input_layernorm.weight",
  "model.layers.19.post_attention_layernorm.weight": "p180.model.layers.19.post_attention_layernorm.weight",
  "model.layers.20.self_attn.q_proj.weight": "p181.model.layers.20.self_attn.q_proj.weight",
  "model.layers.20.self_attn.k_proj.weight": "p182.model.layers.20.self_attn.k_proj.weight",
  "model.layers.20.self_attn.v_proj.weight": "p183.model.layers.20.self_attn.v_proj.weight",
  "model.layers.20.self_attn.o_proj.weight": "p184.model.layers.20.self_attn.o_proj.weight",
  "model.layers.20.mlp.gate_proj.weight": "p185.model.layers.20.mlp.gate_proj.weight",
  "model.layers.20.mlp.up_proj.weight": "p186.model.layers.20.mlp.up_proj.weight",
  "model.layers.20.mlp.down_proj.weight": "p187.model.layers.20.mlp.down_proj.weight",
  "model.layers.20.input_layernorm.weight": "p188.model.layers.20.input_layernorm.weight",
  "model.layers.20.post_attention_layernorm.weight": "p189.model.layers.20.post_attention_layernorm.weight",
  "model.layers.21.self_attn.q_proj.weight": "p190.model.layers.21.self_attn.q_proj.weight",
  "model.layers.21.self_attn.k_proj.weight": "p191.model.layers.21.self_attn.k_proj.weight",
  "model.layers.21.self_attn.v_proj.weight": "p192.model.layers.21.self_attn.v_proj.weight",
  "model.layers.21.self_attn.o_proj.weight": "p193.model.layers.21.self_attn.o_proj.weight",
  "model.layers.21.mlp.gate_proj.weight": "p194.model.layers.21.mlp.gate_proj.weight",
  "model.layers.21.mlp.up_proj.weight": "p195.model.layers.21.mlp.up_proj.weight",
  "model.layers.21.mlp.down_proj.weight": "p196.model.layers.21.mlp.down_proj.weight",
  "model.layers.21.input_layernorm.weight": "p197.model.layers.21.input_layernorm.weight",
  "model.layers.21.post_attention_layernorm.weight": "p198.model.layers.21.post_attention_layernorm.weight",
  "model.layers.22.self_attn.q_proj.weight": "p199.model.layers.22.self_attn.q_proj.weight",
  "model.layers.22.self_attn.k_proj.weight": "p200.model.layers.22.self_attn.k_proj.weight",
  "model.layers.22.self_attn.v_proj.weight": "p201.model.layers.22.self_attn.v_proj.weight",
  "model.layers.22.self_attn.o_proj.weight": "p202.model.layers.22.self_attn.o_proj.weight",
  "model.layers.22.mlp.gate_proj.weight": "p203.model.layers.22.mlp.gate_proj.weight",
  "model.layers.22.mlp.up_proj.weight": "p204.model.layers.22.mlp.up_proj.weight",
  "model.layers.22.mlp.down_proj.weight": "p205.model.layers.22.mlp.down_proj.weight",
  "model.layers.22.input_layernorm.weight": "p206.model.layers.22.input_layernorm.weight",
  "model.layers.22.post_attention_layernorm.weight": "p207.model.layers.22.post_attention_layernorm.weight",
  "model.layers.23.self_attn.q_proj.weight": "p208.model.layers.23.self_attn.q_proj.weight",
  "model.layers.23.self_attn.k_proj.weight": "p209.model.layers.23.self_attn.k_proj.weight",
  "model.layers.23.self_attn.v_proj.weight": "p210.model.layers.23.self_attn.v_proj.weight",
  "model.layers.23.self_attn.o_proj.weight": "p211.model.layers.23.self_attn.o_proj.weight",
  "model.layers.23.mlp.gate_proj.weight": "p212.model.layers.23.mlp.gate_proj.weight",
  "model.layers.23.mlp.up_proj.weight": "p213.model.layers.23.mlp.up_proj.weight",
  "model.layers.23.mlp.down_proj.weight": "p214.model.layers.23.mlp.down_proj.weight",
  "model.layers.23.input_layernorm.weight": "p215.model.layers.23.input_layernorm.weight",
  "model.layers.23.post_attention_layernorm.weight": "p216.model.layers.23.post_attention_layernorm.weight",
  "model.layers.24.self_attn.q_proj.weight": "p217.model.layers.24.self_attn.q_proj.weight",
  "model.layers.24.self_attn.k_proj.weight": "p218.model.layers.24.self_attn.k_proj.weight",
  "model.layers.24.self_attn.v_proj.weight": "p219.model.layers.24.self_attn.v_proj.weight",
  "model.layers.24.self_attn.o_proj.weight": "p220.model.layers.24.self_attn.o_proj.weight",
  "model.layers.24.mlp.gate_proj.weight": "p221.model.layers.24.mlp.gate_proj.weight",
  "model.layers.24.mlp.up_proj.weight": "p222.model.layers.24.mlp.up_proj.weight",
  "model.layers.24.mlp.down_proj.weight": "p223.model.layers.24.mlp.down_proj.weight",
  "model.layers.24.input_layernorm.weight": "p224.model.layers.24.input_layernorm.weight",
  "model.layers.24.post_attention_layernorm.weight": "p225.model.layers.24.post_attention_layernorm.weight",
  "model.layers.25.self_attn.q_proj.weight": "p226.model.layers.25.self_attn.q_proj.weight",
  "model.layers.25.self_attn.k_proj.weight": "p227.model.layers.25.self_attn.k_proj.weight",
  "model.layers.25.self_attn.v_proj.weight": "p228.model.layers.25.self_attn.v_proj.weight",
  "model.layers.25.self_attn.o_proj.weight": "p229.model.layers.25.self_attn.o_proj.weight",
  "model.layers.25.mlp.gate_proj.weight": "p230.model.layers.25.mlp.gate_proj.weight",
  "model.layers.25.mlp.up_proj.weight": "p231.model.layers.25.mlp.up_proj.weight",
  "model.layers.25.mlp.down_proj.weight": "p232.model.layers.25.mlp.down_proj.weight",
  "model.layers.25.input_layernorm.weight": "p233.model.layers.25.input_layernorm.weight",
  "model.layers.25.post_attention_layernorm.weight": "p234.model.layers.25.post_attention_layernorm.weight",
  "model.layers.26.self_attn.q_proj.weight": "p235.model.layers.26.self_attn.q_proj.weight",
  "model.layers.26.self_attn.k_proj.weight": "p236.model.layers.26.self_attn.k_proj.weight",
  "model.layers.26.self_attn.v_proj.weight": "p237.model.layers.26.self_attn.v_proj.weight",
  "model.layers.26.self_attn.o_proj.weight": "p238.model.layers.26.self_attn.o_proj.weight",
  "model.layers.26.mlp.gate_proj.weight": "p239.model.layers.26.mlp.gate_proj.weight",
  "model.layers.26.mlp.up_proj.weight": "p240.model.layers.26.mlp.up_proj.weight",
  "model.layers.26.mlp.down_proj.weight": "p241.model.layers.26.mlp.down_proj.weight",
  "model.layers.26.input_layernorm.weight": "p242.model.layers.26.input_layernorm.weight",
  "model.layers.26.post_attention_layernorm.weight": "p243.model.layers.26.post_attention_layernorm.weight",
  "model.layers.27.self_attn.q_proj.weight": "p244.model.layers.27.self_attn.q_proj.weight",
  "model.layers.27.self_attn.k_proj.weight": "p245.model.layers.27.self_attn.k_proj.weight",
  "model.layers.27.self_attn.v_proj.weight": "p246.model.layers.27.self_attn.v_proj.weight",
  "model.layers.27.self_attn.o_proj.weight": "p247.model.layers.27.self_attn.o_proj.weight",
  "model.layers.27.mlp.gate_proj.weight": "p248.model.layers.27.mlp.gate_proj.weight",
  "model.layers.27.mlp.up_proj.weight": "p249.model.layers.27.mlp.up_proj.weight",
  "model.layers.27.mlp.down_proj.weight": "p250.model.layers.27.mlp.down_proj.weight",
  "model.layers.27.input_layernorm.weight": "p251.model.layers.27.input_layernorm.weight",
  "model.layers.27.post_attention_layernorm.weight": "p252.model.layers.27.post_attention_layernorm.weight",
  "model.layers.28.self_attn.q_proj.weight": "p253.model.layers.28.self_attn.q_proj.weight",
  "model.layers.28.self_attn.k_proj.weight": "p254.model.layers.28.self_attn.k_proj.weight",
  "model.layers.28.self_attn.v_proj.weight": "p255.model.layers.28.self_attn.v_proj.weight",
  "model.layers.28.self_attn.o_proj.weight": "p256.model.layers.28.self_attn.o_proj.weight",
  "model.layers.28.mlp.gate_proj.weight": "p257.model.layers.28.mlp.gate_proj.weight",
  "model.layers.28.mlp.up_proj.weight": "p258.model.layers.28.mlp.up_proj.weight",
  "model.layers.28.mlp.down_proj.weight": "p259.model.layers.28.mlp.down_proj.weight",
  "model.layers.28.input_layernorm.weight": "p260.model.layers.28.input_layernorm.weight",
  "model.layers.28.post_attention_layernorm.weight": "p261.model.layers.28.post_attention_layernorm.weight",
  "model.layers.29.self_attn.q_proj.weight": "p262.model.layers.29.self_attn.q_proj.weight",
  "model.layers.29.self_attn.k_proj.weight": "p263.model.layers.29.self_attn.k_proj.weight",
  "model.layers.29.self_attn.v_proj.weight": "p264.model.layers.29.self_attn.v_proj.weight",
  "model.layers.29.self_attn.o_proj.weight": "p265.model.layers.29.self_attn.o_proj.weight",
  "model.layers.29.mlp.gate_proj.weight": "p266.model.layers.29.mlp.gate_proj.weight",
  "model.layers.29.mlp.up_proj.weight": "p267.model.layers.29.mlp.up_proj.weight",
  "model.layers.29.mlp.down_proj.weight": "p268.model.layers.29.mlp.down_proj.weight",
  "model.layers.29.input_layernorm.weight": "p269.model.layers.29.input_layernorm.weight",
  "model.layers.29.post_attention_layernorm.weight": "p270.model.layers.29.post_attention_layernorm.weight",
  "model.layers.30.self_attn.q_proj.weight": "p271.model.layers.30.self_attn.q_proj.weight",
  "model.layers.30.self_attn.k_proj.weight": "p272.model.layers.30.self_attn.k_proj.weight",
  "model.layers.30.self_attn.v_proj.weight": "p273.model.layers.30.self_attn.v_proj.weight",
  "model.layers.30.self_attn.o_proj.weight": "p274.model.layers.30.self_attn.o_proj.weight",
  "model.layers.30.mlp.gate_proj.weight": "p275.model.layers.30.mlp.gate_proj.weight",
  "model.layers.30.mlp.up_proj.weight": "p276.model.layers.30.mlp.up_proj.weight",
  "model.layers.30.mlp.down_proj.weight": "p277.model.layers.30.mlp.down_proj.weight",
  "model.layers.30.input_layernorm.weight": "p278.model.layers.30.input_layernorm.weight",
  "model.layers.30.post_attention_layernorm.weight": "p279.model.layers.30.post_attention_layernorm.weight",
  "model.layers.31.self_attn.q_proj.weight": "p280.model.layers.31.self_attn.q_proj.weight",
  "model.layers.31.self_attn.k_proj.weight": "p281.model.layers.31.self_attn.k_proj.weight",
  "model.layers.31.self_attn.v_proj.weight": "p282.model.layers.31.self_attn.v_proj.weight",
  "model.layers.31.self_attn.o_proj.weight": "p283.model.layers.31.self_attn.o_proj.weight",
  "model.layers.31.mlp.gate_proj.weight": "p284.model.layers.31.mlp.gate_proj.weight",
  "model.layers.31.mlp.up_proj.weight": "p285.model.layers.31.mlp.up_proj.weight",
  "model.layers.31.mlp.down_proj.weight": "p286.model.layers.31.mlp.down_proj.weight",
  "model.layers.31.input_layernorm.weight": "p287.model.layers.31.input_layernorm.weight",
  "model.layers.31.post_attention_layernorm.weight": "p288.model.layers.31.post_attention_layernorm.weight",
  "model.norm.weight": "p289.model.norm.weight",
  "lm_head.weight": "p290.lm_head.weight"
}