maldv commited on
Commit
6fc9bfa
1 Parent(s): 47bf701

Upload model.safetensors.index.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. model.safetensors.index.json +128 -128
model.safetensors.index.json CHANGED
@@ -17,23 +17,11 @@
17
  "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00017.safetensors",
18
  "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00017.safetensors",
19
  "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00017.safetensors",
20
- "model.layers.1.input_layernorm.weight": "model-00001-of-00017.safetensors",
21
- "model.layers.1.mlp.down_proj.weight": "model-00001-of-00017.safetensors",
22
- "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00017.safetensors",
23
- "model.layers.1.mlp.up_proj.weight": "model-00001-of-00017.safetensors",
24
- "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00017.safetensors",
25
- "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00017.safetensors",
26
- "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00017.safetensors",
27
- "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00017.safetensors",
28
- "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00017.safetensors",
29
- "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00017.safetensors",
30
- "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00017.safetensors",
31
- "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00017.safetensors",
32
- "model.layers.10.input_layernorm.weight": "model-00003-of-00017.safetensors",
33
  "model.layers.10.mlp.down_proj.weight": "model-00004-of-00017.safetensors",
34
- "model.layers.10.mlp.gate_proj.weight": "model-00004-of-00017.safetensors",
35
- "model.layers.10.mlp.up_proj.weight": "model-00003-of-00017.safetensors",
36
- "model.layers.10.post_attention_layernorm.weight": "model-00003-of-00017.safetensors",
37
  "model.layers.10.self_attn.k_proj.bias": "model-00003-of-00017.safetensors",
38
  "model.layers.10.self_attn.k_proj.weight": "model-00003-of-00017.safetensors",
39
  "model.layers.10.self_attn.o_proj.weight": "model-00003-of-00017.safetensors",
@@ -77,11 +65,11 @@
77
  "model.layers.13.self_attn.q_proj.weight": "model-00004-of-00017.safetensors",
78
  "model.layers.13.self_attn.v_proj.bias": "model-00004-of-00017.safetensors",
79
  "model.layers.13.self_attn.v_proj.weight": "model-00004-of-00017.safetensors",
80
- "model.layers.14.input_layernorm.weight": "model-00004-of-00017.safetensors",
81
  "model.layers.14.mlp.down_proj.weight": "model-00005-of-00017.safetensors",
82
- "model.layers.14.mlp.gate_proj.weight": "model-00005-of-00017.safetensors",
83
- "model.layers.14.mlp.up_proj.weight": "model-00004-of-00017.safetensors",
84
- "model.layers.14.post_attention_layernorm.weight": "model-00004-of-00017.safetensors",
85
  "model.layers.14.self_attn.k_proj.bias": "model-00004-of-00017.safetensors",
86
  "model.layers.14.self_attn.k_proj.weight": "model-00004-of-00017.safetensors",
87
  "model.layers.14.self_attn.o_proj.weight": "model-00004-of-00017.safetensors",
@@ -125,11 +113,11 @@
125
  "model.layers.17.self_attn.q_proj.weight": "model-00005-of-00017.safetensors",
126
  "model.layers.17.self_attn.v_proj.bias": "model-00005-of-00017.safetensors",
127
  "model.layers.17.self_attn.v_proj.weight": "model-00005-of-00017.safetensors",
128
- "model.layers.18.input_layernorm.weight": "model-00005-of-00017.safetensors",
129
  "model.layers.18.mlp.down_proj.weight": "model-00006-of-00017.safetensors",
130
- "model.layers.18.mlp.gate_proj.weight": "model-00006-of-00017.safetensors",
131
- "model.layers.18.mlp.up_proj.weight": "model-00005-of-00017.safetensors",
132
- "model.layers.18.post_attention_layernorm.weight": "model-00005-of-00017.safetensors",
133
  "model.layers.18.self_attn.k_proj.bias": "model-00005-of-00017.safetensors",
134
  "model.layers.18.self_attn.k_proj.weight": "model-00005-of-00017.safetensors",
135
  "model.layers.18.self_attn.o_proj.weight": "model-00005-of-00017.safetensors",
@@ -149,18 +137,18 @@
149
  "model.layers.19.self_attn.q_proj.weight": "model-00006-of-00017.safetensors",
150
  "model.layers.19.self_attn.v_proj.bias": "model-00006-of-00017.safetensors",
151
  "model.layers.19.self_attn.v_proj.weight": "model-00006-of-00017.safetensors",
152
- "model.layers.2.input_layernorm.weight": "model-00001-of-00017.safetensors",
153
- "model.layers.2.mlp.down_proj.weight": "model-00002-of-00017.safetensors",
154
- "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00017.safetensors",
155
- "model.layers.2.mlp.up_proj.weight": "model-00001-of-00017.safetensors",
156
- "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00017.safetensors",
157
- "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00017.safetensors",
158
- "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00017.safetensors",
159
- "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00017.safetensors",
160
- "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00017.safetensors",
161
- "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00017.safetensors",
162
- "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00017.safetensors",
163
- "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00017.safetensors",
164
  "model.layers.20.input_layernorm.weight": "model-00006-of-00017.safetensors",
165
  "model.layers.20.mlp.down_proj.weight": "model-00006-of-00017.safetensors",
166
  "model.layers.20.mlp.gate_proj.weight": "model-00006-of-00017.safetensors",
@@ -185,11 +173,11 @@
185
  "model.layers.21.self_attn.q_proj.weight": "model-00006-of-00017.safetensors",
186
  "model.layers.21.self_attn.v_proj.bias": "model-00006-of-00017.safetensors",
187
  "model.layers.21.self_attn.v_proj.weight": "model-00006-of-00017.safetensors",
188
- "model.layers.22.input_layernorm.weight": "model-00006-of-00017.safetensors",
189
  "model.layers.22.mlp.down_proj.weight": "model-00007-of-00017.safetensors",
190
- "model.layers.22.mlp.gate_proj.weight": "model-00007-of-00017.safetensors",
191
- "model.layers.22.mlp.up_proj.weight": "model-00006-of-00017.safetensors",
192
- "model.layers.22.post_attention_layernorm.weight": "model-00006-of-00017.safetensors",
193
  "model.layers.22.self_attn.k_proj.bias": "model-00006-of-00017.safetensors",
194
  "model.layers.22.self_attn.k_proj.weight": "model-00006-of-00017.safetensors",
195
  "model.layers.22.self_attn.o_proj.weight": "model-00006-of-00017.safetensors",
@@ -233,11 +221,11 @@
233
  "model.layers.25.self_attn.q_proj.weight": "model-00007-of-00017.safetensors",
234
  "model.layers.25.self_attn.v_proj.bias": "model-00007-of-00017.safetensors",
235
  "model.layers.25.self_attn.v_proj.weight": "model-00007-of-00017.safetensors",
236
- "model.layers.26.input_layernorm.weight": "model-00007-of-00017.safetensors",
237
  "model.layers.26.mlp.down_proj.weight": "model-00008-of-00017.safetensors",
238
- "model.layers.26.mlp.gate_proj.weight": "model-00008-of-00017.safetensors",
239
- "model.layers.26.mlp.up_proj.weight": "model-00007-of-00017.safetensors",
240
- "model.layers.26.post_attention_layernorm.weight": "model-00007-of-00017.safetensors",
241
  "model.layers.26.self_attn.k_proj.bias": "model-00007-of-00017.safetensors",
242
  "model.layers.26.self_attn.k_proj.weight": "model-00007-of-00017.safetensors",
243
  "model.layers.26.self_attn.o_proj.weight": "model-00007-of-00017.safetensors",
@@ -281,23 +269,23 @@
281
  "model.layers.29.self_attn.q_proj.weight": "model-00008-of-00017.safetensors",
282
  "model.layers.29.self_attn.v_proj.bias": "model-00008-of-00017.safetensors",
283
  "model.layers.29.self_attn.v_proj.weight": "model-00008-of-00017.safetensors",
284
- "model.layers.3.input_layernorm.weight": "model-00002-of-00017.safetensors",
285
- "model.layers.3.mlp.down_proj.weight": "model-00002-of-00017.safetensors",
286
- "model.layers.3.mlp.gate_proj.weight": "model-00002-of-00017.safetensors",
287
- "model.layers.3.mlp.up_proj.weight": "model-00002-of-00017.safetensors",
288
- "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00017.safetensors",
289
- "model.layers.3.self_attn.k_proj.bias": "model-00002-of-00017.safetensors",
290
- "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00017.safetensors",
291
- "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00017.safetensors",
292
- "model.layers.3.self_attn.q_proj.bias": "model-00002-of-00017.safetensors",
293
- "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00017.safetensors",
294
- "model.layers.3.self_attn.v_proj.bias": "model-00002-of-00017.safetensors",
295
- "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00017.safetensors",
296
- "model.layers.30.input_layernorm.weight": "model-00008-of-00017.safetensors",
297
  "model.layers.30.mlp.down_proj.weight": "model-00009-of-00017.safetensors",
298
- "model.layers.30.mlp.gate_proj.weight": "model-00009-of-00017.safetensors",
299
- "model.layers.30.mlp.up_proj.weight": "model-00008-of-00017.safetensors",
300
- "model.layers.30.post_attention_layernorm.weight": "model-00008-of-00017.safetensors",
301
  "model.layers.30.self_attn.k_proj.bias": "model-00008-of-00017.safetensors",
302
  "model.layers.30.self_attn.k_proj.weight": "model-00008-of-00017.safetensors",
303
  "model.layers.30.self_attn.o_proj.weight": "model-00008-of-00017.safetensors",
@@ -341,11 +329,11 @@
341
  "model.layers.33.self_attn.q_proj.weight": "model-00009-of-00017.safetensors",
342
  "model.layers.33.self_attn.v_proj.bias": "model-00009-of-00017.safetensors",
343
  "model.layers.33.self_attn.v_proj.weight": "model-00009-of-00017.safetensors",
344
- "model.layers.34.input_layernorm.weight": "model-00009-of-00017.safetensors",
345
  "model.layers.34.mlp.down_proj.weight": "model-00010-of-00017.safetensors",
346
- "model.layers.34.mlp.gate_proj.weight": "model-00010-of-00017.safetensors",
347
- "model.layers.34.mlp.up_proj.weight": "model-00009-of-00017.safetensors",
348
- "model.layers.34.post_attention_layernorm.weight": "model-00009-of-00017.safetensors",
349
  "model.layers.34.self_attn.k_proj.bias": "model-00009-of-00017.safetensors",
350
  "model.layers.34.self_attn.k_proj.weight": "model-00009-of-00017.safetensors",
351
  "model.layers.34.self_attn.o_proj.weight": "model-00009-of-00017.safetensors",
@@ -389,11 +377,11 @@
389
  "model.layers.37.self_attn.q_proj.weight": "model-00010-of-00017.safetensors",
390
  "model.layers.37.self_attn.v_proj.bias": "model-00010-of-00017.safetensors",
391
  "model.layers.37.self_attn.v_proj.weight": "model-00010-of-00017.safetensors",
392
- "model.layers.38.input_layernorm.weight": "model-00010-of-00017.safetensors",
393
  "model.layers.38.mlp.down_proj.weight": "model-00011-of-00017.safetensors",
394
- "model.layers.38.mlp.gate_proj.weight": "model-00011-of-00017.safetensors",
395
- "model.layers.38.mlp.up_proj.weight": "model-00010-of-00017.safetensors",
396
- "model.layers.38.post_attention_layernorm.weight": "model-00010-of-00017.safetensors",
397
  "model.layers.38.self_attn.k_proj.bias": "model-00010-of-00017.safetensors",
398
  "model.layers.38.self_attn.k_proj.weight": "model-00010-of-00017.safetensors",
399
  "model.layers.38.self_attn.o_proj.weight": "model-00010-of-00017.safetensors",
@@ -413,18 +401,18 @@
413
  "model.layers.39.self_attn.q_proj.weight": "model-00011-of-00017.safetensors",
414
  "model.layers.39.self_attn.v_proj.bias": "model-00011-of-00017.safetensors",
415
  "model.layers.39.self_attn.v_proj.weight": "model-00011-of-00017.safetensors",
416
- "model.layers.4.input_layernorm.weight": "model-00002-of-00017.safetensors",
417
- "model.layers.4.mlp.down_proj.weight": "model-00002-of-00017.safetensors",
418
- "model.layers.4.mlp.gate_proj.weight": "model-00002-of-00017.safetensors",
419
- "model.layers.4.mlp.up_proj.weight": "model-00002-of-00017.safetensors",
420
- "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00017.safetensors",
421
- "model.layers.4.self_attn.k_proj.bias": "model-00002-of-00017.safetensors",
422
- "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00017.safetensors",
423
- "model.layers.4.self_attn.o_proj.weight": "model-00002-of-00017.safetensors",
424
- "model.layers.4.self_attn.q_proj.bias": "model-00002-of-00017.safetensors",
425
- "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00017.safetensors",
426
- "model.layers.4.self_attn.v_proj.bias": "model-00002-of-00017.safetensors",
427
- "model.layers.4.self_attn.v_proj.weight": "model-00002-of-00017.safetensors",
428
  "model.layers.40.input_layernorm.weight": "model-00011-of-00017.safetensors",
429
  "model.layers.40.mlp.down_proj.weight": "model-00011-of-00017.safetensors",
430
  "model.layers.40.mlp.gate_proj.weight": "model-00011-of-00017.safetensors",
@@ -449,11 +437,11 @@
449
  "model.layers.41.self_attn.q_proj.weight": "model-00011-of-00017.safetensors",
450
  "model.layers.41.self_attn.v_proj.bias": "model-00011-of-00017.safetensors",
451
  "model.layers.41.self_attn.v_proj.weight": "model-00011-of-00017.safetensors",
452
- "model.layers.42.input_layernorm.weight": "model-00011-of-00017.safetensors",
453
  "model.layers.42.mlp.down_proj.weight": "model-00012-of-00017.safetensors",
454
- "model.layers.42.mlp.gate_proj.weight": "model-00012-of-00017.safetensors",
455
- "model.layers.42.mlp.up_proj.weight": "model-00011-of-00017.safetensors",
456
- "model.layers.42.post_attention_layernorm.weight": "model-00011-of-00017.safetensors",
457
  "model.layers.42.self_attn.k_proj.bias": "model-00011-of-00017.safetensors",
458
  "model.layers.42.self_attn.k_proj.weight": "model-00011-of-00017.safetensors",
459
  "model.layers.42.self_attn.o_proj.weight": "model-00011-of-00017.safetensors",
@@ -497,11 +485,11 @@
497
  "model.layers.45.self_attn.q_proj.weight": "model-00012-of-00017.safetensors",
498
  "model.layers.45.self_attn.v_proj.bias": "model-00012-of-00017.safetensors",
499
  "model.layers.45.self_attn.v_proj.weight": "model-00012-of-00017.safetensors",
500
- "model.layers.46.input_layernorm.weight": "model-00012-of-00017.safetensors",
501
  "model.layers.46.mlp.down_proj.weight": "model-00013-of-00017.safetensors",
502
- "model.layers.46.mlp.gate_proj.weight": "model-00013-of-00017.safetensors",
503
- "model.layers.46.mlp.up_proj.weight": "model-00012-of-00017.safetensors",
504
- "model.layers.46.post_attention_layernorm.weight": "model-00012-of-00017.safetensors",
505
  "model.layers.46.self_attn.k_proj.bias": "model-00012-of-00017.safetensors",
506
  "model.layers.46.self_attn.k_proj.weight": "model-00012-of-00017.safetensors",
507
  "model.layers.46.self_attn.o_proj.weight": "model-00012-of-00017.safetensors",
@@ -545,23 +533,23 @@
545
  "model.layers.49.self_attn.q_proj.weight": "model-00013-of-00017.safetensors",
546
  "model.layers.49.self_attn.v_proj.bias": "model-00013-of-00017.safetensors",
547
  "model.layers.49.self_attn.v_proj.weight": "model-00013-of-00017.safetensors",
548
- "model.layers.5.input_layernorm.weight": "model-00002-of-00017.safetensors",
549
- "model.layers.5.mlp.down_proj.weight": "model-00002-of-00017.safetensors",
550
- "model.layers.5.mlp.gate_proj.weight": "model-00002-of-00017.safetensors",
551
- "model.layers.5.mlp.up_proj.weight": "model-00002-of-00017.safetensors",
552
- "model.layers.5.post_attention_layernorm.weight": "model-00002-of-00017.safetensors",
553
- "model.layers.5.self_attn.k_proj.bias": "model-00002-of-00017.safetensors",
554
- "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00017.safetensors",
555
- "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00017.safetensors",
556
- "model.layers.5.self_attn.q_proj.bias": "model-00002-of-00017.safetensors",
557
- "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00017.safetensors",
558
- "model.layers.5.self_attn.v_proj.bias": "model-00002-of-00017.safetensors",
559
- "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00017.safetensors",
560
- "model.layers.50.input_layernorm.weight": "model-00013-of-00017.safetensors",
561
  "model.layers.50.mlp.down_proj.weight": "model-00014-of-00017.safetensors",
562
- "model.layers.50.mlp.gate_proj.weight": "model-00014-of-00017.safetensors",
563
- "model.layers.50.mlp.up_proj.weight": "model-00013-of-00017.safetensors",
564
- "model.layers.50.post_attention_layernorm.weight": "model-00013-of-00017.safetensors",
565
  "model.layers.50.self_attn.k_proj.bias": "model-00013-of-00017.safetensors",
566
  "model.layers.50.self_attn.k_proj.weight": "model-00013-of-00017.safetensors",
567
  "model.layers.50.self_attn.o_proj.weight": "model-00013-of-00017.safetensors",
@@ -605,11 +593,11 @@
605
  "model.layers.53.self_attn.q_proj.weight": "model-00014-of-00017.safetensors",
606
  "model.layers.53.self_attn.v_proj.bias": "model-00014-of-00017.safetensors",
607
  "model.layers.53.self_attn.v_proj.weight": "model-00014-of-00017.safetensors",
608
- "model.layers.54.input_layernorm.weight": "model-00014-of-00017.safetensors",
609
  "model.layers.54.mlp.down_proj.weight": "model-00015-of-00017.safetensors",
610
- "model.layers.54.mlp.gate_proj.weight": "model-00015-of-00017.safetensors",
611
- "model.layers.54.mlp.up_proj.weight": "model-00014-of-00017.safetensors",
612
- "model.layers.54.post_attention_layernorm.weight": "model-00014-of-00017.safetensors",
613
  "model.layers.54.self_attn.k_proj.bias": "model-00014-of-00017.safetensors",
614
  "model.layers.54.self_attn.k_proj.weight": "model-00014-of-00017.safetensors",
615
  "model.layers.54.self_attn.o_proj.weight": "model-00014-of-00017.safetensors",
@@ -653,11 +641,11 @@
653
  "model.layers.57.self_attn.q_proj.weight": "model-00015-of-00017.safetensors",
654
  "model.layers.57.self_attn.v_proj.bias": "model-00015-of-00017.safetensors",
655
  "model.layers.57.self_attn.v_proj.weight": "model-00015-of-00017.safetensors",
656
- "model.layers.58.input_layernorm.weight": "model-00015-of-00017.safetensors",
657
  "model.layers.58.mlp.down_proj.weight": "model-00016-of-00017.safetensors",
658
- "model.layers.58.mlp.gate_proj.weight": "model-00016-of-00017.safetensors",
659
- "model.layers.58.mlp.up_proj.weight": "model-00015-of-00017.safetensors",
660
- "model.layers.58.post_attention_layernorm.weight": "model-00015-of-00017.safetensors",
661
  "model.layers.58.self_attn.k_proj.bias": "model-00015-of-00017.safetensors",
662
  "model.layers.58.self_attn.k_proj.weight": "model-00015-of-00017.safetensors",
663
  "model.layers.58.self_attn.o_proj.weight": "model-00015-of-00017.safetensors",
@@ -677,18 +665,18 @@
677
  "model.layers.59.self_attn.q_proj.weight": "model-00016-of-00017.safetensors",
678
  "model.layers.59.self_attn.v_proj.bias": "model-00016-of-00017.safetensors",
679
  "model.layers.59.self_attn.v_proj.weight": "model-00016-of-00017.safetensors",
680
- "model.layers.6.input_layernorm.weight": "model-00002-of-00017.safetensors",
681
- "model.layers.6.mlp.down_proj.weight": "model-00003-of-00017.safetensors",
682
- "model.layers.6.mlp.gate_proj.weight": "model-00003-of-00017.safetensors",
683
- "model.layers.6.mlp.up_proj.weight": "model-00002-of-00017.safetensors",
684
- "model.layers.6.post_attention_layernorm.weight": "model-00002-of-00017.safetensors",
685
- "model.layers.6.self_attn.k_proj.bias": "model-00002-of-00017.safetensors",
686
- "model.layers.6.self_attn.k_proj.weight": "model-00002-of-00017.safetensors",
687
- "model.layers.6.self_attn.o_proj.weight": "model-00002-of-00017.safetensors",
688
- "model.layers.6.self_attn.q_proj.bias": "model-00002-of-00017.safetensors",
689
- "model.layers.6.self_attn.q_proj.weight": "model-00002-of-00017.safetensors",
690
- "model.layers.6.self_attn.v_proj.bias": "model-00002-of-00017.safetensors",
691
- "model.layers.6.self_attn.v_proj.weight": "model-00002-of-00017.safetensors",
692
  "model.layers.60.input_layernorm.weight": "model-00016-of-00017.safetensors",
693
  "model.layers.60.mlp.down_proj.weight": "model-00016-of-00017.safetensors",
694
  "model.layers.60.mlp.gate_proj.weight": "model-00016-of-00017.safetensors",
@@ -713,11 +701,11 @@
713
  "model.layers.61.self_attn.q_proj.weight": "model-00016-of-00017.safetensors",
714
  "model.layers.61.self_attn.v_proj.bias": "model-00016-of-00017.safetensors",
715
  "model.layers.61.self_attn.v_proj.weight": "model-00016-of-00017.safetensors",
716
- "model.layers.62.input_layernorm.weight": "model-00016-of-00017.safetensors",
717
  "model.layers.62.mlp.down_proj.weight": "model-00017-of-00017.safetensors",
718
- "model.layers.62.mlp.gate_proj.weight": "model-00017-of-00017.safetensors",
719
- "model.layers.62.mlp.up_proj.weight": "model-00016-of-00017.safetensors",
720
- "model.layers.62.post_attention_layernorm.weight": "model-00016-of-00017.safetensors",
721
  "model.layers.62.self_attn.k_proj.bias": "model-00016-of-00017.safetensors",
722
  "model.layers.62.self_attn.k_proj.weight": "model-00016-of-00017.safetensors",
723
  "model.layers.62.self_attn.o_proj.weight": "model-00016-of-00017.safetensors",
@@ -737,6 +725,18 @@
737
  "model.layers.63.self_attn.q_proj.weight": "model-00017-of-00017.safetensors",
738
  "model.layers.63.self_attn.v_proj.bias": "model-00017-of-00017.safetensors",
739
  "model.layers.63.self_attn.v_proj.weight": "model-00017-of-00017.safetensors",
 
 
 
 
 
 
 
 
 
 
 
 
740
  "model.layers.7.input_layernorm.weight": "model-00003-of-00017.safetensors",
741
  "model.layers.7.mlp.down_proj.weight": "model-00003-of-00017.safetensors",
742
  "model.layers.7.mlp.gate_proj.weight": "model-00003-of-00017.safetensors",
 
17
  "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00017.safetensors",
18
  "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00017.safetensors",
19
  "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00017.safetensors",
20
+ "model.layers.10.input_layernorm.weight": "model-00004-of-00017.safetensors",
 
 
 
 
 
 
 
 
 
 
 
 
21
  "model.layers.10.mlp.down_proj.weight": "model-00004-of-00017.safetensors",
22
+ "model.layers.10.mlp.gate_proj.weight": "model-00003-of-00017.safetensors",
23
+ "model.layers.10.mlp.up_proj.weight": "model-00004-of-00017.safetensors",
24
+ "model.layers.10.post_attention_layernorm.weight": "model-00004-of-00017.safetensors",
25
  "model.layers.10.self_attn.k_proj.bias": "model-00003-of-00017.safetensors",
26
  "model.layers.10.self_attn.k_proj.weight": "model-00003-of-00017.safetensors",
27
  "model.layers.10.self_attn.o_proj.weight": "model-00003-of-00017.safetensors",
 
65
  "model.layers.13.self_attn.q_proj.weight": "model-00004-of-00017.safetensors",
66
  "model.layers.13.self_attn.v_proj.bias": "model-00004-of-00017.safetensors",
67
  "model.layers.13.self_attn.v_proj.weight": "model-00004-of-00017.safetensors",
68
+ "model.layers.14.input_layernorm.weight": "model-00005-of-00017.safetensors",
69
  "model.layers.14.mlp.down_proj.weight": "model-00005-of-00017.safetensors",
70
+ "model.layers.14.mlp.gate_proj.weight": "model-00004-of-00017.safetensors",
71
+ "model.layers.14.mlp.up_proj.weight": "model-00005-of-00017.safetensors",
72
+ "model.layers.14.post_attention_layernorm.weight": "model-00005-of-00017.safetensors",
73
  "model.layers.14.self_attn.k_proj.bias": "model-00004-of-00017.safetensors",
74
  "model.layers.14.self_attn.k_proj.weight": "model-00004-of-00017.safetensors",
75
  "model.layers.14.self_attn.o_proj.weight": "model-00004-of-00017.safetensors",
 
113
  "model.layers.17.self_attn.q_proj.weight": "model-00005-of-00017.safetensors",
114
  "model.layers.17.self_attn.v_proj.bias": "model-00005-of-00017.safetensors",
115
  "model.layers.17.self_attn.v_proj.weight": "model-00005-of-00017.safetensors",
116
+ "model.layers.18.input_layernorm.weight": "model-00006-of-00017.safetensors",
117
  "model.layers.18.mlp.down_proj.weight": "model-00006-of-00017.safetensors",
118
+ "model.layers.18.mlp.gate_proj.weight": "model-00005-of-00017.safetensors",
119
+ "model.layers.18.mlp.up_proj.weight": "model-00006-of-00017.safetensors",
120
+ "model.layers.18.post_attention_layernorm.weight": "model-00006-of-00017.safetensors",
121
  "model.layers.18.self_attn.k_proj.bias": "model-00005-of-00017.safetensors",
122
  "model.layers.18.self_attn.k_proj.weight": "model-00005-of-00017.safetensors",
123
  "model.layers.18.self_attn.o_proj.weight": "model-00005-of-00017.safetensors",
 
137
  "model.layers.19.self_attn.q_proj.weight": "model-00006-of-00017.safetensors",
138
  "model.layers.19.self_attn.v_proj.bias": "model-00006-of-00017.safetensors",
139
  "model.layers.19.self_attn.v_proj.weight": "model-00006-of-00017.safetensors",
140
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00017.safetensors",
141
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00017.safetensors",
142
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00017.safetensors",
143
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00017.safetensors",
144
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00017.safetensors",
145
+ "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00017.safetensors",
146
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00017.safetensors",
147
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00017.safetensors",
148
+ "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00017.safetensors",
149
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00017.safetensors",
150
+ "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00017.safetensors",
151
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00017.safetensors",
152
  "model.layers.20.input_layernorm.weight": "model-00006-of-00017.safetensors",
153
  "model.layers.20.mlp.down_proj.weight": "model-00006-of-00017.safetensors",
154
  "model.layers.20.mlp.gate_proj.weight": "model-00006-of-00017.safetensors",
 
173
  "model.layers.21.self_attn.q_proj.weight": "model-00006-of-00017.safetensors",
174
  "model.layers.21.self_attn.v_proj.bias": "model-00006-of-00017.safetensors",
175
  "model.layers.21.self_attn.v_proj.weight": "model-00006-of-00017.safetensors",
176
+ "model.layers.22.input_layernorm.weight": "model-00007-of-00017.safetensors",
177
  "model.layers.22.mlp.down_proj.weight": "model-00007-of-00017.safetensors",
178
+ "model.layers.22.mlp.gate_proj.weight": "model-00006-of-00017.safetensors",
179
+ "model.layers.22.mlp.up_proj.weight": "model-00007-of-00017.safetensors",
180
+ "model.layers.22.post_attention_layernorm.weight": "model-00007-of-00017.safetensors",
181
  "model.layers.22.self_attn.k_proj.bias": "model-00006-of-00017.safetensors",
182
  "model.layers.22.self_attn.k_proj.weight": "model-00006-of-00017.safetensors",
183
  "model.layers.22.self_attn.o_proj.weight": "model-00006-of-00017.safetensors",
 
221
  "model.layers.25.self_attn.q_proj.weight": "model-00007-of-00017.safetensors",
222
  "model.layers.25.self_attn.v_proj.bias": "model-00007-of-00017.safetensors",
223
  "model.layers.25.self_attn.v_proj.weight": "model-00007-of-00017.safetensors",
224
+ "model.layers.26.input_layernorm.weight": "model-00008-of-00017.safetensors",
225
  "model.layers.26.mlp.down_proj.weight": "model-00008-of-00017.safetensors",
226
+ "model.layers.26.mlp.gate_proj.weight": "model-00007-of-00017.safetensors",
227
+ "model.layers.26.mlp.up_proj.weight": "model-00008-of-00017.safetensors",
228
+ "model.layers.26.post_attention_layernorm.weight": "model-00008-of-00017.safetensors",
229
  "model.layers.26.self_attn.k_proj.bias": "model-00007-of-00017.safetensors",
230
  "model.layers.26.self_attn.k_proj.weight": "model-00007-of-00017.safetensors",
231
  "model.layers.26.self_attn.o_proj.weight": "model-00007-of-00017.safetensors",
 
269
  "model.layers.29.self_attn.q_proj.weight": "model-00008-of-00017.safetensors",
270
  "model.layers.29.self_attn.v_proj.bias": "model-00008-of-00017.safetensors",
271
  "model.layers.29.self_attn.v_proj.weight": "model-00008-of-00017.safetensors",
272
+ "model.layers.2.input_layernorm.weight": "model-00002-of-00017.safetensors",
273
+ "model.layers.2.mlp.down_proj.weight": "model-00002-of-00017.safetensors",
274
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00017.safetensors",
275
+ "model.layers.2.mlp.up_proj.weight": "model-00002-of-00017.safetensors",
276
+ "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00017.safetensors",
277
+ "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00017.safetensors",
278
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00017.safetensors",
279
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00017.safetensors",
280
+ "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00017.safetensors",
281
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00017.safetensors",
282
+ "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00017.safetensors",
283
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00017.safetensors",
284
+ "model.layers.30.input_layernorm.weight": "model-00009-of-00017.safetensors",
285
  "model.layers.30.mlp.down_proj.weight": "model-00009-of-00017.safetensors",
286
+ "model.layers.30.mlp.gate_proj.weight": "model-00008-of-00017.safetensors",
287
+ "model.layers.30.mlp.up_proj.weight": "model-00009-of-00017.safetensors",
288
+ "model.layers.30.post_attention_layernorm.weight": "model-00009-of-00017.safetensors",
289
  "model.layers.30.self_attn.k_proj.bias": "model-00008-of-00017.safetensors",
290
  "model.layers.30.self_attn.k_proj.weight": "model-00008-of-00017.safetensors",
291
  "model.layers.30.self_attn.o_proj.weight": "model-00008-of-00017.safetensors",
 
329
  "model.layers.33.self_attn.q_proj.weight": "model-00009-of-00017.safetensors",
330
  "model.layers.33.self_attn.v_proj.bias": "model-00009-of-00017.safetensors",
331
  "model.layers.33.self_attn.v_proj.weight": "model-00009-of-00017.safetensors",
332
+ "model.layers.34.input_layernorm.weight": "model-00010-of-00017.safetensors",
333
  "model.layers.34.mlp.down_proj.weight": "model-00010-of-00017.safetensors",
334
+ "model.layers.34.mlp.gate_proj.weight": "model-00009-of-00017.safetensors",
335
+ "model.layers.34.mlp.up_proj.weight": "model-00010-of-00017.safetensors",
336
+ "model.layers.34.post_attention_layernorm.weight": "model-00010-of-00017.safetensors",
337
  "model.layers.34.self_attn.k_proj.bias": "model-00009-of-00017.safetensors",
338
  "model.layers.34.self_attn.k_proj.weight": "model-00009-of-00017.safetensors",
339
  "model.layers.34.self_attn.o_proj.weight": "model-00009-of-00017.safetensors",
 
377
  "model.layers.37.self_attn.q_proj.weight": "model-00010-of-00017.safetensors",
378
  "model.layers.37.self_attn.v_proj.bias": "model-00010-of-00017.safetensors",
379
  "model.layers.37.self_attn.v_proj.weight": "model-00010-of-00017.safetensors",
380
+ "model.layers.38.input_layernorm.weight": "model-00011-of-00017.safetensors",
381
  "model.layers.38.mlp.down_proj.weight": "model-00011-of-00017.safetensors",
382
+ "model.layers.38.mlp.gate_proj.weight": "model-00010-of-00017.safetensors",
383
+ "model.layers.38.mlp.up_proj.weight": "model-00011-of-00017.safetensors",
384
+ "model.layers.38.post_attention_layernorm.weight": "model-00011-of-00017.safetensors",
385
  "model.layers.38.self_attn.k_proj.bias": "model-00010-of-00017.safetensors",
386
  "model.layers.38.self_attn.k_proj.weight": "model-00010-of-00017.safetensors",
387
  "model.layers.38.self_attn.o_proj.weight": "model-00010-of-00017.safetensors",
 
401
  "model.layers.39.self_attn.q_proj.weight": "model-00011-of-00017.safetensors",
402
  "model.layers.39.self_attn.v_proj.bias": "model-00011-of-00017.safetensors",
403
  "model.layers.39.self_attn.v_proj.weight": "model-00011-of-00017.safetensors",
404
+ "model.layers.3.input_layernorm.weight": "model-00002-of-00017.safetensors",
405
+ "model.layers.3.mlp.down_proj.weight": "model-00002-of-00017.safetensors",
406
+ "model.layers.3.mlp.gate_proj.weight": "model-00002-of-00017.safetensors",
407
+ "model.layers.3.mlp.up_proj.weight": "model-00002-of-00017.safetensors",
408
+ "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00017.safetensors",
409
+ "model.layers.3.self_attn.k_proj.bias": "model-00002-of-00017.safetensors",
410
+ "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00017.safetensors",
411
+ "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00017.safetensors",
412
+ "model.layers.3.self_attn.q_proj.bias": "model-00002-of-00017.safetensors",
413
+ "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00017.safetensors",
414
+ "model.layers.3.self_attn.v_proj.bias": "model-00002-of-00017.safetensors",
415
+ "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00017.safetensors",
416
  "model.layers.40.input_layernorm.weight": "model-00011-of-00017.safetensors",
417
  "model.layers.40.mlp.down_proj.weight": "model-00011-of-00017.safetensors",
418
  "model.layers.40.mlp.gate_proj.weight": "model-00011-of-00017.safetensors",
 
437
  "model.layers.41.self_attn.q_proj.weight": "model-00011-of-00017.safetensors",
438
  "model.layers.41.self_attn.v_proj.bias": "model-00011-of-00017.safetensors",
439
  "model.layers.41.self_attn.v_proj.weight": "model-00011-of-00017.safetensors",
440
+ "model.layers.42.input_layernorm.weight": "model-00012-of-00017.safetensors",
441
  "model.layers.42.mlp.down_proj.weight": "model-00012-of-00017.safetensors",
442
+ "model.layers.42.mlp.gate_proj.weight": "model-00011-of-00017.safetensors",
443
+ "model.layers.42.mlp.up_proj.weight": "model-00012-of-00017.safetensors",
444
+ "model.layers.42.post_attention_layernorm.weight": "model-00012-of-00017.safetensors",
445
  "model.layers.42.self_attn.k_proj.bias": "model-00011-of-00017.safetensors",
446
  "model.layers.42.self_attn.k_proj.weight": "model-00011-of-00017.safetensors",
447
  "model.layers.42.self_attn.o_proj.weight": "model-00011-of-00017.safetensors",
 
485
  "model.layers.45.self_attn.q_proj.weight": "model-00012-of-00017.safetensors",
486
  "model.layers.45.self_attn.v_proj.bias": "model-00012-of-00017.safetensors",
487
  "model.layers.45.self_attn.v_proj.weight": "model-00012-of-00017.safetensors",
488
+ "model.layers.46.input_layernorm.weight": "model-00013-of-00017.safetensors",
489
  "model.layers.46.mlp.down_proj.weight": "model-00013-of-00017.safetensors",
490
+ "model.layers.46.mlp.gate_proj.weight": "model-00012-of-00017.safetensors",
491
+ "model.layers.46.mlp.up_proj.weight": "model-00013-of-00017.safetensors",
492
+ "model.layers.46.post_attention_layernorm.weight": "model-00013-of-00017.safetensors",
493
  "model.layers.46.self_attn.k_proj.bias": "model-00012-of-00017.safetensors",
494
  "model.layers.46.self_attn.k_proj.weight": "model-00012-of-00017.safetensors",
495
  "model.layers.46.self_attn.o_proj.weight": "model-00012-of-00017.safetensors",
 
533
  "model.layers.49.self_attn.q_proj.weight": "model-00013-of-00017.safetensors",
534
  "model.layers.49.self_attn.v_proj.bias": "model-00013-of-00017.safetensors",
535
  "model.layers.49.self_attn.v_proj.weight": "model-00013-of-00017.safetensors",
536
+ "model.layers.4.input_layernorm.weight": "model-00002-of-00017.safetensors",
537
+ "model.layers.4.mlp.down_proj.weight": "model-00002-of-00017.safetensors",
538
+ "model.layers.4.mlp.gate_proj.weight": "model-00002-of-00017.safetensors",
539
+ "model.layers.4.mlp.up_proj.weight": "model-00002-of-00017.safetensors",
540
+ "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00017.safetensors",
541
+ "model.layers.4.self_attn.k_proj.bias": "model-00002-of-00017.safetensors",
542
+ "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00017.safetensors",
543
+ "model.layers.4.self_attn.o_proj.weight": "model-00002-of-00017.safetensors",
544
+ "model.layers.4.self_attn.q_proj.bias": "model-00002-of-00017.safetensors",
545
+ "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00017.safetensors",
546
+ "model.layers.4.self_attn.v_proj.bias": "model-00002-of-00017.safetensors",
547
+ "model.layers.4.self_attn.v_proj.weight": "model-00002-of-00017.safetensors",
548
+ "model.layers.50.input_layernorm.weight": "model-00014-of-00017.safetensors",
549
  "model.layers.50.mlp.down_proj.weight": "model-00014-of-00017.safetensors",
550
+ "model.layers.50.mlp.gate_proj.weight": "model-00013-of-00017.safetensors",
551
+ "model.layers.50.mlp.up_proj.weight": "model-00014-of-00017.safetensors",
552
+ "model.layers.50.post_attention_layernorm.weight": "model-00014-of-00017.safetensors",
553
  "model.layers.50.self_attn.k_proj.bias": "model-00013-of-00017.safetensors",
554
  "model.layers.50.self_attn.k_proj.weight": "model-00013-of-00017.safetensors",
555
  "model.layers.50.self_attn.o_proj.weight": "model-00013-of-00017.safetensors",
 
593
  "model.layers.53.self_attn.q_proj.weight": "model-00014-of-00017.safetensors",
594
  "model.layers.53.self_attn.v_proj.bias": "model-00014-of-00017.safetensors",
595
  "model.layers.53.self_attn.v_proj.weight": "model-00014-of-00017.safetensors",
596
+ "model.layers.54.input_layernorm.weight": "model-00015-of-00017.safetensors",
597
  "model.layers.54.mlp.down_proj.weight": "model-00015-of-00017.safetensors",
598
+ "model.layers.54.mlp.gate_proj.weight": "model-00014-of-00017.safetensors",
599
+ "model.layers.54.mlp.up_proj.weight": "model-00015-of-00017.safetensors",
600
+ "model.layers.54.post_attention_layernorm.weight": "model-00015-of-00017.safetensors",
601
  "model.layers.54.self_attn.k_proj.bias": "model-00014-of-00017.safetensors",
602
  "model.layers.54.self_attn.k_proj.weight": "model-00014-of-00017.safetensors",
603
  "model.layers.54.self_attn.o_proj.weight": "model-00014-of-00017.safetensors",
 
641
  "model.layers.57.self_attn.q_proj.weight": "model-00015-of-00017.safetensors",
642
  "model.layers.57.self_attn.v_proj.bias": "model-00015-of-00017.safetensors",
643
  "model.layers.57.self_attn.v_proj.weight": "model-00015-of-00017.safetensors",
644
+ "model.layers.58.input_layernorm.weight": "model-00016-of-00017.safetensors",
645
  "model.layers.58.mlp.down_proj.weight": "model-00016-of-00017.safetensors",
646
+ "model.layers.58.mlp.gate_proj.weight": "model-00015-of-00017.safetensors",
647
+ "model.layers.58.mlp.up_proj.weight": "model-00016-of-00017.safetensors",
648
+ "model.layers.58.post_attention_layernorm.weight": "model-00016-of-00017.safetensors",
649
  "model.layers.58.self_attn.k_proj.bias": "model-00015-of-00017.safetensors",
650
  "model.layers.58.self_attn.k_proj.weight": "model-00015-of-00017.safetensors",
651
  "model.layers.58.self_attn.o_proj.weight": "model-00015-of-00017.safetensors",
 
665
  "model.layers.59.self_attn.q_proj.weight": "model-00016-of-00017.safetensors",
666
  "model.layers.59.self_attn.v_proj.bias": "model-00016-of-00017.safetensors",
667
  "model.layers.59.self_attn.v_proj.weight": "model-00016-of-00017.safetensors",
668
+ "model.layers.5.input_layernorm.weight": "model-00002-of-00017.safetensors",
669
+ "model.layers.5.mlp.down_proj.weight": "model-00002-of-00017.safetensors",
670
+ "model.layers.5.mlp.gate_proj.weight": "model-00002-of-00017.safetensors",
671
+ "model.layers.5.mlp.up_proj.weight": "model-00002-of-00017.safetensors",
672
+ "model.layers.5.post_attention_layernorm.weight": "model-00002-of-00017.safetensors",
673
+ "model.layers.5.self_attn.k_proj.bias": "model-00002-of-00017.safetensors",
674
+ "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00017.safetensors",
675
+ "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00017.safetensors",
676
+ "model.layers.5.self_attn.q_proj.bias": "model-00002-of-00017.safetensors",
677
+ "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00017.safetensors",
678
+ "model.layers.5.self_attn.v_proj.bias": "model-00002-of-00017.safetensors",
679
+ "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00017.safetensors",
680
  "model.layers.60.input_layernorm.weight": "model-00016-of-00017.safetensors",
681
  "model.layers.60.mlp.down_proj.weight": "model-00016-of-00017.safetensors",
682
  "model.layers.60.mlp.gate_proj.weight": "model-00016-of-00017.safetensors",
 
701
  "model.layers.61.self_attn.q_proj.weight": "model-00016-of-00017.safetensors",
702
  "model.layers.61.self_attn.v_proj.bias": "model-00016-of-00017.safetensors",
703
  "model.layers.61.self_attn.v_proj.weight": "model-00016-of-00017.safetensors",
704
+ "model.layers.62.input_layernorm.weight": "model-00017-of-00017.safetensors",
705
  "model.layers.62.mlp.down_proj.weight": "model-00017-of-00017.safetensors",
706
+ "model.layers.62.mlp.gate_proj.weight": "model-00016-of-00017.safetensors",
707
+ "model.layers.62.mlp.up_proj.weight": "model-00017-of-00017.safetensors",
708
+ "model.layers.62.post_attention_layernorm.weight": "model-00017-of-00017.safetensors",
709
  "model.layers.62.self_attn.k_proj.bias": "model-00016-of-00017.safetensors",
710
  "model.layers.62.self_attn.k_proj.weight": "model-00016-of-00017.safetensors",
711
  "model.layers.62.self_attn.o_proj.weight": "model-00016-of-00017.safetensors",
 
725
  "model.layers.63.self_attn.q_proj.weight": "model-00017-of-00017.safetensors",
726
  "model.layers.63.self_attn.v_proj.bias": "model-00017-of-00017.safetensors",
727
  "model.layers.63.self_attn.v_proj.weight": "model-00017-of-00017.safetensors",
728
+ "model.layers.6.input_layernorm.weight": "model-00003-of-00017.safetensors",
729
+ "model.layers.6.mlp.down_proj.weight": "model-00003-of-00017.safetensors",
730
+ "model.layers.6.mlp.gate_proj.weight": "model-00002-of-00017.safetensors",
731
+ "model.layers.6.mlp.up_proj.weight": "model-00003-of-00017.safetensors",
732
+ "model.layers.6.post_attention_layernorm.weight": "model-00003-of-00017.safetensors",
733
+ "model.layers.6.self_attn.k_proj.bias": "model-00002-of-00017.safetensors",
734
+ "model.layers.6.self_attn.k_proj.weight": "model-00002-of-00017.safetensors",
735
+ "model.layers.6.self_attn.o_proj.weight": "model-00002-of-00017.safetensors",
736
+ "model.layers.6.self_attn.q_proj.bias": "model-00002-of-00017.safetensors",
737
+ "model.layers.6.self_attn.q_proj.weight": "model-00002-of-00017.safetensors",
738
+ "model.layers.6.self_attn.v_proj.bias": "model-00002-of-00017.safetensors",
739
+ "model.layers.6.self_attn.v_proj.weight": "model-00002-of-00017.safetensors",
740
  "model.layers.7.input_layernorm.weight": "model-00003-of-00017.safetensors",
741
  "model.layers.7.mlp.down_proj.weight": "model-00003-of-00017.safetensors",
742
  "model.layers.7.mlp.gate_proj.weight": "model-00003-of-00017.safetensors",