danielhanchen commited on
Commit
df815e3
1 Parent(s): f687702

Upload Qwen2ForCausalLM

Browse files
config.json CHANGED
@@ -31,70 +31,13 @@
31
  "multi_modal_projector",
32
  "merger",
33
  "modality_projection",
34
- "model.layers.0.self_attn",
35
- "model.layers.1.self_attn",
36
- "model.layers.2.self_attn",
37
- "model.layers.3.self_attn",
38
- "model.layers.4.self_attn",
39
- "model.layers.5.self_attn",
40
- "model.layers.6.self_attn",
41
- "model.layers.7.self_attn",
42
- "model.layers.8.self_attn",
43
- "model.layers.9.self_attn",
44
- "model.layers.10.self_attn",
45
- "model.layers.11.self_attn",
46
- "model.layers.12.self_attn",
47
- "model.layers.13.self_attn",
48
- "model.layers.14.self_attn",
49
- "model.layers.15.self_attn",
50
- "model.layers.16.self_attn",
51
- "model.layers.17.self_attn",
52
- "model.layers.18.self_attn",
53
- "model.layers.19.self_attn",
54
- "model.layers.20.self_attn",
55
- "model.layers.21.self_attn",
56
- "model.layers.22.self_attn",
57
- "model.layers.23.self_attn",
58
- "model.layers.24.self_attn",
59
- "model.layers.25.self_attn",
60
- "model.layers.26.self_attn",
61
- "model.layers.27.self_attn",
62
- "model.layers.28.self_attn",
63
- "model.layers.29.self_attn",
64
- "model.layers.30.self_attn",
65
- "model.layers.31.self_attn",
66
- "model.layers.32.self_attn",
67
- "model.layers.33.self_attn",
68
- "model.layers.34.self_attn",
69
- "model.layers.35.self_attn",
70
- "model.layers.36.self_attn",
71
- "model.layers.37.self_attn",
72
- "model.layers.38.self_attn",
73
- "model.layers.39.self_attn",
74
- "model.layers.40.self_attn",
75
- "model.layers.41.self_attn",
76
- "model.layers.42.self_attn",
77
- "model.layers.43.self_attn",
78
- "model.layers.44.self_attn",
79
- "model.layers.45.self_attn",
80
- "model.layers.46.self_attn",
81
- "model.layers.47.self_attn",
82
- "model.layers.48.self_attn",
83
- "model.layers.49.self_attn",
84
- "model.layers.50.self_attn",
85
- "model.layers.51.self_attn",
86
- "model.layers.52.self_attn",
87
- "model.layers.53.self_attn",
88
- "model.layers.54.self_attn",
89
- "model.layers.55.self_attn",
90
- "model.layers.56.self_attn",
91
- "model.layers.57.self_attn",
92
- "model.layers.58.self_attn",
93
- "model.layers.59.self_attn",
94
- "model.layers.60.self_attn",
95
- "model.layers.61.self_attn",
96
- "model.layers.62.self_attn",
97
- "model.layers.63.self_attn"
98
  ],
99
  "llm_int8_threshold": 6.0,
100
  "load_in_4bit": true,
 
31
  "multi_modal_projector",
32
  "merger",
33
  "modality_projection",
34
+ "model.layers.5.mlp",
35
+ "model.layers.33.mlp",
36
+ "model.layers.35.mlp",
37
+ "model.layers.38.mlp",
38
+ "model.layers.60.mlp",
39
+ "model.layers.62.mlp",
40
+ "model.layers.43.self_attn.o_proj"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
41
  ],
42
  "llm_int8_threshold": 6.0,
43
  "load_in_4bit": true,
model-00001-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:017df76fec88fdb52b1f74c25b4deb5ace0858ef7dac2c1a6b5f55f62352468a
3
+ size 4987266814
model-00002-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec6022f551d0b35b8e9e720e5bd767417a93b3831998e1d4bb65ef0aecb51d49
3
+ size 4999136434
model-00003-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e469da8418e53b091612f299c74b2f86ebeaa46c1bc7fb2ed595bd9d5315af19
3
+ size 4981146486
model-00004-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5752710c7e9e198e4c0d4933a778aaa78a042ff7f01f268ba4552ec50dc7ff2a
3
+ size 4843103842
model-00005-of-00005.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0499f7b6052622856e466db518fead9d0b59679421e311d755a5e1999442da67
3
+ size 3225289560
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff