bowenbaoamd commited on
Commit
f9c4890
·
verified ·
1 Parent(s): 012e53c

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,34 +1,10 @@
1
  {
 
2
  "architectures": [
3
- "MllamaForCausalLM"
4
  ],
5
- "bos_token_id": 128000,
6
- "cross_attention_layers": [
7
- 3,
8
- 8,
9
- 13,
10
- 18,
11
- 23,
12
- 28,
13
- 33,
14
- 38
15
- ],
16
- "dropout": 0,
17
- "eos_token_id": [
18
- 128001,
19
- 128008,
20
- 128009
21
- ],
22
- "hidden_act": "silu",
23
- "hidden_size": 4096,
24
- "initializer_range": 0.02,
25
- "intermediate_size": 14336,
26
- "max_position_embeddings": 131072,
27
- "model_type": "mllama_text_model",
28
- "num_attention_heads": 32,
29
- "num_hidden_layers": 40,
30
- "num_key_value_heads": 8,
31
- "pad_token_id": 128004,
32
  "quantization_config": {
33
  "activation_scheme": "static",
34
  "ignored_layers": [
@@ -37,18 +13,220 @@
37
  "kv_cache_scheme": "static",
38
  "quant_method": "fp8"
39
  },
40
- "rms_norm_eps": 1e-05,
41
- "rope_scaling": {
42
- "factor": 8.0,
43
- "high_freq_factor": 4.0,
44
- "low_freq_factor": 1.0,
45
- "original_max_position_embeddings": 8192,
46
- "rope_type": "llama3"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  },
48
- "rope_theta": 500000.0,
49
- "tie_word_embeddings": false,
50
  "torch_dtype": "bfloat16",
51
- "transformers_version": "4.45.0",
52
- "use_cache": true,
53
- "vocab_size": 128256
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
54
  }
 
1
  {
2
+ "_name_or_path": "/model_path/Llama-3.2-11B-Vision-Instruct/",
3
  "architectures": [
4
+ "MllamaForConditionalGeneration"
5
  ],
6
+ "image_token_index": 128256,
7
+ "model_type": "mllama",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "quantization_config": {
9
  "activation_scheme": "static",
10
  "ignored_layers": [
 
13
  "kv_cache_scheme": "static",
14
  "quant_method": "fp8"
15
  },
16
+ "text_config": {
17
+ "_name_or_path": "",
18
+ "add_cross_attention": false,
19
+ "architectures": null,
20
+ "bad_words_ids": null,
21
+ "begin_suppress_tokens": null,
22
+ "bos_token_id": 128000,
23
+ "chunk_size_feed_forward": 0,
24
+ "cross_attention_hidden_size": null,
25
+ "cross_attention_layers": [
26
+ 3,
27
+ 8,
28
+ 13,
29
+ 18,
30
+ 23,
31
+ 28,
32
+ 33,
33
+ 38
34
+ ],
35
+ "decoder_start_token_id": null,
36
+ "diversity_penalty": 0.0,
37
+ "do_sample": false,
38
+ "dropout": 0,
39
+ "early_stopping": false,
40
+ "encoder_no_repeat_ngram_size": 0,
41
+ "eos_token_id": [
42
+ 128001,
43
+ 128008,
44
+ 128009
45
+ ],
46
+ "exponential_decay_length_penalty": null,
47
+ "finetuning_task": null,
48
+ "forced_bos_token_id": null,
49
+ "forced_eos_token_id": null,
50
+ "hidden_act": "silu",
51
+ "hidden_size": 4096,
52
+ "id2label": {
53
+ "0": "LABEL_0",
54
+ "1": "LABEL_1"
55
+ },
56
+ "initializer_range": 0.02,
57
+ "intermediate_size": 14336,
58
+ "is_decoder": false,
59
+ "is_encoder_decoder": false,
60
+ "label2id": {
61
+ "LABEL_0": 0,
62
+ "LABEL_1": 1
63
+ },
64
+ "length_penalty": 1.0,
65
+ "max_length": 20,
66
+ "max_position_embeddings": 131072,
67
+ "min_length": 0,
68
+ "model_type": "mllama_text_model",
69
+ "no_repeat_ngram_size": 0,
70
+ "num_attention_heads": 32,
71
+ "num_beam_groups": 1,
72
+ "num_beams": 1,
73
+ "num_hidden_layers": 40,
74
+ "num_key_value_heads": 8,
75
+ "num_return_sequences": 1,
76
+ "output_attentions": false,
77
+ "output_hidden_states": false,
78
+ "output_scores": false,
79
+ "pad_token_id": 128004,
80
+ "prefix": null,
81
+ "problem_type": null,
82
+ "pruned_heads": {},
83
+ "remove_invalid_values": false,
84
+ "repetition_penalty": 1.0,
85
+ "return_dict": true,
86
+ "return_dict_in_generate": false,
87
+ "rms_norm_eps": 1e-05,
88
+ "rope_scaling": {
89
+ "factor": 8.0,
90
+ "high_freq_factor": 4.0,
91
+ "low_freq_factor": 1.0,
92
+ "original_max_position_embeddings": 8192,
93
+ "rope_type": "llama3"
94
+ },
95
+ "rope_theta": 500000.0,
96
+ "sep_token_id": null,
97
+ "suppress_tokens": null,
98
+ "task_specific_params": null,
99
+ "temperature": 1.0,
100
+ "tf_legacy_loss": false,
101
+ "tie_encoder_decoder": false,
102
+ "tie_word_embeddings": false,
103
+ "tokenizer_class": null,
104
+ "top_k": 50,
105
+ "top_p": 1.0,
106
+ "torch_dtype": "bfloat16",
107
+ "torchscript": false,
108
+ "typical_p": 1.0,
109
+ "use_bfloat16": false,
110
+ "use_cache": true,
111
+ "vocab_size": 128256
112
  },
 
 
113
  "torch_dtype": "bfloat16",
114
+ "transformers_version": "4.45.2",
115
+ "vision_config": {
116
+ "_name_or_path": "",
117
+ "add_cross_attention": false,
118
+ "architectures": null,
119
+ "attention_heads": 16,
120
+ "bad_words_ids": null,
121
+ "begin_suppress_tokens": null,
122
+ "bos_token_id": null,
123
+ "chunk_size_feed_forward": 0,
124
+ "cross_attention_hidden_size": null,
125
+ "decoder_start_token_id": null,
126
+ "diversity_penalty": 0.0,
127
+ "do_sample": false,
128
+ "early_stopping": false,
129
+ "encoder_no_repeat_ngram_size": 0,
130
+ "eos_token_id": null,
131
+ "exponential_decay_length_penalty": null,
132
+ "finetuning_task": null,
133
+ "forced_bos_token_id": null,
134
+ "forced_eos_token_id": null,
135
+ "hidden_act": "gelu",
136
+ "hidden_size": 1280,
137
+ "id2label": {
138
+ "0": "LABEL_0",
139
+ "1": "LABEL_1"
140
+ },
141
+ "image_size": 560,
142
+ "initializer_range": 0.02,
143
+ "intermediate_layers_indices": [
144
+ 3,
145
+ 7,
146
+ 15,
147
+ 23,
148
+ 30
149
+ ],
150
+ "intermediate_size": 5120,
151
+ "is_decoder": false,
152
+ "is_encoder_decoder": false,
153
+ "label2id": {
154
+ "LABEL_0": 0,
155
+ "LABEL_1": 1
156
+ },
157
+ "length_penalty": 1.0,
158
+ "max_length": 20,
159
+ "max_num_tiles": 4,
160
+ "min_length": 0,
161
+ "model_type": "mllama_vision_model",
162
+ "no_repeat_ngram_size": 0,
163
+ "norm_eps": 1e-05,
164
+ "num_beam_groups": 1,
165
+ "num_beams": 1,
166
+ "num_channels": 3,
167
+ "num_global_layers": 8,
168
+ "num_hidden_layers": 32,
169
+ "num_return_sequences": 1,
170
+ "output_attentions": false,
171
+ "output_hidden_states": false,
172
+ "output_scores": false,
173
+ "pad_token_id": null,
174
+ "patch_size": 14,
175
+ "prefix": null,
176
+ "problem_type": null,
177
+ "pruned_heads": {},
178
+ "remove_invalid_values": false,
179
+ "repetition_penalty": 1.0,
180
+ "return_dict": true,
181
+ "return_dict_in_generate": false,
182
+ "sep_token_id": null,
183
+ "supported_aspect_ratios": [
184
+ [
185
+ 1,
186
+ 1
187
+ ],
188
+ [
189
+ 1,
190
+ 2
191
+ ],
192
+ [
193
+ 1,
194
+ 3
195
+ ],
196
+ [
197
+ 1,
198
+ 4
199
+ ],
200
+ [
201
+ 2,
202
+ 1
203
+ ],
204
+ [
205
+ 2,
206
+ 2
207
+ ],
208
+ [
209
+ 3,
210
+ 1
211
+ ],
212
+ [
213
+ 4,
214
+ 1
215
+ ]
216
+ ],
217
+ "suppress_tokens": null,
218
+ "task_specific_params": null,
219
+ "temperature": 1.0,
220
+ "tf_legacy_loss": false,
221
+ "tie_encoder_decoder": false,
222
+ "tie_word_embeddings": true,
223
+ "tokenizer_class": null,
224
+ "top_k": 50,
225
+ "top_p": 1.0,
226
+ "torch_dtype": "bfloat16",
227
+ "torchscript": false,
228
+ "typical_p": 1.0,
229
+ "use_bfloat16": false,
230
+ "vision_output_dim": 7680
231
+ }
232
  }
generation_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "bos_token_id": 128000,
3
  "do_sample": true,
4
  "eos_token_id": [
@@ -9,5 +10,5 @@
9
  "pad_token_id": 128004,
10
  "temperature": 0.6,
11
  "top_p": 0.9,
12
- "transformers_version": "4.45.0"
13
  }
 
1
  {
2
+ "attn_implementation": "eager",
3
  "bos_token_id": 128000,
4
  "do_sample": true,
5
  "eos_token_id": [
 
10
  "pad_token_id": 128004,
11
  "temperature": 0.6,
12
  "top_p": 0.9,
13
+ "transformers_version": "4.45.2"
14
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:29d291f289fbbe62614042c9851490c4d8a28973beeedc460575de3f9f8309cb
3
- size 4997924140
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7913886bb1471568324e41653aad932101a8bc86d76b56ff71dade79b0f9afd2
3
+ size 4984341162
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3126f4701c2e4186f8f741150a7e424507bc30970c21e20d8ab3a4e637a12731
3
- size 4777741476
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71f3d10886ee124b629c0813a3d01b675650fe49d085fdce7a1cd09bae34292b
3
+ size 4991669352
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28701ba9489796a953b3732e4a27e6d655fdf08f3d8dc98e44c3032a555ca464
3
- size 1050673280
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d70e8659c127002e674b5c0ab5f56bbfa6b6bd49ed8cade7fd9169548bdbb9bc
3
+ size 2640468236
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff
 
preprocessor_config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": true,
3
+ "do_normalize": true,
4
+ "do_pad": true,
5
+ "do_rescale": true,
6
+ "do_resize": true,
7
+ "image_mean": [
8
+ 0.48145466,
9
+ 0.4578275,
10
+ 0.40821073
11
+ ],
12
+ "image_processor_type": "MllamaImageProcessor",
13
+ "image_std": [
14
+ 0.26862954,
15
+ 0.26130258,
16
+ 0.27577711
17
+ ],
18
+ "max_image_tiles": 4,
19
+ "resample": 2,
20
+ "rescale_factor": 0.00392156862745098,
21
+ "size": {
22
+ "height": 560,
23
+ "width": 560
24
+ }
25
+ }
special_tokens_map.json CHANGED
@@ -13,11 +13,5 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": {
17
- "content": "<|finetune_right_pad_id|>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- }
23
  }
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": "<|eot_id|>"
 
 
 
 
 
 
17
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:47be6519609d58a5f29b3497045b8a2798d0d0978955ea90a893ad80e2ecdd4d
3
- size 17208880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2950f10d620c6db4032082fe810ee88b29d8cda2caabfbcbf30f40eec988741c
3
+ size 17210350
tokenizer_config.json CHANGED
@@ -2065,7 +2065,8 @@
2065
  "input_ids",
2066
  "attention_mask"
2067
  ],
2068
- "model_max_length": 131072,
2069
- "pad_token": "<|finetune_right_pad_id|>",
 
2070
  "tokenizer_class": "PreTrainedTokenizerFast"
2071
  }
 
2065
  "input_ids",
2066
  "attention_mask"
2067
  ],
2068
+ "model_max_length": 512,
2069
+ "pad_token": "<|eot_id|>",
2070
+ "padding_side": "left",
2071
  "tokenizer_class": "PreTrainedTokenizerFast"
2072
  }