patrickvonplaten commited on
Commit
04f177d
·
1 Parent(s): ebb811d

correct safetensors

Browse files
feature_extractor/preprocessor_config.json CHANGED
@@ -1,8 +1,12 @@
1
  {
2
- "crop_size": 224,
 
 
 
3
  "do_center_crop": true,
4
  "do_convert_rgb": true,
5
  "do_normalize": true,
 
6
  "do_resize": true,
7
  "feature_extractor_type": "CLIPFeatureExtractor",
8
  "image_mean": [
@@ -10,11 +14,15 @@
10
  0.4578275,
11
  0.40821073
12
  ],
 
13
  "image_std": [
14
  0.26862954,
15
  0.26130258,
16
  0.27577711
17
  ],
18
  "resample": 3,
19
- "size": 224
 
 
 
20
  }
 
1
  {
2
+ "crop_size": {
3
+ "height": 224,
4
+ "width": 224
5
+ },
6
  "do_center_crop": true,
7
  "do_convert_rgb": true,
8
  "do_normalize": true,
9
+ "do_rescale": true,
10
  "do_resize": true,
11
  "feature_extractor_type": "CLIPFeatureExtractor",
12
  "image_mean": [
 
14
  0.4578275,
15
  0.40821073
16
  ],
17
+ "image_processor_type": "CLIPImageProcessor",
18
  "image_std": [
19
  0.26862954,
20
  0.26130258,
21
  0.27577711
22
  ],
23
  "resample": 3,
24
+ "rescale_factor": 0.00392156862745098,
25
+ "size": {
26
+ "shortest_edge": 224
27
+ }
28
  }
model_index.json CHANGED
@@ -1,10 +1,11 @@
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
- "_diffusers_version": "0.2.2",
4
  "feature_extractor": [
5
  "transformers",
6
- "CLIPFeatureExtractor"
7
  ],
 
8
  "safety_checker": [
9
  "stable_diffusion",
10
  "StableDiffusionSafetyChecker"
 
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.12.0.dev0",
4
  "feature_extractor": [
5
  "transformers",
6
+ "CLIPImageProcessor"
7
  ],
8
+ "requires_safety_checker": true,
9
  "safety_checker": [
10
  "stable_diffusion",
11
  "StableDiffusionSafetyChecker"
safety_checker/config.json CHANGED
@@ -1,5 +1,6 @@
1
  {
2
- "_name_or_path": "./safety_module",
 
3
  "architectures": [
4
  "StableDiffusionSafetyChecker"
5
  ],
@@ -13,6 +14,7 @@
13
  "architectures": null,
14
  "attention_dropout": 0.0,
15
  "bad_words_ids": null,
 
16
  "bos_token_id": 0,
17
  "chunk_size_feed_forward": 0,
18
  "cross_attention_hidden_size": null,
@@ -60,14 +62,17 @@
60
  "pad_token_id": 1,
61
  "prefix": null,
62
  "problem_type": null,
 
63
  "pruned_heads": {},
64
  "remove_invalid_values": false,
65
  "repetition_penalty": 1.0,
66
  "return_dict": true,
67
  "return_dict_in_generate": false,
68
  "sep_token_id": null,
 
69
  "task_specific_params": null,
70
  "temperature": 1.0,
 
71
  "tie_encoder_decoder": false,
72
  "tie_word_embeddings": true,
73
  "tokenizer_class": null,
@@ -75,7 +80,7 @@
75
  "top_p": 1.0,
76
  "torch_dtype": null,
77
  "torchscript": false,
78
- "transformers_version": "4.21.0.dev0",
79
  "typical_p": 1.0,
80
  "use_bfloat16": false,
81
  "vocab_size": 49408
@@ -86,7 +91,7 @@
86
  "num_attention_heads": 12,
87
  "num_hidden_layers": 12
88
  },
89
- "torch_dtype": "float32",
90
  "transformers_version": null,
91
  "vision_config": {
92
  "_name_or_path": "",
@@ -94,6 +99,7 @@
94
  "architectures": null,
95
  "attention_dropout": 0.0,
96
  "bad_words_ids": null,
 
97
  "bos_token_id": null,
98
  "chunk_size_feed_forward": 0,
99
  "cross_attention_hidden_size": null,
@@ -133,6 +139,7 @@
133
  "num_attention_heads": 16,
134
  "num_beam_groups": 1,
135
  "num_beams": 1,
 
136
  "num_hidden_layers": 24,
137
  "num_return_sequences": 1,
138
  "output_attentions": false,
@@ -142,14 +149,17 @@
142
  "patch_size": 14,
143
  "prefix": null,
144
  "problem_type": null,
 
145
  "pruned_heads": {},
146
  "remove_invalid_values": false,
147
  "repetition_penalty": 1.0,
148
  "return_dict": true,
149
  "return_dict_in_generate": false,
150
  "sep_token_id": null,
 
151
  "task_specific_params": null,
152
  "temperature": 1.0,
 
153
  "tie_encoder_decoder": false,
154
  "tie_word_embeddings": true,
155
  "tokenizer_class": null,
@@ -157,7 +167,7 @@
157
  "top_p": 1.0,
158
  "torch_dtype": null,
159
  "torchscript": false,
160
- "transformers_version": "4.21.0.dev0",
161
  "typical_p": 1.0,
162
  "use_bfloat16": false
163
  },
 
1
  {
2
+ "_commit_hash": null,
3
+ "_name_or_path": "stable-diffusion-v1-4/safety_checker",
4
  "architectures": [
5
  "StableDiffusionSafetyChecker"
6
  ],
 
14
  "architectures": null,
15
  "attention_dropout": 0.0,
16
  "bad_words_ids": null,
17
+ "begin_suppress_tokens": null,
18
  "bos_token_id": 0,
19
  "chunk_size_feed_forward": 0,
20
  "cross_attention_hidden_size": null,
 
62
  "pad_token_id": 1,
63
  "prefix": null,
64
  "problem_type": null,
65
+ "projection_dim": 512,
66
  "pruned_heads": {},
67
  "remove_invalid_values": false,
68
  "repetition_penalty": 1.0,
69
  "return_dict": true,
70
  "return_dict_in_generate": false,
71
  "sep_token_id": null,
72
+ "suppress_tokens": null,
73
  "task_specific_params": null,
74
  "temperature": 1.0,
75
+ "tf_legacy_loss": false,
76
  "tie_encoder_decoder": false,
77
  "tie_word_embeddings": true,
78
  "tokenizer_class": null,
 
80
  "top_p": 1.0,
81
  "torch_dtype": null,
82
  "torchscript": false,
83
+ "transformers_version": "4.25.1",
84
  "typical_p": 1.0,
85
  "use_bfloat16": false,
86
  "vocab_size": 49408
 
91
  "num_attention_heads": 12,
92
  "num_hidden_layers": 12
93
  },
94
+ "torch_dtype": "float16",
95
  "transformers_version": null,
96
  "vision_config": {
97
  "_name_or_path": "",
 
99
  "architectures": null,
100
  "attention_dropout": 0.0,
101
  "bad_words_ids": null,
102
+ "begin_suppress_tokens": null,
103
  "bos_token_id": null,
104
  "chunk_size_feed_forward": 0,
105
  "cross_attention_hidden_size": null,
 
139
  "num_attention_heads": 16,
140
  "num_beam_groups": 1,
141
  "num_beams": 1,
142
+ "num_channels": 3,
143
  "num_hidden_layers": 24,
144
  "num_return_sequences": 1,
145
  "output_attentions": false,
 
149
  "patch_size": 14,
150
  "prefix": null,
151
  "problem_type": null,
152
+ "projection_dim": 512,
153
  "pruned_heads": {},
154
  "remove_invalid_values": false,
155
  "repetition_penalty": 1.0,
156
  "return_dict": true,
157
  "return_dict_in_generate": false,
158
  "sep_token_id": null,
159
+ "suppress_tokens": null,
160
  "task_specific_params": null,
161
  "temperature": 1.0,
162
+ "tf_legacy_loss": false,
163
  "tie_encoder_decoder": false,
164
  "tie_word_embeddings": true,
165
  "tokenizer_class": null,
 
167
  "top_p": 1.0,
168
  "torch_dtype": null,
169
  "torchscript": false,
170
+ "transformers_version": "4.25.1",
171
  "typical_p": 1.0,
172
  "use_bfloat16": false
173
  },
safety_checker/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4666d0f9b718a6ed165ce95b8aac0d3d78031b8906fdc88ca8e735af5261788c
3
- size 1215981833
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69c4f8737c7620a738704911a1c7be4dce7bad37bead09e9e499da8810ffb26d
3
+ size 608018446
scheduler/scheduler_config.json CHANGED
@@ -1,13 +1,14 @@
1
  {
2
  "_class_name": "PNDMScheduler",
3
- "_diffusers_version": "0.7.0.dev0",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
 
7
  "num_train_timesteps": 1000,
 
8
  "set_alpha_to_one": false,
9
  "skip_prk_steps": true,
10
  "steps_offset": 1,
11
- "trained_betas": null,
12
- "clip_sample": false
13
  }
 
1
  {
2
  "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.12.0.dev0",
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
7
+ "clip_sample": false,
8
  "num_train_timesteps": 1000,
9
+ "prediction_type": "epsilon",
10
  "set_alpha_to_one": false,
11
  "skip_prk_steps": true,
12
  "steps_offset": 1,
13
+ "trained_betas": null
 
14
  }
text_encoder/config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "openai/clip-vit-large-patch14",
3
  "architectures": [
4
  "CLIPTextModel"
5
  ],
@@ -18,7 +18,8 @@
18
  "num_attention_heads": 12,
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
- "torch_dtype": "float32",
22
- "transformers_version": "4.21.0.dev0",
 
23
  "vocab_size": 49408
24
  }
 
1
  {
2
+ "_name_or_path": "stable-diffusion-v1-4/text_encoder",
3
  "architectures": [
4
  "CLIPTextModel"
5
  ],
 
18
  "num_attention_heads": 12,
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
+ "projection_dim": 512,
22
+ "torch_dtype": "float16",
23
+ "transformers_version": "4.25.1",
24
  "vocab_size": 49408
25
  }
text_encoder/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7b3a12df205cb3c74dd4eae4354d93f606ae6b3bc29d5d06fd97921cb9ad8a81
3
- size 492265879
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc83cf401d930147807e7c44021c164bcc5508c9d4cc0ff35f4e354685ca9cd0
3
+ size 246144867
tokenizer/tokenizer_config.json CHANGED
@@ -19,7 +19,7 @@
19
  },
20
  "errors": "replace",
21
  "model_max_length": 77,
22
- "name_or_path": "openai/clip-vit-large-patch14",
23
  "pad_token": "<|endoftext|>",
24
  "special_tokens_map_file": "./special_tokens_map.json",
25
  "tokenizer_class": "CLIPTokenizer",
 
19
  },
20
  "errors": "replace",
21
  "model_max_length": 77,
22
+ "name_or_path": "stable-diffusion-v1-4/tokenizer",
23
  "pad_token": "<|endoftext|>",
24
  "special_tokens_map_file": "./special_tokens_map.json",
25
  "tokenizer_class": "CLIPTokenizer",
unet/config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
- "_diffusers_version": "0.2.2",
 
4
  "act_fn": "silu",
5
  "attention_head_dim": 8,
6
  "block_out_channels": [
@@ -10,6 +11,7 @@
10
  1280
11
  ],
12
  "center_input_sample": false,
 
13
  "cross_attention_dim": 768,
14
  "down_block_types": [
15
  "CrossAttnDownBlock2D",
@@ -18,19 +20,26 @@
18
  "DownBlock2D"
19
  ],
20
  "downsample_padding": 1,
 
21
  "flip_sin_to_cos": true,
22
  "freq_shift": 0,
23
  "in_channels": 4,
24
  "layers_per_block": 2,
25
  "mid_block_scale_factor": 1,
 
26
  "norm_eps": 1e-05,
27
  "norm_num_groups": 32,
 
 
28
  "out_channels": 4,
 
29
  "sample_size": 64,
30
  "up_block_types": [
31
  "UpBlock2D",
32
  "CrossAttnUpBlock2D",
33
  "CrossAttnUpBlock2D",
34
  "CrossAttnUpBlock2D"
35
- ]
 
 
36
  }
 
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.12.0.dev0",
4
+ "_name_or_path": "stable-diffusion-v1-4/unet",
5
  "act_fn": "silu",
6
  "attention_head_dim": 8,
7
  "block_out_channels": [
 
11
  1280
12
  ],
13
  "center_input_sample": false,
14
+ "class_embed_type": null,
15
  "cross_attention_dim": 768,
16
  "down_block_types": [
17
  "CrossAttnDownBlock2D",
 
20
  "DownBlock2D"
21
  ],
22
  "downsample_padding": 1,
23
+ "dual_cross_attention": false,
24
  "flip_sin_to_cos": true,
25
  "freq_shift": 0,
26
  "in_channels": 4,
27
  "layers_per_block": 2,
28
  "mid_block_scale_factor": 1,
29
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
30
  "norm_eps": 1e-05,
31
  "norm_num_groups": 32,
32
+ "num_class_embeds": null,
33
+ "only_cross_attention": false,
34
  "out_channels": 4,
35
+ "resnet_time_scale_shift": "default",
36
  "sample_size": 64,
37
  "up_block_types": [
38
  "UpBlock2D",
39
  "CrossAttnUpBlock2D",
40
  "CrossAttnUpBlock2D",
41
  "CrossAttnUpBlock2D"
42
+ ],
43
+ "upcast_attention": false,
44
+ "use_linear_projection": false
45
  }
unet/diffusion_pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:62d48b4d841a3178511fa453df0dae59b22089ace64609cc9d5353d0a7f37c26
3
- size 3438354725
 
 
 
 
unet/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:145a07e0f05ec5bbe6e2e9faf608bdb311caf708895cac8c8ed713c59864e1e8
3
- size 3438167534
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18518e7407ace29645a10ea0d1749ee299b28d1fca584336abbcd16abeba744e
3
+ size 1719125271
vae/config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "_class_name": "AutoencoderKL",
3
- "_diffusers_version": "0.2.2",
 
4
  "act_fn": "silu",
5
  "block_out_channels": [
6
  128,
@@ -17,6 +18,7 @@
17
  "in_channels": 3,
18
  "latent_channels": 4,
19
  "layers_per_block": 2,
 
20
  "out_channels": 3,
21
  "sample_size": 512,
22
  "up_block_types": [
 
1
  {
2
  "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.12.0.dev0",
4
+ "_name_or_path": "stable-diffusion-v1-4/vae",
5
  "act_fn": "silu",
6
  "block_out_channels": [
7
  128,
 
18
  "in_channels": 3,
19
  "latent_channels": 4,
20
  "layers_per_block": 2,
21
+ "norm_num_groups": 32,
22
  "out_channels": 3,
23
  "sample_size": 512,
24
  "up_block_types": [
vae/diffusion_pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b134cded8eb78b184aefb8805b6b572f36fa77b255c483665dda931fa0130c5
3
- size 334707217
 
 
 
 
vae/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a2b5134f4dbc140d9c11f11cba3233099e00af40f262f136c691fb7d38d2194c
3
- size 334643276
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11e897bda3163f66848a4abe4613001d27502c985c84d33eb10e7387f9010a2c
3
+ size 167335315